Skip to main content

Apple is tackling one of the most frustrating aspects with AI today

Apple Intelligence on AI
Apple

As companies like Google, Anthropic, and OpenAI update and upgrade their AI models, the way that those LLMs interact with users is sure to change as well. However, getting used to the new system can become a hassle for users who then have to adjust how they pose their queries in order to get the results they’ve come to expect. An Apple research team has developed a new method to streamline that upgrade transition while reducing inconsistencies between the two versions by as much as 40%.

As part of their study, “MUSCLE: A Model Update Strategy for Compatible LLM Evolution,” published July 15, the researchers argue that when upgrading their models, developers tend to focus more on upping the overall performance, rather than making sure that the transition between models is seamless for the user. That includes making sure that negative flips, wherein the new model predicts the incorrect output for a test sample that was correctly predicted by the older model, are kept to a minimum.

This is because, the study authors argue, each user has their own quirks, quibbles, and personalized ways of interacting with chatbots. Having to continually adjust and adapt the manner in which they interact with a model can become an exhausting affair — one that is antithetical to Apple’s desired user experience.

The research team even argues that incorrect predictions by the AI should remain between versions, “There is value in being consistent when both models are incorrect,” they wrote. “A user may have developed coping strategies on how to interact with a model when it is incorrect.”

Apple presents MUSCLE

A Model Update Strategy for Compatible LLM Evolution

Large Language Models (LLMs) are frequently updated due to data or architecture changes to improve their performance. When updating models, developers often focus on increasing overall performance… pic.twitter.com/ATm2zM4Poc

— AK (@_akhaliq) July 15, 2024

To address this, the researchers first developed metrics by which to measure the degree of regression between models and then developed a strategy to minimize their occurrence. The result is MUSCLE, a strategy that doesn’t require developers to retrain the entire base model and instead relies on the use of training adapters. Adapters small AI modules that can integrate at different points along the overall LLM.

Developers can then fine-tune these specific modules instead of the entire model. This enables the model as a whole to perform distinct tasks at a fraction of the training cost and with only a small increase in the number of parameters. They’re essentially plug-ins for large language models that allow us to fine-tune specific sections of the overall AI instead of the whole thing.

The research team upgraded LLMs including Meta’s Llama and Microsoft’s Phi as part of their study, using specific math queries as samples, and found that negative flips occurred as much as 60% of the time. By incorporating the MUSCLE strategy, the team wasn’t able to fully eliminate negative flips, but they did manage to reduce their occurrence by as much as 40% compared to the control.

Andrew Tarantola
Andrew has spent more than a decade reporting on emerging technologies ranging from robotics and machine learning to space…
Apple Intelligence goes against the entire Apple ethos — in a good way
Summarization of notification and emails on iPhone with Apple Intelligence.

ChatGPT isn't the only third-party large language model Apple hopes to incorporate into its upcoming Apple Intelligence system, the Wall Street Journal reports. Apple and Facebook's parent company, Meta -- as well as Anthropic, developers of the Claude AI -- are reportedly in talks to reach a similar deal. What's more, Reuters reports that Apple and Google have also been discussing potentially working together to bring the Gemini AI to Apple devices, as well as AI developers in China, where Google's products are banned. Granted, no official agreement has yet been reached with any of these potential partners, and talks could very well fall through before a bargain is struck.

Apple is taking an interesting approach to structuring these partnerships, in that it is not offering to pay for the integration of any of these AI models. Instead, Apple wants to leverage its massive market share and the reach of its broad portfolio of digital devices to serve as a distribution platform for its partners. Apple would be able to integrate a wide swath of models into its offerings, reducing its reliance on a single partner, while the AI developers will reportedly be able to sell premium subscriptions for their models through Apple Intelligence.

Read more
Apple’s cheaper Vision Pro may come with one big drawback
Apple Vision Pro demo unit.

One of the major issues with Apple’s Vision Pro is its price tag. Many folks aren’t willing or able to drop $3,499 on a device they’re not even sure they’re going to use regularly, while others feel more inclined to see if the tech company might follow through with a more affordable version.

Mark Gurman, a high-profile and generally reliable Apple tipster, wrote in his weekly Power On newsletter on Sunday that the Vision Pro team tasked with building a cheaper Vision Pro is struggling to reduce costs without ruining the experience of a cheaper mixed-reality headset, which Gurman suggests may launch at the end of 2025 at the earliest.

Read more
The Vision Pro just got one of its most anticipated features
A man adjusts an Apple Vision Pro headset over his eyes.

Apple's chess demo showed how to make Vision Pro games in WebXR. Apple

At the recent WWDC 2024 event, Apple previewed visionOS 2, which will arrive later this year. The developer beta version is already available, and now we know about a new way to experience VR in the Vision Pro.

Read more