Skip to main content

Apple is tackling one of the most frustrating aspects with AI today

Apple Intelligence on AI
Apple

As companies like Google, Anthropic, and OpenAI update and upgrade their AI models, the way that those LLMs interact with users is sure to change as well. However, getting used to the new system can become a hassle for users who then have to adjust how they pose their queries in order to get the results they’ve come to expect. An Apple research team has developed a new method to streamline that upgrade transition while reducing inconsistencies between the two versions by as much as 40%.

As part of their study, “MUSCLE: A Model Update Strategy for Compatible LLM Evolution,” published July 15, the researchers argue that when upgrading their models, developers tend to focus more on upping the overall performance, rather than making sure that the transition between models is seamless for the user. That includes making sure that negative flips, wherein the new model predicts the incorrect output for a test sample that was correctly predicted by the older model, are kept to a minimum.

Recommended Videos

This is because, the study authors argue, each user has their own quirks, quibbles, and personalized ways of interacting with chatbots. Having to continually adjust and adapt the manner in which they interact with a model can become an exhausting affair — one that is antithetical to Apple’s desired user experience.

The research team even argues that incorrect predictions by the AI should remain between versions, “There is value in being consistent when both models are incorrect,” they wrote. “A user may have developed coping strategies on how to interact with a model when it is incorrect.”

Apple presents MUSCLE

A Model Update Strategy for Compatible LLM Evolution

Large Language Models (LLMs) are frequently updated due to data or architecture changes to improve their performance. When updating models, developers often focus on increasing overall performance… pic.twitter.com/ATm2zM4Poc

— AK (@_akhaliq) July 15, 2024

To address this, the researchers first developed metrics by which to measure the degree of regression between models and then developed a strategy to minimize their occurrence. The result is MUSCLE, a strategy that doesn’t require developers to retrain the entire base model and instead relies on the use of training adapters. Adapters small AI modules that can integrate at different points along the overall LLM.

Developers can then fine-tune these specific modules instead of the entire model. This enables the model as a whole to perform distinct tasks at a fraction of the training cost and with only a small increase in the number of parameters. They’re essentially plug-ins for large language models that allow us to fine-tune specific sections of the overall AI instead of the whole thing.

The research team upgraded LLMs including Meta’s Llama and Microsoft’s Phi as part of their study, using specific math queries as samples, and found that negative flips occurred as much as 60% of the time. By incorporating the MUSCLE strategy, the team wasn’t able to fully eliminate negative flips, but they did manage to reduce their occurrence by as much as 40% compared to the control.

Andrew Tarantola
Former Digital Trends Contributor
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
10 years ago today, Apple launched a revolutionary MacBook that failed spectacularly
An Apple 12-inch MacBook on a desk.

Ten years ago today, Apple unveiled the 12-inch MacBook to the world, claiming it had “reinvented the notebook” for the better. The laptop almost instantly divided opinion, with fans and detractors at each other’s throats from the start. And sure, it was by no means perfect, but look a little closer and I think you’ll find a device that has had a monumental impact on the world of computing -- not just on Apple, but on the industry as a whole.

The 12-inch MacBook is often seen as a flop and as a product emblematic of the excesses of Apple’s Jony Ive era, where the design guru’s penchant for thinness and lightness ruled all. The fact that this MacBook was discontinued after just four years is seen as proof of this idea.

Read more
Apple Intelligence could solve my App Store pet peeve, but I’m skeptical
The app store open on a MacBook Pro.

It’s no secret that Apple’s App Store has its problems, but it generally works pretty well. Yet there’s one thing about it that just feels absolutely useless: the reviews section.

Apple highlights a few reviews on each app’s page, but infuriatingly, they’re often from many years ago. It’s not uncommon to see reviews complaining about issues that have long-since been fixed, yet they still get highlighted. When your initial impression is based on completely inappropriate information, it makes the review section borderline useless and is a terrible way to sum up information about an app.

Read more
Apple’s new Mac Studio breaks records in more ways than one
2025 Mac Studio

Apple has announced a new Mac Studio to add to the company’s range of high-end Macs. In a surprising twist, the new Mac Studio will come in two flavors, one containing the M4 Max and one containing a previously unreleased chip: The M3 Ultra. It's safe to say that this Mac Studio will break several records, all thanks to the staggering amounts of unified memory and storage that it's equipped with -- but also largely thanks to the eye-watering price.

The M3 Ultra is a new addition to the lineup. Apple says that it's built for AI and promises to deliver up to 2.5x the performance of the M1 Ultra, although there's no mention of how much faster it'll be than the M2 Ultra. M3 Ultra, and thus the Mac Studio, will offer support for up to 512GB of unified memory, which, as Apple says, is "the most ever in a personal computer."

Read more