×
Written by
Published on
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Apple’s AI strategy aims to improve language model consistency and user experience:

Key takeaways: Apple researchers have developed techniques to reduce inconsistencies and negative impacts on user experience when upgrading large language models (LLMs):

  • Updating LLMs can result in unexpected behavior changes and force users to adapt their prompt styles and techniques, which may be unacceptable for mainstream iOS users.
  • Apple’s method, called MUSCLE (Model Update Strategy for Compatible LLM Evolution), reduces negative flips (where a new model gives an incorrect answer while the old model was correct) by up to 40%.
  • The research highlights Apple’s preparation for updating its underlying AI models while ensuring a consistent user experience with features like Siri.

Tackling the challenges of model updates: Apple’s paper addresses the issues that arise when LLMs are frequently updated due to data or architecture changes:

  • Users develop their own system to interact with an LLM, and switching to a newer model can be a draining task that dampens their experience.
  • The researchers created metrics to compare regression and inconsistencies between model versions and developed a training strategy (MUSCLE) to minimize those inconsistencies.
  • MUSCLE does not require changes to the base model’s training and relies on training adapters, or plugins for LLMs, called compatibility adapters.

Testing and results: The research team tested their system by updating LLMs like Llama and Phi, finding significant improvements in model consistency:

  • Tests included asking updated models math questions to see if they still got the correct answer, sometimes finding negative flips of up to 60% in different tasks.
  • Using MUSCLE, the researchers managed to mitigate a significant number of those negative flips, sometimes by up to 40%.
  • The authors argue that there is value in being consistent even when both models are incorrect, as users may have developed coping strategies for incorrect responses.

Broader implications: Apple’s research could make AI chatbots and assistants more dependable and user-friendly as they continue to evolve:

  • Given the rapid pace of updates to chatbots like ChatGPT and Google’s Gemini, Apple’s techniques have the potential to ensure newer versions of these tools maintain a consistent user experience.
  • As AI models become more widely adopted, minimizing unexpected behavior changes will be crucial for mainstream acceptance and satisfaction.
  • While it’s unclear if this research will be directly applied to upcoming iOS features like Apple Intelligence, it demonstrates Apple’s commitment to delivering a stable, intuitive AI experience to its users.
Apple just revealed an AI technique to better compete against ChatGPT

Recent News

Newton AI model learns physics autonomously from raw data

The AI model learns complex physics concepts from raw sensor data, potentially transforming fields from energy management to scientific research.

Anthropic just announced a big update to Claude — here’s what’s inside

The update brings enhanced customization and cross-device functionality to Claude AI, allowing for more personalized and efficient user experiences.

Google enhances NotebookLM with customizable AI podcasts

Google's AI writing tool now allows users to create customized podcast-style discussions based on uploaded content and specific prompts.