back
Get SIGNAL/NOISE in your inbox daily

Apple’s new open-source language models showcase the company’s AI prowess, with the 7B model outperforming leading open models and the 1.4B version surpassing competitors in its category.

Introducing DCLM models: Apple’s research team, as part of the DataComp for Language Models project, released a family of open DCLM models on Hugging Face, including a 7 billion parameter model and a 1.4 billion parameter model:

  • The models were trained using a high-quality dataset, DCLM-Baseline, assembled through model-based filtering, demonstrating the effectiveness of this data curation technique.
  • The project is truly open-source, with the release of model weights, training code, and the pretraining dataset, fostering transparency and collaboration in AI research.

Impressive performance of the 7B model: The larger DCLM model, trained on 2.5 trillion tokens, delivers competitive results on benchmarks compared to leading open models:

  • It achieves 63.7% 5-shot accuracy on MMLU, a 6.6 percentage point improvement over the previous state-of-the-art open-data language model, MAP-Neo, while using 40% less compute for training.
  • The model’s performance is close to that of Mistral-7B-v0.3, Llama3 8B, Google’s Gemma, and Microsoft’s Phi-3, despite these models using closed data.

Smaller model outshines competitors: The 1.4B version of the DCLM model, trained jointly with Toyota Research Institute on 2.6 trillion tokens, also showcases impressive performance:

  • It scores 41.9% on the 5-shot MMLU test, considerably higher than other models in its category, such as Hugging Face’s SmolLM, Qwen-1.5B, and Phi-1.5B.
  • The smaller model is released under the Apache 2.0 license, allowing for commercial use, distribution, and modification.

Broader implications: The release of Apple’s DCLM models highlights the importance of dataset design in training high-performing language models and serves as a starting point for further research on data curation. However, it is essential to note that these models are early research and may exhibit biases or produce harmful responses. As Apple continues to expand its AI offerings, the company’s commitment to open-source development and collaborative research could drive significant advancements in the field, while also raising questions about the potential impact on user privacy and the competitive landscape of AI technology.

Recent Stories

Oct 17, 2025

DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment

The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...

Oct 17, 2025

Tying it all together: Credo’s purple cables power the $4B AI data center boom

Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...

Oct 17, 2025

Vatican launches Latin American AI network for human development

The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...