×
Thanks to Exo Labs you can run powerful open-source LLMs on a Mac
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The emergence of local AI model computing has taken a significant leap forward with Exo Labs enabling powerful open source AI models to run on Apple’s new M4-powered Mac computers.

Breaking new ground: Exo Labs has successfully demonstrated running advanced large language models (LLMs) locally on Apple M4 devices, marking a significant shift away from cloud-based AI computing.

  • A cluster of four Mac Mini M4 devices and one Macbook Pro M4 Max, totaling around $5,000, can now run sophisticated AI models like Qwen 2.5 Coder-32B
  • This setup provides a cost-effective alternative to traditional GPU solutions, as a single Nvidia H100 GPU costs between $25,000-$30,000
  • The cluster achieves impressive performance metrics, processing 18 tokens per second with Qwen 2.5 Coder 32B and 8 tokens per second with Nemotron-70B

Technical innovation and accessibility: Exo Labs’ open-source software enables distributed AI computing across multiple devices, democratizing access to powerful AI capabilities.

  • The software distributes AI workloads across connected devices, allowing users to run models without requiring expensive specialized hardware
  • Anyone with coding experience can access and implement the solution through Exo’s Github repository
  • The system works particularly well with Apple’s M4 chip, which offers what Apple calls “the world’s fastest GPU core” and superior single-threaded performance

Privacy and control benefits: Local AI computing offers significant advantages over web-based solutions in terms of data security and user autonomy.

  • Users can process sensitive information locally without exposing data to cloud services
  • The system allows for complete control over AI model behavior and data handling
  • Enterprises in regulated industries can leverage powerful AI capabilities while maintaining data sovereignty

Future developments: Exo Labs is positioning itself for broader adoption and enterprise implementation.

  • A benchmarking website launch is planned to help users compare different hardware configurations
  • The company is developing enterprise-grade software offerings and providing bespoke services
  • Exo Labs has secured private funding to support its expansion and development efforts

Market implications: The ability to run powerful AI models locally on relatively affordable hardware could reshape the AI computing landscape and challenge the current cloud-centric paradigm.

You can now run the most powerful open source AI models locally on Mac M4 computers, thanks to Exo Labs

Recent News

China-based DeepSeek just released a very powerful ultra large AI model

Chinese startup achieves comparable performance to GPT-4 while cutting typical training costs by 99% through an innovative parameter activation approach.

7 practical tips and tools for using AI to improve your relationships

AI tools offer relationship support through structured communication guidance and conflict management, but experts emphasize they should complement rather than replace human interaction.

How AI-powered tsunami prediction will save lives in future disasters

Emergency response teams are leveraging AI systems to cut tsunami warning times from hours to minutes while improving evacuation planning and damage assessment.