×
Mistral AI launches small, local and open-source alternative to GPT-4o mini
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Mistral AI has released Small 3, a 24B-parameter open-source language model designed to run locally while delivering performance comparable to larger proprietary models.

Key features and capabilities; Small 3 represents a significant advancement in efficient, locally-deployable language models that can operate with minimal computing resources.

  • The model can run on a MacBook with 32GB RAM, making it accessible for individual developers and small organizations
  • Built with fewer layers than comparable models to optimize for speed and latency
  • Achieved over 81% accuracy on the MMLU benchmark test without using reinforcement learning or synthetic data
  • Released under the Apache 2.0 license, allowing for broad commercial and research applications

Performance benchmarks; Independent testing reveals competitive performance against both larger open-source models and proprietary alternatives.

  • Human evaluators preferred Small 3 over Gemma-2 27B and Qwen-2.5 32B in coding and general knowledge tasks
  • Results were more evenly split when compared to Llama-3.3 70B and GPT-4o mini
  • The model shows particular strength in scenarios requiring quick, accurate responses

Practical applications; Mistral AI positions Small 3 as particularly suitable for specific industry use cases requiring rapid response times and local deployment.

  • Recommended for building customer-facing virtual assistants
  • Suitable for time-sensitive applications like fraud detection in financial services
  • Applicable for legal advice and healthcare contexts where quick responses are crucial
  • Valuable for robotics and manufacturing applications
  • Ideal for organizations handling sensitive data that requires local processing

Technical distinctions; The model’s architecture and training approach set it apart from other language models in the market.

  • Developed without reinforcement learning or synthetic data, placing it earlier in the model production pipeline than competitors like DeepSeek R1
  • Optimized layer structure contributes to improved latency and processing speed
  • Can be fine-tuned to create specialized subject matter experts for specific domains

Future developments; The launch of Small 3 appears to be part of a broader strategy at Mistral AI.

  • The company has indicated plans to release additional models of varying sizes
  • Future releases will focus on enhanced reasoning capabilities
  • Expected rollout will occur over the coming weeks

Market implications; Small 3’s release challenges the notion that larger models are always better, potentially shifting industry focus toward efficiency and accessibility.

Mistral AI says its Small 3 model is a local, open-source alternative to GPT-4o mini

Recent News

South Korea’s webcomics industry blends tradition with AI-driven innovation

AI tools help legendary Korean cartoonists extend their creative legacies while preserving artistic integrity in the booming global webcomics market.

Generative AI evolves from automation tool to emotional assistant

Personal assistants powered by generative AI are increasingly being used for emotional support and life coaching rather than just productivity tasks.

How AI perpetuates misinformation through “digital fossils” in scientific literature

Scanning and digitization errors created the nonsensical term "vegetative electron microscopy," which has spread to 22 papers and illustrates how AI systems perpetuate misinformation in scientific literature.