×
Mistral AI launches small, local and open-source alternative to GPT-4o mini
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Mistral AI has released Small 3, a 24B-parameter open-source language model designed to run locally while delivering performance comparable to larger proprietary models.

Key features and capabilities; Small 3 represents a significant advancement in efficient, locally-deployable language models that can operate with minimal computing resources.

  • The model can run on a MacBook with 32GB RAM, making it accessible for individual developers and small organizations
  • Built with fewer layers than comparable models to optimize for speed and latency
  • Achieved over 81% accuracy on the MMLU benchmark test without using reinforcement learning or synthetic data
  • Released under the Apache 2.0 license, allowing for broad commercial and research applications

Performance benchmarks; Independent testing reveals competitive performance against both larger open-source models and proprietary alternatives.

  • Human evaluators preferred Small 3 over Gemma-2 27B and Qwen-2.5 32B in coding and general knowledge tasks
  • Results were more evenly split when compared to Llama-3.3 70B and GPT-4o mini
  • The model shows particular strength in scenarios requiring quick, accurate responses

Practical applications; Mistral AI positions Small 3 as particularly suitable for specific industry use cases requiring rapid response times and local deployment.

  • Recommended for building customer-facing virtual assistants
  • Suitable for time-sensitive applications like fraud detection in financial services
  • Applicable for legal advice and healthcare contexts where quick responses are crucial
  • Valuable for robotics and manufacturing applications
  • Ideal for organizations handling sensitive data that requires local processing

Technical distinctions; The model’s architecture and training approach set it apart from other language models in the market.

  • Developed without reinforcement learning or synthetic data, placing it earlier in the model production pipeline than competitors like DeepSeek R1
  • Optimized layer structure contributes to improved latency and processing speed
  • Can be fine-tuned to create specialized subject matter experts for specific domains

Future developments; The launch of Small 3 appears to be part of a broader strategy at Mistral AI.

  • The company has indicated plans to release additional models of varying sizes
  • Future releases will focus on enhanced reasoning capabilities
  • Expected rollout will occur over the coming weeks

Market implications; Small 3’s release challenges the notion that larger models are always better, potentially shifting industry focus toward efficiency and accessibility.

Mistral AI says its Small 3 model is a local, open-source alternative to GPT-4o mini

Recent News

Say what?! Sesame AI’s hyper-realistic voice assistants garner over $1 billion valuation

Investors flock to startup's AI voice assistants that aim to outperform common virtual assistants like Siri and Alexa with more natural interactions.

Danone pivots marketing to target weight loss medication users, food manufacturers put on blast

As major pharmaceuticals reshape dietary needs, Danone revamps its product positioning to serve the protein-hungry market of GLP-1 medication users.

AI blurs boundaries in advertising skills as generalists master copywriting, graphic design and more

AI tools allow non-specialists to produce passable creative content across disciplines, raising questions about quality standards and the future value of specialized expertise in advertising.