×
Mistral AI launches small, local and open-source alternative to GPT-4o mini
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Mistral AI has released Small 3, a 24B-parameter open-source language model designed to run locally while delivering performance comparable to larger proprietary models.

Key features and capabilities; Small 3 represents a significant advancement in efficient, locally-deployable language models that can operate with minimal computing resources.

  • The model can run on a MacBook with 32GB RAM, making it accessible for individual developers and small organizations
  • Built with fewer layers than comparable models to optimize for speed and latency
  • Achieved over 81% accuracy on the MMLU benchmark test without using reinforcement learning or synthetic data
  • Released under the Apache 2.0 license, allowing for broad commercial and research applications

Performance benchmarks; Independent testing reveals competitive performance against both larger open-source models and proprietary alternatives.

  • Human evaluators preferred Small 3 over Gemma-2 27B and Qwen-2.5 32B in coding and general knowledge tasks
  • Results were more evenly split when compared to Llama-3.3 70B and GPT-4o mini
  • The model shows particular strength in scenarios requiring quick, accurate responses

Practical applications; Mistral AI positions Small 3 as particularly suitable for specific industry use cases requiring rapid response times and local deployment.

  • Recommended for building customer-facing virtual assistants
  • Suitable for time-sensitive applications like fraud detection in financial services
  • Applicable for legal advice and healthcare contexts where quick responses are crucial
  • Valuable for robotics and manufacturing applications
  • Ideal for organizations handling sensitive data that requires local processing

Technical distinctions; The model’s architecture and training approach set it apart from other language models in the market.

  • Developed without reinforcement learning or synthetic data, placing it earlier in the model production pipeline than competitors like DeepSeek R1
  • Optimized layer structure contributes to improved latency and processing speed
  • Can be fine-tuned to create specialized subject matter experts for specific domains

Future developments; The launch of Small 3 appears to be part of a broader strategy at Mistral AI.

  • The company has indicated plans to release additional models of varying sizes
  • Future releases will focus on enhanced reasoning capabilities
  • Expected rollout will occur over the coming weeks

Market implications; Small 3’s release challenges the notion that larger models are always better, potentially shifting industry focus toward efficiency and accessibility.

Mistral AI says its Small 3 model is a local, open-source alternative to GPT-4o mini

Recent News

Anthropic secures $3.5 billion at $61.5 billion valuation amid AI funding surge

Safety-focused AI startup reaches billion-dollar revenue milestone as enterprise adoption of its chatbot Claude drives unprecedented growth.

AI agents could make the internet go dark, top analysts warn

AI assistants acting as digital middlemen could drastically reduce direct human visits to websites and apps, upending current online business models.

OpenAI launches GPT-4.5 with groundbreaking new capabilities, comprehension level-up

Building on recent advancements in unsupervised learning, the model cuts AI hallucination rates nearly in half while improving pattern recognition and creative insight generation.