Content
summary Summary

Mistral AI has released Small 3, a new 24-billion parameter language model that matches the performance of much larger models from Meta and OpenAI. The company is also switching to the more permissive Apache 2.0 license.

Ad

Mistral AI says Small 3 achieves similar performance to models three times its size while using just 24 billion parameters. The model, optimized for low latency and local deployment, follows Mistral's previous release from September 2024.

In benchmarks, Small 3 performed on par with significantly larger models including Meta's Llama-3.3-70B, Qwen2.5-32B, and GPT-4o-mini. The model achieved an 81 percent accuracy score on the MMLU benchmark while processing 150 tokens per second, making it one of the most efficient models in its category, according to Mistral.

Scatter plot: Performance/MMLU-Pro vs. latency for various LLM models with Mistral Small 3, GPT-4o Mini, Gemma-2 27B and Qwen-2.5 32B.
Mistral Small 3 is optimized for quality and latency. | Image: Mistral

Optimized for real-world applications at low cost

The new model is built to handle everyday tasks without requiring expensive hardware. It can manage quick chat responses, run specific functions, adapt to specialized fields, and work on a single GPU - making it practical for real-world use.

Ad
Ad

Several industries are already trying out the technology, Mistral says. Banks are looking at it for spotting fraud, while healthcare providers and manufacturers are testing it for customer service and analyzing customer feedback. Robotics companies are also exploring potential uses.

Bar chart: Performance comparison of 5 AI models in MMLU Pro and GPQA main tests, accuracy values between 0.3 and 0.7.
Mistral-Small-24B shows strong performance in the multimodal understanding MMLU benchmark. | Image: Mistral
Bar chart: Comparison of 5 AI models based on their accuracy in HumanEval and Math Instruct benchmarks.
In benchmarks designed for coding and math tasks, Mistral Small performs similarly to much larger models. | Image: Mistral
Bar chart: Comparison of 5 AI language models using 4 benchmark tests (Wildbench, Arena Hard, MTBench, IFEval), accuracy in percent.
These benchmarks test, among other things, how well the models follow instructions and retain information in their "memory". | Image: Mistral

The model is now available through Mistral's platform and partners including Hugging Face, Ollama, Kaggle, Together AI, and Fireworks AI, with more platforms planned.

Apache license opens the door for commercial use

With this release, Mistral is moving away from its proprietary MRL (Mistral Research License) to the more permissive Apache 2.0 license. This change allows users to freely use, modify, and redistribute the models, even for commercial purposes. The company will continue offering specialized commercial models for specific needs.

Mistral plans to release additional models in the coming weeks, focusing on improved reasoning capabilities. The company positions Small 3 as a complement to larger open-source reasoning models like those from Deepseek, trying to match similar capabilities while using less computing power.

Mistral has carved out its own space in the European AI landscape and is more or less the only relevant AI model provider. While its models don't yet match the capabilities of multimodal systems like Sonnet 3.5, the company continues to build momentum. Its latest offering, Pixtral, marks the company's first step into vision-language models. Anyone can test these developments through Le Chat, Mistral's public chatbot.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Recommendation
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • French AI startup Mistral AI has released Small 3, a language model with 24 billion parameters that performs similarly to larger models from Meta, Qwen, and OpenAI in benchmarks.
  • Mistral sees potential applications for Small 3 in areas such as fast chat responses, function calls, domain-specific fine-tuning, and local inference on individual GPUs.
  • With the release of Small 3, Mistral is switching its general language models from the proprietary MRL license to the open-source Apache 2.0 license, allowing free use, modification, and redistribution for commercial purposes. The company plans to release additional small and large models with improved reasoning capabilities in the coming weeks.
Sources
Jonathan works as a freelance tech journalist for THE DECODER, focusing on AI tools and how GenAI can be used in everyday work.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.