Content
summary Summary

French AI startup Mistral AI has announced two new language models designed for use on edge devices and in edge computing scenarios.

Ad

The new models, called Ministral 3B and Ministral 8B, are part of the "Ministraux" family. According to Mistral, these are currently the most powerful AI systems in their class for edge use cases. Both models support context lengths of up to 128,000 tokens.

Applications range from translation to robotics

The Ministraux models are designed for use cases where local processing and privacy are critical. Mistral says they are well-suited for tasks such as on-device translation, offline intelligent assistants, local data analysis, and autonomous robotics.

When combined with larger language models such as Mistral Large, the company says Ministraux can also act as efficient intermediaries for function calls in multi-step workflows.

Ad
Ad
Table compares the performance of 6 AI models in 11 categories. Ministral 8B leads in most areas.
Ministral 8B shows the best results overall, especially in the areas of knowledge, common-sense and multilingual skills. | Image: Mistral

Benchmarks provided by Mistral indicate that Ministral 3B and 8B outperform comparable models like Google's Gemma 2 2B and Meta's Llama 3.1 8B in most categories. The company highlights the performance of the smaller Ministral 3B, which surpasses its larger predecessor, Mistral 7B, on some benchmarks. Mistral 7B is widely considered one of the most successful open-source models.

The larger Ministral 8B is clearly outperforming the 7B model across all benchmarks. Mistral AI reports that Ministral 8B excels particularly in areas such as knowledge, common-sense, function-calling, and multilingual capabilities. It didn't compare it to the newer Llama 3.2 11B, which is likely a bit better (73 in MMLU), but outside the 10-billion-parameter class.

Pricing and availability

The new models are available now. Mistral AI offers Ministral 8B via API for $0.10 per million tokens, while Ministral 3B costs $0.04 per million tokens. Commercial licenses are available for on-premises use.

For research purposes, Mistral AI provides the model weights for Ministral 8B Instruct. The company says both models will soon be available through cloud partners such as Google Vertex and AWS.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • French AI company Mistral AI has introduced two new language models called Ministral 3B and 8B. These models are specifically optimized for use on edge devices and in edge computing scenarios, supporting context lengths of up to 128,000 tokens.
  • According to the company, these "Ministraux" models are particularly well-suited for applications where local processing and privacy are critical, such as on-device translation, offline intelligent assistants, local data analytics, and autonomous robotics.
  • Benchmarks provided by Mistral show that both the Ministral 3B and 8B outperform comparable models in many categories. The new models are now available via API, with pricing ranging from $0.04 to $0.10 per million tokens. For research purposes, Mistral AI also provides the model weights for Ministral 8B Instruct.
Sources
Online journalist Matthias is the co-founder and publisher of THE DECODER. He believes that artificial intelligence will fundamentally change the relationship between humans and computers.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.