Content
summary Summary

Alibaba has developed a new language model called Qwen2.5-Max that uses what the company says is a record-breaking amount of training data - over 20 trillion tokens.

Ad

Alibaba has unveiled Qwen2.5-Max, a new AI language model trained on what the company claims is a record-breaking 20 trillion tokens of data. The model joins Alibaba's existing Qwen2.5 family, which includes Qwen2.5-VL and Qwen2.5-1M.

Built using a mixture-of-experts (MoE) architecture, Qwen2.5-Max goes head-to-head with and beats some leading AI models like Deepseek-V3, GPT-4o, Claude 3.5 Sonnet, and Llama-3.1-405B in benchmark tests. While the exact training data size of some commercial competitors remains private, Deepseek-V3 and Llama-3.1-405B used approximately 15 trillion tokens each.

Säulendiagramm: Leistungsvergleich von 5 KI-Modellen (Qwen, DeepSeek, Llama, GPT4, Claude) anhand von 5 Benchmark-Tests mit Prozentangaben.
In a comparison of various large language models, Qwen2.5-Max shows top performance in the important Arena-Hard and LiveBench benchmarks, but is at least on a par in others. | Picture: Qwen

The model shows particularly strong results in the Arena-Hard and LiveBench benchmarks, while matching competitors in other tests. Alibaba's team used established training methods including supervised fine-tuning and reinforcement learning from human feedback to develop the model.

Ad
Ad

Qwen2.5-Max available via API and in Qwen Chat

Users can now access Qwen2.5-Max through Alibaba Cloud's API or test it in Qwen Chat, the company's chatbot that offers features like web search and content generation. While Alibaba hasn't disclosed its data sources, experts suggest synthetic data - text generated by other AI models - likely plays a significant role.

Despite the massive investment in training data, the model's performance lead over competitors remains modest. This aligns with recent discussions in the AI community suggesting that improvements in test-time computing power, rather than training data size alone, may be key to advancing language model capabilities.

Unlike other models in the Qwen2.5 family, the Max version will stay API-only and won't be released as open source. Alibaba is targeting developers with competitive pricing and an OpenAI-compatible interface to encourage migration to their cloud platform. However, like other Chinese language models, Qwen2.5-Max operates under Chinese government content restrictions.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Alibaba has released Qwen2.5-Max, a new language model trained on over 20 trillion tokens of data, which the company claims is a record-breaking amount. The model uses a mixture-of-experts (MoE) architecture.
  • In some benchmark tests, Qwen2.5-Max outperforms leading AI models such as Deepseek-V3, GPT-4o, Claude 3.5 Sonnet, and Llama-3.1-405B.
  • Users can access Qwen2.5-Max through Alibaba Cloud's API or test it in the Qwen Chat chatbot.
Sources
Jonathan works as a freelance tech journalist for THE DECODER, focusing on AI tools and how GenAI can be used in everyday work.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.