Alibaba's Qwen AI unit releases capable new models to help developers write and analyze code
Key Points
- Alibaba's research unit Qwen has introduced a new series of AI models called Qwen-2.5-Coder, designed to help programmers write, analyze, and understand code, with model sizes ranging from 0.5 to 32 billion parameters.
- The models demonstrated strong performance in real-world tests, both integrated into the Cursor AI editor and as a web chatbot with artifact support, with the largest model, Qwen-2.5-Coder-32B-Instruct, outperforming available open-source systems on code generation benchmarks.
- Qwen's approach to improving the performance of its code AI models relies on scaling, with models trained on over 5.5 trillion tokens from public source code and programming-related web text. All models except the 3 billion parameter version are available under the Apache 2.0 license on GitHub.
Alibaba's AI research unit Qwen has released a new series of AI models designed specifically for software development.
Called Qwen-2.5-Coder, these models help developers write, analyze, and understand code. The new series includes six different model sizes, ranging from 0.5 to 32 billion parameters, to accommodate various use cases and computing requirements.
Qwen tested these models in two practical applications: the AI-powered code editor Cursor and a web-based chatbot with artifact support similar to ChatGPT or Claude. Alibaba plans to integrate the chatbot functionality into its Tongyi cloud platform soon.
Video: Qwen
Video: Qwen
According to Qwen, their largest model, Qwen-2.5-Coder-32B-Instruct, outperformed other open-source systems like DeepSeek-Coder and Codestral in code generation benchmarks. The model also showed strong performance in general tasks like logical reasoning and language comprehension, though GPT-4o still leads in some benchmark tests.

Massive training datasets set token record
The models were trained on more than 20 trillion tokens of data from two sources: 18.5 trillion tokens from Qwen 2.5's general data mix introduced last September, plus 5.5 trillion tokens from public source code and programming-related web content. This makes it the first open-source model to exceed 20 trillion training tokens.
The top model, Qwen-2.5-Coder-32B-Instruct, supports over 40 programming languages, from common ones like Python, Java, and JavaScript to specialized languages like Haskell and Racket. All models feature context windows of up to 128,000 tokens.

Alibaba has released all models except the three-billion-parameter version under an Apache 2.0 license on GitHub. Developers can test the models through a free demo on Hugging Face.
Qwen researchers found that scaling up both model size and data consistently produced better results across programming tasks. The company says it plans to continue scaling to larger models and improving reasoning capabilities in future releases.
AI News Without the Hype – Curated by Humans
As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.
Subscribe now