Content
summary Summary

Tencent has released its new language model Hunyuan-A13B as open source, introducing a dynamic approach to reasoning that lets the model switch between fast and slow "thinking" depending on the complexity of the task.

Ad

Hunyuan-A13B's key feature is its ability to adapt its depth of reasoning in real time. For simple queries, it uses a fast mode that delivers quick responses with minimal inference.

For more complex prompts, it activates a deeper reasoning process involving multistep thought. Users can toggle this behavior with special commands: "/think" enables the in-depth mode, while "/no_think" turns it off.

The model uses a Mixture of Experts (MoE) architecture with 80 billion total parameters, but only 13 billion are active during inference. It supports context windows of up to 256,000 tokens.

Ad
Ad

A focus on scientific reasoning

According to Tencent's technical report, Hunyuan-A13B was trained on 20 trillion tokens, then tuned for reasoning tasks and refined for broader use cases. Tencent assembled 250 billion tokens from STEM fields to boost the model's reliability in scientific tasks.

The training data includes math textbooks, exams, open-source code from GitHub, collections of logic puzzles, and scientific texts ranging from middle school to university level.

Video: Tencent

Tencent claims Hunyuan-A13B-Instruct can hold its own against leading models from OpenAI, Deepseek, and Alibaba's Qwen. In the 2024 American math competition AIME, the model reportedly achieves 87.3 percent accuracy, compared to OpenAI o1's 74.3 percent.

But a look at results from the 2025 edition shows that such comparisons aren't always clear-cut - o1 leads by nearly three percent in that round. Tencent also bases its comparisons on an outdated January version of Deepseek-R1, even though the May release performed significantly better in AIME 2024 and 2025, scoring 91.4 and 87.5 points, respectively.

Recommendation
Table showing benchmark results for OpenAI-o1-1217, Deepseek-R1-0120, Qwen3-A22B, and Hunyuan-A13B across six categories.
Benchmarks show Hunyuan-A13B keeps pace with Deepseek-R1, OpenAI o1, and Qwen3-A22B, with a clear lead in agent-based tasks. | Image: Tencent

Tencent says the model excels at automated agent tasks and tool use. In internal benchmarks, A13B scored highest in nearly all agent benchmarks. Even as tests pushed up to 128,000-token contexts, Hunyuan-A13B's performance stayed above Deepseek-R1 and Qwen3-A22B, though it fell short of Gemini 2.5 Pro.

Table with RULER benchmark values for LLMs (Gemini 2.5 Pro, DeepSeek R1, Qwen3-A22B, Hunyuan-A13B) across context lengths 0–128K
Performance remains stable even with context windows up to 128,000 tokens. | Bild: Tencent

The model is available under the Apache 2.0 license on Hugging Face and GitHub, with ready-to-use Docker images for various deployment frameworks. API access is offered through Tencent Cloud and a browser demo.

Tencent has also released two new benchmark datasets: ArtifactsBench for code generation and C3-Bench for agent task evaluation.

Tencent's dynamic reasoning strategy is in line with trends seen at other leading AI labs. The concept of switching between different reasoning modes echoes recent models like Claude 3.7 Sonnet and Qwen3, which also support adaptive reasoning.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.

Tencent's shift toward language models follows earlier work on video generation. In December 2024, the company launched HunyuanVideo, followed in March by Hunyuan-T1, a reasoning-focused LLM that Tencent says already matched OpenAI o1's performance.

Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Tencent has released the Hunyuan-A13B language model as open source, featuring two reasoning modes: a fast mode for basic tasks and a deep mode for complex queries, which can be switched using commands like "/think" and "/no_think".
  • The model uses a mixture-of-experts design with 80 billion parameters in total, 13 billion of which are active during inference, and is capable of handling texts up to 256,000 tokens in length.
  • Hunyuan-A13B was trained on 20 trillion tokens, including 250 billion from STEM areas such as mathematics, programming, and science, and is available with test data sets under the Apache 2.0 license on Hugging Face and GitHub.
Sources
Jonathan writes for THE DECODER about how AI tools can make our work and creative lives better.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.