Ad
Skip to content

New Artificial Analysis benchmark shows OpenAI, Anthropic, and Google locked in a three-way tie at the top

Artificial Analysis just released version 4.0 of its Intelligence Index, ranking AI models across multiple benchmarks. OpenAI's GPT-5.2 at its highest reasoning setting takes the top spot, with Anthropic's Claude Opus 4.5 and Google's Gemini 3 Pro close behind.

The index scores models across four equally weighted categories: Agents, Programming, Scientific Reasoning, and General. Results are less saturated this time, with top models peaking at 50 points compared to 73 in the previous version.

Artificial Analysis Intelligence Index v4.0: GPT-5.2 (xhigh) leads with 50 points, followed by Claude Opus 4.5 (49) and Gemini 3 Pro Preview (48). It's a tight race at the top. | Image: Artificial Analysis
At the top of the cost table is GPT-5.2 (xhigh) with a total cost of $2,322, followed by Grok 4 ($1,574) and Claude 4.5 Opus ($1,510). Gemini 3 Pro Preview trails significantly behind at $988. | Image: Artificial Analysis

The updated index swaps three older tests (AIME 2025, LiveCodeBench, and MMLU-Pro) for a fresh set: AA-Omniscience checks model knowledge across 40 topics while flagging hallucinations, GDPval-AA tests models on practical tasks across 44 professions, and CritPt tackles physics research problems. Artificial Analysis says it ran all tests independently using a standardized approach, with full details available on its website.

Ad
DEC_D_Incontent-1

AI News Without the Hype – Curated by Humans

As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.