Content
summary Summary

Anthropic has released its next generation of AI models, Claude Opus 4 and Claude Sonnet 4, and is introducing new safety measures designed to prevent their use in developing chemical, biological, radiological, or nuclear (CBRN) weapons.

Ad

Claude Opus 4 and Claude Sonnet 4 both aim to expand what’s possible for software developers and agent-based applications. According to Anthropic, the two models are better at handling longer reasoning chains, can use tools like web search in parallel, and remember document access with an expanded memory system.

Opus 4 is Anthropic’s most advanced model to date, with a particular focus on coding and complex, multi-hour workflows. It tops benchmarks like SWE-bench (72.5%) and Terminal-bench (43.2%), and also scores highly on MMMLU (87.4%) and GPQA Diamond (74.9%). The "Extended Thinking" feature gives it a performance boost on certain tasks, and Anthropic says Opus 4 is designed for sustainable performance over many steps. When accessing local files, Opus 4 can create "Memory Files" that help it keep track of information—useful in scenarios like navigating complex game worlds such as Pokémon.

Image: Anthropic

Sonnet 4: The all-purpose model for developers

Claude Sonnet 4 is an upgraded version of Sonnet 3.7 built for everyday but demanding developer tasks. It scores 72.7% on SWE-bench and, according to partners like GitHub and Sourcegraph, shows significant improvements in problem-solving, code navigation, and handling complex instructions. GitHub plans to use Sonnet 4 as the foundation for its new Copilot agent.

Ad
Ad

Anthropic says both new models are 65% less likely than Sonnet 3.7 to take shortcuts or exploit loopholes in agent-based tasks. For especially long reasoning chains, they use "Thinking Summaries" to condense steps—a feature Anthropic says is only needed about 5% of the time.

New tools for building agents

Alongside the new models, Anthropic is rolling out updated API features for building more efficient agents. There’s a new code execution tool that can run Python in an isolated environment—including data analysis and visualization in a single step. Agents can now also connect to external systems like Asana or Zapier via the new MCP Connector, removing the need for custom integrations.

The new Files API lets users upload a document once and reference it across multiple sessions. Combined with the code execution tool, Claude can analyze these files and return results such as diagrams directly. For long-running sessions, prompt caching is now available for up to an hour—a twelvefold improvement over the previous five-minute cache.

Claude Code now generally available

Claude Code, previously in testing, is now available to everyone. Developers can integrate it directly into IDEs like VS Code or JetBrains, where Claude suggests changes inline. The Claude Code agent can also be used with GitHub pull requests to implement feedback, fix CI errors, or adjust code. An SDK is available for teams building their own agents.

ASL-3 safety standard activated for the first time

With Claude Opus 4, Anthropic is activating its AI Safety Level 3 (ASL-3) standard from the Responsible Scaling Policy for the first time. This is a precautionary move, as the model shows advanced knowledge of CBRN risks. Claude Sonnet 4, by contrast, does not fall under ASL-3.

Recommendation

ASL-3 has two main components: preventing misuse and protecting model weights from theft. To block misuse, Anthropic uses "Constitutional Classifiers" that monitor inputs and outputs in real time to filter out dangerous CBRN-related information. A bug bounty program and synthetic training with jailbreak data are also used to strengthen security.

For protecting model weights, Anthropic has implemented over 100 security controls, such as two-person authorization, change management protocols, and egress bandwidth monitoring to prevent large data exports from going undetected.

Anthropic notes that it remains unclear whether Claude Opus 4 strictly requires ASL-3, but activating the safeguards now allows the company to test and refine them in practice.

Ad
Ad
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • Anthropic has released Claude Opus 4 and Claude Sonnet 4, its next-generation AI models, with Opus 4 being the company's most advanced model focused on coding and complex workflows, scoring highly on benchmarks like SWE-bench (72.5%) and MMMLU (87.4%).
  • Claude Sonnet 4, an upgraded version of Sonnet 3.7, is designed for everyday developer tasks with significant improvements in problem-solving and code navigation, and will serve as the foundation for GitHub's new Copilot agent.
  • For the first time, Anthropic is activating its AI Safety Level 3 (ASL-3) standard for Opus 4, implementing security measures to prevent misuse related to chemical, biological, radiological, or nuclear weapons, including "Constitutional Classifiers" that filter dangerous information in real-time.
Sources
Max is the managing editor of THE DECODER, bringing his background in philosophy to explore questions of consciousness and whether machines truly think or just pretend to.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.