Ad
Skip to content

OpenAI updates Codex model, adds trusted access program for cyber defense

Image description
Nano Banana Pro prompted by THE DECODER

The new AI model GPT-5.2-Codex is built to solve complex tasks as an autonomous software agent. Because the technology is also effective at finding vulnerabilities, OpenAI is launching an exclusive access program where verified experts get a version with relaxed security filters.

Technically, OpenAI relies on advanced context compression, or "compaction." This method helps the model process long conversation histories and extensive code analyses more efficiently. The system is designed to maintain an overview even in complex projects, building directly on the capabilities of its predecessor, GPT-5.1-Codex-Max, which was already designed to work on tasks for longer than a day.

OpenAI has also optimized image processing, allowing GPT-5.2-Codex to interpret technical diagrams or screenshots of user interfaces more precisely. According to the company, controlling native Windows environments now works more reliably than it did with the previous model.

Benchmarks show only slight gains

In standardized tests, the new model shows only slight improvements over the basic version. In SWE-Bench Pro, where software simulates solving real problems in GitHub repositories, GPT-5.2-Codex achieves a solution rate of 56.4 percent, compared to the standard version's 55.6 percent.

Zwei Balkendiagramme zeigen die Leistungssteigerung von GPT-5.2-Codex. Links: Im SWE-Bench Pro erreicht das Modell 56,4 Prozent Genauigkeit, verglichen mit 55,6 Prozent bei GPT-5.2 und 50,8 Prozent bei GPT-5.1. Rechts: Im Terminal-Bench 2.0 erzielt GPT-5.2-Codex 64,0 Prozent, GPT-5.2 erreicht 62,2 Prozent und GPT-5.1-Codex-Max 58,1 Prozent.
GPT-5.2-Codex achieves only slight gains in the SWE-Bench Pro and Terminal-Bench 2.0 benchmarks compared to the standard version GPT-5.2.

In Terminal-Bench 2.0, accuracy increases slightly more to 64 percent. This test checks how well AI agents can operate command-line tools, set up servers, or compile code.

Dual-use risks for cybersecurity

A major focus of the release is cybersecurity. The increased ability to analyze code can be used for both defense and attack, and OpenAI cites a recent incident as proof. Security researcher Andrew MacPherson reportedly used an earlier version of the model to investigate a vulnerability in the React framework.

The AI discovered unexpected behaviors that, after further analysis, led to three previously unknown vulnerabilities capable of paralyzing services or exposing source code. According to OpenAI, the discovery demonstrates how autonomous AI systems can speed up the work of security researchers.

These capabilities carry risks. OpenAI now rates the model at nearly a "high" level within its Preparedness Framework for cybersecurity. In response, the company is introducing a trusted access program.

Aimed at certified security experts and organizations, the program gives participants access to models that are less restrictive than the public version. This allows experts to search for security vulnerabilities without being blocked by the AI's standard protection filters.

GPT-5.2-Codex is available now to paying ChatGPT users. Integration is handled via the command line, development environments, and the cloud, with an interface for third-party providers coming soon.

AI News Without the Hype – Curated by Humans

As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.

AI news without the hype
Curated by humans.

  • Over 20 percent launch discount.
  • Read without distractions – no Google ads.
  • Access to comments and community discussions.
  • Weekly AI newsletter.
  • 6 times a year: “AI Radar” – deep dives on key AI topics.
  • Up to 25 % off on KI Pro online events.
  • Access to our full ten-year archive.
  • Get the latest AI news from The Decoder.
Subscribe to The Decoder