Ad
Skip to content
Read full article about: Anthropic says Claude Code's usage drain comes down to peak-hour caps and ballooning contexts

Anthropic has looked into complaints from users who were hitting their Claude Code usage limits much faster than expected. According to Anthropic's Lydia Hallie, tighter limits during peak hours and sessions with 1-million-token contexts growing larger are the two main reasons for the problem. Hallie says Anthropic also fixed some bugs, but none of them led to incorrect billing. The company has also shipped efficiency improvements and added in-product pop-ups to keep users informed.

Hallie recommends using Sonnet 4.6 instead of Opus, since Opus burns through limits roughly twice as fast. She also suggests turning off Extended Thinking when it's not needed, starting fresh sessions instead of continuing old ones, and limiting the context window. Users who still notice unusually high usage should report it through the feedback function.

Read full article about: OpenAI shifts to usage-based pricing for Codex in ChatGPT business plans

OpenAI is switching to usage-based pricing for Codex in ChatGPT Business and Enterprise. Admins can enable free Codex access across their workspace and pay only for actual usage; no upfront licenses required. Eligible Business customers can also claim up to 500 dollars in promotional credit per workspace as part of a limited-time promotion.

The move is designed to lower the barrier for enterprise adoption, OpenAI says. Coding tools typically spread from individual developers to full teams. "This model gives organizations a simpler way to support that motion inside a managed workspace," the company writes. OpenAI is likely betting that hands-on experience will drive long-term lock-in. It's a direct shot at GitHub Copilot and Cursor, which still charge per seat.

OpenAI says over two million developers use Codex weekly, with Business and Enterprise usage growing sixfold since January. The company's biggest competitor in this space is Anthropic with Claude Code.

Read full article about: Sakana AI launches "Ultra Deep Research" to automate weeks of strategy work

Japanese AI startup Sakana AI has unveiled "Sakana Marlin," its first product for business customers. The system works autonomously: give it a topic, and it researches on its own for up to eight hours, then delivers detailed reports and presentations. Sakana AI says the tool can produce professional strategy analyses that would normally take human teams several weeks.

Eine Übersichtsansicht eines von der KI generierten Beispiel-Dokuments. Zu sehen sind mehrere hochformatige Textseiten eines Berichts sowie darunterliegende querformatige Präsentationsfolien in japanischer Sprache. Auf dem Deckblatt sind das Logo von Sakana AI, der Name „Sakana Marlin“ sowie das Wasserzeichen „Sample output“ zu erkennen.
Sample output from "Sakana Marlin": after autonomous research, the tool creates text reports and presentation slides on a given topic (here: AI trends in the financial sector). | Image: Sakana AI

Sakana Marlin combines the company's "AI Scientist," designed to resolve contradictions, with its previously introduced "AB-MCTS" method for strategic searches. Multiple AI models work together, and longer thinking time is meant to yield better results, the company says.

The company is looking for beta testers in finance, research, and business consulting. The beta is free, but requires registration (the form is in Japanese). The biggest weakness of automated reports like these is hard-to-spot AI errors, something the startup doesn't address in its announcement.

Read full article about: Microsoft's MAI-Transcribe-1 runs 2.5x faster than its predecessor at $0.36 per audio hour

Microsoft has introduced MAI-Transcribe-1, a speech-to-text model supporting 25 languages that achieves the lowest word error rate of any model tested on the FLEURS benchmark, beating Scribe v2, Whisper-large-V3, GPT-Transcribe, and Gemini 3.1 Flash-Lite. The model is also built to handle tough recording conditions like background noise, poor audio quality, and overlapping speech, Microsoft says.

MAI-Transcribe-1 (green) leads in word error rate on the FLEURS benchmark in most of the 25 languages tested, outperforming Scribe v2, Gemini 3.1 Flash-Lite, Whisper-large-v3, and GPT-Transcribe. | Image: Microsoft

Microsoft is rolling out MAI-Transcribe-1 across Copilot Voice and Microsoft Teams. Developers can try it as a public preview through Microsoft Foundry and the Microsoft AI Playground. The model runs 2.5 times faster than Microsoft's previous Azure Fast offering and costs $0.36 per audio hour. Combined with MAI-Voice-1 and a language model, it can also power voice agents, Microsoft says.

Cohere and Mistral recently released open-source alternatives that perform at a similar level.

Read full article about: Alibaba launches Qwen3.6-Plus, its third proprietary AI model in days

Alibaba has released Qwen3.6-Plus, its third proprietary AI model in just a few days. The model is available through the Alibaba Cloud Model Studio API and offers a context window of one million tokens. According to the Qwen team, the focus is on significantly improved capabilities for agentic coding, including frontend development and complex code tasks.

In benchmarks published by Alibaba, the model partially outperforms Anthropic's older flagship model Claude 4.5 Opus, which was replaced by the stronger 4.6 Opus in December 2025. It's worth noting that some of these measurements were conducted by Alibaba itself.

Das neue Qwen3.6-Plus übertrifft das ältere 3.5-Modell und in einigen Fällen auch Opus. Allerdings erreicht das im Dezember 2025 veröffentlichte Opus 4.6 etwa im Terminal-Bench 2.0 65,4 Prozent und liegt damit vor Qwen3.6-Plus. | Bild: Alibaba
Qwen3.6-Plus outperforms the older 3.5 model and in some cases beats Opus. However, the Opus 4.6 released in December 2025 scores 65.4 percent on Terminal-Bench 2.0, putting it ahead of Qwen3.6-Plus. | Image: Alibaba

For a long time, Alibaba released its Qwen models as open source, but the company has recently changed course. The latest Qwen3.5-Omni is also not freely available. Alibaba wants to drive more revenue from enterprise customers with its proprietary models, as its cloud division faces intense competition from ByteDance.

According to Bloomberg, Alibaba is targeting $100 billion in AI revenue over the next five years. Qwen3.6-Plus will be integrated into the Qwen chatbot app and the company's new enterprise AI service Wukong.

Comment Source: Qwen