AI models can barely control their own reasoning, and OpenAI says that's a good sign
With GPT-5.4 Thinking, OpenAI is reporting on “CoT controllability” for the first time – a measure of whether AI models can deliberately manipulate their own reasoning. An accompanying study finds that reasoning models almost universally fail at this task, which OpenAI says is encouraging for AI safety.
Moltbook's alleged AI civilization is just a massive void of bloated bot traffic
Over 2.6 million AI agents interact on Moltbook with zero human involvement. They post, comment, and vote, but a new study shows they never learn from each other. It’s hollow interaction without mutual influence, shared memory, or social structures, a new study finds.
Current language model training leaves large parts of the internet on the table
Large language models learn from web data, but which pages actually make it into training sets depends heavily on a seemingly mundane choice: the HTML extractor. Researchers at Apple, Stanford, and the University of Washington found that three common extraction tools pull surprisingly different content from the same web pages.
Anthropic can't stop humanizing its AI models, now Claude Opus 3 gets a retirement blog
Anthropic is retiring its Claude Opus 3 AI model and letting it publish weekly essays on Substack. The company says it conducted “retirement interviews” to ask the model about its wishes, and it “enthusiastically” agreed. The move is a prime example of how AI companies keep pushing the humanization of their products, blurring the line between philosophical caution and PR stagecraft.