Short

According to internal documents obtained by TechCrunch, Google has been benchmarking its Gemini AI model against Anthropic's Claude. Google contractors are given up to 30 minutes per prompt to evaluate which model produces better outputs, focusing on criteria like truthfulness and comprehensiveness. Claude tends to be more safety-conscious in its answers compared to Gemini, according to Techcrunch. A Google DeepMind spokesperson confirmed that they do compare results across models, but stressed that they don't use Anthropic's models to directly improve Gemini, which would go against Anthropic's ToS. Also, this kind of competitive benchmarking is common in the AI industry - companies regularly benchmark their models against competitors to understand where they stand. Moreover, Google is an investor in Anthropic.

Ad
Ad
Short

OpenAI CEO Sam Altman recently took to X to gauge what users want from the company in 2025. The most requested features (?) were AGI, AI agents, a "much better 4o upgrade," better memory capabilities, and longer context windows. Users also clamored for an "adult mode" that would dial back OpenAI's content moderation, along with enhanced research capabilities and improvements to the company's Sora videogenerator. One of these wishes - better memory - is already in testing, though OpenAI hasn't revealed any details about the underlying technology. Interestingly, Altman noted that many of the features OpenAI actually plans to release in 2025 barely made it onto users' wishlists - or didn't show up at all.

Ad
Google News