Stability AI, once a major force in open-source AI with its Stable Diffusion image model, is shifting its focus to commercial products. The company's latest release is Brand Studio, a platform built for creative teams that need AI-generated visuals matching their brand identity.
At the core of Brand Studio is "Brand Central," where teams can train their own brand-specific image models and set up campaign templates. A "Producer Mode" turns text descriptions into step-by-step visual production plans and runs them automatically. The platform also includes Curated Model Routing, which picks the best-suited AI model for a given task, whether that's Stable Diffusion or a third-party model. Other additions include "Precision Inpainting" for making targeted edits to specific parts of an image. Brand Studio comes in a free Core version and a paid Enterprise plan.
One in four quotes generated by AI systems comes from journalistic sources. That's the finding of PR database Muckrack, which evaluated 15 million quotes from AI responses across Gemini, Perplexity, Claude, and ChatGPT, as Press Gazette reports.
Trade publications and specialist journalists show up particularly often. Former Business Insider chief Henry Blodget is the most cited journalist worldwide. Reuters leads among publications globally, followed by Forbes. In the UK, The Guardian ranks first, followed by specialist magazine Homes and Gardens.
Muckrack sent millions of queries to all four AI services and tracked how often specific journalists and outlets appeared as linked sources. Based on the results, the company launched a new feature rating the "AI visibility" of journalists and publications across three tiers.
A separate analysis of Google's AI Overviews—AI-generated answers shown directly in search results—found that Facebook and Reddit are among the most cited sources across all queries.
Elon Musk has updated his lawsuit against OpenAI and Microsoft. He's now asking that any damages, potentially more than $150 billion, go not to him but to OpenAI's charitable foundation. He's also pushing for the removal of CEO Sam Altman from the foundation's board, according to the Wall Street Journal. Musk's lawyer, Marc Toberoff, said Musk "is not seeking a single dollar for himself."
Musk accuses OpenAI of abandoning its charitable mission and defrauding him as a donor by exploiting its nonprofit status. He wants Altman and OpenAI President Greg Brockman to turn over their shares and financial benefits to the foundation. The trial is set to begin in April in Oakland, California.
Musk argues OpenAI betrayed the mission he helped fund. However, early interview notes show he agreed to adding a for-profit unit in 2017 and actively discussed the transition while keeping the nonprofit in place.
OpenAI called the lawsuit on X "a harassment campaign driven by ego, jealousy and a desire to slow down a competitor." The company has also asked the attorneys general of Delaware and California to investigate Musk's behavior. OpenAI is currently valued at $852 billion and planning an IPO.
Microsoft's Bing team (yes, really) has released "Harrier," an open-source embedding model. Harrier supports more than 100 languages, offers a 32,000-token context window, and was trained on over two billion examples plus synthetic data from GPT-5. According to the team, Harrier takes the top spot on the multilingual MTEB v2 benchmark and outperforms proprietary models from OpenAI and Amazon.
Embedding models handle the searching, retrieving, and organizing of information AI systems need for accurate answers. According to Microsoft, they're becoming increasingly critical as AI agents independently take on more complex, multi-step tasks.
Meta employees compete for token consumption on an internal AI leaderboard
At Meta, employees compete for titles like “Token Legend,” “Model Connoisseur,” and “Cache Wizard” on an internal leaderboard that ranks AI token consumption. But burning through more tokens doesn’t automatically mean getting more done.
OpenAI's safety brain drain finally gets an explanation and it's just Sam Altman's vibes
“My vibes don’t really fit.” In a new New Yorker profile based on over 100 interviews, Sam Altman explains why safety researchers keep leaving OpenAI and why shifting commitments others might call deception are just part of the job.
Sycophantic AI chatbots can break even ideal rational thinkers, researchers formally prove
A new study by researchers from MIT and the University of Washington shows that even perfectly rational users can be drawn into dangerous delusional spirals by flattering AI chatbots. Fact-checking bots and educated users don’t fully solve the problem.