Reasoning tasks sharply raise AI costs, according to a new analysis by Artificial Analysis. Google's Gemini Flash 2.5 costs 150 times more to run than Flash 2.0, due to using 17 times more tokens and charging $3.50 per million output tokens with reasoning, compared to $0.40 for the earlier model. This makes Flash 2.5 the most expensive model in terms of token use for logic. OpenAI's o4-mini costs more per token but used fewer tokens overall, making it cheaper in the benchmark.

SoundCloud changed its terms of use in February 2024 to allow uploaded music to be used for AI training. AI copyright activist Ed Newton-Rex spotted the change and said that users were not informed. In a statement, SoundCloud said it does not train AI models with artist content, does not build its own AI tools, and blocks third-party scraping. It said AI is only used internally for things like recommendations, fraud detection, and content sorting. Artists keep control of their work, and all AI use follows existing license deals. However, the statement does not clearly rule out general AI training.
Google is now using AI models to protect Chrome users from online scams. On desktop, the company has rolled out its local Gemini Nano language model to quickly spot fraudulent websites, including ones that have never been seen before. On Android, Chrome will now warn users about suspicious notifications sent by websites. Google says these changes are part of a broader effort to improve security, which also includes the "Enhanced Protection" feature in Safe Browsing. The company reports that AI-powered systems in Google Search block hundreds of millions of scam results every day, cutting the number of fake airline support pages by more than 80 percent.
Google introduces "implicit caching" in Gemini 2.5, aiming to cut developer costs by as much as 75 percent. The new feature automatically detects and stores recurring content, so repeated prompts are only processed once. According to Google, this can lead to significant savings compared to the old explicit caching method, where users had to set up their own cache. To maximize the benefits, Google recommends putting the stable part of a prompt—like system instructions—at the start, and adding user-specific input, such as questions, afterwards. Implicit caching kicks in for Gemini 2.5 Flash starting at 1,024 tokens, and for Pro versions from 2,048 tokens onwards. More details and best practices are available in the Gemini API documentation.
OpenAI is rolling out a new GitHub connector for ChatGPT's deep research agent. Users with Plus, Pro, or Team subscriptions can now connect their own GitHub repositories and ask questions about their code. ChatGPT searches through the source code and documentation in the repo, then returns a detailed report with source references. Only content that users already have access to is visible to ChatGPT, so existing permissions apply. The connector will become available to users over the next few days, with support for enterprise customers coming soon. According to OpenAI Product Manager Nate Gonzalez, the goal is to better integrate ChatGPT into internal workflows. OpenAI also plans to add more deep research connectors in the future.
Media error: Format(s) not supported or source(s) not found
Download File: https://the-decoder.com/wp-content/uploads/2025/05/twitter_video_downloader.mp4?_=1