Ad
Skip to content
Read full article about: Google improves "Search Live" with new AI voice

Google has updated the voice for "Search Live." A new Gemini audio model powers the feature, producing responses that sound more natural and fluid, according to a blog post. Search Live lets users have real-time conversations while displaying relevant websites. The feature is part of Google Search's "AI Mode".

The update rolls out to all Search Live users in the US over the coming week. Users can open the Google app on Android or iOS, tap the Live icon, and speak their question.

The update fits into Google's broader push to build a voice-controlled assistant capable of handling everyday tasks—a goal shared by OpenAI and other major AI companies.

Read full article about: Anthropic places $21 billion order for Google chips via Broadcom

AI lab Anthropic has placed orders totaling $21 billion with Broadcom for Google's AI chips. Broadcom CEO Hock Tan confirmed that the startup is purchasing "Ironwood Racks" equipped with Google's Tensor Processing Units (TPUs).

The move follows a massive cloud partnership between Anthropic and Google announced in late October. That deal grants Anthropic access to up to one million TPUs and is expected to bring over one gigawatt of new AI compute capacity online by 2026. Anthropic maintains a multi-cloud strategy, spreading its workloads across Google TPUs, Amazon's Trainium chips, and Nvidia GPUs.

Comment Source: CNBC
Read full article about: Google opens its infrastructure for AI models via MCP

Google is integrating Anthropic's Model Context Protocol (MCP) directly into its cloud infrastructure. MCP serves as a universal standard for connecting AI models with external data and tools, eliminating the need to program new interfaces for every application.

Starting immediately, Google is offering managed servers that give AI agents direct access to services like Google Maps, BigQuery, and both the Compute and Kubernetes Engines. This allows AI to handle tasks like independently managing infrastructure or planning travel routes. Through the Apigee platform, companies can also deploy their own internal APIs as AI tools. Google announced plans to expand support to additional services, such as Cloud Storage and databases, in the near future.

Read full article about: Meta is reportedly ditching open Llama models for Avocado, a closed model built for direct sales

According to Bloomberg's sources, Meta is shifting its focus to a new AI model codenamed "Avocado," with a release potentially coming next spring. Avocado is expected to launch as a closed model, letting the company sell access directly. This marks a major shift from Meta's established open-model strategy. Internally, the open-source approach reportedly lost steam after the disappointing performance of Llama 4. Management is betting big on Alexandr Wang, who joined Meta following the company's deal with Scale AI.

The development process involves some surprising ingredients. According to Bloomberg, the team is training Avocado using several external models, including Google's Gemma, OpenAI's gpt-oss, and Alibaba's Qwen. Using Chinese technology clashes with CEO Mark Zuckerberg's previous warnings about Chinese censorship.