Google is bringing live video analysis to its Gemini assistant. The company announced this at the Mobile World Congress.
Google is presenting AI functions for its Gemini assistant at the Mobile World Congress (MWC) in Barcelona. The company announced that subscribers to the Google One AI Premium plan for Gemini Advanced will have access to live video and screen sharing features later this month.
Gemini Live will have two major new features: Firstly, the ability to analyze live video, and secondly, a screen sharing feature. With both functions, users can share visual content with the AI assistant in real time - in the case of live video via camera images from outside, and in the case of screen sharing via their own smartphone screen in order to have content commented on.
The new functions are initially only available for Android devices and support multiple languages. At the MWC, Google will be demonstrating the integration of these functions on partner devices from various Android manufacturers.
AI assistants arrive in the real world
The addition of visual functions is an important step in the development of AI assistants, which are increasingly expected to act multimodally and interact with the real world.
Google's goal for 2025 is "Project Astra", a universal multimodal AI assistant that can process text, video and audio data in real time and store it in a conversational context for up to ten minutes. Astra will also be able to use Google Search, Lens and Maps.
It is not known whether Google is actually planning to release Astra or, more likely, whether the functions presented for Astra will be integrated into Gemini.
With Gemini Live, Google is positioning itself against its competitor OpenAI and its ChatGPT: ChatGPT's Advanced Voice Mode has supported live and screen sharing since December.