At the Mobile World Congress (MWC) 2025 in Barcelona, Google introduced two new AI-powered features for its Gemini assistant: Live Video and Screen Sharing. These innovations aim to enhance how Android users interact with their devices by enabling real-time, context-aware responses. Both features are set to launch later this month for Gemini Advanced subscribers.
Live Video: Real-Time AI Assistance
Live Video, first previewed at Google I/O 2024, leverages Gemini’s advanced multimodal AI to provide instant insights based on a smartphone’s camera feed. Users can point their camera at objects, scenes, or even printed text and ask Gemini contextual questions. Google demonstrated how the AI could analyze mid-century modern decor and suggest ways to incorporate the style into pottery. This capability makes the AI assistant more interactive, allowing users to seek recommendations, identify objects, or better understand their surroundings.
Read More: Google launches AI tool to simplify career exploration
While this feature closely mirrors ChatGPT’s Advanced Voice Mode, which introduced live video capabilities last year, Google’s implementation integrates seamlessly within the Gemini ecosystem, eliminating the need for third-party apps.
Screen Sharing: On-Screen Assistance
The new screen-sharing feature enables users to share their phone screens with Gemini and ask questions about on-screen content in real time. Whether browsing a webpage, shopping online, or viewing a PDF, users can get immediate AI-powered responses.
For instance, while watching a fitness tutorial on YouTube, a user could ask Gemini about muscle engagement for a specific exercise. Similarly, when viewing a product page, the AI could suggest complementary fashion choices or provide background information on an item. The “Ask about this PDF” feature also streamlines research by generating summaries and clarifications without requiring a desktop computer.
This development is part of Google’s broader Project Astra, which focuses on creating a highly responsive and visually aware AI assistant. The goal is to make Gemini a more intuitive tool for real-time, multimodal interactions.
Seamless Integration with Google’s Ecosystem
One of the standout aspects of Google’s approach is how these features blend into existing apps and services. Gemini’s screen-sharing capability will work across browsers, shopping platforms, and media apps like YouTube, enhancing accessibility and usability.
The AI assistant’s ability to interpret visual inputs in real-time offers a new level of convenience, potentially reducing the need for manual searches. With these advancements, Google is positioning Gemini as an AI assistant that actively enhances daily tasks rather than simply providing static responses.
Availability and Subscription Details
Both Live Video and Screen Sharing will be available later this month for Gemini Advanced subscribers. Access to these features is bundled with Google’s $20 per month Google One AI Premium plan. Additionally, Verizon has recently incorporated this subscription into select customer plans, expanding its availability.
Google’s AI Expansion with SpeciesNet
Beyond Gemini’s new capabilities, Google also recently open-sourced SpeciesNet, an AI model designed to identify wildlife through camera trap images. The model can recognize individual species and classify them into taxonomic categories such as “mammalian” or “Felidae.” It can even identify non-animal objects like vehicles, further showcasing Google’s advancements in AI-driven image recognition.
Read More: Google Gemini AI gets memory recall in major update
By incorporating Live Video and Screen Sharing, Google is setting new benchmarks for AI assistants. These features not only enhance convenience but also shift AI from a passive tool into an interactive assistant capable of real-world engagement. As these updates roll out, Google’s Gemini is expected to redefine how users interact with AI, making smartphones smarter and everyday tasks more intuitive.