Photo by Charlotte May on Pexels
Google is set to upgrade its Gemini Live AI assistant, introducing real-time visual assistance capabilities. Starting next week, Gemini Live will be able to highlight specific elements directly on a user’s screen while they share their camera feed, enabling clearer and more precise interactions.
Imagine pointing your smartphone at a workbench full of tools – Gemini Live will be able to identify and highlight the precise tool you need. This feature will first appear on the Pixel 10 on August 28th and then expand to other Android and iOS devices shortly after.
Beyond visual guidance, Gemini Live will also deepen its integration with apps like Messages, Phone, and Clock. Users will gain the ability to seamlessly interrupt conversations to trigger actions such as sending a quick text or setting a reminder, creating a more fluid user experience.
Google is also improving Gemini’s speech capabilities. A revamped audio model will allow the assistant to better incorporate natural speech patterns like intonation, rhythm, and pitch. Gemini’s tone will become more context-aware, using a gentler voice for sensitive topics, and users can now personalize the assistant’s speaking speed. Looking ahead, Gemini might even use different accents when narrating stories from varied perspectives, adding a touch of personality to its interactions.