Google's Gemini AI Transforms Search into Conversational, Live Experience
Google’s Gemini AI transforms search into a conversational, multimodal assistant with new audio and live voice features.
June 20, 2025

Google is significantly advancing its search capabilities with the introduction of two new AI-powered features, Audio Overviews and Search Live, aimed at creating a more natural and conversational user experience. These additions, available through its experimental Search Labs platform for U.S. users, signal a clear trajectory towards a multimodal and interactive future for information retrieval. At the core of these innovations is Google's powerful Gemini AI model, which enables the search engine to understand and respond to user queries in a more human-like manner. The rollout of these features follows the broader introduction of AI Overviews, formerly known as the Search Generative Experience (SGE), which provides AI-generated summaries at the top of search results pages.[1][2] These initiatives collectively represent a pivotal shift in how Google is approaching its core product, moving beyond a simple list of links to a more intelligent and assistive tool.
The first of the new features, Audio Overviews, offers users the ability to receive spoken summaries of their search results.[3] When a user performs a search for certain topics, they will see an option to generate an audio overview.[4][5] This feature is designed for situations where reading a screen is inconvenient, such as when multitasking, cooking, or exercising.[5] The audio summaries are generated by Gemini and presented in a conversational format, aiming to provide a quick "lay of the land" on a given topic.[4][6] The audio player is integrated directly into the search results page and includes controls for play/pause, volume, and playback speed.[4] Crucially, the player also displays links to the web pages used to generate the summary, allowing users to delve deeper into the source material.[4][5] This feature follows in the footsteps of similar audio functionalities in other Google products like NotebookLM and the Gemini app.[4] While the convenience is a clear benefit, the rollout of AI-powered summaries has raised concerns about the potential for reduced traffic to content creators' websites, as users may find their queries answered without needing to click through to the source.[6][7]
The second major development is Search Live, a real-time, voice-driven conversational search feature available in the Google app for both iOS and Android.[8][9] By tapping a new "Live" icon, users can initiate a spoken dialogue with the search engine, asking questions and receiving AI-generated audio responses.[8][10] Powered by a custom version of Gemini with advanced voice capabilities, Search Live allows for natural back-and-forth conversations, including follow-up questions.[8][11] The system maintains context throughout the conversation, and users can even switch to another app while the conversation continues.[8][11] A transcript of the conversation is also available, allowing users to switch between voice and text input.[8][11] This hands-free experience is positioned as particularly useful for users on the go.[8][12] The introduction of Search Live mirrors capabilities seen in other conversational AI like ChatGPT and points to a broader industry trend towards more interactive and accessible search interfaces.[9]
These new voice and audio features are part of a larger push by Google towards a multimodal search experience, a concept heavily emphasized at its recent I/O conference.[13][14] The company's vision extends beyond text and voice to include visual search capabilities through Google Lens, all integrated within an "AI Mode."[14][15] This mode allows users to ask complex questions about images and receive detailed, context-aware answers, leveraging Gemini's ability to understand both visual and textual information.[14][16][15] The underlying technology for these advancements is complex, utilizing techniques like "query fan-out," where the AI breaks down a complex question into multiple sub-queries to gather a wider range of information.[17][18] Looking ahead, Google plans to further enhance Search Live by enabling users to incorporate their phone's camera in real-time, a development that aligns with its "Project Astra" initiative.[19][20][10] This project aims to create a universal AI agent that can understand and respond to the user's environment through sight and sound.[13][19][20]
In conclusion, the launch of Audio Overviews and Search Live marks a significant step in the evolution of Google Search. By leveraging the advanced multimodal capabilities of its Gemini AI, Google is transforming its search engine from a passive information repository into an active, conversational assistant. While the initial rollout is limited to U.S. users who opt-in through Search Labs, these features provide a clear glimpse into the future of search, a future that is more intuitive, interactive, and deeply integrated with artificial intelligence. However, this evolution is not without its challenges and implications. The initial rollout of the broader AI Overviews was met with some criticism for providing inaccurate or unhelpful information, prompting Google to make technical improvements.[21][22] Furthermore, the move towards providing direct answers and summaries within the search results page continues to fuel debate about the impact on website traffic and the broader content ecosystem.[6][22] As Google continues to refine these AI-powered experiences, the balance between user convenience and the health of the open web will remain a critical point of discussion.
Research Queries Used
Google Audio Overviews Search Labs
Google AI-powered voice search updates
Project Astra Google I/O 2024
Google multimodal search AI
Google AI Overviews rollout and user feedback
Sources
[2]
[8]
[10]
[12]
[15]
[16]
[17]
[18]
[19]
[20]
[22]