Google Search Unleashes AI Agents: Books Tasks, Understands Visuals
From finding to doing: Google's AI Mode transforms search with visual understanding, conversational shopping, and agentic task completion.
September 30, 2025

Google is fundamentally transforming its search experience by integrating sophisticated visual and agent-like capabilities into its AI Mode, signaling a strategic shift from a tool that finds information to a conversational assistant that accomplishes tasks.[1][2] The technology giant has rolled out a significant update that allows users to search using natural, conversational language and receive a rich range of visual results, moving beyond the limitations of text-based queries.[3][4] This enhancement is particularly aimed at searches that are difficult to articulate with words alone, such as finding design inspiration or specific fashion items.[5][6] Concurrently, Google is experimenting with "agentic capabilities" in AI Mode, enabling the AI to take direct actions on behalf of the user, such as booking restaurant reservations, a move that hints at a future where search engines are active participants in completing real-world tasks.[1][2]
The core of the new visual search functionality lies in a proprietary technique Google calls "visual search fan-out."[3][7] This method builds upon the "query fan-out" approach already used in AI Mode's text-based answers, where a single user prompt is broken down into multiple related queries that run in the background.[7][8] For visual search, the system performs a comprehensive analysis of an image, identifying not just the primary subject but also subtle details and secondary objects.[3][9] It then executes several queries simultaneously to grasp the full context of the image and the user's conversational prompt, delivering more relevant and nuanced visual results.[3][9] This powerful multimodal experience is underpinned by a combination of Google's established visual understanding technologies like Lens and Image Search, fused with the advanced capabilities of its Gemini 2.5 AI model.[3][4][10] Users can initiate a search with a text prompt, an uploaded image, or by taking a new photo, and then conversationally refine the results, for example, by asking for "more options with dark tones and bold prints" after an initial search for "maximalist design inspiration."[3][5]
A major beneficiary of this visual evolution is online shopping, which becomes a more intuitive and less rigid process.[6][4] Instead of relying on traditional filters for size, color, and style, shoppers can now describe what they are looking for in conversational terms.[5] For instance, a user can search for "barrel jeans that aren't too baggy" and receive a curated set of shoppable product images.[4] From there, they can further narrow the options with follow-up requests like "show me ankle length."[5] Each image links directly to the retailer's website, streamlining the path to purchase.[4] This enhanced shopping experience is powered by Google's massive Shopping Graph, which encompasses over 50 billion product listings that are constantly updated with details like reviews, deals, and stock availability.[5][11] By making the discovery process more visual and conversational, Google aims to provide the inspiration that text-only results often lack, addressing a key challenge for users who know what they want to see but can't perfectly describe it.[4]
Beyond enhancing visual discovery, Google is pushing the boundaries of its AI's capabilities by introducing agentic functions that actively assist users in completing tasks.[1][2] Available as a Labs experiment in the U.S. for Google AI Ultra subscribers, these new features allow AI Mode to perform actions like searching for and presenting available restaurant reservation slots across multiple platforms.[1][2] A user can make a complex request, such as "find me a dinner reservation for 3 people this Friday after 6pm around Logan square. craving ramen or bibimbap," and the AI will search platforms like OpenTable and Resy for real-time availability that matches the specific criteria.[1][2] The AI then presents a curated list and provides a direct link to the booking page to finalize the reservation.[1] This functionality is powered by Project Mariner's live web browsing capabilities, partner integrations, the Knowledge Graph, and Google Maps.[1] Google has stated its intention to expand these agentic capabilities to include booking local service appointments and event tickets, demonstrating a clear strategy to evolve Google Search into a personal assistant that not only finds information but also acts on it.[2]
The expansion of AI Mode with both advanced visual search and initial agentic capabilities represents a significant step in the evolution of information technology, challenging the traditional paradigm of keyword-based search engines. By enabling a more natural, multimodal, and conversational interaction, Google is making its search platform more intuitive and aligned with human thought processes. The "visual search fan-out" technique, powered by Gemini 2.5, offers a deeper, more contextual understanding of user intent, particularly in visually driven domains like shopping and creative inspiration.[3][10] The simultaneous introduction of AI agents that can execute real-world tasks like booking reservations signals a move towards a more proactive and automated internet experience, where the search engine transitions from a passive repository of links to an active collaborator in a user's daily life.[2][12] This dual advancement underscores a broader industry trend towards more capable and integrated AI assistants, fundamentally reshaping how users interact with the digital world and setting a new standard for what they can expect from a search query.