Search is becoming a conversation, and Google just went first at scale

Google's expansion of Search Live to more than 200 countries this week is easy to frame as a product update. Its flagship AI voice model got an update and wider availability.
But zoom out and the move looks more significant: it is the first time any company has deployed conversational, multimodal search at truly global scale.
Search Live lets users point their phone camera at something, ask a question by voice, and have a back-and-forth dialogue with Google Search about what they are seeing.

It is powered by Gemini 3.1 Flash Live, Google's new real-time audio model, which supports over 90 languages. The feature was previously limited to a handful of markets. Now it is available everywhere Google's AI Mode operates.
Here is a video of Search Live, with voice, in action:
This matters because it changes what "searching" looks like.
For two decades, search meant typing keywords into a box and scanning a list of blue links. AI Overviews (pictured below) began shifting that toward summarized answers, but the input was still text and the output was still a page.

Search Live essentially replaces both by making the input your voice and your camera. The output is a spoken conversation.
The search results page, the thing Google built its entire advertising business around, does not appear.
The competition is converging
Google is not alone in recognizing this shift, but it is the first to deploy it at this scale.
OpenAI recently improved the voice search capabilities of ChatGPT in February 2026, noting improved ability to follow user instructions and use tools like web search for better responses:

Perplexity also offers voice search in its main app and its Comet browser, including citations for its information.

Both are meaningful products, but neither have the sheer distribution advantage of Google. The fact that Search is already on every Android phone and most of the world's browsers, gives it a head start that is hard to replicate with a standalone app.
Google's approach integrates voice, live camera input, and web search into a single interaction built around Google's search infrastructure. ChatGPT also supports live camera sharing during voice conversations and can invoke web search, but these are general-purpose assistant capabilities rather than fully-fleshed out features of a dedicated search product. In Search Live, pointing a camera at an object can surface shopping results, web links, and visual matches alongside the spoken response, tying the visual context directly to Google's search capabilities.
What changes when search becomes a dialogue
The shift from text queries to spoken conversation is more just an interface change, as it rewrites several assumptions that the search industry has operated on for years.
Query structure changes. When people type, they use keywords. When they speak, they use natural sentences, follow-up questions, and references to what they said 30 seconds ago. Search systems built for keyword matching are not optimized for this. Systems built for multi-turn dialogue are.
The "results page" disappears. In a voice conversation, there is one answer, spoken aloud, possibly followed by a clarifying question. While it does attribute where it’s getting the information from (as seen in the image below) it compresses the information funnel dramatically. For users, it is more convenient (as long as the information it provides is valuable of course).

For publishers and advertisers who depend on appearing in search results, it raises hard questions about visibility and attribution.
Context becomes continuous. In traditional search, each query is independent. In a voice conversation, the fifth question depends on the answer to the fourth. The model needs to maintain state across the entire dialogue. Google says 3.1 Flash Live can follow a conversation thread for twice as long as its predecessor, which suggests this is an area where they are actively investing.
The business model question
The most interesting tension in conversational search is economic. In 2024, Google generated $265 billion in ad revenue, most of it from search ads that appear alongside query results. Wouldn’t a voice-first search interface that eliminate the results page also eliminates the primary surface for those ads?
Google has not said how it plans to monetize Search Live specifically. The company has been experimenting with ads in AI Overviews, but inserting advertising into a spoken conversation is a different and more delicate problem. Users are likely to have less tolerance for interruptions in a dialogue than they do for sponsored links on a page.
This is not a problem unique to Google. Any company building conversational search, whether OpenAI, Perplexity, or anyone else, faces the same question: if the answer is the product, where does the revenue come from?
What to watch
The real test is whether people actually use voice as their primary way to search, or whether it remains a secondary mode for specific situations like driving or cooking.
Google has the distribution to find out. And perhaps that’s what it is doing: testing traction. With Search Live now available globally, the company will have data within months on whether conversational search is something billions of people want or something that demos well but does not change daily behavior. That data will shape not just Google's product roadmap but the strategic assumptions of every company building in this space.
For now, the safest observation is that the interface of search is fragmenting. Text boxes, AI summaries, voice conversations, and camera-based queries are all becoming valid ways to find information. Google's decision to roll out Search Live to more markets is that voice and vision will not just coexist with text search but eventually become the default for a significant share of queries.
This week's global launch is the beginning of that experiment.

Written by
Joseph Nordqvist
Joseph founded AI News Home in 2026. He studied marketing and later completed a postgraduate program in AI and machine learning (business applications) at UT Austin’s McCombs School of Business. He is now pursuing an MSc in Computer Science at the University of York.
Editorial Transparency
This article was produced with the assistance of AI tools as part of our editorial workflow. All analysis, conclusions, and editorial decisions were made by human editors. Read our Editorial Guidelines
References
- 1.
Gemini 3.1 Flash Live: Making audio AI more natural and reliable — Valeria Wu,Yifan Ding, Google, March 26, 2026
Primary - 2.
Build real-time conversational agents with Gemini 3.1 Flash Live — Alisa Fortin,Thor Schaeff, Google, March 26, 2026
Primary - 3.
Alphabet Inc. Q4 2024 earnings (Form 8-K), Alphabet Inc. / SEC, February 4, 2025
Google advertising revenue summed across Q1-Q4 2024 SEC filings: approximately $264.6 billion
Primary - 4.
- 5.
Was this useful?