Search is becoming a conversation, and Google just went first at scale

Written by Joseph Nordqvist/March 26, 2026 at 11:13 PM UTC

7 min read
A Google Search Live demo on a phone showing the camera pointed at screws and a metal bracket on a wooden surface. The spoken response reads "it's much safer to screw directly into wall studs if you can." Below the camera feed, a "7 sites" indicator and a web result for "Floating Shelf Installation: A Complete Guide to Wall Anchors" from HomeRepairDaily are visible, alongside Unmute, Video, and Transcript buttons.
(Source: Google)

Google's expansion of Search Live to more than 200 countries this week is easy to frame as a product update. Its flagship AI voice model got an update and wider availability.

But zoom out and the move looks more significant: it is the first time any company has deployed conversational, multimodal search at truly global scale.

Search Live lets users point their phone camera at something, ask a question by voice, and have a back-and-forth dialogue with Google Search about what they are seeing.

A post on X from Google's verified account announcing that Search Live is now global. The post includes a video showing a phone pointed at a white armchair, with a Spanish-language voice query asking what style of chair it is. The response identifies it as contemporary or mid-century style. The post has 828 likes and 59,000 views.
Google announcing the global expansion of Search Live on X. The accompanying video demonstrates the feature in Spanish, with a user pointing their camera at a chair and asking about its style by voice. (Screenshot: @Google on X, March 26, 2026)

It is powered by Gemini 3.1 Flash Live, Google's new real-time audio model, which supports over 90 languages. The feature was previously limited to a handful of markets. Now it is available everywhere Google's AI Mode operates.

Here is a video of Search Live, with voice, in action:

This matters because it changes what "searching" looks like.

For two decades, search meant typing keywords into a box and scanning a list of blue links. AI Overviews (pictured below) began shifting that toward summarized answers, but the input was still text and the output was still a page.

A Google search results page in dark mode showing an AI Overview for the query "How has search changed." The AI-generated summary describes the shift from keyword-based search to conversational answer engines, with bullet points covering AI Overviews, zero-click results, and user intent. A "Dive deeper in AI Mode" button appears at the bottom of the summary.
Google's AI Overview for the query "How has search changed," summarizing the shift to AI-generated answers. (Screenshot: Google Search, March 2026)

Search Live essentially replaces both by making the input your voice and your camera. The output is a spoken conversation.

The search results page, the thing Google built its entire advertising business around, does not appear.

The competition is converging

Google is not alone in recognizing this shift, but it is the first to deploy it at this scale.

OpenAI recently improved the voice search capabilities of ChatGPT in February 2026, noting improved ability to follow user instructions and use tools like web search for better responses:

A screenshot from OpenAI's ChatGPT release notes dated February 12, 2026, titled "ChatGPT Voice Update." The text describes an update improving Voice's ability to follow user instructions and use tools like web search.
OpenAI's February 2026 release notes describing improvements to ChatGPT Voice's web search capabilities. (Screenshot: OpenAI ChatGPT Release Notes)

Perplexity also offers voice search in its main app and its Comet browser, including citations for its information.

A screenshot of Perplexity's voice search mode on iOS. The app displays a spoken response referencing articles about AI's impact on search, with a "10 sources" indicator at the top. A golden animated orb and microphone button sit at the bottom of the screen, indicating an active voice conversation.
Perplexity's voice search in the main iOS app, responding to a query about AI's impact on search. Unlike Google's Search Live, the interaction is voice-and-text only, without integrated camera input. (Screenshot: Perplexity iOS app)

Both are meaningful products, but neither have the sheer distribution advantage of Google. The fact that Search is already on every Android phone and most of the world's browsers, gives it a head start that is hard to replicate with a standalone app.

Google's approach integrates voice, live camera input, and web search into a single interaction built around Google's search infrastructure. ChatGPT also supports live camera sharing during voice conversations and can invoke web search, but these are general-purpose assistant capabilities rather than fully-fleshed out features of a dedicated search product. In Search Live, pointing a camera at an object can surface shopping results, web links, and visual matches alongside the spoken response, tying the visual context directly to Google's search capabilities.

What changes when search becomes a dialogue

The shift from text queries to spoken conversation is more just an interface change, as it rewrites several assumptions that the search industry has operated on for years.

Query structure changes. When people type, they use keywords. When they speak, they use natural sentences, follow-up questions, and references to what they said 30 seconds ago. Search systems built for keyword matching are not optimized for this. Systems built for multi-turn dialogue are.

The "results page" disappears. In a voice conversation, there is one answer, spoken aloud, possibly followed by a clarifying question. While it does attribute where it’s getting the information from (as seen in the image below) it compresses the information funnel dramatically. For users, it is more convenient (as long as the information it provides is valuable of course).

A Google Search Live demo on a phone showing the camera pointed at screws and a metal bracket on a wooden surface. The spoken response reads "it's much safer to screw directly into wall studs if you can." Below the camera feed, a "7 sites" indicator and a web result for "Floating Shelf Installation: A Complete Guide to Wall Anchors" from HomeRepairDaily are visible, alongside Unmute, Video, and Transcript buttons.
Google's Search Live combining camera input, spoken advice, and web results in a single interaction. The user points their camera at mounting hardware and receives both a spoken answer and relevant web links. (Source: Google)

For publishers and advertisers who depend on appearing in search results, it raises hard questions about visibility and attribution.

Context becomes continuous. In traditional search, each query is independent. In a voice conversation, the fifth question depends on the answer to the fourth. The model needs to maintain state across the entire dialogue. Google says 3.1 Flash Live can follow a conversation thread for twice as long as its predecessor, which suggests this is an area where they are actively investing.

The business model question

The most interesting tension in conversational search is economic. In 2024, Google generated $265 billion in ad revenue, most of it from search ads that appear alongside query results. Wouldn’t a voice-first search interface that eliminate the results page also eliminates the primary surface for those ads?

Google has not said how it plans to monetize Search Live specifically. The company has been experimenting with ads in AI Overviews, but inserting advertising into a spoken conversation is a different and more delicate problem. Users are likely to have less tolerance for interruptions in a dialogue than they do for sponsored links on a page.

This is not a problem unique to Google. Any company building conversational search, whether OpenAI, Perplexity, or anyone else, faces the same question: if the answer is the product, where does the revenue come from?

What to watch

The real test is whether people actually use voice as their primary way to search, or whether it remains a secondary mode for specific situations like driving or cooking.

Google has the distribution to find out. And perhaps that’s what it is doing: testing traction. With Search Live now available globally, the company will have data within months on whether conversational search is something billions of people want or something that demos well but does not change daily behavior. That data will shape not just Google's product roadmap but the strategic assumptions of every company building in this space.

For now, the safest observation is that the interface of search is fragmenting. Text boxes, AI summaries, voice conversations, and camera-based queries are all becoming valid ways to find information. Google's decision to roll out Search Live to more markets is that voice and vision will not just coexist with text search but eventually become the default for a significant share of queries.

This week's global launch is the beginning of that experiment.

Joseph Nordqvist

Written by

Joseph Nordqvist

Founder & Editor-in-Chief at AI News Home

View all articles →

Editorial Transparency

This article was produced with the assistance of AI tools as part of our editorial workflow. All analysis, conclusions, and editorial decisions were made by human editors. Read our Editorial Guidelines

References

  1. 1.
    Gemini 3.1 Flash Live: Making audio AI more natural and reliableValeria Wu,Yifan Ding, Google, March 26, 2026
    Primary
  2. 2.
    Build real-time conversational agents with Gemini 3.1 Flash LiveAlisa Fortin,Thor Schaeff, Google, March 26, 2026
    Primary
  3. 3.
    Alphabet Inc. Q4 2024 earnings (Form 8-K), Alphabet Inc. / SEC, February 4, 2025

    Google advertising revenue summed across Q1-Q4 2024 SEC filings: approximately $264.6 billion

    Primary
  4. 4.
    Gemini 3.1 Flash Live model card, Google DeepMind, March 26, 2026
    Primary
  5. 5.

Was this useful?