Search Weight Loss Topics:




May 11

AI is coming to Google search through Search Generative Experience – The Verge

The future of Google Search is AI. But not in the way you think. The company synonymous with web search isnt all in on chatbots (even though its building one, called Bard), and its not redesigning its homepage to look more like a ChatGPT-style messaging system. Instead, Google is putting AI front and center in the most valuable real estate on the internet: its existing search results.

To demonstrate, Liz Reid, Googles VP of Search, flips open her laptop and starts typing into the Google search box. Why is sourdough bread still so popular? she writes and hits enter. Googles normal search results load almost immediately. Above them, a rectangular orange section pulses and glows and shows the phrase Generative AI is experimental. A few seconds later, the glowing is replaced by an AI-generated summary: a few paragraphs detailing how good sourdough tastes, the upsides of its prebiotic abilities, and more. To the right, there are three links to sites with information that Reid says corroborates whats in the summary.

Google calls this the AI snapshot. All of it is by Googles large language models, all of it sourced from the open web. Reid then mouses up to the top right of the box and clicks an icon Googles designers call the bear claw, which looks like a hamburger menu with a vertical line to the left. The bear claw opens a new view: the AI snapshot is now split sentence by sentence, with links underneath to the sources of the information for that specific sentence. This, Reid points out again, is corroboration. And she says its key to the way Googles AI implementation is different. We want [the LLM], when it says something, to tell us as part of its goal: what are some sources to read more about that?

A few seconds later, Reid clicks back and starts another search. This time, she searches for the best Bluetooth speakers for the beach. Again, standard search results appear almost immediately, and again, AI results are generated a few seconds later. This time, theres a short summary at the top detailing what you should care about in such a speaker: battery life, water resistance, sound quality. Links to three buying guides sit off to the right, and below are shopping links for a half-dozen good options, each with an AI-generated summary next to it. I ask Reid to follow up with the phrase under $100, and she does so. The snapshot regenerates with new summaries and new picks.

These AI snapshots will appear at the top of Search and pull information from all over the web. Image: Google

This is the new look of Googles search results page. Its AI-first, its colorful, and its nothing like youre used to. Its powered by some of Googles most advanced LLM work to date, including a new general-purpose model called PaLM 2 and the Multitask Unified Model (MUM) that Google uses to understand multiple types of media. In the demos I saw, its often extremely impressive. And it changes the way youll experience search, especially on mobile, where that AI snapshot often eats up the entire first page of your results.

There are some caveats: to get access to these AI snapshots, youll have to opt in to a new feature called Search Generative Experience (SGE for short), part of an also-new feature called Search Labs. Not all searches will spark an AI answer the AI only appears when Googles algorithms think its more useful than standard results, and sensitive subjects like health and finances are currently set to avoid AI interference altogether. But in my brief demos and testing, it showed up whether I searched for chocolate chip cookies, Adele, nearby coffee shops, or the best movies of 2022. AI may not be killing the 10 blue links, but its definitely pushing them down the page.

SGE, Google executives tell me over and over, is an experiment. But theyre also clear that they see it as a foundational long-term change to the way people search. AI adds another layer of input, helping you ask better and richer questions. And it adds another layer of output, designed to both answer your questions and guide you to new ones.

An opt-in box at the top of search results might sound like a small move from Google compared to Microsofts AI-first Bing redesign or the total newness of ChatGPT. But SGE amounts to the first step in a complete rethinking of how billions of people find information online and how Google makes money. As pixels on the internet go, these are as consequential as it gets.

The AI snapshots borrow colors from the content they discover and change depending on what you search. Image: Google

Asked and answered

Google feels pretty good about the state of its search results. Were long past the 10 blue links era of 25 years ago when you Googled by typing in a box and getting links in return. Now, you can search by asking questions aloud or snapping a picture of the world, and you might get back everything from images and podcasts to TikToks.

Many searches are already well served by these results. If youre going to Google and searching Facebook to land on facebook.com or youre looking for the height of the Empire State Building, youre already good to go.

But theres a set of queries for which Google has never quite worked, which is where the company is hoping AI can come in. Queries like Where should I go in Paris next week? or Whats the best restaurant in Tokyo? These are hard questions to answer because theyre not actually one question. Whats your budget? What days are all the museums open in Paris? How long are you willing to wait? Do you have kids with you? On and on and on.

Theres a set of queries for which Google has never quite worked, which is where the company is hoping AI can come in

The bottleneck turns out to be what I call the orchestration of structure, says Prabhakar Raghavan, the SVP at Google who oversees Search. Much of that data exists somewhere on the internet or even within Google museums post hours on Google Maps, people leave reviews about wait times at restaurants but putting it all together into something like a coherent answer is really hard. People want to say, plan me a seven-day vacation, Raghavan says, and they believe if the language model outputs, it should be right.

One way to think about these is simply as questions with no right answer. A huge percentage of people who come to Google arent looking for a piece of information that exists somewhere. Theyre looking for ideas, looking to explore. And since theres also likely no page on the internet titled Best vacation in Paris for a family with two kids, one of whom has peanut allergies and the other of whom loves soccer, and you definitely want to go to the Louvre on the quietest possible day of the week, the links and podcasts and TikToks wont be much help.

Because theyre trained on a huge corpus of data from all over the internet, large language models can help answer those questions by essentially running lots of disparate searches at once and then combining that information into a few sentences and a few links. Lots of times you have to take a single question and break it into 15 questions to get useful information from search, Reid says. Can you just ask one? How do we change how the information is organized?

Thats the idea, but Raghavan and Reid are both quick to point out that SGE still cant do these completely creative acts very well. Right now, its going to be much more handy for synthesizing all the search data behind questions like what speaker should I buy to take into the pool. Itll do well with what were the best movies of 2022, too, because it has some objective Rotten Tomatoes-style data to pull from along with the internets many rankings and blog posts on the subject. AI appears to make Google a better information-retrieval machine, even if its not quite ready to be your travel agent.

One thing that didnt show up in most SGE demos? Ads. Google is still experimenting with how to put ads into the AI snapshots, though rest assured, theyre coming.Googles going to need to monetize the heck out of AI for any of this to stick.

Right now, AI hasnt really changed how Google ads work. But it will. Image: Google

The Google Bot

At one point in our demo, I asked Reid to search only the word Adele. The AI snapshot contained more or less what youd expect some information about her past, her accolades as a singer, a note about her recent weight loss and then threw in that her live performances are even better than her recorded albums. Googles AI has opinions! Reid quickly clicked the bear claw and sourced that sentence to a music blog but also acknowledged that this was something of a system failure.

Googles search AI is not supposed to have opinions. Its not supposed to use the word I when it answers questions. Unlike Bings multiple-personality chaos or ChatGPTs chipper helper or even Bards whole droll middle school teacher vibe, Googles search AI is not trying to seem human or affable. Its actually trying very hard to not be those things. You want the librarian to really understand you, Reid says. But most of the time, when you go to the library, your goal is for them to help you with something, not to be your friend. Thats the vibe Google is going for.

The reason for this goes beyond just that strange itchy feeling you get talking to a chatbot for too long. And it doesnt seem like Google is just trying to avoid super horny AI responses, either. Its more a recognition of the moment were in: large language models are suddenly everywhere, theyre far more useful than most people would have guessed, and yet they have a worrying tendency to be confidently wrong about just about everything. When that confidence comes in perfectly formed paragraphs that sound good and make sense, people are going to believe the wrong stuff.

A few executives I spoke to mentioned a tension in AI between factual and fluid. You can build a system that is factual, which is to say it offers you lots of good and grounded information. Or you can build a system that is fluid, feeling totally seamless and human. Maybe someday youll be able to have both. But right now, the two are at odds, and Google is trying hard to lean in the direction of factual. The way the company sees it, its better to be right than interesting.

Google projects a lot of confidence in its ability to be factually strong, but recent history seems to suggest otherwise

Google projects a lot of confidence in its ability to be factually strong, but recent history seems to suggest otherwise. Not only is Bard less wacky and fun than ChatGPT or Bing, but its also often less correct it makes basic mistakes in math, information retrieval, and more. The PaLM 2 model should improve some of that, but Google certainly hasnt solved the AI lies problem by a long shot.

Theres also the question of when AI should appear at all. Sometimes its obvious: the snapshots shouldnt appear if you ask sensitive medical questions, Reid says, or if youre looking to do something illegal or harmful. But theres a wide swath of searches where AI may or may not be useful. If I search Adele, some basic summary information at the top helps; if I search Adele music videos, Im much more likely to just want the YouTube videos in the results.

Google can afford to be cautious here, Reid says, because the fail state is just Google search. So whenever the snapshot shouldnt appear, or whenever the models confidence score is low enough that it might not be more useful than the top few results, its easy to just not do anything.

Bold and responsible

Compared to the splashy launch of the new Bing or the breakneck developmental pace of ChatGPT, SGE feels awfully conservative. Its an opt-in, personality-free tool that collates and summarizes your search results. For Google, suddenly in an existential crisis over the fact that AI is changing the way people interact with technology, is that enough?

A couple of executives used the same phrase to describe the companys approach: bold and responsible. Google knows it has to move fast not only are chatbots booming in popularity, but TikTok and other platforms are stealing some of the more exploratory search out from under Google. But it also has to avoid making mistakes, giving people bad information, or creating new problems for users. To do that would be a PR disaster for Google, it would be yet more reason for people to try new products, and it would potentially destroy the business that made Google a trillion-dollar company.

So, for now, SGE remains opt-in and personality-free. Raghavan says hes comfortable playing a longer game: knee-jerk reacting to some trend is not necessarily going to be the way to go. Hes also convinced that AI is not some panacea that changes everything that 10 years from now, well all do everything through chatbots and LLMs. I think its going to be one more step, he says. Its not like, Okay, the old world went away. And were in a whole new world.

In other words, Google Bard is not the future of Google Search. But AI is. Over time, SGE will start to come out of the labs and into search results for billions of users, mingling generated information with links out to the web. It will change Googles business and probably upend parts of how the web works. If Google gets it right, it will trade 10 blue links for all the knowledge on the internet, all in one place. And hopefully telling the truth.

Related:

Read the original post:
AI is coming to Google search through Search Generative Experience - The Verge

Related Posts

    Your Full Name

    Your Email

    Your Phone Number

    Select your age (30+ only)

    Select Your US State

    Program Choice

    Confirm over 30 years old

    Yes

    Confirm that you resident in USA

    Yes

    This is a Serious Inquiry

    Yes

    Message:



    matomo tracker