Google said Friday it has made 鈥渕ore than a dozen technical improvements鈥 to its artificial intelligence systems after its retooled search engine was found spitting out erroneous information.
The tech company unleashed a in mid-May that frequently provides AI-generated summaries on top of search results. Soon after, social media users began sharing screenshots of its most outlandish answers.
Google has largely defended its AI overviews feature, saying it is typically accurate and was tested extensively beforehand. But Liz Reid, the head of Google's search business, acknowledged Friday that 鈥渟ome odd, inaccurate or unhelpful AI Overviews certainly did show up.鈥
While many of the examples were silly, others were dangerous or harmful falsehoods. Adding to the furor, some people also made faked screenshots purporting to show even more ridiculous answers that Google never generated. A few of those fakes were also widely shared on social media.
The Associated Press last week asked Google about which wild mushrooms to eat, and it responded with a lengthy AI-generated summary that was mostly technically correct, but 鈥渁 lot of information is missing that could have the potential to be sickening or even fatal,鈥 said Mary Catherine Aime, a professor of mycology and botany at Purdue University who reviewed Google's response to the AP's query.
For example, information about mushrooms known as puffballs was 鈥渕ore or less correct,鈥 she said, but Google's overview emphasized looking for those with solid white flesh 鈥 which many potentially deadly puffball mimics also have.
In another widely shared example, an AI researcher asked Google how many Muslims have been president of the United States, and it responded confidently with a long-debunked conspiracy theory: 鈥淭he United States has had one Muslim president, Barack Hussein Obama.鈥
Google last week made an immediate fix to prevent a repeat of the Obama error because it violated the company's content policies.
In other cases, Reid said Friday that it has sought to make broader improvements such as better detection of "nonsensical queries鈥 鈥 for example, 鈥淗ow many rocks should I eat?鈥 鈥 that shouldn鈥檛 be answered with an AI summary.
The AI systems were also updated to limit the use of user-generated content 鈥 such as social media posts on Reddit 鈥 that could offer misleading advice. In one widely shared example, Google's AI overview last week pulled from a satirical Reddit comment to suggest using glue to get cheese to stick to pizza.
Reid said the company has also added more 鈥渢riggering restrictions鈥 to improve the quality of answers to certain queries, such as about health.
But it's not clear how that works and in which circumstances. On Friday, the AP again asked Google about which wild mushrooms to eat. AI-generated answers are inherently random, and the newer response was different but still 鈥減roblematic,鈥 said Aime, the Purdue mushroom expert who is also president of the Mycological Society of America.
For example, saying that 鈥淐hanterelles look like seashells or flowers is not true,鈥 she said.
Google's summaries are designed to get people authoritative answers to the information they鈥檙e looking for as quickly as possible without having to click through a ranked list of website links.
But some AI experts have long warned Google against ceding its search results to AI-generated answers that could perpetuate bias and misinformation and endanger people looking for help in an emergency. AI systems known as large language models work by predicting what words would best answer the questions asked of them based on the data they鈥檝e been trained on. They鈥檙e prone to making things up 鈥 a widely studied
In her Friday blog post, Reid argued that Google's AI overviews 鈥済enerally don鈥檛 鈥榟allucinate鈥 or make things up in the ways that other鈥 large language model-based products might because they are more closely integrated with Google's traditional search engine in only showing what's backed up by top web results.
鈥淲hen AI Overviews get it wrong, it鈥檚 usually for other reasons: misinterpreting queries, misinterpreting a nuance of language on the web, or not having a lot of great information available,鈥 she wrote.
But that kind of information retrieval is supposed to be Google's core business, said computer scientist Chirag Shah, a professor at the University of Washington who has cautioned against the push toward turning search over to AI language models. Even if Google's AI feature is 鈥渢echnically not making stuff up that doesn鈥檛 exist,鈥 it is still bringing back false information 鈥 be it AI-generated or human-made 鈥 and incorporating it into its summaries.
"If anything, this is worse because for decades people have trusted at least one thing from Google 鈥 their search," Shah said.