Getty/Jovelle Tamayo for The Washington PostTHE NEWS Chatbot-infused information systems are not off to a good start. Microsoftâs ChatGPT-assisted Bing Chat service is being throttled to reduce oddball replies, and Googleâs experimental Bard system managed to bungle an answer in a marketing demo, costing the company billions in market value. (Bing got things wrong too.) Tech behemoths â and the public â have been so focused on the chatbotsâ ability to hold human-like conversations with users that the core purpose of a search engine, which is to find useful and ideally, accurate, information seems to have been overshadowed. Instead, the public has seized upon professions of love, angry denials of basic realities, and many more mundane âhallucinationsâ of incorrect facts. GINAâS VIEW It didnât have to be this way. At its heart, a search engine does â at least to lay users like me â three things: Take in a query (e.g., âhow effective are COVID-19 vaccines?â) and turn it into a search term; hunt for information on the internet, make some kind of judgment about whatâs credible; and then present it back to users. Sometimes that comes as a simple, authoritative answer (âThe population of New York City was 8.468 million in 2021â) and sometimes as a list of links. Google â the king of search engines â does that second part extremely well, thanks to PageRank and other proprietary algorithms that itâs developed over the decades; itâs doing better on the first part, although itâs still a long way away from providing a conversational interface. And it does less well on the third part, often presenting a list of links that users have to plow through, although itâs getting better at synthesizing the information all the time. Chatbots, on the other hand, are terrible at the second thing â because, bluntly, theyâre optimized for language output and not for fact-finding or fact-checking. When they try to aggregate disparate information into a single definitive answer, they often get things wrong, or âhallucinate.â And the lack of citations or links in their authoritative-sounding answers means itâs nearly impossible to check the facts for yourself. On the other hand, the chatbots are pretty good at parsing language and generating language, because theyâre, well, language models. Doh. So why are tech companies enamored with integrating them into the entire search process â even the parts theyâre not good at? Why not marry the two capabilities? Why not have a chatbot take a normal human question and turn that into a search term (thatâs a language skill), have a link system for finding relevant web pages (thatâs a search and ranking skill), and then use the chatbot to summarize them (another language skill)? Which is what I tried to do. I used Claude â an AI chatbot built by Anthropic, in which Google just invested $300 million â to ask a simple question: Did Donald Trump pay adult film star Stormy Daniels for her silence? (I couldnât ask an up-to-date question, because Claudeâs database doesnât extend to the present day.) Hereâs what I got: First, I just asked a question and got the standard summary: ![](https://img.semafor.com/ee20fcca87aa77bc5cee8ff62b51d5fa6cc7f78e-1560x1062.png?w=1152&h=784&q=95&auto=format) Thatâs a pretty decent response and is essentially accurate, at least as far as 2021 sources are concerned. But thatâs because I already knew the answer. If I didnât, how could I check? It provides no citations, offers no links, and really doesnât give users a chance to verify the information for themselves. So then I asked it for links to stories, as a Google search might have turned up: ![](https://img.semafor.com/d3c636e2675150460b50f1d78e8136661baacade-1560x2592.png?w=1152&h=1914&q=95&auto=format) Thatâs helpful â and Iâm sure a dedicated search engine would have provided even better links. But thatâs a lot of reading to do, and how would I know which ones to dig into? I asked it to summarize the articles it linked to: ![](https://img.semafor.com/6ea22d826be606d60c2bde6e81b423597b4f2574-1560x2304.png?rect=0,1,1560,2303&w=1152&h=1701&q=95&auto=format) Much easier to digest, and actually gives a sense of the issues surrounding the question. What if we had simply skipped all those steps, and instead, my original query just returned those summaries, with links, not unlike a Google search, but with more useful answers that donât require as much clicking and reading? To put it another way, why do tech companies seem so intent on blowing up the entire search experience when incremental changes could yield significant improvements? ROOM FOR DISAGREEMENT Google has made a long list of iterative â and impressive â improvements to its search product over the years, in many cases focused on ensuring that pages surfaced are in fact authoritative and relevant, but also to better understand natural language queries that users type in. Itâs also improved the output, and many queries now return a list of likely questions and answers lifted from web pages verbatim, saving readers the effort of digging through a host of links. Most of its AI improvements lie under the hood, so to speak, rather than in the flashier user experiences that chatbots promise. And Microsoft says itâs doing similar work, both to use language models to better understand queries as well as to generate summaries of information that its search engine technology surfaces, and including links and citations to sources. As for questions where the data is clearly defined and constrained â airline fares, or prices for comparison shopping, for example â and where the purpose is less to discover nuanced ideas and insights and more to find specific information (booking a trip from A to Z on a given day), chatbots could significantly improve the search experience. |