AI Chatbots Are Coming to Search Engines. Can You Trust Them?

2023-02-19 19:34:58
关注

Months after the chatbot ChatGPT wowed the world with its uncanny ability to write essays and answer questions like a human, artificial intelligence (AI) is coming to Internet search.

Three of the world’s biggest search engines — Google, Bing and Baidu — last week said they will be integrating ChatGPT or similar technology into their search products, allowing people to get direct answers or engage in a conversation, rather than merely receiving a list of links after typing in a word or question. How will this change the way people relate to search engines? Are there risks to this form of human–machine interaction?

Microsoft’s Bing uses the same technology as ChatGPT, which was developed by OpenAI of San Francisco, California. But all three companies are using large language models (LLMs). LLMs create convincing sentences by echoing the statistical patterns of text they encounter in a large database. Google’s AI-powered search engine, Bard, announced on 6 February, is currently in use by a small group of testers. Microsoft’s version is widely available now, although there is a waiting list for unfettered access. Baidu’s ERNIE Bot will be available in March.

Before these announcements, a few smaller companies had already released AI-powered search engines. “Search engines are evolving into this new state, where you can actually start talking to them, and converse with them like you would talk to a friend,” says Aravind Srinivas, a computer scientist in San Francisco who last August co-founded Perplexity — an LLM-based search engine that provides answers in conversational English.

Changing trust

The intensely personal nature of a conversation — compared with a classic Internet search — might help to sway perceptions of search results. People might inherently trust the answers from a chatbot that engages in conversation more than those from a detached search engine, says Aleksandra Urman, a computational social scientist at the University of Zurich in Switzerland.

A 2022 study1 by a team based at the University of Florida in Gainesville found that for participants interacting with chatbots used by companies such as Amazon and Best Buy, the more they perceived the conversation to be human-like, the more they trusted the organization.

That could be beneficial, making searching faster and smoother. But an enhanced sense of trust could be problematic given that AI chatbots make mistakes. Google’s Bard flubbed a question about the James Webb Space Telescope in its own tech demo, confidently answering incorrectly. And ChatGPT has a tendency to create fictional answers to questions to which it doesn’t know the answer — known by those in the field as hallucinating.

A Google spokesperson said Bard’s error “highlights the importance of a rigorous testing process, something that we’re kicking off this week with our trusted-tester programme”. But some speculate that, rather than increasing trust, such errors, assuming they are discovered, could cause users to lose confidence in chat-based search. “Early perception can have a very large impact,” says Sridhar Ramaswamy, a computer scientists based in Mountain View, California and chief executive of Neeva, an LLM-powered search engine launched in January. The mistake wiped $100 billion from Google’s value as investors worried about the future and sold stock.

Lack of transparency

Compounding the problem of inaccuracy is a comparative lack of transparency. Typically, search engines present users with their sources — a list of links — and leave them to decide what they trust. By contrast, it’s rarely known what data an LLM trained on — is it Encyclopaedia Britannica or a gossip blog?

“It’s completely untransparent how [AI-powered search] is going to work, which might have major implications if the language model misfires, hallucinates or spreads misinformation,” says Urman.

If search bots make enough errors, then, rather than increasing trust with their conversational ability, they have the potential to unseat users’ perceptions of search engines as impartial arbiters of truth, Urman says.

She has conducted as-yet unpublished research that suggests current trust is high. She examined how people perceive existing features that Google uses to enhance the search experience, known as ‘featured snippets’, in which an extract from a page that is deemed particularly relevant to the search appears above the link, and ‘knowledge panels’ — summaries that Google automatically generates in response to searches about, for example, a person or organization. Almost 80% of people Urman surveyed deemed these features accurate, and around 70% thought they were objective.

Chatbot-powered search blurs the distinction between machines and humans, says Giada Pistilli, principal ethicist at Hugging Face, a data-science platform in Paris that promotes the responsible use of AI. She worries about how quickly companies are adopting AI advances: “We always have these new technologies thrown at us without any control or an educational framework to know how to use them.”

This article is reproduced with permission and was first published on February 13 2023.

参考译文
人工智能聊天机器人即将进入搜索引擎。你能相信他们吗?
几个月前,聊天机器人ChatGPT凭借其不寻常的类人写作和回答问题能力震惊了世界,如今人工智能(AI)正进入网络搜索领域。上周,全球三大搜索引擎——谷歌(Google)、必应(Bing)和百度——宣布,它们将把ChatGPT或类似技术整合到各自的搜索产品中,使用户可以获取直接答案或进行对话,而不再仅仅在输入一个词或问题后看到一串链接。这将如何改变人们与搜索引擎的互动方式?这种人机互动形式是否存在风险?微软的必应使用的是与ChatGPT相同的技术,而ChatGPT由美国旧金山的OpenAI开发。但所有三家公司都在使用大型语言模型(LLMs)。LLMs通过模仿大型数据库中的文本统计模式,生成看起来真实可信的句子。谷歌于2023年2月6日宣布的AI驱动搜索引擎“巴德”(Bard),目前正由一小群测试者使用。微软的版本目前已广泛可用,尽管要获得无限制访问权限仍需排队等候。百度的“文心一言”(ERNIE Bot)将在3月推出。在这些公告之前,一些较小的公司已经推出了AI驱动的搜索引擎。旧金山的计算机科学家阿鲁文·斯里尼瓦斯(Aravind Srinivas)于去年8月共同创立了一家基于LLM的搜索引擎公司Perplexity,他表示:“搜索引擎正在进入一种新状态,你可以像跟朋友聊天那样与它们对话。”信任的转变与传统互联网搜索相比,对话的私密性和互动性可能会影响用户对搜索结果的看法。瑞士苏黎世大学的计算社会科学家阿莱桑德拉·乌曼(Aleksandra Urman)表示,人们可能更信任与他们对话的聊天机器人提供的答案,而不是传统搜索引擎给出的答案。2022年,由美国佛罗里达州盖恩斯维尔大学一支团队进行的一项研究发现,当参与者与亚马逊和百思买等公司使用的聊天机器人互动时,他们越觉得对话像人类,就越信任相关公司。这可能是一个优势,使搜索更加高效顺畅。但考虑到AI聊天机器人也会犯错,这种增强的信任感也可能引发问题。谷歌的Bard在自家技术演示中就曾错误地回答了关于詹姆斯·韦伯太空望远镜的问题,并自信地给出了错误答案。而ChatGPT则倾向于对其不知道的问题编造答案,业内称之为“幻觉”。谷歌的一位发言人表示,Bard的这个错误“突显了严格测试流程的重要性,我们本周正通过可信测试者计划启动相关测试”。不过,一些人认为,这类错误,一旦被发现,反而可能导致用户对基于聊天的搜索失去信心。总部位于加州山景城的计算机科学家斯里达尔·拉马斯瓦米(Sridhar Ramaswamy)是1月推出的LLM驱动搜索引擎Neeva的首席执行官,他表示:“用户最初的印象可能会产生巨大影响。”Bard的这个错误使谷歌市值蒸发了1000亿美元,投资者对未来的担忧促使他们抛售股票。透明度的缺乏除了准确性问题,AI搜索还缺乏透明度。通常,搜索引擎会向用户展示其来源——即链接列表——并让用户自行决定信任哪个。相比之下,LLMs通常不会告诉用户自己是基于什么数据训练的——是百科全书,还是八卦博客?苏黎世大学的乌曼表示:“AI驱动的搜索引擎运作方式完全不透明,这可能会带来重大影响,如果语言模型出现错误、产生幻觉或传播虚假信息。”Urman指出,如果搜索机器人犯的错误足够多,那么,它们的对话能力反而会削弱用户对搜索引擎作为事实中立仲裁者的信任。她目前尚未发表的研究表明,目前人们对搜索引擎的准确性仍抱有较高信任。她研究了用户如何看待谷歌用来提升搜索体验的两种现有功能:一种是“特色摘要”(featured snippets),即从与搜索高度相关的一篇网页中提取的摘要内容,显示在链接之上;另一种是“知识卡片”(knowledge panels),即谷歌自动为对某人或组织的搜索生成的摘要。乌曼调查的受访者中,近80%认为这些功能是准确的,约70%认为它们是客观的。巴黎的人工智能数据科学平台Hugging Face的首席伦理学家吉亚达·皮斯蒂利(Giada Pistilli)表示,聊天机器人驱动的搜索模糊了机器和人类之间的界限。她担心公司采用AI技术的速度太快来不及控制:“我们总是被层出不穷的新技术冲击着,却没有控制手段或教育框架来了解如何使用这些技术。”本文经授权转载,最初于2023年2月13日发布。
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

广告

scientific

这家伙很懒,什么描述也没留下

关注

点击进入下一篇

特斯拉在中国正式推送自动辅助导航驾驶功能

提取码
复制提取码
点击跳转至百度网盘