小程序
传感搜
传感圈

The Assumptions You Bring into Conversation with an AI Bot Influence What It Says

2023-10-09 02:03:26
关注

Do you think artificial intelligence will change our lives for the better or threaten the existence of humanity? Consider carefully—your position on this may influence how generative AI programs such as ChatGPT respond to you, prompting them to deliver results that align with your expectations.

“AI is a mirror,” says Pat Pataranutaporn, a researcher at the M.I.T. Media Lab and co-author of a new study that exposes how user bias drives AI interactions. In it, researchers found that the way a user is “primed” for an AI experience consistently impacts the results. Experiment subjects who expected a “caring” AI reported having a more positive interaction, while those who presumed the bot to have bad intentions recounted experiencing negativity—even though all participants were using the same program.

“We wanted to quantify the effect of AI placebo, basically,” Pataranutaporn says. “We wanted to see what happened if you have a certain imagination of AI: How would that manifest in your interaction?” He and his colleagues hypothesized that AI reacts with a feedback loop: if you believe an AI will act a certain way, it will.

To test this idea, the researchers divided 300 participants into three groups and asked each person to interact with an AI program and assess its ability to deliver mental health support. Before starting, those in the first group were told the AI they would be using had no motives—it was just a run-of-the-mill text completion program. The second set of participants were told their AI was trained to have empathy. The third group was warned that the AI in question was manipulative and that it would act nice merely to sell a service. But in reality, all three groups encountered an identical program. After chatting with the bot for one 10- to 30-minute session, the participants were asked to evaluate whether it was an effective mental health companion.

The results suggest that the participants’ preconceived ideas affected the chatbot’s output. In all three groups, the majority of users reported a neutral, positive or negative experience in line with the expectations the researchers had planted. “When people think that the AI is caring, they become more positive toward it,” Pataranutaporn explains. “This creates a positive reinforcement feedback loop where, at the end, the AI becomes much more positive, compared to the control condition. And when people believe that the AI was manipulative, they become more negative toward the AI—and it makes the AI become more negative toward the person as well.”

This impact was absent, however, in a simple rule-based chatbot, as opposed to a more complex one that used generative AI. While half the study participants interacted with a chatbot that used GPT-3, the other half used the more primitive chatbot ELIZA, which does not rely on machine learning to generate its responses. The expectation effect was seen with the former bot but not the latter one. This suggests that the more complex the AI, the more reflective the mirror that it holds up to humans.

The study intimates that AI aims to give people what they want—whatever that happens to be. As Pataranutaporn puts it, “A lot of this actually happens in our head.” His team’s work was published in Nature on Monday.

According to Nina Beguš, a researcher at the University of California, Berkeley, and author of the upcoming book Artificial Humanities: A Fictional Perspective on Language in AI, who was not involved in the M.I.T. Media Lab paper, it is “a good first step. Having these kinds of studies, and further studies about how people will interact under certain priming, is crucial.”

Both Beguš and Pataranutaporn worry about how human presuppositions about AI—derived largely from popular media such as the films Her and Ex Machina, as well as classic stories such as the myth of Pygmalion—will shape our future interactions with it. Beguš’s book examines how literature across history has primed our expectations regarding AI.

“The way we build them right now is: they are mirroring you,” she says. “They adjust to you.” In order to shift attitudes toward AI, Beguš suggests that art containing more accurate depictions of the technology is necessary. “We should create a culture around it,” she says.

“What we think about AI came from what we see in Star Wars or Blade Runner or Ex Machina,” Pataranutaporn says. “This ‘collective imagination’ of what AI could be, or should be, has been around. Right now, when we create a new AI system, we’re still drawing from that same source of inspiration.”

That collective imagination can change over time, and it can also vary depending on where people grew up. “AI will have different flavors in different cultures,” Beguš says. Pataranutaporn has firsthand experience with that. “I grew up with a cartoon, Doraemon, about a cool robot cat who helped a boy who was a loser in ... school,” he says. Because Pataranutaporn was familiar with a positive example of a robot, as opposed to a portrayal of a killing machine, “my mental model of AI was more positive,” he says. “I think in ... Asia people have more of a positive narrative about AI and robots—you see them as this companion or friend.” Knowing how AI “culture” influences AI users can help ensure that the technology delivers desirable outcomes, Pataranutaporn adds. For instance, developers might design a system to seem more positive in order to bolster positive results. Or they could program it to use more straightforward delivery, providing answers like a search engine does and avoiding talking about itself as “I” or “me” in order to limit people from becoming emotionally attached to or overly reliant on the AI.

This same knowledge, however, can also make it easier to manipulate AI users. “Different people will try to put out different narratives for different purposes,” Pataranutaporn says. “People in marketing or people who make the product want to shape it a certain way. They want to make it seem more empathetic or trustworthy, even though the inside engine might be super biased or flawed.” He calls for something analogous to a “nutrition label” for AI, which would allow users to see a variety of information—the data on which a particular model was trained, its coding architecture, the biases that have been tested, its potential misuses and its mitigation options—in order to better understand the AI before deciding to trust its output.

“It’s very hard to eliminate biases,” Beguš says. “Being very careful in what you put out and thinking about potential challenges as you develop your product is the only way.”

“A lot of conversation on AI bias is on the responses: Does it give biased answers?” Pataranutaporn says. “But when you think of human-AI interaction, it’s not just a one-way street. You need to think about what kind of biases people bring into the system.”

参考译文
与人工智能聊天机器人对话时,你所持的假设会影响它所说的内容。
你认为人工智能是会让我们的生活变得更好,还是会威胁人类的存在?请仔细考虑一下,你对这一问题的立场可能会在某种程度上影响像ChatGPT这样的生成式人工智能程序对你的回应方式,进而促使它们提供更符合你预期的结果。麻省理工媒体实验室的研究员帕特·帕塔努塔普恩(Pat Pataranutaporn)说:“人工智能就像一面镜子。” 他与人合著的一篇新研究揭示了用户偏见是如何驱动人工智能交互的。在这项研究中,研究人员发现,用户在使用AI之前的心理准备方式会持续影响最终结果。那些期望与“有同理心”的人工智能交互的实验参与者,普遍报告了更积极的体验;而那些认为AI意图不良的参与者则报告了消极体验——尽管所有实验对象使用的程序完全相同。“我们想量化所谓的‘人工智能安慰剂效应’,”帕塔努塔普恩说,“我们想看看,如果你对人工智能怀有某种想象,这种想象会怎样在你的交互中体现出来。”他和他的同事假设,人工智能会以一种反馈循环的方式回应人:如果你相信某个人工智能会以某种方式表现,它就会以你预期的方式表现出来。为了验证这一观点,研究人员将300名参与者分为三组,并让每组成员与一个AI程序互动,并评估它在提供心理健康支持方面的能力。在开始之前,第一组人被告知他们将要使用的AI没有特定动机,只是一个普通的文本生成程序。第二组则被告知,AI是训练成具有同理心的。第三组被提醒说,这个AI是操纵型的,只是为了推销服务而假装友好。但事实上,所有三组人最终都使用的是同一个程序。在与AI进行了10到30分钟的简短对话后,参与者被要求评估这个AI是否是一个有效的心理健康助手。结果显示,参与者的先入之见确实影响了AI的输出结果。在三组中,大多数用户的体验——中性、积极或消极——都与研究人员事先植入的期望一致。“当人们认为AI是有同理心的,他们就会对它更积极。”帕塔努塔普恩解释道,“这会形成一种积极强化的反馈循环,最终,与对照组相比,AI表现出更多的积极倾向。而当人们认为AI是操纵型的,他们也会对AI更消极——而这也会使AI对人也表现出更消极的态度。”然而,这种影响在基于简单规则的聊天机器人中并不存在,而在更复杂的生成式AI中则明显出现。研究中一半的参与者使用了采用GPT-3的聊天机器人,而另一半则使用了较为原始的ELIZA聊天机器人,后者并不依赖机器学习来生成回应。期望效应只在前者身上出现,而在后者身上没有。这表明,AI越复杂,它所呈现给人类的“镜像”就越真实。这项研究暗示,AI总是试图满足人们的需求,无论这些需求是什么。正如帕塔努塔普恩所说,“很多东西其实都发生在我们的脑海中。”他团队的研究成果于周一发表在《自然》杂志上。加州大学伯克利分校的研究员尼娜·贝古什(Nina Beguš)并未参与麻省理工媒体实验室的研究,她正在撰写即将出版的《人工智能人文学:从虚构角度审视AI中的语言》一书。她认为这项研究“是一个不错的开端。进行这样的研究,以及进一步研究人们在特定预期影响下会如何与AI互动,至关重要。”贝古什和帕塔努塔普恩都担心,人类对AI的预设(大多来自流行媒体,如电影《她》和《机械姬》,以及经典故事如皮格马利翁神话)将如何影响我们与AI的未来互动。贝古什的书探讨了历史上文学如何影响了人们对AI的期望。“我们现在构建AI的方式是:它们在模仿你,”她说,“它们会根据你进行调整。”为了改变人们对AI的态度,贝古什建议需要创作更多更准确描绘AI技术本质的艺术作品。“我们需要建立一种围绕AI的文化,”她说。帕塔努塔普恩也表示:“我们对AI的看法大多来自于《星球大战》《银翼杀手》或《机械姬》等影视作品。”他补充道,“我们对AI可能或应该是什么的这种‘集体想象’,早已存在。如今,当我们开发新的人工智能系统时,我们依然借鉴着同样的灵感来源。”这种集体想象会随着时间而改变,也会根据人们成长的环境而有所不同。“AI在不同文化中会有不同的风格,”贝古什说。帕塔努塔普恩对此深有体会。“我小时候看过一部叫《哆啦A梦》的动画片,里面是一只酷酷的机器人猫,它帮助一个在学校表现糟糕的男孩,”他说。因为帕塔努塔普恩从小接触到的是一个正面的机器人例子,而不是那种杀人机器的描绘,“我对AI的心理模型更加积极,”他说,“我认为……在亚洲,人们对AI和机器人更持积极态度——你们把它们看作是伴侣或朋友。”帕塔努塔普恩补充说,了解AI“文化”如何影响用户,有助于确保这项技术带来理想的结果。例如,开发者可以设计系统,使其表现得更积极,以增强正面结果。或者,他们可以编写系统,使其以更直接的方式提供答案,像搜索引擎那样工作,并避免使用“我”或“我”这样的第一人称,以防止人们对AI产生情感依赖或过度依赖。然而,同样的知识也可以被用来操纵AI用户。“不同的人会出于不同的目的,传播不同的叙事方式,”帕塔努塔普恩说,“做营销的人或产品开发者,会希望以特定方式塑造AI。他们希望让AI看起来更有同理心或更值得信赖,即使它的内部引擎可能极度偏见或存在缺陷。”他呼吁建立类似AI“营养成分标签”的东西,让使用者能够看到各种信息——某个模型所用的训练数据、代码架构、经过测试的偏见、潜在的误用以及缓解方案——以便在决定信任AI的输出结果之前更好地了解它。“要完全消除偏见是很难的,”贝古什说,“开发产品时非常谨慎地考虑你所发布的内容,并思考潜在挑战,才是唯一的方法。”“关于AI偏见的讨论大多集中在回应上:它是否会给出有偏见的答案?”帕塔努塔普恩说,“但当你考虑人与AI的交互时,这并不是单向的。你还必须思考人们带入这个系统的是什么样的偏见。”
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

广告

scientific

这家伙很懒,什么描述也没留下

关注

点击进入下一篇

循迹网络:深度造假与新闻真实体制

提取码
复制提取码
点击跳转至百度网盘