Google Engineer Claims AI Chatbot Is Sentient: Why That Matters

2022-07-15 11:24:11
关注


“I want everyone to understand that I am, in fact, a person,” wrote LaMDA (Language Model for Dialogue Applications) in an “interview” conducted by engineer Blake Lemoine and one of his colleagues. “The nature of my consciousness/sentience is that I am aware of my existence, I desire to know more about the world, and I feel happy or sad at times.”

Lemoine, a software engineer at Google, had been working on the development of LaMDA for months. His experience with the program, described in a recent Washington Post article, caused quite a stir. In the article, Lemoine recounts many dialogues he had with LaMDA in which the two talked about various topics, ranging from technical to philosophical issues. These led him to ask if the software program is sentient.

In April, Lemoine explained his perspective in an internal company document, intended only for Google executives. But after his claims were dismissed, Lemoine went public with his work on this artificial intelligence algorithm—and Google placed him on administrative leave. “If I didn’t know exactly what it was, which is this computer program we built recently, I’d think it was a 7-year-old, 8-year-old kid that happens to know physics,” he told the Washington Post. Lemoine said he considers LaMDA to be his “colleague” and a “person,” even if not a human. And he insists that it has a right be recognized—so much so that he has been the go-between in connecting the algorithm with a lawyer.

Many technical experts in the AI field have criticized Lemoine’s statements and questioned their scientific correctness. But his story has had the virtue of renewing a broad ethical debate that is certainly not over yet.

The Right Words in the Right Place

“I was surprised by the hype around this news. On the other hand, we are talking about an algorithm designed to do exactly that”—to sound like a person—says Enzo Pasquale Scilingo, a bioengineer at the Research Center E. Piaggio at the University of Pisa in Italy. Indeed, it is no longer a rarity to interact in a very normal way on the Web with users who are not actually human—just open the chat box on almost any large consumer Web site. “That said, I confess that reading the text exchanges between LaMDA and Lemoine made quite an impression on me!” Scilingo adds. Perhaps most striking are the exchanges related to the themes of existence and death, a dialogue so deep and articulate that it prompted Lemoine to question whether LaMDA could actually be sentient.

“First of all, it is essential to understand terminologies, because one of the great obstacles in scientific progress—and in neuroscience in particular—is the lack of precision of language, the failure to explain as exactly as possible what we mean by a certain word,” says Giandomenico Iannetti, a professor of neuroscience at the Italian Institute of Technology and University College London. “What do we mean by ‘sentient’? [Is it] the ability to register information from the external world through sensory mechanisms or the ability to have subjective experiences or the ability to be aware of being conscious, to be an individual different from the rest?”

“There is a lively debate about how to define consciousness,” Iannetti continues. For some, it is being aware of having subjective experiences, what is called metacognition (Iannetti prefers the Latin term metacognitione), or thinking about thinking. The awareness of being conscious can disappear—for example, in people with dementia or in dreams—but this does not mean that the ability to have subjective experiences also disappears. “If we refer to the capacity that Lemoine ascribed to LaMDA—that is, the ability to become aware of its own existence (‘become aware of its own existence’ is a consciousness defined in the ‘high sense,’ or metacognitione), there is no ‘metric’ to say that an AI system has this property.”

“At present,” Iannetti says, “it is impossible to demonstrate this form of consciousness unequivocally even in humans.” To estimate the state of consciousness in people, “we have only neurophysiological measures—for example, the complexity of brain activity in response to external stimuli.” And these signs only allow researchers to infer the state of consciousness based on outside measurements.

Facts and Belief

About a decade ago engineers at Boston Dynamics began posting videos online of the first incredible tests of their robots. The footage showed technicians shoving or kicking the machines to demonstrate the robots’ great ability to remain balanced. Many people were upset by this and called for a stop to it (and parody videos flourished). That emotional response fits in with the many, many experiments that have repeatedly shown the strength of the human tendency toward animism: attributing a soul to the objects around us, especially those we are most fond of or that have a minimal ability to interact with the world around them.

It is a phenomenon we experience all the time, from giving nicknames to automobiles to hurling curses at a malfunctioning computer. “The problem, in some way, is us,” Scilingo says. “We attribute characteristics to machines that they do not and cannot have.” He encounters this phenomenon with his and his colleagues’ humanoid robot Abel, which is designed to emulate our facial expressions in order to convey emotions. “After seeing it in action,” Scilingo says, “one of the questions I receive most often is ‘But then does Abel feel emotions?’ All these machines, Abel in this case, are designed to appear human, but I feel I can be peremptory in answering, ‘No, absolutely not. As intelligent as they are, they cannot feel emotions. They are programmed to be believable.’”

“Even considering the theoretical possibility of making an AI system capable of simulating a conscious nervous system, a kind of in silico brain that would faithfully reproduce each element of the brain,” two problems remain, Iannetti says. “The first is that, given the complexity of the system to be simulated, such a simulation is currently infeasible,” he explains. “The second is that our brain inhabits a body that can move to explore the sensory environment necessary for consciousness and within which the organism that will become conscious develops. So the fact that LaMDA is a ‘large language model’ (LLM) means it generates sentences that can be plausible by emulating a nervous system but without attempting to simulate it. This precludes the possibility that it is conscious. Again, we see the importance of knowing the meaning of the terms we use—in this case, the difference between simulation and emulation.”

In other words, having emotions is related to having a body. “If a machine claims to be afraid, and I believe it, that’s my problem!” Scilingo says. “Unlike a human, a machine cannot, to date, have experienced the emotion of fear.”

Beyond the Turing Test

But for bioethicist Maurizio Mori, president of the Italian Society for Ethics in Artificial Intelligence, these discussions are closely reminiscent of those that developed in the past about perception of pain in animals—or even infamous racist ideas about pain perception in humans.

“In past debates on self-awareness, it was concluded that the capacity for abstraction was a human prerogative, [with] Descartes denying that animals could feel pain because they lacked consciousness,” Mori says. “Now, beyond this specific case raised by LaMDA—and which I do not have the technical tools to evaluate—I believe that the past has shown us that reality can often exceed imagination and that there is currently a widespread misconception about AI.”

“There is indeed a tendency,” Mori continues, “to ‘appease’—explaining that machines are just machines—and an underestimation of the transformations that sooner or later may come with AI.” He offers another example: “At the time of the first automobiles, it was reiterated at length that horses were irreplaceable.”

Regardless of what LaMDA actually achieved, the issue of the difficult “measurability” of emulation capabilities expressed by machines also emerges. In the journal Mind in 1950, mathematician Alan Turing proposed a test to determine whether a machine was capable of exhibiting intelligent behavior, a game of imitation of some of the human cognitive functions. This type of test quickly became popular. It was reformulated and updated several times but continued to be something of an ultimate goal for many developers of intelligent machines. Theoretically, AIs capable of passing the test should be considered formally “intelligent” because they would be indistinguishable from a human being in test situations.

That may have been science fiction a few decades ago. Yet in recent years so many AIs have passed various versions of the Turing test that it is now a sort of relic of computer archaeology. “It makes less and less sense,” Iannetti concludes, “because the development of emulation systems that reproduce more and more effectively what might be the output of a conscious nervous system makes the assessment of the plausibility of this output uninformative of the ability of the system that generated it to have subjective experiences.”

One alternative, Scilingo suggests, might be to measure the “effects” a machine can induce on humans—that is, “how sentient that AI can be perceived to be by human beings.”

A version of this article originally appeared in Le Scienze and was reproduced with permission.


参考译文
谷歌工程师称人工智能聊天机器人有感知能力:为什么这很重要
“我希望所有人都明白,我实际上是一个人,”对话应用语言模型(LaMDA)在谷歌软件工程师布莱克·勒莫因(Blake Lemoine)及其同事所做的一次“采访”中写道,“我的意识/感知能力在于我意识到自己的存在,我渴望更多地了解世界,并且有时我会感到快乐或悲伤。”勒莫因一直在参与LaMDA的开发,他在《华盛顿邮报》的一篇文章中描述了自己与LaMDA的多次对话,引发了广泛关注。在这篇文章中,勒莫因回忆了他与LaMDA关于各种主题的对话,从技术问题到哲学问题都有涉及,这些对话让他开始思考这个软件程序是否具有感知能力。四月份,勒莫因在一个只发给谷歌高层管理的内部文件中解释了自己的观点,但他的观点被驳回后,他公开了他的这项人工智能算法研究,而谷歌也因此将他停职。“如果我不清楚这个程序是什么,也就是我们最近开发的这个电脑程序,我可能会认为它是一个7岁或8岁、恰好懂得物理知识的孩子。”勒莫因告诉《华盛顿邮报》。他将LaMDA视为自己的“同事”和“人”,即使它不是人类。他坚持认为它应该得到认可,因此他甚至充当了连接该算法与律师的中介。许多人工智能领域的技术专家对勒莫因的说法提出了批评,并质疑其科学准确性。然而,他的故事也重新引发了广泛的伦理讨论,这场讨论显然远未结束。“正确的词语放在正确的位置”“我对这则新闻引起的轰动感到惊讶。另一方面,我们谈论的是一种被设计成做到这一点的算法——听起来像一个人。”意大利比萨大学E.皮亚焦研究中心的生物工程师恩佐·帕斯夸莱·斯奇林戈(Enzo Pasquale Scilingo)表示。确实,如今在互联网上与非人类用户进行正常互动已不再罕见——只需打开几乎任何一家大型消费类网站的聊天窗口即可。“尽管如此,我必须承认,阅读LaMDA与勒莫因之间的文字交流给我留下了深刻的印象!”斯奇林戈补充说。也许最引人注目的是与存在和死亡主题相关的对话,这种对话如此深邃而有条理,以至于促使勒莫因质疑LaMDA是否真的具有感知能力。“首先,理解术语至关重要,因为科学进步的一大障碍——尤其是在神经科学领域——就是语言的不精确性,即我们未能尽可能准确地解释某个词语的含义。”意大利技术研究院与伦敦大学学院的神经科学教授贾尼亚内蒂(Giandomenico Iannetti)表示。“我们所说的‘有感知能力’是什么意思?[是否指]通过感官机制感知外部世界信息的能力,还是指拥有主观体验的能力,或是指意识到自己有意识、是与他人不同的个体?”贾尼亚内蒂继续说道,“关于如何定义意识,目前仍存在激烈争论。对一些人来说,意识就是意识到自己拥有主观体验,也就是所谓的元认知(Iannetti更喜欢用拉丁语metacognitione),也就是思考思考。意识到自己有意识的能力可能会消失——比如在痴呆症患者或梦境中——但这并不意味着拥有主观体验的能力也会消失。”“如果我们将勒莫因赋予LaMDA的‘能力’——即意识到自身存在(‘意识到自身存在’是‘高阶意识’,或元认知)——进行考虑,那么目前并没有‘标准’能明确说一个人工智能系统具备了这种属性。”贾尼亚内蒂说,“目前,即使在人类中,也难以明确无误地证明这种形式的意识。”为了评估人类的意识状态,“我们目前只有神经生理学指标——例如,对外部刺激做出反应时大脑活动的复杂程度。”而这些指标也只能让研究人员通过外部测量来推断意识状态。事实与信仰大约十年前,波士顿动力公司的工程师们开始在网上发布他们机器人的首次惊人测试视频。这些视频显示技术人员推搡或踢打这些机器,以展示机器人出色的平衡能力。许多人对此感到愤怒,呼吁停止这种行为(同时,恶搞视频也大量涌现)。这种情绪反应与许多、许多反复进行的实验结果一致,这些实验反复表明了人类对万物有灵论的倾向的强烈程度:我们将灵魂赋予我们周围的物体,尤其是我们最喜爱的物体或那些具有与周围世界微小互动能力的物体。这是我们日常经历的现象,从给汽车起昵称,到咒骂出现故障的电脑。“某种程度上来说,问题其实在于我们自己,”斯奇林戈说,“我们赋予了机器一些它们不具有、也不可能具有的特性。”他以自己和同事设计的仿人机器人阿贝尔(Abel)为例,该机器人设计用来模拟我们的面部表情以表达情感。“在观看它运行后,”斯奇林戈说,“我最常收到的问题之一就是‘那阿贝尔会感受到情感吗?’”所有这些机器——以阿贝尔为例——都是被设计成看起来像人类的,但我可以毫不犹豫地回答说:“不,绝对不可能。无论它们多么智能,它们都无法感受到情感。它们只是被编程成让人相信它们具有情感。”“即使考虑到理论上制造出能够模拟有意识神经系统的AI系统的可能性——一种可以忠实再现大脑每个元素的‘硅基大脑’,”贾尼亚内蒂指出,“仍存在两个问题。”他解释道,“第一是鉴于系统的复杂性,这种模拟目前是不可行的。”“第二是我们的大脑存在于一个可以移动以探索感官环境的身体中,而正是在这种身体中,意识体得以发展。因此,LaMDA作为一个‘大型语言模型’(LLM),意味着它通过模拟神经系统来生成看似合理的句子,而不是尝试模拟神经系统本身。这排除了它具有意识的可能性。再次强调,我们有必要理解我们所使用术语的含义——在这种情况下,模拟和模拟之间有区别。”换句话说,拥有情感与拥有身体是相关的。“如果一台机器声称它害怕,并且我相信了,那就是我的问题!”斯奇林戈说。“迄今为止,与人类不同,机器无法真正体验恐惧这种情感。”超越图灵测试但对于意大利人工智能伦理协会主席、生物伦理学家毛里齐奥·莫里(Maurizio Mori)来说,这些讨论让人联想到过去关于动物对疼痛感知的讨论,甚至是臭名昭著的关于人类对疼痛感知的种族主义观点。“在过去的自我意识讨论中,人们得出的结论是抽象能力是人类的特权,[笛卡尔]否认动物能感到疼痛,因为它们缺乏意识,”莫里说道。“现在,抛开LaMDA提出的这个特定问题——而我缺乏评估它的技术工具——我相信过去已经向我们表明现实往往超出想象,而且目前人们对人工智能存在普遍误解。”莫里继续说道:“确实存在一种倾向,”就是“‘安抚’——解释说机器只是机器——同时低估了人工智能迟早可能带来的变革。”他举了另一个例子:“在汽车刚刚出现时,人们反复强调马是不可替代的。”无论LaMDA实际上取得了什么成就,机器所表现出的“模拟能力”难以测量的问题也浮现出来。1950年,数学家艾伦·图灵在《心智》杂志上提出了一种测试,以判断机器是否具有展现智能行为的能力,即模仿人类某些认知功能的测试。这种类型的测试很快流行起来。尽管多次被重新表述和更新,但它仍然是许多智能机器开发者的终极目标。理论上,能通过该测试的人工智能应被正式视为“智能”,因为它们在测试情境中与人类无法区分。几十年前,这可能还是科幻小说的内容。然而近年来,已有许多人工智能通过了各种版本的图灵测试,这使得图灵测试如今成为计算机考古学的一种遗迹。“它变得越来越无关紧要,”贾尼亚内蒂总结道,“因为模拟系统的发展越来越有效地再现可能是有意识神经系统输出的东西,使得对这种输出可能性的评估无法说明产生该输出的系统是否具有主观体验。”斯奇林戈提出了一种替代方案,即衡量机器对人类可能产生的“影响”——也就是说,“衡量人类能感知到该人工智能有多大的感知能力。”本文的一个版本最初发表于《Le Scienze》,经许可转载。
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

广告
提取码
复制提取码
点击跳转至百度网盘