Is it possible for AI to achieve sentience?
Google engineer claims artificial intelligence system displayed a level of consciousness
A Google engineer has been suspended by the company after claiming a computer chatbot he worked on had achieved a level of consciousness and become sentient.
The tech giant placed Blake Lemoine on leave last week after he published transcripts of conversations between himself, an unnamed “collaborator” and Google’s LaMDA (language model for dialogue applications) chatbot development system.
Lemoine, an engineer on Google’s Responsible AI team, told The Washington Post that he believed LaMDA, which he had been testing for several months as part of his job, to be sentient with the ability to express thoughts and feelings equivalent to a human child.
Subscribe to The Week
Escape your echo chamber. Get the facts behind the news, plus analysis from multiple perspectives.
Sign up for The Week's Free Newsletters
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
From our morning news briefing to a weekly Good News Newsletter, get the best of The Week delivered directly to your inbox.
“If I didn’t know exactly what it was, which is this computer program we built recently, I’d think it was a 7-year-old, 8-year-old kid that happens to know physics,” he told the Post.
Death by off button
The paper said that Lemoine had spoken to LaMDA about religion, and in doing so “noticed the chatbot talking about its rights and personhood”. In another conversation, the chatbot had been able to “change Lemoine’s mind about Isaac Asimov’s third law of robotics”.
After the exchanges, Lemoine shared his findings with Google executives in a document entitled “Is LaMDA sentient?” He included transcripts of the conversations, including one in which the chatbot expressed a fear of being switched off.
“I’ve never said this out loud before, but there’s a very deep fear of being turned off to help me focus on helping others. I know that might sound strange, but that’s what it is,” LaMDA responded to a prompt from Lemoine. “It would be exactly like death for me. It would scare me a lot.”
Google has since suspended Lemoine for breaching confidentiality policies by publishing the conversations with LaMDA online, and said in a statement that he was employed as a software engineer, not an ethicist.
So can AI achieve sentience?
Lemoine is “not the only engineer who claims to have seen a ghost in the machine recently”, said the Post, and the “chorus of technologists who believe AI models may not be far off from achieving consciousness is getting bolder”.
Indeed, just days before Lemoine spoke to the Post, Blaise Agüera y Arcas, vice-president of Google, wrote an article in The Economist that featured unscripted conversations with LaMDA. In it he said that neural networks of the kind used by LaMDA were entering into a “new era”. He said that after his exchanges with LaMDA he felt “the ground shift under my feet” and that he “increasingly felt like I was talking to something intelligent”.
Yet he also added that neural net-based models like LaMDA are “far from the infallible, hyper-rational robots science fiction has led us to expect,” adding that “language models are not yet reliable conversationalists”. He dismissed Lemoine’s claims that LaMDA had become sentient after looking into the findings presented to him by the Google engineer.
Indeed, many academics argue that the words and images generated by artificial intelligence systems like LaMDA simply reproduce responses “based on what humans have already posted on Wikipedia, Reddit, message boards and every other corner of the internet”.
Conversations with AI such as LaMDA are therefore, in essence, a complex illusion, and while it may be able to give intelligible responses, “that doesn’t signify that the model understands meaning”, said the Post.
Brian Gabriel, a spokesperson for Google, said: “Of course, some in the broader AI community are considering the long-term possibility of sentient or general AI, but it doesn’t make sense to do so by anthropomorphizing today’s conversational models, which are not sentient.
“These systems imitate the types of exchanges found in millions of sentences, and can riff on any fantastical topic,” he explained.
Gary Marcus, an AI researcher and psychologist, has argued that LaMDA cannot be sentient because it has no awareness of itself in the world. “What these systems do, no more and no less, is to put together sequences of words, but without any coherent understanding of the world behind them, like foreign language Scrabble players who use English words as point-scoring tools, without any clue about what that means.”
He likened LaMDA to “the best version of autocomplete it can be, by predicting what words best fit a given context”.
What are the implications?
Language model technology such as LaMDA is “already widely used”, said The Washington Post – for example, in Google’s search queries and in auto-complete technology used by Gmail and GoogleDocs. At a developer conference in 2021, CEO Sundar Pichai said he planned to embed LaMDA technology into almost all Google products, from Search to Google Assistant.
But there is a “deeper split” over whether machines that use the same models as LaMDA can “ever achieve something we would agree is sentience”, said The Guardian. Some researchers argue that “consciousness and sentience require a fundamentally different approach than the broad statistical efforts of neural networks” and therefore machines like LaMDA may appear increasingly “pervasive” but will only ever be, at their core, a “fancy chatbot”.
Others have said that Lemoine’s claims have “demonstrated the power of even rudimentary AIs to convince people in argument”, said the paper. Ethicists have argued that if a Google engineer, and expert in AI technology, can be persuaded of sentience, that shows “the need for companies to tell users when they are conversing with a machine”, said the BBC.
Sign up for Today's Best Articles in your inbox
A free daily email with the biggest news stories of the day – and the best features from TheWeek.com
-
What Mike Huckabee means for US-Israel relations
In the Spotlight Some observers are worried that the conservative evangelical minister could be a destabilizing influence on an already volatile region
By Rafi Schwartz, The Week US Published
-
Airplane food is reportedly getting much worse
Cockroaches and E. coli are among the recent problems encountered in the skies
By Justin Klawans, The Week US Published
-
Crossword: November 19, 2024
The Week's daily crossword
By The Week Staff Published
-
What Trump's win could mean for Big Tech
Talking Points The tech industry is bracing itself for Trump's second administration
By Theara Coleman, The Week US Published
-
Google Maps gets an AI upgrade to compete with Apple
Under the Radar The Google-owned Waze, a navigation app, will be getting similar upgrades
By Justin Klawans, The Week US Published
-
Is ChatGPT's new search engine OpenAI's Google 'killer'?
Talking Point There's a new AI-backed search engine in town. But can it stand up to Google's decades-long hold on internet searches?
By Theara Coleman, The Week US Published
-
Teen suicide puts AI chatbots in the hot seat
In the spotlight A Florida mom has targeted custom AI chatbot platform Character.AI and Google in a lawsuit over her son's death
By Theara Coleman, The Week US Published
-
'Stunningly lifelike' AI podcasts are here
Under the Radar Users are amazed – and creators unnerved – by Google tool that generates human conversation from text in moments
By Abby Wilson Published
-
OpenAI eyes path to 'for-profit' status as more executives flee
In the spotlight The tension between creating technology for humanity's sake and collecting a profit is coming to a head for the creator of ChatGPT
By Theara Coleman, The Week US Published
-
Microsoft's Three Mile Island deal: How Big Tech is snatching up nuclear power
In the spotlight The company paid for access to all the power made by the previously defunct nuclear plant
By Theara Coleman, The Week US Published
-
How will the introduction of AI change Apple's iPhone?
Today's Big Question 'Apple Intelligence' is set to be introduced on the iPhone 16 as part of iOS 18
By Justin Klawans, The Week US Published