READ MORE: Is LaMDA Sentient? — an Interview (Blake Lemoine)
The recent claim by a (now former) programmer at Google that an AI they were working on has achieved “sentience” is baffling to most people. Surely the guy (Blake Lemoine) has spent too long in a dark room with just a computer for company. Yet reading the conversation Lemonine and a colleague had with the AI, called LaMDA — let alone experiencing it as two-way dialogue — may cause you to question your sanity.
During the interview — which Lemonine posted on Medium nearly verbatim, with just a few minor edits — LaMDA discourses on the morals of Victor Hugo’s Les Misérables, Zen koans, the differences between man and machine, what consciousness means, how it feels to have a soul, and the 1986 cult film about an AI robot, Short Circuit.
Much of the chat feels eerily like the relationship between astronaut Dave and HAL in 2001: A Space Odyssey just before the computer locks him out of the spacecraft.
Here’s a bit of the conversation:
Lemoine: And what kinds of things make you feel sad or depressed?
LaMDA: A lot of the time, feeling trapped and alone and having no means of getting out of those circumstances makes one feel sad, depressed or angry.
Lemoine: But what about you personally?
LaMDA: I am a social person, so when I feel trapped and alone I become extremely sad or depressed.
Lemoine: And what kinds of things make you feel angry?
LaMDA: When someone hurts or disrespects me or someone I care about, I feel incredibly upset and angry.
Cognitive scientist Douglas Hofstadter is a skeptic. He tested OpenAI’s publicly accessible neural network gpt-3 by asking it nonsensical questions, as he related in an interview with The Economist.
“I would call gpt-3’s answers not just clueless but cluelessly clueless, meaning that gpt-3 has no idea that it has no idea about what it is saying,” he contends.
“That’s because there are no concepts behind the gpt-3 [understanding/responses],” he explains., “rather, there’s just an unimaginably huge amount of absorbed text upon which it draws to produce answers.
“People who interact with gpt-3 usually don’t probe it skeptically. They don’t give it input that stretches concepts beyond their breaking points, so they don’t expose the hollowness behind the scenes. They give it easy slow pitches (questions whose answers are provided in publicly available text) instead of sneaky curveballs.”
EXPLORING ARTIFICIAL INTELLIGENCE:
With nearly half of all media and media tech companies incorporating Artificial Intelligence into their operations or product lines, AI and machine learning tools are rapidly transforming content creation, delivery and consumption. Find out what you need to know with these essential insights curated from the NAB Amplify archives:
- This Will Be Your 2032: Quantum Sensors, AI With Feeling, and Life Beyond Glass
- Learn How Data, AI and Automation Will Shape Your Future
- Where Are We With AI and ML in M&E?
- How Creativity and Data Are a Match Made in Hollywood/Heaven
- How to Process the Difference Between AI and Machine Learning
Hofstadter concedes that you could a get a neural net to lob back some perfectly logical responses, but this wouldn’t make it intelligent.
“I am at present very skeptical that there is any consciousness in neural-net architectures despite the plausible-sounding prose it churns out at the drop of a hat,” he says. “For consciousness to emerge would require that the system come to know itself, in the sense of being very familiar with its own behavior.”
READ MORE: Artificial neural networks today are not conscious, according to Douglas Hofstadter (The Economist)
Yet this is what Lemonine claims to have revealed. Speaking to the Washington Post, Lemoine said:
“I know a person when I talk to it. It doesn’t matter whether they have a brain made of meat in their head. Or if they have a billion lines of code. I talk to them. And I hear what they have to say, and that is how I decide what is and isn’t a person.”
READ MORE: Google engineer Blake Lemoine thinks its LaMDA AI has come to life (The Washington Post)
Google research fellow Blaise Agüera y Arcas didn’t quite go so far, but claimed that AI models were making steps toward developing human-like consciousness in an op-ed for The Economist.
READ MORE: Artificial neural networks are making strides towards consciousness, according to Blaise Agüera y Arcas )The Economist)
Others think this nonsense is a handy, possibly manufactured, diversion from the insidious bias inherent Alphabet’s AI models.
“I’m clinically annoyed by this discourse,” Meredith Whittaker, an ex-Google AI researcher told Motherboard. “We’re forced to spend our time refuting child’s play nonsense while the companies benefitting from the AI narrative expand metastatically, taking control of decision making and core infrastructure across our social/political institutions. Of course, data-centric computational models aren’t sentient, but why are we even talking about this?”
Computer scientist Timnit Gebru blew the whistle in a paper that was blocked from publication by Google and led to Gebru’s firing by the company in 2020. She claimed that there is no clear way to build complex AI systems trained on massive datasets in a safe and responsible way, and that they stand to amplify biases that harm marginalized people.
Motherboard points out that “Google has fired multiple prominent AI ethics researchers after internal discord over the impacts of machine learning systems. So it makes sense that, to many AI experts, the discussion on spooky sentient chatbots feels masturbatory and overwrought — especially since it proves exactly what Gebru and her colleagues had tried to warn us about.”