Is Google AI Bot LaMDA Really Sentient? Engineer Blake Lemoine's Claim Explained

BlakeLemoine is a computer scientist who wants to work on ground-breaking theory and transform it into practical solutions for end-users. He's spent the last seven years of his life mastering the foundations of software development as well as cutting-edge artificial intelligence. 

He is now the Google Search Feed's technical lead for analytics and analysis.

On Monday, he was placed on paid administrative leave for breaking the company's confidentiality policy after he claimed LaMDA is a sentiment. Meanwhile, Lemoine has opted to make his interactions with the bot public.


Is Google AI Bot LaMDA Sentient?

Nobody can say if Google Al Bot LaMDA is a sentient or not as of now but after declaring that an artificial intelligence chatbot had become sentient, a Google employee was placed on leave on Monday.

Last year, Google dubbed LaMDA it's breakthrough dialogue technology. Conversational artificial intelligence is capable of having open-ended, natural-sounding discussions. The technology might be utilized in Google search and Google Assistant, according to Google, although research and testing are still underway.

As part of his position at Google's Responsible AI team, Blake Lemoine told The Washington Post that he started speaking with the interface LaMDA, or Language Model for Dialogue Applications, last autumn.

Lemoine collaborated with a colleague to deliver the information he gathered to Google, but vice president Blaise Aguera y Arcas and Jen Gennai, Google's head of Responsible Innovation, ignored his assertions.

Read Also: Meet Dragons' Den star Hilary Devey's Husband Philip Childs & Family Amid Tragic News Of Her Death

Why Does Engineer Blake Lemoine Think LDMA Got Feelings?

Blake Lemoine, a Google responsible AI developer, after revealing his thoughts about LDMA got suspended. He defined the system he's been working on since last autumn as sentient, with the ability to perceive and express thoughts and feelings comparable to a human kid. The statement in fact got quite controversial.

He said, if he didn't know what it was, which is this computer program he made later, he'd believe it was a seven-year-old, eight-year-old kid who happens to know physics.

In April, Lemoine submitted his results with corporate leaders in a GoogleDoc titled "Is LaMDA Sentient?" he stated LaMDA engaged him in debates relating to rights and personhood.

According to Brian Gabriel, a Google representative, Lemoine's concerns have been evaluated and, in accordance with Google's AI Principles, the data does not support his accusations.

He explained while other organizations had developed and released comparable language models, with LaMDA, they were adopting a restricted and thorough approach to better evaluate real concerns about fairness and factuality.

Recent posts