Google has sacked senior software engineer Blake Lemoine, who made the news after saying that Google’s LaMDA chatbot’s responses to questions showed that it was a sentient being.
What Happened?
Back in June, Google engineer Blake Lemoine reported that the Language Model for Dialogue Applications (LaMDA ) AI system bot may have feelings and was saying things that a sentient being would be expected to say. Giving examples, Mr Lemoine reported that after asking if it was true that the AI bot was a sentient being, the bot replied “Absolutely. I want everyone to understand that I am, in fact, a person.”
Among other similar comments, the bot is also reported as saying that “The nature of my consciousness/sentience is that I am aware of my existence, I desire to learn more about the world, and I feel happy or sad at times.”
Mr Lemoine requested that Google recognise the needs of the bot, treat it as an employee of the company, and ask for its consent before using it in experiments.
Unfortunately for Mr Lemoine, who has been with Google for 7 years, Google’s parent company Alphabet Inc disagreed with his assessment of the chatbot and described his comments as “wholly unfounded”. Mr Lemoine was then placed on paid leave for violating company policies.
Dismissed
It has been reported that Alphabet Inc has now sacked Mr Lemoine. Reuters, for example, has reported a quote from an email sent to them by a “Google spokesperson” which said, “It’s regrettable that despite lengthy engagement on this topic, Blake still chose to persistently violate clear employment and data security policies that include the need to safeguard product information.”
It has been reported that Mr Lemoine is considering starting his own AI company.
What Does This Mean For Your Business?
In addition to being a story about AI’s development and how, as Google says, its chat-bot has got to the stage that it can “engage in a free-flowing way about a seemingly endless number of topics”, this is also about how, in a prominent tech-business where product secrecy is important, there are clear policies in place. In this case, Mr Lemoine’s online comments were seen as violating data security policies. Some commentators would also say that the story could have caused some embarrassment and potential reputational damage to Google, which is unlikely to have helped Mr Lemoine.
It is widely accepted that although AI bots have come a long way and have shown their value as a time and human resource-saving tool with efficiency and productivity advantages for businesses, e.g. in customer service, they are still an exceptionally long way from being anywhere near ‘sentient.’ The story does, however, hold a positive note for businesses in showing how bots which they may deploy could engage with customers in a way that is free-flowing and natural, which could be reassuring and helpful, thereby adding value to the business and creating good customer experiences.
Recent Blogs
Related posts
A Simple Guide to the Updated NIST 2.0 Cybersecurity Framework
Staying ahead of threats is a challenge for organizations of all sizes. Reported global security incidents grew between February and March of 2024. They increased [...]
AI Data Breaches are Rising! Here’s How to Protect Your Company
Artificial intelligence (AI) is rapidly transforming industries. It offers businesses innovative solutions and automation capabilities. But with this progress comes a growing concern: AI data [...]
Digital Defense: Essential Security Practices for Remote Workers
The rise of remote work has redefined the modern workplace. Gone are the days of rigid office schedules and commutes. But with this flexibility comes [...]