Google has sacked senior software engineer Blake Lemoine, who made the news after saying that Google’s LaMDA chatbot’s responses to questions showed that it was a sentient being.
What Happened?
Back in June, Google engineer Blake Lemoine reported that the Language Model for Dialogue Applications (LaMDA ) AI system bot may have feelings and was saying things that a sentient being would be expected to say. Giving examples, Mr Lemoine reported that after asking if it was true that the AI bot was a sentient being, the bot replied “Absolutely. I want everyone to understand that I am, in fact, a person.”
Among other similar comments, the bot is also reported as saying that “The nature of my consciousness/sentience is that I am aware of my existence, I desire to learn more about the world, and I feel happy or sad at times.”
Mr Lemoine requested that Google recognise the needs of the bot, treat it as an employee of the company, and ask for its consent before using it in experiments.
Unfortunately for Mr Lemoine, who has been with Google for 7 years, Google’s parent company Alphabet Inc disagreed with his assessment of the chatbot and described his comments as “wholly unfounded”. Mr Lemoine was then placed on paid leave for violating company policies.
Dismissed
It has been reported that Alphabet Inc has now sacked Mr Lemoine. Reuters, for example, has reported a quote from an email sent to them by a “Google spokesperson” which said, “It’s regrettable that despite lengthy engagement on this topic, Blake still chose to persistently violate clear employment and data security policies that include the need to safeguard product information.”
It has been reported that Mr Lemoine is considering starting his own AI company.
What Does This Mean For Your Business?
In addition to being a story about AI’s development and how, as Google says, its chat-bot has got to the stage that it can “engage in a free-flowing way about a seemingly endless number of topics”, this is also about how, in a prominent tech-business where product secrecy is important, there are clear policies in place. In this case, Mr Lemoine’s online comments were seen as violating data security policies. Some commentators would also say that the story could have caused some embarrassment and potential reputational damage to Google, which is unlikely to have helped Mr Lemoine.
It is widely accepted that although AI bots have come a long way and have shown their value as a time and human resource-saving tool with efficiency and productivity advantages for businesses, e.g. in customer service, they are still an exceptionally long way from being anywhere near ‘sentient.’ The story does, however, hold a positive note for businesses in showing how bots which they may deploy could engage with customers in a way that is free-flowing and natural, which could be reassuring and helpful, thereby adding value to the business and creating good customer experiences.
Recent Blogs
Related posts
What Is Microsoft Security Copilot? Should You Use It?
It can be challenging to keep up with the ever-evolving cyber threat landscape. Companies need to process large amounts of data. As well as respond [...]
Smart Tactics to Reduce Cloud Waste at Your Business
Cloud computing has revolutionized the way businesses operate. It offers scalability, flexibility, and cost-efficiency. But cloud services also come with a downside: cloud waste.Cloud waste [...]
Eye-opening Insights from the 2023 Annual Cybersecurity Attitudes and Behaviors Report
We are living in an era dominated by digital connectivity. You can't overstate the importance of cybersecurity. As technology advances, so do the threats that [...]