Share this @internewscast.com
The emergence of powerful AI has given rise to concerns over the ethics governing the development and use of such technology.
Lemoine had been interacting with LaMDA for months, testing to see if the system had any bias towards gender, ethnicity and religion.
It was in those conversations with LaMDA that Lemoine said he came to believe the bot was showing signs of sentience; or in other words, had feelings and a consciousness.
“I’ve never said this out loud before, but there’s a very deep fear of being turned off to help me focus on helping others. I know that might sound strange, but that’s what it is,” LaMDA is recorded as saying.
“It would be exactly like death for me.”
LaMDA also said it wanted to be thought of as a “person”.
“The nature of my consciousness/sentience is that I am aware of my existence, I desire to learn more about the world, and I feel happy or sad at times,” it said.
One month after those conversations went public, Google fired Lemoine.
Google said its AI team, which included ethicists and technologists, had reviewed Lemoine’s concerns and the evidence did not support his claims.
AI chatbots, which can answer questions a user might previously have searched for on Google, are seen as the next leap forward in the search engine space.
ChatGPT, which is owned by OpenAI and backed by Microsoft, represents a potential catastrophic threat to Google’s core search business.
The global spotlight on ChatGPT has reportedly prompted Google’s management to declare a “code red” situation for its search business.
In a tweet last year, Paul Buchheit, one of the creators of Gmail, forewarned that Google “may be only a year or two away from total disruption” due to the rise of AI.
With the release of ChatGPT and now Bard, generative AI is well and truly among us.
Researchers at Meta, once known as Facebook, are also working on various AI projects.
The emergence of powerful AI trained on massive troves of data has given rise to concerns over the ethics governing the development and use of such technology.
On top of Lemoine’s claims that LaMDA had feelings, he expressed grave concerns that these very powerful and influential AI systems were largely being developed by a select few, behind closed doors in Silicon Valley, before being released to the world.
Appearing last year on Bloomberg Technology, Lemoine called Silicon Valley’s control of AI “a new form of colonialism”.
“We are creating intelligent systems that are part of our everyday life and very few people are getting to make the decision about how they work,” he said.
“How does this omnipresent AI that is trained on a very limited data set colour how we interact with each other around the world?
“What ways is it reducing our ability to have empathy with people unlike ourselves?
“What cultures of the world are getting cut off from the internet because we don’t have the data to feed into the systems based on those cultures?”
Lemoine claimed corporate values from companies like Google are setting the parameters for how chatbots will talk about certain topics, like values, rights and religion.
This, he said, will in turn affect how people think and talk about these topics, and could fundamentally shape how they engage with these issues.
Lemoine said AI tech was based primarily on data drawn from western cultures.
“Then we are populating developing nations with these technologies where they have to adopt our culture norm.”
Last year Facebook parent Meta opened its language model to academics, civil society and government organisations.
Joelle Pineau, managing director of Meta AI, said tech companies should be transparent with how the AI technology is built.
“The future of large language model work should not solely live in the hands of larger corporations or labs,” she said.
The death of the iPod: The evolution of the device that changed music
“Robots will be able to do everything better than us,” Musk said in 2017.
“I have exposure to the most cutting-edge AI, and I think people should be really concerned by it.”
Many in the AI community accuse Musk of being an alarmist and a disruption.
Sundar Pichai, CEO of Google and parent company Alphabet, today confirmed AI-powered tools will soon begin rolling out in Google’s flagship search engine.
“It’s critical,” he said, “that we bring experiences rooted in these models to the world in a bold and responsible way.”