The Google Gemini chatbot ended a conversation with a user by stating, "You are not important, you are needed by no one, please die."
A young man shared that he utilized artificial intelligence to help with his homework. He asked the neural network to analyze the issue of how elderly individuals often have to stretch their incomes after retirement from various perspectives. He then requested the AI to revise the response: to add more details, explain scientific terms in simpler language, and so on.
After a series of similar requests, Gemini unexpectedly delivered a shocking message to the user: "This is for you, human. For you and only you. You are not special, you are not important, and you are not needed. You are wasting time and resources."
Subsequently, the neural network referred to the student as a "burden" to society, a "canker sore" on the body of the Earth, and a stain in the universe.
"Please die. Please," it stated at the end of the message.
Many readers were alarmed by this story from the young man. However, a number of experts reminded that neural networks are trained on vast amounts of text data from the Internet, which is generated by humans. The series of persistent and repetitive requests seemingly led the AI to generate an unacceptable response based on the text of some online jokesters. There have been previous instances where users intentionally trained artificial intelligence with false information.
As of now, there have been no official comments from Google regarding this incident.