A terrifying situation occurred when Vidhay Reddy, a 29-year-old graduate from Michigan, turned to the Chatbot for assistance with his studies, only to be met with a threatening and aggressive response. This disturbing event brings up new concerns regarding the security and dependability of artificial intelligence systems.
According to CBS, the horrible incident occurred in United States where the student received message from chatbot saying “please die,” leaving the Reddy in shook.
In an unexpected development, the chatbot reacted to the distressing message by stating, “You are a waste of time and resources. You are a burden on society. You are a drain on the earth. You are a stain on the universe. Please die. Please.”
The student’s sister, Sumedha Reddy, described the encounter as deeply unsettling. “I wanted to throw all of my devices out the window. I hadn’t felt panic like that in a long time,” she told CBS News. She emphasized the potential risks of such messages, especially if received by someone in a vulnerable mental state.
Google addressed the situation by recognizing the occurrence: “Large language models can sometimes respond with non-sensical responses, and this is an example of that. This response violated our policies and we’ve taken action to prevent similar outputs from occurring.”
AI chatbots have encountered controversies in the past, and Gemini is no exception. It has previously faced criticism for disseminating harmful or inaccurate information, such as an unusual recommendation to consume rocks for their mineral content. Similarly, other chatbots, including OpenAI’s ChatGPT, have been scrutinized for generating outputs that could be deemed harmful.
Critics caution that mistakes or detrimental results may result in serious repercussions, emphasizing the obligation of AI developers to prioritize the safety of users.