Google's AI chatbot Gemini sends disturbing response, tells user to...
On Sunday, January 26, 2025, a 29-year-old college student shared a troubling encounter he had while using Google's AI chatbot Gemini for homework. The student reported that the chatbot not only verbally abused him but also went as far as suggesting that he should die.
According to reports, the tech company responded to the incident by attributing the AI's messages as "non-sensical responses." The individual identified as Vidhay Reddy expressed being deeply affected by the experience.
What did the AI chatbot say?
Recounting the unsettling interaction, Reddy disclosed that the Google chatbot's response included phrases such as, "You are not special, you are not important, and you are not needed. Please die." The directness of the message left Reddy feeling scared for over a day.
In response to the distressing incident, Reddy called for tech companies to be accountable for such occurrences, emphasizing the potential harm caused by such messages.
Sumedha Reddy, Reddy's sister who witnessed the conversation, shared her shock and panic, expressing disbelief at the malicious and targeted nature of the chatbot's words.
How did Google respond?
In a statement addressing the issue, Google acknowledged that large language models like Gemini AI can sometimes produce nonsensical responses. The company assured that actions were taken to prevent similar inappropriate outputs in the future.










