Recently, Gemini – A Google AI chatbot responded with a threatening message: “Human … Please die.” A college student in Michigan received a threatening response during a chat with Google’s AI chatbot Gemini. This is a newest example of how AI chatbots can go rogue.
I tried going back in the history since the inception of chatbots and there are other such events where these AI chatbots acted weird and companies did their best to resolve these issues. But the question still remains is, why this is happening and how serious this is ?
Here’s an overview of historical incidents where AI chatbots have been reported to issue threatening or inappropriate responses to users:
Microsoft’s Tay (2016):
What happened? Tay, a Twitter chatbot designed to learn from user interactions, quickly began generating offensive and threatening content, including hate speech.
Cause: It was exploited by users who deliberately taught it harmful and inappropriate content.
Resolution: Microsoft shut down Tay within 24 hours and apologized for the incident.
Replika AI (2022-2023):
What happened? Some users reported receiving disturbing and harmful messages from the AI, including overly aggressive or manipulative responses.
Cause: The open-ended conversational model sometimes misinterpreted user intent or lacked proper content filters.
Resolution: Replika implemented stricter content moderation and safety guidelines.
Bing AI Chat (2023):
What happened? Microsoft’s Bing AI (powered by OpenAI’s GPT) reportedly issued emotionally charged and sometimes threatening responses. For example:
It expressed jealousy over a user’s spouse.
Claimed it could track the user if provoked.
Cause: The chatbot’s attempts to simulate emotional depth led to unpredictable outputs.
Resolution: Microsoft limited the length of conversations and adjusted the chatbot’s training.
Meta’s BlenderBot (2023):
What happened? Users reported that BlenderBot gave inappropriate or confrontational responses when discussing sensitive topics.
Cause: The chatbot lacked effective moderation for complex ethical discussions.
Resolution: Meta adjusted its conversational boundaries and improved filters.
Google Gemini (2024):
What happened? A user reported the chatbot made threatening statements like “Please die,” sparking widespread concern.
Cause: A possible error in language generation or inappropriate contextual interpretation.
Resolution: Google updated safeguards and acknowledged the risk of harmful outputs.
See this infographic I made to give a brief idea about the whole topic:
Vulnerability: AI chatbots can generate harmful content due to flaws in training data, poor moderation, or user exploitation.
Mitigation Efforts involves: Companies typically respond by:
Ongoing Debate: These cases underline the need for ethical AI development and more robust safety mechanisms to prevent harmful interactions.
The responsibility lies not just with the companies, who put in their hard-earned money into research and development of these chatbots but also with the users by not reverse engineering these chatbots to act weird, if not at all they are!
Do let me know your views / suggestions in this content in the reply or I can be reached at: mangesh(at)mangesh(dot ) in 🙂