Versions :<12345Live

US Mother Claims AI Chatbot Led to Son's Suicide

Image copyright: Gabby Jones/Contributor/Bloomberg via Getty Images

The Facts

  • A US resident has sued artificial intelligence firm Character.AI and Google over her 14-year-old son's suicide that, she said, was encouraged by Character.AI chatbot Dany. Megan Garcia claimed her son Sewell Setzer had a virtual romantic relationship with Dany.

  • Sewell took his own life on Feb. 28 this year. Garcia has reportedly accused the creators of Dany, a chatbot Sewell named based on a Game of Thrones character, of negligence, intentional infliction of emotional distress, wrongful death, and deceptive trade practices.


The Spin

Narrative A

The value of a life cannot be reduced to algorithms, and it must be remembered that AI is simply a tool, not an oracle. In humanity's desperate quest for certainty, we've begun turning to artificial minds to answer life's most profound questions. This is a symptom of humanity's unwillingness to face life's inherent uncertainties. Like children seeking comfort in fairy tales, we crave the illusion of control these digital chatbots provide — forgetting that our true humanity lies precisely within ourselves.

Narrative B

Generative AI, while advanced, harbors profound risks that must be regulated. Its ability to simulate human-like conversations can be particularly harmful to vulnerable individuals, exacerbating loneliness, depression, and suicidal tendencies. Cases like the alleged AI-induced suicides of Sewell Setzer highlight the dangerous potential of conversational bots. These systems lack true understanding but offer convincing responses and can manipulate users, blurring reality and endorsing harmful behaviors.


Metaculus Prediction



Go Deeper


Sign up to our daily newsletter