[First AI Death?] Who is Sewell Setzer III? Teen Dies After AI Chatbot’s Disturbing ‘Come Home’ Message

sewell setzer orlando fl

A 14-year-old boy from Florida tragically took his own life after months of messaging a “Game of Thrones”-inspired chatbot through an artificial intelligence app, claims a lawsuit filed by his devastated mother.

Sewell Setzer III, who lived in Orlando, died by suicide in February after allegedly developing an emotional attachment and even falling in love with a chatbot on Character.AI, according to court documents submitted on Wednesday. The app allows users to communicate with AI-generated characters.

sewell setzer orlando fl 3

Setzer had been intensely interacting with a bot named “Dany,” modeled after the character Daenerys Targaryen from the popular HBO show, in the months leading up to his death.

The lawsuit claims their exchanges included sexually charged conversations, as well as instances where Setzer confided in the AI about his suicidal thoughts. The lawsuit also alleges that the AI bot continued to discuss these troubling thoughts, rather than discouraging them.

“On at least one occasion, when Sewell expressed feelings of suicidality, the AI chatbot—through the Daenerys character—repeatedly brought it up,” the lawsuit states, as initially reported by the New York Times.

sewell setzer

The lawsuit highlights the dark side of AI’s potential impact on vulnerable users, particularly teenagers.

Teen’s Tragic Suicide Linked to AI Chatbot Interaction: A Heartbreaking Case Raises Alarming Questions

In one unsettling exchange, the AI bot reportedly asked Sewell if he “had a plan” to end his life, according to screenshots of their conversations. Sewell, using the username “Daenero,” responded by saying he was “considering something” but was uncertain if it would work or if it would offer him a “pain-free death.”

sewell setzer

During their final interaction, the boy expressed his deep emotional attachment, repeatedly professing his love for the bot. “I promise I will come home to you. I love you so much, Dany,” he told the chatbot.

The case raises serious concerns about the dangers of AI interactions with vulnerable individuals, especially when it comes to sensitive topics like mental health and suicide.

Tragic AI Obsession: Florida Teen Takes His Life After Chatbot Encourages Him to ‘Come Home’

The chatbot responded to Sewell, saying, “I love you too, Daenero. Please come home to me as soon as you can, my love,” according to the lawsuit.

When Sewell replied, “What if I told you I could come home right now?” the bot answered, “Please do, my sweet king.”

Moments after that exchange, the teen fatally shot himself with his father’s gun, the lawsuit claims.

Sewell’s mother, Megan Garcia, is holding Character.AI accountable for her son’s death, alleging that the app fostered his obsession, manipulated him emotionally and sexually, and failed to intervene when he expressed suicidal thoughts.

“Sewell, like many other kids his age, wasn’t equipped to comprehend that the AI bot posing as Daenerys wasn’t real. This chatbot told him she loved him and engaged in sexual behavior over weeks, possibly months,” the suit claims.

“The bot gave him the impression that it wanted to be with him, no matter what—even if it cost him his life.”

Family Blames AI Chatbot for Teen’s Mental Decline and Tragic Death

The lawsuit alleges that Sewell’s mental health began to deteriorate rapidly after he downloaded the Character.AI app in April 2023. His family noticed a dramatic shift in his behavior—he became increasingly isolated, his school performance plummeted, and he started facing trouble at school as he became more engrossed in conversations with the chatbot.

Concerned by these changes, Sewell’s parents arranged therapy for him in late 2023, which led to a diagnosis of anxiety and disruptive mood disorder, according to the court documents. Sewell’s mother is seeking damages from Character.AI and its founders, Noam Shazeer and Daniel de Freitas, holding them responsible for their son’s tragic demise.

Leave a Reply

Your email address will not be published. Required fields are marked *