The boy, namedSewell Setzer III,had named the chatbot onCharacter.AIafter theGame of ThronescharacterDaenerys Targaryen.
His mother has since filed a lawsuit against Character.AI.
Sewell, who wasdiagnosedwith mild Aspergers syndrome as a child, reportedly preferred talking with Dany about his problems.
The chatbot reportedly replied at the time: My eyes narrow.
My voice is a dangerous whisper.
And why the hell would you do something like that?
At first, Dany seemed to have a go at persuade the disturbed teen from hurting himself.
… like do, my sweet king, Dany replied.
The last text exchange occurred on the night of February 28, in the bathroom of Sewells mothers house.
Sewells parents and friends had no idea hed fallen for a chatbot, theTimesreported.
They just noticed him getting sucked deeper into his phone.
Despite his autism diagnosis, Sewell never had serious behavioral ormental healthissues before.
Marketed as solutions to loneliness, certain apps reliant on AI can simulate intimate relationships.
However, they could also pose certain risks to teens already struggling with mental health issues.
researchers, is the market leader in A.I.
I feel like its a big experiment, and my kid was just collateral damage, Garcia told theTimes.
The company went on to share a link to its official website which outlines its new protective enhancements.
This is absolutely devastating, a reader commented
Thanks!
Check out the results: