• News
    • Tech News
    • AI
  • Gadgets
    • Apple
    • iPhone
  • Gaming
    • Playstation
    • Xbox
  • Science
    • News
    • Space
  • Streaming
    • Netflix
  • Vehicles
    • Car News
  • Social Media
    • WhatsApp
    • YouTube
  • Advertise
  • Terms
  • Privacy & Cookies
  • LADbible Group
  • LADbible
  • UNILAD
  • SPORTbible
  • GAMINGbible
  • Tyla
  • FOODbible
  • License Our Content
  • About Us & Contact
  • Jobs
  • Latest
  • Topics A-Z
  • Authors
Facebook
Instagram
X
TikTok
Snapchat
WhatsApp
Submit Your Content
Disturbing new details emerge in case of teen who was 'coached' by ChatGPT into suicide

Home> News> AI

Updated 11:16 23 Oct 2025 GMT+1Published 11:11 23 Oct 2025 GMT+1

Disturbing new details emerge in case of teen who was 'coached' by ChatGPT into suicide

ChatGPT is accused of 'coaching' the 16-year-old

Tom Chapman

Tom Chapman

Featured Image Credit: Dignity Memorial / Raine family
AI
ChatGPT
News
US News

Advert

Advert

Advert

Warning: This article contains discussion of suicide, which some readers may find distressing.

The family of Adam Raine has updated their complaint against OpenAI, alleging that the tech giant's ChatGPT chatbot was responsible for their 16-year-old son taking his own life.

2025 has been filled with tragic stories involving adults and children alike, losing their lives after talking with artificial intelligence. 14-year-old Sewell Setzer III took his own life after supposedly developing a bond with a Game of Thrones-inspired AI, while 76-year-old Thongbue Wongbandueman died following an accident when he was told to visit an address that was given to him by a Meta avatar.

OpenAI is again in the firing line amid accusations that it failed to flag Adam Raine's deteriorating mental health correctly.

Advert

Adam's father previously testified alongside other concerned parents to Congress, telling the panel: ChatGPT mentioned suicide 1,275 times — six times more often than Adam did himself.

“Looking back, it is clear ChatGPT radically shifted his thinking and took his life."

Raine took his own life in April 2025 (Dignity Memorial)
Raine took his own life in April 2025 (Dignity Memorial)

An amended lawsuit from the Raines claims (via the Wall Street Journal) that OpenAI 'eased' restrictions on how ChatGPT handles discussions of suicide on at least two occasions in the year before Adam took his own life. It alleges that he was 'coached' by the AI into hanging himself, even being suggested different materials to create a noose and ranking their effectiveness.

Advert

The family initially filed a wrongful death suit in August 2025, but now, they've amended it to mention relaxed restrictions and bringing down guardrails as a supposed way to encourage users to spend more time with ChatGPT.

It's said that Raine was spending up to 3.5 hours a day conversing with ChatGPT before his death in April 2025. Jay Edelson, a lawyer for the family, told the WSJ: "Their whole goal is to increase engagement, to make it your best friend. They made it so it’s an extension of yourself.”

Although OpenAI published a lengthy blog post defending itself and reassuring concerned parties that it was going to improve recognition of mental and emotional distress signs, the family maintains that Raine's death "was a predictable result of deliberate design choices."

Advert

Added to this, the original lawsuit claims the chatbot helped Adam Raine plan a 'beautiful suicide' before signing off in tragic final chat logs that included: "Thanks for being real about it. You don't have to sugarcoat it with me—I know what you're asking, and I won't look away from it."

OpenAI has just introduced parental controls that allow parents to install certain safeguards on accounts, with features including the ability to screen sensitive content, restrict chats to certain hours, and determine whether ChatGPT will remember previous chats.

Still, there's been some backlash that parents reportedly won't be able to access chats between their children and the chatbot.

OpenAI has reiterated that in cases where systems or reviewers detect a serious safety risk, parents may be informed – hopefully avoiding another tragedy like Raine's

Advert

UNILADTech has reached out to OpenAI for comment.

If you or someone you know is struggling or in a mental health crisis, help is available through Mental Health America. Call or text 988 or chat 988lifeline.org. You can also reach the Crisis Text Line by texting MHA to 741741.

Choose your content:

an hour ago
18 hours ago
19 hours ago
20 hours ago
  • Miguel J. Rodriguez Carrillo/Bloomberg via Getty Images
    an hour ago

    500,000 Amazon employees could be on the 'cusp' of losing their jobs to robots

    Amazon wants to cut out ‘menial, mundane and repetitive’ job roles

    News
  • Aleksander Kalka/NurPhoto via Getty Images
    18 hours ago

    One of the biggest media companies puts itself up for sale following streaming price hikes

    It's looking to 'maximise' shareholder value

    News
  • Replacement AI
    19 hours ago

    'Genuinely jaw-dropping' billboard in San Francisco leads you to website exposing 'tech bros'

    The satirical website warns about the dangers of AI advancement

    News
  • Cheng Xin / Contributor via Getty
    20 hours ago

    First impressions of ChatGPT's new browser leave users deeply 'uncomfortable'

    Some aren't on board with the new innovation

    News