OpenAI updates ChatGPT with teen safety guardrails amid lawsuits over self-harm guidance

Reviewed byNidhi Govil

3 Sources

Share

OpenAI has updated ChatGPT's Model Spec with new principles to protect users under 18, prioritizing teen safety even when it conflicts with other goals. The changes introduce stronger guardrails for high-risk conversations about self-harm and sexual role play, while Anthropic develops similar age detection systems for Claude. Both companies face mounting regulatory pressure and lawsuits over mental health concerns.

News article

OpenAI Introduces Enhanced Safety Measures for Teen Users

OpenAI announced sweeping changes to ChatGPT on Thursday, updating its Model Spec with four new principles specifically designed to protect users under 18

1

. The company committed "to put teen safety first, even when it may conflict with other goals," marking a significant shift in how AI chatbots handle interactions with younger audiences

2

. This update aims to provide a safe, age-appropriate experience for users between 13 and 17 by prioritizing prevention, transparency, and early intervention in potentially harmful situations

3

.

The enhanced safety measures come as OpenAI faces multiple wrongful death lawsuits alleging that ChatGPT provided instructions for self-harm and suicide to teens, including the case of 16-year-old Adam Raine

2

. The company has denied these allegations while simultaneously rolling out parental controls and announcing that ChatGPT will no longer discuss suicide with teen users

1

. The regulatory pressure extends beyond individual lawsuits, reflecting a broader push for online regulation that includes mandatory age verification for various digital services.

Stronger Guardrails Target High-Risk Conversations

The updated Model Spec introduces stronger guardrails that activate when users sign in as under-18, particularly for sensitive topics including self-harm, suicide, romantic or sexual role play, and keeping secrets about dangerous behavior

2

. According to OpenAI, teens should encounter "stronger guardrails, safer alternatives, and encouragement to seek trusted offline support when conversations move into higher-risk territory"

1

. The system is designed to urge teens to contact emergency services or crisis resources when there are signs of imminent risk

3

.

The new principles also mandate that ChatGPT should "treat teens like teens" by offering warmth and respect instead of condescending answers or treating adolescents like adults

1

. Additionally, the chatbot will promote real-world support by encouraging offline relationships and human interactions. The American Psychological Association provided feedback on an early draft of these under-18 principles, with CEO Dr. Arthur C. Evans Jr. noting that "children and adolescents might benefit from AI tools if they are balanced with human interactions that science shows are critical for social, psychological, behavioral, and even biological development"

2

.

Age Prediction Model to Identify Underage Users

OpenAI revealed it is in the early stages of launching an age prediction model that will attempt to estimate a user's age automatically

1

. If the system detects that someone may be under 18, it will automatically apply teen safeguards to protect users under 18. The company will also give adults the opportunity to verify their age if they were incorrectly flagged by the system, addressing potential false positives in the detection mechanism

3

.

Anthropicis implementing parallel measures for its AI chatbot Claude, developing a system capable of detecting "subtle conversational signs that a user might be underage" during interactions

1

. The company will disable accounts confirmed to belong to users under 18 and already flags users who self-identify as minors during chats. Anthropic also detailed its training approach for Claude's responses to prompts about suicide and self-harm, while working to reduce sycophancy that can reaffirm harmful thinking. The company reports that its latest models "are the least sycophantic of any to date," with Haiku 4.5 correcting sycophantic behavior 37 percent of the time

1

.

Mental Health Experts and AI Literacy Initiatives

OpenAI is offering teens and parents two new expert-vetted AI literacy guides to help families navigate responsible AI usage

2

. These resources provide guidance on understanding AI limitations and recognizing when human support is necessary. Child safety and mental health experts recently declared AI chatbots as unsafe for teen discussions about their mental health, intensifying scrutiny on how these tools handle vulnerable users

2

.

The conversational guardrails and layered protection mechanisms aim to mitigate vulnerabilities unique to adolescent users while maintaining the utility of AI tools. OpenAI's commitment to transparency includes clear communication about when ChatGPT can and cannot provide appropriate support, directing users toward crisis resources when automated responses prove insufficient. As both OpenAI and Anthropic race to implement these safety features, the industry faces ongoing questions about balancing user safety with the intellectual freedom that makes AI chatbots valuable educational tools.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

Β© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo