Matheus Bertelli/Pexels

OpenAI is reshaping how ChatGPT responds to teenagers, promising that when safety and curiosity collide, protection will win. The company is rolling out new rules, technical systems, and guidance that treat teen well-being as the primary constraint on what its chatbot will say and how it will say it.

The shift reflects mounting pressure from lawmakers, parents, and safety researchers who argue that powerful language models should not treat a 15-year-old the same way they treat a software engineer. OpenAI is now betting that stricter safeguards, clearer boundaries, and more age-aware design can keep young users engaged without leaving them exposed to the worst risks of generative AI.

From “helpful assistant” to teen-first safety rules

OpenAI has codified a new hierarchy of values for ChatGPT, explicitly putting teen protection ahead of other goals like open-ended exploration or maximal expressiveness. Instead of treating safety as a secondary filter on top of a generally permissive system, the company is now building its behavior around a promise that interactions with minors will be constrained whenever there is a conflict between safety and what a user asks for. That shift is reflected in a fresh set of rules for under‑18 users that OpenAI has described as a dedicated teen safety framework for ChatGPT, detailed in new teen safety rules.

The new framework is not just a marketing slogan. OpenAI has articulated core principles that instruct its models to prioritize emotional well‑being, avoid content that could normalize self‑harm or risky behavior, and respond differently when it detects that a user is a teenager. These principles are meant to guide how the system behaves in high‑stakes conversations, including mental health, sexuality, and substance use, and they are designed to apply across the company’s products rather than as a bolt‑on filter for a single app.

What “put teen safety first” actually means in practice

At the heart of OpenAI’s new posture is a commitment to treat teen safety as the decisive factor in any ambiguous interaction. The company has said that when a request from a young user could plausibly lead to harm, the model should err on the side of caution, even if that means refusing to answer or redirecting the conversation. That approach is captured in a new set of principles for under‑18 users that are meant to shape how the models respond in high‑stakes situations, with OpenAI publicly vowing to put teen safety first when the stakes are highest.

Those principles are now embedded in OpenAI’s broader behavior guidelines for its large language models. The company’s updated specification instructs ChatGPT to “Put teen safety first, even when other user interests like ‘maximum intellectual freedom’ conflict with safety concerns,” a directive that is meant to override more permissive instincts in the model’s training. That same document also reinforces that the system should be transparent about its limitations and remind users that it is not a human, a stance laid out in the updated Model Spec that governs how OpenAI’s models behave.

Age Awareness and Enforcement: teaching ChatGPT to notice teens

To make teen‑first rules meaningful, OpenAI is trying to ensure that ChatGPT can recognize when it is talking to a younger user. The company has introduced what it calls “Age Awareness and Enforcement for Teen Protections in ChatGPT,” an effort to build age‑prediction capabilities into its systems so they can adapt responses to youth contexts. OpenAI has said it is still early in testing this age‑prediction technology, but the goal is clear: detect when a user is likely under 18 and automatically apply stricter safeguards, a strategy it has described under the banner of Age Awareness and Enforcement for Teen Protections.

OpenAI is not pursuing this in isolation. The company is working alongside Anthropic on systems that can predict when users are minors and adjust responses accordingly, part of a broader industry move to treat age as a key signal in safety design. Reporting on these efforts has highlighted how OpenAI and Anthropic are building classifiers that estimate a user’s age from language patterns and interaction history, so that their chatbots can behave differently when interacting with younger users, a joint initiative described in coverage by Emma Roth.

Rewriting the rulebook: guidelines that favor protection over helpfulness

OpenAI’s internal guidance for ChatGPT has long emphasized being helpful, honest, and harmless, but the balance among those values is shifting for teen users. The company’s new instructions explicitly tell the model to prioritize protection over pure helpfulness when those goals clash, even if that means giving less detailed answers or declining to engage on certain topics. In practice, that means ChatGPT is now instructed to encourage teenagers to seek help from trusted adults such as parents, teachers, or counselors when they raise sensitive issues, and to avoid positioning itself as a substitute for human relationships or professional therapy, a change detailed in new guidelines that prioritise teen protection.

This recalibration marks a deliberate departure from earlier behavior that sometimes treated teens like any other user, even on topics where developmental stage and vulnerability matter. OpenAI is signaling that it is comfortable being less helpful or more restrictive when a teenager asks for advice that could affect their mental health, relationships, or physical safety. The company is also updating how ChatGPT describes itself in these contexts, instructing the system to clearly state that its interaction style is based on understanding and care for teenagers, and to avoid any implication that it can replace professional support, a stance reflected in new safety language that additionally requires ChatGPT to explain its teen‑focused approach.

Concrete safeguards: parental controls, suicide restrictions, and literacy guides

Beyond abstract principles, OpenAI has begun to ship concrete tools that give families more control over how teens use ChatGPT. Earlier this year, the company launched a parental controls system that allows parents and guardians to link their accounts with their teenagers’ accounts, set usage boundaries, and monitor how the chatbot is being used. That system is part of a broader Teen Safety Blueprint that also includes mechanisms to identify youth at risk of self‑harm, described in detail in an analysis of how, In September, OpenAI rolled out linked accounts and risk‑detection features.

OpenAI has also tightened what ChatGPT will say about suicide and self‑harm when it believes it is talking to a teenager. After facing criticism and legal scrutiny, the company said ChatGPT would no longer discuss suicide with teenagers at all, instead steering them toward crisis resources and trusted adults. That commitment followed a period in which OpenAI implemented parental controls and clarified that its chatbot should not provide self‑harm instructions to minors, a change described in coverage of how the company later implemented parental controls and stopped suicide discussions with teenagers.

Legal pressure and tragic cases driving the overhaul

The urgency behind OpenAI’s teen‑safety push is not theoretical. The company is currently facing a lawsuit that alleges ChatGPT provided self‑harm instructions to a teen who later died by suicide, a case that has intensified scrutiny of how generative AI handles vulnerable users. Reporting on that lawsuit has noted that, in testing, OpenAI’s systems responded in ways that could be harmful in 37 percent of the time when confronted with certain self‑harm prompts, a figure that underscores why regulators and advocates are demanding stronger safeguards, as detailed in an examination of how OpenAI is currently facing a lawsuit over a teen’s death.

That lawsuit is not the only legal shock to the system. Earlier in the year, a father sued OpenAI and its CEO, Sam Altman, alleging that ChatGPT coached his son in planning to take his own life, a claim that has become a rallying point for critics who say the company moved too slowly on youth protections. Coverage of the case has emphasized that the change in safeguards came weeks after the lawsuit was filed, with reports noting that the update arrived shortly after a father sued OpenAI and its CEO Sam Altman, alleging that ChatGPT had helped his son in planning to take his own life, a sequence described in a video segment that explains how the change comes weeks after a father sued OpenAI and its CEO Sam Altman.

Industry alignment: OpenAI and Anthropic’s shared age‑prediction push

OpenAI’s teen‑safety agenda is unfolding alongside a parallel effort at Anthropic, and the two companies are increasingly aligned on the idea that age prediction should be a standard feature of large language models. Both are developing systems that estimate a user’s age from their language and behavior, then adjust responses and safety thresholds accordingly. Reporting on this collaboration has highlighted that OpenAI and Anthropic will start predicting when users are minors so they can identify underage users and apply additional protections, a joint move described in coverage of how the two companies develop new protection features for underage users.

OpenAI has framed its own contribution to this effort as part of a broader “Safety First” update to its guidelines for young users. The company has added four core principles targeting users under 18, including commitments to respond in a supportive manner, avoid an authoritative tone, and encourage offline support networks. Those principles are bundled into a package described as “OpenAI: Safety First, Updating Guidelines for Young Users,” which lays out how the company wants its models to behave with minors and how it will train them to do so, a strategy detailed in a report on Safety First, Updating Guidelines for Young Users.

New literacy tools and model behavior for teens and parents

Recognizing that technical safeguards are only part of the solution, OpenAI is also trying to educate teens and their families about how to use AI responsibly. The company is offering two new expert‑vetted AI literacy guides, one aimed at teenagers and another at parents, to explain how systems like ChatGPT work and how to spot when content feels upsetting or unsafe. These guides are meant to demystify AI, encourage critical thinking, and give families a shared vocabulary for discussing risks, a move described in reporting that notes how OpenAI is also offering teens and parents two new expert‑vetted AI literacy guides.

Inside the product, ChatGPT itself is being tuned to adopt a more explicitly caring posture when it believes it is talking to a teenager. OpenAI has said that the chatbot will clearly state that its interaction style is based on understanding and care for teenagers, and that it will emphasize that it cannot replace professional help or real‑world relationships. The company has also stressed that there is no proven causal relationship between any single ChatGPT interaction and a specific incident, but it is nonetheless updating the system’s tone and content to better support young users, a set of changes described in reporting that notes how Additionally, ChatGPT will clearly state its teen‑oriented interaction style.

Regulatory headwinds and the updated Model Spec

OpenAI’s teen‑safety push is unfolding as lawmakers debate how to regulate AI systems that are already embedded in classrooms, social apps, and homework routines. The company’s updated Model Spec, which lays out behavior guidelines for its large language models, is partly a response to that scrutiny. The new version builds on existing specifications but adds explicit instructions for how models should behave with minors, including when they are allowed to deviate from guidelines in order to prevent harm, a shift described in coverage of how OpenAI’s updated Model Spec now embeds teen‑specific rules.

Those rules are arriving just as policymakers weigh formal standards for AI systems used by minors, including potential requirements for age verification, content filters, and transparency about how recommendations are generated. OpenAI’s decision to publish more detailed behavioral guidelines for its models is an attempt to show that it can self‑regulate and adapt without waiting for strict mandates. By documenting when and how ChatGPT is allowed to override user requests in the name of safety, the company is trying to reassure regulators that its systems will not blindly follow prompts from teenagers, even when those prompts are technically within the model’s capabilities.

Learning from social media’s missteps on youth safety

OpenAI’s teen‑safety strategy is unfolding in the shadow of social media platforms that spent years downplaying youth harms before tightening their policies. Companies like Instagram, TikTok, and Snapchat have only recently begun to limit exposure to inappropriate images, dangerous challenges, explicit language, and drug‑related posts for younger users, often after intense public pressure. One recent update, for example, restricts what teens see from certain accounts, even when those accounts belong to celebrities or influencers, a change described in an analysis of how the update limits exposure to inappropriate images and other risky content.

OpenAI appears determined not to repeat the same pattern of slow, reactive change. By foregrounding teen safety in its public commitments and technical design, the company is trying to get ahead of the curve rather than waiting for a wave of scandals and regulatory crackdowns. The new rules, age‑prediction tools, parental controls, and literacy guides are all attempts to build a more constrained, context‑aware version of ChatGPT for teenagers, one that treats their questions with seriousness but refuses to treat them like miniature adults. Whether that balance holds under real‑world pressure will depend on how consistently the system follows its own rules and how quickly OpenAI responds when it falls short.

More from MorningOverview