On December 18, 2025, OpenAI updated its ChatGPT Model Spec with new Under‑18 principles that require the assistant to prioritize teen safety over other goals. The Verge reported that OpenAI is also rolling out an age‑prediction system to automatically apply teen safeguards, while Anthropic is introducing similar measures for Claude.
This article aggregates reporting from 4 news sources. The TL;DR is AI-generated from original reporting. Race to AGI's analysis provides editorial context on implications for AGI development.
This update is less about squeezing another point of benchmark performance out of GPT‑5.2 and more about making frontier models survivable in consumer settings. By formally encoding U18 principles in its Model Spec, OpenAI is acknowledging that a system powerful enough to influence teens’ mental health can’t be treated like a generic productivity tool. The focus on routing sensitive conversations toward safer alternatives, steering teens to offline support, and defaulting to teen protections when age is uncertain is a shift in how model behavior is specified and audited.
From a race‑to‑AGI standpoint, this is a reminder that capability and safety are increasingly intertwined. Age prediction and differential policies by cohort are early examples of context‑aware control layers that will later be needed for agents operating across jurisdictions, institutions, and risk levels. If OpenAI and Anthropic can show regulators that they can meaningfully reduce self‑harm and sexual‑content risks for minors, it strengthens the case for allowing faster deployment of more capable systems. The flip side is that such controls add complexity and potential failure modes, especially if age inference or routing is wrong, which may eventually require third‑party oversight.



