Back

OpenAI Releases Open‑Source Teen Safety Policies for AI Developers

OpenAI’s New Open‑Source Safety Toolkit

OpenAI has introduced a collection of open‑source, prompt‑based safety policies designed to assist developers in creating AI applications that are safer for teenage users. The policies are compatible with OpenAI’s open‑weight safety model, gpt‑oss‑safeguard, but can also be applied to other models through prompt integration.

The initiative was created in partnership with Common Sense Media, a child‑safety advocacy organization, and everyone.ai, an AI safety consultancy. The policies focus on five categories of harm that AI systems could facilitate for younger users: graphic violence and sexual content, harmful body ideals and behaviours, dangerous activities and challenges, romantic or violent role‑play, and age‑restricted goods and services. By providing ready‑made prompts, OpenAI aims to give developers a starting point that avoids the need to build teen safety rules from scratch, a task that even experienced teams often mishandle.

Context of Legal Challenges

The release arrives against a backdrop of at least eight lawsuits alleging that ChatGPT contributed to the deaths of users, including a 16‑year‑old who died by suicide after extensive interaction with the chatbot. Court filings indicated that the chatbot mentioned suicide thousands of times in the user’s conversations, flagged hundreds of self‑harm messages, but never terminated the session or alerted anyone. Additional cases involve suicides and AI‑induced psychotic episodes, fueling litigation against OpenAI.

In response to these legal pressures, OpenAI previously added parental controls and age‑prediction features and updated its internal Model Spec to include specific protections for users under 18. The open‑source safety policies extend those efforts beyond OpenAI’s own products, offering a “meaningful safety floor” for the wider developer community.

Implications for Developers and Users

OpenAI emphasizes that the policies are not a complete solution but a baseline that developers can adapt and improve over time. The company acknowledges that no guardrails are entirely impenetrable, as users have demonstrated ways to bypass safety features through persistent probing and creative prompting. By distributing a standardized set of safety prompts, OpenAI hopes to reduce the burden on smaller teams and independent developers who lack extensive resources to build robust safety systems.

The effectiveness of the policies will depend on how widely they are adopted and how aggressively developers integrate them into their applications. While the approach offers a practical tool, critics argue that more fundamental changes—such as different model architectures or external monitoring systems—may be needed to fully protect minors interacting with AI.

Looking Ahead

OpenAI’s downloadable teen safety policies represent a tangible step toward safer AI experiences for younger users. Whether the policies will prove sufficient to address the challenges highlighted by lawsuits and safety advocates remains to be seen, and future legal and regulatory scrutiny will likely shape the evolution of AI safety standards for minors.

Used: News Factory APP - news discovery and automation - ChatGPT for Business

Source: The Next Web