OpenAI's Teen Safety Policy Pack #55
Pinned
ludovic-openai
started this conversation in
gpt-oss-safeguard Implementation
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
We’re releasing prompt-based safety policies to help developers create age-appropriate protections for teens.
Find the policies on OpenAI's GitHub: https://github.com/openai/teen-safety-policy-pack
While safety classifiers like gpt-oss-safeguard can detect harmful content, they depend on clear definitions of what that content is. In practice, one of the biggest challenges developers face is defining policies that accurately capture teen-specific risks and can be consistently applied in real systems.
To address this challenge, we are releasing a set of safety policies on OpenAI’s GitHub, tailored to common risks faced by teens and informed by careful review of existing research about teens’ unique developmental differences. These policies are structured as prompts that can be directly used with gpt-oss-safeguard and other reasoning models, enabling developers to more easily apply consistent safety standards across their systems.
The initial release includes policies covering:
These policies can be used for real-time content filtering, as well as offline analysis of user-generated content.
Beta Was this translation helpful? Give feedback.
All reactions