OpenAI gpt-oss-safeguard
Blog post from Ollama
Ollama, in collaboration with OpenAI and ROOST, is introducing gpt-oss-safeguard reasoning models to enhance safety classification tasks. These models, available in two sizes (20B and 120B) and licensed under Apache 2.0, are designed to reason about safety, enabling use cases like filtering and content labeling. The models utilize a "bring your own policy" approach, allowing customization to various products and use cases with minimal engineering. They emphasize reasoned decisions for greater transparency and trust in policy application, and offer configurable reasoning effort to match specific needs. Evaluated against internal and external datasets, including a moderation dataset and the ToxicChat benchmark, these models offer organizations the flexibility to study, modify, and deploy critical safety technologies. ROOST, established in 2025, is a non-profit dedicated to providing open-source safety tools, underscoring the collaborative effort to advance online safety solutions.