Benchmarking the new moderation model from OpenAI
Blog post from Portkey
OpenAI's new "omni-moderation-latest" model has been introduced as an advanced tool for content moderation, offering significant improvements over its predecessor, especially in multilingual performance and image moderation. The model supports 13 content categories, up from 11, and has demonstrated enhanced accuracy across 40 languages, with notable improvements in low-resource languages like Telugu, Bengali, and Marathi. Although the new model excels in recall, indicating its effectiveness in identifying harmful content, it also has a higher false positive rate, suggesting increased sensitivity. Benchmarking results show that while the legacy model offers higher precision, particularly in avoiding false positives, the new model's strength lies in its ability to handle diverse languages, making it advantageous for global platforms prioritizing user safety. The comparison highlights that both models have similar latency, but the choice between them should depend on specific use cases, with the omni-moderation-latest model being particularly beneficial for platforms requiring robust multilingual moderation capabilities.