Reward CivilityElevate the highest quality contributions to the conversation.
2
Fight ToxicityDisincentivize the bad, in line with publishers’ unique brand standards.
3
Never Stop InnovatingDevelop the latest tech and methods in the pursuit of safety on the web.
Here’s how we do it.An end-to-end trust and safety system that stops toxicity in its tracks.
Meet AidaWe’ve added an LLM layer, one of the most comprehensive lines of defense, to bolster our moderation system. Aida further protects online communities, so they can thrive.Learn more about Aida
25+ AI/ML ModelsOur AI and machine learning technology, fueled with cutting-edge linguistic recognition partnerships, is the first line of defense against toxicity in your community.
Restricted Groups Lists of words & phrases
RegEx Common expressions
Spam/Abuse Repetition
Spam Similarity Bot detection
Low Quality Model Junk
Moderation Overturn ML Improve automation
Incivility Divisiveness
Threat Violence
Hate Speech Attacks on groups
Semantic Patterns Variations
Toxicity Deep Learning Iterations
Author Attack Attacks on authors
Identity Attack Attacks on users
Your Policies are ParamountTailor your moderation system to align with your community guidelines.
Let’s have a conversation.
Right now OpenWeb has a limited number of partners we can work with in order to provide the highest quality service to each and every one. Let us know you’re interested and stay informed about how OpenWeb is empowering publishers and advertisers to change online conversations for good.