Hate speech and extremism should be content moderators top priority for detecting and removing.
Hate speech is defined as an explicit or suggestive attack against a person or group on the basis of race, religion, ethic origin, sex, disability, sexual orientation or gender identity. If hate speech is not detected online, it can escalate to extremism.
Extremism for Trust & Safety and content moderation professionals is a behavior that further moves people into "out groups" on the basis of race, religion, ethic origin, sex, disability, sexual orientation or gender identity.
Extremism does not often happen on a platform, but the severity of the actions can cause serious damage to community health online and offline. Extremist groups use online platforms to spread radical beliefs and recruit people. Even if they only use online platforms, extremist groups can inspire offline actions that have catastrophic consequences.
The problems Trust & Safety professionals can face with detecting hate speech are understanding cultural significance, translation of languages, and being kept up to date with current events.
Platforms need a combination of people and technology to accurately detect hate speech and extremism.
Read this whitepaper to learn: