September 22, 2022

The social networks are failing to detect hate speech

By Charles Cohen
View all posts


Everyone is entitled to their opinion. But when free speech goes to extremes and uses damaging vocabulary, it can become extremely dangerous. As social media use grows, so does online hate. Groups and individuals have become targets because of their ethnicity, religion, gender, and sexual orientation.

And sadly, many social media platforms don't have the tools in place to moderate hate speech before it causes irreparable harm.

The link between social networking sites and hate

Social media is a convenient way to reach millions of users across the world. However, despite the growing use of toxicity to target minority groups, moderation guidelines are still limited in their effectiveness. For example, when there is no moderation, violent marketing campaigns can go viral in moments and affect many people.  

Plus, anyone can post on a social platform anonymously and within seconds the post can go viral. Manually monitoring bullying and aggression on social networks is almost impossible, both in terms of the time it takes and the possibility of human error. Not to mention, human moderators can be severely affected by reading hundreds of racist or homophobic messages every day.

The big platforms are making efforts to moderate content and remove toxicity using machine learning algorithms to detect hate speech.

But currently, social platforms like Facebook detect only 20-40% of hateful content (in English only): which means up to 60% is still there targeting vulnerable people. 

The impact of hate being posted on social media

There are many examples of the destructive impact hateful content can have on communities. This can range from inciting others to commit violence and promoting propaganda to emotional abuse, and even encouraging murder and suicide.

One example is the ‘Fetrah’ campaign that took place on Twitter. This anti-LGBT (lesbian, gay, bisexual, and transgender) movement spread rapidly, despite Facebook and Instagram banning the account. The logo is a blue and pink flag that represents the male and female genders and rejects any other communities. The campaign was created by three Egyptian marketers and seeks to harm gay and transgender people and groups, particularly in the Middle East.

Another example was Facebook’s inability to detect hate in the East African language Kiswahili (Swahili). A series of ads were submitted to Facebook that referred to rape, beheadings, and extreme physical violence by the non-profit group, Global Witness. The ads were approved by Facebook (some were even in English). While the ads were never published, as this was a test, this shows the gap that exists in content moderation on social media. 

These examples demonstrate that giant social platforms are still not monitoring and filtering out hateful content. Every user deserves maximum protection through the use of reliable content moderation systems.

While 40% of users leave a platform after their first exposure to toxic language, many stay and participate, resulting in collective abuse. Research shows that even if a user reports hateful comments, one-third are not deleted. The result is that many marginalized groups and people are left vulnerable to online hate. There's little doubt that improved moderation standards are needed on social media, particularly those that have a major influence on users’ psyches.

How can Bodyguard help you and your communities

Bodyguard believes in freedom of speech and the right for people to use the internet freely, without receiving harmful and hateful comments. 

Bodyguard is a unique moderation solution that is both instantaneous and intelligent, combining the speed of a machine with the subtle ability and nuance of a human to assess and evaluate toxic content.

Bodyguard moderation:

  • Detects and moderates toxic content in real-time
  • Protects brands, businesses, communities, and staff members
  • Prevents negative exposure about your brand (bad buzz…)
  • Is tailored to suit your exact needs

As a more effective method than employing a human moderator, Bodyguard’s objective is to have a positive social impact on our society. Toxic content is detected in real-time and is moderated immediately, eliminating the possibility of human error.

For more information on how Bodyguard can help your business, contact us.

Popular Insights

What is user-generated content and why does it matter?

By Gracia Kahungu|February 26, 2024
Read more

Bodyguard elevates content moderation with new language features

By Gracia Kahungu|December 19, 2023
Read more

TikTok moderation: why it’s important and how to do it

By Lucy Wright|September 15, 2023
Read more

New season, new rules: Why football clubs can’t tolerate online toxicity

By The Bodyguard Team |August 3, 2023
Read more
Threat MonitoringCommunity ProtectionAudience UnderstandingIntegrations
Helpful Links
SupportTechnical Documentation
CompanyCareersMedia KitContact us

© 2024 Bodyguard ai. All rights reserved worldwide.

Terms & Conditions|Privacy Policy