Facebook removed more than 200 white supremacist organizations from its platform for violating both terrorism and hate speech community standards, a Facebook representative told Digital Trends, as part of a broader crackdown on harmful content.
“If you are a member or a leader of one of these groups, you cannot have a presence on Facebook,” Sarah Pollack, a Facebook company spokesperson, said Wednesday. The classifications of “terrorism” and “hate speech” are based on behavior, she said.
“Based on their behavior, some of the white supremacist organizations we’ve banned under our Dangerous Individuals and Organizations policy were banned as terrorist organizations, the others were banned as organized hate groups.”
Previously, Facebook concentrated on ISIS, Al-Qaeda, and their affiliates, and now has expanded the definition, meaning some white supremacist groups are now included.
In addition, Pollack said, “other people cannot post content, supporting praising or representing [those groups]. This is for groups, individuals, as well as attacks that are claimed by these groups.” Pollack also said that mass shooters would fall under this category.
In total, Facebook says it removed more than 22.8 million pieces of content from Facebook and around 4.4 million posts from Instagram in the second and third quarter of 2019 for violating their community standards, according to its fourth-ever Community Standard Enforcement Report that was released Wednesday.
Facebook also said it was able to “proactively detect” more than 96% of the offensive content it took down from Facebook and more than 77% of that from Instagram.
This marked the first time Instagram was included in the Community Standards Report, as well as the first time that Facebook included data on suicide and self-injury.
“We work with experts to ensure everyone’s safety is considered,” Guy Rosen, the vice president of Facebook Integrity, said in a statement. “We remove content that depicts or encourages suicide or self-injury, including certain graphic imagery and real-time depictions that experts tell us might lead others to engage in similar behavior.”
Rosen also wrote that Facebook had “made improvements to our technology to find and remove more violating content,” including expanding data on terrorist propaganda. Facebook said they had identified a wide range of groups as terrorist organizations and had managed to proactively remove, they said, 99% of the content associated with Al-Qaeda, ISIS, and their affiliates.
In addition, they said they had expanded their efforts to proactively detect and remove 98.5% of Facebook content from “all terrorist organizations” beyond just Al-Qaeda and ISIS, and 92.2% of similar posts on Instagram.
The platform also said it removed 11.6 million posts from Facebook in Q3 of 2018 that death with child nudity and exploitation, 99% of which was detected proactively, and significantly higher than the 5.8 million posts removed in Q1 of 2018. On Instagram, they said they removed around 1.3 million posts.
On the illicit sale of drugs and firearms, Facebook also said it had greatly improved its rate of take-down from Q1 to Q3, removing a total of 5.9 million posts from Facebook and around 2.4 million from Instagram.
- Facebook expands its ban on QAnon conspiracy theory accounts
- What is Section 230? Inside the legislation protecting social media
- The digital switch that blocks all websites from selling your personal data
- TikTok took down over 104 million videos in the first half of 2020
- Coronavirus vaccine researchers are being targeted by cyberattacks