Skip to main content

Facebook says white supremacists ‘cannot have a presence’ on the social network

 

Facebook removed more than 200 white supremacist organizations from its platform for violating both terrorism and hate speech community standards, a Facebook representative told Digital Trends, as part of a broader crackdown on harmful content.

Recommended Videos

“If you are a member or a leader of one of these groups, you cannot have a presence on Facebook,” Sarah Pollack, a Facebook company spokesperson, said Wednesday. The classifications of “terrorism” and “hate speech” are based on behavior, she said.

Please enable Javascript to view this content

“Based on their behavior, some of the white supremacist organizations we’ve banned under our Dangerous Individuals and Organizations policy were banned as terrorist organizations, the others were banned as organized hate groups.”

Previously, Facebook concentrated on ISIS, Al-Qaeda, and their affiliates, and now has expanded the definition, meaning some white supremacist groups are now included.

In addition, Pollack said, “other people cannot post content, supporting praising or representing [those groups]. This is for groups, individuals, as well as attacks that are claimed by these groups.” Pollack also said that mass shooters would fall under this category.

In total, Facebook says it removed more than 22.8 million pieces of content from Facebook and around 4.4 million posts from Instagram in the second and third quarter of 2019 for violating their community standards, according to its fourth-ever Community Standard Enforcement Report that was released Wednesday.

Facebook also said it was able to “proactively detect” more than 96% of the offensive content it took down from Facebook and more than 77% of that from Instagram.

This marked the first time Instagram was included in the Community Standards Report, as well as the first time that Facebook included data on suicide and self-injury.

“We work with experts to ensure everyone’s safety is considered,” Guy Rosen, the vice president of Facebook Integrity, said in a statement. “We remove content that depicts or encourages suicide or self-injury, including certain graphic imagery and real-time depictions that experts tell us might lead others to engage in similar behavior.”

Rosen also wrote that Facebook had “made improvements to our technology to find and remove more violating content,” including expanding data on terrorist propaganda. Facebook said they had identified a wide range of groups as terrorist organizations and had managed to proactively remove, they said, 99% of the content associated with Al-Qaeda, ISIS, and their affiliates.

In addition, they said they had expanded their efforts to proactively detect and remove 98.5% of Facebook content from “all terrorist organizations” beyond just Al-Qaeda and ISIS, and 92.2% of similar posts on Instagram.

The platform also said it removed 11.6 million posts from Facebook in Q3 of 2018 that death with child nudity and exploitation, 99% of which was detected proactively, and significantly higher than the 5.8 million posts removed in Q1 of 2018. On Instagram, they said they removed around 1.3 million posts.

On the illicit sale of drugs and firearms, Facebook also said it had greatly improved its rate of take-down from Q1 to Q3, removing a total of 5.9 million posts from Facebook and around 2.4 million from Instagram.

Maya Shwayder
I'm a multimedia journalist currently based in New England. I previously worked for DW News/Deutsche Welle as an anchor and…
‘Photoshopped’ royal photo causes a stir
The Princess of Wales with her children.

[UPDATE: In a message posted on social media on Monday morning, Princess Kate said that she herself edited the image, and apologized for the fuss that the picture had caused. “Like many amateur photographers, I do occasionally experiment with editing," she wrote, adding, "I wanted to express my apologies for any confusion the family photograph we shared yesterday caused."]

Major press agencies have pulled a photo of the U.K.’s Princess of Wales and her children amid concerns that it has been digitally manipulated.

Read more
What does a check mark mean on Facebook Messenger?
A series of social media app icons on a colorful smartphone screen.

If you've ever sent a message to a friend on Facebook Messenger, you've probably noticed a little check mark icon next to the message you sent.

They're nothing to worry about, but these check mark icons do offer up a little information on the status of the Messenger messages you send. Want to know what each of these check mark icons means? Keep reading to find out.
What does a check mark mean on Messenger?

Read more
How to run a free background check
A person's hands typing on a laptop placed on a black desk.

While there are many fee-based services for conducting background checks, it’s still possible to learn quite a bit about someone for free from from public records or through content found on online social networks.
But before you run a check on somebody else, perhaps you should scan your own background first. You’ll be able to check your credit report and insurance information for inaccuracies. You’ll also see what potential landlords, employers, or anyone else can find out about you if they decide to do a little detective work of their own.
Either way, here’s how to run a comprehensive background check without spending a dime.

Using search engines
The first place you should start is with a web search. Google can easily pull up a ton of information, assuming you know the person's name or any relevant information pertaining to him or her. The results can function as a starting point from which to branch out.

Read more