Skip to main content

Facebook removes nearly 800 QAnon-related groups, pages, hashtags, and ads

Facebook took down nearly 800 groups associated with the far-right conspiracy theory group QAnon on Wednesday, as well as more than 1,500 advertisements and 100 pages tied to the group in a move to restrict “violent acts.”

In a blog post, Facebook said the action is part of a broader “Dangerous Individuals and Organizations” policy measure to remove and restrict content that has led to real-world violence. The policy will also impact militia groups and political protest organizations like Antifa.

“While we will allow people to post content that supports these movements and groups, so long as they do not otherwise violate our content policies, we will restrict their ability to organize on our platform,” the company said.

QAnon supporters believe in a widely disproven “deep state” conspiracy that President Donald Trump is working to eradicate pedophilia and Satanism throughout Washington D.C. The conspiracy theorists have recently latched onto the COVID-19 public health crisis, calling it a “bioweapon.”

QAnon theories hit the mainstream after the controversy surrounding #Pizzagate, in which a man brought a gun to a pizzeria, claiming he would find victims of child abuse. The group has also been linked to dozens of other violent incidents that stem from baseless theories shared on private Facebook groups and message boards.

Facebook took action against QAnon earlier this month, when it yanked down an influential group with more than 200,000 members, but Wednesday’s move is perhaps the social media giant’s most substantial move yet.

The company said it will limit QAnon content from appearing in its recommendations tab, reduce its content in search results, and prohibit QAnon-related accounts and groups from monetizing content, selling merchandise, fundraising, and purchasing advertising on both Facebook and Instagram. The company plans to continue to investigate just how QAnon operates on its platform, by observing “specific terminology and symbolism used by supporters to identify the language used by these groups and movements indicating violence and take action accordingly.”

In recent months, other social media sites like Twitter and TikTok have banned and disabled popular QAnon hashtags and accounts for inauthentic, coordinated behavior and for spreading disinformation.

However, do not expect QAnon to disappear quietly: Experts have called QAnon members “really good at adapting” to online ecosystems, and several QAnon supporters have won primaries for public office on platforms that represent the conspiracy theories shared within the group.

Editors' Recommendations

Meira Gebel
Meira Gebel is a freelance reporter based in Portland. She writes about tech, social media, and internet culture for Digital…
Facebook takes down viral ‘Plandemic’ coronavirus conspiracy video
fatal shooting facebook live app

Facebook removed the viral conspiracy video "Plandemic" from its platform Thursday for violating misinformation policies, the company told Digital Trends.

The 25-minute clip from a supposed upcoming documentary was posted on Facebook on Monday and had racked up 1.8 million views, including 17,000 comments and nearly 150,000 shares.

Read more
Facebook shoots down Pages, ads sharing vaccine misinformation

Tired of vaccine myths in your newsfeed? Facebook is making them less prominent. On Thursday, March 7, Facebook shared a list of changes focused on fighting vaccine misinformation on the platform. Like “news” proven false by third-party platforms, Facebook will make debunked vaccine myths less prominent in the news feed.

The changes apply to what Facebook describes as “verifiable vaccine hoaxes,” recognized by organizations like the U.S. Centers for Disease Control and Prevention and the World Health Organization (WHO). Pages and Groups sharing these hoaxes, the platform says, will be penalized in the News Feed.

Read more
X seems to have deleted years of old Twitter images
The new X sign replacing the Twitter logo on the company's headquarters in San Francisco.

The social media platform formerly known as Twitter and recently rebranded as X appears to be having trouble showing images posted on the site between 2011 and 2014.

The issue came to widespread attention on Saturday when X user Tom Coates noted how the famous selfie posted by Ellen DeGeneres at the Oscars in 2014, which quickly broke the “most retweets” record, was no longer displaying. Later reports suggested the image had been restored, though, at the time of writing, we’re not seeing it.

Read more