Skip to main content

Here’s what social media giants are doing to keep extremism off your screen

Social media is a powerful tool for groups engaged in terrorist activities. The extremist content they post have sparked widespread changes across social networks. But, are those changes enough? That’s the question representatives from Facebook, Twitter, and YouTube addressed this week, speaking before the U.S. Senate Committee on Commerce, Science, and Transportation in a hearing in Washington, D.C.

The hearing was designed to look into the social networks’ current efforts into curbing extremist content, opening up a discussion on tech companies’ role in stunting the spread of online propaganda. While the companies have previously testified on Russian interference in the U.S. election, this hearing was the first time the companies spoke to the commerce committee on extremist content.

Recommended Videos

All three networks demonstrated a significant increase in the number of content removed from their respective platforms, as well as preventing the information in the first place. In some cases, the networks’ efforts overlap, including the Global Internet Forum to Counter Terrorism for information sharing, while a database of more than 40,000 “hashes” helps keep content recognized on one network, off another.

Facebook

Facebook’s head of Product Policy and Counterterrorism, Monika Bickert, said that Facebook is now able to remove 99 percent of ISIS and Al Qaeda-related posts before reaching a human flagger, thanks largely to machine learning; Facebook’s AI platform looks through image, video, and text material. The company is also working to teach the system how to recognize posts that support a terrorist organization (rather than generating false positives on posts condoning the behavior, for example).

For Facebook, AI is also being used to prevent some content uploads. Image matching prevents other accounts from uploading videos previously removed by the company. The company also works with experts “to track propaganda released by these groups and proactively insert it into our matching systems,” Bickert wrote in a prepared statement.

Facebook also looks for “clusters” or related Pages, groups, posts, and profiles tied to the removed account. The social network is also improving efforts in keeping users previously removed from creating a new account.

Facebook has already added 3,000 people to the review team and this year will expand to a total of 20,000 people working to identify all content that violates the community standards, including extremist content. Another 180 people, Bickert said, are trained specifically in preventing terrorist content.

At the same time, Facebook is working to further “counterspeech,” or content that fights against extremism and other hateful posts.

“On terrorist content, our view is simple: There is no place on Facebook for terrorism,” Bickert said. “Our longstanding policies, which are posted on our site, make clear that we do not allow terrorists to have any presence on Facebook. Even if they are not posting content that would violate our policies, we remove their accounts as soon as we find them.”

Twitter

Twitter’s director of Public Policy and Philanthropy, Carlos Monje Jr., said the platform has now suspended more than one million accounts for terrorism since mid-2015 — including 574,070 accounts just last year, a jump from the more than 67,000 suspensions in 2015. A big part of that increase is the technology used to detect those accounts, which caught one-third of the accounts in 2015 but is now responsible for 90 percent of the latest suspensions.

“While there is no ‘magic algorithm’ for identifying terrorist content on the internet, we have increasingly tapped technology in efforts to improve the effectiveness of our in-house proprietary anti-spam technology,” Monje said. “This technology supplements reports from our users and dramatically augments our ability to identify and remove violative content from Twitter.”

Extremist content was part of Twitter’s rule overhaul late last year prompted in part by #womenboycotttwitter. Those expanded rules went beyond Tweets to include handles, profile images and other profile information.

On a different note, the platform is also working to prevent election misinformation and will soon show users if they viewed that propaganda — along with donating money from those ads to conduct additional research. While Twitter has already shared updates designed specifically for political ads, verifying all state and federal candidates is part of those changes as well.

YouTube

Juniper Downs, YouTube’s director of Public Policy and Government Relations, said machine learning now removes 98 percent of “violent extremism” videos, up from 40 percent a year ago. Around 70 percent is removed within eight hours and half in under two, Downs said.

Along with the expanded software, YouTube has also added additional organizations to the Trusted Flagger program, including counter-terrorism groups. Within parent company Google itself, the number of staff working with those videos in violation will grow to 10,000 this year. This year will also bring a transparency report on flagged videos.

For videos that fall in a more gray area without an outright violation, YouTube has already announced these types of videos won’t receive monetary compensation or be part of the recommended videos, along with removing options for comments. Like Facebook, counter-speech is also part of the initiative, including the Creators for Change program.

“No single component can solve this problem in isolation,” Downs wrote in her prepared statement. “To get this right, we must all work together.”

Moving forward

While the session has been described as “mostly genial” with each platform reporting higher numbers of removed content and accounts, Clint Watts, a Robert A. Fellow for the Foreign Policy Research Institute, suggested that social networks can do more by reconsidering anonymous accounts and eliminating non-human bot accounts or requiring a CAPTCHA, while federal regulations for political ads should also be extended to social media.

“Social media companies realize the damage of these bad actors far too late,” Watts wrote in a prepared statement. “They race to implement policies to prevent the last information attack, but have yet to anticipate the next abuse of their social media platforms by emerging threats seeking to do bad things to good people.”

A video of the hearing is publicly available from the committee’s website, including prepared statements from each network.

Hillary K. Grigonis
Hillary never planned on becoming a photographer—and then she was handed a camera at her first writing job and she's been…
Bluesky finally adds a feature many had been waiting for
A blue sky with clouds.

Bluesky has been making a lot of progress in recent months by simplifying the process to sign up while at the same time rolling out a steady stream of new features.

As part of those continuing efforts, the social media app has just announced that users can now send direct messages (DMs).

Read more
Reddit just achieved something for the first time in its 20-year history
The Reddit logo.

Reddit’s on a roll. The social media platform has just turned a profit for the first time in its 20-year history, and now boasts a record 97.2 million daily active users, marking a year-over-year increase of 47%. A few times during the quarter, the figure topped 100 million, which Reddit CEO and co-founder Steve Huffman said in a letter to shareholders had been a “long-standing milestone” for the site.

The company, which went public in March, announced the news in its third-quarter earnings results on Tuesday.

Read more
Worried about the TikTok ban? This is how it might look on your phone
TikTok splash screen on an Android phone.

The US Supreme Court has decided to uphold a law that would see TikTok banned in the country on January 19. Now, the platform has issued an official statement, confirming that it will indeed shut down unless it gets some emergency relief from the outgoing president.

“Unless the Biden Administration immediately provides a definitive statement to satisfy the most critical service providers assuring non-enforcement, unfortunately TikTok will be forced to go dark on January 19,” said the company soon after the court’s verdict.
So, what does going dark mean?
So, far, there is no official statement on what exactly TikTok means by “going dark.” There is a lot of speculation out there on how exactly the app or website will look once TikTok shutters in the US.

Read more