Skip to main content

Facebook is training an army of malicious bots to research anti-spam methods

Despite Facebook’s many efforts, bad actors somehow always manage to seep through its safeguards and policies. The social network is now experimenting with a new way to buttress its anti-spam walls and preempt bad behaviors that could potentially breach its safeguards: An army of bots.

Facebook says it’s developing a new system of bots that can simulate bad behaviors and stress-test its platform to unearth any flaws and loopholes. These automated bots are trained and taught how to act like a real person utilizing the treasure trove of behavior models Facebook has acquired from its over two billion users.

To ensure this experiment doesn’t interfere with real users, Facebook has also built a sort of parallel version of its social network. Here, the bots are let loose and allowed to run rampant — they can message each other, comment on dummy posts, send friend requests, visit pages, and more. More importantly, these A.I. bots are programmed to simulate extreme scenarios such as selling drugs and guns to test how Facebook’s algorithms would try to prevent them.

Facebook claims this new system can host “thousands or even millions of bots.” Since it runs on the same code users actually experience, it adds that “the bots’ actions are faithful to the effects that would be witnessed by real people using the platform.”

“While the project is in a research-only stage at the moment, the hope is that one day it will help us improve our services and spot potential reliability or integrity issues before they affect real people using the platform.” wrote the project’s lead, Mark Harman in a blog post.

It’s unclear at the moment how effective Facebook’s new simulation environment will be. As Harman mentioned, it’s still in rather early stages and the company hasn’t put any of its outcomes to use for public-facing updates just yet. Over the last few years, the social network has actively invested and supported artificial intelligence-based research to develop new tools for fighting harassment and spam. At its annual developer conference two years ago, Mark Zuckerberg announced the company is building artificial intelligence tools for tackling posts that feature terrorist content, hate speech, spam, and more.

Editors' Recommendations

Shubham Agarwal
Shubham Agarwal is a freelance technology journalist from Ahmedabad, India. His work has previously appeared in Firstpost…
WhatsApp now lets you send self-destructing voice messages
WhatsApp logo on a phone.

If you’re on WhatsApp and regularly make use of the view once feature for photo and video messages, then you might be interested to learn that the feature has now been expanded to voice messages.

WhatsApp’s view once feature does what it says, deleting a message after it’s been viewed a single time. It’s been available for photos and videos since 2021, but now you can also send voice messages that can only be played once before they, too, disappear from the app.

Read more
X rival Threads could be about to get millions of more users
Instagram Threads app.

Threads -- Meta’s rival to X, formerly Twitter -- has just launched in the European Union (EU), a market with nearly half a billion people.

The app launched in the U.S. to much fanfare in July, with Meta hoping to attract X users disillusioned with the turbulence on the platform since Elon Musk acquired it for $44 billion 14 months ago.

Read more
X (formerly Twitter) returns after global outage
A white X on a black background, which could be Twitter's new logo.

X, formerly known as Twitter, went down for about 90 minutes for users worldwide early on Thursday ET.

Anyone opening the social media app across all platforms was met with a blank timeline. On desktop, users saw a message that simply read, "Welcome to X," while on mobile the app showed suggestions for accounts to follow.

Read more