Skip to main content

Facebook says white supremacists ‘cannot have a presence’ on the social network

Facebook removed more than 200 white supremacist organizations from its platform for violating both terrorism and hate speech community standards, a Facebook representative told Digital Trends, as part of a broader crackdown on harmful content.

Recommended Videos

“If you are a member or a leader of one of these groups, you cannot have a presence on Facebook,” Sarah Pollack, a Facebook company spokesperson, said Wednesday. The classifications of “terrorism” and “hate speech” are based on behavior, she said.

Please enable Javascript to view this content

“Based on their behavior, some of the white supremacist organizations we’ve banned under our Dangerous Individuals and Organizations policy were banned as terrorist organizations, the others were banned as organized hate groups.”

Previously, Facebook concentrated on ISIS, Al-Qaeda, and their affiliates, and now has expanded the definition, meaning some white supremacist groups are now included.

In addition, Pollack said, “other people cannot post content, supporting praising or representing [those groups]. This is for groups, individuals, as well as attacks that are claimed by these groups.” Pollack also said that mass shooters would fall under this category.

In total, Facebook says it removed more than 22.8 million pieces of content from Facebook and around 4.4 million posts from Instagram in the second and third quarter of 2019 for violating their community standards, according to its fourth-ever Community Standard Enforcement Report that was released Wednesday.

Facebook also said it was able to “proactively detect” more than 96% of the offensive content it took down from Facebook and more than 77% of that from Instagram.

This marked the first time Instagram was included in the Community Standards Report, as well as the first time that Facebook included data on suicide and self-injury.

“We work with experts to ensure everyone’s safety is considered,” Guy Rosen, the vice president of Facebook Integrity, said in a statement. “We remove content that depicts or encourages suicide or self-injury, including certain graphic imagery and real-time depictions that experts tell us might lead others to engage in similar behavior.”

Rosen also wrote that Facebook had “made improvements to our technology to find and remove more violating content,” including expanding data on terrorist propaganda. Facebook said they had identified a wide range of groups as terrorist organizations and had managed to proactively remove, they said, 99% of the content associated with Al-Qaeda, ISIS, and their affiliates.

In addition, they said they had expanded their efforts to proactively detect and remove 98.5% of Facebook content from “all terrorist organizations” beyond just Al-Qaeda and ISIS, and 92.2% of similar posts on Instagram.

The platform also said it removed 11.6 million posts from Facebook in Q3 of 2018 that death with child nudity and exploitation, 99% of which was detected proactively, and significantly higher than the 5.8 million posts removed in Q1 of 2018. On Instagram, they said they removed around 1.3 million posts.

On the illicit sale of drugs and firearms, Facebook also said it had greatly improved its rate of take-down from Q1 to Q3, removing a total of 5.9 million posts from Facebook and around 2.4 million from Instagram.

Maya Shwayder
I'm a multimedia journalist currently based in New England. I previously worked for DW News/Deutsche Welle as an anchor and…
I paid Meta to ‘verify’ me — here’s what actually happened
An Instagram profile on an iPhone.

In the fall of 2023 I decided to do a little experiment in the height of the “blue check” hysteria. Twitter had shifted from verifying accounts based (more or less) on merit or importance and instead would let users pay for a blue checkmark. That obviously went (and still goes) badly. Meanwhile, Meta opened its own verification service earlier in the year, called Meta Verified.

Mostly aimed at “creators,” Meta Verified costs $15 a month and helps you “establish your account authenticity and help[s] your community know it’s the real us with a verified badge." It also gives you “proactive account protection” to help fight impersonation by (in part) requiring you to use two-factor authentication. You’ll also get direct account support “from a real person,” and exclusive features like stickers and stars.

Read more
Here’s how to delete your YouTube account on any device
How to delete your YouTube account

Wanting to get out of the YouTube business? If you want to delete your YouTube account, all you need to do is go to your YouTube Studio page, go to the Advanced Settings, and follow the section that will guide you to permanently delete your account. If you need help with these steps, or want to do so on a platform that isn't your computer, you can follow the steps below.

Note that the following steps will delete your YouTube channel, not your associated Google account.

Read more
How to download Instagram photos for free
Instagram app running on the Samsung Galaxy Z Flip 5.

Instagram is amazing, and many of us use it as a record of our lives — uploading the best bits of our trips, adventures, and notable moments. But sometimes you can lose the original files of those moments, leaving the Instagram copy as the only available one . While you may be happy to leave it up there, it's a lot more convenient to have another version of it downloaded onto your phone or computer. While downloading directly from Instagram can be tricky, there are ways around it. Here are a few easy ways to download Instagram photos.

Read more