QAnon is facing a wave of bans by tech companies. These are the platforms that have and haven't responded to the conspiracy theory.

Advertisement
QAnon is facing a wave of bans by tech companies. These are the platforms that have and haven't responded to the conspiracy theory.
  • Social-media platforms have been slow to act against the QAnon conspiracy theory, with many only acting to curb the ballooning movement within the last few weeks.
  • QAnon followers have been linked to several alleged crimes, including killings and attempted kidnappings.
  • Now, tech companies, including Facebook, TikTok, Triller, Etsy, and others, say they're making efforts to rid their platforms of the unfounded conspiracy theory movement.

The QAnon conspiracy theory was born on the internet, and while it's spread to real-life rallies in the US and abroad, it's continued to thrive and spread online. And yet social-media platforms, where the conspiracy theory gains power and radicalizes people in the US abroad, have been generally slow to act on banning it.

QAnon is a baseless far-right conspiracy theory that claims President Donald Trump is fighting a deep-state cabal of elite figures who are involved with human trafficking. It is unfounded, and yet its followers — estimated to be in the millions — have reportedly been linked to several alleged crimes, including killings and attempted kidnappings. In 2019, a bulletin from the FBI field office in Phoenix warned that the conspiracy theory movement could become a domestic terrorism threat.

Many platforms, including Facebook, are finally taking steps to combat the spread of QAnon-related misinformation ahead of the November election.

Here's how major tech companies are handling the spread of the QAnon conspiracy theory online.

Advertisement

Facebook said its companies are cracking down on QAnon.

Facebook said its companies are cracking down on QAnon.
In this photo illustration a Facebook logo seen displayed on a smartphone. Photo Illustration by Rafael Henrique/SOPA Images/LightRocket via Getty Images

On October 6, Facebook announced it would remove all pages, groups, and Instagram accounts that promoted QAnon.

The ban, which the company said would be enacted gradually, comes after the platform previously announced over the summer that it had removed 790 QAnon Facebook groups.

Extremism researchers are tracking how the new ban will play out, as the movement has spread rapidly on Facebook and on Instagram, where many are using "Save the Children" rhetoric to further propagate the movement's misguided focus on human trafficking conspiracy theories.

Facebook has been criticized for its slowness in acting against QAnon.

Twitter announced a moderation plan on QAnon in July.

Twitter announced a moderation plan on QAnon in July.
Rafael Henrique/SOPA Images/LightRocket via Getty Images

In July, Twitter announced it would begin cracking down on QAnon content by suspending accounts that were "engaged in violations of our multi-account policy, coordinating abuse around individual victims, or are attempting to evade a previous suspension."

The platform said it would also stop recommending QAnon-related accounts and trends and block URLs associated with QAnon from being shared on Twitter.

Critics have said the platform was slow to act on the movement and hasn't moderated the community enough. On October 3, The Washington Post reported that there were still 93,000 active Twitter accounts referencing QAnon in their profiles, citing research from Advance Democracy, a nonpartisan research group.

Advertisement

YouTube announced a crackdown on QAnon after significant pressure, but stopped short of an explicit ban.

YouTube announced a crackdown on QAnon after significant pressure, but stopped short of an explicit ban.
Mateusz Slodkowski/Getty Images

YouTube, where QAnon videos that spread the conspiracy theory have thrived and gained millions of views, said on Thursday that it would prohibit QAnon content that threatens violence against a group or individual.

The move is part of an addition to the company's policies against hate and harassment that is focused on "conspiracy theories that have been used to justify real-world violence."

"We will begin enforcing this updated policy today, and will ramp up in the weeks to come," the company said on Thursday. "Due to the evolving nature and shifting tactics of groups promoting these conspiracy theories, we'll continue to adapt our policies to stay current and remain committed to taking the steps needed to live up to this responsibility."

The announcement, which stops short of a ban, already led to the removal of two top QAnon accounts from the platform, NBC News reported.

YouTube CEO Susan Wojcicki had previously told CNN on Monday that the platform had not made a policy banning the conspiracy theory movement. "We're looking very closely at QAnon," Wojcicki said. "We already implemented a large number of different policies that have helped to maintain that in a responsible way."

YouTube has faced criticism for the way its algorithm can send users down rabbit holes of extremism and radicalization, often recommending QAnon-related videos and other videos containing misinformation.

TikTok has disabled popular QAnon hashtags.

TikTok has disabled popular QAnon hashtags.
While President Donald Trump has been the one to target Chinese-based TikTok, Democratic nominee Joe Biden's campaign has steered clear of the platform. Jaap Arriens/NurPhoto via Getty Images

TikTok announced in July that it disabled the hashtag pages for "QAnon" and "WWG1WGA" in an effort to curb the theory's spread.

Both hashtags had amassed tens of millions of views on the app before they were banned, Insider reported.

Advertisement

Triller said it would ban QAnon after it defended the conspiracy theory's presence on the app.

Triller said it would ban QAnon after it defended the conspiracy theory's presence on the app.
Photo Illustration by Rafael Henrique/SOPA Images/LightRocket via Getty Images

Triller, the TikTok rival app that previously hosted a slew of QAnon-related content, told Business Insider on Tuesday that the platform banned QAnon. Hashtag pages for "QAnon" and "QAnonBeliever" were unsearchable as of Tuesday, but "WWG1WGA" remained available as a hashtag page.

The company said its recent decision was based on the FBI's 2019 reference to QAnon as a possible domestic terrorism threat.

"We are a platform that believes in freedom of speech, expression, open discussion and freedom of opinion, however when the government classifies something as a terrorist threat we must take action to protect our community," CEO Mike Lu said in a statement.

The announcement was in contrast with what Triller majority owner Ryan Kavanaugh told The New York Times just one week before. Kavanaugh said Triller did not intend to moderate QAnon, saying that "if it's not illegal, if it's not unethical, it doesn't harm a group, and it's not against our terms of service, we're not going to filter or ban it."

Etsy, which previously hosted QAnon merchandise from third-party sellers, banned QAnon products.

Etsy, which previously hosted QAnon merchandise from third-party sellers, banned QAnon products.
QAnon conspiracy theorists hold signs during the protest at the State Capitol in Salem, Oregon, United States on May 2, 2020. Photo by John Rudoff/Anadolu Agency via Getty Images

An Etsy spokesperson told Insider on October 7 that the company was "in the process of removing QAnon merchandise" from the platform.

"Etsy is firmly committed to the safety of our marketplace and fostering an inclusive environment. Our seller policies prohibit items that promote hate, incite violence, or promote or endorse harmful misinformation. In accordance with these policies, we are removing items related to 'QAnon' from our marketplace," the spokesperson said in a statement.

Etsy had been a hotbed for QAnon merchandise, including stickers, hats, apparel, and other items sold by third-party sellers on the retail marketplace.

Advertisement

Amazon has not taken a stance on QAnon, and continue to host and recommend QAnon products.

Amazon has not taken a stance on QAnon, and continue to host and recommend QAnon products.
Screenshots show QAnon-related products available for sale on Amazon. Amazon; Business Insider

Amazon continues to host more than 1,000 QAnon-related items for sale from third-party sellers.

When asked if the platform would remove these products, which include more than a dozen books filled with misinformation about the theory, an Amazon spokesperson said the company had no comment.

The Amazon QAnon product landscape resembles YouTube's infamous related video recommendation algorithm. A search for "QAnon" on Amazon offers related search terms including "qanon shirt," "qanon flag," "qanon hat" and "qanon stickers and decals."

Fitness company Peloton has removed tags related to QAnon.

Fitness company Peloton has removed tags related to QAnon.
Peloton; Insider

Peloton, the at-home fitness company, has had to moderate its platform for QAnon tags. On Peloton, users can follow tags to see who else in that interest group is in a ride or workout class with them.

A Peloton spokesperson told Insider that the platform had removed QAnon-related tags in accordance with policies against "hateful content."

"We actively moderate our channels and remove anything that violates our policy or does not reflect our company's values of inclusiveness and unity or maintain a respectful environment," the spokesperson said.

Advertisement

Pinterest says it banned QAnon in 2018 and actively seeks out content related to the conspiracy theory to remove.

Pinterest says it banned QAnon in 2018 and actively seeks out content related to the conspiracy theory to remove.
QAnon supporters attend a Trump rally hosted by Long Island and New York City police unions in support of the police on October 4, 2020 in Suffolk County, New York. Andrew Lichtenstein/Corbis via Getty Images

A Pinterest spokesperson told Insider that the platform actually began prohibiting QAnon content in 2018, which is quite early in the conspiracy theory's lifespan, but didn't make an announcement on it.

"We believe in a more inspired internet, and that means being deliberate about creating a safe and positive space for Pinners. Pinterest is not a place for QAnon conspiracy theories or other harmful and misleading content," the spokesperson said.

In addition to seeking QAnon content to remove, the platform disabled search terms for "QAnon" and "WWG1WGA."

Reddit has emerged as one of the few online communities largely devoid of QAnon.

Reddit has emerged as one of the few online communities largely devoid of QAnon.
In this photo illustration, a Reddit logo is seen displayed on a smartphone. Mateusz Slodkowski/SOPA Images/LightRocket via Getty Images

The Atlantic's Kaitlyn Tiffany reported in September that Reddit had successfully rid its platform of most QAnon content, but company leadership wasn't quite sure how.

Chris Slowe, Reddit's chief technology officer, told The Atlantic, "I don't think we've had any focused effort to keep QAnon off the platform."

While QAnon had a big presence on Reddit in its early days, the platform was quick to ban subreddits associated with the theory as early as 2018. "Reddit has plenty of problems, but QAnon isn't one of them," Tiffany reported.

Advertisement

Twitch has managed to keep QAnon content to a minimum with partial suspensions for some far-right accounts.

Twitch has managed to keep QAnon content to a minimum with partial suspensions for some far-right accounts.
Thomas Trutschel/Getty Images

As Kotaku reported in September, Twitch, the streaming platform popular in the gaming community, has managed to keep QAnon communications to a minimum.

The platform temporarily banned Patriots' Soapbox, a notorious QAnon-spreading YouTube channel, from Twitch in September. The channel only has a few thousand followers on Twitch, where it's still active as of Thursday. YouTube removed the group's main channel as part of its new crackdown, though a second version of the channel was still up at publishing time.

Twitch's functionality makes it less of a QAnon target, Nathan Grayson of Kotaku wrote, because there's no recommendation algorithm that leads viewers down rabbit holes.

The platform has not rolled out a specific policy on QAnon, but told Kotaku in September that "the safety of our community is our top priority, and we reserve the right to suspend any account for conduct that violates our rules, or that we determine to be inappropriate, harmful, or puts our community at risk."

A Twitch spokesperson told Insider that the platform's guidelines "prohibit hateful conduct, harassment, and threats of violence," and that Twitch evaluates "all accounts under the same criteria and take action when we have evidence that a user has violated our policies."

Discord said QAnon servers violate its regulations and are removed from the platform.

Discord said QAnon servers violate its regulations and are removed from the platform.
Allmy/Shutterstock

Discord, an instant messaging application, has not made a public statement on its QAnon moderation, but the platform has been suspected of taking down QAnon-related groups. The Discord chat room for listeners of the podcast QAnon Anonymous, which analyzes and debunks the conspiracy theory and its community, has been wrongfully removed from the platform in the past.

There have been several popular Discord chats for QAnon believers, including Patriots' Soapbox, but the platform appeared to have removed Patriots' Soapbox and other groups, according to tweets from members discussing the removal.

A Discord representative told Insider that "QAnon servers violate our community guidelines and we continue to identify and ban these servers and users."

Advertisement

Google banned QAnon merchandise from its shopping tab.

Google banned QAnon merchandise from its shopping tab.
A man wearing a right-wing t-shirt and a couple wearing QAnon shirts face off against riot police on Unter den Linden avenue during protests against coronavirus-related restrictions and government policy on August 29, 2020 in Berlin, Germany. Sean Gallup/Getty Images

Google, which owns YouTube, blocked search results for people seeking QAnon-related merchandise on its "shopping" tab, The Telegraph reported in August.

"We do not allow ads or products that promote hatred, intolerance, discrimination or violence against others," a Google spokesperson told The Telegraph.

A Google representative did not immediately respond to Insider's request for comment regarding QAnon.