기사검색

On Twitter, bots spread conspiracy theories and QAnon talking points

가 -가 +

강승환 기자
기사입력 2020/10/30 [11:16]

Americans who seek political insight and information on Twitter should know how much of what they are seeing is the result of automated propaganda campaigns.

 

Nearly four years after my collaborators and I revealed how automated Twitter accounts were distorting online election discussions in 2016, the situation appears to be no better. That’s despite the efforts of policymakers, technology companies and even the public to root out disinformation campaigns on social media.

 

In our latest study, we collected 240 million election-related tweets mentioning presidential candidates and election-related keywords, posted between June 20 and Sept. 9, 2020. We looked for activity from automated (or bot) accounts, and the spread of distorted or conspiracy theory narratives.

 

We learned that on Twitter, many conspiracy theories, including QAnon, may not be quite as popular among real people as media reports indicate. But automation can significantly increase the distribution of these ideas, inflating their power by reaching unsuspecting users who may be drawn in not by posts from their fellow humans, but from bots programmed to spread the word.

 

Bots amplify conspiracy theories

 

Typically, bots are created by people or groups who want to amplify certain ideas or points of view. We found that bots are roughly equally active in online discussions of both right-wing and left-wing perspectives, making up about 5% of the Twitter accounts active in those threads.

 

Bots appear to thrive in political groups discussing conspiracy theories, making up nearly 13% of the accounts tweeting or retweeting posts with conspiracy theory-related hashtags and keywords.

 

Then we looked more closely at three major categories of conspiracies. One was a category of alleged scandals described using the suffix “-gate,” such as “Pizzagate” and “Obamagate.” The second was COVID-19-related political conspiracies, such as biased claims that the virus was deliberately spread by China or that it could be spread via products imported from China. The third was the QAnon movement, which has been called a “collective delusion” and a “virtual cult.”

 

These three categories overlap: Accounts tweeting about material in one of them were likely to also tweet about material in at least one of the others.

 

The link to right-wing media

 

We found that the accounts that are prone to share conspiratorial narratives are significantly more likely than nonconspirator accounts to tweet links to, or retweet posts from, right-leaning media such as One America News Network, Infowars and Breitbart.

 

Bots play an important role as well: More than 20% of the accounts sharing content from those hyperpartisan platforms are bots. And most of those accounts also distribute conspiracy-related content.

 

Twitter has recently tried to limit the spread of QAnon and other conspiracy theories on its site. But that may not be enough to stem the tide. To contribute to the global effort against social media manipulation, we have publicly released the dataset used in our work to assist future studies.

트위터 페이스북 카카오톡 카카오스토리 band naver URL복사

뉴스레터 구독하기

세상을 바꾸고 있는 블록체인과 IT 관련 이야기를 쉽고 재미있게 만나보세요.

개인정보 수집 및 이용

뉴스레터 발송을 위한 최소한의 개인정보를 수집하고 이용합니다. 수집된 정보는 발송 외 다른 목적으로 이용되지 않으며, 서비스가 종료되거나 구독을 해지할 경우 즉시 파기됩니다.

최신기사

URL 복사
x
  • 위에의 URL을 누르면 복사하실수 있습니다.

PC버전 맨위로

Copyright ⓒ 코인캣미디어. All rights reserved.