Russian web brigades (also called Russian trolls, Russian bots, Kremlinbots, Kremlins or Rustapar) are state-sponsored anonymous Internet political commentators and trolls linked to the Government of Russia.[1][2] Participants report that they are organized into teams and groups of commentators that participate in Russian and international political blogs and Internet forums using sockpuppets, social bots and large-scale orchestrated trolling and disinformation campaigns to promote pro-Putin and pro-Russian propaganda.[1][3][4][5][6] Articles on the Russian Wikipedia concerning the MH17 crash and the Russo-Ukrainian War were targeted by Russian internet propaganda outlets.[1][7][8][9] In June 2019 a group of 12 editors introducing coordinated pro-government and anti-opposition bias was blocked on the Russian-language Wikipedia.[10] During the war by Russia against Ukraine in 2022, Kremlin trolls were still active on many social platforms and were spreading disinformation related to the war events.[11]
Kremlin trolls are closely tied to the Internet Research Agency, a Saint Petersburg-based company run by Yevgeny Prigozhin, who is a close ally to Vladimir Putin and head of the mercenary Wagner Group, known for committing war crimes.
The earliest documented allegations of the existence of "web brigades" appear to be in the April 2003 Vestnik Online article "The Virtual Eye of Big Brother" by French journalist Anna Polyanskaya (a former assistant to assassinated Russian politician Galina Starovoitova[12]) and two other authors, Andrey Krivov and Ivan Lomako. The authors claim that up to 1998, contributions to forums on Russian Internet sites (Runet) predominantly reflected liberal and democratic values, but after 2000, the vast majority of contributions reflected totalitarian values. This sudden change was attributed to the appearance of teams of pro-Russian commenters who appeared to be organized by the Russian state security service.[13][14][15][16] According to the authors, about 70% of Russian Internet posters were of generally liberal views prior to 1998–1999, while a surge of "antidemocratic" posts (about 60–80%) suddenly occurred at many Russian forums in 2000. This could also be a reflection to the fact that access to Internet among the general Russian population soared during this time, which was until then accessible only to some sections of the society.
In January 2012, a hacktivist group calling itself the Russian arm of Anonymous published a massive collection of email allegedly belonging to former and present leaders of the pro-Putin youth organization Nashi (including a number of government officials).[17] Journalists who investigated the leaked information found that the pro-Putin movement had engaged in a range of activities including paying commentators to post content and hijacking blog ratings in the fall of 2011.[18][19] The e-mails indicated that members of the "brigades" were paid 85 rubles (about US$3) or more per comment, depending on whether the comment received replies. Some were paid as much as 600,000 roubles (about US$21,000) for leaving hundreds of comments on negative press articles on the internet, and were presented with iPads. A number of high-profile bloggers were also mentioned as being paid for promoting Nashi and government activities. The Federal Youth Agency, whose head (and the former leader of Nashi) Vasily Yakemenko was the highest-ranking individual targeted by the leaks, refused to comment on the authenticity of the e-mails.[17][20]
In 2013, a Freedom House report stated that 22 of 60 countries examined have been using paid pro-government commentators to manipulate online discussions, and that Russia has been at the forefront of this practice for several years, along with China and Bahrain.[21][22] In the same year, Russian reporters investigated the St. Petersburg Internet Research Agency, which employs at least 400 people. They found that the agency covertly hired young people as "Internet operators" paid to write pro-Russian postings and comments, smearing opposition leader Alexei Navalny and U.S. politics and culture.[23][24]
“ | Each commenter was to write no less than 100 comments a day, while people in the other room were to write four postings a day, which then went to the other employees whose job was to post them on social networks as widely as possible.[23] | ” |
Some Russian opposition journalists state that such practices create a chilling effect on the few independent media outlets remaining in the country.[22]
Further investigations were performed by Russian opposition newspaper Novaya Gazeta and Institute of Modern Russia in 2014–15, inspired by the peak of activity of the pro-Russian brigades during the Russo-Ukrainian War and assassination of Boris Nemtsov.[25][26][27][28] The effort of using "troll armies" to promote Putin's policies is reported to be a multimillion-dollar operation.[29] According to an investigation by the British The Guardian newspaper, the flood of pro-Russian comments is part of a coordinated "informational-psychological war operation".[30] One Twitter bot network was documented to use more than 20,500 fake Twitter accounts to spam negative comments after the death of Boris Nemtsov and events related to the Ukrainian conflict.[31][32]
An article based on the original Polyanskaya article, authored by the Independent Customers' Association, was published in May 2008 at Expertiza.Ru. In this article the term web brigades is replaced by the term Team "G".[33][34]
During his presidency, Donald Trump retweeted a tweet by a fake account operated by Russians. In 2017, he was among almost 40 celebrities and politicians, along with over 3,000 global news outlets, identified to have inadvertently shared content from Russian troll-farm accounts.[35]
Web brigades commentators sometimes leave hundreds of postings a day that criticize the country's opposition and promote Kremlin-backed policymakers.[19][22][23][24][36][37] Commentators simultaneously react to discussions of "taboo" topics, including the historical role of Soviet leader Joseph Stalin , political opposition, dissidents such as Mikhail Khodorkovsky, murdered journalists, and cases of international conflict or rivalry (with countries such as Estonia, Georgia, and Ukraine , but also with the foreign policies of the United States and the European Union).[19] Prominent journalist and Russia expert Peter Pomerantsev believes Russia's efforts are aimed at confusing the audience, rather than convincing it. He states that they cannot censor information but can "trash it with conspiracy theories and rumours".[24]
To avert suspicions, the users sandwich political remarks between neutral articles on travelling, cooking and pets.[24] They overwhelm comment sections of media to render meaningful dialogue impossible.[38][39]
“ | The effect created by such Internet trolls is not very big, but they manage to make certain forums meaningless because people stop commenting on the articles when these trolls sit there and constantly create an aggressive, hostile atmosphere toward those whom they don’t like. The trolls react to certain news with torrents of mud and abuse. This makes it meaningless for a reasonable person to comment on anything there.[23] | ” |
A collection of leaked documents, published by Moy Rayon, suggests that work at the "troll den" is strictly regulated by a set of guidelines. Any blog post written by an agency employee, according to the leaked files, must contain "no fewer than 700 characters" during day shifts and "no fewer than 1,000 characters" on night shifts. Use of graphics and keywords in the post's body and headline is also mandatory. In addition to general guidelines, bloggers are also provided with "technical tasks" – keywords and talking points on specific issues, such as Ukraine, Russia's internal opposition and relations with the West.[24] On an average working day, the workers are to post on news articles 50 times. Each blogger is to maintain six Facebook accounts publishing at least three posts a day and discussing the news in groups at least twice a day. By the end of the first month, they are expected to have won 500 subscribers and get at least five posts on each item a day. On Twitter, the bloggers are expected to manage 10 accounts with up to 2,000 followers and tweet 50 times a day.[37]
In 2015, Lawrence Alexander disclosed a network of propaganda websites sharing the same Google Analytics identifier and domain registration details, allegedly run by Nikita Podgorny from Internet Research Agency. The websites were mostly meme repositories focused on attacking Ukraine, Euromaidan, Russian opposition and Western policies. Other websites from this cluster promoted president Putin and Russian nationalism, and spread alleged news from Syria presenting anti-Western and pro-Bashar al-Assad viewpoints.[40][41]
In August 2015, Russian researchers correlated Google search statistics of specific phrases with their geographic origin, observing increases in specific politically loaded phrases (such as "Poroshenko", "Maidan", "sanctions") starting from 2013 and originating from very small, peripheral locations in Russia, such as Olgino, which also happens to be the headquarters of the Internet Research Agency company.[42] The Internet Research Agency also appears to be the primary sponsor of an anti-Western exhibition Material Evidence.[43]
Since 2015, Finnish reporter Jessikka Aro has inquired into web brigades and Russian trolls.[44] In addition, Western journalists have referred to the phenomenon and have supported traditional media.[45][46]
In May 2019, it was reported that a study from the George Washington University found that Russian Twitter bots had tried to inflame the United States' anti-vaccination debate by posting opinions on both sides in 2018.[47]
In June 2019 a group of 12 editors introducing coordinated pro-government and anti-opposition bias was blocked on the Russian-language Wikipedia.[10] In July 2019 two operatives of the Internet Research Agency were detained in Libya and charged with attempting to influence local elections.[48] They were reportedly employees of Alexander Malkevich, manager of USA Really, a propaganda website.[49]
In 2020, the research firm Graphika published a report detailing one particular Russian disinformation group codenamed "Secondary Infektion" (alluding to 80's Operation Infektion) operating running since 2014. Over 6 years the group published over 2,500 items in seven languages and to over 300 platforms such as social media (Facebook, Twitter, YouTube, Reddit) and discussion forums. The group specialized in highly divisive topics regarding immigration, environment, politics, international relations and frequently used fake images presented as "leaked documents".[50]
Starting in February 2022, a special attempt was made to back the Russian war in Ukraine. Particular effort was made to target Facebook and YouTube.
In May 2022, during the Russian war with Ukraine the trolls allegedly hired by Internet Research Agency (IRA) had reportedly extended their foothold into TikTok, spreading misinformation on war events and attempting to question or sow doubt about the Ukraine war.[11] Authentic-looking profiles had allegedly hundreds of thousands of followers.[lower-alpha 1][11] IRA was reported to be active across different platforms, including Instagram and Telegram.[11]
Other countries and businesses have used paid Internet commenters to influence public opinion in other countries; some examples are below. (Note: most operations in the list below are smaller in scale than Russia's web brigades, and not all have involved the spread of misinformation).
Original source: https://en.wikipedia.org/wiki/Russian web brigades.
Read more |
Categories: [Internet manipulation and propaganda] [Public relations techniques] [Cyberwarfare] [Internet trolling]