This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Information Operations in Turkey:
Manufacturing Resilience with Free Twitter Accounts

Maya Merhi            Sarah Rajtmajer            Dongwon Lee
Abstract

Following the 2016 US elections Twitter launched their Information Operations (IO) hub where they archive account activity connected to state linked information operations. In June 2020, Twitter took down and released a set of accounts linked to Turkey’s ruling political party (AKP). We investigate these accounts in the aftermath of the takedown to explore whether AKP-linked operations are ongoing and to understand the strategies they use to remain resilient to disruption. We collect live accounts that appear to be part of the same network,  30% of which have been suspended by Twitter since our collection. We create a BERT-based classifier that shows similarity between these two networks, develop a taxonomy to categorize these accounts, find direct sequel accounts between the Turkish takedown and the live accounts, and find evidence that Turkish IO actors deliberately construct their network to withstand large-scale shutdown by utilizing explicit and implicit signals of coordination. We compare our findings from the Turkish operation to Russian and Chinese IO on Twitter and find that Turkey’s IO utilizes a unique group structure to remain resilient. Our work highlights the fundamental imbalance between IO actors quickly and easily creating free accounts and the social media platforms spending significant resources on detection and removal, and contributes novel findings about Turkish IO on Twitter.

Introduction

Social networking sites have become the birthplaces of conspiracy theories, battlegrounds for trolling, and the new stage for information operations (IO), which describes “actions taken by governments or organized non-state actors to manipulate public opinion”  (Arif, Stewart, and Starbird 2018). The most widely studied of such has been the Russian Internet Research Agency’s (IRA) campaign aimed at influencing the outcome of the 2016 U.S. Presidential Election. As authorities began investigating this effort in the aftermath of Trump’s victory in 2016, social media corporations were urged to investigate and report on the manipulation that took place  (Starbird, Arif, and Wilson 2019). In their response, Twitter began releasing archives of state-linked IO as they detect and remove them. There are currently archives from 20 different countries that include account information, tweets, and embedded media files  (Twitter 2018).

Amongst these is a large dataset representing an operation Twitter has attributed to the Turkish Justice and Development Party (Adalet ve Kalkınma Partisi; AKP) (Twitter Safety 2020). Turkey has been undergoing a hostile political transformation into an authoritarian regime for the past decade, which became even more severe following a failed coup attempt in 2016  (Yılmaz and Turner 2019). AKP has a well-known history of utilizing social media to spread pro-government content, and utilizing an army of troll accounts commonly known as the AK Trolls to attack those critical of the government  (Saka 2018; Bulut and Yörük 2017; Albayrak and Parkinson 2013). In 2020, Twitter removed 7,340 accounts linked to the AKP from the platform and subsequently shared these accounts and their activity with researchers.

Using these accounts as a starting point, we explore the following research questions in this work: (RQ1) Is there evidence that AKP-linked operations are ongoing following Twitter’s takedown of 7,340 AKP-linked accounts in 2020? (RQ2) What strategies do ongoing AKP-linked operations use to remain resilient to disruption? To address these questions, we collect live accounts that appear to be associated with the operation, including 169 accounts which we suggest have been stood up as direct replacements for accounts suspended in Twitter’s takedown. We find that a rich interaction network underlies the operation and remains robust to node removal over time. Our findings suggest that Turkish IO actors prepare for and adapt to shutdown by creating a strong network through explicit signals of coordination such as hashtags and following patterns, and strategic creation and assignment of new accounts. We compare the Turkish network with known Russian and Chinese IO and find that all three networks exhibit explicit signals of coordination, but the Turkish network utilizes a unique group structure to build resilience. These findings lay groundwork for a broader understanding of how state-linked IO adapt in the face of detection.

Related Work

Information Operations. Social media has inadvertently created direct channels for IO actors to disseminate targeted propaganda to a wide audience. There is evidence of social media manipulation campaigns originating from authoritarian regimes as well as democratic states and targeting both domestic and foreign audiences (Bulut and Yörük 2017; Ferrara 2017; Farkas and Bastos 2018). Authoritarian regimes primarily target domestic audiences through carefully crafted narratives disseminated by fake accounts and legitimate accounts of local government officials (King, Pan, and Roberts 2017). Bradshaw and Howard (2017) found that of 28 countries examined, every authoritarian regime used social media campaigns to target their domestic populations, while almost all democratic nations in their study organize campaigns to target foreign audiences.

IO Strategy and Tactics. Over the past decade, IO have included coordinated efforts within the social media environment in support of strategic aims  (Lin and Kerr 2019; Starbird, Arif, and Wilson 2019; Bradshaw and Howard 2017). These efforts employ fabricated profiles, intricately constructed narratives, and armies of social bots that amplify disinformation, e.g., “troll armies”  (Arif, Stewart, and Starbird 2018; Ferrara 2017; Recuero, Soares, and Gruzd 2020; Bradshaw and Howard 2017; Linvill and Warren 2020).

IO leverage content-based, language-based, cross-platform and multimedia strategies. Content-based strategies include organizing dedicated accounts to post certain content, synchronizing content with real-world events, and fabricating profiles that imitate real people to embed polarizing content in different communities (Farkas and Bastos 2018; Zannettou et al. 2019a; Arif, Stewart, and Starbird 2018). Language-based strategies include the targeting of populations using their native language (i.e., Russian troll accounts tweeting in German (Zannettou et al. 2019a)), and the use of deceptive language and shorter, less-complex language than baseline online users (Addawood et al. 2019). Cross-platform and multimedia strategies include posting links to external websites, posting links to accounts on other social media platforms, and using photos and videos to increase content sharability  (Zannettou et al. 2020, 2019b).

Detecting IO. Few works have used Twitter’s archives of detected IO to classify and detect live IO. Alizadeh et al. (2020) utilize Twitter’s IO archives originating from Russian, China, and Venezuela to build content-based Random Forest classifiers for a number of classification tasks, including using a prior data release to classify trolls in a subsequent data release. Luceri, Giordano, and Ferrara (2020a) employ Inverse Reinforcement Learning (IRL) to identify Russian troll accounts leading up to the 2016 US Presidential Election. They utilize IRL to infer users’ objectives and identify how trolls’ incentives differ from regular online users. Im et al. (2020) utilize Twitter’s first release of ground-truth Russian troll data to build machine learning classifiers of English-speaking Russian trolls targeting US politics.

Turkey: The AK Trolls. In 2013, the Gezi park protests mobilized Turkish citizens to Twitter as a primary news source and mechanism for political discourse and organization due to severe censorship in mainstream media (Yılmaz and Turner 2019; Karatas and Saka 2017). In response, the AKP funded, recruited, and trained an army of 6000 young AKP members to create and disseminate pro-government/AKP content on social media (Albayrak and Parkinson 2013; Saka 2018). These “AK Trolls” use social media to spread AKP ideals through large volumes of messages and images (Saka 2018; Albayrak and Parkinson 2013). They promote unwavering support and praise of President Erdoğan, continuously criticize and demonize opposing political parties, specifically the People’s Republican Party (CHP) and the People’s Democratic Party (HDP), and attack journalists who criticize the government (Grossman et al. 2020; Bulut and Yörük 2017; Saka 2018; Karatas and Saka 2017; Başaran 2020). Twitter’s takedown attributed to the AKP shares these same sentiments and is likely part of the AK Trolls’ larger network of accounts (Grossman et al. 2020). Stanford Internet Observatory released a report on the Turkish takedown and identified tactics such as the use of compromise accounts, highly centralized retweet rings, time-coordinated posting, and coordinated content (Grossman et al. 2020). Karatas and Saka (2017) utilized digital ethnography to examine how the AKP utilizes political trolling strategies, how trolling is coordinated to silence opposing view points, and the effects of AK Trolls on vulnerable and unprotected citizens, e.g., self-censorship and quitting social media. Saka (2018) provides interviews with self-proclaimed AK Trolls; the author finds that political trolling in Turkey “is more decentralized and less institutionalized then generally thought” (Saka 2018, p.1). Bulut and Yörük (2017) provide an analysis of AK Trolls through the lens of mediated populism and demonstrate the increased politicization of Twitter in Turkey following the Gezi Park protests.

Our work contributes a taxonomy of Turkish IO account types and groups memberships, a novel analysis of explicit signals used by IO networks, and evidence that Turkish accounts deliberately construct their network to withstand large scale shutdown.

Dataset

We analyze two datasets and their interactions: first, the set of accounts and content identified by Twitter as a Turkish IO and shared on the company’s IO Hub111https://about.twitter.com/en/our-priorities/civic-integrity“#data (here forward, takedown dataset); second, a network of Turkish accounts that we detected live on Twitter as suspected Turkish IO actors (here forward, live dataset) (see Table 1).

Feature Takedown Live
# of users 7340 7973
# of active users 6270 7502
# of tweets 36,948,536 9,298,325
start date 2009-06-04 2020-01-01
end date 2020-04-21 2021-01-02
Table 1: Overview of takedown and live accounts. Number of active users includes all users who tweeted at least once.

Twitter’s Turkish Takedown

In June 2020, Twitter added 7340 accounts and nearly 37 million tweets attributed to Turkey’s AKP to their archive.222https://blog.twitter.com/en˙us/topics/company/2020/information-operations-june-2020.html The downloadable data hashes the User ID, username, and display name of all users with less than 5000 followers. We use the unhashed data, which researchers can apply to access. Throughout the paper, we suppress any usernames that are hashed in the public version of the dataset.

Live Account Collection

We collect live accounts and tweets suspected to be part of the Turkish IO through analysis of the social networks and common behaviors amongst accounts in Twitter’s takedown dataset. We start with the follow trains and groups observed in the takedown dataset (Grossman et al. 2020). Follow trains are tweets that mention (@) a list of users and request their followers follow these accounts. Of almost 37 million tweets released in Twitter’s takedown dataset, approximately 1.1 million are follow trains.333We calculate this by filtering tweets that contain at least 5 mentions and have a mention to word ratio above 0.8. Many Turkish IO accounts are also members of “groups” and declare their group membership in their profile description. These accounts create “RT” (retweet) and “fav” (favorite) groups to share content and boost followers, engagement, and visibility.

(1) We manually identify 13 “parent” accounts that appear to be part of the operation based on their interactions with and strong similarities to accounts from the takedown. Specifically, these accounts frequently retweeted and mentioned multiple takedown accounts, and likewise were retweeted and mentioned by takedown accounts. They also posted very similar content to accounts in the takedown (similar political images/messages, follow trains, targeted hashtag campaigns). And, they contained other elements common to Turkish IO, e.g., Turkish flag imagery, profile descriptions devoted to the AKP (#AKParti, #RecepTayyipErdoğan), roughly equal followers to friends ratio.444All of the user IDs and tweet IDs we collected are available to the community on Github https://github.com/mayamerhi/Turkey_IO_ICWSM23. Parent accounts are denoted as such.

(2) We collect parent accounts’ followers and friends through Twitter API’s GET friends/list and GET followers/list endpoints. To reduce potential false positives, we remove Recep Tayyip Erdoğan’s account and all accounts he was following at the time– a total of 96 accounts, almost all of which were verified, and all appeared to be real Turkish politicians or organizations.

(3) As our focus was to study operations in response to the takedown, we filter out accounts created before 2020, resulting in 7973 accounts. Twitter announced the takedown on June 11, 2020. Based on date of last tweet, it appears that most accounts were suspended prior to Jan 2020.

(4) To collect tweets from live accounts, we use the Twitter API GET statuses/user_timeline endpoint, which returns up to 3200 of a user’s most recent Tweets. Our tweet collection resulted in 7973 accounts and 9,298,325 associated tweets throughout 2020.

Suspended Accounts

Since our initial data collection in Dec 2020, a significant portion of the accounts we turned up have been suspended. Specifically, as of Dec 6 2021, 2454 (\sim31%) of the accounts in our live dataset been suspended by Twitter.555To determine which, we queried the Twitter API for all users in our live dataset and assigned the label ‘suspended’ for all accounts whose API response was account suspended. Twitter lists the three most common reasons for account suspension as: spam/fake account, account security at risk (hacked or compromised), and abusive tweets or behavior (threats or impersonating other accounts).666https://help.twitter.com/en/managing-your-account/suspended-twitter-accounts As we collected these accounts based on interactions and shared characteristics with the takedown dataset, and as IO routinely engage in these behaviors, we suggest that accounts from our live dataset which have been suspended can be reasonably assumed IO accounts. We denote this subset of 2454 suspended accounts as live suspended.

RQ1: Ongoing Operations

We explore the takedown and live datasets to explore whether the live accounts are part of an ongoing AKP operation. First, we use network analyses to map interactions between users from the takedown and live datasets. Then, we create a BERT-based classifier to distinguish Turkish state linked accounts from “ordinary” Turkish Twitter users and provide further evidence that suspended users in the live dataset are likely Turkish state linked users.

Network Interactions

Figure 1 shows all interactions (mentions, retweets, replies, quotes) between users in both the live and takedown datasets. This includes interactions both within and across datasets. The graph contains a total of 11,551 user nodes and 609,459 directed, weighted interaction edges. The network shows that there are considerably more interactions amongst accounts in each dataset than between them. This is largely due to our collection. Namely, tweets in the takedown dataset span 2009 to early 2020 while the live accounts we pulled were created in 2020, so the two sets of accounts’ activity overlapped only briefly. Specifically, only 1426 users in Twitter’s takedown dataset were active in 2020, and 1422 tweeted their last tweet in Jan 2020. Therefore, most of the overlap between these two datasets happened in only one month (Jan 2020).

Refer to caption
Figure 1: Account interaction graph. Takedown accounts are represented in purple, live accounts (still live as of December 2021) are green, and suspended accounts are red. Edge color represents the color of the target node. If there are bidirectional interactions between users, both edges are present. Edges are weighted by number of interactions.

In Figure 2, we zoom in on this time period to analyze the interactions amongst takedown and live accounts while they were both active. The graph contains 120 nodes and 192 weighted interaction edges. There are 80 nodes from the takedown dataset and 40 nodes from the live dataset, 13 of which have been suspended by Twitter as of Dec 6, 2021. Unlike Figure 1, we do not see two distinct clusters separating the two datasets. Rather, we see a high amount of interconnectivity between the two datasets. This view, we suggest, captures a snapshot of the operation’s activity transition to new accounts in response ongoing shutdown.

Of 1426 users in the takedown dataset active in 2020, 1422 last tweeted in Jan 2020.777Of the remaining users in the takedown set, 2143 were suspended in 2019 and 1239 were suspended in 2018. This indicates two possibilities: Twitter may have suspended many accounts from the takedown dataset at once in Jan 2020, or the Turkish operation was already transitioning to new accounts in anticipation of shutdown. During this time, they began interacting with accounts in the live dataset to boost visibility of newly created accounts through retweets and follow trains. During Jan 2020, takedown accounts interacted with accounts from the live dataset 530 times.888This is greater than the number of edges in Figure 2 because the graph contains weighted edges. Of these 530 tweets, 455 were retweets and 72 were follow trains.

Refer to caption
Figure 2: Transition network. Interaction graph amongst all users in both the takedown and the live datasets during Jan 2020. Node/edge colors follow as in Figure 1.

BERT Classification

We create a BERT-based classifier trained on users from the takedown and evaluated on the live suspended users. We aim to determine whether latent similarities between the two datasets can further substantiate the link between them. We use the pre-trained Turkish language DistilBERTurk model (Schweter 2020). We fine-tune DistilBERTurk using accounts from the takedown as positively labeled samples and we collect negatively labeled samples using the Twitter API. Our classifier yields an F1-score of 0.88 when tested on the takedown dataset, and 0.71 when evaluating the live dataset.

Classifier data. Our positively labeled training dataset contains all tweets associated with the 3569 accounts in the takedown that posted in 2019 and 2020. We collect negatively labeled data to represent ordinary Turkish language Twitter users that have no link to the Turkish state. For a pseudo-random sample of such users, we queried the Twitter API for a list of common Turkish stopwords, while also restricting the search to return tweets written in the Turkish language, between the date ranges of Jan 1 2019 to Jan 1 2021. We collected the timelines for 6571 users. We use 3571 of these users as negatively labeled samples for training, validation, and testing (here after, negative dataset) and 3000 as negatively labeled samples when evaluating the fine-tuned classifier on the live dataset. Since the takedown dataset provides all tweets ever posted by users, and the Twitter API provides only up to 3200 tweets for the live dataset and negative data, we limit the number of tweets for each user to a random sample 2000 of their tweets.

Fine-tuning specifications. We use 3569 users from the takedown dataset as positively labeled samples and 3571 users from the negative dataset as negatively labeled samples to create a balanced dataset of 7140 users for model building. We use a 70-15-15 training, validation, and testing split. We remove only URL links for text pre-processing. We use the auto-tokenizer from the DistilBERTurk model, the Keras implementation of the Adam optimizer with a learning rate of 5e55e^{-5}, binary cross-entropy loss function, batch size of 16, and 2 training epochs. After training, validation, and testing, we run the fine-tuned model on the live suspended dataset to evaluate if the model trained on takedown users distinguishes live suspended users from negative users. Table 2 gives the model’s performance.

BERT results. Our fine-tuned DistilBERTurk model performs well on the takedown users during testing and shows reasonable discernment between these users and the negative users. The model also performs well when evaluated on the live suspended users. These model results provide further evidence that users in the live dataset are likely Turkish state-linked users. The classifier correctly classifies 79% of the live suspended accounts in the positive class, i.e., as Turkish state linked accounts rather than the ordinary Turkish Twitter users.999Of our 7973 collected users, 7502 of were active at the time of collection. Of the 2454 suspended users, 2093 tweeted at least once. We removed all users with no tweets from the classification task. Our classifier predicts that 1648 of the 2093 suspended accounts we predict on are in the positive class. Of the remaining users in the live dataset, the classifier predicts 4181 users in the positive class and 1228 users in the negative class. Of these 1228 users that have not been suspended and which our classifier predicts to be ordinary users, it is likely that some subset are actual AKP followers. See Discussion and Conclusions for further discussion on actual AKP followers.

As detailed in Table 2, there is a slight decrease in classifier performance during evaluation on the live dataset, although we note that 0.71 F1-score is in line with state-of-the-art performance on similar tasks (Dhamani et al. 2019; Islam et al. 2020; Luceri, Giordano, and Ferrara 2020b). We suggest this may be due to a shift in tweet topics over time101010Prior work has shown that supervised classifiers trained on Twitter data from one time period may not perform as well on Twitter data from another time period (Alizadeh et al. 2020). Tweets from the takedown dataset that were used to train the classifier were mostly from 2019 with some tweets in the beginning of 2020; tweets from the live dataset were all from 2020, with more tweets towards the end of the year due to the Twitter API’s rate limits. or a subtle adjustment of strategy by the network after the large-scale shutdown in 2020. To explore this further, we perform topic analysis using an implementation of BERTopic (Grootendorst 2022). We use DistilBERTurk embeddings, UMAP for dimensionality reduction, HDBSCAN for clustering, and Grootendorst’s (2022) class-based variant of TF-IDF (c-TD-IDF) to get topic keywords for interpretation. We apply these methods to the takedown users’ tweets used to train, validate, and test the classifier and the live suspended users’ tweets that we evaluate on. Our analyses suggest topical drift between the two datsets. The top 3 topics from the takedown users’ tweets are related to finance, prison, and agriculture, whereas the top 3 topics from the live suspended users’ tweets are related to society/politics, national accounts, and Turkish cities and names.

Accuracy Precision Recall F1-Score
Training (e1e_{1}) 0.734 0.766 0.574 0.656
Validation (e1e_{1}) 0.872 0.885 0.861 0.873
Training (e2e_{2}) 0.931 0.954 0.911 0.932
Validation (e2e_{2}) 0.867 0.817 0.954 0.880
Testing 0.866 0.814 0.968 0.885
Evaluation 0.733 0.645 0.787 0.709
Table 2: Results from training at epochs 1 and 2, validation at epochs 1 and 2, testing on the takedown users and evaluation on our live users that have been suspended by Twitter as of Dec 6 2021.

Experiment. After evaluating our model on the live suspended data, we run an experiment to determine if further fine-tuning aids our performance. In February 2021, we performed the first check on our live users and found that 1,105 of our 7,973 had been suspended. This provided the first timestamp (t1t_{1}) of suspended users from the live dataset. We use this first set of suspended users at t1t_{1} (hereafer, S1S_{1}) to train an additional layer on our classifier. After removing S1S_{1} from the live suspended data (t2t_{2}), we are left with 1,433 users (hereafter, S2S_{2}). We collect and use an additional sample of pseudo-random Turkish language Twitter users as negatively labeled samples to create a balanced dataset for training, validation, testing on S1S_{1} and evaluation on S2S_{2}. We follow the same BERT specifications as listed above. Table 3 shows the results from this experiment. Further fine-tuning on our model using suspended users from the live dataset at t1t_{1} to predict which users will be suspended at t2t_{2} performs very well.

Before further fine-tuning, the classifier achieves higher recall than precision, indicating that it has a higher rate of predicting false positives than of missing true positives. However, after further fine-tuning, the classifier achieves higher precision than recall. Since this task has a significant cost associated with false positives (predicting ordinary Turkish citizens as Turkish state linked accounts), a classifier with high precision is favored. The experiment shows that further fine-tuning with suspended users from S1S_{1} results in a better-performing classifier.

Accuracy Precision Recall F1-Score
Testing 0.834 0.840 0.801 0.820
Evaluation 0.819 0.866 0.745 0.801
Table 3: Results from experiment using our first check of suspended users in Feb 2021 as additional training, validation, and testing data and evaluating on our live users suspended in Dec 2021.

Results presented in this section provide evidence that the suspended users from our live dataset are very likely part of the AKP operation conducted on Twitter. We found that in Jan 2020 accounts in the the takedown dataset interacted substantially with live accounts through retweets and follow trains, which are common ways to boost engagement and visibility and were heavily used by the takedown accounts throughout the operation (Grossman et al. 2020). Our BERT-based classifier achieves an F1-score of 0.71 when evaluated on the live suspended users and 0.80 after further fine tuning on our first set of suspended live users. These results indicate reasonable discernment between Turkish state-linked accounts and ordinary Turkish Twitter users, and provides evidence that AKP-linked IO are ongoing beyond the accounts shared by Twitter in June 2020. Following, we explore strategies of Turkish IO and provide a taxonomy of Turkish state-linked accounts.

RQ2: Strategies for Resilience

We identify strategies employed by Turkish IO actors to gain visibility and retain their social networks in anticipation of and in the aftermath of detection. Network statistics provide evidence that Turkish IO actors use explicit and implicit signals to construct their network to be robust to shutdown.

Account Types and Role Assignment

We first identify distinct account types based on information disclosed within the user profile description. We introduce a taxonomy for the Turkish-linked accounts across two dimensions: account type and account membership.

We define four mutually exclusive account types observed among both the takedown dataset and the live dataset: (1) main account - the account an actor uses most often to post original content and to spread and amplify content from other accounts; (2) retweet account - an account that an actor uses primarily to amplify own and others’ content through retweets; (3) backup account - an account that an actor creates in addition to their main account in case their main account is suspended and uses mostly to post retweets; and (4) sequel account - an account created by an actor in place of a previously suspended account.

Main, retweet, and backup accounts are explicitly disclosed in two ways: (1) an account’s profile description states in some form “this is my main/retweet/backup account”; or (2) a user’s main/retweet/backup account explicitly mentions (@) its backup/retweet/main account username in its profile description. Similarly, some accounts disclose more than one additional account in their profile description. For example, one user discloses their main account and two distinct retweet accounts:

Profile Description (translated): main account - @ ibrahimk_rael, rt 1 account - @suppressed , rt 2 account @suppressed.

Main, retweet, and backup accounts are created and active during the same time period, whereas sequel accounts are created after a main account is suspended (see Figure 3).

Refer to caption
Figure 3: Exemplary structure of main, retweet, backup and sequel accounts.

Table 4 gives the number of retweets and original tweets for main, retweet, and backup accounts among both datasets.111111We note that our present labelling categorizes main accounts only through explicit mentions of such in the account’s profile description and therefore the group we highlight here represents a subset of such accounts. Main accounts post the highest proportion of original content, retweet accounts post the highest proportion of retweets, and backup accounts post mostly retweets across both datasets.

Dataset Account Type Total Tweets # Retweets # Original % retweets
Takedown Main 162,373 110,106 52,267 67.8
Retweet 1,178,639 1,129,094 49,545 95.8
Backup 980,482 877,104 103,378 89.5
All tweets 36,948,536 27,522,156 9,426,380 74.4
Live Main 58,688 27,492 31,196 46.8
Retweet 117,251 111,479 57,72 95.0
Backup 188,744 162,858 25,886 86.2
All tweets 9,298,325 4,890,405 4,407,920 52.6
Table 4: Original and retweet statistics for main, retweet, and backup accounts across both datasets.
Profile Description (translated)
My old account has been suspended, this is my new account.
Bastard feto guys, you close it, I will make a new account
New account. I came back. You won’t be able to silence it.
New account Only national accounts can be followed immediately. (Either you rise with Islam, or you rot with denial …)
new account, my other account is CLOSED #BACKUPACCOUNT @suppressed
Table 5: Sequel account examples.

Table 5 gives examples of newly created sequel accounts disclosing that their previous accounts have been suspended. We hypothesized that the live dataset may contain direct sequel accounts from the takedown dataset. Following, we describe this analysis.

Direct Sequels Between Datasets

We aimed to identify whether a subset of users in the live dataset are sequels for accounts in the takedown dataset. We calculated similarity scores for usernames, profile descriptions, and display names between the two datasets. First, we lowercased all usernames and calculated the Levenshtein distance ratio between each username in both datasets. For takedown users, we match the user from the live dataset with the maximum Levenshtein distance ratio, which we approximate as the user who is most likely a sequel account. Once we have these user pairs, we calculate the similarity between each pair’s profile descriptions and display names using the Python library difflib’s SequenceMatcher.ratio121212This returns any hashable sequence’s similarity as a float between 0 and 1; 1 if the sequences are identical, and 0 if they have nothing in common. method. We also calculate the number of 3rd party users that both accounts in the user pair interacted with under the assumption that sequel accounts will retweet, reply to, quote, and mention some of the same users. We set a maximum Levenshtein distance ratio >> 0.9 for usernames, or a maximum Levenshtein distance ratio >> 0.6 for usernames and at least one of the following: a similarity score >> 0.5 for all non-null profile descriptions in the user pairs, at least 2 common interactions between the user pairs, or display name similarity score >> 0.8. This gave us 169 user pairs between the takedown dataset and the live dataset that we classify as sequel accounts. Table 6 gives examples of sequel pairs.

Takedown Live Username Similarity Bio Similarity Name Similarity # Common Interactions
suppressed suppressed 0.963 0.327 0.231 0
suppressed suppressed 0.933 0.254 0.920 0
suppressed suppressed 0.929 0.680 1.000 0
suppressed suppressed 0.897 0.328 1.000 0
suppressed suppressed 0.897 1.000 0.919 2
ihsantopbas ihsan_topbas42 0.880 0.114 0.812 5
avhasanteke av_hasanteke27 0.880 0.526 1.000 16
hocaketum hocaket 0.875 0.125 0.930 11
suppressed suppressed 0.783 0.108 0.875 1
Table 6: User pairs that we classify as sequel accounts given their username similarity, user profile description similarity, user display name similarity, and the number of common users with which that they interact. Usernames of users with fewer than 5,000 followers are suppressed for privacy.

Account Membership

In contrast to accounts associated with other IO, e.g., those originating in Russia, Saudia Arabia, and China, which are believed to typically to hide their affiliation with the state (Arif, Stewart, and Starbird 2018; King, Pan, and Roberts 2017; DiResta, Shelby Grossman, and Miller 2019), many AKP-linked accounts explicitly mention their allegiance to AKP groups and their allegiance to the state through national accounts and group accounts.

We define national accounts as those that explicitly mention “this is a national account” or use any of the hashtags #MilliTakipMerkezi (#NationalFollowingCenter), #MilliHesaplarYanyana (#NationalAccountsSideBySide), and #MilliHesaplarBurada (#NationalAccountsHere) (see Table 7). National accounts disclose in their tweets and profile descriptions that they are building an alliance of followers online in support of the Turkish government, specifically the AKP. We find that the these accounts encourage follow trains to increase visibility. For example:

Translated tweet from @suppressed (user from the live dataset): FRIENDS, DON’T HAVE A SMALL ACCOUNT!! In order to strengthen national accounts, We support the work of #NationalAccountsTogether, which started with the instruction of our President #NationalAccountsHere RT and Let those who reply follow each other.

National accounts’ hashtags are some of the most tweeted hashtags among our live tweet data. #NationalAccountsSideBySide (tweeted 54,526 times) is the #1 most tweeted hashtag, followed by #NationalAccountsHere (tweeted 27,835 times). While, #NationalFollowingCenter was tweeted 7,428 times. We suggest that the IO is using national accounts to encourage ordinary citizens to engage with their content and begin posting their own pro-government content.

Account Type Live Takedown Profile Description (translated)
Main 34 23 My main account. RT Account: @ibrahimergin98
Retweet 59 92 RT ACCOUNT. Let the storms stop, give way to the Turkish flag.
Backup 112 68 BACKUP ACCOUNT. MAIN ACCOUNT: @suppressed.
Sequel 47 27 New account, old one is suspended!
National accounts 214 6 THE REPUBLIC ALLIANCE IS FOLLOWING ACCOUNTS. #NationalFollowingCenter. #NationalAccountsTogether. #NationalAccountsHere.
Enderun group 10 10 ENDERUN 5 RT & FAV accounts. Please write DM to join our # EnderunGroups. Founding President @suppressed
Table 7: Number of users and examples by type.

We define group accounts as those which explicitly mention their membership in an AK Twitter group. In their report on the Turkish accounts, Stanford Internet Observatory (SIO) refers to these groups as “retweet rings,” because of their highly organized and centralized pro-AKP retweet structure (Grossman et al. 2020). We identified nine distinct groups within our live dataset that follow the same structure as the retweet rings described in SIO’s report. This structure includes nearly identical profile descriptions declaring group membership, and designated group emojis in the user display names and profile descriptions. For example, accounts belonging to ReisiOsmanlıGrupları all use the same yellow star emoji on both sides of the group name, accounts in EnderunGrupları use the red “anger emoji”, and accounts in REİS61 use the Turkish flag. The Enderun group is present in both the takedown dataset and the live dataset. Table 7 gives the number of samples of each account type and membership across takedown and live datasets.

As indicated by the presence and coordination of backup and sequel accounts, AKP-linked IO actors anticipate shutdown and take measures to continue their agenda after account suspension. We hypothesize that the network may use explicit signals strategically to garner easy visibility while depending on more subtle signals to provide a social network backbone even in the face of shutdown. We explore this in the following section.

Network Experiment: Explicit v.s. Implicit Nodes

Subnetwork Nodes Edges Density Diameter Path Length
Full graph 11,551 609,459 0.005 15 3.525
Implicit only (Explicit removed) 10,635 474,833 0.004 16 3.627
Explicit only (Implicit removed) 915 9,877 0.012 8 3.339
Random implicit only 10,636 varies 0.005 15.2 3.536
Random explicit only 915 varies 0.005 11 3.573
Table 8: Graph statistics (density, diameter, avg. path length) for the network in Figure 1 and subnetworks corresponding to: all nodes, implicit nodes, explicit nodes, random samples for comparison, nodes in the takedown dataset only, nodes in the live dataset only, and explicit and implicit nodes in the takedown dataset and the live dataset only. Number of edges for random samples varies with each sample.

In 2020, Twitter removed thousands of Turkish IO accounts, yet, our findings show that the operation has maintained a robust presence on Twitter. After identifying explicit traces of IO strategy, such as dedicated backup and retweet accounts, and direct sequels for accounts taken down by Twitter, we hypothesize that Turkish IO actors utilize explicit signals to support engagement with their followers. While, accounts with no explicit signals are purposefully kept discreet to avoid detection and serve as a structural backbone of the operation. We test these hypotheses by way of graph statistics for the network in Figure 1 and various subnetworks.

We classify explicit nodes as accounts that are: (1) main, (2) retweet, (3) backup, (4) sequel (including direct sequels), (5) group, or (6) national accounts, or if they mention “add to groups” or “do not add to groups” in their profile description. All others are considered implicit nodes. In total, we have 915 explicit and 10,636 implicit nodes.

We calculate network density, diameter, and average path length over five distinct subnetworks from Figure 1. Our hypothesis suggests that the network structure should remain relatively unchanged when removing explicit nodes. While, if we remove all implicit nodes, we expect graph metrics to significantly change. In this case, we are removing a large majority of the network, and in particular, nodes we suggest are integral to the longevity of the network.

Finally, we also remove five random sets of 915 nodes and five random sets of 10,636 nodes as “random explicit” and “random implicit” baselines, respectively. These random samples follow the same class distribution (takedown vs. live) as their comparable subnetworks.

Table 8 details the results of this experiment. Graph density decreases by 20%, diameter increases by 1 (6.67%), and the average path length increases by 2.8% after removing the explicit nodes. After removing the implicit nodes, the graph density increases by 140%, the diameter decreases by 46.7%, and the average path length decreases by 5.3%. However, we do not find significant changes after removing equally-sized random samples of nodes. These results reveal that the subnetwork of only explicit nodes is a denser, more closely connected network. Whereas, the network of only implicit nodes has similar properties to the full graph. This supports our hypothesis that the operation utilizes explicit signals to gain visibility and retain a loyal set of followers but is likewise supported by a robust subnetwork of discrete accounts structured to withstand detection and shutdown.

Generalizability to Other Countries

We conduct parallel analyses on explicit v.s. implicit signals for archived IO originating from Russia and China to examine if our findings from the Turkish network generalize to other countries. To do so, we use Twitter’s Oct 2018 and Jan 2019 releases of Russian/IRA accounts (here forward, R1 and R2, respectively) and Sept 2019 and May 2020 releases of Chinese accounts (here forward, C1 and C2, respectively).131313As with the Turkish dataset, these accounts and their activities can be downloaded on Twitter’s Elections Integrity Hub in the Information Operations archives. https://transparency.twitter.com/en/reports/information-operations.html#1.3 R1 contains 3613 accounts, R2 contains 416 accounts, C1 contains 4301 accounts, and C2 contains 23,750. We note that, in the case of the Turkish IO, there has been just a single release of state-linked IO shared by Twitter so our collection of live accounts serves as a second time point to explore the ongoing operation. In the cases of Russia and China, Twitter has shared multiple releases of IO over time thus supporting a similar inquiry without live account collection.

We first run the direct sequels analysis on both networks of accounts and find 151 direct sequels between C1 and C2, but only 3 direct sequels between R1 and R2. Next, we aim to find additional explicit signals from the Chinese and Russian operations. We manually label all Russian and Chinese IO for presence of explicit signals and sketch primary findings. We find that while the Russian and Chinese accounts do not exhibit the same explicit signals as the Turkish network, they do exhibit their own, and we find interesting overlap between the two groups’ explicit signals.

Russian IO. We observe five primary categories of explicit signals within the Russian network: (1) local news accounts for US cities (e.g., “Breaking news, weather, traffic and more for St. Louis. DM us anytime. RTs not endorsements”); (2) fake personas portraying black Americans and southern conservative Americans or organizations (first identified in (Arif, Stewart, and Starbird 2018), e.g.: “Slave blood. Melanin. Water. Fire. Honey #BlackLivesMatter bih,” “Christian Conservative. #Trump Supporter #Election2016. Watching the polls. Make America Great Again #WakeUpAmerica #tcot”); (3) accounts containing a list of repeated buzzword characteristics with “USA” as the location (e.g., “Lifelong food aficionado. Hardcore problem solver. Twitter lover. Typical bacon fanatic. Creator.”); (4) accounts with inspirational/motivational/love sayings in the profile description and “USA” as the location (e.g., “It’s not about being the best. It’s about being better than you were yesterday”); and, (5) personas of Russian immigrants living in the US (e.g., “Russian Immigrant currently living in the US”). We find 647 explicit accounts within the Russian IO datasets.

Chinese IO. The Chinese network displays four distinct categories of explicit signals, with some overlap with the Russian network: (1) accounts that list followback hashtags in the profile description (e.g., #autofollowback, #teamfollowback); (2) accounts that have lowercased keyboard mashing in the profile description (e.g., “hgfdsa,” “vdcwqd”); (3) accounts that contain love and marriage sayings in the profile description with location set to a US city (e.g., “A great marriage is not when the ‘perfect couple’ comes together. It is when an imperfect couple learns to enjoy their differences.” Location: “Knoxville, USA”); and, (4) accounts containing a list of repeated buzzword characteristics (e.g., “Food aficionado. Alcohol expert. Introvert. Proud tv maven. Award-winning zombieaholic. Wannabe coffee scholar. Music guru. Thinker.”). The categories containing lists of buzzword characteristics in the Russian and Chinese networks have significant overlap and both use words such as aficionado, expert, lover, junkie, trailblazer, maven, guru, and more, following items such as coffee, alcohol, beer, social media, music, and more. We find 862 explicit accounts within the Chinese IO datasets.

Comparative analysis. The explicit signals identified within the Russian and Chinese networks do not show similarity with explicit signals identified in the Turkish network aside from one interesting overlap. We identify two backup accounts for the Russian account “@TEN_GOP.” @TEN_GOP was an IRA-run fake Tennessee GOP account that amassed over 150,000 followers, including Donald Trump, Jr. This account was highly significant in the IRA’s campaign targeting the 2016 US Presidential Election, and has been the subject of multiple news articles (Mueller et al. 2019; Timberg, Dwoskin, and Entous 2017; Prokop 2018; Kessler 2018; Wadhwani 2019).

We create account interaction graphs for both countries and conduct a network analysis identical to the experiment shown in Table 8 for Russia and China’s networks. Network visualizations and experiment results are included in Supplemental Materials.141414https://github.com/mayamerhi/Turkey_IO_ICWSM23 We do not find significant increase in density and decrease in diameter and average path length when examining the explicit only vs. the random explicit only subnetworks as seen in the Turkish network.

From these findings, we conclude that the presence of explicit signals in IO accounts generalizes across the three countries we have examined. Yet, we find that the nature of explicit signals are largely unique to each country’s operation, aside from the overlap in buzzword characteristics used in both the Russian and Chinese networks. We find that the account types and group memberships we present in this work are, to our knowledge, unique to AKP-linked accounts. We also show that our methodology for identifying direct sequel accounts was successful in finding a significant number of direct sequels in the Chinese network, but not in the Russian network.

Explicit signals identified in the Russian and Chinese networks are individual in nature, whereas the group and national accounts from the Turkish network have a group structure. These explicit group structures are, to our knowledge, unique to Turkey. We suggest that the dense explicit only subnetwork (shown in Table 8) is likely due to this group structure.Our findings highlight the lack of a one-size-fits-all solution for studying and mitigating IO.

Discussion and Conclusions

Turkey has been known to operate state-linked Twitter accounts since at least 2013 using a grassroots approach, including hiring young tech-savy Turkish citizens to create and disseminate pro-government content on social media (Albayrak and Parkinson 2013; Saka 2018). We present 7973 Twitter accounts collected from 13 parent accounts that interacted with and resembled AKP-linked accounts from Twitter’s takedown. The accounts we present here are a subset of what is likely an expansive and dynamic network of Turkish IO accounts. Our subset is dependent upon the initial parent accounts; had we chosen different parent accounts, our resulting set of accounts would have been different. Of these accounts, 2454 have been suspended since collection. Due to their grassroots approach, it is incredibly difficult to distinguish Turkish IO actors from ordinary AKP supporters. We have not studied whether there are ordinary AKP supporters in our live dataset, and we do not claim that every account in our dataset is linked to the operation. However, given Twitter’s subsequent suspension of 2454 of our live accounts, the reasons Twitter states for suspending users, and the results of our classifier, we believe that this subset of our live dataset is likely linked to Turkey’s ongoing IO.

We utilize machine learning techniques to show latent similarities between content from Twitter’s takedown dataset and our live dataset. We identify strategies that Turkish IO actors appear to utilize to remain resilient to shutdown. These include explicit signals to bolster alliance and visibility, such as national account hashtags and groups, which we suspect are also used to engage ordinary citizens and appear as grassroots government support. After accounts are shutdown, IO actors preserve their social networks by standing up highly similar replacement accounts which are easy for their followers to find, even if it makes them more vulnerable to detection again. Our findings suggest that the integrity of the operation relies on a set of discreet accounts to keep the network intact as explicit accounts are detected and shut down. We believe that there is likely also more subtle strategy supplementing operations manifest as implicit relations amongst accounts that may be tied to tweet text and media. Next steps in understanding the strategies employed by the Turkish IO campaign should explore this content and potential shifts in related tactics, e.g., use of memes, cross-platform engagement.

To examine whether our findings in the context of Turkish operations generalize to our countries, we apply similar analyses to two sets of Russian and Chinese IO accounts archived by Twitter. Our methodology for identifying direct sequels generalizes well to China, but not to Russia. Both Russia and China exhibit a robust set of explicit signals, however, they are unique to each country for the most part. Results of comparative network experiments suggest that Turkey uniquely constructs a dense subnetwork of explicit accounts. We attribute this to Turkey’s unique group structure discussed in the Account Membership analysis.

We have suggested that Turkish IO actors began anticipating and preparing for shutdown by opening backup and retweet accounts, and after shutdown, sequel accounts. This is a novel finding that we did not anticipate at the time of data collection when we decided to limit account collection to those created in 2020, which led to a very small window of activity overlap between the two datasets studied here. This is a limitation of data collection and we suggest future work studying IO dynamics include accounts created prior to large scale shutdown.

Our findings highlight the time, effort and cost imbalances between IO actors and the institutions working to combat them. For IO actors, opening a large number of social media accounts is relatively cheap and easy. For Twitter, significant resources are expended to detect them and account shutdown bears risks. Broadly, it seems that Twitter’s efforts to mitigate coordinated manipulation on their platform is a step in the right direction for creating an open, democratic information environment. Their data sharing provides researchers with exceptional volumes of ground-truth IO data and they set valuable industry standards for transparency. However, our findings also suggest that current approaches fall short and highlight the need for fundamental shifts in platform norms and policies.

Ethics

This work provides novel and valuable insights to the research community studying IO and coordinated influence online. We provide our dataset of a suspected live Turkish IO on Twitter, along with our random sample of Turkish-language Twitter users used to train our classifier. Both datasets are available on the first author’s Github page. 151515https://github.com/mayamerhi/Turkey_IO_ICWSM23 Following Twitter’s Terms of Service, we release only the User ID’s and Tweet ID’s for all data presented here  161616https://developer.twitter.com/en/developer-terms/policy. Also following Twitter’s protocols, we suppress the usernames of all users with fewer than 5000 followers.

In this work, we also obtained and analyzed Twitter’s unhashed versions of their archived IO data. Academic researchers must apply for access to the unhashed versions of the data, provide details on the proposed usage, analysis, storage of the data, and sharing of results obtained from analyzing the data, and sign a data agreement. The first author obtained access to the unhashed data, and followed all guidelines laid out by Twitter.

We expect that these datasets may be utilized to build models to detect future Turkish IO as their strategies continue to evolve. We acknowledge that some users in our live collected dataset may not be affiliated with the Turkish state and may not be attempting to influence online users through coordinated operations. We present the live dataset as suspected IO based on our observations and analyses, but we cannot assess the origins and intentions of these users with certainty given the information we have. This is true for Twitter’s archive data as well, which Twitter acknowledges as do we, and which motives the suppression of usernames for users with fewer than 5000 followers.

Both datasets were collected by the first author through their personal Twitter account using the Academic Track Twitter API and stored on a password protected external hard drive. The data were also uploaded to a cloud server for analysis, which is protected with two-factor authentication. After analysis, data were removed from the cloud server and stored only on the password protected external hard drive.

Acknowledgement

This work was in part supported by NSF awards #1820609 and #2131144.

References

  • Addawood et al. (2019) Addawood, A.; Badawy, A.; Lerman, K.; and Ferrara, E. 2019. Linguistic cues to deception: Identifying political trolls on social media. Proceedings of the International AAAI Conference on Web and Social Media 13: 15–25. ISSN 2334-0770. URL https://aaai.org/ojs/index.php/ICWSM/article/view/3205.
  • Albayrak and Parkinson (2013) Albayrak, A.; and Parkinson, J. 2013. Turkey’s Government Forms 6,000-Member Social Media Team - WSJ. URL https://www.wsj.com/articles/turkeys-government-forms-6000member-social-media-team-1379351399.
  • Alizadeh et al. (2020) Alizadeh, M.; Shapiro, J. N.; Buntain, C.; and Tucker, J. A. 2020. Content-based features predict social media influence operations. Science Advances 6(30): eabb5824. ISSN 2375-2548. doi:10.1126/sciadv.abb5824. URL https://advances.sciencemag.org/content/6/30/eabb5824. Publisher: American Association for the Advancement of Science Section: Research Article.
  • Arif, Stewart, and Starbird (2018) Arif, A.; Stewart, L. G.; and Starbird, K. 2018. Acting the part: Examining information operations within #BlackLivesMatter discourse. Proceedings of the ACM on Human-Computer Interaction 2(CSCW): 20:1–20:27. doi:10.1145/3274289. URL https://doi.org/10.1145/3274289.
  • Başaran (2020) Başaran, F. 2020. Digital Authoritarianism and Trolling in Turkey. URL https://publicseminar.org/essays/digital-authoritarianism-and-trolling-in-turkey/.
  • Bradshaw and Howard (2017) Bradshaw, S.; and Howard, P. 2017. Troops, Trolls and Troublemakers: A Global Inventory of Organized Social Media Manipulation 2017.12. URL https://ora.ox.ac.uk/objects/uuid:cef7e8d9-27bf-4ea5-9fd6-855209b3e1f6. Publisher: Oxford Internet Institute.
  • Bulut and Yörük (2017) Bulut, E.; and Yörük, E. 2017. Mediatized Populisms| Digital Populism: Trolls and Political Polarization of Twitter in Turkey. International Journal of Communication 11(0): 25. ISSN 1932-8036. URL https://ijoc.org/index.php/ijoc/article/view/6702. Number: 0.
  • Dhamani et al. (2019) Dhamani, N.; Azunre, P.; Gleason, J. L.; Corcoran, C.; Honke, G.; Kramer, S.; and Morgan, J. 2019. Using Deep Networks and Transfer Learning to Address Disinformation.
  • DiResta, Shelby Grossman, and Miller (2019) DiResta, R.; Shelby Grossman, K.; and Miller, C. 2019. Analysis of Twitter takedown of state-backed operation attributed to saudi arabian digital marketing firm smaat.
  • Farkas and Bastos (2018) Farkas, J.; and Bastos, M. 2018. IRA propaganda on Twitter: Stoking antagonism and tweeting local news. In Proceedings of the 9th International Conference on Social Media and Society, SMSociety ’18, 281–285. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-6334-1. doi:10.1145/3217804.3217929. URL https://doi.org/10.1145/3217804.3217929.
  • Ferrara (2017) Ferrara, E. 2017. Disinformation and social bot operations in the run up to the 2017 French presidential election. First Monday 22(8). doi:10.5210/fm.v22i8.8005. URL https://firstmonday.org/ojs/index.php/fm/article/view/8005.
  • Grootendorst (2022) Grootendorst, M. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv preprint arXiv:2203.05794 .
  • Grossman et al. (2020) Grossman, S.; Akis, F. A.; Alemdaroğlu, A.; Goldstein, J. A.; Jonsson, K.; Garcia-Camargo, I.; Thiel, D.; and Zaheer, A. 2020. Political retweet rings and compromised accounts: A Twitter influence operation linked to the youth wing of Turkey’s ruling party. URL https://cyber. fsi. stanford. edu/io/news/june-2020-twitter-takedown .
  • Im et al. (2020) Im, J.; Chandrasekharan, E.; Sargent, J.; Lighthammer, P.; Denby, T.; Bhargava, A.; Hemphill, L.; Jurgens, D.; and Gilbert, E. 2020. Still out there: Modeling and Identifying Russian Troll Accounts on Twitter. In 12th ACM Conference on Web Science, WebSci ’20, 1–10. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-7989-2. doi:10.1145/3394231.3397889. URL https://doi.org/10.1145/3394231.3397889.
  • Islam et al. (2020) Islam, M. R.; Liu, S.; Wang, X.; and Xu, G. 2020. Deep learning for misinformation detection on online social networks: a survey and new perspectives. Social Network Analysis and Mining 10(1): 1–20.
  • Karatas and Saka (2017) Karatas, D.; and Saka, E. 2017. Online political trolling in the context of post-Gezi social media in Turkey. International Journal of Digital Television 8: 383–401. doi:10.1386/jdtv.8.3.383˙1.
  • Kessler (2018) Kessler, A. 2018. Who is @TEN_GOP from the Russia indictment? Here’s what we found reading 2,000 of its tweets. CNN URL https://www.cnn.com/2018/02/16/politics/who-is-ten-gop/index.html.
  • King, Pan, and Roberts (2017) King, G.; Pan, J.; and Roberts, M. E. 2017. How the Chinese government fabricates social media posts for strategic distraction, not engaged argument. American political science review 111(3): 484–501.
  • Lin and Kerr (2019) Lin, H.; and Kerr, J. 2019. On cyber-enabled information warfare and information operations. forthcoming, Oxford Handbook of Cybersecurity .
  • Linvill and Warren (2020) Linvill, D. L.; and Warren, P. L. 2020. Troll factories: Manufacturing specialized disinformation on Twitter. Political Communication 37(4): 447–467. ISSN 1058-4609. doi:10.1080/10584609.2020.1718257. URL https://doi.org/10.1080/10584609.2020.1718257. Publisher: Routledge _eprint: https://doi.org/10.1080/10584609.2020.1718257.
  • Luceri, Giordano, and Ferrara (2020a) Luceri, L.; Giordano, S.; and Ferrara, E. 2020a. Detecting troll behavior via inverse reinforcement learning: A case study of Russian trolls in the 2016 US Election. Proceedings of the International AAAI Conference on Web and Social Media 14: 417–427. ISSN 2334-0770. URL https://aaai.org/ojs/index.php/ICWSM/article/view/7311.
  • Luceri, Giordano, and Ferrara (2020b) Luceri, L.; Giordano, S.; and Ferrara, E. 2020b. Don’t Feed the Troll: Detecting Troll Behavior via Inverse Reinforcement Learning.
  • Mueller et al. (2019) Mueller, R. S.; Helderman, R. S.; Zapotosky, M.; United States; Department of Justice; and Special Counsel’s Office. 2019. The Mueller report. ISBN 978-1-982129-74-3. URL https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db“linebreak=nlabk&AN=2036126. OCLC: 1097961575.
  • Prokop (2018) Prokop, A. 2018. 23 tweets from @TEN_GOP, one Russian-run Twitter account mentioned in Mueller’s new indictment. Vox URL https://www.vox.com/policy-and-politics/2017/10/19/16504510/ten-gop-twitter-russia.
  • Recuero, Soares, and Gruzd (2020) Recuero, R.; Soares, F. B.; and Gruzd, A. 2020. Hyperpartisanship, disinformation and political conversations on Twitter: The Brazilian Presidential Election of 2018. Proceedings of the International AAAI Conference on Web and Social Media 14: 569–578. ISSN 2334-0770. URL https://aaai.org/ojs/index.php/ICWSM/article/view/7324.
  • Saka (2018) Saka, E. 2018. Social Media in Turkey as a Space for Political Battles: AKTrolls and other Politically motivated trolling. Middle East Critique 27(2): 161–177. ISSN 1943-6149. doi:10.1080/19436149.2018.1439271. URL https://doi.org/10.1080/19436149.2018.1439271. Publisher: Routledge _eprint: https://doi.org/10.1080/19436149.2018.1439271.
  • Schweter (2020) Schweter, S. 2020. BERTurk - BERT models for Turkish. doi:10.5281/zenodo.3770924. URL https://doi.org/10.5281/zenodo.3770924.
  • Starbird, Arif, and Wilson (2019) Starbird, K.; Arif, A.; and Wilson, T. 2019. Disinformation as collaborative work: Surfacing the participatory nature of strategic information operations. Proceedings of the ACM on Human-Computer Interaction 3(CSCW): 127:1–127:26. doi:10.1145/3359229. URL https://doi.org/10.1145/3359229.
  • Timberg, Dwoskin, and Entous (2017) Timberg, C.; Dwoskin, E.; and Entous, A. 2017. Russian Twitter account pretending to be Tennessee GOP fools celebrities, politicians. Chicago Tribune URL https://www.chicagotribune.com/business/blue-sky/ct-russian-twitter-account-tennessee-gop-20171018-story.html.
  • Twitter (2018) Twitter. 2018. Update on Twitter’s review of the 2016 US election. URL https://blog.twitter.com/en˙us/topics/company/2018/2016-election-update.html.
  • Twitter Safety (2020) Twitter Safety. 2020. Disclosing networks of state-linked information operations we’ve removed. URL https://blog.twitter.com/en˙us/topics/company/2020/information-operations-june-2020.html.
  • Wadhwani (2019) Wadhwani, A. 2019. Mueller report outlines activities of fake Tennessee GOP Twitter account linked to Russians. Tennessean URL https://www.tennessean.com/story/news/2019/04/18/mueller-report-release-shows-fake-tennessee-gop-twitter-account-link-russia/3507153002/.
  • Yılmaz and Turner (2019) Yılmaz, Z.; and Turner, B. S. 2019. Turkey’s deepening authoritarianism and the fall of electoral democracy. British Journal of Middle Eastern Studies 46(5): 691–698. ISSN 1353-0194. doi:10.1080/13530194.2019.1642662. URL https://doi.org/10.1080/13530194.2019.1642662. Publisher: Routledge _eprint: https://doi.org/10.1080/13530194.2019.1642662.
  • Zannettou et al. (2020) Zannettou, S.; Caulfield, T.; Bradlyn, B.; Cristofaro, E. D.; Stringhini, G.; and Blackburn, J. 2020. Characterizing the use of images in state-sponsored information warfare operations by Russian trolls on Twitter. Proceedings of the International AAAI Conference on Web and Social Media 14: 774–785. ISSN 2334-0770. URL https://www.aaai.org/ojs/index.php/ICWSM/article/view/7342.
  • Zannettou et al. (2019a) Zannettou, S.; Caulfield, T.; De Cristofaro, E.; Sirivianos, M.; Stringhini, G.; and Blackburn, J. 2019a. Disinformation Warfare: Understanding State-Sponsored Trolls on Twitter and Their Influence on the Web. In Companion Proceedings of The 2019 World Wide Web Conference, WWW ’19, 218–226. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-6675-5. doi:10.1145/3308560.3316495. URL https://doi.org/10.1145/3308560.3316495.
  • Zannettou et al. (2019b) Zannettou, S.; Caulfield, T.; Setzer, W.; Sirivianos, M.; Stringhini, G.; and Blackburn, J. 2019b. Who Let The Trolls Out? Towards Understanding State-Sponsored Trolls. In Proceedings of the 10th ACM Conference on Web Science, WebSci ’19, 353–362. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-6202-3. doi:10.1145/3292522.3326016. URL https://doi.org/10.1145/3292522.3326016.