arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

15
TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective Bilal Ghanem 1 , Davide Buscaldi 2 , and Paolo Rosso 1 1 PRHLT Research Center, Universitat Polit` ecnica de Val` encia, Spain {bigha@doctor, prosso@dsic}.upv.es 2 CNRS / Laboratoire LIPN, Universit´ e Paris 13, France [email protected] Abstract. The online new emerging suspicious users, that usually are called trolls, are one of the main sources of hate, fake, and deceptive on- line messages. Some agendas are utilizing these harmful users to spread incitement tweets, and as a consequence, the audience get deceived. The challenge in detecting such accounts is that they conceal their identi- ties which make them disguised in social media, adding more difficulty to identify them using just their social network information. Therefore, in this paper, we propose a text-based approach to detect the online trolls such as those that were discovered during the US 2016 presidential elections. Our approach is mainly based on textual features which uti- lize thematic information, and profiling features to identify the accounts from their way of writing tweets. We deduced the thematic information in a unsupervised way and we show that coupling them with the textual features enhanced the performance of the proposed model. In addition, we find that the proposed profiling features perform the best comparing to the textual features. Keywords: Russian Trolls · Twitter · Topic Modeling · Profiling. 1 Introduction Recent years have seen a large increase in the amount of disinformation and fake news spread on social media. False information was used to spread fear and anger among people, which in turn, provoked crimes in some countries. The US in the recent years experienced many similar cases during the presidential elections, such as the one commonly known as “Pizzagate” 3 . Later on, Twitter declared that they had detected a suspicious campaign originated in Russia by an organization named Internet Research Agency (IRA), and targeted the US to affect the results of the 2016 presidential elections 4 . The desired goals behind these accounts are to spread fake and hateful news to further polarize the public opinion. Such attempts are not limited to Twitter, since Facebook announced in 3 https://www.rollingstone.com/politics/politics-news/anatomy-of-a-fake-news-scandal-125877/ 4 https://blog.twitter.com/official/en us/topics/company/2018/ 2016-election-update.html arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

Transcript of arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

Page 1: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitterfrom a Textual Perspective

Bilal Ghanem1, Davide Buscaldi2, and Paolo Rosso1

1 PRHLT Research Center, Universitat Politecnica de Valencia, Spain{bigha@doctor, prosso@dsic}.upv.es

2 CNRS / Laboratoire LIPN, Universite Paris 13, [email protected]

Abstract. The online new emerging suspicious users, that usually arecalled trolls, are one of the main sources of hate, fake, and deceptive on-line messages. Some agendas are utilizing these harmful users to spreadincitement tweets, and as a consequence, the audience get deceived. Thechallenge in detecting such accounts is that they conceal their identi-ties which make them disguised in social media, adding more difficultyto identify them using just their social network information. Therefore,in this paper, we propose a text-based approach to detect the onlinetrolls such as those that were discovered during the US 2016 presidentialelections. Our approach is mainly based on textual features which uti-lize thematic information, and profiling features to identify the accountsfrom their way of writing tweets. We deduced the thematic informationin a unsupervised way and we show that coupling them with the textualfeatures enhanced the performance of the proposed model. In addition,we find that the proposed profiling features perform the best comparingto the textual features.

Keywords: Russian Trolls · Twitter · Topic Modeling · Profiling.

1 Introduction

Recent years have seen a large increase in the amount of disinformation andfake news spread on social media. False information was used to spread fearand anger among people, which in turn, provoked crimes in some countries. TheUS in the recent years experienced many similar cases during the presidentialelections, such as the one commonly known as “Pizzagate” 3. Later on, Twitterdeclared that they had detected a suspicious campaign originated in Russia byan organization named Internet Research Agency (IRA), and targeted the USto affect the results of the 2016 presidential elections4. The desired goals behindthese accounts are to spread fake and hateful news to further polarize the publicopinion. Such attempts are not limited to Twitter, since Facebook announced in

3 https://www.rollingstone.com/politics/politics-news/anatomy-of-a-fake-news-scandal-125877/4 https://blog.twitter.com/official/en us/topics/company/2018/

2016-election-update.html

arX

iv:1

910.

0134

0v1

[cs

.CL

] 3

Oct

201

9

Page 2: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

2 Anonymous et al.

mid-2019 that they detected a similar attempt originating from UAE, Egypt andSaudi Arabia and targeting other countries such as Qatar, Palestine, Lebanonand Jordan5. This attempt used Facebook pages, groups, and user accountswith fake identities to spread fake news supporting their ideological agendas.The automatic detection of such attempts is very challenging, since the trueidentity of these suspicious accounts is hidden by imitating the profiles of realpersons from the targeted audience; in addition, sometimes they publish theirsuspicious idea in a vague way through their tweets’ messages.

A previous work [19] showed that such suspicious accounts are not bots in astrict sense and they argue that they could be considered as “software-assistedhuman workers”. According to [9], the online suspicious accounts can be catego-rized into 3 main types: Robots, Cyborgs, and Human Spammers. We considerIRA accounts as another new emerging type called trolls, which is similar toCyborgs except that the former focuses on targeting communities instead ofindividuals6.

In this work, we identify online trolls in Twitter, namely IRA trolls, froma textual perspective. We study the effect of a set of text-based features andwe propose a machine learning model to detect them7. We aim to answer threeresearch questions: RQ1. Does the thematic information improve the detectionperformance?, RQ2. Can we detect IRA trolls from only a textual perspective?and RQ3. How IRA campaign utilized the emotions to affect the public opinions?

The rest of the paper is structured as follows. In the following section, wepresent an overview on the literature work on IRA trolls. In Section 3, we describehow the used dataset was compiled. Section 4 describes our proposed features forour approach. The experiments, results, and analyses are presented in Section5. Finally, we draw some conclusions and discuss possible future work on IRAtrolls.

2 Related Work on IRA Trolls

After the 2016 US elections, Twitter has detected a suspicious attempt by alarge set of accounts to influence the results of the elections. Due to this event,an emerging research works about the Russian troll accounts started to appear[7,32,19,13,2].

The research works studied IRA trolls from several perspectives. The work in[13] studied the links’ domains that were mentioned by IRA trolls and how muchthey overlap with other links used in tweets related to ”Brexit”. In addition, theycompare ”Left” and ”Right” ideological trolls in terms of the number of re-tweetsthey received, number of followers, etc, and the online propaganda strategies theyused. The authors in [7] analyzed IRA campaign in both Twitter and Facebook,and they focus on the evolution of IRA paid advertisements on Facebook beforeand after the US presidential elections from a thematic perspective.

5 https://newsroom.fb.com/news/2019/08/cib-uae-egypt-saudi-arabia/6 https://itstillworks.com/difference-between-troll-cyberbully-5054.html7 Data and model are available at: https://www.github.com/anonymous

Page 3: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 3

The analysis work on IRA trolls not limited only to the tweets content, butit also considered the profile description, screen name, application client, geo-location, timezone, and number of links used per each media domain [32]. Thereis a probability that Twitter has missed some IRA accounts that maybe wereless active than the others. Based on this hypothesis, the work in [19] built amachine learning model based on profile, language distribution, and stop-wordsusage features to detect IRA trolls in a newly sampled data from Twitter. Otherworks tried to model IRA campaign not only by focusing on the trolls accounts,but also by examining who interacted with the trolls by sharing their contents [1].Similarly, the work [2] proposed a model that made use of the political ideologiesof users, bot likelihood, and activity-related account metadata to predict userswho spread the trolls’ contents.

3 Data

To model the identification process of the Russian trolls, we considered a largedataset of both regular users (legitimate accounts) and IRA troll accounts. Fol-lowing we describe the dataset. In Table 1 we summarizes its statistics.

Table 1: Description of the dataset.Russian Trolls Regular Accounts Total

Total # of Accounts 2,023 94,643 96,666

Total # of Tweets 1,840,679 1,945,505 3,786,184

3.1 Russian Trolls (IRA)

We used the IRA dataset8 that was released by Twitter after identifying theRussian trolls. The original dataset contains 3, 841 accounts, but we use a lowernumber of accounts and tweets after filtering them. We focus on accounts thatuse English as main language. In fact, our goal is to detect Russian accountsthat mimic a regular US user. Then, we remove from these accounts non-Englishtweets, and maintain only tweets that were tweeted originally by them. Our finalIRA accounts list contains 2,023 accounts.

3.2 Regular Accounts

To contrast IRA behaviour, we sampled a large set of accounts to represent theordinary behaviour of accounts from US. We collected a random sample of usersthat they post at least 5 tweets between 1st of August and 31 of December,2016 (focusing on the US 2016 debates: first, second, third and vice president

8 https://about.twitter.com/en us/values/elections-integrity.html

Page 4: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

4 Anonymous et al.

debates and the election day) by querying Twitter API hashtags related to theelections and its parties (e.g #trump, #clinton, #election, #debate, #vote,etc.). In addition, we selected the accounts that have location within US anduse English as language of the Twitter interface. We focus on users during thepresidential debates and elections dates because we suppose that the peak oftrolls efforts concentrated during this period.

The final dataset is totally imbalanced (2% for IRA trolls and 98% for theregular users). This class imbalance situation represent a real scenario. FromTable 1, we can notice that the number of total tweets of the IRA trolls issimilar to the one obtained from the regular users. This is due to the fact thatIRA trolls were posting a lot of tweets before and during the elections in anattempt to try to make their messages reach the largest possible audience.

4 Textual Representation

In order to identify IRA trolls, we use a rich set of textual features. With this setof features we aim to model the tweets of the accounts from several perspectives.

4.1 Thematic Information

Previous works [26] have investigated IRA campaign efforts on Facebook, andthey found that IRA pages have posted more than ∼80K posts focused on di-vision issues in US. Later on, the work in [7] has analyzed Facebook advertisedposts by IRA and they specified the main themes that these advertisementsdiscussed. Given the results of the previous works, we applied a topic modelingtechnique on our dataset to extract its main themes. We aim to detect IRA trollsby identifying their suspicious ideological changes across a set of themes.

Given our dataset, we applied Latent Dirichlet Allocation (LDA) topic mod-eling algorithm [5] on the tweets after a prepossessing step where we maintainedonly nouns and proper nouns. In addition, we removed special characters (exceptHASH ”#” sign for the hashtags) and lowercase the final tweet. To ensure thequality of the themes, we removed the hashtags we used in the collecting pro-cess where they may bias the modeling algorithm. We tested multiple numberof themes and we chose seven of them. We manually observed the content ofthese themes to label them. The extracted themes are: Police shootings, Islamand War, Supporting Trump, Black People, Civil Rights, Attacking Hillary, andCrimes. In some themes, like Supporting Trump and Attacking Hillary, we foundcontradicted opinions, in favor and against the main themes, but we chose thefinal stance based on the most representative hashtags and words in each of them(see Figure 1). Also, the themes Police Shooting and Crimes are similar, but wefound that some words such as: police, officers, cops, shooting, gun, shot, etc. arethe most discriminative between these two themes. In addition, we found thatthe Crimes theme focuses more on raping crimes against children and women.Our resulted themes are generally consistent with the ones obtained from the

Page 5: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 5

Facebook advertised posts in [7], and this emphasizes that IRA efforts organizedin a similar manner in both social media platforms.

Based on our thematic information, we model the users textual features w.r.t.each of these themes. In other words, we model a set of textual features inde-pendently for each of the former themes to capture the emotional, stance, andothers changes in the users tweets.

(a)(b)

Fig. 1: (a) Supporting Trump and (b) Attacking Hillary themes words clouds.

For the theme-based features, we use the following features that we believethat they change based on the themes:

– Emotions: Since the results of the previous works [7,26] showed that IRAefforts engineered to seed discord among individuals in US, we use emotionsfeatures to detect their emotional attempts to manipulate the public opinions(e.g. fear spreading behavior). For that, we use the NRC emotions lexicon[25] that contains ∼14K words labeled using the eight Plutchik’s emotions.

– Sentiment: We extract the sentiment of the tweets from NRC [25], positiveand negative.

– Bad & Sexual Cues: During the manual analysis of a sample from IRAtweets, we found that some users use bad slang word to mimic the languageof a US citizen. Thus, we model the presence of such words using a list ofbad and sexual words from [11].

– Stance Cues: Stance detection has been studied in different contexts todetect the stance of a tweet reply with respect to a main tweet/thread [24].Using this feature, we aim to detect the stance of the users regarding thedifferent topics we extracted. To model the stance we use a set of stancelexicons employed in previous works [3,12]. Concretely, we focus on the fol-

Page 6: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

6 Anonymous et al.

lowing categories: belief, denial, doubt, fake, knowledge, negation, question,and report.

– Bias Cues: We rely on a set of lexicons to capture the bias in text. We modelthe presence of the words in one of the following cues categories: assertivesverbs [17], bias [28], factive verbs [21], implicative verbs [20], hedges [18],report verbs [28]. A previous work has used these bias cues to identify biasin suspicious news posts in Twitter [31].

– LIWC: We use a set of linguistic categories from the LIWC linguistic dic-tionary [30]. The used categories are: pronoun, anx, cogmech, insight, cause,discrep, tentat, certain, inhib, incl9.

– Morality: Cues based on the morality foundation theory [15] where wordslabeled in one of a set of categories: care, harm, fairness, cheating, loyalty,betrayal, authority, subversion, sanctity, and degradation.

Given Vi as the concatenation of the previous features vectors of a tweeti, werepresent each user’s tweets by considering the average and standard deviationof her tweets’ V1,2,..N in each theme j independently and we concatenate them.Mathematically, a user x final feature vector is defined as follows:

userx =

T⊙j=1

∑Nj

i=1 Vij

Nj�

√∑Nj

i=1(Vij − Vj)

Nj

(1)

where given the jth theme, Nj is the total number of tweets of the user, Vij isthe ith tweet feature vector, Vj is the mean of the tweets’ feature vectors. Withthis representation we aim at capturing the ”Flip-Flop” behavior of IRA trollsamong the themes (see Section 5.4).

4.2 Profiling IRA Accounts

As Twitter declared, although the IRA campaign was originated in Russia, ithas been found that IRA trolls concealed their identity by tweeting in English.Furthermore, for any possibility of unmasking their identity, the majority of IRAtrolls changed their location to other countries and the language of the Twitterinterface they use. Thus, we propose the following features to identify these usersusing only their tweets text:

– Native Language Identification (NLI): This feature was inspired byearlier works on identifying native language of essays writers [22]. We aim todetect IRA trolls by identifying their way of writing English tweets. As shownin [31], English tweets generated by non-English speakers have a differentsyntactic pattern . Thus, we use state-of-the-art NLI features to detect thisunique pattern [8,23,14]; the feature set consists of bag of stopwords, Part-of-speech tags (POS), and syntactic dependency relations (DEPREL). We

9 Total pronouns, Anxiety, Cognitive processes, Insight, Causation, Discrepancy, Ten-tative, Certainty, Inhibition, and Inclusive respectively.

Page 7: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 7

extract the POS and the DEPREL information using spaCy10, an off-the-shelf POS tagger11. We clean the tweets from the special characters andmaintained dots, commas, and first-letter capitalization of words. We useregular expressions to convert a sequence of dots to a single dot, and similarlyfor sequence of characters.

– Stylistic: We extract a set of stylistic features following previous worksin the authorship attribution domain [33,4,29], such as: the count of spe-cial characters, consecutive characters and letters12, URLs, hashtags, users’mentions. In addition, we extract the uppercase ratio and the tweet length.

Similar to the feature representation of the theme-based features, we representeach user’s tweets by considering the average and standard deviation of hertweets’ V1,2,..N , given Vi as the concatenation of the previous two features vectorsof a tweeti. A user x final feature vector is defined as follows:

userx =

∑Ni=1 Vi

N;

√∑Ni=1(Vi − V )

N

(2)

where N is her total number of tweets, Vi is the ith tweet feature vector, Vis the mean of her tweets feature vectors.

5 Experiments and Analysis

5.1 Experimental Setup

We report precision, recall and F1 score. Given the substantial class imbalancein the dataset, we use the macro weighted version of the F1 metric. We testedseveral classifiers and Logistic Regression showed the best F1macro value. Wekept the default parameters values. We report results for 5-folds cross-validation.

5.2 Baselines

In order to evaluate our feature set, we use Random Selection, Majority Class,and bag-of-words baselines. In the bag-of-words baseline, we aggregate all thetweets of a user into one document. A previous work [6] showed that IRA trollswere playing a hashtag game which is a popular word game played on Twitter,where users add a hashtag to their tweets and then answer an implied question[16]. IRA trolls used this game in a similar way but focusing more on offending orattacking others; an example from IRA tweets: ”#OffendEveryoneIn4Words un-documented immigrants are ILLEGALS”. Thus, we use as a baseline Tweet2vec[10] which is a a character-based Bidirectional Gated Recurrent neural network

10 https://spacy.io/models11 We tested TweetNLP [27], the state-of-the-art POS tagger on tweets, but spaCy

produced higher overall results.12 We considered 2 or more consecutive characters, and 3 or more consecutive letters.

Page 8: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

8 Anonymous et al.

reads tweets and predicts their hashtags. We aim to assess if the tweets hashtagscan help identifying the IRA tweets. The model reads the tweets in a form ofcharacter one-hot encodings and uses them for training with their hashtags aslabels. To train the model, we use our collected dataset which consists of ∼3.7Mtweets13. To represent the tweets in this baseline, we use the decoded embeddingproduced by the model and we feed them to the Logistic Regression classifier.

IRA dataset provided by Twitter contains less information about the ac-counts details, and they limited to: profile description, account creation date,number of followers and followees, location, and account language. Therefore,as another baseline we use the number of followers and followees to assess theiridentification ability (we will mention them as Network Features in the rest ofthe paper).

5.3 Results

Table 2 presents the classification results showing the performance of each featureset independently. Generally, we can see that the thematic information improvesthe performance of the proposed features clearly (RQ1), and with the largestamount in the Emotions features (see −themes and +themes columns). This resultemphasizes the importance of the thematic information. Also, we see that theemotions performance increases with the largest amount considering F1macro

value; this motivates us to analyze the emotions in IRA tweets (see the followingsection).

The result of the NLI feature in the table is interesting; we are able to detectIRA trolls from their writing style with a F1macro value of 0.91. Consideringthe results in Table 2, we can notice that we are able to detect the IRA trollseffectively using only textual features (RQ2).

Finally, the baselines results show us that the Network features do not per-form well. A previous work [32] showed that IRA trolls tend to follow a lot ofusers, and nudging other users to follow them (e.g. by writing ”follow me” intheir profile description) to fuse their identity (account information) with theregular users. Finally, similar to the Network features, the Tweet2vec baselineperforms poorly. This indicates that, although IRA trolls used the hashtag gameextensively in their tweets, the Tweet2vec baseline is not able to identify them.

5.4 Analysis

Given that the Emotions features boosted the F1macro with the highest valuecomparing to the other theme-based features, in Figure 2 we analyze IRA trollsfrom emotional perspective to answer RQ3. The analysis shows that the themesthat were used to attack immigrants (Black People and Islam and War) havethe fear emotion in their top two emotions. While on the other hand, a themelike Supporting Trump has a less amount of fear emotion, and the joy emotionamong the top emotions.

13 We used the default parameters that were provided with the system code.

Page 9: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 9

Table 2: Classification results. We report the results of each feature set indepen-dently.

Method Precisionmacro Recallmacro F1macro

Random Selection 0.5 0.5 0.35

Majority Class 0.49 0.5 0.5

Tweet2vec 0.49 0.5 0.5

Network Features 0.49 0.5 0.5

Theme-based Features -themes +themes

Emotions 0.59 0.5 0.5 0.87

Sentiment 0.49 0.5 0.5 0.64

Bad & Sexual Cues 0.49 0.5 0.5 0.79

Stance cues 0.91 0.53 0.56 0.86

Bias Cues 0.73 0.51 0.51 0.87

LIWC 0.69 0.51 0.51 0.85

Morality 0.90 0.62 0.68 0.86

Profiling Features

Stylistic 0.92 0.84 0.88

NLI 0.96 0.88 0.91

All Features 0.96 0.93 0.94

Fig. 2: Emotional analyzing of IRA trolls from thematic perspective.

Why do the thematic information help? The Flip-Flop behavior. As an example,let’s considering the fear and joy emotions in Figure 2. We can notice thatall the themes that used to nudge the division issues have a decreasing dashed

Page 10: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

10 Anonymous et al.

line, where others such as Supporting Trump theme has an extremely increasingdashed line. Therefore, we manually analyzed the tweets of some IRA accountsand we found this observation clear, as an example from user x:

Islam and War : (A) @RickMad: Questions are a joke, a Muslim askshow SHE will be protected from Islamaphobia! Gmaffb! How will WEbe protected from terrori. . .

Supporting Trump: (B) @realDonaldTrump: That was really exciting.Made all of my points. MAKE AMERICA GREAT AGAIN!

Figure 3 shows the flipping behaviour for user x by extracting the mean valueof the fear and joy emotions. The smaller difference between the fear and joyemotions in the Islam and War theme for this user is due to the ironic way oftweeting for the user (e.g. the beginning of tweet A: ”Questions are a joke”).Even though, the fear emotion is still superior to the joy. We notice a similarpattern in some of the regular users, although much more evident among IRAtrolls.

Fig. 3: Flipping emotions between themes by user x (an IRA troll).

To understand more the NLI features performance, given their high perfor-mance comparing to the other features, we extract the top important tokens foreach of the NLI feature subsets (see Figure 4). Some of the obtained resultsconfirmed what was found previously. For instance, the authors in [31] foundthat Russians write English tweets with more prepositions comparing to nativespeakers of other languages (e.g. as, about, because in (c) Stop-words and RP14

in (a) POS in Figure 4). Further research must be conducted to investigate indepth the rest of the results.Linguistic Analysis. We measure statistically significant differences in the cuesmarkers of Morality, LIWC, Bias and Subjectivity, Stance, and Bad and Sexual

14 RP stands for ”adverb, particle” in the POS tag set.

Page 11: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 11

(a) POS (b) DEPREL (c) Stop-words

Fig. 4: The top 10 important tokens in each of the NLI features.

words across IRA trolls and regular users. These findings presented in Table 3allows for a deeper understanding of IRA trolls.

False Positive Cases. The proposed features showed to be effective in the clas-sification process. We are interested in understanding the causes of misclassifyingsome of IRA trolls. Therefore, we manually investigated the false positive tweetsand we found that there are three main reasons: 1) Some trolls were tweetingin a questioning way by asking about general issues; we examined their tweetsbut we did not find a clear ideological orientation or a suspicious behaviour intheir tweets. 2) Some accounts were sharing traditional social media posts (e.g.”http://t.co/GGpZMvnEAj cat vs trashcan”); the majority of the false positiveIRA trolls are categorized under this reason. In addition, these posts were givena false theme name; the tweet in the previous example assigned to AttackingHillary theme. 3) Lack of content. Some of the misclassified trolls mention onlyexternal links without a clear textual content. This kind of trolls needs a secondstep to investigate the content of the external links. Thus, we tried to read thecontent of these links but we found that the majority of them referred to deletedtweets. Probably this kind of accounts was used to ”raise the voice” of othertrolls, as well as, we argue that the three kinds of IRA trolls were used for ”likesboosting”.

Page 12: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

12 Anonymous et al.

Table 3: Linguistic analysis of Morality, LIWC, Bias and Subjectivity, Stance,and Bad and Sexual cues shown as the percentage of averaged value oftweets with one or more cues across IRA trolls (X) and regular users (Y) in ashape of X(arrows)Y. We report only significant differences: p-value ≤ 0.001↑↑↑,≤ 0.01↑↑, ≤ 0.05↑ estimated using the Mann-Whitney U test. The tweets averagevalue is the mean value across the themes.

Morality LIWC Bias language

category Pvalue category Pvalue category Pvalue

care 1.39↑↑↑0.74 pronoun 53.34↑47.59 assertive 6.53↓↓↓7.05

harm 2.39↑↑↑1.61 anx 1.91↑↑↑0.98 bias ×fairness 0.64↓↓↓0.84 cogmech × factive 5.58↑↑↑3.95

cheating 0.06↓↓0.31 insight 12.16↑↑↑10.08 hedge 10.06↑↑↑9.69

loyalty 0.84↓↓↓1.26 cause 10.76↑↑↑10.27 implicative 9.09↑↑↑6.37

betrayal 0.13↓↓↓0.35 discrep 12.73↑↑↑11.07 report 14.37↓↓↓18.89

authority 1.59↓↓↓1.88 tentat 13.95↑↑↑12.29 strong subj 54.15↑↑↑49.9

subversion 0.38↑↑↑0.33 certain 13.59↑↑↑10.69 weak subj 50.33↑41.96

Sanctity 0.44↑↑↑0.27 inhib 4.19↑↑↑3.87 - -

degradation 0.54↑↑↑0.49 incl 20.69↓↓21.24 - -

Stance Bad and Sexual

category Pvalue category Pvalue

belief 2.97↑↑↑2.49 bad 5.45↑↑↑4.66

denial 0.63↑↑↑0.57 sexual 3.54↑↑↑3.16

doubt 1.34↑↑↑1.25 - -

fake 0.49↓↓↓1.22 - -

knowledge 0.75↓↓↓1.48 - -

negation 11.46↑↑↑9.10 - -

question 3.12↑↑↑2.44 - -

report 2.86↓↓↓3.46 - -

6 Conclusion

In this work, we present a textual approach to detect social media trolls, namelyIRA accounts. Due to the anonymity characteristic that social media provide tousers, these kinds of suspicious behavioural accounts have started to appear. Webuilt a new machine learning model based on theme-based and profiling featuresthat in cross-validation evaluation achieved a F1macro value of 0.94. We applieda topic modeling algorithm to go behind the superficial textual information ofthe tweets. Our experiments showed that the extracted themes boosted the per-formance of the proposed model when coupled with other surface text features.In addition, we proposed NLI features to identify IRA trolls from their writingstyle, which showed to be very effective. Finally, for a better understanding weanalyzed the IRA accounts from emotional and linguistic perspectives.

Through the manually checking of IRA accounts, we noticed that frequentlyirony was employed. As a future work, it would be interesting to identify theseaccounts by integrating an irony detection module.

Page 13: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 13

References

1. Badawy, A., Ferrara, E., Lerman, K.: Analyzing the Digital Traces of Political Ma-nipulation: the 2016 Russian Interference Twitter Campaign. In: 2018 IEEE/ACMInternational Conference on Advances in Social Networks Analysis and Mining(ASONAM). pp. 258–265. IEEE (2018)

2. Badawy, A., Lerman, K., Ferrara, E.: Who Falls for Online Political Manipulation?In: Companion Proceedings of The 2019 World Wide Web Conference. pp. 162–168.ACM (2019)

3. Bahuleyan, H., Vechtomova, O.: UWaterloo at SemEval-2017 Task 8: DetectingStance Towards Rumours with Topic Independent Features. In: Proceedings of the11th International Workshop on Semantic Evaluation (SemEval-2017). pp. 461–464(2017)

4. Bhargava, M., Mehndiratta, P., Asawa, K.: Stylometric Analysis for AuthorshipAttribution on Twitter. In: International Conference on Big Data Analytics. pp.37–47. Springer (2013)

5. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet Allocation. Journal of machineLearning research 3(Jan), 993–1022 (2003)

6. Boatwright, B.C., Linvill, D.L., Warren, P.L.: Troll Factories: The Internet Re-search Agency and State-Sponsored Agenda Building. Resource Centre on MediaFreedom in Europe (2018)

7. Boyd, R.L., Spangher, A., Fourney, A., Nushi, B., Ranade, G., Pennebaker, J.,Horvitz, E.: Characterizing the Internet Research Agency’s Social Media Op-erations During the 2016 US Presidential Election using Linguistic Analyses.PsyArXiv (2018)

8. Cimino, A., Dell’Orletta, F.: Stacked Sentence-Document Classifier Approach forImproving Native Language Identification. In: Proceedings of the 12th Workshopon Innovative Use of NLP for Building Educational Applications. pp. 430–437(2017)

9. Clark, E.M., Williams, J.R., Jones, C.A., Galbraith, R.A., Danforth, C.M., Dodds,P.S.: Sifting Robotic from Organic Text: A Natural Language Approach for De-tecting Automation on Twitter. Journal of Computational Science 16, 1–7 (2016)

10. Dhingra, B., Zhou, Z., Fitzpatrick, D., Muehl, M., Cohen, W.W.: Tweet2Vec:Character-Based Distributed Representations for Social Media. In: The 54th An-nual Meeting of the Association for Computational Linguistics. p. 269 (2016)

11. Frenda, S., Ghanem, B., Montes-y Gomez, M.: Exploration of Misogyny in Spanishand English Tweets. In: 3rd Workshop on Evaluation of Human Language Tech-nologies for Iberian Languages (IberEval 2018). vol. 2150, pp. 260–267. CEUR-WS(2018)

12. Ghanem, B., Cignarella, A.T., Bosco, C., Rosso, P., Pardo, F.M.R.: UPV-28-UNITO at SemEval-2019 Task 7: Exploiting Post’s Nesting and Syntax Infor-mation for Rumor Stance Classification. In: Proceedings of the 13th InternationalWorkshop on Semantic Evaluation. pp. 1125–1131 (2019)

13. Gorrell, G., Bakir, M.E., Roberts, I., Greenwood, M.A., Iavarone, B., Bontcheva,K.: Partisanship, Propaganda and Post-Truth Politics: Quantifying Impact in On-line. arXiv preprint arXiv:1902.01752 (2019)

14. Goutte, C., Leger, S.: Exploring Optimal Voting in Native Language Identification.In: Proceedings of the 12th Workshop on Innovative Use of NLP for BuildingEducational Applications. pp. 367–373 (2017)

Page 14: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

14 Anonymous et al.

15. Graham, J., Haidt, J., Nosek, B.A.: Liberals and Conservatives Rely on DifferentSets of Moral Foundations. Journal of personality and social psychology 96(5),1029 (2009)

16. Haskell, W.: People Explaining their ’Personal Paradise’ is the LatestHashtag to Explode on Twitter (2015), https://www.businessinsider.com.au/hashtag-games-on-twitter-2015-6

17. Hooper, J.B.: On Assertive Predicates, vol. 4. In J. Kimball, editor, Syntax andSemantics (1974)

18. Hyland, K.: Metadiscourse: Exploring Interaction in Writing. Bloomsbury Pub-lishing (2018)

19. Im, J., Chandrasekharan, E., Sargent, J., Lighthammer, P., Denby, T., Bhargava,A., Hemphill, L., Jurgens, D., Gilbert, E.: Still out there: Modeling and IdentifyingRussian Troll Accounts on Twitter. arXiv preprint arXiv:1901.11162 (2019)

20. Karttunen, L.: Implicative Verbs. Language pp. 340–358 (1971)

21. Kiparsky, P., Kiparsky, C.: Fact. Linguistics Club, Indiana University (1968)

22. Malmasi, S., Evanini, K., Cahill, A., Tetreault, J., Pugh, R., Hamill, C., Napolitano,D., Qian, Y.: A Report on the 2017 Native Language Identification Shared Task.In: Proceedings of the 12th Workshop on Innovative Use of NLP for BuildingEducational Applications. pp. 62–75 (2017)

23. Markov, I., Chen, L., Strapparava, C., Sidorov, G.: CIC-FBK Approach to NativeLanguage Identification. In: Proceedings of the 12th Workshop on Innovative Useof NLP for Building Educational Applications. pp. 374–381 (2017)

24. Mohammad, S., Kiritchenko, S., Sobhani, P., Zhu, X., Cherry, C.: Semeval-2016Task 6: Detecting Stance in Tweets. In: Proceedings of the 10th InternationalWorkshop on Semantic Evaluation (SemEval-2016). pp. 31–41 (2016)

25. Mohammad, S.M., Turney, P.D.: Emotions Evoked by Common Words andPhrases: Using Mechanical Turk to Create an Emotion Lexicon. In: Proceedingsof the NAACL HLT 2010 workshop on computational approaches to analysis andgeneration of emotion in text. pp. 26–34. Association for Computational Linguistics(2010)

26. Ng, A.: This was the Most Viewed Facebook ad Boughtby Russian Trolls (2018), https://www.cnet.com/news/this-was-the-most-viewed-facebook-ad-bought-by-russian-trolls/

27. Owoputi, O., O’Connor, B., Dyer, C., Gimpel, K., Schneider, N., Smith, N.A.: Im-proved Part-of-Speech Tagging for Online Conversational Text with Word Clusters.In: Proceedings of the 2013 conference of the North American chapter of the as-sociation for computational linguistics: human language technologies. pp. 380–390(2013)

28. Recasens, M., Danescu-Niculescu-Mizil, C., Jurafsky, D.: Linguistic Models forAnalyzing and Detecting Biased Language. In: Proceedings of the 51st AnnualMeeting of the Association for Computational Linguistics. vol. 1, pp. 1650–1659(2013)

29. Sultana, M., Polash, P., Gavrilova, M.: Authorship Recognition of Tweets: A Com-parison Between Social Behavior and Linguistic Profiles. In: 2017 IEEE Interna-tional Conference on Systems, Man, and Cybernetics (SMC). pp. 471–476. IEEE(2017)

30. Tausczik, Y.R., Pennebaker, J.W.: The Psychological Meaning of Words: LIWCand Computerized Text Analysis Methods. Journal of language and social psychol-ogy 29(1), 24–54 (2010)

Page 15: arXiv:1910.01340v1 [cs.CL] 3 Oct 2019

TexTrolls: Identifying Russian Trolls on Twitter from a Textual Perspective 15

31. Volkova, S., Ranshous, S., Phillips, L.: Predicting Foreign Language Usage fromEnglish-Only Social Media Posts. In: Proceedings of the 2018 Conference of theNorth American Chapter of the Association for Computational Linguistics: HumanLanguage Technologies. vol. 2, pp. 608–614 (2018)

32. Zannettou, S., Caulfield, T., De Cristofaro, E., Sirivianos, M., Stringhini, G., Black-burn, J.: Disinformation Warfare: Understanding State-Sponsored Trolls on Twit-ter and their Influence on the Web. In: Companion Proceedings of The 2019 WorldWide Web Conference. pp. 218–226. ACM (2019)

33. Zheng, R., Li, J., Chen, H., Huang, Z.: A Framework for Authorship Identificationof Online Messages: Writing-Style Features and Classification Techniques. Journalof the American society for information science and technology 57(3), 378–393(2006)