Sentiment analysis of political communication: combining a dictionary approach with crowdcoding

Quality & Quantity, Sep 2016

Sentiment is important in studies of news values, public opinion, negative campaigning or political polarization and an explosive expansion of digital textual data and fast progress in automated text analysis provide vast opportunities for innovative social science research. Unfortunately, tools currently available for automated sentiment analysis are mostly restricted to English texts and require considerable contextual adaption to produce valid results. We present a procedure for collecting fine-grained sentiment scores through crowdcoding to build a negative sentiment dictionary in a language and for a domain of choice. The dictionary enables the analysis of large text corpora that resource-intensive hand-coding struggles to cope with. We calculate the tonality of sentences from dictionary words and we validate these estimates with results from manual coding. The results show that the crowdbased dictionary provides efficient and valid measurement of sentiment. Empirical examples illustrate its use by analyzing the tonality of party statements and media reports.

A PDF file should load here. If you do not see its contents the file may be temporarily unavailable at the journal website or you do not have a PDF plug-in installed and enabled in your browser.

Alternatively, you can download the file locally and open with any standalone PDF reader:

https://link.springer.com/content/pdf/10.1007%2Fs11135-016-0412-4.pdf

Sentiment analysis of political communication: combining a dictionary approach with crowdcoding

Sentiment analysis of political communication: combining a dictionary approach with crowdcoding Martin Haselmayer 0 Marcelo Jenny 0 0 Department of Government, University of Vienna , Rooseveltplatz 3/1, 1090 Vienna , Austria Sentiment is important in studies of news values, public opinion, negative campaigning or political polarization and an explosive expansion of digital textual data and fast progress in automated text analysis provide vast opportunities for innovative social science research. Unfortunately, tools currently available for automated sentiment analysis are mostly restricted to English texts and require considerable contextual adaption to produce valid results. We present a procedure for collecting fine-grained sentiment scores through crowdcoding to build a negative sentiment dictionary in a language and for a domain of choice. The dictionary enables the analysis of large text corpora that resourceintensive hand-coding struggles to cope with. We calculate the tonality of sentences from dictionary words and we validate these estimates with results from manual coding. The results show that the crowdbased dictionary provides efficient and valid measurement of sentiment. Empirical examples illustrate its use by analyzing the tonality of party statements and media reports. Previous versions of this paper have been presented at the 4th EPSA conference (Edinburgh 2014), the Workshop 'Political Context Matters' (Mannheim 2014), and the Symposium on 'New Frontiers of Automated Content Analysis in the Social Sciences' (Zu¨rich 2015). We thank our discussants at these conferences and the participants in our department's research seminar, in particular Thomas M. Meyer, Wolgang C. Mu¨ller, and our reviewers for their helpful feedback. We are grateful to Paul Ameli, Isabella Angermayr, Jakob-Moritz Eberl, Lukas Haselsteiner, Veronika Heider, Lisa Hirsch, Vanessa Kinz, Paul Preuer and Katharina Wurzer, and last, but not least the crowdcoders for their valuable contributions. Sentiment analysis campaigning; Media negativity Crowdcoding Political communication Negative 1 Introduction Sentiment analysis of textual data has manifold applications in the social sciences, among them the study of polarization, public opinion or media tone (e.g., Monroe et al. 2008; Van Atteveldt et al. 2008; Hopkins and King 2010; Soroka 2012; Young and Soroka 2012; Burscher et al. 2015; Gonza´lez-Bailo´ n and Paltoglou 2015; Soroka et al. 2015a, b) . However, a lack of tools or procedures for producing or collecting sentiments ratings of acceptable quality for large-scale data analyses currently hampers progress, in some languages more than in others. Computer-based approaches dominate the field of sentiment analysis, which attempt to produce the same sentiment rating of texts as a human reader. Unfortunately for social scientists interested in phenomena such as political polarization or media tone in nonEnglish countries, automated methods exhibit a strong language bias as they are developed and validated predominantly with textual data in English language. The number of sentiment analysis tools available for other languages is much smaller and their output tends to be of lower quality (Mohammad 2016) . If computer-based sentiment analysis is not available or its results are not good enough, one can resort to traditional content analysis with human coders. However, in ‘big data’ research projects manual content analysis quickly faces the restrictions of limited time, money and small numbers of trained coders. We outline a measurement procedure that (1) alleviates resource constraints (2) produces sentiment ratings that meet conventional quality standards, and (3) allows a researcher to conduct sentiment analyses in his or her language and domain of interest. Applying this procedure, we create a German language sentiment dictionary for the analyses of party statements and media reports. We use crowdcoding, the services of online coders, to produce the sentiment ratings of dictionary words. The sentiment dictionary is available for download,1 but similar to Laver and Garry (2000, p. 626) we want to highlight the procedure rather than a specific product: ‘‘Most important, given changing political meanings of words over time and space, is the procedure for deriving a dictionary, rather than the substantive content of any given dictionary.’’ By presenting our procedure we want to support ‘‘sentiment analysis in the resource poor languages’’ (Mohammad 2016, p. 203) and encourage the creation of customized dictionaries that fit well for the domain (Grimmer and Stewart 2013) and language studied. The structure of the paper is as follows. The next section deals with the measurement of sentiment in political discourse. Section three introduces crowdcoding as a data collection technique. Section four covers the creation of a sentiment dictionary, section five the rating of texts. Then we compare dictionary-based sentiment scores of texts with the results of manual coding. We also compare the scores from our custom-built sentiment dictionary for political communication with scores from existing, non-domain specific sentiment dictionaries. Section six includes two empirical illustrations to show the value of the data produced. The first covers negative campaigning in the 2013 Austrian national elections, the second looks into media tone. Finally, we discuss critical points and outline future uses of the procedure. 1 https://homepage.univie.ac.at/martin.haselmayer/. Sentiment analysis measures the polarity or tonality of texts by identifying and assessing expressions people use to evaluate or appraise persons, entities or events (Pang and Lee 2008; Liu 2012; Soroka 2014; Mohammad 2016) . Analyzing the polarity of texts has a long tradition in the social sciences. A prominent example is media negativity, a concept that captures the over-selection of negative over positive news, the tonality of media stories, and the degree of conflict or confrontation in the news (Esser and Stro¨ mba¨ck 2012) . Its ‘‘measurement in quantitative content analytic research can be defined as the process of linking certain aspects of textual data to numerical values that represent the presence, intensity, and frequency of textual aspects relevant to communication research’’ (Lengauer et al. 2012, p. 183). A number of recent studies demonstrate the benefits of sentiment analysis for such analyses (Van Atteveldt et al. 2008; Soroka 2012; Young and Soroka 2012; Burscher et al. 2015; Soroka et al. 2015a, b) . Sentiment analysis has also been used to establish the level of support for legislative proposals or polarization from the analysis of parliamentary debates (Monroe et al. 2008) , to identify issue positions or public opinion in online debates (Hopkins and King 2010; Ceron et al. 2012; Gonza´lez-Bailo´ n and Paltoglou 2015) , or for studying negative campaigning (Kahn and Kenney 2004; Lau and Pomper 2004; Geer 2006; Nai and Walter 2015) to mention just a few prominent uses. The classification of text as positive, negative, or neutral, is denoted by expressions such as polarity, valence or tone (Wilson et al. 2005; Young and Soroka 2012; Thelwall and Buckley 2013; Gonza´lez-Bailo´n and Paltoglou 2015; Mohammad 2016) . An incomplete list of terms for the gradual or quantitative measurement of sentiment includes potency (Osgood et al. 1957); intensity, sentiment strength (e.g. Thelwall et al. 2010) or emotive force (Macagno and Walton 2014). We will use sentiment strength and tonality as synonymous terms for a fine-grained measure of negativity. We cover only the neutral to negative part of the sentiment scale as psychological research has highlighted asymmetries between positive and negative evaluations of situations, persons or events (Peeters 1971; Peeters and Czapinski 1990; Cacioppo and Berntson 1994; Baumeister et al. 2001; Rozin and Royzman 2001) . We also do not probe into different ‘negative’ emotions (Ekman 1992) nor look at causes of negative evaluations (Soroka 2014; Soroka et al. 2015a) . The field of sentiment analysis is dominated by computer-based, automated approaches whose progress varies strongly by language (Mohammad 2016) . Many social scientists will be still more familiar with human-based content analyses with or without dictionaries (Stone et al. 1966; Budge and Farlie 1983; Baumgartner and Jones 1993; Laver and Garry 2000; Young and Soroka 2012; Krippendorff 2013) . Both manual and automated text analysis require an initial step of coding (or annotating or labelling) the sentiment of a text unit. Supervised and non-supervised automated approaches employ sample texts with coded sentiment ratings to ‘learn’ the sentiment of words. Once that phase of the research process has concluded—which usually includes a considerable amount of ‘fine-tuning’ the procedure—, the algorithms are scalable to large text corpora. Manual coding, in contrast, does not scale well as human coders often have to rate small units of texts such as sentences or words. Compared to unit by unit hand coding creating and using a dictionary of words already coded is a big step towards higher efficiency. An automated search can then find out whether a new text unit contains a dictionary word and retrieve its sentiment value. A basic assumption of using a dictionary is that it contains the most important words required for rating a text. A recent comparison of English language dictionaries and machine learning approaches found that ‘‘dictionaries had exceptional precision, but very low recall, suggesting that the method can be accurate, but that current lexicons are lacking scope. Machine learning systems worked in the opposite manner, exhibiting greater coverage but more error’’ (Soroka et al. 2015a, p. 112) . A large dictionary can provide good scope, but dictionary size on its own misleads about the quality of the output as irrelevant vocabulary produces less discriminating sentiment scores (Gonza´lez-Bailo´ n and Paltoglou 2015) . Related is the problem of domain specificity. Sentiment scores of words extracted from a training set of annotated texts do not generalize well to texts from other domains. Social scientists have accordingly stressed the need for custom-made dictionaries (Loughran and McDonald 2011; Young and Soroka 2012; Grimmer and Stewart 2013; Gonza´lez-Bailo´ n and Paltoglou 2015; Soroka et al. 2015a) . Even some commercial providers advise against using a sentiment dictionary ‘as is’ without thorough customization.2 We have pointed out that creating a customized dictionary or setting up a sample of training texts for machine learning requires an initial step of human coding which will be a procedural bottleneck if unit-by-unit sentiment coding has to be done with a small number of coders. We mitigate this bottleneck through crowdcoding, which offers a cheap and fast way to collect annotations for large amounts of text. 3 Employing crowdcoding3 to create a sentiment dictionary The idea of crowdsourcing draws on ‘‘wisdom of the crowd’’ arguments (e.g., List and Goodin 2001) and evaluations of expert-coded versus crowdcoded data show that for many tasks small aggregates of non-expert annotations are as good as single-expert annotation (Snow et al. 2008; Alonso and Baeza-Yates 2011) . Crowdsourcing online platforms such as Amazon’s Mechanical Turk, CrowdFlower and others provide access to an international large workforce for ‘‘micro tasks’’ requiring human intelligence. These lay coders have identified sentiment in texts with good results (e.g., Hsue et al. 2009; Taboada et al. 2011) . Political scientists have employed crowdsourcing for data generation (Berinsky et al. 2012, 2014) , for instance for content analyses of election manifestos (Benoit et al. 2016) . Using a large anonymous online workforce naturally raises data quality concerns. The best crowdsourcing platforms provide tools for quality control and real-time scrutiny of the data generation process, such as coder recruitment based on previous work record, skills, context knowledge or geographic location. Test questions can be randomly interspersed in a coding task to identify bad performance, and ‘‘screener’’ questions to check the attention of coders during the coding process (Berinsky et al. 2014) . Crowdcoding facilitates the completion of a large coding project at relatively low costs. Yet it still is unit by unit coding, and time and monetary costs increase with the number of units to be coded. If the goal is to code large amounts of texts using a dictionary with a good scope is an economic alternative. Creating the dictionary is a one-time, fixed-coast investment in time and money. Its application to large text corpora incurs few additional costs, apart from some text preprocessing. Large-scale text analyses can be easily repeated 2 http://provalisresearch.com/products/content-analysis-software/wordstat-dictionary/sentiment-dictionaries/. 3 The tasks carried out on crowdsourcing platforms extend far beyond text coding. We use the term crowdcoding for content analysis tasks. whenever a dictionary gets additional entries or modified sentiment scores. The next section shows step by step how to build your own sentiment dictionary. 4 Building a negative sentiment dictionary A negative sentiment dictionary consists of words with sentiment scores. Our procedure contains the following steps: (1) (2) (3) (4) (5) Sampling sentences from the domain of interest Crowdcoding the sentiment strength of sentences Estimating a sentence tonality score Estimating a word tonality score Discriminating between important and unimportant words Note that we move from words to sentences and back to words as relevant units. The reason is that we ask coders to rate complete sentences instead of single words taken out of context. Figure 1 shows a flow chart of the procedure: 4.1 Sampling sentences with negative sentiment from domain of interest We are substantively interested in the tonality of political communication of Austrian parties and media, and assemble a corpus of party press releases, minutes of parliamentary debates and media reports on election campaigns from the years 1995–2013. The texts are available in machine-readable format.4 Sampling of sentences with negative sentiment from domain of interest (1) Sentence ( ) {contain future dictionary words ( 1− )} Crowdcoding sentiment strength of sentences (2) ( ) : coder( 1 − ) | (0,1,2,3,4, uncodable) (3) ( ) = Estimating sentence tonality score ( ) ∑ =1 (4) ( ) = Estimating word tonality score: ∑ =1 ( { }) Selecting relevant words Following Remus et al. (2010) and Liu (2012) we use a small set of common negative words from existing German language sentiment dictionaries as ‘‘seed words’’ to select potentially negative sentences (Remus et al. 2010; Waltinger 2010; Momtazi 2012; Diwisch and Siegel 2014) . The corpus initially consists of about 470,000 sentences. Prefiltering with seed words cuts its size to about 215,000 sentences with negative sentiment. From that corpus, we randomly select 13,000 sentences for crowdcoding.5 Pre-filtering with seed words is not required, but it reduces the coding costs. The alternative is to submit an unfiltered, random sample of sentences to crowdcoding, with many of these subsequently coded as neutral or positive. While this strategy will increase the coding costs, some may view the collecting of negative, positive and neutral sentiment words as an advantage. 4.2 Crowdcoding the sentiment strength of sentences As the texts are in German, we recruit only coders from Austria or Germany through the platform CrowdFlower. We provide the coding instructions in the Appendix. Each sentence is assigned to ten coders to rate its negativity on a 5-point scale ranging from 0 (not negative) to 4 (very strongly negative) or judge it uncodable. Individual coder performance is monitored. Before the actual coding begins four test questions have to be answered correctly, and one out of five sentences of a task is another test question. The selection of appropriate test sentences is crucial for the successful application of crowdsourcing. We (i) first selected a large number of sentences that we judged unanimously, (ii) asked a group of ten colleagues and student assistants to code them too, and (iii) selected only those that showed very strong agreement in the tonality coding. We also pretested the coding jobs to collect feedback from the crowd. After completion, coders evaluate a job, including the quality of the instructions and the fairness of test tasks. Finally, it is possible to monitor the performance of test questions in real time and to remove or adapt them if problems occur. Coding sentiment strength on a five-point ordinal scale is difficult (Pang et al. 2002; Hopkins and King 2010) and for the test questions we accepted two adjacent options on the five-point scale as correct answers. Coders ‘‘usually have difficulty distinguishing between two adjacent ordinal classes whereas distinguishing between two classes which are far away from each other is much easier’’ (Zhou et al. 2014, p. 2). The probability of passing the first test by guessing is only 4 % and it gets smaller with each additional test unit.6 A coder dropping below an accuracy threshold of 75 % during coding is stopped from further contributing and his or her data not included in the data set. 480 coders answered on average 92 % (standard deviation of 0.07) of the tests questions correctly and contributed ratings to the data set. Overall, we collected about 130,000 valid codings (ten ratings per sentence), split into eight tasks, for which we paid 4 Press releases and media reports were collected as part of the Austrian National Election Survey (AUTNES). The parliamentary debates are available online at the Parliament’s Website (https://www. parlament.gv.at). 5 The target sample size corresponded to a budget limit of about 2000 Euros (roughly 2200 US-dollars) for coding, based on estimated task duration and payment of local minimum wages. After deleting duplicate and incomplete sentences from the automatic text preprocessing, the final sample size was 12,713 sentences. 6 An accuracy threshold of 75 % means three of the four initial test questions have to be correct. Including ‘‘uncodable’’ a coder has six options with two accepted as correct. The probability of passing by guessing then is 26 3 = 0.04. 2000 Euros in total (about 2200 US-dollars) through the crowdcoding platform. We provide a list of the crowdcoding jobs in the Appendix. 4.3 Estimating sentence tonality scores For each sentence we collect negativity ratings from ten coders (cj) and calculate a mean sentence score t(si). tðsiÞ ¼ P j j¼1 cjðsiÞ j ð1Þ ð2Þ 4.4 Estimating word tonality scores This is also the initial tonality score of each word or more specifically word form contained in a sentence. We restrict the dictionary to single words and do not consider combinations of words (bigrams, trigrams) or short phrases. We lemmatize the word forms and do partof-speech tagging with the tool Treetagger (Schmidt 1994) , a process that due to the current quality of such tools for the German language requires some manual post-processing of results. An alternative to lemmatizing is to include all existing word forms in the dictionary (e.g. Remus et al. 2010) , which is difficult with languages, such as Finnish (or Hungarian, Turkish, and Russian), where nouns can take up to 2000 different forms. Here, a feasable strategy is to concentrate on the most important word forms (Kettunen and Airio 2006). Then we check the frequency of words. If a word wk appears in more than one sentence, we calculate a mean word score from these sentences t(si). tðwkÞ ¼ Pn l¼1 tlðsiÞ nk Note that the double step of mean aggregation of ordinal scores produces numbers with decimal places, which methodological purists can object to. Alternative aggregation measures exist if the goal is to preserve the original 5-point scale (Dawid and Skene 1979; Zhou et al. 2014; Felt et al. 2015) . A recent crowdsourcing study by Benoit et al. (2016) found that ‘‘means of means’’ gave almost the same results as more complex algorithms. 4.5 Separating relevant and irrelevant words At this stage the complete list or ‘‘bag of words’’ contained in the rated sample of sentences has a sentiment score. However, we want only words in a sentiment dictionary that express negative tonality with a high probability and delete the rest of the list as irrelevant. We start by cleaning the database and remove all words that have less than three characters (n = 960, most of which are due to errors in text pre-processing). Word frequency is a standard indicator of relevance in automated text analyses. The more common a word is the less informative is it about a specific quality such as negativity. There is no gold standard for deleting high-frequency words. We delete words such as articles, pronouns as well as names using part-of-speech-tags (Schmidt 1994) and use stop word lists that identify highly frequent words based on the Leipzig Corpora Collection (Quasthoff et al. 2006) (n = 4518). For a different reason we also delete rare words. We aim at collecting word negativity as a global rather than as a local, highly context-dependent quality. Therefore, we drop unique words (n = 24,511) that appear in a single sentence as containing too much measurement error. We identify and remove positive words from existing sentiment dictionaries (Wolf et al. 2008; Klenner et al. 2010; Remus et al. 2010; Waltinger 2010; Momtazi 2012; Diwisch and Siegel 2014) (n = 3725). Then we delete any remaining named entities from the list. We use online available lists for named entity recognition (Faruqi and Pado 2010; Steinberger Ralf et al. 2011; Benikova et al. 2014) and a set of named entities from the AUTNES project to identify and delete the names of politicians, parties, or organizations (n = 6378). The deletion of named entities, of stop words and rare words reduces the number of words from initially about 40,000 to about 5000 words. Figure 2 shows the distribution of tonality scores for the 5001 words in the dictionary which range from 0.06 to 3.8 on the scale from 0 (not negative) to 4 (very strongly negative). The mean tonality score of the dictionary words is 2.04 (standard deviation of 0.65). 5 Scoring sentences and texts Unit scale in sentiment analysis varies from documents to sentences to smaller textual unit such as word groups or single words. If a procedure estimates sentiment scores for words, one needs an aggregation rule to get a sentence-level or document-level score. Our scoring approach rests on the ‘‘bag of words’’ assumption (Laver et al. 2003; Monroe et al. 2008; Slapin and Proksch 2008) . We equate the tonality of a sentence with the tonality of a dictionary word contained in it. If a sentence (si) contains several different sentiment words (wk), we apply the ‘‘maximum’’ rule of Thelwall and Buckley (2013) which means the most strongly negative word (max(wk)) sets the tonality of the sentence.7 tðsiÞ ¼ maxðwkÞ ð3Þ 7 We test other aggregation algorithms such as mean scores of all words, which resulted in a lower correlation with the expert coding. See the Appendix with the results of different aggregation rules. The dictionary includes negation words (n = 13, such as no, not, never, neither, without) and intensifier words (n = 53, e.g. completely, exceedingly, extremely, heavily, very). If a negation word immediately precedes a dictionary word, we exclude the latter from the calculation of the tonality score for the sentence rather than flipping the polarity of a sentence (see Thelwall et al. 2012) . One could use a similar strategy for the negation of positive words, but we focus on negativity only. Taboada (2016 , pp. 332–335) points out that polarity inversion usually translates into low intensity scores. Following Taboada et al. (2011) , we amplify a dictionary word’s negativity score (by a factor of 1.25) if it is preceded by an intensifier word, up to the maximum value set by the scale’s boundary. 6 Validating the procedure Face validity (e.g., Monroe et al. 2008) and cross-validation (e.g., Laver et al. 2003; Slapin and Proksch 2008) are popular standards used to evaluate results from automated text analyses, but the gold standard is a comparison with results from human coding (Grimmer and Stewart 2013; Lowe and Benoit 2013, p. 13) . To check the validity of our approach we use a random sample of 200 sentences from party press releases from four national election campaigns held between 2002 and 2013 as well as media reports from the most recent campaign. Like Benoit et al. (2016) , we evaluate the validity of our approach by comparing the aggregated, rather than individual coder results obtained through crowdcoding to manual expert annotation. Each of the authors separately coded the sample sentences on a 5-point scale. A group of ten online recruited coders completed the same task. The mean sentence scores, aggregated for two expert ratings on the one hand and the group of lay coders on the other hand, exhibits a Pearson correlation of 0.82 (Fig. 3). Thus in line with previous research, we find that the group of lay coders was able to replicate the expert data (e.g. Snow et al. 2008; Benoit et al. 2016) , with a slight centrist bias in these aggregate ratings (Saal et al. 1980). We score the test sample with our dictionary, which results in a Pearson correlation between manual expert ratings and dictionary ratings of 0.65 with 84 % coverage. The level of correlation is at par with recent English language sentiment analyses employing similar levels of granularity (e.g., Strapparava and Mihalcea 1556; Thelwall and Buckley 2013) . Figure 4 provides a graphical representation of the correlation between dictionarybased scores and expert scores, including the linear regression between these two.8 There is again some degree of a centrist bias of the crowdscores due to mean aggregation. To provide a direct test of our argument that we need context-sensitive dictionaries, we subject our sample to analysis with two other German-language sentiment dictionaries (Remus et al. 2010; Momtazi 2012) . These dictionaries are translations of English sentiment dictionaries.9 Table 1 shows how their size and type of entries differs (e.g. words, word stems, word forms (including conjugation/declination), lemmas and synonyms). Applied to our test sample they exhibit a lower rate of coverage and their sentence scores have almost no 8 We provide an extended validation with a larger set of sentences coded by one of the authors in the Appendix, where we also report on effects of text preprocessing. The binary predictive quality of the dictionary (negative vs. not negative), based on a standard confusion matrix (e.g. Davis and Goadrich 2006; Fawcett 2006) results in an accuracy of 0.85, the F-score is 0.91. 9 SentiWS (Remus et al. 2010) contains positive and negative words along with word flexion (conjugation/ declination). The sentiment score provided by the dictionary indicates the probability of the word appearing in a positive or negative context. The sentiment dictionary created by Momtazi (2012) contains positive and negative words and word stems, as well as valence shifters and negation words. correlation with the manual expert coding of negative sentiment strength. This confirms the point that dictionary-based analysis requires a customized dictionary to begin with (Grimmer and Stewart 2013; Gonza´lez-Bailo´ n and Paltoglou 2015; Mohammad 2016) . We check whether our dictionary is large enough to provide a good coverage of the phenomenon under study. A dictionary with perfect scope should assign a negativity score to all sentences with negative content. Accordingly, the 32 sentences out of 200 without a matching dictionary word should not be negative. We assign these sentences negativity scores of zero and recalculate the correlation with manual coding which results in a slightly lower Pearson correlation of 0.6. Closer inspection shows a few sentences without a dictionary word that coders rated negatively. Most of them contain separable verbs (that we could not match) or express irony or sarcasm. A few sentences are rhetorical questions. A politician’s statement ‘‘Politics for Women is different’’ expressed dissatisfaction without using a manifestly negative word. The rhetorical question ‘‘Where has the Green’s objective environmental policy gone?’’ criticizes the party’s actions without a negative word. Irony, sarcasm and rhetorical questions are common pitfalls in automated text analyses. However, as long as these phenomena do not make up a major portion of the text corpus, the coverage rate of our dictionary appears fine. It is at a par with comparable English language sentiment dictionaries (e.g. Strapparava and Mihalcea 2008; Thelwall and Buckley 2013) . 7 Applications We now use the political sentiment dictionary in two applications. Specifically, we study the parties’ use negative campaigning and the tone of media coverage with data from the Austrian National Election Study. 7.1 Negative campaigning in the Austrian national elections 2013 Research on negative campaigning has predominantly relied on binary classifications of statements as negative or non-negative (e.g., Damore 2002; Lau and Pomper 2004; Walter 2014) which is easier to operationalize than a fine-grained measure of tonality. However, we contend that the degree of negativity matters. Weak expressions of criticism have different effects than strongly worded attacks. Studies find that voters react to the intensity of negative messages (e.g., Mutz and Reeves 2005; Fridkin and Kenney 2011) . Negative campaigning featured prominently in the 2013 Austrian national elections. We analyze rhetorical interaction between parties via press releases during the final 6 weeks of the campaign. As part of the Austrian National Election Study (M u¨ller et al. 2014) , a relational content analysis with human coders of the headlines of 1958 party press releases was done. We use a subset of these press releases, which contain 755 directed, negative relations between two parties. Words from the sentiment dictionary matched 82 % of these statements. Table 2 shows the frequency and tonality of negative campaigning of the six parliamentary parties competing in the 2013 Austrian election. The number of negative press releases sets the three largest parliamentary parties apart from the three smaller parties. The government parties SPO¨ and O¨ VP, and the opposition party FPO¨ account for almost four out of five negative press releases. Studies of negative campaigning in multi-party systems argue that government parties use fewer negative campaign statements than opposition parties (e.g., Walter and Van der Brug 2013) . At the same time, parties in government are expected to be the most important targets of negative campaigning (Walter 2014) . We draw on these arguments to explore the patterns of negative campaigning in the 2013 election campaigns. Additionally, we want to test evidence from a recent study, that the government parties (SPO¨ and O¨ VP) devote most of their negative campaigning on each other (Dolezal et al. 2015) . We expect that coalition partners criticize each other frequently but less strongly than other parties. To test these expectations, we perform an OLS regression using the tonality of a press release as our dependent variable. We have binary indicators for government (SPO¨ , O¨ VP) and opposition parties (FPO¨ , Greens, BZO¨ , Team Stronach) and distinguish negative statements among the coalition partners SPO¨ and O¨ VP from other party pairs. We use the performance of a party in the pre-electoral polls (using the net difference in poll standings at the beginning of the campaign with the election result) and the proximity of the election (in days) as control variables. Empirical research shows that parties that are losing ground in the electoral competition employ more negative campaigning and that campaigns become increasingly negative towards the end (Damore 2002) . The results in Table 3 and Fig. 5 indicate significant differences with regard to the tonality of negative campaign messages made by government and opposition parties. We also find evidence that parties refrain from using aggressive statements against their coalition partner. Negative campaigning from a government party was ceteris paribus 0.1 units less negative than negative statements from opposition parties. Inter-government conflict was on average 0.13 units less negative than criticism exchanged between other party pairs. Parties showing bad electoral performance issued more strongly worded campaign messages, but this effect disappears in the second model. We find no escalation of negativity towards the end of the campaign. 7.2 Media tonality in campaign reporting The second application focusses on how the media cover politics and transmit the parties’ campaign messages. These topics deserve study because mass media are the most important source of information for voters about a specific electoral contest (Str o¨mba¨ck and Kaid 2008) . Our starts off from the classic finding (Galtung and Ruge 1965) that negativity is a highly important factor determining the newsworthiness of an event. A wealth of studies from the United States has established the findings of the media’s focus on negative stories or its cynical reporting on politics (Patterson 1993; Capella and Jamieson 1997; Farnsworth and Lichter 2010) . We want to know whether sentences in media reports that mention a political party or a top candidate are more critical than statements without a reference to these political actors. We use the dictionary to measure the tonality of reporting on the six parliamentary parties and their top candidates in fifteen Austrian daily newspapers. The raw data consist of 15,096 news reports published during the final 8 weeks of the campaign for the national parliamentary elections of 2013. The media reports were collected as part of the Austrian National Election Study (Haselmayer et al. 2013; Kleinenvon Ko¨ nigslo¨ w et al. 2013) . They consist of 439,954 sentences, of which about one in five has a reference to a party or candidate. Slightly more than half (55 per cent) contain a dictionary word. For the rest we assume that the scope of the dictionary is sufficient to identify all overtly negative statements and code them as neutral statements. We compare the tonality of sentences with and without a reference to a party or candidate and find that the mean tonality of sentences with a reference is 1.23 across the fifteen print news outlets compared to 1.00 for the contrast group of statements without actor reference. It indicates that media coverage was slightly negative on average. Figure 6 shows the temporal variation in the last 6 weeks before the election. The slightly more negative tone whenever a political actor is mentioned can be clearly seen. Note that this application is purely illustrative. Whether it constitutes evidence of a critical or cynical perspective of journalists on politics (Patterson 1993; Capella and Jamieson 1997) would require further study. 8 Conclusions This article shows how to create a dictionary-based measurement procedure for negative sentiment in a language of choice that is cheap, fast, reliable and valid when compared to human coding. The English language bias of computer-based sentiment analysis constrains social scientists interested in studying textual data in other languages. For this reason, with the help of crowdcoders, we created a German political language dictionary tailored to party statements and media reports. Our results underscore that crowdcoding is a viable alternative to the use of expert coders or trained coders in the social sciences (Benoit et al. 2016) . Yet, even if the costs (in terms of both, money and time) are low when compared to trained coders, they increase linearly. Thus, large-scale analyses of (comprehensive) text units clearly impose limits to manual coding. Thus, computer-based sentiment analysis dominates the field. Its amazing pace of innovation, low costs and scalability make it a highly attractive, alternative approach. Yet, most tools for (semi-)automated text analyses were developed using English language texts. Convinced of their value social scientists frequently apply them to texts in other languages. Algorithms do not object to a language transfer as long as the strings to be processed appear well formed. Yet, prudent users should require empirical evidence on how well an automated text analysis tool operates in another language, which brings us back to the need for validation (Grimmer and Stewart 2013; Lowe and Benoit 2013) . We have empirically demonstrated the importance of using a customized dictionary. Two alternative sentiment dictionaries had substantially lower coverage and agreement with manual codings of sentiment strength. We would recommend the use of any generalpurpose dictionary with caution (see already Grimmer and Stewart 2013; Gonza´lez-Bailo´ n and Paltoglou 2015; Soroka et al. 2015a) . Our fine-grained dictionary-based sentiment scores move beyond a polarity classification of text. Although the sentences taken from press releases and media reports were a challenging test set our dictionary-based scores reflected the human ratings of crowdcoders and expert coders to a large degree. While the bag-of-words approach has its limitations, for example when confronted with figurative language, it performed well in the vast majority of cases. Sentiment analysis offers many exciting avenues for innovative social science research. We have shown the usefulness of our new sentiment dictionary in two applications: negative campaigning by parties and media tone. Future research could look at the incentives for negative campaigning in multi-party systems, for example rhetorical interaction between potential coalition partners (e.g., Walter and Van der Brug 2013) , and study the effects of campaign tonality on post-electoral government formation. Parallel sentiment analyses of campaigns and its media coverage allow an empirical assessment whether parties can attract media attention through negative campaign messages as hypothesized by Geer (2006) . Comparing the tonality of party campaign messages and their news coverage enables testing the presumed ‘negativity bias’ of the news media (Patterson 1993; Capella and Jamieson 1997; Farnsworth and Lichter 2010) . Sentiment analyses can also add to our understanding of the effects of negative campaigning and media negativity on voting behavior. Studying how voters react to variation in campaign tonality provides new contributions to the debate on potential benefits from negative campaigning with regard to political knowledge and turnout (e.g., Lau and Pomper 2004; Geer 2006) . Sentiment analysis can also contribute to other research topics at the intersection of communication science and political science such as the study of public opinion and political polarization (e.g., Monroe et al. 2008; Hopkins and King 2010; Gonza´lez-Bailo´ n and Paltoglou 2015) . Parliamentary debates, party manifestos, blogs and social media platforms provide rich data sources for sentiment analyses. We have shown how to create and use a dictionary for large-scale sentiment analyses. There are no limits in using the same or similar procedures to create a customized dictionary for other research areas. Finally, we foresee sentiment analyses with multi-language dictionaries. Debate transcripts from the European Parliament or the United Nations General Assembly provide multi-lingual textual data. Comparative sentiment analyses can submit textual data in several languages to crowdcoding or use tools for automated text translation. We have identified several research topics that may benefit from fine-grained sentiment analyses and have described our procedure for German language texts. Hopefully, we convinced readers that undertaking similar analyses with textual data in a language of choice is worthwhile and doable. Acknowledgments Research for this paper was conducted under the auspices of the Austrian National Election Study (AUTNES), a National Research Network (NFN) sponsored by the Austrian Science Fund (FWF) (S10903-G08, S10903-G11). Additional funding was provided by the University of Vienna’s Department of Social Sciences. Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. Appendix Appendix 1: Text preprocessing Automated text analysis requires text preprocessing. We found the performance of current natural language processing tools for lemmatization and part-of-speech tagging of separable verbs, compounds, dialectal variations, and the frequent neologisms in German political and media texts to be rather unsatisfactory and manually cleaned incorrect lemmas. Given imperfect lemmatization, we allow for imperfect word matching to increase the number of matches when applying the dictionary to new texts. Stemming (Porter 1980) is common in English language applications as an alternative to lemmatization, but stemming worked worse in our applications and produced the lowest correlation with manual coding (see Table 4). A combination of lemmatization, part-of-speech-tagging and deletion of stop words and named entities gave the best result (Schmidt 1994; Quasthoff et al. 2006; Faruqi and Pado 2010; Steinberger Ralf et al. 2011; Benikova et al. 2014) . Appendix 2: Effect of different aggregation rules Table 5 shows how different aggregation rules for word and sentence scores such as meanof-means for word scores (see Benoit et al. 2016) and selecting the most negative word as sentence score (Thelwall et al. 2012) as well as accounting for intensifier and negation words (Taboada et al. 2011; Thelwall et al. 2012) affect the correlations with manual expert coding. We obtained the highest correlation using means for calculating word scores, choosing the most negative word as sentence score and by accounting for intensifier and negation words. n = 168 dictionary-based scores for 200 test sentences Mean Yes 0.54 Max Yes 0.65 Appendix 3: Validation with a larger sample and single expert coding An additional validation test uses a larger sample (n = 755) of negative statements in party press releases from the same election campaign which were coded by only one of the authors. The dictionary’s coverage was 82 % and the correlation between manual and dictionary based sentiment coding 0.63. We validate the results of our regression model by re-running the same models as in Table 3 with the single expert coding as dependent variable. Table 6 shows identical signs of coefficients for all effects compared with the dictionary-based sentiment scores. The automated scoring procedure tends to show weaker effect sizes due to the two-step-mean aggregation. We conclude that our dictionary-based scoring produces valid sentiment estimates at the level of analysis that are more conservative than results based on manual coding. Rerunning the regression after assigning zero negativity to sentences that had no matching dictionary word (n = 136) did not change these results (not reported). Standard errors clustered across party pairs in parentheses. Differences in the number of observations are due to sentence with no matching dictionary words ? p \ 0.1, * p \ 0.05, *** p \ 0.001 Appendix 4: CrowdFlower coding jobs T T T T T T T re ,tr U U U U U U U d a s e ir A A A A A A A co ts tn , , , , , , , r e u R R R T R R R R e th o E E E U E E E E p t C G G G A G G G G s a g in % d 0 o 1 0 0 8 c t 0 .3 0 81 .8 4 8 f a .45 .58 4 .4 . 2 .5 .4 o t 6 9 8 7 3 9 r e 2 5 5 5 1 8 6 5 e s t b e s $ $ $ $ $ $ $ $ r o S S S S S S S S m e u C U U U U U U U U n w Appendix 5: Coding instructions (translation) The following coding instructions were pretested by colleagues, student assistants and a few online coders. How negative are these statements? What is this about? We present you sentences from political and media texts. Many, though not all, of these sentences include direct or indirect criticism, allegations or attacks. Task Please read each sentence carefully and decide, whether it includes a positive, neutral or negative statement. In a second step, we ask you to rate the intensity of the statement using the following scale: – – – – – – Not negative (neutral or positive) Very weakly negative Weakly negative Strongly negative Very strongly negative Not codable What should you consider? Only rate the actual content of the text! Stay impartial, your personal preferences towards persons or organizations should not influence your coding decisions. Not negative A sentence should be coded as ‘‘not negative’’ if it contains a neutral or positive statement. Example ‘‘not negative’’: ‘‘I serve the Austrian citizens with passion and commitment.’’ Not codable A sentence is ‘‘not codable’’ if it is incomprehensible or if it does not make any sense to you. Some sentences may be incomplete, as they have been processed automatically. As long as you are able to purposefully decide, whether they are positive, neutral or negative, we ask you to rate them anyhow. Example ‘‘not codable’’: ‘‘Ic$ %$#* we retain %, that & %§’’ Negative Negative sentences contain direct or indirect criticism, allegations or attacks in varying intensity. Examples with increasing negativity: ‘‘We demand that the government finally delivers a better job!’’ ‘‘These are bad actions, which come at the expense of the population.’’ ‘‘This minister promotes corruption and consciously dupes the people.’’ ‘‘This is a scam on all of us: the dishonesty of these politicians stinks to high heavens.’’ Special case: sentences containing specific coding instructions Some sentences may contain instructions, asking you to choose a specific category. In such cases, you should ignore all other textual information and directly follow the instructions. Example: ‘‘The government has failed to address these issues in the past legislative term. Please ignore the previous part of the text and code this unit as ‘‘not codable’’. In case of any question regarding the coding process or if you would like to provide us with feedback, please send us an E-Mail: . Thank you for your contribution! Appendix 6: CrowdFlower task script CrowdFlower jobs have to be designed in CrowdFlower Markup Language (CML), which is entirely based on HTML, but contains a small set of special features (e.g. to link data). Its implementation for simple coding tasks, such as the annotation of sentiment strength is very easy (Fig. 7): Appendix 7: CrowdFlower coding task Figure 8 shows a screenshot of the coding task as presented to our crowdcoders. Instructions (see Appendix 3) are hidden for a better readability. One page usually contains five sentences along with the question ‘‘How negative is this statement?’’ and the labelled coding options. Alonso , O. , Baeza-Yates , R. : Design and implementation of relevance assessments Using Crowdsourcing . In: Clough, P. , Foley , C. , Gurrin , C. , Jones , G. , Kraaj , W. , Lee , H. , Mudoch , V. (eds.) Advances in Information Retrieval , pp. 153 - 164 . Springer, Berlin ( 2011 ) Baumeister , R.A. , Bratlavsky , E. , Finkenauer , C. : Bad is stronger than good . Rev. Gen. Psychol . 5 ( 4 ), 323 - 370 ( 2001 ) Baumgartner , F.R. , Jones , B.D. : Agendas and instability in American politics. American politics and political economy series . University of Chicago Press, Chicago ( 1993 ) Benikova , D. , Biemann , C. , Reznicek , M. : NoSta-D Named Entity Annotation for German: Guidelines and Dataset . Proceedings of Language Resources and Evaluation Conference (LREC) 2014 , 26 -31 May, Reykjavik, Iceland ( 2014 ) Benoit , K. , Conway , D. , Lauderdale , B. , Laver , M. , Mikhaylov , S. : Crowd-sourced text analysis: Reproducible and agile production of political data . Am. Polit. Sci. Rev . 110 ( 2 ), 278 - 295 ( 2016 ) Berinsky , A.J. , Huber , G.A. , Lenz , G.S.: Evaluating online labor markets for experimental research: Amazon.com's mechanical turk . Polit. Anal . 20 ( 3 ), 351 - 368 ( 2012 ) Berinsky , A.J. , Margolis , M. , Sances , M.W. : Separating the shirkers from the workers? Making sure respondents pay attention on internet surveys . Am. J. Polit. Sci . 58 ( 3 ), 739 - 753 ( 2014 ) Budge , I. , Farlie , D. : Explaining and Predicting Elections: Issue Effects and Party Strategies in TwentyThree Democracies . Allen & Unwin , London ( 1983 ) Burscher , B. , Vliegenthart , R., de Vreese, C.H.: Frames Beyond words: applying cluster and sentiment analysis to news coverage of the nuclear power issue . Soc. Sci. Comput. Rev . 34 ( 5 ), 530 - 545 ( 2015 ) Cacioppo , J.T. , Berntson , G.G. : Relationship between attitudes and evaluative space: a critical review, with emphasis on the separability of positive and negative substrates . Psychol. Bull . 115 ( 3 ), 401 - 423 ( 1994 ) Capella , J.N. , Jamieson , K.H. : Spiral of Cynicism: The Press and the Public Good . Oxford University Press, Oxford ( 1997 ) Ceron , A. , Curini , L. , Iacus , S.M.: Every tweet counts? How sentiment analysis of social media can improve our knowledge of citizens' political preferences with an application to Italy and France . New Media Soc. 162 , 340 - 358 ( 2012 ) Damore , D.F. : Candidate strategy and the decision to go negative . Polit. Res. Q . 55 ( 3 ), 669 - 685 ( 2002 ) Davis , J. , Goadrich , M.: The relationship between Precision-Recall and ROC curves . In: Proceedings of the 23rd International Conference on Machine learning , Pittsburgh, PA, 25 - 29 June, pp. 233 - 240 ( 2006 ) Dawid , A.P. , Skene , A.M.: Maximum likelihood estimation of observer error-rates using the EM algorithm . J. R. Stat. Soc . 28 ( 1 ), 20 - 28 ( 1979 ) Diwisch , K. , Siegel , M.: Multi-domain sentiment Lexicon for German . In: University of Applied Sciences Darmstadt. https://sites.google.com/site/iggsahome/downloads ( 2014 ) Dolezal , M. , Ennser-Jedenastik , L. , M u¨ller, W.C.: When do parties attack their competitors? Negative campaigning in Austria, 2002 - 2008 . In: Nai, A. , Walter , A.S. (eds.) New Perspectives on Negative Campaigning: Measures, Causes and Effects , pp. 165 - 181 . ECPR Press, Colchester ( 2015 ) Ekman , P.: An argument for basic emotions . Cogn. Emot . 6 ( 3-4 ), 169 - 200 ( 1992 ) Esser , F. , Stro¨mba¨ck, J.: Comparing news on national elections . In: Esser, F. , Hanitzsch , T. (eds.) The Handbook of Comparative Communication Research , pp. 289 - 326 . Routledge, New York ( 2012 ) Farnsworth , S.J. , Lichter , R.S.: The Nightly News Nightmare: Network Television's Coverage of U.S. Presidential Elections , 1988 - 2000 . Rowman & Littlefield, Lanham ( 2010 ) Faruqi , M. , Pado , S. : Training and evaluating a german named entity recognizer with semantic generalization . In: Proceedings of KONVENS 2010 , 6-8 September, Saarbru¨cken ( 2010 ) Fawcett , T. : An introduction to ROC analysis . Pattern Recogn. Lett . 27 ( 8 ), 861 - 874 ( 2006 ) Felt , P. , Boyd-Graber , J. , Ringger , E.K. , Seppi , K. : Making the most of crowdsourced document annotations: confused supervised LDA . In: Proceedings of the 19th Conference on Computational Language Learning , Beijing, China, July 30 - 31 ( 2015 ) Fridkin , K.L. , Kenney , P.J.: Variability in citizens' reactions to different types of negative campaigns . Am. J. Polit. Sci . 55 ( 2 ), 307 - 325 ( 2011 ) Galtung , J. , Ruge , M.H.: The structure of foreign news: the presentation of the Congo, Cuba and Cyprus crises in four foreign newspapers . J. Peace Res . 2 ( 1 ), 64 - 91 ( 1965 ) Geer , J.G. : In Defence of Negativity. Attack Ads in Presidential Campaigns. University of Chicago Press, Chicago ( 2006 ) Gonza´lez-Bailo´n, S., Paltoglou , G.: Signals of public opinion in online communication: a comparison of methods and data sources . Ann. Am. Acad. Polit. Soc. Sci . 659 ( 1 ), 95 - 107 ( 2015 ) Grimmer , J. , Stewart , B.M. : Text as data: the promise and pitfalls of automatic content analysis methods for political texts . Polit. Anal . 21 ( 3 ), 267 - 297 ( 2013 ) Haselmayer , M. , Jacobi , C. , Eberl , J.-M. , Vonbun , R. , Kleinen-von Ko¨nigslo¨w, K., Scho¨nbach, K. , Boomgaarden , H.: AUTNES MediaAutomatic 2013 : Documentation ( 2015 ) Hopkins , D.J. , King , G. : A method of automated nonparametric content analysis for social science . Am. J. Polit. Sci . 54 ( 1 ), 229 - 247 ( 2010 ) Hsue , P.-Y., Melville , P. , Sindhwei , V. : Data quality from crowdsourcing: a study of annotation selection criteria . In: Proceedings of the NAACL HLT 2009 Workshop on Active Learning for Natural Language Processing , Boulder, Colorado, 5 June, pp. 27 - 35 ( 2009 ) Kahn , K.F. , Kenney , P.J.: When do candidates go negative ? In: Kahn, K.F. , Kenney , P.J. (eds.) No Holds Barred: Negative Campaigning in U.S. Senate Campaigns, pp. 19 - 37 . Pearson, Upper Saddle River ( 2004 ) Kettunen , A. , Airio , E.: Is a morphologically complex language really that complex in full-text retrieval ? In: Salakoski, T. , Ginter , F. , Pyysalo , S. , Pahikkala , T. (eds.) Advances in Natural Language Processing , pp. 411 - 422 . Springer, Heidlerberg ( 2006 ) Kleinen-von Ko ¨nigslo¨w, K., Jacobi , C. , Haselmayer , M. , Eberl , J.-M. , Vonbun , R. , Scho¨nbach, K. , Boomgaarden , H.: AUTNES MediaAutomatic 2013: Data ( 2015 ) Klenner , M. , Clematide , S. , Wiegand , M. , Peters , R. : Polarity lexicon: Nouns, adjectives, verbs, adverbs . In: Institute of Computational Linguistics, U.o.Z. , Switzerland (ed.). http://sentimental.li/german.lex ( 2010 ) Krippendorff , K. : Content Analysis. An Introduction to Its Methodology . Sage, Thousand Oaks ( 2013 ) Lau , R.R. , Pomper , G.M. : Negative Campaigning: An analysis of US Senate Elections . Rowman & Littlefield, Lanahan ( 2004 ) Laver , M. , Benoit , K. , Garry , J.: Extracting policy positions from political texts using words as data . Am. Polit. Sci. Rev . 97 ( 2 ), 311 - 331 ( 2003 ) Laver , M. , Garry , J.: Estimating policy positions from political texts . Am. J. Polit. Sci . 44 ( 3 ), 619 - 634 ( 2000 ) Lengauer , G. , Esser , F. , Berganza , R.: Negativity in political news: a review of concepts, operationalizations and key findings . Journalism 13 ( 2 ), 179 - 202 ( 2012 ) List , C. , Goodin , R.E. : Epistemic democracy: generalizing the condorcet jury theorem . J. Polit. Philos . 9 ( 3 ), 277 - 306 ( 2001 ) Liu , B. : Sentiment Analysis and Opinion Mining . Morgan & Claypool, San Rafael ( 2012 ) Loughran , T. , McDonald , B. : When is a liability not a liability? Textual analysis , dictionaries and 10-Ks. J. Financ . 66 ( 1 ), 35 - 66 ( 2011 ) Lowe , W. , Benoit , K. : Validating estimates of latent traits from textual data using human judgment as a benchmark . Polit. Anal . 21 ( 3 ), 298 - 313 ( 2013 ) Macagno , F. , Walton , D.N. : Emotive Language in Argumentation . Cambridge University Press, New York ( 2014 ) Mohammad , S.M. : Sentiment analysis: detecting valence, emotions, and other affectual states from text . In: Meiselman, H. (ed.) Emotion Measurement, pp. 201 - 238 . Elsevier, Amsterdam ( 2016 ) Momtazi , S. : Fine-grained German sentiment analysis on social media . In: Proceedings of the 8th International Conference on Language Resources and Evaluation , Istanbul, Turkey, 23 -25 May, pp. 1215 - 1220 ( 2012 ) Monroe , B.L. , Colaresi , M.P. , Quinn , K.M.: Fightin' words: lexical feature selection and evaluation for identifying the content of political conflict . Polit. Anal . 16 ( 4 ), 372 - 403 ( 2008 ) M u¨ller , W.C., Dolezal , M. , Ennser-Jedenastik , L. , Winkler , A.K. , Eder , N. , Jenny , M., Meyer, T.M., Schermann , K. : AUTNES Inhaltsanalyse der Presseaussendungen (OTS) von Parteien bei der o¨sterreichischen Nationalratswahl 2013-Dokumentation . Universita¨t Wien, Wien ( 2014 ) Mutz , D.C. , Reeves , B. : The new videomalaise: effects of televised incivility on political trust . Am. Polit. Sci. Rev . 99 ( 1 ), 1 - 15 ( 2005 ) Nai , A. , Walter , A.S. : New perspectives on negative campaigning. Why attack politics matter . ECPR Press, Colchester ( 2015 ) Osgood , C.E. , Suci , G.J. , Tannenbaum , P.H. : The Measurement of Meaning . University of Illinois Press, Urbana ( 1957 ) Pang , B. , Lee , L. : Opinion mining and sentiment analysis . Found. Trends Inf. Retr . 2 ( 1 ), 1 - 90 ( 2008 ) Pang , B. , Lee , L. , Vaithyanathan , S. : Thumbs up? Sentiment classification using machine learning techniques . In: Proceedings of the Conference on Empirical Methods in Natural Language Processing , Philadelphia, Pennsylvania, 6 - 7 July , pp. 79 - 86 ( 2002 ) Patterson , T.E. : Out of Order. Alfred A. Knopf , New York ( 1993 ) Peeters , G. : The positive-negative asymmetry: on cognitive consistency and positivity bias . Eur. J. Soc. Psychol . 1 ( 4 ), 455 - 474 ( 1971 ) Peeters , G. , Czapinski , J.: Positive-negative asymmetry in evaluations: the distinction between affective and informational negativity effects . Eur. Rev. Soc. Psychol . 1 ( 1 ), 33 - 60 ( 1990 ) Porter , M.F. : An algorithm for suffix stripping . Program 14 ( 3 ), 130 - 137 ( 1980 ) Quasthoff , U. , Richter , M. , Biemann , C. : Corpus portal for search in monolingual corpora . In: Proceedings of the Fifth International Conference on Language Resources and Evaluation , LREC 2006 , Genoa, pp. 1799 - 1802 ( 2006 ) Remus , R. , Quasthoff , U. , Heyer , G.: SentiWS-a publicly available German-language resource for sentiment analysis . In: Proceedings of the 7th International Language Ressources and Evaluation , Valletta, Malta, 17 -23 May ( 2010 ) Rozin , P. , Royzman , E.B. : Negativity bias, negativity dominance, and contagion . Personal. Soc. Psychol. Rev . 5 ( 4 ), 296 - 320 ( 2001 ) Saal , F.E. , Downey , R.G. , Lahey , M.A. : Rating the ratings: assessing the psychometric quality of rating data . Psychol. Bull . 88 ( 2 ), 413 - 428 ( 1980 ) Schmidt , H. : Probabilistic part-of-speech tagging using decision trees . In: International Conference on New Methods in Language Processing , 14 - 16 September, Manchester, UK ( 1994 ) Slapin , J.B. , Proksch , S.-O.: A scaling model for estimating time-series party positions from texts . Am. J. Polit. Sci . 52 ( 3 ), 705 - 722 ( 2008 ) Snow , R. , O 'Connor , B. , Jurafsky , D. , Ng , A.Y. : Cheap and fast-but is it good? Evaluating non-expert annotations for natural language tasks . In: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2008 ), Honolulu, Hawaii, 25 - 27 October, pp. 254 - 263 ( 2008 ) Soroka , S. , Young , L. , Balmas , M. : Bad news or mad news? Sentiment scoring of negativity, fear, and anger in news content . Ann. Am. Acad. Polit. Soc. Sci . 659 ( 1 ), 108 - 121 ( 2015a ) Soroka , S.N.: The gatekeeping function: distributions of information in media and the real world . J. Polit . 74 ( 02 ), 514 - 528 ( 2012 ) Soroka , S.N.: Negativity in Democratic Politics: Causes and Consequences . Cambridge University Press, New York ( 2014 ) Soroka , S.N. , Stecula , D.A. , Wlezien , C. : It's (Change in) the (future) economy, stupid: economic indicators, the media, and public opinion . Am. J. Polit. Sci . 59 ( 2 ), 457 - 474 ( 2015b ) Steinberger , R. , Pouliquen , B. , Kabadjov , M. , Belyaeva , J., van der Goot , E.: JRC-names: a freely available, highly multilingual named entity resource . In: Proceedings of the 8th International Conference Recent Advances in Natural Language Processing (RANLP) , Hissar, Bulgaria, 12 -14 September ( 2011 ) Stone , P.J. , Kirsch , J.: Cambridge Computer Associates. The General Inquirer: A Computer Approach to Content Analysis . M.I.T . Press, Cambridge ( 1966 ) Strapparava , C. , Mihalcea , R. : Learning to identify emotions in text . In: Proceedings of the 23rd Annual ACM Symposium on Applied Computing , Fortaleza, Brazil, 16 - 20 March, pp. 1556 - 1560 ( 2008 ) Stro¨mba¨ck , J., Kaid , L.L.: A framework for comparing election news coverage around the world . In: Stro¨mba¨ck, J., Kaid , L.L . (eds.) The Handbook of Election Coverage Around the World , pp. 1 - 18 . Routledge, New York ( 2008 ) Taboada , M. : Sentiment analysis: an overview from linguistics . Annu. Rev. Linguist . 2 , 325 - 347 ( 2016 ) Taboada , M. , Brooke , J. , Tofiloski , M. , Voll , K. , Stede , M. : Lexicon based methods for sentiment analysis . Comput. Linguist . 37 ( 2 ), 267 - 307 ( 2011 ) Thelwall , M. , Buckley , K. : Topic based sentiment analysis for the social web: the role of mood and issuerelated words . J. Am. Soc. Inf. Sci. Technol . 64 ( 8 ), 1608 - 1617 ( 2013 ) Thelwall , M. , Buckley , K. , Paltoglou , G.: Sentiment strength detection for the social web . J. Am. Soc. Inf. Sci. Technol . 63 ( 1 ), 163 - 173 ( 2012 ) Thelwall , M. , Buckley , K. , Paltoglou , G. , Cai , D. , Kappas , A. : Sentiment strength detection in short informal text . J. Am. Soc. Inf. Sci. Technol . 61 ( 12 ), 2544 - 2558 ( 2010 ) Van Atteveldt , W. , Kleinnijenhuis , J. , Ruigrok , N. , Schlobach , S. : Good news or bad news? Conducting sentiment analysis on Dutch text to distinguish between positive and negative relations . J. Inf. Technol. Polit . 5 ( 1 ), 73 - 94 ( 2008 ) Walter , A.S.: Choosing the enemy: attack behaviour in a multiparty system . Party Polit . 20 ( 3 ), 311 - 323 ( 2014 ) Walter , A.S. , Van der Brug , W.: When the Gloves come off: inter-party variation in negative campaigning in Dutch elections, 1981 - 2010 . Acta Polit. 48 , 367 - 388 ( 2013 ) Waltinger , U. : GermanPolarityClues: a lexical resource for German sentiment analysis . In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC) , 19 -21 May, Valletta, Malta ( 2010 ) Wilson , T. , Wiebe , J. , Hoffman , P. : Recognizing contextual polarity in phrase-level sentiment analysis . In: Proceedings of the Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing (HLT/EMNLP) , Vancouver, Canada, 6 -8 October, pp. 347 - 354 . Association for Computational Linguistics ( 2005 ) Wolf , M. , Horn , A.B. , Mehl , M.R. , Haug , S. , Pennebaker , J.W. , Kordy , H.: Computergest u¨tzte quantitative textanalyse . Diagnostica 54 ( 2 ), 85 - 98 ( 2008 ) Young , L. , Soroka , S. : Affective news: the automated coding of sentiment in political texts . Polit. Commun . 29 ( 2 ), 205 - 231 ( 2012 ) Zhou , D. , Liu , Q. , Platt , J.C. , Meek , C. : Aggregating ordinal labels from crowds by minimax conditional entropy . In: Proceedings of the 31st International Conference on Machine Learning , Beijing, China, 21 - 24 June ( 2014 )


This is a preview of a remote PDF: https://link.springer.com/content/pdf/10.1007%2Fs11135-016-0412-4.pdf

Martin Haselmayer, Marcelo Jenny. Sentiment analysis of political communication: combining a dictionary approach with crowdcoding, Quality & Quantity, 2016, 1-24, DOI: 10.1007/s11135-016-0412-4