File size: 71,128 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
{
    "paper_id": "W10-0204",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T05:00:27.980198Z"
    },
    "title": "Emotions Evoked by Common Words and Phrases: Using Mechanical Turk to Create an Emotion Lexicon",
    "authors": [
        {
            "first": "Saif",
            "middle": [
                "M"
            ],
            "last": "Mohammad",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "National Research Council Canada",
                "location": {
                    "postCode": "K1A 0R6",
                    "settlement": "Ottawa",
                    "region": "Ontario",
                    "country": "Canada"
                }
            },
            "email": "[email protected]"
        },
        {
            "first": "Peter",
            "middle": [
                "D"
            ],
            "last": "Turney",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "National Research Council Canada",
                "location": {
                    "postCode": "K1A 0R6",
                    "settlement": "Ottawa",
                    "region": "Ontario",
                    "country": "Canada"
                }
            },
            "email": "[email protected]"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Even though considerable attention has been given to semantic orientation of words and the creation of large polarity lexicons, research in emotion analysis has had to rely on limited and small emotion lexicons. In this paper, we show how we create a high-quality, moderate-sized emotion lexicon using Mechanical Turk. In addition to questions about emotions evoked by terms, we show how the inclusion of a word choice question can discourage malicious data entry, help identify instances where the annotator may not be familiar with the target term (allowing us to reject such annotations), and help obtain annotations at sense level (rather than at word level). We perform an extensive analysis of the annotations to better understand the distribution of emotions evoked by terms of different parts of speech. We identify which emotions tend to be evoked simultaneously by the same term and show that certain emotions indeed go hand in hand.",
    "pdf_parse": {
        "paper_id": "W10-0204",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Even though considerable attention has been given to semantic orientation of words and the creation of large polarity lexicons, research in emotion analysis has had to rely on limited and small emotion lexicons. In this paper, we show how we create a high-quality, moderate-sized emotion lexicon using Mechanical Turk. In addition to questions about emotions evoked by terms, we show how the inclusion of a word choice question can discourage malicious data entry, help identify instances where the annotator may not be familiar with the target term (allowing us to reject such annotations), and help obtain annotations at sense level (rather than at word level). We perform an extensive analysis of the annotations to better understand the distribution of emotions evoked by terms of different parts of speech. We identify which emotions tend to be evoked simultaneously by the same term and show that certain emotions indeed go hand in hand.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "When analyzing text, automatically detecting emotions such as joy, sadness, fear, anger, and surprise is useful for a number of purposes, including identifying blogs that express specific emotions towards the topic of interest, identifying what emotion a newspaper headline is trying to evoke, and devising automatic dialogue systems that respond appropriately to different emotional states of the user. Often different emotions are expressed through different words. For example, delightful and yummy indicate the emotion of joy, gloomy and cry are indicative of sadness, shout and boiling are indicative of anger, and so on. Therefore an emotion lexicon-a list of emotions and words that are indicative of each emotion-is likely to be useful in identifying emotions in text.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Words may evoke different emotions in different contexts, and the emotion evoked by a phrase or a sentence is not simply the sum of emotions conveyed by the words in it, but the emotion lexicon will be a useful component for any sophisticated emotion detecting algorithm. The lexicon will also be useful for evaluating automatic methods that identify the emotions evoked by a word. Such algorithms may then be used to automatically generate emotion lexicons in languages where no such lexicons exist. As of now, high-quality high-coverage emotion lexicons do not exist for any language, although there are a few limited-coverage lexicons for a handful of languages, for example, the WordNet Affect Lexicon (WAL) (Strapparava and Valitutti, 2004) for six basic emotions and the General Inquirer (GI) (Stone et al., 1966) , which categorizes words into a number of categories, including positive and negative semantic orientation.",
                "cite_spans": [
                    {
                        "start": 712,
                        "end": 745,
                        "text": "(Strapparava and Valitutti, 2004)",
                        "ref_id": null
                    },
                    {
                        "start": 799,
                        "end": 819,
                        "text": "(Stone et al., 1966)",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Amazon has an online service called Mechanical Turk that can be used to obtain a large amount of human annotation in an efficient and inexpensive manner (Snow et al., 2008; Callison-Burch, 2009) . 1 However, one must define the task carefully to obtain annotations of high quality. Several checks must be placed to ensure that random and erroneous annotations are discouraged, rejected, and re-annotated.",
                "cite_spans": [
                    {
                        "start": 153,
                        "end": 172,
                        "text": "(Snow et al., 2008;",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 173,
                        "end": 194,
                        "text": "Callison-Burch, 2009)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "In this paper, we show how we compiled a moderate-sized English emotion lexicon by manual annotation through Amazon's Mechanical Turk service. This dataset, which we will call EmoLex, is many times as large as the only other known emotion lexicon, WordNet Affect Lexicon. More importantly, the terms in this lexicon are carefully chosen to include some of the most frequent nouns, verbs, adjectives, and adverbs. Beyond unigrams, it has a large number of commonly used bigrams. We also include some words from the General Inquirer and some from WordNet Affect Lexicon, to allow comparison of annotations between the various resources.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "We perform an extensive analysis of the annotations to answer several questions that have not been properly addressed so far. For instance, how hard is it for humans to annotate words with the emotions they evoke? What percentage of commonly used terms, in each part of speech, evoke an emotion? Are emotions more commonly evoked by nouns, verbs, adjectives, or adverbs? Is there a correlation between the semantic orientation of a word and the emotion it evokes? Which emotions tend to go together; that is, which emotions are evoked simultaneously by the same term? This work is intended to be a pilot study before we create a much larger emotion lexicon with tens of thousands of terms.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "We focus on the emotions of joy, sadness, anger, fear, trust, disgust, surprise, and anticipationargued by many to be the basic and prototypical emotions (Plutchik, 1980) . Complex emotions can be viewed as combinations of these basic emotions.",
                "cite_spans": [
                    {
                        "start": 154,
                        "end": 170,
                        "text": "(Plutchik, 1980)",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "WordNet Affect Lexicon (Strapparava and Valitutti, 2004) has a few hundred words annotated with the emotions they evoke. 2 It was created by manually identifying the emotions of a few seed words and then marking all their WordNet synonyms as having the same emotion. The General Inquirer (Stone et al., 1966) has 11,788 words labeled with 182 categories of word tags, including positive and negative semantic orientation. 3 It also has certain other affect categories, such as pleasure, arousal, feeling, and pain but these have not been exploited to a significant degree by the natural language processing community.",
                "cite_spans": [
                    {
                        "start": 23,
                        "end": 56,
                        "text": "(Strapparava and Valitutti, 2004)",
                        "ref_id": null
                    },
                    {
                        "start": 288,
                        "end": 308,
                        "text": "(Stone et al., 1966)",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related work",
                "sec_num": "2"
            },
            {
                "text": "Work in emotion detection can be roughly classified into that which looks for specific emotion denoting words (Elliott, 1992) , that which determines tendency of terms to co-occur with seed words whose emotions are known (Read, 2004) , that which uses hand-coded rules (Neviarouskaya et al., 2009) , and that which uses machine learning and a number of emotion features, including emotion denoting words (Alm et al., 2005) .",
                "cite_spans": [
                    {
                        "start": 110,
                        "end": 125,
                        "text": "(Elliott, 1992)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 221,
                        "end": 233,
                        "text": "(Read, 2004)",
                        "ref_id": "BIBREF12"
                    },
                    {
                        "start": 269,
                        "end": 297,
                        "text": "(Neviarouskaya et al., 2009)",
                        "ref_id": "BIBREF10"
                    },
                    {
                        "start": 404,
                        "end": 422,
                        "text": "(Alm et al., 2005)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related work",
                "sec_num": "2"
            },
            {
                "text": "Much of this recent work focuses on six emotions studied by Ekman (1992) . These emotionsjoy, sadness, anger, fear, disgust, and surprise-are a subset of the eight proposed in Plutchik (1980) . We focus on the Plutchik emotions because the emotions can be naturally paired into opposites-joysadness, anger-fear, trust-disgust, and anticipationsurprise. Natural symmetry apart, we believe that prior work on automatically computing word-pair antonymy (Lin et al., 2003; Mohammad et al., 2008; Lobanova et al., 2010) can now be leveraged in automatic emotion detection.",
                "cite_spans": [
                    {
                        "start": 60,
                        "end": 72,
                        "text": "Ekman (1992)",
                        "ref_id": "BIBREF4"
                    },
                    {
                        "start": 176,
                        "end": 191,
                        "text": "Plutchik (1980)",
                        "ref_id": "BIBREF11"
                    },
                    {
                        "start": 450,
                        "end": 468,
                        "text": "(Lin et al., 2003;",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 469,
                        "end": 491,
                        "text": "Mohammad et al., 2008;",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 492,
                        "end": 514,
                        "text": "Lobanova et al., 2010)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Related work",
                "sec_num": "2"
            },
            {
                "text": "In the subsections below we present the challenges in obtaining high-quality emotion annotation, how we address those challenges, how we select the target terms, and the questionnaire we created for the annotators.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Emotion annotation",
                "sec_num": "3"
            },
            {
                "text": "Words used in different senses can evoke different emotions. For example, the word shout evokes a different emotion when used in the context of admonishment, than when used in \"Give me a shout if you need any help.\" Getting human annotations on word senses is made complicated by decisions about which sense-inventory to use and what level of granularity the senses must have. On the one hand, we do not want to choose a fine-grained sense-inventory because then the number of word-sense combinations will become too large and difficult to easily distinguish, and on the other hand we do not want to work only at the word level because when used in different senses a word may evoke different emotions.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Key challenges",
                "sec_num": "3.1"
            },
            {
                "text": "Yet another challenge is how best to convey a word sense to the annotator. Long definitions will take time to read and limit the number of annotations we can obtain for the same amount of resources. Further, we do not want to bias the annotator towards an emotion through the definition. We want the users to annotate a word only if they are already familiar with it and know its meanings. And lastly, we must ensure that malicious and erroneous annotations are rejected.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Key challenges",
                "sec_num": "3.1"
            },
            {
                "text": "In order to overcome the challenges described above, before asking the annotators questions about what emotions are evoked by a target term, we first present them with a word choice problem pertaining to the target. They are provided with four different words and asked which word is closest in meaning to the target. This single question serves many purposes. Through this question we convey the word sense for which annotations are to be provided, without actually providing annotators with long definitions. If an annotator is not familiar with the target word and still attempts to answer questions pertaining to the target, or is randomly clicking options in our questionnaire, then there is a 75% chance that they will get the answer to this question wrong, and we can discard all responses pertaining to this target term by the annotator (that is, we discard answers to the emotion questions provided by the annotator for this target term). We generated these word choice problems automatically using the Macquarie Thesaurus (Bernard, 1986) . Published thesauri, such as Roget's and Macquarie, divide the vocabulary into about a thousand categories, which may be interpreted as coarse senses. If a word has more than one sense, then it can be found in more than one thesaurus category. Each category also has a head word which best captures the meaning of the category.",
                "cite_spans": [
                    {
                        "start": 1032,
                        "end": 1047,
                        "text": "(Bernard, 1986)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Our solution",
                "sec_num": "3.2"
            },
            {
                "text": "Most of the target terms chosen for annotation are restricted to those that are listed in exactly one thesaurus category. The word choice question for a target term is automatically generated by selecting the following four alternatives (choices): the head word of the thesaurus category pertaining to the target term (the correct answer); and three other head words of randomly selected categories (the distractors). The four alternatives are presented to the an-notator in random order.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Our solution",
                "sec_num": "3.2"
            },
            {
                "text": "Only a small number of the words in the WordNet Affect Lexicon are listed in exactly one thesaurus category (have one sense), and so we included target terms that occurred in two thesaurus categories as well. For these questions, we listed head words from both the senses (categories) as two of the alternatives (probability of a random choice being correct is 50%). Depending on the alternative chosen, we can thus determine the sense for which the subsequent emotion responses are provided by the annotator.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Our solution",
                "sec_num": "3.2"
            },
            {
                "text": "In order to generate an emotion lexicon, we first identify a list of words and phrases for which we want human annotations. We chose the Macquarie Thesaurus as our source pool for unigrams and bigrams. Any other published dictionary would have worked well too. However, apart from over 57,000 commonly used English word types, the Macquarie Thesaurus also has entries for more than 40,000 commonly used phrases. From this list of unigrams and bigrams we chose those that occur frequently in the Google n-gram corpus (Brants and Franz, 2006 ). Specifically we chose the 200 most frequent n-grams in the following categories: noun unigrams, noun bigrams, verb unigrams, verb bigrams, adverb unigrams, adverb bigrams, adjective unigrams, adjective bigrams, words in the General Inquirer that are marked as having a negative semantic orientation, words in General Inquirer that are marked as having a positive semantic orientation. When selecting these sets, we ignored terms that occurred in more than one Macquarie Thesaurus category. Lastly, we chose all words from each of the six emotion categories in the WordNet Affect Lexicon that had at most two senses in the thesaurus (occurred in at most two thesaurus categories). The first and second column of Table 1 list the various sets of target terms as well as the number of terms in each set for which annotations were requested. EmoLex Uni stands for all the unigrams taken from the thesaurus. EmoLex Bi refers to all the bigrams. EmoLex GI are all the words taken from the General Inquirer. EmoLex WAL are all the words taken from the Word-Net Affect Lexicon.",
                "cite_spans": [
                    {
                        "start": 516,
                        "end": 539,
                        "text": "(Brants and Franz, 2006",
                        "ref_id": "BIBREF2"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 1254,
                        "end": 1261,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Target terms",
                "sec_num": "3.3"
            },
            {
                "text": "An entity submitting a task to Mechanical Turk is called the requester. A requester first breaks the task into small independently solvable units called HITs (Human Intelligence Tasks) and uploads them on the Mechanical Turk website. The requester specifies the compensation that will be paid for solving each HIT. The people who provide responses to these HITs are called Turkers. The requester also specifies the number of different Turkers that are to annotate each HIT. The annotation provided by a Turker for a HIT is called an assignment.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Mechanical Turk HITs",
                "sec_num": "3.4"
            },
            {
                "text": "We created Mechanical Turk HITs for each of the terms specified in Table 1 . Each HIT has a set of questions, all of which are to be answered by the same person. We requested five different assignments for each HIT (each HIT is to be annotated by five different Turkers). Different HITS may be attempted by different Turkers, and a Turker may attempt as many HITs as they wish. Below is an example HIT for the target word \"startle\". Before going live, the survey was approved by the ethics committee at the National Research Council Canada.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 67,
                        "end": 74,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Mechanical Turk HITs",
                "sec_num": "3.4"
            },
            {
                "text": "The first set of emotion annotations on Mechanical Turk were completed in about nine days. The Turkers spent a minute on average to answer the questions in a HIT. This resulted in an hourly pay of slightly more than $2.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Annotation analysis",
                "sec_num": "4"
            },
            {
                "text": "Once the assignments were collected, we used automatic scripts to validate the annotations. Some assignments were discarded because they failed certain tests (described below). A subset of the discarded assignments were officially rejected (the Turkers were not paid for these assignments) because instructions were not followed. About 500 of the 10,880 assignments (2,176 \u00d7 5) included at least one unanswered question. These assignments were discarded and rejected. More than 85% of the remaining assignments had the correct answer for the word choice question. This was a welcome result showing that, largely, the annotations were done in a responsible manner. We discarded all assignments that had the wrong answer for the word choice question. If an annotator obtained an overall score that is less than 66.67% on the word choice questions (that is, got more than one out of three wrong), then we assumed that, contrary to instructions, HITs for words not familiar to the annotator were attempted. We discarded and rejected all assignments by such annotators (not just the assignments for which they got the word choice question wrong).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Annotation analysis",
                "sec_num": "4"
            },
            {
                "text": "HITs pertaining to all the discarded assignments were uploaded for a second time on Mechanical Turk and the validation process was repeated. After the second round, we had three or more valid assignments for 2081 of the 2176 target terms. We will refer to this set of assignments as the master set. We create the emotion lexicon from this master set containing 9892 assignments from about 1000 Turkers who attempted 1 to 450 assignments each. About 100 of them provided 20 or more assignments each (more than 7000 assignments in all). The master set has, on average, about 4.75 assignments for each of the 2081 target terms. (See Table 1 for more details.)",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 630,
                        "end": 637,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Annotation analysis",
                "sec_num": "4"
            },
            {
                "text": "The different emotion annotations for a target term were consolidated by determining the majority class of emotion intensities. For a given termemotion pair, the majority class is that intensity level that is chosen most often by the Turkers to represent the degree of emotion evoked by the word. Ties are broken by choosing the stronger intensity level. Table 2 lists the percent of 2081 target terms assigned a majority class of no, weak, moderate, and strong emotion. For example, it tells us that 7.6% of the tar- get terms strongly evoke joy. The table also presents an average of the numbers in each column (micro average). Observe that the percentages for individual emotions do not vary greatly from the average. The last row lists the percent of target terms that evoke some emotion (any of the eight) at the various intensity levels. We calculated this using the intensity level of the strongest emotion expressed by each target. Observe that 30.1% of the target terms strongly evoke at least one of the eight basic emotions. Even though we asked Turkers to annotate emotions at four levels of intensity, practical NLP applications often require only two levels-evoking particular emotion (evocative) or not (non-evocative). For each target term-emotion pair, we convert the four-level annotations into two-level annotations by placing all no-and weak-intensity assignments in the non-evocative bin, all moderate-and strongintensity assignments in the evocative bin, and then choosing the bin with the majority assignments. Table 3 gives percent of target terms considered to be EmoLex   anger anticipation disgust fear joy sadness surprise trust any  EmoLex Uni :  adjectives  12  21  8  11  30  13  10  19  72  adverbs  12  16  7  8  21  6  11  25  65  nouns  4  21  2  9  16  3  3  21  47  verbs  12  21  7  11  15  12 evocative. The last row in the table gives the percentage of terms evocative of some emotion (any of the eight). Table 4 shows how many terms in each category are evocative of the different emotions. Table 4 shows that a sizable percent of nouns, verbs, adjectives, and adverbs are evocative. Adverbs and adjectives are some of the most emotion inspiring terms and this is not surprising considering that they are used to qualify a noun or a verb. Anticipation, trust, and joy come through as the most common emotions evoked by terms of all four parts of speech. The EmoLex WAL rows are particularly interesting because they serve to determine how much the Turker annotations match annotations in the Wordnet Affect Lexicon (WAL). The most common Turker-determined emotion for each of these rows is marked in bold. Observe that WAL anger terms are mostly marked as anger evocative, joy terms as joy evocative, and so on. The EmoLex WAL rows also indicate which emotions get confused for which, or which emotions tend to be evoked simultaneously by a term. Observe that anger terms tend also to be evocative of disgust. Similarly, fear and sadness go together, as do joy, trust, and anticipation.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 1534,
                        "end": 1831,
                        "text": "Table 3 gives percent of target terms considered to be EmoLex   anger anticipation disgust fear joy sadness surprise trust any  EmoLex Uni :  adjectives  12  21  8  11  30  13  10  19  72  adverbs  12  16  7  8  21  6  11  25  65  nouns  4  21  2  9  16  3  3  21  47  verbs  12  21  7  11  15  12",
                        "ref_id": "TABREF2"
                    },
                    {
                        "start": 1945,
                        "end": 1952,
                        "text": "Table 4",
                        "ref_id": "TABREF5"
                    },
                    {
                        "start": 2032,
                        "end": 2039,
                        "text": "Table 4",
                        "ref_id": "TABREF5"
                    }
                ],
                "eq_spans": [],
                "section": "Emotions evoked by words",
                "sec_num": "4.1"
            },
            {
                "text": "The EmoLex GI rows rightly show that words marked as negative in the General Inquirer, mostly evoke negative emotions (anger, fear, disgust, and sadness). Observe that the percentages for trust and joy are much lower. On the other hand, positive words evoke anticipation, joy, and trust.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Analysis and discussion",
                "sec_num": "4.1.1"
            },
            {
                "text": "In order to analyze how often the annotators agreed with each other, for each term-emotion pair, we calculated the percentage of times the majority class has size 5 (all Turkers agree), size 4 (all but one agree), size 3, and size 2. Observe that for more than 50% of the terms, at least four annotators agree with each other. Table 5 presents these agreement values. Since many NLP systems may rely only on two intensity values (evocative or non-evocative), we also calculate agreement at that level ( Table 6 ). Observe that for more than 50% of the terms, all five annotators agree with each other, and for more than 80% of the terms, at least four annotators agree. This shows a high degree of agreement on emotion annotations despite no real control over the educational background and qualifications of the annotators. Table 6 : Agreement at two intensity levels for emotion (evocative and non-evocative): Percent of 2081 terms for which the majority class size was 3, 4, and 5.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 327,
                        "end": 334,
                        "text": "Table 5",
                        "ref_id": "TABREF6"
                    },
                    {
                        "start": 503,
                        "end": 510,
                        "text": "Table 6",
                        "ref_id": null
                    },
                    {
                        "start": 825,
                        "end": 832,
                        "text": "Table 6",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Agreement",
                "sec_num": "4.1.2"
            },
            {
                "text": "We consolidate the semantic orientation (polarity) annotations in a manner identical to the process for emotion annotations. Table 7 lists the percent of 2081 target terms assigned a majority class of no, weak, moderate, and strong semantic orientation. For example, it tells us that 16% of the target terms are strongly negative. The last row in the table lists the percent of target terms that have some semantic orientation (positive or negative) at the various intensity levels. Observe that 35% of the target terms are strongly evaluative (positively or negatively).",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 125,
                        "end": 132,
                        "text": "Table 7",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Semantic orientation of words",
                "sec_num": "4.2"
            },
            {
                "text": "Just as in the case for emotions, practical NLP applications often require only two levels of semantic orientation-having particular semantic orientation or not (evaluative) or not (non-evaluative). For each target term-emotion pair, we convert the fourlevel semantic orientation annotations into two-level ones, just as we did for the emotions. Table 9 shows how many terms in each category are positively and negatively evaluative.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 346,
                        "end": 353,
                        "text": "Table 9",
                        "ref_id": "TABREF9"
                    }
                ],
                "eq_spans": [],
                "section": "Semantic orientation of words",
                "sec_num": "4.2"
            },
            {
                "text": "Observe in Table 9 that, across the board, a sizable number of terms are evaluative with respect to some semantic orientation. Interestingly unigram nouns have a markedly lower proportion of negative terms, and a much higher proportion of positive terms. It may be argued that the default semantic orientation of noun concepts is positive, and that usually it takes a negative adjective to make the phrase negative. The EmoLex GI rows in the two tables show that words marked as having a negative semantic orientation in the General Inquirer are mostly marked as negative by the Turkers. And similarly, the positives in GI are annotated as positive. Again, this is confirmation that the quality of annotation obtained is high. The EmoLex WAL rows show that anger, disgust, fear, and sadness terms tend not to have a positive semantic orientation and are mostly negative. In contrast, and expectedly, the joy terms are positive. The surprise terms are more than twice as likely to be positive than negative.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 11,
                        "end": 18,
                        "text": "Table 9",
                        "ref_id": "TABREF9"
                    }
                ],
                "eq_spans": [],
                "section": "Analysis and discussion",
                "sec_num": "4.2.1"
            },
            {
                "text": "In order to analyze how often the annotators agreed with each other, for each term-emotion pair, we cal- culated the percentage of times the majority class has size 5 (all Turkers agree), size 4 (all but one agree), size 3, and size 2. Table 10 presents these agreement values. Observe that for more than 50% of the terms, at least four annotators agree with each other. Table 11 gives agreement values at the twointensity level. Observe that for more than 50% of the terms, all five annotators agree with each other, and for more than 80% of the terms, at least four annotators agree.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 236,
                        "end": 244,
                        "text": "Table 10",
                        "ref_id": null
                    },
                    {
                        "start": 371,
                        "end": 379,
                        "text": "Table 11",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Agreement",
                "sec_num": "4.2.2"
            },
            {
                "text": "We showed how Mechanical Turk can be used to create a high-quality, moderate-sized, emotion lexicon for a very small cost (less than US$500). Notably, we used automatically generated word choice questions to detect and reject erroneous annotations and to reject all annotations by unqualified Turkers and those who indulge in malicious data entry. We compared a subset of our lexicon with existing gold standard data to show that the annotations obtained are indeed of high quality. A detailed analysis of the Table 10 : Agreement at four intensity levels for polarity (no, weak, moderate, and strong): Percent of 2081 terms for which the majority class size was 2, 3, 4, and 5.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 510,
                        "end": 518,
                        "text": "Table 10",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Conclusions",
                "sec_num": "5"
            },
            {
                "text": "three four five negative 11.8 21.2 66.9 positive 23.1 26.3 50.5 micro average 17.5 23.8 58.7 Table 11 : Agreement at two intensity levels for polarity (evaluative and non-evaluative): Percent of 2081 terms for which the majority class size was 3, 4, and 5. lexicon revealed insights into how prevalent emotion bearing terms are among common unigrams and bigrams. We also identified which emotions tend to be evoked simultaneously by the same term. The lexicon is available for free download. 4 Since this pilot experiment with about 2000 target terms was successful, we will now obtain emotion annotations for tens of thousands of English terms. We will use the emotion lexicon to identify emotional tone of larger units of text, such as newspaper headlines and blog posts. We will also use it to evaluate automatically generated lexicons, such as the polarity lexicons by Turney and Littman (2003) and Mohammad et al. (2009) . We will explore the variance in emotion evoked by near-synonyms, and also how common it is for words with many meanings to evoke different emotions in different senses.",
                "cite_spans": [
                    {
                        "start": 492,
                        "end": 493,
                        "text": "4",
                        "ref_id": null
                    },
                    {
                        "start": 873,
                        "end": 898,
                        "text": "Turney and Littman (2003)",
                        "ref_id": "BIBREF17"
                    },
                    {
                        "start": 903,
                        "end": 925,
                        "text": "Mohammad et al. (2009)",
                        "ref_id": "BIBREF9"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 93,
                        "end": 101,
                        "text": "Table 11",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Majority class size Polarity",
                "sec_num": null
            },
            {
                "text": "https://www.mturk.com/mturk/welcome",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://wndomains.fbk.eu/wnaffect.html 3 http://www.wjh.harvard.edu/\u223cinquirer",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://www.purl.org/net/emolex",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [
            {
                "text": "This research was funded by the National research Council Canada (NRC). Thanks to Diana Inkpen and Diman Ghazi for early discussions on emotion. Thanks to Joel Martin for encouragement and support. Thanks to Norm Vinson and the Ethics Committee at NRC for examining, guiding, and approving the survey. And last but not least, thanks to the more than 1000 anonymous people who answered the emotion survey with diligence and care.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Acknowledgments",
                "sec_num": null
            }
        ],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Emotions from text: Machine learning for text-based emotion prediction",
                "authors": [
                    {
                        "first": "Cecilia",
                        "middle": [],
                        "last": "Ovesdotter Alm",
                        "suffix": ""
                    },
                    {
                        "first": "Dan",
                        "middle": [],
                        "last": "Roth",
                        "suffix": ""
                    },
                    {
                        "first": "Richard",
                        "middle": [],
                        "last": "Sproat",
                        "suffix": ""
                    }
                ],
                "year": 2005,
                "venue": "Proceedings of the Joint Conference on Human Language Technology / Empirical Methods in Natural Language Processing (HLT/EMNLP-2005)",
                "volume": "",
                "issue": "",
                "pages": "579--586",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Cecilia Ovesdotter Alm, Dan Roth, and Richard Sproat. 2005. Emotions from text: Machine learning for text-based emotion prediction. In Proceedings of the Joint Conference on Human Language Technology / Empirical Methods in Natural Language Process- ing (HLT/EMNLP-2005), pages 579-586, Vancouver, Canada.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "The Macquarie Thesaurus",
                "authors": [],
                "year": 1986,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J.R.L. Bernard, editor. 1986. The Macquarie Thesaurus. Macquarie Library, Sydney, Australia.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Web 1t 5-gram version 1. Linguistic Data Consortium",
                "authors": [
                    {
                        "first": "Thorsten",
                        "middle": [],
                        "last": "Brants",
                        "suffix": ""
                    },
                    {
                        "first": "Alex",
                        "middle": [],
                        "last": "Franz",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Thorsten Brants and Alex Franz. 2006. Web 1t 5-gram version 1. Linguistic Data Consortium.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Fast, cheap and creative: Evaluating translation quality using amazon's mechanical turk",
                "authors": [
                    {
                        "first": "Chris",
                        "middle": [],
                        "last": "Callison",
                        "suffix": ""
                    },
                    {
                        "first": "-",
                        "middle": [],
                        "last": "Burch",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
                "volume": "",
                "issue": "",
                "pages": "286--295",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Chris Callison-Burch. 2009. Fast, cheap and cre- ative: Evaluating translation quality using amazon's mechanical turk. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP-2009), pages 286-295, Singapore.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "An argument for basic emotions",
                "authors": [
                    {
                        "first": "Paul",
                        "middle": [],
                        "last": "Ekman",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "Cognition and Emotion",
                "volume": "6",
                "issue": "3",
                "pages": "169--200",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Paul Ekman. 1992. An argument for basic emotions. Cognition and Emotion, 6(3):169-200.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "The affective reasoner: A process model of emotions in a multi-agent system",
                "authors": [
                    {
                        "first": "Clark",
                        "middle": [
                            "Elliott"
                        ],
                        "last": "",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Clark Elliott. 1992. The affective reasoner: A process model of emotions in a multi-agent system. Ph.D. the- sis, Institute for the Learning Sciences, Northwestern University.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Identifying synonyms among distributionally similar words",
                "authors": [
                    {
                        "first": "Dekang",
                        "middle": [],
                        "last": "Lin",
                        "suffix": ""
                    },
                    {
                        "first": "Shaojun",
                        "middle": [],
                        "last": "Zhao",
                        "suffix": ""
                    },
                    {
                        "first": "Lijuan",
                        "middle": [],
                        "last": "Qin",
                        "suffix": ""
                    },
                    {
                        "first": "Ming",
                        "middle": [],
                        "last": "Zhou",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "Proceedings of the 18th International Joint Conference on Artificial Intelligence (IJCAI-03)",
                "volume": "",
                "issue": "",
                "pages": "1492--1493",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dekang Lin, Shaojun Zhao, Lijuan Qin, and Ming Zhou. 2003. Identifying synonyms among distributionally similar words. In Proceedings of the 18th Inter- national Joint Conference on Artificial Intelligence (IJCAI-03), pages 1492-1493, Acapulco, Mexico.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Defining antonymy: A corpus-based study of opposites by lexico-syntactic patterns",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Lobanova",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Van Der Kleij",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Spenader",
                        "suffix": ""
                    }
                ],
                "year": 2010,
                "venue": "International Journal of Lexicography",
                "volume": "23",
                "issue": "",
                "pages": "19--53",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "A. Lobanova, T. van der Kleij, and J. Spenader. 2010. Defining antonymy: A corpus-based study of oppo- sites by lexico-syntactic patterns. International Jour- nal of Lexicography (in press), 23:19-53.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Computing word-pair antonymy",
                "authors": [
                    {
                        "first": "Saif",
                        "middle": [],
                        "last": "Mohammad",
                        "suffix": ""
                    },
                    {
                        "first": "Bonnie",
                        "middle": [],
                        "last": "Dorr",
                        "suffix": ""
                    },
                    {
                        "first": "Codie",
                        "middle": [],
                        "last": "Dunn",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP-2008)",
                "volume": "",
                "issue": "",
                "pages": "982--991",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Saif Mohammad, Bonnie Dorr, and Codie Dunn. 2008. Computing word-pair antonymy. In Proceedings of the Conference on Empirical Methods in Natural Lan- guage Processing (EMNLP-2008), pages 982-991, Waikiki, Hawaii.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Generating high-coverage semantic orientation lexicons from overtly marked words and a thesaurus",
                "authors": [
                    {
                        "first": "Saif",
                        "middle": [],
                        "last": "Mohammad",
                        "suffix": ""
                    },
                    {
                        "first": "Cody",
                        "middle": [],
                        "last": "Dunne",
                        "suffix": ""
                    },
                    {
                        "first": "Bonnie",
                        "middle": [],
                        "last": "Dorr",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "Proceedings of Empirical Methods in Natural Language Processing",
                "volume": "",
                "issue": "",
                "pages": "599--608",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Saif Mohammad, Cody Dunne, and Bonnie Dorr. 2009. Generating high-coverage semantic orientation lexi- cons from overtly marked words and a thesaurus. In Proceedings of Empirical Methods in Natural Lan- guage Processing (EMNLP-2009), pages 599-608, Singapore.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Compositionality principle in recognition of fine-grained emotions from text",
                "authors": [
                    {
                        "first": "Alena",
                        "middle": [],
                        "last": "Neviarouskaya",
                        "suffix": ""
                    },
                    {
                        "first": "Helmut",
                        "middle": [],
                        "last": "Prendinger",
                        "suffix": ""
                    },
                    {
                        "first": "Mitsuru",
                        "middle": [],
                        "last": "Ishizuka",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "Proceedings of the Proceedings of the Third International Conference on Weblogs and Social Media (ICWSM-09)",
                "volume": "",
                "issue": "",
                "pages": "278--281",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Alena Neviarouskaya, Helmut Prendinger, and Mitsuru Ishizuka. 2009. Compositionality principle in recog- nition of fine-grained emotions from text. In Proceed- ings of the Proceedings of the Third International Con- ference on Weblogs and Social Media (ICWSM-09), pages 278-281, San Jose, California.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "A general psychoevolutionary theory of emotion. Emotion: Theory, research, and experience",
                "authors": [
                    {
                        "first": "",
                        "middle": [],
                        "last": "Plutchik",
                        "suffix": ""
                    }
                ],
                "year": 1980,
                "venue": "",
                "volume": "1",
                "issue": "",
                "pages": "3--33",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "R Plutchik. 1980. A general psychoevolutionary theory of emotion. Emotion: Theory, research, and experi- ence, 1(3):3-33.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Recognising affect in text using pointwise-mutual information",
                "authors": [
                    {
                        "first": "Jonathon",
                        "middle": [],
                        "last": "Read",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Jonathon Read. 2004. Recognising affect in text using pointwise-mutual information. Ph.D. thesis, Depart- ment of Informatics, University of Sussex.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Cheap and fast -but is it good? Evaluating nonexpert annotations for natural language tasks",
                "authors": [
                    {
                        "first": "Rion",
                        "middle": [],
                        "last": "Snow",
                        "suffix": ""
                    },
                    {
                        "first": "O'",
                        "middle": [],
                        "last": "Brendan",
                        "suffix": ""
                    },
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Connor",
                        "suffix": ""
                    },
                    {
                        "first": "Andrew",
                        "middle": [],
                        "last": "Jurafsky",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Ng",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP-2008)",
                "volume": "",
                "issue": "",
                "pages": "254--263",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Rion Snow, Brendan O'Connor, Daniel Jurafsky, and An- drew Ng. 2008. Cheap and fast -but is it good? Evalu- ating nonexpert annotations for natural language tasks. In Proceedings of the Conference on Empirical Meth- ods in Natural Language Processing (EMNLP-2008), pages 254-263, Waikiki, Hawaii.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "The General Inquirer: A Computer Approach to Content Analysis",
                "authors": [
                    {
                        "first": "Philip",
                        "middle": [],
                        "last": "Stone",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Dexter",
                        "suffix": ""
                    },
                    {
                        "first": "Marshall",
                        "middle": [
                            "S"
                        ],
                        "last": "Dunphy",
                        "suffix": ""
                    },
                    {
                        "first": "Daniel",
                        "middle": [
                            "M"
                        ],
                        "last": "Smith",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Ogilvie",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Associates",
                        "suffix": ""
                    }
                ],
                "year": 1966,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Philip Stone, Dexter C. Dunphy, Marshall S. Smith, Daniel M. Ogilvie, and associates. 1966. The General Inquirer: A Computer Approach to Content Analysis. The MIT Press.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "An affective extension of WordNet",
                "authors": [
                    {
                        "first": "",
                        "middle": [],
                        "last": "Wordnet-Affect",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC-2004)",
                "volume": "",
                "issue": "",
                "pages": "1083--1086",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Wordnet-Affect: An affective extension of WordNet. In Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC-2004), pages 1083-1086, Lisbon, Portugal.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Measuring praise and criticism: Inference of semantic orientation from association",
                "authors": [
                    {
                        "first": "Peter",
                        "middle": [],
                        "last": "Turney",
                        "suffix": ""
                    },
                    {
                        "first": "Michael",
                        "middle": [],
                        "last": "Littman",
                        "suffix": ""
                    }
                ],
                "year": 2003,
                "venue": "ACM Transactions on Information Systems (TOIS)",
                "volume": "21",
                "issue": "4",
                "pages": "315--346",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Peter Turney and Michael Littman. 2003. Measuring praise and criticism: Inference of semantic orientation from association. ACM Transactions on Information Systems (TOIS), 21(4):315-346.",
                "links": null
            }
        },
        "ref_entries": {
            "TABREF0": {
                "num": null,
                "type_str": "table",
                "text": "Return HIT if you are not familiar with the prompt word. How much does the word startle evoke or produce the emotion joy (for example, happy and fun may strongly evoke joy)?",
                "content": "<table><tr><td>Prompt word: startle 1. Which word is closest in meaning (most related) to startle? \u2022 automobile \u2022 shake \u2022 honesty \u2022 entertain 2. How positive (good, praising) is the word startle? \u2022 startle is not positive \u2022 startle is weakly positive \u2022 startle is moderately positive \u2022 startle is strongly positive 3. How negative (bad, criticizing) is the word startle? \u2022 startle is not negative \u2022 startle is weakly negative \u2022 startle is moderately negative \u2022 startle is strongly negative 4. # of terms EmoLex Initial Master per word Annotns. EmoLex Uni : adjectives 200 196 4.7 adverbs 200 192 4.7 nouns 200 187 4.6 verbs 200 197 4.7 EmoLex Bi : adjectives 200 182 4.7 adverbs 187 171 4.7 nouns 200 193 4.7 verbs 200 186 4.7 EmoLex GI : negatives in GI 200 196 4.7 positives in GI 200 194 4.8 EmoLex WAL : anger terms in WAL 107 84 4.8 disgust terms in WAL 25 25 4.8 fear terms in WAL 58 58 4.8 joy terms in WAL 109 92 4.8 sadness terms in WAL 86 73 4.7 surprise terms in WAL 39 38 4.7 Union 2176 2081 4.75</td></tr></table>",
                "html": null
            },
            "TABREF2": {
                "num": null,
                "type_str": "table",
                "text": "Percent of 2081 terms assigned a majority class of no, weak, moderate, and strong emotion.",
                "content": "<table><tr><td>Emotion</td><td>% of terms</td></tr><tr><td>anger</td><td>15.4</td></tr><tr><td>anticipation</td><td>20.9</td></tr><tr><td>disgust</td><td>11.0</td></tr><tr><td>fear</td><td>14.5</td></tr><tr><td>joy</td><td>21.9</td></tr><tr><td>sadness</td><td>14.4</td></tr><tr><td>surprise</td><td>9.8</td></tr><tr><td>trust</td><td>20.6</td></tr><tr><td>micro average</td><td>16.1</td></tr><tr><td>any emotion</td><td>67.9</td></tr></table>",
                "html": null
            },
            "TABREF3": {
                "num": null,
                "type_str": "table",
                "text": "Percent of 2081 target terms that are evocative.",
                "content": "<table/>",
                "html": null
            },
            "TABREF5": {
                "num": null,
                "type_str": "table",
                "text": "Percent of terms, in each target set, that are evocative. Highest individual emotion scores for EmoLex WAL are shown bold. Observe that WAL fear terms are marked most as fear evocative, joy terms as joy evocative, and so on.",
                "content": "<table/>",
                "html": null
            },
            "TABREF6": {
                "num": null,
                "type_str": "table",
                "text": "",
                "content": "<table><tr><td colspan=\"2\">: Agreement at four intensity levels for emotion</td></tr><tr><td colspan=\"2\">(no, weak, moderate, and strong): Percent of 2081 terms</td></tr><tr><td colspan=\"2\">for which the majority class size was 2, 3, 4, and 5.</td></tr><tr><td/><td>Majority class size</td></tr><tr><td>Emotion</td><td>three four five</td></tr><tr><td>anger</td><td>15.0 25.9 58.9</td></tr><tr><td>anticipation</td><td>32.3 33.7 33.8</td></tr><tr><td>disgust</td><td>12.8 24.6 62.4</td></tr><tr><td>fear</td><td>14.9 25.6 59.4</td></tr><tr><td>joy</td><td>18.4 27.0 54.5</td></tr><tr><td>sadness</td><td>13.6 22.0 64.2</td></tr><tr><td>surprise</td><td>17.5 31.4 50.9</td></tr><tr><td>trust</td><td>23.9 29.3 46.6</td></tr><tr><td colspan=\"2\">micro average 18.6 27.4 53.8</td></tr></table>",
                "html": null
            },
            "TABREF7": {
                "num": null,
                "type_str": "table",
                "text": "",
                "content": "<table><tr><td/><td/><td colspan=\"2\">Intensity</td><td/></tr><tr><td>Polarity</td><td colspan=\"4\">no weak moderate strong</td></tr><tr><td>negative</td><td>60.8</td><td>10.8</td><td>12.3</td><td>16.0</td></tr><tr><td>positive</td><td>48.3</td><td>11.7</td><td>20.7</td><td>19.0</td></tr><tr><td colspan=\"2\">micro average 54.6</td><td>11.3</td><td>16.5</td><td>17.5</td></tr><tr><td>any polarity</td><td>14.7</td><td>17.4</td><td>32.7</td><td>35.0</td></tr><tr><td colspan=\"5\">Table 7: Percent of 2081 terms assigned a majority class</td></tr><tr><td colspan=\"4\">of no, weak, moderate, and strong polarity.</td><td/></tr><tr><td colspan=\"2\">Polarity</td><td colspan=\"2\">% of terms</td><td/></tr><tr><td colspan=\"2\">negative</td><td/><td>31.3</td><td/></tr><tr><td colspan=\"2\">positive</td><td/><td>45.5</td><td/></tr><tr><td colspan=\"2\">micro average</td><td/><td>38.4</td><td/></tr><tr><td colspan=\"2\">any polarity</td><td/><td>76.1</td><td/></tr></table>",
                "html": null
            },
            "TABREF8": {
                "num": null,
                "type_str": "table",
                "text": "Percent of 2081 target terms that are evaluative. percent of target terms considered to be evaluative. The last row in the table gives the percentage of terms evaluative with respect to some semantic orientation (positive or negative).",
                "content": "<table/>",
                "html": null
            },
            "TABREF9": {
                "num": null,
                "type_str": "table",
                "text": "Percent of terms, in each target set, that are evaluative. The highest individual polarity EmoLex GI row scores are shown bold. Observe that the positive GI terms are marked mostly as positively evaluative and the negative terms are marked mostly as negatively evaluative.",
                "content": "<table/>",
                "html": null
            },
            "TABREF10": {
                "num": null,
                "type_str": "table",
                "text": "",
                "content": "<table><tr><td/><td>Majority class size</td></tr><tr><td>Polarity</td><td>two three four five</td></tr><tr><td>negative</td><td>11.8 28.7 29.4 29.8</td></tr><tr><td>positive</td><td>21.2 30.7 19.0 28.8</td></tr></table>",
                "html": null
            }
        }
    }
}