| { |
| "paper_id": "W16-0303", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T03:55:10.533566Z" |
| }, |
| "title": "Self-Reflective Sentiment Analysis", |
| "authors": [ |
| { |
| "first": "Benjamin", |
| "middle": [], |
| "last": "Shickel", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "shickelb@ufl.edu" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Heesacker", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "heesack@ufl.edu" |
| }, |
| { |
| "first": "Sherry", |
| "middle": [], |
| "last": "Benton", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "sherry.benton@taoconnect.org" |
| }, |
| { |
| "first": "Ashkan", |
| "middle": [], |
| "last": "Ebadi", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "ashkan.ebadi@ufl.edu" |
| }, |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Nickerson", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Parisa", |
| "middle": [], |
| "last": "Rashidi", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "parisa.rashidi@ufl.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "As self-directed online anxiety treatment and e-mental health programs become more prevalent and begin to rapidly scale to a large number of users, the need to develop automated techniques for monitoring patient progress and detecting early warning signs is at an alltime high. While current online therapy systems work based on explicit quantitative feedback from various survey measures, little attention has been paid thus far to the large amount of unstructured free text present in the monitoring logs and journals submitted by patients as part of the treatment process. In this paper, we automatically categorize patients' internal sentiment and emotions using machine learning classifiers based on n-grams, syntactic patterns, sentiment lexicon features, and distributed word embeddings. We report classification metrics on a novel mental health dataset.", |
| "pdf_parse": { |
| "paper_id": "W16-0303", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "As self-directed online anxiety treatment and e-mental health programs become more prevalent and begin to rapidly scale to a large number of users, the need to develop automated techniques for monitoring patient progress and detecting early warning signs is at an alltime high. While current online therapy systems work based on explicit quantitative feedback from various survey measures, little attention has been paid thus far to the large amount of unstructured free text present in the monitoring logs and journals submitted by patients as part of the treatment process. In this paper, we automatically categorize patients' internal sentiment and emotions using machine learning classifiers based on n-grams, syntactic patterns, sentiment lexicon features, and distributed word embeddings. We report classification metrics on a novel mental health dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "As mental health awareness becomes more widespread, especially among at-risk populations such as young adults and college-aged students, many institutions and universities are beginning to offer online anxiety and depression treatment programs to supplement traditional therapy services. A key component of these largely self-directed programs is the regular completion of journals, in which patients describe how they are feeling. These journals contain a wide variety of information, including a patient's specific fears, worries, triggers, reactions, or simply status updates on their emotional state. At current time, these journals are either reviewed by therapists (who are vastly outnumbered by the users) or left unused, with the assumption that simply talking about negative emotions is therapy in and of itself. We see a large and novel opportunity for applying natural language techniques to these unstructured mental health records. In this paper, we focus on analyzing the sentiment of patient text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The largest motivator of existing sentiment analysis research has arguably been the detection of user sentiment towards entities, such as products, companies, or people. We define this type of problem as external sentiment analysis. In contrast, when working in the mental health domain (particularly with self-reflective textual journals), we are trying to gauge a patient's internal sentiment towards their own thoughts, feelings, and emotions. The differences in goals, types of sentiment, and distribution of polarity presents unique challenges for applying sentiment analysis to this new domain.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "One key aspect that sets our task apart from traditional sentiment analysis is our treatment of polarity classes. Traditionally, sentiment is categorized as either positive, negative, or neutral. In contrast, we subdivide the neutral polarity class into two distinct classes: both positive and negative and neither positive nor negative. We justify this decision based on several studies showing the independent dimensions of positive and negative affect in human emotion (Warr et al., 1983; Watson et al., 1988; Diener et al., 1985; Bradburn, 1969) , and feel that is a more appropriate framework for our domain. This choice represents a novel characterization of sentiment analysis in mental health, and is one we hope to see made in future studies in this domain.", |
| "cite_spans": [ |
| { |
| "start": 472, |
| "end": 491, |
| "text": "(Warr et al., 1983;", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 492, |
| "end": 512, |
| "text": "Watson et al., 1988;", |
| "ref_id": "BIBREF36" |
| }, |
| { |
| "start": 513, |
| "end": 533, |
| "text": "Diener et al., 1985;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 534, |
| "end": 549, |
| "text": "Bradburn, 1969)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our primary focus in this paper is on the automatic and reliable categorization of patient responses as positive, negative, both positive and negative, or neither positive nor negative. Such a system has far-reaching implications for the online therapy setting, in which automatic language analysis can be incorporated into existing patient evaluation and progress monitoring, or serve as an early warning indicator for patients with severe cases of depression and/or risk of suicide. Additionally, tools based on this type of internal sentiment analysis can provide immediate feedback on mental health and thought processes, which can become distorted and unclear in patients stuck in anxiety or depression. In the future, sentiment-based mental health models can be incorporated into the characterization and treatment of patients with autism, dementia, or other broadly-defined language disorders.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In short, our main contributions are summarized by the following:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We present a novel sentiment analysis dataset, annotated by psychology experts, specifically targeted towards the mental health domain.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We introduce the notion of subdividing the traditional neutral polarity class into both a dual polarity sentiment (both positive and negative) and a neither positive nor negative sentiment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We identify the unique challenges faced when applying existing sentiment analysis techniques to mental health.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We present an automatic model for classifying the polarity of patient text, and compare our work to models trained on existing sentiment corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "From a technical point of view, our methods fall squarely in the realm of sentiment analysis, a field of computer science and computational linguistics primarily concerned with analyzing people's opinions, sentiments, attitudes, and emotions from written language (Liu, 2010) . In our paper, we apply sentiment analysis and polarity detection techniques to the largely untapped mental health domain.", |
| "cite_spans": [ |
| { |
| "start": 264, |
| "end": 275, |
| "text": "(Liu, 2010)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In the past decade, sentiment analysis techniques have been applied to a wide variety of areas. Although the majority of work has dealt in areas outside of mental health, we must discuss the bulk of previous sentiment analysis research, from which our techniques are derived.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Given the explosive rise in popularity of social media platforms, a large number of studies have focused on user sentiment in microblogs such as Twitter (Barbosa and Feng, 2010; Pak and Paroubek, 2010; Agarwal et al., 2011; Kouloumpis et al., 2011; Nielsen, 2011; Wang et al., 2011; Montejo-R\u00e1ez et al., 2012; Spencer and Uchyigit, 2012; Montejo-R\u00e1ez et al., 2014; Tang et al., 2014) . Other studies have explored user sentiment in web forum opinions (Abbasi et al., 2008) , movie reviews (Agrawal and Siddiqui, 2009) , blogs (Melville et al., 2009) , and Yahoo! Answers (Kucuktunc et al., 2012) . As we will show, the models proposed in all of these works cannot be directly transferred to polarity detection in mental health (as sentiment analysis remains a largely domain-specific task), but our initial techniques are largely based on these previous works.", |
| "cite_spans": [ |
| { |
| "start": 153, |
| "end": 177, |
| "text": "(Barbosa and Feng, 2010;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 178, |
| "end": 201, |
| "text": "Pak and Paroubek, 2010;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 202, |
| "end": 223, |
| "text": "Agarwal et al., 2011;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 224, |
| "end": 248, |
| "text": "Kouloumpis et al., 2011;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 249, |
| "end": 263, |
| "text": "Nielsen, 2011;", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 264, |
| "end": 282, |
| "text": "Wang et al., 2011;", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 283, |
| "end": 309, |
| "text": "Montejo-R\u00e1ez et al., 2012;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 310, |
| "end": 337, |
| "text": "Spencer and Uchyigit, 2012;", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 338, |
| "end": 364, |
| "text": "Montejo-R\u00e1ez et al., 2014;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 365, |
| "end": 383, |
| "text": "Tang et al., 2014)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 451, |
| "end": 472, |
| "text": "(Abbasi et al., 2008)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 489, |
| "end": 517, |
| "text": "(Agrawal and Siddiqui, 2009)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 526, |
| "end": 549, |
| "text": "(Melville et al., 2009)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 571, |
| "end": 595, |
| "text": "(Kucuktunc et al., 2012)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Although the majority of sentiment analysis has focused on user opinions towards entities, there are studies in domains more directly related to our area. One such study analyzed the sentiment of suicide notes (Pestian et al., 2012) . Another mined user sentiment in MOOC discussion forums (Wen et al., 2014) .", |
| "cite_spans": [ |
| { |
| "start": 210, |
| "end": 232, |
| "text": "(Pestian et al., 2012)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 290, |
| "end": 308, |
| "text": "(Wen et al., 2014)", |
| "ref_id": "BIBREF37" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Sentiment analysis and polarity detection techniques are widely varied (Mejova and Srinivasan, 2011; Feldman, 2013) , and as this research area is still garnering a great deal of interest, many studies have proposed novel methods. These include topiclevel sentiment analysis (Nasukawa and Yi, 2003; Kim and Hovy, 2004) , phrase-level sentiment analysis (Wilson et al., 2009) , linguistic approaches (Wiegand et al., 2010; Benamara et al., 2007; Tan et al., 2011) , semantic word vectorization (Maas et al., 2011; Tang et al., 2014) , various lexicon-based approaches (Taboada et al., 2011; Baccianella et al., 2010) , information-theoretic techniques (Lin et al., 2012) , and graph-based methods (Montejo-R\u00e1ez et al., 2014; Pang and Lee, 2004; Wang et al., 2011) . In recent years, approaches based on deep learning architectures have also shown promising results (Glo-rot et al., 2011; Socher et al., 2013) . As a starting point for our new internal sentiment analysis framework, in this paper we apply more straightforward approaches based on linear classifiers.", |
| "cite_spans": [ |
| { |
| "start": 71, |
| "end": 100, |
| "text": "(Mejova and Srinivasan, 2011;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 101, |
| "end": 115, |
| "text": "Feldman, 2013)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 275, |
| "end": 298, |
| "text": "(Nasukawa and Yi, 2003;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 299, |
| "end": 318, |
| "text": "Kim and Hovy, 2004)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 353, |
| "end": 374, |
| "text": "(Wilson et al., 2009)", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 399, |
| "end": 421, |
| "text": "(Wiegand et al., 2010;", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 422, |
| "end": 444, |
| "text": "Benamara et al., 2007;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 445, |
| "end": 462, |
| "text": "Tan et al., 2011)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 493, |
| "end": 512, |
| "text": "(Maas et al., 2011;", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 513, |
| "end": 531, |
| "text": "Tang et al., 2014)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 567, |
| "end": 589, |
| "text": "(Taboada et al., 2011;", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 590, |
| "end": 615, |
| "text": "Baccianella et al., 2010)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 651, |
| "end": 669, |
| "text": "(Lin et al., 2012)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 696, |
| "end": 723, |
| "text": "(Montejo-R\u00e1ez et al., 2014;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 724, |
| "end": 743, |
| "text": "Pang and Lee, 2004;", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 744, |
| "end": 762, |
| "text": "Wang et al., 2011)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 864, |
| "end": 886, |
| "text": "(Glo-rot et al., 2011;", |
| "ref_id": null |
| }, |
| { |
| "start": 887, |
| "end": 907, |
| "text": "Socher et al., 2013)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this section, we detail the construction of our mental health sentiment dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "While not yet publicly available, we plan to release our data in the near future.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In order to build a dataset of real patient responses, we partnered with TAO Connect, Inc. 1 , an online therapy program designed to treat anxiety, depression, and stress. This program is being implemented in several universities around the country, and as such, the primary demographic is collegeaged students.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "As part of the TAO program, patients complete several self-contained content modules designed to teach awareness and coping strategies for anxiety, depression, and stress. Additionally, patients regularly submit several types of journals and logs pertaining to monitoring, anxiety, depression, worries, and relaxation. The free text contained in these logs is the source of our dataset. In total, we collected 4021 textual responses from 342 unique patients, with submission dates ranging from April 2014 to November 2015. Patients were de-identified and the collection process was part of an IRB-approved study. Responses typically range from single sentences to a single paragraph, with an average of 39 words per response. We show a complete word count distribution in Figure 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 772, |
| "end": 780, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To help transform our collection of free text responses into a classification dataset suitable for polarity prediction, we solicited the expertise of three psychology undergraduates (all female) under the supervision of one psychology professor (male) to provide polarity labels for our response documents. The annotators were tasked with reading each individual response, and assigning it a label of positive, negative, both positive and negative, or neither positive nor negative. The inter-rater agreement reliability (Cohen's kappa) between annotators 1 and 2 was 0.5, between annotators 2 and 3 was 0.67, and between annotators 1 and 3 was 0.48. The overall 1 http://www.taoconnect.org/ agreement reliability between all annotators (Fleiss' kappa) was 0.55. We used a majority-vote scheme to assign a single label to each piece of text, where 62% of the documents had full annotator agreement, 35% had a clear label majority, and only 3% had no majority, in which case we picked the label from the annotator with the best aggregate reliability. Table 1 shows label counts for each annotator, as well as the final count after applying the majority-vote process.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1050, |
| "end": 1057, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To provide a clearer picture of the types of responses in our dataset, we present one short concrete example of each polarity class below.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Positive -I tried to say good things for them since I know there was a lot of arguments happening.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Negative -I don't do well at parties, I'm not interesting.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Both Positive and Negative -I shouldn't have taken things so seriously.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 Neither Positive nor Negative -I wrote in my journal, and read till I was tired enough to fall asleep.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In the above examples, the challenges of applying sentiment analysis and traditional text classification techniques to self-reflective text becomes more apparent. For instance, the positive example mentions arguments, typically associated with negative sentiment, while the negative example mentions parties, a word usually associated with a positive connotation. Additionally, the both positive and negative example exhibits subtle cues that differentiate it from the other three polarity classes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To predict polarity from patient text, we employ several established machine learning and text classification techniques. We begin by preprocessing the annotated patient responses, which we refer to interchangeably as documents. We then extract several types of attributes from each response, referred to as features. The extracted features and polarity annotations are used to build a logistic regression classifier, which is a linear machine learning model we use to predict the final polarity label. In this section, we describe each step in detail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Method", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Starting with the raw documents obtained from our data collection process, we apply several traditional preprocessing steps to the text. First, based on experimental feedback, we convert all the text to lowercase and strip all documents of punctuation following a standard tokenization phase. While these are relatively standard steps, it should be explicitly noted that we did not remove stop words from our corpus, which is a common preprocessing technique in other domains, due to lowered classification performance. This can be partially explained by the nature of our domain; for example, the phrase \"what if\" tended to be associated with worrying about the future -traditionally, both of these words are considered stop words and filtered out, losing valuable information for our task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preprocessing", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Next, we extract several types of features from the preprocessed documents. In our experiments, we evaluate classification performance with various feature subsets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Feature Extraction", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "As a starting point for our experiments with this new domain, the most numerous of our extracted features are derived from a traditional \"bag of ngrams\" approach, in which we create document vectors comprised of word unigrams, bigrams, and/or trigram counts. As previous works have shown, this allows the capture of important syntactical information like negation, which would otherwise be missed in a standard \"bag of words\" (i.e., unigrams only) model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "N-Gram Features and POS Tags", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "In order to constrain the scope of later feature subset experiments, we first obtain the n-gram combination resulting in the best performance for our newly created dataset. We denote this optimal n-gram setting as the \"n-grams only\" model in later experiments. In this experiment, we perform a 10-fold cross-validated randomized parameter search using six possible word n-gram combinations: unigrams, bigrams, trigrams, unigrams + bigrams, bigrams + trigrams, and unigrams + bigrams + trigrams. We split cross-validation folds on responses, as we expect patient responses to be independent over time. All extracted n-gram counts are normalized by tfidf (term frequency-inverse document frequency), a common technique used for describing how important particular n-grams are to their respective documents. The results of this n-gram comparison experiment are shown in Figure 2 , where it is clear that using a combination of unigrams and bigrams resulted in the best performance.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 867, |
| "end": 875, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "N-Gram Features and POS Tags", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "In an effort to capture more subtle patterns of grammatical structure, we also experiment with augmenting each document with each word's Penn-Treebank part-of-speech (POS) tag. In these experiments, we augment our documents by appending these tags, in order, to the end of every sentence, allowing for our n-gram extraction methods to capture syntactic language patterns. During the tokenization process, we ignore any n-grams consisting of both words and part-of-speech tags.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "N-Gram Features and POS Tags", |
| "sec_num": "4.2.1" |
| }, |
| { |
| "text": "One of the more rudimentary sentiment analysis techniques stems from the use of a sentiment dictio- Figure 2 : Classification results using only word n-gram features for our 4-class polarity dataset. Results were obtained following a 10-fold cross-validated randomized hyperparameter search. A combination of unigrams and bigrams resulted in the best metrics.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 100, |
| "end": 108, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Sentiment Lexicon Word Counts", |
| "sec_num": "4.2.2" |
| }, |
| { |
| "text": "As seen by the final cluster, adding trigrams to this subset resulted in a performance decrease. Thus, when we use n-gram features in later experiments, we only consider the combination of unigrams and bigrams.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Lexicon Word Counts", |
| "sec_num": "4.2.2" |
| }, |
| { |
| "text": "nary, or lexicon, which is a pre-existing collection of subjective words that are labeled as either positive or negative. Using the sentiment lexicon from (Liu, 2012) 2 , we count the number of positive and negative words occurring in each document and incorporate the counts as two additional features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Lexicon Word Counts", |
| "sec_num": "4.2.2" |
| }, |
| { |
| "text": "In our initial analysis, we discovered that oftentimes the most negative text responses were associated with a larger word count. Although the correlation is relatively weak across the entire corpus, we nonetheless include a word count of each document as a feature.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Document Word Count", |
| "sec_num": "4.2.3" |
| }, |
| { |
| "text": "Based on the recent successes of distributed word representations like Word2Vec (Mikolov et al., 2013) and GloVe (Pennington et al., 2014) , we sought to harness these learned language models for predicting sentiment polarity. Although primarily used in deep learning architectures, we show that these representations can also be useful with linear models. Unlike our other features, the individual features contained in word embeddings are indecipherable; however, as we show in the results section, they contribute to the overall success of our classification.", |
| "cite_spans": [ |
| { |
| "start": 80, |
| "end": 102, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 113, |
| "end": 138, |
| "text": "(Pennington et al., 2014)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word Embeddings", |
| "sec_num": "4.2.4" |
| }, |
| { |
| "text": "In our experiments, we utilize a publicly avail- ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word Embeddings", |
| "sec_num": "4.2.4" |
| }, |
| { |
| "text": "Because our new dataset introduces a clear distinction between text labeled as both positive and negative and neither positive nor negative (traditionally, both of these classes are grouped together as neutral), there are no baselines for which to compare our experiments. We offer our results for this scenario as a launching point for future studies on polarity detection in mental health. For this scenario, we show the results of each feature extraction method individually, as well as the results for the combination of all features. All results are evaluated via 10-fold cross-validation, with folds split on responses. Results are shown in Figure 3 , where it is clear that optimal performance is achieved using the model trained on all features. Our methods gave rise to an overall classification accuracy of 78%.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 647, |
| "end": 655, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Four-Class Polarity Prediction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "From Figure 3 , it is apparent that of all individual features, n-grams perform the best. The relatively strong performance of n-gram features tends to align with our expectations, given the widespread use of n-gram features across all types of text classifica- tion problems. However, what is more surprising is the relatively weak results for the sentiment lexicon features, given their popularity in modern sentiment analysis. Additionally, the word embedding features also gave rise to better performance than expected, especially considering that we used the Word2Vec embeddings with linear models as opposed to the more traditional deep learning architectures. Finally, we see optimal performance across all metrics when using the combination of all features. Using the optimal model from Figure 3 , we show the individual class metrics for precision, recall, F1, and overall accuracy in Table 2 . It is apparent that the both positive and negative class proves especially difficult to classify. This is explained in part by the previously mentioned class imbalance issue -when the majority of the corpus is negative, it becomes difficult for the classifier to differentiate between sentiment comprising of mostly positive polarity, and sentiment comprising of some positive polarity. The low recall of the both positive and negative class clearly points towards the need for more research in this area.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 5, |
| "end": 13, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 795, |
| "end": 803, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 894, |
| "end": 901, |
| "text": "Table 2", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Four-Class Polarity Prediction", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In this section, we experiment with using existing sentiment analysis corpora to perform traditional two-class polarity prediction on our dataset, and compare the results to a cross-validation approach, split on responses, trained on our dataset alone. The primary purpose is to gauge the effectiveness of classifiers trained on existing sentiment corpora as applied to the mental health domain. State of the art sentence-level binary polarity detection accuracy is reported as 85.4% (Socher et al., 2013) using deep learning models and a specialized movie review dataset, and while our models are computationally more simple and use different features, we incorporate such existing corpora in our experiments. Since our full dataset consists of four polarity labels, whereas traditional sentiment analysis only uses two, for these experiments we only consider the responses from our dataset belonging to the positive and negative classes. We begin by training our model on existing sentiment datasets only. The first is a large-scale Twitter sentiment analysis dataset 4 which automatically assigns polarity labels based on emoticons present in user tweets (we denote this dataset as \"Twitter\"). The next is a collection of IMDB movie reviews published by (Maas et al., 2011) at Stanford University 5 (we denote this dataset as \"Stanford\"). We also use two movie reviews datasets from (Pang et 2002) at Cornell University 6 , where one is geared towards document-level sentiment classification (denoted as \"Cornell\"), and the other towards sentencelevel classification (denoted as \"Cornell Sentence\"). Our final dataset is a collection of web forum opinions collected by the University of Michigan as part of a Kaggle competition 7 (which we denote as \"UMich\"). The number of documents of each sentiment class, per dataset, is given in Table 3 . Using all features from the previously outlined extraction process, we train a separate model on each of the five existing sentiment analysis corpora. Optimal hyperparameters for each experiment were selected via a randomized parameter search in conjunction with three-fold cross validation. In each case, the trained models were tested against the binary version of our dataset. Additionally, we perform the same extraction and fine-tuning process to construct a model trained on our new dataset alone. For this experiment, we report the results after a 10-fold cross-validation process split on responses. A summary of accuracy, precision, recall, and F1 score for the binary prediction setting is shown in Figure 4 , where it is apparent that the best performance occurs when using our dataset, pointing towards the need for collecting custom mental health datasets for this new type of internal sentiment analysis. Our binary polarity model resulted in 90% classification accuracy.", |
| "cite_spans": [ |
| { |
| "start": 484, |
| "end": 505, |
| "text": "(Socher et al., 2013)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 1257, |
| "end": 1276, |
| "text": "(Maas et al., 2011)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 1386, |
| "end": 1394, |
| "text": "(Pang et", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1837, |
| "end": 1844, |
| "text": "Table 3", |
| "ref_id": "TABREF8" |
| }, |
| { |
| "start": 2556, |
| "end": 2564, |
| "text": "Figure 4", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Binary Polarity Prediction", |
| "sec_num": "6" |
| }, |
| { |
| "text": "In this section, we wish to understand which features are most discriminative in predicting whether a piece of text is positive, negative, both positive and negative, or neither positive nor negative. These features (all of which are naturally-interpretable aside from the word embeddings) can serve as useful indicators for therapists and future mental health polarity studies.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Important Features", |
| "sec_num": "7" |
| }, |
| { |
| "text": "To evaluate our features, we examine the weight matrix of a randomized logistic regression classifier trained on our full four-class polarity dataset. The feature weights corresponding to each of the four classes give an idea of the relative importance of each feature, and how discriminative they are as compared to the remaining three classes. We summarize the 10 most important features per class in Table 4 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 403, |
| "end": 410, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Important Features", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Much can be gleaned from an informal inspection of these top features. For example, while the words found in the positive and negative polarity classes are clearly characteristic of their respective labels (with negative words pertaining mostly to worry and stress), the words found in the both positive and negative class are more indecisive in nature ('but', 'eventually', 'although', 'actually') . Words from the neither positive nor negative class carry less surface-level emotional significance. The partof-speech patterns are more difficult to interpret, but these results point towards the need for future exploration.", |
| "cite_spans": [ |
| { |
| "start": 353, |
| "end": 398, |
| "text": "('but', 'eventually', 'although', 'actually')", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Important Features", |
| "sec_num": "7" |
| }, |
| { |
| "text": "In this paper, we introduced the notion of applying sentiment analysis to the mental health domain, and show that existing techniques and corpora cannot be simply transferred to this new setting. We developed baseline classification techniques grounded in the results from previous works, and show the benefit of spending resources on creating new mental health datasets explicitly focused on patient sentiment. We introduced the notion of splitting the polarity class traditionally defined as neutral into two sub-classes, and demonstrated the new challenges that decision brings as it pertains to the automatic classification of patient sentiment in mental health text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "8" |
| }, |
| { |
| "text": "https://code.google.com/p/word2vec/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://help.sentiment140.com/for-students/ 5 http://ai.stanford.edu/ amaas/data/sentiment/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://www.cs.cornell.edu/people/pabo/movie-reviewdata/ 7 https://inclass.kaggle.com/c/si650winter11/data", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Sentiment analysis in multiple languages: Feature selection for opinion classification in web forums", |
| "authors": [ |
| { |
| "first": "Ahmed", |
| "middle": [], |
| "last": "Abbasi", |
| "suffix": "" |
| }, |
| { |
| "first": "Hsinchun", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Arab", |
| "middle": [], |
| "last": "Salem", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "ACM Transactions on Information Systems", |
| "volume": "26", |
| "issue": "3", |
| "pages": "1--34", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ahmed Abbasi, Hsinchun Chen, and Arab Salem. 2008. Sentiment analysis in multiple languages: Feature se- lection for opinion classification in web forums. ACM Transactions on Information Systems, 26(3):1-34.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Sentiment analysis of Twitter data", |
| "authors": [ |
| { |
| "first": "Apoorv", |
| "middle": [], |
| "last": "Agarwal", |
| "suffix": "" |
| }, |
| { |
| "first": "Boyi", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilia", |
| "middle": [], |
| "last": "Vovsha", |
| "suffix": "" |
| }, |
| { |
| "first": "Owen", |
| "middle": [], |
| "last": "Rambow", |
| "suffix": "" |
| }, |
| { |
| "first": "Rebecca", |
| "middle": [], |
| "last": "Passonneau", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Workshop on Languages in Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "30--38", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Apoorv Agarwal, Boyi Xie, Ilia Vovsha, Owen Rambow, and Rebecca Passonneau. 2011. Sentiment analysis of Twitter data. In Proceedings of the Workshop on Languages in Social Media, pages 30-38.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Using syntactic and contextual information for sentiment polarity analysis", |
| "authors": [ |
| { |
| "first": "Shaishav", |
| "middle": [], |
| "last": "Agrawal", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Tanveer", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Siddiqui", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the 2nd International Conference on Interaction Sciences: Information Technology", |
| "volume": "", |
| "issue": "", |
| "pages": "620--623", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shaishav Agrawal and Tanveer J Siddiqui. 2009. Us- ing syntactic and contextual information for sentiment polarity analysis. In Proceedings of the 2nd Interna- tional Conference on Interaction Sciences: Informa- tion Technology, Culture and Human, pages 620-623.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "SentiWordNet 3.0: An enhanced lexical resource for sentiment analysis and opinion mining", |
| "authors": [ |
| { |
| "first": "Stefano", |
| "middle": [], |
| "last": "Baccianella", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Esuli", |
| "suffix": "" |
| }, |
| { |
| "first": "Fabrizio", |
| "middle": [], |
| "last": "Sebastiani", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)", |
| "volume": "", |
| "issue": "", |
| "pages": "2200--2204", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stefano Baccianella, Andrea Esuli, and Fabrizio Sebas- tiani. 2010. SentiWordNet 3.0: An enhanced lexical resource for sentiment analysis and opinion mining. In Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10), pages 2200-2204.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Robust sentiment detection on Twitter from biased and noisy data", |
| "authors": [ |
| { |
| "first": "Luciano", |
| "middle": [], |
| "last": "Barbosa", |
| "suffix": "" |
| }, |
| { |
| "first": "Junlan", |
| "middle": [], |
| "last": "Feng", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 23rd International Conference on Computational Linguistics: Posters", |
| "volume": "", |
| "issue": "", |
| "pages": "36--44", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luciano Barbosa and Junlan Feng. 2010. Robust senti- ment detection on Twitter from biased and noisy data. In Proceedings of the 23rd International Conference on Computational Linguistics: Posters, pages 36-44.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Sentiment analysis: adjectives and adverbs are better than adjectives alone", |
| "authors": [ |
| { |
| "first": "Farah", |
| "middle": [], |
| "last": "Benamara", |
| "suffix": "" |
| }, |
| { |
| "first": "Carmine", |
| "middle": [], |
| "last": "Cesarano", |
| "suffix": "" |
| }, |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Picariello", |
| "suffix": "" |
| }, |
| { |
| "first": "Diego", |
| "middle": [], |
| "last": "Reforgiato", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [ |
| "S" |
| ], |
| "last": "Subrahmanian", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the International Conference on Weblogs and Social Media (ICWSM)", |
| "volume": "", |
| "issue": "", |
| "pages": "203--206", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Farah Benamara, Carmine Cesarano, Antonio Picariello, Diego Reforgiato, and V.S. Subrahmanian. 2007. Sen- timent analysis: adjectives and adverbs are better than adjectives alone. In Proceedings of the International Conference on Weblogs and Social Media (ICWSM), pages 203-206.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "The structure of psychological well-being", |
| "authors": [ |
| { |
| "first": "Norman", |
| "middle": [ |
| "M" |
| ], |
| "last": "Bradburn", |
| "suffix": "" |
| } |
| ], |
| "year": 1969, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Norman M. Bradburn. 1969. The structure of psycho- logical well-being.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Intensity and frequency: dimensions underlying positive and negative affect", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Diener", |
| "suffix": "" |
| }, |
| { |
| "first": "R J", |
| "middle": [], |
| "last": "Larsen", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Levine", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Emmons", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "Journal of personality and social psychology", |
| "volume": "48", |
| "issue": "5", |
| "pages": "1253--1265", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "E Diener, R J Larsen, S Levine, and R a Emmons. 1985. Intensity and frequency: dimensions underlying pos- itive and negative affect. Journal of personality and social psychology, 48(5):1253-1265.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Techniques and applications for sentiment analysis", |
| "authors": [ |
| { |
| "first": "Ronen", |
| "middle": [], |
| "last": "Feldman", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Communications of the ACM", |
| "volume": "56", |
| "issue": "4", |
| "pages": "82--89", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronen Feldman. 2013. Techniques and applications for sentiment analysis. Communications of the ACM, 56(4):82-89.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Domain adaptation for large-scale sentiment classification: A deep learning approach", |
| "authors": [ |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Glorot", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Bordes", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "X Glorot, A Bordes, and Y Bengio. 2011. Domain adap- tation for large-scale sentiment classification: A deep learning approach. Proceedings of the.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Determining the sentiment of opinions", |
| "authors": [ |
| { |
| "first": "Min", |
| "middle": [], |
| "last": "Soo", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the 20th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1367--1373", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soo-Min Kim and Eduard Hovy. 2004. Determining the sentiment of opinions. In Proceedings of the 20th In- ternational Conference on Computational Linguistics, pages 1367-1373, Morristown, NJ, USA.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Twitter sentiment analysis: The good the bad and the omg!", |
| "authors": [ |
| { |
| "first": "Efthymios", |
| "middle": [], |
| "last": "Kouloumpis", |
| "suffix": "" |
| }, |
| { |
| "first": "Theresa", |
| "middle": [], |
| "last": "Wilson", |
| "suffix": "" |
| }, |
| { |
| "first": "Johanna", |
| "middle": [], |
| "last": "Moore", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Fifth International AAAI Conference on Weblogs and Social Media (ICWSM 11)", |
| "volume": "", |
| "issue": "", |
| "pages": "538--541", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Efthymios Kouloumpis, Theresa Wilson, and Johanna Moore. 2011. Twitter sentiment analysis: The good the bad and the omg! In Proceedings of the Fifth In- ternational AAAI Conference on Weblogs and Social Media (ICWSM 11), pages 538-541.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "A large-scale sentiment analysis for Yahoo! Answers", |
| "authors": [ |
| { |
| "first": "Onur", |
| "middle": [], |
| "last": "Kucuktunc", |
| "suffix": "" |
| }, |
| { |
| "first": "Ingmar", |
| "middle": [], |
| "last": "Weber", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Barla", |
| "suffix": "" |
| }, |
| { |
| "first": "Hakan", |
| "middle": [], |
| "last": "Cambazoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ferhatosmanoglu", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 5th ACM International Conference on Web Search and Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "633--642", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Onur Kucuktunc, Ingmar Weber, B. Barla Cambazoglu, and Hakan Ferhatosmanoglu. 2012. A large-scale sentiment analysis for Yahoo! Answers. In Proceed- ings of the 5th ACM International Conference on Web Search and Data Mining, pages 633-642.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "An information theoretic approach to sentiment polarity classification", |
| "authors": [ |
| { |
| "first": "Yuming", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Jingwei", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaoling", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Aoying", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 2nd Joint WICOW/AIRWeb Workshop on Web Quality", |
| "volume": "", |
| "issue": "", |
| "pages": "35--40", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yuming Lin, Jingwei Zhang, Xiaoling Wang, and Aoying Zhou. 2012. An information theoretic approach to sentiment polarity classification. In Proceedings of the 2nd Joint WICOW/AIRWeb Workshop on Web Quality, pages 35-40.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Sentiment Analysis and Subjectivity", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Liu. 2010. Sentiment Analysis and Subjectivity. 2 edition.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Sentiment analysis and opinion mining", |
| "authors": [ |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Synthesis Lectures on Human Language Technologies", |
| "volume": "5", |
| "issue": "1", |
| "pages": "1--167", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bing Liu. 2012. Sentiment analysis and opinion mining. Synthesis Lectures on Human Language Technologies, 5(1):1-167.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Learning word vectors for sentiment analysis", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "E" |
| ], |
| "last": "Maas", |
| "suffix": "" |
| }, |
| { |
| "first": "Peter", |
| "middle": [ |
| "T" |
| ], |
| "last": "Daly", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Pham", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [ |
| "Y" |
| ], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Potts", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "142--150", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrew L Maas, Raymond E. Daly, Peter T. Pham, Dan Huang, Andrew Y. Ng, and Christopher Potts. 2011. Learning word vectors for sentiment analysis. In Pro- ceedings of the 49th Annual Meeting of the Associa- tion for Computational Linguistics: Human Language Technologies, pages 142-150. Association for Compu- tational Linguistics.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Exploring feature definition and selection for sentiment classifiers", |
| "authors": [ |
| { |
| "first": "Yelena", |
| "middle": [], |
| "last": "Mejova", |
| "suffix": "" |
| }, |
| { |
| "first": "Padmini", |
| "middle": [], |
| "last": "Srinivasan", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Fifth International AAAI Conference on Weblogs and Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "546--549", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yelena Mejova and Padmini Srinivasan. 2011. Explor- ing feature definition and selection for sentiment clas- sifiers. In Proceedings of the Fifth International AAAI Conference on Weblogs and Social Media, pages 546- 549.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Sentiment analysis of blogs by combining lexical knowledge with text classification", |
| "authors": [ |
| { |
| "first": "Prem", |
| "middle": [], |
| "last": "Melville", |
| "suffix": "" |
| }, |
| { |
| "first": "Wojciech", |
| "middle": [], |
| "last": "Gryc", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [ |
| "D" |
| ], |
| "last": "Lawrence", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of ACM SIGKDD International Conference on Knowledge Discovery and Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "1275--1284", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Prem Melville, Wojciech Gryc, and Richard D. Lawrence. 2009. Sentiment analysis of blogs by com- bining lexical knowledge with text classification. In Proceedings of ACM SIGKDD International Confer- ence on Knowledge Discovery and Data Mining, pages 1275-1284, New York, New York, USA. ACM Press.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of International Conference on Learning Representations (ICLR)", |
| "volume": "", |
| "issue": "", |
| "pages": "1--12", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word repre- sentations in vector space. In Proceedings of In- ternational Conference on Learning Representations (ICLR), pages 1-12.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Random walk weighting over SentiWordNet for sentiment polarity detection on Twitter", |
| "authors": [ |
| { |
| "first": "Arturo", |
| "middle": [], |
| "last": "Montejo-R\u00e1ez", |
| "suffix": "" |
| }, |
| { |
| "first": "Eugenio", |
| "middle": [], |
| "last": "Mart\u00ednez-C\u00e1mara", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Teresa Mart\u00edn-Valdivia", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [ |
| "Alfonso" |
| ], |
| "last": "Ure\u00f1a-L\u00f3pez", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 3rd Workshop in Computational Approaches to Subjectivity and Sentiment Analysis", |
| "volume": "", |
| "issue": "", |
| "pages": "3--10", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arturo Montejo-R\u00e1ez, Eugenio Mart\u00ednez-C\u00e1mara, M. Teresa Mart\u00edn-Valdivia, and L. Alfonso Ure\u00f1a- L\u00f3pez. 2012. Random walk weighting over SentiWordNet for sentiment polarity detection on Twitter. In Proceedings of the 3rd Workshop in Com- putational Approaches to Subjectivity and Sentiment Analysis, pages 3-10. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Ranked WordNet graph for sentiment polarity classification in Twitter", |
| "authors": [ |
| { |
| "first": "Arturo", |
| "middle": [], |
| "last": "Montejo-R\u00e1ez", |
| "suffix": "" |
| }, |
| { |
| "first": "Eugenio", |
| "middle": [], |
| "last": "Mart\u00ednez-C\u00e1mara", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Teresa Mart\u00edn-Valdivia", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [ |
| "Alfonso" |
| ], |
| "last": "Ure\u00f1a-L\u00f3pez", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Computer Speech & Language", |
| "volume": "28", |
| "issue": "1", |
| "pages": "93--107", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arturo Montejo-R\u00e1ez, Eugenio Mart\u00ednez-C\u00e1mara, M. Teresa Mart\u00edn-Valdivia, and L. Alfonso Ure\u00f1a- L\u00f3pez. 2014. Ranked WordNet graph for sentiment polarity classification in Twitter. Computer Speech & Language, 28(1):93-107.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Sentiment analysis: Capturing favorability using natural language processing", |
| "authors": [ |
| { |
| "first": "Tetsuya", |
| "middle": [], |
| "last": "Nasukawa", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeonghee", |
| "middle": [], |
| "last": "Yi", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the 2nd International Conference on Knowledge Capture", |
| "volume": "", |
| "issue": "", |
| "pages": "70--77", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tetsuya Nasukawa and Jeonghee Yi. 2003. Senti- ment analysis: Capturing favorability using natural language processing. In Proceedings of the 2nd In- ternational Conference on Knowledge Capture, pages 70-77, New York, New York, USA.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "A new ANEW: Evaluation of a word list for sentiment analysis in microblogs", |
| "authors": [ |
| { |
| "first": "Finn\u00e5rup", |
| "middle": [], |
| "last": "Nielsen", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "ESWC2011 Workshop on Making Sense of Microposts", |
| "volume": "", |
| "issue": "", |
| "pages": "93--98", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Finn\u00c5rup Nielsen. 2011. A new ANEW: Evaluation of a word list for sentiment analysis in microblogs. In ESWC2011 Workshop on Making Sense of Microposts, pages 93-98.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Twitter as a corpus for sentiment analysis and opinion mining", |
| "authors": [ |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Pak", |
| "suffix": "" |
| }, |
| { |
| "first": "Patrick", |
| "middle": [], |
| "last": "Paroubek", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Seventh Conference on International Language Resources and Evaluation (LREC'10)", |
| "volume": "", |
| "issue": "", |
| "pages": "19--21", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexander Pak and Patrick Paroubek. 2010. Twit- ter as a corpus for sentiment analysis and opinion mining. In Proceedings of the Seventh Conference on International Language Resources and Evaluation (LREC'10), pages 19-21.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts", |
| "authors": [ |
| { |
| "first": "Bo", |
| "middle": [], |
| "last": "Pang", |
| "suffix": "" |
| }, |
| { |
| "first": "Lillian", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "271--278", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bo Pang and Lillian Lee. 2004. A sentimental edu- cation: Sentiment analysis using subjectivity summa- rization based on minimum cuts. In Proceedings of the 42nd Annual Meeting on Association for Compu- tational Linguistics, pages 271-278.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Thumbs up? Sentiment classification using machine learning techniques", |
| "authors": [ |
| { |
| "first": "Bo", |
| "middle": [], |
| "last": "Pang", |
| "suffix": "" |
| }, |
| { |
| "first": "Lilian", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Shivakumar", |
| "middle": [], |
| "last": "Vaithyanathan", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "79--86", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bo Pang, Lilian Lee, and Shivakumar Vaithyanathan. 2002. Thumbs up? Sentiment classification using ma- chine learning techniques. In Proceedings of the Con- ference on Empirical Methods in Natural Language Processing, pages 79-86.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "GloVe: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher D", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Empiricial Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. GloVe: Global vectors for word rep- resentation. In Proceedings of the Empiricial Methods in Natural Language Processing, pages 1532-1543.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Sentiment analysis of suicide notes: A shared task", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Pestian", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Pestian", |
| "suffix": "" |
| }, |
| { |
| "first": "Pawel", |
| "middle": [], |
| "last": "Matykiewicz", |
| "suffix": "" |
| }, |
| { |
| "first": "Brett", |
| "middle": [], |
| "last": "South", |
| "suffix": "" |
| }, |
| { |
| "first": "Ozlem", |
| "middle": [], |
| "last": "Uzuner", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Hurdle", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Biomedical Informatics Insights", |
| "volume": "5", |
| "issue": "1", |
| "pages": "3--16", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Pestian, John Pestian, Pawel Matykiewicz, Brett South, Ozlem Uzuner, and John Hurdle. 2012. Senti- ment analysis of suicide notes: A shared task. Biomed- ical Informatics Insights, 5(1):3-16.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Recursive deep models for semantic compositionality over a sentiment treebank", |
| "authors": [ |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Perelygin", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Jean", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Chuang", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Andrew", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Potts", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1631--1642", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard Socher, Alex Perelygin, Jean Y Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and Christopher Potts. 2013. Recursive deep models for semantic compositionality over a sentiment tree- bank. In Proceedings of the Conference on Empiri- cal Methods in Natural Language Processing, pages 1631-1642.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Sentimentor: Sentiment analysis of Twitter data", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Spencer", |
| "suffix": "" |
| }, |
| { |
| "first": "Gulden", |
| "middle": [], |
| "last": "Uchyigit", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases", |
| "volume": "", |
| "issue": "", |
| "pages": "56--66", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Spencer and Gulden Uchyigit. 2012. Sentimen- tor: Sentiment analysis of Twitter data. In Proceed- ings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, pages 56-66.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Lexicon-based methods for sentiment analysis", |
| "authors": [ |
| { |
| "first": "Maite", |
| "middle": [], |
| "last": "Taboada", |
| "suffix": "" |
| }, |
| { |
| "first": "Julian", |
| "middle": [], |
| "last": "Brooke", |
| "suffix": "" |
| }, |
| { |
| "first": "Milan", |
| "middle": [], |
| "last": "Tofiloski", |
| "suffix": "" |
| }, |
| { |
| "first": "Kimberly", |
| "middle": [], |
| "last": "Voll", |
| "suffix": "" |
| }, |
| { |
| "first": "Manfred", |
| "middle": [], |
| "last": "Stede", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Computational Linguistics", |
| "volume": "37", |
| "issue": "2", |
| "pages": "267--307", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Maite Taboada, Julian Brooke, Milan Tofiloski, Kimberly Voll, and Manfred Stede. 2011. Lexicon-based meth- ods for sentiment analysis. Computational Linguis- tics, 37(2):267-307.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Sentence-level sentiment polarity classification using a linguistic approach", |
| "authors": [ |
| { |
| "first": "Luke Kien Weng", |
| "middle": [], |
| "last": "Tan", |
| "suffix": "" |
| }, |
| { |
| "first": "Jin", |
| "middle": [ |
| "Cheon" |
| ], |
| "last": "Na", |
| "suffix": "" |
| }, |
| { |
| "first": "Yin", |
| "middle": [], |
| "last": "Leng Theng", |
| "suffix": "" |
| }, |
| { |
| "first": "Kuiyu", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of 13th International Conference on Asia-Pacific Digital Libraries", |
| "volume": "", |
| "issue": "", |
| "pages": "77--87", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luke Kien Weng Tan, Jin Cheon Na, Yin Leng Theng, and Kuiyu Chang. 2011. Sentence-level sentiment polarity classification using a linguistic approach. In Proceedings of 13th International Conference on Asia- Pacific Digital Libraries, pages 77-87.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Learning sentiment-specific word embedding for Twitter sentiment classification", |
| "authors": [ |
| { |
| "first": "Duyu", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Nan", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Ting", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1555--1565", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, and Bing Qin. 2014. Learning sentiment-specific word embedding for Twitter sentiment classification. In Proceedings of the 52nd Annual Meeting of the As- sociation for Computational Linguistics, pages 1555- 1565.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Topic sentiment analysis in Twitter: A graph-based hashtag sentiment classification approach", |
| "authors": [ |
| { |
| "first": "Xiaolong", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaohua", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 20th ACM International Conference on Information and Knowledge Management", |
| "volume": "", |
| "issue": "", |
| "pages": "1031--1040", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xiaolong Wang, Furu Wei, Xiaohua Liu, Ming Zhou, and Ming Zhang. 2011. Topic sentiment analysis in Twit- ter: A graph-based hashtag sentiment classification ap- proach. In Proceedings of the 20th ACM International Conference on Information and Knowledge Manage- ment, pages 1031-1040, New York, New York, USA. ACM Press.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "On the Independence of Positive and Negative Affect", |
| "authors": [ |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Warr", |
| "suffix": "" |
| }, |
| { |
| "first": "Joanna", |
| "middle": [], |
| "last": "Barter", |
| "suffix": "" |
| }, |
| { |
| "first": "Garry", |
| "middle": [], |
| "last": "Brownbridge", |
| "suffix": "" |
| } |
| ], |
| "year": 1983, |
| "venue": "Journal of Personality and Social Psychology", |
| "volume": "44", |
| "issue": "3", |
| "pages": "644--651", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter Warr, Joanna Barter, and Garry Brownbridge. 1983. On the Independence of Positive and Negative Affect. Journal of Personality and Social Psychology, 44(3):644-651.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Development and Validation of Brief Measures of Positive and Negative Affect -the Panas Scales", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Watson", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Tellegen", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "Journal of Personality and Social Psychology", |
| "volume": "54", |
| "issue": "6", |
| "pages": "1063--1070", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D Watson, L A Clark, and A Tellegen. 1988. Develop- ment and Validation of Brief Measures of Positive and Negative Affect -the Panas Scales. Journal of Person- ality and Social Psychology, 54(6):1063-1070.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Sentiment analysis in MOOC discussion forums: What does it tell us?", |
| "authors": [ |
| { |
| "first": "Miaomiao", |
| "middle": [], |
| "last": "Wen", |
| "suffix": "" |
| }, |
| { |
| "first": "Diyi", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Cp", |
| "middle": [], |
| "last": "Ros\u00e9", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of Educational Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Miaomiao Wen, Diyi Yang, and Cp Ros\u00e9. 2014. Sen- timent analysis in MOOC discussion forums: What does it tell us? In Proceedings of Educational Data Mining, pages 1-8.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "A survey on the role of negation in sentiment analysis", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Wiegand", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Balahur", |
| "suffix": "" |
| }, |
| { |
| "first": "Benjamin", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| }, |
| { |
| "first": "Dietrich", |
| "middle": [], |
| "last": "Klakow", |
| "suffix": "" |
| }, |
| { |
| "first": "Andr\u00e9s", |
| "middle": [], |
| "last": "Montoyo", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Workshop on Negation and Speculation in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "60--68", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael Wiegand, Alexandra Balahur, Benjamin Roth, Dietrich Klakow, and Andr\u00e9s Montoyo. 2010. A sur- vey on the role of negation in sentiment analysis. In Proceedings of the Workshop on Negation and Specu- lation in Natural Language Processing, pages 60-68.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Recognizing contextual polarity: An exploration of features for phrase-level sentiment analysis", |
| "authors": [ |
| { |
| "first": "Theresa", |
| "middle": [ |
| "A" |
| ], |
| "last": "Wilson", |
| "suffix": "" |
| }, |
| { |
| "first": "Janyce", |
| "middle": [], |
| "last": "Wiebe", |
| "suffix": "" |
| }, |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Hoffmann", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Computational Linguistics", |
| "volume": "35", |
| "issue": "3", |
| "pages": "399--433", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Theresa A. Wilson, Janyce Wiebe, and Paul Hoffmann. 2009. Recognizing contextual polarity: An explo- ration of features for phrase-level sentiment analysis. Computational Linguistics, 35(3):399-433.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "Combining lexiconbased and learning-based methods for Twitter sentiment analysis", |
| "authors": [ |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Riddhiman", |
| "middle": [], |
| "last": "Ghosh", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohamed", |
| "middle": [], |
| "last": "Dekhil", |
| "suffix": "" |
| }, |
| { |
| "first": "Meichun", |
| "middle": [], |
| "last": "Hsu", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lei Zhang, Riddhiman Ghosh, Mohamed Dekhil, Me- ichun Hsu, and Bing Liu. 2011. Combining lexicon- based and learning-based methods for Twitter senti- ment analysis. Technical report.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "uris": null, |
| "type_str": "figure", |
| "text": "Classification performance for the 4-class polarity prediction task. We show results for each feature set individually, as well as the combination of all features. Using all extracted features results in the highest accuracy, F1, precision, and recall.", |
| "num": null |
| }, |
| "FIGREF2": { |
| "uris": null, |
| "type_str": "figure", |
| "text": "Top 10 features per class from a randomized logistic regression model, trained on our mental health dataset. Features with a $ symbol are part-of-speech tags (using our POS n-gram method). All individual word embedding features, obtained via a pre-trained Word2Vec embedding, are denoted as <W2V-X>, where X is the dimension index of the embedding vector. The POS tags shown are are as follows: $RB = adverb, $VBG = present participle verb, $VBN = past participle verb, $IN = preposition, $JJ = adjective, $NNS = plural noun, $PRP = personal pronoun, $VB = base form verb, $NNP = singular proper noun.", |
| "num": null |
| }, |
| "TABREF0": { |
| "type_str": "table", |
| "text": "Figure 1: Distribution of word counts per response for our collected dataset. On average, each response contains 39 words, with a minimum of two words and a maximum of 762 words. 30 responses had more than 200 words, which we do not show.", |
| "num": null, |
| "content": "<table><tr><td colspan=\"5\">Annotator POS NEG BOTH NEITHER</td></tr><tr><td colspan=\"2\">Annotator 1 494</td><td>2569</td><td>556</td><td>402</td></tr><tr><td colspan=\"2\">Annotator 2 321</td><td>2509</td><td>552</td><td>638</td></tr><tr><td colspan=\"2\">Annotator 3 531</td><td>2152</td><td>383</td><td>954</td></tr><tr><td>Final</td><td>414</td><td>2545</td><td>510</td><td>548</td></tr></table>", |
| "html": null |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "text": "Label counts per annotator, as well as the the final dataset label counts obtained via a majority-voting scheme. For brevity, we denote the positive label as POS, negative as NEG, both positive and negative as BOTH, and neither positive nor negative as NEITHER.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF6": { |
| "type_str": "table", |
| "text": "Polarity prediction results for the full 4-class version of our dataset. For brevity, the polarity class both positive and negative is denoted as Both, and the class neither positive nor negative is denoted as Neither.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF7": { |
| "type_str": "table", |
| "text": "al.,", |
| "num": null, |
| "content": "<table><tr><td/><td>1.0</td><td/><td/><td colspan=\"3\">Binary Polarity Prediction with External Corpora</td></tr><tr><td/><td/><td/><td/><td>Accuracy</td><td>Average Precision</td><td colspan=\"2\">Average Recall</td><td>Average F1</td></tr><tr><td/><td>0.9</td><td/><td/><td/><td/><td/></tr><tr><td>Mean Value</td><td>0.7 0.8</td><td/><td/><td/><td/><td/></tr><tr><td/><td>0.6</td><td/><td/><td/><td/><td/></tr><tr><td/><td>0.5</td><td>Twitter</td><td>Stanford</td><td>Cornell</td><td>Cornell Sentence</td><td>UMich</td><td>TAO 2-Class</td></tr><tr><td colspan=\"8\">Figure 4: Classification results for the positive vs. negative prediction setting using 5 external sentiment corpora and cross-</td></tr><tr><td colspan=\"8\">validated results on our own binary dataset (TAO 2-Class). The precision, recall, and F1 scores are reported using a weighted</td></tr><tr><td colspan=\"8\">average incorporating the support of each class label. For all metrics, training on our dataset (TAO 2-Class) yields better results</td></tr><tr><td colspan=\"5\">than using models trained on existing sentiment corpora.</td><td/><td/></tr><tr><td/><td/><td>Dataset</td><td colspan=\"2\"># Positive # Negative</td><td/><td/></tr><tr><td/><td/><td>Twitter</td><td>797792</td><td>798076</td><td/><td/></tr><tr><td/><td/><td>Stanford</td><td>25000</td><td>25000</td><td/><td/></tr><tr><td/><td/><td>Cornell</td><td>1000</td><td>1000</td><td/><td/></tr><tr><td/><td colspan=\"2\">Cornell Sentence</td><td>5221</td><td>5212</td><td/><td/></tr><tr><td/><td/><td>UMich</td><td>3995</td><td>3091</td><td/><td/></tr></table>", |
| "html": null |
| }, |
| "TABREF8": { |
| "type_str": "table", |
| "text": "Existing sentiment corpora summary.", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF10": { |
| "type_str": "table", |
| "text": "", |
| "num": null, |
| "content": "<table/>", |
| "html": null |
| } |
| } |
| } |
| } |