| { |
| "paper_id": "S16-1030", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T15:27:04.322253Z" |
| }, |
| "title": "SENSEI-LIF at SemEval-2016 Task 4: Polarity embedding fusion for robust sentiment analysis", |
| "authors": [ |
| { |
| "first": "Mickael", |
| "middle": [], |
| "last": "Rouvier", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "mickael.rouvier@lif.univ-mrs.fr" |
| }, |
| { |
| "first": "Benoit", |
| "middle": [], |
| "last": "Favre", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "benoit.favre@lif.univ-mrs.fr" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper describes the system developed at LIF for the SemEval-2016 evaluation campaign. The goal of Task 4.A was to identify sentiment polarity in tweets. The system extends the Convolutional Neural Networks (CNN) state of the art approach. We initialize the input representations with embeddings trained on different units: lexical, partof-speech, and sentiment embeddings. Neural networks for each input space are trained separately, and then the representations extracted from their hidden layers are concatenated as input of a fusion neural network. The system ranked 2nd at SemEval-2016 and obtained an average F1 of 63.0%.", |
| "pdf_parse": { |
| "paper_id": "S16-1030", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper describes the system developed at LIF for the SemEval-2016 evaluation campaign. The goal of Task 4.A was to identify sentiment polarity in tweets. The system extends the Convolutional Neural Networks (CNN) state of the art approach. We initialize the input representations with embeddings trained on different units: lexical, partof-speech, and sentiment embeddings. Neural networks for each input space are trained separately, and then the representations extracted from their hidden layers are concatenated as input of a fusion neural network. The system ranked 2nd at SemEval-2016 and obtained an average F1 of 63.0%.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "This paper describes the system developed at LIF for the SemEval-2016 sentiment analysis evaluation task (Nakov et al., 2016) . The goal of our participation was to apply approaches developed for the European FP7 project SENSEI 1 based on the study of human conversations according to feelings, opinions, emotions of the participants, in corpora such as transcripts of telephone speech and web comments.", |
| "cite_spans": [ |
| { |
| "start": 105, |
| "end": 125, |
| "text": "(Nakov et al., 2016)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We have participated in Subtask A: sentiment analysis at the message level. It consists in determining the message polarity of each tweet in the test set. The sentiment polarity classification task is set as a three-class problem: positive, negative and neutral.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The sentiment analysis task is often modeled as a classification problem which relies on features ex-tracted from the text in order to feed a classifier. Recent work has shown that Convolutional Neural Networks (CNN) using word representations as input are well suited for sentence classification problems (Kim, 2014) and have been shown to produce state-of-the-art results for sentiment polarity classification (Tang et al., 2014a; Severyn and Moschitti, 2015) . Pre-trained word embeddings are used to initialize the word representations, which are then taken as input of a text CNN.", |
| "cite_spans": [ |
| { |
| "start": 306, |
| "end": 317, |
| "text": "(Kim, 2014)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 412, |
| "end": 432, |
| "text": "(Tang et al., 2014a;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 433, |
| "end": 461, |
| "text": "Severyn and Moschitti, 2015)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our approach consists in learning polarity classifiers for three types of embeddings, based on the same CNN architecture. Each set of word embedding models the tweet according to a different point of view: lexical, part-of-speech and sentiment. A final fusion step is applied, based on concatenating the hidden layers of the CNNs and training a deep neural network for the fusion.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our contributions are as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We extend the deep CNN architecture proposed in (Poria et al., 2015) and introduce lexical information similar to (Ebert et al., 2015) .", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 70, |
| "text": "(Poria et al., 2015)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 116, |
| "end": 136, |
| "text": "(Ebert et al., 2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We introduce polarity embeddings, tweet representations extracted from the hidden layer of CNNs with different word embeddings as input.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 We fuse polarity embeddings by concatenating them and feeding them to a neural network trained on the final task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "\u2022 The source code of our system, the models trained for the evaluation, and the corpus collected for creating word embeddings are made available to the community to help future research 2 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The paper is structured as follows. Section 2 presents the system architecture. Section 3 reviews the implementation details. Then we detail the different word embeddings and other features used in our system (Section 4). Results and discussion appear in Section 5.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Deep learning models have been shown to produce state-of-the-art performance in various domains (vision, speech, etc...) . Convolutional Neural Networks (CNN) represent one of the most used deep learning model in computer vision (LeCun and Bengio, 1995) . Recent work has shown that CNNs are also well suited for sentence classification problems and can produce state-of-the-art results (Tang et al., 2014a; Severyn and Moschitti, 2015) . The difference between CNNs applied to computer vision and their equivalent in NLP lies in the input dimensionality and format. In computer vision, inputs are usually single-channel (eg. grayscale) or multi-channel (eg. RGB) 2D or 3D matrices, usually of constant dimension. In sentence classification, each input consists of a sequence of words of variable length. Each word w is represented with a n-dimensional vector (word embedding) e w of constant size. All the word representations are then concatenated in their respective order and padded with zero-vectors to a fixed length (maximum possible length of the sentence).", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 120, |
| "text": "(vision, speech, etc...)", |
| "ref_id": null |
| }, |
| { |
| "start": 229, |
| "end": 253, |
| "text": "(LeCun and Bengio, 1995)", |
| "ref_id": null |
| }, |
| { |
| "start": 387, |
| "end": 407, |
| "text": "(Tang et al., 2014a;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 408, |
| "end": 436, |
| "text": "Severyn and Moschitti, 2015)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Polarity embeddings", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Word embeddings are an approach for distributional semantics which represents words as vectors of real numbers. Such representation has useful clustering properties, since it groups together words that are semantically and syntactically similar (Mikolov et al., 2013) . For example, the word \"coffee\" and \"tea\" will be very close in the created space. The goal is to use these features as input to a CNN classifier. However, with the sentiment analysis task in mind, typical word embeddings extracted from lexical context might not be the most accurate because antonyms tend to be placed at the same location in the created space. As exemplified in Table 1 , \"good\" and \"bad\" occur in similar con-texts, and therefore obtain very similar representations. In addition, the model does not differentiate the senses of a word and creates a representation close to the most used sense in the training data. In order to tackle the representation robustness problem, we propose to extract word embeddings with different training regimes, and fuse their contribution to the system.", |
| "cite_spans": [ |
| { |
| "start": 245, |
| "end": 267, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 649, |
| "end": 656, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Polarity embeddings", |
| "sec_num": "2" |
| }, |
| { |
| "text": "embeddings% Sen-ment% embeddings%", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "w 1# w 2# w 3# w i'1# w i# w 1# w 2# w 3# w i'1# w i# w 1# w 2# w 3# w i'1# w i# CNN# Output' Output' CNN# Output' CNN# Output' 28level% 18level%", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "Input' Figure 1 : Overview of the sentiment embedding fusion approach. In a first level, different word representations are used to train CNNs to predict sentiment polarity, in the second level, representations extracted at hidden layers are concatenated and fed to a final classifier.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 7, |
| "end": 15, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "There are two approaches commonly used for fusion: early and late fusion. Late fusion considers that the different systems are independent by first applying classification separately on each system and then merging the output using a high-level classifier. Unfortunately, the classifier cannot model the correlations among modalities. The early fusion approach tackles this problem by learning features and class relationships to model the interaction between modalities. While late fusion cannot benefit from different system feature correlations, early fusion requires lots of training data.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "In previous work (Rouvier et al., 2015) , we have proposed a new fusion framework called embedding fusion which consists in concatenating hidden layers of subsystems trained independently, and input them to an other classifier trained to the actual task targets. This embedding fusion approach goes beyond late fusion and overcomes most of the problems linked to early fusion.", |
| "cite_spans": [ |
| { |
| "start": 17, |
| "end": 39, |
| "text": "(Rouvier et al., 2015)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper, we apply the embedding fusion to Table 1 : Closest words to \"good\" and \"bad\" according to different regimes for creating word embeddings: lexical, part-of-speech and sentiment (described later in the paper). the sentiment polarity prediction task, with a twolevel architecture (Figure 1 ). Given a tweet, the first level extracts input representations based different word embeddings. These embeddings are fed to a CNN with n-gram filters (from 1 to 5). The CNN is followed by a series of fully connected hidden layers which are trained to predict the target (sentiment polarity). Three different sets of word embeddings are used: lexical embeddings, joint lexical-part-ofspeech embeddings, and joint lexical-sentiment embeddings. The training procedure for the embeddings is explained in the following sections. The second level inputs the concatenation of the last hidden layer resulting from each input representation, which we call polarity embeddings. This representation is fed to fully connected hidden layers, and also trained to predict the polarity target. This method allows us to take advantage of both early and late fusion at the same time, which brings an improvement in term of performance over merging the decisions of the independent neural networks.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 48, |
| "end": 55, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 292, |
| "end": 301, |
| "text": "(Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Part8of8speech%", |
| "sec_num": null |
| }, |
| { |
| "text": "The proposed architecture relies on word embeddings as word representation as well as sentiment polarity lexicon features, concatenated to the word representation. An alternative to word-level features captured by CNNs is to extract sentence-level features in order to model global evidence in the tweet. In order to incorporate this source of information into the system, a classical MLP with one hidden layer is trained to predict sentiment polarity from a set of sentence-level features and its hidden layer is concatenated to the other polarity embeddings and fed to the second-level MLP. The CNN and MLP are trained jointly.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Implementation details", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The final complete architecture including CNNs and the sentence-level MLP, presented in Figure 2 , is based on a single convolutional layer followed by a max-over-time pooling layer (Collobert et al., 2011) and two fully-connected layers. In order to learn this kind of model there are two soft-max fully connected layers. The first one is connected to the pooling layer and the second one at the end of fullyconnected layer. The parameters of our model were chosen so as to maximize performance on the development set: the width of the convolution filters is set to 5 and the number of convolutional feature maps is 500. We use ReLU activation functions and a simple maxpooling. The two fully connected hidden-layers are of size 512. For each layer, a standard dropout of 0.4 (40 % of the neurons are disabled in each iteration) is used. The back-propagation algorithm used for training is Adadelta. In our experiments we observed that the weight initialization of the convolution layer can lead to high variation in term of performance. Therefore, we trained 20 models and selected the one that obtained the best results on the development corpus.", |
| "cite_spans": [ |
| { |
| "start": 182, |
| "end": 206, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 88, |
| "end": 96, |
| "text": "Figure 2", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Implementation details", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Feature%Lexicon%@%Word%level% Feature%Lexicon% @%Sentence%level% w 1# w 2# w 3# w i'1# w i# CNN# Output' Word%Representa-on% Sentence%Composi-on% Sentence%Representa-on%", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word%embeddings%", |
| "sec_num": null |
| }, |
| { |
| "text": "In the second part of the system which inputs polarity embeddings and predicts polarity targets, the DNN is composed of two 512-dimensional hidden layers. The non-linearity of the hidden layers is corrected by a ReLU function.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word%embeddings%", |
| "sec_num": null |
| }, |
| { |
| "text": "We propose to make use of word embeddings trained under different regimes, in order to capture different aspects of the relation between words so that it might benefit the polarity classifier. Three representations are explored.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Input features 4.1 Word embeddings", |
| "sec_num": "4" |
| }, |
| { |
| "text": "these embeddings are obtained with the classical skipgram model from (Mikolov et al., 2013) . The representation is created by using the hidden layer of a linear neural network to predict a context window from a central word. For a given context w i\u22122 . . . w i+2 , the input to the model is w i , and the output could be w i\u22122 , w i\u22121 , w i+1 , w i+2 . This method typically extracts a representation which both covers syntax and semantics, to some extent. Part-of-speech embeddings: as stated earlier, the lexical model cannot distinguish between the senses of words and creates a single representation per word form. For example, the word \"apple\" receives an embedding that is a mixture of its different contextual senses: fruit, company... A lot of sophisticated approaches have been proposed to tackle the problem (Guo et al., 2014; Neelakantan et al., 2015; Huang et al., 2012) , by considering senses as latent variables during training, or by conditionning the training documents on topic distributions. In our system we follow a very simple approach which creates joint embeddings for words and their part of speech. Thus, for con-text w i\u22122 . . . w i+2 tagged with the part-of-speech sequence p i\u22122 . . . p i+2 the input to the model is (w i , p i ) and the output is", |
| "cite_spans": [ |
| { |
| "start": 69, |
| "end": 91, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 819, |
| "end": 837, |
| "text": "(Guo et al., 2014;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 838, |
| "end": 863, |
| "text": "Neelakantan et al., 2015;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 864, |
| "end": 883, |
| "text": "Huang et al., 2012)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical embeddings:", |
| "sec_num": null |
| }, |
| { |
| "text": "(w i\u22122 , p i\u22122 ), (w i\u22121 , p i\u22121 ), (w i+1 : p i+1 ), (w i+2 , p i+2 ).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical embeddings:", |
| "sec_num": null |
| }, |
| { |
| "text": "Sentiment embeddings: another problem with the basic skipgram approach (lexical embeddings) is that the model ignores the sentiment polarity of the words. As a result, words with opposite polarity, such as \"good\" and \"bad\", are mapped into close vectors. In (Tang et al., 2014b) , the authors propose to tackle this problem so that sentiment information is encoded in the continuous representation of words. They propose to create a neural network that predicts two tasks: the context of the word and the sentiment label of the whole sentence. Since it is expensive to manually label sentences with a polarity label, the authors propose to use tweets that contain emoticons and rely on the polarity of the emoticon to label the sentences. As they report that best performance is obtained by weighting both tasks equivalently, the model is the same as for lexical embeddings, except that the predicted context is formed of (word, sentiment) couples. For example, if s is the polarity of the sentence where the context w i\u22122 . . . w i+2 is extracted, the model gets w i as input and has to predict", |
| "cite_spans": [ |
| { |
| "start": 258, |
| "end": 278, |
| "text": "(Tang et al., 2014b)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical embeddings:", |
| "sec_num": null |
| }, |
| { |
| "text": "(w i\u22122 , s), (w i\u22121 , s), (w i+1 , s), (w i+2 , s).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lexical embeddings:", |
| "sec_num": null |
| }, |
| { |
| "text": "Word representations are learned from distributional information of words in large corpora. Although such statistics are semantically informative, they disregard the valuable information that is contained in manually curated sentiment lexicons. In (Ebert et al., 2015) , the authors propose to incorporate knowledge from semantic lexicons at the word level. The goal is to extract features based on the overlap between words in the input and sentiment lexicons, and stack these features to the word embedding.", |
| "cite_spans": [ |
| { |
| "start": 248, |
| "end": 268, |
| "text": "(Ebert et al., 2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment lexicon features", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We create two such features per word per lexicon. Both are binary indicators of positive and negative polarity of that word in the lexicons. The lexicons for this feature type are MPQA (Wiebe et al., 2005) , Opinion lexicon (Hu and Liu, 2004) , and NRC Emotion lexicon (Mohammad and Turney, 2013) . The NRC lexicons provide a score for each word instead of just a label. We replace the binary indicators by the scores.", |
| "cite_spans": [ |
| { |
| "start": 185, |
| "end": 205, |
| "text": "(Wiebe et al., 2005)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 224, |
| "end": 242, |
| "text": "(Hu and Liu, 2004)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 269, |
| "end": 296, |
| "text": "(Mohammad and Turney, 2013)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment lexicon features", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The following features are extracted at sentence level and used for training the sentence-level MLP:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence-level features", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "\u2022 Lexicons: frequency of lemmas that are matched in MPQA (Wiebe et al., 2005) , Opinion Lexicon (Hu and Liu, 2004) and NRC Emotion lexicon (Mohammad and Turney, 2013). \u2022 Emoticons: number of emoticons that are grouped in positive, negative and neutral categories. \u2022 All-caps: number of words in all-caps \u2022 Elongated units: number of words in which characters are repeated more than twice (for example: looooool) \u2022 Punctuation: number of contiguous sequences of period, exclamation mark and question mark.", |
| "cite_spans": [ |
| { |
| "start": 57, |
| "end": 77, |
| "text": "(Wiebe et al., 2005)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 96, |
| "end": 114, |
| "text": "(Hu and Liu, 2004)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentence-level features", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "A step of pre-processing is applied to every tweet in the corpus:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 Character encoding: every tweet is encoded in UTF-8 \u2022 XML Entities: all the XML entities are converted back to characters \u2022 Lowercase: all the characters are converted in lowercase \u2022 Lengthening: character lengthening consists in repeating several times a character in a word. It is used in social media as a method to emphasize a fact. This extension is often correlated with the expression of sentiment. If a character is repeated more than three times, we reduce it to three characters. For example, \"looool\" is replaced by \"loool\". \u2022 Tokenization: tokenization is performed by splitting a sentence in pre-lexical units. We used the tokenizer from the macaon toolchain (Nasr et al., 2011) . It is based on a regular grammar that defines a set of types of atoms. A lexical analyzer detects the character sequences (in terms of the grammar) and combines them as a type. We added the atoms for detecting smileys, hashtags and users names (atoms specific to tweets).", |
| "cite_spans": [ |
| { |
| "start": 674, |
| "end": 693, |
| "text": "(Nasr et al., 2011)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 Map generic words: The hashtags, numbers and usertags are mapped to generic tokens.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-processing", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We use the train and dev corpora from Twitter'13 to 16 for training and Twitter'16-dev as a development set. Note that we were unable to download all the training and development data because some tweets were deleted or not available due to modified authorization status. The datasets are summarized in ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "To train the word embeddings, we have created a unannotated corpus of sentiment bearing tweets in English. These tweets were recovered on the Twitter platform by searching for emotion keywords (from the sentiment lexicons) and unigrams, bigrams and trigrams extracted from the SemEval training corpus. This corpus consists of about 90 million tweets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word embedding training", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "A sub-corpus of about 20 million tweets containing at least one emoticon is used for training the sentiment embeddings. Both corpora are made available 3 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word embedding training", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "In our experiments, lexical embeddings and part-of-speech embeddings are estimated using the word2vec toolkit (Mikolov et al., 2013) . Sentiment embeddings are estimated using word2vecf. This toolkit allows to replace linear bag-of-word contexts with arbitrary features. The embeddings are trained using the skipgram approach with a window of size 3 and 5 iterations. The dimension of the embeddings is fixed to 100. Part-of-speech tagging is performed with Tweet NLP (Owoputi et al., 2013; Gimpel et al., 2011) .", |
| "cite_spans": [ |
| { |
| "start": 110, |
| "end": 132, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 468, |
| "end": 490, |
| "text": "(Owoputi et al., 2013;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 491, |
| "end": 511, |
| "text": "Gimpel et al., 2011)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word embedding training", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "Overall performance: The evaluation metric used in the competition is the macro-averaged F-measure calculated over the positive and negative categories. Table 4 presents the overall performance of our system. It achieved the second rank on the Twitter 2016", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 153, |
| "end": 160, |
| "text": "Table 4", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "Feature set", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "Part-of-speech Sentiment SENSEI-LIF all features 61.3 62.0 62.3 63.0 w/o word level lexicon 61.7 (+0.4) 62.4 (+0.4) 61.6 (-0.7) 63.2 (+0.2) w/o sentence level lexicon 60.7 (-0.6) 61.1 (-0.9) 62.0 (-0.3) 62.6 (-0.4) w/o both lexicon 61.0 (-0.3) 61.4 (-0.6) 61.8 (-0.5) 62.8 (-0.2) w/o word embeddings 58.4 (-2.9) 59.1 (-2.9) 59.6 (-2.7) 59.6 (-3.4) Table 2 presents the results of ablation experiments on the Twitter 2016 test set. SENSEI-LIF is the system which participated to the evaluation campaign. We present the results of three contrastive systems: Lexical, Partof-speech and Sentiment. These systems are based on the CNN classifier prior to the concatenation of the hidden layers. They use only one set of word embeddings without any kind of fusion. The different features used in our system are: lexicon features and word embeddings. The ablation of lexicon features removes the lexicon features at the word and sentence level. The ablation of word embeddings feature consists in randomly initializing the word representations.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 348, |
| "end": 355, |
| "text": "Table 2", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Lexical", |
| "sec_num": null |
| }, |
| { |
| "text": "We observe that the most influential features are word embeddings. They provide a gain of 3.4 points. The main advantage of word embeddings is to learn unsupervised representations on very large corpora which capture general semantic properties. The last most important features are lexicon features. We observe that word level lexicon features are not relevant and tend to degrade the performance of the SENSEI-LIF system on the Twitter 2016 dataset. Impact of fusion: Table 5 presents the results us-ing different kinds of fusion: early, late and embedding fusion. We observe that early fusion obtains the worse results. We think that is due to the small training corpus used. Embedding fusion obtains the bests results on the Twitter 2016 dataset, but more generally late and embedding fusions obtain very close results on the other datasets. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 470, |
| "end": 477, |
| "text": "Table 5", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Contribution of features:", |
| "sec_num": null |
| }, |
| { |
| "text": "This paper describes the LIF participation at Se-mEval 2016. Our approach consists in learning polarity classifiers for three types of embeddings, based on the same CNN architecture. Each set of word embeddings models the tweet according to a different point of view: lexical, part-of-speech and sentiment. A final fusion step is applied, based on concatenating the hidden layers of the CNNs and training a deep neural network for the fusion. The fusion system ranked 2nd at the SemEval-2016 evaluation campaign.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "6" |
| }, |
| { |
| "text": "http://www.sensei-conversation.eu/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.github.com/mrouvier/SemEval2016", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.github.com/mrouvier/SemEval2016", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.sensei-conversation.eu", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The research leading to these results has received funding from the European Union -Seventh Framework Programme (FP7/2007-2013) under grant agreement 610916 SENSEI 4 . The Tesla K40 used for this research was donated by the NVIDIA Corporation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "JMLR", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. JMLR, 12:2493-2537.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A linguistically informed convolutional neural network", |
| "authors": [ |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Ebert", |
| "suffix": "" |
| }, |
| { |
| "first": "Ngoc", |
| "middle": [ |
| "Thang" |
| ], |
| "last": "Vu", |
| "suffix": "" |
| }, |
| { |
| "first": "Hinrich", |
| "middle": [], |
| "last": "Sch\u00fctze", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Computational Approaches to Subjectivity, Sentimant and Social Media Analysis WASSA", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sebastian Ebert, Ngoc Thang Vu, and Hinrich Sch\u00fctze. 2015. A linguistically informed convolutional neural network. In Computational Approaches to Subjectiv- ity, Sentimant and Social Media Analysis WASSA, page 109.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Part-of-speech tagging for twitter: Annotation, features, and experiments", |
| "authors": [ |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Gimpel", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Schneider", |
| "suffix": "" |
| }, |
| { |
| "first": "O'", |
| "middle": [], |
| "last": "Brendan", |
| "suffix": "" |
| }, |
| { |
| "first": "Dipanjan", |
| "middle": [], |
| "last": "Connor", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Das", |
| "suffix": "" |
| }, |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Mills", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Eisenstein", |
| "suffix": "" |
| }, |
| { |
| "first": "Dani", |
| "middle": [], |
| "last": "Heilman", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Yogatama", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah A", |
| "middle": [], |
| "last": "Flanigan", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers", |
| "volume": "2", |
| "issue": "", |
| "pages": "42--47", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kevin Gimpel, Nathan Schneider, Brendan O'Connor, Dipanjan Das, Daniel Mills, Jacob Eisenstein, Michael Heilman, Dani Yogatama, Jeffrey Flanigan, and Noah A Smith. 2011. Part-of-speech tagging for twit- ter: Annotation, features, and experiments. In Pro- ceedings of the 49th Annual Meeting of the Associa- tion for Computational Linguistics: Human Language Technologies: short papers-Volume 2, pages 42-47. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Learning sense-specific word embeddings by exploiting bilingual resources", |
| "authors": [ |
| { |
| "first": "Jiang", |
| "middle": [], |
| "last": "Guo", |
| "suffix": "" |
| }, |
| { |
| "first": "Wanxiang", |
| "middle": [], |
| "last": "Che", |
| "suffix": "" |
| }, |
| { |
| "first": "Haifeng", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ting", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "497--507", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiang Guo, Wanxiang Che, Haifeng Wang, and Ting Liu. 2014. Learning sense-specific word embeddings by exploiting bilingual resources. In COLING, pages 497-507.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Mining and summarizing customer reviews", |
| "authors": [ |
| { |
| "first": "Minqing", |
| "middle": [], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "SIGKDD", |
| "volume": "", |
| "issue": "", |
| "pages": "168--177", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Minqing Hu and Bing Liu. 2004. Mining and summariz- ing customer reviews. In SIGKDD, pages 168-177.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Improving word representations via global context and multiple word prototypes", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Eric", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew Y", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers", |
| "volume": "1", |
| "issue": "", |
| "pages": "873--882", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representa- tions via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Asso- ciation for Computational Linguistics: Long Papers- Volume 1, pages 873-882. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Convolutional networks for images, speech, and time series. The handbook of brain theory and neural networks", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Convolutional neural networks for sentence classification", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1408.5882" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoon Kim. 2014. Convolutional neural networks for sen- tence classification. arXiv preprint arXiv:1408.5882. Yann LeCun and Yoshua Bengio. 1995. Convolu- tional networks for images, speech, and time series. The handbook of brain theory and neural networks, 3361(10).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1301.3781" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representa- tions in vector space. arXiv preprint arXiv:1301.3781.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "SemEval-2016 task 4: Sentiment analysis in Twitter", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Saif", |
| "suffix": "" |
| }, |
| { |
| "first": "Peter D Turney ; Alan", |
| "middle": [], |
| "last": "Mohammad", |
| "suffix": "" |
| }, |
| { |
| "first": "Sara", |
| "middle": [], |
| "last": "Ritter", |
| "suffix": "" |
| }, |
| { |
| "first": "Veselin", |
| "middle": [], |
| "last": "Rosenthal", |
| "suffix": "" |
| }, |
| { |
| "first": "Fabrizio", |
| "middle": [], |
| "last": "Stoyanov", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sebastiani", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval '16", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Saif M Mohammad and Peter D Turney. 2013. Nrc emo- tion lexicon. Technical report, NRC Technical Report. Preslav Nakov, Alan Ritter, Sara Rosenthal, Veselin Stoy- anov, and Fabrizio Sebastiani. 2016. SemEval-2016 task 4: Sentiment analysis in Twitter. In Proceedings of the 10th International Workshop on Semantic Eval- uation, SemEval '16, San Diego, California, June. As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Macaon: An nlp tool suite for processing word lattices", |
| "authors": [ |
| { |
| "first": "Alexis", |
| "middle": [], |
| "last": "Nasr", |
| "suffix": "" |
| }, |
| { |
| "first": "Fr\u00e9d\u00e9ric", |
| "middle": [], |
| "last": "B\u00e9chet", |
| "suffix": "" |
| }, |
| { |
| "first": "Jean-Fran\u00e7ois", |
| "middle": [], |
| "last": "Rey", |
| "suffix": "" |
| }, |
| { |
| "first": "Beno\u00eet", |
| "middle": [], |
| "last": "Favre", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [ |
| "Le" |
| ], |
| "last": "Roux", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "86--91", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alexis Nasr, Fr\u00e9d\u00e9ric B\u00e9chet, Jean-Fran\u00e7ois Rey, Beno\u00eet Favre, and Joseph Le Roux. 2011. Macaon: An nlp tool suite for processing word lattices. In ACL, pages 86-91.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Efficient non-parametric estimation of multiple embeddings per word in vector space", |
| "authors": [ |
| { |
| "first": "Arvind", |
| "middle": [], |
| "last": "Neelakantan", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeevan", |
| "middle": [], |
| "last": "Shankar", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandre", |
| "middle": [], |
| "last": "Passos", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1504.06654" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arvind Neelakantan, Jeevan Shankar, Alexandre Pas- sos, and Andrew McCallum. 2015. Effi- cient non-parametric estimation of multiple embed- dings per word in vector space. arXiv preprint arXiv:1504.06654.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Improved part-of-speech tagging for online conversational text with word clusters. Association for Computational Linguistics", |
| "authors": [ |
| { |
| "first": "Olutobi", |
| "middle": [], |
| "last": "Owoputi", |
| "suffix": "" |
| }, |
| { |
| "first": "O'", |
| "middle": [], |
| "last": "Brendan", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Connor", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Gimpel", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah A", |
| "middle": [], |
| "last": "Schneider", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Olutobi Owoputi, Brendan O'Connor, Chris Dyer, Kevin Gimpel, Nathan Schneider, and Noah A Smith. 2013. Improved part-of-speech tagging for online conversa- tional text with word clusters. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis", |
| "authors": [ |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [], |
| "last": "Gelbukh", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "2539--2544", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soujanya Poria, Erik Cambria, and Alexander Gelbukh. 2015. Deep convolutional neural network textual fea- tures and multiple kernel learning for utterance-level multimodal sentiment analysis. In Proceedings of EMNLP, pages 2539-2544.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Multimodal embedding fusion for robust speaker role recognition in video broadcast", |
| "authors": [ |
| { |
| "first": "Mickael", |
| "middle": [], |
| "last": "Rouvier", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastien", |
| "middle": [], |
| "last": "Delecraz", |
| "suffix": "" |
| }, |
| { |
| "first": "Meriem", |
| "middle": [], |
| "last": "Benoit Favre", |
| "suffix": "" |
| }, |
| { |
| "first": "Frederic", |
| "middle": [], |
| "last": "Bendris", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bechet", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "ASRU", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mickael Rouvier, Sebastien Delecraz, Benoit Favre, Meriem Bendris, and Frederic Bechet. 2015. Multi- modal embedding fusion for robust speaker role recog- nition in video broadcast. In ASRU.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Unitn: Training deep convolutional neural network for twitter sentiment classification", |
| "authors": [ |
| { |
| "first": "Aliaksei", |
| "middle": [], |
| "last": "Severyn", |
| "suffix": "" |
| }, |
| { |
| "first": "Alessandro", |
| "middle": [], |
| "last": "Moschitti", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 9th International Workshop on Semantic Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "464--469", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aliaksei Severyn and Alessandro Moschitti. 2015. Unitn: Training deep convolutional neural network for twitter sentiment classification. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), Association for Computational Lin- guistics, Denver, Colorado, pages 464-469.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Coooolll: A deep learning system for twitter sentiment classification", |
| "authors": [ |
| { |
| "first": "Duyu", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ting", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 8th International Workshop on Semantic Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "208--212", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Duyu Tang, Furu Wei, Bing Qin, Ting Liu, and Ming Zhou. 2014a. Coooolll: A deep learning system for twitter sentiment classification. In Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), pages 208-212.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Learning sentiment-specific word embedding for twitter sentiment classification", |
| "authors": [ |
| { |
| "first": "Duyu", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Nan", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Ting", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "ACL (1)", |
| "volume": "", |
| "issue": "", |
| "pages": "1555--1565", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, and Bing Qin. 2014b. Learning sentiment-specific word embedding for twitter sentiment classification. In ACL (1), pages 1555-1565.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Annotating expressions of opinions and emotions in language. Language resources and evaluation", |
| "authors": [ |
| { |
| "first": "Janyce", |
| "middle": [], |
| "last": "Wiebe", |
| "suffix": "" |
| }, |
| { |
| "first": "Theresa", |
| "middle": [], |
| "last": "Wilson", |
| "suffix": "" |
| }, |
| { |
| "first": "Claire", |
| "middle": [], |
| "last": "Cardie", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "", |
| "volume": "39", |
| "issue": "", |
| "pages": "165--210", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Janyce Wiebe, Theresa Wilson, and Claire Cardie. 2005. Annotating expressions of opinions and emotions in language. Language resources and evaluation, 39(2- 3):165-210.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "uris": null, |
| "text": "Actual CNN architecture: word representations are concatenated with lexicon features, and sentence-level lexicon features are concatenated with the polarity embeddings, and also trained to predict polarity targets on its own.", |
| "num": null, |
| "type_str": "figure" |
| }, |
| "TABREF1": { |
| "num": null, |
| "content": "<table><tr><td>Corpus Positive Negative Neutral Train 7.727 2.916 7.837 18.480 Total Dev 884 279 616 1.779</td></tr></table>", |
| "text": "", |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF2": { |
| "num": null, |
| "content": "<table/>", |
| "text": "Statistics of the successfully downloaded part of the SemEval 2016 Twitter sentiment classification dataset.", |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF3": { |
| "num": null, |
| "content": "<table><tr><td colspan=\"2\">data among 34 teams. The system proved to gener-</td></tr><tr><td colspan=\"2\">alize well to other types of short informal texts; it</td></tr><tr><td colspan=\"2\">ranked first and third respectively on the two out-of-</td></tr><tr><td colspan=\"2\">domain datasets: Live Journal 2014 and SMS 2013.</td></tr><tr><td>Corpus Twt2013 SMS2013 Twt2014 TwtSarc2014 LvJn2014 Twt2015 Twt2016</td><td>SENSEI-LIF Rank 70.6 3 63.4 3 74.4 1 46.7 8 74.1 1 66.2 2 63.0 2</td></tr></table>", |
| "text": "Ablation experiment: macro-averaged F-scores obtained on the Twitter 2016 test sets with each of the feature groups removed.", |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF4": { |
| "num": null, |
| "content": "<table/>", |
| "text": "Overall performance of the SENSEI-LIF sentiment analysis systems.", |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF6": { |
| "num": null, |
| "content": "<table/>", |
| "text": "Overall performance using different methods of fusion: early, late and embedding fusion.", |
| "type_str": "table", |
| "html": null |
| } |
| } |
| } |
| } |