| { |
| "paper_id": "K16-1017", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:11:03.288424Z" |
| }, |
| "title": "Modelling Context with User Embeddings for Sarcasm Detection in Social Media", |
| "authors": [ |
| { |
| "first": "Silvio", |
| "middle": [], |
| "last": "Amir", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "samir@inesc-id.pt" |
| }, |
| { |
| "first": "Byron", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wallace", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "byron.wallace@utexas.edu" |
| }, |
| { |
| "first": "Hao", |
| "middle": [], |
| "last": "Lyu", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Paula", |
| "middle": [], |
| "last": "Carvalho", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "M\u00e1rio", |
| "middle": [ |
| "J" |
| ], |
| "last": "Silva", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We introduce a deep neural network for automated sarcasm detection. Recent work has emphasized the need for models to capitalize on contextual features, beyond lexical and syntactic cues present in utterances. For example, different speakers will tend to employ sarcasm regarding different subjects and, thus, sarcasm detection models ought to encode such speaker information. Current methods have achieved this by way of laborious feature engineering. By contrast, we propose to automatically learn and then exploit user embeddings, to be used in concert with lexical signals to recognize sarcasm. Our approach does not require elaborate feature engineering (and concomitant data scraping); fitting user embeddings requires only the text from their previous posts. The experimental results show that the our model outperforms a state-of-the-art approach leveraging an extensive set of carefully crafted features.", |
| "pdf_parse": { |
| "paper_id": "K16-1017", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We introduce a deep neural network for automated sarcasm detection. Recent work has emphasized the need for models to capitalize on contextual features, beyond lexical and syntactic cues present in utterances. For example, different speakers will tend to employ sarcasm regarding different subjects and, thus, sarcasm detection models ought to encode such speaker information. Current methods have achieved this by way of laborious feature engineering. By contrast, we propose to automatically learn and then exploit user embeddings, to be used in concert with lexical signals to recognize sarcasm. Our approach does not require elaborate feature engineering (and concomitant data scraping); fitting user embeddings requires only the text from their previous posts. The experimental results show that the our model outperforms a state-of-the-art approach leveraging an extensive set of carefully crafted features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Existing social media analysis systems are hampered by their inability to accurately detect and interpret figurative language. This is particularly relevant in domains like the social sciences and politics, in which the use of figurative communication devices such as verbal irony (roughly, sarcasm) is common. Sarcasm is often used by individuals to express opinions on complex matters and regarding specific targets (Carvalho et al., 2009) .", |
| "cite_spans": [ |
| { |
| "start": 418, |
| "end": 441, |
| "text": "(Carvalho et al., 2009)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Early computational models for verbal irony and sarcasm detection tended to rely on shallow methods exploiting conditional token count regularities. But lexical clues alone are insufficient to discern ironic intent. Appreciating the context of utterances is critical for this; even for humans (Wallace et al., 2014) . Indeed, the exact same sentence can be interpreted as literal or sarcastic, depending on the speaker. Consider the sarcastic tweet in Figure 1 (ignoring for the moment the attached #sarcasm hashtag). Without knowing the author's political leanings, it would be difficult to conclude with certainty whether the remark was intended sarcastically or in earnest.", |
| "cite_spans": [ |
| { |
| "start": 293, |
| "end": 315, |
| "text": "(Wallace et al., 2014)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 452, |
| "end": 460, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Recent work in sarcasm detection on social media has tried to incorporate contextual information by exploiting the preceding messages of a user, to e.g., detect contrasts in sentiments expressed towards named entities (Khattri et al., 2015) , infer behavioural traits (Rajadesingan et al., 2015) and capture the relationship between authors and the audience (Bamman and Smith, 2015) . However, all of these approaches require the design and implementation of complex features that explicitly encode the content and (relevant) context of messages to be classified. This feature engineering is labor intensive, and depends on external tools and resources. Therefore, deploying such systems in practice is expensive, time-consuming and unwieldy.", |
| "cite_spans": [ |
| { |
| "start": 218, |
| "end": 240, |
| "text": "(Khattri et al., 2015)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 268, |
| "end": 295, |
| "text": "(Rajadesingan et al., 2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 358, |
| "end": 382, |
| "text": "(Bamman and Smith, 2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We propose a novel approach to sarcasm detection on social media that does not require extensive manual feature engineering. Instead, we develop a neural model that learns to represent and exploit embeddings of both content and context. For the former, we induce vector lexical repre-sentations via a convolutional layer; for the latter, our model learns user embeddings. Inference concerning whether an utterance (tweet) was intended ironically (or not) is then modelled as a joint function of lexical representations and corresponding author embeddings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The main contributions of this paper are as follows. (i) We propose a novel convolutional neural network based model that explicitly learns and exploits user embeddings in conjunction with features derived from utterances. (ii) We show that this model outperforms the strong baseline recently proposed by Bamman and Smith (2015) by more than 2% in absolute accuracy, while obviating the need to manually engineer features. (iii) We show that the learned user embeddings can capture relevant user attributes.", |
| "cite_spans": [ |
| { |
| "start": 305, |
| "end": 328, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 423, |
| "end": 428, |
| "text": "(iii)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Verbal irony is a rhetorical device in which speakers say something other than, and often opposite to, what they actually mean. 1 Sarcasm may be viewed as a special case of irony, where the positive literal meaning is perceived as an indirect insult (Dews et al., 1995) .", |
| "cite_spans": [ |
| { |
| "start": 250, |
| "end": 269, |
| "text": "(Dews et al., 1995)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Most of the previously proposed computational models to detect irony and sarcasm have used features similar to those used in sentiment analysis. Carvalho et al. (2009) analyzed comments posted by users on a Portuguese online newspaper and found that oral and gestural cues indicate irony. These included: emoticons, onomatopoeic expressions for laughter, heavy punctuation, quotation marks and positive interjections. Others have used text classifiers with features based on word and character n-grams, sentiment lexicons, surface patterns and textual markers (Davidov et al., 2010; Gonz\u00e1lez-Ib\u00e1nez et al., 2011; Reyes et al., 2013; Lukin and Walker, 2013) . Elsewhere, Barbieri and Saggion (2014) derived new wordfrequency based features to detect irony, e.g., combinations of frequent and rare words, ambiguous words, 'spoken style' words combined with 'written style' words and intensity of adjectives. Riloff et al. (2013) demonstrated that one may exploit the apparent expression of contrasting sentiment in the same utterance as a marker of verbal irony.", |
| "cite_spans": [ |
| { |
| "start": 145, |
| "end": 167, |
| "text": "Carvalho et al. (2009)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 560, |
| "end": 582, |
| "text": "(Davidov et al., 2010;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 583, |
| "end": 612, |
| "text": "Gonz\u00e1lez-Ib\u00e1nez et al., 2011;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 613, |
| "end": 632, |
| "text": "Reyes et al., 2013;", |
| "ref_id": "BIBREF26" |
| }, |
| { |
| "start": 633, |
| "end": 656, |
| "text": "Lukin and Walker, 2013)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 670, |
| "end": 697, |
| "text": "Barbieri and Saggion (2014)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 906, |
| "end": 926, |
| "text": "Riloff et al. (2013)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The aforementioned approaches rely predominantly on features intrinsic to texts, but these will often be insufficient to infer figurative meaning: context is needed. There have been some recent attempts to exploit contextual information, e.g. Khattri et al. (2015) extended the notion of contrasting sentiments beyond the textual content at hand. In particular, they analyzed previous posts to estimate the author's prior sentiment towards specific targets (i.e., entities). A tweet is then predicted to be sarcastic if it expresses a sentiment about an entity that contradicts the author's (estimated) prior sentiment regarding the same. Rajadesingan et al. (2015) built a system based on theories of sarcasm expression from psychology and behavioral sciences. To operationalize such theories, they used several linguistic tools and resources (e.g. lexicons, sentiment classifiers and a PoS tagger), in addition to user profile information and previous posts, to model a range of behavioural aspects (e.g., mood, writing style). developed an approach for classifying posts on reddit 2 as sarcastic or literal, based in part on the interaction between the specific sub-reddit to which a post was made, the entities mentioned, and the (apparent) sentiment expressed. For example, if a post in the (politically) conservative sub-reddit mentions Obama, it is more likely to have been intended ironically than posts mentioning Obama in the progressive sub-reddit. But this approach is limited because it relies on the unique sub-reddit structure. Bamman and Smith (2015) proposed an approach that relied on an extensive, rich set of features capturing various contextual information about authors of tweets and the audience (in addition to lexical cues). We review these at length in Section 5.1.", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 264, |
| "text": "Khattri et al. (2015)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 639, |
| "end": 665, |
| "text": "Rajadesingan et al. (2015)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 1543, |
| "end": 1566, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "A major downside of these and related approaches, however, is the amount of manual effort required to derive these feature sets. A primary goal of this work is to explore whether neural models can effectively learn these rich contextualizing features, thus obviating the need to manually craft them. In particular, the model we propose similarly aims to combine lexical clues with extralinguistic information. Unlike prior work, however, our model attempts to automatically induce representations for the content and the author of a message that are predictive of sarcasm.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our goal is to learn representations (vectors) that encode latent aspects of users and capture homophily, by projecting similar users into nearby regions of the embedding space. We hypothesize that such representations will naturally capture some of the signals that have been described in the literature as important indicators of sarcasm, such as contrasts between what someone believes and what they have ostensibly expressed (Campbell and Katz, 2012) or Kreuz (1996) principle of inferability, stating that sarcasm requires a common ground between parties to be understood.", |
| "cite_spans": [ |
| { |
| "start": 429, |
| "end": 454, |
| "text": "(Campbell and Katz, 2012)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 458, |
| "end": 470, |
| "text": "Kreuz (1996)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "To induce the user embeddings, we adopt an approach similar to that described in the preliminary work of Li et al. (2015) . In particular, we capture relations between users and the content they produce by optimizing the conditional probability of texts, given their authors (or, more precisely, given the vector representations of their authors). This method is akin to Le and Mikolov (2014)'s Paragraph Vector model, which jointly estimates embeddings for words and paragraphs by learning to predict the occurrence of a word w within a paragraph p conditioned on the (learned) representation for p.", |
| "cite_spans": [ |
| { |
| "start": 105, |
| "end": 121, |
| "text": "Li et al. (2015)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Given a sentence S = {w 1 , . . . , w N } where w i denotes a word drawn from a vocabulary V, we aim to maximize the following probability:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (S|user j ) = w i \u2208S log P (w i |u j ) + w i \u2208S w k \u2208C(w i ) log P (w i |e k )", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Where C(w i ) denotes the set of words in a prespecified window around word w i , e k \u2208 R d and u j \u2208 R d denote the embeddings of word k and user j, respectively. This objective function encodes the notion that the occurrence of a word w, depends both on the author of S and it's neighbouring words. The conditional probabilities in Equation 1 can be estimated with log-linear models of the form:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (w i |x) = exp(W i \u2022 x + b i ) Y k=1 exp(W k \u2022 x + b k )", |
| "eq_num": "(2)" |
| } |
| ], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Where x denotes a feature vector, W k and b k are the weight vectors and bias for class k. In our case, we treat words as classes to be predicted.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Calculating the denominator thus requires summing over all of the words in the (large) vocabulary, an expensive operation. To avoid this computational bottleneck, we approximate the term P (w i |e k ) with Morin and Bengio (2005) Hierarchical Softmax. 3 To learn meaningful user embeddings, we seek representations that are predictive of individual word-usage patterns. In light of this motivation, we approximate P (w i |u j ) via the following hinge-loss objective which we aim to minimize:", |
| "cite_spans": [ |
| { |
| "start": 206, |
| "end": 229, |
| "text": "Morin and Bengio (2005)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 252, |
| "end": 253, |
| "text": "3", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "L(w i , user j ) = w l \u2208V,w l \u2208S max(0, 1 \u2212 e i \u2022 u j + e l \u2022 u j ) (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Here, each w l (and corresponding embedding, e l ) is a negative example, i.e., a word not in the sentence under consideration, which was authored by user j. The intuition is that in the aggregate, such words are less likely to be employed by user j than words observed in sentences she has authored. Thus minimizing this objective attempts to induce a representation that is discriminative with respect to word usage.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In practice, V will be very large and hence we approximate the objective via negative sampling, a variant of Noise Contrastive Estimation. The idea is to approximate the objective function in a binary classification task by learning to discriminate between observed positive examples (sampled from the true distribution) and pseudo-negative examples (sampled from a large space of predominantly negative instances). Intuitively, this shifts probability mass to plausible observations. See Dyer (2014) for notes on Negative Sampling and Noise Contrastive Estimation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Previous work by Collobert et al. (2011) showed that this approach works well in representation learning tasks when a sufficient amount of training data is available . However, we have access to only a limited amount of text for each user (see Section 5). We hypothesize that this problem can be alleviated by carefully selecting the negative samples that mostly contribute to 'push' the vectors into the appropriate region of the embedding space (i.e., closer to the words commonly employed by a given user and far from other words).", |
| "cite_spans": [ |
| { |
| "start": 17, |
| "end": 40, |
| "text": "Collobert et al. (2011)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "This requires designing a strategy for selectively sampling negative examples. One straightforward approach would be to sample from a user-specific unigram model, informing which words are less likely to be utilized by that user. But estimating the parameters of such model with scarce data would be prone to overfitting. Instead, we sample from a unigram distribution estimated from the posts authored by all the users. The goal is to select the most commonly used words as the negative samples, thereby forcing the representations to capture the differences between the words a given individual employs and the words that everyone commonly uses.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Learning User Embeddings", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We now present the details of our proposed sarcasm detection model. Given a message S authored by user u, we wish to capture both the relevant aspects of the content and the relevant contextual information about the author. To represent the content, we use pre-trained word embeddings as the input to a convolutional layer that extracts high-level features. More formally, let E \u2208 R d\u00d7|V| be a pre-trained word embedding matrix, where each column represents a word from the vocabulary V as a d dimensional vector. By selecting the columns of E corresponding to the words in S, we form the sentence matrix:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "S = \uf8ee \uf8ef \uf8f0 e 1 . . . e m \uf8f9 \uf8fa \uf8fb", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "A convolutional layer is composed of a set of filters F \u2208 R d\u00d7h where h is the height of the filter. Filters slide across the input, extracting hgram features that constitute a feature map m \u2208 R |S|\u2212h+1 , where each entry is obtained as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "m i = \u03b1(F \u2022 S [i:i\u2212h+1] + b)", |
| "eq_num": "(5)" |
| } |
| ], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "with i = 1, . . . i \u2212 h + 1. Here, S [i:j] denotes a sub-matrix of S (from row i to row j), b \u2208 R is an additive bias and \u03b1(\u2022) denotes a non-linear activation function, applied elementwise. We transform the resultant feature map into a scalar using max-pooling, i.e., we extract the largest value in the map. We use 3 filters (with varying heights) each of which generates M feature maps that are reduced to a vector", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "f k = [max(m 1 ) \u2295 max(m 2 ) . . . \u2295 max(m M )],", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where \u2295 denotes concatenation. We set \u03b1(\u2022) to be the Rectified Linear Unit activation function (Nair and Hinton, 2010) . The output of all the filters is then combined to form the final representation", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 118, |
| "text": "(Nair and Hinton, 2010)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "c = [f 1 \u2295 f 2 \u2295 f 3 ].", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We will denote this feature vector of a specific sentence S by c S . To represent the context, we assume there is a user embedding matrix U \u2208 R d\u00d7N , where each column represents one of N users with a d dimensional vector. The parameters of this embedding matrix can be initialized randomly or using the approach described in Section 3. Then, we simply map the author of the message into the user embedding space by selecting the corresponding column of U. Letting U u denote the user embedding of author u, we formulate our sarcasm detection model as follows:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "P (y = k|s, u; \u03b8) \u221d Y k \u2022 g(c S \u2295 U u ) + b k g(x) = \u03b1(H \u2022 x + h)", |
| "eq_num": "(6)" |
| } |
| ], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "where g(\u2022) denote the activations of a hidden layer, capturing the relations between the content and context representations, and", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "\u03b8 = {Y, b, H, h, F 1 , F 2 , F 3", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": ", E, U} are parameters to be estimated during training. Here, Y \u2208 R 2\u00d7z and b \u2208 R 2 are the weights and bias of the output layer; H \u2208 R z\u00d73M +d and h \u2208 R z are the weights and bias of the hidden layer; and F i are the weights of the convolutional filters. Henceforth, we will refer to this approach as Content and User Embedding Convolutional Neural Network (CUE-CNN). Figure 2 provides an illustrative schematic depicting this model.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 369, |
| "end": 377, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Proposed Model", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We replicated Bamman and Smith (2015) experimental setup using (a subset of) the same Twitter corpus. The labels were inferred from selfdeclarations of sarcasm, i.e., a tweet is considered sarcastic if it contains the hashtag #sarcasm or #sarcastic and deemed non-sarcastic otherwise. 4 To comply with Twitter terms of service, we were given only the tweet ids along with the corresponding labels and had to retrieve the messages ourselves. By the time we tried to retrieve the messages, some of them were not available. We also did not have access to the historical user tweets used by Bamman and Smith, hence, for each author and mentioned user, we scraped additional tweets from their Twitter feed. Due to restrictions in the Twitter API, we were only able to crawl at most 1000 historical tweets per user. 5 Furthermore, we were unable to collect historical tweets for a significant proportion of the users, thus, we discarded messages for which no contextual information was available, resulting in a corpus of 11, 541 tweets involving 12, 500 unique users (authors and mentioned users). It should also be noted that our historical tweets were posted after the ones in the corpus used for the experiments.", |
| "cite_spans": [ |
| { |
| "start": 285, |
| "end": 286, |
| "text": "4", |
| "ref_id": null |
| }, |
| { |
| "start": 810, |
| "end": 811, |
| "text": "5", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We reimplemented Bamman and Smith (2015) 's sarcasm detection model. This a simple, logisticregression based classifier that exploits rich feature sets to achieve strong performance. These are detailed at length in the original paper, but we briefly summarize them here:", |
| "cite_spans": [ |
| { |
| "start": 17, |
| "end": 40, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 tweet-features, encoding attributes of the target tweet text, including: uni-and bigram bag of words (BoW) features; Brown et al. (1992) word clusters indicators; unlabeled dependency bigrams (both BoW and with Brown cluster representations); part-ofspeech, spelling and abbreviation features; inferred sentiment, at both the tweet and word level; and 'intensifier' indicators.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 author-features, aimed at encoding attributes of the author, including: historically 'salient' terms used by the author; the inferred distribution over topics 6 historically tweeted about by the user; inferred sentiment historically expressed by the user; and author profile information (e.g., profile BoW features).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 audience-features, capturing properties of the addressee of tweets, in those cases that a tweet is directed at someone (via the @ symbol). A subset of these, duplicate the aforementioned author features for the addressee. Additionally, author/audience interaction features are introduced, which capture similarity between the author and addressee, w.r.t. inferred topic distributions. Finally, this set includes a feature capturing the frequency of past communication between the author and addressee.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 response-features, for tweets written in response to another tweet. This set of features captures information relating the two, with BoW features of the original tweet and pairwise cluster indicator features, which the encode Brown clusters observed in both the original and response tweet.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We emphasize that implementing this rich set of features took considerable time and effort. This motivates our approach, which aims to effectively induce and exploit contextually-aware representations without manual feature engineering.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "To assess the importance of modelling contextual information in sarcasm detection, we considered two groups of models as baselines: the first only takes into account the content of a target tweet. The second, combines lexical cues with contextual information. The first group includes the following models:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 UNIGRAMS: 2 -regularized logistic regression classifier with binary unigrams as features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 TWEET ONLY: 2 -regularized logistic regression classifier with binary unigrams and tweet-features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 NBOW: Logistic regression with neural word embeddings as features. Given a sentence matrix S (Eq. 4) as input, a d-dimensional feature vector is computed by summing the individual word embeddings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 NLSE: The Non-linear subspace embedding model due to Astudillo et al. (2015) . The NLSE model adapts pre-trained word embeddings for specific tasks by learning a projection into a small subspace. The idea is that this subspace captures the most discriminative latent aspects encoded in the word embeddings. Given a sentence matrix S, each word vector is first projected into the subspace and then transformed through an element-wise sigmoid function. The final sentence representation is obtained by summing the (adapted) word embeddings and passed into a softmax layer that outputs the predictions.", |
| "cite_spans": [ |
| { |
| "start": 55, |
| "end": 78, |
| "text": "Astudillo et al. (2015)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 CNN: The CNN model for text classification proposed by Kim (2014) , using only features extracted from the convolutional layer acting on the lexical content. The input layer was initialized with pre-trained word embeddings.", |
| "cite_spans": [ |
| { |
| "start": 57, |
| "end": 67, |
| "text": "Kim (2014)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "The second group of baselines consists of the following models:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 TWEET+*: 2 -regularized logistic regression classifier with a combination of tweetfeatures and each of the aforementioned Bamman and Smith (2015) feature sets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 SHALLOW CUE-CNN: A simplified version of our neural model for sarcasm detection, without the hidden layer. We evaluated two variants: initializing the user embeddings at random, and initializing the user embeddings with the approach described in Section 3 (SHALLOW CUE-CNN+USER2VEC). In both cases, the (word and user) embeddings weights were updated during training.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "\u2022 CUE-CNN+*: Our proposed neural network for sarcasm detection. We also evaluated the two aforementioned variants: randomly initialized user embeddings and pretrained user embeddings. But here we compared two different approaches for the negative sampling procedure, namely, sampling from a unigram distribution (CUE-CNN+USER2VEC) and sampling uniformly at random from the vocabulary (CUE-CNN+USER2VEC-UNIFRAND).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We first trained Mikolov et al. 2013 To induce user embeddings, we estimated a unigram distribution with maximum likelihood estimation. Then, for each word in a tweet, we extracted 15 negative samples (for the first term in Eq.1) and used the skip-gram model to precompute the conditional probabilities of words occurring in a window of size 5 (for the second term in Eq.1). Finally, Equation 1 was minimized via Stochastic Gradient Descent on 90% of the historical data, holding out the remainder for validation and using the P (tweet text|user) as early stopping criteria.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-Training Word and User Embeddings", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Note that the parameters for each user only depend on their own tweets; this allowed us to perform these computations in parallel to speed-up the training. m", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pre-Training Word and User Embeddings", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Evaluation was performed via 10-fold crossvalidation. For each split, we fit models to 80% of the data, tuned them on 10% and tested on the remaining, held-out 10%. These data splits were kept fixed in all the experiments. For the linear classifiers, in each split, the regularization constant was selected with a linear search over the (a) Performance of the linear classifier baselines. We include the results reported by Bamman and Smith (2015) as a reference. Discrepancies between their reported results and those we achieved with our re-implementation reflect the fact that their experiments were performed using a significantly larger training set and more historical tweets than we had access to.", |
| "cite_spans": [ |
| { |
| "start": 424, |
| "end": 447, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model Training and Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "(b) Performance of the proposed neural models. We compare simple neural models that only consider the lexical content of a message (first 3 bars) with architectures that explicitly model the context. Bars 4 and 5 show the gains obtained by pretraining the user embeddings. The last 2 bars compare different negative sampling procedures for the user embedding pre-training. The left sub-plot shows linear model performance; the right shows the performance of neural variants. The horizontal line corresponds to the best performance achieved via linear models with rich feature sets. Performance was measured in terms of average accuracy over 10-fold cross-validation; error bars depict the variance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model Training and Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "range C = [1e \u22124 , 1e \u22123 , 1e \u22122 , 1e \u22121 , 1, 10] using the training set to fit the model and evaluating on the tuning set. After selecting the best regularization constant, the model was re-trained on the union of the train and tune sets, and evaluated on the test set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model Training and Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "To train our neural model, we first had to choose a suitable architecture and hyperparameter set. However, selecting the optimal network parametrization would require an extensive search over a large configuration space. Therefore, in these experiments, we followed the recommendations in Zhang and Wallace (2015) We performed random search by sampling without replacement over half of the possible configurations. For each data split, 20% of the training set was reserved for early stopping. We compared the sampled configurations by fitting the model on the remaining training data and testing on the tune set. After choosing the best configuration, we retrained the model on the union of the train and tune set (again reserving 20% of the data for early stopping) and evaluated on the test set.", |
| "cite_spans": [ |
| { |
| "start": 289, |
| "end": 313, |
| "text": "Zhang and Wallace (2015)", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model Training and Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "The model was trained by minimizing the crossentropy error between the predictions and true labels, the gradients w.r.t to the network parameters were computed with backpropagation (Rumelhart et al., 1988) and the model weights were updated with the AdaDelta rule (Zeiler, 2012).", |
| "cite_spans": [ |
| { |
| "start": 181, |
| "end": 205, |
| "text": "(Rumelhart et al., 1988)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model Training and Evaluation", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "Figure 3 presents the main experimental results. In Figure 3a , we show the performance of linear classifiers with the manually engineered feature sets proposed by Bamman and Smith (2015) . Our results differ slightly from those originally reported. Nonetheless, we observe the same general trends: namely, that including contextual features significantly improves the performance, and that the biggest gains are attributable to features encoding information about the authors of tweets.", |
| "cite_spans": [ |
| { |
| "start": 164, |
| "end": 187, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 52, |
| "end": 61, |
| "text": "Figure 3a", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Classification Results", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "The results of neural model variants are shown in Figure 3b . Once again, we find that modelling the context (i.e., the author) of a tweet yields significant gains in accuracy. The difference is that here the network jointly learns appropriate user representations, lexical feature extractors and, finally, the classification model. Further improvements are realized by pre-training the user embed-(a) Users colored according to the politicians they follow on Twitter: the blue circles represent users that follow at least one of the (democrats) accounts: @BarackObama, @HillaryClinton and @BernieSanders; the red circles represent users that follow at least one of the (republicans) accounts: @marcorubio, @tedcruz and @realDonaldTrump. Users that follow accounts from both groups were excluded. We can see that users with a similar political leaning tend to have similar vectors.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 50, |
| "end": 59, |
| "text": "Figure 3b", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Classification Results", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "(b) Users colored with respect to the likelihood of following a sports account. The 500 most popular accounts (according to the authors in our training data) were manually inspected and 100 sports related accounts were selected, e.g., @SkySports, @NBA and @cristiano. We should note that users for which the probabilities lied in the range between 0.3 \u2212 0.7 were discarded to emphasize the extremes. dings (we elaborate on this in the following section). We see additional gains when we introduce a hidden layer capturing the interactions between the context (i.e., user vectors) and the content (lexical vectors). This is intuitively agreeable: the recognition of sarcasm is possible when we jointly consider the speaker and the utterance at hand. Interestingly, we observed that our proposed model not only outperforms all the other baselines, but also shows less variance over the cross-validation experiments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Classification Results", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "Finally, we compared the effect of obtaining negative samples uniformly at random with sampling from a unigram distribution. The experimental results show that the latter improves the accuracy of the model by 0.8%. We believe the reason is that considering the most likely words (under the model) as negative samples, helps by pushing the user vectors away from non-informative words and simultaneously closer to the most discriminative words for that user.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Classification Results", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "We now investigate the user embeddings in more detail. In particular, we are interested in two questions: first, what aspects are being captured in these representations; and second, how they contribute to the improved performance of our model. To investigate the first question, we plotted a T-SNE projection (Maaten and Hinton, 2008) of the high-dimensional vector space where the users are represented into two-dimensions. We then colored each point (representing a user) according to their apparent political leaning (Figure 4a) , and according to their interest in sports (Figure 4b ). These attributes were inferred using the Twitter accounts that a user follows, as a proxy. The plots suggest that the user vectors are indeed able to cap-ture latent aspects, such as political preferences and personal interests. Moreover, the embeddings seem to uncover a notion of homophily, i.e. similar users tend to occupy neighbouring regions of the embedding space. Regarding the second question, we examined the influence of the contextual information on the model's predictions. To this end, we measured the response of our model to the same textual content with different hypothetical contexts (authors). We selected two examples that were misclassified by a simple CNN and ran them trough the CUE-CNN model with three different user embeddings. In Figure 5 , we show these examples along with the predicted probabilities of being a sarcastic post, when no user information is considered and when the author is taken into account. We can see that the predictions drastically change when contextual information is available and that two of the authors trigger similar responses on both examples. This example provides evidence that our model captures the intuition that the same utterance can be interpreted as sarcastic or not, depending on the speaker.", |
| "cite_spans": [ |
| { |
| "start": 310, |
| "end": 335, |
| "text": "(Maaten and Hinton, 2008)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 521, |
| "end": 532, |
| "text": "(Figure 4a)", |
| "ref_id": "FIGREF5" |
| }, |
| { |
| "start": 577, |
| "end": 587, |
| "text": "(Figure 4b", |
| "ref_id": "FIGREF5" |
| }, |
| { |
| "start": 1349, |
| "end": 1357, |
| "text": "Figure 5", |
| "ref_id": "FIGREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "User Embedding Analysis", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "We have introduced CUE-CNN, a novel, deep neural network for automatically recognizing sarcastic utterances on social media. Our model jointly learns and exploits embeddings for the content and users, thus integrating information about the speaker and what he or she has said. This is accomplished without manual feature engineering. Nonetheless, our model outperforms (by over 2% in absolute accuracy) a recently proposed stateof-the-art model that exploits an extensive, handcrafted set of features encoding user attributes and other contextual information. Unlike other approaches that explicitly exploit the structure of particular social media services, such as the forum where a message was posted or metadata about the users, learning user embeddings only requires their preceding messages. Yet, the obtained vectors are able to capture relevant user attributes and a soft notion of homophily. This, we believe, makes our model easier to deploy over different social media environments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Our implementation of the proposed method and the datasets used in this paper have been made publicly available 7 . As future work, we intended to further explore the user embeddings for context", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Like other forms of subjective expression, irony and sarcasm are difficult to define precisely.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://reddit.com is a social news aggregation site comprising specific topical sub-reddits.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "As implemented in Gensim (\u0158eh\u016f\u0159ek and Sojka, 2010).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Note that this is a form of noisy supervision, as of course all sarcastic tweets will not be explicitly flagged as such.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The original study(Bamman and Smith, 2015) was done with at most 3, 200 historical tweets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The topics were extracted from Latent Dirichlet Allocation(Blei et al., 2003).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/samiroid/CUE-CNN representation, namely by also incorporating the interaction between the author and the audience into the model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was supported in part by the Army Research Office (grant W911NF-14-1-0442) and by The Foundation for Science and Technology, Portugal (FCT), through contracts UID/CEC/50021/2013, EXCL/EEI-ESS/0257/2012 (DataStorm), grant UTAP-EXPL/EEIESS/0031/2014 and Ph.D. scholarship SFRH/BD/89020/2012. This work was also made possible by the support of the Texas Advanced Computer Center (TACC) at UT Austin.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Learning word representations from scarce and noisy data with embedding subspaces", |
| "authors": [ |
| { |
| "first": "Ram\u00f3n", |
| "middle": [], |
| "last": "Astudillo", |
| "suffix": "" |
| }, |
| { |
| "first": "Silvio", |
| "middle": [], |
| "last": "Amir", |
| "suffix": "" |
| }, |
| { |
| "first": "Wang", |
| "middle": [], |
| "last": "Ling", |
| "suffix": "" |
| }, |
| { |
| "first": "Mario", |
| "middle": [], |
| "last": "Silva", |
| "suffix": "" |
| }, |
| { |
| "first": "Isabel", |
| "middle": [], |
| "last": "Trancoso", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "1", |
| "issue": "", |
| "pages": "1074--1084", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ram\u00f3n Astudillo, Silvio Amir, Wang Ling, Mario Silva, and Isabel Trancoso. 2015. Learning word representations from scarce and noisy data with em- bedding subspaces. In Proceedings of the 53rd An- nual Meeting of the Association for Computational Linguistics and the 7th International Joint Confer- ence on Natural Language Processing (Volume 1: Long Papers), pages 1074-1084, Beijing, China, July.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Contextualized sarcasm detection on twitter", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Bamman", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Noah", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 9th International Conference on Web and Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "574--77", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Bamman and Noah A Smith. 2015. Contextual- ized sarcasm detection on twitter. In Proceedings of the 9th International Conference on Web and Social Media, pages 574-77. AAAI Menlo Park, CA.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Modelling irony in twitter", |
| "authors": [ |
| { |
| "first": "Francesco", |
| "middle": [], |
| "last": "Barbieri", |
| "suffix": "" |
| }, |
| { |
| "first": "Horacio", |
| "middle": [], |
| "last": "Saggion", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "EACL", |
| "volume": "", |
| "issue": "", |
| "pages": "56--64", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Francesco Barbieri and Horacio Saggion. 2014. Mod- elling irony in twitter. In EACL, pages 56-64.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Class-based n-gram models of natural language", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Peter F Brown", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Peter", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Desouza", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Robert", |
| "suffix": "" |
| }, |
| { |
| "first": "Vincent J Della", |
| "middle": [], |
| "last": "Mercer", |
| "suffix": "" |
| }, |
| { |
| "first": "Jenifer C", |
| "middle": [], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Lai", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "Computational linguistics", |
| "volume": "18", |
| "issue": "4", |
| "pages": "467--479", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peter F Brown, Peter V Desouza, Robert L Mercer, Vincent J Della Pietra, and Jenifer C Lai. 1992. Class-based n-gram models of natural language. Computational linguistics, 18(4):467-479.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Are there necessary conditions for inducing a sense of sarcastic irony?", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "John", |
| "suffix": "" |
| }, |
| { |
| "first": "Albert N", |
| "middle": [], |
| "last": "Campbell", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Katz", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Discourse Processes", |
| "volume": "49", |
| "issue": "", |
| "pages": "459--480", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John D Campbell and Albert N Katz. 2012. Are there necessary conditions for inducing a sense of sarcas- tic irony? Discourse Processes, 49(6):459-480.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Clues for detecting irony in user-generated contents: oh", |
| "authors": [ |
| { |
| "first": "Paula", |
| "middle": [], |
| "last": "Carvalho", |
| "suffix": "" |
| }, |
| { |
| "first": "Lu\u00eds", |
| "middle": [], |
| "last": "Sarmento", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "M\u00e1rio", |
| "suffix": "" |
| }, |
| { |
| "first": "Eug\u00e9nio De", |
| "middle": [], |
| "last": "Silva", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Oliveira", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paula Carvalho, Lu\u00eds Sarmento, M\u00e1rio J Silva, and Eug\u00e9nio De Oliveira. 2009. Clues for detect- ing irony in user-generated contents: oh...", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "s so easy;-)", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "!! It", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "Proceedings of the 1st international CIKM workshop on Topic-sentiment analysis for mass opinion", |
| "volume": "", |
| "issue": "", |
| "pages": "53--56", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "!! it's so easy;-). In Proceedings of the 1st international CIKM workshop on Topic-sentiment analysis for mass opinion, pages 53-56. ACM.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "The Journal of Machine Learning Research", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. The Journal of Machine Learning Re- search, 12:2493-2537.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Semi-supervised recognition of sarcastic sentences in twitter and amazon", |
| "authors": [ |
| { |
| "first": "Dmitry", |
| "middle": [], |
| "last": "Davidov", |
| "suffix": "" |
| }, |
| { |
| "first": "Oren", |
| "middle": [], |
| "last": "Tsur", |
| "suffix": "" |
| }, |
| { |
| "first": "Ari", |
| "middle": [], |
| "last": "Rappoport", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Fourteenth Conference on Computational Natural Language Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "107--116", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dmitry Davidov, Oren Tsur, and Ari Rappoport. 2010. Semi-supervised recognition of sarcastic sentences in twitter and amazon. In Proceedings of the Four- teenth Conference on Computational Natural Lan- guage Learning, pages 107-116. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Why not say it directly? the social functions of irony", |
| "authors": [ |
| { |
| "first": "Shelly", |
| "middle": [], |
| "last": "Dews", |
| "suffix": "" |
| }, |
| { |
| "first": "Joan", |
| "middle": [], |
| "last": "Kaplan", |
| "suffix": "" |
| }, |
| { |
| "first": "Ellen", |
| "middle": [], |
| "last": "Winner", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Discourse processes", |
| "volume": "19", |
| "issue": "3", |
| "pages": "347--367", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shelly Dews, Joan Kaplan, and Ellen Winner. 1995. Why not say it directly? the social functions of irony. Discourse processes, 19(3):347-367.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Notes on noise contrastive estimation and negative sampling", |
| "authors": [ |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1410.8251" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chris Dyer. 2014. Notes on noise contrastive es- timation and negative sampling. arXiv preprint arXiv:1410.8251.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Identifying sarcasm in twitter: a closer look", |
| "authors": [ |
| { |
| "first": "Roberto", |
| "middle": [], |
| "last": "Gonz\u00e1lez-Ib\u00e1nez", |
| "suffix": "" |
| }, |
| { |
| "first": "Smaranda", |
| "middle": [], |
| "last": "Muresan", |
| "suffix": "" |
| }, |
| { |
| "first": "Nina", |
| "middle": [], |
| "last": "Wacholder", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers", |
| "volume": "2", |
| "issue": "", |
| "pages": "581--586", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Roberto Gonz\u00e1lez-Ib\u00e1nez, Smaranda Muresan, and Nina Wacholder. 2011. Identifying sarcasm in twit- ter: a closer look. In Proceedings of the 49th An- nual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers-Volume 2, pages 581-586. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Your sentiment precedes you: Using an author's historical tweets to predict sarcasm", |
| "authors": [ |
| { |
| "first": "Anupam", |
| "middle": [], |
| "last": "Khattri", |
| "suffix": "" |
| }, |
| { |
| "first": "Aditya", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Pushpak", |
| "middle": [], |
| "last": "Bhattacharyya", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [ |
| "James" |
| ], |
| "last": "Carman", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "6th Workshop on Computational Approaches To Subjectivity", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anupam Khattri, Aditya Joshi, Pushpak Bhat- tacharyya, and Mark James Carman. 2015. Your sentiment precedes you: Using an author's histor- ical tweets to predict sarcasm. In 6th Workshop on Computational Approaches To Subjectivity, Sen- timent And Social Media Analysis WASSA 2015, page 25.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Convolutional neural networks for sentence classification", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1746--1751", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoon Kim. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natu- ral Language Processing (EMNLP), pages 1746- 1751, Doha, Qatar, October. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "The use of verbal irony: Cues and constraints. Metaphor: Implications and applications", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Roger", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kreuz", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "23--38", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Roger J Kreuz. 1996. The use of verbal irony: Cues and constraints. Metaphor: Implications and appli- cations, pages 23-38.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Distributed representations of sentences and documents", |
| "authors": [ |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Quoc", |
| "suffix": "" |
| }, |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1405.4053" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Quoc V Le and Tomas Mikolov. 2014. Distributed representations of sentences and documents. arXiv preprint arXiv:1405.4053.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Learning multi-faceted representations of individuals from heterogeneous evidence using neural networks", |
| "authors": [ |
| { |
| "first": "Jiwei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Ritter", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1510.05198" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiwei Li, Alan Ritter, and Dan Jurafsky. 2015. Learning multi-faceted representations of individu- als from heterogeneous evidence using neural net- works. arXiv preprint arXiv:1510.05198.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Really? well. apparently bootstrapping improves the performance of sarcasm and nastiness classifiers for online dialogue", |
| "authors": [ |
| { |
| "first": "Stephanie", |
| "middle": [], |
| "last": "Lukin", |
| "suffix": "" |
| }, |
| { |
| "first": "Marilyn", |
| "middle": [], |
| "last": "Walker", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Workshop on Language Analysis in Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "30--40", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stephanie Lukin and Marilyn Walker. 2013. Really? well. apparently bootstrapping improves the perfor- mance of sarcasm and nastiness classifiers for online dialogue. In Proceedings of the Workshop on Lan- guage Analysis in Social Media, pages 30-40.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Visualizing data using t-sne", |
| "authors": [ |
| { |
| "first": "Laurens", |
| "middle": [], |
| "last": "Van Der Maaten", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Journal of Machine Learning Research", |
| "volume": "9", |
| "issue": "", |
| "pages": "2579--2605", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-sne. Journal of Machine Learning Research, 9(Nov):2579-2605.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Distributed representations of words and phrases and their compositionality", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [ |
| "S" |
| ], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeff", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Advances in neural information processing systems", |
| "volume": "", |
| "issue": "", |
| "pages": "3111--3119", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- rado, and Jeff Dean. 2013. Distributed representa- tions of words and phrases and their compositional- ity. In Advances in neural information processing systems, pages 3111-3119.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Hierarchical probabilistic neural network language model", |
| "authors": [ |
| { |
| "first": "Frederic", |
| "middle": [], |
| "last": "Morin", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Aistats", |
| "volume": "5", |
| "issue": "", |
| "pages": "246--252", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Frederic Morin and Yoshua Bengio. 2005. Hierarchi- cal probabilistic neural network language model. In Aistats, volume 5, pages 246-252. Citeseer.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Rectified linear units improve restricted boltzmann machines", |
| "authors": [ |
| { |
| "first": "Vinod", |
| "middle": [], |
| "last": "Nair", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [ |
| "E" |
| ], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 27th International Conference on Machine Learning (ICML-10)", |
| "volume": "", |
| "issue": "", |
| "pages": "807--814", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Vinod Nair and Geoffrey E Hinton. 2010. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), pages 807-814.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Improved part-of-speech tagging for online conversational text with word clusters. Association for Computational Linguistics", |
| "authors": [ |
| { |
| "first": "Olutobi", |
| "middle": [], |
| "last": "Owoputi", |
| "suffix": "" |
| }, |
| { |
| "first": "O'", |
| "middle": [], |
| "last": "Brendan", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Connor", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Gimpel", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah A", |
| "middle": [], |
| "last": "Schneider", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Olutobi Owoputi, Brendan O'Connor, Chris Dyer, Kevin Gimpel, Nathan Schneider, and Noah A Smith. 2013. Improved part-of-speech tagging for online conversational text with word clusters. Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Sarcasm detection on twitter: A behavioral modeling approach", |
| "authors": [ |
| { |
| "first": "Ashwin", |
| "middle": [], |
| "last": "Rajadesingan", |
| "suffix": "" |
| }, |
| { |
| "first": "Reza", |
| "middle": [], |
| "last": "Zafarani", |
| "suffix": "" |
| }, |
| { |
| "first": "Huan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the Eighth ACM International Conference on Web Search and Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "97--106", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ashwin Rajadesingan, Reza Zafarani, and Huan Liu. 2015. Sarcasm detection on twitter: A behavioral modeling approach. In Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, pages 97-106. ACM.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Software Framework for Topic Modelling with Large Corpora", |
| "authors": [ |
| { |
| "first": "Petr", |
| "middle": [], |
| "last": "Radim\u0159eh\u016f\u0159ek", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Sojka", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks", |
| "volume": "", |
| "issue": "", |
| "pages": "45--50", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Radim\u0158eh\u016f\u0159ek and Petr Sojka. 2010. Software Framework for Topic Modelling with Large Cor- pora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks, pages 45- 50, Valletta, Malta, May. ELRA. http://is. muni.cz/publication/884893/en.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "A multidimensional approach for detecting irony in twitter. Language resources and evaluation", |
| "authors": [ |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Reyes", |
| "suffix": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "" |
| }, |
| { |
| "first": "Tony", |
| "middle": [], |
| "last": "Veale", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "47", |
| "issue": "", |
| "pages": "239--268", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonio Reyes, Paolo Rosso, and Tony Veale. 2013. A multidimensional approach for detecting irony in twitter. Language resources and evaluation, 47(1):239-268.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Sarcasm as contrast between a positive sentiment and negative situation", |
| "authors": [ |
| { |
| "first": "Ellen", |
| "middle": [], |
| "last": "Riloff", |
| "suffix": "" |
| }, |
| { |
| "first": "Ashequl", |
| "middle": [], |
| "last": "Qadir", |
| "suffix": "" |
| }, |
| { |
| "first": "Prafulla", |
| "middle": [], |
| "last": "Surve", |
| "suffix": "" |
| }, |
| { |
| "first": "Lalindra De", |
| "middle": [], |
| "last": "Silva", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Gilbert", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruihong", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "704--714", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ellen Riloff, Ashequl Qadir, Prafulla Surve, Lalin- dra De Silva, Nathan Gilbert, and Ruihong Huang. 2013. Sarcasm as contrast between a positive senti- ment and negative situation. In EMNLP, pages 704- 714.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Learning representations by backpropagating errors", |
| "authors": [ |
| { |
| "first": "Geoffrey", |
| "middle": [ |
| "E" |
| ], |
| "last": "David E Rumelhart", |
| "suffix": "" |
| }, |
| { |
| "first": "Ronald J", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Williams", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "Cognitive modeling", |
| "volume": "5", |
| "issue": "3", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. 1988. Learning representations by back- propagating errors. Cognitive modeling, 5(3):1.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Humans require context to infer ironic intent (so computers probably do, too)", |
| "authors": [ |
| { |
| "first": "Byron", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wallace", |
| "suffix": "" |
| }, |
| { |
| "first": "Laura", |
| "middle": [], |
| "last": "Do Kook Choe", |
| "suffix": "" |
| }, |
| { |
| "first": "Eugene", |
| "middle": [], |
| "last": "Kertz", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Charniak", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "512--516", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Byron C. Wallace, Do Kook Choe, Laura Kertz, and Eugene Charniak. 2014. Humans require context to infer ironic intent (so computers probably do, too). In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 512-516, Baltimore, Mary- land, June. Association for Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Sparse, contextually informed models for irony detection: Exploiting user communities, entities and sentiment", |
| "authors": [ |
| { |
| "first": "Byron", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wallace", |
| "suffix": "" |
| }, |
| { |
| "first": "Eugene", |
| "middle": [], |
| "last": "Do Kook Choe", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Charniak", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1035--1044", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Byron C. Wallace, Do Kook Choe, and Eugene Char- niak. 2015. Sparse, contextually informed mod- els for irony detection: Exploiting user communi- ties, entities and sentiment. In Proceedings of the 53rd Annual Meeting of the Association for Compu- tational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol- ume 1: Long Papers), pages 1035-1044, Beijing, China, July. Association for Computational Linguis- tics.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Adadelta: an adaptive learning rate method", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Matthew", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zeiler", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1212.5701" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Matthew D Zeiler. 2012. Adadelta: an adaptive learn- ing rate method. arXiv preprint arXiv:1212.5701.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "A sensitivity analysis of (and practitioners' guide to) convolutional neural networks for sentence classification", |
| "authors": [ |
| { |
| "first": "Ye", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Byron", |
| "middle": [], |
| "last": "Wallace", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1510.03820" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ye Zhang and Byron Wallace. 2015. A sensitiv- ity analysis of (and practitioners' guide to) convo- lutional neural networks for sentence classification. arXiv preprint arXiv:1510.03820.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "An illustrative tweet.", |
| "num": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Illustration of the CUE-CNN model for sarcasm detection. The model learns to represent and exploit embeddings of both content and users in social media.", |
| "num": null |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "'s skip-gram model variant to induce word embeddings using the union of: Owoputi et al. (2013)'s dataset of 52 Million unlabeled tweets, Bamman and Smith (2015) sarcasm dataset and 5 Million historical tweets collected from users.", |
| "num": null |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Comparison of different models.", |
| "num": null |
| }, |
| "FIGREF4": { |
| "type_str": "figure", |
| "uris": null, |
| "text": ", focusing our search over combinations of dropout rates D = [0.0, 0.1, 0.3, 0.5], filter heights H = [(1, 3, 5), (2, 4, 6), (3, 5, 7), (4, 6, 8), (5, 7, 9)], number of feature maps M = [100, 200, 400, 600] and size of the hidden layer Z = [25, 50, 75, 100].", |
| "num": null |
| }, |
| "FIGREF5": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "T-SNE projection of the user embeddings into 2-dimensions. The users are color coded according to their political preferences and interest in sports. The visualization suggests that the learned embeddings capture some notion of homophily.", |
| "num": null |
| }, |
| "FIGREF6": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Two sarcastic examples that were misclassified by a simple CNN (no user). Using the CUE-CNN with contextual information drastically changes the model's predictions on the same examples.", |
| "num": null |
| } |
| } |
| } |
| } |