| { |
| "paper_id": "N16-1016", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T14:37:33.684008Z" |
| }, |
| "title": "A Long Short-Term Memory Framework for Predicting Humor in Dialogues", |
| "authors": [ |
| { |
| "first": "Dario", |
| "middle": [], |
| "last": "Bertero", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "dbertero@connect.ust.hk" |
| }, |
| { |
| "first": "Pascale", |
| "middle": [], |
| "last": "Fung", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "pascale@ece.ust.hk" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "We propose a first-ever attempt to employ a Long Short-Term memory based framework to predict humor in dialogues. We analyze data from a popular TV-sitcom, whose canned laughters give an indication of when the audience would react. We model the setuppunchline relation of conversational humor with a Long Short-Term Memory, with utterance encodings obtained from a Convolutional Neural Network. Out neural network framework is able to improve the F-score of 8% over a Conditional Random Field baseline. We show how the LSTM effectively models the setup-punchline relation reducing the number of false positives and increasing the recall. We aim to employ our humor prediction model to build effective empathetic machine able to understand jokes.", |
| "pdf_parse": { |
| "paper_id": "N16-1016", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "We propose a first-ever attempt to employ a Long Short-Term memory based framework to predict humor in dialogues. We analyze data from a popular TV-sitcom, whose canned laughters give an indication of when the audience would react. We model the setuppunchline relation of conversational humor with a Long Short-Term Memory, with utterance encodings obtained from a Convolutional Neural Network. Out neural network framework is able to improve the F-score of 8% over a Conditional Random Field baseline. We show how the LSTM effectively models the setup-punchline relation reducing the number of false positives and increasing the recall. We aim to employ our humor prediction model to build effective empathetic machine able to understand jokes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "There has been many recent attempts to detect and understand humor, irony and sarcasm from sentences, usually taken from Twitter (Reyes et al., 2013 ; Barbieri and Saggion, 2014; Riloff et al., 2013; Joshi et al., 2015) , customer reviews (Reyes and Rosso, 2012) or generic canned jokes . Bamman and Smith (2015) and Karoui et al. (2015) included the surrounding context.", |
| "cite_spans": [ |
| { |
| "start": 129, |
| "end": 148, |
| "text": "(Reyes et al., 2013", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 151, |
| "end": 178, |
| "text": "Barbieri and Saggion, 2014;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 179, |
| "end": 199, |
| "text": "Riloff et al., 2013;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 200, |
| "end": 219, |
| "text": "Joshi et al., 2015)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 239, |
| "end": 262, |
| "text": "(Reyes and Rosso, 2012)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 289, |
| "end": 312, |
| "text": "Bamman and Smith (2015)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 317, |
| "end": 337, |
| "text": "Karoui et al. (2015)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our work has a different focus from the above. We analyze transcripts of funny dialogues, a genre somehow neglected but important for human-robot interaction. Laughter is the natural reaction of people to a verbal or textual humorous stimulus. We want to predict when the audience would laugh.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Compared to a typical canned joke or a sarcastic Tweet, a dialog utterance is perceived as funny only in relation to the dialog context and the past history. In a spontaneous setting a funny dialog is usually built through a setup which prepares the audience to receive the humorous discourse stimuli, followed by a punchline which releases the tension and triggers the laughter reaction (Attardo, 1997; Taylor and Mazlack, 2005) . Automatic understanding of a humorous dialog is a first step to build an effective empathetic machine fully able to react to the user's humor and to other discourse stimuli. We are ultimately interested in developing robots that can bond with humans better (Devillers et al., 2015) .", |
| "cite_spans": [ |
| { |
| "start": 388, |
| "end": 403, |
| "text": "(Attardo, 1997;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 404, |
| "end": 429, |
| "text": "Taylor and Mazlack, 2005)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 689, |
| "end": 713, |
| "text": "(Devillers et al., 2015)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "As a source of situational humor we study a popular TV sitcom: \"The Big Bang Theory\". The domain of sitcoms is of interest as it provides a full dialog setting, together with an indication of when the audience is expected to laugh, given by the background canned laughters. An example of dialog from this sitcom, as well as of the setup-punchline schema, is shown below (punchlines in bold): LAUGH He started America on a path to the metric system but then just gave up. LAUGH The utterances before the punchline are the setup. Without them, the punchlines may not be perceived as humorous (the last utterance, out of context, may be a political complaint), only with proper setup a laughter would be triggered. The humorous intent is also strengthen by the fact the dialog takes place in a bar (evident from the previous and following utterances), where a request of 40 ml of \"Ethyl Alcohol\" is unusual and weird.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "PENNY: Okay,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our previous attempts on the same corpus (Bertero and Fung, 2016b; Bertero and Fung, 2016a) showed that using a bag-of-ngram representation over a sliding window or a simple RNN to capture the contextual information of the setup was not ideal. For this reason we propose a method based on a Long Short-Term Memory network (Hochreiter and Schmidhuber, 1997), where we encode each sentence through a Convolutional Neural Network (Collobert et al., 2011) . LSTM is successfully used in context-dependent sequential classification tasks such as speech recognition (Graves et al., 2013) , dependency parsing and conversation modelling (Shang et al., 2015) . This is also to our knowledge the first-ever attempt that a LSTM is applied to humor response prediction or general humor detection tasks.", |
| "cite_spans": [ |
| { |
| "start": 67, |
| "end": 91, |
| "text": "Bertero and Fung, 2016a)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 427, |
| "end": 451, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 560, |
| "end": 581, |
| "text": "(Graves et al., 2013)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 630, |
| "end": 650, |
| "text": "(Shang et al., 2015)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We employ a supervised classification method to detect when punchlines occur. The bulk of our classifier is made of a concatenation of a Convolutional Neural Network (Collobert et al., 2011) to encode each utterance, followed by a Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) to model the sequential context of the dialog. Before the output softmax layer we add a vector of higher level syntactic, structural and sentiment features. A framework diagram is shown in Figure 1 .", |
| "cite_spans": [ |
| { |
| "start": 166, |
| "end": 190, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 254, |
| "end": 288, |
| "text": "(Hochreiter and Schmidhuber, 1997)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 478, |
| "end": 486, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The first stage of our classifier is represented by a Convolutional Neural Network (Collobert et al., 2011) . Low-level, high-dimensional input feature vectors are fed into a first embedding layer to obtain a low dimensional dense vector. A sliding window is then moved on these vectors and another layer is applied to each group of token vectors, in order to capture the local context of each token. A max-pooling operation is then applied to extract the most salient features of all the tokens into a single vector for the whole utterance. An additional layer is used to generalize and distribute each feature to its full range before obtaining the final utterance vector. In our task we use three input features:", |
| "cite_spans": [ |
| { |
| "start": 83, |
| "end": 107, |
| "text": "(Collobert et al., 2011)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Neural Network for each utterance", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "1. Word tokens: each utterance token is represented as a one-hot vector. This feature models how much each word is likely to trigger humor in the specific corpus.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Neural Network for each utterance", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "2. Character trigrams: each token is represented as a bag-of-character-trigrams vector. The feature models the role of the word signifier and removes the influence of the word stems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Neural Network for each utterance", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "3. Word2Vec: we extract for each token a word vector from word2vec (Mikolov et al., 2013) , trained on the text9 Wikipedia corpus 1 . This representation models the general semantic meanings, and matches words that do not appear to others similar in meaning.", |
| "cite_spans": [ |
| { |
| "start": 67, |
| "end": 89, |
| "text": "(Mikolov et al., 2013)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Neural Network for each utterance", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The convolution and max-pooling operation is applied individually to each feature, and the three vectors obtained are then concatenated together and fed to the final sentence encoding layer, which combines all the contributions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Neural Network for each utterance", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The LSTM is an improvement over the Recurrent Neural Network aimed to improve its memory capabilities. In a standard RNN the hidden memory layer is updated through a function of the input and the hidden layer at the previous time instant:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "h t = tanh(W x x t + W h h t\u22121 + b)", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "where x is the network input and b the bias term. This kind of connection is not very effective to maintain the information stored for long time instants, as well as it does not allow to forget unneeded information between two time steps. The LSTM enhances the RNN with a series of three multiplicative gates.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The structure is the following:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "i t = \u03c3(W ix x t + W i h h t\u22121 + b i ) (2) f t = \u03c3(W fx x t + W f h h t\u22121 + b f ) (3) o t = \u03c3(W ox x t + W o h h t\u22121 + b o ) (4) s t = tanh(W sx x t + W s h h t\u22121 + b h ) (5) c t = f t c t\u22121 + i t s t (6) h t = tanh(c t ) o t (7)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "where is the element-wise product. Each gate factor is able to let through or suppress a specific update contribution, thus allowing a selective information retaining. The input gate i is applied to the cell input s, the forget gate f to the cell value at the previous time step c t\u22121 , and the output gate o to the cell output for the current time instant h t . In this way a cell value can be retained for multiple time steps when i = 0, ignored in the output when o = 0, and forgotten when f = 0. As dialog utterances are sequential, we feed all utterance vectors of a sitcom scene in sequence into a Long Short-Term Memory block to incorporate contextual information. The memory unit of the LSTM keeps track of the context in each scene, and mimics human memory to accumulate the setup that may trigger a punchline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Before the output we incorporate a set of high level features from our previous work (Bertero and Fung, 2016b) and past literature (Reyes et al., 2013; Barbieri and Saggion, 2014) . They include:", |
| "cite_spans": [ |
| { |
| "start": 85, |
| "end": 110, |
| "text": "(Bertero and Fung, 2016b)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 131, |
| "end": 151, |
| "text": "(Reyes et al., 2013;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 152, |
| "end": 179, |
| "text": "Barbieri and Saggion, 2014)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Structural features: average word length, sentence length, difference in sentence length with the five previous utterances.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Part of speech proportion: noun, verbs, adjectives and adverbs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Antonyms: proportion of antonym words with the previous utterance (from WordNet (Miller, 1995) ).", |
| "cite_spans": [ |
| { |
| "start": 74, |
| "end": 96, |
| "text": "WordNet (Miller, 1995)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Sentiment: positive, negative and average sentiment score among all words (from Senti-WordNet (Esuli and Sebastiani, 2006) ).", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 124, |
| "text": "(Esuli and Sebastiani, 2006)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Speaker and turn: speaker character identity and utterance position in the turn (beginning, middle, end, isolated).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Speaking rate: time duration of the utterance from the subtitle files, divided by the sentence length.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "All these features are concatenated to the LSTM output, and a softmax layer is applied to get the final output probabilities.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "3 Experiments", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Long/Short Term Memory for the utterance sequence", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We built a corpus from the popular TV-sitcom \"The Big Bang Theory\", seasons 1 to 6. We downloaded the subtitle files (annotated with the timestamps of each utterance) and the scripts 2 , used to segment all the episodes into scenes and get the speaker identity of each utterance. We extracted the audio track of each episode in order to retrieve the canned laughters timestamps, with a vocal removal tool followed by a silence/sound detector. We then annotated each utterance as a punchline in case it was followed by a laughter within 1s, assuming that utterances not followed by a laughter would be the setup for the punchline.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We obtained a total of 135 episodes, 1589 overall scenes, 42.8% of punchlines, and an average interval between two punchlines of 2.2 utterances. We built a training set of 80% of the overall episodes, and a development and test set of 10% each. The episodes were drawn from all the seasons with the same proportion. The total number of utterances is 35865 for the training set, 3904 for the development set and 3903 for the test set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Corpus", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In the neural network we set the size to 100 for all the hidden layers of the CNN and the LSTM, and 5 to the convolutional window. We applied a dropout regularization layer (Srivastava et al., 2014) after the output of the LSTM, and L2 regularization on the softmax output layer. The network was trained with standard backpropagation, using each scene as a training unit. The development set was used to tune the hyperparameters, and to determine the early stopping condition. When the error on the development set began to increase for the first time we kept training only the final softmax layer, this improved the overall results. The neural network was implemented with THEANO toolkit (Bergstra et al., 2010) . We ran experiments with and without the extra high-level feature vector.", |
| "cite_spans": [ |
| { |
| "start": 173, |
| "end": 198, |
| "text": "(Srivastava et al., 2014)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 689, |
| "end": 712, |
| "text": "(Bergstra et al., 2010)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup and baseline", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "As a baseline for comparison we used an implementation of the Conditional Random Field (Lafferty et al., 2001 ) from CRFSuite (Okazaki, 2007) , with L2 regularization. We ran experiments using the same high level feature vector added at the end of the neural network, 1-2-3gram features of a window made by the utterance and the four previous, and the two feature sets combined. We also compared the overall system where we replace the CNN with an LSTM sentence encoder (Li et al., 2015) , where we kept the same input features.", |
| "cite_spans": [ |
| { |
| "start": 87, |
| "end": 109, |
| "text": "(Lafferty et al., 2001", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 126, |
| "end": 141, |
| "text": "(Okazaki, 2007)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 470, |
| "end": 487, |
| "text": "(Li et al., 2015)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup and baseline", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Results of our system and our baselines are shown in table 1. The LSTM with the aid of the high level feature vector generally outperformed all the CRF baselines with the highest accuracy of 70.0% and the highest F-score of 62.9%. The biggest improvement of the LSTM is the improvement of the recall without affecting too much the precision. Lexical features given by n-gram from a context window are very useful to recognize more punchlines in our baseline experiment, but they also yield many false positives, when the same n-gram is used in different contexts. A CNN-LSTM network seems to overcome this issue as the CNN stage is better in modeling the lexical and semantic content of the utterance, as the LSTM allows to put each utterance in relation with the past context, filtering out many false positives from wrong contexts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "The choice of the CNN is further justified by the results obtained from the comparison between the CNN and the LSTM sentence encoding input, shown in table 2. The CNN is more effective, obtaining a recall of 10% higher and 6% more in Fscore. The CNN is a simpler model that might benefit more of a small-size corpus. It also required a much shorter training time compared to the LSTM. We may consider in the future to use more data, and try other sentence input encoders, including deeper or bi-directional LSTMs, to find the most effective one.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Predicting humor response from the canned laughters is a particularly challenging task. In some cases canned laughters are inserted by the show producers with the purpose of solicit response to weak jokes, where otherwise people would not laugh. The audience must also be kept constantly amused, extra canned laughters may help in scenes where fewer jokes are used.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and discussion", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "We proposed a Long Short-Term Memory based framework to predict punchlines in a humorous dialog. We showed that our neural network is particularly effective in increasing the F-score to 62.9% over a Conditional Random Field baseline of 58.1%. We furthermore showed that the LSTM is more effective in obtaining an higher recall with fewer false positives compared to simple n-gram shifting context window features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and future work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "As future work we plan to use a virtual agent system to collect a set of human-robot humorous interactions, and adapt our model to predict humor from them.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion and future work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Extension of the text8 corpus, obtained from http://mattmahoney.net/dc/textdata", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "From bigbangtrans.wordpress.com", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work was partially funded by the Hong Kong Phd Fellowship Scheme, and partially by grant #16214415 of the Hong Kong Research Grants Council.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "The semantic foundations of cognitive theories of humor", |
| "authors": [ |
| { |
| "first": "Salvatore", |
| "middle": [], |
| "last": "Attardo", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Humor-International Journal of Humor Research", |
| "volume": "", |
| "issue": "10", |
| "pages": "395--420", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Salvatore Attardo. 1997. The semantic foundations of cognitive theories of humor. Humor-International Journal of Humor Research, (10):395-420.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Contextualized sarcasm detection on twitter", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Bamman", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Noah", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Ninth International AAAI Conference on Web and Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Bamman and Noah A Smith. 2015. Contextual- ized sarcasm detection on twitter. In Ninth Interna- tional AAAI Conference on Web and Social Media.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Modelling irony in twitter", |
| "authors": [ |
| { |
| "first": "Francesco", |
| "middle": [], |
| "last": "Barbieri", |
| "suffix": "" |
| }, |
| { |
| "first": "Horacio", |
| "middle": [], |
| "last": "Saggion", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the Student Research Workshop at the 14th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "56--64", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Francesco Barbieri and Horacio Saggion. 2014. Mod- elling irony in twitter. In Proceedings of the Student Research Workshop at the 14th Conference of the Eu- ropean Chapter of the Association for Computational Linguistics, pages 56-64.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Theano: a CPU and GPU math expression compiler", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Bergstra", |
| "suffix": "" |
| }, |
| { |
| "first": "Olivier", |
| "middle": [], |
| "last": "Breuleux", |
| "suffix": "" |
| }, |
| { |
| "first": "Fr\u00e9d\u00e9ric", |
| "middle": [], |
| "last": "Bastien", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascal", |
| "middle": [], |
| "last": "Lamblin", |
| "suffix": "" |
| }, |
| { |
| "first": "Razvan", |
| "middle": [], |
| "last": "Pascanu", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Desjardins", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Turian", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Warde-Farley", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Python for Scientific Computing Conference (SciPy)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Bergstra, Olivier Breuleux, Fr\u00e9d\u00e9ric Bastien, Pas- cal Lamblin, Razvan Pascanu, Guillaume Desjardins, Joseph Turian, David Warde-Farley, and Yoshua Ben- gio. 2010. Theano: a CPU and GPU math expression compiler. In Proceedings of the Python for Scientific Computing Conference (SciPy), June. Oral Presenta- tion.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Deep learning of audio and language features for humor prediction", |
| "authors": [ |
| { |
| "first": "Dario", |
| "middle": [], |
| "last": "Bertero", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascale", |
| "middle": [], |
| "last": "Fung", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "International Conference on Language Resources and Evaluation (LREC)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dario Bertero and Pascale Fung. 2016a. Deep learning of audio and language features for humor prediction. International Conference on Language Resources and Evaluation (LREC) 2016.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Predicting humor response in dialogues from TV sitcoms", |
| "authors": [ |
| { |
| "first": "Dario", |
| "middle": [], |
| "last": "Bertero", |
| "suffix": "" |
| }, |
| { |
| "first": "Pascale", |
| "middle": [], |
| "last": "Fung", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Acoustics, Speech and Signal Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dario Bertero and Pascale Fung. 2016b. Predicting hu- mor response in dialogues from TV sitcoms. In Acous- tics, Speech and Signal Processing (ICASSP), 2016 IEEE International Conference on.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Natural language processing (almost) from scratch", |
| "authors": [ |
| { |
| "first": "Ronan", |
| "middle": [], |
| "last": "Collobert", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Weston", |
| "suffix": "" |
| }, |
| { |
| "first": "L\u00e9on", |
| "middle": [], |
| "last": "Bottou", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Karlen", |
| "suffix": "" |
| }, |
| { |
| "first": "Koray", |
| "middle": [], |
| "last": "Kavukcuoglu", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Kuksa", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "The Journal of Machine Learning Research", |
| "volume": "12", |
| "issue": "", |
| "pages": "2493--2537", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ronan Collobert, Jason Weston, L\u00e9on Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. The Journal of Machine Learning Research, 12:2493- 2537.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Multimodal data collection of human-robot humorous interactions in the joker project", |
| "authors": [ |
| { |
| "first": "Laurence", |
| "middle": [], |
| "last": "Devillers", |
| "suffix": "" |
| }, |
| { |
| "first": "Sophie", |
| "middle": [], |
| "last": "Rosset", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [ |
| "Dubuisson" |
| ], |
| "last": "Duplessis", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohamed", |
| "middle": [ |
| "A" |
| ], |
| "last": "Sehili", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucile", |
| "middle": [], |
| "last": "B\u00e9chade", |
| "suffix": "" |
| }, |
| { |
| "first": "Agnes", |
| "middle": [], |
| "last": "Delaborde", |
| "suffix": "" |
| }, |
| { |
| "first": "Cl\u00e9ment", |
| "middle": [], |
| "last": "Gossart", |
| "suffix": "" |
| }, |
| { |
| "first": "Vincent", |
| "middle": [], |
| "last": "Letard", |
| "suffix": "" |
| }, |
| { |
| "first": "Fan", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yucel", |
| "middle": [], |
| "last": "Yemez", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on", |
| "volume": "", |
| "issue": "", |
| "pages": "348--354", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Laurence Devillers, Sophie Rosset, Guillaume Dubuis- son Duplessis, Mohamed A Sehili, Lucile B\u00e9chade, Agnes Delaborde, Cl\u00e9ment Gossart, Vincent Letard, Fan Yang, Yucel Yemez, et al. 2015. Multimodal data collection of human-robot humorous interactions in the joker project. In Affective Computing and Intel- ligent Interaction (ACII), 2015 International Confer- ence on, pages 348-354. IEEE.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Transitionbased dependency parsing with stack long short-term memory", |
| "authors": [ |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Miguel", |
| "middle": [], |
| "last": "Ballesteros", |
| "suffix": "" |
| }, |
| { |
| "first": "Wang", |
| "middle": [], |
| "last": "Ling", |
| "suffix": "" |
| }, |
| { |
| "first": "Austin", |
| "middle": [], |
| "last": "Matthews", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah", |
| "middle": [ |
| "A" |
| ], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "1", |
| "issue": "", |
| "pages": "334--343", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chris Dyer, Miguel Ballesteros, Wang Ling, Austin Matthews, and Noah A. Smith. 2015. Transition- based dependency parsing with stack long short-term memory. In Proceedings of the 53rd Annual Meet- ing of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 334-343, Beijing, China, July. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Sentiwordnet: A publicly available lexical resource for opinion mining", |
| "authors": [ |
| { |
| "first": "Andrea", |
| "middle": [], |
| "last": "Esuli", |
| "suffix": "" |
| }, |
| { |
| "first": "Fabrizio", |
| "middle": [], |
| "last": "Sebastiani", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of LREC", |
| "volume": "6", |
| "issue": "", |
| "pages": "417--422", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Andrea Esuli and Fabrizio Sebastiani. 2006. Sentiword- net: A publicly available lexical resource for opinion mining. In Proceedings of LREC, volume 6, pages 417-422.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Speech recognition with deep recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Graves", |
| "suffix": "" |
| }, |
| { |
| "first": "Mohamed", |
| "middle": [], |
| "last": "Abdel-Rahman", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on", |
| "volume": "", |
| "issue": "", |
| "pages": "6645--6649", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alan Graves, Abdel-rahman Mohamed, and Geoffrey Hinton. 2013. Speech recognition with deep recur- rent neural networks. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Con- ference on, pages 6645-6649. IEEE.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Long short-term memory", |
| "authors": [ |
| { |
| "first": "Sepp", |
| "middle": [], |
| "last": "Hochreiter", |
| "suffix": "" |
| }, |
| { |
| "first": "J\u00fcrgen", |
| "middle": [], |
| "last": "Schmidhuber", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Neural computation", |
| "volume": "9", |
| "issue": "8", |
| "pages": "1735--1780", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation, 9(8):1735- 1780.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Harnessing context incongruity for sarcasm detection", |
| "authors": [ |
| { |
| "first": "Aditya", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Vinita", |
| "middle": [], |
| "last": "Sharma", |
| "suffix": "" |
| }, |
| { |
| "first": "Pushpak", |
| "middle": [], |
| "last": "Bhattacharyya", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "2", |
| "issue": "", |
| "pages": "757--762", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Aditya Joshi, Vinita Sharma, and Pushpak Bhat- tacharyya. 2015. Harnessing context incongruity for sarcasm detection. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguis- tics and the 7th International Joint Conference on Nat- ural Language Processing, volume 2, pages 757-762.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Towards a contextual pragmatic model to detect irony in tweets", |
| "authors": [ |
| { |
| "first": "Jihen", |
| "middle": [], |
| "last": "Karoui", |
| "suffix": "" |
| }, |
| { |
| "first": "Benamara", |
| "middle": [], |
| "last": "Farah", |
| "suffix": "" |
| }, |
| { |
| "first": "V\u00e9ronique", |
| "middle": [], |
| "last": "Moriceau", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathalie", |
| "middle": [], |
| "last": "Aussenac-Gilles", |
| "suffix": "" |
| }, |
| { |
| "first": "Lamia", |
| "middle": [], |
| "last": "Hadrich-Belguith", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
| "volume": "2", |
| "issue": "", |
| "pages": "644--650", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jihen Karoui, Benamara Farah, V\u00e9ronique Moriceau, Nathalie Aussenac-Gilles, and Lamia Hadrich- Belguith. 2015. Towards a contextual pragmatic model to detect irony in tweets. In Proceedings of the 53rd Annual Meeting of the Association for Com- putational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 644-650, Beijing, China, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Conditional random fields: Probabilistic models for segmenting and labeling sequence data", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [ |
| "D" |
| ], |
| "last": "Lafferty", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| }, |
| { |
| "first": "Fernando", |
| "middle": [ |
| "C N" |
| ], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the Eighteenth International Conference on Machine Learning, ICML '01", |
| "volume": "", |
| "issue": "", |
| "pages": "282--289", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John D. Lafferty, Andrew McCallum, and Fernando C. N. Pereira. 2001. Conditional random fields: Proba- bilistic models for segmenting and labeling sequence data. In Proceedings of the Eighteenth International Conference on Machine Learning, ICML '01, pages 282-289, San Francisco, CA, USA. Morgan Kauf- mann Publishers Inc.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A hierarchical neural autoencoder for paragraphs and documents", |
| "authors": [ |
| { |
| "first": "Jiwei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Minh-Thang", |
| "middle": [], |
| "last": "Luong", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1506.01057" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiwei Li, Minh-Thang Luong, and Dan Jurafsky. 2015. A hierarchical neural autoencoder for paragraphs and documents. arXiv preprint arXiv:1506.01057.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Efficient estimation of word representations in vector space", |
| "authors": [ |
| { |
| "first": "Tomas", |
| "middle": [], |
| "last": "Mikolov", |
| "suffix": "" |
| }, |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Greg", |
| "middle": [], |
| "last": "Corrado", |
| "suffix": "" |
| }, |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Dean", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "CoRR", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word represen- tations in vector space. CoRR, abs/1301.3781.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Wordnet: a lexical database for english", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "George", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Miller", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Communications of the ACM", |
| "volume": "38", |
| "issue": "11", |
| "pages": "39--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "George A Miller. 1995. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39-41.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Crfsuite: a fast implementation of conditional random fields (crfs", |
| "authors": [ |
| { |
| "first": "Naoaki", |
| "middle": [], |
| "last": "Okazaki", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Naoaki Okazaki. 2007. Crfsuite: a fast implemen- tation of conditional random fields (crfs). URL http://www.chokkan.org/software/crfsuite.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Making objective decisions from subjective data: Detecting irony in customer reviews", |
| "authors": [ |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Reyes", |
| "suffix": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Decision Support Systems", |
| "volume": "53", |
| "issue": "4", |
| "pages": "754--760", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonio Reyes and Paolo Rosso. 2012. Making objec- tive decisions from subjective data: Detecting irony in customer reviews. Decision Support Systems, 53(4):754-760.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "A multidimensional approach for detecting irony in twitter. Language Resources and Evaluation", |
| "authors": [ |
| { |
| "first": "Antonio", |
| "middle": [], |
| "last": "Reyes", |
| "suffix": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "" |
| }, |
| { |
| "first": "Tony", |
| "middle": [], |
| "last": "Veale", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "47", |
| "issue": "", |
| "pages": "239--268", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Antonio Reyes, Paolo Rosso, and Tony Veale. 2013. A multidimensional approach for detecting irony in twit- ter. Language Resources and Evaluation, 47(1):239- 268.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Sarcasm as contrast between a positive sentiment and negative situation", |
| "authors": [ |
| { |
| "first": "Ellen", |
| "middle": [], |
| "last": "Riloff", |
| "suffix": "" |
| }, |
| { |
| "first": "Ashequl", |
| "middle": [], |
| "last": "Qadir", |
| "suffix": "" |
| }, |
| { |
| "first": "Prafulla", |
| "middle": [], |
| "last": "Surve", |
| "suffix": "" |
| }, |
| { |
| "first": "Lalindra De", |
| "middle": [], |
| "last": "Silva", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Gilbert", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruihong", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "704--714", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ellen Riloff, Ashequl Qadir, Prafulla Surve, Lalindra De Silva, Nathan Gilbert, and Ruihong Huang. 2013. Sarcasm as contrast between a positive sentiment and negative situation. In EMNLP, pages 704-714.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Neural responding machine for short-text conversation", |
| "authors": [ |
| { |
| "first": "Lifeng", |
| "middle": [], |
| "last": "Shang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhengdong", |
| "middle": [], |
| "last": "Lu", |
| "suffix": "" |
| }, |
| { |
| "first": "Hang", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1503.02364" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lifeng Shang, Zhengdong Lu, and Hang Li. 2015. Neu- ral responding machine for short-text conversation. arXiv preprint arXiv:1503.02364.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Dropout: A simple way to prevent neural networks from overfitting", |
| "authors": [ |
| { |
| "first": "Nitish", |
| "middle": [], |
| "last": "Srivastava", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Hinton", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Krizhevsky", |
| "suffix": "" |
| }, |
| { |
| "first": "Ilya", |
| "middle": [], |
| "last": "Sutskever", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruslan", |
| "middle": [], |
| "last": "Salakhutdinov", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "The Journal of Machine Learning Research", |
| "volume": "15", |
| "issue": "1", |
| "pages": "1929--1958", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 15(1):1929-1958.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Toward computational recognition of humorous intent", |
| "authors": [ |
| { |
| "first": "Julia", |
| "middle": [], |
| "last": "Taylor", |
| "suffix": "" |
| }, |
| { |
| "first": "Lawrence", |
| "middle": [], |
| "last": "Mazlack", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of Cognitive Science Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "2166--2171", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Julia Taylor and Lawrence Mazlack. 2005. Toward computational recognition of humorous intent. In Proceedings of Cognitive Science Conference, pages 2166-2171.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Humor recognition and humor anchor extraction", |
| "authors": [ |
| { |
| "first": "Diyi", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Alon", |
| "middle": [], |
| "last": "Lavie", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2367--2376", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Diyi Yang, Alon Lavie, Chris Dyer, and Eduard Hovy. 2015. Humor recognition and humor anchor extrac- tion. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 2367-2376, Lisbon, Portugal, September. Asso- ciation for Computational Linguistics.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Framework diagram. at, bt and ct are the CNN three input features (words, word2vec and character trigrams).lt are the high level feature vectors, and yt the outputs for each utterance.", |
| "num": null |
| }, |
| "TABREF2": { |
| "html": null, |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "Comparison between a CNN and a LSTM sentence encoding input." |
| } |
| } |
| } |
| } |