| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T12:52:27.023152Z" |
| }, |
| "title": "CAN-GRU: a Hierarchical Model for Emotion Recognition in Dialogue", |
| "authors": [ |
| { |
| "first": "Ting", |
| "middle": [], |
| "last": "Jiang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Laboratory of Machine Intelligence and Translation", |
| "institution": "Harbin Institute of Technology Harbin", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "jiangtinghit@163.com" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Laboratory of Machine Intelligence and Translation", |
| "institution": "Harbin Institute of Technology Harbin", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Tiejun", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Laboratory of Machine Intelligence and Translation", |
| "institution": "Harbin Institute of Technology Harbin", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "tjzhao@hit.edu.cn" |
| }, |
| { |
| "first": "Sheng", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Laboratory of Machine Intelligence and Translation", |
| "institution": "Harbin Institute of Technology Harbin", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "lisheng@hit.edu.cn" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Emotion recognition in dialogue systems has gained attention in the field of natural language processing recent years, because it can be applied in opinion mining from public conversational data on social media. In this paper, we propose a hierarchical model to recognize emotions in the dialogue. In the first layer, in order to extract textual features of utterances, we propose a convolutional self-attention network(CAN). Convolution is used to capture n-gram information and attention mechanism is used to obtain the relevant semantic information among words in the utterance. In the second layer, a GRU-based network helps to capture contextual information in the conversation. Furthermore, we discuss the effects of unidirectional and bidirectional networks. We conduct experiments on Friends dataset and EmotionPush dataset. The results show that our proposed model(CAN-GRU) and its variants achieve better performance than baselines.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Emotion recognition in dialogue systems has gained attention in the field of natural language processing recent years, because it can be applied in opinion mining from public conversational data on social media. In this paper, we propose a hierarchical model to recognize emotions in the dialogue. In the first layer, in order to extract textual features of utterances, we propose a convolutional self-attention network(CAN). Convolution is used to capture n-gram information and attention mechanism is used to obtain the relevant semantic information among words in the utterance. In the second layer, a GRU-based network helps to capture contextual information in the conversation. Furthermore, we discuss the effects of unidirectional and bidirectional networks. We conduct experiments on Friends dataset and EmotionPush dataset. The results show that our proposed model(CAN-GRU) and its variants achieve better performance than baselines.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "As an important component of human intelligence, emotional intelligence is defined as the ability to perceive, integrate, understand and regulate emotions (Mayer et al., 2008) . Emotion is the essential difference between human and machine, so emotion understanding is an important research direction of artificial intelligence. As the most common way for people to communicate in daily life, dialogue contains a wealth of emotions. Recognising the emotions in the conversation is of great significance in intelligent customer service, medical systems, education systems and other aspects.", |
| "cite_spans": [ |
| { |
| "start": 155, |
| "end": 175, |
| "text": "(Mayer et al., 2008)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "According to (Poria et al., 2015) , textual features usually contain more emotional information than video or audio features, so we focus on the emotion analysis of dialogue text and aims to recognize the emotion of each utterrance in dialogues.", |
| "cite_spans": [ |
| { |
| "start": 13, |
| "end": 33, |
| "text": "(Poria et al., 2015)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "There are some challenges in this task. First, the length of an utterance may be too long, making it difficult to capture contextual information. Furthermore, a dialogue usually contains lots of utterances, therefore, it's hard to grasp long-term contextual relations between utterances. Second, the same word may express different emotions in different contexts. For example, in Table 1 , while in different dialogues, the word 'Yeah' can express three different emotions, that is , joy, neutral and suprise. To tackle these challenges, we propose a hierarchical model based on convolutional attention network and gated recurrent unit (CAN-GRU). Existing works pay little attention to the extraction of semantic information within an utterance. In this work, we focus on this problem. Our proposed model can extract n-gram information by CNNs and use self-attention to capture contextual information within an utterance in the first layer. Moreover, we utilize a GRU-based network to model the sequence of utterances in the second layer, which can fully combine the context when analyzing utterance emotion and solve the problem of long-term dependence between texts at the same time.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 380, |
| "end": 387, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Text emotion recognition is one of the most hot topic in natural language processing. Recent years,a lot of classical neural networks are used to tackle this problem. work (Hochreiter and Schmidhuber, 1997) , Gated Recurrent Unit Network (Cho et al., 2014) and textual Convolutional Neural Network (Kim, 2014) . However, these models don't perform well when the texts are too long, because it's hard to capture the long-range contextual information. Later, attention mechanism (Bahdanau et al., 2015) is proposed to solve this problem. Recently, self-attention (Vaswani et al., 2017) is widely used since it can solve the long-term dependence problem of text effectively.", |
| "cite_spans": [ |
| { |
| "start": 172, |
| "end": 206, |
| "text": "(Hochreiter and Schmidhuber, 1997)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 238, |
| "end": 256, |
| "text": "(Cho et al., 2014)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 298, |
| "end": 309, |
| "text": "(Kim, 2014)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 477, |
| "end": 500, |
| "text": "(Bahdanau et al., 2015)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 561, |
| "end": 583, |
| "text": "(Vaswani et al., 2017)", |
| "ref_id": "BIBREF20" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Recent years, more and more researchers focus on emotion recognition in conversation. This task aims to recognize the emotion of each utterance in dialogues. bcLSTM (Poria et al., 2017) extracts textual features by CNN and model the sequence of utterances by LSTM. Considering inter-speaker dependency relations, conversational memory network(CMN) (Hazarika et al., 2018b) has been proposed to model the speaker-based emotion using memory network and summarize task-specific details by attention mechanisms. ICON (Hazarika et al., 2018a) improves the CMN, it hierarchically models the self-speaker emotion and inter-speaker emotion into global memories. DialogueRNN uses emotion GRU and global GRU to model inter-party relation, and uses party GRU to model relation between two sequential states of the same party. DialogueGCN (Ghosal et al., 2019) improves DialogueRNN by graph convolutional network, and it can hold richer context relevant to emotion. However, these models may be too complex for small textual dialogue datasets.", |
| "cite_spans": [ |
| { |
| "start": 165, |
| "end": 185, |
| "text": "(Poria et al., 2017)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 348, |
| "end": 372, |
| "text": "(Hazarika et al., 2018b)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 513, |
| "end": 537, |
| "text": "(Hazarika et al., 2018a)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 827, |
| "end": 848, |
| "text": "(Ghosal et al., 2019)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this paper, we study on the EmotionX Challenge(Hsu and Ku, 2018), Dialogue Emotion Recognition Challenge, which aims to recognize the emotion of each utterance in dialogues. According to the overview of this task, the best team (Khosla, 2018) proposes a CNN-DCNN auto encoder based model, which includes a convolutional encoder and a deconvolutional decoder. The second place team (Luo et al., 2018) mainly uses BiLSTM with a self-attentive architecture on the top for the classiffication. The third place team (Saxena et al., 2018) proposes a hierarchical network based on attention models and conditional random fields(CRF). For a meaningful comparison, we use the same dataset and metric as the challenge in our study.", |
| "cite_spans": [ |
| { |
| "start": 231, |
| "end": 245, |
| "text": "(Khosla, 2018)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 384, |
| "end": 402, |
| "text": "(Luo et al., 2018)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 514, |
| "end": 535, |
| "text": "(Saxena et al., 2018)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Given a dialogue dia = {u 1 , u 2 , ..., u N }, where N is the number of utterances in the dialogue, u i = {w 1 , w 2 , ..., w L } represents the ith(1 \u2264 i \u2264 N ) utterance in the dialogue that consists of L words, our goal is to analyze the emotion of each utterance in the dialogue. To solve this task, we propose a hierarchical model CAN-GRU and extend three variants, CAN-GRUA, CAN-biGRU and CAN-biGRUA(illustrated in Fig. 1 ).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 421, |
| "end": 427, |
| "text": "Fig. 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Task Definition", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "In this section, we discuss the first layer of the model. Like (Poria et al., 2017) , we use convolutional neural network to extract the features of the utterance. Inspired by (Gao et al., 2018) , in order to capture the contextual information of long text effectively, we use convolutional self-attention network(CAN) instead of traditional CNN network. Figure 1 : The architecture of our proposed CAN-biGRUA. In the first layer, convolutional neural network and self-attention mechanism are used to extract text features. In the second layer, biGRU with an attentive architecture on the top is used to model the sequence of utterances in the dialogue.", |
| "cite_spans": [ |
| { |
| "start": 63, |
| "end": 83, |
| "text": "(Poria et al., 2017)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 176, |
| "end": 194, |
| "text": "(Gao et al., 2018)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 355, |
| "end": 363, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "For query embedding Q, key embedding K and value embedding V involved in attention operation, they may need different effective features. And different effective information can be extracted in different convolution operations. So we obtain the Q, K and V embeddings by convolving the input word embeddings, instead of using the input word embeddings as the Q, K and V embeddings directly:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Q = f (conv(E, W q ) + b q ) (1) K = f (conv(E, W k ) + b k ) (2) V = f (conv(E, W v ) + b v )", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In the equations above, E is the input word embeddings, {E, Q, K, V } \u2208 R l\u00d7d , where l means the length of the sentence and d means the embedding dimension.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "{W q , W k , W v } \u2208 R w\u00d7n\u00d7d ,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "where w is the window size of filters and n is the feature maps of filters.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "{b q , b k , b v } \u2208 R d . conv(E, W )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "means convolution operation between E and W . And f is the activation function. After getting Q, K, V embeddings, we calculate semantic relations among words within the utterance by the scaled dot product attention operation. More specificly, Q and K operate to get the weight matrix. Then we scale this weight matix by \u221a d. After that, softmax operation is conducted to obtain the standardized weight matrix, which is used to express the degree of attention between words in the sentence, then the normalized weight matrix is mutiplied with V to get the result Z \u2208 R l\u00d7d of attention operation:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Z = sof tmax( QK T \u221a d )V", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "As mentioned in (Gao et al., 2018) , attention mechanisms cannot capture complex interactions, because it is designed for creating weighted averages. So we do the equations(1)-(3) twice to create Q a , K a , V a and Q b , K b , V b respectively, and get Z a , Z b by operate equation 4respectively, then perform elementwise multiplication on Z a and Z b to get U \u2208 R l\u00d7d :", |
| "cite_spans": [ |
| { |
| "start": 16, |
| "end": 34, |
| "text": "(Gao et al., 2018)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "U = Z a \u2297 Z b (5) Finally, for each U i (1 \u2264 i \u2264 N ) in dia = {U 1 , U 2 , .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": ".., U N }, we get the individual embedding e u i by max-pooling on the contextual word embeddings within the U i . In this way, we obtain a set of utterance embeddings {e u 1 , e u 2 , ..., e u N } in one dialogue.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Text Feature Extraction", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In this section, we discuss the sencond layer of the model. Considering different networks, we propose the hierarchical model CAN-GRU and its three progressive variants. CAN-GRU: In real life, when we analyze the emotion of the current utterance, we can only refer to the historical information of the past utterances in the conversation. So in our model, we use GRU to model the sequence of utterances in the dialogue, because it can memory and transmit historical information. GRU (Cho et al., 2014 ) is an improved model for the original recurrent neural networks and it performs well with simple calculation. At timestep t, it use reset gate R t and update gate Z t to calculate current hidden state S t with input utterance embedding e ut and hidden state s t\u22121 at the previous time step.", |
| "cite_spans": [ |
| { |
| "start": 483, |
| "end": 500, |
| "text": "(Cho et al., 2014", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "R t = \u03c3(e ut W ur + S t\u22121 W sr + b r ) (6) Z t = \u03c3(e ut W uz + S t\u22121 W sz + b z )", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "H t = tanh(e ut W uh + (R t \u2297 S t\u22121 )W sh + b h )", |
| "eq_num": "(8)" |
| } |
| ], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "S t = Z t \u2297 S t\u22121 + (1 \u2212 Z t ) \u2297 H t (9)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "where W, b are trainable parameters and \u2297 means elementwise mutilication. CAN-GRUA: However, it is difficult to grasp long-term dependence between sentences when there are too many sentences in a conversation. That is, it is hard for the current utterance to capture the historical information contained in the distant utterance. To solve this problem, we connect an attention layer upon the GRU to obtain the influence degree of historical information on the emotion of the current utterance. If the weight calculated by attention mechanism tends to be large, it indicates that the preceding utterance have an important influence on the current utterance, so this preceding utterance should be given more attention.S", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "t = t\u22121 i=1 S i \u03b1 i (10) \u03b1 i = exp(S t S i ) t\u22121 i=1 exp(S t S i )", |
| "eq_num": "(11)" |
| } |
| ], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Here, S t \u2208 R m is the current hidden state, where m is the dimension of the hidden state. S i \u2208 R m is the preceding hidden state at time step i,S t \u2208 R m is the attention result at time step t. CAN-biGRU: In fact, when analyzing the emotion of the utterance, we can not only use the historical information before the utterance, but also the future information after the current utterance. This is because emotional tone is usually maintained and does not shift frequently within a conversation in a short time. If we only pay attention to the historical information, it may be difficult to analyze the emotion of the current utterance, while the future information can be helpful in the analysis. Therefore, using both historical and future information can help to capture a richer context. Bidirectional GRU(biGRU) is used to model the sequence of utterances abstracting contextual features forward and backward, which can provides context for emotion classification more effectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialogue Modeling", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "As mentioned before, biGRU also suffers from the difficulty of obtaining semantic connections between long sequences. So we connect a self-attention layer on the top of the hidden states of biGRU to take full advantage of global contextual information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CAN-biGRUA:", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "S t = N i=1 S i \u03b1 i (12) \u03b1 i = exp(S t S i ) N i=1 exp(S t S i )", |
| "eq_num": "(13)" |
| } |
| ], |
| "section": "CAN-biGRUA:", |
| "sec_num": null |
| }, |
| { |
| "text": "Here, S t \u2208 R m is the current hidden state, S i \u2208 R m is the hidden state at time step i, N is the number of sentences in the dialogue,S t \u2208 R m is the attention result at time step t.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CAN-biGRUA:", |
| "sec_num": null |
| }, |
| { |
| "text": "As mentioned above, we get final representations of utterances {S 1 ,S 2 , ...,S t , ...,S N }. Then we utilize a fully-connected layer and a softmax layer to get the emotion class of each utterance in a dialogue.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "f t = tanh(W fSt + b f )", |
| "eq_num": "(14)" |
| } |
| ], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "o", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "t = sof tmax(W oft + b o ) (15) y t = argmax i (o t [i]), i \u2208 [1, c]", |
| "eq_num": "(16)" |
| } |
| ], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Where", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "W f \u2208 R m\u00d7m , b f \u2208 R m . W o \u2208 R m\u00d7c , c is the number of emotion class, b o \u2208 R c , o t \u2208 R c ,\u0177", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "t is the predicted class for utterance u t .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Emotion Classification", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Like (Khosla, 2018) , in order to solve the problem of emotion class imbalance, we use a weigted cross entropy loss as a minimization target to optimize the parameters in the model. We give higher weight to the loss of minority class data sample in the dataset.", |
| "cite_spans": [ |
| { |
| "start": 5, |
| "end": 19, |
| "text": "(Khosla, 2018)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "Loss = 1 K K k=1", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "weight k loss k (17)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "loss k = \u2212[y k log(p k ) + (1 \u2212 y k )log(1 \u2212 p k )]", |
| "eq_num": "(18)" |
| } |
| ], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "1 weight k = count i c i=1 count i", |
| "eq_num": "(19)" |
| } |
| ], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "Where K is the total number of samples, y k is the ground-truth, p k is the probability calculated in softmax layer, count i is the total number of samples in the same class as sample k.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training", |
| "sec_num": "3.5" |
| }, |
| { |
| "text": "We conduct experiments on two datasets provided by the EmotionX Challenge(Hsu and Ku, 2018).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Friends 0 : The conversations in this dataset are from the Friends TV show transcripts. The dataset contains eight emotion categories: joy, anger, sadness, surprise, fear, disgust, neutral, and non-neutral.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "EmotionPush 1 : The conversations in this dataset are from the facebook messenger logs after processing the private information. Emotion categories are the same as Friends dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "In the challenge(Hsu and Ku, 2018), each dataset is divided into the training set with 720 dilogues, the validation set with 80 dialogues and the test set with 200 dialogues. Since there are few utterances for some emotions, the challenge only evaluate the performence of recognition for four emotions: joy, anger, sadness and neutral. Table 2 shows the distributions (10,733) 80(1202) 200(2807) 1402100 514 9855 2133 Table 2 : Statistics of the datasets.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 336, |
| "end": 367, |
| "text": "Table 2 shows the distributions", |
| "ref_id": null |
| }, |
| { |
| "start": 418, |
| "end": 425, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Datasets", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We use the unweighted accuracy(UWA) as the evaluation metric instead of the weighted accuracy(WA), the same as the challenge. This is because WA is easily influced by the large proportion of neutral emotion and UWA can help to make a meaningful comparision.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metric", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "U W A = 1 c c i=1 a i , W A = c i=1 weight i a i (20)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metric", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Where a i is the accuracy of class i and weight i is the percentage of the class i.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metric", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We use 300-dimensional pre-trained GloVe 2 (Pennington et al., 2014) word-embeddings which is trained from web data. We use three distinct convolution filters of sizes 3, 4, and 5 respectively, each having 100 feature maps. The dimension of the hidden states of the GRU is set to 300. We use adam(Kingma and Ba, 2015) optimizer and set the initial learning rate as 1.0 \u00d7 10 \u22124 . The learning rate is halved every 20 epochs during training. Dropout probability is set to 0.3.", |
| "cite_spans": [ |
| { |
| "start": 43, |
| "end": 68, |
| "text": "(Pennington et al., 2014)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setting", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "In experiments, we compare our proposed model with the following models. CNN-DCNN: The winner of EmotionX Challenge (Khosla, 2018) . The model contains a convolutional encoder and a deconvolutional decoder. The linguistic features enhance the latent feature of the model.", |
| "cite_spans": [ |
| { |
| "start": 116, |
| "end": 130, |
| "text": "(Khosla, 2018)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baselines", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "The second place of the challenge (Luo et al., 2018) . A self-attentive biLSTM network can provide information between utterances and the word dependency in each utterance.", |
| "cite_spans": [ |
| { |
| "start": 34, |
| "end": 52, |
| "text": "(Luo et al., 2018)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SA-LSTM:", |
| "sec_num": null |
| }, |
| { |
| "text": "HAN: The third place of the challenge (Saxena et al., 2018) . LSTM with attention mechanism gets the sentence embedding. Another LSTM and CRF layer model the context dependency between sentence embeddings of the dialogue. scGRU:", |
| "cite_spans": [ |
| { |
| "start": 38, |
| "end": 59, |
| "text": "(Saxena et al., 2018)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SA-LSTM:", |
| "sec_num": null |
| }, |
| { |
| "text": "We implement the basic model proposed by (Poria et al., 2017) , but with a few changes. The same as (Poria et al., 2017) , CNN is used to extract text features, but we use a contextual GRU network instead of a contextual LSTM network to model the sequences.", |
| "cite_spans": [ |
| { |
| "start": 41, |
| "end": 61, |
| "text": "(Poria et al., 2017)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 100, |
| "end": 120, |
| "text": "(Poria et al., 2017)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SA-LSTM:", |
| "sec_num": null |
| }, |
| { |
| "text": "bcGRU:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "SA-LSTM:", |
| "sec_num": null |
| }, |
| { |
| "text": "We implement the variant model proposed by (Poria et al., 2017) , CNN is also used to obtain utterance features, but the biLSTM network used in the author's work is replaced by biGRU network. Table 3 presents the performance of baselines and CAN-GRU along with its variants.", |
| "cite_spans": [ |
| { |
| "start": 43, |
| "end": 63, |
| "text": "(Poria et al., 2017)", |
| "ref_id": "BIBREF17" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 192, |
| "end": 199, |
| "text": "Table 3", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "SA-LSTM:", |
| "sec_num": null |
| }, |
| { |
| "text": "Baselines: Our implemented bcGRU model performes better than scGRU on both datasets. On the Emotionpush dataset, bcGRU's performance has surpassed CNN-DCNN, and it is the best model in baselines. On the Friend dataset, CNN-DCNN remains the best baseline. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Main Results", |
| "sec_num": "4.5" |
| }, |
| { |
| "text": "In the first layer, it uses the convolutional self-attention mechanism to extract utterance features, and in the second layer, GRU is used to model the sequence of utterances. Compared with scGRU, it attains 3.2% and 1.7% improvement on the Friend dataset and EmotionPush dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CAN-GRU:", |
| "sec_num": null |
| }, |
| { |
| "text": "CAN-biGRU: Compared with CAN-GRU, it uses biGRU at the second layer and get improvements on the two datasets. CAN-biGRU achieves 2.8% and 1.9% improvements over bcGRU on the Friend dataset and the Emotionpush dataset respectively. Both the improvements of CAN-GRU and CAN-biGRU over baselines illustrate that the convolutional self-attention mechanism can capture contextual information in long text effectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CAN-GRU:", |
| "sec_num": null |
| }, |
| { |
| "text": "CAN-GRUA: Compared with CAN-GRU, an attention mechanism is connected upon the GRU layer, which can help the model better capture the historical information of utterance and give high weight to important historical information. It gets 1.8% and 2.0% improvements over CAN-GRU on the two datasets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "CAN-GRU:", |
| "sec_num": null |
| }, |
| { |
| "text": "CAN-biGRUA: At the top of biGRU, a self-attention mechanism is added to help calculate the importance of contextual information by using historical and future information when analyzing the current utterance emotion. This model achieves the best results, it improves 2.8% and 4.6% over baseline on the two datasets respectively. Table 4 : Experimental results for BERT and CAN-biGRU(*).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 329, |
| "end": 336, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "CAN-GRU:", |
| "sec_num": null |
| }, |
| { |
| "text": "In addition, we use the pretrained model BERT (Devlin et al., 2019) to get the word embeddings and input the pre-trained word embeddings into our CAN-biGRU, the experimental results are shown as the CAN-biGRU(*) in Table 4 . As we can see, while BERT achieves a high degree of accuracy, our model can be further improved on the basis of BERT. CAN-biGRU(*) gets 1.8% and 1.5% improvements over BERT on the Friend dataset and the Emotionpush dataset respectively.", |
| "cite_spans": [ |
| { |
| "start": 46, |
| "end": 67, |
| "text": "(Devlin et al., 2019)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 215, |
| "end": 222, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "CAN-GRU:", |
| "sec_num": null |
| }, |
| { |
| "text": "In Table 5 , we compare the emotion recognition results of bcGRU and CAN-biGRU. In the first case, 'bad' expresses strong emotion and both two model can recognize the sad emotion successfully. While there is no explicit emotion word in the third utterance, but the word 'ruined' delivers bad information, our CAN-biGRU can extract semantic information among words by CAN and gives the right prediction.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 3, |
| "end": 10, |
| "text": "Table 5", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Case Study", |
| "sec_num": "4.6" |
| }, |
| { |
| "text": "In the second case, the word 'celebrating' in the third utterance express the joy emotion implicitly. Our model obtains the contextual information through the CAN, and makes the correct prediction. As shown in Table 6 , we analyse some cases of the results of emotion recognition by our CAN-biGRUA. In the first two cases, our model can successfully recognise the emotion category of utterances. In the conversation of Monica and Joey, 'Yeah' expresses neutral emotion, while in the conversation of Chloe and Ross, 'Yeah' with '!' expresses stronger emotion and our model analyses its joy emotion effectively. However, in the third case, our model makes wrong classification for the first,third and fourth utterances.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 210, |
| "end": 217, |
| "text": "Table 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Case Study", |
| "sec_num": "4.6" |
| }, |
| { |
| "text": "This dialogue contains three different emotions and emotion shifts frequently. The failure of our model indicates that although considering the context, model's ability to understand emotions in the complicated situation is limited and still needs improvement. In Fig. 2 , we compare the recognition results and attention results of CAN-GRUA and CAN-biGRUA for the sixth utterance in the dialogue. As we can see, CAN-GRUA only uses historical information and focuses on the fourth utterance, and it takes neutral emotion as a result. While CAN-biGRUA takes both historical and future information into account, and it mainly pays attention to the seventh and the eighth utterances which contain strongly anger emotion, so the model finnaly classifies the test utterance as anger emotion. This case shows that considering both historical information and future information can help model make better classifications. Well,... be hurt over something that is so silly. sadness sadness Ross I mean, enough of the silliness! anger sadness Chandler Well, why don't you tell her to stop being silly! anger sadness Table 6 : Some cases of emotion recognition results by CAN-biGRUA", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 264, |
| "end": 270, |
| "text": "Fig. 2", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 1106, |
| "end": 1113, |
| "text": "Table 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Case Study", |
| "sec_num": "4.6" |
| }, |
| { |
| "text": "In the paper, we propose a hierarchical model(CAN-GRU) to tackle emotion recognition in dialogues. Unlike existing works, we focus on semantic information extraction within utterance in the dialogue. N-gram features and relevant semantic information among words in the utterance are learned by the convolutional self-attention network in the first layer and the sequence of utterances is modeled by the GRU-based network in the second layer. We improve CAN-GRU to three variants, CAN-biGRU, CAN-GRUA and CAN-biGRUA. Experimental results show that attention mechanism can help to grasp longterm dependency in the contexts effectively. CAN-biGRUA achieves better results than CAN-GRUA demonstrates that it is necessary to consider both past and future information of the utterance.In the future, we will try to explore deeper semantic information in the context and focus more on emotion shift to solve the problem of poor performance of the model in complex situations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "http://doraemon.iis.sinica.edu.tw/emotionlines 1 http://doraemon.iis.sinica.edu.tw/emotionlines", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://nlp.stanford.edu/projects/glove/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The work of this paper is funded by the project of National key research and development program of China (No. 2018YFC0830700).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Neural machine translation by jointly learning to align and translate", |
| "authors": [ |
| { |
| "first": "Dzmitry", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "3rd International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In 3rd International Conference on Learning Representations.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Learning phrase representations using RNN encoder-decoder for statistical machine translation", |
| "authors": [ |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Bart", |
| "middle": [], |
| "last": "Van Merrienboer", |
| "suffix": "" |
| }, |
| { |
| "first": "Dzmitry", |
| "middle": [], |
| "last": "Aglar G\u00fcl\u00e7ehre", |
| "suffix": "" |
| }, |
| { |
| "first": "Fethi", |
| "middle": [], |
| "last": "Bahdanau", |
| "suffix": "" |
| }, |
| { |
| "first": "Holger", |
| "middle": [], |
| "last": "Bougares", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Schwenk", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1724--1734", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kyunghyun Cho, Bart van Merrienboer, \u00c7 aglar G\u00fcl\u00e7ehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pages 1724-1734. ACL.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "BERT: pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Computational Linguistics Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019", |
| "volume": "1", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: pre-training of deep bidirec- tional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Computational Linguistics Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers).", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Hierarchical convolutional attention networks for text classification", |
| "authors": [ |
| { |
| "first": "Shang", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "Arvind", |
| "middle": [], |
| "last": "Ramanathan", |
| "suffix": "" |
| }, |
| { |
| "first": "Georgia", |
| "middle": [ |
| "D" |
| ], |
| "last": "Tourassi", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of The Third Workshop on Representation Learning for NLP, Rep4NLP@ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "11--23", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shang Gao, Arvind Ramanathan, and Georgia D. Tourassi. 2018. Hierarchical convolutional attention net- works for text classification. In Proceedings of The Third Workshop on Representation Learning for NLP, Rep4NLP@ACL, pages 11-23. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Dialoguegcn: A graph convolutional neural network for emotion recognition in conversation", |
| "authors": [ |
| { |
| "first": "Deepanway", |
| "middle": [], |
| "last": "Ghosal", |
| "suffix": "" |
| }, |
| { |
| "first": "Navonil", |
| "middle": [], |
| "last": "Majumder", |
| "suffix": "" |
| }, |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Niyati", |
| "middle": [], |
| "last": "Chhaya", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [ |
| "F" |
| ], |
| "last": "Gelbukh", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "154--164", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deepanway Ghosal, Navonil Majumder, Soujanya Poria, Niyati Chhaya, and Alexander F. Gelbukh. 2019. Di- aloguegcn: A graph convolutional neural network for emotion recognition in conversation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pages 154-164. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "ICON: interactive conversational memory network for multimodal emotion detection", |
| "authors": [ |
| { |
| "first": "Devamanyu", |
| "middle": [], |
| "last": "Hazarika", |
| "suffix": "" |
| }, |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Rada", |
| "middle": [], |
| "last": "Mihalcea", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| }, |
| { |
| "first": "Roger", |
| "middle": [], |
| "last": "Zimmermann", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2594--2604", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Devamanyu Hazarika, Soujanya Poria, Rada Mihalcea, Erik Cambria, and Roger Zimmermann. 2018a. ICON: interactive conversational memory network for multimodal emotion detection. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2594-2604. Association for Compu- tational Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Conversational memory network for emotion recognition in dyadic dialogue videos", |
| "authors": [ |
| { |
| "first": "Devamanyu", |
| "middle": [], |
| "last": "Hazarika", |
| "suffix": "" |
| }, |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Amir", |
| "middle": [], |
| "last": "Zadeh", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| }, |
| { |
| "first": "Louis-Philippe", |
| "middle": [], |
| "last": "Morency", |
| "suffix": "" |
| }, |
| { |
| "first": "Roger", |
| "middle": [], |
| "last": "Zimmermann", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "2122--2132", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Devamanyu Hazarika, Soujanya Poria, Amir Zadeh, Erik Cambria, Louis-Philippe Morency, and Roger Zimmer- mann. 2018b. Conversational memory network for emotion recognition in dyadic dialogue videos. In Proceed- ings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 2122-2132. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Long short-term memory", |
| "authors": [ |
| { |
| "first": "Sepp", |
| "middle": [], |
| "last": "Hochreiter", |
| "suffix": "" |
| }, |
| { |
| "first": "J\u00fcrgen", |
| "middle": [], |
| "last": "Schmidhuber", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Neural Computation", |
| "volume": "9", |
| "issue": "8", |
| "pages": "1735--1780", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural Computation, 9(8):1735- 1780.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Socialnlp 2018 emotionx challenge overview: Recognizing emotions in dialogues", |
| "authors": [ |
| { |
| "first": "Chun", |
| "middle": [], |
| "last": "Chao", |
| "suffix": "" |
| }, |
| { |
| "first": "Lun-Wei", |
| "middle": [], |
| "last": "Hsu", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ku", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "27--31", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chao-Chun Hsu and Lun-Wei Ku. 2018. Socialnlp 2018 emotionx challenge overview: Recognizing emotions in dialogues. In Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL, pages 27-31. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Emotionx-ar: CNN-DCNN autoencoder based emotion classifier", |
| "authors": [ |
| { |
| "first": "Sopan", |
| "middle": [], |
| "last": "Khosla", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "37--44", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sopan Khosla. 2018. Emotionx-ar: CNN-DCNN autoencoder based emotion classifier. In Lun-Wei Ku and Cheng-Te Li, editors, Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL, pages 37-44. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Convolutional neural networks for sentence classification", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1746--1751", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoon Kim. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Confer- ence on Empirical Methods in Natural Language Processing, pages 1746-1751. ACL.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Adam: A method for stochastic optimization", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Diederik", |
| "suffix": "" |
| }, |
| { |
| "first": "Jimmy", |
| "middle": [], |
| "last": "Kingma", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ba", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "3rd International Conference on Learning Representations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In 3rd International Conference on Learning Representations.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Emotionx-dlc: Self-attentive bilstm for detecting sequential emotions in dialogues", |
| "authors": [ |
| { |
| "first": "Linkai", |
| "middle": [], |
| "last": "Luo", |
| "suffix": "" |
| }, |
| { |
| "first": "Haiqing", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Francis", |
| "middle": [ |
| "Y L" |
| ], |
| "last": "Chin", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "32--36", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Linkai Luo, Haiqing Yang, and Francis Y. L. Chin. 2018. Emotionx-dlc: Self-attentive bilstm for detecting sequential emotions in dialogues. In Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL, pages 32-36. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Dialoguernn: An attentive RNN for emotion detection in conversations", |
| "authors": [ |
| { |
| "first": "Navonil", |
| "middle": [], |
| "last": "Majumder", |
| "suffix": "" |
| }, |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Devamanyu", |
| "middle": [], |
| "last": "Hazarika", |
| "suffix": "" |
| }, |
| { |
| "first": "Rada", |
| "middle": [], |
| "last": "Mihalcea", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [ |
| "F" |
| ], |
| "last": "Gelbukh", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence", |
| "volume": "2019", |
| "issue": "", |
| "pages": "6818--6825", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Navonil Majumder, Soujanya Poria, Devamanyu Hazarika, Rada Mihalcea, Alexander F. Gelbukh, and Erik Cam- bria. 2019. Dialoguernn: An attentive RNN for emotion detection in conversations. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelli- gence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 -February 1, 2019, pages 6818-6825. AAAI Press.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Human abilities: emotional intelligence. Annual review of psychology", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "John", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [ |
| "D" |
| ], |
| "last": "Mayer", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Roberts", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Barsade", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "", |
| "volume": "59", |
| "issue": "", |
| "pages": "507--536", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John D Mayer, Richard D Roberts, and Sigal G Barsade. 2008. Human abilities: emotional intelligence. Annual review of psychology, 59:507-536.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Glove: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. Glove: Global vectors for word rep- resentation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pages 1532-1543. ACL.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis", |
| "authors": [ |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexander", |
| "middle": [ |
| "F" |
| ], |
| "last": "Gelbukh", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "2539--2544", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soujanya Poria, Erik Cambria, and Alexander F. Gelbukh. 2015. Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 2539-2544. ACL.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Context-dependent sentiment analysis in user-generated videos", |
| "authors": [ |
| { |
| "first": "Soujanya", |
| "middle": [], |
| "last": "Poria", |
| "suffix": "" |
| }, |
| { |
| "first": "Erik", |
| "middle": [], |
| "last": "Cambria", |
| "suffix": "" |
| }, |
| { |
| "first": "Devamanyu", |
| "middle": [], |
| "last": "Hazarika", |
| "suffix": "" |
| }, |
| { |
| "first": "Navonil", |
| "middle": [], |
| "last": "Majumder", |
| "suffix": "" |
| }, |
| { |
| "first": "Amir", |
| "middle": [], |
| "last": "Zadeh", |
| "suffix": "" |
| }, |
| { |
| "first": "Louis-Philippe", |
| "middle": [], |
| "last": "Morency", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soujanya Poria, Erik Cambria, Devamanyu Hazarika, Navonil Majumder, Amir Zadeh, and Louis-Philippe Morency. 2017. Context-dependent sentiment analysis in user-generated videos. In Proceedings of the 55th", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Annual Meeting of the Association for Computational Linguistics", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "1", |
| "issue": "", |
| "pages": "873--883", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Annual Meeting of the Association for Computational Linguistics, Volume 1: Long Papers, pages 873-883. Association for Computational Linguistics. Computational Linguistics", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Emotionx-area66: Predicting emotions in dialogues using hierarchical attention network with sequence labeling", |
| "authors": [ |
| { |
| "first": "Rohit", |
| "middle": [], |
| "last": "Saxena", |
| "suffix": "" |
| }, |
| { |
| "first": "Savita", |
| "middle": [], |
| "last": "Bhat", |
| "suffix": "" |
| }, |
| { |
| "first": "Niranjan", |
| "middle": [], |
| "last": "Pedanekar", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "50--55", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rohit Saxena, Savita Bhat, and Niranjan Pedanekar. 2018. Emotionx-area66: Predicting emotions in dialogues using hierarchical attention network with sequence labeling. In Proceedings of the Sixth International Workshop on Natural Language Processing for Social Media, SocialNLP@ACL, pages 50-55. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Attention is all you need", |
| "authors": [ |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| }, |
| { |
| "first": "Llion", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Aidan", |
| "middle": [ |
| "N" |
| ], |
| "last": "Gomez", |
| "suffix": "" |
| }, |
| { |
| "first": "Lukasz", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "Illia", |
| "middle": [], |
| "last": "Polosukhin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "5998--6008", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems, pages 5998-6008.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "(a) Recognition results and attention results of CAN-GRUA (b) Recognition results and attention results of CAN-biGRUA" |
| }, |
| "FIGREF1": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Comparison of recognition results and attention results between CAN-GRUA and CAN-biGRUA. Deeper color means higher attention." |
| }, |
| "TABREF0": { |
| "content": "<table><tr><td>Such as Long Short-Term Memory Net-</td></tr></table>", |
| "text": "The word 'Yeah' expresses different emotions in the different contexts.", |
| "num": null, |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF1": { |
| "content": "<table><tr><td>dataset</td><td>train</td><td colspan=\"2\">Dialogue(Utterance) validation test</td><td colspan=\"2\">anger joy</td><td>Emotion sadness neutral others</td></tr><tr><td>Friend</td><td colspan=\"2\">720(10561)</td><td colspan=\"2\">80(1178) 200(2764) 759</td><td colspan=\"2\">1710 498</td><td>6530</td><td>5006</td></tr><tr><td colspan=\"2\">EmotionPush 720</td><td/><td/><td/><td/></tr></table>", |
| "text": "of train, validation, test samples and the distributions of the emotions for both datasets respectively.", |
| "num": null, |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF3": { |
| "content": "<table/>", |
| "text": "Experimental results on Friend dataset and EmotionPush dataset.", |
| "num": null, |
| "type_str": "table", |
| "html": null |
| }, |
| "TABREF6": { |
| "content": "<table/>", |
| "text": "Some case comparisons of emotion recognition results by bcGRU and CAN-biGRUA in the third case, both two model make false predictions for the utterance said by Phoebe, since the word 'sorry' expresses strong sad emotion. This shows CAN is still limited in such complicated semantic environment.", |
| "num": null, |
| "type_str": "table", |
| "html": null |
| } |
| } |
| } |
| } |