ACL-OCL / Base_JSON /prefixD /json /D16 /D16-1046.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "D16-1046",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T16:36:22.040576Z"
},
"title": "How Transferable are Neural Networks in NLP Applications?",
"authors": [
{
"first": "Lili",
"middle": [],
"last": "Mou",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": ""
},
{
"first": "Zhao",
"middle": [],
"last": "Meng",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": "zhaomeng.pku@outlook.com"
},
{
"first": "Rui",
"middle": [],
"last": "Yan",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Ge",
"middle": [],
"last": "Li",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": "lige@sei.pku.edu.cn"
},
{
"first": "Yan",
"middle": [],
"last": "Xu",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": ""
},
{
"first": "Lu",
"middle": [],
"last": "Zhang",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": "zhanglu@sei.pku.edu.cn"
},
{
"first": "Zhi",
"middle": [],
"last": "Jin",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Key Laboratory of High Confidence Software Technologies (Peking University)",
"location": {
"region": "MoE",
"country": "China"
}
},
"email": "zhijin@sei.pku.edu.cn"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Transfer learning is aimed to make use of valuable knowledge in a source domain to help model performance in a target domain. It is particularly important to neural networks, which are very likely to be overfitting. In some fields like image processing, many studies have shown the effectiveness of neural network-based transfer learning. For neural NLP, however, existing studies have only casually applied transfer learning, and conclusions are inconsistent. In this paper, we conduct systematic case studies and provide an illuminating picture on the transferability of neural networks in NLP. 1",
"pdf_parse": {
"paper_id": "D16-1046",
"_pdf_hash": "",
"abstract": [
{
"text": "Transfer learning is aimed to make use of valuable knowledge in a source domain to help model performance in a target domain. It is particularly important to neural networks, which are very likely to be overfitting. In some fields like image processing, many studies have shown the effectiveness of neural network-based transfer learning. For neural NLP, however, existing studies have only casually applied transfer learning, and conclusions are inconsistent. In this paper, we conduct systematic case studies and provide an illuminating picture on the transferability of neural networks in NLP. 1",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Transfer learning, or sometimes known as domain adaptation, 2 plays an important role in various natural language processing (NLP) applications, especially when we do not have large enough datasets for the task of interest (called the target task T ). In such scenarios, we would like to transfer or adapt knowledge from other domains (called the source domains/tasks S) so as to mitigate the problem of overfitting and to improve model performance in T . For traditional feature-rich or kernel-based models, researchers have developed a variety of elegant methods for domain adaptation; examples include EasyAdapt (Daum\u00e9 III, 2007 ; Daum\u00e9 III et * Yan Xu is currently a research scientist at Inveno Co., Ltd.",
"cite_spans": [
{
"start": 615,
"end": 631,
"text": "(Daum\u00e9 III, 2007",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "1 Code released on https://sites.google.com/site/transfernlp/ 2 In this paper, we do not distinguish the conceptual difference between transfer learning and domain adaptation. Domain-in the sense we use throughout this paper-is defined by datasets.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "al., 2010), instance weighting (Jiang and Zhai, 2007; Foster et al., 2010) , and structural correspondence learning (Blitzer et al., 2006; Prettenhofer and Stein, 2010) .",
"cite_spans": [
{
"start": 31,
"end": 53,
"text": "(Jiang and Zhai, 2007;",
"ref_id": "BIBREF13"
},
{
"start": 54,
"end": 74,
"text": "Foster et al., 2010)",
"ref_id": "BIBREF9"
},
{
"start": 116,
"end": 138,
"text": "(Blitzer et al., 2006;",
"ref_id": "BIBREF1"
},
{
"start": 139,
"end": 168,
"text": "Prettenhofer and Stein, 2010)",
"ref_id": "BIBREF19"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Recently, deep neural networks are emerging as the prevailing technical solution to almost every field in NLP. Although capable of learning highly nonlinear features, deep neural networks are very prone to overfitting, compared with traditional methods. Transfer learning therefore becomes even more important. Fortunately, neural networks can be trained in a transferable way by their incremental learning nature: we can directly use trained (tuned) parameters from a source task to initialize the network in the target task; alternatively, we may also train two tasks simultaneously with some parameters shared. But their performance should be verified by empirical experiments.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Existing studies have already shown some evidence of the transferability of neural features. For example, in image processing, low-level neural layers closely resemble Gabor filters or color blobs (Zeiler and Fergus, 2014; Krizhevsky et al., 2012) ; they can be transferred well to different tasks. Donahue et al. (2014) suggest that high-level layers are also transferable in general visual recognition; Yosinski et al. (2014) further investigate the transferability of neural layers in different levels of abstraction.",
"cite_spans": [
{
"start": 197,
"end": 222,
"text": "(Zeiler and Fergus, 2014;",
"ref_id": "BIBREF23"
},
{
"start": 223,
"end": 247,
"text": "Krizhevsky et al., 2012)",
"ref_id": "BIBREF14"
},
{
"start": 299,
"end": 320,
"text": "Donahue et al. (2014)",
"ref_id": "BIBREF7"
},
{
"start": 405,
"end": 427,
"text": "Yosinski et al. (2014)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Although transfer learning is promising in image processing, conclusions appear to be less clear in NLP applications. Image pixels are low-level signals, which are generally continuous and less related to semantics. By contrast, natural language tokens are discrete: each word well reflects the thought of humans, but neighboring words do not share as much information as pixels in images do. Previous neural NLP studies have casually applied transfer techniques, but their results are not consistent. Collobert and Weston (2008) apply multi-task learning to SRL, NER, POS, and CHK, 3 but obtain only 0.04-0.21% error reduction 4 (out of a base error rate of 16-18%). Bowman et al. 2015, on the contrary, improve a natural language inference task from an accuracy of 71.3% to 80.8% by initializing parameters with an additional dataset of 550,000 samples. Therefore, more systematic studies are needed to shed light on transferring neural networks in the field of NLP.",
"cite_spans": [
{
"start": 502,
"end": 529,
"text": "Collobert and Weston (2008)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In this paper, we investigate the question \"How transferable are neural networks in NLP applications?\"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "We distinguish two scenarios of transfer: (1) transferring knowledge to a semantically similar/equivalent task but with a different dataset; (2) transferring knowledge to a task that is semantically different but shares the same neural topology/architecture so that neural parameters can indeed be transferred. We further distinguish two transfer methods: (1) using the parameters trained on S to initialize T (INIT), and (2) multi-task learning (MULT), i.e., training S and T simultaneously. (Please see Sections 2 and 4). Our study mainly focuses on the following research questions: RQ1: How transferable are neural networks between two tasks with similar or different semantics in NLP applications?",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "RQ2: How transferable are different layers of NLP neural models?",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "RQ3: How transferable are INIT and MULT, respectively? What is the effect of combining these two methods?",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "We conducted extensive experiments over six datasets on classifying sentences and sentence pairs. We leveraged the widely-used convolutional neural network (CNN) and long short term memory (LSTM)-based recurrent neural network (RNN) as our models.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "Based on our experimental results, we have the following main observations, some of which are unexpected.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "\u2022 Whether a neural network is transferable in NLP depends largely on how semantically similar the tasks are, which is different from the consensus in image processing. \u2022 The output layer is mainly specific to the dataset and not transferable. Word embeddings are likely to be transferable to semantically different tasks. \u2022 MULT and INIT appear to be generally comparable to each other; combining these two methods does not result in further gain in our study.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "The rest of this paper is organized as follows. Section 2 introduces the datasets that neural models are transferred across; Section 3 details the neural architectures and experimental settings. We describe two approaches (INIT and MULT) to transfer learning in Section 4. We present experimental results in Sections 5-6 and have concluding remarks in Section 7.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Contributions",
"sec_num": null
},
{
"text": "In our study, we conducted two series of experiments using six open datasets as follows. In each experiment, the large dataset serves as the source domain and small ones are the target domains. Table 1 presents statistics of the above datasets.",
"cite_spans": [],
"ref_spans": [
{
"start": 194,
"end": 201,
"text": "Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2"
},
{
"text": "We distinguish two scenarios of transfer regarding semantic similarity: (1) semantically equivalent transfer (IMDB\u2192MR, SNLI\u2192SICK), that is, the tasks of S and T are defined by the same meaning, and (2) semantically different transfer (IMDB\u2192QC, SNLI\u2192MSRP). Examples are also illustrated in Table 1 to demonstrate semantic relatedness.",
"cite_spans": [],
"ref_spans": [
{
"start": 289,
"end": 296,
"text": "Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2"
},
{
"text": "It should be noticed that in image or speech processing (Yosinski et al., 2014; Wang and Zheng, 2015) , the input of neural networks pretty much consists of raw signals; hence, low-level feature detectors are almost always transferable, even if Yosinski et al. 2014manually distinguish artificial objects and natural ones in an image classification task.",
"cite_spans": [
{
"start": 56,
"end": 79,
"text": "(Yosinski et al., 2014;",
"ref_id": "BIBREF22"
},
{
"start": 80,
"end": 101,
"text": "Wang and Zheng, 2015)",
"ref_id": "BIBREF21"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2"
},
{
"text": "Distinguishing semantic relatedness-which emerges from very low layers of either word embeddings or the successive hidden layer-is specific to NLP and also a new insight of our paper. As we shall see in Sections 5 and 6, the transferability of neural networks in NLP is more sensitive to semantics than in image processing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2"
},
{
"text": "In each group, we used a single neural model to solve three problems in a unified manner. That is to say, the neural architecture is the same among the three datasets, which makes it possible to investigate transfer learning regardless of whether the tasks are semantically equivalent. Concretely, the neural models are as follows.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "\u2022 Experiment I: LSTM-RNN. To classify a sentence according to its sentiment or question type, we use a recurrent neural network (RNN, Figure 1a ) with long short term memory (LSTM) units (Hochreiter and Schmidhuber, 1997) . A softmax layer is added to the last word's hidden state for classification.",
"cite_spans": [
{
"start": 187,
"end": 221,
"text": "(Hochreiter and Schmidhuber, 1997)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [
{
"start": 134,
"end": 143,
"text": "Figure 1a",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "\u2022 Experiment II: CNN-pair. In this group, we use a \"Siamese\" architecture (Bromley et al., 1993) to classify the relation of two sentences. We first apply a convolutional neural network (CNN, Figure 1b ) with a window size of 5 to model local context, and a max pooling layer gathers information to a fixed-size vector. Then the sentence vectors are concatenated and fed to a hidden layer before the softmax output.",
"cite_spans": [
{
"start": 74,
"end": 96,
"text": "(Bromley et al., 1993)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [
{
"start": 192,
"end": 201,
"text": "Figure 1b",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "In our experiments, embeddings were pretrained by word2vec (Mikolov et al., 2013) ; all embeddings and hidden layers were 100 dimensional. We applied stochastic gradient descent with a minibatch size of 50 for optimization. In each setting, we tuned the hyperparameters as follows: learning rate from {3, 1, 0.3, 0.1, 0.03}, power decay of learning rate from {fast, moderate, low} (defined by how much, after one epoch, the learning rate residual is: 0.1x, 0.3x, 0.9x, resp). We regularized our network by dropout with a rate from {0, 0.1, 0.2, 0.3}. Note that we might not run nonsensical settings, e.g., a larger dropout rate if the network has already been underfitting (i.e., accuracy has decreased when the dropout rate increases). We report the test performance associated with the highest validation accuracy.",
"cite_spans": [
{
"start": 59,
"end": 81,
"text": "(Mikolov et al., 2013)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "To setup a baseline, we trained our models without transfer 5 times by different random parameter initializations (Table 2) . We have achieved reasonable performance that is comparable to similar models reported in the literature with all six datasets. Therefore, our implementation is fair and suitable for further study of transfer learning.",
"cite_spans": [],
"ref_spans": [
{
"start": 114,
"end": 123,
"text": "(Table 2)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "It should be mentioned that the goal of this paper is not to outperform state-of-the-art results; instead, Table 2 : Accuracy (%) without transfer. We also include related models for comparison (Dong et al., 2015; Socher et al., 2011; Zhao et al., 2015; Bowman et al., 2015; Hu et al., 2014) , showing that we have achieved comparable results, and thus are ready to investigate transfer learning. The models were run one only once in source domains, because we could only transfer a particular model instead of an average of several models.",
"cite_spans": [
{
"start": 194,
"end": 213,
"text": "(Dong et al., 2015;",
"ref_id": "BIBREF21"
},
{
"start": 214,
"end": 234,
"text": "Socher et al., 2011;",
"ref_id": "BIBREF20"
},
{
"start": 235,
"end": 253,
"text": "Zhao et al., 2015;",
"ref_id": "BIBREF24"
},
{
"start": 254,
"end": 274,
"text": "Bowman et al., 2015;",
"ref_id": "BIBREF2"
},
{
"start": 275,
"end": 291,
"text": "Hu et al., 2014)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [
{
"start": 107,
"end": 114,
"text": "Table 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "we would like to conduct a fair comparison of different methods and settings for transfer learning in NLP.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Neural Models and Settings",
"sec_num": "3"
},
{
"text": "Transfer learning aims to use knowledge in a source domain to aid the target domain. As neural networks are usually trained incrementally with gradient descent (or variants), it is straightforward to use gradient information in both source and target domains for optimization so as to accomplish knowledge transfer. Depending on how samples in source and target domains are scheduled, there are two main approaches to neural network-based transfer learning:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "\u2022 Parameter initialization (INIT). The INIT approach first trains the network on S, and then directly uses the tuned parameters to initialize the network for T . After transfer, we may fix ( \u2642 ) the parameters in the target domain (Glorot et al., 2011) , i.e., no training is performed on T . But when labeled data are available in T , it would be better to fine-tune ( ) the parameters.",
"cite_spans": [
{
"start": 231,
"end": 252,
"text": "(Glorot et al., 2011)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "INIT is also related to unsupervised pretraining such as word embedding learning (Mikolov et al., 2013) and autoencoders (Bengio et al., 2006) . In these approaches, parameters that are (pre)trained in an unsupervised way are transferred to initialize the model for a supervised task (Plank and Moschitti, 2013) . However, our paper focuses on \"supervised pretraining,\" which means we transfer knowledge from a labeled source domain.",
"cite_spans": [
{
"start": 81,
"end": 103,
"text": "(Mikolov et al., 2013)",
"ref_id": "BIBREF16"
},
{
"start": 121,
"end": 142,
"text": "(Bengio et al., 2006)",
"ref_id": "BIBREF0"
},
{
"start": 284,
"end": 311,
"text": "(Plank and Moschitti, 2013)",
"ref_id": "BIBREF18"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "\u2022 Multi-task learning (MULT). MULT, on the other hand, simultaneously trains samples in both domains (Collobert and Weston, 2008; Liu et al., 2016) . The overall cost function is given by",
"cite_spans": [
{
"start": 101,
"end": 129,
"text": "(Collobert and Weston, 2008;",
"ref_id": "BIBREF4"
},
{
"start": 130,
"end": 147,
"text": "Liu et al., 2016)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "J = \u03bbJ T + (1 \u2212 \u03bb)J S (1)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "where J T and J S are the individual cost function of each domain. (Both J T and J S are normalized by the number of training samples.) \u03bb \u2208 (0, 1) is a hyperparameter balancing the two domains.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "It is nontrivial to optimize Equation 1 in practice by gradient-based methods. One may take the partial derivative of J and thus \u03bb goes to the learning rate (Liu et al., 2016 ), but the model is then vulnerable because it is likely to blow up with large learning rates (multiplied by \u03bb or 1 \u2212 \u03bb) and be stuck in local optima with small ones. Collobert and Weston (2008) alternatively choose a data sample from either domain with a certain probability (controlled by \u03bb) and take the derivative for the particular data sample. In this way, domain transfer is independent of learning rates, but we may not be able to fully use the entire dataset of S if \u03bb is large. We adopted the latter approach in our experiment for simplicity. (More in-depth analysis may be needed in future work.) Formally, our multi-task learning strategy is as follows.",
"cite_spans": [
{
"start": 157,
"end": 174,
"text": "(Liu et al., 2016",
"ref_id": "BIBREF15"
},
{
"start": 342,
"end": 369,
"text": "Collobert and Weston (2008)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "1 Switch to T with prob. \u03bb, or to S with prob. 1 \u2212 \u03bb. 2 Compute the gradient of the next data sample in the particular domain.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "Further, INIT and MULT can be combined straightforwardly, and we obtain the third setting:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "\u2022 Combination (MULT+INIT). We first pretrain on the source domain S for parameter initialization, and then train S and T simultaneously.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "From a theoretical perspective, INIT and MULT work in different ways. In the MULT approach, the source domain regularizes the model by \"aliasing\" the error surface of the target domain; hence the neural network is less prone to overfitting. In INIT, T 's error surface remains intact. Before training on the target dataset, the parameters are initialized in such a meaningful way that they contain additional knowledge in the source domain. However, in an extreme case where T 's error surface is convex, INIT is ineffective because the parameters can reach the global optimum regardless of their initialization. In practice, deep neural networks usually have highly complicated, non-convex error surfaces. By properly initializing parameters with the knowledge of S, we can reasonably expect that the parameters are in a better \"catchment basin,\" and that the INIT approach can transfer knowledge from S to T .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Transfer Methods",
"sec_num": "4"
},
{
"text": "We first analyze how INIT behaves in NLP-based transfer learning. In addition to two different transfer scenarios regarding semantic relatedness as described in Section 2, we further evaluated two settings: (1) fine-tuning parameters , and (2) freezing parameters after transfer \u2642 . Existing evidence shows that frozen parameters would generally hurt the performance (Peng et al., 2015) , but this setting provides a more direct understanding on how transferable the features are (because the factor of target domain optimization is ruled out). Therefore, we included it in our experiments. Moreover, we transferred parameters layer by layer to answer our second research question.",
"cite_spans": [
{
"start": 367,
"end": 386,
"text": "(Peng et al., 2015)",
"ref_id": "BIBREF17"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Results of Transferring by INIT",
"sec_num": "5"
},
{
"text": "Through Subsections 5.1-5.3, we initialized the parameters of T with the ones corresponding to the highest validation accuracy of S. In Subsection 5.4, we further investigated when the parameters are ready to be transferred during the training on S. Table 3 shows the main results of INIT. A quick observation is that, in both groups, transfer learning of semantically equivalent tasks (IMDB\u2192MR, SNLI\u2192SICK) appears to be successful with an improvement of \u223c6%. The results are not surprising and also reported in Bowman et al. (2015) .",
"cite_spans": [
{
"start": 512,
"end": 532,
"text": "Bowman et al. (2015)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [
{
"start": 250,
"end": 257,
"text": "Table 3",
"ref_id": "TABREF4"
}
],
"eq_spans": [],
"section": "Results of Transferring by INIT",
"sec_num": "5"
},
{
"text": "For IMDB\u2192QC and SNLI\u2192MSRP, however, there is no improvement of transferring hidden layers (embeddings excluded), namely LSTM-RNN units and CNN feature maps. The E H O setting yields a slight degradation of 0.2-0.4%, \u223c.5x std. The incapability of transferring is also proved by locking embeddings and hidden layers (E \u2642 H \u2642 O ). We see in this setting, the test performance is very low in QC or even worse than majority-class guess in MSRP. By further examining its training accuracy, which is 48.2% and 65.5%, respectively, we conclude that extracted features by LSTM-RNN and CNN models in S are almost irrelevant to the ultimate tasks T (QC and MSRP).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Overall Performance",
"sec_num": "5.1"
},
{
"text": "Although in previous studies, researchers have mainly drawn positive conclusions about transfer learning, we find a negative result similar to ours upon careful examination of Collobert and Weston (2008) , and unfortunately, their results may be somewhat misinterpreted. In that paper, the authors report transferring NER, POS, CHK, and pretrained word embeddings improves the SRL task by 1.91-3.90% accuracy (out of 16.54-18.40% error rate), but their gain is mainly due to word embeddings. In the settings that use pretrained word embeddings (which is common in NLP), NER, POS, and CHK together improve the SRL accuracy by only 0.04-0.21%.",
"cite_spans": [
{
"start": 176,
"end": 203,
"text": "Collobert and Weston (2008)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Overall Performance",
"sec_num": "5.1"
},
{
"text": "The above results are rather frustrating, indicating for RQ1 that neural networks may not be transferable to NLP tasks of different semantics. Transfer learning for NLP is more prone to semantics than the image processing domain, where even highlevel feature detectors are almost always transferable (Donahue et al., 2014; Yosinski et al., 2014) .",
"cite_spans": [
{
"start": 300,
"end": 322,
"text": "(Donahue et al., 2014;",
"ref_id": "BIBREF7"
},
{
"start": 323,
"end": 345,
"text": "Yosinski et al., 2014)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Overall Performance",
"sec_num": "5.1"
},
{
"text": "To answer RQ2, we next analyze the transferability of each layer. First, we freeze both embeddings and hidden layers (E \u2642 H \u2642 ). Even in semantically equivalent settings, if we further freeze the output layer (O \u2642 ), the performance in both IMDB\u2192MR and SNLI\u2192SICK drops, but by randomly initializing the output layer's parameters (O ), we can obtain a similar or higher result compared with the baseline (E H O ). The finding suggests that the output layer is mainly specific to a dataset. Transferring the output layer's parameters yields little (if any) gain.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Layer-by-Layer Analysis",
"sec_num": "5.2"
},
{
"text": "Regarding embeddings and hidden layers (in the settings E H O /E H O vs. E H O ), the IMDB\u2192MR experiment suggests both of embeddings and the hidden layer play an important role, each improving the accuracy by 3%. In SNLI\u2192SICK, however, the main improvement lies in the hidden layer. A plausible explanation is that in sentiment classification tasks (IMDB and MR), information emerges from raw input, i.e., sentiment lexicons and thus their embeddings, but natural language inference tasks (SNLI and SICK) address more on semantic compositionality and thus hidden layers are more important. Moreover, for semantically different tasks (IMDB\u2192QC and SNLI\u2192MSRP), the embeddings are the only parameters that have been observed to be transferable, slightly benefiting the target task by 2.7x and 1.8x std, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Layer-by-Layer Analysis",
"sec_num": "5.2"
},
{
"text": "Bowman et al. (2015) suggest that after transferring, a large learning rate may damage the knowledge stored in the parameters; in their paper, they transfer the learning rate information (AdaDelta) from S to T in addition to the parameters. Although the rule of the thumb is to choose all hyperparameters-including the learning rate-by validation, we are curious whether the above conjecture holds. Estimating a rough range of sensible hyperparameters can ease the burden of model selection; it also provides evidence to better understand how transfer learning actually works.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "How does learning rate affect transfer?",
"sec_num": "5.3"
},
{
"text": "We plot the learning curves of different learning rates \u03b1 in Figure 2 (IMDB\u2192MR and SNLI\u2192SICK, E H O ). (In the figure, no learning rate decay is applied.) As we see, with a large learning rate like \u03b1 = 0.3, the accuracy increases fast and peaks at earlier epochs. Training with a small learning rate (e.g., \u03b1 = 0.01) is slow, but its peak performance is comparable to large learning rates when iterated by, say, 100 epochs. The learning curves in Figure 2 are similar to classic speed/variance trade-off, and we have the following additional discovery:",
"cite_spans": [],
"ref_spans": [
{
"start": 61,
"end": 69,
"text": "Figure 2",
"ref_id": "FIGREF1"
},
{
"start": 447,
"end": 455,
"text": "Figure 2",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "How does learning rate affect transfer?",
"sec_num": "5.3"
},
{
"text": "In INIT, transferring learning rate information is not necessarily useful. A large learning rate does not damage the knowledge stored in the pretrained hyperparameters, but accelerates the training process to a large extent. In all, we may need to perform validation to choose the learning rate if computational resources are available. Learning curve of SNLI ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "How does learning rate affect transfer?",
"sec_num": "5.3"
},
{
"text": "In the above experiments, we transfer the parameters when they achieve the highest validation performance on S. This is a straightforward and intuitive practice. However, we may imagine that the parameters well-tuned to the source dataset may be too specific to it, i.e., the model overfits S and thus may underfit T . Another advantage of early transfer lies in com-putational concerns. If we manage to transfer model parameters after one or a few epochs on S, we can save much time especially when S is large.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "When is it ready to transfer?",
"sec_num": "5.4"
},
{
"text": "We therefore made efforts in studying when the neural model is ready to be transferred. Figures 3a and 3c plot the learning curves of the source tasks. The accuracy increases sharply from epochs 1-5; later, it reaches a plateau but is still growing slowly.",
"cite_spans": [],
"ref_spans": [
{
"start": 88,
"end": 106,
"text": "Figures 3a and 3c",
"ref_id": "FIGREF3"
}
],
"eq_spans": [],
"section": "When is it ready to transfer?",
"sec_num": "5.4"
},
{
"text": "We then transferred the parameters at different stages (epochs) of training to target tasks (also with the setting E H O ). Their accuracies are plotted in Figures 3b and 3d .",
"cite_spans": [],
"ref_spans": [
{
"start": 156,
"end": 173,
"text": "Figures 3b and 3d",
"ref_id": "FIGREF3"
}
],
"eq_spans": [],
"section": "When is it ready to transfer?",
"sec_num": "5.4"
},
{
"text": "In IMDB\u2192MR, the source performance and transferring performance align well. The SNLI\u2192SICK experiment, however, produces interesting yet unexpected results. Using the second epoch of SNLI's training yields the highest transfer performance on SICK, i.e., 78.98%, when the SNLI performance itself is comparatively low (72.65% vs. 76.26% at epoch 23). Later, the transfer performance decreases gradually by \u223c2.7%. The results in these two experiments are inconsistent and lack explanation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "When is it ready to transfer?",
"sec_num": "5.4"
},
{
"text": "To answer RQ3, we investigate how multi-task learning performs in transferring knowledge, as well as the effect of the combination of MULT and INIT. In this section, we applied the setting: sharing embeddings and hidden layers (denoted as E\u2665H\u2665O ), analogous to E H O in INIT. When combining MULT and INIT, we used the pretrained parameters of embeddings and hidden layers on S to initialize the multi-task training of S and T , visually represented by E \u2665H \u2665O .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "MULT, and its Combination with INIT",
"sec_num": "6"
},
{
"text": "In both MULT and MULT+INIT, we had a hyperparameter \u03bb \u2208 (0, 1) balancing the source and target tasks (defined in Section 4). \u03bb was tuned with a granularity of 0.1. As a friendly reminder, \u03bb = 1 refers to using T only; \u03bb = 0 refers to using S only. After finding that a small \u03bb yields high performance of MULT in the IMDB+MR and SNLI+SICK experiments (thick blue lines in Figures 4a and 4c) , we further tuned the \u03bb from 0.01 to 0.09 with a finegrained granularity of 0.02.",
"cite_spans": [],
"ref_spans": [
{
"start": 371,
"end": 389,
"text": "Figures 4a and 4c)",
"ref_id": "FIGREF4"
}
],
"eq_spans": [],
"section": "MULT, and its Combination with INIT",
"sec_num": "6"
},
{
"text": "The results are shown in Figure 4 . From the green curves in the 2nd and 4th subplots, we see MULT (with or without INIT) does not improve the accu- racy of target tasks (QC and MSRP); the inability to transfer is cross-checked by the INIT method in Section 5. For MR and SICK, on the other hand, transferability of the neural model is also consistently positive (blue curves in Figures 4a and 4c) , supporting our conclusion to RQ1 that neural trans-fer learning in NLP depends largely on how similar in semantics the source and target datasets are. Moreover, we see that the peak performance of MULT is slightly lower than INIT in Experiment I (Figure 4a ), but higher in Experiment II ( Figure 4c) ; they are in the same ballpark.",
"cite_spans": [],
"ref_spans": [
{
"start": 25,
"end": 33,
"text": "Figure 4",
"ref_id": "FIGREF4"
},
{
"start": 379,
"end": 397,
"text": "Figures 4a and 4c)",
"ref_id": "FIGREF4"
},
{
"start": 646,
"end": 656,
"text": "(Figure 4a",
"ref_id": "FIGREF4"
},
{
"start": 690,
"end": 700,
"text": "Figure 4c)",
"ref_id": "FIGREF4"
}
],
"eq_spans": [],
"section": "MULT, and its Combination with INIT",
"sec_num": "6"
},
{
"text": "In MULT+INIT (E \u2665H \u2665O ), the transfer performance of MULT+INIT remains high for different values of \u03bb. Because the parameters given by INIT have already conveyed sufficient information about the source task, MULT+INIT consistently outperforms non-transferring by a large margin. Its peak performance, however, is not higher than MULT or INIT. In summary, we answer our RQ3 as follows: in our experiments, MULT and INIT are generally comparable; we do not obtain further gain by combining MULT and INIT.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "MULT, and its Combination with INIT",
"sec_num": "6"
},
{
"text": "In this paper, we addressed the problem of transfer learning in neural network-based NLP applications. We conducted two series of experiments on six datasets, showing that the transferability of neural NLP models depends largely on the semantic relatedness of the source and target tasks, which is different from other domains like image processing. We analyzed the behavior of different neural layers. We also experimented with two transfer methods: parameter initialization (INIT) and multi-task learning (MULT). Besides, we reported two additional studies in Sections 5.3 and 5.4 (not repeated here). Our paper provides insight on the transferability of neural NLP models; the results also help to better understand neural features in general.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Concluding Remarks",
"sec_num": "7"
},
{
"text": "How transferable are the conclusions in this paper? We have to concede that empirical studies are subject to a variety of factors (e.g., models, tasks, datasets), and that conclusions may vary in different scenarios. In our paper, we have tested all results on two groups of experiments involving 6 datasets and 2 neural models (CNN and LSTM-RNN). Both models and tasks are widely studied in the literature, and not chosen deliberately. Results are mostly consistent (except Section 5.4). Along with analyzing our own experimental data, we have also collected related results in previous studies, serving as additional evidence in answering our research questions. Therefore, we think the generality of this work is fair and that the conclusions can be generalized to similar scenarios.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Concluding Remarks",
"sec_num": "7"
},
{
"text": "Future work. Our work also points out some future directions of research. For example, we would like to analyze the effect of different MULT strategies. More efforts are also needed in developing an effective yet robust method for multi-task learning.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Concluding Remarks",
"sec_num": "7"
},
{
"text": "The acronyms refer to semantic role labeling, named entity recognition, part-of-speech tagging, and chunking, respectively.4 Here, we quote the accuracies obtained by using unsupervised pretraining of word embeddings. This is the highest performance in that paper; using pretrained word embeddings is also a common practice in the literature.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "https://drive.google.com/file/d/ 0B8yp1gOBCztyN0JaMDVoeXhHWm8/ 6 https://www.cs.cornell.edu/people/pabo/ movie-review-data/ 7 http://cogcomp.cs.illinois.edu/Data/QA/QC/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "http://nlp.stanford.edu/projects/snli/ 9 http://http://alt.qcri.org/semeval2014/task1/ 10 http://research.microsoft.com/en-us/downloads/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "We thank all reviewers for their constructive comments, Sam Bowman for helpful suggestion, and Vicky Li for discussion on the manuscript. This research is supported by the National Basic Research Program of China (the 973 Program) under Grant No. 2015CB352201 and the National Natural Science Foundation of China under Grant Nos. 61232015, 91318301, 61421091, 61225007, and 61502014. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Greedy layer-wise training of deep networks",
"authors": [
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
},
{
"first": "Pascal",
"middle": [],
"last": "Lamblin",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Popovici",
"suffix": ""
},
{
"first": "Hugo",
"middle": [],
"last": "Larochelle",
"suffix": ""
}
],
"year": 2006,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "153--160",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yoshua Bengio, Pascal Lamblin, Dan Popovici, and Hugo Larochelle. 2006. Greedy layer-wise train- ing of deep networks. In Advances in Neural In- formation Processing Systems, pages 153-160.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Domain adaptation with structural correspondence learning",
"authors": [
{
"first": "John",
"middle": [],
"last": "Blitzer",
"suffix": ""
},
{
"first": "Ryan",
"middle": [],
"last": "Mcdonald",
"suffix": ""
},
{
"first": "Fernando",
"middle": [],
"last": "Pereira",
"suffix": ""
}
],
"year": 2006,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "120--128",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "John Blitzer, Ryan McDonald, and Fernando Pereira. 2006. Domain adaptation with struc- tural correspondence learning. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 120-128.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "A large annotated corpus for learning natural language inference",
"authors": [
{
"first": "R",
"middle": [],
"last": "Samuel",
"suffix": ""
},
{
"first": "Gabor",
"middle": [],
"last": "Bowman",
"suffix": ""
},
{
"first": "Christopher",
"middle": [],
"last": "Angeli",
"suffix": ""
},
{
"first": "Christopher",
"middle": [
"D"
],
"last": "Potts",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Manning",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "632--642",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural lan- guage inference. In Proceedings of the Confer- ence on Empirical Methods in Natural Language Processing, pages 632-642.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Signature verification using a \"Siamese\" time delay neural network",
"authors": [
{
"first": "Jane",
"middle": [],
"last": "Bromley",
"suffix": ""
},
{
"first": "W",
"middle": [],
"last": "James",
"suffix": ""
},
{
"first": "L\u00e9on",
"middle": [],
"last": "Bentz",
"suffix": ""
},
{
"first": "Isabelle",
"middle": [],
"last": "Bottou",
"suffix": ""
},
{
"first": "Yann",
"middle": [],
"last": "Guyon",
"suffix": ""
},
{
"first": "Cliff",
"middle": [],
"last": "Lecun",
"suffix": ""
},
{
"first": "Eduard",
"middle": [],
"last": "Moore",
"suffix": ""
},
{
"first": "Roopak",
"middle": [],
"last": "S\u00e4ckinger",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Shah",
"suffix": ""
}
],
"year": 1993,
"venue": "International Journal of Pattern Recognition and Artificial Intelligence",
"volume": "7",
"issue": "04",
"pages": "669--688",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jane Bromley, James W Bentz, L\u00e9on Bottou, Is- abelle Guyon, Yann LeCun, Cliff Moore, Eduard S\u00e4ckinger, and Roopak Shah. 1993. Signature verification using a \"Siamese\" time delay neural network. International Journal of Pattern Recog- nition and Artificial Intelligence, 7(04):669-688.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "A unified architecture for natural language processing: Deep neural networks with multitask learning",
"authors": [
{
"first": "Ronan",
"middle": [],
"last": "Collobert",
"suffix": ""
},
{
"first": "Jason",
"middle": [],
"last": "Weston",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the 25th International Conference on Machine Learning",
"volume": "",
"issue": "",
"pages": "160--167",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ronan Collobert and Jason Weston. 2008. A uni- fied architecture for natural language processing: Deep neural networks with multitask learning. In Proceedings of the 25th International Conference on Machine Learning, pages 160-167.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Frustratingly easy semi-supervised domain adaptation",
"authors": [
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Abhishek",
"middle": [],
"last": "Kumar",
"suffix": ""
},
{
"first": "Avishek",
"middle": [],
"last": "Saha",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the Workshop on Domain Adaptation for Natural Language Processing",
"volume": "",
"issue": "",
"pages": "53--59",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal Daum\u00e9 III, Abhishek Kumar, and Avishek Saha. 2010. Frustratingly easy semi-supervised domain adaptation. In Proceedings of the Workshop on Domain Adaptation for Natural Language Pro- cessing, pages 53-59.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Frustratingly easy domain adaptation",
"authors": [
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics",
"volume": "",
"issue": "",
"pages": "256--263",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal Daum\u00e9 III. 2007. Frustratingly easy domain adaptation. In Proceedings of the 45th Annual Meeting of the Association of Computational Lin- guistics, pages 256-263.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "DeCAF: A deep convolutional activation feature for generic visual recognition",
"authors": [
{
"first": "Jeff",
"middle": [],
"last": "Donahue",
"suffix": ""
},
{
"first": "Yangqing",
"middle": [],
"last": "Jia",
"suffix": ""
},
{
"first": "Oriol",
"middle": [],
"last": "Vinyals",
"suffix": ""
},
{
"first": "Judy",
"middle": [],
"last": "Hoffman",
"suffix": ""
},
{
"first": "Ning",
"middle": [],
"last": "Zhang",
"suffix": ""
},
{
"first": "Eric",
"middle": [],
"last": "Tzeng",
"suffix": ""
},
{
"first": "Trevor",
"middle": [],
"last": "Darrell",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the 31st International Conference on Machine Learning",
"volume": "",
"issue": "",
"pages": "647--655",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor Darrell. 2014. DeCAF: A deep convolutional ac- tivation feature for generic visual recognition. In Proceedings of the 31st International Conference on Machine Learning, pages 647-655.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "A statistical parsing framework for sentiment classification",
"authors": [
{
"first": "Li",
"middle": [],
"last": "Dong",
"suffix": ""
},
{
"first": "Furu",
"middle": [],
"last": "Wei",
"suffix": ""
},
{
"first": "Shujie",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Ming",
"middle": [],
"last": "Zhou",
"suffix": ""
},
{
"first": "Ke",
"middle": [],
"last": "Xu",
"suffix": ""
}
],
"year": 2015,
"venue": "Computational Linguistics",
"volume": "41",
"issue": "2",
"pages": "293--336",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Li Dong, Furu Wei, Shujie Liu, Ming Zhou, and Ke Xu. 2015. A statistical parsing framework for sentiment classification. Computational Lin- guistics, 41(2):293-336.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Discriminative instance weighting for domain adaptation in statistical machine translation",
"authors": [
{
"first": "George",
"middle": [],
"last": "Foster",
"suffix": ""
},
{
"first": "Cyril",
"middle": [],
"last": "Goutte",
"suffix": ""
},
{
"first": "Roland",
"middle": [],
"last": "Kuhn",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "451--459",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "George Foster, Cyril Goutte, and Roland Kuhn. 2010. Discriminative instance weighting for do- main adaptation in statistical machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 451-459.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Domain adaptation for large-scale sentiment classification: A deep learning approach",
"authors": [
{
"first": "Xavier",
"middle": [],
"last": "Glorot",
"suffix": ""
},
{
"first": "Antoine",
"middle": [],
"last": "Bordes",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 28th International Conference on Machine Learning",
"volume": "",
"issue": "",
"pages": "513--520",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Xavier Glorot, Antoine Bordes, and Yoshua Bengio. 2011. Domain adaptation for large-scale senti- ment classification: A deep learning approach. In Proceedings of the 28th International Conference on Machine Learning, pages 513-520.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Long short-term memory",
"authors": [
{
"first": "Sepp",
"middle": [],
"last": "Hochreiter",
"suffix": ""
},
{
"first": "J\u00fcrgen",
"middle": [],
"last": "Schmidhuber",
"suffix": ""
}
],
"year": 1997,
"venue": "Neural Computation",
"volume": "9",
"issue": "8",
"pages": "1735--1780",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural Computation, 9(8):1735-1780.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Convolutional neural network architectures for matching natural language sentences",
"authors": [
{
"first": "Baotian",
"middle": [],
"last": "Hu",
"suffix": ""
},
{
"first": "Zhengdong",
"middle": [],
"last": "Lu",
"suffix": ""
},
{
"first": "Hang",
"middle": [],
"last": "Li",
"suffix": ""
},
{
"first": "Qingcai",
"middle": [],
"last": "Chen",
"suffix": ""
}
],
"year": 2014,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "2042--2050",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Baotian Hu, Zhengdong Lu, Hang Li, and Qing- cai Chen. 2014. Convolutional neural network architectures for matching natural language sen- tences. In Advances in Neural Information Pro- cessing Systems, pages 2042-2050.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Instance weighting for domain adaptation in NLP",
"authors": [
{
"first": "Jing",
"middle": [],
"last": "Jiang",
"suffix": ""
},
{
"first": "Chengxiang",
"middle": [],
"last": "Zhai",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics",
"volume": "",
"issue": "",
"pages": "264--271",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jing Jiang and ChengXiang Zhai. 2007. Instance weighting for domain adaptation in NLP. In Pro- ceedings of the 45th Annual Meeting of the Asso- ciation of Computational Linguistics, pages 264- 271.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "ImageNet classification with deep convolutional neural networks",
"authors": [
{
"first": "Alex",
"middle": [],
"last": "Krizhevsky",
"suffix": ""
},
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "Geoffrey",
"middle": [
"E"
],
"last": "Hinton",
"suffix": ""
}
],
"year": 2012,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "1097--1105",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. ImageNet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097-1105.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Implicit discourse relation classification via multi-task neural networks",
"authors": [
{
"first": "Yang",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Sujian",
"middle": [],
"last": "Li",
"suffix": ""
},
{
"first": "Xiaodong",
"middle": [],
"last": "Zhang",
"suffix": ""
},
{
"first": "Zhifang",
"middle": [],
"last": "Sui",
"suffix": ""
}
],
"year": 2016,
"venue": "Proceedings of the 30th AAAI Conference on Artificial Intelligence",
"volume": "",
"issue": "",
"pages": "2750--2756",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yang Liu, Sujian Li, Xiaodong Zhang, and Zhifang Sui. 2016. Implicit discourse relation classifica- tion via multi-task neural networks. In Proceed- ings of the 30th AAAI Conference on Artificial In- telligence, pages 2750-2756.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Distributed representations of words and phrases and their compositionality",
"authors": [
{
"first": "Tomas",
"middle": [],
"last": "Mikolov",
"suffix": ""
},
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "Kai",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Greg",
"middle": [
"S"
],
"last": "Corrado",
"suffix": ""
},
{
"first": "Jeff",
"middle": [],
"last": "Dean",
"suffix": ""
}
],
"year": 2013,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "3111--3119",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed rep- resentations of words and phrases and their com- positionality. In Advances in Neural Information Processing Systems, pages 3111-3119.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "A comparative study on regularization strategies for embeddingbased neural networks",
"authors": [
{
"first": "Hao",
"middle": [],
"last": "Peng",
"suffix": ""
},
{
"first": "Lili",
"middle": [],
"last": "Mou",
"suffix": ""
},
{
"first": "Ge",
"middle": [],
"last": "Li",
"suffix": ""
},
{
"first": "Yunchuan",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Yangyang",
"middle": [],
"last": "Lu",
"suffix": ""
},
{
"first": "Zhi",
"middle": [],
"last": "Jin",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "2106--2111",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hao Peng, Lili Mou, Ge Li, Yunchuan Chen, Yangyang Lu, and Zhi Jin. 2015. A comparative study on regularization strategies for embedding- based neural networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 2106-2111.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Embedding semantic similarity in tree kernels for domain adaptation of relation extraction",
"authors": [
{
"first": "Barbara",
"middle": [],
"last": "Plank",
"suffix": ""
},
{
"first": "Alessandro",
"middle": [],
"last": "Moschitti",
"suffix": ""
}
],
"year": 2013,
"venue": "Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "1498--1507",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Barbara Plank and Alessandro Moschitti. 2013. Embedding semantic similarity in tree kernels for domain adaptation of relation extraction. In Pro- ceedings of the 51st Annual Meeting of the As- sociation for Computational Linguistics, pages 1498-1507.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Crosslanguage text classification using structural correspondence learning",
"authors": [
{
"first": "Peter",
"middle": [],
"last": "Prettenhofer",
"suffix": ""
},
{
"first": "Benno",
"middle": [],
"last": "Stein",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "1118--1127",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Peter Prettenhofer and Benno Stein. 2010. Cross- language text classification using structural corre- spondence learning. In Proceedings of the 48th Annual Meeting of the Association for Computa- tional Linguistics, pages 1118-1127.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Semi-supervised recursive autoencoders for predicting sentiment distributions",
"authors": [
{
"first": "Richard",
"middle": [],
"last": "Socher",
"suffix": ""
},
{
"first": "Jeffrey",
"middle": [],
"last": "Pennington",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Eric",
"suffix": ""
},
{
"first": "Andrew",
"middle": [
"Y"
],
"last": "Huang",
"suffix": ""
},
{
"first": "Christopher D",
"middle": [],
"last": "Ng",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Manning",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "151--161",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Richard Socher, Jeffrey Pennington, Eric H Huang, Andrew Y Ng, and Christopher D Manning. 2011. Semi-supervised recursive autoencoders for predicting sentiment distributions. In Pro- ceedings of the Conference on Empirical Methods in Natural Language Processing, pages 151-161.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Transfer learning for speech and language processing",
"authors": [
{
"first": "Dong",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Thomas Fang",
"middle": [],
"last": "Zheng",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the Asia-Pacific Signal and Information Processing Association Annual Summit and Conference",
"volume": "",
"issue": "",
"pages": "1225--1237",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dong Wang and Thomas Fang Zheng. 2015. Trans- fer learning for speech and language processing. In Proceedings of the Asia-Pacific Signal and In- formation Processing Association Annual Summit and Conference, pages 1225-1237.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "How transferable are features in deep neural networks?",
"authors": [
{
"first": "Jason",
"middle": [],
"last": "Yosinski",
"suffix": ""
},
{
"first": "Jeff",
"middle": [],
"last": "Clune",
"suffix": ""
},
{
"first": "Yoshua",
"middle": [],
"last": "Bengio",
"suffix": ""
},
{
"first": "Hod",
"middle": [],
"last": "Lipson",
"suffix": ""
}
],
"year": 2014,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "3320--3328",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. 2014. How transferable are features in deep neural networks? In Advances in Neural In- formation Processing Systems, pages 3320-3328.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Visualizing and understanding convolutional networks",
"authors": [
{
"first": "D",
"middle": [],
"last": "Matthew",
"suffix": ""
},
{
"first": "Rob",
"middle": [],
"last": "Zeiler",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Fergus",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of 13th European Conference on Computer Vision",
"volume": "",
"issue": "",
"pages": "818--833",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Matthew D Zeiler and Rob Fergus. 2014. Visual- izing and understanding convolutional networks. In Proceedings of 13th European Conference on Computer Vision, pages 818-833.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Self-adaptive hierarchical sentence model",
"authors": [
{
"first": "Han",
"middle": [],
"last": "Zhao",
"suffix": ""
},
{
"first": "Zhengdong",
"middle": [],
"last": "Lu",
"suffix": ""
},
{
"first": "Pascal",
"middle": [],
"last": "Poupart",
"suffix": ""
}
],
"year": 2015,
"venue": "International Joint Conference on Artificial Intelligence",
"volume": "",
"issue": "",
"pages": "4069--4076",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Han Zhao, Zhengdong Lu, and Pascal Poupart. 2015. Self-adaptive hierarchical sentence model. In International Joint Conference on Artificial In- telligence, pages 4069-4076.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "The models in our study. (a) Experiment I: RNNs with LSTM units for sentence classification. (b) Experiment II: CNN for sentence pair modeling."
},
"FIGREF1": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "Learning curves of different learning rates (denoted as \u03b1). (a) Experiment I: IMDB\u2192MR; (b) Experiment II: SNLI\u2192SICK."
},
"FIGREF3": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "(a) and (c): Learning curves of S. (b) and (d): Accuracies of T when parameters are transferred at a certain epoch during the training of S. Dotted lines refer to non-transfer, which can be equivalently viewed as transferring before training on S, i.e., epoch = 0. Note that the x-axis shares across different subplots."
},
"FIGREF4": {
"type_str": "figure",
"num": null,
"uris": null,
"text": "Results of MULT and MULT+INIT, where we share word embeddings and hidden layers. Dotted lines are the nontransfer setting; dashed lines are the INIT setting E H O , transferred at the peak performance of IMDB and SNLI."
},
"TABREF1": {
"html": null,
"type_str": "table",
"num": null,
"text": "Statistics and examples of the datasets.",
"content": "<table><tr><td>\u2022 Experiment II: Sentence-pair classification \u2212 SNLI. A large dataset for sentence entail-ment recognition. The classification objec-</td></tr><tr><td>tives are entailment, contradiction, and neutral. 8</td></tr><tr><td>\u2212 SICK. A small dataset with exactly the same classification objective as SNLI. 9</td></tr><tr><td>\u2212 MSRP. A (small) dataset for paraphrase de-tection. The objective is binary classification:</td></tr><tr><td>judging whether two sentences have the same meaning. 10</td></tr></table>"
},
"TABREF2": {
"html": null,
"type_str": "table",
"num": null,
"text": "Non-NN, Dong + ,2015) MR 75.1 \u00b1 0.6 77.7 (RAE, Socher + , 2013) QC 90.8 \u00b1 0.9 90.2 (RNN, Zhao + ,2015) Exp. II SNLI 76.3 77.6 (RNN, Bowman + ,2015) SICK 70.9 \u00b1 1.3 71.3 (RNN, Bowman + ,2015) MSRP 69.0 \u00b1 0.5 69.6 (Arc-I CNN, Hu + ,2014)",
"content": "<table><tr><td colspan=\"3\">Dataset Avg acc.\u00b1std.</td><td>Related model</td></tr><tr><td>Exp. I</td><td>IMDB</td><td>87.0</td><td>89.3 (</td></tr></table>"
},
"TABREF4": {
"html": null,
"type_str": "table",
"num": null,
"text": "Main results of neural transfer learning by INIT. We",
"content": "<table><tr><td colspan=\"5\">report test accuracies (%) in this table. E: embedding layer;</td></tr><tr><td colspan=\"5\">H: hidden layers; O: output layer. : Word embeddings are</td></tr><tr><td colspan=\"5\">pretrained by word2vec; : Parameters are randomly initial-</td></tr><tr><td>ized);</td><td colspan=\"4\">: Parameters are transferred but frozen; : Parame-</td></tr><tr><td colspan=\"2\">ters are transferred and fine-tuned. Notice that the E</td><td>H</td><td>O</td></tr><tr><td colspan=\"5\">and E H O settings are inapplicable to IMDB\u2192QC and</td></tr><tr><td colspan=\"5\">\u2642 SNLI\u2192MSRP, because the output targets do not share same</td></tr><tr><td colspan=\"2\">meanings and numbers of target classes.</td><td>\u2642</td><td>\u2642</td><td>\u2642</td></tr></table>"
}
}
}
}