ACL-OCL / Base_JSON /prefixN /json /N16 /N16-1027.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "N16-1027",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T14:35:59.138617Z"
},
"title": "Supertagging with LSTMs",
"authors": [
{
"first": "Ashish",
"middle": [],
"last": "Vaswani",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Southern California",
"location": {
"addrLine": "2 Kitt.ai"
}
},
"email": "vaswani@usc.edu"
},
{
"first": "Yonatan",
"middle": [],
"last": "Bisk",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Southern California",
"location": {
"addrLine": "2 Kitt.ai"
}
},
"email": "ybisk@isi.edu"
},
{
"first": "Kenji",
"middle": [],
"last": "Sagae",
"suffix": "",
"affiliation": {},
"email": "sagae@kitt.ai"
},
{
"first": "Ryan",
"middle": [],
"last": "Musa",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Illinois at Urbana-Champaign",
"location": {}
},
"email": "ramusa2@illinois.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In this paper we present new state-of-the-art performance on CCG supertagging and parsing. Our model outperforms existing approaches by an absolute gain of 1.5%. We analyze the performance of several neural models and demonstrate that while feed-forward architectures can compete with bidirectional LSTMs on POS tagging, models that encode the complete sentence are necessary for the long range syntactic information encoded in supertags.",
"pdf_parse": {
"paper_id": "N16-1027",
"_pdf_hash": "",
"abstract": [
{
"text": "In this paper we present new state-of-the-art performance on CCG supertagging and parsing. Our model outperforms existing approaches by an absolute gain of 1.5%. We analyze the performance of several neural models and demonstrate that while feed-forward architectures can compete with bidirectional LSTMs on POS tagging, models that encode the complete sentence are necessary for the long range syntactic information encoded in supertags.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Morphosyntactic labels for words are commonly used in a variety of NLP applications. For this reason, part-of-speech (POS) tagging and supertagging have drawn significant attention from the community. Combinatory Categorial Grammar is a lexicalized grammar formalism that is widely used for syntactic and semantic parsing. Supertagging (Clark, 2002; Bangalore and Joshi, 2010) assigns complex syntactic labels to words to enable fast and accurate parsing. The disambiguation of correctly labeling a word with one of over 1,200 CCG labels is difficult compared to choosing on of the 45 POS labels in the Penn Treebank (Marcus et al., 1993) . In addition to the large label space of CCG supertags, labeling a word correctly depends on knowledge of syntactic phenomena arbitrarily far in the sentence (Hockenmaier and Steedman, 2007) . This is because supertags encode highly specific syntactic information (e.g. types and locations of arguments) about a word's usage in a sentence.",
"cite_spans": [
{
"start": 336,
"end": 349,
"text": "(Clark, 2002;",
"ref_id": "BIBREF6"
},
{
"start": 350,
"end": 376,
"text": "Bangalore and Joshi, 2010)",
"ref_id": "BIBREF0"
},
{
"start": 608,
"end": 638,
"text": "Treebank (Marcus et al., 1993)",
"ref_id": null
},
{
"start": 798,
"end": 830,
"text": "(Hockenmaier and Steedman, 2007)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "In this paper, we show that Bidirectional Long Short-Term Memory recurrent neural networks (bi-LSTMs) (Graves, 2013; Zaremba et al., 2014) , which can use information from the entire sentence, are a natural and powerful architecture for CCG supertagging. In addition to the bi-LSTM, we create a simple yet novel model that outperforms the previous state-of-the-art RNN model that uses handcrafted features ) by 1.5%. Concurrent to this work (Lewis et al., 2016) introduced a different training methodology for bi-LSTM for supertagging. We provide a detailed analysis of the quality of various LSTM architectures, forward, backward, and bi-directional, shedding light over the ability of the bi-LSTM to exploit rich sentential context necessary for performing supertagging. We also show that a baseline feed-forward neural network (NN) architecture significantly outperforms previous feed-forward NN baselines, with slightly fewer features, achieving better accuracy than the RNN model from .",
"cite_spans": [
{
"start": 102,
"end": 116,
"text": "(Graves, 2013;",
"ref_id": "BIBREF7"
},
{
"start": 117,
"end": 138,
"text": "Zaremba et al., 2014)",
"ref_id": "BIBREF19"
},
{
"start": 441,
"end": 461,
"text": "(Lewis et al., 2016)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Recently, bi-LSTMs have achieved high accuracies in a simpler sequence labeling task: partof-speech tagging (Wang et al., 2015; Ling et al., 2015) on the Penn treebank, with small improvements over local models. However, we achieve strong accuracies compared to (Wang et al., 2015) using feed-forward neural network model trained on local context, showing that this task does not require bi-LSTMs. Our strong feed-forward NN baselines show the power of feed-forward NNs for some tasks.",
"cite_spans": [
{
"start": 108,
"end": 127,
"text": "(Wang et al., 2015;",
"ref_id": "BIBREF11"
},
{
"start": 128,
"end": 146,
"text": "Ling et al., 2015)",
"ref_id": "BIBREF11"
},
{
"start": 262,
"end": 281,
"text": "(Wang et al., 2015)",
"ref_id": "BIBREF11"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our main contributions are the introduction of a new bi-LSTM model for CCG supertagging that achieves state-of-the-art, on both CCG supertagging and parsing, and a detailed analysis of our results, including a comparison of bi-LSTMs and simpler feed forward NN models for supertagging and POS tagging, which suggests that the added complexity of bi-LSTMs may not be necessary for POS tagging, where local contexts suffice to a much greater extent than in supertagging.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We use feed-forward neural network models and bidirectional LSTM (bi-LSTM) based models in this work.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Models And Training",
"sec_num": "2"
},
{
"text": "For both POS tagging and our baseline supertagging model, we use feed-forward neural networks with two hidden layers of rectified linear units (Nair and Hinton, 2010) . For supertagging, we use a slightly smaller set than Lewis and Steedman (2014a) , using a left and right 3-word window with suffix and capitalization features for the center word. However, unlike them, we train on the full set of supertag categories observed during training.",
"cite_spans": [
{
"start": 143,
"end": 166,
"text": "(Nair and Hinton, 2010)",
"ref_id": "BIBREF13"
},
{
"start": 222,
"end": 248,
"text": "Lewis and Steedman (2014a)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Feed-Forward",
"sec_num": "2.1"
},
{
"text": "In POS tagging, when tagging word w i , we consider only features from a window of five words, with w i at the center. For each w j with i \u2212 2 \u2264 j \u2264 i + 2, we add w j lowercased and a string that encodes the basic \"word shape\" of w j . This is computed by replacing all sequences of uppercase letters with A, all sequences of lowercase letters with a, all sequences of digits with 9, and all sequences of other characters with * . Finally, we add two and three letter suffixes and two letter prefix for w i only.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Feed-Forward",
"sec_num": "2.1"
},
{
"text": "We experiment with two kinds of bi-LSTM models. We train a basic bi-LSTM where the forward and backward LSTMs take input words w i and produce hidden state \u2212 \u2192 h i and \u2190 \u2212 h i . For each position, we produceh i , wher\u1ebd",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "h i = \u03c3(W\u2190 \u2212 h \u2190 \u2212 h T i + W\u2212 \u2192 h \u2212 \u2192 h T i ),",
"eq_num": "(1)"
}
],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "where \u03c3(x) = max(0, x) is a rectifier nonlinearity, and where W\u2190 \u2212 h and W\u2212 \u2192 h are parameters to be learned. The unnormalized likelihood of an output supertag is computed using supertag embeddings",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "D t i and biases b t i as p(t i |h i ) = D t ih T i + b t i .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "The final softmax layer computes normalized supertag probabilities.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "Although bidirectional LSTMs can capture long distance interactions between words, each output label is predicted independently. To explicitly model supertag interactions, our next model combines two models, the bi-LSTM and a LSTM language model (LM) over the supertags (Figure 1 ",
"cite_spans": [],
"ref_spans": [
{
"start": 270,
"end": 279,
"text": "(Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "). At position \u2026 \u2026 \u2026 \u2026 \u2026 Backward LSTM Forward LSTM Combiner Nodes LSTM LM Combiner Nodes Output Words h LM i h LM i+1 h LM i+2 \u2190 \u2212 hi \u2190 \u2212 \u2212 hi+1 \u2190 \u2212 \u2212 hi+2 \u2212 \u2212 \u2192 hi+2 \u2212 \u2212 \u2192 hi+1 \u2212 \u2192 hi t i t i+1 ti+2 t i t i+1 ti+2 \u2026 eat sushi with tuna Figure 1:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "We add a language model between supertags.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "i, the LM accepts an input supertag t i\u22121 producing hidden state h LM i , and a second combiner layer, parametrized by matrices W LM and Wh transforms h i and h LM i to h i similar to the combiner forh i (Equation 1). Output supertag probabilities are computed just as before, replacing replacingh i with h i . We refer to this model as bi-LSTM-LM. For all our LSTM models, we only use words as input features.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LSTM models",
"sec_num": "2.2"
},
{
"text": "We train our models to maximize the log-likelihood of the data with minibatch gradient ascent. Gradients of the models are computed with backpropagation (Chauvin and Rumelhart, 1995) . Since gold supertags are available during training time and not while decoding, a bi-LSTM-LM trained on gold supertags might not recover from errors caused by using incorrectly predicted supertags. This results in the bi-LSTM-LM slightly underperforming the bi-LSTM (we refer to training with gold supertags as g-train in Table 1 ). To bridge this gap between training and testing we also experiment with a sampling training regime in addition to training.",
"cite_spans": [
{
"start": 153,
"end": 182,
"text": "(Chauvin and Rumelhart, 1995)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [
{
"start": 507,
"end": 514,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Training",
"sec_num": "2.3"
},
{
"text": "Scheduled sampling: Following (Bengio et al., 2015; Ranzato et al., 2015) , for each output token, with some probability p, we use the most likely predicted supertag (arg max t i P (t i | h i )) from the model in position i\u22121 as input to the supertag LSTM LM in position i and use the gold supertag with probability 1 \u2212 p. We denote this training as sstrain-1. We also experiment with using the 5-best previous predicted supertags from the output distribution at position i \u2212 1 and feed them to the LM as input in position i as a bit vector. Additionally, we g-train ss-train-1 ss-train-5 1 Figure 2 : Scheduled sampling improves the perplexity of the gold sequence under predicted tags. We see that the perplexity of the gold supertag sequence when using predicted tags for the LM is lower for ss-train-1 and ss-train-5 than with g-train.",
"cite_spans": [
{
"start": 30,
"end": 51,
"text": "(Bengio et al., 2015;",
"ref_id": "BIBREF1"
},
{
"start": 52,
"end": 73,
"text": "Ranzato et al., 2015)",
"ref_id": "BIBREF14"
}
],
"ref_spans": [
{
"start": 591,
"end": 599,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Training",
"sec_num": "2.3"
},
{
"text": "use their probabilities (re-normalized over the 5-best tags) and scale the input supertag embeddings with their re-normalized probability during look-up. We refer to this setting as ss-train-5. In this work, we use an inverse sigmoid schedule to compute p,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "2.3"
},
{
"text": "p = k k + e s k ,",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "2.3"
},
{
"text": "where s is the epoch number and k is a hyperparameter that is tuned. 1 In Figure 2 , we see that for the development set training with scheduled sampling improves the perplexity of the gold supertag sequence when using predicted supertags, indicating better recovery from conditioning on erroneous supertags. For both ss-train and g-train, we use gold supertags for the output layer and train the model to maximize the log-likelihood of the data. 2",
"cite_spans": [],
"ref_spans": [
{
"start": 74,
"end": 82,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Training",
"sec_num": "2.3"
},
{
"text": "Our feed-forward models use 2048 rectifier units in the first hidden layer, 50 and 128 rectifier units in the second hidden layer for POS tagging and Supertagging respectively, and 64 dim. input embeddings.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Architectures",
"sec_num": "2.4"
},
{
"text": "Our LSTM based models use 512 hidden states. We pre-train our word embeddings with a 7gram feed-forward neural language model using the NPLM toolkit 3 on a concatenation of the BLLIP corpus (Charniak et al., 2000) and WSJ sections 02-21 of the Penn Treebank. ",
"cite_spans": [
{
"start": 190,
"end": 213,
"text": "(Charniak et al., 2000)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Architectures",
"sec_num": "2.4"
},
{
"text": "We perform greedy decoding. For each position i, we select the most probable supertag from the output distribution. For the bi-LSTM-LM models trained with g-train and ss-train-1, we feed the most likely supertag from the output distribution as LM input in the next position. We decode with beam search (size 12) for bi-LSTM-LMs trained with g-train and ss-train-1. For the bi-LSTM-LMs trained with ss-train-5, we perform greedy decoding similar to training, feeding the k-best supertags from the output supertag distribution in position i \u2212 1 as input to the LM in position i, along with the renormalized probabilities. We don't perform beam decoding for ss-train-5, as the previous k-best inputs already capture different paths through the network. 4",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Decoding",
"sec_num": "2.5"
},
{
"text": "For supertagging, experiments were run with the standard splits of CCGbank. Unlike previous work no features were extracted for the LSTM models and rare categories were not thresholded. Words were lowercased and digits replaced with @.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Data",
"sec_num": "3"
},
{
"text": "CCGbank's training section contains 1,284 lexical categories (394 in Dev). The distribution of categories has a long tail, with only a third of those cate- gories having a frequency count \u2265 10 (the threshold used by existing literature). Following (Lewis and Steedman, 2014b) , we allow the model to predict all categories for a word, not just those with which the word was observed to co-occur in the training data. Accuracies on these unseen (word, cat) pairs are presented in the third column of Table 1 . Table 3 presents our Feed-Forward POS tagging results. We achieve 97.28% on the development set and 97.4% on test. Although slightly below state-ofthe-art, we approach existing work with bi-LSTMs, and our models are much simpler and faster to train. 5 Table 1 shows a steady increase in performance as the model is provided additional context. The forward and backward models are presented with information that may be arbitrarily far away in the sentence, but only in a specific direction. This yields weaker results than the Feed Forward model which can see in both directions within a small window. The real gains are achieved by the Bidirectional LSTM which incorporates knowledge from the entire sentence. Our addition of a language model and changes to training, further improve the perfor- 5 We use train, dev, and test splits of WSJ sections 00-18, 19-21, and 22-24, for POS tagging. mance. Our final model (bi-LSTM-LM+ss-train-1 model with beam decoding) has a test accuracy of 94.5%, 1.5% above state-of-the-art.",
"cite_spans": [
{
"start": 248,
"end": 275,
"text": "(Lewis and Steedman, 2014b)",
"ref_id": "BIBREF9"
},
{
"start": 1306,
"end": 1307,
"text": "5",
"ref_id": null
}
],
"ref_spans": [
{
"start": 499,
"end": 506,
"text": "Table 1",
"ref_id": null
},
{
"start": 509,
"end": 516,
"text": "Table 3",
"ref_id": "TABREF4"
},
{
"start": 761,
"end": 768,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Data",
"sec_num": "3"
},
{
"text": "Our primary goal in this paper was to demonstrate how a bi-LSTM captures new and different information from uni-directional or feed-forward approaches. This advantage also translates to gains in parsing. Table 4 presents new state-of-the-art parsing results for both and our bi-LSTM-LM +ss-train-1. These results were attained using our part-of-speech tags (Table 3 ) and the Java implementation of the C&C parser (Clark and Curran, 2007) 6 .",
"cite_spans": [
{
"start": 414,
"end": 438,
"text": "(Clark and Curran, 2007)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [
{
"start": 204,
"end": 211,
"text": "Table 4",
"ref_id": "TABREF6"
},
{
"start": 357,
"end": 365,
"text": "(Table 3",
"ref_id": "TABREF4"
}
],
"eq_spans": [],
"section": "Parsing",
"sec_num": "4.1"
},
{
"text": "Our analysis indicates that the information following a word is more informative than what preceded it. Table 2 compares how well our models recover common and syntactically interesting supertags. In particular, the Forward and Backward models, motivate the need for a Bi-directional approach. Table 5 : \"Neighbor\" categories as determined by embedding-based vector similarity for each class of model. As expected for this category, the Backward model captures the argument preference while the Forward model correctly predicts the result.",
"cite_spans": [],
"ref_spans": [
{
"start": 104,
"end": 111,
"text": "Table 2",
"ref_id": "TABREF3"
},
{
"start": 294,
"end": 301,
"text": "Table 5",
"ref_id": "TABREF8"
}
],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "The first two rows show prepositional phrase attachment decisions (noun and verb attaching categories are in rows one and two, respectively). Here the forward model outperforms the backward model, presumably because knowing the word to be modified and the preposition, is more important than observing the object of the prepositional phrase (the information available to the backward model).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "Conversely, the backward model outperforms the forward model in most of the remaining categories. (Di-)transitive verbs (lines 4 & 5) require knowledge of future arguments in the sentence (e.g. separated by a relative clause). Because English has strict SVO word-order, the presence of a subject is more predictable than the presence of an (in-)direct object. It is therefore not surprising that the backward model is often comparable to the Feed Forward model.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "If the information missing from either the forward or backward models were local, the bidirectional model should perform the same as the Feed-Forward model, instead it surpasses it, often by a large margin. This implies there is long range information necessary for choosing a supertag.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "Embeddings In addition, we can visualize the information captured by our models by investigating a category's nearest neighbors based on the learned embeddings. We see see that the forward model learns internal structure with the query category, but the list of arguments is nearly random. In contrast, the backward model clusters categories primarily based on the final argument, perhaps sharing similarities in the subject argument only because of the predictable SVO nature of English text. However, due to its lack of forward context the model incorrectly asso-ciates categories with less-common first arguments (e.g. S[qem]). Finally, the bidirectional embeddings appear to cleanly capture the strengths of both the forward and backward models.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "Consistency and Internal Structure Because supertags are highly structured their co-occurence in a sentence must be permitted by the combinators of CCG. Without encoding this explicitly, the language model dramatically increases the percent of predicted sequences that result in a valid parse by up to 15% (last column of Table 2 ).",
"cite_spans": [],
"ref_spans": [
{
"start": 322,
"end": 329,
"text": "Table 2",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "Sparsity One consideration of our approach is that we do not threshold rare categories or use any tag dictionaries; our models are presented with the full space of CCG categories, despite the long tail. This did not did not hurt performance and the models learned to successfully use several categories which were outside the set of traditionally-thresholded frequent categories. Additionally, the total number of categories used correctly at least once by the bidirectional models was substantially higher than the other models (\u223c270 vs. \u223c220 of 394), though the large number of unused categories (\u2265120) indicates that there is still substantial room for improvement.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error Analysis",
"sec_num": "4.2"
},
{
"text": "Because bi-LSTMs with a language model encode an entire sentence at decision time, we demonstrated large gains in supertagging and parsing. Future work will investigate improving performance on rare categories.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusions and Future Work",
"sec_num": "5"
},
{
"text": "The reader should refer to(Bengio et al., 2015) for details.2 We use dropout for all our feed-forward(Srivastava, 2013) and bi-LSTM based models(Zaremba et al., 2014). We carry out a grid search over dropout probabilities and sampling schedules. We train the LSTMs for 25 epochs and the feed-forward models for 30 epochs, tuning on the development data.3 http://nlg.isi.edu/software/nplm/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Code and supertags for our models can be downloaded here: https://bitbucket.org/ashish_vaswani/ lstm_supertagger",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Results are presented on the standard development and test splits (Section 00 and 23), and with a beam threshold of 10 \u22126 . For a fair comparison to prior work we report results without the skimmer, so no partial credit is given to parse failures. The skimmer boosts performance to 87.91/88.39 for Dev and Test.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "This work was supported by the U.S. DARPA LORELEI Program No. HR0011-15-C-0115. We would like to thank Wenduan Xu for his help.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgements",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Supertagging: Using Complex Lexical Descriptions in Natural Language Processing",
"authors": [
{
"first": "Srinivas",
"middle": [],
"last": "Bangalore",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Aravind",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Joshi",
"suffix": ""
}
],
"year": 2010,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Srinivas Bangalore and Aravind K. Joshi. 2010. Su- pertagging: Using Complex Lexical Descriptions in Natural Language Processing. The MIT Press.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Scheduled sampling for sequence prediction with recurrent neural networks",
"authors": [
{
"first": "Samy",
"middle": [],
"last": "Bengio",
"suffix": ""
},
{
"first": "Oriol",
"middle": [],
"last": "Vinyals",
"suffix": ""
},
{
"first": "Navdeep",
"middle": [],
"last": "Jaitly",
"suffix": ""
},
{
"first": "Noam",
"middle": [],
"last": "Shazeer",
"suffix": ""
}
],
"year": 2015,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "1171--1179",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Samy Bengio, Oriol Vinyals, Navdeep Jaitly, and Noam Shazeer. 2015. Scheduled sampling for sequence pre- diction with recurrent neural networks. In Advances in Neural Information Processing Systems, pages 1171- 1179.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Bllip 1987-89 wsj corpus release 1. Linguistic Data Consortium",
"authors": [
{
"first": "Eugene",
"middle": [],
"last": "Charniak",
"suffix": ""
},
{
"first": "Don",
"middle": [],
"last": "Blaheta",
"suffix": ""
},
{
"first": "Niyu",
"middle": [],
"last": "Ge",
"suffix": ""
},
{
"first": "Keith",
"middle": [],
"last": "Hall",
"suffix": ""
},
{
"first": "John",
"middle": [],
"last": "Hale",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Johnson",
"suffix": ""
}
],
"year": 2000,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Eugene Charniak, Don Blaheta, Niyu Ge, Keith Hall, John Hale, and Mark Johnson. 2000. Bllip 1987- 89 wsj corpus release 1. Linguistic Data Consortium, Philadelphia, 36.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Backpropagation: theory, architectures, and applications",
"authors": [
{
"first": "Yves",
"middle": [],
"last": "Chauvin",
"suffix": ""
},
{
"first": "David",
"middle": [
"E"
],
"last": "Rumelhart",
"suffix": ""
}
],
"year": 1995,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yves Chauvin and David E Rumelhart. 1995. Backprop- agation: theory, architectures, and applications. Psy- chology Press.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Wide-Coverage Efficient Statistical Parsing with CCG and Log-Linear Models",
"authors": [
{
"first": "Stephen",
"middle": [],
"last": "Clark",
"suffix": ""
},
{
"first": "James",
"middle": [],
"last": "Curran",
"suffix": ""
}
],
"year": 2007,
"venue": "Computational Linguistics",
"volume": "33",
"issue": "4",
"pages": "493--552",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stephen Clark and James Curran. 2007. Wide-Coverage Efficient Statistical Parsing with CCG and Log-Linear Models. Computational Linguistics, 33(4):493-552.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "The Java Version of the C&C Parser: Version 0.95",
"authors": [
{
"first": "Stephen",
"middle": [],
"last": "Clark",
"suffix": ""
},
{
"first": "Darren",
"middle": [],
"last": "Foong",
"suffix": ""
},
{
"first": "Luana",
"middle": [],
"last": "Bulat",
"suffix": ""
},
{
"first": "Wenduan",
"middle": [],
"last": "Xu",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stephen Clark, Darren Foong, Luana Bulat, and Wenduan Xu. 2015. The Java Version of the C&C Parser: Ver- sion 0.95. Technical report, University of Cambridge Computer Laboratory, August.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Supertagging for combinatory categorial grammar",
"authors": [
{
"first": "Stephen",
"middle": [],
"last": "Clark",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the 6th International Workshop on Tree Adjoining Grammars and Related Formalisms (TAG+6)",
"volume": "",
"issue": "",
"pages": "19--24",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stephen Clark. 2002. Supertagging for combinatory cat- egorial grammar. In Proceedings of the 6th Interna- tional Workshop on Tree Adjoining Grammars and Re- lated Formalisms (TAG+6), pages 19-24.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Generating sequences with recurrent neural networks",
"authors": [
{
"first": "A",
"middle": [],
"last": "Graves",
"suffix": ""
}
],
"year": 2013,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1308.0850"
]
},
"num": null,
"urls": [],
"raw_text": "A. Graves. 2013. Generating sequences with recurrent neural networks. arXiv preprint arXiv:1308.0850.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "CCGbank: A Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank",
"authors": [
{
"first": "Julia",
"middle": [],
"last": "Hockenmaier",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Steedman",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "33",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Julia Hockenmaier and Mark Steedman. 2007. CCG- bank: A Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank. Com- putational Linguistics, 33:355-396, September. Mike Lewis and Mark Steedman. 2014a. A* ccg pars- ing with a supertag-factored model. In Proceedings of the Conference on Empirical Methods in Natural Lan- guage Processing (EMNLP-2014).",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Improved ccg parsing with semi-supervised supertagging",
"authors": [
{
"first": "Mike",
"middle": [],
"last": "Lewis",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Steedman",
"suffix": ""
}
],
"year": 2014,
"venue": "Transactions of the Association for Computational Linguistics",
"volume": "2",
"issue": "",
"pages": "327--338",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mike Lewis and Mark Steedman. 2014b. Improved ccg parsing with semi-supervised supertagging. Transac- tions of the Association for Computational Linguistics, 2:327-338.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "LSTM CCG Parsing",
"authors": [
{
"first": "Mike",
"middle": [],
"last": "Lewis",
"suffix": ""
},
{
"first": "Kenton",
"middle": [],
"last": "Lee",
"suffix": ""
},
{
"first": "Luke",
"middle": [],
"last": "Zettlemoyer",
"suffix": ""
}
],
"year": 2016,
"venue": "Proceedings of the 15th Annual Conference of the North American Chapter",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mike Lewis, Kenton Lee, and Luke Zettlemoyer. 2016. LSTM CCG Parsing. In Proceedings of the 15th An- nual Conference of the North American Chapter of the Association for Computational Linguistics.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Finding function in form: Compositional character models for open vocabulary word representation",
"authors": [
{
"first": "Wang",
"middle": [],
"last": "Ling",
"suffix": ""
},
{
"first": "Tiago",
"middle": [],
"last": "Lu\u00eds",
"suffix": ""
},
{
"first": "Lu\u00eds",
"middle": [],
"last": "Marujo",
"suffix": ""
},
{
"first": "Ram\u00f3n",
"middle": [],
"last": "Fernandez Astudillo",
"suffix": ""
},
{
"first": "Silvio",
"middle": [],
"last": "Amir",
"suffix": ""
},
{
"first": "Chris",
"middle": [],
"last": "Dyer",
"suffix": ""
},
{
"first": "Alan",
"middle": [
"W"
],
"last": "Black",
"suffix": ""
},
{
"first": "Isabel",
"middle": [],
"last": "Trancoso",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1508.02096"
]
},
"num": null,
"urls": [],
"raw_text": "Wang Ling, Tiago Lu\u00eds, Lu\u00eds Marujo, Ram\u00f3n Fernan- dez Astudillo, Silvio Amir, Chris Dyer, Alan W Black, and Isabel Trancoso. 2015. Finding func- tion in form: Compositional character models for open vocabulary word representation. arXiv preprint arXiv:1508.02096.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Building a Large Annotated Corpus of English: The Penn Treebank",
"authors": [
{
"first": "P",
"middle": [],
"last": "Mitchell",
"suffix": ""
},
{
"first": "Beatrice",
"middle": [],
"last": "Marcus",
"suffix": ""
},
{
"first": "Mary",
"middle": [
"Ann"
],
"last": "Santorini",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Marcinkiewicz",
"suffix": ""
}
],
"year": 1993,
"venue": "Computational Linguistics",
"volume": "19",
"issue": "",
"pages": "313--330",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mitchell P Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1993. Building a Large Annotated Corpus of English: The Penn Treebank. Computa- tional Linguistics, 19:313-330, June.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Rectified linear units improve restricted Boltzmann machines",
"authors": [
{
"first": "Vinod",
"middle": [],
"last": "Nair",
"suffix": ""
},
{
"first": "Geoffrey",
"middle": [
"E"
],
"last": "Hinton",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of ICML",
"volume": "",
"issue": "",
"pages": "807--814",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vinod Nair and Geoffrey E. Hinton. 2010. Rectified lin- ear units improve restricted Boltzmann machines. In Proceedings of ICML, pages 807-814.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Sequence Level Training with Recurrent Neural Networks",
"authors": [
{
"first": "Aurelio",
"middle": [],
"last": "Marc",
"suffix": ""
},
{
"first": "Sumit",
"middle": [],
"last": "Ranzato",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Chopra",
"suffix": ""
},
{
"first": "Wojciech",
"middle": [],
"last": "Auli",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Zaremba",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1511.06732"
]
},
"num": null,
"urls": [],
"raw_text": "Marc'Aurelio Ranzato, Sumit Chopra, Michael Auli, and Wojciech Zaremba. 2015. Sequence Level Train- ing with Recurrent Neural Networks. arXiv preprint arXiv:1511.06732.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Semisupervised condensed nearest neighbor for part-of-speech tagging",
"authors": [
{
"first": "Anders",
"middle": [],
"last": "S\u00f8gaard",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers",
"volume": "2",
"issue": "",
"pages": "48--52",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Anders S\u00f8gaard. 2011. Semisupervised condensed near- est neighbor for part-of-speech tagging. In Proceed- ings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Tech- nologies: short papers-Volume 2, pages 48-52. Asso- ciation for Computational Linguistics.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "Improving neural networks with dropout",
"authors": [
{
"first": "Nitish",
"middle": [],
"last": "Srivastava",
"suffix": ""
}
],
"year": 2013,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Nitish Srivastava. 2013. Improving neural networks with dropout. Ph.D. thesis, University of Toronto.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Part-of-speech tagging with bidirectional long short-term memory recurrent neural network",
"authors": [
{
"first": "Peilu",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Yao",
"middle": [],
"last": "Qian",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Frank",
"suffix": ""
},
{
"first": "Lei",
"middle": [],
"last": "Soong",
"suffix": ""
},
{
"first": "Hai",
"middle": [],
"last": "He",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Zhao",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1510.06168"
]
},
"num": null,
"urls": [],
"raw_text": "Peilu Wang, Yao Qian, Frank K Soong, Lei He, and Hai Zhao. 2015. Part-of-speech tagging with bidirec- tional long short-term memory recurrent neural net- work. arXiv preprint arXiv:1510.06168.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Ccg supertagging with a recurrent neural network",
"authors": [
{
"first": "Wenduan",
"middle": [],
"last": "Xu",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Auli",
"suffix": ""
},
{
"first": "Stephen",
"middle": [],
"last": "Clark",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "2",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Wenduan Xu, Michael Auli, and Stephen Clark. 2015. Ccg supertagging with a recurrent neural network. Volume 2: Short Papers, page 250.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Recurrent neural network regularization",
"authors": [
{
"first": "W",
"middle": [],
"last": "Zaremba",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Sutskever",
"suffix": ""
},
{
"first": "O",
"middle": [],
"last": "Vinyals",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1409.2329"
]
},
"num": null,
"urls": [],
"raw_text": "W. Zaremba, I. Sutskever, and O. Vinyals. 2014. Re- current neural network regularization. arXiv preprint arXiv:1409.2329.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"num": null,
"type_str": "figure",
"text": "Accuracies on the development section. The language model provides a boost in performance, and large gains on the parseability of the sequence (%P). The numbers for bi-LSTM-LM + ss-train-1 and + g-train are with beam decoding. All others use greedy decoding. Interestingly, greedy decoding with ss-train-5 works as well as beam decoding with ss-train-1.",
"uris": null
},
"TABREF3": {
"type_str": "table",
"text": "Prediction accuracy for our models on several common and difficult supertags.",
"content": "<table><tr><td/><td>Architecture</td><td>Test Acc</td></tr><tr><td>Ling et al. (2015)</td><td>Bi-LSTM</td><td>97.36</td></tr><tr><td colspan=\"2\">Wang et al. (2015) Bi-LSTM</td><td>97.78</td></tr><tr><td>S\u00f8gaard (2011)</td><td>SCNN</td><td>97.50</td></tr><tr><td>This work</td><td>Feed-Forward</td><td>97.40</td></tr></table>",
"num": null,
"html": null
},
"TABREF4": {
"type_str": "table",
"text": "",
"content": "<table/>",
"num": null,
"html": null
},
"TABREF6": {
"type_str": "table",
"text": "Parsing at 100% coverage with our new Feed-Forward POS tagger and the Java implementation of C&C. We show both the published and improved results for Wenduan et al.",
"content": "<table/>",
"num": null,
"html": null
},
"TABREF8": {
"type_str": "table",
"text": "shows nearest neighbor categories for (S[dcl]\\NP)/(S[adj]\\NP) under the Forward, Backward, and Bidirectional models.",
"content": "<table/>",
"num": null,
"html": null
}
}
}
}