| { |
| "paper_id": "P19-1042", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T08:28:07.580514Z" |
| }, |
| "title": "Cross-Sentence Grammatical Error Correction", |
| "authors": [ |
| { |
| "first": "Shamil", |
| "middle": [], |
| "last": "Chollampatt", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National University of Singapore", |
| "location": {} |
| }, |
| "email": "shamil@u.nus.edu" |
| }, |
| { |
| "first": "Weiqi", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National University of Singapore", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Hwee", |
| "middle": [ |
| "Tou" |
| ], |
| "last": "Ng", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National University of Singapore", |
| "location": {} |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Automatic grammatical error correction (GEC) research has made remarkable progress in the past decade. However, all existing approaches to GEC correct errors by considering a single sentence alone and ignoring crucial cross-sentence context. Some errors can only be corrected reliably using cross-sentence context and models can also benefit from the additional contextual information in correcting other errors. In this paper, we address this serious limitation of existing approaches and improve strong neural encoder-decoder models by appropriately modeling wider contexts. We employ an auxiliary encoder that encodes previous sentences and incorporate the encoding in the decoder via attention and gating mechanisms. Our approach results in statistically significant improvements in overall GEC performance over strong baselines across multiple test sets. Analysis of our cross-sentence GEC model on a synthetic dataset shows high performance in verb tense corrections that require cross-sentence context.", |
| "pdf_parse": { |
| "paper_id": "P19-1042", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Automatic grammatical error correction (GEC) research has made remarkable progress in the past decade. However, all existing approaches to GEC correct errors by considering a single sentence alone and ignoring crucial cross-sentence context. Some errors can only be corrected reliably using cross-sentence context and models can also benefit from the additional contextual information in correcting other errors. In this paper, we address this serious limitation of existing approaches and improve strong neural encoder-decoder models by appropriately modeling wider contexts. We employ an auxiliary encoder that encodes previous sentences and incorporate the encoding in the decoder via attention and gating mechanisms. Our approach results in statistically significant improvements in overall GEC performance over strong baselines across multiple test sets. Analysis of our cross-sentence GEC model on a synthetic dataset shows high performance in verb tense corrections that require cross-sentence context.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Grammatical error correction (GEC) is the task of correcting errors in input text and producing well-formed output text. GEC models are essential components of writing assistance and proof-reading tools that help both native and non-native speakers. Several adaptations of sophisticated sequence-to-sequence learning models with specialized techniques have been shown to achieve impressive performance (Ge et al., 2018; Lichtarge et al., 2018; Chollampatt and Ng, 2018b; .", |
| "cite_spans": [ |
| { |
| "start": 402, |
| "end": 419, |
| "text": "(Ge et al., 2018;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 420, |
| "end": 443, |
| "text": "Lichtarge et al., 2018;", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 444, |
| "end": 470, |
| "text": "Chollampatt and Ng, 2018b;", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "All prior approaches to GEC consider one sentence at a time and ignore useful contextual information from the document in which it appears, unlike a human proofreader. Cross-sentence context", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "As a result, they are not convenient enough.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "OUT OF CONTEXT:", |
| "sec_num": null |
| }, |
| { |
| "text": "Electric cars have a very obvious shortage in their technique design. The electric cars invented in 1990 did not have a powerful battery. Due to the limitation of its weigh, size and the battery technology, the battery used in the electric cars at that time was limited to a range of 100 miles (Rogers,2003) . As a result, they are were not convenient enough. Instead, Hydrogen fuel cell was brought up to substitute the electric battery. is essential to identify and correct certain errors, mostly involving tense choice, use of the definite article 'the', and use of connectives. The example in Figure 1 , from the NUS Corpus of Learner English or NUCLE (Dahlmeier et al., 2013) , shows a learner-written sentence that seems correct in isolation, but actually involves a verb tense error in context. Sentence-level GEC systems fail to reliably correct such errors. Moreover, models may also benefit from the additional context by being able to disambiguate error corrections better.", |
| "cite_spans": [ |
| { |
| "start": 294, |
| "end": 307, |
| "text": "(Rogers,2003)", |
| "ref_id": null |
| }, |
| { |
| "start": 656, |
| "end": 680, |
| "text": "(Dahlmeier et al., 2013)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 597, |
| "end": 605, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "IN CONTEXT:", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper, we present the first approach to cross-sentence GEC 1 . We build on a state-of-theart convolutional neural encoder-decoder model and incorporate cross-sentence context from previous sentences using an auxiliary encoder. The decoder attends to the representations generated by the auxiliary encoder via separate attention mechanisms and incorporates them via gating that controls the information that goes into the de-coder. Auxiliary encoders have also been used in other sequence generation tasks such as automatic post editing (Libovick\u00fd and Helcl, 2017) , multilingual machine translation (Firat et al., 2016) , and document-level neural machine translation (Jean et al., 2017; Wang et al., 2017) .", |
| "cite_spans": [ |
| { |
| "start": 544, |
| "end": 571, |
| "text": "(Libovick\u00fd and Helcl, 2017)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 607, |
| "end": 627, |
| "text": "(Firat et al., 2016)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 676, |
| "end": 695, |
| "text": "(Jean et al., 2017;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 696, |
| "end": 714, |
| "text": "Wang et al., 2017)", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IN CONTEXT:", |
| "sec_num": null |
| }, |
| { |
| "text": "Our cross-sentence GEC model shows statistically significant improvements over a competitive sentence-level baseline across multiple test sets. We further improve the baseline and crosssentence model by ensembling multiple models and rescoring. We also incorporate probabilities computed by BERT (Devlin et al., 2018) as a feature. The gains in performance with the crosssentence component over the sentence-level baseline still remain even with the improvements. Our final cross-sentence model also outperforms stateof-the-art models trained on the same datasets on the CoNLL-2014 test set. They show notable improvements in correcting determiner and verb tense errors. Our analysis demonstrates that our cross-sentence model is able to accurately correct many cross-sentence verb tense errors when evaluated on a synthetic test set.", |
| "cite_spans": [ |
| { |
| "start": 296, |
| "end": 317, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "IN CONTEXT:", |
| "sec_num": null |
| }, |
| { |
| "text": "Consider a source document S = S 1 , . . . S N made up of N source sentences, where S k represents the kth source sentence. S k comprises of |S k | tokens s k,1 . . . s k,|S k | . In a standard sentence-level encoder-decoder model, the probability of a corrected target hypothesis T k = t k,1 . . . t k,|T k | , given the source sentence S k and the set of model parameters \u0398, is computed by", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Cross-Sentence Context", |
| "sec_num": "2" |
| }, |
| { |
| "text": "P (T k |S k , \u0398) = |T k | i=1 P (t k,i |T k,<i , S k , \u0398) (1)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Cross-Sentence Context", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where T k,<i denotes previous target words t k,1 , . . . , t k,i\u22121 . The above model assumes that the correction of S k is independent of other sentences in the source document S. This assumption may not always hold since cross-sentence context is required or helpful for correcting many errors. In our proposed cross-sentence model, the above conditional probability also depends on other sentences in the source document, denoted by S doc . Equation 1 is rewritten as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Cross-Sentence Context", |
| "sec_num": "2" |
| }, |
| { |
| "text": "P (T k |S, \u0398) = |T k | i=1 P (t k,i |T k,<i , S k , S doc , \u0398) (2)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Cross-Sentence Context", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We make an assumption to simplify modeling and consider only two previous source sentences as the cross-sentence context (S doc = S k\u22121 , S k\u22122 ). We also do not explicitly consider previously corrected target sentences to avoid error propagation (Wang et al., 2017) . We extend a sentencelevel encoder-decoder baseline model to build our cross-sentence GEC model. We describe our baseline encoder-decoder model and the proposed cross-sentence model below.", |
| "cite_spans": [ |
| { |
| "start": 247, |
| "end": 266, |
| "text": "(Wang et al., 2017)", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Cross-Sentence Context", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We use a deep hierarchical convolutional encoderdecoder model (Gehring et al., 2017) as our encoder-decoder framework. Ensembles of convolutional models have achieved high performance for GEC and are used in two recent state-of-theart GEC models (Chollampatt and Ng, 2018b; Ge et al., 2018) .", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 84, |
| "text": "(Gehring et al., 2017)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 246, |
| "end": 273, |
| "text": "(Chollampatt and Ng, 2018b;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 274, |
| "end": 290, |
| "text": "Ge et al., 2018)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "A source sentence S is embedded as S = EMB tok (S) + EMB pos (S) where S \u2208 R |S|\u00d7d . EMB tok (\u2022) and EMB pos (\u2022) are token embeddings and learned position embeddings, respectively. The source embeddings S are then passed through a linear layer (denoted by LIN) before they are fed to the initial encoder layer as H 0 = LIN(S), where H 0 \u2208 R |S|\u00d7h . The output of the lth encoder layer H l \u2208 R |S|\u00d7h , where l = 1, . . . , L, is computed by passing the output of the previous layer H l\u22121 through a convolutional neural network (CONV) followed by gated linear unit (GLU) activation function (Dauphin et al., 2017), and residual connections (He et al., 2016) :", |
| "cite_spans": [ |
| { |
| "start": 638, |
| "end": 655, |
| "text": "(He et al., 2016)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "H l = GLU(CONV(H l\u22121 )) + H l\u22121", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Output of the encoder is a transformation of the final encoder layer output H L to E \u2208 R |S|\u00d7d . The decoder network also consists of multiple layers. During the prediction of the (n+1)th word, the decoder has access to previously predicted n words t 1 , . . . , t n and their embeddings T \u2208 R n\u00d7d . The embeddings are obtained in the same way as the encoder via separate embedding layers in the decoder. T passes through a linear layer to obtain the input to the initial decoder layer G 0 \u2208 R n\u00d7h . The lth decoder layer computes an intermediate representation Y l \u2208 R n\u00d7h by passing the outputs of the previous decoder layer G l\u22121 through a CONV layer followed by a GLU activation: Additionally, each decoder layer has an attention mechanism that utilizes Y l to compute summarized representations C l \u2208 R n\u00d7h of the encoder states E \u2208 R |S|\u00d7d for n decoder states. Attention at layer l is computed as follows", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Y l = GLU(CONV(G l\u22121 ))", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "Z l = LIN(Y l ) + T (5) X l = SOFTMAX(Z l \u2022 E ) \u2022 (E + S) (6) C l = LIN(X l )", |
| "eq_num": "(7)" |
| } |
| ], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Output of the lth decoder layer is computed as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "G l = Y l + C l + G l\u22121", |
| "eq_num": "(8)" |
| } |
| ], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The decoder outputs a linear transformation of the final decoder layer output G L to D \u2208 R n\u00d7d . For predicting the (n+1)th target word, the output softmax is computed after a linear transformation of the last decoder state to the output vocabulary size. The last decoder state corresponds to the last row of D.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Baseline Encoder-Decoder Framework", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Our proposed model ( Figure 2 ) incorporates crosssentence context using an auxiliary encoder that is similar in structure to our source sentence encoder (Section 2.1). For encoding the context consisting of two previous sentences, we simply concatenate these two sentences and pass it to the auxiliary encoder. Let\u015c represent the cross-sentence context consisting of |\u015c| tokens\u015d 1 , ...,\u015d |\u015c| from the previous source sentences. They are embedded as\u015c \u2208 R |\u015c|\u00d7d using separate token and position embedding layers and fed into the auxiliary multilayer encoder (consisting of L identical layers) to obtain the auxiliary encoder output\u00ca \u2208 R |\u015c|\u00d7d .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 21, |
| "end": 29, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "For integrating the auxiliary encoder output during decoding, each decoder layer employs a separate attention mechanism similar to that described previously. We use another LIN layer to comput\u00ea Z l (Equation 5) and use\u00ca and\u015c in place of E and S, respectively, to compute summarized auxiliary encoder representations\u0108 l (Equations 6 and 7).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The summarized auxiliary encoder representation at each layer\u0108 l is added to the output of the layer along with other terms in Equation 8. All corrections do not depend equally on crosssentence context. So, instead of adding\u0108 l directly with an equal weight as the other terms, we add a gating \u039b l \u2208 R n\u00d7h at each layer to control the cross-sentence information that gets passed:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "G l = Y l + C l + \u039b l \u2022\u0108 l + G l\u22121 (9)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "where \u2022 denotes the Hadamard product. The gate \u039b l is determined based on Y l and the summarized context representations of the source sentence C l .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u039b l = \u03c3(LIN(Y l ) + LIN(C l ))", |
| "eq_num": "(10)" |
| } |
| ], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "where \u03c3 represents element-wise sigmoid activation that restricts values to [0, 1]. \u039b l can be regarded as probabilities of retaining values in\u0108 l . We also analyze our proposed approach by comparing against two other ways for modeling cross-sentence context (Section 5.1). One way is to integrate representations from the auxiliary encoder directly in the decoder layers via the attention mechanism, without gating. Another more straightforward way of incorporating crosssentence context is by simply passing the concatenation of the previous source sentences and the current source sentence (Tiedemann and Scherrer, 2017) to the main encoder itself, without employing an auxiliary encoder.", |
| "cite_spans": [ |
| { |
| "start": 593, |
| "end": 623, |
| "text": "(Tiedemann and Scherrer, 2017)", |
| "ref_id": "BIBREF29" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Cross-Sentence GEC Model", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We further improve our models (both sentencelevel baseline and cross-sentence model) with several techniques from prior work that have been shown to be useful for GEC. They include initializing the word embedding vectors with pretrained embeddings, pretraining the decoder on large English corpora using a language modeling objective , label smoothing (Szegedy et al., 2016) , and dropping out entire word embeddings of source words during training. We found that using the technique of training with an edit-weighted objective (Junczys-Dowmunt et al., 2018) results in a higher recall but hurts the overall performance of our baseline model, and hence we do not use it in our models.", |
| "cite_spans": [ |
| { |
| "start": 352, |
| "end": 374, |
| "text": "(Szegedy et al., 2016)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Other Techniques and BERT Rescoring", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "We also rescore the final candidates using feature-based rescoring (with edit operation and language model features) for both our sentencelevel baseline and our cross-sentence model following Chollampatt and Ng (2018a) . We investigate the effectiveness of BERT (Devlin et al., 2018) for GEC. We use masked language model probabilities computed by pretrained BERT model 2 as an additional feature during rescoring to further improve our sentence-level baseline and our cross-sentence GEC model. Specifically, we replace tokens in a hypothesis T with the [MASK] token, one at a time, predicting the log probability of the masked token using BERT. The final BERT feature is computed as", |
| "cite_spans": [ |
| { |
| "start": 192, |
| "end": 218, |
| "text": "Chollampatt and Ng (2018a)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 262, |
| "end": 283, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 554, |
| "end": 560, |
| "text": "[MASK]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Other Techniques and BERT Rescoring", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "f BERT (T ) = |T | i=1 log P BERT (t i |T \u2212i )", |
| "eq_num": "(11)" |
| } |
| ], |
| "section": "Other Techniques and BERT Rescoring", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "where T \u2212i is the target sentence where the ith target word t i is masked.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Other Techniques and BERT Rescoring", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "Similar to most published GEC models (Chollampatt and Ng, 2018b; Junczys-Dowmunt et al., 2018; Grundkiewicz and Junczys-Dowmunt, 2018), we rely on two datasets for training: Lang-8 Learner Corpora 3 v2.0 (Mizumoto et al., 2011) and NUCLE (Dahlmeier et al., 2013) . Both datasets have document context available which make them suitable for cross-sentence GEC. We split training and development datasets based on essay boundaries. We extract development set from a subset of NUCLE. To ensure the development set has a high number of error annotations, we sort the essays in decreasing order of the ratio of corrected sentences per essay. We select 25% of essays from the top (sampling one essay from every four) until we get over 5,000 annotated sentences. The remaining essays from NUCLE are used for training. From Lang-8, we extract essays written by learners whose native language is English and contain at least two English sentences with a minimum of one annotation. We identify the language of a sentence using langid.py (Lui and Baldwin, 2012) . From the extracted essays, we select annotated English sentences (with at most 80 tokens) as our source sentences and their corresponding corrections as our target sentences. Statistics of the datasets are given in Table 1 . For training our cross-sentence GEC models, we select two previous English sentences for each source sentence from its essay as the cross-sentence context. We found that using two previous sentences performed better than using only one previous sentence as the context. Still, in our dataset, the first English sentence of an essay has an empty context and the second English sentence of an essay has only one previous sentence as its context. We evaluate on three test sets which have document-level contexts: CoNLL-2013 and CoNLL-2014 (Ng et al., 2014 shared task test sets, and Cambridge Learner Corpus-First Certificate Exam or FCE test set (Yannakoudakis et al., 2011) . Another recent dataset, JFLEG (Napoles et al., 2017) , does not have documentlevel contexts available and hence we do not use it for evaluation. We use the M 2 scorer (Dahlmeier and Ng, 2012) for evaluation and perform significance tests using one-tailed sign test with bootstrap resampling on 100 samples.", |
| "cite_spans": [ |
| { |
| "start": 204, |
| "end": 227, |
| "text": "(Mizumoto et al., 2011)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 238, |
| "end": 262, |
| "text": "(Dahlmeier et al., 2013)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 1027, |
| "end": 1050, |
| "text": "(Lui and Baldwin, 2012)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 1789, |
| "end": 1799, |
| "text": "CoNLL-2013", |
| "ref_id": null |
| }, |
| { |
| "start": 1804, |
| "end": 1814, |
| "text": "CoNLL-2014", |
| "ref_id": null |
| }, |
| { |
| "start": 1815, |
| "end": 1831, |
| "text": "(Ng et al., 2014", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 1923, |
| "end": 1951, |
| "text": "(Yannakoudakis et al., 2011)", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 1984, |
| "end": 2006, |
| "text": "(Napoles et al., 2017)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 2121, |
| "end": 2145, |
| "text": "(Dahlmeier and Ng, 2012)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1268, |
| "end": 1275, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data and Evaluation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We extend a convolutional encoder-decoder model following recent state-of-the-art sentence-level GEC models (Chollampatt and Ng, 2018b; Ge et al., 2018) with identical architecture and hyperparameters to build our sentence-level baseline. Table 2 : Results of our proposed cross-sentence GEC model (CROSENT). 'avg' row reports average precision (P), recall (R), and F 0.5 score of 4 independently trained single models. 'ens' denotes results of the 4-model ensemble. 'rs' denotes results of feature-based rescoring, and 'rs BERT ' additionally uses BERT feature for rescoring. All CROSENT results are statistically significant compared to the corresponding BASELINE results (p < 0.001).", |
| "cite_spans": [ |
| { |
| "start": 108, |
| "end": 135, |
| "text": "(Chollampatt and Ng, 2018b;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 136, |
| "end": 152, |
| "text": "Ge et al., 2018)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 239, |
| "end": 246, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Model", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The final system of Chollampatt and Ng (2018b) is an ensemble of three variants of this architecture initialized and trained differently. We replicate the best-performing variant using Fairseq 4 v0.5 on our training data. We incorporate the techniques mentioned in Section 2.3 such as source word dropout, pretraining word embedding, decoder pretraining, and label smoothing. We reuse the the pretrained word embeddings and vocabularies from (Chollampatt and Ng, 2018a) and the pretrained decoder from (Chollampatt and Ng, 2018b) . Word embeddings had been pretrained on Wikipedia corpus consisting of 1.78 billion words. The decoder had been pretrained using 100 million sentences (1.42 billion words) from Common Crawl. To fit training of a single convolutional encoder-decoder model efficiently in a single Titan X GPU with 12 GB memory, we restrict each batch to a maximum of 6,000 source or target tokens per batch, apart from setting a maximum batch size of 96 instances. All other hyperparameters, pretrained models, and vocabularies are from (Chollampatt and Ng, 2018b) , with source token dropout of 0.2 and label smoothing parameter of 0.1. We refer to this model as BASELINE in our results. The best model is chosen based on the development set perplexity.", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 46, |
| "text": "Chollampatt and Ng (2018b)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 502, |
| "end": 529, |
| "text": "(Chollampatt and Ng, 2018b)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1050, |
| "end": 1077, |
| "text": "(Chollampatt and Ng, 2018b)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We extend Fairseq and implement an auxiliary encoder for modeling previous sentences. We use the source vocabulary itself as the vocabulary for the auxiliary encoder and initialize its embeddings with the same pretrained word embeddings used to initialize the source encoder. We use embeddings of size 500. We use three layers in the auxil-iary encoder with the output size set to 1024. The cross-sentence context is encoded by the auxiliary encoder. In the case of the first sentence in an essay, an empty context consisting of only padding tokens and the end-of-sentence marker token is passed. We denote our cross-sentence GEC model as CROSENT.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Model", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We evaluate the performance of our cross-sentence GEC model, CROSENT, and compare it to the sentence-level BASELINE on three test sets in Table 2 . Average single model results ('avg') show a notable improvement on CoNLL-2013 and CoNLL-2014 for our CROSENT model. On FCE, a significant improvement of 0.32 F 0.5 is observed (p < 0.001) for a single model. CoNLL-2013 and FCE have only a single set of annotations for each sentence and hence, reference coverage could potentially underrate model performance on these test sets. Interestingly, the performance gap widens for an ensemble of models. CROSENT achieves significant improvements of 2.87, 2.74, and 1.43, respectively, on the three test sets. Feature-based rescoring ('rs') shows further improvements. Adding BERT ('rs BERT ') improves our BASELINE and CROSENT model further. The improvements due to the integration of the crosssentence component still remain notable and significant.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 138, |
| "end": 145, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We compare our CROSENT model to the state of the art.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to the State of the Art", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Our best result of 55.57 for CROSENT (ens+rs BERT ) is competitive to the best published models trained using the same training datasets: NUS 1+2+3 (Chollampatt and Ng, 2018b) , G&J , JGGH , and NUS 1 (Chollampatt and Ng, 2018a) . NUS 1+2+3 is a 12-model ensemble of three sets of models (4 each), with one set reused from NUS 1 . We re-implement this ensemble model ( NUS 1+2+3 ) using identical preprocessing and data used by our BASELINE, without rescoring and spell checking. Our replicated results reach 52.49. We replace the weakest set of 4 models in NUS 1+2+3 , i.e., NUS 1 , in the ensemble with our 4 BASELINE models instead. We reach a result of 52.77. We then use our CROSENT models in place of BASELINE models and observe notable improvements of 2.1 F 0.5 score. When we use BERT feature and rescore, our improved baseline NUS 2+3 +BASELINE (ens+rs BERT ) achieves 55.47 F 0.5 score, and using CROSENT models instead of our BASELINE models achieves a result of 57.16 (+1.69 F 0.5 ). This result is better than that of all prior competing models trained on the same datasets 5 , of which three systems use a spell checker from (Chollampatt and Ng, 2018a) . When 5 After the submission of this paper, improved results on the CoNLL-2014 benchmark have been published (Zhao et al., 2019; Lichtarge et al., 2019; Stahlberg et al., 2019 we add this spell checker in a similar way, our final result reaches 57.30 F 0.5 . With much larger high-quality annotated training datasets, a better result of 60.0 is achieved by MSR (Ge et al., 2018) . Recently, Google (Lichtarge et al., 2018) published a higher result of 58.3 by an ensemble of big Transformers (Vaswani et al., 2017) additionally trained on Wikipedia edits (4.1 billion words). Such large-scale training is very likely to further improve our model as well. We leave it to future work to explore large-scale cross-sentence GEC.", |
| "cite_spans": [ |
| { |
| "start": 148, |
| "end": 175, |
| "text": "(Chollampatt and Ng, 2018b)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 201, |
| "end": 228, |
| "text": "(Chollampatt and Ng, 2018a)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 1139, |
| "end": 1166, |
| "text": "(Chollampatt and Ng, 2018a)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 1174, |
| "end": 1175, |
| "text": "5", |
| "ref_id": null |
| }, |
| { |
| "start": 1277, |
| "end": 1296, |
| "text": "(Zhao et al., 2019;", |
| "ref_id": "BIBREF37" |
| }, |
| { |
| "start": 1297, |
| "end": 1320, |
| "text": "Lichtarge et al., 2019;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 1321, |
| "end": 1343, |
| "text": "Stahlberg et al., 2019", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 1529, |
| "end": 1546, |
| "text": "(Ge et al., 2018)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 1566, |
| "end": 1590, |
| "text": "(Lichtarge et al., 2018)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 1660, |
| "end": 1682, |
| "text": "(Vaswani et al., 2017)", |
| "ref_id": "BIBREF31" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Comparison to the State of the Art", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We analyze our model choices on our development data (5,006 sentences from NUCLE) and held-out CoNLL-2013 test set. We have not used CoNLL-2013 test set directly during training and hence, it can be used to evaluate the generalization performance of our models. We also analyze model outputs on the CoNLL-2013 test set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Analysis", |
| "sec_num": "5" |
| }, |
| { |
| "text": "We investigate different mechanisms of integrating cross-sentence context. Table 4 shows the average single model results of our sentence-level BASELINE compared to two different strategies of integrating cross-sentence context. 'concat' refers to simply prepending the previous source sentences to the current source sentence. The context and the current source sentence is separated by a special token (<CONCAT>). This model does not have an auxiliary encoder. 'aux (no gate)' uses an auxiliary encoder similar to our CROSENT model except for gating. 'aux (+gate)' is our CROSENT model (Section 2.2) which employs the auxiliary encoder with the gating mechanism. The first two variants perform comparably to our sentence-level BASELINE and shows no notable gains from using cross-sentence context. When I agree that RFID technology shall not be made available to the public for easy abuse and distorted usage. First, our privacy is at threat. Though tracking devices such as the applications in our smartphones these days can add fun and entertainment to our fast-living paced livings, this can be a double-edged sword.We revealed our locations, for our friends to catch up with us, however we can never know who is watching us out there secretly. BASELINE: We revealed our locations, for our friends to catch up with us, ... CROSENT: We reveal our locations, for our friends to catch up with us, however ...", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 75, |
| "end": 82, |
| "text": "Table 4", |
| "ref_id": "TABREF6" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Modeling Cross-Sentence Context", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We reveal our locations, for our friends to catch up with us, ... the gating mechanism is added, results improve substantially. Using the gating mechanism is crucial in our CROSENT model, as it has the ability to selectively pass information through. This shows that properly modeling cross-sentence context is essential to improve overall performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "REF:", |
| "sec_num": null |
| }, |
| { |
| "text": "We show an example from the CoNLL-2013 test set which involves a verb tense error that requires cross-sentence context for correction. The original sentence (with its context) and the corrections made by our BASELINE and CROSENT ensemble models are shown in Figure 3 . The verb 'revealed' is to be corrected to its present tense form 'reveal' according to the annotated reference (REF). Our CROSENT model corrects this verb tense error accurately. However, our sentence-level BASELINE is unable to make this correction as it only has access to the sentence-level context. To investigate if cross-sentence context is adequately captured by our model, we adopt a larger scale controlled evaluation. To do this, 795 wellformed sentences are obtained from multiple documents in simple English Wikipedia along with their previous sentences. We create a synthetic dataset of verb tense errors, by corrupting all verbs in these 795 sentences and replacing them by their present tense form 6 , producing 1,090 synthetic verb tense errors. These errors require cross- 6 Adapted from https://github.com/bendichter/tenseflow P / R / F 0.5 n c / n p BASELINE 22.86 / 8.35 / 16.96 91 / 398 CROSENT 44.60 / 20.83 / 36.31 227 / 509 We analyze the performance of our crosssentence model on this dataset by passing the corrupted sentences and their contexts as input. We evaluate the ability of our model to correct the verb tense errors and recover the original sentences. The result of our BASELINE and CROSENT ensemble models on this dataset is shown in Table 5 . In this dataset, we find a sharp increase in both precision and recall for our CROSENT model showing their ability to capture cross-sentence context. While the BASELINE accurately corrects 91 errors, our CROSENT model accurately corrects 227 errors. The number of proposed corrections (n p ) is also significantly higher for CROSENT. The results indicate that our cross-sentence model can identify and correct a significant number of crosssentence errors.", |
| "cite_spans": [ |
| { |
| "start": 1059, |
| "end": 1060, |
| "text": "6", |
| "ref_id": null |
| }, |
| { |
| "start": 1138, |
| "end": 1216, |
| "text": "BASELINE 22.86 / 8.35 / 16.96 91 / 398 CROSENT 44.60 / 20.83 / 36.31 227 / 509", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 258, |
| "end": 266, |
| "text": "Figure 3", |
| "ref_id": "FIGREF2" |
| }, |
| { |
| "start": 1540, |
| "end": 1547, |
| "text": "Table 5", |
| "ref_id": "TABREF7" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Ability to Correct Cross-Sentence Errors", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "We evaluate the overall performance on common error types (Figure 4 ) on the CoNLL-2013 dataset using ERRANT (Bryant et al., 2017) . They include determiner (DET), noun number (NOUN:NUM), preposition (PREP), verb tense (VERB:TENSE), and errors that are not categorized (OTHERS). We find that the largest margins of improvements are observed on verb tense errors and determiner errors. This aligns with our expectation of cross-sentence models. ", |
| "cite_spans": [ |
| { |
| "start": 109, |
| "end": 130, |
| "text": "(Bryant et al., 2017)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 58, |
| "end": 67, |
| "text": "(Figure 4", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Overall Performance across Error Types", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "We also analyze the output words that produce attention distributions on the auxiliary context that deviate the most from a uniform distribution, indicating their dependence on cross-sentence context. For each output word, we find the KL divergence between its auxiliary attention distribution and a uniform distribution on words in the auxiliary context. We average the KL divergences for all instances of this output word in the CoNLL-2013 test set. We then identify the top words for our CROSENT ensemble model. The top words mostly include nouns (such as criminal, chip, economy, and individual). Manual analysis shows that many of these nouns appear in consecutive sentences, which results in higher attention placed on the same words on the source side. We conduct a more coarse-grained analysis to find the top 5 part-of-speech tags of output words (based on Universal tags) that produce the highest average KL divergences (Table 6 ). Nouns and proper nouns (PROPN) attending to the same word in previous source sentences cause them to rank higher. Verbs and determiners (DET) are also among the top tags since they often require cross-sentence context for disambiguation. According to ERRANT (Section 5.3), compared to our BASELINE ensemble, CROSENT ensemble achieves notable improvements in performance on verb tense errors (+7.8% F 0.5 ) and on determiner errors (+4.7%).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 930, |
| "end": 938, |
| "text": "(Table 6", |
| "ref_id": "TABREF9" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Attention on Auxiliary Context", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "Sophisticated sequence-to-sequence architectures (Gehring et al., 2017; Vaswani et al., 2017) 2018b), rescoring (Chollampatt and Ng, 2018a) , iterative decoding strategies (Ge et al., 2018; Lichtarge et al., 2018) , synthetic (Xie et al., 2018) and semi-supervised corpora (Lichtarge et al., 2018) , and other task-specific techniques has achieved impressive results on this task. However, all prior work ignores document-wide context for GEC, and uses sentence-level models. For spell checking, Flor and Futagi (2012) used document-level context to check if a candidate correction for a misspelled word had been used earlier in the document. Zheng et al. (2018) proposed splitting runon sentences into separate sentences. However, they did not use cross-sentence context.", |
| "cite_spans": [ |
| { |
| "start": 49, |
| "end": 71, |
| "text": "(Gehring et al., 2017;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 72, |
| "end": 93, |
| "text": "Vaswani et al., 2017)", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 112, |
| "end": 139, |
| "text": "(Chollampatt and Ng, 2018a)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 172, |
| "end": 189, |
| "text": "(Ge et al., 2018;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 190, |
| "end": 213, |
| "text": "Lichtarge et al., 2018)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 226, |
| "end": 244, |
| "text": "(Xie et al., 2018)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 273, |
| "end": 297, |
| "text": "(Lichtarge et al., 2018)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 496, |
| "end": 518, |
| "text": "Flor and Futagi (2012)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 643, |
| "end": 662, |
| "text": "Zheng et al. (2018)", |
| "ref_id": "BIBREF38" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "On the other hand, there are a number of studies recently on integrating cross-sentence context for neural machine translation (NMT). There are three major approaches for document-level NMT: (1) translating an extended source (context concatenated with the source) to a single or extended target (Tiedemann and Scherrer, 2017; Bawden et al., 2018) ; (2) using an additional encoder to capture document-wide context (Jean et al., 2017; Wang et al., 2017; Bawden et al., 2018; Voita et al., 2018; Miculicich et al., 2018; ; and (3) using discrete (Kuang et al., 2018) or continuous cache (Tu et al., 2018; Maruf and Haffari, 2018) mechanisms during translation for storing and retrieving document-level information. We investigated the first two approaches in this paper as we believe that most of the ambiguities can be resolved by considering a few previous sentences. Since GEC is a monolingual rewriting task, most of the disambiguating information is in the source sentence itself, unlike bilingual NMT. All approaches for document-level NMT extended recurrent models or Transformer models. There is no prior work that extends convolutional sequence-to-sequence models for document-level NMT.", |
| "cite_spans": [ |
| { |
| "start": 296, |
| "end": 326, |
| "text": "(Tiedemann and Scherrer, 2017;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 327, |
| "end": 347, |
| "text": "Bawden et al., 2018)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 415, |
| "end": 434, |
| "text": "(Jean et al., 2017;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 435, |
| "end": 453, |
| "text": "Wang et al., 2017;", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 454, |
| "end": 474, |
| "text": "Bawden et al., 2018;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 475, |
| "end": 494, |
| "text": "Voita et al., 2018;", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 495, |
| "end": 519, |
| "text": "Miculicich et al., 2018;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 545, |
| "end": 565, |
| "text": "(Kuang et al., 2018)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 586, |
| "end": 603, |
| "text": "(Tu et al., 2018;", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 604, |
| "end": 628, |
| "text": "Maruf and Haffari, 2018)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "The aim of this paper is to demonstrate the necessity of modeling cross-sentence context for GEC. It is beyond the scope of this paper to comprehensively evaluate all approaches to exploit document-level context, and we leave it to future work to evaluate more sophisticated models such as memory networks to capture entire documentlevel context or incorporate external knowledge sources.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We present the first approach to cross-sentence GEC, building on a convolutional encoderdecoder architecture. Our cross-sentence models show significant gains over strong sentencelevel baselines. On the CoNLL-2014 benchmark test set, when using larger ensembles and integrating BERT during rescoring, our final crosssentence model achieves 57.30 F 0.5 score which is higher than all prior published F 0.5 scores at the time of paper submission, when trained using the same datasets. We also demonstrate the ability of our models to exploit wider contexts adequately and correct errors on a synthetic test set of crosssentence verb tense errors.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "Our source code is publicly available at https:// github.com/nusnlp/crosentgec.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We use pretrained cased BERT base model from https://github.com/huggingface/pytorch-pretrained-BERT 3 https://sites.google.com/site/naistlang8corpora", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/pytorch/fairseq", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We keep the previous sentences in their actual form to analyze the model's ability in a controlled way. However, in reality, previous sentences may also contain errors.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Evaluating discourse phenomena in neural machine translation", |
| "authors": [ |
| { |
| "first": "Rachel", |
| "middle": [], |
| "last": "Bawden", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "Alexandra", |
| "middle": [], |
| "last": "Birch", |
| "suffix": "" |
| }, |
| { |
| "first": "Barry", |
| "middle": [], |
| "last": "Haddow", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rachel Bawden, Rico Sennrich, Alexandra Birch, and Barry Haddow. 2018. Evaluating discourse phe- nomena in neural machine translation. In Proceed- ings of the 2018 Conference of the North American Chapter of the Association for Computational Lin- guistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Automatic annotation and evaluation of error types for grammatical error correction", |
| "authors": [ |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Bryant", |
| "suffix": "" |
| }, |
| { |
| "first": "Mariano", |
| "middle": [], |
| "last": "Felice", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher Bryant, Mariano Felice, and Ted Briscoe. 2017. Automatic annotation and evaluation of error types for grammatical error correction. In Proceed- ings of the 55th Annual Meeting of the Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "A multilayer convolutional encoder-decoder neural network for grammatical error correction", |
| "authors": [ |
| { |
| "first": "Shamil", |
| "middle": [], |
| "last": "Chollampatt", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 32nd AAAI Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shamil Chollampatt and Hwee Tou Ng. 2018a. A mul- tilayer convolutional encoder-decoder neural net- work for grammatical error correction. In Proceed- ings of the 32nd AAAI Conference on Artificial In- telligence.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Neural quality estimation of grammatical error correction", |
| "authors": [ |
| { |
| "first": "Shamil", |
| "middle": [], |
| "last": "Chollampatt", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shamil Chollampatt and Hwee Tou Ng. 2018b. Neural quality estimation of grammatical error correction. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Better evaluation for grammatical error correction", |
| "authors": [ |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Dahlmeier", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Daniel Dahlmeier and Hwee Tou Ng. 2012. Better evaluation for grammatical error correction. In Pro- ceedings of the 2012 Conference of the North Amer- ican Chapter of the Association for Computational Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Building a large annotated corpus of learner English: The NUS corpus of learner English", |
| "authors": [ |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Dahlmeier", |
| "suffix": "" |
| }, |
| { |
| "first": "Siew Mei", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Eighth Workshop on Innovative Use of NLP for Building Educational Applications", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Daniel Dahlmeier, Hwee Tou Ng, and Siew Mei Wu. 2013. Building a large annotated corpus of learner English: The NUS corpus of learner English. In Proceedings of the Eighth Workshop on Innovative Use of NLP for Building Educational Applications.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Language modeling with gated convolutional networks", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Yann", |
| "suffix": "" |
| }, |
| { |
| "first": "Angela", |
| "middle": [], |
| "last": "Dauphin", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Auli", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Grangier", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 34th International Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier. 2017. Language modeling with gated convolutional networks. In Proceedings of the 34th International Conference on Machine Learn- ing.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1810.04805" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language under- standing. arXiv preprint arXiv:1810.04805.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Multi-way, multilingual neural machine translation with a shared attention mechanism", |
| "authors": [ |
| { |
| "first": "Orhan", |
| "middle": [], |
| "last": "Firat", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshua", |
| "middle": [], |
| "last": "Bengio", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Orhan Firat, Kyunghyun Cho, and Yoshua Bengio. 2016. Multi-way, multilingual neural machine translation with a shared attention mechanism. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "On using context for automatic correction of non-word misspellings in student essays", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Flor", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoko", |
| "middle": [], |
| "last": "Futagi", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the Seventh Workshop on Building Educational Applications Using NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Michael Flor and Yoko Futagi. 2012. On using context for automatic correction of non-word misspellings in student essays. In Proceedings of the Seventh Workshop on Building Educational Applications Us- ing NLP.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Reaching human-level performance in automatic grammatical error correction: An empirical study", |
| "authors": [ |
| { |
| "first": "Tao", |
| "middle": [], |
| "last": "Ge", |
| "suffix": "" |
| }, |
| { |
| "first": "Furu", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1807.01270" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tao Ge, Furu Wei, and Ming Zhou. 2018. Reaching human-level performance in automatic grammatical error correction: An empirical study. arXiv preprint arXiv:1807.01270.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Convolutional sequence to sequence learning", |
| "authors": [ |
| { |
| "first": "Jonas", |
| "middle": [], |
| "last": "Gehring", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Auli", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Grangier", |
| "suffix": "" |
| }, |
| { |
| "first": "Denis", |
| "middle": [], |
| "last": "Yarats", |
| "suffix": "" |
| }, |
| { |
| "first": "Yann", |
| "middle": [ |
| "N" |
| ], |
| "last": "Dauphin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 34th International Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jonas Gehring, Michael Auli, David Grangier, De- nis Yarats, and Yann N. Dauphin. 2017. Convolu- tional sequence to sequence learning. In Proceed- ings of the 34th International Conference on Ma- chine Learning.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Near human-level performance in grammatical error correction with hybrid machine translation", |
| "authors": [ |
| { |
| "first": "Roman", |
| "middle": [], |
| "last": "Grundkiewicz", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcin", |
| "middle": [], |
| "last": "Junczys-Dowmunt", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Roman Grundkiewicz and Marcin Junczys-Dowmunt. 2018. Near human-level performance in grammati- cal error correction with hybrid machine translation. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies (Short Papers).", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Deep residual learning for image recognition", |
| "authors": [ |
| { |
| "first": "Kaiming", |
| "middle": [], |
| "last": "He", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiangyu", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Shaoqing", |
| "middle": [], |
| "last": "Ren", |
| "suffix": "" |
| }, |
| { |
| "first": "Jian", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recog- nition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Does neural machine translation benefit from larger context? arXiv preprint", |
| "authors": [ |
| { |
| "first": "Sebastien", |
| "middle": [], |
| "last": "Jean", |
| "suffix": "" |
| }, |
| { |
| "first": "Stanislas", |
| "middle": [], |
| "last": "Lauly", |
| "suffix": "" |
| }, |
| { |
| "first": "Orhan", |
| "middle": [], |
| "last": "Firat", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyunghyun", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1704.05135" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sebastien Jean, Stanislas Lauly, Orhan Firat, and Kyunghyun Cho. 2017. Does neural machine trans- lation benefit from larger context? arXiv preprint arXiv:1704.05135.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Approaching neural grammatical error correction as a low-resource machine translation task", |
| "authors": [ |
| { |
| "first": "Marcin", |
| "middle": [], |
| "last": "Junczys-Dowmunt", |
| "suffix": "" |
| }, |
| { |
| "first": "Roman", |
| "middle": [], |
| "last": "Grundkiewicz", |
| "suffix": "" |
| }, |
| { |
| "first": "Shubha", |
| "middle": [], |
| "last": "Guha", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenneth", |
| "middle": [], |
| "last": "Heafield", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcin Junczys-Dowmunt, Roman Grundkiewicz, Shubha Guha, and Kenneth Heafield. 2018. Ap- proaching neural grammatical error correction as a low-resource machine translation task. In Proceed- ings of the 2018 Conference of the North American Chapter of the Association for Computational Lin- guistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Modeling coherence for neural machine translation with dynamic and topic caches", |
| "authors": [ |
| { |
| "first": "Shaohui", |
| "middle": [], |
| "last": "Kuang", |
| "suffix": "" |
| }, |
| { |
| "first": "Deyi", |
| "middle": [], |
| "last": "Xiong", |
| "suffix": "" |
| }, |
| { |
| "first": "Weihua", |
| "middle": [], |
| "last": "Luo", |
| "suffix": "" |
| }, |
| { |
| "first": "Guodong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 27th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shaohui Kuang, Deyi Xiong, Weihua Luo, and Guodong Zhou. 2018. Modeling coherence for neural machine translation with dynamic and topic caches. In Proceedings of the 27th International Conference on Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Attention strategies for multi-source sequence-to-sequence learning", |
| "authors": [ |
| { |
| "first": "Jind\u0159ich", |
| "middle": [], |
| "last": "Libovick\u00fd", |
| "suffix": "" |
| }, |
| { |
| "first": "Jind\u0159ich", |
| "middle": [], |
| "last": "Helcl", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jind\u0159ich Libovick\u00fd and Jind\u0159ich Helcl. 2017. Attention strategies for multi-source sequence-to-sequence learning. In Proceedings of the 55th Annual Meet- ing of the Association for Computational Linguistics (Short Papers).", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Corpora generation for grammatical error correction", |
| "authors": [ |
| { |
| "first": "Jared", |
| "middle": [], |
| "last": "Lichtarge", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Alberti", |
| "suffix": "" |
| }, |
| { |
| "first": "Shankar", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Simon", |
| "middle": [], |
| "last": "Tong", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jared Lichtarge, Chris Alberti, Shankar Kumar, Noam Shazeer, Niki Parmar, and Simon Tong. 2019. Cor- pora generation for grammatical error correction. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Weakly supervised grammatical error correction using iterative decoding", |
| "authors": [ |
| { |
| "first": "Jared", |
| "middle": [], |
| "last": "Lichtarge", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Alberti", |
| "suffix": "" |
| }, |
| { |
| "first": "Shankar", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1811.01710" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jared Lichtarge, Christopher Alberti, Shankar Kumar, Noam Shazeer, and Niki Parmar. 2018. Weakly su- pervised grammatical error correction using iterative decoding. arXiv preprint arXiv:1811.01710.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "langid.py: An off-the-shelf language identification tool", |
| "authors": [ |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Lui", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Baldwin", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the ACL 2012 System Demonstrations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marco Lui and Timothy Baldwin. 2012. langid.py: An off-the-shelf language identification tool. In Pro- ceedings of the ACL 2012 System Demonstrations.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Document context neural machine translation with memory networks", |
| "authors": [ |
| { |
| "first": "Sameen", |
| "middle": [], |
| "last": "Maruf", |
| "suffix": "" |
| }, |
| { |
| "first": "Gholamreza", |
| "middle": [], |
| "last": "Haffari", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sameen Maruf and Gholamreza Haffari. 2018. Docu- ment context neural machine translation with mem- ory networks. In Proceedings of the 56th Annual Meeting of the Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Document-level neural machine translation with hierarchical attention networks", |
| "authors": [ |
| { |
| "first": "Lesly", |
| "middle": [], |
| "last": "Miculicich", |
| "suffix": "" |
| }, |
| { |
| "first": "Dhananjay", |
| "middle": [], |
| "last": "Ram", |
| "suffix": "" |
| }, |
| { |
| "first": "Nikolaos", |
| "middle": [], |
| "last": "Pappas", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Henderson", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lesly Miculicich, Dhananjay Ram, Nikolaos Pappas, and James Henderson. 2018. Document-level neural machine translation with hierarchical attention net- works. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Process- ing.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Mining revision log of language learning SNS for automated Japanese error correction of second language learners", |
| "authors": [ |
| { |
| "first": "Tomoya", |
| "middle": [], |
| "last": "Mizumoto", |
| "suffix": "" |
| }, |
| { |
| "first": "Mamoru", |
| "middle": [], |
| "last": "Komachi", |
| "suffix": "" |
| }, |
| { |
| "first": "Masaaki", |
| "middle": [], |
| "last": "Nagata", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuji", |
| "middle": [], |
| "last": "Matsumoto", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the Fifth International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tomoya Mizumoto, Mamoru Komachi, Masaaki Na- gata, and Yuji Matsumoto. 2011. Mining revi- sion log of language learning SNS for automated Japanese error correction of second language learn- ers. In Proceedings of the Fifth International Joint Conference on Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "JFLEG: A fluency corpus and benchmark for grammatical error correction", |
| "authors": [ |
| { |
| "first": "Courtney", |
| "middle": [], |
| "last": "Napoles", |
| "suffix": "" |
| }, |
| { |
| "first": "Keisuke", |
| "middle": [], |
| "last": "Sakaguchi", |
| "suffix": "" |
| }, |
| { |
| "first": "Joel", |
| "middle": [], |
| "last": "Tetreault", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Courtney Napoles, Keisuke Sakaguchi, and Joel Tetreault. 2017. JFLEG: A fluency corpus and benchmark for grammatical error correction. In Pro- ceedings of the 15th Conference of the European Chapter of the Association for Computational Lin- guistics (Short Papers).", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "The CoNLL-2014 shared task on grammatical error correction", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Mei", |
| "middle": [], |
| "last": "Siew", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "Raymond", |
| "middle": [ |
| "Hendy" |
| ], |
| "last": "Hadiwinoto", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Susanto", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bryant", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 18th Conference on Computational Natural Language Learning: Shared Task", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hwee Tou Ng, Siew Mei Wu, Ted Briscoe, Christian Hadiwinoto, Raymond Hendy Susanto, and Christo- pher Bryant. 2014. The CoNLL-2014 shared task on grammatical error correction. In Proceedings of the 18th Conference on Computational Natural Lan- guage Learning: Shared Task.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "The CoNLL-2013 shared task on grammatical error correction", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Mei", |
| "middle": [], |
| "last": "Siew", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuanbin", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Joel", |
| "middle": [], |
| "last": "Hadiwinoto", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Tetreault", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 17th Conference on Computational Natural Language Learning: Shared Task", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hwee Tou Ng, Siew Mei Wu, Yuanbin Wu, Christian Hadiwinoto, and Joel Tetreault. 2013. The CoNLL- 2013 shared task on grammatical error correction. In Proceedings of the 17th Conference on Computa- tional Natural Language Learning: Shared Task.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Neural grammatical error correction with finite state transducers", |
| "authors": [ |
| { |
| "first": "Felix", |
| "middle": [], |
| "last": "Stahlberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Bryant", |
| "suffix": "" |
| }, |
| { |
| "first": "Bill", |
| "middle": [], |
| "last": "Byrne", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Felix Stahlberg, Christopher Bryant, and Bill Byrne. 2019. Neural grammatical error correction with fi- nite state transducers. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Rethinking the Inception architecture for computer vision", |
| "authors": [ |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Szegedy", |
| "suffix": "" |
| }, |
| { |
| "first": "Vincent", |
| "middle": [], |
| "last": "Vanhoucke", |
| "suffix": "" |
| }, |
| { |
| "first": "Sergey", |
| "middle": [], |
| "last": "Ioffe", |
| "suffix": "" |
| }, |
| { |
| "first": "Jon", |
| "middle": [], |
| "last": "Shlens", |
| "suffix": "" |
| }, |
| { |
| "first": "Zbigniew", |
| "middle": [], |
| "last": "Wojna", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. Rethinking the Inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Neural machine translation with extended context", |
| "authors": [ |
| { |
| "first": "J\u00f6rg", |
| "middle": [], |
| "last": "Tiedemann", |
| "suffix": "" |
| }, |
| { |
| "first": "Yves", |
| "middle": [], |
| "last": "Scherrer", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the Third Workshop on Discourse in Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J\u00f6rg Tiedemann and Yves Scherrer. 2017. Neural ma- chine translation with extended context. In Proceed- ings of the Third Workshop on Discourse in Machine Translation.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Learning to remember translation history with a continuous cache", |
| "authors": [ |
| { |
| "first": "Zhaopeng", |
| "middle": [], |
| "last": "Tu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Shuming", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "" |
| }, |
| { |
| "first": "Tong", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Transactions of the Association for Computational Linguistics", |
| "volume": "6", |
| "issue": "", |
| "pages": "407--420", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhaopeng Tu, Yang Liu, Shuming Shi, and Tong Zhang. 2018. Learning to remember translation his- tory with a continuous cache. Transactions of the Association for Computational Linguistics, 6:407- 420.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Attention is all you need", |
| "authors": [ |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| }, |
| { |
| "first": "Llion", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Aidan", |
| "middle": [ |
| "N" |
| ], |
| "last": "Gomez", |
| "suffix": "" |
| }, |
| { |
| "first": "\u0141ukasz", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "Illia", |
| "middle": [], |
| "last": "Polosukhin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Advances in Neural Information Processing Systems", |
| "volume": "30", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Pro- cessing Systems 30.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Context-aware neural machine translation learns anaphora resolution", |
| "authors": [ |
| { |
| "first": "Elena", |
| "middle": [], |
| "last": "Voita", |
| "suffix": "" |
| }, |
| { |
| "first": "Pavel", |
| "middle": [], |
| "last": "Serdyukov", |
| "suffix": "" |
| }, |
| { |
| "first": "Rico", |
| "middle": [], |
| "last": "Sennrich", |
| "suffix": "" |
| }, |
| { |
| "first": "Ivan", |
| "middle": [], |
| "last": "Titov", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Elena Voita, Pavel Serdyukov, Rico Sennrich, and Ivan Titov. 2018. Context-aware neural machine trans- lation learns anaphora resolution. In Proceedings of the 56th Annual Meeting of the Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Exploiting cross-sentence context for neural machine translation", |
| "authors": [ |
| { |
| "first": "Longyue", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhaopeng", |
| "middle": [], |
| "last": "Tu", |
| "suffix": "" |
| }, |
| { |
| "first": "Andy", |
| "middle": [], |
| "last": "Way", |
| "suffix": "" |
| }, |
| { |
| "first": "Qun", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Longyue Wang, Zhaopeng Tu, Andy Way, and Qun Liu. 2017. Exploiting cross-sentence context for neural machine translation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Noising and denoising natural language: Diverse backtranslation for grammar correction", |
| "authors": [ |
| { |
| "first": "Ziang", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Guillaume", |
| "middle": [], |
| "last": "Genthial", |
| "suffix": "" |
| }, |
| { |
| "first": "Stanley", |
| "middle": [], |
| "last": "Xie", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Jurafsky", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziang Xie, Guillaume Genthial, Stanley Xie, Andrew Ng, and Dan Jurafsky. 2018. Noising and denoising natural language: Diverse backtranslation for gram- mar correction. In Proceedings of the 2018 Confer- ence of the North American Chapter of the Associ- ation for Computational Linguistics: Human Lan- guage Technologies.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "A new dataset and method for automatically grading ESOL texts", |
| "authors": [ |
| { |
| "first": "Helen", |
| "middle": [], |
| "last": "Yannakoudakis", |
| "suffix": "" |
| }, |
| { |
| "first": "Ted", |
| "middle": [], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "Ben", |
| "middle": [], |
| "last": "Medlock", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helen Yannakoudakis, Ted Briscoe, and Ben Medlock. 2011. A new dataset and method for automatically grading ESOL texts. In Proceedings of the 49th An- nual Meeting of the Association for Computational Linguistics: Human Language Technologies.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "Improving the Transformer translation model with document-level context", |
| "authors": [ |
| { |
| "first": "Jiacheng", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Huanbo", |
| "middle": [], |
| "last": "Luan", |
| "suffix": "" |
| }, |
| { |
| "first": "Maosong", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Feifei", |
| "middle": [], |
| "last": "Zhai", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jiacheng Zhang, Huanbo Luan, Maosong Sun, and Feifei Zhai. 2018. Improving the Transformer trans- lation model with document-level context. In Pro- ceedings of the 2018 Conference on Empirical Meth- ods in Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Improving grammatical error correction via pre-training a copy-augmented architecture with unlabeled data", |
| "authors": [ |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Liang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kewei", |
| "middle": [], |
| "last": "Shen", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruoyu", |
| "middle": [], |
| "last": "Jia", |
| "suffix": "" |
| }, |
| { |
| "first": "Jingming", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wei Zhao, Liang Wang, Kewei Shen, Ruoyu Jia, and Jingming Liu. 2019. Improving grammatical error correction via pre-training a copy-augmented archi- tecture with unlabeled data. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Hu- man Language Technologies.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "How do you correct run-on sentences its not as easy as it seems", |
| "authors": [ |
| { |
| "first": "Junchao", |
| "middle": [], |
| "last": "Zheng", |
| "suffix": "" |
| }, |
| { |
| "first": "Courtney", |
| "middle": [], |
| "last": "Napoles", |
| "suffix": "" |
| }, |
| { |
| "first": "Joel", |
| "middle": [], |
| "last": "Tetreault", |
| "suffix": "" |
| }, |
| { |
| "first": "Kostiantyn", |
| "middle": [], |
| "last": "Omelianchuk", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 2018 EMNLP Workshop W-NUT: The 4th Workshop on Noisy User-generated Text", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Junchao Zheng, Courtney Napoles, Joel Tetreault, and Kostiantyn Omelianchuk. 2018. How do you cor- rect run-on sentences its not as easy as it seems. In Proceedings of the 2018 EMNLP Workshop W-NUT: The 4th Workshop on Noisy User-generated Text.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "A sentence from NUCLE appears correct out of context, but erroneous in context.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF1": { |
| "text": "Our cross-sentence convolutional encoderdecoder model with auxiliary encoder and gating.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF2": { |
| "text": "Output showing a cross-sentence correction.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "FIGREF3": { |
| "text": "Performance on common error types on the CoNLL-2013 test set.", |
| "num": null, |
| "type_str": "figure", |
| "uris": null |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "Statistics of training and development datasets." |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td/><td colspan=\"2\">CoNLL-2013</td><td/><td colspan=\"2\">CoNLL-2014</td><td/><td>FCE</td><td/></tr><tr><td>P</td><td>R</td><td>F 0.5</td><td>P</td><td>R</td><td>F 0.5</td><td>P</td><td>R</td><td>F 0.5</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "BASELINE (avg) 54.51 15.16 35.88 65.16 27.13 50.88 52.89 26.85 44.29 CROSENT (avg) 55.65 16.93 38.17 65.59 30.07 53.06 52.17 28.25 44.61 BASELINE (ens) 55.67 15.02 36.12 66.45 27.12 51.51 53.71 26.79 44.72 CROSENT (ens) 58.72 16.64 38.99 68.06 29.94 54.25 54.49 28.63 46.15 BASELINE (ens+rs) 51.88 19.15 38.66 62.53 33.62 53.35 52.01 31.19 45.89 CROSENT (ens+rs) 54.43 20.22 40.67 64.15 35.26 55.12 52.93 32.81 47.15 BASELINE (ens+rs BERT ) 52.77 19.01 38.93 64.08 34.21 54.55 53.47 30.91 46.65 CROSENT (ens+rs BERT ) 55.24 20.71 41.43 64.32 35.98 55.57 53.91 32.81 47.77" |
| }, |
| "TABREF4": { |
| "content": "<table><tr><td>: Comparison to the best published results on</td></tr><tr><td>the CoNLL-2014 test set. indicates controlled repli-</td></tr><tr><td>cation under identical setup as ours. All CROSENT re-</td></tr><tr><td>sults are statistically significant compared to the corre-</td></tr><tr><td>sponding baselines (p < 0.001).</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "" |
| }, |
| "TABREF5": { |
| "content": "<table><tr><td>Dev</td><td/><td colspan=\"2\">CoNLL-2013</td></tr><tr><td>F 0.5</td><td>P</td><td>R</td><td>F 0.5</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "). BASELINE 33.21 54.51 15.16 35.88 concat 33.41 55.14 15.28 36.23 aux (no gate) 32.99 55.10 14.83 35.69 aux (+gate) 35.68 55.65 16.93 38.17" |
| }, |
| "TABREF6": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "Average single model results comparing different strategies to model cross-sentence context. 'aux (+gate)' is used in our CROSENT model." |
| }, |
| "TABREF7": { |
| "content": "<table><tr><td>sentence context for correction 7 . An example is</td></tr><tr><td>shown below:</td></tr><tr><td>CONTEXT:</td></tr><tr><td>He got a flat tyre, and had to drive slowly back</td></tr><tr><td>to the pits. He finished the race in fifth place.</td></tr><tr><td>ORIGINAL:</td></tr><tr><td>His championship lead was reduced .</td></tr><tr><td>CORRUPTED:</td></tr><tr><td>His championship lead is reduced.</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "Performance on contextual verb tense errors on a synthetic test set. n c denotes the no. of correct changes and n p denotes the no. of proposed changes." |
| }, |
| "TABREF9": { |
| "content": "<table><tr><td>: Top 5 part-of-speech (POS) tags based on av-</td></tr><tr><td>erage KL-divergence between auxiliary attention and</td></tr><tr><td>uniform distribution.</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "html": null, |
| "text": "" |
| } |
| } |
| } |
| } |