| { |
| "paper_id": "K16-2009", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T07:11:00.832567Z" |
| }, |
| "title": "SoNLP-DP System for ConLL-2016 English Shallow Discourse Parsing", |
| "authors": [ |
| { |
| "first": "Fang", |
| "middle": [], |
| "last": "Kong", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Soochow University", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "kongfang@suda.edu.cn" |
| }, |
| { |
| "first": "Sheng", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Alibaba Inc", |
| "location": { |
| "settlement": "Hangzhou", |
| "country": "China" |
| } |
| }, |
| "email": "lisheng.ls@alibaba-inc.com" |
| }, |
| { |
| "first": "Junhui", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Soochow University", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "lijunhui@suda.edu.cn" |
| }, |
| { |
| "first": "Muhua", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Alibaba Inc", |
| "location": { |
| "settlement": "Hangzhou", |
| "country": "China" |
| } |
| }, |
| "email": "muhua.zmh@alibaba-inc.com" |
| }, |
| { |
| "first": "Guodong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Soochow University", |
| "location": { |
| "country": "China" |
| } |
| }, |
| "email": "gdzhou@suda.edu.cn" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper describes the submitted English shallow discourse parsing system from the natural language processing (NLP) group of Soochow university (SoNLP-DP) to the CoNLL-2016 shared task. Our System classifies discourse relations into explicit and non-explicit relations and uses a pipeline platform to conduct every subtask to form an end-to-end shallow discourse parser in the Penn Discourse Treebank (PDTB). Our system is evaluated on the CoNLL-2016 Shared Task closed track and achieves the 24.31% and 28.78% in F1-measure on the official blind test set and test set, respectively.", |
| "pdf_parse": { |
| "paper_id": "K16-2009", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper describes the submitted English shallow discourse parsing system from the natural language processing (NLP) group of Soochow university (SoNLP-DP) to the CoNLL-2016 shared task. Our System classifies discourse relations into explicit and non-explicit relations and uses a pipeline platform to conduct every subtask to form an end-to-end shallow discourse parser in the Penn Discourse Treebank (PDTB). Our system is evaluated on the CoNLL-2016 Shared Task closed track and achieves the 24.31% and 28.78% in F1-measure on the official blind test set and test set, respectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Discourse parsing determines the internal structure of a text via identifying the discourse relations between its text units and plays an important role in natural language understanding that benefits a wide range of downstream natural language applications, such as coherence modeling (Barzilay and Lapata, 2005; Lin et al., 2011 ), text summarization (Lin et al., 2012) , and statistical machine translation (Meyer and Webber, 2013) .", |
| "cite_spans": [ |
| { |
| "start": 286, |
| "end": 313, |
| "text": "(Barzilay and Lapata, 2005;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 314, |
| "end": 330, |
| "text": "Lin et al., 2011", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 353, |
| "end": 371, |
| "text": "(Lin et al., 2012)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 410, |
| "end": 434, |
| "text": "(Meyer and Webber, 2013)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "As the largest discourse corpus, the Penn Discourse TreeBank (PDTB) corpus (Prasad et al., 2008) adds a layer of discourse annotations on the top of the Penn TreeBank (PTB) corpus (Marcus et al., 1993) and has been attracting more and more attention recently (Elwell and Baldridge, 2008; Pitler and Nenkova, 2009; Prasad et al., 2010; Ghosh et al., 2011; Kong et al., 2014; Lin et al., 2014) . Different from another famous discourse corpus, the Rhetorical Structure Theory(RST) Treebank corpus (Carlson et al., 2001) , the PDTB focuses on shallow discourse relations either lexically grounded in explicit discourse connectives or associated with sentential adjacency. This theory-neutral way makes no commitment to any kind of higher-level discourse structure and can work jointly with high-level topic and functional structuring (Webber et al., 2012) or hierarchial structuring (Asher and Lascarides, 2003) .", |
| "cite_spans": [ |
| { |
| "start": 75, |
| "end": 96, |
| "text": "(Prasad et al., 2008)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 180, |
| "end": 201, |
| "text": "(Marcus et al., 1993)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 259, |
| "end": 287, |
| "text": "(Elwell and Baldridge, 2008;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 288, |
| "end": 313, |
| "text": "Pitler and Nenkova, 2009;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 314, |
| "end": 334, |
| "text": "Prasad et al., 2010;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 335, |
| "end": 354, |
| "text": "Ghosh et al., 2011;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 355, |
| "end": 373, |
| "text": "Kong et al., 2014;", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 374, |
| "end": 391, |
| "text": "Lin et al., 2014)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 495, |
| "end": 517, |
| "text": "(Carlson et al., 2001)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 831, |
| "end": 852, |
| "text": "(Webber et al., 2012)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 880, |
| "end": 908, |
| "text": "(Asher and Lascarides, 2003)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Although much research work has been conducted for certain subtasks since the release of the PDTB corpus, there is still little work on constructing an end-to-end shallow discourse parser. The CoNLL 2016 shared task evaluates endto-end shallow discourse parsing systems for determining and classifying both explicit and nonexplicit discourse relations. A participant system needs to (1)locate all explicit (e.g., \"because\", \"however\", \"and\".) discourse connectives in the text, (2)identify the spans of text that serve as the two arguments for each discourse connective, and (3) predict the sense of the discourse relations (e.g., \"Cause\", \"Condition\", \"Contrast\").", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we describe the system submission from the NLP group of Soochow university (SoNLP-DP). Our shallow discourse parser consists of multiple components in a pipeline architecture, including a connective classifier, argument labeler, explicit classifier, non-explicit classifier. Our system is evaluated on the CoNLL-2016 Shared Task closed track and achieves the 24.31% and 28.78% in F1-measure on the official blind test set and test set, respectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The remainder of this paper is organized as follows. Section 2 presents our shallow discourse parsing system. The experimental results are described in Section 3. Section 4 concludes the paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this section, after a quick overview of our system, we describe the details involved in implementing the end-to-end shallow discourse parser.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Architecture", |
| "sec_num": "2" |
| }, |
| { |
| "text": "A typical text consists of sentences glued together in a systematic way to form a coherent discourse. Referring to the PDTB, shallow discourse parsing focus on shallow discourse relations either lexically grounded in explicit discourse connectives or associated with sentential adjacency. Different from full discourse parsing, shallow discourse parsing transforms a piece of text into a set of discourse relations between two adjacent or nonadjacent discourse units, instead of connecting the relations hierarchically to one another to form a connected structure in the form of tree or graph.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Specifically, given a piece of text, the end-toend shallow discourse parser returns a set of discourse relations in the form of a discourse connective (explicit or implicit) taking two arguments (clauses or sentences) with a discourse sense. That is, a complete end-to-end shallow discourse parser includes:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 connective identification, which identifies all connective candidates and labels them as whether they function as discourse connectives or not,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 argument labeling, which identifies the spans of text that serve as the two arguments for each discourse connective,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 explicit sense classification, which predicts the sense of the explicit discourse relations after achieving the connective and its arguments,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 non-explicit sense classification, for all adjacent sentence pairs within each paragraph without explicit discourse relations, which classify the given pair into EntRel, NoRel, or one of the Implicit/AltLex relation senses. Figure 1 shows the components and the relations among them. Different from traditional approach (i.e., Lin et al. (2014) ), considering the interaction between argument labeler and explicit sense classifier, co-occurrence relation between explicit and non-explicit discourse relations in a text, our system does not employ complete sequential pipeline framework. ", |
| "cite_spans": [ |
| { |
| "start": 329, |
| "end": 346, |
| "text": "Lin et al. (2014)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 226, |
| "end": 234, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "System Overview", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Our connective identifier works in two steps. First, the connective candidates are extracted from the given text referring to the PDTB. There are 100 types of discourse connectives defined in the PDT-B. Then every connective candidate is checked whether it functions as a discourse connective. Pitler and Nenkova (2009) showed that syntactic features extracted from constituent parse trees are very useful in disambiguating discourse connectives. Followed their work, Lin et al. (2014) found that a connective's context and part-ofspeech (POS) are also helpful. Motivated by their work, we get a set of effective features, includes:", |
| "cite_spans": [ |
| { |
| "start": 294, |
| "end": 319, |
| "text": "Pitler and Nenkova (2009)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 468, |
| "end": 485, |
| "text": "Lin et al. (2014)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Connective Identification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Lexical: connective itself, POS of the connective, connective with its previous word, connective with its next word, the location of the connective in the sentence, i.e., start, middle and end of the sentence.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Connective Identification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "\u2022 Syntactic: the highest node in the parse tree that covers only the connective words (dominate node), the context of the dominate node 1 , whether the right sibling contains a VP, the path from the parent node of the connective to the root of the parse tree.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Connective Identification", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Argument labeler need to label the Arg1 and Arg2 spans for every connective determined by connective identifier. Following the work of Kong et al. (2014) , we employ the constituent-based approach to argument labeling by first extracting the constituents from a parse tree are casted as argument candidates, then determining the role of every constituent as part of Arg1, Arg2, or NULL, and finally, merging all the constituents for Arg1 and Arg2 to obtain the Arg1 and Arg2 text spans respectively. Note that, we do not use ILP approach to do joint inference.", |
| "cite_spans": [ |
| { |
| "start": 135, |
| "end": 153, |
| "text": "Kong et al. (2014)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "After extracting the argument candidates, a multi-category classifier is employed to determine the role of every argument candidate (i.e., Arg1, Arg2, or NULL) with features reflecting the properties of the connective, the candidate constituent and relationship between them. Features include,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Connective related features: connective itself, its syntactic category, its sense class 2", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 Number of left/right siblings of the connective.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 The context of the constituent. We use POS combination of the constituent, its parent, left sibling and right sibling to represent the context. When there is no parent or siblings, it is marked NULL.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 The path from the parent node of the connective to the node of the constituent.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "\u2022 The position of the constituent relative to the connective: left, right, or previous.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Argument Labeling", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "After a discourse connective and its two arguments are identified, the sense classifier is proved to decide the sense that the relation conveys. Although the same connective may carry different semantics under different contexts, only a few connectives are ambiguous (Pitler and Nenkova, 2009) . Following the work of Lin et al. (2014), we introduce four features to train a sense classifier: the connective itself, its lower format, its POS and the combination of the previous word and the connective.", |
| "cite_spans": [ |
| { |
| "start": 267, |
| "end": 293, |
| "text": "(Pitler and Nenkova, 2009)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Explicit sense classification", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "Referring to the PDTB, the non-explicit relations 3 are annotated for all adjacent sentence pairs within paragraphs. So non-explicit sense classification only considers the sense of every adjacent sentence pair within a paragraph without explicit discourse relations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Our non-explicit sense classifier includes five traditional features:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Production rules: According to Lin et al. (2009) , the syntactic structure of one argument may constrain the relation type and the syntactic structure of the other argument. Three features are introduced to denote the presence of syntactic productions in Arg1, Arg2 or both. Here, these production rules are extracted from the training data and the rules with frequency less than 5 are ignored.", |
| "cite_spans": [ |
| { |
| "start": 31, |
| "end": 48, |
| "text": "Lin et al. (2009)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Dependency rules: Similar with Production rules, three features denoting the presence of dependency productions in Arg1, Arg2 or both are also introduced in our system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Fisrt/Last and First 3 words: This set of features include the first and last words of Arg1, the first and last words of Arg2, the pair of the first words of Arg1 and Arg2, the pair of the last words as features, and the first three words of each argument.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Word pairs: We include the Cartesian product of words in Arg1 and Arg2. We apply MI (Mutual Information) method to select top 500 word pairs. Brown cluster pairs: We include the Cartesian product of the Brown cluster values of the words in Arg1 and Arg2. In our system, we take 3200 Brown clusters provided by CoNLL shared task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "Besides, we notice that not all adjacent sentences contain relation between them. Therfore, we view these adjacent sentences as NoRel relations like the PDTB.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Non-explicit sense Classification", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "We train our system on the corpora provided in the CoNLL-2016 Shared Task and evaluate our system on the CoNLL-2016 Shared Task closed track. All our classifiers are trained using the OpenNLP maximum entropy package 4 with the default pa-rameters (i.e. without smoothing and with 100 iterations). We firstly report the official score on the CoNLL-2016 shared task on development, test and blind test sets. Then, the supplementary results provided by the shared task organizes are reported. (Wang and Lan, 2015) 46.37 91.86 24.00 Table 1 : the official F1 score of our system.", |
| "cite_spans": [ |
| { |
| "start": 490, |
| "end": 510, |
| "text": "(Wang and Lan, 2015)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 529, |
| "end": 536, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In Table 1 , we present the official results of our system performances on the CoNLL-2016 development, test and blind test sets, respectively. In the blind test, our parser achieve a better result than the best system of last year (Wang and Lan, 2015) . Table 2 : the supplementary F1 score of our system.", |
| "cite_spans": [ |
| { |
| "start": 231, |
| "end": 251, |
| "text": "(Wang and Lan, 2015)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 3, |
| "end": 10, |
| "text": "Table 1", |
| "ref_id": null |
| }, |
| { |
| "start": 254, |
| "end": 261, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In Table 2 , we reported the supplementary results provided by the shared task organizes on the development, test and blind test sets. These additional experiments investigate the performance of our shallow discourse parsing for explicit and non-explicit relations separately. From the results, we can find that the sense classification for both explicit and non-explicit discourse relations are the biggest obstacles to the overall performance of discourse parsing.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 3, |
| "end": 10, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Further, we reports all the official performance in Table 3 on the development, test and blind test set in detail. From the table, we observe:", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 52, |
| "end": 59, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 For argument recognition of explicit discourse relations, the performance of Arg2 is much better than that of Arg1 on all the three datasets. So the performance of Arg1 & Arg2 recognition mainly depends on the performance of Arg1 recognition. With respect to non-explicit discourse relations, the performance gap of argument recognition on Arg1 and Arg2 is very small.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "\u2022 With respect to explicit discourse relations, the sense classification works almost perfectly on development data. It also works well on the test and blind test sets. With respect to non-explicit discourse relations, the sense classification works much worse than that of explicit sense classification. The performance gap caused by non-explicit sense classification reaches 15% 16%.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimentation", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We have presented the SoNLP-DP system from the NLP group of Soochow university that participated in the CoNLL-2016 shared task. Table 3 : Official results (%) of our parser on development, test and blind test sets. Group Explicit indicates the performance with respect to explicit discourse relations; group Non-Explicit indicates the performance with respect to non-explicit discourse relations, and group all indicates the performance with respect to all discourse relations, including both explicit and non-explicit ones.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 128, |
| "end": 135, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We use POS combination of the parent, left sibling and right sibling of the dominate node to represent the context. When no parent or siblings, it is marked NULL.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "In training stage, we extract the gold sense class from the annotated corpus. And in testing stage, the sense classification will be employed to get the automatic sense.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The PDTB provides annotation for Implicit relations, Al-tLex relations, entity transition (EntRel), and otherwise no relation (NoRel), which are lumped together as Non-Explicit relations.4 http://maxent.sourceforge.net/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This research is supported by Key project 61333018 and 61331011 under the National Natural Science Foundation of China, Project 61472264, 61401295, 61305088 and 61402314 under the National Natural Science Foundation of China.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Logics of Conversation", |
| "authors": [ |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Asher", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Lascarides", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicholas Asher and Alex Lascarides. 2003. Logics of Conversation. Cambridge Unversity Press.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Modeling local coherence: An entity-based approach", |
| "authors": [ |
| { |
| "first": "Regina", |
| "middle": [], |
| "last": "Barzilay", |
| "suffix": "" |
| }, |
| { |
| "first": "Mirella", |
| "middle": [], |
| "last": "Lapata", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Regina Barzilay and Mirella Lapata. 2005. Model- ing local coherence: An entity-based approach. In Proceedings of the 43rd Annual Meeting of the As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Building a discourse-tagged corpus in the framework of rhetorical structure theory", |
| "authors": [ |
| { |
| "first": "Lynn", |
| "middle": [], |
| "last": "Carlson", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| }, |
| { |
| "first": "Mary", |
| "middle": [ |
| "Ellen" |
| ], |
| "last": "Okurowski", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of Second SIGdial Workshop on Discourse and Dialogue", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lynn Carlson, Daniel Marcu, and Mary Ellen Okurowski. 2001. Building a discourse-tagged cor- pus in the framework of rhetorical structure theory. In Proceedings of Second SIGdial Workshop on Dis- course and Dialogue.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Discourse connective argument identification with connective specific rankers", |
| "authors": [ |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Elwell", |
| "suffix": "" |
| }, |
| { |
| "first": "Jason", |
| "middle": [], |
| "last": "Baldridge", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Second IEEE International Conference on Semantic Computing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Robert Elwell and Jason Baldridge. 2008. Discourse connective argument identification with connective specific rankers. In Second IEEE International Con- ference on Semantic Computing.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Shallow discourse parsing with conditional random fields", |
| "authors": [ |
| { |
| "first": "Sucheta", |
| "middle": [], |
| "last": "Ghosh", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Johansson", |
| "suffix": "" |
| }, |
| { |
| "first": "Giuseppe", |
| "middle": [], |
| "last": "Riccardi", |
| "suffix": "" |
| }, |
| { |
| "first": "Sara", |
| "middle": [], |
| "last": "Tonelli", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 5th International Joint Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sucheta Ghosh, Richard Johansson, Giuseppe Riccar- di, and Sara Tonelli. 2011. Shallow discourse pars- ing with conditional random fields. In Proceedings of the 5th International Joint Conference on Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "A constituent-based approach to argument labeling with joint inference in discourse parsing", |
| "authors": [ |
| { |
| "first": "Fang", |
| "middle": [], |
| "last": "Kong", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Guodong", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fang Kong, Hwee Tou Ng, and Guodong Zhou. 2014. A constituent-based approach to argument labeling with joint inference in discourse parsing. In Pro- ceedings of the 2014 Conference on Empirical Meth- ods in Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Recognizing implicit discourse relations in the Penn Discourse Treebank", |
| "authors": [ |
| { |
| "first": "Ziheng", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng. 2009. Recognizing implicit discourse relations in the Pen- n Discourse Treebank. In Proceedings of the 2009 Conference on Empirical Methods in Natural Lan- guage Processing.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Automatically evaluating text coherence using discourse relations", |
| "authors": [ |
| { |
| "first": "Ziheng", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziheng Lin, Hwee Tou Ng, and Min-Yen Kan. 2011. Automatically evaluating text coherence using dis- course relations. In Proceedings of the 49th Annual Meeting of the Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Combining coherence models and machine translation evaluation metrics for summarization evaluation", |
| "authors": [ |
| { |
| "first": "Ziheng", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Chang", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Hwee Tou", |
| "middle": [], |
| "last": "Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziheng Lin, Chang Liu, Hwee Tou Ng, and Min-Yen Kan. 2012. Combining coherence models and ma- chine translation evaluation metrics for summariza- tion evaluation. In Proceedings of the 50th Annual Meeting of the Association for Computational Lin- guistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "A PDTB-styled end-to-end discourse parser", |
| "authors": [ |
| { |
| "first": "Ziheng", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "" |
| }, |
| { |
| "first": "Min-Yen", |
| "middle": [], |
| "last": "Hwee Tou Ng", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kan", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Natural Language Engineering", |
| "volume": "20", |
| "issue": "2", |
| "pages": "151--184", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ziheng Lin, Hwee Tou Ng, and Min-Yen Kan. 2014. A PDTB-styled end-to-end discourse parser. Natural Language Engineering, 20(2):151-184.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Building a large annotated corpus of English: The Penn Treebank", |
| "authors": [ |
| { |
| "first": "Mitchell", |
| "middle": [ |
| "P" |
| ], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "Beatrice", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| }, |
| { |
| "first": "Mary", |
| "middle": [ |
| "Ann" |
| ], |
| "last": "Marcinkiewicz", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "2", |
| "pages": "313--330", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mitchell P. Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1993. Building a large annotated corpus of English: The Penn Treebank. Computa- tional Linguistics, 19(2):313-330.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Implicitation of discourse connectives in (machine) translation", |
| "authors": [ |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Meyer", |
| "suffix": "" |
| }, |
| { |
| "first": "Bonnie", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the Workshop on Discourse in Machine Translation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Thomas Meyer and Bonnie Webber. 2013. Implicita- tion of discourse connectives in (machine) transla- tion. In Proceedings of the Workshop on Discourse in Machine Translation.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Using syntax to disambiguate explicit discourse connectives in text", |
| "authors": [ |
| { |
| "first": "Emily", |
| "middle": [], |
| "last": "Pitler", |
| "suffix": "" |
| }, |
| { |
| "first": "Ani", |
| "middle": [], |
| "last": "Nenkova", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "Proceedings of the ACL-IJCNLP 2009 Conference Short Papers", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emily Pitler and Ani Nenkova. 2009. Using syntax to disambiguate explicit discourse connectives in tex- t. In Proceedings of the ACL-IJCNLP 2009 Confer- ence Short Papers.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "The Penn Discourse TreeBank 2.0", |
| "authors": [ |
| { |
| "first": "Rashmi", |
| "middle": [], |
| "last": "Prasad", |
| "suffix": "" |
| }, |
| { |
| "first": "Nikhil", |
| "middle": [], |
| "last": "Dinesh", |
| "suffix": "" |
| }, |
| { |
| "first": "Alan", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Eleni", |
| "middle": [], |
| "last": "Miltsakaki", |
| "suffix": "" |
| }, |
| { |
| "first": "Livio", |
| "middle": [], |
| "last": "Robaldo", |
| "suffix": "" |
| }, |
| { |
| "first": "Aravind", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Bonnie", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the LREC 2008 Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rashmi Prasad, Nikhil Dinesh, Alan Lee, Eleni Milt- sakaki, Livio Robaldo, Aravind Joshi, and Bonnie Webber. 2008. The Penn Discourse TreeBank 2.0. In Proceedings of the LREC 2008 Conference.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Exploiting scope for shallow discourse parsing", |
| "authors": [ |
| { |
| "first": "Rashmi", |
| "middle": [], |
| "last": "Prasad", |
| "suffix": "" |
| }, |
| { |
| "first": "Aravind", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "Bonnie", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the Seventh International Conference on Language Resources and Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Rashmi Prasad, Aravind Joshi, and Bonnie Webber. 2010. Exploiting scope for shallow discourse pars- ing. In Proceedings of the Seventh International Conference on Language Resources and Evaluation.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A refined endto-end discourse parser", |
| "authors": [ |
| { |
| "first": "Jianxiang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Man", |
| "middle": [], |
| "last": "Lan", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the Nineteenth Conference on Computational Natural Language Learning -Shared Task", |
| "volume": "", |
| "issue": "", |
| "pages": "17--24", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jianxiang Wang and Man Lan. 2015. A refined end- to-end discourse parser. In Proceedings of the Nine- teenth Conference on Computational Natural Lan- guage Learning -Shared Task, pages 17-24, Bei- jing, China, July. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Discourse structure and language technology. Natural Language Engineering", |
| "authors": [ |
| { |
| "first": "Bonnie", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcus", |
| "middle": [], |
| "last": "Egg", |
| "suffix": "" |
| }, |
| { |
| "first": "Valia", |
| "middle": [], |
| "last": "Kordoni", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "", |
| "volume": "18", |
| "issue": "", |
| "pages": "437--490", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bonnie Webber, Marcus Egg, and Valia Kordoni. 2012. Discourse structure and language technology. Natu- ral Language Engineering, 18(4):437-490, 10.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "Framework of our end-to-end shallow discourse parser", |
| "num": null, |
| "uris": null, |
| "type_str": "figure" |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td/><td>Dev</td><td/><td/><td>Test</td><td/><td/><td>Blind test</td><td/></tr><tr><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td></tr><tr><td>Explicit 93.53 Non-Explicit Connective Connective -Arg1 47.35</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td><td>-</td></tr><tr><td/><td/><td/><td/><td/><td/><td/><td/><td>Our sys-</td></tr><tr><td/><td/><td/><td colspan=\"6\">tem is evaluated on the CoNLL-2016 Shared Task</td></tr><tr><td/><td/><td/><td colspan=\"6\">closed track and achieves the 24.31% and 28.78%</td></tr><tr><td/><td/><td/><td colspan=\"6\">in F1-measure on the official blind test set and test</td></tr><tr><td/><td/><td/><td colspan=\"2\">set, respectively.</td><td/><td/><td/><td/></tr></table>", |
| "html": null, |
| "text": "94.93 94.22 94.04 95.38 94.71 90.47 92.80 91.62 Arg1 52.50 53.28 52.89 46.80 47.47 47.14 46.58 47.79 47.18 Arg2 74.26 75.37 74.81 70.64 71.65 71.14 67.99 69.74 68.85 Arg1 & Arg2 46.03 46.72 46.37 40.52 41.10 40.81 37.77 38.75 38.25 Overall 42.84 43.10 42.97 36.81 36.33 36.57 31.61 30.76 31.18 75.85 58.31 40.55 70.79 51.56 29.71 72.93 42.22 Arg2 48.54 77.75 59.77 40.55 70.79 51.56 31.55 77.44 44.83 Arg1 & Arg2 40.21 64.41 49.51 33.56 58.59 42.68 23.74 58.27 33.73 Overall 35.67 22.43 27.54 27.19 15.60 19.82 27.82 11.33 16.10 All Connective 93.53 94.93 94.22 94.04 95.38 94.71 90.47 92.80 91.62 Arg1 50.35 63.31 56.09 43.89 57.04 49.61 37.55 56.19 45.02 Arg2 60.72 76.36 67.65 54.87 71.31 62.02 48.30 72.28 57.91 Arg1 & Arg2 42.97 54.03 47.87 36.87 47.92 41.68 30.19 45.17 36.19 Overall 39.86 32.10 35.56 33.07 25.48 28.78 30.36 20.26 24.31", |
| "num": null, |
| "type_str": "table" |
| } |
| } |
| } |
| } |