Dataset Viewer
Auto-converted to Parquet Duplicate
venue
stringclasses
1 value
title
stringlengths
18
162
abstract
stringlengths
252
1.89k
doc_id
stringlengths
32
32
publication_year
int64
2.02k
2.02k
sentences
listlengths
1
13
events
listlengths
1
24
document
listlengths
50
348
ACL
Program Transfer for Answering Complex Questions over Knowledge Bases
Program induction for answering complex questions over knowledge bases (KBs) aims to decompose a question into a multi-step program, whose execution against the KB produces the final answer. Learning to induce programs relies on a large number of parallel question-program pairs for the given KB. However, for most KBs, the gold program annotations are usually lacking, making learning difficult. In this paper, we propose the approach of program transfer, which aims to leverage the valuable program annotations on the rich-resourced KBs as external supervision signals to aid program induction for the low-resourced KBs that lack program annotations. For program transfer, we design a novel two-stage parsing framework with an efficient ontology-guided pruning strategy. First, a sketch parser translates the question into a high-level program sketch, which is the composition of functions. Second, given the question and sketch, an argument parser searches the detailed arguments from the KB for functions. During the searching, we incorporate the KB ontology to prune the search space. The experiments on ComplexWebQuestions and WebQuestionSP show that our method outperforms SOTA methods significantly, demonstrating the effectiveness of program transfer and our framework. Our codes and datasets can be obtained from https://github.com/THU-KEG/ProgramTransfer.
febd573ada9568c635f6d8aeada27ec5
2,022
[ "program induction for answering complex questions over knowledge bases ( kbs ) aims to decompose a question into a multi - step program , whose execution against the kb produces the final answer .", "learning to induce programs relies on a large number of parallel question - program pairs for the given kb .", "however , for most kbs , the gold program annotations are usually lacking , making learning difficult .", "in this paper , we propose the approach of program transfer , which aims to leverage the valuable program annotations on the rich - resourced kbs as external supervision signals to aid program induction for the low - resourced kbs that lack program annotations .", "for program transfer , we design a novel two - stage parsing framework with an efficient ontology - guided pruning strategy .", "first , a sketch parser translates the question into a high - level program sketch , which is the composition of functions .", "second , given the question and sketch , an argument parser searches the detailed arguments from the kb for functions .", "during the searching , we incorporate the kb ontology to prune the search space .", "the experiments on complexwebquestions and webquestionsp show that our method outperforms sota methods significantly , demonstrating the effectiveness of program transfer and our framework .", "our codes and datasets can be obtained from https : / / github . com / thu - keg / programtransfer ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1 ], "text": "program induction", "tokens": [ "program", "induction" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 14 ], "text": "decompose", "tokens": [ "decompose" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 61, 62, 63 ], "text": "gold program annotations", "tokens": [ "gold", "program", "annotations" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 66 ], "text": "lacking", "tokens": [ "lacking" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 66 ], "text": "lacking", "tokens": [ "lacking" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 76 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 79, 80, 81, 82 ], "text": "approach of program transfer", "tokens": [ "approach", "of", "program", "transfer" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 77 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 89, 90, 91, 92, 93, 94, 95, 96, 7, 8 ], "text": "valuable program annotations on the rich - resourced kbs", "tokens": [ "valuable", "program", "annotations", "on", "the", "rich", "-", "resourced", "knowledge", "bases" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 103 ], "text": "aid", "tokens": [ "aid" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 87 ], "text": "leverage", "tokens": [ "leverage" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 104, 105 ], "text": "program induction", "tokens": [ "program", "induction" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 103 ], "text": "aid", "tokens": [ "aid" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 121 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 118, 119 ], "text": "program transfer", "tokens": [ "program", "transfer" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 125, 126, 127, 128, 129 ], "text": "two - stage parsing framework", "tokens": [ "two", "-", "stage", "parsing", "framework" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 122 ], "text": "design", "tokens": [ "design" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 146 ], "text": "question", "tokens": [ "question" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 149, 150, 151, 152, 153 ], "text": "high - level program sketch", "tokens": [ "high", "-", "level", "program", "sketch" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 144 ], "text": "translates", "tokens": [ "translates" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 175, 176 ], "text": "detailed arguments", "tokens": [ "detailed", "arguments" ] }, { "argument_type": "BaseComponent", "nugget_type": "DST", "offsets": [ 7, 8 ], "text": "kb", "tokens": [ "knowledge", "bases" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 181 ], "text": "functions", "tokens": [ "functions" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 173 ], "text": "searches", "tokens": [ "searches" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 166 ], "text": "question", "tokens": [ "question" ] }, { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 168 ], "text": "sketch", "tokens": [ "sketch" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 164 ], "text": "given", "tokens": [ "given" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 183, 184, 185 ], "text": "during the searching", "tokens": [ "during", "the", "searching" ] }, { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 187 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 7, 8, 191 ], "text": "kb ontology", "tokens": [ "knowledge", "bases", "ontology" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 193 ], "text": "prune", "tokens": [ "prune" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 188 ], "text": "incorporate", "tokens": [ "incorporate" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 195, 196 ], "text": "search space", "tokens": [ "search", "space" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 193 ], "text": "prune", "tokens": [ "prune" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 208 ], "text": "outperforms", "tokens": [ "outperforms" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 204 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 125, 126, 127, 128, 129 ], "text": "two - stage parsing framework", "tokens": [ "two", "-", "stage", "parsing", "framework" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 208 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 209, 210 ], "text": "sota methods", "tokens": [ "sota", "methods" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 208 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 213 ], "text": "demonstrating", "tokens": [ "demonstrating" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 204 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 215, 216, 217, 218 ], "text": "effectiveness of program transfer", "tokens": [ "effectiveness", "of", "program", "transfer" ] }, { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 215, 216, 125, 126, 127, 128, 129 ], "text": "effectiveness of our framework", "tokens": [ "effectiveness", "of", "two", "-", "stage", "parsing", "framework" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 213 ], "text": "demonstrating", "tokens": [ "demonstrating" ] } } ]
[ "program", "induction", "for", "answering", "complex", "questions", "over", "knowledge", "bases", "(", "kbs", ")", "aims", "to", "decompose", "a", "question", "into", "a", "multi", "-", "step", "program", ",", "whose", "execution", "against", "the", "kb", "produces", "the", "final", "answer", ".", "learning", "to", "induce", "programs", "relies", "on", "a", "large", "number", "of", "parallel", "question", "-", "program", "pairs", "for", "the", "given", "kb", ".", "however", ",", "for", "most", "kbs", ",", "the", "gold", "program", "annotations", "are", "usually", "lacking", ",", "making", "learning", "difficult", ".", "in", "this", "paper", ",", "we", "propose", "the", "approach", "of", "program", "transfer", ",", "which", "aims", "to", "leverage", "the", "valuable", "program", "annotations", "on", "the", "rich", "-", "resourced", "kbs", "as", "external", "supervision", "signals", "to", "aid", "program", "induction", "for", "the", "low", "-", "resourced", "kbs", "that", "lack", "program", "annotations", ".", "for", "program", "transfer", ",", "we", "design", "a", "novel", "two", "-", "stage", "parsing", "framework", "with", "an", "efficient", "ontology", "-", "guided", "pruning", "strategy", ".", "first", ",", "a", "sketch", "parser", "translates", "the", "question", "into", "a", "high", "-", "level", "program", "sketch", ",", "which", "is", "the", "composition", "of", "functions", ".", "second", ",", "given", "the", "question", "and", "sketch", ",", "an", "argument", "parser", "searches", "the", "detailed", "arguments", "from", "the", "kb", "for", "functions", ".", "during", "the", "searching", ",", "we", "incorporate", "the", "kb", "ontology", "to", "prune", "the", "search", "space", ".", "the", "experiments", "on", "complexwebquestions", "and", "webquestionsp", "show", "that", "our", "method", "outperforms", "sota", "methods", "significantly", ",", "demonstrating", "the", "effectiveness", "of", "program", "transfer", "and", "our", "framework", ".", "our", "codes", "and", "datasets", "can", "be", "obtained", "from", "https", ":", "/", "/", "github", ".", "com", "/", "thu", "-", "keg", "/", "programtransfer", "." ]
ACL
Are Natural Language Inference Models IMPPRESsive? Learning IMPlicature and PRESupposition
Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by “some” as entailments. For some presupposition triggers like “only”, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.
1a6285faf0918175c1ea9e0b7c8ea82e
2,020
[ "natural language inference ( nli ) is an increasingly important task for natural language understanding , which requires one to infer whether a sentence entails another .", "however , the ability of nli models to make pragmatic inferences remains understudied .", "we create an implicature and presupposition diagnostic dataset ( imppres ) , consisting of 32k semi - automatically generated sentence pairs illustrating well - studied pragmatic inference types .", "we use imppres to evaluate whether bert , infersent , and bow nli models trained on multinli ( williams et al . , 2018 ) learn to make pragmatic inferences .", "although multinli appears to contain very few pairs illustrating these inference types , we find that bert learns to draw pragmatic inferences .", "it reliably treats scalar implicatures triggered by “ some ” as entailments .", "for some presupposition triggers like “ only ” , bert reliably recognizes the presupposition as an entailment , even when the trigger is embedded under an entailment canceling operator like negation .", "bow and infersent show weaker evidence of pragmatic reasoning .", "we conclude that nli training encourages models to learn some , but not all , pragmatic inferences ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1, 2 ], "text": "natural language inference", "tokens": [ "natural", "language", "inference" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 10 ], "text": "task", "tokens": [ "task" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 41 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 44, 45, 46, 47, 48 ], "text": "implicature and presupposition diagnostic dataset", "tokens": [ "implicature", "and", "presupposition", "diagnostic", "dataset" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 42 ], "text": "create", "tokens": [ "create" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 70 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 76 ], "text": "bert", "tokens": [ "bert" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 78 ], "text": "infersent", "tokens": [ "infersent" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 81, 0, 1, 2, 83 ], "text": "bow nli models", "tokens": [ "bow", "natural", "language", "inference", "models" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 84, 85, 86 ], "text": "trained on multinli", "tokens": [ "trained", "on", "multinli" ] }, { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 44, 45, 46, 47, 48 ], "text": "implicature and presupposition diagnostic dataset", "tokens": [ "implicature", "and", "presupposition", "diagnostic", "dataset" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 74 ], "text": "evaluate", "tokens": [ "evaluate" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 114 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 118 ], "text": "learns", "tokens": [ "learns" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 125, 126 ], "text": "reliably treats", "tokens": [ "reliably", "treats" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 147, 148 ], "text": "reliably recognizes", "tokens": [ "reliably", "recognizes" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 115 ], "text": "find", "tokens": [ "find" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 117 ], "text": "bert", "tokens": [ "bert" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 120 ], "text": "draw", "tokens": [ "draw" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 118 ], "text": "learns", "tokens": [ "learns" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 121, 122 ], "text": "pragmatic inferences", "tokens": [ "pragmatic", "inferences" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 120 ], "text": "draw", "tokens": [ "draw" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 169 ], "text": "bow", "tokens": [ "bow" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 171 ], "text": "infersent", "tokens": [ "infersent" ] }, { "argument_type": "Object", "nugget_type": "WEA", "offsets": [ 173, 174, 175, 176, 177 ], "text": "weaker evidence of pragmatic reasoning", "tokens": [ "weaker", "evidence", "of", "pragmatic", "reasoning" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 172 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 179 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 184 ], "text": "encourages", "tokens": [ "encourages" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 180 ], "text": "conclude", "tokens": [ "conclude" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "MOD", "offsets": [ 0, 1, 2, 183 ], "text": "nli training", "tokens": [ "natural", "language", "inference", "training" ] }, { "argument_type": "Object", "nugget_type": "APP", "offsets": [ 185 ], "text": "models", "tokens": [ "models" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 187 ], "text": "learn", "tokens": [ "learn" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 184 ], "text": "encourages", "tokens": [ "encourages" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 188, 189, 190, 191, 192, 193, 194, 195 ], "text": "some , but not all , pragmatic inferences", "tokens": [ "some", ",", "but", "not", "all", ",", "pragmatic", "inferences" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 187 ], "text": "learn", "tokens": [ "learn" ] } }, { "arguments": [ { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 39 ], "text": "understudied", "tokens": [ "understudied" ] }, { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 0, 1, 2, 33 ], "text": "nli models", "tokens": [ "natural", "language", "inference", "models" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 38 ], "text": "remains", "tokens": [ "remains" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 117 ], "text": "bert", "tokens": [ "bert" ] }, { "argument_type": "Object", "nugget_type": "FEA", "offsets": [ 127, 128, 129, 130, 131, 132, 133 ], "text": "scalar implicatures triggered by “ some ”", "tokens": [ "scalar", "implicatures", "triggered", "by", "“", "some", "”" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 135 ], "text": "entailments", "tokens": [ "entailments" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 125, 126 ], "text": "reliably treats", "tokens": [ "reliably", "treats" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 137, 138, 139, 140, 141, 142, 143, 144 ], "text": "for some presupposition triggers like “ only ”", "tokens": [ "for", "some", "presupposition", "triggers", "like", "“", "only", "”" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 146 ], "text": "bert", "tokens": [ "bert" ] }, { "argument_type": "Object", "nugget_type": "FEA", "offsets": [ 150 ], "text": "presupposition", "tokens": [ "presupposition" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 153 ], "text": "entailment", "tokens": [ "entailment" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 147, 148 ], "text": "reliably recognizes", "tokens": [ "reliably", "recognizes" ] } } ]
[ "natural", "language", "inference", "(", "nli", ")", "is", "an", "increasingly", "important", "task", "for", "natural", "language", "understanding", ",", "which", "requires", "one", "to", "infer", "whether", "a", "sentence", "entails", "another", ".", "however", ",", "the", "ability", "of", "nli", "models", "to", "make", "pragmatic", "inferences", "remains", "understudied", ".", "we", "create", "an", "implicature", "and", "presupposition", "diagnostic", "dataset", "(", "imppres", ")", ",", "consisting", "of", "32k", "semi", "-", "automatically", "generated", "sentence", "pairs", "illustrating", "well", "-", "studied", "pragmatic", "inference", "types", ".", "we", "use", "imppres", "to", "evaluate", "whether", "bert", ",", "infersent", ",", "and", "bow", "nli", "models", "trained", "on", "multinli", "(", "williams", "et", "al", ".", ",", "2018", ")", "learn", "to", "make", "pragmatic", "inferences", ".", "although", "multinli", "appears", "to", "contain", "very", "few", "pairs", "illustrating", "these", "inference", "types", ",", "we", "find", "that", "bert", "learns", "to", "draw", "pragmatic", "inferences", ".", "it", "reliably", "treats", "scalar", "implicatures", "triggered", "by", "“", "some", "”", "as", "entailments", ".", "for", "some", "presupposition", "triggers", "like", "“", "only", "”", ",", "bert", "reliably", "recognizes", "the", "presupposition", "as", "an", "entailment", ",", "even", "when", "the", "trigger", "is", "embedded", "under", "an", "entailment", "canceling", "operator", "like", "negation", ".", "bow", "and", "infersent", "show", "weaker", "evidence", "of", "pragmatic", "reasoning", ".", "we", "conclude", "that", "nli", "training", "encourages", "models", "to", "learn", "some", ",", "but", "not", "all", ",", "pragmatic", "inferences", "." ]
ACL
Leveraging Task Transferability to Meta-learning for Clinical Section Classification with Limited Data
Identifying sections is one of the critical components of understanding medical information from unstructured clinical notes and developing assistive technologies for clinical note-writing tasks. Most state-of-the-art text classification systems require thousands of in-domain text data to achieve high performance. However, collecting in-domain and recent clinical note data with section labels is challenging given the high level of privacy and sensitivity. The present paper proposes an algorithmic way to improve the task transferability of meta-learning-based text classification in order to address the issue of low-resource target data. Specifically, we explore how to make the best use of the source dataset and propose a unique task transferability measure named Normalized Negative Conditional Entropy (NNCE). Leveraging the NNCE, we develop strategies for selecting clinical categories and sections from source task data to boost cross-domain meta-learning accuracy. Experimental results show that our task selection strategies improve section classification accuracy significantly compared to meta-learning algorithms.
6bab1cf097070e6d457c9c8fd0e74e57
2,022
[ "identifying sections is one of the critical components of understanding medical information from unstructured clinical notes and developing assistive technologies for clinical note - writing tasks .", "most state - of - the - art text classification systems require thousands of in - domain text data to achieve high performance .", "however , collecting in - domain and recent clinical note data with section labels is challenging given the high level of privacy and sensitivity .", "the present paper proposes an algorithmic way to improve the task transferability of meta - learning - based text classification in order to address the issue of low - resource target data .", "specifically , we explore how to make the best use of the source dataset and propose a unique task transferability measure named normalized negative conditional entropy ( nnce ) .", "leveraging the nnce , we develop strategies for selecting clinical categories and sections from source task data to boost cross - domain meta - learning accuracy .", "experimental results show that our task selection strategies improve section classification accuracy significantly compared to meta - learning algorithms ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 28, 29, 30, 31, 32, 33, 34, 35, 36, 37 ], "text": "state - of - the - art text classification systems", "tokens": [ "state", "-", "of", "-", "the", "-", "art", "text", "classification", "systems" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 38 ], "text": "require", "tokens": [ "require" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 81, 82 ], "text": "algorithmic way", "tokens": [ "algorithmic", "way" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 79 ], "text": "proposes", "tokens": [ "proposes" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 86, 87, 88, 89, 90, 91, 92, 93, 94, 95 ], "text": "task transferability of meta - learning - based text classification", "tokens": [ "task", "transferability", "of", "meta", "-", "learning", "-", "based", "text", "classification" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 99 ], "text": "address", "tokens": [ "address" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 84 ], "text": "improve", "tokens": [ "improve" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 101, 102, 103, 104, 105, 106, 107 ], "text": "issue of low - resource target data", "tokens": [ "issue", "of", "low", "-", "resource", "target", "data" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 99 ], "text": "address", "tokens": [ "address" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 111 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 121, 122 ], "text": "source dataset", "tokens": [ "source", "dataset" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 115, 116, 117, 118, 119 ], "text": "make the best use of", "tokens": [ "make", "the", "best", "use", "of" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 111 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 131, 132, 133, 134 ], "text": "normalized negative conditional entropy", "tokens": [ "normalized", "negative", "conditional", "entropy" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 124 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 143 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 145 ], "text": "strategies", "tokens": [ "strategies" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 157 ], "text": "boost", "tokens": [ "boost" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 144 ], "text": "develop", "tokens": [ "develop" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 158, 159, 160, 161, 162, 163, 164 ], "text": "cross - domain meta - learning accuracy", "tokens": [ "cross", "-", "domain", "meta", "-", "learning", "accuracy" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 157 ], "text": "boost", "tokens": [ "boost" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 174 ], "text": "improve", "tokens": [ "improve" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 168 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 174 ], "text": "improve", "tokens": [ "improve" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 175, 176, 177 ], "text": "section classification accuracy", "tokens": [ "section", "classification", "accuracy" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 178 ], "text": "significantly", "tokens": [ "significantly" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 174 ], "text": "improve", "tokens": [ "improve" ] } } ]
[ "identifying", "sections", "is", "one", "of", "the", "critical", "components", "of", "understanding", "medical", "information", "from", "unstructured", "clinical", "notes", "and", "developing", "assistive", "technologies", "for", "clinical", "note", "-", "writing", "tasks", ".", "most", "state", "-", "of", "-", "the", "-", "art", "text", "classification", "systems", "require", "thousands", "of", "in", "-", "domain", "text", "data", "to", "achieve", "high", "performance", ".", "however", ",", "collecting", "in", "-", "domain", "and", "recent", "clinical", "note", "data", "with", "section", "labels", "is", "challenging", "given", "the", "high", "level", "of", "privacy", "and", "sensitivity", ".", "the", "present", "paper", "proposes", "an", "algorithmic", "way", "to", "improve", "the", "task", "transferability", "of", "meta", "-", "learning", "-", "based", "text", "classification", "in", "order", "to", "address", "the", "issue", "of", "low", "-", "resource", "target", "data", ".", "specifically", ",", "we", "explore", "how", "to", "make", "the", "best", "use", "of", "the", "source", "dataset", "and", "propose", "a", "unique", "task", "transferability", "measure", "named", "normalized", "negative", "conditional", "entropy", "(", "nnce", ")", ".", "leveraging", "the", "nnce", ",", "we", "develop", "strategies", "for", "selecting", "clinical", "categories", "and", "sections", "from", "source", "task", "data", "to", "boost", "cross", "-", "domain", "meta", "-", "learning", "accuracy", ".", "experimental", "results", "show", "that", "our", "task", "selection", "strategies", "improve", "section", "classification", "accuracy", "significantly", "compared", "to", "meta", "-", "learning", "algorithms", "." ]
ACL
Generate, Delete and Rewrite: A Three-Stage Framework for Improving Persona Consistency of Dialogue Generation
Maintaining a consistent personality in conversations is quite natural for human beings, but is still a non-trivial task for machines. The persona-based dialogue generation task is thus introduced to tackle the personality-inconsistent problem by incorporating explicit persona text into dialogue generation models. Despite the success of existing persona-based models on generating human-like responses, their one-stage decoding framework can hardly avoid the generation of inconsistent persona words. In this work, we introduce a three-stage framework that employs a generate-delete-rewrite mechanism to delete inconsistent words from a generated response prototype and further rewrite it to a personality-consistent one. We carry out evaluations by both human and automatic metrics. Experiments on the Persona-Chat dataset show that our approach achieves good performance.
71ff0f02bc14a28822f0cdf6c508aae2
2,020
[ "maintaining a consistent personality in conversations is quite natural for human beings , but is still a non - trivial task for machines .", "the persona - based dialogue generation task is thus introduced to tackle the personality - inconsistent problem by incorporating explicit persona text into dialogue generation models .", "despite the success of existing persona - based models on generating human - like responses , their one - stage decoding framework can hardly avoid the generation of inconsistent persona words .", "in this work , we introduce a three - stage framework that employs a generate - delete - rewrite mechanism to delete inconsistent words from a generated response prototype and further rewrite it to a personality - consistent one .", "we carry out evaluations by both human and automatic metrics .", "experiments on the persona - chat dataset show that our approach achieves good performance ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 2, 3 ], "text": "consistent personality", "tokens": [ "consistent", "personality" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 0 ], "text": "maintaining", "tokens": [ "maintaining" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 47, 48, 49 ], "text": "dialogue generation models", "tokens": [ "dialogue", "generation", "models" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 43, 44, 45 ], "text": "explicit persona text", "tokens": [ "explicit", "persona", "text" ] } ], "event_type": "RWS", "trigger": { "offsets": [ 42 ], "text": "incorporating", "tokens": [ "incorporating" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 68, 69, 70, 71, 72 ], "text": "one - stage decoding framework", "tokens": [ "one", "-", "stage", "decoding", "framework" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 74, 75 ], "text": "hardly avoid", "tokens": [ "hardly", "avoid" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 74, 75 ], "text": "hardly avoid", "tokens": [ "hardly", "avoid" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 87 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 90, 91, 92, 93 ], "text": "three - stage framework", "tokens": [ "three", "-", "stage", "framework" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 104 ], "text": "delete", "tokens": [ "delete" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 88 ], "text": "introduce", "tokens": [ "introduce" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 105, 106, 107, 108, 109, 110, 111 ], "text": "inconsistent words from a generated response prototype", "tokens": [ "inconsistent", "words", "from", "a", "generated", "response", "prototype" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 104 ], "text": "delete", "tokens": [ "delete" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 118, 119, 120, 121 ], "text": "personality - consistent one", "tokens": [ "personality", "-", "consistent", "one" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 105, 106 ], "text": "inconsistent words", "tokens": [ "inconsistent", "words" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 113, 114 ], "text": "further rewrite", "tokens": [ "further", "rewrite" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 123 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 129, 132 ], "text": "human metrics", "tokens": [ "human", "metrics" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 131, 132 ], "text": "automatic metrics", "tokens": [ "automatic", "metrics" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 126 ], "text": "evaluations", "tokens": [ "evaluations" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 145 ], "text": "achieves", "tokens": [ "achieves" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 141 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 90, 91, 92, 93 ], "text": "three - stage framework", "tokens": [ "three", "-", "stage", "framework" ] }, { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 146, 147 ], "text": "good performance", "tokens": [ "good", "performance" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 145 ], "text": "achieves", "tokens": [ "achieves" ] } } ]
[ "maintaining", "a", "consistent", "personality", "in", "conversations", "is", "quite", "natural", "for", "human", "beings", ",", "but", "is", "still", "a", "non", "-", "trivial", "task", "for", "machines", ".", "the", "persona", "-", "based", "dialogue", "generation", "task", "is", "thus", "introduced", "to", "tackle", "the", "personality", "-", "inconsistent", "problem", "by", "incorporating", "explicit", "persona", "text", "into", "dialogue", "generation", "models", ".", "despite", "the", "success", "of", "existing", "persona", "-", "based", "models", "on", "generating", "human", "-", "like", "responses", ",", "their", "one", "-", "stage", "decoding", "framework", "can", "hardly", "avoid", "the", "generation", "of", "inconsistent", "persona", "words", ".", "in", "this", "work", ",", "we", "introduce", "a", "three", "-", "stage", "framework", "that", "employs", "a", "generate", "-", "delete", "-", "rewrite", "mechanism", "to", "delete", "inconsistent", "words", "from", "a", "generated", "response", "prototype", "and", "further", "rewrite", "it", "to", "a", "personality", "-", "consistent", "one", ".", "we", "carry", "out", "evaluations", "by", "both", "human", "and", "automatic", "metrics", ".", "experiments", "on", "the", "persona", "-", "chat", "dataset", "show", "that", "our", "approach", "achieves", "good", "performance", "." ]
ACL
An In-depth Study on Internal Structure of Chinese Words
Unlike English letters, Chinese characters have rich and specific meanings. Usually, the meaning of a word can be derived from its constituent characters in some way. Several previous works on syntactic parsing propose to annotate shallow word-internal structures for better utilizing character-level information. This work proposes to model the deep internal structures of Chinese words as dependency trees with 11 labels for distinguishing syntactic relationships. First, based on newly compiled annotation guidelines, we manually annotate a word-internal structure treebank (WIST) consisting of over 30K multi-char words from Chinese Penn Treebank. To guarantee quality, each word is independently annotated by two annotators and inconsistencies are handled by a third senior annotator. Second, we present detailed and interesting analysis on WIST to reveal insights on Chinese word formation. Third, we propose word-internal structure parsing as a new task, and conduct benchmark experiments using a competitive dependency parser. Finally, we present two simple ways to encode word-internal structures, leading to promising gains on the sentence-level syntactic parsing task.
636dd0c8ece0788d40d37b9f500026d8
2,021
[ "unlike english letters , chinese characters have rich and specific meanings .", "usually , the meaning of a word can be derived from its constituent characters in some way .", "several previous works on syntactic parsing propose to annotate shallow word - internal structures for better utilizing character - level information .", "this work proposes to model the deep internal structures of chinese words as dependency trees with 11 labels for distinguishing syntactic relationships .", "first , based on newly compiled annotation guidelines , we manually annotate a word - internal structure treebank ( wist ) consisting of over 30k multi - char words from chinese penn treebank .", "to guarantee quality , each word is independently annotated by two annotators and inconsistencies are handled by a third senior annotator .", "second , we present detailed and interesting analysis on wist to reveal insights on chinese word formation .", "third , we propose word - internal structure parsing as a new task , and conduct benchmark experiments using a competitive dependency parser .", "finally , we present two simple ways to encode word - internal structures , leading to promising gains on the sentence - level syntactic parsing task ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 4, 5 ], "text": "chinese characters", "tokens": [ "chinese", "characters" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 6 ], "text": "have", "tokens": [ "have" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 31, 32, 33, 34, 35 ], "text": "previous works on syntactic parsing", "tokens": [ "previous", "works", "on", "syntactic", "parsing" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 45, 46 ], "text": "better utilizing", "tokens": [ "better", "utilizing" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 39, 40, 41, 42, 43 ], "text": "shallow word - internal structures", "tokens": [ "shallow", "word", "-", "internal", "structures" ] } ], "event_type": "RWS", "trigger": { "offsets": [ 38 ], "text": "annotate", "tokens": [ "annotate" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 47, 48, 49, 50 ], "text": "character - level information", "tokens": [ "character", "-", "level", "information" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 45, 46 ], "text": "better utilizing", "tokens": [ "better", "utilizing" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 71 ], "text": "distinguishing", "tokens": [ "distinguishing" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 58, 59, 60, 61, 62, 63 ], "text": "deep internal structures of chinese words", "tokens": [ "deep", "internal", "structures", "of", "chinese", "words" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 67, 68, 69 ], "text": "with 11 labels", "tokens": [ "with", "11", "labels" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 65, 66 ], "text": "dependency trees", "tokens": [ "dependency", "trees" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 56 ], "text": "model", "tokens": [ "model" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 72, 73 ], "text": "syntactic relationships", "tokens": [ "syntactic", "relationships" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 71 ], "text": "distinguishing", "tokens": [ "distinguishing" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 77, 78, 79, 80, 81, 82 ], "text": "based on newly compiled annotation guidelines", "tokens": [ "based", "on", "newly", "compiled", "annotation", "guidelines" ] }, { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 84 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 88, 89, 90, 91, 92 ], "text": "word - internal structure treebank", "tokens": [ "word", "-", "internal", "structure", "treebank" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 85, 86 ], "text": "manually annotate", "tokens": [ "manually", "annotate" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 142 ], "text": "reveal", "tokens": [ "reveal" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 88, 89, 90, 91, 92 ], "text": "word - internal structure treebank", "tokens": [ "word", "-", "internal", "structure", "treebank" ] }, { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 133 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 135, 138 ], "text": "detailed analysis", "tokens": [ "detailed", "analysis" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 137, 138 ], "text": "interesting analysis", "tokens": [ "interesting", "analysis" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 134 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 143, 144, 145, 146, 147 ], "text": "insights on chinese word formation", "tokens": [ "insights", "on", "chinese", "word", "formation" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 142 ], "text": "reveal", "tokens": [ "reveal" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 151 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 153, 154, 155, 156, 157 ], "text": "word - internal structure parsing", "tokens": [ "word", "-", "internal", "structure", "parsing" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 152 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 151 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 169, 170, 171 ], "text": "competitive dependency parser", "tokens": [ "competitive", "dependency", "parser" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 164 ], "text": "conduct", "tokens": [ "conduct" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 167 ], "text": "using", "tokens": [ "using" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 165, 166 ], "text": "benchmark experiments", "tokens": [ "benchmark", "experiments" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 164 ], "text": "conduct", "tokens": [ "conduct" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 175 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 177, 178, 179 ], "text": "two simple ways", "tokens": [ "two", "simple", "ways" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 181 ], "text": "encode", "tokens": [ "encode" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 176 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 182, 183, 184, 185 ], "text": "word - internal structures", "tokens": [ "word", "-", "internal", "structures" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 181 ], "text": "encode", "tokens": [ "encode" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 177, 178, 179 ], "text": "two simple ways", "tokens": [ "two", "simple", "ways" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 191, 192, 193, 194, 195, 196, 197, 198 ], "text": "on the sentence - level syntactic parsing task", "tokens": [ "on", "the", "sentence", "-", "level", "syntactic", "parsing", "task" ] }, { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 189, 190 ], "text": "promising gains", "tokens": [ "promising", "gains" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 187 ], "text": "leading", "tokens": [ "leading" ] } } ]
[ "unlike", "english", "letters", ",", "chinese", "characters", "have", "rich", "and", "specific", "meanings", ".", "usually", ",", "the", "meaning", "of", "a", "word", "can", "be", "derived", "from", "its", "constituent", "characters", "in", "some", "way", ".", "several", "previous", "works", "on", "syntactic", "parsing", "propose", "to", "annotate", "shallow", "word", "-", "internal", "structures", "for", "better", "utilizing", "character", "-", "level", "information", ".", "this", "work", "proposes", "to", "model", "the", "deep", "internal", "structures", "of", "chinese", "words", "as", "dependency", "trees", "with", "11", "labels", "for", "distinguishing", "syntactic", "relationships", ".", "first", ",", "based", "on", "newly", "compiled", "annotation", "guidelines", ",", "we", "manually", "annotate", "a", "word", "-", "internal", "structure", "treebank", "(", "wist", ")", "consisting", "of", "over", "30k", "multi", "-", "char", "words", "from", "chinese", "penn", "treebank", ".", "to", "guarantee", "quality", ",", "each", "word", "is", "independently", "annotated", "by", "two", "annotators", "and", "inconsistencies", "are", "handled", "by", "a", "third", "senior", "annotator", ".", "second", ",", "we", "present", "detailed", "and", "interesting", "analysis", "on", "wist", "to", "reveal", "insights", "on", "chinese", "word", "formation", ".", "third", ",", "we", "propose", "word", "-", "internal", "structure", "parsing", "as", "a", "new", "task", ",", "and", "conduct", "benchmark", "experiments", "using", "a", "competitive", "dependency", "parser", ".", "finally", ",", "we", "present", "two", "simple", "ways", "to", "encode", "word", "-", "internal", "structures", ",", "leading", "to", "promising", "gains", "on", "the", "sentence", "-", "level", "syntactic", "parsing", "task", "." ]
ACL
Preview, Attend and Review: Schema-Aware Curriculum Learning for Multi-Domain Dialogue State Tracking
Existing dialog state tracking (DST) models are trained with dialog data in a random order, neglecting rich structural information in a dataset. In this paper, we propose to use curriculum learning (CL) to better leverage both the curriculum structure and schema structure for task-oriented dialogs. Specifically, we propose a model-agnostic framework called Schema-aware Curriculum Learning for Dialog State Tracking (SaCLog), which consists of a preview module that pre-trains a DST model with schema information, a curriculum module that optimizes the model with CL, and a review module that augments mispredicted data to reinforce the CL training. We show that our proposed approach improves DST performance over both a transformer-based and RNN-based DST model (TripPy and TRADE) and achieves new state-of-the-art results on WOZ2.0 and MultiWOZ2.1.
a0fd29c17984ed8d2e2b7f86831cb0a4
2,021
[ "existing dialog state tracking ( dst ) models are trained with dialog data in a random order , neglecting rich structural information in a dataset .", "in this paper , we propose to use curriculum learning ( cl ) to better leverage both the curriculum structure and schema structure for task - oriented dialogs .", "specifically , we propose a model - agnostic framework called schema - aware curriculum learning for dialog state tracking ( saclog ) , which consists of a preview module that pre - trains a dst model with schema information , a curriculum module that optimizes the model with cl , and a review module that augments mispredicted data to reinforce the cl training .", "we show that our proposed approach improves dst performance over both a transformer - based and rnn - based dst model ( trippy and trade ) and achieves new state - of - the - art results on woz2 . 0 and multiwoz2 . 1 ." ]
[ { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "APP", "offsets": [ 34, 35 ], "text": "curriculum learning", "tokens": [ "curriculum", "learning" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 44, 45 ], "text": "curriculum structure", "tokens": [ "curriculum", "structure" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 47, 48 ], "text": "schema structure", "tokens": [ "schema", "structure" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 50, 51, 52, 53 ], "text": "task - oriented dialogs", "tokens": [ "task", "-", "oriented", "dialogs" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 40, 41 ], "text": "better leverage", "tokens": [ "better", "leverage" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 57 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 65, 66, 67, 68, 69, 70, 71, 72, 73 ], "text": "schema - aware curriculum learning for dialog state tracking", "tokens": [ "schema", "-", "aware", "curriculum", "learning", "for", "dialog", "state", "tracking" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 58 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 119 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 125 ], "text": "improves", "tokens": [ "improves" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 146 ], "text": "achieves", "tokens": [ "achieves" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 120 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 125 ], "text": "improves", "tokens": [ "improves" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 1, 2, 3, 127 ], "text": "dst performance", "tokens": [ "dialog", "state", "tracking", "performance" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 65, 66, 67, 68, 69, 70, 71, 72, 73 ], "text": "schema - aware curriculum learning for dialog state tracking", "tokens": [ "schema", "-", "aware", "curriculum", "learning", "for", "dialog", "state", "tracking" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 131, 132, 133, 138, 139 ], "text": "transformer - based dst model", "tokens": [ "transformer", "-", "based", "dst", "model" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 125 ], "text": "improves", "tokens": [ "improves" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 138, 7 ], "text": "dialog state tracking ( dst ) models", "tokens": [ "dst", "models" ] }, { "argument_type": "Fault", "nugget_type": "FEA", "offsets": [ 19, 20, 21 ], "text": "rich structural information", "tokens": [ "rich", "structural", "information" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 18 ], "text": "neglecting", "tokens": [ "neglecting" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 147, 148, 149, 150, 151, 152, 153, 154, 155 ], "text": "new state - of - the - art results", "tokens": [ "new", "state", "-", "of", "-", "the", "-", "art", "results" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 156, 157, 158, 159, 160, 161, 162, 163 ], "text": "on woz2 . 0 and multiwoz2 . 1", "tokens": [ "on", "woz2", ".", "0", "and", "multiwoz2", ".", "1" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 65, 66, 67, 68, 69, 70, 71, 72, 73 ], "text": "schema - aware curriculum learning for dialog state tracking", "tokens": [ "schema", "-", "aware", "curriculum", "learning", "for", "dialog", "state", "tracking" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 146 ], "text": "achieves", "tokens": [ "achieves" ] } } ]
[ "existing", "dialog", "state", "tracking", "(", "dst", ")", "models", "are", "trained", "with", "dialog", "data", "in", "a", "random", "order", ",", "neglecting", "rich", "structural", "information", "in", "a", "dataset", ".", "in", "this", "paper", ",", "we", "propose", "to", "use", "curriculum", "learning", "(", "cl", ")", "to", "better", "leverage", "both", "the", "curriculum", "structure", "and", "schema", "structure", "for", "task", "-", "oriented", "dialogs", ".", "specifically", ",", "we", "propose", "a", "model", "-", "agnostic", "framework", "called", "schema", "-", "aware", "curriculum", "learning", "for", "dialog", "state", "tracking", "(", "saclog", ")", ",", "which", "consists", "of", "a", "preview", "module", "that", "pre", "-", "trains", "a", "dst", "model", "with", "schema", "information", ",", "a", "curriculum", "module", "that", "optimizes", "the", "model", "with", "cl", ",", "and", "a", "review", "module", "that", "augments", "mispredicted", "data", "to", "reinforce", "the", "cl", "training", ".", "we", "show", "that", "our", "proposed", "approach", "improves", "dst", "performance", "over", "both", "a", "transformer", "-", "based", "and", "rnn", "-", "based", "dst", "model", "(", "trippy", "and", "trade", ")", "and", "achieves", "new", "state", "-", "of", "-", "the", "-", "art", "results", "on", "woz2", ".", "0", "and", "multiwoz2", ".", "1", "." ]
ACL
Self-Attentional Models for Lattice Inputs
Lattices are an efficient and effective method to encode ambiguity of upstream systems in natural language processing tasks, for example to compactly capture multiple speech recognition hypotheses, or to represent multiple linguistic analyses. Previous work has extended recurrent neural networks to model lattice inputs and achieved improvements in various tasks, but these models suffer from very slow computation speeds. This paper extends the recently proposed paradigm of self-attention to handle lattice inputs. Self-attention is a sequence modeling technique that relates inputs to one another by computing pairwise similarities and has gained popularity for both its strong results and its computational efficiency. To extend such models to handle lattices, we introduce probabilistic reachability masks that incorporate lattice structure into the model and support lattice scores if available. We also propose a method for adapting positional embeddings to lattice structures. We apply the proposed model to a speech translation task and find that it outperforms all examined baselines while being much faster to compute than previous neural lattice models during both training and inference.
8e057b24ffe8ed4a5448b19bb7b9c2bf
2,019
[ "lattices are an efficient and effective method to encode ambiguity of upstream systems in natural language processing tasks , for example to compactly capture multiple speech recognition hypotheses , or to represent multiple linguistic analyses .", "previous work has extended recurrent neural networks to model lattice inputs and achieved improvements in various tasks , but these models suffer from very slow computation speeds .", "this paper extends the recently proposed paradigm of self - attention to handle lattice inputs .", "self - attention is a sequence modeling technique that relates inputs to one another by computing pairwise similarities and has gained popularity for both its strong results and its computational efficiency .", "to extend such models to handle lattices , we introduce probabilistic reachability masks that incorporate lattice structure into the model and support lattice scores if available .", "we also propose a method for adapting positional embeddings to lattice structures .", "we apply the proposed model to a speech translation task and find that it outperforms all examined baselines while being much faster to compute than previous neural lattice models during both training and inference ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0 ], "text": "lattices", "tokens": [ "lattices" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 13, 14, 15, 16, 17 ], "text": "in natural language processing tasks", "tokens": [ "in", "natural", "language", "processing", "tasks" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 6 ], "text": "method", "tokens": [ "method" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 36, 37 ], "text": "previous work", "tokens": [ "previous", "work" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 59, 60, 61, 62 ], "text": "very slow computation speeds", "tokens": [ "very", "slow", "computation", "speeds" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 57 ], "text": "suffer", "tokens": [ "suffer" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 76 ], "text": "handle", "tokens": [ "handle" ] }, { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 68, 69, 70, 71, 72, 73, 74 ], "text": "recently proposed paradigm of self - attention", "tokens": [ "recently", "proposed", "paradigm", "of", "self", "-", "attention" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 66 ], "text": "extends", "tokens": [ "extends" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "MOD", "offsets": [ 77, 78 ], "text": "lattice inputs", "tokens": [ "lattice", "inputs" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 76 ], "text": "handle", "tokens": [ "handle" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 120 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 113 ], "text": "extend", "tokens": [ "extend" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 122, 123, 124 ], "text": "probabilistic reachability masks", "tokens": [ "probabilistic", "reachability", "masks" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 121 ], "text": "introduce", "tokens": [ "introduce" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 146, 147 ], "text": "positional embeddings", "tokens": [ "positional", "embeddings" ] }, { "argument_type": "Target", "nugget_type": "MOD", "offsets": [ 149, 150 ], "text": "lattice structures", "tokens": [ "lattice", "structures" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 145 ], "text": "adapting", "tokens": [ "adapting" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 152 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 166 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 175 ], "text": "compute", "tokens": [ "compute" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 163 ], "text": "find", "tokens": [ "find" ] } }, { "arguments": [ { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 166 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 167, 168, 169 ], "text": "all examined baselines", "tokens": [ "all", "examined", "baselines" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 166 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 177, 178, 179, 180 ], "text": "previous neural lattice models", "tokens": [ "previous", "neural", "lattice", "models" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 181, 182, 183, 184, 185 ], "text": "during both training and inference", "tokens": [ "during", "both", "training", "and", "inference" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 173 ], "text": "faster", "tokens": [ "faster" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 172 ], "text": "much", "tokens": [ "much" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 122, 123, 124 ], "text": "probabilistic reachability masks", "tokens": [ "probabilistic", "reachability", "masks" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 175 ], "text": "compute", "tokens": [ "compute" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 127, 128 ], "text": "lattice structure", "tokens": [ "lattice", "structure" ] }, { "argument_type": "Target", "nugget_type": "APP", "offsets": [ 131 ], "text": "model", "tokens": [ "model" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 126 ], "text": "incorporate", "tokens": [ "incorporate" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 136, 137 ], "text": "if available", "tokens": [ "if", "available" ] }, { "argument_type": "Aim", "nugget_type": "APP", "offsets": [ 115 ], "text": "models", "tokens": [ "models" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 113 ], "text": "extend", "tokens": [ "extend" ] } } ]
[ "lattices", "are", "an", "efficient", "and", "effective", "method", "to", "encode", "ambiguity", "of", "upstream", "systems", "in", "natural", "language", "processing", "tasks", ",", "for", "example", "to", "compactly", "capture", "multiple", "speech", "recognition", "hypotheses", ",", "or", "to", "represent", "multiple", "linguistic", "analyses", ".", "previous", "work", "has", "extended", "recurrent", "neural", "networks", "to", "model", "lattice", "inputs", "and", "achieved", "improvements", "in", "various", "tasks", ",", "but", "these", "models", "suffer", "from", "very", "slow", "computation", "speeds", ".", "this", "paper", "extends", "the", "recently", "proposed", "paradigm", "of", "self", "-", "attention", "to", "handle", "lattice", "inputs", ".", "self", "-", "attention", "is", "a", "sequence", "modeling", "technique", "that", "relates", "inputs", "to", "one", "another", "by", "computing", "pairwise", "similarities", "and", "has", "gained", "popularity", "for", "both", "its", "strong", "results", "and", "its", "computational", "efficiency", ".", "to", "extend", "such", "models", "to", "handle", "lattices", ",", "we", "introduce", "probabilistic", "reachability", "masks", "that", "incorporate", "lattice", "structure", "into", "the", "model", "and", "support", "lattice", "scores", "if", "available", ".", "we", "also", "propose", "a", "method", "for", "adapting", "positional", "embeddings", "to", "lattice", "structures", ".", "we", "apply", "the", "proposed", "model", "to", "a", "speech", "translation", "task", "and", "find", "that", "it", "outperforms", "all", "examined", "baselines", "while", "being", "much", "faster", "to", "compute", "than", "previous", "neural", "lattice", "models", "during", "both", "training", "and", "inference", "." ]
ACL
Joint Effects of Context and User History for Predicting Online Conversation Re-entries
As the online world continues its exponential growth, interpersonal communication has come to play an increasingly central role in opinion formation and change. In order to help users better engage with each other online, we study a challenging problem of re-entry prediction foreseeing whether a user will come back to a conversation they once participated in. We hypothesize that both the context of the ongoing conversations and the users’ previous chatting history will affect their continued interests in future engagement. Specifically, we propose a neural framework with three main layers, each modeling context, user history, and interactions between them, to explore how the conversation context and user chatting history jointly result in their re-entry behavior. We experiment with two large-scale datasets collected from Twitter and Reddit. Results show that our proposed framework with bi-attention achieves an F1 score of 61.1 on Twitter conversations, outperforming the state-of-the-art methods from previous work.
54dc18f3c81976ab42c7f5f4bd591db4
2,019
[ "as the online world continues its exponential growth , interpersonal communication has come to play an increasingly central role in opinion formation and change .", "in order to help users better engage with each other online , we study a challenging problem of re - entry prediction foreseeing whether a user will come back to a conversation they once participated in .", "we hypothesize that both the context of the ongoing conversations and the users ’ previous chatting history will affect their continued interests in future engagement .", "specifically , we propose a neural framework with three main layers , each modeling context , user history , and interactions between them , to explore how the conversation context and user chatting history jointly result in their re - entry behavior .", "we experiment with two large - scale datasets collected from twitter and reddit .", "results show that our proposed framework with bi - attention achieves an f1 score of 61 . 1 on twitter conversations , outperforming the state - of - the - art methods from previous work ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 9, 10 ], "text": "interpersonal communication", "tokens": [ "interpersonal", "communication" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 14 ], "text": "play", "tokens": [ "play" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 28 ], "text": "help", "tokens": [ "help" ] }, { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 37 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 40, 41, 42, 43, 44, 45, 46 ], "text": "challenging problem of re - entry prediction", "tokens": [ "challenging", "problem", "of", "re", "-", "entry", "prediction" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 38 ], "text": "study", "tokens": [ "study" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 29, 30, 31, 32, 33, 34, 35 ], "text": "users better engage with each other online", "tokens": [ "users", "better", "engage", "with", "each", "other", "online" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 28 ], "text": "help", "tokens": [ "help" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 62 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86 ], "text": "both the context of the ongoing conversations and the users ’ previous chatting history will affect their continued interests in future engagement", "tokens": [ "both", "the", "context", "of", "the", "ongoing", "conversations", "and", "the", "users", "’", "previous", "chatting", "history", "will", "affect", "their", "continued", "interests", "in", "future", "engagement" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 63 ], "text": "hypothesize", "tokens": [ "hypothesize" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 90 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 93, 94, 95, 96, 97, 98 ], "text": "neural framework with three main layers", "tokens": [ "neural", "framework", "with", "three", "main", "layers" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 113 ], "text": "explore", "tokens": [ "explore" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 91 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129 ], "text": "how the conversation context and user chatting history jointly result in their re - entry behavior", "tokens": [ "how", "the", "conversation", "context", "and", "user", "chatting", "history", "jointly", "result", "in", "their", "re", "-", "entry", "behavior" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 113 ], "text": "explore", "tokens": [ "explore" ] } }, { "arguments": [ { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 134, 135, 136, 137, 138, 139, 140, 141, 142, 143 ], "text": "two large - scale datasets collected from twitter and reddit", "tokens": [ "two", "large", "-", "scale", "datasets", "collected", "from", "twitter", "and", "reddit" ] }, { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 131 ], "text": "we", "tokens": [ "we" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 132 ], "text": "experiment", "tokens": [ "experiment" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 155 ], "text": "achieves", "tokens": [ "achieves" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 167 ], "text": "outperforming", "tokens": [ "outperforming" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 146 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 93, 94, 95, 96, 97, 98, 151, 152, 153, 154 ], "text": "our proposed framework with bi - attention", "tokens": [ "neural", "framework", "with", "three", "main", "layers", "with", "bi", "-", "attention" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 157, 158 ], "text": "f1 score", "tokens": [ "f1", "score" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 160, 161, 162 ], "text": "61 . 1", "tokens": [ "61", ".", "1" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 163, 164, 165 ], "text": "on twitter conversations", "tokens": [ "on", "twitter", "conversations" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 155 ], "text": "achieves", "tokens": [ "achieves" ] } }, { "arguments": [ { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 167 ], "text": "outperforming", "tokens": [ "outperforming" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 93, 94 ], "text": "neural framework", "tokens": [ "neural", "framework" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 167 ], "text": "outperforming", "tokens": [ "outperforming" ] } } ]
[ "as", "the", "online", "world", "continues", "its", "exponential", "growth", ",", "interpersonal", "communication", "has", "come", "to", "play", "an", "increasingly", "central", "role", "in", "opinion", "formation", "and", "change", ".", "in", "order", "to", "help", "users", "better", "engage", "with", "each", "other", "online", ",", "we", "study", "a", "challenging", "problem", "of", "re", "-", "entry", "prediction", "foreseeing", "whether", "a", "user", "will", "come", "back", "to", "a", "conversation", "they", "once", "participated", "in", ".", "we", "hypothesize", "that", "both", "the", "context", "of", "the", "ongoing", "conversations", "and", "the", "users", "’", "previous", "chatting", "history", "will", "affect", "their", "continued", "interests", "in", "future", "engagement", ".", "specifically", ",", "we", "propose", "a", "neural", "framework", "with", "three", "main", "layers", ",", "each", "modeling", "context", ",", "user", "history", ",", "and", "interactions", "between", "them", ",", "to", "explore", "how", "the", "conversation", "context", "and", "user", "chatting", "history", "jointly", "result", "in", "their", "re", "-", "entry", "behavior", ".", "we", "experiment", "with", "two", "large", "-", "scale", "datasets", "collected", "from", "twitter", "and", "reddit", ".", "results", "show", "that", "our", "proposed", "framework", "with", "bi", "-", "attention", "achieves", "an", "f1", "score", "of", "61", ".", "1", "on", "twitter", "conversations", ",", "outperforming", "the", "state", "-", "of", "-", "the", "-", "art", "methods", "from", "previous", "work", "." ]
ACL
Probing for Predicate Argument Structures in Pretrained Language Models
Thanks to the effectiveness and wide availability of modern pretrained language models (PLMs), recently proposed approaches have achieved remarkable results in dependency- and span-based, multilingual and cross-lingual Semantic Role Labeling (SRL). These results have prompted researchers to investigate the inner workings of modern PLMs with the aim of understanding how, where, and to what extent they encode information about SRL. In this paper, we follow this line of research and probe for predicate argument structures in PLMs. Our study shows that PLMs do encode semantic structures directly into the contextualized representation of a predicate, and also provides insights into the correlation between predicate senses and their structures, the degree of transferability between nominal and verbal structures, and how such structures are encoded across languages. Finally, we look at the practical implications of such insights and demonstrate the benefits of embedding predicate argument structure information into an SRL model.
81cb7fa52062f9d17d9e93a1e4567dec
2,022
[ "thanks to the effectiveness and wide availability of modern pretrained language models ( plms ) , recently proposed approaches have achieved remarkable results in dependency - and span - based , multilingual and cross - lingual semantic role labeling ( srl ) .", "these results have prompted researchers to investigate the inner workings of modern plms with the aim of understanding how , where , and to what extent they encode information about srl .", "in this paper , we follow this line of research and probe for predicate argument structures in plms .", "our study shows that plms do encode semantic structures directly into the contextualized representation of a predicate , and also provides insights into the correlation between predicate senses and their structures , the degree of transferability between nominal and verbal structures , and how such structures are encoded across languages .", "finally , we look at the practical implications of such insights and demonstrate the benefits of embedding predicate argument structure information into an srl model ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 36, 37, 38 ], "text": "semantic role labeling", "tokens": [ "semantic", "role", "labeling" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 20 ], "text": "achieved", "tokens": [ "achieved" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 79 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 88, 89, 90 ], "text": "predicate argument structures", "tokens": [ "predicate", "argument", "structures" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 91, 9, 10, 11 ], "text": "in plms", "tokens": [ "in", "pretrained", "language", "models" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 86 ], "text": "probe", "tokens": [ "probe" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "MOD", "offsets": [ 101, 102 ], "text": "semantic structures", "tokens": [ "semantic", "structures" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 104, 105, 106, 107, 108, 109, 110 ], "text": "into the contextualized representation of a predicate", "tokens": [ "into", "the", "contextualized", "representation", "of", "a", "predicate" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 9, 10, 11 ], "text": "pretrained language models", "tokens": [ "pretrained", "language", "models" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 100 ], "text": "encode", "tokens": [ "encode" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 9, 10, 11 ], "text": "pretrained language models", "tokens": [ "pretrained", "language", "models" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 127, 128, 129 ], "text": "degree of transferability", "tokens": [ "degree", "of", "transferability" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 130, 131, 132, 133, 134 ], "text": "between nominal and verbal structures", "tokens": [ "between", "nominal", "and", "verbal", "structures" ] }, { "argument_type": "Object", "nugget_type": "FEA", "offsets": [ 139 ], "text": "structures", "tokens": [ "structures" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 115, 116, 117, 118 ], "text": "insights into the correlation", "tokens": [ "insights", "into", "the", "correlation" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 119, 120, 121, 122, 123, 124 ], "text": "between predicate senses and their structures", "tokens": [ "between", "predicate", "senses", "and", "their", "structures" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 114 ], "text": "provides", "tokens": [ "provides" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 147 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 151, 152, 153, 154, 155 ], "text": "practical implications of such insights", "tokens": [ "practical", "implications", "of", "such", "insights" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 148, 149 ], "text": "look at", "tokens": [ "look", "at" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 147 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "STR", "offsets": [ 159, 160, 161, 162, 163, 164, 165, 166, 167, 36, 37, 38, 169 ], "text": "benefits of embedding predicate argument structure information into an srl model", "tokens": [ "benefits", "of", "embedding", "predicate", "argument", "structure", "information", "into", "an", "semantic", "role", "labeling", "model" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 157 ], "text": "demonstrate", "tokens": [ "demonstrate" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 100 ], "text": "encode", "tokens": [ "encode" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 96 ], "text": "shows", "tokens": [ "shows" ] } } ]
[ "thanks", "to", "the", "effectiveness", "and", "wide", "availability", "of", "modern", "pretrained", "language", "models", "(", "plms", ")", ",", "recently", "proposed", "approaches", "have", "achieved", "remarkable", "results", "in", "dependency", "-", "and", "span", "-", "based", ",", "multilingual", "and", "cross", "-", "lingual", "semantic", "role", "labeling", "(", "srl", ")", ".", "these", "results", "have", "prompted", "researchers", "to", "investigate", "the", "inner", "workings", "of", "modern", "plms", "with", "the", "aim", "of", "understanding", "how", ",", "where", ",", "and", "to", "what", "extent", "they", "encode", "information", "about", "srl", ".", "in", "this", "paper", ",", "we", "follow", "this", "line", "of", "research", "and", "probe", "for", "predicate", "argument", "structures", "in", "plms", ".", "our", "study", "shows", "that", "plms", "do", "encode", "semantic", "structures", "directly", "into", "the", "contextualized", "representation", "of", "a", "predicate", ",", "and", "also", "provides", "insights", "into", "the", "correlation", "between", "predicate", "senses", "and", "their", "structures", ",", "the", "degree", "of", "transferability", "between", "nominal", "and", "verbal", "structures", ",", "and", "how", "such", "structures", "are", "encoded", "across", "languages", ".", "finally", ",", "we", "look", "at", "the", "practical", "implications", "of", "such", "insights", "and", "demonstrate", "the", "benefits", "of", "embedding", "predicate", "argument", "structure", "information", "into", "an", "srl", "model", "." ]
ACL
Where to Go for the Holidays: Towards Mixed-Type Dialogs for Clarification of User Goals
Most dialog systems posit that users have figured out clear and specific goals before starting an interaction. For example, users have determined the departure, the destination, and the travel time for booking a flight. However, in many scenarios, limited by experience and knowledge, users may know what they need, but still struggle to figure out clear and specific goals by determining all the necessary slots. In this paper, we identify this challenge, and make a step forward by collecting a new human-to-human mixed-type dialog corpus. It contains 5k dialog sessions and 168k utterances for 4 dialog types and 5 domains. Within each session, an agent first provides user-goal-related knowledge to help figure out clear and specific goals, and then help achieve them. Furthermore, we propose a mixed-type dialog model with a novel Prompt-based continual learning mechanism. Specifically, the mechanism enables the model to continually strengthen its ability on any specific type by utilizing existing dialog corpora effectively.
e6819b3ce223923478bb9d3b63e830a6
2,022
[ "most dialog systems posit that users have figured out clear and specific goals before starting an interaction .", "for example , users have determined the departure , the destination , and the travel time for booking a flight .", "however , in many scenarios , limited by experience and knowledge , users may know what they need , but still struggle to figure out clear and specific goals by determining all the necessary slots .", "in this paper , we identify this challenge , and make a step forward by collecting a new human - to - human mixed - type dialog corpus .", "it contains 5k dialog sessions and 168k utterances for 4 dialog types and 5 domains .", "within each session , an agent first provides user - goal - related knowledge to help figure out clear and specific goals , and then help achieve them .", "furthermore , we propose a mixed - type dialog model with a novel prompt - based continual learning mechanism .", "specifically , the mechanism enables the model to continually strengthen its ability on any specific type by utilizing existing dialog corpora effectively ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 1, 2 ], "text": "dialog systems", "tokens": [ "dialog", "systems" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 3 ], "text": "posit", "tokens": [ "posit" ] } }, { "arguments": [ { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 59, 60, 61, 62, 63 ], "text": "still struggle to figure out", "tokens": [ "still", "struggle", "to", "figure", "out" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 64, 65, 66, 67 ], "text": "clear and specific goals", "tokens": [ "clear", "and", "specific", "goals" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 59, 60, 61, 62, 63 ], "text": "still struggle to figure out", "tokens": [ "still", "struggle", "to", "figure", "out" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 79 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "WEA", "offsets": [ 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73 ], "text": "still struggle to figure out clear and specific goals by determining all the necessary slots", "tokens": [ "still", "struggle", "to", "figure", "out", "clear", "and", "specific", "goals", "by", "determining", "all", "the", "necessary", "slots" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 80 ], "text": "identify", "tokens": [ "identify" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 79 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 93, 94, 95, 96, 97, 98, 99, 100, 101, 102 ], "text": "human - to - human mixed - type dialog corpus", "tokens": [ "human", "-", "to", "-", "human", "mixed", "-", "type", "dialog", "corpus" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 90 ], "text": "collecting", "tokens": [ "collecting" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 128, 129, 130, 131, 132, 133 ], "text": "user - goal - related knowledge", "tokens": [ "user", "-", "goal", "-", "related", "knowledge" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 136, 137 ], "text": "figure out", "tokens": [ "figure", "out" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 146 ], "text": "achieve", "tokens": [ "achieve" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 127 ], "text": "provides", "tokens": [ "provides" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 138, 139, 140, 141 ], "text": "clear and specific goals", "tokens": [ "clear", "and", "specific", "goals" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 136, 137 ], "text": "figure out", "tokens": [ "figure", "out" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 138, 139, 140, 141 ], "text": "clear and specific goals", "tokens": [ "clear", "and", "specific", "goals" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 146 ], "text": "achieve", "tokens": [ "achieve" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 151 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167 ], "text": "mixed - type dialog model with a novel prompt - based continual learning mechanism", "tokens": [ "mixed", "-", "type", "dialog", "model", "with", "a", "novel", "prompt", "-", "based", "continual", "learning", "mechanism" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 152 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "MOD", "offsets": [ 162, 163, 164, 165, 166, 167 ], "text": "prompt - based continual learning mechanism", "tokens": [ "prompt", "-", "based", "continual", "learning", "mechanism" ] }, { "argument_type": "Object", "nugget_type": "APP", "offsets": [ 154, 155, 156, 157, 158 ], "text": "mixed - type dialog model", "tokens": [ "mixed", "-", "type", "dialog", "model" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 177, 178 ], "text": "continually strengthen", "tokens": [ "continually", "strengthen" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 181, 182, 183, 184 ], "text": "on any specific type", "tokens": [ "on", "any", "specific", "type" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 186, 187, 188, 189, 190 ], "text": "utilizing existing dialog corpora effectively", "tokens": [ "utilizing", "existing", "dialog", "corpora", "effectively" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 173 ], "text": "enables", "tokens": [ "enables" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 154, 155, 156, 157, 158, 180 ], "text": "its ability", "tokens": [ "mixed", "-", "type", "dialog", "model", "ability" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 177, 178 ], "text": "continually strengthen", "tokens": [ "continually", "strengthen" ] } } ]
[ "most", "dialog", "systems", "posit", "that", "users", "have", "figured", "out", "clear", "and", "specific", "goals", "before", "starting", "an", "interaction", ".", "for", "example", ",", "users", "have", "determined", "the", "departure", ",", "the", "destination", ",", "and", "the", "travel", "time", "for", "booking", "a", "flight", ".", "however", ",", "in", "many", "scenarios", ",", "limited", "by", "experience", "and", "knowledge", ",", "users", "may", "know", "what", "they", "need", ",", "but", "still", "struggle", "to", "figure", "out", "clear", "and", "specific", "goals", "by", "determining", "all", "the", "necessary", "slots", ".", "in", "this", "paper", ",", "we", "identify", "this", "challenge", ",", "and", "make", "a", "step", "forward", "by", "collecting", "a", "new", "human", "-", "to", "-", "human", "mixed", "-", "type", "dialog", "corpus", ".", "it", "contains", "5k", "dialog", "sessions", "and", "168k", "utterances", "for", "4", "dialog", "types", "and", "5", "domains", ".", "within", "each", "session", ",", "an", "agent", "first", "provides", "user", "-", "goal", "-", "related", "knowledge", "to", "help", "figure", "out", "clear", "and", "specific", "goals", ",", "and", "then", "help", "achieve", "them", ".", "furthermore", ",", "we", "propose", "a", "mixed", "-", "type", "dialog", "model", "with", "a", "novel", "prompt", "-", "based", "continual", "learning", "mechanism", ".", "specifically", ",", "the", "mechanism", "enables", "the", "model", "to", "continually", "strengthen", "its", "ability", "on", "any", "specific", "type", "by", "utilizing", "existing", "dialog", "corpora", "effectively", "." ]
ACL
DVD: A Diagnostic Dataset for Multi-step Reasoning in Video Grounded Dialogue
A video-grounded dialogue system is required to understand both dialogue, which contains semantic dependencies from turn to turn, and video, which contains visual cues of spatial and temporal scene variations. Building such dialogue systems is a challenging problem, involving various reasoning types on both visual and language inputs. Existing benchmarks do not have enough annotations to thoroughly analyze dialogue systems and understand their capabilities and limitations in isolation. These benchmarks are also not explicitly designed to minimise biases that models can exploit without actual reasoning. To address these limitations, in this paper, we present DVD, a Diagnostic Dataset for Video-grounded Dialogue. The dataset is designed to contain minimal biases and has detailed annotations for the different types of reasoning over the spatio-temporal space of video. Dialogues are synthesized over multiple question turns, each of which is injected with a set of cross-turn semantic relationships. We use DVD to analyze existing approaches, providing interesting insights into their abilities and limitations. In total, DVD is built from 11k CATER synthetic videos and contains 10 instances of 10-round dialogues for each video, resulting in more than 100k dialogues and 1M question-answer pairs. Our code and dataset are publicly available.
d45cd0ddedda4f5e033a5ce54cd0afb9
2,021
[ "a video - grounded dialogue system is required to understand both dialogue , which contains semantic dependencies from turn to turn , and video , which contains visual cues of spatial and temporal scene variations .", "building such dialogue systems is a challenging problem , involving various reasoning types on both visual and language inputs .", "existing benchmarks do not have enough annotations to thoroughly analyze dialogue systems and understand their capabilities and limitations in isolation .", "these benchmarks are also not explicitly designed to minimise biases that models can exploit without actual reasoning .", "to address these limitations , in this paper , we present dvd , a diagnostic dataset for video - grounded dialogue .", "the dataset is designed to contain minimal biases and has detailed annotations for the different types of reasoning over the spatio - temporal space of video .", "dialogues are synthesized over multiple question turns , each of which is injected with a set of cross - turn semantic relationships .", "we use dvd to analyze existing approaches , providing interesting insights into their abilities and limitations .", "in total , dvd is built from 11k cater synthetic videos and contains 10 instances of 10 - round dialogues for each video , resulting in more than 100k dialogues and 1m question - answer pairs .", "our code and dataset are publicly available ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "APP", "offsets": [ 1, 2, 3, 4, 5 ], "text": "video - grounded dialogue system", "tokens": [ "video", "-", "grounded", "dialogue", "system" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 7 ], "text": "required", "tokens": [ "required" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 56, 57 ], "text": "existing benchmarks", "tokens": [ "existing", "benchmarks" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 59, 60, 61, 62 ], "text": "not have enough annotations", "tokens": [ "not", "have", "enough", "annotations" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 65 ], "text": "analyze", "tokens": [ "analyze" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 59, 60, 61, 62 ], "text": "not have enough annotations", "tokens": [ "not", "have", "enough", "annotations" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "APP", "offsets": [ 66, 67 ], "text": "dialogue systems", "tokens": [ "dialogue", "systems" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 65 ], "text": "analyze", "tokens": [ "analyze" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 74, 75 ], "text": "in isolation", "tokens": [ "in", "isolation" ] }, { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 66, 67, 71 ], "text": "their capabilities", "tokens": [ "dialogue", "systems", "capabilities" ] }, { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 66, 67, 73 ], "text": "their limitations", "tokens": [ "dialogue", "systems", "limitations" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 69 ], "text": "understand", "tokens": [ "understand" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 78 ], "text": "benchmarks", "tokens": [ "benchmarks" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 85, 86 ], "text": "minimise biases", "tokens": [ "minimise", "biases" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 81, 82, 83 ], "text": "not explicitly designed", "tokens": [ "not", "explicitly", "designed" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 104 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 109, 110 ], "text": "diagnostic dataset", "tokens": [ "diagnostic", "dataset" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 112, 113, 114, 115 ], "text": "video - grounded dialogue", "tokens": [ "video", "-", "grounded", "dialogue" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 105 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "DST", "offsets": [ 118 ], "text": "dataset", "tokens": [ "dataset" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 123, 124 ], "text": "minimal biases", "tokens": [ "minimal", "biases" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 127, 128 ], "text": "detailed annotations", "tokens": [ "detailed", "annotations" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 122 ], "text": "contain", "tokens": [ "contain" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 148, 149, 150 ], "text": "multiple question turns", "tokens": [ "multiple", "question", "turns" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 159, 160, 161, 162, 163, 164, 165 ], "text": "set of cross - turn semantic relationships", "tokens": [ "set", "of", "cross", "-", "turn", "semantic", "relationships" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 156 ], "text": "injected", "tokens": [ "injected" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 167 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 176, 177 ], "text": "interesting insights", "tokens": [ "interesting", "insights" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 175 ], "text": "providing", "tokens": [ "providing" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 192, 193, 194 ], "text": "cater synthetic videos", "tokens": [ "cater", "synthetic", "videos" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 200, 201, 202, 203 ], "text": "10 - round dialogues", "tokens": [ "10", "-", "round", "dialogues" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 213 ], "text": "dialogues", "tokens": [ "dialogues" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 216, 217, 218, 219 ], "text": "question - answer pairs", "tokens": [ "question", "-", "answer", "pairs" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 208 ], "text": "resulting", "tokens": [ "resulting" ] } } ]
[ "a", "video", "-", "grounded", "dialogue", "system", "is", "required", "to", "understand", "both", "dialogue", ",", "which", "contains", "semantic", "dependencies", "from", "turn", "to", "turn", ",", "and", "video", ",", "which", "contains", "visual", "cues", "of", "spatial", "and", "temporal", "scene", "variations", ".", "building", "such", "dialogue", "systems", "is", "a", "challenging", "problem", ",", "involving", "various", "reasoning", "types", "on", "both", "visual", "and", "language", "inputs", ".", "existing", "benchmarks", "do", "not", "have", "enough", "annotations", "to", "thoroughly", "analyze", "dialogue", "systems", "and", "understand", "their", "capabilities", "and", "limitations", "in", "isolation", ".", "these", "benchmarks", "are", "also", "not", "explicitly", "designed", "to", "minimise", "biases", "that", "models", "can", "exploit", "without", "actual", "reasoning", ".", "to", "address", "these", "limitations", ",", "in", "this", "paper", ",", "we", "present", "dvd", ",", "a", "diagnostic", "dataset", "for", "video", "-", "grounded", "dialogue", ".", "the", "dataset", "is", "designed", "to", "contain", "minimal", "biases", "and", "has", "detailed", "annotations", "for", "the", "different", "types", "of", "reasoning", "over", "the", "spatio", "-", "temporal", "space", "of", "video", ".", "dialogues", "are", "synthesized", "over", "multiple", "question", "turns", ",", "each", "of", "which", "is", "injected", "with", "a", "set", "of", "cross", "-", "turn", "semantic", "relationships", ".", "we", "use", "dvd", "to", "analyze", "existing", "approaches", ",", "providing", "interesting", "insights", "into", "their", "abilities", "and", "limitations", ".", "in", "total", ",", "dvd", "is", "built", "from", "11k", "cater", "synthetic", "videos", "and", "contains", "10", "instances", "of", "10", "-", "round", "dialogues", "for", "each", "video", ",", "resulting", "in", "more", "than", "100k", "dialogues", "and", "1m", "question", "-", "answer", "pairs", ".", "our", "code", "and", "dataset", "are", "publicly", "available", "." ]
ACL
MATE-KD: Masked Adversarial TExt, a Companion to Knowledge Distillation
The advent of large pre-trained language models has given rise to rapid progress in the field of Natural Language Processing (NLP). While the performance of these models on standard benchmarks has scaled with size, compression techniques such as knowledge distillation have been key in making them practical. We present MATE-KD, a novel text-based adversarial training algorithm which improves the performance of knowledge distillation. MATE-KD first trains a masked language model-based generator to perturb text by maximizing the divergence between teacher and student logits. Then using knowledge distillation a student is trained on both the original and the perturbed training samples. We evaluate our algorithm, using BERT-based models, on the GLUE benchmark and demonstrate that MATE-KD outperforms competitive adversarial learning and data augmentation baselines. On the GLUE test set our 6 layer RoBERTa based model outperforms BERT-large.
bcf2a5086a3b7ab9ae680289f38dad5f
2,021
[ "the advent of large pre - trained language models has given rise to rapid progress in the field of natural language processing ( nlp ) .", "while the performance of these models on standard benchmarks has scaled with size , compression techniques such as knowledge distillation have been key in making them practical .", "we present mate - kd , a novel text - based adversarial training algorithm which improves the performance of knowledge distillation .", "mate - kd first trains a masked language model - based generator to perturb text by maximizing the divergence between teacher and student logits .", "then using knowledge distillation a student is trained on both the original and the perturbed training samples .", "we evaluate our algorithm , using bert - based models , on the glue benchmark and demonstrate that mate - kd outperforms competitive adversarial learning and data augmentation baselines .", "on the glue test set our 6 layer roberta based model outperforms bert - large ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 19, 20, 21 ], "text": "natural language processing", "tokens": [ "natural", "language", "processing" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 11 ], "text": "rise", "tokens": [ "rise" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 54 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 69 ], "text": "improves", "tokens": [ "improves" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 62, 63, 64, 65, 66, 67 ], "text": "text - based adversarial training algorithm", "tokens": [ "text", "-", "based", "adversarial", "training", "algorithm" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 55 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 71, 72, 73, 74 ], "text": "performance of knowledge distillation", "tokens": [ "performance", "of", "knowledge", "distillation" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 69 ], "text": "improves", "tokens": [ "improves" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 89 ], "text": "perturb", "tokens": [ "perturb" ] }, { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 82, 83, 84, 85, 86, 87 ], "text": "masked language model - based generator", "tokens": [ "masked", "language", "model", "-", "based", "generator" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 94, 95, 96, 97, 98, 99 ], "text": "divergence between teacher and student logits", "tokens": [ "divergence", "between", "teacher", "and", "student", "logits" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 92 ], "text": "maximizing", "tokens": [ "maximizing" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 90 ], "text": "text", "tokens": [ "text" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 89 ], "text": "perturb", "tokens": [ "perturb" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 102, 103, 104 ], "text": "using knowledge distillation", "tokens": [ "using", "knowledge", "distillation" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 106 ], "text": "student", "tokens": [ "student" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 112, 116, 117 ], "text": "original training samples", "tokens": [ "original", "training", "samples" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 115, 116, 117 ], "text": "perturbed training samples", "tokens": [ "perturbed", "training", "samples" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 108 ], "text": "trained", "tokens": [ "trained" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 137, 138, 139 ], "text": "mate - kd", "tokens": [ "mate", "-", "kd" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 140 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 141, 142, 143 ], "text": "competitive adversarial learning", "tokens": [ "competitive", "adversarial", "learning" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 140 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 154, 155, 156, 157, 158, 159 ], "text": "our 6 layer roberta based model", "tokens": [ "our", "6", "layer", "roberta", "based", "model" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 160 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 161, 162, 163 ], "text": "bert - large", "tokens": [ "bert", "-", "large" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 160 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 119 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 125, 126, 127, 128 ], "text": "bert - based models", "tokens": [ "bert", "-", "based", "models" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 120 ], "text": "evaluate", "tokens": [ "evaluate" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 130, 131, 132, 133 ], "text": "on the glue benchmark", "tokens": [ "on", "the", "glue", "benchmark" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 124 ], "text": "using", "tokens": [ "using" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "APP", "offsets": [ 62, 63, 64, 65, 66, 67 ], "text": "text - based adversarial training algorithm", "tokens": [ "text", "-", "based", "adversarial", "training", "algorithm" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 120 ], "text": "evaluate", "tokens": [ "evaluate" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 140 ], "text": "outperforms", "tokens": [ "outperforms" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 135 ], "text": "demonstrate", "tokens": [ "demonstrate" ] } } ]
[ "the", "advent", "of", "large", "pre", "-", "trained", "language", "models", "has", "given", "rise", "to", "rapid", "progress", "in", "the", "field", "of", "natural", "language", "processing", "(", "nlp", ")", ".", "while", "the", "performance", "of", "these", "models", "on", "standard", "benchmarks", "has", "scaled", "with", "size", ",", "compression", "techniques", "such", "as", "knowledge", "distillation", "have", "been", "key", "in", "making", "them", "practical", ".", "we", "present", "mate", "-", "kd", ",", "a", "novel", "text", "-", "based", "adversarial", "training", "algorithm", "which", "improves", "the", "performance", "of", "knowledge", "distillation", ".", "mate", "-", "kd", "first", "trains", "a", "masked", "language", "model", "-", "based", "generator", "to", "perturb", "text", "by", "maximizing", "the", "divergence", "between", "teacher", "and", "student", "logits", ".", "then", "using", "knowledge", "distillation", "a", "student", "is", "trained", "on", "both", "the", "original", "and", "the", "perturbed", "training", "samples", ".", "we", "evaluate", "our", "algorithm", ",", "using", "bert", "-", "based", "models", ",", "on", "the", "glue", "benchmark", "and", "demonstrate", "that", "mate", "-", "kd", "outperforms", "competitive", "adversarial", "learning", "and", "data", "augmentation", "baselines", ".", "on", "the", "glue", "test", "set", "our", "6", "layer", "roberta", "based", "model", "outperforms", "bert", "-", "large", "." ]
ACL
An Automated Framework for Fast Cognate Detection and Bayesian Phylogenetic Inference in Computational Historical Linguistics
We present a fully automated workflow for phylogenetic reconstruction on large datasets, consisting of two novel methods, one for fast detection of cognates and one for fast Bayesian phylogenetic inference. Our results show that the methods take less than a few minutes to process language families that have so far required large amounts of time and computational power. Moreover, the cognates and the trees inferred from the method are quite close, both to gold standard cognate judgments and to expert language family trees. Given its speed and ease of application, our framework is specifically useful for the exploration of very large datasets in historical linguistics.
db2fff29a55036937a41cdace0266be9
2,019
[ "we present a fully automated workflow for phylogenetic reconstruction on large datasets , consisting of two novel methods , one for fast detection of cognates and one for fast bayesian phylogenetic inference .", "our results show that the methods take less than a few minutes to process language families that have so far required large amounts of time and computational power .", "moreover , the cognates and the trees inferred from the method are quite close , both to gold standard cognate judgments and to expert language family trees .", "given its speed and ease of application , our framework is specifically useful for the exploration of very large datasets in historical linguistics ." ]
[ { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 0 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 3, 4, 5 ], "text": "fully automated workflow", "tokens": [ "fully", "automated", "workflow" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 7, 8 ], "text": "phylogenetic reconstruction", "tokens": [ "phylogenetic", "reconstruction" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 1 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 39 ], "text": "take", "tokens": [ "take" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 35 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 15, 16, 17 ], "text": "two novel methods", "tokens": [ "two", "novel", "methods" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 46 ], "text": "process", "tokens": [ "process" ] }, { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 40, 41, 42, 43, 44 ], "text": "less than a few minutes", "tokens": [ "less", "than", "a", "few", "minutes" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 39 ], "text": "take", "tokens": [ "take" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 47, 48 ], "text": "language families", "tokens": [ "language", "families" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 46 ], "text": "process", "tokens": [ "process" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 79, 80, 81, 82 ], "text": "gold standard cognate judgments", "tokens": [ "gold", "standard", "cognate", "judgments" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 85, 86, 87, 88 ], "text": "expert language family trees", "tokens": [ "expert", "language", "family", "trees" ] }, { "argument_type": "Subject", "nugget_type": "FEA", "offsets": [ 65 ], "text": "cognates", "tokens": [ "cognates" ] }, { "argument_type": "Subject", "nugget_type": "FEA", "offsets": [ 68 ], "text": "trees", "tokens": [ "trees" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 74, 75 ], "text": "quite close", "tokens": [ "quite", "close" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 3, 4, 5 ], "text": "fully automated workflow", "tokens": [ "fully", "automated", "workflow" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 110, 111, 112 ], "text": "in historical linguistics", "tokens": [ "in", "historical", "linguistics" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 105, 106, 107, 108, 109 ], "text": "exploration of very large datasets", "tokens": [ "exploration", "of", "very", "large", "datasets" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 101, 102 ], "text": "specifically useful", "tokens": [ "specifically", "useful" ] } } ]
[ "we", "present", "a", "fully", "automated", "workflow", "for", "phylogenetic", "reconstruction", "on", "large", "datasets", ",", "consisting", "of", "two", "novel", "methods", ",", "one", "for", "fast", "detection", "of", "cognates", "and", "one", "for", "fast", "bayesian", "phylogenetic", "inference", ".", "our", "results", "show", "that", "the", "methods", "take", "less", "than", "a", "few", "minutes", "to", "process", "language", "families", "that", "have", "so", "far", "required", "large", "amounts", "of", "time", "and", "computational", "power", ".", "moreover", ",", "the", "cognates", "and", "the", "trees", "inferred", "from", "the", "method", "are", "quite", "close", ",", "both", "to", "gold", "standard", "cognate", "judgments", "and", "to", "expert", "language", "family", "trees", ".", "given", "its", "speed", "and", "ease", "of", "application", ",", "our", "framework", "is", "specifically", "useful", "for", "the", "exploration", "of", "very", "large", "datasets", "in", "historical", "linguistics", "." ]
ACL
Faithful or Extractive? On Mitigating the Faithfulness-Abstractiveness Trade-off in Abstractive Summarization
Despite recent progress in abstractive summarization, systems still suffer from faithfulness errors. While prior work has proposed models that improve faithfulness, it is unclear whether the improvement comes from an increased level of extractiveness of the model outputs as one naive way to improve faithfulness is to make summarization models more extractive. In this work, we present a framework for evaluating the effective faithfulness of summarization systems, by generating a faithfulness-abstractiveness trade-off curve that serves as a control at different operating points on the abstractiveness spectrum. We then show that the Maximum Likelihood Estimation (MLE) baseline as well as recently proposed methods for improving faithfulness, fail to consistently improve over the control at the same level of abstractiveness. Finally, we learn a selector to identify the most faithful and abstractive summary for a given document, and show that this system can attain higher faithfulness scores in human evaluations while being more abstractive than the baseline system on two datasets. Moreover, we show that our system is able to achieve a better faithfulness-abstractiveness trade-off than the control at the same level of abstractiveness.
959fbecee82a093efd41a9a4608a4728
2,022
[ "despite recent progress in abstractive summarization , systems still suffer from faithfulness errors .", "while prior work has proposed models that improve faithfulness , it is unclear whether the improvement comes from an increased level of extractiveness of the model outputs as one naive way to improve faithfulness is to make summarization models more extractive .", "in this work , we present a framework for evaluating the effective faithfulness of summarization systems , by generating a faithfulness - abstractiveness trade - off curve that serves as a control at different operating points on the abstractiveness spectrum .", "we then show that the maximum likelihood estimation ( mle ) baseline as well as recently proposed methods for improving faithfulness , fail to consistently improve over the control at the same level of abstractiveness .", "finally , we learn a selector to identify the most faithful and abstractive summary for a given document , and show that this system can attain higher faithfulness scores in human evaluations while being more abstractive than the baseline system on two datasets .", "moreover , we show that our system is able to achieve a better faithfulness - abstractiveness trade - off than the control at the same level of abstractiveness ." ]
[ { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 7 ], "text": "systems", "tokens": [ "systems" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 11, 12 ], "text": "faithfulness errors", "tokens": [ "faithfulness", "errors" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 9 ], "text": "suffer", "tokens": [ "suffer" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 60 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 63 ], "text": "framework", "tokens": [ "framework" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 65 ], "text": "evaluating", "tokens": [ "evaluating" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 61 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 67, 68, 69, 70, 71 ], "text": "effective faithfulness of summarization systems", "tokens": [ "effective", "faithfulness", "of", "summarization", "systems" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 65 ], "text": "evaluating", "tokens": [ "evaluating" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 76, 77, 78, 79, 80, 81, 82 ], "text": "faithfulness - abstractiveness trade - off curve", "tokens": [ "faithfulness", "-", "abstractiveness", "trade", "-", "off", "curve" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 87, 88, 89, 90, 91, 92, 93, 94, 95 ], "text": "control at different operating points on the abstractiveness spectrum", "tokens": [ "control", "at", "different", "operating", "points", "on", "the", "abstractiveness", "spectrum" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 74 ], "text": "generating", "tokens": [ "generating" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 102, 103, 104, 105, 106, 107, 108 ], "text": "maximum likelihood estimation ( mle ) baseline", "tokens": [ "maximum", "likelihood", "estimation", "(", "mle", ")", "baseline" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 112, 113, 114 ], "text": "recently proposed methods", "tokens": [ "recently", "proposed", "methods" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 125 ], "text": "control", "tokens": [ "control" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 126, 127, 128, 129, 130, 131 ], "text": "at the same level of abstractiveness", "tokens": [ "at", "the", "same", "level", "of", "abstractiveness" ] }, { "argument_type": "Result", "nugget_type": "WEA", "offsets": [ 119 ], "text": "fail", "tokens": [ "fail" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 122 ], "text": "improve", "tokens": [ "improve" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 135 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 138 ], "text": "selector", "tokens": [ "selector" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 140 ], "text": "identify", "tokens": [ "identify" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 136 ], "text": "learn", "tokens": [ "learn" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 142, 143, 144, 145, 146 ], "text": "most faithful and abstractive summary", "tokens": [ "most", "faithful", "and", "abstractive", "summary" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 147, 148, 149, 150 ], "text": "for a given document", "tokens": [ "for", "a", "given", "document" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 140 ], "text": "identify", "tokens": [ "identify" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "TAK", "offsets": [ 159, 160, 161 ], "text": "higher faithfulness scores", "tokens": [ "higher", "faithfulness", "scores" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 163, 164 ], "text": "human evaluations", "tokens": [ "human", "evaluations" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 158 ], "text": "attain", "tokens": [ "attain" ] } }, { "arguments": [ { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 171, 172 ], "text": "baseline system", "tokens": [ "baseline", "system" ] }, { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 174, 175 ], "text": "two datasets", "tokens": [ "two", "datasets" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 167, 168 ], "text": "more abstractive", "tokens": [ "more", "abstractive" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 179 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 187 ], "text": "achieve", "tokens": [ "achieve" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 180 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 189, 190, 191, 192, 193, 194, 195 ], "text": "better faithfulness - abstractiveness trade - off", "tokens": [ "better", "faithfulness", "-", "abstractiveness", "trade", "-", "off" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 198 ], "text": "control", "tokens": [ "control" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 199, 200, 201, 202, 203, 204 ], "text": "at the same level of abstractiveness", "tokens": [ "at", "the", "same", "level", "of", "abstractiveness" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 187 ], "text": "achieve", "tokens": [ "achieve" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 4, 5 ], "text": "abstractive summarization", "tokens": [ "abstractive", "summarization" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 2 ], "text": "progress", "tokens": [ "progress" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 97 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 122 ], "text": "improve", "tokens": [ "improve" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 99 ], "text": "show", "tokens": [ "show" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 167, 168 ], "text": "more abstractive", "tokens": [ "more", "abstractive" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 153 ], "text": "show", "tokens": [ "show" ] } } ]
[ "despite", "recent", "progress", "in", "abstractive", "summarization", ",", "systems", "still", "suffer", "from", "faithfulness", "errors", ".", "while", "prior", "work", "has", "proposed", "models", "that", "improve", "faithfulness", ",", "it", "is", "unclear", "whether", "the", "improvement", "comes", "from", "an", "increased", "level", "of", "extractiveness", "of", "the", "model", "outputs", "as", "one", "naive", "way", "to", "improve", "faithfulness", "is", "to", "make", "summarization", "models", "more", "extractive", ".", "in", "this", "work", ",", "we", "present", "a", "framework", "for", "evaluating", "the", "effective", "faithfulness", "of", "summarization", "systems", ",", "by", "generating", "a", "faithfulness", "-", "abstractiveness", "trade", "-", "off", "curve", "that", "serves", "as", "a", "control", "at", "different", "operating", "points", "on", "the", "abstractiveness", "spectrum", ".", "we", "then", "show", "that", "the", "maximum", "likelihood", "estimation", "(", "mle", ")", "baseline", "as", "well", "as", "recently", "proposed", "methods", "for", "improving", "faithfulness", ",", "fail", "to", "consistently", "improve", "over", "the", "control", "at", "the", "same", "level", "of", "abstractiveness", ".", "finally", ",", "we", "learn", "a", "selector", "to", "identify", "the", "most", "faithful", "and", "abstractive", "summary", "for", "a", "given", "document", ",", "and", "show", "that", "this", "system", "can", "attain", "higher", "faithfulness", "scores", "in", "human", "evaluations", "while", "being", "more", "abstractive", "than", "the", "baseline", "system", "on", "two", "datasets", ".", "moreover", ",", "we", "show", "that", "our", "system", "is", "able", "to", "achieve", "a", "better", "faithfulness", "-", "abstractiveness", "trade", "-", "off", "than", "the", "control", "at", "the", "same", "level", "of", "abstractiveness", "." ]
ACL
Enhancing the generalization for Intent Classification and Out-of-Domain Detection in SLU
Intent classification is a major task in spoken language understanding (SLU). Since most models are built with pre-collected in-domain (IND) training utterances, their ability to detect unsupported out-of-domain (OOD) utterances has a critical effect in practical use. Recent works have shown that using extra data and labels can improve the OOD detection performance, yet it could be costly to collect such data. This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection. Our method designs a novel domain-regularized module (DRM) to reduce the overconfident phenomenon of a vanilla classifier, achieving a better generalization in both cases. Besides, DRM can be used as a drop-in replacement for the last layer in any neural network-based intent classifier, providing a low-cost strategy for a significant improvement. The evaluation on four datasets shows that our method built on BERT and RoBERTa models achieves state-of-the-art performance against existing approaches and the strong baselines we created for the comparisons.
bbe87249393bc09725f2b0dcfda04997
2,021
[ "intent classification is a major task in spoken language understanding ( slu ) .", "since most models are built with pre - collected in - domain ( ind ) training utterances , their ability to detect unsupported out - of - domain ( ood ) utterances has a critical effect in practical use .", "recent works have shown that using extra data and labels can improve the ood detection performance , yet it could be costly to collect such data .", "this paper proposes to train a model with only ind data while supporting both ind intent classification and ood detection .", "our method designs a novel domain - regularized module ( drm ) to reduce the overconfident phenomenon of a vanilla classifier , achieving a better generalization in both cases .", "besides , drm can be used as a drop - in replacement for the last layer in any neural network - based intent classifier , providing a low - cost strategy for a significant improvement .", "the evaluation on four datasets shows that our method built on bert and roberta models achieves state - of - the - art performance against existing approaches and the strong baselines we created for the comparisons ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1 ], "text": "intent classification", "tokens": [ "intent", "classification" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 5 ], "text": "task", "tokens": [ "task" ] } }, { "arguments": [ { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 75 ], "text": "costly", "tokens": [ "costly" ] }, { "argument_type": "Concern", "nugget_type": "FEA", "offsets": [ 60, 61 ], "text": "extra data", "tokens": [ "extra", "data" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 75 ], "text": "costly", "tokens": [ "costly" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 87, 88, 89, 90, 91 ], "text": "model with only ind data", "tokens": [ "model", "with", "only", "ind", "data" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 93, 94, 95, 96, 97, 98, 99, 100 ], "text": "supporting both ind intent classification and ood detection", "tokens": [ "supporting", "both", "ind", "intent", "classification", "and", "ood", "detection" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 85 ], "text": "train", "tokens": [ "train" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 106, 107, 108, 109, 110 ], "text": "novel domain - regularized module", "tokens": [ "novel", "domain", "-", "regularized", "module" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 115 ], "text": "reduce", "tokens": [ "reduce" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 104 ], "text": "designs", "tokens": [ "designs" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 117, 118 ], "text": "overconfident phenomenon", "tokens": [ "overconfident", "phenomenon" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 115 ], "text": "reduce", "tokens": [ "reduce" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 128, 129, 130 ], "text": "in both cases", "tokens": [ "in", "both", "cases" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 126, 127 ], "text": "better generalization", "tokens": [ "better", "generalization" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 124 ], "text": "achieving", "tokens": [ "achieving" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "MOD", "offsets": [ 146, 147, 148, 149, 150, 151, 152, 153, 154, 155 ], "text": "last layer in any neural network - based intent classifier", "tokens": [ "last", "layer", "in", "any", "neural", "network", "-", "based", "intent", "classifier" ] }, { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 107, 108, 109, 110 ], "text": "domain - regularized module", "tokens": [ "domain", "-", "regularized", "module" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 137 ], "text": "used", "tokens": [ "used" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 183 ], "text": "achieves", "tokens": [ "achieves" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 173 ], "text": "shows", "tokens": [ "shows" ] } }, { "arguments": [ { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 193, 194 ], "text": "existing approaches", "tokens": [ "existing", "approaches" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 176, 177, 178, 179, 180, 181, 182 ], "text": "method built on bert and roberta models", "tokens": [ "method", "built", "on", "bert", "and", "roberta", "models" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 197, 198 ], "text": "strong baselines", "tokens": [ "strong", "baselines" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 184, 185, 186, 187, 188, 189, 190, 191 ], "text": "state - of - the - art performance", "tokens": [ "state", "-", "of", "-", "the", "-", "art", "performance" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 183 ], "text": "achieves", "tokens": [ "achieves" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 99, 100 ], "text": "ood detection", "tokens": [ "ood", "detection" ] }, { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 95, 96, 97 ], "text": "ind intent classification", "tokens": [ "ind", "intent", "classification" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 93 ], "text": "supporting", "tokens": [ "supporting" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 15, 16 ], "text": "most models", "tokens": [ "most", "models" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30 ], "text": "pre - collected in - domain ( ind ) training utterances", "tokens": [ "pre", "-", "collected", "in", "-", "domain", "(", "ind", ")", "training", "utterances" ] } ], "event_type": "RWS", "trigger": { "offsets": [ 18 ], "text": "built", "tokens": [ "built" ] } } ]
[ "intent", "classification", "is", "a", "major", "task", "in", "spoken", "language", "understanding", "(", "slu", ")", ".", "since", "most", "models", "are", "built", "with", "pre", "-", "collected", "in", "-", "domain", "(", "ind", ")", "training", "utterances", ",", "their", "ability", "to", "detect", "unsupported", "out", "-", "of", "-", "domain", "(", "ood", ")", "utterances", "has", "a", "critical", "effect", "in", "practical", "use", ".", "recent", "works", "have", "shown", "that", "using", "extra", "data", "and", "labels", "can", "improve", "the", "ood", "detection", "performance", ",", "yet", "it", "could", "be", "costly", "to", "collect", "such", "data", ".", "this", "paper", "proposes", "to", "train", "a", "model", "with", "only", "ind", "data", "while", "supporting", "both", "ind", "intent", "classification", "and", "ood", "detection", ".", "our", "method", "designs", "a", "novel", "domain", "-", "regularized", "module", "(", "drm", ")", "to", "reduce", "the", "overconfident", "phenomenon", "of", "a", "vanilla", "classifier", ",", "achieving", "a", "better", "generalization", "in", "both", "cases", ".", "besides", ",", "drm", "can", "be", "used", "as", "a", "drop", "-", "in", "replacement", "for", "the", "last", "layer", "in", "any", "neural", "network", "-", "based", "intent", "classifier", ",", "providing", "a", "low", "-", "cost", "strategy", "for", "a", "significant", "improvement", ".", "the", "evaluation", "on", "four", "datasets", "shows", "that", "our", "method", "built", "on", "bert", "and", "roberta", "models", "achieves", "state", "-", "of", "-", "the", "-", "art", "performance", "against", "existing", "approaches", "and", "the", "strong", "baselines", "we", "created", "for", "the", "comparisons", "." ]
ACL
PromDA: Prompt-based Data Augmentation for Low-Resource NLU Tasks
This paper focuses on the Data Augmentation for low-resource Natural Language Understanding (NLU) tasks. We propose Prompt-based Data Augmentation model (PromDA) which only trains small-scale Soft Prompt (i.e., a set of trainable vectors) in the frozen Pre-trained Language Models (PLMs). This avoids human effort in collecting unlabeled in-domain data and maintains the quality of generated synthetic data. In addition, PromDA generates synthetic data via two different views and filters out the low-quality data using NLU models. Experiments on four benchmarks show that synthetic data produced by PromDA successfully boost up the performance of NLU models which consistently outperform several competitive baseline models, including a state-of-the-art semi-supervised model using unlabeled in-domain data. The synthetic data from PromDA are also complementary with unlabeled in-domain data. The NLU models can be further improved when they are combined for training.
04e5995f7999d5daad821408248f8262
2,022
[ "this paper focuses on the data augmentation for low - resource natural language understanding ( nlu ) tasks .", "we propose prompt - based data augmentation model ( promda ) which only trains small - scale soft prompt ( i . e . , a set of trainable vectors ) in the frozen pre - trained language models ( plms ) .", "this avoids human effort in collecting unlabeled in - domain data and maintains the quality of generated synthetic data .", "in addition , promda generates synthetic data via two different views and filters out the low - quality data using nlu models .", "experiments on four benchmarks show that synthetic data produced by promda successfully boost up the performance of nlu models which consistently outperform several competitive baseline models , including a state - of - the - art semi - supervised model using unlabeled in - domain data .", "the synthetic data from promda are also complementary with unlabeled in - domain data .", "the nlu models can be further improved when they are combined for training ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 8, 9, 10, 11, 12, 13, 17 ], "text": "low - resource natural language understanding ( nlu ) tasks", "tokens": [ "low", "-", "resource", "natural", "language", "understanding", "tasks" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 2 ], "text": "focuses", "tokens": [ "focuses" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 19 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 21, 22, 23, 24, 25, 26 ], "text": "prompt - based data augmentation model", "tokens": [ "prompt", "-", "based", "data", "augmentation", "model" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 20 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 33, 34, 35, 36, 37 ], "text": "small - scale soft prompt", "tokens": [ "small", "-", "scale", "soft", "prompt" ] }, { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 52, 53, 54, 55, 56, 57 ], "text": "frozen pre - trained language models", "tokens": [ "frozen", "pre", "-", "trained", "language", "models" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 32 ], "text": "trains", "tokens": [ "trains" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 64, 65 ], "text": "human effort", "tokens": [ "human", "effort" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 67, 68 ], "text": "collecting unlabeled", "tokens": [ "collecting", "unlabeled" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 69, 70, 71, 72 ], "text": "in - domain data", "tokens": [ "in", "-", "domain", "data" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 63 ], "text": "avoids", "tokens": [ "avoids" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 76, 77, 78, 79, 80 ], "text": "quality of generated synthetic data", "tokens": [ "quality", "of", "generated", "synthetic", "data" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 74 ], "text": "maintains", "tokens": [ "maintains" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "DST", "offsets": [ 87, 88 ], "text": "synthetic data", "tokens": [ "synthetic", "data" ] }, { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 21, 22, 23, 24, 25, 26 ], "text": "prompt - based data augmentation model", "tokens": [ "prompt", "-", "based", "data", "augmentation", "model" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 89, 90, 91, 92 ], "text": "via two different views", "tokens": [ "via", "two", "different", "views" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 86 ], "text": "generates", "tokens": [ "generates" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "DST", "offsets": [ 97, 98, 99, 100 ], "text": "low - quality data", "tokens": [ "low", "-", "quality", "data" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 101, 11, 12, 13, 103 ], "text": "using nlu models", "tokens": [ "using", "natural", "language", "understanding", "models" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 94, 95 ], "text": "filters out", "tokens": [ "filters", "out" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 120, 121, 11, 12, 13, 123 ], "text": "performance of nlu models", "tokens": [ "performance", "of", "natural", "language", "understanding", "models" ] }, { "argument_type": "Subject", "nugget_type": "DST", "offsets": [ 111, 112 ], "text": "synthetic data", "tokens": [ "synthetic", "data" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 113, 114, 115 ], "text": "produced by promda", "tokens": [ "produced", "by", "promda" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 117 ], "text": "boost", "tokens": [ "boost" ] } }, { "arguments": [ { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 128, 129, 130 ], "text": "competitive baseline models", "tokens": [ "competitive", "baseline", "models" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 126 ], "text": "outperform", "tokens": [ "outperform" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 125 ], "text": "consistently", "tokens": [ "consistently" ] }, { "argument_type": "Arg1", "nugget_type": "DST", "offsets": [ 111, 112 ], "text": "synthetic data", "tokens": [ "synthetic", "data" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 126 ], "text": "outperform", "tokens": [ "outperform" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "DST", "offsets": [ 153, 154 ], "text": "synthetic data", "tokens": [ "synthetic", "data" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 159 ], "text": "complementary", "tokens": [ "complementary" ] } }, { "arguments": [ { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 174, 153, 154, 160, 161, 162, 163, 164, 165, 176, 177, 178, 179 ], "text": "when they are combined for training", "tokens": [ "when", "synthetic", "data", "with", "unlabeled", "in", "-", "domain", "data", "are", "combined", "for", "training" ] }, { "argument_type": "Object", "nugget_type": "APP", "offsets": [ 11, 12, 13, 169 ], "text": "nlu models", "tokens": [ "natural", "language", "understanding", "models" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 173 ], "text": "improved", "tokens": [ "improved" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 117 ], "text": "boost", "tokens": [ "boost" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 126 ], "text": "outperform", "tokens": [ "outperform" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 109 ], "text": "show", "tokens": [ "show" ] } } ]
[ "this", "paper", "focuses", "on", "the", "data", "augmentation", "for", "low", "-", "resource", "natural", "language", "understanding", "(", "nlu", ")", "tasks", ".", "we", "propose", "prompt", "-", "based", "data", "augmentation", "model", "(", "promda", ")", "which", "only", "trains", "small", "-", "scale", "soft", "prompt", "(", "i", ".", "e", ".", ",", "a", "set", "of", "trainable", "vectors", ")", "in", "the", "frozen", "pre", "-", "trained", "language", "models", "(", "plms", ")", ".", "this", "avoids", "human", "effort", "in", "collecting", "unlabeled", "in", "-", "domain", "data", "and", "maintains", "the", "quality", "of", "generated", "synthetic", "data", ".", "in", "addition", ",", "promda", "generates", "synthetic", "data", "via", "two", "different", "views", "and", "filters", "out", "the", "low", "-", "quality", "data", "using", "nlu", "models", ".", "experiments", "on", "four", "benchmarks", "show", "that", "synthetic", "data", "produced", "by", "promda", "successfully", "boost", "up", "the", "performance", "of", "nlu", "models", "which", "consistently", "outperform", "several", "competitive", "baseline", "models", ",", "including", "a", "state", "-", "of", "-", "the", "-", "art", "semi", "-", "supervised", "model", "using", "unlabeled", "in", "-", "domain", "data", ".", "the", "synthetic", "data", "from", "promda", "are", "also", "complementary", "with", "unlabeled", "in", "-", "domain", "data", ".", "the", "nlu", "models", "can", "be", "further", "improved", "when", "they", "are", "combined", "for", "training", "." ]
ACL
Analyzing the Limitations of Cross-lingual Word Embedding Mappings
Recent research in cross-lingual word embeddings has almost exclusively focused on offline methods, which independently train word embeddings in different languages and map them to a shared space through linear transformations. While several authors have questioned the underlying isomorphism assumption, which states that word embeddings in different languages have approximately the same structure, it is not clear whether this is an inherent limitation of mapping approaches or a more general issue when learning cross-lingual embeddings. So as to answer this question, we experiment with parallel corpora, which allows us to compare offline mapping to an extension of skip-gram that jointly learns both embedding spaces. We observe that, under these ideal conditions, joint learning yields to more isomorphic embeddings, is less sensitive to hubness, and obtains stronger results in bilingual lexicon induction. We thus conclude that current mapping methods do have strong limitations, calling for further research to jointly learn cross-lingual embeddings with a weaker cross-lingual signal.
c84823d8450a619b600d844943a96c1e
2,019
[ "recent research in cross - lingual word embeddings has almost exclusively focused on offline methods , which independently train word embeddings in different languages and map them to a shared space through linear transformations .", "while several authors have questioned the underlying isomorphism assumption , which states that word embeddings in different languages have approximately the same structure , it is not clear whether this is an inherent limitation of mapping approaches or a more general issue when learning cross - lingual embeddings .", "so as to answer this question , we experiment with parallel corpora , which allows us to compare offline mapping to an extension of skip - gram that jointly learns both embedding spaces .", "we observe that , under these ideal conditions , joint learning yields to more isomorphic embeddings , is less sensitive to hubness , and obtains stronger results in bilingual lexicon induction .", "we thus conclude that current mapping methods do have strong limitations , calling for further research to jointly learn cross - lingual embeddings with a weaker cross - lingual signal ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 3, 4, 5, 6, 7 ], "text": "cross - lingual word embeddings", "tokens": [ "cross", "-", "lingual", "word", "embeddings" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 11 ], "text": "focused", "tokens": [ "focused" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 102, 103 ], "text": "offline mapping", "tokens": [ "offline", "mapping" ] }, { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 106, 107, 108, 109, 110 ], "text": "extension of skip - gram", "tokens": [ "extension", "of", "skip", "-", "gram" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 101 ], "text": "compare", "tokens": [ "compare" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 155, 156 ], "text": "mapping methods", "tokens": [ "mapping", "methods" ] }, { "argument_type": "Object", "nugget_type": "WEA", "offsets": [ 159, 160 ], "text": "strong limitations", "tokens": [ "strong", "limitations" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 158 ], "text": "have", "tokens": [ "have" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 91 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "DST", "offsets": [ 94, 95 ], "text": "parallel corpora", "tokens": [ "parallel", "corpora" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 92 ], "text": "experiment", "tokens": [ "experiment" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 118 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 129 ], "text": "yields", "tokens": [ "yields" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 119 ], "text": "observe", "tokens": [ "observe" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 127, 128 ], "text": "joint learning", "tokens": [ "joint", "learning" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 131 ], "text": "more", "tokens": [ "more" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 132, 133 ], "text": "isomorphic embeddings", "tokens": [ "isomorphic", "embeddings" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 129 ], "text": "yields", "tokens": [ "yields" ] } }, { "arguments": [ { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 139 ], "text": "hubness", "tokens": [ "hubness" ] }, { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 127, 128 ], "text": "joint learning", "tokens": [ "joint", "learning" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 136 ], "text": "less", "tokens": [ "less" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 136 ], "text": "less", "tokens": [ "less" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 127, 128 ], "text": "joint learning", "tokens": [ "joint", "learning" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 143, 144 ], "text": "stronger results", "tokens": [ "stronger", "results" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 146, 147, 148 ], "text": "bilingual lexicon induction", "tokens": [ "bilingual", "lexicon", "induction" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 142 ], "text": "obtains", "tokens": [ "obtains" ] } }, { "arguments": [ { "argument_type": "Finder", "nugget_type": "OG", "offsets": [ 150 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "E-FAC", "offsets": [ 158 ], "text": "have", "tokens": [ "have" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 152 ], "text": "conclude", "tokens": [ "conclude" ] } } ]
[ "recent", "research", "in", "cross", "-", "lingual", "word", "embeddings", "has", "almost", "exclusively", "focused", "on", "offline", "methods", ",", "which", "independently", "train", "word", "embeddings", "in", "different", "languages", "and", "map", "them", "to", "a", "shared", "space", "through", "linear", "transformations", ".", "while", "several", "authors", "have", "questioned", "the", "underlying", "isomorphism", "assumption", ",", "which", "states", "that", "word", "embeddings", "in", "different", "languages", "have", "approximately", "the", "same", "structure", ",", "it", "is", "not", "clear", "whether", "this", "is", "an", "inherent", "limitation", "of", "mapping", "approaches", "or", "a", "more", "general", "issue", "when", "learning", "cross", "-", "lingual", "embeddings", ".", "so", "as", "to", "answer", "this", "question", ",", "we", "experiment", "with", "parallel", "corpora", ",", "which", "allows", "us", "to", "compare", "offline", "mapping", "to", "an", "extension", "of", "skip", "-", "gram", "that", "jointly", "learns", "both", "embedding", "spaces", ".", "we", "observe", "that", ",", "under", "these", "ideal", "conditions", ",", "joint", "learning", "yields", "to", "more", "isomorphic", "embeddings", ",", "is", "less", "sensitive", "to", "hubness", ",", "and", "obtains", "stronger", "results", "in", "bilingual", "lexicon", "induction", ".", "we", "thus", "conclude", "that", "current", "mapping", "methods", "do", "have", "strong", "limitations", ",", "calling", "for", "further", "research", "to", "jointly", "learn", "cross", "-", "lingual", "embeddings", "with", "a", "weaker", "cross", "-", "lingual", "signal", "." ]
ACL
Dependency-driven Relation Extraction with Attentive Graph Convolutional Networks
Syntactic information, especially dependency trees, has been widely used by existing studies to improve relation extraction with better semantic guidance for analyzing the context information associated with the given entities. However, most existing studies suffer from the noise in the dependency trees, especially when they are automatically generated, so that intensively leveraging dependency information may introduce confusions to relation classification and necessary pruning is of great importance in this task. In this paper, we propose a dependency-driven approach for relation extraction with attentive graph convolutional networks (A-GCN). In this approach, an attention mechanism upon graph convolutional networks is applied to different contextual words in the dependency tree obtained from an off-the-shelf dependency parser, to distinguish the importance of different word dependencies. Consider that dependency types among words also contain important contextual guidance, which is potentially helpful for relation extraction, we also include the type information in A-GCN modeling. Experimental results on two English benchmark datasets demonstrate the effectiveness of our A-GCN, which outperforms previous studies and achieves state-of-the-art performance on both datasets.
09e8a58fe50453a2401747d5e9c40e18
2,021
[ "syntactic information , especially dependency trees , has been widely used by existing studies to improve relation extraction with better semantic guidance for analyzing the context information associated with the given entities .", "however , most existing studies suffer from the noise in the dependency trees , especially when they are automatically generated , so that intensively leveraging dependency information may introduce confusions to relation classification and necessary pruning is of great importance in this task .", "in this paper , we propose a dependency - driven approach for relation extraction with attentive graph convolutional networks ( a - gcn ) .", "in this approach , an attention mechanism upon graph convolutional networks is applied to different contextual words in the dependency tree obtained from an off - the - shelf dependency parser , to distinguish the importance of different word dependencies .", "consider that dependency types among words also contain important contextual guidance , which is potentially helpful for relation extraction , we also include the type information in a - gcn modeling .", "experimental results on two english benchmark datasets demonstrate the effectiveness of our a - gcn , which outperforms previous studies and achieves state - of - the - art performance on both datasets ." ]
[ { "arguments": [], "event_type": "ITT", "trigger": { "offsets": [ 10 ], "text": "used", "tokens": [ "used" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 36, 37 ], "text": "existing studies", "tokens": [ "existing", "studies" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 41 ], "text": "noise", "tokens": [ "noise" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 42, 43, 44, 45 ], "text": "in the dependency trees", "tokens": [ "in", "the", "dependency", "trees" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 38 ], "text": "suffer", "tokens": [ "suffer" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 81 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 84, 85, 86, 87 ], "text": "dependency - driven approach", "tokens": [ "dependency", "-", "driven", "approach" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 89, 90 ], "text": "relation extraction", "tokens": [ "relation", "extraction" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 82 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "E-CMP", "offsets": [ 192 ], "text": "outperforms", "tokens": [ "outperforms" ] } ], "event_type": "FIN", "trigger": { "offsets": [ 182 ], "text": "demonstrate", "tokens": [ "demonstrate" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 187, 188, 189 ], "text": "attentive graph convolutional networks", "tokens": [ "a", "-", "gcn" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 193, 194 ], "text": "previous studies", "tokens": [ "previous", "studies" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 192 ], "text": "outperforms", "tokens": [ "outperforms" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 192 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 187, 188, 189 ], "text": "attentive graph convolutional networks", "tokens": [ "a", "-", "gcn" ] }, { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 197, 198, 199, 200, 201, 202, 203, 204 ], "text": "state - of - the - art performance", "tokens": [ "state", "-", "of", "-", "the", "-", "art", "performance" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 177, 178, 179, 180, 181 ], "text": "on two english benchmark datasets", "tokens": [ "on", "two", "english", "benchmark", "datasets" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 196 ], "text": "achieves", "tokens": [ "achieves" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "FEA", "offsets": [ 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132 ], "text": "different contextual words in the dependency tree obtained from an off - the - shelf dependency parser", "tokens": [ "different", "contextual", "words", "in", "the", "dependency", "tree", "obtained", "from", "an", "off", "-", "the", "-", "shelf", "dependency", "parser" ] }, { "argument_type": "TriedComponent", "nugget_type": "MOD", "offsets": [ 107, 108 ], "text": "attention mechanism", "tokens": [ "attention", "mechanism" ] }, { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 110, 111, 112 ], "text": "graph convolutional networks", "tokens": [ "graph", "convolutional", "networks" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 135 ], "text": "distinguish", "tokens": [ "distinguish" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 114 ], "text": "applied", "tokens": [ "applied" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 137, 138, 139, 140, 141 ], "text": "importance of different word dependencies", "tokens": [ "importance", "of", "different", "word", "dependencies" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 135 ], "text": "distinguish", "tokens": [ "distinguish" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 167, 168 ], "text": "type information", "tokens": [ "type", "information" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 169, 170, 171, 172, 173 ], "text": "in a - gcn modeling", "tokens": [ "in", "a", "-", "gcn", "modeling" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 165 ], "text": "include", "tokens": [ "include" ] } } ]
[ "syntactic", "information", ",", "especially", "dependency", "trees", ",", "has", "been", "widely", "used", "by", "existing", "studies", "to", "improve", "relation", "extraction", "with", "better", "semantic", "guidance", "for", "analyzing", "the", "context", "information", "associated", "with", "the", "given", "entities", ".", "however", ",", "most", "existing", "studies", "suffer", "from", "the", "noise", "in", "the", "dependency", "trees", ",", "especially", "when", "they", "are", "automatically", "generated", ",", "so", "that", "intensively", "leveraging", "dependency", "information", "may", "introduce", "confusions", "to", "relation", "classification", "and", "necessary", "pruning", "is", "of", "great", "importance", "in", "this", "task", ".", "in", "this", "paper", ",", "we", "propose", "a", "dependency", "-", "driven", "approach", "for", "relation", "extraction", "with", "attentive", "graph", "convolutional", "networks", "(", "a", "-", "gcn", ")", ".", "in", "this", "approach", ",", "an", "attention", "mechanism", "upon", "graph", "convolutional", "networks", "is", "applied", "to", "different", "contextual", "words", "in", "the", "dependency", "tree", "obtained", "from", "an", "off", "-", "the", "-", "shelf", "dependency", "parser", ",", "to", "distinguish", "the", "importance", "of", "different", "word", "dependencies", ".", "consider", "that", "dependency", "types", "among", "words", "also", "contain", "important", "contextual", "guidance", ",", "which", "is", "potentially", "helpful", "for", "relation", "extraction", ",", "we", "also", "include", "the", "type", "information", "in", "a", "-", "gcn", "modeling", ".", "experimental", "results", "on", "two", "english", "benchmark", "datasets", "demonstrate", "the", "effectiveness", "of", "our", "a", "-", "gcn", ",", "which", "outperforms", "previous", "studies", "and", "achieves", "state", "-", "of", "-", "the", "-", "art", "performance", "on", "both", "datasets", "." ]
ACL
Joint Chinese Word Segmentation and Part-of-speech Tagging via Two-way Attentions of Auto-analyzed Knowledge
Chinese word segmentation (CWS) and part-of-speech (POS) tagging are important fundamental tasks for Chinese language processing, where joint learning of them is an effective one-step solution for both tasks. Previous studies for joint CWS and POS tagging mainly follow the character-based tagging paradigm with introducing contextual information such as n-gram features or sentential representations from recurrent neural models. However, for many cases, the joint tagging needs not only modeling from context features but also knowledge attached to them (e.g., syntactic relations among words); limited efforts have been made by existing research to meet such needs. In this paper, we propose a neural model named TwASP for joint CWS and POS tagging following the character-based sequence labeling paradigm, where a two-way attention mechanism is used to incorporate both context feature and their corresponding syntactic knowledge for each input character. Particularly, we use existing language processing toolkits to obtain the auto-analyzed syntactic knowledge for the context, and the proposed attention module can learn and benefit from them although their quality may not be perfect. Our experiments illustrate the effectiveness of the two-way attentions for joint CWS and POS tagging, where state-of-the-art performance is achieved on five benchmark datasets.
9609778ad9e5f0ef4d2c7c494df6a6dc
2,020
[ "chinese word segmentation ( cws ) and part - of - speech ( pos ) tagging are important fundamental tasks for chinese language processing , where joint learning of them is an effective one - step solution for both tasks .", "previous studies for joint cws and pos tagging mainly follow the character - based tagging paradigm with introducing contextual information such as n - gram features or sentential representations from recurrent neural models .", "however , for many cases , the joint tagging needs not only modeling from context features but also knowledge attached to them ( e . g . , syntactic relations among words ) ; limited efforts have been made by existing research to meet such needs .", "in this paper , we propose a neural model named twasp for joint cws and pos tagging following the character - based sequence labeling paradigm , where a two - way attention mechanism is used to incorporate both context feature and their corresponding syntactic knowledge for each input character .", "particularly , we use existing language processing toolkits to obtain the auto - analyzed syntactic knowledge for the context , and the proposed attention module can learn and benefit from them although their quality may not be perfect .", "our experiments illustrate the effectiveness of the two - way attentions for joint cws and pos tagging , where state - of - the - art performance is achieved on five benchmark datasets ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 21, 22, 23 ], "text": "chinese language processing", "tokens": [ "chinese", "language", "processing" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 19 ], "text": "tasks", "tokens": [ "tasks" ] } }, { "arguments": [ { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 109, 110 ], "text": "limited efforts", "tokens": [ "limited", "efforts" ] }, { "argument_type": "Concern", "nugget_type": "APP", "offsets": [ 115, 116 ], "text": "existing research", "tokens": [ "existing", "research" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 118 ], "text": "meet", "tokens": [ "meet" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 113 ], "text": "made", "tokens": [ "made" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 126 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 132 ], "text": "twasp", "tokens": [ "twasp" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 134, 135, 136, 137, 138 ], "text": "joint cws and pos tagging", "tokens": [ "joint", "cws", "and", "pos", "tagging" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 127 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 174 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 181 ], "text": "obtain", "tokens": [ "obtain" ] }, { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 176, 177, 178, 179 ], "text": "existing language processing toolkits", "tokens": [ "existing", "language", "processing", "toolkits" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 175 ], "text": "use", "tokens": [ "use" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 183, 184, 185, 186, 187, 188, 189, 190 ], "text": "auto - analyzed syntactic knowledge for the context", "tokens": [ "auto", "-", "analyzed", "syntactic", "knowledge", "for", "the", "context" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 181 ], "text": "obtain", "tokens": [ "obtain" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "STR", "offsets": [ 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227 ], "text": "effectiveness of the two - way attentions for joint cws and pos tagging", "tokens": [ "effectiveness", "of", "the", "two", "-", "way", "attentions", "for", "joint", "cws", "and", "pos", "tagging" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 213 ], "text": "illustrate", "tokens": [ "illustrate" ] } }, { "arguments": [ { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 241, 242, 243 ], "text": "five benchmark datasets", "tokens": [ "five", "benchmark", "datasets" ] }, { "argument_type": "Subject", "nugget_type": "STR", "offsets": [ 230, 231, 232, 233, 234, 235, 236, 237 ], "text": "state - of - the - art performance", "tokens": [ "state", "-", "of", "-", "the", "-", "art", "performance" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 239 ], "text": "achieved", "tokens": [ "achieved" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 59, 60 ], "text": "contextual information", "tokens": [ "contextual", "information" ] }, { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 71, 72, 73 ], "text": "recurrent neural models", "tokens": [ "recurrent", "neural", "models" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 41, 42 ], "text": "previous studies", "tokens": [ "previous", "studies" ] } ], "event_type": "RWS", "trigger": { "offsets": [ 58 ], "text": "introducing", "tokens": [ "introducing" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96 ], "text": "such needs", "tokens": [ "needs", "not", "only", "modeling", "from", "context", "features", "but", "also", "knowledge", "attached", "to", "them" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 118 ], "text": "meet", "tokens": [ "meet" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "APP", "offsets": [ 150, 151, 152, 153, 154 ], "text": "two - way attention mechanism", "tokens": [ "two", "-", "way", "attention", "mechanism" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 160, 161, 167, 168, 169, 170 ], "text": "context feature for each input character", "tokens": [ "context", "feature", "for", "each", "input", "character" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 160, 161, 164, 165, 166, 167, 168, 169, 170 ], "text": "their corresponding syntactic knowledge for each input character", "tokens": [ "context", "feature", "corresponding", "syntactic", "knowledge", "for", "each", "input", "character" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 158 ], "text": "incorporate", "tokens": [ "incorporate" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "MOD", "offsets": [ 194, 195, 196 ], "text": "proposed attention module", "tokens": [ "proposed", "attention", "module" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 198, 199, 200 ], "text": "learn and benefit", "tokens": [ "learn", "and", "benefit" ] } } ]
[ "chinese", "word", "segmentation", "(", "cws", ")", "and", "part", "-", "of", "-", "speech", "(", "pos", ")", "tagging", "are", "important", "fundamental", "tasks", "for", "chinese", "language", "processing", ",", "where", "joint", "learning", "of", "them", "is", "an", "effective", "one", "-", "step", "solution", "for", "both", "tasks", ".", "previous", "studies", "for", "joint", "cws", "and", "pos", "tagging", "mainly", "follow", "the", "character", "-", "based", "tagging", "paradigm", "with", "introducing", "contextual", "information", "such", "as", "n", "-", "gram", "features", "or", "sentential", "representations", "from", "recurrent", "neural", "models", ".", "however", ",", "for", "many", "cases", ",", "the", "joint", "tagging", "needs", "not", "only", "modeling", "from", "context", "features", "but", "also", "knowledge", "attached", "to", "them", "(", "e", ".", "g", ".", ",", "syntactic", "relations", "among", "words", ")", ";", "limited", "efforts", "have", "been", "made", "by", "existing", "research", "to", "meet", "such", "needs", ".", "in", "this", "paper", ",", "we", "propose", "a", "neural", "model", "named", "twasp", "for", "joint", "cws", "and", "pos", "tagging", "following", "the", "character", "-", "based", "sequence", "labeling", "paradigm", ",", "where", "a", "two", "-", "way", "attention", "mechanism", "is", "used", "to", "incorporate", "both", "context", "feature", "and", "their", "corresponding", "syntactic", "knowledge", "for", "each", "input", "character", ".", "particularly", ",", "we", "use", "existing", "language", "processing", "toolkits", "to", "obtain", "the", "auto", "-", "analyzed", "syntactic", "knowledge", "for", "the", "context", ",", "and", "the", "proposed", "attention", "module", "can", "learn", "and", "benefit", "from", "them", "although", "their", "quality", "may", "not", "be", "perfect", ".", "our", "experiments", "illustrate", "the", "effectiveness", "of", "the", "two", "-", "way", "attentions", "for", "joint", "cws", "and", "pos", "tagging", ",", "where", "state", "-", "of", "-", "the", "-", "art", "performance", "is", "achieved", "on", "five", "benchmark", "datasets", "." ]
ACL
RankQA: Neural Question Answering with Answer Re-Ranking
The conventional paradigm in neural question answering (QA) for narrative content is limited to a two-stage process: first, relevant text passages are retrieved and, subsequently, a neural network for machine comprehension extracts the likeliest answer. However, both stages are largely isolated in the status quo and, hence, information from the two phases is never properly fused. In contrast, this work proposes RankQA: RankQA extends the conventional two-stage process in neural QA with a third stage that performs an additional answer re-ranking. The re-ranking leverages different features that are directly extracted from the QA pipeline, i.e., a combination of retrieval and comprehension features. While our intentionally simple design allows for an efficient, data-sparse estimation, it nevertheless outperforms more complex QA systems by a significant margin: in fact, RankQA achieves state-of-the-art performance on 3 out of 4 benchmark datasets. Furthermore, its performance is especially superior in settings where the size of the corpus is dynamic. Here the answer re-ranking provides an effective remedy against the underlying noise-information trade-off due to a variable corpus size. As a consequence, RankQA represents a novel, powerful, and thus challenging baseline for future research in content-based QA.
864e901c9c8268c1e32b4e85b4cdda05
2,019
[ "the conventional paradigm in neural question answering ( qa ) for narrative content is limited to a two - stage process : first , relevant text passages are retrieved and , subsequently , a neural network for machine comprehension extracts the likeliest answer .", "however , both stages are largely isolated in the status quo and , hence , information from the two phases is never properly fused .", "in contrast , this work proposes rankqa : rankqa extends the conventional two - stage process in neural qa with a third stage that performs an additional answer re - ranking .", "the re - ranking leverages different features that are directly extracted from the qa pipeline , i . e . , a combination of retrieval and comprehension features .", "while our intentionally simple design allows for an efficient , data - sparse estimation , it nevertheless outperforms more complex qa systems by a significant margin : in fact , rankqa achieves state - of - the - art performance on 3 out of 4 benchmark datasets .", "furthermore , its performance is especially superior in settings where the size of the corpus is dynamic .", "here the answer re - ranking provides an effective remedy against the underlying noise - information trade - off due to a variable corpus size .", "as a consequence , rankqa represents a novel , powerful , and thus challenging baseline for future research in content - based qa ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 4, 5, 6 ], "text": "neural question answering", "tokens": [ "neural", "question", "answering" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 14 ], "text": "limited", "tokens": [ "limited" ] } }, { "arguments": [], "event_type": "RWF", "trigger": { "offsets": [ 50 ], "text": "isolated", "tokens": [ "isolated" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 75 ], "text": "rankqa", "tokens": [ "rankqa" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 74 ], "text": "proposes", "tokens": [ "proposes" ] } }, { "arguments": [ { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 138, 139, 140, 141, 142, 143 ], "text": "efficient , data - sparse estimation", "tokens": [ "efficient", ",", "data", "-", "sparse", "estimation" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 135 ], "text": "allows", "tokens": [ "allows" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 75 ], "text": "rankqa", "tokens": [ "rankqa" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 147 ], "text": "outperforms", "tokens": [ "outperforms" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 154, 155 ], "text": "significant margin", "tokens": [ "significant", "margin" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 147 ], "text": "outperforms", "tokens": [ "outperforms" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 160 ], "text": "rankqa", "tokens": [ "rankqa" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 162, 163, 164, 165, 166, 167, 168 ], "text": "state - of - the - art", "tokens": [ "state", "-", "of", "-", "the", "-", "art" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 169 ], "text": "performance", "tokens": [ "performance" ] }, { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 170, 171, 172, 173, 174, 175, 176 ], "text": "on 3 out of 4 benchmark datasets", "tokens": [ "on", "3", "out", "of", "4", "benchmark", "datasets" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 161 ], "text": "achieves", "tokens": [ "achieves" ] } }, { "arguments": [ { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 183 ], "text": "especially", "tokens": [ "especially" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 184 ], "text": "superior", "tokens": [ "superior" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 181 ], "text": "performance", "tokens": [ "performance" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 187, 188, 189, 190, 191, 192, 193, 194 ], "text": "where the size of the corpus is dynamic", "tokens": [ "where", "the", "size", "of", "the", "corpus", "is", "dynamic" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 184 ], "text": "superior", "tokens": [ "superior" ] } } ]
[ "the", "conventional", "paradigm", "in", "neural", "question", "answering", "(", "qa", ")", "for", "narrative", "content", "is", "limited", "to", "a", "two", "-", "stage", "process", ":", "first", ",", "relevant", "text", "passages", "are", "retrieved", "and", ",", "subsequently", ",", "a", "neural", "network", "for", "machine", "comprehension", "extracts", "the", "likeliest", "answer", ".", "however", ",", "both", "stages", "are", "largely", "isolated", "in", "the", "status", "quo", "and", ",", "hence", ",", "information", "from", "the", "two", "phases", "is", "never", "properly", "fused", ".", "in", "contrast", ",", "this", "work", "proposes", "rankqa", ":", "rankqa", "extends", "the", "conventional", "two", "-", "stage", "process", "in", "neural", "qa", "with", "a", "third", "stage", "that", "performs", "an", "additional", "answer", "re", "-", "ranking", ".", "the", "re", "-", "ranking", "leverages", "different", "features", "that", "are", "directly", "extracted", "from", "the", "qa", "pipeline", ",", "i", ".", "e", ".", ",", "a", "combination", "of", "retrieval", "and", "comprehension", "features", ".", "while", "our", "intentionally", "simple", "design", "allows", "for", "an", "efficient", ",", "data", "-", "sparse", "estimation", ",", "it", "nevertheless", "outperforms", "more", "complex", "qa", "systems", "by", "a", "significant", "margin", ":", "in", "fact", ",", "rankqa", "achieves", "state", "-", "of", "-", "the", "-", "art", "performance", "on", "3", "out", "of", "4", "benchmark", "datasets", ".", "furthermore", ",", "its", "performance", "is", "especially", "superior", "in", "settings", "where", "the", "size", "of", "the", "corpus", "is", "dynamic", ".", "here", "the", "answer", "re", "-", "ranking", "provides", "an", "effective", "remedy", "against", "the", "underlying", "noise", "-", "information", "trade", "-", "off", "due", "to", "a", "variable", "corpus", "size", ".", "as", "a", "consequence", ",", "rankqa", "represents", "a", "novel", ",", "powerful", ",", "and", "thus", "challenging", "baseline", "for", "future", "research", "in", "content", "-", "based", "qa", "." ]
ACL
Domain Adaptation in Multilingual and Multi-Domain Monolingual Settings for Complex Word Identification
Complex word identification (CWI) is a cornerstone process towards proper text simplification. CWI is highly dependent on context, whereas its difficulty is augmented by the scarcity of available datasets which vary greatly in terms of domains and languages. As such, it becomes increasingly more difficult to develop a robust model that generalizes across a wide array of input examples. In this paper, we propose a novel training technique for the CWI task based on domain adaptation to improve the target character and context representations. This technique addresses the problem of working with multiple domains, inasmuch as it creates a way of smoothing the differences between the explored datasets. Moreover, we also propose a similar auxiliary task, namely text simplification, that can be used to complement lexical complexity prediction. Our model obtains a boost of up to 2.42% in terms of Pearson Correlation Coefficients in contrast to vanilla training techniques, when considering the CompLex from the Lexical Complexity Prediction 2021 dataset. At the same time, we obtain an increase of 3% in Pearson scores, while considering a cross-lingual setup relying on the Complex Word Identification 2018 dataset. In addition, our model yields state-of-the-art results in terms of Mean Absolute Error.
0e0d6cc75f98e0e32960341f2f384171
2,022
[ "complex word identification ( cwi ) is a cornerstone process towards proper text simplification .", "cwi is highly dependent on context , whereas its difficulty is augmented by the scarcity of available datasets which vary greatly in terms of domains and languages .", "as such , it becomes increasingly more difficult to develop a robust model that generalizes across a wide array of input examples .", "in this paper , we propose a novel training technique for the cwi task based on domain adaptation to improve the target character and context representations .", "this technique addresses the problem of working with multiple domains , inasmuch as it creates a way of smoothing the differences between the explored datasets .", "moreover , we also propose a similar auxiliary task , namely text simplification , that can be used to complement lexical complexity prediction .", "our model obtains a boost of up to 2 . 42 % in terms of pearson correlation coefficients in contrast to vanilla training techniques , when considering the complex from the lexical complexity prediction 2021 dataset .", "at the same time , we obtain an increase of 3 % in pearson scores , while considering a cross - lingual setup relying on the complex word identification 2018 dataset .", "in addition , our model yields state - of - the - art results in terms of mean absolute error ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1, 2 ], "text": "complex word identification", "tokens": [ "complex", "word", "identification" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 8, 9 ], "text": "cornerstone process", "tokens": [ "cornerstone", "process" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1, 2 ], "text": "complex word identification", "tokens": [ "complex", "word", "identification" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 17, 18 ], "text": "highly dependent", "tokens": [ "highly", "dependent" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1, 2, 24 ], "text": "its difficulty", "tokens": [ "complex", "word", "identification", "difficulty" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 29, 30, 31, 32 ], "text": "scarcity of available datasets", "tokens": [ "scarcity", "of", "available", "datasets" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 26 ], "text": "augmented", "tokens": [ "augmented" ] } }, { "arguments": [ { "argument_type": "Concern", "nugget_type": "DST", "offsets": [ 31, 32 ], "text": "available datasets", "tokens": [ "available", "datasets" ] }, { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 34, 35 ], "text": "vary greatly", "tokens": [ "vary", "greatly" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 34, 35 ], "text": "vary greatly", "tokens": [ "vary", "greatly" ] } }, { "arguments": [ { "argument_type": "Fault", "nugget_type": "WEA", "offsets": [ 49, 50 ], "text": "more difficult", "tokens": [ "more", "difficult" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 52 ], "text": "develop", "tokens": [ "develop" ] } ], "event_type": "RWF", "trigger": { "offsets": [ 47, 48 ], "text": "becomes increasingly", "tokens": [ "becomes", "increasingly" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "APP", "offsets": [ 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64 ], "text": "robust model that generalizes across a wide array of input examples", "tokens": [ "robust", "model", "that", "generalizes", "across", "a", "wide", "array", "of", "input", "examples" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 52 ], "text": "develop", "tokens": [ "develop" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 70 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 74, 75 ], "text": "training technique", "tokens": [ "training", "technique" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 0, 1, 2, 79, 80, 81, 82, 83 ], "text": "cwi task based on domain adaptation", "tokens": [ "complex", "word", "identification", "task", "based", "on", "domain", "adaptation" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 85 ], "text": "improve", "tokens": [ "improve" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 71 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 87, 88 ], "text": "target character", "tokens": [ "target", "character" ] }, { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 90, 91 ], "text": "context representations", "tokens": [ "context", "representations" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 85 ], "text": "improve", "tokens": [ "improve" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 121 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 125, 126, 127 ], "text": "similar auxiliary task", "tokens": [ "similar", "auxiliary", "task" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 138 ], "text": "complement", "tokens": [ "complement" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 123 ], "text": "propose", "tokens": [ "propose" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 139, 140, 141 ], "text": "lexical complexity prediction", "tokens": [ "lexical", "complexity", "prediction" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 138 ], "text": "complement", "tokens": [ "complement" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 143, 144 ], "text": "our model", "tokens": [ "our", "model" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 147 ], "text": "boost", "tokens": [ "boost" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 149, 150, 151, 152, 153, 154 ], "text": "up to 2 . 42 %", "tokens": [ "up", "to", "2", ".", "42", "%" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 155, 156, 157, 158, 159, 160 ], "text": "in terms of pearson correlation coefficients", "tokens": [ "in", "terms", "of", "pearson", "correlation", "coefficients" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 164, 165, 166 ], "text": "vanilla training techniques", "tokens": [ "vanilla", "training", "techniques" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 169, 170, 171, 172, 173, 174, 175, 176, 177, 178 ], "text": "considering the complex from the lexical complexity prediction 2021 dataset", "tokens": [ "considering", "the", "complex", "from", "the", "lexical", "complexity", "prediction", "2021", "dataset" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 145 ], "text": "obtains", "tokens": [ "obtains" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 143, 144 ], "text": "our model", "tokens": [ "our", "model" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 193, 194 ], "text": "pearson scores", "tokens": [ "pearson", "scores" ] }, { "argument_type": "Extent", "nugget_type": "DEG", "offsets": [ 190, 191 ], "text": "3 %", "tokens": [ "3", "%" ] }, { "argument_type": "Object", "nugget_type": "TAK", "offsets": [ 188 ], "text": "increase", "tokens": [ "increase" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210 ], "text": "considering a cross - lingual setup relying on the complex word identification 2018 dataset", "tokens": [ "considering", "a", "cross", "-", "lingual", "setup", "relying", "on", "the", "complex", "word", "identification", "2018", "dataset" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 186 ], "text": "obtain", "tokens": [ "obtain" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 215, 216 ], "text": "our model", "tokens": [ "our", "model" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 218, 219, 220, 221, 222, 223, 224 ], "text": "state - of - the - art", "tokens": [ "state", "-", "of", "-", "the", "-", "art" ] }, { "argument_type": "Metrics", "nugget_type": "TAK", "offsets": [ 225 ], "text": "results", "tokens": [ "results" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 226, 227, 228, 229, 230, 231 ], "text": "in terms of mean absolute error", "tokens": [ "in", "terms", "of", "mean", "absolute", "error" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 217 ], "text": "yields", "tokens": [ "yields" ] } } ]
[ "complex", "word", "identification", "(", "cwi", ")", "is", "a", "cornerstone", "process", "towards", "proper", "text", "simplification", ".", "cwi", "is", "highly", "dependent", "on", "context", ",", "whereas", "its", "difficulty", "is", "augmented", "by", "the", "scarcity", "of", "available", "datasets", "which", "vary", "greatly", "in", "terms", "of", "domains", "and", "languages", ".", "as", "such", ",", "it", "becomes", "increasingly", "more", "difficult", "to", "develop", "a", "robust", "model", "that", "generalizes", "across", "a", "wide", "array", "of", "input", "examples", ".", "in", "this", "paper", ",", "we", "propose", "a", "novel", "training", "technique", "for", "the", "cwi", "task", "based", "on", "domain", "adaptation", "to", "improve", "the", "target", "character", "and", "context", "representations", ".", "this", "technique", "addresses", "the", "problem", "of", "working", "with", "multiple", "domains", ",", "inasmuch", "as", "it", "creates", "a", "way", "of", "smoothing", "the", "differences", "between", "the", "explored", "datasets", ".", "moreover", ",", "we", "also", "propose", "a", "similar", "auxiliary", "task", ",", "namely", "text", "simplification", ",", "that", "can", "be", "used", "to", "complement", "lexical", "complexity", "prediction", ".", "our", "model", "obtains", "a", "boost", "of", "up", "to", "2", ".", "42", "%", "in", "terms", "of", "pearson", "correlation", "coefficients", "in", "contrast", "to", "vanilla", "training", "techniques", ",", "when", "considering", "the", "complex", "from", "the", "lexical", "complexity", "prediction", "2021", "dataset", ".", "at", "the", "same", "time", ",", "we", "obtain", "an", "increase", "of", "3", "%", "in", "pearson", "scores", ",", "while", "considering", "a", "cross", "-", "lingual", "setup", "relying", "on", "the", "complex", "word", "identification", "2018", "dataset", ".", "in", "addition", ",", "our", "model", "yields", "state", "-", "of", "-", "the", "-", "art", "results", "in", "terms", "of", "mean", "absolute", "error", "." ]
ACL
Substructure Distribution Projection for Zero-Shot Cross-Lingual Dependency Parsing
We present substructure distribution projection (SubDP), a technique that projects a distribution over structures in one domain to another, by projecting substructure distributions separately. Models for the target domain can then be trained, using the projected distributions as soft silver labels. We evaluate SubDP on zero shot cross-lingual dependency parsing, taking dependency arcs as substructures: we project the predicted dependency arc distributions in the source language(s) to target language(s), and train a target language parser on the resulting distributions. Given an English tree bank as the only source of human supervision, SubDP achieves better unlabeled attachment score than all prior work on the Universal Dependencies v2.2 (Nivre et al., 2020) test set across eight diverse target languages, as well as the best labeled attachment score on six languages. In addition, SubDP improves zero shot cross-lingual dependency parsing with very few (e.g., 50) supervised bitext pairs, across a broader range of target languages.
a11d23df083ec881504fcaf35594405c
2,022
[ "we present substructure distribution projection ( subdp ) , a technique that projects a distribution over structures in one domain to another , by projecting substructure distributions separately .", "models for the target domain can then be trained , using the projected distributions as soft silver labels .", "we evaluate subdp on zero shot cross - lingual dependency parsing , taking dependency arcs as substructures : we project the predicted dependency arc distributions in the source language ( s ) to target language ( s ) , and train a target language parser on the resulting distributions .", "given an english tree bank as the only source of human supervision , subdp achieves better unlabeled attachment score than all prior work on the universal dependencies v2 . 2 ( nivre et al . , 2020 ) test set across eight diverse target languages , as well as the best labeled attachment score on six languages .", "in addition , subdp improves zero shot cross - lingual dependency parsing with very few ( e . g . , 50 ) supervised bitext pairs , across a broader range of target languages ." ]
[ { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 0 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 2, 3, 4 ], "text": "substructure distribution projection", "tokens": [ "substructure", "distribution", "projection" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 1 ], "text": "present", "tokens": [ "present" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 25, 26 ], "text": "substructure distributions", "tokens": [ "substructure", "distributions" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 12 ], "text": "projects", "tokens": [ "projects" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 27 ], "text": "separately", "tokens": [ "separately" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 24 ], "text": "projecting", "tokens": [ "projecting" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "FEA", "offsets": [ 14, 15, 16 ], "text": "distribution over structures", "tokens": [ "distribution", "over", "structures" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 17, 18, 19, 20, 21 ], "text": "in one domain to another", "tokens": [ "in", "one", "domain", "to", "another" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 12 ], "text": "projects", "tokens": [ "projects" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 69, 70, 71, 72 ], "text": "predicted dependency arc distributions", "tokens": [ "predicted", "dependency", "arc", "distributions" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 81, 82 ], "text": "target language", "tokens": [ "target", "language" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 67 ], "text": "project", "tokens": [ "project" ] } }, { "arguments": [ { "argument_type": "Arg1", "nugget_type": "APP", "offsets": [ 2, 3, 4 ], "text": "substructure distribution projection", "tokens": [ "substructure", "distribution", "projection" ] }, { "argument_type": "Result", "nugget_type": "STR", "offsets": [ 113, 114, 115, 116 ], "text": "better unlabeled attachment score", "tokens": [ "better", "unlabeled", "attachment", "score" ] }, { "argument_type": "Arg2", "nugget_type": "APP", "offsets": [ 118, 119, 120 ], "text": "all prior work", "tokens": [ "all", "prior", "work" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 138, 139, 140, 141, 142 ], "text": "across eight diverse target languages", "tokens": [ "across", "eight", "diverse", "target", "languages" ] }, { "argument_type": "Dataset", "nugget_type": "DST", "offsets": [ 123, 124, 125, 126, 127, 136, 137 ], "text": "universal dependencies v2 . 2 ( nivre et al . , 2020 ) test set", "tokens": [ "universal", "dependencies", "v2", ".", "2", "test", "set" ] } ], "event_type": "CMP", "trigger": { "offsets": [ 112 ], "text": "achieves", "tokens": [ "achieves" ] } }, { "arguments": [ { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 2, 3, 4 ], "text": "substructure distribution projection", "tokens": [ "substructure", "distribution", "projection" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 168, 169, 170, 179, 180, 181 ], "text": "with very few ( e . g . , 50 ) supervised bitext pairs", "tokens": [ "with", "very", "few", "supervised", "bitext", "pairs" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 183, 184, 185, 186, 187, 188, 189 ], "text": "across a broader range of target languages", "tokens": [ "across", "a", "broader", "range", "of", "target", "languages" ] }, { "argument_type": "Object", "nugget_type": "APP", "offsets": [ 161, 162, 163, 164, 165, 166, 167 ], "text": "zero shot cross - lingual dependency parsing", "tokens": [ "zero", "shot", "cross", "-", "lingual", "dependency", "parsing" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 160 ], "text": "improves", "tokens": [ "improves" ] } }, { "arguments": [ { "argument_type": "BaseComponent", "nugget_type": "APP", "offsets": [ 29, 30, 31, 32, 33 ], "text": "models for the target domain", "tokens": [ "models", "for", "the", "target", "domain" ] }, { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 41, 42 ], "text": "projected distributions", "tokens": [ "projected", "distributions" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 43, 44, 45, 46 ], "text": "as soft silver labels", "tokens": [ "as", "soft", "silver", "labels" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 37 ], "text": "trained", "tokens": [ "trained" ] } }, { "arguments": [ { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 49 ], "text": "evaluate", "tokens": [ "evaluate" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 51, 52, 53, 54, 55, 56, 57, 58 ], "text": "on zero shot cross - lingual dependency parsing", "tokens": [ "on", "zero", "shot", "cross", "-", "lingual", "dependency", "parsing" ] }, { "argument_type": "BaseComponent", "nugget_type": "FEA", "offsets": [ 61, 62 ], "text": "dependency arcs", "tokens": [ "dependency", "arcs" ] }, { "argument_type": "TriedComponent", "nugget_type": "MOD", "offsets": [ 64 ], "text": "substructures", "tokens": [ "substructures" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 60 ], "text": "taking", "tokens": [ "taking" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "APP", "offsets": [ 2, 3, 4 ], "text": "subdp", "tokens": [ "substructure", "distribution", "projection" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 49 ], "text": "evaluate", "tokens": [ "evaluate" ] } }, { "arguments": [ { "argument_type": "TriedComponent", "nugget_type": "FEA", "offsets": [ 95, 96 ], "text": "resulting distributions", "tokens": [ "resulting", "distributions" ] }, { "argument_type": "BaseComponent", "nugget_type": "MOD", "offsets": [ 90, 91, 92 ], "text": "target language parser", "tokens": [ "target", "language", "parser" ] } ], "event_type": "MDS", "trigger": { "offsets": [ 88 ], "text": "train", "tokens": [ "train" ] } }, { "arguments": [ { "argument_type": "Object", "nugget_type": "STR", "offsets": [ 148, 149, 150, 151 ], "text": "best labeled attachment score", "tokens": [ "best", "labeled", "attachment", "score" ] }, { "argument_type": "Subject", "nugget_type": "APP", "offsets": [ 2, 3, 4 ], "text": "substructure distribution projection", "tokens": [ "substructure", "distribution", "projection" ] } ], "event_type": "FAC", "trigger": { "offsets": [ 112 ], "text": "achieves", "tokens": [ "achieves" ] } } ]
[ "we", "present", "substructure", "distribution", "projection", "(", "subdp", ")", ",", "a", "technique", "that", "projects", "a", "distribution", "over", "structures", "in", "one", "domain", "to", "another", ",", "by", "projecting", "substructure", "distributions", "separately", ".", "models", "for", "the", "target", "domain", "can", "then", "be", "trained", ",", "using", "the", "projected", "distributions", "as", "soft", "silver", "labels", ".", "we", "evaluate", "subdp", "on", "zero", "shot", "cross", "-", "lingual", "dependency", "parsing", ",", "taking", "dependency", "arcs", "as", "substructures", ":", "we", "project", "the", "predicted", "dependency", "arc", "distributions", "in", "the", "source", "language", "(", "s", ")", "to", "target", "language", "(", "s", ")", ",", "and", "train", "a", "target", "language", "parser", "on", "the", "resulting", "distributions", ".", "given", "an", "english", "tree", "bank", "as", "the", "only", "source", "of", "human", "supervision", ",", "subdp", "achieves", "better", "unlabeled", "attachment", "score", "than", "all", "prior", "work", "on", "the", "universal", "dependencies", "v2", ".", "2", "(", "nivre", "et", "al", ".", ",", "2020", ")", "test", "set", "across", "eight", "diverse", "target", "languages", ",", "as", "well", "as", "the", "best", "labeled", "attachment", "score", "on", "six", "languages", ".", "in", "addition", ",", "subdp", "improves", "zero", "shot", "cross", "-", "lingual", "dependency", "parsing", "with", "very", "few", "(", "e", ".", "g", ".", ",", "50", ")", "supervised", "bitext", "pairs", ",", "across", "a", "broader", "range", "of", "target", "languages", "." ]
ACL
How can NLP Help Revitalize Endangered Languages? A Case Study and Roadmap for the Cherokee Language
More than 43% of the languages spoken in the world are endangered, and language loss currently occurs at an accelerated rate because of globalization and neocolonialism. Saving and revitalizing endangered languages has become very important for maintaining the cultural diversity on our planet. In this work, we focus on discussing how NLP can help revitalize endangered languages. We first suggest three principles that may help NLP practitioners to foster mutual understanding and collaboration with language communities, and we discuss three ways in which NLP can potentially assist in language education. We then take Cherokee, a severely-endangered Native American language, as a case study. After reviewing the language’s history, linguistic features, and existing resources, we (in collaboration with Cherokee community members) arrive at a few meaningful ways NLP practitioners can collaborate with community partners. We suggest two approaches to enrich the Cherokee language’s resources with machine-in-the-loop processing, and discuss several NLP tools that people from the Cherokee community have shown interest in. We hope that our work serves not only to inform the NLP community about Cherokee, but also to provide inspiration for future work on endangered languages in general.
6043d6347900bef1481b47f957a5bdf4
2,022
[ "more than 43 % of the languages spoken in the world are endangered , and language loss currently occurs at an accelerated rate because of globalization and neocolonialism .", "saving and revitalizing endangered languages has become very important for maintaining the cultural diversity on our planet .", "in this work , we focus on discussing how nlp can help revitalize endangered languages .", "we first suggest three principles that may help nlp practitioners to foster mutual understanding and collaboration with language communities , and we discuss three ways in which nlp can potentially assist in language education .", "we then take cherokee , a severely - endangered native american language , as a case study .", "after reviewing the language ’ s history , linguistic features , and existing resources , we ( in collaboration with cherokee community members ) arrive at a few meaningful ways nlp practitioners can collaborate with community partners .", "we suggest two approaches to enrich the cherokee language ’ s resources with machine - in - the - loop processing , and discuss several nlp tools that people from the cherokee community have shown interest in .", "we hope that our work serves not only to inform the nlp community about cherokee , but also to provide inspiration for future work on endangered languages in general ." ]
[ { "arguments": [ { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 41, 42 ], "text": "cultural diversity", "tokens": [ "cultural", "diversity" ] } ], "event_type": "ITT", "trigger": { "offsets": [ 39 ], "text": "maintaining", "tokens": [ "maintaining" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 51 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 56 ], "text": "nlp", "tokens": [ "nlp" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 59 ], "text": "revitalize", "tokens": [ "revitalize" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 54 ], "text": "discussing", "tokens": [ "discussing" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 63 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 66, 67 ], "text": "three principles", "tokens": [ "three", "principles" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 74 ], "text": "foster", "tokens": [ "foster" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 65 ], "text": "suggest", "tokens": [ "suggest" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 84 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 86, 87 ], "text": "three ways", "tokens": [ "three", "ways" ] }, { "argument_type": "Target", "nugget_type": "TAK", "offsets": [ 95, 96 ], "text": "language education", "tokens": [ "language", "education" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 85 ], "text": "discuss", "tokens": [ "discuss" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 98 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 101 ], "text": "cherokee", "tokens": [ "cherokee" ] }, { "argument_type": "Condition", "nugget_type": "LIM", "offsets": [ 111, 112, 113, 114 ], "text": "as a case study", "tokens": [ "as", "a", "case", "study" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 100 ], "text": "take", "tokens": [ "take" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 131 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "TAK", "offsets": [ 119, 120, 121, 122 ], "text": "language ’ s history", "tokens": [ "language", "’", "s", "history" ] }, { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 124, 125 ], "text": "linguistic features", "tokens": [ "linguistic", "features" ] }, { "argument_type": "Content", "nugget_type": "FEA", "offsets": [ 128, 129 ], "text": "existing resources", "tokens": [ "existing", "resources" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 140 ], "text": "arrive", "tokens": [ "arrive" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 117 ], "text": "reviewing", "tokens": [ "reviewing" ] } }, { "arguments": [ { "argument_type": "Proposer", "nugget_type": "OG", "offsets": [ 154 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Target", "nugget_type": "E-PUR", "offsets": [ 159 ], "text": "enrich", "tokens": [ "enrich" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 167, 168, 169, 170, 171, 172, 173, 174 ], "text": "machine - in - the - loop processing", "tokens": [ "machine", "-", "in", "-", "the", "-", "loop", "processing" ] } ], "event_type": "PRP", "trigger": { "offsets": [ 155 ], "text": "suggest", "tokens": [ "suggest" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 161, 162, 163, 164, 165 ], "text": "cherokee language ’ s resources", "tokens": [ "cherokee", "language", "’", "s", "resources" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 159 ], "text": "enrich", "tokens": [ "enrich" ] } }, { "arguments": [ { "argument_type": "Researcher", "nugget_type": "OG", "offsets": [ 154 ], "text": "we", "tokens": [ "we" ] }, { "argument_type": "Content", "nugget_type": "APP", "offsets": [ 178, 179, 180 ], "text": "several nlp tools", "tokens": [ "several", "nlp", "tools" ] } ], "event_type": "WKS", "trigger": { "offsets": [ 177 ], "text": "discuss", "tokens": [ "discuss" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 60, 61 ], "text": "endangered languages", "tokens": [ "endangered", "languages" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 59 ], "text": "revitalize", "tokens": [ "revitalize" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 75, 76 ], "text": "mutual understanding", "tokens": [ "mutual", "understanding" ] }, { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 75, 78 ], "text": "mutual collaboration", "tokens": [ "mutual", "collaboration" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 74 ], "text": "foster", "tokens": [ "foster" ] } }, { "arguments": [ { "argument_type": "Aim", "nugget_type": "TAK", "offsets": [ 143, 144, 145 ], "text": "few meaningful ways", "tokens": [ "few", "meaningful", "ways" ] } ], "event_type": "PUR", "trigger": { "offsets": [ 140 ], "text": "arrive", "tokens": [ "arrive" ] } } ]
[ "more", "than", "43", "%", "of", "the", "languages", "spoken", "in", "the", "world", "are", "endangered", ",", "and", "language", "loss", "currently", "occurs", "at", "an", "accelerated", "rate", "because", "of", "globalization", "and", "neocolonialism", ".", "saving", "and", "revitalizing", "endangered", "languages", "has", "become", "very", "important", "for", "maintaining", "the", "cultural", "diversity", "on", "our", "planet", ".", "in", "this", "work", ",", "we", "focus", "on", "discussing", "how", "nlp", "can", "help", "revitalize", "endangered", "languages", ".", "we", "first", "suggest", "three", "principles", "that", "may", "help", "nlp", "practitioners", "to", "foster", "mutual", "understanding", "and", "collaboration", "with", "language", "communities", ",", "and", "we", "discuss", "three", "ways", "in", "which", "nlp", "can", "potentially", "assist", "in", "language", "education", ".", "we", "then", "take", "cherokee", ",", "a", "severely", "-", "endangered", "native", "american", "language", ",", "as", "a", "case", "study", ".", "after", "reviewing", "the", "language", "’", "s", "history", ",", "linguistic", "features", ",", "and", "existing", "resources", ",", "we", "(", "in", "collaboration", "with", "cherokee", "community", "members", ")", "arrive", "at", "a", "few", "meaningful", "ways", "nlp", "practitioners", "can", "collaborate", "with", "community", "partners", ".", "we", "suggest", "two", "approaches", "to", "enrich", "the", "cherokee", "language", "’", "s", "resources", "with", "machine", "-", "in", "-", "the", "-", "loop", "processing", ",", "and", "discuss", "several", "nlp", "tools", "that", "people", "from", "the", "cherokee", "community", "have", "shown", "interest", "in", ".", "we", "hope", "that", "our", "work", "serves", "not", "only", "to", "inform", "the", "nlp", "community", "about", "cherokee", ",", "but", "also", "to", "provide", "inspiration", "for", "future", "work", "on", "endangered", "languages", "in", "general", "." ]
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
1