id stringlengths 40 40 | pid stringlengths 42 42 | input stringlengths 8.37k 169k | output stringlengths 1 1.63k |
|---|---|---|---|
9a4aa0e4096c73cd2c3b1eab437c1bf24ae7bf03 | 9a4aa0e4096c73cd2c3b1eab437c1bf24ae7bf03_0 | Q: What text sequences are associated with each vertex?
Text: Introduction
Networks are ubiquitous, with prominent examples including social networks (e.g., Facebook, Twitter) or citation networks of research papers (e.g., arXiv). When analyzing data from these real-world networks, traditional methods often represent v... | abstracts, sentences |
1d1ab5d8a24dfd15d95a5a7506ac0456d1192209 | 1d1ab5d8a24dfd15d95a5a7506ac0456d1192209_0 | Q: How long does it take for the model to run?
Text: Introduction
Networks are ubiquitous, with prominent examples including social networks (e.g., Facebook, Twitter) or citation networks of research papers (e.g., arXiv). When analyzing data from these real-world networks, traditional methods often represent vertices (... | Unanswerable |
09a993756d2781a89f7ec5d7992f812d60e24232 | 09a993756d2781a89f7ec5d7992f812d60e24232_0 | Q: Do they report results only on English data?
Text: Introduction
Improving unsupervised learning is of key importance for advancing machine learning methods, as to unlock access to almost unlimited amounts of data to be used as training resources. The majority of recent success stories of deep learning does not fall ... | Yes |
37eba8c3cfe23778498d95a7dfddf8dfb725f8e2 | 37eba8c3cfe23778498d95a7dfddf8dfb725f8e2_0 | Q: Which other unsupervised models are used for comparison?
Text: Introduction
Improving unsupervised learning is of key importance for advancing machine learning methods, as to unlock access to almost unlimited amounts of data to be used as training resources. The majority of recent success stories of deep learning do... | Sequential (Denoising) Autoencoder, TF-IDF BOW, SkipThought, FastSent, Siamese C-BOW, C-BOW, C-PHRASE, ParagraphVector |
cdf1bf4b202576c39e063921f6b63dc9e4d6b1ff | cdf1bf4b202576c39e063921f6b63dc9e4d6b1ff_0 | Q: What metric is used to measure performance?
Text: Introduction
Improving unsupervised learning is of key importance for advancing machine learning methods, as to unlock access to almost unlimited amounts of data to be used as training resources. The majority of recent success stories of deep learning does not fall i... | Accuracy and F1 score for supervised tasks, Pearson's and Spearman's correlation for unsupervised tasks |
03f4e5ac5a9010191098d6d66ed9bbdfafcbd013 | 03f4e5ac5a9010191098d6d66ed9bbdfafcbd013_0 | Q: How do the n-gram features incorporate compositionality?
Text: Introduction
Improving unsupervised learning is of key importance for advancing machine learning methods, as to unlock access to almost unlimited amounts of data to be used as training resources. The majority of recent success stories of deep learning do... | by also learning source embeddings for not only unigrams but also n-grams present in each sentence, and averaging the n-gram embeddings along with the words |
9a9338d0e74fd315af643335e733445031bd7656 | 9a9338d0e74fd315af643335e733445031bd7656_0 | Q: Which dataset do they use?
Text: Introduction
Language models (LMs) are crucial components in many applications, such as speech recognition and machine translation. The aim of language models is to compute the probability of any given sentence INLINEFORM0 , which can be calculated as DISPLAYFORM0
The task of LMs is ... | AMI IHM meeting corpus |
3103502cf07726d3eeda34f31c0bdf1fc0ae964e | 3103502cf07726d3eeda34f31c0bdf1fc0ae964e_0 | Q: How do Zipf and Herdan-Heap's laws differ?
Text: Introduction
Statistical characterization of languages has been a field of study for decadesBIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. Even simple quantities, like letter frequency, can be used to decode simple substitution cryptogramsBIBREF6, BIBREF7, BIBR... | Zipf's law describes change of word frequency rate, while Heaps-Herdan describes different word number in large texts (assumed that Hepas-Herdan is consequence of Zipf's) |
aaec98481defc4c230f84a64cdcf793d89081a76 | aaec98481defc4c230f84a64cdcf793d89081a76_0 | Q: What was the best performing baseline?
Text: Introduction
The goal of text summarization task is to produce a summary from a set of documents. The summary should retain important information and be reasonably shorter than the original documents BIBREF0 . When the set of documents contains only a single document, the... | Lead-3 |
69b41524dc5820143e45f2f3545cd5c0a70e2922 | 69b41524dc5820143e45f2f3545cd5c0a70e2922_0 | Q: Which approaches did they use?
Text: Introduction
The goal of text summarization task is to produce a summary from a set of documents. The summary should retain important information and be reasonably shorter than the original documents BIBREF0 . When the set of documents contains only a single document, the task is... | SumBasic, Lsa, LexRank, TextRank, Bayes, Hmm, MaxEnt, NeuralSum, Lead-N |
72122e0bc5da1d07c0dadb3401aab2acd748424d | 72122e0bc5da1d07c0dadb3401aab2acd748424d_0 | Q: What is the size of the dataset?
Text: Introduction
The goal of text summarization task is to produce a summary from a set of documents. The summary should retain important information and be reasonably shorter than the original documents BIBREF0 . When the set of documents contains only a single document, the task ... | 20K |
1af4d56eeaf74460ca2c621a2ad8a5d8dbac491c | 1af4d56eeaf74460ca2c621a2ad8a5d8dbac491c_0 | Q: Did they use a crowdsourcing platform for the summaries?
Text: Introduction
The goal of text summarization task is to produce a summary from a set of documents. The summary should retain important information and be reasonably shorter than the original documents BIBREF0 . When the set of documents contains only a si... | No |
3f5f74c39a560b5d916496e05641783c58af2c5d | 3f5f74c39a560b5d916496e05641783c58af2c5d_0 | Q: How are the synthetic examples generated?
Text: Introduction
In the last few years, research in natural text generation (NLG) has made significant progress, driven largely by the neural encoder-decoder paradigm BIBREF0, BIBREF1 which can tackle a wide array of tasks including translation BIBREF2, summarization BIBRE... | Random perturbation of Wikipedia sentences using mask-filling with BERT, backtranslation and randomly drop out |
07f5e360e91b99aa2ed0284d7d6688335ed53778 | 07f5e360e91b99aa2ed0284d7d6688335ed53778_0 | Q: Do they measure the number of created No-Arc long sequences?
Text: Introduction
Greedy transition-based parsers are popular in NLP, as they provide competitive accuracy with high efficiency. They syntactically analyze a sentence by greedily applying transitions, which read it from left to right and produce a depende... | No |
11dde2be9a69a025f2fc29ce647201fb5a4df580 | 11dde2be9a69a025f2fc29ce647201fb5a4df580_0 | Q: By how much does the new parser outperform the current state-of-the-art?
Text: Introduction
Greedy transition-based parsers are popular in NLP, as they provide competitive accuracy with high efficiency. They syntactically analyze a sentence by greedily applying transitions, which read it from left to right and produ... | Proposed method achieves 94.5 UAS and 92.4 LAS compared to 94.3 and 92.2 of best state-of-the -art greedy based parser. Best state-of-the art parser overall achieves 95.8 UAS and 94.6 LAS. |
bcce5eef9ddc345177b3c39c469b4f8934700f80 | bcce5eef9ddc345177b3c39c469b4f8934700f80_0 | Q: Do they evaluate only on English datasets?
Text: Introduction
A cryptocurrency is a digital currency designed to work as a medium of exchange that uses strong cryptography to secure financial transactions, control the creation of additional units, and verify the transfer of assets. They are based on decentralized sy... | Yes |
d3092f78bdbe7e741932e3ddf997e8db42fa044c | d3092f78bdbe7e741932e3ddf997e8db42fa044c_0 | Q: What experimental evaluation is used?
Text: Introduction
A cryptocurrency is a digital currency designed to work as a medium of exchange that uses strong cryptography to secure financial transactions, control the creation of additional units, and verify the transfer of assets. They are based on decentralized systems... | root mean square error between the actual and the predicted price of Bitcoin for every minute |
e2427f182d7cda24eb7197f7998a02bc80550f15 | e2427f182d7cda24eb7197f7998a02bc80550f15_0 | Q: How is the architecture fault-tolerant?
Text: Introduction
A cryptocurrency is a digital currency designed to work as a medium of exchange that uses strong cryptography to secure financial transactions, control the creation of additional units, and verify the transfer of assets. They are based on decentralized syste... | By using Apache Spark which stores all executions in a lineage graph and recovers to the previous steady state from any fault |
0457242fb2ec33446799de229ff37eaad9932f2a | 0457242fb2ec33446799de229ff37eaad9932f2a_0 | Q: Which elements of the platform are modular?
Text: Introduction
A cryptocurrency is a digital currency designed to work as a medium of exchange that uses strong cryptography to secure financial transactions, control the creation of additional units, and verify the transfer of assets. They are based on decentralized s... | handling large volume incoming data, sentiment analysis on tweets and predictive online learning |
5e997d4499b18f1ee1ef6fa145cadbc018b8dd87 | 5e997d4499b18f1ee1ef6fa145cadbc018b8dd87_0 | Q: What is the source of memes?
Text: Motivation
The spread of misinformation or hate messages through social media is a central societal challenge given the unprecedented broadcast potential of these tools. While there already exist some moderation mechanisms such as crowd-sourced abuse reports and dedicated human tea... | Google Images, Reddit Memes Dataset |
12c7d79d2a26af2d445229d0c8ba3ba1aab3f5b5 | 12c7d79d2a26af2d445229d0c8ba3ba1aab3f5b5_0 | Q: Is the dataset multimodal?
Text: Motivation
The spread of misinformation or hate messages through social media is a central societal challenge given the unprecedented broadcast potential of these tools. While there already exist some moderation mechanisms such as crowd-sourced abuse reports and dedicated human teams... | Yes |
98daaa9eaa1e1e574be336b8933b861bfd242e5e | 98daaa9eaa1e1e574be336b8933b861bfd242e5e_0 | Q: How is each instance of the dataset annotated?
Text: Motivation
The spread of misinformation or hate messages through social media is a central societal challenge given the unprecedented broadcast potential of these tools. While there already exist some moderation mechanisms such as crowd-sourced abuse reports and d... | weakly labeled into hate or non-hate memes, depending on their source |
a93196fb0fb5f8202912971e14552fd7828976db | a93196fb0fb5f8202912971e14552fd7828976db_0 | Q: Which dataset do they use for text modelling?
Text: Introduction
Variational Autoencoder (VAE) BIBREF1 is a powerful method for learning representations of high-dimensional data. However, recent attempts of applying VAEs to text modelling are still far less successful compared to its application to image and speech ... | Penn Treebank (PTB), end-to-end (E2E) text generation corpus |
983c2fe7bdbf471bb8b15db858fd2cbec86b96a5 | 983c2fe7bdbf471bb8b15db858fd2cbec86b96a5_0 | Q: Do they compare against state of the art text generation?
Text: Introduction
Variational Autoencoder (VAE) BIBREF1 is a powerful method for learning representations of high-dimensional data. However, recent attempts of applying VAEs to text modelling are still far less successful compared to its application to image... | Yes |
a5418e4af99a2cbd6b7a2b8041388a2d01b8efb2 | a5418e4af99a2cbd6b7a2b8041388a2d01b8efb2_0 | Q: How do they evaluate generated text quality?
Text: Introduction
Variational Autoencoder (VAE) BIBREF1 is a powerful method for learning representations of high-dimensional data. However, recent attempts of applying VAEs to text modelling are still far less successful compared to its application to image and speech B... | Loss analysis. To conduct a more thorough evaluation, we further investigate model behaviours in terms of both reconstruction loss and KL loss, as shown in Figure FIGREF14. These plots were obtained based on the E2E training set using the inputless setting. |
b540cd4fe9dc4394f64d5b76b0eaa4d9e30fb728 | b540cd4fe9dc4394f64d5b76b0eaa4d9e30fb728_0 | Q: Could you tell me more about the metrics used for performance evaluation?
Text: Introduction
With the growing amount of biomedical information available in textual form, there have been significant advances in the development of pre-training language representations that can be applied to a range of different tasks ... | BLUE utilizes different metrics for each of the tasks: Pearson correlation coefficient, F-1 scores, micro-averaging, and accuracy |
41173179efa6186eef17c96f7cbd8acb29105b0e | 41173179efa6186eef17c96f7cbd8acb29105b0e_0 | Q: which tasks are used in BLUE benchmark?
Text: Introduction
With the growing amount of biomedical information available in textual form, there have been significant advances in the development of pre-training language representations that can be applied to a range of different tasks in the biomedical domain, such as ... | Inference task
The aim of the inference task is to predict whether the premise sentence entails or contradicts the hypothesis sentence, Document multilabel classification
The multilabel classification task predicts multiple labels from the texts., Relation extraction
The aim of the relation extraction task is to predic... |
0bd683c51a87a110b68b377e9a06f0a3e12c8da0 | 0bd683c51a87a110b68b377e9a06f0a3e12c8da0_0 | Q: What are the tasks that this method has shown improvements?
Text: Introduction
Word embeddings are one of the most widely used resources in NLP, as they have proven to be of enormous importance for modeling linguistic phenomena in both supervised and unsupervised settings. In particular, the representation of words ... | bilingual dictionary induction, monolingual and cross-lingual word similarity, and cross-lingual hypernym discovery |
a979749e59e6e300a453d8a8b1627f97101799de | a979749e59e6e300a453d8a8b1627f97101799de_0 | Q: Why does the model improve in monolingual spaces as well?
Text: Introduction
Word embeddings are one of the most widely used resources in NLP, as they have proven to be of enormous importance for modeling linguistic phenomena in both supervised and unsupervised settings. In particular, the representation of words i... | because word pair similarity increases if the two words translate to similar parts of the cross-lingual embedding space |
b10632eaa0ca48f86522d8ec38b1d702cb0b8c01 | b10632eaa0ca48f86522d8ec38b1d702cb0b8c01_0 | Q: What are the categories being extracted?
Text: Introduction and Background
Electronic Health Records (EHRs) are organized collections of information about individual patients. They are designed such that they can be shared across different settings for providing health care services. The Institute of Medicine commit... | Unanswerable |
8fa7011e7beaa9fb4083bf7dd75d1216f9c7b2eb | 8fa7011e7beaa9fb4083bf7dd75d1216f9c7b2eb_0 | Q: Do the authors test their annotation projection techniques on tasks other than AMR?
Text: Introduction
Abstract Meaning Representation (AMR) parsing is the process of converting natural language sentences into their corresponding AMR representations BIBREF0 . An AMR is a graph with nodes representing the concepts of... | No |
e0b7acf4292b71725b140f089c6850aebf2828d2 | e0b7acf4292b71725b140f089c6850aebf2828d2_0 | Q: How is annotation projection done when languages have different word order?
Text: Introduction
Abstract Meaning Representation (AMR) parsing is the process of converting natural language sentences into their corresponding AMR representations BIBREF0 . An AMR is a graph with nodes representing the concepts of the sen... | Word alignments are generated for parallel text, and aligned words are assumed to also share AMR node alignments. |
b6ffa18d49e188c454188669987b0a4807ca3018 | b6ffa18d49e188c454188669987b0a4807ca3018_0 | Q: What is the reasoning method that is used?
Text: Introduction
Ontology-based knowledge bases (KBs) like DBpedia BIBREF0 are playing an increasingly important role in domains such knowledge management, data analysis and natural language understanding. Although they are very valuable resources, the usefulness and usab... | SPARQL |
2b61893b22ac190c94c2cb129e86086888347079 | 2b61893b22ac190c94c2cb129e86086888347079_0 | Q: What KB is used in this work?
Text: Introduction
Ontology-based knowledge bases (KBs) like DBpedia BIBREF0 are playing an increasingly important role in domains such knowledge management, data analysis and natural language understanding. Although they are very valuable resources, the usefulness and usability of such... | DBpedia |
a996b6aee9be88a3db3f4127f9f77a18ed10caba | a996b6aee9be88a3db3f4127f9f77a18ed10caba_0 | Q: What's the precision of the system?
Text: Introduction
Ontology-based knowledge bases (KBs) like DBpedia BIBREF0 are playing an increasingly important role in domains such knowledge management, data analysis and natural language understanding. Although they are very valuable resources, the usefulness and usability o... | 0.8320 on semantic typing, 0.7194 on entity matching |
65e2f97f2fe8eb5c2fa41cb95c02b577e8d6e5ee | 65e2f97f2fe8eb5c2fa41cb95c02b577e8d6e5ee_0 | Q: How did they measure effectiveness?
Text: Introduction
Modern speech-based assistants, such as Amazon Alexa, Google Home, Microsoft Cortana, and Apple Siri, enable users to complete daily tasks such as shopping, setting reminders, and playing games using voice commands. Such human-like interfaces create a rich exper... | number of dialogs that resulted in launching a skill divided by total number of dialogs |
83f14af3ccca4ab9deb4c6d208f624d1e79dc7eb | 83f14af3ccca4ab9deb4c6d208f624d1e79dc7eb_0 | Q: Which of the two ensembles yields the best performance?
Text: Introduction
Imagine that you have a friend who claims to know a lot of trivia. During a quiz, you ask them about the native language of actor Jean Marais. They correctly answer French. For a moment you are impressed, until you realize that Jean is a typi... | Answer with content missing: (Table 2) CONCAT ensemble |
0154d8be772193bfd70194110f125813057413a4 | 0154d8be772193bfd70194110f125813057413a4_0 | Q: What are the two ways of ensembling BERT and E-BERT?
Text: Introduction
Imagine that you have a friend who claims to know a lot of trivia. During a quiz, you ask them about the native language of actor Jean Marais. They correctly answer French. For a moment you are impressed, until you realize that Jean is a typical... | mean-pooling their outputs (AVG), concatenating the entity and its name with a slash symbol (CONCAT) |
e737cfe0f6cfc6d3ac6bec32231d9c893bfc3fc9 | e737cfe0f6cfc6d3ac6bec32231d9c893bfc3fc9_0 | Q: How is it determined that a fact is easy-to-guess?
Text: Introduction
Imagine that you have a friend who claims to know a lot of trivia. During a quiz, you ask them about the native language of actor Jean Marais. They correctly answer French. For a moment you are impressed, until you realize that Jean is a typical F... | filter deletes all KB triples where the correct answer (e.g., Apple) is a case-insensitive substring of the subject entity name (e.g., Apple Watch), person name filter uses cloze-style questions to elicit name associations inherent in BERT, and deletes KB triples that correlate with them |
42be49b883eba268e3dbc5c3ff4631442657dcbb | 42be49b883eba268e3dbc5c3ff4631442657dcbb_0 | Q: How is dependency parsing empirically verified?
Text: Introduction
Constituent and dependency are two typical syntactic structure representation forms as shown in Figure FIGREF1, which have been well studied from both linguistic and computational perspective BIBREF0, BIBREF1. In earlier time, linguists and NLP resea... | At last, our model is evaluated on two benchmark treebanks for both constituent and dependency parsing. The empirical results show that our parser reaches new state-of-the-art for all parsing tasks. |
8d4f0815f8a23fe45c298c161fc7a27f3bb0d338 | 8d4f0815f8a23fe45c298c161fc7a27f3bb0d338_0 | Q: How are different network components evaluated?
Text: Introduction
Constituent and dependency are two typical syntactic structure representation forms as shown in Figure FIGREF1, which have been well studied from both linguistic and computational perspective BIBREF0, BIBREF1. In earlier time, linguists and NLP resea... | For different numbers of shared layers, the results are in Table TABREF14. We respectively disable the constituent and the dependency parser to obtain a separate learning setting for both parsers in our model. |
a6665074b067abb2676d5464f36b2cb07f6919d3 | a6665074b067abb2676d5464f36b2cb07f6919d3_0 | Q: What are the performances obtained for PTB and CTB?
Text: Introduction
Constituent and dependency are two typical syntactic structure representation forms as shown in Figure FIGREF1, which have been well studied from both linguistic and computational perspective BIBREF0, BIBREF1. In earlier time, linguists and NLP r... | . On PTB, our model achieves 93.90 F1 score of constituent parsing and 95.91 UAS and 93.86 LAS of dependency parsing., On CTB, our model achieves a new state-of-the-art result on both constituent and dependency parsing. |
b0fbd4b0f02b877a0d3df1d8ccc47d90dd49147c | b0fbd4b0f02b877a0d3df1d8ccc47d90dd49147c_0 | Q: What are the models used to perform constituency and dependency parsing?
Text: Introduction
Constituent and dependency are two typical syntactic structure representation forms as shown in Figure FIGREF1, which have been well studied from both linguistic and computational perspective BIBREF0, BIBREF1. In earlier time... | token representation, self-attention encoder,, Constituent Parsing Decoder, Dependency Parsing Decoder |
3288a50701a80303fd71c8c5ede81cbee14fa2c7 | 3288a50701a80303fd71c8c5ede81cbee14fa2c7_0 | Q: Is the proposed layer smaller in parameters than a Transformer?
Text: Introduction
The capability of deep neural models of handling complex dependencies has benefited various artificial intelligence tasks, such as image recognition where test error was reduced by scaling VGG nets BIBREF0 up to hundreds of convolutio... | No |
22b8836cb00472c9780226483b29771ae3ebdc87 | 22b8836cb00472c9780226483b29771ae3ebdc87_0 | Q: What is the new initialization method proposed in this paper?
Text: Introduction
Named Entity Disambiguation (NED) is the task of linking mentions of entities in text to a given knowledge base, such as Freebase or Wikipedia. NED is a key component in Entity Linking (EL) systems, focusing on the disambiguation task i... | They initialize their word and entity embeddings with vectors pre-trained over a large corpus of unlabeled data. |
540e9db5595009629b2af005e3c06610e1901b12 | 540e9db5595009629b2af005e3c06610e1901b12_0 | Q: How was a quality control performed so that the text is noisy but the annotations are accurate?
Text: Introduction
Named Entity Disambiguation (NED) is the task of linking mentions of entities in text to a given knowledge base, such as Freebase or Wikipedia. NED is a key component in Entity Linking (EL) systems, foc... | The authors believe that the Wikilinks corpus contains ground truth annotations while being noisy. They discard mentions that cannot have ground-truth verified by comparison with Wikipedia. |
bd1a3c651ca2b27f283d3f36df507ed4eb24c2b0 | bd1a3c651ca2b27f283d3f36df507ed4eb24c2b0_0 | Q: Is it a neural model? How is it trained?
Text: Introduction
In active machine learning, a learner is able to query an oracle in order to obtain information that is expected to improve performance. Theoretical and empirical results show that active learning can speed acquisition for a variety of learning tasks BIBREF... | No, it is a probabilistic model trained by finding feature weights through gradient ascent |
5a2c0c55a43dcc0b9439d330d2cbe1d5d444bf36 | 5a2c0c55a43dcc0b9439d330d2cbe1d5d444bf36_0 | Q: How do people engage in Twitter threads on different types of news?
Text: Introduction
Twitter is a social network that has been used worldwide as a means of news spreading. In fact, more than 85% of its users use Twitter to be updated with news, and do so on a daily basis BIBREF0. Users behaviour of this social net... | Unanswerable |
0c78d2fe8bc5491b5fd8a2166190c59eba069ced | 0c78d2fe8bc5491b5fd8a2166190c59eba069ced_0 | Q: How are the clusters related to security, violence and crime identified?
Text: Introduction
Twitter is a social network that has been used worldwide as a means of news spreading. In fact, more than 85% of its users use Twitter to be updated with news, and do so on a daily basis BIBREF0. Users behaviour of this socia... | Yes |
d2473c039ab85f8e9e99066894658381ae852e16 | d2473c039ab85f8e9e99066894658381ae852e16_0 | Q: What are the features of used to customize target user interaction?
Text: Introduction
Recent advances in visual language field enabled by deep learning techniques have succeeded in bridging the gap between vision and language in a variety of tasks, ranging from describing the image BIBREF0 , BIBREF1 , BIBREF2 , BI... | image feature, question feature, label vector for the user's answer |
5d6cc65b73f428ea2a499bcf91995ef5441f63d4 | 5d6cc65b73f428ea2a499bcf91995ef5441f63d4_0 | Q: How they evaluate quality of generated output?
Text: Introduction
The growing interest in Machine Reading Comprehension (MRC) has sparked significant research efforts on Question Generation (QG), the dual task to Question Answering (QA). In QA, the objective is to produce an adequate response given a query and a tex... | Through human evaluation where they are asked to evaluate the generated output on a likert scale. |
0a8bc204a76041a25cee7e9f8e2af332a17da67a | 0a8bc204a76041a25cee7e9f8e2af332a17da67a_0 | Q: What automated metrics authors investigate?
Text: Introduction
The growing interest in Machine Reading Comprehension (MRC) has sparked significant research efforts on Question Generation (QG), the dual task to Question Answering (QA). In QA, the objective is to produce an adequate response given a query and a text; ... | BLEU, Self-BLEU, n-gram based score, probability score |
81686454f215e28987c7ad00ddce5ffe84b37195 | 81686454f215e28987c7ad00ddce5ffe84b37195_0 | Q: What supervised models are experimented with?
Text: Introduction
NLP can be extremely useful for enabling scientific inquiry, helping us to quickly and efficiently understand large corpora, gather evidence, and test hypotheses BIBREF0 , BIBREF1 . One domain for which automated analysis is particularly useful is Inte... | Unanswerable |
fc06502fa62803b62f6fd84265bfcfb207c1113b | fc06502fa62803b62f6fd84265bfcfb207c1113b_0 | Q: Who annotated the data?
Text: Introduction
NLP can be extremely useful for enabling scientific inquiry, helping us to quickly and efficiently understand large corpora, gather evidence, and test hypotheses BIBREF0 , BIBREF1 . One domain for which automated analysis is particularly useful is Internet security: researc... | annotators who were not security experts, researchers in either NLP or computer security |
ce807a42370bfca10fa322d6fa772e4a58a8dca1 | ce807a42370bfca10fa322d6fa772e4a58a8dca1_0 | Q: What are the four forums the data comes from?
Text: Introduction
NLP can be extremely useful for enabling scientific inquiry, helping us to quickly and efficiently understand large corpora, gather evidence, and test hypotheses BIBREF0 , BIBREF1 . One domain for which automated analysis is particularly useful is Inte... | Darkode, Hack Forums, Blackhat and Nulled. |
f91835f17c0086baec65ebd99d12326ae1ae87d2 | f91835f17c0086baec65ebd99d12326ae1ae87d2_0 | Q: How do they obtain parsed source sentences?
Text: Introduction
Neural machine translation (NMT) typically makes use of a recurrent neural network (RNN) -based encoder and decoder, along with an attention mechanism BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 . However, it has been shown that RNNs require some supervision t... | Stanford CoreNLP BIBREF11 |
14e78db206a8180ea637774aa572b073e3ffa219 | 14e78db206a8180ea637774aa572b073e3ffa219_0 | Q: What kind of encoders are used for the parsed source sentence?
Text: Introduction
Neural machine translation (NMT) typically makes use of a recurrent neural network (RNN) -based encoder and decoder, along with an attention mechanism BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 . However, it has been shown that RNNs require... | RNN encoders |
bc1e3f67d607bfc7c4c56d6b9763d3ae7f56ad5b | bc1e3f67d607bfc7c4c56d6b9763d3ae7f56ad5b_0 | Q: Whas is the performance drop of their model when there is no parsed input?
Text: Introduction
Neural machine translation (NMT) typically makes use of a recurrent neural network (RNN) -based encoder and decoder, along with an attention mechanism BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 . However, it has been shown that ... | improvements of up to 1.5 BLEU over the seq2seq baseline |
e8e00b4c0673af5ab02ec82563105e4157cc54bb | e8e00b4c0673af5ab02ec82563105e4157cc54bb_0 | Q: How were their results compared to state-of-the-art?
Text: Introduction
Machine Translation, which is a field of concentrate under common language preparing, focuses at deciphering normal language naturally utilizing machines. Information driven machine interpretation has turned into the overwhelming field of concen... | transformer model achieves higher BLEU score than both Attention encoder-decoder and sequence-sequence model |
18ad60f97f53af64cb9db2123c0d8846c57bfa4a | 18ad60f97f53af64cb9db2123c0d8846c57bfa4a_0 | Q: What supports the claim that injected CNN into recurent units will enhance ability of the model to catch local context and reduce ambiguities?
Text: Introduction
Neural network based approaches have become popular frameworks in many machine learning research fields, showing its advantages over traditional methods. I... | word embeddings to generate a new feature, i.e., summarizing a local context |
87357448ce4cae3c59d4570a19c7a9df4c086bd8 | 87357448ce4cae3c59d4570a19c7a9df4c086bd8_0 | Q: How is CNN injected into recurent units?
Text: Introduction
Neural network based approaches have become popular frameworks in many machine learning research fields, showing its advantages over traditional methods. In NLP tasks, two types of neural networks are widely used: Recurrent Neural Network (RNN) and Convolut... | The most simple one is to directly apply a CNN layer after the embedding layer to obtain blended contextual representations. Then a GRU layer is applied afterward. |
1ccc4f63268aa7841cc6fd23535c9cbe85791007 | 1ccc4f63268aa7841cc6fd23535c9cbe85791007_0 | Q: Are there some results better than state of the art on these tasks?
Text: Introduction
Neural network based approaches have become popular frameworks in many machine learning research fields, showing its advantages over traditional methods. In NLP tasks, two types of neural networks are widely used: Recurrent Neural... | Yes |
afe34e553c3c784dbf02add675b15c27638cdd45 | afe34e553c3c784dbf02add675b15c27638cdd45_0 | Q: Do experiment results show consistent significant improvement of new approach over traditional CNN and RNN models?
Text: Introduction
Neural network based approaches have become popular frameworks in many machine learning research fields, showing its advantages over traditional methods. In NLP tasks, two types of ne... | Yes |
3f46d8082a753265ec2a88ae8f1beb6651e281b6 | 3f46d8082a753265ec2a88ae8f1beb6651e281b6_0 | Q: What datasets are used for testing sentiment classification and reading comprehension?
Text: Introduction
Neural network based approaches have become popular frameworks in many machine learning research fields, showing its advantages over traditional methods. In NLP tasks, two types of neural networks are widely use... | CBT NE/CN, MR Movie reviews, IMDB Movie reviews, SUBJ |
63d9b12dc3ff3ceb1aed83ce11371bca8aac4e8f | 63d9b12dc3ff3ceb1aed83ce11371bca8aac4e8f_0 | Q: So we do not use pre-trained embedding in this case?
Text: Introduction
Encoder-decoder models BIBREF0 are effective in tasks such as machine translation ( BIBREF1 , BIBREF1 ; BIBREF2 , BIBREF2 ) and grammatical error correction BIBREF3 . Vocabulary in encoder-decoder models is generally selected from the training c... | Yes |
0bd864f83626a0c60f5e96b73fb269607afc7c09 | 0bd864f83626a0c60f5e96b73fb269607afc7c09_0 | Q: How are sentence embeddings incorporated into the speech recognition system?
Text: Introduction
In a long conversation, there exists a tendency of semantically related words, or phrases reoccur across sentences, or there exists topical coherence. Existing speech recognition systems are built at individual, isolated ... | BERT generates sentence embeddings that represent words in context. These sentence embeddings are merged into a single conversational-context vector that is used to calculate a gated embedding and is later combined with the output of the decoder h to provide the gated activations for the next hidden layer. |
c77d6061d260f627f2a29a63718243bab5a6ed5a | c77d6061d260f627f2a29a63718243bab5a6ed5a_0 | Q: How different is the dataset size of source and target?
Text: Question Answering
One of the most important characteristics of an intelligent system is to understand stories like humans do. A story is a sequence of sentences, and can be in the form of plain text BIBREF3 , BIBREF4 , BIBREF5 , BIBREF6 or spoken content... | the training dataset is large while the target dataset is usually much smaller |
4c7b29f6e3cc1e902959a1985146ccc0b15fe521 | 4c7b29f6e3cc1e902959a1985146ccc0b15fe521_0 | Q: How do you find the entity descriptions?
Text: Introduction
Knowledge about entities is essential for understanding human language. This knowledge can be attributional (e.g., canFly, isEdible), type-based (e.g., isFood, isPolitician, isDisease) or relational (e.g, marriedTo, bornIn). Knowledge bases (KBs) are design... | Wikipedia |
b34c60eb4738e0439523bcc679fe0fe70ceb8bde | b34c60eb4738e0439523bcc679fe0fe70ceb8bde_0 | Q: How is OpenBookQA different from other natural language QA?
Text: Introduction
Natural language based question answering (NLQA) not only involves linguistic understanding, but often involves reasoning with various kinds of knowledge. In recent years, many NLQA datasets and challenges have been proposed, for example,... | in the OpenBookQA setup the open book part is much larger, the open book part is much larger (than a small paragraph) and is not complete as additional common knowledge may be required |
9623884915b125d26e13e8eeebe9a0f79d56954b | 9623884915b125d26e13e8eeebe9a0f79d56954b_0 | Q: At what text unit/level were documents processed?
Text: Introduction
Business documents broadly characterize a large class of documents that are central to the operation of business. These include legal contracts, purchase orders, financial statements, regulatory filings, and more. Such documents have a number of ch... | documents are segmented into paragraphs and processed at the paragraph level |
77db56fee07b01015a74413ca31f19bea7203f0b | 77db56fee07b01015a74413ca31f19bea7203f0b_0 | Q: What evaluation metric were used for presenting results?
Text: Introduction
Business documents broadly characterize a large class of documents that are central to the operation of business. These include legal contracts, purchase orders, financial statements, regulatory filings, and more. Such documents have a numb... | F$_1$, precision, and recall |
c309e87c9e08cf847f31e554577d6366faec1ea0 | c309e87c9e08cf847f31e554577d6366faec1ea0_0 | Q: Was the structure of regulatory filings exploited when training the model?
Text: Introduction
Business documents broadly characterize a large class of documents that are central to the operation of business. These include legal contracts, purchase orders, financial statements, regulatory filings, and more. Such doc... | No |
81cee2fc6edd9b7bc65bbf6b4aa35782339e6cff | 81cee2fc6edd9b7bc65bbf6b4aa35782339e6cff_0 | Q: What type of documents are supported by the annotation platform?
Text: Introduction
Business documents broadly characterize a large class of documents that are central to the operation of business. These include legal contracts, purchase orders, financial statements, regulatory filings, and more. Such documents have... | Variety of formats supported (PDF, Word...), user can define content elements of document |
79620a2b4b121b6d3edd0f7b1d4a8cc7ada0b516 | 79620a2b4b121b6d3edd0f7b1d4a8cc7ada0b516_0 | Q: What are the state-of-the-art models for the task?
Text: Introduction
Disinformation presents a serious threat to society, as the proliferation of fake news can have a significant impact on an individual's perception of reality. Fake news is a claim or story that is fabricated, with the intention to deceive, often f... | To the best of our knowledge, our method achieves state-of-the-art results in weighted-accuracy and standard accuracy on the dataset |
2555ca85ff6b56bd09c3919aa6b277eb7a4d4631 | 2555ca85ff6b56bd09c3919aa6b277eb7a4d4631_0 | Q: Which datasets are used for evaluation?
Text: Introduction
Semantic composition plays an important role in sentiment analysis of phrases and sentences. This includes detecting the scope and impact of negation in reversing a sentiment's polarity, as well as quantifying the influence of modifiers, such as degree adver... | Stanford Sentiment Treebank |
d028dcef22cdf0e86f62455d083581d025db1955 | d028dcef22cdf0e86f62455d083581d025db1955_0 | Q: What are the strong baselines you have?
Text: Introduction
One of the main challenges in building a Natural Language Understanding (NLU) component for a specific task is the necessary human effort to encode the task's specific knowledge. In traditional NLU components, this was done by creating hand-written rules. In... | optimize single task with no synthetic data |
593e307d9a9d7361eba49484099c7a8147d3dade | 593e307d9a9d7361eba49484099c7a8147d3dade_0 | Q: What are causal attribution networks?
Text: Causal attribution datasets
In this work we compare causal attribution networks derived from three datasets. A causal attribution dataset is a collection of text pairs that reflect cause-effect relationships proposed by humans (for example, “virus causes sickness”). These ... | networks where nodes represent causes and effects, and directed edges represent cause-effect relationships proposed by humans |
6f8881e60fdaca7c1b35a5acc7125994bb1206a3 | 6f8881e60fdaca7c1b35a5acc7125994bb1206a3_0 | Q: How accurate is their predictive model?
Text: Introduction
Urban legends are a genre of modern folklore consisting of stories told as true – and plausible enough to be believed – about some rare and exceptional events that supposedly happened to a real person or in a real place.
Whether urban legends are produced by... | Unanswerable |
6a7370dd12682434248d006ffe0a72228c439693 | 6a7370dd12682434248d006ffe0a72228c439693_0 | Q: How large language sets are able to be explored using this approach?
Text: Introduction
The need to uncover presumed underlying linguistic evolutionary principles and analyse correlation between world's languages has entailed this research. For centuries people have been speculating about the origins of language, ho... | Unanswerable |
a71ebd8dc907d470f6bd3829fa949b15b29a0631 | a71ebd8dc907d470f6bd3829fa949b15b29a0631_0 | Q: how did they ask if a tweet was racist?
Text: 1.1em
Stéphan Tulkens, Lisa Hilte, Elise Lodewyckx, Ben Verhoeven, Walter Daelemans
CLiPS Research Center, University of Antwerp
Prinsstraat 13, 2000, Antwerpen, Belgium
{stephan.tulkens, lisa.hilte, ben.verhoeven, walter.daelemans}@uantwerpen.be,
elise.lodewyckx@student... | if it includes negative utterances, negative generalizations and insults concerning ethnicity, nationality, religion and culture. |
1546356a8c5893dc2d298dcbd96d0307731dd54d | 1546356a8c5893dc2d298dcbd96d0307731dd54d_0 | Q: What other cross-lingual approaches is the model compared to?
Text: Introduction
Morphological analysis (hajivc1998tagging, oflazer1994tagging, inter alia) is the task of predicting fine-grained annotations about the syntactic properties of tokens in a language such as part-of-speech, case, or tense. For instance, i... | The baseline model BIBREF5 we compare with regards the output space of the model as a subset INLINEFORM2 where INLINEFORM3 is the set of all tag sets seen in this training data. |
9f5507a8c835c4671020d7d310fff2930d44e75a | 9f5507a8c835c4671020d7d310fff2930d44e75a_0 | Q: What languages are explored?
Text: Introduction
Morphological analysis (hajivc1998tagging, oflazer1994tagging, inter alia) is the task of predicting fine-grained annotations about the syntactic properties of tokens in a language such as part-of-speech, case, or tense. For instance, in Figure FIGREF2 , the given Port... | Danish/Swedish (da/sv), Russian/Bulgarian (ru/bg), Finnish/Hungarian (fi/hu), Spanish/Portuguese (es/pt) |
96ee62407b1ca2a6538c218781e73e8fbf45094a | 96ee62407b1ca2a6538c218781e73e8fbf45094a_0 | Q: How many human subjects were used in the study?
Text: Introduction
Recent years have seen a rapid increase of robotic deployment, beyond traditional applications in cordoned-off workcells in factories, into new, more collaborative use-cases. For example, social robotics and service robotics have targeted scenarios l... | Unanswerable |
ad0a7fe75db5553652cd25555c6980f497e08113 | ad0a7fe75db5553652cd25555c6980f497e08113_0 | Q: How does the model compute the likelihood of executing to the correction semantic denotation?
Text: Introduction
Semantic parsing is the task of converting natural language utterances into machine-understandable meaning representations or logical forms. The task has attracted much attention in the literature due to ... | By treating logical forms as a latent variable and training a discriminative log-linear model over logical form y given x. |
f268b70b08bd0436de5310e390ca5f38f7636612 | f268b70b08bd0436de5310e390ca5f38f7636612_0 | Q: Which conventional alignment models do they use as guidance?
Text: Introduction
Neural Machine Translation (NMT) has achieved great successes on machine translation tasks recently BIBREF0 , BIBREF1 . Generally, it relies on a recurrent neural network under the Encode-Decode framework: it firstly encodes a source sen... | GIZA++ BIBREF3 or fast_align BIBREF4 |
7aae4533dbf097992f23fb2e0574ec5c891ca236 | 7aae4533dbf097992f23fb2e0574ec5c891ca236_0 | Q: Which dataset do they use?
Text: Introduction
Neural Machine Translation (NMT) has achieved great successes on machine translation tasks recently BIBREF0 , BIBREF1 . Generally, it relies on a recurrent neural network under the Encode-Decode framework: it firstly encodes a source sentence into context vectors and the... | BTEC corpus, the CSTAR03 and IWSLT04 held out sets, the NIST2008 Open Machine Translation Campaign |
c80669cb444a6ec6249b971213b0226f59940a82 | c80669cb444a6ec6249b971213b0226f59940a82_0 | Q: On average, by how much do they reduce the diarization error?
Text: Introduction
Speaker diarization is the task of segmenting an audio recording in time, indexing each segment by speaker identity. In the standard version of the task BIBREF0, the goal is not to identify known speakers, but to co-index segments that ... | Unanswerable |
10045d7dac063013a8447b5a4bc3a3c2f18f9e82 | 10045d7dac063013a8447b5a4bc3a3c2f18f9e82_0 | Q: Do they compare their algorithm to voting without weights?
Text: Introduction
Speaker diarization is the task of segmenting an audio recording in time, indexing each segment by speaker identity. In the standard version of the task BIBREF0, the goal is not to identify known speakers, but to co-index segments that are... | No |
4e4946c023211712c782637fcca523deb126e519 | 4e4946c023211712c782637fcca523deb126e519_0 | Q: How do they assign weights between votes in their DOVER algorithm?
Text: Introduction
Speaker diarization is the task of segmenting an audio recording in time, indexing each segment by speaker identity. In the standard version of the task BIBREF0, the goal is not to identify known speakers, but to co-index segments ... | Unanswerable |
144714fe0d5a2bb7e21a7bf50df39d790ff12916 | 144714fe0d5a2bb7e21a7bf50df39d790ff12916_0 | Q: What are state of the art methods authors compare their work with?
Text: Introduction
Flexibility and ease of access to social media have resulted in the use of online channels for news access by a great number of people. For example, nearly two-thirds of American adults have access to news by online channels BIBRE... | ISOT dataset: LLVM
Liar dataset: Hybrid CNN and LSTM with attention |
f01aa192d97fa3cc36b6e316355dc5da0e9b97dc | f01aa192d97fa3cc36b6e316355dc5da0e9b97dc_0 | Q: What are the baselines model?
Text: Introduction
With more than one hundred thousand new scholarly articles being published each year, there is a rapid growth in the number of citations for the relevant scientific articles. In this context, we highlight the following interesting facts about the process of citing sci... | (i) Uniform, (ii) SVR+W, (iii) SVR+O, (iv) C4.5SSL, (v) GLM |
3d583a0675ad34eb7a46767ef5eba5f0ea898aa9 | 3d583a0675ad34eb7a46767ef5eba5f0ea898aa9_0 | Q: What is the architecture of the model?
Text: Introduction
Code-switching has received a lot of attention from speech and computational linguistic communities especially on how to automatically recognize text from speech and understand the structure within it. This phenomenon is very common in bilingual and multiling... | LSTM |
d7d41a1b8bbb1baece89b28962d23ee4457b9c3a | d7d41a1b8bbb1baece89b28962d23ee4457b9c3a_0 | Q: What languages are explored in the work?
Text: Introduction
Code-switching has received a lot of attention from speech and computational linguistic communities especially on how to automatically recognize text from speech and understand the structure within it. This phenomenon is very common in bilingual and multili... | Mandarin, English |
b458ebca72e3013da3b4064293a0a2b4b5ef1fa6 | b458ebca72e3013da3b4064293a0a2b4b5ef1fa6_0 | Q: What is the state-of-the-art neural coreference resolution model?
Text: Introduction
Natural language processing (NLP) with neural networks has grown in importance over the last few years. They provide state-of-the-art models for tasks like coreference resolution, language modeling, and machine translation BIBREF0 ,... | BIBREF2 , BIBREF1 |
1cbca15405632a2e9d0a7061855642d661e3b3a7 | 1cbca15405632a2e9d0a7061855642d661e3b3a7_0 | Q: How much improvement do they get?
Text: Introduction
Satirical news, which uses parody characterized in a conventional news style, has now become an entertainment on social media. While news satire is claimed to be pure comedic and of amusement, it makes statements on real events often with the aim of attaining soci... | Their GTRS approach got an improvement of 3.89% compared to SVM and 27.91% compared to Pawlak. |
018ef092ffc356a2c0e970ae64ad3c2cf8443288 | 018ef092ffc356a2c0e970ae64ad3c2cf8443288_0 | Q: How large is the dataset?
Text: Introduction
Satirical news, which uses parody characterized in a conventional news style, has now become an entertainment on social media. While news satire is claimed to be pure comedic and of amusement, it makes statements on real events often with the aim of attaining social criti... | 8757 news records |
de4e180f49ff187abc519d01eff14ebcd8149cad | de4e180f49ff187abc519d01eff14ebcd8149cad_0 | Q: What features do they extract?
Text: Introduction
Satirical news, which uses parody characterized in a conventional news style, has now become an entertainment on social media. While news satire is claimed to be pure comedic and of amusement, it makes statements on real events often with the aim of attaining social ... | Inconsistency in Noun Phrase Structures, Inconsistency Between Clauses, Inconsistency Between Named Entities and Noun Phrases, Word Level Feature Using TF-IDF |
bdc1f37c8b5e96e3c29cc02dae4ce80087d83284 | bdc1f37c8b5e96e3c29cc02dae4ce80087d83284_0 | Q: What they use as a metric of finding hot spots in meeting?
Text: Introduction and Prior Work
A definition of the meeting “hot spots” was first introduced in BIBREF2, where it was investigated whether human annotators could reliably identify regions in which participants are “highly involved in the discussion”. The m... | unweighted average recall (UAR) metric |
c54de73b36ab86534d18a295f3711591ce9e1784 | c54de73b36ab86534d18a295f3711591ce9e1784_0 | Q: Is this approach compared to some baseline?
Text: Introduction and Prior Work
A definition of the meeting “hot spots” was first introduced in BIBREF2, where it was investigated whether human annotators could reliably identify regions in which participants are “highly involved in the discussion”. The motivation was t... | No |
fdd9dea06550a2fd0df7a1e6a5109facf3601d76 | fdd9dea06550a2fd0df7a1e6a5109facf3601d76_0 | Q: How big is ICSI meeting corpus?
Text: Introduction and Prior Work
A definition of the meeting “hot spots” was first introduced in BIBREF2, where it was investigated whether human annotators could reliably identify regions in which participants are “highly involved in the discussion”. The motivation was that meetings... | 75 meetings and about 70 hours of real-time audio duration |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.