id
stringclasses
179 values
question
stringlengths
8.75k
85.9k
answer
dict
1909.00694
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is the seed lexicon? Context: <<<Title>>> Minimally Supervised Learning of Affective Events Using Discourse Relations <<<Abstract>>> Recognizing affective events that trigger positive ...
{ "references": [ "seed lexicon consists of positive and negative predicates" ], "type": "extractive" }
1909.00694
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are labels available in dataset for supervision? Context: <<<Title>>> Minimally Supervised Learning of Affective Events Using Discourse Relations <<<Abstract>>> Recognizing affective e...
{ "references": [ "negative,positive" ], "type": "extractive" }
1909.00694
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How large is raw corpus used for training? Context: <<<Title>>> Minimally Supervised Learning of Affective Events Using Discourse Relations <<<Abstract>>> Recognizing affective events that ...
{ "references": [ "100 million sentences" ], "type": "extractive" }
1910.14497
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is embedding quality assessed? Context: <<<Title>>> Probabilistic Bias Mitigation in Word Embeddings <<<Abstract>>> It has been shown that word embeddings derived from large corpora ten...
{ "references": [ "We compare this method of bias mitigation with the no bias mitigation (\"Orig\"), geometric bias mitigation (\"Geo\"), the two pieces of our method alone (\"Prob\" and \"KNN\") and the composite method (\"KNN+Prob\"). We note that the composite method performs reasonably well according the the ...
1912.02481
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What turn out to be more important high volume or high quality data? Context: <<<Title>>> Massive vs. Curated Word Embeddings for Low-Resourced Languages. The Case of Yor\`ub\'a and Twi <<<...
{ "references": [ "only high-quality data helps", "high-quality" ], "type": "extractive" }
1912.02481
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What two architectures are used? Context: <<<Title>>> Massive vs. Curated Word Embeddings for Low-Resourced Languages. The Case of Yor\`ub\'a and Twi <<<Abstract>>> The success of several a...
{ "references": [ "fastText,CWE-LP" ], "type": "extractive" }
2002.02224
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is quality of the citation measured? Context: <<<Title>>> Citation Data of Czech Apex Courts <<<Abstract>>> In this paper, we introduce the citation data of the Czech apex courts (Supre...
{ "references": [ "it is necessary to evaluate the performance of the above mentioned part of the pipeline before proceeding further. The evaluation of the performance is summarised in Table TABREF11. It shows that organising the two models into the pipeline boosted the performance of the reference recognition mo...
2003.06651
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Is the method described in this work a clustering-based method? Context: <<<Title>>> Word Sense Disambiguation for 158 Languages using Word Embeddings Only <...
{ "references": [ "Yes" ], "type": "boolean" }
2003.06651
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How are the different senses annotated/labeled? Context: <<<Title>>> Word Sense Disambiguation for 158 Languages using Word Embeddings Only <<<Abstract>>> Disambiguation of word senses in ...
{ "references": [ "The contexts are manually labelled with WordNet senses of the target words" ], "type": "extractive" }
2003.06651
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Was any extrinsic evaluation carried out? Context: <<<Title>>> Word Sense Disambiguation for 158 Languages using Word Embeddings Only <<<Abstract>>> Disambig...
{ "references": [ "Yes" ], "type": "boolean" }
1910.04269
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Does the model use both spectrogram images and raw waveforms as features? Context: <<<Title>>> Spoken Language Identification using ConvNets <<<Abstract>>> L...
{ "references": [ "No" ], "type": "boolean" }
2001.00137
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they report results only on English datasets? Context: <<<Title>>> Stacked DeBERT: All Attention in Incomplete Data for Text Classification <<<Abstract>>>...
{ "references": [ "Yes" ], "type": "boolean" }
2001.00137
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they test their approach on a dataset without incomplete data? Context: <<<Title>>> Stacked DeBERT: All Attention in Incomplete Data for Text Classificati...
{ "references": [ "No" ], "type": "boolean" }
2001.00137
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Should their approach be applied only when dealing with incomplete data? Context: <<<Title>>> Stacked DeBERT: All Attention in Incomplete Data for Text Class...
{ "references": [ "No" ], "type": "boolean" }
2003.08529
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Did they propose other metrics? Context: <<<Title>>> Diversity, Density, and Homogeneity: Quantitative Characteristic Metrics for Text Collections <<<Abstrac...
{ "references": [ "Yes" ], "type": "boolean" }
2003.08529
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Which real-world datasets did they use? Context: <<<Title>>> Diversity, Density, and Homogeneity: Quantitative Characteristic Metrics for Text Collections <<<Abstract>>> Summarizing data sa...
{ "references": [ "SST-2 (Stanford Sentiment Treebank, version 2),Snips", "SST-2,Snips" ], "type": "extractive" }
2003.08553
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What experiments do the authors present to validate their system? Context: <<<Title>>> QnAMaker: Data to Bot in 2 Minutes <<<Abstract>>> Having a bot for seamless conversations is a much-de...
{ "references": [ " we measure our system's performance for datasets across various domains,evaluations are done by managed judges who understands the knowledge base and then judge user queries relevance to the QA pairs" ], "type": "extractive" }
2003.08553
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What components is the QnAMaker composed of? Context: <<<Title>>> QnAMaker: Data to Bot in 2 Minutes <<<Abstract>>> Having a bot for seamless conversations is a much-desired feature that pr...
{ "references": [ "QnAMaker Portal,QnaMaker Management APIs,Azure Search Index,QnaMaker WebApp,Bot", "QnAMaker Portal,QnaMaker Management APIs,Azure Search Index,QnaMaker WebApp,Bot" ], "type": "extractive" }
1909.12140
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Is the semantic hierarchy representation used for any task? Context: <<<Title>>> DisSim: A Discourse-Aware Syntactic Text Simplification Frameworkfor English...
{ "references": [ "Yes" ], "type": "boolean" }
1909.12140
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are the corpora used for the task? Context: <<<Title>>> DisSim: A Discourse-Aware Syntactic Text Simplification Frameworkfor English and German <<<Abstract>>> We introduce DisSim, a di...
{ "references": [ "For the English version, we performed both a thorough manual analysis and automatic evaluation across three commonly used TS datasets from two different domains,The evaluation of the German version is in progress." ], "type": "extractive" }
2002.11893
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How was the dataset collected? Context: <<<Title>>> CrossWOZ: A Large-Scale Chinese Cross-Domain Task-Oriented Dialogue Dataset <<<Abstract>>> To advance multi-domain (cross-domain) dialogu...
{ "references": [ "Database Construction: we crawled travel information in Beijing from the Web, including Hotel, Attraction, and Restaurant domains (hereafter we name the three domains as HAR domains). Then, we used the metro information of entities in HAR domains to build the metro database. ,Goal Generation: a...
2002.11893
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are the benchmark models? Context: <<<Title>>> CrossWOZ: A Large-Scale Chinese Cross-Domain Task-Oriented Dialogue Dataset <<<Abstract>>> To advance multi-domain (cross-domain) dialogu...
{ "references": [ "BERTNLU from ConvLab-2,a rule-based model (RuleDST) ,TRADE (Transferable Dialogue State Generator) ,a vanilla policy trained in a supervised fashion from ConvLab-2 (SL policy)" ], "type": "extractive" }
2002.11893
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How was the corpus annotated? Context: <<<Title>>> CrossWOZ: A Large-Scale Chinese Cross-Domain Task-Oriented Dialogue Dataset <<<Abstract>>> To advance multi-domain (cross-domain) dialogue...
{ "references": [ "The workers were also asked to annotate both user states and system states,we used some rules to automatically annotate dialogue acts according to user states, system states, and dialogue histories" ], "type": "extractive" }
1909.02764
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Does the paper evaluate any adjustment to improve the predicion accuracy of face and audio features? Context: <<<Title>>> Towards Multimodal Emotion Recognit...
{ "references": [ "No" ], "type": "boolean" }
1909.02764
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is face and audio data analysis evaluated? Context: <<<Title>>> Towards Multimodal Emotion Recognition in German Speech Events in Cars using Transfer Learning <<<Abstract>>> The recogni...
{ "references": [ "confusion matrices,$\\text{F}_1$ score" ], "type": "extractive" }
1909.02764
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are the emotion detection tools used for audio and face input? Context: <<<Title>>> Towards Multimodal Emotion Recognition in German Speech Events in Cars using Transfer Learning <<<Ab...
{ "references": [ "We apply an off-the-shelf tool for emotion recognition (the manufacturer cannot be disclosed due to licensing restrictions)", "cannot be disclosed due to licensing restrictions" ], "type": "extractive" }
1912.01252
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are the causal mapping methods employed? Context: <<<Title>>> Facilitating on-line opinion dynamics by mining expressions of causation. The case of climate change debates on The Guardi...
{ "references": [ "Axelrod's causal mapping method" ], "type": "extractive" }
1909.00578
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What dataset do they use? Context: <<<Title>>> SUM-QE: a BERT-based Summary Quality Estimation Model <<<Abstract>>> We propose SumQE, a novel Quality Estimation model for summarization base...
{ "references": [ "datasets from the NIST DUC-05, DUC-06 and DUC-07 shared tasks" ], "type": "extractive" }
1909.00578
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What simpler models do they look at? Context: <<<Title>>> SUM-QE: a BERT-based Summary Quality Estimation Model <<<Abstract>>> We propose SumQE, a novel Quality Estimation model for summari...
{ "references": [ "BiGRU s with attention,ROUGE,Language model (LM),Next sentence prediction" ], "type": "extractive" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What additional techniques are incorporated? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making computer programming l...
{ "references": [ "incorporating coding syntax tree model" ], "type": "extractive" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What dataset do they use? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making computer programming language more unders...
{ "references": [ " text-code parallel corpus" ], "type": "extractive" }
1910.11471
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they compare to other models? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making co...
{ "references": [ "No" ], "type": "boolean" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is the architecture of the system? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making computer programming langua...
{ "references": [ "seq2seq translation" ], "type": "extractive" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What additional techniques could be incorporated to further improve accuracy? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract...
{ "references": [ "phrase-based word embedding,Abstract Syntax Tree(AST)" ], "type": "extractive" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What programming language is target language? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making computer programming ...
{ "references": [ "Python" ], "type": "extractive" }
1910.11471
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What dataset is used to measure accuracy? Context: <<<Title>>> Machine Translation from Natural Language to Code using Long-Short Term Memory <<<Abstract>>> Making computer programming lang...
{ "references": [ "validation data" ], "type": "extractive" }
1910.09399
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Is text-to-image synthesis trained is suppervized or unsuppervized manner? Context: <<<Title>>> A Survey and Taxonomy of Adversarial Neural Networks for Text-to-Image Synthesis <<<Abstract>...
{ "references": [ "unsupervised ", "Even though natural language and image synthesis were part of several contributions on the supervised side of deep learning, unsupervised learning saw recently a tremendous rise in input from the research community specially on two subproblems: text-based natural language a...
1910.09399
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What challenges remain unresolved? Context: <<<Title>>> A Survey and Taxonomy of Adversarial Neural Networks for Text-to-Image Synthesis <<<Abstract>>> Text-to-image synthesis refers to com...
{ "references": [ "give more independence to the several learning methods (e.g. less human intervention) involved in the studies,increasing the size of the output images" ], "type": "extractive" }
1910.09399
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is the conclusion of comparison of proposed solution? Context: <<<Title>>> A Survey and Taxonomy of Adversarial Neural Networks for Text-to-Image Synthesis <<<Abstract>>> Text-to-image...
{ "references": [ "HDGAN produced relatively better visual results on the CUB and Oxford datasets while AttnGAN produced far more impressive results than the rest on the more complex COCO dataset,In terms of inception score (IS), which is the metric that was applied to majority models except DC-GAN, the results i...
1910.04601
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is the baseline? Context: <<<Title>>> RC-QED: Evaluating Natural Language Derivations in Multi-Hop Reading Comprehension <<<Abstract>>> Recent studies revealed that reading comprehensi...
{ "references": [ " path ranking-based KGC (PRKGC)" ], "type": "extractive" }
1910.04601
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What dataset was used in the experiment? Context: <<<Title>>> RC-QED: Evaluating Natural Language Derivations in Multi-Hop Reading Comprehension <<<Abstract>>> Recent studies revealed that ...
{ "references": [ "WikiHop" ], "type": "extractive" }
1910.04601
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Did they use any crowdsourcing platform? Context: <<<Title>>> RC-QED: Evaluating Natural Language Derivations in Multi-Hop Reading Comprehension <<<Abstract>...
{ "references": [ "Yes" ], "type": "boolean" }
1910.04601
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How was the dataset annotated? Context: <<<Title>>> RC-QED: Evaluating Natural Language Derivations in Multi-Hop Reading Comprehension <<<Abstract>>> Recent studies revealed that reading co...
{ "references": [ "True, Likely (i.e. Answerable), or Unsure (i.e. Unanswerable),why they are unsure from two choices (“Not stated in the article” or “Other”),The “summary” text boxes" ], "type": "extractive" }
1912.05066
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How many label options are there in the multi-label task? Context: <<<Title>>> Event Outcome Prediction using Sentiment Analysis and Crowd Wisdom in Microblog Feeds <<<Abstract>>> Sentiment...
{ "references": [ " two labels " ], "type": "extractive" }
1912.05066
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Who are the experts? Context: <<<Title>>> Event Outcome Prediction using Sentiment Analysis and Crowd Wisdom in Microblog Feeds <<<Abstract>>> Sentiment Analysis of microblog feeds has attr...
{ "references": [ "political pundits of the Washington Post", "the experts in the field" ], "type": "extractive" }
1912.05066
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Who is the crowd in these experiments? Context: <<<Title>>> Event Outcome Prediction using Sentiment Analysis and Crowd Wisdom in Microblog Feeds <<<Abstract>>> Sentiment Analysis of microb...
{ "references": [ " peoples' sentiments expressed over social media" ], "type": "extractive" }
1912.05066
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How do you establish the ground truth of who won a debate? Context: <<<Title>>> Event Outcome Prediction using Sentiment Analysis and Crowd Wisdom in Microblog Feeds <<<Abstract>>> Sentimen...
{ "references": [ "experts in Washington Post" ], "type": "extractive" }
1910.03891
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What further analysis is done? Context: <<<Title>>> Learning High-order Structural and Attribute information by Knowledge Graph Attention Networks for Enhancing Knowledge Graph Embedding <<...
{ "references": [ "we use t-SNE tool BIBREF27 to visualize the learned embedding" ], "type": "extractive" }
1910.03891
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What seven state-of-the-art methods are used for comparison? Context: <<<Title>>> Learning High-order Structural and Attribute information by Knowledge Graph Attention Networks for Enhancin...
{ "references": [ "TransE, TransR and TransH,PTransE, and ALL-PATHS,R-GCN BIBREF24 and KR-EAR BIBREF26" ], "type": "extractive" }
1910.03891
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What three datasets are used to measure performance? Context: <<<Title>>> Learning High-order Structural and Attribute information by Knowledge Graph Attention Networks for Enhancing Knowle...
{ "references": [ "FB24K,DBP24K,Game30K", "Freebase BIBREF0, DBpedia BIBREF1 and a self-construction game knowledge graph" ], "type": "extractive" }
1910.03891
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How does KANE capture both high-order structural and attribute information of KGs in an efficient, explicit and unified manner? Context: <<<Title>>> Learning High-order Structural and Attri...
{ "references": [ "To capture both high-order structural information of KGs, we used an attention-based embedding propagation method." ], "type": "extractive" }
1910.03891
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are recent works on knowedge graph embeddings authors mention? Context: <<<Title>>> Learning High-order Structural and Attribute information by Knowledge Graph Attention Networks for E...
{ "references": [ "entity types or concepts BIBREF13,relations paths BIBREF17, textual descriptions BIBREF11, BIBREF12,logical rules BIBREF23,deep neural network models BIBREF24" ], "type": "extractive" }
1909.13375
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How they use sequence tagging to answer multi-span questions? Context: <<<Title>>> Tag-based Multi-Span Extraction in Reading Comprehension <<<Abstract>>> With models reaching human perform...
{ "references": [ "To model an answer which is a collection of spans, the multi-span head uses the $\\mathtt {BIO}$ tagging format BIBREF8: $\\mathtt {B}$ is used to mark the beginning of a span, $\\mathtt {I}$ is used to mark the inside of a span and $\\mathtt {O}$ is used to mark tokens not included in a span" ...
1909.13375
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is the previous model that attempted to tackle multi-span questions as a part of its design? Context: <<<Title>>> Tag-based Multi-Span Extraction in Reading Comprehension <<<Abstract>>...
{ "references": [ "MTMSN BIBREF4" ], "type": "extractive" }
1910.00912
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Which publicly available NLU dataset is used? Context: <<<Title>>> Hierarchical Multi-Task Natural Language Understanding for Cross-domain Conversational AI: HERMIT NLU <<<Abstract>>> We pr...
{ "references": [ "ROMULUS dataset,NLU-Benchmark dataset" ], "type": "extractive" }
1910.00912
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What metrics other than entity tagging are compared? Context: <<<Title>>> Hierarchical Multi-Task Natural Language Understanding for Cross-domain Conversational AI: HERMIT NLU <<<Abstract>>...
{ "references": [ "We also report the metrics in BIBREF7 for consistency,we report the span F1, Exact Match (EM) accuracy of the entire sequence of labels,metric that combines intent and entities" ], "type": "extractive" }
1908.10449
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they provide decision sequences as supervision while training models? Context: <<<Title>>> Interactive Machine Comprehension with Information Seeking Agen...
{ "references": [ "No" ], "type": "boolean" }
1908.10449
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How do they train models in this setup? Context: <<<Title>>> Interactive Machine Comprehension with Information Seeking Agents <<<Abstract>>> Existing machine reading comprehension (MRC) mo...
{ "references": [ "Thus, our task requires models to `feed themselves' rather than spoon-feeding them with information. This casts MRC as a sequential decision-making problem amenable to reinforcement learning (RL)." ], "type": "extractive" }
1908.10449
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What commands does their setup provide to models seeking information? Context: <<<Title>>> Interactive Machine Comprehension with Information Seeking Agents <<<Abstract>>> Existing machine ...
{ "references": [ "previous,next,Ctrl+F $<$query$>$,stop" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What models do they propose? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of hate speech detection in mu...
{ "references": [ "Feature Concatenation Model (FCM),Spatial Concatenation Model (SCM),Textual Kernels Model (TKM)" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How large is the dataset? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of hate speech detection in multi...
{ "references": [ " $150,000$ tweets" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is author's opinion on why current multimodal models cannot outperform models analyzing only text? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Ab...
{ "references": [ "Noisy data,Complexity and diversity of multimodal relations,Small set of multimodal examples" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What metrics are used to benchmark the results? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of hate spe...
{ "references": [ "F-score,Area Under the ROC Curve (AUC),mean accuracy (ACC),Precision vs Recall plot,ROC curve (which plots the True Positive Rate vs the False Positive Rate)" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is data collected, manual collection or Twitter api? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of...
{ "references": [ "Twitter API" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How many tweats does MMHS150k contains, 150000? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of hate spe...
{ "references": [ "$150,000$ tweets" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What unimodal detection models were used? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the problem of hate speech de...
{ "references": [ " single layer LSTM with a 150-dimensional hidden state for hate / not hate classification" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What different models for multimodal detection were proposed? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we target the probl...
{ "references": [ "Feature Concatenation Model (FCM),Spatial Concatenation Model (SCM),Textual Kernels Model (TKM)" ], "type": "extractive" }
1910.03814
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What annotations are available in the dataset - tweat used hate speach or not? Context: <<<Title>>> Exploring Hate Speech Detection in Multimodal Publications <<<Abstract>>> In this work we...
{ "references": [ "No attacks to any community, racist,sexist,homophobic,religion based attacks,attacks to other communities" ], "type": "extractive" }
1912.00871
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Does pre-training on general text corpus improve performance? Context: <<<Title>>> Solving Arithmetic Word Problems Automatically Using Transformer and Unamb...
{ "references": [ "No" ], "type": "boolean" }
1912.00871
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What neural configurations are explored? Context: <<<Title>>> Solving Arithmetic Word Problems Automatically Using Transformer and Unambiguous Representations <<<Abstract>>> Constructing ac...
{ "references": [ "tried many configurations of our network models, but report results with only three configurations,Transformer Type 1,Transformer Type 2,Transformer Type 3" ], "type": "extractive" }
1912.00871
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Are the Transformers masked? Context: <<<Title>>> Solving Arithmetic Word Problems Automatically Using Transformer and Unambiguous Representations <<<Abstrac...
{ "references": [ "Yes" ], "type": "boolean" }
1912.00871
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is this problem evaluated? Context: <<<Title>>> Solving Arithmetic Word Problems Automatically Using Transformer and Unambiguous Representations <<<Abstract>>> Constructing accurate and...
{ "references": [ "BLEU-2,average accuracies over 3 test trials on different randomly sampled test sets" ], "type": "extractive" }
1912.00871
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What datasets do they use? Context: <<<Title>>> Solving Arithmetic Word Problems Automatically Using Transformer and Unambiguous Representations <<<Abstract>>> Constructing accurate and aut...
{ "references": [ "AI2 BIBREF2,CC BIBREF19,IL BIBREF4,MAWPS BIBREF20" ], "type": "extractive" }
1911.11750
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What representations for textual documents do they use? Context: <<<Title>>> A Measure of Similarity in Textual Data Using Spearman's Rank Correlation Coefficient <<<Abstract>>> In the last...
{ "references": [ "finite sequence of terms" ], "type": "extractive" }
1911.11750
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Which dataset(s) do they use? Context: <<<Title>>> A Measure of Similarity in Textual Data Using Spearman's Rank Correlation Coefficient <<<Abstract>>> In the last decade, many diverse adva...
{ "references": [ "14 TDs,BIBREF15" ], "type": "extractive" }
1911.11750
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How do they evaluate knowledge extraction performance? Context: <<<Title>>> A Measure of Similarity in Textual Data Using Spearman's Rank Correlation Coefficient <<<Abstract>>> In the last ...
{ "references": [ "SRCC" ], "type": "extractive" }
1911.03894
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What is CamemBERT trained on? Context: <<<Title>>> CamemBERT: a Tasty French Language Model <<<Abstract>>> Pretrained language models are now ubiquitous in Natural Language Processing. Desp...
{ "references": [ "unshuffled version of the French OSCAR corpus" ], "type": "extractive" }
1911.03894
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Which tasks does CamemBERT not improve on? Context: <<<Title>>> CamemBERT: a Tasty French Language Model <<<Abstract>>> Pretrained language models are now ubiquitous in Natural Language Pro...
{ "references": [ "its performance still lags behind models trained on the original English training set in the TRANSLATE-TEST setting, 81.2 vs. 82.91 for RoBERTa" ], "type": "extractive" }
1911.03894
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How much better was results of CamemBERT than previous results on these tasks? Context: <<<Title>>> CamemBERT: a Tasty French Language Model <<<Abstract>>> Pretrained language models are no...
{ "references": [ "2.36 point increase in the F1 score with respect to the best SEM architecture,on the TRANSLATE-TRAIN setting (81.2 vs. 80.2 for XLM),lags behind models trained on the original English training set in the TRANSLATE-TEST setting, 81.2 vs. 82.91 for RoBERTa,For POS tagging, we observe error reduct...
1911.03894
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Was CamemBERT compared against multilingual BERT on these tasks? Context: <<<Title>>> CamemBERT: a Tasty French Language Model <<<Abstract>>> Pretrained lang...
{ "references": [ "Yes" ], "type": "boolean" }
1911.03894
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What data is used for training CamemBERT? Context: <<<Title>>> CamemBERT: a Tasty French Language Model <<<Abstract>>> Pretrained language models are now ubiquitous in Natural Language Proc...
{ "references": [ "unshuffled version of the French OSCAR corpus" ], "type": "extractive" }
1912.01673
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Is this dataset publicly available? Context: <<<Title>>> COSTRA 1.0: A Dataset of Complex Sentence Transformations <<<Abstract>>> COSTRA 1.0 is a dataset of ...
{ "references": [ "Yes" ], "type": "boolean" }
1912.01673
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Are some baseline models trained on this dataset? Context: <<<Title>>> COSTRA 1.0: A Dataset of Complex Sentence Transformations <<<Abstract>>> COSTRA 1.0 is...
{ "references": [ "Yes" ], "type": "boolean" }
1912.01673
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they do any analysis of of how the modifications changed the starting set of sentences? Context: <<<Title>>> COSTRA 1.0: A Dataset of Complex Sentence Tra...
{ "references": [ "Yes" ], "type": "boolean" }
1912.01673
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How do they introduce language variation? Context: <<<Title>>> COSTRA 1.0: A Dataset of Complex Sentence Transformations <<<Abstract>>> COSTRA 1.0 is a dataset of Czech complex sentence tra...
{ "references": [ " we were looking for original and uncommon sentence change suggestions" ], "type": "extractive" }
1912.01673
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Do they use external resources to make modifications to sentences? Context: <<<Title>>> COSTRA 1.0: A Dataset of Complex Sentence Transformations <<<Abstract...
{ "references": [ "No" ], "type": "boolean" }
1909.00088
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Does the model proposed beat the baseline models for all the values of the masking parameter tested? Context: <<<Title>>> Keep Calm and Switch On! Preserving...
{ "references": [ "No" ], "type": "boolean" }
1909.00088
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Has STES been previously used in the literature to evaluate similar tasks? Context: <<<Title>>> Keep Calm and Switch On! Preserving Sentiment and Fluency in ...
{ "references": [ "No" ], "type": "boolean" }
1909.00088
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are the baseline models mentioned in the paper? Context: <<<Title>>> Keep Calm and Switch On! Preserving Sentiment and Fluency in Semantic Text Exchange <<<Abstract>>> In this paper, w...
{ "references": [ "Noun WordNet Semantic Text Exchange Model (NWN-STEM),General WordNet Semantic Text Exchange Model (GWN-STEM),Word2Vec Semantic Text Exchange Model (W2V-STEM)" ], "type": "extractive" }
1911.03385
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Is this style generator compared to some baseline? Context: <<<Title>>> Low-Level Linguistic Controls for Style Transfer and Content Preservation <<<Abstract...
{ "references": [ "Yes" ], "type": "boolean" }
1911.03385
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How they perform manual evaluation, what is criteria? Context: <<<Title>>> Low-Level Linguistic Controls for Style Transfer and Content Preservation <<<Abstract>>> Despite the success of st...
{ "references": [ "accuracy" ], "type": "extractive" }
1911.03385
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What metrics are used for automatic evaluation? Context: <<<Title>>> Low-Level Linguistic Controls for Style Transfer and Content Preservation <<<Abstract>>> Despite the success of style tr...
{ "references": [ "classification accuracy,BLEU scores,model perplexities of the reconstruction" ], "type": "extractive" }
1911.03385
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How they know what are content words? Context: <<<Title>>> Low-Level Linguistic Controls for Style Transfer and Content Preservation <<<Abstract>>> Despite the success of style transfer in ...
{ "references": [ " words found in the control word lists are then removed,The remaining words, which represent the content" ], "type": "extractive" }
1911.03385
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How they model style as a suite of low-level linguistic controls, such as frequency of pronouns, prepositions, and subordinate clause constructions? Context: <<<Title>>> Low-Level Linguisti...
{ "references": [ "style of a sentence is represented as a vector of counts of closed word classes (like personal pronouns) as well as counts of syntactic features like the number of SBAR non-terminals in its constituency parse, since clause structure has been shown to be indicative of style" ], "type": "extr...
2001.07209
Please answer the following question with yes or no based on the given text. You only need to output 'Yes' or 'No' without any additional explanation. Question: Does the paper discuss previous models which have been applied to the same task? Context: <<<Title>>> Text-based inference of moral sentiment change <<<Abstr...
{ "references": [ "Yes" ], "type": "boolean" }
2001.07209
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How does the parameter-free model work? Context: <<<Title>>> Text-based inference of moral sentiment change <<<Abstract>>> We present a text-based framework for investigating moral sentimen...
{ "references": [ "A Centroid model summarizes each set of seed words by its expected vector in embedding space, and classifies concepts into the class of closest expected embedding in Euclidean distance following a softmax rule;,A Naïve Bayes model considers both mean and variance, under the assumption of indepe...
2001.07209
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: Which fine-grained moral dimension examples do they showcase? Context: <<<Title>>> Text-based inference of moral sentiment change <<<Abstract>>> We present a text-based framework for invest...
{ "references": [ "Care / Harm, Fairness / Cheating, Loyalty / Betrayal, Authority / Subversion, and Sanctity / Degradation" ], "type": "extractive" }
2001.10161
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How well did the system do? Context: <<<Title>>> Bringing Stories Alive: Generating Interactive Fiction Worlds <<<Abstract>>> World building forms the foundation of any task that requires n...
{ "references": [ "the neural approach is generally preferred by a greater percentage of participants than the rules or random,human-made game outperforms them all" ], "type": "extractive" }
2001.10161
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: How is the information extracted? Context: <<<Title>>> Bringing Stories Alive: Generating Interactive Fiction Worlds <<<Abstract>>> World building forms the foundation of any task that requ...
{ "references": [ "neural question-answering technique to extract relations from a story text,OpenIE5, a commonly used rule-based information extraction technique" ], "type": "extractive" }
1909.00279
Please extract a concise answer without any additional explanation for the following question based on the given text. Question: What are some guidelines in writing input vernacular so model can generate Context: <<<Title>>> Generating Classical Chinese Poems from Vernacular Chinese <<<Abstract>>> Classical Chinese ...
{ "references": [ " if a vernacular paragraph contains more poetic images used in classical literature, its generated poem usually achieves higher score,poems generated from descriptive paragraphs achieve higher scores than from logical or philosophical paragraphs" ], "type": "extractive" }