prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the SPN model in the Joint Entity and Relation Extraction with Set Prediction Networks paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the TDEER model in the TDEER: An Efficient Translating Decoding Schema for Joint Extraction of Entities and Relations paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the RIFRE model in the Representation Iterative Fusion based on Heterogeneous Graph Neural Network for Joint Entity and Relation Extraction paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the TPLinker model in the TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the HBT (CasRel) model in the A Novel Cascade Binary Tagging Framework for Relational Triple Extraction paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the RIN (BERT, K=2) model in the Recurrent Interaction Network for Jointly Extracting Entities and Classifying Relations paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the CGT(UniLM) model in the Contrastive Triple Extraction with Generative Transformer paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the ETL-Span model in the Joint Extraction of Entities and Relations Based on a Novel Decomposition Strategy paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the RSAN model in the A Relation-Specific Attention Network for Joint Entity and Relation Extraction paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the CopyRE' OneDecoder model in the CopyMTL: Copy Mechanism for Joint Extraction of Entities and Relations with Multi-Task Learning paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the CopyRE MultiDecoder model in the Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the NovelTagging model in the Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the ETL-Span model in the Joint Extraction of Entities and Relations Based on a Novel Decomposition Strategy paper on the NYT-single dataset? | F1 |
What metrics were used to measure the PA-LSTM model in the Joint extraction of entities and overlapping relations using position-attentive sequence labeling paper on the NYT-single dataset? | F1 |
What metrics were used to measure the NovelTagging model in the Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme paper on the NYT-single dataset? | F1 |
What metrics were used to measure the GeoLayoutLM model in the GeoLayoutLM: Geometric Pre-training for Visual Information Extraction paper on the FUNSD dataset? | F1 |
What metrics were used to measure the LayoutLMv3 large model in the LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking paper on the FUNSD dataset? | F1 |
What metrics were used to measure the BROS model in the BROS: A Pre-trained Language Model Focusing on Text and Layout for Better Key Information Extraction from Documents paper on the FUNSD dataset? | F1 |
What metrics were used to measure the LayoutLMv2 large model in the LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding paper on the FUNSD dataset? | F1 |
What metrics were used to measure the LayoutLM model in the LayoutLM: Pre-training of Text and Layout for Document Image Understanding paper on the FUNSD dataset? | F1 |
What metrics were used to measure the SciBERT (SciVocab) model in the SciBERT: A Pretrained Language Model for Scientific Text paper on the SciERC dataset? | F1, NER Micro F1, RE+ Micro F1 |
What metrics were used to measure the SciBERT (Base Vocab) model in the SciBERT: A Pretrained Language Model for Scientific Text paper on the SciERC dataset? | F1, NER Micro F1, RE+ Micro F1 |
What metrics were used to measure the PFN model in the A Partition Filter Network for Joint Entity and Relation Extraction paper on the SciERC dataset? | F1, NER Micro F1, RE+ Micro F1 |
What metrics were used to measure the SciBERT (SciVocab) model in the SciBERT: A Pretrained Language Model for Scientific Text paper on the JNLPBA dataset? | F1 |
What metrics were used to measure the PL-Marker model in the Packed Levitated Marker for Entity and Relation Extraction paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the PFN model in the A Partition Filter Network for Joint Entity and Relation Extraction paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Ours: cross-sentence ALB model in the A Frustratingly Easy Approach for Entity and Relation Extraction paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Table-Sequence model in the Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Multi-turn QA model in the Entity-Relation Extraction as Multi-Turn Question Answering paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the SPTree model in the End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the multi-head + AT model in the Adversarial training for multi-context joint entity and relation extraction paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the multi-head model in the Joint entity recognition and relation extraction as a multi-head selection problem paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Attention model in the Going out on a limb: Joint Extraction of Entity Mentions and Relations without Dependency Trees paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Joint w/ Global model in the Incremental Joint Extraction of Entity Mentions and Relations paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the DyGIE model in the A General Framework for Information Extraction using Dynamic Span Graphs paper on the ACE 2004 dataset? | RE+ Micro F1, RE Micro F1, NER Micro F1, Cross Sentence |
What metrics were used to measure the Baseline model in the WNUT-2020 Task 1 Overview: Extracting Entities and Relations from Wet Lab Protocols paper on the WNUT 2020 dataset? | F1, Precision, Recall |
What metrics were used to measure the SVM with GloVe model in the SUNNYNLP at SemEval-2018 Task 10: A Support-Vector-Machine-Based Method for Detecting Semantic Difference using Taxonomy and Word Embedding Features paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the SVM with ConceptNet, Wikipedia articles and WordNet synonyms model in the Luminoso at SemEval-2018 Task 10: Distinguishing Attributes Using Text Corpora and Relational Knowledge paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the Gradient boosting with co-occurrence count features and JoBimText features model in the BomJi at SemEval-2018 Task 10: Combining Vector-, Pattern- and Graph-based Information to Identify Discriminative Attributes paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the LexVec, word co-occurrence, and ConceptNet data combined using maximum entropy classifier model in the UWB at SemEval-2018 Task 10: Capturing Discriminative Attributes from Word Distributions paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the Composes explicit vector spaces from WordNet Definitions, ConceptNet and Visual Genome model in the Identifying and Explaining Discriminative Attributes paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the Use of Wikipedia and ConceptNet Transp. (No expl.) model in the ELiRF-UPV at SemEval-2018 Task 10: Capturing Discriminative Attributes with Knowledge Graphs and Wikipedia paper on the SemEval 2018 Task 10 dataset? | F1-Score |
What metrics were used to measure the Spark NLP model in the Deeper Clinical Document Understanding Using Relation Extraction paper on the 2012 i2b2 Temporal Relations dataset? | Macro F1 |
What metrics were used to measure the Spark NLP model in the Deeper Clinical Document Understanding Using Relation Extraction paper on the 2018 n2c2 posology dataset? | Macro F1 |
What metrics were used to measure the RERE model in the Revisiting the Negative Data of Distantly Supervised Relation Extraction paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the TPLinker model in the TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the ReRe (exact) model in the Revisiting the Negative Data of Distantly Supervised Relation Extraction paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the TPLinker model in the TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the CasRel model in the A Novel Cascade Binary Tagging Framework for Relational Triple Extraction paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the HRL model in the A Hierarchical Framework for Relation Extraction with Reinforcement Learning paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the HRL model in the Revisiting the Negative Data of Distantly Supervised Relation Extraction paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the SPTree model in the End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the CasRel (exact) model in the A Novel Cascade Binary Tagging Framework for Relational Triple Extraction paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the NovelTagging model in the Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the Cotype model in the CoType: Joint Extraction of Typed Entities and Relations with Knowledge Bases paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the CopyR model in the Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism paper on the NYT11-HRL dataset? | F1 |
What metrics were used to measure the UniRel model in the UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the REBEL model in the REBEL: Relation Extraction By End-to-end Language generation paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the REBEL (no pre-training) model in the REBEL: Relation Extraction By End-to-end Language generation paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the DIRECT model in the Adjacency List Oriented Relational Fact Extraction via Adaptive Multi-task Learning paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the SPN model in the Joint Entity and Relation Extraction with Set Prediction Networks paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the TDEER model in the TDEER: An Efficient Translating Decoding Schema for Joint Extraction of Entities and Relations paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the PFN model in the A Partition Filter Network for Joint Entity and Relation Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the RIFRE model in the Representation Iterative Fusion based on Heterogeneous Graph Neural Network for Joint Entity and Relation Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the TPLinker model in the TPLinker: Single-stage Joint Extraction of Entities and Relations Through Token Pair Linking paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the PCNN+RL+HME model in the RH-Net: Improving Neural Relation Extraction via Reinforcement Learning and Hierarchical Relational Searching paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the HBT(CasRel) model in the A Novel Cascade Binary Tagging Framework for Relational Triple Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the CGT(UniLM) model in the Contrastive Triple Extraction with Generative Transformer paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the BiTT model in the A Bidirectional Tree Tagging Scheme for Joint Medical Relation Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the RIN (BERT, K=2) model in the Recurrent Interaction Network for Jointly Extracting Entities and Classifying Relations paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the RSAN model in the A Relation-Specific Attention Network for Joint Entity and Relation Extraction paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the ETL-Span model in the Joint Extraction of Entities and Relations Based on a Novel Decomposition Strategy paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the CopyRE' OneDecoder model in the CopyMTL: Copy Mechanism for Joint Extraction of Entities and Relations with Multi-Task Learning paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the CopyRE MultiDecoder model in the Extracting Relational Facts by an End-to-End Neural Model with Copy Mechanism paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the PA model in the Joint extraction of entities and overlapping relations using position-attentive sequence labeling paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the NovelTagging model in the Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme paper on the NYT dataset? | F1, F1 (strict), NER Micro F1 |
What metrics were used to measure the REBEL model in the REBEL: Relation Extraction By End-to-end Language generation paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Table-Sequence model in the Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the SpERT model in the Span-based Joint Entity and Relation Extraction with Transformer Pre-training paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Deeper model in the Deeper Task-Specificity Improves Joint Entity and Relation Extraction paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Biaffine attention model in the End-to-end neural relation extraction using deep biaffine attention paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Relation-Metric with AT model in the Neural Metric Learning for Fast End-to-End Relation Extraction paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the multi-head model in the Joint entity recognition and relation extraction as a multi-head selection problem paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the multi-head + AT model in the Adversarial training for multi-context joint entity and relation extraction paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the ASP+T0-3B model in the Autoregressive Structured Prediction with Language Models paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the TablERT model in the Named Entity Recognition and Relation Extraction using Enhanced Table Filling by Contextualized Representations paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the TANL model in the Structured Prediction as Translation between Augmented Natural Languages paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the TriMF model in the A Trigger-Sense Memory Flow Framework for Joint Entity and Relation Extraction paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Multi-turn QA model in the Entity-Relation Extraction as Multi-Turn Question Answering paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Global model in the End-to-End Neural Relation Extraction with Global Optimization paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the Table Representation model in the Modeling Joint Entity and Relation Extraction with Table Representation paper on the CoNLL04 dataset? | RE+ Macro F1 , RE+ Micro F1, NER Macro F1, NER Micro F1, RE+ Macro F1 |
What metrics were used to measure the RoCORE model in the A Relation-Oriented Clustering Method for Open Relation Extraction paper on the FewRel dataset? | F1, Precision, Recall |
What metrics were used to measure the ERNIE model in the ERNIE: Enhanced Language Representation with Informative Entities paper on the FewRel dataset? | F1, Precision, Recall |
What metrics were used to measure the Spark NLP model in the Deeper Clinical Document Understanding Using Relation Extraction paper on the 2010 i2b2/VA dataset? | Macro F1 |
What metrics were used to measure the Fellini model in the TräumerAI: Dreaming Music with StyleGAN paper on the TimeTravel dataset? | 0..5sec |
What metrics were used to measure the CLMR model in the Contrastive Learning of Musical Representations paper on the MagnaTagATune dataset? | PR-AUC, ROC AUC |
What metrics were used to measure the CLMR model in the Contrastive Learning of Musical Representations paper on the Million Song Dataset dataset? | PR-AUC, ROC-AUC |
What metrics were used to measure the CLMR (ours) model in the Contrastive Learning of Musical Representations paper on the Million Song Dataset dataset? | PR-AUC, ROC-AUC |
What metrics were used to measure the CRIM model in the CRIM at SemEval-2018 Task 9: A Hybrid Approach to Hypernym Discovery paper on the General dataset? | MAP, MRR, P@5 |
What metrics were used to measure the vTE model in the Supervised Distributional Hypernym Discovery via Domain Adaptation paper on the General dataset? | MAP, MRR, P@5 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.