prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the R-BERT model in the Enriching Pre-trained Language Model with Entity Information for Relation Classification paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the C-GCN + PA-LSTM model in the Graph Convolution over Pruned Dependency Trees Improves Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the C-AGGCN model in the Attention Guided Graph Convolutional Networks for Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the ERNIE model in the ERNIE: Enhanced Language Representation with Informative Entities paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the BERT-LSTM-base model in the Simple BERT Models for Relation Extraction and Semantic Role Labeling paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the SA-LSTM+D model in the Beyond Word Attention: Using Segment Attention in Neural Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the TRE model in the Improving Relation Extraction by Pre-trained Language Representations paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the Alt et al. (2019) model in the Improving Relation Extraction by Pre-trained Language Representations paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the GCN + PA-LSTM model in the Graph Convolution over Pruned Dependency Trees Improves Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the C-SGC model in the Simplifying Graph Convolutional Networks paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the C-GCN model in the Graph Convolution over Pruned Dependency Trees Improves Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the PA-LSTM model in the Position-aware Attention and Supervised Data Improve Slot Filling paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the AGGCN model in the Attention Guided Graph Convolutional Networks for Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the GCN model in the Graph Convolution over Pruned Dependency Trees Improves Relation Extraction paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the LLM-QA4R (Zero-shot) model in the Aligning Instruction Tasks Unlocks Large Language Models as Zero-Shot Relation Extractors paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the LUKE model in the LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention paper on the TACRED dataset? | F1, F1 (10% Few-Shot), F1 (5% Few-Shot), F1 (1% Few-Shot), F1 (Zero-Shot) |
What metrics were used to measure the RELA model in the Sequence Generation with Label Augmentation for Relation Extraction paper on the sciERC-sent dataset? | F1 |
What metrics were used to measure the Stacked_LinkedBERT model in the Exploiting Unary Relations with Stacked Learning for Relation Extraction paper on the LPSC-contains dataset? | F1 (micro) |
What metrics were used to measure the DREEAM model in the DREEAM: Guiding Attention with Evidence for Improving Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the KD-Rb-l model in the Document-Level Relation Extraction with Adaptive Focal Loss and Knowledge Distillation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SSAN-RoBERTa-large+Adaptation model in the Entity Structure Within and Throughout: Modeling Mention Dependencies for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SAIS-RoBERTa-large model in the SAIS: Supervising and Augmenting Intermediate Steps for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the Eider-RoBERTa-large model in the Eider: Empowering Document-level Relation Extraction with Efficient Evidence Extraction and Inference-stage Fusion paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DocuNet-RoBERTa-large model in the Document-level Relation Extraction as Semantic Segmentation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CGM2IR-RoBERTalarge model in the Document-level Relation Extraction with Context Guided Mention Integration and Inter-pair Reasoning paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SETE-Roberta-large model in the Document-Level Relation Extraction with Structure Enhanced Transformer Encoder paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the ATLOP-RoBERTa-large model in the Document-Level Relation Extraction with Adaptive Thresholding and Localized Context Pooling paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DRE-MIR-BERTbase model in the A Masked Image Reconstruction Network for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SAIS-BERT-base model in the SAIS: Supervising and Augmenting Intermediate Steps for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GAIN-BERT-large model in the Double Graph Based Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the Dense-CCNet-BERTbase model in the A Densely Connected Criss-Cross Attention Network for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the E2GRE-RoBERTa-large model in the Entity and Evidence Guided Relation Extraction for DocRED paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the Eider-BERT-base model in the Eider: Empowering Document-level Relation Extraction with Efficient Evidence Extraction and Inference-stage Fusion paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SagDRE model in the SagDRE: Sequence-Aware Graph-Based Document-Level Relation Extraction with Adaptive Margin Loss paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GAIN+SIEF model in the Document-Level Relation Extraction with Sentences Importance Estimation and Focusing paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CGM2IR-BERTbase model in the Document-level Relation Extraction with Context Guided Mention Integration and Inter-pair Reasoning paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SIRE-BERT-base model in the SIRE: Separate Intra- and Inter-sentential Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the MRN+BERT model in the MRN: A Locally and Globally Mention-Based Reasoning Network for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the ATLOP + LogiRE model in the Learning Logic Rules for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SSAN-RoBERTa-large model in the Entity Structure Within and Throughout: Modeling Mention Dependencies for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the ATLOP + KIRE model in the Enhancing Document-level Relation Extraction by Entity Knowledge Injection paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DRN-BERT-base model in the Discriminative Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the ERACL model in the Improving Long Tailed Document-Level Relation Extraction via Easy Relation Augmentation and Contrastive Learning paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the ATLOP-BERT-base model in the Document-Level Relation Extraction with Adaptive Thresholding and Localized Context Pooling paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GAIN-BERT model in the Double Graph Based Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CorefDRE- BERT model in the CorefDRE: Document-level Relation Extraction with coreference resolution paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GAIN + LogiRE model in the Learning Logic Rules for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the JEREX-BERT-base model in the An End-to-end Model for Entity-level Relation Extraction using Multi-instance Learning paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CorefRoBERTa-large model in the Coreferential Reasoning Learning for Language Representation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the MIUK model in the Multi-view Inference for Relation Extraction with Uncertain Knowledge paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SSAN-RoBERTa-base model in the Entity Structure Within and Throughout: Modeling Mention Dependencies for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CFER-BERT-base model in the Coarse-to-Fine Entity Representations for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the HeterGSAN+Reconstruction+BERT-base model in the Document-Level Relation Extraction with Reconstruction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the FineCL model in the Fine-grained Contrastive Learning for Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SSAN+RSMAN model in the Relation-Specific Attentions over Entity Mentions for Enhanced Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the LSR+BERT-base model in the Reasoning with Latent Structure Refinement for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GLRE-XLNet-Large model in the Global-to-Local Neural Networks for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CorefBERT-large model in the Coreferential Reasoning Learning for Language Representation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the E2GRE-BERT-base model in the Entity and Evidence Guided Relation Extraction for DocRED paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the EncAttAgg model in the Improving Document-level Relation Extraction via Contextualizing Mention Representations and Weighting Mention Pairs paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the MRN model in the MRN: A Locally and Globally Mention-Based Reasoning Network for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SSAN-BERT-base model in the Entity Structure Within and Throughout: Modeling Mention Dependencies for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DUAL+BERT-base model in the Dual Supervision Framework for Relation Extraction with Distant Supervision and Human Annotation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CorefBERT-base model in the Coreferential Reasoning Learning for Language Representation paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DRN-GloVe model in the Discriminative Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the Paths+BiLSTM-GloVe model in the Three Sentences Are All You Need: Local Path Enhanced Document Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the SIRE-GloVe model in the SIRE: Separate Intra- and Inter-sentential Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the CFER-GloVe model in the Coarse-to-Fine Entity Representations for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the HIN-BERT-base model in the HIN: Hierarchical Inference Network for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the HeterGSAN+Reconstruction model in the Document-Level Relation Extraction with Reconstruction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GAIN-GloVe model in the Double Graph Based Reasoning for Document-level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the LSR+GloVe model in the Reasoning with Latent Structure Refinement for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the Two-Step+BERT-base model in the Fine-tune Bert for DocRED with Two-step Process paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the HIN-GloVe model in the HIN: Hierarchical Inference Network for Document-Level Relation Extraction paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the BERT-base model in the Fine-tune Bert for DocRED with Two-step Process paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the GREG model in the GREG: A Global Level Relation Extraction with Knowledge Graph Embedding paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the BiLSTM model in the DocRED: A Large-Scale Document-Level Relation Extraction Dataset paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DocRED-Context-Aware model in the DocRED: A Large-Scale Document-Level Relation Extraction Dataset paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the BiLSTM model in the DocRED: A Large-Scale Document-Level Relation Extraction Dataset paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the DocRED-CNN model in the DocRED: A Large-Scale Document-Level Relation Extraction Dataset paper on the DocRED dataset? | F1, Ign F1 |
What metrics were used to measure the PubMedBERT model in the BioRED: A Rich Biomedical Relation Extraction Dataset paper on the BioRED dataset? | F1 |
What metrics were used to measure the BERT-GT model in the BERT-GT: Cross-sentence n-ary relation extraction with BERT and Graph Transformer paper on the BioRED dataset? | F1 |
What metrics were used to measure the iDepNN model in the Neural Relation Extraction Within and Across Sentence Boundaries paper on the MUC6 dataset? | Average F1 |
What metrics were used to measure the aimped model in the paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the PFN (ALBERT XXL, average aggregation) model in the An Information Extraction Study: Take In Mind the Tokenization! paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the Deeper model in the Deeper Task-Specificity Improves Joint Entity and Relation Extraction paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the PFN (ALBERT XXL, no aggregation) model in the A Partition Filter Network for Joint Entity and Relation Extraction paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the SpERT.PL (without overlap and BioBERT) model in the Joint Entity and Relation Extraction from Scientific Documents: Role of Linguistic Information and Entity Types paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the REBEL (including overlapping entities) model in the REBEL: Relation Extraction By End-to-end Language generation paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the SpERT.PL (with overlap and BioBERT) model in the Joint Entity and Relation Extraction from Scientific Documents: Role of Linguistic Information and Entity Types paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the CMAN model in the Modeling Dense Cross-Modal Interactions for Joint Entity-Relation Extraction paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the Table-Sequence model in the Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the CLDR + CLNER model in the Imposing Relation Structure in Language-Model Embeddings Using Contrastive Learning paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the SpERT (without overlap) model in the Span-based Joint Entity and Relation Extraction with Transformer Pre-training paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the SpERT (with overlap) model in the Span-based Joint Entity and Relation Extraction with Transformer Pre-training paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the Relation-Metric model in the Neural Metric Learning for Fast End-to-End Relation Extraction paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the multi-head + AT model in the Adversarial training for multi-context joint entity and relation extraction paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the multi-head model in the Joint entity recognition and relation extraction as a multi-head selection problem paper on the Adverse Drug Events (ADE) Corpus dataset? | RE+ Macro F1, RE Macro F1, NER Macro F1 |
What metrics were used to measure the UniRel model in the UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction paper on the WebNLG dataset? | F1, NER Micro F1 |
What metrics were used to measure the PFN model in the A Partition Filter Network for Joint Entity and Relation Extraction paper on the WebNLG dataset? | F1, NER Micro F1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.