prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the AUCO ResNet model in the AUCO ResNet: an end-to-end network for Covid-19 pre-screening from cough and breath paper on the Covid-19 Cough Cambridge dataset? | AUC |
What metrics were used to measure the Sanskar et al. model in the A Novel Approach for detecting Normal, COVID-19 and Pneumonia patient using only binary classifications from chest CT-Scans paper on the COVIDx dataset? | 3-class test accuracy, AUC |
What metrics were used to measure the COVID-WideNet model in the COVID-WideNet—A capsule network for COVID-19 detection paper on the COVIDx dataset? | 3-class test accuracy, AUC |
What metrics were used to measure the CovidExpert model in the CovidExpert: A Triplet Siamese Neural Network framework for the detection of COVID-19 paper on the Large COVID-19 CT scan slice dataset dataset? | AUC-ROC, Accuracy, Macro F1, Macro Precision, Macro Recall, Micro Precision, Specificity |
What metrics were used to measure the Bhowal et al. model in the Choquet Integral and Coalition Game-based Ensemble of Deep Learning Models for COVID-19 Screening from Chest X-ray Images paper on the Novel COVID-19 Chestxray Repository dataset? | ACCURACY |
What metrics were used to measure the SS-CXR model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the DenseNet-169 model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the EfficientNet-B2 model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the Inception Resnet V2 model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the Inception ResNet model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the DenseNet-121 model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the ViT-S model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the COVIDx CXR-3 dataset? | Per-Class Accuracy |
What metrics were used to measure the CovidCTNet model in the CovidCTNet: An Open-Source Deep Learning Approach to Identify Covid-19 Using CT Image paper on the dataset? | 10 fold Cross validation, Per-Class Accuracy |
What metrics were used to measure the Thoracic Radiologist model in the SPCXR: Self-supervised Pretraining using Chest X-rays Towards a Domain Specific Foundation Model paper on the dataset? | 10 fold Cross validation, Per-Class Accuracy |
What metrics were used to measure the CBHG model model in the Effective Deep Learning Models for Automatic Diacritization of Arabic Text paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the Shakkelha model in the Neural Arabic Text Diacritization: State of the Art Results and a Novel Approach for Machine Translation paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the D3 (D2 + decoder) model in the Deep Diacritization: Efficient Hierarchical Recurrence for Improved Arabic Diacritization paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the D2 model in the Deep Diacritization: Efficient Hierarchical Recurrence for Improved Arabic Diacritization paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the MC model in the Multi-components System for Automatic Arabic Diacritization paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the Shakkala model in the Arabic Text Diacritization Using Deep Neural Networks paper on the Tashkeela dataset? | Diacritic Error Rate, Word Error Rate (WER) |
What metrics were used to measure the SPSN (k=2, VGG) model in the Parallel Spiking Neurons with High Efficiency and Ability to Learn Long-term Dependencies paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the tdBN + NDA (VGG11) model in the Neuromorphic Data Augmentation for Training Spiking Neural Networks paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the OTTT model in the Online Training Through Time for Spiking Neural Networks paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the Dspike (ResNet-18) model in the Differentiable Spike: Rethinking Gradient-Descent for Training Spiking Neural Networks paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the IM-Loss (ResNet-19) model in the IM-Loss: Information Maximization Loss for Spiking Neural Networks paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the STS-ResNet model in the Convolutional Spiking Neural Networks for Spatio-Temporal Feature Extraction paper on the CIFAR10-DVS dataset? | Accuracy |
What metrics were used to measure the mm model in the SwapNet: Garment Transfer in Single View Images paper on the FashionIQ dataset? | 10 fold Cross validation |
What metrics were used to measure the MG-VTON model in the Towards Multi-pose Guided Virtual Try-on Network paper on the Deep-Fashion dataset? | IS, SSIM |
What metrics were used to measure the Poly-GAN model in the Poly-GAN: Multi-Conditioned GAN for Fashion Synthesis paper on the Deep-Fashion dataset? | IS, SSIM |
What metrics were used to measure the HR-VITON model in the High-Resolution Virtual Try-On with Misalignment and Occlusion-Handled Conditions paper on the VITON-HD dataset? | FID |
What metrics were used to measure the VITON-HD model in the VITON-HD: High-Resolution Virtual Try-On via Misalignment-Aware Normalization paper on the VITON-HD dataset? | FID |
What metrics were used to measure the Flow-Style-VTON model in the Style-Based Global Appearance Flow for Virtual Try-On paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the PF-AFN model in the Parser-Free Virtual Try-on via Distilling Appearance Flows paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the SDAFN model in the Single Stage Virtual Try-on via Deformable Attention Flows paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the RT-VTON model in the Full-Range Virtual Try-On With Recurrent Tri-Level Transform paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the PSAD model in the Dress Code: High-Resolution Multi-Category Virtual Try-On paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the C-VTON model in the C-VTON: Context-Driven Image-Based Virtual Try-On Network paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the CloTH-VTON+ model in the CloTH-VTON+: Clothing Three-dimensional reconstruction for Hybrid image-based Virtual Try-ON paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the ACGPN model in the Towards Photo-Realistic Virtual Try-On by Adaptively Generating$\leftrightarrow$Preserving Image Content paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the ClothFlow model in the ClothFlow: A Flow-Based Model for Clothed Person Generation paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the CP-VTON+ model in the CP-VTON+: Clothing Shape and Texture Preserving Image-Based Virtual Try-On paper on the VITON dataset? | FID, SSIM, LPIPS, IS, KID, PSNR |
What metrics were used to measure the PF-AFN model in the Parser-Free Virtual Try-on via Distilling Appearance Flows paper on the MPV dataset? | FID, SWD |
What metrics were used to measure the DGP model in the Weakly Supervised High-Fidelity Clothing Model Generation paper on the MPV dataset? | FID, SWD |
What metrics were used to measure the Tran-BERT-MS-ML-R model in the On the Use of BERT for Automated Essay Scoring: Joint Learning of Multi-Scale Essay Representation paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the Considering-Content-XLNet model in the Countering the Influence of Essay Length in Neural Essay Scoring paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the HISK+BOSWE model in the Automated essay scoring with string kernels and word embeddings paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the SkipFlow model in the SkipFlow: Incorporating Neural Coherence Features for End-to-End Automatic Text Scoring paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the MHMLW model in the Many Hands Make Light Work: Using Essay Traits to Automatically Score Essays paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the AF model in the Automatic Features for Essay Scoring -- An Empirical Study paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the FDA model in the Flexible Domain Adaptation for Automated Essay Scoring Using Correlated Linear Regression paper on the ASAP dataset? | Quadratic Weighted Kappa |
What metrics were used to measure the DFSud model in the Code Prediction by Feeding Trees to Transformers paper on the Py150 dataset? | MRR |
What metrics were used to measure the GraphCodeBERT model in the GraphCodeBERT: Pre-training Code Representations with Data Flow paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the CodeBERT model in the CodeBERT: A Pre-Trained Model for Programming and Natural Languages paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the PolyGot model in the Multilingual training for Software Engineering paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the GraphPolyGot model in the Multilingual training for Software Engineering paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the RoBERTa model in the RoBERTa: A Robustly Optimized BERT Pretraining Approach paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the CodeBERTa model in the paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the BERT model in the BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding paper on the ManyTypes4TypeScript dataset? | Average Precision, Average Recall, Average F1, Average Accuracy |
What metrics were used to measure the DFSud model in the Code Prediction by Feeding Trees to Transformers paper on the Py150 dataset? | MRR |
What metrics were used to measure the ContraCode model in the Contrastive Code Representation Learning paper on the DeepTyper dataset? | Accuracy@5 |
What metrics were used to measure the Sim2-CNN model in the Exploiting Redundancy: Separable Group Convolutional Networks on Lie Groups paper on the Rotated MNIST dataset? | Test error |
What metrics were used to measure the E2FCNN (D16 |5 C16) model in the General E(2)-Equivariant Steerable CNNs paper on the Rotated MNIST dataset? | Test error |
What metrics were used to measure the Steerable Filter CNN model in the Learning Steerable Filters for Rotation Equivariant CNNs paper on the Rotated MNIST dataset? | Test error |
What metrics were used to measure the FarSeg++@MiT-B2 model in the FarSeg++: Foreground-Aware Relation Network for Geospatial Object Segmentation in High Spatial Resolution Remote Sensing Imagery paper on the UV6K dataset? | IoU (%) |
What metrics were used to measure the FarSeg++@Swin-T model in the FarSeg++: Foreground-Aware Relation Network for Geospatial Object Segmentation in High Spatial Resolution Remote Sensing Imagery paper on the UV6K dataset? | IoU (%) |
What metrics were used to measure the FarSeg++@ResNet-50 model in the FarSeg++: Foreground-Aware Relation Network for Geospatial Object Segmentation in High Spatial Resolution Remote Sensing Imagery paper on the UV6K dataset? | IoU (%) |
What metrics were used to measure the ResNet-50 model in the FireRisk: A Remote Sensing Dataset for Fire Risk Assessment with Benchmarks Using Supervised and Self-supervised Learning paper on the FireRisk dataset? | Accuracy (%) |
What metrics were used to measure the ViT-B/16 model in the FireRisk: A Remote Sensing Dataset for Fire Risk Assessment with Benchmarks Using Supervised and Self-supervised Learning paper on the FireRisk dataset? | Accuracy (%) |
What metrics were used to measure the DINO (ViT-B/16) model in the FireRisk: A Remote Sensing Dataset for Fire Risk Assessment with Benchmarks Using Supervised and Self-supervised Learning paper on the FireRisk dataset? | Accuracy (%) |
What metrics were used to measure the MAE (ViT-B/16) model in the FireRisk: A Remote Sensing Dataset for Fire Risk Assessment with Benchmarks Using Supervised and Self-supervised Learning paper on the FireRisk dataset? | Accuracy (%) |
What metrics were used to measure the GPT-2 (small) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the XLNet (large) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the GPT-2 (medium) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the BERT (base) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the GPT-2 (large) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the BERT (large) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the RoBERTa (base) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the GAL 120B model in the Galactica: A Large Language Model for Science paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the XLNet (base) model in the StereoSet: Measuring stereotypical bias in pretrained language models paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the GPT-3 (text-davinci-002) model in the Galactica: A Large Language Model for Science paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the OPT 175B model in the Galactica: A Large Language Model for Science paper on the StereoSet dataset? | ICAT Score, LMS, SS |
What metrics were used to measure the RoBERTa+ALBERT model in the Towards Detection of Subjective Bias using Contextualized Word Embeddings paper on the Wiki Neutrality Corpus dataset? | F1 |
What metrics were used to measure the gpt-4-temp-0 model in the BAD: BiAs Detection for Large Language Models in the context of candidate screening paper on the ICAT LLM bias dataset? | ICAT Score |
What metrics were used to measure the RandomForest_default_hyperparameters model in the Uncovering bias in the PlantVillage dataset paper on the PlantVillage_8px dataset? | Accuracy (%) |
What metrics were used to measure the Relational Reasoning model in the Self-Supervised Relational Reasoning for Representation Learning paper on the STL-10 dataset? | Accuracy (%), Accuracy |
What metrics were used to measure the MV-MR model in the MV-MR: multi-views and multi-representations for self-supervised learning and knowledge distillation paper on the STL-10 dataset? | Accuracy (%), Accuracy |
What metrics were used to measure the Hybrid BYOL-S/CvT model in the BYOL-S: Learning Self-supervised Speech Representations by Bootstrapping paper on the CREMA-D dataset? | Accuracy |
What metrics were used to measure the Pretraining: None model in the DABS: A Domain-Agnostic Benchmark for Self-Supervised Learning paper on the DABS dataset? | Images & Text, Med. Imaging, Natural Images, Sensors, Speech, Text |
What metrics were used to measure the Pretraining: ShED model in the DABS: A Domain-Agnostic Benchmark for Self-Supervised Learning paper on the DABS dataset? | Images & Text, Med. Imaging, Natural Images, Sensors, Speech, Text |
What metrics were used to measure the Pretraining: e-Mix model in the DABS: A Domain-Agnostic Benchmark for Self-Supervised Learning paper on the DABS dataset? | Images & Text, Med. Imaging, Natural Images, Sensors, Speech, Text |
What metrics were used to measure the Hate-CLIPper model in the Hate-CLIPper: Multimodal Hateful Meme Classification based on Cross-modal Interaction of CLIP Features paper on the Tamil Memes dataset? | Micro-F1 |
What metrics were used to measure the UVCE-IIITT model in the UVCE-IIITT@DravidianLangTech-EACL2021: Tamil Troll Meme Classification: You need to Pay more Attention paper on the Tamil Memes dataset? | Micro-F1 |
What metrics were used to measure the RGCL - HateCLIPper model in the Improving hateful memes detection via learning hatefulness-aware embedding space through retrieval-guided contrastive learning paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the Hate-CLIPper - Align model in the Hate-CLIPper: Multimodal Hateful Meme Classification based on Cross-modal Interaction of CLIP Features paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the Vilio model in the Vilio: State-of-the-art Visio-Linguistic Models applied to Hateful Memes paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the HateDetectron27 model in the Detecting Hate Speech in Memes Using Multimodal Deep Learning Approaches: Prize-winning solution to Hateful Memes Challenge paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the SEER (RegNet10B) model in the Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without Supervision paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the Flamingo (few-shot:32) model in the Flamingo: a Visual Language Model for Few-Shot Learning paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the CLIP (zero-shot) model in the Learning Transferable Visual Models From Natural Language Supervision paper on the Hateful Memes dataset? | ROC-AUC |
What metrics were used to measure the LAFF model in the Lightweight Attentional Feature Fusion: A New Baseline for Text-to-Video Retrieval paper on the TRECVID-AVS17 (IACC.3) dataset? | infAP |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.