prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the BART MNLI zero-shot model in the RAFT: A Real-World Few-Shot Text Classification Benchmark paper on the RAFT dataset? | Avg, ADE, B77, NIS, OSE, Over, SOT, SRI, TAI, ToS, TEH, TC, Over |
What metrics were used to measure the Plurality-class model in the RAFT: A Real-World Few-Shot Text Classification Benchmark paper on the RAFT dataset? | Avg, ADE, B77, NIS, OSE, Over, SOT, SRI, TAI, ToS, TEH, TC, Over |
What metrics were used to measure the GPT-3 zero-shot model in the RAFT: A Real-World Few-Shot Text Classification Benchmark paper on the RAFT dataset? | Avg, ADE, B77, NIS, OSE, Over, SOT, SRI, TAI, ToS, TEH, TC, Over |
What metrics were used to measure the Induction Networks model in the Induction Networks for Few-Shot Text Classification paper on the ODIC 5-way (5-shot) dataset? | Accuracy |
What metrics were used to measure the Induction Networks model in the Induction Networks for Few-Shot Text Classification paper on the ODIC 10-way (10-shot) dataset? | Accuracy |
What metrics were used to measure the CFEAR-3-s4 model in the CFEAR Radarodometry - Conservative Filtering for Efficient and Accurate Radar Odometry paper on the Oxford Radar RobotCar Dataset dataset? | translation error [%] |
What metrics were used to measure the Achelous-FV-RDF-S2 model in the Achelous: A Fast Unified Water-surface Panoptic Perception Framework based on Fusion of Monocular Camera and 4D mmWave Radar paper on the WaterScenes dataset? | mIoU |
What metrics were used to measure the Modular U-Net (2D) model in the A modular U-Net for automated segmentation of X-ray tomography images in composite materials paper on the GF-PA66 3D XCT dataset? | Jaccard (Mean) |
What metrics were used to measure the LSTM model in the (0,2) hybrid models paper on the Time Series Prediction Benchmarks dataset? | 1:3 Accuracy |
What metrics were used to measure the 7x7 dense-morphological + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 5x5 dense-morphological + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 7x7 linear + ibf+sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 7x7 linear-ratio + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 7x7 polynomial + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 3x3 universal-function + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 5x5 polynomial + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 5x5 universal-function + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 7x7 universal-function + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 3x3 dense-morphological + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 1x1 universal-function + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 5x5 linear-ratio + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 1x1 polynomial + ibf model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the 1x1 dense-morphological + ibf + sprb model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Modified Triangular Vegetation Index 1 model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Modified Chlorophyll Absorption In Reflectance Index 1 model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Enhanced Vegetation Index 2 model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Soil Adjusted Vegetation Index model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Soil And Atmospherically Resistant VI 3 model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Enhanced Vegetation Index 3 model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Global Environment Monitoring Index model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the Adjusted Transformed Soil Adjusted VI model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the NDVI model in the DeepIndices: Remote Sensing Indices Based on Approximation of Functions through Deep-Learning, Application to Uncalibrated Vegetation Images paper on the Deep Indices dataset? | mIoU |
What metrics were used to measure the BDANet model in the BDANet: Multiscale Convolutional Neural Network with Cross-directional Attention for Building Damage Assessment from Satellite Images paper on the xBD dataset? | Weighted Average F1-score |
What metrics were used to measure the Double branch U-Net model in the Cross-directional Feature Fusion Network for Building Damage Assessment from Satellite Imagery paper on the xBD dataset? | Weighted Average F1-score |
What metrics were used to measure the Double branch U-Net model in the Building Disaster Damage Assessment in Satellite Imagery with Multi-Temporal Fusion paper on the xBD dataset? | Weighted Average F1-score |
What metrics were used to measure the Baseline Model model in the xBD: A Dataset for Assessing Building Damage from Satellite Imagery paper on the xBD dataset? | Weighted Average F1-score |
What metrics were used to measure the UNet model in the Methods for the frugal labeler: Multi-class semantic segmentation on heterogeneous labels paper on the Extended heartSeg dataset? | Average IOU |
What metrics were used to measure the TempGen model in the Document-level Entity-based Extraction as Template Generation paper on the MUC-4 dataset? | Avg. F1 |
What metrics were used to measure the Ning et al. model in the A Structured Learning Approach to Temporal Relation Extraction paper on the TempEval-3 dataset? | Temporal awareness |
What metrics were used to measure the ClearTK model in the ClearTK-TimeML: A minimalist approach to TempEval 2013 paper on the TempEval-3 dataset? | Temporal awareness |
What metrics were used to measure the Catena model in the CATENA: CAusal and TEmporal relation extraction from NAtural language texts paper on the TimeBank dataset? | F1 score |
What metrics were used to measure the CAEVO model in the Dense Event Ordering with a Multi-Pass Architecture paper on the TimeBank dataset? | F1 score |
What metrics were used to measure the ReviewKD++(T: faster rcnn(resnet101), S:faster rcnn(resnet50)) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the COCO 2017 val dataset? | AP@0.5, AP@0.75, mAP |
What metrics were used to measure the ReviewKD++(T: faster rcnn(resnet101), S:faster rcnn(resnet18)) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the COCO 2017 val dataset? | AP@0.5, AP@0.75, mAP |
What metrics were used to measure the ReviewKD++(T: faster rcnn(resnet101), S:faster rcnn(mobilenet-v2)) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the COCO 2017 val dataset? | AP@0.5, AP@0.75, mAP |
What metrics were used to measure the ADLIK-Faster (T: Faster R-CNN vit-base S: Faster R-CNN deit-small) model in the Focal and Global Knowledge Distillation for Detectors paper on the COCO dataset? | box AP, mask AP, mAP |
What metrics were used to measure the ADLIK-Mask (T: Mask R-CNN vit-base S: Mask R-CNN deit-small) model in the Focal and Global Knowledge Distillation for Detectors paper on the COCO dataset? | box AP, mask AP, mAP |
What metrics were used to measure the LSHFM (T: ResNet101 S: ResNet50) model in the Distilling Knowledge by Mimicking Features paper on the COCO dataset? | box AP, mask AP, mAP |
What metrics were used to measure the LSHFM (T: ResNet101 S: MobileNetV2) model in the Distilling Knowledge by Mimicking Features paper on the COCO dataset? | box AP, mask AP, mAP |
What metrics were used to measure the KD++(T: regnety-16GF S:ViT-B) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the DIST (T: Swin-L S: Swin-T) model in the Knowledge Distillation from A Stronger Teacher paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the Closer Look (T:RegNety 160 S:DeIT-S) model in the A closer look at the training dynamics of knowledge distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:resnet-152 S:resnet-101) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ADLIK-MO-P25(T:SeNet154, ResNet152b S:ResNet-50-prune25%) model in the Ensemble Knowledge Distillation for Learning Improved and Efficient Networks paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ADLIK-MO-P375(T:SeNet154, ResNet152b S:ResNet-50-prune37.5) model in the Ensemble Knowledge Distillation for Learning Improved and Efficient Networks paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:resnet-152 S:resnet-50) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the Closer Look (T:RegNety 160 S:DeIT-Ti) model in the A closer look at the training dynamics of knowledge distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ADLIK-MO(T: ResNet101 S: ResNet50) model in the Distilling the Knowledge in a Neural Network paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ADLIK-MO-P50(T:SeNet154, ResNet152b S:ResNet-50-half) model in the Ensemble Knowledge Distillation for Learning Improved and Efficient Networks paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:resnet152 S:resnet34) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ReviewKD++(T:resnet50, S:mobilenet-v1) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:resnet-152 S:resnet18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:renset101 S:resnet18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:resnet50 S:resnet18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the HSAKD (T: ResNet-34 S:ResNet-18) model in the Hierarchical Self-supervised Augmented Knowledge Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ICKD (T: ResNet-34 S:ResNet-18) model in the Exploring Inter-Channel Correlation for Diversity-Preserved Knowledge Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the DIST (T: ResNet-34 S:ResNet-18) model in the Knowledge Distillation from A Stronger Teacher paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T: ResNet-34 S:ResNet-18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the WSL (T: ResNet-34 S:ResNet-18) model in the Rethinking Soft Labels for Knowledge Distillation: A Bias-Variance Tradeoff Perspective paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the CRCD (T: ResNet-34 S:ResNet-18) model in the Complementary Relation Contrastive Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T:ViT-B, S:resnet18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the LSHFM (T: ResNet-34 S:ResNet-18) model in the Distilling Knowledge by Mimicking Features paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the ITRD (T: ResNet-34 S:ResNet-18) model in the Information Theoretic Representation Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the GLD (T: ResNet-34 S:ResNet-18) model in the Distilling Global and Local Logits With Densely Connected Relations paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the SSKD (T: ResNet-34 S:ResNet-18) model in the Knowledge Distillation Meets Self-Supervision paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the Knowledge Review (T: ResNet-34 S:ResNet-18) model in the Distilling Knowledge via Knowledge Review paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the Adaptive (T:ResNet-50 S:ResNet-18) model in the Adaptive Distillation: Aggregating Knowledge from Multiple Paths for Efficient Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD++(T: ViT-S, S:resnet18) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the AFD (T: ResNet-34 S:ResNet-18) model in the Show, Attend and Distill:Knowledge Distillation via Attention-based Feature Matching paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the CRD (T: ResNet-34 S:ResNet-18) model in the Contrastive Representation Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the Overhual (T: ResNet-34 S:ResNet-18) model in the A Comprehensive Overhaul of Feature Distillation paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the AT (T: ResNet-34 S:ResNet-18) model in the Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the KD (T: ResNet-34 S:ResNet-18) model in the Distilling the Knowledge in a Neural Network paper on the ImageNet dataset? | Top-1 accuracy %, model size, CRD training setting |
What metrics were used to measure the LSHFM (T: ResNet101 S: ResNet50) model in the Distilling Knowledge by Mimicking Features paper on the PASCAL VOC dataset? | mAP |
What metrics were used to measure the LSHFM (T: ResNet101 S: MobileNetV2) model in the Distilling Knowledge by Mimicking Features paper on the PASCAL VOC dataset? | mAP |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4 [modified]) model in the Knowledge Distillation with the Reused Teacher Classifier paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the ReviewKD++(T:resnet-32x4, S:shufflenet-v2) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the ReviewKD++(T:resnet-32x4, S:shufflenet-v1) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Information Theoretic Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Knowledge Distillation from A Stronger Teacher paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the DKD++(T:resnet-32x4, S:resnet-8x4) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Wasserstein Contrastive Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the ReviewKD++(T:WRN-40-2, S:WRN-40-1) model in the Improving Knowledge Distillation via Regularizing Feature Norm and Direction paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Distilling Knowledge via Knowledge Review paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Contrastive Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the vgg8 (T:vgg13 S:vgg8) model in the Information Theoretic Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the vgg8 (T:vgg13 S:vgg8) model in the Distilling Knowledge via Knowledge Review paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the vgg8 (T:vgg13 S:vgg8) model in the Wasserstein Contrastive Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the vgg8 (T:vgg13 S:vgg8) model in the Contrastive Representation Distillation paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
What metrics were used to measure the resnet8x4 (T: resnet32x4 S: resnet8x4) model in the Distilling the Knowledge in a Neural Network paper on the CIFAR-100 dataset? | Top-1 Accuracy (%) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.