prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the PaLM 2 model in the PaLM 2 Technical Report paper on the FRMT (Chinese - Mainland) dataset? | BLEURT |
What metrics were used to measure the Google Translate model in the PaLM 2 Technical Report paper on the FRMT (Chinese - Mainland) dataset? | BLEURT |
What metrics were used to measure the PaLM model in the PaLM 2 Technical Report paper on the FRMT (Chinese - Mainland) dataset? | BLEURT |
What metrics were used to measure the fast-noisy-channel-modeling model in the Language Models not just for Pre-training: Fast Online Neural Noisy Channel Modeling paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FLAN 137B zero-shot model in the Finetuned Language Models Are Zero-Shot Learners paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the MLM pretraining model in the Cross-lingual Language Model Pretraining paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the Attentional encoder-decoder + BPE model in the Edinburgh Neural Machine Translation Systems for WMT 16 paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the Levenshtein Transformer (distillation) model in the Levenshtein Transformer paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the CMLM+LAT+4 iterations model in the Incorporating a Local Translation Mechanism into Non-autoregressive Translation paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the Adaptively Sparse Transformer (1.5-entmax) model in the Adaptively Sparse Transformers paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the HeadMask (Impt-18) model in the Alleviating the Inequality of Attention Heads for Neural Machine Translation paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FlowSeq-large (NPD n = 30) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the Adaptively Sparse Transformer (alpha-entmax) model in the Adaptively Sparse Transformers paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the HeadMask (Random-18) model in the Alleviating the Inequality of Attention Heads for Neural Machine Translation paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FlowSeq-large (NPD n = 15) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FlowSeq-large (IWD n = 15) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the NAT +FT + NPD model in the Non-Autoregressive Neural Machine Translation paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the CMLM+LAT+1 iterations model in the Incorporating a Local Translation Mechanism into Non-autoregressive Translation paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FlowSeq-large model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the Denoising autoencoders (non-autoregressive) model in the Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the FlowSeq-base model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the BART (TextBox 2.0) model in the TextBox 2.0: A Text Generation Library with Pre-trained Language Models paper on the WMT2016 Romanian-English dataset? | BLEU score, BLEU-4 |
What metrics were used to measure the OmniNetP model in the OmniNet: Omnidirectional Representations from Transformers paper on the WMT2017 English-French dataset? | BLEU |
What metrics were used to measure the Transformer trained on highly filtered data model in the Impact of Corpora Quality on Neural Machine Translation paper on the WMT 2017 English-Latvian dataset? | BLEU |
What metrics were used to measure the Baseline (en->sn) model in the Itihasa: A large-scale corpus for Sanskrit to English translation paper on the Itihasa dataset? | SacreBLEU |
What metrics were used to measure the Baseline (sn->en) model in the Itihasa: A large-scale corpus for Sanskrit to English translation paper on the Itihasa dataset? | SacreBLEU |
What metrics were used to measure the Transformer Cycle (Rev) model in the Lessons on Parameter Sharing across Layers in Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Noisy back-translation model in the Understanding Back-Translation at Scale paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer+Rep(Uni) model in the Rethinking Perturbations in Encoder-Decoders for Fast Training paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the T5-11B model in the Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the BiBERT model in the BERT, mBERT, or BiBERT? A Study on Contextualized Embeddings for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer + R-Drop model in the R-Drop: Regularized Dropout for Neural Networks paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Bi-SimCut model in the Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the BERT-fused NMT model in the Incorporating BERT into Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Data Diversification - Transformer model in the Data Diversification: A Simple Strategy For Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the SimCut model in the Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Mask Attention Network (big) model in the Mask Attention Networks: Rethinking and Strengthen Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer (ADMIN init) model in the Very Deep Transformers for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the PowerNorm (Transformer) model in the PowerNorm: Rethinking Batch Normalization in Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Depth Growing model in the Depth Growing for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the MUSE(Parallel Multi-scale Attention) model in the MUSE: Parallel Multi-Scale Attention for Sequence to Sequence Learning paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Evolved Transformer model in the The Evolved Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the OmniNetP model in the OmniNet: Omnidirectional Representations from Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the DynamicConv model in the Pay Less Attention with Lightweight and Dynamic Convolutions paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Local Joint Self-attention model in the Joint Source-Target Self Attention with Locality Constraints paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the TaLK Convolutions model in the Time-aware Large Kernel Convolutions paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Big + MoS model in the Fast and Simple Mixture of Softmaxes with BPE and Hybrid-LightRNN for Language Generation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the AdvAug (aut+adv) model in the AdvAug: Robust Adversarial Augmentation for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the PartialFormer model in the PartialFormer: Modeling Part Instead of Whole paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Big + adversarial MLE model in the Improving Neural Language Modeling via Adversarial Training paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Evolved Transformer Big model in the The Evolved Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Big model in the Scaling Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Subformer-xlarge model in the Subformer: A Parameter Reduced Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the SB-NMT model in the Synchronous Bidirectional Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer (big) + Relative Position Representations model in the Self-Attention with Relative Position Representations paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the FLOATER-large model in the Learning to Encode Position for Transformer with Continuous Dynamical Model paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Local Transformer model in the Modeling Localness for Self-Attention Networks paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Big with FRAGE model in the FRAGE: Frequency-Agnostic Word Representation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Mask Attention Network (base) model in the Mask Attention Networks: Rethinking and Strengthen Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Mega model in the Mega: Moving Average Equipped Gated Attention paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the adequacy-oriented NMT model in the Neural Machine Translation with Adequacy-Oriented Learning paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the LightConv model in the Pay Less Attention with Lightweight and Dynamic Convolutions paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Weighted Transformer (large) model in the Weighted Transformer Network for Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the universal transformer base model in the Universal Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the KERMIT model in the KERMIT: Generative Insertion-Based Modeling for Sequences paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the T2R + Pretrain model in the Finetuning Pretrained Transformers into RNNs paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the AdvAug (aut) model in the AdvAug: Robust Adversarial Augmentation for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the RNMT+ model in the The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Synthesizer (Random + Vanilla) model in the Synthesizer: Rethinking Self-Attention in Transformer Models paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Hardware Aware Transformer model in the HAT: Hardware-Aware Transformers for Efficient Natural Language Processing paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Big model in the Attention Is All You Need paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer + SRU model in the Simple Recurrent Units for Highly Parallelizable Recurrence paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Evolved Transformer Base model in the The Evolved Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Rfa-Gate-arccos model in the Random Feature Attention paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer-DRILL Base model in the Deep Residual Output Layers for Neural Language Generation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the AdvAug (mixup) model in the AdvAug: Robust Adversarial Augmentation for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the CMLM+LAT+4 iterations model in the Incorporating a Local Translation Mechanism into Non-autoregressive Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Transformer Base model in the Attention Is All You Need paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Levenshtein Transformer (distillation) model in the Levenshtein Transformer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Adaptively Sparse Transformer (alpha-entmax) model in the Adaptively Sparse Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the ResMLP-12 model in the ResMLP: Feedforward networks for image classification with data-efficient training paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the CNAT model in the Non-Autoregressive Translation by Learning Target Categorical Codes paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Lite Transformer model in the Lite Transformer with Long-Short Range Attention paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the ConvS2S (ensemble) model in the Convolutional Sequence to Sequence Learning paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the ResMLP-6 model in the ResMLP: Feedforward networks for image classification with data-efficient training paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Average Attention Network model in the Accelerating Neural Transformer via an Average Attention Network paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the GNMT+RL model in the Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the SliceNet model in the Depthwise Separable Convolutions for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Average Attention Network (w/o FFN) model in the Accelerating Neural Transformer via an Average Attention Network paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the MoE model in the Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Average Attention Network (w/o gate) model in the Accelerating Neural Transformer via an Average Attention Network paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the Adaptively Sparse Transformer (1.5-entmax) model in the Adaptively Sparse Transformers paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the DenseNMT model in the Dense Information Flow for Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the GLAT model in the Glancing Transformer for Non-Autoregressive Neural Machine Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the CMLM+LAT+1 iterations model in the Incorporating a Local Translation Mechanism into Non-autoregressive Translation paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the ConvS2S model in the Convolutional Sequence to Sequence Learning paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the ByteNet model in the Neural Machine Translation in Linear Time paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the FlowSeq-large (NPD n = 30) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the FlowSeq-large (NPD n = 15) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
What metrics were used to measure the FlowSeq-large (IWD n = 15) model in the FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow paper on the WMT2014 English-German dataset? | BLEU score, SacreBLEU, Number of Params, Hardware Burden, Operations per network pass |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.