prompts stringlengths 81 413 | metrics_response stringlengths 0 371 |
|---|---|
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DIQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the The StarCraft Multi-Agent Challenge paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the The StarCraft Multi-Agent Challenge paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the The StarCraft Multi-Agent Challenge paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the Heuristic model in the The StarCraft Multi-Agent Challenge paper on the SMAC MMM2 dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the VDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_4s6z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the VDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 26m_vs_30m dataset? | Average Score, Median Win Rate |
What metrics were used to measure the ACE model in the ACE: Cooperative Multi-agent Q-learning with Bidirectional Action-Dependency paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DIQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the The StarCraft Multi-Agent Challenge paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the The StarCraft Multi-Agent Challenge paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the Heuristic model in the The StarCraft Multi-Agent Challenge paper on the SMAC corridor dataset? | Median Win Rate, Average Score |
What metrics were used to measure the ACE model in the ACE: Cooperative Multi-agent Q-learning with Bidirectional Action-Dependency paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DIQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the The StarCraft Multi-Agent Challenge paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the The StarCraft Multi-Agent Challenge paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the Heuristic model in the The StarCraft Multi-Agent Challenge paper on the SMAC 3s5z_vs_3s6z dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the VDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 6h_vs_9z dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the VDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC corridor_2z_vs_24zg dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the VDN model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the The StarCraft Multi-Agent Challenge paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the QMIX model in the Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DIQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the IQL model in the DFAC Framework: Factorizing the Value Function via Quantile Mixture for Multi-Agent Distributional Q-Learning paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the Heuristic model in the The StarCraft Multi-Agent Challenge paper on the SMAC 27m_vs_30m dataset? | Median Win Rate, Average Score |
What metrics were used to measure the DDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QPLEX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the QMIX model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the VDN model in the A Unified Framework for Factorizing Distributional Value Functions for Multi-Agent Reinforcement Learning paper on the SMAC MMM2_7m2M1M_vs_9m3M1M dataset? | Average Score, Median Win Rate |
What metrics were used to measure the DePlot+FlanPaLM+Codex (PoT Self-Consistency) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the DePlot+Codex (PoT Self-Consistency) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the PaLI-X (Single-task FT w/ OCR) model in the PaLI-X: On Scaling up a Multilingual Vision and Language Model paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the PaLI-X (Single-task FT) model in the PaLI-X: On Scaling up a Multilingual Vision and Language Model paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the PaLI-X (Multi-task FT) model in the PaLI-X: On Scaling up a Multilingual Vision and Language Model paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the DePlot+FlanPaLM (Self-Consistency) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the PaLI-3 model in the PaLI-3 Vision Language Models: Smaller, Faster, Stronger paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the PaLI-3 (w/ OCR) model in the PaLI-3 Vision Language Models: Smaller, Faster, Stronger paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the DePlot+FlanPaLM (CoT) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the Qwen-VL-Chat model in the Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the UniChart model in the UniChart: A Universal Vision-language Pretrained Model for Chart Comprehension and Reasoning paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the Qwen-VL model in the Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the StructChart+GPT3.5 (STR ChartQA+SimChart9K) model in the StructChart: Perception, Structuring, Reasoning for Visual Chart Understanding paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the MatCha model in the MatCha: Enhancing Visual Language Pretraining with Math Reasoning and Chart Derendering paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the StructChart+GPT3.5 (STR) model in the StructChart: Perception, Structuring, Reasoning for Visual Chart Understanding paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the Pix2Struct-large model in the Pix2Struct: Screenshot Parsing as Pretraining for Visual Language Understanding paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the Pix2Struct-base model in the Pix2Struct: Screenshot Parsing as Pretraining for Visual Language Understanding paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the VisionTapas-OCR model in the ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the DePlot+GPT3 (Self-Consistency) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the DePlot+GPT3 (CoT) model in the DePlot: One-shot visual language reasoning by plot-to-table translation paper on the ChartQA dataset? | 1:1 Accuracy |
What metrics were used to measure the crct - baseline model in the ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning paper on the RealCQA dataset? | 1:1 Accuracy |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.