--- dataset_info: features: - name: source dtype: string - name: question dtype: string - name: answer dtype: string - name: constraint_desc list: string - name: key dtype: string - name: instruction_id_list list: string - name: kwargs list: - name: capital_frequency dtype: int64 - name: capital_relation dtype: string - name: num_words dtype: int64 - name: relation dtype: string - name: keyword dtype: string - name: frequency dtype: int64 - name: prompt_to_repeat dtype: string - name: keywords list: string - name: forbidden_words list: string - name: num_highlights dtype: int64 - name: end_phrase dtype: string - name: num_bullets dtype: int64 - name: section_spliter dtype: string - name: num_sections dtype: int64 - name: language dtype: string - name: prompt dtype: string splits: - name: dev num_bytes: 123575 num_examples: 90 - name: test num_bytes: 478304 num_examples: 332 download_size: 223359 dataset_size: 601879 configs: - config_name: default data_files: - split: dev path: data/dev-* - split: test path: data/test-* license: apache-2.0 language: - en size_categories: - n<1K --- # Math-IF Dataset Card ## Dataset Description Math-IF (MathIF) is an instruction-following benchmark built on top of math word problems. Each example includes a math question together with explicit, verifiable instructions about how the model should respond (e.g., format, style, or structural constraints). The benchmark is designed to jointly test: - instruction following in the **reasoning trace (RT)** and - instruction following and correctness in the **final answer (FA)**. In this repository, Math-IF is used as both a development set and a test benchmark for controllable reasoning models. ## Intended Use - Evaluate how well models follow explicit instructions when solving math problems. The dataset is intended for **research and benchmarking** only. ## Dataset Structure From the accompanying paper in this repository (see `paper/`): - **Size**: - **Dev**: 90 examples - **Test**: 332 examples - **Splits used here**: - The **GSM8K partition** is used as dev set for model selection. - The remaining partition is used as test set. Each instance conceptually includes: - **`prompt`**: the user prompt with the math question and instruction. - **`answer`**: the ground-truth final answer. - **`question`**: the underlying math word problem (without instructions). - **metadata for evaluation**: information needed to compute instruction-following metrics and answer accuracy. ## Tasks and Evaluation - **Main task**: Instruction-following on math problems. - **Metrics**: - *Instruction-level loose-accuracy* (as defined in the Math-IF paper) for both RTs and FAs, yielding **IF-RT** and **IF-FA**. - **Answer accuracy** measuring whether the final numeric answer is correct. ## Data Source Math-IF was introduced to study the trade-off between reasoning performance and instruction-following in large reasoning models. For complete details, examples, and official evaluation scripts, please see the original Math-IF paper and repository. ## License - **License**: Apache 2.0 ## Known Limitations and Considerations - The dataset focuses on **math word problems**, so instruction-following performance may differ on other domains (e.g., open-ended dialogue, code generation). - The benchmark size is modest (422 examples total in the dev+test configuration used here), which can make very fine-grained comparisons noisy. - Instructions are in **English**, so the benchmark does not directly evaluate multilingual behavior. ## Citation ```bibtex @article{fu2025scaling, title={Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models}, author={Fu, Tingchen and Gu, Jiawei and Li, Yafu and Qu, Xiaoye and Cheng, Yu}, journal={arXiv preprint arXiv:2505.14810}, year={2025} } ```