| | --- |
| | dataset_info: |
| | features: |
| | - name: source |
| | dtype: string |
| | - name: question |
| | dtype: string |
| | - name: answer |
| | dtype: string |
| | - name: constraint_desc |
| | list: string |
| | - name: key |
| | dtype: string |
| | - name: instruction_id_list |
| | list: string |
| | - name: kwargs |
| | list: |
| | - name: capital_frequency |
| | dtype: int64 |
| | - name: capital_relation |
| | dtype: string |
| | - name: num_words |
| | dtype: int64 |
| | - name: relation |
| | dtype: string |
| | - name: keyword |
| | dtype: string |
| | - name: frequency |
| | dtype: int64 |
| | - name: prompt_to_repeat |
| | dtype: string |
| | - name: keywords |
| | list: string |
| | - name: forbidden_words |
| | list: string |
| | - name: num_highlights |
| | dtype: int64 |
| | - name: end_phrase |
| | dtype: string |
| | - name: num_bullets |
| | dtype: int64 |
| | - name: section_spliter |
| | dtype: string |
| | - name: num_sections |
| | dtype: int64 |
| | - name: language |
| | dtype: string |
| | - name: prompt |
| | dtype: string |
| | splits: |
| | - name: dev |
| | num_bytes: 123575 |
| | num_examples: 90 |
| | - name: test |
| | num_bytes: 478304 |
| | num_examples: 332 |
| | download_size: 223359 |
| | dataset_size: 601879 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: dev |
| | path: data/dev-* |
| | - split: test |
| | path: data/test-* |
| | license: apache-2.0 |
| | language: |
| | - en |
| | size_categories: |
| | - n<1K |
| | --- |
| | |
| |
|
| |
|
| | # Math-IF Dataset Card |
| |
|
| | ## Dataset Description |
| |
|
| | Math-IF (MathIF) is an instruction-following benchmark built on top of math word problems. Each example includes a math question together with explicit, verifiable instructions about how the model should respond (e.g., format, style, or structural constraints). The benchmark is designed to jointly test: |
| |
|
| | - instruction following in the **reasoning trace (RT)** and |
| | - instruction following and correctness in the **final answer (FA)**. |
| |
|
| | In this repository, Math-IF is used as both a development set and a test benchmark for controllable reasoning models. |
| |
|
| | ## Intended Use |
| |
|
| | - Evaluate how well models follow explicit instructions when solving math problems. |
| |
|
| |
|
| | The dataset is intended for **research and benchmarking** only. |
| |
|
| | ## Dataset Structure |
| |
|
| | From the accompanying paper in this repository (see `paper/`): |
| |
|
| | - **Size**: |
| | - **Dev**: 90 examples |
| | - **Test**: 332 examples |
| | - **Splits used here**: |
| | - The **GSM8K partition** is used as dev set for model selection. |
| | - The remaining partition is used as test set. |
| |
|
| | Each instance conceptually includes: |
| |
|
| | - **`prompt`**: the user prompt with the math question and instruction. |
| | - **`answer`**: the ground-truth final answer. |
| | - **`question`**: the underlying math word problem (without instructions). |
| | - **metadata for evaluation**: information needed to compute instruction-following metrics and answer accuracy. |
| |
|
| | ## Tasks and Evaluation |
| |
|
| | - **Main task**: Instruction-following on math problems. |
| | - **Metrics**: |
| | - *Instruction-level loose-accuracy* (as defined in the Math-IF paper) for both RTs and FAs, yielding **IF-RT** and **IF-FA**. |
| | - **Answer accuracy** measuring whether the final numeric answer is correct. |
| |
|
| | ## Data Source |
| |
|
| | Math-IF was introduced to study the trade-off between reasoning performance and instruction-following in large reasoning models. For complete details, examples, and official evaluation scripts, please see the original Math-IF paper and repository. |
| |
|
| | ## License |
| |
|
| | - **License**: Apache 2.0 |
| |
|
| | ## Known Limitations and Considerations |
| |
|
| | - The dataset focuses on **math word problems**, so instruction-following performance may differ on other domains (e.g., open-ended dialogue, code generation). |
| | - The benchmark size is modest (422 examples total in the dev+test configuration used here), which can make very fine-grained comparisons noisy. |
| | - Instructions are in **English**, so the benchmark does not directly evaluate multilingual behavior. |
| |
|
| | ## Citation |
| |
|
| |
|
| | ```bibtex |
| | @article{fu2025scaling, |
| | title={Scaling Reasoning, Losing Control: Evaluating Instruction Following in Large Reasoning Models}, |
| | author={Fu, Tingchen and Gu, Jiawei and Li, Yafu and Qu, Xiaoye and Cheng, Yu}, |
| | journal={arXiv preprint arXiv:2505.14810}, |
| | year={2025} |
| | } |
| | ``` |