--- license: apache-2.0 datasets: - Miaow-Lab/RLVR-Linearity-Dataset base_model: - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B pipeline_tag: text-generation --- # Model Card ## 1. Model Details This model is the fine-tuned checkpoint described in the paper **"Not All Steps are Informative: On the Linearity of LLMs’ RLVR Training"**. It was trained using Reinforcement Learning (RL) to enhance reasoning capabilities. - **Paper:** [ArXiv](https://arxiv.org/pdf/2601.04537v1) - **Code:** [Github](https://github.com/Miaow-Lab/RLVR-Linearity) - **Base Model:** [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) - **Training Method:** GRPO ## 2. Training Details - **Hyperparameters:** - Learning Rate: `1e-6` - Train Batch Size: `128` - PPO Mini Batch Size: `64` - RL Algorithm: `GRPO` - Rollout Temperature: 1.0 - Group Size: 16 - **Compute:** Trained on `32 x H100` GPUs for about `150` hours. For full training configurations, please refer to the `config.json` or the training scripts in our [GitHub](https://github.com/Miaow-Lab/RLVR-Linearity). ## 3. Citation If you use this model in your research, please cite our paper: ```bibtex @misc{wang2026stepsinformativelinearityllms, title={Not All Steps are Informative: On the Linearity of LLMs' RLVR Training}, author={Tianle Wang and Zhongyuan Wu and Shenghao Jin and Hao Xu and Wei Chen and Ning Miao}, year={2026}, eprint={2601.04537}, archivePrefix={arXiv}, primaryClass={cs.LG}, url={https://arxiv.org/abs/2601.04537}, } ``` > [!TIP] > **Motivation for this Model** > This checkpoint is released primarily as a research artifact to facilitate the analysis of linearity in model outputs and weight updates during RLVR fine‑tuning.