| --- |
| language: |
| - en |
| - zh |
| license: apache-2.0 |
| library_name: transformers |
| tags: |
| - qwen3 |
| - text-generation |
| - casual-lm |
| base_model: Qwen/Qwen3-8B |
| pipeline_tag: text-generation |
| arxiv: 2601.21912 |
| --- |
| |
| # Model Card for ProRAG |
|
|
| This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) based on the methodology described in the paper associated with arXiv ID: **2601.21912**. |
|
|
| ## Model Details |
|
|
| - **Base Model:** Qwen3-8B |
| - **Language:** English, Chinese (and others supported by Qwen3) |
| - **Paper:** [View on arXiv](https://arxiv.org/abs/2601.21912) |
| - **Library:** Transformers |
|
|
| ## 💻 Code & Inference |
|
|
| For inference code, usage examples, and reproduction scripts, please refer to our GitHub repository: |
|
|
| 👉 **[Click here to view the GitHub Repository](https://github.com/lilinwz/ProRAG/tree/main)** |
|
|
| *(Please verify the details and instructions on the GitHub page.)* |
|
|
| ## Citation |
|
|
| If you use this model or the associated paper in your research, please cite: |
|
|
| ```bibtex |
| @misc{wang2026proragprocesssupervisedreinforcementlearning, |
| title={ProRAG: Process-Supervised Reinforcement Learning for Retrieval-Augmented Generation}, |
| author={Zhao Wang and Ziliang Zhao and Zhicheng Dou}, |
| year={2026}, |
| eprint={2601.21912}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.AI}, |
| url={https://arxiv.org/abs/2601.21912}, |
| } |
| ``` |