bmbgsj commited on
Commit
699561d
·
verified ·
1 Parent(s): b7a5eac

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - zh
5
+ license: apache-2.0
6
+ library_name: transformers
7
+ tags:
8
+ - qwen3
9
+ - text-generation
10
+ - casual-lm
11
+ base_model: Qwen/Qwen3-8B
12
+ pipeline_tag: text-generation
13
+ arxiv: 2601.21912
14
+ ---
15
+
16
+ # Model Card for ProRAG
17
+
18
+ This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) based on the methodology described in the paper associated with arXiv ID: **2601.21912**.
19
+
20
+ ## Model Details
21
+
22
+ - **Base Model:** Qwen3-8B
23
+ - **Language:** English, Chinese (and others supported by Qwen3)
24
+ - **Paper:** [View on arXiv](https://arxiv.org/abs/2601.21912)
25
+ - **Library:** Transformers
26
+
27
+ ## 💻 Code & Inference
28
+
29
+ For inference code, usage examples, and reproduction scripts, please refer to our GitHub repository:
30
+
31
+ 👉 **[Click here to view the GitHub Repository](https://github.com/lilinwz/ProRAG/tree/main)**
32
+
33
+ *(Please verify the details and instructions on the GitHub page.)*
34
+
35
+ ## Citation
36
+
37
+ If you use this model or the associated paper in your research, please cite:
38
+
39
+ ```bibtex
40
+ @misc{wang2026proragprocesssupervisedreinforcementlearning,
41
+ title={ProRAG: Process-Supervised Reinforcement Learning for Retrieval-Augmented Generation},
42
+ author={Zhao Wang and Ziliang Zhao and Zhicheng Dou},
43
+ year={2026},
44
+ eprint={2601.21912},
45
+ archivePrefix={arXiv},
46
+ primaryClass={cs.AI},
47
+ url={https://arxiv.org/abs/2601.21912},
48
+ }
49
+ ```