Add paper link, fix repository URL, and update metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +15 -9
README.md CHANGED
@@ -1,19 +1,25 @@
1
  ---
2
- license: apache-2.0
3
  base_model:
4
  - Qwen/Qwen3-8B
5
  library_name: transformers
 
 
 
 
6
  tags:
7
  - reward-model
8
  - rlhf
9
  - dpo
10
  - alignment
11
  - wildchat
 
12
  ---
13
 
14
  # WildReward
15
 
16
- WildReward is a reward model trained on in-the-wild human-LLM interactions from the WildChat dataset. Unlike conventional reward models that rely on expensive human-annotated preference pairs, WildReward extracts implicit reward signals from real-world user feedback through an automated pipeline.
 
 
17
 
18
  ## Model Details
19
 
@@ -96,12 +102,6 @@ with torch.no_grad():
96
  print(f"Reward score: {reward:.2f} (scale: 1-5)")
97
  ```
98
 
99
-
100
- **Architecture:**
101
- - Router on port 9000 with round-robin load balancing
102
- - Multiple workers on dedicated GPUs (ports 8004-8007)
103
- - FP16 inference with batch processing
104
-
105
  ## Performance
106
 
107
  WildReward achieves competitive results on standard reward model benchmarks while demonstrating superior calibration properties. When applied to Online DPO, it significantly improves performance in mathematical reasoning, instruction following, and creative writing tasks.
@@ -109,6 +109,12 @@ WildReward achieves competitive results on standard reward model benchmarks whil
109
  ## Citation
110
 
111
  ```bibtex
 
 
 
 
 
 
112
  ```
113
 
114
  ## License
@@ -117,4 +123,4 @@ Apache License 2.0
117
 
118
  ---
119
 
120
- **Note:** This model card provides a brief overview. For detailed documentation on data collection, training, and deployment, please visit the [GitHub repository](https://github.com/yourusername/WildReward).
 
1
  ---
 
2
  base_model:
3
  - Qwen/Qwen3-8B
4
  library_name: transformers
5
+ license: apache-2.0
6
+ pipeline_tag: text-classification
7
+ datasets:
8
+ - THU-KEG/WildFB
9
  tags:
10
  - reward-model
11
  - rlhf
12
  - dpo
13
  - alignment
14
  - wildchat
15
+ arxiv: 2602.08829
16
  ---
17
 
18
  # WildReward
19
 
20
+ WildReward is a reward model presented in the paper [WildReward: Learning Reward Models from In-the-Wild Human Interactions](https://huggingface.co/papers/2602.08829).
21
+
22
+ It is trained on in-the-wild human-LLM interactions from the WildChat dataset. Unlike conventional reward models that rely on expensive human-annotated preference pairs, WildReward extracts implicit reward signals from real-world user feedback through an automated pipeline.
23
 
24
  ## Model Details
25
 
 
102
  print(f"Reward score: {reward:.2f} (scale: 1-5)")
103
  ```
104
 
 
 
 
 
 
 
105
  ## Performance
106
 
107
  WildReward achieves competitive results on standard reward model benchmarks while demonstrating superior calibration properties. When applied to Online DPO, it significantly improves performance in mathematical reasoning, instruction following, and creative writing tasks.
 
109
  ## Citation
110
 
111
  ```bibtex
112
+ @article{peng2026wildreward,
113
+ title={WildReward: Learning Reward Models from In-the-Wild Human Interactions},
114
+ author={Peng, Hao and Qi, Yunjia and Wang, Xiaozhi and Yao, Zijun and Hou, Lei and Li, Juanzi},
115
+ journal={arXiv preprint arXiv:2602.08829},
116
+ year={2026}
117
+ }
118
  ```
119
 
120
  ## License
 
123
 
124
  ---
125
 
126
+ **Note:** This model card provides a brief overview. For detailed documentation on data collection, training, and deployment, please visit the [GitHub repository](https://github.com/THU-KEG/WildReward).