Add paper link, fix repository URL, and update metadata
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,19 +1,25 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- Qwen/Qwen3-8B
|
| 5 |
library_name: transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
tags:
|
| 7 |
- reward-model
|
| 8 |
- rlhf
|
| 9 |
- dpo
|
| 10 |
- alignment
|
| 11 |
- wildchat
|
|
|
|
| 12 |
---
|
| 13 |
|
| 14 |
# WildReward
|
| 15 |
|
| 16 |
-
WildReward is a reward model
|
|
|
|
|
|
|
| 17 |
|
| 18 |
## Model Details
|
| 19 |
|
|
@@ -96,12 +102,6 @@ with torch.no_grad():
|
|
| 96 |
print(f"Reward score: {reward:.2f} (scale: 1-5)")
|
| 97 |
```
|
| 98 |
|
| 99 |
-
|
| 100 |
-
**Architecture:**
|
| 101 |
-
- Router on port 9000 with round-robin load balancing
|
| 102 |
-
- Multiple workers on dedicated GPUs (ports 8004-8007)
|
| 103 |
-
- FP16 inference with batch processing
|
| 104 |
-
|
| 105 |
## Performance
|
| 106 |
|
| 107 |
WildReward achieves competitive results on standard reward model benchmarks while demonstrating superior calibration properties. When applied to Online DPO, it significantly improves performance in mathematical reasoning, instruction following, and creative writing tasks.
|
|
@@ -109,6 +109,12 @@ WildReward achieves competitive results on standard reward model benchmarks whil
|
|
| 109 |
## Citation
|
| 110 |
|
| 111 |
```bibtex
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
```
|
| 113 |
|
| 114 |
## License
|
|
@@ -117,4 +123,4 @@ Apache License 2.0
|
|
| 117 |
|
| 118 |
---
|
| 119 |
|
| 120 |
-
**Note:** This model card provides a brief overview. For detailed documentation on data collection, training, and deployment, please visit the [GitHub repository](https://github.com/
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen3-8B
|
| 4 |
library_name: transformers
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
pipeline_tag: text-classification
|
| 7 |
+
datasets:
|
| 8 |
+
- THU-KEG/WildFB
|
| 9 |
tags:
|
| 10 |
- reward-model
|
| 11 |
- rlhf
|
| 12 |
- dpo
|
| 13 |
- alignment
|
| 14 |
- wildchat
|
| 15 |
+
arxiv: 2602.08829
|
| 16 |
---
|
| 17 |
|
| 18 |
# WildReward
|
| 19 |
|
| 20 |
+
WildReward is a reward model presented in the paper [WildReward: Learning Reward Models from In-the-Wild Human Interactions](https://huggingface.co/papers/2602.08829).
|
| 21 |
+
|
| 22 |
+
It is trained on in-the-wild human-LLM interactions from the WildChat dataset. Unlike conventional reward models that rely on expensive human-annotated preference pairs, WildReward extracts implicit reward signals from real-world user feedback through an automated pipeline.
|
| 23 |
|
| 24 |
## Model Details
|
| 25 |
|
|
|
|
| 102 |
print(f"Reward score: {reward:.2f} (scale: 1-5)")
|
| 103 |
```
|
| 104 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
## Performance
|
| 106 |
|
| 107 |
WildReward achieves competitive results on standard reward model benchmarks while demonstrating superior calibration properties. When applied to Online DPO, it significantly improves performance in mathematical reasoning, instruction following, and creative writing tasks.
|
|
|
|
| 109 |
## Citation
|
| 110 |
|
| 111 |
```bibtex
|
| 112 |
+
@article{peng2026wildreward,
|
| 113 |
+
title={WildReward: Learning Reward Models from In-the-Wild Human Interactions},
|
| 114 |
+
author={Peng, Hao and Qi, Yunjia and Wang, Xiaozhi and Yao, Zijun and Hou, Lei and Li, Juanzi},
|
| 115 |
+
journal={arXiv preprint arXiv:2602.08829},
|
| 116 |
+
year={2026}
|
| 117 |
+
}
|
| 118 |
```
|
| 119 |
|
| 120 |
## License
|
|
|
|
| 123 |
|
| 124 |
---
|
| 125 |
|
| 126 |
+
**Note:** This model card provides a brief overview. For detailed documentation on data collection, training, and deployment, please visit the [GitHub repository](https://github.com/THU-KEG/WildReward).
|