Update README.md
Browse files
README.md
CHANGED
|
@@ -31,7 +31,7 @@ WildReward is trained using **ordinal regression** (CORAL-like approach) on the
|
|
| 31 |
- **Source:** WildChat - large-scale human-LLM interactions
|
| 32 |
- **Labeling:** 5-point ordinal scale based on user satisfaction signals
|
| 33 |
- **Filtering:** Two-stage refinement including implicit feedback mining and refusal validation
|
| 34 |
-
- **License:**
|
| 35 |
|
| 36 |
## Usage
|
| 37 |
|
|
@@ -41,7 +41,7 @@ WildReward is trained using **ordinal regression** (CORAL-like approach) on the
|
|
| 41 |
import torch
|
| 42 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 43 |
|
| 44 |
-
model_name = "
|
| 45 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 46 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 47 |
|
|
|
|
| 31 |
- **Source:** WildChat - large-scale human-LLM interactions
|
| 32 |
- **Labeling:** 5-point ordinal scale based on user satisfaction signals
|
| 33 |
- **Filtering:** Two-stage refinement including implicit feedback mining and refusal validation
|
| 34 |
+
- **License:** MIT
|
| 35 |
|
| 36 |
## Usage
|
| 37 |
|
|
|
|
| 41 |
import torch
|
| 42 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 43 |
|
| 44 |
+
model_name = "THU-KEG/WildReward-4B"
|
| 45 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 46 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 47 |
|