End of training
Browse files- README.md +1 -1
- all_results.json +1 -1
- eval_results.json +1 -1
README.md
CHANGED
|
@@ -26,7 +26,7 @@ print(output["generated_text"])
|
|
| 26 |
|
| 27 |
## Training procedure
|
| 28 |
|
| 29 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shahrad_m/AIFGen-ppo-continual-test/runs/
|
| 30 |
|
| 31 |
|
| 32 |
This model was trained with PPO, a method introduced in [Fine-Tuning Language Models from Human Preferences](https://huggingface.co/papers/1909.08593).
|
|
|
|
| 26 |
|
| 27 |
## Training procedure
|
| 28 |
|
| 29 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shahrad_m/AIFGen-ppo-continual-test/runs/3pa7x10a)
|
| 30 |
|
| 31 |
|
| 32 |
This model was trained with PPO, a method introduced in [Fine-Tuning Language Models from Human Preferences](https://huggingface.co/papers/1909.08593).
|
all_results.json
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
{
|
| 2 |
"dataset": 0,
|
| 3 |
-
"eval_score":
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"dataset": 0,
|
| 3 |
+
"eval_score": 0.616264283657074
|
| 4 |
}
|
eval_results.json
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
{
|
| 2 |
"dataset": 0,
|
| 3 |
-
"eval_score":
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"dataset": 0,
|
| 3 |
+
"eval_score": 0.616264283657074
|
| 4 |
}
|