Lansechen commited on
Commit
d954f6a
·
verified ·
1 Parent(s): 4d35c6d

Model save

Browse files
Files changed (4) hide show
  1. README.md +6 -5
  2. all_results.json +3 -3
  3. train_results.json +3 -3
  4. trainer_state.json +0 -0
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-7B
3
- datasets: DigitalLearningGmbH/MATH-lighteval
4
  library_name: transformers
 
5
  tags:
6
  - generated_from_trainer
7
- - open-r1
 
8
  licence: license
9
  ---
10
 
11
- # Model Card for None
12
 
13
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) on the [DigitalLearningGmbH/MATH-lighteval](https://huggingface.co/datasets/DigitalLearningGmbH/MATH-lighteval) dataset.
14
  It has been trained using [TRL](https://github.com/huggingface/trl).
15
 
16
  ## Quick start
@@ -26,7 +27,7 @@ print(output["generated_text"])
26
 
27
  ## Training procedure
28
 
29
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/chenran1995-the-chinese-university-of-hong-kong/huggingface/runs/ljeo7hai)
30
 
31
 
32
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
1
  ---
2
  base_model: Qwen/Qwen2.5-7B
 
3
  library_name: transformers
4
+ model_name: Qwen2.5-7B-Open-R1-GRPO-math-lighteval-weighted-sync
5
  tags:
6
  - generated_from_trainer
7
+ - trl
8
+ - grpo
9
  licence: license
10
  ---
11
 
12
+ # Model Card for Qwen2.5-7B-Open-R1-GRPO-math-lighteval-weighted-sync
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/chenran1995-the-chinese-university-of-hong-kong/huggingface/runs/plhx359x)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.02047422327097703,
4
- "train_runtime": 34869.9992,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.43,
7
  "train_steps_per_second": 0.004
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.019874975508476684,
4
+ "train_runtime": 35069.7154,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.428,
7
  "train_steps_per_second": 0.004
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.02047422327097703,
4
- "train_runtime": 34869.9992,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.43,
7
  "train_steps_per_second": 0.004
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.019874975508476684,
4
+ "train_runtime": 35069.7154,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.428,
7
  "train_steps_per_second": 0.004
8
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff