hzk17 commited on
Commit
9f98cc8
·
verified ·
1 Parent(s): 9a95b41

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
- datasets: open-r1/OpenR1-Math-220k
4
  library_name: transformers
5
  model_name: Qwen2.5-1.5B-Open-R1-Distill
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - sft
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for Qwen2.5-1.5B-Open-R1-Distill
15
 
16
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the [open-r1/OpenR1-Math-220k](https://huggingface.co/datasets/open-r1/OpenR1-Math-220k) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/haozhongkai/huggingface/runs/2jjht3cx)
33
 
34
 
35
  This model was trained with SFT.
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
 
3
  library_name: transformers
4
  model_name: Qwen2.5-1.5B-Open-R1-Distill
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for Qwen2.5-1.5B-Open-R1-Distill
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/haozhongkai/huggingface/runs/8jf13p5l)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 488621249396736.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 2.0029,
5
  "train_samples": 93733,
6
- "train_samples_per_second": 17125.887,
7
- "train_steps_per_second": 133.808
8
  }
 
1
  {
2
  "total_flos": 488621249396736.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 1.9989,
5
  "train_samples": 93733,
6
+ "train_samples_per_second": 17160.018,
7
+ "train_steps_per_second": 134.074
8
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0",
26
- "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0",
26
+ "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 488621249396736.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 2.0029,
5
  "train_samples": 93733,
6
- "train_samples_per_second": 17125.887,
7
- "train_steps_per_second": 133.808
8
  }
 
1
  {
2
  "total_flos": 488621249396736.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 1.9989,
5
  "train_samples": 93733,
6
+ "train_samples_per_second": 17160.018,
7
+ "train_steps_per_second": 134.074
8
  }
trainer_state.json CHANGED
@@ -384,9 +384,9 @@
384
  "step": 268,
385
  "total_flos": 488621249396736.0,
386
  "train_loss": 0.0,
387
- "train_runtime": 2.0029,
388
- "train_samples_per_second": 17125.887,
389
- "train_steps_per_second": 133.808
390
  }
391
  ],
392
  "logging_steps": 5,
 
384
  "step": 268,
385
  "total_flos": 488621249396736.0,
386
  "train_loss": 0.0,
387
+ "train_runtime": 1.9989,
388
+ "train_samples_per_second": 17160.018,
389
+ "train_steps_per_second": 134.074
390
  }
391
  ],
392
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cc7e92627a95e87b9fda238e9948f04fed77920a0b9bf1ef729ff479e24bdb7
3
  size 7480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf0c3f4a11cfed8e3235a87da29a911f5ba3a4380bd6fc7ba8c12ff643fb2d5
3
  size 7480