ChenDRAG commited on
Commit
7d73a24
·
verified ·
1 Parent(s): 39fed3e

Model save

Browse files
README.md CHANGED
@@ -1,17 +1,16 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
 
3
  library_name: transformers
4
- model_name: Qwen2.5-1.5B-Open-R1-Distill
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
- - sft
9
  licence: license
10
  ---
11
 
12
- # Model Card for Qwen2.5-1.5B-Open-R1-Distill
13
 
14
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -20,14 +19,14 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
20
  from transformers import pipeline
21
 
22
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="ChenDRAG/Qwen2.5-1.5B-Open-R1-Distill", device="cuda")
24
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
  print(output["generated_text"])
26
  ```
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/noteam2235/huggingface/runs/ojbm1qwt)
31
 
32
 
33
  This model was trained with SFT.
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
+ datasets: ChenDRAG/OM220k
4
  library_name: transformers
 
5
  tags:
6
  - generated_from_trainer
7
+ - open-r1
 
8
  licence: license
9
  ---
10
 
11
+ # Model Card for None
12
 
13
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the [ChenDRAG/OM220k](https://huggingface.co/datasets/ChenDRAG/OM220k) dataset.
14
  It has been trained using [TRL](https://github.com/huggingface/trl).
15
 
16
  ## Quick start
 
19
  from transformers import pipeline
20
 
21
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
+ generator = pipeline("text-generation", model="None", device="cuda")
23
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
  print(output["generated_text"])
25
  ```
26
 
27
  ## Training procedure
28
 
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/noteam2235/huggingface/runs/6bxdmvd8)
30
 
31
 
32
  This model was trained with SFT.
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 2937605436813312.0,
3
- "train_loss": 0.8297107517719269,
4
- "train_runtime": 42.4708,
5
  "train_samples": 100,
6
- "train_samples_per_second": 2.355,
7
- "train_steps_per_second": 0.141
8
  }
 
1
  {
2
+ "total_flos": 427315691520.0,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 2.1723,
5
  "train_samples": 100,
6
+ "train_samples_per_second": 46.034,
7
+ "train_steps_per_second": 0.46
8
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
- "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
+ "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4734070617070f746b58b580fa2ddf9444e9a2b0b21faaeff0f0233be1222769
3
  size 3087467144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ebae36cd2df3923a2524f695d1a18f61dea8aaf923959966556af36d6e43ff
3
  size 3087467144
runs/Feb17_02-30-57_8f59999e8611/events.out.tfevents.1739759469.8f59999e8611.119099.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:977e270f83280329b32b2a7172282796430e10f43e6527c3c35e3fe5881fba77
3
+ size 6272
runs/Feb17_02-33-36_8f59999e8611/events.out.tfevents.1739759635.8f59999e8611.123913.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22d7c11cd4029576ff26572239db75ce96d5efff623465bf328b8149fa59e5b3
3
+ size 6209
runs/Feb17_02-36-20_8f59999e8611/events.out.tfevents.1739759798.8f59999e8611.130416.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04335c80070cf72380ba3ef095e9e5e71aed4745b09ebdb10a4f58030a936d59
3
+ size 6210
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "total_flos": 2937605436813312.0,
3
- "train_loss": 0.8297107517719269,
4
- "train_runtime": 42.4708,
5
  "train_samples": 100,
6
- "train_samples_per_second": 2.355,
7
- "train_steps_per_second": 0.141
8
  }
 
1
  {
2
+ "total_flos": 427315691520.0,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 2.1723,
5
  "train_samples": 100,
6
+ "train_samples_per_second": 46.034,
7
+ "train_steps_per_second": 0.46
8
  }
trainer_state.json CHANGED
@@ -1,34 +1,25 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.96,
5
- "eval_steps": 100,
6
- "global_step": 6,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.8,
13
- "grad_norm": 1.609375,
14
- "learning_rate": 1.9098300562505266e-06,
15
- "loss": 0.8248,
16
- "mean_token_accuracy": 0.7740710348407838,
17
- "step": 5
18
- },
19
- {
20
- "epoch": 0.96,
21
- "mean_token_accuracy": 0.7635029501317351,
22
- "step": 6,
23
- "total_flos": 2937605436813312.0,
24
- "train_loss": 0.8297107517719269,
25
- "train_runtime": 42.4708,
26
- "train_samples_per_second": 2.355,
27
- "train_steps_per_second": 0.141
28
  }
29
  ],
30
- "logging_steps": 5,
31
- "max_steps": 6,
32
  "num_input_tokens_seen": 0,
33
  "num_train_epochs": 1,
34
  "save_steps": 500,
@@ -38,13 +29,13 @@
38
  "should_epoch_stop": false,
39
  "should_evaluate": false,
40
  "should_log": false,
41
- "should_save": false,
42
- "should_training_stop": false
43
  },
44
  "attributes": {}
45
  }
46
  },
47
- "total_flos": 2937605436813312.0,
48
  "train_batch_size": 2,
49
  "trial_name": null,
50
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 2,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 1.0,
13
+ "step": 2,
14
+ "total_flos": 427315691520.0,
15
+ "train_loss": 0.0,
16
+ "train_runtime": 2.1723,
17
+ "train_samples_per_second": 46.034,
18
+ "train_steps_per_second": 0.46
 
 
 
 
 
 
 
 
 
19
  }
20
  ],
21
+ "logging_steps": 500,
22
+ "max_steps": 1,
23
  "num_input_tokens_seen": 0,
24
  "num_train_epochs": 1,
25
  "save_steps": 500,
 
29
  "should_epoch_stop": false,
30
  "should_evaluate": false,
31
  "should_log": false,
32
+ "should_save": true,
33
+ "should_training_stop": true
34
  },
35
  "attributes": {}
36
  }
37
  },
38
+ "total_flos": 427315691520.0,
39
  "train_batch_size": 2,
40
  "trial_name": null,
41
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65529b6d979e05bcd41a7fc218b22003e53c740b022f317f959e3bad0ee633bd
3
  size 6008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a4d6fd11ad02d7543ed5a479292a62376452da8c54cfaec33b422758a4027fa
3
  size 6008