ChenDRAG commited on
Commit
b13d672
·
verified ·
1 Parent(s): 7d73a24

Model save

Browse files
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
- datasets: ChenDRAG/OM220k
4
  library_name: transformers
 
5
  tags:
6
  - generated_from_trainer
7
- - open-r1
 
8
  licence: license
9
  ---
10
 
11
- # Model Card for None
12
 
13
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the [ChenDRAG/OM220k](https://huggingface.co/datasets/ChenDRAG/OM220k) dataset.
14
  It has been trained using [TRL](https://github.com/huggingface/trl).
15
 
16
  ## Quick start
@@ -19,14 +20,14 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
19
  from transformers import pipeline
20
 
21
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
22
- generator = pipeline("text-generation", model="None", device="cuda")
23
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
24
  print(output["generated_text"])
25
  ```
26
 
27
  ## Training procedure
28
 
29
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/noteam2235/huggingface/runs/6bxdmvd8)
30
 
31
 
32
  This model was trained with SFT.
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
 
3
  library_name: transformers
4
+ model_name: Qwen2.5-1.5B-Open-R1-Distill
5
  tags:
6
  - generated_from_trainer
7
+ - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
12
+ # Model Card for Qwen2.5-1.5B-Open-R1-Distill
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
20
  from transformers import pipeline
21
 
22
  question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="ChenDRAG/Qwen2.5-1.5B-Open-R1-Distill", device="cuda")
24
  output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
  print(output["generated_text"])
26
  ```
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/noteam2235/huggingface/runs/ctc7qeil)
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 427315691520.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 2.1723,
5
  "train_samples": 100,
6
- "train_samples_per_second": 46.034,
7
- "train_steps_per_second": 0.46
8
  }
 
1
  {
2
  "total_flos": 427315691520.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 3.3707,
5
  "train_samples": 100,
6
+ "train_samples_per_second": 10.977,
7
+ "train_steps_per_second": 0.593
8
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
- "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
+ "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151936
29
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92ebae36cd2df3923a2524f695d1a18f61dea8aaf923959966556af36d6e43ff
3
  size 3087467144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:081546ce9a9403a817de3a122319a951510c3396cd7631c26afdac41207fedc4
3
  size 3087467144
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 427315691520.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 2.1723,
5
  "train_samples": 100,
6
- "train_samples_per_second": 46.034,
7
- "train_steps_per_second": 0.46
8
  }
 
1
  {
2
  "total_flos": 427315691520.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 3.3707,
5
  "train_samples": 100,
6
+ "train_samples_per_second": 10.977,
7
+ "train_steps_per_second": 0.593
8
  }
trainer_state.json CHANGED
@@ -13,13 +13,13 @@
13
  "step": 2,
14
  "total_flos": 427315691520.0,
15
  "train_loss": 0.0,
16
- "train_runtime": 2.1723,
17
- "train_samples_per_second": 46.034,
18
- "train_steps_per_second": 0.46
19
  }
20
  ],
21
  "logging_steps": 500,
22
- "max_steps": 1,
23
  "num_input_tokens_seen": 0,
24
  "num_train_epochs": 1,
25
  "save_steps": 500,
 
13
  "step": 2,
14
  "total_flos": 427315691520.0,
15
  "train_loss": 0.0,
16
+ "train_runtime": 3.3707,
17
+ "train_samples_per_second": 10.977,
18
+ "train_steps_per_second": 0.593
19
  }
20
  ],
21
  "logging_steps": 500,
22
+ "max_steps": 2,
23
  "num_input_tokens_seen": 0,
24
  "num_train_epochs": 1,
25
  "save_steps": 500,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a4d6fd11ad02d7543ed5a479292a62376452da8c54cfaec33b422758a4027fa
3
  size 6008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61d29d81d710536ac7d16eccea481e3bdb272b565be4989aeb12e5534122506f
3
  size 6008