kangdawei commited on
Commit
2d19634
·
verified ·
1 Parent(s): adc2c55

End of training

Browse files
README.md CHANGED
@@ -1,17 +1,19 @@
1
  ---
2
  base_model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
 
3
  library_name: transformers
4
  model_name: DAPO-8B
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - dapo
 
9
  licence: license
10
  ---
11
 
12
  # Model Card for DAPO-8B
13
 
14
- This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
1
  ---
2
  base_model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
3
+ datasets: knoveleng/open-rs
4
  library_name: transformers
5
  model_name: DAPO-8B
6
  tags:
7
  - generated_from_trainer
8
+ - open-r1
9
  - dapo
10
+ - trl
11
  licence: license
12
  ---
13
 
14
  # Model Card for DAPO-8B
15
 
16
+ This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B) on the [knoveleng/open-rs](https://huggingface.co/datasets/knoveleng/open-rs) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed5057e58d9c86f27f25bff17f93ec31710991515dde331e78b702be3b250793
3
- size 335605144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60d95b10b6e140a9626a7058d5038528f2ff80148dc4569b881db56052046509
3
+ size 40
config.json CHANGED
@@ -18,6 +18,7 @@
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
 
21
  "pretraining_tp": 1,
22
  "rms_norm_eps": 1e-05,
23
  "rope_scaling": {
 
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
21
+ "pad_token_id": 128001,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
generation_config.json CHANGED
@@ -2,7 +2,10 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
- "eos_token_id": 128001,
 
 
 
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
  "transformers_version": "4.57.1"
 
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
+ "eos_token_id": [
6
+ 128001
7
+ ],
8
+ "pad_token_id": 128001,
9
  "temperature": 0.6,
10
  "top_p": 0.95,
11
  "transformers_version": "4.57.1"
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36c984f65c9eedc58fe094e25c972d3709f8cddf1340fed3ebce67c17afbe06c
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f30f112920d1e7b4852aa1a8523b5318ceaeadc79ccbb259c8ce88affedff3f
3
  size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a030a1b0b2b180923fa99a8b0296488c5e16daa53c33ebb8f0d6fbc2a6b2dc1
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad8d9d161c36affea92f78179b228855b0d53f921b4bb4c773cf06fd8d657391
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b546fd3f6d8c9d12ac6ff2199aa18b6f14b3970dda812f2325b38cda62b1c91
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2427ec8e0f491bb265bd37a808b85655ce1bdd7d17f223a48ef547154cbd8f5
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ce2f3c51b0f3f003535366a56f9348f58eeb46c3b8eee8217b7346bd5352ec9
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0622c5d71a8f922509f93ebaceb413ad5277c8bce1b81800d679262b910e8d3
3
  size 1168138808