kangdawei commited on
Commit
44fe5e1
·
verified ·
1 Parent(s): ffd2409

Training in progress, step 110

Browse files
adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "v_proj",
33
  "k_proj",
 
34
  "gate_proj",
35
- "q_proj",
36
  "o_proj",
37
- "up_proj",
38
- "down_proj"
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
32
  "k_proj",
33
+ "v_proj",
34
  "gate_proj",
 
35
  "o_proj",
36
+ "q_proj",
37
+ "down_proj",
38
+ "up_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60d95b10b6e140a9626a7058d5038528f2ff80148dc4569b881db56052046509
3
- size 40
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dc2767d43ff64fdc473d3da757398d1dbefb6920eb8b0523fec8c503ee7e7c8
3
+ size 323014560
config.json CHANGED
@@ -3,7 +3,7 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "dtype": "bfloat16",
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
@@ -46,6 +46,7 @@
46
  "num_attention_heads": 28,
47
  "num_hidden_layers": 28,
48
  "num_key_value_heads": 4,
 
49
  "rms_norm_eps": 1e-06,
50
  "rope_scaling": null,
51
  "rope_theta": 10000,
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 151646,
7
  "dtype": "bfloat16",
8
  "eos_token_id": 151643,
9
  "hidden_act": "silu",
 
46
  "num_attention_heads": 28,
47
  "num_hidden_layers": 28,
48
  "num_key_value_heads": 4,
49
+ "pad_token_id": 151643,
50
  "rms_norm_eps": 1e-06,
51
  "rope_scaling": null,
52
  "rope_theta": 10000,
generation_config.json CHANGED
@@ -2,7 +2,10 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 151646,
4
  "do_sample": true,
5
- "eos_token_id": 151643,
 
 
 
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
  "transformers_version": "4.57.1"
 
2
  "_from_model_config": true,
3
  "bos_token_id": 151646,
4
  "do_sample": true,
5
+ "eos_token_id": [
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
  "temperature": 0.6,
10
  "top_p": 0.95,
11
  "transformers_version": "4.57.1"
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64d2c6edd65794c79413c57f7b9ecb464a193a01e21e6db35b0b8149223abc58
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f33d404b24cd4296db6593cef4a200fbf897d26091bb29a925f44f5048212822
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:547cb05578bea1cd4dcea2fbbe29d92a5af252c312b74833ff2bcafbec22c19d
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5839f0133080f959019114927e11dfc999fc8c2eb4d58fc4ca73bc257921c67e
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8b4c6b367ca75c3616aa92be096b0cca836aafc7c53c81e5615466bd0a16ae4
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be74946458476a35cf284f213926802097c0b9a7f7731d9b8785e19712c95cfb
3
  size 4330865200
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e20ddafc659ba90242154b55275402edeca0715e5dbb30f56815a4ce081f4893
3
- size 11422778
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4256422650d141f228fe954acee98679da412984c29a569877eefd3af69315a
3
+ size 11422959
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c61c93517da2b72eeb50829ab91aebcb7b8a3bc71c7559e58a3ed702f9d2d42c
3
  size 8760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce150e95afc759cfdbebc809595f0ddd7785d8982327752b175192eea4eda04e
3
  size 8760