lfhe commited on
Commit
3e3b045
·
1 Parent(s): 3236567
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Qwen/Qwen1.5-7B-Chat
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: Qwen/Qwen2-7B-Instruct
3
  library_name: peft
4
  ---
5
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen1.5-7B-Chat",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -13,30 +13,27 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 64,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
- "modules_to_save": [
22
- "embed_tokens",
23
- "lm_head"
24
- ],
25
  "peft_type": "LORA",
26
- "r": 8,
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "down_proj",
31
- "up_proj",
32
  "o_proj",
33
  "k_proj",
 
34
  "gate_proj",
 
35
  "q_proj",
36
- "v_proj"
37
  ],
38
  "task_type": "CAUSAL_LM",
39
  "trainable_token_indices": null,
40
- "use_dora": false,
41
- "use_rslora": true
42
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2-7B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 96,
17
  "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
 
 
 
22
  "peft_type": "LORA",
23
+ "r": 48,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
27
  "o_proj",
28
  "k_proj",
29
+ "v_proj",
30
  "gate_proj",
31
+ "down_proj",
32
  "q_proj",
33
+ "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
37
+ "use_dora": true,
38
+ "use_rslora": false
39
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:241ad965acba9b36aef3bbf9dc4897ad29d4c148ad5b89bd5da32009c1829c45
3
- size 5058654112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28d93242ca2bdd155a5a07f13fd33548ccfd9e8178b045302b3237344fc85925
3
+ size 490084704
added_tokens.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "<|endoftext|>": 151643,
3
- "<|end|>": 151646,
4
  "<|im_end|>": 151645,
5
  "<|im_start|>": 151644
6
  }
 
1
  {
2
  "<|endoftext|>": 151643,
 
3
  "<|im_end|>": 151645,
4
  "<|im_start|>": 151644
5
  }
chat_template.jinja DELETED
@@ -1,6 +0,0 @@
1
- {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
2
- You are a helpful assistant.<|im_end|>
3
- ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
4
- ' + message['content'] + '<|im_end|>' + '
5
- '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
6
- ' }}{% endif %}
 
 
 
 
 
 
 
special_tokens_map.json CHANGED
@@ -4,7 +4,7 @@
4
  "<|im_end|>"
5
  ],
6
  "eos_token": {
7
- "content": "<|end|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
 
4
  "<|im_end|>"
5
  ],
6
  "eos_token": {
7
+ "content": "<|im_end|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:803259e51eee062d0ffb48046ad7f8420e415a25028e85a56c06a9cd29416999
3
- size 11418450
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
tokenizer_config.json CHANGED
@@ -24,14 +24,6 @@
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
- },
28
- "151646": {
29
- "content": "<|end|>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
  }
36
  },
37
  "additional_special_tokens": [
@@ -39,11 +31,12 @@
39
  "<|im_end|>"
40
  ],
41
  "bos_token": null,
 
42
  "clean_up_tokenization_spaces": false,
43
- "eos_token": "<|end|>",
44
  "errors": "replace",
45
  "extra_special_tokens": {},
46
- "model_max_length": 32768,
47
  "pad_token": "<|endoftext|>",
48
  "padding_side": "right",
49
  "split_special_tokens": false,
 
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
 
 
 
 
 
 
 
 
27
  }
28
  },
29
  "additional_special_tokens": [
 
31
  "<|im_end|>"
32
  ],
33
  "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
  "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
  "errors": "replace",
38
  "extra_special_tokens": {},
39
+ "model_max_length": 131072,
40
  "pad_token": "<|endoftext|>",
41
  "padding_side": "right",
42
  "split_special_tokens": false,