lfhe commited on
Commit
403f8c2
·
verified ·
1 Parent(s): 06d9b52
README.md CHANGED
@@ -1,6 +1,11 @@
1
  ---
2
  base_model: microsoft/Phi-4-mini-instruct
3
  library_name: peft
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
@@ -199,4 +204,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
- - PEFT 0.15.0
 
1
  ---
2
  base_model: microsoft/Phi-4-mini-instruct
3
  library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:microsoft/Phi-4-mini-instruct
7
+ - lora
8
+ - transformers
9
  ---
10
 
11
  # Model Card for Model ID
 
204
  [More Information Needed]
205
  ### Framework versions
206
 
207
+ - PEFT 0.17.1
adapter_config.json CHANGED
@@ -13,24 +13,27 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 64,
17
  "lora_bias": false,
18
- "lora_dropout": 0.15,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "r": 64,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "down_proj",
28
- "o_proj",
29
  "gate_up_proj",
30
- "qkv_proj"
 
 
31
  ],
 
32
  "task_type": "CAUSAL_LM",
33
  "trainable_token_indices": null,
34
  "use_dora": false,
35
- "use_rslora": true
 
36
  }
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 96,
17
  "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 96,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
 
28
  "gate_up_proj",
29
+ "qkv_proj",
30
+ "down_proj",
31
+ "o_proj"
32
  ],
33
+ "target_parameters": null,
34
  "task_type": "CAUSAL_LM",
35
  "trainable_token_indices": null,
36
  "use_dora": false,
37
+ "use_qalora": false,
38
+ "use_rslora": false
39
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68d045fb03654a8204309dd246669c066b2656f45943c27079a36b94ac59f6f6
3
- size 369133568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1632f20a104b454c5896af4f6239ac6b4d5e9fd0446cbce3e9bd0f4e533bea58
3
+ size 553683000
added_tokens.json CHANGED
@@ -3,6 +3,7 @@
3
  "<|/tool|>": 200024,
4
  "<|assistant|>": 200019,
5
  "<|end|>": 200020,
 
6
  "<|system|>": 200022,
7
  "<|tag|>": 200028,
8
  "<|tool_call|>": 200025,
 
3
  "<|/tool|>": 200024,
4
  "<|assistant|>": 200019,
5
  "<|end|>": 200020,
6
+ "<|im_end|>": 200029,
7
  "<|system|>": 200022,
8
  "<|tag|>": 200028,
9
  "<|tool_call|>": 200025,
special_tokens_map.json CHANGED
@@ -7,7 +7,7 @@
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "<|endoftext|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
 
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|im_end|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:382cc235b56c725945e149cc25f191da667c836655efd0857b004320e90e91ea
3
- size 15524095
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8799b7ce62bf40e724f80af87b9545f9d9043778f435040a6c08e834dc047b69
3
+ size 15524282
tokenizer_config.json CHANGED
@@ -98,15 +98,25 @@
98
  "rstrip": true,
99
  "single_word": false,
100
  "special": true
 
 
 
 
 
 
 
 
101
  }
102
  },
103
  "bos_token": "<|endoftext|>",
104
  "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and 'tools' in message and message['tools'] is not none %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|tool|>' + message['tools'] + '<|/tool|>' + '<|end|>' }}{% else %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|end|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>' }}{% else %}{{ eos_token }}{% endif %}",
105
  "clean_up_tokenization_spaces": false,
106
- "eos_token": "<|endoftext|>",
107
  "extra_special_tokens": {},
108
  "model_max_length": 131072,
109
  "pad_token": "<|endoftext|>",
 
 
110
  "tokenizer_class": "GPT2Tokenizer",
111
  "unk_token": "<|endoftext|>"
112
  }
 
98
  "rstrip": true,
99
  "single_word": false,
100
  "special": true
101
+ },
102
+ "200029": {
103
+ "content": "<|im_end|>",
104
+ "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": false,
107
+ "single_word": false,
108
+ "special": true
109
  }
110
  },
111
  "bos_token": "<|endoftext|>",
112
  "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and 'tools' in message and message['tools'] is not none %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|tool|>' + message['tools'] + '<|/tool|>' + '<|end|>' }}{% else %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|end|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>' }}{% else %}{{ eos_token }}{% endif %}",
113
  "clean_up_tokenization_spaces": false,
114
+ "eos_token": "<|im_end|>",
115
  "extra_special_tokens": {},
116
  "model_max_length": 131072,
117
  "pad_token": "<|endoftext|>",
118
+ "padding_side": "right",
119
+ "split_special_tokens": false,
120
  "tokenizer_class": "GPT2Tokenizer",
121
  "unk_token": "<|endoftext|>"
122
  }