lfhe commited on
Commit
cc43c8c
·
verified ·
1 Parent(s): b293d93
README.md CHANGED
@@ -1,11 +1,6 @@
1
  ---
2
  base_model: microsoft/Phi-4-mini-instruct
3
  library_name: peft
4
- pipeline_tag: text-generation
5
- tags:
6
- - base_model:adapter:microsoft/Phi-4-mini-instruct
7
- - lora
8
- - transformers
9
  ---
10
 
11
  # Model Card for Model ID
@@ -204,4 +199,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
204
  [More Information Needed]
205
  ### Framework versions
206
 
207
- - PEFT 0.17.1
 
1
  ---
2
  base_model: microsoft/Phi-4-mini-instruct
3
  library_name: peft
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
 
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
+ - PEFT 0.15.0
adapter_config.json CHANGED
@@ -15,25 +15,22 @@
15
  "loftq_config": {},
16
  "lora_alpha": 96,
17
  "lora_bias": false,
18
- "lora_dropout": 0.0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "qalora_group_size": 16,
24
- "r": 96,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "gate_up_proj",
29
- "qkv_proj",
30
  "down_proj",
31
- "o_proj"
32
  ],
33
- "target_parameters": null,
34
  "task_type": "CAUSAL_LM",
35
  "trainable_token_indices": null,
36
  "use_dora": false,
37
- "use_qalora": false,
38
- "use_rslora": false
39
  }
 
15
  "loftq_config": {},
16
  "lora_alpha": 96,
17
  "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
+ "r": 64,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "o_proj",
28
  "gate_up_proj",
 
29
  "down_proj",
30
+ "qkv_proj"
31
  ],
 
32
  "task_type": "CAUSAL_LM",
33
  "trainable_token_indices": null,
34
  "use_dora": false,
35
+ "use_rslora": true
 
36
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65260ebca6f70c9996e6bb9839e4748828e37c9fb6bb101009b8e851ff5fd948
3
- size 553683000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71789a667f3e6e410dc27af9e562a71024345eadf3739cccacd6e917e2320296
3
+ size 369133600
tokenizer_config.json CHANGED
@@ -109,7 +109,6 @@
109
  }
110
  },
111
  "bos_token": "<|endoftext|>",
112
- "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and 'tools' in message and message['tools'] is not none %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|tool|>' + message['tools'] + '<|/tool|>' + '<|end|>' }}{% else %}{{ '<|' + message['role'] + '|>' + message['content'] + '<|end|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>' }}{% else %}{{ eos_token }}{% endif %}",
113
  "clean_up_tokenization_spaces": false,
114
  "eos_token": "<|im_end|>",
115
  "extra_special_tokens": {},
 
109
  }
110
  },
111
  "bos_token": "<|endoftext|>",
 
112
  "clean_up_tokenization_spaces": false,
113
  "eos_token": "<|im_end|>",
114
  "extra_special_tokens": {},