mgp77 commited on
Commit
c2f04ec
·
verified ·
1 Parent(s): 039f063

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +2 -2
  2. adapter_config.json +4 -4
  3. adapter_model.safetensors +2 -2
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
- base_model: Qwen/Qwen2.5-1.5B-Instruct
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
- - base_model:adapter:Qwen/Qwen2.5-1.5B-Instruct
7
  - lora
8
  - transformers
9
  ---
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-3B-Instruct
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
+ - base_model:adapter:Qwen/Qwen2.5-3B-Instruct
7
  - lora
8
  - transformers
9
  ---
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -26,12 +26,12 @@
26
  "revision": null,
27
  "target_modules": [
28
  "q_proj",
29
- "gate_proj",
30
  "k_proj",
31
- "down_proj",
32
  "o_proj",
33
  "v_proj",
34
- "up_proj"
 
 
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-3B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
26
  "revision": null,
27
  "target_modules": [
28
  "q_proj",
 
29
  "k_proj",
 
30
  "o_proj",
31
  "v_proj",
32
+ "down_proj",
33
+ "up_proj",
34
+ "gate_proj"
35
  ],
36
  "task_type": "CAUSAL_LM",
37
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b29b9d8cfbce09b204803ce24f3c9b46b9d381a28f58a86e9a39792e8fab7761
3
- size 36981072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeedb6d485f19c20d375b062f06900a22c64b1263f70ccf546371bbeefa5ab46
3
+ size 59933632