gradientdegen commited on
Commit
b3152e4
·
verified ·
1 Parent(s): bdc3c64

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. adapter_config.json +5 -5
  2. adapter_model.safetensors +1 -1
adapter_config.json CHANGED
@@ -3,13 +3,13 @@
3
  "peft_type": "LORA",
4
  "use_rslora": false,
5
  "target_modules": [
6
- "up_proj",
7
- "gate_proj",
8
- "k_proj",
9
  "v_proj",
10
  "down_proj",
11
- "q_proj",
12
- "o_proj"
 
 
 
13
  ],
14
  "task_type": "CAUSAL_LM",
15
  "r": 128,
 
3
  "peft_type": "LORA",
4
  "use_rslora": false,
5
  "target_modules": [
 
 
 
6
  "v_proj",
7
  "down_proj",
8
+ "k_proj",
9
+ "gate_proj",
10
+ "o_proj",
11
+ "up_proj",
12
+ "q_proj"
13
  ],
14
  "task_type": "CAUSAL_LM",
15
  "r": 128,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad7c0eec04535a1db7f1de2774a471a65646201487aeab3b55f1d8f8d8446ce2
3
  size 1736241568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba74813d8d062726299d9ebc1aba007d83cc0b55498ad20eb48047fb56b2a124
3
  size 1736241568