tomal66 commited on
Commit
bbec7cd
·
verified ·
1 Parent(s): ec8116f

[S2] Upload LoRA adapters

Browse files
Files changed (2) hide show
  1. adapter_config.json +3 -3
  2. adapter_model.safetensors +2 -2
adapter_config.json CHANGED
@@ -21,13 +21,13 @@
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
  "qalora_group_size": 16,
24
- "r": 4,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "c_proj",
29
- "c_attn",
30
- "c_fc"
31
  ],
32
  "target_parameters": null,
33
  "task_type": "CAUSAL_LM",
 
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
  "qalora_group_size": 16,
24
+ "r": 8,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "c_fc",
29
  "c_proj",
30
+ "c_attn"
 
31
  ],
32
  "target_parameters": null,
33
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4a986f25892a4970e1c775c255ee207d44b8486a8bdc190acf5ab0e770b2e7c
3
- size 2371296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df790686f9f2371c173e964ee0ff9107990227efb8c71ad02172b34e2841d8ea
3
+ size 4730632