asherisaac commited on
Commit
bbe2e5d
·
verified ·
1 Parent(s): f1b4749

Finetuner.ai deployment ID: YOT3R_GWS94

Browse files
Files changed (2) hide show
  1. adapter_config.json +32 -32
  2. adapter_model.safetensors +1 -1
adapter_config.json CHANGED
@@ -20,49 +20,49 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "model.layers.1.self_attn.v_proj",
24
- "model.layers.2.mlp.up_proj",
25
  "model.layers.1.mlp.up_proj",
26
- "model.layers.4.self_attn.q_proj",
27
- "model.layers.1.mlp.gate_proj",
28
- "model.layers.5.mlp.up_proj",
29
- "model.layers.0.self_attn.v_proj",
30
  "model.layers.1.self_attn.o_proj",
31
- "model.layers.5.self_attn.k_proj",
32
- "model.layers.2.self_attn.q_proj",
33
- "model.layers.5.mlp.down_proj",
34
- "model.layers.1.self_attn.q_proj",
35
- "model.layers.3.mlp.down_proj",
36
- "model.layers.2.self_attn.o_proj",
37
- "model.layers.3.self_attn.q_proj",
38
- "model.layers.2.self_attn.k_proj",
39
- "model.layers.0.mlp.gate_proj",
40
- "model.layers.2.self_attn.v_proj",
41
  "model.layers.0.self_attn.q_proj",
42
- "model.layers.0.self_attn.k_proj",
43
- "model.layers.0.mlp.up_proj",
44
  "model.layers.1.self_attn.k_proj",
 
45
  "model.layers.3.self_attn.k_proj",
46
- "model.layers.4.self_attn.o_proj",
47
- "model.layers.5.self_attn.q_proj",
48
  "model.layers.5.self_attn.v_proj",
49
- "model.layers.4.mlp.up_proj",
50
- "model.layers.4.self_attn.v_proj",
 
 
51
  "model.layers.3.self_attn.v_proj",
52
- "model.layers.4.mlp.down_proj",
53
- "model.layers.1.mlp.down_proj",
 
54
  "model.layers.3.mlp.up_proj",
55
- "model.layers.5.self_attn.o_proj",
 
 
 
 
 
56
  "model.layers.2.mlp.gate_proj",
57
- "model.layers.3.mlp.gate_proj",
58
  "model.layers.0.self_attn.o_proj",
 
 
 
 
 
 
 
 
59
  "model.layers.0.mlp.down_proj",
60
- "model.layers.2.mlp.down_proj",
61
- "model.layers.4.mlp.gate_proj",
62
- "model.layers.5.mlp.gate_proj",
63
- "model.layers.4.self_attn.k_proj",
64
- "lm_head",
65
- "model.layers.3.self_attn.o_proj"
66
  ],
67
  "task_type": "CAUSAL_LM",
68
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "model.layers.1.mlp.up_proj",
24
+ "model.layers.3.mlp.gate_proj",
25
+ "model.layers.4.mlp.up_proj",
26
+ "model.layers.4.self_attn.o_proj",
 
27
  "model.layers.1.self_attn.o_proj",
 
 
 
 
 
 
 
 
 
 
28
  "model.layers.0.self_attn.q_proj",
 
 
29
  "model.layers.1.self_attn.k_proj",
30
+ "model.layers.4.self_attn.k_proj",
31
  "model.layers.3.self_attn.k_proj",
32
+ "lm_head",
 
33
  "model.layers.5.self_attn.v_proj",
34
+ "model.layers.5.self_attn.k_proj",
35
+ "model.layers.0.self_attn.k_proj",
36
+ "model.layers.2.mlp.up_proj",
37
+ "model.layers.2.self_attn.q_proj",
38
  "model.layers.3.self_attn.v_proj",
39
+ "model.layers.4.mlp.gate_proj",
40
+ "model.layers.5.self_attn.q_proj",
41
+ "model.layers.5.mlp.gate_proj",
42
  "model.layers.3.mlp.up_proj",
43
+ "model.layers.2.mlp.down_proj",
44
+ "model.layers.2.self_attn.k_proj",
45
+ "model.layers.4.self_attn.q_proj",
46
+ "model.layers.0.self_attn.v_proj",
47
+ "model.layers.1.mlp.down_proj",
48
+ "model.layers.3.self_attn.q_proj",
49
  "model.layers.2.mlp.gate_proj",
 
50
  "model.layers.0.self_attn.o_proj",
51
+ "model.layers.5.self_attn.o_proj",
52
+ "model.layers.1.mlp.gate_proj",
53
+ "model.layers.2.self_attn.o_proj",
54
+ "model.layers.5.mlp.up_proj",
55
+ "model.layers.4.self_attn.v_proj",
56
+ "model.layers.2.self_attn.v_proj",
57
+ "model.layers.0.mlp.up_proj",
58
+ "model.layers.0.mlp.gate_proj",
59
  "model.layers.0.mlp.down_proj",
60
+ "model.layers.3.self_attn.o_proj",
61
+ "model.layers.1.self_attn.v_proj",
62
+ "model.layers.5.mlp.down_proj",
63
+ "model.layers.3.mlp.down_proj",
64
+ "model.layers.4.mlp.down_proj",
65
+ "model.layers.1.self_attn.q_proj"
66
  ],
67
  "task_type": "CAUSAL_LM",
68
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c8ca7c1818278409305ed2719c3a11e0ba26d9470bee1f5089102218641274c
3
  size 38959168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dcdabedbb78c575ef6a62db49de3015a94bc5a2b524ea738393363e64b95591
3
  size 38959168