NetherQuartz commited on
Commit
4511577
·
verified ·
1 Parent(s): d7283df

Try lr 6e-5 with 3 epochs

Browse files
adapter_config.json CHANGED
@@ -25,13 +25,13 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "o_proj",
29
- "down_proj",
30
  "q_proj",
 
 
31
  "v_proj",
32
  "gate_proj",
33
- "k_proj",
34
- "up_proj"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
 
28
  "q_proj",
29
+ "down_proj",
30
+ "up_proj",
31
  "v_proj",
32
  "gate_proj",
33
+ "o_proj",
34
+ "k_proj"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f1528491ae8e1396fa5aa18db741f2366fb45d40155ad7e6585c33a3aceba05
3
  size 41581360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf15bd1c329b674f846473375ff305d791f84d2e1ecdcb9c825bf98f06e19b6e
3
  size 41581360
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1f2af16ba8915b9f87cd87f2a6ed0309c0a31a3d2972f79a8875699505eb7ae
3
  size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c95bd3374bc392ec67264e2f20b7a952043f3ddc7718e141de2269505647c79
3
  size 6161