gprashant17 commited on
Commit
f2cb05b
·
verified ·
1 Parent(s): d753265

Training in progress, step 100

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_attn",
24
  "c_proj",
25
  "c_attn",
26
- "c_fc"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "c_fc",
24
  "c_proj",
25
  "c_attn",
26
+ "q_attn"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5adc457169018f1d8c6847898ab37910e20c239b28253767fd1752777a732b7f
3
  size 29506408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3808496b69e81916845c243b3f631affb4c599b9db5e96990cc003749fe5c240
3
  size 29506408
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:685e76a3aa4ee8e4031db09190d776063582d4557ef9e0a453d1edc046d89208
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55be39195d43417c48c79aab054c0dc2ede062ead72e3d89339a36ef89567b7d
3
  size 5112