error577 commited on
Commit
d50e6d9
·
verified ·
1 Parent(s): a542e28

Training in progress, step 400

Browse files
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "fc2",
25
- "q_proj",
26
  "fc1",
27
  "dense",
28
- "v_proj"
 
 
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "fc1",
24
  "dense",
25
+ "fc2",
26
+ "v_proj",
27
+ "q_proj",
28
+ "k_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:376b9dd19dc0ab1fcd39adfa230f0743aca1b30a4242b6b18f0bcf0c85a77b28
3
  size 113284112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba83219601ab88cc71f0be490e4a2b89cc0531c9d17474fed2f2d9dc0e2dff57
3
  size 113284112
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7b2b003dd5212b13ff20f39e64cf314b55dc58061d234f47b1d0a005744c530
3
- size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62dc5607f4b311b629bd491f1e8feae0d5a56fbdbaa42feb5d077b975d21ab3c
3
+ size 6712