pharaouk commited on
Commit
ea6a256
·
1 Parent(s): 601ba8b

Training in progress, step 1000, checkpoint

Browse files
checkpoint-1000/adapter_config.json CHANGED
@@ -14,12 +14,12 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "up_proj",
18
  "k_proj",
 
 
19
  "down_proj",
20
  "gate_proj",
21
  "q_proj",
22
- "v_proj",
23
  "o_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
 
17
  "k_proj",
18
+ "v_proj",
19
+ "up_proj",
20
  "down_proj",
21
  "gate_proj",
22
  "q_proj",
 
23
  "o_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
checkpoint-1000/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:821e0ce5a7a01b1f815e345fe67d7a4afd3bb5d38d722f154edc75d0d4df8f85
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b669123c077c6a76eede1c4f7711ab7ad0cb92e65ef1ad7bdd14310d018bd11
3
  size 335706314
checkpoint-1000/adapter_model/adapter_model/adapter_config.json CHANGED
@@ -14,12 +14,12 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "up_proj",
18
  "k_proj",
 
 
19
  "down_proj",
20
  "gate_proj",
21
  "q_proj",
22
- "v_proj",
23
  "o_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
 
17
  "k_proj",
18
+ "v_proj",
19
+ "up_proj",
20
  "down_proj",
21
  "gate_proj",
22
  "q_proj",
 
23
  "o_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
checkpoint-1000/adapter_model/adapter_model/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:821e0ce5a7a01b1f815e345fe67d7a4afd3bb5d38d722f154edc75d0d4df8f85
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b669123c077c6a76eede1c4f7711ab7ad0cb92e65ef1ad7bdd14310d018bd11
3
  size 335706314
checkpoint-1000/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9d245d8e896372a036d47d5685d8b11f119e6753280bb6d7d7564e8e105fe0a
3
  size 1342453434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7f0475433f4b5810b8bf8bfc1642c09ba6e2077e57e6688c0acc473fd12e4dd
3
  size 1342453434
checkpoint-1000/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4888bd21cd14896443d95549de20b3d15cb9c5edc042017cf37ce382033409e5
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1642e6cad7fffd8d25723a2897b94fcaab5a887057516d48412986a1f4f600
3
  size 14180
checkpoint-1000/trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33eae242384e626b4f6ecff3a308c1a811f1b81e477cc7196ec4e5dd7f41f001
3
- size 6392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd5bb9a2c78d80517f60f05f862b74c6d4f0f6baad6cc6e2b5b20b44363d6c05
3
+ size 6456