pharaouk commited on
Commit
da78ca6
·
1 Parent(s): 33bb7bc

Training in progress, step 1200, checkpoint

Browse files
checkpoint-1200/adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "q_proj",
18
- "v_proj",
19
  "down_proj",
20
- "k_proj",
21
- "o_proj",
22
  "up_proj",
23
- "gate_proj"
 
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "gate_proj",
 
18
  "down_proj",
19
+ "v_proj",
 
20
  "up_proj",
21
+ "q_proj",
22
+ "o_proj",
23
+ "k_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
checkpoint-1200/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bb002fec1a007c80ceac4f816d203920e40ef81c39151b5feb1050c30e9b2e5
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:398ca32bc3837fddb1912c0c9e68f383c57718a22269189b0052f168502a57da
3
  size 335706314
checkpoint-1200/adapter_model/adapter_model/adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
- "q_proj",
18
- "v_proj",
19
  "down_proj",
20
- "k_proj",
21
- "o_proj",
22
  "up_proj",
23
- "gate_proj"
 
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "gate_proj",
 
18
  "down_proj",
19
+ "v_proj",
 
20
  "up_proj",
21
+ "q_proj",
22
+ "o_proj",
23
+ "k_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
checkpoint-1200/adapter_model/adapter_model/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bb002fec1a007c80ceac4f816d203920e40ef81c39151b5feb1050c30e9b2e5
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:398ca32bc3837fddb1912c0c9e68f383c57718a22269189b0052f168502a57da
3
  size 335706314
checkpoint-1200/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46b56c8b3d7d311c4542a639314d6a53b2f069c9038689cb069d9ed90a67764d
3
  size 1342453434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de1d073aebe90338d8bb858f744260d16dbc94c2493ff6299188b47ba2195c41
3
  size 1342453434
checkpoint-1200/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b94dc38ca2a367b87cc379508d1eedcf4b1ec5642c67e1bb1a20f7e59db55976
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d301983a035e46cacf9969a288eb2f7691a3df7ba3baab4782369e331af31fb8
3
  size 14180
checkpoint-1200/trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
checkpoint-1200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfdb50f52090ac3e9996d411c5880a744db662f16e3b201fb304c8bcc37e571e
3
- size 6456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cb003a7b99948d23dd98c2fac241488db19f1ea276420dd1c11b73a24708a8f
3
+ size 6392