pharaouk commited on
Commit
a32fa30
·
1 Parent(s): 0f14381

Training in progress, step 1400, checkpoint

Browse files
checkpoint-1400/adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
 
 
 
 
17
  "k_proj",
18
  "down_proj",
19
- "v_proj",
20
- "q_proj",
21
- "o_proj",
22
- "up_proj",
23
- "gate_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "gate_proj",
18
+ "up_proj",
19
+ "q_proj",
20
+ "v_proj",
21
  "k_proj",
22
  "down_proj",
23
+ "o_proj"
 
 
 
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
checkpoint-1400/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d4dd2812760687c9a4c37040d019e420ed7995a01a7041654a86bd07ba0ded9
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fead452006cb1b165a89af4cdae2c451523c83d45f82db12568eb5818686869a
3
  size 335706314
checkpoint-1400/adapter_model/adapter_model/adapter_config.json CHANGED
@@ -14,13 +14,13 @@
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
 
 
 
 
17
  "k_proj",
18
  "down_proj",
19
- "v_proj",
20
- "q_proj",
21
- "o_proj",
22
- "up_proj",
23
- "gate_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
14
  "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
+ "gate_proj",
18
+ "up_proj",
19
+ "q_proj",
20
+ "v_proj",
21
  "k_proj",
22
  "down_proj",
23
+ "o_proj"
 
 
 
 
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
checkpoint-1400/adapter_model/adapter_model/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d4dd2812760687c9a4c37040d019e420ed7995a01a7041654a86bd07ba0ded9
3
  size 335706314
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fead452006cb1b165a89af4cdae2c451523c83d45f82db12568eb5818686869a
3
  size 335706314
checkpoint-1400/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f3a13648fb3e96572d0a1c3f6fb9b6872437e16f8e10d85378dbb8d36352842
3
  size 1342453434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:744e86c95a87117ea6cdaf56141f69cd3f6496aaa2a6c9490f804e96ad7e93b0
3
  size 1342453434
checkpoint-1400/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a30ab784c90109db1c644d705b02ba26960912fc2ecab8da4c4875c4df7fb2af
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c2f8a3b0cc11f90dc012dbd7939c7055958ca37c42d77b214a449bba38a3fac
3
  size 14180
checkpoint-1400/trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
checkpoint-1400/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5ab8ef2a53a353c81e0e27307b4f9b37e7b4be919328eb7405b3b561276e06a
3
- size 6456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8584a03439a91de296fc372191c445a686a0ab74c08cca7d76e9ffa983c30d18
3
+ size 6392