jkjjbk commited on
Commit
81d766b
·
verified ·
1 Parent(s): 167a2fd

Run 6. Outer Step 43. Inner Step 5.

Browse files
Files changed (4) hide show
  1. config.json +16 -16
  2. gradients.pt +1 -1
  3. inner_optimizer.pt +1 -1
  4. model.safetensors +1 -1
config.json CHANGED
@@ -76,7 +76,7 @@
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
- "166": "NOT_ALIVE",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
@@ -104,7 +104,7 @@
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
- "191": "NOT_ALIVE",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
@@ -113,7 +113,7 @@
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
- "2": "SUCCESS",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
@@ -177,22 +177,22 @@
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
- "27": "NON_PARTICIPATING",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
- "31": "SUCCESS",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
- "35": "SUCCESS",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
- "40": "NOT_ALIVE",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
@@ -209,24 +209,24 @@
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
- "56": "NON_PARTICIPATING",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
- "60": "NON_PARTICIPATING",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
- "65": "NOT_ALIVE",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
- "71": "NON_PARTICIPATING",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
@@ -253,7 +253,7 @@
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
- "96": "NOT_ALIVE",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
@@ -264,17 +264,17 @@
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
- 6157509,
268
- 6157525
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
- "inner_step": 84,
276
  "intermediate_size": 5632,
277
- "last_allreduce_block": 6156907,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
 
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
+ "166": "SUCCESS",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
 
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
+ "191": "NON_PARTICIPATING",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
 
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
+ "2": "NOT_ALIVE",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
 
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
+ "27": "NOT_ALIVE",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
+ "31": "NOT_ALIVE",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
+ "35": "NOT_ALIVE",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
+ "40": "SUCCESS",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
 
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
+ "56": "NOT_ALIVE",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
+ "60": "NOT_ALIVE",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
+ "65": "NON_PARTICIPATING",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
+ "71": "FAIL",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
 
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
+ "96": "NON_PARTICIPATING",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
 
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
+ 6157627,
268
+ 6157641
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
+ "inner_step": 4,
276
  "intermediate_size": 5632,
277
+ "last_allreduce_block": 6157557,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
gradients.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9e052ccd335251f2a477bbd6cc8064d237621b7e2092effabeeae38193f2406
3
  size 88389626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d9121f3aed653eb9202cb7c38c126154914a64cff533aecbdcf55e17a64cd37
3
  size 88389626
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9734e012ddb5a7367d6f53c3170657591d928aaa2b3564976ff9922b995a8d18
3
  size 8800564076
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae0a8c545afd93f9ec25908b806e71fab66dcdf23eea42b1d6df18aa8e4e3ce1
3
  size 8800564076
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8aabb99cb2671f39a4a7ac7937e74643370dcdd8a3c6047fb9341201cded3bd
3
  size 4400216536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7b24c0f0dcc647f10d9ec004f12249fb83456cac9c35d3eb4fccdc3bdcef27d
3
  size 4400216536