mizzaay commited on
Commit
3cd7146
·
verified ·
1 Parent(s): 201f1ec

Run 6. Outer Step 43. Inner Step 5.

Browse files
Files changed (4) hide show
  1. config.json +16 -16
  2. gradients.pt +1 -1
  3. inner_optimizer.pt +1 -1
  4. model.safetensors +1 -1
config.json CHANGED
@@ -76,7 +76,7 @@
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
- "166": "NOT_ALIVE",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
@@ -104,7 +104,7 @@
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
- "191": "NOT_ALIVE",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
@@ -113,7 +113,7 @@
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
- "2": "SUCCESS",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
@@ -177,22 +177,22 @@
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
- "27": "NON_PARTICIPATING",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
- "31": "SUCCESS",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
- "35": "SUCCESS",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
- "40": "NOT_ALIVE",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
@@ -209,24 +209,24 @@
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
- "56": "NON_PARTICIPATING",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
- "60": "NON_PARTICIPATING",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
- "65": "NOT_ALIVE",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
- "71": "NON_PARTICIPATING",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
@@ -253,7 +253,7 @@
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
- "96": "NOT_ALIVE",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
@@ -264,17 +264,17 @@
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
- 6157543,
268
- 6157560
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
- "inner_step": 76,
276
  "intermediate_size": 5632,
277
- "last_allreduce_block": 6156907,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
 
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
+ "166": "SUCCESS",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
 
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
+ "191": "NON_PARTICIPATING",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
 
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
+ "2": "NOT_ALIVE",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
 
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
+ "27": "NOT_ALIVE",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
+ "31": "NOT_ALIVE",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
+ "35": "NOT_ALIVE",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
+ "40": "SUCCESS",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
 
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
+ "56": "NOT_ALIVE",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
+ "60": "NOT_ALIVE",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
+ "65": "NON_PARTICIPATING",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
+ "71": "FAIL",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
 
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
+ "96": "NON_PARTICIPATING",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
 
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
+ 6157626,
268
+ 6157645
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
+ "inner_step": 4,
276
  "intermediate_size": 5632,
277
+ "last_allreduce_block": 6157557,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
gradients.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29cde1b921f0191f4957c5a01d5ae7fb6377266feeffa5d2d07d14f56c2ecca2
3
  size 88389626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1142d1b09426a6d9ef0f5eefc464f84ee36e2308e469a3776652183503fe03c
3
  size 88389626
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c13424b4c440dee5ac6ead0518ea40fedecd61368d002091e24ccc37cb3da3e
3
  size 8800564076
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40ee7147079520808e069835b742fc95cd14ff2cd23e6e026d639547940c08f6
3
  size 8800564076
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a63d6b8122b7d6b26d04c1bb703b90d4219e6e7ade99c7a6a64dfd081b74548
3
  size 4400216536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d5640c0d18836414defe5b0e5931f8bd79af9517181100b41e52b6d05763504
3
  size 4400216536