mizzaay commited on
Commit
e6aff7f
·
verified ·
1 Parent(s): 6c69c0d

Run 6. Outer Step 43. Inner Step 4.

Browse files
Files changed (4) hide show
  1. config.json +16 -16
  2. gradients.pt +1 -1
  3. inner_optimizer.pt +1 -1
  4. model.safetensors +1 -1
config.json CHANGED
@@ -76,7 +76,7 @@
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
- "166": "NOT_ALIVE",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
@@ -104,7 +104,7 @@
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
- "191": "NOT_ALIVE",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
@@ -113,7 +113,7 @@
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
- "2": "SUCCESS",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
@@ -177,22 +177,22 @@
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
- "27": "NON_PARTICIPATING",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
- "31": "SUCCESS",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
- "35": "SUCCESS",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
- "40": "NOT_ALIVE",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
@@ -209,24 +209,24 @@
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
- "56": "NON_PARTICIPATING",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
- "60": "NON_PARTICIPATING",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
- "65": "NOT_ALIVE",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
- "71": "NON_PARTICIPATING",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
@@ -253,7 +253,7 @@
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
- "96": "NOT_ALIVE",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
@@ -264,17 +264,17 @@
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
- 6157525,
268
- 6157542
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
- "inner_step": 76,
276
  "intermediate_size": 5632,
277
- "last_allreduce_block": 6156907,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
 
76
  "163": "NON_PARTICIPATING",
77
  "164": "NOT_ALIVE",
78
  "165": "NON_PARTICIPATING",
79
+ "166": "SUCCESS",
80
  "167": "NOT_ALIVE",
81
  "168": "NOT_ALIVE",
82
  "169": "NOT_ALIVE",
 
104
  "189": "NOT_ALIVE",
105
  "19": "NOT_ALIVE",
106
  "190": "SUCCESS",
107
+ "191": "NON_PARTICIPATING",
108
  "192": "NOT_ALIVE",
109
  "193": "SUCCESS",
110
  "194": "NOT_ALIVE",
 
113
  "197": "NOT_ALIVE",
114
  "198": "NOT_ALIVE",
115
  "199": "SUCCESS",
116
+ "2": "NOT_ALIVE",
117
  "20": "NOT_ALIVE",
118
  "200": "NOT_ALIVE",
119
  "201": "SUCCESS",
 
177
  "254": "NOT_ALIVE",
178
  "255": "NOT_ALIVE",
179
  "26": "NON_PARTICIPATING",
180
+ "27": "NOT_ALIVE",
181
  "28": "NON_PARTICIPATING",
182
  "29": "NOT_ALIVE",
183
  "3": "SUCCESS",
184
  "30": "NOT_ALIVE",
185
+ "31": "NOT_ALIVE",
186
  "32": "SUCCESS",
187
  "33": "NOT_ALIVE",
188
  "34": "NOT_ALIVE",
189
+ "35": "NOT_ALIVE",
190
  "36": "NOT_ALIVE",
191
  "37": "NOT_ALIVE",
192
  "38": "NOT_ALIVE",
193
  "39": "NOT_ALIVE",
194
  "4": "NOT_ALIVE",
195
+ "40": "SUCCESS",
196
  "41": "NOT_ALIVE",
197
  "42": "NOT_ALIVE",
198
  "43": "NOT_ALIVE",
 
209
  "53": "SUCCESS",
210
  "54": "NOT_ALIVE",
211
  "55": "NOT_ALIVE",
212
+ "56": "NOT_ALIVE",
213
  "57": "NON_PARTICIPATING",
214
  "58": "NOT_ALIVE",
215
  "59": "NOT_ALIVE",
216
  "6": "NOT_ALIVE",
217
+ "60": "NOT_ALIVE",
218
  "61": "NOT_ALIVE",
219
  "62": "NOT_ALIVE",
220
  "63": "NOT_ALIVE",
221
  "64": "NOT_ALIVE",
222
+ "65": "NON_PARTICIPATING",
223
  "66": "NOT_ALIVE",
224
  "67": "NOT_ALIVE",
225
  "68": "NOT_ALIVE",
226
  "69": "NOT_ALIVE",
227
  "7": "NOT_ALIVE",
228
  "70": "SUCCESS",
229
+ "71": "FAIL",
230
  "72": "NOT_ALIVE",
231
  "73": "NOT_ALIVE",
232
  "74": "NOT_ALIVE",
 
253
  "93": "NOT_ALIVE",
254
  "94": "NOT_ALIVE",
255
  "95": "NON_PARTICIPATING",
256
+ "96": "NON_PARTICIPATING",
257
  "97": "NOT_ALIVE",
258
  "98": "NOT_ALIVE",
259
  "99": "SUCCESS"
 
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
+ 6157626,
268
+ 6157644
269
  ],
270
  "bos_token_id": 1,
271
  "eos_token_id": 2,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
+ "inner_step": 4,
276
  "intermediate_size": 5632,
277
+ "last_allreduce_block": 6157557,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
gradients.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4963cbccff8c5493eec4413f16d5be56f17aa2af43aafff20682e9b9e981268c
3
  size 88389626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:332e2900d104ede530ddff3993f90cb1d0428d43da489607fa3d3c7702177d58
3
  size 88389626
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3f15727d922e2bfb5ec2f46e8c88a2fead97c42f472c95c9ad11c50f8396982
3
  size 8800564076
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc4125ee66076d1e9b6200433724dfb95d4c4fa2a1dcbc0459a5d81731e9997b
3
  size 8800564076
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21d3f84ea3e8910f022d780731633a4b7a11caa5b44d7e7f20c82ce313e92cca
3
  size 4400216536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c6c4c5b50ed603c09ad1c70f7150eaf17068beec447cef9592cdf553974376e
3
  size 4400216536