crypt0trading commited on
Commit
960fe77
·
verified ·
1 Parent(s): 03a1d71

Run 3. Outer Step 8. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +20 -20
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "crypt0trading/c66-h4",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
@@ -115,7 +115,7 @@
115
  "198": "NON_PARTICIPATING",
116
  "199": "NON_PARTICIPATING",
117
  "2": "SUCCESS",
118
- "20": "SUCCESS",
119
  "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
@@ -137,7 +137,7 @@
137
  "217": "NON_PARTICIPATING",
138
  "218": "NON_PARTICIPATING",
139
  "219": "NON_PARTICIPATING",
140
- "22": "FAIL",
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
@@ -189,14 +189,14 @@
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
  "36": "NON_PARTICIPATING",
192
- "37": "SUCCESS",
193
  "38": "NON_PARTICIPATING",
194
- "39": "SUCCESS",
195
  "4": "NON_PARTICIPATING",
196
  "40": "NON_PARTICIPATING",
197
- "41": "NON_PARTICIPATING",
198
  "42": "NON_PARTICIPATING",
199
- "43": "FAIL",
200
  "44": "SUCCESS",
201
  "45": "NON_PARTICIPATING",
202
  "46": "NON_PARTICIPATING",
@@ -217,17 +217,17 @@
217
  "6": "NON_PARTICIPATING",
218
  "60": "NON_PARTICIPATING",
219
  "61": "NON_PARTICIPATING",
220
- "62": "SUCCESS",
221
  "63": "NON_PARTICIPATING",
222
  "64": "NON_PARTICIPATING",
223
  "65": "NON_PARTICIPATING",
224
  "66": "NON_PARTICIPATING",
225
- "67": "FAIL",
226
  "68": "NON_PARTICIPATING",
227
  "69": "NON_PARTICIPATING",
228
  "7": "NON_PARTICIPATING",
229
  "70": "NON_PARTICIPATING",
230
- "71": "SUCCESS",
231
  "72": "NON_PARTICIPATING",
232
  "73": "NON_PARTICIPATING",
233
  "74": "NON_PARTICIPATING",
@@ -237,11 +237,11 @@
237
  "78": "NON_PARTICIPATING",
238
  "79": "NON_PARTICIPATING",
239
  "8": "NON_PARTICIPATING",
240
- "80": "FAIL",
241
  "81": "NON_PARTICIPATING",
242
  "82": "NON_PARTICIPATING",
243
  "83": "NON_PARTICIPATING",
244
- "84": "SUCCESS",
245
  "85": "NON_PARTICIPATING",
246
  "86": "NON_PARTICIPATING",
247
  "87": "NON_PARTICIPATING",
@@ -256,7 +256,7 @@
256
  "95": "NON_PARTICIPATING",
257
  "96": "SUCCESS",
258
  "97": "NON_PARTICIPATING",
259
- "98": "FAIL",
260
  "99": "NON_PARTICIPATING"
261
  },
262
  "architectures": [
@@ -268,20 +268,20 @@
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
- 5395522,
272
- 5395526,
273
- 5395530,
274
- 5395534,
275
- 5395538
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 79,
283
  "inner_steps": 0,
284
- "last_allreduce_block": 5392742,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
 
115
  "198": "NON_PARTICIPATING",
116
  "199": "NON_PARTICIPATING",
117
  "2": "SUCCESS",
118
+ "20": "NON_PARTICIPATING",
119
  "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
 
137
  "217": "NON_PARTICIPATING",
138
  "218": "NON_PARTICIPATING",
139
  "219": "NON_PARTICIPATING",
140
+ "22": "NON_PARTICIPATING",
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
 
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
  "36": "NON_PARTICIPATING",
192
+ "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
194
+ "39": "NON_PARTICIPATING",
195
  "4": "NON_PARTICIPATING",
196
  "40": "NON_PARTICIPATING",
197
+ "41": "SUCCESS",
198
  "42": "NON_PARTICIPATING",
199
+ "43": "SUCCESS",
200
  "44": "SUCCESS",
201
  "45": "NON_PARTICIPATING",
202
  "46": "NON_PARTICIPATING",
 
217
  "6": "NON_PARTICIPATING",
218
  "60": "NON_PARTICIPATING",
219
  "61": "NON_PARTICIPATING",
220
+ "62": "NON_PARTICIPATING",
221
  "63": "NON_PARTICIPATING",
222
  "64": "NON_PARTICIPATING",
223
  "65": "NON_PARTICIPATING",
224
  "66": "NON_PARTICIPATING",
225
+ "67": "SUCCESS",
226
  "68": "NON_PARTICIPATING",
227
  "69": "NON_PARTICIPATING",
228
  "7": "NON_PARTICIPATING",
229
  "70": "NON_PARTICIPATING",
230
+ "71": "NON_PARTICIPATING",
231
  "72": "NON_PARTICIPATING",
232
  "73": "NON_PARTICIPATING",
233
  "74": "NON_PARTICIPATING",
 
237
  "78": "NON_PARTICIPATING",
238
  "79": "NON_PARTICIPATING",
239
  "8": "NON_PARTICIPATING",
240
+ "80": "NON_PARTICIPATING",
241
  "81": "NON_PARTICIPATING",
242
  "82": "NON_PARTICIPATING",
243
  "83": "NON_PARTICIPATING",
244
+ "84": "NON_PARTICIPATING",
245
  "85": "NON_PARTICIPATING",
246
  "86": "NON_PARTICIPATING",
247
  "87": "NON_PARTICIPATING",
 
256
  "95": "NON_PARTICIPATING",
257
  "96": "SUCCESS",
258
  "97": "NON_PARTICIPATING",
259
+ "98": "NON_PARTICIPATING",
260
  "99": "NON_PARTICIPATING"
261
  },
262
  "architectures": [
 
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
+ 5395806,
272
+ 5395810,
273
+ 5395815,
274
+ 5395819,
275
+ 5395823
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
+ "inner_step": 0,
283
  "inner_steps": 0,
284
+ "last_allreduce_block": 5395632,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcfea18d8dcabfacac1fe54346f6e6c281870b21da29e59ccc1c2d53ad8f7c42
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a9de59582f1efe499f7f8c2f7a2f7264304beba633b0bb929e53fe31d401577
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a37fae9bd2469fa44f3ab96c72a99fa01c7cf76f3ff675a88efb9c80b4bc7990
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22f339a6f66964efc25710cefa4bf7100e0613a2b429de3960b6d9fa873a1c3a
3
  size 4040701744