crypt0trading commited on
Commit
2f5f935
·
verified ·
1 Parent(s): 1c19671

Run 2. Outer Step 3. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +17 -23
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "crypt0trading/c66-h13",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
@@ -23,7 +23,7 @@
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
- "117": "SUCCESS",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "NON_PARTICIPATING",
@@ -85,7 +85,7 @@
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
  "172": "NON_PARTICIPATING",
88
- "173": "NON_PARTICIPATING",
89
  "174": "NON_PARTICIPATING",
90
  "175": "NON_PARTICIPATING",
91
  "176": "NON_PARTICIPATING",
@@ -94,10 +94,10 @@
94
  "179": "NON_PARTICIPATING",
95
  "18": "NON_PARTICIPATING",
96
  "180": "NON_PARTICIPATING",
97
- "181": "SUCCESS",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
- "184": "SUCCESS",
101
  "185": "NON_PARTICIPATING",
102
  "186": "NON_PARTICIPATING",
103
  "187": "NON_PARTICIPATING",
@@ -114,9 +114,9 @@
114
  "197": "NON_PARTICIPATING",
115
  "198": "NON_PARTICIPATING",
116
  "199": "NON_PARTICIPATING",
117
- "2": "SUCCESS",
118
  "20": "NON_PARTICIPATING",
119
- "200": "SUCCESS",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
122
  "203": "NON_PARTICIPATING",
@@ -141,12 +141,12 @@
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
144
- "223": "NON_PARTICIPATING",
145
  "224": "NON_PARTICIPATING",
146
  "225": "NON_PARTICIPATING",
147
  "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
- "228": "NON_PARTICIPATING",
150
  "229": "NON_PARTICIPATING",
151
  "23": "SUCCESS",
152
  "230": "NON_PARTICIPATING",
@@ -188,17 +188,17 @@
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
- "36": "NON_PARTICIPATING",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
194
- "39": "NON_PARTICIPATING",
195
  "4": "NON_PARTICIPATING",
196
  "40": "NON_PARTICIPATING",
197
  "41": "NON_PARTICIPATING",
198
  "42": "NON_PARTICIPATING",
199
  "43": "NON_PARTICIPATING",
200
  "44": "NON_PARTICIPATING",
201
- "45": "NON_PARTICIPATING",
202
  "46": "NON_PARTICIPATING",
203
  "47": "NON_PARTICIPATING",
204
  "48": "NON_PARTICIPATING",
@@ -212,7 +212,7 @@
212
  "55": "NON_PARTICIPATING",
213
  "56": "NON_PARTICIPATING",
214
  "57": "NON_PARTICIPATING",
215
- "58": "NON_PARTICIPATING",
216
  "59": "NON_PARTICIPATING",
217
  "6": "NON_PARTICIPATING",
218
  "60": "NON_PARTICIPATING",
@@ -234,7 +234,7 @@
234
  "75": "NON_PARTICIPATING",
235
  "76": "NON_PARTICIPATING",
236
  "77": "NON_PARTICIPATING",
237
- "78": "SUCCESS",
238
  "79": "NON_PARTICIPATING",
239
  "8": "NON_PARTICIPATING",
240
  "80": "NON_PARTICIPATING",
@@ -267,21 +267,15 @@
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
- "block_list": [
271
- 5330554,
272
- 5330558,
273
- 5330562,
274
- 5330566,
275
- 5330570
276
- ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 117,
283
  "inner_steps": 0,
284
- "last_allreduce_block": 5323221,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
 
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
+ "117": "NON_PARTICIPATING",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "NON_PARTICIPATING",
 
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
  "172": "NON_PARTICIPATING",
88
+ "173": "SUCCESS",
89
  "174": "NON_PARTICIPATING",
90
  "175": "NON_PARTICIPATING",
91
  "176": "NON_PARTICIPATING",
 
94
  "179": "NON_PARTICIPATING",
95
  "18": "NON_PARTICIPATING",
96
  "180": "NON_PARTICIPATING",
97
+ "181": "NON_PARTICIPATING",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
+ "184": "NON_PARTICIPATING",
101
  "185": "NON_PARTICIPATING",
102
  "186": "NON_PARTICIPATING",
103
  "187": "NON_PARTICIPATING",
 
114
  "197": "NON_PARTICIPATING",
115
  "198": "NON_PARTICIPATING",
116
  "199": "NON_PARTICIPATING",
117
+ "2": "NON_PARTICIPATING",
118
  "20": "NON_PARTICIPATING",
119
+ "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
122
  "203": "NON_PARTICIPATING",
 
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
144
+ "223": "SUCCESS",
145
  "224": "NON_PARTICIPATING",
146
  "225": "NON_PARTICIPATING",
147
  "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
+ "228": "SUCCESS",
150
  "229": "NON_PARTICIPATING",
151
  "23": "SUCCESS",
152
  "230": "NON_PARTICIPATING",
 
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
+ "36": "SUCCESS",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
194
+ "39": "SUCCESS",
195
  "4": "NON_PARTICIPATING",
196
  "40": "NON_PARTICIPATING",
197
  "41": "NON_PARTICIPATING",
198
  "42": "NON_PARTICIPATING",
199
  "43": "NON_PARTICIPATING",
200
  "44": "NON_PARTICIPATING",
201
+ "45": "SUCCESS",
202
  "46": "NON_PARTICIPATING",
203
  "47": "NON_PARTICIPATING",
204
  "48": "NON_PARTICIPATING",
 
212
  "55": "NON_PARTICIPATING",
213
  "56": "NON_PARTICIPATING",
214
  "57": "NON_PARTICIPATING",
215
+ "58": "SUCCESS",
216
  "59": "NON_PARTICIPATING",
217
  "6": "NON_PARTICIPATING",
218
  "60": "NON_PARTICIPATING",
 
234
  "75": "NON_PARTICIPATING",
235
  "76": "NON_PARTICIPATING",
236
  "77": "NON_PARTICIPATING",
237
+ "78": "NON_PARTICIPATING",
238
  "79": "NON_PARTICIPATING",
239
  "8": "NON_PARTICIPATING",
240
  "80": "NON_PARTICIPATING",
 
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
+ "block_list": [],
 
 
 
 
 
 
271
  "block_size": 1024,
272
  "bos_token_id": 50256,
273
  "embd_pdrop": 0.1,
274
  "eos_token_id": 50256,
275
  "initializer_range": 0.02,
276
+ "inner_step": 0,
277
  "inner_steps": 0,
278
+ "last_allreduce_block": 5330632,
279
  "layer_norm_epsilon": 1e-05,
280
  "model_type": "gpt_optimized",
281
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fa75e0b8e6d0c03acf813da2644fd9b5ad7fb336d9ad77229c48875555f1ce3
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:898b0a36c2505d6758769a80676b4795fcab5f81ef52c3f9a1e115538eec578d
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:958310cb9221d13f6559c726d2fa5a78d892f8cee3799e4940b87e877bee47be
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d627d8a20903d8137d02511fb1a23ea4261d41dc04f46a5426df8e48bcd8f67
3
  size 4040701744