crypt0trading commited on
Commit
a263288
·
verified ·
1 Parent(s): 52d7841

Run 2. Outer Step 7. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +14 -20
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "_name_or_path": "crypt0trading/c66-h5",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
- "1": "NON_PARTICIPATING",
7
  "10": "NON_PARTICIPATING",
8
  "100": "NON_PARTICIPATING",
9
  "101": "NON_PARTICIPATING",
@@ -14,7 +14,7 @@
14
  "106": "NON_PARTICIPATING",
15
  "107": "NON_PARTICIPATING",
16
  "108": "NON_PARTICIPATING",
17
- "109": "NON_PARTICIPATING",
18
  "11": "NON_PARTICIPATING",
19
  "110": "NON_PARTICIPATING",
20
  "111": "NON_PARTICIPATING",
@@ -23,7 +23,7 @@
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
- "117": "SUCCESS",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "NON_PARTICIPATING",
@@ -106,7 +106,7 @@
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
- "192": "SUCCESS",
110
  "193": "NON_PARTICIPATING",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
@@ -116,7 +116,7 @@
116
  "199": "NON_PARTICIPATING",
117
  "2": "NON_PARTICIPATING",
118
  "20": "NON_PARTICIPATING",
119
- "200": "SUCCESS",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
122
  "203": "NON_PARTICIPATING",
@@ -137,7 +137,7 @@
137
  "217": "NON_PARTICIPATING",
138
  "218": "NON_PARTICIPATING",
139
  "219": "NON_PARTICIPATING",
140
- "22": "NON_PARTICIPATING",
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
@@ -146,7 +146,7 @@
146
  "225": "NON_PARTICIPATING",
147
  "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
- "228": "SUCCESS",
150
  "229": "NON_PARTICIPATING",
151
  "23": "SUCCESS",
152
  "230": "NON_PARTICIPATING",
@@ -169,7 +169,7 @@
169
  "246": "NON_PARTICIPATING",
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
- "249": "SUCCESS",
173
  "25": "SUCCESS",
174
  "250": "NON_PARTICIPATING",
175
  "251": "NON_PARTICIPATING",
@@ -188,7 +188,7 @@
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
- "36": "SUCCESS",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
194
  "39": "SUCCESS",
@@ -198,7 +198,7 @@
198
  "42": "NON_PARTICIPATING",
199
  "43": "NON_PARTICIPATING",
200
  "44": "NON_PARTICIPATING",
201
- "45": "SUCCESS",
202
  "46": "NON_PARTICIPATING",
203
  "47": "NON_PARTICIPATING",
204
  "48": "NON_PARTICIPATING",
@@ -267,21 +267,15 @@
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
- "block_list": [
271
- 5338057,
272
- 5338062,
273
- 5338066,
274
- 5338070,
275
- 5338075
276
- ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 558,
283
  "inner_steps": 0,
284
- "last_allreduce_block": 5332469,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
+ "1": "SUCCESS",
7
  "10": "NON_PARTICIPATING",
8
  "100": "NON_PARTICIPATING",
9
  "101": "NON_PARTICIPATING",
 
14
  "106": "NON_PARTICIPATING",
15
  "107": "NON_PARTICIPATING",
16
  "108": "NON_PARTICIPATING",
17
+ "109": "SUCCESS",
18
  "11": "NON_PARTICIPATING",
19
  "110": "NON_PARTICIPATING",
20
  "111": "NON_PARTICIPATING",
 
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
+ "117": "NON_PARTICIPATING",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "NON_PARTICIPATING",
 
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
+ "192": "NON_PARTICIPATING",
110
  "193": "NON_PARTICIPATING",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
 
116
  "199": "NON_PARTICIPATING",
117
  "2": "NON_PARTICIPATING",
118
  "20": "NON_PARTICIPATING",
119
+ "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
  "202": "NON_PARTICIPATING",
122
  "203": "NON_PARTICIPATING",
 
137
  "217": "NON_PARTICIPATING",
138
  "218": "NON_PARTICIPATING",
139
  "219": "NON_PARTICIPATING",
140
+ "22": "SUCCESS",
141
  "220": "NON_PARTICIPATING",
142
  "221": "NON_PARTICIPATING",
143
  "222": "NON_PARTICIPATING",
 
146
  "225": "NON_PARTICIPATING",
147
  "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
+ "228": "NON_PARTICIPATING",
150
  "229": "NON_PARTICIPATING",
151
  "23": "SUCCESS",
152
  "230": "NON_PARTICIPATING",
 
169
  "246": "NON_PARTICIPATING",
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
+ "249": "NON_PARTICIPATING",
173
  "25": "SUCCESS",
174
  "250": "NON_PARTICIPATING",
175
  "251": "NON_PARTICIPATING",
 
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
  "35": "NON_PARTICIPATING",
191
+ "36": "NON_PARTICIPATING",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
194
  "39": "SUCCESS",
 
198
  "42": "NON_PARTICIPATING",
199
  "43": "NON_PARTICIPATING",
200
  "44": "NON_PARTICIPATING",
201
+ "45": "NON_PARTICIPATING",
202
  "46": "NON_PARTICIPATING",
203
  "47": "NON_PARTICIPATING",
204
  "48": "NON_PARTICIPATING",
 
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
+ "block_list": [],
 
 
 
 
 
 
271
  "block_size": 1024,
272
  "bos_token_id": 50256,
273
  "embd_pdrop": 0.1,
274
  "eos_token_id": 50256,
275
  "initializer_range": 0.02,
276
+ "inner_step": 0,
277
  "inner_steps": 0,
278
+ "last_allreduce_block": 5339896,
279
  "layer_norm_epsilon": 1e-05,
280
  "model_type": "gpt_optimized",
281
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f07f922908cbe3583e954b54018b0d3adbe4157195cbd3725550f6efefbfab4d
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed94bc36dce83a7e9f7b2a0fa58b6d3de86f514816a5452bc24b18d96b3944e1
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f3dd47ac89203fbd6eb6eb4e07f6c3b8df76fb7cb3e2911934218fcecfb946d
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcdfb45c1c7b670e719f10e78091cec2c601ac898763796339f16712e03c08c1
3
  size 4040701744