crypt0trading commited on
Commit
7647a0b
·
verified ·
1 Parent(s): cba0499

Outer Step 4. Inner Step 0. Batch Size 0

Browse files
Files changed (3) hide show
  1. config.json +7 -23
  2. inner_optimizer.pt +2 -2
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "crypt0trading/c66-h13",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
@@ -106,7 +106,7 @@
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
- "192": "NON_PARTICIPATING",
110
  "193": "NON_PARTICIPATING",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
@@ -170,7 +170,7 @@
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
  "249": "SUCCESS",
173
- "25": "NON_PARTICIPATING",
174
  "250": "NON_PARTICIPATING",
175
  "251": "NON_PARTICIPATING",
176
  "252": "NON_PARTICIPATING",
@@ -187,7 +187,7 @@
187
  "32": "NON_PARTICIPATING",
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
- "35": "SUCCESS",
191
  "36": "NON_PARTICIPATING",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
@@ -267,31 +267,15 @@
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
- "block_list": [
271
- 5227552,
272
- 5227557,
273
- 5227561,
274
- 5227565,
275
- 5227570,
276
- 5227574,
277
- 5227578,
278
- 5227582,
279
- 5227586,
280
- 5227590,
281
- 5227598,
282
- 5227609,
283
- 5227620,
284
- 5227630,
285
- 5227641
286
- ],
287
  "block_size": 1024,
288
  "bos_token_id": 50256,
289
  "embd_pdrop": 0.1,
290
  "eos_token_id": 50256,
291
  "initializer_range": 0.02,
292
- "inner_step": 132,
293
  "inner_steps": 0,
294
- "last_allreduce_block": 5226437,
295
  "layer_norm_epsilon": 1e-05,
296
  "model_type": "gpt_optimized",
297
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
 
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
+ "192": "SUCCESS",
110
  "193": "NON_PARTICIPATING",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
 
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
  "249": "SUCCESS",
173
+ "25": "SUCCESS",
174
  "250": "NON_PARTICIPATING",
175
  "251": "NON_PARTICIPATING",
176
  "252": "NON_PARTICIPATING",
 
187
  "32": "NON_PARTICIPATING",
188
  "33": "NON_PARTICIPATING",
189
  "34": "NON_PARTICIPATING",
190
+ "35": "NON_PARTICIPATING",
191
  "36": "NON_PARTICIPATING",
192
  "37": "NON_PARTICIPATING",
193
  "38": "NON_PARTICIPATING",
 
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
+ "block_list": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
271
  "block_size": 1024,
272
  "bos_token_id": 50256,
273
  "embd_pdrop": 0.1,
274
  "eos_token_id": 50256,
275
  "initializer_range": 0.02,
276
+ "inner_step": 0,
277
  "inner_steps": 0,
278
+ "last_allreduce_block": 5227728,
279
  "layer_norm_epsilon": 1e-05,
280
  "model_type": "gpt_optimized",
281
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a15fdd4605333479978e1728c76a845d4f52109ce14a19e0333e446bfed6c13
3
- size 8081781770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4cbaa53e58e44c7721a2bbb748a09d64dd416d5abb1c596a9bca2f5085b65c1
3
+ size 2752
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f55e78fb16bf253edd69adf0271706f2597ba5b5dbad821891e626a061fd8673
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbacd9c5aa5c72199f4b0f44e26f8da971b38e44f9497137ecd97afcb379f0fd
3
  size 4040701744