jockibaba commited on
Commit
02cae82
·
verified ·
1 Parent(s): f6d5fe7

Run 4. Outer Step 0. Inner Step 3.

Browse files
Files changed (3) hide show
  1. config.json +5 -15
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -268,27 +268,17 @@
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
- 5609166,
272
- 5609170,
273
- 5609175,
274
- 5609179,
275
- 5609184,
276
- 5609188,
277
- 5609193,
278
- 5609197,
279
- 5609202,
280
- 5609206,
281
- 5609211,
282
- 5609215,
283
- 5609219,
284
- 5609223
285
  ],
286
  "block_size": 1024,
287
  "bos_token_id": 50256,
288
  "embd_pdrop": 0.1,
289
  "eos_token_id": 50256,
290
  "initializer_range": 0.02,
291
- "inner_step": 2,
292
  "inner_steps": 0,
293
  "last_allreduce_block": 5470731,
294
  "layer_norm_epsilon": 1e-05,
 
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
+ 5609229,
272
+ 5609234,
273
+ 5609239,
274
+ 5609244
 
 
 
 
 
 
 
 
 
 
275
  ],
276
  "block_size": 1024,
277
  "bos_token_id": 50256,
278
  "embd_pdrop": 0.1,
279
  "eos_token_id": 50256,
280
  "initializer_range": 0.02,
281
+ "inner_step": 3,
282
  "inner_steps": 0,
283
  "last_allreduce_block": 5470731,
284
  "layer_norm_epsilon": 1e-05,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c26c5891d268e84c07f659a8ff1b7392c90189b255580ebaa0ef440de24ed1c
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1706debd5143351e7a9f71d01cd9b28f8696ccdae2c24c7f0aae36bd8349ab26
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e68a816e593fdae4f7dd816dae1353dc7d7a30448966f973289744657c8eb3bb
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b1a4d4c0617e3878feb8c3e738c232965424044bf9fefd795bfa073e7e2eaf9
3
  size 4040701744