jockibaba commited on
Commit
06421bb
·
verified ·
1 Parent(s): bffd22e

Run 4. Outer Step 44. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +2 -8
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -267,19 +267,13 @@
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
- "block_list": [
271
- 5563202,
272
- 5563206,
273
- 5563210,
274
- 5563214,
275
- 5563218
276
- ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 80,
283
  "inner_steps": 0,
284
  "last_allreduce_block": 5543976,
285
  "layer_norm_epsilon": 1e-05,
 
267
  "AutoConfig": "distributed/optimized-gpt2-500m--configuration_gpt_optimized.GPTOptimConfig",
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
+ "block_list": [],
 
 
 
 
 
 
271
  "block_size": 1024,
272
  "bos_token_id": 50256,
273
  "embd_pdrop": 0.1,
274
  "eos_token_id": 50256,
275
  "initializer_range": 0.02,
276
+ "inner_step": 0,
277
  "inner_steps": 0,
278
  "last_allreduce_block": 5543976,
279
  "layer_norm_epsilon": 1e-05,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72fccec7191ee1f8983e7426a42546e0573fae9d681ac12c50a508756ad8a314
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fa32ae54655675d10cb46f1d8acd8bd46db2a5ba4721fb67f318ced3776079e
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:523af28a30011b064cf1ebff15dd60bf938ffcfd7fe4147990aba4ee99014cce
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e18b2b6fb366410e9de4a989d7b215596c0c6542a1ca528a189af865a9eb5f25
3
  size 4040701744