Jacksss123 commited on
Commit
9022dfb
·
verified ·
1 Parent(s): 1637397

Run 3. Outer Step 11. Inner Step 196.

Browse files
Files changed (3) hide show
  1. config.json +6 -6
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -268,18 +268,18 @@
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
- 5443031,
272
- 5443035,
273
- 5443038,
274
- 5443042,
275
- 5443045
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 195,
283
  "inner_steps": 0,
284
  "last_allreduce_block": 5422706,
285
  "layer_norm_epsilon": 1e-05,
 
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
+ 5443075,
272
+ 5443078,
273
+ 5443082,
274
+ 5443086,
275
+ 5443089
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
+ "inner_step": 196,
283
  "inner_steps": 0,
284
  "last_allreduce_block": 5422706,
285
  "layer_norm_epsilon": 1e-05,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f352865d5d6eb82b575053c54cacb8de5045523aee24b74ca28a5f7f8c12dc6c
3
  size 8081782026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:320ffbe8b6437ca1149adc23b609342c9f5b3bbc29b001478937d93026c57090
3
  size 8081782026
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc63bebaf0d68d2b849570f5f72f0e41bb4b20872b9d08ba6077f94d2011923a
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3faeb98aa70a7f0fc8684ebdf9015d5593d0b1a10a3da6d5133f8b15c6a37f08
3
  size 4040701744