wandb02 commited on
Commit
a211ba4
·
verified ·
1 Parent(s): 7ab2059

Run 4. Outer Step 6. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +20 -20
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "_name_or_path": "wandb02/c66-h13",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
- "1": "NON_PARTICIPATING",
7
  "10": "SUCCESS",
8
  "100": "NON_PARTICIPATING",
9
  "101": "SUCCESS",
10
  "102": "NON_PARTICIPATING",
11
  "103": "NON_PARTICIPATING",
12
  "104": "NON_PARTICIPATING",
13
- "105": "NON_PARTICIPATING",
14
  "106": "NON_PARTICIPATING",
15
  "107": "SUCCESS",
16
  "108": "NON_PARTICIPATING",
@@ -20,7 +20,7 @@
20
  "111": "NON_PARTICIPATING",
21
  "112": "NON_PARTICIPATING",
22
  "113": "NON_PARTICIPATING",
23
- "114": "SUCCESS",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
  "117": "NON_PARTICIPATING",
@@ -58,7 +58,7 @@
58
  "146": "NON_PARTICIPATING",
59
  "147": "SUCCESS",
60
  "148": "SUCCESS",
61
- "149": "NON_PARTICIPATING",
62
  "15": "SUCCESS",
63
  "150": "NON_PARTICIPATING",
64
  "151": "NON_PARTICIPATING",
@@ -71,7 +71,7 @@
71
  "158": "NON_PARTICIPATING",
72
  "159": "NON_PARTICIPATING",
73
  "16": "SUCCESS",
74
- "160": "SUCCESS",
75
  "161": "NON_PARTICIPATING",
76
  "162": "NON_PARTICIPATING",
77
  "163": "NON_PARTICIPATING",
@@ -84,7 +84,7 @@
84
  "17": "SUCCESS",
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
- "172": "SUCCESS",
88
  "173": "NON_PARTICIPATING",
89
  "174": "NON_PARTICIPATING",
90
  "175": "SUCCESS",
@@ -97,16 +97,16 @@
97
  "181": "NON_PARTICIPATING",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
- "184": "NON_PARTICIPATING",
101
  "185": "NON_PARTICIPATING",
102
- "186": "SUCCESS",
103
  "187": "NON_PARTICIPATING",
104
  "188": "NON_PARTICIPATING",
105
  "189": "NON_PARTICIPATING",
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
- "192": "SUCCESS",
110
  "193": "NON_PARTICIPATING",
111
  "194": "SUCCESS",
112
  "195": "SUCCESS",
@@ -208,7 +208,7 @@
208
  "51": "SUCCESS",
209
  "52": "NON_PARTICIPATING",
210
  "53": "SUCCESS",
211
- "54": "SUCCESS",
212
  "55": "SUCCESS",
213
  "56": "SUCCESS",
214
  "57": "NON_PARTICIPATING",
@@ -218,7 +218,7 @@
218
  "60": "SUCCESS",
219
  "61": "NON_PARTICIPATING",
220
  "62": "SUCCESS",
221
- "63": "NON_PARTICIPATING",
222
  "64": "NON_PARTICIPATING",
223
  "65": "NON_PARTICIPATING",
224
  "66": "NON_PARTICIPATING",
@@ -254,7 +254,7 @@
254
  "93": "NON_PARTICIPATING",
255
  "94": "NON_PARTICIPATING",
256
  "95": "SUCCESS",
257
- "96": "NON_PARTICIPATING",
258
  "97": "NON_PARTICIPATING",
259
  "98": "NON_PARTICIPATING",
260
  "99": "SUCCESS"
@@ -268,20 +268,20 @@
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
- 5489987,
272
- 5489991,
273
- 5489996,
274
- 5490000,
275
- 5490003
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 77,
283
  "inner_steps": 0,
284
- "last_allreduce_block": 5487765,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
+ "1": "SUCCESS",
7
  "10": "SUCCESS",
8
  "100": "NON_PARTICIPATING",
9
  "101": "SUCCESS",
10
  "102": "NON_PARTICIPATING",
11
  "103": "NON_PARTICIPATING",
12
  "104": "NON_PARTICIPATING",
13
+ "105": "SUCCESS",
14
  "106": "NON_PARTICIPATING",
15
  "107": "SUCCESS",
16
  "108": "NON_PARTICIPATING",
 
20
  "111": "NON_PARTICIPATING",
21
  "112": "NON_PARTICIPATING",
22
  "113": "NON_PARTICIPATING",
23
+ "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
  "117": "NON_PARTICIPATING",
 
58
  "146": "NON_PARTICIPATING",
59
  "147": "SUCCESS",
60
  "148": "SUCCESS",
61
+ "149": "SUCCESS",
62
  "15": "SUCCESS",
63
  "150": "NON_PARTICIPATING",
64
  "151": "NON_PARTICIPATING",
 
71
  "158": "NON_PARTICIPATING",
72
  "159": "NON_PARTICIPATING",
73
  "16": "SUCCESS",
74
+ "160": "FAIL",
75
  "161": "NON_PARTICIPATING",
76
  "162": "NON_PARTICIPATING",
77
  "163": "NON_PARTICIPATING",
 
84
  "17": "SUCCESS",
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
+ "172": "NON_PARTICIPATING",
88
  "173": "NON_PARTICIPATING",
89
  "174": "NON_PARTICIPATING",
90
  "175": "SUCCESS",
 
97
  "181": "NON_PARTICIPATING",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
+ "184": "SUCCESS",
101
  "185": "NON_PARTICIPATING",
102
+ "186": "NON_PARTICIPATING",
103
  "187": "NON_PARTICIPATING",
104
  "188": "NON_PARTICIPATING",
105
  "189": "NON_PARTICIPATING",
106
  "19": "NON_PARTICIPATING",
107
  "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
+ "192": "NON_PARTICIPATING",
110
  "193": "NON_PARTICIPATING",
111
  "194": "SUCCESS",
112
  "195": "SUCCESS",
 
208
  "51": "SUCCESS",
209
  "52": "NON_PARTICIPATING",
210
  "53": "SUCCESS",
211
+ "54": "NON_PARTICIPATING",
212
  "55": "SUCCESS",
213
  "56": "SUCCESS",
214
  "57": "NON_PARTICIPATING",
 
218
  "60": "SUCCESS",
219
  "61": "NON_PARTICIPATING",
220
  "62": "SUCCESS",
221
+ "63": "SUCCESS",
222
  "64": "NON_PARTICIPATING",
223
  "65": "NON_PARTICIPATING",
224
  "66": "NON_PARTICIPATING",
 
254
  "93": "NON_PARTICIPATING",
255
  "94": "NON_PARTICIPATING",
256
  "95": "SUCCESS",
257
+ "96": "SUCCESS",
258
  "97": "NON_PARTICIPATING",
259
  "98": "NON_PARTICIPATING",
260
  "99": "SUCCESS"
 
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
+ 5490071,
272
+ 5490138,
273
+ 5490142,
274
+ 5490146,
275
+ 5490150
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
+ "inner_step": 0,
283
  "inner_steps": 0,
284
+ "last_allreduce_block": 5490074,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f27157348c0fccc69e9112f8a498b939234e8b0a26574b9263d6b2adfbe1091
3
  size 8081782503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c1187b6fd4a15fdb7e0d7205e470d360d291e5680e4a17cac88e01b34083242
3
  size 8081782503
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0df88b051ce4e4c2cb94df8323894cb32775153e99046c9341d8cbf0337d0428
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feeff45d1021a7e9ff5bba2e352dbf923b3e2978877f5c2c450c6c4f9fdce4d5
3
  size 4040701744