jockibaba commited on
Commit
90b9017
·
verified ·
1 Parent(s): 8aaaccd

Run 4. Outer Step 42. Inner Step 0.

Browse files
Files changed (3) hide show
  1. config.json +47 -47
  2. inner_optimizer.pt +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -1,21 +1,21 @@
1
  {
2
- "_name_or_path": "jockibaba/jockibaba7",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
  "1": "SUCCESS",
7
  "10": "NON_PARTICIPATING",
8
  "100": "NON_PARTICIPATING",
9
- "101": "NON_PARTICIPATING",
10
- "102": "SUCCESS",
11
  "103": "NON_PARTICIPATING",
12
  "104": "NON_PARTICIPATING",
13
  "105": "NON_PARTICIPATING",
14
  "106": "NON_PARTICIPATING",
15
- "107": "NON_PARTICIPATING",
16
  "108": "NON_PARTICIPATING",
17
- "109": "SUCCESS",
18
- "11": "NON_PARTICIPATING",
19
  "110": "NON_PARTICIPATING",
20
  "111": "NON_PARTICIPATING",
21
  "112": "NON_PARTICIPATING",
@@ -23,18 +23,18 @@
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
- "117": "SUCCESS",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "SUCCESS",
30
  "120": "NON_PARTICIPATING",
31
- "121": "NON_PARTICIPATING",
32
  "122": "NON_PARTICIPATING",
33
- "123": "SUCCESS",
34
  "124": "NON_PARTICIPATING",
35
  "125": "NON_PARTICIPATING",
36
- "126": "NON_PARTICIPATING",
37
- "127": "NON_PARTICIPATING",
38
  "128": "NON_PARTICIPATING",
39
  "129": "NON_PARTICIPATING",
40
  "13": "SUCCESS",
@@ -44,9 +44,9 @@
44
  "133": "NON_PARTICIPATING",
45
  "134": "NON_PARTICIPATING",
46
  "135": "NON_PARTICIPATING",
47
- "136": "SUCCESS",
48
  "137": "NON_PARTICIPATING",
49
- "138": "NON_PARTICIPATING",
50
  "139": "NON_PARTICIPATING",
51
  "14": "SUCCESS",
52
  "140": "NON_PARTICIPATING",
@@ -63,11 +63,11 @@
63
  "150": "NON_PARTICIPATING",
64
  "151": "NON_PARTICIPATING",
65
  "152": "NON_PARTICIPATING",
66
- "153": "SUCCESS",
67
  "154": "NON_PARTICIPATING",
68
  "155": "NON_PARTICIPATING",
69
- "156": "SUCCESS",
70
- "157": "SUCCESS",
71
  "158": "NON_PARTICIPATING",
72
  "159": "NON_PARTICIPATING",
73
  "16": "SUCCESS",
@@ -85,7 +85,7 @@
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
  "172": "NON_PARTICIPATING",
88
- "173": "NON_PARTICIPATING",
89
  "174": "NON_PARTICIPATING",
90
  "175": "NON_PARTICIPATING",
91
  "176": "NON_PARTICIPATING",
@@ -94,7 +94,7 @@
94
  "179": "NON_PARTICIPATING",
95
  "18": "SUCCESS",
96
  "180": "NON_PARTICIPATING",
97
- "181": "NON_PARTICIPATING",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
  "184": "NON_PARTICIPATING",
@@ -103,11 +103,11 @@
103
  "187": "NON_PARTICIPATING",
104
  "188": "NON_PARTICIPATING",
105
  "189": "NON_PARTICIPATING",
106
- "19": "NON_PARTICIPATING",
107
- "190": "NON_PARTICIPATING",
108
  "191": "NON_PARTICIPATING",
109
- "192": "NON_PARTICIPATING",
110
- "193": "NON_PARTICIPATING",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
113
  "196": "NON_PARTICIPATING",
@@ -118,14 +118,14 @@
118
  "20": "SUCCESS",
119
  "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
- "202": "SUCCESS",
122
  "203": "NON_PARTICIPATING",
123
  "204": "NON_PARTICIPATING",
124
  "205": "NON_PARTICIPATING",
125
  "206": "NON_PARTICIPATING",
126
  "207": "NON_PARTICIPATING",
127
  "208": "NON_PARTICIPATING",
128
- "209": "SUCCESS",
129
  "21": "SUCCESS",
130
  "210": "NON_PARTICIPATING",
131
  "211": "NON_PARTICIPATING",
@@ -143,8 +143,8 @@
143
  "222": "NON_PARTICIPATING",
144
  "223": "NON_PARTICIPATING",
145
  "224": "NON_PARTICIPATING",
146
- "225": "NON_PARTICIPATING",
147
- "226": "NON_PARTICIPATING",
148
  "227": "NON_PARTICIPATING",
149
  "228": "NON_PARTICIPATING",
150
  "229": "NON_PARTICIPATING",
@@ -154,24 +154,24 @@
154
  "232": "NON_PARTICIPATING",
155
  "233": "NON_PARTICIPATING",
156
  "234": "NON_PARTICIPATING",
157
- "235": "SUCCESS",
158
  "236": "NON_PARTICIPATING",
159
  "237": "NON_PARTICIPATING",
160
- "238": "NON_PARTICIPATING",
161
  "239": "NON_PARTICIPATING",
162
  "24": "SUCCESS",
163
- "240": "NON_PARTICIPATING",
164
- "241": "NON_PARTICIPATING",
165
- "242": "SUCCESS",
166
- "243": "SUCCESS",
167
- "244": "SUCCESS",
168
  "245": "SUCCESS",
169
  "246": "NON_PARTICIPATING",
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
  "249": "SUCCESS",
173
  "25": "SUCCESS",
174
- "250": "NON_PARTICIPATING",
175
  "251": "NON_PARTICIPATING",
176
  "252": "NON_PARTICIPATING",
177
  "253": "NON_PARTICIPATING",
@@ -183,7 +183,7 @@
183
  "29": "SUCCESS",
184
  "3": "SUCCESS",
185
  "30": "SUCCESS",
186
- "31": "NON_PARTICIPATING",
187
  "32": "SUCCESS",
188
  "33": "NON_PARTICIPATING",
189
  "34": "SUCCESS",
@@ -203,13 +203,13 @@
203
  "47": "SUCCESS",
204
  "48": "NON_PARTICIPATING",
205
  "49": "SUCCESS",
206
- "5": "SUCCESS",
207
  "50": "SUCCESS",
208
  "51": "SUCCESS",
209
  "52": "SUCCESS",
210
  "53": "SUCCESS",
211
  "54": "SUCCESS",
212
- "55": "SUCCESS",
213
  "56": "SUCCESS",
214
  "57": "SUCCESS",
215
  "58": "NON_PARTICIPATING",
@@ -230,7 +230,7 @@
230
  "71": "NON_PARTICIPATING",
231
  "72": "SUCCESS",
232
  "73": "NON_PARTICIPATING",
233
- "74": "NON_PARTICIPATING",
234
  "75": "SUCCESS",
235
  "76": "SUCCESS",
236
  "77": "SUCCESS",
@@ -243,11 +243,11 @@
243
  "83": "NON_PARTICIPATING",
244
  "84": "SUCCESS",
245
  "85": "NON_PARTICIPATING",
246
- "86": "SUCCESS",
247
  "87": "SUCCESS",
248
  "88": "SUCCESS",
249
  "89": "SUCCESS",
250
- "9": "NON_PARTICIPATING",
251
  "90": "SUCCESS",
252
  "91": "SUCCESS",
253
  "92": "SUCCESS",
@@ -268,20 +268,20 @@
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
- 5559556,
272
- 5559560,
273
- 5559565,
274
- 5559569,
275
- 5559573
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
- "inner_step": 32,
283
  "inner_steps": 0,
284
- "last_allreduce_block": 5552497,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
 
1
  {
2
+ "_name_or_path": "distributed/optimized-gpt2-1b",
3
  "activation_function": "gelu_new",
4
  "all_reduce_scores": {
5
  "0": "NON_PARTICIPATING",
6
  "1": "SUCCESS",
7
  "10": "NON_PARTICIPATING",
8
  "100": "NON_PARTICIPATING",
9
+ "101": "FAIL",
10
+ "102": "NON_PARTICIPATING",
11
  "103": "NON_PARTICIPATING",
12
  "104": "NON_PARTICIPATING",
13
  "105": "NON_PARTICIPATING",
14
  "106": "NON_PARTICIPATING",
15
+ "107": "SUCCESS",
16
  "108": "NON_PARTICIPATING",
17
+ "109": "NON_PARTICIPATING",
18
+ "11": "SUCCESS",
19
  "110": "NON_PARTICIPATING",
20
  "111": "NON_PARTICIPATING",
21
  "112": "NON_PARTICIPATING",
 
23
  "114": "NON_PARTICIPATING",
24
  "115": "NON_PARTICIPATING",
25
  "116": "NON_PARTICIPATING",
26
+ "117": "NON_PARTICIPATING",
27
  "118": "NON_PARTICIPATING",
28
  "119": "NON_PARTICIPATING",
29
  "12": "SUCCESS",
30
  "120": "NON_PARTICIPATING",
31
+ "121": "SUCCESS",
32
  "122": "NON_PARTICIPATING",
33
+ "123": "NON_PARTICIPATING",
34
  "124": "NON_PARTICIPATING",
35
  "125": "NON_PARTICIPATING",
36
+ "126": "SUCCESS",
37
+ "127": "SUCCESS",
38
  "128": "NON_PARTICIPATING",
39
  "129": "NON_PARTICIPATING",
40
  "13": "SUCCESS",
 
44
  "133": "NON_PARTICIPATING",
45
  "134": "NON_PARTICIPATING",
46
  "135": "NON_PARTICIPATING",
47
+ "136": "NON_PARTICIPATING",
48
  "137": "NON_PARTICIPATING",
49
+ "138": "SUCCESS",
50
  "139": "NON_PARTICIPATING",
51
  "14": "SUCCESS",
52
  "140": "NON_PARTICIPATING",
 
63
  "150": "NON_PARTICIPATING",
64
  "151": "NON_PARTICIPATING",
65
  "152": "NON_PARTICIPATING",
66
+ "153": "NON_PARTICIPATING",
67
  "154": "NON_PARTICIPATING",
68
  "155": "NON_PARTICIPATING",
69
+ "156": "NON_PARTICIPATING",
70
+ "157": "NON_PARTICIPATING",
71
  "158": "NON_PARTICIPATING",
72
  "159": "NON_PARTICIPATING",
73
  "16": "SUCCESS",
 
85
  "170": "NON_PARTICIPATING",
86
  "171": "NON_PARTICIPATING",
87
  "172": "NON_PARTICIPATING",
88
+ "173": "SUCCESS",
89
  "174": "NON_PARTICIPATING",
90
  "175": "NON_PARTICIPATING",
91
  "176": "NON_PARTICIPATING",
 
94
  "179": "NON_PARTICIPATING",
95
  "18": "SUCCESS",
96
  "180": "NON_PARTICIPATING",
97
+ "181": "SUCCESS",
98
  "182": "NON_PARTICIPATING",
99
  "183": "NON_PARTICIPATING",
100
  "184": "NON_PARTICIPATING",
 
103
  "187": "NON_PARTICIPATING",
104
  "188": "NON_PARTICIPATING",
105
  "189": "NON_PARTICIPATING",
106
+ "19": "SUCCESS",
107
+ "190": "SUCCESS",
108
  "191": "NON_PARTICIPATING",
109
+ "192": "SUCCESS",
110
+ "193": "SUCCESS",
111
  "194": "NON_PARTICIPATING",
112
  "195": "NON_PARTICIPATING",
113
  "196": "NON_PARTICIPATING",
 
118
  "20": "SUCCESS",
119
  "200": "NON_PARTICIPATING",
120
  "201": "NON_PARTICIPATING",
121
+ "202": "NON_PARTICIPATING",
122
  "203": "NON_PARTICIPATING",
123
  "204": "NON_PARTICIPATING",
124
  "205": "NON_PARTICIPATING",
125
  "206": "NON_PARTICIPATING",
126
  "207": "NON_PARTICIPATING",
127
  "208": "NON_PARTICIPATING",
128
+ "209": "NON_PARTICIPATING",
129
  "21": "SUCCESS",
130
  "210": "NON_PARTICIPATING",
131
  "211": "NON_PARTICIPATING",
 
143
  "222": "NON_PARTICIPATING",
144
  "223": "NON_PARTICIPATING",
145
  "224": "NON_PARTICIPATING",
146
+ "225": "SUCCESS",
147
+ "226": "SUCCESS",
148
  "227": "NON_PARTICIPATING",
149
  "228": "NON_PARTICIPATING",
150
  "229": "NON_PARTICIPATING",
 
154
  "232": "NON_PARTICIPATING",
155
  "233": "NON_PARTICIPATING",
156
  "234": "NON_PARTICIPATING",
157
+ "235": "NON_PARTICIPATING",
158
  "236": "NON_PARTICIPATING",
159
  "237": "NON_PARTICIPATING",
160
+ "238": "SUCCESS",
161
  "239": "NON_PARTICIPATING",
162
  "24": "SUCCESS",
163
+ "240": "SUCCESS",
164
+ "241": "SUCCESS",
165
+ "242": "NON_PARTICIPATING",
166
+ "243": "NON_PARTICIPATING",
167
+ "244": "NON_PARTICIPATING",
168
  "245": "SUCCESS",
169
  "246": "NON_PARTICIPATING",
170
  "247": "NON_PARTICIPATING",
171
  "248": "NON_PARTICIPATING",
172
  "249": "SUCCESS",
173
  "25": "SUCCESS",
174
+ "250": "SUCCESS",
175
  "251": "NON_PARTICIPATING",
176
  "252": "NON_PARTICIPATING",
177
  "253": "NON_PARTICIPATING",
 
183
  "29": "SUCCESS",
184
  "3": "SUCCESS",
185
  "30": "SUCCESS",
186
+ "31": "SUCCESS",
187
  "32": "SUCCESS",
188
  "33": "NON_PARTICIPATING",
189
  "34": "SUCCESS",
 
203
  "47": "SUCCESS",
204
  "48": "NON_PARTICIPATING",
205
  "49": "SUCCESS",
206
+ "5": "NON_PARTICIPATING",
207
  "50": "SUCCESS",
208
  "51": "SUCCESS",
209
  "52": "SUCCESS",
210
  "53": "SUCCESS",
211
  "54": "SUCCESS",
212
+ "55": "NON_PARTICIPATING",
213
  "56": "SUCCESS",
214
  "57": "SUCCESS",
215
  "58": "NON_PARTICIPATING",
 
230
  "71": "NON_PARTICIPATING",
231
  "72": "SUCCESS",
232
  "73": "NON_PARTICIPATING",
233
+ "74": "SUCCESS",
234
  "75": "SUCCESS",
235
  "76": "SUCCESS",
236
  "77": "SUCCESS",
 
243
  "83": "NON_PARTICIPATING",
244
  "84": "SUCCESS",
245
  "85": "NON_PARTICIPATING",
246
+ "86": "NON_PARTICIPATING",
247
  "87": "SUCCESS",
248
  "88": "SUCCESS",
249
  "89": "SUCCESS",
250
+ "9": "SUCCESS",
251
  "90": "SUCCESS",
252
  "91": "SUCCESS",
253
  "92": "SUCCESS",
 
268
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
269
  },
270
  "block_list": [
271
+ 5559709,
272
+ 5559713,
273
+ 5559717,
274
+ 5559723,
275
+ 5559727
276
  ],
277
  "block_size": 1024,
278
  "bos_token_id": 50256,
279
  "embd_pdrop": 0.1,
280
  "eos_token_id": 50256,
281
  "initializer_range": 0.02,
282
+ "inner_step": 0,
283
  "inner_steps": 0,
284
+ "last_allreduce_block": 5559633,
285
  "layer_norm_epsilon": 1e-05,
286
  "model_type": "gpt_optimized",
287
  "n_embd": 1280,
inner_optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:773d6ed23dfeb7c275849725c8b4b2129d3c550fbfd47398e3ac59db49dcd47f
3
  size 8081782503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e35ae2857fe712e85a356c3e5c9737ea298b698b7d6eeff4f0c0c9e85cdc65d8
3
  size 8081782503
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9fe7dcad5c25482395f3d4000eac32a9bb732263682d633bb7862eca8f447af
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2981500704542d902389fe986676a2269fb03e83ae68abe9ee34df0b10360aa
3
  size 4040701744