Run 3. Outer Step 8. Inner Step 0.
Browse files- config.json +20 -20
- inner_optimizer.pt +1 -1
- model.safetensors +1 -1
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"activation_function": "gelu_new",
|
| 4 |
"all_reduce_scores": {
|
| 5 |
"0": "NON_PARTICIPATING",
|
|
@@ -115,7 +115,7 @@
|
|
| 115 |
"198": "NON_PARTICIPATING",
|
| 116 |
"199": "NON_PARTICIPATING",
|
| 117 |
"2": "SUCCESS",
|
| 118 |
-
"20": "
|
| 119 |
"200": "NON_PARTICIPATING",
|
| 120 |
"201": "NON_PARTICIPATING",
|
| 121 |
"202": "NON_PARTICIPATING",
|
|
@@ -137,7 +137,7 @@
|
|
| 137 |
"217": "NON_PARTICIPATING",
|
| 138 |
"218": "NON_PARTICIPATING",
|
| 139 |
"219": "NON_PARTICIPATING",
|
| 140 |
-
"22": "
|
| 141 |
"220": "NON_PARTICIPATING",
|
| 142 |
"221": "NON_PARTICIPATING",
|
| 143 |
"222": "NON_PARTICIPATING",
|
|
@@ -189,14 +189,14 @@
|
|
| 189 |
"34": "NON_PARTICIPATING",
|
| 190 |
"35": "NON_PARTICIPATING",
|
| 191 |
"36": "NON_PARTICIPATING",
|
| 192 |
-
"37": "
|
| 193 |
"38": "NON_PARTICIPATING",
|
| 194 |
-
"39": "
|
| 195 |
"4": "NON_PARTICIPATING",
|
| 196 |
"40": "NON_PARTICIPATING",
|
| 197 |
-
"41": "
|
| 198 |
"42": "NON_PARTICIPATING",
|
| 199 |
-
"43": "
|
| 200 |
"44": "SUCCESS",
|
| 201 |
"45": "NON_PARTICIPATING",
|
| 202 |
"46": "NON_PARTICIPATING",
|
|
@@ -217,17 +217,17 @@
|
|
| 217 |
"6": "NON_PARTICIPATING",
|
| 218 |
"60": "NON_PARTICIPATING",
|
| 219 |
"61": "NON_PARTICIPATING",
|
| 220 |
-
"62": "
|
| 221 |
"63": "NON_PARTICIPATING",
|
| 222 |
"64": "NON_PARTICIPATING",
|
| 223 |
"65": "NON_PARTICIPATING",
|
| 224 |
"66": "NON_PARTICIPATING",
|
| 225 |
-
"67": "
|
| 226 |
"68": "NON_PARTICIPATING",
|
| 227 |
"69": "NON_PARTICIPATING",
|
| 228 |
"7": "NON_PARTICIPATING",
|
| 229 |
"70": "NON_PARTICIPATING",
|
| 230 |
-
"71": "
|
| 231 |
"72": "NON_PARTICIPATING",
|
| 232 |
"73": "NON_PARTICIPATING",
|
| 233 |
"74": "NON_PARTICIPATING",
|
|
@@ -237,11 +237,11 @@
|
|
| 237 |
"78": "NON_PARTICIPATING",
|
| 238 |
"79": "NON_PARTICIPATING",
|
| 239 |
"8": "NON_PARTICIPATING",
|
| 240 |
-
"80": "
|
| 241 |
"81": "NON_PARTICIPATING",
|
| 242 |
"82": "NON_PARTICIPATING",
|
| 243 |
"83": "NON_PARTICIPATING",
|
| 244 |
-
"84": "
|
| 245 |
"85": "NON_PARTICIPATING",
|
| 246 |
"86": "NON_PARTICIPATING",
|
| 247 |
"87": "NON_PARTICIPATING",
|
|
@@ -256,7 +256,7 @@
|
|
| 256 |
"95": "NON_PARTICIPATING",
|
| 257 |
"96": "SUCCESS",
|
| 258 |
"97": "NON_PARTICIPATING",
|
| 259 |
-
"98": "
|
| 260 |
"99": "NON_PARTICIPATING"
|
| 261 |
},
|
| 262 |
"architectures": [
|
|
@@ -268,20 +268,20 @@
|
|
| 268 |
"AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
|
| 269 |
},
|
| 270 |
"block_list": [
|
| 271 |
-
|
| 272 |
-
|
| 273 |
-
|
| 274 |
-
|
| 275 |
-
|
| 276 |
],
|
| 277 |
"block_size": 1024,
|
| 278 |
"bos_token_id": 50256,
|
| 279 |
"embd_pdrop": 0.1,
|
| 280 |
"eos_token_id": 50256,
|
| 281 |
"initializer_range": 0.02,
|
| 282 |
-
"inner_step":
|
| 283 |
"inner_steps": 0,
|
| 284 |
-
"last_allreduce_block":
|
| 285 |
"layer_norm_epsilon": 1e-05,
|
| 286 |
"model_type": "gpt_optimized",
|
| 287 |
"n_embd": 1280,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "distributed/optimized-gpt2-1b",
|
| 3 |
"activation_function": "gelu_new",
|
| 4 |
"all_reduce_scores": {
|
| 5 |
"0": "NON_PARTICIPATING",
|
|
|
|
| 115 |
"198": "NON_PARTICIPATING",
|
| 116 |
"199": "NON_PARTICIPATING",
|
| 117 |
"2": "SUCCESS",
|
| 118 |
+
"20": "NON_PARTICIPATING",
|
| 119 |
"200": "NON_PARTICIPATING",
|
| 120 |
"201": "NON_PARTICIPATING",
|
| 121 |
"202": "NON_PARTICIPATING",
|
|
|
|
| 137 |
"217": "NON_PARTICIPATING",
|
| 138 |
"218": "NON_PARTICIPATING",
|
| 139 |
"219": "NON_PARTICIPATING",
|
| 140 |
+
"22": "NON_PARTICIPATING",
|
| 141 |
"220": "NON_PARTICIPATING",
|
| 142 |
"221": "NON_PARTICIPATING",
|
| 143 |
"222": "NON_PARTICIPATING",
|
|
|
|
| 189 |
"34": "NON_PARTICIPATING",
|
| 190 |
"35": "NON_PARTICIPATING",
|
| 191 |
"36": "NON_PARTICIPATING",
|
| 192 |
+
"37": "NON_PARTICIPATING",
|
| 193 |
"38": "NON_PARTICIPATING",
|
| 194 |
+
"39": "NON_PARTICIPATING",
|
| 195 |
"4": "NON_PARTICIPATING",
|
| 196 |
"40": "NON_PARTICIPATING",
|
| 197 |
+
"41": "SUCCESS",
|
| 198 |
"42": "NON_PARTICIPATING",
|
| 199 |
+
"43": "SUCCESS",
|
| 200 |
"44": "SUCCESS",
|
| 201 |
"45": "NON_PARTICIPATING",
|
| 202 |
"46": "NON_PARTICIPATING",
|
|
|
|
| 217 |
"6": "NON_PARTICIPATING",
|
| 218 |
"60": "NON_PARTICIPATING",
|
| 219 |
"61": "NON_PARTICIPATING",
|
| 220 |
+
"62": "NON_PARTICIPATING",
|
| 221 |
"63": "NON_PARTICIPATING",
|
| 222 |
"64": "NON_PARTICIPATING",
|
| 223 |
"65": "NON_PARTICIPATING",
|
| 224 |
"66": "NON_PARTICIPATING",
|
| 225 |
+
"67": "SUCCESS",
|
| 226 |
"68": "NON_PARTICIPATING",
|
| 227 |
"69": "NON_PARTICIPATING",
|
| 228 |
"7": "NON_PARTICIPATING",
|
| 229 |
"70": "NON_PARTICIPATING",
|
| 230 |
+
"71": "NON_PARTICIPATING",
|
| 231 |
"72": "NON_PARTICIPATING",
|
| 232 |
"73": "NON_PARTICIPATING",
|
| 233 |
"74": "NON_PARTICIPATING",
|
|
|
|
| 237 |
"78": "NON_PARTICIPATING",
|
| 238 |
"79": "NON_PARTICIPATING",
|
| 239 |
"8": "NON_PARTICIPATING",
|
| 240 |
+
"80": "NON_PARTICIPATING",
|
| 241 |
"81": "NON_PARTICIPATING",
|
| 242 |
"82": "NON_PARTICIPATING",
|
| 243 |
"83": "NON_PARTICIPATING",
|
| 244 |
+
"84": "NON_PARTICIPATING",
|
| 245 |
"85": "NON_PARTICIPATING",
|
| 246 |
"86": "NON_PARTICIPATING",
|
| 247 |
"87": "NON_PARTICIPATING",
|
|
|
|
| 256 |
"95": "NON_PARTICIPATING",
|
| 257 |
"96": "SUCCESS",
|
| 258 |
"97": "NON_PARTICIPATING",
|
| 259 |
+
"98": "NON_PARTICIPATING",
|
| 260 |
"99": "NON_PARTICIPATING"
|
| 261 |
},
|
| 262 |
"architectures": [
|
|
|
|
| 268 |
"AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
|
| 269 |
},
|
| 270 |
"block_list": [
|
| 271 |
+
5395806,
|
| 272 |
+
5395810,
|
| 273 |
+
5395815,
|
| 274 |
+
5395819,
|
| 275 |
+
5395823
|
| 276 |
],
|
| 277 |
"block_size": 1024,
|
| 278 |
"bos_token_id": 50256,
|
| 279 |
"embd_pdrop": 0.1,
|
| 280 |
"eos_token_id": 50256,
|
| 281 |
"initializer_range": 0.02,
|
| 282 |
+
"inner_step": 0,
|
| 283 |
"inner_steps": 0,
|
| 284 |
+
"last_allreduce_block": 5395632,
|
| 285 |
"layer_norm_epsilon": 1e-05,
|
| 286 |
"model_type": "gpt_optimized",
|
| 287 |
"n_embd": 1280,
|
inner_optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 8081782026
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a9de59582f1efe499f7f8c2f7a2f7264304beba633b0bb929e53fe31d401577
|
| 3 |
size 8081782026
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4040701744
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22f339a6f66964efc25710cefa4bf7100e0613a2b429de3960b6d9fa873a1c3a
|
| 3 |
size 4040701744
|