Run 10. Outer Step 3. Inner Step 108.
Browse files- config.json +39 -3
- gradients.pt +1 -1
- inner_optimizer.pt +1 -1
- model.safetensors +1 -1
config.json
CHANGED
|
@@ -264,15 +264,51 @@
|
|
| 264 |
"attention_bias": false,
|
| 265 |
"attention_dropout": 0.0,
|
| 266 |
"block_list": [
|
| 267 |
-
|
| 268 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 269 |
],
|
| 270 |
"bos_token_id": 1,
|
| 271 |
"eos_token_id": 2,
|
| 272 |
"hidden_act": "silu",
|
| 273 |
"hidden_size": 2048,
|
| 274 |
"initializer_range": 0.02,
|
| 275 |
-
"inner_step":
|
| 276 |
"intermediate_size": 5632,
|
| 277 |
"last_allreduce_block": 6464980,
|
| 278 |
"max_position_embeddings": 2048,
|
|
|
|
| 264 |
"attention_bias": false,
|
| 265 |
"attention_dropout": 0.0,
|
| 266 |
"block_list": [
|
| 267 |
+
6465120,
|
| 268 |
+
6465134,
|
| 269 |
+
6465143,
|
| 270 |
+
6465153,
|
| 271 |
+
6465166,
|
| 272 |
+
6465179,
|
| 273 |
+
6465188,
|
| 274 |
+
6465200,
|
| 275 |
+
6465217,
|
| 276 |
+
6465233,
|
| 277 |
+
6465243,
|
| 278 |
+
6465258,
|
| 279 |
+
6465271,
|
| 280 |
+
6465284,
|
| 281 |
+
6465296,
|
| 282 |
+
6465310,
|
| 283 |
+
6465326,
|
| 284 |
+
6465337,
|
| 285 |
+
6465352,
|
| 286 |
+
6465369,
|
| 287 |
+
6465381,
|
| 288 |
+
6465392,
|
| 289 |
+
6465407,
|
| 290 |
+
6465423,
|
| 291 |
+
6465437,
|
| 292 |
+
6465448,
|
| 293 |
+
6465463,
|
| 294 |
+
6465477,
|
| 295 |
+
6465488,
|
| 296 |
+
6465505,
|
| 297 |
+
6465517,
|
| 298 |
+
6465528,
|
| 299 |
+
6465541,
|
| 300 |
+
6465556,
|
| 301 |
+
6465571,
|
| 302 |
+
6465580,
|
| 303 |
+
6465593,
|
| 304 |
+
6465607
|
| 305 |
],
|
| 306 |
"bos_token_id": 1,
|
| 307 |
"eos_token_id": 2,
|
| 308 |
"hidden_act": "silu",
|
| 309 |
"hidden_size": 2048,
|
| 310 |
"initializer_range": 0.02,
|
| 311 |
+
"inner_step": 107,
|
| 312 |
"intermediate_size": 5632,
|
| 313 |
"last_allreduce_block": 6464980,
|
| 314 |
"max_position_embeddings": 2048,
|
gradients.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 88389626
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b9943caf01bc4938dd07ced7ffec5b91af7683247202a563e61db86923ac03a
|
| 3 |
size 88389626
|
inner_optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4924926906
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e427eb170d887378316c6616fb6c3fa36bd53f74b326e29b353549cd26567dfc
|
| 3 |
size 4924926906
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4400216536
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:80bd11952ba6f31e452c6a7b677630f7b9aaabe85a98e7f113a9474e59bac983
|
| 3 |
size 4400216536
|