Golden Checkpoint: Step 1397 (Logic Giant Saturation)
Browse files- config.json +15 -0
- pytorch_model.bin +3 -0
config.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"nacre_tip": "8586164c7d4b4a88e325b9986570f0bf296880586df545c8c445092b5008014a",
|
| 3 |
+
"step": 1397,
|
| 4 |
+
"hash": "8586164c7d4b4a88e325b9986570f0bf296880586df545c8c445092b5008014a",
|
| 5 |
+
"weights_file": "/data/rabbit/scholar-efs/pearl/weights_step1397.pt",
|
| 6 |
+
"checkpoint_digest": "a260f132fb7c74b059178d273c14a2e7258e6c100c5a2f8b57026a39643a068e",
|
| 7 |
+
"checkpoint_layout": "full",
|
| 8 |
+
"base_weights_file": "/data/rabbit/scholar-efs/pearl/weights_step0.pt",
|
| 9 |
+
"surface_file": "",
|
| 10 |
+
"tokenizer_id": "EleutherAI/gpt-neox-20b",
|
| 11 |
+
"tokenizer_hash": "",
|
| 12 |
+
"schema_version": "2.0",
|
| 13 |
+
"gurukul_version": "phase2_hardened",
|
| 14 |
+
"stage_id": "stage00_sanskrit"
|
| 15 |
+
}
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a260f132fb7c74b059178d273c14a2e7258e6c100c5a2f8b57026a39643a068e
|
| 3 |
+
size 12264396835
|