Upload folder using huggingface_hub
Browse files- README.md +2 -2
- checkpoints/checkpoint_17000_step.pth +3 -0
README.md
CHANGED
|
@@ -45,8 +45,8 @@ Due to the sparse nature of the MoE layers, approximately **43% of the parameter
|
|
| 45 |
|
| 46 |
Training is currently in progress on a single NVIDIA RTX 5000.
|
| 47 |
|
| 48 |
-
* **Current Step:**
|
| 49 |
-
* **Current Loss:** ~
|
| 50 |
* **Dataset:** A subset of SlimPajama-627B
|
| 51 |
|
| 52 |
> **⚠️ Performance Notice:** Aetheris is currently in a "proof-of-concept" state. While it generates coherent English syntax, it does not yet possess strong reasoning capabilities. It is intended for architectural analysis rather than downstream tasks.
|
|
|
|
| 45 |
|
| 46 |
Training is currently in progress on a single NVIDIA RTX 5000.
|
| 47 |
|
| 48 |
+
* **Current Step:** 11,000
|
| 49 |
+
* **Current Loss:** ~1.4167
|
| 50 |
* **Dataset:** A subset of SlimPajama-627B
|
| 51 |
|
| 52 |
> **⚠️ Performance Notice:** Aetheris is currently in a "proof-of-concept" state. While it generates coherent English syntax, it does not yet possess strong reasoning capabilities. It is intended for architectural analysis rather than downstream tasks.
|
checkpoints/checkpoint_17000_step.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:86130f7b206e603a0777d82ce482d5cd08fd3168cdd5dd0a106fc71a145479f7
|
| 3 |
+
size 3533562641
|