Update README.md
Browse files
README.md
CHANGED
|
@@ -20,7 +20,7 @@ library_name: transformers
|
|
| 20 |
|
| 21 |
**Monad** is a 56 million parameters generalist Small Reasoning Model, trained on 200 billions tokens from <a href="https://huggingface.co/PleIAs/Baguettotron">SYNTH</a>, a fully open generalist dataset.
|
| 22 |
|
| 23 |
-
As of 2025, Monad is the best contender for the smallest viable language models. Despite being less than half of gpt-2, Monad not answers in consistent English but
|
| 24 |
|
| 25 |
<p align="center">
|
| 26 |
<img width="80%" src="figures/training_efficiency.jpeg">
|
|
|
|
| 20 |
|
| 21 |
**Monad** is a 56 million parameters generalist Small Reasoning Model, trained on 200 billions tokens from <a href="https://huggingface.co/PleIAs/Baguettotron">SYNTH</a>, a fully open generalist dataset.
|
| 22 |
|
| 23 |
+
As of 2025, Monad is the best contender for the smallest viable language models. Despite being less than half of gpt-2, Monad not only answers in consistent English but performs significanly beyond chance on MMLU and other major industry benchmarks.
|
| 24 |
|
| 25 |
<p align="center">
|
| 26 |
<img width="80%" src="figures/training_efficiency.jpeg">
|