Update README.md
Browse files
README.md
CHANGED
|
@@ -30,8 +30,8 @@ This model was evaluated using the `lm-evaluation-harness` against OpenAI's GPT-
|
|
| 30 |
| **HellaSwag** (acc_norm) | **27.00%** | 31.14% | **86.7%** |
|
| 31 |
|
| 32 |
> **Key Takeaway:** With only **12% of the parameters**, this model achieves over **80% of the reasoning performance** of GPT-2, proving that modern architectures combined with curated data can drastically reduce model size.
|
| 33 |
-
|
| 34 |
-
|
| 35 |
|
| 36 |
The model is based on the **Llama-2 architecture** with several modern optimizations:
|
| 37 |
|
|
|
|
| 30 |
| **HellaSwag** (acc_norm) | **27.00%** | 31.14% | **86.7%** |
|
| 31 |
|
| 32 |
> **Key Takeaway:** With only **12% of the parameters**, this model achieves over **80% of the reasoning performance** of GPT-2, proving that modern architectures combined with curated data can drastically reduce model size.
|
| 33 |
+
|
| 34 |
+
## Model Architecture
|
| 35 |
|
| 36 |
The model is based on the **Llama-2 architecture** with several modern optimizations:
|
| 37 |
|