Update README.md
Browse files
README.md
CHANGED
|
@@ -10,31 +10,28 @@ tags:
|
|
| 10 |
- sft
|
| 11 |
---
|
| 12 |
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
**
|
| 18 |
-
|
| 19 |
-
**
|
| 20 |
-
|
| 21 |
-
**
|
| 22 |
-
|
| 23 |
-
**
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
**V1: 20.40 %**
|
| 36 |
-
**V2: 20.65 %**
|
| 37 |
-
----------------
|
| 38 |
# Gladiator-Mini-exp-1221-Instruct
|
| 39 |
|
| 40 |
**Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.
|
|
|
|
| 10 |
- sft
|
| 11 |
---
|
| 12 |
|
| 13 |
+
## Gladiator-Mini-Exp-1221-3B-Instruct - V2: Enhanced Performance
|
| 14 |
+
|
| 15 |
+
**This is V2, an improved iteration of our Gladiator-Mini-Exp-1221-3B-Instruct model, fine-tuned from Llama-3.2-3B-Instruct with a slightly expanded dataset and increased training epochs.**
|
| 16 |
+
|
| 17 |
+
**Major Improvements in V2:**
|
| 18 |
+
|
| 19 |
+
* **Superior Performance:** V2 demonstrates enhanced performance across multiple benchmarks compared to V1.
|
| 20 |
+
* **Mathematics Boost:** Notably, V2 surpasses both the 1211 model and V1 in mathematical reasoning, as evidenced by the MATH benchmark results below.
|
| 21 |
+
* **Model Size:** 3.21 Billion parameters
|
| 22 |
+
|
| 23 |
+
**Benchmark Highlights:**
|
| 24 |
+
|
| 25 |
+
| Benchmark | 1211 | V1 | V2 |
|
| 26 |
+
| :-------- | :-------- | :-------- | :-------- |
|
| 27 |
+
| **MATH** | **13.44%** | **13.07%** | **13.75%** |
|
| 28 |
+
| IFEval | | 60.79% | 62.15% |
|
| 29 |
+
| BBH | | 20.40% | 20.65% |
|
| 30 |
+
|
| 31 |
+
[Link to V1](https://huggingface.co/MultivexAI/Gladiator-Mini-Exp-1221-3B-Instruct)
|
| 32 |
+
|
| 33 |
+
**In summary, V2 offers a noticeable performance upgrade over V1, particularly in mathematical tasks. Explore the model and experience the improvements!**
|
| 34 |
+
|
|
|
|
|
|
|
|
|
|
| 35 |
# Gladiator-Mini-exp-1221-Instruct
|
| 36 |
|
| 37 |
**Gladiator-Mini-exp-1221** is a 3-billion parameter language model focused on **complex reasoning**. Built upon the foundation of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct), this experimental model is designed to explore what's achievable with smaller models in analytical thinking. It's all about pushing boundaries and learning what's possible in resource-efficient AI. We believe small models represent the future of open source language models, making AI more accessible and adaptable for a wider range of users and applications.
|