Update README.md
Browse files
README.md
CHANGED
|
@@ -34,7 +34,7 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
|
|
| 34 |
|
| 35 |
| Model | IFEval<br>French | GPQA-Diamond<br>French | MMLU<br>French | Math500<br>French | Arc-Challenge<br>French | Hellaswag<br>French |
|
| 36 |
|------------------------|-----------------|-----------------------|----------------|-----------------|------------------------|-------------------|
|
| 37 |
-
| Luth-0.6B-Instruct | <u>48.24</u> | <u>34.52</u> | <u>40.12</u> | <u>44.00</u> | <u>33.88</u> | 45.58 |
|
| 38 |
| Llama-3.2-1B | 27.79 | 25.38 | 25.49 | 15.80 | 29.34 | 25.09 |
|
| 39 |
| Qwen3-0.6B | 44.86 | 26.90 | 27.13 | 29.20 | 31.57 | 25.10 |
|
| 40 |
| Qwen2.5-0.5B-Instruct | 22.00 | 25.89 | 35.04 | 12.00 | 28.23 | <u>51.45</u> |
|
|
@@ -44,7 +44,7 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
|
|
| 44 |
|
| 45 |
| Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
|
| 46 |
|------------------------|-----------------|------------------------|----------------|------------------|-------------------------|--------------------|
|
| 47 |
-
| Luth-0.6B-Instruct | 53.73 | 25.76 | <u>48.12</u> | <u>48.80</u> | <u>36.09</u> | 47.03 |
|
| 48 |
| Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
|
| 49 |
| Qwen3-0.6B | <u>57.18</u> | <u>29.29</u> | 36.79 | 43.40 | 33.70 | 42.92 |
|
| 50 |
| Qwen2.5-0.5B-Instruct | 29.70 | <u>29.29</u> | 43.80 | 32.00 | 32.17 | 49.56 |
|
|
|
|
| 34 |
|
| 35 |
| Model | IFEval<br>French | GPQA-Diamond<br>French | MMLU<br>French | Math500<br>French | Arc-Challenge<br>French | Hellaswag<br>French |
|
| 36 |
|------------------------|-----------------|-----------------------|----------------|-----------------|------------------------|-------------------|
|
| 37 |
+
| **Luth-0.6B-Instruct** | <u>48.24</u> | <u>34.52</u> | <u>40.12</u> | <u>44.00</u> | <u>33.88</u> | 45.58 |
|
| 38 |
| Llama-3.2-1B | 27.79 | 25.38 | 25.49 | 15.80 | 29.34 | 25.09 |
|
| 39 |
| Qwen3-0.6B | 44.86 | 26.90 | 27.13 | 29.20 | 31.57 | 25.10 |
|
| 40 |
| Qwen2.5-0.5B-Instruct | 22.00 | 25.89 | 35.04 | 12.00 | 28.23 | <u>51.45</u> |
|
|
|
|
| 44 |
|
| 45 |
| Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
|
| 46 |
|------------------------|-----------------|------------------------|----------------|------------------|-------------------------|--------------------|
|
| 47 |
+
| **Luth-0.6B-Instruct** | 53.73 | 25.76 | <u>48.12</u> | <u>48.80</u> | <u>36.09</u> | 47.03 |
|
| 48 |
| Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
|
| 49 |
| Qwen3-0.6B | <u>57.18</u> | <u>29.29</u> | 36.79 | 43.40 | 33.70 | 42.92 |
|
| 50 |
| Qwen2.5-0.5B-Instruct | 29.70 | <u>29.29</u> | 43.80 | 32.00 | 32.17 | 49.56 |
|