MaxLSB commited on
Commit
f03f562
·
verified ·
1 Parent(s): 6e6bbe5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -34,7 +34,7 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
34
 
35
  | Model | IFEval<br>French | GPQA-Diamond<br>French | MMLU<br>French | Math500<br>French | Arc-Challenge<br>French | Hellaswag<br>French |
36
  |------------------------|-----------------|-----------------------|----------------|-----------------|------------------------|-------------------|
37
- | Luth-0.6B-Instruct | <u>48.24</u> | <u>34.52</u> | <u>40.12</u> | <u>44.00</u> | <u>33.88</u> | 45.58 |
38
  | Llama-3.2-1B | 27.79 | 25.38 | 25.49 | 15.80 | 29.34 | 25.09 |
39
  | Qwen3-0.6B | 44.86 | 26.90 | 27.13 | 29.20 | 31.57 | 25.10 |
40
  | Qwen2.5-0.5B-Instruct | 22.00 | 25.89 | 35.04 | 12.00 | 28.23 | <u>51.45</u> |
@@ -44,7 +44,7 @@ We used LightEval for evaluation, with custom tasks for the French benchmarks. T
44
 
45
  | Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
46
  |------------------------|-----------------|------------------------|----------------|------------------|-------------------------|--------------------|
47
- | Luth-0.6B-Instruct | 53.73 | 25.76 | <u>48.12</u> | <u>48.80</u> | <u>36.09</u> | 47.03 |
48
  | Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
49
  | Qwen3-0.6B | <u>57.18</u> | <u>29.29</u> | 36.79 | 43.40 | 33.70 | 42.92 |
50
  | Qwen2.5-0.5B-Instruct | 29.70 | <u>29.29</u> | 43.80 | 32.00 | 32.17 | 49.56 |
 
34
 
35
  | Model | IFEval<br>French | GPQA-Diamond<br>French | MMLU<br>French | Math500<br>French | Arc-Challenge<br>French | Hellaswag<br>French |
36
  |------------------------|-----------------|-----------------------|----------------|-----------------|------------------------|-------------------|
37
+ | **Luth-0.6B-Instruct** | <u>48.24</u> | <u>34.52</u> | <u>40.12</u> | <u>44.00</u> | <u>33.88</u> | 45.58 |
38
  | Llama-3.2-1B | 27.79 | 25.38 | 25.49 | 15.80 | 29.34 | 25.09 |
39
  | Qwen3-0.6B | 44.86 | 26.90 | 27.13 | 29.20 | 31.57 | 25.10 |
40
  | Qwen2.5-0.5B-Instruct | 22.00 | 25.89 | 35.04 | 12.00 | 28.23 | <u>51.45</u> |
 
44
 
45
  | Model | IFEval<br>English | GPQA-Diamond<br>English | MMLU<br>English | Math500<br>English | Arc-Challenge<br>English | Hellaswag<br>English |
46
  |------------------------|-----------------|------------------------|----------------|------------------|-------------------------|--------------------|
47
+ | **Luth-0.6B-Instruct** | 53.73 | 25.76 | <u>48.12</u> | <u>48.80</u> | <u>36.09</u> | 47.03 |
48
  | Llama-3.2-1B | 44.05 | 25.25 | 31.02 | 26.40 | 34.30 | <u>55.84</u> |
49
  | Qwen3-0.6B | <u>57.18</u> | <u>29.29</u> | 36.79 | 43.40 | 33.70 | 42.92 |
50
  | Qwen2.5-0.5B-Instruct | 29.70 | <u>29.29</u> | 43.80 | 32.00 | 32.17 | 49.56 |