Update README.md
#1
by
tl-hyungguk
- opened
README.md
CHANGED
|
@@ -138,8 +138,8 @@ We evaluated Tri-7B across a comprehensive suite of benchmarks assessing general
|
|
| 138 |
| • CLIcK | Korean | 0-shot | accuracy |
|
| 139 |
| • KoBEST | Korean | 5-shot | accuracy |
|
| 140 |
| **Knowledge and Reasoning** | | | |
|
| 141 |
-
| • KMMLU | Korean | 5-shot | accuracy |
|
| 142 |
-
| • MMLU | English | 5-shot | accuracy |
|
| 143 |
| • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
|
| 144 |
| **Coding** | | | |
|
| 145 |
| • HumanEval | English | 0-shot | pass@1 |
|
|
@@ -176,8 +176,8 @@ Models compared:
|
|
| 176 |
| HAERAE | 82.49 | 80.02 | +2.47 |
|
| 177 |
| KoBEST | 82.72 | 79.61 | +3.11 |
|
| 178 |
| CLIcK | 64.43 | 60.41 | +4.02 |
|
| 179 |
-
| KMMLU | 51.74 | 48.09 | +3.65 |
|
| 180 |
-
| MMLU | 68.16 | 63.52 | +4.64 |
|
| 181 |
| Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
|
| 182 |
|
| 183 |
|
|
|
|
| 138 |
| • CLIcK | Korean | 0-shot | accuracy |
|
| 139 |
| • KoBEST | Korean | 5-shot | accuracy |
|
| 140 |
| **Knowledge and Reasoning** | | | |
|
| 141 |
+
| • KMMLU | Korean | 5-shot (0-shot, CoT) | accuracy |
|
| 142 |
+
| • MMLU | English | 5-shot (0-shot, CoT) | accuracy |
|
| 143 |
| • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
|
| 144 |
| **Coding** | | | |
|
| 145 |
| • HumanEval | English | 0-shot | pass@1 |
|
|
|
|
| 176 |
| HAERAE | 82.49 | 80.02 | +2.47 |
|
| 177 |
| KoBEST | 82.72 | 79.61 | +3.11 |
|
| 178 |
| CLIcK | 64.43 | 60.41 | +4.02 |
|
| 179 |
+
| KMMLU | 51.74 (53.51) | 48.09 | +3.65 |
|
| 180 |
+
| MMLU | 68.16 (74.67) | 63.52 | +4.64 |
|
| 181 |
| Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
|
| 182 |
|
| 183 |
|