Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -138,8 +138,8 @@ We evaluated Tri-7B across a comprehensive suite of benchmarks assessing general
138
  | • CLIcK | Korean | 0-shot | accuracy |
139
  | • KoBEST | Korean | 5-shot | accuracy |
140
  | **Knowledge and Reasoning** | | | |
141
- | • KMMLU | Korean | 5-shot | accuracy |
142
- | • MMLU | English | 5-shot | accuracy |
143
  | • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
144
  | **Coding** | | | |
145
  | • HumanEval | English | 0-shot | pass@1 |
@@ -176,8 +176,8 @@ Models compared:
176
  | HAERAE | 82.49 | 80.02 | +2.47 |
177
  | KoBEST | 82.72 | 79.61 | +3.11 |
178
  | CLIcK | 64.43 | 60.41 | +4.02 |
179
- | KMMLU | 51.74 | 48.09 | +3.65 |
180
- | MMLU | 68.16 | 63.52 | +4.64 |
181
  | Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
182
 
183
 
 
138
  | • CLIcK | Korean | 0-shot | accuracy |
139
  | • KoBEST | Korean | 5-shot | accuracy |
140
  | **Knowledge and Reasoning** | | | |
141
+ | • KMMLU | Korean | 5-shot (0-shot, CoT) | accuracy |
142
+ | • MMLU | English | 5-shot (0-shot, CoT) | accuracy |
143
  | • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
144
  | **Coding** | | | |
145
  | • HumanEval | English | 0-shot | pass@1 |
 
176
  | HAERAE | 82.49 | 80.02 | +2.47 |
177
  | KoBEST | 82.72 | 79.61 | +3.11 |
178
  | CLIcK | 64.43 | 60.41 | +4.02 |
179
+ | KMMLU | 51.74 (53.51) | 48.09 | +3.65 |
180
+ | MMLU | 68.16 (74.67) | 63.52 | +4.64 |
181
  | Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
182
 
183