juyoung-trl tl-hyungguk commited on
Commit
94e2cc8
·
verified ·
1 Parent(s): 21f21e7

Update README.md (#1)

Browse files

- Update README.md (654d09e3818e917c081334a9faf923f88ba51fc8)


Co-authored-by: hyungguk kim <tl-hyungguk@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -138,8 +138,8 @@ We evaluated Tri-7B across a comprehensive suite of benchmarks assessing general
138
  | • CLIcK | Korean | 0-shot | accuracy |
139
  | • KoBEST | Korean | 5-shot | accuracy |
140
  | **Knowledge and Reasoning** | | | |
141
- | • KMMLU | Korean | 5-shot | accuracy |
142
- | • MMLU | English | 5-shot | accuracy |
143
  | • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
144
  | **Coding** | | | |
145
  | • HumanEval | English | 0-shot | pass@1 |
@@ -176,8 +176,8 @@ Models compared:
176
  | HAERAE | 82.49 | 80.02 | +2.47 |
177
  | KoBEST | 82.72 | 79.61 | +3.11 |
178
  | CLIcK | 64.43 | 60.41 | +4.02 |
179
- | KMMLU | 51.74 | 48.09 | +3.65 |
180
- | MMLU | 68.16 | 63.52 | +4.64 |
181
  | Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
182
 
183
 
 
138
  | • CLIcK | Korean | 0-shot | accuracy |
139
  | • KoBEST | Korean | 5-shot | accuracy |
140
  | **Knowledge and Reasoning** | | | |
141
+ | • KMMLU | Korean | 5-shot (0-shot, CoT) | accuracy |
142
+ | • MMLU | English | 5-shot (0-shot, CoT) | accuracy |
143
  | • Global-MMLU-Lite-ja | English | 5-shot | accuracy |
144
  | **Coding** | | | |
145
  | • HumanEval | English | 0-shot | pass@1 |
 
176
  | HAERAE | 82.49 | 80.02 | +2.47 |
177
  | KoBEST | 82.72 | 79.61 | +3.11 |
178
  | CLIcK | 64.43 | 60.41 | +4.02 |
179
+ | KMMLU | 51.74 (53.51) | 48.09 | +3.65 |
180
+ | MMLU | 68.16 (74.67) | 63.52 | +4.64 |
181
  | Global-MMLU-Lite-ja | 59.25 | 60.75 | -1.50 |
182
 
183