File size: 1,866 Bytes
61ab573
 
2691a4a
61ab573
2691a4a
61ab573
 
 
 
2691a4a
 
 
357d83b
2691a4a
0f79f24
357d83b
 
1b2b3f1
0f79f24
 
 
357d83b
 
1b2b3f1
2691a4a
357d83b
2691a4a
357d83b
88460de
357d83b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
---
title: README
emoji: πŸ”₯
colorFrom: indigo
colorTo: blue
sdk: static
pinned: false
---

# Welcome to HAERAE


We are a non-profit research lab focused on understanding and building better Korean language models. See below for an overview of our projects.

**Benchmarks**  
We have built _the_ most-widely used korean benchmarks including HAE-RAE Bench (cultural knowledge, [dataset](https://huggingface.co/datasets/HAERAE-HUB/HAE_RAE_BENCH_1.0), [paper](https://arxiv.org/abs/2309.02706)), 
KMMLU (general knowledge, [dataset](https://huggingface.co/datasets/HAERAE-HUB/KMMLU), [paper](https://arxiv.org/abs/2402.11548)), HRM8K (math, [dataset](https://huggingface.co/datasets/HAERAE-HUB/HRM8K), [paper](https://www.arxiv.org/abs/2501.02448)), and KMMLU-Redux/Pro  (general knowledge, [dataset](https://huggingface.co/datasets/LGAI-EXAONE/KMMLU-Pro), [paper](https://arxiv.org/abs/2507.08924)).

**Evaluation**  
We developed the [haerae-evaluation-toolkit](https://github.com/HAE-RAE/haerae-evaluation-toolkit), a unified LLM evaluation framework designed to provide consistent and reproducible benchmarking for Korean and multilingual models.

**Reasoning Language Models**  
With cooperation with [KISTI-KONI](https://huggingface.co/KISTI-KONI) we released the [KO-REAson](https://huggingface.co/KOREAson) series, <10B reasoning language models trained for Korean.


# News

2025.08.31: We release six [KO-REAson-0831 models](https://huggingface.co/collections/KoReason/koreason-0831-68b1363e1b3726b041a0a638)  πŸ”₯πŸ”₯πŸ”₯  
2025.07.11: We've collaborated with LG AI Research to build [KMMLU-PRO](https://huggingface.co/datasets/LGAI-EXAONE/KMMLU-Pro) an major update to our KMMLU franchise.  
2025.01.05: We are releasing the first public korean math (πŸ“e = βˆ‘βˆžβΏβΌβ° ΒΉβ‚™πŸ€“) benchmark [HRM8K](https://huggingface.co/datasets/HAERAE-HUB/HRM8K)