MangoLab commited on
Commit
f7c532c
ยท
verified ยท
1 Parent(s): 0af086d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -28
README.md CHANGED
@@ -17,31 +17,3 @@ configs:
17
  - split: train
18
  path: mmlu_refined.csv
19
  ---
20
-
21
- # ๐Ÿš€ EXAONE-4.0-1.2B-Quantized-MMLU Evaluation Results
22
-
23
- This repository contains the refined evaluation results for the **EXAONE-4.0-1.2B-GPTQ** model using the MMLU benchmark.
24
-
25
- ## ๐Ÿ“Š Overview
26
- The evaluation was conducted to measure the model's multitask language understanding capabilities across 57 different subjects. To ensure stability and formatting consistency, we used a **5-shot** prompting approach.
27
-
28
- ### ๐Ÿ’ป Hardware & Software
29
- - **Model:** [MangoLab/EXAONE-4.0-1.2B-GPTQ]
30
- - **Accelerator:** AMD Instinct MI325 OAM
31
- - **Quantization:** GPTQ (W8A16 / fp8_w8a8)
32
- - **Framework:** `lm-evaluation-harness`
33
-
34
- ## ๐Ÿ“ˆ Performance Summary
35
- The table below shows a summary of the evaluation environment. Detailed per-subject accuracy can be viewed in the **Dataset Viewer** above.
36
-
37
- | Category | Details |
38
- | :--- | :--- |
39
- | **Benchmark** | MMLU (Massive Multitask Language Understanding) |
40
- | **Prompting** | 5-shot |
41
- | **Evaluation Date** | 2026-01-24 |
42
-
43
- ## ๐Ÿ“‚ Directory Structure
44
- - `mmlu_refined.csv`: The main result file containing subject-wise accuracy (%).
45
- - `raw_data/`: Original JSON output files from the evaluation process for reproducibility.
46
-
47
- ---
 
17
  - split: train
18
  path: mmlu_refined.csv
19
  ---