DirectionAI-BIT commited on
Commit
c868e73
·
verified ·
1 Parent(s): 09644de

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -3
README.md CHANGED
@@ -1,3 +1,54 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ metrics:
6
+ - name: accuracy
7
+ valud: 75.39
8
+ base_model:
9
+ - Qwen/Qwen3-Reranker-0.6B
10
+ pipeline_tag: text-classification
11
+ ---
12
+
13
+ # EduBenchEvaluator
14
+
15
+ This is a fine-tuned evaluator designed to assess LLM on the **EduBench** benchmark.
16
+
17
+ - 📄 **[Paper](https://arxiv.org/abs/2505.16160)**
18
+ - 💻 **[GitHub Repository](https://github.com/DIRECT-BIT/EduBench)**
19
+
20
+ ## Model Details
21
+
22
+ * **Model Name**: EduBenchEvaluator
23
+ * **Model Type**: Fine-tuned language model (0.6B parameters)
24
+ * **Base Model**: [Qwen3-Reranker-0.6B](https://huggingface.co/Qwen/Qwen3-Reranker-0.6B)
25
+
26
+ ## Training & Methodology
27
+
28
+ The base model, `Qwen3-Reranker-0.6B`, was fine-tuned to align with human evaluations on the EduBench dataset.
29
+
30
+ We approached the fine-tuning process as a text classification task. The model evaluates a given response by taking a `<question, answer, metric>` triplet as input. Based on this context, it is trained to output a precise evaluation score ranging from **1 to 5**.
31
+
32
+ This evaluator is specifically constructed to measure an LLM's capability across the diverse educational tasks presented in EduBench.
33
+
34
+ ## Performance
35
+
36
+ * **Accuracy**: The model achieves a satisfactory accuracy of **75.28%** on the test set.
37
+ * **Human Alignment**: In addition to standard accuracy, we calculated the correlation between the model's predictions and actual human scorers, demonstrating that the model closely mirrors human judgment.
38
+
39
+ *Note: Further evaluation results and comparisons are reported on our [GitHub](https://github.com/DIRECT-BIT/EduBench).*
40
+
41
+ ## 🫣 Citation
42
+
43
+ If you find our benchmark, evaluation pipeline, or models useful or interesting, please cite our paper:
44
+
45
+ ```bibtex
46
+ @misc{xu2025edubenchcomprehensivebenchmarkingdataset,
47
+ title={EduBench: A Comprehensive Benchmarking Dataset for Evaluating Large Language Models in Diverse Educational Scenarios},
48
+ author={Bin Xu and Yu Bai and Huashan Sun and Yiguan Lin and Siming Liu and Xinyue Liang and Yaolin Li and Yang Gao and Heyan Huang},
49
+ year={2025},
50
+ eprint={2505.16160},
51
+ archivePrefix={arXiv},
52
+ primaryClass={cs.CL},
53
+ url={[https://arxiv.org/abs/2505.16160](https://arxiv.org/abs/2505.16160)},
54
+ }