--- license: apache-2.0 --- **Want to fine-tune this dataset on LLaMA-Factory? Check this repository for preprocessing: [llm-merging datasets](https://github.com/Fzkuji/llm-merging/tree/main/datasets)** (Unofficial!) I preprocessed the original data of the CMExam dataset on GitHub so that it can be visualized on huggingface. The dataset loading code for subsequent training (such as LLaMA-Factory) is in the ipynb file in the file directory. --- This paper was presented at NeurIPS 2023, New Orleans, Louisana. See here for the [poster](conference_material/poster.pdf) and [slides](conference_material/presentation.pdf). # Benchmarking Large Language Models on CMExam - A Comprehensive Chinese Medical Exam Dataset ## Introduction CMExam is a dataset sourced from the Chinese National Medical Licensing Examination. It consists of 60K+ multiple-choice questions and five additional question-wise annotations, including disease groups, clinical departments, medical disciplines, areas of competency, and question difficulty levels. Alongside the dataset, comprehensive benchmarks were conducted on representative LLMs on CMExam. example ## Dataset Statistics | | Train | Val | Test | Total | |----------------------------|---------------|---------------|---------------|---------------| | Question | 54,497 | 6,811 | 6,811 | 68,119 | | Vocab | 4,545 | 3,620 | 3,599 | 4,629 | | Max Q tokens | 676 | 500 | 585 | 676 | | Max A tokens | 5 | 5 | 5 | 5 | | Max E tokens | 2,999 | 2,678 | 2,680 | 2,999 | | Avg Q tokens | 29.78 | 30.07 | 32.63 | 30.83 | | Avg A tokens | 1.08 | 1.07 | 1.07 | 1.07 | | Avg E tokens | 186.24 | 188.95 | 201.44 | 192.21 | | Median (Q1, Q3) Q tokens | 17 (12, 32) | 18 (12, 32) | 18 (12, 37) | 18 (12, 32) | | Median (Q1, Q3) A tokens | 1 (1, 1) | 1 (1, 1) | 1 (1, 1) | 1 (1, 1) | | Median (Q1, Q3) E tokens | 146 (69, 246) | 143 (65, 247) | 158 (80, 263) | 146 (69, 247) | \*Q: Question; A: Answer; E: Explanation ## Annotation Characteristics | Annotation Content | References | Unique values | |----------------------------|-----------------------------|---------------| | Disease Groups | The 11th revision of ICD-11 | 27 | | Clinical Departments | The Directory of Medical Institution Diagnostic and Therapeutic Categories (DMIDTC) | 36 | | Medical Disciplines | List of Graduate Education Disciplinary Majors (2022) | 7 | | Medical Competencies | Medical Professionals | 4 | | Difficulty Level | Human Performance | 5 | ## Benchmarks Alongside the dataset, we further conducted thorough experiments with representative LLMs and QA algorithms on CMExam. overall_comparison ## Side notes ### Limitations: - Excluding non-textual questions may introduce biases. - BLEU and ROUGE metrics are inadequate for fully assessing explanations; better expert analysis needed in future. - ### Ethics in Data Collection: - Adheres to legal and ethical guidelines. - Authenticated and accurate for evaluating LLMs. - Intended for academic/research use only; commercial misuse prohibited. - Users should acknowledge dataset limitations and specific context. - Not for assessing individual medical competence or patient diagnosis. - ### Future directions: - Translate to English (in-progress) - Include multimodal information (our new dataset ChiMed-Vision-Language-Instruction - 469,441 QA pairs: [https://paperswithcode.com/dataset/qilin-med-vl](https://paperswithcode.com/dataset/qilin-med-vl)) ## Citation Benchmarking Large Language Models on CMExam -- A Comprehensive Chinese Medical Exam Dataset https://arxiv.org/abs/2306.03030 ``` @article{liu2023benchmarking, title={Benchmarking Large Language Models on CMExam--A Comprehensive Chinese Medical Exam Dataset}, author={Liu, Junling and Zhou, Peilin and Hua, Yining and Chong, Dading and Tian, Zhongyu and Liu, Andrew and Wang, Helin and You, Chenyu and Guo, Zhenhua and Zhu, Lei and others}, journal={arXiv preprint arXiv:2306.03030}, year={2023} } ```