File size: 4,124 Bytes
f416e2c
 
 
 
4391020
 
 
 
 
f416e2c
 
 
4391020
f416e2c
 
141e274
f416e2c
 
 
141e274
 
 
 
 
 
f416e2c
da92982
4391020
 
b78f5db
da92982
b78f5db
4391020
b78f5db
141e274
 
 
 
 
da92982
d6c34da
da92982
141e274
 
d6c34da
141e274
 
 
 
 
 
 
 
 
 
 
 
 
b78f5db
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d6c34da
 
 
 
 
141e274
 
e132332
141e274
 
 
b78f5db
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4391020
d6c34da
4391020
d6c34da
4391020
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
---
language:
- en
- zh
license: c-uda
size_categories:
- 1K<n<10K
task_categories:
- question-answering
tags:
- biology
- medical
- anesthesiology
viewer: true
configs:
- config_name: en
  data_files:
  - split: test
    path:
    - anesbench_en.json
- config_name: zh
  data_files:
  - split: test
    path:
    - anesbench_zh.json
---

# AnesBench

[**Paper**](https://huggingface.co/papers/2504.02404) | [**GitHub**](https://github.com/mililab/anesbench)

# Dataset Description

**AnesBench** is designed to assess anesthesiology-related reasoning capabilities of Large Language Models (LLMs). It provides bilingual (English and Chinese) anesthesiology questions across two separate files. Each question is labeled with a three-level categorization of cognitive demands based on dual-process theory (System 1, System 1.x, and System 2), enabling evaluation of LLMs' knowledge, application, and clinical reasoning abilities across diverse linguistic contexts.

| Subset | File | Total | System 1 | System 1.x | System 2 |
|--------|------|-------|----------|-------------|----------|
| English | `anesbench_en.json` | 4,343 | 2,960 | 1,028 | 355 |
| Chinese | `anesbench_zh.json` | 3,529 | 2,784 | 534 | 211 |

## JSON Sample

**English** (`anesbench_en.json`):

```json
{
    "id": "91b5e145-57f2-5307-99e4-eafd75643de4",
    "question": "The concentration of a specific gas in solution depends on which of the following?",
    "A": "Temperature of the solution",
    "B": "Volume of the system",
    "C": "Solubility of the specific gas in that solution",
    "D": "Molecular weight of the gas",
    "choice_num": 4,
    "target": "C",
    "category": 1
}
```

**Chinese** (`anesbench_zh.json`):

```json
{
    "A": "替代治疗",
    "B": "手术治疗",
    "C": "对症治疗",
    "D": "静脉输注糖皮质激素",
    "E": "补充盐皮质激素",
    "id": "78587bd9-f3f6-4118-b6eb-95ed7c91a0ec",
    "question": "Addison病抢救的主要措施是",
    "choice_num": 5,
    "target": "D",
    "category": 1
}
```

## Field Explanations

| Field         | Type   | Description                                                                 |
|------------------|----------|-----------------------------------------------------------------------------|
| `id`             | string   | A randomly generated ID using UUID                                          |
| `question`       | string   | The question stem                                                           |
| `A``I`          | string   | Answer options (from `A` up to the total number of options)                 |
| `choice_num`     | int      | The number of options in this question                                      |
| `target`         | string   | The correct answer to this question                                         |
| `category`       | int      | The cognitive demand category of the question (`1` = System 1, `2` = System 1.x, `3` = System 2) |

### Cognitive Demand Categories

| Category | Label | Description |
|----------|-------|-------------|
| 1 | **System 1** | Fast, intuitive recall of factual knowledge |
| 2 | **System 1.x** | Pattern recognition and application of learned rules |
| 3 | **System 2** | Deliberate, analytical clinical reasoning |

## Recommended Usage

- **Question Answering**: QA in a zero-shot or few-shot setting, where the question is fed into a QA system. Accuracy should be used as the evaluation metric.

## Usage

To evaluate a model on AnesBench, you can use the evaluation code provided in the [official repository](https://github.com/MiliLab/AnesSuite).


## Citation

If you find AnesBench helpful, please consider citing the following paper:

```latex
@inproceedings{
  feng2026anessuite,
  title={AnesSuite: A Comprehensive Benchmark and Dataset Suite for Anesthesiology Reasoning in {LLM}s},
  author={Xiang Feng and Wentao Jiang and Zengmao Wang and Yong Luo and Pingbo Xu and Baosheng Yu and Hua Jin and Jing Zhang},
  booktitle={The Fourteenth International Conference on Learning Representations},
  year={2026},
  url={https://openreview.net/forum?id=iKRQMeC7yO}
}
```