File size: 3,193 Bytes
b2695b7
 
 
 
 
 
 
 
 
 
 
 
 
 
164d78b
b2695b7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
---
license: mit
task_categories:
  - visual-question-answering
  - image-text-to-text
language:
  - en
pretty_name: Med Eval Data
size_categories:
  - 10K<n<100K
---

# Med Eval Data

This dataset contains evaluation data for the **Med** project. Its data format is the same as [`Med2026/Med_training_data`](https://huggingface.co/datasets/Med2026/Med_training_data), and it can be loaded with the same codebase from [`GAIR-NLP/Med`](https://github.com/GAIR-NLP/Med).

## Overview

Each example is stored in the same JSON / parquet schema as the training data, with the following top-level fields:

- `images`
- `data_source`
- `prompt`
- `ability`
- `reward_model`
- `extra_info`
- `agent_name`

This means the dataset is directly compatible with the data loading pipeline used in the Med codebase.

## Compatibility

This dataset has the **same format** as [`Med2026/Med_training_data`](https://huggingface.co/datasets/Med2026/Med_training_data).

You can use the same loading logic and preprocessing pipeline from:

- [`GAIR-NLP/Med`](https://github.com/GAIR-NLP/Med)

No format conversion is required.

## Data Split by File Naming

The evaluation data is divided into two settings according to the file name:

- Files with `single_turn_agent` in the filename correspond to **evaluation without tool use**
- Files with `tool_agent` in the filename correspond to **evaluation with tool use**

In other words:

- `*single_turn_agent*` → without-tool evaluation
- `*tool_agent*` → with-tool evaluation

## Data Format

Each sample is a JSON object with the following structure:

```python
{
  "images": [PIL.Image],
  "data_source": "vstar_bench_single_turn_agent",
  "prompt": [
    {
      "content": "<image>\nWhat is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
      "role": "user"
    }
  ],
  "ability": "direct_attributes",
  "reward_model": {
    "answer": "A",
    "format_ratio": 0.0,
    "ground_truth": "\\boxed{A}",
    "length_ratio": 0.0,
    "style": "multiple_choice",
    "verifier": "mathverify",
    "verifier_parm": {
      "det_verifier_normalized": null,
      "det_reward_ratio": {
        "iou_max_label_first": null,
        "iou_max_iou_first": null,
        "iou_completeness": null,
        "map": null,
        "map50": null,
        "map75": null
      }
    }
  },
  "extra_info": {
    "answer": "A",
    "data_source": "vstar_bench_single_turn_agent",
    "id": "vstar_bench_0",
    "image_path": "direct_attributes/sa_4690.jpg",
    "question": "<image>\nWhat is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
    "split": "test",
    "index": "0",
    "prompt_length": null,
    "tools_kwargs": {
      "crop_and_zoom": {
        "create_kwargs": {
          "raw_query": "What is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
          "image": "PIL.Image"
        }
      }
    },
    "need_tools_kwargs": false
  },
  "agent_name": "single_turn_agent"
}