Med_eval_data / README.md
ManTle's picture
Update README.md
164d78b verified
---
license: mit
task_categories:
- visual-question-answering
- image-text-to-text
language:
- en
pretty_name: Med Eval Data
size_categories:
- 10K<n<100K
---
# Med Eval Data
This dataset contains evaluation data for the **Med** project. Its data format is the same as [`Med2026/Med_training_data`](https://huggingface.co/datasets/Med2026/Med_training_data), and it can be loaded with the same codebase from [`GAIR-NLP/Med`](https://github.com/GAIR-NLP/Med).
## Overview
Each example is stored in the same JSON / parquet schema as the training data, with the following top-level fields:
- `images`
- `data_source`
- `prompt`
- `ability`
- `reward_model`
- `extra_info`
- `agent_name`
This means the dataset is directly compatible with the data loading pipeline used in the Med codebase.
## Compatibility
This dataset has the **same format** as [`Med2026/Med_training_data`](https://huggingface.co/datasets/Med2026/Med_training_data).
You can use the same loading logic and preprocessing pipeline from:
- [`GAIR-NLP/Med`](https://github.com/GAIR-NLP/Med)
No format conversion is required.
## Data Split by File Naming
The evaluation data is divided into two settings according to the file name:
- Files with `single_turn_agent` in the filename correspond to **evaluation without tool use**
- Files with `tool_agent` in the filename correspond to **evaluation with tool use**
In other words:
- `*single_turn_agent*` → without-tool evaluation
- `*tool_agent*` → with-tool evaluation
## Data Format
Each sample is a JSON object with the following structure:
```python
{
"images": [PIL.Image],
"data_source": "vstar_bench_single_turn_agent",
"prompt": [
{
"content": "<image>\nWhat is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
"role": "user"
}
],
"ability": "direct_attributes",
"reward_model": {
"answer": "A",
"format_ratio": 0.0,
"ground_truth": "\\boxed{A}",
"length_ratio": 0.0,
"style": "multiple_choice",
"verifier": "mathverify",
"verifier_parm": {
"det_verifier_normalized": null,
"det_reward_ratio": {
"iou_max_label_first": null,
"iou_max_iou_first": null,
"iou_completeness": null,
"map": null,
"map50": null,
"map75": null
}
}
},
"extra_info": {
"answer": "A",
"data_source": "vstar_bench_single_turn_agent",
"id": "vstar_bench_0",
"image_path": "direct_attributes/sa_4690.jpg",
"question": "<image>\nWhat is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
"split": "test",
"index": "0",
"prompt_length": null,
"tools_kwargs": {
"crop_and_zoom": {
"create_kwargs": {
"raw_query": "What is the material of the glove?\n(A) rubber\n(B) cotton\n(C) kevlar\n(D) leather\nAnswer with the option's letter from the given choices directly.",
"image": "PIL.Image"
}
}
},
"need_tools_kwargs": false
},
"agent_name": "single_turn_agent"
}