Datasets:
IDEA-AI4S
/

Modalities:
Image
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
License:
File size: 3,459 Bytes
5c0bfbc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
---
license: apache-2.0
size_categories:
- 1K<n<10K
task_categories:
- question-answering
- image-text-to-text
tags:
- chemistry
- agent
- olympaid
- benchmark
- llm-evaluation
- science
- multimodal
language:
- en
---
# πŸ§ͺ **ChemO Dataset**
[![Hugging Face](https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Paper-blue)](https://huggingface.co/papers/2511.16205)
[![arXiv](https://img.shields.io/badge/arXiv-2511.16205-b31b1b.svg)](https://arxiv.org/abs/2511.16205)

πŸ“„ **Paper**: [ChemLabs on ChemO: A Multi-Agent System for Multimodal Reasoning on IChO 2025](https://huggingface.co/papers/2511.16205)

# ChemO Version 1.1

Now with CDXML Files! πŸŽ‰

The ChemO dataset has been officially released after meticulous proofreading and preparation. This benchmark is built from the **International Chemistry Olympiad (IChO) 2025** and represents a new frontier in automated chemical problem-solving.

## 🌟 Key Features

- **πŸ† Olympic-Level Benchmark** - Challenging problems from IChO 2025 for advanced AI reasoning
- **πŸ”¬ Multimodal Symbolic Language** - Addresses chemistry's unique combination of text, formulas, and molecular structures
- **πŸ“Š Two Novel Assessment Methods**:
  - **AER (Assessment-Equivalent Reformulation)** - Converts visual output requirements (e.g., drawing molecules) into computationally tractable formats
  - **SVE (Structured Visual Enhancement)** - Diagnostic mechanism to separate visual perception from core chemical reasoning capabilities

## πŸ“¦ What's Included

The current release includes:

- βœ… **Original Problems** - Complete problem sets with additional chapter markers for Problems and Solutions sections (no other modifications to the original content)
- βœ… **Well-structured JSON Files** - Clean, organized data designed for:
  - πŸ€– **MLLM Benchmarking** - Olympic-level chemistry reasoning evaluation
  - πŸ”— **Multi-Agent System Testing** - Hierarchical agent collaboration assessment
  - 🎯 **Multimodal Reasoning** - Text, formula, and molecular structure understanding
- βœ… **CDXML Files** - Molecular structure files now available in `JSON/cdxml/`

## πŸ“‹ Dataset Structure

The ChemO dataset consists of **9 problems** from IChO 2025, with each problem provided as a structured JSON file (1.json ~ 9.json in `JSON/`). 
```
JSON/
β”œβ”€β”€ 1.json ~ 9.json           # Problem and solution data in structured JSON format
β”œβ”€β”€ images/                   # All referenced images indexed in JSON files
└── cdxml/                    # Molecular structure files in CDXML format
```

## πŸ“š Data Source

All problems are sourced from **ICHO 2025**: https://www.icho2025.ae/problems

## πŸš€ State-of-the-Art Results

Our ChemLabs multi-agent system combined with SVE achieves **93.6/100** on ChemO, surpassing the estimated human gold medal threshold and establishing a new benchmark in automated chemical problem-solving.

## 🀝 Community

We appreciate your patience and look forward to your feedback as we continue to improve this resource for the community. Feel free to reach out to us at jerry.sy.bai@gmail.com.

## πŸ“„ Citation

If you use ChemO in your research, please cite our paper:
```bibtex
@article{qiang2025chemlabs,
  title={ChemLabs on ChemO: A Multi-Agent System for Multimodal Reasoning on IChO 2025},
  author={Xu, Qiang and Bai, Shengyuan and Chen, Leqing and Liu, Zijing and Li, Yu},
  journal={arXiv preprint arXiv:2511.16205},
  year={2025}
}
```