Datasets:

Modalities:
Tabular
Text
Formats:
parquet
Languages:
English
Size:
< 1K
ArXiv:
Tags:
code
Libraries:
Datasets
Dask
License:
effibench-x / README.md
qyhfrank's picture
Update README
49babbe
---
dataset_info:
features:
- name: id
dtype: string
- name: title
dtype: string
- name: title_slug
dtype: string
- name: description
dtype: string
- name: description_md
dtype: string
- name: difficulty
dtype: string
- name: tags
list: string
- name: source
dtype: string
- name: url
dtype: string
- name: type
dtype: string
- name: release_timestamp
dtype: int64
- name: release_date
dtype: string
- name: time_limit_nanos
dtype: int64
- name: memory_limit_bytes
dtype: int64
- name: starter_code
struct:
- name: c
dtype: string
- name: cpp
dtype: string
- name: csharp
dtype: string
- name: dart
dtype: string
- name: elixir
dtype: string
- name: erlang
dtype: string
- name: golang
dtype: string
- name: java
dtype: string
- name: javascript
dtype: string
- name: kotlin
dtype: string
- name: php
dtype: string
- name: python
dtype: string
- name: python3
dtype: string
- name: racket
dtype: string
- name: ruby
dtype: string
- name: rust
dtype: string
- name: scala
dtype: string
- name: swift
dtype: string
- name: typescript
dtype: string
- name: solutions
struct:
- name: cpp
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: golang
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: java
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: javascript
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: python3
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: ruby
struct:
- name: code
dtype: string
- name: memory
dtype: int64
- name: memoryDistribution
dtype: string
- name: runtime
dtype: int64
- name: runtimeDistribution
dtype: string
- name: test_case_generator
dtype: string
- name: evaluator
dtype: string
- name: generated_tests
dtype: string
- name: test_runners
struct:
- name: cpp
dtype: string
- name: golang
dtype: string
- name: java
dtype: string
- name: javascript
dtype: string
- name: python3
dtype: string
- name: ruby
dtype: string
splits:
- name: test
num_bytes: 3865548641
num_examples: 623
download_size: 2341977516
dataset_size: 3865548641
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
license: apache-2.0
task_categories:
- question-answering
language:
- en
tags:
- code
pretty_name: EffiBench-X
size_categories:
- n<1K
---
# Dataset Card for EffiBench-X
**EffiBench-X** is the first multi-language benchmark designed specifically to evaluate the efficiency of LLM-generated code across six programming languages: Python, C++, Java, JavaScript, Ruby, and Golang. The dataset comprises 623 competitive programming problems paired with human expert solutions as efficiency baselines.
## Dataset Details
### Dataset Description
EffiBench-X addresses critical limitations in existing code generation benchmarks by providing:
- **Multi-language evaluation** across Python, C++, Java, JavaScript, Ruby, and Golang
- **Efficiency-focused metrics** including execution time, memory peak, and memory integral
- **Human expert baselines** for reliable efficiency comparison
- **Curated by:** Yuhao Qing, Boyu Zhu, Mingzhe Du, Zhijiang Guo, Terry Yue Zhuo, Qianru Zhang, Jie M. Zhang, Heming Cui, Siu-Ming Yiu, Dong Huang, See-Kiong Ng, Luu Anh Tuan
- **Institutions:** HKU, UCL, NTU, NUS, HKUST, Monash University, CSIRO's Data61, KCL
- **Language(s) (NLP):** English
- **License:** Apache License 2.0
### Dataset Sources
- **Repository:** [EffiBench-X (GitHub)](https://github.com/EffiBench/EffiBench-X)
- **Dataset:** [EffiBench/effibench-x](https://huggingface.co/datasets/EffiBench/effibench-x)
- **Paper:** [arXiv:2505.13004](https://arxiv.org/abs/2505.13004)
- **Problem Sources:**
- [LeetCode](https://leetcode.com)
- [Aizu Online Judge](https://onlinejudge.u-aizu.ac.jp/)
- [AtCoder](https://atcoder.jp)
- [CodeChef](https://www.codechef.com)
- [Codeforces](https://codeforces.com)
## Uses
### Direct Use
- **Benchmarking LLM code generation efficiency**: Evaluate models on runtime performance, memory usage, and correctness across multiple languages
- **Cross-language performance analysis**: Compare model capabilities across different programming paradigms
- **Model development**: Train and fine-tune models for efficient code generation
- **Research**: Study efficiency gaps between LLM-generated and human expert code
### Out-of-Scope Use
- **Production deployment without validation**: Solutions should be verified before production use
- **Security-critical applications**: The dataset focuses on algorithmic efficiency, not security
- **Non-competitive programming domains**: Problems are algorithmic in nature and may not represent all software engineering contexts
## Dataset Structure
The dataset contains 623 problems categorized into:
- **Functional problems**: Implement specific functions/classes with I/O handled by test templates
- **Standard I/O problems**: Complete programs reading from stdin and writing to stdout
Key fields per record include:
- `id`, `title`, `title_slug`, `description`, `description_md`, `difficulty`, `tags`, `source`, `url`, `type`
- Limits: `time_limit_nanos`, `memory_limit_bytes`
- Code artifacts:
- `starter_code`: language-keyed starter snippets
- `solutions`: language-keyed canonical solutions (e.g., for `cpp`, `golang`, `java`, `javascript`, `python3`, `ruby`)
- `test_case_generator`: executable code string that programmatically produces tests
- `evaluator`: executable code string to determine pass/fail given expected vs. program output
- `generated_tests`: serialized tests produced by the generator
- `test_runners`: language-keyed runner templates for executing solutions
All problems are from competitive programming platforms.
## Dataset Creation
### Curation Rationale
Existing code generation benchmarks primarily focus on functional correctness with limited attention to efficiency, often restricted to Python. EffiBench-X addresses three critical limitations:
1. **Language diversity**: Extends beyond Python to include statically-typed (C++, Java, Go) and dynamically-typed languages (Python, JavaScript, Ruby)
2. **Data contamination**: Uses recent problems (post-October 2023) to avoid memorization effects
3. **Complexity**: Features algorithmically challenging problems requiring optimization techniques
### Source Data
#### Data Collection and Processing
Problems are curated from competitive programming platforms. Each problem includes:
- Human expert solutions verified for correctness and efficiency
- 100 programmatically generated test cases
- Test runners and evaluators for automated assessment
- Cross-language validation to ensure consistency
#### Who are the source data producers?
- **Problem creators**: Competitive programming platforms and contest organizers
- **Solution authors**: Human expert programmers from competitive programming communities
- **Dataset curators**: EffiBench research team
## Citation
Please cite our paper if you use this dataset:
```bibtex
@article{qing2025effibench,
title={EffiBench-X: A Multi-Language Benchmark for Measuring Efficiency of LLM-Generated Code},
author={Qing, Yuhao and Zhu, Boyu and Du, Mingzhe and Guo, Zhijiang and Zhuo, Terry Yue and Zhang, Qianru and Zhang, Jie M and Cui, Heming and Yiu, Siu-Ming and Huang, Dong and Ng, See-Kiong and Tuan, Luu Anh},
journal={Advances in neural information processing systems},
year={2025}
}
```
## More Information
- **Dataset Statistics**: 623 problems, 100 test cases per problem, 6 languages
- **Evaluation**: Sandboxed execution environment for consistent performance measurements
- For detailed information and benchmark results, please refer to the [paper](https://arxiv.org/abs/2505.13004) and [GitHub repository](https://github.com/EffiBench/EffiBench-X)
## Dataset Card Contact
For questions and feedback, please open an issue on our [GitHub repository](https://github.com/EffiBench/EffiBench-X).