--- language: - en license: mit size_categories: - n<100 task_categories: - text-generation dataset_info: features: - name: instance_id dtype: string - name: patch dtype: string - name: test_patch dtype: string - name: FAIL_TO_PASS list: string - name: PASS_TO_PASS list: string - name: image_name dtype: string - name: repo dtype: string - name: base_commit dtype: string - name: problem_statement dtype: string - name: repo_settings dtype: string splits: - name: level1 num_bytes: 68545 num_examples: 1 - name: level2 num_bytes: 60238 num_examples: 1 download_size: 60937 dataset_size: 128783 configs: - config_name: default data_files: - split: level1 path: data/level1-* - split: level2 path: data/level2-* tags: - code - agents - software-engineering --- # ACE-Bench: Agent Coding Evaluation Benchmark ## Dataset Description ACE-Bench is a comprehensive benchmark designed to evaluate AI agents' capabilities in end-to-end feature-level code generation. Unlike traditional benchmarks that focus on function-level or algorithm-specific tasks, ACE-Bench challenges agents to implement complete features within real-world software projects. ### Key Characteristics - **Feature-Level Tasks**: Each task requires implementing a complete feature, including multiple functions, classes, and their interactions - **Real-World Codebases**: Tasks are derived from actual open-source projects, preserving the complexity and context of production code - **End-to-End Evaluation**: Agents must understand requirements, generate code, and pass comprehensive test suites - **Two Difficulty Levels**: - **Level 1**: Agents receive masked code with interface signatures and must implement the complete functionality - **Level 2**: Agents receive only test files and must implement both the interface and functionality from scratch ### Dataset Statistics - **Total Instances**: 2 - **Level 1 Instances**: 1 - **Level 2 Instances**: 1 - **Total Size**: 125.76 KB - **Download Size**: 59.51 KB ## Dataset Structure Each instance in ACE-Bench contains: - `instance_id`: Unique identifier for the task - `patch`: Git diff showing the implementation (Level 1) or empty string (Level 2) - `test_patch`: Git diff showing test file modifications - `FAIL_TO_PASS`: List of test files that must pass after implementation - `PASS_TO_PASS`: List of test files that must continue passing (Level 1 only) - `image_name`: Docker image containing the development environment - `repo`: Source repository (e.g., "owner/repo-name") - `base_commit`: Git commit hash of the base version - `problem_statement`: Detailed task description and requirements - `repo_settings`: Repository configuration settings as JSON string (from python.py) ## Usage ```python import json from datasets import load_dataset # Load Level 1 tasks dataset_lv1 = load_dataset("BamChil/ACE-Bench", split="level1") # Load Level 2 tasks dataset_lv2 = load_dataset("BamChil/ACE-Bench", split="level2") # Example: Access a task task = dataset_lv1[0] print(task['instance_id']) print(task['problem_statement']) # Parse repo_settings from JSON string repo_settings = json.loads(task['repo_settings']) print(repo_settings['repository']) print(repo_settings['base_image']) ```