--- pretty_name: "Humaneval" license: "mit" language: - en tags: - code-generation - python - text task_categories: - text-generation --- # Humaneval This repository hosts a copy of the widely used **Humaneval** dataset, a benchmark for evaluating the code generation capabilities and performance of Large Language Models (LLMs). Humaneval consists of programming tasks designed to evaluate the ability of models to generate working Python code based on textual descriptions. It is frequently used in research papers assessing LLMs' code generation performance, particularly in the context of automated programming. ## Contents - `humaneval.jsonl` (or your actual filename): the standard set of programming tasks. Each entry contains: ``` { "task_id": "...", "prompt": "...", "canonical_solution": "...", "test": "...", "entry_point": "..." } ``` ## Usage ``` from datasets import load_dataset ds = load_dataset("S3IC/humaneval") ``` ## Source This dataset is taken from the public Humaneval release: https://github.com/openai/human-eval ## License Humaneval is released under the original license provided by the authors. See the LICENSE file for more details.