--- language: - en task_categories: - text-generation tags: - code-reasoning - benchmark - python - c - java - software-engineering - llm-evaluation license: unknown --- # CodeSense: A Real-World Benchmark and Dataset for Code Semantic Reasoning This repository contains the dataset and resources for **CodeSense**, the first benchmark for evaluating Large Language Models (LLMs) on fine-grained code semantic reasoning tasks in real-world software engineering contexts. The benchmark was presented in the paper [CodeSense: a Real-World Benchmark and Dataset for Code Semantic Reasoning](https://huggingface.co/papers/2506.00750). CodeSense aims to bridge the gap between existing synthetic or educational coding problems and the practical demands of software engineering. It utilizes Python, C, and Java software projects from real-world repositories, collecting execution traces to construct a ground truth dataset for detailed semantic reasoning tasks. **Paper:** [https://huggingface.co/papers/2506.00750](https://huggingface.co/papers/2506.00750) **Project Page:** [https://codesense-bench.github.io/](https://codesense-bench.github.io/) **Code Repository:** [https://github.com/codesense-bench/codesense-codes](https://github.com/codesense-bench/codesense-codes) ## Codebase Overview The associated code repository ([codesense-bench/codesense-codes](https://github.com/codesense-bench/codesense-codes)) contains three main components related to execution tracing, benchmark dataset creation, and LLM evaluation: ### Benchmark Collection - **Purpose:** Contains scripts to process and clean raw execution traces. - **Description:** Converts raw traces into task-specific datasets suitable for various code understanding and reasoning benchmarks. ### Tracing Framework - **Purpose:** Tools for collecting execution traces. - **Description:** Supports tracing of Python, C, and Java programs to capture their runtime behavior and execution steps. ### LLM Evaluation - **Purpose:** Scripts for evaluating Large Language Models (LLMs) on the task-specific datasets. - **Description:** Runs evaluations, computes metrics, and benchmarks model performance on the curated datasets.