|
|
--- |
|
|
language: |
|
|
- en |
|
|
task_categories: |
|
|
- text-generation |
|
|
tags: |
|
|
- code-reasoning |
|
|
- benchmark |
|
|
- python |
|
|
- c |
|
|
- java |
|
|
- software-engineering |
|
|
- llm-evaluation |
|
|
license: unknown |
|
|
--- |
|
|
|
|
|
# CodeSense: A Real-World Benchmark and Dataset for Code Semantic Reasoning |
|
|
|
|
|
This repository contains the dataset and resources for **CodeSense**, the first benchmark for evaluating Large Language Models (LLMs) on fine-grained code semantic reasoning tasks in real-world software engineering contexts. The benchmark was presented in the paper [CodeSense: a Real-World Benchmark and Dataset for Code Semantic Reasoning](https://huggingface.co/papers/2506.00750). |
|
|
|
|
|
CodeSense aims to bridge the gap between existing synthetic or educational coding problems and the practical demands of software engineering. It utilizes Python, C, and Java software projects from real-world repositories, collecting execution traces to construct a ground truth dataset for detailed semantic reasoning tasks. |
|
|
|
|
|
**Paper:** [https://huggingface.co/papers/2506.00750](https://huggingface.co/papers/2506.00750) |
|
|
**Project Page:** [https://codesense-bench.github.io/](https://codesense-bench.github.io/) |
|
|
**Code Repository:** [https://github.com/codesense-bench/codesense-codes](https://github.com/codesense-bench/codesense-codes) |
|
|
|
|
|
## Codebase Overview |
|
|
The associated code repository ([codesense-bench/codesense-codes](https://github.com/codesense-bench/codesense-codes)) contains three main components related to execution tracing, benchmark dataset creation, and LLM evaluation: |
|
|
|
|
|
### Benchmark Collection |
|
|
- **Purpose:** Contains scripts to process and clean raw execution traces. |
|
|
- **Description:** Converts raw traces into task-specific datasets suitable for various code understanding and reasoning benchmarks. |
|
|
|
|
|
### Tracing Framework |
|
|
- **Purpose:** Tools for collecting execution traces. |
|
|
- **Description:** Supports tracing of Python, C, and Java programs to capture their runtime behavior and execution steps. |
|
|
|
|
|
### LLM Evaluation |
|
|
- **Purpose:** Scripts for evaluating Large Language Models (LLMs) on the task-specific datasets. |
|
|
- **Description:** Runs evaluations, computes metrics, and benchmarks model performance on the curated datasets. |