Datasets:
Tasks:
Text Classification
Modalities:
Text
Languages:
code
Size:
10K - 100K
ArXiv:
Tags:
code-comprehension
llm-evaluation
software-metrics
input-output-prediction
code-understanding
benchmark
License:
File size: 5,721 Bytes
6c7a057 44583d0 6c7a057 44583d0 6c7a057 44583d0 6c7a057 44583d0 3276745 44583d0 3276745 44583d0 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 | ---
license: apache-2.0
task_categories:
- text-classification
language:
- code
tags:
- code-comprehension
- llm-evaluation
- software-metrics
- input-output-prediction
- code-understanding
- benchmark
pretty_name: "Beyond Accuracy: Code Comprehension"
size_categories:
- 10K<n<100K
dataset_info:
features:
- name: sample_id
dtype: string
- name: code
dtype: string
- name: genome
dtype: string
- name: io_pairs
dtype: string
- name: correct_io_input
dtype: string
- name: correct_io_output
dtype: string
- name: incorrect_io_input
dtype: string
- name: incorrect_io_output
dtype: string
splits:
- name: test
num_examples: 12584
---
# Beyond Accuracy: Code Comprehension Dataset
Dataset for the paper **"Beyond Accuracy: Characterizing Code Comprehension Capabilities in (Large) Language Models"** by Machtle, Serr, Loose & Eisenbarth (University of Luebeck).
[[Paper]](https://arxiv.org/abs/2601.12951) | [[Code]](https://github.com/UzL-ITS/code-comprehension-capabilities-llms/)
## Task
**Binary I/O consistency**: given a Python program `p`, an input `x`, and a candidate output `y`, determine whether `y` is the correct output of running `p(x)`.
Each sample contains a **correct** I/O pair (label=1) and an **incorrect** I/O pair (label=0). Incorrect pairs are generated via *in-program shuffling* — pairing an input with the output of a *different* input to the same program — preserving lexical and stylistic characteristics while being semantically wrong.
## Quick Start
```python
from datasets import load_dataset
ds = load_dataset("Felix6326727/beyond-accuracy-code-comprehension", split="test")
sample = ds[0]
print(sample["code"][:200])
print(f"Correct: {sample['correct_io_input']!r} -> {sample['correct_io_output']!r}")
print(f"Incorrect: {sample['incorrect_io_input']!r} -> {sample['incorrect_io_output']!r}")
print(f"GPT-OSS 120B success: {sample['llm_gpt_oss_120b_success']}")
print(f"Cyclomatic complexity: {sample['metric_cyclomatic_complexity']}")
```
## Dataset Summary
| | |
|---|---|
| **Columns** | 249 |
| **Source** | Python subset of [Project CodeNet](https://github.com/IBM/Project_CodeNet) |
| **I/O generation** | Type-aware fuzzing with hill-climbing type inference |
| **Models evaluated** | 5 LLMs |
| **Code metrics** | 224 static analysis features |
## Column Groups
### Core Columns (10)
| Column | Type | Description |
|---|---|---|
| `sample_id` | string | Unique identifier (`{problem_id}.{solution_id}`) |
| `code` | string | Python source code |
| `source_file` | string | Original CodeNet file path |
| `genome` | string | Inferred input type signature (e.g. `"is"` = integer + string) |
| `io_pairs` | string | JSON array of all generated `[input, output]` pairs |
| `num_io_pairs` | int | Number of I/O pairs generated |
| `correct_io_input` | string | Input for the correct I/O test case |
| `correct_io_output` | string | Expected output (ground truth) |
| `incorrect_io_input` | string | Input for the incorrect I/O test case |
| `incorrect_io_output` | string | Shuffled (wrong) output |
### LLM Evaluation Columns (15)
Per-model results from the binary I/O consistency evaluation. Each model has 3 columns:
| Column pattern | Type | Description |
|---|---|---|
| `llm_{model}_success` | bool | `True` if the model answered all test cases correctly for this sample |
| `llm_{model}_num_correct` | int | Number of test cases answered correctly (out of `num_total`) |
| `llm_{model}_num_total` | int | Total test cases for this sample (typically 2: one correct, one incorrect) |
### Code Metric Columns (224)
All prefixed with `metric_`. Values are floats (or null if unavailable).
**Size & Complexity (67 columns)** — includes `cyclomatic_complexity`, `loc`, `sloc`, `lloc`, `maintainability_index`, `code_length`, Halstead metrics (`h1`, `h2`, `N1`, `N2`, `vocabulary`, `length`, `volume`, `difficulty`, `effort`, `bugs`), `num_branches`, `num_loops`, `num_identifiers`, `num_literals`, `num_data_flows`, `parameter_count`, `variable_count`, and more.
**AST / Graph Structure (118 columns)** — `metric_graph_nodes_*` columns counting occurrences of each AST node type: `if_statement`, `for_statement`, `call`, `assignment`, `binary_operator`, `identifier`, `block`, etc. Also includes graph-level metrics: `num_nodes`, `num_edges`, `density`, `diameter`, `average_shortest_path_length`, `average_clustering`.
**Opcode Statistics (39 columns)** — Python bytecode features: `num_opcodes`, `sum_opcodes`, `avg_opcode_count`, `min_opcode_count`, `max_opcode_count`, individual opcode counts (`opcode_1`, `opcode_83`, ...), `opcodes_used0`–`opcodes_used3`, and `top_0_opcode_name` through `top_19_opcode_name`.
## Data Generation Pipeline
```
Python files (CodeNet)
|
v
hill_climb.py ─── infer input types ("genome") via coverage-guided search
|
v
fuzzer.py ──────── generate & shrink minimal I/O pairs
|
v
export_io.py ───── create correct + incorrect (shuffled) I/O pairs
|
v
This dataset
```
See the [GitHub repository](https://github.com/UzL-ITS/code-comprehension-capabilities-llms/) for the full pipeline code.
## Citation
```bibtex
@article{machtle2025beyond,
title={Beyond Accuracy: Characterizing Code Comprehension Capabilities in (Large) Language Models},
author={Machtle, Felix and Serr, Jan-Niclas and Loose, Nils and Eisenbarth, Thomas},
journal={arXiv preprint arXiv:2601.12951},
year={2025}
}
```
## License
Derived from [Project CodeNet](https://github.com/IBM/Project_CodeNet) (Apache 2.0).
|