--- license: apache-2.0 task_categories: - text-classification language: - code tags: - code-comprehension - llm-evaluation - software-metrics - input-output-prediction - code-understanding - benchmark pretty_name: "Beyond Accuracy: Code Comprehension" size_categories: - 10K {sample['correct_io_output']!r}") print(f"Incorrect: {sample['incorrect_io_input']!r} -> {sample['incorrect_io_output']!r}") print(f"GPT-OSS 120B success: {sample['llm_gpt_oss_120b_success']}") print(f"Cyclomatic complexity: {sample['metric_cyclomatic_complexity']}") ``` ## Dataset Summary | | | |---|---| | **Columns** | 249 | | **Source** | Python subset of [Project CodeNet](https://github.com/IBM/Project_CodeNet) | | **I/O generation** | Type-aware fuzzing with hill-climbing type inference | | **Models evaluated** | 5 LLMs | | **Code metrics** | 224 static analysis features | ## Column Groups ### Core Columns (10) | Column | Type | Description | |---|---|---| | `sample_id` | string | Unique identifier (`{problem_id}.{solution_id}`) | | `code` | string | Python source code | | `source_file` | string | Original CodeNet file path | | `genome` | string | Inferred input type signature (e.g. `"is"` = integer + string) | | `io_pairs` | string | JSON array of all generated `[input, output]` pairs | | `num_io_pairs` | int | Number of I/O pairs generated | | `correct_io_input` | string | Input for the correct I/O test case | | `correct_io_output` | string | Expected output (ground truth) | | `incorrect_io_input` | string | Input for the incorrect I/O test case | | `incorrect_io_output` | string | Shuffled (wrong) output | ### LLM Evaluation Columns (15) Per-model results from the binary I/O consistency evaluation. Each model has 3 columns: | Column pattern | Type | Description | |---|---|---| | `llm_{model}_success` | bool | `True` if the model answered all test cases correctly for this sample | | `llm_{model}_num_correct` | int | Number of test cases answered correctly (out of `num_total`) | | `llm_{model}_num_total` | int | Total test cases for this sample (typically 2: one correct, one incorrect) | ### Code Metric Columns (224) All prefixed with `metric_`. Values are floats (or null if unavailable). **Size & Complexity (67 columns)** — includes `cyclomatic_complexity`, `loc`, `sloc`, `lloc`, `maintainability_index`, `code_length`, Halstead metrics (`h1`, `h2`, `N1`, `N2`, `vocabulary`, `length`, `volume`, `difficulty`, `effort`, `bugs`), `num_branches`, `num_loops`, `num_identifiers`, `num_literals`, `num_data_flows`, `parameter_count`, `variable_count`, and more. **AST / Graph Structure (118 columns)** — `metric_graph_nodes_*` columns counting occurrences of each AST node type: `if_statement`, `for_statement`, `call`, `assignment`, `binary_operator`, `identifier`, `block`, etc. Also includes graph-level metrics: `num_nodes`, `num_edges`, `density`, `diameter`, `average_shortest_path_length`, `average_clustering`. **Opcode Statistics (39 columns)** — Python bytecode features: `num_opcodes`, `sum_opcodes`, `avg_opcode_count`, `min_opcode_count`, `max_opcode_count`, individual opcode counts (`opcode_1`, `opcode_83`, ...), `opcodes_used0`–`opcodes_used3`, and `top_0_opcode_name` through `top_19_opcode_name`. ## Data Generation Pipeline ``` Python files (CodeNet) | v hill_climb.py ─── infer input types ("genome") via coverage-guided search | v fuzzer.py ──────── generate & shrink minimal I/O pairs | v export_io.py ───── create correct + incorrect (shuffled) I/O pairs | v This dataset ``` See the [GitHub repository](https://github.com/UzL-ITS/code-comprehension-capabilities-llms/) for the full pipeline code. ## Citation ```bibtex @article{machtle2025beyond, title={Beyond Accuracy: Characterizing Code Comprehension Capabilities in (Large) Language Models}, author={Machtle, Felix and Serr, Jan-Niclas and Loose, Nils and Eisenbarth, Thomas}, journal={arXiv preprint arXiv:2601.12951}, year={2025} } ``` ## License Derived from [Project CodeNet](https://github.com/IBM/Project_CodeNet) (Apache 2.0).