File size: 6,793 Bytes
cfbb122
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
---

dataset_info:
  features:
    - name: instance_id
      dtype: string
    - name: repo
      dtype: string
    - name: base_commit
      dtype: string
    - name: problem_statement
      dtype: string
    - name: test_patch
      dtype: string
    - name: human_patch
      dtype: string
    - name: pr_number
      dtype: int64
    - name: pr_url
      dtype: string
    - name: pr_merged_at
      dtype: string
    - name: issue_number
      dtype: int64
    - name: issue_url
      dtype: string
    - name: human_changed_lines
      dtype: int64
    - name: FAIL_TO_PASS
      dtype: string
    - name: PASS_TO_PASS
      dtype: string
    - name: version
      dtype: string
  splits:
    - name: test
      num_examples: 119
license: mit
task_categories:
  - other
language:
  - en
tags:
  - code-generation
  - software-engineering
  - complexity
  - swe-bench
  - contamination-free
  - post-training-cutoff
pretty_name: SWE-bench Complex
size_categories:
  - n<1K
---


# SWE-bench Complex

**A contamination-free, complexity-focused evaluation set for AI coding agents.**

SWE-bench Complex is a curated dataset of **119 real-world GitHub issues** from major Python open-source projects, designed specifically for studying code complexity in AI-generated patches. All tasks were merged between **January–March 2026**, guaranteeing they postdate the training cutoff of current frontier models.

## Why SWE-bench Complex?

Existing benchmarks like [SWE-bench Verified](https://huggingface.co/datasets/princeton-nlp/SWE-bench_Verified) suffer from two problems for complexity research:

### 1. Data Contamination

Over 94% of SWE-bench issues predate current LLM training cutoffs. Aleithan et al. found that **32.67% of successful patches involve "cheating"** through solution leakage, and resolution rates dropped from 12.47% to 3.97% when leaked instances were filtered out ([SWE-bench+, 2024](https://arxiv.org/abs/2410.06992)).

All SWE-bench Complex instances postdate the training cutoffs of:

| Model | Provider | Training Cutoff | Gap |
|---|---|---|---|
| Claude Opus 4.6 | Anthropic | Oct 2025 | 3+ months |
| GPT-5.3-Codex | OpenAI | Sep 2025 | 4+ months |
| GPT-5.4 | OpenAI | Nov 2025 | 2+ months |
| Gemini 3.1 Pro | Google | Oct 2025 | 3+ months |

### 2. Trivial Patches

SWE-bench Verified has a median patch size of just **7 changed lines** — 44.6% of tasks require only 1–5 lines. These trivial patches yield near-zero complexity deltas, reducing statistical power for quality studies.

SWE-bench Complex targets **substantive patches** with a median of **48 changed lines** — 6.9× larger than SWE-bench Verified.

## Dataset Comparison

| Characteristic | SWE-bench Verified | SWE-bench Complex |
|---|---|---|
| Tasks | 500 | 119 |
| Repositories | 12 | 8 |
| Median changed lines | 7 | **48** |
| Mean changed lines | 14.3 | **74.9** |
| Mean Python files changed | 1.2 | **3.9** |
| Human ΔCC (mean) | +1.14 | **+4.06** |
| Human ΔLLOC (mean) | +2.77 | **+19.08** |
| Human ΔMI (mean) | −0.230 | **−0.417** |
| Human ΔCogC (mean) | N/A | **+3.63** |
| Post-training-cutoff | <6% | **100%** |

Complexity metrics measured using [Wily v2](https://github.com/tonybaloney/wily):
- **ΔCC**: Cyclomatic Complexity change (McCabe, 1976)
- **ΔLLOC**: Logical Lines of Code change
- **ΔMI**: Maintainability Index change (Oman & Hagemeister, 1992)
- **ΔCogC**: Cognitive Complexity change (Campbell, 2018)

## Repository Distribution

| Repository | Instances |
|---|---|
| django/django | 38 |
| astropy/astropy | 22 |
| pydata/xarray | 17 |
| scikit-learn/scikit-learn | 14 |
| pylint-dev/pylint | 10 |
| matplotlib/matplotlib | 9 |
| sympy/sympy | 8 |
| pallets/flask | 1 |

## Selection Criteria

Instances were collected from merged pull requests in the SWE-bench ecosystem repositories with the following filters:

1. **Date range**: Merged January 1 – March 10, 2026 (post-training-cutoff)
2. **Issue linkage**: PR explicitly references a GitHub issue via "fixes #N" or equivalent
3. **Test coverage**: PR includes both implementation and test changes to Python files
4. **Minimum complexity**: Implementation patch modifies ≥4 changed lines
5. **Python files**: Only `.py` file changes retained
6. **Manual review**: Each candidate reviewed for solvability — documentation-only changes, large-scale refactors (>300 lines or >10 files), and tasks requiring external domain knowledge were excluded

From 1,043 scraped PRs → 712 with issue references → 224 after automated filters → **119 after manual review**.

## Schema

Each instance contains:

| Field | Type | Description |
|---|---|---|
| `instance_id` | string | Unique identifier (`{owner}__{repo}-{pr_number}`) |
| `repo` | string | GitHub repository (`owner/repo`) |
| `base_commit` | string | Parent commit SHA |
| `problem_statement` | string | GitHub issue text (title + body) |
| `test_patch` | string | Unified diff of test-file changes |
| `human_patch` | string | Unified diff of implementation-file changes |
| `pr_number` | int | Pull request number |
| `pr_url` | string | Pull request URL |
| `pr_merged_at` | string | Merge timestamp (ISO 8601) |
| `issue_number` | int | Referenced issue number |
| `issue_url` | string | Issue URL |
| `human_changed_lines` | int | Total changed lines in the human patch |
| `FAIL_TO_PASS` | string | JSON array of test IDs that must go FAIL→PASS |
| `PASS_TO_PASS` | string | JSON array of test IDs that must remain PASS |

## SWE-bench Compatibility

SWE-bench Complex uses the same schema as SWE-bench Verified and can be evaluated using the standard [SWE-bench harness](https://github.com/princeton-nlp/SWE-bench):

```bash

python -m swebench.harness.run_evaluation \

    -d anthonypjshaw/SWE-bench_Complex \

    -s test \

    -p predictions.jsonl \

    -id my_run \

    --max_workers 4

```

## Usage

```python

from datasets import load_dataset



dataset = load_dataset("anthonypjshaw/SWE-bench_Complex", split="test")

print(f"Tasks: {len(dataset)}")

print(f"Repos: {len(set(dataset['repo']))}")

```

## Citation

If you use SWE-bench Complex in your research, please cite:

```bibtex

@inproceedings{Shaw2026SWEbenchComplex,

  author    = {Shaw, Anthony},

  title     = {Beyond the Benchmark: A Contamination-Free Study of {AI} Code Complexity Across Four Frontier Models},

  booktitle = {Proceedings of the IEEE International Conference on Software Engineering (SSE)},

  year      = {2026},

}

```

## License

MIT License. The dataset contains references to publicly available open-source code under their respective licenses.