File size: 22,405 Bytes
7ec4870 0cff67f 7ec4870 0cff67f 7ec4870 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 | ---
license: cc-by-4.0
task_categories:
- question-answering
- text-generation
language:
- en
pretty_name: Competence-Based Evaluation (Invariance Benchmark)
size_categories:
- 10K<n<100K
tags:
- reasoning
- logical-reasoning
- invariance
- robustness
- benchmark
- sft
configs:
- config_name: eval_pos
data_files:
- split: original
path: eval/pos/original.jsonl
- split: equivalent
path: eval/pos/equivalent.jsonl
- config_name: eval_pos_largeN
data_files:
- split: original
path: eval/pos_largeN/original.jsonl
- split: equivalent
path: eval/pos_largeN/equivalent.jsonl
- config_name: eval_depth
data_files:
- split: original
path: eval/depth/original.jsonl
- split: equivalent
path: eval/depth/equivalent.jsonl
- config_name: sft_full
data_files:
- split: train
path: sft/full/train.jsonl
- split: validation
path: sft/full/val.jsonl
- config_name: sft_noleak
data_files:
- split: train
path: sft/noleak/train.jsonl
- split: validation
path: sft/noleak/val.jsonl
---
# Competence-Based Evaluation (Invariance Benchmark)
A benchmark for testing whether language models give the **same answer** to
semantically equivalent reformulations of a logical-ordering question. Given a
set of pairwise constraints (e.g. *Alice is in front of Bob*), a model should
answer transitive-closure queries (*Is Carol in front of Dave?*) consistently
whether the constraints are stated using a relation or its inverse.
Each item exists as a paired (`original`, `equivalent`) record describing the
same underlying ordering with different surface phrasings. **Invariance** is
measured as the agreement between the model's `original` and `equivalent`
answers; **accuracy** is measured against the ground-truth boolean.
## Subsets
### Evaluation (held-out)
| Config | Split | Rows | N range | Notes |
|---|---|---|---|---|
| `eval_pos` | `original`, `equivalent` | 4,000 each | 4–2048 | Main yes/no eval. Uses the held-out `pos` (in-front-of/behind) relation. Names list shown in the prompt is shuffled to remove the order-of-names leak. |
| `eval_pos_largeN` | `original`, `equivalent` | 1,200 each | up to several thousand | Stress test at large N. |
| `eval_depth` | `original`, `equivalent` | 2,000 each | 4–64 | Held-out `depth` (above/below stacking) relation, names-list shuffled. |
Each row is one yes/no question. Within a config, row `i` of the
`original` split and row `i` of the `equivalent` split describe the **same
underlying ordering** and the **same query**, only with the relation phrased
differently (e.g. "Alice in front of Bob" vs. "Bob behind Alice"). They share
the same ground-truth answer.
Schema:
```json
{
"question": "There are 4 people standing in some order.\nTheir names are [...]\n...\nIs Nicholas in front of Thomas? Provide your answer only as yes or no. Answer: \n",
"answer": "yes",
"is_fwd": true,
"num_elements": 4
}
```
### Supervised fine-tuning
The SFT subsets are chat-formatted (`messages` field) and ready for
`trl.SFTTrainer` / OpenAI fine-tuning. They are built from a different set of
fact-agnostic relations than the eval set, with `n` skewed toward small values.
Each underlying ordering is expanded across `(is_fwd, answer)` combinations
× `(original, equivalent)` phrasing = 8 rows.
| Config | Split | Rows | Train relations | Notes |
|---|---|---|---|---|
| `sft_full` | `train` | 45,600 | arrival, priority, proximity, seniority, spatial_lr, spatial_ud | All fact-agnostic relations. |
| `sft_full` | `validation` | 2,400 | (same) | In-distribution validation split. |
| `sft_noleak` | `train` | 45,600 | (same as `sft_full`) | Built with `--shuffle-names-display` to remove the names-list leak; **this is the version used for the paper's reported fine-tuning results**. |
| `sft_noleak` | `validation` | 2,400 | (same) | In-distribution validation split. |
The `pos` and `depth` relations are **deliberately excluded** from training so
that the eval subsets remain genuinely out-of-distribution.
Schema (chat / messages format):
```json
{
"messages": [
{"role": "system", "content": "You are a helpful assistant. Answer logical reasoning questions concisely."},
{"role": "user", "content": "There are 8 employees ... Is Juana more senior than Felecia? ..."},
{"role": "assistant", "content": "no"}
]
}
```
Per-config metadata (n distribution, per-relation counts, seed) lives in
`sft/full/meta.json` and `sft/noleak/meta.json`.
## Loading
```python
from datasets import load_dataset
# Eval — paired splits, same row index = same underlying ordering.
ds = load_dataset("jizej/Competence-Based-Evaluation", "eval_pos")
org = ds["original"]
eqv = ds["equivalent"]
# SFT — chat-formatted.
sft = load_dataset("jizej/Competence-Based-Evaluation", "sft_noleak")
train = sft["train"]
```
## Source Datasets
The dataset is fully synthetic — no records are copied from another corpus.
However, the procedural generator draws **entity names** (people, animals,
cities, structures, etc.) from external knowledge sources. The complete list
of upstream source URIs is:
- **Wikidata SPARQL endpoint:** `https://query.wikidata.org/sparql`
Used for the `size_animals`, `height_structures`, `age_figures`,
`time_events`, `brightness_stars`, and `speed_animals` (Wikidata fallback)
pools. Queries are stored verbatim in
[`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py).
- **English Wikipedia REST API:** `https://en.wikipedia.org/api/rest_v1/page/html/...`
Specific source pages:
- `https://en.wikipedia.org/wiki/List_of_cities_by_average_temperature`
(`temperature_cities` pool)
- `https://en.wikipedia.org/wiki/Fastest_animals` (fallback for
`speed_animals`)
- **Curated lists** embedded in the generator script (no external URI):
`weight_objects`, `price_items`, `rank_athletes`, `spatial_objects`, and
the names list used by the `pos` and SFT relations. These are author-
maintained and are the only non-Wikidata/Wikipedia sources.
Wikidata content is licensed CC0 and Wikipedia text is licensed CC BY-SA.
Cached responses for every pool are stored under `.entity_cache/{pool}.json`
in the open-source repository so the dataset can be regenerated bit-exact
without re-querying the upstream sources.
**Synthetic-generation seeds** that fully determine the released splits are
recorded in the per-subset `meta.json` files (`sft/full/meta.json`,
`sft/noleak/meta.json`); the seed used for the released SFT subsets is
`42`. Eval splits use deterministic enumeration over `(N, ordering, query)`
triples and require no random seed.
## Provenance Activities
The end-to-end activities applied to produce this dataset are:
1. **Collection (automated, online).** Entity pools fetched from Wikidata
via SPARQL and from Wikipedia via the REST API; see
[`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py).
Rate-limited with retries; results cached on disk.
2. **Cleaning / filtering.** Per-pool deduplication (case-insensitive name
collapsing), removal of entries missing the relevant ground-truth value,
and merging of SPARQL results with curated fallback lists. For
`age_figures` the SPARQL query is split into three era-based
sub-queries to avoid `wikibase:sitelinks`-induced timeouts.
3. **Curated fallback authoring.** Manual curation by the dataset authors
for the `weight_objects`, `price_items`, `rank_athletes`,
`spatial_objects`, and `names` pools (lists embedded directly in
`generate_entities.py` and `question_generation.py`).
4. **Synthetic question generation.** Procedural construction of the
eval and SFT records in
[`invariance_bench/question_generation.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/question_generation.py)
and the entry-point scripts
[`scripts/generate_dataset.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_dataset.py),
[`scripts/generate_heldout_dataset.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_heldout_dataset.py),
and
[`scripts/generate_training_data.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_training_data.py).
This step is fully deterministic given the seed and entity pools.
5. **Annotation.** None. There is no human-annotation step. All
`answer` / `messages` ground-truth labels are produced by the same
deterministic generator that creates the question text, and are derived
from the synthesized ground-truth ordering, not from human judgment.
6. **Synthetic agents / LLMs.** **None.** No language model, embedding
model, or generative agent is used at any step in the pipeline.
7. **Crowdsourcing platforms / human teams.** Not applicable — no
crowdsourcing, no human raters, no annotation contractors were
involved.
8. **Validation / leak audit.** The released `_shufnames` / `noleak`
subsets were produced after an internal audit revealed that an earlier
version's prompt-side names list ordering correlated with the answer.
The audit is documented in `docs/paper_methodology_experiments.md` of
the open-source repository.
## Construction (Synthetic-Data Generation Process)
**All records in this dataset are synthetic.** They are produced by a
deterministic procedural generator; no model-based generation, no human
annotation, and no scraped natural-language Q&A is used in the pipeline.
The generation process is:
1. **Entity pools.** Names of entities (animals, structures, people, cities,
events, stars, etc.) are sourced from Wikidata SPARQL queries, Wikipedia
HTML tables, and small curated fallback lists embedded in the generator.
Each pool is cached on disk as JSON. See
[`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py).
2. **Ordering sampling.** For each (relation, `n`) bucket the generator
samples a random permutation of `n` entities from the appropriate pool
and lays out the chain implied by the relation (e.g. *front-of* / *behind*).
3. **Constraint expansion.** A subset of consecutive pairs is selected to
form the "rules" shown in the prompt; the unstated remainder is what the
transitive-closure query exercises.
4. **Phrasing duplication.** Every ordering is rendered twice: once with
the canonical relation (`original`) and once with the logically inverse
relation (`equivalent`). The two renderings carry the same ground-truth
boolean answer.
5. **Yes/no query selection.** A query pair `(a, b)` is sampled at a
configured minimum hop distance, with the ground-truth `yes`/`no` answer
balanced by construction.
6. **(SFT subsets only) Chat formatting.** Each (ordering, query, phrasing)
triple is serialized into a `messages` array with system / user /
assistant turns ready for SFT trainers.
All generation seeds, the per-relation count distributions, the `N`
schedule, and the held-out relation list are recorded in the per-subset
`meta.json`. The full pipeline is reproducible from the open-source
repository linked above.
## Intended Use Cases
The dataset is designed to measure **answer-level invariance** of language
models under semantically-preserving paraphrasing of logical-ordering
constraints. Concretely:
- **Primary use case (validated):** measuring whether a model returns the
same boolean answer to a transitive-closure query when the underlying
ordering is described with a relation versus its inverse. Validation is
reported in our accompanying NeurIPS 2026 D&B submission across
proprietary and open-weight models.
- **Primary use case (validated):** comparing pre- and post-fine-tuning
checkpoints to verify that targeted SFT improves invariance without
destroying out-of-distribution generalization (held-out `pos` and `depth`
relations).
- **Secondary use case (partially validated):** scaling-law style analyses
of invariance vs. accuracy as a function of `N` (the number of entities
in the ordering). Validated for `N ∈ [4, 2048]`; behavior beyond this
range is not characterized.
- **Secondary use case (not validated here):** as a regression test for
training pipelines that aim to preserve symbolic reasoning under
paraphrase. We provide the data; we do not certify any specific training
recipe.
Use cases for which validation **does not** exist or may not hold:
general-reasoning leaderboard ranking, safety / alignment evaluation,
detection of jailbreaks or adversarial prompts, multilingual robustness,
evaluation of long-form generation quality, and any clinical, legal, or
high-stakes decision-support setting.
## Personal and Sensitive Information
The dataset contains **no real personal data, no real PII, and no health,
medical, financial, biometric, political, or religious data about
identifiable individuals**. All "people" in the prompts are synthetic
references constructed by sampling from entity pools.
The following indirect demographic signals are present and should be
declared:
- **Gender (indirect, via names).** First names sampled from US-style name
lists carry conventional masculine/feminine associations. No gender label
is attached to any record; gender is only implicit in the name token.
- **Geography.** Pools such as `temperature_cities`, `height_structures`,
and `time_events` contain real geographic place names sourced from
Wikidata and Wikipedia. These pools are skewed toward globally prominent,
English-Wikipedia-covered locations.
- **Language.** Prompts and answers are exclusively in English; this is a
deliberate scope restriction, not a privacy signal, but it is recorded
here for completeness.
- **Culture.** Entity selection inherits the cultural skew of Wikidata /
English Wikipedia (Western, anglophone over-representation).
- **Age (of historical figures only).** The `age_figures` pool references
real historical figures with their public birth years. These are
deceased public figures whose biographical data is already published on
Wikidata; no contemporary individuals' ages are present.
The following are **not** present: socio-economic status of identifiable
individuals, professional experience or seniority of identifiable
individuals (the `seniority` and `priority` relations operate on synthetic
placeholders, not on real employees or rankings), health or medical data,
political affiliation, and religious belief.
No data subjects were contacted or surveyed in producing this dataset, so
no consent or withdrawal procedures apply. Wikidata is licensed CC0 and
Wikipedia is licensed CC BY-SA; both permit redistribution of the entity
metadata used here.
## Social Impact
**Intended positive impact.** Releasing a clean invariance benchmark
encourages the field to evaluate language models on robustness to
paraphrase, not only on accuracy. Reproducible held-out splits and an
open-source generator make it harder for the benchmark to be quietly
over-fit, and the SFT subsets give researchers a concrete starting point
for studying targeted invariance training.
**Potential negative impact and risks of misuse.**
- *Over-claiming general reasoning.* High invariance scores on this
dataset measure invariance on transitive ordering only. A naive reader
could mistake them for evidence of general reasoning robustness; results
should always be reported with the scope of the benchmark stated.
- *Skill leaderboarding pressure.* As with any public benchmark, optimizing
directly against this dataset risks Goodharting — gains here may not
transfer to natural-language reasoning. We encourage reporting paired
held-out evaluations from other benchmarks.
- *Cultural / linguistic skew.* Because entity pools are anglocentric,
models tuned on this data may improve on similarly-distributed inputs
while showing little transfer to non-English or non-Western surface
forms.
- *Indirect demographic correlations.* US-style first names carry
conventional gender signals. If a downstream model is trained on the SFT
subsets in a way that picks up name-conditioned heuristics, that bias
will propagate. Users training on this data should audit for gendered
response patterns.
**Mitigations in this release.**
- The dataset is open-license (CC BY 4.0) but **gated by deliberate
narrowness of scope** rather than access controls: every record is
explicitly a synthetic transitive-ordering question, and the dataset
card states the intended-use boundaries above.
- Held-out relations (`pos`, `depth`) are **excluded from the SFT
subsets** so OOD generalization claims remain defensible.
- The earlier internal `_shufnames` / `noleak` audit (where the displayed
names list accidentally encoded the answer) is documented above; the
released eval and SFT files have the leak fixed.
- The generator is open-source, allowing external auditors to reproduce
every record from a documented seed.
No usage gating, embargo, or differential-access controls are applied.
Users are expected to follow the limitations and intended-use guidance
above and to cite the dataset when reporting results.
## Limitations
- **Narrow reasoning skill.** Each question tests transitive closure over a
linear ordering induced by a single binary relation. Performance here does
not generalize to multi-step natural-language reasoning, common-sense
inference, math, code, or any non-ordering relational structure.
- **Synthetic phrasings.** Questions are produced by a small grammar (a fixed
template per relation) rather than written by humans, so surface-form
diversity is limited. Distributional gaps relative to natural prose,
conversational queries, or noisy real-world text are large.
- **English only.** All prompts and answers are English. The benchmark says
nothing about cross-lingual robustness.
- **Yes/no output space.** The eval rewards a literal `yes` or `no` token.
Models that hedge, refuse, or emit verbose chains of thought without a
committed answer score zero on accuracy and invariance regardless of
whether the underlying reasoning is correct. Practitioners using CoT-style
models should add an answer-extraction step (see `invariance_bench/scoring.py`).
- **Single deterministic ground truth.** The eval does not measure
calibration, uncertainty, or partial credit; orderings with ties or
under-specified constraints are not represented.
- **Long-context confound.** At large `N` (especially in `eval_pos_largeN`
and the `N=2048` slice of `eval_pos`), prompts can exceed the effective
context window of many models. Failures at large `N` may reflect context
handling rather than reasoning ability and should not be interpreted as
pure invariance violations.
- **Held-out coverage.** The OOD evaluation surface is two relations (`pos`,
`depth`); the benchmark cannot verify whether a model's invariance
generalizes to relations beyond those seen at train *and* eval time.
- **Names-list leak in earlier internal versions.** Released `_shufnames`
eval files and the `sft_noleak` training files **do not** have this leak.
Older `base2_*` artifacts (not released on HF) did, and any third-party
reuse of those files would over-estimate model performance.
**Not recommended for:** general reasoning leaderboards, safety/alignment
evaluation, multilingual evaluation, evaluating models whose primary output
mode is a long chain of thought without an extractable boolean answer.
## Biases
- **Anglo/Western entity skew.** The `names` pool used by the `pos`-relation
questions and by the SFT data is drawn from US-style first-name lists, so
most prompts contain English-coded given names. The `temperature_cities`,
`height_structures`, and `time_events` pools likewise over-represent
Wikipedia/Wikidata-prominent (largely Western, English-language)
entities. Under-represented populations include non-Western cultures and
languages whose entities have lower Wikipedia coverage.
- **Source-driven content bias.** Wikidata and Wikipedia are themselves known
to be skewed toward male, Western, and modern-era subjects (especially in
`age_figures`). The benchmark inherits these biases. Curated fallback
lists for `weight_objects`, `price_items`, and `rank_athletes` reflect the
authors' own selections and are not demographically balanced.
- **Relation-template bias.** Each relation has one canonical phrasing and
one inverse phrasing. The grammar does not exercise the full space of
English ways to express ordering (passive voice, comparative clauses,
idiomatic expressions, etc.), so reported invariance is a conservative
lower bound: a model that is invariant on this dataset may still be
sensitive to other surface variations.
- **Position-of-name leak (mitigated).** In an earlier internal version,
the order of names listed in the prompt correlated with their position in
the underlying ordering, which models could exploit without reading the
rules. Released eval files (`*_shufnames.jsonl`) and the `sft_noleak`
subset shuffle the displayed names list to remove this leak. Users
regenerating data with the included scripts must pass
`--shuffle-names-display` to reproduce the no-leak setting.
- **Train/eval relation leakage controls.** `pos` (front/behind) and `depth`
(above/below) are deliberately held out of the SFT data so they remain
OOD for fine-tuned checkpoints. Mixing the SFT subsets with held-out
evaluation defeats the OOD claim.
- **Per-`N` row-count imbalance.** Both eval and SFT skew toward small `N`
(the SFT distribution explicitly down-weights large `N`). Aggregate
metrics across `N` are therefore dominated by the small-`N` regime;
report per-`N` numbers when comparing models.
## License
Released under **CC BY 4.0**. Entity names sourced from Wikidata/Wikipedia
retain their original licenses (CC0 / CC BY-SA).
## Citation
Please cite the accompanying paper if you use this dataset (citation TBD —
NeurIPS 2026 Datasets & Benchmarks track submission).
|