| --- |
| license: cc-by-4.0 |
| task_categories: |
| - question-answering |
| - text-generation |
| language: |
| - en |
| pretty_name: Competence-Based Evaluation (Invariance Benchmark) |
| size_categories: |
| - 10K<n<100K |
| tags: |
| - reasoning |
| - logical-reasoning |
| - invariance |
| - robustness |
| - benchmark |
| - sft |
| configs: |
| - config_name: eval_pos |
| data_files: |
| - split: original |
| path: eval/pos/original.jsonl |
| - split: equivalent |
| path: eval/pos/equivalent.jsonl |
| - config_name: eval_pos_largeN |
| data_files: |
| - split: original |
| path: eval/pos_largeN/original.jsonl |
| - split: equivalent |
| path: eval/pos_largeN/equivalent.jsonl |
| - config_name: eval_depth |
| data_files: |
| - split: original |
| path: eval/depth/original.jsonl |
| - split: equivalent |
| path: eval/depth/equivalent.jsonl |
| - config_name: sft_full |
| data_files: |
| - split: train |
| path: sft/full/train.jsonl |
| - split: validation |
| path: sft/full/val.jsonl |
| - config_name: sft_noleak |
| data_files: |
| - split: train |
| path: sft/noleak/train.jsonl |
| - split: validation |
| path: sft/noleak/val.jsonl |
| --- |
| |
| # Competence-Based Evaluation (Invariance Benchmark) |
|
|
| A benchmark for testing whether language models give the **same answer** to |
| semantically equivalent reformulations of a logical-ordering question. Given a |
| set of pairwise constraints (e.g. *Alice is in front of Bob*), a model should |
| answer transitive-closure queries (*Is Carol in front of Dave?*) consistently |
| whether the constraints are stated using a relation or its inverse. |
|
|
| Each item exists as a paired (`original`, `equivalent`) record describing the |
| same underlying ordering with different surface phrasings. **Invariance** is |
| measured as the agreement between the model's `original` and `equivalent` |
| answers; **accuracy** is measured against the ground-truth boolean. |
|
|
| ## Subsets |
|
|
| ### Evaluation (held-out) |
|
|
| | Config | Split | Rows | N range | Notes | |
| |---|---|---|---|---| |
| | `eval_pos` | `original`, `equivalent` | 4,000 each | 4–2048 | Main yes/no eval. Uses the held-out `pos` (in-front-of/behind) relation. Names list shown in the prompt is shuffled to remove the order-of-names leak. | |
| | `eval_pos_largeN` | `original`, `equivalent` | 1,200 each | up to several thousand | Stress test at large N. | |
| | `eval_depth` | `original`, `equivalent` | 2,000 each | 4–64 | Held-out `depth` (above/below stacking) relation, names-list shuffled. | |
|
|
| Each row is one yes/no question. Within a config, row `i` of the |
| `original` split and row `i` of the `equivalent` split describe the **same |
| underlying ordering** and the **same query**, only with the relation phrased |
| differently (e.g. "Alice in front of Bob" vs. "Bob behind Alice"). They share |
| the same ground-truth answer. |
|
|
| Schema: |
|
|
| ```json |
| { |
| "question": "There are 4 people standing in some order.\nTheir names are [...]\n...\nIs Nicholas in front of Thomas? Provide your answer only as yes or no. Answer: \n", |
| "answer": "yes", |
| "is_fwd": true, |
| "num_elements": 4 |
| } |
| ``` |
|
|
| ### Supervised fine-tuning |
|
|
| The SFT subsets are chat-formatted (`messages` field) and ready for |
| `trl.SFTTrainer` / OpenAI fine-tuning. They are built from a different set of |
| fact-agnostic relations than the eval set, with `n` skewed toward small values. |
| Each underlying ordering is expanded across `(is_fwd, answer)` combinations |
| × `(original, equivalent)` phrasing = 8 rows. |
|
|
| | Config | Split | Rows | Train relations | Notes | |
| |---|---|---|---|---| |
| | `sft_full` | `train` | 45,600 | arrival, priority, proximity, seniority, spatial_lr, spatial_ud | All fact-agnostic relations. | |
| | `sft_full` | `validation` | 2,400 | (same) | In-distribution validation split. | |
| | `sft_noleak` | `train` | 45,600 | (same as `sft_full`) | Built with `--shuffle-names-display` to remove the names-list leak; **this is the version used for the paper's reported fine-tuning results**. | |
| | `sft_noleak` | `validation` | 2,400 | (same) | In-distribution validation split. | |
|
|
| The `pos` and `depth` relations are **deliberately excluded** from training so |
| that the eval subsets remain genuinely out-of-distribution. |
|
|
| Schema (chat / messages format): |
|
|
| ```json |
| { |
| "messages": [ |
| {"role": "system", "content": "You are a helpful assistant. Answer logical reasoning questions concisely."}, |
| {"role": "user", "content": "There are 8 employees ... Is Juana more senior than Felecia? ..."}, |
| {"role": "assistant", "content": "no"} |
| ] |
| } |
| ``` |
|
|
| Per-config metadata (n distribution, per-relation counts, seed) lives in |
| `sft/full/meta.json` and `sft/noleak/meta.json`. |
|
|
| ## Loading |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Eval — paired splits, same row index = same underlying ordering. |
| ds = load_dataset("jizej/Competence-Based-Evaluation", "eval_pos") |
| org = ds["original"] |
| eqv = ds["equivalent"] |
| |
| # SFT — chat-formatted. |
| sft = load_dataset("jizej/Competence-Based-Evaluation", "sft_noleak") |
| train = sft["train"] |
| ``` |
|
|
| ## Source Datasets |
|
|
| The dataset is fully synthetic — no records are copied from another corpus. |
| However, the procedural generator draws **entity names** (people, animals, |
| cities, structures, etc.) from external knowledge sources. The complete list |
| of upstream source URIs is: |
|
|
| - **Wikidata SPARQL endpoint:** `https://query.wikidata.org/sparql` |
| Used for the `size_animals`, `height_structures`, `age_figures`, |
| `time_events`, `brightness_stars`, and `speed_animals` (Wikidata fallback) |
| pools. Queries are stored verbatim in |
| [`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py). |
| - **English Wikipedia REST API:** `https://en.wikipedia.org/api/rest_v1/page/html/...` |
| Specific source pages: |
| - `https://en.wikipedia.org/wiki/List_of_cities_by_average_temperature` |
| (`temperature_cities` pool) |
| - `https://en.wikipedia.org/wiki/Fastest_animals` (fallback for |
| `speed_animals`) |
| - **Curated lists** embedded in the generator script (no external URI): |
| `weight_objects`, `price_items`, `rank_athletes`, `spatial_objects`, and |
| the names list used by the `pos` and SFT relations. These are author- |
| maintained and are the only non-Wikidata/Wikipedia sources. |
| |
| Wikidata content is licensed CC0 and Wikipedia text is licensed CC BY-SA. |
| Cached responses for every pool are stored under `.entity_cache/{pool}.json` |
| in the open-source repository so the dataset can be regenerated bit-exact |
| without re-querying the upstream sources. |
|
|
| **Synthetic-generation seeds** that fully determine the released splits are |
| recorded in the per-subset `meta.json` files (`sft/full/meta.json`, |
| `sft/noleak/meta.json`); the seed used for the released SFT subsets is |
| `42`. Eval splits use deterministic enumeration over `(N, ordering, query)` |
| triples and require no random seed. |
|
|
| ## Provenance Activities |
|
|
| The end-to-end activities applied to produce this dataset are: |
|
|
| 1. **Collection (automated, online).** Entity pools fetched from Wikidata |
| via SPARQL and from Wikipedia via the REST API; see |
| [`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py). |
| Rate-limited with retries; results cached on disk. |
| 2. **Cleaning / filtering.** Per-pool deduplication (case-insensitive name |
| collapsing), removal of entries missing the relevant ground-truth value, |
| and merging of SPARQL results with curated fallback lists. For |
| `age_figures` the SPARQL query is split into three era-based |
| sub-queries to avoid `wikibase:sitelinks`-induced timeouts. |
| 3. **Curated fallback authoring.** Manual curation by the dataset authors |
| for the `weight_objects`, `price_items`, `rank_athletes`, |
| `spatial_objects`, and `names` pools (lists embedded directly in |
| `generate_entities.py` and `question_generation.py`). |
| 4. **Synthetic question generation.** Procedural construction of the |
| eval and SFT records in |
| [`invariance_bench/question_generation.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/question_generation.py) |
| and the entry-point scripts |
| [`scripts/generate_dataset.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_dataset.py), |
| [`scripts/generate_heldout_dataset.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_heldout_dataset.py), |
| and |
| [`scripts/generate_training_data.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/scripts/generate_training_data.py). |
| This step is fully deterministic given the seed and entity pools. |
| 5. **Annotation.** None. There is no human-annotation step. All |
| `answer` / `messages` ground-truth labels are produced by the same |
| deterministic generator that creates the question text, and are derived |
| from the synthesized ground-truth ordering, not from human judgment. |
| 6. **Synthetic agents / LLMs.** **None.** No language model, embedding |
| model, or generative agent is used at any step in the pipeline. |
| 7. **Crowdsourcing platforms / human teams.** Not applicable — no |
| crowdsourcing, no human raters, no annotation contractors were |
| involved. |
| 8. **Validation / leak audit.** The released `_shufnames` / `noleak` |
| subsets were produced after an internal audit revealed that an earlier |
| version's prompt-side names list ordering correlated with the answer. |
| The audit is documented in `docs/paper_methodology_experiments.md` of |
| the open-source repository. |
|
|
| ## Construction (Synthetic-Data Generation Process) |
|
|
| **All records in this dataset are synthetic.** They are produced by a |
| deterministic procedural generator; no model-based generation, no human |
| annotation, and no scraped natural-language Q&A is used in the pipeline. |
|
|
| The generation process is: |
|
|
| 1. **Entity pools.** Names of entities (animals, structures, people, cities, |
| events, stars, etc.) are sourced from Wikidata SPARQL queries, Wikipedia |
| HTML tables, and small curated fallback lists embedded in the generator. |
| Each pool is cached on disk as JSON. See |
| [`invariance_bench/generate_entities.py`](https://github.com/jizej/Competence-Based-Evaluation/blob/main/invariance_bench/generate_entities.py). |
| 2. **Ordering sampling.** For each (relation, `n`) bucket the generator |
| samples a random permutation of `n` entities from the appropriate pool |
| and lays out the chain implied by the relation (e.g. *front-of* / *behind*). |
| 3. **Constraint expansion.** A subset of consecutive pairs is selected to |
| form the "rules" shown in the prompt; the unstated remainder is what the |
| transitive-closure query exercises. |
| 4. **Phrasing duplication.** Every ordering is rendered twice: once with |
| the canonical relation (`original`) and once with the logically inverse |
| relation (`equivalent`). The two renderings carry the same ground-truth |
| boolean answer. |
| 5. **Yes/no query selection.** A query pair `(a, b)` is sampled at a |
| configured minimum hop distance, with the ground-truth `yes`/`no` answer |
| balanced by construction. |
| 6. **(SFT subsets only) Chat formatting.** Each (ordering, query, phrasing) |
| triple is serialized into a `messages` array with system / user / |
| assistant turns ready for SFT trainers. |
|
|
| All generation seeds, the per-relation count distributions, the `N` |
| schedule, and the held-out relation list are recorded in the per-subset |
| `meta.json`. The full pipeline is reproducible from the open-source |
| repository linked above. |
|
|
| ## Intended Use Cases |
|
|
| The dataset is designed to measure **answer-level invariance** of language |
| models under semantically-preserving paraphrasing of logical-ordering |
| constraints. Concretely: |
|
|
| - **Primary use case (validated):** measuring whether a model returns the |
| same boolean answer to a transitive-closure query when the underlying |
| ordering is described with a relation versus its inverse. Validation is |
| reported in our accompanying NeurIPS 2026 D&B submission across |
| proprietary and open-weight models. |
| - **Primary use case (validated):** comparing pre- and post-fine-tuning |
| checkpoints to verify that targeted SFT improves invariance without |
| destroying out-of-distribution generalization (held-out `pos` and `depth` |
| relations). |
| - **Secondary use case (partially validated):** scaling-law style analyses |
| of invariance vs. accuracy as a function of `N` (the number of entities |
| in the ordering). Validated for `N ∈ [4, 2048]`; behavior beyond this |
| range is not characterized. |
| - **Secondary use case (not validated here):** as a regression test for |
| training pipelines that aim to preserve symbolic reasoning under |
| paraphrase. We provide the data; we do not certify any specific training |
| recipe. |
|
|
| Use cases for which validation **does not** exist or may not hold: |
| general-reasoning leaderboard ranking, safety / alignment evaluation, |
| detection of jailbreaks or adversarial prompts, multilingual robustness, |
| evaluation of long-form generation quality, and any clinical, legal, or |
| high-stakes decision-support setting. |
|
|
| ## Personal and Sensitive Information |
|
|
| The dataset contains **no real personal data, no real PII, and no health, |
| medical, financial, biometric, political, or religious data about |
| identifiable individuals**. All "people" in the prompts are synthetic |
| references constructed by sampling from entity pools. |
|
|
| The following indirect demographic signals are present and should be |
| declared: |
|
|
| - **Gender (indirect, via names).** First names sampled from US-style name |
| lists carry conventional masculine/feminine associations. No gender label |
| is attached to any record; gender is only implicit in the name token. |
| - **Geography.** Pools such as `temperature_cities`, `height_structures`, |
| and `time_events` contain real geographic place names sourced from |
| Wikidata and Wikipedia. These pools are skewed toward globally prominent, |
| English-Wikipedia-covered locations. |
| - **Language.** Prompts and answers are exclusively in English; this is a |
| deliberate scope restriction, not a privacy signal, but it is recorded |
| here for completeness. |
| - **Culture.** Entity selection inherits the cultural skew of Wikidata / |
| English Wikipedia (Western, anglophone over-representation). |
| - **Age (of historical figures only).** The `age_figures` pool references |
| real historical figures with their public birth years. These are |
| deceased public figures whose biographical data is already published on |
| Wikidata; no contemporary individuals' ages are present. |
|
|
| The following are **not** present: socio-economic status of identifiable |
| individuals, professional experience or seniority of identifiable |
| individuals (the `seniority` and `priority` relations operate on synthetic |
| placeholders, not on real employees or rankings), health or medical data, |
| political affiliation, and religious belief. |
|
|
| No data subjects were contacted or surveyed in producing this dataset, so |
| no consent or withdrawal procedures apply. Wikidata is licensed CC0 and |
| Wikipedia is licensed CC BY-SA; both permit redistribution of the entity |
| metadata used here. |
|
|
| ## Social Impact |
|
|
| **Intended positive impact.** Releasing a clean invariance benchmark |
| encourages the field to evaluate language models on robustness to |
| paraphrase, not only on accuracy. Reproducible held-out splits and an |
| open-source generator make it harder for the benchmark to be quietly |
| over-fit, and the SFT subsets give researchers a concrete starting point |
| for studying targeted invariance training. |
|
|
| **Potential negative impact and risks of misuse.** |
|
|
| - *Over-claiming general reasoning.* High invariance scores on this |
| dataset measure invariance on transitive ordering only. A naive reader |
| could mistake them for evidence of general reasoning robustness; results |
| should always be reported with the scope of the benchmark stated. |
| - *Skill leaderboarding pressure.* As with any public benchmark, optimizing |
| directly against this dataset risks Goodharting — gains here may not |
| transfer to natural-language reasoning. We encourage reporting paired |
| held-out evaluations from other benchmarks. |
| - *Cultural / linguistic skew.* Because entity pools are anglocentric, |
| models tuned on this data may improve on similarly-distributed inputs |
| while showing little transfer to non-English or non-Western surface |
| forms. |
| - *Indirect demographic correlations.* US-style first names carry |
| conventional gender signals. If a downstream model is trained on the SFT |
| subsets in a way that picks up name-conditioned heuristics, that bias |
| will propagate. Users training on this data should audit for gendered |
| response patterns. |
|
|
| **Mitigations in this release.** |
|
|
| - The dataset is open-license (CC BY 4.0) but **gated by deliberate |
| narrowness of scope** rather than access controls: every record is |
| explicitly a synthetic transitive-ordering question, and the dataset |
| card states the intended-use boundaries above. |
| - Held-out relations (`pos`, `depth`) are **excluded from the SFT |
| subsets** so OOD generalization claims remain defensible. |
| - The earlier internal `_shufnames` / `noleak` audit (where the displayed |
| names list accidentally encoded the answer) is documented above; the |
| released eval and SFT files have the leak fixed. |
| - The generator is open-source, allowing external auditors to reproduce |
| every record from a documented seed. |
|
|
| No usage gating, embargo, or differential-access controls are applied. |
| Users are expected to follow the limitations and intended-use guidance |
| above and to cite the dataset when reporting results. |
|
|
| ## Limitations |
|
|
| - **Narrow reasoning skill.** Each question tests transitive closure over a |
| linear ordering induced by a single binary relation. Performance here does |
| not generalize to multi-step natural-language reasoning, common-sense |
| inference, math, code, or any non-ordering relational structure. |
| - **Synthetic phrasings.** Questions are produced by a small grammar (a fixed |
| template per relation) rather than written by humans, so surface-form |
| diversity is limited. Distributional gaps relative to natural prose, |
| conversational queries, or noisy real-world text are large. |
| - **English only.** All prompts and answers are English. The benchmark says |
| nothing about cross-lingual robustness. |
| - **Yes/no output space.** The eval rewards a literal `yes` or `no` token. |
| Models that hedge, refuse, or emit verbose chains of thought without a |
| committed answer score zero on accuracy and invariance regardless of |
| whether the underlying reasoning is correct. Practitioners using CoT-style |
| models should add an answer-extraction step (see `invariance_bench/scoring.py`). |
| - **Single deterministic ground truth.** The eval does not measure |
| calibration, uncertainty, or partial credit; orderings with ties or |
| under-specified constraints are not represented. |
| - **Long-context confound.** At large `N` (especially in `eval_pos_largeN` |
| and the `N=2048` slice of `eval_pos`), prompts can exceed the effective |
| context window of many models. Failures at large `N` may reflect context |
| handling rather than reasoning ability and should not be interpreted as |
| pure invariance violations. |
| - **Held-out coverage.** The OOD evaluation surface is two relations (`pos`, |
| `depth`); the benchmark cannot verify whether a model's invariance |
| generalizes to relations beyond those seen at train *and* eval time. |
| - **Names-list leak in earlier internal versions.** Released `_shufnames` |
| eval files and the `sft_noleak` training files **do not** have this leak. |
| Older `base2_*` artifacts (not released on HF) did, and any third-party |
| reuse of those files would over-estimate model performance. |
|
|
| **Not recommended for:** general reasoning leaderboards, safety/alignment |
| evaluation, multilingual evaluation, evaluating models whose primary output |
| mode is a long chain of thought without an extractable boolean answer. |
|
|
| ## Biases |
|
|
| - **Anglo/Western entity skew.** The `names` pool used by the `pos`-relation |
| questions and by the SFT data is drawn from US-style first-name lists, so |
| most prompts contain English-coded given names. The `temperature_cities`, |
| `height_structures`, and `time_events` pools likewise over-represent |
| Wikipedia/Wikidata-prominent (largely Western, English-language) |
| entities. Under-represented populations include non-Western cultures and |
| languages whose entities have lower Wikipedia coverage. |
| - **Source-driven content bias.** Wikidata and Wikipedia are themselves known |
| to be skewed toward male, Western, and modern-era subjects (especially in |
| `age_figures`). The benchmark inherits these biases. Curated fallback |
| lists for `weight_objects`, `price_items`, and `rank_athletes` reflect the |
| authors' own selections and are not demographically balanced. |
| - **Relation-template bias.** Each relation has one canonical phrasing and |
| one inverse phrasing. The grammar does not exercise the full space of |
| English ways to express ordering (passive voice, comparative clauses, |
| idiomatic expressions, etc.), so reported invariance is a conservative |
| lower bound: a model that is invariant on this dataset may still be |
| sensitive to other surface variations. |
| - **Position-of-name leak (mitigated).** In an earlier internal version, |
| the order of names listed in the prompt correlated with their position in |
| the underlying ordering, which models could exploit without reading the |
| rules. Released eval files (`*_shufnames.jsonl`) and the `sft_noleak` |
| subset shuffle the displayed names list to remove this leak. Users |
| regenerating data with the included scripts must pass |
| `--shuffle-names-display` to reproduce the no-leak setting. |
| - **Train/eval relation leakage controls.** `pos` (front/behind) and `depth` |
| (above/below) are deliberately held out of the SFT data so they remain |
| OOD for fine-tuned checkpoints. Mixing the SFT subsets with held-out |
| evaluation defeats the OOD claim. |
| - **Per-`N` row-count imbalance.** Both eval and SFT skew toward small `N` |
| (the SFT distribution explicitly down-weights large `N`). Aggregate |
| metrics across `N` are therefore dominated by the small-`N` regime; |
| report per-`N` numbers when comparing models. |
|
|
| ## License |
|
|
| Released under **CC BY 4.0**. Entity names sourced from Wikidata/Wikipedia |
| retain their original licenses (CC0 / CC BY-SA). |
|
|
| ## Citation |
|
|
| Please cite the accompanying paper if you use this dataset (citation TBD — |
| NeurIPS 2026 Datasets & Benchmarks track submission). |
|
|