tensorbench-1.0 / croissant.json
tensorbench's picture
tensorbench dataset
0e74d35
{
"@context": {
"@language": "en",
"@vocab": "https://schema.org/",
"citeAs": "cr:citeAs",
"column": "cr:column",
"conformsTo": "dct:conformsTo",
"cr": "http://mlcommons.org/croissant/",
"rai": "http://mlcommons.org/croissant/RAI/",
"data": {
"@id": "cr:data",
"@type": "@json"
},
"dataType": {
"@id": "cr:dataType",
"@type": "@vocab"
},
"dct": "http://purl.org/dc/terms/",
"equivalentProperty": {
"@id": "cr:equivalentProperty",
"@type": "@vocab"
},
"samplingRate": "cr:samplingRate",
"examples": {
"@id": "cr:examples",
"@type": "@json"
},
"extract": "cr:extract",
"field": "cr:field",
"fileProperty": "cr:fileProperty",
"fileObject": "cr:fileObject",
"fileSet": "cr:fileSet",
"format": "cr:format",
"includes": "cr:includes",
"isLiveDataset": "cr:isLiveDataset",
"jsonPath": "cr:jsonPath",
"key": "cr:key",
"md5": "cr:md5",
"parentField": "cr:parentField",
"path": "cr:path",
"recordSet": "cr:recordSet",
"references": "cr:references",
"regex": "cr:regex",
"repeated": "cr:repeated",
"replace": "cr:replace",
"sc": "https://schema.org/",
"separator": "cr:separator",
"source": "cr:source",
"subField": "cr:subField",
"transform": "cr:transform"
},
"@type": "sc:Dataset",
"conformsTo": "http://mlcommons.org/croissant/1.0",
"name": "TensorBench",
"description": "Feature-addition benchmark for LLMs and coding agents, evaluated against the Scorch codebase (a sparse+dense PyTorch compiler). Each task asks a model to add a feature; success is the full pytest suite passing after the patch is applied inside a Docker container. 199 tasks total (194 feature-addition, 5 refactor) across 5 base commits on the upstream `bench` branch.",
"url": "https://huggingface.co/datasets/tensorbench/tensorbench-1.0",
"license": "https://opensource.org/licenses/MIT",
"version": "1.0.0",
"citeAs": "(Anonymized for double-blind review.)",
"datePublished": "2026-05-06",
"keywords": [
"benchmark",
"code-generation",
"llm-evaluation",
"feature-addition",
"python",
"pytorch",
"sparse-tensors"
],
"rai:dataCollection": "Each task is a natural-language prompt paired with a `base_commit` SHA from the upstream Scorch repository's `bench` branch. Prompts were drafted with LLM assistance against the target codebase and then reviewed and curated by the authors. The upstream codebase, tests, and SHAs are pre-existing artifacts of the Scorch project.",
"rai:annotationsPerItem": "Each task has a single author-curated description and a fixed `base_commit`. There are no per-item human annotations beyond the prompt itself; success is computed automatically by running the patched repo's pytest suite.",
"rai:dataLimitations": "All tasks target a single Python codebase (Scorch). Performance does not generalize across languages or unrelated repositories. Test suites authored by the agent are accepted at face value, which can permit a small number of vacuous tests; the paper's adversarial-behavior audit (see supplementary code) characterizes this rate. Container builds depend on a network clone of the upstream repository and may be affected by upstream availability.",
"rai:dataBiases": "All tasks were curated by the same set of authors. The category distribution skews toward runtime/dispatch, scheduler/loop transformations, linear-algebra ops, sparse formats, and IR/codegen surface area — i.e. the parts of a sparse-tensor compiler the authors found tractable to specify. Because prompts were drafted with LLM assistance, phrasing and structure also reflect the drafting model's stylistic patterns. The benchmark therefore measures a slice of `extend an existing PyTorch-extension codebase`-style work, not all of `software engineering with LLMs`.",
"rai:dataUseCases": "Evaluating coding agents and LLMs on extending a real, non-trivial Python+C++ codebase. Useful for: comparing agent frameworks (e.g. Claude Code, OpenAI Codex CLI, Gemini CLI, OpenHands), comparing model capabilities at fixed agent harness, and characterizing failure modes (test fail, patch apply fail, timeout, empty patch, vacuous added tests).",
"rai:personalSensitiveInformation": "None. Tasks describe code changes; no human-subject data is involved.",
"rai:dataSocialImpact": "The benchmark is intended for AI/ML research on code-generation evaluation. It does not target sensitive domains (medical, legal, financial). Like other code-evaluation benchmarks, it could in principle be used to filter or rank developers, which is not its intended use.",
"rai:hasSyntheticData": "Task prompts were drafted with LLM assistance against the target codebase and then reviewed and curated by the authors. The underlying `base_commit` SHAs, source code, and pytest infrastructure are pre-existing upstream artifacts and were not generated.",
"rai:dataReleaseMaintenancePlan": "The dataset will be made public under the authors' real names at the camera-ready deadline. Future revisions will be tagged via the `version` field. Bug reports and corrections will be tracked on the public repository.",
"distribution": [
{
"@type": "cr:FileObject",
"@id": "tensorbench.json",
"name": "tensorbench.json",
"description": "JSON array of 199 task records.",
"contentUrl": "https://huggingface.co/datasets/tensorbench/tensorbench-1.0/resolve/main/tensorbench.json",
"encodingFormat": "application/json",
"sha256": "41b9cf73e37f8458990584a21882575040c360d08935c135eedfc5f474d155f7"
},
{
"@type": "cr:FileObject",
"@id": "Dockerfile",
"name": "Dockerfile",
"description": "Eval image: python:3.11-slim plus Scorch's C++ build deps and a clone of the upstream `bench` branch.",
"contentUrl": "https://huggingface.co/datasets/tensorbench/tensorbench-1.0/resolve/main/Dockerfile",
"encodingFormat": "text/plain",
"sha256": "130e9738bd764e26f5b655d9e3b72fcfc4c7d9df9687cead2b7d53b930289873"
},
{
"@type": "cr:FileObject",
"@id": "run_tests.sh",
"name": "run_tests.sh",
"description": "Container CMD: rebuild the C++ extension and run pytest -v.",
"contentUrl": "https://huggingface.co/datasets/tensorbench/tensorbench-1.0/resolve/main/run_tests.sh",
"encodingFormat": "application/x-sh",
"sha256": "5bcab82be4065de1002995baea922492ba77063f52f4da191df2bf8d23096dc1"
}
],
"recordSet": [
{
"@type": "cr:RecordSet",
"@id": "tasks",
"name": "tasks",
"description": "One record per benchmark task.",
"field": [
{
"@type": "cr:Field",
"@id": "tasks/instance_id",
"name": "instance_id",
"description": "Unique task id, e.g. bobbyyyan__scorch-feature_kernel_fusion. The feature_/refactor_ suffix follows the original taxonomy; semantics for both is feature-addition.",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].instance_id"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/repo_id",
"name": "repo_id",
"description": "Always bobbyyyan__scorch — used by the grading registry to route to the project-specific strategy.",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].repo_id"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/repo_url",
"name": "repo_url",
"description": "Git clone URL for the upstream Scorch repository. Identical for every task: https://github.com/bobbyyyan/scorch.git",
"dataType": "sc:URL",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].repo_url"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/base_commit",
"name": "base_commit",
"description": "Commit SHA the task is anchored at. The harness git reset --hards to this before applying patches.",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].base_commit"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/language",
"name": "language",
"description": "Source language. Always python for this dataset.",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].language"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/setup_commands",
"name": "setup_commands",
"description": "Optional list of shell commands run inside the container before the test command. Empty for all current Scorch tasks (setup is baked into the Docker image); the field is present for harness compatibility with other consumers.",
"dataType": "sc:Text",
"repeated": true,
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].setup_commands"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/test_command",
"name": "test_command",
"description": "Shell command that runs the pytest suite (typically /testbed/run_tests.sh).",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].test_command"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/test_timeout",
"name": "test_timeout",
"description": "Test-runner timeout in seconds. 3000 for every current task.",
"dataType": "sc:Integer",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].test_timeout"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/refactor_type",
"name": "refactor_type",
"description": "Mostly empty (legacy field from the original taxonomy; only a handful of tasks have extract/consolidate populated).",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].refactor_type"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/description",
"name": "description",
"description": "Natural-language task prompt the model receives.",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].description"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/files",
"name": "files",
"description": "Optional list of upstream-codebase files relevant to the task, surfaced as a hint to the model. Empty for all current Scorch tasks; the field is present for harness compatibility with other consumers.",
"dataType": "sc:Text",
"repeated": true,
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].files"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/task_type",
"name": "task_type",
"description": "feature or refactor. Both are evaluated as feature-addition (success = pytest passes).",
"dataType": "sc:Text",
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].task_type"}
}
},
{
"@type": "cr:Field",
"@id": "tasks/categories",
"name": "categories",
"description": "Hierarchical category paths (e.g. `Runtime/Caching & dispatch`, `API/Linear Algebra/Matmul variants`, `Scheduler/Loop transformations/Tiling`). A task may carry multiple categories.",
"dataType": "sc:Text",
"repeated": true,
"source": {
"fileObject": {"@id": "tensorbench.json"},
"extract": {"jsonPath": "$[*].categories"}
}
}
]
}
]
}