codesense / croissant_metadata /loop_iteration_dataset_python_croissant.json
codesense-bench's picture
Added metadata files
201c16d verified
{
"@context": {
"ml": "http://mlcommons.org/croissant/",
"schema": "https://schema.org/",
"sc": "https://schema.org/",
"codesense": "https://codesense-bench.github.io/schema/"
},
"@type": "schema:Dataset",
"schema:name": "CodeSense-Bench-loop_iteration_dataset_python",
"schema:description": "Dataset for loop iteration values of LLMs",
"schema:license": "CC-BY-4.0",
"schema:url": "https://huggingface.co/datasets/codesense-bench/codesense/blob/main/Task%20Specific%20Datasets/loop_iteration_dataset_python.jsonl",
"schema:creator": {
"@type": "schema:Organization",
"schema:name": "CodeSense Research Team"
},
"schema:distribution": [
{
"@type": "schema:FileObject",
"schema:contentUrl": "https://huggingface.co/datasets/codesense-bench/codesense/resolve/main/Task%20Specific%20Datasets/loop_iteration_dataset_python.jsonl",
"schema:encodingFormat": "application/jsonl",
"schema:name": "loop_iteration_dataset_python",
"schema:sha256": "12961156c536e2e21c7e2d42c06814bafe5e2fa652a971d0700f308c2f830bd6"
}
],
"recordSet": [
{
"name": "loop_iteration_dataset_python",
"description": "A collection of Python code examples designed to evaluate a model’s ability to reason about loop iteration.",
"field": [
{
"name": "idx",
"dataType": "sc:Integer",
"description": "Unique identifier for each record in the dataset."
},
{
"name": "scratchpad_format",
"dataType": "sc:Text",
"description": "code in scratchpad format, the variable states defined in comments."
},
{
"name": "loop_code",
"dataType": "sc:Text",
"description": "code used for LLM prompting."
},
{
"name": "question",
"dataType": "sc:Text",
"description": "Query regarding a specific loop line for number of iterations."
},
{
"name": "answer",
"dataType": "sc:Text",
"description": "Ground-Truth, contains the exact value of the number of loop iterations."
}
]
}
]
}