{ "@context": { "@vocab": "https://schema.org/", "dct": "http://purl.org/dc/terms/", "conformsTo": "dct:conformsTo" }, "@type": "Dataset", "name": "WebDS", "description": "WebDS is a reproducible benchmark dataset for evaluating LLM agents on complex web-based data science tasks. It comprises 870 tasks across 29 containerized websites and 10 domains. Each task simulates realistic end-to-end workflows including data acquisition, tool use, analysis, reasoning, and reporting. Tasks span multiple modalities, structured and unstructured data, and multihop reasoning, and are annotated with fine-grained attributes and difficulty levels.", "conformsTo": "http://mlcommons.org/croissant/1.0", "license": "https://creativecommons.org/licenses/by/4.0/", "keywords": [ "web-based data science", "benchmark", "LLM agents", "multi-hop reasoning", "tool use", "agent evaluation" ], "creator": { "@type": "Organization", "name": "Stanford University, UC Berkeley, SUTD, USC" }, "includedInDataCatalog": { "@type": "DataCatalog", "name": "MLCommons Croissant Benchmarks" }, "spatialCoverage": { "@type": "Place", "name": "Web-based / Global" }, "temporalCoverage": "2025", "datePublished": "2025-05-15", "version": "1.0", "isAccessibleForFree": true }