# eval.yaml — registers ClawBench as a Benchmark dataset on the Hub. # See: https://huggingface.co/docs/hub/eval-results#benchmark-datasets # This file unlocks the native "🏆 Leaderboard" tab and "Official benchmark" # badge once the dataset is added to HF's benchmark allow-list. name: ClawBench description: > ClawBench is an open benchmark for AI web agents — the systems that drive a real browser to complete a user's task end-to-end. It scores agents on real, everyday online tasks (booking flights, ordering groceries, submitting job applications) across live websites. V1 ships 153 tasks across 144 websites (the original frontier-model leaderboard); V2 ships 130 newer tasks (expanded coverage). For each run we capture five layers of behavioral data — session replay, HTTP traffic, browser actions, agent reasoning, and the final intercepted request — plus human ground-truth, then score with an agentic evaluator that produces step-level traceable diagnostics. # Pending — needs PR to https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/src/eval.ts # to add `clawbench-eval` as a canonical framework before this validates. evaluation_framework: clawbench-eval tasks: - id: v1 config: default split: test - id: v2 config: default split: test