Spaces:
Running
Running
| # HF-Agent Eval | |
| Rubric-based evaluation pipeline implementing [Rubrics as Rewards](https://arxiv.org/abs/2507.17746) paper (RaR-Explicit formula). | |
| ## Components | |
| | Component | Purpose | Long Term Goal | | |
| |-----------|---------|----------------| | |
| | **`generate_rubrics.py`** | Generates instance-specific evaluation criteria (7-20 weighted rubrics) from QA pairs using LLM, following the RaR paper methodology | Improve rubric quality with few-shot examples, domain-specific templates, and iterative refinement | | |
| | **`rubric_eval.py`** | Scores responses using RaR-Explicit formula: checks each criterion independently via LLM judge, computes weighted normalized score | Support batch evaluation, caching, and alternative scoring formulas (RaR-Holistic) | | |
| | **`task.py`** | Defines Inspect AI task `hf-benchmark-with-rubrics` that wires dataset, solver, and rubric scorer into a single evaluation pipeline | Add more task variants for different benchmarks (code generation, tool use, multi-turn) | | |
| | **`solvers.py`** | Registry of solver implementations (`hf_agent`, `claude_code`, `claude_code+hf_mcp`) that can be swapped via CLI args | Expand solver library to benchmark more agents (OpenAI Codex, Gemini, open-source agents) | | |
| | **`hf_agent_connector.py`** | Lightweight bridge that spins up the hf-agent stack (tools, MCP, LiteLLM loop) and returns the final assistant response | Enable streaming, intermediate step logging, and cost tracking per evaluation | | |
| | **`leaderboard.py`** | Utilities to build records and append scores to a HuggingFace dataset for tracking performance over time | Add score breakdowns, visualizations, and automatic regression detection | | |
| | **`run_eval_with_leaderboard.py`** | CLI wrapper that runs `inspect eval`, parses scores from logs, and pushes results to the leaderboard dataset | Support scheduled CI runs, PR-gated benchmarks, and multi-dataset aggregation | | |
| | **`hf_io.py`** | Helper utilities for pushing DataFrames to HuggingFace Hub | Extend with dataset versioning and diff tracking | | |
| | **`models.py`** | Shared Pydantic models for evaluation data structures | Centralize all eval schemas for consistency across components | | |
| ## Pipeline | |
| ``` | |
| QA pairs → generate_rubrics.py → run `inspect-ai eval eval/task.py@hf-benchmark-with-rubrics` → scores | |
| ``` | |
| ### 1. Generate Rubrics (if not already generated) | |
| Creates instance-specific evaluation criteria from question + reference answer. | |
| ```bash | |
| python eval/generate_rubrics.py \ | |
| --infile qa_pairs.jsonl \ | |
| --outfile qa_rubrics.jsonl \ | |
| --model anthropic/claude-sonnet-4-5-20250929 \ | |
| --push-to-hub akseljoonas/hf-agent-benchmark@rubrics | |
| ``` | |
| **Input format:** | |
| ```json | |
| {"question": "...", "solution": "...", "thread": [...]} | |
| ``` | |
| **Output:** 7-20 weighted criteria per question (Essential: +5, Important: +3-4, Optional: +1-2, Pitfall: -1 to -2) | |
| ### 2. Response evaluation | |
| Files: | |
| - `eval/hf_agent_connector.py` contains a lightweight bridge that spins up | |
| the existing hf-agent stack in `agent/` (tools, MCP, LiteLLM loop) and returns the assistant reply. | |
| - `eval/solvers.py` keeps the solver implementations (e.g. `hf_agent`, | |
| `claude_code`). If additional solvers are needed, register them there and pass | |
| `-T solver_name=<name>` to swap them in without touching the task. | |
| - `eval/task.py` registers `hf-benchmark-with-rubrics`, which wires | |
| the dataset, solver, and rubric scorer into a single Inspect task and does the eval. | |
| ### Running the hf-agent (implemented in `agent/`) (args are optional) | |
| ```bash | |
| uv run inspect eval eval/task.py@hf-benchmark-with-rubrics \ | |
| -T dataset_name=akseljoonas/hf-agent-rubrics \ | |
| -T dataset_split=train \ | |
| -T limit=25 \ | |
| -T solver_name=hf_agent \ | |
| -T solver_kwargs='{"config_path":"agent/config_mcp_example.json","max_iterations":10}' \ | |
| --log-dir logs/inspect | |
| ``` | |
| Different benchmarks can be used by making/running a new task in `eval/task.py`. | |
| ### Running Claude Code headlessly | |
| The `claude_code` solver shell-outs to the `claude` CLI (`claude -p ... --output-format json`) | |
| so you can benchmark Claude Code without any interactive UI. Example: | |
| Claude Code command example (kwargs are optional): | |
| ```bash | |
| uv run inspect eval eval/task.py@hf-benchmark-with-rubrics \ | |
| -T solver_name=claude_code \ | |
| -T solver_kwargs='{"allowed_tools":"Bash,Read","output_format":"json"}' | |
| ``` | |
| ### Leaderboard | |
| Scores can be pushed to a Hugging Face dataset automatically by wrapping the run | |
| with `eval/run_eval_with_leaderboard.py` (it executes `inspect eval ...` under the hood | |
| and only appends results when the command succeeds): | |
| ```bash | |
| uv run python eval/run_eval_with_leaderboard.py \ | |
| --hf-dataset akseljoonas/hf-agent-leaderboard \ | |
| --hf-token $HF_TOKEN \ | |
| --solver-name hf_agent \ | |
| --solver-kwargs '{"config_path":"agent/config_mcp_example.json","max_iterations":10}' \ | |
| --dataset akseljoonas/hf-agent-rubrics@train \ | |
| --limit 25 | |
| ``` | |
| ## Scoring (implemented in `eval/rubric_eval.py`) | |
| The scoring is implemented in `eval/rubric_eval.py` and is based on the RaR-Explicit formula: `score = Σ(weight × satisfied) / Σ(positive_weights)`. | |
| The score is normalized to [0, 1] and clipped if pitfalls make it negative. | |