teambench / README.md
ybkim95's picture
Upload README.md with huggingface_hub
29c59b5 verified
---
dataset_info:
features:
- name: task_id
dtype: string
- name: title
dtype: string
- name: category
dtype: string
- name: difficulty
dtype: string
- name: has_generator
dtype: bool
- name: ablation_scores
dtype: string
- name: tni
dtype: float32
- name: classification
dtype: string
splits:
- name: train
num_examples: 153
- name: test
num_examples: 120
license: mit
task_categories:
- text-generation
tags:
- multi-agent
- benchmark
- software-engineering
- teamwork
- evaluation
- llm-evaluation
pretty_name: TeamBench
---
# TeamBench: A Multi-Agent Teamwork Benchmark
## Overview
TeamBench is a rigorous benchmark for evaluating whether LLM-based agent *teams* outperform single oracle agents on realistic software engineering tasks. Each task is executed under five ablation conditions (oracle, restricted, full team, team without planning, team without verification), enabling fine-grained measurement of when and why teamwork helps. The core metric is the **Teamwork Necessity Index (TNI)**, which quantifies how much a task requires coordinated multi-agent effort beyond what a single capable agent can achieve alone.
TeamBench uses OS-enforced role separation — Planner, Executor, and Verifier agents operate in isolated sandboxes with distinct tool allow-lists — ensuring that role boundaries are structurally enforced rather than prompt-based. Tasks span 18 categories including security, data engineering, adversarial specification traps, distributed systems, and long-horizon planning, with 153 tasks totalling 459 parameterized instances (seeds 0–2).
---
## Task Categories
| Category | Tasks | Description |
|---|---|---|
| Software Eng. | 22 | Hidden specs, backward compatibility, refactoring |
| Security | 17 | Vulnerability patching, cryptographic correctness, audit triage |
| Operations | 16 | Incident root-cause, container debugging, monitoring |
| Data Engineering | 12 | Schema drift, ETL repair, query optimization |
| Testing | 11 | Spec-to-tests, mutation resistance, property-based testing |
| Incident Response | 11 | Cascade failure, memory leak, rollback planning |
| Information Retrieval | 8 | Evidence QA, misinformation traps, multi-source retrieval |
| Policy | 8 | Access control, data retention, license compliance |
| Distributed Systems | 7 | Race conditions, Raft consensus, idempotency |
| Adversarial | 7 | Spec conflicts, false bug reports, security theater |
| Code Review | 6 | API review, style enforcement, test coverage |
| Multi-language | 6 | Go concurrency, JavaScript XSS, polyglot debugging |
| Long-Horizon | 6 | Multi-step migrations, staged deployments, audit trails |
| Pipeline | 6 | ETL fix, API gateway, message queues |
| Cross-System Integration | 5 | API contract mismatches, schema evolution, auth federation |
| Specification | 3 | Feature implementation from RFC, config schema design |
| Integration | 1 | Pipeline repair, API versioning |
| Negotiation | 1 | Trade-off configuration under competing constraints |
**Difficulty breakdown:** 104 hard, 26 medium, 16 expert, 7 easy (78% hard or expert).
---
## Ablation Conditions
Each task is evaluated under five conditions:
| Condition | Description |
|---|---|
| `oracle` | Single powerful agent with full tool access and no role restrictions |
| `restricted` | Single agent with executor-only tool access (no planning/verification tools) |
| `team` | Full three-role team: Planner → Executor → Verifier |
| `team_no_plan` | Two-role team: Executor → Verifier (planning phase skipped) |
| `team_no_verify` | Two-role team: Planner → Executor (verification phase skipped) |
Scores are in [0, 1] and represent the fraction of grader checks passed.
---
## TNI Metric
The **Teamwork Necessity Index (TNI)** measures how much a task *requires* teamwork:
```
TNI = team_uplift / necessity_gap
= (team - oracle) / (1 - restricted)
```
- `TNI > 0`: team outperforms oracle relative to the task's difficulty ceiling
- `TNI = 1.0`: team achieves the maximum possible improvement over oracle
- `TNI > 1.0`: team substantially exceeds oracle (rare; indicates strong synergy)
- `TNI < 0`: team underperforms oracle (teamwork overhead hurts)
**Classification thresholds:**
- `HIGH-TNI`: TNI ≥ 0.5 and team > oracle
- `TEAM-HELPS`: team > oracle but TNI < 0.5
- `NEUTRAL`: |team - oracle| ≤ 0.05
- `TEAM-HURTS`: team < oracle
Of the 153 tasks: TEAM-HELPS 53, NEUTRAL 51, TEAM-HURTS 28, HIGH-TNI 16.
---
## Usage
```python
import json
with open("teambench_dataset.json") as f:
tasks = json.load(f)
# Filter to tasks where teamwork helps
team_helps = [t for t in tasks if t.get("classification") in ("TEAM-HELPS", "HIGH-TNI")]
print(f"Tasks where team > oracle: {len(team_helps)}")
# Get hard tasks with ablation scores
hard_with_scores = [
t for t in tasks
if t["difficulty"] in ("hard", "expert") and "ablation_scores" in t
]
print(f"Hard/expert tasks with ablation data: {len(hard_with_scores)}")
# Compute average team uplift
uplifts = [
t["ablation_scores"]["team"] - t["ablation_scores"]["oracle"]
for t in tasks
if "ablation_scores" in t
]
print(f"Mean team uplift: {sum(uplifts)/len(uplifts):.3f}")
```
### Loading via HuggingFace datasets
```python
from datasets import load_dataset
ds = load_dataset("ybkim95/teambench", split="train") # all 153 tasks
# ds = load_dataset("ybkim95/teambench", split="test") # 120 hard/expert tasks only
print(ds[0])
```
---
## Benchmark Results
Cross-model evaluation across Gemini and OpenAI model families shows that:
- Team outperforms oracle on **43.9%** of tasks (68/155 with full ablation data)
- Average TNI: **0.744** across tasks with a measurable necessity gap
- Weaker models benefit *more* from teamwork (larger relative uplift)
- `team_no_verify` is often the strongest condition — verifier overhead can hurt on average
- The **Expertise-Asymmetry (EA)** variant (5 tasks) shows TNI > 1.0 with capable models, meaning specialized role knowledge pushes teams beyond what any single oracle achieves
Full results and leaderboard: [GitHub](https://github.com/ybkim95/TeamBench)
---
## Citation
```bibtex
@dataset{kim2026teambench,
author = {Yubin Kim},
title = {TeamBench: A Multi-Agent Teamwork Benchmark for LLM Evaluation},
year = {2026},
url = {https://huggingface.co/datasets/ybkim95/teambench},
note = {153 tasks across 18 categories with OS-enforced role separation}
}
```
---
## License
MIT License. See [LICENSE](https://github.com/ybkim95/TeamBench/blob/main/LICENSE) for details.