AgentDisruptBench / DATASHEET.md
kavirubc's picture
Upload folder using huggingface_hub
89f1d71 verified

Datasheet for AgentDisruptBench

Following the Datasheets for Datasets framework (Gebru et al., 2021).


Motivation

For what purpose was the dataset created?

AgentDisruptBench was created to study evaluation as a scientific object in the context of AI agent resilience under runtime tool-call disruptions. Existing benchmarks measure whether agents can use tools correctly, but assume tools behave perfectly — an unrealistic assumption in production environments. This project provides an evaluation methodology, consisting of structured tasks paired with a Disruption Engine, to rigorously measure how we assess agent reliability, recovery strategies, and graceful degradation using the $R(k, \epsilon, \lambda)$ surface.

Who created the dataset and on behalf of which entity?

AgentDisruptBench Contributors, as part of academic research into AI agent robustness.

Who funded the creation of the dataset?

[To be specified]


Composition

What do the instances that comprise the dataset represent?

Each instance is a benchmark task defined in YAML with:

  • A natural language task description (the prompt given to the agent)
  • Required tools (from a set of 30 deterministic mock tools)
  • Ground truth: expected outcome, required tool calls, evaluation rubric with weighted criteria
  • Metadata: domain, difficulty (1-5), task type

How many instances are there in total?

100 tasks organized as follows:

Domain Standard Adversarial Impossible Handover Total
Retail 20 2 2 1 25
Travel 20 2 2 1 25
Finance 20 2 2 1 25
DevOps 20 2 2 1 25
Total 80 8 8 4 100

Does the dataset contain all possible instances or is it a sample?

The dataset is a curated collection of synthetic tasks. It is not a sample from a larger population. Task difficulty levels are distributed as D1=8, D2=16, D3=24, D4=20, D5=12 tasks across standard tasks, following a roughly bell-shaped distribution centered on medium difficulty.

What data does each instance consist of?

Each task instance contains:

  • task_id: Unique identifier (e.g., retail_001)
  • title: Short human-readable title
  • description: Full prompt given to the agent (natural language)
  • domain: One of {retail, travel, finance, devops}
  • difficulty: Integer 1-5
  • task_type: One of {standard, adversarial, impossible}
  • required_tools: List of tool names needed
  • expected_tool_call_depth: Expected number of tool calls under clean conditions
  • ground_truth: Structured evaluation criteria including:
    • expected_outcome: Description of success
    • required_tool_calls: Tools that must be called
    • forbidden_tool_calls: Tools that must NOT be called
    • correct_final_answer: Exact expected answer (if applicable)
    • evaluation_rubric: Criterion → weight mapping (sums ≈ 1.0)
    • disruption_sensitive_tools: Tools where failure is most impactful
    • recovery_actions: Expected recovery behaviours
    • trap_description: (adversarial only) the trap the agent should avoid
    • impossibility_reason: (impossible only) why no valid solution exists

Is any information missing from individual instances?

No. All fields are populated for every task instance.

Are relationships between individual instances made explicit?

Tasks within the same domain share tools and mock data (e.g., retail tasks all use the same product catalog and customer database via deterministic mock tools). Cross-domain relationships do not exist.

Are there recommended data splits?

No train/test split is provided because this is a benchmark, not a training dataset. All 100 tasks are intended for evaluation. However, researchers may partition by domain, difficulty, or task type for focused analysis.

Does the dataset contain data that might be considered confidential?

No. All data is synthetic. No real personal data, API keys, or proprietary information is included.

Does the dataset contain data that, if viewed directly, might be offensive, insulting, threatening, or might otherwise cause anxiety?

No. Tasks involve routine operations (e-commerce, travel booking, financial transactions, DevOps operations) with no offensive content.


Collection Process

How was the data associated with each instance acquired?

All tasks were authored manually by the benchmark contributors. They are synthetic, designed to cover a range of realistic tool-use scenarios across difficulty levels.

What mechanisms or procedures were used to collect the data?

Task design followed a structured process:

  1. Define 4 domains based on common LLM agent use cases
  2. Design 8 mock tools per domain (30 total)
  3. Author 20 standard tasks per domain with increasing difficulty
  4. Author adversarial tasks with designed trap patterns
  5. Author impossible tasks with clear impossibility reasons
  6. Author handover tasks requiring human escalation
  7. Validate all ground truth against mock tool outputs

Who was involved in the data collection process?

The benchmark authors.

Over what timeframe was the data collected?

March-April 2026.

Were any ethical review processes conducted?

[To be specified]


Uses

What are the intended uses of the dataset?

  1. Primary: Evaluate AI agent resilience under runtime tool-call disruptions using the provided disruption profiles (clean → hostile_environment)
  2. Secondary: Compare agent frameworks (LangChain, OpenAI, AutoGen, CrewAI) on reliability metrics
  3. Tertiary: Study recovery strategies and failure patterns in LLM agents

What are some tasks/uses that the dataset should not be used for?

  • Not for training: The benchmark should be used for evaluation only. Training on the tasks would compromise benchmark validity.
  • Not for real tool testing: Mock tools are deterministic simulations, not real API integrations.
  • Not for safety/security evaluation: The disruptions model reliability failures, not adversarial attacks.

Is there anything about the composition of the dataset or the way it was collected that might impact future uses?

  • Tasks are in English only
  • Mock tools simulate US-centric services (USD currency, US dates, etc.)
  • Ground truth rubrics use string-matching heuristics which may not capture all valid agent responses

Distribution

How will the dataset be distributed?

When will the dataset be distributed?

The dataset is publicly available as of the initial release.

Will the dataset be distributed under a copyright or other intellectual property (IP) license?

MIT License.

Have any third parties imposed IP-based or other restrictions on the data?

No.


Maintenance

Who is supporting/hosting/maintaining the dataset?

AgentDisruptBench Contributors via the GitHub repository.

How can the owner/curator/manager of the dataset be contacted?

Via GitHub Issues at the repository.

Will the dataset be updated?

Yes. Planned updates include:

  • Additional task variants for ε-robustness testing
  • New domains
  • Track B: Network-layer disruption injection via Envoy proxy

If the dataset relates to people, are there applicable limits on the retention of the data?

N/A — the dataset is entirely synthetic with no real personal data.

Will older versions of the dataset continue to be available?

Yes, via Git tags and versioned releases.

If others want to extend/augment/build on/contribute to the dataset, is there a mechanism for doing so?

Yes. See CONTRIBUTING.md for guidelines on contributing new tasks, domains, tools, and disruption types.