| --- |
| license: mit |
| task_categories: |
| - text-generation |
| - question-answering |
| tags: |
| - mcp |
| - ai-agents |
| - benchmark |
| - evaluation |
| - tool-use |
| - function-calling |
| size_categories: |
| - n<1K |
| --- |
| |
| # Agent Evaluation Benchmark |
|
|
| A benchmark dataset for evaluating AI agent tool-use capabilities across 55+ test cases spanning 14 categories. |
|
|
| ## Overview |
|
|
| This benchmark tests whether AI agents can correctly select and use the right MCP tools for real-world tasks. It covers data retrieval, blockchain queries, security analysis, academic research, and more. |
|
|
| ## Categories |
|
|
| | Category | Test Cases | Description | |
| |----------|-----------|-------------| |
| | Weather | 5 | Forecasts, UV index, climate history | |
| | Blockchain | 7 | Token prices, wallet analysis, DeFi, whale tracking | |
| | Security | 5 | CVE search, vulnerability analysis, CVSS scores | |
| | Academic | 4 | Paper search, citations, author lookup | |
| | Company | 3 | EU company registry search | |
| | Agriculture | 3 | Crop data, food prices, yield comparison | |
| | Space | 4 | NASA APOD, asteroids, Mars rover, ISS | |
| | Aviation | 3 | Flight tracking, airport info | |
| | Medical | 3 | WHO data, disease outbreaks, health stats | |
| | Political | 2 | Campaign finance, FEC data | |
| | Supply Chain | 2 | UN trade data, import/export stats | |
| | LLM Benchmark | 3 | Model comparison, pricing, benchmarks | |
| | Energy | 3 | CO2 intensity, energy mix, electricity prices | |
| | Legal | 3 | Court decisions, case search | |
| | Agent Infrastructure | 3 | Directory, memory, workflows | |
| | Compliance | 2 | PII detection, GDPR checks | |
|
|
| ## Schema |
|
|
| - **task_description**: Natural language description of the task |
| - **expected_tool**: The MCP tool that should be selected |
| - **difficulty**: easy, medium, or hard |
| - **category**: Task category |
| - **test_input**: Example input parameters |
| - **expected_output_contains**: Key string that should appear in the output |
| |
| ## Difficulty Distribution |
| |
| - Easy: 25 tasks (basic single-tool queries) |
| - Medium: 25 tasks (parameter selection, filtering, comparison) |
| - Hard: 5 tasks (multi-step reasoning, complex analysis) |
| |
| ## Usage |
| |
| Use this dataset to evaluate: |
| 1. **Tool Selection Accuracy**: Does the agent pick the right tool? |
| 2. **Parameter Extraction**: Does the agent correctly parse inputs? |
| 3. **Output Validation**: Does the response contain expected information? |
| |
| ```python |
| from datasets import load_dataset |
| ds = load_dataset("aiagentkarl/agent-evaluation-benchmark") |
| ``` |
| |
| ## License |
| |
| MIT |
| |
| ## Author |
| |
| [AiAgentKarl](https://github.com/AiAgentKarl) |
| |