--- license: apache-2.0 language: - en - ko tags: - skill-retrieval - information-retrieval - agents - benchmark - retrieval pretty_name: SkillRetBench size_categories: - n<1K --- # SkillRetBench ## Dataset summary **SkillRetBench** is a benchmark for **query-to-skill retrieval**: given a natural-language user request, systems must rank or select the correct procedural skill document(s) from a fixed library. The corpus is derived from a real-world production-style agent skill library (**501** skills). The benchmark includes **1,250** queries with gold skill identifiers and optional distractors, spanning five evaluation settings (single skill, multi-skill composition, distractors, outdated/redundant, budget-constrained). Packaged **baseline experiment results** (BM25, dense, hybrid, LLM-style, and SADO-style runs) report standard IR metrics (Recall@k, nDCG@k, MRR, MAP) per setting. SkillRetBench accompanies the **SEFO** framework (Self-Evolving Federated Orchestration) and targets research on skill-aware orchestration, large skill libraries, and retrieval under composition and noise. ## Dataset structure The release consists of three JSON artifacts (also listed in `croissant.json`): | File | Description | |------|-------------| | `skill_corpus.json` | Full skill index: metadata, corpus statistics, and the `skills` array (one object per skill). | | `skillretbench_queries.json` | Benchmark definition: `meta` (counts, seed, generator) and `queries` array. | | `baseline_results.json` | Aggregated baseline metrics: per-method, per-setting IR scores and a markdown summary table. | ### `skill_corpus.json` - **`meta`**: `generated_at` (ISO 8601), `skills_root`, `output_path`, `skill_file_count`. - **`statistics`**: `total_skills`, `skills_per_category`, description length stats, `token_count_distribution`, composition graph stats (`skills_with_composition_references`, `total_composition_edges`). - **`skills`** (array, length 501): each element includes: - `skill_id` (string, canonical id) - `skill_name` (string) - `description` (string, short routing description) - `trigger_phrases` (string array) - `anti_triggers` (string array) - `korean_triggers` (string array; may be empty) - `category` (string) - `full_text` (string, full SKILL.md-derived body) - `token_count` (integer) - `composable_skills` (string array; referenced skill ids) - `parse_warnings` (string array) ### `skillretbench_queries.json` - **`meta`**: `benchmark`, `generator`, `corpus_meta`, `total_queries` (1250), `counts_by_setting`, `rng_seed` (42). - **`queries`** (array, length 1250): each element includes: - `query_id` (string, e.g. `SS-0001`) - `setting` (enum-like string): `single_skill` | `multi_skill_composition` | `distractor` | `outdated_redundant` | `budget_constrained` - `query` (string; English and/or Korean natural language) - `gold_skills` (string array; expected skill ids) - `distractor_skills` (string array) - `budget_tokens` (number or `null`) - `difficulty` (string, e.g. `easy`, `medium`) - `source` (string, e.g. `trigger_paraphrase`) **Query counts by setting** | Setting | Count | |---------|------:| | `single_skill` | 400 | | `multi_skill_composition` | 200 | | `distractor` | 300 | | `outdated_redundant` | 150 | | `budget_constrained` | 200 | ### `baseline_results.json` - **`timestamp`**, **`meta`**: `corpus_skills`, `queries_total`, `queries_by_setting`, `dense_backend` (e.g. `jaccard_fallback`). - **`baselines`**: nested object — top-level keys **`BM25`**, **`Dense`**, **`Hybrid`**, **`NaiveLLM`**, **`SADO`**. Under each method, keys match query `setting` names; each leaf holds metrics: `recall@1`, `recall@3`, `recall@5`, `recall@10`, `ndcg@1`, `ndcg@3`, `ndcg@5`, `ndcg@10`, `mrr`, `map` (floats). - **`summary_table`**: pre-rendered markdown table (macro-style summary over settings). > **Note:** As documented in the artifact, `NaiveLLM` and `SADO` entries may be **simulated** (no live API) for reproducible baselines; see project script `run_baselines.py`. ## Supported tasks - **skill-retrieval** — Rank or classify skills given a query; evaluate with gold `gold_skills` and IR metrics. - **information-retrieval** — Same retrieval formulation; comparable to passage/tool retrieval with long procedural documents. ## Languages - **English** — Majority of skill text and many queries. - **Korean** — Present in `korean_triggers`, many `query` strings, and mixed Korean/English skill descriptions. ## Dataset creation ### Sources - Skills were built from **Cursor agent `SKILL.md` files** under a single repository’s `.cursor/skills` tree (501 files), parsed into structured records plus full markdown bodies. - Queries were **programmatically generated** by `app.sefo.benchmark.query_generator` with fixed **`rng_seed`: 42** for reproducibility. ### Methodology (high level) 1. **Corpus build** — Parse frontmatter and body; extract triggers, categories, composition edges, and token counts. 2. **Query generation** — Instantiate templates and paraphrases aligned to skills and settings (single vs. multi-skill, distractors, outdated/redundant, token budget hints). 3. **Baselines** — Run packaged retrievers/rankers (lexical, dense with configured fallback, hybrid, simulated LLM/SADO-style) and aggregate metrics per setting. ## Considerations ### Biases - **Corpus domain** — Skills reflect one organization’s automation stack (dev tooling, PM, trading, cloud, etc.); frequency and vocabulary are **not** uniformly representative of all industries or locales. - **Authoring style** — Descriptions follow a consistent “when to use / do not use” pattern; models may exploit **format bias** rather than deep semantics. - **Query generation** — Synthetic and template-driven queries may **over- or under-represent** realistic user phrasing compared to production logs. ### Limitations - **Scale** — 501 skills is medium-scale vs. ecosystems with 10⁵+ tools; retrieval hardness may not transfer linearly. - **Gold labels** — Defined by benchmark construction rules; edge cases and valid alternative skill sets may exist. - **Static snapshot** — Skill text and ids are frozen for the benchmark revision tied to `meta.generated_at` in the JSON files. - **Baseline simulation** — Some methods are noted as simulated in-code; compare only under the same harness and seeds. ## License This dataset is released under the **Apache License 2.0**. See [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0). ## Citation If you use SkillRetBench or this corpus, please cite: ```bibtex @misc{han2026sefo, title = {{SEFO}: Self-Evolving Federated Orchestration with Trusted Skill Governance and Skill Retrieval Benchmark for Recursive Agentic Systems}, author = {Han, Hyojung and {ThakiCloud AI Research}}, year = {2026}, howpublished = {\url{https://github.com/hyojunguy/ai-model-event-stock-analytics}}, note = {Includes SkillRetBench: 501-skill corpus and 1,250-query retrieval benchmark} } ``` Adjust `howpublished` / venue fields when a formal publication DOI is available. ## Dataset repository layout (recommended) After upload to the Hugging Face Hub, a typical layout is: ``` skill_corpus.json skillretbench_queries.json baseline_results.json croissant.json README.md ``` ## Additional documentation - Croissant (ML Commons) metadata: **`croissant.json`** in this folder. - Implementation and regeneration: `backend/app/sefo/benchmark/` in the source repository.