File size: 7,584 Bytes
d10ec49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
---
license: apache-2.0
language:
  - en
  - ko
tags:
  - skill-retrieval
  - information-retrieval
  - agents
  - benchmark
  - retrieval
pretty_name: SkillRetBench
size_categories:
  - n<1K
---

# SkillRetBench

## Dataset summary

**SkillRetBench** is a benchmark for **query-to-skill retrieval**: given a natural-language user request, systems must rank or select the correct procedural skill document(s) from a fixed library. The corpus is derived from a real-world production-style agent skill library (**501** skills). The benchmark includes **1,250** queries with gold skill identifiers and optional distractors, spanning five evaluation settings (single skill, multi-skill composition, distractors, outdated/redundant, budget-constrained). Packaged **baseline experiment results** (BM25, dense, hybrid, LLM-style, and SADO-style runs) report standard IR metrics (Recall@k, nDCG@k, MRR, MAP) per setting.

SkillRetBench accompanies the **SEFO** framework (Self-Evolving Federated Orchestration) and targets research on skill-aware orchestration, large skill libraries, and retrieval under composition and noise.

## Dataset structure

The release consists of three JSON artifacts (also listed in `croissant.json`):

| File | Description |
|------|-------------|
| `skill_corpus.json` | Full skill index: metadata, corpus statistics, and the `skills` array (one object per skill). |
| `skillretbench_queries.json` | Benchmark definition: `meta` (counts, seed, generator) and `queries` array. |
| `baseline_results.json` | Aggregated baseline metrics: per-method, per-setting IR scores and a markdown summary table. |

### `skill_corpus.json`

- **`meta`**: `generated_at` (ISO 8601), `skills_root`, `output_path`, `skill_file_count`.
- **`statistics`**: `total_skills`, `skills_per_category`, description length stats, `token_count_distribution`, composition graph stats (`skills_with_composition_references`, `total_composition_edges`).
- **`skills`** (array, length 501): each element includes:
  - `skill_id` (string, canonical id)
  - `skill_name` (string)
  - `description` (string, short routing description)
  - `trigger_phrases` (string array)
  - `anti_triggers` (string array)
  - `korean_triggers` (string array; may be empty)
  - `category` (string)
  - `full_text` (string, full SKILL.md-derived body)
  - `token_count` (integer)
  - `composable_skills` (string array; referenced skill ids)
  - `parse_warnings` (string array)

### `skillretbench_queries.json`

- **`meta`**: `benchmark`, `generator`, `corpus_meta`, `total_queries` (1250), `counts_by_setting`, `rng_seed` (42).
- **`queries`** (array, length 1250): each element includes:
  - `query_id` (string, e.g. `SS-0001`)
  - `setting` (enum-like string): `single_skill` | `multi_skill_composition` | `distractor` | `outdated_redundant` | `budget_constrained`
  - `query` (string; English and/or Korean natural language)
  - `gold_skills` (string array; expected skill ids)
  - `distractor_skills` (string array)
  - `budget_tokens` (number or `null`)
  - `difficulty` (string, e.g. `easy`, `medium`)
  - `source` (string, e.g. `trigger_paraphrase`)

**Query counts by setting**

| Setting | Count |
|---------|------:|
| `single_skill` | 400 |
| `multi_skill_composition` | 200 |
| `distractor` | 300 |
| `outdated_redundant` | 150 |
| `budget_constrained` | 200 |

### `baseline_results.json`

- **`timestamp`**, **`meta`**: `corpus_skills`, `queries_total`, `queries_by_setting`, `dense_backend` (e.g. `jaccard_fallback`).
- **`baselines`**: nested object — top-level keys **`BM25`**, **`Dense`**, **`Hybrid`**, **`NaiveLLM`**, **`SADO`**. Under each method, keys match query `setting` names; each leaf holds metrics: `recall@1`, `recall@3`, `recall@5`, `recall@10`, `ndcg@1`, `ndcg@3`, `ndcg@5`, `ndcg@10`, `mrr`, `map` (floats).
- **`summary_table`**: pre-rendered markdown table (macro-style summary over settings).

> **Note:** As documented in the artifact, `NaiveLLM` and `SADO` entries may be **simulated** (no live API) for reproducible baselines; see project script `run_baselines.py`.

## Supported tasks

- **skill-retrieval** — Rank or classify skills given a query; evaluate with gold `gold_skills` and IR metrics.
- **information-retrieval** — Same retrieval formulation; comparable to passage/tool retrieval with long procedural documents.

## Languages

- **English** — Majority of skill text and many queries.
- **Korean** — Present in `korean_triggers`, many `query` strings, and mixed Korean/English skill descriptions.

## Dataset creation

### Sources

- Skills were built from **Cursor agent `SKILL.md` files** under a single repository’s `.cursor/skills` tree (501 files), parsed into structured records plus full markdown bodies.
- Queries were **programmatically generated** by `app.sefo.benchmark.query_generator` with fixed **`rng_seed`: 42** for reproducibility.

### Methodology (high level)

1. **Corpus build** — Parse frontmatter and body; extract triggers, categories, composition edges, and token counts.
2. **Query generation** — Instantiate templates and paraphrases aligned to skills and settings (single vs. multi-skill, distractors, outdated/redundant, token budget hints).
3. **Baselines** — Run packaged retrievers/rankers (lexical, dense with configured fallback, hybrid, simulated LLM/SADO-style) and aggregate metrics per setting.

## Considerations

### Biases

- **Corpus domain** — Skills reflect one organization’s automation stack (dev tooling, PM, trading, cloud, etc.); frequency and vocabulary are **not** uniformly representative of all industries or locales.
- **Authoring style** — Descriptions follow a consistent “when to use / do not use” pattern; models may exploit **format bias** rather than deep semantics.
- **Query generation** — Synthetic and template-driven queries may **over- or under-represent** realistic user phrasing compared to production logs.

### Limitations

- **Scale** — 501 skills is medium-scale vs. ecosystems with 10⁵+ tools; retrieval hardness may not transfer linearly.
- **Gold labels** — Defined by benchmark construction rules; edge cases and valid alternative skill sets may exist.
- **Static snapshot** — Skill text and ids are frozen for the benchmark revision tied to `meta.generated_at` in the JSON files.
- **Baseline simulation** — Some methods are noted as simulated in-code; compare only under the same harness and seeds.

## License

This dataset is released under the **Apache License 2.0**. See [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).

## Citation

If you use SkillRetBench or this corpus, please cite:

```bibtex
@misc{han2026sefo,
  title        = {{SEFO}: Self-Evolving Federated Orchestration with Trusted Skill Governance and Skill Retrieval Benchmark for Recursive Agentic Systems},
  author       = {Han, Hyojung and {ThakiCloud AI Research}},
  year         = {2026},
  howpublished = {\url{https://github.com/hyojunguy/ai-model-event-stock-analytics}},
  note         = {Includes SkillRetBench: 501-skill corpus and 1,250-query retrieval benchmark}
}
```

Adjust `howpublished` / venue fields when a formal publication DOI is available.

## Dataset repository layout (recommended)

After upload to the Hugging Face Hub, a typical layout is:

```
skill_corpus.json
skillretbench_queries.json
baseline_results.json
croissant.json
README.md
```

## Additional documentation

- Croissant (ML Commons) metadata: **`croissant.json`** in this folder.
- Implementation and regeneration: `backend/app/sefo/benchmark/` in the source repository.