Repair dataset card configs after MSE merge

#6
Files changed (1) hide show
  1. README.md +75 -18
README.md CHANGED
@@ -12,7 +12,7 @@ tags:
12
  - sglang
13
  - agentic-workloads
14
  size_categories:
15
- - 1K<n<10K
16
  pretty_name: AgentPerfBench
17
  version: "1.0"
18
  configs:
@@ -20,10 +20,14 @@ configs:
20
  data_files:
21
  - split: summary
22
  path: trace_replay/summary.parquet
23
- - config_name: distributional
24
  data_files:
25
  - split: summary
26
- path: distributional/summary.parquet
 
 
 
 
27
  - config_name: mse_validation
28
  data_files:
29
  - split: summary
@@ -99,7 +103,7 @@ dataset_info:
99
  - name: summary
100
  num_examples: 3147
101
  num_bytes: 694254
102
- - config_name: distributional
103
  features:
104
  - name: run_id
105
  dtype: string
@@ -169,6 +173,55 @@ dataset_info:
169
  - name: summary
170
  num_examples: 245
171
  num_bytes: 70836
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
172
  - config_name: mse_validation
173
  features:
174
  - name: validation_id
@@ -279,11 +332,11 @@ dataset_info:
279
 
280
  # AgentPerfBench
281
 
282
- LLM inference benchmark: 3,392 main sweep runs measuring TTFT, TPOT, ITL, and throughput across 9 models, up to 14 GPU configurations, and 2 serving engines (vLLM 0.19.0, SGLang 0.5.9). All models served in BF16 except gpt-oss, which uses mxfp4 for projection weights. The dataset also includes 28 curated MSE validation rows for the distributional synthetic replay generator.
283
 
284
  ## Dataset configurations
285
 
286
- The dataset provides three configurations. *trace_replay* replays exact input/output sequences from recorded agent sessions. *distributional* samples from statistical distributions fitted to those same workloads, trading fidelity for faster sweeps across the hardware matrix. *mse_validation* contains paired synthetic-vs-real validation runs and ablations for the final APC-aware synthetic generator.
287
 
288
  ### trace_replay (3,147 rows)
289
 
@@ -291,12 +344,16 @@ Replays exact ISL/OSL sequences from recorded agent sessions (SWE-Bench, Termina
291
 
292
  17 profiles: `chat-medium`, `chat-multiturn-long`, `chat-multiturn-medium`, `chat-multiturn-short`, `chat-short`, `chat-singleturn`, `coding-singleturn`, `decode-heavy`, `osworld-multiturn-long`, `osworld-multiturn-medium`, `osworld-multiturn-short`, `prefill-heavy`, `random-1k`, `swebench-multiturn-medium`, `swebench-multiturn-short`, `terminalbench-multiturn-medium`, `terminalbench-multiturn-short`
293
 
294
- ### distributional (245 rows)
295
 
296
  Samples ISL/OSL from lognormal distributions fitted to real workload statistics. Covers 42 unique (model, hardware, engine) combinations across 6 profiles and 7 concurrency levels (3.0% matrix fill). `gpt-oss-120b`, `3090x8`, and `A100-40GBx8` are excluded from this configuration.
297
 
298
  6 profiles: `chat-multiturn`, `chat-singleturn`, `coding-singleturn`, `osworld-multiturn`, `swebench-multiturn`, `terminalbench-multiturn`
299
 
 
 
 
 
300
  ### mse_validation (28 rows)
301
 
302
  Curated H100 / Llama-3.1-8B / vLLM validation artifacts for the distributional synthetic replay generator. The main rows keep paired MSE distributional replay and real trace replay runs with success rate at least 75%; supplementary rows preserve no-replacement and high-concurrency debug runs.
@@ -317,18 +374,18 @@ See `mse_validation/README.md`, `mse_validation/manifest.csv`, and `mse_validati
317
  The benchmark harness capped actual concurrent connections at the session pool size. Rows where declared concurrency exceeded the pool were excluded:
318
 
319
  - trace_replay: concurrency > 100 removed (session pool was 100). Remaining values: {1, 5, 10, 20, 40, 80}.
320
- - distributional (pre-fix): concurrency > 10 removed (session pool was 10). Post-fix data has no cap. Remaining values: {1, 5, 10, 40, 80, 200, 320}.
321
 
322
  | Config | Rows |
323
  |--------|------|
324
  | trace_replay | 3,147 |
325
- | distributional | 245 |
326
  | mse_validation | 28 |
327
- | **Total** | **3,420** |
328
 
329
  ### Failed requests
330
 
331
- Some runs produce request failures, typically at high concurrency where the engine hits memory or timeout limits. 30.8% of trace_replay rows and 42% of distributional rows have `failed_requests > 0`. Summary metrics (TTFT, TPOT, throughput) are computed from successful requests only.
332
 
333
  ## Coverage
334
 
@@ -370,7 +427,7 @@ Model names in this table match the `model` column in the parquet files.
370
 
371
  ## Schema
372
 
373
- Each row in `summary.parquet` (both configs):
374
 
375
  | Column | Type | Description |
376
  |--------|------|-------------|
@@ -401,13 +458,13 @@ Each row in `summary.parquet` (both configs):
401
  from datasets import load_dataset
402
 
403
  ds = load_dataset("agent-perf-bench/AgentPerfBench", "trace_replay")
404
- # or "distributional" / "mse_validation"
405
  ```
406
 
407
  ## Benchmark methodology
408
 
409
  - Closed-loop concurrency with semaphore control.
410
- - Concurrency levels: {1, 5, 10, 20, 40, 80} (trace_replay), {1, 5, 10, 40, 80, 200, 320} (distributional).
411
  - 3-request warmup before each configuration.
412
  - Metrics: TTFT, TPOT, ITL, E2EL, request throughput, token throughput.
413
  - Summary statistics: mean, median, p90, p99.
@@ -417,7 +474,7 @@ ds = load_dataset("agent-perf-bench/AgentPerfBench", "trace_replay")
417
  ## Future releases
418
 
419
  - Full per-request and multi-turn granularity data for the main sweep (pending raw JSON availability from collection infrastructure). Curated raw JSONs are included for `mse_validation`.
420
- - Per-kernel CUDA roofline profiles (PyTorch profiler, 2-layer forward passes, batch sizes 1/4/8/32/64).
421
  - This is version 1.0. Updates will be tagged with semantic versions.
422
 
423
  ## Intended uses
@@ -429,10 +486,10 @@ ds = load_dataset("agent-perf-bench/AgentPerfBench", "trace_replay")
429
  ## Limitations
430
 
431
  - Results are specific to tested hardware and software versions (vLLM 0.19.0, SGLang 0.5.9, PyTorch 2.10.0, CUDA 12.8).
432
- - Distributional profiles approximate but do not replicate production traffic patterns.
433
- - No consumer GPUs beyond RTX 3090; no non-NVIDIA accelerators.
434
  - Closed-loop concurrency only; no open-loop (Poisson) arrivals.
435
- - The model-hardware-concurrency matrix is sparse (12.2% fill for trace_replay, 3.0% for distributional). Not all model-hardware combinations are represented.
436
  - No model quality metrics. This is a systems benchmark.
437
 
438
  ## Ethical considerations
 
12
  - sglang
13
  - agentic-workloads
14
  size_categories:
15
+ - 100K<n<1M
16
  pretty_name: AgentPerfBench
17
  version: "1.0"
18
  configs:
 
20
  data_files:
21
  - split: summary
22
  path: trace_replay/summary.parquet
23
+ - config_name: synthetic_distributional
24
  data_files:
25
  - split: summary
26
+ path: synthetic_distributional/summary.parquet
27
+ - config_name: kernels_labeled
28
+ data_files:
29
+ - split: train
30
+ path: kernel_profiles/kernels_labeled.parquet
31
  - config_name: mse_validation
32
  data_files:
33
  - split: summary
 
103
  - name: summary
104
  num_examples: 3147
105
  num_bytes: 694254
106
+ - config_name: synthetic_distributional
107
  features:
108
  - name: run_id
109
  dtype: string
 
173
  - name: summary
174
  num_examples: 245
175
  num_bytes: 70836
176
+ - config_name: kernels_labeled
177
+ features:
178
+ - name: source
179
+ dtype: string
180
+ - name: gpu
181
+ dtype: string
182
+ - name: model
183
+ dtype: string
184
+ - name: kernel_family
185
+ dtype: string
186
+ - name: kernel_name
187
+ dtype: string
188
+ - name: dtype
189
+ dtype: string
190
+ - name: held_out
191
+ dtype: bool
192
+ - name: M
193
+ dtype: float64
194
+ - name: N
195
+ dtype: float64
196
+ - name: K
197
+ dtype: float64
198
+ - name: bs
199
+ dtype: float64
200
+ - name: seq
201
+ dtype: float64
202
+ - name: n_heads
203
+ dtype: float64
204
+ - name: head_dim
205
+ dtype: float64
206
+ - name: kv_heads
207
+ dtype: float64
208
+ - name: numel
209
+ dtype: float64
210
+ - name: op_type
211
+ dtype: string
212
+ - name: gpu_time_duration_ms
213
+ dtype: float64
214
+ - name: launch_block_size
215
+ dtype: float64
216
+ - name: launch_grid_size
217
+ dtype: float64
218
+ - name: dram_bytes_sum
219
+ dtype: float64
220
+ - name: launch_registers_per_thread
221
+ dtype: float64
222
+ splits:
223
+ - name: train
224
+ num_examples: 148077
225
  - config_name: mse_validation
226
  features:
227
  - name: validation_id
 
332
 
333
  # AgentPerfBench
334
 
335
+ LLM inference benchmark: 3,392 main sweep rows measuring TTFT, TPOT, ITL, and throughput across 9 models, up to 14 GPU configurations, and 2 serving engines (vLLM 0.19.0, SGLang 0.5.9). All models served in BF16 except gpt-oss, which uses mxfp4 for projection weights. The dataset also includes 148,077 per-kernel NCU profiles and 28 curated MSE validation rows for the distributional synthetic replay generator.
336
 
337
  ## Dataset configurations
338
 
339
+ The dataset provides four configurations. `trace_replay` replays exact input/output sequences from recorded agent sessions. `synthetic_distributional` samples from statistical distributions fitted to those same workloads, trading fidelity for faster sweeps across the hardware matrix. `kernels_labeled` contains per-kernel Nsight Compute labels. `mse_validation` contains paired synthetic-vs-real validation runs and ablations for the final APC-aware synthetic generator.
340
 
341
  ### trace_replay (3,147 rows)
342
 
 
344
 
345
  17 profiles: `chat-medium`, `chat-multiturn-long`, `chat-multiturn-medium`, `chat-multiturn-short`, `chat-short`, `chat-singleturn`, `coding-singleturn`, `decode-heavy`, `osworld-multiturn-long`, `osworld-multiturn-medium`, `osworld-multiturn-short`, `prefill-heavy`, `random-1k`, `swebench-multiturn-medium`, `swebench-multiturn-short`, `terminalbench-multiturn-medium`, `terminalbench-multiturn-short`
346
 
347
+ ### synthetic_distributional (245 rows)
348
 
349
  Samples ISL/OSL from lognormal distributions fitted to real workload statistics. Covers 42 unique (model, hardware, engine) combinations across 6 profiles and 7 concurrency levels (3.0% matrix fill). `gpt-oss-120b`, `3090x8`, and `A100-40GBx8` are excluded from this configuration.
350
 
351
  6 profiles: `chat-multiturn`, `chat-singleturn`, `coding-singleturn`, `osworld-multiturn`, `swebench-multiturn`, `terminalbench-multiturn`
352
 
353
+ ### kernels_labeled (148,077 rows)
354
+
355
+ Per-kernel Nsight Compute (ncu) profiles across 4 GPUs (A100, H100, RTX 3090, RTX 2080 Ti) and 13 model/sweep sources.
356
+
357
  ### mse_validation (28 rows)
358
 
359
  Curated H100 / Llama-3.1-8B / vLLM validation artifacts for the distributional synthetic replay generator. The main rows keep paired MSE distributional replay and real trace replay runs with success rate at least 75%; supplementary rows preserve no-replacement and high-concurrency debug runs.
 
374
  The benchmark harness capped actual concurrent connections at the session pool size. Rows where declared concurrency exceeded the pool were excluded:
375
 
376
  - trace_replay: concurrency > 100 removed (session pool was 100). Remaining values: {1, 5, 10, 20, 40, 80}.
377
+ - synthetic_distributional (pre-fix): concurrency > 10 removed (session pool was 10). Post-fix data has no cap. Remaining values: {1, 5, 10, 40, 80, 200, 320}.
378
 
379
  | Config | Rows |
380
  |--------|------|
381
  | trace_replay | 3,147 |
382
+ | synthetic_distributional | 245 |
383
  | mse_validation | 28 |
384
+ | kernels_labeled | 148,077 |
385
 
386
  ### Failed requests
387
 
388
+ Some runs produce request failures, typically at high concurrency where the engine hits memory or timeout limits. 30.8% of trace_replay rows and 42% of synthetic_distributional rows have `failed_requests > 0`. Summary metrics (TTFT, TPOT, throughput) are computed from successful requests only.
389
 
390
  ## Coverage
391
 
 
427
 
428
  ## Schema
429
 
430
+ Each row in the serving `summary.parquet` configs:
431
 
432
  | Column | Type | Description |
433
  |--------|------|-------------|
 
458
  from datasets import load_dataset
459
 
460
  ds = load_dataset("agent-perf-bench/AgentPerfBench", "trace_replay")
461
+ # or "synthetic_distributional", "kernels_labeled", "mse_validation"
462
  ```
463
 
464
  ## Benchmark methodology
465
 
466
  - Closed-loop concurrency with semaphore control.
467
+ - Concurrency levels: {1, 5, 10, 20, 40, 80} (trace_replay), {1, 5, 10, 40, 80, 200, 320} (synthetic_distributional).
468
  - 3-request warmup before each configuration.
469
  - Metrics: TTFT, TPOT, ITL, E2EL, request throughput, token throughput.
470
  - Summary statistics: mean, median, p90, p99.
 
474
  ## Future releases
475
 
476
  - Full per-request and multi-turn granularity data for the main sweep (pending raw JSON availability from collection infrastructure). Curated raw JSONs are included for `mse_validation`.
477
+ - Additional per-kernel roofline profiles beyond the included `kernels_labeled` and `per_layer_oi_cf` artifacts.
478
  - This is version 1.0. Updates will be tagged with semantic versions.
479
 
480
  ## Intended uses
 
486
  ## Limitations
487
 
488
  - Results are specific to tested hardware and software versions (vLLM 0.19.0, SGLang 0.5.9, PyTorch 2.10.0, CUDA 12.8).
489
+ - Synthetic distributional profiles approximate but do not replicate production traffic patterns.
490
+ - Consumer GPU coverage is limited to RTX 3090 and RTX 2080 Ti; no non-NVIDIA accelerators.
491
  - Closed-loop concurrency only; no open-loop (Poisson) arrivals.
492
+ - The model-hardware-concurrency matrix is sparse (12.2% fill for trace_replay, 3.0% for synthetic_distributional). Not all model-hardware combinations are represented.
493
  - No model quality metrics. This is a systems benchmark.
494
 
495
  ## Ethical considerations