File size: 5,879 Bytes
6c38604
9feed8c
6c38604
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3506923
6c38604
 
 
 
 
d0510f4
3506923
6c38604
 
3506923
 
 
6c38604
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9feed8c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
---
license: other
language:
  - en
task_categories:
  - visual-question-answering
  - automatic-speech-recognition
  - text-generation
tags:
  - evaluation
  - benchmark
  - multimodal
  - edge-inference
  - on-device
  - litert-lm
  - image
  - audio
  - text
  - multi-turn
size_categories:
  - n<1K
configs:
  - config_name: default
    data_files:
      - split: test
        path: data/test/metadata.jsonl
pretty_name: EdgeMMEval
---

# EdgeMMEval

Minimal multimodal evaluation dataset for on-device inference testing.
Covers functional correctness, accuracy, latency stress, and memory
pressure across image, audio, text, multi-turn, combination, structured
output, and tool-calling cases.

## Dataset summary

The test split is defined in `data/test/metadata.jsonl` (**200** rows). Each
row has a `test_id` (for example `IMG-001`, `STO-020`) and a `modality`.

| Modality             | Samples | Focus |
|----------------------|--------:|--------|
| Image                | 34      | VQA, OCR, description, classification, resolution / loop stress |
| Audio                | 27      | Transcription, spoken QA, translation, noise and edge cases |
| Text                 | 41      | QA, translation, summarization, reasoning-style prompts |
| Multi-turn           | 24      | Context retention, KV-cache stress |
| Combination          | 28      | Cross-modal alignment |
| Structured output    | 32      | JSON schema, regex, grammar-style constraints (`constraint_type`) |
| Tool call            | 14      | Correct tool name, arguments, or valid refusal text |
| **Total**            | **200** | |

## Usage

```python
from datasets import load_dataset

ds = load_dataset("CortexSwarm/EdgeMMEval", split="test")
print(ds[0])
```

## Scoring

Each sample includes a `reference` field and usually `reference_variants` for
automatic scoring. The scorer lives in this repo at `scripts/score.py`.

**Pipeline**

1. Run your model on each `test_id` and collect the model’s output string.
2. Write a JSON object to **`results.json`** at the repo root: keys are
   `test_id` values, values are the raw model outputs (strings).
3. Run `python scripts/score.py`. It reads `data/test/metadata.jsonl` and
   `results.json`, then writes **`report.json`**.

**Metrics and pass rules** (see constants at the top of `score.py`)

- Most tasks: **BLEU-1** vs `reference_variants`; pass if score **≥ 0.5**.
- Summarization-style tasks (`task` in `summarization` / `summarize`): **ROUGE-L**;
  pass if **≥ 0.4**.
- **Structured output**: format check (JSON Schema, regex, or grammar-style
  heuristic) plus content BLEU; pass if format is valid and BLEU **≥ 0.3**.
- **Tool call**: compares expected tool/args or valid text-only refusal;
  separate logic in `score_tool_call`.

**`report.json` shape**

- `summary`: `total_scored`, `total_passed`, `overall_avg`, `pass_rate_pct`,
  `verdict` (`✓ INFERENCE WORKING` if pass rate ≥ 80%, else
  `✗ ISSUES DETECTED`), and `skipped` (test IDs with **no** entry in
  `results.json`).
- `by_modality`: average score and pass counts per modality (empty if nothing
  was scored).
- `samples`: per-test rows with scores, metrics, and pass/fail.

If **`total_scored` is 0**, every test ID was skipped—typically **`results.json`
is missing or does not map test IDs to outputs**. Fix the results file and
re-run the scorer.

## License

The majority of this dataset is **CC BY 4.0**. A small subset of image files
comes from [Unsplash](https://unsplash.com/) and is governed by the
**[Unsplash License](https://unsplash.com/license)** instead.

### CC BY 4.0 (metadata, text tasks, original media, tooling, and most images)

The following are licensed under **Creative Commons Attribution 4.0**
([CC BY 4.0](https://creativecommons.org/licenses/by/4.0/)):

- `data/test/metadata.jsonl` (prompts, references, labels, structure).
- All **text**, **multi-turn**, **combination**, **structured output**, and
  **tool call** samples.
- **Audio** clips produced from project `samples/` recordings (see
  `scripts/collect_all_audio.sh`).
- **Images** generated in-repo by `scripts/collect_all_images.py` using PIL—
  synthetic shapes, UI mockups, charts, QR patterns, blank canvases, sequential
  frames, and all derived images (blur, crop, rotation, collage, meme, watermark,
  overexposed) whose source is a PIL-generated file rather than an Unsplash photo.
  This covers `IMG-001`–`IMG-005`, `IMG-007`–`IMG-009`, `IMG-015`–`IMG-017`,
  `IMG-021`–`IMG-024`, `IMG-026`–`IMG-033`.
- `IMG-020` (built from `samples/v2/real-receipt.webp`, author-provided).
- Scripts and scorer logic (`scripts/`, `upload.py`).

Reuse requires **attribution** to **EdgeMMEval** and a link to this dataset or
source repository.

### Unsplash License (specific image files)

The following files under `data/test/images/` are photographs downloaded from
[Unsplash](https://unsplash.com/) (URLs in `scripts/collect_all_images.py`)
and remain under the **[Unsplash License](https://unsplash.com/license)**:

**Direct downloads:** `IMG-006.jpg` (4K mountain; if the download failed and
the script used its generated fallback, that copy is CC BY 4.0 instead),
`IMG-010.jpg`, `IMG-011.jpg`, `IMG-012.jpg`, `IMG-013.jpg`, `IMG-014.jpg`,
`IMG-025.jpg`, `IMG-034.jpg`.

**Derivatives of those photos:** `IMG-018.jpg` (180° rotation of `IMG-010`),
`IMG-019.jpg` (JPEG-compressed from `IMG-010`).

The Unsplash License permits free use and modification; you may not sell
unmodified copies or build a competing image-service from the content—see the
[full license text](https://unsplash.com/license) for details.

### Summary

| Part | License |
|------|---------|
| Metadata, text tasks, scripts, audio, PIL-generated images, receipt | **CC BY 4.0** |
| Unsplash photos and two derivatives listed above | **[Unsplash License](https://unsplash.com/license)** |