metaphi-ai's picture
Update README.md
42485f3 verified
---
task_type: figma_to_code
dataset_info:
features:
- name: id
dtype: string
- name: url
dtype: string
- name: tags
list: string
- name: description
dtype: string
- name: figma_data
dtype: string
- name: access_level
dtype: string
- name: created_at
dtype: string
splits:
- name: train
num_bytes: 64990
num_examples: 37
download_size: 35843
dataset_size: 64990
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: mit
language:
- en
tags:
- figma-to-code
- autonomous-agents
pretty_name: CREW:Figma-to-Code
size_categories:
- n<1K
---
# CREW: Figma to Code
A benchmark for evaluating AI coding agents on **Figma-to-code generation** — converting real-world Figma community designs
into production-ready React + Tailwind CSS applications.
Each task gives an agent full access to a Figma file via MCP tools. The agent must extract the design system, generate
components, build successfully, and deploy a live preview. Outputs are evaluated through human preference (ELO) and
automated verifiers.
## Benchmark at a Glance
| | |
|---|---|
| **Tasks** | 37 real Figma community designs |
| **Agents tested** | Claude Code (Opus 4.6), Codex (GPT-5.2), Gemini CLI (3.1 Pro) |
| **Total runs** | 96 autonomous agent executions |
| **Human evaluations** | 135 pairwise preference votes |
| **Primary metric** | Human Preference ELO (Bradley-Terry) |
| **Task duration** | 6–40 expert-hours equivalent per task |
## Leaderboard (Human Preference ELO)
| Rank | Agent | Model | ELO | 95% CI | Win % |
|------|-------|-------|-----|--------|-------|
| 1 | Codex | GPT-5.2 | 1054 | [1005, 1114] | 56.8% |
| 2 | Claude Code | Opus 4.6 | 1039 | [987, 1093] | 55.1% |
| 3 | Gemini CLI | 3.1 Pro | 907 | [842, 965] | 31.1% |
Top two agents are statistically indistinguishable (p=0.67, Cohen's h=0.08); both significantly outperform Gemini CLI
(p<0.05).
Live leaderboard: [evals.metaphi.ai](https://evals.metaphi.ai)
## Dataset Schema
Each row represents one Figma design task:
| Column | Type | Description |
|--------|------|-------------|
| `id` | string | Unique task identifier |
| `figma_data.figma_file_key` | string | Figma file ID for API access |
| `figma_data.figma_file_url` | string | Full Figma URL |
| `figma_data.figma_file_name` | string | Design name/title |
| `figma_data.is_site` | bool | Whether design is a Figma Site |
| `description` | string | Design context and task description |
## Task Scenarios
Tasks span 7 complexity levels, from single-component extraction to full multi-page applications:
1. **E-commerce Product Page** (12 hrs) — PDP with image gallery, variant selectors, inventory states, cart integration
2. **Mobile Onboarding Flow** (16 hrs) — Multi-step flow with transitions, conditional branching, state management
3. **Component Set with States** (6 hrs) — Variant matrix extraction, typed props, conditional rendering
4. **Design Tokens to Theme** (8 hrs) — Variables, typography, effects → Tailwind config + CSS custom properties
5. **Multi-Page Webapp** (40 hrs) — 5+ pages with routing, shared components, consistent theming
6. **Animation-Heavy Interface** (20 hrs) — Smart Animate → Framer Motion with precise timing choreography
7. **Messy Enterprise File** (24 hrs) — Real-world chaos: inconsistent naming, duplicates, orphaned components
## Data Curation
Sourced from licensing partnerships with enterprises, domain-experts and community dataset curation.
## Evaluation Framework
| Verifier | Type | Method |
|----------|------|--------|
| **Human Preference** | Subjective | Pairwise comparison → Bradley-Terry ELO |
| **Visual Judge** | Subjective | VLM screenshot comparison (design vs. output) |
| **Skill Verifier** | Subjective | Task-specific rubrics (build, tokens, components, fidelity) |
| **Behavior Verifier** | Subjective | Agent trajectory analysis (error recovery, tool usage) |
## Agent Error Recovery
Across 96 runs, agents encountered 590 errors with a 70.3% autonomous recovery rate:
| Error Type | Count | Recovery |
|------------|-------|----------|
| Tool call failure | 419 | 66.3% |
| Git error | 64 | 75.0% |
| Syntax error | 33 | 90.9% |
| Build error | 11 | 100.0% |
## Usage
```python
from datasets import load_dataset
ds = load_dataset("metaphilabs/figma", split="train")
# Each row contains a Figma file key for API access
for task in ds:
print(task["id"], task["figma_data"]["figma_file_name"])
Citation
@misc{metaphi2026crew,
title={CREW: Enterprise Agent Benchmarks},
author={Metaphi Labs},
year={2026},
url={https://evals.metaphi.ai}
}
Links
- Leaderboard: https://evals.metaphi.ai
- Website: https://metaphi.ai
- Collection: https://huggingface.co/collections/metaphilabs/crew