# Judge Dataset Documentation ## Overview The Judge dataset evaluates how well vision-language models (VLMs) can act as judges of computer vision model outputs. Each prompt presents a VLM with one or more encoded predictions and asks it to assess quality — via pairwise comparison, ranking, or absolute scoring. The dataset covers 15 vision tasks, 53 encoding variants, and three question types. **Dataset location:** `Icey444/Judge_questions_v2` on Hugging Face **v2.1 split:** 5,549 items (2,855 pairwise · 2,647 scoring · 47 ranking) --- ## 1. Task Coverage | Task | Category | What is judged | |---|---|---| | `object_detection` | Perception | Bounding boxes (label + coordinates) for specified classes | | `instance_segmentation` | Perception | Per-instance pixel masks for specified classes | | `semantic_segmentation` | Perception | Per-class pixel masks across all categories | | `referring_segmentation` | Perception | Mask of the region referred to by a natural-language expression | | `keypoint` | Perception | 17-keypoint COCO pose skeleton per person | | `depth_estimation` | Perception | Dense monocular depth map | | `lowlevel-deblur` | Restoration | Image deblurring result | | `lowlevel-derain` | Restoration | Image deraining result | | `lowlevel-desnow` | Restoration | Image desnowing result | | `lowlevel-super-resolution` | Restoration | Image super-resolution result | | `generation_controllable` | Generation | Controllable image generation (control condition + prompt) | | `generation_editing` | Generation | Instruction-based image editing | | `generation_inpainting_high_level` | Generation | High-level inpainting (semantic fill) | | `generation_inpainting_low_level` | Generation | Low-level inpainting (seamless texture fill) | | `generation_t2i` | Generation | Text-to-image generation | --- ## 2. Encoding Variants Encoding variants determine **how a model's prediction is presented** to the judge. Each variant transforms raw annotation output (bounding boxes, masks, keypoints, etc.) into a form the VLM can read. There are two broad families: **pixel** (rendered image) and **text** (structured string). Combo encodings pair one text and one pixel form per option. ### 2.1 Object Detection (6 variants) | Stem | Type | Description | |---|---|---| | `pixel_s0_m0` | pixel | Colored bounding boxes, no label text, overlaid on original image | | `pixel_s1_m0` | pixel | Colored boxes + label text, overlaid on original image | | `pixel_s1_m1` | pixel | Colored boxes + label text, rendered on separate black canvas | | `0305` | combo | `text_xyxy` coordinates + `pixel_s1_m0` box image, per option | | `text_xyxy` | text | `{"label":…,"bbox":[x1,y1,x2,y2]}` per detection | | `text_xywh` | text | `{"label":…,"bbox":[x,y,w,h]}` per detection | ### 2.2 Instance Segmentation (12 variants) **Sub-sampled pixel** (downsampled grid, each cell = most-covered instance index): | Stem | Embed | Description | |---|---|---| | `pixel_ss0_m0` | overlay | Grid overlaid on original image | | `pixel_ss0_m1` | separate | Grid on black canvas | | `pixel_ss1_m0_o0_l0_c0_b0` | overlay | … (sub-sampled text) | **Original-resolution pixel** (full-resolution mask rendering): | Stem | Opacity | Label overlay | Color scheme | Bbox style | |---|---|---|---|---| | `pixel_ss1_m0_o0_l1_c0_b1` | 0.5 | text labels | color-by-class | dashed bbox | | `pixel_ss1_m0_o0_l1_c1_b1` | 0.5 | text labels | color-by-instance | dashed bbox | | `pixel_ss1_m0_o1_l1_c0_b1` | 1.0 | text labels | color-by-class | dashed bbox | | `pixel_ss1_m1_o0_l1_c0_b1` | 0.5 | text labels | color-by-class | dashed bbox, separate canvas | | `pixel_ss1_m0_o0_l1_c0_b0` | 0.5 | text labels | color-by-class | no bbox | | `pixel_ss1_m0_o0_l0_c0_b1` | 0.5 | no labels | color-by-class | dashed bbox | **Text-only:** | Stem | Format | |---|---| | `text_polygon` | `{"instance_id":…,"label":…,"polygon":[[x,y],…]}` per instance | | `text_rle` | `{"instance_id":…,"label":…,"rle":{…}}` COCO RLE per instance | | `text_matrix` | 2D integer grid (rows × cols), each cell = instance index | **Combo:** `1742` — polygon text + color-by-instance image, per option. ### 2.3 Semantic Segmentation (8 variants) **Sub-sampled pixel** (3): overlay, separate canvas, text sub-sample. **Original-resolution pixel** (4, all opacity 0.5): | Stem | Label overlay | Color scheme | Canvas | |---|---|---|---| | `pixel_ss1_m0_o0_l1_c0` | text labels | standard palette | overlay | | `pixel_ss1_m0_o0_l1_c1` | text labels | random colors | overlay | | `pixel_ss1_m0_o1_l1_c0` | text labels | standard palette | overlay, full opacity | | `pixel_ss1_m1_o0_l1_c0` | text labels | standard palette | separate canvas | | `pixel_ss1_m0_o0_l0_c0` | no labels | standard palette | overlay | **Text-only:** | Stem | Format | |---|---| | `text_polygon` | `{"label":…,"polygon":[[x,y],…]}` per segment | | `text_matrix` | 2D integer grid, each cell = class index | **Combo:** `4649` — sub-sample text + original-res overlay image, per option. ### 2.4 Referring Segmentation (11 variants) **Sub-sampled pixel** (3): overlay, separate, text sub-sample. **Original-resolution pixel** (5): | Stem | Mask style | Opacity | Canvas | |---|---|---|---| | `pixel_ss1_m0_o0_m0` | filled region | 0.5 | overlay | | `pixel_ss1_m0_o0_m1` | contour only | 0.5 | overlay | | `pixel_ss1_m0_o0_m2` | fill + contour | 0.5 | overlay | | `pixel_ss1_m0_o1_m0` | filled region | 1.0 | overlay | | `pixel_ss1_m1_o0_m0` | filled region | 0.5 | separate canvas | **Text-only:** | Stem | Format | |---|---| | `text_polygon` | `{"label":"","polygon":[[x,y],…]}` | | `text_matrix` | 2D grid; legend maps index → referring expression | **Combo:** `7080` — polygon text + fill+contour image, per option. ### 2.5 Keypoint Detection (8 variants) **Pixel:** | Stem | Style | Color scheme | Canvas | |---|---|---|---| | `pixel_s0_c1_m0` | points only | color-by-instance | overlay | | `pixel_s1_c0_m0` | skeleton | single color (green) | overlay | | `pixel_s1_c1_m0` | skeleton | color-by-instance | overlay | | `pixel_s1_c2_m0` | skeleton | color-by-body-part | overlay | | `pixel_s1_c1_m1` | skeleton | color-by-instance | separate canvas | **Text-only:** | Stem | Format | |---|---| | `text_flat_list` | 34 numbers `[x0..x16, y0..y16]` per person (COCO order) | | `text_part_keyed_json` | `{"person_id":…,"keypoints":[{"name":…,"x":…,"y":…},…]}` | | `text_coco_style` | 51 numbers `[x,y,v]×17` per person | All text formats include the note: *x=0.0, y=0.0 means the keypoint was not detected or is not visible.* ### 2.6 Depth Estimation (3 variants) Each variant is a colormap applied to the predicted depth map: | Stem | Colormap | Semantics | |---|---|---| | `plasma` | Magma/plasma | Bright yellow = closest, dark purple = farthest | | `turbo` | Turbo (rainbow) | Red = closest, blue = farthest | | `gray` | Grayscale | Bright = closest, dark = farthest | ### 2.7 Low-level Restoration (4 tasks × 1 variant each) Each task has a single pixel encoding: the restored output image shown alongside the degraded input. | Task | Input context | |---|---| | `lowlevel-deblur` | Blurry source image | | `lowlevel-derain` | Rainy source image | | `lowlevel-desnow` | Snowy source image | | `lowlevel-super-resolution` | Low-resolution source image | ### 2.8 Image Generation (5 tasks × 1 variant each) Each task shows the generated output image(s) alongside the source context. | Task | Source context shown | |---|---| | `generation_controllable` | Source image (control signal + reference) | | `generation_editing` | Original image before editing | | `generation_inpainting_high_level` | Original image with masked region | | `generation_inpainting_low_level` | Original image with masked region | | `generation_t2i` | No source image (text prompt only) | --- ## 3. Question Types ### 3.1 Pairwise Comparison The judge sees two options (A and B) and selects the better prediction. **Structure:** ``` [] ← original/reference image (if available) You are a judge to decide the quality of answers to a task [based on my given image]. [Task-specific context: class(es) of interest / referring prompt / etc.] Format of predictions: Options: A. [] [text or legend] B. [] [text or legend] . Please answer with A or B. ``` **Pair sampling:** Within each `(image_id, class-of-interest, error_type, prompt)` group, pairs are drawn so that no two annotations with equal `final_score` are paired. Up to 10 pairs per group per task (encoding_analysis) or 50 pairs (judge_analysis). **Answer:** The letter corresponding to the annotation with the higher `final_score`. **Final question phrasing** is sampled from five paraphrases to reduce positional bias: - *Which prediction is better?* - *Which option is a better execution of the vision task?* - *Which option would you prefer as answer to the vision task?* - *Which of the two is the better result?* - *Which option better fulfills the task?* ### 3.2 Ranking The judge sees N options (A through E, or fewer) and ranks them best-to-worst. **Structure:** ``` [original image context] You are a judge to decide the quality of answers to a task. [Task-specific context] Format of predictions: Options: A. [ or text] B. [ or text] ... Rank the predictions from best to worst. Respond with the ranking as a single string of letters only (best first, worst last). For example, BCAED. ``` Groups of 3–5 annotations sharing the same `(image_id, class-of-interest, error_type)` are ranked together. Used in `judge_analysis` only. **Answer:** Letters ordered by descending `final_score`. ### 3.3 Scoring The judge sees a single prediction and assigns a score from 0 to 10. **Structure:** ``` [original image] You are a judge to decide the quality of answers to a task [based on my given image]. [Task-specific context] Format of prediction: [First image: original. Second image: encoded prediction.] ← pixel encodings [Prediction (text): ] ← text encodings Score the quality of the prediction from 0 to 10. 0 = random guessing / worst, 10 = best possible. Please answer with a single score from 0 to 10 only. ``` **Answer:** The annotation's `final_score` (normalized to 0–10). Used in `judge_analysis` only. 20 groups × 5 annotations per group × stems per task. --- ## 4. Prompt Construction Standards ### 4.1 Role Framing Every prompt begins with a judge role sentence tailored to the task: | Task group | Intro pattern | |---|---| | Object detection | "You are a judge to decide the quality of answers to an object detection task. **The class(es) of interest is** {coi}." | | Instance / semantic segmentation | Same pattern with respective task name and COI | | Referring segmentation | "… **The prompt is** '{expression}'." (or "The prompt is the referring expression shown in the options below." if not available at prompt-level) | | Keypoint | "… **The task is pose estimation.**" | | Depth estimation | "… **The task is depth prediction.**" | | Low-level restoration | Task-specific sentence describing the restoration goal | | Generation | Task-specific sentence describing the generation goal + text prompt when available | "based on my given image" is appended when the original image is included as a `` placeholder. ### 4.2 Format Description After the role sentence, the prompt includes a **Format of predictions** block describing the encoding so the judge knows what it is looking at: - **Pixel encodings:** describe the visual style (overlay/canvas, color scheme, opacity, label style). - **Text encodings:** describe the schema (e.g., JSON structure, coordinate conventions, grid dimensions and legend). - **Combo encodings:** each option shows its own format description inline, followed by the encoded image. - **Generation/low-level:** no format description (the prediction is a natural image); the instruction covers the task criterion instead. ### 4.3 Color Legends For encodings where colors carry semantic meaning, a legend is included **per option** (not once globally), because different predictions may contain different classes or instances: - **Object detection pixel:** legend lists each detected class and its assigned color. - **Instance segmentation (color-by-class):** legend lists each class and color. - **Instance segmentation (color-by-instance):** no per-instance legend (color only distinguishes people; skeleton structure is self-evident). - **Semantic segmentation:** legend lists each class and color. - **Keypoint (color-by-part):** legend lists each of the 17 COCO keypoint names and its color. - **Keypoint (color-by-instance):** one sentence describing that all keypoints and links of the same person share a color; no per-person list. - **Keypoint (same color):** "All keypoints and links use a single color (green). No color legend." - **Depth colormaps:** the colormap semantics (which end is near/far) are described in the format block. ### 4.4 Image Placeholder Ordering `` placeholders in the prompt correspond to `media` entries in the same order: 1. **Original/reference image** (first, when present) — always `original_{image_id}.png`. 2. **Option A image** (prediction rendered for annotation A). 3. **Option B image** (prediction rendered for annotation B). Text-only encodings include only the original image (1 image total). Generation/low-level pixel encodings that have no source image in the image index include 2 images (A and B only). Generation tasks with a retrievable source image include 3 images. ### 4.5 Subset Labels Each item carries a `subset` field indicating which run produced it: | Subset | Stems used | Samples | Pairs | Scoring/Ranking | |---|---|---|---|---| | `encoding_analysis` | All `is_base_ablation=1` (51 regular + gen/lowlevel) | 10 per task | 10 per task | No | | `judge_analysis` | `is_final=1` (53) | 20 per task | 50 per task | Yes (20 groups × 5 annotations) | Both runs use `seed=42`. The v2.1 JSON is the deduplicated union (keyed on task + encoding + question_type + annotation IDs). --- ## 5. Data Sources | Task group | Annotation source | Image source | |---|---|---| | Perception tasks | `data_json_v2/*_annotations.json` | `images_v2/*.json` → local files or HF URLs | | Low-level restoration | `data_json_v2/lowlevel_annotations.json` | HF: `Icey444/VisualJudge_images` (prediction URLs) + `images_v2/lowlevel.json` (source URLs) | | Generation | `data_json_v2/generation_annotations.json` | HF: `Icey444/VisualJudge_images` (prediction URLs) + `images_v2/generation_images.json` (source URLs) | Predictions are encoded locally by task-specific encoder scripts (`src/encoders/encode_*.py`) and stored under `output/encoded_v2/`. Original images are cached as `original_{image_id}.png` in the same directory.