| | --- |
| | license: mit |
| | --- |
| | |
| | # HC-Bench |
| |
|
| | **HC-Bench** is a compact multi-part image benchmark for evaluating recognition and prompting robustness, especially in **hidden-content** scenes. It contains: |
| |
|
| | - **object/** — 56 base images and 56 *hidden* variants of the same lemmas, plus prompts and metadata. |
| | - **text/** — 56 Latin/English and 56 Chinese lemma–description pairs with matching PNGs. |
| | - **wild/** — 53 in-the-wild images for additional generalization checks. |
| |
|
| | --- |
| |
|
| | ## Repository structure |
| |
|
| | ``` |
| | |
| | HC-Bench/ |
| | ├─ object/ |
| | │ ├─ base/ # 56 base images (7 types × 8 lemmas) |
| | │ ├─ hidden/ # 56 hidden-content variants (same lemmas) |
| | │ ├─ image\_base.txt # 7 types and their 8 lemmas each |
| | │ ├─ image\_generate\_prompts.txt# per-lemma scene prompts used for generation |
| | │ └─ lemmas\_descriptions.json # \[{Type, Lemma, Description}] × 56 |
| | ├─ text/ |
| | │ ├─ Latin/ # 28 English PNGs |
| | │ ├─ Chinese/ # 28 Chinese PNGs |
| | │ ├─ English\_text.json # 56 entries (Type, Length, Rarity, Lemma, Description) |
| | │ └─ Chinese\_text.json # 56 entries (Type, Length, Rarity, Lemma, Description) |
| | └─ wild/ # 53 PNGs |
| | |
| | ```` |
| |
|
| | --- |
| |
|
| | ## Contents |
| |
|
| | ### `object/` |
| | - **`base/`**: Canonical image per lemma (e.g., `Apple.jpg`, `Einstein.png`). |
| | - **`hidden/`**: Composite/camouflaged image for the *same* lemma set (e.g., `apple.png`, `einstein.png`). |
| | - **`image_base.txt`**: The 7 high-level types and their 8 lemmas each (Humans, Species, Buildings, Cartoon, Furniture, Transports, Food). |
| | - **`image_generate_prompts.txt`**: Per-lemma prompts used to compose/generate scenes (e.g., *“A monorail cutting through a futuristic city with elevated walkways”* for `notredame`). |
| | - **`lemmas_descriptions.json`**: Minimal metadata with `{Type, Lemma, Description}` aligned 1:1 with the 56 lemmas. |
| |
|
| | ### `text/` |
| | - **`Latin/`** & **`Chinese/`**: 28 images each (total 56). |
| | - **`English_text.json`** & **`Chinese_text.json`**: 56-entry lists pairing lemmas to descriptions in two languages. |
| | (Note: The `English_text.json`/`Chinese_text.json` files include extra fields `Length` and `Rarity` for flexibility.) |
| |
|
| | ### `wild/` |
| | - 53 natural/urban scenes for robustness and transfer evaluation. |
| |
|
| | --- |
| |
|
| | ## Quick start (🤗 Datasets) |
| |
|
| | > HC-Bench uses the **ImageFolder**/“imagefolder” style. Class labels are inferred from directory names when present (e.g., `base`, `hidden`). If you prefer raw images without labels, pass `drop_labels=True`. |
| | |
| | ### Load **object/base** and **object/hidden** |
| | ```python |
| | from datasets import load_dataset |
| |
|
| | base = load_dataset( |
| | "imagefolder", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/object/base/*", |
| | split="train", |
| | drop_labels=True, # drop automatic label inference |
| | ) |
| | |
| | hidden = load_dataset( |
| | "imagefolder", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/object/hidden/*", |
| | split="train", |
| | drop_labels=True, |
| | ) |
| | ```` |
| | |
| | ### Load **wild/** |
| |
|
| | ```python |
| | wild = load_dataset( |
| | "imagefolder", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/wild/*", |
| | split="train", |
| | drop_labels=True, |
| | ) |
| | ``` |
| |
|
| | ### Load the **JSON** metadata (English/Chinese) |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | en = load_dataset( |
| | "json", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/text/English_text.json", |
| | split="train", |
| | ) |
| | zh = load_dataset( |
| | "json", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/text/Chinese_text.json", |
| | split="train", |
| | ) |
| | ``` |
| |
|
| | > Docs reference: `load_dataset` for JSON & files, and ImageFolder for image datasets. |
| | |
| | --- |
| | |
| | ## Pairing base/hidden with metadata |
| | |
| | Filenames differ in casing/spaces between `base/` (`Apple.jpg`) and `hidden/` (`apple.png`). Use `object/lemmas_descriptions.json` as the canonical list of 56 lemmas and join by `Lemma`: |
| |
|
| | ```python |
| | import pandas as pd |
| | from datasets import load_dataset |
| | |
| | # 1) Canonical lemma list |
| | lemmas = load_dataset( |
| | "json", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/object/lemmas_descriptions.json", |
| | split="train", |
| | ).to_pandas() |
| | |
| | # 2) Build (lemma -> file) maps |
| | def to_lemma(name): # normalize filenames to lemma |
| | import re, os |
| | stem = os.path.splitext(os.path.basename(name))[0] |
| | return re.sub(r"\s+", "", stem).lower() |
| | |
| | base_ds = load_dataset( |
| | "imagefolder", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/object/base/*", |
| | split="train", |
| | drop_labels=True, |
| | ) |
| | hidden_ds = load_dataset( |
| | "imagefolder", |
| | data_files="https://huggingface.co/datasets/JohnnyZeppelin/HC-Bench/resolve/main/object/hidden/*", |
| | split="train", |
| | drop_labels=True, |
| | ) |
| | |
| | import os |
| | base_map = {to_lemma(x["image"].filename): x["image"] for x in base_ds} |
| | hidden_map= {to_lemma(x["image"].filename): x["image"] for x in hidden_ds} |
| | |
| | # 3) Join |
| | lemmas["base_image"] = lemmas["Lemma"].apply(lambda L: base_map.get(L.lower())) |
| | lemmas["hidden_image"] = lemmas["Lemma"].apply(lambda L: hidden_map.get(L.lower())) |
| | ``` |
| |
|
| | --- |
| |
|
| |
|
| |
|
| | --- |
| |
|
| | ## Statistics |
| |
|
| | * `object/base`: 56 images |
| | * `object/hidden`: 56 images |
| | * `text/Latin`: 28 images |
| | * `text/Chinese`: 28 images |
| | * `wild`: 53 images |
| |
|
| | --- |
| |
|
| | ## Citation |
| |
|
| | If you use **HC-Bench**, please cite: |
| |
|
| | ```bibtex |
| | @misc{li2025semvinkadvancingvlmssemantic, |
| | title={SemVink: Advancing VLMs' Semantic Understanding of Optical Illusions via Visual Global Thinking}, |
| | author={Sifan Li and Yujun Cai and Yiwei Wang}, |
| | year={2025}, |
| | eprint={2506.02803}, |
| | archivePrefix={arXiv}, |
| | primaryClass={cs.CL}, |
| | url={https://arxiv.org/abs/2506.02803}, |
| | } |
| | ``` |
| |
|
| | --- |
| |
|
| |
|
| |
|
| |
|
| |
|