small-norb / README.md
haodoz0118's picture
Update README.md
40c88d4 verified
---
license: other
---
# Dataset Card for SmallNORB
## Dataset Description
The **SmallNORB dataset** is a **real-world stereo image dataset** designed for benchmarking algorithms in **disentangled representation learning** and **unsupervised representation learning**. It was introduced by **LeCun et al. (2004)** for evaluating **generic object recognition** with **invariance to pose and lighting**.
Unlike synthetic datasets such as **dSprites** or **MPI3D**, which are generated as a **complete Cartesian product of factors** (i.e. every possible combination is present), SmallNORB consists of **real photographs** of physical toy objects under controlled variations, but **not every combination of factors is present** — for example, object instances are sampled randomly and the views (azimuth, elevation, lighting) do not form an exact grid.
Each sample contains **two views**:
- **Left image** (96x96 grayscale)
- **Right image** (96x96 grayscale)
Each image pair is associated with **4 known factors of variation** and **instance index**:
- **category** (object type)
- **instance** (specific object instance)
- **elevation** (camera tilt angle)
- **azimuth** (camera rotation angle)
- **lighting** (lighting condition)
The dataset allows researchers to evaluate **representation learning on real-world 3D objects**, under complex lighting and pose variations. SmallNORB provides an **official train/test split**. Typically, **instance** is not considered as a factor.
![Dataset visualization](https://huggingface.co/datasets/randall-lab/small-norb/resolve/main/animation0.gif)
## Dataset Source
- **Homepage**: [https://cs.nyu.edu/~ylclab/data/norb-v1.0-small/](https://cs.nyu.edu/~ylclab/data/norb-v1.0-small/)
- **License**: other. Small NORB is public domain, for research use.
- **Paper**: Yann LeCun et al. _Learning methods for generic object recognition with invariance to pose and lighting_. CVPR 2004.
## Dataset Structure
|Factors|Possible Classes (Indices)|Values|
|---|---|---|
|category|0,...,4| airplane=0, car=1, truck=2, human=3, animal=4 |
|instance|0,...,9| specific instance of object |
|elevation|0,...,8| 9 elevation angles |
|azimuth|0,...,17| azimuth originally 0,2,...,34 → scaled to 0-17 |
|lighting|0,...,5| 6 lighting conditions |
**Note:** The dataset is not a complete Cartesian product — **instances and views are sampled** in the original design. Each sample contains a **left image** and a **right image**, both corresponding to the same factors.
## Example Usage
Below is a quick example of how to load this dataset via the Hugging Face Datasets library:
```python
from datasets import load_dataset
# Load train set
train_ds = load_dataset("randall-lab/small-norb", split="train", trust_remote_code=True)
# Load test set
# test_ds = load_dataset("randall-lab/small-norb", split="test", trust_remote_code=True)
# Access a sample
example = train_ds[0]
left_image = example["left_image"]
right_image = example["right_image"]
label = example["label"] # [category, elevation, azimuth, lighting]
# Label breakdown
category = example["category"] # 0-4
instance = example["instance"] # 0-9
elevation = example["elevation"] # 0-8
azimuth = example["azimuth"] # 0-17
lighting = example["lighting"] # 0-5
# Visualize
left_image.show()
right_image.show()
print(f"Label (factors): {label}")
```
If you are using colab, you should update datasets to avoid errors
```
pip install -U datasets
```
## Citation
```
@inproceedings{lecun2004learning,
title={Learning methods for generic object recognition with invariance to pose and lighting},
author={LeCun, Yann and Huang, Fu Jie and Bottou, Leon},
booktitle={Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004.},
volume={2},
pages={II--104},
year={2004},
organization={IEEE}
}
```