File size: 3,887 Bytes
a729b9c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c74e593
a729b9c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40c88d4
a729b9c
 
40c88d4
a729b9c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
---
license: other
---
# Dataset Card for SmallNORB

## Dataset Description

The **SmallNORB dataset** is a **real-world stereo image dataset** designed for benchmarking algorithms in **disentangled representation learning** and **unsupervised representation learning**. It was introduced by **LeCun et al. (2004)** for evaluating **generic object recognition** with **invariance to pose and lighting**.

Unlike synthetic datasets such as **dSprites** or **MPI3D**, which are generated as a **complete Cartesian product of factors** (i.e. every possible combination is present), SmallNORB consists of **real photographs** of physical toy objects under controlled variations, but **not every combination of factors is present** — for example, object instances are sampled randomly and the views (azimuth, elevation, lighting) do not form an exact grid.

Each sample contains **two views**:
- **Left image** (96x96 grayscale)
- **Right image** (96x96 grayscale)

Each image pair is associated with **4 known factors of variation** and **instance index**:
- **category** (object type)
- **instance** (specific object instance)
- **elevation** (camera tilt angle)
- **azimuth** (camera rotation angle)
- **lighting** (lighting condition)

The dataset allows researchers to evaluate **representation learning on real-world 3D objects**, under complex lighting and pose variations. SmallNORB provides an **official train/test split**. Typically, **instance** is not considered as a factor.
![Dataset visualization](https://huggingface.co/datasets/randall-lab/small-norb/resolve/main/animation0.gif)

## Dataset Source
- **Homepage**: [https://cs.nyu.edu/~ylclab/data/norb-v1.0-small/](https://cs.nyu.edu/~ylclab/data/norb-v1.0-small/)
- **License**: other. Small NORB is public domain, for research use.
- **Paper**: Yann LeCun et al. _Learning methods for generic object recognition with invariance to pose and lighting_. CVPR 2004.

## Dataset Structure
|Factors|Possible Classes (Indices)|Values|
|---|---|---|
|category|0,...,4| airplane=0, car=1, truck=2, human=3, animal=4 |
|instance|0,...,9| specific instance of object |
|elevation|0,...,8| 9 elevation angles |
|azimuth|0,...,17| azimuth originally 0,2,...,34 → scaled to 0-17 |
|lighting|0,...,5| 6 lighting conditions |

**Note:** The dataset is not a complete Cartesian product — **instances and views are sampled** in the original design. Each sample contains a **left image** and a **right image**, both corresponding to the same factors.

## Example Usage

Below is a quick example of how to load this dataset via the Hugging Face Datasets library:

```python
from datasets import load_dataset

# Load train set
train_ds = load_dataset("randall-lab/small-norb", split="train", trust_remote_code=True)

# Load test set
# test_ds = load_dataset("randall-lab/small-norb", split="test", trust_remote_code=True)

# Access a sample
example = train_ds[0]
left_image = example["left_image"]
right_image = example["right_image"]
label = example["label"]  # [category, elevation, azimuth, lighting]

# Label breakdown
category = example["category"]      # 0-4
instance = example["instance"]      # 0-9
elevation = example["elevation"]    # 0-8
azimuth = example["azimuth"]        # 0-17
lighting = example["lighting"]      # 0-5

# Visualize
left_image.show()
right_image.show()

print(f"Label (factors): {label}")
```
If you are using colab, you should update datasets to avoid errors
```
pip install -U datasets
```
## Citation
```
@inproceedings{lecun2004learning,
  title={Learning methods for generic object recognition with invariance to pose and lighting},
  author={LeCun, Yann and Huang, Fu Jie and Bottou, Leon},
  booktitle={Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004.},
  volume={2},
  pages={II--104},
  year={2004},
  organization={IEEE}
}
```