Datasets:
url stringlengths 18 9.49k | natural_score float32 0 1 ⌀ |
|---|---|
http://brightbazaaar.wpengine.netdna-cdn.com/wp-content/uploads/2013/06/colorful-home-in-mexico.jpg | 0.202178 |
http://ichef.bbci.co.uk/images/ic/336xn/p036k3pp.jpg | 0.89958 |
0.04968 | |
0.171863 | |
https://images.ulta.com/is/image/Ulta/2302320?$detail$ | 0.30632 |
0.935128 | |
0.868931 | |
0.220399 | |
0.124186 | |
0.873063 | |
0.556558 | |
0.259737 | |
0.488934 | |
0.075599 | |
0.075599 | |
http://lr-assets.storage.googleapis.com/gardimg/400/9780435049607.jpg | 0.055118 |
0.192867 | |
0.392829 | |
0.012658 | |
0.158149 | |
0.075137 | |
0.644879 | |
0.256467 | |
0.797264 | |
0.073661 | |
0.200006 | |
https://chairish-prod.global.ssl.fastly.net/image/product/sized/5044fac4-2745-4311-9dc3-f9d640b1204d/helmut-lubke-sculptural-bar-stools-set-of-3-9404?aspect=fit&width=320&height=320 | 0.267561 |
0.200006 | |
0.200006 | |
http://www.jewelsforme.com/productimages/large/y/10/2374e.jpg | 0.371927 |
0.096557 | |
http://s7d2.scene7.com/is/image/Motosport/MOS-BAG-003B_is?$productdetail264$ | 0.224823 |
0.036101 | |
http://s7d2.scene7.com/is/image/Motosport/MOS-BAG-003B_is?$productdetail264$ | 0.224823 |
0.453489 | |
http://decorstainless.com/uploadfiles/image/201911/1239.png | 0.087862 |
0.11193 | |
http://image1.slideserve.com/1582656/slide11-n.jpg | 0.015946 |
https://tse3.mm.bing.net/th?id=OIP.gs-55Dlc8KYT9XTfrAGOnQEsDH&pid=15.1&P=0&w=300&h=300 | 0.565278 |
0.31788 | |
0.060966 | |
0.914777 | |
0.923865 | |
0.007298 | |
0.973418 | |
http://image.lampsplus.com/is/image/R4996.fpx?qlt=65&wid=236&hei=236&fmt=jpeg | 0.295743 |
0.17519 | |
http://img.omni7.jp/co/productimage/0001/product/42/1106416942/image/1106416942_main_m.jpg | 0.012136 |
http://www.magment.com/wp-content/uploads/2016/10/Brown-Copper-and-Gold-Christmas-Tree.jpg | 0.930325 |
http://st.depositphotos.com/1401847/2610/i/110/depositphotos_26107209-Beekeepers.jpg | 0.944331 |
http://tse2.mm.bing.net/th?id=OIP.b37NMGP3NFDLaQMEYqn-9wHaJ4 | 0.894893 |
https://lf.lids.com/hwl?set=sku[20952141],c[2],w[400],h[300]&call=url[file:product] | 0.245889 |
https://images.carpages.ca/inventory/3056997.92439747?w=320&h=240&q=75&s=19dee924cabd2c8147ce310d91ede192 | 0.391937 |
http://www.lovablequote.com/wp-content/uploads/2017/09/i-promise-i-will-always-do-whatever-i-can-love-lovable-quote.jpg | 0.041598 |
0.940367 | |
0.057772 | |
http://sc02.alicdn.com/kf/HTB12ADaKpXXXXaUXVXXq6xXFXXXx/custom-made-metal-dog-tag-with-printed.jpg_200x200.jpg | 0.186894 |
http://sc01.alicdn.com/kf/HTB1TcfEj22H8KJjy1zkq6xr7pXa3/193612510/HTB1TcfEj22H8KJjy1zkq6xr7pXa3.jpg | 0.130072 |
http://sc01.alicdn.com/kf/HTB1TcfEj22H8KJjy1zkq6xr7pXa3/193612510/HTB1TcfEj22H8KJjy1zkq6xr7pXa3.jpg | 0.130072 |
0.252772 | |
0.004496 | |
0.42363 | |
0.781965 | |
0.087165 | |
0.250631 | |
0.1262 | |
0.683728 | |
0.077801 | |
0.036165 | |
0.040446 | |
0.071291 | |
0.879757 | |
0.19366 | |
http://www.davidsanger.com/images/sanfrancisco/5-620-9915.hongkongshow.x.jpg | 0.533288 |
0.827446 | |
0.877989 | |
0.006249 | |
http://images.crestock.com/5050000-5059999/5058344-xs.jpg | 0.570638 |
0.653473 | |
0.095641 | |
https://process.fs.grailed.com/AJdAgnqCST4iPtnUxiGtTz/cache=expiry:max/rotate=deg:exif/resize=width:2400,fit:crop/output=quality:70/compress/https://process.fs.grailed.com/uYDIyR47T2yelP0xVSUh | 0.343877 |
0.481573 | |
0.91981 | |
0.389525 | |
http://i0.wp.com/venueeventartist.com/imateq/event/446/1126/366730/900SC0/419292.jpeg?strip=all | 0.206685 |
0.244809 | |
0.008085 | |
http://images.shopflowers.net/images/products/SW0_512290.jpg | 0.397581 |
0.019927 | |
0.792261 | |
0.030368 | |
0.046875 | |
http://m.olokaustos.org/uploaded_images/c1597594-dansion-kyrgyzstan-p080-series-pump-p080-03r5c-h8p-00.jpg | 0.212421 |
0.886475 | |
0.037512 | |
http://images.fineartamerica.com/images-small-5/1-golden-sunset-over-farm-field-with-hay-bales-elena-elisseeva.jpg | 0.251134 |
0.344486 | |
0.356114 | |
0.441749 | |
http://brookeandyara.com/wp-content/uploads/2017/08/how-to-write-an-awesome-college-essay.png | 0.047471 |
ReLAION-2B Natural Scores
Naturalness scores for 2.1 billion images from ReLAION-2B-en-research-safe, predicting how "natural" or "photographic" an image looks versus artificial/rendered content.
Quick Start
from datasets import load_dataset
# Load the dataset (streaming recommended due to size)
ds = load_dataset("andropar/relaion2b-natural", streaming=True)
# Filter to natural images only
for row in ds['train']:
if row['natural_score'] and row['natural_score'] > 0.7:
print(row['url']) # Natural photograph URL
Overview
| Total rows | ~2.1 billion |
| Score range | 0.0 (artificial) to 1.0 (natural) |
| Recommended threshold | > 0.7 for natural photographs |
| Format | Parquet (Snappy compressed) |
| Source | ReLAION-2B-en-research-safe |
Example Images
Examples of images at different natural score ranges:
Non-natural (score < 0.3): Graphics, logos, text overlays, screenshots

Low (score 0.3 - 0.5): Mixed content, product images, some editing

Medium (score 0.5 - 0.7): Mostly natural with some artifacts

High (score 0.7 - 0.85): Natural photographs

Very high (score 0.85 - 1.0): Clean natural photographs

Thumbnails shown solely to illustrate dataset characteristics. Source: ReLAION-2B-en-research-safe (Apache 2.0). Underlying images remain under the copyright of their original creators.
Dataset Structure
| Column | Type | Description |
|---|---|---|
url |
string | Image URL from ReLAION-2B |
natural_score |
float32 | Naturalness prediction (0-1), null if no match found in original LAION-2B-en |
Files are named relaion2b_natural_part-*.snappy.parquet.
How the Scores Were Created
Manual labeling: About 26k images were labeled via active learning over a pool of roughly 200k candidates from LAION-2B-en (21k usable after filtering broken URLs). Selection criteria for "natural" images:
- No watermarks, logos, or banners
- No heavy editing (B&W filters, high saturation, photoshopping)
- Must be a real-world scene or object
Classifier training: A logistic regression classifier was trained on CLIP ViT-L/14 image embeddings (768-dim). This simple linear model was chosen deliberately - we verified that nonlinear models (MLPs) do not improve over logistic regression on these features, indicating that the linear classifier efficiently captures the available signal.
Scoring: The classifier was applied to pre-computed CLIP ViT-L/14 embeddings for all of LAION-2B-en (~2.1B images).
Matching: Predicted scores were matched to ReLAION-2B-en-research-safe by URL. Some URLs have null scores where no match was found in the original dataset.
Classifier
The trained classifier is included in the classifier/ directory and can be used to score new images. Since it operates on standard CLIP ViT-L/14 features, it can be applied to any image that can be embedded with CLIP.
Performance
| Metric | Value |
|---|---|
| ROC AUC | 0.89 |
| Average Precision | 0.89 |
| Precision @ threshold 0.7 | 0.89 |
| Recall @ threshold 0.7 | 0.59 |
| Accuracy @ threshold 0.5 | 0.80 |
The classifier was evaluated on a held-out test set of 4,200 labeled images. At the recommended threshold of 0.7, precision is high (89%). The tradeoff is lower recall (59%), meaning some natural images will be missed, but we decided that this trade-off was acceptable for our use case.
Detailed diagnostics (click to expand)
ROC and Precision-Recall curves:
Score distributions by true label:
Confusion matrices:
Typical errors - most misclassifications occur on genuinely ambiguous images:
| False positives (predicted natural, actually non-natural) | False negatives (predicted non-natural, actually natural) |
|---|---|
![]() |
![]() |
False positives tend to be product or studio photography with subtle watermarks/overlays. False negatives tend to be natural scenes with text, heavy cropping, or unusual framing.
Files
| File | Description |
|---|---|
classifier/classifier_weights.json |
Portable weights (JSON) - use this for framework-agnostic inference |
classifier/classifier_weights.npz |
Weights as numpy arrays (coef + intercept) |
classifier/laion_natural_img_clf_vitl14.pkl |
Original scikit-learn pickle |
classifier/diagnostics.json |
Full evaluation metrics |
Usage
Option 1: Framework-agnostic (recommended)
import json
import numpy as np
# Load weights
with open("classifier/classifier_weights.json") as f:
weights = json.load(f)
coef = np.array(weights["coef"], dtype=np.float32)
intercept = weights["intercept"]
def predict_natural_score(clip_embedding):
"""Score a CLIP ViT-L/14 embedding (768-dim, L2-normalized)."""
logit = np.dot(clip_embedding, coef) + intercept
return 1.0 / (1.0 + np.exp(-logit))
# Example: extract features with CLIP and score
import clip, torch
from PIL import Image
model, preprocess = clip.load("ViT-L/14")
image = preprocess(Image.open("photo.jpg")).unsqueeze(0)
with torch.no_grad():
embedding = model.encode_image(image)
embedding /= embedding.norm(dim=-1, keepdim=True)
embedding = embedding.cpu().numpy().squeeze()
score = predict_natural_score(embedding)
print(f"Natural score: {score:.3f}") # > 0.7 = likely a natural photograph
Option 2: scikit-learn
import pickle
with open("classifier/laion_natural_img_clf_vitl14.pkl", "rb") as f:
clf = pickle.load(f)
# clf.predict_proba(embeddings)[:, 1] gives natural scores
Usage Examples
Filter a subset with pandas:
import pandas as pd
df = pd.read_parquet("relaion2b_natural_part-000.snappy.parquet")
# High-quality natural images
natural = df[df['natural_score'] > 0.7]
print(f"Found {len(natural):,} natural images")
# Very high confidence
very_natural = df[df['natural_score'] > 0.9]
Load all files:
from datasets import load_dataset
# Full dataset (streaming)
ds = load_dataset("andropar/relaion2b-natural", streaming=True)
# Or load specific files
import glob
files = glob.glob("relaion2b_natural_part-*.snappy.parquet")
df_all = pd.concat([pd.read_parquet(f) for f in files])
Combine with image downloading:
import requests
from PIL import Image
from io import BytesIO
def download_image(url):
resp = requests.get(url, timeout=10)
return Image.open(BytesIO(resp.content))
# Get natural image URLs and download
natural_urls = df[df['natural_score'] > 0.8]['url'].tolist()
images = [download_image(url) for url in natural_urls[:100]]
Use Cases
- Dataset filtering: Remove non-photographic content from web-scraped image datasets
- Quality assessment: Score images for naturalness before model training
- Research: Study distribution of natural vs. artificial images on the web
- Preprocessing: Filter training data for vision models that need natural photographs
Related Datasets
- andropar/relaion2b-natural-embeddings - CLIP ViT-H/14 embeddings for the ~500M images with natural_score > 0.7
Licensing / Content
This repository contains only metadata (URLs and natural scores). No images are distributed.
- The underlying images are hosted by third-party websites and remain under their original copyrights and terms of use.
- Our additions (naturalness scores, classifier, documentation) are released under CC-BY 4.0.
- This dataset is based on ReLAION-2B-en-research-safe, which is licensed under Apache 2.0.
- Please check license compatibility for any commercial usage.
Limitations
- "Naturalness" reflects our specific labeling criteria - may not match your definition
- These are ML predictions, not ground truth labels
- The classifier was trained on CLIP ViT-L/14 features; performance may differ with other embedding models
- Some URLs may be broken or point to different/removed images
- Null scores indicate URLs not found in original LAION-2B-en dataset
Citation
@inproceedings{
roth2025how,
title={How to sample the world for understanding the visual system},
author={Johannes Roth and Martin N Hebart},
booktitle={8th Annual Conference on Cognitive Computational Neuroscience},
year={2025},
url={https://openreview.net/forum?id=T9k6KkZoca}
}
Questions or issues? Open a discussion!
This dataset is intended for research purposes. Verify license compatibility before commercial use.
- Downloads last month
- 829






