imagenet-100 / scripts /pytorch_dataloader.py
basavyr's picture
update data dir as variable
fe2aefa
import torch
import pandas as pd
from torch.utils.data import Dataset, DataLoader
from PIL import Image
import io
from pathlib import Path
from torchvision import transforms
from typing import Optional, Callable, Tuple, Union
class ImageNet100Parquet(Dataset):
def __init__(self, data_dir: str, split: str = "train",
transform: Optional[Callable] = None) -> None:
"""
Initialize ImageNet-100 dataset from parquet files.
Args:
data_dir: Path to directory containing parquet files.
split: Dataset split - either "train" or "validation".
transform: Optional torchvision transforms to apply to images.
Raises:
FileNotFoundError: If data directory doesn't exist.
ValueError: If no parquet files found for the specified split.
"""
self.data_dir = Path(data_dir)
self.transform = transform
# Validate data directory
if not self.data_dir.exists():
raise FileNotFoundError(
f"Data directory not found: {self.data_dir}")
# Validate split
valid_splits = ["train", "validation"]
if split not in valid_splits:
raise ValueError(
f"Invalid split '{split}'. Must be one of: {valid_splits}")
# Load all parquet files for the split
parquet_files = sorted(self.data_dir.glob(f"{split}-*.parquet"))
if not parquet_files:
raise ValueError(
f"No parquet files found for split '{split}' in {self.data_dir}")
try:
self.data = pd.concat([pd.read_parquet(f)
for f in parquet_files], ignore_index=True)
except Exception as e:
raise RuntimeError(f"Failed to load parquet files: {e}")
def __len__(self) -> int:
"""Return the total number of samples in the dataset."""
return len(self.data)
def __getitem__(self, idx: int) -> Tuple[Union[Image.Image, torch.Tensor], torch.Tensor]:
"""
Get a sample from the dataset at the given index.
Args:
idx: Index of the sample to retrieve.
Returns:
Tuple of (image, label) where image is a PIL Image or transformed tensor,
and label is a torch tensor with dtype long.
Raises:
IndexError: If idx is out of range.
RuntimeError: If image decoding fails.
"""
if idx >= len(self.data):
raise IndexError(
f"Index {idx} out of range for dataset of size {len(self.data)}")
try:
row = self.data.iloc[idx]
# Decode image from bytes (stored in dict)
image_bytes = row['image']['bytes']
image = Image.open(io.BytesIO(image_bytes)).convert('RGB')
label = torch.tensor(row['label'], dtype=torch.long)
if self.transform:
image = self.transform(image)
return image, label
except Exception as e:
raise RuntimeError(f"Failed to load sample at index {idx}: {e}")
def main() -> None:
data_dir = "../data"
# Define transforms
transform = transforms.Compose([
transforms.Resize((224, 224)),
transforms.ToTensor(),
])
# Create datasets
train_dataset = ImageNet100Parquet(data_dir, "train", transform)
test_dataset = ImageNet100Parquet(data_dir, "validation", transform)
print(f"Train dataset size: {len(train_dataset)}")
print(f"Test dataset size: {len(test_dataset)}")
# Create dataloaders
train_loader = DataLoader(train_dataset, batch_size=8, shuffle=True)
test_loader = DataLoader(test_dataset, batch_size=8, shuffle=False)
# Test iteration
print("\nTesting train loader iteration...")
stop_idx = 0
for x, y_true in train_loader:
print(f"Batch shape: {x.shape}, Labels shape: {y_true.shape}")
stop_idx += 1
if stop_idx > 10:
break
stop_idx = 0
print("\nTesting test loader iteration...")
for x, y_true in test_loader:
print(f"Batch shape: {x.shape}, Labels shape: {y_true.shape}")
stop_idx += 1
if stop_idx > 10:
break
if __name__ == "__main__":
main()