language:
- en
size_categories:
- 100K<n<1M
Dataset Summary
DeepShade is a multimodal dataset designed for shade simulation via text-conditioned image generation. It captures realistic outdoor scenes and their corresponding shade conditions over time, enabling supervised training of diffusion-based models that can simulate sun-shade transitions based on spatial layout and temporal context.
The dataset was introduced in the DeepShade project (IJCAI 2025 submission), and it supports research in text-to-image generation.
Dataset Files Structure:
Each city has a dedicated zip folder which contains source, target and the test and train json files There is one common zip file for all the satellite images of all the cities
Dataset Structure
Data Modality: Multimodal (Image, Text, Time) Number of samples: ~100,000 Resolution: 1024×1024 images
Formats: .png images (rendered) .json metadata files with the following fields:
- Source image file path
- Target image file path
- Prompt
