--- license: mit task_categories: - text-generation - image-segmentation - image-to-text - image-to-image language: - en size_categories: - n<1K --- # Forest-Change ## Overview Forest-Change is the first benchmark dataset specifically designed for joint forest change detection and captioning in remote sensing imagery. It provides bi-temporal satellite images, pixel-level deforestation masks, and multi-granularity semantic captions describing forest cover changes in tropical and subtropical regions. ## Dataset Details - **Total Examples**: 334 annotated bi-temporal image pairs - **Spatial Resolution**: ~30m/pixel (medium resolution) - **Original Image Size**: 480×480 pixels (cropped from larger scenes) - **Processed Image Size**: 256×256 pixels (resized for model training) - **Temporal Resolution**: 1 year between image pairs - **Geographic Focus**: Tropical and subtropical deforestation fronts ## Dataset Splits - **Training**: 270 examples (~80%) - **Validation**: 31 examples (~10%) - **Test**: 33 examples (~10%) ## Data Format Each example contains: - **Image A**: Pre-change RGB satellite image - **Image B**: Post-change RGB satellite image - **Change Mask**: Binary segmentation mask (0=no change, 1=deforestation) - **Captions**: Five captions describing the forest change event with varied granularity ## Data Sources - **Imagery Source**: Google Earth Engine (GEE) - **Base Dataset**: Derived from Hewarathna et al. (2024) forest ecosystem change detection dataset - **Validation**: Forest cover changes verified through Global Forest Watch (GFW) platform - **Geographic Selection**: Based on WWF 2015 Deforestation Fronts report ## Caption Generation Captions are generated through a hybrid two-stage approach: 1. **Human Annotation**: One caption per example manually created by domain annotators describing observed changes 2. **Rule-Based Generation**: Four additional captions automatically generated based on quantitative mask properties: - Percentage of newly deforested area (binned into descriptive severity levels) - Size and number of individual change patches - Spatial distribution patterns of deforestation - Variation in patch sizes This approach ensures both semantic richness from human expertise and consistent structural variation across captions. ## Key Characteristics - **Change Coverage**: - Mean: <5% deforestation per image - Maximum: 40% deforestation - Distribution: Heavily skewed toward lower deforestation percentages - **Caption Length**: Bimodal distribution with both concise and detailed descriptions - **Change Patterns**: Diverse deforestation manifestations including: - Scattered small patches across forest areas - Concentrated clearing zones - Edge-of-clearing expansion patterns - Highly variable patch sizes and configurations - **Caption Content**: Descriptions emphasize: - Degree/severity of forest loss - Spatial location within the image - Patch characteristics (size, number, distribution) ## Preprocessing - All images resized to 256×256 pixels for consistency - Change masks binarized (0=no change, 1=change) - Bi-temporal image pairs pre-aligned - Per-channel normalization using dataset-specific mean and standard deviation statistics - No atmospheric correction applied - No cloud masking applied (some samples contain partial cloud occlusion) ## Use Cases - Forest change detection and monitoring - Deforestation segmentation in natural environments - Change captioning for ecological applications - Multi-task learning for remote sensing - Benchmarking vision-language models on forest imagery - Training interactive forest analysis systems - Developing automated forest monitoring workflows ## Evaluation Metrics Due to severe class imbalance (most pixels are no-change), evaluation requires: - **Per-class IoU**: Separate metrics for change and no-change classes - **Mean IoU (mIoU)**: Average of both class IoUs - **Caption metrics**: BLEU-n (n=1,2,3,4), METEOR, ROUGE-L, CIDEr-D - **Note**: Overall accuracy is not recommended due to class imbalance ## Limitations - **Dataset Scale**: Limited to 334 examples, restricting model generalization - **Scene Diversity**: Limited number of unique geographic sites due to cropping and augmentation strategy - **Class Imbalance**: Severe imbalance with most pixels representing no-change, challenging for detection models - **Caption Quality**: Majority of captions are rule-based, limiting linguistic variation and naturalness - **Geographic Grounding**: Limited incorporation of geographic features and contextual information in captions - **Spatial Resolution**: Medium resolution (~30m/pixel) limits detection of very small-scale changes - **Temporal Coverage**: Fixed 1-year intervals between image pairs - **Atmospheric Effects**: Some samples affected by partial cloud occlusion - **Edge Boundaries**: Fuzzy boundaries at deforestation patch edges complicate precise segmentation ## Citation If you use this dataset, please cite: ```bibtex @article{brock2024forestchat, title={Forest-Chat: Adapting Vision-Language Agents for Interactive Forest Change Analysis}, author={Brock, James and Zhang, Ce and Anantrasirichai, Nantheera}, journal={Ecological Informatics}, year={2024} } @article{hewarathna2024change, title={Change detection for forest ecosystems using remote sensing images with siamese attention u-net}, author={Hewarathna, AI and Hamlin, L and Charles, J and Vigneshwaran, P and George, R and Thuseethan, S and Wimalasooriya, C and Shanmugam, B}, journal={Technologies}, volume={12}, number={9}, pages={160}, year={2024} } ``` Paper page can be found at: https://huggingface.co/papers/2601.04497 ## License MIT License - Academic re-use purpose only ## Contact For questions or issues regarding this dataset, please contact: - James Brock: james.brock@bristol.ac.uk - School of Computer Science, University of Bristol