Datasets:
Search is not available for this dataset
image
imagewidth (px) 480
480
| label
class label 3
classes |
|---|---|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
|
0A
|
End of preview. Expand
in Data Studio
Forest-Change
Overview
Forest-Change is the first benchmark dataset specifically designed for joint forest change detection and captioning in remote sensing imagery. It provides bi-temporal satellite images, pixel-level deforestation masks, and multi-granularity semantic captions describing forest cover changes in tropical and subtropical regions.
Dataset Details
- Total Examples: 334 annotated bi-temporal image pairs
- Spatial Resolution: ~30m/pixel (medium resolution)
- Original Image Size: 480×480 pixels (cropped from larger scenes)
- Processed Image Size: 256×256 pixels (resized for model training)
- Temporal Resolution: 1 year between image pairs
- Geographic Focus: Tropical and subtropical deforestation fronts
Dataset Splits
- Training: 270 examples (~80%)
- Validation: 31 examples (~10%)
- Test: 33 examples (~10%)
Data Format
Each example contains:
- Image A: Pre-change RGB satellite image
- Image B: Post-change RGB satellite image
- Change Mask: Binary segmentation mask (0=no change, 1=deforestation)
- Captions: Five captions describing the forest change event with varied granularity
Data Sources
- Imagery Source: Google Earth Engine (GEE)
- Base Dataset: Derived from Hewarathna et al. (2024) forest ecosystem change detection dataset
- Validation: Forest cover changes verified through Global Forest Watch (GFW) platform
- Geographic Selection: Based on WWF 2015 Deforestation Fronts report
Caption Generation
Captions are generated through a hybrid two-stage approach:
- Human Annotation: One caption per example manually created by domain annotators describing observed changes
- Rule-Based Generation: Four additional captions automatically generated based on quantitative mask properties:
- Percentage of newly deforested area (binned into descriptive severity levels)
- Size and number of individual change patches
- Spatial distribution patterns of deforestation
- Variation in patch sizes
This approach ensures both semantic richness from human expertise and consistent structural variation across captions.
Key Characteristics
- Change Coverage:
- Mean: <5% deforestation per image
- Maximum: 40% deforestation
- Distribution: Heavily skewed toward lower deforestation percentages
- Caption Length: Bimodal distribution with both concise and detailed descriptions
- Change Patterns: Diverse deforestation manifestations including:
- Scattered small patches across forest areas
- Concentrated clearing zones
- Edge-of-clearing expansion patterns
- Highly variable patch sizes and configurations
- Caption Content: Descriptions emphasize:
- Degree/severity of forest loss
- Spatial location within the image
- Patch characteristics (size, number, distribution)
Preprocessing
- All images resized to 256×256 pixels for consistency
- Change masks binarized (0=no change, 1=change)
- Bi-temporal image pairs pre-aligned
- Per-channel normalization using dataset-specific mean and standard deviation statistics
- No atmospheric correction applied
- No cloud masking applied (some samples contain partial cloud occlusion)
Use Cases
- Forest change detection and monitoring
- Deforestation segmentation in natural environments
- Change captioning for ecological applications
- Multi-task learning for remote sensing
- Benchmarking vision-language models on forest imagery
- Training interactive forest analysis systems
- Developing automated forest monitoring workflows
Evaluation Metrics
Due to severe class imbalance (most pixels are no-change), evaluation requires:
- Per-class IoU: Separate metrics for change and no-change classes
- Mean IoU (mIoU): Average of both class IoUs
- Caption metrics: BLEU-n (n=1,2,3,4), METEOR, ROUGE-L, CIDEr-D
- Note: Overall accuracy is not recommended due to class imbalance
Limitations
- Dataset Scale: Limited to 334 examples, restricting model generalization
- Scene Diversity: Limited number of unique geographic sites due to cropping and augmentation strategy
- Class Imbalance: Severe imbalance with most pixels representing no-change, challenging for detection models
- Caption Quality: Majority of captions are rule-based, limiting linguistic variation and naturalness
- Geographic Grounding: Limited incorporation of geographic features and contextual information in captions
- Spatial Resolution: Medium resolution (~30m/pixel) limits detection of very small-scale changes
- Temporal Coverage: Fixed 1-year intervals between image pairs
- Atmospheric Effects: Some samples affected by partial cloud occlusion
- Edge Boundaries: Fuzzy boundaries at deforestation patch edges complicate precise segmentation
Citation
If you use this dataset, please cite:
@article{brock2024forestchat,
title={Forest-Chat: Adapting Vision-Language Agents for Interactive Forest Change Analysis},
author={Brock, James and Zhang, Ce and Anantrasirichai, Nantheera},
journal={Ecological Informatics},
year={2024}
}
@article{hewarathna2024change,
title={Change detection for forest ecosystems using remote sensing images with siamese attention u-net},
author={Hewarathna, AI and Hamlin, L and Charles, J and Vigneshwaran, P and George, R and Thuseethan, S and Wimalasooriya, C and Shanmugam, B},
journal={Technologies},
volume={12},
number={9},
pages={160},
year={2024}
}
License
MIT License - Academic re-use purpose only
Contact
For questions or issues regarding this dataset, please contact:
- James Brock: james.brock@bristol.ac.uk
- School of Computer Science, University of Bristol
- Downloads last month
- 20