TerraMesh: A Planetary Mosaic of Multimodal Earth Observation Data
By: Benedikt Blumenstiel , Paolo Fraccaro , Valerio Marsocci and more
Potential Business Impact:
Maps Earth better using many kinds of pictures.
Large-scale foundation models in Earth Observation can learn versatile, label-efficient representations by leveraging massive amounts of unlabeled data. However, existing public datasets are often limited in scale, geographic coverage, or sensor variety. We introduce TerraMesh, a new globally diverse, multimodal dataset combining optical, synthetic aperture radar, elevation, and land-cover modalities in an Analysis-Ready Data format. TerraMesh includes over 9~million samples with eight spatiotemporal aligned modalities, enabling large-scale pre-training. We provide detailed data processing steps, comprehensive statistics, and empirical evidence demonstrating improved model performance when pre-trained on TerraMesh. The dataset is hosted at https://huggingface.co/datasets/ibm-esa-geospatial/TerraMesh.
Similar Papers
TerraMind: Large-Scale Generative Multimodality for Earth Observation
CV and Pattern Recognition
Helps computers understand Earth from any picture.
TerraFM: A Scalable Foundation Model for Unified Multisensor Earth Observation
CV and Pattern Recognition
Helps satellites understand Earth better from space.
The Transparent Earth: A Multimodal Foundation Model for the Earth's Subsurface
Machine Learning (CS)
Maps Earth's hidden layers from many clues.