language:
- en
license: other
size_categories:
- 100K<n<1M
pretty_name: Place Pulse 2.0 Multimodal
tags:
- GeoFM
- PlacePulse
- SpatialRepresentationLearning
- Multimodal
- OpenStreetMap
- StreetView
- geospatial
task_categories:
- other
PP2-M: Place Pulse 2.0 - Multimodal
Paper: UrbanFusion: Stochastic Multimodal Fusion for Contrastive Learning of Robust Spatial Representations Code: https://github.com/DominikM198/UrbanFusion
PP2-M (Place Pulse 2.0 - Multimodal) is a dataset based on the original Place Pulse 2.0 dataset [1], enriched with additional geospatial modalities for training multimodal Geo-Foundation Models (GeoFM).
The dataset includes aligned pairs of the following modalities:
- π Geographical coordinates (lat, lon) from Place Pulse 2.0 [1]
- π Street view images from Place Pulse 2.0 [1]
- π° Remote sensing images from Sentinel-2 [2]
- πΊ Cartographic basemaps from OpenStreetMap [3]
- π Points of interest (POIs) from OpenStreetMap [3]
π Sample Usage
Using pretrained models for location encoding is straightforward. The example below demonstrates how to load the model and generate representations based solely on geographic coordinates (latitude and longitude), without requiring any additional input modalities.
import torch
from huggingface_hub import hf_hub_download
from srl.multi_modal_encoder.load import get_urbanfusion
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
# Coordinates: batch of 32 (lat, lon) pairs
coords = torch.randn(32, 2).to(device)
# Placeholders for other modalities (SV, RS, OSM, POI)
placeholder = torch.empty(32).to(device)
inputs = [coords, placeholder, placeholder, placeholder, placeholder]
# Mask all but coordinates (indices: 0=coords, 1=SV, 2=RS, 3=OSM, 4=POI)
mask_indices = [1, 2, 3, 4]
# Load pretrained UrbanFusion model
ckpt = hf_hub_download("DominikM198/UrbanFusion", "UrbanFusion/UrbanFusion.ckpt")
model = get_urbanfusion(ckpt, device=device).eval()
# Encode inputs (output shape: [32, 768])
with torch.no_grad():
embeddings = model(inputs, mask_indices=mask_indices, return_representations=True).cpu()
For a more comprehensive guideβincluding instructions on applying the model to downstream tasks and incorporating additional modalities (with options for downloading, preprocessing, and using contextual prompts with or without precomputed features)βsee the following tutorials:
π License
Due to its multimodality, PP2-M comes with different licenses per modality, as described in the folder LICENSES.
π Modalities Description
π Coordinates
- 110,988 locations, each with associated geographic coordinates.
π Street View Images (SVI)
- Obtained from **Google Street View.
- Resolution: 400 Γ 300 pixels.
π° Remote Sensing Images (Sentinel-2)
- Sentinel-2 Level-2A images.
- Acquisition period: Jan 1 β Dec 31, 2024.
- Filtered for minimal cloud coverage.
- Each patch includes spectral bands:
B01, B02, B03, B04, B05, B06, B07, B08, B08A, B09, B11, B12 - Resolution: 256 Γ 256 pixels.
πΊ Cartographic Basemaps (OSM_basemaps)
- Tiles from OpenStreetMap tile server.
- Zoom levels: 15, 16, 17 β resolutions of 1200 m, 600 m, 300 m.
- Downloaded: May 2025.
- Rendered at 256 Γ 256 pixels.
π Points of Interest (OSM_pois)
- Extracted from OpenStreetMap.
- For each location: up to 15 nearest POIs within 200 m.
- Adaptive search radius ensures coverage in sparse areas.
- Retained POIs with tags:
amenity, shop, leisure, tourism, healthcare, theatre, cinema, building=religious, building=transportation, public_transport=station - Excluded:
parking, parking_space, bench, bicycle_parking, motorcycle_parking, post_box, toilets - Each POI is assigned a representative category (priority order:
amenity β leisure β religion β public_transport β shop β tourism). - Special cases:
healthcareif substring matchesmuseumif name contains "museum"
- Final POIs are used to construct textual prompts describing each POIβs name, category, and distance.
π Folder Structure
PP2-M/
β
βββ LICENSES/ β Licenses for all modalities
βββ Tables_statistics/ β Statistics & tables (based on Place Pulse 2.0)
βββ SVI/ β Street View Images
βββ sentinel2/ β Sentinel-2 images
βββ OSM_basemaps/ β OSM basemaps (zoom 15, 16, 17)
βββ OSM_pois/ β Raw POIs + generated text prompts
βββ Precomputed_features/ β Pre-extracted modality-specific features
π Dataset Splits
- training β samples used for training.
- validation_in_region β interpolation evaluation.
- validation_out_region β extrapolation evaluation (unseen cities).
π Precomputed Features
In addition to raw data, we provide pre-extracted features from each modality using modality-specific models. See details in our paper: UrbanFusion.
π Citation
If you use PP2-M, please cite our work:
@article{muehlematter2025urbanfusion,
title = {UrbanFusion: Stochastic Multimodal Fusion for Contrastive Learning of Robust Spatial Representations},
author = {Dominik J. MΓΌhlematter and Lin Che and Ye Hong and Martin Raubal and Nina Wiedemann},
year = {2025},
journal = {arXiv preprint arXiv:2510.13774},
url = {https://huggingface.co/papers/2510.13774}
}
π References
[1] Dubey, A., Naik, N., Parikh, D., Raskar, R., and Hidalgo, C. A. (2016). Deep learning the city: Quantifying urban perception at a global scale. In ECCV, pp. 196β212.
[2] Drusch, M., Del Bello, U., Carlier, S., Colin, O., Fernandez, V., Gascon, F., ... Bargellini, P. (2012). Sentinel-2: ESAβs optical high-resolution mission for GMES operational services. Remote Sensing of Environment, 120:25β36.
[3] OpenStreetMap contributors (2017). Planet dump retrieved from https://planet.osm.org