PP2-M / README.md
nielsr's picture
nielsr HF Staff
Improve dataset card: Add paper/code links, sample usage, and update metadata/citation
39cf597 verified
|
raw
history blame
6.58 kB
metadata
language:
  - en
license: other
size_categories:
  - 100K<n<1M
pretty_name: Place Pulse 2.0 Multimodal
tags:
  - GeoFM
  - PlacePulse
  - SpatialRepresentationLearning
  - Multimodal
  - OpenStreetMap
  - StreetView
  - geospatial
task_categories:
  - other

PP2-M: Place Pulse 2.0 - Multimodal

Paper: UrbanFusion: Stochastic Multimodal Fusion for Contrastive Learning of Robust Spatial Representations Code: https://github.com/DominikM198/UrbanFusion

PP2-M (Place Pulse 2.0 - Multimodal) is a dataset based on the original Place Pulse 2.0 dataset [1], enriched with additional geospatial modalities for training multimodal Geo-Foundation Models (GeoFM).

The dataset includes aligned pairs of the following modalities:

  • 🌍 Geographical coordinates (lat, lon) from Place Pulse 2.0 [1]
  • πŸ™ Street view images from Place Pulse 2.0 [1]
  • πŸ›° Remote sensing images from Sentinel-2 [2]
  • πŸ—Ί Cartographic basemaps from OpenStreetMap [3]
  • πŸ“ Points of interest (POIs) from OpenStreetMap [3]

πŸš€ Sample Usage

Using pretrained models for location encoding is straightforward. The example below demonstrates how to load the model and generate representations based solely on geographic coordinates (latitude and longitude), without requiring any additional input modalities.

import torch
from huggingface_hub import hf_hub_download
from srl.multi_modal_encoder.load import get_urbanfusion

device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

# Coordinates: batch of 32 (lat, lon) pairs
coords = torch.randn(32, 2).to(device)

# Placeholders for other modalities (SV, RS, OSM, POI)
placeholder = torch.empty(32).to(device)
inputs = [coords, placeholder, placeholder, placeholder, placeholder]

# Mask all but coordinates (indices: 0=coords, 1=SV, 2=RS, 3=OSM, 4=POI)
mask_indices = [1, 2, 3, 4]

# Load pretrained UrbanFusion model
ckpt = hf_hub_download("DominikM198/UrbanFusion", "UrbanFusion/UrbanFusion.ckpt")
model = get_urbanfusion(ckpt, device=device).eval()

# Encode inputs (output shape: [32, 768])
with torch.no_grad():
    embeddings = model(inputs, mask_indices=mask_indices, return_representations=True).cpu()

For a more comprehensive guideβ€”including instructions on applying the model to downstream tasks and incorporating additional modalities (with options for downloading, preprocessing, and using contextual prompts with or without precomputed features)β€”see the following tutorials:


πŸ“œ License

Due to its multimodality, PP2-M comes with different licenses per modality, as described in the folder LICENSES.


πŸ“‘ Modalities Description

πŸ“Œ Coordinates

  • 110,988 locations, each with associated geographic coordinates.

πŸ™ Street View Images (SVI)

  • Obtained from **Google Street View.
  • Resolution: 400 Γ— 300 pixels.

πŸ›° Remote Sensing Images (Sentinel-2)

  • Sentinel-2 Level-2A images.
  • Acquisition period: Jan 1 – Dec 31, 2024.
  • Filtered for minimal cloud coverage.
  • Each patch includes spectral bands: B01, B02, B03, B04, B05, B06, B07, B08, B08A, B09, B11, B12
  • Resolution: 256 Γ— 256 pixels.

πŸ—Ί Cartographic Basemaps (OSM_basemaps)

  • Tiles from OpenStreetMap tile server.
  • Zoom levels: 15, 16, 17 β†’ resolutions of 1200 m, 600 m, 300 m.
  • Downloaded: May 2025.
  • Rendered at 256 Γ— 256 pixels.

πŸ“ Points of Interest (OSM_pois)

  • Extracted from OpenStreetMap.
  • For each location: up to 15 nearest POIs within 200 m.
  • Adaptive search radius ensures coverage in sparse areas.
  • Retained POIs with tags: amenity, shop, leisure, tourism, healthcare, theatre, cinema, building=religious, building=transportation, public_transport=station
  • Excluded: parking, parking_space, bench, bicycle_parking, motorcycle_parking, post_box, toilets
  • Each POI is assigned a representative category (priority order: amenity β†’ leisure β†’ religion β†’ public_transport β†’ shop β†’ tourism).
  • Special cases:
    • healthcare if substring matches
    • museum if name contains "museum"
  • Final POIs are used to construct textual prompts describing each POI’s name, category, and distance.

πŸ“‚ Folder Structure

PP2-M/
β”‚
β”œβ”€β”€ LICENSES/ β†’ Licenses for all modalities
β”œβ”€β”€ Tables_statistics/ β†’ Statistics & tables (based on Place Pulse 2.0)
β”œβ”€β”€ SVI/ β†’ Street View Images
β”œβ”€β”€ sentinel2/ β†’ Sentinel-2 images
β”œβ”€β”€ OSM_basemaps/ β†’ OSM basemaps (zoom 15, 16, 17)
β”œβ”€β”€ OSM_pois/ β†’ Raw POIs + generated text prompts
└── Precomputed_features/ β†’ Pre-extracted modality-specific features

πŸ”€ Dataset Splits

  • training – samples used for training.
  • validation_in_region – interpolation evaluation.
  • validation_out_region – extrapolation evaluation (unseen cities).

πŸ“Š Precomputed Features

In addition to raw data, we provide pre-extracted features from each modality using modality-specific models. See details in our paper: UrbanFusion.


πŸ“– Citation

If you use PP2-M, please cite our work:

@article{muehlematter2025urbanfusion,
  title   = {UrbanFusion: Stochastic Multimodal Fusion for Contrastive Learning of Robust Spatial Representations},
  author  = {Dominik J. MΓΌhlematter and Lin Che and Ye Hong and Martin Raubal and Nina Wiedemann},
  year    = {2025},
  journal = {arXiv preprint arXiv:2510.13774},
  url     = {https://huggingface.co/papers/2510.13774}
}

πŸ“Š References

[1] Dubey, A., Naik, N., Parikh, D., Raskar, R., and Hidalgo, C. A. (2016). Deep learning the city: Quantifying urban perception at a global scale. In ECCV, pp. 196–212.
[2] Drusch, M., Del Bello, U., Carlier, S., Colin, O., Fernandez, V., Gascon, F., ... Bargellini, P. (2012). Sentinel-2: ESA’s optical high-resolution mission for GMES operational services. Remote Sensing of Environment, 120:25–36.
[3] OpenStreetMap contributors (2017). Planet dump retrieved from https://planet.osm.org