File size: 1,636 Bytes
fa6d93e 51c1744 fa6d93e 51c1744 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 | ---
license: cc-by-4.0
task_categories:
- robotics
tags:
- robot
- ogbench
- rl
- imitation
- learning
- simulation
- manipulation
---
# OGBench Data for Latent Particle World Models (LPWM)
This repository contains pre-processed 64x64 frames for the `scene` and `cube` tasks from the [OGBench benchmark](https://github.com/seohongpark/ogbench). The dataset includes actions and frames used for training and evaluating **Latent Particle World Models (LPWM)**.
LPWM is a self-supervised object-centric world model that autonomously discovers keypoints, bounding boxes, and object masks directly from video data. It is designed to scale to real-world multi-object datasets and is applicable in decision-making tasks such as goal-conditioned imitation learning.
- **Paper:** [Latent Particle World Models: Self-supervised Object-centric Stochastic Dynamics Modeling](https://huggingface.co/papers/2603.04553)
- **Project Page:** [https://taldatech.github.io/lpwm-web](https://taldatech.github.io/lpwm-web)
- **GitHub Repository:** [https://github.com/taldatech/lpwm](https://github.com/taldatech/lpwm)
## Citation
If you use this data or the LPWM model in your research, please cite the following paper:
```bibtex
@inproceedings{
daniel2026latent,
title={Latent Particle World Models: Self-supervised Object-centric Stochastic Dynamics Modeling},
author={Tal Daniel and Carl Qi and Dan Haramati and Amir Zadeh and Chuan Li and Aviv Tamar and Deepak Pathak and David Held},
booktitle={The Fourteenth International Conference on Learning Representations},
year={2026},
url={https://openreview.net/forum?id=lTaPtGiUUc}
}
``` |