|
|
--- |
|
|
pipeline_tag: robotics |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- reinforcement-learning |
|
|
- robotic-manipulation |
|
|
- action-chunking |
|
|
--- |
|
|
|
|
|
# This repository provided the torch-version openpi checkpoints (pi0base, pi0.5base) coverted from JAX-version! |
|
|
|
|
|
# [Mixture of Horizons in Action Chunking](https://huggingface.co/papers/2511.19433) |
|
|
|
|
|
Vision-language-action (VLA) models for robotic manipulation are highly sensitive to the chosen **action chunk length**, termed **horizon** in this work. A fixed horizon presents an inherent trade-off: longer horizons offer superior global foresight but compromise fine-grained accuracy, while shorter ones provide precise local control but struggle with long-term tasks. |
|
|
|
|
|
To address this challenge, we propose **Mixture of Horizons (MoH)**, a novel, plug-and-play strategy that fuses multiple horizons within a single policy. MoH processes action chunks in parallel segments with different horizons and integrates their outputs. This approach simultaneously leverages long-term foresight and short-term precision with minimal overhead, and enables **Dynamic Inference** through cross-horizon consensus for enhanced efficiency and robustness in complex robotic tasks. |
|
|
|
|
|
- π [Paper](https://huggingface.co/papers/2511.19433) |
|
|
- π [Project Page](https://timsty1.github.io/moh/) |
|
|
- π» [Code](https://github.com/Timsty1/MixtureOfHorizons/tree/main) |
|
|
|
|
|
## Introduction |
|
|
|
|
|
<div align="center"> |
|
|
<table border="0" cellspacing="0" cellpadding="0"> |
|
|
<tr> |
|
|
<td align="center" width="50%"> |
|
|
<img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/study_of_horizons_pi0.png" alt="Trade-off Effect" width="100%"> |
|
|
</td> |
|
|
<td align="center" width="50%"> |
|
|
<img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/intro_motivation_v2.png" alt="Mixture of Horizons" width="100%"> |
|
|
</td> |
|
|
</tr> |
|
|
<tr> |
|
|
<td align="center" valign="top"> |
|
|
Figure 1: Trade-off between long-term foresight and short-term precision induced by single horizon |
|
|
</td> |
|
|
<td align="center" valign="top"> |
|
|
Figure 2: Overview of the proposed mixture-of-horizons strategy |
|
|
</td> |
|
|
</tr> |
|
|
</table> |
|
|
</div> |
|
|
|
|
|
<br> |
|
|
|
|
|
* **Mitigates Trade-off**: Addresses the inherent trade-off between long-term foresight and short-term precision induced by single action chunk horizons. |
|
|
* **Plug-and-Play**: Easily integrates into existing full-attention action modules with minimal training or inference overhead. |
|
|
* **Dynamic Inference**: Achieves higher efficiency and robustness by selecting stable actions through cross-horizon consensus. |
|
|
|
|
|
#### More results on LIBERO |
|
|
<div align="center"> |
|
|
<img src="https://github.com/Timsty1/MixtureOfHorizons/raw/main/figure/libero_main.jpg" width="90%" /> |
|
|
</div> |
|
|
|
|
|
## Usage |
|
|
|
|
|
For detailed instructions on environment setup, training, and evaluation, please refer to the [GitHub repository](https://github.com/Timsty1/MixtureOfHorizons/tree/main). |
|
|
|
|
|
## β€οΈ Acknowledgment |
|
|
|
|
|
We express our gratitude to [OpenPi](https://github.com/Physical-Intelligence/openpi/tree/main), [LIBERO](https://github.com/Lifelong-Robot-Learning/LIBERO), and [RoboTwin](https://robotwin-platform.github.io/) for their open-source contributions. |
|
|
|
|
|
## π Citation |
|
|
|
|
|
If you feel that this paper, models, or codes are helpful, please cite our paper, thanks for your support! |
|
|
|
|
|
```bibtex |
|
|
@article{jing2025mixture_of_horizons, |
|
|
title={Mixture of Horizons in Action Chunking}, |
|
|
author={Jing, Dong and Wang, Gang and Liu, Jiaqi and Tang, Weiliang and Sun, Zelong and Yao, Yunchao and Wei, Zhenyu and Liu, Yunhui and Lu, Zhiwu and Ding, Mingyu}, |
|
|
journal={arXiv preprint arXiv:2511.19433}, |
|
|
year={2025} |
|
|
} |
|
|
``` |