Add robotics metadata and link to paper/code

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +45 -3
README.md CHANGED
@@ -1,10 +1,52 @@
1
  ---
2
  datasets:
3
  - behavior-1k/2025-challenge-demos
 
 
 
 
 
 
4
  ---
5
 
6
- This is the model weights of Team Comet for the [2025 BEHAVIOR Challenge](https://behavior.stanford.edu/index.html). Our [submission](https://behavior.stanford.edu/challenge/leaderboard.html#privileged-information-track) achieved a Q-score of 0.2514, securing 2nd place overall and finishing just behind the winning team by a narrow margin—highlighting both the strong competitiveness of our approach and the effectiveness of our end-to-end VLA training strategy.
7
 
8
- Please refer to [this repository](https://github.com/mli0603/openpi-comet) for detailed usage. It provides a unified framework for pre-training, post-training, data generation and evaluation of π0.5 (Pi05) models on BEHAVIOR-1K.
9
 
10
- The [tech report](https://arxiv.org/abs/2512.10071) is available now.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  datasets:
3
  - behavior-1k/2025-challenge-demos
4
+ pipeline_tag: robotics
5
+ license: apache-2.0
6
+ tags:
7
+ - embodied-ai
8
+ - vla
9
+ - behavior-1k
10
  ---
11
 
12
+ # Openpi Comet: Competition Solution for 2025 BEHAVIOR Challenge
13
 
14
+ This repository contains the model weights of Team Comet for the [2025 BEHAVIOR Challenge](https://behavior.stanford.edu/index.html). Our [submission](https://behavior.stanford.edu/challenge/leaderboard.html#privileged-information-track) achieved a Q-score of 0.2514, securing 2nd place overall and finishing just behind the winning team by a narrow margin—highlighting both the strong competitiveness of our approach and the effectiveness of our end-to-end Vision-Language-Action (VLA) training strategy.
15
 
16
+ - **Paper:** [Openpi Comet: Competition Solution For 2025 BEHAVIOR Challenge](https://arxiv.org/abs/2512.10071)
17
+ - **Repository:** [https://github.com/mli0603/openpi-comet](https://github.com/mli0603/openpi-comet)
18
+ - **Project Page:** [BEHAVIOR Challenge](https://behavior.stanford.edu/index.html)
19
+
20
+ ## Model Description
21
+
22
+ OpenPi Comet is a foundation model designed for long-horizon mobile manipulation tasks in realistic household settings, specifically developed for the BEHAVIOR-1K dataset. Building on $\pi_{0.5}$ (Pi05), the model incorporates hierarchical instructions (global, subtask, skill) and multimodal observations (RGB, depth, point cloud, etc.). Through systematic training techniques and data scaling, including Rejection Sampling Fine-Tuning (RFT), the team achieved a validation Q-score of 0.345.
23
+
24
+ ## Usage
25
+
26
+ Please refer to [the official GitHub repository](https://github.com/mli0603/openpi-comet) for detailed installation and usage instructions. The codebase provides a unified framework for:
27
+ - Distributed pre-training and fine-tuning of OpenPi models.
28
+ - Data generation (teleoperation and simulation rollouts).
29
+ - Post-training via Rejection Sampling Fine-Tuning (RFT).
30
+ - Evaluation within the BEHAVIOR-1K simulator.
31
+
32
+ ### Quick Setup
33
+ ```bash
34
+ git clone https://github.com/mli0603/openpi-comet.git
35
+ cd openpi-comet
36
+ uv sync
37
+ uv pip install -e .
38
+ ```
39
+
40
+ ## Citation
41
+
42
+ If you find this work useful, please consider citing:
43
+
44
+ ```bibtex
45
+ @article{bai2025openpicometcompetitionsolution,
46
+ title={Openpi Comet: Competition Solution For 2025 BEHAVIOR Challenge},
47
+ author={Junjie Bai and Yu-Wei Chao and Qizhi Chen and Jinwei Gu and Moo Jin Kim and Zhaoshuo Li and Xuan Li and Tsung-Yi Lin and Ming-Yu Liu and Nic Ma and Kaichun Mo and Delin Qu and Shangkun Sun and Hongchi Xia and Fangyin Wei and Xiaohui Zeng},
48
+ journal={arXiv preprint arXiv:2512.10071},
49
+ year={2025},
50
+ url={https://arxiv.org/abs/2512.10071},
51
+ }
52
+ ```