Improve model card: Add robotics pipeline tag, lerobot library, and tutorial links

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +31 -11
README.md CHANGED
@@ -1,21 +1,35 @@
 
 
 
 
 
 
1
  # Ο€β‚€ (Pi0)
2
 
3
  These weights directly come from the Pytorch conversion script of openpi and their `pi0_base` model.
4
 
5
  Ο€β‚€ is a **Vision-Language-Action model for general robot control**, from Physical Intelligence. The LeRobot implementation is adapted from their open source [OpenPI](https://github.com/Physical-Intelligence/openpi) repository.
6
 
 
 
 
 
 
 
 
 
7
  ## Model Overview
8
 
9
- Ο€β‚€ represents a breakthrough in robotics as the first general-purpose robot foundation model developed by [Physical Intelligence](https://www.physicalintelligence.company/blog/pi0). Unlike traditional robots that are narrow specialists programmed for repetitive motions, Ο€β‚€ is designed to be a generalist policy that can understand visual inputs, interpret natural language instructions, and control a variety of different robots across diverse tasks.
10
 
11
  ### Architecture and Approach
12
 
13
  Ο€β‚€ combines several key innovations:
14
 
15
- - **Flow Matching**: Uses a novel method to augment pre-trained VLMs with continuous action outputs via flow matching (a variant of diffusion models)
16
- - **Cross-Embodiment Training**: Trained on data from 8 distinct robot platforms including UR5e, Bimanual UR5e, Franka, Bimanual Trossen, Bimanual ARX, Mobile Trossen, and Mobile Fibocom
17
- - **Internet-Scale Pre-training**: Inherits semantic knowledge from a pre-trained 3B parameter Vision-Language Model
18
- - **High-Frequency Control**: Outputs motor commands at up to 50 Hz for real-time dexterous manipulation
19
 
20
  ## Training
21
 
@@ -40,7 +54,7 @@ python src/lerobot/scripts/train.py \
40
 
41
  ## Citation
42
 
43
- If you use this model, please cite the original OpenPI work:
44
 
45
  ```bibtex
46
  @article{openpi2024,
@@ -49,12 +63,18 @@ If you use this model, please cite the original OpenPI work:
49
  year={2024},
50
  url={https://github.com/Physical-Intelligence/openpi}
51
  }
52
- ```
53
 
54
- ## Original Repository
55
-
56
- [OpenPI GitHub Repository](https://github.com/Physical-Intelligence/openpi)
 
 
 
 
 
 
 
57
 
58
  ## License
59
 
60
- This model follows the same license as the original OpenPI repository.
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: robotics
4
+ library_name: lerobot
5
+ ---
6
+
7
  # Ο€β‚€ (Pi0)
8
 
9
  These weights directly come from the Pytorch conversion script of openpi and their `pi0_base` model.
10
 
11
  Ο€β‚€ is a **Vision-Language-Action model for general robot control**, from Physical Intelligence. The LeRobot implementation is adapted from their open source [OpenPI](https://github.com/Physical-Intelligence/openpi) repository.
12
 
13
+ ---
14
+ **Paper:** [Robot Learning: A Tutorial](https://huggingface.co/papers/2510.12403)
15
+ **Abstract:** Robot learning is at an inflection point, driven by rapid advancements in machine learning and the growing availability of large-scale robotics data. This shift from classical, model-based methods to data-driven, learning-based paradigms is unlocking unprecedented capabilities in autonomous systems. This tutorial navigates the landscape of modern robot learning, charting a course from the foundational principles of Reinforcement Learning and Behavioral Cloning to generalist, language-conditioned models capable of operating across diverse tasks and even robot embodiments. This work is intended as a guide for researchers and practitioners, and our goal is to equip the reader with the conceptual understanding and practical tools necessary to contribute to developments in robot learning, with ready-to-use examples implemented in `lerobot`.
16
+ **Project Page:** [https://huggingface.co/spaces/lerobot/robot-learning-tutorial](https://huggingface.co/spaces/lerobot/robot-learning-tutorial)
17
+ **Code for Tutorial:** [https://github.com/fracapuano/robot-learning-tutorial](https://github.com/fracapuano/robot-learning-tutorial)
18
+ **Original Repository (OpenPI):** [https://github.com/Physical-Intelligence/openpi](https://github.com/Physical-Intelligence/openpi)
19
+ ---
20
+
21
  ## Model Overview
22
 
23
+ Ο€β‚€ represents a breakthrough in robotics as the first general-purpose robot foundation model developed by [Physical Intelligence](https://www.physicalintelligence.company/blog/pi0). Unlike traditional robots that are narrow specialists programmed for repetitive motions, Ο€β‚€ is designed to be a generalist policy that can understand visual inputs, interpret natural language instructions, and control a variety of different robots across diverse tasks. This model is featured as an example in the "Robot Learning: A Tutorial" paper.
24
 
25
  ### Architecture and Approach
26
 
27
  Ο€β‚€ combines several key innovations:
28
 
29
+ - **Flow Matching**: Uses a novel method to augment pre-trained VLMs with continuous action outputs via flow matching (a variant of diffusion models)
30
+ - **Cross-Embodiment Training**: Trained on data from 8 distinct robot platforms including UR5e, Bimanual UR5e, Franka, Bimanual Trossen, Bimanual ARX, Mobile Trossen, and Mobile Fibocom
31
+ - **Internet-Scale Pre-training**: Inherits semantic knowledge from a pre-trained 3B parameter Vision-Language Model
32
+ - **High-Frequency Control**: Outputs motor commands at up to 50 Hz for real-time dexterous manipulation
33
 
34
  ## Training
35
 
 
54
 
55
  ## Citation
56
 
57
+ If you use this model, please cite the original OpenPI work and the tutorial paper:
58
 
59
  ```bibtex
60
  @article{openpi2024,
 
63
  year={2024},
64
  url={https://github.com/Physical-Intelligence/openpi}
65
  }
 
66
 
67
+ @misc{tutorial2025robotlearning,
68
+ title={Robot Learning: A Tutorial},
69
+ author={Francisco Cruz and Niels Rogge and Victor Dibia and Sasha Bozhkov and Thomas Wolf},
70
+ year={2025},
71
+ eprint={2510.12403},
72
+ archivePrefix={arXiv},
73
+ primaryClass={cs.RO},
74
+ url={https://arxiv.org/abs/2510.12403},
75
+ }
76
+ ```
77
 
78
  ## License
79
 
80
+ This model follows the same license as the original OpenPI repository, which is Apache 2.0.