Add metadata and link paper/code
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,31 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
# A Pragmatic VLA Foundation Model
|
|
|
|
| 2 |
<p align="center">
|
| 3 |
-
<img src="assets/Teaser.png" width="100%">
|
| 4 |
</p>
|
| 5 |
|
|
|
|
| 6 |
|
| 7 |
-
|
| 8 |
-
- **
|
| 9 |
-
|
| 10 |
-
- **Strong Performance**: Achieve clear superiority over competitors on simulation and real-world benchmarks.
|
| 11 |
-
- **Training Efficiency**: Represent a 1.5 ∼ 2.8× (depending on the relied VLM base model) speedup over existing VLA-oriented codebases.
|
| 12 |
|
| 13 |
-
|
| 14 |
|
| 15 |
-
|
| 16 |
-
-
|
| 17 |
-
-
|
| 18 |
-
|
|
|
|
| 19 |
|
| 20 |
## Related Models
|
| 21 |
|
| 22 |
-
| Model Name |
|
| 23 |
| :--- | :---: | :---: | :---: |
|
| 24 |
| LingBot-VLA-4B | [🤗 lingbot-vla-4b](https://huggingface.co/robbyant/lingbot-vla-4b) | [🤖 lingbot-vla-4b](https://modelscope.cn/models/Robbyant/lingbot-vla-4b) | LingBot-VLA *w/o* Depth|
|
| 25 |
| LingBot-VLA-4B-Depth | [🤗 lingbot-vla-4b-depth](https://huggingface.co/robbyant/lingbot-vla-4b-depth) | [🤖 lingbot-vla-4b-depth](https://modelscope.cn/models/Robbyant/lingbot-vla-4b-depth) | LingBot-VLA *w/* Depth |
|
| 26 |
|
| 27 |
-
|
| 28 |
-
|
| 29 |
---
|
| 30 |
|
| 31 |
## Citation
|
|
@@ -46,4 +50,4 @@ This project is licensed under the [Apache-2.0 License](LICENSE).
|
|
| 46 |
|
| 47 |
## Acknowledgement
|
| 48 |
|
| 49 |
-
This codebase is
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
pipeline_tag: robotics
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
# A Pragmatic VLA Foundation Model
|
| 7 |
+
|
| 8 |
<p align="center">
|
| 9 |
+
<img src="https://huggingface.co/robbyant/lingbot-vla-4b/resolve/main/assets/Teaser.png" width="100%">
|
| 10 |
</p>
|
| 11 |
|
| 12 |
+
**LingBot-VLA** is a Vision-Language-Action (VLA) foundation model designed for robotic manipulation, emphasizing pragmatic deployment, efficiency, and strong generalization across tasks and platforms.
|
| 13 |
|
| 14 |
+
- **Paper:** [A Pragmatic VLA Foundation Model](https://huggingface.co/papers/2601.18692)
|
| 15 |
+
- **Repository:** [https://github.com/robbyant/lingbot-vla](https://github.com/robbyant/lingbot-vla)
|
| 16 |
+
- **Project Page:** [https://technology.robbyant.com/lingbot-vla](https://technology.robbyant.com/lingbot-vla)
|
|
|
|
|
|
|
| 17 |
|
| 18 |
+
## Highlights
|
| 19 |
|
| 20 |
+
- **Large-scale Pre-training Data**: Trained on 20,000 hours of real-world data from 9 popular dual-arm robot configurations.
|
| 21 |
+
- **Strong Performance**: Achieves clear superiority over competitors on simulation and real-world benchmarks (GM-100 and RoboTwin 2.0).
|
| 22 |
+
- **Training Efficiency**: Offers a 1.5 ~ 2.8× speedup over existing VLA-oriented codebases, ensuring it is well-suited for real-world deployment.
|
| 23 |
+
|
| 24 |
+
---
|
| 25 |
|
| 26 |
## Related Models
|
| 27 |
|
| 28 |
+
| Model Name | Hugging Face | ModelScope | Description |
|
| 29 |
| :--- | :---: | :---: | :---: |
|
| 30 |
| LingBot-VLA-4B | [🤗 lingbot-vla-4b](https://huggingface.co/robbyant/lingbot-vla-4b) | [🤖 lingbot-vla-4b](https://modelscope.cn/models/Robbyant/lingbot-vla-4b) | LingBot-VLA *w/o* Depth|
|
| 31 |
| LingBot-VLA-4B-Depth | [🤗 lingbot-vla-4b-depth](https://huggingface.co/robbyant/lingbot-vla-4b-depth) | [🤖 lingbot-vla-4b-depth](https://modelscope.cn/models/Robbyant/lingbot-vla-4b-depth) | LingBot-VLA *w/* Depth |
|
| 32 |
|
|
|
|
|
|
|
| 33 |
---
|
| 34 |
|
| 35 |
## Citation
|
|
|
|
| 50 |
|
| 51 |
## Acknowledgement
|
| 52 |
|
| 53 |
+
This codebase is built on the [VeOmni](https://arxiv.org/abs/2508.02317) and [LeRobot](https://github.com/huggingface/lerobot) projects. We thank the authors for their excellent work!
|