Add metadata and link paper/code

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +19 -15
README.md CHANGED
@@ -1,31 +1,35 @@
 
 
 
 
 
1
  # A Pragmatic VLA Foundation Model
 
2
  <p align="center">
3
- <img src="assets/Teaser.png" width="100%">
4
  </p>
5
 
 
6
 
7
- **LingBot-VLA** has focused on **Pragmatic**:
8
- - **Large-scale Pre-training Data**: 20,000 hours of real-world
9
- data from 9 popular dual-arm robot configurations.
10
- - **Strong Performance**: Achieve clear superiority over competitors on simulation and real-world benchmarks.
11
- - **Training Efficiency**: Represent a 1.5 ∼ 2.8× (depending on the relied VLM base model) speedup over existing VLA-oriented codebases.
12
 
13
- ---
14
 
15
- ## Model Sources
16
- - Repository: https://github.com/robbyant/lingbot-vla
17
- - Paper: A Pragmatic VLA Foundation Model
18
- - Project Page: https://technology.robbyant.com/lingbot-vla
 
19
 
20
  ## Related Models
21
 
22
- | Model Name | Huggingface | ModelScope | Description |
23
  | :--- | :---: | :---: | :---: |
24
  | LingBot-VLA-4B &nbsp; | [🤗 lingbot-vla-4b](https://huggingface.co/robbyant/lingbot-vla-4b) | [🤖 lingbot-vla-4b](https://modelscope.cn/models/Robbyant/lingbot-vla-4b) | LingBot-VLA *w/o* Depth|
25
  | LingBot-VLA-4B-Depth | [🤗 lingbot-vla-4b-depth](https://huggingface.co/robbyant/lingbot-vla-4b-depth) | [🤖 lingbot-vla-4b-depth](https://modelscope.cn/models/Robbyant/lingbot-vla-4b-depth) | LingBot-VLA *w/* Depth |
26
 
27
-
28
-
29
  ---
30
 
31
  ## Citation
@@ -46,4 +50,4 @@ This project is licensed under the [Apache-2.0 License](LICENSE).
46
 
47
  ## Acknowledgement
48
 
49
- This codebase is builded on the [VeOmni](https://arxiv.org/abs/2508.02317) project. Thanks for their excellent work!
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: robotics
4
+ ---
5
+
6
  # A Pragmatic VLA Foundation Model
7
+
8
  <p align="center">
9
+ <img src="https://huggingface.co/robbyant/lingbot-vla-4b/resolve/main/assets/Teaser.png" width="100%">
10
  </p>
11
 
12
+ **LingBot-VLA** is a Vision-Language-Action (VLA) foundation model designed for robotic manipulation, emphasizing pragmatic deployment, efficiency, and strong generalization across tasks and platforms.
13
 
14
+ - **Paper:** [A Pragmatic VLA Foundation Model](https://huggingface.co/papers/2601.18692)
15
+ - **Repository:** [https://github.com/robbyant/lingbot-vla](https://github.com/robbyant/lingbot-vla)
16
+ - **Project Page:** [https://technology.robbyant.com/lingbot-vla](https://technology.robbyant.com/lingbot-vla)
 
 
17
 
18
+ ## Highlights
19
 
20
+ - **Large-scale Pre-training Data**: Trained on 20,000 hours of real-world data from 9 popular dual-arm robot configurations.
21
+ - **Strong Performance**: Achieves clear superiority over competitors on simulation and real-world benchmarks (GM-100 and RoboTwin 2.0).
22
+ - **Training Efficiency**: Offers a 1.5 ~ 2.8× speedup over existing VLA-oriented codebases, ensuring it is well-suited for real-world deployment.
23
+
24
+ ---
25
 
26
  ## Related Models
27
 
28
+ | Model Name | Hugging Face | ModelScope | Description |
29
  | :--- | :---: | :---: | :---: |
30
  | LingBot-VLA-4B &nbsp; | [🤗 lingbot-vla-4b](https://huggingface.co/robbyant/lingbot-vla-4b) | [🤖 lingbot-vla-4b](https://modelscope.cn/models/Robbyant/lingbot-vla-4b) | LingBot-VLA *w/o* Depth|
31
  | LingBot-VLA-4B-Depth | [🤗 lingbot-vla-4b-depth](https://huggingface.co/robbyant/lingbot-vla-4b-depth) | [🤖 lingbot-vla-4b-depth](https://modelscope.cn/models/Robbyant/lingbot-vla-4b-depth) | LingBot-VLA *w/* Depth |
32
 
 
 
33
  ---
34
 
35
  ## Citation
 
50
 
51
  ## Acknowledgement
52
 
53
+ This codebase is built on the [VeOmni](https://arxiv.org/abs/2508.02317) and [LeRobot](https://github.com/huggingface/lerobot) projects. We thank the authors for their excellent work!