Update README.md
Browse files
README.md
CHANGED
|
@@ -1,23 +1,23 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
base_model:
|
| 4 |
-
- google/paligemma2-3b-pt-224
|
| 5 |
-
tags:
|
| 6 |
-
- VLA
|
| 7 |
-
- Foundation Vision-language-action Model
|
| 8 |
-
- Generalist Robot Policy
|
| 9 |
-
- robotics
|
| 10 |
-
language:
|
| 11 |
-
- en
|
| 12 |
-
pipeline_tag: image-text-to-text
|
| 13 |
-
library_name: transformers
|
| 14 |
-
---
|
| 15 |
|
| 16 |
# SpatialVLA
|
| 17 |
|
| 18 |
SpatialVLA is a spatial-enhanced vision-language-action model trained on 1.1 Million real robot episodes. The code is purely huggingFace-based and concise, with efficient performance.
|
| 19 |
|
| 20 |
-
All SpatialVLA checkpoints, as well as our [training codebase](https://github.com/
|
| 21 |
|
| 22 |
For full details, please read [our paper](https://arxiv.org/abs/2501.15830) and see [our project page](https://spatialvla.github.io/).
|
| 23 |
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
base_model:
|
| 4 |
+
- google/paligemma2-3b-pt-224
|
| 5 |
+
tags:
|
| 6 |
+
- VLA
|
| 7 |
+
- Foundation Vision-language-action Model
|
| 8 |
+
- Generalist Robot Policy
|
| 9 |
+
- robotics
|
| 10 |
+
language:
|
| 11 |
+
- en
|
| 12 |
+
pipeline_tag: image-text-to-text
|
| 13 |
+
library_name: transformers
|
| 14 |
+
---
|
| 15 |
|
| 16 |
# SpatialVLA
|
| 17 |
|
| 18 |
SpatialVLA is a spatial-enhanced vision-language-action model trained on 1.1 Million real robot episodes. The code is purely huggingFace-based and concise, with efficient performance.
|
| 19 |
|
| 20 |
+
All SpatialVLA checkpoints, as well as our [training codebase](https://github.com/SpatialVLA/SpatialVLA) are released under an MIT License.
|
| 21 |
|
| 22 |
For full details, please read [our paper](https://arxiv.org/abs/2501.15830) and see [our project page](https://spatialvla.github.io/).
|
| 23 |
|