Tavish9 commited on
Commit
cf960b5
·
verified ·
1 Parent(s): 01f5ea2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -1,23 +1,23 @@
1
- ---
2
- license: mit
3
- base_model:
4
- - google/paligemma2-3b-pt-224
5
- tags:
6
- - VLA
7
- - Foundation Vision-language-action Model
8
- - Generalist Robot Policy
9
- - robotics
10
- language:
11
- - en
12
- pipeline_tag: image-text-to-text
13
- library_name: transformers
14
- ---
15
 
16
  # SpatialVLA
17
 
18
  SpatialVLA is a spatial-enhanced vision-language-action model trained on 1.1 Million real robot episodes. The code is purely huggingFace-based and concise, with efficient performance.
19
 
20
- All SpatialVLA checkpoints, as well as our [training codebase](https://github.com/openvla/openvla) are released under an MIT License.
21
 
22
  For full details, please read [our paper](https://arxiv.org/abs/2501.15830) and see [our project page](https://spatialvla.github.io/).
23
 
 
1
+ ---
2
+ license: mit
3
+ base_model:
4
+ - google/paligemma2-3b-pt-224
5
+ tags:
6
+ - VLA
7
+ - Foundation Vision-language-action Model
8
+ - Generalist Robot Policy
9
+ - robotics
10
+ language:
11
+ - en
12
+ pipeline_tag: image-text-to-text
13
+ library_name: transformers
14
+ ---
15
 
16
  # SpatialVLA
17
 
18
  SpatialVLA is a spatial-enhanced vision-language-action model trained on 1.1 Million real robot episodes. The code is purely huggingFace-based and concise, with efficient performance.
19
 
20
+ All SpatialVLA checkpoints, as well as our [training codebase](https://github.com/SpatialVLA/SpatialVLA) are released under an MIT License.
21
 
22
  For full details, please read [our paper](https://arxiv.org/abs/2501.15830) and see [our project page](https://spatialvla.github.io/).
23