bambooboom commited on
Commit
4cb2aab
·
verified ·
1 Parent(s): 0beff28

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -3
README.md CHANGED
@@ -1,3 +1,50 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # XR-1-Stage2
3
+
4
+ [[Project Page](https://github.com/Open-X-Humanoid/XR-1)] [[Paper](https://www.google.com/search?q=https://arxiv.org/abs/2411.02776v1)] [[GitHub](https://github.com/Open-X-Humanoid/XR-1)]
5
+
6
+ This repository contains the **Stage 2** weights for the **XR-1 (X Robotic Model 1)** project. This is the large-scale **Vision-Language-Action (VLA)** policy trained on top of the Unified Vision-Motion Codes (UVMC).
7
+
8
+ ## 🤖 Model Description
9
+
10
+ **XR-1-Stage2** is a generative Transformer-based policy. While Stage 1 focused on tokenizing sensory and action data, Stage 2 performs the high-level reasoning required to map natural language instructions and visual observations into executable robotic actions.
11
+
12
+ ### Key Features
13
+
14
+ * **VLA Architecture**: Integrates vision, language, and action into a unified Transformer framework.
15
+ * **Token-Based Policy**: Operates on the discrete latent space defined by the [XR-1-Stage1-UVMC](https://huggingface.co/X-Humanoid/XR-1-Stage1-UVMC) tokenizer.
16
+ * **End-to-End Control**: Predicts the next set of action tokens based on the current visual state and text prompt.
17
+
18
+ ## 🧱 Model Hierarchy
19
+
20
+ 1. **Stage 1 (UVMC)**: Encodes images and actions into discrete tokens.
21
+ 2. **Stage 2 (VLA)**: **[This Model]** Predicts action tokens given instruction and image tokens.
22
+
23
+ ## 🛠 Usage
24
+
25
+ To deploy this policy, you need both the Stage 1 tokenizer and these Stage 2 weights. Please refer to the [XR-1 GitHub Repository](https://github.com/Open-X-Humanoid/XR-1) for inference scripts and environment setup.
26
+
27
+
28
+ ## 📝 Citation
29
+
30
+ If you use the XR-1 model in your research, please cite our work:
31
+
32
+ ```bibtex
33
+ @article{fan2025xr,
34
+ title={XR-1: Towards Versatile Vision-Language-Action Models via Learning Unified Vision-Motion Representations},
35
+ author={Fan, Shichao and Wu, Kun and Che, Zhengping and Wang, Xinhua and Wu, Di and Liao, Fei and Liu, Ning and Zhang, Yixue and Zhao, Zhen and Xu, Zhiyuan and others},
36
+ journal={arXiv preprint arXiv:2411.02776},
37
+ year={2025}
38
+ }
39
+
40
+ ```
41
+
42
+ ## 📜 License
43
+
44
+ This project is licensed under the [MIT License](https://github.com/Open-X-Humanoid/XR-1/blob/main/LICENSE).
45
+
46
+ ---
47
+
48
+ **Contact**: For questions, please open an issue on our [GitHub](https://github.com/Open-X-Humanoid/XR-1) or contact us at opensource@x-humanoid.com.
49
+
50
+