X-VLA: Soft-Prompted Transformer as Scalable Cross-Embodiment Vision-Language-Action Model
Paper • 2510.10274 • Published • 16
observation.images.image dict | observation.images.image2 dict | observation.state listlengths 8 8 | action listlengths 7 7 | timestamp float64 0 45.4 | frame_index int64 0 454 | episode_index int64 0 146 | index int64 0 40k | task_index int64 0 9 |
|---|---|---|---|---|---|---|---|---|
{
"bytes": [
137,
80,
78,
71,
13,
10,
26,
10,
0,
0,
0,
13,
73,
72,
68,
82,
0,
0,
1,
0,
0,
0,
1,
0,
8,
2,
0,
0,
0,
211,
16,
63,
49,
0,
0,
254,
108,
73,
68,
65,... | {
"bytes": [
137,
80,
78,
71,
13,
10,
26,
10,
0,
0,
0,
13,
73,
72,
68,
82,
0,
0,
1,
0,
0,
0,
1,
0,
8,
2,
0,
0,
0,
211,
16,
63,
49,
0,
0,
254,
191,
73,
68,
65,... | [
-0.05338004603981972,
0.007029631175100803,
0.6783280968666077,
3.1407692432403564,
0.0017593271331861615,
-0.08994418382644653,
0.03878866136074066,
-0.03878721222281456
] | [
-0.6162847365161563,
0.2636697775018787,
1.7497075536669522,
-2.192049980163574,
-2.193129777908325,
0.0622796006500721,
-1
] | 0 | 0 | 0 | 0 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+fElEQVR4nOz9WbAtyXUYiq21MqtqT2e+c98e0BOABr(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+f0lEQVR4nJT92bOs2XUfiP1+a+/Mc+6tW7fmQhUmAi(...TRUNCATED) | [-0.05331140756607056,0.007027116138488054,0.6783228516578674,3.140770196914673,0.001722074346616864(...TRUNCATED) | [-0.6833599166079287,0.11436976120195748,1.729176878116946,-2.192049980163574,-2.193129777908325,0.0(...TRUNCATED) | 0.1 | 1 | 0 | 1 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+uklEQVR4nOz9W7Blx3UYCK61MvfjPO67qlCFwhskCI(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD/k0lEQVR4nHz96a932XUeiD3P2vvc+w41DxxKVZxEUh(...TRUNCATED) | [-0.05320742353796959,0.007024175021797419,0.6782967448234558,3.140775442123413,0.00169978360645473,(...TRUNCATED) | [-0.7476034447036296,0.016543411542094724,1.6625444069102184,-2.192049980163574,-2.193129777908325,0(...TRUNCATED) | 0.2 | 2 | 0 | 2 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+pUlEQVR4nOz9WbBl2XUYBq619j7DHd6ULzMrM2tETS(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD9x0lEQVR4nHT917O16XUfiP1+63nefc4XOncjNFKjkQ(...TRUNCATED) | [-0.05315941572189331,0.007020609453320503,0.6782735586166382,3.140772581100464,0.001685390714555978(...TRUNCATED) | [-0.8012059043211467,-0.050593769284297666,1.5957494085051471,-2.192049980163574,-2.193129777908325,(...TRUNCATED) | 0.3 | 3 | 0 | 3 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD++UlEQVR4nOz9WbAlyXUYCJ5z3GO5y1tzq6zMqqwNVV(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD970lEQVR4nHT92a9t23kfiP1+3xhz7dPdjre/vGzFnh(...TRUNCATED) | [-0.053153958171606064,0.006965538952499628,0.6782670021057129,3.1407337188720703,0.0015781606780365(...TRUNCATED) | [-0.8109599450892988,-0.069275359450036,1.5913911302209016,-2.192049980163574,-2.193129777908325,0.0(...TRUNCATED) | 0.4 | 4 | 0 | 4 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD/FElEQVR4nOz919M9R3YYCJ5zMstc89mfN8APaKDRaA(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD990lEQVR4nHT9WbCsWXYehn3f2jvznHtv3Rq6qqvnec(...TRUNCATED) | [-0.053149230778217316,0.00663584191352129,0.678267240524292,3.140554428100586,0.0010026784148067236(...TRUNCATED) | [-0.8117838533681199,-0.08151683468343629,1.5974907011658468,-2.192049980163574,-2.193129777908325,0(...TRUNCATED) | 0.5 | 5 | 0 | 5 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD/E0lEQVR4nOz9V9ctyXUYCO69I9Ic87lr65ryhSq4Mi(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD9BUlEQVR4nHz96bNtW1YfiP1+Y869z7n3vr7LvodMEl(...TRUNCATED) | [-0.053141187876462936,0.005687231197953224,0.6782097816467285,3.140174388885498,-0.0000849748685141(...TRUNCATED) | [-0.8122408580244523,-0.09527638202906903,1.5950059283273978,-2.192049980163574,-2.193129777908325,0(...TRUNCATED) | 0.6 | 6 | 0 | 6 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+e0lEQVR4nOz9abMlyXUYCJ5z3GO5y9vyZWZlVWXthU(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+L0lEQVR4nHz9adNs2XUeiD1r7Z35vvfemucqAAUOYB(...TRUNCATED) | [-0.05314163118600845,0.003786786925047636,0.6779237985610962,3.1396303176879883,-0.0014228069921955(...TRUNCATED) | [-0.8159662620877657,-0.10898831239295687,1.590373245288339,-2.192049980163574,-2.193129777908325,0.(...TRUNCATED) | 0.7 | 7 | 0 | 7 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD/KElEQVR4nOz9WdMlyXUYCJ5z3GO5y7fnUlmZVYlakI(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD8GElEQVR4nHT957Nl2XUfCP7W2vvc9zKrshzKAQVTsC(...TRUNCATED) | [-0.0532541386783123,0.0008253292180597782,0.6773104667663574,3.1390726566314697,-0.0026678470894694(...TRUNCATED) | [-0.8209503808811313,-0.12129978308131847,1.5862895829331805,-2.192049980163574,-2.193129777908325,0(...TRUNCATED) | 0.8 | 8 | 0 | 8 | 0 |
{"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD+lklEQVR4nOz9+dMtx3UYCJ5zMmu5y7e+He/hYSNAkA(...TRUNCATED) | {"bytes":"iVBORw0KGgoAAAANSUhEUgAAAQAAAAEACAIAAADTED8xAAD9A0lEQVR4nHz96bNl2XUfiP3W2vve916ONRfGQhUmAi(...TRUNCATED) | [-0.053471047431230545,-0.0030237147584557533,0.6764023303985596,3.1386992931365967,-0.0034863611217(...TRUNCATED) | [-0.8252032499344313,-0.13277420678156746,1.5818177289912008,-2.192049980163574,-2.193129777908325,0(...TRUNCATED) | 0.9 | 9 | 0 | 9 | 0 |
This dataset is a converted version of the LIBERO robot learning benchmark, where actions have been converted from relative (delta) commands to absolute end-effector poses.
[x, y, z, axis_angle_x, axis_angle_y, axis_angle_z, gripper](x, y, z): Absolute position in world coordinates(axis_angle_x, axis_angle_y, axis_angle_z): Rotation as axis-angle representationgripper: Gripper state (-1: open, 1: closed)The conversion was performed by simulating each episode in the LIBERO environment and recording the absolute end-effector states.
Same structure as the original LeRobot LIBERO dataset:
data/chunk-000/: Main dataset parquet filesmeta/: Metadata including info.json, stats.json, tasks.parquet, and episode statisticsfrom datasets import load_dataset
dataset = load_dataset("albus2024/libero_absolute", split="train")
# Access absolute actions
for episode in dataset:
action = episode['action'] # [x, y, z, ax, ay, az, gripper]
# action is now in absolute coordinates
If you use this dataset, please cite both LIBERO and this conversion:
@inproceedings{liu2023libero,
title={LIBERO: Benchmarking Knowledge Transfer in Lifelong Robot Learning},
author={Liu, Bo and Zhu, Yifeng and Gao, Chongkai and Feng, Yihao and Liu, Qiang and Zhu, Yuke and Stone, Peter},
booktitle={NeurIPS 2023 Datasets and Benchmarks Track},
year={2023}
}
@article{zheng2025x,
title = {X-VLA: Soft-Prompted Transformer as Scalable Cross-Embodiment Vision-Language-Action Model},
author = {Zheng, Jinliang and Li, Jianxiong and Wang, Zhihao and Liu, Dongxiu and Kang, Xirui
and Feng, Yuchun and Zheng, Yinan and Zou, Jiayin and Chen, Yilun and Zeng, Jia and others},
journal = {arXiv preprint arXiv:2510.10274},
year = {2025}
}
Apache 2.0 (same as original LIBERO dataset)