Lift3D Foundation Policy: Lifting 2D Large-Scale Pretrained Models for Robust 3D Robotic Manipulation Paper • 2411.18623 • Published Nov 27, 2024 • 1
RoboMIND: Benchmark on Multi-embodiment Intelligence Normative Data for Robot Manipulation Paper • 2412.13877 • Published Dec 18, 2024
HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model Paper • 2503.10631 • Published Mar 13, 2025
MLA: A Multisensory Language-Action Model for Multimodal Understanding and Forecasting in Robotic Manipulation Paper • 2509.26642 • Published Sep 30, 2025
TwinRL-VLA: Digital Twin-Driven Reinforcement Learning for Real-World Robotic Manipulation Paper • 2602.09023 • Published Feb 9
LaST$_{0}$: Latent Spatio-Temporal Chain-of-Thought for Robotic Vision-Language-Action Model Paper • 2601.05248 • Published Jan 8
Look Before Acting: Enhancing Vision Foundation Representations for Vision-Language-Action Models Paper • 2603.15618 • Published 1 day ago • 7