OpenVLA: An Open-Source Vision-Language-Action Model
Paper
•
2406.09246
•
Published
•
41
CogACT: A Foundational Vision-Language-Action Model for Synergizing
Cognition and Action in Robotic Manipulation
Paper
•
2411.19650
•
Published
Octo: An Open-Source Generalist Robot Policy
Paper
•
2405.12213
•
Published
•
29
Diffusion-VLA: Scaling Robot Foundation Models via Unified Diffusion and
Autoregression
Paper
•
2412.03293
•
Published
robotics-diffusion-transformer/rdt-1b
Robotics
•
Updated
•
734
•
98
OmniManip: Towards General Robotic Manipulation via Object-Centric
Interaction Primitives as Spatial Constraints
Paper
•
2501.03841
•
Published
•
56
Beyond Sight: Finetuning Generalist Robot Policies with Heterogeneous
Sensors via Language Grounding
Paper
•
2501.04693
•
Published
•
3
Spatial Forcing: Implicit Spatial Representation Alignment for
Vision-language-action Model
Paper
•
2510.12276
•
Published
•
145
Robot Learning: A Tutorial
Paper
•
2510.12403
•
Published
•
118
RoboOmni: Proactive Robot Manipulation in Omni-modal Context
Paper
•
2510.23763
•
Published
•
53
RynnVLA-002: A Unified Vision-Language-Action and World Model
Paper
•
2511.17502
•
Published
•
25
MobileVLA-R1: Reinforcing Vision-Language-Action for Mobile Robots
Paper
•
2511.17889
•
Published
•
5
MIND-V: Hierarchical Video Generation for Long-Horizon Robotic Manipulation with RL-based Physical Alignment
Paper
•
2512.06628
•
Published
•
12