[IROS 2025 Best Paper Award Finalist & IEEE TRO 2026] The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems
-
Updated
Dec 16, 2025 - Python
[IROS 2025 Best Paper Award Finalist & IEEE TRO 2026] The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems
VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model
StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing
A comprehensive list of papers about Robot Manipulation, including papers, codes, and related websites.
Official code of Motus: A Unified Latent Action World Model
InternRobotics' open platform for building generalized navigation foundation models.
[AAAI 2026] OpenDriveVLA: Towards End-to-end Autonomous Driving with Large Vision Language Action Model
[ICLR 2026] The offical Implementation of "Soft-Prompted Transformer as Scalable Cross-Embodiment Vision-Language-Action Model"
InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy
OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulation
InternVLA-A1: Unifying Understanding, Generation, and Action for Robotic Manipulation
NORA: A Small Open-Sourced Generalist Vision Language Action Model for Embodied Tasks
Official implementation of ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver.
LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintained🔥]
WAM-Flow: Parallel Coarse-to-Fine Motion Planning via Discrete Flow Matching for Autonomous Driving
WAM-Diff: A Masked Diffusion VLA Framework with MoE and Online Reinforcement Learning for Autonomous Driving
🔥 The first open-sourced diffusion vision-langauge-action model.
Open & Reproducible Research for Tracking VLAs
🔥This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the repository, so follow us to keep up with the latest developments!!!
A collection of vision-language-action model post-training methods.
Add a description, image, and links to the vision-language-action-model topic page so that developers can more easily learn about it.
To associate your repository with the vision-language-action-model topic, visit your repo's landing page and select "manage topics."