VLA
Vision-Language-Action models (VLA)
VLA models
Paper: Vision-Language-Action Models: Concepts, Progress, Applications and Challenges




RT-1
Paper: RT-1: Robotics Transformer for Real-World Control at Scale

RT-2
Paper: RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control

BridgeData V2
Paper: BridgeData V2: A Dataset for Robot Learning at Scale
Code: https://github.com/rail-berkeley/bridge_data_v2
Open X-Embodiment
Paper: Open X-Embodiment: Robotic Learning Datasets and RT-X Models

OpenVLA
Paper: OpenVLA: An Open-Source Vision-Language-Action Model
OpenVLA-OFT
Paper: Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success
Code: https://github.com/moojink/openvla-oft
SimpleVLA-RL
Paper: SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning
Code: https://github.com/PRIME-RL/SimpleVLA-RL

SmolVLA
Model: lerobot/smolvla_base
Paper: SmolVLA: A Vision-Language-Action Model for Affordable and Efficient Robotics

SO-101 robot
SO-ARM101 AI 機器手臂PRO套件 for LeRobot

Robot Learning
Paper: Robot Learning: A Tutorial
Code: https://github.com/fracapuano/robot-learning-tutorial
HIL-SERL


Generalist Robot Policies

RLVR-World
Paper RLVR-World: Training World Models with Reinforcement Learning
Code: https://github.com/thuml/RLVR-World

Agentic Robot
Paper: Agentic Robot: A Brain-Inspired Framework for Vision-Language-Action Models in Embodied Agents
Github:https://github.com/Agentic-Robot/agentic-robot

SAP: Standardized Action Procedure for Coordinated Agentic Control

ACoT-VLA
Paper: ACoT-VLA: Action Chain-of-Thought for Vision-Language-Action Models

This site was last updated April 15, 2026.