InternVLA-M1

Latent Spatial Grounding for Instruction-Following Robotic Manipulation

Abstract

We introduce InternVLA-M1, a unified framework for spatial grounding and robot control that advances instruction-following robots toward general-purpose intelligence. Its core idea is spatially guided vision-language-action training, where spatial grounding serves as the critical link between instructions and robot actions. InternVLA-M1 employs a two-stage pipeline: (i) spatial grounding pre-training on over 2.3M spatial reasoning data to determine “where to act” by aligning instructions with visual, embodiment-agnostic positions, and (ii) spatially guided action post-training to decide “how to act” by generating embodiment-aware actions through plug-and-play spatial prompting. This spatially guided training recepit yields consistent gains: InternVLA-M1 outperforms its variant without spatial guidance by +13.6% on SimplerEnv Google Robot, +17% on WidowX, and +4.3% on LIBERO Franka. To further scale instruction following, we built a simulation engine to collect 244K pick-and-place episodes, enabling a 6.2% average improvement across 200 tasks and 3K+ objects. In real-world clustered pick-and-place, InternVLA-M1 improved by 7.3%, and with synthetic co-training, achieved +20.6% on unseen objects and novel configurations. Moreover, in long-horizon reasoning-intensive scenarios, it surpassed existing works by over 10 points. These results highlight spatially guided training as a unifying principle for scalable and resilient generalist robots.

Model Overview

InternVLA-M1 Model Architecture

InternVLA-M1 integrates spatial grounding into the vision–language–action training pipeline. Given a task instruction, the VLM planner produces latent plans through explicit spatial prompting, which then effectively guides the action expert to generate control signals.

Simulation Data Generation

Simulation Data Pipeline

The pipeline automatically generates diverse instruction-following robotic manipulation data from a large asset library, incorporating intermediate representations such as Box, Point, and Trajectory, which can be further converted into VLM spatial grounding data.

Results

Watch InternVLA-M1 perform instruction-following manipulation tasks in both large-scale simulated environments and real-world tasks.

Instruction-Following Manipulation

Cluttered-scene Pick-and-Place

Long-horizon and Reasoning Manipulation

Experimental Results

InternVLA-M1 demonstrates superior performance across various challenging scenarios

Performance Comparison on Simpler Env and Libero Benchmark

0 20 40 60 80 Success Rate (%) Google Robot VM 35.2 58.8 80.7 Google Robot VA 44.5 54.8 76 WidowX VM 61.9 27.1 71.7 Libero 93.9 94.2 95.9
GR00T
π0​
InternVLA-M1

Effects of Spatially Guided VLA Training

0 20 40 60 80 Success Rate (%) Google Robot VM 66.1 80.7 Google Robot VA 63.5 76 WidowX VM 54.7 71.7 Libero 91.6 95.9
Vanilla VLA
InternVLA-M1

System2 Spatial Reasoning Results

Demonstrating InternVLA's System 2 capabilities in box detection, point localization, and visual trace prediction.

📦 Box Detection

Precise bounding box detection and object localization

📍 Point Localization

Precise keypoint localization and spatial analysis

📈 Trajectory Prediction

Intelligent trajectory prediction and motion path planning

VLM Pre-training Data Distribution

Comprehensive dataset composition for spatial grounding pre-training

Pre-training Data 3,032K General VQA Spatial Grounding QA VQA Trajectory-QA Point-QA BOX-QA
VLM Training Data
General VQA 21.0%
Spatial Grounding QA 79.0%
Detailed Components
VQA 21.0%
Trajectory-QA 22.6%
Point-QA 27.4%
BOX-QA 29.0%

Reference

@article{2025internvlam1,
  title = {InternVLA-M1: Latent Spatial Grounding for Instruction-Following Robotic Manipulation},
  author = {Intern Robotics},
  booktitle = {Arxiv},
  year = {2025},
}