← Back to Benchmarks
simmediumsim-to-realmetric · varies

InternVLA-A1: Unifying Understanding, Generation and Action for Robotic Manipulation

Description

Prevalent Vision-Language-Action (VLA) models are typically built upon Multimodal Large Language Models (MLLMs) and demonstrate exceptional proficiency in semantic understanding, but they inherently lack the capability to deduce physical world dynamics. Consequently, recent approaches have shifted toward World Models, typically formulated via video prediction; however, these methods often suffer from a lack of semantic grounding and exhibit brittleness in the presence of video prediction errors.

Source

http://arxiv.org/abs/2601.02456v2