alphaXiv

Explore

State of the Art

Sign In

Labs

Feedback

Browser Extension

We're hiring
PaperBlogResources

IA-VLA: Input Augmentation for Vision-Language-Action models in settings with semantically complex tasks