🔥[ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy
-
Updated
Mar 29, 2025 - Python
🔥[ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy
🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes.
Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning
Pipeline which takes raw VLA data and yields science-ready data products, i.e. star formation rates
Very Large Array Image Download and Time Lapse Builder
Add a description, image, and links to the vla topic page so that developers can more easily learn about it.
To associate your repository with the vla topic, visit your repo's landing page and select "manage topics."