[CVPR 2025] Open-source, End-to-end, Vision-Language-Action model for GUI Agent & Computer Use.
-
Updated
Mar 13, 2025 - Python
[CVPR 2025] Open-source, End-to-end, Vision-Language-Action model for GUI Agent & Computer Use.
[CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"
A simple and scalable codebase for training and fine-tuning vision-language-action models (VLAs) for generalist robotic manipulation:
VLAGen: Automated Data Collection for Generalizing Robotic Policies
Add a description, image, and links to the vision-language-action topic page so that developers can more easily learn about it.
To associate your repository with the vision-language-action topic, visit your repo's landing page and select "manage topics."