Vision-Language-Action Model

From GM-RKB
Jump to navigation Jump to search

A Vision-Language-Action Model is a multimodal AI model that integrates visual perception, language understanding, and physical action for embodied AI tasks.