Post
3946
Hello, amazing robotics people π π π We have FINALLY delivered on your major request! Ark just got a major upgrade:
Weβve now integrated Vision-Language-Action Models (VLAs) into Ark π VLAs = models that connect vision + language β robot actions (see image)
What does this mean?
π£οΈ Give robots natural language instructions β they act
π Combine perception + language for real-world control
π¦Ύ Powered by pi0 pretrained models for fast prototyping
β‘ Supports easy data collection and fine-tuning within Ark within a couple of lines of code
Next, we plan to go into the world of designing worlds π
Who knows, maybe those video models are actually zero-shot learners and reasoners?
Check it out here π https://github.com/Robotics-Ark/ark_framework
Check out the tutorial π https://arkrobotics.notion.site/VLA-Pi0-with-Ark-279e053d9c6f800ab0a2d498835dd96b
β Star the repo, try it with your robots, and let us together make robots great (again?)!
Weβve now integrated Vision-Language-Action Models (VLAs) into Ark π VLAs = models that connect vision + language β robot actions (see image)
What does this mean?
π£οΈ Give robots natural language instructions β they act
π Combine perception + language for real-world control
π¦Ύ Powered by pi0 pretrained models for fast prototyping
β‘ Supports easy data collection and fine-tuning within Ark within a couple of lines of code
Next, we plan to go into the world of designing worlds π
Who knows, maybe those video models are actually zero-shot learners and reasoners?
Check it out here π https://github.com/Robotics-Ark/ark_framework
Check out the tutorial π https://arkrobotics.notion.site/VLA-Pi0-with-Ark-279e053d9c6f800ab0a2d498835dd96b
β Star the repo, try it with your robots, and let us together make robots great (again?)!