Skip to content
/ vlax Public

Flexible vision-language-action models for robotics in JAX

License

Notifications You must be signed in to change notification settings

kylestach/vlax

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VLAx: Vision-Language-Action models in JAX

This repository is an implementation of a vision-language-action model for robotics, including both training and inference code. It's currently heavily under development, so check back soon for more updates!

The current VLA implementation is based on PaliGemma but supports some nice features such as masked multi-modal inputs, flexible outputs, etc.

VLAx uses the experimental grain-oxe format for robotics data.

About

Flexible vision-language-action models for robotics in JAX

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published