keivalya/mini-vla

a minimal, beginner-friendly VLA to show how robot policies can fuse images, text, and states to generate actions

57
/ 100
Established

This project helps robotics students and researchers understand how to build models that enable robots to follow instructions. It takes robot sensor data (images, internal state) and text commands, then outputs continuous actions for the robot to perform. This is for anyone learning or prototyping robot control policies, particularly those interested in vision-language-action (VLA) models.

204 stars.

Use this if you are a student or researcher looking for a clear, minimalist example to learn or prototype vision-language-action models for robotics.

Not ideal if you need a production-ready, state-of-the-art robot control system or a robust solution for real-world industrial applications.

robotics education robot control systems robot learning robot policy prototyping automation research
No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 13 / 25
Community 21 / 25

How are scores calculated?

Stars

204

Forks

40

Language

Python

License

MIT

Last pushed

Mar 17, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/keivalya/mini-vla"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.