sou350121/VLA-Handbook
本项目旨在为致力于进入VLA(Vision-Language-Action)领域的算法工程师提供一份全中文、实战导向的学习/面试手册。 不同于通用的 CV/NLP 面试指南,本项目聚焦于 Robotics 特有的挑战
This handbook provides a practical, Chinese-language guide for algorithm engineers working in Vision-Language-Action (VLA) robotics. It helps bridge the gap between understanding academic papers and successfully implementing VLA models by detailing engineering challenges and offering real-world deployment insights. The resource integrates daily updates from cutting-edge research, community discussions, and GitHub issue experiences, making it ideal for those developing and deploying robot control systems.
Use this if you are an algorithm engineer focused on robotics and need concrete, executable details to implement Vision-Language-Action models, especially if you're navigating the complexities of multi-modal synchronization, Sim2Real transitions, or action space alignment.
Not ideal if you are looking for high-level overviews of VLA concepts without delving into the specific engineering challenges and implementation details.
Stars
73
Forks
5
Language
HTML
License
—
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/sou350121/VLA-Handbook"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
chrisliu298/awesome-llm-unlearning
A resource repository for machine unlearning in large language models
worldbench/awesome-vla-for-ad
🌐 Vision-Language-Action Models for Autonomous Driving: Past, Present, and Future
hijkzzz/Awesome-LLM-Strawberry
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
zjukg/KG-MM-Survey
Knowledge Graphs Meet Multi-Modal Learning: A Comprehensive Survey
worldbench/awesome-spatial-intelligence
🌐 Forging Spatial Intelligence: A Roadmap of Multi-Modal Data Pre-Training for Autonomous Systems