OpenGVLab/Instruct2Act
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Instruct2Act helps robotics engineers and researchers translate complex, multi-modal instructions (like "Put the polka dot block into the green container" combined with pointing gestures) into precise, sequential actions for robotic arms. It takes in human-like commands, visual cues, and robot operating environments, then outputs the necessary code to execute those commands, making robotic manipulation tasks more intuitive. This is for professionals working with robotic systems that need to perform varied physical tasks.
373 stars. No commits in the last 6 months.
Use this if you need to program robotic arms to perform manipulation tasks based on natural language and visual input, without writing all the low-level code yourself.
Not ideal if your robotic tasks are fixed, repetitive, and don't require dynamic interpretation of diverse, high-level commands.
Stars
373
Forks
22
Language
Python
License
—
Category
Last pushed
Jun 23, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/OpenGVLab/Instruct2Act"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
MantisAI/sieves
Plug-and-play document AI with zero-shot models.
xiaoya-li/Instruction-Tuning-Survey
Project for the paper entitled `Instruction Tuning for Large Language Models: A Survey`
TencentARC-QQ/TagGPT
TagGPT: Large Language Models are Zero-shot Multimodal Taggers
rafaelpierre/bullet
bullet: A Zero-Shot / Few-Shot Learning, LLM Based, text classification framework
amazon-science/adaptive-in-context-learning
AdaICL: Which Examples to Annotate of In-Context Learning? Towards Effective and Efficient Selection