kyegomez/AutoRT
Implementation of AutoRT: "AutoRT: Embodied Foundation Models for Large Scale Orchestration of Robotic Agents"
This tool helps robotics researchers and engineers turn visual observations of a scene into a prioritized list of tasks that a robot can then execute. You provide an image or visual input describing a scene and a general goal, and it outputs a sequence of actionable instructions for a robotic system. It's designed for those orchestrating single or multiple robotic agents.
No commits in the last 6 months.
Use this if you need to automatically generate, filter, and prioritize tasks for a robotic agent based on visual input, streamlining the process of getting robots to perform complex actions in real-world environments.
Not ideal if you are looking for a tool to design robot hardware or control low-level motor functions directly, as this focuses on high-level task generation and orchestration.
Stars
42
Forks
3
Language
Python
License
MIT
Category
Last pushed
Nov 11, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kyegomez/AutoRT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch