autonomousvision/transfuser
[PAMI'23] TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving; [CVPR'21] Multi-Modal Fusion Transformer for End-to-End Autonomous Driving
This project helps self-driving car developers and researchers create and test models for autonomous vehicles. It takes in raw sensor data like camera images, depth images, LiDAR, and semantic segmentation maps, then processes it to output driving commands for a simulated car. The primary users are engineers and scientists working on perception and control systems for autonomous driving.
1,516 stars.
Use this if you are developing or evaluating end-to-end autonomous driving systems and need a robust framework for sensor fusion and imitation learning.
Not ideal if you are looking for a plug-and-play solution for physical self-driving cars, as this is a research framework for simulated environments.
Stars
1,516
Forks
233
Language
Python
License
MIT
Category
Last pushed
Oct 19, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/autonomousvision/transfuser"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
NVlabs/MambaVision
[CVPR 2025] Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
sign-language-translator/sign-language-translator
Python library & framework to build custom translators for the hearing-impaired and translate...
kyegomez/Jamba
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
kyegomez/MultiModalMamba
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance...
dali92002/DocEnTR
DocEnTr: An end-to-end document image enhancement transformer - ICPR 2022