shrimo/SLAMBox
Education, research and development using the Simultaneous Localization and Mapping (SLAM) method.
SLAMBOX helps robotics engineers, researchers, and students quickly experiment with Simultaneous Localization and Mapping (SLAM) systems. You input raw sensor data (like video, LIDAR, or stereo camera feeds) and configure various SLAM components visually, receiving an optimized 3D map of an environment and the agent's tracked location within it. It's designed for anyone needing to prototype or understand SLAM without deep programming knowledge.
Use this if you need to rapidly prototype, test, or learn about different Simultaneous Localization and Mapping (SLAM) algorithms using a visual, node-based interface.
Not ideal if you need to deploy a highly optimized, production-ready SLAM system that requires custom code integration or maximum performance.
Stars
98
Forks
14
Language
Python
License
MIT
Category
Last pushed
Mar 22, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/shrimo/SLAMBox"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
alicevision/AliceVision
3D Computer Vision Framework
colmap/colmap
COLMAP - Structure-from-Motion and Multi-View Stereo
ANTsX/ANTs
Advanced Normalization Tools (ANTs)
alicevision/Meshroom
Node-based Visual Programming Toolbox
MOLAorg/mola
A Modular Optimization framework for Localization and mApping (MOLA)