heisenberg141/Pointcloud-Segmentation

This repository contains sensor fusion between a lidar and camera, semantic segmentation on point clouds and ICP registration of multiple point clouds.

19
/ 100
Experimental

This project helps operations engineers and robotics developers create detailed 3D maps of environments by combining data from cameras and LiDAR sensors. It takes raw camera images and 3D point cloud data, then processes them to produce semantically labeled, colorized point clouds and a unified, registered 3D map of the scene. This is useful for anyone building perception systems for autonomous vehicles or robotics.

No commits in the last 6 months.

Use this if you need to merge 2D image information with 3D LiDAR scans to understand and map complex environments, especially for autonomous navigation.

Not ideal if you only have a single type of sensor data (e.g., just a camera or just LiDAR) and don't need to combine them for semantic understanding.

autonomous-navigation robotics-perception 3D-mapping sensor-fusion environmental-modeling
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 6 / 25

How are scores calculated?

Stars

13

Forks

1

Language

Python

License

Last pushed

Jun 03, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/heisenberg141/Pointcloud-Segmentation"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.