SamirAbouHaidar/HARP-NeXt

[IROS 2025] HARP-NeXt: High-Speed and Accurate Range-Point Fusion Network for 3D LiDAR Semantic Segmentation

37
/ 100
Emerging

This project helps self-driving car engineers and robotics researchers quickly and accurately identify objects in 3D environments using LiDAR sensor data. It takes raw 3D point cloud data from LiDAR scans as input and outputs a segmented scene where each point is labeled with what it represents (e.g., car, pedestrian, road, tree). This allows for robust perception and scene understanding in autonomous systems.

Use this if you need to process 3D LiDAR data in real-time on both powerful workstations and embedded systems like NVIDIA Jetson AGX Orin for tasks like autonomous navigation or robotic sensing.

Not ideal if your primary data source is visual camera imagery, as this tool is specifically designed for LiDAR 3D point cloud segmentation.

autonomous-driving robotics 3D-scene-understanding LiDAR-processing real-time-perception
No Package No Dependents
Maintenance 6 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

14

Forks

2

Language

Python

License

Last pushed

Dec 22, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/SamirAbouHaidar/HARP-NeXt"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.