Toytiny/RaFlow

[RA-L & IROS'22] Self-Supervised Scene Flow Estimation with 4-D Automotive Radar

41
/ 100
Emerging

This project helps autonomous vehicles understand the movement of objects around them using 4-D automotive radar data. It takes raw radar point clouds as input and outputs a 'scene flow' map, which shows how each object in the scene is moving. This is especially useful for self-driving car engineers and researchers developing robust navigation systems in challenging weather.

No commits in the last 6 months.

Use this if you need to determine the arbitrary motion of multiple independent objects from 4-D radar data for autonomous driving applications.

Not ideal if you are working with LiDAR data or require annotated datasets for training, as this method focuses on self-supervised learning with radar.

autonomous-driving 4d-radar motion-estimation robotics perception-systems
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

75

Forks

12

Language

Python

License

MIT

Last pushed

Mar 18, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/Toytiny/RaFlow"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.