PardisTaghavi/SwinMTL

[IROS24]A Shared Architecture for Simultaneous Depth Estimation and Semantic Segmentation from Monocular Camera Images

33
/ 100
Emerging

This project helps roboticists and autonomous vehicle developers process single camera images to understand both how far away objects are and what those objects are (e.g., road, car, pedestrian). You input a standard monocular camera image, and it outputs two enhanced images: one showing depth perception and another with objects clearly segmented and labeled. This is ideal for those building navigation systems or perception stacks for robots.

No commits in the last 6 months.

Use this if you need to rapidly extract both depth information and semantic understanding from a single camera feed for robotic navigation or environmental perception.

Not ideal if your application requires extremely high-precision depth sensing that might be better suited for stereo cameras, LiDAR, or other dedicated depth sensors.

robot-perception autonomous-navigation robotics-computer-vision mobile-robot-mapping scene-understanding
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 9 / 25

How are scores calculated?

Stars

48

Forks

4

Language

Jupyter Notebook

License

GPL-3.0

Last pushed

Mar 11, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/PardisTaghavi/SwinMTL"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.