fuenwang/BiFusev2

BiFuse++: Self-supervised and Efficient Bi-projection Fusion for 360 Depth Estimation

37
/ 100
Emerging

This project helps researchers working with 360-degree cameras generate detailed depth maps from panoramic images. It takes a single 360-degree image as input and outputs a corresponding depth map, indicating the distance of objects from the camera. This is ideal for computer vision researchers, robotics engineers, or anyone building virtual environments who needs accurate spatial understanding from omnidirectional imagery.

No commits in the last 6 months.

Use this if you need to extract precise depth information from 360-degree panoramic photos for tasks like 3D reconstruction or environmental perception.

Not ideal if you are working with standard perspective images (non-360 photos) or if your primary goal is object detection or classification rather than depth estimation.

360-degree vision depth estimation computer vision research robotics perception virtual environment modeling
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

72

Forks

6

Language

Python

License

MIT

Last pushed

May 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/fuenwang/BiFusev2"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.