tthinking/MATR
[IEEE TIP 2022] Official implementation of MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer
This project helps medical professionals like radiologists or diagnostic specialists combine information from multiple medical image scans, such as MRI and CT. It takes two or more different types of medical images of the same area and merges them into a single, enhanced image that provides a more comprehensive view. The output is a fused image that reveals details not visible in individual scans.
No commits in the last 6 months.
Use this if you need to merge different types of medical scans (e.g., MRI, CT, PET) into a single, high-information image for improved diagnosis or analysis.
Not ideal if you are working with non-medical images or only have single-modality medical images to analyze.
Stars
99
Forks
10
Language
Python
License
—
Category
Last pushed
Mar 16, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tthinking/MATR"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
NVlabs/MambaVision
[CVPR 2025] Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
sign-language-translator/sign-language-translator
Python library & framework to build custom translators for the hearing-impaired and translate...
kyegomez/Jamba
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
autonomousvision/transfuser
[PAMI'23] TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving;...
kyegomez/MultiModalMamba
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance...