GerrySant/multimodalhugs
MultimodalHugs is an extension of Hugging Face that offers a generalized framework for training, evaluating, and using multimodal AI models with minimal code differences, ensuring seamless compatibility with Hugging Face pipelines.
Stars
15
Forks
7
Language
Python
License
MIT
Last pushed
Apr 10, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/multimodal/GerrySant/multimodalhugs"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
starVLA/starVLA
StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing
vortex-data/vortex
An extensible, state-of-the-art framework for columnar compression, and the fastest FOSS...
motis-project/motis
multimodal routing, geocoding, and map tiles
zai-org/GLM-V
GLM-4.6V/4.5V/4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning
neka-nat/cad3dify
2D to 3D CAD Conversion Using VLM