TinyLLaVA/TinyLLaVA_Factory
A Framework of Small-scale Large Multimodal Models
This project offers a specialized framework for creating and customizing small-scale Large Multimodal Models (LMMs). It takes raw image and text data, along with configuration choices for language models, vision models, and training methods, to produce a finely tuned LMM. This is for machine learning researchers and practitioners who want to build efficient LMMs without extensive coding.
962 stars. Actively maintained with 1 commit in the last 30 days.
Use this if you are a machine learning researcher or engineer looking to develop or experiment with custom, compact multimodal AI models that can understand both images and text.
Not ideal if you are an end-user simply looking to use an existing multimodal AI model off-the-shelf without any customization or training.
Stars
962
Forks
96
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 11, 2026
Commits (30d)
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/TinyLLaVA/TinyLLaVA_Factory"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
zjunlp/EasyInstruct
[ACL 2024] An Easy-to-use Instruction Processing Framework for LLMs.
rese1f/MovieChat
[CVPR 2024] MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
NVlabs/Eagle
Eagle: Frontier Vision-Language Models with Data-Centric Strategies
DAMO-NLP-SG/Video-LLaMA
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding