zchoi/Multi-Modal-Large-Language-Learning
Awesome multi-modal large language paper/project, collections of popular training strategies, e.g., PEFT, LoRA.
This is a curated collection of cutting-edge research papers and open-source projects related to Multi-modal Large Language Models (MLLMs). It gathers information on different MLLM architectures, evaluation benchmarks, foundational models that underpin them, and efficient methods for adapting these models to specific tasks. It's for researchers, AI developers, and academics who are actively working with or studying advanced AI models that process and understand both text and other forms of data like images or video.
No commits in the last 6 months.
Use this if you need a comprehensive overview and direct links to the latest research and implementations in multi-modal AI, particularly for developing or evaluating MLLMs.
Not ideal if you are a non-technical end-user looking for a ready-to-use AI application, as this project focuses on the underlying models and research.
Stars
27
Forks
—
Language
—
License
—
Category
Last pushed
Aug 02, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/zchoi/Multi-Modal-Large-Language-Learning"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
FoundationVision/Liquid
(Accepted by IJCV) Liquid: Language Models are Scalable and Unified Multi-modal Generators
Paranioar/Awesome_Matching_Pretraining_Transfering
The Paper List of Large Multi-Modality Model (Perception, Generation, Unification),...
Yangyi-Chen/Multimodal-AND-Large-Language-Models
Paper list about multimodal and large language models, only used to record papers I read in the...
thuml/AutoTimes
Official implementation for "AutoTimes: Autoregressive Time Series Forecasters via Large Language Models"