zchoi/Multi-Modal-Large-Language-Learning

Awesome multi-modal large language paper/project, collections of popular training strategies, e.g., PEFT, LoRA.

15
/ 100
Experimental

This is a curated collection of cutting-edge research papers and open-source projects related to Multi-modal Large Language Models (MLLMs). It gathers information on different MLLM architectures, evaluation benchmarks, foundational models that underpin them, and efficient methods for adapting these models to specific tasks. It's for researchers, AI developers, and academics who are actively working with or studying advanced AI models that process and understand both text and other forms of data like images or video.

No commits in the last 6 months.

Use this if you need a comprehensive overview and direct links to the latest research and implementations in multi-modal AI, particularly for developing or evaluating MLLMs.

Not ideal if you are a non-technical end-user looking for a ready-to-use AI application, as this project focuses on the underlying models and research.

AI-research natural-language-processing computer-vision machine-learning-engineering foundation-models
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 0 / 25

How are scores calculated?

Stars

27

Forks

Language

License

Last pushed

Aug 02, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/zchoi/Multi-Modal-Large-Language-Learning"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.