Awesome-Large-Vision-Language-Model and Awesome-Multimodal-LLM
About Awesome-Large-Vision-Language-Model
SuperBruceJia/Awesome-Large-Vision-Language-Model
Awesome Large Vision-Language Model: A Curated List of Large Vision-Language Model
This resource provides a curated collection of materials for anyone exploring or working with large vision-language models, including medical foundation models. It centralizes key papers, presentations, books, and benchmarks related to integrating visual and linguistic data. Researchers and AI practitioners focused on developing or applying advanced AI systems that understand and process both images and text will find this helpful.
About Awesome-Multimodal-LLM
vincentlux/Awesome-Multimodal-LLM
Reading list for Multimodal Large Language Models
This is a reading list for anyone deeply involved in or studying advanced AI, specifically focusing on how large language models (LLMs) can process and understand multiple types of data, like images, video, and text, simultaneously. It provides a structured collection of the latest academic papers, tutorials, and datasets in the field. Researchers, academics, and AI practitioners looking to stay current or explore specific areas within multimodal AI would use this resource.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work