Awesome-Multimodal-Large-Language-Models and Awesome-VLA
A is a comprehensive collection of resources on multimodal large language models, including Vision Language Action (VLA) models, making B, which specifically focuses on VLA advancements, a specialized subset or a more focused alternative to A within the broader multimodal LLM ecosystem.
About Awesome-Multimodal-Large-Language-Models
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
This resource helps AI researchers and practitioners stay current with the rapidly evolving field of Multimodal Large Language Models (MLLMs). It provides curated lists of significant research papers, comprehensive surveys, and evaluation benchmarks for MLLMs. The intended users are researchers, students, and engineers who are actively working on or studying advanced AI models that integrate different data types like text, images, and audio.
About Awesome-VLA
Orlando-CS/Awesome-VLA
✨✨latest advancements in VLA models(VIsion Language Action)
This collection provides an overview of the latest research and advancements in Vision-Language-Action (VLA) models. It helps researchers and engineers quickly find information on cutting-edge models, relevant papers, and available datasets related to training robots and embodied AI to understand and act based on visual and linguistic input. The main users are AI researchers, robotics engineers, and deep learning practitioners focused on developing autonomous systems.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work