kyegomez/PaLM2-VAdapter
Implementation of "PaLM2-VAdapter:" from the multi-modal model paper: "PaLM2-VAdapter: Progressively Aligned Language Model Makes a Strong Vision-language Adapter"
This is a developer tool that implements a specific research paper to help connect large language models with image processing capabilities. It takes image and text data as input and produces an output that aligns these different data types, enabling further processing by other large language models. This is for machine learning engineers and researchers working on advanced AI models that understand both text and images.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher building multi-modal AI systems that need to process and understand both visual and textual information together.
Not ideal if you are looking for a ready-to-use application or a high-level API for general image or text analysis.
Stars
17
Forks
—
Language
Python
License
MIT
Category
Last pushed
Nov 11, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/kyegomez/PaLM2-VAdapter"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
adapter-hub/adapters
A Unified Library for Parameter-Efficient and Modular Transfer Learning
gaussalgo/adaptor
ACL 2022: Adaptor: a library to easily adapt a language model to your own task, domain, or...
ylsung/VL_adapter
PyTorch code for "VL-Adapter: Parameter-Efficient Transfer Learning for Vision-and-Language...
intersun/LightningDOT
source code and pre-trained/fine-tuned checkpoint for NAACL 2021 paper LightningDOT
kyegomez/M2PT
Implementation of M2PT in PyTorch from the paper: "Multimodal Pathway: Improve Transformers with...