kyegomez/PaLM2-VAdapter

Implementation of "PaLM2-VAdapter:" from the multi-modal model paper: "PaLM2-VAdapter: Progressively Aligned Language Model Makes a Strong Vision-language Adapter"

22
/ 100
Experimental

This is a developer tool that implements a specific research paper to help connect large language models with image processing capabilities. It takes image and text data as input and produces an output that aligns these different data types, enabling further processing by other large language models. This is for machine learning engineers and researchers working on advanced AI models that understand both text and images.

No commits in the last 6 months.

Use this if you are a machine learning engineer or researcher building multi-modal AI systems that need to process and understand both visual and textual information together.

Not ideal if you are looking for a ready-to-use application or a high-level API for general image or text analysis.

AI model development multi-modal AI large language models computer vision research natural language processing research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

17

Forks

Language

Python

License

MIT

Last pushed

Nov 11, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/kyegomez/PaLM2-VAdapter"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.