AviSoori1x/seemore

From scratch implementation of a vision language model in pure PyTorch

42
/ 100
Emerging

This is a detailed, from-scratch implementation of a vision language model (VLM) in PyTorch. It takes an image and a text prompt as input and generates human-like text outputs, similar to how advanced AI models understand both images and text. It's designed for machine learning researchers, students, or practitioners who want to deeply understand how these multimodal AI models work by building one from its fundamental components.

255 stars. No commits in the last 6 months.

Use this if you are a machine learning researcher or student who wants to learn the foundational principles of vision language models by examining a complete, transparent, and hackable implementation.

Not ideal if you are looking for an off-the-shelf, production-ready vision language model for immediate application, as this project prioritizes educational value and readability over performance.

deep-learning-research ai-model-development multimodal-ai natural-language-processing computer-vision
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

255

Forks

31

Language

Jupyter Notebook

License

MIT

Last pushed

May 06, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AviSoori1x/seemore"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.