moein-shariatnia/OpenAI-CLIP
Simple implementation of OpenAI CLIP model in PyTorch.
This project helps researchers and engineers build models that understand both images and text together. It takes a collection of images and their descriptive captions, processing them to create a model that can connect what's seen in a picture with what's said in a sentence. This is useful for anyone working on tasks like searching images using text descriptions, or classifying images based on natural language.
720 stars.
Use this if you need to train a model to find relationships between visual content and its textual descriptions, especially for tasks like cross-modal search or image classification guided by language.
Not ideal if you are looking for an out-of-the-box solution for a specific image analysis task without needing to train or fine-tune a model yourself.
Stars
720
Forks
104
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Oct 18, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/moein-shariatnia/OpenAI-CLIP"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
mlfoundations/open_clip
An open source implementation of CLIP.
noxdafox/clipspy
Python CFFI bindings for the 'C' Language Integrated Production System CLIPS
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
BioMedIA-MBZUAI/FetalCLIP
Official repository of FetalCLIP: A Visual-Language Foundation Model for Fetal Ultrasound Image Analysis
filipbasara0/simple-clip
A minimal, but effective implementation of CLIP (Contrastive Language-Image Pretraining) in PyTorch