moein-shariatnia/OpenAI-CLIP

Simple implementation of OpenAI CLIP model in PyTorch.

53
/ 100
Established

This project helps researchers and engineers build models that understand both images and text together. It takes a collection of images and their descriptive captions, processing them to create a model that can connect what's seen in a picture with what's said in a sentence. This is useful for anyone working on tasks like searching images using text descriptions, or classifying images based on natural language.

720 stars.

Use this if you need to train a model to find relationships between visual content and its textual descriptions, especially for tasks like cross-modal search or image classification guided by language.

Not ideal if you are looking for an out-of-the-box solution for a specific image analysis task without needing to train or fine-tune a model yourself.

multimodal-AI image-retrieval natural-language-processing computer-vision data-labeling
No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 21 / 25

How are scores calculated?

Stars

720

Forks

104

Language

Jupyter Notebook

License

MIT

Last pushed

Oct 18, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/moein-shariatnia/OpenAI-CLIP"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.