rom1504/clip-retrieval
Easily compute clip embeddings and build a clip retrieval system with them
This project helps you build a search engine for images and text. You input a collection of images and their descriptions, and it creates a system where you can search for visuals using either text descriptions or another image. This is useful for content managers, researchers, or anyone who needs to quickly find specific visual content within a large dataset.
2,733 stars. No commits in the last 6 months. Available on PyPI.
Use this if you have a large collection of images and associated text, and you want to create a semantic search system to easily find relevant content.
Not ideal if you only need a simple keyword search, or if your dataset consists purely of text documents without corresponding images.
Stars
2,733
Forks
240
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Aug 15, 2025
Commits (30d)
0
Dependencies
28
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/rom1504/clip-retrieval"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
unum-cloud/UForm
Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts,...
mazzzystar/Queryable
Run OpenAI's CLIP and Apple's MobileCLIP model on iOS to search photos.
s-emanuilov/litepali
LitePali is a minimal, efficient implementation of ColPali for image retrieval and indexing,...
slavabarkov/tidy
Offline semantic Text-to-Image and Image-to-Image search on Android powered by quantized...
cloudera/CML_AMP_Image_Analysis
Build a semantic search application with deep learning models.