DRSY/MoTIS
[NAACL 2022]Mobile Text-to-Image search powered by multimodal semantic representation models(e.g., OpenAI's CLIP)
This project offers a way to quickly find images on your mobile device using descriptive text. You provide a phrase, and it searches your photos to bring up the most relevant matches, even if the exact words aren't in a caption. It's ideal for anyone who needs to locate specific images in their personal or professional photo library without tedious manual scrolling or tagging.
126 stars. No commits in the last 6 months.
Use this if you need to perform fast, semantic searches through a large collection of images on a mobile device by simply typing what you're looking for.
Not ideal if you primarily need to organize images by file metadata or require highly precise keyword matches rather than conceptual similarity.
Stars
126
Forks
10
Language
Swift
License
—
Category
Last pushed
May 11, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/DRSY/MoTIS"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
unum-cloud/UForm
Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts,...
rom1504/clip-retrieval
Easily compute clip embeddings and build a clip retrieval system with them
mazzzystar/Queryable
Run OpenAI's CLIP and Apple's MobileCLIP model on iOS to search photos.
s-emanuilov/litepali
LitePali is a minimal, efficient implementation of ColPali for image retrieval and indexing,...
slavabarkov/tidy
Offline semantic Text-to-Image and Image-to-Image search on Android powered by quantized...