rhysdg/vision-at-a-clip
Low-latency ONNX and TensorRT based zero-shot classification and detection with contrastive language-image pre-training based prompts
This tool helps non-technical users quickly analyze and understand what's in their images using natural language. You input images and text descriptions (like "a photo of a dog" or "spaceman"), and it tells you what objects are present or the likelihood of different descriptions matching the image. It's designed for anyone needing fast visual search, classification, or object detection without complex machine learning setup.
No commits in the last 6 months.
Use this if you need to rapidly identify objects, classify images, or search through visual content using simple text prompts, especially when you need high performance and low latency.
Not ideal if you require highly specialized object recognition that is not described by everyday language or if your images contain extremely fine-grained, niche details that require extensive pre-training on specific datasets.
Stars
44
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
Aug 31, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/rhysdg/vision-at-a-clip"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
mlfoundations/open_clip
An open source implementation of CLIP.
noxdafox/clipspy
Python CFFI bindings for the 'C' Language Integrated Production System CLIPS
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
moein-shariatnia/OpenAI-CLIP
Simple implementation of OpenAI CLIP model in PyTorch.
BioMedIA-MBZUAI/FetalCLIP
Official repository of FetalCLIP: A Visual-Language Foundation Model for Fetal Ultrasound Image Analysis