jaisidhsingh/CoN-CLIP
Implementation of the "Learn No to Say Yes Better" paper.
This project offers an improved way to classify images and understand complex visual relationships by making vision-language models better at handling negative statements. It takes an image and a list of text descriptions (some of which might be negative, like 'this is NOT a cat') and outputs the likelihood that the image matches each description. This is useful for AI researchers and machine learning engineers who are working with advanced image recognition and multimodal understanding tasks.
Use this if you need your image recognition models to more accurately interpret what an image is not, in addition to what it is, improving performance on tasks requiring nuanced understanding.
Not ideal if you are looking for a simple, out-of-the-box image classification tool without needing to integrate it into existing vision-language model pipelines.
Stars
40
Forks
4
Language
Python
License
—
Category
Last pushed
Oct 30, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/jaisidhsingh/CoN-CLIP"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
mlfoundations/open_clip
An open source implementation of CLIP.
noxdafox/clipspy
Python CFFI bindings for the 'C' Language Integrated Production System CLIPS
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
moein-shariatnia/OpenAI-CLIP
Simple implementation of OpenAI CLIP model in PyTorch.
BioMedIA-MBZUAI/FetalCLIP
Official repository of FetalCLIP: A Visual-Language Foundation Model for Fetal Ultrasound Image Analysis