open_clip and CoN-CLIP

Tool B is an ecosystem sibling to tool A, as it implements a specific method building upon the core open-source CLIP framework provided by tool A, rather than offering a competing or complementary base CLIP implementation itself.

open_clip
73
Verified
CoN-CLIP
39
Emerging
Maintenance 13/25
Adoption 15/25
Maturity 25/25
Community 20/25
Maintenance 6/25
Adoption 7/25
Maturity 16/25
Community 10/25
Stars: 13,496
Forks: 1,253
Downloads:
Commits (30d): 1
Language: Python
License:
Stars: 40
Forks: 4
Downloads:
Commits (30d): 0
Language: Python
License:
No risk flags
No Package No Dependents

About open_clip

mlfoundations/open_clip

An open source implementation of CLIP.

This project provides pre-trained models that understand both images and text, allowing you to connect what you see with descriptive phrases. You can input an image and a list of text descriptions to get back probabilities of which description best matches the image. This is ideal for researchers or developers building applications that need to categorize images based on natural language or search for images using text.

image-text-matching zero-shot-classification multimodal-search computer-vision natural-language-processing

About CoN-CLIP

jaisidhsingh/CoN-CLIP

Implementation of the "Learn No to Say Yes Better" paper.

This project offers an improved way to classify images and understand complex visual relationships by making vision-language models better at handling negative statements. It takes an image and a list of text descriptions (some of which might be negative, like 'this is NOT a cat') and outputs the likelihood that the image matches each description. This is useful for AI researchers and machine learning engineers who are working with advanced image recognition and multimodal understanding tasks.

AI research image classification natural language understanding computer vision multimodal AI

Scores updated daily from GitHub, PyPI, and npm data. How scores work