open_clip and AlphaCLIP

AlphaCLIP builds upon the open-source CLIP implementation by adding spatial attention mechanisms to focus on user-specified regions, making it an enhanced variant rather than a direct competitor.

open_clip
73
Verified
AlphaCLIP
43
Emerging
Maintenance 13/25
Adoption 15/25
Maturity 25/25
Community 20/25
Maintenance 2/25
Adoption 10/25
Maturity 16/25
Community 15/25
Stars: 13,496
Forks: 1,253
Downloads:
Commits (30d): 1
Language: Python
License:
Stars: 869
Forks: 58
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: Apache-2.0
No risk flags
Stale 6m No Package No Dependents

About open_clip

mlfoundations/open_clip

An open source implementation of CLIP.

This project provides pre-trained models that understand both images and text, allowing you to connect what you see with descriptive phrases. You can input an image and a list of text descriptions to get back probabilities of which description best matches the image. This is ideal for researchers or developers building applications that need to categorize images based on natural language or search for images using text.

image-text-matching zero-shot-classification multimodal-search computer-vision natural-language-processing

About AlphaCLIP

SunzeY/AlphaCLIP

[CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want

This tool helps creative professionals and researchers direct AI models to focus on specific parts of an image. By providing an image along with a mask highlighting an area of interest, the AI will prioritize that region when generating descriptions or creating new images. This is ideal for designers, marketers, or researchers working with visual content who need precise control over AI interpretations.

generative-art image-editing visual-content-creation AI-assisted-design image-analysis

Scores updated daily from GitHub, PyPI, and npm data. How scores work