jmisilo/clip-gpt-captioning
CLIPxGPT Captioner is Image Captioning Model based on OpenAI's CLIP and GPT-2.
This tool automatically generates descriptive captions for your images. You provide an image, and it outputs a concise sentence summarizing its content. It's designed for anyone who needs quick, accurate descriptions for visual assets, such as content creators, digital marketers, or photographers.
118 stars. No commits in the last 6 months.
Use this if you need to quickly generate text descriptions for a collection of images without manual effort.
Not ideal if you require highly specialized captions with nuanced context that a general image model might miss.
Stars
118
Forks
32
Language
Python
License
MIT
Category
Last pushed
Feb 17, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/jmisilo/clip-gpt-captioning"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
leaderj1001/CLIP
CLIP: Connecting Text and Image (Learning Transferable Visual Models From Natural Language Supervision)
PathologyFoundation/plip
Pathology Language and Image Pre-Training (PLIP) is the first vision and language foundation...
kesimeg/turkish-clip
OpenAI's clip model training for Turkish language using pretrained Resnet and DistilBERT
Lahdhirim/CV-image-captioning-clip-gpt2
Image caption generation using a hybrid CLIP-GPT2 architecture. CLIP encodes the image while...