imSanko/Image_Caption_Generator_With_Transformers
This repository contains code for generating captions for images using a Transformer-based model. The model used is the `VisionEncoderDecoderModel` from the Hugging Face Transformers library, specifically the `nlpconnect/vit-gpt2-image-captioning` model.
ArchivedNo commits in the last 6 months.
Stars
12
Forks
1
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Sep 02, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/imSanko/Image_Caption_Generator_With_Transformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
hydropix/AutoDescribe-Images
Tool to automatically generate text descriptions for images using Ollama vision models (LLaVA,...
SkAndMl/captiongpt
Image Captioning using ViT and GPT. Notebook version in the following link
adityakapole/Caption-Generator
Captionify is a versatile and user-friendly image captioning project that harnesses the power of...