tranquoctrinh/Image-Captioning-EfficientNet-Transformer
Image Captioning with Encoder as Efficientnet and Decoder as Decoder of Transformer combined with the attention mechanism.
This tool automatically generates descriptive text captions for images. It takes raw image files as input and outputs a human-readable sentence that describes the visual content of each image. This is useful for anyone needing to quickly label or organize large collections of photos, such as content managers, digital archivists, or e-commerce professionals.
No commits in the last 6 months.
Use this if you need to generate concise, accurate text descriptions for individual images or batches of images.
Not ideal if you need highly nuanced, subjective, or creative captions, or if you don't have access to image datasets for training.
Stars
7
Forks
3
Language
Python
License
MIT
Category
Last pushed
Apr 07, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/tranquoctrinh/Image-Captioning-EfficientNet-Transformer"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ntrang086/image_captioning
generate captions for images using a CNN-RNN model that is trained on the Microsoft Common...
fregu856/CS224n_project
Neural Image Captioning in TensorFlow.
vacancy/SceneGraphParser
A python toolkit for parsing captions (in natural language) into scene graphs (as symbolic...
ltguo19/VSUA-Captioning
Code for "Aligning Linguistic Words and Visual Semantic Units for Image Captioning", ACM MM 2019
Abdelrhman-Yasser/video-content-description
Video content description model for generating descriptions for unconstrained videos