mxagar/image_captioning
Image captioning project with a deep learning model. Given an image, it is processed by an encoder CNN and then passed to a decoder RNN. The decoder outputs a text caption of the image content. The work is based on the paper "Show and Tell" by Vinyals et al. Pytorch is used for the implementation and the COCO dataset for training.
No commits in the last 6 months.
Stars
2
Forks
—
Language
Jupyter Notebook
License
—
Category
Last pushed
Oct 14, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/mxagar/image_captioning"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ntrang086/image_captioning
generate captions for images using a CNN-RNN model that is trained on the Microsoft Common...
fregu856/CS224n_project
Neural Image Captioning in TensorFlow.
vacancy/SceneGraphParser
A python toolkit for parsing captions (in natural language) into scene graphs (as symbolic...
ltguo19/VSUA-Captioning
Code for "Aligning Linguistic Words and Visual Semantic Units for Image Captioning", ACM MM 2019
Abdelrhman-Yasser/video-content-description
Video content description model for generating descriptions for unconstrained videos