image_captioning_with_transformers and pytorch-image-captioning

These are competitors—both implement standalone PyTorch solutions for transformer-based image captioning without dependency relationships, so a user would select one based on implementation details and code quality rather than using them together.

Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 14/25
Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 11/25
Stars: 68
Forks: 9
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
Stars: 44
Forks: 5
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About image_captioning_with_transformers

zarzouram/image_captioning_with_transformers

Pytorch implementation of image captioning using transformer-based model.

Implements an encoder-decoder transformer architecture with per-head attention visualization capabilities, modified from PyTorch's standard multi-head attention to enable detailed attention analysis. Trained on MS COCO 2017 with beam search inference and comprehensive NLG evaluation metrics (BLEU, METEOR, GLEU). Includes preprocessing pipeline for image-caption dataset creation with HDF5 storage and Tensorboard integration for training monitoring.

About pytorch-image-captioning

senadkurtisi/pytorch-image-captioning

Transformer & CNN Image Captioning model in PyTorch.

Scores updated daily from GitHub, PyPI, and npm data. How scores work