ThilinaRajapakse/simpletransformers
Transformers for Information Retrieval, Text Classification, NER, QA, Language Modelling, Language Generation, T5, Multi-Modal, and Conversational AI
Wraps HuggingFace Transformers with task-specific model classes that standardize the train/eval/predict workflow across NLP and multi-modal applications. Built-in integrations with Weights & Biases enable experiment tracking, while support for any HuggingFace pretrained model (BERT, RoBERTa, T5, etc.) provides flexibility without lock-in. Dense retrieval, conversational AI, and encoder fine-tuning extend beyond typical classification pipelines.
4,234 stars and 52,813 monthly downloads. Used by 4 other packages. No commits in the last 6 months. Available on PyPI.
Stars
4,234
Forks
721
Language
Python
License
Apache-2.0
Category
Last pushed
Aug 25, 2025
Monthly downloads
52,813
Commits (30d)
0
Dependencies
16
Reverse dependents
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ThilinaRajapakse/simpletransformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
explosion/curated-transformers
🤖 A PyTorch library of curated Transformer models and their composable components
jsksxs360/How-to-use-Transformers
Transformers 库快速入门教程
google/deepconsensus
DeepConsensus uses gap-aware sequence transformers to correct errors in Pacific Biosciences...
ruanchaves/hashformers
Accurate word segmentation for hashtags and text, powered by Transformers and Beam Search. A...
EricFillion/happy-transformer
Happy Transformer makes it easy to fine-tune and perform inference with NLP Transformer models.