jaketae/ensemble-transformers
Ensembling Hugging Face transformers made easy
Provides task-specific ensemble classes (e.g., `EnsembleModelForSequenceClassification`) that automatically detect and instantiate matching preprocessors for each backbone model, eliminating manual tokenizer management. Supports multi-device distribution via `to_multiple()` to load different models across GPUs, and includes modality validation to prevent mixing incompatible model types. Returns per-model predictions alongside aggregated outputs via mean pooling, enabling flexible inference strategies across heterogeneous Hugging Face transformer architectures.
No commits in the last 6 months. Available on PyPI.
Stars
61
Forks
5
Language
Python
License
MIT
Category
Last pushed
Dec 24, 2022
Monthly downloads
42
Commits (30d)
0
Dependencies
2
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/jaketae/ensemble-transformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
lvapeab/nmt-keras
Neural Machine Translation with Keras
dair-ai/Transformers-Recipe
🧠A study guide to learn about Transformers
SirawitC/Transformer_from_scratch_pytorch
Build a transformer model from scratch using pytorch to understand its inner workings and gain...
lof310/transformer
PyTorch implementation of the current SOTA Transformer. Configurable, efficient, and...
jiangtaoxie/SoT
SoT: Delving Deeper into Classification Head for Transformer