huggingface/optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
Supports hardware-specific backends including ONNX Runtime, OpenVINO, TensorRT-LLM, AWS Neuron, and Intel Gaudi through modular installations, enabling optimized inference across diverse accelerators. Provides unified APIs for model export, quantization, and graph optimization while maintaining compatibility with PyTorch, enabling deployment from research to production without refactoring model code.
3,325 stars and 1,613,657 monthly downloads. Used by 29 other packages. Actively maintained with 4 commits in the last 30 days. Available on PyPI.
Stars
3,325
Forks
624
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 12, 2026
Monthly downloads
1,613,657
Commits (30d)
4
Dependencies
5
Reverse dependents
29
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/huggingface/optimum"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
openvinotoolkit/nncf
Neural Network Compression Framework for enhanced OpenVINOâ„¢ inference
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
RBLN-SW/optimum-rbln
âš¡ A seamless integration of HuggingFace Transformers & Diffusers with RBLN SDK for efficient...
eole-nlp/eole
Open language modeling toolkit based on PyTorch