quic/aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Supports both post-training quantization (PTQ) and quantization-aware training (QAT) with specialized techniques like AdaRound, SeqMSE, and Data-Free Quantization that minimize accuracy loss during INT8 conversion. Integrates directly with PyTorch and ONNX model pipelines, offering automated layer-wise optimization and compression techniques including spatial SVD and channel pruning. Provides visualization tools for weight ranges and per-layer compression sensitivity to guide optimization decisions.
2,566 stars. Actively maintained with 87 commits in the last 30 days.
Stars
2,566
Forks
448
Language
Python
License
—
Category
Last pushed
Mar 12, 2026
Commits (30d)
87
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/quic/aimet"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related frameworks
InterDigitalInc/CompressAI
A PyTorch library and evaluation platform for end-to-end compression research
tensorflow/compression
Data compression in TensorFlow
baler-collaboration/baler
Repository of Baler, a machine learning based data compression tool
Orange-OpenSource/Cool-Chic
Low-complexity neural image & video codec.
thulab/DeepHash
An Open-Source Package for Deep Learning to Hash (DeepHash)