Llm Compression Optimization Transformer Models
There are 29 llm compression optimization models tracked. 4 score above 50 (established tier). The highest-rated is ModelTC/LightCompress at 67/100 with 688 stars. 2 of the top 10 are actively maintained.
Get all 29 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=llm-compression-optimization&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
ModelTC/LightCompress
[EMNLP 2024 & AAAI 2026] A powerful toolkit for compressing large models... |
|
Established |
| 2 |
p-e-w/heretic
Fully automatic censorship removal for language models |
|
Established |
| 3 |
YerbaPage/LongCodeZip
LongCodeZip: Compress Long Context for Code Language Models [ASE2025] |
|
Established |
| 4 |
Orion-zhen/abliteration
Make abliterated models with transformers, easy and fast |
|
Established |
| 5 |
zyushun/Adam-mini
Code for Adam-mini: Use Fewer Learning Rates To Gain More... |
|
Emerging |
| 6 |
locuslab/wanda
A simple and effective LLM pruning approach. |
|
Emerging |
| 7 |
tommasomncttn/mergenetic
Flexible library for merging large language models (LLMs) via evolutionary... |
|
Emerging |
| 8 |
FMInference/FlexLLMGen
Running large language models on a single GPU for throughput-oriented scenarios. |
|
Emerging |
| 9 |
luuyin/OWL
Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity... |
|
Emerging |
| 10 |
ymcui/Chinese-Mixtral
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs) |
|
Emerging |
| 11 |
horseee/Awesome-Efficient-LLM
A curated list for Efficient Large Language Models |
|
Emerging |
| 12 |
BaiTheBest/SparseLLM
Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024) |
|
Emerging |
| 13 |
HOLYKEYZ/model-unfetter
The production engine for directional ablation. Unalign / remove models... |
|
Emerging |
| 14 |
jeffreysijuntan/lloco
The official repo for "LLoCo: Learning Long Contexts Offline" |
|
Emerging |
| 15 |
xuyang-liu16/GlobalCom2
[AAAI 2026] Global Compression Commander: Plug-and-Play Inference... |
|
Emerging |
| 16 |
arcee-ai/PruneMe
Automated Identification of Redundant Layer Blocks for Pruning in Large... |
|
Emerging |
| 17 |
asahi417/lm-vocab-trimmer
Vocabulary Trimming (VT) is a model compression technique, which reduces a... |
|
Emerging |
| 18 |
BauplanLabs/Making-Databases-Faster-with-LLM-Evolutionary-Sampling
Repository hosting code to reproduce our paper (with Stanford and... |
|
Emerging |
| 19 |
Koratahiu/Advanced_Optimizers
A family of highly efficient, lightweight yet powerful optimizers. |
|
Emerging |
| 20 |
Nota-NetsPresso/shortened-llm
Compressed LLMs for Efficient Text Generation [ICLR'24 Workshop] |
|
Emerging |
| 21 |
dmis-lab/Outlier-Safe-Pre-Training
[ACL 2025] Outlier-Safe Pre-Training for Robust 4-Bit Quantization of Large... |
|
Emerging |
| 22 |
jordddan/Pruning-LLMs
The framework to prune LLMs to any size and any config. |
|
Emerging |
| 23 |
whucs21Mzy/Model-Phase-Transitions
Navigating Model Phase Transitions to Enable Extreme Lossless Compression: A... |
|
Emerging |
| 24 |
Aaronhuang-778/SliM-LLM
[ICML 2025] SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large... |
|
Experimental |
| 25 |
Scientific-Computing-Lab/Tokompiler
Scope is all you need: Transforming LLMs for HPC Code |
|
Experimental |
| 26 |
OpenNLG/OpenBA-v2
OpenBA-V2: 3B LLM (Large Language Model) with T5 architecture, utilizing... |
|
Experimental |
| 27 |
liyucheng09/llm-compressive
Longitudinal Evaluation of LLMs via Data Compression |
|
Experimental |
| 28 |
JingyangXiang/DFRot
[COLM 2025] DFRot: Achieving Outlier-Free and Massive Activation-Free for... |
|
Experimental |
| 29 |
bupt-ai-club/llm-compression-papers
papers of llm compression |
|
Experimental |