Llm Knowledge Distillation Transformer Models
There are 28 llm knowledge distillation models tracked. 1 score above 50 (established tier). The highest-rated is scaleapi/llm-engine at 56/100 with 821 stars. 2 of the top 10 are actively maintained.
Get all 28 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=llm-knowledge-distillation&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
scaleapi/llm-engine
Scale LLM Engine public repository |
|
Established |
| 2 |
modelscope/easydistill
a toolkit on knowledge distillation for large language models |
|
Emerging |
| 3 |
AGI-Arena/MARS
The official implementation of MARS: Unleashing the Power of Variance... |
|
Emerging |
| 4 |
Wang-ML-Lab/bayesian-peft
Bayesian Low-Rank Adaptation of LLMs: BLoB [NeurIPS 2024] and TFB [NeurIPS 2025] |
|
Emerging |
| 5 |
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for... |
|
Emerging |
| 6 |
yifanzhang-pro/HLA
Official Project Page for HLA: Higher-order Linear Attention... |
|
Emerging |
| 7 |
sangmichaelxie/doremi
Pytorch implementation of DoReMi, a method for optimizing the data mixture... |
|
Emerging |
| 8 |
ZO-Bench/ZO-LLM
[ICML‘24] Official code for the paper "Revisiting Zeroth-Order Optimization... |
|
Emerging |
| 9 |
Liuhong99/Sophia
The official implementation of “Sophia: A Scalable Stochastic Second-order... |
|
Emerging |
| 10 |
yang-ai-lab/OSF-Open-Sleep-FM
OSF: On Pre-training and Scaling of Sleep Foundation Models |
|
Emerging |
| 11 |
ShiZhengyan/DePT
[ICLR 2024] This is the repository for the paper titled "DePT: Decomposed... |
|
Emerging |
| 12 |
YJiangcm/Lion
[EMNLP 2023] Lion: Adversarial Distillation of Proprietary Large Language Models |
|
Emerging |
| 13 |
golololologol/LLM-Distillery
A pipeline for LLM knowledge distillation |
|
Emerging |
| 14 |
shufangxun/LLaVA-MoD
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation |
|
Emerging |
| 15 |
horus-ai-labs/DistillFlow
Library for model distillation |
|
Emerging |
| 16 |
OatmealLiu/FineR
[ICLR'24] Democratizing Fine-grained Visual Recognition with Large Language Models |
|
Experimental |
| 17 |
Tebmer/Awesome-Knowledge-Distillation-of-LLMs
This repository collects papers for "A Survey on Knowledge Distillation of... |
|
Experimental |
| 18 |
Qwen-Applications/STAR
STAR: Similarity-guided Teacher-Assisted Refinement for Super-Tiny Function... |
|
Experimental |
| 19 |
ROIM1998/APT
[ICML'24 Oral] APT: Adaptive Pruning and Tuning Pretrained Language Models... |
|
Experimental |
| 20 |
pittisl/GreenTrainer
Code for paper "Towards Green AI in Fine-tuning Large Language Models via... |
|
Experimental |
| 21 |
iboing/CorDA
CorDA: Context-Oriented Decomposition Adaptation of Large Language Models... |
|
Experimental |
| 22 |
pdaicode/awesome-LLMs-finetuning
Collection of resources for finetuning Large Language Models (LLMs). |
|
Experimental |
| 23 |
teilomillet/retrain
a Python library that uses Reinforcement Learning (RL) to train LLMs. |
|
Experimental |
| 24 |
TamSiuhin/OPPU
Official Implementation of "Democratizing Large Language Models via... |
|
Experimental |
| 25 |
wshi83/MedAdapter
[EMNLP'24] MedAdapter: Efficient Test-Time Adaptation of Large Language... |
|
Experimental |
| 26 |
amazon-science/mezo_svrg
Code the ICML 2024 paper: "Variance-reduced Zeroth-Order Methods for... |
|
Experimental |
| 27 |
BaohaoLiao/mefts
[NeurIPS 2023] Make Your Pre-trained Model Reversible: From Parameter to... |
|
Experimental |
| 28 |
kriskrisliu/PAT
[AAAI 2025] PAT: Pruning-Aware Tuning for Large Language Models |
|
Experimental |