Tebmer/Awesome-Knowledge-Distillation-of-LLMs
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
Organized taxonomically across Knowledge Elicitation (labeling, expansion, curation, feedback, self-knowledge) and Distillation Algorithms (SFT, divergence-based, RL, rank optimization), the collection maps how to extract and transfer both general capabilities and domain-specific skills. Covers practical applications across skill dimensions (instruction-following, alignment, agent behavior, task specialization, multimodality) and vertical domains (legal, medical, finance, science), alongside encoder-based KD approaches. Integrates with open-source LLM ecosystems like LLaMA and Mistral, addressing capability transfer from proprietary models (GPT-4, Claude) through synthetic data generation and self-improvement techniques.
1,264 stars. No commits in the last 6 months.
Stars
1,264
Forks
71
Language
—
License
—
Category
Last pushed
Mar 09, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Tebmer/Awesome-Knowledge-Distillation-of-LLMs"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
scaleapi/llm-engine
Scale LLM Engine public repository
AGI-Arena/MARS
The official implementation of MARS: Unleashing the Power of Variance Reduction for Training Large Models
modelscope/easydistill
a toolkit on knowledge distillation for large language models
Wang-ML-Lab/bayesian-peft
Bayesian Low-Rank Adaptation of LLMs: BLoB [NeurIPS 2024] and TFB [NeurIPS 2025]
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient...