AIoT-MLSys-Lab/Efficient-LLMs-Survey
[TMLR 2024] Efficient Large Language Models: A Survey
Provides a comprehensive taxonomy organizing efficient LLM techniques across model-centric (compression, architecture optimization, inference acceleration), data-centric (selection, curation), and framework-centric perspectives. Covers specific methods including quantization, pruning, low-rank adaptation, parameter-efficient fine-tuning (LoRA, adapters), speculative decoding, KV-cache optimization, and efficient attention mechanisms like grouped-query attention. Actively maintained repository with curated paper collection and taxonomy designed to be updated with emerging research.
1,256 stars. No commits in the last 6 months.
Stars
1,256
Forks
98
Language
—
License
—
Category
Last pushed
Jun 23, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AIoT-MLSys-Lab/Efficient-LLMs-Survey"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
USC-FORTIS/AD-LLM
[ACL Findings 2025] A benchmark for anomaly detection using large language models. It supports...
FairyFali/SLMs-Survey
Survey of Small Language Models from Penn State, ...
swordlidev/Efficient-Multimodal-LLMs-Survey
Efficient Multimodal Large Language Models: A Survey
Wang-ML-Lab/llm-continual-learning-survey
[CSUR 2025] Continual Learning of Large Language Models: A Comprehensive Survey
wang2226/Awesome-LLM-Decoding
📜 Paper list on decoding methods for LLMs and LVLMs