Efficient-Multimodal-LLMs-Survey and Efficient-LLMs-Survey

Maintenance 2/25
Adoption 10/25
Maturity 9/25
Community 12/25
Maintenance 2/25
Adoption 10/25
Maturity 1/25
Community 18/25
Stars: 389
Forks: 21
Downloads:
Commits (30d): 0
Language:
License: Apache-2.0
Stars: 1,256
Forks: 98
Downloads:
Commits (30d): 0
Language:
License:
Stale 6m No Package No Dependents
No License Stale 6m No Package No Dependents

About Efficient-Multimodal-LLMs-Survey

swordlidev/Efficient-Multimodal-LLMs-Survey

Efficient Multimodal Large Language Models: A Survey

About Efficient-LLMs-Survey

AIoT-MLSys-Lab/Efficient-LLMs-Survey

[TMLR 2024] Efficient Large Language Models: A Survey

Provides a comprehensive taxonomy organizing efficient LLM techniques across model-centric (compression, architecture optimization, inference acceleration), data-centric (selection, curation), and framework-centric perspectives. Covers specific methods including quantization, pruning, low-rank adaptation, parameter-efficient fine-tuning (LoRA, adapters), speculative decoding, KV-cache optimization, and efficient attention mechanisms like grouped-query attention. Actively maintained repository with curated paper collection and taxonomy designed to be updated with emerging research.

Scores updated daily from GitHub, PyPI, and npm data. How scores work