LLM-Pruner and LLaMA-Pruning

LLM-Pruner is a generalized structural pruning framework that evolved from and supersedes the earlier LLaMA-Pruning project, extending the same pruning methodology across multiple model architectures beyond just LLaMA.

LLM-Pruner
47
Emerging
LLaMA-Pruning
33
Emerging
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 21/25
Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 9/25
Stars: 1,109
Forks: 130
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
Stars: 54
Forks: 4
Downloads:
Commits (30d): 0
Language: Python
License: GPL-3.0
Stale 6m No Package No Dependents
Archived Stale 6m No Package No Dependents

About LLM-Pruner

horseee/LLM-Pruner

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.

About LLaMA-Pruning

horseee/LLaMA-Pruning

Structural Pruning for LLaMA

Scores updated daily from GitHub, PyPI, and npm data. How scores work