LLM-Pruner and LLaMA-Pruning

LLM-Pruner is a generalized structural pruning framework that evolved from and supersedes the earlier LLaMA-Pruning project, extending the same pruning methodology across multiple model architectures beyond just LLaMA.

LLM-Pruner
47
Emerging
LLaMA-Pruning
33
Emerging
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 21/25
Maintenance 0/25
Adoption 8/25
Maturity 16/25
Community 9/25
Stars: 1,109
Forks: 130
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
Stars: 54
Forks: 4
Downloads:
Commits (30d): 0
Language: Python
License: GPL-3.0
Stale 6m No Package No Dependents
Archived Stale 6m No Package No Dependents

About LLM-Pruner

horseee/LLM-Pruner

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.

This project helps machine learning engineers and researchers reduce the size of large language models (LLMs) like Llama, BLOOM, and Vicuna. By taking an existing LLM as input, it prunes unnecessary components while aiming to maintain its multi-task abilities. The output is a smaller, more efficient LLM that uses less computational resources, allowing for easier deployment and faster inference.

Large Language Models Model Compression Deep Learning Deployment AI Efficiency Resource Optimization

About LLaMA-Pruning

horseee/LLaMA-Pruning

Structural Pruning for LLaMA

Scores updated daily from GitHub, PyPI, and npm data. How scores work