bupt-ai-club/llm-compression-papers
papers of llm compression
This repository helps AI researchers and machine learning engineers keep up with the latest advancements in making large language models (LLMs) smaller and more efficient. It provides a curated collection of academic and industry papers on LLM compression techniques. You can find research papers and sometimes accompanying code to understand how to reduce the size and computational demands of LLMs.
No commits in the last 6 months.
Use this if you are an AI/ML researcher or practitioner looking for comprehensive resources on optimizing large language models for better performance and reduced resource usage.
Not ideal if you are an end-user simply looking to use an LLM and are not concerned with its underlying technical architecture or compression.
Stars
13
Forks
—
Language
—
License
Apache-2.0
Category
Last pushed
Mar 06, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/bupt-ai-club/llm-compression-papers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ModelTC/LightCompress
[EMNLP 2024 & AAAI 2026] A powerful toolkit for compressing large models including LLMs, VLMs,...
p-e-w/heretic
Fully automatic censorship removal for language models
Orion-zhen/abliteration
Make abliterated models with transformers, easy and fast
YerbaPage/LongCodeZip
LongCodeZip: Compress Long Context for Code Language Models [ASE2025]
locuslab/wanda
A simple and effective LLM pruning approach.