October2001/Awesome-KV-Cache-Compression
๐ฐ Must-read papers on KV Cache Compression (constantly updating ๐ค).
This resource provides a curated collection of research papers and projects focused on optimizing the memory usage of Large Language Models (LLMs). It gathers various techniques to make LLMs run more efficiently, specifically by managing their 'KV Cache' โ a memory component crucial for generating responses. This helps AI researchers and practitioners identify and implement methods to reduce the computational demands and costs associated with deploying and operating LLMs.
668 stars.
Use this if you are a researcher, engineer, or practitioner working with Large Language Models and want to understand or implement methods to reduce their memory footprint and improve inference efficiency.
Not ideal if you are looking for a plug-and-play software solution or a general introduction to LLMs without a technical background in their architecture and optimization.
Stars
668
Forks
22
Language
—
License
MIT
Category
Last pushed
Feb 24, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/October2001/Awesome-KV-Cache-Compression"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ModelEngine-Group/unified-cache-management
Persist and reuse KV Cache to speedup your LLM.
reloadware/reloadium
Hot Reloading and Profiling for Python
alibaba/tair-kvcache
Alibaba Cloud's high-performance KVCache system for LLM inference, with components for global...
Zefan-Cai/Awesome-LLM-KV-Cache
Awesome-LLM-KV-Cache: A curated list of ๐Awesome LLM KV Cache Papers with Codes.
xcena-dev/maru
High-Performance KV Cache Storage Engine on CXL Shared Memory for LLM Inference