TreeAI-Lab/Awesome-KV-Cache-Management

This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding code links.

33
/ 100
Emerging

This project is for developers who work with Large Language Models (LLMs) and need to improve their performance, particularly regarding memory usage and speed. It collects and categorizes research papers on "KV Cache Management," which is a technique to optimize how LLMs process information. The output is a curated list of research papers and their code, helping developers find methods to make their LLMs run faster and more efficiently.

291 stars.

Use this if you are an LLM developer or researcher looking for state-of-the-art techniques to accelerate LLM inference and manage memory more effectively.

Not ideal if you are an end-user of an LLM application and do not work directly with LLM infrastructure or model development.

Large Language Models LLM development AI model optimization AI research Machine Learning engineering
No License No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 9 / 25

How are scores calculated?

Stars

291

Forks

9

Language

License

Last pushed

Dec 05, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/TreeAI-Lab/Awesome-KV-Cache-Management"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.