Siddhant-K-code/tokenvm
TokenVM is a high-performance runtime that treats LLM KV cache and activations as a virtual memory working set across GPU VRAM β pinned host RAM β NVMe storage, with intelligent paging, prefetching, and compute-copy overlap.
This high-performance runtime helps developers working with large language models (LLMs) to use much longer text sequences than typically possible. It intelligently manages the model's working memory across GPU, host RAM, and NVMe storage. As input, you provide your existing LLM code, and it outputs an LLM that can process significantly longer contexts more efficiently.
No commits in the last 6 months.
Use this if you are an ML engineer or researcher who needs to run LLMs with very long context windows (e.g., 32,000 to 64,000 tokens) but are constrained by GPU memory.
Not ideal if you are working with smaller LLMs or short context windows where memory optimization is not a primary concern, or if you require a stable, production-ready solution without further testing.
Stars
9
Forks
—
Language
Go
License
MIT
Category
Last pushed
Aug 21, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Siddhant-K-code/tokenvm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ModelEngine-Group/unified-cache-management
Persist and reuse KV Cache to speedup your LLM.
reloadware/reloadium
Hot Reloading and Profiling for Python
alibaba/tair-kvcache
Alibaba Cloud's high-performance KVCache system for LLM inference, with components for global...
October2001/Awesome-KV-Cache-Compression
π° Must-read papers on KV Cache Compression (constantly updating π€).
Zefan-Cai/Awesome-LLM-KV-Cache
Awesome-LLM-KV-Cache: A curated list of πAwesome LLM KV Cache Papers with Codes.