vcache-project/vCache

Reliable and Efficient Semantic Prompt Caching with vCache

37
/ 100
Emerging

vCache helps you efficiently manage the costs and speed of your large language model (LLM) applications. It intelligently reuses past LLM responses for similar user requests, ensuring you don't pay for or wait for the same answer twice, even if the phrasing is slightly different. This tool is ideal for developers and engineers building applications that frequently interact with LLMs, such as AI-powered chatbots, data analysis tools, or content generation systems.

Use this if you need to reduce the operational cost and improve the response time of your LLM-powered applications while strictly controlling accuracy.

Not ideal if your application primarily uses LLMs for unique, non-repetitive tasks where caching similar prompts offers no benefit.

LLM-ops AI-application-development cost-optimization performance-engineering chatbot-infrastructure
No Package No Dependents
Maintenance 6 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 7 / 25

How are scores calculated?

Stars

60

Forks

3

Language

Python

License

Last pushed

Dec 17, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/vcache-project/vCache"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.