vcache-project/vCache
Reliable and Efficient Semantic Prompt Caching with vCache
vCache helps you efficiently manage the costs and speed of your large language model (LLM) applications. It intelligently reuses past LLM responses for similar user requests, ensuring you don't pay for or wait for the same answer twice, even if the phrasing is slightly different. This tool is ideal for developers and engineers building applications that frequently interact with LLMs, such as AI-powered chatbots, data analysis tools, or content generation systems.
Use this if you need to reduce the operational cost and improve the response time of your LLM-powered applications while strictly controlling accuracy.
Not ideal if your application primarily uses LLMs for unique, non-repetitive tasks where caching similar prompts offers no benefit.
Stars
60
Forks
3
Language
Python
License
—
Category
Last pushed
Dec 17, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/vcache-project/vCache"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
RediSearch/RediSearch
A query and indexing engine for Redis, providing secondary indexing, full-text search, vector...
redis/redis-vl-python
Redis Vector Library (RedisVL) -- the AI-native Python client for Redis.
redis-developer/redis-ai-resources
✨ A curated list of awesome community resources, integrations, and examples of Redis in the AI ecosystem.
redis-developer/redis-product-search
Visual and semantic vector similarity with Redis Stack, FastAPI, PyTorch and Huggingface.
luyug/GradCache
Run Effective Large Batch Contrastive Learning Beyond GPU/TPU Memory Constraint