messkan/prompt-cache

Cut LLM costs by up to 80% and unlock sub-millisecond responses with intelligent semantic caching.A drop-in, provider-agnostic LLM proxy written in Go with sub-millisecond response

47
/ 100
Emerging

PromptCache helps development teams reduce expenses and speed up applications that use large language models (LLMs). By sitting between your application and the LLM provider, it detects semantically similar user requests and serves cached answers instantly. This is ideal for developers building AI-powered applications that experience repetitive queries, like customer support bots or AI agents.

209 stars.

Use this if you are developing an application that uses LLMs and you frequently see similar user prompts resulting in duplicate, costly, and slow API calls to your LLM provider.

Not ideal if your application primarily handles unique, non-repetitive user prompts where caching would offer minimal benefit.

AI-application-development LLM-cost-optimization developer-tooling API-management
No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 13 / 25
Community 14 / 25

How are scores calculated?

Stars

209

Forks

19

Language

Go

License

MIT

Last pushed

Jan 25, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/messkan/prompt-cache"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.