fellanH/context-vault
Persistent memory for AI agents — save and search knowledge across sessions via MCP. Local-first, markdown + SQLite + embeddings.
Implements hybrid full-text and semantic search via embeddings, with MCP tools for saving structured entry types (insights, decisions, patterns) and ingesting external content from URLs or projects. Runs as an auto-configured shared daemon that detects Claude, Cursor, and other AI tools, storing all data as plain markdown in `~/vault/` with SQLite indexing for search and optional web dashboard access.
Available on npm.
Stars
2
Forks
1
Language
JavaScript
License
—
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Dependencies
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/vector-db/fellanH/context-vault"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
topoteretes/cognee
Knowledge Engine for AI Agent Memory in 6 lines of code
CaviraOSS/OpenMemory
Local persistent memory store for LLM applications including claude desktop, github copilot,...
verygoodplugins/automem
AutoMem is a graph-vector memory service that gives AI assistants durable, relational memory:
CortexReach/memory-lancedb-pro
Enhanced LanceDB memory plugin for OpenClaw — Hybrid Retrieval (Vector + BM25), Cross-Encoder...
divagr18/memlayer
Plug-and-play memory for LLMs in 3 lines of code. Add persistent, intelligent, human-like memory...