loading…
Search for a command to run...
loading…
Persistent AI memory for Claude Code, Cursor, GitHub Copilot & Windsurf — sessions, lessons learned, semantic search, and team brain. 38 MCP tools. Free tier, E
Persistent AI memory for Claude Code, Cursor, GitHub Copilot & Windsurf — sessions, lessons learned, semantic search, and team brain. 38 MCP tools. Free tier, EU servers.
Persistent memory for Claude Code, Cursor, GitHub Copilot & Windsurf.
Your AI remembers every lesson, every fix, every architecture decision — forever.
Every morning, you open your AI coding assistant. It doesn't remember yesterday.
You explain your architecture. You explain the deployment process. You explain the bug you fixed last week.
The average developer wastes 45 minutes/day re-establishing context. That's €15,000+ in lost productivity per engineer per year.
npx @cachly-dev/mcp-server@latest setup
The interactive wizard:
CLAUDE.md with memory rules pre-filledResult: Your AI remembers everything. Always.
Step 1 — Get your free credentials at cachly.dev (25 MB free, forever, no credit card).
Step 2 — Add to your editor's MCP config:
~/.claude/mcp.json or .mcp.json){
"mcpServers": {
"cachly": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@cachly-dev/mcp-server@latest"],
"env": {
"CACHLY_JWT": "your-jwt-token",
"CACHLY_INSTANCE_ID": "your-instance-id"
}
}
}
}
.cursor/mcp.json / .mcp.json){
"servers": {
"cachly": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@cachly-dev/mcp-server@latest"],
"env": {
"CACHLY_JWT": "your-jwt-token",
"CACHLY_INSTANCE_ID": "your-instance-id"
}
}
}
}
Step 3 — Add to CLAUDE.md / .github/copilot-instructions.md:
## AI Brain Rules (cachly)
- Call session_start BEFORE reading any files or making changes
- Call learn_from_attempts AFTER every fix, deploy, or discovery
- Call session_end when closing the window
| Situation | Without cachly | With cachly |
|---|---|---|
| Session start | "What's your architecture?" | "Ready. 23 lessons, last session: deployed API." |
| Known bug hits again | Re-researches from scratch | "You fixed this on March 12, here's the exact command" |
| After holiday / team handoff | Context dead | Fully briefed in < 10 seconds |
| Repeated LLM patterns | Pays for each re-discovery | Cached by meaning, ~1,200 tokens saved per hit |
| New team member | Weeks to onboard | session_start gives full context instantly |
| Tool | What it does |
|---|---|
session_start |
Full briefing: last session summary, open failures, recent lessons, brain health |
session_end |
Save what you built, auto-extract lessons from summary + ambient git log |
session_handoff |
Hand off remaining tasks to next window, with context |
auto_learn_session |
Batch-learn from a list of observations |
sync_file_changes |
Sync changed files into brain index |
learn_from_attempts |
Store structured lessons after any fix, deploy, or discovery |
recall_best_solution |
Best known solution for a topic — with success/failure history |
recall_at |
Recall a lesson at a specific point in time |
remember_context |
Cache architecture findings, decisions, file summaries |
recall_context |
Get exact context by key (supports glob) |
list_remembered |
See all cached context entries |
forget_context |
Remove stale context |
smart_recall |
BM25+ full-text search across all brain data |
setup_ai_memory |
Interactive wizard to configure AI memory |
| Feature | How it works |
|---|---|
| Team Telepathy | Multi-dev Brain: the author param on learn_from_attempts shares fixes across the whole team. Every session_start shows colleagues' lessons. |
| Ambient Git Learning | Pass workspace_path to session_end — it auto-reads git log, stores recent commits as Brain lessons. Zero extra calls. |
| Memory Crystals | memory_crystalize distills all lessons into a compact Crystal injected at every session_start. AI arrives pre-loaded with team knowledge. |
| IQ Boost | brain_doctor reports iq_boost_pct — how much smarter your AI gets vs. baseline from cached lessons. |
| Tool | What it does |
|---|---|
list_instances |
List all your cache instances |
create_instance |
Spin up a new instance (free or paid) |
get_connection_string |
Get the redis:// URL for your app |
delete_instance |
Remove an instance |
get_real_time_stats |
Memory, hit rate, ops/sec |
| Tool | What it does |
|---|---|
cache_get / cache_set / cache_delete |
Standard cache operations |
cache_mget / cache_mset |
Bulk pipeline (single round-trip) |
cache_lock_acquire / cache_lock_release |
Distributed Redlock-lite |
cache_stream_set / cache_stream_get |
LLM token stream caching |
| Tool | What it does |
|---|---|
semantic_search |
Find cached entries by meaning (pgvector HNSW, EU-hosted) |
semantic_warmup |
Pre-warm cache with prompt/response pairs |
detect_namespace |
Auto-classify prompt into code/qa/summary/translation/creative |
| Tool | What it does |
|---|---|
team_learn / team_recall |
Share lessons across the team |
team_synthesize |
Consolidate multiple lessons into one authoritative version |
memory_crystalize |
Distill all lessons into a Crystal snapshot for instant team context |
brain_doctor |
Health check: lesson count, IQ boost %, open failures, quality score |
global_learn / global_recall |
Cross-project universal lessons |
publish_lesson / import_public_brain |
Share/import community knowledge |
trace_dependency |
Causal chain — find lessons affected by a dependency change |
list_orgs / create_org |
Manage team organizations |
invite_member / get_org_plan |
Invite a developer to your org by email |
| Tier | RAM | Price | Best for |
|---|---|---|---|
| Free | 25 MB | €0/mo forever | Dev & side projects |
| Dev | 200 MB | €19/mo | Individual developers |
| Pro | 900 MB | €49/mo | Teams |
| Speed | 900 MB + Dragonfly + Semantic Cache | €79/mo | AI-heavy workloads |
| Business | 7 GB | €199/mo | Scale-ups |
✅ All plans: German servers · GDPR-compliant · 99.9% SLA · No credit card for Free tier
| Variable | Default | Description |
|---|---|---|
CACHLY_JWT |
— | Required. Your API token from cachly.dev |
CACHLY_INSTANCE_ID |
— | Default instance UUID (optional if you pass per-call) |
CACHLY_API_URL |
https://api.cachly.dev |
Override for self-hosted |
CACHLY_NO_TELEMETRY |
unset | Set to 1 to disable anonymous usage pings |
CACHLY_NO_UPDATE_CHECK |
unset | Set to 1 to disable the version-check on startup |
| Package | What it does |
|---|---|
| @cachly-dev/mcp-server | ← you are here — AI Brain MCP tools for Claude Code, Cursor, Copilot, Windsurf |
| @cachly-dev/openclaw | JS/TS SDK — cut LLM costs 60–90% with semantic cache + BM25 in your own app, 3 lines of code |
| @cachly-dev/init | One-command setup wizard: auto-detects all your editors and writes the correct MCP config — npx @cachly-dev/init |
| @cachly-dev/cli | Terminal CLI — manage instances, cache data, brain lessons: npm i -g @cachly-dev/cli |
| @cachly-dev/sdk | Node.js/TypeScript SDK — use managed Redis, semantic cache & AI memory directly in your app |
Добавь это в claude_desktop_config.json и перезапусти Claude Desktop.
{
"mcpServers": {
"cachly-mcp-server": {
"command": "npx",
"args": []
}
}
}