๐ง cachly AI Brain โ MCP Server
Persistent memory for Claude Code, Cursor, GitHub Copilot & Windsurf. Your AI remembers every lesson, every fix, every architecture decision โ forever.
The Problem
Every morning, you open your AI coding assistant. It doesn't remember yesterday. You explain your architecture. You explain the deployment process. You explain the bug you fixed last week.
The average developer wastes 45 minutes/day re-establishing context. That's โฌ15,000+ in lost productivity per engineer per year.
The Fix โ One Command
npx @cachly-dev/mcp-server@latest setup
The interactive wizard:
- Signs you in (free, no credit card required)
- Picks or creates your AI Brain instance
- Auto-detects Cursor, Windsurf, VS Code, Claude Code, Continue.dev
- Writes the correct MCP config for every detected editor
- Creates
CLAUDE.mdwith memory rules pre-filled
Result: Your AI remembers everything. Always.
Quick Start (Manual)
Step 1 โ Get your free credentials at cachly.dev (25 MB free, forever, no credit card).
Step 2 โ Add to your editor's MCP config:
Claude Code (~/.claude/mcp.json or .mcp.json)
{
"mcpServers": {
"cachly": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@cachly-dev/mcp-server"],
"env": {
"CACHLY_JWT": "your-jwt-token",
"CACHLY_INSTANCE_ID": "your-instance-id"
}
}
}
}
Cursor / Windsurf / VS Code (.cursor/mcp.json / .mcp.json)
{
"servers": {
"cachly": {
"type": "stdio",
"command": "npx",
"args": ["-y", "@cachly-dev/mcp-server"],
"env": {
"CACHLY_JWT": "your-jwt-token",
"CACHLY_INSTANCE_ID": "your-instance-id"
}
}
}
}
Step 3 โ Add to CLAUDE.md / .github/copilot-instructions.md:
## AI Brain Rules (cachly)
- Call session_start BEFORE reading any files or making changes
- Call learn_from_attempts AFTER every fix, deploy, or discovery
- Call session_end when closing the window
With vs. Without cachly
| Situation | Without cachly | With cachly |
|---|---|---|
| Session start | "What's your architecture?" | "Ready. 23 lessons, last session: deployed API." |
| Known bug hits again | Re-researches from scratch | "You fixed this on March 12, here's the exact command" |
| After holiday / team handoff | Context dead | Fully briefed in < 10 seconds |
| Repeated LLM patterns | Pays for each re-discovery | Cached by meaning, ~1,200 tokens saved per hit |
| New team member | Weeks to onboard | session_start gives full context instantly |
38 MCP Tools
๐ง Session & Memory (most used)
| Tool | What it does |
|---|---|
session_start |
Full briefing: last session summary, open failures, recent lessons, brain health |
session_end |
Save what you built, auto-extract lessons from summary |
learn_from_attempts |
Store structured lessons after any fix, deploy, or discovery |
recall_best_solution |
Best known solution for a topic โ with success/failure history |
remember_context |
Cache architecture findings, decisions, file summaries |
smart_recall |
BM25+ full-text search across all brain data |
session_handoff |
Hand off remaining tasks to next window, with context |
โ๏ธ Instance Management
| Tool | What it does |
|---|---|
list_instances |
List all your cache instances |
create_instance |
Spin up a new instance (free or paid) |
get_connection_string |
Get the redis:// URL for your app |
delete_instance |
Remove an instance |
get_real_time_stats |
Memory, hit rate, ops/sec |
๐๏ธ Cache Operations
| Tool | What it does |
|---|---|
cache_get / cache_set / cache_delete |
Standard cache operations |
cache_mget / cache_mset |
Bulk pipeline (single round-trip) |
cache_lock_acquire / cache_lock_release |
Distributed Redlock-lite |
cache_stream_set / cache_stream_get |
LLM token stream caching |
๐ Semantic Cache
| Tool | What it does |
|---|---|
semantic_search |
Find cached entries by meaning (pgvector HNSW, EU-hosted) |
semantic_warmup |
Pre-warm cache with prompt/response pairs |
detect_namespace |
Auto-classify prompt into code/qa/summary/translation/creative |
๐ฅ Team Brain
| Tool | What it does |
|---|---|
team_learn / team_recall |
Share lessons across the team |
global_learn / global_recall |
Cross-project universal lessons |
list_orgs / create_org |
Manage team organizations |
invite_member |
Invite a developer to your org by email |
Pricing
| Tier | RAM | Price | Best for |
|---|---|---|---|
| Free | 25 MB | โฌ0/mo forever | Dev & side projects |
| Dev | 200 MB | โฌ19/mo | Individual developers |
| Pro | 900 MB | โฌ49/mo | Teams |
| Speed | 900 MB + Dragonfly + Semantic Cache | โฌ79/mo | AI-heavy workloads |
| Business | 7 GB | โฌ199/mo | Scale-ups |
โ All plans: German servers ยท GDPR-compliant ยท 99.9% SLA ยท No credit card for Free tier
Environment Variables
| Variable | Default | Description |
|---|---|---|
CACHLY_JWT |
โ | Required. Your API token from cachly.dev |
CACHLY_INSTANCE_ID |
โ | Default instance UUID (optional if you pass per-call) |
CACHLY_API_URL |
https://api.cachly.dev |
Override for self-hosted |
CACHLY_NO_TELEMETRY |
unset | Set to 1 to disable anonymous usage pings |
CACHLY_NO_UPDATE_CHECK |
unset | Set to 1 to disable the version-check on startup |
Links
- ๐ cachly.dev โ Dashboard & free signup
- ๐ AI Brain docs โ Full documentation
- ๐ฌ GitHub Issues โ Bug reports & feature requests
- โญ Star on GitHub โ If cachly saves you time, a star means a lot!
- ๐ฆ npm