Your AI Finally Remembers You
β‘ Created & Architected by Varun Pratap Bhardwaj β‘
Solution Architect β’ Original Creator β’ 2026
Stop re-explaining your codebase every session. 100% local. Zero setup. Completely free.
superlocalmemory.com β’ Quick Start β’ Why This? β’ Features β’ vs Alternatives β’ Docs β’ Issues
Created by Varun Pratap Bhardwaj β’ π Sponsor β’ π Attribution Required
SuperLocalMemory is now production-hardened with trust enforcement, rate limiting, and accelerated graph building.
| What's New in v2.6 | Why It Matters |
|---|---|
| Trust Enforcement | Agents with trust below 0.3 are blocked from write/delete β Bayesian scoring now actively protects your memory. |
| Profile Isolation | Memory profiles are fully sandboxed β no cross-profile data leakage. |
| Rate Limiting | Protects against memory flooding and spam from misbehaving agents. |
| SSRF Protection | Webhook dispatcher validates URLs to prevent server-side request forgery. |
| HNSW-Accelerated Graphs | Knowledge graph edge building uses HNSW index for faster construction at scale. |
| Hybrid Search Engine | Combined semantic + FTS5 + graph retrieval for maximum accuracy. |
v2.5 highlights (included): Real-time event stream, WAL-mode concurrent writes, agent tracking, memory provenance, 28 API endpoints.
Upgrade: npm install -g superlocalmemory@latest
Dashboard: python3 ~/.claude-memory/ui_server.py then open http://localhost:8765
Interactive Architecture Diagram | Architecture Doc | Full Changelog
Use SuperLocalMemory as a memory backend in your LangChain and LlamaIndex applications β 100% local, zero cloud.
pip install langchain-superlocalmemoryfrom langchain_superlocalmemory import SuperLocalMemoryChatMessageHistory
from langchain_core.runnables.history import RunnableWithMessageHistory
history = SuperLocalMemoryChatMessageHistory(session_id="my-session")
# Messages persist across sessions, stored locally in ~/.claude-memory/memory.dbpip install llama-index-storage-chat-store-superlocalmemoryfrom llama_index.storage.chat_store.superlocalmemory import SuperLocalMemoryChatStore
from llama_index.core.memory import ChatMemoryBuffer
chat_store = SuperLocalMemoryChatStore()
memory = ChatMemoryBuffer.from_defaults(chat_store=chat_store, chat_store_key="user-1")LangChain Guide | LlamaIndex Guide
npm install -g superlocalmemoryOr clone manually:
git clone https://github.com/varun369/SuperLocalMemoryV2.git && cd SuperLocalMemoryV2 && ./install.shBoth methods auto-detect and configure 17+ IDEs and AI tools β Cursor, VS Code/Copilot, Codex, Claude, Windsurf, Gemini CLI, JetBrains, and more.
Every time you start a new Claude session:
You: "Remember that authentication bug we fixed last week?"
Claude: "I don't have access to previous conversations..."
You: *sighs and explains everything again*
AI assistants forget everything between sessions. You waste time re-explaining your:
- Project architecture
- Coding preferences
- Previous decisions
- Debugging history
# Install in one command
npm install -g superlocalmemory
# Save a memory
superlocalmemoryv2:remember "Fixed auth bug - JWT tokens were expiring too fast, increased to 24h"
# Later, in a new session...
superlocalmemoryv2:recall "auth bug"
# β Found: "Fixed auth bug - JWT tokens were expiring too fast, increased to 24h"Your AI now remembers everything. Forever. Locally. For free.
npm install -g superlocalmemorygit clone https://github.com/varun369/SuperLocalMemoryV2.git
cd SuperLocalMemoryV2
./install.shgit clone https://github.com/varun369/SuperLocalMemoryV2.git
cd SuperLocalMemoryV2
.\install.ps1superlocalmemoryv2:status
# β Database: OK (0 memories)
# β Graph: Ready
# β Patterns: ReadyThat's it. No Docker. No API keys. No cloud accounts. No configuration.
npm users:
# Update to latest version
npm update -g superlocalmemory
# Or force latest
npm install -g superlocalmemory@latest
# Install specific version
npm install -g superlocalmemory@latestManual install users:
cd SuperLocalMemoryV2
git pull origin main
./install.sh # Mac/Linux
# or
.\install.ps1 # WindowsYour data is safe: Updates preserve your database and all memories.
# Launch the interactive web UI
python3 ~/.claude-memory/ui_server.py
# Opens at http://localhost:8765
# Features: Timeline view, search explorer, graph visualizationNEW in v2.2.0: Interactive web-based dashboard for exploring your memories visually.
| Feature | Description |
|---|---|
| π Timeline View | See your memories chronologically with importance indicators |
| π Search Explorer | Real-time semantic search with score visualization |
| πΈοΈ Graph Visualization | Interactive knowledge graph with clusters and relationships |
| π Statistics Dashboard | Memory trends, tag clouds, pattern insights |
| π― Advanced Filters | Filter by tags, importance, date range, clusters |
# 1. Start dashboard
python ~/.claude-memory/ui_server.py
# 2. Navigate to http://localhost:8765
# 3. Explore your memories:
# - Timeline: See memories over time
# - Search: Find with semantic scoring
# - Graph: Visualize relationships
# - Stats: Analyze patterns[[Complete Dashboard Guide β|Visualization-Dashboard]]
| Feature | Description |
|---|---|
| Hierarchical Leiden | Recursive community detection β clusters within clusters up to 3 levels. "Python" β "FastAPI" β "Auth patterns" |
| Community Summaries | TF-IDF structured reports per cluster: key topics, projects, categories at a glance |
| MACLA Confidence | Bayesian Beta-Binomial scoring (arXiv:2512.18950) β calibrated confidence, not raw frequency |
| Auto-Backup | Configurable SQLite backups with retention policies, restore from any backup via CLI |
| Profile UI | Create, switch, delete profiles from the web dashboard β full isolation per context |
| Profile Isolation | All API endpoints (graph, clusters, patterns, timeline) scoped to active profile |
SuperLocalMemory V2.2.0 implements hybrid search combining multiple strategies for maximum accuracy.
| Strategy | Method | Best For |
|---|---|---|
| Semantic Search | TF-IDF vectors + cosine similarity | Conceptual queries ("authentication patterns") |
| Full-Text Search | SQLite FTS5 with ranking | Exact phrases ("JWT tokens expire") |
| Graph-Enhanced | Knowledge graph traversal | Related concepts ("show auth-related") |
| Hybrid Mode | All three combined | General queries (default) |
# Semantic: finds conceptually similar
slm recall "security best practices"
# Matches: "JWT implementation", "OAuth flow", "CSRF protection"
# Exact: finds literal text
slm recall "PostgreSQL 15"
# Matches: exactly "PostgreSQL 15"
# Graph: finds related via clusters
slm recall "authentication" --use-graph
# Matches: JWT, OAuth, sessions (via "Auth & Security" cluster)
# Hybrid: best of all worlds (default)
slm recall "API design patterns"
# Combines semantic + exact + graph for optimal results| Database Size | Median | P95 | P99 |
|---|---|---|---|
| 100 memories | 10.6ms | 14.9ms | 15.8ms |
| 500 memories | 65.2ms | 101.7ms | 112.5ms |
| 1,000 memories | 124.3ms | 190.1ms | 219.5ms |
For typical personal databases (under 500 memories), search returns faster than you blink. Full benchmarks β
All numbers measured on real hardware (Apple M4 Pro, 24GB RAM). No estimates β real benchmarks.
| Database Size | Median Latency | P95 Latency |
|---|---|---|
| 100 memories | 10.6ms | 14.9ms |
| 500 memories | 65.2ms | 101.7ms |
| 1,000 memories | 124.3ms | 190.1ms |
For typical personal use (under 500 memories), search results return faster than you blink.
| Scenario | Writes/sec | Errors |
|---|---|---|
| 1 AI tool writing | 204/sec | 0 |
| 2 AI tools simultaneously | 220/sec | 0 |
| 5 AI tools simultaneously | 130/sec | 0 |
| 10 AI tools simultaneously | 25/sec | 0 |
WAL mode + serialized write queue = zero "database is locked" errors, ever.
10,000 memories = 13.6 MB on disk (~1.9 KB per memory). Your entire AI memory history takes less space than a photo.
Bayesian trust scoring achieves perfect separation (trust gap = 1.0) between honest and malicious agents. Detects "sleeper" attacks with 74.7% trust drop. Zero false positives.
| Memories | Build Time |
|---|---|
| 100 | 0.28s |
| 1,000 | 10.6s |
Leiden clustering discovers 6-7 natural topic communities automatically.
Graph Scaling: Knowledge graph features work best with up to 10,000 memories. For larger databases, the system uses intelligent sampling (most recent + highest importance memories) for graph construction. Core search and memory storage have no upper limit.
LoCoMo benchmark results coming soon β evaluation against the standardized LoCoMo long-conversation memory benchmark (Snap Research, ACL 2024).
SuperLocalMemory V2 is the ONLY memory system that works across ALL your tools:
| Tool | Integration | How It Works |
|---|---|---|
| Claude Code | β Skills + MCP | /superlocalmemoryv2:remember |
| Cursor | β MCP + Skills | AI uses memory tools natively |
| Windsurf | β MCP + Skills | Native memory access |
| Claude Desktop | β MCP | Built-in support |
| OpenAI Codex | β MCP + Skills | Auto-configured (TOML) |
| VS Code / Copilot | β MCP + Skills | .vscode/mcp.json |
| Continue.dev | β MCP + Skills | /slm-remember |
| Cody | β Custom Commands | /slm-remember |
| Gemini CLI | β MCP + Skills | Native MCP + skills |
| JetBrains IDEs | β MCP | Via AI Assistant settings |
| Zed Editor | β MCP | Native MCP tools |
| OpenCode | β MCP | Native MCP tools |
| Perplexity | β MCP | Native MCP tools |
| Antigravity | β MCP + Skills | Native MCP tools |
| ChatGPT | β MCP Connector | search() + fetch() via HTTP tunnel |
| Aider | β Smart Wrapper | aider-smart with context |
| Any Terminal | β Universal CLI | slm remember "content" |
-
MCP (Model Context Protocol) - Auto-configured for Cursor, Windsurf, Claude Desktop
- AI assistants get natural access to your memory
- No manual commands needed
- "Remember that we use FastAPI" just works
-
Skills & Commands - For Claude Code, Continue.dev, Cody
/superlocalmemoryv2:rememberin Claude Code/slm-rememberin Continue.dev and Cody- Familiar slash command interface
-
Universal CLI - Works in any terminal or script
slm remember "content"- Simple, clean syntaxslm recall "query"- Search from anywhereaider-smart- Aider with auto-context injection
All three methods use the SAME local database. No data duplication, no conflicts.
Installation automatically detects and configures:
- Existing IDEs (Cursor, Windsurf, VS Code)
- Installed tools (Aider, Continue, Cody)
- Shell environment (bash, zsh)
Zero manual configuration required. It just works.
Want to use SuperLocalMemory in ChatGPT, Perplexity, Zed, or other MCP-compatible tools?
π Complete setup guide: docs/MCP-MANUAL-SETUP.md
Covers:
- ChatGPT Desktop - Add via Settings β MCP
- Perplexity - Configure via app settings
- Zed Editor - JSON configuration
- Cody - VS Code/JetBrains setup
- Custom MCP clients - Python/HTTP integration
All tools connect to the same local database - no data duplication.
| Scenario | Without Memory | With SuperLocalMemory |
|---|---|---|
| New Claude session | Re-explain entire project | recall "project context" β instant context |
| Debugging | "We tried X last week..." starts over | Knowledge graph shows related past fixes |
| Code preferences | "I prefer React..." every time | Pattern learning knows your style |
| Multi-project | Context constantly bleeds | Separate profiles per project |
Not another simple key-value store. SuperLocalMemory implements cutting-edge memory architecture:
- PageIndex (Meta AI) β Hierarchical memory organization
- GraphRAG (Microsoft) β Knowledge graph with auto-clustering
- xMemory (Stanford) β Identity pattern learning
- A-RAG β Multi-level retrieval with context awareness
The only open-source implementation combining all four approaches.
| Solution | Free Tier Limits | Paid Price | What's Missing |
|---|---|---|---|
| Mem0 | 10K memories, limited API | Usage-based | No pattern learning, not local |
| Zep | Limited credits | $50/month | Credit system, cloud-only |
| Supermemory | 1M tokens, 10K queries | $19-399/mo | Not local, no graphs |
| Personal.AI | β No free tier | $33/month | Cloud-only, closed ecosystem |
| Letta/MemGPT | Self-hosted (complex) | TBD | Requires significant setup |
| SuperLocalMemory V2 | Unlimited | $0 forever | Nothing. |
| Feature | Mem0 | Zep | Khoj | Letta | SuperLocalMemory V2 |
|---|---|---|---|---|---|
| Works in Cursor | Cloud Only | β | β | β | β Local |
| Works in Windsurf | Cloud Only | β | β | β | β Local |
| Works in VS Code | 3rd Party | β | Partial | β | β Native |
| Works in Claude | β | β | β | β | β |
| Works with Aider | β | β | β | β | β |
| Universal CLI | β | β | β | β | β |
| 7-Layer Universal Architecture | β | β | β | β | β |
| Pattern Learning | β | β | β | β | β |
| Multi-Profile Support | β | β | β | Partial | β |
| Knowledge Graphs | β | β | β | β | β |
| 100% Local | β | β | Partial | Partial | β |
| Zero Setup | β | β | β | β | β |
| Progressive Compression | β | β | β | β | β |
| Completely Free | Limited | Limited | Partial | β | β |
SuperLocalMemory V2 is the ONLY solution that:
- β Works across 17+ IDEs and CLI tools
- β Remains 100% local (no cloud dependencies)
- β Completely free with unlimited memories
See full competitive analysis β
View Interactive Architecture Diagram β Click any layer for details, research references, and file paths.
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β Layer 9: VISUALIZATION (NEW v2.2.0) β
β Interactive dashboard: timeline, search, graph explorer β
β Real-time analytics and visual insights β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 8: HYBRID SEARCH (NEW v2.2.0) β
β Combines: Semantic + FTS5 + Graph traversal β
β 80ms response time with maximum accuracy β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 7: UNIVERSAL ACCESS β
β MCP + Skills + CLI (works everywhere) β
β 17+ IDEs with single database β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 6: MCP INTEGRATION β
β Model Context Protocol: 6 tools, 4 resources, 2 prompts β
β Auto-configured for Cursor, Windsurf, Claude β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 5: SKILLS LAYER β
β 6 universal slash-commands for AI assistants β
β Compatible with Claude Code, Continue, Cody β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 4: PATTERN LEARNING + MACLA (v2.4.0) β
β Bayesian Beta-Binomial confidence (arXiv:2512.18950) β
β "You prefer React over Vue" (73% confidence) β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 3: KNOWLEDGE GRAPH + HIERARCHICAL LEIDEN (v2.4.1) β
β Recursive clustering: "Python" β "FastAPI" β "Auth" β
β Community summaries + TF-IDF structured reports β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 2: HIERARCHICAL INDEX β
β Tree structure for fast navigation β
β O(log n) lookups instead of O(n) scans β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β Layer 1: RAW STORAGE β
β SQLite + Full-text search + TF-IDF vectors β
β Compression: 60-96% space savings β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
# Build the graph from your memories
python ~/.claude-memory/graph_engine.py build
# Output:
# β Processed 47 memories
# β Created 12 clusters:
# - "Authentication & Tokens" (8 memories)
# - "Performance Optimization" (6 memories)
# - "React Components" (11 memories)
# - "Database Queries" (5 memories)
# ...The graph automatically discovers relationships. Ask "what relates to auth?" and get JWT, session management, token refreshβeven if you never tagged them together.
# Learn patterns from your memories
python ~/.claude-memory/pattern_learner.py update
# Get your coding identity
python ~/.claude-memory/pattern_learner.py context 0.5
# Output:
# Your Coding Identity:
# - Framework preference: React (73% confidence)
# - Style: Performance over readability (58% confidence)
# - Testing: Jest + React Testing Library (65% confidence)
# - API style: REST over GraphQL (81% confidence)Your AI assistant can now match your preferences automatically.
MACLA Confidence Scoring (v2.4.0): Confidence uses a Bayesian Beta-Binomial posterior (Forouzandeh et al., arXiv:2512.18950). Pattern-specific priors, log-scaled competition, recency bonus. Range: 0.0β0.95 (hard cap prevents overconfidence).
# Work profile
superlocalmemoryv2:profile create work --description "Day job"
superlocalmemoryv2:profile switch work
# Personal projects
superlocalmemoryv2:profile create personal
superlocalmemoryv2:profile switch personal
# Client projects (completely isolated)
superlocalmemoryv2:profile create client-acmeEach profile has isolated memories, graphs, and patterns. No context bleeding.
| Guide | Description |
|---|---|
| Quick Start | Get running in 5 minutes |
| Installation | Detailed setup instructions |
| Visualization Dashboard | Interactive web UI guide (NEW v2.2.0) |
| CLI Reference | All commands explained |
| Knowledge Graph | How clustering works |
| Pattern Learning | Identity extraction |
| Profiles Guide | Multi-context management |
| API Reference | Python API documentation |
# Memory Operations
superlocalmemoryv2:remember "content" --tags tag1,tag2 # Save memory
superlocalmemoryv2:recall "search query" # Search
superlocalmemoryv2:list # Recent memories
superlocalmemoryv2:status # System health
# Profile Management
superlocalmemoryv2:profile list # Show all profiles
superlocalmemoryv2:profile create <name> # New profile
superlocalmemoryv2:profile switch <name> # Switch context
# Knowledge Graph
python ~/.claude-memory/graph_engine.py build # Build graph (+ hierarchical + summaries)
python ~/.claude-memory/graph_engine.py stats # View clusters
python ~/.claude-memory/graph_engine.py related --id 5 # Find related
python ~/.claude-memory/graph_engine.py hierarchical # Sub-cluster large communities
python ~/.claude-memory/graph_engine.py summaries # Generate cluster summaries
# Pattern Learning
python ~/.claude-memory/pattern_learner.py update # Learn patterns
python ~/.claude-memory/pattern_learner.py context 0.5 # Get identity
# Auto-Backup (v2.4.0)
python ~/.claude-memory/auto_backup.py backup # Manual backup
python ~/.claude-memory/auto_backup.py list # List backups
python ~/.claude-memory/auto_backup.py status # Backup status
# Reset (Use with caution!)
superlocalmemoryv2:reset soft # Clear memories
superlocalmemoryv2:reset hard --confirm # Nuclear option| Metric | Measured Result |
|---|---|
| Search latency | 10.6ms median (100 memories) |
| Concurrent writes | 220/sec with 2 agents, zero errors |
| Storage | 1.9 KB per memory at scale (13.6 MB for 10K) |
| Trust defense | 1.0 trust gap (perfect separation) |
| Graph build | 0.28s for 100 memories |
| Search quality | MRR 0.90 (first result correct 9/10 times) |
We welcome contributions! See CONTRIBUTING.md for guidelines.
Areas for contribution:
- Additional pattern categories
- Graph visualization UI
- Integration with more AI assistants
- Performance optimizations
- Documentation improvements
If SuperLocalMemory saves you time, consider supporting its development:
- β Star this repo β helps others discover it
- π Report bugs β open an issue
- π‘ Suggest features β start a discussion
- β Buy me a coffee β buymeacoffee.com/varunpratah
- πΈ PayPal β paypal.me/varunpratapbhardwaj
- π Sponsor β GitHub Sponsors
MIT License β use freely, even commercially. Just include the license.
Varun Pratap Bhardwaj β Solution Architect
Building tools that make AI actually useful for developers.
100% local. 100% private. 100% yours.