You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
LEANN is a graph-based RAG system that achieves 97% storage savings through selective recomputation. Instead of storing all embedding vectors, it:
Stores a compressed knowledge graph (CSR format)
Computes embeddings on-demand during search using graph traversal
Uses HNSW or DiskANN indexing backends
Already supports MCP servers for live data integration
Potential Benefits
✅ Storage Efficiency: 97% reduction in storage requirements
✅ Privacy-First: Local data processing (aligns with our philosophy)
✅ Graph-Based Queries: Enable relationship-based memory retrieval
✅ Portability: Smaller data transfers for multi-device sync
✅ MCP Compatible: Designed to work with MCP protocol
Trade-offs & Concerns
❌ Performance Impact: On-demand computation vs our current 5ms pre-computed reads
❌ Query Latency: Trades storage for compute time during queries
❌ Complexity: Graph maintenance overhead vs current vector storage
❌ Feature Parity: Need to support tags, time-based search, metadata
❌ Migration: 2,530+ production memories, backward compatibility
Current Architecture Context
Our current backends optimize for query performance:
SQLite-vec: 5ms reads, single-file, <150MB memory
Cloudflare: Edge-deployed vector search
Hybrid: Fast local + background cloud sync (recommended)
LEANN optimizes for storage efficiency, which is a different trade-off.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
LEANN Integration Discussion
I'd like to open a discussion about potentially integrating LEANN (Learning-Enhanced Augmented Neural Network) as a storage backend option for mcp-memory-service.
What is LEANN?
LEANN is a graph-based RAG system that achieves 97% storage savings through selective recomputation. Instead of storing all embedding vectors, it:
Potential Benefits
✅ Storage Efficiency: 97% reduction in storage requirements
✅ Privacy-First: Local data processing (aligns with our philosophy)
✅ Graph-Based Queries: Enable relationship-based memory retrieval
✅ Portability: Smaller data transfers for multi-device sync
✅ MCP Compatible: Designed to work with MCP protocol
Trade-offs & Concerns
❌ Performance Impact: On-demand computation vs our current 5ms pre-computed reads
❌ Query Latency: Trades storage for compute time during queries
❌ Complexity: Graph maintenance overhead vs current vector storage
❌ Feature Parity: Need to support tags, time-based search, metadata
❌ Migration: 2,530+ production memories, backward compatibility
Current Architecture Context
Our current backends optimize for query performance:
LEANN optimizes for storage efficiency, which is a different trade-off.
Possible Integration Approaches
Option 1: Experimental 4th Backend
Option 2: Tiered Storage Architecture
Option 3: Document Ingestion Layer
Use LEANN specifically for compressing large document collections while keeping current backends for active memory.
Questions for Discussion
My Initial Assessment
This seems like a research prototype rather than immediate priority:
Community Input Needed
What do you think? Should we pursue LEANN integration? If so, which approach makes the most sense for your use cases?
Reference: https://github.com/yichuan-w/LEANN
Beta Was this translation helpful? Give feedback.
All reactions