ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β β
β ββββββ βββ ββββββββββββ βββββββ ββββ ββββββββββββββββ ββββ β
β βββββββββββ ββββββββββββββββββββββββββ ββββββββββββββββββ βββββ β
β βββββββββββ βββ βββ βββ ββββββββββββββββββββ βββββββββββ β
β βββββββββββ βββ βββ βββ ββββββββββββββββββββ βββββββββββ β
β βββ ββββββββββββ βββ ββββββββββββ βββ ββββββββββββββ βββ βββ β
β βββ βββ βββββββ βββ βββββββ βββ ββββββββββββββ βββ β
β β
β State-of-the-Art Conversational Memory β’ 90.53% LoCoMo Score β
β Graph + Vector Architecture β’ Research-Validated β
β β
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
AutoMem is the world's best-performing long-term memory system for AI assistants. On December 2, 2025, AutoMem achieved 90.53% accuracy on the LoCoMo benchmark (ACL 2024), beating the previous state-of-the-art by 2.29 points.
Deploy production-grade AI memory in 60 seconds:
railway upAutoMem achieved 90.53% accuracy on the LoCoMo benchmark (ACL 2024)βthe academic standard for long-term conversational memory. This makes AutoMem the highest-performing memory system in the world, beating:
- CORE (heysol.ai): 88.24% (previous SOTA)
- OpenAI's baseline: 39%
- +14.45 percentage points improvement since baseline
Key breakthroughs:
- π Entity-to-entity expansion - Multi-hop reasoning via entity tag linking
- π Multi-hop bridge discovery - Finds connecting memories across conversation threads
- β° Temporal alignment scoring - Understands time-aware queries ("what happened last year?")
- π― 9-component hybrid search - Semantic + lexical + graph + temporal signals
- π― 100% complex reasoning - Perfect score on multi-step reasoning tasks
Your AI forgets everything between sessions. RAG dumps similar documents. Vector databases match keywords but miss meaning. None of them learn.
AutoMem gives AI assistants the ability to remember, connect, and evolve their understanding over timeβjust like human long-term memory.
December 2, 2025: AutoMem scored 90.53% on LoCoMoβthe academic benchmark for long-term conversational memory (ACL 2024). This beats:
- CORE (heysol.ai): 88.24% (previous SOTA)
- OpenAI's implementation: 39%
- +14.45 points improvement since baseline
AutoMem is a graph-vector memory service built on peer-reviewed neuroscience and validated by academic benchmarks:
- π§ Stores memories with metadata, importance scores, temporal context, and semantic embeddings
- π Recalls via 9-component hybrid search (vector + keyword + graph + temporal + lexical)
- π Connects memories with 11 typed relationships (RELATES_TO, LEADS_TO, CONTRADICTS, etc.)
- π― Learns through automatic entity extraction, pattern detection, and consolidation cycles
- π Reasons with multi-hop bridge discoveryβfinds connecting memories across conversation threads
- β‘ Performs with sub-100ms recall across thousands of memories
AutoMem implements breakthroughs from:
- HippoRAG 2 (Ohio State, 2025): Graph-vector hybrid matches human associative memory
- A-MEM (2025): Dynamic memory organization with Zettelkasten-inspired clustering
- MELODI (DeepMind, 2024): 8x compression without quality loss via gist representations
- ReadAgent (DeepMind, 2024): 20x context extension through episodic memory
We didn't just read the papers. We built the system they describeβand proved it works.
βββββββββββββββββββββββββββββββββββββββββββββββ
β AutoMem Service (Flask) β
β β’ REST API for memory lifecycle β
β β’ Background enrichment pipeline β
β β’ Consolidation engine β
β β’ Automated backups (optional) β
ββββββββββββββββ¬βββββββββββββββ¬ββββββββββββββββ
β β
ββββββββΌβββββββ βββββΌβββββββββ
β FalkorDB β β Qdrant β
β (Graph) β β (Vectors) β
β β β β
β β’ 11 edge β β β’ Semantic β
β types β β search β
β β’ Pattern β β β’ 768-d β
β nodes β β vectors β
βββββββββββββββ ββββββββββββββ
FalkorDB (graph) = canonical record, relationships, consolidation
Qdrant (vectors) = semantic recall, similarity search
Dual storage = Built-in redundancy and disaster recovery
Memory: "Chose PostgreSQL for reliability"
Query: "What database should I use?"
Result: β
Finds the memory
β Doesn't know WHY you chose it
β Can't connect to related decisions
Memory: "Chose PostgreSQL for reliability"
Graph: PREFERS_OVER MongoDB
RELATES_TO "team expertise" memory
DERIVED_FROM "boring technology" principle
Query: "What database should I use?"
Result: β
Finds the memory
β
Knows your decision factors
β
Shows related preferences
β
Explains your reasoning pattern
The innovation that pushed AutoMem to SOTA: path-based memory expansion that discovers "bridge" memories connecting disparate conversation threads.
User asks: "Why did we choose the boring tech approach for Kafka?"
Traditional RAG: Returns "Kafka" memories (misses the connection)
AutoMem bridge discovery:
- Seed 1: "Migrated to PostgreSQL for operational simplicity"
- Seed 2: "Evaluating Kafka vs RabbitMQ for message queue"
- Bridge: "Team prefers boring technologyβproven, debuggable systems"
AutoMem finds the bridge that connects both decisions
β Result: AI understands your architectural philosophy, not just isolated choices
Technical details:
- Graph traversal finds memories connecting multiple seed results
- Ranked by relation strength, temporal relevance, importance
- Configurable with
expand_paths=true(enabled by default) - Drives the 37.5% β ongoing improvement in multi-hop reasoning
# After storing: "Migrated to PostgreSQL for operational simplicity"
AutoMem automatically:
βββ Extracts entities (PostgreSQL, operational simplicity)
βββ Auto-tags (entity:tool:postgresql, entity:concept:ops-simplicity)
βββ Detects pattern ("prefers boring technology")
βββ Links temporally (PRECEDED_BY migration planning)
βββ Connects semantically (SIMILAR_TO "Redis deployment")
# Next query: "Should we use Kafka?"
AI recalls:
- PostgreSQL decision
- "Boring tech" pattern (reinforced across memories)
- Operational simplicity preference
β Suggests: "Based on your pattern, consider RabbitMQ instead"
GET /recall?query=database%20migration&tags=decision&time_query=last%20month
# AutoMem combines nine signals:
score = vectorΓ0.25 # Semantic similarity
+ keywordΓ0.15 # TF-IDF text matching
+ relationΓ0.25 # Graph relationship strength
+ contentΓ0.25 # Direct token overlap
+ temporalΓ0.15 # Time alignment with query
+ tagΓ0.10 # Tag matching
+ importanceΓ0.05 # User-assigned priority
+ confidenceΓ0.05 # Memory confidence
+ recencyΓ0.10 # Freshness boost
# Result: Memories ranked by meaning, not just similarity- Store - Rich memories with metadata, importance, timestamps, embeddings
- Recall - Hybrid search (vector + keyword + tags + time windows)
- Update - Modify memories, auto-regenerate embeddings
- Delete - Remove from both graph and vector stores
- Associate - Create typed relationships between memories
- Filter - Tag-based queries with prefix/exact matching
AutoMem uses neuroscience-inspired consolidation cyclesβlike human sleepβto keep memories relevant:
| Cycle | Frequency | Purpose |
|---|---|---|
| Decay | Hourly | Exponential relevance scoring (age, access, connections, importance) |
| Creative | Hourly | REM-like processing that discovers non-obvious connections |
| Cluster | 6 hours | Groups similar memories, generates meta-patterns |
| Forget | Daily | Archives low-relevance (<0.2), deletes very old (<0.05) |
How it works:
- Wrong rabbit holes fade naturally (~30-45 days without access)
- Important memories with strong connections stay indefinitely
- Memories archive before deletion (0.05-0.2 relevance range)
- The system learns what matters to you, not what you explicitly tag
Every memory gets automatically enhanced in the background (doesn't block your API calls):
Enrichment Pipeline (runs immediately after storage):
- Entity extraction - Identifies people, projects, tools, concepts (spaCy NLP)
- Auto-tagging - Generates
entity:<type>:<slug>for structured queries - Summaries - Lightweight gist representations for quick scanning
- Temporal links - Connects to recent memories with
PRECEDED_BYrelationships - Semantic neighbors - Finds similar memories via cosine similarity (
SIMILAR_TO) - Pattern detection - Reinforces emerging themes across your memory graph
Consolidation Engine (runs on configurable schedules):
- See Memory Consolidation section above
Build rich knowledge graphs:
| Type | Use Case | Example |
|---|---|---|
RELATES_TO |
General connection | Bug report β Related issue |
LEADS_TO |
Causal relationship | Problem β Solution |
OCCURRED_BEFORE |
Temporal sequence | Planning β Execution |
PREFERS_OVER |
User preferences | PostgreSQL β MongoDB |
EXEMPLIFIES |
Pattern examples | Code review β Best practice |
CONTRADICTS |
Conflicting info | Old approach β New approach |
REINFORCES |
Supporting evidence | Decision β Validation |
INVALIDATED_BY |
Outdated info | Legacy docs β Current docs |
EVOLVED_INTO |
Knowledge evolution | Initial design β Final design |
DERIVED_FROM |
Source tracking | Implementation β Spec |
PART_OF |
Hierarchical structure | Feature β Epic |
Deploy AutoMem + FalkorDB to Railway in 60 seconds:
# Install Railway CLI
npm i -g @railway/cli
# Deploy
railway login
railway init
railway upπ Deployment Guide for detailed Railway setup
Run everything locally:
# Clone and start services
git clone https://github.com/verygoodplugins/automem.git
cd automem
make dev
# API: http://localhost:8001
# FalkorDB: localhost:6379
# Qdrant: localhost:6333Run API without Docker:
python3 -m venv venv
source venv/bin/activate
pip install -r requirements-dev.txt
PORT=8001 python app.pycurl -X POST http://localhost:8001/memory \
-H "Authorization: Bearer YOUR_TOKEN" \
-H "Content-Type: application/json" \
-d '{
"content": "Chose PostgreSQL over MongoDB for ACID compliance",
"type": "Decision",
"confidence": 0.95,
"tags": ["database", "architecture"],
"importance": 0.9,
"metadata": {
"source": "architecture-meeting",
"alternatives": ["MongoDB", "MySQL"],
"deciding_factors": ["ACID", "team_expertise"]
}
}'Available memory types: Decision, Pattern, Preference, Style, Habit, Insight, Context (default)
- Explicit
typerecommended when you know the classification - Omit
typeto let enrichment auto-classify from content
# Hybrid search with tags and time
GET /recall?query=database&tags=decision&time_query=last%20month
# Semantic search with vector
GET /recall?embedding=0.12,0.56,...&limit=10
# Tag prefix matching (finds slack:U123:*, slack:channel-ops, etc.)
GET /recall?tags=slack&tag_match=prefixcurl -X POST http://localhost:8001/associate \
-H "Authorization: Bearer YOUR_TOKEN" \
-H "Content-Type: application/json" \
-d '{
"memory1_id": "uuid-postgres-decision",
"memory2_id": "uuid-mongodb-evaluation",
"type": "PREFERS_OVER",
"strength": 0.9
}'AutoMem works with any AI platform via:
Local MCP Bridge (Claude Desktop, Cursor, Claude Code):
# Install official MCP bridge
npm install -g @verygoodplugins/mcp-automem
# Configure for local AI tools
npx @verygoodplugins/mcp-automem setupSSE Sidecar (Cloud AI Platforms):
Connect AutoMem to cloud AI platforms via HTTPS. Works with:
- ChatGPT (requires developer mode)
- Claude.ai web interface
- Claude mobile app
- ElevenLabs Agents
See MCP over SSE documentation for setup instructions.
π Resources:
- NPM bridge (local): https://www.npmjs.com/package/@verygoodplugins/mcp-automem
- SSE setup guide: docs/MCP_SSE.md
Any language, any framework:
import requests
response = requests.post(
"https://your-automem.railway.app/memory",
headers={"Authorization": f"Bearer {token}"},
json={"content": "Memory content", "importance": 0.8}
)Traditional RAG retrieves similar documents. AutoMem understands relationships:
RAG: "Here are 5 documents about PostgreSQL"
AutoMem: "You chose PostgreSQL over MongoDB because you prefer boring technology for operational simplicity. This pattern also influenced your Redis and RabbitMQ decisions."
- β Typed relationships - Not just "similar", but "causes", "contradicts", "evolved from"
- β Temporal awareness - Knows what preceded, what invalidated, what emerged
- β Pattern learning - Discovers your preferences and decision-making style
- β Consolidation - Memories strengthen or fade based on useβlike human memory
Vector databases match embeddings. AutoMem builds knowledge graphs:
- β Multi-hop reasoning - Bridge discovery connects memories across conversation threads
- β 11 relationship types - Structured semantics vs. cosine similarity alone
- β Background intelligence - Auto-enrichment, pattern detection, decay cycles
- β 9-component scoring - Combines semantic, lexical, graph, temporal, and importance signals
AutoMem delivers what took 12 days of iteration to achieve SOTA performance:
- β Benchmark-proven - 90.53% on LoCoMo (ACL 2024), beats funded competitors
- β Research-validated - Implements HippoRAG 2, A-MEM, MELODI, ReadAgent principles
- β Production-ready - Auth, admin tools, health monitoring, automated backups
- β Battle-tested - Enrichment pipeline, consolidation engine, retry logic, dual storage
- β Open source - MIT license, deploy anywhere, no vendor lock-in
90.53% overall accuracy across 1,986 questions:
| Category | AutoMem | Notes |
|---|---|---|
| Complex Reasoning | 100% | Perfect score on multi-step reasoning |
| Open Domain | 95.84% | General knowledge recall |
| Temporal Understanding | 85.05% | Time-aware queries |
| Single-hop Recall | 79.79% | Basic fact retrieval |
| Multi-hop Reasoning | 50.00% | Connecting disparate memories (+12.5pp) |
Comparison:
- CORE (previous SOTA): 88.24%
- AutoMem: 90.53% (+2.29 points)
- OpenAI baseline: 39%
Run the benchmark yourself: make test-locomo
- β‘ Sub-100ms recall - Even with 100k+ memories
- π Concurrent writes - Background enrichment doesn't block API
- π‘οΈ Graceful degradation - Works without Qdrant (graph-only mode)
- β»οΈ Automatic retries - Failed enrichments queue for reprocessing
- π Health monitoring -
/healthand/enrichment/statusendpoints - πΎ Dual storage redundancy - Data persists in both FalkorDB and Qdrant
- π¦ Automated backups - Optional backup service for disaster recovery
AUTOMEM_API_TOKEN- Authentication for all endpoints (except/health)FALKORDB_HOST/FALKORDB_PORT- Graph database connection
QDRANT_URL/QDRANT_API_KEY- Enable semantic searchOPENAI_API_KEY- Real embeddings (otherwise deterministic placeholders)ADMIN_API_TOKEN- Required for/admin/reembedand enrichment controls- Consolidation tuning:
CONSOLIDATION_*_INTERVAL_SECONDS - Enrichment tuning:
ENRICHMENT_*(similarity threshold, retry limits, etc.)
- π¦ Installation Guide - Railway, Docker, development setup
- π MCP over SSE Sidecar - Expose AutoMem as an MCP server over SSE for ChatGPT/ElevenLabs
- πΎ Monitoring & Backups - Health monitoring and automated backups
- π§ API Reference - All endpoints with examples
- π§ͺ Testing Guide - Unit, integration, live server, and LoCoMo benchmark tests
- π LoCoMo Benchmark - Validate against ACL 2024 long-term memory benchmark
- π Migration Guide - Move from MCP SQLite
- π automem.ai - Official website and guides
- π automem.ai - Official website
- π GitHub - Source code
- π¦ NPM MCP Bridge - MCP integration
- π Issues - Bug reports and feature requests
AutoMem's 90.53% LoCoMo score didn't come from hypeβit came from implementing peer-reviewed neuroscience and graph theory:
HippoRAG 2 (Ohio State, June 2025)
Finding: Graph-vector hybrid achieves 7% better associative memory than pure vector RAG, approaching human long-term memory performance.
AutoMem implementation: Dual FalkorDB (graph) + Qdrant (vector) architecture with 11 typed relationship edges.
A-MEM (July 2025)
Finding: Dynamic memory organization with Zettelkasten principles enables emergent knowledge structures.
AutoMem implementation: Pattern detection, clustering cycles, and automatic entity linking that builds knowledge graphs from conversation.
MELODI (DeepMind, 2024)
Finding: 8x memory compression without quality loss through gist representations and selective preservation.
AutoMem implementation: Summary generation, importance scoring, and consolidation cycles that strengthen relevant memories while fading noise.
ReadAgent (DeepMind, 2024)
Finding: 20x context extension via episodic memory and temporal organization.
AutoMem implementation: Temporal relationship types (PRECEDED_BY, OCCURRED_BEFORE) and time-aware scoring that preserves conversation flow.
We didn't just read the papers. We built the system they describeβthen validated it on academic benchmarks.
We welcome contributions! Please:
- Fork the repository
- Create a feature branch
- Add tests for your changes
- Submit a pull request
See TESTING.md for running the test suite.
MIT - Because AI memory should be free.
railway upAutoMem is the world's best-performing long-term memory system.
90.53% LoCoMo score. Open source. Research-validated. Production-ready.
Transform your AI from a chatbot into a thinking partner that actually remembers.
Built by a solo developer. Validated by academic benchmarks. Beats well-funded competitors.
MIT License. Deploy anywhere. No vendor lock-in.