Show HN: Engram – Memory for AI coding agents (2.5K installs, 80% on LOCOMO)
1 points
2 hours ago
| 0 comments
| engram.fyi
| HN
Engram gives AI coding agents persistent memory across sessions. 2,500 npm installs in the first five days. The problem: every Claude Code or Cursor session starts from zero. You re-explain your architecture, conventions, past decisions. CLAUDE.md helps but it's manual and doesn't scale. Engram runs as a native MCP server. One command to install, SQLite backend, no infrastructure. Three tiers of memory:

Explicit: things you tell it to remember Implicit: behavioral patterns extracted from how you work (low confidence, reinforced over time) Synthesized: meta-observations generated during consolidation

The key architectural bet: intelligence at read time, not write time. Most memory systems extract and classify on ingest. Engram stores broadly and invests compute when you query, because that's when you actually know what matters. This is why it scores 80% on LOCOMO (arXiv:2402.17753) while using 30x fewer tokens than full-context retrieval. Stack: TypeScript, SQLite + sqlite-vec, Gemini embeddings by default (any OpenAI-compatible provider works via ENGRAM_LLM_BASE_URL). Zero external dependencies.

Install: npm install -g engram-sdk && engram init Comparison with Mem0, Zep/Graphiti, Letta/MemGPT: https://www.engram.fyi/compare GitHub: https://github.com/tstockham96/engram

No one has commented on this post.