Engram
Engram is an event - sourcing memory system designed for AI agents, adopting an architecture without an LLM write path, and achieving reliable semantic search memory storage through local vector embeddings and DuckDB.
2 points
5.2K

What is Engram?

Engram is an innovative AI memory system specifically designed for AI assistants and agents. Different from traditional memory systems, Engram decouples memory storage and semantic search, ensuring that write operations are always reliable, even when there is no network connection or external APIs are unavailable. The core idea of Engram is to first reliably store memory fragments (called "events") and then perform semantic search. This design avoids relying on potentially unstable external services in the write path, ensuring that your AI assistant can always remember important information.

How to use Engram?

Using Engram is very simple: 1. Install the Engram binary or run it via Docker. 2. Configure the connection to the local Ollama service (for generating text vectors). 3. Integrate with Claude Desktop, Claude Code, or Cursor via the MCP protocol. 4. Your AI assistant can start storing and retrieving memories. Engram will automatically handle all technical details. You only need to focus on the conversation with your AI assistant.

Use cases

Engram is particularly suitable for the following scenarios: - **Long - term conversations**: Let the AI assistant remember important information across multiple sessions. - **Project collaboration**: Store project requirements, decisions, and progress. - **Personal assistant**: Remember your preferences, schedules, and important matters. - **Research assistant**: Organize research materials, references, and notes. - **Code development**: Remember code structures, API documentation, and development decisions.

Main features

Semantic search
Use vector similarity for intelligent search instead of simple keyword matching. This means that Engram can understand the semantic meaning of the query and find the most relevant content, even without an exact keyword match.
Graceful fallback
Engram can still work normally even if the vector generation service is unavailable. It will first store the text content and generate vectors after the service is restored, ensuring that write operations never fail.
Fast queries
Use DuckDB's HNSW indexing technology to achieve millisecond - level vector search responses. Even when a large amount of memory is stored, the search speed remains fast.
Local vector generation
All text vectors are generated locally via Ollama, without the need to call external APIs, protecting your privacy and reducing latency.
Single - file deployment
Engram is an independent executable file that does not require the installation of complex dependencies. Just download one file and configure a few environment variables to run it.
MCP native support
It can be directly integrated into Claude Desktop, Claude Code, and Cursor without additional configuration. Your AI assistant can use Engram just like a built - in function.
Advantages
100% reliable write operations: Does not rely on external LLM APIs. It can write as long as the database is working properly.
Privacy protection: All data is processed locally and will not be sent to the cloud.
Fast response: Local vector generation and search with extremely low latency.
Easy to use: Single - file deployment with simple configuration.
Cost - effective: No need to pay for API calls.
Offline work: Can store memories even without a network connection.
Limitations
Requires local resources: Needs to run the Ollama service, which occupies local computing resources.
Initial setup: Requires manual configuration of environment variables and integration.
Relatively basic functions: Focuses on reliable storage and search, without complex memory organization functions.
Depends on Ollama: If the Ollama service stops, new memories cannot generate vectors (but can still be stored).

How to use

Install Engram
Download the pre - compiled binary file suitable for your operating system from the GitHub Releases page, or build it from the source code.
Install and configure Ollama
Install Ollama and download the vector generation model. Ollama is a locally run large language model service.
Configure environment variables
Set the environment variables required for Engram to run, including the database path and Ollama connection information.
Integrate into the AI assistant
Add Engram to the MCP configuration of Claude Desktop, Claude Code, or Cursor.
Start using
Restart your AI assistant. Now it can use Engram to store and retrieve memories.

Usage examples

Remember project requirements
When discussing project requirements with the AI assistant, let the assistant remember important functional requirements and design decisions.
Cross - session memory
Let the AI assistant remember your personal preferences and work habits mentioned in multiple sessions.
Research material organization
When researching a topic, let the AI assistant organize and remember important reference materials and key points.
Meeting minutes
During a meeting discussion, let the AI assistant record important decisions, to - do items, and responsible persons.

Frequently Asked Questions

What is the difference between Engram and an ordinary note - taking app?
Do I need to run Ollama all the time?
Will Engram store my private conversations?
Can I access my memories from other devices?
Which AI assistants does Engram support?
If I have too many memories, will the search slow down?
Can I export or back up my memories?
Is Engram free?

Related resources

Official documentation
Complete Engram technical documentation and usage guide
GitHub repository
Source code, issue tracking, and release versions
MCP integration guide
Detailed MCP client integration instructions
Deployment guide
Deployment guides for Docker, Kubernetes, and production environments
Ollama official website
A tool for running large language models locally
Model Context Protocol
Official specification of the MCP protocol

Installation

Copy the following command to your Client for configuration
{
  "mcpServers": {
    "engram-memory": {
      "command": "/absolute/path/to/engram",
      "args": [],
      "env": {
        "DUCKDB_PATH": "/absolute/path/to/engram.duckdb",
        "OLLAMA_URL": "http://localhost:11434",
        "EMBEDDING_MODEL": "nomic-embed-text"
      }
    }
  }
}
Note: Your key is sensitive information, do not share it with anyone.

Alternatives

A
Airweave
Airweave is an open - source context retrieval layer for AI agents and RAG systems. It connects and synchronizes data from various applications, tools, and databases, and provides relevant, real - time, multi - source contextual information to AI agents through a unified search interface.
Python
16.2K
5 points
V
Vestige
Vestige is an AI memory engine based on cognitive science. By implementing 29 neuroscience modules such as prediction error gating, FSRS - 6 spaced repetition, and memory dreaming, it provides long - term memory capabilities for AI. It includes a 3D visualization dashboard and 21 MCP tools, runs completely locally, and does not require the cloud.
Rust
10.5K
4.5 points
M
Moltbrain
MoltBrain is a long-term memory layer plugin designed for OpenClaw, MoltBook, and Claude Code, capable of automatically learning and recalling project context, providing intelligent search, observation recording, analysis statistics, and persistent storage functions.
TypeScript
10.1K
4.5 points
B
Bm.md
A feature-rich Markdown typesetting tool that supports multiple style themes and platform adaptation, providing real-time editing preview, image export, and API integration capabilities
TypeScript
14.8K
5 points
S
Security Detections MCP
Security Detections MCP is a server based on the Model Context Protocol that allows LLMs to query a unified security detection rule database covering Sigma, Splunk ESCU, Elastic, and KQL formats. The latest version 3.0 is upgraded to an autonomous detection engineering platform that can automatically extract TTPs from threat intelligence, analyze coverage gaps, generate SIEM-native format detection rules, run tests, and verify. The project includes over 71 tools, 11 pre-built workflow prompts, and a knowledge graph system, supporting multiple SIEM platforms.
TypeScript
6.7K
4 points
P
Paperbanana
Python
8.9K
5 points
B
Better Icons
An MCP server and CLI tool that provides search and retrieval of over 200,000 icons, supports more than 150 icon libraries, and helps AI assistants and developers quickly obtain and use icons.
TypeScript
9.7K
4.5 points
A
Assistant Ui
assistant - ui is an open - source TypeScript/React library for quickly building production - grade AI chat interfaces, providing composable UI components, streaming responses, accessibility, etc., and supporting multiple AI backends and models.
TypeScript
10.0K
5 points
M
Markdownify MCP
Markdownify is a multi-functional file conversion service that supports converting multiple formats such as PDFs, images, audio, and web page content into Markdown format.
TypeScript
39.1K
5 points
N
Notion Api MCP
Certified
A Python-based MCP Server that provides advanced to-do list management and content organization functions through the Notion API, enabling seamless integration between AI models and Notion.
Python
24.8K
4.5 points
D
Duckduckgo MCP Server
Certified
The DuckDuckGo Search MCP Server provides web search and content scraping services for LLMs such as Claude.
Python
81.4K
4.3 points
G
Gitlab MCP Server
Certified
The GitLab MCP server is a project based on the Model Context Protocol that provides a comprehensive toolset for interacting with GitLab accounts, including code review, merge request management, CI/CD configuration, and other functions.
TypeScript
28.4K
4.3 points
U
Unity
Certified
UnityMCP is a Unity editor plugin that implements the Model Context Protocol (MCP), providing seamless integration between Unity and AI assistants, including real - time state monitoring, remote command execution, and log functions.
C#
38.4K
5 points
F
Figma Context MCP
Framelink Figma MCP Server is a server that provides access to Figma design data for AI programming tools (such as Cursor). By simplifying the Figma API response, it helps AI more accurately achieve one - click conversion from design to code.
TypeScript
69.4K
4.5 points
G
Gmail MCP Server
A Gmail automatic authentication MCP server designed for Claude Desktop, supporting Gmail management through natural language interaction, including complete functions such as sending emails, label management, and batch operations.
TypeScript
24.9K
4.5 points
M
Minimax MCP Server
The MiniMax Model Context Protocol (MCP) is an official server that supports interaction with powerful text-to-speech, video/image generation APIs, and is suitable for various client tools such as Claude Desktop and Cursor.
Python
55.3K
4.8 points
AIBase
Zhiqi Future, Your AI Solution Think Tank
© 2026AIBase