loading…
Search for a command to run...
loading…
A privacy-focused local memory server that provides long-term semantic storage and retrieval for AI agents using SQLite and ChromaDB. It enables LLMs to persist
A privacy-focused local memory server that provides long-term semantic storage and retrieval for AI agents using SQLite and ChromaDB. It enables LLMs to persist and query text, chat histories, and PDF documents across sessions through the Model Context Protocol.
Your Eternal Second Brain, Running Locally.
Eternity MCP is a lightweight, privacy-focused memory server designed to provide long-term memory for LLMs and AI agents using the Model Context Protocol (MCP).
It combines structured storage (SQLite) with semantic vector search (ChromaDB), enabling agents to persist and retrieve text, PDF documents, and chat histories across sessions using natural language queries.
Built to run fully locally, Eternity integrates seamlessly with MCP-compatible clients, LangChain, LangGraph, and custom LLM pipelines, giving agents a durable and private memory layer.
Building agents that "remember" is hard. Most solutions rely on expensive cloud vector databases or complex setups. Eternity solves this by being:

You can install Eternity directly from PyPI (coming soon) or from source:
# From source
git clone https://github.com/danttis/eternity-mcp.git
cd eternity
Run the server in a terminal. It will host the API and the Memory UI.
eternity
Server runs at http://localhost:8000
You can interact with Eternity using simple HTTP requests.
import requests
ETERNITY_URL = "http://localhost:8000"
# 💾 Store a memory
requests.post("{ETERNITY_URL}/add", data={
"content": "The project deadline is next Friday.",
"tags": "work,deadline"
})
# 🔍 Search memory
response = requests.get("{ETERNITY_URL}/search", params={"q": "When is the deadline?"})
print(response.json())
Eternity shines when connected to an LLM. Here is a simple pattern for an agent with long-term memory:
(See langgraph_agent.py in the repo for a full, working example using Ollama/Groq).
| Method | Endpoint | Description |
|---|---|---|
GET |
/ |
Web UI to view recent memories. |
POST |
/add |
Add text or file (PDF). Params: content, tags, file. |
GET |
/search |
Semantic search. Params: q (query text). |
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.
This project was inspired by Supermemory. We admire their vision for a second brain and their open-source spirit.
Created by Junior Dantas with a little help from AI :)
Добавь это в claude_desktop_config.json и перезапусти Claude Desktop.
{
"mcpServers": {
"eternity-mcp": {
"command": "npx",
"args": []
}
}
}