Give AI memory that lasts

Recalla is an open-source, self-hosted memory infrastructure for AI systems. Store, search, and sync context across models, tools, and devices.

Get Started How it works
$ git clone https://github.com/gentpan/recalla.git
$ cd recalla && docker compose up -d
# That's it. Memory layer running.
AI forgets everything
Every session starts from zero. No memory, no continuity. You keep repeating yourself across devices and tools.

Context lost

Switch from MacBook to desktop — AI has no idea what you were working on. Decisions forgotten, progress lost.

Repeated work

Explain the same architecture, same codebase, same constraints — every single session, every single device.

No continuity

Token limits, session breaks, tool switching. Context is a short-term cache that vanishes instantly.

Recalla fixes that
A persistent, searchable, cross-device memory layer for your AI workflow.

Cross-device sync

Your memories live on your server. Switch devices, AI picks up exactly where you left off. Branch, device, decisions — all remembered.

Semantic search

Find relevant context with natural language. Powered by vector search (Qdrant). Supports OpenAI, Qwen, Ollama embeddings.

MCP protocol

One server, all tools. Works with Claude Code Cursor VS Code Codex via standard MCP.

AI compression

Long sessions compressed into structured summaries. Key decisions, code changes, and TODOs preserved. Context limits handled.

GitHub integration

Auto-record pushes and PR events as memories via webhook. Browse repos and commits from dashboard.

Team collaboration

Create teams, invite members by username, share memories across the team. Team-wide semantic search finds knowledge from any member.

Knowledge graph

Entity facts with temporal validity. Track who works on what, which tech stack is used, when decisions were made. Contradiction detection built-in.

Auto-tagging & scoring

Memories automatically tagged (decision/bug/deploy/code) and scored by importance. Decisions ranked higher. Zero manual effort.

Config sync

Push CLAUDE.md and AI configs to server. New device pulls the latest version automatically. No Syncthing needed.

Telegram & CLI

Search memories, ask AI questions, get daily briefings — all from Telegram or terminal. recalla search "login bug"

Self-hosted

Your data on your server. Zero third-party dependency. Docker one-command deploy. MIT licensed.

Architecture
Simple, fast, self-contained.
ANY DEVICE Claude Code Cursor Codex VS Code MCP Protocol Recalla Server Go | REST API | MCP | Dashboard | Webhook PostgreSQL Memories / Teams Knowledge Graph Qdrant Vector Search Semantic Retrieval AI Provider OpenAI / Qwen / DeepSeek Embedding + Compression Save Memory → Embed → Store Vector → Search → Inject Context → AI Output docker compose up
MCP Tools
12 tools that give your AI persistent memory and team collaboration.
ToolDescriptionWhen to use
context_restoreRestore project context — last device, branch, recent memoriesEvery conversation start
memory_saveSave a memory — decisions, code changes, bug fixes, deploymentsAfter important actions
memory_searchSemantic search across all memoriesBefore starting new tasks
session_syncSync current session to cloud with device and branch infoEnd of conversation
session_compressAI-compress long sessions into structured summariesWhen context gets long
project_listList all recorded projectsDiscovery
config_pushPush local AI config files to serverAfter config changes
config_pullPull latest config from server to localNew device setup
team_searchSearch across all team members' memoriesTeam collaboration
team_shareShare a memory with your teamKnowledge sharing
add_factAdd entity fact to knowledge graphRecording facts
query_factsQuery knowledge graph with temporal validityWho works on what

Start in 30 seconds

Clone, deploy, connect. Your AI remembers everything from now on.

GitHub Documentation