Self-Hosted AI Memory Server

A private AI memory engine that runs entirely on your infrastructure. Single Rust binary, on-device embeddings, AES-256-GCM encryption. Your data never leaves your machines.

Why Self-Host Your AI Memory?

How It Works

Deploy the Binary

Choose Docker, Kubernetes (Helm), or systemd. A single binary runs the full memory engine with REST (port 3300) and gRPC (port 50051).

# Docker
docker run -d -p 3300:3300 -p 50051:50051 \
  -v dakera_data:/data \
  ghcr.io/dakera-ai/dakera:latest

# Or download the binary
curl -fsSL https://get.dakera.ai | sh
dakera serve

Configure Encryption & Access

All data is encrypted with AES-256-GCM at rest. Create scoped API keys for each agent or service.

Connect Your Agents

Use the Python, TypeScript, Go, or Rust SDK — or connect via 83 MCP tools for Claude Desktop, Cursor, and Windsurf.

from dakera import Dakera

client = Dakera(base_url="http://localhost:3300", api_key="dk-...")

# Store a memory
client.memory.store(
    content="User requires HIPAA-compliant data handling",
    namespace="compliance",
    metadata={"importance": 1.0}
)

# Recall with hybrid retrieval
results = client.memory.recall(
    query="What compliance requirements exist?",
    namespace="compliance",
    top_k=5
)

Deployment Options

MethodCommandBest For
Dockerdocker run ghcr.io/dakera-ai/dakera:latestDevelopment, single-node production
Kuberneteshelm install dakera dakera/dakeraScalable production clusters
systemddakera serveBare-metal, VPS, edge deployment

Security Architecture

LayerImplementation
Encryption at RestAES-256-GCM with user-controlled keys
Access ControlScoped API keys per namespace/agent
Rate LimitingConfigurable per-key request limits
NetworkBind to localhost by default; TLS termination at your reverse proxy
IsolationNamespace-level data isolation, multi-agent separation

Frequently Asked Questions

What does self-hosted AI memory mean? +

Self-hosted AI memory means your agent memory engine runs entirely on your own infrastructure. No data is sent to third-party clouds. Dakera runs as a single Rust binary with on-device embedding inference via ONNX, so even vector generation happens locally.

How is data encrypted? +

Dakera encrypts all memory data at rest using AES-256-GCM. API access is controlled via scoped API keys with rate limiting. You maintain full control of your encryption keys.

What are the deployment options? +

Dakera can be deployed via Docker (ghcr.io/dakera-ai/dakera:latest), Kubernetes with Helm charts, or as a systemd service using the standalone binary. It exposes REST on port 3300 and gRPC on port 50051.

Does it require external services like OpenAI for embeddings? +

No. Dakera uses on-device ONNX inference with models like MiniLM, BGE, and E5. All embedding generation happens locally without external API calls.

Deploy Private AI Memory in Minutes

Single binary, zero cloud dependencies, full data sovereignty.

Get Started