OpenClaw Advanced Memory OpenClaw Skill

Provides persistent, searchable AI agent memory with real-time capture, vector search, and nightly LLM curation for long-term recall on local hardware.

v1.0.0 Recently Updated Updated 1 day ago

Installation

clawhub install openclaw-advanced-memory

Requires npm i -g clawhub

344

Downloads

1

Stars

0

current installs

0 all-time

1

Versions

EU EU-Hosted Inference API

Power your OpenClaw skills with the best open-source models.

Drop-in OpenAI-compatible API. No data leaves Europe.

Explore Inference API

GLM

GLM 5

$1.00 / $3.20

per M tokens

Kimi

Kimi K2.5

$0.60 / $2.80

per M tokens

MiniMax

MiniMax M2.5

$0.30 / $1.20

per M tokens

Qwen

Qwen3.5 122B

$0.40 / $3.00

per M tokens

OpenClaw Advanced Memory

Three-tier AI agent memory system — real-time capture, vector search, and LLM-curated long-term recall.

What It Does

Gives your OpenClaw agent persistent, searchable memory that survives across sessions:

  • HOT tier — Redis buffer captures conversation turns in real-time (every 30s)
  • WARM tier — Qdrant vector store with chunked, embedded conversations (searchable, 7-day retention)
  • COLD tier — LLM-curated "gems" extracted nightly (decisions, lessons, milestones — stored forever)

Requirements

  • Qdrant — vector database (Docker recommended)
  • Redis — buffer queue (Docker recommended)
  • Ollama — local embeddings (snowflake-arctic-embed2) + curation LLM (qwen2.5:7b)
  • Python 3.10+ with qdrant-client, redis, requests

No cloud APIs. No subscriptions. Runs entirely on your own hardware.

Setup

# 1. Start Qdrant + Redis (Docker)
docker compose up -d

# 2. Pull Ollama models
ollama pull snowflake-arctic-embed2
ollama pull qwen2.5:7b

# 3. Run the installer
bash scripts/install.sh

The installer sets up Qdrant collections, installs a systemd capture service, and configures cron jobs.

Edit connection hosts at the top of each script if your infra isn't on localhost.

Usage

# Search your memory
./recall "what did we decide about pricing"
./recall "deployment" --project myproject --tier cold -v

# Check system status
./mem-status

# Force a warm flush or curation run
./warm-now
./curate-now 2026-03-01

Schedules

Component Schedule What It Does
mem-capture Always running (systemd) Watches transcripts → Redis
mem-warm Every 30 min (cron) Redis → Qdrant warm
mem-curate Nightly 2 AM (cron) Warm → LLM curation → Qdrant cold

How Curation Works

Every night, a local LLM (qwen2.5:7b via Ollama) reads the day's conversations and extracts structured gems:

{
  "gem": "Chose DistilBERT over TinyBERT — 99.69% F1, zero false positives",
  "context": "A/B tested both architectures on red team suite",
  "categories": ["decision", "technical"],
  "project": "guardian",
  "importance": "high"
}

Only decisions, milestones, lessons, and people info make the cut. Casual banter and debugging noise get filtered out.

Links

Statistics

Downloads 344
Stars 1
Current installs 0
All-time installs 0
Versions 1
Comments 0
Created Mar 3, 2026
Updated Apr 20, 2026

Latest Changes

v1.0.0 · Mar 3, 2026

Initial release: Three-tier memory system (Hot/Warm/Cold) with real-time capture, vector search, and LLM-curated long-term recall.

Quick Install

clawhub install openclaw-advanced-memory
EU Made in Europe

Chat with 100+ AI Models in one App.

Use Claude, ChatGPT, Gemini alongside with EU-Hosted Models like Deepseek, GLM-5, Kimi K2.5 and many more.

Customer Support