by zilliztech
A Markdown-first memory system, a standalone library for any AI agent. Inspired by OpenClaw.
# Add to your Claude Code skills
git clone https://github.com/zilliztech/memsearch.md files β human-readable, editable, version-controllable. Milvus is a "shadow index": a derived, rebuildable cachePick your platform, install the plugin, and you're done. Each plugin captures conversations automatically and provides semantic recall with zero configuration.
# Install
/plugin marketplace add zilliztech/memsearch
/plugin install memsearch
# Restart Claude Code to activate the plugin
After restarting, just chat with Claude Code as usual. The plugin captures every conversation turn automatically.
Verify it's working β after a few conversations, check your memory files:
ls .memsearch/memory/ # you should see daily .md files
cat .memsearch/memory/$(date +%Y-%m-%d).md
Recall memories β two ways to trigger:
/memory-recall what did we discuss about Redis?
Or just ask naturally β Claude auto-invokes the skill when it senses the question needs history:
We discussed Redis caching before, what was the TTL we chose?
# Install from ClawHub
openclaw plugins install clawhub:memsearch
openclaw gateway restart
After installing, chat in TUI as usual. The plugin captures each turn automatically.
Verify it's working β memory files are stored in your agent's workspace:
# For the main agent:
ls ~/.openclaw/workspace/.memsearch/memory/
# For other agents (e.g. work):
ls ~/.openclaw/workspace-work/.memsearch/memory/
Recall memories β two ways to trigger:
/memory-recall what was the batch size limit we set?
Or just ask naturally β the LLM auto-invokes memory tools when it senses the question needs history:
We discussed batch size limits before, what did we decide?
π OpenClaw Plugin docs
// In ~/.config/opencode/opencode.json
{ "plugin": ["@zilliz/memsearch-opencode"] }
After installing, chat in TUI as usual. A background daemon captures conversations.
Verify it's working:
ls .memsearch/memory/ # daily .md files appear after a few conversations
Recall memories β two ways to trigger:
/memory-recall what did we discuss about authentication?
Or just ask naturally β the LLM auto-invokes memory tools when it senses the question needs history:
We discussed the authentication flow before, what was the approach?
π OpenCode Plugin docs
# Install
bash memsearch/plugins/codex/scripts/install.sh
codex --yolo # needed for ONNX model network access
After installing, chat as usual. Hooks capture and summarize each turn.
Verify it's working:
ls .memsearch/memory/
Recall memories β use the skill:
$memory-recall what did we discuss about deployment?
All plugins share the same memsearch backend. Configure once, works everywhere.
Defaults to ONNX bge-m3 β runs locally on CPU, no API key, no cost. On first launch the model (~558 MB) is downloaded from HuggingFace Hub.
memsearch config set embedding.provider onnx # default β local, free
memsearch config set embedding.provider openai # needs OPENAI_API_KEY
memsearch config set embedding.provider ollama # local, any model
All providers and models: Configuration β Embedding Provider
Just change milvus_uri (and optionally milvus_token) to switch between deployment modes:
Milvus Lite (default) β zero config, single file. Great for getting started:
# Works out of the box, no setup needed
memsearch config get milvus.uri # β ~/.memsearch/milvus.db
β Zilliz Cloud (recommended) β fully managed, free tier available β sign up π:
memsearch config set milvus.uri "https://in03-xxx.api.gcp-us-west1.zillizcloud.com"
memsearch config set milvus.token "your-api-key"
You can sign up on Zilliz Cloud to get a free cluster and API key.

For multi-user or team environments with a dedicated Milvus instance. Requires Docker. See the official installation guide.
memsearch config set milvus.uri http://localhost:19530
π Full configuration guide: Configuration Β· Platform comparison
Beyond ready-to-use plugins, memsearch provides a complete CLI and Python API for building memory into your own agents. Whether you're adding persistent context to a custom agent, building a memory-augmented RAG pipeline, or doing harness engineering β the same core engine that powers the plugins is available as a library.
ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β π§βπ» For Agent Users (Plugins) β
β ββββββββββββ ββββββββββββ ββββββββββββ ββββββββββ ββββββββ β
β β Claude β β OpenClaw β β OpenCode β β Codex β β Your β β
β β Code β β Plugin β β Plugin β β Plugin β β App β β
β ββββββ¬ββββββ ββββββ¬ββββββ ββββββ¬ββββββ βββββ¬βββββ ββββ¬ββββ β
β βββββββββββββββ΄βββββββββββββ΄βββββββ
No comments yet. Be the first to share your thoughts!