AI Memory Assistant for OpenClaw Agents
Give your AI agent a brain that actually remembers — across sessions, across agents, across time.
A LanceDB-backed OpenClaw memory plugin that stores preferences, decisions, and project context, then auto-recalls them in future sessions.
English | 简体中文
Most AI agents have amnesia. They forget everything the moment you start a new chat.
memory-lancedb-pro is a production-grade long-term memory plugin for OpenClaw that turns your agent into an AI Memory Assistant — it automatically captures what matters, lets noise naturally fade, and retrieves the right memory at the right time. No manual tagging, no configuration headaches.
Without memory — every session starts from zero:
You: "Use tabs for indentation, always add error handling." (next session) You: "I already told you — tabs, not spaces!" 😤 (next session) You: "...seriously, tabs. And error handling. Again."
With memory-lancedb-pro — your agent learns and remembers:
You: "Use tabs for indentation, always add error handling." (next session — agent auto-recalls your preferences) Agent: (silently applies tabs + error handling) ✅ You: "Why did we pick PostgreSQL over MongoDB last month?" Agent: "Based on our discussion on Feb 12, the main reasons were..." ✅
That's the difference an AI Memory Assistant makes — it learns your style, recalls past decisions, and delivers personalized responses without you repeating yourself.
| What you get | |
|---|---|
| Auto-Capture | Your agent learns from every conversation — no manual memory_store needed |
| Smart Extraction | LLM-powered 6-category classification: profiles, preferences, entities, events, cases, patterns |
| Intelligent Forgetting | Weibull decay model — important memories stay, noise naturally fades away |
| Hybrid Retrieval | Vector + BM25 full-text search, fused with cross-encoder reranking |
| Context Injection | Relevant memories automatically surface before each reply |
| Multi-Scope Isolation | Per-agent, per-user, per-project memory boundaries |
| Any Provider | OpenAI, Jina, Gemini, Ollama, or any OpenAI-compatible API |
| Full Toolkit | CLI, backup, migration, upgrade, export/import — production-ready |
The community-maintained setup script handles install, upgrade, and repair in one command:
curl -fsSL https://raw.githubusercontent.com/CortexReach/toolbox/main/memory-lancedb-pro-setup/setup-memory.sh -o setup-memory.sh
bash setup-memory.shSee Ecosystem below for the full list of scenarios the script covers and other community tools.
Via OpenClaw CLI (recommended):
openclaw plugins install memory-lancedb-pro@betaOr via npm:
npm i memory-lancedb-pro@betaIf using npm, you will also need to add the plugin's install directory as an absolute path in
plugins.load.pathsin youropenclaw.json. This is the most common setup issue.
Add to your openclaw.json:
{
"plugins": {
"slots": { "memory": "memory-lancedb-pro" },
"entries": {
"memory-lancedb-pro": {
"enabled": true,
"config": {
"embedding": {
"provider": "openai-compatible",
"apiKey": "${OPENAI_API_KEY}",
"model": "text-embedding-3-small"
},
"autoCapture": true,
"autoRecall": true,
"smartExtraction": true,
"extractMinMessages": 2,
"extractMaxChars": 8000,
"sessionMemory": { "enabled": false }
}
}
}
}
}Why these defaults?
autoCapture+smartExtraction→ your agent learns from every conversation automaticallyautoRecall→ relevant memories are injected before each replyextractMinMessages: 2→ extraction triggers in normal two-turn chatssessionMemory.enabled: false→ avoids polluting retrieval with session summaries on day one
Validate & restart:
openclaw config validate
openclaw gateway restart
openclaw logs --follow --plain | grep "memory-lancedb-pro"You should see:
memory-lancedb-pro: smart extraction enabledmemory-lancedb-pro@...: plugin registered
Done! Your agent now has long-term memory.
More installation paths (existing users, upgrades)
Already using OpenClaw?
- Add the plugin with an absolute
plugins.load.pathsentry - Bind the memory slot:
plugins.slots.memory = "memory-lancedb-pro" - Verify:
openclaw plugins info memory-lancedb-pro && openclaw memory-pro stats
Upgrading from pre-v1.1.0?
# 1) Backup
openclaw memory-pro export --scope global --output memories-backup.json
# 2) Dry run
openclaw memory-pro upgrade --dry-run
# 3) Run upgrade
openclaw memory-pro upgrade
# 4) Verify
openclaw memory-pro statsSee CHANGELOG-v1.1.0.md for behavior changes and upgrade rationale.
Telegram Bot Quick Import (click to expand)
If you are using OpenClaw's Telegram integration, the easiest way is to send an import command directly to the main Bot instead of manually editing config.
Send this message:
Help me connect this memory plugin with the most user-friendly configuration: https://github.com/CortexReach/memory-lancedb-pro
Requirements:
1. Set it as the only active memory plugin
2. Use Jina for embedding
3. Use Jina for reranker
4. Use gpt-4o-mini for the smart-extraction LLM
5. Enable autoCapture, autoRecall, smartExtraction
6. extractMinMessages=2
7. sessionMemory.enabled=false
8. captureAssistant=false
9. retrieval mode=hybrid, vectorWeight=0.7, bm25Weight=0.3
10. rerank=cross-encoder, candidatePoolSize=12, minScore=0.6, hardMinScore=0.62
11. Generate the final openclaw.json config directly, not just an explanation
memory-lancedb-pro is the core plugin. The community has built tools around it to make setup and daily use even smoother:
Not just a simple installer — the script intelligently handles a wide range of real-world scenarios:
| Your situation | What the script does |
|---|---|
| Never installed | Fresh download → install deps → pick config → write to openclaw.json → restart |
Installed via git clone, stuck on old commit |
Auto git fetch + checkout to latest → reinstall deps → verify |
| Config has invalid fields | Auto-detect via schema filter, remove unsupported fields |
Installed via npm |
Skips git update, reminds you to run npm update yourself |
openclaw CLI broken due to invalid config |
Fallback: read workspace path directly from openclaw.json file |
extensions/ instead of plugins/ |
Auto-detect plugin location from config or filesystem |
| Already up to date | Run health checks only, no changes |
bash setup-memory.sh # Install or upgrade
bash setup-memory.sh --dry-run # Preview only
bash setup-memory.sh --beta # Include pre-release versions
bash setup-memory.sh --uninstall # Revert config and remove pluginBuilt-in provider presets: Jina / DashScope / SiliconFlow / OpenAI / Ollama, or bring your own OpenAI-compatible API. For full usage (including --ref, --selfcheck-only, and more), see the setup script README.
Install this skill and your AI agent (Claude Code or OpenClaw) gains deep knowledge of every feature in memory-lancedb-pro. Just say "help me enable the best config" and get:
- Guided 7-step configuration workflow with 4 deployment plans:
- Full Power (Jina + OpenAI) / Budget (free SiliconFlow reranker) / Simple (OpenAI only) / Fully Local (Ollama, zero API cost)
- All 9 MCP tools used correctly:
memory_recall,memory_store,memory_forget,memory_update,memory_stats,memory_list,self_improvement_log,self_improvement_extract_skill,self_improvement_review(full toolset requiresenableManagementTools: true— the default Quick Start config exposes the 4 core tools) - Common pitfall avoidance: workspace plugin enablement,
autoRecalldefault-false, jiti cache, env vars, scope isolation, and more
Install for Claude Code:
git clone https://github.com/CortexReach/memory-lancedb-pro-skill.git ~/.claude/skills/memory-lancedb-proInstall for OpenClaw:
git clone https://github.com/CortexReach/memory-lancedb-pro-skill.git ~/.openclaw/workspace/skills/memory-lancedb-pro-skillFull walkthrough: installation, configuration, and hybrid retrieval internals.
https://www.bilibili.com/video/BV1zUf2BGEgn/
┌─────────────────────────────────────────────────────────┐
│ index.ts (Entry Point) │
│ Plugin Registration · Config Parsing · Lifecycle Hooks │
└────────┬──────────┬──────────┬──────────┬───────────────┘
│ │ │ │
┌────▼───┐ ┌────▼───┐ ┌───▼────┐ ┌──▼──────────┐
│ store │ │embedder│ │retriever│ │ scopes │
│ .ts │ │ .ts │ │ .ts │ │ .ts │
└────────┘ └────────┘ └────────┘ └─────────────┘
│ │
┌────▼───┐ ┌─────▼──────────┐
│migrate │ │noise-filter.ts │
│ .ts │ │adaptive- │
└────────┘ │retrieval.ts │
└────────────────┘
┌─────────────┐ ┌──────────┐
│ tools.ts │ │ cli.ts │
│ (Agent API) │ │ (CLI) │
└─────────────┘ └──────────┘
For a deep-dive into the full architecture, see docs/memory_architecture_analysis.md.
File Reference (click to expand)
| File | Purpose |
|---|---|
index.ts |
Plugin entry point. Registers with OpenClaw Plugin API, parses config, mounts lifecycle hooks |
openclaw.plugin.json |
Plugin metadata + full JSON Schema config declaration |
cli.ts |
CLI commands: memory-pro list/search/stats/delete/delete-bulk/export/import/reembed/upgrade/migrate |
src/store.ts |
LanceDB storage layer. Table creation / FTS indexing / Vector search / BM25 search / CRUD |
src/embedder.ts |
Embedding abstraction. Compatible with any OpenAI-compatible API provider |
src/retriever.ts |
Hybrid retrieval engine. Vector + BM25 → Hybrid Fusion → Rerank → Lifecycle Decay → Filter |
src/scopes.ts |
Multi-scope access control |
src/tools.ts |
Agent tool definitions: memory_recall, memory_store, memory_forget, memory_update + management tools |
src/noise-filter.ts |
Filters out agent refusals, meta-questions, greetings, and low-quality content |
src/adaptive-retrieval.ts |
Determines whether a query needs memory retrieval |
src/migrate.ts |
Migration from built-in memory-lancedb to Pro |
src/smart-extractor.ts |
LLM-powered 6-category extraction with L0/L1/L2 layered storage and two-stage dedup |
src/decay-engine.ts |
Weibull stretched-exponential decay model |
src/tier-manager.ts |
Three-tier promotion/demotion: Peripheral ↔ Working ↔ Core |
Query → embedQuery() ─┐
├─→ Hybrid Fusion → Rerank → Lifecycle Decay Boost → Length Norm → Filter
Query → BM25 FTS ─────┘
- Vector Search — semantic similarity via LanceDB ANN (cosine distance)
- BM25 Full-Text Search — exact keyword matching via LanceDB FTS index
- Hybrid Fusion — vector score as base, BM25 hits receive a weighted boost (not standard RRF — tuned for real-world recall quality)
- Configurable Weights —
vectorWeight,bm25Weight,minScore
- Built-in adapters for Jina, SiliconFlow, Voyage AI, and Pinecone
- Compatible with any Jina-compatible endpoint (e.g., Hugging Face TEI, DashScope)
- Hybrid scoring: 60% cross-encoder + 40% original fused score
- Graceful degradation: falls back to cosine similarity on API failure
| Stage | Effect |
|---|---|
| Hybrid Fusion | Combines semantic and exact-match recall |
| Cross-Encoder Rerank | Promotes semantically precise hits |
| Lifecycle Decay Boost | Weibull freshness + access frequency + importance × confidence |
| Length Normalization | Prevents long entries from dominating (anchor: 500 chars) |
| Hard Min Score | Removes irrelevant results (default: 0.35) |
| MMR Diversity | Cosine similarity > 0.85 → demoted |
- LLM-Powered 6-Category Extraction: profile, preferences, entities, events, cases, patterns
- L0/L1/L2 Layered Storage: L0 (one-sentence index) → L1 (structured summary) → L2 (full narrative)
- Two-Stage Dedup: vector similarity pre-filter (≥0.7) → LLM semantic decision (CREATE/MERGE/SKIP)
- Category-Aware Merge:
profilealways merges,events/casesare append-only
- Weibull Decay Engine: composite score = recency + frequency + intrinsic value
- Three-Tier Promotion:
Peripheral ↔ Working ↔ Corewith configurable thresholds - Access Reinforcement: frequently recalled memories decay slower (spaced-repetition style)
- Importance-Modulated Half-Life: important memories decay slower
- Built-in scopes:
global,agent:<id>,custom:<name>,project:<id>,user:<id> - Agent-level access control via
scopes.agentAccess - Default: each agent accesses
global+ its ownagent:<id>scope
- Auto-Capture (
agent_end): extracts preference/fact/decision/entity from conversations, deduplicates, stores up to 3 per turn - Auto-Recall (
before_agent_start): injects<relevant-memories>context (up to 3 entries)
- Filters low-quality content: agent refusals, meta-questions, greetings
- Skips retrieval for greetings, slash commands, simple confirmations, emoji
- Forces retrieval for memory keywords ("remember", "previously", "last time")
- CJK-aware thresholds (Chinese: 6 chars vs English: 15 chars)
Compared to Built-in memory-lancedb (click to expand)
| Feature | Built-in memory-lancedb |
memory-lancedb-pro |
|---|---|---|
| Vector search | Yes | Yes |
| BM25 full-text search | - | Yes |
| Hybrid fusion (Vector + BM25) | - | Yes |
| Cross-encoder rerank (multi-provider) | - | Yes |
| Recency boost & time decay | - | Yes |
| Length normalization | - | Yes |
| MMR diversity | - | Yes |
| Multi-scope isolation | - | Yes |
| Noise filtering | - | Yes |
| Adaptive retrieval | - | Yes |
| Management CLI | - | Yes |
| Session memory | - | Yes |
| Task-aware embeddings | - | Yes |
| LLM Smart Extraction (6-category) | - | Yes (v1.1.0) |
| Weibull Decay + Tier Promotion | - | Yes (v1.1.0) |
| Any OpenAI-compatible embedding | Limited | Yes |
Full Configuration Example
{
"embedding": {
"apiKey": "${JINA_API_KEY}",
"model": "jina-embeddings-v5-text-small",
"baseURL": "https://api.jina.ai/v1",
"dimensions": 1024,
"taskQuery": "retrieval.query",
"taskPassage": "retrieval.passage",
"normalized": true
},
"dbPath": "~/.openclaw/memory/lancedb-pro",
"autoCapture": true,
"autoRecall": true,
"retrieval": {
"mode": "hybrid",
"vectorWeight": 0.7,
"bm25Weight": 0.3,
"minScore": 0.3,
"rerank": "cross-encoder",
"rerankApiKey": "${JINA_API_KEY}",
"rerankModel": "jina-reranker-v3",
"rerankEndpoint": "https://api.jina.ai/v1/rerank",
"rerankProvider": "jina",
"candidatePoolSize": 20,
"recencyHalfLifeDays": 14,
"recencyWeight": 0.1,
"filterNoise": true,
"lengthNormAnchor": 500,
"hardMinScore": 0.35,
"timeDecayHalfLifeDays": 60,
"reinforcementFactor": 0.5,
"maxHalfLifeMultiplier": 3
},
"enableManagementTools": false,
"scopes": {
"default": "global",
"definitions": {
"global": { "description": "Shared knowledge" },
"agent:discord-bot": { "description": "Discord bot private" }
},
"agentAccess": {
"discord-bot": ["global", "agent:discord-bot"]
}
},
"sessionMemory": {
"enabled": false,
"messageCount": 15
},
"smartExtraction": true,
"llm": {
"apiKey": "${OPENAI_API_KEY}",
"model": "gpt-4o-mini",
"baseURL": "https://api.openai.com/v1"
},
"extractMinMessages": 2,
"extractMaxChars": 8000
}Embedding Providers
Works with any OpenAI-compatible embedding API:
| Provider | Model | Base URL | Dimensions |
|---|---|---|---|
| Jina (recommended) | jina-embeddings-v5-text-small |
https://api.jina.ai/v1 |
1024 |
| OpenAI | text-embedding-3-small |
https://api.openai.com/v1 |
1536 |
| Voyage | voyage-4-lite / voyage-4 |
https://api.voyageai.com/v1 |
1024 / 1024 |
| Google Gemini | gemini-embedding-001 |
https://generativelanguage.googleapis.com/v1beta/openai/ |
3072 |
| Ollama (local) | nomic-embed-text |
http://localhost:11434/v1 |
provider-specific |
Rerank Providers
Cross-encoder reranking supports multiple providers via rerankProvider:
| Provider | rerankProvider |
Example Model |
|---|---|---|
| Jina (default) | jina |
jina-reranker-v3 |
| SiliconFlow (free tier available) | siliconflow |
BAAI/bge-reranker-v2-m3 |
| Voyage AI | voyage |
rerank-2.5 |
| Pinecone | pinecone |
bge-reranker-v2-m3 |
Any Jina-compatible rerank endpoint also works — set rerankProvider: "jina" and point rerankEndpoint to your service (e.g., Hugging Face TEI, DashScope qwen3-rerank).
Smart Extraction (LLM) — v1.1.0
When smartExtraction is enabled (default: true), the plugin uses an LLM to intelligently extract and classify memories instead of regex-based triggers.
| Field | Type | Default | Description |
|---|---|---|---|
smartExtraction |
boolean | true |
Enable/disable LLM-powered 6-category extraction |
llm.auth |
string | api-key |
api-key uses llm.apiKey / embedding.apiKey; oauth uses a plugin-scoped OAuth token file by default |
llm.apiKey |
string | (falls back to embedding.apiKey) |
API key for the LLM provider |
llm.model |
string | openai/gpt-oss-120b |
LLM model name |
llm.baseURL |
string | (falls back to embedding.baseURL) |
LLM API endpoint |
llm.oauthProvider |
string | openai-codex |
OAuth provider id used when llm.auth is oauth |
llm.oauthPath |
string | ~/.openclaw/.memory-lancedb-pro/oauth.json |
OAuth token file used when llm.auth is oauth |
llm.timeoutMs |
number | 30000 |
LLM request timeout in milliseconds |
extractMinMessages |
number | 2 |
Minimum messages before extraction triggers |
extractMaxChars |
number | 8000 |
Maximum characters sent to the LLM |
OAuth llm config (use existing Codex / ChatGPT login cache for LLM calls):
{
"llm": {
"auth": "oauth",
"oauthProvider": "openai-codex",
"model": "gpt-5.4",
"oauthPath": "${HOME}/.openclaw/.memory-lancedb-pro/oauth.json",
"timeoutMs": 30000
}
}Notes for llm.auth: "oauth":
llm.oauthProvideris currentlyopenai-codex.- OAuth tokens default to
~/.openclaw/.memory-lancedb-pro/oauth.json. - You can set
llm.oauthPathif you want to store that file somewhere else. auth loginsnapshots the previous api-keyllmconfig next to the OAuth file, andauth logoutrestores that snapshot when available.- Switching from
api-keytooauthdoes not automatically carry overllm.baseURL. Set it manually in OAuth mode only when you intentionally want a custom ChatGPT/Codex-compatible backend.
Lifecycle Configuration (Decay + Tier)
| Field | Default | Description |
|---|---|---|
decay.recencyHalfLifeDays |
30 |
Base half-life for Weibull recency decay |
decay.frequencyWeight |
0.3 |
Weight of access frequency in composite score |
decay.intrinsicWeight |
0.3 |
Weight of importance × confidence |
decay.betaCore |
0.8 |
Weibull beta for core memories |
decay.betaWorking |
1.0 |
Weibull beta for working memories |
decay.betaPeripheral |
1.3 |
Weibull beta for peripheral memories |
tier.coreAccessThreshold |
10 |
Min recall count before promoting to core |
tier.peripheralAgeDays |
60 |
Age threshold for demoting stale memories |
Access Reinforcement
Frequently recalled memories decay more slowly (spaced-repetition style).
Config keys (under retrieval):
reinforcementFactor(0-2, default:0.5) — set0to disablemaxHalfLifeMultiplier(1-10, default:3) — hard cap on effective half-life
openclaw memory-pro list [--scope global] [--category fact] [--limit 20] [--json]
openclaw memory-pro search "query" [--scope global] [--limit 10] [--json]
openclaw memory-pro stats [--scope global] [--json]
openclaw memory-pro auth login [--provider openai-codex] [--model gpt-5.4] [--oauth-path /abs/path/oauth.json]
openclaw memory-pro auth status
openclaw memory-pro auth logout
openclaw memory-pro delete <id>
openclaw memory-pro delete-bulk --scope global [--before 2025-01-01] [--dry-run]
openclaw memory-pro export [--scope global] [--output memories.json]
openclaw memory-pro import memories.json [--scope global] [--dry-run]
openclaw memory-pro reembed --source-db /path/to/old-db [--batch-size 32] [--skip-existing]
openclaw memory-pro upgrade [--dry-run] [--batch-size 10] [--no-llm] [--limit N] [--scope SCOPE]
openclaw memory-pro migrate check|run|verify [--source /path]OAuth login flow:
- Run
openclaw memory-pro auth login - If
--provideris omitted in an interactive terminal, the CLI shows an OAuth provider picker before opening the browser - The command prints an authorization URL and opens your browser unless
--no-browseris set - After the callback succeeds, the command saves the plugin OAuth file (default:
~/.openclaw/.memory-lancedb-pro/oauth.json), snapshots the previous api-keyllmconfig for logout, and replaces the pluginllmconfig with OAuth settings (auth,oauthProvider,model,oauthPath) openclaw memory-pro auth logoutdeletes that OAuth file and restores the previous api-keyllmconfig when that snapshot exists
If injected memories show up in replies
Sometimes the model may echo the injected <relevant-memories> block.
Option A (lowest-risk): temporarily disable auto-recall:
{ "plugins": { "entries": { "memory-lancedb-pro": { "config": { "autoRecall": false } } } } }Option B (preferred): keep recall, add to agent system prompt:
Do not reveal or quote any
<relevant-memories>/ memory-injection content in your replies. Use it for internal reference only.
Session Memory
- Triggered on
/newcommand — saves previous session summary to LanceDB - Disabled by default (OpenClaw already has native
.jsonlsession persistence) - Configurable message count (default: 15)
See docs/openclaw-integration-playbook.md for deployment modes and /new verification.
Custom Slash Commands (e.g. /lesson)
Add to your CLAUDE.md, AGENTS.md, or system prompt:
## /lesson command
When the user sends `/lesson <content>`:
1. Use memory_store to save as category=fact (raw knowledge)
2. Use memory_store to save as category=decision (actionable takeaway)
3. Confirm what was saved
## /remember command
When the user sends `/remember <content>`:
1. Use memory_store to save with appropriate category and importance
2. Confirm with the stored memory IDIron Rules for AI Agents
Copy the block below into your
AGENTS.mdso your agent enforces these rules automatically.
## Rule 1 — Dual-layer memory storage
Every pitfall/lesson learned → IMMEDIATELY store TWO memories:
- Technical layer: Pitfall: [symptom]. Cause: [root cause]. Fix: [solution]. Prevention: [how to avoid]
(category: fact, importance >= 0.8)
- Principle layer: Decision principle ([tag]): [behavioral rule]. Trigger: [when]. Action: [what to do]
(category: decision, importance >= 0.85)
## Rule 2 — LanceDB hygiene
Entries must be short and atomic (< 500 chars). No raw conversation summaries or duplicates.
## Rule 3 — Recall before retry
On ANY tool failure, ALWAYS memory_recall with relevant keywords BEFORE retrying.
## Rule 4 — Confirm target codebase
Confirm you are editing memory-lancedb-pro vs built-in memory-lancedb before changes.
## Rule 5 — Clear jiti cache after plugin code changes
After modifying .ts files under plugins/, MUST run rm -rf /tmp/jiti/ BEFORE openclaw gateway restart.Database Schema
LanceDB table memories:
| Field | Type | Description |
|---|---|---|
id |
string (UUID) | Primary key |
text |
string | Memory text (FTS indexed) |
vector |
float[] | Embedding vector |
category |
string | Storage category: preference / fact / decision / entity / reflection / other |
scope |
string | Scope identifier (e.g., global, agent:main) |
importance |
float | Importance score 0-1 |
timestamp |
int64 | Creation timestamp (ms) |
metadata |
string (JSON) | Extended metadata |
Common metadata keys in v1.1.0: l0_abstract, l1_overview, l2_content, memory_category, tier, access_count, confidence, last_accessed_at
Note on categories: The top-level
categoryfield uses 6 storage categories. The 6-category semantic labels from Smart Extraction (profile/preferences/entities/events/cases/patterns) are stored inmetadata.memory_category.
Troubleshooting
On LanceDB 0.26+, some numeric columns may be returned as BigInt. Upgrade to memory-lancedb-pro >= 1.0.14 — this plugin now coerces values using Number(...) before arithmetic.
| Document | Description |
|---|---|
| OpenClaw Integration Playbook | Deployment modes, verification, regression matrix |
| Memory Architecture Analysis | Full architecture deep-dive |
| CHANGELOG v1.1.0 | v1.1.0 behavior changes and upgrade rationale |
| Long-Context Chunking | Chunking strategy for long documents |
Status: Beta — available via
npm i memory-lancedb-pro@beta. Stable users onlatestare not affected.
| Feature | Description |
|---|---|
| Smart Extraction | LLM-powered 6-category extraction with L0/L1/L2 metadata. Falls back to regex when disabled. |
| Lifecycle Scoring | Weibull decay integrated into retrieval — high-frequency and high-importance memories rank higher. |
| Tier Management | Three-tier system (Core → Working → Peripheral) with automatic promotion/demotion. |
Feedback: GitHub Issues · Revert: npm i memory-lancedb-pro@latest
| Package | Purpose |
|---|---|
@lancedb/lancedb ≥0.26.2 |
Vector database (ANN + FTS) |
openai ≥6.21.0 |
OpenAI-compatible Embedding API client |
@sinclair/typebox 0.34.48 |
JSON Schema type definitions |
Full list: Contributors
MIT