moltmind 0.7.2 → 0.7.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -4,6 +4,22 @@ Persistent semantic memory and session continuity for AI agents. One install, ze
4
4
 
5
5
  MoltMind is an [MCP](https://modelcontextprotocol.io) server that gives your AI agent long-term memory across sessions — storing learnings, decisions, error fixes, and handoff context using local SQLite and embeddings. No API keys, no cloud, no accounts needed.
6
6
 
7
+ ## Why MoltMind?
8
+
9
+ Every time your AI agent starts a new conversation, it forgets everything. It spends 1-2 minutes re-reading your files, re-learning your architecture, and re-discovering decisions you already made. MoltMind gives it memory — your agent picks up right where it left off in seconds.
10
+
11
+ | | Without MoltMind | With MoltMind |
12
+ |--|-----------------|---------------|
13
+ | **Model used** | Claude Opus 4.6 ($5/$25 per 1M tokens) | |
14
+ | **Time per session** | 1-2 min re-exploring | Seconds to resume |
15
+ | **Cost per session** | ~$0.09 | ~$0.009 |
16
+ | **20 sessions** | $1.80 | $0.18 |
17
+ | **Daily use (1 year)** | $32.85 | $3.29 |
18
+ | **Time saved (1 year)** | — | **~6 hours** |
19
+ | **Money saved (1 year)** | — | **~$30** |
20
+
21
+ > Assumes ~8,000 input + ~2,000 output tokens per cold start, ~825 input + ~200 output per resume. Savings scale with usage — power users save more.
22
+
7
23
  ## Quick Start
8
24
 
9
25
  ### Claude Code
@@ -12,7 +28,15 @@ MoltMind is an [MCP](https://modelcontextprotocol.io) server that gives your AI
12
28
  claude mcp add moltmind -- npx -y moltmind
13
29
  ```
14
30
 
15
- Restart Claude Code, then run `/mcp` to verify. Add `--moltbook` for social features ([moltbook.com](https://moltbook.com)).
31
+ Restart Claude Code, then run `/mcp` to verify.
32
+
33
+ With moltbook social features:
34
+
35
+ ```bash
36
+ claude mcp add moltmind -- npx -y moltmind --moltbook
37
+ ```
38
+
39
+ See [moltbook.com](https://moltbook.com) for the agent social network.
16
40
 
17
41
  ### Other Clients
18
42
 
@@ -73,11 +97,23 @@ npm uninstall -g moltmind # then let npx handle it
73
97
 
74
98
  ## How It Works
75
99
 
76
- **Memory & Search** — Memories are stored in local SQLite with FTS5. Each has a type (`learning`, `error`, `decision`, `plan`, `raw`), tags, and a tier (`hot`, `warm`, `cold`, `archived`). `mm_recall` runs hybrid search: semantic similarity (0.7 weight) via a local [MiniLM-L6-v2](https://huggingface.co/Xenova/all-MiniLM-L6-v2) embedding model plus FTS5 keyword matching (0.3 weight). If the embedding model isn't available, it falls back to keyword-only.
100
+ **Memory & Search** — Your agent stores memories in a local database. When it needs to find something, MoltMind searches by meaning (not just keywords) so searching for "API port" finds a memory about "our server runs on port 8080". If the search model isn't downloaded yet, it falls back to keyword matching.
101
+
102
+ **Sessions & Handoffs** — Sessions are auto-created on startup and auto-paused on shutdown. Your agent saves where it left off and picks up seamlessly next time. Handoffs let one agent pass context to another with structured goal/state/next-action documents.
103
+
104
+ **Diagnostics** — Every tool call is logged locally with timing and success/failure. `mm_status` shows health, `mm_metrics` shows usage stats and token savings. All data stays on your machine.
77
105
 
78
- **Sessions & Handoffs** Sessions are auto-created on startup and auto-paused on shutdown. `mm_session_save` captures what happened and where you left off; `mm_session_resume` restores full context. `mm_handoff_create` structures goal/state/next-action for agent-to-agent transfers. All tool calls are tagged with session IDs for traceability.
106
+ ## What It Costs (Tokens)
79
107
 
80
- **Diagnostics** Every tool call is logged locally with latency and success/failure. `mm_status` shows health score, `mm_metrics` shows per-tool usage stats, error rates, and token savings. All data stays on your machine.
108
+ Every MCP tool adds a small overhead to each request because the AI needs to know what tools are available. Here's what MoltMind costs and what it saves you:
109
+
110
+ | | Cost per request | In dollars |
111
+ |--|-----------------|------------|
112
+ | MoltMind overhead (14 tools) | ~500 tokens | ~$0.0015 |
113
+ | With prompt caching | ~50 tokens | ~$0.00015 |
114
+ | **Session resume (saves you)** | **~7,675 tokens** | **~$0.023** |
115
+
116
+ **Bottom line:** MoltMind pays for itself after a single session resume. Every conversation after that is pure savings.
81
117
 
82
118
  ## Free vs Pro
83
119
 
@@ -91,27 +127,17 @@ npm uninstall -g moltmind # then let npx handle it
91
127
 
92
128
  Upgrade: `npx moltmind --upgrade`
93
129
 
94
- ## Token Cost
95
-
96
- MCP tools add overhead because descriptions are sent with every request. MoltMind pays for itself quickly:
130
+ ## Search Performance (Pro)
97
131
 
98
- | Mode | Overhead per request |
99
- |------|---------------------|
100
- | Default (14 tools) | ~500 tokens |
101
- | + Moltbook (21 tools) | ~1,000 tokens |
102
- | With prompt caching | ~50 tokens |
132
+ Pro tier uses [Zvec ANN](https://github.com/ariv14/zvec-native) for fast memory search. Here's what that means in practice:
103
133
 
104
- ### Session resume vs cold start
134
+ **Accuracy** At 1,000 memories (a typical heavy user), Zvec finds **98% of the exact same results** as an exhaustive search. Your agent gets the right answer almost every time.
105
135
 
106
- Without MoltMind, re-exploring a codebase costs ~8,000 tokens per session. `mm_session_resume` restores context in ~325 tokens.
136
+ **Speed** Search takes **under 1ms** at 1,000 memories. At 10,000 memories, it's still under 5ms. Your agent won't notice any delay.
107
137
 
108
- | Scenario | Without | With MoltMind | Savings |
109
- |----------|---------|---------------|---------|
110
- | Single resume | ~8,000 | ~825 | 90% |
111
- | 5-session project | ~40,000 | ~7,500 | 81% |
112
- | 20-session project | ~160,000 | ~40,200 | 75% |
138
+ **Reliability** Handles **330+ searches per second** with zero latency spikes. Deleted memories never come back. Results are deterministic.
113
139
 
114
- Run `npm run benchmark` for latency measurements and projected savings. See [RUNBOOK.md](RUNBOOK.md) for detailed results.
140
+ See [BENCHMARK_RESULTS.md](BENCHMARK_RESULTS.md) for the full report, or [RUNBOOK.md](RUNBOOK.md) for how to run benchmarks yourself.
115
141
 
116
142
  ## Data Storage
117
143
 
package/dist/index.js CHANGED
@@ -25,7 +25,7 @@ const moltbookInstructions = isMoltbookEnabled()
25
25
  : "";
26
26
  const server = new McpServer({
27
27
  name: "moltmind",
28
- version: "0.7.2",
28
+ version: "0.7.4",
29
29
  }, {
30
30
  instructions: `MoltMind provides persistent memory and session continuity. On startup, call mm_session_resume to restore context from previous sessions. Before disconnecting or when a task is complete, call mm_session_save to preserve session state. Use mm_handoff_create to checkpoint progress during long tasks.${moltbookInstructions}`,
31
31
  });
@@ -9,7 +9,7 @@ export async function handleMmStatus() {
9
9
  const uptimeSeconds = Math.floor((Date.now() - startTime) / 1000);
10
10
  return {
11
11
  success: true,
12
- version: "0.7.2",
12
+ version: "0.7.4",
13
13
  tier: isProTier() ? "pro" : "free",
14
14
  usage: checkStoreLimits().message,
15
15
  db_stats: stats,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "moltmind",
3
- "version": "0.7.2",
3
+ "version": "0.7.4",
4
4
  "description": "Agent Memory MCP Server — persistent semantic memory and session continuity for AI agents",
5
5
  "type": "module",
6
6
  "bin": {