claude-memory-agent 2.0.0 → 2.0.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +12 -6
- package/bin/cli.js +11 -7
- package/bin/postinstall.js +74 -20
- package/install.py +16 -1
- package/package.json +2 -10
package/README.md
CHANGED
|
@@ -13,16 +13,22 @@ A persistent semantic memory system for Claude Code sessions. Stores memories, t
|
|
|
13
13
|
|
|
14
14
|
## Quick Start
|
|
15
15
|
|
|
16
|
-
### 1. Prerequisites
|
|
16
|
+
### 1. Prerequisites (Install These First!)
|
|
17
17
|
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
18
|
+
| Requirement | Download | Why Needed |
|
|
19
|
+
|-------------|----------|------------|
|
|
20
|
+
| **Python 3.9+** | https://python.org | Runs the agent |
|
|
21
|
+
| **Ollama** | https://ollama.ai/download | Provides embeddings |
|
|
22
|
+
| **Node.js 16+** | https://nodejs.org | For npm install |
|
|
23
|
+
|
|
24
|
+
**After installing Ollama, run these commands:**
|
|
21
25
|
|
|
22
26
|
```bash
|
|
23
|
-
#
|
|
27
|
+
# Download the embedding model (required!)
|
|
24
28
|
ollama pull nomic-embed-text
|
|
25
|
-
|
|
29
|
+
|
|
30
|
+
# Start Ollama (runs in background)
|
|
31
|
+
ollama serve
|
|
26
32
|
```
|
|
27
33
|
|
|
28
34
|
### 2. Installation
|
package/bin/cli.js
CHANGED
|
@@ -113,16 +113,20 @@ COMMANDS:
|
|
|
113
113
|
help Show this help message
|
|
114
114
|
|
|
115
115
|
QUICK START:
|
|
116
|
-
1.
|
|
117
|
-
2.
|
|
118
|
-
3.
|
|
116
|
+
1. Install Ollama: https://ollama.ai/download
|
|
117
|
+
2. Pull embedding model: ollama pull nomic-embed-text
|
|
118
|
+
3. Start Ollama: ollama serve
|
|
119
|
+
4. Configure: claude-memory-agent install
|
|
120
|
+
5. Start agent: claude-memory-agent start
|
|
121
|
+
6. Open dashboard: claude-memory-agent dashboard
|
|
119
122
|
|
|
120
123
|
REQUIREMENTS:
|
|
121
|
-
- Python 3.9+
|
|
122
|
-
- Ollama
|
|
123
|
-
-
|
|
124
|
+
- Python 3.9+ (https://python.org)
|
|
125
|
+
- Ollama (https://ollama.ai)
|
|
126
|
+
- nomic-embed-text (ollama pull nomic-embed-text)
|
|
127
|
+
- Claude Code (npm install -g @anthropic-ai/claude-code)
|
|
124
128
|
|
|
125
|
-
For more info: https://
|
|
129
|
+
For more info: https://www.npmjs.com/package/claude-memory-agent
|
|
126
130
|
`);
|
|
127
131
|
}
|
|
128
132
|
|
package/bin/postinstall.js
CHANGED
|
@@ -175,36 +175,90 @@ async function main() {
|
|
|
175
175
|
// Create .env file
|
|
176
176
|
createEnvFile();
|
|
177
177
|
|
|
178
|
-
// Check Ollama
|
|
178
|
+
// Check Ollama
|
|
179
179
|
log('Checking Ollama...');
|
|
180
180
|
const ollamaRunning = await checkOllama();
|
|
181
|
+
|
|
181
182
|
if (ollamaRunning) {
|
|
182
183
|
success('Ollama is running');
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
184
|
+
// Check if model is installed
|
|
185
|
+
const modelInstalled = await checkOllamaModel();
|
|
186
|
+
if (modelInstalled) {
|
|
187
|
+
success('Embedding model is ready');
|
|
188
|
+
} else {
|
|
189
|
+
warn('Embedding model not found');
|
|
190
|
+
}
|
|
187
191
|
}
|
|
188
192
|
|
|
189
|
-
// Print next steps
|
|
193
|
+
// Print next steps based on what's missing
|
|
190
194
|
header('Installation Complete!');
|
|
191
|
-
console.log('Next steps:');
|
|
192
|
-
console.log('');
|
|
193
|
-
console.log(' 1. Run the setup wizard:');
|
|
194
|
-
console.log(' claude-memory-agent install');
|
|
195
|
-
console.log('');
|
|
196
|
-
console.log(' 2. Start the agent:');
|
|
197
|
-
console.log(' claude-memory-agent start');
|
|
198
|
-
console.log('');
|
|
199
|
-
console.log(' 3. Open the dashboard:');
|
|
200
|
-
console.log(' claude-memory-agent dashboard');
|
|
201
|
-
console.log('');
|
|
202
195
|
|
|
203
196
|
if (!ollamaRunning) {
|
|
204
|
-
|
|
205
|
-
console.log(
|
|
206
|
-
console.log(
|
|
197
|
+
// Ollama not installed - show big warning
|
|
198
|
+
console.log(`${colors.red}${colors.bold}╔════════════════════════════════════════════════════════════╗${colors.reset}`);
|
|
199
|
+
console.log(`${colors.red}${colors.bold}║ REQUIRED: Install Ollama for semantic search to work ║${colors.reset}`);
|
|
200
|
+
console.log(`${colors.red}${colors.bold}╚════════════════════════════════════════════════════════════╝${colors.reset}`);
|
|
201
|
+
console.log('');
|
|
202
|
+
console.log(' Ollama is required for the memory agent to work properly.');
|
|
203
|
+
console.log(' Without it, semantic search will be disabled.');
|
|
204
|
+
console.log('');
|
|
205
|
+
console.log(` ${colors.cyan}Step 1:${colors.reset} Download and install Ollama:`);
|
|
206
|
+
console.log(` ${colors.bold}https://ollama.ai/download${colors.reset}`);
|
|
207
|
+
console.log('');
|
|
208
|
+
console.log(` ${colors.cyan}Step 2:${colors.reset} After installing, open a terminal and run:`);
|
|
209
|
+
console.log(` ${colors.bold}ollama pull nomic-embed-text${colors.reset}`);
|
|
210
|
+
console.log('');
|
|
211
|
+
console.log(` ${colors.cyan}Step 3:${colors.reset} Start Ollama (it runs in background):`);
|
|
212
|
+
console.log(` ${colors.bold}ollama serve${colors.reset}`);
|
|
213
|
+
console.log('');
|
|
214
|
+
console.log(` ${colors.cyan}Step 4:${colors.reset} Then run the setup:`);
|
|
215
|
+
console.log(` ${colors.bold}claude-memory-agent install${colors.reset}`);
|
|
207
216
|
console.log('');
|
|
217
|
+
} else {
|
|
218
|
+
// Ollama running - show normal next steps
|
|
219
|
+
console.log('Next steps:');
|
|
220
|
+
console.log('');
|
|
221
|
+
console.log(' 1. Run the setup wizard:');
|
|
222
|
+
console.log(` ${colors.bold}claude-memory-agent install${colors.reset}`);
|
|
223
|
+
console.log('');
|
|
224
|
+
console.log(' 2. Start the agent:');
|
|
225
|
+
console.log(` ${colors.bold}claude-memory-agent start${colors.reset}`);
|
|
226
|
+
console.log('');
|
|
227
|
+
console.log(' 3. Open the dashboard:');
|
|
228
|
+
console.log(` ${colors.bold}claude-memory-agent dashboard${colors.reset}`);
|
|
229
|
+
console.log('');
|
|
230
|
+
}
|
|
231
|
+
}
|
|
232
|
+
|
|
233
|
+
// Check if embedding model is installed in Ollama
|
|
234
|
+
async function checkOllamaModel() {
|
|
235
|
+
try {
|
|
236
|
+
const http = require('http');
|
|
237
|
+
return new Promise((resolve) => {
|
|
238
|
+
const req = http.get('http://localhost:11434/api/tags', { timeout: 2000 }, (res) => {
|
|
239
|
+
let data = '';
|
|
240
|
+
res.on('data', chunk => data += chunk);
|
|
241
|
+
res.on('end', () => {
|
|
242
|
+
try {
|
|
243
|
+
const json = JSON.parse(data);
|
|
244
|
+
const models = json.models || [];
|
|
245
|
+
const hasModel = models.some(m =>
|
|
246
|
+
m.name && m.name.includes('nomic-embed-text')
|
|
247
|
+
);
|
|
248
|
+
resolve(hasModel);
|
|
249
|
+
} catch (e) {
|
|
250
|
+
resolve(false);
|
|
251
|
+
}
|
|
252
|
+
});
|
|
253
|
+
});
|
|
254
|
+
req.on('error', () => resolve(false));
|
|
255
|
+
req.on('timeout', () => {
|
|
256
|
+
req.destroy();
|
|
257
|
+
resolve(false);
|
|
258
|
+
});
|
|
259
|
+
});
|
|
260
|
+
} catch (e) {
|
|
261
|
+
return false;
|
|
208
262
|
}
|
|
209
263
|
}
|
|
210
264
|
|
package/install.py
CHANGED
|
@@ -276,7 +276,22 @@ def check_ollama() -> bool:
|
|
|
276
276
|
return True
|
|
277
277
|
except Exception:
|
|
278
278
|
pass
|
|
279
|
-
|
|
279
|
+
|
|
280
|
+
print_warning("Ollama not detected")
|
|
281
|
+
print("")
|
|
282
|
+
print(" " + "="*56)
|
|
283
|
+
print(" OLLAMA REQUIRED FOR SEMANTIC SEARCH")
|
|
284
|
+
print(" " + "="*56)
|
|
285
|
+
print("")
|
|
286
|
+
print(" The memory agent needs Ollama for embeddings.")
|
|
287
|
+
print(" Without it, semantic search will not work.")
|
|
288
|
+
print("")
|
|
289
|
+
print(" To install Ollama:")
|
|
290
|
+
print(" 1. Download from: https://ollama.ai/download")
|
|
291
|
+
print(" 2. Install and run: ollama pull nomic-embed-text")
|
|
292
|
+
print(" 3. Start Ollama: ollama serve")
|
|
293
|
+
print(" 4. Re-run this installer")
|
|
294
|
+
print("")
|
|
280
295
|
return False
|
|
281
296
|
|
|
282
297
|
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "claude-memory-agent",
|
|
3
|
-
"version": "2.0.
|
|
3
|
+
"version": "2.0.1",
|
|
4
4
|
"description": "Persistent semantic memory system for Claude Code sessions with anti-hallucination grounding",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"claude",
|
|
@@ -12,16 +12,8 @@
|
|
|
12
12
|
"embeddings",
|
|
13
13
|
"ollama"
|
|
14
14
|
],
|
|
15
|
-
"author": "
|
|
15
|
+
"author": "Claude Memory Agent Contributors",
|
|
16
16
|
"license": "MIT",
|
|
17
|
-
"repository": {
|
|
18
|
-
"type": "git",
|
|
19
|
-
"url": "https://github.com/yourusername/claude-memory-agent.git"
|
|
20
|
-
},
|
|
21
|
-
"homepage": "https://github.com/yourusername/claude-memory-agent#readme",
|
|
22
|
-
"bugs": {
|
|
23
|
-
"url": "https://github.com/yourusername/claude-memory-agent/issues"
|
|
24
|
-
},
|
|
25
17
|
"bin": {
|
|
26
18
|
"claude-memory-agent": "./bin/cli.js",
|
|
27
19
|
"claude-memory": "./bin/cli.js"
|