grepmax 0.7.37 → 0.7.39
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +37 -28
- package/dist/commands/setup.js +6 -24
- package/package.json +1 -1
- package/plugins/grepmax/.claude-plugin/plugin.json +1 -1
package/README.md
CHANGED
|
@@ -29,44 +29,51 @@ Natural-language search that works like `grep`. Fast, local, and built for codin
|
|
|
29
29
|
|
|
30
30
|
## Quick Start
|
|
31
31
|
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
2. **Setup (Recommended)**
|
|
32
|
+
```bash
|
|
33
|
+
npm install -g grepmax # 1. Install
|
|
34
|
+
cd my-repo && gmax index # 2. Index (models download automatically)
|
|
35
|
+
gmax "where do we handle auth?" # 3. Search
|
|
36
|
+
```
|
|
38
37
|
|
|
39
|
-
|
|
40
|
-
gmax setup
|
|
41
|
-
```
|
|
38
|
+
That's it. No setup required — gmax auto-detects your platform (GPU on Apple Silicon, CPU elsewhere) and downloads models on first use.
|
|
42
39
|
|
|
43
|
-
|
|
40
|
+
### Optional: Interactive Setup
|
|
44
41
|
|
|
45
|
-
|
|
42
|
+
```bash
|
|
43
|
+
gmax setup # Choose model tier + embedding mode interactively
|
|
44
|
+
```
|
|
46
45
|
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
```
|
|
46
|
+
Run this if you want to:
|
|
47
|
+
- Switch between **CPU** (ONNX, works everywhere) and **GPU** (MLX, Apple Silicon only, ~3x faster)
|
|
48
|
+
- Choose between **small** model (384d, 47M params, fast) and **standard** model (768d, 149M params, better quality)
|
|
51
49
|
|
|
52
|
-
|
|
50
|
+
### Quick Config (Non-Interactive)
|
|
53
51
|
|
|
54
|
-
|
|
52
|
+
```bash
|
|
53
|
+
gmax config # View current settings
|
|
54
|
+
gmax config --embed-mode cpu # Switch to CPU
|
|
55
|
+
gmax config --embed-mode gpu # Switch to GPU (Apple Silicon only)
|
|
56
|
+
gmax config --model-tier standard # Switch to larger model
|
|
57
|
+
```
|
|
55
58
|
|
|
56
|
-
|
|
57
|
-
gmax "where do we handle authentication?"
|
|
58
|
-
```
|
|
59
|
+
### Verify Installation
|
|
59
60
|
|
|
60
|
-
|
|
61
|
+
```bash
|
|
62
|
+
gmax doctor # Check models, index, servers
|
|
63
|
+
```
|
|
61
64
|
|
|
62
|
-
|
|
63
|
-
gmax trace "function_name"
|
|
64
|
-
```
|
|
65
|
-
See who calls a function (upstream dependencies) and what it calls (downstream dependencies).
|
|
65
|
+
### Core Commands
|
|
66
66
|
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
67
|
+
```bash
|
|
68
|
+
gmax "where do we handle auth?" # Semantic search
|
|
69
|
+
gmax "VectorDB" --symbol --agent # Search + call graph (compact output)
|
|
70
|
+
gmax trace handleAuth -d 2 # Call graph (2-hop)
|
|
71
|
+
gmax skeleton src/lib/search/ # File structure (directory)
|
|
72
|
+
gmax project # Project overview
|
|
73
|
+
gmax related src/lib/syncer.ts # Dependencies + dependents
|
|
74
|
+
gmax recent # Recently modified files
|
|
75
|
+
gmax symbols auth # List indexed symbols
|
|
76
|
+
```
|
|
70
77
|
|
|
71
78
|
In our public benchmarks, `grepmax` can save about 20% of your LLM tokens and deliver a 30% speedup.
|
|
72
79
|
|
|
@@ -82,6 +89,8 @@ In our public benchmarks, `grepmax` can save about 20% of your LLM tokens and de
|
|
|
82
89
|
2. Open Claude Code — the plugin auto-starts the MLX GPU server and a background file watcher.
|
|
83
90
|
3. Claude uses `gmax` for semantic searches automatically via MCP tools.
|
|
84
91
|
|
|
92
|
+
Plugin files (skill instructions, hooks) auto-update when you run `npm update -g grepmax` — no need to re-run `install-claude-code`.
|
|
93
|
+
|
|
85
94
|
### Opencode
|
|
86
95
|
1. Run `gmax install-opencode`
|
|
87
96
|
2. OC uses `gmax` for semantic searches via MCP.
|
package/dist/commands/setup.js
CHANGED
|
@@ -118,30 +118,12 @@ exports.setup = new commander_1.Command("setup")
|
|
|
118
118
|
return;
|
|
119
119
|
}
|
|
120
120
|
const selectedTier = config_1.MODEL_TIERS[modelTier];
|
|
121
|
-
// Step 5: Embed mode
|
|
122
|
-
const
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
|
|
127
|
-
label: "CPU only",
|
|
128
|
-
hint: "ONNX — works everywhere",
|
|
129
|
-
},
|
|
130
|
-
{
|
|
131
|
-
value: "gpu",
|
|
132
|
-
label: "GPU (MLX)",
|
|
133
|
-
hint: "Apple Silicon only, faster indexing + search",
|
|
134
|
-
},
|
|
135
|
-
],
|
|
136
|
-
initialValue: (_c = existingConfig === null || existingConfig === void 0 ? void 0 : existingConfig.embedMode) !== null && _c !== void 0 ? _c : (process.arch === "arm64" && process.platform === "darwin"
|
|
137
|
-
? "gpu"
|
|
138
|
-
: "cpu"),
|
|
139
|
-
});
|
|
140
|
-
if (p.isCancel(embedMode)) {
|
|
141
|
-
p.cancel("Setup cancelled");
|
|
142
|
-
yield (0, exit_1.gracefulExit)();
|
|
143
|
-
return;
|
|
144
|
-
}
|
|
121
|
+
// Step 5: Embed mode — auto-detect, no prompt needed
|
|
122
|
+
const isAppleSilicon = process.arch === "arm64" && process.platform === "darwin";
|
|
123
|
+
const embedMode = (_c = existingConfig === null || existingConfig === void 0 ? void 0 : existingConfig.embedMode) !== null && _c !== void 0 ? _c : (isAppleSilicon ? "gpu" : "cpu");
|
|
124
|
+
p.log.info(isAppleSilicon
|
|
125
|
+
? "Apple Silicon detected — using GPU acceleration (MLX)"
|
|
126
|
+
: "Using CPU embeddings (ONNX)");
|
|
145
127
|
const mlxModel = embedMode === "gpu" ? selectedTier.mlxModel : undefined;
|
|
146
128
|
// Step 6: Write configs
|
|
147
129
|
(0, index_config_1.writeSetupConfig)(paths.configPath, {
|
package/package.json
CHANGED