@goondocks/myco 0.4.2 → 0.4.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude-plugin/marketplace.json +1 -1
- package/.claude-plugin/plugin.json +1 -1
- package/dist/chunk-2AMAOSRF.js +105 -0
- package/dist/chunk-2AMAOSRF.js.map +1 -0
- package/dist/chunk-3F63SFZZ.js +381 -0
- package/dist/chunk-3F63SFZZ.js.map +1 -0
- package/dist/{chunk-WBT5DWGC.js → chunk-42R7KVAW.js} +2 -2
- package/dist/{chunk-GFBG73P4.js → chunk-5FIIK27E.js} +3 -3
- package/dist/{chunk-XCPQHC4X.js → chunk-6CAKKNGD.js} +2 -2
- package/dist/{chunk-I7PNZEBO.js → chunk-6LTNFMXO.js} +12 -1
- package/dist/{chunk-I7PNZEBO.js.map → chunk-6LTNFMXO.js.map} +1 -1
- package/dist/{chunk-V2OWD2VV.js → chunk-DKHYIA2V.js} +24 -146
- package/dist/chunk-DKHYIA2V.js.map +1 -0
- package/dist/{chunk-BNIYWCST.js → chunk-EQVQEFOA.js} +2 -2
- package/dist/{chunk-FPEDTLQ6.js → chunk-JJL6AMDA.js} +3 -101
- package/dist/chunk-JJL6AMDA.js.map +1 -0
- package/dist/{chunk-OUFSLZTX.js → chunk-KDWBZSOB.js} +21 -9
- package/dist/chunk-KDWBZSOB.js.map +1 -0
- package/dist/{chunk-67R6EMYD.js → chunk-OPO47BVS.js} +31 -52
- package/dist/chunk-OPO47BVS.js.map +1 -0
- package/dist/{chunk-IYFKPSRP.js → chunk-OSZRLHIJ.js} +3 -3
- package/dist/chunk-PD7LV22R.js +150 -0
- package/dist/chunk-PD7LV22R.js.map +1 -0
- package/dist/{chunk-JBD5KP5G.js → chunk-TDLQBGKA.js} +6 -2
- package/dist/chunk-TDLQBGKA.js.map +1 -0
- package/dist/{chunk-2GJFTIWX.js → chunk-TK2ZYIAL.js} +2 -2
- package/dist/{chunk-ZCBL5HER.js → chunk-XIIVIMFC.js} +2 -2
- package/dist/{cli-PMOFCZQL.js → cli-WOM4Z2Z4.js} +21 -18
- package/dist/cli-WOM4Z2Z4.js.map +1 -0
- package/dist/{client-5SUO2UYH.js → client-XCNF6NFT.js} +5 -5
- package/dist/{detect-providers-IRL2TTLK.js → detect-providers-CQSPTW2B.js} +3 -3
- package/dist/digest-WTS6S4XP.js +96 -0
- package/dist/digest-WTS6S4XP.js.map +1 -0
- package/dist/{init-NUF5UBUJ.js → init-VPLUEULI.js} +5 -5
- package/dist/{main-2XEBVUR6.js → main-OGXH6XWO.js} +230 -575
- package/dist/main-OGXH6XWO.js.map +1 -0
- package/dist/{rebuild-E6YFIRYZ.js → rebuild-Z4YUY6HT.js} +8 -7
- package/dist/{rebuild-E6YFIRYZ.js.map → rebuild-Z4YUY6HT.js.map} +1 -1
- package/dist/{reprocess-7G7KQWCN.js → reprocess-DMGPZTLC.js} +91 -20
- package/dist/reprocess-DMGPZTLC.js.map +1 -0
- package/dist/{restart-ABW4ZK3P.js → restart-QCQQ55KX.js} +6 -6
- package/dist/{search-MPD7SFK6.js → search-ACEFQOUW.js} +6 -6
- package/dist/{server-NZLZRITH.js → server-BQ3DWKZ6.js} +16 -14
- package/dist/{server-NZLZRITH.js.map → server-BQ3DWKZ6.js.map} +1 -1
- package/dist/{session-start-YB4A4PZB.js → session-start-BXRTKS4X.js} +6 -6
- package/dist/{setup-digest-K732MGOJ.js → setup-digest-EJXSQGZ5.js} +5 -5
- package/dist/{setup-llm-XCCH5LYD.js → setup-llm-P3MLWUDR.js} +5 -5
- package/dist/src/cli.js +4 -4
- package/dist/src/daemon/main.js +4 -4
- package/dist/src/hooks/post-tool-use.js +5 -5
- package/dist/src/hooks/session-end.js +5 -5
- package/dist/src/hooks/session-start.js +4 -4
- package/dist/src/hooks/stop.js +6 -6
- package/dist/src/hooks/stop.js.map +1 -1
- package/dist/src/hooks/user-prompt-submit.js +5 -5
- package/dist/src/mcp/server.js +4 -4
- package/dist/src/prompts/extraction.md +1 -1
- package/dist/src/prompts/summary.md +1 -11
- package/dist/{stats-6G7SN5YZ.js → stats-3FAP5FKV.js} +5 -5
- package/dist/{verify-JFHQH55Z.js → verify-3FTCOULE.js} +4 -4
- package/dist/{version-5B2TWXQJ.js → version-AL67JH7X.js} +4 -4
- package/package.json +1 -1
- package/skills/myco/SKILL.md +4 -0
- package/skills/myco/references/reconfiguration.md +92 -0
- package/skills/setup/SKILL.md +59 -31
- package/skills/setup/references/model-recommendations.md +49 -43
- package/dist/chunk-67R6EMYD.js.map +0 -1
- package/dist/chunk-FPEDTLQ6.js.map +0 -1
- package/dist/chunk-JBD5KP5G.js.map +0 -1
- package/dist/chunk-OUFSLZTX.js.map +0 -1
- package/dist/chunk-V2OWD2VV.js.map +0 -1
- package/dist/cli-PMOFCZQL.js.map +0 -1
- package/dist/main-2XEBVUR6.js.map +0 -1
- package/dist/reprocess-7G7KQWCN.js.map +0 -1
- /package/dist/{chunk-WBT5DWGC.js.map → chunk-42R7KVAW.js.map} +0 -0
- /package/dist/{chunk-GFBG73P4.js.map → chunk-5FIIK27E.js.map} +0 -0
- /package/dist/{chunk-XCPQHC4X.js.map → chunk-6CAKKNGD.js.map} +0 -0
- /package/dist/{chunk-BNIYWCST.js.map → chunk-EQVQEFOA.js.map} +0 -0
- /package/dist/{chunk-IYFKPSRP.js.map → chunk-OSZRLHIJ.js.map} +0 -0
- /package/dist/{chunk-2GJFTIWX.js.map → chunk-TK2ZYIAL.js.map} +0 -0
- /package/dist/{chunk-ZCBL5HER.js.map → chunk-XIIVIMFC.js.map} +0 -0
- /package/dist/{client-5SUO2UYH.js.map → client-XCNF6NFT.js.map} +0 -0
- /package/dist/{detect-providers-IRL2TTLK.js.map → detect-providers-CQSPTW2B.js.map} +0 -0
- /package/dist/{init-NUF5UBUJ.js.map → init-VPLUEULI.js.map} +0 -0
- /package/dist/{restart-ABW4ZK3P.js.map → restart-QCQQ55KX.js.map} +0 -0
- /package/dist/{search-MPD7SFK6.js.map → search-ACEFQOUW.js.map} +0 -0
- /package/dist/{session-start-YB4A4PZB.js.map → session-start-BXRTKS4X.js.map} +0 -0
- /package/dist/{setup-digest-K732MGOJ.js.map → setup-digest-EJXSQGZ5.js.map} +0 -0
- /package/dist/{setup-llm-XCCH5LYD.js.map → setup-llm-P3MLWUDR.js.map} +0 -0
- /package/dist/{stats-6G7SN5YZ.js.map → stats-3FAP5FKV.js.map} +0 -0
- /package/dist/{verify-JFHQH55Z.js.map → verify-3FTCOULE.js.map} +0 -0
- /package/dist/{version-5B2TWXQJ.js.map → version-AL67JH7X.js.map} +0 -0
|
@@ -12,21 +12,21 @@ import {
|
|
|
12
12
|
import "./chunk-6UJWI4IW.js";
|
|
13
13
|
import {
|
|
14
14
|
readStdin
|
|
15
|
-
} from "./chunk-
|
|
15
|
+
} from "./chunk-XIIVIMFC.js";
|
|
16
16
|
import {
|
|
17
17
|
DaemonClient
|
|
18
|
-
} from "./chunk-
|
|
19
|
-
import "./chunk-
|
|
18
|
+
} from "./chunk-KDWBZSOB.js";
|
|
19
|
+
import "./chunk-TK2ZYIAL.js";
|
|
20
20
|
import {
|
|
21
21
|
resolveVaultDir
|
|
22
22
|
} from "./chunk-N33KUCFP.js";
|
|
23
|
-
import "./chunk-
|
|
23
|
+
import "./chunk-EQVQEFOA.js";
|
|
24
24
|
import {
|
|
25
25
|
CONTEXT_PLAN_PREVIEW_CHARS,
|
|
26
26
|
CONTEXT_SESSION_PREVIEW_CHARS,
|
|
27
27
|
CONTEXT_SPORE_PREVIEW_CHARS,
|
|
28
28
|
estimateTokens
|
|
29
|
-
} from "./chunk-
|
|
29
|
+
} from "./chunk-TDLQBGKA.js";
|
|
30
30
|
import "./chunk-PZUWP5VK.js";
|
|
31
31
|
|
|
32
32
|
// src/context/relevance.ts
|
|
@@ -189,4 +189,4 @@ async function main() {
|
|
|
189
189
|
}
|
|
190
190
|
}
|
|
191
191
|
main();
|
|
192
|
-
//# sourceMappingURL=session-start-
|
|
192
|
+
//# sourceMappingURL=session-start-BXRTKS4X.js.map
|
|
@@ -2,14 +2,14 @@ import { createRequire as __cr } from 'node:module'; const require = __cr(import
|
|
|
2
2
|
import {
|
|
3
3
|
run
|
|
4
4
|
} from "./chunk-UKWO26VI.js";
|
|
5
|
-
import "./chunk-
|
|
5
|
+
import "./chunk-5FIIK27E.js";
|
|
6
6
|
import "./chunk-SAKJMNSR.js";
|
|
7
|
-
import "./chunk-
|
|
7
|
+
import "./chunk-OPO47BVS.js";
|
|
8
8
|
import "./chunk-6UJWI4IW.js";
|
|
9
|
-
import "./chunk-
|
|
10
|
-
import "./chunk-
|
|
9
|
+
import "./chunk-EQVQEFOA.js";
|
|
10
|
+
import "./chunk-TDLQBGKA.js";
|
|
11
11
|
import "./chunk-PZUWP5VK.js";
|
|
12
12
|
export {
|
|
13
13
|
run
|
|
14
14
|
};
|
|
15
|
-
//# sourceMappingURL=setup-digest-
|
|
15
|
+
//# sourceMappingURL=setup-digest-EJXSQGZ5.js.map
|
|
@@ -2,14 +2,14 @@ import { createRequire as __cr } from 'node:module'; const require = __cr(import
|
|
|
2
2
|
import {
|
|
3
3
|
run
|
|
4
4
|
} from "./chunk-T7OC6GH5.js";
|
|
5
|
-
import "./chunk-
|
|
5
|
+
import "./chunk-5FIIK27E.js";
|
|
6
6
|
import "./chunk-SAKJMNSR.js";
|
|
7
|
-
import "./chunk-
|
|
7
|
+
import "./chunk-OPO47BVS.js";
|
|
8
8
|
import "./chunk-6UJWI4IW.js";
|
|
9
|
-
import "./chunk-
|
|
10
|
-
import "./chunk-
|
|
9
|
+
import "./chunk-EQVQEFOA.js";
|
|
10
|
+
import "./chunk-TDLQBGKA.js";
|
|
11
11
|
import "./chunk-PZUWP5VK.js";
|
|
12
12
|
export {
|
|
13
13
|
run
|
|
14
14
|
};
|
|
15
|
-
//# sourceMappingURL=setup-llm-
|
|
15
|
+
//# sourceMappingURL=setup-llm-P3MLWUDR.js.map
|
package/dist/src/cli.js
CHANGED
|
@@ -2,12 +2,12 @@
|
|
|
2
2
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
3
3
|
import {
|
|
4
4
|
ensureNativeDeps
|
|
5
|
-
} from "../chunk-
|
|
6
|
-
import "../chunk-
|
|
7
|
-
import "../chunk-
|
|
5
|
+
} from "../chunk-6CAKKNGD.js";
|
|
6
|
+
import "../chunk-EQVQEFOA.js";
|
|
7
|
+
import "../chunk-TDLQBGKA.js";
|
|
8
8
|
import "../chunk-PZUWP5VK.js";
|
|
9
9
|
|
|
10
10
|
// src/entries/cli.ts
|
|
11
11
|
ensureNativeDeps();
|
|
12
|
-
await import("../cli-
|
|
12
|
+
await import("../cli-WOM4Z2Z4.js");
|
|
13
13
|
//# sourceMappingURL=cli.js.map
|
package/dist/src/daemon/main.js
CHANGED
|
@@ -1,13 +1,13 @@
|
|
|
1
1
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
2
2
|
import {
|
|
3
3
|
ensureNativeDeps
|
|
4
|
-
} from "../../chunk-
|
|
5
|
-
import "../../chunk-
|
|
6
|
-
import "../../chunk-
|
|
4
|
+
} from "../../chunk-6CAKKNGD.js";
|
|
5
|
+
import "../../chunk-EQVQEFOA.js";
|
|
6
|
+
import "../../chunk-TDLQBGKA.js";
|
|
7
7
|
import "../../chunk-PZUWP5VK.js";
|
|
8
8
|
|
|
9
9
|
// src/entries/daemon.ts
|
|
10
10
|
ensureNativeDeps();
|
|
11
|
-
var { main } = await import("../../main-
|
|
11
|
+
var { main } = await import("../../main-OGXH6XWO.js");
|
|
12
12
|
await main();
|
|
13
13
|
//# sourceMappingURL=main.js.map
|
|
@@ -4,18 +4,18 @@ import {
|
|
|
4
4
|
} from "../../chunk-HIN3UVOG.js";
|
|
5
5
|
import {
|
|
6
6
|
readStdin
|
|
7
|
-
} from "../../chunk-
|
|
7
|
+
} from "../../chunk-XIIVIMFC.js";
|
|
8
8
|
import {
|
|
9
9
|
DaemonClient
|
|
10
|
-
} from "../../chunk-
|
|
11
|
-
import "../../chunk-
|
|
10
|
+
} from "../../chunk-KDWBZSOB.js";
|
|
11
|
+
import "../../chunk-TK2ZYIAL.js";
|
|
12
12
|
import {
|
|
13
13
|
resolveVaultDir
|
|
14
14
|
} from "../../chunk-N33KUCFP.js";
|
|
15
|
-
import "../../chunk-
|
|
15
|
+
import "../../chunk-EQVQEFOA.js";
|
|
16
16
|
import {
|
|
17
17
|
TOOL_OUTPUT_PREVIEW_CHARS
|
|
18
|
-
} from "../../chunk-
|
|
18
|
+
} from "../../chunk-TDLQBGKA.js";
|
|
19
19
|
import "../../chunk-PZUWP5VK.js";
|
|
20
20
|
|
|
21
21
|
// src/hooks/post-tool-use.ts
|
|
@@ -1,16 +1,16 @@
|
|
|
1
1
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
2
2
|
import {
|
|
3
3
|
readStdin
|
|
4
|
-
} from "../../chunk-
|
|
4
|
+
} from "../../chunk-XIIVIMFC.js";
|
|
5
5
|
import {
|
|
6
6
|
DaemonClient
|
|
7
|
-
} from "../../chunk-
|
|
8
|
-
import "../../chunk-
|
|
7
|
+
} from "../../chunk-KDWBZSOB.js";
|
|
8
|
+
import "../../chunk-TK2ZYIAL.js";
|
|
9
9
|
import {
|
|
10
10
|
resolveVaultDir
|
|
11
11
|
} from "../../chunk-N33KUCFP.js";
|
|
12
|
-
import "../../chunk-
|
|
13
|
-
import "../../chunk-
|
|
12
|
+
import "../../chunk-EQVQEFOA.js";
|
|
13
|
+
import "../../chunk-TDLQBGKA.js";
|
|
14
14
|
import "../../chunk-PZUWP5VK.js";
|
|
15
15
|
|
|
16
16
|
// src/hooks/session-end.ts
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
2
2
|
import {
|
|
3
3
|
ensureNativeDeps
|
|
4
|
-
} from "../../chunk-
|
|
5
|
-
import "../../chunk-
|
|
6
|
-
import "../../chunk-
|
|
4
|
+
} from "../../chunk-6CAKKNGD.js";
|
|
5
|
+
import "../../chunk-EQVQEFOA.js";
|
|
6
|
+
import "../../chunk-TDLQBGKA.js";
|
|
7
7
|
import "../../chunk-PZUWP5VK.js";
|
|
8
8
|
|
|
9
9
|
// src/entries/session-start.ts
|
|
10
10
|
ensureNativeDeps();
|
|
11
|
-
await import("../../session-start-
|
|
11
|
+
await import("../../session-start-BXRTKS4X.js");
|
|
12
12
|
//# sourceMappingURL=session-start.js.map
|
package/dist/src/hooks/stop.js
CHANGED
|
@@ -5,16 +5,16 @@ import {
|
|
|
5
5
|
import "../../chunk-6UJWI4IW.js";
|
|
6
6
|
import {
|
|
7
7
|
readStdin
|
|
8
|
-
} from "../../chunk-
|
|
8
|
+
} from "../../chunk-XIIVIMFC.js";
|
|
9
9
|
import {
|
|
10
10
|
DaemonClient
|
|
11
|
-
} from "../../chunk-
|
|
12
|
-
import "../../chunk-
|
|
11
|
+
} from "../../chunk-KDWBZSOB.js";
|
|
12
|
+
import "../../chunk-TK2ZYIAL.js";
|
|
13
13
|
import {
|
|
14
14
|
resolveVaultDir
|
|
15
15
|
} from "../../chunk-N33KUCFP.js";
|
|
16
|
-
import "../../chunk-
|
|
17
|
-
import "../../chunk-
|
|
16
|
+
import "../../chunk-EQVQEFOA.js";
|
|
17
|
+
import "../../chunk-TDLQBGKA.js";
|
|
18
18
|
import "../../chunk-PZUWP5VK.js";
|
|
19
19
|
|
|
20
20
|
// src/hooks/stop.ts
|
|
@@ -29,7 +29,7 @@ async function main() {
|
|
|
29
29
|
if (!sessionId) return;
|
|
30
30
|
const config = loadConfig(VAULT_DIR);
|
|
31
31
|
const client = new DaemonClient(VAULT_DIR);
|
|
32
|
-
await client.ensureRunning();
|
|
32
|
+
await client.ensureRunning({ checkStale: false });
|
|
33
33
|
await client.post("/events/stop", {
|
|
34
34
|
session_id: sessionId,
|
|
35
35
|
user: config.team.user || void 0,
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../src/hooks/stop.ts"],"sourcesContent":["import { DaemonClient } from './client.js';\nimport { readStdin } from './read-stdin.js';\nimport { loadConfig } from '../config/loader.js';\nimport { resolveVaultDir } from '../vault/resolve.js';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\nasync function main() {\n const VAULT_DIR = resolveVaultDir();\n if (!fs.existsSync(path.join(VAULT_DIR, 'myco.yaml'))) return;\n\n try {\n const input = JSON.parse(await readStdin());\n const sessionId = input.session_id ?? process.env.MYCO_SESSION_ID;\n if (!sessionId) return;\n\n const config = loadConfig(VAULT_DIR);\n const client = new DaemonClient(VAULT_DIR);\n\n await client.ensureRunning();\n\n // Pass transcript_path and last_assistant_message from Claude Code.\n // These are provided by the hook system and eliminate the need to\n // scan directories or mine the transcript for the AI response.\n await client.post('/events/stop', {\n session_id: sessionId,\n user: config.team.user || undefined,\n transcript_path: input.transcript_path,\n last_assistant_message: input.last_assistant_message,\n });\n } catch (error) {\n process.stderr.write(`[myco] stop error: ${(error as Error).message}\\n`);\n }\n}\n\nmain();\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAIA,OAAO,QAAQ;AACf,OAAO,UAAU;AAEjB,eAAe,OAAO;AACpB,QAAM,YAAY,gBAAgB;AAClC,MAAI,CAAC,GAAG,WAAW,KAAK,KAAK,WAAW,WAAW,CAAC,EAAG;AAEvD,MAAI;AACF,UAAM,QAAQ,KAAK,MAAM,MAAM,UAAU,CAAC;AAC1C,UAAM,YAAY,MAAM,cAAc,QAAQ,IAAI;AAClD,QAAI,CAAC,UAAW;AAEhB,UAAM,SAAS,WAAW,SAAS;AACnC,UAAM,SAAS,IAAI,aAAa,SAAS;AAEzC,UAAM,OAAO,cAAc;
|
|
1
|
+
{"version":3,"sources":["../../../src/hooks/stop.ts"],"sourcesContent":["import { DaemonClient } from './client.js';\nimport { readStdin } from './read-stdin.js';\nimport { loadConfig } from '../config/loader.js';\nimport { resolveVaultDir } from '../vault/resolve.js';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\nasync function main() {\n const VAULT_DIR = resolveVaultDir();\n if (!fs.existsSync(path.join(VAULT_DIR, 'myco.yaml'))) return;\n\n try {\n const input = JSON.parse(await readStdin());\n const sessionId = input.session_id ?? process.env.MYCO_SESSION_ID;\n if (!sessionId) return;\n\n const config = loadConfig(VAULT_DIR);\n const client = new DaemonClient(VAULT_DIR);\n\n await client.ensureRunning({ checkStale: false });\n\n // Pass transcript_path and last_assistant_message from Claude Code.\n // These are provided by the hook system and eliminate the need to\n // scan directories or mine the transcript for the AI response.\n await client.post('/events/stop', {\n session_id: sessionId,\n user: config.team.user || undefined,\n transcript_path: input.transcript_path,\n last_assistant_message: input.last_assistant_message,\n });\n } catch (error) {\n process.stderr.write(`[myco] stop error: ${(error as Error).message}\\n`);\n }\n}\n\nmain();\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAIA,OAAO,QAAQ;AACf,OAAO,UAAU;AAEjB,eAAe,OAAO;AACpB,QAAM,YAAY,gBAAgB;AAClC,MAAI,CAAC,GAAG,WAAW,KAAK,KAAK,WAAW,WAAW,CAAC,EAAG;AAEvD,MAAI;AACF,UAAM,QAAQ,KAAK,MAAM,MAAM,UAAU,CAAC;AAC1C,UAAM,YAAY,MAAM,cAAc,QAAQ,IAAI;AAClD,QAAI,CAAC,UAAW;AAEhB,UAAM,SAAS,WAAW,SAAS;AACnC,UAAM,SAAS,IAAI,aAAa,SAAS;AAEzC,UAAM,OAAO,cAAc,EAAE,YAAY,MAAM,CAAC;AAKhD,UAAM,OAAO,KAAK,gBAAgB;AAAA,MAChC,YAAY;AAAA,MACZ,MAAM,OAAO,KAAK,QAAQ;AAAA,MAC1B,iBAAiB,MAAM;AAAA,MACvB,wBAAwB,MAAM;AAAA,IAChC,CAAC;AAAA,EACH,SAAS,OAAO;AACd,YAAQ,OAAO,MAAM,sBAAuB,MAAgB,OAAO;AAAA,CAAI;AAAA,EACzE;AACF;AAEA,KAAK;","names":[]}
|
|
@@ -4,16 +4,16 @@ import {
|
|
|
4
4
|
} from "../../chunk-HIN3UVOG.js";
|
|
5
5
|
import {
|
|
6
6
|
readStdin
|
|
7
|
-
} from "../../chunk-
|
|
7
|
+
} from "../../chunk-XIIVIMFC.js";
|
|
8
8
|
import {
|
|
9
9
|
DaemonClient
|
|
10
|
-
} from "../../chunk-
|
|
11
|
-
import "../../chunk-
|
|
10
|
+
} from "../../chunk-KDWBZSOB.js";
|
|
11
|
+
import "../../chunk-TK2ZYIAL.js";
|
|
12
12
|
import {
|
|
13
13
|
resolveVaultDir
|
|
14
14
|
} from "../../chunk-N33KUCFP.js";
|
|
15
|
-
import "../../chunk-
|
|
16
|
-
import "../../chunk-
|
|
15
|
+
import "../../chunk-EQVQEFOA.js";
|
|
16
|
+
import "../../chunk-TDLQBGKA.js";
|
|
17
17
|
import "../../chunk-PZUWP5VK.js";
|
|
18
18
|
|
|
19
19
|
// src/hooks/user-prompt-submit.ts
|
package/dist/src/mcp/server.js
CHANGED
|
@@ -1,13 +1,13 @@
|
|
|
1
1
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
2
2
|
import {
|
|
3
3
|
ensureNativeDeps
|
|
4
|
-
} from "../../chunk-
|
|
5
|
-
import "../../chunk-
|
|
6
|
-
import "../../chunk-
|
|
4
|
+
} from "../../chunk-6CAKKNGD.js";
|
|
5
|
+
import "../../chunk-EQVQEFOA.js";
|
|
6
|
+
import "../../chunk-TDLQBGKA.js";
|
|
7
7
|
import "../../chunk-PZUWP5VK.js";
|
|
8
8
|
|
|
9
9
|
// src/entries/mcp-server.ts
|
|
10
10
|
ensureNativeDeps();
|
|
11
|
-
var { main } = await import("../../server-
|
|
11
|
+
var { main } = await import("../../server-BQ3DWKZ6.js");
|
|
12
12
|
await main();
|
|
13
13
|
//# sourceMappingURL=server.js.map
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
You are analyzing a coding session buffer for session "{{sessionId}}".
|
|
2
|
-
You have a budget of ~{{maxTokens}} tokens for your response.
|
|
2
|
+
You have a budget of ~{{maxTokens}} tokens for your response.
|
|
3
3
|
|
|
4
4
|
## Events ({{eventCount}} total)
|
|
5
5
|
{{toolSummary}}
|
|
@@ -1,19 +1,9 @@
|
|
|
1
1
|
You are summarizing a coding session for user "{{user}}" (session "{{sessionId}}").
|
|
2
|
-
You have a budget of ~{{maxTokens}} tokens. Use the full budget to produce a rich, detailed narrative.
|
|
3
2
|
|
|
4
3
|
## Session Content
|
|
5
4
|
{{content}}
|
|
6
5
|
|
|
7
6
|
## Task
|
|
8
|
-
Write a
|
|
9
|
-
|
|
10
|
-
Cover:
|
|
11
|
-
- **What was accomplished** — features built, bugs fixed, refactors completed
|
|
12
|
-
- **Key decisions made** — what was chosen and why, including alternatives that were rejected
|
|
13
|
-
- **Problems encountered** — what went wrong, how it was debugged, what the root cause was
|
|
14
|
-
- **Discoveries and learnings** — anything surprising or non-obvious that was learned
|
|
15
|
-
- **Current state** — where things stand at the end of the session, what's next
|
|
16
|
-
|
|
17
|
-
Focus on outcomes and reasoning rather than individual tool calls. Include enough context that someone reading this summary months later would understand what happened and why.
|
|
7
|
+
Write a concise summary of this session in 2-4 sentences. Cover what was accomplished, key decisions made, and where things stand. Skip routine details — focus on outcomes someone would need to know months later.
|
|
18
8
|
|
|
19
9
|
Respond with plain text only, no JSON or markdown fences.
|
|
@@ -7,11 +7,11 @@ import {
|
|
|
7
7
|
} from "./chunk-AK6GNLPV.js";
|
|
8
8
|
import {
|
|
9
9
|
isProcessAlive
|
|
10
|
-
} from "./chunk-
|
|
10
|
+
} from "./chunk-5FIIK27E.js";
|
|
11
11
|
import "./chunk-SAKJMNSR.js";
|
|
12
|
-
import "./chunk-
|
|
13
|
-
import "./chunk-
|
|
14
|
-
import "./chunk-
|
|
12
|
+
import "./chunk-OPO47BVS.js";
|
|
13
|
+
import "./chunk-EQVQEFOA.js";
|
|
14
|
+
import "./chunk-TDLQBGKA.js";
|
|
15
15
|
import "./chunk-PZUWP5VK.js";
|
|
16
16
|
|
|
17
17
|
// src/cli/stats.ts
|
|
@@ -74,4 +74,4 @@ Vectors: error \u2014 ${e.message}`);
|
|
|
74
74
|
export {
|
|
75
75
|
run
|
|
76
76
|
};
|
|
77
|
-
//# sourceMappingURL=stats-
|
|
77
|
+
//# sourceMappingURL=stats-3FAP5FKV.js.map
|
|
@@ -2,13 +2,13 @@ import { createRequire as __cr } from 'node:module'; const require = __cr(import
|
|
|
2
2
|
import {
|
|
3
3
|
createEmbeddingProvider,
|
|
4
4
|
createLlmProvider
|
|
5
|
-
} from "./chunk-
|
|
6
|
-
import "./chunk-
|
|
5
|
+
} from "./chunk-OSZRLHIJ.js";
|
|
6
|
+
import "./chunk-OPO47BVS.js";
|
|
7
7
|
import {
|
|
8
8
|
loadConfig
|
|
9
9
|
} from "./chunk-TBRZAJ7W.js";
|
|
10
10
|
import "./chunk-6UJWI4IW.js";
|
|
11
|
-
import "./chunk-
|
|
11
|
+
import "./chunk-TDLQBGKA.js";
|
|
12
12
|
import "./chunk-PZUWP5VK.js";
|
|
13
13
|
|
|
14
14
|
// src/cli/verify.ts
|
|
@@ -47,4 +47,4 @@ async function run(_args, vaultDir) {
|
|
|
47
47
|
export {
|
|
48
48
|
run
|
|
49
49
|
};
|
|
50
|
-
//# sourceMappingURL=verify-
|
|
50
|
+
//# sourceMappingURL=verify-3FTCOULE.js.map
|
|
@@ -1,11 +1,11 @@
|
|
|
1
1
|
import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
|
|
2
2
|
import {
|
|
3
3
|
getPluginVersion
|
|
4
|
-
} from "./chunk-
|
|
5
|
-
import "./chunk-
|
|
6
|
-
import "./chunk-
|
|
4
|
+
} from "./chunk-TK2ZYIAL.js";
|
|
5
|
+
import "./chunk-EQVQEFOA.js";
|
|
6
|
+
import "./chunk-TDLQBGKA.js";
|
|
7
7
|
import "./chunk-PZUWP5VK.js";
|
|
8
8
|
export {
|
|
9
9
|
getPluginVersion
|
|
10
10
|
};
|
|
11
|
-
//# sourceMappingURL=version-
|
|
11
|
+
//# sourceMappingURL=version-AL67JH7X.js.map
|
package/package.json
CHANGED
package/skills/myco/SKILL.md
CHANGED
|
@@ -207,6 +207,10 @@ The vault should get sharper over time, not just bigger. Every session should le
|
|
|
207
207
|
2. If you find relevant context, factor it into your recommendation
|
|
208
208
|
3. After the decision is made, `myco_remember` the rationale
|
|
209
209
|
|
|
210
|
+
## Reconfiguration
|
|
211
|
+
|
|
212
|
+
To change LLM providers, models, or digest settings on an existing vault, see `references/reconfiguration.md`. It covers the exact CLI commands, flag names, and order of operations (setup-llm → restart → rebuild if needed → verify).
|
|
213
|
+
|
|
210
214
|
## Maintenance
|
|
211
215
|
|
|
212
216
|
For the full CLI reference with all flags, see `references/cli-usage.md`.
|
|
@@ -0,0 +1,92 @@
|
|
|
1
|
+
# Reconfiguration
|
|
2
|
+
|
|
3
|
+
Workflows for changing LLM providers, models, and digest settings on an existing vault. **Use the AskUserQuestion tool** to ask which settings to change — do not guess.
|
|
4
|
+
|
|
5
|
+
## Changing LLM or Embedding Provider/Model
|
|
6
|
+
|
|
7
|
+
Follow this exact order:
|
|
8
|
+
|
|
9
|
+
```bash
|
|
10
|
+
# 1. Detect what's available
|
|
11
|
+
node <plugin-root>/dist/src/cli.js detect-providers
|
|
12
|
+
|
|
13
|
+
# 2. Apply the change (use the correct --llm- or --embedding- prefixed flags)
|
|
14
|
+
node <plugin-root>/dist/src/cli.js setup-llm \
|
|
15
|
+
--llm-provider <provider> --llm-model <model> \
|
|
16
|
+
--embedding-provider <provider> --embedding-model <model>
|
|
17
|
+
|
|
18
|
+
# 3. ALWAYS restart daemon after any config change
|
|
19
|
+
node <plugin-root>/dist/src/cli.js restart
|
|
20
|
+
|
|
21
|
+
# 4. Only rebuild if the EMBEDDING model changed (not needed for LLM-only changes)
|
|
22
|
+
node <plugin-root>/dist/src/cli.js rebuild
|
|
23
|
+
|
|
24
|
+
# 5. Verify connectivity
|
|
25
|
+
node <plugin-root>/dist/src/cli.js verify
|
|
26
|
+
```
|
|
27
|
+
|
|
28
|
+
### Critical Flags
|
|
29
|
+
|
|
30
|
+
The `setup-llm` command uses `--llm-provider`, `--llm-model`, `--embedding-provider`, `--embedding-model` — NOT `--provider` or `--model`. Only pass flags for settings the user explicitly wants to change.
|
|
31
|
+
|
|
32
|
+
### Order Matters
|
|
33
|
+
|
|
34
|
+
1. `setup-llm` writes config
|
|
35
|
+
2. `restart` loads the new config into the daemon
|
|
36
|
+
3. `rebuild` re-embeds with the new embedding model (skip if embedding didn't change)
|
|
37
|
+
4. `verify` confirms everything works
|
|
38
|
+
|
|
39
|
+
### Embedding Model Warning
|
|
40
|
+
|
|
41
|
+
If the embedding model changed, tell the user: "Changing the embedding model requires a full vector index rebuild. This may take a few minutes."
|
|
42
|
+
|
|
43
|
+
## Changing Digest Settings
|
|
44
|
+
|
|
45
|
+
```bash
|
|
46
|
+
node <plugin-root>/dist/src/cli.js setup-digest \
|
|
47
|
+
--context-window <number> --inject-tier <tier>
|
|
48
|
+
node <plugin-root>/dist/src/cli.js restart
|
|
49
|
+
```
|
|
50
|
+
|
|
51
|
+
For all available `setup-digest` flags (tiers, provider override, metabolism tuning, token budgets), see `cli-usage.md`.
|
|
52
|
+
|
|
53
|
+
## Viewing Current Settings
|
|
54
|
+
|
|
55
|
+
```bash
|
|
56
|
+
node <plugin-root>/dist/src/cli.js setup-llm --show
|
|
57
|
+
node <plugin-root>/dist/src/cli.js setup-digest --show
|
|
58
|
+
```
|
|
59
|
+
|
|
60
|
+
## Common Scenarios
|
|
61
|
+
|
|
62
|
+
### "Change my LLM model" (same provider)
|
|
63
|
+
|
|
64
|
+
```bash
|
|
65
|
+
node <plugin-root>/dist/src/cli.js setup-llm --llm-model qwen3.5:35b
|
|
66
|
+
node <plugin-root>/dist/src/cli.js restart
|
|
67
|
+
node <plugin-root>/dist/src/cli.js verify
|
|
68
|
+
```
|
|
69
|
+
|
|
70
|
+
No rebuild needed — embedding didn't change.
|
|
71
|
+
|
|
72
|
+
### "Switch from Ollama to LM Studio"
|
|
73
|
+
|
|
74
|
+
```bash
|
|
75
|
+
node <plugin-root>/dist/src/cli.js detect-providers
|
|
76
|
+
node <plugin-root>/dist/src/cli.js setup-llm \
|
|
77
|
+
--llm-provider lm-studio --llm-model "qwen/qwen3.5-35b-a3b"
|
|
78
|
+
node <plugin-root>/dist/src/cli.js restart
|
|
79
|
+
node <plugin-root>/dist/src/cli.js verify
|
|
80
|
+
```
|
|
81
|
+
|
|
82
|
+
### "Change everything" (provider, model, and embedding)
|
|
83
|
+
|
|
84
|
+
```bash
|
|
85
|
+
node <plugin-root>/dist/src/cli.js detect-providers
|
|
86
|
+
node <plugin-root>/dist/src/cli.js setup-llm \
|
|
87
|
+
--llm-provider ollama --llm-model qwen3.5:35b \
|
|
88
|
+
--embedding-provider ollama --embedding-model bge-m3
|
|
89
|
+
node <plugin-root>/dist/src/cli.js restart
|
|
90
|
+
node <plugin-root>/dist/src/cli.js rebuild
|
|
91
|
+
node <plugin-root>/dist/src/cli.js verify
|
|
92
|
+
```
|
package/skills/setup/SKILL.md
CHANGED
|
@@ -19,7 +19,7 @@ Run:
|
|
|
19
19
|
node ${CLAUDE_PLUGIN_ROOT}/dist/src/cli.js stats
|
|
20
20
|
```
|
|
21
21
|
|
|
22
|
-
- If the command **succeeds** (exit code 0):
|
|
22
|
+
- If the command **succeeds** (exit code 0): the vault already exists. Tell the user "Myco is already configured at `<vault-path>`." Then invoke the `myco` skill using the Skill tool — the `myco` skill handles all reconfiguration, status checks, and ongoing management. **Stop here. Do not continue with the setup flow. Do not attempt reconfiguration yourself.**
|
|
23
23
|
- If the command **fails** (exit code non-zero or vault not found): proceed to Step 2.
|
|
24
24
|
|
|
25
25
|
## Step 2: Detect System
|
|
@@ -39,22 +39,22 @@ For RAM detection, run the appropriate command for the OS:
|
|
|
39
39
|
|
|
40
40
|
Use the RAM value to determine the recommended tier from `references/model-recommendations.md`:
|
|
41
41
|
|
|
42
|
-
| RAM |
|
|
43
|
-
|
|
44
|
-
| 64GB+ | `qwen3.5:35b` | 65536 | 3000 |
|
|
45
|
-
|
|
|
46
|
-
|
|
|
47
|
-
|
|
|
42
|
+
| RAM | Processor Model | Digest Model | Digest Context | Inject Tier |
|
|
43
|
+
|-----|----------------|--------------|----------------|-------------|
|
|
44
|
+
| 64GB+ | `qwen3.5:latest` | `qwen3.5:35b` | 65536 | 3000 |
|
|
45
|
+
| 48GB | `qwen3.5:latest` | `qwen3.5:27b` | 32768 | 3000 |
|
|
46
|
+
| 32GB | `qwen3.5:4b` | `qwen3.5:latest` | 16384 | 1500 |
|
|
47
|
+
| 16GB | `qwen3.5:4b` | `qwen3.5:4b` | 8192 | 1500 |
|
|
48
48
|
|
|
49
|
-
Record: detected RAM (GB), recommended model, digest context window, and default inject tier.
|
|
49
|
+
Record: detected RAM (GB), recommended processor model, recommended digest model, digest context window, and default inject tier.
|
|
50
50
|
|
|
51
51
|
## Step 3: Ask Questions
|
|
52
52
|
|
|
53
|
-
|
|
53
|
+
**Use the AskUserQuestion tool for every question.** Present choices as selectable options. Do not ask questions in plain text — always use AskUserQuestion so the user can select from options. Wait for each answer before asking the next.
|
|
54
54
|
|
|
55
55
|
### Question 1: Vault Location
|
|
56
56
|
|
|
57
|
-
|
|
57
|
+
Use AskUserQuestion to ask the user where to store the vault. Present three choices:
|
|
58
58
|
|
|
59
59
|
- **Project-local** — `.myco/` in the current directory
|
|
60
60
|
- **Centralized** — `~/.myco/vaults/<project-name>/` (where `<project-name>` is the current directory's basename)
|
|
@@ -62,40 +62,54 @@ Ask the user where to store the vault. Present three choices:
|
|
|
62
62
|
|
|
63
63
|
Record the resolved vault path.
|
|
64
64
|
|
|
65
|
-
### Question 2:
|
|
65
|
+
### Question 2: Processor Model (extraction, summaries, titles)
|
|
66
66
|
|
|
67
|
-
|
|
67
|
+
Present the recommended processor model from the RAM table as the default. Show available models from the detected providers, grouped by provider.
|
|
68
68
|
|
|
69
|
-
|
|
69
|
+
Explain: "The processor model handles session extraction, summaries, and titles. Smaller, faster models work well here — speed matters more than depth."
|
|
70
70
|
|
|
71
|
-
|
|
72
|
-
- **LM Studio:** tell the user to open LM Studio, search for `<recommended-model>`, and download it. Offer to wait or to let the user choose a different available model.
|
|
71
|
+
If the recommended model is not installed:
|
|
73
72
|
|
|
74
|
-
|
|
73
|
+
- **Ollama:** offer to run `ollama pull <recommended-model>` before continuing.
|
|
74
|
+
- **LM Studio:** tell the user to download it from the model browser.
|
|
75
75
|
|
|
76
|
-
|
|
76
|
+
Record the chosen provider and processor model.
|
|
77
77
|
|
|
78
|
-
|
|
78
|
+
### Question 3: Digest Model (vault synthesis)
|
|
79
|
+
|
|
80
|
+
Present the recommended digest model from the RAM table as the default. Show available models from the detected providers.
|
|
81
|
+
|
|
82
|
+
Explain: "The digest model synthesizes your vault into context extracts. Larger models produce better results here — quality matters more than speed. This can be the same as the processor model on smaller machines."
|
|
83
|
+
|
|
84
|
+
If the recommended model is not installed, offer to pull/download as above.
|
|
85
|
+
|
|
86
|
+
Record the chosen provider and digest model.
|
|
87
|
+
|
|
88
|
+
### Question 4: Embedding Model
|
|
89
|
+
|
|
90
|
+
List embedding models from available providers. Exclude Anthropic — it does not support embeddings.
|
|
79
91
|
|
|
80
92
|
If no embedding models are installed:
|
|
81
93
|
|
|
82
94
|
- **Ollama:** offer to run `ollama pull bge-m3`. If the user accepts, run it before continuing.
|
|
83
|
-
- **LM Studio:** tell the user to search for and download an embedding model
|
|
95
|
+
- **LM Studio:** tell the user to search for and download an embedding model.
|
|
84
96
|
|
|
85
97
|
Recommend `bge-m3` as the default. Record the chosen embedding provider and model.
|
|
86
98
|
|
|
87
|
-
### Question
|
|
99
|
+
### Question 5: Inject Tier
|
|
88
100
|
|
|
89
|
-
|
|
101
|
+
Ask the user which coding agent they primarily use, then recommend an inject tier based on the agent's context window (see `references/model-recommendations.md`). Show all tiers:
|
|
90
102
|
|
|
91
103
|
| Tier | Description |
|
|
92
104
|
|------|-------------|
|
|
93
105
|
| 1500 | Executive briefing — fastest, lightest |
|
|
94
|
-
| 3000 | Team standup —
|
|
95
|
-
| 5000 | Deep onboarding |
|
|
96
|
-
| 10000 | Institutional knowledge — richest
|
|
106
|
+
| 3000 | Team standup — balanced context |
|
|
107
|
+
| 5000 | Deep onboarding — good for 200K agents |
|
|
108
|
+
| 10000 | Institutional knowledge — richest, best for 1M+ agents |
|
|
109
|
+
|
|
110
|
+
Explain: "The inject tier controls how much vault context is injected at session start. Larger tiers give the agent more project history but use more of its context window. All tiers work regardless of local hardware."
|
|
97
111
|
|
|
98
|
-
|
|
112
|
+
Pre-select the default based on the agent's context window. Tell the user: "Agents can always request a different tier on-demand via the `myco_context` MCP tool."
|
|
99
113
|
|
|
100
114
|
Record the chosen inject tier.
|
|
101
115
|
|
|
@@ -106,14 +120,16 @@ Run the following commands in sequence, substituting the recorded values. Show e
|
|
|
106
120
|
```bash
|
|
107
121
|
node ${CLAUDE_PLUGIN_ROOT}/dist/src/cli.js init \
|
|
108
122
|
--vault <chosen-vault-path> \
|
|
109
|
-
--llm-provider <provider> \
|
|
110
|
-
--llm-model <model> \
|
|
123
|
+
--llm-provider <processor-provider> \
|
|
124
|
+
--llm-model <processor-model> \
|
|
111
125
|
--embedding-provider <embedding-provider> \
|
|
112
126
|
--embedding-model <embedding-model>
|
|
113
127
|
```
|
|
114
128
|
|
|
115
129
|
```bash
|
|
116
130
|
node ${CLAUDE_PLUGIN_ROOT}/dist/src/cli.js setup-digest \
|
|
131
|
+
--provider <digest-provider> \
|
|
132
|
+
--model <digest-model> \
|
|
117
133
|
--context-window <digest-context-window-from-ram-table> \
|
|
118
134
|
--inject-tier <chosen-inject-tier>
|
|
119
135
|
```
|
|
@@ -124,16 +140,28 @@ node ${CLAUDE_PLUGIN_ROOT}/dist/src/cli.js verify
|
|
|
124
140
|
|
|
125
141
|
If any command fails, report the error and stop. Do not continue to the next command on failure. Show the full error output to the user and ask how to proceed.
|
|
126
142
|
|
|
127
|
-
## Step 5:
|
|
143
|
+
## Step 5: Ollama Performance Tips
|
|
144
|
+
|
|
145
|
+
If the user is using Ollama, recommend adding these to their Ollama service configuration:
|
|
146
|
+
|
|
147
|
+
```
|
|
148
|
+
OLLAMA_FLASH_ATTENTION=1 # Required for KV cache quantization
|
|
149
|
+
OLLAMA_KV_CACHE_TYPE=q8_0 # Halves KV cache memory
|
|
150
|
+
```
|
|
151
|
+
|
|
152
|
+
Explain: "These settings halve the memory used for large context windows, making digest much more efficient. They're Ollama-wide settings — on macOS, add them to your Ollama launchd plist."
|
|
153
|
+
|
|
154
|
+
## Step 6: Report
|
|
128
155
|
|
|
129
156
|
Display a summary table:
|
|
130
157
|
|
|
131
158
|
| Setting | Value |
|
|
132
159
|
|---------|-------|
|
|
133
160
|
| Vault path | `<resolved path>` |
|
|
134
|
-
|
|
|
161
|
+
| Processor | `<provider>` / `<processor-model>` |
|
|
162
|
+
| Digest | `<provider>` / `<digest-model>` (context: `<context-window>`) |
|
|
135
163
|
| Embedding | `<embedding-provider>` / `<embedding-model>` |
|
|
136
|
-
|
|
|
164
|
+
| Inject tier | `<inject-tier>` |
|
|
137
165
|
| RAM detected | `<X>` GB |
|
|
138
166
|
|
|
139
167
|
Tell the user: "Myco is ready. Start a new session to begin capturing knowledge."
|
|
@@ -143,4 +171,4 @@ Tell the user: "Myco is ready. Start a new session to begin capturing knowledge.
|
|
|
143
171
|
- All writes via CLI commands — never read or modify `myco.yaml` directly.
|
|
144
172
|
- All provider detection via `detect-providers` — no raw HTTP calls to provider APIs.
|
|
145
173
|
- One question at a time — do not batch questions or present them together.
|
|
146
|
-
-
|
|
174
|
+
- Three model choices in guided setup: processor, digest, and embedding.
|