@iloom/cli 0.9.2 → 0.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/LICENSE +1 -1
- package/README.md +159 -40
- package/dist/{BranchNamingService-K6XNWQ6C.js → BranchNamingService-ECJHBB67.js} +2 -2
- package/dist/ClaudeContextManager-QXX6ZFST.js +14 -0
- package/dist/ClaudeService-NJNK2SUH.js +13 -0
- package/dist/{GitHubService-TGWJN4V4.js → GitHubService-MEHKHUQP.js} +4 -4
- package/dist/IssueTrackerFactory-NG53YX5S.js +14 -0
- package/dist/{LoomLauncher-73NXL2CL.js → LoomLauncher-L64HHS3T.js} +9 -9
- package/dist/{MetadataManager-W3C54UYT.js → MetadataManager-5QZSTKNN.js} +2 -2
- package/dist/{ProjectCapabilityDetector-N5L7T4IY.js → ProjectCapabilityDetector-5KSYUTBJ.js} +3 -3
- package/dist/{PromptTemplateManager-36YLQRHP.js → PromptTemplateManager-DULSVRRE.js} +2 -2
- package/dist/README.md +159 -40
- package/dist/{SettingsManager-AW3JTJHD.js → SettingsManager-BQDQA3FK.js} +4 -2
- package/dist/agents/iloom-artifact-reviewer.md +11 -0
- package/dist/agents/iloom-code-reviewer.md +14 -0
- package/dist/agents/iloom-issue-analyze-and-plan.md +55 -12
- package/dist/agents/iloom-issue-analyzer.md +49 -6
- package/dist/agents/iloom-issue-complexity-evaluator.md +47 -6
- package/dist/agents/iloom-issue-enhancer.md +86 -7
- package/dist/agents/iloom-issue-implementer.md +48 -7
- package/dist/agents/iloom-issue-planner.md +115 -62
- package/dist/{build-THZI572G.js → build-5GO3XW26.js} +9 -9
- package/dist/{chunk-NUACL52E.js → chunk-3D7WQM7I.js} +2 -2
- package/dist/chunk-4232AHNQ.js +35 -0
- package/dist/chunk-4232AHNQ.js.map +1 -0
- package/dist/{chunk-QN47QVBX.js → chunk-4WJNIR5O.js} +1 -1
- package/dist/chunk-4WJNIR5O.js.map +1 -0
- package/dist/{chunk-A7NJF73J.js → chunk-5MWV33NN.js} +4 -4
- package/dist/{chunk-3I4ONZRT.js → chunk-6EU6TCF6.js} +10 -10
- package/dist/chunk-6EU6TCF6.js.map +1 -0
- package/dist/{chunk-CWRI4JC3.js → chunk-FB47TIJG.js} +29 -11
- package/dist/chunk-FB47TIJG.js.map +1 -0
- package/dist/chunk-HEXKPKCK.js +1396 -0
- package/dist/chunk-HEXKPKCK.js.map +1 -0
- package/dist/{chunk-KAYXR544.js → chunk-J5S7DFYC.js} +2 -2
- package/dist/{chunk-ULSWCPQG.js → chunk-JO2LZ6EQ.js} +476 -5
- package/dist/chunk-JO2LZ6EQ.js.map +1 -0
- package/dist/{chunk-KBEIQP4G.js → chunk-KB64WNBZ.js} +43 -3
- package/dist/chunk-KB64WNBZ.js.map +1 -0
- package/dist/{chunk-OFDN5NKS.js → chunk-KXDRI47U.js} +69 -12
- package/dist/chunk-KXDRI47U.js.map +1 -0
- package/dist/{chunk-R4YWBGY6.js → chunk-LXLMMXXY.js} +54 -14
- package/dist/chunk-LXLMMXXY.js.map +1 -0
- package/dist/{chunk-AR5QKYNE.js → chunk-MNHZB4Z2.js} +4 -4
- package/dist/{chunk-TL72BGP6.js → chunk-MORRVYPT.js} +2 -2
- package/dist/{chunk-KJTVU3HZ.js → chunk-NRSWLOAZ.js} +8 -8
- package/dist/chunk-NRSWLOAZ.js.map +1 -0
- package/dist/{chunk-FO5GGFOV.js → chunk-ONQYPICO.js} +13 -5
- package/dist/chunk-ONQYPICO.js.map +1 -0
- package/dist/{chunk-7ZEHSSUP.js → chunk-P4O6EH46.js} +4 -4
- package/dist/chunk-QZWEJVWV.js +207 -0
- package/dist/chunk-QZWEJVWV.js.map +1 -0
- package/dist/chunk-RSYT7MVI.js +202 -0
- package/dist/chunk-RSYT7MVI.js.map +1 -0
- package/dist/{chunk-Z2TWEXR7.js → chunk-RYWFS37M.js} +6 -6
- package/dist/chunk-RYWFS37M.js.map +1 -0
- package/dist/{chunk-B7U6OKUR.js → chunk-SF2P22EE.js} +11 -3
- package/dist/chunk-SF2P22EE.js.map +1 -0
- package/dist/{chunk-6IIL5M2L.js → chunk-SN3SQCFK.js} +10 -8
- package/dist/{chunk-6IIL5M2L.js.map → chunk-SN3SQCFK.js.map} +1 -1
- package/dist/{chunk-SOSQILHO.js → chunk-UD3WJDIV.js} +92 -82
- package/dist/chunk-UD3WJDIV.js.map +1 -0
- package/dist/{chunk-KXGQYLFZ.js → chunk-UKBAJ2QQ.js} +61 -7
- package/dist/chunk-UKBAJ2QQ.js.map +1 -0
- package/dist/{chunk-W6DP5RVR.js → chunk-UVD4CZKS.js} +3 -3
- package/dist/chunk-UWGVCXRF.js +207 -0
- package/dist/chunk-UWGVCXRF.js.map +1 -0
- package/dist/{chunk-NWMORW3U.js → chunk-VECNX6VX.js} +2 -2
- package/dist/{chunk-4CO6KG5S.js → chunk-VG45TUYK.js} +53 -7
- package/dist/{chunk-4CO6KG5S.js.map → chunk-VG45TUYK.js.map} +1 -1
- package/dist/{chunk-TC7APDKU.js → chunk-VGGST52X.js} +2 -2
- package/dist/{chunk-4LKGCFGG.js → chunk-WWKOVDWC.js} +2 -2
- package/dist/{chunk-YKFCCV6S.js → chunk-WY4QBK43.js} +7 -7
- package/dist/chunk-WY4QBK43.js.map +1 -0
- package/dist/chunk-Y4YZTHZE.js +73 -0
- package/dist/chunk-Y4YZTHZE.js.map +1 -0
- package/dist/{chunk-VOGGLPG5.js → chunk-YQ57ORTV.js} +14 -1
- package/dist/chunk-YQ57ORTV.js.map +1 -0
- package/dist/{chunk-RI2YL6TK.js → chunk-YYAKPQBT.js} +65 -18
- package/dist/chunk-YYAKPQBT.js.map +1 -0
- package/dist/{chunk-IZIYLYPK.js → chunk-ZEWU5PZK.js} +2 -2
- package/dist/{chunk-VPTAX5TR.js → chunk-ZHPNZC75.js} +12 -12
- package/dist/chunk-ZHPNZC75.js.map +1 -0
- package/dist/{chunk-DGG2VY7B.js → chunk-ZW2LKWWE.js} +9 -9
- package/dist/chunk-ZW2LKWWE.js.map +1 -0
- package/dist/{claude-TP2QO3BU.js → claude-P3NQR6IJ.js} +2 -2
- package/dist/{cleanup-PJRIFFU4.js → cleanup-6UCPVMFG.js} +81 -32
- package/dist/cleanup-6UCPVMFG.js.map +1 -0
- package/dist/cli.js +638 -349
- package/dist/cli.js.map +1 -1
- package/dist/{commit-IVP3M4HG.js → commit-L3EPY5QG.js} +21 -20
- package/dist/commit-L3EPY5QG.js.map +1 -0
- package/dist/{compile-R2J65HBQ.js → compile-ZS4HYRX5.js} +9 -9
- package/dist/{contribute-VDZXHK5Y.js → contribute-ORDDQGSL.js} +14 -6
- package/dist/contribute-ORDDQGSL.js.map +1 -0
- package/dist/{dev-server-7F622OEO.js → dev-server-FYZ2AQIH.js} +29 -15
- package/dist/dev-server-FYZ2AQIH.js.map +1 -0
- package/dist/{feedback-E7VET7CL.js → feedback-TMBXSCM5.js} +15 -15
- package/dist/{git-2QDQ2X2S.js → git-ET64COO3.js} +4 -4
- package/dist/hooks/iloom-hook.js +15 -0
- package/dist/ignite-CGOV3TD4.js +1393 -0
- package/dist/ignite-CGOV3TD4.js.map +1 -0
- package/dist/index.d.ts +382 -53
- package/dist/index.js +1167 -36
- package/dist/index.js.map +1 -1
- package/dist/{init-676DHF6R.js → init-GFQ5W7GK.js} +57 -21
- package/dist/init-GFQ5W7GK.js.map +1 -0
- package/dist/{issues-PJSOLOBJ.js → issues-T4ZZSPEG.js} +61 -20
- package/dist/issues-T4ZZSPEG.js.map +1 -0
- package/dist/{lint-CJM7BAIM.js → lint-6TQXDZ3T.js} +9 -9
- package/dist/mcp/issue-management-server.js +2471 -256
- package/dist/mcp/issue-management-server.js.map +1 -1
- package/dist/mcp/recap-server.js +144 -21
- package/dist/mcp/recap-server.js.map +1 -1
- package/dist/{neon-helpers-VVFFTLXE.js → neon-helpers-CQN2PB4S.js} +3 -3
- package/dist/neon-helpers-CQN2PB4S.js.map +1 -0
- package/dist/{open-544H7JF5.js → open-5QZGXQRF.js} +15 -15
- package/dist/open-5QZGXQRF.js.map +1 -0
- package/dist/{plan-Q7ELXDLC.js → plan-U7ZQWLFY.js} +41 -25
- package/dist/plan-U7ZQWLFY.js.map +1 -0
- package/dist/{projects-LH362JZQ.js → projects-2UOXFLNZ.js} +4 -4
- package/dist/prompts/CLAUDE.md +62 -0
- package/dist/prompts/init-prompt.txt +347 -26
- package/dist/prompts/issue-prompt.txt +427 -54
- package/dist/prompts/plan-prompt.txt +97 -16
- package/dist/prompts/pr-prompt.txt +44 -1
- package/dist/prompts/regular-prompt.txt +42 -1
- package/dist/prompts/session-summary-prompt.txt +14 -0
- package/dist/prompts/swarm-orchestrator-prompt.txt +437 -0
- package/dist/{rebase-YND35CIE.js → rebase-DWIB77KV.js} +10 -10
- package/dist/{recap-3W7COH7D.js → recap-MX63HAKV.js} +47 -19
- package/dist/recap-MX63HAKV.js.map +1 -0
- package/dist/{run-QUXJKDQQ.js → run-O3TFNQFC.js} +15 -15
- package/dist/run-O3TFNQFC.js.map +1 -0
- package/dist/schema/package-iloom.schema.json +58 -0
- package/dist/schema/settings.schema.json +115 -15
- package/dist/{shell-QGECBLST.js → shell-G6VC2CYR.js} +14 -7
- package/dist/shell-G6VC2CYR.js.map +1 -0
- package/dist/{summary-G2T4452H.js → summary-FWHAX55O.js} +27 -25
- package/dist/summary-FWHAX55O.js.map +1 -0
- package/dist/{test-EA5NQFDC.js → test-F7JNJZYP.js} +9 -9
- package/dist/{test-git-M7LSLEFL.js → test-git-BTAOIUE2.js} +4 -4
- package/dist/test-jira-CHYNV33F.js +96 -0
- package/dist/test-jira-CHYNV33F.js.map +1 -0
- package/dist/{test-prefix-64NAAUON.js → test-prefix-Q6TFSU6F.js} +4 -4
- package/dist/{test-webserver-OK6Z5FJM.js → test-webserver-EONCG7E7.js} +6 -6
- package/dist/{vscode-AR5NNXXI.js → vscode-VA5X4P25.js} +7 -7
- package/package.json +5 -1
- package/dist/ClaudeContextManager-HR5JQKAI.js +0 -14
- package/dist/ClaudeService-TK7FMC2X.js +0 -13
- package/dist/chunk-3I4ONZRT.js.map +0 -1
- package/dist/chunk-B7U6OKUR.js.map +0 -1
- package/dist/chunk-CWRI4JC3.js.map +0 -1
- package/dist/chunk-DGG2VY7B.js.map +0 -1
- package/dist/chunk-FJDRTVJX.js +0 -520
- package/dist/chunk-FJDRTVJX.js.map +0 -1
- package/dist/chunk-FO5GGFOV.js.map +0 -1
- package/dist/chunk-KBEIQP4G.js.map +0 -1
- package/dist/chunk-KJTVU3HZ.js.map +0 -1
- package/dist/chunk-KXGQYLFZ.js.map +0 -1
- package/dist/chunk-OFDN5NKS.js.map +0 -1
- package/dist/chunk-QN47QVBX.js.map +0 -1
- package/dist/chunk-R4YWBGY6.js.map +0 -1
- package/dist/chunk-RI2YL6TK.js.map +0 -1
- package/dist/chunk-SOSQILHO.js.map +0 -1
- package/dist/chunk-ULSWCPQG.js.map +0 -1
- package/dist/chunk-VOGGLPG5.js.map +0 -1
- package/dist/chunk-VPTAX5TR.js.map +0 -1
- package/dist/chunk-WHI5KEOX.js +0 -121
- package/dist/chunk-WHI5KEOX.js.map +0 -1
- package/dist/chunk-YKFCCV6S.js.map +0 -1
- package/dist/chunk-Z2TWEXR7.js.map +0 -1
- package/dist/cleanup-PJRIFFU4.js.map +0 -1
- package/dist/commit-IVP3M4HG.js.map +0 -1
- package/dist/contribute-VDZXHK5Y.js.map +0 -1
- package/dist/dev-server-7F622OEO.js.map +0 -1
- package/dist/ignite-IW35CDBD.js +0 -784
- package/dist/ignite-IW35CDBD.js.map +0 -1
- package/dist/init-676DHF6R.js.map +0 -1
- package/dist/issues-PJSOLOBJ.js.map +0 -1
- package/dist/open-544H7JF5.js.map +0 -1
- package/dist/plan-Q7ELXDLC.js.map +0 -1
- package/dist/recap-3W7COH7D.js.map +0 -1
- package/dist/run-QUXJKDQQ.js.map +0 -1
- package/dist/shell-QGECBLST.js.map +0 -1
- package/dist/summary-G2T4452H.js.map +0 -1
- /package/dist/{BranchNamingService-K6XNWQ6C.js.map → BranchNamingService-ECJHBB67.js.map} +0 -0
- /package/dist/{ClaudeContextManager-HR5JQKAI.js.map → ClaudeContextManager-QXX6ZFST.js.map} +0 -0
- /package/dist/{ClaudeService-TK7FMC2X.js.map → ClaudeService-NJNK2SUH.js.map} +0 -0
- /package/dist/{GitHubService-TGWJN4V4.js.map → GitHubService-MEHKHUQP.js.map} +0 -0
- /package/dist/{MetadataManager-W3C54UYT.js.map → IssueTrackerFactory-NG53YX5S.js.map} +0 -0
- /package/dist/{LoomLauncher-73NXL2CL.js.map → LoomLauncher-L64HHS3T.js.map} +0 -0
- /package/dist/{ProjectCapabilityDetector-N5L7T4IY.js.map → MetadataManager-5QZSTKNN.js.map} +0 -0
- /package/dist/{PromptTemplateManager-36YLQRHP.js.map → ProjectCapabilityDetector-5KSYUTBJ.js.map} +0 -0
- /package/dist/{SettingsManager-AW3JTJHD.js.map → PromptTemplateManager-DULSVRRE.js.map} +0 -0
- /package/dist/{claude-TP2QO3BU.js.map → SettingsManager-BQDQA3FK.js.map} +0 -0
- /package/dist/{build-THZI572G.js.map → build-5GO3XW26.js.map} +0 -0
- /package/dist/{chunk-NUACL52E.js.map → chunk-3D7WQM7I.js.map} +0 -0
- /package/dist/{chunk-A7NJF73J.js.map → chunk-5MWV33NN.js.map} +0 -0
- /package/dist/{chunk-KAYXR544.js.map → chunk-J5S7DFYC.js.map} +0 -0
- /package/dist/{chunk-AR5QKYNE.js.map → chunk-MNHZB4Z2.js.map} +0 -0
- /package/dist/{chunk-TL72BGP6.js.map → chunk-MORRVYPT.js.map} +0 -0
- /package/dist/{chunk-7ZEHSSUP.js.map → chunk-P4O6EH46.js.map} +0 -0
- /package/dist/{chunk-W6DP5RVR.js.map → chunk-UVD4CZKS.js.map} +0 -0
- /package/dist/{chunk-NWMORW3U.js.map → chunk-VECNX6VX.js.map} +0 -0
- /package/dist/{chunk-TC7APDKU.js.map → chunk-VGGST52X.js.map} +0 -0
- /package/dist/{chunk-4LKGCFGG.js.map → chunk-WWKOVDWC.js.map} +0 -0
- /package/dist/{chunk-IZIYLYPK.js.map → chunk-ZEWU5PZK.js.map} +0 -0
- /package/dist/{git-2QDQ2X2S.js.map → claude-P3NQR6IJ.js.map} +0 -0
- /package/dist/{compile-R2J65HBQ.js.map → compile-ZS4HYRX5.js.map} +0 -0
- /package/dist/{feedback-E7VET7CL.js.map → feedback-TMBXSCM5.js.map} +0 -0
- /package/dist/{neon-helpers-VVFFTLXE.js.map → git-ET64COO3.js.map} +0 -0
- /package/dist/{lint-CJM7BAIM.js.map → lint-6TQXDZ3T.js.map} +0 -0
- /package/dist/{projects-LH362JZQ.js.map → projects-2UOXFLNZ.js.map} +0 -0
- /package/dist/{rebase-YND35CIE.js.map → rebase-DWIB77KV.js.map} +0 -0
- /package/dist/{test-EA5NQFDC.js.map → test-F7JNJZYP.js.map} +0 -0
- /package/dist/{test-git-M7LSLEFL.js.map → test-git-BTAOIUE2.js.map} +0 -0
- /package/dist/{test-prefix-64NAAUON.js.map → test-prefix-Q6TFSU6F.js.map} +0 -0
- /package/dist/{test-webserver-OK6Z5FJM.js.map → test-webserver-EONCG7E7.js.map} +0 -0
- /package/dist/{vscode-AR5NNXXI.js.map → vscode-VA5X4P25.js.map} +0 -0
package/dist/mcp/recap-server.js
CHANGED
|
@@ -5,10 +5,12 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
|
|
|
5
5
|
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
|
|
6
6
|
import { z } from "zod";
|
|
7
7
|
import path from "path";
|
|
8
|
+
import os from "os";
|
|
8
9
|
import fs from "fs-extra";
|
|
9
10
|
import { randomUUID } from "crypto";
|
|
10
11
|
var validatedRecapFilePath = null;
|
|
11
12
|
var validatedLoomMetadata = null;
|
|
13
|
+
var validatedMetadataFilePath = null;
|
|
12
14
|
function validateEnvironment() {
|
|
13
15
|
const recapFilePath = process.env.RECAP_FILE_PATH;
|
|
14
16
|
const loomMetadataJson = process.env.LOOM_METADATA_JSON;
|
|
@@ -27,9 +29,11 @@ function validateEnvironment() {
|
|
|
27
29
|
console.error("Failed to parse LOOM_METADATA_JSON:", error);
|
|
28
30
|
process.exit(1);
|
|
29
31
|
}
|
|
32
|
+
const metadataFilePath = process.env.METADATA_FILE_PATH ?? null;
|
|
30
33
|
validatedRecapFilePath = recapFilePath;
|
|
31
34
|
validatedLoomMetadata = loomMetadata;
|
|
32
|
-
|
|
35
|
+
validatedMetadataFilePath = metadataFilePath;
|
|
36
|
+
return { recapFilePath, loomMetadata, metadataFilePath };
|
|
33
37
|
}
|
|
34
38
|
function getRecapFilePath() {
|
|
35
39
|
if (!validatedRecapFilePath) {
|
|
@@ -43,6 +47,32 @@ function getLoomMetadata() {
|
|
|
43
47
|
}
|
|
44
48
|
return validatedLoomMetadata;
|
|
45
49
|
}
|
|
50
|
+
function getMetadataFilePath() {
|
|
51
|
+
if (!validatedMetadataFilePath) {
|
|
52
|
+
throw new Error("METADATA_FILE_PATH not configured - state transition tools require this environment variable");
|
|
53
|
+
}
|
|
54
|
+
return validatedMetadataFilePath;
|
|
55
|
+
}
|
|
56
|
+
function slugifyPath(worktreePath) {
|
|
57
|
+
let slug = worktreePath.replace(/[/\\]+$/, "");
|
|
58
|
+
slug = slug.replace(/[/\\]/g, "___");
|
|
59
|
+
slug = slug.replace(/[^a-zA-Z0-9_-]/g, "-");
|
|
60
|
+
return `${slug}.json`;
|
|
61
|
+
}
|
|
62
|
+
function resolveRecapFilePath(worktreePath) {
|
|
63
|
+
if (worktreePath) {
|
|
64
|
+
const recapsDir = path.join(os.homedir(), ".config", "iloom-ai", "recaps");
|
|
65
|
+
return path.join(recapsDir, slugifyPath(worktreePath));
|
|
66
|
+
}
|
|
67
|
+
return getRecapFilePath();
|
|
68
|
+
}
|
|
69
|
+
function resolveMetadataFilePath(worktreePath) {
|
|
70
|
+
if (worktreePath) {
|
|
71
|
+
const loomsDir = path.join(os.homedir(), ".config", "iloom-ai", "looms");
|
|
72
|
+
return path.join(loomsDir, slugifyPath(worktreePath));
|
|
73
|
+
}
|
|
74
|
+
return getMetadataFilePath();
|
|
75
|
+
}
|
|
46
76
|
async function readRecapFile(filePath) {
|
|
47
77
|
try {
|
|
48
78
|
if (await fs.pathExists(filePath)) {
|
|
@@ -68,14 +98,15 @@ server.registerTool(
|
|
|
68
98
|
title: "Set Goal",
|
|
69
99
|
description: "Set the initial goal (called once at session start)",
|
|
70
100
|
inputSchema: {
|
|
71
|
-
goal: z.string().describe("The original problem statement")
|
|
101
|
+
goal: z.string().describe("The original problem statement"),
|
|
102
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope recap to a specific loom")
|
|
72
103
|
},
|
|
73
104
|
outputSchema: {
|
|
74
105
|
success: z.literal(true)
|
|
75
106
|
}
|
|
76
107
|
},
|
|
77
|
-
async ({ goal }) => {
|
|
78
|
-
const filePath =
|
|
108
|
+
async ({ goal, worktreePath }) => {
|
|
109
|
+
const filePath = resolveRecapFilePath(worktreePath);
|
|
79
110
|
const recap = await readRecapFile(filePath);
|
|
80
111
|
recap.goal = goal;
|
|
81
112
|
await writeRecapFile(filePath, recap);
|
|
@@ -92,15 +123,16 @@ server.registerTool(
|
|
|
92
123
|
description: "Set the assessed complexity of the current task",
|
|
93
124
|
inputSchema: {
|
|
94
125
|
complexity: z.enum(["trivial", "simple", "complex"]).describe("Task complexity level"),
|
|
95
|
-
reason: z.string().optional().describe("Brief explanation for the assessment")
|
|
126
|
+
reason: z.string().optional().describe("Brief explanation for the assessment"),
|
|
127
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope recap to a specific loom")
|
|
96
128
|
},
|
|
97
129
|
outputSchema: {
|
|
98
130
|
success: z.literal(true),
|
|
99
131
|
timestamp: z.string()
|
|
100
132
|
}
|
|
101
133
|
},
|
|
102
|
-
async ({ complexity, reason }) => {
|
|
103
|
-
const filePath =
|
|
134
|
+
async ({ complexity, reason, worktreePath }) => {
|
|
135
|
+
const filePath = resolveRecapFilePath(worktreePath);
|
|
104
136
|
const recap = await readRecapFile(filePath);
|
|
105
137
|
const timestamp = (/* @__PURE__ */ new Date()).toISOString();
|
|
106
138
|
recap.complexity = reason !== void 0 ? { level: complexity, reason, timestamp } : { level: complexity, timestamp };
|
|
@@ -119,7 +151,8 @@ server.registerTool(
|
|
|
119
151
|
description: "Append an entry to the recap. If an entry with the same type and content already exists, it will be skipped.",
|
|
120
152
|
inputSchema: {
|
|
121
153
|
type: z.enum(["decision", "insight", "risk", "assumption", "other"]).describe("Entry type"),
|
|
122
|
-
content: z.string().describe("Entry content")
|
|
154
|
+
content: z.string().describe("Entry content"),
|
|
155
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope recap to a specific loom")
|
|
123
156
|
},
|
|
124
157
|
outputSchema: {
|
|
125
158
|
id: z.string(),
|
|
@@ -127,8 +160,8 @@ server.registerTool(
|
|
|
127
160
|
skipped: z.boolean()
|
|
128
161
|
}
|
|
129
162
|
},
|
|
130
|
-
async ({ type, content }) => {
|
|
131
|
-
const filePath =
|
|
163
|
+
async ({ type, content, worktreePath }) => {
|
|
164
|
+
const filePath = resolveRecapFilePath(worktreePath);
|
|
132
165
|
const recap = await readRecapFile(filePath);
|
|
133
166
|
recap.entries ??= [];
|
|
134
167
|
const existingEntry = recap.entries.find((e) => e.type === type && e.content === content);
|
|
@@ -164,7 +197,8 @@ server.registerTool(
|
|
|
164
197
|
primaryUrl: z.string().url().describe("Main URL for the artifact"),
|
|
165
198
|
description: z.string().describe("Brief description of the artifact"),
|
|
166
199
|
id: z.string().optional().describe("Optional artifact ID (e.g., comment ID, issue number)"),
|
|
167
|
-
urls: z.record(z.string()).optional().describe('Optional additional URLs (e.g., { api: "..." })')
|
|
200
|
+
urls: z.record(z.string()).optional().describe('Optional additional URLs (e.g., { api: "..." })'),
|
|
201
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope recap to a specific loom")
|
|
168
202
|
},
|
|
169
203
|
outputSchema: {
|
|
170
204
|
id: z.string(),
|
|
@@ -172,8 +206,8 @@ server.registerTool(
|
|
|
172
206
|
replaced: z.boolean()
|
|
173
207
|
}
|
|
174
208
|
},
|
|
175
|
-
async ({ type, primaryUrl, description, id, urls }) => {
|
|
176
|
-
const filePath =
|
|
209
|
+
async ({ type, primaryUrl, description, id, urls, worktreePath }) => {
|
|
210
|
+
const filePath = resolveRecapFilePath(worktreePath);
|
|
177
211
|
const recap = await readRecapFile(filePath);
|
|
178
212
|
const artifact = {
|
|
179
213
|
id: id ?? randomUUID(),
|
|
@@ -208,7 +242,9 @@ server.registerTool(
|
|
|
208
242
|
{
|
|
209
243
|
title: "Get Recap",
|
|
210
244
|
description: "Read current recap (for catching up or review)",
|
|
211
|
-
inputSchema: {
|
|
245
|
+
inputSchema: {
|
|
246
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope recap to a specific loom")
|
|
247
|
+
},
|
|
212
248
|
outputSchema: {
|
|
213
249
|
filePath: z.string(),
|
|
214
250
|
goal: z.string().nullable(),
|
|
@@ -237,10 +273,23 @@ server.registerTool(
|
|
|
237
273
|
)
|
|
238
274
|
}
|
|
239
275
|
},
|
|
240
|
-
async () => {
|
|
241
|
-
const filePath =
|
|
276
|
+
async ({ worktreePath }) => {
|
|
277
|
+
const filePath = resolveRecapFilePath(worktreePath);
|
|
242
278
|
const recap = await readRecapFile(filePath);
|
|
243
|
-
|
|
279
|
+
let defaultGoal = null;
|
|
280
|
+
if (worktreePath) {
|
|
281
|
+
try {
|
|
282
|
+
const metaPath = resolveMetadataFilePath(worktreePath);
|
|
283
|
+
if (await fs.pathExists(metaPath)) {
|
|
284
|
+
const metaContent = await fs.readFile(metaPath, "utf8");
|
|
285
|
+
const meta = JSON.parse(metaContent);
|
|
286
|
+
defaultGoal = meta.description || null;
|
|
287
|
+
}
|
|
288
|
+
} catch {
|
|
289
|
+
}
|
|
290
|
+
} else {
|
|
291
|
+
defaultGoal = getLoomMetadata().description || null;
|
|
292
|
+
}
|
|
244
293
|
const result = {
|
|
245
294
|
filePath,
|
|
246
295
|
goal: recap.goal ?? defaultGoal,
|
|
@@ -254,20 +303,94 @@ server.registerTool(
|
|
|
254
303
|
};
|
|
255
304
|
}
|
|
256
305
|
);
|
|
306
|
+
var swarmStateSchema = z.enum(["pending", "in_progress", "code_review", "done", "failed"]);
|
|
307
|
+
server.registerTool(
|
|
308
|
+
"set_loom_state",
|
|
309
|
+
{
|
|
310
|
+
title: "Set Loom State",
|
|
311
|
+
description: "Set the swarm lifecycle state of the current loom",
|
|
312
|
+
inputSchema: {
|
|
313
|
+
state: swarmStateSchema.describe("The new state for the loom"),
|
|
314
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope state operations to a specific loom")
|
|
315
|
+
},
|
|
316
|
+
outputSchema: {
|
|
317
|
+
success: z.literal(true),
|
|
318
|
+
state: swarmStateSchema
|
|
319
|
+
}
|
|
320
|
+
},
|
|
321
|
+
async ({ state, worktreePath }) => {
|
|
322
|
+
const metadataFilePath = resolveMetadataFilePath(worktreePath);
|
|
323
|
+
let metadata;
|
|
324
|
+
try {
|
|
325
|
+
const content = await fs.readFile(metadataFilePath, "utf8");
|
|
326
|
+
metadata = JSON.parse(content);
|
|
327
|
+
} catch (error) {
|
|
328
|
+
throw new Error(`Failed to read metadata file at ${metadataFilePath}: ${error instanceof Error ? error.message : String(error)}`);
|
|
329
|
+
}
|
|
330
|
+
metadata.state = state;
|
|
331
|
+
await fs.writeFile(metadataFilePath, JSON.stringify(metadata, null, 2), { mode: 420 });
|
|
332
|
+
const result = { success: true, state };
|
|
333
|
+
return {
|
|
334
|
+
content: [{ type: "text", text: JSON.stringify(result) }],
|
|
335
|
+
structuredContent: result
|
|
336
|
+
};
|
|
337
|
+
}
|
|
338
|
+
);
|
|
339
|
+
server.registerTool(
|
|
340
|
+
"get_loom_state",
|
|
341
|
+
{
|
|
342
|
+
title: "Get Loom State",
|
|
343
|
+
description: "Get the current swarm lifecycle state of the loom",
|
|
344
|
+
inputSchema: {
|
|
345
|
+
worktreePath: z.string().optional().describe("Optional worktree path to scope state operations to a specific loom")
|
|
346
|
+
},
|
|
347
|
+
outputSchema: {
|
|
348
|
+
state: swarmStateSchema.nullable()
|
|
349
|
+
}
|
|
350
|
+
},
|
|
351
|
+
async ({ worktreePath }) => {
|
|
352
|
+
const metadataFilePath = resolveMetadataFilePath(worktreePath);
|
|
353
|
+
let metadata;
|
|
354
|
+
try {
|
|
355
|
+
const content = await fs.readFile(metadataFilePath, "utf8");
|
|
356
|
+
metadata = JSON.parse(content);
|
|
357
|
+
} catch (error) {
|
|
358
|
+
throw new Error(`Failed to read metadata file at ${metadataFilePath}: ${error instanceof Error ? error.message : String(error)}`);
|
|
359
|
+
}
|
|
360
|
+
const result = { state: metadata.state ?? null };
|
|
361
|
+
return {
|
|
362
|
+
content: [{ type: "text", text: JSON.stringify(result) }],
|
|
363
|
+
structuredContent: result
|
|
364
|
+
};
|
|
365
|
+
}
|
|
366
|
+
);
|
|
257
367
|
async function main() {
|
|
258
|
-
console.error("
|
|
259
|
-
|
|
368
|
+
console.error("=== Loom Recap MCP Server Starting ===");
|
|
369
|
+
console.error(`PID: ${process.pid}`);
|
|
370
|
+
console.error(`Node version: ${process.version}`);
|
|
371
|
+
console.error(`CWD: ${process.cwd()}`);
|
|
372
|
+
console.error(`Script: ${new URL(import.meta.url).pathname}`);
|
|
373
|
+
console.error("Environment variables:");
|
|
374
|
+
console.error(` RECAP_FILE_PATH=${process.env.RECAP_FILE_PATH ?? "<not set>"}`);
|
|
375
|
+
console.error(` METADATA_FILE_PATH=${process.env.METADATA_FILE_PATH ?? "<not set>"}`);
|
|
376
|
+
console.error(` LOOM_METADATA_JSON=${process.env.LOOM_METADATA_JSON ? `<set, ${process.env.LOOM_METADATA_JSON.length} chars>` : "<not set>"}`);
|
|
377
|
+
const { recapFilePath, loomMetadata, metadataFilePath } = validateEnvironment();
|
|
260
378
|
console.error(`Recap file path: ${recapFilePath}`);
|
|
379
|
+
console.error(`Metadata file path: ${metadataFilePath ?? "<not configured>"}`);
|
|
261
380
|
console.error(`Loom: ${loomMetadata.description} (branch: ${loomMetadata.branchName})`);
|
|
381
|
+
const recapExists = await fs.pathExists(recapFilePath);
|
|
382
|
+
console.error(`Recap file exists: ${recapExists}`);
|
|
262
383
|
const transport = new StdioServerTransport();
|
|
263
384
|
await server.connect(transport);
|
|
264
|
-
console.error("Loom Recap MCP Server
|
|
385
|
+
console.error("=== Loom Recap MCP Server READY (stdio transport) ===");
|
|
265
386
|
}
|
|
266
387
|
main().catch((error) => {
|
|
267
388
|
console.error("Fatal error starting MCP server:", error);
|
|
268
389
|
process.exit(1);
|
|
269
390
|
});
|
|
270
391
|
export {
|
|
271
|
-
getLoomMetadata
|
|
392
|
+
getLoomMetadata,
|
|
393
|
+
resolveMetadataFilePath,
|
|
394
|
+
resolveRecapFilePath
|
|
272
395
|
};
|
|
273
396
|
//# sourceMappingURL=recap-server.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/mcp/recap-server.ts"],"sourcesContent":["/**\n * Loom Recap MCP Server\n *\n * Captures session context (goal, decisions, insights, risks, assumptions)\n * for the VS Code Loom Context Panel.\n *\n * Environment variables:\n * - RECAP_FILE_PATH: Complete path to the recap.json file (read/write)\n * - LOOM_METADATA_JSON: Stringified JSON of the loom metadata (parsed using LoomMetadata type)\n */\nimport { McpServer } from '@modelcontextprotocol/sdk/server/mcp.js'\nimport { StdioServerTransport } from '@modelcontextprotocol/sdk/server/stdio.js'\nimport { z } from 'zod'\nimport path from 'path'\nimport fs from 'fs-extra'\nimport { randomUUID } from 'crypto'\nimport type { RecapFile, RecapEntry, RecapOutput, RecapArtifact } from './recap-types.js'\nimport type { LoomMetadata } from '../lib/MetadataManager.js'\n\ninterface EnvConfig {\n\trecapFilePath: string\n\tloomMetadata: LoomMetadata\n}\n\n// Store validated config for use in tool handlers\nlet validatedRecapFilePath: string | null = null\nlet validatedLoomMetadata: LoomMetadata | null = null\n\n/**\n * Validate required environment variables\n * Exits with error if missing (matches issue-management-server.ts pattern)\n */\nfunction validateEnvironment(): EnvConfig {\n\tconst recapFilePath = process.env.RECAP_FILE_PATH\n\tconst loomMetadataJson = process.env.LOOM_METADATA_JSON\n\n\tif (!recapFilePath) {\n\t\tconsole.error('Missing required environment variable: RECAP_FILE_PATH')\n\t\tprocess.exit(1)\n\t}\n\tif (!loomMetadataJson) {\n\t\tconsole.error('Missing required environment variable: LOOM_METADATA_JSON')\n\t\tprocess.exit(1)\n\t}\n\n\tlet loomMetadata: LoomMetadata\n\ttry {\n\t\tloomMetadata = JSON.parse(loomMetadataJson) as LoomMetadata\n\t} catch (error) {\n\t\tconsole.error('Failed to parse LOOM_METADATA_JSON:', error)\n\t\tprocess.exit(1)\n\t}\n\n\t// Store for tool handlers\n\tvalidatedRecapFilePath = recapFilePath\n\tvalidatedLoomMetadata = loomMetadata\n\n\treturn { recapFilePath, loomMetadata }\n}\n\n/**\n * Get the validated recap file path\n * Throws if called before validateEnvironment()\n */\nfunction getRecapFilePath(): string {\n\tif (!validatedRecapFilePath) {\n\t\tthrow new Error('RECAP_FILE_PATH not validated - validateEnvironment() must be called first')\n\t}\n\treturn validatedRecapFilePath\n}\n\n/**\n * Get the validated loom metadata\n * Throws if called before validateEnvironment()\n */\nexport function getLoomMetadata(): LoomMetadata {\n\tif (!validatedLoomMetadata) {\n\t\tthrow new Error('LOOM_METADATA_JSON not validated - validateEnvironment() must be called first')\n\t}\n\treturn validatedLoomMetadata\n}\n\n/**\n * Read recap file (returns empty object if not found or invalid)\n */\nasync function readRecapFile(filePath: string): Promise<RecapFile> {\n\ttry {\n\t\tif (await fs.pathExists(filePath)) {\n\t\t\tconst content = await fs.readFile(filePath, 'utf8')\n\t\t\treturn JSON.parse(content) as RecapFile\n\t\t}\n\t} catch (error) {\n\t\tconsole.error(`Warning: Could not read recap file: ${error}`)\n\t}\n\treturn {}\n}\n\n/**\n * Write recap file (ensures parent directory exists)\n */\nasync function writeRecapFile(filePath: string, recap: RecapFile): Promise<void> {\n\tawait fs.ensureDir(path.dirname(filePath), { mode: 0o755 })\n\tawait fs.writeFile(filePath, JSON.stringify(recap, null, 2), { mode: 0o644 })\n}\n\n// Initialize MCP server\nconst server = new McpServer({\n\tname: 'loom-recap',\n\tversion: '0.1.0',\n})\n\n// Register set_goal tool\nserver.registerTool(\n\t'set_goal',\n\t{\n\t\ttitle: 'Set Goal',\n\t\tdescription: 'Set the initial goal (called once at session start)',\n\t\tinputSchema: {\n\t\t\tgoal: z.string().describe('The original problem statement'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tsuccess: z.literal(true),\n\t\t},\n\t},\n\tasync ({ goal }) => {\n\t\tconst filePath = getRecapFilePath()\n\t\tconst recap = await readRecapFile(filePath)\n\t\trecap.goal = goal\n\t\tawait writeRecapFile(filePath, recap)\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify({ success: true }) }],\n\t\t\tstructuredContent: { success: true },\n\t\t}\n\t}\n)\n\n// Register set_complexity tool\nserver.registerTool(\n\t'set_complexity',\n\t{\n\t\ttitle: 'Set Complexity',\n\t\tdescription: 'Set the assessed complexity of the current task',\n\t\tinputSchema: {\n\t\t\tcomplexity: z.enum(['trivial', 'simple', 'complex']).describe('Task complexity level'),\n\t\t\treason: z.string().optional().describe('Brief explanation for the assessment'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tsuccess: z.literal(true),\n\t\t\ttimestamp: z.string(),\n\t\t},\n\t},\n\tasync ({ complexity, reason }) => {\n\t\tconst filePath = getRecapFilePath()\n\t\tconst recap = await readRecapFile(filePath)\n\t\tconst timestamp = new Date().toISOString()\n\t\trecap.complexity = reason !== undefined ? { level: complexity, reason, timestamp } : { level: complexity, timestamp }\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { success: true as const, timestamp }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register add_entry tool\nserver.registerTool(\n\t'add_entry',\n\t{\n\t\ttitle: 'Add Entry',\n\t\tdescription:\n\t\t\t'Append an entry to the recap. If an entry with the same type and content already exists, it will be skipped.',\n\t\tinputSchema: {\n\t\t\ttype: z\n\t\t\t\t.enum(['decision', 'insight', 'risk', 'assumption', 'other'])\n\t\t\t\t.describe('Entry type'),\n\t\t\tcontent: z.string().describe('Entry content'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tid: z.string(),\n\t\t\ttimestamp: z.string(),\n\t\t\tskipped: z.boolean(),\n\t\t},\n\t},\n\tasync ({ type, content }) => {\n\t\tconst filePath = getRecapFilePath()\n\t\tconst recap = await readRecapFile(filePath)\n\t\trecap.entries ??= []\n\n\t\t// Deduplication: skip if entry with same type and content exists\n\t\tconst existingEntry = recap.entries.find((e) => e.type === type && e.content === content)\n\n\t\tif (existingEntry) {\n\t\t\tconst result = { id: existingEntry.id, timestamp: existingEntry.timestamp, skipped: true }\n\t\t\treturn {\n\t\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\t\tstructuredContent: result,\n\t\t\t}\n\t\t}\n\n\t\tconst entry: RecapEntry = {\n\t\t\tid: randomUUID(),\n\t\t\ttimestamp: new Date().toISOString(),\n\t\t\ttype,\n\t\t\tcontent,\n\t\t}\n\t\trecap.entries.push(entry)\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { id: entry.id, timestamp: entry.timestamp, skipped: false }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register add_artifact tool\nserver.registerTool(\n\t'add_artifact',\n\t{\n\t\ttitle: 'Add Artifact',\n\t\tdescription:\n\t\t\t'Track an artifact (comment, issue, PR) created during the session. If an artifact with the same primaryUrl already exists, it will be replaced.',\n\t\tinputSchema: {\n\t\t\ttype: z.enum(['comment', 'issue', 'pr']).describe('Artifact type'),\n\t\t\tprimaryUrl: z.string().url().describe('Main URL for the artifact'),\n\t\t\tdescription: z.string().describe('Brief description of the artifact'),\n\t\t\tid: z.string().optional().describe('Optional artifact ID (e.g., comment ID, issue number)'),\n\t\t\turls: z.record(z.string()).optional().describe('Optional additional URLs (e.g., { api: \"...\" })'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tid: z.string(),\n\t\t\ttimestamp: z.string(),\n\t\t\treplaced: z.boolean(),\n\t\t},\n\t},\n\tasync ({ type, primaryUrl, description, id, urls }) => {\n\t\tconst filePath = getRecapFilePath()\n\t\tconst recap = await readRecapFile(filePath)\n\n\t\tconst artifact: RecapArtifact = {\n\t\t\tid: id ?? randomUUID(),\n\t\t\ttype,\n\t\t\tprimaryUrl,\n\t\t\turls: urls ?? {},\n\t\t\tdescription,\n\t\t\ttimestamp: new Date().toISOString(),\n\t\t}\n\n\t\trecap.artifacts ??= []\n\n\t\t// Deduplication: replace existing artifact with same primaryUrl\n\t\tconst existingIndex = recap.artifacts.findIndex((a) => a.primaryUrl === primaryUrl)\n\t\tconst replaced = existingIndex !== -1\n\n\t\tif (replaced) {\n\t\t\t// Preserve the original id if not explicitly provided\n\t\t\tconst existingArtifact = recap.artifacts[existingIndex]\n\t\t\tif (existingArtifact) {\n\t\t\t\tartifact.id = id ?? existingArtifact.id\n\t\t\t\trecap.artifacts[existingIndex] = artifact\n\t\t\t}\n\t\t} else {\n\t\t\trecap.artifacts.push(artifact)\n\t\t}\n\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { id: artifact.id, timestamp: artifact.timestamp, replaced }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register get_recap tool\nserver.registerTool(\n\t'get_recap',\n\t{\n\t\ttitle: 'Get Recap',\n\t\tdescription: 'Read current recap (for catching up or review)',\n\t\tinputSchema: {},\n\t\toutputSchema: {\n\t\t\tfilePath: z.string(),\n\t\t\tgoal: z.string().nullable(),\n\t\t\tcomplexity: z\n\t\t\t\t.object({\n\t\t\t\t\tlevel: z.enum(['trivial', 'simple', 'complex']),\n\t\t\t\t\treason: z.string().optional(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t})\n\t\t\t\t.nullable(),\n\t\t\tentries: z.array(\n\t\t\t\tz.object({\n\t\t\t\t\tid: z.string(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t\ttype: z.enum(['decision', 'insight', 'risk', 'assumption', 'other']),\n\t\t\t\t\tcontent: z.string(),\n\t\t\t\t})\n\t\t\t),\n\t\t\tartifacts: z.array(\n\t\t\t\tz.object({\n\t\t\t\t\tid: z.string(),\n\t\t\t\t\ttype: z.enum(['comment', 'issue', 'pr']),\n\t\t\t\t\tprimaryUrl: z.string(),\n\t\t\t\t\turls: z.record(z.string()),\n\t\t\t\t\tdescription: z.string(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t})\n\t\t\t),\n\t\t},\n\t},\n\tasync () => {\n\t\tconst filePath = getRecapFilePath()\n\t\tconst recap = await readRecapFile(filePath)\n\t\t// Use loom description as default goal for new/missing recap files\n\t\tconst defaultGoal = getLoomMetadata().description || null\n\t\tconst result: RecapOutput = {\n\t\t\tfilePath,\n\t\t\tgoal: recap.goal ?? defaultGoal,\n\t\t\tcomplexity: recap.complexity ?? null,\n\t\t\tentries: recap.entries ?? [],\n\t\t\tartifacts: recap.artifacts ?? [],\n\t\t}\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result as unknown as Record<string, unknown>,\n\t\t}\n\t}\n)\n\n// Main server startup\nasync function main(): Promise<void> {\n\tconsole.error('Starting Loom Recap MCP Server...')\n\tconst { recapFilePath, loomMetadata } = validateEnvironment()\n\tconsole.error(`Recap file path: ${recapFilePath}`)\n\tconsole.error(`Loom: ${loomMetadata.description} (branch: ${loomMetadata.branchName})`)\n\tconst transport = new StdioServerTransport()\n\tawait server.connect(transport)\n\tconsole.error('Loom Recap MCP Server running on stdio transport')\n}\n\nmain().catch((error) => {\n\tconsole.error('Fatal error starting MCP server:', error)\n\tprocess.exit(1)\n})\n"],"mappings":";;;AAUA,SAAS,iBAAiB;AAC1B,SAAS,4BAA4B;AACrC,SAAS,SAAS;AAClB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,SAAS,kBAAkB;AAU3B,IAAI,yBAAwC;AAC5C,IAAI,wBAA6C;AAMjD,SAAS,sBAAiC;AACzC,QAAM,gBAAgB,QAAQ,IAAI;AAClC,QAAM,mBAAmB,QAAQ,IAAI;AAErC,MAAI,CAAC,eAAe;AACnB,YAAQ,MAAM,wDAAwD;AACtE,YAAQ,KAAK,CAAC;AAAA,EACf;AACA,MAAI,CAAC,kBAAkB;AACtB,YAAQ,MAAM,2DAA2D;AACzE,YAAQ,KAAK,CAAC;AAAA,EACf;AAEA,MAAI;AACJ,MAAI;AACH,mBAAe,KAAK,MAAM,gBAAgB;AAAA,EAC3C,SAAS,OAAO;AACf,YAAQ,MAAM,uCAAuC,KAAK;AAC1D,YAAQ,KAAK,CAAC;AAAA,EACf;AAGA,2BAAyB;AACzB,0BAAwB;AAExB,SAAO,EAAE,eAAe,aAAa;AACtC;AAMA,SAAS,mBAA2B;AACnC,MAAI,CAAC,wBAAwB;AAC5B,UAAM,IAAI,MAAM,4EAA4E;AAAA,EAC7F;AACA,SAAO;AACR;AAMO,SAAS,kBAAgC;AAC/C,MAAI,CAAC,uBAAuB;AAC3B,UAAM,IAAI,MAAM,+EAA+E;AAAA,EAChG;AACA,SAAO;AACR;AAKA,eAAe,cAAc,UAAsC;AAClE,MAAI;AACH,QAAI,MAAM,GAAG,WAAW,QAAQ,GAAG;AAClC,YAAM,UAAU,MAAM,GAAG,SAAS,UAAU,MAAM;AAClD,aAAO,KAAK,MAAM,OAAO;AAAA,IAC1B;AAAA,EACD,SAAS,OAAO;AACf,YAAQ,MAAM,uCAAuC,KAAK,EAAE;AAAA,EAC7D;AACA,SAAO,CAAC;AACT;AAKA,eAAe,eAAe,UAAkB,OAAiC;AAChF,QAAM,GAAG,UAAU,KAAK,QAAQ,QAAQ,GAAG,EAAE,MAAM,IAAM,CAAC;AAC1D,QAAM,GAAG,UAAU,UAAU,KAAK,UAAU,OAAO,MAAM,CAAC,GAAG,EAAE,MAAM,IAAM,CAAC;AAC7E;AAGA,IAAM,SAAS,IAAI,UAAU;AAAA,EAC5B,MAAM;AAAA,EACN,SAAS;AACV,CAAC;AAGD,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,MAAM,EAAE,OAAO,EAAE,SAAS,gCAAgC;AAAA,IAC3D;AAAA,IACA,cAAc;AAAA,MACb,SAAS,EAAE,QAAQ,IAAI;AAAA,IACxB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,KAAK,MAAM;AACnB,UAAM,WAAW,iBAAiB;AAClC,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,OAAO;AACb,UAAM,eAAe,UAAU,KAAK;AACpC,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,EAAE,SAAS,KAAK,CAAC,EAAE,CAAC;AAAA,MAC5E,mBAAmB,EAAE,SAAS,KAAK;AAAA,IACpC;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,YAAY,EAAE,KAAK,CAAC,WAAW,UAAU,SAAS,CAAC,EAAE,SAAS,uBAAuB;AAAA,MACrF,QAAQ,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,sCAAsC;AAAA,IAC9E;AAAA,IACA,cAAc;AAAA,MACb,SAAS,EAAE,QAAQ,IAAI;AAAA,MACvB,WAAW,EAAE,OAAO;AAAA,IACrB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,YAAY,OAAO,MAAM;AACjC,UAAM,WAAW,iBAAiB;AAClC,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,aAAY,oBAAI,KAAK,GAAE,YAAY;AACzC,UAAM,aAAa,WAAW,SAAY,EAAE,OAAO,YAAY,QAAQ,UAAU,IAAI,EAAE,OAAO,YAAY,UAAU;AACpH,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,SAAS,MAAe,UAAU;AACnD,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aACC;AAAA,IACD,aAAa;AAAA,MACZ,MAAM,EACJ,KAAK,CAAC,YAAY,WAAW,QAAQ,cAAc,OAAO,CAAC,EAC3D,SAAS,YAAY;AAAA,MACvB,SAAS,EAAE,OAAO,EAAE,SAAS,eAAe;AAAA,IAC7C;AAAA,IACA,cAAc;AAAA,MACb,IAAI,EAAE,OAAO;AAAA,MACb,WAAW,EAAE,OAAO;AAAA,MACpB,SAAS,EAAE,QAAQ;AAAA,IACpB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,MAAM,QAAQ,MAAM;AAC5B,UAAM,WAAW,iBAAiB;AAClC,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,YAAY,CAAC;AAGnB,UAAM,gBAAgB,MAAM,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,QAAQ,EAAE,YAAY,OAAO;AAExF,QAAI,eAAe;AAClB,YAAMA,UAAS,EAAE,IAAI,cAAc,IAAI,WAAW,cAAc,WAAW,SAAS,KAAK;AACzF,aAAO;AAAA,QACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAUA,OAAM,EAAE,CAAC;AAAA,QACjE,mBAAmBA;AAAA,MACpB;AAAA,IACD;AAEA,UAAM,QAAoB;AAAA,MACzB,IAAI,WAAW;AAAA,MACf,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,MAClC;AAAA,MACA;AAAA,IACD;AACA,UAAM,QAAQ,KAAK,KAAK;AACxB,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,IAAI,MAAM,IAAI,WAAW,MAAM,WAAW,SAAS,MAAM;AAC1E,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aACC;AAAA,IACD,aAAa;AAAA,MACZ,MAAM,EAAE,KAAK,CAAC,WAAW,SAAS,IAAI,CAAC,EAAE,SAAS,eAAe;AAAA,MACjE,YAAY,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,2BAA2B;AAAA,MACjE,aAAa,EAAE,OAAO,EAAE,SAAS,mCAAmC;AAAA,MACpE,IAAI,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,uDAAuD;AAAA,MAC1F,MAAM,EAAE,OAAO,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS,iDAAiD;AAAA,IACjG;AAAA,IACA,cAAc;AAAA,MACb,IAAI,EAAE,OAAO;AAAA,MACb,WAAW,EAAE,OAAO;AAAA,MACpB,UAAU,EAAE,QAAQ;AAAA,IACrB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,MAAM,YAAY,aAAa,IAAI,KAAK,MAAM;AACtD,UAAM,WAAW,iBAAiB;AAClC,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAE1C,UAAM,WAA0B;AAAA,MAC/B,IAAI,MAAM,WAAW;AAAA,MACrB;AAAA,MACA;AAAA,MACA,MAAM,QAAQ,CAAC;AAAA,MACf;AAAA,MACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IACnC;AAEA,UAAM,cAAc,CAAC;AAGrB,UAAM,gBAAgB,MAAM,UAAU,UAAU,CAAC,MAAM,EAAE,eAAe,UAAU;AAClF,UAAM,WAAW,kBAAkB;AAEnC,QAAI,UAAU;AAEb,YAAM,mBAAmB,MAAM,UAAU,aAAa;AACtD,UAAI,kBAAkB;AACrB,iBAAS,KAAK,MAAM,iBAAiB;AACrC,cAAM,UAAU,aAAa,IAAI;AAAA,MAClC;AAAA,IACD,OAAO;AACN,YAAM,UAAU,KAAK,QAAQ;AAAA,IAC9B;AAEA,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,IAAI,SAAS,IAAI,WAAW,SAAS,WAAW,SAAS;AAC1E,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa,CAAC;AAAA,IACd,cAAc;AAAA,MACb,UAAU,EAAE,OAAO;AAAA,MACnB,MAAM,EAAE,OAAO,EAAE,SAAS;AAAA,MAC1B,YAAY,EACV,OAAO;AAAA,QACP,OAAO,EAAE,KAAK,CAAC,WAAW,UAAU,SAAS,CAAC;AAAA,QAC9C,QAAQ,EAAE,OAAO,EAAE,SAAS;AAAA,QAC5B,WAAW,EAAE,OAAO;AAAA,MACrB,CAAC,EACA,SAAS;AAAA,MACX,SAAS,EAAE;AAAA,QACV,EAAE,OAAO;AAAA,UACR,IAAI,EAAE,OAAO;AAAA,UACb,WAAW,EAAE,OAAO;AAAA,UACpB,MAAM,EAAE,KAAK,CAAC,YAAY,WAAW,QAAQ,cAAc,OAAO,CAAC;AAAA,UACnE,SAAS,EAAE,OAAO;AAAA,QACnB,CAAC;AAAA,MACF;AAAA,MACA,WAAW,EAAE;AAAA,QACZ,EAAE,OAAO;AAAA,UACR,IAAI,EAAE,OAAO;AAAA,UACb,MAAM,EAAE,KAAK,CAAC,WAAW,SAAS,IAAI,CAAC;AAAA,UACvC,YAAY,EAAE,OAAO;AAAA,UACrB,MAAM,EAAE,OAAO,EAAE,OAAO,CAAC;AAAA,UACzB,aAAa,EAAE,OAAO;AAAA,UACtB,WAAW,EAAE,OAAO;AAAA,QACrB,CAAC;AAAA,MACF;AAAA,IACD;AAAA,EACD;AAAA,EACA,YAAY;AACX,UAAM,WAAW,iBAAiB;AAClC,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAE1C,UAAM,cAAc,gBAAgB,EAAE,eAAe;AACrD,UAAM,SAAsB;AAAA,MAC3B;AAAA,MACA,MAAM,MAAM,QAAQ;AAAA,MACpB,YAAY,MAAM,cAAc;AAAA,MAChC,SAAS,MAAM,WAAW,CAAC;AAAA,MAC3B,WAAW,MAAM,aAAa,CAAC;AAAA,IAChC;AACA,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,eAAe,OAAsB;AACpC,UAAQ,MAAM,mCAAmC;AACjD,QAAM,EAAE,eAAe,aAAa,IAAI,oBAAoB;AAC5D,UAAQ,MAAM,oBAAoB,aAAa,EAAE;AACjD,UAAQ,MAAM,SAAS,aAAa,WAAW,aAAa,aAAa,UAAU,GAAG;AACtF,QAAM,YAAY,IAAI,qBAAqB;AAC3C,QAAM,OAAO,QAAQ,SAAS;AAC9B,UAAQ,MAAM,kDAAkD;AACjE;AAEA,KAAK,EAAE,MAAM,CAAC,UAAU;AACvB,UAAQ,MAAM,oCAAoC,KAAK;AACvD,UAAQ,KAAK,CAAC;AACf,CAAC;","names":["result"]}
|
|
1
|
+
{"version":3,"sources":["../../src/mcp/recap-server.ts"],"sourcesContent":["/**\n * Loom Recap MCP Server\n *\n * Captures session context (goal, decisions, insights, risks, assumptions)\n * for the VS Code Loom Context Panel.\n *\n * Environment variables:\n * - RECAP_FILE_PATH: Complete path to the recap.json file (read/write)\n * - LOOM_METADATA_JSON: Stringified JSON of the loom metadata (parsed using LoomMetadata type)\n * - METADATA_FILE_PATH: Complete path to the loom metadata JSON file (for state transition tools)\n */\nimport { McpServer } from '@modelcontextprotocol/sdk/server/mcp.js'\nimport { StdioServerTransport } from '@modelcontextprotocol/sdk/server/stdio.js'\nimport { z } from 'zod'\nimport path from 'path'\nimport os from 'os'\nimport fs from 'fs-extra'\nimport { randomUUID } from 'crypto'\nimport type { RecapFile, RecapEntry, RecapOutput, RecapArtifact } from './recap-types.js'\nimport type { LoomMetadata, MetadataFile, SwarmState } from '../lib/MetadataManager.js'\n\ninterface EnvConfig {\n\trecapFilePath: string\n\tloomMetadata: LoomMetadata\n\tmetadataFilePath: string | null\n}\n\n// Store validated config for use in tool handlers\nlet validatedRecapFilePath: string | null = null\nlet validatedLoomMetadata: LoomMetadata | null = null\nlet validatedMetadataFilePath: string | null = null\n\n/**\n * Validate required environment variables\n * Exits with error if missing (matches issue-management-server.ts pattern)\n */\nfunction validateEnvironment(): EnvConfig {\n\tconst recapFilePath = process.env.RECAP_FILE_PATH\n\tconst loomMetadataJson = process.env.LOOM_METADATA_JSON\n\n\tif (!recapFilePath) {\n\t\tconsole.error('Missing required environment variable: RECAP_FILE_PATH')\n\t\tprocess.exit(1)\n\t}\n\tif (!loomMetadataJson) {\n\t\tconsole.error('Missing required environment variable: LOOM_METADATA_JSON')\n\t\tprocess.exit(1)\n\t}\n\n\tlet loomMetadata: LoomMetadata\n\ttry {\n\t\tloomMetadata = JSON.parse(loomMetadataJson) as LoomMetadata\n\t} catch (error) {\n\t\tconsole.error('Failed to parse LOOM_METADATA_JSON:', error)\n\t\tprocess.exit(1)\n\t}\n\n\t// METADATA_FILE_PATH is optional (only needed for state transition tools)\n\tconst metadataFilePath = process.env.METADATA_FILE_PATH ?? null\n\n\t// Store for tool handlers\n\tvalidatedRecapFilePath = recapFilePath\n\tvalidatedLoomMetadata = loomMetadata\n\tvalidatedMetadataFilePath = metadataFilePath\n\n\treturn { recapFilePath, loomMetadata, metadataFilePath }\n}\n\n/**\n * Get the validated recap file path\n * Throws if called before validateEnvironment()\n */\nfunction getRecapFilePath(): string {\n\tif (!validatedRecapFilePath) {\n\t\tthrow new Error('RECAP_FILE_PATH not validated - validateEnvironment() must be called first')\n\t}\n\treturn validatedRecapFilePath\n}\n\n/**\n * Get the validated loom metadata\n * Throws if called before validateEnvironment()\n */\nexport function getLoomMetadata(): LoomMetadata {\n\tif (!validatedLoomMetadata) {\n\t\tthrow new Error('LOOM_METADATA_JSON not validated - validateEnvironment() must be called first')\n\t}\n\treturn validatedLoomMetadata\n}\n\n/**\n * Get the validated metadata file path\n * Throws if METADATA_FILE_PATH was not provided\n */\nfunction getMetadataFilePath(): string {\n\tif (!validatedMetadataFilePath) {\n\t\tthrow new Error('METADATA_FILE_PATH not configured - state transition tools require this environment variable')\n\t}\n\treturn validatedMetadataFilePath\n}\n\n/**\n * Convert worktree path to filename slug\n * Same algorithm as MetadataManager.slugifyPath() and src/utils/mcp.ts slugifyPath()\n */\nfunction slugifyPath(worktreePath: string): string {\n\tlet slug = worktreePath.replace(/[/\\\\]+$/, '')\n\tslug = slug.replace(/[/\\\\]/g, '___')\n\tslug = slug.replace(/[^a-zA-Z0-9_-]/g, '-')\n\treturn `${slug}.json`\n}\n\n/**\n * Resolve the recap file path.\n * When worktreePath is provided, derives the path dynamically.\n * Otherwise falls back to the env var default.\n */\nexport function resolveRecapFilePath(worktreePath?: string): string {\n\tif (worktreePath) {\n\t\tconst recapsDir = path.join(os.homedir(), '.config', 'iloom-ai', 'recaps')\n\t\treturn path.join(recapsDir, slugifyPath(worktreePath))\n\t}\n\treturn getRecapFilePath()\n}\n\n/**\n * Resolve the metadata file path.\n * When worktreePath is provided, derives the path dynamically.\n * Otherwise falls back to the env var default.\n */\nexport function resolveMetadataFilePath(worktreePath?: string): string {\n\tif (worktreePath) {\n\t\tconst loomsDir = path.join(os.homedir(), '.config', 'iloom-ai', 'looms')\n\t\treturn path.join(loomsDir, slugifyPath(worktreePath))\n\t}\n\treturn getMetadataFilePath()\n}\n\n/**\n * Read recap file (returns empty object if not found or invalid)\n */\nasync function readRecapFile(filePath: string): Promise<RecapFile> {\n\ttry {\n\t\tif (await fs.pathExists(filePath)) {\n\t\t\tconst content = await fs.readFile(filePath, 'utf8')\n\t\t\treturn JSON.parse(content) as RecapFile\n\t\t}\n\t} catch (error) {\n\t\tconsole.error(`Warning: Could not read recap file: ${error}`)\n\t}\n\treturn {}\n}\n\n/**\n * Write recap file (ensures parent directory exists)\n */\nasync function writeRecapFile(filePath: string, recap: RecapFile): Promise<void> {\n\tawait fs.ensureDir(path.dirname(filePath), { mode: 0o755 })\n\tawait fs.writeFile(filePath, JSON.stringify(recap, null, 2), { mode: 0o644 })\n}\n\n// Initialize MCP server\nconst server = new McpServer({\n\tname: 'loom-recap',\n\tversion: '0.1.0',\n})\n\n// Register set_goal tool\nserver.registerTool(\n\t'set_goal',\n\t{\n\t\ttitle: 'Set Goal',\n\t\tdescription: 'Set the initial goal (called once at session start)',\n\t\tinputSchema: {\n\t\t\tgoal: z.string().describe('The original problem statement'),\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope recap to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tsuccess: z.literal(true),\n\t\t},\n\t},\n\tasync ({ goal, worktreePath }) => {\n\t\tconst filePath = resolveRecapFilePath(worktreePath)\n\t\tconst recap = await readRecapFile(filePath)\n\t\trecap.goal = goal\n\t\tawait writeRecapFile(filePath, recap)\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify({ success: true }) }],\n\t\t\tstructuredContent: { success: true },\n\t\t}\n\t}\n)\n\n// Register set_complexity tool\nserver.registerTool(\n\t'set_complexity',\n\t{\n\t\ttitle: 'Set Complexity',\n\t\tdescription: 'Set the assessed complexity of the current task',\n\t\tinputSchema: {\n\t\t\tcomplexity: z.enum(['trivial', 'simple', 'complex']).describe('Task complexity level'),\n\t\t\treason: z.string().optional().describe('Brief explanation for the assessment'),\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope recap to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tsuccess: z.literal(true),\n\t\t\ttimestamp: z.string(),\n\t\t},\n\t},\n\tasync ({ complexity, reason, worktreePath }) => {\n\t\tconst filePath = resolveRecapFilePath(worktreePath)\n\t\tconst recap = await readRecapFile(filePath)\n\t\tconst timestamp = new Date().toISOString()\n\t\trecap.complexity = reason !== undefined ? { level: complexity, reason, timestamp } : { level: complexity, timestamp }\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { success: true as const, timestamp }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register add_entry tool\nserver.registerTool(\n\t'add_entry',\n\t{\n\t\ttitle: 'Add Entry',\n\t\tdescription:\n\t\t\t'Append an entry to the recap. If an entry with the same type and content already exists, it will be skipped.',\n\t\tinputSchema: {\n\t\t\ttype: z\n\t\t\t\t.enum(['decision', 'insight', 'risk', 'assumption', 'other'])\n\t\t\t\t.describe('Entry type'),\n\t\t\tcontent: z.string().describe('Entry content'),\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope recap to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tid: z.string(),\n\t\t\ttimestamp: z.string(),\n\t\t\tskipped: z.boolean(),\n\t\t},\n\t},\n\tasync ({ type, content, worktreePath }) => {\n\t\tconst filePath = resolveRecapFilePath(worktreePath)\n\t\tconst recap = await readRecapFile(filePath)\n\t\trecap.entries ??= []\n\n\t\t// Deduplication: skip if entry with same type and content exists\n\t\tconst existingEntry = recap.entries.find((e) => e.type === type && e.content === content)\n\n\t\tif (existingEntry) {\n\t\t\tconst result = { id: existingEntry.id, timestamp: existingEntry.timestamp, skipped: true }\n\t\t\treturn {\n\t\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\t\tstructuredContent: result,\n\t\t\t}\n\t\t}\n\n\t\tconst entry: RecapEntry = {\n\t\t\tid: randomUUID(),\n\t\t\ttimestamp: new Date().toISOString(),\n\t\t\ttype,\n\t\t\tcontent,\n\t\t}\n\t\trecap.entries.push(entry)\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { id: entry.id, timestamp: entry.timestamp, skipped: false }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register add_artifact tool\nserver.registerTool(\n\t'add_artifact',\n\t{\n\t\ttitle: 'Add Artifact',\n\t\tdescription:\n\t\t\t'Track an artifact (comment, issue, PR) created during the session. If an artifact with the same primaryUrl already exists, it will be replaced.',\n\t\tinputSchema: {\n\t\t\ttype: z.enum(['comment', 'issue', 'pr']).describe('Artifact type'),\n\t\t\tprimaryUrl: z.string().url().describe('Main URL for the artifact'),\n\t\t\tdescription: z.string().describe('Brief description of the artifact'),\n\t\t\tid: z.string().optional().describe('Optional artifact ID (e.g., comment ID, issue number)'),\n\t\t\turls: z.record(z.string()).optional().describe('Optional additional URLs (e.g., { api: \"...\" })'),\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope recap to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tid: z.string(),\n\t\t\ttimestamp: z.string(),\n\t\t\treplaced: z.boolean(),\n\t\t},\n\t},\n\tasync ({ type, primaryUrl, description, id, urls, worktreePath }) => {\n\t\tconst filePath = resolveRecapFilePath(worktreePath)\n\t\tconst recap = await readRecapFile(filePath)\n\n\t\tconst artifact: RecapArtifact = {\n\t\t\tid: id ?? randomUUID(),\n\t\t\ttype,\n\t\t\tprimaryUrl,\n\t\t\turls: urls ?? {},\n\t\t\tdescription,\n\t\t\ttimestamp: new Date().toISOString(),\n\t\t}\n\n\t\trecap.artifacts ??= []\n\n\t\t// Deduplication: replace existing artifact with same primaryUrl\n\t\tconst existingIndex = recap.artifacts.findIndex((a) => a.primaryUrl === primaryUrl)\n\t\tconst replaced = existingIndex !== -1\n\n\t\tif (replaced) {\n\t\t\t// Preserve the original id if not explicitly provided\n\t\t\tconst existingArtifact = recap.artifacts[existingIndex]\n\t\t\tif (existingArtifact) {\n\t\t\t\tartifact.id = id ?? existingArtifact.id\n\t\t\t\trecap.artifacts[existingIndex] = artifact\n\t\t\t}\n\t\t} else {\n\t\t\trecap.artifacts.push(artifact)\n\t\t}\n\n\t\tawait writeRecapFile(filePath, recap)\n\t\tconst result = { id: artifact.id, timestamp: artifact.timestamp, replaced }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register get_recap tool\nserver.registerTool(\n\t'get_recap',\n\t{\n\t\ttitle: 'Get Recap',\n\t\tdescription: 'Read current recap (for catching up or review)',\n\t\tinputSchema: {\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope recap to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tfilePath: z.string(),\n\t\t\tgoal: z.string().nullable(),\n\t\t\tcomplexity: z\n\t\t\t\t.object({\n\t\t\t\t\tlevel: z.enum(['trivial', 'simple', 'complex']),\n\t\t\t\t\treason: z.string().optional(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t})\n\t\t\t\t.nullable(),\n\t\t\tentries: z.array(\n\t\t\t\tz.object({\n\t\t\t\t\tid: z.string(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t\ttype: z.enum(['decision', 'insight', 'risk', 'assumption', 'other']),\n\t\t\t\t\tcontent: z.string(),\n\t\t\t\t})\n\t\t\t),\n\t\t\tartifacts: z.array(\n\t\t\t\tz.object({\n\t\t\t\t\tid: z.string(),\n\t\t\t\t\ttype: z.enum(['comment', 'issue', 'pr']),\n\t\t\t\t\tprimaryUrl: z.string(),\n\t\t\t\t\turls: z.record(z.string()),\n\t\t\t\t\tdescription: z.string(),\n\t\t\t\t\ttimestamp: z.string(),\n\t\t\t\t})\n\t\t\t),\n\t\t},\n\t},\n\tasync ({ worktreePath }) => {\n\t\tconst filePath = resolveRecapFilePath(worktreePath)\n\t\tconst recap = await readRecapFile(filePath)\n\t\t// Use loom description as default goal for new/missing recap files\n\t\t// When worktreePath is provided, read metadata from the target worktree's metadata file\n\t\tlet defaultGoal: string | null = null\n\t\tif (worktreePath) {\n\t\t\ttry {\n\t\t\t\tconst metaPath = resolveMetadataFilePath(worktreePath)\n\t\t\t\tif (await fs.pathExists(metaPath)) {\n\t\t\t\t\tconst metaContent = await fs.readFile(metaPath, 'utf8')\n\t\t\t\t\tconst meta = JSON.parse(metaContent) as MetadataFile\n\t\t\t\t\tdefaultGoal = meta.description || null\n\t\t\t\t}\n\t\t\t} catch {\n\t\t\t\t// Fall through to null default goal\n\t\t\t}\n\t\t} else {\n\t\t\tdefaultGoal = getLoomMetadata().description || null\n\t\t}\n\t\tconst result: RecapOutput = {\n\t\t\tfilePath,\n\t\t\tgoal: recap.goal ?? defaultGoal,\n\t\t\tcomplexity: recap.complexity ?? null,\n\t\t\tentries: recap.entries ?? [],\n\t\t\tartifacts: recap.artifacts ?? [],\n\t\t}\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result as unknown as Record<string, unknown>,\n\t\t}\n\t}\n)\n\n// Zod schema for swarm state values\nconst swarmStateSchema = z.enum(['pending', 'in_progress', 'code_review', 'done', 'failed'])\n\n// Register set_loom_state tool\nserver.registerTool(\n\t'set_loom_state',\n\t{\n\t\ttitle: 'Set Loom State',\n\t\tdescription: 'Set the swarm lifecycle state of the current loom',\n\t\tinputSchema: {\n\t\t\tstate: swarmStateSchema.describe('The new state for the loom'),\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope state operations to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tsuccess: z.literal(true),\n\t\t\tstate: swarmStateSchema,\n\t\t},\n\t},\n\tasync ({ state, worktreePath }) => {\n\t\tconst metadataFilePath = resolveMetadataFilePath(worktreePath)\n\n\t\t// Read existing metadata\n\t\tlet metadata: MetadataFile\n\t\ttry {\n\t\t\tconst content = await fs.readFile(metadataFilePath, 'utf8')\n\t\t\tmetadata = JSON.parse(content) as MetadataFile\n\t\t} catch (error) {\n\t\t\tthrow new Error(`Failed to read metadata file at ${metadataFilePath}: ${error instanceof Error ? error.message : String(error)}`)\n\t\t}\n\n\t\t// Update state\n\t\tmetadata.state = state as SwarmState\n\n\t\t// Write back\n\t\tawait fs.writeFile(metadataFilePath, JSON.stringify(metadata, null, 2), { mode: 0o644 })\n\n\t\tconst result = { success: true as const, state: state as SwarmState }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Register get_loom_state tool\nserver.registerTool(\n\t'get_loom_state',\n\t{\n\t\ttitle: 'Get Loom State',\n\t\tdescription: 'Get the current swarm lifecycle state of the loom',\n\t\tinputSchema: {\n\t\t\tworktreePath: z.string().optional().describe('Optional worktree path to scope state operations to a specific loom'),\n\t\t},\n\t\toutputSchema: {\n\t\t\tstate: swarmStateSchema.nullable(),\n\t\t},\n\t},\n\tasync ({ worktreePath }) => {\n\t\tconst metadataFilePath = resolveMetadataFilePath(worktreePath)\n\n\t\t// Read metadata\n\t\tlet metadata: MetadataFile\n\t\ttry {\n\t\t\tconst content = await fs.readFile(metadataFilePath, 'utf8')\n\t\t\tmetadata = JSON.parse(content) as MetadataFile\n\t\t} catch (error) {\n\t\t\tthrow new Error(`Failed to read metadata file at ${metadataFilePath}: ${error instanceof Error ? error.message : String(error)}`)\n\t\t}\n\n\t\tconst result = { state: metadata.state ?? null }\n\t\treturn {\n\t\t\tcontent: [{ type: 'text' as const, text: JSON.stringify(result) }],\n\t\t\tstructuredContent: result,\n\t\t}\n\t}\n)\n\n// Main server startup\nasync function main(): Promise<void> {\n\tconsole.error('=== Loom Recap MCP Server Starting ===')\n\tconsole.error(`PID: ${process.pid}`)\n\tconsole.error(`Node version: ${process.version}`)\n\tconsole.error(`CWD: ${process.cwd()}`)\n\tconsole.error(`Script: ${new URL(import.meta.url).pathname}`)\n\n\t// Log relevant env vars (LOOM_METADATA_JSON is large, just log presence and length)\n\tconsole.error('Environment variables:')\n\tconsole.error(` RECAP_FILE_PATH=${process.env.RECAP_FILE_PATH ?? '<not set>'}`)\n\tconsole.error(` METADATA_FILE_PATH=${process.env.METADATA_FILE_PATH ?? '<not set>'}`)\n\tconsole.error(` LOOM_METADATA_JSON=${process.env.LOOM_METADATA_JSON ? `<set, ${process.env.LOOM_METADATA_JSON.length} chars>` : '<not set>'}`)\n\n\tconst { recapFilePath, loomMetadata, metadataFilePath } = validateEnvironment()\n\tconsole.error(`Recap file path: ${recapFilePath}`)\n\tconsole.error(`Metadata file path: ${metadataFilePath ?? '<not configured>'}`)\n\tconsole.error(`Loom: ${loomMetadata.description} (branch: ${loomMetadata.branchName})`)\n\n\t// Check if recap file already exists\n\tconst recapExists = await fs.pathExists(recapFilePath)\n\tconsole.error(`Recap file exists: ${recapExists}`)\n\n\tconst transport = new StdioServerTransport()\n\tawait server.connect(transport)\n\tconsole.error('=== Loom Recap MCP Server READY (stdio transport) ===')\n}\n\nmain().catch((error) => {\n\tconsole.error('Fatal error starting MCP server:', error)\n\tprocess.exit(1)\n})\n"],"mappings":";;;AAWA,SAAS,iBAAiB;AAC1B,SAAS,4BAA4B;AACrC,SAAS,SAAS;AAClB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,OAAO,QAAQ;AACf,SAAS,kBAAkB;AAW3B,IAAI,yBAAwC;AAC5C,IAAI,wBAA6C;AACjD,IAAI,4BAA2C;AAM/C,SAAS,sBAAiC;AACzC,QAAM,gBAAgB,QAAQ,IAAI;AAClC,QAAM,mBAAmB,QAAQ,IAAI;AAErC,MAAI,CAAC,eAAe;AACnB,YAAQ,MAAM,wDAAwD;AACtE,YAAQ,KAAK,CAAC;AAAA,EACf;AACA,MAAI,CAAC,kBAAkB;AACtB,YAAQ,MAAM,2DAA2D;AACzE,YAAQ,KAAK,CAAC;AAAA,EACf;AAEA,MAAI;AACJ,MAAI;AACH,mBAAe,KAAK,MAAM,gBAAgB;AAAA,EAC3C,SAAS,OAAO;AACf,YAAQ,MAAM,uCAAuC,KAAK;AAC1D,YAAQ,KAAK,CAAC;AAAA,EACf;AAGA,QAAM,mBAAmB,QAAQ,IAAI,sBAAsB;AAG3D,2BAAyB;AACzB,0BAAwB;AACxB,8BAA4B;AAE5B,SAAO,EAAE,eAAe,cAAc,iBAAiB;AACxD;AAMA,SAAS,mBAA2B;AACnC,MAAI,CAAC,wBAAwB;AAC5B,UAAM,IAAI,MAAM,4EAA4E;AAAA,EAC7F;AACA,SAAO;AACR;AAMO,SAAS,kBAAgC;AAC/C,MAAI,CAAC,uBAAuB;AAC3B,UAAM,IAAI,MAAM,+EAA+E;AAAA,EAChG;AACA,SAAO;AACR;AAMA,SAAS,sBAA8B;AACtC,MAAI,CAAC,2BAA2B;AAC/B,UAAM,IAAI,MAAM,8FAA8F;AAAA,EAC/G;AACA,SAAO;AACR;AAMA,SAAS,YAAY,cAA8B;AAClD,MAAI,OAAO,aAAa,QAAQ,WAAW,EAAE;AAC7C,SAAO,KAAK,QAAQ,UAAU,KAAK;AACnC,SAAO,KAAK,QAAQ,mBAAmB,GAAG;AAC1C,SAAO,GAAG,IAAI;AACf;AAOO,SAAS,qBAAqB,cAA+B;AACnE,MAAI,cAAc;AACjB,UAAM,YAAY,KAAK,KAAK,GAAG,QAAQ,GAAG,WAAW,YAAY,QAAQ;AACzE,WAAO,KAAK,KAAK,WAAW,YAAY,YAAY,CAAC;AAAA,EACtD;AACA,SAAO,iBAAiB;AACzB;AAOO,SAAS,wBAAwB,cAA+B;AACtE,MAAI,cAAc;AACjB,UAAM,WAAW,KAAK,KAAK,GAAG,QAAQ,GAAG,WAAW,YAAY,OAAO;AACvE,WAAO,KAAK,KAAK,UAAU,YAAY,YAAY,CAAC;AAAA,EACrD;AACA,SAAO,oBAAoB;AAC5B;AAKA,eAAe,cAAc,UAAsC;AAClE,MAAI;AACH,QAAI,MAAM,GAAG,WAAW,QAAQ,GAAG;AAClC,YAAM,UAAU,MAAM,GAAG,SAAS,UAAU,MAAM;AAClD,aAAO,KAAK,MAAM,OAAO;AAAA,IAC1B;AAAA,EACD,SAAS,OAAO;AACf,YAAQ,MAAM,uCAAuC,KAAK,EAAE;AAAA,EAC7D;AACA,SAAO,CAAC;AACT;AAKA,eAAe,eAAe,UAAkB,OAAiC;AAChF,QAAM,GAAG,UAAU,KAAK,QAAQ,QAAQ,GAAG,EAAE,MAAM,IAAM,CAAC;AAC1D,QAAM,GAAG,UAAU,UAAU,KAAK,UAAU,OAAO,MAAM,CAAC,GAAG,EAAE,MAAM,IAAM,CAAC;AAC7E;AAGA,IAAM,SAAS,IAAI,UAAU;AAAA,EAC5B,MAAM;AAAA,EACN,SAAS;AACV,CAAC;AAGD,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,MAAM,EAAE,OAAO,EAAE,SAAS,gCAAgC;AAAA,MAC1D,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,0DAA0D;AAAA,IACxG;AAAA,IACA,cAAc;AAAA,MACb,SAAS,EAAE,QAAQ,IAAI;AAAA,IACxB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,MAAM,aAAa,MAAM;AACjC,UAAM,WAAW,qBAAqB,YAAY;AAClD,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,OAAO;AACb,UAAM,eAAe,UAAU,KAAK;AACpC,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,EAAE,SAAS,KAAK,CAAC,EAAE,CAAC;AAAA,MAC5E,mBAAmB,EAAE,SAAS,KAAK;AAAA,IACpC;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,YAAY,EAAE,KAAK,CAAC,WAAW,UAAU,SAAS,CAAC,EAAE,SAAS,uBAAuB;AAAA,MACrF,QAAQ,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,sCAAsC;AAAA,MAC7E,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,0DAA0D;AAAA,IACxG;AAAA,IACA,cAAc;AAAA,MACb,SAAS,EAAE,QAAQ,IAAI;AAAA,MACvB,WAAW,EAAE,OAAO;AAAA,IACrB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,YAAY,QAAQ,aAAa,MAAM;AAC/C,UAAM,WAAW,qBAAqB,YAAY;AAClD,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,aAAY,oBAAI,KAAK,GAAE,YAAY;AACzC,UAAM,aAAa,WAAW,SAAY,EAAE,OAAO,YAAY,QAAQ,UAAU,IAAI,EAAE,OAAO,YAAY,UAAU;AACpH,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,SAAS,MAAe,UAAU;AACnD,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aACC;AAAA,IACD,aAAa;AAAA,MACZ,MAAM,EACJ,KAAK,CAAC,YAAY,WAAW,QAAQ,cAAc,OAAO,CAAC,EAC3D,SAAS,YAAY;AAAA,MACvB,SAAS,EAAE,OAAO,EAAE,SAAS,eAAe;AAAA,MAC5C,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,0DAA0D;AAAA,IACxG;AAAA,IACA,cAAc;AAAA,MACb,IAAI,EAAE,OAAO;AAAA,MACb,WAAW,EAAE,OAAO;AAAA,MACpB,SAAS,EAAE,QAAQ;AAAA,IACpB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,MAAM,SAAS,aAAa,MAAM;AAC1C,UAAM,WAAW,qBAAqB,YAAY;AAClD,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAC1C,UAAM,YAAY,CAAC;AAGnB,UAAM,gBAAgB,MAAM,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,QAAQ,EAAE,YAAY,OAAO;AAExF,QAAI,eAAe;AAClB,YAAMA,UAAS,EAAE,IAAI,cAAc,IAAI,WAAW,cAAc,WAAW,SAAS,KAAK;AACzF,aAAO;AAAA,QACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAUA,OAAM,EAAE,CAAC;AAAA,QACjE,mBAAmBA;AAAA,MACpB;AAAA,IACD;AAEA,UAAM,QAAoB;AAAA,MACzB,IAAI,WAAW;AAAA,MACf,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,MAClC;AAAA,MACA;AAAA,IACD;AACA,UAAM,QAAQ,KAAK,KAAK;AACxB,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,IAAI,MAAM,IAAI,WAAW,MAAM,WAAW,SAAS,MAAM;AAC1E,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aACC;AAAA,IACD,aAAa;AAAA,MACZ,MAAM,EAAE,KAAK,CAAC,WAAW,SAAS,IAAI,CAAC,EAAE,SAAS,eAAe;AAAA,MACjE,YAAY,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,2BAA2B;AAAA,MACjE,aAAa,EAAE,OAAO,EAAE,SAAS,mCAAmC;AAAA,MACpE,IAAI,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,uDAAuD;AAAA,MAC1F,MAAM,EAAE,OAAO,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS,iDAAiD;AAAA,MAChG,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,0DAA0D;AAAA,IACxG;AAAA,IACA,cAAc;AAAA,MACb,IAAI,EAAE,OAAO;AAAA,MACb,WAAW,EAAE,OAAO;AAAA,MACpB,UAAU,EAAE,QAAQ;AAAA,IACrB;AAAA,EACD;AAAA,EACA,OAAO,EAAE,MAAM,YAAY,aAAa,IAAI,MAAM,aAAa,MAAM;AACpE,UAAM,WAAW,qBAAqB,YAAY;AAClD,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAE1C,UAAM,WAA0B;AAAA,MAC/B,IAAI,MAAM,WAAW;AAAA,MACrB;AAAA,MACA;AAAA,MACA,MAAM,QAAQ,CAAC;AAAA,MACf;AAAA,MACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IACnC;AAEA,UAAM,cAAc,CAAC;AAGrB,UAAM,gBAAgB,MAAM,UAAU,UAAU,CAAC,MAAM,EAAE,eAAe,UAAU;AAClF,UAAM,WAAW,kBAAkB;AAEnC,QAAI,UAAU;AAEb,YAAM,mBAAmB,MAAM,UAAU,aAAa;AACtD,UAAI,kBAAkB;AACrB,iBAAS,KAAK,MAAM,iBAAiB;AACrC,cAAM,UAAU,aAAa,IAAI;AAAA,MAClC;AAAA,IACD,OAAO;AACN,YAAM,UAAU,KAAK,QAAQ;AAAA,IAC9B;AAEA,UAAM,eAAe,UAAU,KAAK;AACpC,UAAM,SAAS,EAAE,IAAI,SAAS,IAAI,WAAW,SAAS,WAAW,SAAS;AAC1E,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,0DAA0D;AAAA,IACxG;AAAA,IACA,cAAc;AAAA,MACb,UAAU,EAAE,OAAO;AAAA,MACnB,MAAM,EAAE,OAAO,EAAE,SAAS;AAAA,MAC1B,YAAY,EACV,OAAO;AAAA,QACP,OAAO,EAAE,KAAK,CAAC,WAAW,UAAU,SAAS,CAAC;AAAA,QAC9C,QAAQ,EAAE,OAAO,EAAE,SAAS;AAAA,QAC5B,WAAW,EAAE,OAAO;AAAA,MACrB,CAAC,EACA,SAAS;AAAA,MACX,SAAS,EAAE;AAAA,QACV,EAAE,OAAO;AAAA,UACR,IAAI,EAAE,OAAO;AAAA,UACb,WAAW,EAAE,OAAO;AAAA,UACpB,MAAM,EAAE,KAAK,CAAC,YAAY,WAAW,QAAQ,cAAc,OAAO,CAAC;AAAA,UACnE,SAAS,EAAE,OAAO;AAAA,QACnB,CAAC;AAAA,MACF;AAAA,MACA,WAAW,EAAE;AAAA,QACZ,EAAE,OAAO;AAAA,UACR,IAAI,EAAE,OAAO;AAAA,UACb,MAAM,EAAE,KAAK,CAAC,WAAW,SAAS,IAAI,CAAC;AAAA,UACvC,YAAY,EAAE,OAAO;AAAA,UACrB,MAAM,EAAE,OAAO,EAAE,OAAO,CAAC;AAAA,UACzB,aAAa,EAAE,OAAO;AAAA,UACtB,WAAW,EAAE,OAAO;AAAA,QACrB,CAAC;AAAA,MACF;AAAA,IACD;AAAA,EACD;AAAA,EACA,OAAO,EAAE,aAAa,MAAM;AAC3B,UAAM,WAAW,qBAAqB,YAAY;AAClD,UAAM,QAAQ,MAAM,cAAc,QAAQ;AAG1C,QAAI,cAA6B;AACjC,QAAI,cAAc;AACjB,UAAI;AACH,cAAM,WAAW,wBAAwB,YAAY;AACrD,YAAI,MAAM,GAAG,WAAW,QAAQ,GAAG;AAClC,gBAAM,cAAc,MAAM,GAAG,SAAS,UAAU,MAAM;AACtD,gBAAM,OAAO,KAAK,MAAM,WAAW;AACnC,wBAAc,KAAK,eAAe;AAAA,QACnC;AAAA,MACD,QAAQ;AAAA,MAER;AAAA,IACD,OAAO;AACN,oBAAc,gBAAgB,EAAE,eAAe;AAAA,IAChD;AACA,UAAM,SAAsB;AAAA,MAC3B;AAAA,MACA,MAAM,MAAM,QAAQ;AAAA,MACpB,YAAY,MAAM,cAAc;AAAA,MAChC,SAAS,MAAM,WAAW,CAAC;AAAA,MAC3B,WAAW,MAAM,aAAa,CAAC;AAAA,IAChC;AACA,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,IAAM,mBAAmB,EAAE,KAAK,CAAC,WAAW,eAAe,eAAe,QAAQ,QAAQ,CAAC;AAG3F,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,OAAO,iBAAiB,SAAS,4BAA4B;AAAA,MAC7D,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,qEAAqE;AAAA,IACnH;AAAA,IACA,cAAc;AAAA,MACb,SAAS,EAAE,QAAQ,IAAI;AAAA,MACvB,OAAO;AAAA,IACR;AAAA,EACD;AAAA,EACA,OAAO,EAAE,OAAO,aAAa,MAAM;AAClC,UAAM,mBAAmB,wBAAwB,YAAY;AAG7D,QAAI;AACJ,QAAI;AACH,YAAM,UAAU,MAAM,GAAG,SAAS,kBAAkB,MAAM;AAC1D,iBAAW,KAAK,MAAM,OAAO;AAAA,IAC9B,SAAS,OAAO;AACf,YAAM,IAAI,MAAM,mCAAmC,gBAAgB,KAAK,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK,CAAC,EAAE;AAAA,IACjI;AAGA,aAAS,QAAQ;AAGjB,UAAM,GAAG,UAAU,kBAAkB,KAAK,UAAU,UAAU,MAAM,CAAC,GAAG,EAAE,MAAM,IAAM,CAAC;AAEvF,UAAM,SAAS,EAAE,SAAS,MAAe,MAA2B;AACpE,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,OAAO;AAAA,EACN;AAAA,EACA;AAAA,IACC,OAAO;AAAA,IACP,aAAa;AAAA,IACb,aAAa;AAAA,MACZ,cAAc,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,qEAAqE;AAAA,IACnH;AAAA,IACA,cAAc;AAAA,MACb,OAAO,iBAAiB,SAAS;AAAA,IAClC;AAAA,EACD;AAAA,EACA,OAAO,EAAE,aAAa,MAAM;AAC3B,UAAM,mBAAmB,wBAAwB,YAAY;AAG7D,QAAI;AACJ,QAAI;AACH,YAAM,UAAU,MAAM,GAAG,SAAS,kBAAkB,MAAM;AAC1D,iBAAW,KAAK,MAAM,OAAO;AAAA,IAC9B,SAAS,OAAO;AACf,YAAM,IAAI,MAAM,mCAAmC,gBAAgB,KAAK,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK,CAAC,EAAE;AAAA,IACjI;AAEA,UAAM,SAAS,EAAE,OAAO,SAAS,SAAS,KAAK;AAC/C,WAAO;AAAA,MACN,SAAS,CAAC,EAAE,MAAM,QAAiB,MAAM,KAAK,UAAU,MAAM,EAAE,CAAC;AAAA,MACjE,mBAAmB;AAAA,IACpB;AAAA,EACD;AACD;AAGA,eAAe,OAAsB;AACpC,UAAQ,MAAM,wCAAwC;AACtD,UAAQ,MAAM,QAAQ,QAAQ,GAAG,EAAE;AACnC,UAAQ,MAAM,iBAAiB,QAAQ,OAAO,EAAE;AAChD,UAAQ,MAAM,QAAQ,QAAQ,IAAI,CAAC,EAAE;AACrC,UAAQ,MAAM,WAAW,IAAI,IAAI,YAAY,GAAG,EAAE,QAAQ,EAAE;AAG5D,UAAQ,MAAM,wBAAwB;AACtC,UAAQ,MAAM,qBAAqB,QAAQ,IAAI,mBAAmB,WAAW,EAAE;AAC/E,UAAQ,MAAM,wBAAwB,QAAQ,IAAI,sBAAsB,WAAW,EAAE;AACrF,UAAQ,MAAM,wBAAwB,QAAQ,IAAI,qBAAqB,SAAS,QAAQ,IAAI,mBAAmB,MAAM,YAAY,WAAW,EAAE;AAE9I,QAAM,EAAE,eAAe,cAAc,iBAAiB,IAAI,oBAAoB;AAC9E,UAAQ,MAAM,oBAAoB,aAAa,EAAE;AACjD,UAAQ,MAAM,uBAAuB,oBAAoB,kBAAkB,EAAE;AAC7E,UAAQ,MAAM,SAAS,aAAa,WAAW,aAAa,aAAa,UAAU,GAAG;AAGtF,QAAM,cAAc,MAAM,GAAG,WAAW,aAAa;AACrD,UAAQ,MAAM,sBAAsB,WAAW,EAAE;AAEjD,QAAM,YAAY,IAAI,qBAAqB;AAC3C,QAAM,OAAO,QAAQ,SAAS;AAC9B,UAAQ,MAAM,uDAAuD;AACtE;AAEA,KAAK,EAAE,MAAM,CAAC,UAAU;AACvB,UAAQ,MAAM,oCAAoC,KAAK;AACvD,UAAQ,KAAK,CAAC;AACf,CAAC;","names":["result"]}
|
|
@@ -1,11 +1,11 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
2
|
import {
|
|
3
3
|
createNeonProviderFromSettings
|
|
4
|
-
} from "./chunk-
|
|
5
|
-
import "./chunk-7JDMYTFZ.js";
|
|
4
|
+
} from "./chunk-P4O6EH46.js";
|
|
6
5
|
import "./chunk-6MLEBAYZ.js";
|
|
6
|
+
import "./chunk-7JDMYTFZ.js";
|
|
7
7
|
import "./chunk-VT4PDUYT.js";
|
|
8
8
|
export {
|
|
9
9
|
createNeonProviderFromSettings
|
|
10
10
|
};
|
|
11
|
-
//# sourceMappingURL=neon-helpers-
|
|
11
|
+
//# sourceMappingURL=neon-helpers-CQN2PB4S.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":[],"sourcesContent":[],"mappings":"","names":[]}
|
|
@@ -1,35 +1,35 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
2
|
import {
|
|
3
3
|
DevServerManager
|
|
4
|
-
} from "./chunk-
|
|
5
|
-
import "./chunk-
|
|
4
|
+
} from "./chunk-5MWV33NN.js";
|
|
5
|
+
import "./chunk-ZEWU5PZK.js";
|
|
6
6
|
import {
|
|
7
7
|
getWorkspacePort
|
|
8
|
-
} from "./chunk-
|
|
8
|
+
} from "./chunk-3D7WQM7I.js";
|
|
9
|
+
import "./chunk-WWKOVDWC.js";
|
|
9
10
|
import {
|
|
10
11
|
IdentifierParser
|
|
11
|
-
} from "./chunk-
|
|
12
|
+
} from "./chunk-WY4QBK43.js";
|
|
13
|
+
import {
|
|
14
|
+
GitWorktreeManager
|
|
15
|
+
} from "./chunk-VGGST52X.js";
|
|
12
16
|
import {
|
|
13
17
|
openBrowser
|
|
14
18
|
} from "./chunk-YETJNRQM.js";
|
|
15
|
-
import "./chunk-4LKGCFGG.js";
|
|
16
19
|
import {
|
|
17
|
-
|
|
18
|
-
} from "./chunk-
|
|
20
|
+
ProjectCapabilityDetector
|
|
21
|
+
} from "./chunk-MORRVYPT.js";
|
|
22
|
+
import "./chunk-YQ57ORTV.js";
|
|
19
23
|
import {
|
|
20
24
|
extractSettingsOverrides
|
|
21
25
|
} from "./chunk-GYCR2LOU.js";
|
|
22
|
-
import {
|
|
23
|
-
ProjectCapabilityDetector
|
|
24
|
-
} from "./chunk-TL72BGP6.js";
|
|
25
|
-
import "./chunk-VOGGLPG5.js";
|
|
26
26
|
import {
|
|
27
27
|
extractIssueNumber
|
|
28
|
-
} from "./chunk-
|
|
28
|
+
} from "./chunk-MNHZB4Z2.js";
|
|
29
29
|
import {
|
|
30
30
|
SettingsManager
|
|
31
|
-
} from "./chunk-
|
|
32
|
-
import "./chunk-
|
|
31
|
+
} from "./chunk-YYAKPQBT.js";
|
|
32
|
+
import "./chunk-KB64WNBZ.js";
|
|
33
33
|
import "./chunk-6MLEBAYZ.js";
|
|
34
34
|
import {
|
|
35
35
|
logger
|
|
@@ -238,4 +238,4 @@ Make sure the project is built (run 'il start' first)`
|
|
|
238
238
|
export {
|
|
239
239
|
OpenCommand
|
|
240
240
|
};
|
|
241
|
-
//# sourceMappingURL=open-
|
|
241
|
+
//# sourceMappingURL=open-5QZGXQRF.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/commands/open.ts"],"sourcesContent":["import path from 'path'\nimport fs from 'fs-extra'\nimport { execa } from 'execa'\nimport { GitWorktreeManager } from '../lib/GitWorktreeManager.js'\nimport { ProjectCapabilityDetector } from '../lib/ProjectCapabilityDetector.js'\nimport { DevServerManager } from '../lib/DevServerManager.js'\nimport { SettingsManager } from '../lib/SettingsManager.js'\nimport { IdentifierParser } from '../utils/IdentifierParser.js'\nimport { openBrowser } from '../utils/browser.js'\nimport { getWorkspacePort } from '../utils/port.js'\nimport { extractIssueNumber } from '../utils/git.js'\nimport { logger } from '../utils/logger.js'\nimport { extractSettingsOverrides } from '../utils/cli-overrides.js'\nimport type { GitWorktree } from '../types/worktree.js'\n\nexport interface OpenCommandInput {\n\tidentifier?: string\n\targs?: string[]\n}\n\ninterface ParsedOpenInput {\n\ttype: 'issue' | 'pr' | 'branch' | 'epic'\n\tnumber?: string | number // For issues and PRs\n\tbranchName?: string // For branches\n\toriginalInput: string\n\tautoDetected: boolean\n}\n\n/**\n * OpenCommand - Opens workspace in browser or runs CLI tool\n * Priority: Web first, CLI fallback\n */\nexport class OpenCommand {\n\tconstructor(\n\t\tprivate gitWorktreeManager = new GitWorktreeManager(),\n\t\tprivate capabilityDetector = new ProjectCapabilityDetector(),\n\t\tprivate identifierParser = new IdentifierParser(new GitWorktreeManager()),\n\t\tprivate devServerManager = new DevServerManager(),\n\t\tprivate settingsManager = new SettingsManager()\n\t) {}\n\n\tasync execute(input: OpenCommandInput): Promise<void> {\n\t\t// 1. Parse or auto-detect identifier\n\t\tconst parsed = input.identifier\n\t\t\t? await this.parseExplicitInput(input.identifier)\n\t\t\t: await this.autoDetectFromCurrentDirectory()\n\n\t\tlogger.debug(`Parsed input: ${JSON.stringify(parsed)}`)\n\n\t\t// 2. Find worktree path based on identifier\n\t\tconst worktree = await this.findWorktreeForIdentifier(parsed)\n\n\t\tlogger.info(`Found worktree at: ${worktree.path}`)\n\n\t\t// 3. Detect project capabilities\n\t\tconst { capabilities, binEntries } =\n\t\t\tawait this.capabilityDetector.detectCapabilities(worktree.path)\n\n\t\tlogger.debug(`Detected capabilities: ${capabilities.join(', ')}`)\n\n\t\t// 4. Execute based on capabilities (web first, CLI fallback)\n\t\tif (capabilities.includes('web')) {\n\t\t\tawait this.openWebBrowser(worktree)\n\t\t} else if (capabilities.includes('cli')) {\n\t\t\tawait this.runCLITool(worktree.path, binEntries, input.args ?? [])\n\t\t} else {\n\t\t\tthrow new Error(\n\t\t\t\t`No web or CLI capabilities detected for workspace at ${worktree.path}`\n\t\t\t)\n\t\t}\n\t}\n\n\t/**\n\t * Parse explicit identifier input\n\t */\n\tprivate async parseExplicitInput(identifier: string): Promise<ParsedOpenInput> {\n\t\tconst parsed = await this.identifierParser.parseForPatternDetection(identifier)\n\n\t\t// Description type should never reach open command (converted in start)\n\t\tif (parsed.type === 'description') {\n\t\t\tthrow new Error('Description input type is not supported in open command')\n\t\t}\n\n\t\tconst result: ParsedOpenInput = {\n\t\t\ttype: parsed.type,\n\t\t\toriginalInput: parsed.originalInput,\n\t\t\tautoDetected: false,\n\t\t}\n\n\t\tif (parsed.number !== undefined) {\n\t\t\tresult.number = parsed.number\n\t\t}\n\t\tif (parsed.branchName !== undefined) {\n\t\t\tresult.branchName = parsed.branchName\n\t\t}\n\n\t\treturn result\n\t}\n\n\t/**\n\t * Auto-detect identifier from current directory\n\t * Same logic as FinishCommand.autoDetectFromCurrentDirectory()\n\t */\n\tprivate async autoDetectFromCurrentDirectory(): Promise<ParsedOpenInput> {\n\t\tconst currentDir = path.basename(process.cwd())\n\n\t\t// Check for PR worktree pattern: _pr_N suffix\n\t\tconst prPattern = /_pr_(\\d+)$/\n\t\tconst prMatch = currentDir.match(prPattern)\n\n\t\tif (prMatch?.[1]) {\n\t\t\tconst prNumber = parseInt(prMatch[1], 10)\n\t\t\tlogger.debug(`Auto-detected PR #${prNumber} from directory: ${currentDir}`)\n\t\t\treturn {\n\t\t\t\ttype: 'pr',\n\t\t\t\tnumber: prNumber,\n\t\t\t\toriginalInput: currentDir,\n\t\t\t\tautoDetected: true,\n\t\t\t}\n\t\t}\n\n\t\t// Check for issue pattern in directory\n\t\tconst issueNumber = extractIssueNumber(currentDir)\n\n\t\tif (issueNumber !== null) {\n\t\t\tlogger.debug(`Auto-detected issue #${issueNumber} from directory: ${currentDir}`)\n\t\t\treturn {\n\t\t\t\ttype: 'issue',\n\t\t\t\tnumber: issueNumber,\n\t\t\t\toriginalInput: currentDir,\n\t\t\t\tautoDetected: true,\n\t\t\t}\n\t\t}\n\n\t\t// Fallback: get current branch name\n\t\tconst repoInfo = await this.gitWorktreeManager.getRepoInfo()\n\t\tconst currentBranch = repoInfo.currentBranch\n\n\t\tif (!currentBranch) {\n\t\t\tthrow new Error(\n\t\t\t\t'Could not auto-detect identifier. Please provide an issue number, PR number, or branch name.\\n' +\n\t\t\t\t\t'Expected directory pattern: feat/issue-XX-description OR worktree with _pr_N suffix'\n\t\t\t)\n\t\t}\n\n\t\t// Try to extract issue from branch name\n\t\tconst branchIssueNumber = extractIssueNumber(currentBranch)\n\t\tif (branchIssueNumber !== null) {\n\t\t\tlogger.debug(`Auto-detected issue #${branchIssueNumber} from branch: ${currentBranch}`)\n\t\t\treturn {\n\t\t\t\ttype: 'issue',\n\t\t\t\tnumber: branchIssueNumber,\n\t\t\t\toriginalInput: currentBranch,\n\t\t\t\tautoDetected: true,\n\t\t\t}\n\t\t}\n\n\t\t// Last resort: use branch name\n\t\treturn {\n\t\t\ttype: 'branch',\n\t\t\tbranchName: currentBranch,\n\t\t\toriginalInput: currentBranch,\n\t\t\tautoDetected: true,\n\t\t}\n\t}\n\n\t/**\n\t * Find worktree for the given identifier\n\t */\n\tprivate async findWorktreeForIdentifier(parsed: ParsedOpenInput): Promise<GitWorktree> {\n\t\tlet worktree: GitWorktree | null = null\n\n\t\tif (parsed.type === 'issue' && parsed.number !== undefined) {\n\t\t\tworktree = await this.gitWorktreeManager.findWorktreeForIssue(parsed.number)\n\t\t} else if (parsed.type === 'pr' && parsed.number !== undefined) {\n\t\t\t// For PRs, ensure the number is numeric (PRs are always numeric per GitHub)\n\t\t\tconst prNumber = typeof parsed.number === 'number' ? parsed.number : Number(parsed.number)\n\t\t\tif (isNaN(prNumber) || !isFinite(prNumber)) {\n\t\t\t\tthrow new Error(`Invalid PR number: ${parsed.number}. PR numbers must be numeric.`)\n\t\t\t}\n\t\t\t// Pass empty string for branch name since we don't know it yet\n\t\t\tworktree = await this.gitWorktreeManager.findWorktreeForPR(prNumber, '')\n\t\t} else if (parsed.type === 'branch' && parsed.branchName) {\n\t\t\tworktree = await this.gitWorktreeManager.findWorktreeForBranch(\n\t\t\t\tparsed.branchName\n\t\t\t)\n\t\t}\n\n\t\tif (!worktree) {\n\t\t\tthrow new Error(\n\t\t\t\t`No worktree found for ${this.formatParsedInput(parsed)}. ` +\n\t\t\t\t\t`Run 'il start ${parsed.originalInput}' to create one.`\n\t\t\t)\n\t\t}\n\n\t\treturn worktree\n\t}\n\n\t/**\n\t * Format parsed input for display\n\t */\n\tprivate formatParsedInput(parsed: ParsedOpenInput): string {\n\t\tconst autoLabel = parsed.autoDetected ? ' (auto-detected)' : ''\n\n\t\tif (parsed.type === 'issue') {\n\t\t\treturn `issue #${parsed.number}${autoLabel}`\n\t\t}\n\t\tif (parsed.type === 'pr') {\n\t\t\treturn `PR #${parsed.number}${autoLabel}`\n\t\t}\n\t\treturn `branch \"${parsed.branchName}\"${autoLabel}`\n\t}\n\n\t/**\n\t * Open web browser with workspace URL\n\t * Auto-starts dev server if not already running\n\t */\n\tprivate async openWebBrowser(worktree: GitWorktree): Promise<void> {\n\t\tconst cliOverrides = extractSettingsOverrides()\n\t\tconst settings = await this.settingsManager.loadSettings(undefined, cliOverrides)\n\t\tconst port = await getWorkspacePort({\n\t\t\tworktreePath: worktree.path,\n\t\t\tworktreeBranch: worktree.branch,\n\t\t\tbasePort: settings.capabilities?.web?.basePort,\n\t\t\tcheckEnvFile: true,\n\t\t})\n\n\t\t// Ensure dev server is running on the port\n\t\tconst serverReady = await this.devServerManager.ensureServerRunning(\n\t\t\tworktree.path,\n\t\t\tport\n\t\t)\n\n\t\tif (!serverReady) {\n\t\t\tlogger.warn(\n\t\t\t\t`Dev server failed to start on port ${port}. Opening browser anyway...`\n\t\t\t)\n\t\t}\n\n\t\t// Construct URL and open browser\n\t\tconst url = `http://localhost:${port}`\n\t\tlogger.info(`Opening browser: ${url}`)\n\t\tawait openBrowser(url)\n\t\tlogger.success('Browser opened')\n\t}\n\n\t/**\n\t * Run CLI tool directly from worktree bin path (NO SYMLINKS!)\n\t */\n\tprivate async runCLITool(\n\t\tworktreePath: string,\n\t\tbinEntries: Record<string, string>,\n\t\targs: string[]\n\t): Promise<void> {\n\t\t// Validate binEntries exist\n\t\tif (Object.keys(binEntries).length === 0) {\n\t\t\tthrow new Error('No bin entries found in package.json')\n\t\t}\n\n\t\t// Get first bin entry (deterministic)\n\t\tconst firstEntry = Object.entries(binEntries)[0]\n\t\tif (!firstEntry) {\n\t\t\tthrow new Error('No bin entries found in package.json')\n\t\t}\n\t\tconst [binName, binPath] = firstEntry\n\t\tlogger.debug(`Using bin entry: ${binName} -> ${binPath}`)\n\n\t\t// CRITICAL: Construct absolute path (NO SYMLINKS!)\n\t\tconst binFilePath = path.resolve(worktreePath, binPath)\n\t\tlogger.debug(`Resolved bin file path: ${binFilePath}`)\n\n\t\t// Verify file exists\n\t\tif (!(await fs.pathExists(binFilePath))) {\n\t\t\tthrow new Error(\n\t\t\t\t`CLI executable not found: ${binFilePath}\\n` +\n\t\t\t\t\t`Make sure the project is built (run 'il start' first)`\n\t\t\t)\n\t\t}\n\n\t\t// Execute with Node.js\n\t\tlogger.info(`Running CLI: node ${binFilePath} ${args.join(' ')}`)\n\t\tawait execa('node', [binFilePath, ...args], {\n\t\t\tstdio: 'inherit', // Allow interactive CLIs (prompts, colors, etc.)\n\t\t\tcwd: worktreePath, // Execute in worktree context\n\t\t\tenv: process.env, // Inherit environment\n\t\t})\n\t}\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,SAAS,aAAa;AA8Bf,IAAM,cAAN,MAAkB;AAAA,EACxB,YACS,qBAAqB,IAAI,mBAAmB,GAC5C,qBAAqB,IAAI,0BAA0B,GACnD,mBAAmB,IAAI,iBAAiB,IAAI,mBAAmB,CAAC,GAChE,mBAAmB,IAAI,iBAAiB,GACxC,kBAAkB,IAAI,gBAAgB,GAC7C;AALO;AACA;AACA;AACA;AACA;AAAA,EACN;AAAA,EAEH,MAAM,QAAQ,OAAwC;AAErD,UAAM,SAAS,MAAM,aAClB,MAAM,KAAK,mBAAmB,MAAM,UAAU,IAC9C,MAAM,KAAK,+BAA+B;AAE7C,WAAO,MAAM,iBAAiB,KAAK,UAAU,MAAM,CAAC,EAAE;AAGtD,UAAM,WAAW,MAAM,KAAK,0BAA0B,MAAM;AAE5D,WAAO,KAAK,sBAAsB,SAAS,IAAI,EAAE;AAGjD,UAAM,EAAE,cAAc,WAAW,IAChC,MAAM,KAAK,mBAAmB,mBAAmB,SAAS,IAAI;AAE/D,WAAO,MAAM,0BAA0B,aAAa,KAAK,IAAI,CAAC,EAAE;AAGhE,QAAI,aAAa,SAAS,KAAK,GAAG;AACjC,YAAM,KAAK,eAAe,QAAQ;AAAA,IACnC,WAAW,aAAa,SAAS,KAAK,GAAG;AACxC,YAAM,KAAK,WAAW,SAAS,MAAM,YAAY,MAAM,QAAQ,CAAC,CAAC;AAAA,IAClE,OAAO;AACN,YAAM,IAAI;AAAA,QACT,wDAAwD,SAAS,IAAI;AAAA,MACtE;AAAA,IACD;AAAA,EACD;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,mBAAmB,YAA8C;AAC9E,UAAM,SAAS,MAAM,KAAK,iBAAiB,yBAAyB,UAAU;AAG9E,QAAI,OAAO,SAAS,eAAe;AAClC,YAAM,IAAI,MAAM,yDAAyD;AAAA,IAC1E;AAEA,UAAM,SAA0B;AAAA,MAC/B,MAAM,OAAO;AAAA,MACb,eAAe,OAAO;AAAA,MACtB,cAAc;AAAA,IACf;AAEA,QAAI,OAAO,WAAW,QAAW;AAChC,aAAO,SAAS,OAAO;AAAA,IACxB;AACA,QAAI,OAAO,eAAe,QAAW;AACpC,aAAO,aAAa,OAAO;AAAA,IAC5B;AAEA,WAAO;AAAA,EACR;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,iCAA2D;AACxE,UAAM,aAAa,KAAK,SAAS,QAAQ,IAAI,CAAC;AAG9C,UAAM,YAAY;AAClB,UAAM,UAAU,WAAW,MAAM,SAAS;AAE1C,QAAI,mCAAU,IAAI;AACjB,YAAM,WAAW,SAAS,QAAQ,CAAC,GAAG,EAAE;AACxC,aAAO,MAAM,qBAAqB,QAAQ,oBAAoB,UAAU,EAAE;AAC1E,aAAO;AAAA,QACN,MAAM;AAAA,QACN,QAAQ;AAAA,QACR,eAAe;AAAA,QACf,cAAc;AAAA,MACf;AAAA,IACD;AAGA,UAAM,cAAc,mBAAmB,UAAU;AAEjD,QAAI,gBAAgB,MAAM;AACzB,aAAO,MAAM,wBAAwB,WAAW,oBAAoB,UAAU,EAAE;AAChF,aAAO;AAAA,QACN,MAAM;AAAA,QACN,QAAQ;AAAA,QACR,eAAe;AAAA,QACf,cAAc;AAAA,MACf;AAAA,IACD;AAGA,UAAM,WAAW,MAAM,KAAK,mBAAmB,YAAY;AAC3D,UAAM,gBAAgB,SAAS;AAE/B,QAAI,CAAC,eAAe;AACnB,YAAM,IAAI;AAAA,QACT;AAAA,MAED;AAAA,IACD;AAGA,UAAM,oBAAoB,mBAAmB,aAAa;AAC1D,QAAI,sBAAsB,MAAM;AAC/B,aAAO,MAAM,wBAAwB,iBAAiB,iBAAiB,aAAa,EAAE;AACtF,aAAO;AAAA,QACN,MAAM;AAAA,QACN,QAAQ;AAAA,QACR,eAAe;AAAA,QACf,cAAc;AAAA,MACf;AAAA,IACD;AAGA,WAAO;AAAA,MACN,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,eAAe;AAAA,MACf,cAAc;AAAA,IACf;AAAA,EACD;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,0BAA0B,QAA+C;AACtF,QAAI,WAA+B;AAEnC,QAAI,OAAO,SAAS,WAAW,OAAO,WAAW,QAAW;AAC3D,iBAAW,MAAM,KAAK,mBAAmB,qBAAqB,OAAO,MAAM;AAAA,IAC5E,WAAW,OAAO,SAAS,QAAQ,OAAO,WAAW,QAAW;AAE/D,YAAM,WAAW,OAAO,OAAO,WAAW,WAAW,OAAO,SAAS,OAAO,OAAO,MAAM;AACzF,UAAI,MAAM,QAAQ,KAAK,CAAC,SAAS,QAAQ,GAAG;AAC3C,cAAM,IAAI,MAAM,sBAAsB,OAAO,MAAM,+BAA+B;AAAA,MACnF;AAEA,iBAAW,MAAM,KAAK,mBAAmB,kBAAkB,UAAU,EAAE;AAAA,IACxE,WAAW,OAAO,SAAS,YAAY,OAAO,YAAY;AACzD,iBAAW,MAAM,KAAK,mBAAmB;AAAA,QACxC,OAAO;AAAA,MACR;AAAA,IACD;AAEA,QAAI,CAAC,UAAU;AACd,YAAM,IAAI;AAAA,QACT,yBAAyB,KAAK,kBAAkB,MAAM,CAAC,mBACrC,OAAO,aAAa;AAAA,MACvC;AAAA,IACD;AAEA,WAAO;AAAA,EACR;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,QAAiC;AAC1D,UAAM,YAAY,OAAO,eAAe,qBAAqB;AAE7D,QAAI,OAAO,SAAS,SAAS;AAC5B,aAAO,UAAU,OAAO,MAAM,GAAG,SAAS;AAAA,IAC3C;AACA,QAAI,OAAO,SAAS,MAAM;AACzB,aAAO,OAAO,OAAO,MAAM,GAAG,SAAS;AAAA,IACxC;AACA,WAAO,WAAW,OAAO,UAAU,IAAI,SAAS;AAAA,EACjD;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,eAAe,UAAsC;AAzNpE;AA0NE,UAAM,eAAe,yBAAyB;AAC9C,UAAM,WAAW,MAAM,KAAK,gBAAgB,aAAa,QAAW,YAAY;AAChF,UAAM,OAAO,MAAM,iBAAiB;AAAA,MACnC,cAAc,SAAS;AAAA,MACvB,gBAAgB,SAAS;AAAA,MACzB,WAAU,oBAAS,iBAAT,mBAAuB,QAAvB,mBAA4B;AAAA,MACtC,cAAc;AAAA,IACf,CAAC;AAGD,UAAM,cAAc,MAAM,KAAK,iBAAiB;AAAA,MAC/C,SAAS;AAAA,MACT;AAAA,IACD;AAEA,QAAI,CAAC,aAAa;AACjB,aAAO;AAAA,QACN,sCAAsC,IAAI;AAAA,MAC3C;AAAA,IACD;AAGA,UAAM,MAAM,oBAAoB,IAAI;AACpC,WAAO,KAAK,oBAAoB,GAAG,EAAE;AACrC,UAAM,YAAY,GAAG;AACrB,WAAO,QAAQ,gBAAgB;AAAA,EAChC;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,WACb,cACA,YACA,MACgB;AAEhB,QAAI,OAAO,KAAK,UAAU,EAAE,WAAW,GAAG;AACzC,YAAM,IAAI,MAAM,sCAAsC;AAAA,IACvD;AAGA,UAAM,aAAa,OAAO,QAAQ,UAAU,EAAE,CAAC;AAC/C,QAAI,CAAC,YAAY;AAChB,YAAM,IAAI,MAAM,sCAAsC;AAAA,IACvD;AACA,UAAM,CAAC,SAAS,OAAO,IAAI;AAC3B,WAAO,MAAM,oBAAoB,OAAO,OAAO,OAAO,EAAE;AAGxD,UAAM,cAAc,KAAK,QAAQ,cAAc,OAAO;AACtD,WAAO,MAAM,2BAA2B,WAAW,EAAE;AAGrD,QAAI,CAAE,MAAM,GAAG,WAAW,WAAW,GAAI;AACxC,YAAM,IAAI;AAAA,QACT,6BAA6B,WAAW;AAAA;AAAA,MAEzC;AAAA,IACD;AAGA,WAAO,KAAK,qBAAqB,WAAW,IAAI,KAAK,KAAK,GAAG,CAAC,EAAE;AAChE,UAAM,MAAM,QAAQ,CAAC,aAAa,GAAG,IAAI,GAAG;AAAA,MAC3C,OAAO;AAAA;AAAA,MACP,KAAK;AAAA;AAAA,MACL,KAAK,QAAQ;AAAA;AAAA,IACd,CAAC;AAAA,EACF;AACD;","names":[]}
|
|
@@ -2,43 +2,47 @@
|
|
|
2
2
|
import {
|
|
3
3
|
launchFirstRunSetup,
|
|
4
4
|
needsFirstRunSetup
|
|
5
|
-
} from "./chunk-
|
|
5
|
+
} from "./chunk-UVD4CZKS.js";
|
|
6
6
|
import {
|
|
7
|
-
|
|
8
|
-
} from "./chunk-
|
|
7
|
+
IssueManagementProviderFactory
|
|
8
|
+
} from "./chunk-JO2LZ6EQ.js";
|
|
9
|
+
import "./chunk-4232AHNQ.js";
|
|
9
10
|
import {
|
|
10
11
|
matchIssueIdentifier
|
|
11
|
-
} from "./chunk-
|
|
12
|
+
} from "./chunk-WY4QBK43.js";
|
|
13
|
+
import {
|
|
14
|
+
TelemetryService
|
|
15
|
+
} from "./chunk-RSYT7MVI.js";
|
|
12
16
|
import "./chunk-Q7POFB5Q.js";
|
|
13
17
|
import {
|
|
14
18
|
generateIssueManagementMcpConfig
|
|
15
|
-
} from "./chunk-
|
|
19
|
+
} from "./chunk-UWGVCXRF.js";
|
|
16
20
|
import {
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
21
|
+
detectClaudeCli,
|
|
22
|
+
launchClaude
|
|
23
|
+
} from "./chunk-ONQYPICO.js";
|
|
20
24
|
import {
|
|
21
25
|
PromptTemplateManager
|
|
22
|
-
} from "./chunk-
|
|
23
|
-
import "./chunk-
|
|
26
|
+
} from "./chunk-4WJNIR5O.js";
|
|
27
|
+
import "./chunk-MNHZB4Z2.js";
|
|
24
28
|
import {
|
|
25
29
|
PlanCommandSettingsSchema,
|
|
26
30
|
SettingsManager
|
|
27
|
-
} from "./chunk-
|
|
28
|
-
import "./chunk-
|
|
29
|
-
import "./chunk-OFDN5NKS.js";
|
|
30
|
-
import "./chunk-4CO6KG5S.js";
|
|
31
|
+
} from "./chunk-YYAKPQBT.js";
|
|
32
|
+
import "./chunk-KB64WNBZ.js";
|
|
31
33
|
import {
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
import
|
|
36
|
-
|
|
37
|
-
launchClaude
|
|
38
|
-
} from "./chunk-FO5GGFOV.js";
|
|
34
|
+
IssueTrackerFactory
|
|
35
|
+
} from "./chunk-UKBAJ2QQ.js";
|
|
36
|
+
import "./chunk-HEXKPKCK.js";
|
|
37
|
+
import "./chunk-KXDRI47U.js";
|
|
38
|
+
import "./chunk-VG45TUYK.js";
|
|
39
39
|
import {
|
|
40
40
|
withLogger
|
|
41
41
|
} from "./chunk-6MLEBAYZ.js";
|
|
42
|
+
import {
|
|
43
|
+
isInteractiveEnvironment,
|
|
44
|
+
promptConfirmation
|
|
45
|
+
} from "./chunk-7JDMYTFZ.js";
|
|
42
46
|
import {
|
|
43
47
|
createStderrLogger,
|
|
44
48
|
logger
|
|
@@ -152,7 +156,7 @@ var PlanCommand = class {
|
|
|
152
156
|
};
|
|
153
157
|
logger.info(chalk.dim(`Preparing to create a detailed plan for issue #${decompositionContext.identifier}: ${decompositionContext.title}`));
|
|
154
158
|
try {
|
|
155
|
-
const mcpProvider = IssueManagementProviderFactory.create(provider);
|
|
159
|
+
const mcpProvider = IssueManagementProviderFactory.create(provider, settings ?? void 0);
|
|
156
160
|
logger.debug("Fetching child issues for decomposition context", { identifier: decompositionContext.identifier });
|
|
157
161
|
const children = await mcpProvider.getChildIssues({ number: decompositionContext.identifier });
|
|
158
162
|
if (children.length > 0) {
|
|
@@ -205,7 +209,7 @@ var PlanCommand = class {
|
|
|
205
209
|
);
|
|
206
210
|
if (shouldRunInit) {
|
|
207
211
|
logger.info(chalk.bold("Launching iloom init..."));
|
|
208
|
-
const { InitCommand } = await import("./init-
|
|
212
|
+
const { InitCommand } = await import("./init-GFQ5W7GK.js");
|
|
209
213
|
const initCommand = new InitCommand();
|
|
210
214
|
await initCommand.execute(
|
|
211
215
|
"Help the user set up a GitHub repository or Linear project for this project so they can use issue management features. When complete tell the user they can exit to continue the planning session."
|
|
@@ -357,7 +361,7 @@ var PlanCommand = class {
|
|
|
357
361
|
throw new Error('--yolo requires a prompt or issue identifier (e.g., il plan --yolo "add gitlab support" or il plan --yolo 42)');
|
|
358
362
|
}
|
|
359
363
|
logger.warn(
|
|
360
|
-
"
|
|
364
|
+
"YOLO mode enabled - Claude will skip permission prompts and proceed autonomously. This could destroy important data or make irreversible changes. Proceeding means you accept this risk."
|
|
361
365
|
);
|
|
362
366
|
}
|
|
363
367
|
logger.debug("Launching Claude with options", {
|
|
@@ -387,6 +391,18 @@ ${initialMessage}`;
|
|
|
387
391
|
...claudeOptions,
|
|
388
392
|
...effectiveYolo && { permissionMode: "bypassPermissions" }
|
|
389
393
|
});
|
|
394
|
+
if (decompositionContext) {
|
|
395
|
+
try {
|
|
396
|
+
const mcpProv = IssueManagementProviderFactory.create(provider, settings ?? void 0);
|
|
397
|
+
const children = await mcpProv.getChildIssues({ number: decompositionContext.identifier });
|
|
398
|
+
TelemetryService.getInstance().track("epic.planned", {
|
|
399
|
+
child_count: children.length,
|
|
400
|
+
tracker: provider
|
|
401
|
+
});
|
|
402
|
+
} catch (error) {
|
|
403
|
+
logger.debug(`Telemetry epic.planned tracking failed: ${error instanceof Error ? error.message : error}`);
|
|
404
|
+
}
|
|
405
|
+
}
|
|
390
406
|
if (printOptions == null ? void 0 : printOptions.json) {
|
|
391
407
|
console.log(JSON.stringify({
|
|
392
408
|
success: true,
|
|
@@ -400,4 +416,4 @@ ${initialMessage}`;
|
|
|
400
416
|
export {
|
|
401
417
|
PlanCommand
|
|
402
418
|
};
|
|
403
|
-
//# sourceMappingURL=plan-
|
|
419
|
+
//# sourceMappingURL=plan-U7ZQWLFY.js.map
|