@goondocks/myco 0.3.7 → 0.4.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (121) hide show
  1. package/.claude-plugin/marketplace.json +1 -1
  2. package/.claude-plugin/plugin.json +1 -1
  3. package/README.md +9 -4
  4. package/commands/init.md +63 -39
  5. package/commands/setup-llm.md +69 -44
  6. package/commands/status.md +28 -10
  7. package/dist/{chunk-YFG2O5HR.js → chunk-2GJFTIWX.js} +2 -2
  8. package/dist/chunk-4FCFRJIQ.js +147 -0
  9. package/dist/chunk-4FCFRJIQ.js.map +1 -0
  10. package/dist/{chunk-PA3VMINE.js → chunk-AK6GNLPV.js} +6 -1
  11. package/dist/chunk-AK6GNLPV.js.map +1 -0
  12. package/dist/{chunk-JKOALBZC.js → chunk-BNIYWCST.js} +2 -2
  13. package/dist/{chunk-ISCT2SI6.js → chunk-G6ZMTQMJ.js} +7357 -60
  14. package/dist/chunk-G6ZMTQMJ.js.map +1 -0
  15. package/dist/{chunk-7WNE22W7.js → chunk-IVS5MYBL.js} +3 -3
  16. package/dist/{chunk-7WNE22W7.js.map → chunk-IVS5MYBL.js.map} +1 -1
  17. package/dist/{chunk-7VPJK56U.js → chunk-JBD5KP5G.js} +31 -16
  18. package/dist/chunk-JBD5KP5G.js.map +1 -0
  19. package/dist/chunk-NUA7UTIY.js +37 -0
  20. package/dist/chunk-NUA7UTIY.js.map +1 -0
  21. package/dist/{chunk-NYAWCMRZ.js → chunk-OUFSLZTX.js} +4 -4
  22. package/dist/chunk-P7RNAYU7.js +242 -0
  23. package/dist/chunk-P7RNAYU7.js.map +1 -0
  24. package/dist/chunk-QQ36XEJP.js +38 -0
  25. package/dist/chunk-QQ36XEJP.js.map +1 -0
  26. package/dist/chunk-RDXTQ436.js +49 -0
  27. package/dist/chunk-RDXTQ436.js.map +1 -0
  28. package/dist/{chunk-AWF3M57N.js → chunk-S7EIHYE7.js} +8 -8
  29. package/dist/{chunk-AWF3M57N.js.map → chunk-S7EIHYE7.js.map} +1 -1
  30. package/dist/{chunk-QWU7QLZI.js → chunk-TZDDXRHG.js} +10 -10
  31. package/dist/chunk-TZDDXRHG.js.map +1 -0
  32. package/dist/chunk-VYV5IFD6.js +99 -0
  33. package/dist/chunk-VYV5IFD6.js.map +1 -0
  34. package/dist/{chunk-LR7RQCOB.js → chunk-XCPQHC4X.js} +2 -2
  35. package/dist/{chunk-CCIV47S4.js → chunk-XHWIIU5D.js} +8 -9
  36. package/dist/chunk-XHWIIU5D.js.map +1 -0
  37. package/dist/{chunk-FFQNE6CT.js → chunk-YZO22BBI.js} +45 -31
  38. package/dist/chunk-YZO22BBI.js.map +1 -0
  39. package/dist/{chunk-ZBNT6E22.js → chunk-ZCBL5HER.js} +2 -2
  40. package/dist/{cli-3WQSDSW6.js → cli-ZN6VBA7V.js} +23 -17
  41. package/dist/cli-ZN6VBA7V.js.map +1 -0
  42. package/dist/{client-5T4M42UQ.js → client-5SUO2UYH.js} +5 -5
  43. package/dist/{config-MD4XMLUS.js → config-4GGMWGAF.js} +4 -4
  44. package/dist/{detect-providers-LNOLBICR.js → detect-providers-5FU3BN5Q.js} +3 -3
  45. package/dist/{init-RALMQKOQ.js → init-7UXGDOFS.js} +51 -60
  46. package/dist/init-7UXGDOFS.js.map +1 -0
  47. package/dist/{main-S3WSUF5T.js → main-6UPAIDGS.js} +648 -228
  48. package/dist/main-6UPAIDGS.js.map +1 -0
  49. package/dist/{rebuild-JW6BCHHZ.js → rebuild-QDSYYCS7.js} +10 -10
  50. package/dist/rebuild-QDSYYCS7.js.map +1 -0
  51. package/dist/{reprocess-SNXFNKBN.js → reprocess-ZNUQCIS3.js} +18 -18
  52. package/dist/reprocess-ZNUQCIS3.js.map +1 -0
  53. package/dist/{restart-YE2IGOYT.js → restart-5UY2KV54.js} +6 -6
  54. package/dist/{search-2HMG3ON7.js → search-2VEN3XIG.js} +9 -9
  55. package/dist/{server-JM3TM7D2.js → server-OR5B4B7K.js} +77 -54
  56. package/dist/{server-JM3TM7D2.js.map → server-OR5B4B7K.js.map} +1 -1
  57. package/dist/{session-5GI2YU6R.js → session-QF6MILAC.js} +2 -2
  58. package/dist/{session-start-2UEEEO52.js → session-start-TUITIUMB.js} +29 -28
  59. package/dist/session-start-TUITIUMB.js.map +1 -0
  60. package/dist/setup-digest-ETCZAUIU.js +15 -0
  61. package/dist/setup-llm-DWEJE3JE.js +15 -0
  62. package/dist/setup-llm-DWEJE3JE.js.map +1 -0
  63. package/dist/src/cli.js +4 -4
  64. package/dist/src/daemon/main.js +4 -4
  65. package/dist/src/hooks/post-tool-use.js +5 -5
  66. package/dist/src/hooks/session-end.js +5 -5
  67. package/dist/src/hooks/session-start.js +4 -4
  68. package/dist/src/hooks/stop.js +7 -7
  69. package/dist/src/hooks/user-prompt-submit.js +5 -5
  70. package/dist/src/hooks/user-prompt-submit.js.map +1 -1
  71. package/dist/src/mcp/server.js +4 -4
  72. package/dist/src/prompts/classification.md +1 -0
  73. package/dist/src/prompts/digest-10000.md +74 -0
  74. package/dist/src/prompts/digest-1500.md +25 -0
  75. package/dist/src/prompts/digest-3000.md +32 -0
  76. package/dist/src/prompts/digest-5000.md +43 -0
  77. package/dist/src/prompts/digest-system.md +32 -0
  78. package/dist/src/prompts/extraction.md +11 -10
  79. package/dist/src/prompts/summary.md +11 -1
  80. package/dist/src/prompts/title.md +1 -1
  81. package/dist/{stats-IOWXG576.js → stats-IVIXIKTS.js} +12 -12
  82. package/dist/stats-IVIXIKTS.js.map +1 -0
  83. package/dist/{verify-7MWOV72E.js → verify-4H6CEE5T.js} +6 -6
  84. package/dist/{version-S7MHLD5P.js → version-5B2TWXQJ.js} +4 -4
  85. package/dist/version-5B2TWXQJ.js.map +1 -0
  86. package/package.json +1 -1
  87. package/skills/myco/SKILL.md +20 -20
  88. package/skills/myco/references/wisdom.md +14 -14
  89. package/skills/rules/SKILL.md +4 -4
  90. package/dist/chunk-7VPJK56U.js.map +0 -1
  91. package/dist/chunk-BA23DROX.js +0 -160
  92. package/dist/chunk-BA23DROX.js.map +0 -1
  93. package/dist/chunk-CCIV47S4.js.map +0 -1
  94. package/dist/chunk-EF4JVH24.js +0 -7299
  95. package/dist/chunk-EF4JVH24.js.map +0 -1
  96. package/dist/chunk-FFQNE6CT.js.map +0 -1
  97. package/dist/chunk-ISCT2SI6.js.map +0 -1
  98. package/dist/chunk-PA3VMINE.js.map +0 -1
  99. package/dist/chunk-QWU7QLZI.js.map +0 -1
  100. package/dist/chunk-YMYJ7FNH.js +0 -19
  101. package/dist/chunk-YMYJ7FNH.js.map +0 -1
  102. package/dist/cli-3WQSDSW6.js.map +0 -1
  103. package/dist/init-RALMQKOQ.js.map +0 -1
  104. package/dist/main-S3WSUF5T.js.map +0 -1
  105. package/dist/rebuild-JW6BCHHZ.js.map +0 -1
  106. package/dist/reprocess-SNXFNKBN.js.map +0 -1
  107. package/dist/session-start-2UEEEO52.js.map +0 -1
  108. package/dist/stats-IOWXG576.js.map +0 -1
  109. /package/dist/{chunk-YFG2O5HR.js.map → chunk-2GJFTIWX.js.map} +0 -0
  110. /package/dist/{chunk-JKOALBZC.js.map → chunk-BNIYWCST.js.map} +0 -0
  111. /package/dist/{chunk-NYAWCMRZ.js.map → chunk-OUFSLZTX.js.map} +0 -0
  112. /package/dist/{chunk-LR7RQCOB.js.map → chunk-XCPQHC4X.js.map} +0 -0
  113. /package/dist/{chunk-ZBNT6E22.js.map → chunk-ZCBL5HER.js.map} +0 -0
  114. /package/dist/{client-5T4M42UQ.js.map → client-5SUO2UYH.js.map} +0 -0
  115. /package/dist/{config-MD4XMLUS.js.map → config-4GGMWGAF.js.map} +0 -0
  116. /package/dist/{detect-providers-LNOLBICR.js.map → detect-providers-5FU3BN5Q.js.map} +0 -0
  117. /package/dist/{restart-YE2IGOYT.js.map → restart-5UY2KV54.js.map} +0 -0
  118. /package/dist/{search-2HMG3ON7.js.map → search-2VEN3XIG.js.map} +0 -0
  119. /package/dist/{session-5GI2YU6R.js.map → session-QF6MILAC.js.map} +0 -0
  120. /package/dist/{version-S7MHLD5P.js.map → setup-digest-ETCZAUIU.js.map} +0 -0
  121. /package/dist/{verify-7MWOV72E.js.map → verify-4H6CEE5T.js.map} +0 -0
@@ -1,7 +1,7 @@
1
1
  import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
2
2
  import {
3
3
  require_dist
4
- } from "./chunk-EF4JVH24.js";
4
+ } from "./chunk-G6ZMTQMJ.js";
5
5
  import {
6
6
  __toESM
7
7
  } from "./chunk-PZUWP5VK.js";
@@ -89,7 +89,7 @@ function formatSessionBody(input) {
89
89
  sections.push(fields.join("\n"));
90
90
  if (input.relatedMemories?.length) {
91
91
  const links = input.relatedMemories.map((m) => `- ${wikilink(m.id, m.title)}`);
92
- sections.push(`## Related Memories
92
+ sections.push(`## Related Spores
93
93
  ${links.join("\n")}`);
94
94
  }
95
95
  if (input.turns.length > 0) {
@@ -124,7 +124,7 @@ ${turnLines.join("\n\n")}`);
124
124
  sections.push(footerTags(allTags));
125
125
  return sections.join("\n\n");
126
126
  }
127
- function formatMemoryBody(input) {
127
+ function formatSporeBody(input) {
128
128
  const sections = [];
129
129
  const calloutType = observationCalloutType(input.observationType);
130
130
  const calloutTitle = capitalize(tagNormalize(input.observationType));
@@ -148,7 +148,7 @@ ${input.alternatives_rejected}`);
148
148
  ${input.gained}`);
149
149
  if (input.sacrificed) sections.push(`## Sacrificed
150
150
  ${input.sacrificed}`);
151
- const allTags = buildTags("memory", input.observationType, input.tags ?? []);
151
+ const allTags = buildTags("spore", input.observationType, input.tags ?? []);
152
152
  sections.push(footerTags(allTags));
153
153
  return sections.join("\n\n");
154
154
  }
@@ -281,9 +281,9 @@ var VaultWriter = class {
281
281
  this.writeMarkdown(relativePath, frontmatter, body);
282
282
  return relativePath;
283
283
  }
284
- writeMemory(input) {
284
+ writeSpore(input) {
285
285
  const normalizedType = input.observation_type.replace(/_/g, "-");
286
- const relativePath = `memories/${normalizedType}/${input.id}.md`;
286
+ const relativePath = `spores/${normalizedType}/${input.id}.md`;
287
287
  const fullPath = path.join(this.vaultDir, relativePath);
288
288
  const now = (/* @__PURE__ */ new Date()).toISOString();
289
289
  let created = now;
@@ -297,14 +297,14 @@ var VaultWriter = class {
297
297
  } catch {
298
298
  }
299
299
  const frontmatter = {
300
- type: "memory",
300
+ type: "spore",
301
301
  id: input.id,
302
302
  observation_type: input.observation_type,
303
303
  created
304
304
  };
305
305
  if (input.session) frontmatter.session = input.session;
306
306
  if (input.plan) frontmatter.plan = input.plan;
307
- frontmatter.tags = buildTags("memory", input.observation_type, input.tags ?? []);
307
+ frontmatter.tags = buildTags("spore", input.observation_type, input.tags ?? []);
308
308
  this.writeMarkdown(relativePath, frontmatter, input.content);
309
309
  return relativePath;
310
310
  }
@@ -419,7 +419,7 @@ export {
419
419
  sessionWikilink,
420
420
  sessionRelativePath,
421
421
  formatSessionBody,
422
- formatMemoryBody,
422
+ formatSporeBody,
423
423
  VaultWriter
424
424
  };
425
- //# sourceMappingURL=chunk-QWU7QLZI.js.map
425
+ //# sourceMappingURL=chunk-TZDDXRHG.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/vault/session-id.ts","../src/obsidian/formatter.ts","../src/vault/writer.ts"],"sourcesContent":["/**\n * Session ID convention utilities.\n *\n * The runtime session ID is a bare UUID (e.g., \"abc123\").\n * The vault stores it with a \"session-\" prefix (e.g., \"session-abc123\").\n * These helpers convert between the two forms.\n */\nconst SESSION_PREFIX = 'session-';\n\n/** Convert a bare session ID to a vault note ID: \"abc123\" → \"session-abc123\" */\nexport function sessionNoteId(bareId: string): string {\n if (bareId.startsWith(SESSION_PREFIX)) return bareId;\n return `${SESSION_PREFIX}${bareId}`;\n}\n\n/** Convert a vault note ID to a bare session ID: \"session-abc123\" → \"abc123\" */\nexport function bareSessionId(noteId: string): string {\n if (noteId.startsWith(SESSION_PREFIX)) return noteId.slice(SESSION_PREFIX.length);\n return noteId;\n}\n\n/** Convert a bare session ID to an Obsidian wikilink: \"abc123\" → \"[[session-abc123]]\" */\nexport function sessionWikilink(bareId: string): string {\n return `[[${sessionNoteId(bareId)}]]`;\n}\n\n/** Build the relative vault path for a session note */\nexport function sessionRelativePath(bareId: string, date: string): string {\n return `sessions/${date}/${sessionNoteId(bareId)}.md`;\n}\n\n","/**\n * Pure formatting functions for Obsidian-native vault notes.\n * No I/O, no external dependencies — just string transforms.\n */\nimport type { ArtifactType } from '../vault/types.js';\nimport { sessionNoteId } from '../vault/session-id.js';\n\n// Callout type mapping for observation types\nconst CALLOUT_MAP: Record<string, string> = {\n gotcha: 'warning',\n bug_fix: 'bug',\n decision: 'info',\n discovery: 'tip',\n trade_off: 'question',\n};\n\nexport function observationCalloutType(observationType: string): string {\n return CALLOUT_MAP[observationType] ?? 'note';\n}\n\n/**\n * Escape angle brackets that Obsidian would interpret as HTML tags.\n * Matches `<` followed by a letter, `/`, or `!` (opening/closing tags, comments).\n */\nexport function escapeHtmlTags(text: string): string {\n return text.replace(/<(?=[a-zA-Z/!])/g, '\\\\<');\n}\n\nexport function callout(type: string, title: string, content: string): string {\n const safe = escapeHtmlTags(content);\n const indented = safe.split('\\n').map((line) => `> ${line}`).join('\\n');\n return `> [!${type}] ${title}\\n${indented}`;\n}\n\nexport function inlineField(key: string, value: string): string {\n return `${key}:: ${value}`;\n}\n\nexport function wikilink(target: string, display?: string): string {\n return display ? `[[${target}|${display}]]` : `[[${target}]]`;\n}\n\n/**\n * Normalize an observation_type to a tag-safe form.\n * Frontmatter keeps underscores; tags use hyphens per Obsidian convention.\n */\nfunction tagNormalize(s: string): string {\n return s.replace(/_/g, '-');\n}\n\n/**\n * Sanitize a user/LLM-provided tag for Obsidian compatibility.\n * Obsidian tags cannot contain spaces — replace with slash (nested tag).\n * Strips leading # if present.\n */\nfunction sanitizeTag(raw: string): string {\n const stripped = raw.startsWith('#') ? raw.slice(1) : raw;\n return stripped.replace(/\\s+/g, '/');\n}\n\nexport function buildTags(type: string, subtype: string, extraTags: string[] = []): string[] {\n const tags: string[] = [`type/${type}`];\n\n if (subtype) {\n tags.push(`${type}/${tagNormalize(subtype)}`);\n }\n\n for (const tag of extraTags) {\n const normalized = sanitizeTag(tag);\n if (normalized && !tags.includes(normalized)) {\n tags.push(normalized);\n }\n }\n\n return tags;\n}\n\nexport function footerTags(tags: string[]): string {\n return tags.map((t) => (t.startsWith('#') ? t : `#${t}`)).join(' ');\n}\n\n// --- Session formatting ---\n\nexport interface SessionBodyInput {\n title: string;\n narrative: string;\n sessionId: string;\n user?: string;\n started?: string;\n ended?: string;\n branch?: string;\n relatedMemories?: Array<{ id: string; title: string }>;\n turns: Array<{\n prompt: string;\n toolCount: number;\n aiResponse?: string;\n /** Filenames of images in the vault attachments folder */\n images?: string[];\n }>;\n tags?: string[];\n}\n\nexport function formatSessionBody(input: SessionBodyInput): string {\n const sections: string[] = [];\n\n sections.push(`# ${input.title}`);\n\n if (input.narrative) {\n sections.push(callout('abstract', 'Summary', input.narrative));\n }\n\n // Inline fields\n const fields: string[] = [];\n fields.push(inlineField('Session', wikilink(sessionNoteId(input.sessionId))));\n if (input.user) fields.push(inlineField('User', input.user));\n if (input.started && input.ended) {\n const duration = formatDuration(input.started, input.ended);\n if (duration) fields.push(inlineField('Duration', duration));\n }\n if (input.branch) fields.push(inlineField('Branch', `\\`${input.branch}\\``));\n sections.push(fields.join('\\n'));\n\n // Related spores\n if (input.relatedMemories?.length) {\n const links = input.relatedMemories.map((m) => `- ${wikilink(m.id, m.title)}`);\n sections.push(`## Related Spores\\n${links.join('\\n')}`);\n }\n\n // Conversation turns — always rebuilt from the full transcript.\n // The transcript is the source of truth for the complete conversation.\n if (input.turns.length > 0) {\n const turnLines: string[] = [];\n for (let i = 0; i < input.turns.length; i++) {\n const turn = input.turns[i];\n const turnNum = i + 1;\n turnLines.push(`### Turn ${turnNum}`);\n if (turn.prompt || turn.images?.length) {\n // Build prompt content: text + images + tool count\n const parts: string[] = [];\n if (turn.prompt) parts.push(turn.prompt);\n if (turn.images?.length) {\n parts.push(turn.images.map((f) => `![[${f}]]`).join('\\n'));\n }\n if (turn.toolCount > 0) parts.push(`*${turn.toolCount} tool calls*`);\n turnLines.push(callout('user', 'Prompt', parts.join('\\n\\n')));\n } else if (turn.toolCount > 0) {\n turnLines.push(callout('user', 'Prompt', `*${turn.toolCount} tool calls*`));\n }\n if (turn.aiResponse) {\n turnLines.push(callout('assistant', 'Response', turn.aiResponse));\n }\n }\n sections.push(`## Conversation\\n\\n${turnLines.join('\\n\\n')}`);\n }\n\n // Footer tags\n const allTags = buildTags('session', 'ended', [\n ...(input.user ? [`user/${input.user}`] : []),\n ...(input.tags ?? []),\n ]);\n sections.push(footerTags(allTags));\n\n return sections.join('\\n\\n');\n}\n\n// --- Spore formatting ---\n\nexport interface SporeBodyInput {\n title: string;\n observationType: string;\n content: string;\n sessionId?: string;\n root_cause?: string;\n fix?: string;\n rationale?: string;\n alternatives_rejected?: string;\n gained?: string;\n sacrificed?: string;\n tags?: string[];\n}\n\nexport function formatSporeBody(input: SporeBodyInput): string {\n const sections: string[] = [];\n const calloutType = observationCalloutType(input.observationType);\n const calloutTitle = capitalize(tagNormalize(input.observationType));\n\n sections.push(`# ${input.title}`);\n sections.push(callout(calloutType, calloutTitle, input.content));\n\n // Inline fields\n const fields: string[] = [];\n if (input.sessionId) {\n fields.push(inlineField('Session', wikilink(sessionNoteId(input.sessionId))));\n }\n fields.push(inlineField('Observation', input.observationType));\n if (fields.length > 0) sections.push(fields.join('\\n'));\n\n // Type-specific sub-sections\n if (input.root_cause) sections.push(`## Root Cause\\n${input.root_cause}`);\n if (input.fix) sections.push(`## Fix\\n${input.fix}`);\n if (input.rationale) sections.push(`## Rationale\\n${input.rationale}`);\n if (input.alternatives_rejected) sections.push(`## Alternatives Rejected\\n${input.alternatives_rejected}`);\n if (input.gained) sections.push(`## Gained\\n${input.gained}`);\n if (input.sacrificed) sections.push(`## Sacrificed\\n${input.sacrificed}`);\n\n // Footer tags\n const allTags = buildTags('spore', input.observationType, input.tags ?? []);\n sections.push(footerTags(allTags));\n\n return sections.join('\\n\\n');\n}\n\n// --- Plan formatting ---\n\nexport interface PlanBodyInput {\n id: string;\n status: string;\n author?: string;\n created?: string;\n sessions?: Array<{ id: string; title: string }>;\n content: string;\n tags?: string[];\n}\n\nexport function formatPlanBody(input: PlanBodyInput): string {\n const sections: string[] = [];\n\n // Inline fields block\n const fields: string[] = [];\n fields.push(inlineField('Plan', wikilink(input.id)));\n fields.push(inlineField('Status', input.status));\n if (input.author) fields.push(inlineField('Author', input.author));\n if (input.created) fields.push(inlineField('Created', input.created));\n sections.push(fields.join('\\n'));\n\n // User-provided content body (don't restructure)\n sections.push(input.content);\n\n // Sessions section\n if (input.sessions?.length) {\n const links = input.sessions.map((s) => `- ${wikilink(sessionNoteId(s.id), s.title)}`);\n sections.push(`## Sessions\\n${links.join('\\n')}`);\n }\n\n // Footer tags\n const statusTag = tagNormalize(input.status);\n const allTags = buildTags('plan', statusTag, input.tags ?? []);\n sections.push(footerTags(allTags));\n\n return sections.join('\\n\\n');\n}\n\n// --- Artifact formatting ---\n\nexport interface ArtifactBodyInput {\n id: string;\n title: string;\n artifact_type: ArtifactType;\n source_path: string;\n sessionId: string;\n content: string;\n tags?: string[];\n}\n\nexport function formatArtifactBody(input: ArtifactBodyInput): string {\n const sections: string[] = [];\n\n // Inline fields\n const fields: string[] = [];\n fields.push(inlineField('Artifact', wikilink(input.id)));\n fields.push(inlineField('Source', `\\`${input.source_path}\\``));\n fields.push(inlineField('Type', input.artifact_type));\n fields.push(inlineField('Session', wikilink(sessionNoteId(input.sessionId))));\n sections.push(fields.join('\\n'));\n\n // Body: full content from disk\n sections.push(input.content);\n\n // Footer tags\n const allTags = buildTags('artifact', input.artifact_type, input.tags ?? []);\n sections.push(footerTags(allTags));\n\n return sections.join('\\n\\n');\n}\n\n// --- Team formatting ---\n\nexport interface TeamBodyInput {\n user: string;\n role?: string;\n recentSessions?: Array<{ id: string; title: string }>;\n}\n\nexport function formatTeamBody(input: TeamBodyInput): string {\n const sections: string[] = [];\n\n sections.push(`# ${input.user}`);\n sections.push(callout('info', 'Team Member', input.role ?? 'Contributor'));\n\n // Inline fields\n const fields: string[] = [];\n fields.push(inlineField('User', input.user));\n if (input.role) fields.push(inlineField('Role', input.role));\n sections.push(fields.join('\\n'));\n\n // Recent sessions\n if (input.recentSessions?.length) {\n const links = input.recentSessions.map((s) => `- ${wikilink(sessionNoteId(s.id), s.title)}`);\n sections.push(`## Recent Sessions\\n${links.join('\\n')}`);\n }\n\n // Footer tags\n const allTags = buildTags('team', '', [`user/${input.user}`]);\n sections.push(footerTags(allTags));\n\n return sections.join('\\n\\n');\n}\n\n// --- Helpers ---\n\nfunction capitalize(s: string): string {\n return s.charAt(0).toUpperCase() + s.slice(1);\n}\n\nfunction formatDuration(started: string, ended: string): string {\n const ms = new Date(ended).getTime() - new Date(started).getTime();\n if (isNaN(ms) || ms < 0) return '';\n const totalMinutes = Math.floor(ms / 60000);\n if (totalMinutes < 1) return '<1m';\n const hours = Math.floor(totalMinutes / 60);\n const minutes = totalMinutes % 60;\n if (hours === 0) return `${minutes}m`;\n if (minutes === 0) return `${hours}h`;\n return `${hours}h ${minutes}m`;\n}\n","import fs from 'node:fs';\nimport path from 'node:path';\nimport YAML from 'yaml';\nimport { buildTags, formatTeamBody, formatPlanBody, formatArtifactBody } from '../obsidian/formatter.js';\nimport type { ArtifactType } from './types.js';\nimport { sessionNoteId, sessionRelativePath } from './session-id.js';\n\ninterface WriteSessionInput {\n id: string;\n agent?: string;\n user?: string;\n started: string;\n ended?: string;\n parent?: string;\n parent_reason?: string;\n plans?: string[];\n branch?: string;\n tags?: string[];\n tools_used?: number;\n files_changed?: number;\n summary: string;\n}\n\ninterface WritePlanInput {\n id: string;\n status?: string;\n author?: string;\n tags?: string[];\n content: string;\n}\n\ninterface WriteSporeInput {\n id: string;\n observation_type: string;\n session?: string;\n plan?: string;\n tags?: string[];\n content: string;\n}\n\ninterface WriteArtifactInput {\n id: string;\n artifact_type: ArtifactType;\n source_path: string;\n title: string;\n session: string;\n tags?: string[];\n content: string;\n}\n\ninterface WriteTeamMemberInput {\n user: string;\n role?: string;\n}\n\nexport class VaultWriter {\n constructor(private vaultDir: string) {}\n\n writeSession(input: WriteSessionInput): string {\n const date = input.started.slice(0, 10);\n const relativePath = sessionRelativePath(input.id, date);\n\n const frontmatter: Record<string, unknown> = {\n type: 'session',\n id: input.id,\n agent: input.agent ?? 'claude-code',\n user: input.user ?? '',\n started: input.started,\n };\n if (input.ended) frontmatter.ended = input.ended;\n if (input.parent) frontmatter.parent = input.parent;\n if (input.parent_reason) frontmatter.parent_reason = input.parent_reason;\n if (input.plans?.length) frontmatter.plans = input.plans;\n if (input.branch) frontmatter.branch = input.branch;\n frontmatter.tags = buildTags('session', 'ended', [\n ...(input.user ? [`user/${input.user}`] : []),\n ...(input.tags ?? []),\n ]);\n if (input.tools_used != null) frontmatter.tools_used = input.tools_used;\n if (input.files_changed != null) frontmatter.files_changed = input.files_changed;\n\n this.writeMarkdown(relativePath, frontmatter, input.summary);\n return relativePath;\n }\n\n writePlan(input: WritePlanInput): string {\n const relativePath = `plans/${input.id}.md`;\n const fullPath = path.join(this.vaultDir, relativePath);\n\n const status = input.status ?? 'active';\n let created = new Date().toISOString();\n\n // Preserve created from existing file (idempotent)\n try {\n const existing = fs.readFileSync(fullPath, 'utf-8');\n const fmMatch = existing.match(/^---\\n([\\s\\S]*?)\\n---/);\n if (fmMatch) {\n const parsed = YAML.parse(fmMatch[1]) as Record<string, unknown>;\n if (typeof parsed.created === 'string') created = parsed.created;\n }\n } catch {\n // File doesn't exist yet\n }\n const frontmatter: Record<string, unknown> = {\n type: 'plan',\n id: input.id,\n status,\n created,\n };\n if (input.author) frontmatter.author = input.author;\n frontmatter.tags = buildTags('plan', status, input.tags ?? []);\n\n const body = formatPlanBody({\n id: input.id,\n status,\n author: input.author,\n created,\n content: input.content,\n tags: input.tags,\n });\n\n this.writeMarkdown(relativePath, frontmatter, body);\n return relativePath;\n }\n\n writeSpore(input: WriteSporeInput): string {\n const normalizedType = input.observation_type.replace(/_/g, '-');\n const relativePath = `spores/${normalizedType}/${input.id}.md`;\n const fullPath = path.join(this.vaultDir, relativePath);\n const now = new Date().toISOString();\n\n // Preserve created from existing file (idempotent)\n let created = now;\n try {\n const existing = fs.readFileSync(fullPath, 'utf-8');\n const fmMatch = existing.match(/^---\\n([\\s\\S]*?)\\n---/);\n if (fmMatch) {\n const parsed = YAML.parse(fmMatch[1]) as Record<string, unknown>;\n if (typeof parsed.created === 'string') created = parsed.created;\n }\n } catch {\n // File doesn't exist yet — created = now\n }\n\n const frontmatter: Record<string, unknown> = {\n type: 'spore',\n id: input.id,\n observation_type: input.observation_type,\n created,\n };\n if (input.session) frontmatter.session = input.session;\n if (input.plan) frontmatter.plan = input.plan;\n frontmatter.tags = buildTags('spore', input.observation_type, input.tags ?? []);\n\n this.writeMarkdown(relativePath, frontmatter, input.content);\n return relativePath;\n }\n\n writeArtifact(input: WriteArtifactInput): string {\n const relativePath = `artifacts/${input.id}.md`;\n const fullPath = path.join(this.vaultDir, relativePath);\n const now = new Date().toISOString();\n\n let created = now;\n\n // Preserve created from existing file (latest-wins update)\n try {\n const existing = fs.readFileSync(fullPath, 'utf-8');\n const fmMatch = existing.match(/^---\\n([\\s\\S]*?)\\n---/);\n if (fmMatch) {\n const parsed = YAML.parse(fmMatch[1]) as Record<string, unknown>;\n if (typeof parsed.created === 'string') created = parsed.created;\n }\n } catch {\n // File doesn't exist yet — created = now\n }\n\n const frontmatter: Record<string, unknown> = {\n type: 'artifact',\n id: input.id,\n artifact_type: input.artifact_type,\n source_path: input.source_path,\n title: input.title,\n last_captured_by: sessionNoteId(input.session),\n created,\n updated: now,\n tags: buildTags('artifact', input.artifact_type, input.tags ?? []),\n };\n\n const body = formatArtifactBody({\n id: input.id,\n title: input.title,\n artifact_type: input.artifact_type,\n source_path: input.source_path,\n sessionId: input.session,\n content: input.content,\n tags: input.tags,\n });\n\n this.writeMarkdown(relativePath, frontmatter, body);\n return relativePath;\n }\n\n writeTeamMember(input: WriteTeamMemberInput): string {\n const relativePath = `team/${input.user}.md`;\n\n const frontmatter: Record<string, unknown> = {\n type: 'team-member',\n user: input.user,\n joined: new Date().toISOString(),\n tags: buildTags('team', '', [`user/${input.user}`]),\n };\n if (input.role) frontmatter.role = input.role;\n\n const body = formatTeamBody({\n user: input.user,\n role: input.role,\n });\n\n this.writeMarkdown(relativePath, frontmatter, body);\n return relativePath;\n }\n\n /**\n * Update frontmatter fields on an existing note without touching the body.\n * By default only adds fields that don't exist. Set overwrite=true to replace existing values.\n * Returns true if the update was applied, false if the file doesn't exist.\n */\n updateNoteFrontmatter(relativePath: string, fields: Record<string, unknown>, overwrite = false): boolean {\n const fullPath = path.join(this.vaultDir, relativePath);\n let fileContent: string;\n try {\n fileContent = fs.readFileSync(fullPath, 'utf-8');\n } catch {\n return false;\n }\n\n const fmMatch = fileContent.match(/^---\\n([\\s\\S]*?)\\n---/);\n if (!fmMatch) return false;\n\n const parsed = YAML.parse(fmMatch[1]) as Record<string, unknown>;\n for (const [key, value] of Object.entries(fields)) {\n if (overwrite || parsed[key] === undefined) {\n parsed[key] = value;\n }\n }\n\n const body = fileContent.slice(fmMatch[0].length);\n const fmYaml = YAML.stringify(parsed, { defaultStringType: 'QUOTE_DOUBLE', defaultKeyType: 'PLAIN' }).trim();\n this.atomicWrite(fullPath, `---\\n${fmYaml}\\n---${body}`);\n return true;\n }\n\n /** @deprecated Use updateNoteFrontmatter instead */\n updateSessionFrontmatter(relativePath: string, fields: Record<string, unknown>): boolean {\n return this.updateNoteFrontmatter(relativePath, fields);\n }\n\n private writeMarkdown(relativePath: string, frontmatter: Record<string, unknown>, content: string): void {\n const fullPath = path.join(this.vaultDir, relativePath);\n fs.mkdirSync(path.dirname(fullPath), { recursive: true });\n\n const fmYaml = YAML.stringify(frontmatter, { defaultStringType: 'QUOTE_DOUBLE', defaultKeyType: 'PLAIN' }).trim();\n const file = `---\\n${fmYaml}\\n---\\n\\n${content}\\n`;\n this.atomicWrite(fullPath, file);\n }\n\n /** Write to a temp file then rename — prevents Obsidian from seeing a truncated file mid-write. */\n private atomicWrite(fullPath: string, content: string): void {\n const tmp = `${fullPath}.tmp`;\n fs.writeFileSync(tmp, content, 'utf-8');\n fs.renameSync(tmp, fullPath);\n }\n}\n"],"mappings":";;;;;;;;;AAOA,IAAM,iBAAiB;AAGhB,SAAS,cAAc,QAAwB;AACpD,MAAI,OAAO,WAAW,cAAc,EAAG,QAAO;AAC9C,SAAO,GAAG,cAAc,GAAG,MAAM;AACnC;AAGO,SAAS,cAAc,QAAwB;AACpD,MAAI,OAAO,WAAW,cAAc,EAAG,QAAO,OAAO,MAAM,eAAe,MAAM;AAChF,SAAO;AACT;AAGO,SAAS,gBAAgB,QAAwB;AACtD,SAAO,KAAK,cAAc,MAAM,CAAC;AACnC;AAGO,SAAS,oBAAoB,QAAgB,MAAsB;AACxE,SAAO,YAAY,IAAI,IAAI,cAAc,MAAM,CAAC;AAClD;;;ACrBA,IAAM,cAAsC;AAAA,EAC1C,QAAQ;AAAA,EACR,SAAS;AAAA,EACT,UAAU;AAAA,EACV,WAAW;AAAA,EACX,WAAW;AACb;AAEO,SAAS,uBAAuB,iBAAiC;AACtE,SAAO,YAAY,eAAe,KAAK;AACzC;AAMO,SAAS,eAAe,MAAsB;AACnD,SAAO,KAAK,QAAQ,oBAAoB,KAAK;AAC/C;AAEO,SAAS,QAAQ,MAAc,OAAe,SAAyB;AAC5E,QAAM,OAAO,eAAe,OAAO;AACnC,QAAM,WAAW,KAAK,MAAM,IAAI,EAAE,IAAI,CAAC,SAAS,KAAK,IAAI,EAAE,EAAE,KAAK,IAAI;AACtE,SAAO,OAAO,IAAI,KAAK,KAAK;AAAA,EAAK,QAAQ;AAC3C;AAEO,SAAS,YAAY,KAAa,OAAuB;AAC9D,SAAO,GAAG,GAAG,MAAM,KAAK;AAC1B;AAEO,SAAS,SAAS,QAAgB,SAA0B;AACjE,SAAO,UAAU,KAAK,MAAM,IAAI,OAAO,OAAO,KAAK,MAAM;AAC3D;AAMA,SAAS,aAAa,GAAmB;AACvC,SAAO,EAAE,QAAQ,MAAM,GAAG;AAC5B;AAOA,SAAS,YAAY,KAAqB;AACxC,QAAM,WAAW,IAAI,WAAW,GAAG,IAAI,IAAI,MAAM,CAAC,IAAI;AACtD,SAAO,SAAS,QAAQ,QAAQ,GAAG;AACrC;AAEO,SAAS,UAAU,MAAc,SAAiB,YAAsB,CAAC,GAAa;AAC3F,QAAM,OAAiB,CAAC,QAAQ,IAAI,EAAE;AAEtC,MAAI,SAAS;AACX,SAAK,KAAK,GAAG,IAAI,IAAI,aAAa,OAAO,CAAC,EAAE;AAAA,EAC9C;AAEA,aAAW,OAAO,WAAW;AAC3B,UAAM,aAAa,YAAY,GAAG;AAClC,QAAI,cAAc,CAAC,KAAK,SAAS,UAAU,GAAG;AAC5C,WAAK,KAAK,UAAU;AAAA,IACtB;AAAA,EACF;AAEA,SAAO;AACT;AAEO,SAAS,WAAW,MAAwB;AACjD,SAAO,KAAK,IAAI,CAAC,MAAO,EAAE,WAAW,GAAG,IAAI,IAAI,IAAI,CAAC,EAAG,EAAE,KAAK,GAAG;AACpE;AAuBO,SAAS,kBAAkB,OAAiC;AACjE,QAAM,WAAqB,CAAC;AAE5B,WAAS,KAAK,KAAK,MAAM,KAAK,EAAE;AAEhC,MAAI,MAAM,WAAW;AACnB,aAAS,KAAK,QAAQ,YAAY,WAAW,MAAM,SAAS,CAAC;AAAA,EAC/D;AAGA,QAAM,SAAmB,CAAC;AAC1B,SAAO,KAAK,YAAY,WAAW,SAAS,cAAc,MAAM,SAAS,CAAC,CAAC,CAAC;AAC5E,MAAI,MAAM,KAAM,QAAO,KAAK,YAAY,QAAQ,MAAM,IAAI,CAAC;AAC3D,MAAI,MAAM,WAAW,MAAM,OAAO;AAChC,UAAM,WAAW,eAAe,MAAM,SAAS,MAAM,KAAK;AAC1D,QAAI,SAAU,QAAO,KAAK,YAAY,YAAY,QAAQ,CAAC;AAAA,EAC7D;AACA,MAAI,MAAM,OAAQ,QAAO,KAAK,YAAY,UAAU,KAAK,MAAM,MAAM,IAAI,CAAC;AAC1E,WAAS,KAAK,OAAO,KAAK,IAAI,CAAC;AAG/B,MAAI,MAAM,iBAAiB,QAAQ;AACjC,UAAM,QAAQ,MAAM,gBAAgB,IAAI,CAAC,MAAM,KAAK,SAAS,EAAE,IAAI,EAAE,KAAK,CAAC,EAAE;AAC7E,aAAS,KAAK;AAAA,EAAsB,MAAM,KAAK,IAAI,CAAC,EAAE;AAAA,EACxD;AAIA,MAAI,MAAM,MAAM,SAAS,GAAG;AAC1B,UAAM,YAAsB,CAAC;AAC7B,aAAS,IAAI,GAAG,IAAI,MAAM,MAAM,QAAQ,KAAK;AAC3C,YAAM,OAAO,MAAM,MAAM,CAAC;AAC1B,YAAM,UAAU,IAAI;AACpB,gBAAU,KAAK,YAAY,OAAO,EAAE;AACpC,UAAI,KAAK,UAAU,KAAK,QAAQ,QAAQ;AAEtC,cAAM,QAAkB,CAAC;AACzB,YAAI,KAAK,OAAQ,OAAM,KAAK,KAAK,MAAM;AACvC,YAAI,KAAK,QAAQ,QAAQ;AACvB,gBAAM,KAAK,KAAK,OAAO,IAAI,CAAC,MAAM,MAAM,CAAC,IAAI,EAAE,KAAK,IAAI,CAAC;AAAA,QAC3D;AACA,YAAI,KAAK,YAAY,EAAG,OAAM,KAAK,IAAI,KAAK,SAAS,cAAc;AACnE,kBAAU,KAAK,QAAQ,QAAQ,UAAU,MAAM,KAAK,MAAM,CAAC,CAAC;AAAA,MAC9D,WAAW,KAAK,YAAY,GAAG;AAC7B,kBAAU,KAAK,QAAQ,QAAQ,UAAU,IAAI,KAAK,SAAS,cAAc,CAAC;AAAA,MAC5E;AACA,UAAI,KAAK,YAAY;AACnB,kBAAU,KAAK,QAAQ,aAAa,YAAY,KAAK,UAAU,CAAC;AAAA,MAClE;AAAA,IACF;AACA,aAAS,KAAK;AAAA;AAAA,EAAsB,UAAU,KAAK,MAAM,CAAC,EAAE;AAAA,EAC9D;AAGA,QAAM,UAAU,UAAU,WAAW,SAAS;AAAA,IAC5C,GAAI,MAAM,OAAO,CAAC,QAAQ,MAAM,IAAI,EAAE,IAAI,CAAC;AAAA,IAC3C,GAAI,MAAM,QAAQ,CAAC;AAAA,EACrB,CAAC;AACD,WAAS,KAAK,WAAW,OAAO,CAAC;AAEjC,SAAO,SAAS,KAAK,MAAM;AAC7B;AAkBO,SAAS,gBAAgB,OAA+B;AAC7D,QAAM,WAAqB,CAAC;AAC5B,QAAM,cAAc,uBAAuB,MAAM,eAAe;AAChE,QAAM,eAAe,WAAW,aAAa,MAAM,eAAe,CAAC;AAEnE,WAAS,KAAK,KAAK,MAAM,KAAK,EAAE;AAChC,WAAS,KAAK,QAAQ,aAAa,cAAc,MAAM,OAAO,CAAC;AAG/D,QAAM,SAAmB,CAAC;AAC1B,MAAI,MAAM,WAAW;AACnB,WAAO,KAAK,YAAY,WAAW,SAAS,cAAc,MAAM,SAAS,CAAC,CAAC,CAAC;AAAA,EAC9E;AACA,SAAO,KAAK,YAAY,eAAe,MAAM,eAAe,CAAC;AAC7D,MAAI,OAAO,SAAS,EAAG,UAAS,KAAK,OAAO,KAAK,IAAI,CAAC;AAGtD,MAAI,MAAM,WAAY,UAAS,KAAK;AAAA,EAAkB,MAAM,UAAU,EAAE;AACxE,MAAI,MAAM,IAAK,UAAS,KAAK;AAAA,EAAW,MAAM,GAAG,EAAE;AACnD,MAAI,MAAM,UAAW,UAAS,KAAK;AAAA,EAAiB,MAAM,SAAS,EAAE;AACrE,MAAI,MAAM,sBAAuB,UAAS,KAAK;AAAA,EAA6B,MAAM,qBAAqB,EAAE;AACzG,MAAI,MAAM,OAAQ,UAAS,KAAK;AAAA,EAAc,MAAM,MAAM,EAAE;AAC5D,MAAI,MAAM,WAAY,UAAS,KAAK;AAAA,EAAkB,MAAM,UAAU,EAAE;AAGxE,QAAM,UAAU,UAAU,SAAS,MAAM,iBAAiB,MAAM,QAAQ,CAAC,CAAC;AAC1E,WAAS,KAAK,WAAW,OAAO,CAAC;AAEjC,SAAO,SAAS,KAAK,MAAM;AAC7B;AAcO,SAAS,eAAe,OAA8B;AAC3D,QAAM,WAAqB,CAAC;AAG5B,QAAM,SAAmB,CAAC;AAC1B,SAAO,KAAK,YAAY,QAAQ,SAAS,MAAM,EAAE,CAAC,CAAC;AACnD,SAAO,KAAK,YAAY,UAAU,MAAM,MAAM,CAAC;AAC/C,MAAI,MAAM,OAAQ,QAAO,KAAK,YAAY,UAAU,MAAM,MAAM,CAAC;AACjE,MAAI,MAAM,QAAS,QAAO,KAAK,YAAY,WAAW,MAAM,OAAO,CAAC;AACpE,WAAS,KAAK,OAAO,KAAK,IAAI,CAAC;AAG/B,WAAS,KAAK,MAAM,OAAO;AAG3B,MAAI,MAAM,UAAU,QAAQ;AAC1B,UAAM,QAAQ,MAAM,SAAS,IAAI,CAAC,MAAM,KAAK,SAAS,cAAc,EAAE,EAAE,GAAG,EAAE,KAAK,CAAC,EAAE;AACrF,aAAS,KAAK;AAAA,EAAgB,MAAM,KAAK,IAAI,CAAC,EAAE;AAAA,EAClD;AAGA,QAAM,YAAY,aAAa,MAAM,MAAM;AAC3C,QAAM,UAAU,UAAU,QAAQ,WAAW,MAAM,QAAQ,CAAC,CAAC;AAC7D,WAAS,KAAK,WAAW,OAAO,CAAC;AAEjC,SAAO,SAAS,KAAK,MAAM;AAC7B;AAcO,SAAS,mBAAmB,OAAkC;AACnE,QAAM,WAAqB,CAAC;AAG5B,QAAM,SAAmB,CAAC;AAC1B,SAAO,KAAK,YAAY,YAAY,SAAS,MAAM,EAAE,CAAC,CAAC;AACvD,SAAO,KAAK,YAAY,UAAU,KAAK,MAAM,WAAW,IAAI,CAAC;AAC7D,SAAO,KAAK,YAAY,QAAQ,MAAM,aAAa,CAAC;AACpD,SAAO,KAAK,YAAY,WAAW,SAAS,cAAc,MAAM,SAAS,CAAC,CAAC,CAAC;AAC5E,WAAS,KAAK,OAAO,KAAK,IAAI,CAAC;AAG/B,WAAS,KAAK,MAAM,OAAO;AAG3B,QAAM,UAAU,UAAU,YAAY,MAAM,eAAe,MAAM,QAAQ,CAAC,CAAC;AAC3E,WAAS,KAAK,WAAW,OAAO,CAAC;AAEjC,SAAO,SAAS,KAAK,MAAM;AAC7B;AAUO,SAAS,eAAe,OAA8B;AAC3D,QAAM,WAAqB,CAAC;AAE5B,WAAS,KAAK,KAAK,MAAM,IAAI,EAAE;AAC/B,WAAS,KAAK,QAAQ,QAAQ,eAAe,MAAM,QAAQ,aAAa,CAAC;AAGzE,QAAM,SAAmB,CAAC;AAC1B,SAAO,KAAK,YAAY,QAAQ,MAAM,IAAI,CAAC;AAC3C,MAAI,MAAM,KAAM,QAAO,KAAK,YAAY,QAAQ,MAAM,IAAI,CAAC;AAC3D,WAAS,KAAK,OAAO,KAAK,IAAI,CAAC;AAG/B,MAAI,MAAM,gBAAgB,QAAQ;AAChC,UAAM,QAAQ,MAAM,eAAe,IAAI,CAAC,MAAM,KAAK,SAAS,cAAc,EAAE,EAAE,GAAG,EAAE,KAAK,CAAC,EAAE;AAC3F,aAAS,KAAK;AAAA,EAAuB,MAAM,KAAK,IAAI,CAAC,EAAE;AAAA,EACzD;AAGA,QAAM,UAAU,UAAU,QAAQ,IAAI,CAAC,QAAQ,MAAM,IAAI,EAAE,CAAC;AAC5D,WAAS,KAAK,WAAW,OAAO,CAAC;AAEjC,SAAO,SAAS,KAAK,MAAM;AAC7B;AAIA,SAAS,WAAW,GAAmB;AACrC,SAAO,EAAE,OAAO,CAAC,EAAE,YAAY,IAAI,EAAE,MAAM,CAAC;AAC9C;AAEA,SAAS,eAAe,SAAiB,OAAuB;AAC9D,QAAM,KAAK,IAAI,KAAK,KAAK,EAAE,QAAQ,IAAI,IAAI,KAAK,OAAO,EAAE,QAAQ;AACjE,MAAI,MAAM,EAAE,KAAK,KAAK,EAAG,QAAO;AAChC,QAAM,eAAe,KAAK,MAAM,KAAK,GAAK;AAC1C,MAAI,eAAe,EAAG,QAAO;AAC7B,QAAM,QAAQ,KAAK,MAAM,eAAe,EAAE;AAC1C,QAAM,UAAU,eAAe;AAC/B,MAAI,UAAU,EAAG,QAAO,GAAG,OAAO;AAClC,MAAI,YAAY,EAAG,QAAO,GAAG,KAAK;AAClC,SAAO,GAAG,KAAK,KAAK,OAAO;AAC7B;;;AC5UA,kBAAiB;AAFjB,OAAO,QAAQ;AACf,OAAO,UAAU;AAsDV,IAAM,cAAN,MAAkB;AAAA,EACvB,YAAoB,UAAkB;AAAlB;AAAA,EAAmB;AAAA,EAEvC,aAAa,OAAkC;AAC7C,UAAM,OAAO,MAAM,QAAQ,MAAM,GAAG,EAAE;AACtC,UAAM,eAAe,oBAAoB,MAAM,IAAI,IAAI;AAEvD,UAAM,cAAuC;AAAA,MAC3C,MAAM;AAAA,MACN,IAAI,MAAM;AAAA,MACV,OAAO,MAAM,SAAS;AAAA,MACtB,MAAM,MAAM,QAAQ;AAAA,MACpB,SAAS,MAAM;AAAA,IACjB;AACA,QAAI,MAAM,MAAO,aAAY,QAAQ,MAAM;AAC3C,QAAI,MAAM,OAAQ,aAAY,SAAS,MAAM;AAC7C,QAAI,MAAM,cAAe,aAAY,gBAAgB,MAAM;AAC3D,QAAI,MAAM,OAAO,OAAQ,aAAY,QAAQ,MAAM;AACnD,QAAI,MAAM,OAAQ,aAAY,SAAS,MAAM;AAC7C,gBAAY,OAAO,UAAU,WAAW,SAAS;AAAA,MAC/C,GAAI,MAAM,OAAO,CAAC,QAAQ,MAAM,IAAI,EAAE,IAAI,CAAC;AAAA,MAC3C,GAAI,MAAM,QAAQ,CAAC;AAAA,IACrB,CAAC;AACD,QAAI,MAAM,cAAc,KAAM,aAAY,aAAa,MAAM;AAC7D,QAAI,MAAM,iBAAiB,KAAM,aAAY,gBAAgB,MAAM;AAEnE,SAAK,cAAc,cAAc,aAAa,MAAM,OAAO;AAC3D,WAAO;AAAA,EACT;AAAA,EAEA,UAAU,OAA+B;AACvC,UAAM,eAAe,SAAS,MAAM,EAAE;AACtC,UAAM,WAAW,KAAK,KAAK,KAAK,UAAU,YAAY;AAEtD,UAAM,SAAS,MAAM,UAAU;AAC/B,QAAI,WAAU,oBAAI,KAAK,GAAE,YAAY;AAGrC,QAAI;AACF,YAAM,WAAW,GAAG,aAAa,UAAU,OAAO;AAClD,YAAM,UAAU,SAAS,MAAM,uBAAuB;AACtD,UAAI,SAAS;AACX,cAAM,SAAS,YAAAA,QAAK,MAAM,QAAQ,CAAC,CAAC;AACpC,YAAI,OAAO,OAAO,YAAY,SAAU,WAAU,OAAO;AAAA,MAC3D;AAAA,IACF,QAAQ;AAAA,IAER;AACA,UAAM,cAAuC;AAAA,MAC3C,MAAM;AAAA,MACN,IAAI,MAAM;AAAA,MACV;AAAA,MACA;AAAA,IACF;AACA,QAAI,MAAM,OAAQ,aAAY,SAAS,MAAM;AAC7C,gBAAY,OAAO,UAAU,QAAQ,QAAQ,MAAM,QAAQ,CAAC,CAAC;AAE7D,UAAM,OAAO,eAAe;AAAA,MAC1B,IAAI,MAAM;AAAA,MACV;AAAA,MACA,QAAQ,MAAM;AAAA,MACd;AAAA,MACA,SAAS,MAAM;AAAA,MACf,MAAM,MAAM;AAAA,IACd,CAAC;AAED,SAAK,cAAc,cAAc,aAAa,IAAI;AAClD,WAAO;AAAA,EACT;AAAA,EAEA,WAAW,OAAgC;AACzC,UAAM,iBAAiB,MAAM,iBAAiB,QAAQ,MAAM,GAAG;AAC/D,UAAM,eAAe,UAAU,cAAc,IAAI,MAAM,EAAE;AACzD,UAAM,WAAW,KAAK,KAAK,KAAK,UAAU,YAAY;AACtD,UAAM,OAAM,oBAAI,KAAK,GAAE,YAAY;AAGnC,QAAI,UAAU;AACd,QAAI;AACF,YAAM,WAAW,GAAG,aAAa,UAAU,OAAO;AAClD,YAAM,UAAU,SAAS,MAAM,uBAAuB;AACtD,UAAI,SAAS;AACX,cAAM,SAAS,YAAAA,QAAK,MAAM,QAAQ,CAAC,CAAC;AACpC,YAAI,OAAO,OAAO,YAAY,SAAU,WAAU,OAAO;AAAA,MAC3D;AAAA,IACF,QAAQ;AAAA,IAER;AAEA,UAAM,cAAuC;AAAA,MAC3C,MAAM;AAAA,MACN,IAAI,MAAM;AAAA,MACV,kBAAkB,MAAM;AAAA,MACxB;AAAA,IACF;AACA,QAAI,MAAM,QAAS,aAAY,UAAU,MAAM;AAC/C,QAAI,MAAM,KAAM,aAAY,OAAO,MAAM;AACzC,gBAAY,OAAO,UAAU,SAAS,MAAM,kBAAkB,MAAM,QAAQ,CAAC,CAAC;AAE9E,SAAK,cAAc,cAAc,aAAa,MAAM,OAAO;AAC3D,WAAO;AAAA,EACT;AAAA,EAEA,cAAc,OAAmC;AAC/C,UAAM,eAAe,aAAa,MAAM,EAAE;AAC1C,UAAM,WAAW,KAAK,KAAK,KAAK,UAAU,YAAY;AACtD,UAAM,OAAM,oBAAI,KAAK,GAAE,YAAY;AAEnC,QAAI,UAAU;AAGd,QAAI;AACF,YAAM,WAAW,GAAG,aAAa,UAAU,OAAO;AAClD,YAAM,UAAU,SAAS,MAAM,uBAAuB;AACtD,UAAI,SAAS;AACX,cAAM,SAAS,YAAAA,QAAK,MAAM,QAAQ,CAAC,CAAC;AACpC,YAAI,OAAO,OAAO,YAAY,SAAU,WAAU,OAAO;AAAA,MAC3D;AAAA,IACF,QAAQ;AAAA,IAER;AAEA,UAAM,cAAuC;AAAA,MAC3C,MAAM;AAAA,MACN,IAAI,MAAM;AAAA,MACV,eAAe,MAAM;AAAA,MACrB,aAAa,MAAM;AAAA,MACnB,OAAO,MAAM;AAAA,MACb,kBAAkB,cAAc,MAAM,OAAO;AAAA,MAC7C;AAAA,MACA,SAAS;AAAA,MACT,MAAM,UAAU,YAAY,MAAM,eAAe,MAAM,QAAQ,CAAC,CAAC;AAAA,IACnE;AAEA,UAAM,OAAO,mBAAmB;AAAA,MAC9B,IAAI,MAAM;AAAA,MACV,OAAO,MAAM;AAAA,MACb,eAAe,MAAM;AAAA,MACrB,aAAa,MAAM;AAAA,MACnB,WAAW,MAAM;AAAA,MACjB,SAAS,MAAM;AAAA,MACf,MAAM,MAAM;AAAA,IACd,CAAC;AAED,SAAK,cAAc,cAAc,aAAa,IAAI;AAClD,WAAO;AAAA,EACT;AAAA,EAEA,gBAAgB,OAAqC;AACnD,UAAM,eAAe,QAAQ,MAAM,IAAI;AAEvC,UAAM,cAAuC;AAAA,MAC3C,MAAM;AAAA,MACN,MAAM,MAAM;AAAA,MACZ,SAAQ,oBAAI,KAAK,GAAE,YAAY;AAAA,MAC/B,MAAM,UAAU,QAAQ,IAAI,CAAC,QAAQ,MAAM,IAAI,EAAE,CAAC;AAAA,IACpD;AACA,QAAI,MAAM,KAAM,aAAY,OAAO,MAAM;AAEzC,UAAM,OAAO,eAAe;AAAA,MAC1B,MAAM,MAAM;AAAA,MACZ,MAAM,MAAM;AAAA,IACd,CAAC;AAED,SAAK,cAAc,cAAc,aAAa,IAAI;AAClD,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOA,sBAAsB,cAAsB,QAAiC,YAAY,OAAgB;AACvG,UAAM,WAAW,KAAK,KAAK,KAAK,UAAU,YAAY;AACtD,QAAI;AACJ,QAAI;AACF,oBAAc,GAAG,aAAa,UAAU,OAAO;AAAA,IACjD,QAAQ;AACN,aAAO;AAAA,IACT;AAEA,UAAM,UAAU,YAAY,MAAM,uBAAuB;AACzD,QAAI,CAAC,QAAS,QAAO;AAErB,UAAM,SAAS,YAAAA,QAAK,MAAM,QAAQ,CAAC,CAAC;AACpC,eAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,MAAM,GAAG;AACjD,UAAI,aAAa,OAAO,GAAG,MAAM,QAAW;AAC1C,eAAO,GAAG,IAAI;AAAA,MAChB;AAAA,IACF;AAEA,UAAM,OAAO,YAAY,MAAM,QAAQ,CAAC,EAAE,MAAM;AAChD,UAAM,SAAS,YAAAA,QAAK,UAAU,QAAQ,EAAE,mBAAmB,gBAAgB,gBAAgB,QAAQ,CAAC,EAAE,KAAK;AAC3G,SAAK,YAAY,UAAU;AAAA,EAAQ,MAAM;AAAA,KAAQ,IAAI,EAAE;AACvD,WAAO;AAAA,EACT;AAAA;AAAA,EAGA,yBAAyB,cAAsB,QAA0C;AACvF,WAAO,KAAK,sBAAsB,cAAc,MAAM;AAAA,EACxD;AAAA,EAEQ,cAAc,cAAsB,aAAsC,SAAuB;AACvG,UAAM,WAAW,KAAK,KAAK,KAAK,UAAU,YAAY;AACtD,OAAG,UAAU,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAExD,UAAM,SAAS,YAAAA,QAAK,UAAU,aAAa,EAAE,mBAAmB,gBAAgB,gBAAgB,QAAQ,CAAC,EAAE,KAAK;AAChH,UAAM,OAAO;AAAA,EAAQ,MAAM;AAAA;AAAA;AAAA,EAAY,OAAO;AAAA;AAC9C,SAAK,YAAY,UAAU,IAAI;AAAA,EACjC;AAAA;AAAA,EAGQ,YAAY,UAAkB,SAAuB;AAC3D,UAAM,MAAM,GAAG,QAAQ;AACvB,OAAG,cAAc,KAAK,SAAS,OAAO;AACtC,OAAG,WAAW,KAAK,QAAQ;AAAA,EAC7B;AACF;","names":["YAML"]}
@@ -0,0 +1,99 @@
1
+ import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
2
+ import {
3
+ parseStringFlag
4
+ } from "./chunk-SAKJMNSR.js";
5
+ import {
6
+ MycoConfigSchema,
7
+ require_dist
8
+ } from "./chunk-G6ZMTQMJ.js";
9
+ import {
10
+ __toESM
11
+ } from "./chunk-PZUWP5VK.js";
12
+
13
+ // src/cli/setup-llm.ts
14
+ var import_yaml = __toESM(require_dist(), 1);
15
+ import fs from "fs";
16
+ import path from "path";
17
+ var CONFIG_FILENAME = "myco.yaml";
18
+ var DAEMON_STATE_FILENAME = "daemon.json";
19
+ var USAGE = `Usage: myco setup-llm [options]
20
+
21
+ Configure LLM and embedding provider settings.
22
+
23
+ Options:
24
+ --llm-provider <name> LLM provider (ollama, lm-studio, anthropic)
25
+ --llm-model <name> LLM model name
26
+ --llm-url <url> LLM provider base URL
27
+ --llm-context-window <number> LLM context window (tokens)
28
+ --llm-max-tokens <number> LLM max output tokens
29
+ --embedding-provider <name> Embedding provider (ollama, lm-studio)
30
+ --embedding-model <name> Embedding model name
31
+ --embedding-url <url> Embedding provider base URL
32
+ --show Show current settings and exit
33
+ `;
34
+ async function run(args, vaultDir) {
35
+ const configPath = path.join(vaultDir, CONFIG_FILENAME);
36
+ const raw = fs.readFileSync(configPath, "utf-8");
37
+ const doc = import_yaml.default.parse(raw);
38
+ if (args.includes("--show")) {
39
+ const config = MycoConfigSchema.parse(doc);
40
+ console.log(JSON.stringify(config.intelligence, null, 2));
41
+ return;
42
+ }
43
+ if (args.length === 0) {
44
+ console.log(USAGE);
45
+ return;
46
+ }
47
+ if (!doc.intelligence || typeof doc.intelligence !== "object") {
48
+ doc.intelligence = {};
49
+ }
50
+ const intelligence = doc.intelligence;
51
+ if (!intelligence.llm || typeof intelligence.llm !== "object") {
52
+ intelligence.llm = {};
53
+ }
54
+ if (!intelligence.embedding || typeof intelligence.embedding !== "object") {
55
+ intelligence.embedding = {};
56
+ }
57
+ const llm = intelligence.llm;
58
+ const embedding = intelligence.embedding;
59
+ const llmProvider = parseStringFlag(args, "--llm-provider");
60
+ if (llmProvider !== void 0) llm.provider = llmProvider;
61
+ const llmModel = parseStringFlag(args, "--llm-model");
62
+ if (llmModel !== void 0) llm.model = llmModel;
63
+ const llmUrl = parseStringFlag(args, "--llm-url");
64
+ if (llmUrl !== void 0) llm.base_url = llmUrl;
65
+ const llmContextWindow = parseStringFlag(args, "--llm-context-window");
66
+ if (llmContextWindow !== void 0) llm.context_window = parseInt(llmContextWindow, 10);
67
+ const llmMaxTokens = parseStringFlag(args, "--llm-max-tokens");
68
+ if (llmMaxTokens !== void 0) llm.max_tokens = parseInt(llmMaxTokens, 10);
69
+ const embeddingProvider = parseStringFlag(args, "--embedding-provider");
70
+ if (embeddingProvider !== void 0) embedding.provider = embeddingProvider;
71
+ const embeddingModel = parseStringFlag(args, "--embedding-model");
72
+ if (embeddingModel !== void 0) embedding.model = embeddingModel;
73
+ const embeddingUrl = parseStringFlag(args, "--embedding-url");
74
+ if (embeddingUrl !== void 0) embedding.base_url = embeddingUrl;
75
+ const result = MycoConfigSchema.safeParse(doc);
76
+ if (!result.success) {
77
+ console.error("Validation error:");
78
+ for (const issue of result.error.issues) {
79
+ console.error(` ${issue.path.join(".")}: ${issue.message}`);
80
+ }
81
+ process.exit(1);
82
+ }
83
+ fs.writeFileSync(configPath, import_yaml.default.stringify(doc), "utf-8");
84
+ console.log("Intelligence configuration updated.");
85
+ const updated = MycoConfigSchema.parse(doc);
86
+ console.log(JSON.stringify(updated.intelligence, null, 2));
87
+ if (embeddingModel !== void 0) {
88
+ console.log("\nWarning: changing the embedding model requires a full vector index rebuild.");
89
+ console.log("Run: node dist/src/cli.js rebuild");
90
+ }
91
+ if (fs.existsSync(path.join(vaultDir, DAEMON_STATE_FILENAME))) {
92
+ console.log("\nNote: restart the daemon for changes to take effect (myco restart)");
93
+ }
94
+ }
95
+
96
+ export {
97
+ run
98
+ };
99
+ //# sourceMappingURL=chunk-VYV5IFD6.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/cli/setup-llm.ts"],"sourcesContent":["import fs from 'node:fs';\nimport path from 'node:path';\nimport YAML from 'yaml';\nimport { MycoConfigSchema } from '../config/schema.js';\nimport { parseStringFlag } from './shared.js';\n\nconst CONFIG_FILENAME = 'myco.yaml';\nconst DAEMON_STATE_FILENAME = 'daemon.json';\n\nconst USAGE = `Usage: myco setup-llm [options]\n\nConfigure LLM and embedding provider settings.\n\nOptions:\n --llm-provider <name> LLM provider (ollama, lm-studio, anthropic)\n --llm-model <name> LLM model name\n --llm-url <url> LLM provider base URL\n --llm-context-window <number> LLM context window (tokens)\n --llm-max-tokens <number> LLM max output tokens\n --embedding-provider <name> Embedding provider (ollama, lm-studio)\n --embedding-model <name> Embedding model name\n --embedding-url <url> Embedding provider base URL\n --show Show current settings and exit\n`;\n\nexport async function run(args: string[], vaultDir: string): Promise<void> {\n const configPath = path.join(vaultDir, CONFIG_FILENAME);\n const raw = fs.readFileSync(configPath, 'utf-8');\n const doc = YAML.parse(raw) as Record<string, unknown>;\n\n // Show current settings\n if (args.includes('--show')) {\n const config = MycoConfigSchema.parse(doc);\n console.log(JSON.stringify(config.intelligence, null, 2));\n return;\n }\n\n // No flags = show usage\n if (args.length === 0) {\n console.log(USAGE);\n return;\n }\n\n // Ensure intelligence section exists\n if (!doc.intelligence || typeof doc.intelligence !== 'object') {\n doc.intelligence = {};\n }\n const intelligence = doc.intelligence as Record<string, unknown>;\n\n if (!intelligence.llm || typeof intelligence.llm !== 'object') {\n intelligence.llm = {};\n }\n if (!intelligence.embedding || typeof intelligence.embedding !== 'object') {\n intelligence.embedding = {};\n }\n\n const llm = intelligence.llm as Record<string, unknown>;\n const embedding = intelligence.embedding as Record<string, unknown>;\n\n // Parse and apply flags\n const llmProvider = parseStringFlag(args, '--llm-provider');\n if (llmProvider !== undefined) llm.provider = llmProvider;\n\n const llmModel = parseStringFlag(args, '--llm-model');\n if (llmModel !== undefined) llm.model = llmModel;\n\n const llmUrl = parseStringFlag(args, '--llm-url');\n if (llmUrl !== undefined) llm.base_url = llmUrl;\n\n const llmContextWindow = parseStringFlag(args, '--llm-context-window');\n if (llmContextWindow !== undefined) llm.context_window = parseInt(llmContextWindow, 10);\n\n const llmMaxTokens = parseStringFlag(args, '--llm-max-tokens');\n if (llmMaxTokens !== undefined) llm.max_tokens = parseInt(llmMaxTokens, 10);\n\n const embeddingProvider = parseStringFlag(args, '--embedding-provider');\n if (embeddingProvider !== undefined) embedding.provider = embeddingProvider;\n\n const embeddingModel = parseStringFlag(args, '--embedding-model');\n if (embeddingModel !== undefined) embedding.model = embeddingModel;\n\n const embeddingUrl = parseStringFlag(args, '--embedding-url');\n if (embeddingUrl !== undefined) embedding.base_url = embeddingUrl;\n\n // Validate the full config\n const result = MycoConfigSchema.safeParse(doc);\n if (!result.success) {\n console.error('Validation error:');\n for (const issue of result.error.issues) {\n console.error(` ${issue.path.join('.')}: ${issue.message}`);\n }\n process.exit(1);\n }\n\n // Write back\n fs.writeFileSync(configPath, YAML.stringify(doc), 'utf-8');\n console.log('Intelligence configuration updated.');\n\n // Show what was set\n const updated = MycoConfigSchema.parse(doc);\n console.log(JSON.stringify(updated.intelligence, null, 2));\n\n // Warn about embedding model changes\n if (embeddingModel !== undefined) {\n console.log('\\nWarning: changing the embedding model requires a full vector index rebuild.');\n console.log('Run: node dist/src/cli.js rebuild');\n }\n\n if (fs.existsSync(path.join(vaultDir, DAEMON_STATE_FILENAME))) {\n console.log('\\nNote: restart the daemon for changes to take effect (myco restart)');\n }\n}\n"],"mappings":";;;;;;;;;;;;;AAEA,kBAAiB;AAFjB,OAAO,QAAQ;AACf,OAAO,UAAU;AAKjB,IAAM,kBAAkB;AACxB,IAAM,wBAAwB;AAE9B,IAAM,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgBd,eAAsB,IAAI,MAAgB,UAAiC;AACzE,QAAM,aAAa,KAAK,KAAK,UAAU,eAAe;AACtD,QAAM,MAAM,GAAG,aAAa,YAAY,OAAO;AAC/C,QAAM,MAAM,YAAAA,QAAK,MAAM,GAAG;AAG1B,MAAI,KAAK,SAAS,QAAQ,GAAG;AAC3B,UAAM,SAAS,iBAAiB,MAAM,GAAG;AACzC,YAAQ,IAAI,KAAK,UAAU,OAAO,cAAc,MAAM,CAAC,CAAC;AACxD;AAAA,EACF;AAGA,MAAI,KAAK,WAAW,GAAG;AACrB,YAAQ,IAAI,KAAK;AACjB;AAAA,EACF;AAGA,MAAI,CAAC,IAAI,gBAAgB,OAAO,IAAI,iBAAiB,UAAU;AAC7D,QAAI,eAAe,CAAC;AAAA,EACtB;AACA,QAAM,eAAe,IAAI;AAEzB,MAAI,CAAC,aAAa,OAAO,OAAO,aAAa,QAAQ,UAAU;AAC7D,iBAAa,MAAM,CAAC;AAAA,EACtB;AACA,MAAI,CAAC,aAAa,aAAa,OAAO,aAAa,cAAc,UAAU;AACzE,iBAAa,YAAY,CAAC;AAAA,EAC5B;AAEA,QAAM,MAAM,aAAa;AACzB,QAAM,YAAY,aAAa;AAG/B,QAAM,cAAc,gBAAgB,MAAM,gBAAgB;AAC1D,MAAI,gBAAgB,OAAW,KAAI,WAAW;AAE9C,QAAM,WAAW,gBAAgB,MAAM,aAAa;AACpD,MAAI,aAAa,OAAW,KAAI,QAAQ;AAExC,QAAM,SAAS,gBAAgB,MAAM,WAAW;AAChD,MAAI,WAAW,OAAW,KAAI,WAAW;AAEzC,QAAM,mBAAmB,gBAAgB,MAAM,sBAAsB;AACrE,MAAI,qBAAqB,OAAW,KAAI,iBAAiB,SAAS,kBAAkB,EAAE;AAEtF,QAAM,eAAe,gBAAgB,MAAM,kBAAkB;AAC7D,MAAI,iBAAiB,OAAW,KAAI,aAAa,SAAS,cAAc,EAAE;AAE1E,QAAM,oBAAoB,gBAAgB,MAAM,sBAAsB;AACtE,MAAI,sBAAsB,OAAW,WAAU,WAAW;AAE1D,QAAM,iBAAiB,gBAAgB,MAAM,mBAAmB;AAChE,MAAI,mBAAmB,OAAW,WAAU,QAAQ;AAEpD,QAAM,eAAe,gBAAgB,MAAM,iBAAiB;AAC5D,MAAI,iBAAiB,OAAW,WAAU,WAAW;AAGrD,QAAM,SAAS,iBAAiB,UAAU,GAAG;AAC7C,MAAI,CAAC,OAAO,SAAS;AACnB,YAAQ,MAAM,mBAAmB;AACjC,eAAW,SAAS,OAAO,MAAM,QAAQ;AACvC,cAAQ,MAAM,KAAK,MAAM,KAAK,KAAK,GAAG,CAAC,KAAK,MAAM,OAAO,EAAE;AAAA,IAC7D;AACA,YAAQ,KAAK,CAAC;AAAA,EAChB;AAGA,KAAG,cAAc,YAAY,YAAAA,QAAK,UAAU,GAAG,GAAG,OAAO;AACzD,UAAQ,IAAI,qCAAqC;AAGjD,QAAM,UAAU,iBAAiB,MAAM,GAAG;AAC1C,UAAQ,IAAI,KAAK,UAAU,QAAQ,cAAc,MAAM,CAAC,CAAC;AAGzD,MAAI,mBAAmB,QAAW;AAChC,YAAQ,IAAI,+EAA+E;AAC3F,YAAQ,IAAI,mCAAmC;AAAA,EACjD;AAEA,MAAI,GAAG,WAAW,KAAK,KAAK,UAAU,qBAAqB,CAAC,GAAG;AAC7D,YAAQ,IAAI,sEAAsE;AAAA,EACpF;AACF;","names":["YAML"]}
@@ -1,7 +1,7 @@
1
1
  import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
2
2
  import {
3
3
  AgentRegistry
4
- } from "./chunk-JKOALBZC.js";
4
+ } from "./chunk-BNIYWCST.js";
5
5
 
6
6
  // src/native-deps.ts
7
7
  import { execFileSync } from "child_process";
@@ -53,4 +53,4 @@ function ensureNativeDeps() {
53
53
  export {
54
54
  ensureNativeDeps
55
55
  };
56
- //# sourceMappingURL=chunk-LR7RQCOB.js.map
56
+ //# sourceMappingURL=chunk-XCPQHC4X.js.map
@@ -2,10 +2,10 @@ import { createRequire as __cr } from 'node:module'; const require = __cr(import
2
2
  import {
3
3
  LmStudioBackend,
4
4
  OllamaBackend
5
- } from "./chunk-BA23DROX.js";
5
+ } from "./chunk-P7RNAYU7.js";
6
6
  import {
7
7
  AgentRegistry
8
- } from "./chunk-JKOALBZC.js";
8
+ } from "./chunk-BNIYWCST.js";
9
9
 
10
10
  // src/cli/shared.ts
11
11
  import fs from "fs";
@@ -48,22 +48,22 @@ TABLE user, started, tools_used FROM #type/session
48
48
  SORT started DESC LIMIT 10
49
49
  \`\`\`
50
50
 
51
- ## Recent Memories
51
+ ## Recent Spores
52
52
  \`\`\`dataview
53
- TABLE observation_type AS "Type", created FROM #type/memory
53
+ TABLE observation_type AS "Type", created FROM #type/spore
54
54
  SORT created DESC LIMIT 15
55
55
  \`\`\`
56
56
 
57
- ## Memories by Type
57
+ ## Spores by Type
58
58
  \`\`\`dataview
59
59
  TABLE WITHOUT ID observation_type AS "Type", length(rows) AS "Count"
60
- FROM #type/memory GROUP BY observation_type
60
+ FROM #type/spore GROUP BY observation_type
61
61
  SORT length(rows) DESC
62
62
  \`\`\`
63
63
 
64
64
  ## Gotchas
65
65
  \`\`\`dataview
66
- LIST FROM #memory/gotcha SORT created DESC LIMIT 10
66
+ LIST FROM #spore/gotcha SORT created DESC LIMIT 10
67
67
  \`\`\`
68
68
  `;
69
69
  var VAULT_GITIGNORE = `# Runtime \u2014 rebuilt on daemon startup
@@ -112,9 +112,8 @@ For other agents, add to your shell profile:`);
112
112
  export {
113
113
  loadEnv,
114
114
  isProcessAlive,
115
- PROVIDER_DEFAULTS,
116
115
  DASHBOARD_CONTENT,
117
116
  VAULT_GITIGNORE,
118
117
  configureVaultEnv
119
118
  };
120
- //# sourceMappingURL=chunk-CCIV47S4.js.map
119
+ //# sourceMappingURL=chunk-XHWIIU5D.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/cli/shared.ts"],"sourcesContent":["import fs from 'node:fs';\nimport path from 'node:path';\nimport os from 'node:os';\nimport { AgentRegistry } from '../agents/registry.js';\nimport { OllamaBackend } from '../intelligence/ollama.js';\nimport { LmStudioBackend } from '../intelligence/lm-studio.js';\n\nexport { parseStringFlag, parseIntFlag } from '../logs/format.js';\n\n/** Load .env from cwd (not script location — that's the plugin install dir). */\nexport function loadEnv(): void {\n const envPath = path.resolve(process.cwd(), '.env');\n if (!fs.existsSync(envPath)) return;\n for (const line of fs.readFileSync(envPath, 'utf-8').split('\\n')) {\n const match = line.match(/^\\s*([^#=]+?)\\s*=\\s*(.*?)\\s*$/);\n if (match && !process.env[match[1]]) {\n process.env[match[1]] = match[2];\n }\n }\n}\n\nexport function isProcessAlive(pid: number): boolean {\n try { process.kill(pid, 0); return true; } catch { return false; }\n}\n\n// --- Provider defaults (sourced from backend classes) ---\nexport const PROVIDER_DEFAULTS: Record<string, { base_url: string }> = {\n ollama: { base_url: OllamaBackend.DEFAULT_BASE_URL },\n 'lm-studio': { base_url: LmStudioBackend.DEFAULT_BASE_URL },\n};\n\nexport const DASHBOARD_CONTENT = `# Myco Vault\n\n## Active Plans\n\\`\\`\\`dataview\nTABLE status, tags FROM #type/plan\nWHERE status = \"active\" OR status = \"in_progress\"\nSORT created DESC\n\\`\\`\\`\n\n## Recent Sessions\n\\`\\`\\`dataview\nTABLE user, started, tools_used FROM #type/session\nSORT started DESC LIMIT 10\n\\`\\`\\`\n\n## Recent Spores\n\\`\\`\\`dataview\nTABLE observation_type AS \"Type\", created FROM #type/spore\nSORT created DESC LIMIT 15\n\\`\\`\\`\n\n## Spores by Type\n\\`\\`\\`dataview\nTABLE WITHOUT ID observation_type AS \"Type\", length(rows) AS \"Count\"\nFROM #type/spore GROUP BY observation_type\nSORT length(rows) DESC\n\\`\\`\\`\n\n## Gotchas\n\\`\\`\\`dataview\nLIST FROM #spore/gotcha SORT created DESC LIMIT 10\n\\`\\`\\`\n`;\n\nexport const VAULT_GITIGNORE = `# Runtime — rebuilt on daemon startup\nindex.db\nindex.db-wal\nindex.db-shm\nvectors.db\n\n# Daemon state — per-machine, ephemeral\ndaemon.json\nbuffer/\nlogs/\n\n# Obsidian — per-user workspace config\n.obsidian/\n`;\n\n/** Collapse an absolute home-dir path to its `~/` form for portable config storage. */\nexport function collapseHomePath(absPath: string): string {\n const home = os.homedir();\n if (absPath.startsWith(home + path.sep) || absPath === home) {\n return '~' + absPath.slice(home.length);\n }\n return absPath;\n}\n\n/** Set MYCO_VAULT_DIR in the active agent's config, falling back to all known agents. */\nexport function configureVaultEnv(projectRoot: string, vaultDir: string): void {\n const registry = new AgentRegistry();\n const active = registry.detectActiveAgent();\n // Store the portable ~/... form so config files don't leak the username\n const portableDir = collapseHomePath(vaultDir);\n\n if (active) {\n if (active.configureVaultEnv(projectRoot, portableDir)) {\n console.log(`Set MYCO_VAULT_DIR for ${active.displayName}`);\n }\n } else {\n // No active agent detected — try all adapters\n for (const name of registry.adapterNames) {\n const adapter = registry.getAdapter(name);\n if (adapter?.configureVaultEnv(projectRoot, portableDir)) {\n console.log(`Set MYCO_VAULT_DIR for ${adapter.displayName}`);\n }\n }\n }\n\n console.log(`\\nFor other agents, add to your shell profile:`);\n console.log(` export MYCO_VAULT_DIR=\"${portableDir}\"\\n`);\n}\n"],"mappings":";;;;;;;;;;AAAA,OAAO,QAAQ;AACf,OAAO,UAAU;AACjB,OAAO,QAAQ;AAQR,SAAS,UAAgB;AAC9B,QAAM,UAAU,KAAK,QAAQ,QAAQ,IAAI,GAAG,MAAM;AAClD,MAAI,CAAC,GAAG,WAAW,OAAO,EAAG;AAC7B,aAAW,QAAQ,GAAG,aAAa,SAAS,OAAO,EAAE,MAAM,IAAI,GAAG;AAChE,UAAM,QAAQ,KAAK,MAAM,+BAA+B;AACxD,QAAI,SAAS,CAAC,QAAQ,IAAI,MAAM,CAAC,CAAC,GAAG;AACnC,cAAQ,IAAI,MAAM,CAAC,CAAC,IAAI,MAAM,CAAC;AAAA,IACjC;AAAA,EACF;AACF;AAEO,SAAS,eAAe,KAAsB;AACnD,MAAI;AAAE,YAAQ,KAAK,KAAK,CAAC;AAAG,WAAO;AAAA,EAAM,QAAQ;AAAE,WAAO;AAAA,EAAO;AACnE;AAGO,IAAM,oBAA0D;AAAA,EACrE,QAAQ,EAAE,UAAU,cAAc,iBAAiB;AAAA,EACnD,aAAa,EAAE,UAAU,gBAAgB,iBAAiB;AAC5D;AAEO,IAAM,oBAAoB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAkC1B,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgBxB,SAAS,iBAAiB,SAAyB;AACxD,QAAM,OAAO,GAAG,QAAQ;AACxB,MAAI,QAAQ,WAAW,OAAO,KAAK,GAAG,KAAK,YAAY,MAAM;AAC3D,WAAO,MAAM,QAAQ,MAAM,KAAK,MAAM;AAAA,EACxC;AACA,SAAO;AACT;AAGO,SAAS,kBAAkB,aAAqB,UAAwB;AAC7E,QAAM,WAAW,IAAI,cAAc;AACnC,QAAM,SAAS,SAAS,kBAAkB;AAE1C,QAAM,cAAc,iBAAiB,QAAQ;AAE7C,MAAI,QAAQ;AACV,QAAI,OAAO,kBAAkB,aAAa,WAAW,GAAG;AACtD,cAAQ,IAAI,0BAA0B,OAAO,WAAW,EAAE;AAAA,IAC5D;AAAA,EACF,OAAO;AAEL,eAAW,QAAQ,SAAS,cAAc;AACxC,YAAM,UAAU,SAAS,WAAW,IAAI;AACxC,UAAI,SAAS,kBAAkB,aAAa,WAAW,GAAG;AACxD,gBAAQ,IAAI,0BAA0B,QAAQ,WAAW,EAAE;AAAA,MAC7D;AAAA,IACF;AAAA,EACF;AAEA,UAAQ,IAAI;AAAA,6CAAgD;AAC5D,UAAQ,IAAI,4BAA4B,WAAW;AAAA,CAAK;AAC1D;","names":[]}
@@ -1,29 +1,26 @@
1
1
  import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
2
2
  import {
3
- formatMemoryBody,
3
+ formatSporeBody,
4
4
  sessionNoteId
5
- } from "./chunk-QWU7QLZI.js";
5
+ } from "./chunk-TZDDXRHG.js";
6
6
  import {
7
7
  ARTIFACT_TYPES,
8
8
  indexNote
9
- } from "./chunk-AWF3M57N.js";
9
+ } from "./chunk-S7EIHYE7.js";
10
10
  import {
11
11
  external_exports
12
- } from "./chunk-ISCT2SI6.js";
12
+ } from "./chunk-G6ZMTQMJ.js";
13
13
  import {
14
14
  AgentRegistry
15
- } from "./chunk-JKOALBZC.js";
15
+ } from "./chunk-BNIYWCST.js";
16
16
  import {
17
17
  AI_RESPONSE_PREVIEW_CHARS,
18
18
  CANDIDATE_CONTENT_PREVIEW,
19
19
  CHARS_PER_TOKEN,
20
- CLASSIFICATION_MAX_TOKENS,
21
20
  COMMAND_PREVIEW_CHARS,
22
- EXTRACTION_MAX_TOKENS,
23
21
  PROMPT_PREVIEW_CHARS,
24
- SUMMARY_MAX_TOKENS,
25
- TITLE_MAX_TOKENS
26
- } from "./chunk-7VPJK56U.js";
22
+ estimateTokens
23
+ } from "./chunk-JBD5KP5G.js";
27
24
 
28
25
  // src/prompts/index.ts
29
26
  import fs from "fs";
@@ -59,18 +56,20 @@ function interpolate(template, vars) {
59
56
  }
60
57
  return result;
61
58
  }
62
- function buildExtractionPrompt(sessionId, eventCount, toolSummary) {
59
+ function buildExtractionPrompt(sessionId, eventCount, toolSummary, maxTokens) {
63
60
  return interpolate(loadPrompt("extraction"), {
64
61
  sessionId,
65
62
  eventCount: String(eventCount),
66
- toolSummary
63
+ toolSummary,
64
+ maxTokens: String(maxTokens ?? 2048)
67
65
  });
68
66
  }
69
- function buildSummaryPrompt(sessionId, user, content) {
67
+ function buildSummaryPrompt(sessionId, user, content, maxTokens) {
70
68
  return interpolate(loadPrompt("summary"), {
71
69
  sessionId,
72
70
  user,
73
- content
71
+ content,
72
+ maxTokens: String(maxTokens ?? 1024)
74
73
  });
75
74
  }
76
75
  function buildTitlePrompt(summary, sessionId) {
@@ -92,7 +91,7 @@ function buildSimilarityPrompt(currentSummary, candidateSummary) {
92
91
  candidateSummary
93
92
  });
94
93
  }
95
- function buildClassificationPrompt(sessionId, candidates) {
94
+ function buildClassificationPrompt(sessionId, candidates, maxTokens) {
96
95
  const fileList = candidates.map((c) => {
97
96
  const truncated = c.content.slice(0, CANDIDATE_CONTENT_PREVIEW);
98
97
  return `### ${c.path}
@@ -104,7 +103,8 @@ ${truncated}
104
103
  sessionId,
105
104
  fileList,
106
105
  artifactTypes: ARTIFACT_TYPE_DESCRIPTIONS.map((d) => `- ${d}`).join("\n"),
107
- validTypes: ARTIFACT_TYPES.join("|")
106
+ validTypes: ARTIFACT_TYPES.join("|"),
107
+ maxTokens: String(maxTokens ?? 1024)
108
108
  });
109
109
  }
110
110
 
@@ -117,7 +117,11 @@ var REASONING_PATTERNS = [
117
117
  // <reasoning>...</reasoning>answer
118
118
  /<reasoning>[\s\S]*?<\/reasoning>\s*/gi,
119
119
  // <|thinking|>...<|/thinking|>answer
120
- /<\|thinking\|>[\s\S]*?<\|\/thinking\|>\s*/gi
120
+ /<\|thinking\|>[\s\S]*?<\|\/thinking\|>\s*/gi,
121
+ // Plain-text "Thinking Process:" block followed by actual content
122
+ // (Qwen 3.5 via LM Studio without native thinking mode)
123
+ // Matches from "Thinking Process:" up to the last numbered step, then the synthesis follows
124
+ /^Thinking Process:[\s\S]*?(?=\n(?:## |# |\*\*[A-Z]))/i
121
125
  ];
122
126
  function stripReasoningTokens(text) {
123
127
  if (!text) return text;
@@ -158,22 +162,30 @@ var ClassificationResponseSchema = external_exports.object({
158
162
  })).default([])
159
163
  });
160
164
  var BufferProcessor = class {
161
- constructor(backend, contextWindow = 8192) {
165
+ constructor(backend, contextWindow = 8192, captureConfig) {
162
166
  this.backend = backend;
163
167
  this.contextWindow = contextWindow;
168
+ this.extractionMaxTokens = captureConfig?.extraction_max_tokens ?? 2048;
169
+ this.summaryMaxTokens = captureConfig?.summary_max_tokens ?? 512;
170
+ this.titleMaxTokens = captureConfig?.title_max_tokens ?? 32;
171
+ this.classificationMaxTokens = captureConfig?.classification_max_tokens ?? 1024;
164
172
  }
173
+ extractionMaxTokens;
174
+ summaryMaxTokens;
175
+ titleMaxTokens;
176
+ classificationMaxTokens;
165
177
  truncateForContext(data, maxTokens) {
166
178
  const available = this.contextWindow - maxTokens;
167
- const dataTokens = Math.ceil(data.length / CHARS_PER_TOKEN);
179
+ const dataTokens = estimateTokens(data);
168
180
  if (dataTokens <= available) return data;
169
181
  const charBudget = available * CHARS_PER_TOKEN;
170
182
  return data.slice(0, charBudget);
171
183
  }
172
184
  async process(events, sessionId) {
173
185
  const rawPrompt = this.buildPromptForExtraction(events, sessionId);
174
- const prompt = this.truncateForContext(rawPrompt, EXTRACTION_MAX_TOKENS);
186
+ const prompt = this.truncateForContext(rawPrompt, this.extractionMaxTokens);
175
187
  try {
176
- const response = await this.backend.summarize(prompt, { maxTokens: EXTRACTION_MAX_TOKENS });
188
+ const response = await this.backend.summarize(prompt, { maxTokens: this.extractionMaxTokens });
177
189
  const parsed = extractJson(response.text);
178
190
  return {
179
191
  summary: parsed.summary,
@@ -190,14 +202,14 @@ var BufferProcessor = class {
190
202
  }
191
203
  buildPromptForExtraction(events, sessionId) {
192
204
  const toolSummary = this.summarizeEvents(events);
193
- return buildExtractionPrompt(sessionId, events.length, toolSummary);
205
+ return buildExtractionPrompt(sessionId, events.length, toolSummary, this.extractionMaxTokens);
194
206
  }
195
207
  async summarizeSession(conversationMarkdown, sessionId, user) {
196
- const truncatedContent = this.truncateForContext(conversationMarkdown, SUMMARY_MAX_TOKENS);
197
- const summaryPrompt = buildSummaryPrompt(sessionId, user ?? "unknown", truncatedContent);
208
+ const truncatedContent = this.truncateForContext(conversationMarkdown, this.summaryMaxTokens);
209
+ const summaryPrompt = buildSummaryPrompt(sessionId, user ?? "unknown", truncatedContent, this.summaryMaxTokens);
198
210
  let summaryText;
199
211
  try {
200
- const response = await this.backend.summarize(summaryPrompt, { maxTokens: SUMMARY_MAX_TOKENS });
212
+ const response = await this.backend.summarize(summaryPrompt, { maxTokens: this.summaryMaxTokens });
201
213
  summaryText = stripReasoningTokens(response.text);
202
214
  } catch (error) {
203
215
  summaryText = `Session ${sessionId} \u2014 summarization failed: ${error.message}`;
@@ -205,7 +217,7 @@ var BufferProcessor = class {
205
217
  const titlePrompt = buildTitlePrompt(summaryText, sessionId);
206
218
  let title;
207
219
  try {
208
- const response = await this.backend.summarize(titlePrompt, { maxTokens: TITLE_MAX_TOKENS });
220
+ const response = await this.backend.summarize(titlePrompt, { maxTokens: this.titleMaxTokens });
209
221
  title = stripReasoningTokens(response.text).trim();
210
222
  } catch {
211
223
  title = `Session ${sessionId}`;
@@ -215,13 +227,13 @@ var BufferProcessor = class {
215
227
  async classifyArtifacts(candidates, sessionId) {
216
228
  if (candidates.length === 0) return [];
217
229
  const prompt = this.buildPromptForClassification(candidates, sessionId);
218
- const response = await this.backend.summarize(prompt, { maxTokens: CLASSIFICATION_MAX_TOKENS });
230
+ const response = await this.backend.summarize(prompt, { maxTokens: this.classificationMaxTokens });
219
231
  const raw = extractJson(response.text);
220
232
  const parsed = ClassificationResponseSchema.parse(raw);
221
233
  return parsed.artifacts;
222
234
  }
223
235
  buildPromptForClassification(candidates, sessionId) {
224
- return buildClassificationPrompt(sessionId, candidates);
236
+ return buildClassificationPrompt(sessionId, candidates, this.classificationMaxTokens);
225
237
  }
226
238
  summarizeEvents(events) {
227
239
  const toolCounts = /* @__PURE__ */ new Map();
@@ -325,7 +337,7 @@ function writeObservationNotes(observations, sessionId, writer, index, vaultDir)
325
337
  const results = [];
326
338
  for (const obs of observations) {
327
339
  const obsId = `${obs.type}-${sessionId.slice(-6)}-${Date.now()}`;
328
- const body = formatMemoryBody({
340
+ const body = formatSporeBody({
329
341
  title: obs.title,
330
342
  observationType: obs.type,
331
343
  content: obs.content,
@@ -338,7 +350,7 @@ function writeObservationNotes(observations, sessionId, writer, index, vaultDir)
338
350
  sacrificed: obs.sacrificed,
339
351
  tags: obs.tags
340
352
  });
341
- const relativePath = writer.writeMemory({
353
+ const relativePath = writer.writeSpore({
342
354
  id: obsId,
343
355
  observation_type: obs.type,
344
356
  session: sessionNoteId(sessionId),
@@ -352,11 +364,13 @@ function writeObservationNotes(observations, sessionId, writer, index, vaultDir)
352
364
  }
353
365
 
354
366
  export {
367
+ loadPrompt,
355
368
  buildSimilarityPrompt,
369
+ stripReasoningTokens,
356
370
  extractNumber,
357
371
  BufferProcessor,
358
372
  TranscriptMiner,
359
373
  extractTurnsFromBuffer,
360
374
  writeObservationNotes
361
375
  };
362
- //# sourceMappingURL=chunk-FFQNE6CT.js.map
376
+ //# sourceMappingURL=chunk-YZO22BBI.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/prompts/index.ts","../src/intelligence/response.ts","../src/daemon/processor.ts","../src/capture/transcript-miner.ts","../src/vault/observations.ts"],"sourcesContent":["/**\n * Prompt loader — reads .md templates from disk and interpolates variables.\n * Prompts are markdown files in this directory, not TypeScript strings.\n */\n\nimport fs from 'node:fs';\nimport path from 'node:path';\nimport { fileURLToPath } from 'node:url';\nimport { ARTIFACT_TYPES } from '../vault/types.js';\nimport { CANDIDATE_CONTENT_PREVIEW } from '../constants.js';\n\n/**\n * Resolve the prompts directory. With tsup code-splitting, import.meta.url\n * points to a chunk file (dist/chunk-XXXX.js), not dist/src/prompts/.\n * Walk up from the current file to find package.json, then use dist/src/prompts/.\n */\nfunction resolvePromptsDir(): string {\n let dir = path.dirname(fileURLToPath(import.meta.url));\n for (let i = 0; i < 5; i++) {\n if (fs.existsSync(path.join(dir, 'package.json'))) {\n return path.join(dir, 'dist', 'src', 'prompts');\n }\n // Also check if we're already in the right place (tsc output or dev mode)\n if (fs.existsSync(path.join(dir, 'extraction.md'))) {\n return dir;\n }\n dir = path.dirname(dir);\n }\n // Final fallback: adjacent to current file (works with tsc)\n return path.dirname(fileURLToPath(import.meta.url));\n}\n\nconst PROMPTS_DIR = resolvePromptsDir();\n\nconst promptCache = new Map<string, string>();\n\nexport function loadPrompt(name: string): string {\n let cached = promptCache.get(name);\n if (!cached) {\n cached = fs.readFileSync(path.join(PROMPTS_DIR, `${name}.md`), 'utf-8').trim();\n promptCache.set(name, cached);\n }\n return cached;\n}\n\nfunction interpolate(template: string, vars: Record<string, string>): string {\n let result = template;\n for (const [key, value] of Object.entries(vars)) {\n result = result.replaceAll(`{{${key}}}`, value);\n }\n return result;\n}\n\n// --- Prompt builders ---\n\nexport function buildExtractionPrompt(\n sessionId: string,\n eventCount: number,\n toolSummary: string,\n maxTokens?: number,\n): string {\n return interpolate(loadPrompt('extraction'), {\n sessionId,\n eventCount: String(eventCount),\n toolSummary,\n maxTokens: String(maxTokens ?? 2048),\n });\n}\n\nexport function buildSummaryPrompt(\n sessionId: string,\n user: string,\n content: string,\n maxTokens?: number,\n): string {\n return interpolate(loadPrompt('summary'), {\n sessionId,\n user,\n content,\n maxTokens: String(maxTokens ?? 1024),\n });\n}\n\nexport function buildTitlePrompt(\n summary: string,\n sessionId: string,\n): string {\n return interpolate(loadPrompt('title'), {\n summary,\n sessionId,\n });\n}\n\nconst ARTIFACT_TYPE_DESCRIPTIONS = [\n '\"spec\" — Design specifications, architecture documents',\n '\"plan\" — Implementation plans, roadmaps',\n '\"rfc\" — Requests for comment, proposals',\n '\"doc\" — Documentation, guides, READMEs',\n '\"other\" — Other substantive documents',\n];\n\nexport function buildSimilarityPrompt(\n currentSummary: string,\n candidateSummary: string,\n): string {\n return interpolate(loadPrompt('session-similarity'), {\n currentSummary,\n candidateSummary,\n });\n}\n\nexport function buildClassificationPrompt(\n sessionId: string,\n candidates: Array<{ path: string; content: string }>,\n maxTokens?: number,\n): string {\n const fileList = candidates\n .map((c) => {\n const truncated = c.content.slice(0, CANDIDATE_CONTENT_PREVIEW);\n return `### ${c.path}\\n\\`\\`\\`\\n${truncated}\\n\\`\\`\\``;\n })\n .join('\\n\\n');\n\n return interpolate(loadPrompt('classification'), {\n sessionId,\n fileList,\n artifactTypes: ARTIFACT_TYPE_DESCRIPTIONS.map((d) => `- ${d}`).join('\\n'),\n validTypes: ARTIFACT_TYPES.join('|'),\n maxTokens: String(maxTokens ?? 1024),\n });\n}\n","/**\n * Clean LLM response text before parsing.\n *\n * Reasoning models (DeepSeek, Qwen, GLM, etc.) embed chain-of-thought\n * in the response using special tags. These must be stripped before\n * JSON parsing or value extraction.\n */\n\n// Patterns for reasoning model chain-of-thought tokens.\n// Order matters: most specific patterns first.\nconst REASONING_PATTERNS = [\n // <think>...</think>answer (DeepSeek, Qwen, GLM, many others)\n /<think>[\\s\\S]*?<\\/think>\\s*/gi,\n // Implicit opening: reasoning...</think>answer (GLM-4.7 observed)\n /^[\\s\\S]*?<\\/think>\\s*/i,\n // <reasoning>...</reasoning>answer\n /<reasoning>[\\s\\S]*?<\\/reasoning>\\s*/gi,\n // <|thinking|>...<|/thinking|>answer\n /<\\|thinking\\|>[\\s\\S]*?<\\|\\/thinking\\|>\\s*/gi,\n // Plain-text \"Thinking Process:\" block followed by actual content\n // (Qwen 3.5 via LM Studio without native thinking mode)\n // Matches from \"Thinking Process:\" up to the last numbered step, then the synthesis follows\n /^Thinking Process:[\\s\\S]*?(?=\\n(?:## |# |\\*\\*[A-Z]))/i,\n];\n\n/**\n * Strip reasoning/chain-of-thought tokens from LLM response text.\n * Returns the final answer without the thinking process.\n */\nexport function stripReasoningTokens(text: string): string {\n if (!text) return text;\n\n for (const pattern of REASONING_PATTERNS) {\n const stripped = text.replace(pattern, '').trim();\n if (stripped && stripped !== text.trim()) {\n return stripped;\n }\n }\n\n return text;\n}\n\n/**\n * Extract JSON from an LLM response that may contain markdown fences,\n * reasoning tokens, or other wrapper text.\n *\n * Tries in order:\n * 1. Strip reasoning tokens\n * 2. Extract from ```json ... ``` code fences\n * 3. Find bare {...} JSON object\n * 4. Parse the cleaned text directly\n */\nexport function extractJson(text: string): unknown {\n const cleaned = stripReasoningTokens(text);\n\n // Try code fence extraction\n const fenceMatch = cleaned.match(/```(?:json)?\\s*\\n?([\\s\\S]*?)\\n?```/);\n if (fenceMatch) {\n return JSON.parse(fenceMatch[1].trim());\n }\n\n // Try bare JSON object\n const objectMatch = cleaned.match(/\\{[\\s\\S]*\\}/);\n if (objectMatch) {\n return JSON.parse(objectMatch[0]);\n }\n\n // Try direct parse\n return JSON.parse(cleaned);\n}\n\n/**\n * Extract a numeric value from an LLM response that may contain\n * reasoning tokens or extra text around the number.\n */\nexport function extractNumber(text: string): number {\n const cleaned = stripReasoningTokens(text).trim();\n const match = cleaned.match(/(\\d+\\.?\\d*)/);\n if (match) return parseFloat(match[1]);\n return parseFloat(cleaned);\n}\n","import { z } from 'zod';\nimport type { LlmProvider } from '../intelligence/llm.js';\nimport { ARTIFACT_TYPES } from '../vault/types.js';\nimport { estimateTokens, CHARS_PER_TOKEN, PROMPT_PREVIEW_CHARS, AI_RESPONSE_PREVIEW_CHARS, COMMAND_PREVIEW_CHARS } from '../constants.js';\nimport type { MycoConfig } from '../config/schema.js';\nimport type { ObservationType, ArtifactType } from '../vault/types.js';\nimport { buildExtractionPrompt, buildSummaryPrompt, buildTitlePrompt, buildClassificationPrompt } from '../prompts/index.js';\nimport { extractJson, stripReasoningTokens } from '../intelligence/response.js';\n\nexport interface Observation {\n type: ObservationType;\n title: string;\n content: string;\n tags: string[];\n root_cause?: string;\n fix?: string;\n rationale?: string;\n alternatives_rejected?: string;\n gained?: string;\n sacrificed?: string;\n}\n\nexport interface ProcessorResult {\n summary: string;\n observations: Observation[];\n degraded: boolean;\n}\n\nexport interface ClassifiedArtifact {\n source_path: string;\n artifact_type: ArtifactType;\n title: string;\n tags: string[];\n}\n\nconst ClassificationResponseSchema = z.object({\n artifacts: z.array(z.object({\n source_path: z.string(),\n artifact_type: z.enum(ARTIFACT_TYPES),\n title: z.string(),\n tags: z.array(z.string()).default([]),\n })).default([]),\n});\n\nexport class BufferProcessor {\n private extractionMaxTokens: number;\n private summaryMaxTokens: number;\n private titleMaxTokens: number;\n private classificationMaxTokens: number;\n\n constructor(private backend: LlmProvider, private contextWindow: number = 8192, captureConfig?: MycoConfig['capture']) {\n this.extractionMaxTokens = captureConfig?.extraction_max_tokens ?? 2048;\n this.summaryMaxTokens = captureConfig?.summary_max_tokens ?? 512;\n this.titleMaxTokens = captureConfig?.title_max_tokens ?? 32;\n this.classificationMaxTokens = captureConfig?.classification_max_tokens ?? 1024;\n }\n\n private truncateForContext(data: string, maxTokens: number): string {\n const available = this.contextWindow - maxTokens;\n const dataTokens = estimateTokens(data);\n if (dataTokens <= available) return data;\n const charBudget = available * CHARS_PER_TOKEN;\n return data.slice(0, charBudget);\n }\n\n async process(\n events: Array<Record<string, unknown>>,\n sessionId: string,\n ): Promise<ProcessorResult> {\n const rawPrompt = this.buildPromptForExtraction(events, sessionId);\n const prompt = this.truncateForContext(rawPrompt, this.extractionMaxTokens);\n\n try {\n const response = await this.backend.summarize(prompt, { maxTokens: this.extractionMaxTokens });\n const parsed = extractJson(response.text) as {\n summary: string;\n observations: Observation[];\n };\n\n return {\n summary: parsed.summary,\n observations: parsed.observations ?? [],\n degraded: false,\n };\n } catch (error) {\n return {\n summary: `LLM processing failed for session ${sessionId}. ${events.length} events captured. Error: ${(error as Error).message}`,\n observations: [],\n degraded: true,\n };\n }\n }\n\n private buildPromptForExtraction(\n events: Array<Record<string, unknown>>,\n sessionId: string,\n ): string {\n const toolSummary = this.summarizeEvents(events);\n return buildExtractionPrompt(sessionId, events.length, toolSummary, this.extractionMaxTokens);\n }\n\n async summarizeSession(\n conversationMarkdown: string,\n sessionId: string,\n user?: string,\n ): Promise<{ summary: string; title: string }> {\n const truncatedContent = this.truncateForContext(conversationMarkdown, this.summaryMaxTokens);\n const summaryPrompt = buildSummaryPrompt(sessionId, user ?? 'unknown', truncatedContent, this.summaryMaxTokens);\n\n let summaryText: string;\n try {\n const response = await this.backend.summarize(summaryPrompt, { maxTokens: this.summaryMaxTokens });\n summaryText = stripReasoningTokens(response.text);\n } catch (error) {\n summaryText = `Session ${sessionId} — summarization failed: ${(error as Error).message}`;\n }\n\n const titlePrompt = buildTitlePrompt(summaryText, sessionId);\n let title: string;\n try {\n const response = await this.backend.summarize(titlePrompt, { maxTokens: this.titleMaxTokens });\n title = stripReasoningTokens(response.text).trim();\n } catch {\n title = `Session ${sessionId}`;\n }\n\n return { summary: summaryText, title };\n }\n\n async classifyArtifacts(\n candidates: Array<{ path: string; content: string }>,\n sessionId: string,\n ): Promise<ClassifiedArtifact[]> {\n if (candidates.length === 0) return [];\n\n const prompt = this.buildPromptForClassification(candidates, sessionId);\n const response = await this.backend.summarize(prompt, { maxTokens: this.classificationMaxTokens });\n const raw = extractJson(response.text);\n const parsed = ClassificationResponseSchema.parse(raw);\n return parsed.artifacts;\n }\n\n private buildPromptForClassification(\n candidates: Array<{ path: string; content: string }>,\n sessionId: string,\n ): string {\n return buildClassificationPrompt(sessionId, candidates, this.classificationMaxTokens);\n }\n\n private summarizeEvents(events: Array<Record<string, unknown>>): string {\n const toolCounts = new Map<string, number>();\n const filesAccessed = new Set<string>();\n const prompts: string[] = [];\n const aiResponses: string[] = [];\n\n for (const event of events) {\n if (event.type === 'user_prompt') {\n const prompt = String(event.prompt ?? '');\n if (prompt) prompts.push(prompt.slice(0, PROMPT_PREVIEW_CHARS));\n continue;\n }\n\n if (event.type === 'ai_response') {\n const content = String(event.content ?? '');\n if (content) aiResponses.push(content.slice(0, AI_RESPONSE_PREVIEW_CHARS));\n continue;\n }\n\n // Hooks send tool_name/tool_input; also support legacy tool/input\n const tool = String(event.tool_name ?? event.tool ?? 'unknown');\n toolCounts.set(tool, (toolCounts.get(tool) ?? 0) + 1);\n\n const input = (event.tool_input ?? event.input) as Record<string, unknown> | undefined;\n if (input?.path) filesAccessed.add(String(input.path));\n if (input?.file_path) filesAccessed.add(String(input.file_path));\n if (input?.command) filesAccessed.add(`[cmd] ${String(input.command).slice(0, COMMAND_PREVIEW_CHARS)}`);\n }\n\n const lines: string[] = [];\n\n if (prompts.length > 0) {\n lines.push('### User Prompts');\n for (const p of prompts) {\n lines.push(`- \"${p}\"`);\n }\n }\n\n lines.push('\\n### Tool Usage');\n for (const [tool, count] of toolCounts) {\n lines.push(`- ${tool}: ${count} calls`);\n }\n\n if (filesAccessed.size > 0) {\n lines.push('\\n### Files Accessed');\n for (const file of filesAccessed) {\n lines.push(`- ${file}`);\n }\n }\n\n if (aiResponses.length > 0) {\n lines.push('\\n### AI Responses');\n for (const r of aiResponses) {\n lines.push(`- \"${r}\"`);\n }\n }\n\n return lines.join('\\n');\n }\n}\n","import { AgentRegistry } from '../agents/registry.js';\nimport type { AgentAdapter } from '../agents/adapter.js';\nimport { PROMPT_PREVIEW_CHARS } from '../constants.js';\nimport fs from 'node:fs';\n\n// Re-export TranscriptTurn from its canonical home in agents/adapter.ts\nexport type { TranscriptTurn } from '../agents/adapter.js';\nimport type { TranscriptTurn } from '../agents/adapter.js';\n\ninterface TranscriptConfig {\n /** Additional agent adapters to register (useful for testing or custom agents) */\n additionalAdapters?: AgentAdapter[];\n}\n\nexport class TranscriptMiner {\n private registry: AgentRegistry;\n\n constructor(config?: TranscriptConfig) {\n this.registry = new AgentRegistry(config?.additionalAdapters);\n }\n\n /**\n * Extract all conversation turns for a session.\n * Convenience wrapper — delegates to getAllTurnsWithSource.\n */\n getAllTurns(sessionId: string): TranscriptTurn[] {\n return this.getAllTurnsWithSource(sessionId).turns;\n }\n\n /**\n * Extract turns using the hook-provided transcript path first (fast, no scanning),\n * then fall back to adapter registry scanning if the path isn't provided.\n */\n getAllTurnsWithSource(sessionId: string, transcriptPath?: string): { turns: TranscriptTurn[]; source: string } {\n // Primary: use the path provided by the hook (no directory scanning needed)\n if (transcriptPath) {\n const result = this.registry.parseTurnsFromPath(transcriptPath);\n if (result) return result;\n }\n\n // Fallback: scan known agent directories\n const result = this.registry.getTranscriptTurns(sessionId);\n if (result) return result;\n return { turns: [], source: 'none' };\n }\n}\n\n/**\n * Build turns from buffer events — the fallback when no agent transcript is available.\n * Buffer events come from hooks (user_prompt, tool_use) and lack AI responses.\n * Turns will have prompts and tool counts but no aiResponse.\n */\nexport function extractTurnsFromBuffer(events: Array<Record<string, unknown>>): TranscriptTurn[] {\n const turns: TranscriptTurn[] = [];\n let current: TranscriptTurn | null = null;\n\n for (const event of events) {\n const type = event.type as string;\n if (type === 'user_prompt') {\n if (current) turns.push(current);\n current = {\n prompt: String(event.prompt ?? '').slice(0, PROMPT_PREVIEW_CHARS),\n toolCount: 0,\n timestamp: String(event.timestamp ?? new Date().toISOString()),\n };\n } else if (type === 'tool_use') {\n if (current) current.toolCount++;\n }\n }\n if (current) turns.push(current);\n return turns;\n}\n","import { formatSporeBody } from '../obsidian/formatter.js';\nimport { sessionNoteId } from './session-id.js';\nimport { indexNote } from '../index/rebuild.js';\nimport type { Observation } from '../daemon/processor.js';\nimport type { VaultWriter } from './writer.js';\nimport type { MycoIndex } from '../index/sqlite.js';\n\nexport interface WrittenNote {\n id: string;\n path: string;\n observation: Observation;\n}\n\nexport function writeObservationNotes(\n observations: Observation[],\n sessionId: string,\n writer: VaultWriter,\n index: MycoIndex,\n vaultDir: string,\n): WrittenNote[] {\n const results: WrittenNote[] = [];\n\n for (const obs of observations) {\n const obsId = `${obs.type}-${sessionId.slice(-6)}-${Date.now()}`;\n const body = formatSporeBody({\n title: obs.title,\n observationType: obs.type,\n content: obs.content,\n sessionId,\n root_cause: obs.root_cause,\n fix: obs.fix,\n rationale: obs.rationale,\n alternatives_rejected: obs.alternatives_rejected,\n gained: obs.gained,\n sacrificed: obs.sacrificed,\n tags: obs.tags,\n });\n const relativePath = writer.writeSpore({\n id: obsId,\n observation_type: obs.type,\n session: sessionNoteId(sessionId),\n tags: obs.tags,\n content: body,\n });\n indexNote(index, vaultDir, relativePath);\n results.push({ id: obsId, path: relativePath, observation: obs });\n }\n\n return results;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;AAKA,OAAO,QAAQ;AACf,OAAO,UAAU;AACjB,SAAS,qBAAqB;AAS9B,SAAS,oBAA4B;AACnC,MAAI,MAAM,KAAK,QAAQ,cAAc,YAAY,GAAG,CAAC;AACrD,WAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,QAAI,GAAG,WAAW,KAAK,KAAK,KAAK,cAAc,CAAC,GAAG;AACjD,aAAO,KAAK,KAAK,KAAK,QAAQ,OAAO,SAAS;AAAA,IAChD;AAEA,QAAI,GAAG,WAAW,KAAK,KAAK,KAAK,eAAe,CAAC,GAAG;AAClD,aAAO;AAAA,IACT;AACA,UAAM,KAAK,QAAQ,GAAG;AAAA,EACxB;AAEA,SAAO,KAAK,QAAQ,cAAc,YAAY,GAAG,CAAC;AACpD;AAEA,IAAM,cAAc,kBAAkB;AAEtC,IAAM,cAAc,oBAAI,IAAoB;AAErC,SAAS,WAAW,MAAsB;AAC/C,MAAI,SAAS,YAAY,IAAI,IAAI;AACjC,MAAI,CAAC,QAAQ;AACX,aAAS,GAAG,aAAa,KAAK,KAAK,aAAa,GAAG,IAAI,KAAK,GAAG,OAAO,EAAE,KAAK;AAC7E,gBAAY,IAAI,MAAM,MAAM;AAAA,EAC9B;AACA,SAAO;AACT;AAEA,SAAS,YAAY,UAAkB,MAAsC;AAC3E,MAAI,SAAS;AACb,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,IAAI,GAAG;AAC/C,aAAS,OAAO,WAAW,KAAK,GAAG,MAAM,KAAK;AAAA,EAChD;AACA,SAAO;AACT;AAIO,SAAS,sBACd,WACA,YACA,aACA,WACQ;AACR,SAAO,YAAY,WAAW,YAAY,GAAG;AAAA,IAC3C;AAAA,IACA,YAAY,OAAO,UAAU;AAAA,IAC7B;AAAA,IACA,WAAW,OAAO,aAAa,IAAI;AAAA,EACrC,CAAC;AACH;AAEO,SAAS,mBACd,WACA,MACA,SACA,WACQ;AACR,SAAO,YAAY,WAAW,SAAS,GAAG;AAAA,IACxC;AAAA,IACA;AAAA,IACA;AAAA,IACA,WAAW,OAAO,aAAa,IAAI;AAAA,EACrC,CAAC;AACH;AAEO,SAAS,iBACd,SACA,WACQ;AACR,SAAO,YAAY,WAAW,OAAO,GAAG;AAAA,IACtC;AAAA,IACA;AAAA,EACF,CAAC;AACH;AAEA,IAAM,6BAA6B;AAAA,EACjC;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,SAAS,sBACd,gBACA,kBACQ;AACR,SAAO,YAAY,WAAW,oBAAoB,GAAG;AAAA,IACnD;AAAA,IACA;AAAA,EACF,CAAC;AACH;AAEO,SAAS,0BACd,WACA,YACA,WACQ;AACR,QAAM,WAAW,WACd,IAAI,CAAC,MAAM;AACV,UAAM,YAAY,EAAE,QAAQ,MAAM,GAAG,yBAAyB;AAC9D,WAAO,OAAO,EAAE,IAAI;AAAA;AAAA,EAAa,SAAS;AAAA;AAAA,EAC5C,CAAC,EACA,KAAK,MAAM;AAEd,SAAO,YAAY,WAAW,gBAAgB,GAAG;AAAA,IAC/C;AAAA,IACA;AAAA,IACA,eAAe,2BAA2B,IAAI,CAAC,MAAM,KAAK,CAAC,EAAE,EAAE,KAAK,IAAI;AAAA,IACxE,YAAY,eAAe,KAAK,GAAG;AAAA,IACnC,WAAW,OAAO,aAAa,IAAI;AAAA,EACrC,CAAC;AACH;;;ACxHA,IAAM,qBAAqB;AAAA;AAAA,EAEzB;AAAA;AAAA,EAEA;AAAA;AAAA,EAEA;AAAA;AAAA,EAEA;AAAA;AAAA;AAAA;AAAA,EAIA;AACF;AAMO,SAAS,qBAAqB,MAAsB;AACzD,MAAI,CAAC,KAAM,QAAO;AAElB,aAAW,WAAW,oBAAoB;AACxC,UAAM,WAAW,KAAK,QAAQ,SAAS,EAAE,EAAE,KAAK;AAChD,QAAI,YAAY,aAAa,KAAK,KAAK,GAAG;AACxC,aAAO;AAAA,IACT;AAAA,EACF;AAEA,SAAO;AACT;AAYO,SAAS,YAAY,MAAuB;AACjD,QAAM,UAAU,qBAAqB,IAAI;AAGzC,QAAM,aAAa,QAAQ,MAAM,oCAAoC;AACrE,MAAI,YAAY;AACd,WAAO,KAAK,MAAM,WAAW,CAAC,EAAE,KAAK,CAAC;AAAA,EACxC;AAGA,QAAM,cAAc,QAAQ,MAAM,aAAa;AAC/C,MAAI,aAAa;AACf,WAAO,KAAK,MAAM,YAAY,CAAC,CAAC;AAAA,EAClC;AAGA,SAAO,KAAK,MAAM,OAAO;AAC3B;AAMO,SAAS,cAAc,MAAsB;AAClD,QAAM,UAAU,qBAAqB,IAAI,EAAE,KAAK;AAChD,QAAM,QAAQ,QAAQ,MAAM,aAAa;AACzC,MAAI,MAAO,QAAO,WAAW,MAAM,CAAC,CAAC;AACrC,SAAO,WAAW,OAAO;AAC3B;;;AC7CA,IAAM,+BAA+B,iBAAE,OAAO;AAAA,EAC5C,WAAW,iBAAE,MAAM,iBAAE,OAAO;AAAA,IAC1B,aAAa,iBAAE,OAAO;AAAA,IACtB,eAAe,iBAAE,KAAK,cAAc;AAAA,IACpC,OAAO,iBAAE,OAAO;AAAA,IAChB,MAAM,iBAAE,MAAM,iBAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;AAAA,EACtC,CAAC,CAAC,EAAE,QAAQ,CAAC,CAAC;AAChB,CAAC;AAEM,IAAM,kBAAN,MAAsB;AAAA,EAM3B,YAAoB,SAA8B,gBAAwB,MAAM,eAAuC;AAAnG;AAA8B;AAChD,SAAK,sBAAsB,eAAe,yBAAyB;AACnE,SAAK,mBAAmB,eAAe,sBAAsB;AAC7D,SAAK,iBAAiB,eAAe,oBAAoB;AACzD,SAAK,0BAA0B,eAAe,6BAA6B;AAAA,EAC7E;AAAA,EAVQ;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EASA,mBAAmB,MAAc,WAA2B;AAClE,UAAM,YAAY,KAAK,gBAAgB;AACvC,UAAM,aAAa,eAAe,IAAI;AACtC,QAAI,cAAc,UAAW,QAAO;AACpC,UAAM,aAAa,YAAY;AAC/B,WAAO,KAAK,MAAM,GAAG,UAAU;AAAA,EACjC;AAAA,EAEA,MAAM,QACJ,QACA,WAC0B;AAC1B,UAAM,YAAY,KAAK,yBAAyB,QAAQ,SAAS;AACjE,UAAM,SAAS,KAAK,mBAAmB,WAAW,KAAK,mBAAmB;AAE1E,QAAI;AACF,YAAM,WAAW,MAAM,KAAK,QAAQ,UAAU,QAAQ,EAAE,WAAW,KAAK,oBAAoB,CAAC;AAC7F,YAAM,SAAS,YAAY,SAAS,IAAI;AAKxC,aAAO;AAAA,QACL,SAAS,OAAO;AAAA,QAChB,cAAc,OAAO,gBAAgB,CAAC;AAAA,QACtC,UAAU;AAAA,MACZ;AAAA,IACF,SAAS,OAAO;AACd,aAAO;AAAA,QACL,SAAS,qCAAqC,SAAS,KAAK,OAAO,MAAM,4BAA6B,MAAgB,OAAO;AAAA,QAC7H,cAAc,CAAC;AAAA,QACf,UAAU;AAAA,MACZ;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,yBACN,QACA,WACQ;AACR,UAAM,cAAc,KAAK,gBAAgB,MAAM;AAC/C,WAAO,sBAAsB,WAAW,OAAO,QAAQ,aAAa,KAAK,mBAAmB;AAAA,EAC9F;AAAA,EAEA,MAAM,iBACJ,sBACA,WACA,MAC6C;AAC7C,UAAM,mBAAmB,KAAK,mBAAmB,sBAAsB,KAAK,gBAAgB;AAC5F,UAAM,gBAAgB,mBAAmB,WAAW,QAAQ,WAAW,kBAAkB,KAAK,gBAAgB;AAE9G,QAAI;AACJ,QAAI;AACF,YAAM,WAAW,MAAM,KAAK,QAAQ,UAAU,eAAe,EAAE,WAAW,KAAK,iBAAiB,CAAC;AACjG,oBAAc,qBAAqB,SAAS,IAAI;AAAA,IAClD,SAAS,OAAO;AACd,oBAAc,WAAW,SAAS,iCAA6B,MAAgB,OAAO;AAAA,IACxF;AAEA,UAAM,cAAc,iBAAiB,aAAa,SAAS;AAC3D,QAAI;AACJ,QAAI;AACF,YAAM,WAAW,MAAM,KAAK,QAAQ,UAAU,aAAa,EAAE,WAAW,KAAK,eAAe,CAAC;AAC7F,cAAQ,qBAAqB,SAAS,IAAI,EAAE,KAAK;AAAA,IACnD,QAAQ;AACN,cAAQ,WAAW,SAAS;AAAA,IAC9B;AAEA,WAAO,EAAE,SAAS,aAAa,MAAM;AAAA,EACvC;AAAA,EAEA,MAAM,kBACJ,YACA,WAC+B;AAC/B,QAAI,WAAW,WAAW,EAAG,QAAO,CAAC;AAErC,UAAM,SAAS,KAAK,6BAA6B,YAAY,SAAS;AACtE,UAAM,WAAW,MAAM,KAAK,QAAQ,UAAU,QAAQ,EAAE,WAAW,KAAK,wBAAwB,CAAC;AACjG,UAAM,MAAM,YAAY,SAAS,IAAI;AACrC,UAAM,SAAS,6BAA6B,MAAM,GAAG;AACrD,WAAO,OAAO;AAAA,EAChB;AAAA,EAEQ,6BACN,YACA,WACQ;AACR,WAAO,0BAA0B,WAAW,YAAY,KAAK,uBAAuB;AAAA,EACtF;AAAA,EAEQ,gBAAgB,QAAgD;AACtE,UAAM,aAAa,oBAAI,IAAoB;AAC3C,UAAM,gBAAgB,oBAAI,IAAY;AACtC,UAAM,UAAoB,CAAC;AAC3B,UAAM,cAAwB,CAAC;AAE/B,eAAW,SAAS,QAAQ;AAC1B,UAAI,MAAM,SAAS,eAAe;AAChC,cAAM,SAAS,OAAO,MAAM,UAAU,EAAE;AACxC,YAAI,OAAQ,SAAQ,KAAK,OAAO,MAAM,GAAG,oBAAoB,CAAC;AAC9D;AAAA,MACF;AAEA,UAAI,MAAM,SAAS,eAAe;AAChC,cAAM,UAAU,OAAO,MAAM,WAAW,EAAE;AAC1C,YAAI,QAAS,aAAY,KAAK,QAAQ,MAAM,GAAG,yBAAyB,CAAC;AACzE;AAAA,MACF;AAGA,YAAM,OAAO,OAAO,MAAM,aAAa,MAAM,QAAQ,SAAS;AAC9D,iBAAW,IAAI,OAAO,WAAW,IAAI,IAAI,KAAK,KAAK,CAAC;AAEpD,YAAM,QAAS,MAAM,cAAc,MAAM;AACzC,UAAI,OAAO,KAAM,eAAc,IAAI,OAAO,MAAM,IAAI,CAAC;AACrD,UAAI,OAAO,UAAW,eAAc,IAAI,OAAO,MAAM,SAAS,CAAC;AAC/D,UAAI,OAAO,QAAS,eAAc,IAAI,SAAS,OAAO,MAAM,OAAO,EAAE,MAAM,GAAG,qBAAqB,CAAC,EAAE;AAAA,IACxG;AAEA,UAAM,QAAkB,CAAC;AAEzB,QAAI,QAAQ,SAAS,GAAG;AACtB,YAAM,KAAK,kBAAkB;AAC7B,iBAAW,KAAK,SAAS;AACvB,cAAM,KAAK,MAAM,CAAC,GAAG;AAAA,MACvB;AAAA,IACF;AAEA,UAAM,KAAK,kBAAkB;AAC7B,eAAW,CAAC,MAAM,KAAK,KAAK,YAAY;AACtC,YAAM,KAAK,KAAK,IAAI,KAAK,KAAK,QAAQ;AAAA,IACxC;AAEA,QAAI,cAAc,OAAO,GAAG;AAC1B,YAAM,KAAK,sBAAsB;AACjC,iBAAW,QAAQ,eAAe;AAChC,cAAM,KAAK,KAAK,IAAI,EAAE;AAAA,MACxB;AAAA,IACF;AAEA,QAAI,YAAY,SAAS,GAAG;AAC1B,YAAM,KAAK,oBAAoB;AAC/B,iBAAW,KAAK,aAAa;AAC3B,cAAM,KAAK,MAAM,CAAC,GAAG;AAAA,MACvB;AAAA,IACF;AAEA,WAAO,MAAM,KAAK,IAAI;AAAA,EACxB;AACF;;;AClMO,IAAM,kBAAN,MAAsB;AAAA,EACnB;AAAA,EAER,YAAY,QAA2B;AACrC,SAAK,WAAW,IAAI,cAAc,QAAQ,kBAAkB;AAAA,EAC9D;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,YAAY,WAAqC;AAC/C,WAAO,KAAK,sBAAsB,SAAS,EAAE;AAAA,EAC/C;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,sBAAsB,WAAmB,gBAAsE;AAE7G,QAAI,gBAAgB;AAClB,YAAMA,UAAS,KAAK,SAAS,mBAAmB,cAAc;AAC9D,UAAIA,QAAQ,QAAOA;AAAA,IACrB;AAGA,UAAM,SAAS,KAAK,SAAS,mBAAmB,SAAS;AACzD,QAAI,OAAQ,QAAO;AACnB,WAAO,EAAE,OAAO,CAAC,GAAG,QAAQ,OAAO;AAAA,EACrC;AACF;AAOO,SAAS,uBAAuB,QAA0D;AAC/F,QAAM,QAA0B,CAAC;AACjC,MAAI,UAAiC;AAErC,aAAW,SAAS,QAAQ;AAC1B,UAAM,OAAO,MAAM;AACnB,QAAI,SAAS,eAAe;AAC1B,UAAI,QAAS,OAAM,KAAK,OAAO;AAC/B,gBAAU;AAAA,QACR,QAAQ,OAAO,MAAM,UAAU,EAAE,EAAE,MAAM,GAAG,oBAAoB;AAAA,QAChE,WAAW;AAAA,QACX,WAAW,OAAO,MAAM,cAAa,oBAAI,KAAK,GAAE,YAAY,CAAC;AAAA,MAC/D;AAAA,IACF,WAAW,SAAS,YAAY;AAC9B,UAAI,QAAS,SAAQ;AAAA,IACvB;AAAA,EACF;AACA,MAAI,QAAS,OAAM,KAAK,OAAO;AAC/B,SAAO;AACT;;;AC1DO,SAAS,sBACd,cACA,WACA,QACA,OACA,UACe;AACf,QAAM,UAAyB,CAAC;AAEhC,aAAW,OAAO,cAAc;AAC9B,UAAM,QAAQ,GAAG,IAAI,IAAI,IAAI,UAAU,MAAM,EAAE,CAAC,IAAI,KAAK,IAAI,CAAC;AAC9D,UAAM,OAAO,gBAAgB;AAAA,MAC3B,OAAO,IAAI;AAAA,MACX,iBAAiB,IAAI;AAAA,MACrB,SAAS,IAAI;AAAA,MACb;AAAA,MACA,YAAY,IAAI;AAAA,MAChB,KAAK,IAAI;AAAA,MACT,WAAW,IAAI;AAAA,MACf,uBAAuB,IAAI;AAAA,MAC3B,QAAQ,IAAI;AAAA,MACZ,YAAY,IAAI;AAAA,MAChB,MAAM,IAAI;AAAA,IACZ,CAAC;AACD,UAAM,eAAe,OAAO,WAAW;AAAA,MACrC,IAAI;AAAA,MACJ,kBAAkB,IAAI;AAAA,MACtB,SAAS,cAAc,SAAS;AAAA,MAChC,MAAM,IAAI;AAAA,MACV,SAAS;AAAA,IACX,CAAC;AACD,cAAU,OAAO,UAAU,YAAY;AACvC,YAAQ,KAAK,EAAE,IAAI,OAAO,MAAM,cAAc,aAAa,IAAI,CAAC;AAAA,EAClE;AAEA,SAAO;AACT;","names":["result"]}
@@ -1,7 +1,7 @@
1
1
  import { createRequire as __cr } from 'node:module'; const require = __cr(import.meta.url);
2
2
  import {
3
3
  STDIN_TIMEOUT_MS
4
- } from "./chunk-7VPJK56U.js";
4
+ } from "./chunk-JBD5KP5G.js";
5
5
 
6
6
  // src/hooks/read-stdin.ts
7
7
  function readStdin() {
@@ -18,4 +18,4 @@ function readStdin() {
18
18
  export {
19
19
  readStdin
20
20
  };
21
- //# sourceMappingURL=chunk-ZBNT6E22.js.map
21
+ //# sourceMappingURL=chunk-ZCBL5HER.js.map