@goondocks/myco 0.20.2 → 0.21.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (216) hide show
  1. package/bin/myco-run +68 -7
  2. package/dist/agent-eval-YK2VP2S4.js +356 -0
  3. package/dist/agent-eval-YK2VP2S4.js.map +1 -0
  4. package/dist/{agent-run-X25Q2A6T.js → agent-run-GEJBD2YD.js} +10 -8
  5. package/dist/{agent-run-X25Q2A6T.js.map → agent-run-GEJBD2YD.js.map} +1 -1
  6. package/dist/{agent-tasks-7B6OFERB.js → agent-tasks-5XSRGTRX.js} +10 -8
  7. package/dist/{agent-tasks-7B6OFERB.js.map → agent-tasks-5XSRGTRX.js.map} +1 -1
  8. package/dist/{chunk-OD4AA7PV.js → chunk-53RPGOEN.js} +56 -8
  9. package/dist/chunk-53RPGOEN.js.map +1 -0
  10. package/dist/chunk-54SXG5HF.js +26 -0
  11. package/dist/chunk-54SXG5HF.js.map +1 -0
  12. package/dist/{chunk-XATDZX7U.js → chunk-6ALVMIB4.js} +19 -5
  13. package/dist/{chunk-XATDZX7U.js.map → chunk-6ALVMIB4.js.map} +1 -1
  14. package/dist/{chunk-FLLBJLHM.js → chunk-6C6QZ4PM.js} +9 -5
  15. package/dist/chunk-6C6QZ4PM.js.map +1 -0
  16. package/dist/{chunk-CCRGY3QW.js → chunk-AUIXX33A.js} +24 -95
  17. package/dist/chunk-AUIXX33A.js.map +1 -0
  18. package/dist/chunk-CISWUP5W.js +101 -0
  19. package/dist/chunk-CISWUP5W.js.map +1 -0
  20. package/dist/{chunk-MYOZLMB2.js → chunk-DTWUHHFI.js} +576 -20
  21. package/dist/chunk-DTWUHHFI.js.map +1 -0
  22. package/dist/chunk-EEOJWLMP.js +582 -0
  23. package/dist/chunk-EEOJWLMP.js.map +1 -0
  24. package/dist/{chunk-XG5RRUYF.js → chunk-ENZR5NG7.js} +2 -2
  25. package/dist/{chunk-6RFZWV4R.js → chunk-FCJ5JV54.js} +1 -1
  26. package/dist/{chunk-6RFZWV4R.js.map → chunk-FCJ5JV54.js.map} +1 -1
  27. package/dist/{chunk-US4LNCAT.js → chunk-IPPMYQ2Y.js} +5 -1
  28. package/dist/chunk-IPPMYQ2Y.js.map +1 -0
  29. package/dist/{chunk-VVNL26WX.js → chunk-KTTSXYEK.js} +22 -10
  30. package/dist/chunk-KTTSXYEK.js.map +1 -0
  31. package/dist/chunk-LQIPXVDH.js +17 -0
  32. package/dist/chunk-LQIPXVDH.js.map +1 -0
  33. package/dist/{chunk-DCSGJ7W4.js → chunk-N2DGFACQ.js} +3 -3
  34. package/dist/chunk-N7Z3LUEZ.js +858 -0
  35. package/dist/chunk-N7Z3LUEZ.js.map +1 -0
  36. package/dist/{chunk-2PDWCDKY.js → chunk-NFO7BRCO.js} +10 -7
  37. package/dist/{chunk-2PDWCDKY.js.map → chunk-NFO7BRCO.js.map} +1 -1
  38. package/dist/{chunk-Q36VMZST.js → chunk-OTQH5KZW.js} +89 -38
  39. package/dist/chunk-OTQH5KZW.js.map +1 -0
  40. package/dist/chunk-OUJSQSKE.js +113 -0
  41. package/dist/chunk-OUJSQSKE.js.map +1 -0
  42. package/dist/chunk-OZ3FBAK5.js +50 -0
  43. package/dist/chunk-OZ3FBAK5.js.map +1 -0
  44. package/dist/chunk-QATYARI5.js +408 -0
  45. package/dist/chunk-QATYARI5.js.map +1 -0
  46. package/dist/{chunk-KESLPBKV.js → chunk-QDLVIW2O.js} +4 -4
  47. package/dist/{chunk-5XIVBO25.js → chunk-QLLBJEM7.js} +6 -2
  48. package/dist/chunk-QLLBJEM7.js.map +1 -0
  49. package/dist/{chunk-EVDQKYCG.js → chunk-RQSJLWP4.js} +13 -2
  50. package/dist/chunk-RQSJLWP4.js.map +1 -0
  51. package/dist/{chunk-BPRIYNLE.js → chunk-TKAJ3JVF.js} +3 -3
  52. package/dist/chunk-TSM6VESW.js +25 -0
  53. package/dist/chunk-TSM6VESW.js.map +1 -0
  54. package/dist/{chunk-6X2ERTQV.js → chunk-USVFEWYL.js} +6 -4
  55. package/dist/{chunk-6X2ERTQV.js.map → chunk-USVFEWYL.js.map} +1 -1
  56. package/dist/{chunk-JZGN33AY.js → chunk-VRI56337.js} +4 -4
  57. package/dist/chunk-X2IRGXGF.js +14103 -0
  58. package/dist/chunk-X2IRGXGF.js.map +1 -0
  59. package/dist/{chunk-FMRZ26U5.js → chunk-X3IGT5RV.js} +5 -2
  60. package/dist/{chunk-FMRZ26U5.js.map → chunk-X3IGT5RV.js.map} +1 -1
  61. package/dist/{chunk-KHT24OWC.js → chunk-YDUOSRGD.js} +8 -94
  62. package/dist/{chunk-KHT24OWC.js.map → chunk-YDUOSRGD.js.map} +1 -1
  63. package/dist/{chunk-NGROSFOH.js → chunk-Z66IT5KL.js} +14 -9
  64. package/dist/chunk-Z66IT5KL.js.map +1 -0
  65. package/dist/{cli-GGPWH4UO.js → cli-HSLIG7EX.js} +50 -43
  66. package/dist/cli-HSLIG7EX.js.map +1 -0
  67. package/dist/{client-YXQUTXVZ.js → client-Z43DNLJH.js} +4 -4
  68. package/dist/{config-OMCYHG2S.js → config-VC4ACP42.js} +6 -4
  69. package/dist/{config-OMCYHG2S.js.map → config-VC4ACP42.js.map} +1 -1
  70. package/dist/{detect-PXNM6TA7.js → detect-7NUD5B5R.js} +2 -2
  71. package/dist/{detect-providers-5KOPZ7J2.js → detect-providers-ILLQZROY.js} +4 -4
  72. package/dist/{doctor-5JXJ36KA.js → doctor-HJCWHAU4.js} +49 -16
  73. package/dist/doctor-HJCWHAU4.js.map +1 -0
  74. package/dist/executor-DO6QFC6G.js +45 -0
  75. package/dist/{init-LMYOVZAV.js → init-4KVK7W2E.js} +16 -14
  76. package/dist/{init-LMYOVZAV.js.map → init-4KVK7W2E.js.map} +1 -1
  77. package/dist/{installer-FS257JRZ.js → installer-N4UTEACX.js} +6 -4
  78. package/dist/{llm-TH4NLIRM.js → llm-AGVEF5XD.js} +5 -4
  79. package/dist/{loader-CQYTFHEW.js → loader-LX7TFRM6.js} +5 -3
  80. package/dist/{loader-NOMBJUPW.js → loader-UDNUMEDA.js} +5 -3
  81. package/dist/{main-YTBVRTBI.js → main-4J4QZZTZ.js} +2518 -656
  82. package/dist/main-4J4QZZTZ.js.map +1 -0
  83. package/dist/{open-HG2DX6RN.js → open-7TXJQM3H.js} +10 -8
  84. package/dist/{open-HG2DX6RN.js.map → open-7TXJQM3H.js.map} +1 -1
  85. package/dist/{post-compact-JSECI44W.js → post-compact-7AEFVCZS.js} +8 -8
  86. package/dist/{post-tool-use-POGPTJBA.js → post-tool-use-TZINWWDH.js} +11 -9
  87. package/dist/post-tool-use-TZINWWDH.js.map +1 -0
  88. package/dist/{post-tool-use-failure-OT7BFWQW.js → post-tool-use-failure-TCFEU2GI.js} +8 -8
  89. package/dist/{pre-compact-OXVODKH4.js → pre-compact-LO2VZCGR.js} +8 -8
  90. package/dist/{provider-check-43LAMSMH.js → provider-check-ZEV5P4KM.js} +4 -4
  91. package/dist/{registry-U4CHXK6R.js → registry-F3THYC5M.js} +6 -4
  92. package/dist/{remove-N7ZPELFU.js → remove-F77AAALE.js} +12 -10
  93. package/dist/{remove-N7ZPELFU.js.map → remove-F77AAALE.js.map} +1 -1
  94. package/dist/{restart-ADG5GBTB.js → restart-UEFDPMLT.js} +11 -9
  95. package/dist/{restart-ADG5GBTB.js.map → restart-UEFDPMLT.js.map} +1 -1
  96. package/dist/{search-AHZEUNRR.js → search-NHNVUAQQ.js} +11 -9
  97. package/dist/{search-AHZEUNRR.js.map → search-NHNVUAQQ.js.map} +1 -1
  98. package/dist/{server-AGVYZVP5.js → server-AZJSTQEK.js} +369 -270
  99. package/dist/server-AZJSTQEK.js.map +1 -0
  100. package/dist/{session-6IU4AXYP.js → session-3HLC5KOD.js} +11 -9
  101. package/dist/{session-6IU4AXYP.js.map → session-3HLC5KOD.js.map} +1 -1
  102. package/dist/{session-end-FT27DWYZ.js → session-end-FS46UARX.js} +7 -7
  103. package/dist/session-start-46KPFV2H.js +134 -0
  104. package/dist/session-start-46KPFV2H.js.map +1 -0
  105. package/dist/{setup-llm-77MP4I2G.js → setup-llm-JMWSNQ2C.js} +11 -9
  106. package/dist/{setup-llm-77MP4I2G.js.map → setup-llm-JMWSNQ2C.js.map} +1 -1
  107. package/dist/src/agent/definitions/agent.yaml +9 -5
  108. package/dist/src/agent/definitions/tasks/cortex-instructions.yaml +115 -0
  109. package/dist/src/agent/definitions/tasks/cortex-prompt-builder.yaml +67 -0
  110. package/dist/src/agent/definitions/tasks/digest-only.yaml +1 -1
  111. package/dist/src/agent/definitions/tasks/extract-only.yaml +1 -1
  112. package/dist/src/agent/definitions/tasks/review-session.yaml +10 -39
  113. package/dist/src/agent/definitions/tasks/skill-evolve.yaml +181 -25
  114. package/dist/src/agent/definitions/tasks/skill-generate.yaml +21 -7
  115. package/dist/src/agent/definitions/tasks/skill-survey.yaml +2 -6
  116. package/dist/src/agent/definitions/tasks/supersession-sweep.yaml +1 -1
  117. package/dist/src/agent/definitions/tasks/title-summary.yaml +12 -19
  118. package/dist/src/agent/definitions/tasks/{full-intelligence.yaml → vault-evolve.yaml} +74 -129
  119. package/dist/src/agent/definitions/tasks/vault-seed.yaml +370 -0
  120. package/dist/src/agent/prompts/agent.md +12 -38
  121. package/dist/src/cli.js +1 -1
  122. package/dist/src/daemon/main.js +1 -1
  123. package/dist/src/hooks/post-tool-use.js +1 -1
  124. package/dist/src/hooks/session-end.js +1 -1
  125. package/dist/src/hooks/session-start.js +1 -1
  126. package/dist/src/hooks/stop.js +1 -1
  127. package/dist/src/hooks/user-prompt-submit.js +1 -1
  128. package/dist/src/mcp/server.js +1 -1
  129. package/dist/src/symbionts/manifests/claude-code.yaml +4 -0
  130. package/dist/src/symbionts/manifests/opencode.yaml +7 -0
  131. package/dist/src/symbionts/manifests/pi.yaml +22 -0
  132. package/dist/src/symbionts/templates/agents-starter.md +1 -1
  133. package/dist/src/symbionts/templates/pi/package.json +6 -0
  134. package/dist/src/symbionts/templates/pi/plugin.ts +559 -0
  135. package/dist/{stats-NVPWOYTE.js → stats-MKMETHMA.js} +11 -9
  136. package/dist/{stats-NVPWOYTE.js.map → stats-MKMETHMA.js.map} +1 -1
  137. package/dist/{stop-ZPIKVLH4.js → stop-OUEX6KA4.js} +7 -7
  138. package/dist/{stop-failure-2PX67YJC.js → stop-failure-2BWVNZEG.js} +8 -8
  139. package/dist/{subagent-start-UUE6EHQD.js → subagent-start-J4VV6DEE.js} +8 -8
  140. package/dist/{subagent-stop-KQWWWPE6.js → subagent-stop-JMLVEPIA.js} +8 -8
  141. package/dist/{task-completed-WMHOFQ7B.js → task-completed-65CHMMKA.js} +8 -8
  142. package/dist/{team-LRZ6GTQK.js → team-U2LDKIS4.js} +7 -5
  143. package/dist/{turns-YFNI5CQC.js → turns-HU2CTZAP.js} +2 -2
  144. package/dist/ui/assets/index-BUGor9dk.js +842 -0
  145. package/dist/ui/assets/index-_OP4ifzH.css +1 -0
  146. package/dist/ui/index.html +2 -2
  147. package/dist/{update-O6V4RC4W.js → update-ZSHVXWSQ.js} +12 -10
  148. package/dist/{update-O6V4RC4W.js.map → update-ZSHVXWSQ.js.map} +1 -1
  149. package/dist/{user-prompt-submit-N36KUPHI.js → user-prompt-submit-APMO6FVU.js} +10 -9
  150. package/dist/{user-prompt-submit-N36KUPHI.js.map → user-prompt-submit-APMO6FVU.js.map} +1 -1
  151. package/dist/{verify-LXPV7NYG.js → verify-R76ZFJSZ.js} +8 -5
  152. package/dist/{verify-LXPV7NYG.js.map → verify-R76ZFJSZ.js.map} +1 -1
  153. package/dist/{version-XMPPJQHR.js → version-TXPPS3L5.js} +2 -2
  154. package/dist/version-TXPPS3L5.js.map +1 -0
  155. package/package.json +3 -1
  156. package/skills/myco/SKILL.md +16 -1
  157. package/skills/myco/references/cli-usage.md +1 -1
  158. package/skills/myco-rules/SKILL.md +94 -0
  159. package/skills/{rules → myco-rules}/references/rules-bad-example.md +1 -1
  160. package/skills/{rules → myco-rules}/references/rules-good-example.md +1 -1
  161. package/dist/chunk-4YFKBL3F.js +0 -195
  162. package/dist/chunk-4YFKBL3F.js.map +0 -1
  163. package/dist/chunk-5XIVBO25.js.map +0 -1
  164. package/dist/chunk-CCRGY3QW.js.map +0 -1
  165. package/dist/chunk-CUDIZJY7.js +0 -36
  166. package/dist/chunk-CUDIZJY7.js.map +0 -1
  167. package/dist/chunk-EVDQKYCG.js.map +0 -1
  168. package/dist/chunk-FLLBJLHM.js.map +0 -1
  169. package/dist/chunk-MYOZLMB2.js.map +0 -1
  170. package/dist/chunk-NGROSFOH.js.map +0 -1
  171. package/dist/chunk-OD4AA7PV.js.map +0 -1
  172. package/dist/chunk-Q36VMZST.js.map +0 -1
  173. package/dist/chunk-US4LNCAT.js.map +0 -1
  174. package/dist/chunk-UYMFCYBF.js +0 -2326
  175. package/dist/chunk-UYMFCYBF.js.map +0 -1
  176. package/dist/chunk-VVNL26WX.js.map +0 -1
  177. package/dist/cli-GGPWH4UO.js.map +0 -1
  178. package/dist/doctor-5JXJ36KA.js.map +0 -1
  179. package/dist/executor-HWW2QNZQ.js +0 -2472
  180. package/dist/executor-HWW2QNZQ.js.map +0 -1
  181. package/dist/main-YTBVRTBI.js.map +0 -1
  182. package/dist/post-tool-use-POGPTJBA.js.map +0 -1
  183. package/dist/server-AGVYZVP5.js.map +0 -1
  184. package/dist/session-start-LAFICHII.js +0 -189
  185. package/dist/session-start-LAFICHII.js.map +0 -1
  186. package/dist/src/agent/definitions/tasks/graph-maintenance.yaml +0 -93
  187. package/dist/ui/assets/index-C2JuNtRB.css +0 -1
  188. package/dist/ui/assets/index-JLVaQKV2.js +0 -832
  189. package/skills/myco-curate/SKILL.md +0 -86
  190. package/skills/rules/SKILL.md +0 -214
  191. /package/dist/{chunk-XG5RRUYF.js.map → chunk-ENZR5NG7.js.map} +0 -0
  192. /package/dist/{chunk-DCSGJ7W4.js.map → chunk-N2DGFACQ.js.map} +0 -0
  193. /package/dist/{chunk-KESLPBKV.js.map → chunk-QDLVIW2O.js.map} +0 -0
  194. /package/dist/{chunk-BPRIYNLE.js.map → chunk-TKAJ3JVF.js.map} +0 -0
  195. /package/dist/{chunk-JZGN33AY.js.map → chunk-VRI56337.js.map} +0 -0
  196. /package/dist/{client-YXQUTXVZ.js.map → client-Z43DNLJH.js.map} +0 -0
  197. /package/dist/{detect-PXNM6TA7.js.map → detect-7NUD5B5R.js.map} +0 -0
  198. /package/dist/{detect-providers-5KOPZ7J2.js.map → detect-providers-ILLQZROY.js.map} +0 -0
  199. /package/dist/{installer-FS257JRZ.js.map → executor-DO6QFC6G.js.map} +0 -0
  200. /package/dist/{llm-TH4NLIRM.js.map → installer-N4UTEACX.js.map} +0 -0
  201. /package/dist/{loader-CQYTFHEW.js.map → llm-AGVEF5XD.js.map} +0 -0
  202. /package/dist/{loader-NOMBJUPW.js.map → loader-LX7TFRM6.js.map} +0 -0
  203. /package/dist/{provider-check-43LAMSMH.js.map → loader-UDNUMEDA.js.map} +0 -0
  204. /package/dist/{post-compact-JSECI44W.js.map → post-compact-7AEFVCZS.js.map} +0 -0
  205. /package/dist/{post-tool-use-failure-OT7BFWQW.js.map → post-tool-use-failure-TCFEU2GI.js.map} +0 -0
  206. /package/dist/{pre-compact-OXVODKH4.js.map → pre-compact-LO2VZCGR.js.map} +0 -0
  207. /package/dist/{registry-U4CHXK6R.js.map → provider-check-ZEV5P4KM.js.map} +0 -0
  208. /package/dist/{team-LRZ6GTQK.js.map → registry-F3THYC5M.js.map} +0 -0
  209. /package/dist/{session-end-FT27DWYZ.js.map → session-end-FS46UARX.js.map} +0 -0
  210. /package/dist/{stop-ZPIKVLH4.js.map → stop-OUEX6KA4.js.map} +0 -0
  211. /package/dist/{stop-failure-2PX67YJC.js.map → stop-failure-2BWVNZEG.js.map} +0 -0
  212. /package/dist/{subagent-start-UUE6EHQD.js.map → subagent-start-J4VV6DEE.js.map} +0 -0
  213. /package/dist/{subagent-stop-KQWWWPE6.js.map → subagent-stop-JMLVEPIA.js.map} +0 -0
  214. /package/dist/{task-completed-WMHOFQ7B.js.map → task-completed-65CHMMKA.js.map} +0 -0
  215. /package/dist/{turns-YFNI5CQC.js.map → team-U2LDKIS4.js.map} +0 -0
  216. /package/dist/{version-XMPPJQHR.js.map → turns-HU2CTZAP.js.map} +0 -0
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/db/queries/cortex-instructions.ts","../src/db/queries/digest-extracts.ts","../src/db/queries/plans.ts","../src/db/queries/spores.ts","../src/context/cortex-brief.ts"],"sourcesContent":["import { getDatabase } from '@myco/db/client.js';\nimport { getTeamMachineId } from '@myco/daemon/team-context.js';\n\nconst CORTEX_INSTRUCTION_COLUMNS = [\n 'id',\n 'agent_id',\n 'content',\n 'input_hash',\n 'source_run_id',\n 'generated_at',\n 'machine_id',\n 'synced_at',\n] as const;\n\nconst SELECT_COLUMNS = CORTEX_INSTRUCTION_COLUMNS.join(', ');\nconst DEFAULT_CORTEX_INSTRUCTIONS_ID = 'session-start';\n\nexport interface CortexInstructionsUpsert {\n agent_id: string;\n content: string;\n input_hash: string;\n generated_at: number;\n id?: string;\n machine_id?: string;\n source_run_id?: string | null;\n}\n\nexport interface CortexInstructionsRow {\n id: string;\n agent_id: string;\n content: string;\n input_hash: string;\n source_run_id: string | null;\n generated_at: number;\n machine_id: string;\n synced_at: number | null;\n}\n\nfunction toCortexInstructionsRow(row: Record<string, unknown>): CortexInstructionsRow {\n return {\n id: row.id as string,\n agent_id: row.agent_id as string,\n content: row.content as string,\n input_hash: row.input_hash as string,\n source_run_id: (row.source_run_id as string) ?? null,\n generated_at: row.generated_at as number,\n machine_id: (row.machine_id as string) ?? 'local',\n synced_at: (row.synced_at as number) ?? null,\n };\n}\n\nexport function upsertCortexInstructions(input: CortexInstructionsUpsert): CortexInstructionsRow {\n const db = getDatabase();\n const id = input.id ?? `${input.agent_id}:${DEFAULT_CORTEX_INSTRUCTIONS_ID}`;\n\n const row = db.prepare(\n `INSERT INTO cortex_instructions (\n id, agent_id, content, input_hash, source_run_id, generated_at, machine_id\n ) VALUES (\n ?, ?, ?, ?, ?, ?, ?\n )\n ON CONFLICT (id) DO UPDATE SET\n content = EXCLUDED.content,\n input_hash = EXCLUDED.input_hash,\n source_run_id = EXCLUDED.source_run_id,\n generated_at = EXCLUDED.generated_at,\n machine_id = EXCLUDED.machine_id\n RETURNING ${SELECT_COLUMNS}`,\n ).get(\n id,\n input.agent_id,\n input.content,\n input.input_hash,\n input.source_run_id ?? null,\n input.generated_at,\n input.machine_id ?? getTeamMachineId(),\n ) as Record<string, unknown>;\n\n return toCortexInstructionsRow(row);\n}\n\nexport function getCortexInstructions(agentId: string): CortexInstructionsRow | null {\n const db = getDatabase();\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS}\n FROM cortex_instructions\n WHERE agent_id = ?\n ORDER BY generated_at DESC\n LIMIT 1`,\n ).get(agentId) as Record<string, unknown> | undefined;\n return row ? toCortexInstructionsRow(row) : null;\n}\n","/**\n * Digest extract CRUD query helpers.\n *\n * All functions obtain the SQLite instance internally via `getDatabase()`.\n * Queries use positional `?` placeholders throughout (better-sqlite3).\n */\n\nimport { getDatabase } from '@myco/db/client.js';\nimport { DIGEST_TIERS, epochSeconds } from '@myco/constants.js';\nimport { getTeamMachineId } from '@myco/daemon/team-context.js';\n\n// ---------------------------------------------------------------------------\n// Types\n// ---------------------------------------------------------------------------\n\n/** Fields required when upserting a digest extract. */\nexport interface DigestExtractUpsert {\n agent_id: string;\n tier: number;\n content: string;\n generated_at: number;\n machine_id?: string;\n}\n\n/**\n * Options that control whether the upsert actually writes and how the\n * revision history is recorded. Added in schema v15.\n */\nexport interface DigestExtractUpsertOptions {\n /**\n * When true, the upsert is a no-op: nothing is written, no revision is\n * recorded, and `null` is returned. Used by dry-run tooling so we can\n * preview writes without touching persistent state.\n */\n dryRun?: boolean;\n /**\n * Id of the agent_run that produced this write. Recorded on the\n * revision row so operators can roll a specific run back.\n */\n runId?: string | null;\n /**\n * Optional JSON-encoded metadata to store with the revision.\n */\n metadata?: string | null;\n}\n\n/** Row shape for entries in digest_extract_revisions. */\nexport interface DigestExtractRevisionRow {\n id: number;\n agent_id: string;\n tier: number;\n content: string;\n metadata: string | null;\n run_id: string | null;\n parent_revision_id: number | null;\n created_at: number;\n}\n\n/** Options accepted by rollbackDigestExtract. */\nexport interface RollbackDigestExtractOptions {\n revisionId: number;\n /** Id of the run performing the rollback (recorded on the new revision). */\n runId?: string | null;\n}\n\n/** Row shape returned from digest_extracts queries (all columns). */\nexport interface DigestExtractRow {\n id: number;\n agent_id: string;\n tier: number;\n content: string;\n substrate_hash: string | null;\n generated_at: number;\n machine_id: string;\n synced_at: number | null;\n}\n\n// ---------------------------------------------------------------------------\n// Column list\n// ---------------------------------------------------------------------------\n\nconst EXTRACT_COLUMNS = [\n 'id',\n 'agent_id',\n 'tier',\n 'content',\n 'substrate_hash',\n 'generated_at',\n 'machine_id',\n 'synced_at',\n] as const;\n\nconst SELECT_COLUMNS = EXTRACT_COLUMNS.join(', ');\n\n// ---------------------------------------------------------------------------\n// Helpers\n// ---------------------------------------------------------------------------\n\n/** Normalize a SQLite result row into a typed DigestExtractRow. */\nfunction toDigestExtractRow(row: Record<string, unknown>): DigestExtractRow {\n return {\n id: row.id as number,\n agent_id: row.agent_id as string,\n tier: row.tier as number,\n content: row.content as string,\n substrate_hash: (row.substrate_hash as string) ?? null,\n generated_at: row.generated_at as number,\n machine_id: (row.machine_id as string) ?? 'local',\n synced_at: (row.synced_at as number) ?? null,\n };\n}\n\n// ---------------------------------------------------------------------------\n// Public API\n// ---------------------------------------------------------------------------\n\n/**\n * Upsert a digest extract. Uses ON CONFLICT on (agent_id, tier).\n *\n * Schema v15 behaviour: when an existing row would be overwritten, the\n * prior content is copied into digest_extract_revisions (linked to the\n * previous revision if any) *before* the upsert runs. This makes the\n * revision log append-only and preserves the state the agent is replacing.\n *\n * When `options.dryRun === true`, the function is a no-op: nothing is\n * written to digest_extracts or digest_extract_revisions, and `null` is\n * returned. Call sites that care about the hydrated row should skip\n * follow-up reads when dry-running.\n */\nexport function upsertDigestExtract(\n data: DigestExtractUpsert,\n options: DigestExtractUpsertOptions = {},\n): DigestExtractRow | null {\n if (options.dryRun) return null;\n\n const db = getDatabase();\n\n // The revision snapshot and the live-row upsert MUST be atomic. Without\n // a transaction, a crash between the two writes would leave the revision\n // log out of sync with `digest_extracts` — the exact invariant this log\n // exists to guarantee. Matches the pattern used in sessions.ts /\n // skill-records.ts for multi-table writes.\n return db.transaction(() => {\n // Capture the row we're about to overwrite (if any) so we can copy it\n // into the revision history before mutating the live table.\n const existingRow = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM digest_extracts WHERE agent_id = ? AND tier = ?`,\n ).get(data.agent_id, data.tier) as Record<string, unknown> | undefined;\n\n if (existingRow) {\n const priorRevisionId = db.prepare(\n `SELECT id FROM digest_extract_revisions\n WHERE agent_id = ? AND tier = ?\n ORDER BY id DESC\n LIMIT 1`,\n ).get(data.agent_id, data.tier) as { id: number } | undefined;\n\n db.prepare(\n `INSERT INTO digest_extract_revisions\n (agent_id, tier, content, metadata, run_id, parent_revision_id, created_at)\n VALUES (?, ?, ?, ?, ?, ?, ?)`,\n ).run(\n data.agent_id,\n data.tier,\n existingRow.content as string,\n options.metadata ?? null,\n options.runId ?? null,\n priorRevisionId?.id ?? null,\n epochSeconds(),\n );\n }\n\n db.prepare(\n `INSERT INTO digest_extracts (agent_id, tier, content, generated_at)\n VALUES (?, ?, ?, ?)\n ON CONFLICT (agent_id, tier) DO UPDATE SET\n content = EXCLUDED.content,\n generated_at = EXCLUDED.generated_at`,\n ).run(data.agent_id, data.tier, data.content, data.generated_at);\n\n // Always look up by composite unique key — works for both insert and update cases.\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM digest_extracts WHERE agent_id = ? AND tier = ?`,\n ).get(data.agent_id, data.tier);\n\n return toDigestExtractRow(row as Record<string, unknown>);\n })();\n}\n\n/**\n * Get a digest extract for a specific agent and tier.\n *\n * @returns the extract row, or null if not found.\n */\nexport function getDigestExtract(\n agentId: string,\n tier: number,\n): DigestExtractRow | null {\n const db = getDatabase();\n\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM digest_extracts\n WHERE agent_id = ? AND tier = ?`,\n ).get(agentId, tier) as Record<string, unknown> | undefined;\n\n if (!row) return null;\n return toDigestExtractRow(row);\n}\n\n/**\n * List digest extracts for an agent, filtered to configured tiers, ordered by tier ASC.\n */\nexport function listDigestExtracts(\n agentId: string,\n): DigestExtractRow[] {\n const db = getDatabase();\n const tierPlaceholders = DIGEST_TIERS.map(() => '?').join(', ');\n\n const rows = db.prepare(\n `SELECT ${SELECT_COLUMNS}\n FROM digest_extracts\n WHERE agent_id = ? AND tier IN (${tierPlaceholders})\n ORDER BY tier ASC`,\n ).all(agentId, ...DIGEST_TIERS) as Record<string, unknown>[];\n\n return rows.map(toDigestExtractRow);\n}\n\n// ---------------------------------------------------------------------------\n// Revision history (schema v15)\n// ---------------------------------------------------------------------------\n\nconst REVISION_COLUMNS = [\n 'id',\n 'agent_id',\n 'tier',\n 'content',\n 'metadata',\n 'run_id',\n 'parent_revision_id',\n 'created_at',\n] as const;\n\nconst REVISION_SELECT = REVISION_COLUMNS.join(', ');\n\nfunction toRevisionRow(row: Record<string, unknown>): DigestExtractRevisionRow {\n return {\n id: row.id as number,\n agent_id: row.agent_id as string,\n tier: row.tier as number,\n content: row.content as string,\n metadata: (row.metadata as string) ?? null,\n run_id: (row.run_id as string) ?? null,\n parent_revision_id: (row.parent_revision_id as number) ?? null,\n created_at: row.created_at as number,\n };\n}\n\n/**\n * List revisions for a specific (agent_id, tier) pair, newest first.\n * Used by operators who want to roll back a digest to an earlier state.\n */\nexport function listDigestRevisions(\n options: { agentId: string; tier: number; limit?: number },\n): DigestExtractRevisionRow[] {\n const db = getDatabase();\n const limit = options.limit ?? 50;\n const rows = db.prepare(\n `SELECT ${REVISION_SELECT}\n FROM digest_extract_revisions\n WHERE agent_id = ? AND tier = ?\n ORDER BY created_at DESC, id DESC\n LIMIT ?`,\n ).all(options.agentId, options.tier, limit) as Record<string, unknown>[];\n return rows.map(toRevisionRow);\n}\n\n/** Result of a successful rollback. */\nexport interface RollbackDigestExtractResult {\n /** The restored digest_extracts row (content now matches the target revision). */\n row: DigestExtractRow;\n /**\n * Id of the newly-appended revision that captures the pre-rollback live\n * content (so the rollback itself is reversible). `null` when no live row\n * existed before the rollback (nothing to preserve).\n */\n newRevisionId: number | null;\n}\n\n/**\n * Restore an earlier revision's content back into digest_extracts, and\n * append a *new* revision row so the revision history remains append-only.\n *\n * The newly-appended revision captures what was live before the rollback\n * (so the rollback itself is reversible), with its parent set to the last\n * revision for (agent_id, tier).\n *\n * Returns the restored digest_extracts row plus the newly-minted revision\n * id, or null if the revision id doesn't exist.\n */\nexport function rollbackDigestExtract(\n options: RollbackDigestExtractOptions,\n): RollbackDigestExtractResult | null {\n const db = getDatabase();\n\n const revision = db.prepare(\n `SELECT ${REVISION_SELECT}\n FROM digest_extract_revisions\n WHERE id = ?`,\n ).get(options.revisionId) as Record<string, unknown> | undefined;\n\n if (!revision) return null;\n\n const agentId = revision.agent_id as string;\n const tier = revision.tier as number;\n const targetContent = revision.content as string;\n const now = epochSeconds();\n\n // Preservation of the pre-rollback state and the live-row restore must\n // be atomic — same invariant as `upsertDigestExtract`.\n return db.transaction(() => {\n // 1) Append a new revision that preserves the *current* live content\n // (pre-rollback state) so the rollback itself is reversible.\n const currentRow = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM digest_extracts WHERE agent_id = ? AND tier = ?`,\n ).get(agentId, tier) as Record<string, unknown> | undefined;\n\n let newRevisionId: number | null = null;\n if (currentRow) {\n const priorRevisionId = db.prepare(\n `SELECT id FROM digest_extract_revisions\n WHERE agent_id = ? AND tier = ?\n ORDER BY id DESC\n LIMIT 1`,\n ).get(agentId, tier) as { id: number } | undefined;\n\n const info = db.prepare(\n `INSERT INTO digest_extract_revisions\n (agent_id, tier, content, metadata, run_id, parent_revision_id, created_at)\n VALUES (?, ?, ?, ?, ?, ?, ?)`,\n ).run(\n agentId,\n tier,\n currentRow.content as string,\n JSON.stringify({ rollback_of: options.revisionId }),\n options.runId ?? null,\n priorRevisionId?.id ?? null,\n now,\n );\n newRevisionId = Number(info.lastInsertRowid);\n }\n\n // 2) Restore the target revision's content into the live row.\n db.prepare(\n `INSERT INTO digest_extracts (agent_id, tier, content, generated_at)\n VALUES (?, ?, ?, ?)\n ON CONFLICT (agent_id, tier) DO UPDATE SET\n content = EXCLUDED.content,\n generated_at = EXCLUDED.generated_at`,\n ).run(agentId, tier, targetContent, now);\n\n const restored = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM digest_extracts WHERE agent_id = ? AND tier = ?`,\n ).get(agentId, tier) as Record<string, unknown>;\n\n return {\n row: toDigestExtractRow(restored),\n newRevisionId,\n };\n })();\n}\n","/**\n * Plan CRUD query helpers.\n *\n * All functions obtain the SQLite instance internally via `getDatabase()`.\n * Queries use positional `?` placeholders throughout (better-sqlite3).\n */\n\nimport { getDatabase } from '@myco/db/client.js';\nimport { epochSeconds } from '@myco/constants.js';\nimport { getTeamMachineId, isTeamSyncEnabled } from '@myco/daemon/team-context.js';\nimport { enqueueOutbox } from '@myco/db/queries/team-outbox.js';\nimport { syncRow } from '@myco/db/queries/team-outbox.js';\n\n// ---------------------------------------------------------------------------\n// Constants\n// ---------------------------------------------------------------------------\n\n/** Default number of plans returned by listPlans when no limit given. */\nconst DEFAULT_LIST_LIMIT = 100;\n\n/** Default plan status for new plans. */\nconst DEFAULT_STATUS = 'active';\n\n/** Default processed flag for new plans. */\nconst DEFAULT_PROCESSED = 0;\n\n// ---------------------------------------------------------------------------\n// Types\n// ---------------------------------------------------------------------------\n\n/** Fields required (or optional) when inserting/upserting a plan. */\nexport interface PlanInsert {\n id: string;\n logical_key: string;\n created_at: number;\n status?: string;\n author?: string | null;\n title?: string | null;\n content?: string | null;\n source_path?: string | null;\n tags?: string | null;\n session_id?: string | null;\n prompt_batch_id?: number | null;\n content_hash?: string | null;\n processed?: number;\n updated_at?: number | null;\n machine_id?: string;\n}\n\n/** Row shape returned from plan queries. */\nexport interface PlanRow {\n id: string;\n logical_key: string;\n status: string;\n author: string | null;\n title: string | null;\n content: string | null;\n source_path: string | null;\n tags: string | null;\n session_id: string | null;\n prompt_batch_id: number | null;\n content_hash: string | null;\n processed: number;\n embedded: number;\n created_at: number;\n updated_at: number | null;\n machine_id: string;\n synced_at: number | null;\n}\n\n/** Filter options for `listPlans`. */\nexport interface ListPlansOptions {\n status?: string;\n limit?: number;\n}\n\n// ---------------------------------------------------------------------------\n// Column list\n// ---------------------------------------------------------------------------\n\nconst PLAN_COLUMNS = [\n 'id',\n 'logical_key',\n 'status',\n 'author',\n 'title',\n 'content',\n 'source_path',\n 'tags',\n 'session_id',\n 'prompt_batch_id',\n 'content_hash',\n 'processed',\n 'embedded',\n 'created_at',\n 'updated_at',\n 'machine_id',\n 'synced_at',\n] as const;\n\nconst SELECT_COLUMNS = PLAN_COLUMNS.join(', ');\n\n// ---------------------------------------------------------------------------\n// Helpers\n// ---------------------------------------------------------------------------\n\n/** Normalize a SQLite result row into a typed PlanRow. */\nfunction toPlanRow(row: Record<string, unknown>): PlanRow {\n return {\n id: row.id as string,\n logical_key: row.logical_key as string,\n status: row.status as string,\n author: (row.author as string) ?? null,\n title: (row.title as string) ?? null,\n content: (row.content as string) ?? null,\n source_path: (row.source_path as string) ?? null,\n tags: (row.tags as string) ?? null,\n session_id: (row.session_id as string) ?? null,\n prompt_batch_id: (row.prompt_batch_id as number) ?? null,\n content_hash: (row.content_hash as string) ?? null,\n processed: row.processed as number,\n embedded: (row.embedded as number) ?? 0,\n created_at: row.created_at as number,\n updated_at: (row.updated_at as number) ?? null,\n machine_id: (row.machine_id as string) ?? 'local',\n synced_at: (row.synced_at as number) ?? null,\n };\n}\n\n// ---------------------------------------------------------------------------\n// Public API\n// ---------------------------------------------------------------------------\n\n/**\n * Insert a plan or update it if the id already exists.\n *\n * On conflict the row is updated with the values from `data`.\n */\nexport function upsertPlan(data: PlanInsert): PlanRow {\n const db = getDatabase();\n\n db.prepare(\n `INSERT INTO plans (\n id, logical_key, status, author, title, content,\n source_path, tags, session_id, prompt_batch_id, content_hash,\n processed, created_at, updated_at, machine_id\n ) VALUES (\n ?, ?, ?, ?, ?, ?,\n ?, ?, ?, ?, ?,\n ?, ?, ?, ?\n )\n ON CONFLICT (logical_key) DO UPDATE SET\n id = EXCLUDED.id,\n status = EXCLUDED.status,\n author = EXCLUDED.author,\n title = EXCLUDED.title,\n content = EXCLUDED.content,\n source_path = EXCLUDED.source_path,\n tags = EXCLUDED.tags,\n session_id = EXCLUDED.session_id,\n prompt_batch_id = EXCLUDED.prompt_batch_id,\n content_hash = EXCLUDED.content_hash,\n processed = EXCLUDED.processed,\n updated_at = EXCLUDED.updated_at,\n embedded = CASE\n WHEN EXCLUDED.content_hash != plans.content_hash THEN 0\n ELSE plans.embedded\n END`,\n ).run(\n data.id,\n data.logical_key,\n data.status ?? DEFAULT_STATUS,\n data.author ?? null,\n data.title ?? null,\n data.content ?? null,\n data.source_path ?? null,\n data.tags ?? null,\n data.session_id ?? null,\n data.prompt_batch_id ?? null,\n data.content_hash ?? null,\n data.processed ?? DEFAULT_PROCESSED,\n data.created_at,\n data.updated_at ?? null,\n data.machine_id ?? getTeamMachineId(),\n );\n\n const row = toPlanRow(\n db.prepare(`SELECT ${SELECT_COLUMNS} FROM plans WHERE logical_key = ?`).get(data.logical_key) as Record<string, unknown>,\n );\n\n syncRow('plans', row);\n\n return row;\n}\n\n/**\n * Retrieve a single plan by id.\n *\n * @returns the plan row, or null if not found.\n */\nexport function getPlan(id: string): PlanRow | null {\n const db = getDatabase();\n\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM plans WHERE id = ?`,\n ).get(id) as Record<string, unknown> | undefined;\n\n if (!row) return null;\n return toPlanRow(row);\n}\n\n/**\n * Retrieve a single plan by logical key.\n *\n * @returns the plan row, or null if not found.\n */\nexport function getPlanByLogicalKey(logicalKey: string): PlanRow | null {\n const db = getDatabase();\n\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM plans WHERE logical_key = ?`,\n ).get(logicalKey) as Record<string, unknown> | undefined;\n\n if (!row) return null;\n return toPlanRow(row);\n}\n\n/**\n * Delete a single plan by id and enqueue a team-sync tombstone when enabled.\n *\n * @returns the deleted plan row, or null if not found.\n */\nexport function deletePlan(id: string): PlanRow | null {\n const db = getDatabase();\n const row = getPlan(id);\n if (!row) return null;\n\n const info = db.prepare(`DELETE FROM plans WHERE id = ?`).run(id);\n if (info.changes === 0) return null;\n\n if (isTeamSyncEnabled()) {\n enqueueOutbox({\n table_name: 'plans',\n row_id: row.id,\n operation: 'delete',\n payload: JSON.stringify({\n id: row.id,\n logical_key: row.logical_key,\n title: row.title,\n }),\n machine_id: getTeamMachineId(),\n created_at: epochSeconds(),\n });\n }\n\n return row;\n}\n\n/**\n * List plans with optional filters, ordered by created_at DESC.\n */\nexport function listPlans(\n options: ListPlansOptions = {},\n): PlanRow[] {\n const db = getDatabase();\n\n const conditions: string[] = [];\n const params: unknown[] = [];\n\n if (options.status !== undefined) {\n conditions.push(`status = ?`);\n params.push(options.status);\n }\n\n const where = conditions.length > 0 ? `WHERE ${conditions.join(' AND ')}` : '';\n const limit = options.limit ?? DEFAULT_LIST_LIMIT;\n\n params.push(limit);\n\n const rows = db.prepare(\n `SELECT ${SELECT_COLUMNS}\n FROM plans\n ${where}\n ORDER BY created_at DESC\n LIMIT ?`,\n ).all(...params) as Record<string, unknown>[];\n\n return rows.map(toPlanRow);\n}\n\n/**\n * List all plans associated with a specific session, ordered by created_at DESC.\n */\nexport function listPlansBySession(sessionId: string): PlanRow[] {\n const db = getDatabase();\n\n const rows = db.prepare(\n `SELECT ${SELECT_COLUMNS}\n FROM plans\n WHERE session_id = ?\n ORDER BY created_at DESC`,\n ).all(sessionId) as Record<string, unknown>[];\n\n return rows.map(toPlanRow);\n}\n","/**\n * Spore CRUD query helpers.\n *\n * All functions obtain the SQLite instance internally via `getDatabase()`.\n * Queries use positional `?` placeholders throughout (better-sqlite3).\n */\n\nimport { getDatabase } from '@myco/db/client.js';\nimport { getTeamMachineId } from '@myco/daemon/team-context.js';\nimport { syncRow } from '@myco/db/queries/team-outbox.js';\n\n// ---------------------------------------------------------------------------\n// Constants\n// ---------------------------------------------------------------------------\n\n/** Default number of spores returned by listSpores when no limit given. */\nconst DEFAULT_LIST_LIMIT = 100;\n\n/** Default spore status for new spores. */\nconst DEFAULT_STATUS = 'active';\n\n/** Default importance score for new spores. */\nexport const DEFAULT_IMPORTANCE = 5;\n\n// ---------------------------------------------------------------------------\n// Types\n// ---------------------------------------------------------------------------\n\n/** Fields required (or optional) when inserting a spore. */\nexport interface SporeInsert {\n id: string;\n agent_id: string;\n observation_type: string;\n content: string;\n created_at: number;\n session_id?: string | null;\n prompt_batch_id?: number | null;\n status?: string;\n context?: string | null;\n importance?: number;\n file_path?: string | null;\n tags?: string | null;\n content_hash?: string | null;\n properties?: string | null;\n updated_at?: number | null;\n machine_id?: string;\n}\n\n/** Row shape returned from spore queries (all columns). */\nexport interface SporeRow {\n id: string;\n agent_id: string;\n session_id: string | null;\n prompt_batch_id: number | null;\n observation_type: string;\n status: string;\n content: string;\n context: string | null;\n importance: number;\n file_path: string | null;\n tags: string | null;\n content_hash: string | null;\n properties: string | null;\n embedded: number;\n created_at: number;\n updated_at: number | null;\n machine_id: string;\n synced_at: number | null;\n}\n\n/** Filter options for `listSpores`. */\nexport interface ListSporesOptions {\n agent_id?: string;\n observation_type?: string;\n status?: string;\n session_id?: string;\n search?: string;\n /** Only return spores created after this epoch-seconds timestamp. */\n since?: number;\n limit?: number;\n offset?: number;\n /**\n * When explicitly `false`, exclude spores whose source session is still\n * `status = 'active'` — intelligence-task reads (agent tools, context\n * queries) should opt in to this. Defaults to permissive so UI listings\n * and prompt-time context injection keep seeing in-flight work. Explicit\n * `session_id` filters bypass this check: a direct lookup of one session's\n * spores is always permitted.\n */\n includeActive?: boolean;\n}\n\n// ---------------------------------------------------------------------------\n// Column list\n// ---------------------------------------------------------------------------\n\nconst SPORE_COLUMNS = [\n 'id',\n 'agent_id',\n 'session_id',\n 'prompt_batch_id',\n 'observation_type',\n 'status',\n 'content',\n 'context',\n 'importance',\n 'file_path',\n 'tags',\n 'content_hash',\n 'properties',\n 'embedded',\n 'created_at',\n 'updated_at',\n 'machine_id',\n 'synced_at',\n] as const;\n\nconst SELECT_COLUMNS = SPORE_COLUMNS.join(', ');\n\n// ---------------------------------------------------------------------------\n// Helpers\n// ---------------------------------------------------------------------------\n\n/** Normalize a SQLite result row into a typed SporeRow. */\nfunction toSporeRow(row: Record<string, unknown>): SporeRow {\n return {\n id: row.id as string,\n agent_id: row.agent_id as string,\n session_id: (row.session_id as string) ?? null,\n prompt_batch_id: (row.prompt_batch_id as number) ?? null,\n observation_type: row.observation_type as string,\n status: row.status as string,\n content: row.content as string,\n context: (row.context as string) ?? null,\n importance: row.importance as number,\n file_path: (row.file_path as string) ?? null,\n tags: (row.tags as string) ?? null,\n content_hash: (row.content_hash as string) ?? null,\n properties: (row.properties as string) ?? null,\n embedded: (row.embedded as number) ?? 0,\n created_at: row.created_at as number,\n updated_at: (row.updated_at as number) ?? null,\n machine_id: (row.machine_id as string) ?? 'local',\n synced_at: (row.synced_at as number) ?? null,\n };\n}\n\n// ---------------------------------------------------------------------------\n// Public API\n// ---------------------------------------------------------------------------\n\n/**\n * Insert a new spore.\n *\n * Requires a valid `agent_id` (foreign key to agents table).\n */\nexport function insertSpore(data: SporeInsert): SporeRow {\n const db = getDatabase();\n\n db.prepare(\n `INSERT INTO spores (\n id, agent_id, session_id, prompt_batch_id,\n observation_type, status, content, context,\n importance, file_path, tags, content_hash,\n properties, created_at, updated_at, machine_id\n ) VALUES (\n ?, ?, ?, ?,\n ?, ?, ?, ?,\n ?, ?, ?, ?,\n ?, ?, ?, ?\n )`,\n ).run(\n data.id,\n data.agent_id,\n data.session_id ?? null,\n data.prompt_batch_id ?? null,\n data.observation_type,\n data.status ?? DEFAULT_STATUS,\n data.content,\n data.context ?? null,\n data.importance ?? DEFAULT_IMPORTANCE,\n data.file_path ?? null,\n data.tags ?? null,\n data.content_hash ?? null,\n data.properties ?? null,\n data.created_at,\n data.updated_at ?? null,\n data.machine_id ?? getTeamMachineId(),\n );\n\n const row = toSporeRow(\n db.prepare(`SELECT ${SELECT_COLUMNS} FROM spores WHERE id = ?`).get(data.id) as Record<string, unknown>,\n );\n\n syncRow('spores', row);\n\n return row;\n}\n\n/**\n * Retrieve a single spore by id.\n *\n * @returns the spore row, or null if not found.\n */\nexport function getSpore(id: string): SporeRow | null {\n const db = getDatabase();\n\n const row = db.prepare(\n `SELECT ${SELECT_COLUMNS} FROM spores WHERE id = ?`,\n ).get(id) as Record<string, unknown> | undefined;\n\n if (!row) return null;\n return toSporeRow(row);\n}\n\n/**\n * List spores with optional filters, ordered by created_at DESC.\n */\n/** Build WHERE clause and bound params from spore filter options. */\nfunction buildSporeWhere(\n options: Omit<ListSporesOptions, 'limit' | 'offset'>,\n): { where: string; params: unknown[] } {\n const conditions: string[] = [];\n const params: unknown[] = [];\n\n if (options.agent_id !== undefined) {\n conditions.push(`agent_id = ?`);\n params.push(options.agent_id);\n }\n if (options.observation_type !== undefined) {\n conditions.push(`observation_type = ?`);\n params.push(options.observation_type);\n }\n if (options.status !== undefined) {\n conditions.push(`status = ?`);\n params.push(options.status);\n }\n if (options.session_id !== undefined) {\n conditions.push(`session_id = ?`);\n params.push(options.session_id);\n }\n if (options.search !== undefined && options.search.length > 0) {\n conditions.push(`(content LIKE ? OR observation_type LIKE ?)`);\n const pattern = `%${options.search}%`;\n params.push(pattern, pattern);\n }\n if (options.since !== undefined) {\n conditions.push('created_at > ?');\n params.push(options.since);\n }\n\n // Only exclude spores from in-flight sessions when the caller explicitly\n // asks for it (intelligence tasks). UI and hook-level context injection\n // leave this unset so they see everything. A direct session_id filter\n // bypasses the gate — that lookup is always permitted.\n if (options.includeActive === false && options.session_id === undefined) {\n conditions.push(\n `(session_id IS NULL OR EXISTS (SELECT 1 FROM sessions s WHERE s.id = spores.session_id AND s.status != 'active'))`,\n );\n }\n\n return {\n where: conditions.length > 0 ? `WHERE ${conditions.join(' AND ')}` : '',\n params,\n };\n}\n\n/**\n * List spores with optional filters, ordered by created_at DESC.\n */\nexport function listSpores(\n options: ListSporesOptions = {},\n): SporeRow[] {\n const db = getDatabase();\n const { where, params } = buildSporeWhere(options);\n const limit = options.limit ?? DEFAULT_LIST_LIMIT;\n const offset = options.offset ?? 0;\n\n const rows = db.prepare(\n `SELECT ${SELECT_COLUMNS}\n FROM spores\n ${where}\n ORDER BY created_at DESC\n LIMIT ?\n OFFSET ?`,\n ).all(...params, limit, offset) as Record<string, unknown>[];\n\n return rows.map(toSporeRow);\n}\n\n/**\n * Count spores matching optional filters (for pagination totals).\n */\nexport function countSpores(\n options: Omit<ListSporesOptions, 'limit' | 'offset'> = {},\n): number {\n const db = getDatabase();\n const { where, params } = buildSporeWhere(options);\n\n const row = db.prepare(\n `SELECT COUNT(*) as count FROM spores ${where}`,\n ).get(...params) as { count: number };\n\n return row.count;\n}\n\n/**\n * Count active spores created after a given timestamp.\n * Used by skill-evolve to detect new knowledge since last assessment.\n */\nexport function countSporesSince(sinceEpoch: number): number {\n const db = getDatabase();\n const row = db.prepare(\n `SELECT COUNT(*) as count FROM spores WHERE created_at > ? AND status = 'active'`,\n ).get(sinceEpoch) as { count: number };\n return row.count;\n}\n\n/**\n * List active spore IDs created after a given timestamp, ordered newest first.\n */\nexport function listSporeIdsSince(sinceEpoch: number, limit = 20): string[] {\n const db = getDatabase();\n const rows = db.prepare(\n `SELECT id FROM spores WHERE created_at > ? AND status = 'active' ORDER BY created_at DESC LIMIT ?`,\n ).all(sinceEpoch, limit) as Array<{ id: string }>;\n return rows.map(r => r.id);\n}\n\n/**\n * Update the status and updated_at timestamp of a spore.\n *\n * @returns the updated row, or null if the spore does not exist.\n */\nexport function updateSporeStatus(\n id: string,\n status: string,\n updatedAt: number,\n): SporeRow | null {\n const db = getDatabase();\n\n const info = db.prepare(\n `UPDATE spores\n SET status = ?, updated_at = ?\n WHERE id = ?`,\n ).run(status, updatedAt, id);\n\n if (info.changes === 0) return null;\n\n const row = toSporeRow(\n db.prepare(`SELECT ${SELECT_COLUMNS} FROM spores WHERE id = ?`).get(id) as Record<string, unknown>,\n );\n\n syncRow('spores', row);\n\n return row;\n}\n","/**\n * Cortex content assembly.\n *\n * Builds the material that the Cortex agent consumes and emits:\n * - Capability resolution (team/collective availability)\n * - Delivery-decision logic (inline vs session-start injection)\n * - Retrieval guidance derived from MCP tool definitions\n * - Instruction-input prompt for the `cortex-instructions` agent task\n *\n * Pure content layer — orchestration (agent run launch, snapshot reads,\n * prompt builder) lives in `@myco/daemon/cortex`.\n */\nimport { createHash } from 'node:crypto';\nimport type { MycoConfig } from '@myco/config/schema.js';\nimport {\n CONTENT_HASH_ALGORITHM,\n DEFAULT_AGENT_ID,\n DIGEST_FALLBACK_TIER,\n} from '@myco/constants.js';\nimport { getCortexInstructions } from '@myco/db/queries/cortex-instructions.js';\nimport { getDigestExtract } from '@myco/db/queries/digest-extracts.js';\nimport { listPlans } from '@myco/db/queries/plans.js';\nimport { listSessions } from '@myco/db/queries/sessions.js';\nimport { listSpores } from '@myco/db/queries/spores.js';\nimport type { TeamSyncClient } from '../daemon/team-sync.js';\nimport {\n TOOL_DEFINITIONS,\n COLLECTIVE_TOOL_DEFINITIONS,\n getToolCortexPriority,\n type ToolDefinition,\n} from '../mcp/tool-definitions.js';\n\nconst MAX_COLLECTIVE_CAPABILITY_LABELS = 4;\nconst ALL_CORTEX_TOOL_DEFINITIONS = [...TOOL_DEFINITIONS, ...COLLECTIVE_TOOL_DEFINITIONS];\n\nconst RECENT_SESSION_LIMIT = 5;\nconst RECENT_WISDOM_SPORE_LIMIT = 3;\nconst RECENT_DECISION_SPORE_LIMIT = 3;\nconst RECENT_DISCOVERY_SPORE_LIMIT = 3;\nconst RECENT_PLAN_LIMIT = 3;\nconst CONTENT_PREVIEW_MAX_CHARS = 360;\nconst DIGEST_EXCERPT_MAX_CHARS = 1800;\nconst JSON_INDENT = 2;\n\nexport const CORTEX_SKILLS_NOTE = 'Project and Myco skills are already registered with the agent separately. Tell the agent to use those skills directly when relevant, and do not instruct it to call `myco_skills`.';\n\n// ---------------------------------------------------------------------------\n// Capability resolution\n// ---------------------------------------------------------------------------\n\nexport interface CortexCapabilities {\n teamEnabled: boolean;\n collectiveConnected: boolean;\n collectiveCapabilities: string[];\n}\n\nexport interface CortexToolGuidance {\n tool: string;\n guidance: string;\n requiresTeam?: boolean;\n requiresCollective?: boolean;\n priority: number;\n}\n\nexport interface DeliveryDecision {\n inlineInstructions: boolean;\n reason: 'missing-symbiont' | 'session-start-supported' | 'session-start-disabled' | 'no-session-start';\n}\n\nfunction toCortexToolGuidance(\n tool: Pick<ToolDefinition, 'name' | 'cortex'>,\n): CortexToolGuidance | null {\n const cortex = tool.cortex;\n if (!cortex) return null;\n return {\n tool: tool.name,\n guidance: cortex.guidance,\n requiresTeam: cortex.requiresTeam,\n requiresCollective: cortex.requiresCollective,\n priority: getToolCortexPriority(tool),\n };\n}\n\nexport const RETRIEVAL_GUIDANCE: CortexToolGuidance[] = ALL_CORTEX_TOOL_DEFINITIONS\n .map(toCortexToolGuidance)\n .filter((entry): entry is CortexToolGuidance => entry !== null)\n .sort((left, right) => left.priority - right.priority);\n\nexport async function resolveCortexCapabilities(\n config: Pick<MycoConfig, 'team'>,\n getTeamClient?: () => TeamSyncClient | null,\n): Promise<CortexCapabilities> {\n const teamClient = getTeamClient?.() ?? null;\n const teamEnabled = Boolean(config.team.enabled && teamClient);\n let collectiveConnected = false;\n let collectiveCapabilities: string[] = [];\n\n if (teamEnabled && teamClient) {\n try {\n const status = await teamClient.getCollectiveStatus();\n collectiveConnected = Boolean(status?.connected);\n collectiveCapabilities = status?.capabilities ?? [];\n } catch {\n collectiveConnected = false;\n collectiveCapabilities = [];\n }\n }\n\n return {\n teamEnabled,\n collectiveConnected,\n collectiveCapabilities,\n };\n}\n\nexport function shouldInjectCortex(\n config: MycoConfig['context'],\n): boolean {\n return config.cortex_enabled;\n}\n\nexport function resolveInstructionDelivery(\n config: MycoConfig['context'],\n symbiont: {\n supportsSessionStartInjection: boolean;\n } | null,\n): DeliveryDecision {\n if (!symbiont) {\n return { inlineInstructions: true, reason: 'missing-symbiont' };\n }\n if (!config.cortex_enabled) {\n return { inlineInstructions: true, reason: 'session-start-disabled' };\n }\n if (symbiont.supportsSessionStartInjection) {\n return { inlineInstructions: false, reason: 'session-start-supported' };\n }\n return { inlineInstructions: true, reason: 'no-session-start' };\n}\n\nexport function buildCapabilitySummary(capabilities: CortexCapabilities): string[] {\n const summary = [\n capabilities.collectiveConnected\n ? 'Myco can retrieve local, team, and collective knowledge in this project.'\n : capabilities.teamEnabled\n ? 'Myco can retrieve local and shared team knowledge in this project.'\n : 'Myco can retrieve local project knowledge in this project.',\n 'Use only the currently available Myco MCP tools described below, and omit any surfaces that are offline.',\n ];\n\n if (capabilities.collectiveConnected && capabilities.collectiveCapabilities.length > 0) {\n const labels = capabilities.collectiveCapabilities.slice(0, MAX_COLLECTIVE_CAPABILITY_LABELS);\n const remaining = Math.max(\n 0,\n capabilities.collectiveCapabilities.length - MAX_COLLECTIVE_CAPABILITY_LABELS,\n );\n const suffix = remaining > 0 ? ` (+${remaining} more)` : '';\n summary.push(`Collective capabilities online: ${labels.join(', ')}${suffix}.`);\n }\n\n return summary;\n}\n\nexport function buildRetrievalGuidanceLines(capabilities: CortexCapabilities): string[] {\n const lines: string[] = [];\n\n for (const entry of RETRIEVAL_GUIDANCE) {\n if (entry.requiresTeam && !capabilities.teamEnabled) continue;\n if (entry.requiresCollective && !capabilities.collectiveConnected) continue;\n lines.push(`- \\`${entry.tool}\\`: ${entry.guidance}`);\n }\n\n return lines;\n}\n\n// ---------------------------------------------------------------------------\n// Instruction-input prompt assembly (for the `cortex-instructions` task)\n// ---------------------------------------------------------------------------\n\nfunction hashInput(value: unknown): string {\n return createHash(CONTENT_HASH_ALGORITHM)\n .update(JSON.stringify(value))\n .digest('hex');\n}\n\nfunction truncatePreview(text: string | null, maxChars: number = CONTENT_PREVIEW_MAX_CHARS): string | null {\n if (!text) return null;\n return text.length > maxChars\n ? `${text.slice(0, maxChars)}...`\n : text;\n}\n\nfunction formatRecentSessions(): string {\n const sessions = listSessions({\n includeActive: false,\n limit: RECENT_SESSION_LIMIT,\n });\n if (sessions.length === 0) return 'No recent sessions are available.';\n\n return sessions.map((session) => {\n const head = `- ${session.title ?? session.id}${session.branch ? ` (branch=${session.branch})` : ''}`;\n const body = truncatePreview(session.summary);\n return body ? `${head}\\n ${body}` : head;\n }).join('\\n');\n}\n\nfunction formatSporesOfType(\n observationType: 'wisdom' | 'decision' | 'discovery',\n limit: number,\n): string {\n const spores = listSpores({\n observation_type: observationType,\n includeActive: false,\n status: 'active',\n limit,\n });\n if (spores.length === 0) return `No recent ${observationType} spores are available.`;\n\n return spores.map((spore) => {\n const parts = [\n `- ${truncatePreview(spore.content)}`,\n spore.session_id ? `session=${spore.session_id}` : null,\n ].filter(Boolean);\n return parts.join(' — ');\n }).join('\\n');\n}\n\nfunction formatRecentPlans(): string {\n const plans = listPlans({\n status: 'active',\n limit: RECENT_PLAN_LIMIT,\n });\n if (plans.length === 0) return 'No active plans are available.';\n\n return plans.map((plan) => {\n const parts = [\n `- ${plan.title ?? plan.id}`,\n `status=${plan.status}`,\n truncatePreview(plan.content),\n ].filter(Boolean);\n return parts.join(' — ');\n }).join('\\n');\n}\n\nfunction formatDigestExcerpt(config: MycoConfig): string {\n const preferredTier = config.context.digest_tier;\n const extract =\n getDigestExtract(DEFAULT_AGENT_ID, preferredTier) ??\n getDigestExtract(DEFAULT_AGENT_ID, DIGEST_FALLBACK_TIER);\n if (!extract) return 'No current digest extract is available.';\n\n const excerpt = truncatePreview(extract.content, DIGEST_EXCERPT_MAX_CHARS) ?? '';\n return excerpt\n ? `Tier ${extract.tier} digest excerpt:\\n${excerpt}`\n : `Tier ${extract.tier} digest extract is empty.`;\n}\n\nexport interface CortexInstructionPayload {\n inputHash: string;\n instruction: string;\n}\n\nexport async function buildCortexInstructionsInput(\n config: MycoConfig,\n getTeamClient?: () => TeamSyncClient | null,\n): Promise<CortexInstructionPayload> {\n const capabilities = await resolveCortexCapabilities(config, getTeamClient);\n const capabilitySummary = buildCapabilitySummary(capabilities);\n const retrievalGuidance = buildRetrievalGuidanceLines(capabilities);\n const recentSessions = formatRecentSessions();\n const recentWisdomSpores = formatSporesOfType('wisdom', RECENT_WISDOM_SPORE_LIMIT);\n const recentDecisionSpores = formatSporesOfType('decision', RECENT_DECISION_SPORE_LIMIT);\n const recentDiscoverySpores = formatSporesOfType('discovery', RECENT_DISCOVERY_SPORE_LIMIT);\n const recentPlans = formatRecentPlans();\n const digestExcerpt = formatDigestExcerpt(config);\n const input = {\n context: {\n digest_tier: config.context.digest_tier,\n cortex_enabled: config.context.cortex_enabled,\n prompt_search: config.context.prompt_search,\n prompt_max_spores: config.context.prompt_max_spores,\n },\n capabilities,\n digestExcerpt,\n recentSessions,\n recentWisdomSpores,\n recentDecisionSpores,\n recentDiscoverySpores,\n recentPlans,\n skillsNote: CORTEX_SKILLS_NOTE,\n };\n\n return {\n inputHash: hashInput(input),\n instruction: [\n 'Author compact session-start instructions for another coding agent.',\n 'Focus on teaching how to use the highest-signal Myco tools correctly, especially retrieval and plan persistence.',\n 'Do not restate AGENTS.md or static installation details.',\n '',\n '## Runtime config',\n JSON.stringify(input.context, null, JSON_INDENT),\n '',\n '## Authoring requirements',\n '- Start with the heading `## Myco-Enabled Project`.',\n '- Follow the heading with one brief sentence explaining that Myco provides project memory, prior decisions, plans, and retrieval tools for this repository.',\n '- Teach the most useful current Myco MCP tool behavior, especially retrieval and plan persistence.',\n '- Use the recent vault activity below to mention live project hotspots when that improves usefulness.',\n `- ${CORTEX_SKILLS_NOTE}`,\n '- Keep the heading and description brief so most of the budget goes to retrieval guidance.',\n '- Keep the output compact and ready for direct injection.',\n '',\n '## Capability summary',\n ...capabilitySummary,\n '',\n '## Tool guidance to encode',\n ...retrievalGuidance,\n '',\n '## Current digest excerpt',\n digestExcerpt,\n '',\n '## Recent sessions',\n recentSessions,\n '',\n '## Recent wisdom spores',\n recentWisdomSpores,\n '',\n '## Recent decision spores',\n recentDecisionSpores,\n '',\n '## Recent discovery spores',\n recentDiscoverySpores,\n '',\n '## Active plans',\n recentPlans,\n ].join('\\n'),\n };\n}\n\nexport async function buildScheduledCortexInstruction(\n config: MycoConfig,\n getTeamClient?: () => TeamSyncClient | null,\n): Promise<CortexInstructionPayload | undefined> {\n const built = await buildCortexInstructionsInput(config, getTeamClient);\n const existing = getCortexInstructions(DEFAULT_AGENT_ID);\n if (existing?.input_hash === built.inputHash) {\n return undefined;\n }\n return built;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;AAGA,IAAM,6BAA6B;AAAA,EACjC;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAM,iBAAiB,2BAA2B,KAAK,IAAI;AAC3D,IAAM,iCAAiC;AAuBvC,SAAS,wBAAwB,KAAqD;AACpF,SAAO;AAAA,IACL,IAAI,IAAI;AAAA,IACR,UAAU,IAAI;AAAA,IACd,SAAS,IAAI;AAAA,IACb,YAAY,IAAI;AAAA,IAChB,eAAgB,IAAI,iBAA4B;AAAA,IAChD,cAAc,IAAI;AAAA,IAClB,YAAa,IAAI,cAAyB;AAAA,IAC1C,WAAY,IAAI,aAAwB;AAAA,EAC1C;AACF;AAEO,SAAS,yBAAyB,OAAwD;AAC/F,QAAM,KAAK,YAAY;AACvB,QAAM,KAAK,MAAM,MAAM,GAAG,MAAM,QAAQ,IAAI,8BAA8B;AAE1E,QAAM,MAAM,GAAG;AAAA,IACb;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iBAWa,cAAc;AAAA,EAC7B,EAAE;AAAA,IACA;AAAA,IACA,MAAM;AAAA,IACN,MAAM;AAAA,IACN,MAAM;AAAA,IACN,MAAM,iBAAiB;AAAA,IACvB,MAAM;AAAA,IACN,MAAM,cAAc,iBAAiB;AAAA,EACvC;AAEA,SAAO,wBAAwB,GAAG;AACpC;AAEO,SAAS,sBAAsB,SAA+C;AACnF,QAAM,KAAK,YAAY;AACvB,QAAM,MAAM,GAAG;AAAA,IACb,UAAU,cAAc;AAAA;AAAA;AAAA;AAAA;AAAA,EAK1B,EAAE,IAAI,OAAO;AACb,SAAO,MAAM,wBAAwB,GAAG,IAAI;AAC9C;;;ACVA,IAAM,kBAAkB;AAAA,EACtB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAMA,kBAAiB,gBAAgB,KAAK,IAAI;AAOhD,SAAS,mBAAmB,KAAgD;AAC1E,SAAO;AAAA,IACL,IAAI,IAAI;AAAA,IACR,UAAU,IAAI;AAAA,IACd,MAAM,IAAI;AAAA,IACV,SAAS,IAAI;AAAA,IACb,gBAAiB,IAAI,kBAA6B;AAAA,IAClD,cAAc,IAAI;AAAA,IAClB,YAAa,IAAI,cAAyB;AAAA,IAC1C,WAAY,IAAI,aAAwB;AAAA,EAC1C;AACF;AAmBO,SAAS,oBACd,MACA,UAAsC,CAAC,GACd;AACzB,MAAI,QAAQ,OAAQ,QAAO;AAE3B,QAAM,KAAK,YAAY;AAOvB,SAAO,GAAG,YAAY,MAAM;AAG1B,UAAM,cAAc,GAAG;AAAA,MACrB,UAAUA,eAAc;AAAA,IAC1B,EAAE,IAAI,KAAK,UAAU,KAAK,IAAI;AAE9B,QAAI,aAAa;AACf,YAAM,kBAAkB,GAAG;AAAA,QACzB;AAAA;AAAA;AAAA;AAAA,MAIF,EAAE,IAAI,KAAK,UAAU,KAAK,IAAI;AAE9B,SAAG;AAAA,QACD;AAAA;AAAA;AAAA,MAGF,EAAE;AAAA,QACA,KAAK;AAAA,QACL,KAAK;AAAA,QACL,YAAY;AAAA,QACZ,QAAQ,YAAY;AAAA,QACpB,QAAQ,SAAS;AAAA,QACjB,iBAAiB,MAAM;AAAA,QACvB,aAAa;AAAA,MACf;AAAA,IACF;AAEA,OAAG;AAAA,MACD;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EAAE,IAAI,KAAK,UAAU,KAAK,MAAM,KAAK,SAAS,KAAK,YAAY;AAG/D,UAAM,MAAM,GAAG;AAAA,MACb,UAAUA,eAAc;AAAA,IAC1B,EAAE,IAAI,KAAK,UAAU,KAAK,IAAI;AAE9B,WAAO,mBAAmB,GAA8B;AAAA,EAC1D,CAAC,EAAE;AACL;AAOO,SAAS,iBACd,SACA,MACyB;AACzB,QAAM,KAAK,YAAY;AAEvB,QAAM,MAAM,GAAG;AAAA,IACb,UAAUA,eAAc;AAAA;AAAA,EAE1B,EAAE,IAAI,SAAS,IAAI;AAEnB,MAAI,CAAC,IAAK,QAAO;AACjB,SAAO,mBAAmB,GAAG;AAC/B;AAKO,SAAS,mBACd,SACoB;AACpB,QAAM,KAAK,YAAY;AACvB,QAAM,mBAAmB,aAAa,IAAI,MAAM,GAAG,EAAE,KAAK,IAAI;AAE9D,QAAM,OAAO,GAAG;AAAA,IACd,UAAUA,eAAc;AAAA;AAAA,uCAEW,gBAAgB;AAAA;AAAA,EAErD,EAAE,IAAI,SAAS,GAAG,YAAY;AAE9B,SAAO,KAAK,IAAI,kBAAkB;AACpC;AAMA,IAAM,mBAAmB;AAAA,EACvB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAM,kBAAkB,iBAAiB,KAAK,IAAI;AAElD,SAAS,cAAc,KAAwD;AAC7E,SAAO;AAAA,IACL,IAAI,IAAI;AAAA,IACR,UAAU,IAAI;AAAA,IACd,MAAM,IAAI;AAAA,IACV,SAAS,IAAI;AAAA,IACb,UAAW,IAAI,YAAuB;AAAA,IACtC,QAAS,IAAI,UAAqB;AAAA,IAClC,oBAAqB,IAAI,sBAAiC;AAAA,IAC1D,YAAY,IAAI;AAAA,EAClB;AACF;AAMO,SAAS,oBACd,SAC4B;AAC5B,QAAM,KAAK,YAAY;AACvB,QAAM,QAAQ,QAAQ,SAAS;AAC/B,QAAM,OAAO,GAAG;AAAA,IACd,UAAU,eAAe;AAAA;AAAA;AAAA;AAAA;AAAA,EAK3B,EAAE,IAAI,QAAQ,SAAS,QAAQ,MAAM,KAAK;AAC1C,SAAO,KAAK,IAAI,aAAa;AAC/B;AAyBO,SAAS,sBACd,SACoC;AACpC,QAAM,KAAK,YAAY;AAEvB,QAAM,WAAW,GAAG;AAAA,IAClB,UAAU,eAAe;AAAA;AAAA;AAAA,EAG3B,EAAE,IAAI,QAAQ,UAAU;AAExB,MAAI,CAAC,SAAU,QAAO;AAEtB,QAAM,UAAU,SAAS;AACzB,QAAM,OAAO,SAAS;AACtB,QAAM,gBAAgB,SAAS;AAC/B,QAAM,MAAM,aAAa;AAIzB,SAAO,GAAG,YAAY,MAAM;AAG1B,UAAM,aAAa,GAAG;AAAA,MACpB,UAAUA,eAAc;AAAA,IAC1B,EAAE,IAAI,SAAS,IAAI;AAEnB,QAAI,gBAA+B;AACnC,QAAI,YAAY;AACd,YAAM,kBAAkB,GAAG;AAAA,QACzB;AAAA;AAAA;AAAA;AAAA,MAIF,EAAE,IAAI,SAAS,IAAI;AAEnB,YAAM,OAAO,GAAG;AAAA,QACd;AAAA;AAAA;AAAA,MAGF,EAAE;AAAA,QACA;AAAA,QACA;AAAA,QACA,WAAW;AAAA,QACX,KAAK,UAAU,EAAE,aAAa,QAAQ,WAAW,CAAC;AAAA,QAClD,QAAQ,SAAS;AAAA,QACjB,iBAAiB,MAAM;AAAA,QACvB;AAAA,MACF;AACA,sBAAgB,OAAO,KAAK,eAAe;AAAA,IAC7C;AAGA,OAAG;AAAA,MACD;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EAAE,IAAI,SAAS,MAAM,eAAe,GAAG;AAEvC,UAAM,WAAW,GAAG;AAAA,MAClB,UAAUA,eAAc;AAAA,IAC1B,EAAE,IAAI,SAAS,IAAI;AAEnB,WAAO;AAAA,MACL,KAAK,mBAAmB,QAAQ;AAAA,MAChC;AAAA,IACF;AAAA,EACF,CAAC,EAAE;AACL;;;AChWA,IAAM,qBAAqB;AAG3B,IAAM,iBAAiB;AAGvB,IAAM,oBAAoB;AAwD1B,IAAM,eAAe;AAAA,EACnB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAMC,kBAAiB,aAAa,KAAK,IAAI;AAO7C,SAAS,UAAU,KAAuC;AACxD,SAAO;AAAA,IACL,IAAI,IAAI;AAAA,IACR,aAAa,IAAI;AAAA,IACjB,QAAQ,IAAI;AAAA,IACZ,QAAS,IAAI,UAAqB;AAAA,IAClC,OAAQ,IAAI,SAAoB;AAAA,IAChC,SAAU,IAAI,WAAsB;AAAA,IACpC,aAAc,IAAI,eAA0B;AAAA,IAC5C,MAAO,IAAI,QAAmB;AAAA,IAC9B,YAAa,IAAI,cAAyB;AAAA,IAC1C,iBAAkB,IAAI,mBAA8B;AAAA,IACpD,cAAe,IAAI,gBAA2B;AAAA,IAC9C,WAAW,IAAI;AAAA,IACf,UAAW,IAAI,YAAuB;AAAA,IACtC,YAAY,IAAI;AAAA,IAChB,YAAa,IAAI,cAAyB;AAAA,IAC1C,YAAa,IAAI,cAAyB;AAAA,IAC1C,WAAY,IAAI,aAAwB;AAAA,EAC1C;AACF;AAWO,SAAS,WAAW,MAA2B;AACpD,QAAM,KAAK,YAAY;AAEvB,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EA0BF,EAAE;AAAA,IACA,KAAK;AAAA,IACL,KAAK;AAAA,IACL,KAAK,UAAU;AAAA,IACf,KAAK,UAAU;AAAA,IACf,KAAK,SAAS;AAAA,IACd,KAAK,WAAW;AAAA,IAChB,KAAK,eAAe;AAAA,IACpB,KAAK,QAAQ;AAAA,IACb,KAAK,cAAc;AAAA,IACnB,KAAK,mBAAmB;AAAA,IACxB,KAAK,gBAAgB;AAAA,IACrB,KAAK,aAAa;AAAA,IAClB,KAAK;AAAA,IACL,KAAK,cAAc;AAAA,IACnB,KAAK,cAAc,iBAAiB;AAAA,EACtC;AAEA,QAAM,MAAM;AAAA,IACV,GAAG,QAAQ,UAAUA,eAAc,mCAAmC,EAAE,IAAI,KAAK,WAAW;AAAA,EAC9F;AAEA,UAAQ,SAAS,GAAG;AAEpB,SAAO;AACT;AAOO,SAAS,QAAQ,IAA4B;AAClD,QAAM,KAAK,YAAY;AAEvB,QAAM,MAAM,GAAG;AAAA,IACb,UAAUA,eAAc;AAAA,EAC1B,EAAE,IAAI,EAAE;AAER,MAAI,CAAC,IAAK,QAAO;AACjB,SAAO,UAAU,GAAG;AACtB;AAOO,SAAS,oBAAoB,YAAoC;AACtE,QAAM,KAAK,YAAY;AAEvB,QAAM,MAAM,GAAG;AAAA,IACb,UAAUA,eAAc;AAAA,EAC1B,EAAE,IAAI,UAAU;AAEhB,MAAI,CAAC,IAAK,QAAO;AACjB,SAAO,UAAU,GAAG;AACtB;AAOO,SAAS,WAAW,IAA4B;AACrD,QAAM,KAAK,YAAY;AACvB,QAAM,MAAM,QAAQ,EAAE;AACtB,MAAI,CAAC,IAAK,QAAO;AAEjB,QAAM,OAAO,GAAG,QAAQ,gCAAgC,EAAE,IAAI,EAAE;AAChE,MAAI,KAAK,YAAY,EAAG,QAAO;AAE/B,MAAI,kBAAkB,GAAG;AACvB,kBAAc;AAAA,MACZ,YAAY;AAAA,MACZ,QAAQ,IAAI;AAAA,MACZ,WAAW;AAAA,MACX,SAAS,KAAK,UAAU;AAAA,QACtB,IAAI,IAAI;AAAA,QACR,aAAa,IAAI;AAAA,QACjB,OAAO,IAAI;AAAA,MACb,CAAC;AAAA,MACD,YAAY,iBAAiB;AAAA,MAC7B,YAAY,aAAa;AAAA,IAC3B,CAAC;AAAA,EACH;AAEA,SAAO;AACT;AAKO,SAAS,UACd,UAA4B,CAAC,GAClB;AACX,QAAM,KAAK,YAAY;AAEvB,QAAM,aAAuB,CAAC;AAC9B,QAAM,SAAoB,CAAC;AAE3B,MAAI,QAAQ,WAAW,QAAW;AAChC,eAAW,KAAK,YAAY;AAC5B,WAAO,KAAK,QAAQ,MAAM;AAAA,EAC5B;AAEA,QAAM,QAAQ,WAAW,SAAS,IAAI,SAAS,WAAW,KAAK,OAAO,CAAC,KAAK;AAC5E,QAAM,QAAQ,QAAQ,SAAS;AAE/B,SAAO,KAAK,KAAK;AAEjB,QAAM,OAAO,GAAG;AAAA,IACd,UAAUA,eAAc;AAAA;AAAA,OAErB,KAAK;AAAA;AAAA;AAAA,EAGV,EAAE,IAAI,GAAG,MAAM;AAEf,SAAO,KAAK,IAAI,SAAS;AAC3B;AAKO,SAAS,mBAAmB,WAA8B;AAC/D,QAAM,KAAK,YAAY;AAEvB,QAAM,OAAO,GAAG;AAAA,IACd,UAAUA,eAAc;AAAA;AAAA;AAAA;AAAA,EAI1B,EAAE,IAAI,SAAS;AAEf,SAAO,KAAK,IAAI,SAAS;AAC3B;;;AChSA,IAAMC,sBAAqB;AAG3B,IAAMC,kBAAiB;AAGhB,IAAM,qBAAqB;AA0ElC,IAAM,gBAAgB;AAAA,EACpB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAMC,kBAAiB,cAAc,KAAK,IAAI;AAO9C,SAAS,WAAW,KAAwC;AAC1D,SAAO;AAAA,IACL,IAAI,IAAI;AAAA,IACR,UAAU,IAAI;AAAA,IACd,YAAa,IAAI,cAAyB;AAAA,IAC1C,iBAAkB,IAAI,mBAA8B;AAAA,IACpD,kBAAkB,IAAI;AAAA,IACtB,QAAQ,IAAI;AAAA,IACZ,SAAS,IAAI;AAAA,IACb,SAAU,IAAI,WAAsB;AAAA,IACpC,YAAY,IAAI;AAAA,IAChB,WAAY,IAAI,aAAwB;AAAA,IACxC,MAAO,IAAI,QAAmB;AAAA,IAC9B,cAAe,IAAI,gBAA2B;AAAA,IAC9C,YAAa,IAAI,cAAyB;AAAA,IAC1C,UAAW,IAAI,YAAuB;AAAA,IACtC,YAAY,IAAI;AAAA,IAChB,YAAa,IAAI,cAAyB;AAAA,IAC1C,YAAa,IAAI,cAAyB;AAAA,IAC1C,WAAY,IAAI,aAAwB;AAAA,EAC1C;AACF;AAWO,SAAS,YAAY,MAA6B;AACvD,QAAM,KAAK,YAAY;AAEvB,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAWF,EAAE;AAAA,IACA,KAAK;AAAA,IACL,KAAK;AAAA,IACL,KAAK,cAAc;AAAA,IACnB,KAAK,mBAAmB;AAAA,IACxB,KAAK;AAAA,IACL,KAAK,UAAUD;AAAA,IACf,KAAK;AAAA,IACL,KAAK,WAAW;AAAA,IAChB,KAAK,cAAc;AAAA,IACnB,KAAK,aAAa;AAAA,IAClB,KAAK,QAAQ;AAAA,IACb,KAAK,gBAAgB;AAAA,IACrB,KAAK,cAAc;AAAA,IACnB,KAAK;AAAA,IACL,KAAK,cAAc;AAAA,IACnB,KAAK,cAAc,iBAAiB;AAAA,EACtC;AAEA,QAAM,MAAM;AAAA,IACV,GAAG,QAAQ,UAAUC,eAAc,2BAA2B,EAAE,IAAI,KAAK,EAAE;AAAA,EAC7E;AAEA,UAAQ,UAAU,GAAG;AAErB,SAAO;AACT;AAOO,SAAS,SAAS,IAA6B;AACpD,QAAM,KAAK,YAAY;AAEvB,QAAM,MAAM,GAAG;AAAA,IACb,UAAUA,eAAc;AAAA,EAC1B,EAAE,IAAI,EAAE;AAER,MAAI,CAAC,IAAK,QAAO;AACjB,SAAO,WAAW,GAAG;AACvB;AAMA,SAAS,gBACP,SACsC;AACtC,QAAM,aAAuB,CAAC;AAC9B,QAAM,SAAoB,CAAC;AAE3B,MAAI,QAAQ,aAAa,QAAW;AAClC,eAAW,KAAK,cAAc;AAC9B,WAAO,KAAK,QAAQ,QAAQ;AAAA,EAC9B;AACA,MAAI,QAAQ,qBAAqB,QAAW;AAC1C,eAAW,KAAK,sBAAsB;AACtC,WAAO,KAAK,QAAQ,gBAAgB;AAAA,EACtC;AACA,MAAI,QAAQ,WAAW,QAAW;AAChC,eAAW,KAAK,YAAY;AAC5B,WAAO,KAAK,QAAQ,MAAM;AAAA,EAC5B;AACA,MAAI,QAAQ,eAAe,QAAW;AACpC,eAAW,KAAK,gBAAgB;AAChC,WAAO,KAAK,QAAQ,UAAU;AAAA,EAChC;AACA,MAAI,QAAQ,WAAW,UAAa,QAAQ,OAAO,SAAS,GAAG;AAC7D,eAAW,KAAK,6CAA6C;AAC7D,UAAM,UAAU,IAAI,QAAQ,MAAM;AAClC,WAAO,KAAK,SAAS,OAAO;AAAA,EAC9B;AACA,MAAI,QAAQ,UAAU,QAAW;AAC/B,eAAW,KAAK,gBAAgB;AAChC,WAAO,KAAK,QAAQ,KAAK;AAAA,EAC3B;AAMA,MAAI,QAAQ,kBAAkB,SAAS,QAAQ,eAAe,QAAW;AACvE,eAAW;AAAA,MACT;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL,OAAO,WAAW,SAAS,IAAI,SAAS,WAAW,KAAK,OAAO,CAAC,KAAK;AAAA,IACrE;AAAA,EACF;AACF;AAKO,SAAS,WACd,UAA6B,CAAC,GAClB;AACZ,QAAM,KAAK,YAAY;AACvB,QAAM,EAAE,OAAO,OAAO,IAAI,gBAAgB,OAAO;AACjD,QAAM,QAAQ,QAAQ,SAASF;AAC/B,QAAM,SAAS,QAAQ,UAAU;AAEjC,QAAM,OAAO,GAAG;AAAA,IACd,UAAUE,eAAc;AAAA;AAAA,OAErB,KAAK;AAAA;AAAA;AAAA;AAAA,EAIV,EAAE,IAAI,GAAG,QAAQ,OAAO,MAAM;AAE9B,SAAO,KAAK,IAAI,UAAU;AAC5B;AAKO,SAAS,YACd,UAAuD,CAAC,GAChD;AACR,QAAM,KAAK,YAAY;AACvB,QAAM,EAAE,OAAO,OAAO,IAAI,gBAAgB,OAAO;AAEjD,QAAM,MAAM,GAAG;AAAA,IACb,wCAAwC,KAAK;AAAA,EAC/C,EAAE,IAAI,GAAG,MAAM;AAEf,SAAO,IAAI;AACb;AAiBO,SAAS,kBAAkB,YAAoB,QAAQ,IAAc;AAC1E,QAAM,KAAK,YAAY;AACvB,QAAM,OAAO,GAAG;AAAA,IACd;AAAA,EACF,EAAE,IAAI,YAAY,KAAK;AACvB,SAAO,KAAK,IAAI,OAAK,EAAE,EAAE;AAC3B;AAOO,SAAS,kBACd,IACA,QACA,WACiB;AACjB,QAAM,KAAK,YAAY;AAEvB,QAAM,OAAO,GAAG;AAAA,IACd;AAAA;AAAA;AAAA,EAGF,EAAE,IAAI,QAAQ,WAAW,EAAE;AAE3B,MAAI,KAAK,YAAY,EAAG,QAAO;AAE/B,QAAM,MAAM;AAAA,IACV,GAAG,QAAQ,UAAUC,eAAc,2BAA2B,EAAE,IAAI,EAAE;AAAA,EACxE;AAEA,UAAQ,UAAU,GAAG;AAErB,SAAO;AACT;;;ACxVA,SAAS,kBAAkB;AAoB3B,IAAM,mCAAmC;AACzC,IAAM,8BAA8B,CAAC,GAAG,kBAAkB,GAAG,2BAA2B;AAExF,IAAM,uBAAuB;AAC7B,IAAM,4BAA4B;AAClC,IAAM,8BAA8B;AACpC,IAAM,+BAA+B;AACrC,IAAM,oBAAoB;AAC1B,IAAM,4BAA4B;AAClC,IAAM,2BAA2B;AACjC,IAAM,cAAc;AAEb,IAAM,qBAAqB;AAyBlC,SAAS,qBACP,MAC2B;AAC3B,QAAM,SAAS,KAAK;AACpB,MAAI,CAAC,OAAQ,QAAO;AACpB,SAAO;AAAA,IACL,MAAM,KAAK;AAAA,IACX,UAAU,OAAO;AAAA,IACjB,cAAc,OAAO;AAAA,IACrB,oBAAoB,OAAO;AAAA,IAC3B,UAAU,sBAAsB,IAAI;AAAA,EACtC;AACF;AAEO,IAAM,qBAA2C,4BACrD,IAAI,oBAAoB,EACxB,OAAO,CAAC,UAAuC,UAAU,IAAI,EAC7D,KAAK,CAAC,MAAM,UAAU,KAAK,WAAW,MAAM,QAAQ;AAEvD,eAAsB,0BACpB,QACA,eAC6B;AAC7B,QAAM,aAAa,gBAAgB,KAAK;AACxC,QAAM,cAAc,QAAQ,OAAO,KAAK,WAAW,UAAU;AAC7D,MAAI,sBAAsB;AAC1B,MAAI,yBAAmC,CAAC;AAExC,MAAI,eAAe,YAAY;AAC7B,QAAI;AACF,YAAM,SAAS,MAAM,WAAW,oBAAoB;AACpD,4BAAsB,QAAQ,QAAQ,SAAS;AAC/C,+BAAyB,QAAQ,gBAAgB,CAAC;AAAA,IACpD,QAAQ;AACN,4BAAsB;AACtB,+BAAyB,CAAC;AAAA,IAC5B;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;AAEO,SAAS,mBACd,QACS;AACT,SAAO,OAAO;AAChB;AAEO,SAAS,2BACd,QACA,UAGkB;AAClB,MAAI,CAAC,UAAU;AACb,WAAO,EAAE,oBAAoB,MAAM,QAAQ,mBAAmB;AAAA,EAChE;AACA,MAAI,CAAC,OAAO,gBAAgB;AAC1B,WAAO,EAAE,oBAAoB,MAAM,QAAQ,yBAAyB;AAAA,EACtE;AACA,MAAI,SAAS,+BAA+B;AAC1C,WAAO,EAAE,oBAAoB,OAAO,QAAQ,0BAA0B;AAAA,EACxE;AACA,SAAO,EAAE,oBAAoB,MAAM,QAAQ,mBAAmB;AAChE;AAEO,SAAS,uBAAuB,cAA4C;AACjF,QAAM,UAAU;AAAA,IACd,aAAa,sBACT,6EACA,aAAa,cACX,uEACA;AAAA,IACN;AAAA,EACF;AAEA,MAAI,aAAa,uBAAuB,aAAa,uBAAuB,SAAS,GAAG;AACtF,UAAM,SAAS,aAAa,uBAAuB,MAAM,GAAG,gCAAgC;AAC5F,UAAM,YAAY,KAAK;AAAA,MACrB;AAAA,MACA,aAAa,uBAAuB,SAAS;AAAA,IAC/C;AACA,UAAM,SAAS,YAAY,IAAI,MAAM,SAAS,WAAW;AACzD,YAAQ,KAAK,mCAAmC,OAAO,KAAK,IAAI,CAAC,GAAG,MAAM,GAAG;AAAA,EAC/E;AAEA,SAAO;AACT;AAEO,SAAS,4BAA4B,cAA4C;AACtF,QAAM,QAAkB,CAAC;AAEzB,aAAW,SAAS,oBAAoB;AACtC,QAAI,MAAM,gBAAgB,CAAC,aAAa,YAAa;AACrD,QAAI,MAAM,sBAAsB,CAAC,aAAa,oBAAqB;AACnE,UAAM,KAAK,OAAO,MAAM,IAAI,OAAO,MAAM,QAAQ,EAAE;AAAA,EACrD;AAEA,SAAO;AACT;AAMA,SAAS,UAAU,OAAwB;AACzC,SAAO,WAAW,sBAAsB,EACrC,OAAO,KAAK,UAAU,KAAK,CAAC,EAC5B,OAAO,KAAK;AACjB;AAEA,SAAS,gBAAgB,MAAqB,WAAmB,2BAA0C;AACzG,MAAI,CAAC,KAAM,QAAO;AAClB,SAAO,KAAK,SAAS,WACjB,GAAG,KAAK,MAAM,GAAG,QAAQ,CAAC,QAC1B;AACN;AAEA,SAAS,uBAA+B;AACtC,QAAM,WAAW,aAAa;AAAA,IAC5B,eAAe;AAAA,IACf,OAAO;AAAA,EACT,CAAC;AACD,MAAI,SAAS,WAAW,EAAG,QAAO;AAElC,SAAO,SAAS,IAAI,CAAC,YAAY;AAC/B,UAAM,OAAO,KAAK,QAAQ,SAAS,QAAQ,EAAE,GAAG,QAAQ,SAAS,YAAY,QAAQ,MAAM,MAAM,EAAE;AACnG,UAAM,OAAO,gBAAgB,QAAQ,OAAO;AAC5C,WAAO,OAAO,GAAG,IAAI;AAAA,IAAO,IAAI,KAAK;AAAA,EACvC,CAAC,EAAE,KAAK,IAAI;AACd;AAEA,SAAS,mBACP,iBACA,OACQ;AACR,QAAM,SAAS,WAAW;AAAA,IACxB,kBAAkB;AAAA,IAClB,eAAe;AAAA,IACf,QAAQ;AAAA,IACR;AAAA,EACF,CAAC;AACD,MAAI,OAAO,WAAW,EAAG,QAAO,aAAa,eAAe;AAE5D,SAAO,OAAO,IAAI,CAAC,UAAU;AAC3B,UAAM,QAAQ;AAAA,MACZ,KAAK,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACnC,MAAM,aAAa,WAAW,MAAM,UAAU,KAAK;AAAA,IACrD,EAAE,OAAO,OAAO;AAChB,WAAO,MAAM,KAAK,UAAK;AAAA,EACzB,CAAC,EAAE,KAAK,IAAI;AACd;AAEA,SAAS,oBAA4B;AACnC,QAAM,QAAQ,UAAU;AAAA,IACtB,QAAQ;AAAA,IACR,OAAO;AAAA,EACT,CAAC;AACD,MAAI,MAAM,WAAW,EAAG,QAAO;AAE/B,SAAO,MAAM,IAAI,CAAC,SAAS;AACzB,UAAM,QAAQ;AAAA,MACZ,KAAK,KAAK,SAAS,KAAK,EAAE;AAAA,MAC1B,UAAU,KAAK,MAAM;AAAA,MACrB,gBAAgB,KAAK,OAAO;AAAA,IAC9B,EAAE,OAAO,OAAO;AAChB,WAAO,MAAM,KAAK,UAAK;AAAA,EACzB,CAAC,EAAE,KAAK,IAAI;AACd;AAEA,SAAS,oBAAoB,QAA4B;AACvD,QAAM,gBAAgB,OAAO,QAAQ;AACrC,QAAM,UACJ,iBAAiB,kBAAkB,aAAa,KAChD,iBAAiB,kBAAkB,oBAAoB;AACzD,MAAI,CAAC,QAAS,QAAO;AAErB,QAAM,UAAU,gBAAgB,QAAQ,SAAS,wBAAwB,KAAK;AAC9E,SAAO,UACH,QAAQ,QAAQ,IAAI;AAAA,EAAqB,OAAO,KAChD,QAAQ,QAAQ,IAAI;AAC1B;AAOA,eAAsB,6BACpB,QACA,eACmC;AACnC,QAAM,eAAe,MAAM,0BAA0B,QAAQ,aAAa;AAC1E,QAAM,oBAAoB,uBAAuB,YAAY;AAC7D,QAAM,oBAAoB,4BAA4B,YAAY;AAClE,QAAM,iBAAiB,qBAAqB;AAC5C,QAAM,qBAAqB,mBAAmB,UAAU,yBAAyB;AACjF,QAAM,uBAAuB,mBAAmB,YAAY,2BAA2B;AACvF,QAAM,wBAAwB,mBAAmB,aAAa,4BAA4B;AAC1F,QAAM,cAAc,kBAAkB;AACtC,QAAM,gBAAgB,oBAAoB,MAAM;AAChD,QAAM,QAAQ;AAAA,IACZ,SAAS;AAAA,MACP,aAAa,OAAO,QAAQ;AAAA,MAC5B,gBAAgB,OAAO,QAAQ;AAAA,MAC/B,eAAe,OAAO,QAAQ;AAAA,MAC9B,mBAAmB,OAAO,QAAQ;AAAA,IACpC;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,YAAY;AAAA,EACd;AAEA,SAAO;AAAA,IACL,WAAW,UAAU,KAAK;AAAA,IAC1B,aAAa;AAAA,MACX;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,KAAK,UAAU,MAAM,SAAS,MAAM,WAAW;AAAA,MAC/C;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,KAAK,kBAAkB;AAAA,MACvB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,GAAG;AAAA,MACH;AAAA,MACA;AAAA,MACA,GAAG;AAAA,MACH;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,EAAE,KAAK,IAAI;AAAA,EACb;AACF;AAEA,eAAsB,gCACpB,QACA,eAC+C;AAC/C,QAAM,QAAQ,MAAM,6BAA6B,QAAQ,aAAa;AACtE,QAAM,WAAW,sBAAsB,gBAAgB;AACvD,MAAI,UAAU,eAAe,MAAM,WAAW;AAC5C,WAAO;AAAA,EACT;AACA,SAAO;AACT;","names":["SELECT_COLUMNS","SELECT_COLUMNS","DEFAULT_LIST_LIMIT","DEFAULT_STATUS","SELECT_COLUMNS","SELECT_COLUMNS"]}
@@ -5,13 +5,13 @@ import {
5
5
  import {
6
6
  normalizeHookInput,
7
7
  readStdin
8
- } from "./chunk-DCSGJ7W4.js";
8
+ } from "./chunk-N2DGFACQ.js";
9
9
  import {
10
10
  resolveVaultDir
11
- } from "./chunk-CUDIZJY7.js";
11
+ } from "./chunk-TSM6VESW.js";
12
12
  import {
13
13
  DaemonClient
14
- } from "./chunk-VVNL26WX.js";
14
+ } from "./chunk-KTTSXYEK.js";
15
15
 
16
16
  // src/hooks/send-event.ts
17
17
  import fs from "fs";
@@ -23,12 +23,15 @@ async function sendEvent(hookName, buildEvent) {
23
23
  const rawInput = JSON.parse(await readStdin());
24
24
  const input = normalizeHookInput(rawInput);
25
25
  const event = buildEvent(input);
26
+ const eventWithContext = {
27
+ ...event,
28
+ transcript_path: input.transcriptPath
29
+ };
26
30
  const client = new DaemonClient(VAULT_DIR);
27
- const result = await client.post("/events", { ...event, session_id: input.sessionId, agent: input.agent });
31
+ const result = await client.post("/events", { ...eventWithContext, session_id: input.sessionId, agent: input.agent });
28
32
  if (!result.ok) {
29
33
  const buffer = new EventBuffer(path.join(VAULT_DIR, "buffer"), input.sessionId);
30
- const { session_id: _, ...bufferPayload } = event;
31
- buffer.append(bufferPayload);
34
+ buffer.append(eventWithContext);
32
35
  }
33
36
  } catch (error) {
34
37
  process.stderr.write(`[myco] ${hookName} error: ${error.message}
@@ -39,4 +42,4 @@ async function sendEvent(hookName, buildEvent) {
39
42
  export {
40
43
  sendEvent
41
44
  };
42
- //# sourceMappingURL=chunk-2PDWCDKY.js.map
45
+ //# sourceMappingURL=chunk-NFO7BRCO.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/hooks/send-event.ts"],"sourcesContent":["/**\n * Shared hook helper — sends an event to the daemon, buffers on failure.\n *\n * Every hook follows the same pattern: read stdin, POST to daemon /events,\n * buffer to disk if the daemon is unreachable. This helper extracts that\n * skeleton so each hook is a one-liner mapping input fields to event fields.\n */\n\nimport { DaemonClient } from './client.js';\nimport { readStdin } from './read-stdin.js';\nimport { normalizeHookInput, type NormalizedHookInput } from './normalize.js';\nimport { EventBuffer } from '../capture/buffer.js';\nimport { resolveVaultDir } from '../vault/resolve.js';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Read hook stdin, POST event to daemon, buffer on failure.\n *\n * @param hookName — used for error logging (e.g., 'subagent-start')\n * @param buildEvent — maps the normalized hook input to the event payload.\n * Receives a NormalizedHookInput with canonical field names.\n * Return the full event object (must include `type`).\n */\nexport async function sendEvent(\n hookName: string,\n buildEvent: (input: NormalizedHookInput) => Record<string, unknown>,\n): Promise<void> {\n const VAULT_DIR = resolveVaultDir();\n if (!fs.existsSync(path.join(VAULT_DIR, 'myco.yaml'))) return;\n\n try {\n const rawInput = JSON.parse(await readStdin()) as Record<string, unknown>;\n const input = normalizeHookInput(rawInput);\n\n const event = buildEvent(input);\n\n const client = new DaemonClient(VAULT_DIR);\n const result = await client.post('/events', { ...event, session_id: input.sessionId, agent: input.agent });\n\n if (!result.ok) {\n const buffer = new EventBuffer(path.join(VAULT_DIR, 'buffer'), input.sessionId);\n // Strip session_id from buffer entry — it's in the filename\n const { session_id: _, ...bufferPayload } = event;\n buffer.append(bufferPayload);\n }\n } catch (error) {\n process.stderr.write(`[myco] ${hookName} error: ${(error as Error).message}\\n`);\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;AAaA,OAAO,QAAQ;AACf,OAAO,UAAU;AAUjB,eAAsB,UACpB,UACA,YACe;AACf,QAAM,YAAY,gBAAgB;AAClC,MAAI,CAAC,GAAG,WAAW,KAAK,KAAK,WAAW,WAAW,CAAC,EAAG;AAEvD,MAAI;AACF,UAAM,WAAW,KAAK,MAAM,MAAM,UAAU,CAAC;AAC7C,UAAM,QAAQ,mBAAmB,QAAQ;AAEzC,UAAM,QAAQ,WAAW,KAAK;AAE9B,UAAM,SAAS,IAAI,aAAa,SAAS;AACzC,UAAM,SAAS,MAAM,OAAO,KAAK,WAAW,EAAE,GAAG,OAAO,YAAY,MAAM,WAAW,OAAO,MAAM,MAAM,CAAC;AAEzG,QAAI,CAAC,OAAO,IAAI;AACd,YAAM,SAAS,IAAI,YAAY,KAAK,KAAK,WAAW,QAAQ,GAAG,MAAM,SAAS;AAE9E,YAAM,EAAE,YAAY,GAAG,GAAG,cAAc,IAAI;AAC5C,aAAO,OAAO,aAAa;AAAA,IAC7B;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,OAAO,MAAM,UAAU,QAAQ,WAAY,MAAgB,OAAO;AAAA,CAAI;AAAA,EAChF;AACF;","names":[]}
1
+ {"version":3,"sources":["../src/hooks/send-event.ts"],"sourcesContent":["/**\n * Shared hook helper — sends an event to the daemon, buffers on failure.\n *\n * Every hook follows the same pattern: read stdin, POST to daemon /events,\n * buffer to disk if the daemon is unreachable. This helper extracts that\n * skeleton so each hook is a one-liner mapping input fields to event fields.\n */\n\nimport { DaemonClient } from './client.js';\nimport { readStdin } from './read-stdin.js';\nimport { normalizeHookInput, type NormalizedHookInput } from './normalize.js';\nimport { EventBuffer } from '../capture/buffer.js';\nimport { resolveVaultDir } from '../vault/resolve.js';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Read hook stdin, POST event to daemon, buffer on failure.\n *\n * @param hookName — used for error logging (e.g., 'subagent-start')\n * @param buildEvent — maps the normalized hook input to the event payload.\n * Receives a NormalizedHookInput with canonical field names.\n * Return the full event object (must include `type`).\n */\nexport async function sendEvent(\n hookName: string,\n buildEvent: (input: NormalizedHookInput) => Record<string, unknown>,\n): Promise<void> {\n const VAULT_DIR = resolveVaultDir();\n if (!fs.existsSync(path.join(VAULT_DIR, 'myco.yaml'))) return;\n\n try {\n const rawInput = JSON.parse(await readStdin()) as Record<string, unknown>;\n const input = normalizeHookInput(rawInput);\n\n const event = buildEvent(input);\n const eventWithContext = {\n ...event,\n transcript_path: input.transcriptPath,\n };\n\n const client = new DaemonClient(VAULT_DIR);\n const result = await client.post('/events', { ...eventWithContext, session_id: input.sessionId, agent: input.agent });\n\n if (!result.ok) {\n const buffer = new EventBuffer(path.join(VAULT_DIR, 'buffer'), input.sessionId);\n buffer.append(eventWithContext);\n }\n } catch (error) {\n process.stderr.write(`[myco] ${hookName} error: ${(error as Error).message}\\n`);\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;AAaA,OAAO,QAAQ;AACf,OAAO,UAAU;AAUjB,eAAsB,UACpB,UACA,YACe;AACf,QAAM,YAAY,gBAAgB;AAClC,MAAI,CAAC,GAAG,WAAW,KAAK,KAAK,WAAW,WAAW,CAAC,EAAG;AAEvD,MAAI;AACF,UAAM,WAAW,KAAK,MAAM,MAAM,UAAU,CAAC;AAC7C,UAAM,QAAQ,mBAAmB,QAAQ;AAEzC,UAAM,QAAQ,WAAW,KAAK;AAC9B,UAAM,mBAAmB;AAAA,MACvB,GAAG;AAAA,MACH,iBAAiB,MAAM;AAAA,IACzB;AAEA,UAAM,SAAS,IAAI,aAAa,SAAS;AACzC,UAAM,SAAS,MAAM,OAAO,KAAK,WAAW,EAAE,GAAG,kBAAkB,YAAY,MAAM,WAAW,OAAO,MAAM,MAAM,CAAC;AAEpH,QAAI,CAAC,OAAO,IAAI;AACd,YAAM,SAAS,IAAI,YAAY,KAAK,KAAK,WAAW,QAAQ,GAAG,MAAM,SAAS;AAC9E,aAAO,OAAO,gBAAgB;AAAA,IAChC;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,OAAO,MAAM,UAAU,QAAQ,WAAY,MAAgB,OAAO;AAAA,CAAI;AAAA,EAChF;AACF;","names":[]}
@@ -3,7 +3,7 @@ import {
3
3
  deepMerge,
4
4
  isPlainObject,
5
5
  loadMergedConfig
6
- } from "./chunk-OD4AA7PV.js";
6
+ } from "./chunk-53RPGOEN.js";
7
7
 
8
8
  // src/symbionts/installer.ts
9
9
  import fs3 from "fs";
@@ -200,7 +200,7 @@ var DEFAULT_AGENTS_STARTER = `# Project Rules
200
200
 
201
201
  <!-- This starter file was created by Myco. Replace it with your project's rules and conventions. -->
202
202
 
203
- Rules haven't been defined yet. Use the /rules skill to generate project rules, or edit this file directly.
203
+ Rules haven't been defined yet. Use the /myco-rules skill to add durable project rules, or edit this file directly.
204
204
  `;
205
205
  function isMycoHookCommand(command) {
206
206
  return command.includes(".agents/myco-run.cjs") || command.includes(".agents/myco-hook.cjs") || command.startsWith("myco-run");
@@ -259,9 +259,18 @@ var HOOK_GUARD_PROJECT_PATH = `.agents/${HOOK_GUARD_INSTALLED_FILENAME}`;
259
259
  var LEGACY_HOOK_GUARD_PATH = ".agents/myco-hook.cjs";
260
260
  var SKILLS_SUBDIR = "skills";
261
261
  var CANONICAL_SKILLS_DIR = ".agents/skills";
262
+ var LEGACY_BUILTIN_SKILL_NAMES = ["myco-curate", "rules"];
262
263
  var MYCO_MCP_SERVER_NAME = "myco";
263
- var MCP_ENV_PROJECT_ROOT_TOKEN = "{projectRoot}";
264
- var MCP_ENV_VAULT_DIR_TOKEN = "{vaultDir}";
264
+ function substituteMycoRunCommand(server, alias) {
265
+ const command = server.command;
266
+ if (command === "myco-run") {
267
+ return { ...server, command: alias };
268
+ }
269
+ if (Array.isArray(command) && command[0] === "myco-run") {
270
+ return { ...server, command: [alias, ...command.slice(1)] };
271
+ }
272
+ return server;
273
+ }
265
274
  var MYCO_PLUGIN_FILE_MARKER = "myco:plugin-marker";
266
275
  var HOOKS_FORMAT_PLUGIN_FILE = "plugin-file";
267
276
  var INSTRUCTIONS_STUB_MARKER = "Edit AGENTS.md, not this file";
@@ -771,11 +780,42 @@ var SymbiontInstaller = class {
771
780
  /** List skill directory names from the package root. Returns empty array if not found. */
772
781
  listSkillDirs() {
773
782
  try {
774
- return fs3.readdirSync(path3.join(this.packageRoot, SKILLS_SUBDIR), { withFileTypes: true }).filter((d) => d.isDirectory()).map((d) => d.name);
783
+ const skillsRoot = path3.join(this.packageRoot, SKILLS_SUBDIR);
784
+ return fs3.readdirSync(skillsRoot, { withFileTypes: true }).filter((d) => d.isDirectory()).filter((d) => fs3.existsSync(path3.join(skillsRoot, d.name, "SKILL.md"))).map((d) => d.name);
775
785
  } catch {
776
786
  return [];
777
787
  }
778
788
  }
789
+ /** Remove symlinks for retired built-in skills from older installs. */
790
+ cleanupLegacySkillSymlinks(currentSkillNames) {
791
+ const reg = this.manifest.registration;
792
+ if (!reg?.skillsTarget) return;
793
+ const staleSkillNames = LEGACY_BUILTIN_SKILL_NAMES.filter((name) => !currentSkillNames.includes(name));
794
+ if (staleSkillNames.length === 0) return;
795
+ const canonicalDir = path3.join(this.projectRoot, CANONICAL_SKILLS_DIR);
796
+ for (const name of staleSkillNames) {
797
+ try {
798
+ fs3.unlinkSync(path3.join(canonicalDir, name));
799
+ } catch {
800
+ }
801
+ if (reg.skillsTarget !== CANONICAL_SKILLS_DIR) {
802
+ try {
803
+ fs3.unlinkSync(path3.join(this.projectRoot, reg.skillsTarget, name));
804
+ } catch {
805
+ }
806
+ }
807
+ }
808
+ if (reg.skillsTarget !== CANONICAL_SKILLS_DIR) {
809
+ try {
810
+ fs3.rmdirSync(path3.join(this.projectRoot, reg.skillsTarget));
811
+ } catch {
812
+ }
813
+ }
814
+ try {
815
+ fs3.rmdirSync(canonicalDir);
816
+ } catch {
817
+ }
818
+ }
779
819
  /**
780
820
  * Reconcile Myco-owned skill entries in project .gitignore.
781
821
  * Computes the desired entry set, strips any existing Myco block
@@ -832,6 +872,12 @@ ${desired.join("\n")}
832
872
  legacyOwnedLines.add(`${reg.skillsTarget}/${name}`);
833
873
  }
834
874
  }
875
+ for (const name of LEGACY_BUILTIN_SKILL_NAMES) {
876
+ legacyOwnedLines.add(`${CANONICAL_SKILLS_DIR}/${name}`);
877
+ if (reg?.skillsTarget && reg.skillsTarget !== CANONICAL_SKILLS_DIR) {
878
+ legacyOwnedLines.add(`${reg.skillsTarget}/${name}`);
879
+ }
880
+ }
835
881
  const filtered = stripped.split("\n").filter((line) => !legacyOwnedLines.has(line));
836
882
  return filtered.join("\n").replace(/\n{3,}/g, "\n\n").trimEnd() + (filtered.length > 0 ? "\n" : "");
837
883
  }
@@ -940,44 +986,47 @@ ${desired.join("\n")}
940
986
  }
941
987
  buildMcpTemplate(template) {
942
988
  if (!template) return null;
943
- const overrides = this.resolveMcpLaunchOverrides();
944
- if (!overrides.cwd && Object.keys(overrides.env).length === 0) return template;
989
+ const reg = this.manifest.registration;
990
+ const cwd = reg?.mcpCwd;
991
+ const alias = reg?.substituteRuntimeCommand ? this.readRuntimeCommandAlias() : null;
992
+ if (!cwd && !alias) return template;
945
993
  return Object.fromEntries(
946
994
  Object.entries(template).map(([name, def]) => {
947
995
  if (!def || typeof def !== "object" || Array.isArray(def)) return [name, def];
948
- const server = def;
949
- const mergedEnv = {
950
- ...server.env && typeof server.env === "object" && !Array.isArray(server.env) ? server.env : {},
951
- ...overrides.env
952
- };
953
- return [
954
- name,
955
- {
956
- ...server,
957
- ...overrides.cwd ? { cwd: overrides.cwd } : {},
958
- ...Object.keys(mergedEnv).length > 0 ? { env: mergedEnv } : {}
959
- }
960
- ];
996
+ let next = { ...def };
997
+ if (cwd) next.cwd = cwd;
998
+ if (alias) next = substituteMycoRunCommand(next, alias);
999
+ return [name, next];
961
1000
  })
962
1001
  );
963
1002
  }
964
- resolveMcpLaunchOverrides() {
965
- const vaultDir = path3.join(this.projectRoot, ".myco");
966
- const registration = this.manifest.registration;
967
- const envEntries = Object.entries(registration?.mcpEnv ?? {});
968
- return {
969
- cwd: this.resolveMcpPlaceholderValue(registration?.mcpCwd, vaultDir),
970
- env: Object.fromEntries(
971
- envEntries.flatMap(([key, value]) => {
972
- const resolved = this.resolveMcpPlaceholderValue(value, vaultDir);
973
- return resolved ? [[key, resolved]] : [];
974
- })
975
- )
976
- };
977
- }
978
- resolveMcpPlaceholderValue(value, vaultDir) {
979
- if (!value) return value;
980
- return value.replaceAll(MCP_ENV_PROJECT_ROOT_TOKEN, this.projectRoot).replaceAll(MCP_ENV_VAULT_DIR_TOKEN, vaultDir);
1003
+ /**
1004
+ * Read `.myco/runtime.command` and return the alias (e.g. `myco-dev`).
1005
+ *
1006
+ * Same file the hook guard and the `bin/myco-run` launcher consult to
1007
+ * decide which binary answers for the project. Only symbionts whose
1008
+ * manifest opts in via `registration.substituteRuntimeCommand = true`
1009
+ * reach this path — most symbionts rely on runtime PATH resolution
1010
+ * inside `bin/myco-run` instead, which keeps the alias truly dynamic
1011
+ * (change `runtime.command`, next spawn picks it up, no re-install).
1012
+ * Opt-in is for hosts that reorder PATH so `myco-run` can't be reached
1013
+ * via `~/.local/bin` (opencode, and whatever future host shares that
1014
+ * behavior).
1015
+ *
1016
+ * Returns null when the file is missing, empty, or contains the
1017
+ * default (`myco`) — no substitution is needed in those cases.
1018
+ */
1019
+ readRuntimeCommandAlias() {
1020
+ try {
1021
+ const raw = fs3.readFileSync(
1022
+ path3.join(this.projectRoot, ".myco", "runtime.command"),
1023
+ "utf-8"
1024
+ ).trim();
1025
+ if (!raw || raw === "myco") return null;
1026
+ return raw;
1027
+ } catch {
1028
+ return null;
1029
+ }
981
1030
  }
982
1031
  /**
983
1032
  * Write MCP servers to a JSON config file under the manifest-configured key.
@@ -1021,6 +1070,7 @@ ${desired.join("\n")}
1021
1070
  if (!reg?.skillsTarget) return false;
1022
1071
  const skillNames = this.listSkillDirs();
1023
1072
  if (skillNames.length === 0) return false;
1073
+ this.cleanupLegacySkillSymlinks(skillNames);
1024
1074
  const skillsSrc = path3.join(this.packageRoot, SKILLS_SUBDIR);
1025
1075
  const canonicalDir = path3.join(this.projectRoot, CANONICAL_SKILLS_DIR);
1026
1076
  fs3.mkdirSync(canonicalDir, { recursive: true });
@@ -1337,8 +1387,9 @@ function ensureLocalSkillsGitignore(agentSkillsDir) {
1337
1387
  }
1338
1388
 
1339
1389
  export {
1390
+ isMycoHookGroup,
1340
1391
  MYCO_MCP_SERVER_NAME,
1341
1392
  SymbiontInstaller,
1342
1393
  syncSkillSymlinks
1343
1394
  };
1344
- //# sourceMappingURL=chunk-Q36VMZST.js.map
1395
+ //# sourceMappingURL=chunk-OTQH5KZW.js.map