@vpxa/kb 0.1.15 → 0.1.16

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (449) hide show
  1. package/README.md +62 -15
  2. package/package.json +5 -1
  3. package/packages/analyzers/dist/blast-radius-analyzer.d.ts +1 -2
  4. package/packages/analyzers/dist/blast-radius-analyzer.js +1 -2
  5. package/packages/analyzers/dist/dependency-analyzer.d.ts +2 -3
  6. package/packages/analyzers/dist/dependency-analyzer.js +3 -4
  7. package/packages/analyzers/dist/diagram-generator.d.ts +1 -2
  8. package/packages/analyzers/dist/diagram-generator.js +1 -2
  9. package/packages/analyzers/dist/entry-point-analyzer.d.ts +1 -2
  10. package/packages/analyzers/dist/entry-point-analyzer.js +3 -4
  11. package/packages/analyzers/dist/knowledge-producer.d.ts +6 -2
  12. package/packages/analyzers/dist/knowledge-producer.js +4 -3
  13. package/packages/analyzers/dist/pattern-analyzer.d.ts +1 -2
  14. package/packages/analyzers/dist/pattern-analyzer.js +2 -3
  15. package/packages/analyzers/dist/regex-call-graph.d.ts +1 -2
  16. package/packages/analyzers/dist/regex-call-graph.js +1 -2
  17. package/packages/analyzers/dist/structure-analyzer.d.ts +1 -2
  18. package/packages/analyzers/dist/structure-analyzer.js +1 -2
  19. package/packages/analyzers/dist/symbol-analyzer.d.ts +3 -3
  20. package/packages/analyzers/dist/symbol-analyzer.js +8 -7
  21. package/packages/analyzers/dist/ts-call-graph.d.ts +1 -2
  22. package/packages/analyzers/dist/ts-call-graph.js +1 -2
  23. package/packages/analyzers/dist/types.d.ts +1 -2
  24. package/packages/chunker/dist/call-graph-extractor.d.ts +3 -3
  25. package/packages/chunker/dist/call-graph-extractor.js +1 -2
  26. package/packages/chunker/dist/chunker-factory.d.ts +12 -4
  27. package/packages/chunker/dist/chunker-factory.js +1 -2
  28. package/packages/chunker/dist/chunker.interface.d.ts +1 -2
  29. package/packages/chunker/dist/code-chunker.d.ts +1 -2
  30. package/packages/chunker/dist/code-chunker.js +1 -2
  31. package/packages/chunker/dist/extractors/call-extractor.d.ts +24 -0
  32. package/packages/chunker/dist/extractors/call-extractor.js +1 -0
  33. package/packages/chunker/dist/extractors/entry-point-detector.d.ts +14 -0
  34. package/packages/chunker/dist/extractors/entry-point-detector.js +1 -0
  35. package/packages/chunker/dist/extractors/import-extractor.d.ts +14 -0
  36. package/packages/chunker/dist/extractors/import-extractor.js +1 -0
  37. package/packages/chunker/dist/extractors/pattern-detector.d.ts +14 -0
  38. package/packages/chunker/dist/extractors/pattern-detector.js +1 -0
  39. package/packages/chunker/dist/extractors/scope-resolver.d.ts +26 -0
  40. package/packages/chunker/dist/extractors/scope-resolver.js +1 -0
  41. package/packages/chunker/dist/extractors/symbol-extractor.d.ts +14 -0
  42. package/packages/chunker/dist/extractors/symbol-extractor.js +1 -0
  43. package/packages/chunker/dist/extractors/types.d.ts +36 -0
  44. package/packages/chunker/dist/extractors/types.js +1 -0
  45. package/packages/chunker/dist/generic-chunker.d.ts +1 -2
  46. package/packages/chunker/dist/generic-chunker.js +1 -2
  47. package/packages/chunker/dist/index.d.ts +15 -4
  48. package/packages/chunker/dist/index.js +1 -1
  49. package/packages/chunker/dist/markdown-chunker.d.ts +1 -2
  50. package/packages/chunker/dist/markdown-chunker.js +1 -2
  51. package/packages/chunker/dist/wasm/languages.d.ts +18 -0
  52. package/packages/chunker/dist/wasm/languages.js +1 -0
  53. package/packages/chunker/dist/wasm/query-executor.d.ts +70 -0
  54. package/packages/chunker/dist/wasm/query-executor.js +1 -0
  55. package/packages/chunker/dist/wasm/runtime.d.ts +44 -0
  56. package/packages/chunker/dist/wasm/runtime.js +1 -0
  57. package/packages/chunker/dist/wasm/types.d.ts +84 -0
  58. package/packages/chunker/dist/wasm/types.js +1 -0
  59. package/packages/chunker/dist/wasm-chunker.d.ts +23 -0
  60. package/packages/chunker/dist/wasm-chunker.js +6 -0
  61. package/packages/chunker/src/queries/go/calls.scm +11 -0
  62. package/packages/chunker/src/queries/go/entry-points.scm +20 -0
  63. package/packages/chunker/src/queries/go/imports.scm +6 -0
  64. package/packages/chunker/src/queries/go/patterns.scm +25 -0
  65. package/packages/chunker/src/queries/go/symbols.scm +26 -0
  66. package/packages/chunker/src/queries/java/calls.scm +10 -0
  67. package/packages/chunker/src/queries/java/entry-points.scm +27 -0
  68. package/packages/chunker/src/queries/java/imports.scm +11 -0
  69. package/packages/chunker/src/queries/java/patterns.scm +27 -0
  70. package/packages/chunker/src/queries/java/symbols.scm +28 -0
  71. package/packages/chunker/src/queries/javascript/calls.scm +21 -0
  72. package/packages/chunker/src/queries/javascript/entry-points.scm +31 -0
  73. package/packages/chunker/src/queries/javascript/imports.scm +32 -0
  74. package/packages/chunker/src/queries/javascript/patterns.scm +28 -0
  75. package/packages/chunker/src/queries/javascript/symbols.scm +52 -0
  76. package/packages/chunker/src/queries/python/calls.scm +11 -0
  77. package/packages/chunker/src/queries/python/entry-points.scm +21 -0
  78. package/packages/chunker/src/queries/python/imports.scm +14 -0
  79. package/packages/chunker/src/queries/python/patterns.scm +25 -0
  80. package/packages/chunker/src/queries/python/symbols.scm +17 -0
  81. package/packages/chunker/src/queries/rust/calls.scm +20 -0
  82. package/packages/chunker/src/queries/rust/entry-points.scm +7 -0
  83. package/packages/chunker/src/queries/rust/imports.scm +26 -0
  84. package/packages/chunker/src/queries/rust/patterns.scm +18 -0
  85. package/packages/chunker/src/queries/rust/symbols.scm +73 -0
  86. package/packages/chunker/src/queries/typescript/calls.scm +21 -0
  87. package/packages/chunker/src/queries/typescript/entry-points.scm +48 -0
  88. package/packages/chunker/src/queries/typescript/imports.scm +35 -0
  89. package/packages/chunker/src/queries/typescript/patterns.scm +47 -0
  90. package/packages/chunker/src/queries/typescript/symbols.scm +79 -0
  91. package/packages/chunker/wasm/tree-sitter-go.wasm +0 -0
  92. package/packages/chunker/wasm/tree-sitter-java.wasm +0 -0
  93. package/packages/chunker/wasm/tree-sitter-javascript.wasm +0 -0
  94. package/packages/chunker/wasm/tree-sitter-python.wasm +0 -0
  95. package/packages/chunker/wasm/tree-sitter-rust.wasm +0 -0
  96. package/packages/chunker/wasm/tree-sitter-typescript.wasm +0 -0
  97. package/packages/chunker/wasm/tree-sitter.wasm +0 -0
  98. package/packages/cli/dist/commands/analyze.d.ts +1 -2
  99. package/packages/cli/dist/commands/analyze.js +1 -2
  100. package/packages/cli/dist/commands/context-cmds.d.ts +1 -2
  101. package/packages/cli/dist/commands/context-cmds.js +1 -2
  102. package/packages/cli/dist/commands/environment.d.ts +1 -2
  103. package/packages/cli/dist/commands/environment.js +1 -2
  104. package/packages/cli/dist/commands/execution.d.ts +1 -2
  105. package/packages/cli/dist/commands/execution.js +1 -2
  106. package/packages/cli/dist/commands/graph.d.ts +1 -2
  107. package/packages/cli/dist/commands/graph.js +1 -2
  108. package/packages/cli/dist/commands/init/adapters.d.ts +4 -3
  109. package/packages/cli/dist/commands/init/adapters.js +1 -2
  110. package/packages/cli/dist/commands/init/config.d.ts +1 -2
  111. package/packages/cli/dist/commands/init/config.js +3 -4
  112. package/packages/cli/dist/commands/init/constants.d.ts +18 -0
  113. package/packages/cli/dist/commands/init/constants.js +1 -0
  114. package/packages/cli/dist/commands/init/curated.d.ts +1 -2
  115. package/packages/cli/dist/commands/init/curated.js +1 -2
  116. package/packages/cli/dist/commands/init/global.d.ts +34 -0
  117. package/packages/cli/dist/commands/init/global.js +5 -0
  118. package/packages/cli/dist/commands/init/index.d.ts +7 -2
  119. package/packages/cli/dist/commands/init/index.js +5 -3
  120. package/packages/cli/dist/commands/init/scaffold.d.ts +1 -2
  121. package/packages/cli/dist/commands/init/scaffold.js +1 -2
  122. package/packages/cli/dist/commands/init/templates.d.ts +3 -4
  123. package/packages/cli/dist/commands/init/templates.js +106 -243
  124. package/packages/cli/dist/commands/knowledge.d.ts +1 -2
  125. package/packages/cli/dist/commands/knowledge.js +1 -2
  126. package/packages/cli/dist/commands/search.d.ts +1 -2
  127. package/packages/cli/dist/commands/search.js +1 -2
  128. package/packages/cli/dist/commands/system.d.ts +1 -2
  129. package/packages/cli/dist/commands/system.js +3 -4
  130. package/packages/cli/dist/commands/workspace.d.ts +1 -2
  131. package/packages/cli/dist/commands/workspace.js +1 -2
  132. package/packages/cli/dist/context.d.ts +1 -2
  133. package/packages/cli/dist/context.js +1 -2
  134. package/packages/cli/dist/helpers.d.ts +1 -2
  135. package/packages/cli/dist/helpers.js +1 -2
  136. package/packages/cli/dist/index.d.ts +1 -2
  137. package/packages/cli/dist/index.js +1 -2
  138. package/packages/cli/dist/kb-init.d.ts +1 -2
  139. package/packages/cli/dist/kb-init.js +1 -2
  140. package/packages/cli/dist/types.d.ts +1 -2
  141. package/packages/core/dist/constants.d.ts +23 -2
  142. package/packages/core/dist/constants.js +1 -2
  143. package/packages/core/dist/content-detector.d.ts +1 -2
  144. package/packages/core/dist/content-detector.js +1 -2
  145. package/packages/core/dist/errors.d.ts +1 -2
  146. package/packages/core/dist/errors.js +1 -2
  147. package/packages/core/dist/global-registry.d.ts +62 -0
  148. package/packages/core/dist/global-registry.js +1 -0
  149. package/packages/core/dist/index.d.ts +4 -3
  150. package/packages/core/dist/index.js +1 -1
  151. package/packages/core/dist/logger.d.ts +4 -2
  152. package/packages/core/dist/logger.js +1 -2
  153. package/packages/core/dist/types.d.ts +1 -2
  154. package/packages/core/dist/types.js +1 -2
  155. package/packages/embeddings/dist/embedder.interface.d.ts +1 -2
  156. package/packages/embeddings/dist/onnx-embedder.d.ts +1 -2
  157. package/packages/embeddings/dist/onnx-embedder.js +1 -2
  158. package/packages/enterprise-bridge/dist/cache.d.ts +1 -2
  159. package/packages/enterprise-bridge/dist/cache.js +1 -2
  160. package/packages/enterprise-bridge/dist/er-client.d.ts +1 -2
  161. package/packages/enterprise-bridge/dist/er-client.js +1 -2
  162. package/packages/enterprise-bridge/dist/evolution-collector.d.ts +1 -2
  163. package/packages/enterprise-bridge/dist/evolution-collector.js +1 -2
  164. package/packages/enterprise-bridge/dist/policy-store.d.ts +1 -2
  165. package/packages/enterprise-bridge/dist/policy-store.js +1 -2
  166. package/packages/enterprise-bridge/dist/push-adapter.d.ts +1 -2
  167. package/packages/enterprise-bridge/dist/push-adapter.js +1 -2
  168. package/packages/enterprise-bridge/dist/result-merger.d.ts +1 -2
  169. package/packages/enterprise-bridge/dist/result-merger.js +1 -2
  170. package/packages/enterprise-bridge/dist/types.d.ts +1 -2
  171. package/packages/enterprise-bridge/dist/types.js +1 -2
  172. package/packages/indexer/dist/file-hasher.d.ts +1 -2
  173. package/packages/indexer/dist/file-hasher.js +1 -2
  174. package/packages/indexer/dist/filesystem-crawler.d.ts +1 -2
  175. package/packages/indexer/dist/filesystem-crawler.js +1 -2
  176. package/packages/indexer/dist/graph-extractor.d.ts +1 -2
  177. package/packages/indexer/dist/graph-extractor.js +1 -2
  178. package/packages/indexer/dist/incremental-indexer.d.ts +1 -2
  179. package/packages/indexer/dist/incremental-indexer.js +1 -2
  180. package/packages/server/dist/config.d.ts +1 -2
  181. package/packages/server/dist/config.js +1 -2
  182. package/packages/server/dist/cross-workspace.d.ts +43 -0
  183. package/packages/server/dist/cross-workspace.js +1 -0
  184. package/packages/server/dist/curated-manager.d.ts +2 -3
  185. package/packages/server/dist/curated-manager.js +4 -5
  186. package/packages/server/dist/index.js +1 -2
  187. package/packages/server/dist/replay-interceptor.d.ts +1 -2
  188. package/packages/server/dist/replay-interceptor.js +1 -2
  189. package/packages/server/dist/resources/resources.d.ts +1 -2
  190. package/packages/server/dist/resources/resources.js +1 -2
  191. package/packages/server/dist/server.d.ts +4 -2
  192. package/packages/server/dist/server.js +1 -2
  193. package/packages/server/dist/tools/analyze.tools.d.ts +1 -2
  194. package/packages/server/dist/tools/analyze.tools.js +1 -2
  195. package/packages/server/dist/tools/audit.tool.d.ts +1 -2
  196. package/packages/server/dist/tools/audit.tool.js +1 -2
  197. package/packages/server/dist/tools/bridge.tools.d.ts +1 -2
  198. package/packages/server/dist/tools/bridge.tools.js +1 -2
  199. package/packages/server/dist/tools/evolution.tools.d.ts +1 -2
  200. package/packages/server/dist/tools/evolution.tools.js +1 -2
  201. package/packages/server/dist/tools/forge.tools.d.ts +1 -2
  202. package/packages/server/dist/tools/forge.tools.js +5 -6
  203. package/packages/server/dist/tools/forget.tool.d.ts +1 -2
  204. package/packages/server/dist/tools/forget.tool.js +1 -2
  205. package/packages/server/dist/tools/graph.tool.d.ts +1 -2
  206. package/packages/server/dist/tools/graph.tool.js +1 -2
  207. package/packages/server/dist/tools/list.tool.d.ts +1 -2
  208. package/packages/server/dist/tools/list.tool.js +1 -2
  209. package/packages/server/dist/tools/lookup.tool.d.ts +1 -2
  210. package/packages/server/dist/tools/lookup.tool.js +1 -2
  211. package/packages/server/dist/tools/onboard.tool.d.ts +1 -2
  212. package/packages/server/dist/tools/onboard.tool.js +1 -2
  213. package/packages/server/dist/tools/policy.tools.d.ts +1 -2
  214. package/packages/server/dist/tools/policy.tools.js +1 -2
  215. package/packages/server/dist/tools/produce.tool.d.ts +1 -2
  216. package/packages/server/dist/tools/produce.tool.js +1 -2
  217. package/packages/server/dist/tools/read.tool.d.ts +1 -2
  218. package/packages/server/dist/tools/read.tool.js +2 -3
  219. package/packages/server/dist/tools/reindex.tool.d.ts +1 -2
  220. package/packages/server/dist/tools/reindex.tool.js +3 -3
  221. package/packages/server/dist/tools/remember.tool.d.ts +1 -2
  222. package/packages/server/dist/tools/remember.tool.js +1 -2
  223. package/packages/server/dist/tools/replay.tool.d.ts +1 -2
  224. package/packages/server/dist/tools/replay.tool.js +1 -2
  225. package/packages/server/dist/tools/search.tool.d.ts +1 -2
  226. package/packages/server/dist/tools/search.tool.js +4 -5
  227. package/packages/server/dist/tools/status.tool.d.ts +7 -3
  228. package/packages/server/dist/tools/status.tool.js +2 -3
  229. package/packages/server/dist/tools/toolkit.tools.d.ts +1 -2
  230. package/packages/server/dist/tools/toolkit.tools.js +19 -19
  231. package/packages/server/dist/tools/update.tool.d.ts +1 -2
  232. package/packages/server/dist/tools/update.tool.js +1 -2
  233. package/packages/server/dist/tools/utility.tools.d.ts +1 -2
  234. package/packages/server/dist/tools/utility.tools.js +2 -3
  235. package/packages/server/dist/version-check.d.ts +1 -2
  236. package/packages/server/dist/version-check.js +1 -2
  237. package/packages/store/dist/graph-store.interface.d.ts +1 -2
  238. package/packages/store/dist/lance-store.d.ts +1 -2
  239. package/packages/store/dist/lance-store.js +1 -2
  240. package/packages/store/dist/sqlite-graph-store.d.ts +1 -2
  241. package/packages/store/dist/sqlite-graph-store.js +6 -7
  242. package/packages/store/dist/store-factory.d.ts +1 -2
  243. package/packages/store/dist/store-factory.js +1 -2
  244. package/packages/store/dist/store.interface.d.ts +1 -2
  245. package/packages/tools/dist/audit.d.ts +1 -2
  246. package/packages/tools/dist/audit.js +1 -2
  247. package/packages/tools/dist/batch.d.ts +1 -2
  248. package/packages/tools/dist/batch.js +1 -2
  249. package/packages/tools/dist/changelog.d.ts +1 -2
  250. package/packages/tools/dist/changelog.js +2 -3
  251. package/packages/tools/dist/check.d.ts +1 -2
  252. package/packages/tools/dist/check.js +2 -3
  253. package/packages/tools/dist/checkpoint.d.ts +1 -2
  254. package/packages/tools/dist/checkpoint.js +1 -2
  255. package/packages/tools/dist/codemod.d.ts +1 -2
  256. package/packages/tools/dist/codemod.js +1 -2
  257. package/packages/tools/dist/compact.d.ts +1 -2
  258. package/packages/tools/dist/compact.js +1 -2
  259. package/packages/tools/dist/data-transform.d.ts +1 -2
  260. package/packages/tools/dist/data-transform.js +1 -2
  261. package/packages/tools/dist/dead-symbols.d.ts +1 -2
  262. package/packages/tools/dist/dead-symbols.js +2 -3
  263. package/packages/tools/dist/delegate.d.ts +1 -2
  264. package/packages/tools/dist/delegate.js +1 -2
  265. package/packages/tools/dist/diff-parse.d.ts +1 -2
  266. package/packages/tools/dist/diff-parse.js +1 -2
  267. package/packages/tools/dist/digest.d.ts +1 -2
  268. package/packages/tools/dist/digest.js +1 -2
  269. package/packages/tools/dist/dogfood-log.d.ts +49 -0
  270. package/packages/tools/dist/dogfood-log.js +2 -0
  271. package/packages/tools/dist/encode.d.ts +1 -2
  272. package/packages/tools/dist/encode.js +1 -2
  273. package/packages/tools/dist/env-info.d.ts +1 -2
  274. package/packages/tools/dist/env-info.js +1 -2
  275. package/packages/tools/dist/eval.d.ts +1 -2
  276. package/packages/tools/dist/eval.js +1 -2
  277. package/packages/tools/dist/evidence-map.d.ts +1 -2
  278. package/packages/tools/dist/evidence-map.js +2 -3
  279. package/packages/tools/dist/file-cache.d.ts +1 -2
  280. package/packages/tools/dist/file-cache.js +1 -2
  281. package/packages/tools/dist/file-summary.d.ts +17 -2
  282. package/packages/tools/dist/file-summary.js +2 -3
  283. package/packages/tools/dist/file-walk.d.ts +1 -2
  284. package/packages/tools/dist/file-walk.js +1 -2
  285. package/packages/tools/dist/find-examples.d.ts +1 -2
  286. package/packages/tools/dist/find-examples.js +1 -2
  287. package/packages/tools/dist/find.d.ts +1 -2
  288. package/packages/tools/dist/find.js +1 -2
  289. package/packages/tools/dist/forge-classify.d.ts +1 -2
  290. package/packages/tools/dist/forge-classify.js +2 -3
  291. package/packages/tools/dist/forge-ground.d.ts +1 -2
  292. package/packages/tools/dist/forge-ground.js +1 -2
  293. package/packages/tools/dist/git-context.d.ts +1 -2
  294. package/packages/tools/dist/git-context.js +1 -2
  295. package/packages/tools/dist/graph-query.d.ts +1 -2
  296. package/packages/tools/dist/graph-query.js +1 -2
  297. package/packages/tools/dist/guide.d.ts +1 -2
  298. package/packages/tools/dist/guide.js +1 -2
  299. package/packages/tools/dist/health.d.ts +1 -2
  300. package/packages/tools/dist/health.js +1 -2
  301. package/packages/tools/dist/http-request.d.ts +1 -2
  302. package/packages/tools/dist/http-request.js +1 -2
  303. package/packages/tools/dist/index.d.ts +2 -1
  304. package/packages/tools/dist/index.js +1 -1
  305. package/packages/tools/dist/lane.d.ts +1 -2
  306. package/packages/tools/dist/lane.js +3 -4
  307. package/packages/tools/dist/measure.d.ts +4 -3
  308. package/packages/tools/dist/measure.js +2 -3
  309. package/packages/tools/dist/onboard.d.ts +1 -2
  310. package/packages/tools/dist/onboard.js +14 -15
  311. package/packages/tools/dist/parse-output.d.ts +1 -2
  312. package/packages/tools/dist/parse-output.js +2 -3
  313. package/packages/tools/dist/path-resolver.d.ts +1 -2
  314. package/packages/tools/dist/path-resolver.js +1 -2
  315. package/packages/tools/dist/process-manager.d.ts +1 -2
  316. package/packages/tools/dist/process-manager.js +1 -2
  317. package/packages/tools/dist/queue.d.ts +1 -2
  318. package/packages/tools/dist/queue.js +1 -2
  319. package/packages/tools/dist/regex-test.d.ts +1 -2
  320. package/packages/tools/dist/regex-test.js +1 -2
  321. package/packages/tools/dist/rename.d.ts +1 -2
  322. package/packages/tools/dist/rename.js +1 -2
  323. package/packages/tools/dist/replay.d.ts +2 -3
  324. package/packages/tools/dist/replay.js +4 -5
  325. package/packages/tools/dist/response-envelope.d.ts +1 -2
  326. package/packages/tools/dist/response-envelope.js +1 -2
  327. package/packages/tools/dist/schema-validate.d.ts +1 -2
  328. package/packages/tools/dist/schema-validate.js +1 -2
  329. package/packages/tools/dist/scope-map.d.ts +1 -2
  330. package/packages/tools/dist/scope-map.js +1 -2
  331. package/packages/tools/dist/snippet.d.ts +1 -2
  332. package/packages/tools/dist/snippet.js +1 -2
  333. package/packages/tools/dist/stash.d.ts +1 -2
  334. package/packages/tools/dist/stash.js +1 -2
  335. package/packages/tools/dist/stratum-card.d.ts +1 -2
  336. package/packages/tools/dist/stratum-card.js +1 -2
  337. package/packages/tools/dist/symbol.d.ts +3 -2
  338. package/packages/tools/dist/symbol.js +3 -4
  339. package/packages/tools/dist/test-run.d.ts +1 -2
  340. package/packages/tools/dist/test-run.js +2 -3
  341. package/packages/tools/dist/text-utils.d.ts +1 -2
  342. package/packages/tools/dist/text-utils.js +1 -2
  343. package/packages/tools/dist/time-utils.d.ts +1 -2
  344. package/packages/tools/dist/time-utils.js +1 -2
  345. package/packages/tools/dist/trace.d.ts +3 -2
  346. package/packages/tools/dist/trace.js +2 -3
  347. package/packages/tools/dist/truncation.d.ts +1 -2
  348. package/packages/tools/dist/truncation.js +1 -2
  349. package/packages/tools/dist/watch.d.ts +1 -2
  350. package/packages/tools/dist/watch.js +1 -2
  351. package/packages/tools/dist/web-fetch.d.ts +1 -2
  352. package/packages/tools/dist/web-fetch.js +1 -2
  353. package/packages/tools/dist/web-search.d.ts +1 -2
  354. package/packages/tools/dist/web-search.js +1 -2
  355. package/packages/tools/dist/workset.d.ts +1 -2
  356. package/packages/tools/dist/workset.js +1 -2
  357. package/packages/tui/dist/{App-BAlmxCCw.js → App-CYLNJLr6.js} +1 -2
  358. package/packages/tui/dist/App.d.ts +1 -2
  359. package/packages/tui/dist/App.js +1 -1
  360. package/packages/tui/dist/CuratedPanel-sYdZAICX.js +1 -2
  361. package/packages/tui/dist/LogPanel-DtMnoyXT.js +3 -0
  362. package/packages/tui/dist/SearchPanel-DREo6zgt.js +1 -2
  363. package/packages/tui/dist/StatusPanel-2ex8fLOO.js +1 -2
  364. package/packages/tui/dist/devtools-DUyj952l.js +1 -2
  365. package/packages/tui/dist/embedder.interface-D4ew0HPW.d.ts +1 -2
  366. package/packages/tui/dist/index-B9VpfVPP.d.ts +1 -2
  367. package/packages/tui/dist/index.js +1 -2
  368. package/packages/tui/dist/jsx-runtime-Cof-kwFn.js +1 -2
  369. package/packages/tui/dist/panels/CuratedPanel.d.ts +1 -2
  370. package/packages/tui/dist/panels/LogPanel.d.ts +1 -2
  371. package/packages/tui/dist/panels/LogPanel.js +1 -1
  372. package/packages/tui/dist/panels/SearchPanel.d.ts +1 -2
  373. package/packages/tui/dist/panels/StatusPanel.d.ts +1 -2
  374. package/packages/tui/dist/store.interface-CnY6SPOH.d.ts +1 -2
  375. package/scaffold/copilot/agents/Architect-Reviewer-Alpha.agent.md +7 -0
  376. package/scaffold/copilot/agents/Architect-Reviewer-Beta.agent.md +7 -0
  377. package/scaffold/copilot/agents/Documenter.agent.md +7 -0
  378. package/scaffold/copilot/agents/Orchestrator.agent.md +8 -0
  379. package/scaffold/copilot/agents/Planner.agent.md +9 -0
  380. package/scaffold/copilot/agents/Refactor.agent.md +6 -0
  381. package/scaffold/copilot/agents/Researcher-Alpha.agent.md +8 -0
  382. package/scaffold/copilot/agents/Researcher-Beta.agent.md +8 -0
  383. package/scaffold/copilot/agents/Researcher-Delta.agent.md +8 -0
  384. package/scaffold/copilot/agents/Researcher-Gamma.agent.md +8 -0
  385. package/scaffold/general/agents/Architect-Reviewer-Alpha.agent.md +21 -0
  386. package/scaffold/general/agents/Architect-Reviewer-Beta.agent.md +21 -0
  387. package/scaffold/general/agents/Documenter.agent.md +42 -0
  388. package/scaffold/general/agents/Orchestrator.agent.md +104 -0
  389. package/scaffold/general/agents/Planner.agent.md +55 -0
  390. package/scaffold/general/agents/Refactor.agent.md +36 -0
  391. package/scaffold/general/agents/Researcher-Alpha.agent.md +20 -0
  392. package/scaffold/general/agents/Researcher-Beta.agent.md +20 -0
  393. package/scaffold/general/agents/Researcher-Delta.agent.md +20 -0
  394. package/scaffold/general/agents/Researcher-Gamma.agent.md +20 -0
  395. package/scaffold/{copilot → general}/agents/_shared/code-agent-base.md +18 -0
  396. package/skills/adr-skill/SKILL.md +329 -0
  397. package/skills/adr-skill/assets/templates/adr-madr.md +89 -0
  398. package/skills/adr-skill/assets/templates/adr-readme.md +20 -0
  399. package/skills/adr-skill/assets/templates/adr-simple.md +46 -0
  400. package/skills/adr-skill/references/adr-conventions.md +95 -0
  401. package/skills/adr-skill/references/examples.md +193 -0
  402. package/skills/adr-skill/references/review-checklist.md +77 -0
  403. package/skills/adr-skill/references/template-variants.md +52 -0
  404. package/skills/adr-skill/scripts/bootstrap_adr.js +259 -0
  405. package/skills/adr-skill/scripts/new_adr.js +391 -0
  406. package/skills/adr-skill/scripts/set_adr_status.js +169 -0
  407. package/skills/c4-architecture/SKILL.md +295 -0
  408. package/skills/c4-architecture/references/advanced-patterns.md +552 -0
  409. package/skills/c4-architecture/references/c4-syntax.md +492 -0
  410. package/skills/c4-architecture/references/common-mistakes.md +437 -0
  411. package/skills/knowledge-base/SKILL.md +78 -1
  412. package/skills/lesson-learned/SKILL.md +105 -0
  413. package/skills/lesson-learned/references/anti-patterns.md +55 -0
  414. package/skills/lesson-learned/references/se-principles.md +109 -0
  415. package/skills/requirements-clarity/SKILL.md +324 -0
  416. package/skills/session-handoff/SKILL.md +189 -0
  417. package/skills/session-handoff/references/handoff-template.md +139 -0
  418. package/skills/session-handoff/references/resume-checklist.md +80 -0
  419. package/skills/session-handoff/scripts/check_staleness.js +269 -0
  420. package/skills/session-handoff/scripts/create_handoff.js +299 -0
  421. package/skills/session-handoff/scripts/list_handoffs.js +113 -0
  422. package/skills/session-handoff/scripts/validate_handoff.js +241 -0
  423. package/packages/chunker/dist/treesitter-chunker.d.ts +0 -44
  424. package/packages/chunker/dist/treesitter-chunker.js +0 -7
  425. package/packages/cli/dist/commands/init.d.ts +0 -15
  426. package/packages/cli/dist/commands/init.js +0 -305
  427. package/packages/tui/dist/LogPanel-DVB8Sv46.js +0 -4
  428. /package/scaffold/{copilot → general}/agents/Code-Reviewer-Alpha.agent.md +0 -0
  429. /package/scaffold/{copilot → general}/agents/Code-Reviewer-Beta.agent.md +0 -0
  430. /package/scaffold/{copilot → general}/agents/Debugger.agent.md +0 -0
  431. /package/scaffold/{copilot → general}/agents/Explorer.agent.md +0 -0
  432. /package/scaffold/{copilot → general}/agents/Frontend.agent.md +0 -0
  433. /package/scaffold/{copilot → general}/agents/Implementer.agent.md +0 -0
  434. /package/scaffold/{copilot → general}/agents/README.md +0 -0
  435. /package/scaffold/{copilot → general}/agents/Security.agent.md +0 -0
  436. /package/scaffold/{copilot → general}/agents/_shared/adr-protocol.md +0 -0
  437. /package/scaffold/{copilot → general}/agents/_shared/architect-reviewer-base.md +0 -0
  438. /package/scaffold/{copilot → general}/agents/_shared/code-reviewer-base.md +0 -0
  439. /package/scaffold/{copilot → general}/agents/_shared/decision-protocol.md +0 -0
  440. /package/scaffold/{copilot → general}/agents/_shared/forge-protocol.md +0 -0
  441. /package/scaffold/{copilot → general}/agents/_shared/researcher-base.md +0 -0
  442. /package/scaffold/{copilot → general}/agents/templates/adr-template.md +0 -0
  443. /package/scaffold/{copilot → general}/agents/templates/execution-state.md +0 -0
  444. /package/scaffold/{copilot → general}/prompts/ask.prompt.md +0 -0
  445. /package/scaffold/{copilot → general}/prompts/debug.prompt.md +0 -0
  446. /package/scaffold/{copilot → general}/prompts/design.prompt.md +0 -0
  447. /package/scaffold/{copilot → general}/prompts/implement.prompt.md +0 -0
  448. /package/scaffold/{copilot → general}/prompts/plan.prompt.md +0 -0
  449. /package/scaffold/{copilot → general}/prompts/review.prompt.md +0 -0
@@ -0,0 +1,43 @@
1
+ import { RegistryEntry, SearchResult } from "@kb/core";
2
+ import { IKnowledgeStore, SearchOptions } from "@kb/store";
3
+
4
+ //#region packages/server/src/cross-workspace.d.ts
5
+ interface CrossWorkspaceOptions {
6
+ /** Workspace partition names or folder basenames to include. `"*"` means all. */
7
+ workspaces: string[];
8
+ /** Current workspace partition (excluded from cross-workspace queries to avoid double results) */
9
+ currentPartition?: string;
10
+ }
11
+ /**
12
+ * Resolve workspace identifiers to partition entries.
13
+ * Accepts partition keys (exact) or folder basenames (fuzzy match).
14
+ * Special value `"*"` resolves to all registered workspaces.
15
+ */
16
+ declare function resolveWorkspaces(identifiers: string[], currentPartition?: string): RegistryEntry[];
17
+ /**
18
+ * Open temporary read-only store instances for the specified workspace partitions.
19
+ * Caller MUST call closeAll() when done.
20
+ */
21
+ declare function openWorkspaceStores(entries: RegistryEntry[]): Promise<{
22
+ stores: Map<string, IKnowledgeStore>;
23
+ closeAll: () => Promise<void>;
24
+ }>;
25
+ /**
26
+ * Fan-out a vector search to multiple workspace stores in parallel.
27
+ * Returns merged results sorted by score descending, with workspace labels.
28
+ */
29
+ declare function fanOutSearch(stores: Map<string, IKnowledgeStore>, queryVector: Float32Array, options: SearchOptions & {
30
+ limit: number;
31
+ }): Promise<Array<SearchResult & {
32
+ workspace: string;
33
+ }>>;
34
+ /**
35
+ * Fan-out a full-text search to multiple workspace stores in parallel.
36
+ */
37
+ declare function fanOutFtsSearch(stores: Map<string, IKnowledgeStore>, query: string, options: SearchOptions & {
38
+ limit: number;
39
+ }): Promise<Array<SearchResult & {
40
+ workspace: string;
41
+ }>>;
42
+ //#endregion
43
+ export { CrossWorkspaceOptions, fanOutFtsSearch, fanOutSearch, openWorkspaceStores, resolveWorkspaces };
@@ -0,0 +1 @@
1
+ import{createLogger as e,getPartitionDir as t,isGlobalInstalled as n,listWorkspaces as r}from"../../core/dist/index.js";import{createStore as i}from"../../store/dist/index.js";const a=e(`cross-workspace`);function o(e,t){if(!n())return[];let i=r();if(i.length===0)return[];if(e.includes(`*`))return t?i.filter(e=>e.partition!==t):i;let a=[];for(let n of e){let e=i.find(e=>e.partition===n);if(e){e.partition!==t&&a.push(e);continue}let r=i.filter(e=>e.partition!==t&&e.partition.replace(/-[a-f0-9]{8}$/,``)===n.toLowerCase());a.push(...r)}let o=new Set;return a.filter(e=>o.has(e.partition)?!1:(o.add(e.partition),!0))}async function s(e){let n=new Map;for(let r of e)try{let e=await i({backend:`lancedb`,path:t(r.partition)});await e.initialize(),n.set(r.partition,e)}catch(e){a.warn(`Failed to open workspace store`,{partition:r.partition,err:e})}return{stores:n,closeAll:async()=>{for(let[,e]of n)try{await e.close()}catch{}}}}async function c(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.search(t,n)).map(t=>({...t,workspace:e}))}catch(t){return a.warn(`Cross-workspace search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}async function l(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.ftsSearch(t,n)).map(t=>({...t,workspace:e}))}catch(t){return a.warn(`Cross-workspace FTS search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}export{l as fanOutFtsSearch,c as fanOutSearch,s as openWorkspaceStores,o as resolveWorkspaces};
@@ -29,7 +29,7 @@ interface CuratedFrontmatter {
29
29
  }
30
30
  /**
31
31
  * Manages curated knowledge files — the LLM's persistent memory.
32
- * Files are stored as markdown in a curated/ directory with YAML frontmatter.
32
+ * Files are stored as markdown in a .ai/curated/ directory with YAML frontmatter.
33
33
  */
34
34
  declare class CuratedKnowledgeManager {
35
35
  private readonly curatedDir;
@@ -85,5 +85,4 @@ declare class CuratedKnowledgeManager {
85
85
  };
86
86
  }
87
87
  //#endregion
88
- export { CuratedCategory, CuratedEntry, CuratedKnowledgeManager };
89
- //# sourceMappingURL=curated-manager.d.ts.map
88
+ export { CuratedCategory, CuratedEntry, CuratedKnowledgeManager };
@@ -1,6 +1,5 @@
1
- import{dirname as e,isAbsolute as t,join as n}from"node:path";import{createLogger as r,serializeError as i}from"../../core/dist/index.js";import{createHash as a}from"node:crypto";import{mkdir as o,readFile as s,readdir as c,stat as l,unlink as u,writeFile as d}from"node:fs/promises";const f=50*1024,p=r(`server`);var m=class{constructor(e,t,n){this.curatedDir=e,this.store=t,this.embedder=n}async remember(t,r,i,a=[]){this.validateCategoryName(i),this.validateContentSize(r);let s=this.slugify(t),c=await this.uniqueRelativePath(i,s),l=n(this.curatedDir,c),f=new Date().toISOString(),p={title:t,category:i,tags:a,created:f,updated:f,version:1,origin:`curated`,changelog:[{version:1,date:f,reason:`Initial creation`}]},m=this.serializeFile(r,p);await o(e(l),{recursive:!0});try{await d(l,m,{encoding:`utf-8`,flag:`wx`})}catch(e){throw e.code===`EEXIST`?Error(`Concurrent write collision for "${c}" — retry the operation`):e}try{await this.indexCuratedFile(c,r,p)}catch(e){throw await u(l).catch(()=>{}),Error(`Remember failed: wrote file but indexing failed — rolled back. ${e.message}`)}return{path:c}}async update(e,t,r){this.guardPath(e),this.validateContentSize(t);let i=n(this.curatedDir,e),a=await s(i,`utf-8`),{frontmatter:o}=this.parseFile(a),c=(o.version??1)+1,l=new Date().toISOString();return o.version=c,o.updated=l,o.changelog=[...o.changelog??[],{version:c,date:l,reason:r}],await d(i,this.serializeFile(t,o),`utf-8`),await this.indexCuratedFile(e,t,o),{path:e,version:c}}async forget(e,t){this.guardPath(e),await u(n(this.curatedDir,e));let r=`curated/${e}`;return await this.store.deleteBySourcePath(r).catch(e=>{p.warn(`File deleted but vector cleanup failed`,{sourcePath:r,...i(e)})}),{path:e}}async read(e){this.guardPath(e);let t=await s(n(this.curatedDir,e),`utf-8`),{frontmatter:r,content:i}=this.parseFile(t),a=e.split(`/`)[0];return{path:e,title:r.title??e,category:a,tags:r.tags??[],version:r.version??1,created:r.created??``,updated:r.updated??``,contentPreview:i.slice(0,200),content:i}}async list(e){let t=[],r=e?.category?[e.category]:await this.discoverCategories();for(let i of r){let r=n(this.curatedDir,i);try{let a=await c(r);for(let o of a){if(!o.endsWith(`.md`))continue;let a=await s(n(r,o),`utf-8`),{frontmatter:c,content:l}=this.parseFile(a);e?.tag&&!(c.tags??[]).includes(e.tag)||t.push({path:`${i}/${o}`,title:c.title??o,category:i,tags:c.tags??[],version:c.version??1,created:c.created??``,updated:c.updated??``,contentPreview:l.slice(0,200)})}}catch{}}return t}async reindexAll(){let e=await this.discoverCategories(),t=[],r=[];for(let a of e){let e=n(this.curatedDir,a),o;try{o=(await c(e)).filter(e=>e.endsWith(`.md`))}catch{continue}for(let c of o){let o=`${a}/${c}`,l=n(e,c);try{let e=await s(l,`utf-8`),{frontmatter:t,content:n}=this.parseFile(e);r.push({relativePath:o,content:n,frontmatter:t})}catch(e){p.error(`Failed to read curated file`,{relativePath:o,...i(e)}),t.push(`${o}: read failed`)}}}if(r.length===0)return{indexed:0,errors:t};let a=await this.embedder.embedBatch(r.map(e=>e.content)),o=new Date().toISOString(),l=r.map(e=>{let t=`curated/${e.relativePath}`;return{id:this.hashId(t,0),content:e.content,sourcePath:t,contentType:`curated-knowledge`,headingPath:e.frontmatter.title,chunkIndex:0,totalChunks:1,startLine:1,endLine:e.content.split(`
2
- `).length,fileHash:this.hash(e.content),indexedAt:o,origin:`curated`,tags:e.frontmatter.tags,category:e.frontmatter.category,version:e.frontmatter.version}});return await this.store.upsert(l,a),{indexed:r.length,errors:t}}async indexCuratedFile(e,t,n){let r=await this.embedder.embed(t),i=`curated/${e}`,a=new Date().toISOString(),o={id:this.hashId(i,0),content:t,sourcePath:i,contentType:`curated-knowledge`,headingPath:n.title,chunkIndex:0,totalChunks:1,startLine:1,endLine:t.split(`
3
- `).length,fileHash:this.hash(t),indexedAt:a,origin:`curated`,tags:n.tags,category:n.category,version:n.version};await this.store.upsert([o],[r])}async discoverCategories(){try{return(await c(this.curatedDir,{withFileTypes:!0})).filter(e=>e.isDirectory()&&/^[a-z][a-z0-9-]*$/.test(e.name)).map(e=>e.name)}catch{return[]}}guardPath(e){if(e.includes(`..`)||t(e))throw Error(`Invalid path: ${e}. Must be relative within curated/ directory.`);let n=e.split(`/`)[0];this.validateCategoryName(n)}validateCategoryName(e){if(!/^[a-z][a-z0-9-]*$/.test(e))throw Error(`Invalid category name: "${e}". Must be lowercase kebab-case (e.g., "decisions", "api-contracts").`)}validateContentSize(e){if(Buffer.byteLength(e,`utf-8`)>f)throw Error(`Content exceeds maximum size of ${f/1024}KB`)}slugify(e){return e.toLowerCase().replace(/[^a-z0-9]+/g,`-`).replace(/^-|-$/g,``).slice(0,80)}async uniqueRelativePath(e,t){let r=`${e}/${t}.md`,i=n(this.curatedDir,r);try{await l(i)}catch{return r}for(let r=2;r<=100;r++){let i=`${e}/${t}-${r}.md`;try{await l(n(this.curatedDir,i))}catch{return i}}throw Error(`Too many entries with slug "${t}" in category "${e}"`)}hash(e){return a(`sha256`).update(e).digest(`hex`).slice(0,16)}hashId(e,t){return this.hash(`${e}::${t}`)}serializeFile(e,t){return`${[`---`,`title: "${t.title.replace(/"/g,`\\"`)}"`,`category: ${t.category}`,`tags: [${t.tags.map(e=>`"${e}"`).join(`, `)}]`,`created: ${t.created}`,`updated: ${t.updated}`,`version: ${t.version}`,`origin: ${t.origin}`,`changelog:`,...t.changelog.map(e=>` - version: ${e.version}\n date: ${e.date}\n reason: "${e.reason.replace(/"/g,`\\"`)}"`),`---`].join(`
1
+ import{dirname as e,isAbsolute as t,join as n}from"node:path";import{createLogger as r,serializeError as i}from"../../core/dist/index.js";import{createHash as a}from"node:crypto";import{mkdir as o,readFile as s,readdir as c,stat as l,unlink as u,writeFile as d}from"node:fs/promises";const f=50*1024,p=r(`server`);var m=class{constructor(e,t,n){this.curatedDir=e,this.store=t,this.embedder=n}async remember(t,r,i,a=[]){this.validateCategoryName(i),this.validateContentSize(r);let s=this.slugify(t),c=await this.uniqueRelativePath(i,s),l=n(this.curatedDir,c),f=new Date().toISOString(),p={title:t,category:i,tags:a,created:f,updated:f,version:1,origin:`curated`,changelog:[{version:1,date:f,reason:`Initial creation`}]},m=this.serializeFile(r,p);await o(e(l),{recursive:!0});try{await d(l,m,{encoding:`utf-8`,flag:`wx`})}catch(e){throw e.code===`EEXIST`?Error(`Concurrent write collision for "${c}" — retry the operation`):e}try{await this.indexCuratedFile(c,r,p)}catch(e){throw await u(l).catch(()=>{}),Error(`Remember failed: wrote file but indexing failed — rolled back. ${e.message}`)}return{path:c}}async update(e,t,r){this.guardPath(e),this.validateContentSize(t);let i=n(this.curatedDir,e),a=await s(i,`utf-8`),{frontmatter:o}=this.parseFile(a),c=(o.version??1)+1,l=new Date().toISOString();return o.version=c,o.updated=l,o.changelog=[...o.changelog??[],{version:c,date:l,reason:r}],await d(i,this.serializeFile(t,o),`utf-8`),await this.indexCuratedFile(e,t,o),{path:e,version:c}}async forget(e,t){this.guardPath(e),await u(n(this.curatedDir,e));let r=`.ai/curated/${e}`;return await this.store.deleteBySourcePath(r).catch(e=>{p.warn(`File deleted but vector cleanup failed`,{sourcePath:r,...i(e)})}),{path:e}}async read(e){this.guardPath(e);let t=await s(n(this.curatedDir,e),`utf-8`),{frontmatter:r,content:i}=this.parseFile(t),a=e.split(`/`)[0];return{path:e,title:r.title??e,category:a,tags:r.tags??[],version:r.version??1,created:r.created??``,updated:r.updated??``,contentPreview:i.slice(0,200),content:i}}async list(e){let t=[],r=e?.category?[e.category]:await this.discoverCategories();for(let i of r){let r=n(this.curatedDir,i);try{let a=await c(r);for(let o of a){if(!o.endsWith(`.md`))continue;let a=await s(n(r,o),`utf-8`),{frontmatter:c,content:l}=this.parseFile(a);e?.tag&&!(c.tags??[]).includes(e.tag)||t.push({path:`${i}/${o}`,title:c.title??o,category:i,tags:c.tags??[],version:c.version??1,created:c.created??``,updated:c.updated??``,contentPreview:l.slice(0,200)})}}catch{}}return t}async reindexAll(){let e=await this.discoverCategories(),t=[],r=[];for(let a of e){let e=n(this.curatedDir,a),o;try{o=(await c(e)).filter(e=>e.endsWith(`.md`))}catch{continue}for(let c of o){let o=`${a}/${c}`,l=n(e,c);try{let e=await s(l,`utf-8`),{frontmatter:t,content:n}=this.parseFile(e);r.push({relativePath:o,content:n,frontmatter:t})}catch(e){p.error(`Failed to read curated file`,{relativePath:o,...i(e)}),t.push(`${o}: read failed`)}}}if(r.length===0)return{indexed:0,errors:t};let a=await this.embedder.embedBatch(r.map(e=>e.content)),o=new Date().toISOString(),l=r.map(e=>{let t=`.ai/curated/${e.relativePath}`;return{id:this.hashId(t,0),content:e.content,sourcePath:t,contentType:`curated-knowledge`,headingPath:e.frontmatter.title,chunkIndex:0,totalChunks:1,startLine:1,endLine:e.content.split(`
2
+ `).length,fileHash:this.hash(e.content),indexedAt:o,origin:`curated`,tags:e.frontmatter.tags,category:e.frontmatter.category,version:e.frontmatter.version}});return await this.store.upsert(l,a),{indexed:r.length,errors:t}}async indexCuratedFile(e,t,n){let r=await this.embedder.embed(t),i=`.ai/curated/${e}`,a=new Date().toISOString(),o={id:this.hashId(i,0),content:t,sourcePath:i,contentType:`curated-knowledge`,headingPath:n.title,chunkIndex:0,totalChunks:1,startLine:1,endLine:t.split(`
3
+ `).length,fileHash:this.hash(t),indexedAt:a,origin:`curated`,tags:n.tags,category:n.category,version:n.version};await this.store.upsert([o],[r])}async discoverCategories(){try{return(await c(this.curatedDir,{withFileTypes:!0})).filter(e=>e.isDirectory()&&/^[a-z][a-z0-9-]*$/.test(e.name)).map(e=>e.name)}catch{return[]}}guardPath(e){if(e.includes(`..`)||t(e))throw Error(`Invalid path: ${e}. Must be relative within .ai/curated/ directory.`);let n=e.split(`/`)[0];this.validateCategoryName(n)}validateCategoryName(e){if(!/^[a-z][a-z0-9-]*$/.test(e))throw Error(`Invalid category name: "${e}". Must be lowercase kebab-case (e.g., "decisions", "api-contracts").`)}validateContentSize(e){if(Buffer.byteLength(e,`utf-8`)>f)throw Error(`Content exceeds maximum size of ${f/1024}KB`)}slugify(e){return e.toLowerCase().replace(/[^a-z0-9]+/g,`-`).replace(/^-|-$/g,``).slice(0,80)}async uniqueRelativePath(e,t){let r=`${e}/${t}.md`,i=n(this.curatedDir,r);try{await l(i)}catch{return r}for(let r=2;r<=100;r++){let i=`${e}/${t}-${r}.md`;try{await l(n(this.curatedDir,i))}catch{return i}}throw Error(`Too many entries with slug "${t}" in category "${e}"`)}hash(e){return a(`sha256`).update(e).digest(`hex`).slice(0,16)}hashId(e,t){return this.hash(`${e}::${t}`)}serializeFile(e,t){return`${[`---`,`title: "${t.title.replace(/"/g,`\\"`)}"`,`category: ${t.category}`,`tags: [${t.tags.map(e=>`"${e}"`).join(`, `)}]`,`created: ${t.created}`,`updated: ${t.updated}`,`version: ${t.version}`,`origin: ${t.origin}`,`changelog:`,...t.changelog.map(e=>` - version: ${e.version}\n date: ${e.date}\n reason: "${e.reason.replace(/"/g,`\\"`)}"`),`---`].join(`
4
4
  `)}\n\n${e}\n`}parseFile(e){let t=e.match(/^---\n([\s\S]*?)\n---\n\n?([\s\S]*)$/);if(!t)return{frontmatter:{title:`Untitled`,category:`notes`,tags:[],created:``,updated:``,version:1,origin:`curated`,changelog:[]},content:e};let n=t[1],r=t[2].trim(),i={},a=[],o=n.split(`
5
- `),s=!1,c={};for(let e of o){if(/^changelog:\s*$/.test(e)){s=!0;continue}if(s){let t=e.match(/^\s+-\s+version:\s*(\d+)$/);if(t){c.version!=null&&a.push(c),c={version:parseInt(t[1],10)};continue}let n=e.match(/^\s+date:\s*(.+)$/);if(n){c.date=n[1].trim();continue}let r=e.match(/^\s+reason:\s*"?(.*?)"?\s*$/);if(r){c.reason=r[1];continue}/^\w/.test(e)&&(s=!1,c.version!=null&&a.push(c),c={});continue}let t=e.match(/^(\w+):\s*(.*)$/);if(t){let e=t[1],n=t[2];typeof n==`string`&&n.startsWith(`[`)&&n.endsWith(`]`)?n=n.slice(1,-1).split(`,`).map(e=>e.trim().replace(/^"|"$/g,``)).filter(e=>e.length>0):typeof n==`string`&&/^\d+$/.test(n)?n=parseInt(n,10):typeof n==`string`&&n.startsWith(`"`)&&n.endsWith(`"`)&&(n=n.slice(1,-1)),i[e]=n}}return c.version!=null&&a.push(c),{frontmatter:{title:i.title??`Untitled`,category:i.category??`notes`,tags:i.tags??[],created:i.created??``,updated:i.updated??``,version:i.version??1,origin:`curated`,changelog:a},content:r}}};export{m as CuratedKnowledgeManager};
6
- //# sourceMappingURL=curated-manager.js.map
5
+ `),s=!1,c={};for(let e of o){if(/^changelog:\s*$/.test(e)){s=!0;continue}if(s){let t=e.match(/^\s+-\s+version:\s*(\d+)$/);if(t){c.version!=null&&a.push(c),c={version:parseInt(t[1],10)};continue}let n=e.match(/^\s+date:\s*(.+)$/);if(n){c.date=n[1].trim();continue}let r=e.match(/^\s+reason:\s*"?(.*?)"?\s*$/);if(r){c.reason=r[1];continue}/^\w/.test(e)&&(s=!1,c.version!=null&&a.push(c),c={});continue}let t=e.match(/^(\w+):\s*(.*)$/);if(t){let e=t[1],n=t[2];typeof n==`string`&&n.startsWith(`[`)&&n.endsWith(`]`)?n=n.slice(1,-1).split(`,`).map(e=>e.trim().replace(/^"|"$/g,``)).filter(e=>e.length>0):typeof n==`string`&&/^\d+$/.test(n)?n=parseInt(n,10):typeof n==`string`&&n.startsWith(`"`)&&n.endsWith(`"`)&&(n=n.slice(1,-1)),i[e]=n}}return c.version!=null&&a.push(c),{frontmatter:{title:i.title??`Untitled`,category:i.category??`notes`,tags:i.tags??[],created:i.created??``,updated:i.updated??``,version:i.version??1,origin:`curated`,changelog:a},content:r}}};export{m as CuratedKnowledgeManager};
@@ -1,2 +1 @@
1
- import{loadConfig as e}from"./config.js";import{checkForUpdates as t}from"./version-check.js";import{createLazyServer as n,createMcpServer as r,initializeKnowledgeBase as i}from"./server.js";import{createLogger as a,serializeError as o}from"../../core/dist/index.js";import{parseArgs as s}from"node:util";const c=a(`server`),{values:l}=s({options:{transport:{type:`string`,default:process.env.KB_TRANSPORT??`stdio`},port:{type:`string`,default:process.env.KB_PORT??`3210`}}});async function u(){process.on(`unhandledRejection`,e=>{c.error(`Unhandled rejection`,o(e))}),c.info(`Starting MCP Knowledge Base server`);let a=e();if(c.info(`Config loaded`,{sourceCount:a.sources.length,storePath:a.store.path}),t(),l.transport===`http`){let{StreamableHTTPServerTransport:e}=await import(`@modelcontextprotocol/sdk/server/streamableHttp.js`),t=(await import(`express`)).default,n=await i(a),s=r(n,a);c.info(`MCP server configured`,{toolCount:46,resourceCount:2});let u=t();u.use(t.json()),u.use((e,t,n)=>{if(t.setHeader(`Access-Control-Allow-Origin`,process.env.KB_CORS_ORIGIN??`*`),t.setHeader(`Access-Control-Allow-Methods`,`GET, POST, DELETE, OPTIONS`),t.setHeader(`Access-Control-Allow-Headers`,`Content-Type, Authorization`),e.method===`OPTIONS`){t.status(204).end();return}n()}),u.get(`/health`,(e,t)=>{t.json({status:`ok`})}),u.post(`/mcp`,async(t,n)=>{try{let r=new e({sessionIdGenerator:void 0});await s.connect(r),await r.handleRequest(t,n,t.body),n.on(`close`,()=>{r.close()})}catch(e){c.error(`MCP handler error`,o(e)),n.headersSent||n.status(500).json({jsonrpc:`2.0`,error:{code:-32603,message:`Internal server error`},id:null})}}),u.get(`/mcp`,(e,t)=>{t.writeHead(405).end(JSON.stringify({jsonrpc:`2.0`,error:{code:-32e3,message:`Method not allowed.`},id:null}))}),u.delete(`/mcp`,(e,t)=>{t.writeHead(405).end(JSON.stringify({jsonrpc:`2.0`,error:{code:-32e3,message:`Method not allowed.`},id:null}))});let d=Number(l.port),f=u.listen(d,()=>{c.info(`MCP server listening`,{url:`http://0.0.0.0:${d}/mcp`,port:d}),(async()=>{try{let e=a.sources.map(e=>e.path).join(`, `);c.info(`Running initial index`,{sourcePaths:e});let t=await n.indexer.index(a,e=>{e.phase===`crawling`||e.phase===`done`||e.phase===`chunking`&&e.currentFile&&c.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile})});c.info(`Initial index complete`,{filesProcessed:t.filesProcessed,filesSkipped:t.filesSkipped,chunksCreated:t.chunksCreated,durationMs:t.durationMs});try{let e=await n.curated.reindexAll();c.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){c.error(`Curated re-index failed`,o(e))}}catch(e){c.error(`Initial index failed; will retry on kb_reindex`,o(e))}})().catch(e=>c.error(`Initial index failed`,o(e)))}),p=async e=>{c.info(`Shutdown signal received`,{signal:e}),f.close(),await s.close(),await n.graphStore.close().catch(()=>{}),await n.store.close(),await n.embedder.shutdown(),process.exit(0)};process.on(`SIGINT`,()=>p(`SIGINT`)),process.on(`SIGTERM`,()=>p(`SIGTERM`))}else{let{server:e,ready:t,runInitialIndex:r}=n(a),{StdioServerTransport:i}=await import(`@modelcontextprotocol/sdk/server/stdio.js`),s=new i;await e.connect(s),c.info(`MCP server started`,{transport:`stdio`}),t.catch(e=>{c.error(`Initialization failed`,o(e)),process.exit(1)}),process.env.KB_AUTO_INDEX===`false`?c.warn(`Auto-index disabled; use kb_reindex to index manually`):r().catch(e=>c.error(`Initial index failed`,o(e)))}}u().catch(e=>{c.error(`Fatal error`,o(e)),process.exit(1)});export{};
2
- //# sourceMappingURL=index.js.map
1
+ import{loadConfig as e}from"./config.js";import{checkForUpdates as t}from"./version-check.js";import{ALL_TOOL_NAMES as n,createLazyServer as r,createMcpServer as i,initializeKnowledgeBase as a}from"./server.js";import{createLogger as o,serializeError as s}from"../../core/dist/index.js";import{parseArgs as c}from"node:util";const l=o(`server`),{values:u}=c({options:{transport:{type:`string`,default:process.env.KB_TRANSPORT??`stdio`},port:{type:`string`,default:process.env.KB_PORT??`3210`}}});async function d(){process.on(`unhandledRejection`,e=>{l.error(`Unhandled rejection`,s(e))}),l.info(`Starting MCP Knowledge Base server`);let o=e();if(l.info(`Config loaded`,{sourceCount:o.sources.length,storePath:o.store.path}),t(),u.transport===`http`){let{StreamableHTTPServerTransport:e}=await import(`@modelcontextprotocol/sdk/server/streamableHttp.js`),t=(await import(`express`)).default,r=await a(o),c=i(r,o);l.info(`MCP server configured`,{toolCount:n.length,resourceCount:2});let d=t();d.use(t.json()),d.use((e,t,n)=>{if(t.setHeader(`Access-Control-Allow-Origin`,process.env.KB_CORS_ORIGIN??`*`),t.setHeader(`Access-Control-Allow-Methods`,`GET, POST, DELETE, OPTIONS`),t.setHeader(`Access-Control-Allow-Headers`,`Content-Type, Authorization`),e.method===`OPTIONS`){t.status(204).end();return}n()}),d.get(`/health`,(e,t)=>{t.json({status:`ok`})}),d.post(`/mcp`,async(t,n)=>{try{let r=new e({sessionIdGenerator:void 0});await c.connect(r),await r.handleRequest(t,n,t.body),n.on(`close`,()=>{r.close()})}catch(e){l.error(`MCP handler error`,s(e)),n.headersSent||n.status(500).json({jsonrpc:`2.0`,error:{code:-32603,message:`Internal server error`},id:null})}}),d.get(`/mcp`,(e,t)=>{t.writeHead(405).end(JSON.stringify({jsonrpc:`2.0`,error:{code:-32e3,message:`Method not allowed.`},id:null}))}),d.delete(`/mcp`,(e,t)=>{t.writeHead(405).end(JSON.stringify({jsonrpc:`2.0`,error:{code:-32e3,message:`Method not allowed.`},id:null}))});let f=Number(u.port),p=d.listen(f,()=>{l.info(`MCP server listening`,{url:`http://0.0.0.0:${f}/mcp`,port:f}),(async()=>{try{let e=o.sources.map(e=>e.path).join(`, `);l.info(`Running initial index`,{sourcePaths:e});let t=await r.indexer.index(o,e=>{e.phase===`crawling`||e.phase===`done`||e.phase===`chunking`&&e.currentFile&&l.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile})});l.info(`Initial index complete`,{filesProcessed:t.filesProcessed,filesSkipped:t.filesSkipped,chunksCreated:t.chunksCreated,durationMs:t.durationMs});try{let e=await r.curated.reindexAll();l.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){l.error(`Curated re-index failed`,s(e))}}catch(e){l.error(`Initial index failed; will retry on kb_reindex`,s(e))}})().catch(e=>l.error(`Initial index failed`,s(e)))}),m=async e=>{l.info(`Shutdown signal received`,{signal:e}),p.close(),await c.close(),await r.graphStore.close().catch(()=>{}),await r.store.close(),await r.embedder.shutdown(),process.exit(0)};process.on(`SIGINT`,()=>m(`SIGINT`)),process.on(`SIGTERM`,()=>m(`SIGTERM`))}else{let{server:e,ready:t,runInitialIndex:n}=r(o),{StdioServerTransport:i}=await import(`@modelcontextprotocol/sdk/server/stdio.js`),a=new i;await e.connect(a),l.info(`MCP server started`,{transport:`stdio`}),t.catch(e=>{l.error(`Initialization failed`,s(e)),process.exit(1)}),process.env.KB_AUTO_INDEX===`false`?l.warn(`Auto-index disabled; use kb_reindex to index manually`):n().catch(e=>l.error(`Initial index failed`,s(e)))}}d().catch(e=>{l.error(`Fatal error`,s(e)),process.exit(1)});export{};
@@ -7,5 +7,4 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
7
7
  */
8
8
  declare function installReplayInterceptor(server: McpServer): void;
9
9
  //#endregion
10
- export { installReplayInterceptor };
11
- //# sourceMappingURL=replay-interceptor.d.ts.map
10
+ export { installReplayInterceptor };
@@ -1,2 +1 @@
1
- import{replayAppend as e}from"../../tools/dist/index.js";const t=/key|token|secret|auth|password|bearer/i,n=new Set([`eval`,`env`]);function r(e,r){if(n.has(e))return JSON.stringify({_redacted:!0,tool:e});if(e===`http`&&r.headers&&typeof r.headers==`object`){let e={...r},n={};for(let[e,i]of Object.entries(r.headers))n[e]=t.test(e)?`[REDACTED]`:i;return e.headers=n,JSON.stringify(e).slice(0,200)}return JSON.stringify(r).slice(0,200)}function i(t){let n=t.registerTool.bind(t);t.registerTool=(t,i,a)=>n(t,i,async(n,i)=>{let o=Date.now();try{let s=await a(n,i);return e({ts:new Date().toISOString(),source:`mcp`,tool:t,input:r(t,n),durationMs:Date.now()-o,status:`ok`,output:JSON.stringify(s).slice(0,200)}),s}catch(i){throw e({ts:new Date().toISOString(),source:`mcp`,tool:t,input:r(t,n),durationMs:Date.now()-o,status:`error`,output:i instanceof Error?i.message:String(i)}),i}})}export{i as installReplayInterceptor};
2
- //# sourceMappingURL=replay-interceptor.js.map
1
+ import{replayAppend as e}from"../../tools/dist/index.js";const t=/key|token|secret|auth|password|bearer/i,n=new Set([`eval`,`env`]);function r(e,r){if(n.has(e))return JSON.stringify({_redacted:!0,tool:e});if(e===`http`&&r.headers&&typeof r.headers==`object`){let e={...r},n={};for(let[e,i]of Object.entries(r.headers))n[e]=t.test(e)?`[REDACTED]`:i;return e.headers=n,JSON.stringify(e).slice(0,200)}return JSON.stringify(r).slice(0,200)}function i(t){let n=t.registerTool.bind(t);t.registerTool=(t,i,a)=>n(t,i,async(n,i)=>{let o=Date.now();try{let s=await a(n,i);return e({ts:new Date().toISOString(),source:`mcp`,tool:t,input:r(t,n),durationMs:Date.now()-o,status:`ok`,output:JSON.stringify(s).slice(0,200)}),s}catch(i){throw e({ts:new Date().toISOString(),source:`mcp`,tool:t,input:r(t,n),durationMs:Date.now()-o,status:`error`,output:i instanceof Error?i.message:String(i)}),i}})}export{i as installReplayInterceptor};
@@ -4,5 +4,4 @@ import { IKnowledgeStore } from "@kb/store";
4
4
  //#region packages/server/src/resources/resources.d.ts
5
5
  declare function registerResources(server: McpServer, store: IKnowledgeStore): void;
6
6
  //#endregion
7
- export { registerResources };
8
- //# sourceMappingURL=resources.d.ts.map
7
+ export { registerResources };
@@ -1,3 +1,2 @@
1
1
  function e(e,t){e.resource(`kb-status`,`kb://status`,{description:`Current knowledge base status and statistics`,mimeType:`text/plain`},async()=>{let e=await t.getStats();return{contents:[{uri:`kb://status`,text:`Knowledge Base: ${e.totalRecords} records from ${e.totalFiles} files. Last indexed: ${e.lastIndexedAt??`Never`}`,mimeType:`text/plain`}]}}),e.resource(`kb-file-tree`,`kb://file-tree`,{description:`List of all indexed source files`,mimeType:`text/plain`},async()=>({contents:[{uri:`kb://file-tree`,text:(await t.listSourcePaths()).sort().join(`
2
- `),mimeType:`text/plain`}]}))}export{e as registerResources};
3
- //# sourceMappingURL=resources.js.map
2
+ `),mimeType:`text/plain`}]}))}export{e as registerResources};
@@ -19,6 +19,8 @@ interface KnowledgeBaseComponents {
19
19
  bridge?: BridgeComponents;
20
20
  policyStore?: PolicyStore;
21
21
  evolutionCollector?: EvolutionCollector;
22
+ onboardComplete: boolean;
23
+ onboardTimestamp?: string;
22
24
  }
23
25
  declare function initializeKnowledgeBase(config: KBConfig): Promise<KnowledgeBaseComponents>;
24
26
  declare function createMcpServer(kb: KnowledgeBaseComponents, config: KBConfig): McpServer;
@@ -31,11 +33,11 @@ declare function createServer(config: KBConfig): Promise<{
31
33
  runInitialIndex: () => Promise<void>;
32
34
  shutdown: () => Promise<void>;
33
35
  }>;
36
+ declare const ALL_TOOL_NAMES: readonly ["analyze_dependencies", "analyze_diagram", "analyze_entry_points", "analyze_patterns", "analyze_structure", "analyze_symbols", "audit", "batch", "blast_radius", "changelog", "check", "checkpoint", "codemod", "compact", "data_transform", "dead_symbols", "delegate", "diff_parse", "digest", "encode", "env", "eval", "evidence_map", "file_summary", "find", "forge_classify", "forge_ground", "forget", "git_context", "graph", "guide", "health", "http", "lane", "list", "lookup", "measure", "onboard", "parse_output", "process", "produce_knowledge", "queue", "read", "regex_test", "reindex", "remember", "rename", "replay", "schema_validate", "scope_map", "search", "snippet", "stash", "status", "stratum_card", "symbol", "test_run", "time", "trace", "update", "watch", "web_fetch", "web_search", "workset"];
34
37
  declare function createLazyServer(config: KBConfig): {
35
38
  server: McpServer;
36
39
  ready: Promise<void>;
37
40
  runInitialIndex: () => Promise<void>;
38
41
  };
39
42
  //#endregion
40
- export { KnowledgeBaseComponents, createLazyServer, createMcpServer, createServer, initializeKnowledgeBase, registerMcpTools };
41
- //# sourceMappingURL=server.d.ts.map
43
+ export { ALL_TOOL_NAMES, KnowledgeBaseComponents, createLazyServer, createMcpServer, createServer, initializeKnowledgeBase, registerMcpTools };
@@ -1,2 +1 @@
1
- import{CuratedKnowledgeManager as e}from"./curated-manager.js";import{installReplayInterceptor as t}from"./replay-interceptor.js";import{registerResources as n}from"./resources/resources.js";import{registerAnalyzeDependenciesTool as r,registerAnalyzeDiagramTool as i,registerAnalyzeEntryPointsTool as a,registerAnalyzePatternsTool as o,registerAnalyzeStructureTool as s,registerAnalyzeSymbolsTool as c,registerBlastRadiusTool as l}from"./tools/analyze.tools.js";import{registerAuditTool as u}from"./tools/audit.tool.js";import{initBridgeComponents as d,registerErPullTool as ee,registerErPushTool as te,registerErSyncStatusTool as ne}from"./tools/bridge.tools.js";import{registerErEvolveReviewTool as re}from"./tools/evolution.tools.js";import{registerDigestTool as ie,registerEvidenceMapTool as ae,registerForgeClassifyTool as oe,registerForgeGroundTool as se,registerStratumCardTool as ce}from"./tools/forge.tools.js";import{registerForgetTool as le}from"./tools/forget.tool.js";import{registerGraphTool as f}from"./tools/graph.tool.js";import{registerListTool as p}from"./tools/list.tool.js";import{registerLookupTool as m}from"./tools/lookup.tool.js";import{registerOnboardTool as h}from"./tools/onboard.tool.js";import{registerErUpdatePolicyTool as g}from"./tools/policy.tools.js";import{registerProduceKnowledgeTool as _}from"./tools/produce.tool.js";import{registerReadTool as v}from"./tools/read.tool.js";import{registerReindexTool as y}from"./tools/reindex.tool.js";import{registerRememberTool as b}from"./tools/remember.tool.js";import{registerReplayTool as x}from"./tools/replay.tool.js";import{registerSearchTool as S}from"./tools/search.tool.js";import{registerStatusTool as C}from"./tools/status.tool.js";import{registerBatchTool as w,registerCheckTool as T,registerCheckpointTool as E,registerCodemodTool as D,registerCompactTool as O,registerDataTransformTool as k,registerDeadSymbolsTool as A,registerDelegateTool as j,registerDiffParseTool as M,registerEvalTool as N,registerFileSummaryTool as P,registerFindTool as F,registerGitContextTool as I,registerGuideTool as L,registerHealthTool as R,registerLaneTool as z,registerParseOutputTool as B,registerProcessTool as V,registerQueueTool as H,registerRenameTool as U,registerScopeMapTool as W,registerStashTool as ue,registerSymbolTool as de,registerTestRunTool as fe,registerTraceTool as pe,registerWatchTool as me,registerWebFetchTool as he,registerWorksetTool as ge}from"./tools/toolkit.tools.js";import{registerUpdateTool as _e}from"./tools/update.tool.js";import{registerChangelogTool as ve,registerEncodeTool as ye,registerEnvTool as be,registerHttpTool as xe,registerMeasureTool as Se,registerRegexTestTool as Ce,registerSchemaValidateTool as we,registerSnippetTool as Te,registerTimeTool as Ee,registerWebSearchTool as De}from"./tools/utility.tools.js";import{getCurrentVersion as G}from"./version-check.js";import{createLogger as Oe,serializeError as K}from"../../core/dist/index.js";import{initializeTreeSitter as ke}from"../../chunker/dist/index.js";import{OnnxEmbedder as Ae}from"../../embeddings/dist/index.js";import{EvolutionCollector as je,PolicyStore as Me}from"../../enterprise-bridge/dist/index.js";import{IncrementalIndexer as Ne}from"../../indexer/dist/index.js";import{SqliteGraphStore as q,createStore as Pe}from"../../store/dist/index.js";import{FileCache as Fe}from"../../tools/dist/index.js";import{McpServer as J}from"@modelcontextprotocol/sdk/server/mcp.js";const Y=Oe(`server`);async function X(t){Y.info(`Initializing knowledge base components`);let n=new Ae({model:t.embedding.model,dimensions:t.embedding.dimensions});await n.initialize(),Y.info(`Embedder loaded`,{modelId:n.modelId,dimensions:n.dimensions});let r=await Pe({backend:t.store.backend,path:t.store.path});await r.initialize(),Y.info(`Store initialized`);let i=new Ne(n,r),a=t.curated.path,o=new e(a,r,n),s=new q({path:t.store.path});await s.initialize(),Y.info(`Graph store initialized`),i.setGraphStore(s),await ke()?Y.info(`Tree-sitter chunking enabled`):Y.warn(`Tree-sitter not available; using regex-based code chunking`);let c=d(t.er),l=c?new Me(t.curated.path):void 0;l&&Y.info(`Policy store initialized`,{ruleCount:l.getRules().length});let u=c?new je:void 0;return{embedder:n,store:r,indexer:i,curated:o,graphStore:s,fileCache:new Fe,bridge:c,policyStore:l,evolutionCollector:u}}function Z(e,t){let n=new J({name:t.serverName??`knowledge-base`,version:G()});return Q(n,e,t),n}function Q(e,d,G){t(e),S(e,d.embedder,d.store,d.graphStore,d.bridge,d.evolutionCollector),m(e,d.store),C(e,d.store,d.graphStore),y(e,d.indexer,G,d.curated,d.store),b(e,d.curated,d.policyStore,d.evolutionCollector),_e(e,d.curated),le(e,d.curated),v(e,d.curated),p(e,d.curated),s(e,d.store,d.embedder),r(e,d.store,d.embedder),c(e,d.store,d.embedder),o(e,d.store,d.embedder),a(e,d.store,d.embedder),i(e,d.store,d.embedder),l(e,d.store,d.embedder),_(e),h(e,d.store,d.embedder),f(e,d.graphStore),u(e,d.store,d.embedder),O(e,d.embedder,d.fileCache),W(e,d.embedder,d.store),F(e,d.embedder,d.store),B(e),ge(e),T(e),w(e,d.embedder,d.store),de(e,d.embedder,d.store),N(e),fe(e),ue(e),I(e),M(e),U(e),D(e),P(e,d.fileCache),E(e),k(e),pe(e,d.embedder,d.store),V(e),me(e),A(e,d.embedder,d.store),j(e),R(e),z(e),H(e),he(e),L(e),ae(e),ie(e,d.embedder),oe(e),ce(e,d.embedder,d.fileCache),se(e,d.embedder,d.store),De(e),xe(e),Ce(e),ye(e),Se(e),ve(e),we(e),Te(e),be(e),Ee(e),d.bridge&&(te(e,d.bridge,d.evolutionCollector),ee(e,d.bridge),ne(e,d.bridge)),d.policyStore&&g(e,d.policyStore),d.evolutionCollector&&re(e,d.evolutionCollector),n(e,d.store),x(e)}async function Ie(e){let t=await X(e),n=Z(t,e);Y.info(`MCP server configured`,{toolCount:$.length,resourceCount:2});let r=async()=>{try{let n=e.sources.map(e=>e.path).join(`, `);Y.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Y.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Y.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});Y.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Y.warn(`FTS index creation failed`,K(e))}try{let e=await t.curated.reindexAll();Y.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Y.error(`Curated re-index failed`,K(e))}}catch(e){Y.error(`Initial index failed; will retry on kb_reindex`,K(e))}},i=async()=>{Y.info(`Shutting down`),await t.embedder.shutdown().catch(()=>{}),await t.graphStore.close().catch(()=>{}),await t.store.close(),process.exit(0)};process.on(`SIGINT`,i),process.on(`SIGTERM`,i);let a=process.ppid,o=setInterval(()=>{try{process.kill(a,0)}catch{Y.info(`Parent process died; shutting down`,{parentPid:a}),clearInterval(o),i()}},5e3);return o.unref(),{server:n,runInitialIndex:r,shutdown:i}}const Le=new Set(`batch.changelog.check.checkpoint.codemod.data_transform.delegate.diff_parse.encode.env.eval.evidence_map.file_summary.forge_classify.git_context.guide.health.http.lane.measure.parse_output.process.queue.read.regex_test.reindex.remember.rename.replay.schema_validate.snippet.stash.test_run.time.update.forget.list.watch.web_fetch.web_search.workset`.split(`.`)),$=`analyze_dependencies.analyze_diagram.analyze_entry_points.analyze_patterns.analyze_structure.analyze_symbols.audit.batch.blast_radius.changelog.check.checkpoint.codemod.compact.data_transform.dead_symbols.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.find.forge_classify.forge_ground.forget.git_context.graph.guide.health.http.lane.list.lookup.measure.onboard.parse_output.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.schema_validate.scope_map.search.snippet.stash.status.stratum_card.symbol.test_run.time.trace.update.watch.web_fetch.web_search.workset`.split(`.`);function Re(e){let t=new J({name:e.serverName??`knowledge-base`,version:G()}),n=$.map(e=>t.registerTool(e,{description:`${e} (initializing...)`,inputSchema:{}},async()=>({content:[{type:`text`,text:`KB is still initializing, please retry in a few seconds.`}]}))),r=t.resource(`kb-status`,`kb://status`,{description:`Knowledge base status (initializing...)`,mimeType:`text/plain`},async()=>({contents:[{uri:`kb://status`,text:`KB is initializing...`,mimeType:`text/plain`}]})),i,a=new Promise(e=>{i=e}),o=(async()=>{let a=await X(e);for(let e of n)e.remove();r.remove(),Q(t,a,e);let o=t._registeredTools??{};for(let[e,t]of Object.entries(o)){if(Le.has(e))continue;let n=t.handler;t.handler=async(...t)=>{if(!a.indexer.isIndexing)return n(...t);let r=new Promise(t=>setTimeout(()=>t({content:[{type:`text`,text:`⏳ KB is currently indexing. The tool "${e}" requires indexed data and cannot complete right now.\n\nPlease retry in a minute, or use index-independent tools (eval, check, encode, regex_test, stash, etc.) in the meantime.\n\nTo check progress, try again shortly — indexing will complete automatically.`}]}),5e3));return Promise.race([n(...t),r])}}let s=Object.keys(o).length;s!==$.length&&Y.warn(`ALL_TOOL_NAMES count mismatch`,{expectedToolCount:$.length,registeredToolCount:s}),Y.info(`MCP server configured`,{toolCount:$.length,resourceCount:2}),i?.(a)})(),s=async()=>{let t=await a;try{let n=e.sources.map(e=>e.path).join(`, `);Y.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Y.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Y.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});Y.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Y.warn(`FTS index creation failed`,K(e))}try{let e=await t.curated.reindexAll();Y.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Y.error(`Curated re-index failed`,K(e))}}catch(e){Y.error(`Initial index failed; will retry on kb_reindex`,K(e))}},c=process.ppid,l=setInterval(()=>{try{process.kill(c,0)}catch{Y.info(`Parent process died; shutting down`,{parentPid:c}),clearInterval(l),a.then(async e=>{await e.embedder.shutdown().catch(()=>{}),await e.graphStore.close().catch(()=>{}),await e.store.close().catch(()=>{})}).catch(()=>{}).finally(()=>process.exit(0))}},5e3);return l.unref(),{server:t,ready:o,runInitialIndex:s}}export{Re as createLazyServer,Z as createMcpServer,Ie as createServer,X as initializeKnowledgeBase,Q as registerMcpTools};
2
- //# sourceMappingURL=server.js.map
1
+ import{CuratedKnowledgeManager as e}from"./curated-manager.js";import{installReplayInterceptor as t}from"./replay-interceptor.js";import{registerResources as n}from"./resources/resources.js";import{registerAnalyzeDependenciesTool as r,registerAnalyzeDiagramTool as i,registerAnalyzeEntryPointsTool as a,registerAnalyzePatternsTool as o,registerAnalyzeStructureTool as s,registerAnalyzeSymbolsTool as c,registerBlastRadiusTool as l}from"./tools/analyze.tools.js";import{registerAuditTool as u}from"./tools/audit.tool.js";import{initBridgeComponents as d,registerErPullTool as f,registerErPushTool as p,registerErSyncStatusTool as m}from"./tools/bridge.tools.js";import{registerErEvolveReviewTool as ee}from"./tools/evolution.tools.js";import{registerDigestTool as te,registerEvidenceMapTool as ne,registerForgeClassifyTool as re,registerForgeGroundTool as ie,registerStratumCardTool as ae}from"./tools/forge.tools.js";import{registerForgetTool as oe}from"./tools/forget.tool.js";import{registerGraphTool as se}from"./tools/graph.tool.js";import{registerListTool as ce}from"./tools/list.tool.js";import{registerLookupTool as le}from"./tools/lookup.tool.js";import{registerOnboardTool as ue}from"./tools/onboard.tool.js";import{registerErUpdatePolicyTool as h}from"./tools/policy.tools.js";import{registerProduceKnowledgeTool as g}from"./tools/produce.tool.js";import{registerReadTool as _}from"./tools/read.tool.js";import{registerReindexTool as v}from"./tools/reindex.tool.js";import{registerRememberTool as y}from"./tools/remember.tool.js";import{registerReplayTool as b}from"./tools/replay.tool.js";import{registerSearchTool as x}from"./tools/search.tool.js";import{registerStatusTool as S}from"./tools/status.tool.js";import{registerBatchTool as C,registerCheckTool as w,registerCheckpointTool as T,registerCodemodTool as E,registerCompactTool as D,registerDataTransformTool as O,registerDeadSymbolsTool as k,registerDelegateTool as A,registerDiffParseTool as j,registerEvalTool as M,registerFileSummaryTool as N,registerFindTool as P,registerGitContextTool as F,registerGuideTool as I,registerHealthTool as L,registerLaneTool as R,registerParseOutputTool as z,registerProcessTool as B,registerQueueTool as V,registerRenameTool as H,registerScopeMapTool as U,registerStashTool as de,registerSymbolTool as fe,registerTestRunTool as pe,registerTraceTool as me,registerWatchTool as he,registerWebFetchTool as ge,registerWorksetTool as _e}from"./tools/toolkit.tools.js";import{registerUpdateTool as W}from"./tools/update.tool.js";import{registerChangelogTool as ve,registerEncodeTool as ye,registerEnvTool as be,registerHttpTool as xe,registerMeasureTool as Se,registerRegexTestTool as Ce,registerSchemaValidateTool as we,registerSnippetTool as Te,registerTimeTool as Ee,registerWebSearchTool as De}from"./tools/utility.tools.js";import{getCurrentVersion as G}from"./version-check.js";import{existsSync as Oe,statSync as ke}from"node:fs";import{resolve as Ae}from"node:path";import{KB_PATHS as je,createLogger as Me,serializeError as K}from"../../core/dist/index.js";import{initializeWasm as Ne}from"../../chunker/dist/index.js";import{OnnxEmbedder as Pe}from"../../embeddings/dist/index.js";import{EvolutionCollector as Fe,PolicyStore as Ie}from"../../enterprise-bridge/dist/index.js";import{IncrementalIndexer as Le}from"../../indexer/dist/index.js";import{SqliteGraphStore as Re,createStore as ze}from"../../store/dist/index.js";import{FileCache as Be}from"../../tools/dist/index.js";import{McpServer as q}from"@modelcontextprotocol/sdk/server/mcp.js";const J=Me(`server`);async function Y(t){J.info(`Initializing knowledge base components`);let n=new Pe({model:t.embedding.model,dimensions:t.embedding.dimensions});await n.initialize(),J.info(`Embedder loaded`,{modelId:n.modelId,dimensions:n.dimensions});let r=await ze({backend:t.store.backend,path:t.store.path});await r.initialize(),J.info(`Store initialized`);let i=new Le(n,r),a=t.curated.path,o=new e(a,r,n),s=new Re({path:t.store.path});await s.initialize(),J.info(`Graph store initialized`),i.setGraphStore(s),await Ne()?J.info(`WASM tree-sitter enabled for AST analysis`):J.warn(`WASM tree-sitter not available; analyzers will use regex fallback`);let c=d(t.er),l=c?new Ie(t.curated.path):void 0;l&&J.info(`Policy store initialized`,{ruleCount:l.getRules().length});let u=c?new Fe:void 0,f=Ae(process.cwd(),je.aiKb),p=Oe(f),m;if(p)try{m=ke(f).mtime.toISOString()}catch{}return J.info(`Onboard state detected`,{onboardComplete:p,onboardTimestamp:m}),{embedder:n,store:r,indexer:i,curated:o,graphStore:s,fileCache:new Be,bridge:c,policyStore:l,evolutionCollector:u,onboardComplete:p,onboardTimestamp:m}}function X(e,t){let n=new q({name:t.serverName??`knowledge-base`,version:G()});return Z(n,e,t),n}function Z(e,d,G){t(e),x(e,d.embedder,d.store,d.graphStore,d.bridge,d.evolutionCollector),le(e,d.store),S(e,d.store,d.graphStore,d.curated,{onboardComplete:d.onboardComplete,onboardTimestamp:d.onboardTimestamp}),v(e,d.indexer,G,d.curated,d.store),y(e,d.curated,d.policyStore,d.evolutionCollector),W(e,d.curated),oe(e,d.curated),_(e,d.curated),ce(e,d.curated),s(e,d.store,d.embedder),r(e,d.store,d.embedder),c(e,d.store,d.embedder),o(e,d.store,d.embedder),a(e,d.store,d.embedder),i(e,d.store,d.embedder),l(e,d.store,d.embedder),g(e),ue(e,d.store,d.embedder),se(e,d.graphStore),u(e,d.store,d.embedder),D(e,d.embedder,d.fileCache),U(e,d.embedder,d.store),P(e,d.embedder,d.store),z(e),_e(e),w(e),C(e,d.embedder,d.store),fe(e,d.embedder,d.store),M(e),pe(e),de(e),F(e),j(e),H(e),E(e),N(e,d.fileCache),T(e),O(e),me(e,d.embedder,d.store),B(e),he(e),k(e,d.embedder,d.store),A(e),L(e),R(e),V(e),ge(e),I(e),ne(e),te(e,d.embedder),re(e),ae(e,d.embedder,d.fileCache),ie(e,d.embedder,d.store),De(e),xe(e),Ce(e),ye(e),Se(e),ve(e),we(e),Te(e),be(e),Ee(e),d.bridge&&(p(e,d.bridge,d.evolutionCollector),f(e,d.bridge),m(e,d.bridge)),d.policyStore&&h(e,d.policyStore),d.evolutionCollector&&ee(e,d.evolutionCollector),n(e,d.store),b(e)}async function Ve(e){let t=await Y(e),n=X(t,e);J.info(`MCP server configured`,{toolCount:$.length,resourceCount:2});let r=async()=>{try{let n=e.sources.map(e=>e.path).join(`, `);J.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&J.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&J.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});J.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){J.warn(`FTS index creation failed`,K(e))}try{let e=await t.curated.reindexAll();J.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){J.error(`Curated re-index failed`,K(e))}}catch(e){J.error(`Initial index failed; will retry on kb_reindex`,K(e))}},i=async()=>{J.info(`Shutting down`),await t.embedder.shutdown().catch(()=>{}),await t.graphStore.close().catch(()=>{}),await t.store.close(),process.exit(0)};process.on(`SIGINT`,i),process.on(`SIGTERM`,i);let a=process.ppid,o=setInterval(()=>{try{process.kill(a,0)}catch{J.info(`Parent process died; shutting down`,{parentPid:a}),clearInterval(o),i()}},5e3);return o.unref(),{server:n,runInitialIndex:r,shutdown:i}}const He=new Set(`batch.changelog.check.checkpoint.codemod.compact.data_transform.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.forge_classify.git_context.graph.guide.health.http.lane.measure.onboard.parse_output.process.queue.read.regex_test.reindex.remember.rename.replay.schema_validate.scope_map.snippet.stash.status.stratum_card.test_run.time.update.forget.list.watch.web_fetch.web_search.workset`.split(`.`)),Q=5e3,$=`analyze_dependencies.analyze_diagram.analyze_entry_points.analyze_patterns.analyze_structure.analyze_symbols.audit.batch.blast_radius.changelog.check.checkpoint.codemod.compact.data_transform.dead_symbols.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.find.forge_classify.forge_ground.forget.git_context.graph.guide.health.http.lane.list.lookup.measure.onboard.parse_output.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.schema_validate.scope_map.search.snippet.stash.status.stratum_card.symbol.test_run.time.trace.update.watch.web_fetch.web_search.workset`.split(`.`);function Ue(e){let t=new q({name:e.serverName??`knowledge-base`,version:G()}),n=t.sendToolListChanged.bind(t);t.sendToolListChanged=()=>{};let r=$.map(e=>t.registerTool(e,{description:`${e} (initializing...)`,inputSchema:{}},async()=>({content:[{type:`text`,text:`KB is still initializing, please retry in a few seconds.`}]})));t.sendToolListChanged=n;let i=t.resource(`kb-status`,`kb://status`,{description:`Knowledge base status (initializing...)`,mimeType:`text/plain`},async()=>({contents:[{uri:`kb://status`,text:`KB is initializing...`,mimeType:`text/plain`}]})),a,o=new Promise(e=>{a=e}),s=(async()=>{let n=await Y(e),o=t.sendToolListChanged.bind(t);t.sendToolListChanged=()=>{};for(let e of r)e.remove();i.remove(),Z(t,n,e),t.sendToolListChanged=o,t.sendToolListChanged();let s=t._registeredTools??{};for(let[e,t]of Object.entries(s)){if(He.has(e))continue;let r=t.handler;t.handler=async(...t)=>{if(!n.indexer.isIndexing)return r(...t);let i=new Promise(t=>setTimeout(()=>t({content:[{type:`text`,text:`⏳ KB is re-indexing. The tool "${e}" timed out waiting for index data (${Q/1e3}s).\n\nThe existing index may be temporarily locked. Please retry shortly — indexing will complete automatically.`}]}),Q));return Promise.race([r(...t),i])}}let c=Object.keys(s).length;c!==$.length&&J.warn(`ALL_TOOL_NAMES count mismatch`,{expectedToolCount:$.length,registeredToolCount:c}),J.info(`MCP server configured`,{toolCount:$.length,resourceCount:2}),a?.(n)})(),c=async()=>{let t=await o;try{let n=e.sources.map(e=>e.path).join(`, `);J.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&J.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&J.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});J.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){J.warn(`FTS index creation failed`,K(e))}try{let e=await t.curated.reindexAll();J.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){J.error(`Curated re-index failed`,K(e))}}catch(e){J.error(`Initial index failed; will retry on kb_reindex`,K(e))}},l=process.ppid,u=setInterval(()=>{try{process.kill(l,0)}catch{J.info(`Parent process died; shutting down`,{parentPid:l}),clearInterval(u),o.then(async e=>{await e.embedder.shutdown().catch(()=>{}),await e.graphStore.close().catch(()=>{}),await e.store.close().catch(()=>{})}).catch(()=>{}).finally(()=>process.exit(0))}},5e3);return u.unref(),{server:t,ready:s,runInitialIndex:c}}export{$ as ALL_TOOL_NAMES,Ue as createLazyServer,X as createMcpServer,Ve as createServer,Y as initializeKnowledgeBase,Z as registerMcpTools};
@@ -11,5 +11,4 @@ declare function registerAnalyzeEntryPointsTool(server: McpServer, store: IKnowl
11
11
  declare function registerAnalyzeDiagramTool(server: McpServer, store: IKnowledgeStore, embedder: IEmbedder): void;
12
12
  declare function registerBlastRadiusTool(server: McpServer, store: IKnowledgeStore, embedder: IEmbedder): void;
13
13
  //#endregion
14
- export { registerAnalyzeDependenciesTool, registerAnalyzeDiagramTool, registerAnalyzeEntryPointsTool, registerAnalyzePatternsTool, registerAnalyzeStructureTool, registerAnalyzeSymbolsTool, registerBlastRadiusTool };
15
- //# sourceMappingURL=analyze.tools.d.ts.map
14
+ export { registerAnalyzeDependenciesTool, registerAnalyzeDiagramTool, registerAnalyzeEntryPointsTool, registerAnalyzePatternsTool, registerAnalyzeStructureTool, registerAnalyzeSymbolsTool, registerBlastRadiusTool };
@@ -1,2 +1 @@
1
- import{createHash as e}from"node:crypto";import{BlastRadiusAnalyzer as t,DependencyAnalyzer as n,DiagramGenerator as r,EntryPointAnalyzer as i,PatternAnalyzer as a,StructureAnalyzer as o,SymbolAnalyzer as s}from"../../../analyzers/dist/index.js";import{TreeSitterRuntime as c}from"../../../chunker/dist/index.js";import{createLogger as l,serializeError as u}from"../../../core/dist/index.js";import{truncateToTokenBudget as d}from"../../../tools/dist/index.js";import{z as f}from"zod";const p=l(`tools`),m=f.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`);function h(e,t){return t?d(e,t):e}function g(){let e=[];return c.get()||e.push(`Tree-sitter unavailable — using regex fallback, symbol/pattern confidence reduced`),e.length===0?``:`\n\n> **⚠ Caveats:** ${e.join(`; `)}`}async function _(t,n,r,i,a){try{let o=`produced/analysis/${r}/${e(`sha256`).update(i).digest(`hex`).slice(0,12)}.md`,s=e(`sha256`).update(a).digest(`hex`).slice(0,16),c=new Date().toISOString(),l=a.length>2e3?a.split(/(?=^## )/m).filter(e=>e.trim().length>0):[a],u=l.map((t,n)=>({id:e(`sha256`).update(`${o}::${n}`).digest(`hex`).slice(0,16),content:t.trim(),sourcePath:o,contentType:`produced-knowledge`,chunkIndex:n,totalChunks:l.length,startLine:0,endLine:0,fileHash:s,indexedAt:c,origin:`produced`,tags:[`analysis`,r],category:`analysis`,version:1})),d=await n.embedBatch(u.map(e=>e.content));await t.upsert(u,d),p.info(`Auto-persisted analysis`,{analyzerName:r,chunkCount:u.length})}catch(e){p.warn(`Auto-persist analysis failed`,{analyzerName:r,...u(e)})}}function v(e,t,n){let r=new o;e.registerTool(`analyze_structure`,{description:`Analyze the file/directory structure of a codebase. Returns an annotated tree with language stats.`,inputSchema:{path:f.string().describe(`Root path to analyze`),max_depth:f.number().min(1).max(10).default(6).describe(`Maximum directory depth`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,max_depth:i,format:a,max_tokens:o})=>{try{let s=await r.analyze(e,{format:a,maxDepth:i});return _(t,n,`structure`,e,s.output),{content:[{type:`text`,text:h(s.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` for import graphs, or `analyze_patterns` to detect architecture patterns._",o)}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed. Check server logs for details.`}],isError:!0}}})}function y(e,t,r){let i=new n;e.registerTool(`analyze_dependencies`,{description:`Analyze import/require dependencies across a codebase. Shows external packages and internal module graph.`,inputSchema:{path:f.string().describe(`Root path to analyze`),format:f.enum([`json`,`markdown`,`mermaid`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,format:n,max_tokens:a})=>{try{let o=await i.analyze(e,{format:n});return _(t,r,`dependencies`,e,o.output),{content:[{type:`text`,text:h(o.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_symbols` to explore exported symbols, or `analyze_diagram` for visual representation._",a)}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed. Check server logs for details.`}],isError:!0}}})}function b(e,t,n){let r=new s;e.registerTool(`analyze_symbols`,{description:`Extract exported and local symbols (functions, classes, interfaces, types, constants) from a codebase.`,inputSchema:{path:f.string().describe(`Root path to analyze`),filter:f.string().optional().describe(`Filter symbols by name substring`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`)}},async({path:e,filter:i,format:a})=>{try{let o=await r.analyze(e,{format:a,filter:i});return _(t,n,`symbols`,e,o.output),{content:[{type:`text`,text:o.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see import relationships, or `search` to find usage patterns._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed. Check server logs for details.`}],isError:!0}}})}function x(e,t,n){let r=new a;e.registerTool(`analyze_patterns`,{description:`Detect architectural patterns, frameworks, and conventions in a codebase using directory structure and code heuristics.`,inputSchema:{path:f.string().describe(`Root path to analyze`)}},async({path:e})=>{try{let i=await r.analyze(e);return _(t,n,`patterns`,e,i.output),{content:[{type:`text`,text:i.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_entry_points` to find Lambda handlers and main exports, or `produce_knowledge` for full analysis._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed. Check server logs for details.`}],isError:!0}}})}function S(e,t,n){let r=new i;e.registerTool(`analyze_entry_points`,{description:`Find entry points: Lambda handlers, main exports, CLI bins, and server start scripts.`,inputSchema:{path:f.string().describe(`Root path to analyze`)}},async({path:e})=>{try{let i=await r.analyze(e);return _(t,n,`entry-points`,e,i.output),{content:[{type:`text`,text:i.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see what each entry point imports, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed. Check server logs for details.`}],isError:!0}}})}function C(e,t,n){let i=new r;e.registerTool(`analyze_diagram`,{description:`Generate a Mermaid diagram of the codebase architecture or dependency graph.`,inputSchema:{path:f.string().describe(`Root path to analyze`),diagram_type:f.enum([`architecture`,`dependencies`]).default(`architecture`).describe(`Type of diagram`)}},async({path:e,diagram_type:r})=>{try{let a=await i.analyze(e,{diagramType:r});return _(t,n,`diagram`,e,a.output),{content:[{type:`text`,text:a.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_structure` for detailed file tree, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return p.error(`Diagram generation failed`,u(e)),{content:[{type:`text`,text:`Diagram generation failed. Check server logs for details.`}],isError:!0}}})}function w(e,n,r){let i=new t;e.registerTool(`blast_radius`,{description:`Given a list of changed files, trace the dependency graph to find all affected files (direct + transitive importers) and their tests. Useful for scoping code reviews and impact analysis.`,inputSchema:{path:f.string().describe(`Root path of the codebase`),files:f.array(f.string()).min(1).describe(`Changed file paths (relative to root)`),max_depth:f.number().min(1).max(20).default(5).describe(`Maximum transitive dependency depth`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,files:t,max_depth:a,format:o,max_tokens:s})=>{try{let c=await i.analyze(e,{files:t,maxDepth:a,format:o});return _(n,r,`blast-radius`,e,c.output),{content:[{type:`text`,text:h(c.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see the full import graph, or `analyze_symbols` to inspect affected exports._",s)}]}}catch(e){return p.error(`Blast radius analysis failed`,u(e)),{content:[{type:`text`,text:`Blast radius analysis failed. Check server logs for details.`}],isError:!0}}})}export{y as registerAnalyzeDependenciesTool,C as registerAnalyzeDiagramTool,S as registerAnalyzeEntryPointsTool,x as registerAnalyzePatternsTool,v as registerAnalyzeStructureTool,b as registerAnalyzeSymbolsTool,w as registerBlastRadiusTool};
2
- //# sourceMappingURL=analyze.tools.js.map
1
+ import{createHash as e}from"node:crypto";import{BlastRadiusAnalyzer as t,DependencyAnalyzer as n,DiagramGenerator as r,EntryPointAnalyzer as i,PatternAnalyzer as a,StructureAnalyzer as o,SymbolAnalyzer as s}from"../../../analyzers/dist/index.js";import{WasmRuntime as c}from"../../../chunker/dist/index.js";import{createLogger as l,serializeError as u}from"../../../core/dist/index.js";import{truncateToTokenBudget as d}from"../../../tools/dist/index.js";import{z as f}from"zod";const p=l(`tools`),m=f.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`);function h(e,t){return t?d(e,t):e}function g(){let e=[];return c.get()||e.push(`Tree-sitter unavailable — using regex fallback, symbol/pattern confidence reduced`),e.length===0?``:`\n\n> **⚠ Caveats:** ${e.join(`; `)}`}async function _(t,n,r,i,a){try{let o=`produced/analysis/${r}/${e(`sha256`).update(i).digest(`hex`).slice(0,12)}.md`,s=e(`sha256`).update(a).digest(`hex`).slice(0,16),c=new Date().toISOString(),l=a.length>2e3?a.split(/(?=^## )/m).filter(e=>e.trim().length>0):[a],u=l.map((t,n)=>({id:e(`sha256`).update(`${o}::${n}`).digest(`hex`).slice(0,16),content:t.trim(),sourcePath:o,contentType:`produced-knowledge`,chunkIndex:n,totalChunks:l.length,startLine:0,endLine:0,fileHash:s,indexedAt:c,origin:`produced`,tags:[`analysis`,r],category:`analysis`,version:1})),d=await n.embedBatch(u.map(e=>e.content));await t.upsert(u,d),p.info(`Auto-persisted analysis`,{analyzerName:r,chunkCount:u.length})}catch(e){p.warn(`Auto-persist analysis failed`,{analyzerName:r,...u(e)})}}function v(e,t,n){let r=new o;e.registerTool(`analyze_structure`,{description:`Analyze the file/directory structure of a codebase. Returns an annotated tree with language stats.`,inputSchema:{path:f.string().describe(`Root path to analyze`),max_depth:f.number().min(1).max(10).default(6).describe(`Maximum directory depth`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,max_depth:i,format:a,max_tokens:o})=>{try{let s=await r.analyze(e,{format:a,maxDepth:i});return _(t,n,`structure`,e,s.output),{content:[{type:`text`,text:h(s.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` for import graphs, or `analyze_patterns` to detect architecture patterns._",o)}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function y(e,t,r){let i=new n;e.registerTool(`analyze_dependencies`,{description:`Analyze import/require dependencies across a codebase. Shows external packages and internal module graph.`,inputSchema:{path:f.string().describe(`Root path to analyze`),format:f.enum([`json`,`markdown`,`mermaid`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,format:n,max_tokens:a})=>{try{let o=await i.analyze(e,{format:n});return _(t,r,`dependencies`,e,o.output),{content:[{type:`text`,text:h(o.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_symbols` to explore exported symbols, or `analyze_diagram` for visual representation._",a)}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function b(e,t,n){let r=new s;e.registerTool(`analyze_symbols`,{description:`Extract exported and local symbols (functions, classes, interfaces, types, constants) from a codebase.`,inputSchema:{path:f.string().describe(`Root path to analyze`),filter:f.string().optional().describe(`Filter symbols by name substring`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`)}},async({path:e,filter:i,format:a})=>{try{let o=await r.analyze(e,{format:a,filter:i});return _(t,n,`symbols`,e,o.output),{content:[{type:`text`,text:o.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see import relationships, or `search` to find usage patterns._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function x(e,t,n){let r=new a;e.registerTool(`analyze_patterns`,{description:`Detect architectural patterns, frameworks, and conventions in a codebase using directory structure and code heuristics.`,inputSchema:{path:f.string().describe(`Root path to analyze`)}},async({path:e})=>{try{let i=await r.analyze(e);return _(t,n,`patterns`,e,i.output),{content:[{type:`text`,text:i.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_entry_points` to find Lambda handlers and main exports, or `produce_knowledge` for full analysis._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function S(e,t,n){let r=new i;e.registerTool(`analyze_entry_points`,{description:`Find entry points: Lambda handlers, main exports, CLI bins, and server start scripts.`,inputSchema:{path:f.string().describe(`Root path to analyze`)}},async({path:e})=>{try{let i=await r.analyze(e);return _(t,n,`entry-points`,e,i.output),{content:[{type:`text`,text:i.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see what each entry point imports, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return p.error(`Analysis failed`,u(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function C(e,t,n){let i=new r;e.registerTool(`analyze_diagram`,{description:`Generate a Mermaid diagram of the codebase architecture or dependency graph.`,inputSchema:{path:f.string().describe(`Root path to analyze`),diagram_type:f.enum([`architecture`,`dependencies`]).default(`architecture`).describe(`Type of diagram`)}},async({path:e,diagram_type:r})=>{try{let a=await i.analyze(e,{diagramType:r});return _(t,n,`diagram`,e,a.output),{content:[{type:`text`,text:a.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_structure` for detailed file tree, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return p.error(`Diagram generation failed`,u(e)),{content:[{type:`text`,text:`Diagram generation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function w(e,n,r){let i=new t;e.registerTool(`blast_radius`,{description:`Given a list of changed files, trace the dependency graph to find all affected files (direct + transitive importers) and their tests. Useful for scoping code reviews and impact analysis.`,inputSchema:{path:f.string().describe(`Root path of the codebase`),files:f.array(f.string()).min(1).describe(`Changed file paths (relative to root)`),max_depth:f.number().min(1).max(20).default(5).describe(`Maximum transitive dependency depth`),format:f.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:m}},async({path:e,files:t,max_depth:a,format:o,max_tokens:s})=>{try{let c=await i.analyze(e,{files:t,maxDepth:a,format:o});return _(n,r,`blast-radius`,e,c.output),{content:[{type:`text`,text:h(c.output+g()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see the full import graph, or `analyze_symbols` to inspect affected exports._",s)}]}}catch(e){return p.error(`Blast radius analysis failed`,u(e)),{content:[{type:`text`,text:`Blast radius analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{y as registerAnalyzeDependenciesTool,C as registerAnalyzeDiagramTool,S as registerAnalyzeEntryPointsTool,x as registerAnalyzePatternsTool,v as registerAnalyzeStructureTool,b as registerAnalyzeSymbolsTool,w as registerBlastRadiusTool};
@@ -5,5 +5,4 @@ import { IKnowledgeStore } from "@kb/store";
5
5
  //#region packages/server/src/tools/audit.tool.d.ts
6
6
  declare function registerAuditTool(server: McpServer, store: IKnowledgeStore, embedder: IEmbedder): void;
7
7
  //#endregion
8
- export { registerAuditTool };
9
- //# sourceMappingURL=audit.tool.d.ts.map
8
+ export { registerAuditTool };
@@ -1,2 +1 @@
1
- import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{audit as n}from"../../../tools/dist/index.js";import{z as r}from"zod";const i=e(`tools`);function a(e,a,o){e.registerTool(`audit`,{description:`Run a unified audit: structure, dependencies, patterns, health, dead_symbols, check, entry_points. Returns synthesized report with score, recommendations, and next steps. 6 round-trips → 1.`,inputSchema:{path:r.string().default(`.`).describe(`Root path to audit`),checks:r.array(r.enum([`structure`,`dependencies`,`patterns`,`health`,`dead_symbols`,`check`,`entry_points`])).optional().describe(`Which checks to run (default: all)`),detail:r.enum([`summary`,`full`]).default(`summary`).describe(`'summary' for overview (~500 tokens), 'full' includes pattern table`)}},async({path:e,checks:r,detail:s})=>{try{let t=await n(a,o,{path:e,checks:r,detail:s});return{content:[{type:`text`,text:t.ok?`${t.summary}\n\n---\n_Audit score: ${t.data?.score}/100 | ${t.meta.durationMs}ms${t.next&&t.next.length>0?` | Next: ${t.next.map(e=>`\`${e.tool}\` (${e.reason})`).join(`, `)}`:``}_`:t.error?.message??`Audit failed`}],isError:!t.ok}}catch(e){return i.error(`Audit failed`,t(e)),{content:[{type:`text`,text:`Audit failed. Check server logs for details.`}],isError:!0}}})}export{a as registerAuditTool};
2
- //# sourceMappingURL=audit.tool.js.map
1
+ import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{audit as n}from"../../../tools/dist/index.js";import{z as r}from"zod";const i=e(`tools`);function a(e,a,o){e.registerTool(`audit`,{description:`Run a unified audit: structure, dependencies, patterns, health, dead_symbols, check, entry_points. Returns synthesized report with score, recommendations, and next steps. 6 round-trips → 1.`,inputSchema:{path:r.string().default(`.`).describe(`Root path to audit`),checks:r.array(r.enum([`structure`,`dependencies`,`patterns`,`health`,`dead_symbols`,`check`,`entry_points`])).optional().describe(`Which checks to run (default: all)`),detail:r.enum([`summary`,`full`]).default(`summary`).describe(`'summary' for overview (~500 tokens), 'full' includes pattern table`)}},async({path:e,checks:r,detail:s})=>{try{let t=await n(a,o,{path:e,checks:r,detail:s});return{content:[{type:`text`,text:t.ok?`${t.summary}\n\n---\n_Audit score: ${t.data?.score}/100 | ${t.meta.durationMs}ms${t.next&&t.next.length>0?` | Next: ${t.next.map(e=>`\`${e.tool}\` (${e.reason})`).join(`, `)}`:``}_`:t.error?.message??`Audit failed`}],isError:!t.ok}}catch(e){return i.error(`Audit failed`,t(e)),{content:[{type:`text`,text:`Audit failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{a as registerAuditTool};
@@ -31,5 +31,4 @@ declare function registerErPushTool(server: McpServer, bridge: BridgeComponents,
31
31
  declare function registerErPullTool(server: McpServer, bridge: BridgeComponents): void;
32
32
  declare function registerErSyncStatusTool(server: McpServer, bridge: BridgeComponents): void;
33
33
  //#endregion
34
- export { BridgeComponents, initBridgeComponents, registerErPullTool, registerErPushTool, registerErSyncStatusTool, resolveErBridgeConfig };
35
- //# sourceMappingURL=bridge.tools.d.ts.map
34
+ export { BridgeComponents, initBridgeComponents, registerErPullTool, registerErPushTool, registerErSyncStatusTool, resolveErBridgeConfig };
@@ -12,5 +12,4 @@ _The local KB is unaffected. You can retry or check ER health with \`er_sync_sta
12
12
 
13
13
  _Enterprise RAG may be unavailable. Local KB search is unaffected._`}],isError:!0}}})}function d(e,r){e.registerTool(`er_sync_status`,{description:`Show the status of the Enterprise RAG bridge: health, cache stats, and push history.`,inputSchema:{check_health:n.boolean().default(!0).describe(`Whether to ping ER health endpoint (adds latency)`)}},async({check_health:e})=>{try{let t=r.cache.stats(),n=r.pushAdapter.getStatus(),i=new URL(r.config.baseUrl),a=`${i.protocol}//${i.hostname}${i.port?`:${i.port}`:``}`,o=`_Health check skipped_`;if(e){let e=await r.client.health();o=e.healthy?`✅ Healthy (HTTP ${e.status})`:`❌ Unhealthy${e.status?` (HTTP ${e.status})`:``}`}let s=[`## ER Bridge Status
14
14
  `,`**Endpoint**: \`${a}\``,`**Health**: ${o}`,`**Threshold**: ${r.config.fallbackThreshold} (vector similarity for auto-fallback)`,``,`### Cache`,`- Entries: ${t.size} / ${t.maxEntries}`,`- TTL: ${Math.floor(t.defaultTtlMs/36e5)}h`,``,`### Push History`,`- Total pushed: ${n.totalPushed}`,`- Successful: ${n.successCount}`,`- Failed: ${n.failCount}`];return n.lastPush&&s.push(`- Last push: "${n.lastPush.title}" at ${n.lastPush.pushedAt} (${n.lastPush.status})`),s.push("\n---\n_Next: Use `er_push` to send knowledge to ER, or `er_pull` to search ER._"),{content:[{type:`text`,text:s.join(`
15
- `)}]}}catch(e){return o.error(`ER sync status failed`,t(e)),{content:[{type:`text`,text:`ER sync status failed. Check server logs for details.`}],isError:!0}}})}export{c as initBridgeComponents,u as registerErPullTool,l as registerErPushTool,d as registerErSyncStatusTool,s as resolveErBridgeConfig};
16
- //# sourceMappingURL=bridge.tools.js.map
15
+ `)}]}}catch(e){return o.error(`ER sync status failed`,t(e)),{content:[{type:`text`,text:`ER sync status failed. Check server logs for details.`}],isError:!0}}})}export{c as initBridgeComponents,u as registerErPullTool,l as registerErPushTool,d as registerErSyncStatusTool,s as resolveErBridgeConfig};
@@ -4,5 +4,4 @@ import { EvolutionCollector } from "@kb/enterprise-bridge";
4
4
  //#region packages/server/src/tools/evolution.tools.d.ts
5
5
  declare function registerErEvolveReviewTool(server: McpServer, collector: EvolutionCollector): void;
6
6
  //#endregion
7
- export { registerErEvolveReviewTool };
8
- //# sourceMappingURL=evolution.tools.d.ts.map
7
+ export { registerErEvolveReviewTool };
@@ -2,5 +2,4 @@ import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";i
2
2
  `,`**Period**: ${t.period.startedAt} → ${t.period.queriedAt}`,`**Total events**: ${t.period.totalEvents}\n`,`### Search`,`- Total searches: ${t.search.totalSearches}`,`- ER fallback triggered: ${t.search.erFallbackCount} (${(t.search.erFallbackRate*100).toFixed(1)}%)`,`- ER cache hits: ${t.search.erCacheHitCount} (${(t.search.erCacheHitRate*100).toFixed(1)}% of fallbacks)`];if(e&&t.search.topMissedQueries.length>0){a.push(`
3
3
  **Top missed queries** (triggered ER fallback):`);for(let e of t.search.topMissedQueries.slice(0,10)){let t=e.query.length>60?`${e.query.slice(0,57)}...`:e.query;a.push(` - "${t}" (${e.count}x)`)}}if(a.push(``,`### Push`,`- Total pushes: ${t.push.totalPushes} (${t.push.successCount} ok, ${t.push.failCount} failed)`,`- Classification match rate: ${(t.push.classificationMatchRate*100).toFixed(1)}%`,`- Push acceptance rate: ${(t.push.pushAcceptanceRate*100).toFixed(1)}%`),a.push(``,`### Rule Effectiveness`),Object.keys(t.rules.matchCounts).length>0)for(let[e,n]of Object.entries(t.rules.matchCounts)){let r=t.rules.pushCounts[e]??0,i=n>0?(r/n*100).toFixed(0):`0`;a.push(`- **${e}**: ${n} matches → ${r} pushes (${i}% conversion)`)}else a.push(`- _No rule activity recorded yet_`);if(e&&t.rules.lowConversionRules.length>0){a.push(``,`### ⚠️ Low Conversion Rules (potential false positives)`);for(let e of t.rules.lowConversionRules)a.push(`- **${e.ruleId}**: ${e.matchCount} matches, ${e.pushCount} pushes (${(e.conversionRate*100).toFixed(0)}% conversion) — consider tightening patterns`)}return a.push(``,`---`,"_Next: Use `er_update_policy` to refine rules based on these metrics, or `er_push` to share high-value knowledge._"),n&&(r.info(`Evolution metrics reset requested`,{requestedAt:new Date().toISOString(),clearedEvents:t.period.totalEvents}),i.reset(),a.push(`
4
4
  _Metrics have been reset._`)),{content:[{type:`text`,text:a.join(`
5
- `)}]}}catch(e){return r.error(`Evolution review failed`,t(e)),{content:[{type:`text`,text:`Evolution review failed: unable to compute metrics`}],isError:!0}}})}export{i as registerErEvolveReviewTool};
6
- //# sourceMappingURL=evolution.tools.js.map
5
+ `)}]}}catch(e){return r.error(`Evolution review failed`,t(e)),{content:[{type:`text`,text:`Evolution review failed: unable to compute metrics`}],isError:!0}}})}export{i as registerErEvolveReviewTool};
@@ -10,5 +10,4 @@ declare function registerForgeClassifyTool(server: McpServer): void;
10
10
  declare function registerStratumCardTool(server: McpServer, embedder: IEmbedder, cache: FileCache): void;
11
11
  declare function registerForgeGroundTool(server: McpServer, embedder: IEmbedder, store: IKnowledgeStore): void;
12
12
  //#endregion
13
- export { registerDigestTool, registerEvidenceMapTool, registerForgeClassifyTool, registerForgeGroundTool, registerStratumCardTool };
14
- //# sourceMappingURL=forge.tools.d.ts.map
13
+ export { registerDigestTool, registerEvidenceMapTool, registerForgeClassifyTool, registerForgeGroundTool, registerStratumCardTool };
@@ -3,9 +3,8 @@ import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";i
3
3
  `)}]}}case`get`:{if(!n)throw Error(`task_id required for get`);let e=r({action:`get`,taskId:n},p);return e.state?{content:[{type:`text`,text:[`## Evidence Map: ${n} (${e.state.tier})`,``,e.formattedMap??`No entries.`,``,`_${e.state.entries.length} entries — created ${e.state.createdAt}_`].join(`
4
4
  `)}]}:{content:[{type:`text`,text:`Evidence map "${n}" not found.`}]}}case`gate`:{if(!n)throw Error(`task_id required for gate`);let e=r({action:`gate`,taskId:n,retryCount:f},p);if(!e.gate)return{content:[{type:`text`,text:`Evidence map "${n}" not found.`}]};let t=e.gate,i=[`## FORGE Gate: **${t.decision}**`,``,`**Reason:** ${t.reason}`,``,`**Stats:** ${t.stats.verified}V / ${t.stats.assumed}A / ${t.stats.unresolved}U (${t.stats.total} total)`];return t.warnings.length>0&&i.push(``,`**Warnings:**`,...t.warnings.map(e=>`- ⚠️ ${e}`)),t.unresolvedCritical.length>0&&i.push(``,`**Blocking entries:**`,...t.unresolvedCritical.map(e=>`- #${e.id}: ${e.claim} [${e.unknownType??`untyped`}]`)),t.annotation&&i.push(``,`**Annotation:**`,t.annotation),e.formattedMap&&i.push(``,`---`,``,e.formattedMap),i.push(``,`---`,`_Next: ${t.decision===`YIELD`?`Proceed to implementation.`:t.decision===`HOLD`?`Resolve blocking entries, then re-run gate.`:t.decision===`HARD_BLOCK`?`Contract unknowns MUST be resolved. Use search or schema_validate.`:`Proceed with annotation — document unresolved items in delivery.`}_`),{content:[{type:`text`,text:i.join(`
5
5
  `)}]}}case`list`:{let e=r({action:`list`},p);return!e.states||e.states.length===0?{content:[{type:`text`,text:`No evidence maps found.`}]}:{content:[{type:`text`,text:e.states.map(e=>`- **${e.taskId}** (${e.tier}) — ${e.entries.length} entries — ${e.updatedAt}`).join(`
6
- `)}]}}case`delete`:if(!n)throw Error(`task_id required for delete`);return{content:[{type:`text`,text:r({action:`delete`,taskId:n},p).deleted?`Deleted evidence map "${n}".`:`Evidence map "${n}" not found.`}]}}}catch(e){return c.error(`Evidence map error`,t(e)),{content:[{type:`text`,text:`Evidence map error. Check server logs for details.`}],isError:!0}}})}function u(e,r){e.registerTool(`digest`,{description:`Compress multiple text sources (handoffs, debates, phase summaries) into a token-budgeted digest. Jointly ranks across all sources, pins structured fields (status, files, decisions, blockers), and allocates budget by priority weight.`,inputSchema:{sources:s.array(s.object({id:s.string().describe(`Source identifier (e.g., "phase-2-handoff")`),text:s.string().max(5e5).describe(`Source text to compress`),weight:s.number().min(0).default(1).describe(`Priority weight (higher = more budget)`)})).min(1).max(20).describe(`Text sources to compress`),query:s.string().describe(`Focus query — what matters for the next step?`),max_chars:s.number().min(100).max(5e4).default(4e3).describe(`Target budget in characters`),pin_fields:s.array(s.string()).optional().describe(`Key fields to always extract (default: status, files, decisions, blockers, next)`),segmentation:s.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split text for scoring`)}},async({sources:e,query:i,max_chars:a,pin_fields:o,segmentation:s})=>{try{let t=await n(r,{sources:e,query:i,maxChars:a,pinFields:o,segmentation:s}),c=[`## Digest (${t.totalOriginalChars} → ${t.totalCompressedChars} chars, ${(t.ratio*100).toFixed(0)}%)`,``],l=Object.keys(t.fields);if(l.length>0){c.push(`### Extracted Fields`);for(let e of l){let n=t.fields[e];c.push(`**${e}:**`);for(let e of n)c.push(` - [${e.sourceId}] ${e.value}`)}c.push(``)}c.push(`### Compressed Content`,``,t.text),c.push(``,`### Source Stats`);for(let e of t.sourceStats)c.push(`- **${e.id}**: ${e.originalChars} → ${e.keptChars} chars (${e.segmentsKept}/${e.segmentsTotal} segments)`);return c.push(``,`---`,"_Next: Use the digest as compressed context for the next phase. Use `stash` to persist it if needed._"),{content:[{type:`text`,text:c.join(`
7
- `)}]}}catch(e){return c.error(`Digest failed`,t(e)),{content:[{type:`text`,text:`Digest failed. Check server logs for details.`}],isError:!0}}})}function d(e){e.registerTool(`forge_classify`,{description:`Classify FORGE tier (Floor/Standard/Critical) from target files and task description. Checks blast radius, cross-package boundaries, schema/contract patterns, and security signals. Returns tier, triggers, typed unknown seeds, and ceremony guidance.`,inputSchema:{files:s.array(s.string()).min(1).max(100).describe(`Files being modified (paths)`),task:s.string().describe(`Task description`),root_path:s.string().describe(`Root path of the codebase`)}},async({files:e,task:n,root_path:r})=>{try{let t=await i({files:e,task:n,rootPath:r}),a=[`## FORGE Classification: **${t.tier.toUpperCase()}**`,``];if(t.triggers.length>0){a.push(`### Triggers`);for(let e of t.triggers)a.push(`- **${e.rule}** (${e.source}): ${e.detail}`);a.push(``)}if(t.packagesCrossed.length>0&&a.push(`**Packages crossed:** ${t.packagesCrossed.join(`, `)}`),t.typedUnknownSeeds.length>0){a.push(``,`### Typed Unknown Seeds`);for(let e of t.typedUnknownSeeds)a.push(`- [${e.type}] ${e.description} → use \`${e.suggestedTool}\``)}a.push(``,`### Ceremony`);let o=t.ceremony;return a.push(`- **Ground:** ${o.ground}`,`- **Build:** ${o.build}`,`- **Break:** ${o.break}`,`- **Evidence Map:** ${o.evidenceMap}`,`- **Gate:** ${o.gate}`),a.push(``,`---`,`_Next: ${t.tier===`floor`?`Proceed directly to implementation.`:"Run `forge_ground` to execute the full Ground phase."}_`),{content:[{type:`text`,text:a.join(`
8
- `)}]}}catch(e){return c.error(`FORGE classify failed`,t(e)),{content:[{type:`text`,text:`FORGE classify failed. Check server logs for details.`}],isError:!0}}})}function f(e,n,r){e.registerTool(`stratum_card`,{description:`Generate STRATUM context cards from files. T1 = structural metadata (~100 tokens/file). T2 = T1 + compressed content (~300 tokens/file). Replaces reading full files — 10-100x token reduction.`,inputSchema:{files:s.array(s.string()).min(1).max(50).describe(`Absolute file paths to generate cards for`),query:s.string().describe(`Current task query — guides relevance scoring`),tier:s.enum([`T1`,`T2`]).default(`T1`).describe(`Card tier: T1 = structural only, T2 = T1 + compressed content`),max_content_chars:s.number().min(100).max(5e3).default(800).describe(`For T2: max chars for compressed content section`)}},async({files:e,query:i,tier:a,max_content_chars:s})=>{try{let t=await o(n,{files:e,query:i,tier:a,maxContentChars:s,cache:r}),c=[`## STRATUM Cards (${a}) — ${t.cards.length} files`,`Total: ~${t.totalTokenEstimate} tokens (was ~${t.totalOriginalTokenEstimate}, ${(t.compressionRatio*100).toFixed(0)}% of original)`,``];for(let e of t.cards)c.push(e.card,``);return c.push(`---`,"_Next: Use these cards as context instead of reading full files. Use `compact` for deeper content extraction on specific files._"),{content:[{type:`text`,text:c.join(`
9
- `)}]}}catch(e){return c.error(`STRATUM cards failed`,t(e)),{content:[{type:`text`,text:`STRATUM cards failed. Check server logs for details.`}],isError:!0}}})}function p(e,n,r){e.registerTool(`forge_ground`,{description:`Execute the complete FORGE Ground phase in a single call. Chains: tier classification → scope map → typed unknown seeds → constraint loading → file summaries → evidence map creation. Replaces 5-15 manual tool calls.`,inputSchema:{task:s.string().describe(`Task description`),files:s.array(s.string()).min(1).max(100).describe(`Target files being modified (absolute paths)`),root_path:s.string().describe(`Root path of the codebase`),max_constraints:s.number().min(0).max(10).default(3).describe(`Max constraint entries to load from KB`),force_tier:s.enum([`floor`,`standard`,`critical`]).optional().describe(`Force a specific tier (skips auto-classification)`),task_id:s.string().optional().describe(`Custom task ID for evidence map (auto-generated if omitted)`)}},async({task:e,files:i,root_path:o,max_constraints:s,force_tier:l,task_id:u})=>{try{let t=await a(n,r,{task:e,files:i,rootPath:o,maxConstraints:s,forceTier:l,taskId:u}),c=[`## FORGE Ground: **${t.tier.toUpperCase()}**`,`Estimated output: ~${t.estimatedTokens} tokens`,``];if(t.classifyTriggers.length>0){c.push(`### Classification Triggers`);for(let e of t.classifyTriggers)c.push(`- **${e.rule}** (${e.source}): ${e.detail}`);c.push(``)}if(t.fileSummaries.length>0){c.push(`### Target Files`);for(let e of t.fileSummaries)c.push(`- **${e.path}** (${e.lines} lines) — exports: ${e.exports.join(`, `)||`none`} — functions: ${e.functions.join(`, `)||`none`}`);c.push(``)}if(t.typedUnknownSeeds.length>0){c.push(`### Typed Unknown Seeds`);for(let e of t.typedUnknownSeeds)c.push(`- [${e.type}] ${e.description} → \`${e.suggestedTool}\``);c.push(``)}if(t.constraints.length>0){c.push(`### Constraint Seed`);for(let e of t.constraints)c.push(`- **${e.source}** (${(e.relevance*100).toFixed(0)}%): ${e.snippet}`);c.push(``)}if(t.scopeMap){c.push(`### Scope Map: ${t.scopeMap.files.length} files (~${t.scopeMap.totalEstimatedTokens} tokens)`);for(let e of t.scopeMap.files.slice(0,5))c.push(`- ${e.path} (${(e.relevance*100).toFixed(0)}%) — ${e.reason}`);t.scopeMap.files.length>5&&c.push(`- _...and ${t.scopeMap.files.length-5} more_`),c.push(``)}t.evidenceMapTaskId&&(c.push(`**Evidence Map:** \`${t.evidenceMapTaskId}\` (initialized)`),c.push(``)),c.push(`### Ceremony Guidance`);let d=t.ceremony;return c.push(`- **Ground:** ${d.ground}`,`- **Build:** ${d.build}`,`- **Break:** ${d.break}`,`- **Evidence Map:** ${d.evidenceMap}`,`- **Gate:** ${d.gate}`),c.push(``,`---`,`_Next: ${t.tier===`floor`?`Proceed to Build phase.`:"Use `evidence_map` to track claims during Build, then `evidence_map` gate after Break."}_`),{content:[{type:`text`,text:c.join(`
10
- `)}]}}catch(e){return c.error(`FORGE Ground failed`,t(e)),{content:[{type:`text`,text:`FORGE Ground failed. Check server logs for details.`}],isError:!0}}})}export{u as registerDigestTool,l as registerEvidenceMapTool,d as registerForgeClassifyTool,p as registerForgeGroundTool,f as registerStratumCardTool};
11
- //# sourceMappingURL=forge.tools.js.map
6
+ `)}]}}case`delete`:if(!n)throw Error(`task_id required for delete`);return{content:[{type:`text`,text:r({action:`delete`,taskId:n},p).deleted?`Deleted evidence map "${n}".`:`Evidence map "${n}" not found.`}]}}}catch(e){return c.error(`Evidence map error`,t(e)),{content:[{type:`text`,text:`Evidence map error: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function u(e,r){e.registerTool(`digest`,{description:`Compress multiple text sources (handoffs, debates, phase summaries) into a token-budgeted digest. Jointly ranks across all sources, pins structured fields (status, files, decisions, blockers), and allocates budget by priority weight.`,inputSchema:{sources:s.array(s.object({id:s.string().describe(`Source identifier (e.g., "phase-2-handoff")`),text:s.string().max(5e5).describe(`Source text to compress`),weight:s.number().min(0).default(1).describe(`Priority weight (higher = more budget)`)})).min(1).max(20).describe(`Text sources to compress`),query:s.string().describe(`Focus query — what matters for the next step?`),max_chars:s.number().min(100).max(5e4).default(4e3).describe(`Target budget in characters`),pin_fields:s.array(s.string()).optional().describe(`Key fields to always extract (default: status, files, decisions, blockers, next)`),segmentation:s.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split text for scoring`)}},async({sources:e,query:i,max_chars:a,pin_fields:o,segmentation:s})=>{try{let t=await n(r,{sources:e,query:i,maxChars:a,pinFields:o,segmentation:s}),c=[`## Digest (${t.totalOriginalChars} → ${t.totalCompressedChars} chars, ${(t.ratio*100).toFixed(0)}%)`,``],l=Object.keys(t.fields);if(l.length>0){c.push(`### Extracted Fields`);for(let e of l){let n=t.fields[e];c.push(`**${e}:**`);for(let e of n)c.push(` - [${e.sourceId}] ${e.value}`)}c.push(``)}c.push(`### Compressed Content`,``,t.text),c.push(``,`### Source Stats`);for(let e of t.sourceStats)c.push(`- **${e.id}**: ${e.originalChars} → ${e.keptChars} chars (${e.segmentsKept}/${e.segmentsTotal} segments)`);return c.push(``,`---`,"_Next: Use the digest as compressed context for the next phase. Use `stash` to persist it if needed._"),{content:[{type:`text`,text:c.join(`
7
+ `)}]}}catch(e){return c.error(`Digest failed`,t(e)),{content:[{type:`text`,text:`Digest failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function d(e){e.registerTool(`forge_classify`,{description:`Classify FORGE tier (Floor/Standard/Critical) from target files and task description. Checks blast radius, cross-package boundaries, schema/contract patterns, and security signals. Returns tier, triggers, typed unknown seeds, and ceremony guidance.`,inputSchema:{files:s.array(s.string()).min(1).max(100).describe(`Files being modified (paths)`),task:s.string().describe(`Task description`),root_path:s.string().describe(`Root path of the codebase`)}},async({files:e,task:n,root_path:r})=>{try{let t=await i({files:e,task:n,rootPath:r}),a=[`## FORGE Classification: **${t.tier.toUpperCase()}**`,``];if(t.triggers.length>0){a.push(`### Triggers`);for(let e of t.triggers)a.push(`- **${e.rule}** (${e.source}): ${e.detail}`);a.push(``)}if(t.packagesCrossed.length>0&&a.push(`**Packages crossed:** ${t.packagesCrossed.join(`, `)}`),t.typedUnknownSeeds.length>0){a.push(``,`### Typed Unknown Seeds`);for(let e of t.typedUnknownSeeds)a.push(`- [${e.type}] ${e.description} → use \`${e.suggestedTool}\``)}a.push(``,`### Ceremony`);let o=t.ceremony;return a.push(`- **Ground:** ${o.ground}`,`- **Build:** ${o.build}`,`- **Break:** ${o.break}`,`- **Evidence Map:** ${o.evidenceMap}`,`- **Gate:** ${o.gate}`),a.push(``,`---`,`_Next: ${t.tier===`floor`?`Proceed directly to implementation.`:"Run `forge_ground` to execute the full Ground phase."}_`),{content:[{type:`text`,text:a.join(`
8
+ `)}]}}catch(e){return c.error(`FORGE classify failed`,t(e)),{content:[{type:`text`,text:`FORGE classify failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function f(e,n,r){e.registerTool(`stratum_card`,{description:`Generate STRATUM context cards from files. T1 = structural metadata (~100 tokens/file). T2 = T1 + compressed content (~300 tokens/file). Replaces reading full files — 10-100x token reduction.`,inputSchema:{files:s.array(s.string()).min(1).max(50).describe(`Absolute file paths to generate cards for`),query:s.string().describe(`Current task query — guides relevance scoring`),tier:s.enum([`T1`,`T2`]).default(`T1`).describe(`Card tier: T1 = structural only, T2 = T1 + compressed content`),max_content_chars:s.number().min(100).max(5e3).default(800).describe(`For T2: max chars for compressed content section`)}},async({files:e,query:i,tier:a,max_content_chars:s})=>{try{let t=await o(n,{files:e,query:i,tier:a,maxContentChars:s,cache:r}),c=[`## STRATUM Cards (${a}) — ${t.cards.length} files`,`Total: ~${t.totalTokenEstimate} tokens (was ~${t.totalOriginalTokenEstimate}, ${(t.compressionRatio*100).toFixed(0)}% of original)`,``];for(let e of t.cards)c.push(e.card,``);return c.push(`---`,"_Next: Use these cards as context instead of reading full files. Use `compact` for deeper content extraction on specific files._"),{content:[{type:`text`,text:c.join(`
9
+ `)}]}}catch(e){return c.error(`STRATUM cards failed`,t(e)),{content:[{type:`text`,text:`STRATUM cards failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function p(e,n,r){e.registerTool(`forge_ground`,{description:`Execute the complete FORGE Ground phase in a single call. Chains: tier classification → scope map → typed unknown seeds → constraint loading → file summaries → evidence map creation. Replaces 5-15 manual tool calls.`,inputSchema:{task:s.string().describe(`Task description`),files:s.array(s.string()).min(1).max(100).describe(`Target files being modified (absolute paths)`),root_path:s.string().describe(`Root path of the codebase`),max_constraints:s.number().min(0).max(10).default(3).describe(`Max constraint entries to load from KB`),force_tier:s.enum([`floor`,`standard`,`critical`]).optional().describe(`Force a specific tier (skips auto-classification)`),task_id:s.string().optional().describe(`Custom task ID for evidence map (auto-generated if omitted)`)}},async({task:e,files:i,root_path:o,max_constraints:s,force_tier:l,task_id:u})=>{try{let t=await a(n,r,{task:e,files:i,rootPath:o,maxConstraints:s,forceTier:l,taskId:u}),c=[`## FORGE Ground: **${t.tier.toUpperCase()}**`,`Estimated output: ~${t.estimatedTokens} tokens`,``];if(t.classifyTriggers.length>0){c.push(`### Classification Triggers`);for(let e of t.classifyTriggers)c.push(`- **${e.rule}** (${e.source}): ${e.detail}`);c.push(``)}if(t.fileSummaries.length>0){c.push(`### Target Files`);for(let e of t.fileSummaries)c.push(`- **${e.path}** (${e.lines} lines) — exports: ${e.exports.join(`, `)||`none`} — functions: ${e.functions.join(`, `)||`none`}`);c.push(``)}if(t.typedUnknownSeeds.length>0){c.push(`### Typed Unknown Seeds`);for(let e of t.typedUnknownSeeds)c.push(`- [${e.type}] ${e.description} → \`${e.suggestedTool}\``);c.push(``)}if(t.constraints.length>0){c.push(`### Constraint Seed`);for(let e of t.constraints)c.push(`- **${e.source}** (${(e.relevance*100).toFixed(0)}%): ${e.snippet}`);c.push(``)}if(t.scopeMap){c.push(`### Scope Map: ${t.scopeMap.files.length} files (~${t.scopeMap.totalEstimatedTokens} tokens)`);for(let e of t.scopeMap.files.slice(0,5))c.push(`- ${e.path} (${(e.relevance*100).toFixed(0)}%) — ${e.reason}`);t.scopeMap.files.length>5&&c.push(`- _...and ${t.scopeMap.files.length-5} more_`),c.push(``)}t.evidenceMapTaskId&&(c.push(`**Evidence Map:** \`${t.evidenceMapTaskId}\` (initialized)`),c.push(``)),c.push(`### Ceremony Guidance`);let d=t.ceremony;return c.push(`- **Ground:** ${d.ground}`,`- **Build:** ${d.build}`,`- **Break:** ${d.break}`,`- **Evidence Map:** ${d.evidenceMap}`,`- **Gate:** ${d.gate}`),c.push(``,`---`,`_Next: ${t.tier===`floor`?`Proceed to Build phase.`:"Use `evidence_map` to track claims during Build, then `evidence_map` gate after Break."}_`),{content:[{type:`text`,text:c.join(`
10
+ `)}]}}catch(e){return c.error(`FORGE Ground failed`,t(e)),{content:[{type:`text`,text:`FORGE Ground failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{u as registerDigestTool,l as registerEvidenceMapTool,d as registerForgeClassifyTool,p as registerForgeGroundTool,f as registerStratumCardTool};
@@ -4,5 +4,4 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
4
4
  //#region packages/server/src/tools/forget.tool.d.ts
5
5
  declare function registerForgetTool(server: McpServer, curated: CuratedKnowledgeManager): void;
6
6
  //#endregion
7
- export { registerForgetTool };
8
- //# sourceMappingURL=forget.tool.d.ts.map
7
+ export { registerForgetTool };
@@ -1,2 +1 @@
1
- import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`forget`,{description:`Remove a curated knowledge entry. Deletes the file and removes it from the vector store.`,inputSchema:{path:n.string().describe(`Relative path within curated/ (e.g., "decisions/deprecated-approach.md")`),reason:n.string().min(3).describe(`Why this entry is being removed`)}},async({path:e,reason:n})=>{try{let t=await i.forget(e,n);return r.info(`Forgot entry`,{path:t.path,reason:n}),{content:[{type:`text`,text:`Forgotten: \`curated/${t.path}\`\n\nRemoved from disk and vector store.\nReason: ${n}\n\n---\n_Next: Use \`list\` to verify removal, or \`search\` to confirm the entry is no longer returned._`}]}}catch(e){return r.error(`Forget failed`,t(e)),{content:[{type:`text`,text:`Forget failed. Check server logs for details.`}],isError:!0}}})}export{i as registerForgetTool};
2
- //# sourceMappingURL=forget.tool.js.map
1
+ import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`forget`,{description:`Remove a curated knowledge entry. Deletes the file and removes it from the vector store.`,inputSchema:{path:n.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/deprecated-approach.md")`),reason:n.string().min(3).describe(`Why this entry is being removed`)}},async({path:e,reason:n})=>{try{let t=await i.forget(e,n);return r.info(`Forgot entry`,{path:t.path,reason:n}),{content:[{type:`text`,text:`Forgotten: \`.ai/curated/${t.path}\`\n\nRemoved from disk and vector store.\nReason: ${n}\n\n---\n_Next: Use \`list\` to verify removal, or \`search\` to confirm the entry is no longer returned._`}]}}catch(e){return r.error(`Forget failed`,t(e)),{content:[{type:`text`,text:`Forget failed. Check server logs for details.`}],isError:!0}}})}export{i as registerForgetTool};
@@ -4,5 +4,4 @@ import { IGraphStore } from "@kb/store";
4
4
  //#region packages/server/src/tools/graph.tool.d.ts
5
5
  declare function registerGraphTool(server: McpServer, graphStore: IGraphStore): void;
6
6
  //#endregion
7
- export { registerGraphTool };
8
- //# sourceMappingURL=graph.tool.d.ts.map
7
+ export { registerGraphTool };
@@ -1,5 +1,4 @@
1
1
  import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{graphQuery as n}from"../../../tools/dist/index.js";import{z as r}from"zod";const i=e(`tools`),a=r.object({id:r.string().optional().describe(`Node ID (auto-generated if omitted)`),type:r.string().describe(`Node type (entity, service, api, concept, decision)`),name:r.string().describe(`Display name`),properties:r.record(r.string(),r.unknown()).optional().describe(`Arbitrary properties`),sourceRecordId:r.string().optional().describe(`Back-link to knowledge record ID`),sourcePath:r.string().optional().describe(`Source file path`)}),o=r.object({id:r.string().optional().describe(`Edge ID (auto-generated if omitted)`),fromId:r.string().describe(`Source node ID`),toId:r.string().describe(`Target node ID`),type:r.string().describe(`Relationship type (depends-on, implements, calls, affects)`),weight:r.number().min(0).max(1).optional().describe(`Relationship weight`),properties:r.record(r.string(),r.unknown()).optional().describe(`Arbitrary properties`)});function s(e,s){e.registerTool(`graph`,{description:`Query and manage the knowledge graph. Find nodes/edges, traverse connections, add entities and relationships. The graph captures structural relationships between concepts discovered in the codebase.`,inputSchema:{action:r.enum([`find_nodes`,`find_edges`,`neighbors`,`traverse`,`stats`,`add`,`delete`,`clear`]).describe(`Action: find_nodes (search nodes), find_edges (search edges), neighbors (direct connections), traverse (multi-hop), stats (graph overview), add (insert nodes/edges), delete (remove nodes), clear (remove all)`),node_type:r.string().optional().describe(`Node type filter (for find_nodes)`),name_pattern:r.string().optional().describe(`Name substring match (for find_nodes)`),source_path:r.string().optional().describe(`Source path filter`),node_id:r.string().optional().describe(`Node ID (for neighbors, traverse, delete)`),edge_type:r.string().optional().describe(`Edge type filter`),from_id:r.string().optional().describe(`Source node ID (for find_edges)`),to_id:r.string().optional().describe(`Target node ID (for find_edges)`),direction:r.enum([`outgoing`,`incoming`,`both`]).default(`both`).describe(`Traversal direction`),max_depth:r.number().min(1).max(5).default(2).describe(`Max traversal depth`),limit:r.number().min(1).max(100).default(50).describe(`Max results`),nodes:r.array(a).max(500).optional().describe(`Nodes to add (for action=add)`),edges:r.array(o).max(500).optional().describe(`Edges to add (for action=add)`)}},async e=>{try{let t=await n(s,{action:e.action,nodeType:e.node_type,namePattern:e.name_pattern,sourcePath:e.source_path,nodeId:e.node_id,edgeType:e.edge_type,fromId:e.from_id,toId:e.to_id,direction:e.direction,maxDepth:e.max_depth,limit:e.limit,nodes:e.nodes,edges:e.edges}),r=[t.summary];if(t.nodes&&t.nodes.length>0){r.push(`
2
2
  ### Nodes`);for(let e of t.nodes){let t=Object.keys(e.properties).length>0?` — ${JSON.stringify(e.properties)}`:``;r.push(`- **${e.name}** (${e.type}, id: \`${e.id}\`)${t}`)}}if(t.edges&&t.edges.length>0){r.push(`
3
3
  ### Edges`);for(let e of t.edges)r.push(`- \`${e.fromId}\` —[${e.type}]→ \`${e.toId}\`${e.weight===1?``:` (weight: ${e.weight})`}`)}return t.stats&&(r.push(`\nNode types: ${JSON.stringify(t.stats.nodeTypes)}`),r.push(`Edge types: ${JSON.stringify(t.stats.edgeTypes)}`)),r.push("\n---\n_Next: Use `graph(traverse)` to explore connections, `graph(add)` to insert entities, or `search` with graph-augmented results._"),{content:[{type:`text`,text:r.join(`
4
- `)}]}}catch(e){return i.error(`Graph query failed`,t(e)),{content:[{type:`text`,text:`Graph query failed. Check server logs for details.`}],isError:!0}}})}export{s as registerGraphTool};
5
- //# sourceMappingURL=graph.tool.js.map
4
+ `)}]}}catch(e){return i.error(`Graph query failed`,t(e)),{content:[{type:`text`,text:`Graph query failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{s as registerGraphTool};
@@ -4,5 +4,4 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
4
4
  //#region packages/server/src/tools/list.tool.d.ts
5
5
  declare function registerListTool(server: McpServer, curated: CuratedKnowledgeManager): void;
6
6
  //#endregion
7
- export { registerListTool };
8
- //# sourceMappingURL=list.tool.d.ts.map
7
+ export { registerListTool };
@@ -1,3 +1,2 @@
1
1
  import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`list`,{description:`List curated knowledge entries. Optionally filter by category or tag.`,inputSchema:{category:n.string().optional().describe(`Filter by category (e.g., "decisions", "patterns")`),tag:n.string().optional().describe(`Filter by tag`)}},async({category:e,tag:n})=>{try{let t=await i.list({category:e,tag:n});if(t.length===0)return{content:[{type:`text`,text:`No curated knowledge entries found.`+(e?` (category: ${e})`:``)+(n?` (tag: ${n})`:``)}]};let r=t.map(e=>{let t=e.tags.length>0?` [${e.tags.join(`, `)}]`:``;return`- **${e.title}** (v${e.version})${t}\n \`${e.path}\` — ${e.contentPreview.slice(0,80)}…`}).join(`
2
- `);return{content:[{type:`text`,text:`## Curated Knowledge (${t.length} entries)\n\n${r}\n\n---\n_Next: Use \`read\` to view full content of any entry, or \`remember\` to store new knowledge._`}]}}catch(e){return r.error(`List failed`,t(e)),{content:[{type:`text`,text:`List failed. Check server logs for details.`}],isError:!0}}})}export{i as registerListTool};
3
- //# sourceMappingURL=list.tool.js.map
2
+ `);return{content:[{type:`text`,text:`## Curated Knowledge (${t.length} entries)\n\n${r}\n\n---\n_Next: Use \`read\` to view full content of any entry, or \`remember\` to store new knowledge._`}]}}catch(e){return r.error(`List failed`,t(e)),{content:[{type:`text`,text:`List failed. Check server logs for details.`}],isError:!0}}})}export{i as registerListTool};
@@ -4,5 +4,4 @@ import { IKnowledgeStore } from "@kb/store";
4
4
  //#region packages/server/src/tools/lookup.tool.d.ts
5
5
  declare function registerLookupTool(server: McpServer, store: IKnowledgeStore): void;
6
6
  //#endregion
7
- export { registerLookupTool };
8
- //# sourceMappingURL=lookup.tool.d.ts.map
7
+ export { registerLookupTool };
@@ -1,4 +1,3 @@
1
1
  import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`lookup`,{description:`Look up all indexed chunks for a specific file path. Returns all chunks sorted by position.`,inputSchema:{path:n.string().describe(`Relative file path to look up (e.g., "src/index.ts")`)}},async({path:e})=>{try{let t=await i.getBySourcePath(e);return t.length===0?{content:[{type:`text`,text:`No indexed content found for: ${e}`}]}:(t.sort((e,t)=>e.chunkIndex-t.chunkIndex),{content:[{type:`text`,text:`${`## ${e}\n**Chunks**: ${t.length} | **Type**: ${t[0].contentType}\n`}\n${t.map(e=>{let t=e.startLine?` (lines ${e.startLine}-${e.endLine})`:``;return`### Chunk ${e.chunkIndex+1}/${e.totalChunks}${t}\n${e.content}`}).join(`
2
2
 
3
- `)}\n\n---\n_Next: Use \`search\` to find related content, or \`analyze_dependencies\` to see what this file imports._`}]})}catch(e){return r.error(`Lookup failed`,t(e)),{content:[{type:`text`,text:`Lookup failed. Check server logs for details.`}],isError:!0}}})}export{i as registerLookupTool};
4
- //# sourceMappingURL=lookup.tool.js.map
3
+ `)}\n\n---\n_Next: Use \`search\` to find related content, or \`analyze_dependencies\` to see what this file imports._`}]})}catch(e){return r.error(`Lookup failed`,t(e)),{content:[{type:`text`,text:`Lookup failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{i as registerLookupTool};
@@ -5,5 +5,4 @@ import { IKnowledgeStore } from "@kb/store";
5
5
  //#region packages/server/src/tools/onboard.tool.d.ts
6
6
  declare function registerOnboardTool(server: McpServer, store: IKnowledgeStore, embedder: IEmbedder): void;
7
7
  //#endregion
8
- export { registerOnboardTool };
9
- //# sourceMappingURL=onboard.tool.d.ts.map
8
+ export { registerOnboardTool };
@@ -1,3 +1,2 @@
1
1
  import{createHash as e}from"node:crypto";import{createLogger as t,serializeError as n}from"../../../core/dist/index.js";import{onboard as r}from"../../../tools/dist/index.js";import{z as i}from"zod";const a=t(`tools`);async function o(t,r,i){for(let o of i.steps)if(!(o.status!==`success`||!o.output))try{let n=e(`sha256`).update(i.path).digest(`hex`).slice(0,12),a=`produced/onboard/${o.name}/${n}.md`,s=e(`sha256`).update(o.output).digest(`hex`).slice(0,16),c=new Date().toISOString(),l=o.output.length>2e3?o.output.split(/(?=^## )/m).filter(e=>e.trim().length>0):[o.output],u=l.map((t,n)=>({id:e(`sha256`).update(`${a}::${n}`).digest(`hex`).slice(0,16),content:t.trim(),sourcePath:a,contentType:`produced-knowledge`,chunkIndex:n,totalChunks:l.length,startLine:0,endLine:0,fileHash:s,indexedAt:c,origin:`produced`,tags:[`onboard`,o.name],category:`analysis`,version:1})),d=await r.embedBatch(u.map(e=>e.content));await t.upsert(u,d)}catch(e){a.warn(`Auto-persist onboard step failed`,{stepName:o.name,...n(e)})}}function s(e,t,s){e.registerTool(`onboard`,{description:`First-time codebase onboarding: runs all analysis tools (structure, dependencies, entry-points, symbols, patterns, diagram) in one command. Results are auto-persisted to KB. Use mode=generate to also write structured output to .ai/kb/ directory.`,inputSchema:{path:i.string().describe(`Root path of the codebase to onboard`),mode:i.enum([`memory`,`generate`]).default(`memory`).describe(`Output mode: memory = KB vector store only; generate = also write to .ai/kb/ directory`),out_dir:i.string().optional().describe(`Custom output directory for generate mode (default: <path>/.ai/kb)`)}},async({path:e,mode:i,out_dir:c})=>{try{a.info(`Starting onboard`,{path:e,mode:i});let n=await r({path:e,mode:i,outDir:c});o(t,s,n);let l=[`## Onboard Complete`,``,`**Path:** \`${n.path}\``,`**Mode:** ${n.mode}`,`**Duration:** ${n.totalDurationMs}ms`,``];n.outDir&&(l.push(`**Output directory:** \`${n.outDir}\``),l.push(``)),l.push(`### Analysis Results`,``);let u=[],d=[];for(let e of n.steps)e.status===`success`?u.push(`- ✓ **${e.name}** (${e.durationMs}ms) — ${e.output.length} chars`):d.push(`- ✗ **${e.name}** — ${e.error}`);l.push(...u),d.length>0&&l.push(``,`### Failed`,``,...d),l.push(``,`---`,``);for(let e of n.steps)e.status===`success`&&l.push(`### ${e.name}`,``,e.output,``,`---`,``);return l.push(`_All results auto-saved to KB.`,n.mode===`generate`?` Files written to \`${n.outDir}\`.`:``," Next: Use `search` to query the knowledge, or `remember` to add custom insights._"),{content:[{type:`text`,text:l.join(`
2
- `)}]}}catch(e){return a.error(`Onboard failed`,n(e)),{content:[{type:`text`,text:`Onboard failed. Check server logs for details.`}],isError:!0}}})}export{s as registerOnboardTool};
3
- //# sourceMappingURL=onboard.tool.js.map
2
+ `)}]}}catch(e){return a.error(`Onboard failed`,n(e)),{content:[{type:`text`,text:`Onboard failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}export{s as registerOnboardTool};
@@ -4,5 +4,4 @@ import { PolicyStore } from "@kb/enterprise-bridge";
4
4
  //#region packages/server/src/tools/policy.tools.d.ts
5
5
  declare function registerErUpdatePolicyTool(server: McpServer, policyStore: PolicyStore): void;
6
6
  //#endregion
7
- export { registerErUpdatePolicyTool };
8
- //# sourceMappingURL=policy.tools.d.ts.map
7
+ export { registerErUpdatePolicyTool };
@@ -1,3 +1,2 @@
1
1
  import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`er_update_policy`,{description:`Manage ER push classification rules. Supports listing, updating, creating, and deleting rules that determine when knowledge should be pushed to Enterprise RAG.`,inputSchema:{action:n.enum([`list`,`get`,`update`,`create`,`delete`]).describe(`Action to perform on classification rules`),rule_id:n.string().optional().describe(`Rule ID (required for get, update, delete)`),changes:n.object({patterns:n.array(n.string()).optional(),category:n.string().optional(),pushWeight:n.number().min(0).max(1).optional(),description:n.string().optional(),examples:n.array(n.string()).optional(),enabled:n.boolean().optional()}).optional().describe(`Changes to apply (for update action)`),new_rule:n.object({id:n.string().regex(/^[a-z][a-z0-9-]*$/),patterns:n.array(n.string()).min(1),category:n.string(),pushWeight:n.number().min(0).max(1),description:n.string(),examples:n.array(n.string()).default([]),enabled:n.boolean().default(!0)}).optional().describe(`New rule definition (for create action)`)}},async({action:e,rule_id:n,changes:a,new_rule:o})=>{try{if(e===`list`){let e=i.getRules();return{content:[{type:`text`,text:`## Classification Rules\n\n${e.map(e=>`- **${e.id}** (${e.enabled?`enabled`:`disabled`}) — ${e.description}\n Category: ${e.category} | Weight: ${e.pushWeight} | Patterns: ${e.patterns.join(`, `)}`).join(`
2
- `)}\n\n---\n_${e.length} rules total. Use \`action: "update"\` to modify a rule._`}]}}if(e===`get`){if(!n)return{content:[{type:`text`,text:'`rule_id` is required for "get" action.'}],isError:!0};let e=i.getRule(n);return e?{content:[{type:`text`,text:`## Rule: ${e.id}\n\n${JSON.stringify(e,null,2)}`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}if(e===`update`){if(!n||!a)return{content:[{type:`text`,text:'`rule_id` and `changes` are required for "update" action.'}],isError:!0};let e=i.updateRule(n,a);return e?{content:[{type:`text`,text:`Updated rule **${n}**.\n\n${JSON.stringify(e,null,2)}\n\n---\n_Next: Use \`action: "list"\` to verify all rules._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}if(e===`create`){if(!o)return{content:[{type:`text`,text:'`new_rule` is required for "create" action.'}],isError:!0};let e=i.addRule(o);return{content:[{type:`text`,text:`Created rule **${e.id}**.\n\n${JSON.stringify(e,null,2)}\n\n---\n_Next: Test classification with \`remember\` and check signals._`}]}}return e===`delete`?n?i.deleteRule(n)?{content:[{type:`text`,text:`Deleted rule **${n}**.\n\n---\n_Next: Use \`action: "list"\` to verify._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}:{content:[{type:`text`,text:'`rule_id` is required for "delete" action.'}],isError:!0}:{content:[{type:`text`,text:`Unknown action: ${e}`}],isError:!0}}catch(e){return r.error(`Policy update failed`,t(e)),{content:[{type:`text`,text:`Policy update failed. Check server logs for details.`}],isError:!0}}})}export{i as registerErUpdatePolicyTool};
3
- //# sourceMappingURL=policy.tools.js.map
2
+ `)}\n\n---\n_${e.length} rules total. Use \`action: "update"\` to modify a rule._`}]}}if(e===`get`){if(!n)return{content:[{type:`text`,text:'`rule_id` is required for "get" action.'}],isError:!0};let e=i.getRule(n);return e?{content:[{type:`text`,text:`## Rule: ${e.id}\n\n${JSON.stringify(e,null,2)}`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}if(e===`update`){if(!n||!a)return{content:[{type:`text`,text:'`rule_id` and `changes` are required for "update" action.'}],isError:!0};let e=i.updateRule(n,a);return e?{content:[{type:`text`,text:`Updated rule **${n}**.\n\n${JSON.stringify(e,null,2)}\n\n---\n_Next: Use \`action: "list"\` to verify all rules._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}if(e===`create`){if(!o)return{content:[{type:`text`,text:'`new_rule` is required for "create" action.'}],isError:!0};let e=i.addRule(o);return{content:[{type:`text`,text:`Created rule **${e.id}**.\n\n${JSON.stringify(e,null,2)}\n\n---\n_Next: Test classification with \`remember\` and check signals._`}]}}return e===`delete`?n?i.deleteRule(n)?{content:[{type:`text`,text:`Deleted rule **${n}**.\n\n---\n_Next: Use \`action: "list"\` to verify._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}:{content:[{type:`text`,text:'`rule_id` is required for "delete" action.'}],isError:!0}:{content:[{type:`text`,text:`Unknown action: ${e}`}],isError:!0}}catch(e){return r.error(`Policy update failed`,t(e)),{content:[{type:`text`,text:`Policy update failed. Check server logs for details.`}],isError:!0}}})}export{i as registerErUpdatePolicyTool};
@@ -3,5 +3,4 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
3
3
  //#region packages/server/src/tools/produce.tool.d.ts
4
4
  declare function registerProduceKnowledgeTool(server: McpServer): void;
5
5
  //#endregion
6
- export { registerProduceKnowledgeTool };
7
- //# sourceMappingURL=produce.tool.d.ts.map
6
+ export { registerProduceKnowledgeTool };
@@ -1,5 +1,4 @@
1
1
  import{DependencyAnalyzer as e,DiagramGenerator as t,EntryPointAnalyzer as n,KnowledgeProducer as r,PatternAnalyzer as i,StructureAnalyzer as a,SymbolAnalyzer as o}from"../../../analyzers/dist/index.js";import{createLogger as s,serializeError as c}from"../../../core/dist/index.js";import{z as l}from"zod";const u=s(`tools`);function d(s){let d=new r({structure:new a,dependencies:new e,symbols:new o,patterns:new i,entryPoints:new n,diagrams:new t});s.registerTool(`produce_knowledge`,{description:`Run automated codebase analysis and produce synthesis instructions. Executes Tier 1 deterministic analyzers, then returns structured baselines and instructions for you to synthesize knowledge using remember.`,inputSchema:{scope:l.string().optional().describe(`Root path to analyze (defaults to workspace root)`),aspects:l.array(l.enum([`all`,`structure`,`dependencies`,`symbols`,`patterns`,`entry-points`,`diagrams`])).default([`all`]).describe(`Which analysis aspects to run`)}},async({scope:e,aspects:t})=>{try{let n=e??`.`;u.info(`Running knowledge production`,{rootPath:n,aspects:t});let r=await d.runExtraction(n,t);return{content:[{type:`text`,text:d.buildSynthesisInstructions(r,t)+`
2
2
 
3
3
  ---
4
- _Next: Review the baselines above and use \`remember\` to store synthesized knowledge entries._`}]}}catch(e){return u.error(`Knowledge production failed`,c(e)),{content:[{type:`text`,text:`Knowledge production failed. Check server logs for details.`}],isError:!0}}})}export{d as registerProduceKnowledgeTool};
5
- //# sourceMappingURL=produce.tool.js.map
4
+ _Next: Review the baselines above and use \`remember\` to store synthesized knowledge entries._`}]}}catch(e){return u.error(`Knowledge production failed`,c(e)),{content:[{type:`text`,text:`Knowledge production failed. Check server logs for details.`}],isError:!0}}})}export{d as registerProduceKnowledgeTool};
@@ -4,5 +4,4 @@ import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
4
4
  //#region packages/server/src/tools/read.tool.d.ts
5
5
  declare function registerReadTool(server: McpServer, curated: CuratedKnowledgeManager): void;
6
6
  //#endregion
7
- export { registerReadTool };
8
- //# sourceMappingURL=read.tool.d.ts.map
7
+ export { registerReadTool };
@@ -1,3 +1,2 @@
1
- import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`read`,{description:`Read the full content of a specific curated knowledge entry by its path. Use list first to discover available entries.`,inputSchema:{path:n.string().describe(`Relative path within curated/ (e.g., "decisions/use-lancedb.md")`)}},async({path:e})=>{try{let t=await i.read(e);return{content:[{type:`text`,text:`${[`## ${t.title}`,`- **Path**: curated/${e}`,`- **Category**: ${t.category}`,t.tags.length?`- **Tags**: ${t.tags.join(`, `)}`:null,`- **Version**: ${t.version}`,`- **Created**: ${t.created}`,t.updated===t.created?null:`- **Updated**: ${t.updated}`,``].filter(e=>e!==null).join(`
2
- `)}\n${t.content}\n\n---\n_Next: Use \`update\` to modify this entry, or \`search\` to find related entries._`}]}}catch(e){return r.error(`Read failed`,t(e)),{content:[{type:`text`,text:`Read failed. Check server logs for details.`}],isError:!0}}})}export{i as registerReadTool};
3
- //# sourceMappingURL=read.tool.js.map
1
+ import{createLogger as e,serializeError as t}from"../../../core/dist/index.js";import{z as n}from"zod";const r=e(`tools`);function i(e,i){e.registerTool(`read`,{description:`Read the full content of a specific curated knowledge entry by its path. Use list first to discover available entries.`,inputSchema:{path:n.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/use-lancedb.md")`)}},async({path:e})=>{try{let t=await i.read(e);return{content:[{type:`text`,text:`${[`## ${t.title}`,`- **Path**: .ai/curated/${e}`,`- **Category**: ${t.category}`,t.tags.length?`- **Tags**: ${t.tags.join(`, `)}`:null,`- **Version**: ${t.version}`,`- **Created**: ${t.created}`,t.updated===t.created?null:`- **Updated**: ${t.updated}`,``].filter(e=>e!==null).join(`
2
+ `)}\n${t.content}\n\n---\n_Next: Use \`update\` to modify this entry, or \`search\` to find related entries._`}]}}catch(e){return r.error(`Read failed`,t(e)),{content:[{type:`text`,text:`Read failed. Check server logs for details.`}],isError:!0}}})}export{i as registerReadTool};