@vpxa/aikit 0.1.61 → 0.1.63

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,7 +1,7 @@
1
1
  import{n as e,t}from"./curated-manager-DX-_oJg0.js";import{autoUpgradeScaffold as n,getCurrentVersion as r,getUpgradeState as i}from"./version-check-D4j0Pykd.js";import{existsSync as a,mkdirSync as o,readFileSync as s,readdirSync as c,renameSync as l,rmSync as u,statSync as d,writeFileSync as f}from"node:fs";import{basename as p,dirname as m,isAbsolute as h,join as g,resolve as _}from"node:path";import{fileURLToPath as v}from"node:url";import{AIKIT_PATHS as y,CONTENT_TYPES as b,EMBEDDING_DEFAULTS as x,KNOWLEDGE_ORIGINS as S,SOURCE_TYPES as C,addLogListener as w,computePartitionKey as T,createLogger as E,getGlobalDataDir as D,getPartitionDir as O,isUserInstalled as k,listWorkspaces as ee,serializeError as A}from"../../core/dist/index.js";import{createHash as j,randomUUID as M}from"node:crypto";import{readFile as te,stat as ne}from"node:fs/promises";import{homedir as N}from"node:os";import{WasmRuntime as re,initializeWasm as ie}from"../../chunker/dist/index.js";import{OnnxEmbedder as ae}from"../../embeddings/dist/index.js";import{ERCache as oe,ERClient as se,EvolutionCollector as ce,PolicyStore as le,PushAdapter as ue,mergeResults as de}from"../../enterprise-bridge/dist/index.js";import{FileHashCache as fe,IncrementalIndexer as pe}from"../../indexer/dist/index.js";import{SqliteGraphStore as me,createStore as he}from"../../store/dist/index.js";import{FileCache as ge,addToWorkset as _e,audit as ve,batch as ye,bookendReorder as be,changelog as xe,check as Se,checkpointLatest as Ce,checkpointList as we,checkpointLoad as Te,checkpointSave as Ee,codemod as De,compact as Oe,compressTerminalOutput as ke,createRestorePoint as Ae,dataTransform as je,delegate as Me,delegateListModels as Ne,deleteWorkset as Pe,diffParse as Fe,digest as Ie,encode as Le,envInfo as Re,evaluate as ze,evidenceMap as Be,fileSummary as Ve,find as He,findDeadSymbols as Ue,findExamples as We,forgeClassify as Ge,forgeGround as Ke,getWorkset as qe,gitContext as Je,graphAugmentSearch as Ye,graphQuery as Xe,guide as Ze,health as Qe,httpRequest as $e,laneCreate as et,laneDiff as tt,laneDiscard as nt,laneList as rt,laneMerge as it,laneStatus as at,listRestorePoints as ot,listWorksets as st,measure as ct,onboard as lt,parseOutput as ut,processList as dt,processLogs as ft,processStart as pt,processStatus as mt,processStop as ht,queueClear as gt,queueCreate as _t,queueDelete as vt,queueDone as yt,queueFail as bt,queueGet as xt,queueList as St,queueNext as Ct,queuePush as wt,regexTest as Tt,removeFromWorkset as Et,rename as Dt,replayAppend as Ot,replayClear as kt,replayList as At,replayTrim as jt,restoreFromPoint as Mt,saveWorkset as Nt,schemaValidate as Pt,scopeMap as Ft,sessionDigest as It,sessionDigestSampling as Lt,snippet as Rt,stashClear as zt,stashDelete as Bt,stashGet as Vt,stashList as Ht,stashSet as Ut,stratumCard as Wt,summarizeCheckResult as Gt,symbol as Kt,testRun as qt,timeUtils as Jt,trace as Yt,truncateToTokenBudget as Xt,watchList as Zt,watchStart as Qt,watchStop as $t,webFetch as en,webSearch as tn}from"../../tools/dist/index.js";import{completable as P}from"@modelcontextprotocol/sdk/server/completable.js";import{McpServer as nn,ResourceTemplate as rn}from"@modelcontextprotocol/sdk/server/mcp.js";import{z as F}from"zod";import{buildFormSchema as an,field as on,normalizeResponse as sn}from"../../elicitation/dist/index.js";import{BlastRadiusAnalyzer as cn,DependencyAnalyzer as ln,DiagramGenerator as un,EntryPointAnalyzer as dn,KnowledgeProducer as fn,PatternAnalyzer as pn,StructureAnalyzer as mn,SymbolAnalyzer as hn}from"../../analyzers/dist/index.js";import{RESOURCE_MIME_TYPE as gn,registerAppResource as _n,registerAppTool as vn}from"@modelcontextprotocol/ext-apps/server";import{exec as yn,execSync as bn}from"node:child_process";import{createServer as xn}from"node:http";import{createUIResource as Sn}from"@mcp-ui/server";import{marked as Cn}from"marked";const wn={enabled:!0,batchSize:10,debounceMs:5e3,maxFactsPerSession:50};function Tn(e){return(e.content??[]).filter(e=>e.type===`text`&&typeof e.text==`string`).map(e=>e.text??``).join(`
2
2
  `)}function En(e){return j(`sha256`).update(e).digest(`hex`).slice(0,16)}function Dn(e,t){return e.match(t)?.[1]?.trim()||null}function On(e,t,n,r){return{title:e,content:t,category:`context`,tags:n,hash:En(r)}}const kn={name:`environment-facts`,tools:[`env`,`config`,`status`],extract(e){let t=Tn(e.result);if(!t)return[];let n=[],r=Dn(t,/node(?:\.js)?\s*[:=]\s*([^\r\n]+)/i);r&&n.push(On(`Node.js version`,`Detected Node.js version: ${r}`,[`environment`,`node`],`node-version-${r}`));let i=Dn(t,/(?:^|\n)os\s*[:=]\s*([^\r\n]+)/i);i&&n.push(On(`Operating system`,`Detected OS: ${i}`,[`environment`,`os`],`os-${i}`));let a=Dn(t,/workspace(?: path)?\s*[:=]\s*([^\r\n]+)/i);a&&n.push(On(`Workspace path`,`Primary workspace path: ${a}`,[`environment`,`workspace`],`workspace-${a}`));let o=Dn(t,/store(?: backend)?\s*[:=]\s*([^\r\n]+)/i);o&&n.push(On(`Store backend`,`Configured store backend: ${o}`,[`environment`,`storage`],`store-${o}`));let s=Dn(t,/embedding(?: model)?\s*[:=]\s*([^\r\n]+)/i);return s&&n.push(On(`Embedding model`,`Configured embedding model: ${s}`,[`environment`,`embedding`],`embedding-${s}`)),e.toolName===`status`&&n.push({title:`AI Kit status snapshot`,content:`Latest status: ${t.slice(0,300)}`,category:`context`,tags:[`environment`,`status`],hash:En(`status-snapshot`)}),n}},An={name:`error-patterns`,tools:[],extract(e){if(!e.isError)return[];let t=Tn(e.result);return!t||t.length<20?[]:[{title:`Error pattern: ${e.toolName}`,content:`The ${e.toolName} tool encountered an error:\n${t.slice(0,500)}`,category:`context`,tags:[`errors`,e.toolName],hash:En(`error-${e.toolName}-${t.slice(0,100)}`)}]}};function jn(e){let t=e.toLowerCase();return t.includes(`vitest`)?`vitest`:t.includes(`jest`)?`jest`:t.includes(`mocha`)?`mocha`:null}function Mn(e,t){let n=e.split(/\r?\n/).map(e=>e.trim()).filter(e=>/\b(error|fail(?:ed|ure)?|exception)\b/i.test(e));return n.length===0?null:{title:`Build/test error pattern`,content:`Recurring error pattern detected during ${t}:\n${n.slice(0,5).join(`
3
3
  `)}`,category:`context`,tags:[`errors`,t],hash:En(n.slice(0,3).join(`
4
- `))}}const Nn={name:`test-results`,tools:[`check`,`test_run`],extract(e){let t=Tn(e.result);if(!t)return[];let n=[],r=jn(t);r&&n.push({title:`Test framework: ${r}`,content:`This project uses ${r} for testing. Detected from ${e.toolName} output.`,category:`context`,tags:[`testing`,`environment`,r],hash:En(`test-framework-${r}`)});let i=t.toLowerCase();if(e.isError||i.includes(`fail`)){let r=Mn(t,e.toolName);r&&n.push(r)}return n}};var Pn=class{extractors=[];register(e){this.extractors.push(e)}getExtractors(e){return this.extractors.filter(t=>t.tools.length===0||t.tools.includes(e))}extractAll(e){return this.getExtractors(e.toolName).flatMap(t=>t.extract(e))}};function Fn(e){return{content:[{type:`text`,text:(e instanceof Error,String(e))}],isError:!0}}function In(e,t){try{e.push(t)}catch{}}function Ln(e){return async(t,n)=>{let r=Date.now();try{let i=await n();return In(e,{toolName:t.toolName,args:t.args,result:i,durationMs:Date.now()-r,isError:i.isError===!0,timestamp:new Date().toISOString()}),i}catch(n){throw In(e,{toolName:t.toolName,args:t.args,result:Fn(n),durationMs:Date.now()-r,isError:!0,timestamp:new Date().toISOString()}),n}}}function Rn(e,t){console.info(`[auto-knowledge] ${e}`,t)}function zn(e,t){console.warn(`[auto-knowledge] ${e}`,t)}var Bn=class{queue=[];timer=null;seenHashes=new Set;storedCount=0;config;constructor(e,t,n={}){this.registry=e,this.curated=t,this.config={...wn,...n}}push(e){if(!(!this.config.enabled||!this.curated)&&!(this.storedCount>=this.config.maxFactsPerSession)){if(this.queue.push(e),this.queue.length>=this.config.batchSize){this.flush().catch(e=>{zn(`Auto-knowledge flush error`,{error:String(e)})});return}this.scheduleBatch()}}async flush(){this.timer&&=(clearTimeout(this.timer),null);let e=this.queue.splice(0);if(e.length===0||!this.curated)return;let t=new Map;for(let n of e)for(let e of this.registry.extractAll(n))!this.seenHashes.has(e.hash)&&!t.has(e.hash)&&t.set(e.hash,e);for(let e of t.values()){if(this.storedCount>=this.config.maxFactsPerSession)break;try{await this.curated.remember(e.title,e.content,e.category,e.tags),this.seenHashes.add(e.hash),this.storedCount++,Rn(`Auto-knowledge stored`,{title:e.title,category:e.category})}catch(t){zn(`Auto-knowledge store failed`,{title:e.title,error:String(t)})}}}reset(){this.queue=[],this.seenHashes.clear(),this.storedCount=0,this.timer&&=(clearTimeout(this.timer),null)}getStats(){return{queued:this.queue.length,stored:this.storedCount,seen:this.seenHashes.size}}scheduleBatch(){this.timer||=setTimeout(()=>{this.timer=null,this.flush().catch(e=>{zn(`Auto-knowledge flush error`,{error:String(e)})})},this.config.debounceMs)}};const Vn=E(`background-task`);var Hn=class{queue=[];running=null;get isRunning(){return this.running!==null}get currentTask(){return this.running}get pendingCount(){return this.queue.length}schedule(e){return new Promise((t,n)=>{this.queue.push({...e,resolve:t,reject:n}),this.running||this.processQueue()})}async processQueue(){for(;this.queue.length>0;){let e=this.queue.shift();if(!e)break;this.running=e.name,Vn.info(`Background task started`,{task:e.name,pending:this.queue.length});let t=Date.now();try{await e.fn();let n=Date.now()-t;Vn.info(`Background task completed`,{task:e.name,durationMs:n}),e.resolve()}catch(n){let r=Date.now()-t;Vn.error(`Background task failed`,{task:e.name,durationMs:r,err:n}),e.reject(n instanceof Error?n:Error(String(n)))}}this.running=null}};const Un=3e4,Wn=new Map;function Gn(e,t,n){let r=Wn.get(e);if(r&&Date.now()<r.expires)return Promise.resolve(r.data);let i=n();return Promise.resolve(i).then(n=>(Wn.set(e,{data:n,expires:Date.now()+t}),n))}function Kn(){Wn.clear()}function qn(e,t){return Gn(`curated-paths`,Un,async()=>(await e.list()).map(e=>e.path)).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Jn(e,t){return Gn(`file-paths`,Un,()=>e.listSourcePaths()).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Yn(e,t){return Gn(`symbol-names`,Un,async()=>(await e.findNodes({type:`symbol`,limit:500})).map(e=>e.name)).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Xn(e){return Ht().map(e=>e.key).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function Zn(e){return st().map(e=>e.name).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function Qn(e){return we().map(e=>e.label).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function $n(e){return typeof e==`object`&&!!e&&!Array.isArray(e)}function er(e,t){return e.tools?(Array.isArray(e.tools)?e.tools:[e.tools]).includes(t):!0}function tr(e){return e.middleware.critical===!0}var nr=class{entries=[];use(e,t){this.entries.push({middleware:e,order:t?.order??100,tools:t?.tools,name:t?.name??e.name??`anonymous`})}useFor(e,t,n){this.use(t,{order:n?.order,tools:e,name:n?.name})}wrap(e,t,n,r){let i=this.entries.filter(t=>er(t,e)).sort((e,t)=>e.order-t.order);return async(a,o)=>{let s={toolName:e,args:$n(a)?a:{},requestId:M(),meta:r,extra:o,toolConfig:n},c=async n=>{let r=i[n];if(!r)return await t(s.args,o);let a=!1,l=!1,u,d=async()=>{a=!0;let e=await c(n+1);return l=!0,u=e,e};try{return await r.middleware(s,d)}catch(t){if(tr(r))throw t;if(console.warn(`[tool-pipeline] middleware "${r.name}" failed for tool "${e}"`,t),a){if(l&&u)return u;throw t}return c(n+1)}};return c(0)}}getMiddlewareNames(){return this.entries.map(e=>e.name)}};const rr={search:{title:`Hybrid Search`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},find:{title:`Federated Find`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},symbol:{title:`Symbol Resolver — Cross-Module Definitions & References`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},trace:{title:`Data Flow Tracer — Cross-Module Call & Import Chains`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},scope_map:{title:`Task Scope Map`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},lookup:{title:`Chunk Lookup`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},dead_symbols:{title:`Dead Symbol Finder`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},file_summary:{title:`File Summary`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},analyze_structure:{title:`Analyze Structure`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_dependencies:{title:`Analyze Dependencies`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_symbols:{title:`Analyze Symbols`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_patterns:{title:`Analyze Patterns`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_entry_points:{title:`Analyze Entry Points`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_diagram:{title:`Analyze Diagram`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},blast_radius:{title:`Blast Radius`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`,`search`]},brainstorm:{title:`Brainstorm Session`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},remember:{title:`Remember Knowledge`,annotations:{readOnlyHint:!1},category:[`knowledge`]},read:{title:`Read Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},update:{title:`Update Knowledge`,annotations:{readOnlyHint:!1},category:[`knowledge`]},forget:{title:`Forget Knowledge`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`knowledge`]},list:{title:`List Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},produce_knowledge:{title:`Produce Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},compact:{title:`Semantic Compactor`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},digest:{title:`Multi-Source Digest`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},stratum_card:{title:`Stratum Card`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},forge_ground:{title:`FORGE Ground`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`forge`]},forge_classify:{title:`FORGE Classify`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`forge`]},evidence_map:{title:`Evidence Map`,annotations:{readOnlyHint:!1},category:[`forge`]},present:{title:`Rich Content Presenter`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`presentation`]},check:{title:`Typecheck & Lint`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`execution`]},test_run:{title:`Run Tests`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`execution`]},eval:{title:`Evaluate Code`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`execution`]},batch:{title:`Batch Operations`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`execution`]},audit:{title:`Project Audit`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`execution`]},rename:{title:`Rename Symbol`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`manipulation`]},restore:{title:`Restore`,annotations:{readOnlyHint:!1},category:[`manipulation`]},codemod:{title:`Codemod`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`manipulation`]},data_transform:{title:`Data Transform`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`manipulation`]},stash:{title:`Stash Values`,annotations:{readOnlyHint:!1},category:[`session`]},checkpoint:{title:`Session Checkpoint`,annotations:{readOnlyHint:!1},category:[`session`]},session_digest:{title:`Session Digest`,annotations:{readOnlyHint:!0,idempotentHint:!1},category:[`session`]},workset:{title:`Workset Manager`,annotations:{readOnlyHint:!1},category:[`session`]},lane:{title:`Exploration Lane`,annotations:{readOnlyHint:!1},category:[`session`]},git_context:{title:`Git Context`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},diff_parse:{title:`Diff Parser`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},parse_output:{title:`Parse Build Output`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},process:{title:`Process Manager`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},watch:{title:`File Watcher`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},delegate:{title:`Delegate Task`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},config:{title:`Configuration Manager`,annotations:{readOnlyHint:!1},category:[`system`]},status:{title:`AI Kit Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},health:{title:`Health Check`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},reindex:{title:`Reindex`,annotations:{readOnlyHint:!1},category:[`system`]},onboard:{title:`Onboard Codebase`,annotations:{readOnlyHint:!1},category:[`system`]},graph:{title:`Code Knowledge Graph — Module & Symbol Relationships`,annotations:{readOnlyHint:!1},category:[`system`]},guide:{title:`Tool Guide`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},replay:{title:`Replay History`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},list_tools:{title:`List Available Tools`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},describe_tool:{title:`Describe Tool`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},search_tools:{title:`Search Tools`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},changelog:{title:`Generate Changelog`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},regex_test:{title:`Regex Tester`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},encode:{title:`Encode / Decode`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},measure:{title:`Code Metrics`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},schema_validate:{title:`Schema Validator`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},snippet:{title:`Code Snippets`,annotations:{readOnlyHint:!1},category:[`utilities`]},env:{title:`Environment Info`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},time:{title:`Date & Time`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},web_fetch:{title:`Web Fetch`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`web`]},web_search:{title:`Web Search`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`web`]},http:{title:`HTTP Request`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`web`]},queue:{title:`Operation Queue`,annotations:{readOnlyHint:!1},category:[`queue`]},bridge_push:{title:`Bridge Push`,annotations:{readOnlyHint:!1},category:[`system`]},bridge_pull:{title:`Bridge Pull`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},bridge_sync:{title:`Bridge Sync Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_push:{title:`Enterprise Push`,annotations:{readOnlyHint:!1},category:[`system`]},er_pull:{title:`Enterprise Pull`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_sync_status:{title:`Enterprise Sync Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_update_policy:{title:`Enterprise Update Policy`,annotations:{readOnlyHint:!1},category:[`system`]},er_evolve_review:{title:`Enterprise Evolution Review`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},evolution_state:{title:`Evolution State`,annotations:{readOnlyHint:!1},category:[`system`]},policy_check:{title:`Policy Check`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},flow_list:{title:`Flow List`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_info:{title:`Flow Info`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_start:{title:`Flow Start`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_step:{title:`Flow Step`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_status:{title:`Flow Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_reset:{title:`Flow Reset`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_read_instruction:{title:`Flow Read Instruction`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_add:{title:`Flow Add`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_remove:{title:`Flow Remove`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_update:{title:`Flow Update`,annotations:{readOnlyHint:!1},category:[`flow`]}};function I(e){return rr[e]??{title:e,annotations:{readOnlyHint:!1},category:[]}}const ir=Symbol(`toolPipelineState`);function ar(e){return e[ir]}function or(e,t){e[ir]=t}function sr(e,t,n=``){let r=ar(e);if(r){r.pipeline=t,r.prefix=n;return}or(e,{originalRegisterTool:e.registerTool.bind(e),pipeline:t,prefix:n}),e.registerTool=(n,r,i)=>{let a=ar(e),o=a?.prefix?`${a.prefix}${n}`:n;if(!i)return a?.originalRegisterTool(o,r);if((a?.pipeline??t).getMiddlewareNames().length===0)return a?.originalRegisterTool(o,r,i);let s=I(n),c=(a?.pipeline??t).wrap(n,i,r,s);return a?.originalRegisterTool(o,r,c)}}function cr(e,t){t&&sr(e,ar(e)?.pipeline??new nr,t)}const lr=new Set([`compact`,`digest`,`file_summary`,`stratum_card`,`present`,`web_fetch`,`web_search`]);function ur(e){let t=e?.threshold??4e3,n=e?.tokenBudget??2e3;return async(e,r)=>{if(lr.has(e.toolName))return r();let i=await r();if(!Array.isArray(i.content))return i;let a=0;for(let e of i.content)e.type===`text`&&typeof e.text==`string`&&(a+=e.text.length);if(a<t)return i;let o=i.content.map(r=>{if(r.type!==`text`||typeof r.text!=`string`||r.text.length<t)return r;let i=ke({text:r.text,tool:e.toolName,tokenBudget:n});return{...r,text:i.text}});return{...i,content:o}}}function dr(e){function t(){return!!e.server?.getClientCapabilities?.()?.elicitation}async function n(n,r){if(t())try{let t=await e.server.elicitInput({message:n,requestedSchema:r});return sn(t?{action:t.action,content:t.content}:void 0)}catch{return}}return{get available(){return t()},async ask(e){return await n(e.message,e.schema)||{action:`decline`}},async confirm(e){let t=await n(e,an({confirmed:on.confirm(e)}));return t?.action===`accept`&&t.content?.confirmed===!0},async selectOne(e,t){let r=await n(e,an({selection:on.select(`Choose one`,t)}));if(r?.action!==`accept`)return null;let i=r.content?.selection;return typeof i==`string`?i:null},async selectMany(e,t){let r=await n(e,an({selections:on.multi(`Choose one or more`,t)}));if(r?.action!==`accept`)return[];let i=r.content?.selections;return Array.isArray(i)?i:[]},async promptText(e,t){let r=await n(e,an({text:on.text(e,{description:t})}));if(r?.action!==`accept`)return null;let i=r.content?.text;return typeof i==`string`?i:null}}}const fr={available:!1,async ask(){return{action:`decline`}},async confirm(){return!1},async selectOne(){return null},async selectMany(){return[]},async promptText(){return null}},pr=E(`idle-timer`);var mr=class{timer=null;cleanupFns=[];idleMs;disposed=!1;_busy=!1;constructor(e){this.idleMs=e?.idleMs??3e5}setBusy(e){this._busy=e,e?this.cancel():this.touch()}onIdle(e){this.cleanupFns.push(e)}touch(){this.disposed||this._busy||(this.cancel(),this.timer=setTimeout(()=>{this.runCleanup()},this.idleMs),this.timer.unref&&this.timer.unref())}cancel(){this.timer&&=(clearTimeout(this.timer),null)}dispose(){this.cancel(),this.cleanupFns.length=0,this.disposed=!0}async runCleanup(){if(this._busy){pr.info(`Skipping idle cleanup — background work in progress`);return}pr.info(`Idle for ${this.idleMs/1e3}s — running cleanup`);let e=await Promise.allSettled(this.cleanupFns.map(e=>e()));for(let t of e)t.status===`rejected`&&pr.warn(`Idle cleanup callback failed`,{error:String(t.reason)})}};const hr={debug:`debug`,info:`info`,warn:`warning`,error:`error`};function gr(e){return w(({level:t,component:n,message:r,data:i})=>{try{Promise.resolve(e.sendLoggingMessage({level:hr[t],logger:n,data:i?{message:r,...i}:r})).catch(()=>{})}catch{}})}const _r=E(`memory-monitor`);var vr=class{timer=null;warningBytes;criticalBytes;intervalMs;pressureFns=[];lastLevel=`normal`;constructor(e){this.warningBytes=e?.warningBytes??4294967296,this.criticalBytes=e?.criticalBytes??8589934592,this.intervalMs=e?.intervalMs??6e4}onPressure(e){this.pressureFns.push(e)}start(){this.timer||(this.timer=setInterval(()=>this.check(),this.intervalMs),this.timer.unref&&this.timer.unref(),_r.info(`Memory monitor started`,{warningMB:Math.round(this.warningBytes/1024/1024),criticalMB:Math.round(this.criticalBytes/1024/1024),intervalSec:Math.round(this.intervalMs/1e3)}))}stop(){this.timer&&=(clearInterval(this.timer),null)}getRssBytes(){return process.memoryUsage.rss()}check(){let e=this.getRssBytes(),t=`normal`;if(e>=this.criticalBytes?t=`critical`:e>=this.warningBytes&&(t=`warning`),t!==this.lastLevel||t===`critical`){let n=Math.round(e/1024/1024);t===`critical`?_r.warn(`Memory CRITICAL: ${n}MB RSS — consider restarting the server`):t===`warning`?_r.warn(`Memory WARNING: ${n}MB RSS`):this.lastLevel!==`normal`&&_r.info(`Memory returned to normal: ${n}MB RSS`),this.lastLevel=t}if(t!==`normal`)for(let n of this.pressureFns)try{n(t,e)}catch{}return t===`critical`&&typeof globalThis.gc==`function`&&globalThis.gc(),t}};function yr(e,t,n){if(e.registerPrompt(`ready`,{title:`AI Kit Ready`,description:`AI Kit is ready — quick-start guide for search, onboard, and workflows`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`AI Kit is ready. Quick start:`,``,'• **New project?** → `onboard({ path: "." })` for full codebase analysis','• **Returning?** → `status({})` then `search({ query: "SESSION CHECKPOINT", origin: "curated" })`','• **Exploring?** → `guide({ goal: "your task" })` for workflow recommendations','• **Quick lookup?** → `search({ query: "your question" })`'].join(`
4
+ `))}}const Nn={name:`test-results`,tools:[`check`,`test_run`],extract(e){let t=Tn(e.result);if(!t)return[];let n=[],r=jn(t);r&&n.push({title:`Test framework: ${r}`,content:`This project uses ${r} for testing. Detected from ${e.toolName} output.`,category:`context`,tags:[`testing`,`environment`,r],hash:En(`test-framework-${r}`)});let i=t.toLowerCase();if(e.isError||i.includes(`fail`)){let r=Mn(t,e.toolName);r&&n.push(r)}return n}};var Pn=class{extractors=[];register(e){this.extractors.push(e)}getExtractors(e){return this.extractors.filter(t=>t.tools.length===0||t.tools.includes(e))}extractAll(e){return this.getExtractors(e.toolName).flatMap(t=>t.extract(e))}};function Fn(e){return{content:[{type:`text`,text:(e instanceof Error,String(e))}],isError:!0}}function In(e,t){try{e.push(t)}catch{}}function Ln(e){return async(t,n)=>{let r=Date.now();try{let i=await n();return In(e,{toolName:t.toolName,args:t.args,result:i,durationMs:Date.now()-r,isError:i.isError===!0,timestamp:new Date().toISOString()}),i}catch(n){throw In(e,{toolName:t.toolName,args:t.args,result:Fn(n),durationMs:Date.now()-r,isError:!0,timestamp:new Date().toISOString()}),n}}}function Rn(e,t){console.info(`[auto-knowledge] ${e}`,t)}function zn(e,t){console.warn(`[auto-knowledge] ${e}`,t)}var Bn=class{queue=[];timer=null;seenHashes=new Set;storedCount=0;config;constructor(e,t,n={}){this.registry=e,this.curated=t,this.config={...wn,...n}}push(e){if(!(!this.config.enabled||!this.curated)&&!(this.storedCount>=this.config.maxFactsPerSession)){if(this.queue.push(e),this.queue.length>=this.config.batchSize){this.flush().catch(e=>{zn(`Auto-knowledge flush error`,{error:String(e)})});return}this.scheduleBatch()}}async flush(){this.timer&&=(clearTimeout(this.timer),null);let e=this.queue.splice(0);if(e.length===0||!this.curated)return;let t=new Map;for(let n of e)for(let e of this.registry.extractAll(n))!this.seenHashes.has(e.hash)&&!t.has(e.hash)&&t.set(e.hash,e);for(let e of t.values()){if(this.storedCount>=this.config.maxFactsPerSession)break;try{await this.curated.remember(e.title,e.content,e.category,e.tags),this.seenHashes.add(e.hash),this.storedCount++,Rn(`Auto-knowledge stored`,{title:e.title,category:e.category})}catch(t){zn(`Auto-knowledge store failed`,{title:e.title,error:String(t)})}}}reset(){this.queue=[],this.seenHashes.clear(),this.storedCount=0,this.timer&&=(clearTimeout(this.timer),null)}getStats(){return{queued:this.queue.length,stored:this.storedCount,seen:this.seenHashes.size}}scheduleBatch(){this.timer||=setTimeout(()=>{this.timer=null,this.flush().catch(e=>{zn(`Auto-knowledge flush error`,{error:String(e)})})},this.config.debounceMs)}};const Vn=E(`background-task`);var Hn=class{queue=[];running=null;get isRunning(){return this.running!==null}get currentTask(){return this.running}get pendingCount(){return this.queue.length}schedule(e){return new Promise((t,n)=>{this.queue.push({...e,resolve:t,reject:n}),this.running||this.processQueue()})}async processQueue(){for(;this.queue.length>0;){let e=this.queue.shift();if(!e)break;this.running=e.name,Vn.info(`Background task started`,{task:e.name,pending:this.queue.length});let t=Date.now();try{await e.fn();let n=Date.now()-t;Vn.info(`Background task completed`,{task:e.name,durationMs:n}),e.resolve()}catch(n){let r=Date.now()-t;Vn.error(`Background task failed`,{task:e.name,durationMs:r,err:n}),e.reject(n instanceof Error?n:Error(String(n)))}}this.running=null}};const Un=3e4,Wn=new Map;function Gn(e,t,n){let r=Wn.get(e);if(r&&Date.now()<r.expires)return Promise.resolve(r.data);let i=n();return Promise.resolve(i).then(n=>(Wn.set(e,{data:n,expires:Date.now()+t}),n))}function Kn(){Wn.clear()}function qn(e,t){return Gn(`curated-paths`,Un,async()=>(await e.list()).map(e=>e.path)).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Jn(e,t){return Gn(`file-paths`,Un,()=>e.listSourcePaths()).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Yn(e,t){return Gn(`symbol-names`,Un,async()=>(await e.findNodes({type:`symbol`,limit:500})).map(e=>e.name)).then(e=>e.filter(e=>e.toLowerCase().includes(t.toLowerCase())).slice(0,20))}function Xn(e){return Ht().map(e=>e.key).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function Zn(e){return st().map(e=>e.name).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function Qn(e){return we().map(e=>e.label).filter(t=>t.toLowerCase().includes(e.toLowerCase())).slice(0,20)}function $n(e){return typeof e==`object`&&!!e&&!Array.isArray(e)}function er(e,t){return e.tools?(Array.isArray(e.tools)?e.tools:[e.tools]).includes(t):!0}function tr(e){return e.middleware.critical===!0}var nr=class{entries=[];use(e,t){this.entries.push({middleware:e,order:t?.order??100,tools:t?.tools,name:t?.name??e.name??`anonymous`})}useFor(e,t,n){this.use(t,{order:n?.order,tools:e,name:n?.name})}wrap(e,t,n,r){let i=this.entries.filter(t=>er(t,e)).sort((e,t)=>e.order-t.order);return async(a,o)=>{let s={toolName:e,args:$n(a)?a:{},requestId:M(),meta:r,extra:o,toolConfig:n},c=async n=>{let r=i[n];if(!r)return await t(s.args,o);let a=!1,l=!1,u,d=async()=>{a=!0;let e=await c(n+1);return l=!0,u=e,e};try{return await r.middleware(s,d)}catch(t){if(tr(r))throw t;if(console.warn(`[tool-pipeline] middleware "${r.name}" failed for tool "${e}"`,t),a){if(l&&u)return u;throw t}return c(n+1)}};return c(0)}}getMiddlewareNames(){return this.entries.map(e=>e.name)}};const I={search:{title:`Hybrid Search`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},find:{title:`Federated Find`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},symbol:{title:`Symbol Resolver — Cross-Module Definitions & References`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},trace:{title:`Data Flow Tracer — Cross-Module Call & Import Chains`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},scope_map:{title:`Task Scope Map`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},lookup:{title:`Chunk Lookup`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},dead_symbols:{title:`Dead Symbol Finder`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},file_summary:{title:`File Summary`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`search`]},analyze_structure:{title:`Analyze Structure`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_dependencies:{title:`Analyze Dependencies`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_symbols:{title:`Analyze Symbols`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_patterns:{title:`Analyze Patterns`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_entry_points:{title:`Analyze Entry Points`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},analyze_diagram:{title:`Analyze Diagram`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},blast_radius:{title:`Blast Radius`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`,`search`]},brainstorm:{title:`Brainstorm Session`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`analysis`]},remember:{title:`Remember Knowledge`,annotations:{readOnlyHint:!1},category:[`knowledge`]},read:{title:`Read Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},update:{title:`Update Knowledge`,annotations:{readOnlyHint:!1},category:[`knowledge`]},forget:{title:`Forget Knowledge`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`knowledge`]},list:{title:`List Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},produce_knowledge:{title:`Produce Knowledge`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`knowledge`]},compact:{title:`Semantic Compactor`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},digest:{title:`Multi-Source Digest`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},stratum_card:{title:`Stratum Card`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`compression`]},forge_ground:{title:`FORGE Ground`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`forge`]},forge_classify:{title:`FORGE Classify`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`forge`]},evidence_map:{title:`Evidence Map`,annotations:{readOnlyHint:!1},category:[`forge`]},present:{title:`Rich Content Presenter`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`presentation`]},check:{title:`Typecheck & Lint`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`execution`]},test_run:{title:`Run Tests`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`execution`]},eval:{title:`Evaluate Code`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`execution`]},batch:{title:`Batch Operations`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`execution`]},audit:{title:`Project Audit`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`execution`]},rename:{title:`Rename Symbol`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`manipulation`]},restore:{title:`Restore`,annotations:{readOnlyHint:!1},category:[`manipulation`]},codemod:{title:`Codemod`,annotations:{readOnlyHint:!1,destructiveHint:!0},category:[`manipulation`]},data_transform:{title:`Data Transform`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`manipulation`]},stash:{title:`Stash Values`,annotations:{readOnlyHint:!1},category:[`session`]},checkpoint:{title:`Session Checkpoint`,annotations:{readOnlyHint:!1},category:[`session`]},session_digest:{title:`Session Digest`,annotations:{readOnlyHint:!0,idempotentHint:!1},category:[`session`]},workset:{title:`Workset Manager`,annotations:{readOnlyHint:!1},category:[`session`]},lane:{title:`Exploration Lane`,annotations:{readOnlyHint:!1},category:[`session`]},git_context:{title:`Git Context`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},diff_parse:{title:`Diff Parser`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},parse_output:{title:`Parse Build Output`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`git`]},process:{title:`Process Manager`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},watch:{title:`File Watcher`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},delegate:{title:`Delegate Task`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`process`]},config:{title:`Configuration Manager`,annotations:{readOnlyHint:!1},category:[`system`]},status:{title:`AI Kit Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},health:{title:`Health Check`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},reindex:{title:`Reindex`,annotations:{readOnlyHint:!1},category:[`system`]},onboard:{title:`Onboard Codebase`,annotations:{readOnlyHint:!1},category:[`system`]},graph:{title:`Code Knowledge Graph — Module & Symbol Relationships`,annotations:{readOnlyHint:!1},category:[`system`]},guide:{title:`Tool Guide`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},replay:{title:`Replay History`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},list_tools:{title:`List Available Tools`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},describe_tool:{title:`Describe Tool`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},search_tools:{title:`Search Tools`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`meta`]},changelog:{title:`Generate Changelog`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},regex_test:{title:`Regex Tester`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},encode:{title:`Encode / Decode`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},measure:{title:`Code Metrics`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},schema_validate:{title:`Schema Validator`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},snippet:{title:`Code Snippets`,annotations:{readOnlyHint:!1},category:[`utilities`]},env:{title:`Environment Info`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},time:{title:`Date & Time`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`utilities`]},web_fetch:{title:`Web Fetch`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`web`]},web_search:{title:`Web Search`,annotations:{readOnlyHint:!0,openWorldHint:!0},category:[`web`]},http:{title:`HTTP Request`,annotations:{readOnlyHint:!1,openWorldHint:!0},category:[`web`]},queue:{title:`Operation Queue`,annotations:{readOnlyHint:!1},category:[`queue`]},bridge_push:{title:`Bridge Push`,annotations:{readOnlyHint:!1},category:[`system`]},bridge_pull:{title:`Bridge Pull`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},bridge_sync:{title:`Bridge Sync Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_push:{title:`Enterprise Push`,annotations:{readOnlyHint:!1},category:[`system`]},er_pull:{title:`Enterprise Pull`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_sync_status:{title:`Enterprise Sync Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},er_update_policy:{title:`Enterprise Update Policy`,annotations:{readOnlyHint:!1},category:[`system`]},er_evolve_review:{title:`Enterprise Evolution Review`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},evolution_state:{title:`Evolution State`,annotations:{readOnlyHint:!1},category:[`system`]},policy_check:{title:`Policy Check`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`system`]},flow_list:{title:`Flow List`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_info:{title:`Flow Info`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_start:{title:`Flow Start`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_step:{title:`Flow Step`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_status:{title:`Flow Status`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_reset:{title:`Flow Reset`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_read_instruction:{title:`Flow Read Instruction`,annotations:{readOnlyHint:!0,idempotentHint:!0},category:[`flow`]},flow_add:{title:`Flow Add`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_remove:{title:`Flow Remove`,annotations:{readOnlyHint:!1},category:[`flow`]},flow_update:{title:`Flow Update`,annotations:{readOnlyHint:!1},category:[`flow`]}};function L(e){return I[e]??{title:e,annotations:{readOnlyHint:!1},category:[]}}const rr=Symbol(`toolPipelineState`);function ir(e){return e[rr]}function ar(e,t){e[rr]=t}function or(e,t,n=``){let r=ir(e);if(r){r.pipeline=t,r.prefix=n;return}ar(e,{originalRegisterTool:e.registerTool.bind(e),pipeline:t,prefix:n}),e.registerTool=(n,r,i)=>{let a=ir(e),o=a?.prefix?`${a.prefix}${n}`:n;if(!i)return a?.originalRegisterTool(o,r);if((a?.pipeline??t).getMiddlewareNames().length===0)return a?.originalRegisterTool(o,r,i);let s=L(n),c=(a?.pipeline??t).wrap(n,i,r,s);return a?.originalRegisterTool(o,r,c)}}function sr(e,t){t&&or(e,ir(e)?.pipeline??new nr,t)}const cr=new Set([`compact`,`digest`,`file_summary`,`stratum_card`,`present`,`web_fetch`,`web_search`]);function lr(e){let t=e?.threshold??4e3,n=e?.tokenBudget??2e3;return async(e,r)=>{if(cr.has(e.toolName))return r();let i=await r();if(!Array.isArray(i.content))return i;let a=0;for(let e of i.content)e.type===`text`&&typeof e.text==`string`&&(a+=e.text.length);if(a<t)return i;let o=i.content.map(r=>{if(r.type!==`text`||typeof r.text!=`string`||r.text.length<t)return r;let i=ke({text:r.text,tool:e.toolName,tokenBudget:n});return{...r,text:i.text}});return{...i,content:o}}}function ur(e){function t(){return!!e.server?.getClientCapabilities?.()?.elicitation}async function n(n,r){if(t())try{let t=await e.server.elicitInput({message:n,requestedSchema:r});return sn(t?{action:t.action,content:t.content}:void 0)}catch{return}}return{get available(){return t()},async ask(e){return await n(e.message,e.schema)||{action:`decline`}},async confirm(e){let t=await n(e,an({confirmed:on.confirm(e)}));return t?.action===`accept`&&t.content?.confirmed===!0},async selectOne(e,t){let r=await n(e,an({selection:on.select(`Choose one`,t)}));if(r?.action!==`accept`)return null;let i=r.content?.selection;return typeof i==`string`?i:null},async selectMany(e,t){let r=await n(e,an({selections:on.multi(`Choose one or more`,t)}));if(r?.action!==`accept`)return[];let i=r.content?.selections;return Array.isArray(i)?i:[]},async promptText(e,t){let r=await n(e,an({text:on.text(e,{description:t})}));if(r?.action!==`accept`)return null;let i=r.content?.text;return typeof i==`string`?i:null}}}const dr={available:!1,async ask(){return{action:`decline`}},async confirm(){return!1},async selectOne(){return null},async selectMany(){return[]},async promptText(){return null}},fr=E(`idle-timer`);var pr=class{timer=null;cleanupFns=[];idleMs;disposed=!1;_busy=!1;constructor(e){this.idleMs=e?.idleMs??3e5}setBusy(e){this._busy=e,e?this.cancel():this.touch()}onIdle(e){this.cleanupFns.push(e)}touch(){this.disposed||this._busy||(this.cancel(),this.timer=setTimeout(()=>{this.runCleanup()},this.idleMs),this.timer.unref&&this.timer.unref())}cancel(){this.timer&&=(clearTimeout(this.timer),null)}dispose(){this.cancel(),this.cleanupFns.length=0,this.disposed=!0}async runCleanup(){if(this._busy){fr.info(`Skipping idle cleanup — background work in progress`);return}fr.info(`Idle for ${this.idleMs/1e3}s — running cleanup`);let e=await Promise.allSettled(this.cleanupFns.map(e=>e()));for(let t of e)t.status===`rejected`&&fr.warn(`Idle cleanup callback failed`,{error:String(t.reason)})}};const mr={debug:`debug`,info:`info`,warn:`warning`,error:`error`};function hr(e){return w(({level:t,component:n,message:r,data:i})=>{try{Promise.resolve(e.sendLoggingMessage({level:mr[t],logger:n,data:i?{message:r,...i}:r})).catch(()=>{})}catch{}})}const gr=E(`memory-monitor`);var _r=class{timer=null;warningBytes;criticalBytes;intervalMs;pressureFns=[];lastLevel=`normal`;constructor(e){this.warningBytes=e?.warningBytes??4294967296,this.criticalBytes=e?.criticalBytes??8589934592,this.intervalMs=e?.intervalMs??6e4}onPressure(e){this.pressureFns.push(e)}start(){this.timer||(this.timer=setInterval(()=>this.check(),this.intervalMs),this.timer.unref&&this.timer.unref(),gr.info(`Memory monitor started`,{warningMB:Math.round(this.warningBytes/1024/1024),criticalMB:Math.round(this.criticalBytes/1024/1024),intervalSec:Math.round(this.intervalMs/1e3)}))}stop(){this.timer&&=(clearInterval(this.timer),null)}getRssBytes(){return process.memoryUsage.rss()}check(){let e=this.getRssBytes(),t=`normal`;if(e>=this.criticalBytes?t=`critical`:e>=this.warningBytes&&(t=`warning`),t!==this.lastLevel||t===`critical`){let n=Math.round(e/1024/1024);t===`critical`?gr.warn(`Memory CRITICAL: ${n}MB RSS — consider restarting the server`):t===`warning`?gr.warn(`Memory WARNING: ${n}MB RSS`):this.lastLevel!==`normal`&&gr.info(`Memory returned to normal: ${n}MB RSS`),this.lastLevel=t}if(t!==`normal`)for(let n of this.pressureFns)try{n(t,e)}catch{}return t===`critical`&&typeof globalThis.gc==`function`&&globalThis.gc(),t}};function vr(e,t,n){if(e.registerPrompt(`ready`,{title:`AI Kit Ready`,description:`AI Kit is ready — quick-start guide for search, onboard, and workflows`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`AI Kit is ready. Quick start:`,``,'• **New project?** → `onboard({ path: "." })` for full codebase analysis','• **Returning?** → `status({})` then `search({ query: "SESSION CHECKPOINT", origin: "curated" })`','• **Exploring?** → `guide({ goal: "your task" })` for workflow recommendations','• **Quick lookup?** → `search({ query: "your question" })`'].join(`
5
5
  `)}}]})),e.registerPrompt(`onboard`,{title:`Onboard Codebase`,description:`Analyze the codebase for first-time onboarding — runs all analyzers and produces a knowledge summary`,argsSchema:{path:F.string().optional().describe(`Path to analyze (default: workspace root)`)}},async({path:e})=>({messages:[{role:`user`,content:{type:`text`,text:[`Run the full onboarding workflow for "${e??`.`}"`,``,`1. \`onboard({ path: "${e??`.`}" })\` — full codebase analysis`,`2. \`produce_knowledge({ path: "${e??`.`}" })\` — generate synthesis`,"3. `remember` key findings as curated entries","4. `status` to verify index health"].join(`
6
6
  `)}}]})),e.registerPrompt(`sessionStart`,{title:`Start AI Kit Session`,description:`Initialize an AI Kit session — check status, list knowledge, and resume from last checkpoint`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`Run the session start protocol:`,``,"1. `status({})` — check AI Kit health and onboard state","2. `list()` — see stored knowledge entries",'3. `search({ query: "SESSION CHECKPOINT", origin: "curated" })` — resume prior work'].join(`
7
7
  `)}}]})),e.registerPrompt(`sessionEnd`,{title:`End AI Kit Session`,description:`Persist decisions and create a session checkpoint before ending`,argsSchema:{summary:F.string().describe(`Brief summary of decisions made, blockers encountered, and next steps`)}},async({summary:e})=>({messages:[{role:`user`,content:{type:`text`,text:[`Run the session end protocol:`,``,'1. `remember({ title: "Session checkpoint: '+e.slice(0,60)+`...", content: "`+e.replace(/"/g,`\\"`)+'", category: "conventions" })` — persist findings',n===`smart`?`2. Smart indexing is active — index refreshes automatically`:"2. `reindex({})` — refresh search index if files changed",`3. Confirm session data saved`].join(`
@@ -13,27 +13,27 @@ import{n as e,t}from"./curated-manager-DX-_oJg0.js";import{autoUpgradeScaffold a
13
13
  `)}}]})),e.registerPrompt(`investigate_workflow`,{title:`Investigate Issue`,description:`Bug investigation workflow: trace → analyze → find root cause → fix`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`## Bug Investigation Workflow`,``,"1. `parse_output({ output })` — Parse error messages if available","2. `symbol({ name })` — Find the failing symbol",'3. `trace({ symbol, direction: "backward" })` — Trace data flow to root cause',"4. `blast_radius({ changed_files })` — Assess impact of fix",`5. Implement fix`,"6. `check({})` + `test_run({})` — Verify fix works",``,`Describe the issue:`].join(`
14
14
  `)}}]})),e.registerPrompt(`context_workflow`,{title:`Build Context`,description:`Progressive context building: file_summary → compact → digest for deep understanding`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`## Context Building Workflow`,``,`Build understanding progressively:`,``,"1. `file_summary({ path })` — Structure overview (~50 tokens)","2. `compact({ path, query })` — Extract relevant sections (5-20x reduction)","3. `digest({ sources })` — Multi-file compressed summary","4. `stratum_card({ files })` — Reusable reference card",``,`For project-wide context:`,"- `scope_map({ task })` — Reading plan for a specific task",'- `workset({ action: "save" })` — Save working set for later',``,`What do you need to understand?`].join(`
15
15
  `)}}]})),e.registerPrompt(`migrate_workflow`,{title:`Migration Workflow`,description:`Code migration workflow: analyze → plan → migrate → verify → document`},async()=>({messages:[{role:`user`,content:{type:`text`,text:[`## Migration Workflow`,``,"1. `analyze_structure({ path })` — Current codebase state","2. `analyze_patterns({ path })` — Identify patterns to migrate","3. `dead_symbols({ path })` — Find unused code to remove",'4. `scope_map({ task: "migration" })` — Plan migration scope',"5. Use `codemod` for bulk transformations","6. `check({})` + `test_run({})` — Verify after each step","7. `changelog({ from })` — Document changes",``,`What are you migrating from/to?`].join(`
16
- `)}}]}))}const br=E(`auto-gc`);let xr=`warming`,Sr=null,Cr=0;function wr(){let e=[...Pr()];if(e.length<10)return 0;e.sort((e,t)=>e-t);let t=Math.floor(e.length*.95);return e[Math.min(t,e.length-1)]}function Tr(e){let t=wr();if(t===0){xr=`warming`;return}if(t>500){xr=`degraded`;let n=Date.now();if(Sr&&n-Sr<36e5){br.debug(`GC cooldown active — skipping`,{lastGcAgoMs:n-Sr});return}Sr=n,Cr++,br.warn(`p95 latency exceeds threshold, triggering GC`,{p95:t,cycle:Cr}),e&&e().catch(e=>br.error(`GC callback failed`,{err:String(e)}))}else t<200&&(xr=`healthy`)}function Er(){return{state:xr,p95:wr(),lastGcAt:Sr,gcCount:Cr,bufferSize:Pr().length}}const Dr=/key|token|secret|auth|password|bearer/i,Or=new Set([`eval`,`env`]);function kr(e,t){if(Or.has(e))return JSON.stringify({_redacted:!0,tool:e});if(e===`http`&&t.headers&&typeof t.headers==`object`){let e={...t},n={};for(let[e,r]of Object.entries(t.headers))n[e]=Dr.test(e)?`[REDACTED]`:r;return e.headers=n,JSON.stringify(e).slice(0,200)}return JSON.stringify(t).slice(0,200)}const Ar=new Map,jr=[];function Mr(e,t,n,r,i){let a=Ar.get(e);a||(a={callCount:0,totalDurationMs:0,totalInputChars:0,totalOutputChars:0,errorCount:0},Ar.set(e,a)),a.callCount++,a.totalDurationMs+=t,a.totalInputChars+=n,a.totalOutputChars+=r,i&&a.errorCount++,jr.push(t),jr.length>100&&jr.shift(),jr.length>=10&&jr.length%20==0&&Tr()}function Nr(){return Array.from(Ar.entries()).map(([e,t])=>({tool:e,...t}))}function Pr(){return jr}function Fr(){return async(e,t)=>{let n=Date.now(),r=JSON.stringify(e.args??{}).length;try{let i=await t(),a=Date.now()-n,o=JSON.stringify(i);return Ot({ts:new Date().toISOString(),source:`mcp`,tool:e.toolName,input:kr(e.toolName,e.args),durationMs:a,status:`ok`,output:o.slice(0,200)}),Mr(e.toolName,a,r,o.length,!1),i}catch(t){let i=Date.now()-n;throw Ot({ts:new Date().toISOString(),source:`mcp`,tool:e.toolName,input:kr(e.toolName,e.args),durationMs:i,status:`error`,output:t instanceof Error?t.message:String(t)}),Mr(e.toolName,i,r,0,!0),t}}}const Ir=E(`resource-notifier`);var Lr=class{constructor(e){this.mcpServer=e}async notifyStatusChanged(){await this.sendUpdate(`aikit://status`)}async notifyFileTreeChanged(){await this.sendUpdate(`aikit://file-tree`)}async notifyCuratedIndexChanged(){await this.sendUpdate(`aikit://curated`)}async notifyCuratedEntryChanged(e){await this.sendUpdate(`aikit://curated/${e}`)}async notifyResourceListChanged(){try{await this.mcpServer.server.sendResourceListChanged()}catch(e){Ir.debug(`sendResourceListChanged failed`,{error:String(e)})}}async notifyAfterReindex(){await Promise.allSettled([this.notifyStatusChanged(),this.notifyFileTreeChanged()])}async notifyAfterCuratedWrite(e){let t=[this.notifyStatusChanged(),this.notifyCuratedIndexChanged()];e&&t.push(this.notifyCuratedEntryChanged(e)),await Promise.allSettled(t)}async sendUpdate(e){try{await this.mcpServer.server.sendResourceUpdated({uri:e})}catch(t){Ir.debug(`sendResourceUpdated failed`,{uri:e,error:String(t)})}}};function Rr(e,t){e.resource(`aikit-curated-index`,`aikit://curated`,{description:`Index of all curated knowledge entries`,mimeType:`text/markdown`},async()=>{let e=(await t.list()).map(e=>`- [${e.title}](aikit://curated/${e.path}) — ${e.category}`);return{contents:[{uri:`aikit://curated`,text:`# Curated Knowledge Index\n\n${e.length>0?e.join(`
17
- `):`_No curated entries yet._`}`,mimeType:`text/markdown`}]}});let n=new rn(`aikit://curated/{+path}`,{list:async()=>({resources:(await t.list()).map(e=>({uri:`aikit://curated/${e.path}`,name:e.title,description:`[${e.category}] ${e.contentPreview?.slice(0,80)??``}`,mimeType:`text/markdown`}))})});e.resource(`aikit-curated-entry`,n,{description:`A curated knowledge entry`,mimeType:`text/markdown`},async(e,n)=>{let r=n.path;if(!r)throw Error(`Missing path variable in curated resource URI`);let i=await t.read(r);return{contents:[{uri:e.toString(),text:`---\ntitle: ${i.title}\ncategory: ${i.category}\ntags: ${i.tags?.join(`, `)??``}\nversion: ${i.version??1}\n---\n\n${i.content??i.contentPreview??``}`,mimeType:`text/markdown`}]}})}function zr(e,t,n){e.resource(`aikit-status`,`aikit://status`,{description:`Current AI Kit status and statistics`,mimeType:`text/plain`},async()=>{let e=await t.getStats();return{contents:[{uri:`aikit://status`,text:`AI Kit: ${e.totalRecords} records from ${e.totalFiles} files. Last indexed: ${e.lastIndexedAt??`Never`}`,mimeType:`text/plain`}]}}),e.resource(`aikit-file-tree`,`aikit://file-tree`,{description:`List of all indexed source files`,mimeType:`text/plain`},async()=>({contents:[{uri:`aikit://file-tree`,text:(await t.listSourcePaths()).sort().join(`
18
- `),mimeType:`text/plain`}]})),Rr(e,n)}const Br=E(`sampling`);function Vr(e){if(!e.modelPreferences)return;let t={...e.modelPreferences.costPriority===void 0?{}:{costPriority:e.modelPreferences.costPriority},...e.modelPreferences.speedPriority===void 0?{}:{speedPriority:e.modelPreferences.speedPriority},...e.modelPreferences.intelligencePriority===void 0?{}:{intelligencePriority:e.modelPreferences.intelligencePriority}};return Object.keys(t).length>0?t:void 0}function Hr(e){if(Array.isArray(e))return e.map(e=>Hr(e)).filter(e=>e.length>0).join(`
19
- `);if(e&&typeof e==`object`&&`type`in e){let t=e;if(t.type===`text`&&typeof t.text==`string`)return t.text}return JSON.stringify(e)}function Ur(e){let t=e.server,n=typeof t?.createMessage==`function`;return{get available(){return n},async createMessage(e){if(!n)throw Error(`Sampling not available: client does not support createMessage`);let r=e.context?`${e.context}\n\n---\n\n${e.prompt}`:e.prompt;try{let n=await t.createMessage({messages:[{role:`user`,content:{type:`text`,text:r}}],systemPrompt:e.systemPrompt,modelPreferences:Vr(e),maxTokens:e.maxTokens??4e3});return{text:Hr(n.content),model:n.model,stopReason:n.stopReason}}catch(e){throw Br.warn(`Sampling createMessage failed`,{error:String(e)}),e}}}}function Wr(){let e=(async(e,t)=>{let n=await t(),r=e.toolConfig??{};return r.outputSchema==null||n.structuredContent!=null||(n.structuredContent=Gr(r.outputSchema)),n});return e.critical=!0,e}function Gr(e){if(!e)return{};if(e.anyOf){let t=e.anyOf.find(e=>e.type!==`null`);return t?Gr(t):null}switch(e.type){case`object`:{let t={},n=e.properties??e.shape;if(n)for(let[e,r]of Object.entries(n))t[e]=Gr(r);return t}case`array`:return[];case`string`:return``;case`number`:case`integer`:return 0;case`boolean`:return!1;case`nullable`:return null;case`optional`:return;case`record`:return{};default:return{}}}const Kr=new Set([`status`,`config`,`guide`,`health`]),qr={full:{description:`All tools enabled (default)`,includeCategories:[]},safe:{description:`Read-only tools — no file/state modifications`,includeCategories:[`search`,`analysis`,`compression`,`utilities`,`system`,`git`,`flow`],excludeTools:[`reindex`,`onboard`]},research:{description:`Search, analysis, knowledge, and web access`,includeCategories:[`search`,`analysis`,`knowledge`,`compression`,`web`,`system`,`flow`]},minimal:{description:`Essential tools only — search, status, basic operations`,includeCategories:[`search`,`system`],includeTools:[`compact`,`file_summary`,`check`,`test_run`]},discovery:{description:`Full toolset plus discovery-oriented meta-tools for guided tool exploration`,includeCategories:[],includeTools:[`list_tools`,`describe_tool`,`search_tools`]}};function Jr(e,t){let n=qr[e];if(n)return n;let r=t?.[e];if(r)return r;throw Error(`Unknown tool profile: ${e}`)}function Yr(e,t,n){return t.includes(`*`)?!0:(n[e]?.category??[]).some(e=>t.includes(e))}function Xr(e,t,n,r){if(t.includes(`*`)){for(let t of n)e.add(t);return}for(let i of n)Yr(i,t,r)&&e.add(i)}function Zr(e,t,n){if(t.includes(`*`)){e.clear();return}for(let r of[...e])Yr(r,t,n)&&e.delete(r)}function Qr(e,t,n){if(!e||e.length===0||e.includes(`*`))return new Set(t);let r=new Set;return Xr(r,e,t,n),r}function $r(e,t,n,r){let i=(e,a,o)=>{if(a>10)throw Error(`Tool profile inheritance exceeded max depth 10: ${[...o,e].join(` -> `)}`);if(o.includes(e))throw Error(`Tool profile inheritance cycle detected: ${[...o,e].join(` -> `)}`);let s=Jr(e,r),c=[...o,e],l=s.extends?i(s.extends,a+1,c):Qr(s.includeCategories,t,n);s.extends&&s.includeCategories?.length&&Xr(l,s.includeCategories,t,n),s.excludeCategories?.length&&Zr(l,s.excludeCategories,n);for(let e of s.includeTools??[])l.add(e);for(let e of s.excludeTools??[])l.delete(e);return l},a=i(e,0,[]);for(let e of Kr)a.add(e);return a}const ei=E(`tool-timeout`),ti=new Set([`onboard`,`reindex`,`produce_knowledge`,`analyze_structure`,`analyze_dependencies`,`analyze_symbols`,`analyze_patterns`,`analyze_entry_points`,`analyze_diagram`,`codemod`,`batch`,`audit`]);var ni=class extends Error{constructor(e,t){super(`Tool "${e}" timed out after ${t}ms`),this.toolName=e,this.timeoutMs=t,this.name=`ToolTimeoutError`}};function ri(e){return ti.has(e)?6e5:12e4}function ii(e,t,n){return new Promise((r,i)=>{let a=!1,o=setTimeout(()=>{if(!a){a=!0;let e=new ni(n,t);ei.warn(e.message),i(e)}},t);o.unref&&o.unref(),e().then(e=>{a||(a=!0,clearTimeout(o),r(e))},e=>{a||(a=!0,clearTimeout(o),i(e))})})}const ai=F.object({totalRecords:F.number(),totalFiles:F.number(),lastIndexedAt:F.string().nullable(),onboarded:F.boolean(),onboardDir:F.string(),contentTypes:F.record(F.string(),F.number()),wasmAvailable:F.boolean(),graphStats:F.object({nodes:F.number(),edges:F.number()}).nullable(),curatedCount:F.number(),serverVersion:F.string(),scaffoldVersion:F.string().nullable(),workspaceScaffoldVersion:F.string().nullable(),upgradeAvailable:F.boolean(),contextPressure:F.number().min(0).max(100).describe(`0–100 score indicating knowledge base saturation`)}),oi=F.object({entries:F.array(F.object({path:F.string(),title:F.string(),category:F.string(),tags:F.array(F.string()),version:F.number(),preview:F.string()})),totalCount:F.number()}),si=F.object({ok:F.boolean(),checks:F.array(F.object({name:F.string(),ok:F.boolean(),message:F.string().optional()}))}),ci=F.object({summary:F.object({totalFiles:F.number(),totalLines:F.number(),totalCodeLines:F.number(),totalFunctions:F.number(),avgComplexity:F.number(),maxComplexity:F.object({value:F.number(),file:F.string()})}),files:F.array(F.object({path:F.string(),lines:F.number(),code:F.number(),complexity:F.number(),functions:F.number()}))}),li=F.object({platform:F.string(),arch:F.string(),nodeVersion:F.string(),cwd:F.string(),cpus:F.number(),memoryFreeGb:F.number(),memoryTotalGb:F.number()}),ui=F.object({iso:F.string(),unix:F.number(),timezone:F.string(),formatted:F.string()}),di=F.object({passed:F.boolean(),errorCount:F.number(),warningCount:F.number(),topErrors:F.array(F.string())}),fi=F.object({passed:F.boolean(),tsc:di,biome:di}),pi=F.object({name:F.string(),definedIn:F.object({path:F.string(),line:F.number(),kind:F.string(),signature:F.string().optional()}).nullable(),importedBy:F.array(F.object({path:F.string(),line:F.number(),importStatement:F.string()})),referencedIn:F.array(F.object({path:F.string(),line:F.number(),context:F.string(),scope:F.string().optional()})),graphContext:F.object({definingModule:F.string().optional(),importedByModules:F.array(F.string()),siblingSymbols:F.array(F.string())}).nullable()}),mi=F.object({sourcePath:F.string(),contentType:F.string(),score:F.number(),headingPath:F.string().optional(),startLine:F.number().optional(),endLine:F.number().optional(),origin:F.string().optional(),category:F.string().optional(),tags:F.array(F.string()).optional()}),hi=F.object({results:F.array(mi),totalResults:F.number(),searchMode:F.string(),query:F.string()}),gi=F.object({path:F.string(),line:F.number().optional(),matchType:F.string(),preview:F.string()}),_i=F.object({matches:F.array(gi),totalMatches:F.number(),pattern:F.string(),truncated:F.boolean()}),vi=F.object({path:F.string(),relevance:F.number(),estimatedTokens:F.number(),focusLines:F.array(F.string()).optional()}),yi=F.object({files:F.array(vi),totalFiles:F.number(),totalEstimatedTokens:F.number(),task:F.string()}),bi=F.object({path:F.string(),impact:F.string(),reason:F.string()}),xi=F.object({changedFiles:F.array(F.string()),affectedFiles:F.array(bi),totalAffected:F.number(),riskLevel:F.string()}),Si=F.object({name:F.string(),passed:F.boolean(),message:F.string().optional(),severity:F.string().optional()}),Ci=F.object({passed:F.boolean(),score:F.number(),checks:F.array(Si),summary:F.string()}),wi=F.object({id:F.string(),name:F.string(),type:F.string(),sourcePath:F.string().optional()}),Ti=F.object({fromId:F.string(),toId:F.string(),type:F.string()}),Ei=F.object({nodes:F.array(wi),edges:F.array(Ti),totalNodes:F.number(),totalEdges:F.number(),query:F.string()}),Di=F.object({symbols:F.array(F.object({name:F.string(),path:F.string(),line:F.number().optional(),kind:F.string()})),totalDead:F.number()}),Oi=F.object({files:F.number(),packages:F.number(),languages:F.record(F.string(),F.number()),tree:F.string()}),ki=F.object({path:F.string(),language:F.string(),lines:F.number(),imports:F.number(),exports:F.number(),functions:F.number(),classes:F.number()}),Ai=F.object({gitRoot:F.string(),branch:F.string(),commitCount:F.number(),hasUncommitted:F.boolean(),recentCommits:F.array(F.object({hash:F.string(),message:F.string(),author:F.string(),date:F.string()}))}),ji=F.object({originalChars:F.number(),compressedChars:F.number(),ratio:F.number(),segmentsKept:F.number(),segmentsTotal:F.number()}),L=E(`tools`),Mi=F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`);function Ni(e,t){return t?Xt(e,t):e}function Pi(){let e=[];return re.get()||e.push(`Tree-sitter unavailable — using regex fallback, symbol/pattern confidence reduced`),e.length===0?``:`\n\n> **⚠ Caveats:** ${e.join(`; `)}`}function Fi(e){return(e??[]).map(e=>{if(typeof e==`string`)return e;if(e&&typeof e==`object`&&`path`in e)return typeof e.path==`string`?e.path:void 0}).filter(e=>!!e)}function Ii(e){let t=[],n=e.filter(e=>/\.(ts|tsx|js|jsx)$/.test(e)&&/(service|store|model|schema|migration)/i.test(e)),r=e.filter(e=>/\.(ts|tsx|js|jsx)$/.test(e)&&!n.includes(e)),i=e.filter(e=>!/\.(ts|tsx|js|jsx)$/.test(e));return(n.length>0||r.length>0||i.length>0)&&(t.push(`
16
+ `)}}]}))}const yr=E(`auto-gc`);let br=`warming`,xr=null,Sr=0;function Cr(){let e=[...Nr()];if(e.length<10)return 0;e.sort((e,t)=>e-t);let t=Math.floor(e.length*.95);return e[Math.min(t,e.length-1)]}function wr(e){let t=Cr();if(t===0){br=`warming`;return}if(t>500){br=`degraded`;let n=Date.now();if(xr&&n-xr<36e5){yr.debug(`GC cooldown active — skipping`,{lastGcAgoMs:n-xr});return}xr=n,Sr++,yr.warn(`p95 latency exceeds threshold, triggering GC`,{p95:t,cycle:Sr}),e&&e().catch(e=>yr.error(`GC callback failed`,{err:String(e)}))}else t<200&&(br=`healthy`)}function Tr(){return{state:br,p95:Cr(),lastGcAt:xr,gcCount:Sr,bufferSize:Nr().length}}const Er=/key|token|secret|auth|password|bearer/i,Dr=new Set([`eval`,`env`]);function Or(e,t){if(Dr.has(e))return JSON.stringify({_redacted:!0,tool:e});if(e===`http`&&t.headers&&typeof t.headers==`object`){let e={...t},n={};for(let[e,r]of Object.entries(t.headers))n[e]=Er.test(e)?`[REDACTED]`:r;return e.headers=n,JSON.stringify(e).slice(0,200)}return JSON.stringify(t).slice(0,200)}const kr=new Map,Ar=[];function jr(e,t,n,r,i){let a=kr.get(e);a||(a={callCount:0,totalDurationMs:0,totalInputChars:0,totalOutputChars:0,errorCount:0},kr.set(e,a)),a.callCount++,a.totalDurationMs+=t,a.totalInputChars+=n,a.totalOutputChars+=r,i&&a.errorCount++,Ar.push(t),Ar.length>100&&Ar.shift(),Ar.length>=10&&Ar.length%20==0&&wr()}function Mr(){return Array.from(kr.entries()).map(([e,t])=>({tool:e,...t}))}function Nr(){return Ar}function Pr(){return async(e,t)=>{let n=Date.now(),r=JSON.stringify(e.args??{}).length;try{let i=await t(),a=Date.now()-n,o=JSON.stringify(i);return Ot({ts:new Date().toISOString(),source:`mcp`,tool:e.toolName,input:Or(e.toolName,e.args),durationMs:a,status:`ok`,output:o.slice(0,200)}),jr(e.toolName,a,r,o.length,!1),i}catch(t){let i=Date.now()-n;throw Ot({ts:new Date().toISOString(),source:`mcp`,tool:e.toolName,input:Or(e.toolName,e.args),durationMs:i,status:`error`,output:t instanceof Error?t.message:String(t)}),jr(e.toolName,i,r,0,!0),t}}}const Fr=E(`resource-notifier`);var Ir=class{constructor(e){this.mcpServer=e}async notifyStatusChanged(){await this.sendUpdate(`aikit://status`)}async notifyFileTreeChanged(){await this.sendUpdate(`aikit://file-tree`)}async notifyCuratedIndexChanged(){await this.sendUpdate(`aikit://curated`)}async notifyCuratedEntryChanged(e){await this.sendUpdate(`aikit://curated/${e}`)}async notifyResourceListChanged(){try{await this.mcpServer.server.sendResourceListChanged()}catch(e){Fr.debug(`sendResourceListChanged failed`,{error:String(e)})}}async notifyAfterReindex(){await Promise.allSettled([this.notifyStatusChanged(),this.notifyFileTreeChanged()])}async notifyAfterCuratedWrite(e){let t=[this.notifyStatusChanged(),this.notifyCuratedIndexChanged()];e&&t.push(this.notifyCuratedEntryChanged(e)),await Promise.allSettled(t)}async sendUpdate(e){try{await this.mcpServer.server.sendResourceUpdated({uri:e})}catch(t){Fr.debug(`sendResourceUpdated failed`,{uri:e,error:String(t)})}}};function Lr(e,t){e.resource(`aikit-curated-index`,`aikit://curated`,{description:`Index of all curated knowledge entries`,mimeType:`text/markdown`},async()=>{let e=(await t.list()).map(e=>`- [${e.title}](aikit://curated/${e.path}) — ${e.category}`);return{contents:[{uri:`aikit://curated`,text:`# Curated Knowledge Index\n\n${e.length>0?e.join(`
17
+ `):`_No curated entries yet._`}`,mimeType:`text/markdown`}]}});let n=new rn(`aikit://curated/{+path}`,{list:async()=>({resources:(await t.list()).map(e=>({uri:`aikit://curated/${e.path}`,name:e.title,description:`[${e.category}] ${e.contentPreview?.slice(0,80)??``}`,mimeType:`text/markdown`}))})});e.resource(`aikit-curated-entry`,n,{description:`A curated knowledge entry`,mimeType:`text/markdown`},async(e,n)=>{let r=n.path;if(!r)throw Error(`Missing path variable in curated resource URI`);let i=await t.read(r);return{contents:[{uri:e.toString(),text:`---\ntitle: ${i.title}\ncategory: ${i.category}\ntags: ${i.tags?.join(`, `)??``}\nversion: ${i.version??1}\n---\n\n${i.content??i.contentPreview??``}`,mimeType:`text/markdown`}]}})}function Rr(e,t,n){e.resource(`aikit-status`,`aikit://status`,{description:`Current AI Kit status and statistics`,mimeType:`text/plain`},async()=>{let e=await t.getStats();return{contents:[{uri:`aikit://status`,text:`AI Kit: ${e.totalRecords} records from ${e.totalFiles} files. Last indexed: ${e.lastIndexedAt??`Never`}`,mimeType:`text/plain`}]}}),e.resource(`aikit-file-tree`,`aikit://file-tree`,{description:`List of all indexed source files`,mimeType:`text/plain`},async()=>({contents:[{uri:`aikit://file-tree`,text:(await t.listSourcePaths()).sort().join(`
18
+ `),mimeType:`text/plain`}]})),Lr(e,n)}const zr=E(`sampling`);function Br(e){if(!e.modelPreferences)return;let t={...e.modelPreferences.costPriority===void 0?{}:{costPriority:e.modelPreferences.costPriority},...e.modelPreferences.speedPriority===void 0?{}:{speedPriority:e.modelPreferences.speedPriority},...e.modelPreferences.intelligencePriority===void 0?{}:{intelligencePriority:e.modelPreferences.intelligencePriority}};return Object.keys(t).length>0?t:void 0}function Vr(e){if(Array.isArray(e))return e.map(e=>Vr(e)).filter(e=>e.length>0).join(`
19
+ `);if(e&&typeof e==`object`&&`type`in e){let t=e;if(t.type===`text`&&typeof t.text==`string`)return t.text}return JSON.stringify(e)}function Hr(e){let t=e.server,n=typeof t?.createMessage==`function`;return{get available(){return n},async createMessage(e){if(!n)throw Error(`Sampling not available: client does not support createMessage`);let r=e.context?`${e.context}\n\n---\n\n${e.prompt}`:e.prompt;try{let n=await t.createMessage({messages:[{role:`user`,content:{type:`text`,text:r}}],systemPrompt:e.systemPrompt,modelPreferences:Br(e),maxTokens:e.maxTokens??4e3});return{text:Vr(n.content),model:n.model,stopReason:n.stopReason}}catch(e){throw zr.warn(`Sampling createMessage failed`,{error:String(e)}),e}}}}function Ur(){let e=(async(e,t)=>{let n=await t(),r=e.toolConfig??{};return r.outputSchema==null||n.structuredContent!=null||(n.structuredContent=Wr(r.outputSchema)),n});return e.critical=!0,e}function Wr(e){if(!e)return{};if(e.anyOf){let t=e.anyOf.find(e=>e.type!==`null`);return t?Wr(t):null}switch(e.type){case`object`:{let t={},n=e.properties??e.shape;if(n)for(let[e,r]of Object.entries(n))t[e]=Wr(r);return t}case`array`:return[];case`string`:return``;case`number`:case`integer`:return 0;case`boolean`:return!1;case`nullable`:return null;case`optional`:return;case`record`:return{};default:return{}}}const Gr=new Set([`status`,`config`,`guide`,`health`]),Kr={full:{description:`All tools enabled (default)`,includeCategories:[]},safe:{description:`Read-only tools — no file/state modifications`,includeCategories:[`search`,`analysis`,`compression`,`utilities`,`system`,`git`,`flow`],excludeTools:[`reindex`,`onboard`]},research:{description:`Search, analysis, knowledge, and web access`,includeCategories:[`search`,`analysis`,`knowledge`,`compression`,`web`,`system`,`flow`]},minimal:{description:`Essential tools only — search, status, basic operations`,includeCategories:[`search`,`system`],includeTools:[`compact`,`file_summary`,`check`,`test_run`]},discovery:{description:`Full toolset plus discovery-oriented meta-tools for guided tool exploration`,includeCategories:[],includeTools:[`list_tools`,`describe_tool`,`search_tools`]}};function qr(e,t){let n=Kr[e];if(n)return n;let r=t?.[e];if(r)return r;throw Error(`Unknown tool profile: ${e}`)}function Jr(e,t,n){return t.includes(`*`)?!0:(n[e]?.category??[]).some(e=>t.includes(e))}function Yr(e,t,n,r){if(t.includes(`*`)){for(let t of n)e.add(t);return}for(let i of n)Jr(i,t,r)&&e.add(i)}function Xr(e,t,n){if(t.includes(`*`)){e.clear();return}for(let r of[...e])Jr(r,t,n)&&e.delete(r)}function Zr(e,t,n){if(!e||e.length===0||e.includes(`*`))return new Set(t);let r=new Set;return Yr(r,e,t,n),r}function Qr(e,t,n,r){let i=(e,a,o)=>{if(a>10)throw Error(`Tool profile inheritance exceeded max depth 10: ${[...o,e].join(` -> `)}`);if(o.includes(e))throw Error(`Tool profile inheritance cycle detected: ${[...o,e].join(` -> `)}`);let s=qr(e,r),c=[...o,e],l=s.extends?i(s.extends,a+1,c):Zr(s.includeCategories,t,n);s.extends&&s.includeCategories?.length&&Yr(l,s.includeCategories,t,n),s.excludeCategories?.length&&Xr(l,s.excludeCategories,n);for(let e of s.includeTools??[])l.add(e);for(let e of s.excludeTools??[])l.delete(e);return l},a=i(e,0,[]);for(let e of Gr)a.add(e);return a}const $r=E(`tool-timeout`),ei=new Set([`onboard`,`reindex`,`produce_knowledge`,`analyze_structure`,`analyze_dependencies`,`analyze_symbols`,`analyze_patterns`,`analyze_entry_points`,`analyze_diagram`,`codemod`,`batch`,`audit`]);var ti=class extends Error{constructor(e,t){super(`Tool "${e}" timed out after ${t}ms`),this.toolName=e,this.timeoutMs=t,this.name=`ToolTimeoutError`}};function ni(e){return ei.has(e)?6e5:12e4}function ri(e,t,n){return new Promise((r,i)=>{let a=!1,o=setTimeout(()=>{if(!a){a=!0;let e=new ti(n,t);$r.warn(e.message),i(e)}},t);o.unref&&o.unref(),e().then(e=>{a||(a=!0,clearTimeout(o),r(e))},e=>{a||(a=!0,clearTimeout(o),i(e))})})}const ii=F.object({totalRecords:F.number(),totalFiles:F.number(),lastIndexedAt:F.string().nullable(),onboarded:F.boolean(),onboardDir:F.string(),contentTypes:F.record(F.string(),F.number()),wasmAvailable:F.boolean(),graphStats:F.object({nodes:F.number(),edges:F.number()}).nullable(),curatedCount:F.number(),serverVersion:F.string(),scaffoldVersion:F.string().nullable(),workspaceScaffoldVersion:F.string().nullable(),upgradeAvailable:F.boolean(),contextPressure:F.number().min(0).max(100).describe(`0–100 score indicating knowledge base saturation`)}),ai=F.object({entries:F.array(F.object({path:F.string(),title:F.string(),category:F.string(),tags:F.array(F.string()),version:F.number(),preview:F.string()})),totalCount:F.number()}),oi=F.object({ok:F.boolean(),checks:F.array(F.object({name:F.string(),ok:F.boolean(),message:F.string().optional()}))}),si=F.object({summary:F.object({totalFiles:F.number(),totalLines:F.number(),totalCodeLines:F.number(),totalFunctions:F.number(),avgComplexity:F.number(),maxComplexity:F.object({value:F.number(),file:F.string()})}),files:F.array(F.object({path:F.string(),lines:F.number(),code:F.number(),complexity:F.number(),functions:F.number()}))}),ci=F.object({platform:F.string(),arch:F.string(),nodeVersion:F.string(),cwd:F.string(),cpus:F.number(),memoryFreeGb:F.number(),memoryTotalGb:F.number()}),li=F.object({iso:F.string(),unix:F.number(),timezone:F.string(),formatted:F.string()}),ui=F.object({passed:F.boolean(),errorCount:F.number(),warningCount:F.number(),topErrors:F.array(F.string())}),di=F.object({passed:F.boolean(),tsc:ui,biome:ui}),fi=F.object({name:F.string(),definedIn:F.object({path:F.string(),line:F.number(),kind:F.string(),signature:F.string().optional()}).nullable(),importedBy:F.array(F.object({path:F.string(),line:F.number(),importStatement:F.string()})),referencedIn:F.array(F.object({path:F.string(),line:F.number(),context:F.string(),scope:F.string().optional()})),graphContext:F.object({definingModule:F.string().optional(),importedByModules:F.array(F.string()),siblingSymbols:F.array(F.string())}).nullable()}),pi=F.object({sourcePath:F.string(),contentType:F.string(),score:F.number(),headingPath:F.string().optional(),startLine:F.number().optional(),endLine:F.number().optional(),origin:F.string().optional(),category:F.string().optional(),tags:F.array(F.string()).optional()}),mi=F.object({results:F.array(pi),totalResults:F.number(),searchMode:F.string(),query:F.string()}),hi=F.object({path:F.string(),line:F.number().optional(),matchType:F.string(),preview:F.string()}),gi=F.object({matches:F.array(hi),totalMatches:F.number(),pattern:F.string(),truncated:F.boolean()}),_i=F.object({path:F.string(),relevance:F.number(),estimatedTokens:F.number(),focusLines:F.array(F.string()).optional()}),vi=F.object({files:F.array(_i),totalFiles:F.number(),totalEstimatedTokens:F.number(),task:F.string()}),yi=F.object({path:F.string(),impact:F.string(),reason:F.string()}),bi=F.object({changedFiles:F.array(F.string()),affectedFiles:F.array(yi),totalAffected:F.number(),riskLevel:F.string()}),xi=F.object({name:F.string(),passed:F.boolean(),message:F.string().optional(),severity:F.string().optional()}),Si=F.object({passed:F.boolean(),score:F.number(),checks:F.array(xi),summary:F.string()}),Ci=F.object({id:F.string(),name:F.string(),type:F.string(),sourcePath:F.string().optional()}),wi=F.object({fromId:F.string(),toId:F.string(),type:F.string()}),Ti=F.object({nodes:F.array(Ci),edges:F.array(wi),totalNodes:F.number(),totalEdges:F.number(),query:F.string()}),Ei=F.object({symbols:F.array(F.object({name:F.string(),path:F.string(),line:F.number().optional(),kind:F.string()})),totalDead:F.number()}),Di=F.object({files:F.number(),packages:F.number(),languages:F.record(F.string(),F.number()),tree:F.string()}),Oi=F.object({path:F.string(),language:F.string(),lines:F.number(),imports:F.number(),exports:F.number(),functions:F.number(),classes:F.number()}),ki=F.object({gitRoot:F.string(),branch:F.string(),commitCount:F.number(),hasUncommitted:F.boolean(),recentCommits:F.array(F.object({hash:F.string(),message:F.string(),author:F.string(),date:F.string()}))}),Ai=F.object({originalChars:F.number(),compressedChars:F.number(),ratio:F.number(),segmentsKept:F.number(),segmentsTotal:F.number()}),R=E(`tools`),ji=F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`);function Mi(e,t){return t?Xt(e,t):e}function Ni(){let e=[];return re.get()||e.push(`Tree-sitter unavailable — using regex fallback, symbol/pattern confidence reduced`),e.length===0?``:`\n\n> **⚠ Caveats:** ${e.join(`; `)}`}function Pi(e){return(e??[]).map(e=>{if(typeof e==`string`)return e;if(e&&typeof e==`object`&&`path`in e)return typeof e.path==`string`?e.path:void 0}).filter(e=>!!e)}function Fi(e){let t=[],n=e.filter(e=>/\.(ts|tsx|js|jsx)$/.test(e)&&/(service|store|model|schema|migration)/i.test(e)),r=e.filter(e=>/\.(ts|tsx|js|jsx)$/.test(e)&&!n.includes(e)),i=e.filter(e=>!/\.(ts|tsx|js|jsx)$/.test(e));return(n.length>0||r.length>0||i.length>0)&&(t.push(`
20
20
 
21
21
  ### Risk Assessment`),n.length>0&&t.push(`- 🔴 **High risk** (${n.length}): ${n.slice(0,5).map(e=>`\`${e}\``).join(`, `)}`),r.length>0&&t.push(`- 🟡 **Medium risk** (${r.length}): source files`),i.length>0&&t.push(`- 🟢 **Low risk** (${i.length}): non-source files`)),t.join(`
22
- `)}function Li(e){let t=e.replace(/\\/g,`/`);return/(^|\/)__tests__\/|\.(test|spec)\.[jt]sx?$/i.test(t)?`Tests`:/(controller|route|handler|api)/i.test(t)?`API`:/(service|store|model|schema|migration|repo|repository|db|database)/i.test(t)?`Core/Data`:/\.(ts|tsx|js|jsx)$/.test(t)?`Source`:`Config/Docs`}function Ri(e){if(e.length===0)return``;let t=[`Core/Data`,`API`,`Source`,`Tests`,`Config/Docs`],n=new Map;for(let t of e){let e=Li(t),r=n.get(e)??[];r.push(t),n.set(e,r)}let r=[`
22
+ `)}function Ii(e){let t=e.replace(/\\/g,`/`);return/(^|\/)__tests__\/|\.(test|spec)\.[jt]sx?$/i.test(t)?`Tests`:/(controller|route|handler|api)/i.test(t)?`API`:/(service|store|model|schema|migration|repo|repository|db|database)/i.test(t)?`Core/Data`:/\.(ts|tsx|js|jsx)$/.test(t)?`Source`:`Config/Docs`}function Li(e){if(e.length===0)return``;let t=[`Core/Data`,`API`,`Source`,`Tests`,`Config/Docs`],n=new Map;for(let t of e){let e=Ii(t),r=n.get(e)??[];r.push(t),n.set(e,r)}let r=[`
23
23
 
24
24
  ### Layer Classification`];for(let e of t){let t=n.get(e);if(!t?.length)continue;let i=t.slice(0,5).map(e=>`\`${e}\``).join(`, `),a=t.length>5?`, ... and ${t.length-5} more`:``;r.push(`- **${e}** (${t.length}): ${i}${a}`)}return r.join(`
25
- `)}async function zi(e,t,n,r,i){try{let a=`produced/analysis/${n}/${j(`sha256`).update(r).digest(`hex`).slice(0,12)}.md`,o=j(`sha256`).update(i).digest(`hex`).slice(0,16),s=new Date().toISOString(),c=i.length>2e3?i.split(/(?=^## )/m).filter(e=>e.trim().length>0):[i],l=c.map((e,t)=>({id:j(`sha256`).update(`${a}::${t}`).digest(`hex`).slice(0,16),content:e.trim(),sourcePath:a,contentType:`produced-knowledge`,chunkIndex:t,totalChunks:c.length,startLine:0,endLine:0,fileHash:o,indexedAt:s,origin:`produced`,tags:[`analysis`,n],category:`analysis`,version:1})),u=await t.embedBatch(l.map(e=>e.content));await e.upsert(l,u),L.info(`Auto-persisted analysis`,{analyzerName:n,chunkCount:l.length})}catch(e){L.warn(`Auto-persist analysis failed`,{analyzerName:n,...A(e)})}}function Bi(e,t,n){let r=new mn,i=I(`analyze_structure`);e.registerTool(`analyze_structure`,{title:i.title,description:`Analyze the file/directory structure of a codebase. Returns an annotated tree with language stats.`,outputSchema:Oi,inputSchema:{path:F.string().describe(`Root path to analyze`),max_depth:F.number().min(1).max(10).default(6).describe(`Maximum directory depth`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:Mi},annotations:i.annotations},async({path:e,max_depth:i,format:a,max_tokens:o})=>{try{let s=await r.analyze(e,{format:a,maxDepth:i,maxTokens:o});zi(t,n,`structure`,e,s.output);let c=s.data;return{content:[{type:`text`,text:Ni(s.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` for import graphs, or `analyze_patterns` to detect architecture patterns._",o)}],structuredContent:{files:c.fileCount??s.meta.fileCount,packages:c.packageCount??0,languages:c.languages??{},tree:s.output.slice(0,2e3)}}}catch(e){return L.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Vi(e,t,n){let r=new ln,i=I(`analyze_dependencies`);e.registerTool(`analyze_dependencies`,{title:i.title,description:`Analyze import/require dependencies across a codebase. Shows external packages and internal module graph.`,inputSchema:{path:F.string().describe(`Root path to analyze`),format:F.enum([`json`,`markdown`,`mermaid`]).default(`markdown`).describe(`Output format`),max_tokens:Mi},annotations:i.annotations},async({path:e,format:i,max_tokens:a})=>{try{let o=await r.analyze(e,{format:i});return zi(t,n,`dependencies`,e,o.output),{content:[{type:`text`,text:Ni(o.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_symbols` to explore exported symbols, or `analyze_diagram` for visual representation._",a)}]}}catch(e){return L.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Hi(e,t,n){let r=new hn,i=I(`analyze_symbols`);e.registerTool(`analyze_symbols`,{title:i.title,description:`Extract exported and local symbols (functions, classes, interfaces, types, constants) from a codebase.`,inputSchema:{path:F.string().describe(`Root path to analyze`),filter:F.string().optional().describe(`Filter symbols by name substring`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`)},annotations:i.annotations},async({path:e,filter:i,format:a})=>{try{let o=await r.analyze(e,{format:a,filter:i});return zi(t,n,`symbols`,e,o.output),{content:[{type:`text`,text:o.output+Pi()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see import relationships, or `search` to find usage patterns._"}]}}catch(e){return L.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ui(e,t,n){let r=new pn,i=I(`analyze_patterns`);e.registerTool(`analyze_patterns`,{title:i.title,description:`Detect architectural patterns, frameworks, and conventions in a codebase using directory structure and code heuristics.`,inputSchema:{path:F.string().describe(`Root path to analyze`)},annotations:i.annotations},async({path:e})=>{try{let i=await r.analyze(e);return zi(t,n,`patterns`,e,i.output),{content:[{type:`text`,text:i.output+Pi()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_entry_points` to find Lambda handlers and main exports, or `produce_knowledge` for full analysis._"}]}}catch(e){return L.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wi(e,t,n){let r=new dn,i=I(`analyze_entry_points`);e.registerTool(`analyze_entry_points`,{title:i.title,description:`Find entry points: Lambda handlers, main exports, CLI bins, and server start scripts.`,inputSchema:{path:F.string().describe(`Root path to analyze`)},annotations:i.annotations},async({path:e})=>{try{let i=await r.analyze(e);return zi(t,n,`entry-points`,e,i.output),{content:[{type:`text`,text:i.output+Pi()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see what each entry point imports, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return L.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Gi(e,t,n){let r=new un,i=I(`analyze_diagram`);e.registerTool(`analyze_diagram`,{title:i.title,description:`Generate a Mermaid diagram of the codebase architecture or dependency graph. Returns Mermaid flowchart syntax — paste into any Markdown renderer or Mermaid editor to visualize.`,inputSchema:{path:F.string().describe(`Root path to analyze`),diagram_type:F.enum([`architecture`,`dependencies`]).default(`architecture`).describe(`Type of diagram`)},annotations:i.annotations},async({path:e,diagram_type:i})=>{try{let a=await r.analyze(e,{diagramType:i});return zi(t,n,`diagram`,e,a.output),{content:[{type:`text`,text:a.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_structure` for detailed file tree, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return L.error(`Diagram generation failed`,A(e)),{content:[{type:`text`,text:`Diagram generation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ki(e,t,n,r){let i=new cn,a=I(`blast_radius`);e.registerTool(`blast_radius`,{title:a.title,description:`Given a list of changed files, trace the dependency graph to find all affected files (direct + transitive importers) and their tests. Useful for scoping code reviews and impact analysis.`,outputSchema:xi,inputSchema:{path:F.string().describe(`Root path of the codebase`),files:F.array(F.string()).optional().describe(`Changed file paths (relative to root). If omitted, auto-detects from git status.`),max_depth:F.number().min(1).max(20).default(5).describe(`Maximum transitive dependency depth`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:Mi},annotations:a.annotations},async({path:e,files:a,max_depth:o,format:s,max_tokens:c})=>{try{let l=a??[];if(l.length===0)try{let t=await Je({cwd:e,includeDiff:!1});if(l=Array.from(new Set([...Fi(t.status?.staged),...Fi(t.status?.modified),...Fi(t.status?.untracked)])),l.length===0)return{content:[{type:`text`,text:"No changed files detected from git status. Provide `files` explicitly or make changes first."}]}}catch{return{content:[{type:`text`,text:"Could not detect changed files from git. Provide `files` explicitly."}],isError:!0}}let u=await i.analyze(e,{files:l,maxDepth:o,format:s}),d=``;if(r)try{let e=new Set;for(let t of l){let n=await r.findNodes({sourcePath:t,limit:10});for(let t of n){let n=await r.getNeighbors(t.id,{direction:`incoming`,edgeType:`imports`});for(let t of n.nodes){let n=t.sourcePath??t.name;l.includes(n)||e.add(n)}}}e.size>0&&(d=`\n\n### Graph-discovered importers (${e.size})\n`+[...e].slice(0,20).map(e=>`- \`${e}\``).join(`
26
- `),e.size>20&&(d+=`\n- ... and ${e.size-20} more`))}catch{}let f=Ri(l),p=Ii(l),m=u.output+d+f+p+Pi()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see the full import graph, or `analyze_symbols` to inspect affected exports._";zi(t,n,`blast-radius`,e,m);let h=new Set;if(r)for(let e of l)try{let t=await r.findNodes({sourcePath:e,limit:10});for(let e of t){let t=await r.getNeighbors(e.id,{direction:`incoming`,edgeType:`imports`});for(let e of t.nodes){let t=e.sourcePath??e.name;l.includes(t)||h.add(t)}}}catch{}let g=[...h].map(e=>({path:e,impact:`transitive`,reason:`imports changed file`})),_=l.length>10?`high`:l.length>3?`medium`:`low`;return{content:[{type:`text`,text:Ni(m,c)}],structuredContent:{changedFiles:l,affectedFiles:g,totalAffected:g.length,riskLevel:_}}}catch(e){return L.error(`Blast radius analysis failed`,A(e)),{content:[{type:`text`,text:`Blast radius analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const qi=E(`task-manager`);let Ji=0;function Yi(e){let t=typeof e?.sendNotification==`function`;return{createTask(n,r){let i=`aikit-task-${++Ji}`,a=(n,r,a)=>{t&&e.sendNotification?.({method:`notifications/progress`,params:{progressToken:i,progress:n,total:r,...a?{message:a}:{}}}).catch(e=>{qi.debug(`Progress notification failed`,{token:i,error:String(e)})})};return t&&qi.debug(`Task created`,{token:i,title:n,totalSteps:r}),{progress(e,t){a(e,r,t??`${n}: step ${e+1}/${r}`)},complete(e){a(r,r,e??`${n}: complete`)},fail(e){a(r,r,e??`${n}: failed`)}}}}}const Xi=E(`tools`);function Zi(e,t,n){let r=I(`audit`);e.registerTool(`audit`,{title:r.title,description:`Run a unified audit: structure, dependencies, patterns, health, dead_symbols, check, entry_points. Returns synthesized report with score, recommendations, and next steps. 6 round-trips → 1.`,outputSchema:Ci,inputSchema:{path:F.string().default(`.`).describe(`Root path to audit`),checks:F.array(F.enum([`structure`,`dependencies`,`patterns`,`health`,`dead_symbols`,`check`,`entry_points`])).optional().describe(`Which checks to run (default: all)`),detail:F.enum([`summary`,`full`]).default(`summary`).describe(`'summary' for overview (~500 tokens), 'full' includes pattern table`)},annotations:r.annotations},async({path:e,checks:r,detail:i},a)=>{try{let o=await ve(t,n,{path:e,checks:r,detail:i});return Yi(a).createTask(`Audit`,1).complete(`Audit complete: ${o.data?.recommendations?.length??0} recommendations`),{content:[{type:`text`,text:o.ok?`${o.summary}\n\n---\n_Audit score: ${o.data?.score}/100 | ${o.meta.durationMs}ms${o.next&&o.next.length>0?` | Next: ${o.next.map(e=>`\`${e.tool}\` (${e.reason})`).join(`, `)}`:``}_`:o.error?.message??`Audit failed`}],isError:!o.ok,...o.ok&&o.data?{structuredContent:{passed:o.ok,score:o.data.score??0,checks:(o.data.recommendations??[]).map(e=>({name:e.area,passed:e.priority!==`high`,message:e.message,severity:e.priority})),summary:o.summary??``}}:{}}}catch(e){return Xi.error(`Audit failed`,A(e)),{content:[{type:`text`,text:`Audit failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Qi=E(`tools`);function $i(e,t){let n=I(`brainstorm`);e.registerTool(`brainstorm`,{title:n.title,description:`Interactive brainstorming tool for exploring ideas, making design decisions, and gathering requirements. Supports multiple actions: ask_choice (single selection), ask_multi (multiple selections), ask_freeform (text input), ask_confirm (yes/no), present_design (show summary and confirm). Falls back to text-only mode when elicitation is unavailable.`,inputSchema:{action:F.enum([`ask_choice`,`ask_multi`,`ask_freeform`,`ask_confirm`,`present_design`]).describe(`The type of brainstorm interaction`),message:F.string().describe(`The question or prompt to present`),options:F.array(F.string()).optional().describe(`Options for ask_choice/ask_multi actions`),context:F.string().optional().describe(`Additional context to include in the prompt`)},annotations:n.annotations},async({action:e,message:n,options:r,context:i})=>{let a=i?`${i}\n\n${n}`:n;if(!t.available)return ea(e,a,r);try{switch(e){case`ask_choice`:{if(!r?.length)return R(`Error: ask_choice requires options array`);let e=await t.selectOne(a,r);return R(e===null?`No selection made (declined or cancelled).`:`Selected: **${e}**`)}case`ask_multi`:{if(!r?.length)return R(`Error: ask_multi requires options array`);let e=await t.selectMany(a,r);return e.length===0?R(`No selections made (declined or cancelled).`):R(`Selected:\n${e.map(e=>`- ${e}`).join(`
27
- `)}`)}case`ask_freeform`:{let e=await t.promptText(a);return R(e===null?`No input provided (declined or cancelled).`:`Response:\n\n${e}`)}case`ask_confirm`:return R(await t.confirm(a)?`Confirmed: **Yes**`:`Confirmed: **No**`);case`present_design`:return R(await t.confirm(`## Design Summary\n\n${a}\n\nDo you approve this design?`)?`## Design Approved ✓
25
+ `)}async function Ri(e,t,n,r,i){try{let a=`produced/analysis/${n}/${j(`sha256`).update(r).digest(`hex`).slice(0,12)}.md`,o=j(`sha256`).update(i).digest(`hex`).slice(0,16),s=new Date().toISOString(),c=i.length>2e3?i.split(/(?=^## )/m).filter(e=>e.trim().length>0):[i],l=c.map((e,t)=>({id:j(`sha256`).update(`${a}::${t}`).digest(`hex`).slice(0,16),content:e.trim(),sourcePath:a,contentType:`produced-knowledge`,chunkIndex:t,totalChunks:c.length,startLine:0,endLine:0,fileHash:o,indexedAt:s,origin:`produced`,tags:[`analysis`,n],category:`analysis`,version:1})),u=await t.embedBatch(l.map(e=>e.content));await e.upsert(l,u),R.info(`Auto-persisted analysis`,{analyzerName:n,chunkCount:l.length})}catch(e){R.warn(`Auto-persist analysis failed`,{analyzerName:n,...A(e)})}}function zi(e,t,n){let r=new mn,i=L(`analyze_structure`);e.registerTool(`analyze_structure`,{title:i.title,description:`Analyze the file/directory structure of a codebase. Returns an annotated tree with language stats.`,outputSchema:Di,inputSchema:{path:F.string().describe(`Root path to analyze`),max_depth:F.number().min(1).max(10).default(6).describe(`Maximum directory depth`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:ji},annotations:i.annotations},async({path:e,max_depth:i,format:a,max_tokens:o})=>{try{let s=await r.analyze(e,{format:a,maxDepth:i,maxTokens:o});Ri(t,n,`structure`,e,s.output);let c=s.data;return{content:[{type:`text`,text:Mi(s.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` for import graphs, or `analyze_patterns` to detect architecture patterns._",o)}],structuredContent:{files:c.fileCount??s.meta.fileCount,packages:c.packageCount??0,languages:c.languages??{},tree:s.output.slice(0,2e3)}}}catch(e){return R.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Bi(e,t,n){let r=new ln,i=L(`analyze_dependencies`);e.registerTool(`analyze_dependencies`,{title:i.title,description:`Analyze import/require dependencies across a codebase. Shows external packages and internal module graph.`,inputSchema:{path:F.string().describe(`Root path to analyze`),format:F.enum([`json`,`markdown`,`mermaid`]).default(`markdown`).describe(`Output format`),max_tokens:ji},annotations:i.annotations},async({path:e,format:i,max_tokens:a})=>{try{let o=await r.analyze(e,{format:i});return Ri(t,n,`dependencies`,e,o.output),{content:[{type:`text`,text:Mi(o.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_symbols` to explore exported symbols, or `analyze_diagram` for visual representation._",a)}]}}catch(e){return R.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Vi(e,t,n){let r=new hn,i=L(`analyze_symbols`);e.registerTool(`analyze_symbols`,{title:i.title,description:`Extract exported and local symbols (functions, classes, interfaces, types, constants) from a codebase.`,inputSchema:{path:F.string().describe(`Root path to analyze`),filter:F.string().optional().describe(`Filter symbols by name substring`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`)},annotations:i.annotations},async({path:e,filter:i,format:a})=>{try{let o=await r.analyze(e,{format:a,filter:i});return Ri(t,n,`symbols`,e,o.output),{content:[{type:`text`,text:o.output+Ni()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see import relationships, or `search` to find usage patterns._"}]}}catch(e){return R.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Hi(e,t,n){let r=new pn,i=L(`analyze_patterns`);e.registerTool(`analyze_patterns`,{title:i.title,description:`Detect architectural patterns, frameworks, and conventions in a codebase using directory structure and code heuristics.`,inputSchema:{path:F.string().describe(`Root path to analyze`)},annotations:i.annotations},async({path:e})=>{try{let i=await r.analyze(e);return Ri(t,n,`patterns`,e,i.output),{content:[{type:`text`,text:i.output+Ni()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_entry_points` to find Lambda handlers and main exports, or `produce_knowledge` for full analysis._"}]}}catch(e){return R.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ui(e,t,n){let r=new dn,i=L(`analyze_entry_points`);e.registerTool(`analyze_entry_points`,{title:i.title,description:`Find entry points: Lambda handlers, main exports, CLI bins, and server start scripts.`,inputSchema:{path:F.string().describe(`Root path to analyze`)},annotations:i.annotations},async({path:e})=>{try{let i=await r.analyze(e);return Ri(t,n,`entry-points`,e,i.output),{content:[{type:`text`,text:i.output+Ni()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see what each entry point imports, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return R.error(`Analysis failed`,A(e)),{content:[{type:`text`,text:`Analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wi(e,t,n){let r=new un,i=L(`analyze_diagram`);e.registerTool(`analyze_diagram`,{title:i.title,description:`Generate a Mermaid diagram of the codebase architecture or dependency graph. Returns Mermaid flowchart syntax — paste into any Markdown renderer or Mermaid editor to visualize.`,inputSchema:{path:F.string().describe(`Root path to analyze`),diagram_type:F.enum([`architecture`,`dependencies`]).default(`architecture`).describe(`Type of diagram`)},annotations:i.annotations},async({path:e,diagram_type:i})=>{try{let a=await r.analyze(e,{diagramType:i});return Ri(t,n,`diagram`,e,a.output),{content:[{type:`text`,text:a.output+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_structure` for detailed file tree, or `produce_knowledge` for comprehensive analysis._"}]}}catch(e){return R.error(`Diagram generation failed`,A(e)),{content:[{type:`text`,text:`Diagram generation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Gi(e,t,n,r){let i=new cn,a=L(`blast_radius`);e.registerTool(`blast_radius`,{title:a.title,description:`Given a list of changed files, trace the dependency graph to find all affected files (direct + transitive importers) and their tests. Useful for scoping code reviews and impact analysis.`,outputSchema:bi,inputSchema:{path:F.string().describe(`Root path of the codebase`),files:F.array(F.string()).optional().describe(`Changed file paths (relative to root). If omitted, auto-detects from git status.`),max_depth:F.number().min(1).max(20).default(5).describe(`Maximum transitive dependency depth`),format:F.enum([`json`,`markdown`]).default(`markdown`).describe(`Output format`),max_tokens:ji},annotations:a.annotations},async({path:e,files:a,max_depth:o,format:s,max_tokens:c})=>{try{let l=a??[];if(l.length===0)try{let t=await Je({cwd:e,includeDiff:!1});if(l=Array.from(new Set([...Pi(t.status?.staged),...Pi(t.status?.modified),...Pi(t.status?.untracked)])),l.length===0)return{content:[{type:`text`,text:"No changed files detected from git status. Provide `files` explicitly or make changes first."}]}}catch{return{content:[{type:`text`,text:"Could not detect changed files from git. Provide `files` explicitly."}],isError:!0}}let u=await i.analyze(e,{files:l,maxDepth:o,format:s}),d=``;if(r)try{let e=new Set;for(let t of l){let n=await r.findNodes({sourcePath:t,limit:10});for(let t of n){let n=await r.getNeighbors(t.id,{direction:`incoming`,edgeType:`imports`});for(let t of n.nodes){let n=t.sourcePath??t.name;l.includes(n)||e.add(n)}}}e.size>0&&(d=`\n\n### Graph-discovered importers (${e.size})\n`+[...e].slice(0,20).map(e=>`- \`${e}\``).join(`
26
+ `),e.size>20&&(d+=`\n- ... and ${e.size-20} more`))}catch{}let f=Li(l),p=Fi(l),m=u.output+d+f+p+Ni()+"\n\n---\n_Analysis auto-saved to KB. Next: Use `analyze_dependencies` to see the full import graph, or `analyze_symbols` to inspect affected exports._";Ri(t,n,`blast-radius`,e,m);let h=new Set;if(r)for(let e of l)try{let t=await r.findNodes({sourcePath:e,limit:10});for(let e of t){let t=await r.getNeighbors(e.id,{direction:`incoming`,edgeType:`imports`});for(let e of t.nodes){let t=e.sourcePath??e.name;l.includes(t)||h.add(t)}}}catch{}let g=[...h].map(e=>({path:e,impact:`transitive`,reason:`imports changed file`})),_=l.length>10?`high`:l.length>3?`medium`:`low`;return{content:[{type:`text`,text:Mi(m,c)}],structuredContent:{changedFiles:l,affectedFiles:g,totalAffected:g.length,riskLevel:_}}}catch(e){return R.error(`Blast radius analysis failed`,A(e)),{content:[{type:`text`,text:`Blast radius analysis failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Ki=E(`task-manager`);let qi=0;function Ji(e){let t=typeof e?.sendNotification==`function`;return{createTask(n,r){let i=`aikit-task-${++qi}`,a=(n,r,a)=>{t&&e.sendNotification?.({method:`notifications/progress`,params:{progressToken:i,progress:n,total:r,...a?{message:a}:{}}}).catch(e=>{Ki.debug(`Progress notification failed`,{token:i,error:String(e)})})};return t&&Ki.debug(`Task created`,{token:i,title:n,totalSteps:r}),{progress(e,t){a(e,r,t??`${n}: step ${e+1}/${r}`)},complete(e){a(r,r,e??`${n}: complete`)},fail(e){a(r,r,e??`${n}: failed`)}}}}}const Yi=E(`tools`);function Xi(e,t,n){let r=L(`audit`);e.registerTool(`audit`,{title:r.title,description:`Run a unified audit: structure, dependencies, patterns, health, dead_symbols, check, entry_points. Returns synthesized report with score, recommendations, and next steps. 6 round-trips → 1.`,outputSchema:Si,inputSchema:{path:F.string().default(`.`).describe(`Root path to audit`),checks:F.array(F.enum([`structure`,`dependencies`,`patterns`,`health`,`dead_symbols`,`check`,`entry_points`])).optional().describe(`Which checks to run (default: all)`),detail:F.enum([`summary`,`full`]).default(`summary`).describe(`'summary' for overview (~500 tokens), 'full' includes pattern table`)},annotations:r.annotations},async({path:e,checks:r,detail:i},a)=>{try{let o=await ve(t,n,{path:e,checks:r,detail:i});return Ji(a).createTask(`Audit`,1).complete(`Audit complete: ${o.data?.recommendations?.length??0} recommendations`),{content:[{type:`text`,text:o.ok?`${o.summary}\n\n---\n_Audit score: ${o.data?.score}/100 | ${o.meta.durationMs}ms${o.next&&o.next.length>0?` | Next: ${o.next.map(e=>`\`${e.tool}\` (${e.reason})`).join(`, `)}`:``}_`:o.error?.message??`Audit failed`}],isError:!o.ok,...o.ok&&o.data?{structuredContent:{passed:o.ok,score:o.data.score??0,checks:(o.data.recommendations??[]).map(e=>({name:e.area,passed:e.priority!==`high`,message:e.message,severity:e.priority})),summary:o.summary??``}}:{}}}catch(e){return Yi.error(`Audit failed`,A(e)),{content:[{type:`text`,text:`Audit failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Zi=E(`tools`);function Qi(e,t){let n=L(`brainstorm`);e.registerTool(`brainstorm`,{title:n.title,description:`Interactive brainstorming tool for exploring ideas, making design decisions, and gathering requirements. Supports multiple actions: ask_choice (single selection), ask_multi (multiple selections), ask_freeform (text input), ask_confirm (yes/no), present_design (show summary and confirm). Falls back to text-only mode when elicitation is unavailable.`,inputSchema:{action:F.enum([`ask_choice`,`ask_multi`,`ask_freeform`,`ask_confirm`,`present_design`]).describe(`The type of brainstorm interaction`),message:F.string().describe(`The question or prompt to present`),options:F.array(F.string()).optional().describe(`Options for ask_choice/ask_multi actions`),context:F.string().optional().describe(`Additional context to include in the prompt`)},annotations:n.annotations},async({action:e,message:n,options:r,context:i})=>{let a=i?`${i}\n\n${n}`:n;if(!t.available)return $i(e,a,r);try{switch(e){case`ask_choice`:{if(!r?.length)return z(`Error: ask_choice requires options array`);let e=await t.selectOne(a,r);return z(e===null?`No selection made (declined or cancelled).`:`Selected: **${e}**`)}case`ask_multi`:{if(!r?.length)return z(`Error: ask_multi requires options array`);let e=await t.selectMany(a,r);return e.length===0?z(`No selections made (declined or cancelled).`):z(`Selected:\n${e.map(e=>`- ${e}`).join(`
27
+ `)}`)}case`ask_freeform`:{let e=await t.promptText(a);return z(e===null?`No input provided (declined or cancelled).`:`Response:\n\n${e}`)}case`ask_confirm`:return z(await t.confirm(a)?`Confirmed: **Yes**`:`Confirmed: **No**`);case`present_design`:return z(await t.confirm(`## Design Summary\n\n${a}\n\nDo you approve this design?`)?`## Design Approved ✓
28
28
 
29
29
  Proceeding with the design as described.`:`## Design Not Approved
30
30
 
31
- Please provide feedback for revisions.`);default:return R(`Unknown action: ${e}`)}}catch(t){return Qi.warn(`Brainstorm elicitation failed, falling back to text`,{action:e,error:String(t)}),ea(e,a,r)}})}function R(e){return{content:[{type:`text`,text:e}]}}function ea(e,t,n){let r=`## Brainstorm: ${e}\n\n${t}`;return n?.length&&(r+=`\n\n**Options:**\n${n.map((e,t)=>`${t+1}. ${e}`).join(`
31
+ Please provide feedback for revisions.`);default:return z(`Unknown action: ${e}`)}}catch(t){return Zi.warn(`Brainstorm elicitation failed, falling back to text`,{action:e,error:String(t)}),$i(e,a,r)}})}function z(e){return{content:[{type:`text`,text:e}]}}function $i(e,t,n){let r=`## Brainstorm: ${e}\n\n${t}`;return n?.length&&(r+=`\n\n**Options:**\n${n.map((e,t)=>`${t+1}. ${e}`).join(`
32
32
  `)}`),r+=`
33
33
 
34
- > *Elicitation unavailable — please respond with your choice in the next message.*`,{content:[{type:`text`,text:r}]}}const ta=E(`tools`);function na(e){if(!e?.enabled)return;let t=process.env.AIKIT_ER_API_KEY;if(!t){ta.warn(`ER bridge enabled but AIKIT_ER_API_KEY not set; disabling`);return}return{enabled:!0,baseUrl:e.baseUrl,apiKey:t,timeoutMs:e.timeoutMs??5e3,cacheTtlMs:e.cacheTtlMs??360*60*1e3,cacheMaxEntries:e.cacheMaxEntries??100,fallbackThreshold:e.fallbackThreshold??.45}}function ra(e){let t=na(e);if(t)try{let e=new se(t),n=new oe({maxEntries:t.cacheMaxEntries,defaultTtl:t.cacheTtlMs}),r=new ue(e);try{let e=new URL(t.baseUrl).hostname;ta.info(`ER bridge initialized`,{host:e})}catch{ta.info(`ER bridge initialized`)}return{client:e,cache:n,pushAdapter:r,config:t}}catch(e){ta.warn(`ER bridge initialization failed`,A(e));return}}function ia(e,t,n){let r=I(`er_push`);e.registerTool(`er_push`,{title:r.title,description:`Push a curated knowledge entry to Enterprise RAG. The entry is stored via ER's curated_remember tool and becomes immediately searchable in the enterprise AI Kit.`,inputSchema:{title:F.string().min(3).max(120).describe(`Title for the knowledge entry`),content:F.string().min(10).max(1e5).describe(`Markdown content to push (max 100KB)`),category:F.string().regex(/^[a-z][a-z0-9-]*$/).default(`conventions`).describe(`Category slug (e.g., "decisions", "patterns", "conventions")`),tags:F.array(F.string()).default([]).describe(`Optional tags`),rule_id:F.string().optional().describe(`ID of the classification rule that motivated this push (from remember classification signals). Improves rule effectiveness tracking.`)},annotations:r.annotations},async({title:e,content:r,category:i,tags:a,rule_id:o})=>{try{let s=`${i}/${e.toLowerCase().replace(/[^a-z0-9]+/g,`-`)}`,c=await t.pushAdapter.push(s,{title:e,content:r,category:i,tags:a});return c.pushed?(n&&n.recordPush(s,c.pushed,o),{content:[{type:`text`,text:`Pushed to ER: **${e}**\n\nRemote path: \`${c.remotePath??`unknown`}\`\nTimestamp: ${c.timestamp}\n\n---\n_Next: Use \`er_pull\` to verify the entry is searchable in ER, or \`er_sync_status\` to see push history._`}]}):(n&&n.recordPush(s,!1,o),{content:[{type:`text`,text:`ER push failed: ${c.error??`Unknown error`}\n\n_The local AI Kit is unaffected. You can retry or check ER health with \`er_sync_status\`._`}],isError:!0})}catch(t){if(ta.error(`ER push failed`,A(t)),n){let t=`${i}/${e.toLowerCase().replace(/[^a-z0-9]+/g,`-`)}`;n.recordPush(t,!1,o)}return{content:[{type:`text`,text:`ER push failed: operation unsuccessful
34
+ > *Elicitation unavailable — please respond with your choice in the next message.*`,{content:[{type:`text`,text:r}]}}const ea=E(`tools`);function ta(e){if(!e?.enabled)return;let t=process.env.AIKIT_ER_API_KEY;if(!t){ea.warn(`ER bridge enabled but AIKIT_ER_API_KEY not set; disabling`);return}return{enabled:!0,baseUrl:e.baseUrl,apiKey:t,timeoutMs:e.timeoutMs??5e3,cacheTtlMs:e.cacheTtlMs??360*60*1e3,cacheMaxEntries:e.cacheMaxEntries??100,fallbackThreshold:e.fallbackThreshold??.45}}function na(e){let t=ta(e);if(t)try{let e=new se(t),n=new oe({maxEntries:t.cacheMaxEntries,defaultTtl:t.cacheTtlMs}),r=new ue(e);try{let e=new URL(t.baseUrl).hostname;ea.info(`ER bridge initialized`,{host:e})}catch{ea.info(`ER bridge initialized`)}return{client:e,cache:n,pushAdapter:r,config:t}}catch(e){ea.warn(`ER bridge initialization failed`,A(e));return}}function ra(e,t,n){let r=L(`er_push`);e.registerTool(`er_push`,{title:r.title,description:`Push a curated knowledge entry to Enterprise RAG. The entry is stored via ER's curated_remember tool and becomes immediately searchable in the enterprise AI Kit.`,inputSchema:{title:F.string().min(3).max(120).describe(`Title for the knowledge entry`),content:F.string().min(10).max(1e5).describe(`Markdown content to push (max 100KB)`),category:F.string().regex(/^[a-z][a-z0-9-]*$/).default(`conventions`).describe(`Category slug (e.g., "decisions", "patterns", "conventions")`),tags:F.array(F.string()).default([]).describe(`Optional tags`),rule_id:F.string().optional().describe(`ID of the classification rule that motivated this push (from remember classification signals). Improves rule effectiveness tracking.`)},annotations:r.annotations},async({title:e,content:r,category:i,tags:a,rule_id:o})=>{try{let s=`${i}/${e.toLowerCase().replace(/[^a-z0-9]+/g,`-`)}`,c=await t.pushAdapter.push(s,{title:e,content:r,category:i,tags:a});return c.pushed?(n&&n.recordPush(s,c.pushed,o),{content:[{type:`text`,text:`Pushed to ER: **${e}**\n\nRemote path: \`${c.remotePath??`unknown`}\`\nTimestamp: ${c.timestamp}\n\n---\n_Next: Use \`er_pull\` to verify the entry is searchable in ER, or \`er_sync_status\` to see push history._`}]}):(n&&n.recordPush(s,!1,o),{content:[{type:`text`,text:`ER push failed: ${c.error??`Unknown error`}\n\n_The local AI Kit is unaffected. You can retry or check ER health with \`er_sync_status\`._`}],isError:!0})}catch(t){if(ea.error(`ER push failed`,A(t)),n){let t=`${i}/${e.toLowerCase().replace(/[^a-z0-9]+/g,`-`)}`;n.recordPush(t,!1,o)}return{content:[{type:`text`,text:`ER push failed: operation unsuccessful
35
35
 
36
- _The local AI Kit is unaffected. You can retry or check ER health with \`er_sync_status\`._`}],isError:!0}}})}function aa(e,t){let n=I(`er_pull`);e.registerTool(`er_pull`,{title:n.title,description:`Explicitly search the Enterprise RAG AI Kit. Returns results from the enterprise system only (not local AI Kit). Uses a cache with 6-hour TTL to reduce API calls.`,inputSchema:{query:F.string().min(1).max(2e3).describe(`Search query for Enterprise RAG AI Kit (max 2000 chars)`),max_results:F.number().min(1).max(20).default(5).describe(`Maximum results to return`),bypass_cache:F.boolean().default(!0).describe(`Skip cache and fetch fresh results from ER (default: true per DR-001)`)},annotations:n.annotations},async({query:e,max_results:n,bypass_cache:r})=>{try{if(!r){let r=t.cache.get(e);if(r)return{content:[{type:`text`,text:`${r.slice(0,n).map((e,t)=>`### Result ${t+1} (score: ${e.score.toFixed(3)}, source: ER cached)\n- **Source**: ${e.sourcePath}\n\n${e.content}`).join(`
36
+ _The local AI Kit is unaffected. You can retry or check ER health with \`er_sync_status\`._`}],isError:!0}}})}function ia(e,t){let n=L(`er_pull`);e.registerTool(`er_pull`,{title:n.title,description:`Explicitly search the Enterprise RAG AI Kit. Returns results from the enterprise system only (not local AI Kit). Uses a cache with 6-hour TTL to reduce API calls.`,inputSchema:{query:F.string().min(1).max(2e3).describe(`Search query for Enterprise RAG AI Kit (max 2000 chars)`),max_results:F.number().min(1).max(20).default(5).describe(`Maximum results to return`),bypass_cache:F.boolean().default(!0).describe(`Skip cache and fetch fresh results from ER (default: true per DR-001)`)},annotations:n.annotations},async({query:e,max_results:n,bypass_cache:r})=>{try{if(!r){let r=t.cache.get(e);if(r)return{content:[{type:`text`,text:`${r.slice(0,n).map((e,t)=>`### Result ${t+1} (score: ${e.score.toFixed(3)}, source: ER cached)\n- **Source**: ${e.sourcePath}\n\n${e.content}`).join(`
37
37
 
38
38
  ---
39
39
 
@@ -41,11 +41,11 @@ _The local AI Kit is unaffected. You can retry or check ER health with \`er_sync
41
41
 
42
42
  ---
43
43
 
44
- `)}\n\n---\n_Source: Enterprise RAG (${i.length} results) | Results cached for ${Math.floor(t.config.cacheTtlMs/36e5)}h_`}]}}catch(e){return ta.error(`ER pull failed`,A(e)),{content:[{type:`text`,text:`ER pull failed: search request unsuccessful
44
+ `)}\n\n---\n_Source: Enterprise RAG (${i.length} results) | Results cached for ${Math.floor(t.config.cacheTtlMs/36e5)}h_`}]}}catch(e){return ea.error(`ER pull failed`,A(e)),{content:[{type:`text`,text:`ER pull failed: search request unsuccessful
45
45
 
46
- _Enterprise RAG may be unavailable. Local AI Kit search is unaffected._`}],isError:!0}}})}function oa(e,t){let n=I(`er_sync_status`);e.registerTool(`er_sync_status`,{title:n.title,description:`Show the status of the Enterprise RAG bridge: health, cache stats, and push history.`,inputSchema:{check_health:F.boolean().default(!0).describe(`Whether to ping ER health endpoint (adds latency)`)},annotations:n.annotations},async({check_health:e})=>{try{let n=t.cache.stats(),r=t.pushAdapter.getStatus(),i=new URL(t.config.baseUrl),a=`${i.protocol}//${i.hostname}${i.port?`:${i.port}`:``}`,o=`_Health check skipped_`;if(e){let e=await t.client.health();o=e.healthy?`✅ Healthy (HTTP ${e.status})`:`❌ Unhealthy${e.status?` (HTTP ${e.status})`:``}`}let s=[`## ER Bridge Status
46
+ _Enterprise RAG may be unavailable. Local AI Kit search is unaffected._`}],isError:!0}}})}function aa(e,t){let n=L(`er_sync_status`);e.registerTool(`er_sync_status`,{title:n.title,description:`Show the status of the Enterprise RAG bridge: health, cache stats, and push history.`,inputSchema:{check_health:F.boolean().default(!0).describe(`Whether to ping ER health endpoint (adds latency)`)},annotations:n.annotations},async({check_health:e})=>{try{let n=t.cache.stats(),r=t.pushAdapter.getStatus(),i=new URL(t.config.baseUrl),a=`${i.protocol}//${i.hostname}${i.port?`:${i.port}`:``}`,o=`_Health check skipped_`;if(e){let e=await t.client.health();o=e.healthy?`✅ Healthy (HTTP ${e.status})`:`❌ Unhealthy${e.status?` (HTTP ${e.status})`:``}`}let s=[`## ER Bridge Status
47
47
  `,`**Endpoint**: \`${a}\``,`**Health**: ${o}`,`**Threshold**: ${t.config.fallbackThreshold} (vector similarity for auto-fallback)`,``,`### Cache`,`- Entries: ${n.size} / ${n.maxEntries}`,`- TTL: ${Math.floor(n.defaultTtlMs/36e5)}h`,``,`### Push History`,`- Total pushed: ${r.totalPushed}`,`- Successful: ${r.successCount}`,`- Failed: ${r.failCount}`];return r.lastPush&&s.push(`- Last push: "${r.lastPush.title}" at ${r.lastPush.pushedAt} (${r.lastPush.status})`),s.push("\n---\n_Next: Use `er_push` to send knowledge to ER, or `er_pull` to search ER._"),{content:[{type:`text`,text:s.join(`
48
- `)}]}}catch(e){return ta.error(`ER sync status failed`,A(e)),{content:[{type:`text`,text:`ER sync status failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function z(e){return String(e??``).replace(/&/g,`&amp;`).replace(/</g,`&lt;`).replace(/>/g,`&gt;`).replace(/"/g,`&quot;`)}function sa(e,t){let n=typeof t==`string`?JSON.parse(t):t;switch(e){case`list-sort`:return ca(n);case`data-table`:return la(n);case`picker`:return ua(n);case`flame-graph`:return da(n);case`form`:return fa(n);case`timeline`:return pa(n);case`kanban`:return ma(n);case`tree`:return ha(n);case`diff-view`:return ga(n);case`dashboard`:return _a(n);default:return`<pre>${z(JSON.stringify(n,null,2))}</pre>`}}function ca(e){let t=e.items??[];return`
48
+ `)}]}}catch(e){return ea.error(`ER sync status failed`,A(e)),{content:[{type:`text`,text:`ER sync status failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function B(e){return String(e??``).replace(/&/g,`&amp;`).replace(/</g,`&lt;`).replace(/>/g,`&gt;`).replace(/"/g,`&quot;`)}function oa(e,t){let n=typeof t==`string`?JSON.parse(t):t;switch(e){case`list-sort`:return sa(n);case`data-table`:return ca(n);case`picker`:return la(n);case`flame-graph`:return ua(n);case`form`:return da(n);case`timeline`:return fa(n);case`kanban`:return pa(n);case`tree`:return ma(n);case`diff-view`:return ha(n);case`dashboard`:return ga(n);default:return`<pre>${B(JSON.stringify(n,null,2))}</pre>`}}function sa(e){let t=e.items??[];return`
49
49
  <style>
50
50
  .sort-list{list-style:none;padding:0;margin:0}
51
51
  .sort-item{display:flex;align-items:center;gap:10px;padding:12px 16px;margin:4px 0;
@@ -63,9 +63,9 @@ _Enterprise RAG may be unavailable. Local AI Kit search is unaffected._`}],isErr
63
63
  border-radius:6px;font-size:12px;color:var(--aikit-muted)}
64
64
  </style>
65
65
  <ul class="sort-list" id="sortList">${t.map((e,t)=>`
66
- <li class="sort-item" draggable="true" data-idx="${t}" data-id="${z(e.id)}">
66
+ <li class="sort-item" draggable="true" data-idx="${t}" data-id="${B(e.id)}">
67
67
  <span class="drag-handle">⠿</span>
68
- <span class="sort-label">${z(e.label)}</span>
68
+ <span class="sort-label">${B(e.label)}</span>
69
69
  <span class="sort-pos">#${t+1}</span>
70
70
  </li>`).join(``)}</ul>
71
71
  <div class="sort-status" id="sortStatus">Drag items to reorder. ${t.length} items.</div>
@@ -116,7 +116,7 @@ _Enterprise RAG may be unavailable. Local AI Kit search is unaffected._`}],isErr
116
116
  }
117
117
  });
118
118
  })();
119
- <\/script>`}function la(e){let t=e.columns??[],n=e.rows??[],r=e.stats??[],i=r.length?`<div class="dt-stats">${r.map(e=>`<div class="dt-stat"><div class="dt-stat-value"${e.color?` style="color:${e.color}"`:``}>${z(String(e.value))}</div><div class="dt-stat-label">${z(e.label)}</div></div>`).join(``)}</div>`:``,a=t.map(e=>`<th data-key="${z(e.key)}">${z(e.label)} <span class="sort-icon">⇅</span></th>`).join(``),o=n.map(e=>`<tr>${t.map(t=>`<td>${z(String(e[t.key]??``))}</td>`).join(``)}</tr>`).join(``);return`
119
+ <\/script>`}function ca(e){let t=e.columns??[],n=e.rows??[],r=e.stats??[],i=r.length?`<div class="dt-stats">${r.map(e=>`<div class="dt-stat"><div class="dt-stat-value"${e.color?` style="color:${e.color}"`:``}>${B(String(e.value))}</div><div class="dt-stat-label">${B(e.label)}</div></div>`).join(``)}</div>`:``,a=t.map(e=>`<th data-key="${B(e.key)}">${B(e.label)} <span class="sort-icon">⇅</span></th>`).join(``),o=n.map(e=>`<tr>${t.map(t=>`<td>${B(String(e[t.key]??``))}</td>`).join(``)}</tr>`).join(``);return`
120
120
  <style>
121
121
  .dt-stats{display:flex;gap:20px;padding:14px 0;flex-wrap:wrap}
122
122
  .dt-stat{text-align:center;min-width:90px;padding:10px 16px;background:var(--aikit-surface);border-radius:8px}
@@ -193,7 +193,7 @@ ${i}
193
193
  a.download = 'export.csv'; a.click();
194
194
  });
195
195
  })();
196
- <\/script>`}function ua(e){let t=e.categories??[],n=e.items??[];return`
196
+ <\/script>`}function la(e){let t=e.categories??[],n=e.items??[];return`
197
197
  <style>
198
198
  .pk-tabs{display:flex;gap:4px;padding:4px;background:var(--aikit-surface);border-radius:10px;
199
199
  margin-bottom:12px;overflow-x:auto;flex-wrap:wrap}
@@ -221,13 +221,13 @@ ${i}
221
221
  border:none;border-radius:8px;cursor:pointer;font-size:14px;font-weight:600}
222
222
  .pk-apply:hover{filter:brightness(1.1)}
223
223
  </style>
224
- <div class="pk-tabs" id="pkTabs">${[`<button class="pk-tab active" data-cat="all">All (${n.length})</button>`].concat(t.map(e=>{let t=n.filter(t=>t.category===e.id).length;return`<button class="pk-tab" data-cat="${z(e.id)}">${z(e.label)} (${t})</button>`})).join(``)}</div>
224
+ <div class="pk-tabs" id="pkTabs">${[`<button class="pk-tab active" data-cat="all">All (${n.length})</button>`].concat(t.map(e=>{let t=n.filter(t=>t.category===e.id).length;return`<button class="pk-tab" data-cat="${B(e.id)}">${B(e.label)} (${t})</button>`})).join(``)}</div>
225
225
  <input class="pk-search" id="pkSearch" placeholder="Search items...">
226
226
  <div class="pk-summary" id="pkSummary">0 selected</div>
227
- <div class="pk-list" id="pkList">${n.map(e=>{let t=e.status&&[`active`,`enabled`,`on`].includes(e.status.toLowerCase())?`success`:``,n=e.tags?.length?`<div class="pk-tags">${e.tags.map(e=>`<span class="pk-tag">${z(e)}</span>`).join(``)}</div>`:``;return`<div class="pk-item" data-cat="${z(e.category??``)}" data-label="${z(e.label.toLowerCase())}">
228
- <input type="checkbox" data-id="${z(e.id)}">
229
- <span class="pk-item-label">${z(e.label)}</span>
230
- ${e.status?`<span class="pk-item-status ${t}">${z(e.status)}</span>`:``}
227
+ <div class="pk-list" id="pkList">${n.map(e=>{let t=e.status&&[`active`,`enabled`,`on`].includes(e.status.toLowerCase())?`success`:``,n=e.tags?.length?`<div class="pk-tags">${e.tags.map(e=>`<span class="pk-tag">${B(e)}</span>`).join(``)}</div>`:``;return`<div class="pk-item" data-cat="${B(e.category??``)}" data-label="${B(e.label.toLowerCase())}">
228
+ <input type="checkbox" data-id="${B(e.id)}">
229
+ <span class="pk-item-label">${B(e.label)}</span>
230
+ ${e.status?`<span class="pk-item-status ${t}">${B(e.status)}</span>`:``}
231
231
  ${n}
232
232
  </div>`}).join(``)}</div>
233
233
  <button class="pk-apply" id="pkApply">Apply Selection</button>
@@ -264,7 +264,7 @@ ${i}
264
264
  }
265
265
  });
266
266
  })();
267
- <\/script>`}function da(e){let t=e.profile;return t?`
267
+ <\/script>`}function ua(e){let t=e.profile;return t?`
268
268
  <style>
269
269
  .fg-breadcrumb{display:flex;align-items:center;gap:4px;padding:10px 0;flex-wrap:wrap}
270
270
  .fg-crumb{background:var(--aikit-surface);border:1px solid var(--aikit-border);color:var(--aikit-text);
@@ -335,7 +335,7 @@ ${i}
335
335
 
336
336
  renderFlame();
337
337
  })();
338
- <\/script>`:`<p>No profile data provided</p>`}function fa(e){let t=e.fields??[];return`
338
+ <\/script>`:`<p>No profile data provided</p>`}function da(e){let t=e.fields??[];return`
339
339
  <style>
340
340
  .fm-container{display:grid;grid-template-columns:1fr 280px;gap:24px}
341
341
  @media(max-width:640px){.fm-container{grid-template-columns:1fr}}
@@ -361,7 +361,7 @@ ${i}
361
361
  </style>
362
362
  <div class="fm-container">
363
363
  <div class="fm-fields" id="fmFields">
364
- ${t.map(e=>{let t=e.type??`text`,n;if(t===`select`&&e.options){let t=e.options.map(t=>`<option value="${z(t)}"${t===e.value?` selected`:``}>${z(t)}</option>`).join(``);n=`<select class="fm-input" data-field="${z(e.name)}">${t}</select>`}else n=t===`textarea`?`<textarea class="fm-input fm-textarea" data-field="${z(e.name)}" rows="4">${z(e.value??``)}</textarea>`:t===`multi-select`&&e.options?`<div class="fm-multi">${e.options.map(t=>`<label class="fm-check-label"><input type="checkbox" data-field="${z(e.name)}" value="${z(t)}"${e.value?.includes(t)?` checked`:``}> ${z(t)}</label>`).join(``)}</div>`:`<input class="fm-input" type="${z(t)}" data-field="${z(e.name)}" value="${z(e.value??``)}">`;return`<div class="fm-group"><label class="fm-label">${z(e.label)}</label>${n}</div>`}).join(``)}
364
+ ${t.map(e=>{let t=e.type??`text`,n;if(t===`select`&&e.options){let t=e.options.map(t=>`<option value="${B(t)}"${t===e.value?` selected`:``}>${B(t)}</option>`).join(``);n=`<select class="fm-input" data-field="${B(e.name)}">${t}</select>`}else n=t===`textarea`?`<textarea class="fm-input fm-textarea" data-field="${B(e.name)}" rows="4">${B(e.value??``)}</textarea>`:t===`multi-select`&&e.options?`<div class="fm-multi">${e.options.map(t=>`<label class="fm-check-label"><input type="checkbox" data-field="${B(e.name)}" value="${B(t)}"${e.value?.includes(t)?` checked`:``}> ${B(t)}</label>`).join(``)}</div>`:`<input class="fm-input" type="${B(t)}" data-field="${B(e.name)}" value="${B(e.value??``)}">`;return`<div class="fm-group"><label class="fm-label">${B(e.label)}</label>${n}</div>`}).join(``)}
365
365
  <button class="fm-submit" id="fmSubmit">Submit</button>
366
366
  </div>
367
367
  <div class="fm-preview">
@@ -400,7 +400,7 @@ ${i}
400
400
  }
401
401
  });
402
402
  })();
403
- <\/script>`}function pa(e){let t=e.events??[],n={complete:`#22c55e`,active:`#3b82f6`,pending:`#9ca3af`,error:`#ef4444`};return`
403
+ <\/script>`}function fa(e){let t=e.events??[],n={complete:`#22c55e`,active:`#3b82f6`,pending:`#9ca3af`,error:`#ef4444`};return`
404
404
  <style>
405
405
  .tl-wrap{position:relative;padding-left:28px}
406
406
  .tl-wrap::before{content:'';position:absolute;top:0;bottom:0;left:10px;width:2px;background:var(--aikit-border,#3c3c3c)}
@@ -416,14 +416,14 @@ ${i}
416
416
  </style>
417
417
  <div class="tl-wrap">${t.map(e=>`
418
418
  <div class="tl-event">
419
- <span class="tl-dot" style="background:${n[e.status??`pending`]??n.pending}">${z(e.icon??(e.status===`complete`?`✓`:e.status===`error`?`✗`:`●`))}</span>
419
+ <span class="tl-dot" style="background:${n[e.status??`pending`]??n.pending}">${B(e.icon??(e.status===`complete`?`✓`:e.status===`error`?`✗`:`●`))}</span>
420
420
  <div class="tl-body">
421
- <div class="tl-title">${z(e.title)}</div>
422
- ${e.timestamp?`<div class="tl-time">${z(e.timestamp)}</div>`:``}
423
- ${e.description?`<div class="tl-desc">${z(e.description)}</div>`:``}
424
- ${e.category?`<span class="tl-cat">${z(e.category)}</span>`:``}
421
+ <div class="tl-title">${B(e.title)}</div>
422
+ ${e.timestamp?`<div class="tl-time">${B(e.timestamp)}</div>`:``}
423
+ ${e.description?`<div class="tl-desc">${B(e.description)}</div>`:``}
424
+ ${e.category?`<span class="tl-cat">${B(e.category)}</span>`:``}
425
425
  </div>
426
- </div>`).join(``)}</div>`}function ma(e){let t=e.columns??[],n=e.cards??[];return`
426
+ </div>`).join(``)}</div>`}function pa(e){let t=e.columns??[],n=e.cards??[];return`
427
427
  <style>
428
428
  .kbn-board{display:flex;gap:14px;overflow-x:auto;padding-bottom:4px}
429
429
  .kbn-col{min-width:220px;flex:1;border:1px solid var(--aikit-border,#3c3c3c);border-radius:8px;background:var(--aikit-surface,#252526)}
@@ -443,18 +443,18 @@ ${i}
443
443
  </style>
444
444
  <div class="kbn-board">${t.map(e=>{let t=n.filter(t=>t.column===e.id),r=t.map(e=>`
445
445
  <div class="kbn-card ${e.priority?`priority-${e.priority}`:``}">
446
- <div class="kbn-card-title">${z(e.title)}</div>
447
- ${e.description?`<div class="kbn-card-desc">${z(e.description)}</div>`:``}
448
- ${e.tags?.length?`<div class="kbn-tags">${e.tags.map(e=>`<span class="kbn-tag">${z(e)}</span>`).join(``)}</div>`:``}
446
+ <div class="kbn-card-title">${B(e.title)}</div>
447
+ ${e.description?`<div class="kbn-card-desc">${B(e.description)}</div>`:``}
448
+ ${e.tags?.length?`<div class="kbn-tags">${e.tags.map(e=>`<span class="kbn-tag">${B(e)}</span>`).join(``)}</div>`:``}
449
449
  </div>`).join(``);return`
450
450
  <div class="kbn-col">
451
451
  <div class="kbn-col-hdr">
452
452
  <span class="kbn-col-dot" style="background:${e.color??`var(--aikit-accent,#3b82f6)`}"></span>
453
- <span class="kbn-col-title">${z(e.label)}</span>
453
+ <span class="kbn-col-title">${B(e.label)}</span>
454
454
  <span class="kbn-col-count">${t.length}</span>
455
455
  </div>
456
456
  <div class="kbn-cards">${r}</div>
457
- </div>`}).join(``)}</div>`}function ha(e){let t=Array.isArray(e.root)?e.root:e.root?[e.root]:e.roots??[],n=0;function r(e,t){let i=`tn-${n++}`,a=e.children&&e.children.length>0,o=e.icon?`<span class="tr-icon">${z(e.icon)}</span>`:``,s=a?`<button class="tr-toggle" data-target="${i}">▸</button>`:`<span class="tr-spacer"></span>`,c=e.metadata?Object.entries(e.metadata).map(([e,t])=>`<span class="tr-meta">${z(e)}: ${z(t)}</span>`).join(` `):``,l=`<div class="tr-row">${s}${o}<span class="tr-label">${z(e.label)}</span>${c}</div>`;return a&&(l+=`<div class="tr-children" id="${i}" hidden>${(e.children??[]).map(e=>r(e,t+1)).join(``)}</div>`),`<div class="tr-node">${l}</div>`}return`
457
+ </div>`}).join(``)}</div>`}function ma(e){let t=Array.isArray(e.root)?e.root:e.root?[e.root]:e.roots??[],n=0;function r(e,t){let i=`tn-${n++}`,a=e.children&&e.children.length>0,o=e.icon?`<span class="tr-icon">${B(e.icon)}</span>`:``,s=a?`<button class="tr-toggle" data-target="${i}">▸</button>`:`<span class="tr-spacer"></span>`,c=e.metadata?Object.entries(e.metadata).map(([e,t])=>`<span class="tr-meta">${B(e)}: ${B(t)}</span>`).join(` `):``,l=`<div class="tr-row">${s}${o}<span class="tr-label">${B(e.label)}</span>${c}</div>`;return a&&(l+=`<div class="tr-children" id="${i}" hidden>${(e.children??[]).map(e=>r(e,t+1)).join(``)}</div>`),`<div class="tr-node">${l}</div>`}return`
458
458
  <style>
459
459
  .tr-wrap{font:13px var(--aikit-font-sans,system-ui,sans-serif)}
460
460
  .tr-node{margin-left:16px}
@@ -479,7 +479,7 @@ ${i}
479
479
  });
480
480
  });
481
481
  })();
482
- <\/script>`}function ga(e){let t=e.files??[],n=e.stats,r={added:`🟢`,modified:`🟡`,deleted:`🔴`,renamed:`🔵`};return`
482
+ <\/script>`}function ha(e){let t=e.files??[],n=e.stats,r={added:`🟢`,modified:`🟡`,deleted:`🔴`,renamed:`🔵`};return`
483
483
  <style>
484
484
  .dv-stats{padding:8px 12px;background:var(--aikit-surface,#252526);border-radius:6px;color:var(--aikit-text,#e5e7eb);font:600 13px var(--aikit-font-sans,system-ui,sans-serif);margin-bottom:12px}
485
485
  .dv-add{color:#22c55e}.dv-del{color:#ef4444}
@@ -494,12 +494,12 @@ ${i}
494
494
  .dv-ln{display:inline-block;width:40px;text-align:right;margin-right:8px;color:var(--aikit-muted,#9ca3af);font-size:11px}
495
495
  </style>
496
496
  ${n?`<div class="dv-stats"><span class="dv-add">+${n.additions}</span> <span class="dv-del">-${n.deletions}</span> in ${n.filesChanged} files</div>`:``}
497
- <div class="dv-wrap">${t.map(e=>{let t=e.hunks.map(e=>{let t=e.changes.map(e=>{let t=e.type===`add`?`dv-line-add`:e.type===`delete`?`dv-line-del`:`dv-line-ctx`,n=e.type===`add`?`+`:e.type===`delete`?`-`:` `;return`<div class="${t}">${e.line==null?``:`<span class="dv-ln">${e.line}</span>`}<code>${n} ${z(e.content)}</code></div>`}).join(``);return`<div class="dv-hunk"><div class="dv-hunk-hdr">${z(e.header)}</div>${t}</div>`}).join(``),n=r[e.status]??`⚪`,i=e.oldPath?` ← ${z(e.oldPath)}`:``;return`
497
+ <div class="dv-wrap">${t.map(e=>{let t=e.hunks.map(e=>{let t=e.changes.map(e=>{let t=e.type===`add`?`dv-line-add`:e.type===`delete`?`dv-line-del`:`dv-line-ctx`,n=e.type===`add`?`+`:e.type===`delete`?`-`:` `;return`<div class="${t}">${e.line==null?``:`<span class="dv-ln">${e.line}</span>`}<code>${n} ${B(e.content)}</code></div>`}).join(``);return`<div class="dv-hunk"><div class="dv-hunk-hdr">${B(e.header)}</div>${t}</div>`}).join(``),n=r[e.status]??`⚪`,i=e.oldPath?` ← ${B(e.oldPath)}`:``;return`
498
498
  <details class="dv-file" open>
499
- <summary class="dv-file-hdr">${n} <strong>${z(e.path)}</strong>${i}
499
+ <summary class="dv-file-hdr">${n} <strong>${B(e.path)}</strong>${i}
500
500
  <span class="dv-add">+${e.additions}</span> <span class="dv-del">-${e.deletions}</span></summary>
501
501
  ${t}
502
- </details>`}).join(``)}</div>`}function _a(e){let t=e.metrics??[],n={success:`#22c55e`,warning:`#f59e0b`,error:`#ef4444`,info:`#38bdf8`},r={up:`↑`,down:`↓`,neutral:`→`};return`
502
+ </details>`}).join(``)}</div>`}function ga(e){let t=e.metrics??[],n={success:`#22c55e`,warning:`#f59e0b`,error:`#ef4444`,info:`#38bdf8`},r={up:`↑`,down:`↓`,neutral:`→`};return`
503
503
  <style>
504
504
  .db-grid{display:grid;grid-template-columns:repeat(auto-fit,minmax(200px,1fr));gap:14px}
505
505
  .db-card{padding:14px;border:1px solid var(--aikit-border,#3c3c3c);border-left-width:4px;border-radius:8px;background:var(--aikit-surface,#252526)}
@@ -512,14 +512,14 @@ ${n?`<div class="dv-stats"><span class="dv-add">+${n.additions}</span> <span cla
512
512
  .db-list{margin-top:8px;display:grid;gap:4px}
513
513
  .db-list-row{display:flex;justify-content:space-between;padding:4px 0;color:var(--aikit-muted,#9ca3af);font:400 12px var(--aikit-font-sans,system-ui,sans-serif);border-bottom:1px solid var(--aikit-border,#3c3c3c)}
514
514
  </style>
515
- <div class="db-grid">${t.map(e=>{let t=n[e.status??`info`]??n.info,i=e.trend?`<span class="db-trend ${e.trend.direction}">${r[e.trend.direction]??``} ${z(e.trend.value)}</span>`:``,a=`<div class="db-value">${z(String(e.value))}</div>${i}`;return e.type===`progress`&&e.progress!=null&&(a+=`<div class="db-progress"><div class="db-progress-bar" style="width:${Math.min(100,Math.max(0,e.progress))}%;background:${t}"></div></div>`),e.type===`list`&&e.items?.length&&(a+=`<div class="db-list">${e.items.map(e=>`<div class="db-list-row"><span>${z(e.label)}</span><span>${z(e.value)}</span></div>`).join(``)}</div>`),`
515
+ <div class="db-grid">${t.map(e=>{let t=n[e.status??`info`]??n.info,i=e.trend?`<span class="db-trend ${e.trend.direction}">${r[e.trend.direction]??``} ${B(e.trend.value)}</span>`:``,a=`<div class="db-value">${B(String(e.value))}</div>${i}`;return e.type===`progress`&&e.progress!=null&&(a+=`<div class="db-progress"><div class="db-progress-bar" style="width:${Math.min(100,Math.max(0,e.progress))}%;background:${t}"></div></div>`),e.type===`list`&&e.items?.length&&(a+=`<div class="db-list">${e.items.map(e=>`<div class="db-list-row"><span>${B(e.label)}</span><span>${B(e.value)}</span></div>`).join(``)}</div>`),`
516
516
  <div class="db-card" style="border-left-color:${t}">
517
- <div class="db-label">${z(e.label)}</div>
517
+ <div class="db-label">${B(e.label)}</div>
518
518
  ${a}
519
- </div>`}).join(``)}</div>`}const va=F.object({action:F.enum([`view`,`update`]).default(`view`).describe(`Action to perform: view current configuration or update it`),updates:F.record(F.string(),F.unknown()).optional().describe(`Partial configuration to merge when action is "update". Supports nested paths like { "indexing": { "chunkSize": 2000 } }`)}),ya=E(`config-tool`),B=`ui://aikit/config.html`,ba=[`serverName`,`toolPrefix`,`autoIndex`,`sources`,`indexing`];function xa(){let e=process.env.AIKIT_CONFIG_PATH;if(e&&a(e))return e;let t=_(process.cwd(),`aikit.config.json`);return a(t)?t:``}function V(e){return typeof e==`object`&&!!e&&!Array.isArray(e)}function Sa(e,t){let n={...e};for(let[e,r]of Object.entries(t)){let t=n[e];if(V(t)&&V(r)){n[e]=Sa(t,r);continue}n[e]=r}return n}function Ca(e){return e===void 0?`undefined`:JSON.stringify(e)}function wa(e,t,n,r=``){let i=[];for(let[a,o]of Object.entries(n)){let n=r?`${r}.${a}`:a,s=e[a],c=t[a];if(V(o)&&V(s)&&V(c)){i.push(...wa(s,c,o,n));continue}Ca(s)!==Ca(c)&&i.push(`- ${n}: ${Ca(s)} -> ${Ca(c)}`)}return i}function Ta(e){let t=[];for(let[n,r]of Object.entries(e)){if(n===`store`||n===`embedding`||n===`er`){t.push(`Field "${n}" cannot be updated through this tool because it requires a restart.`);continue}if(n===`onboardDir`||n===`stateDir`||n===`curated`){t.push(`Field "${n}" cannot be updated through this tool because it requires a restart.`);continue}if(n===`autoIndex`){typeof r!=`boolean`&&t.push(`autoIndex must be a boolean.`);continue}if(n===`serverName`||n===`toolPrefix`){typeof r!=`string`&&t.push(`${n} must be a string.`);continue}if(n===`sources`){(!Array.isArray(r)||r.length===0)&&t.push(`sources must be a non-empty array.`);continue}if(n===`indexing`){if(!V(r)){t.push(`indexing must be an object.`);continue}for(let[e,n]of Object.entries(r)){if(e!==`chunkSize`&&e!==`chunkOverlap`&&e!==`minChunkSize`&&e!==`concurrency`){t.push(`indexing.${e} is not updatable through this tool.`);continue}(typeof n!=`number`||!Number.isFinite(n)||n<=0)&&t.push(`indexing.${e} must be a positive number.`)}continue}t.push(`Field "${n}" is not updatable through this tool.`)}return t}function Ea(e){let t=[];if(`autoIndex`in e&&typeof e.autoIndex!=`boolean`&&t.push(`autoIndex must be a boolean.`),`serverName`in e&&e.serverName!==void 0&&typeof e.serverName!=`string`&&t.push(`serverName must be a string.`),`toolPrefix`in e&&e.toolPrefix!==void 0&&typeof e.toolPrefix!=`string`&&t.push(`toolPrefix must be a string.`),(!Array.isArray(e.sources)||e.sources.length===0)&&t.push(`sources must be a non-empty array.`),!V(e.indexing))return t.push(`indexing must be an object.`),t;for(let n of[`chunkSize`,`chunkOverlap`,`minChunkSize`,`concurrency`]){let r=e.indexing[n];r!==void 0&&(typeof r!=`number`||!Number.isFinite(r)||r<=0)&&t.push(`indexing.${n} must be a positive number.`)}return t}function Da(e){let t=JSON.parse(s(e,`utf-8`));if(!V(t))throw Error(`Config file must contain a JSON object.`);return t}function Oa(e,t){return[`AI Kit Configuration`,`Config file: ${t}`,``,JSON.stringify(e,null,2)].join(`
520
- `)}function ka(e){return e===void 0?``:V(e)||Array.isArray(e)?JSON.stringify(e,null,2):String(e)}function Aa(e){return _a({metrics:Object.entries(e).map(([e,t])=>({label:e,value:V(t)||Array.isArray(t)?JSON.stringify(t,null,2):String(t),status:`info`}))})}function ja(e){return fa({fields:ba.map(t=>{let n=e[t];return typeof n==`boolean`?{name:t,label:t,type:`select`,options:[`true`,`false`],value:String(n)}:{name:t,label:t,type:V(n)||Array.isArray(n)?`textarea`:`text`,value:ka(n)}})})}function Ma(e,t){return`<!DOCTYPE html><html><head><meta charset="utf-8"><title>${e}</title><style>:root{--aikit-bg:#1e1e1e;--aikit-surface:#252526;--aikit-surface2:#2d2d30;--aikit-border:#3c3c3c;--aikit-text:#e5e7eb;--aikit-muted:#9ca3af;--aikit-accent:#60a5fa;--aikit-success:#22c55e;font-family:system-ui,sans-serif}body{margin:0;padding:20px;background:var(--aikit-bg);color:var(--aikit-text);font-family:system-ui,sans-serif}h2{margin:0 0 16px}</style></head><body><h2>${e}</h2>${t}</body></html>`}function Na(e,t,n){_n(e,`AI Kit Config App`,B,{description:`View and update AI Kit configuration`},async()=>({contents:[{uri:B,mimeType:gn,text:Ma(t,n)}]}))}function Pa(e,t){let n=I(`config`);vn(e,`config`,{title:n.title,description:`View and update AI Kit server configuration. Use action "view" to see current settings, or "update" to modify configuration values. Changes are written to aikit.config.json and take effect on next server restart. Some settings (store, embedding) require a full restart.`,inputSchema:va,annotations:n.annotations,_meta:{ui:{resourceUri:B}}},async n=>{let{action:r,updates:i}=n;if(r===`view`){let n=xa(),r=n||`(runtime config only; no aikit.config.json found)`;try{let i=n?Da(n):t,a=Oa(i,r);return Na(e,`Configuration`,Aa(i)),{content:[{type:`text`,text:a}],ui:{type:`resource`,uri:B}}}catch(e){return ya.error(`Failed to read config for view`,{configPath:n,error:e instanceof Error?e.message:String(e)}),{content:[{type:`text`,text:`Failed to read configuration from ${r}: ${e instanceof Error?e.message:String(e)}`}]}}}if(!i)return Na(e,`Configuration Form`,ja(t)),{content:[{type:`text`,text:`No updates provided. Pass an updates object when action is "update".`}],ui:{type:`resource`,uri:B}};let a=xa();if(!a)return Na(e,`Configuration Form`,ja(t)),{content:[{type:`text`,text:`No config file found to update.`}],ui:{type:`resource`,uri:B}};let o=Ta(i);if(o.length>0)return Na(e,`Configuration Form`,ja(Da(a))),{content:[{type:`text`,text:`Configuration update rejected:\n${o.map(e=>`- ${e}`).join(`
521
- `)}`}],ui:{type:`resource`,uri:B}};try{let t=Da(a),n=Sa(t,i),r=Ea(n);if(r.length>0)return Na(e,`Configuration Form`,ja(t)),{content:[{type:`text`,text:`Configuration update rejected:\n${r.map(e=>`- ${e}`).join(`
522
- `)}`}],ui:{type:`resource`,uri:B}};f(a,JSON.stringify(n,null,2),`utf-8`);let o=wa(t,n,i),s=`Configuration updated successfully.
519
+ </div>`}).join(``)}</div>`}const _a=F.object({action:F.enum([`view`,`update`]).default(`view`).describe(`Action to perform: view current configuration or update it`),updates:F.record(F.string(),F.unknown()).optional().describe(`Partial configuration to merge when action is "update". Supports nested paths like { "indexing": { "chunkSize": 2000 } }`)}),va=E(`config-tool`),V=`ui://aikit/config.html`,ya=[`serverName`,`toolPrefix`,`autoIndex`,`sources`,`indexing`];function ba(){let e=process.env.AIKIT_CONFIG_PATH;if(e&&a(e))return e;let t=_(process.cwd(),`aikit.config.json`);return a(t)?t:``}function H(e){return typeof e==`object`&&!!e&&!Array.isArray(e)}function xa(e,t){let n={...e};for(let[e,r]of Object.entries(t)){let t=n[e];if(H(t)&&H(r)){n[e]=xa(t,r);continue}n[e]=r}return n}function Sa(e){return e===void 0?`undefined`:JSON.stringify(e)}function Ca(e,t,n,r=``){let i=[];for(let[a,o]of Object.entries(n)){let n=r?`${r}.${a}`:a,s=e[a],c=t[a];if(H(o)&&H(s)&&H(c)){i.push(...Ca(s,c,o,n));continue}Sa(s)!==Sa(c)&&i.push(`- ${n}: ${Sa(s)} -> ${Sa(c)}`)}return i}function wa(e){let t=[];for(let[n,r]of Object.entries(e)){if(n===`store`||n===`embedding`||n===`er`){t.push(`Field "${n}" cannot be updated through this tool because it requires a restart.`);continue}if(n===`onboardDir`||n===`stateDir`||n===`curated`){t.push(`Field "${n}" cannot be updated through this tool because it requires a restart.`);continue}if(n===`autoIndex`){typeof r!=`boolean`&&t.push(`autoIndex must be a boolean.`);continue}if(n===`serverName`||n===`toolPrefix`){typeof r!=`string`&&t.push(`${n} must be a string.`);continue}if(n===`sources`){(!Array.isArray(r)||r.length===0)&&t.push(`sources must be a non-empty array.`);continue}if(n===`indexing`){if(!H(r)){t.push(`indexing must be an object.`);continue}for(let[e,n]of Object.entries(r)){if(e!==`chunkSize`&&e!==`chunkOverlap`&&e!==`minChunkSize`&&e!==`concurrency`){t.push(`indexing.${e} is not updatable through this tool.`);continue}(typeof n!=`number`||!Number.isFinite(n)||n<=0)&&t.push(`indexing.${e} must be a positive number.`)}continue}t.push(`Field "${n}" is not updatable through this tool.`)}return t}function Ta(e){let t=[];if(`autoIndex`in e&&typeof e.autoIndex!=`boolean`&&t.push(`autoIndex must be a boolean.`),`serverName`in e&&e.serverName!==void 0&&typeof e.serverName!=`string`&&t.push(`serverName must be a string.`),`toolPrefix`in e&&e.toolPrefix!==void 0&&typeof e.toolPrefix!=`string`&&t.push(`toolPrefix must be a string.`),(!Array.isArray(e.sources)||e.sources.length===0)&&t.push(`sources must be a non-empty array.`),!H(e.indexing))return t.push(`indexing must be an object.`),t;for(let n of[`chunkSize`,`chunkOverlap`,`minChunkSize`,`concurrency`]){let r=e.indexing[n];r!==void 0&&(typeof r!=`number`||!Number.isFinite(r)||r<=0)&&t.push(`indexing.${n} must be a positive number.`)}return t}function Ea(e){let t=JSON.parse(s(e,`utf-8`));if(!H(t))throw Error(`Config file must contain a JSON object.`);return t}function Da(e,t){return[`AI Kit Configuration`,`Config file: ${t}`,``,JSON.stringify(e,null,2)].join(`
520
+ `)}function Oa(e){return e===void 0?``:H(e)||Array.isArray(e)?JSON.stringify(e,null,2):String(e)}function ka(e){return ga({metrics:Object.entries(e).map(([e,t])=>({label:e,value:H(t)||Array.isArray(t)?JSON.stringify(t,null,2):String(t),status:`info`}))})}function Aa(e){return da({fields:ya.map(t=>{let n=e[t];return typeof n==`boolean`?{name:t,label:t,type:`select`,options:[`true`,`false`],value:String(n)}:{name:t,label:t,type:H(n)||Array.isArray(n)?`textarea`:`text`,value:Oa(n)}})})}function ja(e,t){return`<!DOCTYPE html><html><head><meta charset="utf-8"><title>${e}</title><style>:root{--aikit-bg:#1e1e1e;--aikit-surface:#252526;--aikit-surface2:#2d2d30;--aikit-border:#3c3c3c;--aikit-text:#e5e7eb;--aikit-muted:#9ca3af;--aikit-accent:#60a5fa;--aikit-success:#22c55e;font-family:system-ui,sans-serif}body{margin:0;padding:20px;background:var(--aikit-bg);color:var(--aikit-text);font-family:system-ui,sans-serif}h2{margin:0 0 16px}</style></head><body><h2>${e}</h2>${t}</body></html>`}function Ma(e,t,n){_n(e,`AI Kit Config App`,V,{description:`View and update AI Kit configuration`},async()=>({contents:[{uri:V,mimeType:gn,text:ja(t,n)}]}))}function Na(e,t){let n=L(`config`);vn(e,`config`,{title:n.title,description:`View and update AI Kit server configuration. Use action "view" to see current settings, or "update" to modify configuration values. Changes are written to aikit.config.json and take effect on next server restart. Some settings (store, embedding) require a full restart.`,inputSchema:_a,annotations:n.annotations,_meta:{ui:{resourceUri:V}}},async n=>{let{action:r,updates:i}=n;if(r===`view`){let n=ba(),r=n||`(runtime config only; no aikit.config.json found)`;try{let i=n?Ea(n):t,a=Da(i,r);return Ma(e,`Configuration`,ka(i)),{content:[{type:`text`,text:a}],ui:{type:`resource`,uri:V}}}catch(e){return va.error(`Failed to read config for view`,{configPath:n,error:e instanceof Error?e.message:String(e)}),{content:[{type:`text`,text:`Failed to read configuration from ${r}: ${e instanceof Error?e.message:String(e)}`}]}}}if(!i)return Ma(e,`Configuration Form`,Aa(t)),{content:[{type:`text`,text:`No updates provided. Pass an updates object when action is "update".`}],ui:{type:`resource`,uri:V}};let a=ba();if(!a)return Ma(e,`Configuration Form`,Aa(t)),{content:[{type:`text`,text:`No config file found to update.`}],ui:{type:`resource`,uri:V}};let o=wa(i);if(o.length>0)return Ma(e,`Configuration Form`,Aa(Ea(a))),{content:[{type:`text`,text:`Configuration update rejected:\n${o.map(e=>`- ${e}`).join(`
521
+ `)}`}],ui:{type:`resource`,uri:V}};try{let t=Ea(a),n=xa(t,i),r=Ta(n);if(r.length>0)return Ma(e,`Configuration Form`,Aa(t)),{content:[{type:`text`,text:`Configuration update rejected:\n${r.map(e=>`- ${e}`).join(`
522
+ `)}`}],ui:{type:`resource`,uri:V}};f(a,JSON.stringify(n,null,2),`utf-8`);let o=Ca(t,n,i),s=`Configuration updated successfully.
523
523
 
524
524
  Config file: ${a}
525
525
 
@@ -527,23 +527,23 @@ Changed values:
527
527
  ${o.length>0?o.join(`
528
528
  `):`- No effective changes; requested values already matched the file.`}
529
529
 
530
- These changes take effect on the next server restart.`;return Na(e,`Configuration`,Aa(n)),{content:[{type:`text`,text:s}],ui:{type:`resource`,uri:B}}}catch(e){return ya.error(`Failed to update config`,{configPath:a,error:e instanceof Error?e.message:String(e)}),{content:[{type:`text`,text:`Failed to update configuration in ${a}: ${e instanceof Error?e.message:String(e)}`}]}}})}const Fa=E(`cross-workspace`);function Ia(e,t){if(!k())return[];let n=ee();if(n.length===0)return[];if(e.includes(`*`))return t?n.filter(e=>e.partition!==t):n;let r=[];for(let i of e){let e=n.find(e=>e.partition===i);if(e){e.partition!==t&&r.push(e);continue}let a=n.filter(e=>e.partition!==t&&e.partition.replace(/-[a-f0-9]{8}$/,``)===i.toLowerCase());r.push(...a)}let i=new Set;return r.filter(e=>i.has(e.partition)?!1:(i.add(e.partition),!0))}async function La(e){let t=new Map;for(let n of e)try{let e=await he({backend:`lancedb`,path:O(n.partition)});await e.initialize(),t.set(n.partition,e)}catch(e){Fa.warn(`Failed to open workspace store`,{partition:n.partition,err:e})}return{stores:t,closeAll:async()=>{for(let[,e]of t)try{await e.close()}catch{}}}}async function Ra(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.search(t,n)).map(t=>({...t,workspace:e}))}catch(t){return Fa.warn(`Cross-workspace search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}async function za(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.ftsSearch(t,n)).map(t=>({...t,workspace:e}))}catch(t){return Fa.warn(`Cross-workspace FTS search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}const Ba=E(`tools:context`);function Va(e,t,n,r){let i=I(`compact`);e.registerTool(`compact`,{title:i.title,description:"Compress text to relevant sections using embedding similarity (no LLM). Provide either `text` or `path` (server reads the file — saves a round-trip). Segments by paragraph/sentence/line.",outputSchema:ji,inputSchema:{text:F.string().optional().describe(`The text to compress (provide this OR path, not both)`),path:F.string().optional().describe(`File path to read server-side — avoids read_file round-trip + token doubling (provide this OR text)`),query:F.string().describe(`Focus query — what are you trying to understand?`),max_chars:F.number().min(100).max(5e4).default(3e3).describe(`Target output size in characters`),segmentation:F.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split the text for scoring`),token_budget:F.number().min(50).max(12500).optional().describe(`Token budget — overrides max_chars (approx 4 chars per token). Use to fit output into a specific context window.`)},annotations:i.annotations},async({text:e,path:i,query:a,max_chars:o,segmentation:s,token_budget:c})=>{try{let l=i&&!h(i)?_(r,i):i;if(!e&&!l)return{content:[{type:`text`,text:`Error: Either "text" or "path" must be provided.`}],isError:!0};let u=await Oe(t,{text:e,path:l,query:a,maxChars:o,tokenBudget:c,segmentation:s,cache:n});return{content:[{type:`text`,text:[`Compressed ${u.originalChars} → ${u.compressedChars} chars (${(u.ratio*100).toFixed(0)}%)`,`Kept ${u.segmentsKept}/${u.segmentsTotal} segments`,``,u.text].join(`
531
- `)}],structuredContent:{originalChars:u.originalChars,compressedChars:u.compressedChars,ratio:u.ratio,segmentsKept:u.segmentsKept,segmentsTotal:u.segmentsTotal}}}catch(e){return Ba.error(`Compact failed`,A(e)),{content:[{type:`text`,text:`Compact failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ha(e,t,n){let r=I(`scope_map`);e.registerTool(`scope_map`,{title:r.title,description:`Generate a task-scoped reading plan. Given a task description, identifies which files and sections are relevant, with estimated token counts and suggested reading order.`,outputSchema:yi,inputSchema:{task:F.string().describe(`Description of the task to scope`),max_files:F.number().min(1).max(50).default(15).describe(`Maximum files to include`),content_type:F.enum(b).optional().describe(`Filter by content type`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`)},annotations:r.annotations},async({task:e,max_files:r,content_type:i,max_tokens:a})=>{try{let o=await Ft(t,n,{task:e,maxFiles:r,contentType:i}),s=[`## Scope Map: ${e}`,`Total estimated tokens: ~${o.totalEstimatedTokens}`,``,`### Files (by relevance)`,...o.files.map((e,t)=>`${t+1}. **${e.path}** (~${e.estimatedTokens} tokens, ${(e.relevance*100).toFixed(0)}% relevant)\n ${e.reason}\n Focus: ${e.focusRanges.map(e=>`L${e.start}-${e.end}`).join(`, `)}`),``,`### Suggested Reading Order`,...o.readingOrder.map((e,t)=>`${t+1}. ${e}`),``,`### Suggested Compact Calls`,`_Estimated compressed total: ~${Math.ceil(o.totalEstimatedTokens/5)} tokens_`,...o.compactCommands.map((e,t)=>`${t+1}. ${e}`)].join(`
532
- `)+"\n\n---\n_Next: Use `search` to dive into specific files, or `compact` to compress file contents for context._";return{content:[{type:`text`,text:a?Xt(s,a):s}],structuredContent:{files:o.files.map(e=>({path:e.path,relevance:e.relevance,estimatedTokens:e.estimatedTokens,...e.focusRanges.length>0?{focusLines:e.focusRanges.map(e=>`L${e.start}-${e.end}`)}:{}})),totalFiles:o.files.length,totalEstimatedTokens:o.totalEstimatedTokens,task:e}}}catch(e){return Ba.error(`Scope map failed`,A(e)),{content:[{type:`text`,text:`Scope map failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ua(e,t,n){let r=I(`find`);e.registerTool(`find`,{title:r.title,description:`Multi-strategy search combining vector, FTS, glob, and regex. Use for precise queries needing multiple strategies. mode=examples finds real usage of a symbol. For general discovery use search instead.`,outputSchema:_i,inputSchema:{query:F.string().optional().describe(`Semantic/keyword search query (required for mode "examples")`),glob:F.string().optional().describe(`File glob pattern (search mode only)`),pattern:F.string().optional().describe(`Regex pattern to match in content (search mode only)`),limit:F.number().min(1).max(50).default(10).describe(`Max results`),content_type:F.enum(b).optional().describe(`Filter by content type`),mode:F.enum([`search`,`examples`]).default(`search`).describe(`Mode: "search" (default) for federated search, "examples" to find usage examples of a symbol/pattern`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all. User-level mode only.`)},annotations:r.annotations},async({query:e,glob:r,pattern:i,limit:a,content_type:o,mode:s,max_tokens:c,workspaces:l})=>{try{if(s===`examples`){if(!e)return{content:[{type:`text`,text:`Error: "query" is required for mode "examples".`}],isError:!0};let r=await We(t,n,{query:e,limit:a,contentType:o}),i=JSON.stringify(r);return{content:[{type:`text`,text:c?Xt(i,c):i}]}}let u=await He(t,n,{query:e,glob:r,pattern:i,limit:a,contentType:o}),d=``;if(l&&l.length>0&&e){let n=Ia(l,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:i}=await La(n);try{let i=await Ra(r,await t.embedQuery(e),{limit:a,contentType:o});for(let e of i)u.results.push({path:`[${e.workspace}] ${e.record.sourcePath}`,score:e.score,source:`cross-workspace`,lineRange:e.record.startLine?{start:e.record.startLine,end:e.record.endLine}:void 0,preview:e.record.content.slice(0,200)});u.results.sort((e,t)=>t.score-e.score),u.results=u.results.slice(0,a),u.totalFound=u.results.length,d=` + ${n.length} workspace(s)`}finally{await i()}}}if(u.results.length===0)return{content:[{type:`text`,text:`No results found.${u.failedStrategies?.length?`\nStrategies attempted: ${u.strategies.join(`, `)}\nFailed: ${u.failedStrategies.map(e=>`${e.strategy} (${e.reason})`).join(`, `)}`:u.strategies.length===0?`
530
+ These changes take effect on the next server restart.`;return Ma(e,`Configuration`,ka(n)),{content:[{type:`text`,text:s}],ui:{type:`resource`,uri:V}}}catch(e){return va.error(`Failed to update config`,{configPath:a,error:e instanceof Error?e.message:String(e)}),{content:[{type:`text`,text:`Failed to update configuration in ${a}: ${e instanceof Error?e.message:String(e)}`}]}}})}const Pa=E(`cross-workspace`);function Fa(e,t){if(!k())return[];let n=ee();if(n.length===0)return[];if(e.includes(`*`))return t?n.filter(e=>e.partition!==t):n;let r=[];for(let i of e){let e=n.find(e=>e.partition===i);if(e){e.partition!==t&&r.push(e);continue}let a=n.filter(e=>e.partition!==t&&e.partition.replace(/-[a-f0-9]{8}$/,``)===i.toLowerCase());r.push(...a)}let i=new Set;return r.filter(e=>i.has(e.partition)?!1:(i.add(e.partition),!0))}async function Ia(e){let t=new Map;for(let n of e)try{let e=await he({backend:`lancedb`,path:O(n.partition)});await e.initialize(),t.set(n.partition,e)}catch(e){Pa.warn(`Failed to open workspace store`,{partition:n.partition,err:e})}return{stores:t,closeAll:async()=>{for(let[,e]of t)try{await e.close()}catch{}}}}async function La(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.search(t,n)).map(t=>({...t,workspace:e}))}catch(t){return Pa.warn(`Cross-workspace search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}async function Ra(e,t,n){let r=[...e.entries()].map(async([e,r])=>{try{return(await r.ftsSearch(t,n)).map(t=>({...t,workspace:e}))}catch(t){return Pa.warn(`Cross-workspace FTS search failed for partition`,{partition:e,err:t}),[]}});return(await Promise.all(r)).flat().sort((e,t)=>t.score-e.score).slice(0,n.limit)}const za=E(`tools:context`);function Ba(e,t,n,r){let i=L(`compact`);e.registerTool(`compact`,{title:i.title,description:"Compress text to relevant sections using embedding similarity (no LLM). Provide either `text` or `path` (server reads the file — saves a round-trip). Segments by paragraph/sentence/line.",outputSchema:Ai,inputSchema:{text:F.string().optional().describe(`The text to compress (provide this OR path, not both)`),path:F.string().optional().describe(`File path to read server-side — avoids read_file round-trip + token doubling (provide this OR text)`),query:F.string().describe(`Focus query — what are you trying to understand?`),max_chars:F.number().min(100).max(5e4).default(3e3).describe(`Target output size in characters`),segmentation:F.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split the text for scoring`),token_budget:F.number().min(50).max(12500).optional().describe(`Token budget — overrides max_chars (approx 4 chars per token). Use to fit output into a specific context window.`)},annotations:i.annotations},async({text:e,path:i,query:a,max_chars:o,segmentation:s,token_budget:c})=>{try{let l=i&&!h(i)?_(r,i):i;if(!e&&!l)return{content:[{type:`text`,text:`Error: Either "text" or "path" must be provided.`}],isError:!0};let u=await Oe(t,{text:e,path:l,query:a,maxChars:o,tokenBudget:c,segmentation:s,cache:n});return{content:[{type:`text`,text:[`Compressed ${u.originalChars} → ${u.compressedChars} chars (${(u.ratio*100).toFixed(0)}%)`,`Kept ${u.segmentsKept}/${u.segmentsTotal} segments`,``,u.text].join(`
531
+ `)}],structuredContent:{originalChars:u.originalChars,compressedChars:u.compressedChars,ratio:u.ratio,segmentsKept:u.segmentsKept,segmentsTotal:u.segmentsTotal}}}catch(e){return za.error(`Compact failed`,A(e)),{content:[{type:`text`,text:`Compact failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Va(e,t,n){let r=L(`scope_map`);e.registerTool(`scope_map`,{title:r.title,description:`Generate a task-scoped reading plan. Given a task description, identifies which files and sections are relevant, with estimated token counts and suggested reading order.`,outputSchema:vi,inputSchema:{task:F.string().describe(`Description of the task to scope`),max_files:F.number().min(1).max(50).default(15).describe(`Maximum files to include`),content_type:F.enum(b).optional().describe(`Filter by content type`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`)},annotations:r.annotations},async({task:e,max_files:r,content_type:i,max_tokens:a})=>{try{let o=await Ft(t,n,{task:e,maxFiles:r,contentType:i}),s=[`## Scope Map: ${e}`,`Total estimated tokens: ~${o.totalEstimatedTokens}`,``,`### Files (by relevance)`,...o.files.map((e,t)=>`${t+1}. **${e.path}** (~${e.estimatedTokens} tokens, ${(e.relevance*100).toFixed(0)}% relevant)\n ${e.reason}\n Focus: ${e.focusRanges.map(e=>`L${e.start}-${e.end}`).join(`, `)}`),``,`### Suggested Reading Order`,...o.readingOrder.map((e,t)=>`${t+1}. ${e}`),``,`### Suggested Compact Calls`,`_Estimated compressed total: ~${Math.ceil(o.totalEstimatedTokens/5)} tokens_`,...o.compactCommands.map((e,t)=>`${t+1}. ${e}`)].join(`
532
+ `)+"\n\n---\n_Next: Use `search` to dive into specific files, or `compact` to compress file contents for context._";return{content:[{type:`text`,text:a?Xt(s,a):s}],structuredContent:{files:o.files.map(e=>({path:e.path,relevance:e.relevance,estimatedTokens:e.estimatedTokens,...e.focusRanges.length>0?{focusLines:e.focusRanges.map(e=>`L${e.start}-${e.end}`)}:{}})),totalFiles:o.files.length,totalEstimatedTokens:o.totalEstimatedTokens,task:e}}}catch(e){return za.error(`Scope map failed`,A(e)),{content:[{type:`text`,text:`Scope map failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ha(e,t,n){let r=L(`find`);e.registerTool(`find`,{title:r.title,description:`Multi-strategy search combining vector, FTS, glob, and regex. Use for precise queries needing multiple strategies. mode=examples finds real usage of a symbol. For general discovery use search instead.`,outputSchema:gi,inputSchema:{query:F.string().optional().describe(`Semantic/keyword search query (required for mode "examples")`),glob:F.string().optional().describe(`File glob pattern (search mode only)`),pattern:F.string().optional().describe(`Regex pattern to match in content (search mode only)`),limit:F.number().min(1).max(50).default(10).describe(`Max results`),content_type:F.enum(b).optional().describe(`Filter by content type`),mode:F.enum([`search`,`examples`]).default(`search`).describe(`Mode: "search" (default) for federated search, "examples" to find usage examples of a symbol/pattern`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all. User-level mode only.`)},annotations:r.annotations},async({query:e,glob:r,pattern:i,limit:a,content_type:o,mode:s,max_tokens:c,workspaces:l})=>{try{if(s===`examples`){if(!e)return{content:[{type:`text`,text:`Error: "query" is required for mode "examples".`}],isError:!0};let r=await We(t,n,{query:e,limit:a,contentType:o}),i=JSON.stringify(r);return{content:[{type:`text`,text:c?Xt(i,c):i}]}}let u=await He(t,n,{query:e,glob:r,pattern:i,limit:a,contentType:o}),d=``;if(l&&l.length>0&&e){let n=Fa(l,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:i}=await Ia(n);try{let i=await La(r,await t.embedQuery(e),{limit:a,contentType:o});for(let e of i)u.results.push({path:`[${e.workspace}] ${e.record.sourcePath}`,score:e.score,source:`cross-workspace`,lineRange:e.record.startLine?{start:e.record.startLine,end:e.record.endLine}:void 0,preview:e.record.content.slice(0,200)});u.results.sort((e,t)=>t.score-e.score),u.results=u.results.slice(0,a),u.totalFound=u.results.length,d=` + ${n.length} workspace(s)`}finally{await i()}}}if(u.results.length===0)return{content:[{type:`text`,text:`No results found.${u.failedStrategies?.length?`\nStrategies attempted: ${u.strategies.join(`, `)}\nFailed: ${u.failedStrategies.map(e=>`${e.strategy} (${e.reason})`).join(`, `)}`:u.strategies.length===0?`
533
533
  No search strategies were activated. Provide at least one of: query, glob, or pattern.`:``}`}],structuredContent:{matches:[],totalMatches:0,pattern:e??r??i??``,truncated:!1}};let f=[`Found ${u.totalFound} results via ${u.strategies.join(` + `)}${d}`,``,...u.results.map(e=>{let t=e.lineRange?`:${e.lineRange.start}-${e.lineRange.end}`:``,n=e.preview?`\n ${e.preview.slice(0,100)}...`:``;return`- [${e.source}] ${e.path}${t} (${(e.score*100).toFixed(0)}%)${n}`})];return{content:[{type:`text`,text:c?Xt(f.join(`
534
534
  `),c):f.join(`
535
- `)}],structuredContent:{matches:u.results.map(e=>({path:e.path,...e.lineRange?{line:e.lineRange.start}:{},matchType:e.source,preview:e.preview?.slice(0,200)??``})),totalMatches:u.totalFound,pattern:e??r??i??``,truncated:u.results.length<u.totalFound}}}catch(e){return Ba.error(`Find failed`,A(e)),{content:[{type:`text`,text:`Find failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wa(e,t,n,r){let i=I(`symbol`);e.registerTool(`symbol`,{title:i.title,description:`Find definition, imports, and references of a named symbol (function, class, type). For tracing data flow across call sites use trace instead.`,inputSchema:{name:F.string().describe(`Symbol name to look up (function, class, type, etc.)`),limit:F.number().min(1).max(50).default(20).describe(`Max results per category`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all. User-level mode only.`)},outputSchema:pi,annotations:i.annotations},async({name:e,limit:i,workspaces:a})=>{try{let o=await Kt(t,n,{name:e,limit:i,graphStore:r});if(a&&a.length>0){let n=Ia(a,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:a}=await La(n);try{for(let[n,a]of r){let r=await Kt(t,a,{name:e,limit:i});r.definedIn&&!o.definedIn&&(o.definedIn={...r.definedIn,path:`[${n}] ${r.definedIn.path}`});for(let e of r.referencedIn)o.referencedIn.push({...e,path:`[${n}] ${e.path}`});if(r.importedBy){o.importedBy=o.importedBy??[];for(let e of r.importedBy)o.importedBy.push({...e,path:`[${n}] ${e.path}`})}}}finally{await a()}}}let s={name:o.name,definedIn:o.definedIn??null,importedBy:o.importedBy??[],referencedIn:o.referencedIn??[],graphContext:o.graphContext??null};return{content:[{type:`text`,text:Ja(o)}],structuredContent:s}}catch(e){return Ba.error(`Symbol lookup failed`,A(e)),{content:[{type:`text`,text:`Symbol lookup failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ga(e,t,n){let r=new mn,i=I(`file_summary`);e.registerTool(`file_summary`,{title:i.title,description:`Create a concise structural summary of a source file: imports, exports, functions, classes, interfaces, and types.`,outputSchema:ki,inputSchema:{path:F.string().describe(`Absolute path to the file to summarize`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response.`)},annotations:i.annotations},async({path:e,max_tokens:i})=>{try{let a=h(e)?e:_(n,e);if((await ne(a)).isDirectory()){let t=await r.analyze(a,{format:`markdown`,maxDepth:3,maxTokens:i}),n=t.data.stats?.languages??{},o=Object.entries(n).sort((e,t)=>t[1]-e[1])[0]?.[0],s=`📁 **Directory**: \`${e}\`\n\n${t.output}\n\n---\n_Path is a directory. Showing structure analysis. Use \`analyze_structure\` for deeper analysis with custom depth._`;return{content:[{type:`text`,text:i?Xt(s,i):s}],structuredContent:{path:a,language:o??`directory`,lines:0,imports:0,exports:0,functions:0,classes:0}}}let o=await Ve({path:a,content:(await t.get(a)).content});return{content:[{type:`text`,text:Ya(o)}],structuredContent:{path:o.path,language:o.language,lines:o.lines,imports:o.imports?.length??0,exports:o.exports?.length??0,functions:o.functions?.length??0,classes:o.classes?.length??0}}}catch(e){return Ba.error(`File summary failed`,A(e)),{content:[{type:`text`,text:`File summary failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ka(e,t,n,r){let i=I(`trace`);e.registerTool(`trace`,{title:i.title,description:`Follow data flow forward/backward across imports and call sites from a starting symbol or file:line. For finding a single symbol definition use symbol instead.`,inputSchema:{start:F.string().describe(`Starting point — symbol name or file:line reference`),direction:F.enum([`forward`,`backward`,`both`]).describe(`Which direction to trace relationships`),max_depth:F.number().min(1).max(10).default(3).optional().describe(`Maximum trace depth`)},annotations:i.annotations},async({start:e,direction:i,max_depth:a})=>{try{let o=await Yt(t,n,{start:e,direction:i,maxDepth:a,graphStore:r}),s=[`## Trace: ${o.start}`,`Direction: ${o.direction} | Depth: ${o.depth}`,``];if(o.graphContext&&(s.push(`### Graph Context`),o.graphContext.definingModule&&s.push(`- Module: ${o.graphContext.definingModule}`),o.graphContext.community&&s.push(`- Community: ${o.graphContext.community}`),o.graphContext.importedByModules.length>0&&s.push(`- Imported by modules: ${o.graphContext.importedByModules.join(`, `)}`),o.graphContext.siblingSymbols.length>0&&s.push(`- Sibling symbols: ${o.graphContext.siblingSymbols.join(`, `)}`),s.push(``)),o.nodes.length===0)s.push(`No connections found.`);else{let e=o.nodes.filter(e=>e.relationship===`calls`),t=o.nodes.filter(e=>e.relationship===`called-by`),n=o.nodes.filter(e=>e.relationship===`imports`),r=o.nodes.filter(e=>e.relationship===`imported-by`),i=o.nodes.filter(e=>e.relationship===`references`);if(e.length>0){s.push(`### Calls (${e.length})`);for(let t of e){let e=t.scope?` (from ${t.scope}())`:``;s.push(`- ${t.symbol}() — ${t.path}:${t.line}${e}`)}s.push(``)}if(t.length>0){s.push(`### Called by (${t.length})`);for(let e of t){let t=e.scope?` in ${e.scope}()`:``;s.push(`- ${e.symbol}()${t} — ${e.path}:${e.line}`)}s.push(``)}if(n.length>0){s.push(`### Imports (${n.length})`);for(let e of n)s.push(`- ${e.symbol} — ${e.path}:${e.line}`);s.push(``)}if(r.length>0){s.push(`### Imported by (${r.length})`);for(let e of r)s.push(`- ${e.path}:${e.line}`);s.push(``)}if(i.length>0){s.push(`### References (${i.length})`);for(let e of i)s.push(`- ${e.path}:${e.line}`);s.push(``)}}return s.push(`---`,"_Next: `symbol` for definition details | `compact` to read a referenced file | `blast_radius` for impact analysis_"),{content:[{type:`text`,text:s.join(`
536
- `)}],structuredContent:{start:o.start,direction:o.direction,depth:o.depth,nodes:o.nodes,graphContext:o.graphContext??null}}}catch(e){return Ba.error(`Trace failed`,A(e)),{content:[{type:`text`,text:`Trace failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function qa(e,t,n){let r=I(`dead_symbols`);e.registerTool(`dead_symbols`,{title:r.title,description:`Find exported symbols that appear to be unused because they are never imported or re-exported.`,outputSchema:Di,inputSchema:{path:F.string().optional().describe(`Root path to scope the search (default: cwd)`),limit:F.number().min(1).max(500).default(100).optional().describe(`Maximum exported symbols to scan`)},annotations:r.annotations},async({path:e,limit:r})=>{try{let i=await Ue(t,n,{rootPath:e,limit:r}),a=[`## Dead Symbol Analysis`,``,`**Exports scanned:** ${i.totalExports}`,`**Dead in source:** ${i.totalDeadSource} (actionable)`,`**Dead in docs:** ${i.totalDeadDocs} (informational — code samples in .md files)`,``];if(i.deadInSource.length>0){a.push(`### Dead in Source (actionable)`);for(let e of i.deadInSource)a.push(`- \`${e.name}\` (${e.kind}) — ${e.path}:${e.line}`);a.push(``)}if(i.deadInDocs.length>0){a.push(`### Dead in Docs (informational)`),a.push(`_${i.totalDeadDocs} symbol(s) found only in documentation code samples — not actionable dead code._`);for(let e of i.deadInDocs.slice(0,5))a.push(`- \`${e.name}\` — ${e.path}:${e.line}`);i.deadInDocs.length>5&&a.push(`- _... ${i.deadInDocs.length-5} more omitted_`)}return i.totalDeadSource>0?a.push(``,`---`,`_Next: \`codemod\` to remove ${i.totalDeadSource} unused exports | \`symbol\` to verify usage before removing_`):a.push(``,`---`,"_Next: `check` — no dead symbols found, validate types and lint_"),{content:[{type:`text`,text:a.join(`
537
- `)}],structuredContent:{symbols:[...i.deadInSource,...i.deadInDocs].map(e=>({name:e.name,path:e.path,...e.line===void 0?{}:{line:e.line},kind:e.kind})),totalDead:i.totalDeadSource+i.totalDeadDocs}}}catch(e){return Ba.error(`Dead symbol scan failed`,A(e)),{content:[{type:`text`,text:`Dead symbol scan failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ja(e){let t=[`Symbol: ${e.name}`];if(e.definedIn){let n=`Defined in: ${e.definedIn.path}:${e.definedIn.line} (${e.definedIn.kind})`;e.definedIn.signature&&(n+=`\nSignature: ${e.definedIn.signature}`),t.push(n)}else t.push(`Defined in: not found`);if(t.push(``,`Imported by:`),e.importedBy.length===0)t.push(` none`);else for(let n of e.importedBy)t.push(` - ${n.path}:${n.line} ${n.importStatement}`);if(t.push(``,`Referenced in:`),e.referencedIn.length===0)t.push(` none`);else for(let n of e.referencedIn){let e=`scope`in n&&n.scope?` in ${n.scope}()`:``;t.push(` - ${n.path}:${n.line}${e} ${n.context}`)}if(e.graphContext){let n=e.graphContext;t.push(``,`Graph context:`),n.definingModule&&t.push(` Module: ${n.definingModule}`),n.importedByModules.length>0&&t.push(` Imported by modules: ${n.importedByModules.join(`, `)}`),n.siblingSymbols.length>0&&t.push(` Sibling symbols: ${n.siblingSymbols.join(`, `)}`)}return t.join(`
538
- `)}function Ya(e){let t=[e.path,`Language: ${e.language}`,`Lines: ${e.lines}`,`Estimated tokens: ~${e.estimatedTokens}`,``,`Imports (${e.imports.length}):`,...Xa(e.imports),``,`Exports (${e.exports.length}):`,...Xa(e.exports),``,`Functions (${e.functions.length}):`,...Xa(e.functions.map(e=>`${e.name} @ line ${e.line}${e.exported?` [exported]`:``}${`signature`in e&&e.signature?` — ${e.signature}`:``}`)),``,`Classes (${e.classes.length}):`,...Xa(e.classes.map(e=>`${e.name} @ line ${e.line}${e.exported?` [exported]`:``}${e.signature?` — ${e.signature}`:``}`)),``,`Interfaces (${e.interfaces.length}):`,...Xa(e.interfaces.map(e=>`${e.name} @ line ${e.line}${`exported`in e&&e.exported?` [exported]`:``}`)),``,`Types (${e.types.length}):`,...Xa(e.types.map(e=>`${e.name} @ line ${e.line}${`exported`in e&&e.exported?` [exported]`:``}`))];if(`importDetails`in e&&e.importDetails&&e.importDetails.length>0){let n=e.importDetails.filter(e=>e.isExternal).length,r=e.importDetails.length-n;t.push(``,`Import breakdown: ${n} external, ${r} internal`)}if(`callEdges`in e&&e.callEdges&&e.callEdges.length>0){t.push(``,`Call edges (${e.callEdges.length} intra-file):`);for(let n of e.callEdges.slice(0,30))t.push(` - ${n.caller}() → ${n.callee}() @ line ${n.line}`);e.callEdges.length>30&&t.push(` - ... ${e.callEdges.length-30} more`)}return t.join(`
539
- `)}function Xa(e){return e.length===0?[` none`]:e.map(e=>` - ${e}`)}const Za=E(`tools`);function Qa(e,t){let n=I(`er_evolve_review`);e.registerTool(`er_evolve_review`,{title:n.title,description:`Review evolution metrics for the AI Kit ↔ ER bridge. Shows local miss rates, push acceptance, rule effectiveness, and top missed queries. Use this to reason about improving classification rules and knowledge flow.`,inputSchema:{include_details:F.boolean().default(!0).describe(`Include detailed breakdowns (top missed queries, rule stats)`),reset_after:F.boolean().default(!1).describe(`Reset collected metrics after review`)},annotations:n.annotations},async({include_details:e,reset_after:n})=>{try{let r=t.getMetrics(),i=[`## ER Evolution Review
535
+ `)}],structuredContent:{matches:u.results.map(e=>({path:e.path,...e.lineRange?{line:e.lineRange.start}:{},matchType:e.source,preview:e.preview?.slice(0,200)??``})),totalMatches:u.totalFound,pattern:e??r??i??``,truncated:u.results.length<u.totalFound}}}catch(e){return za.error(`Find failed`,A(e)),{content:[{type:`text`,text:`Find failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ua(e,t,n,r){let i=L(`symbol`);e.registerTool(`symbol`,{title:i.title,description:`Find definition, imports, and references of a named symbol (function, class, type). For tracing data flow across call sites use trace instead.`,inputSchema:{name:F.string().describe(`Symbol name to look up (function, class, type, etc.)`),limit:F.number().min(1).max(50).default(20).describe(`Max results per category`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all. User-level mode only.`)},outputSchema:fi,annotations:i.annotations},async({name:e,limit:i,workspaces:a})=>{try{let o=await Kt(t,n,{name:e,limit:i,graphStore:r});if(a&&a.length>0){let n=Fa(a,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:a}=await Ia(n);try{for(let[n,a]of r){let r=await Kt(t,a,{name:e,limit:i});r.definedIn&&!o.definedIn&&(o.definedIn={...r.definedIn,path:`[${n}] ${r.definedIn.path}`});for(let e of r.referencedIn)o.referencedIn.push({...e,path:`[${n}] ${e.path}`});if(r.importedBy){o.importedBy=o.importedBy??[];for(let e of r.importedBy)o.importedBy.push({...e,path:`[${n}] ${e.path}`})}}}finally{await a()}}}let s={name:o.name,definedIn:o.definedIn??null,importedBy:o.importedBy??[],referencedIn:o.referencedIn??[],graphContext:o.graphContext??null};return{content:[{type:`text`,text:qa(o)}],structuredContent:s}}catch(e){return za.error(`Symbol lookup failed`,A(e)),{content:[{type:`text`,text:`Symbol lookup failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wa(e,t,n){let r=new mn,i=L(`file_summary`);e.registerTool(`file_summary`,{title:i.title,description:`Create a concise structural summary of a source file: imports, exports, functions, classes, interfaces, and types.`,outputSchema:Oi,inputSchema:{path:F.string().describe(`Absolute path to the file to summarize`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response.`)},annotations:i.annotations},async({path:e,max_tokens:i})=>{try{let a=h(e)?e:_(n,e);if((await ne(a)).isDirectory()){let t=await r.analyze(a,{format:`markdown`,maxDepth:3,maxTokens:i}),n=t.data.stats?.languages??{},o=Object.entries(n).sort((e,t)=>t[1]-e[1])[0]?.[0],s=`📁 **Directory**: \`${e}\`\n\n${t.output}\n\n---\n_Path is a directory. Showing structure analysis. Use \`analyze_structure\` for deeper analysis with custom depth._`;return{content:[{type:`text`,text:i?Xt(s,i):s}],structuredContent:{path:a,language:o??`directory`,lines:0,imports:0,exports:0,functions:0,classes:0}}}let o=await Ve({path:a,content:(await t.get(a)).content});return{content:[{type:`text`,text:Ja(o)}],structuredContent:{path:o.path,language:o.language,lines:o.lines,imports:o.imports?.length??0,exports:o.exports?.length??0,functions:o.functions?.length??0,classes:o.classes?.length??0}}}catch(e){return za.error(`File summary failed`,A(e)),{content:[{type:`text`,text:`File summary failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ga(e,t,n,r){let i=L(`trace`);e.registerTool(`trace`,{title:i.title,description:`Follow data flow forward/backward across imports and call sites from a starting symbol or file:line. For finding a single symbol definition use symbol instead.`,inputSchema:{start:F.string().describe(`Starting point — symbol name or file:line reference`),direction:F.enum([`forward`,`backward`,`both`]).describe(`Which direction to trace relationships`),max_depth:F.number().min(1).max(10).default(3).optional().describe(`Maximum trace depth`)},annotations:i.annotations},async({start:e,direction:i,max_depth:a})=>{try{let o=await Yt(t,n,{start:e,direction:i,maxDepth:a,graphStore:r}),s=[`## Trace: ${o.start}`,`Direction: ${o.direction} | Depth: ${o.depth}`,``];if(o.graphContext&&(s.push(`### Graph Context`),o.graphContext.definingModule&&s.push(`- Module: ${o.graphContext.definingModule}`),o.graphContext.community&&s.push(`- Community: ${o.graphContext.community}`),o.graphContext.importedByModules.length>0&&s.push(`- Imported by modules: ${o.graphContext.importedByModules.join(`, `)}`),o.graphContext.siblingSymbols.length>0&&s.push(`- Sibling symbols: ${o.graphContext.siblingSymbols.join(`, `)}`),s.push(``)),o.nodes.length===0)s.push(`No connections found.`);else{let e=o.nodes.filter(e=>e.relationship===`calls`),t=o.nodes.filter(e=>e.relationship===`called-by`),n=o.nodes.filter(e=>e.relationship===`imports`),r=o.nodes.filter(e=>e.relationship===`imported-by`),i=o.nodes.filter(e=>e.relationship===`references`);if(e.length>0){s.push(`### Calls (${e.length})`);for(let t of e){let e=t.scope?` (from ${t.scope}())`:``;s.push(`- ${t.symbol}() — ${t.path}:${t.line}${e}`)}s.push(``)}if(t.length>0){s.push(`### Called by (${t.length})`);for(let e of t){let t=e.scope?` in ${e.scope}()`:``;s.push(`- ${e.symbol}()${t} — ${e.path}:${e.line}`)}s.push(``)}if(n.length>0){s.push(`### Imports (${n.length})`);for(let e of n)s.push(`- ${e.symbol} — ${e.path}:${e.line}`);s.push(``)}if(r.length>0){s.push(`### Imported by (${r.length})`);for(let e of r)s.push(`- ${e.path}:${e.line}`);s.push(``)}if(i.length>0){s.push(`### References (${i.length})`);for(let e of i)s.push(`- ${e.path}:${e.line}`);s.push(``)}}return s.push(`---`,"_Next: `symbol` for definition details | `compact` to read a referenced file | `blast_radius` for impact analysis_"),{content:[{type:`text`,text:s.join(`
536
+ `)}],structuredContent:{start:o.start,direction:o.direction,depth:o.depth,nodes:o.nodes,graphContext:o.graphContext??null}}}catch(e){return za.error(`Trace failed`,A(e)),{content:[{type:`text`,text:`Trace failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ka(e,t,n){let r=L(`dead_symbols`);e.registerTool(`dead_symbols`,{title:r.title,description:`Find exported symbols that appear to be unused because they are never imported or re-exported.`,outputSchema:Ei,inputSchema:{path:F.string().optional().describe(`Root path to scope the search (default: cwd)`),limit:F.number().min(1).max(500).default(100).optional().describe(`Maximum exported symbols to scan`)},annotations:r.annotations},async({path:e,limit:r})=>{try{let i=await Ue(t,n,{rootPath:e,limit:r}),a=[`## Dead Symbol Analysis`,``,`**Exports scanned:** ${i.totalExports}`,`**Dead in source:** ${i.totalDeadSource} (actionable)`,`**Dead in docs:** ${i.totalDeadDocs} (informational — code samples in .md files)`,``];if(i.deadInSource.length>0){a.push(`### Dead in Source (actionable)`);for(let e of i.deadInSource)a.push(`- \`${e.name}\` (${e.kind}) — ${e.path}:${e.line}`);a.push(``)}if(i.deadInDocs.length>0){a.push(`### Dead in Docs (informational)`),a.push(`_${i.totalDeadDocs} symbol(s) found only in documentation code samples — not actionable dead code._`);for(let e of i.deadInDocs.slice(0,5))a.push(`- \`${e.name}\` — ${e.path}:${e.line}`);i.deadInDocs.length>5&&a.push(`- _... ${i.deadInDocs.length-5} more omitted_`)}return i.totalDeadSource>0?a.push(``,`---`,`_Next: \`codemod\` to remove ${i.totalDeadSource} unused exports | \`symbol\` to verify usage before removing_`):a.push(``,`---`,"_Next: `check` — no dead symbols found, validate types and lint_"),{content:[{type:`text`,text:a.join(`
537
+ `)}],structuredContent:{symbols:[...i.deadInSource,...i.deadInDocs].map(e=>({name:e.name,path:e.path,...e.line===void 0?{}:{line:e.line},kind:e.kind})),totalDead:i.totalDeadSource+i.totalDeadDocs}}}catch(e){return za.error(`Dead symbol scan failed`,A(e)),{content:[{type:`text`,text:`Dead symbol scan failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function qa(e){let t=[`Symbol: ${e.name}`];if(e.definedIn){let n=`Defined in: ${e.definedIn.path}:${e.definedIn.line} (${e.definedIn.kind})`;e.definedIn.signature&&(n+=`\nSignature: ${e.definedIn.signature}`),t.push(n)}else t.push(`Defined in: not found`);if(t.push(``,`Imported by:`),e.importedBy.length===0)t.push(` none`);else for(let n of e.importedBy)t.push(` - ${n.path}:${n.line} ${n.importStatement}`);if(t.push(``,`Referenced in:`),e.referencedIn.length===0)t.push(` none`);else for(let n of e.referencedIn){let e=`scope`in n&&n.scope?` in ${n.scope}()`:``;t.push(` - ${n.path}:${n.line}${e} ${n.context}`)}if(e.graphContext){let n=e.graphContext;t.push(``,`Graph context:`),n.definingModule&&t.push(` Module: ${n.definingModule}`),n.importedByModules.length>0&&t.push(` Imported by modules: ${n.importedByModules.join(`, `)}`),n.siblingSymbols.length>0&&t.push(` Sibling symbols: ${n.siblingSymbols.join(`, `)}`)}return t.join(`
538
+ `)}function Ja(e){let t=[e.path,`Language: ${e.language}`,`Lines: ${e.lines}`,`Estimated tokens: ~${e.estimatedTokens}`,``,`Imports (${e.imports.length}):`,...Ya(e.imports),``,`Exports (${e.exports.length}):`,...Ya(e.exports),``,`Functions (${e.functions.length}):`,...Ya(e.functions.map(e=>`${e.name} @ line ${e.line}${e.exported?` [exported]`:``}${`signature`in e&&e.signature?` — ${e.signature}`:``}`)),``,`Classes (${e.classes.length}):`,...Ya(e.classes.map(e=>`${e.name} @ line ${e.line}${e.exported?` [exported]`:``}${e.signature?` — ${e.signature}`:``}`)),``,`Interfaces (${e.interfaces.length}):`,...Ya(e.interfaces.map(e=>`${e.name} @ line ${e.line}${`exported`in e&&e.exported?` [exported]`:``}`)),``,`Types (${e.types.length}):`,...Ya(e.types.map(e=>`${e.name} @ line ${e.line}${`exported`in e&&e.exported?` [exported]`:``}`))];if(`importDetails`in e&&e.importDetails&&e.importDetails.length>0){let n=e.importDetails.filter(e=>e.isExternal).length,r=e.importDetails.length-n;t.push(``,`Import breakdown: ${n} external, ${r} internal`)}if(`callEdges`in e&&e.callEdges&&e.callEdges.length>0){t.push(``,`Call edges (${e.callEdges.length} intra-file):`);for(let n of e.callEdges.slice(0,30))t.push(` - ${n.caller}() → ${n.callee}() @ line ${n.line}`);e.callEdges.length>30&&t.push(` - ... ${e.callEdges.length-30} more`)}return t.join(`
539
+ `)}function Ya(e){return e.length===0?[` none`]:e.map(e=>` - ${e}`)}const Xa=E(`tools`);function Za(e,t){let n=L(`er_evolve_review`);e.registerTool(`er_evolve_review`,{title:n.title,description:`Review evolution metrics for the AI Kit ↔ ER bridge. Shows local miss rates, push acceptance, rule effectiveness, and top missed queries. Use this to reason about improving classification rules and knowledge flow.`,inputSchema:{include_details:F.boolean().default(!0).describe(`Include detailed breakdowns (top missed queries, rule stats)`),reset_after:F.boolean().default(!1).describe(`Reset collected metrics after review`)},annotations:n.annotations},async({include_details:e,reset_after:n})=>{try{let r=t.getMetrics(),i=[`## ER Evolution Review
540
540
  `,`**Period**: ${r.period.startedAt} → ${r.period.queriedAt}`,`**Total events**: ${r.period.totalEvents}\n`,`### Search`,`- Total searches: ${r.search.totalSearches}`,`- ER fallback triggered: ${r.search.erFallbackCount} (${(r.search.erFallbackRate*100).toFixed(1)}%)`,`- ER cache hits: ${r.search.erCacheHitCount} (${(r.search.erCacheHitRate*100).toFixed(1)}% of fallbacks)`];if(e&&r.search.topMissedQueries.length>0){i.push(`
541
- **Top missed queries** (triggered ER fallback):`);for(let e of r.search.topMissedQueries.slice(0,10)){let t=e.query.length>60?`${e.query.slice(0,57)}...`:e.query;i.push(` - "${t}" (${e.count}x)`)}}if(i.push(``,`### Push`,`- Total pushes: ${r.push.totalPushes} (${r.push.successCount} ok, ${r.push.failCount} failed)`,`- Classification match rate: ${(r.push.classificationMatchRate*100).toFixed(1)}%`,`- Push acceptance rate: ${(r.push.pushAcceptanceRate*100).toFixed(1)}%`),i.push(``,`### Rule Effectiveness`),Object.keys(r.rules.matchCounts).length>0)for(let[e,t]of Object.entries(r.rules.matchCounts)){let n=r.rules.pushCounts[e]??0,a=t>0?(n/t*100).toFixed(0):`0`;i.push(`- **${e}**: ${t} matches → ${n} pushes (${a}% conversion)`)}else i.push(`- _No rule activity recorded yet_`);if(e&&r.rules.lowConversionRules.length>0){i.push(``,`### ⚠️ Low Conversion Rules (potential false positives)`);for(let e of r.rules.lowConversionRules)i.push(`- **${e.ruleId}**: ${e.matchCount} matches, ${e.pushCount} pushes (${(e.conversionRate*100).toFixed(0)}% conversion) — consider tightening patterns`)}return i.push(``,`---`,"_Next: Use `er_update_policy` to refine rules based on these metrics, or `er_push` to share high-value knowledge._"),n&&(Za.info(`Evolution metrics reset requested`,{requestedAt:new Date().toISOString(),clearedEvents:r.period.totalEvents}),t.reset(),i.push(`
541
+ **Top missed queries** (triggered ER fallback):`);for(let e of r.search.topMissedQueries.slice(0,10)){let t=e.query.length>60?`${e.query.slice(0,57)}...`:e.query;i.push(` - "${t}" (${e.count}x)`)}}if(i.push(``,`### Push`,`- Total pushes: ${r.push.totalPushes} (${r.push.successCount} ok, ${r.push.failCount} failed)`,`- Classification match rate: ${(r.push.classificationMatchRate*100).toFixed(1)}%`,`- Push acceptance rate: ${(r.push.pushAcceptanceRate*100).toFixed(1)}%`),i.push(``,`### Rule Effectiveness`),Object.keys(r.rules.matchCounts).length>0)for(let[e,t]of Object.entries(r.rules.matchCounts)){let n=r.rules.pushCounts[e]??0,a=t>0?(n/t*100).toFixed(0):`0`;i.push(`- **${e}**: ${t} matches → ${n} pushes (${a}% conversion)`)}else i.push(`- _No rule activity recorded yet_`);if(e&&r.rules.lowConversionRules.length>0){i.push(``,`### ⚠️ Low Conversion Rules (potential false positives)`);for(let e of r.rules.lowConversionRules)i.push(`- **${e.ruleId}**: ${e.matchCount} matches, ${e.pushCount} pushes (${(e.conversionRate*100).toFixed(0)}% conversion) — consider tightening patterns`)}return i.push(``,`---`,"_Next: Use `er_update_policy` to refine rules based on these metrics, or `er_push` to share high-value knowledge._"),n&&(Xa.info(`Evolution metrics reset requested`,{requestedAt:new Date().toISOString(),clearedEvents:r.period.totalEvents}),t.reset(),i.push(`
542
542
  _Metrics have been reset._`)),{content:[{type:`text`,text:i.join(`
543
- `)}]}}catch(e){return Za.error(`Evolution review failed`,A(e)),{content:[{type:`text`,text:`Evolution review failed: unable to compute metrics`}],isError:!0}}})}const $a=E(`tools:execution`);function eo(e){let t=I(`check`);e.registerTool(`check`,{title:t.title,description:`Run incremental typecheck (tsc) and lint (biome) on the project or specific files. Returns structured error and warning lists. Default detail level is "summary" (~300 tokens).`,inputSchema:{files:F.array(F.string()).optional().describe(`Specific files to check (if omitted, checks all)`),cwd:F.string().optional().describe(`Working directory`),skip_types:F.boolean().default(!1).describe(`Skip TypeScript typecheck`),skip_lint:F.boolean().default(!1).describe(`Skip Biome lint`),detail:F.enum([`summary`,`errors`,`full`]).default(`summary`).describe(`Output detail level: summary (default, ~300 tokens — pass/fail + counts + top errors), errors (parsed error objects), full (includes raw terminal output)`)},outputSchema:fi,annotations:t.annotations},async({files:e,cwd:t,skip_types:n,skip_lint:r,detail:i},a)=>{try{let o=await Se({files:e,cwd:t,skipTypes:n,skipLint:r,detail:i===`summary`?`errors`:i}),s=Yi(a).createTask(`Check`,2);if(s.progress(0,`tsc: ${o.tsc.errors.length} errors`),s.progress(1,`biome: ${o.biome.errors.length} errors`),s.complete(`Check ${o.passed?`passed`:`failed`}`),i===`summary`){let e=Gt(o),t=[];if(o.passed)t.push({tool:`test_run`,reason:`Types and lint clean — run tests next`});else{let e=o.tsc.errors[0]?.file??o.biome.errors[0]?.file;e&&t.push({tool:`symbol`,reason:`Resolve failing symbol in ${e}`,suggested_args:{name:e}}),t.push({tool:`check`,reason:`Re-check after fixing errors`,suggested_args:{detail:`errors`}})}let n=[`## Check ${e.passed?`✅ PASS`:`❌ FAIL`}`,``,`**tsc**: ${e.tsc.passed?`✅`:`❌`} ${e.tsc.errorCount} errors, ${e.tsc.warningCount} warnings`];if(e.tsc.topErrors.length>0)for(let t of e.tsc.topErrors)n.push(` - ${t}`);if(n.push(`**biome**: ${e.biome.passed?`✅`:`❌`} ${e.biome.errorCount} errors, ${e.biome.warningCount} warnings`),e.biome.topErrors.length>0)for(let t of e.biome.topErrors)n.push(` - ${t}`);if(t.length>0){n.push(``,`**Next steps:**`);for(let e of t)n.push(`- \`${e.tool}\`: ${e.reason}`)}return{content:[{type:`text`,text:n.join(`
544
- `)}],structuredContent:{...e}}}return{content:[{type:`text`,text:JSON.stringify(o)}]}}catch(e){return $a.error(`Check failed`,A(e)),{content:[{type:`text`,text:`Check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function to(e,t,n){let r=I(`batch`);e.registerTool(`batch`,{title:r.title,description:`Execute multiple built-in operations in parallel with concurrency control. Supported operation types: search, find, and check.`,inputSchema:{operations:F.array(F.object({id:F.string().describe(`Unique ID for this operation`),type:F.enum([`search`,`find`,`check`]).describe(`Built-in operation type`),args:F.record(F.string(),F.unknown()).describe(`Arguments for the operation`)})).min(1).max(100).describe(`Operations to execute`),concurrency:F.number().min(1).max(20).default(4).describe(`Max concurrent operations`)},annotations:r.annotations},async({operations:e,concurrency:r},i)=>{try{let a=await ye(e,async e=>lo(e,t,n),{concurrency:r}),o=Yi(i).createTask(`Batch`,e.length);for(let e=0;e<a.length;e++)o.progress(e,`${a[e].id}: ${a[e].status}`);return o.complete(`Batch complete: ${a.length} operations`),{content:[{type:`text`,text:JSON.stringify(a)}]}}catch(e){return $a.error(`Batch failed`,A(e)),{content:[{type:`text`,text:`Batch failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function no(e){let t=I(`eval`);e.registerTool(`eval`,{title:t.title,description:`Execute a JavaScript or TypeScript snippet in a constrained VM sandbox with a timeout. Captures console output and returned values.`,inputSchema:{code:F.string().max(1e5).describe(`Code snippet to execute`),lang:F.enum([`js`,`ts`]).default(`js`).optional().describe(`Language mode: js executes directly, ts strips common type syntax first`),timeout:F.number().min(1).max(6e4).default(5e3).optional().describe(`Execution timeout in milliseconds`)},annotations:t.annotations},async({code:e,lang:t,timeout:n})=>{try{let r=ze({code:e,lang:t,timeout:n});return r.success?{content:[{type:`text`,text:`Eval succeeded in ${r.durationMs}ms\n\n${r.output}`}]}:{content:[{type:`text`,text:`Eval failed in ${r.durationMs}ms: ${r.error??`Unknown error`}`}],isError:!0}}catch(e){return $a.error(`Eval failed`,A(e)),{content:[{type:`text`,text:`Eval failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ro(e){let t=I(`test_run`);e.registerTool(`test_run`,{title:t.title,description:`Run Vitest for the current project or a subset of files, then return a structured summary of passing and failing tests.`,inputSchema:{files:F.array(F.string()).optional().describe(`Specific test files or patterns to run`),grep:F.string().optional().describe(`Only run tests whose names match this pattern`),cwd:F.string().optional().describe(`Working directory for the test run`)},annotations:t.annotations},async({files:e,grep:t,cwd:n},r)=>{try{let i=await qt({files:e,grep:t,cwd:n}),a=Yi(r).createTask(`Test Run`,1),o=i.summary.passed+i.summary.failed+i.summary.skipped;return a.complete(`Tests: ${i.passed?`passed`:`failed`} (${o} tests)`),{content:[{type:`text`,text:po(i)}],isError:!i.passed}}catch(e){return $a.error(`Test run failed`,A(e)),{content:[{type:`text`,text:`Test run failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function io(e){let t=I(`parse_output`);e.registerTool(`parse_output`,{title:t.title,description:`Parse structured data from build tool output. Supports tsc, vitest, biome, and git status. Auto-detects the tool or specify explicitly.`,inputSchema:{output:F.string().max(5e5).describe(`Raw output text from a build tool`),tool:F.enum([`tsc`,`vitest`,`biome`,`git-status`]).optional().describe(`Tool to parse as (auto-detects if omitted)`)},annotations:t.annotations},async({output:e,tool:t})=>{try{let n=ut(e.replace(/\\n/g,`
545
- `).replace(/\\t/g,` `),t);return{content:[{type:`text`,text:JSON.stringify(n)}]}}catch(e){return $a.error(`Parse failed`,A(e)),{content:[{type:`text`,text:`Parse failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ao(e,t){let n=I(`delegate`);e.registerTool(`delegate`,{title:n.title,description:`Delegate a subtask to a local Ollama model. Use for summarization, classification, naming, or any task that can offload work from the host agent. Fails fast if Ollama is not running.`,inputSchema:{prompt:F.string().max(2e5).describe(`The task or question to send to the local model`),model:F.string().optional().describe(`Ollama model name (default: first available model)`),system:F.string().optional().describe(`System prompt for the model`),context:F.string().max(5e5).optional().describe(`Context text to include before the prompt (e.g. file contents)`),temperature:F.number().min(0).max(2).default(.3).optional().describe(`Sampling temperature (0=deterministic, default 0.3)`),timeout:F.number().min(1e3).max(6e5).default(12e4).optional().describe(`Timeout in milliseconds (default 120000)`),action:F.enum([`generate`,`list_models`]).default(`generate`).optional().describe(`Action: generate a response or list available models`)},annotations:n.annotations},async({prompt:e,model:n,system:r,context:i,temperature:a,timeout:o,action:s})=>{try{if(s===`list_models`){let e=await Ne();return{content:[{type:`text`,text:JSON.stringify({models:e,count:e.length,_Next:`Use delegate with a model name`},null,2)}]}}if(t?.available)try{let n=await t.createMessage({prompt:e,systemPrompt:r,context:i,maxTokens:4e3,modelPreferences:{intelligencePriority:.8,speedPriority:.5,costPriority:.3}});return{content:[{type:`text`,text:JSON.stringify({model:n.model??`sampling`,response:n.text,provider:`mcp-sampling`,_Next:`Use the response in your workflow. stash to save it.`},null,2)}]}}catch{$a.debug(`Sampling failed, falling back to Ollama`)}let c=await Me({prompt:e,model:n,system:r,context:i,temperature:a,timeout:o});return c.error?{content:[{type:`text`,text:JSON.stringify({error:c.error,model:c.model,durationMs:c.durationMs},null,2)}],isError:!0}:{content:[{type:`text`,text:JSON.stringify({model:c.model,response:c.response,durationMs:c.durationMs,tokenCount:c.tokenCount,_Next:`Use the response in your workflow. stash to save it.`},null,2)}]}}catch(e){return $a.error(`Delegate failed`,A(e)),{content:[{type:`text`,text:`Delegate failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const oo=F.object({query:F.string(),limit:F.number().min(1).max(20).default(5).optional(),search_mode:F.enum([`hybrid`,`semantic`,`keyword`]).default(`hybrid`).optional(),content_type:F.enum(b).optional(),origin:F.enum([`indexed`,`curated`,`produced`]).optional(),category:F.string().optional(),tags:F.array(F.string()).optional(),min_score:F.number().min(0).max(1).default(.25).optional()}),so=F.object({query:F.string().optional(),glob:F.string().optional(),pattern:F.string().optional(),limit:F.number().min(1).max(50).default(10).optional(),content_type:F.enum(b).optional(),cwd:F.string().optional()}),co=F.object({files:F.array(F.string()).optional(),cwd:F.string().optional(),skip_types:F.boolean().optional(),skip_lint:F.boolean().optional()});async function lo(e,t,n){switch(e.type){case`search`:return uo(t,n,oo.parse(e.args));case`find`:{let r=so.parse(e.args);if(!r.query&&!r.glob&&!r.pattern)throw Error(`find operation requires query, glob, or pattern`);return He(t,n,{query:r.query,glob:r.glob,pattern:r.pattern,limit:r.limit,contentType:r.content_type,cwd:r.cwd})}case`check`:{let t=co.parse(e.args);return Se({files:t.files,cwd:t.cwd,skipTypes:t.skip_types,skipLint:t.skip_lint})}default:throw Error(`Unsupported batch operation type: ${e.type}`)}}async function uo(e,t,n){let r=n.limit??5,i={limit:r,minScore:n.min_score??.25,contentType:n.content_type,origin:n.origin,category:n.category,tags:n.tags},a=e.embedQuery?.bind(e)??e.embed.bind(e);if(n.search_mode===`keyword`)return(await t.ftsSearch(n.query,i)).slice(0,r);let o=await a(n.query);if(n.search_mode===`semantic`)return t.search(o,i);let[s,c]=await Promise.all([t.search(o,{...i,limit:r*2}),t.ftsSearch(n.query,{...i,limit:r*2})]);return fo(s,c).slice(0,r)}function fo(e,t,n=60){let r=new Map;for(let t=0;t<e.length;t++){let i=e[t];r.set(i.record.id,{record:i.record,score:1/(n+t+1)})}for(let e=0;e<t.length;e++){let i=t[e],a=r.get(i.record.id);if(a){a.score+=1/(n+e+1);continue}r.set(i.record.id,{record:i.record,score:1/(n+e+1)})}return[...r.values()].sort((e,t)=>t.score-e.score)}function po(e){let t=[`Vitest run: ${e.passed?`passed`:`failed`}`,`Duration: ${e.durationMs}ms`,`Passed: ${e.summary.passed}`,`Failed: ${e.summary.failed}`,`Skipped: ${e.summary.skipped}`];e.summary.suites!==void 0&&t.push(`Suites: ${e.summary.suites}`);let n=e.summary.tests.filter(e=>e.status===`fail`);if(n.length>0){t.push(``,`Failed tests:`);for(let e of n)t.push(`- ${e.name}${e.file?` (${e.file})`:``}`),e.error&&t.push(` ${e.error}`)}return t.join(`
546
- `)}const H=E(`flow-tools`);function mo(e){return Array.isArray(e)?e.filter(e=>typeof e==`string`&&e.trim().length>0):[]}function ho(e){let t=Ur(e);return async({content:e,sourceFormat:n,stepName:r,metadata:i})=>{let a=r.charAt(0).toUpperCase()+r.slice(1),o=i?JSON.stringify(i,null,2):void 0,s=`You are converting a flow step definition from "${n}" format to aikit native flow step README.md format.
543
+ `)}]}}catch(e){return Xa.error(`Evolution review failed`,A(e)),{content:[{type:`text`,text:`Evolution review failed: unable to compute metrics`}],isError:!0}}})}const Qa=E(`tools:execution`);function $a(e){let t=L(`check`);e.registerTool(`check`,{title:t.title,description:`Run incremental typecheck (tsc) and lint (biome) on the project or specific files. Returns structured error and warning lists. Default detail level is "summary" (~300 tokens).`,inputSchema:{files:F.array(F.string()).optional().describe(`Specific files to check (if omitted, checks all)`),cwd:F.string().optional().describe(`Working directory`),skip_types:F.boolean().default(!1).describe(`Skip TypeScript typecheck`),skip_lint:F.boolean().default(!1).describe(`Skip Biome lint`),detail:F.enum([`summary`,`errors`,`full`]).default(`summary`).describe(`Output detail level: summary (default, ~300 tokens — pass/fail + counts + top errors), errors (parsed error objects), full (includes raw terminal output)`)},outputSchema:di,annotations:t.annotations},async({files:e,cwd:t,skip_types:n,skip_lint:r,detail:i},a)=>{try{let o=await Se({files:e,cwd:t,skipTypes:n,skipLint:r,detail:i===`summary`?`errors`:i}),s=Ji(a).createTask(`Check`,2);if(s.progress(0,`tsc: ${o.tsc.errors.length} errors`),s.progress(1,`biome: ${o.biome.errors.length} errors`),s.complete(`Check ${o.passed?`passed`:`failed`}`),i===`summary`){let e=Gt(o),t=[];if(o.passed)t.push({tool:`test_run`,reason:`Types and lint clean — run tests next`});else{let e=o.tsc.errors[0]?.file??o.biome.errors[0]?.file;e&&t.push({tool:`symbol`,reason:`Resolve failing symbol in ${e}`,suggested_args:{name:e}}),t.push({tool:`check`,reason:`Re-check after fixing errors`,suggested_args:{detail:`errors`}})}let n=[`## Check ${e.passed?`✅ PASS`:`❌ FAIL`}`,``,`**tsc**: ${e.tsc.passed?`✅`:`❌`} ${e.tsc.errorCount} errors, ${e.tsc.warningCount} warnings`];if(e.tsc.topErrors.length>0)for(let t of e.tsc.topErrors)n.push(` - ${t}`);if(n.push(`**biome**: ${e.biome.passed?`✅`:`❌`} ${e.biome.errorCount} errors, ${e.biome.warningCount} warnings`),e.biome.topErrors.length>0)for(let t of e.biome.topErrors)n.push(` - ${t}`);if(t.length>0){n.push(``,`**Next steps:**`);for(let e of t)n.push(`- \`${e.tool}\`: ${e.reason}`)}return{content:[{type:`text`,text:n.join(`
544
+ `)}],structuredContent:{...e}}}return{content:[{type:`text`,text:JSON.stringify(o)}]}}catch(e){return Qa.error(`Check failed`,A(e)),{content:[{type:`text`,text:`Check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function eo(e,t,n){let r=L(`batch`);e.registerTool(`batch`,{title:r.title,description:`Execute multiple built-in operations in parallel with concurrency control. Supported operation types: search, find, and check.`,inputSchema:{operations:F.array(F.object({id:F.string().describe(`Unique ID for this operation`),type:F.enum([`search`,`find`,`check`]).describe(`Built-in operation type`),args:F.record(F.string(),F.unknown()).describe(`Arguments for the operation`)})).min(1).max(100).describe(`Operations to execute`),concurrency:F.number().min(1).max(20).default(4).describe(`Max concurrent operations`)},annotations:r.annotations},async({operations:e,concurrency:r},i)=>{try{let a=await ye(e,async e=>co(e,t,n),{concurrency:r}),o=Ji(i).createTask(`Batch`,e.length);for(let e=0;e<a.length;e++)o.progress(e,`${a[e].id}: ${a[e].status}`);return o.complete(`Batch complete: ${a.length} operations`),{content:[{type:`text`,text:JSON.stringify(a)}]}}catch(e){return Qa.error(`Batch failed`,A(e)),{content:[{type:`text`,text:`Batch failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function to(e){let t=L(`eval`);e.registerTool(`eval`,{title:t.title,description:`Execute a JavaScript or TypeScript snippet in a constrained VM sandbox with a timeout. Captures console output and returned values.`,inputSchema:{code:F.string().max(1e5).describe(`Code snippet to execute`),lang:F.enum([`js`,`ts`]).default(`js`).optional().describe(`Language mode: js executes directly, ts strips common type syntax first`),timeout:F.number().min(1).max(6e4).default(5e3).optional().describe(`Execution timeout in milliseconds`)},annotations:t.annotations},async({code:e,lang:t,timeout:n})=>{try{let r=ze({code:e,lang:t,timeout:n});return r.success?{content:[{type:`text`,text:`Eval succeeded in ${r.durationMs}ms\n\n${r.output}`}]}:{content:[{type:`text`,text:`Eval failed in ${r.durationMs}ms: ${r.error??`Unknown error`}`}],isError:!0}}catch(e){return Qa.error(`Eval failed`,A(e)),{content:[{type:`text`,text:`Eval failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function no(e){let t=L(`test_run`);e.registerTool(`test_run`,{title:t.title,description:`Run Vitest for the current project or a subset of files, then return a structured summary of passing and failing tests.`,inputSchema:{files:F.array(F.string()).optional().describe(`Specific test files or patterns to run`),grep:F.string().optional().describe(`Only run tests whose names match this pattern`),cwd:F.string().optional().describe(`Working directory for the test run`)},annotations:t.annotations},async({files:e,grep:t,cwd:n},r)=>{try{let i=await qt({files:e,grep:t,cwd:n}),a=Ji(r).createTask(`Test Run`,1),o=i.summary.passed+i.summary.failed+i.summary.skipped;return a.complete(`Tests: ${i.passed?`passed`:`failed`} (${o} tests)`),{content:[{type:`text`,text:fo(i)}],isError:!i.passed}}catch(e){return Qa.error(`Test run failed`,A(e)),{content:[{type:`text`,text:`Test run failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ro(e){let t=L(`parse_output`);e.registerTool(`parse_output`,{title:t.title,description:`Parse structured data from build tool output. Supports tsc, vitest, biome, and git status. Auto-detects the tool or specify explicitly.`,inputSchema:{output:F.string().max(5e5).describe(`Raw output text from a build tool`),tool:F.enum([`tsc`,`vitest`,`biome`,`git-status`]).optional().describe(`Tool to parse as (auto-detects if omitted)`)},annotations:t.annotations},async({output:e,tool:t})=>{try{let n=ut(e.replace(/\\n/g,`
545
+ `).replace(/\\t/g,` `),t);return{content:[{type:`text`,text:JSON.stringify(n)}]}}catch(e){return Qa.error(`Parse failed`,A(e)),{content:[{type:`text`,text:`Parse failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function io(e,t){let n=L(`delegate`);e.registerTool(`delegate`,{title:n.title,description:`Delegate a subtask to a local Ollama model. Use for summarization, classification, naming, or any task that can offload work from the host agent. Fails fast if Ollama is not running.`,inputSchema:{prompt:F.string().max(2e5).describe(`The task or question to send to the local model`),model:F.string().optional().describe(`Ollama model name (default: first available model)`),system:F.string().optional().describe(`System prompt for the model`),context:F.string().max(5e5).optional().describe(`Context text to include before the prompt (e.g. file contents)`),temperature:F.number().min(0).max(2).default(.3).optional().describe(`Sampling temperature (0=deterministic, default 0.3)`),timeout:F.number().min(1e3).max(6e5).default(12e4).optional().describe(`Timeout in milliseconds (default 120000)`),action:F.enum([`generate`,`list_models`]).default(`generate`).optional().describe(`Action: generate a response or list available models`)},annotations:n.annotations},async({prompt:e,model:n,system:r,context:i,temperature:a,timeout:o,action:s})=>{try{if(s===`list_models`){let e=await Ne();return{content:[{type:`text`,text:JSON.stringify({models:e,count:e.length,_Next:`Use delegate with a model name`},null,2)}]}}if(t?.available)try{let n=await t.createMessage({prompt:e,systemPrompt:r,context:i,maxTokens:4e3,modelPreferences:{intelligencePriority:.8,speedPriority:.5,costPriority:.3}});return{content:[{type:`text`,text:JSON.stringify({model:n.model??`sampling`,response:n.text,provider:`mcp-sampling`,_Next:`Use the response in your workflow. stash to save it.`},null,2)}]}}catch{Qa.debug(`Sampling failed, falling back to Ollama`)}let c=await Me({prompt:e,model:n,system:r,context:i,temperature:a,timeout:o});return c.error?{content:[{type:`text`,text:JSON.stringify({error:c.error,model:c.model,durationMs:c.durationMs},null,2)}],isError:!0}:{content:[{type:`text`,text:JSON.stringify({model:c.model,response:c.response,durationMs:c.durationMs,tokenCount:c.tokenCount,_Next:`Use the response in your workflow. stash to save it.`},null,2)}]}}catch(e){return Qa.error(`Delegate failed`,A(e)),{content:[{type:`text`,text:`Delegate failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const ao=F.object({query:F.string(),limit:F.number().min(1).max(20).default(5).optional(),search_mode:F.enum([`hybrid`,`semantic`,`keyword`]).default(`hybrid`).optional(),content_type:F.enum(b).optional(),origin:F.enum([`indexed`,`curated`,`produced`]).optional(),category:F.string().optional(),tags:F.array(F.string()).optional(),min_score:F.number().min(0).max(1).default(.25).optional()}),oo=F.object({query:F.string().optional(),glob:F.string().optional(),pattern:F.string().optional(),limit:F.number().min(1).max(50).default(10).optional(),content_type:F.enum(b).optional(),cwd:F.string().optional()}),so=F.object({files:F.array(F.string()).optional(),cwd:F.string().optional(),skip_types:F.boolean().optional(),skip_lint:F.boolean().optional()});async function co(e,t,n){switch(e.type){case`search`:return lo(t,n,ao.parse(e.args));case`find`:{let r=oo.parse(e.args);if(!r.query&&!r.glob&&!r.pattern)throw Error(`find operation requires query, glob, or pattern`);return He(t,n,{query:r.query,glob:r.glob,pattern:r.pattern,limit:r.limit,contentType:r.content_type,cwd:r.cwd})}case`check`:{let t=so.parse(e.args);return Se({files:t.files,cwd:t.cwd,skipTypes:t.skip_types,skipLint:t.skip_lint})}default:throw Error(`Unsupported batch operation type: ${e.type}`)}}async function lo(e,t,n){let r=n.limit??5,i={limit:r,minScore:n.min_score??.25,contentType:n.content_type,origin:n.origin,category:n.category,tags:n.tags},a=e.embedQuery?.bind(e)??e.embed.bind(e);if(n.search_mode===`keyword`)return(await t.ftsSearch(n.query,i)).slice(0,r);let o=await a(n.query);if(n.search_mode===`semantic`)return t.search(o,i);let[s,c]=await Promise.all([t.search(o,{...i,limit:r*2}),t.ftsSearch(n.query,{...i,limit:r*2})]);return uo(s,c).slice(0,r)}function uo(e,t,n=60){let r=new Map;for(let t=0;t<e.length;t++){let i=e[t];r.set(i.record.id,{record:i.record,score:1/(n+t+1)})}for(let e=0;e<t.length;e++){let i=t[e],a=r.get(i.record.id);if(a){a.score+=1/(n+e+1);continue}r.set(i.record.id,{record:i.record,score:1/(n+e+1)})}return[...r.values()].sort((e,t)=>t.score-e.score)}function fo(e){let t=[`Vitest run: ${e.passed?`passed`:`failed`}`,`Duration: ${e.durationMs}ms`,`Passed: ${e.summary.passed}`,`Failed: ${e.summary.failed}`,`Skipped: ${e.summary.skipped}`];e.summary.suites!==void 0&&t.push(`Suites: ${e.summary.suites}`);let n=e.summary.tests.filter(e=>e.status===`fail`);if(n.length>0){t.push(``,`Failed tests:`);for(let e of n)t.push(`- ${e.name}${e.file?` (${e.file})`:``}`),e.error&&t.push(` ${e.error}`)}return t.join(`
546
+ `)}const U=E(`flow-tools`);function po(e){return Array.isArray(e)?e.filter(e=>typeof e==`string`&&e.trim().length>0):[]}function mo(e){let t=Hr(e);return async({content:e,sourceFormat:n,stepName:r,metadata:i})=>{let a=r.charAt(0).toUpperCase()+r.slice(1),o=i?JSON.stringify(i,null,2):void 0,s=`You are converting a flow step definition from "${n}" format to aikit native flow step README.md format.
547
547
 
548
548
  Convert the provided content into a well-structured README.md for a flow step named "${r}".
549
549
 
@@ -556,18 +556,18 @@ The output must:
556
556
  6. Preserve ALL tool restrictions, constraints, and requirements from the source
557
557
  7. Be formatted as clean Markdown
558
558
 
559
- Output ONLY the README.md content, nothing else.`;if(t.available)try{let i=(await t.createMessage({prompt:e,systemPrompt:s,context:o?`Step name: ${r}\nSource format: ${n}\nMetadata:\n${o}`:`Step name: ${r}\nSource format: ${n}`,maxTokens:4096,modelPreferences:{intelligencePriority:.8,speedPriority:.4,costPriority:.2}})).text.trim();if(i.length>0)return i}catch(e){H.debug(`Flow transform sampling failed; using fallback formatting`,{error:W(e),sourceFormat:n,stepName:r})}let c=typeof i?.description==`string`&&i.description.trim().length>0?i.description.trim():null,l=mo(i?.produces),u=mo(i?.requires);return[`# ${a}`,``,`## Purpose`,``,c??`${a} step.`,``,`## Instructions`,``,e.trim(),``,`## Artifacts`,``,...l.length>0?l.map(e=>`- ${e}`):[`- ${r}.md`],...u.length>0?[``,`## Prerequisites`,``,...u.map(e=>`- ${e}`)]:[],``].join(`
559
+ Output ONLY the README.md content, nothing else.`;if(t.available)try{let i=(await t.createMessage({prompt:e,systemPrompt:s,context:o?`Step name: ${r}\nSource format: ${n}\nMetadata:\n${o}`:`Step name: ${r}\nSource format: ${n}`,maxTokens:4096,modelPreferences:{intelligencePriority:.8,speedPriority:.4,costPriority:.2}})).text.trim();if(i.length>0)return i}catch(e){U.debug(`Flow transform sampling failed; using fallback formatting`,{error:G(e),sourceFormat:n,stepName:r})}let c=typeof i?.description==`string`&&i.description.trim().length>0?i.description.trim():null,l=po(i?.produces),u=po(i?.requires);return[`# ${a}`,``,`## Purpose`,``,c??`${a} step.`,``,`## Instructions`,``,e.trim(),``,`## Artifacts`,``,...l.length>0?l.map(e=>`- ${e}`):[`- ${r}.md`],...u.length>0?[``,`## Prerequisites`,``,...u.map(e=>`- ${e}`)]:[],``].join(`
560
560
  `).trimEnd().concat(`
561
- `)}}function U(e){return{content:[{type:`text`,text:e}]}}function W(e){return e instanceof Error?e.message:String(e)}function go(e,t){function n(){return t.sources?.[0]?.path??process.cwd()}function r(){return g(D(),`flows`,`registry.json`)}function i(){return g(t.stateDir??g(n(),`.aikit-state`),`flows`,`state.json`)}function o(e,t){let r=t?.installPath?p(t.installPath):e.replaceAll(`:`,`-`),i=_(N(),`.copilot`,`flows`,r).replaceAll(`\\`,`/`);if(a(i))return i;let o=_(N(),`.claude`,`flows`,r).replaceAll(`\\`,`/`);if(a(o))return o;let s=_(n(),`.github`,`flows`,r).replaceAll(`\\`,`/`);return a(s)?s:t?.installPath&&a(t.installPath)?t.installPath.replaceAll(`\\`,`/`):null}function d(e,t){let n=o(e.name,e);return n?_(n,t).replaceAll(`\\`,`/`):t}function f(){try{let e=g(bn(`npm root -g`,{encoding:`utf-8`}).trim(),`@fission-ai`,`openspec`);if(a(e))return{path:e,sourceType:`npm-global`,isGit:!1}}catch{}return{path:`https://github.com/Fission-AI/OpenSpec.git`,sourceType:`git`,isGit:!0}}function m(e){let t=_(e);if(a(g(t,`.claude-plugin`,`plugin.json`)))return{path:t,sourceType:`claude-plugin`};let n=N(),r=g(n,`.claude`,`plugins`,e);if(a(g(r,`.claude-plugin`,`plugin.json`)))return{path:r,sourceType:`claude-plugin`};let i=g(n,`.claude`,`plugins`);if(!a(i))return null;for(let t of c(i,{withFileTypes:!0})){if(!t.isDirectory())continue;let n=g(i,t.name,`.claude-plugin`,`plugin.json`);if(a(n))try{if(JSON.parse(s(n,`utf-8`)).name===e)return{path:g(i,t.name),sourceType:`claude-plugin`}}catch{}}return null}async function h(){let{FlowRegistryManager:e,FlowStateMachine:a,FlowLoader:o,GitInstaller:s}=await import(`../../flows/dist/index.js`),c=g(t.stateDir??g(n(),`.aikit-state`),`flows`,`installed`);return{registry:new e(r()),stateMachine:new a(i()),loader:new o,installer:new s(c)}}let v=I(`flow_list`);e.registerTool(`flow_list`,{title:v.title,description:`List all installed flows and their steps`,annotations:v.annotations,inputSchema:{}},async()=>{try{let{registry:e,stateMachine:t,installer:n}=await h(),r=e.list(),i=t.getStatus(),a={flows:r.map(e=>{let t={name:e.name,version:e.version,source:e.source,sourceType:e.sourceType,format:e.format,steps:e.manifest.steps.map(e=>e.id)};if(e.sourceType===`git`&&e.commitSha){let r=n.hasUpdates(e.installPath);return{...t,commitSha:e.commitSha,updateAvailable:r.success&&r.data?r.data.hasUpdates:void 0}}return t}),activeFlow:i.success&&i.data?{flow:i.data.flow,status:i.data.status,currentStep:i.data.currentStep}:null};return U(JSON.stringify(a,null,2))}catch(e){return H.error(`flow_list failed`,A(e)),U(`Error: ${W(e)}`)}});let y=I(`flow_info`);e.registerTool(`flow_info`,{title:y.title,description:`Show detailed information about a specific flow`,annotations:y.annotations,inputSchema:{name:F.string().describe(`Flow name to get info for`)}},async({name:e})=>{try{let{registry:t,installer:n}=await h(),r=t.get(e);if(!r)return U(`Flow "${e}" not found. Use flow_list to see available flows.`);let i=r.commitSha??null,a;if(r.sourceType===`git`&&r.installPath){let e=n.hasUpdates(r.installPath);a=e.success&&e.data?e.data.hasUpdates:void 0}let s={name:r.name,version:r.version,description:r.manifest.description,source:r.source,sourceType:r.sourceType,format:r.format,commitSha:i,updateAvailable:a,installPath:o(r.name,r),registeredAt:r.registeredAt,updatedAt:r.updatedAt,steps:r.manifest.steps.map(e=>({id:e.id,name:e.name,instruction:d(r,e.instruction),produces:e.produces,requires:e.requires,description:e.description})),agents:r.manifest.agents,artifactsDir:r.manifest.artifacts_dir,install:r.manifest.install};return U(JSON.stringify(s,null,2))}catch(e){return H.error(`flow_info failed`,A(e)),U(`Error: ${W(e)}`)}});let b=I(`flow_start`);e.registerTool(`flow_start`,{title:b.title,description:`Start a flow. Sets the active flow and positions at the first step.`,annotations:b.annotations,inputSchema:{flow:F.string().describe(`Flow name to start (use flow_list to see options)`)}},async({flow:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=t.get(e);if(!r)return U(`Flow "${e}" not found. Use flow_list to see available flows.`);let i=n.start(r.name,r.manifest);if(!i.success||!i.data)return U(`Cannot start: ${i.error}`);let a=i.data,o=r.manifest.steps.find(e=>e.id===a.currentStep),s={started:!0,flow:a.flow,currentStep:a.currentStep,currentStepInstruction:r&&o?d(r,o.instruction):null,currentStepDescription:o?.description??null,totalSteps:r.manifest.steps.length,stepSequence:r.manifest.steps.map(e=>e.id),artifactsDir:r.manifest.artifacts_dir};return U(JSON.stringify(s,null,2))}catch(e){return H.error(`flow_start failed`,A(e)),U(`Error: ${W(e)}`)}});let x=I(`flow_step`);e.registerTool(`flow_step`,{title:x.title,description:`Advance the active flow: complete current step and move to next, skip current step, or redo current step.`,annotations:x.annotations,inputSchema:{action:F.enum([`next`,`skip`,`redo`]).describe(`next: mark current step done and advance. skip: skip current step. redo: repeat current step.`)}},async({action:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=n.load();if(!r)return U(`No active flow. Use flow_start first.`);let i=t.get(r.flow);if(!i)return U(`Flow "${r.flow}" not found in registry.`);let a=n.step(e,i.manifest);if(!a.success||!a.data)return U(`Cannot ${e}: ${a.error}`);let o=a.data,s=o.currentStep?i.manifest.steps.find(e=>e.id===o.currentStep):null,c={flow:o.flow,status:o.status,action:e,currentStep:o.currentStep,currentStepInstruction:i&&s?d(i,s.instruction):null,currentStepDescription:s?.description??null,completedSteps:o.completedSteps,skippedSteps:o.skippedSteps,totalSteps:i.manifest.steps.length,remaining:i.manifest.steps.filter(e=>!o.completedSteps.includes(e.id)&&!o.skippedSteps.includes(e.id)&&e.id!==o.currentStep).map(e=>e.id)};return U(JSON.stringify(c,null,2))}catch(e){return H.error(`flow_step failed`,A(e)),U(`Error: ${W(e)}`)}});let S=I(`flow_status`);e.registerTool(`flow_status`,{title:S.title,description:`Show the current flow execution state — which flow is active, current step, completed steps, and artifacts.`,annotations:S.annotations,inputSchema:{}},async()=>{try{let{registry:e,stateMachine:t}=await h(),n=t.getStatus();if(!n.success||!n.data)return U(`No active flow. Use flow_start to begin one, or flow_list to see available flows.`);let r=n.data,i=e.get(r.flow),a=i?.manifest.steps.find(e=>e.id===r.currentStep),o=i&&a?d(i,a.instruction):null,s={flow:r.flow,status:r.status,currentStep:r.currentStep,currentStepInstruction:o,instructionPath:o,currentStepDescription:a?.description??null,completedSteps:r.completedSteps,skippedSteps:r.skippedSteps,artifacts:r.artifacts,startedAt:r.startedAt,updatedAt:r.updatedAt,totalSteps:i?.manifest.steps.length??0,progress:i?`${r.completedSteps.length+r.skippedSteps.length}/${i.manifest.steps.length}`:`unknown`};return U(JSON.stringify(s,null,2))}catch(e){return H.error(`flow_status failed`,A(e)),U(`Error: ${W(e)}`)}});let C=I(`flow_read_instruction`);e.registerTool(`flow_read_instruction`,{title:C.title===`flow_read_instruction`?`Flow Read Instruction`:C.title,description:`Read the instruction content for a flow step. If step is omitted, reads the current step.`,annotations:C.title===`flow_read_instruction`?{readOnlyHint:!0,idempotentHint:!0}:C.annotations,inputSchema:{step:F.string().optional().describe(`Step id or name to read. Defaults to the current step.`)}},async({step:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=n.getStatus();if(!r.success||!r.data)return U(`No active flow. Use flow_start to begin one, or flow_list to see available flows.`);let i=r.data,a=t.get(i.flow);if(!a)return U(`Flow "${i.flow}" not found in registry.`);let o=e??i.currentStep;if(!o)return U(`No current step is available for the active flow.`);let s=a.manifest.steps.find(e=>e.id===o||e.name===o);return U(s?await te(d(a,s.instruction),`utf-8`):`Step "${o}" not found in flow "${i.flow}".`)}catch(e){return H.error(`flow_read_instruction failed`,A(e)),e instanceof Error&&`code`in e&&e.code===`ENOENT`?U(`Could not read instruction file: ${e.message}`):U(`Error: ${W(e)}`)}});let w=I(`flow_reset`);e.registerTool(`flow_reset`,{title:w.title,description:`Reset the active flow, clearing all state. Use to start over or switch to a different flow.`,annotations:w.annotations,inputSchema:{}},async()=>{try{let{stateMachine:e}=await h(),t=e.reset();return t.success?U(`Flow state reset. Use flow_start to begin a new flow.`):U(`Reset failed: ${t.error}`)}catch(e){return H.error(`flow_reset failed`,A(e)),U(`Error: ${W(e)}`)}});let T=I(`flow_add`);e.registerTool(`flow_add`,{title:T.title,description:`Install a new development flow from a git repository URL or local directory path. Use when the user wants to add, install, import, or onboard a new workflow — for example: "use this as a flow", "add this flow", "add this flow URL", "install a flow", or "onboard a flow". Accepts git URLs (https://..., git@...) and local filesystem paths. Also accepts the shorthand "openspec" to install the OpenSpec flow (auto-detects local npm global install or clones from GitHub).`,annotations:T.annotations,inputSchema:{source:F.string().describe(`Git repository URL (https://... or git@...) or absolute/local directory path containing a flow definition. Use "openspec" as a shorthand to auto-resolve the OpenSpec flow.`),name:F.string().optional().describe(`Optional override for the installed flow name. If omitted, the name comes from the flow manifest.`),token:F.string().optional().describe(`Authentication token for private/GHE repositories`)}},async({source:t,name:n,token:r})=>{try{let{registry:i,loader:a,installer:o}=await h(),s=ho(e),c=t===`openspec`||t.startsWith(`openspec:`),l,u=t,d;if(c){let e=f();u=e.path,l=e.sourceType}let g=c?null:m(u);g&&(u=g.path,l=g.sourceType);let v=/^https?:\/\/|^git@/.test(u),y;if(v){let e=o.clone(u,r);if(!e.success||!e.data)return U(`Failed to clone flow: ${e.error}`);y=e.data}else{let e=_(u);d=e;let t=n||p(e)||`custom-flow`,r=o.copyLocal(e,t);if(!r.success||!r.data)return U(`Failed to copy flow: ${r.error}`);y=r.data}let b=await a.load(y,{forceAssetSync:!0,transform:s});if(!b.success||!b.data)return o.remove(y),U(`Failed to load flow manifest: ${b.error}`);let{manifest:x,format:S}=b.data,C=n||x.name,w=l??(v?`git`:`local`);if(i.has(C))return o.remove(y),U(`Flow "${C}" is already installed. Use flow_update to update it, or flow_remove then flow_add to replace.`);if(x.install.length>0){let e=o.runInstallDeps(x.install);if(!e.success)return o.remove(y),U(`Dependency install failed: ${e.error}`)}let T=new Date().toISOString(),E=v?o.getLocalCommit(y):void 0,D=i.register({name:C,version:x.version,source:w===`local`||w===`claude-plugin`?d??t:t,sourceType:w,installPath:y,format:S,registeredAt:T,updatedAt:T,manifest:x,...E?{commitSha:E}:{}});if(!D.success)return o.remove(y),U(`Failed to register flow: ${D.error}`);let O=x.steps.length;return U(`Flow "${C}" installed successfully (${O} steps). Use flow_start({ flow: "${C}" }) to begin.`)}catch(e){return H.error(`flow_add failed`,A(e)),U(`Error: ${W(e)}`)}});let E=I(`flow_remove`);e.registerTool(`flow_remove`,{title:E.title,description:`Remove an installed flow by name. Unregisters it and deletes its files when applicable. Builtin flows cannot be removed.`,annotations:E.annotations,inputSchema:{name:F.string().describe(`Name of the flow to remove.`)}},async({name:e})=>{try{let{registry:t,installer:n}=await h();if(!t.has(e))return U(`Flow "${e}" is not installed.`);let r=t.get(e);if(!r)return U(`Flow "${e}" is not installed.`);if(r.sourceType===`builtin`)return U(`Cannot remove builtin flow "${e}".`);let i=n.remove(r.installPath);if(!i.success)return U(`Failed to remove flow files: ${i.error}`);let a=t.unregister(e);return a.success?U(`Flow "${e}" removed successfully.`):U(`Failed to unregister flow: ${a.error}`)}catch(e){return H.error(`flow_remove failed`,A(e)),U(`Error: ${W(e)}`)}});let O=I(`flow_update`);e.registerTool(`flow_update`,{title:O.title,description:`Update an installed flow to its latest version. For git-based flows, pulls the latest changes. For npm-global flows (e.g. OpenSpec), runs npm update.`,annotations:O.annotations,inputSchema:{name:F.string().describe(`Name of the flow to update.`)}},async({name:t})=>{try{let{registry:n,loader:r,installer:i}=await h(),o=ho(e);if(!n.has(t))return U(`Flow "${t}" is not installed.`);let s=n.get(t);if(!s||s.sourceType!==`git`&&s.sourceType!==`npm-global`&&s.sourceType!==`local`&&s.sourceType!==`claude-plugin`)return U(`Flow "${t}" was not installed from git, npm, or a local source — cannot update. Remove and re-add instead.`);if(s.sourceType===`local`||s.sourceType===`claude-plugin`){let e=_(s.source);if(!a(e))return U(`Source path no longer exists: ${e}`);let t=`${s.installPath}.updating`,c=`${s.installPath}.backup`;a(t)&&u(t,{recursive:!0,force:!0}),a(c)&&u(c,{recursive:!0,force:!0});let d=i.copyLocal(e,p(t));if(!d.success||!d.data)return U(`Failed to copy flow: ${d.error}`);let f=await r.load(t,{forceAssetSync:!0,transform:o});if(!f.success||!f.data)return i.remove(t),U(`Failed to load flow manifest: ${f.error}`);let m=!1;try{l(s.installPath,c),l(t,s.installPath);let e=await r.load(s.installPath,{forceAssetSync:!0,transform:o});if(!e.success||!e.data)throw Error(`Failed to load flow manifest: ${e.error}`);let a=e.data.manifest,d=e.data.format,f=n.register({...s,version:a.version,format:d,installPath:s.installPath,manifest:a,updatedAt:new Date().toISOString()});if(!f.success)throw Error(`Failed to refresh flow registry entry: ${f.error}`);if(m=!0,a.install.length>0){let e=i.runInstallDeps(a.install);if(!e.success)throw Error(`Dependency install failed: ${e.error}`)}u(c,{recursive:!0,force:!0})}catch(e){return a(t)&&u(t,{recursive:!0,force:!0}),a(c)&&(a(s.installPath)&&u(s.installPath,{recursive:!0,force:!0}),l(c,s.installPath)),m&&n.register(s),U(W(e))}return U(`Flow "${s.name}" updated from ${s.sourceType} source successfully.`)}if(s.sourceType===`npm-global`){try{bn(`npm update -g @fission-ai/openspec`,{encoding:`utf-8`,stdio:`pipe`})}catch(e){return U(`npm update failed: ${W(e)}`)}let e=f(),i=await r.load(e.path,{forceAssetSync:!0,transform:o}),a=i.success&&i.data?i.data.manifest:null,c=i.success&&i.data?i.data.format:s.format;if(a){let t=n.register({...s,version:a.version,format:c,installPath:e.path,manifest:a,updatedAt:new Date().toISOString()});if(!t.success)return U(`Failed to refresh flow registry entry: ${t.error}`)}return U(`Flow "${t}" updated via npm successfully.`)}let c=i.update(s.installPath);if(!c.success)return U(`Update failed: ${c.error}`);let d=await r.load(s.installPath,{forceAssetSync:!0,transform:o}),m=d.success&&d.data?d.data.manifest:null,g=d.success&&d.data?d.data.format:s.format;if(m){let e=i.getLocalCommit(s.installPath),t=n.register({...s,version:m.version,format:g,manifest:m,updatedAt:new Date().toISOString(),...e?{commitSha:e}:{}});if(!t.success)return U(`Failed to refresh flow registry entry: ${t.error}`)}let v=m?.install??s.manifest.install;if(v.length>0){let e=i.runInstallDeps(v);if(!e.success)return U(`Dependency install failed: ${e.error}`)}return U(`Flow "${t}" updated successfully.`)}catch(e){return H.error(`flow_update failed`,A(e)),U(`Error: ${W(e)}`)}})}const _o=E(`tools`);function vo(e){let t=I(`evidence_map`);e.registerTool(`evidence_map`,{title:t.title,description:`Track verified/assumed/unresolved claims for complex tasks. Gate readiness: YIELD (proceed), HOLD (unknowns), HARD_BLOCK (critical gaps). Persists across calls.`,inputSchema:{action:F.enum([`create`,`add`,`update`,`get`,`gate`,`list`,`delete`]).describe(`Operation to perform`),task_id:F.string().optional().describe(`Task identifier (required for all except list)`),tier:F.enum([`floor`,`standard`,`critical`]).optional().describe(`FORGE tier (required for create)`),claim:F.string().optional().describe(`Critical-path claim text (for add)`),status:F.enum([`V`,`A`,`U`]).optional().describe(`Evidence status: V=Verified, A=Assumed, U=Unresolved`),receipt:F.string().optional().describe(`Evidence receipt: tool→ref for V, reasoning for A, attempts for U`),id:F.number().optional().describe(`Entry ID (for update)`),critical_path:F.boolean().default(!0).describe(`Whether this claim is on the critical path`),unknown_type:F.enum([`contract`,`convention`,`freshness`,`runtime`,`data-flow`,`impact`]).optional().describe(`Typed unknown classification`),safety_gate:F.enum([`provenance`,`commitment`,`coverage`]).optional().describe(`Safety gate tag: provenance (claim→evidence), commitment (user-approved), coverage (nothing dropped)`),retry_count:F.number().default(0).describe(`Retry count for gate evaluation (0 = first attempt)`),cwd:F.string().optional().describe(`Working directory for evidence map storage (default: server cwd). Use root_path from forge_ground to match.`)},annotations:t.annotations},async({action:e,task_id:t,tier:n,claim:r,status:i,receipt:a,id:o,critical_path:s,unknown_type:c,safety_gate:l,retry_count:u,cwd:d})=>{try{switch(e){case`create`:if(!t)throw Error(`task_id required for create`);if(!n)throw Error(`tier required for create`);return Be({action:`create`,taskId:t,tier:n},d),{content:[{type:`text`,text:`Created evidence map "${t}" (tier: ${n}).\n\n---\n_Next: Use \`evidence_map\` with action "add" to record critical-path claims._`}]};case`add`:{if(!t)throw Error(`task_id required for add`);if(!r)throw Error(`claim required for add`);if(!i)throw Error(`status required for add`);let e=Be({action:`add`,taskId:t,claim:r,status:i,receipt:a??``,criticalPath:s,unknownType:c,safetyGate:l},d),n=[`Added entry #${e.entry?.id} to "${t}": [${i}] ${r}`];return e.formattedMap&&n.push(``,e.formattedMap),{content:[{type:`text`,text:n.join(`
561
+ `)}}function W(e){return{content:[{type:`text`,text:e}]}}function G(e){return e instanceof Error?e.message:String(e)}function ho(e,t){function n(){return t.sources?.[0]?.path??process.cwd()}function r(){return g(D(),`flows`,`registry.json`)}function i(){return g(t.stateDir??g(n(),`.aikit-state`),`flows`,`state.json`)}function o(e,t){let r=t?.installPath?p(t.installPath):e.replaceAll(`:`,`-`),i=_(N(),`.copilot`,`flows`,r).replaceAll(`\\`,`/`);if(a(i))return i;let o=_(N(),`.claude`,`flows`,r).replaceAll(`\\`,`/`);if(a(o))return o;let s=_(n(),`.github`,`flows`,r).replaceAll(`\\`,`/`);return a(s)?s:t?.installPath&&a(t.installPath)?t.installPath.replaceAll(`\\`,`/`):null}function d(e,t){let n=o(e.name,e);return n?_(n,t).replaceAll(`\\`,`/`):t}function f(){try{let e=g(bn(`npm root -g`,{encoding:`utf-8`}).trim(),`@fission-ai`,`openspec`);if(a(e))return{path:e,sourceType:`npm-global`,isGit:!1}}catch{}return{path:`https://github.com/Fission-AI/OpenSpec.git`,sourceType:`git`,isGit:!0}}function m(e){let t=_(e);if(a(g(t,`.claude-plugin`,`plugin.json`)))return{path:t,sourceType:`claude-plugin`};let n=N(),r=g(n,`.claude`,`plugins`,e);if(a(g(r,`.claude-plugin`,`plugin.json`)))return{path:r,sourceType:`claude-plugin`};let i=g(n,`.claude`,`plugins`);if(!a(i))return null;for(let t of c(i,{withFileTypes:!0})){if(!t.isDirectory())continue;let n=g(i,t.name,`.claude-plugin`,`plugin.json`);if(a(n))try{if(JSON.parse(s(n,`utf-8`)).name===e)return{path:g(i,t.name),sourceType:`claude-plugin`}}catch{}}return null}async function h(){let{FlowRegistryManager:e,FlowStateMachine:a,FlowLoader:o,GitInstaller:s}=await import(`../../flows/dist/index.js`),c=g(t.stateDir??g(n(),`.aikit-state`),`flows`,`installed`);return{registry:new e(r()),stateMachine:new a(i()),loader:new o,installer:new s(c)}}let v=L(`flow_list`);e.registerTool(`flow_list`,{title:v.title,description:`List all installed flows and their steps`,annotations:v.annotations,inputSchema:{}},async()=>{try{let{registry:e,stateMachine:t,installer:n}=await h(),r=e.list(),i=t.getStatus(),a={flows:r.map(e=>{let t={name:e.name,version:e.version,source:e.source,sourceType:e.sourceType,format:e.format,steps:e.manifest.steps.map(e=>e.id)};if(e.sourceType===`git`&&e.commitSha){let r=n.hasUpdates(e.installPath);return{...t,commitSha:e.commitSha,updateAvailable:r.success&&r.data?r.data.hasUpdates:void 0}}return t}),activeFlow:i.success&&i.data?{flow:i.data.flow,status:i.data.status,currentStep:i.data.currentStep}:null};return W(JSON.stringify(a,null,2))}catch(e){return U.error(`flow_list failed`,A(e)),W(`Error: ${G(e)}`)}});let y=L(`flow_info`);e.registerTool(`flow_info`,{title:y.title,description:`Show detailed information about a specific flow`,annotations:y.annotations,inputSchema:{name:F.string().describe(`Flow name to get info for`)}},async({name:e})=>{try{let{registry:t,installer:n}=await h(),r=t.get(e);if(!r)return W(`Flow "${e}" not found. Use flow_list to see available flows.`);let i=r.commitSha??null,a;if(r.sourceType===`git`&&r.installPath){let e=n.hasUpdates(r.installPath);a=e.success&&e.data?e.data.hasUpdates:void 0}let s={name:r.name,version:r.version,description:r.manifest.description,source:r.source,sourceType:r.sourceType,format:r.format,commitSha:i,updateAvailable:a,installPath:o(r.name,r),registeredAt:r.registeredAt,updatedAt:r.updatedAt,steps:r.manifest.steps.map(e=>({id:e.id,name:e.name,instruction:d(r,e.instruction),produces:e.produces,requires:e.requires,description:e.description})),agents:r.manifest.agents,artifactsDir:r.manifest.artifacts_dir,install:r.manifest.install};return W(JSON.stringify(s,null,2))}catch(e){return U.error(`flow_info failed`,A(e)),W(`Error: ${G(e)}`)}});let b=L(`flow_start`);e.registerTool(`flow_start`,{title:b.title,description:`Start a flow. Sets the active flow and positions at the first step.`,annotations:b.annotations,inputSchema:{flow:F.string().describe(`Flow name to start (use flow_list to see options)`)}},async({flow:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=t.get(e);if(!r)return W(`Flow "${e}" not found. Use flow_list to see available flows.`);let i=n.start(r.name,r.manifest);if(!i.success||!i.data)return W(`Cannot start: ${i.error}`);let a=i.data,o=r.manifest.steps.find(e=>e.id===a.currentStep),s={started:!0,flow:a.flow,currentStep:a.currentStep,currentStepInstruction:r&&o?d(r,o.instruction):null,currentStepDescription:o?.description??null,totalSteps:r.manifest.steps.length,stepSequence:r.manifest.steps.map(e=>e.id),artifactsDir:r.manifest.artifacts_dir};return W(JSON.stringify(s,null,2))}catch(e){return U.error(`flow_start failed`,A(e)),W(`Error: ${G(e)}`)}});let x=L(`flow_step`);e.registerTool(`flow_step`,{title:x.title,description:`Advance the active flow: complete current step and move to next, skip current step, or redo current step.`,annotations:x.annotations,inputSchema:{action:F.enum([`next`,`skip`,`redo`]).describe(`next: mark current step done and advance. skip: skip current step. redo: repeat current step.`)}},async({action:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=n.load();if(!r)return W(`No active flow. Use flow_start first.`);let i=t.get(r.flow);if(!i)return W(`Flow "${r.flow}" not found in registry.`);let a=n.step(e,i.manifest);if(!a.success||!a.data)return W(`Cannot ${e}: ${a.error}`);let o=a.data,s=o.currentStep?i.manifest.steps.find(e=>e.id===o.currentStep):null,c={flow:o.flow,status:o.status,action:e,currentStep:o.currentStep,currentStepInstruction:i&&s?d(i,s.instruction):null,currentStepDescription:s?.description??null,completedSteps:o.completedSteps,skippedSteps:o.skippedSteps,totalSteps:i.manifest.steps.length,remaining:i.manifest.steps.filter(e=>!o.completedSteps.includes(e.id)&&!o.skippedSteps.includes(e.id)&&e.id!==o.currentStep).map(e=>e.id)};return W(JSON.stringify(c,null,2))}catch(e){return U.error(`flow_step failed`,A(e)),W(`Error: ${G(e)}`)}});let S=L(`flow_status`);e.registerTool(`flow_status`,{title:S.title,description:`Show the current flow execution state — which flow is active, current step, completed steps, and artifacts.`,annotations:S.annotations,inputSchema:{}},async()=>{try{let{registry:e,stateMachine:t}=await h(),n=t.getStatus();if(!n.success||!n.data)return W(`No active flow. Use flow_start to begin one, or flow_list to see available flows.`);let r=n.data,i=e.get(r.flow),a=i?.manifest.steps.find(e=>e.id===r.currentStep),o=i&&a?d(i,a.instruction):null,s={flow:r.flow,status:r.status,currentStep:r.currentStep,currentStepInstruction:o,instructionPath:o,currentStepDescription:a?.description??null,completedSteps:r.completedSteps,skippedSteps:r.skippedSteps,artifacts:r.artifacts,startedAt:r.startedAt,updatedAt:r.updatedAt,totalSteps:i?.manifest.steps.length??0,progress:i?`${r.completedSteps.length+r.skippedSteps.length}/${i.manifest.steps.length}`:`unknown`};return W(JSON.stringify(s,null,2))}catch(e){return U.error(`flow_status failed`,A(e)),W(`Error: ${G(e)}`)}});let C=L(`flow_read_instruction`);e.registerTool(`flow_read_instruction`,{title:C.title===`flow_read_instruction`?`Flow Read Instruction`:C.title,description:`Read the instruction content for a flow step. If step is omitted, reads the current step.`,annotations:C.title===`flow_read_instruction`?{readOnlyHint:!0,idempotentHint:!0}:C.annotations,inputSchema:{step:F.string().optional().describe(`Step id or name to read. Defaults to the current step.`)}},async({step:e})=>{try{let{registry:t,stateMachine:n}=await h(),r=n.getStatus();if(!r.success||!r.data)return W(`No active flow. Use flow_start to begin one, or flow_list to see available flows.`);let i=r.data,a=t.get(i.flow);if(!a)return W(`Flow "${i.flow}" not found in registry.`);let o=e??i.currentStep;if(!o)return W(`No current step is available for the active flow.`);let s=a.manifest.steps.find(e=>e.id===o||e.name===o);return W(s?await te(d(a,s.instruction),`utf-8`):`Step "${o}" not found in flow "${i.flow}".`)}catch(e){return U.error(`flow_read_instruction failed`,A(e)),e instanceof Error&&`code`in e&&e.code===`ENOENT`?W(`Could not read instruction file: ${e.message}`):W(`Error: ${G(e)}`)}});let w=L(`flow_reset`);e.registerTool(`flow_reset`,{title:w.title,description:`Reset the active flow, clearing all state. Use to start over or switch to a different flow.`,annotations:w.annotations,inputSchema:{}},async()=>{try{let{stateMachine:e}=await h(),t=e.reset();return t.success?W(`Flow state reset. Use flow_start to begin a new flow.`):W(`Reset failed: ${t.error}`)}catch(e){return U.error(`flow_reset failed`,A(e)),W(`Error: ${G(e)}`)}});let T=L(`flow_add`);e.registerTool(`flow_add`,{title:T.title,description:`Install a new development flow from a git repository URL or local directory path. Use when the user wants to add, install, import, or onboard a new workflow — for example: "use this as a flow", "add this flow", "add this flow URL", "install a flow", or "onboard a flow". Accepts git URLs (https://..., git@...) and local filesystem paths. Also accepts the shorthand "openspec" to install the OpenSpec flow (auto-detects local npm global install or clones from GitHub).`,annotations:T.annotations,inputSchema:{source:F.string().describe(`Git repository URL (https://... or git@...) or absolute/local directory path containing a flow definition. Use "openspec" as a shorthand to auto-resolve the OpenSpec flow.`),name:F.string().optional().describe(`Optional override for the installed flow name. If omitted, the name comes from the flow manifest.`),token:F.string().optional().describe(`Authentication token for private/GHE repositories`)}},async({source:t,name:n,token:r})=>{try{let{registry:i,loader:a,installer:o}=await h(),s=mo(e),c=t===`openspec`||t.startsWith(`openspec:`),l,u=t,d;if(c){let e=f();u=e.path,l=e.sourceType}let g=c?null:m(u);g&&(u=g.path,l=g.sourceType);let v=/^https?:\/\/|^git@/.test(u),y;if(v){let e=o.clone(u,r);if(!e.success||!e.data)return W(`Failed to clone flow: ${e.error}`);y=e.data}else{let e=_(u);d=e;let t=n||p(e)||`custom-flow`,r=o.copyLocal(e,t);if(!r.success||!r.data)return W(`Failed to copy flow: ${r.error}`);y=r.data}let b=await a.load(y,{forceAssetSync:!0,transform:s});if(!b.success||!b.data)return o.remove(y),W(`Failed to load flow manifest: ${b.error}`);let{manifest:x,format:S}=b.data,C=n||x.name,w=l??(v?`git`:`local`);if(i.has(C))return o.remove(y),W(`Flow "${C}" is already installed. Use flow_update to update it, or flow_remove then flow_add to replace.`);if(x.install.length>0){let e=o.runInstallDeps(x.install);if(!e.success)return o.remove(y),W(`Dependency install failed: ${e.error}`)}let T=new Date().toISOString(),E=v?o.getLocalCommit(y):void 0,D=i.register({name:C,version:x.version,source:w===`local`||w===`claude-plugin`?d??t:t,sourceType:w,installPath:y,format:S,registeredAt:T,updatedAt:T,manifest:x,...E?{commitSha:E}:{}});if(!D.success)return o.remove(y),W(`Failed to register flow: ${D.error}`);let O=x.steps.length;return W(`Flow "${C}" installed successfully (${O} steps). Use flow_start({ flow: "${C}" }) to begin.`)}catch(e){return U.error(`flow_add failed`,A(e)),W(`Error: ${G(e)}`)}});let E=L(`flow_remove`);e.registerTool(`flow_remove`,{title:E.title,description:`Remove an installed flow by name. Unregisters it and deletes its files when applicable. Builtin flows cannot be removed.`,annotations:E.annotations,inputSchema:{name:F.string().describe(`Name of the flow to remove.`)}},async({name:e})=>{try{let{registry:t,installer:n}=await h();if(!t.has(e))return W(`Flow "${e}" is not installed.`);let r=t.get(e);if(!r)return W(`Flow "${e}" is not installed.`);if(r.sourceType===`builtin`)return W(`Cannot remove builtin flow "${e}".`);let i=n.remove(r.installPath);if(!i.success)return W(`Failed to remove flow files: ${i.error}`);let a=t.unregister(e);return a.success?W(`Flow "${e}" removed successfully.`):W(`Failed to unregister flow: ${a.error}`)}catch(e){return U.error(`flow_remove failed`,A(e)),W(`Error: ${G(e)}`)}});let O=L(`flow_update`);e.registerTool(`flow_update`,{title:O.title,description:`Update an installed flow to its latest version. For git-based flows, pulls the latest changes. For npm-global flows (e.g. OpenSpec), runs npm update.`,annotations:O.annotations,inputSchema:{name:F.string().describe(`Name of the flow to update.`)}},async({name:t})=>{try{let{registry:n,loader:r,installer:i}=await h(),o=mo(e);if(!n.has(t))return W(`Flow "${t}" is not installed.`);let s=n.get(t);if(!s||s.sourceType!==`git`&&s.sourceType!==`npm-global`&&s.sourceType!==`local`&&s.sourceType!==`claude-plugin`)return W(`Flow "${t}" was not installed from git, npm, or a local source — cannot update. Remove and re-add instead.`);if(s.sourceType===`local`||s.sourceType===`claude-plugin`){let e=_(s.source);if(!a(e))return W(`Source path no longer exists: ${e}`);let t=`${s.installPath}.updating`,c=`${s.installPath}.backup`;a(t)&&u(t,{recursive:!0,force:!0}),a(c)&&u(c,{recursive:!0,force:!0});let d=i.copyLocal(e,p(t));if(!d.success||!d.data)return W(`Failed to copy flow: ${d.error}`);let f=await r.load(t,{forceAssetSync:!0,transform:o});if(!f.success||!f.data)return i.remove(t),W(`Failed to load flow manifest: ${f.error}`);let m=!1;try{l(s.installPath,c),l(t,s.installPath);let e=await r.load(s.installPath,{forceAssetSync:!0,transform:o});if(!e.success||!e.data)throw Error(`Failed to load flow manifest: ${e.error}`);let a=e.data.manifest,d=e.data.format,f=n.register({...s,version:a.version,format:d,installPath:s.installPath,manifest:a,updatedAt:new Date().toISOString()});if(!f.success)throw Error(`Failed to refresh flow registry entry: ${f.error}`);if(m=!0,a.install.length>0){let e=i.runInstallDeps(a.install);if(!e.success)throw Error(`Dependency install failed: ${e.error}`)}u(c,{recursive:!0,force:!0})}catch(e){return a(t)&&u(t,{recursive:!0,force:!0}),a(c)&&(a(s.installPath)&&u(s.installPath,{recursive:!0,force:!0}),l(c,s.installPath)),m&&n.register(s),W(G(e))}return W(`Flow "${s.name}" updated from ${s.sourceType} source successfully.`)}if(s.sourceType===`npm-global`){try{bn(`npm update -g @fission-ai/openspec`,{encoding:`utf-8`,stdio:`pipe`})}catch(e){return W(`npm update failed: ${G(e)}`)}let e=f(),i=await r.load(e.path,{forceAssetSync:!0,transform:o}),a=i.success&&i.data?i.data.manifest:null,c=i.success&&i.data?i.data.format:s.format;if(a){let t=n.register({...s,version:a.version,format:c,installPath:e.path,manifest:a,updatedAt:new Date().toISOString()});if(!t.success)return W(`Failed to refresh flow registry entry: ${t.error}`)}return W(`Flow "${t}" updated via npm successfully.`)}let c=i.update(s.installPath);if(!c.success)return W(`Update failed: ${c.error}`);let d=await r.load(s.installPath,{forceAssetSync:!0,transform:o}),m=d.success&&d.data?d.data.manifest:null,g=d.success&&d.data?d.data.format:s.format;if(m){let e=i.getLocalCommit(s.installPath),t=n.register({...s,version:m.version,format:g,manifest:m,updatedAt:new Date().toISOString(),...e?{commitSha:e}:{}});if(!t.success)return W(`Failed to refresh flow registry entry: ${t.error}`)}let v=m?.install??s.manifest.install;if(v.length>0){let e=i.runInstallDeps(v);if(!e.success)return W(`Dependency install failed: ${e.error}`)}return W(`Flow "${t}" updated successfully.`)}catch(e){return U.error(`flow_update failed`,A(e)),W(`Error: ${G(e)}`)}})}const go=E(`tools`);function _o(e){let t=L(`evidence_map`);e.registerTool(`evidence_map`,{title:t.title,description:`Track verified/assumed/unresolved claims for complex tasks. Gate readiness: YIELD (proceed), HOLD (unknowns), HARD_BLOCK (critical gaps). Persists across calls.`,inputSchema:{action:F.enum([`create`,`add`,`update`,`get`,`gate`,`list`,`delete`]).describe(`Operation to perform`),task_id:F.string().optional().describe(`Task identifier (required for all except list)`),tier:F.enum([`floor`,`standard`,`critical`]).optional().describe(`FORGE tier (required for create)`),claim:F.string().optional().describe(`Critical-path claim text (for add)`),status:F.enum([`V`,`A`,`U`]).optional().describe(`Evidence status: V=Verified, A=Assumed, U=Unresolved`),receipt:F.string().optional().describe(`Evidence receipt: tool→ref for V, reasoning for A, attempts for U`),id:F.number().optional().describe(`Entry ID (for update)`),critical_path:F.boolean().default(!0).describe(`Whether this claim is on the critical path`),unknown_type:F.enum([`contract`,`convention`,`freshness`,`runtime`,`data-flow`,`impact`]).optional().describe(`Typed unknown classification`),safety_gate:F.enum([`provenance`,`commitment`,`coverage`]).optional().describe(`Safety gate tag: provenance (claim→evidence), commitment (user-approved), coverage (nothing dropped)`),retry_count:F.number().default(0).describe(`Retry count for gate evaluation (0 = first attempt)`),cwd:F.string().optional().describe(`Working directory for evidence map storage (default: server cwd). Use root_path from forge_ground to match.`)},annotations:t.annotations},async({action:e,task_id:t,tier:n,claim:r,status:i,receipt:a,id:o,critical_path:s,unknown_type:c,safety_gate:l,retry_count:u,cwd:d})=>{try{switch(e){case`create`:if(!t)throw Error(`task_id required for create`);if(!n)throw Error(`tier required for create`);return Be({action:`create`,taskId:t,tier:n},d),{content:[{type:`text`,text:`Created evidence map "${t}" (tier: ${n}).\n\n---\n_Next: Use \`evidence_map\` with action "add" to record critical-path claims._`}]};case`add`:{if(!t)throw Error(`task_id required for add`);if(!r)throw Error(`claim required for add`);if(!i)throw Error(`status required for add`);let e=Be({action:`add`,taskId:t,claim:r,status:i,receipt:a??``,criticalPath:s,unknownType:c,safetyGate:l},d),n=[`Added entry #${e.entry?.id} to "${t}": [${i}] ${r}`];return e.formattedMap&&n.push(``,e.formattedMap),{content:[{type:`text`,text:n.join(`
562
562
  `)}]}}case`update`:{if(!t)throw Error(`task_id required for update`);if(o===void 0)throw Error(`id required for update`);if(!i)throw Error(`status required for update`);let e=Be({action:`update`,taskId:t,id:o,status:i,receipt:a??``},d),n=[`Updated entry #${o} in "${t}" → ${i}`];return e.formattedMap&&n.push(``,e.formattedMap),{content:[{type:`text`,text:n.join(`
563
563
  `)}]}}case`get`:{if(!t)throw Error(`task_id required for get`);let e=Be({action:`get`,taskId:t},d);return e.state?{content:[{type:`text`,text:[`## Evidence Map: ${t} (${e.state.tier})`,``,e.formattedMap??`No entries.`,``,`_${e.state.entries.length} entries — created ${e.state.createdAt}_`].join(`
564
564
  `)}]}:{content:[{type:`text`,text:`Evidence map "${t}" not found.`}]}}case`gate`:{if(!t)throw Error(`task_id required for gate`);let e=Be({action:`gate`,taskId:t,retryCount:u},d);if(!e.gate)return{content:[{type:`text`,text:`Evidence map "${t}" not found.`}]};let n=e.gate,r=[`## FORGE Gate: **${n.decision}**`,``,`**Reason:** ${n.reason}`,``,`**Stats:** ${n.stats.verified}V / ${n.stats.assumed}A / ${n.stats.unresolved}U (${n.stats.total} total)`];return n.warnings.length>0&&r.push(``,`**Warnings:**`,...n.warnings.map(e=>`- ⚠️ ${e}`)),n.unresolvedCritical.length>0&&r.push(``,`**Blocking entries:**`,...n.unresolvedCritical.map(e=>`- #${e.id}: ${e.claim} [${e.unknownType??`untyped`}]`)),n.safetyGates&&(r.push(``,`**Safety Gates:**`,`- Provenance: ${n.safetyGates.provenance}`,`- Commitment: ${n.safetyGates.commitment}`,`- Coverage: ${n.safetyGates.coverage}`),n.safetyGates.failures.length>0&&r.push(``,`**Safety failures:**`,...n.safetyGates.failures.map(e=>`- \u{1F6D1} ${e}`))),n.annotation&&r.push(``,`**Annotation:**`,n.annotation),e.formattedMap&&r.push(``,`---`,``,e.formattedMap),r.push(``,`---`,`_Next: ${n.decision===`YIELD`?`Proceed to implementation.`:n.decision===`HOLD`?`Resolve blocking entries, then re-run gate.`:n.decision===`HARD_BLOCK`?`Contract unknowns MUST be resolved. Use search or schema_validate.`:`Proceed with annotation — document unresolved items in delivery.`}_`),{content:[{type:`text`,text:r.join(`
565
565
  `)}]}}case`list`:{let e=Be({action:`list`},d);return!e.states||e.states.length===0?{content:[{type:`text`,text:`No evidence maps found.`}]}:{content:[{type:`text`,text:e.states.map(e=>`- **${e.taskId}** (${e.tier}) — ${e.entries.length} entries — ${e.updatedAt}`).join(`
566
- `)}]}}case`delete`:if(!t)throw Error(`task_id required for delete`);return{content:[{type:`text`,text:Be({action:`delete`,taskId:t},d).deleted?`Deleted evidence map "${t}".`:`Evidence map "${t}" not found.`}]}}}catch(e){return _o.error(`Evidence map error`,A(e)),{content:[{type:`text`,text:`Evidence map error: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function yo(e,t){let n=I(`digest`);e.registerTool(`digest`,{title:n.title,description:`Compress multiple text sources (handoffs, debates, phase summaries) into a token-budgeted digest. Jointly ranks across all sources, pins structured fields (status, files, decisions, blockers), and allocates budget by priority weight.`,inputSchema:{sources:F.array(F.object({id:F.string().describe(`Source identifier (e.g., "phase-2-handoff")`),text:F.string().max(5e5).describe(`Source text to compress`),weight:F.number().min(0).default(1).describe(`Priority weight (higher = more budget)`)})).min(1).max(20).describe(`Text sources to compress`),query:F.string().describe(`Focus query — what matters for the next step?`),max_chars:F.number().min(100).max(5e4).default(4e3).describe(`Target budget in characters`),pin_fields:F.array(F.string()).optional().describe(`Key fields to always extract (default: status, files, decisions, blockers, next)`),segmentation:F.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split text for scoring`),token_budget:F.number().min(50).max(12500).optional().describe(`Token budget — overrides max_chars (approx 4 chars per token). Use to fit output into a specific context window.`)},annotations:n.annotations},async({sources:e,query:n,max_chars:r,pin_fields:i,segmentation:a,token_budget:o})=>{try{let s=await Ie(t,{sources:e,query:n,maxChars:o?o*4:r,pinFields:i,segmentation:a}),c=[`## Digest (${s.totalOriginalChars} → ${s.totalCompressedChars} chars, ${(s.ratio*100).toFixed(0)}%)`,``],l=Object.keys(s.fields);if(l.length>0){c.push(`### Extracted Fields`);for(let e of l){let t=s.fields[e];c.push(`**${e}:**`);for(let e of t)c.push(` - [${e.sourceId}] ${e.value}`)}c.push(``)}c.push(`### Compressed Content`,``,s.text),c.push(``,`### Source Stats`);for(let e of s.sourceStats)c.push(`- **${e.id}**: ${e.originalChars} → ${e.keptChars} chars (${e.segmentsKept}/${e.segmentsTotal} segments)`);return c.push(``,`---`,"_Next: Use the digest as compressed context for the next phase. Use `stash` to persist it if needed._"),{content:[{type:`text`,text:c.join(`
567
- `)}]}}catch(e){return _o.error(`Digest failed`,A(e)),{content:[{type:`text`,text:`Digest failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function bo(e){let t=I(`forge_classify`);e.registerTool(`forge_classify`,{title:t.title,description:`Classify FORGE tier (Floor/Standard/Critical) from target files and task description. Checks blast radius, cross-package boundaries, schema/contract patterns, and security signals. Returns tier, triggers, typed unknown seeds, and ceremony guidance.`,inputSchema:{files:F.array(F.string()).min(1).max(100).describe(`Files being modified (paths)`),task:F.string().describe(`Task description`),root_path:F.string().describe(`Root path of the codebase`)},annotations:t.annotations},async({files:e,task:t,root_path:n})=>{try{let r=await Ge({files:e,task:t,rootPath:n}),i=[`## FORGE Classification: **${r.tier.toUpperCase()}**`,``];if(r.triggers.length>0){i.push(`### Triggers`);for(let e of r.triggers)i.push(`- **${e.rule}** (${e.source}): ${e.detail}`);i.push(``)}if(r.packagesCrossed.length>0&&i.push(`**Packages crossed:** ${r.packagesCrossed.join(`, `)}`),r.typedUnknownSeeds.length>0){i.push(``,`### Typed Unknown Seeds`);for(let e of r.typedUnknownSeeds)i.push(`- [${e.type}] ${e.description} → use \`${e.suggestedTool}\``)}i.push(``,`### Ceremony`);let a=r.ceremony;return i.push(`- **Ground:** ${a.ground}`,`- **Build:** ${a.build}`,`- **Break:** ${a.break}`,`- **Evidence Map:** ${a.evidenceMap}`,`- **Gate:** ${a.gate}`),i.push(``,`---`,`_Next: ${r.tier===`floor`?`Proceed directly to implementation.`:"Run `forge_ground` to execute the full Ground phase."}_`),{content:[{type:`text`,text:i.join(`
568
- `)}]}}catch(e){return _o.error(`FORGE classify failed`,A(e)),{content:[{type:`text`,text:`FORGE classify failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function xo(e,t,n){let r=I(`stratum_card`);e.registerTool(`stratum_card`,{title:r.title,description:`Generate context cards from files — a compressed alternative to reading full files (10-100x token reduction). Tier 1 (T1): structure only (~100 tokens/file — imports, exports, functions). Tier 2 (T2): structure + query-relevant content (~300 tokens/file).`,inputSchema:{files:F.array(F.string()).min(1).max(50).describe(`Absolute file paths to generate cards for`),query:F.string().describe(`Current task query — guides relevance scoring`),tier:F.enum([`T1`,`T2`]).default(`T1`).describe(`Card tier: T1 = structural only, T2 = T1 + compressed content`),max_content_chars:F.number().min(100).max(5e3).default(800).describe(`For T2: max chars for compressed content section`)},annotations:r.annotations},async({files:e,query:r,tier:i,max_content_chars:a})=>{try{let o=await Wt(t,{files:e,query:r,tier:i,maxContentChars:a,cache:n}),s=[`## STRATUM Cards (${i}) — ${o.cards.length} files`,`Total: ~${o.totalTokenEstimate} tokens (was ~${o.totalOriginalTokenEstimate}, ${(o.compressionRatio*100).toFixed(0)}% of original)`,``];for(let e of o.cards)s.push(e.card,``);return s.push(`---`,"_Next: Use these cards as context instead of reading full files. Use `compact` for deeper content extraction on specific files._"),{content:[{type:`text`,text:s.join(`
569
- `)}]}}catch(e){return _o.error(`STRATUM cards failed`,A(e)),{content:[{type:`text`,text:`STRATUM cards failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function So(e,t,n){let r=I(`forge_ground`);e.registerTool(`forge_ground`,{title:r.title,description:`Execute the complete FORGE Ground phase in a single call. Chains: tier classification → scope map → typed unknown seeds → constraint loading → file summaries → evidence map creation. Replaces 5-15 manual tool calls.`,inputSchema:{task:F.string().describe(`Task description`),files:F.array(F.string()).min(1).max(100).describe(`Target files being modified (absolute paths)`),root_path:F.string().describe(`Root path of the codebase`),max_constraints:F.number().min(0).max(10).default(3).describe(`Max constraint entries to load from KB`),force_tier:F.enum([`floor`,`standard`,`critical`]).optional().describe(`Force a specific tier (skips auto-classification)`),task_id:F.string().optional().describe(`Custom task ID for evidence map (auto-generated if omitted)`)},annotations:r.annotations},async({task:e,files:r,root_path:i,max_constraints:a,force_tier:o,task_id:s},c)=>{try{let l=await Ke(t,n,{task:e,files:r,rootPath:i,maxConstraints:a,forceTier:o,taskId:s}),u=Yi(c).createTask(`FORGE Ground`,3);u.progress(0,`Classification: ${l.tier}`),u.progress(1,`Files analyzed: ${l.fileSummaries.length}`),u.progress(2,`Unknowns: ${l.typedUnknownSeeds.length}`),u.complete(`FORGE Ground complete: tier=${l.tier}`);let d=[`## FORGE Ground: **${l.tier.toUpperCase()}**`,`Estimated output: ~${l.estimatedTokens} tokens`,``];if(l.classifyTriggers.length>0){d.push(`### Classification Triggers`);for(let e of l.classifyTriggers)d.push(`- **${e.rule}** (${e.source}): ${e.detail}`);d.push(``)}if(l.fileSummaries.length>0){d.push(`### Target Files`);for(let e of l.fileSummaries)d.push(`- **${e.path}** (${e.lines} lines) — exports: ${e.exports.join(`, `)||`none`} — functions: ${e.functions.join(`, `)||`none`}`);d.push(``)}if(l.typedUnknownSeeds.length>0){d.push(`### Typed Unknown Seeds`);for(let e of l.typedUnknownSeeds)d.push(`- [${e.type}] ${e.description} → \`${e.suggestedTool}\``);d.push(``)}if(l.constraints.length>0){d.push(`### Constraint Seed`);for(let e of l.constraints)d.push(`- **${e.source}** (${(e.relevance*100).toFixed(0)}%): ${e.snippet}`);d.push(``)}if(l.scopeMap){d.push(`### Scope Map: ${l.scopeMap.files.length} files (~${l.scopeMap.totalEstimatedTokens} tokens)`);for(let e of l.scopeMap.files.slice(0,5))d.push(`- ${e.path} (${(e.relevance*100).toFixed(0)}%) — ${e.reason}`);l.scopeMap.files.length>5&&d.push(`- _...and ${l.scopeMap.files.length-5} more_`),d.push(``)}l.evidenceMapTaskId&&(d.push(`**Evidence Map:** \`${l.evidenceMapTaskId}\` (initialized)`),d.push(``)),d.push(`### Ceremony Guidance`);let f=l.ceremony;return d.push(`- **Ground:** ${f.ground}`,`- **Build:** ${f.build}`,`- **Break:** ${f.break}`,`- **Evidence Map:** ${f.evidenceMap}`,`- **Gate:** ${f.gate}`),d.push(``,`---`,`_Next: ${l.tier===`floor`?`Proceed to Build phase.`:"Use `evidence_map` to track claims during Build, then `evidence_map` gate after Break."}_`),{content:[{type:`text`,text:d.join(`
570
- `)}]}}catch(e){return _o.error(`FORGE Ground failed`,A(e)),{content:[{type:`text`,text:`FORGE Ground failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Co=E(`tools`);function wo(e,t,n){let r=I(`forget`);e.registerTool(`forget`,{title:r.title,description:`Remove a curated knowledge entry. Deletes the file and removes it from the vector store.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/deprecated-approach.md")`),reason:F.string().min(3).describe(`Why this entry is being removed`)},annotations:r.annotations},async({path:e,reason:r})=>{try{let i=_(process.cwd(),y.aiCurated,e);if(a(i))try{Ae(`forget`,[{path:i,content:s(i,`utf-8`)}],`forget: ${e}`)}catch{}let o=await t.forget(e,r);return Co.info(`Forgot entry`,{path:o.path,reason:r}),n&&(n.notifyAfterCuratedWrite().catch(()=>{}),n.notifyResourceListChanged().catch(()=>{})),{content:[{type:`text`,text:`Forgotten: \`.ai/curated/${o.path}\`\n\nRemoved from disk and vector store.\nReason: ${r}\n\n---\n_Next: Use \`list\` to verify removal, or \`search\` to confirm the entry is no longer returned._`}]}}catch(e){return Co.error(`Forget failed`,A(e)),{content:[{type:`text`,text:`Forget failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const To=E(`tools`),Eo=F.object({id:F.string().optional().describe(`Node ID (auto-generated if omitted)`),type:F.string().describe(`Node type (entity, service, api, concept, decision)`),name:F.string().describe(`Display name`),properties:F.record(F.string(),F.unknown()).optional().describe(`Arbitrary properties`),sourceRecordId:F.string().optional().describe(`Back-link to knowledge record ID`),sourcePath:F.string().optional().describe(`Source file path`)}),Do=F.object({id:F.string().optional().describe(`Edge ID (auto-generated if omitted)`),fromId:F.string().describe(`Source node ID`),toId:F.string().describe(`Target node ID`),type:F.string().describe(`Relationship type (depends-on, implements, calls, affects)`),weight:F.number().min(0).max(1).optional().describe(`Relationship weight`),properties:F.record(F.string(),F.unknown()).optional().describe(`Arbitrary properties`)});function Oo(e,t){let n=I(`graph`);e.registerTool(`graph`,{title:n.title,description:`Query the auto-populated code knowledge graph (modules, symbols, imports) to answer
566
+ `)}]}}case`delete`:if(!t)throw Error(`task_id required for delete`);return{content:[{type:`text`,text:Be({action:`delete`,taskId:t},d).deleted?`Deleted evidence map "${t}".`:`Evidence map "${t}" not found.`}]}}}catch(e){return go.error(`Evidence map error`,A(e)),{content:[{type:`text`,text:`Evidence map error: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function vo(e,t){let n=L(`digest`);e.registerTool(`digest`,{title:n.title,description:`Compress multiple text sources (handoffs, debates, phase summaries) into a token-budgeted digest. Jointly ranks across all sources, pins structured fields (status, files, decisions, blockers), and allocates budget by priority weight.`,inputSchema:{sources:F.array(F.object({id:F.string().describe(`Source identifier (e.g., "phase-2-handoff")`),text:F.string().max(5e5).describe(`Source text to compress`),weight:F.number().min(0).default(1).describe(`Priority weight (higher = more budget)`)})).min(1).max(20).describe(`Text sources to compress`),query:F.string().describe(`Focus query — what matters for the next step?`),max_chars:F.number().min(100).max(5e4).default(4e3).describe(`Target budget in characters`),pin_fields:F.array(F.string()).optional().describe(`Key fields to always extract (default: status, files, decisions, blockers, next)`),segmentation:F.enum([`paragraph`,`sentence`,`line`]).default(`paragraph`).describe(`How to split text for scoring`),token_budget:F.number().min(50).max(12500).optional().describe(`Token budget — overrides max_chars (approx 4 chars per token). Use to fit output into a specific context window.`)},annotations:n.annotations},async({sources:e,query:n,max_chars:r,pin_fields:i,segmentation:a,token_budget:o})=>{try{let s=await Ie(t,{sources:e,query:n,maxChars:o?o*4:r,pinFields:i,segmentation:a}),c=[`## Digest (${s.totalOriginalChars} → ${s.totalCompressedChars} chars, ${(s.ratio*100).toFixed(0)}%)`,``],l=Object.keys(s.fields);if(l.length>0){c.push(`### Extracted Fields`);for(let e of l){let t=s.fields[e];c.push(`**${e}:**`);for(let e of t)c.push(` - [${e.sourceId}] ${e.value}`)}c.push(``)}c.push(`### Compressed Content`,``,s.text),c.push(``,`### Source Stats`);for(let e of s.sourceStats)c.push(`- **${e.id}**: ${e.originalChars} → ${e.keptChars} chars (${e.segmentsKept}/${e.segmentsTotal} segments)`);return c.push(``,`---`,"_Next: Use the digest as compressed context for the next phase. Use `stash` to persist it if needed._"),{content:[{type:`text`,text:c.join(`
567
+ `)}]}}catch(e){return go.error(`Digest failed`,A(e)),{content:[{type:`text`,text:`Digest failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function yo(e){let t=L(`forge_classify`);e.registerTool(`forge_classify`,{title:t.title,description:`Classify FORGE tier (Floor/Standard/Critical) from target files and task description. Checks blast radius, cross-package boundaries, schema/contract patterns, and security signals. Returns tier, triggers, typed unknown seeds, and ceremony guidance.`,inputSchema:{files:F.array(F.string()).min(1).max(100).describe(`Files being modified (paths)`),task:F.string().describe(`Task description`),root_path:F.string().describe(`Root path of the codebase`)},annotations:t.annotations},async({files:e,task:t,root_path:n})=>{try{let r=await Ge({files:e,task:t,rootPath:n}),i=[`## FORGE Classification: **${r.tier.toUpperCase()}**`,``];if(r.triggers.length>0){i.push(`### Triggers`);for(let e of r.triggers)i.push(`- **${e.rule}** (${e.source}): ${e.detail}`);i.push(``)}if(r.packagesCrossed.length>0&&i.push(`**Packages crossed:** ${r.packagesCrossed.join(`, `)}`),r.typedUnknownSeeds.length>0){i.push(``,`### Typed Unknown Seeds`);for(let e of r.typedUnknownSeeds)i.push(`- [${e.type}] ${e.description} → use \`${e.suggestedTool}\``)}i.push(``,`### Ceremony`);let a=r.ceremony;return i.push(`- **Ground:** ${a.ground}`,`- **Build:** ${a.build}`,`- **Break:** ${a.break}`,`- **Evidence Map:** ${a.evidenceMap}`,`- **Gate:** ${a.gate}`),i.push(``,`---`,`_Next: ${r.tier===`floor`?`Proceed directly to implementation.`:"Run `forge_ground` to execute the full Ground phase."}_`),{content:[{type:`text`,text:i.join(`
568
+ `)}]}}catch(e){return go.error(`FORGE classify failed`,A(e)),{content:[{type:`text`,text:`FORGE classify failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function bo(e,t,n){let r=L(`stratum_card`);e.registerTool(`stratum_card`,{title:r.title,description:`Generate context cards from files — a compressed alternative to reading full files (10-100x token reduction). Tier 1 (T1): structure only (~100 tokens/file — imports, exports, functions). Tier 2 (T2): structure + query-relevant content (~300 tokens/file).`,inputSchema:{files:F.array(F.string()).min(1).max(50).describe(`Absolute file paths to generate cards for`),query:F.string().describe(`Current task query — guides relevance scoring`),tier:F.enum([`T1`,`T2`]).default(`T1`).describe(`Card tier: T1 = structural only, T2 = T1 + compressed content`),max_content_chars:F.number().min(100).max(5e3).default(800).describe(`For T2: max chars for compressed content section`)},annotations:r.annotations},async({files:e,query:r,tier:i,max_content_chars:a})=>{try{let o=await Wt(t,{files:e,query:r,tier:i,maxContentChars:a,cache:n}),s=[`## STRATUM Cards (${i}) — ${o.cards.length} files`,`Total: ~${o.totalTokenEstimate} tokens (was ~${o.totalOriginalTokenEstimate}, ${(o.compressionRatio*100).toFixed(0)}% of original)`,``];for(let e of o.cards)s.push(e.card,``);return s.push(`---`,"_Next: Use these cards as context instead of reading full files. Use `compact` for deeper content extraction on specific files._"),{content:[{type:`text`,text:s.join(`
569
+ `)}]}}catch(e){return go.error(`STRATUM cards failed`,A(e)),{content:[{type:`text`,text:`STRATUM cards failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function xo(e,t,n){let r=L(`forge_ground`);e.registerTool(`forge_ground`,{title:r.title,description:`Execute the complete FORGE Ground phase in a single call. Chains: tier classification → scope map → typed unknown seeds → constraint loading → file summaries → evidence map creation. Replaces 5-15 manual tool calls.`,inputSchema:{task:F.string().describe(`Task description`),files:F.array(F.string()).min(1).max(100).describe(`Target files being modified (absolute paths)`),root_path:F.string().describe(`Root path of the codebase`),max_constraints:F.number().min(0).max(10).default(3).describe(`Max constraint entries to load from KB`),force_tier:F.enum([`floor`,`standard`,`critical`]).optional().describe(`Force a specific tier (skips auto-classification)`),task_id:F.string().optional().describe(`Custom task ID for evidence map (auto-generated if omitted)`)},annotations:r.annotations},async({task:e,files:r,root_path:i,max_constraints:a,force_tier:o,task_id:s},c)=>{try{let l=await Ke(t,n,{task:e,files:r,rootPath:i,maxConstraints:a,forceTier:o,taskId:s}),u=Ji(c).createTask(`FORGE Ground`,3);u.progress(0,`Classification: ${l.tier}`),u.progress(1,`Files analyzed: ${l.fileSummaries.length}`),u.progress(2,`Unknowns: ${l.typedUnknownSeeds.length}`),u.complete(`FORGE Ground complete: tier=${l.tier}`);let d=[`## FORGE Ground: **${l.tier.toUpperCase()}**`,`Estimated output: ~${l.estimatedTokens} tokens`,``];if(l.classifyTriggers.length>0){d.push(`### Classification Triggers`);for(let e of l.classifyTriggers)d.push(`- **${e.rule}** (${e.source}): ${e.detail}`);d.push(``)}if(l.fileSummaries.length>0){d.push(`### Target Files`);for(let e of l.fileSummaries)d.push(`- **${e.path}** (${e.lines} lines) — exports: ${e.exports.join(`, `)||`none`} — functions: ${e.functions.join(`, `)||`none`}`);d.push(``)}if(l.typedUnknownSeeds.length>0){d.push(`### Typed Unknown Seeds`);for(let e of l.typedUnknownSeeds)d.push(`- [${e.type}] ${e.description} → \`${e.suggestedTool}\``);d.push(``)}if(l.constraints.length>0){d.push(`### Constraint Seed`);for(let e of l.constraints)d.push(`- **${e.source}** (${(e.relevance*100).toFixed(0)}%): ${e.snippet}`);d.push(``)}if(l.scopeMap){d.push(`### Scope Map: ${l.scopeMap.files.length} files (~${l.scopeMap.totalEstimatedTokens} tokens)`);for(let e of l.scopeMap.files.slice(0,5))d.push(`- ${e.path} (${(e.relevance*100).toFixed(0)}%) — ${e.reason}`);l.scopeMap.files.length>5&&d.push(`- _...and ${l.scopeMap.files.length-5} more_`),d.push(``)}l.evidenceMapTaskId&&(d.push(`**Evidence Map:** \`${l.evidenceMapTaskId}\` (initialized)`),d.push(``)),d.push(`### Ceremony Guidance`);let f=l.ceremony;return d.push(`- **Ground:** ${f.ground}`,`- **Build:** ${f.build}`,`- **Break:** ${f.break}`,`- **Evidence Map:** ${f.evidenceMap}`,`- **Gate:** ${f.gate}`),d.push(``,`---`,`_Next: ${l.tier===`floor`?`Proceed to Build phase.`:"Use `evidence_map` to track claims during Build, then `evidence_map` gate after Break."}_`),{content:[{type:`text`,text:d.join(`
570
+ `)}]}}catch(e){return go.error(`FORGE Ground failed`,A(e)),{content:[{type:`text`,text:`FORGE Ground failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const So=E(`tools`);function Co(e,t,n){let r=L(`forget`);e.registerTool(`forget`,{title:r.title,description:`Remove a curated knowledge entry. Deletes the file and removes it from the vector store.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/deprecated-approach.md")`),reason:F.string().min(3).describe(`Why this entry is being removed`)},annotations:r.annotations},async({path:e,reason:r})=>{try{let i=_(process.cwd(),y.aiCurated,e);if(a(i))try{Ae(`forget`,[{path:i,content:s(i,`utf-8`)}],`forget: ${e}`)}catch{}let o=await t.forget(e,r);return So.info(`Forgot entry`,{path:o.path,reason:r}),n&&(n.notifyAfterCuratedWrite().catch(()=>{}),n.notifyResourceListChanged().catch(()=>{})),{content:[{type:`text`,text:`Forgotten: \`.ai/curated/${o.path}\`\n\nRemoved from disk and vector store.\nReason: ${r}\n\n---\n_Next: Use \`list\` to verify removal, or \`search\` to confirm the entry is no longer returned._`}]}}catch(e){return So.error(`Forget failed`,A(e)),{content:[{type:`text`,text:`Forget failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const wo=E(`tools`),To=F.object({id:F.string().optional().describe(`Node ID (auto-generated if omitted)`),type:F.string().describe(`Node type (entity, service, api, concept, decision)`),name:F.string().describe(`Display name`),properties:F.record(F.string(),F.unknown()).optional().describe(`Arbitrary properties`),sourceRecordId:F.string().optional().describe(`Back-link to knowledge record ID`),sourcePath:F.string().optional().describe(`Source file path`)}),Eo=F.object({id:F.string().optional().describe(`Edge ID (auto-generated if omitted)`),fromId:F.string().describe(`Source node ID`),toId:F.string().describe(`Target node ID`),type:F.string().describe(`Relationship type (depends-on, implements, calls, affects)`),weight:F.number().min(0).max(1).optional().describe(`Relationship weight`),properties:F.record(F.string(),F.unknown()).optional().describe(`Arbitrary properties`)});function Do(e,t){let n=L(`graph`);e.registerTool(`graph`,{title:n.title,description:`Query the auto-populated code knowledge graph (modules, symbols, imports) to answer
571
571
  structural questions vector search cannot.
572
572
 
573
573
  Common flow (when you don't have a node_id yet — TWO STEPS):
@@ -586,7 +586,7 @@ Single-step read actions:
586
586
  Mutating actions (use deliberately): add, delete, clear, set_community,
587
587
  detect_communities, trace_process, delete_process.
588
588
 
589
- Complements: \`symbol\` (single lookup), \`trace\` (call-chain AST), \`blast_radius\` (change impact).`,outputSchema:Ei,inputSchema:{action:F.enum([`find_nodes`,`find_edges`,`neighbors`,`traverse`,`stats`,`validate`,`add`,`delete`,`clear`,`detect_communities`,`set_community`,`trace_process`,`list_processes`,`delete_process`,`depth_traverse`,`cohesion`,`symbol360`]).describe(`Action: find_nodes (search nodes), find_edges (search edges), neighbors (direct connections), traverse (multi-hop), stats (graph overview), validate (check graph integrity), add (insert nodes/edges), delete (remove nodes), clear (remove all), detect_communities (find clusters), set_community (label a node), trace_process (persist execution flow), list_processes (list traced flows), delete_process (remove a flow), depth_traverse (bucket traversal by depth), cohesion (score a community), symbol360 (full node context)`),node_type:F.string().optional().describe(`Node type filter (for find_nodes)`),name_pattern:F.string().optional().describe(`Name substring match (for find_nodes)`),source_path:F.string().optional().describe(`Source path filter`),node_id:F.string().optional().describe(`Node ID (for neighbors, traverse, delete)`),edge_type:F.string().optional().describe(`Edge type filter`),from_id:F.string().optional().describe(`Source node ID (for find_edges)`),to_id:F.string().optional().describe(`Target node ID (for find_edges)`),direction:F.enum([`outgoing`,`incoming`,`both`]).default(`both`).describe(`Traversal direction`),max_depth:F.number().min(1).max(5).default(2).describe(`Max traversal depth`),limit:F.number().min(1).max(100).default(50).describe(`Max results`),nodes:F.array(Eo).max(500).optional().describe(`Nodes to add (for action=add)`),edges:F.array(Do).max(500).optional().describe(`Edges to add (for action=add)`),community:F.string().optional().describe(`Community label (for set_community, cohesion)`),process_id:F.string().optional().describe(`Process ID (for delete_process)`),label:F.string().optional().describe(`Label for process tracing (for trace_process)`)},annotations:n.annotations},async e=>{try{let n=await Xe(t,{action:e.action,nodeType:e.node_type,namePattern:e.name_pattern,sourcePath:e.source_path,nodeId:e.node_id,edgeType:e.edge_type,fromId:e.from_id,toId:e.to_id,direction:e.direction,maxDepth:e.max_depth,limit:e.limit,nodes:e.nodes,edges:e.edges,community:e.community,processId:e.process_id,label:e.label}),r=[n.summary];if(n.nodes&&n.nodes.length>0){r.push(`
589
+ Complements: \`symbol\` (single lookup), \`trace\` (call-chain AST), \`blast_radius\` (change impact).`,outputSchema:Ti,inputSchema:{action:F.enum([`find_nodes`,`find_edges`,`neighbors`,`traverse`,`stats`,`validate`,`add`,`delete`,`clear`,`detect_communities`,`set_community`,`trace_process`,`list_processes`,`delete_process`,`depth_traverse`,`cohesion`,`symbol360`]).describe(`Action: find_nodes (search nodes), find_edges (search edges), neighbors (direct connections), traverse (multi-hop), stats (graph overview), validate (check graph integrity), add (insert nodes/edges), delete (remove nodes), clear (remove all), detect_communities (find clusters), set_community (label a node), trace_process (persist execution flow), list_processes (list traced flows), delete_process (remove a flow), depth_traverse (bucket traversal by depth), cohesion (score a community), symbol360 (full node context)`),node_type:F.string().optional().describe(`Node type filter (for find_nodes)`),name_pattern:F.string().optional().describe(`Name substring match (for find_nodes)`),source_path:F.string().optional().describe(`Source path filter`),node_id:F.string().optional().describe(`Node ID (for neighbors, traverse, delete)`),edge_type:F.string().optional().describe(`Edge type filter`),from_id:F.string().optional().describe(`Source node ID (for find_edges)`),to_id:F.string().optional().describe(`Target node ID (for find_edges)`),direction:F.enum([`outgoing`,`incoming`,`both`]).default(`both`).describe(`Traversal direction`),max_depth:F.number().min(1).max(5).default(2).describe(`Max traversal depth`),limit:F.number().min(1).max(100).default(50).describe(`Max results`),nodes:F.array(To).max(500).optional().describe(`Nodes to add (for action=add)`),edges:F.array(Eo).max(500).optional().describe(`Edges to add (for action=add)`),community:F.string().optional().describe(`Community label (for set_community, cohesion)`),process_id:F.string().optional().describe(`Process ID (for delete_process)`),label:F.string().optional().describe(`Label for process tracing (for trace_process)`)},annotations:n.annotations},async e=>{try{let n=await Xe(t,{action:e.action,nodeType:e.node_type,namePattern:e.name_pattern,sourcePath:e.source_path,nodeId:e.node_id,edgeType:e.edge_type,fromId:e.from_id,toId:e.to_id,direction:e.direction,maxDepth:e.max_depth,limit:e.limit,nodes:e.nodes,edges:e.edges,community:e.community,processId:e.process_id,label:e.label}),r=[n.summary];if(n.nodes&&n.nodes.length>0){r.push(`
590
590
  ### Nodes`);for(let e of n.nodes){let t=Object.keys(e.properties).length>0?` — ${JSON.stringify(e.properties)}`:``;r.push(`- **${e.name}** (${e.type}, id: \`${e.id}\`)${t}`)}}if(n.edges&&n.edges.length>0){r.push(`
591
591
  ### Edges`);for(let e of n.edges)r.push(`- \`${e.fromId}\` —[${e.type}]→ \`${e.toId}\`${e.weight===1?``:` (weight: ${e.weight})`}`)}if(n.stats&&(r.push(`\nNode types: ${JSON.stringify(n.stats.nodeTypes)}`),r.push(`Edge types: ${JSON.stringify(n.stats.edgeTypes)}`)),n.validation){if(r.push(`
592
592
  ### Validation`),r.push(`- **Valid**: ${n.validation.valid?`yes`:`no`}`),n.validation.danglingEdges.length>0){r.push(`- **Dangling edges**:`);for(let e of n.validation.danglingEdges)r.push(` - \`${e.edgeId}\` references missing node \`${e.missingNodeId}\``)}n.validation.orphanNodes.length>0&&r.push(`- **Orphan nodes**: ${n.validation.orphanNodes.map(e=>`\`${e}\``).join(`, `)}`)}if(n.communities){r.push(`
@@ -595,25 +595,25 @@ Complements: \`symbol\` (single lookup), \`trace\` (call-chain AST), \`blast_rad
595
595
  ### Processes`);for(let e of n.processes)r.push(`- **${e.label}** (id: \`${e.id}\`) — ${e.steps.length} step(s)`)}if(n.depthGroups){r.push(`
596
596
  ### Depth Groups`);for(let[e,t]of Object.entries(n.depthGroups))r.push(`- **Depth ${e}**: ${t.map(e=>`${e.name} (${e.type})`).join(`, `)}`)}if(n.cohesionScore!==void 0&&r.push(`\n**Cohesion Score**: ${(n.cohesionScore*100).toFixed(1)}%`),n.symbol360){let e=n.symbol360;r.push(`
597
597
  ### 360° Symbol View`),r.push(`- **Node**: ${e.node.name} (${e.node.type})`),r.push(`- **Community**: ${e.community??`none`}`),r.push(`- **Incoming**: ${e.incoming.length} edge(s)`),r.push(`- **Outgoing**: ${e.outgoing.length} edge(s)`),r.push(`- **Processes**: ${e.processes.length}`)}return r.push("\n---\n_Next: Use `graph(traverse)` to explore connections, `graph(add)` to insert entities, `graph(symbol360)` for full node context, or `graph(detect_communities)` to find clusters._"),{content:[{type:`text`,text:r.join(`
598
- `)}],structuredContent:{nodes:(n.nodes??[]).map(e=>({id:e.id,name:e.name,type:e.type,...e.sourcePath?{sourcePath:e.sourcePath}:{}})),edges:(n.edges??[]).map(e=>({fromId:e.fromId,toId:e.toId,type:e.type})),totalNodes:n.stats?.nodeCount??n.nodes?.length??0,totalEdges:n.stats?.edgeCount??n.edges?.length??0,query:e.action}}}catch(e){return To.error(`Graph query failed`,A(e)),{content:[{type:`text`,text:`Graph query failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const G=E(`tools:infra`);function ko(e){let t=I(`process`);e.registerTool(`process`,{title:t.title,description:`Start, stop, inspect, list, and tail logs for in-memory managed child processes.`,inputSchema:{action:F.enum([`start`,`stop`,`status`,`list`,`logs`]).describe(`Process action to perform`),id:F.string().optional().describe(`Managed process ID`),command:F.string().optional().describe(`Executable to start`),args:F.array(F.string()).optional().describe(`Arguments for start actions`),tail:F.number().min(1).max(500).optional().describe(`Log lines to return for logs actions`)},annotations:t.annotations},async({action:e,id:t,command:n,args:r,tail:i})=>{try{switch(e){case`start`:if(!t||!n)throw Error(`id and command are required for start`);return{content:[{type:`text`,text:JSON.stringify(pt(t,n,r??[]))}]};case`stop`:if(!t)throw Error(`id is required for stop`);return{content:[{type:`text`,text:JSON.stringify(ht(t)??null)}]};case`status`:if(!t)throw Error(`id is required for status`);return{content:[{type:`text`,text:JSON.stringify(mt(t)??null)}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(dt())}]};case`logs`:if(!t)throw Error(`id is required for logs`);return{content:[{type:`text`,text:JSON.stringify(ft(t,i))}]}}}catch(e){return G.error(`Process action failed`,A(e)),{content:[{type:`text`,text:`Process action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ao(e){let t=I(`watch`);e.registerTool(`watch`,{title:t.title,description:`Watch a directory for file changes (create/modify/delete). Actions: start (begin watching), stop (by ID), list (show active watchers). Events are emitted as structured JSON with path, event type, and timestamp.`,inputSchema:{action:F.enum([`start`,`stop`,`list`]).describe(`Watch action to perform`),path:F.string().optional().describe(`Directory path to watch for start actions`),id:F.string().optional().describe(`Watcher ID for stop actions`)},annotations:t.annotations},async({action:e,path:t,id:n})=>{try{switch(e){case`start`:if(!t)throw Error(`path is required for start`);return{content:[{type:`text`,text:JSON.stringify(Qt({path:t}))}]};case`stop`:if(!n)throw Error(`id is required for stop`);return{content:[{type:`text`,text:JSON.stringify({stopped:$t(n)})}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(Zt())}]}}}catch(e){return G.error(`Watch action failed`,A(e)),{content:[{type:`text`,text:`Watch action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function jo(e){let t=I(`health`);e.registerTool(`health`,{title:t.title,description:`Run project health checks — verifies package.json, tsconfig, scripts, lockfile, README, LICENSE, .gitignore.`,outputSchema:si,inputSchema:{path:F.string().optional().describe(`Root directory to check (defaults to cwd)`)},annotations:t.annotations},async({path:e})=>{try{let t=Qe(e),n={ok:t.checks.every(e=>e.status!==`fail`),checks:t.checks.map(e=>({name:e.name,ok:e.status===`pass`,message:e.message}))};return{content:[{type:`text`,text:JSON.stringify(t)}],structuredContent:n}}catch(e){return G.error(`Health check failed`,A(e)),{content:[{type:`text`,text:`Health check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Mo(e){let t=I(`web_fetch`);e.registerTool(`web_fetch`,{title:t.title,description:`PREFERRED web fetcher — fetch one or many URLs and convert to LLM-optimized markdown. Pass one URL or multiple for parallel fetching. Supports CSS selectors, 4 output modes (markdown/raw/links/outline), smart paragraph-boundary truncation. Strips scripts/styles/nav automatically.`,inputSchema:{urls:F.array(F.string().url()).min(1).max(10).describe('URLs to fetch (1–10). Single URL: `["https://..."]`. Multiple fetched in parallel.'),mode:F.enum([`markdown`,`raw`,`links`,`outline`]).default(`markdown`).describe(`Output mode: markdown (clean content), raw (HTML), links (extracted URLs), outline (heading hierarchy)`),selector:F.string().optional().describe(`CSS selector to extract a specific element instead of auto-detecting main content`),max_length:F.number().min(500).max(1e5).default(15e3).describe(`Max characters in output — truncates at paragraph boundaries`),include_metadata:F.boolean().default(!0).describe(`Include page title, description, and URL as a header`),include_links:F.boolean().default(!1).describe(`Append extracted links list at the end`),include_images:F.boolean().default(!1).describe(`Include image alt texts inline`),timeout:F.number().min(1e3).max(12e4).default(3e4).describe(`Request timeout in milliseconds`)},annotations:t.annotations},async({urls:e,mode:t,selector:n,max_length:r,include_metadata:i,include_links:a,include_images:o,timeout:s})=>{let c=e,l=async(e,c)=>{let l=await en({url:e,mode:t,selector:n,maxLength:r,includeMetadata:i,includeLinks:a,includeImages:o,timeout:s}),u=[c?`## ${c} ${l.title||`Web Page`}\n> Source: ${e}`:`## ${l.title||`Web Page`}`,``,l.content];return l.truncated&&u.push(``,`_Original length: ${l.originalLength.toLocaleString()} chars_`),u.join(`
599
- `)};if(c.length===1)try{return{content:[{type:`text`,text:await l(c[0])+"\n\n---\n_Next: Use `remember` to save key findings, or `web_fetch` with a `selector` to extract a specific section._"}]}}catch(e){let t=e instanceof Error?e.message:String(e);return/HTTP [45]\d{2}/.test(t)?G.warn(`Web fetch failed`,{error:t}):G.error(`Web fetch failed`,A(e)),{content:[{type:`text`,text:`Web fetch failed: ${t}`}],isError:!0}}let u=c.length,d=await Promise.allSettled(c.map((e,t)=>l(e,`[${t+1}/${u}]`))),f=[],p=0;for(let e=0;e<d.length;e++){let t=d[e];if(t.status===`fulfilled`)f.push(t.value);else{p++;let n=t.reason instanceof Error?t.reason.message:String(t.reason);/HTTP [45]\d{2}/.test(n)?G.warn(`Web fetch failed`,{url:c[e],error:n}):G.error(`Web fetch failed`,{url:c[e],...A(t.reason)}),f.push(`## ❌ Failed: ${c[e]}\n\n${n}`)}}let m=`_Fetched ${d.length-p}/${d.length} URLs successfully._`;return f.push(``,`---`,m,"_Next: Use `remember` to save key findings, or `web_fetch` with a `selector` to extract a specific section._"),{content:[{type:`text`,text:f.join(`
598
+ `)}],structuredContent:{nodes:(n.nodes??[]).map(e=>({id:e.id,name:e.name,type:e.type,...e.sourcePath?{sourcePath:e.sourcePath}:{}})),edges:(n.edges??[]).map(e=>({fromId:e.fromId,toId:e.toId,type:e.type})),totalNodes:n.stats?.nodeCount??n.nodes?.length??0,totalEdges:n.stats?.edgeCount??n.edges?.length??0,query:e.action}}}catch(e){return wo.error(`Graph query failed`,A(e)),{content:[{type:`text`,text:`Graph query failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Oo=E(`tools:infra`);function ko(e){let t=L(`process`);e.registerTool(`process`,{title:t.title,description:`Start, stop, inspect, list, and tail logs for in-memory managed child processes.`,inputSchema:{action:F.enum([`start`,`stop`,`status`,`list`,`logs`]).describe(`Process action to perform`),id:F.string().optional().describe(`Managed process ID`),command:F.string().optional().describe(`Executable to start`),args:F.array(F.string()).optional().describe(`Arguments for start actions`),tail:F.number().min(1).max(500).optional().describe(`Log lines to return for logs actions`)},annotations:t.annotations},async({action:e,id:t,command:n,args:r,tail:i})=>{try{switch(e){case`start`:if(!t||!n)throw Error(`id and command are required for start`);return{content:[{type:`text`,text:JSON.stringify(pt(t,n,r??[]))}]};case`stop`:if(!t)throw Error(`id is required for stop`);return{content:[{type:`text`,text:JSON.stringify(ht(t)??null)}]};case`status`:if(!t)throw Error(`id is required for status`);return{content:[{type:`text`,text:JSON.stringify(mt(t)??null)}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(dt())}]};case`logs`:if(!t)throw Error(`id is required for logs`);return{content:[{type:`text`,text:JSON.stringify(ft(t,i))}]}}}catch(e){return Oo.error(`Process action failed`,A(e)),{content:[{type:`text`,text:`Process action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ao(e){let t=L(`watch`);e.registerTool(`watch`,{title:t.title,description:`Watch a directory for file changes (create/modify/delete). Actions: start (begin watching), stop (by ID), list (show active watchers). Events are emitted as structured JSON with path, event type, and timestamp.`,inputSchema:{action:F.enum([`start`,`stop`,`list`]).describe(`Watch action to perform`),path:F.string().optional().describe(`Directory path to watch for start actions`),id:F.string().optional().describe(`Watcher ID for stop actions`)},annotations:t.annotations},async({action:e,path:t,id:n})=>{try{switch(e){case`start`:if(!t)throw Error(`path is required for start`);return{content:[{type:`text`,text:JSON.stringify(Qt({path:t}))}]};case`stop`:if(!n)throw Error(`id is required for stop`);return{content:[{type:`text`,text:JSON.stringify({stopped:$t(n)})}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(Zt())}]}}}catch(e){return Oo.error(`Watch action failed`,A(e)),{content:[{type:`text`,text:`Watch action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function jo(e){let t=L(`health`);e.registerTool(`health`,{title:t.title,description:`Run project health checks — verifies package.json, tsconfig, scripts, lockfile, README, LICENSE, .gitignore.`,outputSchema:oi,inputSchema:{path:F.string().optional().describe(`Root directory to check (defaults to cwd)`)},annotations:t.annotations},async({path:e})=>{try{let t=Qe(e),n={ok:t.checks.every(e=>e.status!==`fail`),checks:t.checks.map(e=>({name:e.name,ok:e.status===`pass`,message:e.message}))};return{content:[{type:`text`,text:JSON.stringify(t)}],structuredContent:n}}catch(e){return Oo.error(`Health check failed`,A(e)),{content:[{type:`text`,text:`Health check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Mo(e){let t=L(`web_fetch`);e.registerTool(`web_fetch`,{title:t.title,description:`PREFERRED web fetcher — fetch one or many URLs and convert to LLM-optimized markdown. Pass one URL or multiple for parallel fetching. Supports CSS selectors, 4 output modes (markdown/raw/links/outline), smart paragraph-boundary truncation. Strips scripts/styles/nav automatically.`,inputSchema:{urls:F.array(F.string().url()).min(1).max(10).describe('URLs to fetch (1–10). Single URL: `["https://..."]`. Multiple fetched in parallel.'),mode:F.enum([`markdown`,`raw`,`links`,`outline`]).default(`markdown`).describe(`Output mode: markdown (clean content), raw (HTML), links (extracted URLs), outline (heading hierarchy)`),selector:F.string().optional().describe(`CSS selector to extract a specific element instead of auto-detecting main content`),max_length:F.number().min(500).max(1e5).default(15e3).describe(`Max characters in output — truncates at paragraph boundaries`),include_metadata:F.boolean().default(!0).describe(`Include page title, description, and URL as a header`),include_links:F.boolean().default(!1).describe(`Append extracted links list at the end`),include_images:F.boolean().default(!1).describe(`Include image alt texts inline`),timeout:F.number().min(1e3).max(12e4).default(3e4).describe(`Request timeout in milliseconds`)},annotations:t.annotations},async({urls:e,mode:t,selector:n,max_length:r,include_metadata:i,include_links:a,include_images:o,timeout:s})=>{let c=e,l=async(e,c)=>{let l=await en({url:e,mode:t,selector:n,maxLength:r,includeMetadata:i,includeLinks:a,includeImages:o,timeout:s}),u=[c?`## ${c} ${l.title||`Web Page`}\n> Source: ${e}`:`## ${l.title||`Web Page`}`,``,l.content];return l.truncated&&u.push(``,`_Original length: ${l.originalLength.toLocaleString()} chars_`),u.join(`
599
+ `)};if(c.length===1)try{return{content:[{type:`text`,text:await l(c[0])+"\n\n---\n_Next: Use `remember` to save key findings, or `web_fetch` with a `selector` to extract a specific section._"}]}}catch(e){let t=e instanceof Error?e.message:String(e);return/HTTP [45]\d{2}/.test(t)?Oo.warn(`Web fetch failed`,{error:t}):Oo.error(`Web fetch failed`,A(e)),{content:[{type:`text`,text:`Web fetch failed: ${t}`}],isError:!0}}let u=c.length,d=await Promise.allSettled(c.map((e,t)=>l(e,`[${t+1}/${u}]`))),f=[],p=0;for(let e=0;e<d.length;e++){let t=d[e];if(t.status===`fulfilled`)f.push(t.value);else{p++;let n=t.reason instanceof Error?t.reason.message:String(t.reason);/HTTP [45]\d{2}/.test(n)?Oo.warn(`Web fetch failed`,{url:c[e],error:n}):Oo.error(`Web fetch failed`,{url:c[e],...A(t.reason)}),f.push(`## ❌ Failed: ${c[e]}\n\n${n}`)}}let m=`_Fetched ${d.length-p}/${d.length} URLs successfully._`;return f.push(``,`---`,m,"_Next: Use `remember` to save key findings, or `web_fetch` with a `selector` to extract a specific section._"),{content:[{type:`text`,text:f.join(`
600
600
 
601
- `)}],...p===d.length?{isError:!0}:{}}})}function No(e,t){let n=I(`guide`);e.registerTool(`guide`,{title:n.title,description:`Tool discovery — given a goal description, recommends which AI Kit tools to use and in what order. Matches against 10 predefined workflows: onboard, audit, bugfix, implement, refactor, search, context, memory, validate, analyze.`,inputSchema:{goal:F.string().describe(`What you want to accomplish (e.g., "audit this monorepo", "fix a failing test")`),max_recommendations:F.number().min(1).max(10).default(5).describe(`Maximum number of tool recommendations`)},annotations:n.annotations},async({goal:e,max_recommendations:n})=>{try{let r=Ze(e,n,t),i=[`## Recommended Workflow: **${r.workflow}**`,r.description,``,`### Tools`,...r.tools.map(e=>{let t=e.suggestedArgs?` — \`${JSON.stringify(e.suggestedArgs)}\``:``;return`${e.order}. **${e.tool}** — ${e.reason}${t}`})];return r.alternativeWorkflows.length>0&&i.push(``,`_Alternative workflows: ${r.alternativeWorkflows.join(`, `)}_`),i.push(``,`---`,"_Next: Run the first recommended tool, or use `guide` again with a more specific goal._"),{content:[{type:`text`,text:i.join(`
602
- `)}]}}catch(e){return G.error(`Guide failed`,A(e)),{content:[{type:`text`,text:`Guide failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Po(e,t,n){if(!(!e||typeof e!=`string`)&&!(e.includes(`..`)||e.startsWith(`/`)||e.startsWith(`[`)))return{type:`resource_link`,uri:`aikit://curated/${e}`,name:t??e,mimeType:`text/markdown`,...n?{description:n}:{}}}function Fo(e){let t=new Set,n=[];for(let r of e){let e=Po(r.path,r.title,r.category?`[${r.category}]`:void 0);e&&!t.has(e.uri)&&(t.add(e.uri),n.push(e))}return n}function Io(e){if(e.startsWith(`.ai/curated/`))return e.slice(12)}const Lo=E(`tools`);function Ro(e,t){let n=I(`list`);e.registerTool(`list`,{title:n.title,description:`List curated knowledge entries stored via remember. Returns path, title, category, tags, and content preview for each entry. Filter by category (decisions, patterns, conventions) or tag.`,outputSchema:oi,inputSchema:{category:F.string().optional().describe(`Filter by category (e.g., "decisions", "patterns")`),tag:F.string().optional().describe(`Filter by tag`)},annotations:n.annotations},async({category:e,tag:n})=>{try{let r=await t.list({category:e,tag:n}),i={entries:r.map(e=>({path:e.path,title:e.title,category:e.category,tags:e.tags,version:e.version,preview:e.contentPreview?.slice(0,120)??``})),totalCount:r.length};if(r.length===0)return{content:[{type:`text`,text:`No curated knowledge entries found.`+(e?` (category: ${e})`:``)+(n?` (tag: ${n})`:``)}],structuredContent:i};let a=r.map(e=>{let t=e.tags.length>0?` [${e.tags.join(`, `)}]`:``;return`- **${e.title}** (v${e.version})${t}\n \`${e.path}\` — ${e.contentPreview.slice(0,80)}…`}).join(`
603
- `),o=Fo(r);return{content:[{type:`text`,text:`## Curated Knowledge (${r.length} entries)\n\n${a}\n\n---\n_Next: Use \`read\` to view full content of any entry, or \`remember\` to store new knowledge._`},...o],structuredContent:i}}catch(e){return Lo.error(`List failed`,A(e)),{content:[{type:`text`,text:`List failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const zo=E(`tools`);function Bo(e,t){let n=I(`lookup`);e.registerTool(`lookup`,{title:n.title,description:`Get all indexed content for a known file path, sorted by position. Use when you know the exact file. For discovery across files use search.`,inputSchema:{path:F.string().describe(`Relative file path to look up (e.g., "src/index.ts")`)},annotations:n.annotations},async({path:e})=>{try{let n=await t.getBySourcePath(e);if(n.length===0)return{content:[{type:`text`,text:`No indexed content found for: ${e}`}]};n.sort((e,t)=>e.chunkIndex-t.chunkIndex);let r=`## ${e}\n**Chunks**: ${n.length} | **Type**: ${n[0].contentType}\n`,i=n.map(e=>{let t=e.startLine?` (lines ${e.startLine}-${e.endLine})`:``;return`### Chunk ${e.chunkIndex+1}/${e.totalChunks}${t}\n${e.content}`}).join(`
601
+ `)}],...p===d.length?{isError:!0}:{}}})}function No(e,t){let n=L(`guide`);e.registerTool(`guide`,{title:n.title,description:`Tool discovery — given a goal description, recommends which AI Kit tools to use and in what order. Matches against 10 predefined workflows: onboard, audit, bugfix, implement, refactor, search, context, memory, validate, analyze.`,inputSchema:{goal:F.string().describe(`What you want to accomplish (e.g., "audit this monorepo", "fix a failing test")`),max_recommendations:F.number().min(1).max(10).default(5).describe(`Maximum number of tool recommendations`)},annotations:n.annotations},async({goal:e,max_recommendations:n})=>{try{let r=Ze(e,n,t),i=[`## Recommended Workflow: **${r.workflow}**`,r.description,``,`### Tools`,...r.tools.map(e=>{let t=e.suggestedArgs?` — \`${JSON.stringify(e.suggestedArgs)}\``:``;return`${e.order}. **${e.tool}** — ${e.reason}${t}`})];return r.alternativeWorkflows.length>0&&i.push(``,`_Alternative workflows: ${r.alternativeWorkflows.join(`, `)}_`),i.push(``,`---`,"_Next: Run the first recommended tool, or use `guide` again with a more specific goal._"),{content:[{type:`text`,text:i.join(`
602
+ `)}]}}catch(e){return Oo.error(`Guide failed`,A(e)),{content:[{type:`text`,text:`Guide failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Po(e,t,n){if(!(!e||typeof e!=`string`)&&!(e.includes(`..`)||e.startsWith(`/`)||e.startsWith(`[`)))return{type:`resource_link`,uri:`aikit://curated/${e}`,name:t??e,mimeType:`text/markdown`,...n?{description:n}:{}}}function Fo(e){let t=new Set,n=[];for(let r of e){let e=Po(r.path,r.title,r.category?`[${r.category}]`:void 0);e&&!t.has(e.uri)&&(t.add(e.uri),n.push(e))}return n}function Io(e){if(e.startsWith(`.ai/curated/`))return e.slice(12)}const Lo=E(`tools`);function Ro(e,t){let n=L(`list`);e.registerTool(`list`,{title:n.title,description:`List curated knowledge entries stored via remember. Returns path, title, category, tags, and content preview for each entry. Filter by category (decisions, patterns, conventions) or tag.`,outputSchema:ai,inputSchema:{category:F.string().optional().describe(`Filter by category (e.g., "decisions", "patterns")`),tag:F.string().optional().describe(`Filter by tag`)},annotations:n.annotations},async({category:e,tag:n})=>{try{let r=await t.list({category:e,tag:n}),i={entries:r.map(e=>({path:e.path,title:e.title,category:e.category,tags:e.tags,version:e.version,preview:e.contentPreview?.slice(0,120)??``})),totalCount:r.length};if(r.length===0)return{content:[{type:`text`,text:`No curated knowledge entries found.`+(e?` (category: ${e})`:``)+(n?` (tag: ${n})`:``)}],structuredContent:i};let a=r.map(e=>{let t=e.tags.length>0?` [${e.tags.join(`, `)}]`:``;return`- **${e.title}** (v${e.version})${t}\n \`${e.path}\` — ${e.contentPreview.slice(0,80)}…`}).join(`
603
+ `),o=Fo(r);return{content:[{type:`text`,text:`## Curated Knowledge (${r.length} entries)\n\n${a}\n\n---\n_Next: Use \`read\` to view full content of any entry, or \`remember\` to store new knowledge._`},...o],structuredContent:i}}catch(e){return Lo.error(`List failed`,A(e)),{content:[{type:`text`,text:`List failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const zo=E(`tools`);function Bo(e,t){let n=L(`lookup`);e.registerTool(`lookup`,{title:n.title,description:`Get all indexed content for a known file path, sorted by position. Use when you know the exact file. For discovery across files use search.`,inputSchema:{path:F.string().describe(`Relative file path to look up (e.g., "src/index.ts")`)},annotations:n.annotations},async({path:e})=>{try{let n=await t.getBySourcePath(e);if(n.length===0)return{content:[{type:`text`,text:`No indexed content found for: ${e}`}]};n.sort((e,t)=>e.chunkIndex-t.chunkIndex);let r=`## ${e}\n**Chunks**: ${n.length} | **Type**: ${n[0].contentType}\n`,i=n.map(e=>{let t=e.startLine?` (lines ${e.startLine}-${e.endLine})`:``;return`### Chunk ${e.chunkIndex+1}/${e.totalChunks}${t}\n${e.content}`}).join(`
604
604
 
605
- `),a=Io(e),o=a?Po(a):void 0;return{content:[{type:`text`,text:`${r}\n${i}\n\n---\n_Next: Use \`search\` to find related content, or \`analyze_dependencies\` to see what this file imports._`},...o?[o]:[]]}}catch(e){return zo.error(`Lookup failed`,A(e)),{content:[{type:`text`,text:`Lookup failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Vo=E(`tools:manipulation`);function Ho(e){let t=I(`git_context`);e.registerTool(`git_context`,{title:t.title,description:`Summarize the current Git branch, working tree state, recent commits, and optional diff statistics for the repository.`,inputSchema:{cwd:F.string().optional().describe(`Repository root or working directory`),commit_count:F.number().min(1).max(50).default(5).optional().describe(`How many recent commits to include`),include_diff:F.boolean().default(!1).optional().describe(`Include diff stat for working tree changes`)},annotations:t.annotations,outputSchema:Ai},async({cwd:e,commit_count:t,include_diff:n})=>{try{let r=await Je({cwd:e,commitCount:t,includeDiff:n});return{content:[{type:`text`,text:qo(r)}],structuredContent:{gitRoot:r.gitRoot,branch:r.branch,commitCount:r.recentCommits.length,hasUncommitted:r.status.staged.length>0||r.status.modified.length>0||r.status.untracked.length>0,recentCommits:r.recentCommits.map(e=>({hash:e.hash,message:e.message,author:e.author,date:e.date}))}}}catch(e){return Vo.error(`Git context failed`,A(e)),{content:[{type:`text`,text:`Git context failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Uo(e){let t=I(`diff_parse`);e.registerTool(`diff_parse`,{title:t.title,description:`Parse raw unified diff text into file-level and hunk-level structural changes.`,inputSchema:{diff:F.string().max(1e6).describe(`Raw unified diff text`)},annotations:t.annotations},async({diff:e})=>{try{return{content:[{type:`text`,text:Jo(Fe({diff:e.replace(/\\n/g,`
606
- `).replace(/\\t/g,` `)}))}]}}catch(e){return Vo.error(`Diff parse failed`,A(e)),{content:[{type:`text`,text:`Diff parse failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wo(e){let t=I(`rename`);e.registerTool(`rename`,{title:t.title,description:`Rename a symbol across files using whole-word regex matching for exports, imports, and general usage references.`,inputSchema:{old_name:F.string().describe(`Existing symbol name to replace`),new_name:F.string().describe(`New symbol name to use`),root_path:F.string().describe(`Root directory to search within`),extensions:F.array(F.string()).optional().describe(`Optional file extensions to include, such as .ts,.tsx,.js,.jsx`),dry_run:F.boolean().default(!0).describe(`Preview changes without writing files`)},annotations:t.annotations},async({old_name:e,new_name:t,root_path:n,extensions:r,dry_run:i})=>{try{let a=await Dt({oldName:e,newName:t,rootPath:n,extensions:r,dryRun:i});return{content:[{type:`text`,text:JSON.stringify(a)}]}}catch(e){return Vo.error(`Rename failed`,A(e)),{content:[{type:`text`,text:`Rename failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Go(e){let t=I(`codemod`);e.registerTool(`codemod`,{title:t.title,description:`Apply regex-based codemod rules across files and return structured before/after changes for each affected line.`,inputSchema:{root_path:F.string().describe(`Root directory to transform within`),rules:F.array(F.object({description:F.string().describe(`What the codemod rule does`),pattern:F.string().describe(`Regex pattern in string form`),replacement:F.string().describe(`Replacement string with optional capture groups`)})).min(1).describe(`Codemod rules to apply`),dry_run:F.boolean().default(!0).describe(`Preview changes without writing files`)},annotations:t.annotations},async({root_path:e,rules:t,dry_run:n})=>{try{let r=await De({rootPath:e,rules:t,dryRun:n});return{content:[{type:`text`,text:JSON.stringify(r)}]}}catch(e){return Vo.error(`Codemod failed`,A(e)),{content:[{type:`text`,text:`Codemod failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ko(e){let t=I(`data_transform`);e.registerTool(`data_transform`,{title:t.title,description:`Apply small jq-like transforms to JSON input for filtering, projection, grouping, and path extraction.`,inputSchema:{input:F.string().max(5e5).describe(`Input JSON string`),expression:F.string().max(1e4).describe(`Transform expression to apply`)},annotations:t.annotations},async({input:e,expression:t})=>{try{return{content:[{type:`text`,text:je({input:e,expression:t}).outputString}]}}catch(e){return Vo.error(`Data transform failed`,A(e)),{content:[{type:`text`,text:`Data transform failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function qo(e){let t=[`Git root: ${e.gitRoot}`,`Branch: ${e.branch}`,`Staged: ${e.status.staged.length}`,...e.status.staged.map(e=>` - ${e}`),`Modified: ${e.status.modified.length}`,...e.status.modified.map(e=>` - ${e}`),`Untracked: ${e.status.untracked.length}`,...e.status.untracked.map(e=>` - ${e}`),``,`Recent commits:`];if(e.recentCommits.length===0)t.push(` none`);else for(let n of e.recentCommits)t.push(` - ${n.hash} ${n.message}`),t.push(` ${n.author} @ ${n.date}`);return e.diff&&t.push(``,`Diff stat:`,e.diff),t.join(`
605
+ `),a=Io(e),o=a?Po(a):void 0;return{content:[{type:`text`,text:`${r}\n${i}\n\n---\n_Next: Use \`search\` to find related content, or \`analyze_dependencies\` to see what this file imports._`},...o?[o]:[]]}}catch(e){return zo.error(`Lookup failed`,A(e)),{content:[{type:`text`,text:`Lookup failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Vo=E(`tools:manipulation`);function Ho(e){let t=L(`git_context`);e.registerTool(`git_context`,{title:t.title,description:`Summarize the current Git branch, working tree state, recent commits, and optional diff statistics for the repository.`,inputSchema:{cwd:F.string().optional().describe(`Repository root or working directory`),commit_count:F.number().min(1).max(50).default(5).optional().describe(`How many recent commits to include`),include_diff:F.boolean().default(!1).optional().describe(`Include diff stat for working tree changes`)},annotations:t.annotations,outputSchema:ki},async({cwd:e,commit_count:t,include_diff:n})=>{try{let r=await Je({cwd:e,commitCount:t,includeDiff:n});return{content:[{type:`text`,text:qo(r)}],structuredContent:{gitRoot:r.gitRoot,branch:r.branch,commitCount:r.recentCommits.length,hasUncommitted:r.status.staged.length>0||r.status.modified.length>0||r.status.untracked.length>0,recentCommits:r.recentCommits.map(e=>({hash:e.hash,message:e.message,author:e.author,date:e.date}))}}}catch(e){return Vo.error(`Git context failed`,A(e)),{content:[{type:`text`,text:`Git context failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Uo(e){let t=L(`diff_parse`);e.registerTool(`diff_parse`,{title:t.title,description:`Parse raw unified diff text into file-level and hunk-level structural changes.`,inputSchema:{diff:F.string().max(1e6).describe(`Raw unified diff text`)},annotations:t.annotations},async({diff:e})=>{try{return{content:[{type:`text`,text:Jo(Fe({diff:e.replace(/\\n/g,`
606
+ `).replace(/\\t/g,` `)}))}]}}catch(e){return Vo.error(`Diff parse failed`,A(e)),{content:[{type:`text`,text:`Diff parse failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Wo(e){let t=L(`rename`);e.registerTool(`rename`,{title:t.title,description:`Rename a symbol across files using whole-word regex matching for exports, imports, and general usage references.`,inputSchema:{old_name:F.string().describe(`Existing symbol name to replace`),new_name:F.string().describe(`New symbol name to use`),root_path:F.string().describe(`Root directory to search within`),extensions:F.array(F.string()).optional().describe(`Optional file extensions to include, such as .ts,.tsx,.js,.jsx`),dry_run:F.boolean().default(!0).describe(`Preview changes without writing files`)},annotations:t.annotations},async({old_name:e,new_name:t,root_path:n,extensions:r,dry_run:i})=>{try{let a=await Dt({oldName:e,newName:t,rootPath:n,extensions:r,dryRun:i});return{content:[{type:`text`,text:JSON.stringify(a)}]}}catch(e){return Vo.error(`Rename failed`,A(e)),{content:[{type:`text`,text:`Rename failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Go(e){let t=L(`codemod`);e.registerTool(`codemod`,{title:t.title,description:`Apply regex-based codemod rules across files and return structured before/after changes for each affected line.`,inputSchema:{root_path:F.string().describe(`Root directory to transform within`),rules:F.array(F.object({description:F.string().describe(`What the codemod rule does`),pattern:F.string().describe(`Regex pattern in string form`),replacement:F.string().describe(`Replacement string with optional capture groups`)})).min(1).describe(`Codemod rules to apply`),dry_run:F.boolean().default(!0).describe(`Preview changes without writing files`)},annotations:t.annotations},async({root_path:e,rules:t,dry_run:n})=>{try{let r=await De({rootPath:e,rules:t,dryRun:n});return{content:[{type:`text`,text:JSON.stringify(r)}]}}catch(e){return Vo.error(`Codemod failed`,A(e)),{content:[{type:`text`,text:`Codemod failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function Ko(e){let t=L(`data_transform`);e.registerTool(`data_transform`,{title:t.title,description:`Apply small jq-like transforms to JSON input for filtering, projection, grouping, and path extraction.`,inputSchema:{input:F.string().max(5e5).describe(`Input JSON string`),expression:F.string().max(1e4).describe(`Transform expression to apply`)},annotations:t.annotations},async({input:e,expression:t})=>{try{return{content:[{type:`text`,text:je({input:e,expression:t}).outputString}]}}catch(e){return Vo.error(`Data transform failed`,A(e)),{content:[{type:`text`,text:`Data transform failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function qo(e){let t=[`Git root: ${e.gitRoot}`,`Branch: ${e.branch}`,`Staged: ${e.status.staged.length}`,...e.status.staged.map(e=>` - ${e}`),`Modified: ${e.status.modified.length}`,...e.status.modified.map(e=>` - ${e}`),`Untracked: ${e.status.untracked.length}`,...e.status.untracked.map(e=>` - ${e}`),``,`Recent commits:`];if(e.recentCommits.length===0)t.push(` none`);else for(let n of e.recentCommits)t.push(` - ${n.hash} ${n.message}`),t.push(` ${n.author} @ ${n.date}`);return e.diff&&t.push(``,`Diff stat:`,e.diff),t.join(`
607
607
  `)}function Jo(e){if(e.length===0)return`No diff files found.`;let t=[];for(let n of e){let e=n.oldPath?` (from ${n.oldPath})`:``;t.push(`${n.path}${e} [${n.status}] +${n.additions} -${n.deletions} (${n.hunks.length} hunks)`);for(let e of n.hunks){let n=e.header?` ${e.header}`:``;t.push(` @@ -${e.oldStart},${e.oldLines} +${e.newStart},${e.newLines} @@${n}`)}}return t.join(`
608
- `)}const Yo=[`list_tools`,`describe_tool`,`search_tools`];function Xo(e,t,n){let r=[...new Set(n)].sort();e.registerTool(`list_tools`,{title:I(`list_tools`).title,description:`List the available AI Kit tools with names, titles, and categories. Use this before describe_tool when you need to discover the active toolset.`,inputSchema:{category:F.string().optional().describe(`Optional category filter`)},annotations:I(`list_tools`).annotations},async({category:e})=>{let t=e?.trim().toLowerCase();return{content:[{type:`text`,text:r.filter(e=>t?(rr[e]?.category??[]).some(e=>e.toLowerCase()===t):!0).map(e=>{let t=rr[e]??{title:e,category:[]},n=t.category?.join(`, `)||`uncategorized`;return`${e}: ${t.title} [${n}]`}).join(`
609
- `)||`No tools matched the filter.`}]}}),e.registerTool(`describe_tool`,{title:I(`describe_tool`).title,description:`Describe a specific active AI Kit tool using centralized metadata. Use this after list_tools when you need a tool title, categories, and annotations before calling it directly.`,inputSchema:{tool_name:F.string().describe(`The active tool name to describe`)},annotations:I(`describe_tool`).annotations},async({tool_name:e})=>{if(!r.includes(e))return{content:[{type:`text`,text:`Unknown or inactive tool: ${e}`}],isError:!0};let t=rr[e],n={name:e,title:t?.title??e,categories:t?.category??[],annotations:t?.annotations??{},note:`Call the tool directly after discovery. AI Kit does not expose an execute_tool proxy.`};return{content:[{type:`text`,text:JSON.stringify(n,null,2)}]}}),e.registerTool(`search_tools`,{title:I(`search_tools`).title,description:`Search active AI Kit tools by keyword across tool names, titles, and categories.`,inputSchema:{query:F.string().min(1).describe(`Keyword query to match against active tools`)},annotations:I(`search_tools`).annotations},async({query:e})=>{let t=e.trim().toLowerCase(),n=r.filter(e=>{let n=rr[e];return e.toLowerCase().includes(t)||n?.title.toLowerCase().includes(t)||(n?.category??[]).some(e=>e.toLowerCase().includes(t))}).map(e=>{let t=rr[e]??{title:e,category:[]},n=t.category?.join(`, `)||`uncategorized`;return`${e}: ${t.title} [${n}]`});return{content:[{type:`text`,text:n.length>0?n.join(`
610
- `):`No tools found matching "${e}"`}]}})}const Zo=E(`tools`);let Qo=!1;async function $o(e,t,n){for(let r of n.steps)if(!(r.status!==`success`||!r.output))try{let i=j(`sha256`).update(n.path).digest(`hex`).slice(0,12),a=`produced/onboard/${r.name}/${i}.md`,o=j(`sha256`).update(r.output).digest(`hex`).slice(0,16),s=new Date().toISOString(),c=r.output.length>2e3?r.output.split(/(?=^## )/m).filter(e=>e.trim().length>0):[r.output],l=c.map((e,t)=>({id:j(`sha256`).update(`${a}::${t}`).digest(`hex`).slice(0,16),content:e.trim(),sourcePath:a,contentType:`produced-knowledge`,chunkIndex:t,totalChunks:c.length,startLine:0,endLine:0,fileHash:o,indexedAt:s,origin:`produced`,tags:[`onboard`,r.name],category:`analysis`,version:1})),u=await t.embedBatch(l.map(e=>e.content));await e.upsert(l,u)}catch(e){Zo.warn(`Auto-persist onboard step failed`,{stepName:r.name,...A(e)})}}async function es(e,t,n){if(n.autoRemember?.length)for(let r of n.autoRemember)try{let n=j(`sha256`).update(`onboard-remember::${r.title}`).digest(`hex`).slice(0,16),i=new Date().toISOString(),a={id:n,content:`# ${r.title}\n\n${r.content}`,sourcePath:`curated/onboard/${r.category}/${n}.md`,contentType:`curated`,chunkIndex:0,totalChunks:1,startLine:0,endLine:0,fileHash:j(`sha256`).update(r.content).digest(`hex`).slice(0,16),indexedAt:i,origin:`curated`,tags:r.tags,category:r.category,version:1},[o]=await t.embedBatch([a.content]);await e.upsert([a],[o])}catch(e){Zo.warn(`Auto-persist remember entry failed`,{title:r.title,...A(e)})}}function ts(e,t,n,r,i){let a=I(`onboard`);e.registerTool(`onboard`,{title:a.title,description:`First-time codebase onboarding: runs all analysis tools (structure, dependencies, entry-points, symbols, patterns, diagram) in one command. Results are auto-persisted to KB. Use mode=generate to also write structured output to .ai/context/ directory.`,inputSchema:{path:F.string().describe(`Root path of the codebase to onboard`),mode:F.enum([`memory`,`generate`]).default(`generate`).describe(`Output mode: generate (default) = persist to AI Kit + write .ai/context/ files; memory = AI Kit vector store only`),out_dir:F.string().optional().describe(`Custom output directory for generate mode (default: <path>/.ai/context)`)},annotations:a.annotations},async({path:e,mode:a,out_dir:o},s)=>{try{if(Qo)return{content:[{type:`text`,text:`Onboard is already running. Please wait for it to complete before starting another.`}]};Qo=!0,Zo.info(`Starting onboard`,{path:e,mode:a});let c=await lt({path:e,mode:a,outDir:o??r?.onboardDir}),l=Yi(s).createTask(`Onboard`,c.steps.length);for(let e=0;e<c.steps.length;e++){let t=c.steps[e];l.progress(e,`${t.name}: ${t.status}`)}l.complete(`Onboard complete: ${c.steps.filter(e=>e.status===`success`).length}/${c.steps.length} steps succeeded`),$o(t,n,c),c.autoRemember?.length&&es(t,n,c).catch(e=>{Zo.warn(`Auto-persist autoRemember failed`,A(e))}),i&&(i.onboardComplete=!0,i.onboardTimestamp=new Date().toISOString());let u=[`## Onboard Complete`,``,`**Path:** \`${c.path}\``,`**Mode:** ${c.mode}`,`**Duration:** ${c.totalDurationMs}ms`,``];c.outDir&&(u.push(`**Output directory:** \`${c.outDir}\``),u.push(``)),u.push(`### Analysis Results`,``);let d=[],f=[];for(let e of c.steps)e.status===`success`?d.push(`- ✓ **${e.name}** (${e.durationMs}ms) — ${e.output.length} chars`):f.push(`- ✗ **${e.name}** — ${e.error}`);u.push(...d),f.length>0&&u.push(``,`### Failed`,``,...f),u.push(``,`---`,``);for(let e of c.steps)e.status===`success`&&u.push(`### ${e.name}`,``,e.output,``,`---`,``);return u.push(`_All results auto-saved to KB.`,c.mode===`generate`?` Files written to \`${c.outDir}\`.`:``," Next: Use `search` to query the knowledge, or `remember` to add custom insights._"),{content:[{type:`text`,text:u.join(`
611
- `)}]}}catch(e){return Zo.error(`Onboard failed`,A(e)),{content:[{type:`text`,text:`Onboard failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}finally{Qo=!1}})}const ns=E(`tools:persistence`);function rs(e){let t=I(`workset`);e.registerTool(`workset`,{title:t.title,description:`Manage named file sets (worksets). Save, load, list, add/remove files. Worksets persist across sessions in .aikit-state/worksets.json.`,inputSchema:{action:F.enum([`save`,`get`,`list`,`delete`,`add`,`remove`]).describe(`Operation to perform`),name:F.string().optional().describe(`Workset name (required for all except list)`),files:F.array(F.string()).optional().describe(`File paths (required for save, add, remove)`),description:F.string().optional().describe(`Description (for save)`)},annotations:t.annotations},async({action:e,name:t,files:n,description:r})=>{try{switch(e){case`save`:{if(!t||!n)throw Error(`name and files required for save`);let e=Nt(t,n,{description:r});return{content:[{type:`text`,text:`Saved workset "${e.name}" with ${e.files.length} files.`}]}}case`get`:{if(!t)throw Error(`name required for get`);let e=qe(t);return e?{content:[{type:`text`,text:JSON.stringify(e)}]}:{content:[{type:`text`,text:`Workset "${t}" not found.`}]}}case`list`:{let e=st();return e.length===0?{content:[{type:`text`,text:`No worksets.`}]}:{content:[{type:`text`,text:e.map(e=>`- **${e.name}** (${e.files.length} files) — ${e.description??`no description`}`).join(`
612
- `)}]}}case`delete`:if(!t)throw Error(`name required for delete`);return{content:[{type:`text`,text:Pe(t)?`Deleted workset "${t}".`:`Workset "${t}" not found.`}]};case`add`:{if(!t||!n)throw Error(`name and files required for add`);let e=_e(t,n);return{content:[{type:`text`,text:`Added to workset "${e.name}": now ${e.files.length} files.`}]}}case`remove`:{if(!t||!n)throw Error(`name and files required for remove`);let e=Et(t,n);return e?{content:[{type:`text`,text:`Removed from workset "${e.name}": now ${e.files.length} files.`}]}:{content:[{type:`text`,text:`Workset "${t}" not found.`}]}}}}catch(e){return ns.error(`Workset operation failed`,A(e)),{content:[{type:`text`,text:`Workset operation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function is(e){let t=I(`stash`);e.registerTool(`stash`,{title:t.title,description:`Persist and retrieve named values in .aikit-state/stash.json for intermediate results between tool calls.`,inputSchema:{action:F.enum([`set`,`get`,`list`,`delete`,`clear`]).describe(`Operation to perform on the stash`),key:F.string().optional().describe(`Entry key for set/get/delete operations`),value:F.string().optional().describe(`String or JSON value for set operations`)},annotations:t.annotations},async({action:e,key:t,value:n})=>{try{switch(e){case`set`:{if(!t)throw Error(`key required for set`);let e=Ut(t,ls(n??``));return{content:[{type:`text`,text:`Stored stash entry "${e.key}" (${e.type}) at ${e.storedAt}.`}]}}case`get`:{if(!t)throw Error(`key required for get`);let e=Vt(t);return{content:[{type:`text`,text:e?JSON.stringify(e):`Stash entry "${t}" not found.`}]}}case`list`:{let e=Ht();return{content:[{type:`text`,text:e.length===0?`Stash is empty.`:e.map(e=>`- ${e.key} (${e.type}) — ${e.storedAt}`).join(`
613
- `)}]}}case`delete`:if(!t)throw Error(`key required for delete`);return{content:[{type:`text`,text:Bt(t)?`Deleted stash entry "${t}".`:`Stash entry "${t}" not found.`}]};case`clear`:{let e=zt();return{content:[{type:`text`,text:`Cleared ${e} stash entr${e===1?`y`:`ies`}.`}]}}}}catch(e){return ns.error(`Stash operation failed`,A(e)),{content:[{type:`text`,text:`Stash operation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function as(e){let t=I(`checkpoint`);e.registerTool(`checkpoint`,{title:t.title,description:`Save and restore lightweight session checkpoints in .aikit-state/checkpoints for cross-session continuity.`,inputSchema:{action:F.enum([`save`,`load`,`list`,`latest`]).describe(`Checkpoint action to perform`),label:F.string().optional().describe(`Checkpoint label for save, or checkpoint id for load`),data:F.string().max(5e5).optional().describe(`JSON object string for save actions`),notes:F.string().max(1e4).optional().describe(`Optional notes for save actions`)},annotations:t.annotations},async({action:e,label:t,data:n,notes:r})=>{try{switch(e){case`save`:if(!t)throw Error(`label required for save`);return{content:[{type:`text`,text:cs(Ee(t,us(n),{notes:r}))}]};case`load`:{if(!t)throw Error(`label required for load`);let e=Te(t);return{content:[{type:`text`,text:e?cs(e):`Checkpoint "${t}" not found.`}]}}case`list`:{let e=we();return{content:[{type:`text`,text:e.length===0?`No checkpoints saved.`:e.map(e=>`- ${e.id} — ${e.label} (${e.createdAt})`).join(`
614
- `)}]}}case`latest`:{let e=Ce();return{content:[{type:`text`,text:e?cs(e):`No checkpoints saved.`}]}}}}catch(e){return ns.error(`Checkpoint failed`,A(e)),{content:[{type:`text`,text:`Checkpoint failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function os(e){let t=I(`lane`);e.registerTool(`lane`,{title:t.title,description:`Manage verified lanes — isolated file copies for parallel exploration. Create a lane, make changes, diff, merge back, or discard.`,inputSchema:{action:F.enum([`create`,`list`,`status`,`diff`,`merge`,`discard`]).describe(`Lane action to perform`),name:F.string().optional().describe(`Lane name (required for create/status/diff/merge/discard)`),files:F.array(F.string()).optional().describe(`File paths to copy into the lane (required for create)`)},annotations:t.annotations},async({action:e,name:t,files:n})=>{try{switch(e){case`create`:{if(!t)throw Error(`name is required for create`);if(!n||n.length===0)throw Error(`files are required for create`);let e=et(t,n);return{content:[{type:`text`,text:JSON.stringify(e)}]}}case`list`:return{content:[{type:`text`,text:JSON.stringify(rt())}]};case`status`:if(!t)throw Error(`name is required for status`);return{content:[{type:`text`,text:JSON.stringify(at(t))}]};case`diff`:if(!t)throw Error(`name is required for diff`);return{content:[{type:`text`,text:JSON.stringify(tt(t))}]};case`merge`:if(!t)throw Error(`name is required for merge`);return{content:[{type:`text`,text:JSON.stringify(it(t))}]};case`discard`:if(!t)throw Error(`name is required for discard`);return{content:[{type:`text`,text:JSON.stringify({discarded:nt(t)})}]}}}catch(e){return ns.error(`Lane action failed`,A(e)),{content:[{type:`text`,text:`Lane action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ss(e){let t=I(`queue`);e.registerTool(`queue`,{title:t.title,description:`Manage task queues for sequential agent operations. Push items, take next, mark done/failed, list queues.`,inputSchema:{action:F.enum([`create`,`push`,`next`,`done`,`fail`,`get`,`list`,`clear`,`delete`]).describe(`Queue action`),name:F.string().optional().describe(`Queue name (required for most actions)`),title:F.string().optional().describe(`Item title (required for push)`),id:F.string().optional().describe(`Item ID (required for done/fail)`),data:F.unknown().optional().describe(`Arbitrary data to attach to a queue item`),error:F.string().optional().describe(`Error message (required for fail)`)},annotations:t.annotations},async({action:e,name:t,title:n,id:r,data:i,error:a})=>{try{switch(e){case`create`:if(!t)throw Error(`name is required for create`);return{content:[{type:`text`,text:JSON.stringify(_t(t))}]};case`push`:if(!t)throw Error(`name is required for push`);if(!n)throw Error(`title is required for push`);return{content:[{type:`text`,text:JSON.stringify(wt(t,n,i))}]};case`next`:{if(!t)throw Error(`name is required for next`);let e=Ct(t);return{content:[{type:`text`,text:JSON.stringify(e)}]}}case`done`:if(!t)throw Error(`name is required for done`);if(!r)throw Error(`id is required for done`);return{content:[{type:`text`,text:JSON.stringify(yt(t,r))}]};case`fail`:if(!t)throw Error(`name is required for fail`);if(!r)throw Error(`id is required for fail`);if(!a)throw Error(`error is required for fail`);return{content:[{type:`text`,text:JSON.stringify(bt(t,r,a))}]};case`get`:if(!t)throw Error(`name is required for get`);return{content:[{type:`text`,text:JSON.stringify(xt(t))}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(St())}]};case`clear`:if(!t)throw Error(`name is required for clear`);return{content:[{type:`text`,text:JSON.stringify({cleared:gt(t)})}]};case`delete`:if(!t)throw Error(`name is required for delete`);return{content:[{type:`text`,text:JSON.stringify({deleted:vt(t)})}]}}}catch(e){return ns.error(`Queue action failed`,A(e)),{content:[{type:`text`,text:`Queue action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function cs(e){let t=[e.id,`Label: ${e.label}`,`Created: ${e.createdAt}`];if(e.notes&&t.push(`Notes: ${e.notes}`),e.files?.length){t.push(`Files: ${e.files.length}`);for(let n of e.files)t.push(` - ${n}`)}return t.push(``,`Data:`,JSON.stringify(e.data)),t.join(`
608
+ `)}const Yo=[`list_tools`,`describe_tool`,`search_tools`];function Xo(e,t,n){let r=[...new Set(n)].sort();e.registerTool(`list_tools`,{title:L(`list_tools`).title,description:`List the available AI Kit tools with names, titles, and categories. Use this before describe_tool when you need to discover the active toolset.`,inputSchema:{category:F.string().optional().describe(`Optional category filter`)},annotations:L(`list_tools`).annotations},async({category:e})=>{let t=e?.trim().toLowerCase();return{content:[{type:`text`,text:r.filter(e=>t?(I[e]?.category??[]).some(e=>e.toLowerCase()===t):!0).map(e=>{let t=I[e]??{title:e,category:[]},n=t.category?.join(`, `)||`uncategorized`;return`${e}: ${t.title} [${n}]`}).join(`
609
+ `)||`No tools matched the filter.`}]}}),e.registerTool(`describe_tool`,{title:L(`describe_tool`).title,description:`Describe a specific active AI Kit tool using centralized metadata. Use this after list_tools when you need a tool title, categories, and annotations before calling it directly.`,inputSchema:{tool_name:F.string().describe(`The active tool name to describe`)},annotations:L(`describe_tool`).annotations},async({tool_name:e})=>{if(!r.includes(e))return{content:[{type:`text`,text:`Unknown or inactive tool: ${e}`}],isError:!0};let t=I[e],n={name:e,title:t?.title??e,categories:t?.category??[],annotations:t?.annotations??{},note:`Call the tool directly after discovery. AI Kit does not expose an execute_tool proxy.`};return{content:[{type:`text`,text:JSON.stringify(n,null,2)}]}}),e.registerTool(`search_tools`,{title:L(`search_tools`).title,description:`Search active AI Kit tools by keyword across tool names, titles, and categories.`,inputSchema:{query:F.string().min(1).describe(`Keyword query to match against active tools`)},annotations:L(`search_tools`).annotations},async({query:e})=>{let t=e.trim().toLowerCase(),n=r.filter(e=>{let n=I[e];return e.toLowerCase().includes(t)||n?.title.toLowerCase().includes(t)||(n?.category??[]).some(e=>e.toLowerCase().includes(t))}).map(e=>{let t=I[e]??{title:e,category:[]},n=t.category?.join(`, `)||`uncategorized`;return`${e}: ${t.title} [${n}]`});return{content:[{type:`text`,text:n.length>0?n.join(`
610
+ `):`No tools found matching "${e}"`}]}})}const Zo=E(`tools`);let Qo=!1;async function $o(e,t,n){for(let r of n.steps)if(!(r.status!==`success`||!r.output))try{let i=j(`sha256`).update(n.path).digest(`hex`).slice(0,12),a=`produced/onboard/${r.name}/${i}.md`,o=j(`sha256`).update(r.output).digest(`hex`).slice(0,16),s=new Date().toISOString(),c=r.output.length>2e3?r.output.split(/(?=^## )/m).filter(e=>e.trim().length>0):[r.output],l=c.map((e,t)=>({id:j(`sha256`).update(`${a}::${t}`).digest(`hex`).slice(0,16),content:e.trim(),sourcePath:a,contentType:`produced-knowledge`,chunkIndex:t,totalChunks:c.length,startLine:0,endLine:0,fileHash:o,indexedAt:s,origin:`produced`,tags:[`onboard`,r.name],category:`analysis`,version:1})),u=await t.embedBatch(l.map(e=>e.content));await e.upsert(l,u)}catch(e){Zo.warn(`Auto-persist onboard step failed`,{stepName:r.name,...A(e)})}}async function es(e,t,n){if(n.autoRemember?.length)for(let r of n.autoRemember)try{let n=j(`sha256`).update(`onboard-remember::${r.title}`).digest(`hex`).slice(0,16),i=new Date().toISOString(),a={id:n,content:`# ${r.title}\n\n${r.content}`,sourcePath:`curated/onboard/${r.category}/${n}.md`,contentType:`curated`,chunkIndex:0,totalChunks:1,startLine:0,endLine:0,fileHash:j(`sha256`).update(r.content).digest(`hex`).slice(0,16),indexedAt:i,origin:`curated`,tags:r.tags,category:r.category,version:1},[o]=await t.embedBatch([a.content]);await e.upsert([a],[o])}catch(e){Zo.warn(`Auto-persist remember entry failed`,{title:r.title,...A(e)})}}function ts(e,t,n,r,i){let a=L(`onboard`);e.registerTool(`onboard`,{title:a.title,description:`First-time codebase onboarding: runs all analysis tools (structure, dependencies, entry-points, symbols, patterns, diagram) in one command. Results are auto-persisted to KB. Use mode=generate to also write structured output to .ai/context/ directory.`,inputSchema:{path:F.string().describe(`Root path of the codebase to onboard`),mode:F.enum([`memory`,`generate`]).default(`generate`).describe(`Output mode: generate (default) = persist to AI Kit + write .ai/context/ files; memory = AI Kit vector store only`),out_dir:F.string().optional().describe(`Custom output directory for generate mode (default: <path>/.ai/context)`)},annotations:a.annotations},async({path:e,mode:a,out_dir:o},s)=>{try{if(Qo)return{content:[{type:`text`,text:`Onboard is already running. Please wait for it to complete before starting another.`}]};Qo=!0,Zo.info(`Starting onboard`,{path:e,mode:a});let c=await lt({path:e,mode:a,outDir:o??r?.onboardDir}),l=Ji(s).createTask(`Onboard`,c.steps.length);for(let e=0;e<c.steps.length;e++){let t=c.steps[e];l.progress(e,`${t.name}: ${t.status}`)}l.complete(`Onboard complete: ${c.steps.filter(e=>e.status===`success`).length}/${c.steps.length} steps succeeded`),$o(t,n,c),c.autoRemember?.length&&es(t,n,c).catch(e=>{Zo.warn(`Auto-persist autoRemember failed`,A(e))}),i&&(i.onboardComplete=!0,i.onboardTimestamp=new Date().toISOString());let u=[`## Onboard Complete`,``,`**Path:** \`${c.path}\``,`**Mode:** ${c.mode}`,`**Duration:** ${c.totalDurationMs}ms`,``];c.outDir&&(u.push(`**Output directory:** \`${c.outDir}\``),u.push(``)),u.push(`### Analysis Results`,``);let d=[],f=[];for(let e of c.steps)e.status===`success`?d.push(`- ✓ **${e.name}** (${e.durationMs}ms) — ${e.output.length} chars`):f.push(`- ✗ **${e.name}** — ${e.error}`);u.push(...d),f.length>0&&u.push(``,`### Failed`,``,...f),u.push(``,`---`,``);for(let e of c.steps)e.status===`success`&&u.push(`### ${e.name}`,``,e.output,``,`---`,``);return u.push(`_All results auto-saved to KB.`,c.mode===`generate`?` Files written to \`${c.outDir}\`.`:``," Next: Use `search` to query the knowledge, or `remember` to add custom insights._"),{content:[{type:`text`,text:u.join(`
611
+ `)}]}}catch(e){return Zo.error(`Onboard failed`,A(e)),{content:[{type:`text`,text:`Onboard failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}finally{Qo=!1}})}const ns=E(`tools:persistence`);function rs(e){let t=L(`workset`);e.registerTool(`workset`,{title:t.title,description:`Manage named file sets (worksets). Save, load, list, add/remove files. Worksets persist across sessions in .aikit-state/worksets.json.`,inputSchema:{action:F.enum([`save`,`get`,`list`,`delete`,`add`,`remove`]).describe(`Operation to perform`),name:F.string().optional().describe(`Workset name (required for all except list)`),files:F.array(F.string()).optional().describe(`File paths (required for save, add, remove)`),description:F.string().optional().describe(`Description (for save)`)},annotations:t.annotations},async({action:e,name:t,files:n,description:r})=>{try{switch(e){case`save`:{if(!t||!n)throw Error(`name and files required for save`);let e=Nt(t,n,{description:r});return{content:[{type:`text`,text:`Saved workset "${e.name}" with ${e.files.length} files.`}]}}case`get`:{if(!t)throw Error(`name required for get`);let e=qe(t);return e?{content:[{type:`text`,text:JSON.stringify(e)}]}:{content:[{type:`text`,text:`Workset "${t}" not found.`}]}}case`list`:{let e=st();return e.length===0?{content:[{type:`text`,text:`No worksets.`}]}:{content:[{type:`text`,text:e.map(e=>`- **${e.name}** (${e.files.length} files) — ${e.description??`no description`}`).join(`
612
+ `)}]}}case`delete`:if(!t)throw Error(`name required for delete`);return{content:[{type:`text`,text:Pe(t)?`Deleted workset "${t}".`:`Workset "${t}" not found.`}]};case`add`:{if(!t||!n)throw Error(`name and files required for add`);let e=_e(t,n);return{content:[{type:`text`,text:`Added to workset "${e.name}": now ${e.files.length} files.`}]}}case`remove`:{if(!t||!n)throw Error(`name and files required for remove`);let e=Et(t,n);return e?{content:[{type:`text`,text:`Removed from workset "${e.name}": now ${e.files.length} files.`}]}:{content:[{type:`text`,text:`Workset "${t}" not found.`}]}}}}catch(e){return ns.error(`Workset operation failed`,A(e)),{content:[{type:`text`,text:`Workset operation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function is(e){let t=L(`stash`);e.registerTool(`stash`,{title:t.title,description:`Persist and retrieve named values in .aikit-state/stash.json for intermediate results between tool calls.`,inputSchema:{action:F.enum([`set`,`get`,`list`,`delete`,`clear`]).describe(`Operation to perform on the stash`),key:F.string().optional().describe(`Entry key for set/get/delete operations`),value:F.string().optional().describe(`String or JSON value for set operations`)},annotations:t.annotations},async({action:e,key:t,value:n})=>{try{switch(e){case`set`:{if(!t)throw Error(`key required for set`);let e=Ut(t,ls(n??``));return{content:[{type:`text`,text:`Stored stash entry "${e.key}" (${e.type}) at ${e.storedAt}.`}]}}case`get`:{if(!t)throw Error(`key required for get`);let e=Vt(t);return{content:[{type:`text`,text:e?JSON.stringify(e):`Stash entry "${t}" not found.`}]}}case`list`:{let e=Ht();return{content:[{type:`text`,text:e.length===0?`Stash is empty.`:e.map(e=>`- ${e.key} (${e.type}) — ${e.storedAt}`).join(`
613
+ `)}]}}case`delete`:if(!t)throw Error(`key required for delete`);return{content:[{type:`text`,text:Bt(t)?`Deleted stash entry "${t}".`:`Stash entry "${t}" not found.`}]};case`clear`:{let e=zt();return{content:[{type:`text`,text:`Cleared ${e} stash entr${e===1?`y`:`ies`}.`}]}}}}catch(e){return ns.error(`Stash operation failed`,A(e)),{content:[{type:`text`,text:`Stash operation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function as(e){let t=L(`checkpoint`);e.registerTool(`checkpoint`,{title:t.title,description:`Save and restore lightweight session checkpoints in .aikit-state/checkpoints for cross-session continuity.`,inputSchema:{action:F.enum([`save`,`load`,`list`,`latest`]).describe(`Checkpoint action to perform`),label:F.string().optional().describe(`Checkpoint label for save, or checkpoint id for load`),data:F.string().max(5e5).optional().describe(`JSON object string for save actions`),notes:F.string().max(1e4).optional().describe(`Optional notes for save actions`)},annotations:t.annotations},async({action:e,label:t,data:n,notes:r})=>{try{switch(e){case`save`:if(!t)throw Error(`label required for save`);return{content:[{type:`text`,text:cs(Ee(t,us(n),{notes:r}))}]};case`load`:{if(!t)throw Error(`label required for load`);let e=Te(t);return{content:[{type:`text`,text:e?cs(e):`Checkpoint "${t}" not found.`}]}}case`list`:{let e=we();return{content:[{type:`text`,text:e.length===0?`No checkpoints saved.`:e.map(e=>`- ${e.id} — ${e.label} (${e.createdAt})`).join(`
614
+ `)}]}}case`latest`:{let e=Ce();return{content:[{type:`text`,text:e?cs(e):`No checkpoints saved.`}]}}}}catch(e){return ns.error(`Checkpoint failed`,A(e)),{content:[{type:`text`,text:`Checkpoint failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function os(e){let t=L(`lane`);e.registerTool(`lane`,{title:t.title,description:`Manage verified lanes — isolated file copies for parallel exploration. Create a lane, make changes, diff, merge back, or discard.`,inputSchema:{action:F.enum([`create`,`list`,`status`,`diff`,`merge`,`discard`]).describe(`Lane action to perform`),name:F.string().optional().describe(`Lane name (required for create/status/diff/merge/discard)`),files:F.array(F.string()).optional().describe(`File paths to copy into the lane (required for create)`)},annotations:t.annotations},async({action:e,name:t,files:n})=>{try{switch(e){case`create`:{if(!t)throw Error(`name is required for create`);if(!n||n.length===0)throw Error(`files are required for create`);let e=et(t,n);return{content:[{type:`text`,text:JSON.stringify(e)}]}}case`list`:return{content:[{type:`text`,text:JSON.stringify(rt())}]};case`status`:if(!t)throw Error(`name is required for status`);return{content:[{type:`text`,text:JSON.stringify(at(t))}]};case`diff`:if(!t)throw Error(`name is required for diff`);return{content:[{type:`text`,text:JSON.stringify(tt(t))}]};case`merge`:if(!t)throw Error(`name is required for merge`);return{content:[{type:`text`,text:JSON.stringify(it(t))}]};case`discard`:if(!t)throw Error(`name is required for discard`);return{content:[{type:`text`,text:JSON.stringify({discarded:nt(t)})}]}}}catch(e){return ns.error(`Lane action failed`,A(e)),{content:[{type:`text`,text:`Lane action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ss(e){let t=L(`queue`);e.registerTool(`queue`,{title:t.title,description:`Manage task queues for sequential agent operations. Push items, take next, mark done/failed, list queues.`,inputSchema:{action:F.enum([`create`,`push`,`next`,`done`,`fail`,`get`,`list`,`clear`,`delete`]).describe(`Queue action`),name:F.string().optional().describe(`Queue name (required for most actions)`),title:F.string().optional().describe(`Item title (required for push)`),id:F.string().optional().describe(`Item ID (required for done/fail)`),data:F.unknown().optional().describe(`Arbitrary data to attach to a queue item`),error:F.string().optional().describe(`Error message (required for fail)`)},annotations:t.annotations},async({action:e,name:t,title:n,id:r,data:i,error:a})=>{try{switch(e){case`create`:if(!t)throw Error(`name is required for create`);return{content:[{type:`text`,text:JSON.stringify(_t(t))}]};case`push`:if(!t)throw Error(`name is required for push`);if(!n)throw Error(`title is required for push`);return{content:[{type:`text`,text:JSON.stringify(wt(t,n,i))}]};case`next`:{if(!t)throw Error(`name is required for next`);let e=Ct(t);return{content:[{type:`text`,text:JSON.stringify(e)}]}}case`done`:if(!t)throw Error(`name is required for done`);if(!r)throw Error(`id is required for done`);return{content:[{type:`text`,text:JSON.stringify(yt(t,r))}]};case`fail`:if(!t)throw Error(`name is required for fail`);if(!r)throw Error(`id is required for fail`);if(!a)throw Error(`error is required for fail`);return{content:[{type:`text`,text:JSON.stringify(bt(t,r,a))}]};case`get`:if(!t)throw Error(`name is required for get`);return{content:[{type:`text`,text:JSON.stringify(xt(t))}]};case`list`:return{content:[{type:`text`,text:JSON.stringify(St())}]};case`clear`:if(!t)throw Error(`name is required for clear`);return{content:[{type:`text`,text:JSON.stringify({cleared:gt(t)})}]};case`delete`:if(!t)throw Error(`name is required for delete`);return{content:[{type:`text`,text:JSON.stringify({deleted:vt(t)})}]}}}catch(e){return ns.error(`Queue action failed`,A(e)),{content:[{type:`text`,text:`Queue action failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function cs(e){let t=[e.id,`Label: ${e.label}`,`Created: ${e.createdAt}`];if(e.notes&&t.push(`Notes: ${e.notes}`),e.files?.length){t.push(`Files: ${e.files.length}`);for(let n of e.files)t.push(` - ${n}`)}return t.push(``,`Data:`,JSON.stringify(e.data)),t.join(`
615
615
  `)}function ls(e){let t=e.trim();if(!t)return``;try{return JSON.parse(t)}catch{return e}}function us(e){let t=e?.trim();if(!t)return{};let n;try{n=JSON.parse(t)}catch{throw Error(`data must be a valid JSON object string`)}if(!n||typeof n!=`object`||Array.isArray(n))throw Error(`data must be a JSON object string`);return n}const ds=E(`tools`);function fs(e){let t=[`## Rule: ${e.id}`,``,`- **Status**: ${e.enabled?`enabled`:`disabled`}`,`- **Description**: ${e.description??`—`}`,`- **Category**: ${e.category??`—`}`,`- **Push weight**: ${e.pushWeight??`—`}`];return e.patterns?.length&&t.push(`- **Patterns**: ${e.patterns.join(`, `)}`),e.examples?.length&&t.push(`- **Examples**: ${e.examples.join(`, `)}`),t.join(`
616
- `)}function ps(e,t){let n=I(`er_update_policy`);e.registerTool(`er_update_policy`,{title:n.title,description:`Manage ER push classification rules. Supports listing, updating, creating, and deleting rules that determine when knowledge should be pushed to Enterprise RAG.`,inputSchema:{action:F.enum([`list`,`get`,`update`,`create`,`delete`]).describe(`Action to perform on classification rules`),rule_id:F.string().optional().describe(`Rule ID (required for get, update, delete)`),changes:F.object({patterns:F.array(F.string()).optional(),category:F.string().optional(),pushWeight:F.number().min(0).max(1).optional(),description:F.string().optional(),examples:F.array(F.string()).optional(),enabled:F.boolean().optional()}).optional().describe(`Changes to apply (for update action)`),new_rule:F.object({id:F.string().regex(/^[a-z][a-z0-9-]*$/),patterns:F.array(F.string()).min(1),category:F.string(),pushWeight:F.number().min(0).max(1),description:F.string(),examples:F.array(F.string()).default([]),enabled:F.boolean().default(!0)}).optional().describe(`New rule definition (for create action)`)},annotations:n.annotations},async({action:e,rule_id:n,changes:r,new_rule:i})=>{try{if(e===`list`){let e=t.getRules();return{content:[{type:`text`,text:`## Classification Rules\n\n${e.map(e=>`- **${e.id}** (${e.enabled?`enabled`:`disabled`}) — ${e.description}\n Category: ${e.category} | Weight: ${e.pushWeight} | Patterns: ${e.patterns.join(`, `)}`).join(`
616
+ `)}function ps(e,t){let n=L(`er_update_policy`);e.registerTool(`er_update_policy`,{title:n.title,description:`Manage ER push classification rules. Supports listing, updating, creating, and deleting rules that determine when knowledge should be pushed to Enterprise RAG.`,inputSchema:{action:F.enum([`list`,`get`,`update`,`create`,`delete`]).describe(`Action to perform on classification rules`),rule_id:F.string().optional().describe(`Rule ID (required for get, update, delete)`),changes:F.object({patterns:F.array(F.string()).optional(),category:F.string().optional(),pushWeight:F.number().min(0).max(1).optional(),description:F.string().optional(),examples:F.array(F.string()).optional(),enabled:F.boolean().optional()}).optional().describe(`Changes to apply (for update action)`),new_rule:F.object({id:F.string().regex(/^[a-z][a-z0-9-]*$/),patterns:F.array(F.string()).min(1),category:F.string(),pushWeight:F.number().min(0).max(1),description:F.string(),examples:F.array(F.string()).default([]),enabled:F.boolean().default(!0)}).optional().describe(`New rule definition (for create action)`)},annotations:n.annotations},async({action:e,rule_id:n,changes:r,new_rule:i})=>{try{if(e===`list`){let e=t.getRules();return{content:[{type:`text`,text:`## Classification Rules\n\n${e.map(e=>`- **${e.id}** (${e.enabled?`enabled`:`disabled`}) — ${e.description}\n Category: ${e.category} | Weight: ${e.pushWeight} | Patterns: ${e.patterns.join(`, `)}`).join(`
617
617
  `)}\n\n---\n_${e.length} rules total. Use \`action: "update"\` to modify a rule._`}]}}if(e===`get`){if(!n)return{content:[{type:`text`,text:'`rule_id` is required for "get" action.'}],isError:!0};let e=t.getRule(n);return e?{content:[{type:`text`,text:fs(e)}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}if(e===`update`){if(!n||!r)return{content:[{type:`text`,text:'`rule_id` and `changes` are required for "update" action.'}],isError:!0};let e=t.updateRule(n,r);return e?{content:[{type:`text`,text:`${fs(e)}\n\n---\n_Updated. Next: Use \`action: "list"\` to verify._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}}return e===`create`?i?{content:[{type:`text`,text:`${fs(t.addRule(i))}\n\n---\n_Created. Next: Test classification with \`remember\`._`}]}:{content:[{type:`text`,text:'`new_rule` is required for "create" action.'}],isError:!0}:e===`delete`?n?t.deleteRule(n)?{content:[{type:`text`,text:`Deleted rule **${n}**.\n\n---\n_Next: Use \`action: "list"\` to verify._`}]}:{content:[{type:`text`,text:`Rule "${n}" not found.`}],isError:!0}:{content:[{type:`text`,text:'`rule_id` is required for "delete" action.'}],isError:!0}:{content:[{type:`text`,text:`Unknown action: ${e}`}],isError:!0}}catch(e){return ds.error(`Policy update failed`,A(e)),{content:[{type:`text`,text:`Policy update failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ms(){return`
618
618
  /* ── Reset ────────────────────────────────────────────────────────── */
619
619
  *{box-sizing:border-box;margin:0;padding:0}
@@ -1009,7 +1009,7 @@ li{margin:4px 0}
1009
1009
  ::-webkit-scrollbar-track{background:var(--card)}
1010
1010
  ::-webkit-scrollbar-thumb{background:var(--border);border-radius:3px}
1011
1011
  ::-webkit-scrollbar-thumb:hover{background:var(--muted-foreground)}
1012
- `}const K=z;function hs(e){let t=e;return!t?.items||!Array.isArray(t.items)?``:`<div class="timeline">${t.items.map(e=>{let t=e.status??`pending`,n=e.phase?`<div class="timeline-phase">${K(e.phase)}</div>`:``,r=e.description?`<div class="timeline-desc">${K(e.description)}</div>`:``;return`<div class="timeline-item">
1012
+ `}const K=B;function hs(e){let t=e;return!t?.items||!Array.isArray(t.items)?``:`<div class="timeline">${t.items.map(e=>{let t=e.status??`pending`,n=e.phase?`<div class="timeline-phase">${K(e.phase)}</div>`:``,r=e.description?`<div class="timeline-desc">${K(e.description)}</div>`:``;return`<div class="timeline-item">
1013
1013
  <div class="timeline-dot ${t}"></div>
1014
1014
  ${n}
1015
1015
  <div class="timeline-title">${K(e.title)}</div>
@@ -1035,7 +1035,7 @@ li{margin:4px 0}
1035
1035
  <div class="progress-track">
1036
1036
  <div class="progress-fill" style="width:${r.toFixed(1)}%;background:${i}"></div>
1037
1037
  </div>
1038
- </div>`}).join(``)}</div>`}const q=z;function xs(e){return Math.abs(e)>=1e6?`${(e/1e6).toFixed(1).replace(/\.0$/,``)}M`:Math.abs(e)>=1e3?`${(e/1e3).toFixed(1).replace(/\.0$/,``)}k`:Number.isInteger(e)?String(e):e.toFixed(1)}function Ss(e){return`var(--chart-${e%12+1})`}let Cs=0;function ws(e){let t=e.value;if(!t?.data?.length||!t.chartType)return`<pre>${q(JSON.stringify(e.value,null,2))}</pre>`;Cs++;let n=e.title||t.title?`<div class="chart-title">${q(String(e.title||t.title))}</div>`:``,r;switch(t.chartType){case`line`:r=Ms(t,!1);break;case`area`:r=Ms(t,!0);break;case`bar`:r=Ns(t,!1);break;case`horizontal-bar`:r=Ns(t,!0);break;case`pie`:r=Ps(t,!1);break;case`donut`:r=Ps(t,!0);break;case`sparkline`:r=Fs(t);break;case`heatmap`:r=Is(t);break;default:r=`<pre>${q(JSON.stringify(t,null,2))}</pre>`}let i=t.showLegend!==!1&&t.yKeys.length>1&&t.chartType!==`sparkline`?Ts(t.yKeys):``;return`<div class="chart-container">${n}${r}${i}</div>`}function Ts(e){return`<div class="chart-legend">${e.map((e,t)=>`<span class="chart-legend-item"><span class="chart-legend-swatch" style="background:${Ss(t)}"></span>${q(e)}</span>`).join(``)}</div>`}const J={top:20,right:20,bottom:35,left:55};function Es(e,t){return e.map(e=>Number(e[t])||0)}function Ds(e,t){return e.map(e=>String(e[t]??``))}function Os(e){let t=1/0,n=-1/0;for(let r of e)r<t&&(t=r),r>n&&(n=r);return t===n&&(t=t===0?0:t*.9,n=n===0?1:n*1.1),[t,n]}function ks(e){return e>=0?0:e}function As(e,t,n,r=5){let i=[];for(let a=0;a<=r;a++){let o=a/r,s=J.top+n-o*n,c=e+o*(t-e);i.push(`<line class="chart-grid-line" x1="${J.left}" x2="${600-J.right}" y1="${s}" y2="${s}"/>`),i.push(`<text class="chart-axis-label" x="${J.left-8}" y="${s+4}" text-anchor="end">${xs(c)}</text>`)}return i.join(``)}function js(e,t,n,r){let i=Math.floor(t/50),a=Math.max(1,Math.ceil(e.length/i));return e.map((e,t)=>t%a===0?`<text class="chart-axis-label" x="${J.left+t*r+r/2}" y="${J.top+n+18}" text-anchor="middle">${q(e.length>10?`${e.slice(0,9)}…`:e)}</text>`:``).join(``)}function Ms(e,t){let n=(e.height??300)-J.top-J.bottom,r=600-J.left-J.right,i=Ds(e.data,e.xKey),a=e.data.length;if(a===0)return``;let o=r/Math.max(a-1,1),s=1/0,c=-1/0;for(let t of e.yKeys){let[n,r]=Os(Es(e.data,t));n<s&&(s=n),r>c&&(c=r)}s=ks(s);let l=c-s||1,u=e=>J.top+n-(e-s)/l*n,d=e=>J.left+e*o,f=``,p=[];for(let r=0;r<e.yKeys.length;r++){let i=Es(e.data,e.yKeys[r]),o=Ss(r),s=`M${i.map((e,t)=>`${d(t).toFixed(1)},${u(e).toFixed(1)}`).join(`L`)}`;if(t){let e=`ag${Cs}_${r}`;f+=`<linearGradient id="${e}" x1="0" y1="0" x2="0" y2="1">
1038
+ </div>`}).join(``)}</div>`}const q=B;function xs(e){return Math.abs(e)>=1e6?`${(e/1e6).toFixed(1).replace(/\.0$/,``)}M`:Math.abs(e)>=1e3?`${(e/1e3).toFixed(1).replace(/\.0$/,``)}k`:Number.isInteger(e)?String(e):e.toFixed(1)}function Ss(e){return`var(--chart-${e%12+1})`}let Cs=0;function ws(e){let t=e.value;if(!t?.data?.length||!t.chartType)return`<pre>${q(JSON.stringify(e.value,null,2))}</pre>`;Cs++;let n=e.title||t.title?`<div class="chart-title">${q(String(e.title||t.title))}</div>`:``,r;switch(t.chartType){case`line`:r=Ms(t,!1);break;case`area`:r=Ms(t,!0);break;case`bar`:r=Ns(t,!1);break;case`horizontal-bar`:r=Ns(t,!0);break;case`pie`:r=Ps(t,!1);break;case`donut`:r=Ps(t,!0);break;case`sparkline`:r=Fs(t);break;case`heatmap`:r=Is(t);break;default:r=`<pre>${q(JSON.stringify(t,null,2))}</pre>`}let i=t.showLegend!==!1&&t.yKeys.length>1&&t.chartType!==`sparkline`?Ts(t.yKeys):``;return`<div class="chart-container">${n}${r}${i}</div>`}function Ts(e){return`<div class="chart-legend">${e.map((e,t)=>`<span class="chart-legend-item"><span class="chart-legend-swatch" style="background:${Ss(t)}"></span>${q(e)}</span>`).join(``)}</div>`}const J={top:20,right:20,bottom:35,left:55};function Es(e,t){return e.map(e=>Number(e[t])||0)}function Ds(e,t){return e.map(e=>String(e[t]??``))}function Os(e){let t=1/0,n=-1/0;for(let r of e)r<t&&(t=r),r>n&&(n=r);return t===n&&(t=t===0?0:t*.9,n=n===0?1:n*1.1),[t,n]}function ks(e){return e>=0?0:e}function As(e,t,n,r=5){let i=[];for(let a=0;a<=r;a++){let o=a/r,s=J.top+n-o*n,c=e+o*(t-e);i.push(`<line class="chart-grid-line" x1="${J.left}" x2="${600-J.right}" y1="${s}" y2="${s}"/>`),i.push(`<text class="chart-axis-label" x="${J.left-8}" y="${s+4}" text-anchor="end">${xs(c)}</text>`)}return i.join(``)}function js(e,t,n,r){let i=Math.floor(t/50),a=Math.max(1,Math.ceil(e.length/i));return e.map((e,t)=>t%a===0?`<text class="chart-axis-label" x="${J.left+t*r+r/2}" y="${J.top+n+18}" text-anchor="middle">${q(e.length>10?`${e.slice(0,9)}…`:e)}</text>`:``).join(``)}function Ms(e,t){let n=(e.height??300)-J.top-J.bottom,r=600-J.left-J.right,i=Ds(e.data,e.xKey),a=e.data.length;if(a===0)return``;let o=r/Math.max(a-1,1),s=1/0,c=-1/0;for(let t of e.yKeys){let[n,r]=Os(Es(e.data,t));n<s&&(s=n),r>c&&(c=r)}s=ks(s);let l=c-s||1,u=e=>J.top+n-(e-s)/l*n,d=e=>J.left+e*o,f=``,p=[];for(let r=0;r<e.yKeys.length;r++){let i=Es(e.data,e.yKeys[r]),o=Ss(r),s=`M${i.map((e,t)=>`${d(t).toFixed(1)},${u(e).toFixed(1)}`).join(`L`)}`;if(t){let e=`ag${Cs}_${r}`;f+=`<linearGradient id="${e}" x1="0" y1="0" x2="0" y2="1">
1039
1039
  <stop offset="0%" stop-color="${o}" stop-opacity="0.3"/>
1040
1040
  <stop offset="100%" stop-color="${o}" stop-opacity="0"/>
1041
1041
  </linearGradient>`;let t=`${s}L${d(a-1).toFixed(1)},${(J.top+n).toFixed(1)}L${d(0).toFixed(1)},${(J.top+n).toFixed(1)}Z`;p.push(`<path d="${t}" fill="url(#${e})" stroke="none"/>`)}p.push(`<path d="${s}" fill="none" stroke="${o}" stroke-width="2" stroke-linejoin="round" stroke-linecap="round"/>`);for(let e=0;e<i.length;e++)p.push(`<circle cx="${d(e).toFixed(1)}" cy="${u(i[e]).toFixed(1)}" r="3" fill="${o}" stroke="var(--card)" stroke-width="1.5"/>`)}let m=e.height??300,h=e.showGrid===!1?``:As(s,c,n),g=js(i,r,n,a>1?o:r);return`<svg class="chart-svg" viewBox="0 0 600 ${m}" preserveAspectRatio="xMidYMid meet">
@@ -1083,30 +1083,30 @@ li{margin:4px 0}
1083
1083
  `)}`}).filter(Boolean);if(r.length>0)return{type:`markdown`,title:e.title,value:r.join(`
1084
1084
 
1085
1085
  `)}}let i=n.map(e=>{if(!e||typeof e!=`object`)return null;let n=e,r=typeof n.label==`string`?n.label:typeof n.title==`string`?n.title:null;return r?`- ${typeof n.status==`string`?`${t[n.status.toLowerCase()]??`●`} `:``}${r}${typeof n.description==`string`&&n.description.length>0?` — ${n.description}`:``}`:null}).filter(Boolean);if(i.length>0)return{type:`markdown`,title:e.title,value:i.join(`
1086
- `)}}}return e.type===`table`&&typeof t==`string`?{type:`markdown`,title:e.title,value:t}:new Set([`markdown`,`table`,`code`,`chart`,`metrics`,`tree`,`graph`,`mermaid`]).has(e.type)?e:Hs(e.value,e.title)||{type:`code`,title:e.title,value:typeof e.value==`string`?e.value:JSON.stringify(e.value,null,2)}}function Ws(e){return Array.isArray(e)?e.map(e=>zs(e)?Us(Vs(e)):e):e}Cn.setOptions({async:!1,gfm:!0,breaks:!0}),Cn.use({renderer:{html(e){return z(e.text)}}});function Gs(e){return e.replace(/<table\b/g,`<div class="table-wrap"><table`).replace(/<\/table>/g,`</table></div>`)}function Ks(e,t){let n=[];if(e&&n.push(`<h1>${z(e)}</h1>`),typeof t==`string`)n.push(`<div class="md-content">${Gs(Cn.parse(t))}</div>`);else if(Array.isArray(t))if(t.length===0)n.push(`<p><em>empty</em></p>`);else if(zs(t[0]))for(let e of t)n.push(qs(Vs(e)));else typeof t[0]==`object`&&t[0]!==null?n.push(Js(t)):n.push(`<ul>${t.map(e=>`<li>${z(String(e))}</li>`).join(``)}</ul>`);else if(typeof t==`object`&&t){let e=t;if(Array.isArray(e.blocks)&&e.blocks.length>0)for(let t of e.blocks)zs(t)?n.push(qs(Vs(t))):n.push(`<p>${z(String(t))}</p>`);else Array.isArray(e.metrics)?n.push(Ys(e.metrics)):Array.isArray(e.nodes)&&Array.isArray(e.edges)?(n.push(`<pre class="mermaid">${z($s(e))}</pre>`),n.push(`<script src="https://cdn.jsdelivr.net/npm/mermaid@11/dist/mermaid.min.js"><\/script>`),n.push(`<script>mermaid.run();<\/script>`)):n.push(Zs(e))}else n.push(`<p>${z(String(t))}</p>`);return n.join(`
1087
- `)}function qs(e){let t=[];switch(e.title&&t.push(`<h2>${z(e.title)}</h2>`),e.type){case`markdown`:t.push(`<div class="md-content">${Gs(Cn.parse(String(e.value??``)))}</div>`);break;case`code`:t.push(`<pre><code>${z(String(e.value??``))}</code></pre>`);break;case`mermaid`:t.push(`<pre class="mermaid">${z(String(e.value??``))}</pre>`);break;case`table`:if(Array.isArray(e.value)){let n=e.value;if(n.length>0&&Array.isArray(n[0])){let e=n,r=e[0].map(String),i=e.slice(1).map(e=>Object.fromEntries(r.map((t,n)=>[t,e[n]])));t.push(Js(i))}else t.push(Js(n))}else if(e.value&&typeof e.value==`object`&&`headers`in e.value&&`rows`in e.value){let{headers:n,rows:r}=e.value,i=r.map(e=>Object.fromEntries(n.map((t,n)=>[t,e[n]])));t.push(Js(i))}break;case`metrics`:{let n;Array.isArray(e.value)?n=e.value:e.value&&typeof e.value==`object`&&(n=Object.entries(e.value).map(([e,t])=>({label:e,value:String(t)}))),n&&t.push(Ys(n));break}case`cards`:Array.isArray(e.value)&&t.push(Xs(e.value));break;case`tree`:e.value&&typeof e.value==`object`&&t.push(Zs(e.value));break;case`graph`:e.value&&typeof e.value==`object`&&t.push(`<pre class="mermaid">${z($s(e.value))}</pre>`);break;case`chart`:t.push(ws(e));break;case`timeline`:{let n=e.value;Array.isArray(n)&&(n={items:n.map(e=>({title:String(e.event??e.title??``),phase:e.date==null?e.phase==null?void 0:String(e.phase):String(e.date),description:e.description==null?void 0:String(e.description),status:e.status??`done`}))}),n&&t.push(hs(n));break}case`checklist`:e.value&&t.push(gs(e.value));break;case`comparison`:e.value&&t.push(_s(e.value));break;case`status-board`:e.value&&t.push(vs(e.value));break;case`prompt`:e.value&&t.push(ys(e.value));break;case`progress`:e.value&&t.push(bs(e.value));break;case`text`:t.push(`<div class="md-content">${Cn.parse(String(e.value??``))}</div>`);break;case`heading`:{let n=Math.min(Math.max(Number(e.level)||2,1),6);t.push(`<h${n}>${z(String(e.value??``))}</h${n}>`);break}case`paragraph`:t.push(`<p>${z(String(e.value??``))}</p>`);break;case`separator`:t.push(`<hr class="separator">`);break;case`actions`:{let n=(Array.isArray(e.value)?e.value:[]).map(e=>{if(e.type===`select`){let t=Array.isArray(e.options)?e.options.map(e=>`<option value="${z(String(e.value??e.label??``))}">${z(String(e.label??e.value??``))}</option>`).join(``):``;return`<select class="action-select"><option value="">${z(String(e.label??`Select...`))}</option>${t}</select>`}return`<button class="action-btn action-${String(e.variant??`secondary`)}">${z(String(e.label??``))}</button>`}).join(``);t.push(`<div class="action-bar">${n}</div>`);break}default:t.push(`<pre>${z(JSON.stringify(e.value,null,2))}</pre>`)}return t.join(`
1088
- `)}function Js(e){if(e.length===0)return`<p><em>empty table</em></p>`;let t=Object.keys(e[0]);return`<div class="table-wrap"><table><thead><tr>${t.map(e=>`<th>${z(e)}</th>`).join(``)}</tr></thead><tbody>${e.map(e=>`<tr>${t.map(t=>`<td>${z(String(e[t]??``))}</td>`).join(``)}</tr>`).join(`
1089
- `)}</tbody></table></div>`}function Ys(e){return`<div class="metric-grid">${e.map(e=>`<div class="metric"><div class="metric-value">${z(String(e.value))}</div><div class="metric-label">${z(e.label)}</div></div>`).join(``)}</div>`}function Xs(e){return`<div class="card-grid">${e.map(e=>{let t=[`<div class="card">`];return e.title&&t.push(`<div class="card-title">${z(String(e.title))}</div>`),(e.body||e.description)&&t.push(`<div class="card-body">${z(String(e.body??e.description))}</div>`),(e.badge||e.status)&&t.push(`<span class="badge">${z(String(e.badge??e.status))}</span>`),t.push(`</div>`),t.join(``)}).join(``)}</div>`}function Zs(e){if(`name`in e&&`children`in e&&Array.isArray(e.children))return Qs(e);if(`name`in e&&!(`children`in e))return`<div class="tree-node"><span class="tree-key">${z(String(e.name))}</span></div>`;let t=[];for(let[n,r]of Object.entries(e))typeof r==`object`&&r&&!Array.isArray(r)?t.push(`<div class="tree-node"><span class="tree-key">${z(n)}:</span><div class="tree-children">${Zs(r)}</div></div>`):t.push(`<div class="tree-node"><span class="tree-key">${z(n)}:</span> ${z(Ls(r))}</div>`);return t.join(``)}function Qs(e){let t=z(String(e.name));return!e.children||e.children.length===0?`<div class="tree-node"><span class="tree-key">${t}</span></div>`:`<div class="tree-node"><span class="tree-key">${t}</span><div class="tree-children">${e.children.map(e=>typeof e==`object`&&e?Qs(e):`<div class="tree-node"><span class="tree-key">${z(String(e))}</span></div>`).join(``)}</div></div>`}function $s(e){let t=[`graph LR`];for(let n of e.nodes){let e=Rs(String(n.id??n.name??``)),r=String(n.label??n.name??e);t.push(` ${e}["${r}"]`)}for(let n of e.edges){let e=Rs(String(n.source??n.from??``)),r=Rs(String(n.target??n.to??``)),i=n.label?`|${String(n.label)}|`:``;t.push(` ${e} -->${i} ${r}`)}return t.join(`
1090
- `)}function ec(e,t,n,r){let i=r&&r!==`auto`?sa(r,t):Ks(e,t);return`<!DOCTYPE html>
1086
+ `)}}}return e.type===`table`&&typeof t==`string`?{type:`markdown`,title:e.title,value:t}:new Set([`markdown`,`table`,`code`,`chart`,`metrics`,`tree`,`graph`,`mermaid`]).has(e.type)?e:Hs(e.value,e.title)||{type:`code`,title:e.title,value:typeof e.value==`string`?e.value:JSON.stringify(e.value,null,2)}}function Ws(e){return Array.isArray(e)?e.map(e=>zs(e)?Us(Vs(e)):e):e}Cn.setOptions({async:!1,gfm:!0,breaks:!0}),Cn.use({renderer:{html(e){return B(e.text)}}});function Gs(e){return e.replace(/<table\b/g,`<div class="table-wrap"><table`).replace(/<\/table>/g,`</table></div>`)}function Ks(e,t){let n=[];if(e&&n.push(`<h1>${B(e)}</h1>`),typeof t==`string`)n.push(`<div class="md-content">${Gs(Cn.parse(t))}</div>`);else if(Array.isArray(t))if(t.length===0)n.push(`<p><em>empty</em></p>`);else if(zs(t[0]))for(let e of t)n.push(qs(Vs(e)));else typeof t[0]==`object`&&t[0]!==null?n.push(Js(t)):n.push(`<ul>${t.map(e=>`<li>${B(String(e))}</li>`).join(``)}</ul>`);else if(typeof t==`object`&&t){let e=t;if(Array.isArray(e.blocks)&&e.blocks.length>0)for(let t of e.blocks)zs(t)?n.push(qs(Vs(t))):n.push(`<p>${B(String(t))}</p>`);else Array.isArray(e.metrics)?n.push(Ys(e.metrics)):Array.isArray(e.nodes)&&Array.isArray(e.edges)?(n.push(`<pre class="mermaid">${B($s(e))}</pre>`),n.push(`<script src="https://cdn.jsdelivr.net/npm/mermaid@11/dist/mermaid.min.js"><\/script>`),n.push(`<script>mermaid.run();<\/script>`)):n.push(Zs(e))}else n.push(`<p>${B(String(t))}</p>`);return n.join(`
1087
+ `)}function qs(e){let t=[];switch(e.title&&t.push(`<h2>${B(e.title)}</h2>`),e.type){case`markdown`:t.push(`<div class="md-content">${Gs(Cn.parse(String(e.value??``)))}</div>`);break;case`code`:t.push(`<pre><code>${B(String(e.value??``))}</code></pre>`);break;case`mermaid`:t.push(`<pre class="mermaid">${B(String(e.value??``))}</pre>`);break;case`table`:if(Array.isArray(e.value)){let n=e.value;if(n.length>0&&Array.isArray(n[0])){let e=n,r=e[0].map(String),i=e.slice(1).map(e=>Object.fromEntries(r.map((t,n)=>[t,e[n]])));t.push(Js(i))}else t.push(Js(n))}else if(e.value&&typeof e.value==`object`&&`headers`in e.value&&`rows`in e.value){let{headers:n,rows:r}=e.value,i=r.map(e=>Object.fromEntries(n.map((t,n)=>[t,e[n]])));t.push(Js(i))}break;case`metrics`:{let n;Array.isArray(e.value)?n=e.value:e.value&&typeof e.value==`object`&&(n=Object.entries(e.value).map(([e,t])=>({label:e,value:String(t)}))),n&&t.push(Ys(n));break}case`cards`:Array.isArray(e.value)&&t.push(Xs(e.value));break;case`tree`:e.value&&typeof e.value==`object`&&t.push(Zs(e.value));break;case`graph`:e.value&&typeof e.value==`object`&&t.push(`<pre class="mermaid">${B($s(e.value))}</pre>`);break;case`chart`:t.push(ws(e));break;case`timeline`:{let n=e.value;Array.isArray(n)&&(n={items:n.map(e=>({title:String(e.event??e.title??``),phase:e.date==null?e.phase==null?void 0:String(e.phase):String(e.date),description:e.description==null?void 0:String(e.description),status:e.status??`done`}))}),n&&t.push(hs(n));break}case`checklist`:e.value&&t.push(gs(e.value));break;case`comparison`:e.value&&t.push(_s(e.value));break;case`status-board`:e.value&&t.push(vs(e.value));break;case`prompt`:e.value&&t.push(ys(e.value));break;case`progress`:e.value&&t.push(bs(e.value));break;case`text`:t.push(`<div class="md-content">${Cn.parse(String(e.value??``))}</div>`);break;case`heading`:{let n=Math.min(Math.max(Number(e.level)||2,1),6);t.push(`<h${n}>${B(String(e.value??``))}</h${n}>`);break}case`paragraph`:t.push(`<p>${B(String(e.value??``))}</p>`);break;case`separator`:t.push(`<hr class="separator">`);break;case`actions`:{let n=(Array.isArray(e.value)?e.value:[]).map(e=>{if(e.type===`select`){let t=Array.isArray(e.options)?e.options.map(e=>`<option value="${B(String(e.value??e.label??``))}">${B(String(e.label??e.value??``))}</option>`).join(``):``;return`<select class="action-select"><option value="">${B(String(e.label??`Select...`))}</option>${t}</select>`}return`<button class="action-btn action-${String(e.variant??`secondary`)}">${B(String(e.label??``))}</button>`}).join(``);t.push(`<div class="action-bar">${n}</div>`);break}default:t.push(`<pre>${B(JSON.stringify(e.value,null,2))}</pre>`)}return t.join(`
1088
+ `)}function Js(e){if(e.length===0)return`<p><em>empty table</em></p>`;let t=Object.keys(e[0]);return`<div class="table-wrap"><table><thead><tr>${t.map(e=>`<th>${B(e)}</th>`).join(``)}</tr></thead><tbody>${e.map(e=>`<tr>${t.map(t=>`<td>${B(String(e[t]??``))}</td>`).join(``)}</tr>`).join(`
1089
+ `)}</tbody></table></div>`}function Ys(e){return`<div class="metric-grid">${e.map(e=>`<div class="metric"><div class="metric-value">${B(String(e.value))}</div><div class="metric-label">${B(e.label)}</div></div>`).join(``)}</div>`}function Xs(e){return`<div class="card-grid">${e.map(e=>{let t=[`<div class="card">`];return e.title&&t.push(`<div class="card-title">${B(String(e.title))}</div>`),(e.body||e.description)&&t.push(`<div class="card-body">${B(String(e.body??e.description))}</div>`),(e.badge||e.status)&&t.push(`<span class="badge">${B(String(e.badge??e.status))}</span>`),t.push(`</div>`),t.join(``)}).join(``)}</div>`}function Zs(e){if(`name`in e&&`children`in e&&Array.isArray(e.children))return Qs(e);if(`name`in e&&!(`children`in e))return`<div class="tree-node"><span class="tree-key">${B(String(e.name))}</span></div>`;let t=[];for(let[n,r]of Object.entries(e))typeof r==`object`&&r&&!Array.isArray(r)?t.push(`<div class="tree-node"><span class="tree-key">${B(n)}:</span><div class="tree-children">${Zs(r)}</div></div>`):t.push(`<div class="tree-node"><span class="tree-key">${B(n)}:</span> ${B(Ls(r))}</div>`);return t.join(``)}function Qs(e){let t=B(String(e.name));return!e.children||e.children.length===0?`<div class="tree-node"><span class="tree-key">${t}</span></div>`:`<div class="tree-node"><span class="tree-key">${t}</span><div class="tree-children">${e.children.map(e=>typeof e==`object`&&e?Qs(e):`<div class="tree-node"><span class="tree-key">${B(String(e))}</span></div>`).join(``)}</div></div>`}function $s(e){let t=[`graph LR`];for(let n of e.nodes){let e=Rs(String(n.id??n.name??``)),r=String(n.label??n.name??e);t.push(` ${e}["${r}"]`)}for(let n of e.edges){let e=Rs(String(n.source??n.from??``)),r=Rs(String(n.target??n.to??``)),i=n.label?`|${String(n.label)}|`:``;t.push(` ${e} -->${i} ${r}`)}return t.join(`
1090
+ `)}function ec(e,t,n,r){let i=r&&r!==`auto`?oa(r,t):Ks(e,t);return`<!DOCTYPE html>
1091
1091
  <html lang="en">
1092
1092
  <head>
1093
1093
  <meta charset="UTF-8">
1094
1094
  <meta name="viewport" content="width=device-width,initial-scale=1.0">
1095
- <title>${z(e??`AI Kit Dashboard`)}</title>
1095
+ <title>${B(e??`AI Kit Dashboard`)}</title>
1096
1096
  <link href="https://fonts.googleapis.com/css2?family=Plus+Jakarta+Sans:wght@400;500;600;700&family=JetBrains+Mono:wght@400;500&display=swap" rel="stylesheet">
1097
1097
  <style>${ms()}</style>
1098
1098
  </head>
1099
1099
  <body>
1100
1100
  <div class="dashboard">
1101
1101
  <div class="header">
1102
- <h1>${z(e??`AI Kit Dashboard`)}</h1>
1102
+ <h1>${B(e??`AI Kit Dashboard`)}</h1>
1103
1103
  <div class="subtitle">AI Kit</div>
1104
1104
  </div>
1105
1105
  ${i}
1106
1106
  ${(()=>{let e=Array.isArray(n)?n:[];return e.length===0?``:`
1107
1107
  <div class="actions-bar">
1108
1108
  <h2>Actions</h2>
1109
- <div class="actions-grid">${e.map(e=>{let t=String(e.id??``);if(e.type===`select`&&Array.isArray(e.options)){let n=e.options.map(e=>{let t=typeof e==`string`?e:e.label;return`<option value="${z(typeof e==`string`?e:e.value)}">${z(t)}</option>`}).join(``);return`<div class="action-group"><label>${z(String(e.label??``))}</label><select data-action-id="${z(t)}" onchange="sendCallback(${z(JSON.stringify(t))},this.value)">${n}</select></div>`}return`<button class="action-btn action-${String(e.variant??`default`)}" onclick="sendCallback(${z(JSON.stringify(t))},'clicked')">${z(String(e.label??``))}</button>`}).join(`
1109
+ <div class="actions-grid">${e.map(e=>{let t=String(e.id??``);if(e.type===`select`&&Array.isArray(e.options)){let n=e.options.map(e=>{let t=typeof e==`string`?e:e.label;return`<option value="${B(typeof e==`string`?e:e.value)}">${B(t)}</option>`}).join(``);return`<div class="action-group"><label>${B(String(e.label??``))}</label><select data-action-id="${B(t)}" onchange="sendCallback(${B(JSON.stringify(t))},this.value)">${n}</select></div>`}return`<button class="action-btn action-${String(e.variant??`default`)}" onclick="sendCallback(${B(JSON.stringify(t))},'clicked')">${B(String(e.label??``))}</button>`}).join(`
1110
1110
  `)}</div>
1111
1111
  <div id="action-feedback" class="action-feedback"></div>
1112
1112
  </div>
@@ -1189,7 +1189,7 @@ document.querySelectorAll('.table-wrap').forEach(wrap=>{
1189
1189
  `)}function oc(e){return e.map(e=>`- **${e.label}**: ${e.value}`).join(`
1190
1190
  `)}function sc(e){let t=["```mermaid",`graph LR`];for(let n of e.nodes){let e=Rs(String(n.id??n.name??``)),r=String(n.label??n.name??e);t.push(` ${e}["${r}"]`)}for(let n of e.edges){let e=Rs(String(n.source??n.from??``)),r=Rs(String(n.target??n.to??``)),i=n.label?`|${String(n.label)}|`:``;t.push(` ${e} -->${i} ${r}`)}return t.push("```"),t.join(`
1191
1191
  `)}function cc(e,t=0){let n=` `.repeat(t),r=[];for(let[i,a]of Object.entries(e))typeof a==`object`&&a&&!Array.isArray(a)?(r.push(`${n}- **${i}**:`),r.push(cc(a,t+1))):r.push(`${n}- **${i}**: ${Ls(a)}`);return r.join(`
1192
- `)}const lc=import.meta.dirname??m(v(import.meta.url)),uc=`ui://aikit/present.html`,dc=F.object({type:F.enum([`button`,`select`]).describe(`Action type`),id:F.string().describe(`Unique action identifier`),label:F.string().describe(`Display label`),variant:F.enum([`primary`,`danger`,`default`]).optional().describe(`Button style variant`),options:F.array(F.union([F.string(),F.object({label:F.string(),value:F.string()})])).optional().describe(`Select options (for type=select)`)}),fc={format:F.enum([`html`,`browser`]).default(`html`).describe(`Output format.
1192
+ `)}const lc=import.meta.dirname??m(v(import.meta.url));function uc(e){let t=e;for(let e=0;e<10;e++){try{let e=g(t,`package.json`);if(a(e)&&JSON.parse(s(e,`utf8`)).name===`@vpxa/aikit`)return t}catch{}let e=m(t);if(e===t)break;t=e}return _(e,`..`,`..`,`..`)}const dc=`ui://aikit/present.html`,fc=F.object({type:F.enum([`button`,`select`]).describe(`Action type`),id:F.string().describe(`Unique action identifier`),label:F.string().describe(`Display label`),variant:F.enum([`primary`,`danger`,`default`]).optional().describe(`Button style variant`),options:F.array(F.union([F.string(),F.object({label:F.string(),value:F.string()})])).optional().describe(`Select options (for type=select)`)}),pc={format:F.enum([`html`,`browser`]).default(`html`).describe(`Output format.
1193
1193
  - "html" (default): Rich markdown in chat + embedded UIResource for MCP-UI hosts. Actions shown as numbered choices.
1194
1194
  - "browser": Rich markdown in chat + opens beautiful themed dashboard in the browser. When actions are provided, the browser shows interactive buttons and the tool blocks until user clicks, returning their selection.`),title:F.string().optional().describe(`Optional heading`),content:F.any().describe(`Content to present. Accepts these shapes:
1195
1195
  • markdown string — rendered as rich text
@@ -1211,7 +1211,7 @@ document.querySelectorAll('.table-wrap').forEach(wrap=>{
1211
1211
  - type:"comparison" → value: [{title,items:string[]}] (columns)
1212
1212
  - type:"progress" → value: {label,value:number,max?:number}
1213
1213
  • any JSON — rendered as collapsible tree
1214
- IMPORTANT: For charts, use the ChartValue format above. Do NOT use Chart.js format ({labels, datasets}) — it will be auto-converted but the native format is preferred.`),actions:F.array(dc).optional().describe(`Interactive actions (buttons/selects). In html mode, shown as numbered choices. In browser mode, rendered as clickable buttons and the tool blocks until user clicks.`),template:F.enum([`auto`,`list-sort`,`data-table`,`picker`,`flame-graph`,`form`,`timeline`,`kanban`,`tree`,`diff-view`,`dashboard`]).optional().describe(`UI template for interactive display in MCP Apps hosts.
1214
+ IMPORTANT: For charts, use the ChartValue format above. Do NOT use Chart.js format ({labels, datasets}) — it will be auto-converted but the native format is preferred.`),actions:F.array(fc).optional().describe(`Interactive actions (buttons/selects). In html mode, shown as numbered choices. In browser mode, rendered as clickable buttons and the tool blocks until user clicks.`),template:F.enum([`auto`,`list-sort`,`data-table`,`picker`,`flame-graph`,`form`,`timeline`,`kanban`,`tree`,`diff-view`,`dashboard`]).optional().describe(`UI template for interactive display in MCP Apps hosts.
1215
1215
  - auto (default): detect from content shape
1216
1216
  - list-sort: drag-drop reorderable list — content: { items: [{id, label}] }
1217
1217
  - data-table: filterable sortable table — content: { columns: [{key, label}], rows: [{...}], stats?: [{label, value}] }
@@ -1222,16 +1222,16 @@ IMPORTANT: For charts, use the ChartValue format above. Do NOT use Chart.js form
1222
1222
  - kanban: drag-drop board — content: { columns: [{id, label, color?}], cards: [{id, title, column, tags?[], priority?}] }
1223
1223
  - tree: collapsible tree view — content: { root: {label, children?:[...]} | [...] }
1224
1224
  - diff-view: side-by-side diff — content: { files: [{path, status, additions, deletions, hunks:[{header, changes:[{type, content}]}]}] }
1225
- - dashboard: metric cards grid — content: { metrics: [{label, value, trend?, status?, type?}] }`)};let pc,mc=!1,Y=null;process.on(`exit`,()=>{if(Y){try{Y.close()}catch{}Y=null}});function hc(){return g(lc,`..`,`..`,`..`,`..`,`present`,`dist`,`index.html`)}function gc(){if(!pc)try{pc=s(hc(),`utf-8`)}catch{pc=``}return pc}function _c(e,t){let n=I(`present`),r=gc(),i=`Present content to the user in the best format. Two modes:
1225
+ - dashboard: metric cards grid — content: { metrics: [{label, value, trend?, status?, type?}] }`)};let mc,hc=!1,Y=null;process.on(`exit`,()=>{if(Y){try{Y.close()}catch{}Y=null}});function gc(){return g(uc(lc),`packages`,`present`,`dist`,`index.html`)}function _c(){if(!mc)try{mc=s(gc(),`utf-8`)}catch{mc=``}return mc}function vc(e,t){let n=L(`present`),r=_c(),i=`Present content to the user in the best format. Two modes:
1226
1226
  - "html" (default): Rich markdown in chat + embedded UIResource. Use for display-only content (tables, charts, reports, status boards) where no user interaction is needed.
1227
1227
  - "browser": Serves a themed dashboard on a local URL. Use ONLY when you need user interaction back (confirmations, selections, form input). The tool blocks until user clicks an action button, then returns their selection.
1228
1228
  FORMAT RULE: If no user interaction is needed → use "html". If you need user input back → use "browser".
1229
1229
  CONTENT GUIDE: Pass typed blocks [{ type, title?, value }] for structured content. See the \`content\` parameter description for all supported block types and their value shapes. For charts: use {chartType, data:Record[], xKey, yKeys} — NOT Chart.js format.
1230
- BROWSER WORKFLOW: After calling present with format "browser", you MUST extract the URL from the response and call openBrowserPage({ url }) tool - PlayWright MCP to open it in VS Code Simple Browser. A system browser fallback also opens automatically, but always call openBrowserPage tool - PlayWright MCP yourself.`,a=async({format:e,title:n,content:r,actions:i,template:a})=>(e??`html`)===`browser`||Array.isArray(i)&&i.length>0?await vc(n,r,i,t,a):yc(n,r,i,a);if(r){mc||=(_n(e,`AI Kit Present App`,uc,{description:`Rich interactive content viewer for AI Kit tools`},async()=>({contents:[{uri:uc,mimeType:gn,text:r}]})),!0),vn(e,`present`,{title:n.title,description:i,annotations:n.annotations,inputSchema:fc,_meta:{ui:{resourceUri:uc}}},a);return}e.tool(`present`,i,fc,n.annotations,a)}async function vc(e,t,n,r,i){let a=tc(e,t,{compactTables:!0}),o=ec(e,t,n,i),s=Sn({uri:`ui://aikit/present-browser.html`,content:{type:`rawHtml`,htmlString:o},encoding:`text`,adapters:{mcpApps:{enabled:!0}}}),c,l,u=Array.isArray(n)?n:[],d=``,f;try{Y&&=(Y.close(),null),u.length>0&&(c=new Promise(e=>{l=e}));let e=!1;d=await new Promise((t,n)=>{let r=xn((t,n)=>{if(e||(e=!0,f&&clearTimeout(f)),t.method===`POST`&&t.url===`/callback`){let e=``;t.on(`data`,t=>{e+=t.toString()}),t.on(`end`,()=>{n.writeHead(200,{"Content-Type":`application/json`,"Access-Control-Allow-Origin":`*`}),n.end(`{"ok":true}`);try{let t=JSON.parse(e);l?.(t)}catch{}});return}if(t.method===`OPTIONS`){n.writeHead(204,{"Access-Control-Allow-Origin":`*`,"Access-Control-Allow-Methods":`POST`,"Access-Control-Allow-Headers":`Content-Type`}),n.end();return}n.writeHead(200,{"Content-Type":`text/html; charset=utf-8`}),n.end(o)});r.listen(0,`127.0.0.1`,()=>{let i=r.address();if(typeof i==`object`&&i){Y=r;let n=`http://127.0.0.1:${i.port}`;f=setTimeout(()=>{if(!e)try{yn(process.platform===`win32`?`start "" "${n}"`:process.platform===`darwin`?`open "${n}"`:`xdg-open "${n}"`)}catch{}},8e3),f.unref(),t(n)}else n(Error(`Failed to start present server`))}),setTimeout(()=>{r.close(),Y===r&&(Y=null)},300*1e3).unref()})}catch{}let p=d?`${a}\n\n---\n🌐 **Dashboard opened in browser:** ${d}\n\n**IMPORTANT:** You MUST now open this URL using the \`openBrowserPage\` tool - PlayWright MCP. If that tool is unavailable, open the URL in the system browser as a fallback. If neither works, ask the user to open the URL manually.`:a;if(u.length>0&&r?.available&&c)try{let e=await Promise.race([c,new Promise((e,t)=>setTimeout(()=>t(Error(`timeout`)),300*1e3))]);return{content:[{type:`text`,text:`${p}\n\n✅ **Selected:** ${e.actionId} = \`${e.value}\``},s]}}catch{return{content:[{type:`text`,text:`${p}\n\n⚠️ *No selection received (timed out).*`},s]}}return{content:[{type:`text`,text:p},s]}}function yc(e,t,n,r){let i=Array.isArray(n)?n:[],a=tc(e,t,{compactTables:!0}),o=Sn({uri:`ui://aikit/present-static.html`,content:{type:`rawHtml`,htmlString:ec(e,t,n,r)},encoding:`text`,adapters:{mcpApps:{enabled:!0}}});return{content:[{type:`text`,text:a},o],structuredContent:{title:e,content:Ws(t),actions:i}}}const bc=E(`tools`);function xc(e,t){let n=new fn({structure:new mn,dependencies:new ln,symbols:new hn,patterns:new pn,entryPoints:new dn,diagrams:new un}),r=I(`produce_knowledge`);e.registerTool(`produce_knowledge`,{title:r.title,description:`Run automated codebase analysis and produce synthesis instructions. Executes Tier 1 deterministic analyzers, then returns structured baselines and instructions for you to synthesize knowledge using remember.`,inputSchema:{scope:F.string().optional().describe(`Root path to analyze (defaults to workspace root)`),aspects:F.array(F.enum([`all`,`structure`,`dependencies`,`symbols`,`patterns`,`entry-points`,`diagrams`])).default([`all`]).describe(`Which analysis aspects to run`)},annotations:r.annotations},async({scope:e,aspects:r})=>{try{let i=e??`.`;bc.info(`Running knowledge production`,{rootPath:i,aspects:r});let a=await n.runExtraction(i,r);try{let e=t?.onboardDir??g(i,`.ai`,`context`);o(e,{recursive:!0});let n=`<!-- Generated by produce_knowledge at ${new Date().toISOString()} -->\n\n`;for(let[t,r]of Object.entries(a))r&&typeof r==`string`&&f(g(e,`${t}.md`),n+r,`utf-8`);bc.info(`Knowledge persisted to .ai/context/`,{files:Object.keys(a).length})}catch(e){bc.warn(`Failed to persist knowledge to .ai/context/`,{error:A(e)})}return{content:[{type:`text`,text:n.buildSynthesisInstructions(a,r)+`
1230
+ BROWSER WORKFLOW: After calling present with format "browser", you MUST extract the URL from the response and call openBrowserPage({ url }) tool - PlayWright MCP to open it in VS Code Simple Browser. A system browser fallback also opens automatically, but always call openBrowserPage tool - PlayWright MCP yourself.`,a=async({format:e,title:n,content:r,actions:i,template:a})=>(e??`html`)===`browser`||Array.isArray(i)&&i.length>0?await yc(n,r,i,t,a):bc(n,r,i,a);if(r){hc||=(_n(e,`AI Kit Present App`,dc,{description:`Rich interactive content viewer for AI Kit tools`},async()=>({contents:[{uri:dc,mimeType:gn,text:r}]})),!0),vn(e,`present`,{title:n.title,description:i,annotations:n.annotations,inputSchema:pc,_meta:{ui:{resourceUri:dc}}},a);return}e.tool(`present`,i,pc,n.annotations,a)}async function yc(e,t,n,r,i){let a=tc(e,t,{compactTables:!0}),o=ec(e,t,n,i),s=Sn({uri:`ui://aikit/present-browser.html`,content:{type:`rawHtml`,htmlString:o},encoding:`text`,adapters:{mcpApps:{enabled:!0}}}),c,l,u=Array.isArray(n)?n:[],d=``,f;try{Y&&=(Y.close(),null),u.length>0&&(c=new Promise(e=>{l=e}));let e=!1;d=await new Promise((t,n)=>{let r=xn((t,n)=>{if(e||(e=!0,f&&clearTimeout(f)),t.method===`POST`&&t.url===`/callback`){let e=``;t.on(`data`,t=>{e+=t.toString()}),t.on(`end`,()=>{n.writeHead(200,{"Content-Type":`application/json`,"Access-Control-Allow-Origin":`*`}),n.end(`{"ok":true}`);try{let t=JSON.parse(e);l?.(t)}catch{}});return}if(t.method===`OPTIONS`){n.writeHead(204,{"Access-Control-Allow-Origin":`*`,"Access-Control-Allow-Methods":`POST`,"Access-Control-Allow-Headers":`Content-Type`}),n.end();return}n.writeHead(200,{"Content-Type":`text/html; charset=utf-8`}),n.end(o)});r.listen(0,`127.0.0.1`,()=>{let i=r.address();if(typeof i==`object`&&i){Y=r;let n=`http://127.0.0.1:${i.port}`;f=setTimeout(()=>{if(!e)try{yn(process.platform===`win32`?`start "" "${n}"`:process.platform===`darwin`?`open "${n}"`:`xdg-open "${n}"`)}catch{}},8e3),f.unref(),t(n)}else n(Error(`Failed to start present server`))}),setTimeout(()=>{r.close(),Y===r&&(Y=null)},300*1e3).unref()})}catch{}let p=d?`${a}\n\n---\n🌐 **Dashboard opened in browser:** ${d}\n\n**IMPORTANT:** You MUST now open this URL using the \`openBrowserPage\` tool - PlayWright MCP. If that tool is unavailable, open the URL in the system browser as a fallback. If neither works, ask the user to open the URL manually.`:a;if(u.length>0&&r?.available&&c)try{let e=await Promise.race([c,new Promise((e,t)=>setTimeout(()=>t(Error(`timeout`)),300*1e3))]);return{content:[{type:`text`,text:`${p}\n\n✅ **Selected:** ${e.actionId} = \`${e.value}\``},s]}}catch{return{content:[{type:`text`,text:`${p}\n\n⚠️ *No selection received (timed out).*`},s]}}return{content:[{type:`text`,text:p},s]}}function bc(e,t,n,r){let i=Array.isArray(n)?n:[],a=tc(e,t,{compactTables:!0}),o=Sn({uri:`ui://aikit/present-static.html`,content:{type:`rawHtml`,htmlString:ec(e,t,n,r)},encoding:`text`,adapters:{mcpApps:{enabled:!0}}});return{content:[{type:`text`,text:a},o],structuredContent:{title:e,content:Ws(t),actions:i}}}const xc=E(`tools`);function Sc(e,t){let n=new fn({structure:new mn,dependencies:new ln,symbols:new hn,patterns:new pn,entryPoints:new dn,diagrams:new un}),r=L(`produce_knowledge`);e.registerTool(`produce_knowledge`,{title:r.title,description:`Run automated codebase analysis and produce synthesis instructions. Executes Tier 1 deterministic analyzers, then returns structured baselines and instructions for you to synthesize knowledge using remember.`,inputSchema:{scope:F.string().optional().describe(`Root path to analyze (defaults to workspace root)`),aspects:F.array(F.enum([`all`,`structure`,`dependencies`,`symbols`,`patterns`,`entry-points`,`diagrams`])).default([`all`]).describe(`Which analysis aspects to run`)},annotations:r.annotations},async({scope:e,aspects:r})=>{try{let i=e??`.`;xc.info(`Running knowledge production`,{rootPath:i,aspects:r});let a=await n.runExtraction(i,r);try{let e=t?.onboardDir??g(i,`.ai`,`context`);o(e,{recursive:!0});let n=`<!-- Generated by produce_knowledge at ${new Date().toISOString()} -->\n\n`;for(let[t,r]of Object.entries(a))r&&typeof r==`string`&&f(g(e,`${t}.md`),n+r,`utf-8`);xc.info(`Knowledge persisted to .ai/context/`,{files:Object.keys(a).length})}catch(e){xc.warn(`Failed to persist knowledge to .ai/context/`,{error:A(e)})}return{content:[{type:`text`,text:n.buildSynthesisInstructions(a,r)+`
1231
1231
 
1232
1232
  ---
1233
- _Next: Review the baselines above and use \`remember\` to store synthesized knowledge entries._`}]}}catch(e){return bc.error(`Knowledge production failed`,A(e)),{content:[{type:`text`,text:`Knowledge production failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Sc=E(`tools`);function Cc(e,t){let n=I(`read`);e.registerTool(`read`,{title:n.title,description:`Read the full content of a specific curated knowledge entry by its path. Use list first to discover available entries.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/use-lancedb.md")`)},annotations:n.annotations},async({path:e})=>{try{let n=await t.read(e),r=[`## ${n.title}`,`- **Path**: .ai/curated/${e}`,`- **Category**: ${n.category}`,n.tags.length?`- **Tags**: ${n.tags.join(`, `)}`:null,`- **Version**: ${n.version}`,`- **Created**: ${n.created}`,n.updated===n.created?null:`- **Updated**: ${n.updated}`,``].filter(e=>e!==null).join(`
1234
- `),i=Po(e,n.title,`[${n.category}]`);return{content:[{type:`text`,text:`${r}\n${n.content}\n\n---\n_Next: Use \`update\` to modify this entry, or \`search\` to find related entries._`},...i?[i]:[]]}}catch(e){return Sc.error(`Read failed`,A(e)),{content:[{type:`text`,text:`Read failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const X=E(`tools`);function wc(e,t,n,r,i,a,o){let s=I(`reindex`);e.registerTool(`reindex`,{title:s.title,description:`Trigger re-indexing of the AI Kit index. Can do incremental (only changed files) or full re-index. When smart indexing is active, use force: true to override the automatic trickle indexer.`,inputSchema:{full:F.boolean().default(!1).describe(`If true, force full re-index ignoring file hashes`),force:F.boolean().default(!1).describe(`If true, override smart indexing guard and run reindex anyway`)},annotations:s.annotations},async({full:e,force:s},c)=>{try{if(t.isIndexing)return{content:[{type:`text`,text:`## Reindex Already in Progress
1233
+ _Next: Review the baselines above and use \`remember\` to store synthesized knowledge entries._`}]}}catch(e){return xc.error(`Knowledge production failed`,A(e)),{content:[{type:`text`,text:`Knowledge production failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Cc=E(`tools`);function wc(e,t){let n=L(`read`);e.registerTool(`read`,{title:n.title,description:`Read the full content of a specific curated knowledge entry by its path. Use list first to discover available entries.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/use-lancedb.md")`)},annotations:n.annotations},async({path:e})=>{try{let n=await t.read(e),r=[`## ${n.title}`,`- **Path**: .ai/curated/${e}`,`- **Category**: ${n.category}`,n.tags.length?`- **Tags**: ${n.tags.join(`, `)}`:null,`- **Version**: ${n.version}`,`- **Created**: ${n.created}`,n.updated===n.created?null:`- **Updated**: ${n.updated}`,``].filter(e=>e!==null).join(`
1234
+ `),i=Po(e,n.title,`[${n.category}]`);return{content:[{type:`text`,text:`${r}\n${n.content}\n\n---\n_Next: Use \`update\` to modify this entry, or \`search\` to find related entries._`},...i?[i]:[]]}}catch(e){return Cc.error(`Read failed`,A(e)),{content:[{type:`text`,text:`Read failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const X=E(`tools`);function Tc(e,t,n,r,i,a,o){let s=L(`reindex`);e.registerTool(`reindex`,{title:s.title,description:`Trigger re-indexing of the AI Kit index. Can do incremental (only changed files) or full re-index. When smart indexing is active, use force: true to override the automatic trickle indexer.`,inputSchema:{full:F.boolean().default(!1).describe(`If true, force full re-index ignoring file hashes`),force:F.boolean().default(!1).describe(`If true, override smart indexing guard and run reindex anyway`)},annotations:s.annotations},async({full:e,force:s},c)=>{try{if(t.isIndexing)return{content:[{type:`text`,text:`## Reindex Already in Progress
1235
1235
 
1236
1236
  A reindex operation is currently running. Search and other tools continue to work with existing data. Use \`status({})\` to check when it completes.`}]};if(o===`smart`&&!s)return{content:[{type:`text`,text:`## Smart Indexing Active
1237
1237
 
@@ -1239,36 +1239,36 @@ Smart indexing (trickle mode) is enabled — files are automatically indexed as
1239
1239
 
1240
1240
  **If the index is severely outdated**, use \`reindex({ force: true })\` to override.
1241
1241
 
1242
- Use \`status({})\` to check smart indexing queue status.`}]};let l=Yi(c).createTask(`Reindex`,1);l.progress(0,`Starting ${e?`full`:`incremental`} reindex`),X.info(`Starting background re-index`,{mode:e?`full`:`incremental`});let u=e=>t=>{t.phase===`chunking`&&t.currentFile&&X.debug(`Re-index progress`,{prefix:e,current:t.filesProcessed+1,total:t.filesTotal,file:t.currentFile})};return(e?t.reindexAll(n,u(`Reindex`)):t.index(n,u(`Index`))).then(async e=>{if(X.info(`Background re-index complete`,{filesProcessed:e.filesProcessed,chunksCreated:e.chunksCreated,durationMs:e.durationMs}),l.complete(`Reindex complete: ${e.filesProcessed} files, ${e.chunksCreated} chunks in ${e.durationMs}ms`),i)try{await i.createFtsIndex(),X.info(`FTS index rebuilt after reindex`)}catch(e){X.warn(`FTS index rebuild failed`,A(e))}try{let e=await r.reindexAll();X.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){X.warn(`Curated re-index failed`,A(e))}if(a)try{await a.notifyAfterReindex()}catch(e){X.warn(`Post-reindex resource notification failed`,A(e))}}).catch(e=>{l.fail(`Reindex failed: ${e instanceof Error?e.message:String(e)}`),X.error(`Background reindex failed`,A(e))}),{content:[{type:`text`,text:`## Reindex Started (Background)\n\n- **Mode**: ${e?`Full`:`Incremental`}\n- Search and other tools continue to work with existing data during reindex.\n- Completion will be logged. Use \`status({})\` to check index stats afterward.\n\n---\n_Next: Continue working — the reindex runs in the background._`}]}}catch(e){return X.error(`Reindex failed`,A(e)),{content:[{type:`text`,text:`Reindex failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Tc=E(`tools`);function Ec(e,t,n,r,i){let a=I(`remember`);e.registerTool(`remember`,{title:a.title,description:`Store a new piece of curated knowledge. Use this to persist decisions, patterns, conventions, or any insight worth remembering across sessions.`,inputSchema:{title:F.string().min(3).max(120).describe(`Short descriptive title for the knowledge entry`),content:F.string().min(10).max(1e5).describe(`The markdown content to store`),category:F.string().regex(/^[a-z][a-z0-9-]*$/).describe(`Category slug (e.g., "decisions", "patterns", "conventions", "api-contracts")`),tags:F.array(F.string()).default([]).describe(`Optional tags for filtering`)},annotations:a.annotations},async({title:e,content:a,category:o,tags:s})=>{try{let c=await t.remember(e,a,o,s),l=``;if(n){let t=n.classify(e,a,s);r&&r.recordClassification(e,t.matchingRules.map(e=>e.ruleId),t.pushRecommended),t.matchingRules.length>0&&(l=`\n\n### Classification Signals\n${t.matchingRules.map(e=>` - **${e.ruleId}** (${e.category}, weight: ${e.pushWeight}) — matched: ${e.matchedPatterns.join(`, `)}`).join(`
1242
+ Use \`status({})\` to check smart indexing queue status.`}]};let l=Ji(c).createTask(`Reindex`,1);l.progress(0,`Starting ${e?`full`:`incremental`} reindex`),X.info(`Starting background re-index`,{mode:e?`full`:`incremental`});let u=e=>t=>{t.phase===`chunking`&&t.currentFile&&X.debug(`Re-index progress`,{prefix:e,current:t.filesProcessed+1,total:t.filesTotal,file:t.currentFile})};return(e?t.reindexAll(n,u(`Reindex`)):t.index(n,u(`Index`))).then(async e=>{if(X.info(`Background re-index complete`,{filesProcessed:e.filesProcessed,chunksCreated:e.chunksCreated,durationMs:e.durationMs}),l.complete(`Reindex complete: ${e.filesProcessed} files, ${e.chunksCreated} chunks in ${e.durationMs}ms`),i)try{await i.createFtsIndex(),X.info(`FTS index rebuilt after reindex`)}catch(e){X.warn(`FTS index rebuild failed`,A(e))}try{let e=await r.reindexAll();X.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){X.warn(`Curated re-index failed`,A(e))}if(a)try{await a.notifyAfterReindex()}catch(e){X.warn(`Post-reindex resource notification failed`,A(e))}}).catch(e=>{l.fail(`Reindex failed: ${e instanceof Error?e.message:String(e)}`),X.error(`Background reindex failed`,A(e))}),{content:[{type:`text`,text:`## Reindex Started (Background)\n\n- **Mode**: ${e?`Full`:`Incremental`}\n- Search and other tools continue to work with existing data during reindex.\n- Completion will be logged. Use \`status({})\` to check index stats afterward.\n\n---\n_Next: Continue working — the reindex runs in the background._`}]}}catch(e){return X.error(`Reindex failed`,A(e)),{content:[{type:`text`,text:`Reindex failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Ec=E(`tools`);function Dc(e,t,n,r,i){let a=L(`remember`);e.registerTool(`remember`,{title:a.title,description:`Store a new piece of curated knowledge. Use this to persist decisions, patterns, conventions, or any insight worth remembering across sessions.`,inputSchema:{title:F.string().min(3).max(120).describe(`Short descriptive title for the knowledge entry`),content:F.string().min(10).max(1e5).describe(`The markdown content to store`),category:F.string().regex(/^[a-z][a-z0-9-]*$/).describe(`Category slug (e.g., "decisions", "patterns", "conventions", "api-contracts")`),tags:F.array(F.string()).default([]).describe(`Optional tags for filtering`)},annotations:a.annotations},async({title:e,content:a,category:o,tags:s})=>{try{let c=await t.remember(e,a,o,s),l=``;if(n){let t=n.classify(e,a,s);r&&r.recordClassification(e,t.matchingRules.map(e=>e.ruleId),t.pushRecommended),t.matchingRules.length>0&&(l=`\n\n### Classification Signals\n${t.matchingRules.map(e=>` - **${e.ruleId}** (${e.category}, weight: ${e.pushWeight}) — matched: ${e.matchedPatterns.join(`, `)}`).join(`
1243
1243
  `)}\n- **Push recommended**: ${t.pushRecommended?`yes`:`no`} (max weight: ${t.maxPushWeight})`,t.pushRecommended&&(l+=`
1244
1244
 
1245
- > 💡 This entry matches push rules. Consider \`er_push\` to share with Enterprise RAG.`))}i&&i.notifyAfterCuratedWrite(c.path).catch(()=>{});let u=Po(c.path,e,`[${o}]`);return{content:[{type:`text`,text:`Remembered: **${e}**\n\nStored at \`.ai/curated/${c.path}\` and indexed for semantic search.${l}\n\n---\n_Next: Use \`search\` to verify the entry is findable, or \`list\` to see all curated entries._`},...u?[u]:[]]}}catch(e){return Tc.error(`Remember failed`,A(e)),{content:[{type:`text`,text:`Remember failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Dc=E(`tools`);function Oc(e){let t=I(`replay`);e.registerTool(`replay`,{title:t.title,description:`View or clear the audit trail of recent MCP tool and CLI invocations. Shows tool name, duration, status, and input/output summaries.`,inputSchema:{action:F.enum([`list`,`clear`]).default(`list`).describe(`Action: "list" (default) to view entries, "clear" to wipe the log`),last:F.number().optional().describe(`Number of entries to return (default: 20, list only)`),tool:F.string().optional().describe(`Filter by tool name (list only)`),source:F.enum([`mcp`,`cli`]).optional().describe(`Filter by source: "mcp" or "cli" (list only)`),since:F.string().optional().describe(`ISO timestamp — only show entries after this time (list only)`)},annotations:t.annotations},async({action:e,last:t,tool:n,source:r,since:i})=>{try{if(e===`clear`)return kt(),{content:[{type:`text`,text:`Replay log cleared.`}]};let a=At({last:t,tool:n,source:r,since:i});if(a.length===0)return{content:[{type:`text`,text:`No replay entries found. Activity is logged when tools are invoked via MCP or CLI.`}]};let o=a.map(e=>`${e.ts.split(`T`)[1]?.split(`.`)[0]??e.ts} ${e.status===`ok`?`✓`:`✗`} ${e.tool} (${e.durationMs}ms) [${e.source}]\n in: ${e.input}\n out: ${e.output}`);return jt().catch(()=>{}),{content:[{type:`text`,text:`**Replay Log** (${a.length} entries)\n\n${o.join(`
1245
+ > 💡 This entry matches push rules. Consider \`er_push\` to share with Enterprise RAG.`))}i&&i.notifyAfterCuratedWrite(c.path).catch(()=>{});let u=Po(c.path,e,`[${o}]`);return{content:[{type:`text`,text:`Remembered: **${e}**\n\nStored at \`.ai/curated/${c.path}\` and indexed for semantic search.${l}\n\n---\n_Next: Use \`search\` to verify the entry is findable, or \`list\` to see all curated entries._`},...u?[u]:[]]}}catch(e){return Ec.error(`Remember failed`,A(e)),{content:[{type:`text`,text:`Remember failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Oc=E(`tools`);function kc(e){let t=L(`replay`);e.registerTool(`replay`,{title:t.title,description:`View or clear the audit trail of recent MCP tool and CLI invocations. Shows tool name, duration, status, and input/output summaries.`,inputSchema:{action:F.enum([`list`,`clear`]).default(`list`).describe(`Action: "list" (default) to view entries, "clear" to wipe the log`),last:F.number().optional().describe(`Number of entries to return (default: 20, list only)`),tool:F.string().optional().describe(`Filter by tool name (list only)`),source:F.enum([`mcp`,`cli`]).optional().describe(`Filter by source: "mcp" or "cli" (list only)`),since:F.string().optional().describe(`ISO timestamp — only show entries after this time (list only)`)},annotations:t.annotations},async({action:e,last:t,tool:n,source:r,since:i})=>{try{if(e===`clear`)return kt(),{content:[{type:`text`,text:`Replay log cleared.`}]};let a=At({last:t,tool:n,source:r,since:i});if(a.length===0)return{content:[{type:`text`,text:`No replay entries found. Activity is logged when tools are invoked via MCP or CLI.`}]};let o=a.map(e=>`${e.ts.split(`T`)[1]?.split(`.`)[0]??e.ts} ${e.status===`ok`?`✓`:`✗`} ${e.tool} (${e.durationMs}ms) [${e.source}]\n in: ${e.input}\n out: ${e.output}`);return jt().catch(()=>{}),{content:[{type:`text`,text:`**Replay Log** (${a.length} entries)\n\n${o.join(`
1246
1246
 
1247
- `)}`}]}}catch(e){return Dc.error(`Replay failed`,A(e)),{content:[{type:`text`,text:`Replay failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const kc=E(`tools`);function Ac(e){let t=I(`restore`);e.registerTool(`restore`,{title:t.title,description:`List and restore file snapshots taken before destructive operations (codemod, rename, forget). Use action=list to see available restore points, action=restore with an id to undo.`,inputSchema:{action:F.enum([`list`,`restore`]).describe(`list: show restore points, restore: apply a restore point`),id:F.string().optional().describe(`Restore point ID (required for action=restore)`),limit:F.number().min(1).max(50).default(10).describe(`Max restore points to list`)},annotations:t.annotations},async({action:e,id:t,limit:n})=>{try{if(e===`list`){let e=ot().slice(0,n);return e.length===0?{content:[{type:`text`,text:`No restore points found.`}]}:{content:[{type:`text`,text:`## Restore Points\n\n${e.map(e=>`- **${e.id}** (${e.timestamp}) — ${e.operation}: ${e.description} (${e.files.length} files)`).join(`
1247
+ `)}`}]}}catch(e){return Oc.error(`Replay failed`,A(e)),{content:[{type:`text`,text:`Replay failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Ac=E(`tools`);function jc(e){let t=L(`restore`);e.registerTool(`restore`,{title:t.title,description:`List and restore file snapshots taken before destructive operations (codemod, rename, forget). Use action=list to see available restore points, action=restore with an id to undo.`,inputSchema:{action:F.enum([`list`,`restore`]).describe(`list: show restore points, restore: apply a restore point`),id:F.string().optional().describe(`Restore point ID (required for action=restore)`),limit:F.number().min(1).max(50).default(10).describe(`Max restore points to list`)},annotations:t.annotations},async({action:e,id:t,limit:n})=>{try{if(e===`list`){let e=ot().slice(0,n);return e.length===0?{content:[{type:`text`,text:`No restore points found.`}]}:{content:[{type:`text`,text:`## Restore Points\n\n${e.map(e=>`- **${e.id}** (${e.timestamp}) — ${e.operation}: ${e.description} (${e.files.length} files)`).join(`
1248
1248
  `)}`}]}}if(!t)throw Error(`id is required for restore action`);let r=await Mt(t);return{content:[{type:`text`,text:`Restored ${r.length} files:\n${r.map(e=>`- \`${e}\``).join(`
1249
- `)}`}]}}catch(e){return kc.error(`Restore failed`,A(e)),{content:[{type:`text`,text:`Restore failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const jc=/<\/?curated-context>/gi;function Mc(e){return e.replace(jc,``)}function Nc(e){return`<curated-context>\n${Mc(e)}\n</curated-context>`}const Pc=E(`tools`);function Fc(e){let t=[],n=p(process.cwd());n&&t.push(`[project: ${n}]`);let r=Vt(`__context_boost`);return r&&t.push(`[focus: ${r.value}]`),t.length===0?e:`${t.join(` `)} ${e}`}async function Ic(e,t,n,r,i){if(!e||t>=e.config.fallbackThreshold&&n.length>0)return{results:n,triggered:!1,cacheHit:!1};let a=!1;try{let t=e.cache.get(r);return t?a=!0:(t=await e.client.search(r,i),t.length>0&&e.cache.set(r,t)),t.length>0?{results:de(n,t,i).map(e=>({record:{id:`er:${e.sourcePath}`,content:e.content,sourcePath:e.source===`er`?`[ER] ${e.sourcePath}`:e.sourcePath,startLine:e.startLine??0,endLine:e.endLine??0,contentType:e.contentType??`documentation`,headingPath:e.headingPath,origin:e.source===`er`?`curated`:e.origin??`indexed`,category:e.category,tags:e.tags??[],chunkIndex:0,totalChunks:1,fileHash:``,indexedAt:new Date().toISOString(),version:1},score:e.score})),triggered:!0,cacheHit:a}:{results:n,triggered:!0,cacheHit:a}}catch(e){return Pc.warn(`ER fallback failed`,A(e)),{results:n,triggered:!0,cacheHit:a}}}function Lc(e,t,n=60){let r=new Map;for(let t=0;t<e.length;t++){let i=e[t];r.set(i.record.id,{record:i.record,score:1/(n+t+1)})}for(let e=0;e<t.length;e++){let i=t[e],a=r.get(i.record.id);a?a.score+=1/(n+e+1):r.set(i.record.id,{record:i.record,score:1/(n+e+1)})}return[...r.values()].sort((e,t)=>t.score-e.score).map(({record:e,score:t})=>({record:e,score:t}))}function Rc(e,t){let n=t.toLowerCase().split(/\s+/).filter(e=>e.length>=2);return n.length<2?e:e.map(e=>{let t=e.record.content.toLowerCase();if(t.length>5e3)return e;let r=n.map(e=>{let n=[],r=t.indexOf(e);for(;r!==-1&&n.length<10;)n.push(r),r=t.indexOf(e,r+1);return n});if(r.some(e=>e.length===0))return e;let i=t.length;for(let e of r[0]){let t=e,a=e+n[0].length;for(let i=1;i<r.length;i++){let o=r[i][0],s=Math.abs(o-e);for(let t=1;t<r[i].length;t++){let n=Math.abs(r[i][t]-e);n<s&&(s=n,o=r[i][t])}t=Math.min(t,o),a=Math.max(a,o+n[i].length)}i=Math.min(i,a-t)}let a=1+.25/(1+i/200);return{record:e.record,score:e.score*a}}).sort((e,t)=>t.score-e.score)}function zc(e,t,n=8){let r=new Set(t.toLowerCase().split(/\s+/).filter(e=>e.length>=2)),i=new Map,a=e.length;for(let t of e){let e=t.record.content.split(/[^a-zA-Z0-9_]+/).filter(e=>e.length>=3&&!Bc.has(e.toLowerCase())),n=new Set;for(let t of e){let e=t.toLowerCase();/[_A-Z]/.test(t)&&i.set(`__id__${e}`,1),n.has(e)||(n.add(e),i.set(e,(i.get(e)??0)+1))}}let o=[];for(let[e,t]of i){if(e.startsWith(`__id__`)||r.has(e)||t>a*.8)continue;let n=Math.log(a/t),s=+!!i.has(`__id__${e}`),c=e.length>8?.5:0;o.push({term:e,score:n+s+c})}return o.sort((e,t)=>t.score-e.score).slice(0,n).map(e=>e.term)}const Bc=new Set(`the.and.for.are.but.not.you.all.can.had.her.was.one.our.out.has.have.from.this.that.with.they.been.said.each.which.their.will.other.about.many.then.them.these.some.would.make.like.into.could.time.very.when.come.just.know.take.people.also.back.after.only.more.than.over.such.import.export.const.function.return.true.false.null.undefined.string.number.boolean.void.type.interface`.split(`.`));async function Vc(e,t){try{let n=await e.getStats();if(!n.lastIndexedAt)return;let r=new Date(n.lastIndexedAt).getTime(),i=Date.now(),a=[...new Set(t.map(e=>e.record.sourcePath))].filter(e=>!e.startsWith(`[ER]`)).slice(0,5);if(a.length===0)return;let o=0;for(let e of a)try{(await ne(e)).mtimeMs>r&&o++}catch{o++}if(o>0){let e=i-r,t=Math.floor(e/6e4),n=t<1?`<1 min`:`${t} min`;return`> ⚠️ **Index may be stale** — ${o} file(s) modified since last index (${n} ago). Use \`reindex\` to refresh.`}}catch{}}function Hc(e,t,n,r,i,a,o){let s=I(`search`);e.registerTool(`search`,{title:s.title,description:`Search AI Kit for code, docs, and prior decisions. Default choice for discovery. Modes: hybrid (default), semantic, keyword. For multi-strategy precision queries use find; for a known file path use lookup.`,outputSchema:hi,inputSchema:{query:F.string().max(5e3).describe(`Natural language search query`),limit:F.number().min(1).max(20).default(5).describe(`Maximum results to return`),search_mode:F.enum([`hybrid`,`semantic`,`keyword`]).default(`hybrid`).describe(`Search strategy: hybrid (vector + FTS + RRF fusion, default), semantic (vector only), keyword (FTS only)`),content_type:F.enum(b).optional().describe(`Filter by content type`),source_type:F.enum(C).optional().describe(`Coarse filter: "source" (code only), "documentation" (md, curated), "test", "config". Overrides content_type if both set.`),origin:F.enum(S).optional().describe(`Filter by knowledge origin`),category:F.string().optional().describe(`Filter by category (e.g., decisions, patterns, conventions)`),tags:F.array(F.string()).optional().describe(`Filter by tags (returns results matching ANY of the specified tags)`),min_score:F.number().min(0).max(1).default(.25).describe(`Minimum similarity score`),graph_hops:F.number().min(0).max(3).default(1).describe(`Number of graph hops to augment results with connected entities (0 = disabled, 1 = direct connections, 2-3 = deeper traversal). Default 1 provides module/symbol context automatically.`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`),dedup:F.enum([`file`,`chunk`]).default(`chunk`).describe(`Deduplication mode: "chunk" (default, show all matching chunks) or "file" (collapse chunks from same file into single result with merged line ranges)`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all registered workspaces. Only works in user-level install mode.`)},annotations:s.annotations},async({query:e,limit:s,search_mode:c,content_type:l,source_type:u,origin:d,category:f,tags:p,min_score:m,graph_hops:h,max_tokens:g,dedup:_,workspaces:v})=>{try{let y={limit:s,minScore:m,contentType:l,sourceType:u,origin:d,category:f,tags:p},b,x=!1,S=!1,C=``,w=Fc(e);if(c===`keyword`)b=await n.ftsSearch(e,y),b=b.slice(0,s);else if(c===`semantic`){let r=await t.embedQuery(w);b=await n.search(r,y);let a=await Ic(i,b[0]?.score??0,b,e,s);b=a.results,x=a.triggered,S=a.cacheHit}else{let r=await t.embedQuery(w),[a,o]=await Promise.all([n.search(r,{...y,limit:s*2}),n.ftsSearch(e,{...y,limit:s*2}).catch(()=>[])]);b=Lc(a,o).slice(0,s);let c=await Ic(i,a[0]?.score??0,b,e,s);b=c.results,x=c.triggered,S=c.cacheHit}a&&a.recordSearch(e,x,S),b.length>1&&(b=Rc(b,e));let E=``;if(v&&v.length>0){let n=Ia(v,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:i}=await La(n);try{let i;i=c===`keyword`?await za(r,e,{...y,limit:s}):await Ra(r,await t.embedQuery(e),{...y,limit:s});for(let e of i)b.push({record:{...e.record,sourcePath:`[${e.workspace}] ${e.record.sourcePath}`},score:e.score});b=b.sort((e,t)=>t.score-e.score).slice(0,s),E=` + ${n.length} workspace(s)`}finally{await i()}}}if(_===`file`&&b.length>1){let e=new Map;for(let t of b){let n=t.record.sourcePath,r=e.get(n);r?(t.score>r.best.score&&(r.best=t),r.ranges.push({start:t.record.startLine,end:t.record.endLine})):e.set(n,{best:t,ranges:[{start:t.record.startLine,end:t.record.endLine}]})}b=[...e.values()].sort((e,t)=>t.best.score-e.best.score).map(({best:e,ranges:t})=>({record:{...e.record,content:t.length>1?`${e.record.content}\n\n_Matched ${t.length} sections: ${t.sort((e,t)=>e.start-t.start).map(e=>`L${e.start}-${e.end}`).join(`, `)}_`:e.record.content},score:e.score}))}if(b.length===0){if(o?.available)try{let r=(await o.createMessage({prompt:`The search query "${e}" returned 0 results in AI Kit code search. Suggest ONE alternative search query that might find relevant results. Reply with ONLY the alternative query, nothing else.`,systemPrompt:`You are a search query optimizer for AI Kit code search. Generate a single alternative query.`,maxTokens:100})).text.trim().split(`
1250
- `)[0].slice(0,500);if(r&&r!==e){let i=await t.embedQuery(r),a=await n.search(i,y);a.length>0&&(b=a,C=`> _Original query "${e}" returned 0 results. Auto-reformulated to "${r}"._\n\n`,Pc.info(`Smart search fallback succeeded`,{originalQuery:e,altQuery:r,resultCount:a.length}))}}catch(e){Pc.debug(`Smart search fallback failed`,{error:String(e)})}if(b.length===0)return{content:[{type:`text`,text:`No results found for the given query.`}],structuredContent:{results:[],totalResults:0,searchMode:c,query:e}}}let D,O;if(h>0&&!r&&(O="> **Note:** `graph_hops` was set but no graph store is available. Graph augmentation skipped."),h>0&&r)try{let e=await Ye(r,b.map(e=>({recordId:e.record.id,score:e.score,sourcePath:e.record.sourcePath})),{hops:h,maxPerHit:5});D=new Map;for(let t of e)if(t.graphContext.nodes.length>0){let e=t.graphContext.nodes.slice(0,5).map(e=>` - **${e.name}** (${e.type})`).join(`
1249
+ `)}`}]}}catch(e){return Ac.error(`Restore failed`,A(e)),{content:[{type:`text`,text:`Restore failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Mc=/<\/?curated-context>/gi;function Nc(e){return e.replace(Mc,``)}function Pc(e){return`<curated-context>\n${Nc(e)}\n</curated-context>`}const Fc=E(`tools`);function Ic(e){let t=[],n=p(process.cwd());n&&t.push(`[project: ${n}]`);let r=Vt(`__context_boost`);return r&&t.push(`[focus: ${r.value}]`),t.length===0?e:`${t.join(` `)} ${e}`}async function Lc(e,t,n,r,i){if(!e||t>=e.config.fallbackThreshold&&n.length>0)return{results:n,triggered:!1,cacheHit:!1};let a=!1;try{let t=e.cache.get(r);return t?a=!0:(t=await e.client.search(r,i),t.length>0&&e.cache.set(r,t)),t.length>0?{results:de(n,t,i).map(e=>({record:{id:`er:${e.sourcePath}`,content:e.content,sourcePath:e.source===`er`?`[ER] ${e.sourcePath}`:e.sourcePath,startLine:e.startLine??0,endLine:e.endLine??0,contentType:e.contentType??`documentation`,headingPath:e.headingPath,origin:e.source===`er`?`curated`:e.origin??`indexed`,category:e.category,tags:e.tags??[],chunkIndex:0,totalChunks:1,fileHash:``,indexedAt:new Date().toISOString(),version:1},score:e.score})),triggered:!0,cacheHit:a}:{results:n,triggered:!0,cacheHit:a}}catch(e){return Fc.warn(`ER fallback failed`,A(e)),{results:n,triggered:!0,cacheHit:a}}}function Rc(e,t,n=60){let r=new Map;for(let t=0;t<e.length;t++){let i=e[t];r.set(i.record.id,{record:i.record,score:1/(n+t+1)})}for(let e=0;e<t.length;e++){let i=t[e],a=r.get(i.record.id);a?a.score+=1/(n+e+1):r.set(i.record.id,{record:i.record,score:1/(n+e+1)})}return[...r.values()].sort((e,t)=>t.score-e.score).map(({record:e,score:t})=>({record:e,score:t}))}function zc(e,t){let n=t.toLowerCase().split(/\s+/).filter(e=>e.length>=2);return n.length<2?e:e.map(e=>{let t=e.record.content.toLowerCase();if(t.length>5e3)return e;let r=n.map(e=>{let n=[],r=t.indexOf(e);for(;r!==-1&&n.length<10;)n.push(r),r=t.indexOf(e,r+1);return n});if(r.some(e=>e.length===0))return e;let i=t.length;for(let e of r[0]){let t=e,a=e+n[0].length;for(let i=1;i<r.length;i++){let o=r[i][0],s=Math.abs(o-e);for(let t=1;t<r[i].length;t++){let n=Math.abs(r[i][t]-e);n<s&&(s=n,o=r[i][t])}t=Math.min(t,o),a=Math.max(a,o+n[i].length)}i=Math.min(i,a-t)}let a=1+.25/(1+i/200);return{record:e.record,score:e.score*a}}).sort((e,t)=>t.score-e.score)}function Bc(e,t,n=8){let r=new Set(t.toLowerCase().split(/\s+/).filter(e=>e.length>=2)),i=new Map,a=e.length;for(let t of e){let e=t.record.content.split(/[^a-zA-Z0-9_]+/).filter(e=>e.length>=3&&!Vc.has(e.toLowerCase())),n=new Set;for(let t of e){let e=t.toLowerCase();/[_A-Z]/.test(t)&&i.set(`__id__${e}`,1),n.has(e)||(n.add(e),i.set(e,(i.get(e)??0)+1))}}let o=[];for(let[e,t]of i){if(e.startsWith(`__id__`)||r.has(e)||t>a*.8)continue;let n=Math.log(a/t),s=+!!i.has(`__id__${e}`),c=e.length>8?.5:0;o.push({term:e,score:n+s+c})}return o.sort((e,t)=>t.score-e.score).slice(0,n).map(e=>e.term)}const Vc=new Set(`the.and.for.are.but.not.you.all.can.had.her.was.one.our.out.has.have.from.this.that.with.they.been.said.each.which.their.will.other.about.many.then.them.these.some.would.make.like.into.could.time.very.when.come.just.know.take.people.also.back.after.only.more.than.over.such.import.export.const.function.return.true.false.null.undefined.string.number.boolean.void.type.interface`.split(`.`));async function Hc(e,t){try{let n=await e.getStats();if(!n.lastIndexedAt)return;let r=new Date(n.lastIndexedAt).getTime(),i=Date.now(),a=[...new Set(t.map(e=>e.record.sourcePath))].filter(e=>!e.startsWith(`[ER]`)).slice(0,5);if(a.length===0)return;let o=0;for(let e of a)try{(await ne(e)).mtimeMs>r&&o++}catch{o++}if(o>0){let e=i-r,t=Math.floor(e/6e4),n=t<1?`<1 min`:`${t} min`;return`> ⚠️ **Index may be stale** — ${o} file(s) modified since last index (${n} ago). Use \`reindex\` to refresh.`}}catch{}}function Uc(e,t,n,r,i,a,o){let s=L(`search`);e.registerTool(`search`,{title:s.title,description:`Search AI Kit for code, docs, and prior decisions. Default choice for discovery. Modes: hybrid (default), semantic, keyword. For multi-strategy precision queries use find; for a known file path use lookup.`,outputSchema:mi,inputSchema:{query:F.string().max(5e3).describe(`Natural language search query`),limit:F.number().min(1).max(20).default(5).describe(`Maximum results to return`),search_mode:F.enum([`hybrid`,`semantic`,`keyword`]).default(`hybrid`).describe(`Search strategy: hybrid (vector + FTS + RRF fusion, default), semantic (vector only), keyword (FTS only)`),content_type:F.enum(b).optional().describe(`Filter by content type`),source_type:F.enum(C).optional().describe(`Coarse filter: "source" (code only), "documentation" (md, curated), "test", "config". Overrides content_type if both set.`),origin:F.enum(S).optional().describe(`Filter by knowledge origin`),category:F.string().optional().describe(`Filter by category (e.g., decisions, patterns, conventions)`),tags:F.array(F.string()).optional().describe(`Filter by tags (returns results matching ANY of the specified tags)`),min_score:F.number().min(0).max(1).default(.25).describe(`Minimum similarity score`),graph_hops:F.number().min(0).max(3).default(1).describe(`Number of graph hops to augment results with connected entities (0 = disabled, 1 = direct connections, 2-3 = deeper traversal). Default 1 provides module/symbol context automatically.`),max_tokens:F.number().min(100).max(5e4).optional().describe(`Maximum token budget for the response. When set, output is truncated to fit.`),dedup:F.enum([`file`,`chunk`]).default(`chunk`).describe(`Deduplication mode: "chunk" (default, show all matching chunks) or "file" (collapse chunks from same file into single result with merged line ranges)`),workspaces:F.array(F.string()).optional().describe(`Cross-workspace search: partition names or folder basenames to include. Use ["*"] for all registered workspaces. Only works in user-level install mode.`)},annotations:s.annotations},async({query:e,limit:s,search_mode:c,content_type:l,source_type:u,origin:d,category:f,tags:p,min_score:m,graph_hops:h,max_tokens:g,dedup:_,workspaces:v})=>{try{let y={limit:s,minScore:m,contentType:l,sourceType:u,origin:d,category:f,tags:p},b,x=!1,S=!1,C=``,w=Ic(e);if(c===`keyword`)b=await n.ftsSearch(e,y),b=b.slice(0,s);else if(c===`semantic`){let r=await t.embedQuery(w);b=await n.search(r,y);let a=await Lc(i,b[0]?.score??0,b,e,s);b=a.results,x=a.triggered,S=a.cacheHit}else{let r=await t.embedQuery(w),[a,o]=await Promise.all([n.search(r,{...y,limit:s*2}),n.ftsSearch(e,{...y,limit:s*2}).catch(()=>[])]);b=Rc(a,o).slice(0,s);let c=await Lc(i,a[0]?.score??0,b,e,s);b=c.results,x=c.triggered,S=c.cacheHit}a&&a.recordSearch(e,x,S),b.length>1&&(b=zc(b,e));let E=``;if(v&&v.length>0){let n=Fa(v,T(process.cwd()));if(n.length>0){let{stores:r,closeAll:i}=await Ia(n);try{let i;i=c===`keyword`?await Ra(r,e,{...y,limit:s}):await La(r,await t.embedQuery(e),{...y,limit:s});for(let e of i)b.push({record:{...e.record,sourcePath:`[${e.workspace}] ${e.record.sourcePath}`},score:e.score});b=b.sort((e,t)=>t.score-e.score).slice(0,s),E=` + ${n.length} workspace(s)`}finally{await i()}}}if(_===`file`&&b.length>1){let e=new Map;for(let t of b){let n=t.record.sourcePath,r=e.get(n);r?(t.score>r.best.score&&(r.best=t),r.ranges.push({start:t.record.startLine,end:t.record.endLine})):e.set(n,{best:t,ranges:[{start:t.record.startLine,end:t.record.endLine}]})}b=[...e.values()].sort((e,t)=>t.best.score-e.best.score).map(({best:e,ranges:t})=>({record:{...e.record,content:t.length>1?`${e.record.content}\n\n_Matched ${t.length} sections: ${t.sort((e,t)=>e.start-t.start).map(e=>`L${e.start}-${e.end}`).join(`, `)}_`:e.record.content},score:e.score}))}if(b.length===0){if(o?.available)try{let r=(await o.createMessage({prompt:`The search query "${e}" returned 0 results in AI Kit code search. Suggest ONE alternative search query that might find relevant results. Reply with ONLY the alternative query, nothing else.`,systemPrompt:`You are a search query optimizer for AI Kit code search. Generate a single alternative query.`,maxTokens:100})).text.trim().split(`
1250
+ `)[0].slice(0,500);if(r&&r!==e){let i=await t.embedQuery(r),a=await n.search(i,y);a.length>0&&(b=a,C=`> _Original query "${e}" returned 0 results. Auto-reformulated to "${r}"._\n\n`,Fc.info(`Smart search fallback succeeded`,{originalQuery:e,altQuery:r,resultCount:a.length}))}}catch(e){Fc.debug(`Smart search fallback failed`,{error:String(e)})}if(b.length===0)return{content:[{type:`text`,text:`No results found for the given query.`}],structuredContent:{results:[],totalResults:0,searchMode:c,query:e}}}let D,O;if(h>0&&!r&&(O="> **Note:** `graph_hops` was set but no graph store is available. Graph augmentation skipped."),h>0&&r)try{let e=await Ye(r,b.map(e=>({recordId:e.record.id,score:e.score,sourcePath:e.record.sourcePath})),{hops:h,maxPerHit:5});D=new Map;for(let t of e)if(t.graphContext.nodes.length>0){let e=t.graphContext.nodes.slice(0,5).map(e=>` - **${e.name}** (${e.type})`).join(`
1251
1251
  `),n=t.graphContext.edges.slice(0,5).map(e=>` - ${e.fromId} —[${e.type}]→ ${e.toId}`).join(`
1252
1252
  `),r=[`- **Graph Context** (${h} hop${h>1?`s`:``}):`];e&&r.push(` Entities:\n${e}`),n&&r.push(` Relationships:\n${n}`),D.set(t.recordId,r.join(`
1253
- `))}}catch(e){Pc.warn(`Graph augmentation failed`,A(e)),O=`> **Note:** Graph augmentation failed. Results shown without graph context.`}let k=Date.now();for(let e of b)if(e.record.origin===`curated`&&e.record.indexedAt){let t=k-new Date(e.record.indexedAt).getTime(),n=Math.max(0,t/864e5);e.score*=.95**n}b.sort((e,t)=>t.score-e.score),b=be(b);let ee=b.map((e,t)=>{let n=e.record;return`${`### Result ${t+1} (score: ${e.score.toFixed(3)})`}\n${[`- **Source**: ${n.sourcePath}`,n.headingPath?`- **Section**: ${n.headingPath}`:null,`- **Type**: ${n.contentType}`,n.startLine?`- **Lines**: ${n.startLine}-${n.endLine}`:null,n.origin===`indexed`?null:`- **Origin**: ${n.origin}`,n.category?`- **Category**: ${n.category}`:null,n.tags?.length?`- **Tags**: ${n.tags.join(`, `)}`:null,D?.get(n.id)??null].filter(Boolean).join(`
1254
- `)}\n\n${n.origin===`curated`?Nc(n.content):n.content}`}).join(`
1253
+ `))}}catch(e){Fc.warn(`Graph augmentation failed`,A(e)),O=`> **Note:** Graph augmentation failed. Results shown without graph context.`}let k=Date.now();for(let e of b)if(e.record.origin===`curated`&&e.record.indexedAt){let t=k-new Date(e.record.indexedAt).getTime(),n=Math.max(0,t/864e5);e.score*=.95**n}b.sort((e,t)=>t.score-e.score),b=be(b);let ee=b.map((e,t)=>{let n=e.record;return`${`### Result ${t+1} (score: ${e.score.toFixed(3)})`}\n${[`- **Source**: ${n.sourcePath}`,n.headingPath?`- **Section**: ${n.headingPath}`:null,`- **Type**: ${n.contentType}`,n.startLine?`- **Lines**: ${n.startLine}-${n.endLine}`:null,n.origin===`indexed`?null:`- **Origin**: ${n.origin}`,n.category?`- **Category**: ${n.category}`:null,n.tags?.length?`- **Tags**: ${n.tags.join(`, `)}`:null,D?.get(n.id)??null].filter(Boolean).join(`
1254
+ `)}\n\n${n.origin===`curated`?Pc(n.content):n.content}`}).join(`
1255
1255
 
1256
1256
  ---
1257
1257
 
1258
- `),j=(c===`hybrid`?`hybrid (vector + keyword RRF)`:c===`keyword`?`keyword (FTS)`:`semantic (vector)`)+E,M=zc(b,e),te=M.length>0?`\n_Distinctive terms: ${M.map(e=>`\`${e}\``).join(`, `)}_`:``,ne=await Vc(n,b),N=[];if(b.length===0)N.push("`reindex` — no results found, index may be stale"),N.push("`find` — try federated search with glob/regex");else{let e=b[0]?.record.sourcePath;e&&N.push(`\`lookup\` — see all chunks from \`${e}\``),N.push("`symbol` — resolve a specific symbol from the results"),N.push("`compact` — compress a result file for focused reading")}let re=[O?`${O}\n\n`:``,ne?`${ne}\n\n`:``,ee,`\n\n---\n_Search mode: ${j} | ${b.length} results_${te}`,`\n_Next: ${N.join(` | `)}_`],ie=C+re.join(``);g&&(ie=Xt(ie,g));let ae=new Set,oe=[];for(let e of b){if(e.record.origin!==`curated`||e.record.sourcePath.startsWith(`[`))continue;let t=Io(e.record.sourcePath);if(!t)continue;let n=Po(t,e.record.headingPath??t);n&&!ae.has(n.uri)&&(ae.add(n.uri),oe.push(n))}return{content:[{type:`text`,text:ie},...oe],structuredContent:{results:b.map(e=>({sourcePath:e.record.sourcePath,contentType:e.record.contentType,score:e.score,...e.record.headingPath?{headingPath:e.record.headingPath}:{},...e.record.startLine?{startLine:e.record.startLine}:{},...e.record.endLine?{endLine:e.record.endLine}:{},...e.record.origin===`indexed`?{}:{origin:e.record.origin},...e.record.category?{category:e.record.category}:{},...e.record.tags?.length?{tags:e.record.tags}:{}})),totalResults:b.length,searchMode:c,query:e}}}catch(e){return Pc.error(`Search failed`,A(e)),{content:[{type:`text`,text:`Search failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Uc=E(`tools`);function Wc(e,t){let n=I(`session_digest`);e.registerTool(`session_digest`,{title:n.title,description:`Compress the current session's tool trajectory into a focused digest. Aggregates replay log, stash state, and checkpoints. Supports deterministic and LLM-assisted (sampling) modes with configurable token budget.`,inputSchema:{scope:F.enum([`tools`,`stash`,`all`]).default(`all`).describe(`What to include: tools (replay only), stash (stash + checkpoints only), or all`),since:F.string().optional().describe(`ISO timestamp — only include activity after this time`),last:F.number().min(1).max(500).optional().describe(`Max replay entries to consider (default: 50)`),focus:F.string().optional().describe(`Focus query — prioritize entries matching this topic`),mode:F.enum([`deterministic`,`sampling`]).default(`deterministic`).describe(`Summary mode: deterministic (fast, template-based) or sampling (LLM-assisted, richer)`),token_budget:F.number().min(100).max(8e3).default(2e3).describe(`Target token budget for the digest output`),persist:F.boolean().default(!0).describe(`Auto-save digest to stash for crash recovery`)},annotations:n.annotations},async({scope:e,since:n,last:r,focus:i,mode:a,token_budget:o,persist:s})=>{try{let c={scope:e,since:n,last:r,focus:i,mode:a,tokenBudget:o,persist:s};return{content:[{type:`text`,text:(a===`sampling`&&t?.available?await Lt(c,async(e,n,r)=>(await t.createMessage({prompt:e,systemPrompt:n,maxTokens:r})).text):It(c)).digest}]}}catch(e){return Uc.error(`Session digest failed`,A(e)),{content:[{type:`text`,text:`Session digest failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Gc=E(`tools`);function Kc(e,t,n,r=15e3){let i,a=new Promise(e=>{i=setTimeout(()=>{Gc.warn(`Status sub-operation "${n}" timed out after ${r}ms`),e({value:t,timedOut:!0})},r)});return Promise.race([e.then(e=>(clearTimeout(i),{value:e,timedOut:!1}),e=>(clearTimeout(i),Gc.warn(`Status sub-operation "${n}" failed: ${e instanceof Error?e.message:String(e)}`),{value:t,timedOut:!1})),a])}const qc=5*6e4;let Jc=null,Yc=null;function Xc(e,t,n,r){let i=Math.min(e/2e4,1),a=Math.min((t+n)/5e4,1),o=Math.min(r/200,1);return Math.round(i*40+a*35+o*25)}function Zc(){let e=Date.now();if(Jc&&e-Jc.ts<qc)return Jc.value;try{let t=_(N(),`.copilot`,`.aikit-scaffold.json`);if(!a(t))return Jc={value:null,ts:e},null;let n=JSON.parse(s(t,`utf-8`)).version??null;return Jc={value:n,ts:e},n}catch{return Jc={value:null,ts:e},null}}function Qc(){let e=Date.now();if(Yc&&e-Yc.ts<qc)return Yc.value;try{let t=_(process.cwd(),`.github`,`.aikit-scaffold.json`);if(!a(t))return Yc={value:null,ts:e},null;let n=JSON.parse(s(t,`utf-8`)).version??null;return Yc={value:n,ts:e},n}catch{return Yc={value:null,ts:e},null}}function $c(e){let t=I(`status`);e.registerTool(`status`,{title:t.title,description:`Get the current status and statistics of the AI Kit index.`,outputSchema:ai,annotations:t.annotations},async()=>{let e=r(),t=Zc(),a=Qc(),o=t!=null&&t!==e,s=a!=null&&a!==e,c=[`## AI Kit Status`,``,`⏳ **AI Kit is initializing** — index stats will be available shortly.`,``,`### Runtime`,`- **Tree-sitter (WASM)**: ${re.get()?`✅ Available (AST analysis)`:`⚠ Unavailable (regex fallback)`}`,``,`### Version`,`- **Server**: ${e}`,`- **Scaffold (user)**: ${t??`not installed`}`,`- **Scaffold (workspace)**: ${a??`not installed`}`];if(o||s){let r=i(),l=[];o&&l.push(`user scaffold v${t}`),s&&l.push(`workspace scaffold v${a}`);let u=l.join(`, `);r.state===`success`?c.push(``,`### ✅ Upgrade Applied`,`- Server v${e} — ${u} auto-upgraded successfully.`,`- _Restart the MCP server to use the updated version._`):r.state===`pending`?c.push(``,`### ⏳ Upgrade In Progress`,`- Server v${e} ≠ ${u}`,`- Auto-upgrade is running in the background…`):r.state===`failed`?(n(),c.push(``,`### ⬆ Upgrade Available (auto-upgrade failed, retrying)`,`- Server v${e} ≠ ${u}`,`- Error: ${r.error??`unknown`}`)):(n(),c.push(``,`### ⬆ Upgrade Available`,`- Server v${e} ≠ ${u}`,`- Auto-upgrade triggered — check again shortly.`))}let l={totalRecords:0,totalFiles:0,lastIndexedAt:null,onboarded:!1,onboardDir:``,contentTypes:{},wasmAvailable:!!re.get(),graphStats:null,curatedCount:0,serverVersion:e,scaffoldVersion:t??null,workspaceScaffoldVersion:a??null,upgradeAvailable:o||s,contextPressure:0};return{content:[{type:`text`,text:c.join(`
1259
- `)}],structuredContent:l}})}function el(e,t,o,s,c,l,u,f){let p=I(`status`);e.registerTool(`status`,{title:p.title,description:`Get the current status and statistics of the AI Kit index.`,outputSchema:ai,annotations:p.annotations},async()=>{let e=[];try{let[p,m]=await Promise.all([Kc(t.getStats(),{totalRecords:0,totalFiles:0,lastIndexedAt:null,contentTypeBreakdown:{}},`store.getStats`),Kc(t.listSourcePaths(),[],`store.listSourcePaths`)]),h=p.value;p.timedOut&&e.push(`⚠ Index stats timed out — values may be incomplete`);let g=m.value;m.timedOut&&e.push(`⚠ File listing timed out`);let v=null,b=0,x=[`## AI Kit Status`,``,`- **Total Records**: ${h.totalRecords}`,`- **Total Files**: ${h.totalFiles}`,`- **Last Indexed**: ${h.lastIndexedAt??`Never`}`,``,`### Content Types`,...Object.entries(h.contentTypeBreakdown).map(([e,t])=>`- ${e}: ${t}`),``,`### Indexed Files`,...g.slice(0,50).map(e=>`- ${e}`),g.length>50?`\n... and ${g.length-50} more files`:``];if(o)try{let t=await Kc(o.getStats(),{nodeCount:0,edgeCount:0,nodeTypes:{},edgeTypes:{}},`graphStore.getStats`);if(t.timedOut)e.push(`⚠ Graph stats timed out`),x.push(``,`### Knowledge Graph`,`- Graph stats timed out`);else{let e=t.value;v={nodes:e.nodeCount,edges:e.edgeCount},x.push(``,`### Knowledge Graph`,`- **Nodes**: ${e.nodeCount}`,`- **Edges**: ${e.edgeCount}`,...Object.entries(e.nodeTypes).map(([e,t])=>` - ${e}: ${t}`));try{let e=await Kc(o.validate(),{valid:!0,danglingEdges:[],orphanNodes:[],stats:{nodeCount:0,edgeCount:0,nodeTypes:{},edgeTypes:{}}},`graphStore.validate`);if(!e.timedOut){let t=e.value;t.valid||x.push(`- **⚠ Integrity Issues**: ${t.danglingEdges.length} dangling edges`),t.orphanNodes.length>0&&x.push(`- **Orphan nodes**: ${t.orphanNodes.length}`)}}catch{}}}catch{x.push(``,`### Knowledge Graph`,`- Graph store unavailable`)}let S=l?.onboardDir??_(process.cwd(),y.aiContext),C=a(S),w=c?.onboardComplete||C;if(x.push(``,`### Onboard Status`,w?`- ✅ Complete${c?.onboardTimestamp?` (last: ${c.onboardTimestamp})`:``}`:'- ❌ Not run — call `onboard({ path: "." })` to analyze the codebase',`- **Onboard Directory**: \`${S}\``),s)try{let t=await Kc(s.list(),[],`curated.list`);if(t.timedOut)e.push(`⚠ Curated knowledge listing timed out`),x.push(``,`### Curated Knowledge`,`- Listing timed out`);else{let e=t.value;b=e.length,x.push(``,`### Curated Knowledge`,e.length>0?`- ${e.length} entries`:"- Empty — use `remember()` to persist decisions")}}catch{x.push(``,`### Curated Knowledge`,`- Unable to read curated entries`)}let T=Xc(h.totalRecords,v?.nodes??0,v?.edges??0,b);x.push(``),x.push(`## 📊 Context Pressure: ${T}/100`),T>=80?x.push(`⚠️ High pressure — consider pruning stale entries or compacting context.`):T>=50?x.push(`ℹ️ Moderate pressure — knowledge base is well-populated.`):x.push(`✅ Low pressure — plenty of headroom for more content.`);let E=0;if(h.lastIndexedAt){E=new Date(h.lastIndexedAt).getTime();let e=(Date.now()-E)/(1e3*60*60);x.push(``,`### Index Freshness`,e>24?u===`smart`?`- ⚠ Last indexed ${Math.floor(e)}h ago — smart indexing will refresh automatically`:`- ⚠ Last indexed ${Math.floor(e)}h ago — may be stale. Run \`reindex({})\``:`- ✅ Last indexed ${e<1?`less than 1h`:`${Math.floor(e)}h`} ago`)}if(u===`smart`)if(x.push(``,`### Smart Indexing`),f){let e=f();e?x.push(`- **Mode**: Smart (trickle)`,`- **Status**: ${e.running?`✅ Running`:`⏸ Stopped`}`,`- **Queue**: ${e.queueSize} files pending`,`- **Changed files**: ${e.changedFilesSize} detected`,`- **Interval**: ${Math.round(e.intervalMs/1e3)}s per batch of ${e.batchSize}`):x.push(`- **Mode**: Smart (trickle)`,`- **Status**: scheduler state unavailable (init may have failed)`)}else x.push(`- **Mode**: Smart (trickle) — scheduler state unavailable`);{try{let e=_(process.cwd(),y.data,`stash`);if(a(e)){let t=d(e).mtimeMs;t>E&&(E=t)}}catch{}let e=[];if(s)try{let t=b>0?await s.list():[];for(let e of t){let t=new Date(e.updated||e.created).getTime();t>E&&(E=t)}e.push(...t.sort((e,t)=>new Date(t.updated).getTime()-new Date(e.updated).getTime()).slice(0,5))}catch{}let t=E>0?Date.now()-E:0;if(t>=144e5){let n=Math.floor(t/36e5);if(x.push(``,`### 🌅 Session Briefing`,`_${n}+ hours since last activity — here's what to pick up:_`,``),e.length>0){x.push(`**Recent decisions/notes:**`);for(let t of e)x.push(`- **${t.title}** (${t.category??`note`}) — ${(t.contentPreview??``).slice(0,80)}…`)}x.push(``,`**Suggested next steps:**`,'- `search({ query: "SESSION CHECKPOINT", origin: "curated" })` — find your last checkpoint',"- `restore({})` — resume from a saved checkpoint","- `list()` — browse all stored knowledge")}}x.push(``,`### Runtime`,`- **Tree-sitter (WASM)**: ${re.get()?`✅ Available (AST analysis)`:`⚠ Unavailable (regex fallback)`}`);let D=Zc(),O=Qc(),k=r(),ee=D!=null&&D!==k,A=O!=null&&O!==k;if(ee||A){let e=i(),t=[];ee&&t.push(`user scaffold v${D}`),A&&t.push(`workspace scaffold v${O}`);let r=t.join(`, `);e.state===`success`?x.push(``,`### ✅ Upgrade Applied`,`- Server v${k} — ${r} auto-upgraded successfully.`,`- _Restart the MCP server to use the updated version._`):e.state===`pending`?x.push(``,`### ⏳ Upgrade In Progress`,`- Server v${k} ≠ ${r}`,`- Auto-upgrade is running in the background…`):e.state===`failed`?(n(),x.push(``,`### ⬆ Upgrade Available (auto-upgrade failed, retrying)`,`- Server v${k} ≠ ${r}`,`- Error: ${e.error??`unknown`}`)):(n(),x.push(``,`### ⬆ Upgrade Available`,`- Server v${k} ≠ ${r}`,`- Auto-upgrade triggered — check again shortly.`))}e.length>0&&x.push(``,`### ⚠ Warnings`,...e.map(e=>`- ${e}`));let j=Nr();if(j.length>0){let e=j.sort((e,t)=>t.callCount-e.callCount);x.push(``,`### Tool Usage This Session`,``),x.push(`| Tool | Calls | Tokens In | Tokens Out | Errors | Avg Latency |`),x.push(`|------|-------|-----------|------------|--------|-------------|`);for(let t of e.slice(0,15)){let e=Math.round(t.totalInputChars/4),n=Math.round(t.totalOutputChars/4),r=Math.round(t.totalDurationMs/t.callCount);x.push(`| ${t.tool} | ${t.callCount} | ${e.toLocaleString()} | ${n.toLocaleString()} | ${t.errorCount} | ${r}ms |`)}}let M=Er();if(M.bufferSize>=10){let e=M.state===`healthy`?`🟢`:M.state===`degraded`?`🔴`:`🟡`;x.push(``,`### Auto-GC: ${e} ${M.state}`),x.push(`- p95 latency: ${M.p95}ms | buffer: ${M.bufferSize} samples`),M.gcCount>0&&x.push(`- GC cycles triggered: ${M.gcCount}`)}let te=x.join(`
1260
- `),ne={totalRecords:h.totalRecords,totalFiles:h.totalFiles,lastIndexedAt:h.lastIndexedAt??null,onboarded:w,onboardDir:S,contentTypes:h.contentTypeBreakdown,wasmAvailable:!!re.get(),graphStats:v,curatedCount:b,serverVersion:k,scaffoldVersion:D??null,workspaceScaffoldVersion:O??null,upgradeAvailable:ee||A,contextPressure:T};return{content:[{type:`text`,text:te+(u===`smart`?"\n\n---\n_Next: Use `search` to query indexed content or `graph({action:'find_nodes', name_pattern:'<top-level-module>'})` → then `graph({action:'neighbors', node_id})` for relationships. Smart indexing handles updates automatically._":"\n\n---\n_Next: Use `search` to query indexed content, `graph({action:'find_nodes', name_pattern:'<top-level-module>'})` → then `graph({action:'neighbors', node_id})` for relationships, or `reindex` to refresh the index._")}],structuredContent:ne}}catch(e){return Gc.error(`Status failed`,A(e)),{content:[{type:`text`,text:`Status check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const tl=E(`tools`);function nl(e,t,n){let r=I(`update`);e.registerTool(`update`,{title:r.title,description:`Update an existing curated knowledge entry. Increments version and records the reason in the changelog.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/use-lancedb.md")`),content:F.string().min(10).max(1e5).describe(`New markdown content to replace existing content`),reason:F.string().min(3).max(1e3).describe(`Why this update is being made (recorded in changelog)`)},annotations:r.annotations},async({path:e,content:r,reason:i})=>{try{let a=await t.update(e,r,i);return n&&n.notifyAfterCuratedWrite(e).catch(()=>{}),{content:[{type:`text`,text:`Updated: \`.ai/curated/${a.path}\` → version ${a.version}\n\nReason: ${i}\n\n---\n_Next: Use \`read\` to verify the updated content, or \`search\` to test searchability._`}]}}catch(e){return tl.error(`Update failed`,A(e)),{content:[{type:`text`,text:`Update failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Z=E(`tools`);function rl(e){let t=I(`web_search`);e.registerTool(`web_search`,{title:t.title,description:`PREFERRED web search — search the web via DuckDuckGo (no API key). Pass one query or multiple for parallel searching. Returns structured results with title, URL, and snippet.`,inputSchema:{queries:F.array(F.string().max(2e3)).min(1).max(5).describe('Search queries (1–5). Single: `["react hooks"]`. Multiple searched in parallel.'),limit:F.number().min(1).max(20).default(5).describe(`Max results per query`),site:F.string().optional().describe(`Restrict to domain (e.g., "docs.aws.amazon.com")`)},annotations:t.annotations},async({queries:e,limit:t,site:n})=>{let r=e,i=async e=>{let r=await tn({query:e,limit:t,site:n}),i=[`## Search: ${r.query}`,``];if(r.results.length===0)i.push(`No results found.`);else for(let e of r.results)i.push(`### [${e.title}](${e.url})`,e.snippet,``);return i.join(`
1258
+ `),j=(c===`hybrid`?`hybrid (vector + keyword RRF)`:c===`keyword`?`keyword (FTS)`:`semantic (vector)`)+E,M=Bc(b,e),te=M.length>0?`\n_Distinctive terms: ${M.map(e=>`\`${e}\``).join(`, `)}_`:``,ne=await Hc(n,b),N=[];if(b.length===0)N.push("`reindex` — no results found, index may be stale"),N.push("`find` — try federated search with glob/regex");else{let e=b[0]?.record.sourcePath;e&&N.push(`\`lookup\` — see all chunks from \`${e}\``),N.push("`symbol` — resolve a specific symbol from the results"),N.push("`compact` — compress a result file for focused reading")}let re=[O?`${O}\n\n`:``,ne?`${ne}\n\n`:``,ee,`\n\n---\n_Search mode: ${j} | ${b.length} results_${te}`,`\n_Next: ${N.join(` | `)}_`],ie=C+re.join(``);g&&(ie=Xt(ie,g));let ae=new Set,oe=[];for(let e of b){if(e.record.origin!==`curated`||e.record.sourcePath.startsWith(`[`))continue;let t=Io(e.record.sourcePath);if(!t)continue;let n=Po(t,e.record.headingPath??t);n&&!ae.has(n.uri)&&(ae.add(n.uri),oe.push(n))}return{content:[{type:`text`,text:ie},...oe],structuredContent:{results:b.map(e=>({sourcePath:e.record.sourcePath,contentType:e.record.contentType,score:e.score,...e.record.headingPath?{headingPath:e.record.headingPath}:{},...e.record.startLine?{startLine:e.record.startLine}:{},...e.record.endLine?{endLine:e.record.endLine}:{},...e.record.origin===`indexed`?{}:{origin:e.record.origin},...e.record.category?{category:e.record.category}:{},...e.record.tags?.length?{tags:e.record.tags}:{}})),totalResults:b.length,searchMode:c,query:e}}}catch(e){return Fc.error(`Search failed`,A(e)),{content:[{type:`text`,text:`Search failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Wc=E(`tools`);function Gc(e,t){let n=L(`session_digest`);e.registerTool(`session_digest`,{title:n.title,description:`Compress the current session's tool trajectory into a focused digest. Aggregates replay log, stash state, and checkpoints. Supports deterministic and LLM-assisted (sampling) modes with configurable token budget.`,inputSchema:{scope:F.enum([`tools`,`stash`,`all`]).default(`all`).describe(`What to include: tools (replay only), stash (stash + checkpoints only), or all`),since:F.string().optional().describe(`ISO timestamp — only include activity after this time`),last:F.number().min(1).max(500).optional().describe(`Max replay entries to consider (default: 50)`),focus:F.string().optional().describe(`Focus query — prioritize entries matching this topic`),mode:F.enum([`deterministic`,`sampling`]).default(`deterministic`).describe(`Summary mode: deterministic (fast, template-based) or sampling (LLM-assisted, richer)`),token_budget:F.number().min(100).max(8e3).default(2e3).describe(`Target token budget for the digest output`),persist:F.boolean().default(!0).describe(`Auto-save digest to stash for crash recovery`)},annotations:n.annotations},async({scope:e,since:n,last:r,focus:i,mode:a,token_budget:o,persist:s})=>{try{let c={scope:e,since:n,last:r,focus:i,mode:a,tokenBudget:o,persist:s};return{content:[{type:`text`,text:(a===`sampling`&&t?.available?await Lt(c,async(e,n,r)=>(await t.createMessage({prompt:e,systemPrompt:n,maxTokens:r})).text):It(c)).digest}]}}catch(e){return Wc.error(`Session digest failed`,A(e)),{content:[{type:`text`,text:`Session digest failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Kc=E(`tools`);function qc(e,t,n,r=15e3){let i,a=new Promise(e=>{i=setTimeout(()=>{Kc.warn(`Status sub-operation "${n}" timed out after ${r}ms`),e({value:t,timedOut:!0})},r)});return Promise.race([e.then(e=>(clearTimeout(i),{value:e,timedOut:!1}),e=>(clearTimeout(i),Kc.warn(`Status sub-operation "${n}" failed: ${e instanceof Error?e.message:String(e)}`),{value:t,timedOut:!1})),a])}const Jc=5*6e4;let Yc=null,Xc=null;function Zc(e,t,n,r){let i=Math.min(e/2e4,1),a=Math.min((t+n)/5e4,1),o=Math.min(r/200,1);return Math.round(i*40+a*35+o*25)}function Qc(){let e=Date.now();if(Yc&&e-Yc.ts<Jc)return Yc.value;try{let t=_(N(),`.copilot`,`.aikit-scaffold.json`);if(!a(t))return Yc={value:null,ts:e},null;let n=JSON.parse(s(t,`utf-8`)).version??null;return Yc={value:n,ts:e},n}catch{return Yc={value:null,ts:e},null}}function $c(){let e=Date.now();if(Xc&&e-Xc.ts<Jc)return Xc.value;try{let t=_(process.cwd(),`.github`,`.aikit-scaffold.json`);if(!a(t))return Xc={value:null,ts:e},null;let n=JSON.parse(s(t,`utf-8`)).version??null;return Xc={value:n,ts:e},n}catch{return Xc={value:null,ts:e},null}}function el(e){let t=L(`status`);e.registerTool(`status`,{title:t.title,description:`Get the current status and statistics of the AI Kit index.`,outputSchema:ii,annotations:t.annotations},async()=>{let e=r(),t=Qc(),a=$c(),o=t!=null&&t!==e,s=a!=null&&a!==e,c=[`## AI Kit Status`,``,`⏳ **AI Kit is initializing** — index stats will be available shortly.`,``,`### Runtime`,`- **Tree-sitter (WASM)**: ${re.get()?`✅ Available (AST analysis)`:`⚠ Unavailable (regex fallback)`}`,``,`### Version`,`- **Server**: ${e}`,`- **Scaffold (user)**: ${t??`not installed`}`,`- **Scaffold (workspace)**: ${a??`not installed`}`];if(o||s){let r=i(),l=[];o&&l.push(`user scaffold v${t}`),s&&l.push(`workspace scaffold v${a}`);let u=l.join(`, `);r.state===`success`?c.push(``,`### ✅ Upgrade Applied`,`- Server v${e} — ${u} auto-upgraded successfully.`,`- _Restart the MCP server to use the updated version._`):r.state===`pending`?c.push(``,`### ⏳ Upgrade In Progress`,`- Server v${e} ≠ ${u}`,`- Auto-upgrade is running in the background…`):r.state===`failed`?(n(),c.push(``,`### ⬆ Upgrade Available (auto-upgrade failed, retrying)`,`- Server v${e} ≠ ${u}`,`- Error: ${r.error??`unknown`}`)):(n(),c.push(``,`### ⬆ Upgrade Available`,`- Server v${e} ≠ ${u}`,`- Auto-upgrade triggered — check again shortly.`))}let l={totalRecords:0,totalFiles:0,lastIndexedAt:null,onboarded:!1,onboardDir:``,contentTypes:{},wasmAvailable:!!re.get(),graphStats:null,curatedCount:0,serverVersion:e,scaffoldVersion:t??null,workspaceScaffoldVersion:a??null,upgradeAvailable:o||s,contextPressure:0};return{content:[{type:`text`,text:c.join(`
1259
+ `)}],structuredContent:l}})}function tl(e,t,o,s,c,l,u,f){let p=L(`status`);e.registerTool(`status`,{title:p.title,description:`Get the current status and statistics of the AI Kit index.`,outputSchema:ii,annotations:p.annotations},async()=>{let e=[];try{let[p,m]=await Promise.all([qc(t.getStats(),{totalRecords:0,totalFiles:0,lastIndexedAt:null,contentTypeBreakdown:{}},`store.getStats`),qc(t.listSourcePaths(),[],`store.listSourcePaths`)]),h=p.value;p.timedOut&&e.push(`⚠ Index stats timed out — values may be incomplete`);let g=m.value;m.timedOut&&e.push(`⚠ File listing timed out`);let v=null,b=0,x=[`## AI Kit Status`,``,`- **Total Records**: ${h.totalRecords}`,`- **Total Files**: ${h.totalFiles}`,`- **Last Indexed**: ${h.lastIndexedAt??`Never`}`,``,`### Content Types`,...Object.entries(h.contentTypeBreakdown).map(([e,t])=>`- ${e}: ${t}`),``,`### Indexed Files`,...g.slice(0,50).map(e=>`- ${e}`),g.length>50?`\n... and ${g.length-50} more files`:``];if(o)try{let t=await qc(o.getStats(),{nodeCount:0,edgeCount:0,nodeTypes:{},edgeTypes:{}},`graphStore.getStats`);if(t.timedOut)e.push(`⚠ Graph stats timed out`),x.push(``,`### Knowledge Graph`,`- Graph stats timed out`);else{let e=t.value;v={nodes:e.nodeCount,edges:e.edgeCount},x.push(``,`### Knowledge Graph`,`- **Nodes**: ${e.nodeCount}`,`- **Edges**: ${e.edgeCount}`,...Object.entries(e.nodeTypes).map(([e,t])=>` - ${e}: ${t}`));try{let e=await qc(o.validate(),{valid:!0,danglingEdges:[],orphanNodes:[],stats:{nodeCount:0,edgeCount:0,nodeTypes:{},edgeTypes:{}}},`graphStore.validate`);if(!e.timedOut){let t=e.value;t.valid||x.push(`- **⚠ Integrity Issues**: ${t.danglingEdges.length} dangling edges`),t.orphanNodes.length>0&&x.push(`- **Orphan nodes**: ${t.orphanNodes.length}`)}}catch{}}}catch{x.push(``,`### Knowledge Graph`,`- Graph store unavailable`)}let S=l?.onboardDir??_(process.cwd(),y.aiContext),C=a(S),w=c?.onboardComplete||C;if(x.push(``,`### Onboard Status`,w?`- ✅ Complete${c?.onboardTimestamp?` (last: ${c.onboardTimestamp})`:``}`:'- ❌ Not run — call `onboard({ path: "." })` to analyze the codebase',`- **Onboard Directory**: \`${S}\``),s)try{let t=await qc(s.list(),[],`curated.list`);if(t.timedOut)e.push(`⚠ Curated knowledge listing timed out`),x.push(``,`### Curated Knowledge`,`- Listing timed out`);else{let e=t.value;b=e.length,x.push(``,`### Curated Knowledge`,e.length>0?`- ${e.length} entries`:"- Empty — use `remember()` to persist decisions")}}catch{x.push(``,`### Curated Knowledge`,`- Unable to read curated entries`)}let T=Zc(h.totalRecords,v?.nodes??0,v?.edges??0,b);x.push(``),x.push(`## 📊 Context Pressure: ${T}/100`),T>=80?x.push(`⚠️ High pressure — consider pruning stale entries or compacting context.`):T>=50?x.push(`ℹ️ Moderate pressure — knowledge base is well-populated.`):x.push(`✅ Low pressure — plenty of headroom for more content.`);let E=0;if(h.lastIndexedAt){E=new Date(h.lastIndexedAt).getTime();let e=(Date.now()-E)/(1e3*60*60);x.push(``,`### Index Freshness`,e>24?u===`smart`?`- ⚠ Last indexed ${Math.floor(e)}h ago — smart indexing will refresh automatically`:`- ⚠ Last indexed ${Math.floor(e)}h ago — may be stale. Run \`reindex({})\``:`- ✅ Last indexed ${e<1?`less than 1h`:`${Math.floor(e)}h`} ago`)}if(u===`smart`)if(x.push(``,`### Smart Indexing`),f){let e=f();e?x.push(`- **Mode**: Smart (trickle)`,`- **Status**: ${e.running?`✅ Running`:`⏸ Stopped`}`,`- **Queue**: ${e.queueSize} files pending`,`- **Changed files**: ${e.changedFilesSize} detected`,`- **Interval**: ${Math.round(e.intervalMs/1e3)}s per batch of ${e.batchSize}`):x.push(`- **Mode**: Smart (trickle)`,`- **Status**: scheduler state unavailable (init may have failed)`)}else x.push(`- **Mode**: Smart (trickle) — scheduler state unavailable`);{try{let e=_(process.cwd(),y.data,`stash`);if(a(e)){let t=d(e).mtimeMs;t>E&&(E=t)}}catch{}let e=[];if(s)try{let t=b>0?await s.list():[];for(let e of t){let t=new Date(e.updated||e.created).getTime();t>E&&(E=t)}e.push(...t.sort((e,t)=>new Date(t.updated).getTime()-new Date(e.updated).getTime()).slice(0,5))}catch{}let t=E>0?Date.now()-E:0;if(t>=144e5){let n=Math.floor(t/36e5);if(x.push(``,`### 🌅 Session Briefing`,`_${n}+ hours since last activity — here's what to pick up:_`,``),e.length>0){x.push(`**Recent decisions/notes:**`);for(let t of e)x.push(`- **${t.title}** (${t.category??`note`}) — ${(t.contentPreview??``).slice(0,80)}…`)}x.push(``,`**Suggested next steps:**`,'- `search({ query: "SESSION CHECKPOINT", origin: "curated" })` — find your last checkpoint',"- `restore({})` — resume from a saved checkpoint","- `list()` — browse all stored knowledge")}}x.push(``,`### Runtime`,`- **Tree-sitter (WASM)**: ${re.get()?`✅ Available (AST analysis)`:`⚠ Unavailable (regex fallback)`}`);let D=Qc(),O=$c(),k=r(),ee=D!=null&&D!==k,A=O!=null&&O!==k;if(ee||A){let e=i(),t=[];ee&&t.push(`user scaffold v${D}`),A&&t.push(`workspace scaffold v${O}`);let r=t.join(`, `);e.state===`success`?x.push(``,`### ✅ Upgrade Applied`,`- Server v${k} — ${r} auto-upgraded successfully.`,`- _Restart the MCP server to use the updated version._`):e.state===`pending`?x.push(``,`### ⏳ Upgrade In Progress`,`- Server v${k} ≠ ${r}`,`- Auto-upgrade is running in the background…`):e.state===`failed`?(n(),x.push(``,`### ⬆ Upgrade Available (auto-upgrade failed, retrying)`,`- Server v${k} ≠ ${r}`,`- Error: ${e.error??`unknown`}`)):(n(),x.push(``,`### ⬆ Upgrade Available`,`- Server v${k} ≠ ${r}`,`- Auto-upgrade triggered — check again shortly.`))}e.length>0&&x.push(``,`### ⚠ Warnings`,...e.map(e=>`- ${e}`));let j=Mr();if(j.length>0){let e=j.sort((e,t)=>t.callCount-e.callCount);x.push(``,`### Tool Usage This Session`,``),x.push(`| Tool | Calls | Tokens In | Tokens Out | Errors | Avg Latency |`),x.push(`|------|-------|-----------|------------|--------|-------------|`);for(let t of e.slice(0,15)){let e=Math.round(t.totalInputChars/4),n=Math.round(t.totalOutputChars/4),r=Math.round(t.totalDurationMs/t.callCount);x.push(`| ${t.tool} | ${t.callCount} | ${e.toLocaleString()} | ${n.toLocaleString()} | ${t.errorCount} | ${r}ms |`)}}let M=Tr();if(M.bufferSize>=10){let e=M.state===`healthy`?`🟢`:M.state===`degraded`?`🔴`:`🟡`;x.push(``,`### Auto-GC: ${e} ${M.state}`),x.push(`- p95 latency: ${M.p95}ms | buffer: ${M.bufferSize} samples`),M.gcCount>0&&x.push(`- GC cycles triggered: ${M.gcCount}`)}let te=x.join(`
1260
+ `),ne={totalRecords:h.totalRecords,totalFiles:h.totalFiles,lastIndexedAt:h.lastIndexedAt??null,onboarded:w,onboardDir:S,contentTypes:h.contentTypeBreakdown,wasmAvailable:!!re.get(),graphStats:v,curatedCount:b,serverVersion:k,scaffoldVersion:D??null,workspaceScaffoldVersion:O??null,upgradeAvailable:ee||A,contextPressure:T};return{content:[{type:`text`,text:te+(u===`smart`?"\n\n---\n_Next: Use `search` to query indexed content or `graph({action:'find_nodes', name_pattern:'<top-level-module>'})` → then `graph({action:'neighbors', node_id})` for relationships. Smart indexing handles updates automatically._":"\n\n---\n_Next: Use `search` to query indexed content, `graph({action:'find_nodes', name_pattern:'<top-level-module>'})` → then `graph({action:'neighbors', node_id})` for relationships, or `reindex` to refresh the index._")}],structuredContent:ne}}catch(e){return Kc.error(`Status failed`,A(e)),{content:[{type:`text`,text:`Status check failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const nl=E(`tools`);function rl(e,t,n){let r=L(`update`);e.registerTool(`update`,{title:r.title,description:`Update an existing curated knowledge entry. Increments version and records the reason in the changelog.`,inputSchema:{path:F.string().describe(`Relative path within .ai/curated/ (e.g., "decisions/use-lancedb.md")`),content:F.string().min(10).max(1e5).describe(`New markdown content to replace existing content`),reason:F.string().min(3).max(1e3).describe(`Why this update is being made (recorded in changelog)`)},annotations:r.annotations},async({path:e,content:r,reason:i})=>{try{let a=await t.update(e,r,i);return n&&n.notifyAfterCuratedWrite(e).catch(()=>{}),{content:[{type:`text`,text:`Updated: \`.ai/curated/${a.path}\` → version ${a.version}\n\nReason: ${i}\n\n---\n_Next: Use \`read\` to verify the updated content, or \`search\` to test searchability._`}]}}catch(e){return nl.error(`Update failed`,A(e)),{content:[{type:`text`,text:`Update failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Z=E(`tools`);function il(e){let t=L(`web_search`);e.registerTool(`web_search`,{title:t.title,description:`PREFERRED web search — search the web via DuckDuckGo (no API key). Pass one query or multiple for parallel searching. Returns structured results with title, URL, and snippet.`,inputSchema:{queries:F.array(F.string().max(2e3)).min(1).max(5).describe('Search queries (1–5). Single: `["react hooks"]`. Multiple searched in parallel.'),limit:F.number().min(1).max(20).default(5).describe(`Max results per query`),site:F.string().optional().describe(`Restrict to domain (e.g., "docs.aws.amazon.com")`)},annotations:t.annotations},async({queries:e,limit:t,site:n})=>{let r=e,i=async e=>{let r=await tn({query:e,limit:t,site:n}),i=[`## Search: ${r.query}`,``];if(r.results.length===0)i.push(`No results found.`);else for(let e of r.results)i.push(`### [${e.title}](${e.url})`,e.snippet,``);return i.join(`
1261
1261
  `)};if(r.length===1)try{return{content:[{type:`text`,text:`${await i(r[0])}\n---\n_Next: Use \`web_fetch\` to read any of these pages in full._`}]}}catch(e){return Z.error(`Web search failed`,A(e)),{content:[{type:`text`,text:`Web search failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}let a=await Promise.allSettled(r.map(e=>i(e))),o=[],s=0;for(let e=0;e<a.length;e++){let t=a[e];if(t.status===`fulfilled`)o.push(t.value);else{s++;let n=t.reason instanceof Error?t.reason.message:String(t.reason);Z.error(`Web search failed`,{query:r[e],error:n}),o.push(`## ❌ Search failed: ${r[e]}\n\n${n}`)}}let c=`_Searched ${a.length-s}/${a.length} queries successfully._`;return o.push(``,`---`,c,"_Next: Use `web_fetch` to read any of these pages in full._"),{content:[{type:`text`,text:o.join(`
1262
1262
 
1263
- `)}],...s===a.length?{isError:!0}:{}}})}function il(e){let t=I(`http`);e.registerTool(`http`,{title:t.title,description:`Make HTTP requests (GET/POST/PUT/PATCH/DELETE/HEAD) for API testing. Returns status, headers, and formatted body with timing info.`,inputSchema:{url:F.string().url().describe(`Request URL (http/https only)`),method:F.enum([`GET`,`POST`,`PUT`,`PATCH`,`DELETE`,`HEAD`]).default(`GET`).describe(`HTTP method`),headers:F.record(F.string(),F.string()).optional().describe(`Request headers as key-value pairs`),body:F.string().optional().describe(`Request body (for POST/PUT/PATCH)`),timeout:F.number().min(1e3).max(6e4).default(15e3).describe(`Timeout in milliseconds`)},annotations:t.annotations},async({url:e,method:t,headers:n,body:r,timeout:i})=>{try{let a=await $e({url:e,method:t,headers:n,body:r,timeout:i}),o=[`## ${t} ${e}`,``,`**Status:** ${a.status} ${a.statusText}`,`**Time:** ${a.durationMs}ms`,`**Size:** ${a.sizeBytes} bytes`,`**Content-Type:** ${a.contentType}`,``,`### Headers`,"```json",JSON.stringify(a.headers),"```",``,`### Body`,a.contentType.includes(`json`)?"```json":"```",a.body,"```"];return a.truncated&&o.push(``,`_Response truncated — total size: ${a.sizeBytes} bytes_`),{content:[{type:`text`,text:o.join(`
1264
- `)}]}}catch(e){return Z.error(`HTTP request failed`,A(e)),{content:[{type:`text`,text:`HTTP request failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function al(e){let t=I(`regex_test`);e.registerTool(`regex_test`,{title:t.title,description:`Test a regex pattern against sample strings. Supports match, replace, and split modes.`,inputSchema:{pattern:F.string().max(500).describe(`Regex pattern (without delimiters)`),flags:F.string().max(10).regex(/^[gimsuy]*$/).default(``).describe(`Regex flags (g, i, m, s, etc.)`),test_strings:F.array(F.string().max(1e4)).max(50).describe(`Strings to test the pattern against`),mode:F.enum([`match`,`replace`,`split`]).default(`match`).describe(`Test mode`),replacement:F.string().optional().describe(`Replacement string (for replace mode)`)},annotations:t.annotations},async({pattern:e,flags:t,test_strings:n,mode:r,replacement:i})=>{let a=Tt({pattern:e,flags:t,testStrings:n,mode:r,replacement:i});if(!a.valid)return{content:[{type:`text`,text:`Invalid regex: ${a.error}`}],isError:!0};let o=[`## Regex: \`/${a.pattern}/${a.flags}\``,``,`Mode: ${r}`,``];for(let e of a.results){if(o.push(`**Input:** \`${e.input}\``),o.push(`**Matched:** ${e.matched}`),e.matches)for(let t of e.matches){let e=t.groups.length>0?` groups: [${t.groups.join(`, `)}]`:``;o.push(` - "${t.full}" at index ${t.index}${e}`)}e.replaced!==void 0&&o.push(`**Result:** \`${e.replaced}\``),e.split&&o.push(`**Split:** ${JSON.stringify(e.split)}`),o.push(``)}return{content:[{type:`text`,text:o.join(`
1265
- `)}]}})}function ol(e){let t=I(`encode`);e.registerTool(`encode`,{title:t.title,description:`Encode, decode, or hash text. Supports base64, URL encoding, SHA-256, MD5, JWT decode, hex.`,inputSchema:{operation:F.enum([`base64_encode`,`base64_decode`,`url_encode`,`url_decode`,`sha256`,`md5`,`jwt_decode`,`hex_encode`,`hex_decode`]).describe(`Operation to perform`),input:F.string().max(1e6).describe(`Input text`)},annotations:t.annotations},async({operation:e,input:t})=>{try{let n=Le({operation:e,input:t});return{content:[{type:`text`,text:`## ${e}\n\n**Input:** \`${t.length>100?`${t.slice(0,100)}...`:t}\`\n**Output:**\n\`\`\`\n${n.output}\n\`\`\``}]}}catch(e){return Z.error(`Encode failed`,A(e)),{content:[{type:`text`,text:`Encode failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function sl(e){let t=I(`measure`);e.registerTool(`measure`,{title:t.title,description:`Measure code complexity, line counts, and function counts for a file or directory. Returns per-file metrics sorted by complexity.`,outputSchema:ci,inputSchema:{path:F.string().describe(`File or directory path to measure`),extensions:F.array(F.string()).optional().describe(`File extensions to include (default: .ts,.tsx,.js,.jsx)`)},annotations:t.annotations},async({path:e,extensions:t})=>{try{let n=await ct({path:e,extensions:t}),r=[`## Code Metrics`,``,`**Files:** ${n.summary.totalFiles}`,`**Total lines:** ${n.summary.totalLines} (${n.summary.totalCodeLines} code)`,`**Functions:** ${n.summary.totalFunctions}`,`**Avg complexity:** ${n.summary.avgComplexity}`,`**Max complexity:** ${n.summary.maxComplexity.value} (${n.summary.maxComplexity.file})`,``,`### Top files by complexity`,``,`| File | Lines | Code | Complexity | Cognitive | Functions | Imports |`,`|------|-------|------|------------|-----------|-----------|---------|`];for(let e of n.files.slice(0,20)){let t=e.cognitiveComplexity===void 0?`—`:String(e.cognitiveComplexity);r.push(`| ${e.path} | ${e.lines.total} | ${e.lines.code} | ${e.complexity} | ${t} | ${e.functions} | ${e.imports} |`)}n.files.length>20&&r.push(``,`_...and ${n.files.length-20} more files_`);let i={summary:{totalFiles:n.summary.totalFiles,totalLines:n.summary.totalLines,totalCodeLines:n.summary.totalCodeLines,totalFunctions:n.summary.totalFunctions,avgComplexity:n.summary.avgComplexity,maxComplexity:{value:n.summary.maxComplexity.value,file:n.summary.maxComplexity.file}},files:n.files.map(e=>({path:e.path,lines:e.lines.total,code:e.lines.code,complexity:e.complexity,functions:e.functions}))};return{content:[{type:`text`,text:r.join(`
1266
- `)}],structuredContent:i}}catch(e){return Z.error(`Measure failed`,A(e)),{content:[{type:`text`,text:`Measure failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function cl(e){let t=I(`changelog`);e.registerTool(`changelog`,{title:t.title,description:`Generate a changelog from git history between two refs. Groups by conventional commit type.`,inputSchema:{from:F.string().max(200).describe(`Start ref (tag, SHA, HEAD~N)`),to:F.string().max(200).default(`HEAD`).describe(`End ref (default: HEAD)`),format:F.enum([`grouped`,`chronological`,`per-scope`]).default(`grouped`).describe(`Output format`),include_breaking:F.boolean().default(!0).describe(`Highlight breaking changes`),cwd:F.string().optional().describe(`Repository root or working directory`)},annotations:t.annotations},async({from:e,to:t,format:n,include_breaking:r,cwd:i})=>{try{let a=xe({from:e,to:t,format:n,includeBreaking:r,cwd:i}),o=`${a.stats.total} commits (${Object.entries(a.stats.types).map(([e,t])=>`${t} ${e}`).join(`, `)})`;return{content:[{type:`text`,text:`${a.markdown}\n---\n_${o}_`}]}}catch(e){return Z.error(`Changelog failed`,A(e)),{content:[{type:`text`,text:`Changelog failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ll(e){let t=I(`schema_validate`);e.registerTool(`schema_validate`,{title:t.title,description:`Validate JSON data against a JSON Schema. Supports type, required, properties, items, enum, pattern, min/max.`,inputSchema:{data:F.string().max(5e5).describe(`JSON data to validate (as string)`),schema:F.string().max(5e5).describe(`JSON Schema to validate against (as string)`)},annotations:t.annotations},async({data:e,schema:t})=>{try{let n=Pt({data:JSON.parse(e),schema:JSON.parse(t)});if(n.valid)return{content:[{type:`text`,text:`## Validation: PASSED
1263
+ `)}],...s===a.length?{isError:!0}:{}}})}function al(e){let t=L(`http`);e.registerTool(`http`,{title:t.title,description:`Make HTTP requests (GET/POST/PUT/PATCH/DELETE/HEAD) for API testing. Returns status, headers, and formatted body with timing info.`,inputSchema:{url:F.string().url().describe(`Request URL (http/https only)`),method:F.enum([`GET`,`POST`,`PUT`,`PATCH`,`DELETE`,`HEAD`]).default(`GET`).describe(`HTTP method`),headers:F.record(F.string(),F.string()).optional().describe(`Request headers as key-value pairs`),body:F.string().optional().describe(`Request body (for POST/PUT/PATCH)`),timeout:F.number().min(1e3).max(6e4).default(15e3).describe(`Timeout in milliseconds`)},annotations:t.annotations},async({url:e,method:t,headers:n,body:r,timeout:i})=>{try{let a=await $e({url:e,method:t,headers:n,body:r,timeout:i}),o=[`## ${t} ${e}`,``,`**Status:** ${a.status} ${a.statusText}`,`**Time:** ${a.durationMs}ms`,`**Size:** ${a.sizeBytes} bytes`,`**Content-Type:** ${a.contentType}`,``,`### Headers`,"```json",JSON.stringify(a.headers),"```",``,`### Body`,a.contentType.includes(`json`)?"```json":"```",a.body,"```"];return a.truncated&&o.push(``,`_Response truncated — total size: ${a.sizeBytes} bytes_`),{content:[{type:`text`,text:o.join(`
1264
+ `)}]}}catch(e){return Z.error(`HTTP request failed`,A(e)),{content:[{type:`text`,text:`HTTP request failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ol(e){let t=L(`regex_test`);e.registerTool(`regex_test`,{title:t.title,description:`Test a regex pattern against sample strings. Supports match, replace, and split modes.`,inputSchema:{pattern:F.string().max(500).describe(`Regex pattern (without delimiters)`),flags:F.string().max(10).regex(/^[gimsuy]*$/).default(``).describe(`Regex flags (g, i, m, s, etc.)`),test_strings:F.array(F.string().max(1e4)).max(50).describe(`Strings to test the pattern against`),mode:F.enum([`match`,`replace`,`split`]).default(`match`).describe(`Test mode`),replacement:F.string().optional().describe(`Replacement string (for replace mode)`)},annotations:t.annotations},async({pattern:e,flags:t,test_strings:n,mode:r,replacement:i})=>{let a=Tt({pattern:e,flags:t,testStrings:n,mode:r,replacement:i});if(!a.valid)return{content:[{type:`text`,text:`Invalid regex: ${a.error}`}],isError:!0};let o=[`## Regex: \`/${a.pattern}/${a.flags}\``,``,`Mode: ${r}`,``];for(let e of a.results){if(o.push(`**Input:** \`${e.input}\``),o.push(`**Matched:** ${e.matched}`),e.matches)for(let t of e.matches){let e=t.groups.length>0?` groups: [${t.groups.join(`, `)}]`:``;o.push(` - "${t.full}" at index ${t.index}${e}`)}e.replaced!==void 0&&o.push(`**Result:** \`${e.replaced}\``),e.split&&o.push(`**Split:** ${JSON.stringify(e.split)}`),o.push(``)}return{content:[{type:`text`,text:o.join(`
1265
+ `)}]}})}function sl(e){let t=L(`encode`);e.registerTool(`encode`,{title:t.title,description:`Encode, decode, or hash text. Supports base64, URL encoding, SHA-256, MD5, JWT decode, hex.`,inputSchema:{operation:F.enum([`base64_encode`,`base64_decode`,`url_encode`,`url_decode`,`sha256`,`md5`,`jwt_decode`,`hex_encode`,`hex_decode`]).describe(`Operation to perform`),input:F.string().max(1e6).describe(`Input text`)},annotations:t.annotations},async({operation:e,input:t})=>{try{let n=Le({operation:e,input:t});return{content:[{type:`text`,text:`## ${e}\n\n**Input:** \`${t.length>100?`${t.slice(0,100)}...`:t}\`\n**Output:**\n\`\`\`\n${n.output}\n\`\`\``}]}}catch(e){return Z.error(`Encode failed`,A(e)),{content:[{type:`text`,text:`Encode failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function cl(e){let t=L(`measure`);e.registerTool(`measure`,{title:t.title,description:`Measure code complexity, line counts, and function counts for a file or directory. Returns per-file metrics sorted by complexity.`,outputSchema:si,inputSchema:{path:F.string().describe(`File or directory path to measure`),extensions:F.array(F.string()).optional().describe(`File extensions to include (default: .ts,.tsx,.js,.jsx)`)},annotations:t.annotations},async({path:e,extensions:t})=>{try{let n=await ct({path:e,extensions:t}),r=[`## Code Metrics`,``,`**Files:** ${n.summary.totalFiles}`,`**Total lines:** ${n.summary.totalLines} (${n.summary.totalCodeLines} code)`,`**Functions:** ${n.summary.totalFunctions}`,`**Avg complexity:** ${n.summary.avgComplexity}`,`**Max complexity:** ${n.summary.maxComplexity.value} (${n.summary.maxComplexity.file})`,``,`### Top files by complexity`,``,`| File | Lines | Code | Complexity | Cognitive | Functions | Imports |`,`|------|-------|------|------------|-----------|-----------|---------|`];for(let e of n.files.slice(0,20)){let t=e.cognitiveComplexity===void 0?`—`:String(e.cognitiveComplexity);r.push(`| ${e.path} | ${e.lines.total} | ${e.lines.code} | ${e.complexity} | ${t} | ${e.functions} | ${e.imports} |`)}n.files.length>20&&r.push(``,`_...and ${n.files.length-20} more files_`);let i={summary:{totalFiles:n.summary.totalFiles,totalLines:n.summary.totalLines,totalCodeLines:n.summary.totalCodeLines,totalFunctions:n.summary.totalFunctions,avgComplexity:n.summary.avgComplexity,maxComplexity:{value:n.summary.maxComplexity.value,file:n.summary.maxComplexity.file}},files:n.files.map(e=>({path:e.path,lines:e.lines.total,code:e.lines.code,complexity:e.complexity,functions:e.functions}))};return{content:[{type:`text`,text:r.join(`
1266
+ `)}],structuredContent:i}}catch(e){return Z.error(`Measure failed`,A(e)),{content:[{type:`text`,text:`Measure failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ll(e){let t=L(`changelog`);e.registerTool(`changelog`,{title:t.title,description:`Generate a changelog from git history between two refs. Groups by conventional commit type.`,inputSchema:{from:F.string().max(200).describe(`Start ref (tag, SHA, HEAD~N)`),to:F.string().max(200).default(`HEAD`).describe(`End ref (default: HEAD)`),format:F.enum([`grouped`,`chronological`,`per-scope`]).default(`grouped`).describe(`Output format`),include_breaking:F.boolean().default(!0).describe(`Highlight breaking changes`),cwd:F.string().optional().describe(`Repository root or working directory`)},annotations:t.annotations},async({from:e,to:t,format:n,include_breaking:r,cwd:i})=>{try{let a=xe({from:e,to:t,format:n,includeBreaking:r,cwd:i}),o=`${a.stats.total} commits (${Object.entries(a.stats.types).map(([e,t])=>`${t} ${e}`).join(`, `)})`;return{content:[{type:`text`,text:`${a.markdown}\n---\n_${o}_`}]}}catch(e){return Z.error(`Changelog failed`,A(e)),{content:[{type:`text`,text:`Changelog failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ul(e){let t=L(`schema_validate`);e.registerTool(`schema_validate`,{title:t.title,description:`Validate JSON data against a JSON Schema. Supports type, required, properties, items, enum, pattern, min/max.`,inputSchema:{data:F.string().max(5e5).describe(`JSON data to validate (as string)`),schema:F.string().max(5e5).describe(`JSON Schema to validate against (as string)`)},annotations:t.annotations},async({data:e,schema:t})=>{try{let n=Pt({data:JSON.parse(e),schema:JSON.parse(t)});if(n.valid)return{content:[{type:`text`,text:`## Validation: PASSED
1267
1267
 
1268
1268
  Data matches the schema.`}]};let r=[`## Validation: FAILED`,``,`**${n.errors.length} error(s):**`,``];for(let e of n.errors){let t=e.expected?` (expected: ${e.expected}, got: ${e.received})`:``;r.push(`- \`${e.path}\`: ${e.message}${t}`)}return{content:[{type:`text`,text:r.join(`
1269
- `)}]}}catch(e){return Z.error(`Schema validation failed`,A(e)),{content:[{type:`text`,text:`Schema validation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function ul(e){let t=I(`snippet`);e.registerTool(`snippet`,{title:t.title,description:`Save, retrieve, search, and manage persistent code snippets/templates.`,inputSchema:{action:F.enum([`save`,`get`,`list`,`search`,`delete`]).describe(`Operation to perform`),name:F.string().optional().describe(`Snippet name (required for save/get/delete)`),language:F.string().optional().describe(`Language tag (for save)`),code:F.string().max(1e5).optional().describe(`Code content (for save)`),tags:F.array(F.string()).optional().describe(`Tags for categorization (for save)`),query:F.string().optional().describe(`Search query (for search)`)},annotations:t.annotations},async({action:e,name:t,language:n,code:r,tags:i,query:a})=>{try{let o=Rt({action:e,name:t,language:n,code:r,tags:i,query:a});if(`deleted`in o)return{content:[{type:`text`,text:o.deleted?`Snippet "${t}" deleted.`:`Snippet "${t}" not found.`}]};if(`snippets`in o){if(o.snippets.length===0)return{content:[{type:`text`,text:`No snippets found.`}]};let e=[`## Snippets`,``];for(let t of o.snippets){let n=t.tags.length>0?` [${t.tags.join(`, `)}]`:``;e.push(`- **${t.name}** (${t.language})${n}`)}return{content:[{type:`text`,text:e.join(`
1270
- `)}]}}let s=o,c=s.tags.length>0?`\nTags: ${s.tags.join(`, `)}`:``;return{content:[{type:`text`,text:`## ${s.name} (${s.language})${c}\n\n\`\`\`${s.language}\n${s.code}\n\`\`\``}]}}catch(e){return Z.error(`Snippet failed`,A(e)),{content:[{type:`text`,text:`Snippet failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function dl(e){let t=I(`env`);e.registerTool(`env`,{title:t.title,description:`Get system and runtime environment info. Sensitive env vars are redacted by default.`,outputSchema:li,inputSchema:{include_env:F.boolean().default(!1).describe(`Include environment variables`),filter_env:F.string().optional().describe(`Filter env vars by name substring`),show_sensitive:F.boolean().default(!1).describe(`Show sensitive values (keys, tokens, etc.) — redacted by default`)},annotations:t.annotations},async({include_env:e,filter_env:t,show_sensitive:n})=>{let r=Re({includeEnv:e,filterEnv:t,showSensitive:n}),i=[`## Environment`,``,`**Platform:** ${r.system.platform} ${r.system.arch}`,`**OS:** ${r.system.type} ${r.system.release}`,`**Host:** ${r.system.hostname}`,`**CPUs:** ${r.system.cpus}`,`**Memory:** ${r.system.memoryFreeGb}GB free / ${r.system.memoryTotalGb}GB total`,``,`**Node:** ${r.runtime.node}`,`**V8:** ${r.runtime.v8}`,`**CWD:** ${r.cwd}`];if(r.env){i.push(``,`### Environment Variables`,``);for(let[e,t]of Object.entries(r.env))i.push(`- \`${e}\`: ${t}`)}let a={platform:r.system.platform,arch:r.system.arch,nodeVersion:r.runtime.node,cwd:r.cwd,cpus:r.system.cpus,memoryFreeGb:r.system.memoryFreeGb,memoryTotalGb:r.system.memoryTotalGb};return{content:[{type:`text`,text:i.join(`
1271
- `)}],structuredContent:a}})}function fl(e){let t=I(`time`);e.registerTool(`time`,{title:t.title,description:`Parse dates, convert timezones, calculate durations, add time. Supports ISO 8601, unix timestamps, and human-readable formats.`,outputSchema:ui,inputSchema:{operation:F.enum([`now`,`parse`,`convert`,`diff`,`add`]).describe(`now: current time | parse: parse a date string | convert: timezone conversion | diff: duration between two dates | add: add duration to date`),input:F.string().optional().describe(`Date input (ISO, unix timestamp, or parseable string). For diff: two comma-separated dates`),timezone:F.string().optional().describe(`Target timezone (e.g., "America/New_York", "Asia/Tokyo")`),duration:F.string().optional().describe(`Duration to add (e.g., "2h30m", "1d", "30s") — for add operation`)},annotations:t.annotations},async({operation:e,input:t,timezone:n,duration:r})=>{try{let i=Jt({operation:e,input:t,timezone:n,duration:r}),a=[`**${i.output}**`,``,`ISO: ${i.iso}`,`Unix: ${i.unix}`];i.details&&a.push(``,"```json",JSON.stringify(i.details),"```");let o={iso:i.iso,unix:i.unix,timezone:n??Intl.DateTimeFormat().resolvedOptions().timeZone,formatted:i.output};return{content:[{type:`text`,text:a.join(`
1272
- `)}],structuredContent:o}}catch(e){return Z.error(`Time failed`,A(e)),{content:[{type:`text`,text:`Time failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Q=E(`server`);function pl(e){let t=e.toLowerCase();return[`protobuf`,`invalid model`,`invalid onnx`,`unexpected end`,`unexpected token`,`failed to load`,`failed to initialize embedding`,`checksum`,`corrupt`,`malformed`,`could not load`,`onnx`,`database disk image is malformed`,`file is not a database`,`lance`].some(e=>t.includes(e))}async function ml(e,t){let n=t.toLowerCase(),r;try{({rm:r}=await import(`node:fs/promises`))}catch{return}if(n.includes(`embedding`)||n.includes(`onnx`)||n.includes(`protobuf`)||n.includes(`model`)){let t=e.embedding?.model??x.model,n=g(N(),`.cache`,`huggingface`,`transformers-js`,t);try{await r(n,{recursive:!0,force:!0}),Q.info(`Auto-heal: cleared embedding model cache`,{path:n})}catch{}}if(n.includes(`lance`)||n.includes(`database`)||n.includes(`store`)){let t=g(e.store.path,`lance`);try{await r(t,{recursive:!0,force:!0}),Q.info(`Auto-heal: cleared LanceDB store`,{path:t})}catch{}}if(n.includes(`sqlite`)||n.includes(`database disk image`)||n.includes(`graph`)){let t=g(e.store.path,`graph.db`);try{await r(t,{force:!0}),Q.info(`Auto-heal: cleared graph database`,{path:t})}catch{}}}async function hl(n){Q.info(`Initializing AI Kit components`);let[r,i,o,s]=await Promise.all([(async()=>{let e=new ae({model:n.embedding.model,dimensions:n.embedding.dimensions});return await e.initialize(),Q.info(`Embedder loaded`,{modelId:e.modelId,dimensions:e.dimensions}),e})(),(async()=>{let e=await he({backend:n.store.backend,path:n.store.path});return await e.initialize(),Q.info(`Store initialized`),e})(),(async()=>{let e=new me({path:n.store.path});return await e.initialize(),Q.info(`Graph store initialized`),e})(),(async()=>{let e=await ie();return e?Q.info(`WASM tree-sitter enabled for AST analysis`):Q.warn(`WASM tree-sitter not available; analyzers will use regex fallback`),e})()]),c=new pe(r,i),l=new fe(n.store.path);l.load(),c.setHashCache(l);let u=n.curated.path,f=new e(u);await f.initialize();let p=new t(u,i,r,f);c.setGraphStore(o);let m=ra(n.er),h=m?new le(n.curated.path):void 0;h&&Q.info(`Policy store initialized`,{ruleCount:h.getRules().length});let g=m?new ce:void 0,v=_(n.sources[0]?.path??process.cwd(),y.aiContext),b=a(v),x=n.onboardDir?a(n.onboardDir):!1,S=b||x,C,w=b?v:n.onboardDir;if(S&&w)try{C=d(w).mtime.toISOString()}catch{}return Q.info(`Onboard state detected`,{onboardComplete:S,onboardTimestamp:C,aiKbExists:b,onboardDirExists:x}),{embedder:r,store:i,indexer:c,curated:p,graphStore:o,fileCache:new ge,bridge:m,policyStore:h,evolutionCollector:g,onboardComplete:S,onboardTimestamp:C}}function gl(e,t){let n=new nn({name:t.serverName??`aikit`,version:r()},{capabilities:{logging:{},completions:{},prompts:{}}});return gr(n),Sl(n,e,t,dr(n),new Lr(n),Ur(n)),yr(n,{curated:e.curated,store:e.store,graphStore:e.graphStore},t.indexMode),n}function _l(e){return e.toolProfiles}const vl=[`flow_list`,`flow_info`,`flow_start`,`flow_step`,`flow_status`,`flow_reset`,`flow_read_instruction`,`flow_add`,`flow_remove`,`flow_update`],yl=[`er_push`,`er_pull`,`er_sync_status`],bl=[...yl,`er_update_policy`,`er_evolve_review`],xl=new Set(bl);function Sl(e,t,n,r,i,a,o,s){let c=new nr,l=new Pn;l.register(Nn),l.register(An),l.register(kn);let u=new Bn(l,t.curated);c.use(Ln(u),{order:5,name:`auto-knowledge`}),c.use(Fr(),{order:10,name:`replay`}),c.use(Wr(),{order:50,name:`structured-content-guard`}),c.use(ur(),{order:90,name:`compression`}),sr(e,c,n.toolPrefix??``);let d=$r(process.env.AIKIT_TOOLSET||n.toolProfile||`full`,[...$,...bl],rr,_l(n)),f=e=>d.has(e),p=[...d].filter(e=>xl.has(e)?yl.includes(e)?!!t.bridge:e===`er_update_policy`?!!t.policyStore:e===`er_evolve_review`?!!t.evolutionCollector:!1:!0);f(`search`)&&Hc(e,t.embedder,t.store,t.graphStore,t.bridge,t.evolutionCollector,a),f(`lookup`)&&Bo(e,t.store);let m={onboardComplete:t.onboardComplete,onboardTimestamp:t.onboardTimestamp};f(`status`)&&el(e,t.store,t.graphStore,t.curated,m,n,o,s),f(`config`)&&Pa(e,n),f(`reindex`)&&wc(e,t.indexer,n,t.curated,t.store,i,o),f(`remember`)&&Ec(e,t.curated,t.policyStore,t.evolutionCollector,i),f(`update`)&&nl(e,t.curated,i),f(`forget`)&&wo(e,t.curated,i),f(`read`)&&Cc(e,t.curated),f(`list`)&&Ro(e,t.curated),f(`analyze_structure`)&&Bi(e,t.store,t.embedder),f(`analyze_dependencies`)&&Vi(e,t.store,t.embedder),f(`analyze_symbols`)&&Hi(e,t.store,t.embedder),f(`analyze_patterns`)&&Ui(e,t.store,t.embedder),f(`analyze_entry_points`)&&Wi(e,t.store,t.embedder),f(`analyze_diagram`)&&Gi(e,t.store,t.embedder),f(`blast_radius`)&&Ki(e,t.store,t.embedder,t.graphStore),f(`produce_knowledge`)&&xc(e,n),f(`onboard`)&&ts(e,t.store,t.embedder,n,m),f(`graph`)&&Oo(e,t.graphStore),f(`audit`)&&Zi(e,t.store,t.embedder);let h=n.sources[0]?.path??process.cwd();f(`compact`)&&Va(e,t.embedder,t.fileCache,h),f(`scope_map`)&&Ha(e,t.embedder,t.store),f(`find`)&&Ua(e,t.embedder,t.store),f(`parse_output`)&&io(e),f(`workset`)&&rs(e),f(`check`)&&eo(e),f(`batch`)&&to(e,t.embedder,t.store),f(`symbol`)&&Wa(e,t.embedder,t.store,t.graphStore),f(`eval`)&&no(e),f(`test_run`)&&ro(e),f(`stash`)&&is(e),f(`git_context`)&&Ho(e),f(`diff_parse`)&&Uo(e),f(`rename`)&&Wo(e),f(`codemod`)&&Go(e),f(`restore`)&&Ac(e),f(`file_summary`)&&Ga(e,t.fileCache,h),f(`checkpoint`)&&as(e),f(`data_transform`)&&Ko(e),f(`trace`)&&Ka(e,t.embedder,t.store,t.graphStore),f(`process`)&&ko(e),f(`watch`)&&Ao(e),f(`dead_symbols`)&&qa(e,t.embedder,t.store),f(`delegate`)&&ao(e,a),f(`health`)&&jo(e),f(`lane`)&&os(e),f(`queue`)&&ss(e),f(`web_fetch`)&&Mo(e),f(`guide`)&&No(e,o),Yo.some(e=>f(e))&&Xo(e,n,p),f(`evidence_map`)&&vo(e),f(`digest`)&&yo(e,t.embedder),f(`forge_classify`)&&bo(e),f(`stratum_card`)&&xo(e,t.embedder,t.fileCache),f(`forge_ground`)&&So(e,t.embedder,t.store),f(`present`)&&_c(e,r),r&&f(`brainstorm`)&&$i(e,r),f(`web_search`)&&rl(e),f(`http`)&&il(e),f(`regex_test`)&&al(e),f(`encode`)&&ol(e),f(`measure`)&&sl(e),f(`changelog`)&&cl(e),f(`schema_validate`)&&ll(e),f(`snippet`)&&ul(e),f(`env`)&&dl(e),f(`time`)&&fl(e),vl.some(e=>f(e))&&go(e,n),t.bridge&&yl.some(e=>f(e))&&(ia(e,t.bridge,t.evolutionCollector),aa(e,t.bridge),oa(e,t.bridge)),t.policyStore&&f(`er_update_policy`)&&ps(e,t.policyStore),t.evolutionCollector&&f(`er_evolve_review`)&&Qa(e,t.evolutionCollector),zr(e,t.store,t.curated),Oc(e),f(`session_digest`)&&Wc(e,a)}async function Cl(e){let t=await hl(e),n=gl(t,e);Q.info(`MCP server configured`,{toolCount:$.length,resourceCount:2});let r=async()=>{try{let n=e.sources.map(e=>e.path).join(`, `);Q.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Q.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Q.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});Q.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Q.warn(`FTS index creation failed`,A(e))}try{let e=await t.curated.reindexAll();Q.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Q.error(`Curated re-index failed`,A(e))}}catch(e){Q.error(`Initial index failed; will retry on aikit_reindex`,A(e))}},i=async()=>{Q.info(`Shutting down`),await Promise.all([t.embedder.shutdown().catch(()=>{}),t.graphStore.close().catch(()=>{}),t.store.close().catch(()=>{})]),process.exit(0)};process.on(`SIGINT`,i),process.on(`SIGTERM`,i);let a=process.ppid,o=setInterval(()=>{try{process.kill(a,0)}catch{Q.info(`Parent process died; shutting down`,{parentPid:a}),clearInterval(o),i()}},5e3);return o.unref(),{server:n,runInitialIndex:r,shutdown:i}}const wl=new Set(`batch.brainstorm.changelog.check.checkpoint.codemod.compact.config.data_transform.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.forge_classify.git_context.graph.guide.health.http.lane.measure.onboard.parse_output.present.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.restore.schema_validate.session_digest.scope_map.snippet.stash.status.stratum_card.test_run.time.update.forget.list.watch.web_fetch.web_search.workset`.split(`.`)),Tl=5e3,El=new Set(`brainstorm.changelog.check.checkpoint.codemod.data_transform.delegate.diff_parse.encode.env.eval.evidence_map.flow_info.flow_list.flow_reset.flow_start.flow_status.flow_step.flow_add.flow_update.flow_remove.flow_read_instruction.describe_tool.list_tools.search_tools.forge_classify.git_context.guide.present.health.http.lane.measure.parse_output.process.produce_knowledge.queue.regex_test.rename.replay.restore.schema_validate.session_digest.snippet.stash.status.test_run.time.watch.web_fetch.web_search.workset`.split(`.`));function Dl(e,t,n){let r=e=>!n||n.has(e);r(`check`)&&eo(e),r(`eval`)&&no(e),r(`test_run`)&&ro(e),r(`parse_output`)&&io(e),r(`delegate`)&&ao(e),r(`git_context`)&&Ho(e),r(`diff_parse`)&&Uo(e),r(`rename`)&&Wo(e),r(`codemod`)&&Go(e),r(`data_transform`)&&Ko(e),r(`workset`)&&rs(e),r(`stash`)&&is(e),r(`checkpoint`)&&as(e),r(`restore`)&&Ac(e),r(`lane`)&&os(e),r(`queue`)&&ss(e),r(`session_digest`)&&Wc(e),r(`health`)&&jo(e),r(`process`)&&ko(e),r(`watch`)&&Ao(e),r(`web_fetch`)&&Mo(e),r(`guide`)&&No(e),Yo.some(e=>r(e))&&Xo(e,t,[...n??new Set($)]),r(`evidence_map`)&&vo(e),r(`forge_classify`)&&bo(e),r(`present`)&&_c(e),r(`brainstorm`)&&$i(e,fr),r(`produce_knowledge`)&&xc(e),r(`replay`)&&Oc(e),r(`status`)&&$c(e),vl.some(e=>r(e))&&go(e,t),r(`web_search`)&&rl(e),r(`http`)&&il(e),r(`regex_test`)&&al(e),r(`encode`)&&ol(e),r(`measure`)&&sl(e),r(`changelog`)&&cl(e),r(`schema_validate`)&&ll(e),r(`snippet`)&&ul(e),r(`env`)&&dl(e),r(`time`)&&fl(e)}const $=`analyze_dependencies.analyze_diagram.analyze_entry_points.analyze_patterns.analyze_structure.analyze_symbols.audit.batch.blast_radius.brainstorm.changelog.check.checkpoint.codemod.compact.config.data_transform.dead_symbols.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.find.flow_info.flow_list.flow_reset.flow_start.flow_status.flow_step.flow_add.flow_update.flow_remove.flow_read_instruction.forge_classify.forge_ground.forget.git_context.graph.guide.health.http.lane.describe_tool.list_tools.list.lookup.measure.onboard.parse_output.present.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.restore.schema_validate.scope_map.search.search_tools.session_digest.snippet.stash.status.stratum_card.symbol.test_run.time.trace.update.watch.web_fetch.web_search.workset`.split(`.`);function Ol(e,t){let n=new nn({name:e.serverName??`aikit`,version:r()},{capabilities:{logging:{},completions:{},prompts:{}}}),i=`initializing`,a=``,o=!1,s=null,c=$r(process.env.AIKIT_TOOLSET||e.toolProfile||`full`,$,rr,_l(e)),l=null,u=null;function d(e){if(!e||typeof e!=`object`)return[];let t=e,n=[];for(let e of[`path`,`file`,`source_path`,`sourcePath`,`filePath`]){let r=t[e];typeof r==`string`&&r&&n.push(r)}for(let e of[`changed_files`,`paths`,`files`]){let r=t[e];if(Array.isArray(r))for(let e of r){if(typeof e==`string`){n.push(e);continue}e&&typeof e==`object`&&typeof e.path==`string`&&n.push(e.path)}}if(Array.isArray(t.sources))for(let e of t.sources)e&&typeof e==`object`&&typeof e.path==`string`&&n.push(e.path);return n}let f=()=>i===`failed`?[`❌ AI Kit initialization failed — this tool is unavailable.`,``,a?`Error: ${a}`:``,``,`**${El.size} tools are still available** and fully functional:`,`check, eval, test_run, git_context, health, measure, web_fetch, web_search,`,`flow_list, flow_status, flow_start, flow_read_instruction, regex_test, encode,`,`stash, checkpoint, lane, process, time, env, and more.`,``,`To fix embedding errors, try deleting the cached model:`,` rm -rf ~/.cache/huggingface/transformers-js/mixedbread-ai/`,`Then restart the server to re-download a fresh copy.`,``,`Try restarting the MCP server to retry initialization.`].filter(Boolean).join(`
1273
- `):[`AI Kit is still initializing (loading embeddings model & store).`,``,`**${El.size} tools are already available** while initialization completes — including:`,`check, eval, test_run, git_context, health, measure, web_fetch, web_search,`,`flow_list, flow_status, flow_start, flow_read_instruction, regex_test, encode,`,`stash, checkpoint, lane, process, time, env, and more.`,``,`This tool requires the AI Kit index. Please retry in a few seconds,`,`or use one of the available tools above in the meantime.`].join(`
1274
- `);gr(n),cr(n,e.toolPrefix??``);let p=n.sendToolListChanged.bind(n);n.sendToolListChanged=()=>{};let m=[];for(let e of $){if(!c.has(e))continue;let t=I(e),r=n.registerTool(e,{title:t.title,description:`${t.title} — initializing, available shortly`,inputSchema:{},annotations:t.annotations},async()=>({content:[{type:`text`,text:f()}]}));El.has(e)?r.remove():m.push(r)}Dl(n,e,c),n.sendToolListChanged=p;let h=n.registerResource(`aikit-status`,`aikit://status`,{description:`AI Kit status (initializing...)`,mimeType:`text/plain`},async()=>({contents:[{uri:`aikit://status`,text:`AI Kit is initializing...`,mimeType:`text/plain`}]})),g=n.registerPrompt(`_init`,{description:`Initializing AI Kit…`,argsSchema:{_dummy:P(F.string(),()=>[])}},async()=>({messages:[]})),_,v=new Promise(e=>{_=e}),y,b=new Promise(e=>{y=e}),x=()=>y?.(),S=(async()=>{await b;let r;try{r=await hl(e)}catch(t){let n=t instanceof Error?t.message:String(t);if(pl(n)){Q.warn(`AI Kit initialization failed with recoverable error — attempting auto-heal retry`,{error:n}),await ml(e,n);try{r=await hl(e),Q.info(`AI Kit auto-heal successful — initialization recovered after retry`)}catch(e){i=`failed`,a=e instanceof Error?e.message:String(e),Q.error(`AI Kit initialization failed after auto-heal attempt — server continuing with zero-dep tools only`,{error:a,originalError:n});return}}else{i=`failed`,a=n,Q.error(`AI Kit initialization failed — server continuing with zero-dep tools only`,{error:a});return}}let c=n.sendToolListChanged.bind(n);n.sendToolListChanged=()=>{};let f=n.sendPromptListChanged.bind(n);n.sendPromptListChanged=()=>{};let p=n.sendResourceListChanged.bind(n);n.sendResourceListChanged=()=>{};for(let e of m)e.remove();h.remove(),g.remove();let v=n._registeredTools??{};for(let e of El)v[e]?.remove();let y=new Lr(n),x=Ur(n);Sl(n,r,e,dr(n),y,x,t,t===`smart`?(()=>{let e=u;return e?.getState?e.getState():null}):null),yr(n,{curated:r.curated,store:r.store,graphStore:r.graphStore},t),n.sendToolListChanged=c,n.sendPromptListChanged=f,n.sendResourceListChanged=p,Promise.resolve(n.sendToolListChanged()).catch(()=>{}),Promise.resolve(n.sendPromptListChanged()).catch(()=>{}),Promise.resolve(n.sendResourceListChanged()).catch(()=>{});let S=n._registeredTools??{};for(let[e,t]of Object.entries(S)){if(wl.has(e))continue;let n=t.handler;t.handler=async(...t)=>{if(!r.indexer.isIndexing)return n(...t);let i=o?`re-indexing`:`running initial index`,a=new Promise(t=>setTimeout(()=>t({content:[{type:`text`,text:`⏳ AI Kit is ${i}. The tool "${e}" timed out waiting for index data (${Tl/1e3}s).\n\nThe existing index may be temporarily locked. Please retry shortly — indexing will complete automatically.`}]}),Tl));return Promise.race([n(...t),a])}}for(let[e,t]of Object.entries(S)){let n=t.handler,r=ri(e);t.handler=async(...t)=>{try{return await ii(()=>n(...t),r,e)}catch(t){if(t instanceof ni)return{content:[{type:`text`,text:`⏳ Tool "${e}" timed out after ${r/1e3}s. This may indicate a long-running operation. Please retry or break the task into smaller steps.`}]};throw t}}}let C=Object.keys(S).length;C<$.length&&Q.warn(`ALL_TOOL_NAMES count mismatch`,{expectedToolCount:$.length,registeredToolCount:C}),Q.info(`MCP server configured`,{toolCount:$.length,resourceCount:4});let T=new vr;T.onPressure((e,t)=>{e===`warning`&&Kn(),e===`critical`&&(Q.warn(`Memory pressure critical — consider restarting`,{rssMB:Math.round(t/1024/1024)}),Kn())}),T.start();let E=new mr;l=E,E.onIdle(async()=>{if(w.isRunning||r.indexer.isIndexing){Q.info(`Idle cleanup deferred — background tasks still running`),E.touch();return}Q.info(`Idle cleanup: closing store and graph connections`);try{await Promise.all([r.store.close().catch(()=>{}),r.graphStore.close().catch(()=>{})])}catch{}}),E.touch();for(let e of Object.values(S)){let t=e.handler;e.handler=async(...e)=>{if(E.touch(),u){let t=d(e[0]);t.length>0&&u.prioritize(...t)}return t(...e)}}s=r,_?.(r)})(),C=async()=>{let t=await v;l?.setBusy(!0);try{let n=e.sources.map(e=>e.path).join(`, `);Q.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Q.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Q.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});o=!0,Q.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Q.warn(`FTS index creation failed`,A(e))}try{let e=await t.curated.reindexAll();Q.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Q.error(`Curated re-index failed`,A(e))}}catch(e){Q.error(`Initial index failed; will retry on aikit_reindex`,A(e))}finally{l?.setBusy(!1)}},w=new Hn,T=()=>w.schedule({name:`initial-index`,fn:C}),E=process.ppid,D=setInterval(()=>{try{process.kill(E,0)}catch{Q.info(`Parent process died; shutting down`,{parentPid:E}),clearInterval(D),v.then(async e=>{await Promise.all([e.embedder.shutdown().catch(()=>{}),e.graphStore.close().catch(()=>{}),e.store.close().catch(()=>{})])}).catch(()=>{}).finally(()=>process.exit(0))}},5e3);return D.unref(),{server:n,startInit:x,ready:S,runInitialIndex:T,get kb(){return s},scheduler:w,setSmartScheduler(e){u=e}}}export{$ as ALL_TOOL_NAMES,Ol as createLazyServer,gl as createMcpServer,Cl as createServer,hl as initializeKnowledgeBase,Sl as registerMcpTools};
1269
+ `)}]}}catch(e){return Z.error(`Schema validation failed`,A(e)),{content:[{type:`text`,text:`Schema validation failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function dl(e){let t=L(`snippet`);e.registerTool(`snippet`,{title:t.title,description:`Save, retrieve, search, and manage persistent code snippets/templates.`,inputSchema:{action:F.enum([`save`,`get`,`list`,`search`,`delete`]).describe(`Operation to perform`),name:F.string().optional().describe(`Snippet name (required for save/get/delete)`),language:F.string().optional().describe(`Language tag (for save)`),code:F.string().max(1e5).optional().describe(`Code content (for save)`),tags:F.array(F.string()).optional().describe(`Tags for categorization (for save)`),query:F.string().optional().describe(`Search query (for search)`)},annotations:t.annotations},async({action:e,name:t,language:n,code:r,tags:i,query:a})=>{try{let o=Rt({action:e,name:t,language:n,code:r,tags:i,query:a});if(`deleted`in o)return{content:[{type:`text`,text:o.deleted?`Snippet "${t}" deleted.`:`Snippet "${t}" not found.`}]};if(`snippets`in o){if(o.snippets.length===0)return{content:[{type:`text`,text:`No snippets found.`}]};let e=[`## Snippets`,``];for(let t of o.snippets){let n=t.tags.length>0?` [${t.tags.join(`, `)}]`:``;e.push(`- **${t.name}** (${t.language})${n}`)}return{content:[{type:`text`,text:e.join(`
1270
+ `)}]}}let s=o,c=s.tags.length>0?`\nTags: ${s.tags.join(`, `)}`:``;return{content:[{type:`text`,text:`## ${s.name} (${s.language})${c}\n\n\`\`\`${s.language}\n${s.code}\n\`\`\``}]}}catch(e){return Z.error(`Snippet failed`,A(e)),{content:[{type:`text`,text:`Snippet failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}function fl(e){let t=L(`env`);e.registerTool(`env`,{title:t.title,description:`Get system and runtime environment info. Sensitive env vars are redacted by default.`,outputSchema:ci,inputSchema:{include_env:F.boolean().default(!1).describe(`Include environment variables`),filter_env:F.string().optional().describe(`Filter env vars by name substring`),show_sensitive:F.boolean().default(!1).describe(`Show sensitive values (keys, tokens, etc.) — redacted by default`)},annotations:t.annotations},async({include_env:e,filter_env:t,show_sensitive:n})=>{let r=Re({includeEnv:e,filterEnv:t,showSensitive:n}),i=[`## Environment`,``,`**Platform:** ${r.system.platform} ${r.system.arch}`,`**OS:** ${r.system.type} ${r.system.release}`,`**Host:** ${r.system.hostname}`,`**CPUs:** ${r.system.cpus}`,`**Memory:** ${r.system.memoryFreeGb}GB free / ${r.system.memoryTotalGb}GB total`,``,`**Node:** ${r.runtime.node}`,`**V8:** ${r.runtime.v8}`,`**CWD:** ${r.cwd}`];if(r.env){i.push(``,`### Environment Variables`,``);for(let[e,t]of Object.entries(r.env))i.push(`- \`${e}\`: ${t}`)}let a={platform:r.system.platform,arch:r.system.arch,nodeVersion:r.runtime.node,cwd:r.cwd,cpus:r.system.cpus,memoryFreeGb:r.system.memoryFreeGb,memoryTotalGb:r.system.memoryTotalGb};return{content:[{type:`text`,text:i.join(`
1271
+ `)}],structuredContent:a}})}function pl(e){let t=L(`time`);e.registerTool(`time`,{title:t.title,description:`Parse dates, convert timezones, calculate durations, add time. Supports ISO 8601, unix timestamps, and human-readable formats.`,outputSchema:li,inputSchema:{operation:F.enum([`now`,`parse`,`convert`,`diff`,`add`]).describe(`now: current time | parse: parse a date string | convert: timezone conversion | diff: duration between two dates | add: add duration to date`),input:F.string().optional().describe(`Date input (ISO, unix timestamp, or parseable string). For diff: two comma-separated dates`),timezone:F.string().optional().describe(`Target timezone (e.g., "America/New_York", "Asia/Tokyo")`),duration:F.string().optional().describe(`Duration to add (e.g., "2h30m", "1d", "30s") — for add operation`)},annotations:t.annotations},async({operation:e,input:t,timezone:n,duration:r})=>{try{let i=Jt({operation:e,input:t,timezone:n,duration:r}),a=[`**${i.output}**`,``,`ISO: ${i.iso}`,`Unix: ${i.unix}`];i.details&&a.push(``,"```json",JSON.stringify(i.details),"```");let o={iso:i.iso,unix:i.unix,timezone:n??Intl.DateTimeFormat().resolvedOptions().timeZone,formatted:i.output};return{content:[{type:`text`,text:a.join(`
1272
+ `)}],structuredContent:o}}catch(e){return Z.error(`Time failed`,A(e)),{content:[{type:`text`,text:`Time failed: ${e instanceof Error?e.message:String(e)}`}],isError:!0}}})}const Q=E(`server`);function ml(e){let t=e.toLowerCase();return[`protobuf`,`invalid model`,`invalid onnx`,`unexpected end`,`unexpected token`,`failed to load`,`failed to initialize embedding`,`checksum`,`corrupt`,`malformed`,`could not load`,`onnx`,`database disk image is malformed`,`file is not a database`,`lance`].some(e=>t.includes(e))}async function hl(e,t){let n=t.toLowerCase(),r;try{({rm:r}=await import(`node:fs/promises`))}catch{return}if(n.includes(`embedding`)||n.includes(`onnx`)||n.includes(`protobuf`)||n.includes(`model`)){let t=e.embedding?.model??x.model,n=g(N(),`.cache`,`huggingface`,`transformers-js`,t);try{await r(n,{recursive:!0,force:!0}),Q.info(`Auto-heal: cleared embedding model cache`,{path:n})}catch{}}if(n.includes(`lance`)||n.includes(`database`)||n.includes(`store`)){let t=g(e.store.path,`lance`);try{await r(t,{recursive:!0,force:!0}),Q.info(`Auto-heal: cleared LanceDB store`,{path:t})}catch{}}if(n.includes(`sqlite`)||n.includes(`database disk image`)||n.includes(`graph`)){let t=g(e.store.path,`graph.db`);try{await r(t,{force:!0}),Q.info(`Auto-heal: cleared graph database`,{path:t})}catch{}}}async function gl(n){Q.info(`Initializing AI Kit components`);let[r,i,o,s]=await Promise.all([(async()=>{let e=new ae({model:n.embedding.model,dimensions:n.embedding.dimensions});return await e.initialize(),Q.info(`Embedder loaded`,{modelId:e.modelId,dimensions:e.dimensions}),e})(),(async()=>{let e=await he({backend:n.store.backend,path:n.store.path});return await e.initialize(),Q.info(`Store initialized`),e})(),(async()=>{let e=new me({path:n.store.path});return await e.initialize(),Q.info(`Graph store initialized`),e})(),(async()=>{let e=await ie();return e?Q.info(`WASM tree-sitter enabled for AST analysis`):Q.warn(`WASM tree-sitter not available; analyzers will use regex fallback`),e})()]),c=new pe(r,i),l=new fe(n.store.path);l.load(),c.setHashCache(l);let u=n.curated.path,f=new e(u);await f.initialize();let p=new t(u,i,r,f);c.setGraphStore(o);let m=na(n.er),h=m?new le(n.curated.path):void 0;h&&Q.info(`Policy store initialized`,{ruleCount:h.getRules().length});let g=m?new ce:void 0,v=_(n.sources[0]?.path??process.cwd(),y.aiContext),b=a(v),x=n.onboardDir?a(n.onboardDir):!1,S=b||x,C,w=b?v:n.onboardDir;if(S&&w)try{C=d(w).mtime.toISOString()}catch{}return Q.info(`Onboard state detected`,{onboardComplete:S,onboardTimestamp:C,aiKbExists:b,onboardDirExists:x}),{embedder:r,store:i,indexer:c,curated:p,graphStore:o,fileCache:new ge,bridge:m,policyStore:h,evolutionCollector:g,onboardComplete:S,onboardTimestamp:C}}function _l(e,t){let n=new nn({name:t.serverName??`aikit`,version:r()},{capabilities:{logging:{},completions:{},prompts:{}}});return hr(n),Cl(n,e,t,ur(n),new Ir(n),Hr(n)),vr(n,{curated:e.curated,store:e.store,graphStore:e.graphStore},t.indexMode),n}function vl(e){return e.toolProfiles}const yl=[`flow_list`,`flow_info`,`flow_start`,`flow_step`,`flow_status`,`flow_reset`,`flow_read_instruction`,`flow_add`,`flow_remove`,`flow_update`],bl=[`er_push`,`er_pull`,`er_sync_status`],xl=[...bl,`er_update_policy`,`er_evolve_review`],Sl=new Set(xl);function Cl(e,t,n,r,i,a,o,s){let c=new nr,l=new Pn;l.register(Nn),l.register(An),l.register(kn);let u=new Bn(l,t.curated);c.use(Ln(u),{order:5,name:`auto-knowledge`}),c.use(Pr(),{order:10,name:`replay`}),c.use(Ur(),{order:50,name:`structured-content-guard`}),c.use(lr(),{order:90,name:`compression`}),or(e,c,n.toolPrefix??``);let d=Qr(process.env.AIKIT_TOOLSET||n.toolProfile||`full`,[...$,...xl],I,vl(n)),f=e=>d.has(e),p=[...d].filter(e=>Sl.has(e)?bl.includes(e)?!!t.bridge:e===`er_update_policy`?!!t.policyStore:e===`er_evolve_review`?!!t.evolutionCollector:!1:!0);f(`search`)&&Uc(e,t.embedder,t.store,t.graphStore,t.bridge,t.evolutionCollector,a),f(`lookup`)&&Bo(e,t.store);let m={onboardComplete:t.onboardComplete,onboardTimestamp:t.onboardTimestamp};f(`status`)&&tl(e,t.store,t.graphStore,t.curated,m,n,o,s),f(`config`)&&Na(e,n),f(`reindex`)&&Tc(e,t.indexer,n,t.curated,t.store,i,o),f(`remember`)&&Dc(e,t.curated,t.policyStore,t.evolutionCollector,i),f(`update`)&&rl(e,t.curated,i),f(`forget`)&&Co(e,t.curated,i),f(`read`)&&wc(e,t.curated),f(`list`)&&Ro(e,t.curated),f(`analyze_structure`)&&zi(e,t.store,t.embedder),f(`analyze_dependencies`)&&Bi(e,t.store,t.embedder),f(`analyze_symbols`)&&Vi(e,t.store,t.embedder),f(`analyze_patterns`)&&Hi(e,t.store,t.embedder),f(`analyze_entry_points`)&&Ui(e,t.store,t.embedder),f(`analyze_diagram`)&&Wi(e,t.store,t.embedder),f(`blast_radius`)&&Gi(e,t.store,t.embedder,t.graphStore),f(`produce_knowledge`)&&Sc(e,n),f(`onboard`)&&ts(e,t.store,t.embedder,n,m),f(`graph`)&&Do(e,t.graphStore),f(`audit`)&&Xi(e,t.store,t.embedder);let h=n.sources[0]?.path??process.cwd();f(`compact`)&&Ba(e,t.embedder,t.fileCache,h),f(`scope_map`)&&Va(e,t.embedder,t.store),f(`find`)&&Ha(e,t.embedder,t.store),f(`parse_output`)&&ro(e),f(`workset`)&&rs(e),f(`check`)&&$a(e),f(`batch`)&&eo(e,t.embedder,t.store),f(`symbol`)&&Ua(e,t.embedder,t.store,t.graphStore),f(`eval`)&&to(e),f(`test_run`)&&no(e),f(`stash`)&&is(e),f(`git_context`)&&Ho(e),f(`diff_parse`)&&Uo(e),f(`rename`)&&Wo(e),f(`codemod`)&&Go(e),f(`restore`)&&jc(e),f(`file_summary`)&&Wa(e,t.fileCache,h),f(`checkpoint`)&&as(e),f(`data_transform`)&&Ko(e),f(`trace`)&&Ga(e,t.embedder,t.store,t.graphStore),f(`process`)&&ko(e),f(`watch`)&&Ao(e),f(`dead_symbols`)&&Ka(e,t.embedder,t.store),f(`delegate`)&&io(e,a),f(`health`)&&jo(e),f(`lane`)&&os(e),f(`queue`)&&ss(e),f(`web_fetch`)&&Mo(e),f(`guide`)&&No(e,o),Yo.some(e=>f(e))&&Xo(e,n,p),f(`evidence_map`)&&_o(e),f(`digest`)&&vo(e,t.embedder),f(`forge_classify`)&&yo(e),f(`stratum_card`)&&bo(e,t.embedder,t.fileCache),f(`forge_ground`)&&xo(e,t.embedder,t.store),f(`present`)&&vc(e,r),r&&f(`brainstorm`)&&Qi(e,r),f(`web_search`)&&il(e),f(`http`)&&al(e),f(`regex_test`)&&ol(e),f(`encode`)&&sl(e),f(`measure`)&&cl(e),f(`changelog`)&&ll(e),f(`schema_validate`)&&ul(e),f(`snippet`)&&dl(e),f(`env`)&&fl(e),f(`time`)&&pl(e),yl.some(e=>f(e))&&ho(e,n),t.bridge&&bl.some(e=>f(e))&&(ra(e,t.bridge,t.evolutionCollector),ia(e,t.bridge),aa(e,t.bridge)),t.policyStore&&f(`er_update_policy`)&&ps(e,t.policyStore),t.evolutionCollector&&f(`er_evolve_review`)&&Za(e,t.evolutionCollector),Rr(e,t.store,t.curated),kc(e),f(`session_digest`)&&Gc(e,a)}async function wl(e){let t=await gl(e),n=_l(t,e);Q.info(`MCP server configured`,{toolCount:$.length,resourceCount:2});let r=async()=>{try{let n=e.sources.map(e=>e.path).join(`, `);Q.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Q.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Q.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});Q.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Q.warn(`FTS index creation failed`,A(e))}try{let e=await t.curated.reindexAll();Q.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Q.error(`Curated re-index failed`,A(e))}}catch(e){Q.error(`Initial index failed; will retry on aikit_reindex`,A(e))}},i=async()=>{Q.info(`Shutting down`),await Promise.all([t.embedder.shutdown().catch(()=>{}),t.graphStore.close().catch(()=>{}),t.store.close().catch(()=>{})]),process.exit(0)};process.on(`SIGINT`,i),process.on(`SIGTERM`,i);let a=process.ppid,o=setInterval(()=>{try{process.kill(a,0)}catch{Q.info(`Parent process died; shutting down`,{parentPid:a}),clearInterval(o),i()}},5e3);return o.unref(),{server:n,runInitialIndex:r,shutdown:i}}const Tl=new Set(`batch.brainstorm.changelog.check.checkpoint.codemod.compact.config.data_transform.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.forge_classify.git_context.graph.guide.health.http.lane.measure.onboard.parse_output.present.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.restore.schema_validate.session_digest.scope_map.snippet.stash.status.stratum_card.test_run.time.update.forget.list.watch.web_fetch.web_search.workset`.split(`.`)),El=5e3,Dl=new Set(`brainstorm.changelog.check.checkpoint.codemod.data_transform.delegate.diff_parse.encode.env.eval.evidence_map.flow_info.flow_list.flow_reset.flow_start.flow_status.flow_step.flow_add.flow_update.flow_remove.flow_read_instruction.describe_tool.list_tools.search_tools.forge_classify.git_context.guide.present.health.http.lane.measure.parse_output.process.produce_knowledge.queue.regex_test.rename.replay.restore.schema_validate.session_digest.snippet.stash.status.test_run.time.watch.web_fetch.web_search.workset`.split(`.`));function Ol(e,t,n){let r=e=>!n||n.has(e);r(`check`)&&$a(e),r(`eval`)&&to(e),r(`test_run`)&&no(e),r(`parse_output`)&&ro(e),r(`delegate`)&&io(e),r(`git_context`)&&Ho(e),r(`diff_parse`)&&Uo(e),r(`rename`)&&Wo(e),r(`codemod`)&&Go(e),r(`data_transform`)&&Ko(e),r(`workset`)&&rs(e),r(`stash`)&&is(e),r(`checkpoint`)&&as(e),r(`restore`)&&jc(e),r(`lane`)&&os(e),r(`queue`)&&ss(e),r(`session_digest`)&&Gc(e),r(`health`)&&jo(e),r(`process`)&&ko(e),r(`watch`)&&Ao(e),r(`web_fetch`)&&Mo(e),r(`guide`)&&No(e),Yo.some(e=>r(e))&&Xo(e,t,[...n??new Set($)]),r(`evidence_map`)&&_o(e),r(`forge_classify`)&&yo(e),r(`present`)&&vc(e),r(`brainstorm`)&&Qi(e,dr),r(`produce_knowledge`)&&Sc(e),r(`replay`)&&kc(e),r(`status`)&&el(e),yl.some(e=>r(e))&&ho(e,t),r(`web_search`)&&il(e),r(`http`)&&al(e),r(`regex_test`)&&ol(e),r(`encode`)&&sl(e),r(`measure`)&&cl(e),r(`changelog`)&&ll(e),r(`schema_validate`)&&ul(e),r(`snippet`)&&dl(e),r(`env`)&&fl(e),r(`time`)&&pl(e)}const $=`analyze_dependencies.analyze_diagram.analyze_entry_points.analyze_patterns.analyze_structure.analyze_symbols.audit.batch.blast_radius.brainstorm.changelog.check.checkpoint.codemod.compact.config.data_transform.dead_symbols.delegate.diff_parse.digest.encode.env.eval.evidence_map.file_summary.find.flow_info.flow_list.flow_reset.flow_start.flow_status.flow_step.flow_add.flow_update.flow_remove.flow_read_instruction.forge_classify.forge_ground.forget.git_context.graph.guide.health.http.lane.describe_tool.list_tools.list.lookup.measure.onboard.parse_output.present.process.produce_knowledge.queue.read.regex_test.reindex.remember.rename.replay.restore.schema_validate.scope_map.search.search_tools.session_digest.snippet.stash.status.stratum_card.symbol.test_run.time.trace.update.watch.web_fetch.web_search.workset`.split(`.`);function kl(e,t){let n=new nn({name:e.serverName??`aikit`,version:r()},{capabilities:{logging:{},completions:{},prompts:{}}}),i=`initializing`,a=``,o=!1,s=null,c=Qr(process.env.AIKIT_TOOLSET||e.toolProfile||`full`,$,I,vl(e)),l=null,u=null;function d(e){if(!e||typeof e!=`object`)return[];let t=e,n=[];for(let e of[`path`,`file`,`source_path`,`sourcePath`,`filePath`]){let r=t[e];typeof r==`string`&&r&&n.push(r)}for(let e of[`changed_files`,`paths`,`files`]){let r=t[e];if(Array.isArray(r))for(let e of r){if(typeof e==`string`){n.push(e);continue}e&&typeof e==`object`&&typeof e.path==`string`&&n.push(e.path)}}if(Array.isArray(t.sources))for(let e of t.sources)e&&typeof e==`object`&&typeof e.path==`string`&&n.push(e.path);return n}let f=()=>i===`failed`?[`❌ AI Kit initialization failed — this tool is unavailable.`,``,a?`Error: ${a}`:``,``,`**${Dl.size} tools are still available** and fully functional:`,`check, eval, test_run, git_context, health, measure, web_fetch, web_search,`,`flow_list, flow_status, flow_start, flow_read_instruction, regex_test, encode,`,`stash, checkpoint, lane, process, time, env, and more.`,``,`To fix embedding errors, try deleting the cached model:`,` rm -rf ~/.cache/huggingface/transformers-js/mixedbread-ai/`,`Then restart the server to re-download a fresh copy.`,``,`Try restarting the MCP server to retry initialization.`].filter(Boolean).join(`
1273
+ `):[`AI Kit is still initializing (loading embeddings model & store).`,``,`**${Dl.size} tools are already available** while initialization completes — including:`,`check, eval, test_run, git_context, health, measure, web_fetch, web_search,`,`flow_list, flow_status, flow_start, flow_read_instruction, regex_test, encode,`,`stash, checkpoint, lane, process, time, env, and more.`,``,`This tool requires the AI Kit index. Please retry in a few seconds,`,`or use one of the available tools above in the meantime.`].join(`
1274
+ `);hr(n),sr(n,e.toolPrefix??``);let p=n.sendToolListChanged.bind(n);n.sendToolListChanged=()=>{};let m=[];for(let e of $){if(!c.has(e))continue;let t=L(e),r=n.registerTool(e,{title:t.title,description:`${t.title} — initializing, available shortly`,inputSchema:{},annotations:t.annotations},async()=>({content:[{type:`text`,text:f()}]}));Dl.has(e)?r.remove():m.push(r)}Ol(n,e,c),n.sendToolListChanged=p;let h=n.registerResource(`aikit-status`,`aikit://status`,{description:`AI Kit status (initializing...)`,mimeType:`text/plain`},async()=>({contents:[{uri:`aikit://status`,text:`AI Kit is initializing...`,mimeType:`text/plain`}]})),g=n.registerPrompt(`_init`,{description:`Initializing AI Kit…`,argsSchema:{_dummy:P(F.string(),()=>[])}},async()=>({messages:[]})),_,v=new Promise(e=>{_=e}),y,b=new Promise(e=>{y=e}),x=()=>y?.(),S=(async()=>{await b;let r;try{r=await gl(e)}catch(t){let n=t instanceof Error?t.message:String(t);if(ml(n)){Q.warn(`AI Kit initialization failed with recoverable error — attempting auto-heal retry`,{error:n}),await hl(e,n);try{r=await gl(e),Q.info(`AI Kit auto-heal successful — initialization recovered after retry`)}catch(e){i=`failed`,a=e instanceof Error?e.message:String(e),Q.error(`AI Kit initialization failed after auto-heal attempt — server continuing with zero-dep tools only`,{error:a,originalError:n});return}}else{i=`failed`,a=n,Q.error(`AI Kit initialization failed — server continuing with zero-dep tools only`,{error:a});return}}let c=n.sendToolListChanged.bind(n);n.sendToolListChanged=()=>{};let f=n.sendPromptListChanged.bind(n);n.sendPromptListChanged=()=>{};let p=n.sendResourceListChanged.bind(n);n.sendResourceListChanged=()=>{};for(let e of m)e.remove();h.remove(),g.remove();let v=n._registeredTools??{};for(let e of Dl)v[e]?.remove();let y=new Ir(n),x=Hr(n);Cl(n,r,e,ur(n),y,x,t,t===`smart`?(()=>{let e=u;return e?.getState?e.getState():null}):null),vr(n,{curated:r.curated,store:r.store,graphStore:r.graphStore},t),n.sendToolListChanged=c,n.sendPromptListChanged=f,n.sendResourceListChanged=p,Promise.resolve(n.sendToolListChanged()).catch(()=>{}),Promise.resolve(n.sendPromptListChanged()).catch(()=>{}),Promise.resolve(n.sendResourceListChanged()).catch(()=>{});let S=n._registeredTools??{};for(let[e,t]of Object.entries(S)){if(Tl.has(e))continue;let n=t.handler;t.handler=async(...t)=>{if(!r.indexer.isIndexing)return n(...t);let i=o?`re-indexing`:`running initial index`,a=new Promise(t=>setTimeout(()=>t({content:[{type:`text`,text:`⏳ AI Kit is ${i}. The tool "${e}" timed out waiting for index data (${El/1e3}s).\n\nThe existing index may be temporarily locked. Please retry shortly — indexing will complete automatically.`}]}),El));return Promise.race([n(...t),a])}}for(let[e,t]of Object.entries(S)){let n=t.handler,r=ni(e);t.handler=async(...t)=>{try{return await ri(()=>n(...t),r,e)}catch(t){if(t instanceof ti)return{content:[{type:`text`,text:`⏳ Tool "${e}" timed out after ${r/1e3}s. This may indicate a long-running operation. Please retry or break the task into smaller steps.`}]};throw t}}}let C=Object.keys(S).length;C<$.length&&Q.warn(`ALL_TOOL_NAMES count mismatch`,{expectedToolCount:$.length,registeredToolCount:C}),Q.info(`MCP server configured`,{toolCount:$.length,resourceCount:4});let T=new _r;T.onPressure((e,t)=>{e===`warning`&&Kn(),e===`critical`&&(Q.warn(`Memory pressure critical — consider restarting`,{rssMB:Math.round(t/1024/1024)}),Kn())}),T.start();let E=new pr;l=E,E.onIdle(async()=>{if(w.isRunning||r.indexer.isIndexing){Q.info(`Idle cleanup deferred — background tasks still running`),E.touch();return}Q.info(`Idle cleanup: closing store and graph connections`);try{await Promise.all([r.store.close().catch(()=>{}),r.graphStore.close().catch(()=>{})])}catch{}}),E.touch();for(let e of Object.values(S)){let t=e.handler;e.handler=async(...e)=>{if(E.touch(),u){let t=d(e[0]);t.length>0&&u.prioritize(...t)}return t(...e)}}s=r,_?.(r)})(),C=async()=>{let t=await v;l?.setBusy(!0);try{let n=e.sources.map(e=>e.path).join(`, `);Q.info(`Running initial index`,{sourcePaths:n});let r=await t.indexer.index(e,e=>{e.phase===`crawling`||e.phase===`done`||(e.phase===`chunking`&&e.currentFile&&Q.debug(`Indexing file`,{current:e.filesProcessed+1,total:e.filesTotal,file:e.currentFile}),e.phase===`cleanup`&&Q.debug(`Index cleanup`,{staleEntries:e.filesTotal-e.filesProcessed}))});o=!0,Q.info(`Initial index complete`,{filesProcessed:r.filesProcessed,filesSkipped:r.filesSkipped,chunksCreated:r.chunksCreated,durationMs:r.durationMs});try{await t.store.createFtsIndex()}catch(e){Q.warn(`FTS index creation failed`,A(e))}try{let e=await t.curated.reindexAll();Q.info(`Curated re-index complete`,{indexed:e.indexed})}catch(e){Q.error(`Curated re-index failed`,A(e))}}catch(e){Q.error(`Initial index failed; will retry on aikit_reindex`,A(e))}finally{l?.setBusy(!1)}},w=new Hn,T=()=>w.schedule({name:`initial-index`,fn:C}),E=process.ppid,D=setInterval(()=>{try{process.kill(E,0)}catch{Q.info(`Parent process died; shutting down`,{parentPid:E}),clearInterval(D),v.then(async e=>{await Promise.all([e.embedder.shutdown().catch(()=>{}),e.graphStore.close().catch(()=>{}),e.store.close().catch(()=>{})])}).catch(()=>{}).finally(()=>process.exit(0))}},5e3);return D.unref(),{server:n,startInit:x,ready:S,runInitialIndex:T,get kb(){return s},scheduler:w,setSmartScheduler(e){u=e}}}export{$ as ALL_TOOL_NAMES,kl as createLazyServer,_l as createMcpServer,wl as createServer,gl as initializeKnowledgeBase,Cl as registerMcpTools};