skilld 1.5.1 → 1.5.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (47) hide show
  1. package/dist/_chunks/assemble.mjs +1 -0
  2. package/dist/_chunks/assemble.mjs.map +1 -1
  3. package/dist/_chunks/author.mjs +10 -10
  4. package/dist/_chunks/author.mjs.map +1 -1
  5. package/dist/_chunks/cache.mjs +4 -3
  6. package/dist/_chunks/cache.mjs.map +1 -1
  7. package/dist/_chunks/cache2.mjs +1 -0
  8. package/dist/_chunks/cache2.mjs.map +1 -1
  9. package/dist/_chunks/cli-helpers.mjs +26 -75
  10. package/dist/_chunks/cli-helpers.mjs.map +1 -1
  11. package/dist/_chunks/cli-helpers2.mjs +2 -1
  12. package/dist/_chunks/embedding-cache.mjs +1 -0
  13. package/dist/_chunks/index.d.mts.map +1 -1
  14. package/dist/_chunks/index3.d.mts.map +1 -1
  15. package/dist/_chunks/install.mjs +7 -6
  16. package/dist/_chunks/install.mjs.map +1 -1
  17. package/dist/_chunks/list.mjs +2 -1
  18. package/dist/_chunks/list.mjs.map +1 -1
  19. package/dist/_chunks/lockfile.mjs +14 -1
  20. package/dist/_chunks/lockfile.mjs.map +1 -1
  21. package/dist/_chunks/package-json.mjs +107 -0
  22. package/dist/_chunks/package-json.mjs.map +1 -0
  23. package/dist/_chunks/prepare2.mjs +2 -1
  24. package/dist/_chunks/prepare2.mjs.map +1 -1
  25. package/dist/_chunks/search-interactive.mjs +1 -0
  26. package/dist/_chunks/search-interactive.mjs.map +1 -1
  27. package/dist/_chunks/search.mjs +1 -0
  28. package/dist/_chunks/search2.mjs +1 -1
  29. package/dist/_chunks/setup.mjs +2 -1
  30. package/dist/_chunks/setup.mjs.map +1 -1
  31. package/dist/_chunks/sources.mjs +15 -18
  32. package/dist/_chunks/sources.mjs.map +1 -1
  33. package/dist/_chunks/sync-shared.mjs +1 -0
  34. package/dist/_chunks/sync-shared2.mjs +5 -4
  35. package/dist/_chunks/sync-shared2.mjs.map +1 -1
  36. package/dist/_chunks/sync.mjs +1 -1
  37. package/dist/_chunks/sync2.mjs +1 -0
  38. package/dist/_chunks/uninstall.mjs +2 -1
  39. package/dist/_chunks/uninstall.mjs.map +1 -1
  40. package/dist/_chunks/wizard.mjs +1 -1
  41. package/dist/agent/index.mjs +1 -0
  42. package/dist/cache/index.mjs +1 -0
  43. package/dist/cli.mjs +16 -5
  44. package/dist/cli.mjs.map +1 -1
  45. package/dist/index.mjs +1 -0
  46. package/dist/sources/index.mjs +1 -0
  47. package/package.json +1 -1
@@ -44,7 +44,14 @@ function parseSkillFrontmatter(skillPath) {
44
44
  for (const key of SKILL_FM_KEYS) if (fm[key]) info[key] = fm[key];
45
45
  return info;
46
46
  }
47
+ const lockCache = /* @__PURE__ */ new Map();
48
+ function invalidateLockCache(skillsDir) {
49
+ if (skillsDir) lockCache.delete(skillsDir);
50
+ else lockCache.clear();
51
+ }
47
52
  function readLock(skillsDir) {
53
+ const cached = lockCache.get(skillsDir);
54
+ if (cached) return cached;
48
55
  const lockPath = join(skillsDir, "skilld-lock.yaml");
49
56
  if (!existsSync(lockPath)) return null;
50
57
  const content = readFileSync(lockPath, "utf-8");
@@ -62,7 +69,9 @@ function readLock(skillsDir) {
62
69
  if (kv && isSkillInfoKey(kv[0])) skills[currentSkill][kv[0]] = kv[1];
63
70
  }
64
71
  }
65
- return { skills };
72
+ const lock = { skills };
73
+ lockCache.set(skillsDir, lock);
74
+ return { skills: { ...lock.skills } };
66
75
  }
67
76
  function serializeLock(lock) {
68
77
  let yaml = "skills:\n";
@@ -98,6 +107,7 @@ function writeLock(skillsDir, skillName, info) {
98
107
  }
99
108
  lock.skills[skillName] = info;
100
109
  writeFileSync(lockPath, serializeLock(lock));
110
+ invalidateLockCache(skillsDir);
101
111
  }
102
112
  /**
103
113
  * Merge multiple lockfiles, preferring the most recently synced entry per skill.
@@ -121,6 +131,7 @@ function syncLockfilesToDirs(sourceLock, dirs) {
121
131
  const existing = readLock(dir);
122
132
  if (!existing) continue;
123
133
  writeFileSync(lockPath, serializeLock(mergeLocks([existing, sourceLock])));
134
+ invalidateLockCache(dir);
124
135
  }
125
136
  }
126
137
  function removeLockEntry(skillsDir, skillName) {
@@ -130,9 +141,11 @@ function removeLockEntry(skillsDir, skillName) {
130
141
  delete lock.skills[skillName];
131
142
  if (Object.keys(lock.skills).length === 0) {
132
143
  unlinkSync(lockPath);
144
+ invalidateLockCache(skillsDir);
133
145
  return;
134
146
  }
135
147
  writeFileSync(lockPath, serializeLock(lock));
148
+ invalidateLockCache(skillsDir);
136
149
  }
137
150
  //#endregion
138
151
  export { removeLockEntry as a, readLock as i, parsePackages as n, syncLockfilesToDirs as o, parseSkillFrontmatter as r, writeLock as s, mergeLocks as t };
@@ -1 +1 @@
1
- {"version":3,"file":"lockfile.mjs","names":[],"sources":["../../src/core/lockfile.ts"],"sourcesContent":["import { existsSync, readFileSync, unlinkSync, writeFileSync } from 'node:fs'\nimport { join } from 'pathe'\nimport { parseFrontmatter } from './markdown.ts'\nimport { yamlEscape, yamlParseKV } from './yaml.ts'\n\nexport interface SkillInfo {\n packageName?: string\n version?: string\n /** All tracked packages as comma-separated \"name@version\" pairs (multi-package skills) */\n packages?: string\n repo?: string\n source?: string\n syncedAt?: string\n generator?: string\n /** Skill path within repo (git-sourced skills) */\n path?: string\n /** Git ref tracked for updates */\n ref?: string\n /** Git commit SHA at install time */\n commit?: string\n}\n\nexport function parsePackages(packages?: string): Array<{ name: string, version: string }> {\n if (!packages)\n return []\n return packages.split(',').map((s) => {\n const trimmed = s.trim()\n const atIdx = trimmed.lastIndexOf('@')\n if (atIdx <= 0)\n return { name: trimmed, version: '' }\n return { name: trimmed.slice(0, atIdx), version: trimmed.slice(atIdx + 1) }\n }).filter(p => p.name)\n}\n\nexport function serializePackages(pkgs: Array<{ name: string, version: string }>): string {\n return pkgs.map(p => `${p.name}@${p.version}`).join(', ')\n}\n\nexport interface SkilldLock {\n skills: Record<string, SkillInfo>\n}\n\nconst SKILL_FM_KEYS: (keyof SkillInfo)[] = ['packageName', 'version', 'packages', 'repo', 'source', 'syncedAt', 'generator', 'path', 'ref', 'commit']\n\nfunction isSkillInfoKey(key: string): key is keyof SkillInfo {\n return (SKILL_FM_KEYS as readonly string[]).includes(key)\n}\n\nexport function parseSkillFrontmatter(skillPath: string): SkillInfo | null {\n if (!existsSync(skillPath))\n return null\n const content = readFileSync(skillPath, 'utf-8')\n const fm = parseFrontmatter(content)\n if (Object.keys(fm).length === 0)\n return null\n\n const info: SkillInfo = {}\n for (const key of SKILL_FM_KEYS) {\n if (fm[key])\n info[key] = fm[key]\n }\n return info\n}\n\nexport function readLock(skillsDir: string): SkilldLock | null {\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n if (!existsSync(lockPath))\n return null\n const content = readFileSync(lockPath, 'utf-8')\n\n const skills: Record<string, SkillInfo> = {}\n let currentSkill: string | null = null\n\n for (const line of content.split('\\n')) {\n const skillMatch = line.match(/^ {2}(\\S+):$/)\n if (skillMatch) {\n currentSkill = skillMatch[1]!\n skills[currentSkill] = {}\n continue\n }\n if (currentSkill && line.startsWith(' ')) {\n const kv = yamlParseKV(line)\n if (kv && isSkillInfoKey(kv[0]))\n skills[currentSkill]![kv[0]] = kv[1]\n }\n }\n return { skills }\n}\n\nfunction serializeLock(lock: SkilldLock): string {\n let yaml = 'skills:\\n'\n for (const [name, skill] of Object.entries(lock.skills)) {\n yaml += ` ${name}:\\n`\n for (const key of SKILL_FM_KEYS) {\n if (skill[key])\n yaml += ` ${key}: ${yamlEscape(skill[key])}\\n`\n }\n }\n return yaml\n}\n\nexport function writeLock(skillsDir: string, skillName: string, info: SkillInfo): void {\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n let lock: SkilldLock = { skills: {} }\n if (existsSync(lockPath)) {\n lock = readLock(skillsDir) || { skills: {} }\n }\n\n const existing = lock.skills[skillName]\n if (existing && info.packageName) {\n // Merge packages list\n const existingPkgs = parsePackages(existing.packages)\n // Also include existing primary if not yet in packages list\n if (existing.packageName && !existingPkgs.some(p => p.name === existing.packageName)) {\n existingPkgs.unshift({ name: existing.packageName, version: existing.version || '' })\n }\n // Add/update new package\n const idx = existingPkgs.findIndex(p => p.name === info.packageName)\n if (idx >= 0) {\n existingPkgs[idx]!.version = info.version || ''\n }\n else {\n existingPkgs.push({ name: info.packageName, version: info.version || '' })\n }\n info.packages = serializePackages(existingPkgs)\n // Keep primary as first package\n info.packageName = existingPkgs[0]!.name\n info.version = existingPkgs[0]!.version\n // Preserve fields from existing entry that aren't in new info\n if (!info.repo && existing.repo)\n info.repo = existing.repo\n if (!info.source && existing.source)\n info.source = existing.source\n if (!info.generator && existing.generator)\n info.generator = existing.generator\n }\n\n lock.skills[skillName] = info\n writeFileSync(lockPath, serializeLock(lock))\n}\n\n/**\n * Merge multiple lockfiles, preferring the most recently synced entry per skill.\n */\nexport function mergeLocks(locks: SkilldLock[]): SkilldLock {\n const merged: Record<string, SkillInfo> = {}\n for (const lock of locks) {\n for (const [name, info] of Object.entries(lock.skills)) {\n const existing = merged[name]\n if (!existing || (info.syncedAt && (!existing.syncedAt || info.syncedAt > existing.syncedAt)))\n merged[name] = info\n }\n }\n return { skills: merged }\n}\n\n/**\n * Sync a lockfile to all other dirs that already have a skilld-lock.yaml.\n * Only updates existing lockfiles — does not create new ones.\n */\nexport function syncLockfilesToDirs(sourceLock: SkilldLock, dirs: string[]): void {\n for (const dir of dirs) {\n const lockPath = join(dir, 'skilld-lock.yaml')\n if (!existsSync(lockPath))\n continue\n const existing = readLock(dir)\n if (!existing)\n continue\n // Merge source into existing\n const merged = mergeLocks([existing, sourceLock])\n writeFileSync(lockPath, serializeLock(merged))\n }\n}\n\nexport function removeLockEntry(skillsDir: string, skillName: string): void {\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n const lock = readLock(skillsDir)\n if (!lock)\n return\n\n delete lock.skills[skillName]\n\n if (Object.keys(lock.skills).length === 0) {\n unlinkSync(lockPath)\n return\n }\n\n writeFileSync(lockPath, serializeLock(lock))\n}\n"],"mappings":";;;;;AAsBA,SAAgB,cAAc,UAA6D;AACzF,KAAI,CAAC,SACH,QAAO,EAAE;AACX,QAAO,SAAS,MAAM,IAAI,CAAC,KAAK,MAAM;EACpC,MAAM,UAAU,EAAE,MAAM;EACxB,MAAM,QAAQ,QAAQ,YAAY,IAAI;AACtC,MAAI,SAAS,EACX,QAAO;GAAE,MAAM;GAAS,SAAS;GAAI;AACvC,SAAO;GAAE,MAAM,QAAQ,MAAM,GAAG,MAAM;GAAE,SAAS,QAAQ,MAAM,QAAQ,EAAA;GAAI;GAC3E,CAAC,QAAO,MAAK,EAAE,KAAK;;AAGxB,SAAgB,kBAAkB,MAAwD;AACxF,QAAO,KAAK,KAAI,MAAK,GAAG,EAAE,KAAK,GAAG,EAAE,UAAU,CAAC,KAAK,KAAK;;AAO3D,MAAM,gBAAqC;CAAC;CAAe;CAAW;CAAY;CAAQ;CAAU;CAAY;CAAa;CAAQ;CAAO;CAAS;AAErJ,SAAS,eAAe,KAAqC;AAC3D,QAAQ,cAAoC,SAAS,IAAI;;AAG3D,SAAgB,sBAAsB,WAAqC;AACzE,KAAI,CAAC,WAAW,UAAU,CACxB,QAAO;CAET,MAAM,KAAK,iBADK,aAAa,WAAW,QAAQ,CACZ;AACpC,KAAI,OAAO,KAAK,GAAG,CAAC,WAAW,EAC7B,QAAO;CAET,MAAM,OAAkB,EAAE;AAC1B,MAAK,MAAM,OAAO,cAChB,KAAI,GAAG,KACL,MAAK,OAAO,GAAG;AAEnB,QAAO;;AAGT,SAAgB,SAAS,WAAsC;CAC7D,MAAM,WAAW,KAAK,WAAW,mBAAmB;AACpD,KAAI,CAAC,WAAW,SAAS,CACvB,QAAO;CACT,MAAM,UAAU,aAAa,UAAU,QAAQ;CAE/C,MAAM,SAAoC,EAAE;CAC5C,IAAI,eAA8B;AAElC,MAAK,MAAM,QAAQ,QAAQ,MAAM,KAAK,EAAE;EACtC,MAAM,aAAa,KAAK,MAAM,eAAe;AAC7C,MAAI,YAAY;AACd,kBAAe,WAAW;AAC1B,UAAO,gBAAgB,EAAE;AACzB;;AAEF,MAAI,gBAAgB,KAAK,WAAW,OAAO,EAAE;GAC3C,MAAM,KAAK,YAAY,KAAK;AAC5B,OAAI,MAAM,eAAe,GAAG,GAAG,CAC7B,QAAO,cAAe,GAAG,MAAM,GAAG;;;AAGxC,QAAO,EAAE,QAAQ;;AAGnB,SAAS,cAAc,MAA0B;CAC/C,IAAI,OAAO;AACX,MAAK,MAAM,CAAC,MAAM,UAAU,OAAO,QAAQ,KAAK,OAAO,EAAE;AACvD,UAAQ,KAAK,KAAK;AAClB,OAAK,MAAM,OAAO,cAChB,KAAI,MAAM,KACR,SAAQ,OAAO,IAAI,IAAI,WAAW,MAAM,KAAK,CAAC;;AAGpD,QAAO;;AAGT,SAAgB,UAAU,WAAmB,WAAmB,MAAuB;CACrF,MAAM,WAAW,KAAK,WAAW,mBAAmB;CACpD,IAAI,OAAmB,EAAE,QAAQ,EAAE,EAAE;AACrC,KAAI,WAAW,SAAS,CACtB,QAAO,SAAS,UAAU,IAAI,EAAE,QAAQ,EAAE,EAAE;CAG9C,MAAM,WAAW,KAAK,OAAO;AAC7B,KAAI,YAAY,KAAK,aAAa;EAEhC,MAAM,eAAe,cAAc,SAAS,SAAS;AAErD,MAAI,SAAS,eAAe,CAAC,aAAa,MAAK,MAAK,EAAE,SAAS,SAAS,YAAY,CAClF,cAAa,QAAQ;GAAE,MAAM,SAAS;GAAa,SAAS,SAAS,WAAW;GAAI,CAAC;EAGvF,MAAM,MAAM,aAAa,WAAU,MAAK,EAAE,SAAS,KAAK,YAAY;AACpE,MAAI,OAAO,EACT,cAAa,KAAM,UAAU,KAAK,WAAW;MAG7C,cAAa,KAAK;GAAE,MAAM,KAAK;GAAa,SAAS,KAAK,WAAW;GAAI,CAAC;AAE5E,OAAK,WAAW,kBAAkB,aAAa;AAE/C,OAAK,cAAc,aAAa,GAAI;AACpC,OAAK,UAAU,aAAa,GAAI;AAEhC,MAAI,CAAC,KAAK,QAAQ,SAAS,KACzB,MAAK,OAAO,SAAS;AACvB,MAAI,CAAC,KAAK,UAAU,SAAS,OAC3B,MAAK,SAAS,SAAS;AACzB,MAAI,CAAC,KAAK,aAAa,SAAS,UAC9B,MAAK,YAAY,SAAS;;AAG9B,MAAK,OAAO,aAAa;AACzB,eAAc,UAAU,cAAc,KAAK,CAAC;;;;;AAM9C,SAAgB,WAAW,OAAiC;CAC1D,MAAM,SAAoC,EAAE;AAC5C,MAAK,MAAM,QAAQ,MACjB,MAAK,MAAM,CAAC,MAAM,SAAS,OAAO,QAAQ,KAAK,OAAO,EAAE;EACtD,MAAM,WAAW,OAAO;AACxB,MAAI,CAAC,YAAa,KAAK,aAAa,CAAC,SAAS,YAAY,KAAK,WAAW,SAAS,UACjF,QAAO,QAAQ;;AAGrB,QAAO,EAAE,QAAQ,QAAQ;;;;;;AAO3B,SAAgB,oBAAoB,YAAwB,MAAsB;AAChF,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,WAAW,KAAK,KAAK,mBAAmB;AAC9C,MAAI,CAAC,WAAW,SAAS,CACvB;EACF,MAAM,WAAW,SAAS,IAAI;AAC9B,MAAI,CAAC,SACH;AAGF,gBAAc,UAAU,cADT,WAAW,CAAC,UAAU,WAAW,CAAC,CACJ,CAAC;;;AAIlD,SAAgB,gBAAgB,WAAmB,WAAyB;CAC1E,MAAM,WAAW,KAAK,WAAW,mBAAmB;CACpD,MAAM,OAAO,SAAS,UAAU;AAChC,KAAI,CAAC,KACH;AAEF,QAAO,KAAK,OAAO;AAEnB,KAAI,OAAO,KAAK,KAAK,OAAO,CAAC,WAAW,GAAG;AACzC,aAAW,SAAS;AACpB;;AAGF,eAAc,UAAU,cAAc,KAAK,CAAC"}
1
+ {"version":3,"file":"lockfile.mjs","names":[],"sources":["../../src/core/lockfile.ts"],"sourcesContent":["import { existsSync, readFileSync, unlinkSync, writeFileSync } from 'node:fs'\nimport { join } from 'pathe'\nimport { parseFrontmatter } from './markdown.ts'\nimport { yamlEscape, yamlParseKV } from './yaml.ts'\n\nexport interface SkillInfo {\n packageName?: string\n version?: string\n /** All tracked packages as comma-separated \"name@version\" pairs (multi-package skills) */\n packages?: string\n repo?: string\n source?: string\n syncedAt?: string\n generator?: string\n /** Skill path within repo (git-sourced skills) */\n path?: string\n /** Git ref tracked for updates */\n ref?: string\n /** Git commit SHA at install time */\n commit?: string\n}\n\nexport function parsePackages(packages?: string): Array<{ name: string, version: string }> {\n if (!packages)\n return []\n return packages.split(',').map((s) => {\n const trimmed = s.trim()\n const atIdx = trimmed.lastIndexOf('@')\n if (atIdx <= 0)\n return { name: trimmed, version: '' }\n return { name: trimmed.slice(0, atIdx), version: trimmed.slice(atIdx + 1) }\n }).filter(p => p.name)\n}\n\nexport function serializePackages(pkgs: Array<{ name: string, version: string }>): string {\n return pkgs.map(p => `${p.name}@${p.version}`).join(', ')\n}\n\nexport interface SkilldLock {\n skills: Record<string, SkillInfo>\n}\n\nconst SKILL_FM_KEYS: (keyof SkillInfo)[] = ['packageName', 'version', 'packages', 'repo', 'source', 'syncedAt', 'generator', 'path', 'ref', 'commit']\n\nfunction isSkillInfoKey(key: string): key is keyof SkillInfo {\n return (SKILL_FM_KEYS as readonly string[]).includes(key)\n}\n\nexport function parseSkillFrontmatter(skillPath: string): SkillInfo | null {\n if (!existsSync(skillPath))\n return null\n const content = readFileSync(skillPath, 'utf-8')\n const fm = parseFrontmatter(content)\n if (Object.keys(fm).length === 0)\n return null\n\n const info: SkillInfo = {}\n for (const key of SKILL_FM_KEYS) {\n if (fm[key])\n info[key] = fm[key]\n }\n return info\n}\n\nconst lockCache = new Map<string, SkilldLock>()\n\nexport function invalidateLockCache(skillsDir?: string): void {\n if (skillsDir)\n lockCache.delete(skillsDir)\n else\n lockCache.clear()\n}\n\nexport function readLock(skillsDir: string): SkilldLock | null {\n const cached = lockCache.get(skillsDir)\n if (cached)\n return cached\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n if (!existsSync(lockPath))\n return null\n const content = readFileSync(lockPath, 'utf-8')\n\n const skills: Record<string, SkillInfo> = {}\n let currentSkill: string | null = null\n\n for (const line of content.split('\\n')) {\n const skillMatch = line.match(/^ {2}(\\S+):$/)\n if (skillMatch) {\n currentSkill = skillMatch[1]!\n skills[currentSkill] = {}\n continue\n }\n if (currentSkill && line.startsWith(' ')) {\n const kv = yamlParseKV(line)\n if (kv && isSkillInfoKey(kv[0]))\n skills[currentSkill]![kv[0]] = kv[1]\n }\n }\n const lock = { skills }\n lockCache.set(skillsDir, lock)\n return { skills: { ...lock.skills } }\n}\n\nfunction serializeLock(lock: SkilldLock): string {\n let yaml = 'skills:\\n'\n for (const [name, skill] of Object.entries(lock.skills)) {\n yaml += ` ${name}:\\n`\n for (const key of SKILL_FM_KEYS) {\n if (skill[key])\n yaml += ` ${key}: ${yamlEscape(skill[key])}\\n`\n }\n }\n return yaml\n}\n\nexport function writeLock(skillsDir: string, skillName: string, info: SkillInfo): void {\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n let lock: SkilldLock = { skills: {} }\n if (existsSync(lockPath)) {\n lock = readLock(skillsDir) || { skills: {} }\n }\n\n const existing = lock.skills[skillName]\n if (existing && info.packageName) {\n // Merge packages list\n const existingPkgs = parsePackages(existing.packages)\n // Also include existing primary if not yet in packages list\n if (existing.packageName && !existingPkgs.some(p => p.name === existing.packageName)) {\n existingPkgs.unshift({ name: existing.packageName, version: existing.version || '' })\n }\n // Add/update new package\n const idx = existingPkgs.findIndex(p => p.name === info.packageName)\n if (idx >= 0) {\n existingPkgs[idx]!.version = info.version || ''\n }\n else {\n existingPkgs.push({ name: info.packageName, version: info.version || '' })\n }\n info.packages = serializePackages(existingPkgs)\n // Keep primary as first package\n info.packageName = existingPkgs[0]!.name\n info.version = existingPkgs[0]!.version\n // Preserve fields from existing entry that aren't in new info\n if (!info.repo && existing.repo)\n info.repo = existing.repo\n if (!info.source && existing.source)\n info.source = existing.source\n if (!info.generator && existing.generator)\n info.generator = existing.generator\n }\n\n lock.skills[skillName] = info\n writeFileSync(lockPath, serializeLock(lock))\n invalidateLockCache(skillsDir)\n}\n\n/**\n * Merge multiple lockfiles, preferring the most recently synced entry per skill.\n */\nexport function mergeLocks(locks: SkilldLock[]): SkilldLock {\n const merged: Record<string, SkillInfo> = {}\n for (const lock of locks) {\n for (const [name, info] of Object.entries(lock.skills)) {\n const existing = merged[name]\n if (!existing || (info.syncedAt && (!existing.syncedAt || info.syncedAt > existing.syncedAt)))\n merged[name] = info\n }\n }\n return { skills: merged }\n}\n\n/**\n * Sync a lockfile to all other dirs that already have a skilld-lock.yaml.\n * Only updates existing lockfiles — does not create new ones.\n */\nexport function syncLockfilesToDirs(sourceLock: SkilldLock, dirs: string[]): void {\n for (const dir of dirs) {\n const lockPath = join(dir, 'skilld-lock.yaml')\n if (!existsSync(lockPath))\n continue\n const existing = readLock(dir)\n if (!existing)\n continue\n // Merge source into existing\n const merged = mergeLocks([existing, sourceLock])\n writeFileSync(lockPath, serializeLock(merged))\n invalidateLockCache(dir)\n }\n}\n\nexport function removeLockEntry(skillsDir: string, skillName: string): void {\n const lockPath = join(skillsDir, 'skilld-lock.yaml')\n const lock = readLock(skillsDir)\n if (!lock)\n return\n\n delete lock.skills[skillName]\n\n if (Object.keys(lock.skills).length === 0) {\n unlinkSync(lockPath)\n invalidateLockCache(skillsDir)\n return\n }\n\n writeFileSync(lockPath, serializeLock(lock))\n invalidateLockCache(skillsDir)\n}\n"],"mappings":";;;;;AAsBA,SAAgB,cAAc,UAA6D;AACzF,KAAI,CAAC,SACH,QAAO,EAAE;AACX,QAAO,SAAS,MAAM,IAAI,CAAC,KAAK,MAAM;EACpC,MAAM,UAAU,EAAE,MAAM;EACxB,MAAM,QAAQ,QAAQ,YAAY,IAAI;AACtC,MAAI,SAAS,EACX,QAAO;GAAE,MAAM;GAAS,SAAS;GAAI;AACvC,SAAO;GAAE,MAAM,QAAQ,MAAM,GAAG,MAAM;GAAE,SAAS,QAAQ,MAAM,QAAQ,EAAA;GAAI;GAC3E,CAAC,QAAO,MAAK,EAAE,KAAK;;AAGxB,SAAgB,kBAAkB,MAAwD;AACxF,QAAO,KAAK,KAAI,MAAK,GAAG,EAAE,KAAK,GAAG,EAAE,UAAU,CAAC,KAAK,KAAK;;AAO3D,MAAM,gBAAqC;CAAC;CAAe;CAAW;CAAY;CAAQ;CAAU;CAAY;CAAa;CAAQ;CAAO;CAAS;AAErJ,SAAS,eAAe,KAAqC;AAC3D,QAAQ,cAAoC,SAAS,IAAI;;AAG3D,SAAgB,sBAAsB,WAAqC;AACzE,KAAI,CAAC,WAAW,UAAU,CACxB,QAAO;CAET,MAAM,KAAK,iBADK,aAAa,WAAW,QAAQ,CACZ;AACpC,KAAI,OAAO,KAAK,GAAG,CAAC,WAAW,EAC7B,QAAO;CAET,MAAM,OAAkB,EAAE;AAC1B,MAAK,MAAM,OAAO,cAChB,KAAI,GAAG,KACL,MAAK,OAAO,GAAG;AAEnB,QAAO;;AAGT,MAAM,4BAAY,IAAI,KAAyB;AAE/C,SAAgB,oBAAoB,WAA0B;AAC5D,KAAI,UACF,WAAU,OAAO,UAAU;KAE3B,WAAU,OAAO;;AAGrB,SAAgB,SAAS,WAAsC;CAC7D,MAAM,SAAS,UAAU,IAAI,UAAU;AACvC,KAAI,OACF,QAAO;CACT,MAAM,WAAW,KAAK,WAAW,mBAAmB;AACpD,KAAI,CAAC,WAAW,SAAS,CACvB,QAAO;CACT,MAAM,UAAU,aAAa,UAAU,QAAQ;CAE/C,MAAM,SAAoC,EAAE;CAC5C,IAAI,eAA8B;AAElC,MAAK,MAAM,QAAQ,QAAQ,MAAM,KAAK,EAAE;EACtC,MAAM,aAAa,KAAK,MAAM,eAAe;AAC7C,MAAI,YAAY;AACd,kBAAe,WAAW;AAC1B,UAAO,gBAAgB,EAAE;AACzB;;AAEF,MAAI,gBAAgB,KAAK,WAAW,OAAO,EAAE;GAC3C,MAAM,KAAK,YAAY,KAAK;AAC5B,OAAI,MAAM,eAAe,GAAG,GAAG,CAC7B,QAAO,cAAe,GAAG,MAAM,GAAG;;;CAGxC,MAAM,OAAO,EAAE,QAAQ;AACvB,WAAU,IAAI,WAAW,KAAK;AAC9B,QAAO,EAAE,QAAQ,EAAE,GAAG,KAAK,QAAQ,EAAE;;AAGvC,SAAS,cAAc,MAA0B;CAC/C,IAAI,OAAO;AACX,MAAK,MAAM,CAAC,MAAM,UAAU,OAAO,QAAQ,KAAK,OAAO,EAAE;AACvD,UAAQ,KAAK,KAAK;AAClB,OAAK,MAAM,OAAO,cAChB,KAAI,MAAM,KACR,SAAQ,OAAO,IAAI,IAAI,WAAW,MAAM,KAAK,CAAC;;AAGpD,QAAO;;AAGT,SAAgB,UAAU,WAAmB,WAAmB,MAAuB;CACrF,MAAM,WAAW,KAAK,WAAW,mBAAmB;CACpD,IAAI,OAAmB,EAAE,QAAQ,EAAE,EAAE;AACrC,KAAI,WAAW,SAAS,CACtB,QAAO,SAAS,UAAU,IAAI,EAAE,QAAQ,EAAE,EAAE;CAG9C,MAAM,WAAW,KAAK,OAAO;AAC7B,KAAI,YAAY,KAAK,aAAa;EAEhC,MAAM,eAAe,cAAc,SAAS,SAAS;AAErD,MAAI,SAAS,eAAe,CAAC,aAAa,MAAK,MAAK,EAAE,SAAS,SAAS,YAAY,CAClF,cAAa,QAAQ;GAAE,MAAM,SAAS;GAAa,SAAS,SAAS,WAAW;GAAI,CAAC;EAGvF,MAAM,MAAM,aAAa,WAAU,MAAK,EAAE,SAAS,KAAK,YAAY;AACpE,MAAI,OAAO,EACT,cAAa,KAAM,UAAU,KAAK,WAAW;MAG7C,cAAa,KAAK;GAAE,MAAM,KAAK;GAAa,SAAS,KAAK,WAAW;GAAI,CAAC;AAE5E,OAAK,WAAW,kBAAkB,aAAa;AAE/C,OAAK,cAAc,aAAa,GAAI;AACpC,OAAK,UAAU,aAAa,GAAI;AAEhC,MAAI,CAAC,KAAK,QAAQ,SAAS,KACzB,MAAK,OAAO,SAAS;AACvB,MAAI,CAAC,KAAK,UAAU,SAAS,OAC3B,MAAK,SAAS,SAAS;AACzB,MAAI,CAAC,KAAK,aAAa,SAAS,UAC9B,MAAK,YAAY,SAAS;;AAG9B,MAAK,OAAO,aAAa;AACzB,eAAc,UAAU,cAAc,KAAK,CAAC;AAC5C,qBAAoB,UAAU;;;;;AAMhC,SAAgB,WAAW,OAAiC;CAC1D,MAAM,SAAoC,EAAE;AAC5C,MAAK,MAAM,QAAQ,MACjB,MAAK,MAAM,CAAC,MAAM,SAAS,OAAO,QAAQ,KAAK,OAAO,EAAE;EACtD,MAAM,WAAW,OAAO;AACxB,MAAI,CAAC,YAAa,KAAK,aAAa,CAAC,SAAS,YAAY,KAAK,WAAW,SAAS,UACjF,QAAO,QAAQ;;AAGrB,QAAO,EAAE,QAAQ,QAAQ;;;;;;AAO3B,SAAgB,oBAAoB,YAAwB,MAAsB;AAChF,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,WAAW,KAAK,KAAK,mBAAmB;AAC9C,MAAI,CAAC,WAAW,SAAS,CACvB;EACF,MAAM,WAAW,SAAS,IAAI;AAC9B,MAAI,CAAC,SACH;AAGF,gBAAc,UAAU,cADT,WAAW,CAAC,UAAU,WAAW,CAAC,CACJ,CAAC;AAC9C,sBAAoB,IAAI;;;AAI5B,SAAgB,gBAAgB,WAAmB,WAAyB;CAC1E,MAAM,WAAW,KAAK,WAAW,mBAAmB;CACpD,MAAM,OAAO,SAAS,UAAU;AAChC,KAAI,CAAC,KACH;AAEF,QAAO,KAAK,OAAO;AAEnB,KAAI,OAAO,KAAK,KAAK,OAAO,CAAC,WAAW,GAAG;AACzC,aAAW,SAAS;AACpB,sBAAoB,UAAU;AAC9B;;AAGF,eAAc,UAAU,cAAc,KAAK,CAAC;AAC5C,qBAAoB,UAAU"}
@@ -0,0 +1,107 @@
1
+ import { existsSync, readFileSync, writeFileSync } from "node:fs";
2
+ import { applyEdits, modify, parseTree } from "jsonc-parser";
3
+ //#region src/core/package-json.ts
4
+ const defaultEditOptions = {
5
+ tabSize: 2,
6
+ insertSpaces: true
7
+ };
8
+ const cache = /* @__PURE__ */ new Map();
9
+ /**
10
+ * Read and parse a package.json, returning cached result on repeat calls.
11
+ * Throws if the file does not exist.
12
+ */
13
+ function readPackageJson(pkgPath) {
14
+ const hit = cache.get(pkgPath);
15
+ if (hit) return hit;
16
+ const raw = readFileSync(pkgPath, "utf-8");
17
+ const entry = {
18
+ raw,
19
+ parsed: JSON.parse(raw)
20
+ };
21
+ cache.set(pkgPath, entry);
22
+ return entry;
23
+ }
24
+ /**
25
+ * Same as readPackageJson but returns null when the file is missing or unparseable.
26
+ */
27
+ function readPackageJsonSafe(pkgPath) {
28
+ if (cache.has(pkgPath)) return cache.get(pkgPath);
29
+ if (!existsSync(pkgPath)) return null;
30
+ try {
31
+ return readPackageJson(pkgPath);
32
+ } catch {
33
+ return null;
34
+ }
35
+ }
36
+ /**
37
+ * Drop any cached entry so the next read hits disk.
38
+ */
39
+ function invalidatePackageJson(pkgPath) {
40
+ cache.delete(pkgPath);
41
+ }
42
+ /**
43
+ * Set a value at a JSON path, preserving all surrounding formatting.
44
+ * Returns the modified file content as a string.
45
+ */
46
+ function editJsonProperty(raw, path, value, options) {
47
+ const opts = {
48
+ ...defaultEditOptions,
49
+ ...options
50
+ };
51
+ return applyEdits(raw, modify(raw, path, value, { formattingOptions: {
52
+ tabSize: opts.tabSize,
53
+ insertSpaces: opts.insertSpaces
54
+ } }));
55
+ }
56
+ /**
57
+ * Read a package.json, apply an edit function, write it back, and invalidate the cache.
58
+ * The edit function receives the raw text and parsed object,
59
+ * and returns the new raw text (or null to skip writing).
60
+ */
61
+ function patchPackageJson(pkgPath, editFn) {
62
+ const { raw, parsed } = readPackageJson(pkgPath);
63
+ const result = editFn(raw, parsed);
64
+ if (result === null) return false;
65
+ writeFileSync(pkgPath, result);
66
+ invalidatePackageJson(pkgPath);
67
+ return true;
68
+ }
69
+ /**
70
+ * Append a value to a JSON array at the given path, preserving formatting.
71
+ * Inserts in sorted order if the array contains strings.
72
+ */
73
+ function appendToJsonArray(raw, path, value, options) {
74
+ const opts = {
75
+ ...defaultEditOptions,
76
+ ...options
77
+ };
78
+ const tree = parseTree(raw);
79
+ if (!tree) return editJsonProperty(raw, path, [value], opts);
80
+ let node = tree;
81
+ for (const key of path) {
82
+ const child = node.children?.find((c) => c.type === "property" && c.children?.[0]?.value === key);
83
+ if (!child?.children?.[1]) return editJsonProperty(raw, path, [value], opts);
84
+ node = child.children[1];
85
+ }
86
+ if (node.type !== "array" || !node.children) return editJsonProperty(raw, path, [value], opts);
87
+ const allStrings = node.children.every((c) => typeof c.value === "string");
88
+ let idx = node.children.length;
89
+ if (allStrings) {
90
+ const items = node.children.map((c) => c.value);
91
+ for (let i = 0; i < items.length; i++) if (value.localeCompare(items[i]) < 0) {
92
+ idx = i;
93
+ break;
94
+ }
95
+ }
96
+ return applyEdits(raw, modify(raw, [...path, idx], value, {
97
+ formattingOptions: {
98
+ tabSize: opts.tabSize,
99
+ insertSpaces: opts.insertSpaces
100
+ },
101
+ isArrayInsertion: true
102
+ }));
103
+ }
104
+ //#endregion
105
+ export { readPackageJsonSafe as i, editJsonProperty as n, patchPackageJson as r, appendToJsonArray as t };
106
+
107
+ //# sourceMappingURL=package-json.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"package-json.mjs","names":[],"sources":["../../src/core/package-json.ts"],"sourcesContent":["import { existsSync, readFileSync, writeFileSync } from 'node:fs'\nimport { applyEdits, modify, parseTree } from 'jsonc-parser'\n\nexport interface EditOptions {\n /** Formatting options for inserted content */\n tabSize?: number\n insertSpaces?: boolean\n}\n\nconst defaultEditOptions: EditOptions = { tabSize: 2, insertSpaces: true }\n\n// ── Cached reader ──────────────────────────────────────────────\n\nconst cache = new Map<string, { raw: string, parsed: Record<string, unknown> }>()\n\n/**\n * Read and parse a package.json, returning cached result on repeat calls.\n * Throws if the file does not exist.\n */\nexport function readPackageJson(pkgPath: string): { raw: string, parsed: Record<string, unknown> } {\n const hit = cache.get(pkgPath)\n if (hit)\n return hit\n const raw = readFileSync(pkgPath, 'utf-8')\n const parsed = JSON.parse(raw) as Record<string, unknown>\n const entry = { raw, parsed }\n cache.set(pkgPath, entry)\n return entry\n}\n\n/**\n * Same as readPackageJson but returns null when the file is missing or unparseable.\n */\nexport function readPackageJsonSafe(pkgPath: string): { raw: string, parsed: Record<string, unknown> } | null {\n if (cache.has(pkgPath))\n return cache.get(pkgPath)!\n if (!existsSync(pkgPath))\n return null\n try {\n return readPackageJson(pkgPath)\n }\n catch {\n return null\n }\n}\n\n/**\n * Drop any cached entry so the next read hits disk.\n */\nexport function invalidatePackageJson(pkgPath: string): void {\n cache.delete(pkgPath)\n}\n\n/**\n * Clear all cached entries. Useful in tests.\n */\nexport function clearPackageJsonCache(): void {\n cache.clear()\n}\n\n// ── JSON editing helpers ───────────────────────────────────────\n\n/**\n * Set a value at a JSON path, preserving all surrounding formatting.\n * Returns the modified file content as a string.\n */\nexport function editJsonProperty(raw: string, path: (string | number)[], value: unknown, options?: EditOptions): string {\n const opts = { ...defaultEditOptions, ...options }\n const edits = modify(raw, path, value, {\n formattingOptions: { tabSize: opts.tabSize!, insertSpaces: opts.insertSpaces! },\n })\n return applyEdits(raw, edits)\n}\n\n/**\n * Remove a value at a JSON path, preserving all surrounding formatting.\n */\nexport function removeJsonProperty(raw: string, path: (string | number)[]): string {\n const edits = modify(raw, path, undefined, {})\n return applyEdits(raw, edits)\n}\n\n/**\n * Read a package.json, apply an edit function, write it back, and invalidate the cache.\n * The edit function receives the raw text and parsed object,\n * and returns the new raw text (or null to skip writing).\n */\nexport function patchPackageJson(\n pkgPath: string,\n editFn: (raw: string, pkg: Record<string, unknown>) => string | null,\n): boolean {\n const { raw, parsed } = readPackageJson(pkgPath)\n const result = editFn(raw, parsed)\n if (result === null)\n return false\n writeFileSync(pkgPath, result)\n invalidatePackageJson(pkgPath)\n return true\n}\n\n/**\n * Append a value to a JSON array at the given path, preserving formatting.\n * Inserts in sorted order if the array contains strings.\n */\nexport function appendToJsonArray(raw: string, path: string[], value: string, options?: EditOptions): string {\n const opts = { ...defaultEditOptions, ...options }\n const tree = parseTree(raw)\n if (!tree)\n return editJsonProperty(raw, path, [value], opts)\n\n // Walk to the target array node\n let node = tree\n for (const key of path) {\n const child = node.children?.find(c =>\n c.type === 'property' && c.children?.[0]?.value === key,\n )\n if (!child?.children?.[1])\n return editJsonProperty(raw, path, [value], opts)\n node = child.children[1]\n }\n\n if (node.type !== 'array' || !node.children)\n return editJsonProperty(raw, path, [value], opts)\n\n // Find sorted insertion index (only for string-only arrays)\n const allStrings = node.children.every(c => typeof c.value === 'string')\n let idx = node.children.length\n if (allStrings) {\n const items = node.children.map(c => c.value as string)\n for (let i = 0; i < items.length; i++) {\n if (value.localeCompare(items[i]!) < 0) {\n idx = i\n break\n }\n }\n }\n\n const edits = modify(raw, [...path, idx], value, {\n formattingOptions: { tabSize: opts.tabSize!, insertSpaces: opts.insertSpaces! },\n isArrayInsertion: true,\n })\n return applyEdits(raw, edits)\n}\n"],"mappings":";;;AASA,MAAM,qBAAkC;CAAE,SAAS;CAAG,cAAc;CAAM;AAI1E,MAAM,wBAAQ,IAAI,KAA+D;;;;;AAMjF,SAAgB,gBAAgB,SAAmE;CACjG,MAAM,MAAM,MAAM,IAAI,QAAQ;AAC9B,KAAI,IACF,QAAO;CACT,MAAM,MAAM,aAAa,SAAS,QAAQ;CAE1C,MAAM,QAAQ;EAAE;EAAK,QADN,KAAK,MAAM,IAAA;EACG;AAC7B,OAAM,IAAI,SAAS,MAAM;AACzB,QAAO;;;;;AAMT,SAAgB,oBAAoB,SAA0E;AAC5G,KAAI,MAAM,IAAI,QAAQ,CACpB,QAAO,MAAM,IAAI,QAAQ;AAC3B,KAAI,CAAC,WAAW,QAAQ,CACtB,QAAO;AACT,KAAI;AACF,SAAO,gBAAgB,QAAQ;SAE3B;AACJ,SAAO;;;;;;AAOX,SAAgB,sBAAsB,SAAuB;AAC3D,OAAM,OAAO,QAAQ;;;;;;AAgBvB,SAAgB,iBAAiB,KAAa,MAA2B,OAAgB,SAA+B;CACtH,MAAM,OAAO;EAAE,GAAG;EAAoB,GAAG;EAAS;AAIlD,QAAO,WAAW,KAHJ,OAAO,KAAK,MAAM,OAAO,EACrC,mBAAmB;EAAE,SAAS,KAAK;EAAU,cAAc,KAAK;EAAe,EAChF,CAAC,CAC2B;;;;;;;AAgB/B,SAAgB,iBACd,SACA,QACS;CACT,MAAM,EAAE,KAAK,WAAW,gBAAgB,QAAQ;CAChD,MAAM,SAAS,OAAO,KAAK,OAAO;AAClC,KAAI,WAAW,KACb,QAAO;AACT,eAAc,SAAS,OAAO;AAC9B,uBAAsB,QAAQ;AAC9B,QAAO;;;;;;AAOT,SAAgB,kBAAkB,KAAa,MAAgB,OAAe,SAA+B;CAC3G,MAAM,OAAO;EAAE,GAAG;EAAoB,GAAG;EAAS;CAClD,MAAM,OAAO,UAAU,IAAI;AAC3B,KAAI,CAAC,KACH,QAAO,iBAAiB,KAAK,MAAM,CAAC,MAAM,EAAE,KAAK;CAGnD,IAAI,OAAO;AACX,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,QAAQ,KAAK,UAAU,MAAK,MAChC,EAAE,SAAS,cAAc,EAAE,WAAW,IAAI,UAAU,IACrD;AACD,MAAI,CAAC,OAAO,WAAW,GACrB,QAAO,iBAAiB,KAAK,MAAM,CAAC,MAAM,EAAE,KAAK;AACnD,SAAO,MAAM,SAAS;;AAGxB,KAAI,KAAK,SAAS,WAAW,CAAC,KAAK,SACjC,QAAO,iBAAiB,KAAK,MAAM,CAAC,MAAM,EAAE,KAAK;CAGnD,MAAM,aAAa,KAAK,SAAS,OAAM,MAAK,OAAO,EAAE,UAAU,SAAS;CACxE,IAAI,MAAM,KAAK,SAAS;AACxB,KAAI,YAAY;EACd,MAAM,QAAQ,KAAK,SAAS,KAAI,MAAK,EAAE,MAAgB;AACvD,OAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,IAChC,KAAI,MAAM,cAAc,MAAM,GAAI,GAAG,GAAG;AACtC,SAAM;AACN;;;AASN,QAAO,WAAW,KAJJ,OAAO,KAAK,CAAC,GAAG,MAAM,IAAI,EAAE,OAAO;EAC/C,mBAAmB;GAAE,SAAS,KAAK;GAAU,cAAc,KAAK;GAAe;EAC/E,kBAAkB;EACnB,CAAC,CAC2B"}
@@ -1,5 +1,6 @@
1
1
  import "./agent.mjs";
2
2
  import "./config.mjs";
3
+ import "./package-json.mjs";
3
4
  import { i as restorePkgSymlink, n as linkShippedSkill, t as getShippedSkills } from "./prepare.mjs";
4
5
  import "./sanitize.mjs";
5
6
  import "./cache.mjs";
@@ -9,7 +10,7 @@ import { n as getSharedSkillsDir } from "./shared.mjs";
9
10
  import "./sources.mjs";
10
11
  import { a as targets } from "./detect.mjs";
11
12
  import { i as linkSkillToAgents } from "./prompts.mjs";
12
- import { y as resolveAgent } from "./cli-helpers.mjs";
13
+ import { b as resolveAgent } from "./cli-helpers.mjs";
13
14
  import { i as readLock, s as writeLock } from "./lockfile.mjs";
14
15
  import { t as getProjectState } from "./skills.mjs";
15
16
  import { join } from "pathe";
@@ -1 +1 @@
1
- {"version":3,"file":"prepare2.mjs","names":["agents"],"sources":["../../src/commands/prepare.ts"],"sourcesContent":["/**\n * Prepare command — lightweight hook for package.json \"prepare\" script.\n *\n * Designed to run on every `pnpm install` / `npm install`. Blocking, fast, no LLM calls.\n * 1. Restore broken symlinks from lockfile (like `install` but skips doc fetching)\n * 2. Auto-install shipped skills from deps (just symlinks + lockfile writes)\n * 3. Report outdated skills count and suggest `skilld update`\n */\n\nimport { existsSync, mkdirSync } from 'node:fs'\nimport * as p from '@clack/prompts'\nimport { defineCommand } from 'citty'\nimport { join } from 'pathe'\nimport { agents, linkSkillToAgents } from '../agent/index.ts'\nimport { resolveAgent } from '../cli-helpers.ts'\nimport { readLock, writeLock } from '../core/lockfile.ts'\nimport { getShippedSkills, linkShippedSkill, restorePkgSymlink } from '../core/prepare.ts'\nimport { getSharedSkillsDir } from '../core/shared.ts'\nimport { getProjectState } from '../core/skills.ts'\n\nexport const prepareCommandDef = defineCommand({\n meta: { name: 'prepare', description: 'Restore references and sync shipped skills (for package.json hooks)' },\n args: {\n agent: {\n type: 'enum' as const,\n options: Object.keys(agents),\n alias: 'a',\n description: 'Target agent',\n },\n },\n async run({ args }) {\n const cwd = process.cwd()\n\n const agent = resolveAgent(args.agent)\n if (!agent || agent === 'none')\n return\n\n const agentConfig = agents[agent]\n const shared = getSharedSkillsDir(cwd)\n const skillsDir = shared || join(cwd, agentConfig.skillsDir)\n\n // ── Fast path: read primary lockfile, check all skills intact ──\n\n const lock = readLock(skillsDir)\n if (lock && Object.keys(lock.skills).length > 0) {\n let allIntact = true\n\n for (const [name, info] of Object.entries(lock.skills)) {\n if (!info.version)\n continue\n\n const skillDir = join(skillsDir, name)\n if (existsSync(skillDir)) {\n // Skill dir exists; for non-shipped, also check .skilld/pkg symlink\n if (info.source !== 'shipped')\n restorePkgSymlink(skillsDir, name, info, cwd)\n continue\n }\n\n // Skill dir missing, needs restore\n allIntact = false\n\n if (info.source === 'shipped') {\n const pkgName = info.packageName || name\n const shipped = getShippedSkills(pkgName, cwd, info.version)\n const match = shipped.find(s => s.skillName === name)\n if (match)\n linkShippedSkill(skillsDir, name, match.skillDir)\n }\n }\n\n // If all skills intact, skip expensive getProjectState entirely\n if (allIntact)\n return\n }\n\n // ── Slow path: discover new shipped skills + report outdated ──\n\n const state = await getProjectState(cwd)\n let shippedCount = 0\n\n if (state.shipped.length > 0) {\n mkdirSync(skillsDir, { recursive: true })\n\n for (const entry of state.shipped) {\n const version = state.deps.get(entry.packageName)?.replace(/^[\\^~>=<]+/, '') || '0.0.0'\n\n for (const skill of entry.skills) {\n linkShippedSkill(skillsDir, skill.skillName, skill.skillDir)\n writeLock(skillsDir, skill.skillName, {\n packageName: entry.packageName,\n version,\n source: 'shipped',\n syncedAt: new Date().toISOString().split('T')[0],\n generator: 'skilld',\n })\n\n if (shared)\n linkSkillToAgents(skill.skillName, shared, cwd, agent)\n\n shippedCount++\n }\n }\n\n if (shippedCount > 0)\n p.log.success(`Installed ${shippedCount} shipped skill${shippedCount > 1 ? 's' : ''}`)\n }\n\n if (state.outdated.length > 0) {\n const n = state.outdated.length\n p.log.info(`${n} package${n > 1 ? 's' : ''} ha${n > 1 ? 've' : 's'} new features and/or breaking changes. Run \\`skilld update\\` to sync.`)\n }\n },\n})\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;AAoBA,MAAa,oBAAoB,cAAc;CAC7C,MAAM;EAAE,MAAM;EAAW,aAAa;EAAuE;CAC7G,MAAM,EACJ,OAAO;EACL,MAAM;EACN,SAAS,OAAO,KAAKA,QAAO;EAC5B,OAAO;EACP,aAAa;EACd,EACF;CACD,MAAM,IAAI,EAAE,QAAQ;EAClB,MAAM,MAAM,QAAQ,KAAK;EAEzB,MAAM,QAAQ,aAAa,KAAK,MAAM;AACtC,MAAI,CAAC,SAAS,UAAU,OACtB;EAEF,MAAM,cAAcA,QAAO;EAC3B,MAAM,SAAS,mBAAmB,IAAI;EACtC,MAAM,YAAY,UAAU,KAAK,KAAK,YAAY,UAAU;EAI5D,MAAM,OAAO,SAAS,UAAU;AAChC,MAAI,QAAQ,OAAO,KAAK,KAAK,OAAO,CAAC,SAAS,GAAG;GAC/C,IAAI,YAAY;AAEhB,QAAK,MAAM,CAAC,MAAM,SAAS,OAAO,QAAQ,KAAK,OAAO,EAAE;AACtD,QAAI,CAAC,KAAK,QACR;AAGF,QAAI,WADa,KAAK,WAAW,KAAK,CACd,EAAE;AAExB,SAAI,KAAK,WAAW,UAClB,mBAAkB,WAAW,MAAM,MAAM,IAAI;AAC/C;;AAIF,gBAAY;AAEZ,QAAI,KAAK,WAAW,WAAW;KAG7B,MAAM,QADU,iBADA,KAAK,eAAe,MACM,KAAK,KAAK,QAAQ,CACtC,MAAK,MAAK,EAAE,cAAc,KAAK;AACrD,SAAI,MACF,kBAAiB,WAAW,MAAM,MAAM,SAAS;;;AAKvD,OAAI,UACF;;EAKJ,MAAM,QAAQ,MAAM,gBAAgB,IAAI;EACxC,IAAI,eAAe;AAEnB,MAAI,MAAM,QAAQ,SAAS,GAAG;AAC5B,aAAU,WAAW,EAAE,WAAW,MAAM,CAAC;AAEzC,QAAK,MAAM,SAAS,MAAM,SAAS;IACjC,MAAM,UAAU,MAAM,KAAK,IAAI,MAAM,YAAY,EAAE,QAAQ,cAAc,GAAG,IAAI;AAEhF,SAAK,MAAM,SAAS,MAAM,QAAQ;AAChC,sBAAiB,WAAW,MAAM,WAAW,MAAM,SAAS;AAC5D,eAAU,WAAW,MAAM,WAAW;MACpC,aAAa,MAAM;MACnB;MACA,QAAQ;MACR,2BAAU,IAAI,MAAM,EAAC,aAAa,CAAC,MAAM,IAAI,CAAC;MAC9C,WAAW;MACZ,CAAC;AAEF,SAAI,OACF,mBAAkB,MAAM,WAAW,QAAQ,KAAK,MAAM;AAExD;;;AAIJ,OAAI,eAAe,EACjB,GAAE,IAAI,QAAQ,aAAa,aAAa,gBAAgB,eAAe,IAAI,MAAM,KAAK;;AAG1F,MAAI,MAAM,SAAS,SAAS,GAAG;GAC7B,MAAM,IAAI,MAAM,SAAS;AACzB,KAAE,IAAI,KAAK,GAAG,EAAE,UAAU,IAAI,IAAI,MAAM,GAAG,KAAK,IAAI,IAAI,OAAO,IAAI,uEAAuE;;;CAG/I,CAAC"}
1
+ {"version":3,"file":"prepare2.mjs","names":["agents"],"sources":["../../src/commands/prepare.ts"],"sourcesContent":["/**\n * Prepare command — lightweight hook for package.json \"prepare\" script.\n *\n * Designed to run on every `pnpm install` / `npm install`. Blocking, fast, no LLM calls.\n * 1. Restore broken symlinks from lockfile (like `install` but skips doc fetching)\n * 2. Auto-install shipped skills from deps (just symlinks + lockfile writes)\n * 3. Report outdated skills count and suggest `skilld update`\n */\n\nimport { existsSync, mkdirSync } from 'node:fs'\nimport * as p from '@clack/prompts'\nimport { defineCommand } from 'citty'\nimport { join } from 'pathe'\nimport { agents, linkSkillToAgents } from '../agent/index.ts'\nimport { resolveAgent } from '../cli-helpers.ts'\nimport { readLock, writeLock } from '../core/lockfile.ts'\nimport { getShippedSkills, linkShippedSkill, restorePkgSymlink } from '../core/prepare.ts'\nimport { getSharedSkillsDir } from '../core/shared.ts'\nimport { getProjectState } from '../core/skills.ts'\n\nexport const prepareCommandDef = defineCommand({\n meta: { name: 'prepare', description: 'Restore references and sync shipped skills (for package.json hooks)' },\n args: {\n agent: {\n type: 'enum' as const,\n options: Object.keys(agents),\n alias: 'a',\n description: 'Target agent',\n },\n },\n async run({ args }) {\n const cwd = process.cwd()\n\n const agent = resolveAgent(args.agent)\n if (!agent || agent === 'none')\n return\n\n const agentConfig = agents[agent]\n const shared = getSharedSkillsDir(cwd)\n const skillsDir = shared || join(cwd, agentConfig.skillsDir)\n\n // ── Fast path: read primary lockfile, check all skills intact ──\n\n const lock = readLock(skillsDir)\n if (lock && Object.keys(lock.skills).length > 0) {\n let allIntact = true\n\n for (const [name, info] of Object.entries(lock.skills)) {\n if (!info.version)\n continue\n\n const skillDir = join(skillsDir, name)\n if (existsSync(skillDir)) {\n // Skill dir exists; for non-shipped, also check .skilld/pkg symlink\n if (info.source !== 'shipped')\n restorePkgSymlink(skillsDir, name, info, cwd)\n continue\n }\n\n // Skill dir missing, needs restore\n allIntact = false\n\n if (info.source === 'shipped') {\n const pkgName = info.packageName || name\n const shipped = getShippedSkills(pkgName, cwd, info.version)\n const match = shipped.find(s => s.skillName === name)\n if (match)\n linkShippedSkill(skillsDir, name, match.skillDir)\n }\n }\n\n // If all skills intact, skip expensive getProjectState entirely\n if (allIntact)\n return\n }\n\n // ── Slow path: discover new shipped skills + report outdated ──\n\n const state = await getProjectState(cwd)\n let shippedCount = 0\n\n if (state.shipped.length > 0) {\n mkdirSync(skillsDir, { recursive: true })\n\n for (const entry of state.shipped) {\n const version = state.deps.get(entry.packageName)?.replace(/^[\\^~>=<]+/, '') || '0.0.0'\n\n for (const skill of entry.skills) {\n linkShippedSkill(skillsDir, skill.skillName, skill.skillDir)\n writeLock(skillsDir, skill.skillName, {\n packageName: entry.packageName,\n version,\n source: 'shipped',\n syncedAt: new Date().toISOString().split('T')[0],\n generator: 'skilld',\n })\n\n if (shared)\n linkSkillToAgents(skill.skillName, shared, cwd, agent)\n\n shippedCount++\n }\n }\n\n if (shippedCount > 0)\n p.log.success(`Installed ${shippedCount} shipped skill${shippedCount > 1 ? 's' : ''}`)\n }\n\n if (state.outdated.length > 0) {\n const n = state.outdated.length\n p.log.info(`${n} package${n > 1 ? 's' : ''} ha${n > 1 ? 've' : 's'} new features and/or breaking changes. Run \\`skilld update\\` to sync.`)\n }\n },\n})\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAoBA,MAAa,oBAAoB,cAAc;CAC7C,MAAM;EAAE,MAAM;EAAW,aAAa;EAAuE;CAC7G,MAAM,EACJ,OAAO;EACL,MAAM;EACN,SAAS,OAAO,KAAKA,QAAO;EAC5B,OAAO;EACP,aAAa;EACd,EACF;CACD,MAAM,IAAI,EAAE,QAAQ;EAClB,MAAM,MAAM,QAAQ,KAAK;EAEzB,MAAM,QAAQ,aAAa,KAAK,MAAM;AACtC,MAAI,CAAC,SAAS,UAAU,OACtB;EAEF,MAAM,cAAcA,QAAO;EAC3B,MAAM,SAAS,mBAAmB,IAAI;EACtC,MAAM,YAAY,UAAU,KAAK,KAAK,YAAY,UAAU;EAI5D,MAAM,OAAO,SAAS,UAAU;AAChC,MAAI,QAAQ,OAAO,KAAK,KAAK,OAAO,CAAC,SAAS,GAAG;GAC/C,IAAI,YAAY;AAEhB,QAAK,MAAM,CAAC,MAAM,SAAS,OAAO,QAAQ,KAAK,OAAO,EAAE;AACtD,QAAI,CAAC,KAAK,QACR;AAGF,QAAI,WADa,KAAK,WAAW,KAAK,CACd,EAAE;AAExB,SAAI,KAAK,WAAW,UAClB,mBAAkB,WAAW,MAAM,MAAM,IAAI;AAC/C;;AAIF,gBAAY;AAEZ,QAAI,KAAK,WAAW,WAAW;KAG7B,MAAM,QADU,iBADA,KAAK,eAAe,MACM,KAAK,KAAK,QAAQ,CACtC,MAAK,MAAK,EAAE,cAAc,KAAK;AACrD,SAAI,MACF,kBAAiB,WAAW,MAAM,MAAM,SAAS;;;AAKvD,OAAI,UACF;;EAKJ,MAAM,QAAQ,MAAM,gBAAgB,IAAI;EACxC,IAAI,eAAe;AAEnB,MAAI,MAAM,QAAQ,SAAS,GAAG;AAC5B,aAAU,WAAW,EAAE,WAAW,MAAM,CAAC;AAEzC,QAAK,MAAM,SAAS,MAAM,SAAS;IACjC,MAAM,UAAU,MAAM,KAAK,IAAI,MAAM,YAAY,EAAE,QAAQ,cAAc,GAAG,IAAI;AAEhF,SAAK,MAAM,SAAS,MAAM,QAAQ;AAChC,sBAAiB,WAAW,MAAM,WAAW,MAAM,SAAS;AAC5D,eAAU,WAAW,MAAM,WAAW;MACpC,aAAa,MAAM;MACnB;MACA,QAAQ;MACR,2BAAU,IAAI,MAAM,EAAC,aAAa,CAAC,MAAM,IAAI,CAAC;MAC9C,WAAW;MACZ,CAAC;AAEF,SAAI,OACF,mBAAkB,MAAM,WAAW,QAAQ,KAAK,MAAM;AAExD;;;AAIJ,OAAI,eAAe,EACjB,GAAE,IAAI,QAAQ,aAAa,aAAa,gBAAgB,eAAe,IAAI,MAAM,KAAK;;AAG1F,MAAI,MAAM,SAAS,SAAS,GAAG;GAC7B,MAAM,IAAI,MAAM,SAAS;AACzB,KAAE,IAAI,KAAK,GAAG,EAAE,UAAU,IAAI,IAAI,MAAM,GAAG,KAAK,IAAI,IAAI,OAAO,IAAI,uEAAuE;;;CAG/I,CAAC"}
@@ -1,5 +1,6 @@
1
1
  import "./agent.mjs";
2
2
  import "./config.mjs";
3
+ import "./package-json.mjs";
3
4
  import "./prepare.mjs";
4
5
  import { n as sanitizeMarkdown } from "./sanitize.mjs";
5
6
  import "./cache.mjs";
@@ -1 +1 @@
1
- {"version":3,"file":"search-interactive.mjs","names":[],"sources":["../../src/commands/search-interactive.ts"],"sourcesContent":["import type { SearchFilter, SearchSnippet } from '../retriv/index.ts'\nimport { createLogUpdate } from 'log-update'\nimport { formatCompactSnippet, highlightTerms, normalizeScores, sanitizeMarkdown, scoreLabel } from '../core/index.ts'\nimport { closePool, openPool, SearchDepsUnavailableError, searchPooled } from '../retriv/index.ts'\nimport { findPackageDbs, getPackageVersions, listLockPackages, parseFilterPrefix } from './search.ts'\n\nconst FILTER_CYCLE = [undefined, 'docs', 'issues', 'releases'] as const\ntype FilterLabel = typeof FILTER_CYCLE[number]\n\nfunction filterToSearchFilter(label: FilterLabel): SearchFilter | undefined {\n if (!label)\n return undefined\n if (label === 'issues')\n return { type: 'issue' }\n if (label === 'releases')\n return { type: 'release' }\n return { type: { $in: ['doc', 'docs'] } }\n}\n\nconst SPINNER_FRAMES = ['◐', '◓', '◑', '◒']\n\nexport async function interactiveSearch(packageFilter?: string): Promise<void> {\n const dbs = findPackageDbs(packageFilter)\n const versions = getPackageVersions()\n if (dbs.length === 0) {\n let msg: string\n if (packageFilter) {\n const available = listLockPackages()\n msg = available.length > 0\n ? `No docs indexed for \"${packageFilter}\". Available: ${available.join(', ')}`\n : `No docs indexed for \"${packageFilter}\". Run \\`skilld add ${packageFilter}\\` first.`\n }\n else {\n msg = 'No docs indexed yet. Run `skilld add <package>` first.'\n }\n process.stderr.write(`\\x1B[33m${msg}\\x1B[0m\\n`)\n return\n }\n\n const logUpdate = createLogUpdate(process.stderr, { showCursor: true })\n let pool: Awaited<ReturnType<typeof openPool>>\n try {\n pool = await openPool(dbs)\n }\n catch (err) {\n if (err instanceof SearchDepsUnavailableError) {\n process.stderr.write('\\x1B[31mSearch requires native dependencies (sqlite-vec) that are not installed.\\nInstall skilld globally or in a project to use search: npm i -g skilld\\x1B[0m\\n')\n return\n }\n throw err\n }\n\n // State\n let query = ''\n let results: SearchSnippet[] = []\n let selectedIndex = 0\n let isSearching = false\n let searchId = 0\n let filterIndex = 0\n let error = ''\n let elapsed = 0\n let spinFrame = 0\n let debounceTimer: ReturnType<typeof setTimeout> | null = null\n\n const cols = process.stdout.columns || 80\n const maxResults = 7\n const titleLabel = packageFilter ? `Search ${packageFilter} docs` : 'Search docs'\n\n function getFilterLabel(): string {\n const f = FILTER_CYCLE[filterIndex]\n if (!f)\n return ''\n return `\\x1B[36m${f}:\\x1B[0m`\n }\n\n function render() {\n const lines: string[] = []\n\n // Title\n lines.push('')\n lines.push(` \\x1B[1m${titleLabel}\\x1B[0m`)\n lines.push('')\n\n // Input line\n const filterPrefix = getFilterLabel()\n const prefix = filterPrefix ? `${filterPrefix}` : ''\n lines.push(` \\x1B[36m❯\\x1B[0m ${prefix}${query}\\x1B[7m \\x1B[0m`)\n\n // Separator / spinner\n if (isSearching) {\n const frame = SPINNER_FRAMES[spinFrame % SPINNER_FRAMES.length]\n lines.push(` \\x1B[36m${frame}\\x1B[0m \\x1B[90mSearching…\\x1B[0m`)\n }\n else {\n lines.push(` \\x1B[90m${'─'.repeat(Math.min(cols - 4, 40))}\\x1B[0m`)\n }\n\n // Results or empty state\n if (error) {\n lines.push('')\n lines.push(` \\x1B[31m${error}\\x1B[0m`)\n }\n else if (query.length === 0) {\n lines.push('')\n lines.push(' \\x1B[90mType to search…\\x1B[0m')\n }\n else if (query.length < 2 && !isSearching) {\n lines.push('')\n lines.push(' \\x1B[90mKeep typing…\\x1B[0m')\n }\n else if (results.length === 0 && !isSearching) {\n lines.push('')\n lines.push(' \\x1B[90mNo results\\x1B[0m')\n }\n else {\n lines.push('')\n const shown = results.slice(0, maxResults)\n const scores = normalizeScores(results)\n for (let i = 0; i < shown.length; i++) {\n const r = shown[i]!\n const selected = i === selectedIndex\n const bullet = selected ? '\\x1B[36m●\\x1B[0m' : '\\x1B[90m○\\x1B[0m'\n const sc = scoreLabel(scores.get(r) ?? 0)\n const { title, path, preview } = formatCompactSnippet(r, cols)\n const highlighted = highlightTerms(preview, r.highlights)\n\n const ver = versions.get(r.package)\n const pkgLabel = ver ? `${r.package}@${ver}` : r.package\n\n if (selected) {\n lines.push(` ${bullet} \\x1B[1m${pkgLabel}\\x1B[0m ${sc} \\x1B[36m${title}\\x1B[0m`)\n lines.push(` \\x1B[90m${path}\\x1B[0m`)\n lines.push(` ${highlighted}`)\n }\n else {\n lines.push(` ${bullet} \\x1B[90m${pkgLabel}\\x1B[0m ${sc} \\x1B[90m${title}\\x1B[0m`)\n }\n }\n }\n\n // Footer\n lines.push('')\n const parts: string[] = []\n if (results.length > 0)\n parts.push(`${results.length} results`)\n if (elapsed > 0 && !isSearching)\n parts.push(`${elapsed.toFixed(2)}s`)\n const footer = parts.length > 0 ? `${parts.join(' · ')} ` : ''\n lines.push(` \\x1B[90m${footer}↑↓ navigate ↵ select tab filter esc quit\\x1B[0m`)\n lines.push('')\n\n logUpdate(lines.join('\\n'))\n }\n\n async function doSearch() {\n const id = ++searchId\n const fullQuery = query.trim()\n if (fullQuery.length < 2) {\n results = []\n isSearching = false\n render()\n return\n }\n\n isSearching = true\n error = ''\n render()\n\n // Spin animation\n const spinInterval = setInterval(() => {\n spinFrame++\n if (isSearching)\n render()\n }, 80)\n\n const { query: parsed, filter: parsedFilter } = parseFilterPrefix(fullQuery)\n const filter = parsedFilter || filterToSearchFilter(FILTER_CYCLE[filterIndex])\n const start = performance.now()\n\n const res = await searchPooled(parsed, pool, { limit: maxResults, filter }).catch((e) => {\n if (id === searchId)\n error = e instanceof Error ? e.message : String(e)\n return [] as SearchSnippet[]\n })\n\n clearInterval(spinInterval)\n\n // Discard stale results\n if (id !== searchId)\n return\n\n results = res\n elapsed = (performance.now() - start) / 1000\n selectedIndex = 0\n isSearching = false\n render()\n }\n\n function scheduleSearch() {\n if (debounceTimer)\n clearTimeout(debounceTimer)\n debounceTimer = setTimeout(doSearch, 100)\n }\n\n // Show initial state\n render()\n\n // Raw stdin for keystroke handling\n const { stdin } = process\n if (stdin.isTTY)\n stdin.setRawMode(true)\n stdin.resume()\n stdin.setEncoding('utf-8')\n\n return new Promise<void>((resolve) => {\n function cleanup() {\n if (debounceTimer)\n clearTimeout(debounceTimer)\n if (stdin.isTTY)\n stdin.setRawMode(false)\n stdin.removeListener('data', onData)\n stdin.pause()\n closePool(pool)\n }\n\n function exit() {\n cleanup()\n logUpdate.done()\n resolve()\n }\n\n function selectResult() {\n if (results.length === 0 || selectedIndex >= results.length)\n return\n const r = results[selectedIndex]!\n cleanup()\n logUpdate.done()\n\n // Print full result\n const refPath = `.claude/skills/${r.package}/.skilld/${r.source}`\n const lineRange = r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`\n const highlighted = highlightTerms(sanitizeMarkdown(r.content), r.highlights)\n const rVer = versions.get(r.package)\n const rLabel = rVer ? `${r.package}@${rVer}` : r.package\n const rScores = normalizeScores(results)\n const out = [\n '',\n ` \\x1B[1m${rLabel}\\x1B[0m ${scoreLabel(rScores.get(r) ?? 0)}`,\n ` \\x1B[90m${refPath}:${lineRange}\\x1B[0m`,\n '',\n ` ${highlighted.replace(/\\n/g, '\\n ')}`,\n '',\n ].join('\\n')\n process.stdout.write(`${out}\\n`)\n resolve()\n }\n\n function onData(data: string) {\n // Ctrl+C\n if (data === '\\x03') {\n exit()\n return\n }\n\n // Escape\n if (data === '\\x1B' || data === '\\x1B\\x1B') {\n exit()\n return\n }\n\n // Enter\n if (data === '\\r' || data === '\\n') {\n selectResult()\n return\n }\n\n // Tab — cycle filter\n if (data === '\\t') {\n filterIndex = (filterIndex + 1) % FILTER_CYCLE.length\n if (query.length >= 2)\n scheduleSearch()\n render()\n return\n }\n\n // Backspace\n if (data === '\\x7F' || data === '\\b') {\n if (query.length > 0) {\n query = query.slice(0, -1)\n scheduleSearch()\n render()\n }\n return\n }\n\n // Arrow keys (escape sequences)\n if (data === '\\x1B[A' || data === '\\x1BOA') {\n // Up\n if (selectedIndex > 0) {\n selectedIndex--\n render()\n }\n return\n }\n if (data === '\\x1B[B' || data === '\\x1BOB') {\n // Down\n if (selectedIndex < results.length - 1) {\n selectedIndex++\n render()\n }\n return\n }\n\n // Ignore other escape sequences\n if (data.startsWith('\\x1B'))\n return\n\n // Printable characters\n query += data\n scheduleSearch()\n render()\n }\n\n stdin.on('data', onData)\n })\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAMA,MAAM,eAAe;CAAC,KAAA;CAAW;CAAQ;CAAU;CAAW;AAG9D,SAAS,qBAAqB,OAA8C;AAC1E,KAAI,CAAC,MACH,QAAO,KAAA;AACT,KAAI,UAAU,SACZ,QAAO,EAAE,MAAM,SAAS;AAC1B,KAAI,UAAU,WACZ,QAAO,EAAE,MAAM,WAAW;AAC5B,QAAO,EAAE,MAAM,EAAE,KAAK,CAAC,OAAO,OAAO,EAAE,EAAE;;AAG3C,MAAM,iBAAiB;CAAC;CAAK;CAAK;CAAK;CAAI;AAE3C,eAAsB,kBAAkB,eAAuC;CAC7E,MAAM,MAAM,eAAe,cAAc;CACzC,MAAM,WAAW,oBAAoB;AACrC,KAAI,IAAI,WAAW,GAAG;EACpB,IAAI;AACJ,MAAI,eAAe;GACjB,MAAM,YAAY,kBAAkB;AACpC,SAAM,UAAU,SAAS,IACrB,wBAAwB,cAAc,gBAAgB,UAAU,KAAK,KAAK,KAC1E,wBAAwB,cAAc,sBAAsB,cAAc;QAG9E,OAAM;AAER,UAAQ,OAAO,MAAM,WAAW,IAAI,WAAW;AAC/C;;CAGF,MAAM,YAAY,gBAAgB,QAAQ,QAAQ,EAAE,YAAY,MAAM,CAAC;CACvE,IAAI;AACJ,KAAI;AACF,SAAO,MAAM,SAAS,IAAI;UAErB,KAAK;AACV,MAAI,eAAe,4BAA4B;AAC7C,WAAQ,OAAO,MAAM,oKAAoK;AACzL;;AAEF,QAAM;;CAIR,IAAI,QAAQ;CACZ,IAAI,UAA2B,EAAE;CACjC,IAAI,gBAAgB;CACpB,IAAI,cAAc;CAClB,IAAI,WAAW;CACf,IAAI,cAAc;CAClB,IAAI,QAAQ;CACZ,IAAI,UAAU;CACd,IAAI,YAAY;CAChB,IAAI,gBAAsD;CAE1D,MAAM,OAAO,QAAQ,OAAO,WAAW;CACvC,MAAM,aAAa;CACnB,MAAM,aAAa,gBAAgB,UAAU,cAAc,SAAS;CAEpE,SAAS,iBAAyB;EAChC,MAAM,IAAI,aAAa;AACvB,MAAI,CAAC,EACH,QAAO;AACT,SAAO,WAAW,EAAE;;CAGtB,SAAS,SAAS;EAChB,MAAM,QAAkB,EAAE;AAG1B,QAAM,KAAK,GAAG;AACd,QAAM,KAAK,YAAY,WAAW,SAAS;AAC3C,QAAM,KAAK,GAAG;EAGd,MAAM,eAAe,gBAAgB;EACrC,MAAM,SAAS,eAAe,GAAG,iBAAiB;AAClD,QAAM,KAAK,sBAAsB,SAAS,MAAM,iBAAiB;AAGjE,MAAI,aAAa;GACf,MAAM,QAAQ,eAAe,YAAY,eAAe;AACxD,SAAM,KAAK,aAAa,MAAM,mCAAmC;QAGjE,OAAM,KAAK,aAAa,IAAI,OAAO,KAAK,IAAI,OAAO,GAAG,GAAG,CAAC,CAAC,SAAS;AAItE,MAAI,OAAO;AACT,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,aAAa,MAAM,SAAS;aAEhC,MAAM,WAAW,GAAG;AAC3B,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,mCAAmC;aAEvC,MAAM,SAAS,KAAK,CAAC,aAAa;AACzC,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,gCAAgC;aAEpC,QAAQ,WAAW,KAAK,CAAC,aAAa;AAC7C,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,8BAA8B;SAEtC;AACH,SAAM,KAAK,GAAG;GACd,MAAM,QAAQ,QAAQ,MAAM,GAAG,WAAW;GAC1C,MAAM,SAAS,gBAAgB,QAAQ;AACvC,QAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;IACrC,MAAM,IAAI,MAAM;IAChB,MAAM,WAAW,MAAM;IACvB,MAAM,SAAS,WAAW,qBAAqB;IAC/C,MAAM,KAAK,WAAW,OAAO,IAAI,EAAE,IAAI,EAAE;IACzC,MAAM,EAAE,OAAO,MAAM,YAAY,qBAAqB,GAAG,KAAK;IAC9D,MAAM,cAAc,eAAe,SAAS,EAAE,WAAW;IAEzD,MAAM,MAAM,SAAS,IAAI,EAAE,QAAQ;IACnC,MAAM,WAAW,MAAM,GAAG,EAAE,QAAQ,GAAG,QAAQ,EAAE;AAEjD,QAAI,UAAU;AACZ,WAAM,KAAK,KAAK,OAAO,UAAU,SAAS,UAAU,GAAG,YAAY,MAAM,SAAS;AAClF,WAAM,KAAK,eAAe,KAAK,SAAS;AACxC,WAAM,KAAK,OAAO,cAAc;UAGhC,OAAM,KAAK,KAAK,OAAO,WAAW,SAAS,UAAU,GAAG,YAAY,MAAM,SAAS;;;AAMzF,QAAM,KAAK,GAAG;EACd,MAAM,QAAkB,EAAE;AAC1B,MAAI,QAAQ,SAAS,EACnB,OAAM,KAAK,GAAG,QAAQ,OAAO,UAAU;AACzC,MAAI,UAAU,KAAK,CAAC,YAClB,OAAM,KAAK,GAAG,QAAQ,QAAQ,EAAE,CAAC,GAAG;EACtC,MAAM,SAAS,MAAM,SAAS,IAAI,GAAG,MAAM,KAAK,MAAM,CAAC,QAAQ;AAC/D,QAAM,KAAK,aAAa,OAAO,oDAAoD;AACnF,QAAM,KAAK,GAAG;AAEd,YAAU,MAAM,KAAK,KAAK,CAAC;;CAG7B,eAAe,WAAW;EACxB,MAAM,KAAK,EAAE;EACb,MAAM,YAAY,MAAM,MAAM;AAC9B,MAAI,UAAU,SAAS,GAAG;AACxB,aAAU,EAAE;AACZ,iBAAc;AACd,WAAQ;AACR;;AAGF,gBAAc;AACd,UAAQ;AACR,UAAQ;EAGR,MAAM,eAAe,kBAAkB;AACrC;AACA,OAAI,YACF,SAAQ;KACT,GAAG;EAEN,MAAM,EAAE,OAAO,QAAQ,QAAQ,iBAAiB,kBAAkB,UAAU;EAC5E,MAAM,SAAS,gBAAgB,qBAAqB,aAAa,aAAa;EAC9E,MAAM,QAAQ,YAAY,KAAK;EAE/B,MAAM,MAAM,MAAM,aAAa,QAAQ,MAAM;GAAE,OAAO;GAAY;GAAQ,CAAC,CAAC,OAAO,MAAM;AACvF,OAAI,OAAO,SACT,SAAQ,aAAa,QAAQ,EAAE,UAAU,OAAO,EAAE;AACpD,UAAO,EAAE;IACT;AAEF,gBAAc,aAAa;AAG3B,MAAI,OAAO,SACT;AAEF,YAAU;AACV,aAAW,YAAY,KAAK,GAAG,SAAS;AACxC,kBAAgB;AAChB,gBAAc;AACd,UAAQ;;CAGV,SAAS,iBAAiB;AACxB,MAAI,cACF,cAAa,cAAc;AAC7B,kBAAgB,WAAW,UAAU,IAAI;;AAI3C,SAAQ;CAGR,MAAM,EAAE,UAAU;AAClB,KAAI,MAAM,MACR,OAAM,WAAW,KAAK;AACxB,OAAM,QAAQ;AACd,OAAM,YAAY,QAAQ;AAE1B,QAAO,IAAI,SAAe,YAAY;EACpC,SAAS,UAAU;AACjB,OAAI,cACF,cAAa,cAAc;AAC7B,OAAI,MAAM,MACR,OAAM,WAAW,MAAM;AACzB,SAAM,eAAe,QAAQ,OAAO;AACpC,SAAM,OAAO;AACb,aAAU,KAAK;;EAGjB,SAAS,OAAO;AACd,YAAS;AACT,aAAU,MAAM;AAChB,YAAS;;EAGX,SAAS,eAAe;AACtB,OAAI,QAAQ,WAAW,KAAK,iBAAiB,QAAQ,OACnD;GACF,MAAM,IAAI,QAAQ;AAClB,YAAS;AACT,aAAU,MAAM;GAGhB,MAAM,UAAU,kBAAkB,EAAE,QAAQ,WAAW,EAAE;GACzD,MAAM,YAAY,EAAE,cAAc,EAAE,UAAU,IAAI,EAAE,cAAc,IAAI,EAAE,UAAU,GAAG,EAAE;GACvF,MAAM,cAAc,eAAe,iBAAiB,EAAE,QAAQ,EAAE,EAAE,WAAW;GAC7E,MAAM,OAAO,SAAS,IAAI,EAAE,QAAQ;GAGpC,MAAM,MAAM;IACV;IACA,YAJa,OAAO,GAAG,EAAE,QAAQ,GAAG,SAAS,EAAE,QAI5B,UAAU,WAHf,gBAAgB,QAAQ,CAGU,IAAI,EAAE,IAAI,EAAE;IAC5D,aAAa,QAAQ,GAAG,UAAU;IAClC;IACA,KAAK,YAAY,QAAQ,OAAO,OAAO;IACvC;IACD,CAAC,KAAK,KAAK;AACZ,WAAQ,OAAO,MAAM,GAAG,IAAI,IAAI;AAChC,YAAS;;EAGX,SAAS,OAAO,MAAc;AAE5B,OAAI,SAAS,KAAQ;AACnB,UAAM;AACN;;AAIF,OAAI,SAAS,UAAU,SAAS,YAAY;AAC1C,UAAM;AACN;;AAIF,OAAI,SAAS,QAAQ,SAAS,MAAM;AAClC,kBAAc;AACd;;AAIF,OAAI,SAAS,KAAM;AACjB,mBAAe,cAAc,KAAK,aAAa;AAC/C,QAAI,MAAM,UAAU,EAClB,iBAAgB;AAClB,YAAQ;AACR;;AAIF,OAAI,SAAS,OAAU,SAAS,MAAM;AACpC,QAAI,MAAM,SAAS,GAAG;AACpB,aAAQ,MAAM,MAAM,GAAG,GAAG;AAC1B,qBAAgB;AAChB,aAAQ;;AAEV;;AAIF,OAAI,SAAS,YAAY,SAAS,UAAU;AAE1C,QAAI,gBAAgB,GAAG;AACrB;AACA,aAAQ;;AAEV;;AAEF,OAAI,SAAS,YAAY,SAAS,UAAU;AAE1C,QAAI,gBAAgB,QAAQ,SAAS,GAAG;AACtC;AACA,aAAQ;;AAEV;;AAIF,OAAI,KAAK,WAAW,OAAO,CACzB;AAGF,YAAS;AACT,mBAAgB;AAChB,WAAQ;;AAGV,QAAM,GAAG,QAAQ,OAAO;GACxB"}
1
+ {"version":3,"file":"search-interactive.mjs","names":[],"sources":["../../src/commands/search-interactive.ts"],"sourcesContent":["import type { SearchFilter, SearchSnippet } from '../retriv/index.ts'\nimport { createLogUpdate } from 'log-update'\nimport { formatCompactSnippet, highlightTerms, normalizeScores, sanitizeMarkdown, scoreLabel } from '../core/index.ts'\nimport { closePool, openPool, SearchDepsUnavailableError, searchPooled } from '../retriv/index.ts'\nimport { findPackageDbs, getPackageVersions, listLockPackages, parseFilterPrefix } from './search.ts'\n\nconst FILTER_CYCLE = [undefined, 'docs', 'issues', 'releases'] as const\ntype FilterLabel = typeof FILTER_CYCLE[number]\n\nfunction filterToSearchFilter(label: FilterLabel): SearchFilter | undefined {\n if (!label)\n return undefined\n if (label === 'issues')\n return { type: 'issue' }\n if (label === 'releases')\n return { type: 'release' }\n return { type: { $in: ['doc', 'docs'] } }\n}\n\nconst SPINNER_FRAMES = ['◐', '◓', '◑', '◒']\n\nexport async function interactiveSearch(packageFilter?: string): Promise<void> {\n const dbs = findPackageDbs(packageFilter)\n const versions = getPackageVersions()\n if (dbs.length === 0) {\n let msg: string\n if (packageFilter) {\n const available = listLockPackages()\n msg = available.length > 0\n ? `No docs indexed for \"${packageFilter}\". Available: ${available.join(', ')}`\n : `No docs indexed for \"${packageFilter}\". Run \\`skilld add ${packageFilter}\\` first.`\n }\n else {\n msg = 'No docs indexed yet. Run `skilld add <package>` first.'\n }\n process.stderr.write(`\\x1B[33m${msg}\\x1B[0m\\n`)\n return\n }\n\n const logUpdate = createLogUpdate(process.stderr, { showCursor: true })\n let pool: Awaited<ReturnType<typeof openPool>>\n try {\n pool = await openPool(dbs)\n }\n catch (err) {\n if (err instanceof SearchDepsUnavailableError) {\n process.stderr.write('\\x1B[31mSearch requires native dependencies (sqlite-vec) that are not installed.\\nInstall skilld globally or in a project to use search: npm i -g skilld\\x1B[0m\\n')\n return\n }\n throw err\n }\n\n // State\n let query = ''\n let results: SearchSnippet[] = []\n let selectedIndex = 0\n let isSearching = false\n let searchId = 0\n let filterIndex = 0\n let error = ''\n let elapsed = 0\n let spinFrame = 0\n let debounceTimer: ReturnType<typeof setTimeout> | null = null\n\n const cols = process.stdout.columns || 80\n const maxResults = 7\n const titleLabel = packageFilter ? `Search ${packageFilter} docs` : 'Search docs'\n\n function getFilterLabel(): string {\n const f = FILTER_CYCLE[filterIndex]\n if (!f)\n return ''\n return `\\x1B[36m${f}:\\x1B[0m`\n }\n\n function render() {\n const lines: string[] = []\n\n // Title\n lines.push('')\n lines.push(` \\x1B[1m${titleLabel}\\x1B[0m`)\n lines.push('')\n\n // Input line\n const filterPrefix = getFilterLabel()\n const prefix = filterPrefix ? `${filterPrefix}` : ''\n lines.push(` \\x1B[36m❯\\x1B[0m ${prefix}${query}\\x1B[7m \\x1B[0m`)\n\n // Separator / spinner\n if (isSearching) {\n const frame = SPINNER_FRAMES[spinFrame % SPINNER_FRAMES.length]\n lines.push(` \\x1B[36m${frame}\\x1B[0m \\x1B[90mSearching…\\x1B[0m`)\n }\n else {\n lines.push(` \\x1B[90m${'─'.repeat(Math.min(cols - 4, 40))}\\x1B[0m`)\n }\n\n // Results or empty state\n if (error) {\n lines.push('')\n lines.push(` \\x1B[31m${error}\\x1B[0m`)\n }\n else if (query.length === 0) {\n lines.push('')\n lines.push(' \\x1B[90mType to search…\\x1B[0m')\n }\n else if (query.length < 2 && !isSearching) {\n lines.push('')\n lines.push(' \\x1B[90mKeep typing…\\x1B[0m')\n }\n else if (results.length === 0 && !isSearching) {\n lines.push('')\n lines.push(' \\x1B[90mNo results\\x1B[0m')\n }\n else {\n lines.push('')\n const shown = results.slice(0, maxResults)\n const scores = normalizeScores(results)\n for (let i = 0; i < shown.length; i++) {\n const r = shown[i]!\n const selected = i === selectedIndex\n const bullet = selected ? '\\x1B[36m●\\x1B[0m' : '\\x1B[90m○\\x1B[0m'\n const sc = scoreLabel(scores.get(r) ?? 0)\n const { title, path, preview } = formatCompactSnippet(r, cols)\n const highlighted = highlightTerms(preview, r.highlights)\n\n const ver = versions.get(r.package)\n const pkgLabel = ver ? `${r.package}@${ver}` : r.package\n\n if (selected) {\n lines.push(` ${bullet} \\x1B[1m${pkgLabel}\\x1B[0m ${sc} \\x1B[36m${title}\\x1B[0m`)\n lines.push(` \\x1B[90m${path}\\x1B[0m`)\n lines.push(` ${highlighted}`)\n }\n else {\n lines.push(` ${bullet} \\x1B[90m${pkgLabel}\\x1B[0m ${sc} \\x1B[90m${title}\\x1B[0m`)\n }\n }\n }\n\n // Footer\n lines.push('')\n const parts: string[] = []\n if (results.length > 0)\n parts.push(`${results.length} results`)\n if (elapsed > 0 && !isSearching)\n parts.push(`${elapsed.toFixed(2)}s`)\n const footer = parts.length > 0 ? `${parts.join(' · ')} ` : ''\n lines.push(` \\x1B[90m${footer}↑↓ navigate ↵ select tab filter esc quit\\x1B[0m`)\n lines.push('')\n\n logUpdate(lines.join('\\n'))\n }\n\n async function doSearch() {\n const id = ++searchId\n const fullQuery = query.trim()\n if (fullQuery.length < 2) {\n results = []\n isSearching = false\n render()\n return\n }\n\n isSearching = true\n error = ''\n render()\n\n // Spin animation\n const spinInterval = setInterval(() => {\n spinFrame++\n if (isSearching)\n render()\n }, 80)\n\n const { query: parsed, filter: parsedFilter } = parseFilterPrefix(fullQuery)\n const filter = parsedFilter || filterToSearchFilter(FILTER_CYCLE[filterIndex])\n const start = performance.now()\n\n const res = await searchPooled(parsed, pool, { limit: maxResults, filter }).catch((e) => {\n if (id === searchId)\n error = e instanceof Error ? e.message : String(e)\n return [] as SearchSnippet[]\n })\n\n clearInterval(spinInterval)\n\n // Discard stale results\n if (id !== searchId)\n return\n\n results = res\n elapsed = (performance.now() - start) / 1000\n selectedIndex = 0\n isSearching = false\n render()\n }\n\n function scheduleSearch() {\n if (debounceTimer)\n clearTimeout(debounceTimer)\n debounceTimer = setTimeout(doSearch, 100)\n }\n\n // Show initial state\n render()\n\n // Raw stdin for keystroke handling\n const { stdin } = process\n if (stdin.isTTY)\n stdin.setRawMode(true)\n stdin.resume()\n stdin.setEncoding('utf-8')\n\n return new Promise<void>((resolve) => {\n function cleanup() {\n if (debounceTimer)\n clearTimeout(debounceTimer)\n if (stdin.isTTY)\n stdin.setRawMode(false)\n stdin.removeListener('data', onData)\n stdin.pause()\n closePool(pool)\n }\n\n function exit() {\n cleanup()\n logUpdate.done()\n resolve()\n }\n\n function selectResult() {\n if (results.length === 0 || selectedIndex >= results.length)\n return\n const r = results[selectedIndex]!\n cleanup()\n logUpdate.done()\n\n // Print full result\n const refPath = `.claude/skills/${r.package}/.skilld/${r.source}`\n const lineRange = r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`\n const highlighted = highlightTerms(sanitizeMarkdown(r.content), r.highlights)\n const rVer = versions.get(r.package)\n const rLabel = rVer ? `${r.package}@${rVer}` : r.package\n const rScores = normalizeScores(results)\n const out = [\n '',\n ` \\x1B[1m${rLabel}\\x1B[0m ${scoreLabel(rScores.get(r) ?? 0)}`,\n ` \\x1B[90m${refPath}:${lineRange}\\x1B[0m`,\n '',\n ` ${highlighted.replace(/\\n/g, '\\n ')}`,\n '',\n ].join('\\n')\n process.stdout.write(`${out}\\n`)\n resolve()\n }\n\n function onData(data: string) {\n // Ctrl+C\n if (data === '\\x03') {\n exit()\n return\n }\n\n // Escape\n if (data === '\\x1B' || data === '\\x1B\\x1B') {\n exit()\n return\n }\n\n // Enter\n if (data === '\\r' || data === '\\n') {\n selectResult()\n return\n }\n\n // Tab — cycle filter\n if (data === '\\t') {\n filterIndex = (filterIndex + 1) % FILTER_CYCLE.length\n if (query.length >= 2)\n scheduleSearch()\n render()\n return\n }\n\n // Backspace\n if (data === '\\x7F' || data === '\\b') {\n if (query.length > 0) {\n query = query.slice(0, -1)\n scheduleSearch()\n render()\n }\n return\n }\n\n // Arrow keys (escape sequences)\n if (data === '\\x1B[A' || data === '\\x1BOA') {\n // Up\n if (selectedIndex > 0) {\n selectedIndex--\n render()\n }\n return\n }\n if (data === '\\x1B[B' || data === '\\x1BOB') {\n // Down\n if (selectedIndex < results.length - 1) {\n selectedIndex++\n render()\n }\n return\n }\n\n // Ignore other escape sequences\n if (data.startsWith('\\x1B'))\n return\n\n // Printable characters\n query += data\n scheduleSearch()\n render()\n }\n\n stdin.on('data', onData)\n })\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;AAMA,MAAM,eAAe;CAAC,KAAA;CAAW;CAAQ;CAAU;CAAW;AAG9D,SAAS,qBAAqB,OAA8C;AAC1E,KAAI,CAAC,MACH,QAAO,KAAA;AACT,KAAI,UAAU,SACZ,QAAO,EAAE,MAAM,SAAS;AAC1B,KAAI,UAAU,WACZ,QAAO,EAAE,MAAM,WAAW;AAC5B,QAAO,EAAE,MAAM,EAAE,KAAK,CAAC,OAAO,OAAO,EAAE,EAAE;;AAG3C,MAAM,iBAAiB;CAAC;CAAK;CAAK;CAAK;CAAI;AAE3C,eAAsB,kBAAkB,eAAuC;CAC7E,MAAM,MAAM,eAAe,cAAc;CACzC,MAAM,WAAW,oBAAoB;AACrC,KAAI,IAAI,WAAW,GAAG;EACpB,IAAI;AACJ,MAAI,eAAe;GACjB,MAAM,YAAY,kBAAkB;AACpC,SAAM,UAAU,SAAS,IACrB,wBAAwB,cAAc,gBAAgB,UAAU,KAAK,KAAK,KAC1E,wBAAwB,cAAc,sBAAsB,cAAc;QAG9E,OAAM;AAER,UAAQ,OAAO,MAAM,WAAW,IAAI,WAAW;AAC/C;;CAGF,MAAM,YAAY,gBAAgB,QAAQ,QAAQ,EAAE,YAAY,MAAM,CAAC;CACvE,IAAI;AACJ,KAAI;AACF,SAAO,MAAM,SAAS,IAAI;UAErB,KAAK;AACV,MAAI,eAAe,4BAA4B;AAC7C,WAAQ,OAAO,MAAM,oKAAoK;AACzL;;AAEF,QAAM;;CAIR,IAAI,QAAQ;CACZ,IAAI,UAA2B,EAAE;CACjC,IAAI,gBAAgB;CACpB,IAAI,cAAc;CAClB,IAAI,WAAW;CACf,IAAI,cAAc;CAClB,IAAI,QAAQ;CACZ,IAAI,UAAU;CACd,IAAI,YAAY;CAChB,IAAI,gBAAsD;CAE1D,MAAM,OAAO,QAAQ,OAAO,WAAW;CACvC,MAAM,aAAa;CACnB,MAAM,aAAa,gBAAgB,UAAU,cAAc,SAAS;CAEpE,SAAS,iBAAyB;EAChC,MAAM,IAAI,aAAa;AACvB,MAAI,CAAC,EACH,QAAO;AACT,SAAO,WAAW,EAAE;;CAGtB,SAAS,SAAS;EAChB,MAAM,QAAkB,EAAE;AAG1B,QAAM,KAAK,GAAG;AACd,QAAM,KAAK,YAAY,WAAW,SAAS;AAC3C,QAAM,KAAK,GAAG;EAGd,MAAM,eAAe,gBAAgB;EACrC,MAAM,SAAS,eAAe,GAAG,iBAAiB;AAClD,QAAM,KAAK,sBAAsB,SAAS,MAAM,iBAAiB;AAGjE,MAAI,aAAa;GACf,MAAM,QAAQ,eAAe,YAAY,eAAe;AACxD,SAAM,KAAK,aAAa,MAAM,mCAAmC;QAGjE,OAAM,KAAK,aAAa,IAAI,OAAO,KAAK,IAAI,OAAO,GAAG,GAAG,CAAC,CAAC,SAAS;AAItE,MAAI,OAAO;AACT,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,aAAa,MAAM,SAAS;aAEhC,MAAM,WAAW,GAAG;AAC3B,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,mCAAmC;aAEvC,MAAM,SAAS,KAAK,CAAC,aAAa;AACzC,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,gCAAgC;aAEpC,QAAQ,WAAW,KAAK,CAAC,aAAa;AAC7C,SAAM,KAAK,GAAG;AACd,SAAM,KAAK,8BAA8B;SAEtC;AACH,SAAM,KAAK,GAAG;GACd,MAAM,QAAQ,QAAQ,MAAM,GAAG,WAAW;GAC1C,MAAM,SAAS,gBAAgB,QAAQ;AACvC,QAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;IACrC,MAAM,IAAI,MAAM;IAChB,MAAM,WAAW,MAAM;IACvB,MAAM,SAAS,WAAW,qBAAqB;IAC/C,MAAM,KAAK,WAAW,OAAO,IAAI,EAAE,IAAI,EAAE;IACzC,MAAM,EAAE,OAAO,MAAM,YAAY,qBAAqB,GAAG,KAAK;IAC9D,MAAM,cAAc,eAAe,SAAS,EAAE,WAAW;IAEzD,MAAM,MAAM,SAAS,IAAI,EAAE,QAAQ;IACnC,MAAM,WAAW,MAAM,GAAG,EAAE,QAAQ,GAAG,QAAQ,EAAE;AAEjD,QAAI,UAAU;AACZ,WAAM,KAAK,KAAK,OAAO,UAAU,SAAS,UAAU,GAAG,YAAY,MAAM,SAAS;AAClF,WAAM,KAAK,eAAe,KAAK,SAAS;AACxC,WAAM,KAAK,OAAO,cAAc;UAGhC,OAAM,KAAK,KAAK,OAAO,WAAW,SAAS,UAAU,GAAG,YAAY,MAAM,SAAS;;;AAMzF,QAAM,KAAK,GAAG;EACd,MAAM,QAAkB,EAAE;AAC1B,MAAI,QAAQ,SAAS,EACnB,OAAM,KAAK,GAAG,QAAQ,OAAO,UAAU;AACzC,MAAI,UAAU,KAAK,CAAC,YAClB,OAAM,KAAK,GAAG,QAAQ,QAAQ,EAAE,CAAC,GAAG;EACtC,MAAM,SAAS,MAAM,SAAS,IAAI,GAAG,MAAM,KAAK,MAAM,CAAC,QAAQ;AAC/D,QAAM,KAAK,aAAa,OAAO,oDAAoD;AACnF,QAAM,KAAK,GAAG;AAEd,YAAU,MAAM,KAAK,KAAK,CAAC;;CAG7B,eAAe,WAAW;EACxB,MAAM,KAAK,EAAE;EACb,MAAM,YAAY,MAAM,MAAM;AAC9B,MAAI,UAAU,SAAS,GAAG;AACxB,aAAU,EAAE;AACZ,iBAAc;AACd,WAAQ;AACR;;AAGF,gBAAc;AACd,UAAQ;AACR,UAAQ;EAGR,MAAM,eAAe,kBAAkB;AACrC;AACA,OAAI,YACF,SAAQ;KACT,GAAG;EAEN,MAAM,EAAE,OAAO,QAAQ,QAAQ,iBAAiB,kBAAkB,UAAU;EAC5E,MAAM,SAAS,gBAAgB,qBAAqB,aAAa,aAAa;EAC9E,MAAM,QAAQ,YAAY,KAAK;EAE/B,MAAM,MAAM,MAAM,aAAa,QAAQ,MAAM;GAAE,OAAO;GAAY;GAAQ,CAAC,CAAC,OAAO,MAAM;AACvF,OAAI,OAAO,SACT,SAAQ,aAAa,QAAQ,EAAE,UAAU,OAAO,EAAE;AACpD,UAAO,EAAE;IACT;AAEF,gBAAc,aAAa;AAG3B,MAAI,OAAO,SACT;AAEF,YAAU;AACV,aAAW,YAAY,KAAK,GAAG,SAAS;AACxC,kBAAgB;AAChB,gBAAc;AACd,UAAQ;;CAGV,SAAS,iBAAiB;AACxB,MAAI,cACF,cAAa,cAAc;AAC7B,kBAAgB,WAAW,UAAU,IAAI;;AAI3C,SAAQ;CAGR,MAAM,EAAE,UAAU;AAClB,KAAI,MAAM,MACR,OAAM,WAAW,KAAK;AACxB,OAAM,QAAQ;AACd,OAAM,YAAY,QAAQ;AAE1B,QAAO,IAAI,SAAe,YAAY;EACpC,SAAS,UAAU;AACjB,OAAI,cACF,cAAa,cAAc;AAC7B,OAAI,MAAM,MACR,OAAM,WAAW,MAAM;AACzB,SAAM,eAAe,QAAQ,OAAO;AACpC,SAAM,OAAO;AACb,aAAU,KAAK;;EAGjB,SAAS,OAAO;AACd,YAAS;AACT,aAAU,MAAM;AAChB,YAAS;;EAGX,SAAS,eAAe;AACtB,OAAI,QAAQ,WAAW,KAAK,iBAAiB,QAAQ,OACnD;GACF,MAAM,IAAI,QAAQ;AAClB,YAAS;AACT,aAAU,MAAM;GAGhB,MAAM,UAAU,kBAAkB,EAAE,QAAQ,WAAW,EAAE;GACzD,MAAM,YAAY,EAAE,cAAc,EAAE,UAAU,IAAI,EAAE,cAAc,IAAI,EAAE,UAAU,GAAG,EAAE;GACvF,MAAM,cAAc,eAAe,iBAAiB,EAAE,QAAQ,EAAE,EAAE,WAAW;GAC7E,MAAM,OAAO,SAAS,IAAI,EAAE,QAAQ;GAGpC,MAAM,MAAM;IACV;IACA,YAJa,OAAO,GAAG,EAAE,QAAQ,GAAG,SAAS,EAAE,QAI5B,UAAU,WAHf,gBAAgB,QAAQ,CAGU,IAAI,EAAE,IAAI,EAAE;IAC5D,aAAa,QAAQ,GAAG,UAAU;IAClC;IACA,KAAK,YAAY,QAAQ,OAAO,OAAO;IACvC;IACD,CAAC,KAAK,KAAK;AACZ,WAAQ,OAAO,MAAM,GAAG,IAAI,IAAI;AAChC,YAAS;;EAGX,SAAS,OAAO,MAAc;AAE5B,OAAI,SAAS,KAAQ;AACnB,UAAM;AACN;;AAIF,OAAI,SAAS,UAAU,SAAS,YAAY;AAC1C,UAAM;AACN;;AAIF,OAAI,SAAS,QAAQ,SAAS,MAAM;AAClC,kBAAc;AACd;;AAIF,OAAI,SAAS,KAAM;AACjB,mBAAe,cAAc,KAAK,aAAa;AAC/C,QAAI,MAAM,UAAU,EAClB,iBAAgB;AAClB,YAAQ;AACR;;AAIF,OAAI,SAAS,OAAU,SAAS,MAAM;AACpC,QAAI,MAAM,SAAS,GAAG;AACpB,aAAQ,MAAM,MAAM,GAAG,GAAG;AAC1B,qBAAgB;AAChB,aAAQ;;AAEV;;AAIF,OAAI,SAAS,YAAY,SAAS,UAAU;AAE1C,QAAI,gBAAgB,GAAG;AACrB;AACA,aAAQ;;AAEV;;AAEF,OAAI,SAAS,YAAY,SAAS,UAAU;AAE1C,QAAI,gBAAgB,QAAQ,SAAS,GAAG;AACtC;AACA,aAAQ;;AAEV;;AAIF,OAAI,KAAK,WAAW,OAAO,CACzB;AAGF,YAAS;AACT,mBAAgB;AAChB,WAAQ;;AAGV,QAAM,GAAG,QAAQ,OAAO;GACxB"}
@@ -1,5 +1,6 @@
1
1
  import "./agent.mjs";
2
2
  import "./config.mjs";
3
+ import "./package-json.mjs";
3
4
  import "./prepare.mjs";
4
5
  import "./sanitize.mjs";
5
6
  import "./cache.mjs";
@@ -3,7 +3,7 @@ import { n as sanitizeMarkdown } from "./sanitize.mjs";
3
3
  import { d as searchSnippets, t as SearchDepsUnavailableError } from "./retriv.mjs";
4
4
  import { i as resolveSkilldCommand, n as getSharedSkillsDir } from "./shared.mjs";
5
5
  import { a as targets, r as detectTargetAgent } from "./detect.mjs";
6
- import { f as isInteractive } from "./cli-helpers.mjs";
6
+ import { p as isInteractive } from "./cli-helpers.mjs";
7
7
  import { i as readLock } from "./lockfile.mjs";
8
8
  import { o as normalizeScores, r as formatSnippet } from "./formatting.mjs";
9
9
  import { join } from "pathe";
@@ -1,5 +1,6 @@
1
1
  import "./agent.mjs";
2
2
  import "./config.mjs";
3
+ import "./package-json.mjs";
3
4
  import "./prepare.mjs";
4
5
  import "./sanitize.mjs";
5
6
  import "./cache.mjs";
@@ -7,7 +8,7 @@ import "./yaml.mjs";
7
8
  import "./shared.mjs";
8
9
  import "./detect.mjs";
9
10
  import "./prompts.mjs";
10
- import { b as sharedArgs, y as resolveAgent } from "./cli-helpers.mjs";
11
+ import { b as resolveAgent, x as sharedArgs } from "./cli-helpers.mjs";
11
12
  import { t as runWizard } from "./wizard.mjs";
12
13
  import { defineCommand } from "citty";
13
14
  //#region src/commands/setup.ts
@@ -1 +1 @@
1
- {"version":3,"file":"setup.mjs","names":[],"sources":["../../src/commands/setup.ts"],"sourcesContent":["import type { AgentType } from '../agent/index.ts'\nimport { defineCommand } from 'citty'\nimport { resolveAgent, sharedArgs } from '../cli-helpers.ts'\nimport { runWizard } from './wizard.ts'\n\nexport const setupCommandDef = defineCommand({\n meta: {\n name: 'setup',\n description: 'Re-run the setup wizard to configure features and model',\n },\n args: {\n agent: sharedArgs.agent,\n },\n async run({ args }) {\n const agent = resolveAgent(args.agent)\n await runWizard({\n agent: agent && agent !== 'none' ? agent as AgentType : undefined,\n })\n },\n})\n"],"mappings":";;;;;;;;;;;;;AAKA,MAAa,kBAAkB,cAAc;CAC3C,MAAM;EACJ,MAAM;EACN,aAAa;EACd;CACD,MAAM,EACJ,OAAO,WAAW,OACnB;CACD,MAAM,IAAI,EAAE,QAAQ;EAClB,MAAM,QAAQ,aAAa,KAAK,MAAM;AACtC,QAAM,UAAU,EACd,OAAO,SAAS,UAAU,SAAS,QAAqB,KAAA,GACzD,CAAC;;CAEL,CAAC"}
1
+ {"version":3,"file":"setup.mjs","names":[],"sources":["../../src/commands/setup.ts"],"sourcesContent":["import type { AgentType } from '../agent/index.ts'\nimport { defineCommand } from 'citty'\nimport { resolveAgent, sharedArgs } from '../cli-helpers.ts'\nimport { runWizard } from './wizard.ts'\n\nexport const setupCommandDef = defineCommand({\n meta: {\n name: 'setup',\n description: 'Re-run the setup wizard to configure features and model',\n },\n args: {\n agent: sharedArgs.agent,\n },\n async run({ args }) {\n const agent = resolveAgent(args.agent)\n await runWizard({\n agent: agent && agent !== 'none' ? agent as AgentType : undefined,\n })\n },\n})\n"],"mappings":";;;;;;;;;;;;;;AAKA,MAAa,kBAAkB,cAAc;CAC3C,MAAM;EACJ,MAAM;EACN,aAAa;EACd;CACD,MAAM,EACJ,OAAO,WAAW,OACnB;CACD,MAAM,IAAI,EAAE,QAAQ;EAClB,MAAM,QAAQ,aAAa,KAAK,MAAM;AACtC,QAAM,UAAU,EACd,OAAO,SAAS,UAAU,SAAS,QAAqB,KAAA,GACzD,CAAC;;CAEL,CAAC"}
@@ -1,4 +1,5 @@
1
1
  import { o as getCacheDir } from "./config.mjs";
2
+ import { i as readPackageJsonSafe } from "./package-json.mjs";
2
3
  import { t as yamlEscape } from "./yaml.mjs";
3
4
  import { i as parseFrontmatter, n as extractLinks, r as extractTitle, t as extractDescription } from "./markdown.mjs";
4
5
  import { c as getBlogPreset, l as getCrawlUrl, r as mapInsert, u as getDocOverride } from "./shared.mjs";
@@ -2491,14 +2492,11 @@ async function resolvePackageDocsWithAttempts(packageName, options = {}) {
2491
2492
  */
2492
2493
  function parseVersionSpecifier(name, version, cwd) {
2493
2494
  if (version.startsWith("link:")) {
2494
- const linkedPkgPath = join(resolve(cwd, version.slice(5)), "package.json");
2495
- if (existsSync(linkedPkgPath)) {
2496
- const linkedPkg = JSON.parse(readFileSync(linkedPkgPath, "utf-8"));
2497
- return {
2498
- name: linkedPkg.name || name,
2499
- version: linkedPkg.version || "0.0.0"
2500
- };
2501
- }
2495
+ const linkedPkg = readPackageJsonSafe(join(resolve(cwd, version.slice(5)), "package.json"));
2496
+ if (linkedPkg) return {
2497
+ name: linkedPkg.parsed.name || name,
2498
+ version: linkedPkg.parsed.version || "0.0.0"
2499
+ };
2502
2500
  return null;
2503
2501
  }
2504
2502
  if (version.startsWith("npm:")) {
@@ -2532,14 +2530,13 @@ function parseVersionSpecifier(name, version, cwd) {
2532
2530
  */
2533
2531
  function resolveInstalledVersion(name, cwd) {
2534
2532
  try {
2535
- const resolved = resolvePathSync(`${name}/package.json`, { url: cwd });
2536
- return JSON.parse(readFileSync(resolved, "utf-8")).version || null;
2533
+ return readPackageJsonSafe(resolvePathSync(`${name}/package.json`, { url: cwd }))?.parsed.version || null;
2537
2534
  } catch {
2538
2535
  try {
2539
2536
  let dir = dirname(resolvePathSync(name, { url: cwd }));
2540
2537
  while (dir && basename(dir) !== "node_modules") {
2541
- const pkgPath = join(dir, "package.json");
2542
- if (existsSync(pkgPath)) return JSON.parse(readFileSync(pkgPath, "utf-8")).version || null;
2538
+ const pkg = readPackageJsonSafe(join(dir, "package.json"));
2539
+ if (pkg) return pkg.parsed.version || null;
2543
2540
  const parent = dirname(dir);
2544
2541
  if (parent === dir) break;
2545
2542
  dir = parent;
@@ -2552,9 +2549,9 @@ function resolveInstalledVersion(name, cwd) {
2552
2549
  * Read package.json dependencies with versions
2553
2550
  */
2554
2551
  async function readLocalDependencies(cwd) {
2555
- const pkgPath = join(cwd, "package.json");
2556
- if (!existsSync(pkgPath)) throw new Error("No package.json found in current directory");
2557
- const pkg = JSON.parse(readFileSync(pkgPath, "utf-8"));
2552
+ const result = readPackageJsonSafe(join(cwd, "package.json"));
2553
+ if (!result) throw new Error("No package.json found in current directory");
2554
+ const pkg = result.parsed;
2558
2555
  const deps = {
2559
2556
  ...pkg.dependencies,
2560
2557
  ...pkg.devDependencies
@@ -2570,9 +2567,9 @@ async function readLocalDependencies(cwd) {
2570
2567
  * Read package info from a local path (for link: deps)
2571
2568
  */
2572
2569
  function readLocalPackageInfo(localPath) {
2573
- const pkgPath = join(localPath, "package.json");
2574
- if (!existsSync(pkgPath)) return null;
2575
- const pkg = JSON.parse(readFileSync(pkgPath, "utf-8"));
2570
+ const result = readPackageJsonSafe(join(localPath, "package.json"));
2571
+ if (!result) return null;
2572
+ const pkg = result.parsed;
2576
2573
  let repoUrl;
2577
2574
  if (pkg.repository?.url) repoUrl = normalizeRepoUrl(pkg.repository.url);
2578
2575
  else if (typeof pkg.repository === "string") repoUrl = normalizeRepoUrl(pkg.repository);