skilld 0.15.2 → 0.15.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (72) hide show
  1. package/README.md +1 -0
  2. package/dist/_chunks/{detect-imports.mjs → agent.mjs} +42 -6
  3. package/dist/_chunks/agent.mjs.map +1 -0
  4. package/dist/_chunks/{storage.mjs → cache.mjs} +81 -1
  5. package/dist/_chunks/cache.mjs.map +1 -0
  6. package/dist/_chunks/cache2.mjs +6 -0
  7. package/dist/_chunks/config.mjs +23 -0
  8. package/dist/_chunks/config.mjs.map +1 -1
  9. package/dist/_chunks/config2.mjs +12 -0
  10. package/dist/_chunks/{embedding-cache2.mjs → embedding-cache.mjs} +1 -1
  11. package/dist/_chunks/embedding-cache.mjs.map +1 -0
  12. package/dist/_chunks/formatting.mjs +86 -0
  13. package/dist/_chunks/formatting.mjs.map +1 -0
  14. package/dist/_chunks/{version.d.mts → index.d.mts} +1 -1
  15. package/dist/_chunks/index.d.mts.map +1 -0
  16. package/dist/_chunks/{utils.d.mts → index2.d.mts} +1 -1
  17. package/dist/_chunks/index2.d.mts.map +1 -0
  18. package/dist/_chunks/install.mjs +15 -0
  19. package/dist/_chunks/list.mjs +13 -0
  20. package/dist/_chunks/markdown.mjs +7 -0
  21. package/dist/_chunks/markdown.mjs.map +1 -1
  22. package/dist/_chunks/{pool2.mjs → pool.mjs} +1 -1
  23. package/dist/_chunks/pool.mjs.map +1 -0
  24. package/dist/_chunks/prompts.mjs +232 -0
  25. package/dist/_chunks/prompts.mjs.map +1 -1
  26. package/dist/_chunks/remove.mjs +12 -0
  27. package/dist/_chunks/sanitize.mjs +71 -0
  28. package/dist/_chunks/sanitize.mjs.map +1 -1
  29. package/dist/_chunks/search-interactive.mjs +14 -0
  30. package/dist/_chunks/search-interactive2.mjs +236 -0
  31. package/dist/_chunks/search-interactive2.mjs.map +1 -0
  32. package/dist/_chunks/search.mjs +171 -0
  33. package/dist/_chunks/search.mjs.map +1 -0
  34. package/dist/_chunks/search2.mjs +13 -0
  35. package/dist/_chunks/shared.mjs +4 -0
  36. package/dist/_chunks/shared.mjs.map +1 -1
  37. package/dist/_chunks/skills.mjs +552 -0
  38. package/dist/_chunks/skills.mjs.map +1 -0
  39. package/dist/_chunks/{npm.mjs → sources.mjs} +401 -4
  40. package/dist/_chunks/sources.mjs.map +1 -0
  41. package/dist/_chunks/status.mjs +13 -0
  42. package/dist/_chunks/sync.mjs +2026 -0
  43. package/dist/_chunks/sync.mjs.map +1 -0
  44. package/dist/_chunks/sync2.mjs +14 -0
  45. package/dist/_chunks/uninstall.mjs +15 -0
  46. package/dist/_chunks/validate.mjs +3 -0
  47. package/dist/_chunks/validate.mjs.map +1 -1
  48. package/dist/_chunks/yaml.mjs +19 -0
  49. package/dist/_chunks/yaml.mjs.map +1 -1
  50. package/dist/agent/index.d.mts +1 -1
  51. package/dist/agent/index.mjs +4 -3
  52. package/dist/cache/index.d.mts +2 -2
  53. package/dist/cache/index.mjs +2 -1
  54. package/dist/cli.mjs +173 -3082
  55. package/dist/cli.mjs.map +1 -1
  56. package/dist/index.d.mts +2 -3
  57. package/dist/index.mjs +4 -4
  58. package/dist/retriv/index.d.mts.map +1 -1
  59. package/dist/retriv/index.mjs +26 -5
  60. package/dist/retriv/index.mjs.map +1 -1
  61. package/dist/retriv/worker.mjs +3 -3
  62. package/dist/sources/index.d.mts +2 -2
  63. package/dist/sources/index.mjs +2 -1
  64. package/dist/types.d.mts +2 -3
  65. package/package.json +10 -10
  66. package/dist/_chunks/detect-imports.mjs.map +0 -1
  67. package/dist/_chunks/embedding-cache2.mjs.map +0 -1
  68. package/dist/_chunks/npm.mjs.map +0 -1
  69. package/dist/_chunks/pool2.mjs.map +0 -1
  70. package/dist/_chunks/storage.mjs.map +0 -1
  71. package/dist/_chunks/utils.d.mts.map +0 -1
  72. package/dist/_chunks/version.d.mts.map +0 -1
@@ -1 +1 @@
1
- {"version":3,"file":"config.mjs","names":[],"sources":["../../src/cache/version.ts","../../src/cache/config.ts"],"sourcesContent":["/**\n * Version utilities\n */\n\nimport { resolve } from 'pathe'\nimport { REFERENCES_DIR } from './config.ts'\n\n/** Validate npm package name (scoped or unscoped) */\nconst VALID_PKG_NAME = /^(?:@[a-z0-9][-a-z0-9._]*\\/)?[a-z0-9][-a-z0-9._]*$/\n\n/** Validate version string (semver-ish, no path separators) */\nconst VALID_VERSION = /^[a-z0-9][-\\w.+]*$/i\n\n/**\n * Get exact version key for cache keying\n */\nexport function getVersionKey(version: string): string {\n return version\n}\n\n/**\n * Get cache key for a package: name@version\n */\nexport function getCacheKey(name: string, version: string): string {\n return `${name}@${getVersionKey(version)}`\n}\n\n/**\n * Get path to cached package references.\n * Validates name/version to prevent path traversal.\n */\nexport function getCacheDir(name: string, version: string): string {\n if (!VALID_PKG_NAME.test(name))\n throw new Error(`Invalid package name: ${name}`)\n if (!VALID_VERSION.test(version))\n throw new Error(`Invalid version: ${version}`)\n\n const dir = resolve(REFERENCES_DIR, getCacheKey(name, version))\n if (!dir.startsWith(REFERENCES_DIR))\n throw new Error(`Path traversal detected: ${dir}`)\n return dir\n}\n","/**\n * Cache configuration\n */\n\nimport { homedir } from 'node:os'\nimport { join } from 'pathe'\nimport { getCacheKey } from './version.ts'\n\n/** Global cache directory */\nexport const CACHE_DIR = join(homedir(), '.skilld')\n\n/** References subdirectory */\nexport const REFERENCES_DIR = join(CACHE_DIR, 'references')\n\n/** Repo-level cache (issues, discussions, releases shared across monorepo packages) */\nexport const REPOS_DIR = join(CACHE_DIR, 'repos')\n\n/** Get repo cache dir for owner/repo with path traversal validation */\nexport function getRepoCacheDir(owner: string, repo: string): string {\n if (owner.includes('..') || repo.includes('..') || owner.includes('/') || repo.includes('/'))\n throw new Error(`Invalid repo path: ${owner}/${repo}`)\n return join(REPOS_DIR, owner, repo)\n}\n\n/** Get search DB path for a specific package@version */\nexport function getPackageDbPath(name: string, version: string): string {\n return join(REFERENCES_DIR, getCacheKey(name, version), 'search.db')\n}\n"],"mappings":";;AAQA,MAAM,iBAAiB;AAGvB,MAAM,gBAAgB;AAKtB,SAAgB,cAAc,SAAyB;AACrD,QAAO;;AAMT,SAAgB,YAAY,MAAc,SAAyB;AACjE,QAAO,GAAG,KAAK,GAAG,cAAc,QAAQ;;AAO1C,SAAgB,YAAY,MAAc,SAAyB;AACjE,KAAI,CAAC,eAAe,KAAK,KAAK,CAC5B,OAAM,IAAI,MAAM,yBAAyB,OAAO;AAClD,KAAI,CAAC,cAAc,KAAK,QAAQ,CAC9B,OAAM,IAAI,MAAM,oBAAoB,UAAU;CAEhD,MAAM,MAAM,QAAQ,gBAAgB,YAAY,MAAM,QAAQ,CAAC;AAC/D,KAAI,CAAC,IAAI,WAAW,eAAe,CACjC,OAAM,IAAI,MAAM,4BAA4B,MAAM;AACpD,QAAO;;AC/BT,MAAa,YAAY,KAAK,SAAS,EAAE,UAAU;AAGnD,MAAa,iBAAiB,KAAK,WAAW,aAAa;AAG3D,MAAa,YAAY,KAAK,WAAW,QAAQ;AAGjD,SAAgB,gBAAgB,OAAe,MAAsB;AACnE,KAAI,MAAM,SAAS,KAAK,IAAI,KAAK,SAAS,KAAK,IAAI,MAAM,SAAS,IAAI,IAAI,KAAK,SAAS,IAAI,CAC1F,OAAM,IAAI,MAAM,sBAAsB,MAAM,GAAG,OAAO;AACxD,QAAO,KAAK,WAAW,OAAO,KAAK;;AAIrC,SAAgB,iBAAiB,MAAc,SAAyB;AACtE,QAAO,KAAK,gBAAgB,YAAY,MAAM,QAAQ,EAAE,YAAY"}
1
+ {"version":3,"file":"config.mjs","names":[],"sources":["../../src/cache/version.ts","../../src/cache/config.ts"],"sourcesContent":["/**\n * Version utilities\n */\n\nimport { resolve } from 'pathe'\nimport { REFERENCES_DIR } from './config.ts'\n\n/** Validate npm package name (scoped or unscoped) */\nconst VALID_PKG_NAME = /^(?:@[a-z0-9][-a-z0-9._]*\\/)?[a-z0-9][-a-z0-9._]*$/\n\n/** Validate version string (semver-ish, no path separators) */\nconst VALID_VERSION = /^[a-z0-9][-\\w.+]*$/i\n\n/**\n * Get exact version key for cache keying\n */\nexport function getVersionKey(version: string): string {\n return version\n}\n\n/**\n * Get cache key for a package: name@version\n */\nexport function getCacheKey(name: string, version: string): string {\n return `${name}@${getVersionKey(version)}`\n}\n\n/**\n * Get path to cached package references.\n * Validates name/version to prevent path traversal.\n */\nexport function getCacheDir(name: string, version: string): string {\n if (!VALID_PKG_NAME.test(name))\n throw new Error(`Invalid package name: ${name}`)\n if (!VALID_VERSION.test(version))\n throw new Error(`Invalid version: ${version}`)\n\n const dir = resolve(REFERENCES_DIR, getCacheKey(name, version))\n if (!dir.startsWith(REFERENCES_DIR))\n throw new Error(`Path traversal detected: ${dir}`)\n return dir\n}\n","/**\n * Cache configuration\n */\n\nimport { homedir } from 'node:os'\nimport { join } from 'pathe'\nimport { getCacheKey } from './version.ts'\n\n/** Global cache directory */\nexport const CACHE_DIR = join(homedir(), '.skilld')\n\n/** References subdirectory */\nexport const REFERENCES_DIR = join(CACHE_DIR, 'references')\n\n/** Repo-level cache (issues, discussions, releases shared across monorepo packages) */\nexport const REPOS_DIR = join(CACHE_DIR, 'repos')\n\n/** Get repo cache dir for owner/repo with path traversal validation */\nexport function getRepoCacheDir(owner: string, repo: string): string {\n if (owner.includes('..') || repo.includes('..') || owner.includes('/') || repo.includes('/'))\n throw new Error(`Invalid repo path: ${owner}/${repo}`)\n return join(REPOS_DIR, owner, repo)\n}\n\n/** Get search DB path for a specific package@version */\nexport function getPackageDbPath(name: string, version: string): string {\n return join(REFERENCES_DIR, getCacheKey(name, version), 'search.db')\n}\n"],"mappings":";;;;;;AAQA,MAAM,iBAAiB;;AAGvB,MAAM,gBAAgB;;;;AAKtB,SAAgB,cAAc,SAAyB;AACrD,QAAO;;;;;AAMT,SAAgB,YAAY,MAAc,SAAyB;AACjE,QAAO,GAAG,KAAK,GAAG,cAAc,QAAQ;;;;;;AAO1C,SAAgB,YAAY,MAAc,SAAyB;AACjE,KAAI,CAAC,eAAe,KAAK,KAAK,CAC5B,OAAM,IAAI,MAAM,yBAAyB,OAAO;AAClD,KAAI,CAAC,cAAc,KAAK,QAAQ,CAC9B,OAAM,IAAI,MAAM,oBAAoB,UAAU;CAEhD,MAAM,MAAM,QAAQ,gBAAgB,YAAY,MAAM,QAAQ,CAAC;AAC/D,KAAI,CAAC,IAAI,WAAW,eAAe,CACjC,OAAM,IAAI,MAAM,4BAA4B,MAAM;AACpD,QAAO;;;;;;AC/BT,MAAa,YAAY,KAAK,SAAS,EAAE,UAAU;;AAGnD,MAAa,iBAAiB,KAAK,WAAW,aAAa;;AAG3D,MAAa,YAAY,KAAK,WAAW,QAAQ;;AAGjD,SAAgB,gBAAgB,OAAe,MAAsB;AACnE,KAAI,MAAM,SAAS,KAAK,IAAI,KAAK,SAAS,KAAK,IAAI,MAAM,SAAS,IAAI,IAAI,KAAK,SAAS,IAAI,CAC1F,OAAM,IAAI,MAAM,sBAAsB,MAAM,GAAG,OAAO;AACxD,QAAO,KAAK,WAAW,OAAO,KAAK;;;AAIrC,SAAgB,iBAAiB,MAAc,SAAyB;AACtE,QAAO,KAAK,gBAAgB,YAAY,MAAM,QAAQ,EAAE,YAAY"}
@@ -0,0 +1,12 @@
1
+ import "./config.mjs";
2
+ import "./sanitize.mjs";
3
+ import "./cache.mjs";
4
+ import "./yaml.mjs";
5
+ import "./markdown.mjs";
6
+ import "./shared.mjs";
7
+ import "./sources.mjs";
8
+ import "./prompts.mjs";
9
+ import "./agent.mjs";
10
+ import "./skills.mjs";
11
+ import { f as configCommandDef } from "../cli.mjs";
12
+ export { configCommandDef };
@@ -51,4 +51,4 @@ function clearEmbeddingCache() {
51
51
  }
52
52
  export { clearEmbeddingCache as n, embedding_cache_exports as r, cachedEmbeddings as t };
53
53
 
54
- //# sourceMappingURL=embedding-cache2.mjs.map
54
+ //# sourceMappingURL=embedding-cache.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"embedding-cache.mjs","names":[],"sources":["../../src/retriv/embedding-cache.ts"],"sourcesContent":["import type { DatabaseSync } from 'node:sqlite'\nimport type { Embedding } from 'retriv'\nimport { rmSync } from 'node:fs'\nimport { join } from 'pathe'\nimport { CACHE_DIR } from '../cache/index.ts'\n\ninterface EmbeddingConfig {\n resolve: () => Promise<{ embedder: (texts: string[]) => Promise<Embedding[]>, dimensions: number, maxTokens?: number }>\n}\n\nconst EMBEDDINGS_DB_PATH = join(CACHE_DIR, 'embeddings.db')\n\nfunction openDb(): DatabaseSync {\n // eslint-disable-next-line ts/no-require-imports\n const { DatabaseSync: DB } = require('node:sqlite') as typeof import('node:sqlite')\n const db = new DB(EMBEDDINGS_DB_PATH)\n db.exec('PRAGMA journal_mode=WAL')\n db.exec('PRAGMA busy_timeout=5000')\n db.exec(`CREATE TABLE IF NOT EXISTS embeddings (text_hash TEXT PRIMARY KEY, embedding BLOB NOT NULL)`)\n db.exec(`CREATE TABLE IF NOT EXISTS meta (key TEXT PRIMARY KEY, value TEXT NOT NULL)`)\n return db\n}\n\nfunction createSqliteStorage(db: DatabaseSync) {\n const getStmt = db.prepare('SELECT embedding FROM embeddings WHERE text_hash = ?')\n const setStmt = db.prepare('INSERT OR IGNORE INTO embeddings (text_hash, embedding) VALUES (?, ?)')\n\n return {\n get: (hash: string): Embedding | null => {\n const row = getStmt.get(hash) as { embedding: Buffer } | undefined\n if (!row)\n return null\n return new Float32Array(row.embedding.buffer, row.embedding.byteOffset, row.embedding.byteLength / 4)\n },\n set: (hash: string, embedding: Embedding): void => {\n const arr = embedding instanceof Float32Array ? embedding : new Float32Array(embedding)\n setStmt.run(hash, Buffer.from(arr.buffer, arr.byteOffset, arr.byteLength))\n },\n }\n}\n\nexport async function cachedEmbeddings(config: EmbeddingConfig): Promise<EmbeddingConfig> {\n const { cachedEmbeddings: retrivCached } = await import('retriv/embeddings/cached')\n const db = openDb()\n const storage = createSqliteStorage(db)\n\n const originalResolve = config.resolve\n const validatedConfig: EmbeddingConfig = {\n async resolve() {\n const resolved = await originalResolve()\n const getMetaStmt = db.prepare('SELECT value FROM meta WHERE key = ?')\n const setMetaStmt = db.prepare('INSERT OR REPLACE INTO meta (key, value) VALUES (?, ?)')\n\n const storedDims = getMetaStmt.get('dimensions') as { value: string } | undefined\n if (storedDims && Number(storedDims.value) !== resolved.dimensions) {\n db.exec('DELETE FROM embeddings')\n }\n setMetaStmt.run('dimensions', String(resolved.dimensions))\n\n return resolved\n },\n }\n\n return retrivCached(validatedConfig, { storage })\n}\n\nexport function clearEmbeddingCache(): void {\n rmSync(EMBEDDINGS_DB_PATH, { force: true })\n}\n"],"mappings":";;;;;;;;AAUA,MAAM,qBAAqB,KAAK,WAAW,gBAAgB;AAE3D,SAAS,SAAuB;CAE9B,MAAM,EAAE,cAAc,OAAA,UAAe,cAAc;CACnD,MAAM,KAAK,IAAI,GAAG,mBAAmB;AACrC,IAAG,KAAK,0BAA0B;AAClC,IAAG,KAAK,2BAA2B;AACnC,IAAG,KAAK,8FAA8F;AACtG,IAAG,KAAK,8EAA8E;AACtF,QAAO;;AAGT,SAAS,oBAAoB,IAAkB;CAC7C,MAAM,UAAU,GAAG,QAAQ,uDAAuD;CAClF,MAAM,UAAU,GAAG,QAAQ,wEAAwE;AAEnG,QAAO;EACL,MAAM,SAAmC;GACvC,MAAM,MAAM,QAAQ,IAAI,KAAK;AAC7B,OAAI,CAAC,IACH,QAAO;AACT,UAAO,IAAI,aAAa,IAAI,UAAU,QAAQ,IAAI,UAAU,YAAY,IAAI,UAAU,aAAa,EAAE;;EAEvG,MAAM,MAAc,cAA+B;GACjD,MAAM,MAAM,qBAAqB,eAAe,YAAY,IAAI,aAAa,UAAU;AACvF,WAAQ,IAAI,MAAM,OAAO,KAAK,IAAI,QAAQ,IAAI,YAAY,IAAI,WAAW,CAAC;;EAE7E;;AAGH,eAAsB,iBAAiB,QAAmD;CACxF,MAAM,EAAE,kBAAkB,iBAAiB,MAAM,OAAO;CACxD,MAAM,KAAK,QAAQ;CACnB,MAAM,UAAU,oBAAoB,GAAG;CAEvC,MAAM,kBAAkB,OAAO;AAiB/B,QAAO,aAhBkC,EACvC,MAAM,UAAU;EACd,MAAM,WAAW,MAAM,iBAAiB;EACxC,MAAM,cAAc,GAAG,QAAQ,uCAAuC;EACtE,MAAM,cAAc,GAAG,QAAQ,yDAAyD;EAExF,MAAM,aAAa,YAAY,IAAI,aAAa;AAChD,MAAI,cAAc,OAAO,WAAW,MAAM,KAAK,SAAS,WACtD,IAAG,KAAK,yBAAyB;AAEnC,cAAY,IAAI,cAAc,OAAO,SAAS,WAAW,CAAC;AAE1D,SAAO;IAEV,EAEoC,EAAE,SAAS,CAAC;;AAGnD,SAAgB,sBAA4B;AAC1C,QAAO,oBAAoB,EAAE,OAAO,MAAM,CAAC"}
@@ -0,0 +1,86 @@
1
+ import * as p from "@clack/prompts";
2
+ function timeAgo(iso) {
3
+ if (!iso) return "";
4
+ const diff = Date.now() - new Date(iso).getTime();
5
+ const days = Math.floor(diff / 864e5);
6
+ if (days <= 0) return "today";
7
+ if (days === 1) return "1d ago";
8
+ if (days < 7) return `${days}d ago`;
9
+ if (days < 30) return `${Math.floor(days / 7)}w ago`;
10
+ return `${Math.floor(days / 30)}mo ago`;
11
+ }
12
+ function formatSource(source) {
13
+ if (!source) return "";
14
+ if (source === "shipped") return "shipped";
15
+ if (source.includes("llms.txt")) return "llms.txt";
16
+ if (source.includes("github.com")) return source.replace(/https?:\/\/github\.com\//, "");
17
+ return source;
18
+ }
19
+ function formatDuration(ms) {
20
+ if (ms < 1e3) return `${Math.round(ms)}ms`;
21
+ return `${(ms / 1e3).toFixed(1)}s`;
22
+ }
23
+ /** Spinner wrapper that shows elapsed time via built-in timer indicator */
24
+ function timedSpinner() {
25
+ const spin = p.spinner({ indicator: "timer" });
26
+ return {
27
+ start(msg) {
28
+ spin.start(msg);
29
+ },
30
+ message(msg) {
31
+ spin.message(msg);
32
+ },
33
+ stop(msg) {
34
+ spin.stop(msg);
35
+ }
36
+ };
37
+ }
38
+ function highlightTerms(content, terms) {
39
+ if (terms.length === 0) return content;
40
+ const sorted = [...terms].sort((a, b) => b.length - a.length);
41
+ const pattern = new RegExp(`(${sorted.map((t) => t.replace(/[.*+?^${}()|[\]\\]/g, "\\$&")).join("|")})`, "gi");
42
+ return content.replace(pattern, "\x1B[33m$1\x1B[0m");
43
+ }
44
+ /** Format a normalized score (0-100) with color */
45
+ function scoreLabel(pct) {
46
+ return `${pct >= 70 ? "\x1B[32m" : pct >= 40 ? "\x1B[33m" : "\x1B[90m"}${pct}%\x1B[0m`;
47
+ }
48
+ /** Normalize raw cosine similarity scores to 0-100 relative to the best match */
49
+ function normalizeScores(results) {
50
+ const map = /* @__PURE__ */ new Map();
51
+ const max = results.reduce((m, r) => Math.max(m, r.score), 0);
52
+ for (const r of results) map.set(r, max > 0 ? Math.round(r.score / max * 100) : 0);
53
+ return map;
54
+ }
55
+ function formatSnippet(r, versions, pct) {
56
+ const refPath = `.claude/skills/${r.package}/.skilld/${r.source}`;
57
+ const lineRange = r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`;
58
+ const score = pct != null ? scoreLabel(pct) : `\x1B[90m${r.score.toFixed(2)}\x1B[0m`;
59
+ const version = versions?.get(r.package);
60
+ const pkgLabel = version ? `${r.package}@${version}` : r.package;
61
+ const scopeStr = r.scope?.length ? `${r.scope.map((e) => e.name).join(".")} → ` : "";
62
+ const entityStr = r.entities?.map((e) => e.signature || `${e.type} ${e.name}`).join(", ");
63
+ const highlighted = highlightTerms(r.content, r.highlights);
64
+ return [
65
+ `${pkgLabel} ${score}${entityStr ? ` \x1B[36m${scopeStr}${entityStr}\x1B[0m` : ""}`,
66
+ `\x1B[90m${refPath}:${lineRange}\x1B[0m`,
67
+ ` ${highlighted.replace(/\n/g, "\n ")}`
68
+ ].join("\n");
69
+ }
70
+ /** Compact 2-line format for interactive search list */
71
+ function formatCompactSnippet(r, cols) {
72
+ const entityStr = r.entities?.length ? r.entities.map((e) => e.signature || e.name).join(", ") : "";
73
+ const scopeStr = r.scope?.length ? `${r.scope.map((e) => e.name).join(".")} → ` : "";
74
+ const title = entityStr ? `${scopeStr}${entityStr}` : r.source.split("/").pop() || r.source;
75
+ const path = `${`.claude/skills/${r.package}/.skilld/${r.source}`}:${r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`}`;
76
+ const maxPreview = cols - 6;
77
+ const firstLine = r.content.split("\n").find((l) => l.trim() && l.trim() !== "---" && !/^#+\s*$/.test(l.trim())) || "";
78
+ return {
79
+ title,
80
+ path,
81
+ preview: firstLine.length > maxPreview ? `${firstLine.slice(0, maxPreview - 1)}…` : firstLine
82
+ };
83
+ }
84
+ export { highlightTerms as a, timeAgo as c, formatSource as i, timedSpinner as l, formatDuration as n, normalizeScores as o, formatSnippet as r, scoreLabel as s, formatCompactSnippet as t };
85
+
86
+ //# sourceMappingURL=formatting.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"formatting.mjs","names":[],"sources":["../../src/core/formatting.ts"],"sourcesContent":["import type { SearchSnippet } from '../retriv/index.ts'\nimport type { ProjectState } from './skills.ts'\nimport * as p from '@clack/prompts'\n\nexport function timeAgo(iso?: string): string {\n if (!iso)\n return ''\n const diff = Date.now() - new Date(iso).getTime()\n const days = Math.floor(diff / 86400000)\n if (days <= 0)\n return 'today'\n if (days === 1)\n return '1d ago'\n if (days < 7)\n return `${days}d ago`\n if (days < 30)\n return `${Math.floor(days / 7)}w ago`\n return `${Math.floor(days / 30)}mo ago`\n}\n\nexport function formatSource(source?: string): string {\n if (!source)\n return ''\n if (source === 'shipped')\n return 'shipped'\n if (source.includes('llms.txt'))\n return 'llms.txt'\n if (source.includes('github.com'))\n return source.replace(/https?:\\/\\/github\\.com\\//, '')\n return source\n}\n\nexport function formatDuration(ms: number): string {\n if (ms < 1000)\n return `${Math.round(ms)}ms`\n return `${(ms / 1000).toFixed(1)}s`\n}\n\n/** Spinner wrapper that shows elapsed time via built-in timer indicator */\nexport function timedSpinner() {\n const spin = p.spinner({ indicator: 'timer' })\n return {\n start(msg: string) {\n spin.start(msg)\n },\n message(msg: string) {\n spin.message(msg)\n },\n stop(msg: string) {\n spin.stop(msg)\n },\n }\n}\n\nexport function formatSkillStatus(state: ProjectState): void {\n const { missing, outdated, synced } = state\n\n if (synced.length > 0)\n p.log.success(`${synced.length} synced`)\n if (outdated.length > 0)\n p.log.warn(`${outdated.length} outdated: ${outdated.map(s => s.name).join(', ')}`)\n if (missing.length > 0)\n p.log.info(`${missing.length} missing: ${missing.slice(0, 5).join(', ')}${missing.length > 5 ? '...' : ''}`)\n}\n\nexport function highlightTerms(content: string, terms: string[]): string {\n if (terms.length === 0)\n return content\n // Sort by length desc to match longer terms first\n const sorted = [...terms].sort((a, b) => b.length - a.length)\n const pattern = new RegExp(`(${sorted.map(t => t.replace(/[.*+?^${}()|[\\]\\\\]/g, '\\\\$&')).join('|')})`, 'gi')\n return content.replace(pattern, '\\x1B[33m$1\\x1B[0m')\n}\n\n/** Format a normalized score (0-100) with color */\nexport function scoreLabel(pct: number): string {\n const color = pct >= 70 ? '\\x1B[32m' : pct >= 40 ? '\\x1B[33m' : '\\x1B[90m'\n return `${color}${pct}%\\x1B[0m`\n}\n\n/** Normalize raw cosine similarity scores to 0-100 relative to the best match */\nexport function normalizeScores(results: SearchSnippet[]): Map<SearchSnippet, number> {\n const map = new Map<SearchSnippet, number>()\n const max = results.reduce((m, r) => Math.max(m, r.score), 0)\n for (const r of results)\n map.set(r, max > 0 ? Math.round((r.score / max) * 100) : 0)\n return map\n}\n\nexport function formatSnippet(r: SearchSnippet, versions?: Map<string, string>, pct?: number): string {\n const refPath = `.claude/skills/${r.package}/.skilld/${r.source}`\n const lineRange = r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`\n const score = pct != null ? scoreLabel(pct) : `\\x1B[90m${r.score.toFixed(2)}\\x1B[0m`\n const version = versions?.get(r.package)\n const pkgLabel = version ? `${r.package}@${version}` : r.package\n\n const scopeStr = r.scope?.length ? `${r.scope.map(e => e.name).join('.')} → ` : ''\n const entityStr = r.entities?.map(e => e.signature || `${e.type} ${e.name}`).join(', ')\n const highlighted = highlightTerms(r.content, r.highlights)\n\n return [\n `${pkgLabel} ${score}${entityStr ? ` \\x1B[36m${scopeStr}${entityStr}\\x1B[0m` : ''}`,\n `\\x1B[90m${refPath}:${lineRange}\\x1B[0m`,\n ` ${highlighted.replace(/\\n/g, '\\n ')}`,\n ].join('\\n')\n}\n\n/** Compact 2-line format for interactive search list */\nexport function formatCompactSnippet(r: SearchSnippet, cols: number): { title: string, path: string, preview: string } {\n const entityStr = r.entities?.length\n ? r.entities.map(e => e.signature || e.name).join(', ')\n : ''\n const scopeStr = r.scope?.length ? `${r.scope.map(e => e.name).join('.')} → ` : ''\n const title = entityStr ? `${scopeStr}${entityStr}` : r.source.split('/').pop() || r.source\n\n const refPath = `.claude/skills/${r.package}/.skilld/${r.source}`\n const lineRange = r.lineStart === r.lineEnd ? `L${r.lineStart}` : `L${r.lineStart}-${r.lineEnd}`\n const path = `${refPath}:${lineRange}`\n\n // First meaningful line as preview (skip empty, frontmatter delimiters, headings-only)\n const maxPreview = cols - 6\n const firstLine = r.content.split('\\n').find(l => l.trim() && l.trim() !== '---' && !/^#+\\s*$/.test(l.trim())) || ''\n const preview = firstLine.length > maxPreview ? `${firstLine.slice(0, maxPreview - 1)}…` : firstLine\n\n return { title, path, preview }\n}\n"],"mappings":";AAIA,SAAgB,QAAQ,KAAsB;AAC5C,KAAI,CAAC,IACH,QAAO;CACT,MAAM,OAAO,KAAK,KAAK,GAAG,IAAI,KAAK,IAAI,CAAC,SAAS;CACjD,MAAM,OAAO,KAAK,MAAM,OAAO,MAAS;AACxC,KAAI,QAAQ,EACV,QAAO;AACT,KAAI,SAAS,EACX,QAAO;AACT,KAAI,OAAO,EACT,QAAO,GAAG,KAAK;AACjB,KAAI,OAAO,GACT,QAAO,GAAG,KAAK,MAAM,OAAO,EAAE,CAAC;AACjC,QAAO,GAAG,KAAK,MAAM,OAAO,GAAG,CAAC;;AAGlC,SAAgB,aAAa,QAAyB;AACpD,KAAI,CAAC,OACH,QAAO;AACT,KAAI,WAAW,UACb,QAAO;AACT,KAAI,OAAO,SAAS,WAAW,CAC7B,QAAO;AACT,KAAI,OAAO,SAAS,aAAa,CAC/B,QAAO,OAAO,QAAQ,4BAA4B,GAAG;AACvD,QAAO;;AAGT,SAAgB,eAAe,IAAoB;AACjD,KAAI,KAAK,IACP,QAAO,GAAG,KAAK,MAAM,GAAG,CAAC;AAC3B,QAAO,IAAI,KAAK,KAAM,QAAQ,EAAE,CAAC;;;AAInC,SAAgB,eAAe;CAC7B,MAAM,OAAO,EAAE,QAAQ,EAAE,WAAW,SAAS,CAAC;AAC9C,QAAO;EACL,MAAM,KAAa;AACjB,QAAK,MAAM,IAAI;;EAEjB,QAAQ,KAAa;AACnB,QAAK,QAAQ,IAAI;;EAEnB,KAAK,KAAa;AAChB,QAAK,KAAK,IAAI;;EAEjB;;AAcH,SAAgB,eAAe,SAAiB,OAAyB;AACvE,KAAI,MAAM,WAAW,EACnB,QAAO;CAET,MAAM,SAAS,CAAC,GAAG,MAAM,CAAC,MAAM,GAAG,MAAM,EAAE,SAAS,EAAE,OAAO;CAC7D,MAAM,UAAU,IAAI,OAAO,IAAI,OAAO,KAAI,MAAK,EAAE,QAAQ,uBAAuB,OAAO,CAAC,CAAC,KAAK,IAAI,CAAC,IAAI,KAAK;AAC5G,QAAO,QAAQ,QAAQ,SAAS,oBAAoB;;;AAItD,SAAgB,WAAW,KAAqB;AAE9C,QAAO,GADO,OAAO,KAAK,aAAa,OAAO,KAAK,aAAa,aAC9C,IAAI;;;AAIxB,SAAgB,gBAAgB,SAAsD;CACpF,MAAM,sBAAM,IAAI,KAA4B;CAC5C,MAAM,MAAM,QAAQ,QAAQ,GAAG,MAAM,KAAK,IAAI,GAAG,EAAE,MAAM,EAAE,EAAE;AAC7D,MAAK,MAAM,KAAK,QACd,KAAI,IAAI,GAAG,MAAM,IAAI,KAAK,MAAO,EAAE,QAAQ,MAAO,IAAI,GAAG,EAAE;AAC7D,QAAO;;AAGT,SAAgB,cAAc,GAAkB,UAAgC,KAAsB;CACpG,MAAM,UAAU,kBAAkB,EAAE,QAAQ,WAAW,EAAE;CACzD,MAAM,YAAY,EAAE,cAAc,EAAE,UAAU,IAAI,EAAE,cAAc,IAAI,EAAE,UAAU,GAAG,EAAE;CACvF,MAAM,QAAQ,OAAO,OAAO,WAAW,IAAI,GAAG,WAAW,EAAE,MAAM,QAAQ,EAAE,CAAC;CAC5E,MAAM,UAAU,UAAU,IAAI,EAAE,QAAQ;CACxC,MAAM,WAAW,UAAU,GAAG,EAAE,QAAQ,GAAG,YAAY,EAAE;CAEzD,MAAM,WAAW,EAAE,OAAO,SAAS,GAAG,EAAE,MAAM,KAAI,MAAK,EAAE,KAAK,CAAC,KAAK,IAAI,CAAC,OAAO;CAChF,MAAM,YAAY,EAAE,UAAU,KAAI,MAAK,EAAE,aAAa,GAAG,EAAE,KAAK,GAAG,EAAE,OAAO,CAAC,KAAK,KAAK;CACvF,MAAM,cAAc,eAAe,EAAE,SAAS,EAAE,WAAW;AAE3D,QAAO;EACL,GAAG,SAAS,GAAG,QAAQ,YAAY,aAAa,WAAW,UAAU,WAAW;EAChF,WAAW,QAAQ,GAAG,UAAU;EAChC,KAAK,YAAY,QAAQ,OAAO,OAAO;EACxC,CAAC,KAAK,KAAK;;;AAId,SAAgB,qBAAqB,GAAkB,MAAgE;CACrH,MAAM,YAAY,EAAE,UAAU,SAC1B,EAAE,SAAS,KAAI,MAAK,EAAE,aAAa,EAAE,KAAK,CAAC,KAAK,KAAK,GACrD;CACJ,MAAM,WAAW,EAAE,OAAO,SAAS,GAAG,EAAE,MAAM,KAAI,MAAK,EAAE,KAAK,CAAC,KAAK,IAAI,CAAC,OAAO;CAChF,MAAM,QAAQ,YAAY,GAAG,WAAW,cAAc,EAAE,OAAO,MAAM,IAAI,CAAC,KAAK,IAAI,EAAE;CAIrF,MAAM,OAAO,GAFG,kBAAkB,EAAE,QAAQ,WAAW,EAAE,SAEjC,GADN,EAAE,cAAc,EAAE,UAAU,IAAI,EAAE,cAAc,IAAI,EAAE,UAAU,GAAG,EAAE;CAIvF,MAAM,aAAa,OAAO;CAC1B,MAAM,YAAY,EAAE,QAAQ,MAAM,KAAK,CAAC,MAAK,MAAK,EAAE,MAAM,IAAI,EAAE,MAAM,KAAK,SAAS,CAAC,UAAU,KAAK,EAAE,MAAM,CAAC,CAAC,IAAI;AAGlH,QAAO;EAAE;EAAO;EAAM,SAFN,UAAU,SAAS,aAAa,GAAG,UAAU,MAAM,GAAG,aAAa,EAAE,CAAC,KAAK;EAE5D"}
@@ -163,4 +163,4 @@ declare function getCacheKey(name: string, version: string): string;
163
163
  declare function getCacheDir(name: string, version: string): string;
164
164
  //#endregion
165
165
  export { REPOS_DIR as A, writeToCache as C, CachedPackage as D, CachedDoc as E, getRepoCacheDir as M, CACHE_DIR as O, writeSections as S, CacheConfig as T, listCached as _, clearAllCache as a, readCachedSection as b, getPkgKeyFiles as c, isCached as d, linkCachedDir as f, linkShippedSkill as g, linkRepoCachedDir as h, ShippedSkill as i, getPackageDbPath as j, REFERENCES_DIR as k, getShippedSkills as l, linkPkgNamed as m, getCacheKey as n, clearCache as o, linkPkg as p, getVersionKey as r, ensureCacheDir as s, getCacheDir as t, hasShippedDocs as u, listReferenceFiles as v, writeToRepoCache as w, resolvePkgDir as x, readCachedDocs as y };
166
- //# sourceMappingURL=version.d.mts.map
166
+ //# sourceMappingURL=index.d.mts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.mts","names":[],"sources":["../../src/cache/config.ts","../../src/cache/types.ts","../../src/cache/storage.ts","../../src/cache/version.ts"],"mappings":";;AASA;;;cAAa,SAAA;;cAGA,cAAA;;cAGA,SAAA;;iBAGG,eAAA,CAAgB,KAAA,UAAe,IAAA;AAH/C;AAAA,iBAUgB,gBAAA,CAAiB,IAAA,UAAc,OAAA;;;;AAhB/C;;UCLiB,WAAA;EDKkC;ECHjD,IAAA;EDMW;ECJX,OAAA;AAAA;AAAA,UAGe,aAAA;EACf,IAAA;EACA,OAAA;EACA,GAAA;AAAA;AAAA,UAGe,SAAA;EACf,IAAA;EACA,OAAA;AAAA;;;;;ADPF;iBEiBgB,QAAA,CAAS,IAAA,UAAc,OAAA;;;;iBAOvB,cAAA,CAAA;;;;iBAQA,YAAA,CACd,IAAA,UACA,OAAA,UACA,IAAA,EAAM,SAAA;AF7BR;;;AAAA,iBE8CgB,gBAAA,CACd,KAAA,UACA,IAAA,UACA,IAAA,EAAM,SAAA;;AF1CR;;;iBE4DgB,iBAAA,CAAkB,QAAA,UAAkB,KAAA,UAAe,IAAA,UAAc,MAAA;;;;;ADjFjF;;;;;iBCuGgB,aAAA,CAAc,QAAA,UAAkB,IAAA,UAAc,OAAA,UAAiB,MAAA;;;;;iBAiB/D,aAAA,CAAc,IAAA,UAAc,GAAA,UAAa,OAAA;;;;AD3GzD;;;;;iBCkIgB,OAAA,CAAQ,QAAA,UAAkB,IAAA,UAAc,GAAA,UAAa,OAAA;;;AAtHrE;;;;;iBA4IgB,YAAA,CAAa,QAAA,UAAkB,IAAA,UAAc,GAAA,UAAa,OAAA;;;;;iBAmB1D,cAAA,CAAe,IAAA,UAAc,GAAA,UAAa,OAAA;;;;UAmCzC,YAAA;EACf,SAAA;EACA,QAAA;AAAA;;;AAjKF;iBAuKgB,gBAAA,CAAiB,IAAA,UAAc,GAAA,UAAa,OAAA,YAAmB,YAAA;;;;;;;iBAoB/D,aAAA,CAAc,IAAA,UAAc,OAAA,UAAiB,QAAA,EAAU,KAAA;EAAQ,IAAA;EAAc,OAAA;AAAA;;;;iBAY7E,iBAAA,CAAkB,IAAA,UAAc,OAAA,UAAiB,IAAA;;;;iBAUjD,gBAAA,CAAiB,OAAA,UAAiB,SAAA,UAAmB,SAAA;AAAA,iBAWrD,cAAA,CAAe,IAAA,UAAc,GAAA,UAAa,OAAA;;;;iBAiB1C,UAAA,CAAA,GAAc,aAAA;;;;iBAed,cAAA,CAAe,IAAA,UAAc,OAAA,WAAkB,SAAA;;AAhM/D;;iBA+NgB,UAAA,CAAW,IAAA,UAAc,OAAA;;;;iBAYzB,aAAA,CAAA;;;AApNhB;;iBAmOgB,kBAAA,CAAmB,QAAA,UAAkB,QAAA;;;;AF7WrD;;;;;iBGOgB,aAAA,CAAc,OAAA;;;;iBAOd,WAAA,CAAY,IAAA,UAAc,OAAA;AHR1C;;;;AAAA,iBGgBgB,WAAA,CAAY,IAAA,UAAc,OAAA"}
@@ -601,4 +601,4 @@ declare function parsePackageSpec(spec: string): {
601
601
  declare function extractBranchHint(url: string): string | undefined;
602
602
  //#endregion
603
603
  export { generateIssueIndex as $, ResolveStep as A, generateDiscussionIndex as At, resolveLocalPackageDocs as B, getFilePatterns as C, parseSkillFrontmatterName as Ct, getRepoKeyForPackage as D, GitHubDiscussion as Dt, getRepoEntry as E, generateDocsIndex as Et, getInstalledSkillVersion as F, extractSections as G, resolvePackageDocsWithAttempts as H, parseVersionSpecifier as I, normalizeLlmsLinks as J, fetchLlmsTxt as K, readLocalDependencies as L, fetchNpmPackage as M, toCrawlPattern as Mt, fetchNpmRegistryMeta as N, fetchBlogReleases as Nt, LocalPackageInfo as O, fetchGitHubDiscussions as Ot, fetchPkgDist as P, formatIssueAsMarkdown as Q, readLocalPackageInfo as R, getDocOverride as S, parseGitSkillInput as St, getRelatedPackages as T, resolveEntryFiles as Tt, searchNpmPackages as U, resolvePackageDocs as V, downloadLlmsDocs as W, GitHubIssue as X, parseMarkdownLinks as Y, fetchGitHubIssues as Z, BlogPreset as _, ResolveResult as _t, normalizeRepoUrl as a, fetchReadme as at, getBlogPreset as b, RemoteSkill as bt, verifyUrl as c, isShallowGitDocs as ct, SemVer as d, FetchedDoc as dt, isGhAvailable as et, compareSemver as f, LlmsContent as ft, parseSemver as g, ResolveAttempt as gt, isPrerelease as h, NpmPackageInfo as ht, isGitHubRepoUrl as i, fetchGitHubRepoMeta as it, fetchLatestVersion as j, fetchCrawledDocs as jt, ResolveOptions as k, formatDiscussionAsMarkdown as kt, GitHubRelease as l, resolveGitHubRepo as lt, generateReleaseIndex as m, LocalDependency as mt, extractBranchHint as n, MIN_GIT_DOCS as nt, parseGitHubUrl as o, fetchReadmeContent as ot, fetchReleaseNotes as p, LlmsLink as pt, fetchLlmsUrl as q, fetchText as r, fetchGitDocs as rt, parsePackageSpec as s, filterFrameworkDocs as st, $fetch as t, GitDocsResult as tt, ReleaseIndexOptions as u, validateGitDocsWithLlms as ut, BlogRelease as v, ResolvedPackage as vt, getPrereleaseChangelogRef as w, EntryFile as wt, getCrawlUrl as x, fetchGitSkills as xt, DocOverride as y, GitSkillSource as yt, resolveInstalledVersion as z };
604
- //# sourceMappingURL=utils.d.mts.map
604
+ //# sourceMappingURL=index2.d.mts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index2.d.mts","names":[],"sources":["../../src/sources/blog-releases.ts","../../src/sources/crawl.ts","../../src/sources/discussions.ts","../../src/sources/docs.ts","../../src/sources/entries.ts","../../src/sources/git-skills.ts","../../src/sources/types.ts","../../src/sources/github.ts","../../src/sources/issues.ts","../../src/sources/llms.ts","../../src/sources/npm.ts","../../src/sources/package-registry.ts","../../src/sources/releases.ts","../../src/sources/utils.ts"],"mappings":";;;UAmBU,WAAA;EACR,IAAA;EACA,OAAA;AAAA;;;;;;iBAkFoB,iBAAA,CACpB,WAAA,UACA,gBAAA,WACC,OAAA,CAAQ,WAAA;;;;;;AAzFV;;;;;AAsFD;;;AAtFC,iBCAqB,gBAAA,CACpB,GAAA,UACA,UAAA,IAAc,OAAA,mBACd,QAAA,YACC,OAAA,CAAQ,KAAA;EAAQ,IAAA;EAAc,OAAA;AAAA;;iBAmDjB,cAAA,CAAe,OAAA;;;;;;ADvD9B;;UEQgB,iBAAA;EACf,IAAA;EACA,MAAA;EACA,SAAA;EACA,YAAA;AAAA;AAAA,UAGe,gBAAA;EACf,MAAA;EACA,KAAA;EACA,IAAA;EACA,QAAA;EACA,SAAA;EACA,GAAA;EACA,WAAA;EACA,QAAA;EACA,YAAA;EACA,MAAA;EACA,WAAA,EAAa,iBAAA;AAAA;;;;;;iBA4GO,sBAAA,CACpB,KAAA,UACA,IAAA,UACA,KAAA,WACA,UAAA,WACA,QAAA,YACC,OAAA,CAAQ,gBAAA;ADrFX;;;AAAA,iBC6LgB,0BAAA,CAA2B,CAAA,EAAG,gBAAA;;;;;iBAuC9B,uBAAA,CAAwB,WAAA,EAAa,gBAAA;;;;;;AF3RpD;;;;;AAAA,iBGNe,iBAAA,CAAkB,IAAA,EAAM,KAAA;EAAQ,IAAA;EAAc,OAAA;AAAA;;;UCH7C,SAAA;EACf,IAAA;EACA,OAAA;EACA,IAAA;AAAA;;;;iBAoCoB,iBAAA,CAAkB,UAAA,WAAqB,OAAA,CAAQ,SAAA;;;;;;AJ9BpE;;;UKHgB,cAAA;EACf,IAAA;EACA,KAAA;EACA,IAAA;;EAEA,SAAA;ELqFA;EKnFA,GAAA;ELqFC;EKnFD,SAAA;AAAA;AAAA,UAGe,WAAA;;EAEf,IAAA;;EAEA,WAAA;EJboC;EIepC,IAAA;EJXQ;EIaR,OAAA;EJfc;EIiBd,KAAA,EAAO,KAAA;IAAQ,IAAA;IAAc,OAAA;EAAA;AAAA;;;;AJoC/B;iBI7BgB,kBAAA,CAAmB,KAAA,WAAgB,cAAA;;;;iBA8EnC,yBAAA,CAA0B,OAAA;EAAoB,IAAA;EAAe,WAAA;AAAA;;;;iBA0BvD,cAAA,CACpB,MAAA,EAAQ,cAAA,EACR,UAAA,IAAc,GAAA,oBACb,OAAA;EAAU,MAAA,EAAQ,WAAA;AAAA;;;;;;UClJJ,cAAA;EACf,IAAA;EACA,OAAA;EACA,WAAA;EACA,QAAA;EACA,UAAA;IACE,IAAA;IACA,GAAA;IACA,SAAA;EAAA;EAEF,MAAA;EACA,YAAA,GAAe,MAAA;EACf,eAAA,GAAkB,MAAA;EAClB,gBAAA,GAAmB,MAAA;AAAA;AAAA,UAGJ,eAAA;EACf,IAAA;EACA,OAAA;ELLoB;EKOpB,UAAA;EACA,WAAA;ELJQ;EKMR,YAAA,GAAe,MAAA;ELRD;EKUd,QAAA,GAAW,MAAA;IAAiB,OAAA;IAAiB,UAAA;EAAA;EAC7C,OAAA;EACA,OAAA;EACA,SAAA;EACA,OAAA;ELuCc;EKrCd,UAAA;;EAEA,MAAA;ELmC4C;EKjC5C,eAAA;;EAEA,QAAA;AAAA;AAAA,UAGe,eAAA;EACf,IAAA;EACA,OAAA;AAAA;AAAA,UAGe,WAAA;EACf,GAAA;EJrBA;EIuBA,KAAA,EAAO,QAAA;AAAA;AAAA,UAGQ,QAAA;EACf,KAAA;EACA,GAAA;AAAA;AAAA,UAGe,UAAA;EACf,GAAA;EACA,KAAA;EACA,OAAA;AAAA;AAAA,UAGe,cAAA;EACf,MAAA;EACA,GAAA;EACA,MAAA;EACA,OAAA;AAAA;AAAA,UAGe,aAAA;EACf,OAAA,EAAS,eAAA;EACT,QAAA,EAAU,cAAA;AAAA;;;;cC5DC,YAAA;;cAGA,gBAAA,GAAoB,CAAA;AAAA,UAEhB,aAAA;EPmFsB;EOjFrC,OAAA;EPoFQ;EOlFR,GAAA;EPiFA;EO/EA,KAAA;EPgFS;EO9ET,UAAA;EP8EkB;EO5ElB,QAAA;;EAEA,QAAA;AAAA;;;;;;;iBA0Fc,mBAAA,CAAoB,KAAA,YAAiB,WAAA;;;;;iBAyL/B,YAAA,CAAa,KAAA,UAAe,IAAA,UAAc,OAAA,UAAiB,WAAA,WAAsB,OAAA,YAAmB,OAAA,CAAQ,aAAA;;AN3OlI;;;;;iBM6SgB,uBAAA,CACd,SAAA,EAAW,QAAA,IACX,SAAA;EACG,OAAA;EAAkB,UAAA;AAAA;;;;;iBA6ID,mBAAA,CAAoB,KAAA,UAAe,IAAA,UAAc,WAAA,YAAuB,OAAA;EAAU,QAAA;AAAA;;ALrexG;;iBKogBsB,WAAA,CAAY,KAAA,UAAe,IAAA,UAAc,MAAA,WAAiB,GAAA,YAAe,OAAA;AL7Y/F;;;AAAA,iBK0fsB,kBAAA,CAAmB,GAAA,WAAc,OAAA;;;;;iBAiDjC,iBAAA,CACpB,KAAA,UACA,IAAA,UACA,UAAA,IAAc,GAAA,oBACb,OAAA,CAAQ,eAAA;;;;;;APrrBV;;KQNW,SAAA;AAAA,UAEK,YAAA;EACf,IAAA;EACA,MAAA;EACA,SAAA;EACA,YAAA;AAAA;AAAA,UAGe,WAAA;EACf,MAAA;EACA,KAAA;EACA,KAAA;EACA,MAAA;EACA,IAAA;EACA,SAAA;EACA,GAAA;EACA,SAAA;EACA,QAAA;EACA,IAAA,EAAM,SAAA;EACN,WAAA,EAAa,YAAA;EPVL;EOYR,KAAA;EPdc;EOgBd,UAAA;AAAA;;;;iBAQc,aAAA,CAAA;ANXhB;;;;;AAAA,iBMobsB,iBAAA,CACpB,KAAA,UACA,IAAA,UACA,KAAA,WACA,UAAA,WACA,QAAA,YACC,OAAA,CAAQ,WAAA;;;;iBAwBK,qBAAA,CAAsB,KAAA,EAAO,WAAA;;;;;iBA0C7B,kBAAA,CAAmB,MAAA,EAAQ,WAAA;;;;;;iBChhBrB,YAAA,CAAa,OAAA,WAAkB,OAAA;AT2FrD;;;AAAA,iBShFsB,YAAA,CAAa,GAAA,WAAc,OAAA,CAAQ,WAAA;;;;iBAczC,kBAAA,CAAmB,OAAA,WAAkB,QAAA;AAAA,iBA4B/B,gBAAA,CACpB,WAAA,EAAa,WAAA,EACb,OAAA,UACA,UAAA,IAAc,GAAA,UAAa,KAAA,UAAe,KAAA,oBACzC,OAAA,CAAQ,UAAA;;;;;iBAgCK,kBAAA,CAAmB,OAAA,UAAiB,OAAA;;;;;iBAuBpC,eAAA,CAAgB,OAAA,UAAiB,QAAA;;;;;;;iBCvG3B,iBAAA,CAAkB,KAAA,UAAe,IAAA,YAAW,OAAA,CAAQ,KAAA;EAAQ,IAAA;EAAc,WAAA;EAAsB,OAAA;AAAA;;;;iBAkBhG,eAAA,CAAgB,WAAA,WAAsB,OAAA,CAAQ,cAAA;AAAA,UAUnD,WAAA;EACf,OAAA;EACA,UAAA;AAAA;AAAA,UAGe,eAAA;EACf,UAAA;EACA,QAAA,GAAW,MAAA,SAAe,WAAA;AAAA;;;;iBAMN,oBAAA,CAAqB,WAAA,UAAqB,OAAA,WAAkB,OAAA,CAAQ,eAAA;AAAA,KA2B9E,WAAA;AAAA,UAEK,cAAA;ETtEE;ESwEjB,OAAA;;EAEA,GAAA;ETvBc;ESyBd,UAAA,IAAc,IAAA,EAAM,WAAA;AAAA;;;;iBA+FA,kBAAA,CAAmB,WAAA,UAAqB,OAAA,GAAS,cAAA,GAAsB,OAAA,CAAQ,eAAA;;ARvKrG;;iBQ+KsB,8BAAA,CAA+B,WAAA,UAAqB,OAAA,GAAS,cAAA,GAAsB,OAAA,CAAQ,aAAA;;;;iBAwLjG,qBAAA,CACd,IAAA,UACA,OAAA,UACA,GAAA,WACC,eAAA;;;;ARpWH;iBQwZgB,uBAAA,CAAwB,IAAA,UAAc,GAAA;;;;iBA6BhC,qBAAA,CAAsB,GAAA,WAAc,OAAA,CAAQ,eAAA;AAAA,UAwBjD,gBAAA;EACf,IAAA;EACA,OAAA;EACA,WAAA;EACA,OAAA;EACA,SAAA;AAAA;;;;iBAMc,oBAAA,CAAqB,SAAA,WAAoB,gBAAA;;ARjWzD;;iBQ4XsB,uBAAA,CAAwB,SAAA,WAAoB,OAAA,CAAQ,eAAA;;;;;;;;iBAsDpD,YAAA,CAAa,IAAA,UAAc,OAAA,WAAkB,OAAA;;;ARpUnE;iBQ4YsB,kBAAA,CAAmB,WAAA,WAAsB,OAAA;;;;iBAU/C,wBAAA,CAAyB,QAAA;;;;;;AV1oBxC;;UWXgB,WAAA;EACf,OAAA;EACA,GAAA;EACA,IAAA;EACA,KAAA;AAAA;AAAA,UAGe,YAAA;EACf,YAAA;EACA,OAAA;EX2FC;EWzFD,KAAA;AAAA;AAAA,UAGe,SAAA;EACf,KAAA;EACA,IAAA;;EAEA,QAAA;EVPoC;EUSpC,QAAA;EVLQ;EUOR,OAAA;EVTc;EUWd,QAAA;EVVA;EUYA,QAAA;EVXS;EUaT,sBAAA;EVb+B;EUe/B,QAAA,EAAU,MAAA,SAAe,YAAA;EVfa;EUiBtC,YAAA,GAAe,WAAA;AAAA;AAAA,UAIA,WAAA;EACf,KAAA;EACA,IAAA;EACA,IAAA;EACA,GAAA;EACA,QAAA;AAAA;AAAA,UAGe,UAAA;EACf,WAAA;EACA,QAAA,EAAU,WAAA;AAAA;AAAA,iBAkYI,cAAA,CAAe,WAAA,WAAsB,WAAA;AAAA,iBAiBrC,aAAA,CAAc,WAAA,WAAsB,UAAA;AAAA,iBAcpC,eAAA,CAAgB,WAAA;AAAA,iBAShB,YAAA,CAAa,OAAA,WAAkB,SAAA;AAAA,iBAI/B,oBAAA,CAAqB,WAAA;AAAA,iBAWrB,yBAAA,CAA0B,WAAA;AAAA,iBAO1B,WAAA,CAAY,WAAA;AAAA,iBAOZ,kBAAA,CAAmB,WAAA;;;;;;UClflB,aAAA;EACf,EAAA;EACA,GAAA;EACA,IAAA;EACA,UAAA;EACA,SAAA;EACA,WAAA;EACA,QAAA;AAAA;AAAA,UAOQ,SAAA;EACR,IAAA;EACA,OAAA;AAAA;AAAA,UAGe,MAAA;EACf,KAAA;EACA,KAAA;EACA,KAAA;EACA,GAAA;AAAA;AAAA,iBAGc,WAAA,CAAY,OAAA,WAAkB,MAAA;;;;iBAiD9B,YAAA,CAAa,OAAA;AAAA,iBAIb,aAAA,CAAc,CAAA,EAAG,MAAA,EAAQ,CAAA,EAAG,MAAA;AAAA,UAoI3B,mBAAA;EACf,QAAA,EAAU,aAAA;EACV,WAAA;EACA,YAAA,GAAe,KAAA;IAAQ,OAAA;IAAiB,KAAA;IAAe,IAAA;EAAA;EACvD,YAAA;AAAA;;;;;iBAOc,oBAAA,CAAqB,cAAA,EAAgB,aAAA,KAAkB,mBAAA,EAAqB,WAAA;;;;;;;;iBA6GtE,iBAAA,CACpB,KAAA,UACA,IAAA,UACA,gBAAA,UACA,MAAA,WACA,WAAA,WACA,QAAA,WACA,YAAA,YACC,OAAA,CAAQ,SAAA;;;;;;cCtVE,MAAA,EAKX,QAAA,CALiB,MAAA;;;;iBAUG,SAAA,CAAU,GAAA,WAAc,OAAA;AbuF9C;;;AAAA,iBahFsB,SAAA,CAAU,GAAA,WAAc,OAAA;;;;iBAkC9B,eAAA,CAAgB,GAAA;;;;iBAahB,cAAA,CAAe,GAAA;EAAgB,KAAA;EAAe,IAAA;AAAA;;;;iBAU9C,gBAAA,CAAiB,GAAA;;;;;iBAejB,gBAAA,CAAiB,IAAA;EAAiB,IAAA;EAAc,GAAA;AAAA;;;;iBAqBhD,iBAAA,CAAkB,GAAA"}
@@ -0,0 +1,15 @@
1
+ import "./config.mjs";
2
+ import "./sanitize.mjs";
3
+ import "./cache.mjs";
4
+ import "./yaml.mjs";
5
+ import "./markdown.mjs";
6
+ import "./shared.mjs";
7
+ import "./sources.mjs";
8
+ import "./prompts.mjs";
9
+ import "./agent.mjs";
10
+ import "./skills.mjs";
11
+ import "./formatting.mjs";
12
+ import "./pool.mjs";
13
+ import "./sync.mjs";
14
+ import { u as installCommandDef } from "../cli.mjs";
15
+ export { installCommandDef };
@@ -0,0 +1,13 @@
1
+ import "./config.mjs";
2
+ import "./sanitize.mjs";
3
+ import "./cache.mjs";
4
+ import "./yaml.mjs";
5
+ import "./markdown.mjs";
6
+ import "./shared.mjs";
7
+ import "./sources.mjs";
8
+ import "./prompts.mjs";
9
+ import "./agent.mjs";
10
+ import "./skills.mjs";
11
+ import "./formatting.mjs";
12
+ import { c as listCommandDef } from "../cli.mjs";
13
+ export { listCommandDef };
@@ -4,6 +4,7 @@ import { frontmatterFromMarkdown } from "mdast-util-frontmatter";
4
4
  import { toString } from "mdast-util-to-string";
5
5
  import { frontmatter } from "micromark-extension-frontmatter";
6
6
  import { visit } from "unist-util-visit";
7
+ /** Parse markdown string to AST + frontmatter key-values */
7
8
  function parseMd(content) {
8
9
  const tree = fromMarkdown(content, {
9
10
  extensions: [frontmatter(["yaml"])],
@@ -21,12 +22,15 @@ function parseMd(content) {
21
22
  frontmatter: fm
22
23
  };
23
24
  }
25
+ /** Extract frontmatter key-value pairs only */
24
26
  function parseFrontmatter(content) {
25
27
  return parseMd(content).frontmatter;
26
28
  }
29
+ /** Strip custom heading anchors like {#some-id} */
27
30
  function stripHeadingAnchors(text) {
28
31
  return text.replace(/\s*\{#[^}]+\}\s*$/, "").trim();
29
32
  }
33
+ /** Extract title: frontmatter title > first h1 > null */
30
34
  function extractTitle(content) {
31
35
  const { tree, frontmatter: fm } = parseMd(content);
32
36
  if (fm.title) return fm.title;
@@ -39,6 +43,7 @@ function extractTitle(content) {
39
43
  });
40
44
  return title;
41
45
  }
46
+ /** Extract first paragraph text, 150 char max */
42
47
  function extractDescription(content) {
43
48
  const { tree } = parseMd(content);
44
49
  let desc = null;
@@ -52,6 +57,7 @@ function extractDescription(content) {
52
57
  });
53
58
  return desc;
54
59
  }
60
+ /** Extract all links (deduped by url) */
55
61
  function extractLinks(content) {
56
62
  const { tree } = parseMd(content);
57
63
  const links = [];
@@ -67,6 +73,7 @@ function extractLinks(content) {
67
73
  });
68
74
  return links;
69
75
  }
76
+ /** Strip frontmatter block, return body only */
70
77
  function stripFrontmatter(content) {
71
78
  const match = content.match(/^---\r?\n[\s\S]*?\r?\n---\r?\n/);
72
79
  return match ? content.slice(match[0].length).trim() : content;
@@ -1 +1 @@
1
- {"version":3,"file":"markdown.mjs","names":[],"sources":["../../src/core/markdown.ts"],"sourcesContent":["/**\n * AST-based markdown parsing using mdast/micromark.\n * Replaces scattered regex-based frontmatter/heading/link extraction.\n */\n\nimport type { Nodes, Root } from 'mdast'\nimport { fromMarkdown } from 'mdast-util-from-markdown'\nimport { frontmatterFromMarkdown } from 'mdast-util-frontmatter'\nimport { toString } from 'mdast-util-to-string'\nimport { frontmatter } from 'micromark-extension-frontmatter'\nimport { visit } from 'unist-util-visit'\nimport { yamlParseKV } from './yaml.ts'\n\nexport interface MdHeading {\n depth: number\n text: string\n}\n\nexport interface MdLink {\n title: string\n url: string\n}\n\nexport interface ParsedMd {\n tree: Root\n frontmatter: Record<string, string>\n}\n\n/** Parse markdown string to AST + frontmatter key-values */\nexport function parseMd(content: string): ParsedMd {\n const tree = fromMarkdown(content, {\n extensions: [frontmatter(['yaml'])],\n mdastExtensions: [frontmatterFromMarkdown(['yaml'])],\n })\n\n const fm: Record<string, string> = {}\n visit(tree, 'yaml', (node: Nodes) => {\n if (node.type === 'yaml') {\n for (const line of (node as any).value.split('\\n')) {\n const kv = yamlParseKV(line)\n if (kv)\n fm[kv[0]] = kv[1]\n }\n }\n })\n\n return { tree, frontmatter: fm }\n}\n\n/** Extract frontmatter key-value pairs only */\nexport function parseFrontmatter(content: string): Record<string, string> {\n return parseMd(content).frontmatter\n}\n\n/** Strip custom heading anchors like {#some-id} */\nfunction stripHeadingAnchors(text: string): string {\n return text.replace(/\\s*\\{#[^}]+\\}\\s*$/, '').trim()\n}\n\n/** Extract title: frontmatter title > first h1 > null */\nexport function extractTitle(content: string): string | null {\n const { tree, frontmatter: fm } = parseMd(content)\n if (fm.title)\n return fm.title\n\n let title: string | null = null\n visit(tree, 'heading', (node) => {\n if (node.depth === 1 && !title) {\n // Strip {#id} anchors and leading backslash escapes (e.g. `# \\`)\n const text = stripHeadingAnchors(toString(node)).replace(/^\\\\+\\s*/, '').trim()\n if (text.length > 0)\n title = text\n }\n })\n\n return title\n}\n\n/** Extract first paragraph text, 150 char max */\nexport function extractDescription(content: string): string | null {\n const { tree } = parseMd(content)\n\n let desc: string | null = null\n visit(tree, 'paragraph', (node, _index, parent) => {\n // Only top-level paragraphs (skip blockquote children, list items, etc.)\n if (desc || parent?.type !== 'root')\n return\n\n const text = toString(node).trim()\n if (text.length === 0)\n return\n\n // Strip markdown link syntax remnants and formatting chars\n let clean = text.replace(/\\[([^\\]]+)\\]\\([^)]+\\)/g, '$1').replace(/[`*_~]/g, '')\n if (clean.length > 150)\n clean = `${clean.slice(0, 147)}...`\n desc = clean\n })\n\n return desc\n}\n\n/** Extract all headings with depth and text */\nexport function extractHeadings(content: string): MdHeading[] {\n const { tree } = parseMd(content)\n const headings: MdHeading[] = []\n\n visit(tree, 'heading', (node) => {\n headings.push({ depth: node.depth, text: stripHeadingAnchors(toString(node)) })\n })\n\n return headings\n}\n\n/** Extract all links (deduped by url) */\nexport function extractLinks(content: string): MdLink[] {\n const { tree } = parseMd(content)\n const links: MdLink[] = []\n const seen = new Set<string>()\n\n visit(tree, 'link', (node) => {\n if (!seen.has(node.url)) {\n seen.add(node.url)\n links.push({ title: toString(node), url: node.url })\n }\n })\n\n return links\n}\n\n/** Strip frontmatter block, return body only */\nexport function stripFrontmatter(content: string): string {\n const match = content.match(/^---\\r?\\n[\\s\\S]*?\\r?\\n---\\r?\\n/)\n return match ? content.slice(match[0].length).trim() : content\n}\n"],"mappings":";;;;;;AA6BA,SAAgB,QAAQ,SAA2B;CACjD,MAAM,OAAO,aAAa,SAAS;EACjC,YAAY,CAAC,YAAY,CAAC,OAAO,CAAC,CAAC;EACnC,iBAAiB,CAAC,wBAAwB,CAAC,OAAO,CAAC,CAAA;EACpD,CAAC;CAEF,MAAM,KAA6B,EAAE;AACrC,OAAM,MAAM,SAAS,SAAgB;AACnC,MAAI,KAAK,SAAS,OAChB,MAAK,MAAM,QAAS,KAAa,MAAM,MAAM,KAAK,EAAE;GAClD,MAAM,KAAK,YAAY,KAAK;AAC5B,OAAI,GACF,IAAG,GAAG,MAAM,GAAG;;GAGrB;AAEF,QAAO;EAAE;EAAM,aAAa;EAAI;;AAIlC,SAAgB,iBAAiB,SAAyC;AACxE,QAAO,QAAQ,QAAQ,CAAC;;AAI1B,SAAS,oBAAoB,MAAsB;AACjD,QAAO,KAAK,QAAQ,qBAAqB,GAAG,CAAC,MAAM;;AAIrD,SAAgB,aAAa,SAAgC;CAC3D,MAAM,EAAE,MAAM,aAAa,OAAO,QAAQ,QAAQ;AAClD,KAAI,GAAG,MACL,QAAO,GAAG;CAEZ,IAAI,QAAuB;AAC3B,OAAM,MAAM,YAAY,SAAS;AAC/B,MAAI,KAAK,UAAU,KAAK,CAAC,OAAO;GAE9B,MAAM,OAAO,oBAAoB,SAAS,KAAK,CAAC,CAAC,QAAQ,WAAW,GAAG,CAAC,MAAM;AAC9E,OAAI,KAAK,SAAS,EAChB,SAAQ;;GAEZ;AAEF,QAAO;;AAIT,SAAgB,mBAAmB,SAAgC;CACjE,MAAM,EAAE,SAAS,QAAQ,QAAQ;CAEjC,IAAI,OAAsB;AAC1B,OAAM,MAAM,cAAc,MAAM,QAAQ,WAAW;AAEjD,MAAI,QAAQ,QAAQ,SAAS,OAC3B;EAEF,MAAM,OAAO,SAAS,KAAK,CAAC,MAAM;AAClC,MAAI,KAAK,WAAW,EAClB;EAGF,IAAI,QAAQ,KAAK,QAAQ,0BAA0B,KAAK,CAAC,QAAQ,WAAW,GAAG;AAC/E,MAAI,MAAM,SAAS,IACjB,SAAQ,GAAG,MAAM,MAAM,GAAG,IAAI,CAAC;AACjC,SAAO;GACP;AAEF,QAAO;;AAgBT,SAAgB,aAAa,SAA2B;CACtD,MAAM,EAAE,SAAS,QAAQ,QAAQ;CACjC,MAAM,QAAkB,EAAE;CAC1B,MAAM,uBAAO,IAAI,KAAa;AAE9B,OAAM,MAAM,SAAS,SAAS;AAC5B,MAAI,CAAC,KAAK,IAAI,KAAK,IAAI,EAAE;AACvB,QAAK,IAAI,KAAK,IAAI;AAClB,SAAM,KAAK;IAAE,OAAO,SAAS,KAAK;IAAE,KAAK,KAAK;IAAK,CAAC;;GAEtD;AAEF,QAAO;;AAIT,SAAgB,iBAAiB,SAAyB;CACxD,MAAM,QAAQ,QAAQ,MAAM,iCAAiC;AAC7D,QAAO,QAAQ,QAAQ,MAAM,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG"}
1
+ {"version":3,"file":"markdown.mjs","names":[],"sources":["../../src/core/markdown.ts"],"sourcesContent":["/**\n * AST-based markdown parsing using mdast/micromark.\n * Replaces scattered regex-based frontmatter/heading/link extraction.\n */\n\nimport type { Nodes, Root } from 'mdast'\nimport { fromMarkdown } from 'mdast-util-from-markdown'\nimport { frontmatterFromMarkdown } from 'mdast-util-frontmatter'\nimport { toString } from 'mdast-util-to-string'\nimport { frontmatter } from 'micromark-extension-frontmatter'\nimport { visit } from 'unist-util-visit'\nimport { yamlParseKV } from './yaml.ts'\n\nexport interface MdHeading {\n depth: number\n text: string\n}\n\nexport interface MdLink {\n title: string\n url: string\n}\n\nexport interface ParsedMd {\n tree: Root\n frontmatter: Record<string, string>\n}\n\n/** Parse markdown string to AST + frontmatter key-values */\nexport function parseMd(content: string): ParsedMd {\n const tree = fromMarkdown(content, {\n extensions: [frontmatter(['yaml'])],\n mdastExtensions: [frontmatterFromMarkdown(['yaml'])],\n })\n\n const fm: Record<string, string> = {}\n visit(tree, 'yaml', (node: Nodes) => {\n if (node.type === 'yaml') {\n for (const line of (node as any).value.split('\\n')) {\n const kv = yamlParseKV(line)\n if (kv)\n fm[kv[0]] = kv[1]\n }\n }\n })\n\n return { tree, frontmatter: fm }\n}\n\n/** Extract frontmatter key-value pairs only */\nexport function parseFrontmatter(content: string): Record<string, string> {\n return parseMd(content).frontmatter\n}\n\n/** Strip custom heading anchors like {#some-id} */\nfunction stripHeadingAnchors(text: string): string {\n return text.replace(/\\s*\\{#[^}]+\\}\\s*$/, '').trim()\n}\n\n/** Extract title: frontmatter title > first h1 > null */\nexport function extractTitle(content: string): string | null {\n const { tree, frontmatter: fm } = parseMd(content)\n if (fm.title)\n return fm.title\n\n let title: string | null = null\n visit(tree, 'heading', (node) => {\n if (node.depth === 1 && !title) {\n // Strip {#id} anchors and leading backslash escapes (e.g. `# \\`)\n const text = stripHeadingAnchors(toString(node)).replace(/^\\\\+\\s*/, '').trim()\n if (text.length > 0)\n title = text\n }\n })\n\n return title\n}\n\n/** Extract first paragraph text, 150 char max */\nexport function extractDescription(content: string): string | null {\n const { tree } = parseMd(content)\n\n let desc: string | null = null\n visit(tree, 'paragraph', (node, _index, parent) => {\n // Only top-level paragraphs (skip blockquote children, list items, etc.)\n if (desc || parent?.type !== 'root')\n return\n\n const text = toString(node).trim()\n if (text.length === 0)\n return\n\n // Strip markdown link syntax remnants and formatting chars\n let clean = text.replace(/\\[([^\\]]+)\\]\\([^)]+\\)/g, '$1').replace(/[`*_~]/g, '')\n if (clean.length > 150)\n clean = `${clean.slice(0, 147)}...`\n desc = clean\n })\n\n return desc\n}\n\n/** Extract all headings with depth and text */\nexport function extractHeadings(content: string): MdHeading[] {\n const { tree } = parseMd(content)\n const headings: MdHeading[] = []\n\n visit(tree, 'heading', (node) => {\n headings.push({ depth: node.depth, text: stripHeadingAnchors(toString(node)) })\n })\n\n return headings\n}\n\n/** Extract all links (deduped by url) */\nexport function extractLinks(content: string): MdLink[] {\n const { tree } = parseMd(content)\n const links: MdLink[] = []\n const seen = new Set<string>()\n\n visit(tree, 'link', (node) => {\n if (!seen.has(node.url)) {\n seen.add(node.url)\n links.push({ title: toString(node), url: node.url })\n }\n })\n\n return links\n}\n\n/** Strip frontmatter block, return body only */\nexport function stripFrontmatter(content: string): string {\n const match = content.match(/^---\\r?\\n[\\s\\S]*?\\r?\\n---\\r?\\n/)\n return match ? content.slice(match[0].length).trim() : content\n}\n"],"mappings":";;;;;;;AA6BA,SAAgB,QAAQ,SAA2B;CACjD,MAAM,OAAO,aAAa,SAAS;EACjC,YAAY,CAAC,YAAY,CAAC,OAAO,CAAC,CAAC;EACnC,iBAAiB,CAAC,wBAAwB,CAAC,OAAO,CAAC,CAAA;EACpD,CAAC;CAEF,MAAM,KAA6B,EAAE;AACrC,OAAM,MAAM,SAAS,SAAgB;AACnC,MAAI,KAAK,SAAS,OAChB,MAAK,MAAM,QAAS,KAAa,MAAM,MAAM,KAAK,EAAE;GAClD,MAAM,KAAK,YAAY,KAAK;AAC5B,OAAI,GACF,IAAG,GAAG,MAAM,GAAG;;GAGrB;AAEF,QAAO;EAAE;EAAM,aAAa;EAAI;;;AAIlC,SAAgB,iBAAiB,SAAyC;AACxE,QAAO,QAAQ,QAAQ,CAAC;;;AAI1B,SAAS,oBAAoB,MAAsB;AACjD,QAAO,KAAK,QAAQ,qBAAqB,GAAG,CAAC,MAAM;;;AAIrD,SAAgB,aAAa,SAAgC;CAC3D,MAAM,EAAE,MAAM,aAAa,OAAO,QAAQ,QAAQ;AAClD,KAAI,GAAG,MACL,QAAO,GAAG;CAEZ,IAAI,QAAuB;AAC3B,OAAM,MAAM,YAAY,SAAS;AAC/B,MAAI,KAAK,UAAU,KAAK,CAAC,OAAO;GAE9B,MAAM,OAAO,oBAAoB,SAAS,KAAK,CAAC,CAAC,QAAQ,WAAW,GAAG,CAAC,MAAM;AAC9E,OAAI,KAAK,SAAS,EAChB,SAAQ;;GAEZ;AAEF,QAAO;;;AAIT,SAAgB,mBAAmB,SAAgC;CACjE,MAAM,EAAE,SAAS,QAAQ,QAAQ;CAEjC,IAAI,OAAsB;AAC1B,OAAM,MAAM,cAAc,MAAM,QAAQ,WAAW;AAEjD,MAAI,QAAQ,QAAQ,SAAS,OAC3B;EAEF,MAAM,OAAO,SAAS,KAAK,CAAC,MAAM;AAClC,MAAI,KAAK,WAAW,EAClB;EAGF,IAAI,QAAQ,KAAK,QAAQ,0BAA0B,KAAK,CAAC,QAAQ,WAAW,GAAG;AAC/E,MAAI,MAAM,SAAS,IACjB,SAAQ,GAAG,MAAM,MAAM,GAAG,IAAI,CAAC;AACjC,SAAO;GACP;AAEF,QAAO;;;AAgBT,SAAgB,aAAa,SAA2B;CACtD,MAAM,EAAE,SAAS,QAAQ,QAAQ;CACjC,MAAM,QAAkB,EAAE;CAC1B,MAAM,uBAAO,IAAI,KAAa;AAE9B,OAAM,MAAM,SAAS,SAAS;AAC5B,MAAI,CAAC,KAAK,IAAI,KAAK,IAAI,EAAE;AACvB,QAAK,IAAI,KAAK,IAAI;AAClB,SAAM,KAAK;IAAE,OAAO,SAAS,KAAK;IAAE,KAAK,KAAK;IAAK,CAAC;;GAEtD;AAEF,QAAO;;;AAIT,SAAgB,iBAAiB,SAAyB;CACxD,MAAM,QAAQ,QAAQ,MAAM,iCAAiC;AAC7D,QAAO,QAAQ,QAAQ,MAAM,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG"}
@@ -117,4 +117,4 @@ async function shutdownWorker() {
117
117
  }
118
118
  export { shutdownWorker as n, pool_exports as t };
119
119
 
120
- //# sourceMappingURL=pool2.mjs.map
120
+ //# sourceMappingURL=pool.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"pool.mjs","names":[],"sources":["../../src/retriv/pool.ts"],"sourcesContent":["import type { IndexConfig, Document as RetrivDocument } from './types.ts'\nimport type { WorkerMessage, WorkerResponse } from './worker.ts'\nimport { existsSync } from 'node:fs'\nimport { fileURLToPath } from 'node:url'\nimport { Worker } from 'node:worker_threads'\nimport { dirname, join } from 'pathe'\n\ninterface PendingTask {\n id: number\n resolve: () => void\n reject: (err: Error) => void\n onProgress?: IndexConfig['onProgress']\n}\n\nlet worker: Worker | null = null\nlet taskId = 0\nconst pending = new Map<number, PendingTask>()\nconst queue: Array<() => void> = []\nlet running = false\n\nfunction resolveWorkerPath(): { path: string, execArgv?: string[] } {\n const dir = dirname(fileURLToPath(import.meta.url))\n\n // Bundled: dist/retriv/worker.mjs (resolve from package root, not chunk dir)\n for (const candidate of [join(dir, 'worker.mjs'), join(dir, '..', 'retriv', 'worker.mjs')]) {\n if (existsSync(candidate))\n return { path: candidate }\n }\n\n // Dev stub: src/retriv/pool.ts → src/retriv/worker.ts\n return { path: join(dir, 'worker.ts'), execArgv: ['--experimental-strip-types'] }\n}\n\nfunction ensureWorker(): Worker {\n if (worker)\n return worker\n\n const config = resolveWorkerPath()\n const w = new Worker(config.path, {\n execArgv: config.execArgv,\n })\n\n w.on('message', (msg: WorkerResponse) => {\n const task = pending.get(msg.id)\n if (!task)\n return\n\n if (msg.type === 'progress') {\n task.onProgress?.({ phase: msg.phase as any, current: msg.current, total: msg.total })\n }\n else if (msg.type === 'done') {\n pending.delete(msg.id)\n task.resolve()\n }\n else if (msg.type === 'error') {\n pending.delete(msg.id)\n task.reject(new Error(msg.message))\n }\n })\n\n w.on('error', (err: Error) => {\n for (const task of pending.values())\n task.reject(err)\n pending.clear()\n worker = null\n })\n\n w.on('exit', (code) => {\n if (pending.size > 0) {\n const err = new Error(`Worker exited (code ${code}) with ${pending.size} pending tasks`)\n for (const task of pending.values())\n task.reject(err)\n pending.clear()\n }\n worker = null\n })\n\n worker = w\n return w\n}\n\nfunction drainQueue() {\n if (running || queue.length === 0)\n return\n const next = queue.shift()!\n next()\n}\n\nexport async function createIndexInWorker(\n documents: RetrivDocument[],\n config: IndexConfig,\n): Promise<void> {\n return new Promise<void>((resolve, reject) => {\n const run = () => {\n running = true\n const id = ++taskId\n\n let w: Worker\n try {\n w = ensureWorker()\n }\n catch (err) {\n running = false\n drainQueue()\n reject(err instanceof Error ? err : new Error(String(err)))\n return\n }\n\n pending.set(id, {\n id,\n resolve: () => {\n running = false\n drainQueue()\n resolve()\n },\n reject: (err) => {\n running = false\n drainQueue()\n reject(err)\n },\n onProgress: config.onProgress,\n })\n\n const msg: WorkerMessage = {\n type: 'index',\n id,\n documents,\n dbPath: config.dbPath,\n }\n\n w.postMessage(msg)\n }\n\n if (running) {\n queue.push(run)\n }\n else {\n run()\n }\n })\n}\n\nexport async function shutdownWorker(): Promise<void> {\n if (!worker)\n return\n\n const w = worker\n worker = null\n\n return new Promise<void>((resolve) => {\n const timeout = setTimeout(() => {\n w.terminate().then(() => resolve(), () => resolve())\n }, 5000)\n\n w.once('exit', () => {\n clearTimeout(timeout)\n resolve()\n })\n\n w.postMessage({ type: 'shutdown' } satisfies WorkerMessage)\n })\n}\n"],"mappings":";;;;;;;;;AAcA,IAAI,SAAwB;AAC5B,IAAI,SAAS;AACb,MAAM,0BAAU,IAAI,KAA0B;AAC9C,MAAM,QAA2B,EAAE;AACnC,IAAI,UAAU;AAEd,SAAS,oBAA2D;CAClE,MAAM,MAAM,QAAQ,cAAc,OAAO,KAAK,IAAI,CAAC;AAGnD,MAAK,MAAM,aAAa,CAAC,KAAK,KAAK,aAAa,EAAE,KAAK,KAAK,MAAM,UAAU,aAAa,CAAC,CACxF,KAAI,WAAW,UAAU,CACvB,QAAO,EAAE,MAAM,WAAW;AAI9B,QAAO;EAAE,MAAM,KAAK,KAAK,YAAY;EAAE,UAAU,CAAC,6BAAA;EAA+B;;AAGnF,SAAS,eAAuB;AAC9B,KAAI,OACF,QAAO;CAET,MAAM,SAAS,mBAAmB;CAClC,MAAM,IAAI,IAAI,OAAO,OAAO,MAAM,EAChC,UAAU,OAAO,UAClB,CAAC;AAEF,GAAE,GAAG,YAAY,QAAwB;EACvC,MAAM,OAAO,QAAQ,IAAI,IAAI,GAAG;AAChC,MAAI,CAAC,KACH;AAEF,MAAI,IAAI,SAAS,WACf,MAAK,aAAa;GAAE,OAAO,IAAI;GAAc,SAAS,IAAI;GAAS,OAAO,IAAI;GAAO,CAAC;WAE/E,IAAI,SAAS,QAAQ;AAC5B,WAAQ,OAAO,IAAI,GAAG;AACtB,QAAK,SAAS;aAEP,IAAI,SAAS,SAAS;AAC7B,WAAQ,OAAO,IAAI,GAAG;AACtB,QAAK,OAAO,IAAI,MAAM,IAAI,QAAQ,CAAC;;GAErC;AAEF,GAAE,GAAG,UAAU,QAAe;AAC5B,OAAK,MAAM,QAAQ,QAAQ,QAAQ,CACjC,MAAK,OAAO,IAAI;AAClB,UAAQ,OAAO;AACf,WAAS;GACT;AAEF,GAAE,GAAG,SAAS,SAAS;AACrB,MAAI,QAAQ,OAAO,GAAG;GACpB,MAAM,sBAAM,IAAI,MAAM,uBAAuB,KAAK,SAAS,QAAQ,KAAK,gBAAgB;AACxF,QAAK,MAAM,QAAQ,QAAQ,QAAQ,CACjC,MAAK,OAAO,IAAI;AAClB,WAAQ,OAAO;;AAEjB,WAAS;GACT;AAEF,UAAS;AACT,QAAO;;AAGT,SAAS,aAAa;AACpB,KAAI,WAAW,MAAM,WAAW,EAC9B;AACW,OAAM,OAAO,EACpB;;AAGR,eAAsB,oBACpB,WACA,QACe;AACf,QAAO,IAAI,SAAe,SAAS,WAAW;EAC5C,MAAM,YAAY;AAChB,aAAU;GACV,MAAM,KAAK,EAAE;GAEb,IAAI;AACJ,OAAI;AACF,QAAI,cAAc;YAEb,KAAK;AACV,cAAU;AACV,gBAAY;AACZ,WAAO,eAAe,QAAQ,MAAM,IAAI,MAAM,OAAO,IAAI,CAAC,CAAC;AAC3D;;AAGF,WAAQ,IAAI,IAAI;IACd;IACA,eAAe;AACb,eAAU;AACV,iBAAY;AACZ,cAAS;;IAEX,SAAS,QAAQ;AACf,eAAU;AACV,iBAAY;AACZ,YAAO,IAAI;;IAEb,YAAY,OAAO;IACpB,CAAC;GAEF,MAAM,MAAqB;IACzB,MAAM;IACN;IACA;IACA,QAAQ,OAAO;IAChB;AAED,KAAE,YAAY,IAAI;;AAGpB,MAAI,QACF,OAAM,KAAK,IAAI;MAGf,MAAK;GAEP;;AAGJ,eAAsB,iBAAgC;AACpD,KAAI,CAAC,OACH;CAEF,MAAM,IAAI;AACV,UAAS;AAET,QAAO,IAAI,SAAe,YAAY;EACpC,MAAM,UAAU,iBAAiB;AAC/B,KAAE,WAAW,CAAC,WAAW,SAAS,QAAQ,SAAS,CAAC;KACnD,IAAK;AAER,IAAE,KAAK,cAAc;AACnB,gBAAa,QAAQ;AACrB,YAAS;IACT;AAEF,IAAE,YAAY,EAAE,MAAM,YAAY,CAAyB;GAC3D"}