panopticon-cli 0.5.4 → 0.5.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (119) hide show
  1. package/dist/{agents-HNMF52RM.js → agents-QXVDAW2M.js} +12 -9
  2. package/dist/archive-planning-U3AZAKWI.js +16 -0
  3. package/dist/{chunk-KBHRXV5T.js → chunk-43F4LDZ4.js} +3 -3
  4. package/dist/{chunk-KY2E2Q3T.js → chunk-4XR62WWV.js} +105 -46
  5. package/dist/chunk-4XR62WWV.js.map +1 -0
  6. package/dist/chunk-6OYUJ4AJ.js +146 -0
  7. package/dist/chunk-6OYUJ4AJ.js.map +1 -0
  8. package/dist/{chunk-MOPGR3CL.js → chunk-AAP4G6U7.js} +1 -1
  9. package/dist/chunk-AAP4G6U7.js.map +1 -0
  10. package/dist/{chunk-4HST45MO.js → chunk-BYWVPPAZ.js} +19 -12
  11. package/dist/chunk-BYWVPPAZ.js.map +1 -0
  12. package/dist/{chunk-CFCUOV3Q.js → chunk-DMRTN432.js} +4 -1
  13. package/dist/chunk-DMRTN432.js.map +1 -0
  14. package/dist/{chunk-HOGYHJ2G.js → chunk-DW3PKGIS.js} +2 -2
  15. package/dist/chunk-GUV2EPBG.js +692 -0
  16. package/dist/chunk-GUV2EPBG.js.map +1 -0
  17. package/dist/{chunk-44EOY2ZL.js → chunk-HHL3AWXA.js} +46 -2
  18. package/dist/chunk-HHL3AWXA.js.map +1 -0
  19. package/dist/{chunk-6N2KBSJA.js → chunk-IZIXJYXZ.js} +40 -6
  20. package/dist/chunk-IZIXJYXZ.js.map +1 -0
  21. package/dist/chunk-MJXYTGK5.js +64 -0
  22. package/dist/chunk-MJXYTGK5.js.map +1 -0
  23. package/dist/chunk-OJF4QS3S.js +269 -0
  24. package/dist/chunk-OJF4QS3S.js.map +1 -0
  25. package/dist/{chunk-FQ66DECN.js → chunk-QAJAJBFW.js} +1 -1
  26. package/dist/chunk-QAJAJBFW.js.map +1 -0
  27. package/dist/chunk-R4KPLLRB.js +36 -0
  28. package/dist/chunk-R4KPLLRB.js.map +1 -0
  29. package/dist/{chunk-DFNVHK3N.js → chunk-SUM2WVPF.js} +4 -4
  30. package/dist/{chunk-ID4OYXVH.js → chunk-TFPJD2I2.js} +112 -45
  31. package/dist/chunk-TFPJD2I2.js.map +1 -0
  32. package/dist/{chunk-T7BBPDEJ.js → chunk-UKSGE6RH.js} +45 -15
  33. package/dist/chunk-UKSGE6RH.js.map +1 -0
  34. package/dist/chunk-W2OTF6OS.js +201 -0
  35. package/dist/chunk-W2OTF6OS.js.map +1 -0
  36. package/dist/chunk-WEQW3EAT.js +78 -0
  37. package/dist/chunk-WEQW3EAT.js.map +1 -0
  38. package/dist/chunk-YAAT66RT.js +70 -0
  39. package/dist/chunk-YAAT66RT.js.map +1 -0
  40. package/dist/{chunk-RLZQB7HS.js → chunk-ZMJFEHGF.js} +13 -1
  41. package/dist/chunk-ZMJFEHGF.js.map +1 -0
  42. package/dist/{chunk-HRU7S4TA.js → chunk-ZN5RHWGR.js} +18 -208
  43. package/dist/{chunk-HRU7S4TA.js.map → chunk-ZN5RHWGR.js.map} +1 -1
  44. package/dist/{chunk-ZTYHZMEC.js → chunk-ZWZNEA26.js} +2 -2
  45. package/dist/clean-planning-7Z5YY64X.js +9 -0
  46. package/dist/cli/index.js +1338 -2226
  47. package/dist/cli/index.js.map +1 -1
  48. package/dist/close-issue-CTZK777I.js +9 -0
  49. package/dist/compact-beads-72SHALOL.js +9 -0
  50. package/dist/{config-4CJNUE3O.js → config-FFTMBVHM.js} +2 -2
  51. package/dist/dashboard/public/assets/{index-DSvt5pPn.css → index-Bx4NCn9A.css} +1 -1
  52. package/dist/dashboard/public/assets/index-C7hJ5-o1.js +756 -0
  53. package/dist/dashboard/public/index.html +3 -2
  54. package/dist/dashboard/server.js +34720 -34297
  55. package/dist/{feedback-writer-T43PI5S2.js → feedback-writer-T2WCT6EZ.js} +2 -2
  56. package/dist/{hume-CKJJ3OUU.js → hume-GVTB5BKW.js} +3 -3
  57. package/dist/index.d.ts +24 -16
  58. package/dist/index.js +4 -4
  59. package/dist/label-cleanup-4HJVX6NP.js +103 -0
  60. package/dist/label-cleanup-4HJVX6NP.js.map +1 -0
  61. package/dist/merge-agent-O3TSBTLC.js +1725 -0
  62. package/dist/merge-agent-O3TSBTLC.js.map +1 -0
  63. package/dist/{projects-KVM3MN3Y.js → projects-3CRF57ZU.js} +2 -2
  64. package/dist/{rally-RKFSWC7E.js → rally-LBY24P4C.js} +2 -2
  65. package/dist/{remote-agents-ULPD6C5U.js → remote-agents-3NZPSHYG.js} +2 -3
  66. package/dist/{remote-workspace-XX6ARE6I.js → remote-workspace-M4IULGFZ.js} +24 -49
  67. package/dist/remote-workspace-M4IULGFZ.js.map +1 -0
  68. package/dist/{review-status-XKUKZF6J.js → review-status-J2YJGL3E.js} +2 -2
  69. package/dist/{specialist-context-C66TEMXS.js → specialist-context-IKG6VMNH.js} +7 -5
  70. package/dist/{specialist-context-C66TEMXS.js.map → specialist-context-IKG6VMNH.js.map} +1 -1
  71. package/dist/{specialist-logs-CJKXM3SR.js → specialist-logs-GFKUXCFG.js} +6 -4
  72. package/dist/{specialists-NXYD4Z62.js → specialists-XMFCFGYQ.js} +6 -4
  73. package/dist/specialists-XMFCFGYQ.js.map +1 -0
  74. package/dist/tmux-X2I5SAIJ.js +31 -0
  75. package/dist/tmux-X2I5SAIJ.js.map +1 -0
  76. package/dist/{traefik-5GL3Q7DJ.js → traefik-QXLZ4PO2.js} +4 -4
  77. package/dist/traefik-QXLZ4PO2.js.map +1 -0
  78. package/dist/{tunnel-BKC7KLBX.js → tunnel-7IOSRZVH.js} +3 -3
  79. package/dist/tunnel-7IOSRZVH.js.map +1 -0
  80. package/dist/{workspace-manager-ALBR62AS.js → workspace-manager-G6TTBPC3.js} +6 -6
  81. package/dist/workspace-manager-G6TTBPC3.js.map +1 -0
  82. package/package.json +2 -2
  83. package/scripts/build-cost-script.mjs +17 -0
  84. package/scripts/heartbeat-hook +28 -8
  85. package/scripts/record-cost-event.js +46 -7
  86. package/scripts/record-cost-event.ts +2 -1
  87. package/dist/chunk-44EOY2ZL.js.map +0 -1
  88. package/dist/chunk-4HST45MO.js.map +0 -1
  89. package/dist/chunk-565HZ6VV.js +0 -159
  90. package/dist/chunk-565HZ6VV.js.map +0 -1
  91. package/dist/chunk-6N2KBSJA.js.map +0 -1
  92. package/dist/chunk-CFCUOV3Q.js.map +0 -1
  93. package/dist/chunk-FQ66DECN.js.map +0 -1
  94. package/dist/chunk-ID4OYXVH.js.map +0 -1
  95. package/dist/chunk-KY2E2Q3T.js.map +0 -1
  96. package/dist/chunk-MOPGR3CL.js.map +0 -1
  97. package/dist/chunk-RLZQB7HS.js.map +0 -1
  98. package/dist/chunk-T7BBPDEJ.js.map +0 -1
  99. package/dist/chunk-ZDNQFWR5.js +0 -650
  100. package/dist/chunk-ZDNQFWR5.js.map +0 -1
  101. package/dist/dashboard/public/assets/index-DA6pnizT.js +0 -767
  102. package/dist/remote-workspace-XX6ARE6I.js.map +0 -1
  103. /package/dist/{agents-HNMF52RM.js.map → agents-QXVDAW2M.js.map} +0 -0
  104. /package/dist/{config-4CJNUE3O.js.map → archive-planning-U3AZAKWI.js.map} +0 -0
  105. /package/dist/{chunk-KBHRXV5T.js.map → chunk-43F4LDZ4.js.map} +0 -0
  106. /package/dist/{chunk-HOGYHJ2G.js.map → chunk-DW3PKGIS.js.map} +0 -0
  107. /package/dist/{chunk-DFNVHK3N.js.map → chunk-SUM2WVPF.js.map} +0 -0
  108. /package/dist/{chunk-ZTYHZMEC.js.map → chunk-ZWZNEA26.js.map} +0 -0
  109. /package/dist/{hume-CKJJ3OUU.js.map → clean-planning-7Z5YY64X.js.map} +0 -0
  110. /package/dist/{projects-KVM3MN3Y.js.map → close-issue-CTZK777I.js.map} +0 -0
  111. /package/dist/{rally-RKFSWC7E.js.map → compact-beads-72SHALOL.js.map} +0 -0
  112. /package/dist/{remote-agents-ULPD6C5U.js.map → config-FFTMBVHM.js.map} +0 -0
  113. /package/dist/{feedback-writer-T43PI5S2.js.map → feedback-writer-T2WCT6EZ.js.map} +0 -0
  114. /package/dist/{review-status-XKUKZF6J.js.map → hume-GVTB5BKW.js.map} +0 -0
  115. /package/dist/{specialist-logs-CJKXM3SR.js.map → projects-3CRF57ZU.js.map} +0 -0
  116. /package/dist/{specialists-NXYD4Z62.js.map → rally-LBY24P4C.js.map} +0 -0
  117. /package/dist/{traefik-5GL3Q7DJ.js.map → remote-agents-3NZPSHYG.js.map} +0 -0
  118. /package/dist/{tunnel-BKC7KLBX.js.map → review-status-J2YJGL3E.js.map} +0 -0
  119. /package/dist/{workspace-manager-ALBR62AS.js.map → specialist-logs-GFKUXCFG.js.map} +0 -0
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/lib/skills-merge.ts","../src/lib/workspace-manager.ts"],"sourcesContent":["import {\n existsSync,\n readdirSync,\n mkdirSync,\n readFileSync,\n writeFileSync,\n copyFileSync,\n statSync,\n} from 'fs';\nimport { join, relative, dirname } from 'path';\nimport { SKILLS_DIR, CACHE_AGENTS_DIR, CACHE_RULES_DIR } from './paths.js';\nimport {\n readManifest,\n writeManifest,\n collectSourceFiles,\n hashFile,\n setManifestEntry,\n compareFileToManifest,\n type Manifest,\n} from './manifest.js';\n\nexport interface MergeResult {\n added: string[];\n updated: string[];\n skipped: string[];\n overlayed: string[];\n}\n\n/**\n * Copy all files from a source directory into a target directory,\n * preserving subdirectory structure. Returns the list of relative paths copied.\n */\nfunction copyTree(sourceDir: string, targetDir: string): string[] {\n const copied: string[] = [];\n if (!existsSync(sourceDir)) return copied;\n\n function walk(dir: string): void {\n const entries = readdirSync(dir, { withFileTypes: true });\n for (const entry of entries) {\n const fullPath = join(dir, entry.name);\n if (entry.isDirectory()) {\n walk(fullPath);\n } else if (entry.isFile()) {\n const rel = relative(sourceDir, fullPath);\n const targetPath = join(targetDir, rel);\n mkdirSync(dirname(targetPath), { recursive: true });\n copyFileSync(fullPath, targetPath);\n copied.push(rel);\n }\n }\n }\n\n walk(sourceDir);\n return copied;\n}\n\n/**\n * Merge Panopticon skills, agents, and rules into a workspace using file copies.\n *\n * Flow:\n * 1. Copy from cache (skills, agent-definitions, rules) → workspace/.claude/\n * 2. Write manifest tracking what was placed\n *\n * Project template overlay is handled separately by workspace-manager.ts\n * (processTemplates + createSymlinks → now also copy-based).\n */\nexport function mergeSkillsIntoWorkspace(workspacePath: string): MergeResult {\n const claudeDir = join(workspacePath, '.claude');\n const manifestPath = join(claudeDir, '.panopticon-manifest.json');\n const manifest = readManifest(manifestPath);\n\n const result: MergeResult = {\n added: [],\n updated: [],\n skipped: [],\n overlayed: [],\n };\n\n // Ensure base directories exist\n mkdirSync(join(claudeDir, 'skills'), { recursive: true });\n mkdirSync(join(claudeDir, 'agents'), { recursive: true });\n\n // Sources to copy: category → source cache directory\n const sources: Array<{ category: string; sourceDir: string; targetSubdir: string }> = [\n { category: 'skills', sourceDir: SKILLS_DIR, targetSubdir: 'skills' },\n { category: 'agents', sourceDir: CACHE_AGENTS_DIR, targetSubdir: 'agents' },\n { category: 'rules', sourceDir: CACHE_RULES_DIR, targetSubdir: 'rules' },\n ];\n\n for (const { category, sourceDir, targetSubdir } of sources) {\n if (!existsSync(sourceDir)) continue;\n\n const prefix = targetSubdir ? `${targetSubdir}/` : '';\n const files = collectSourceFiles(sourceDir, '');\n\n for (const file of files) {\n const relativePath = `${prefix}${file.relativePath}`;\n const targetPath = join(claudeDir, relativePath);\n const sourceHash = hashFile(file.absolutePath);\n\n // Check status against manifest\n const status = compareFileToManifest(targetPath, relativePath, manifest);\n\n switch (status.action) {\n case 'new':\n // File doesn't exist at target — copy it\n mkdirSync(dirname(targetPath), { recursive: true });\n copyFileSync(file.absolutePath, targetPath);\n setManifestEntry(manifest, relativePath, sourceHash, 'panopticon');\n result.added.push(relativePath);\n break;\n\n case 'update':\n // File exists and matches manifest — safe to overwrite with latest\n copyFileSync(file.absolutePath, targetPath);\n setManifestEntry(manifest, relativePath, sourceHash, 'panopticon');\n result.updated.push(relativePath);\n break;\n\n case 'modified':\n // User modified the file — skip to preserve their changes\n result.skipped.push(`${relativePath} (modified by user)`);\n break;\n\n case 'user-owned':\n // File exists but wasn't placed by us — never touch\n result.skipped.push(`${relativePath} (user-owned)`);\n break;\n }\n }\n }\n\n // Write updated manifest\n writeManifest(manifestPath, manifest);\n\n return result;\n}\n\n/**\n * Apply project template overlay on top of Panopticon base files in a workspace.\n *\n * This copies files from the project's agent template directory into\n * workspace/.claude/, overwriting Panopticon files where the project\n * provides its own version. Updates the manifest with source=\"project-template\".\n *\n * @param workspacePath - Path to the workspace\n * @param templateDir - Absolute path to the project's agent template directory\n * @param templates - Optional list of specific template files to process (source → target mappings)\n */\nexport function applyProjectTemplateOverlay(\n workspacePath: string,\n templateDir: string,\n templates?: Array<{ source: string; target: string }>,\n): string[] {\n const claudeDir = join(workspacePath, '.claude');\n const manifestPath = join(claudeDir, '.panopticon-manifest.json');\n const manifest = readManifest(manifestPath);\n const overlayed: string[] = [];\n\n if (!existsSync(templateDir)) return overlayed;\n\n if (templates && templates.length > 0) {\n // Process specific template mappings\n for (const { source, target } of templates) {\n const sourcePath = join(templateDir, source);\n if (!existsSync(sourcePath)) continue;\n\n const targetPath = join(workspacePath, target);\n mkdirSync(dirname(targetPath), { recursive: true });\n\n // Read template content and check if it's a template file\n if (source.endsWith('.template')) {\n // Template files are handled by workspace-manager's processTemplates\n // We just track them in the manifest after they're processed\n continue;\n }\n\n copyFileSync(sourcePath, targetPath);\n\n // Track in manifest if it's under .claude/\n if (target.startsWith('.claude/')) {\n const relativePath = target.slice('.claude/'.length);\n const hash = hashFile(targetPath);\n setManifestEntry(manifest, relativePath, hash, 'project-template');\n overlayed.push(relativePath);\n }\n }\n } else {\n // Copy all .claude/ subdirectories from template dir\n const claudeInTemplate = join(templateDir, '.claude');\n if (existsSync(claudeInTemplate)) {\n const copied = copyTree(claudeInTemplate, claudeDir);\n for (const rel of copied) {\n const targetPath = join(claudeDir, rel);\n const hash = hashFile(targetPath);\n setManifestEntry(manifest, rel, hash, 'project-template');\n overlayed.push(rel);\n }\n }\n }\n\n // Write updated manifest\n writeManifest(manifestPath, manifest);\n\n return overlayed;\n}\n\n// ─── Legacy exports (kept for migration, to be removed in future) ───\n\n/**\n * @deprecated No longer needed — skills are copies, not symlinks. Kept for migration.\n */\nexport function cleanupGitignore(gitignorePath: string): {\n cleaned: boolean;\n duplicatesRemoved: number;\n entriesAfter: number;\n} {\n if (!existsSync(gitignorePath)) {\n return { cleaned: false, duplicatesRemoved: 0, entriesAfter: 0 };\n }\n\n const PANOPTICON_HEADER = '# Panopticon-managed symlinks (not committed)';\n let content: string;\n try {\n content = readFileSync(gitignorePath, 'utf-8');\n } catch {\n return { cleaned: false, duplicatesRemoved: 0, entriesAfter: 0 };\n }\n\n // If no Panopticon section, nothing to clean\n if (!content.includes(PANOPTICON_HEADER)) {\n return { cleaned: false, duplicatesRemoved: 0, entriesAfter: 0 };\n }\n\n // Remove the entire Panopticon section (skills are copies now, not symlinks)\n const lines = content.split('\\n');\n const newLines: string[] = [];\n let inPanopticonSection = false;\n\n for (const line of lines) {\n const trimmed = line.trim();\n if (trimmed === PANOPTICON_HEADER) {\n inPanopticonSection = true;\n continue;\n }\n if (inPanopticonSection) {\n if (trimmed.startsWith('#') && trimmed !== '') {\n inPanopticonSection = false;\n newLines.push(line);\n } else if (trimmed === '') {\n // Skip blank lines in Panopticon section\n continue;\n }\n // Skip entries in Panopticon section\n continue;\n }\n newLines.push(line);\n }\n\n // Write cleaned file\n try {\n writeFileSync(gitignorePath, newLines.join('\\n'), 'utf-8');\n return { cleaned: true, duplicatesRemoved: 0, entriesAfter: 0 };\n } catch {\n return { cleaned: false, duplicatesRemoved: 0, entriesAfter: 0 };\n }\n}\n\n/**\n * @deprecated No longer needed — skills are copies, not symlinks. Kept for migration.\n */\nexport function cleanupWorkspaceGitignore(workspacePath: string): {\n cleaned: boolean;\n duplicatesRemoved: number;\n entriesAfter: number;\n} {\n const gitignorePath = join(workspacePath, '.claude', 'skills', '.gitignore');\n return cleanupGitignore(gitignorePath);\n}\n","/**\n * Workspace Manager\n *\n * Handles workspace creation and removal for both monorepo and polyrepo projects.\n */\n\nimport { existsSync, mkdirSync, writeFileSync, readFileSync, readdirSync, copyFileSync, symlinkSync, chmodSync, realpathSync } from 'fs';\nimport { join, dirname, basename, extname } from 'path';\nimport { homedir } from 'os';\nimport { exec } from 'child_process';\nimport { promisify } from 'util';\nimport {\n ProjectConfig,\n WorkspaceConfig,\n TemplatePlaceholders,\n replacePlaceholders,\n getDefaultWorkspaceConfig,\n} from './workspace-config.js';\nimport { addDnsEntry, removeDnsEntry, syncDnsToWindows } from './dns.js';\nimport { addTunnelIngress, removeTunnelIngress } from './tunnel.js';\nimport { createHumeConfig, deleteHumeConfig } from './hume.js';\nimport { mergeSkillsIntoWorkspace } from './skills-merge.js';\n\nconst execAsync = promisify(exec);\n\nexport interface WorkspaceCreateOptions {\n projectConfig: ProjectConfig;\n featureName: string;\n startDocker?: boolean;\n dryRun?: boolean;\n}\n\nexport interface WorkspaceCreateResult {\n success: boolean;\n workspacePath: string;\n errors: string[];\n steps: string[];\n}\n\n/**\n * Create placeholders for template substitution\n */\nfunction createPlaceholders(\n projectConfig: ProjectConfig,\n featureName: string,\n workspacePath: string\n): TemplatePlaceholders {\n const featureFolder = `feature-${featureName}`;\n const domain = projectConfig.workspace?.dns?.domain || 'localhost';\n\n return {\n FEATURE_NAME: featureName,\n FEATURE_FOLDER: featureFolder,\n BRANCH_NAME: `feature/${featureName}`,\n COMPOSE_PROJECT: `${basename(projectConfig.path)}-${featureFolder}`,\n DOMAIN: domain,\n PROJECT_NAME: basename(projectConfig.path),\n PROJECT_PATH: projectConfig.path,\n PROJECTS_DIR: dirname(projectConfig.path),\n WORKSPACE_PATH: workspacePath,\n HOME: homedir(),\n };\n}\n\n/**\n * Sanitize docker-compose files to use platform-agnostic paths\n * Replaces hardcoded /home/username paths with ${HOME}\n */\nfunction sanitizeComposeFile(filePath: string): void {\n if (!existsSync(filePath)) return;\n\n let content = readFileSync(filePath, 'utf-8');\n const originalContent = content;\n\n // Pattern to match hardcoded home paths like /home/username or /Users/username\n // Replace with ${HOME} which docker-compose expands\n const homePatterns = [\n /\\/home\\/[a-zA-Z0-9_-]+\\//g, // Linux: /home/username/\n /\\/Users\\/[a-zA-Z0-9_-]+\\//g, // macOS: /Users/username/\n ];\n\n for (const pattern of homePatterns) {\n content = content.replace(pattern, '${HOME}/');\n }\n\n if (content !== originalContent) {\n writeFileSync(filePath, content, 'utf-8');\n }\n}\n\n/**\n * Validate feature name (alphanumeric and hyphens only)\n */\nfunction validateFeatureName(name: string): boolean {\n return /^[a-zA-Z0-9-]+$/.test(name);\n}\n\n/**\n * Create a git worktree\n * @param repoPath Path to the source git repository\n * @param targetPath Where to create the worktree\n * @param branchName Name of the feature branch to create/checkout\n * @param defaultBranch Base branch to create new branches from (default: 'main')\n */\nasync function createWorktree(\n repoPath: string,\n targetPath: string,\n branchName: string,\n defaultBranch: string = 'main'\n): Promise<{ success: boolean; message: string }> {\n try {\n // Fetch latest from origin\n await execAsync('git fetch origin', { cwd: repoPath });\n\n // Prune stale worktree entries (e.g., from deleted workspaces)\n await execAsync('git worktree prune', { cwd: repoPath });\n\n // Check if branch exists locally or remotely\n const { stdout: localBranches } = await execAsync('git branch --list', { cwd: repoPath });\n const { stdout: remoteBranches } = await execAsync('git branch -r --list', { cwd: repoPath });\n\n const branchExists =\n localBranches.includes(branchName) ||\n remoteBranches.includes(`origin/${branchName}`);\n\n if (branchExists) {\n await execAsync(`git worktree add \"${targetPath}\" \"${branchName}\"`, { cwd: repoPath });\n } else {\n // Create new branch from the configured default branch\n await execAsync(`git worktree add \"${targetPath}\" -b \"${branchName}\" \"${defaultBranch}\"`, { cwd: repoPath });\n }\n\n // Configure beads role so agents don't get \"beads.role not configured\" warnings\n await execAsync('git config beads.role agent', { cwd: targetPath }).catch(() => {});\n\n return { success: true, message: `Created worktree at ${targetPath}` };\n } catch (error) {\n return { success: false, message: `Failed to create worktree: ${error}` };\n }\n}\n\n/**\n * Remove a git worktree\n */\nasync function removeWorktree(\n repoPath: string,\n targetPath: string,\n branchName: string\n): Promise<{ success: boolean; message: string }> {\n try {\n // Remove worktree\n await execAsync(`git worktree remove \"${targetPath}\" --force`, { cwd: repoPath }).catch(() => {});\n\n // Optionally delete the branch\n await execAsync(`git branch -D \"${branchName}\"`, { cwd: repoPath }).catch(() => {});\n\n return { success: true, message: `Removed worktree at ${targetPath}` };\n } catch (error) {\n return { success: false, message: `Failed to remove worktree: ${error}` };\n }\n}\n\n// DNS functions (addWsl2HostEntry, removeWsl2HostEntry, syncDnsToWindows)\n// are now in src/lib/dns.ts and imported above\n\n/**\n * Assign a port from a range\n */\nfunction assignPort(\n portFile: string,\n featureFolder: string,\n range: [number, number]\n): number {\n // Ensure port file exists\n if (!existsSync(portFile)) {\n mkdirSync(dirname(portFile), { recursive: true });\n writeFileSync(portFile, '');\n }\n\n const content = readFileSync(portFile, 'utf-8');\n const lines = content.split('\\n').filter(Boolean);\n\n // Check if already assigned\n for (const line of lines) {\n const [folder, port] = line.split(':');\n if (folder === featureFolder) {\n return parseInt(port, 10);\n }\n }\n\n // Find next available port\n const usedPorts = new Set(lines.map(l => parseInt(l.split(':')[1], 10)));\n for (let port = range[0]; port <= range[1]; port++) {\n if (!usedPorts.has(port)) {\n writeFileSync(portFile, content + (content.endsWith('\\n') ? '' : '\\n') + `${featureFolder}:${port}\\n`);\n return port;\n }\n }\n\n throw new Error(`No available ports in range ${range[0]}-${range[1]}`);\n}\n\n/**\n * Release a port assignment\n */\nfunction releasePort(portFile: string, featureFolder: string): boolean {\n try {\n if (!existsSync(portFile)) return true;\n\n let content = readFileSync(portFile, 'utf-8');\n const lines = content.split('\\n').filter(line => !line.startsWith(`${featureFolder}:`));\n writeFileSync(portFile, lines.join('\\n'));\n return true;\n } catch {\n return false;\n }\n}\n\n/**\n * Process template files with placeholder replacement\n */\nfunction processTemplates(\n templateDir: string,\n targetDir: string,\n placeholders: TemplatePlaceholders,\n templates?: Array<{ source: string; target: string }>\n): string[] {\n const steps: string[] = [];\n\n if (!existsSync(templateDir)) {\n return steps;\n }\n\n // If specific templates are defined, process those\n if (templates && templates.length > 0) {\n for (const { source, target } of templates) {\n const sourcePath = join(templateDir, source);\n const targetPath = join(targetDir, target);\n\n if (existsSync(sourcePath)) {\n const content = readFileSync(sourcePath, 'utf-8');\n const processed = replacePlaceholders(content, placeholders);\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, processed);\n steps.push(`Processed template: ${source} -> ${target}`);\n }\n }\n } else {\n // Process all .template files\n const files = readdirSync(templateDir);\n for (const file of files) {\n if (file.endsWith('.template')) {\n const sourcePath = join(templateDir, file);\n const targetPath = join(targetDir, file.replace('.template', ''));\n\n const content = readFileSync(sourcePath, 'utf-8');\n const processed = replacePlaceholders(content, placeholders);\n writeFileSync(targetPath, processed);\n steps.push(`Processed template: ${file}`);\n }\n }\n }\n\n return steps;\n}\n\n/**\n * @deprecated Use copyProjectTemplateDirs instead. Kept for non-.claude paths.\n */\nfunction createSymlinks(\n sourceDir: string,\n targetDir: string,\n symlinks: string[]\n): string[] {\n const steps: string[] = [];\n\n for (const symlink of symlinks) {\n const sourcePath = join(sourceDir, symlink);\n const targetPath = join(targetDir, symlink);\n\n if (existsSync(sourcePath)) {\n mkdirSync(dirname(targetPath), { recursive: true });\n try {\n symlinkSync(sourcePath, targetPath);\n steps.push(`Created symlink: ${symlink}`);\n } catch {\n // Symlink might already exist\n }\n }\n }\n\n return steps;\n}\n\n/**\n * Copy project template directories into workspace (replaces symlinks).\n * Recursively copies all files from each source directory.\n */\nconst TEXT_EXTENSIONS = new Set([\n '.md', '.sh', '.yml', '.yaml', '.json', '.ts', '.js', '.env', '.txt', '.toml', '.template',\n]);\n\nfunction copyProjectTemplateDirs(\n sourceDir: string,\n targetDir: string,\n dirs: string[],\n placeholders?: TemplatePlaceholders\n): string[] {\n const steps: string[] = [];\n\n for (const dir of dirs) {\n const sourcePath = join(sourceDir, dir);\n const targetPath = join(targetDir, dir);\n\n if (!existsSync(sourcePath)) continue;\n\n // Recursively copy all files, applying placeholder substitution to text files\n function copyDir(src: string, dest: string): number {\n let count = 0;\n mkdirSync(dest, { recursive: true });\n const entries = readdirSync(src, { withFileTypes: true });\n for (const entry of entries) {\n const srcEntry = join(src, entry.name);\n const destEntry = join(dest, entry.name);\n if (entry.isDirectory()) {\n count += copyDir(srcEntry, destEntry);\n } else if (entry.isFile()) {\n const ext = extname(entry.name).toLowerCase();\n if (placeholders && TEXT_EXTENSIONS.has(ext)) {\n const content = readFileSync(srcEntry, 'utf-8');\n writeFileSync(destEntry, replacePlaceholders(content, placeholders));\n } else {\n copyFileSync(srcEntry, destEntry);\n }\n count++;\n }\n }\n return count;\n }\n\n const count = copyDir(sourcePath, targetPath);\n steps.push(`Copied ${count} files from project template: ${dir}`);\n }\n\n return steps;\n}\n\n/**\n * Create a workspace\n */\nexport async function createWorkspace(options: WorkspaceCreateOptions): Promise<WorkspaceCreateResult> {\n const { projectConfig, featureName, startDocker, dryRun } = options;\n const result: WorkspaceCreateResult = {\n success: true,\n workspacePath: '',\n errors: [],\n steps: [],\n };\n\n // Validate feature name\n if (!validateFeatureName(featureName)) {\n result.success = false;\n result.errors.push('Invalid feature name. Use alphanumeric and hyphens only.');\n return result;\n }\n\n // Reject 'main' as feature name\n if (featureName === 'main') {\n result.success = false;\n result.errors.push('Cannot create workspace for \"main\". Use base repos directly.');\n return result;\n }\n\n const workspaceConfig = projectConfig.workspace || getDefaultWorkspaceConfig();\n const workspacesDir = join(projectConfig.path, workspaceConfig.workspaces_dir || 'workspaces');\n const featureFolder = `feature-${featureName}`;\n const workspacePath = join(workspacesDir, featureFolder);\n result.workspacePath = workspacePath;\n\n // Check if workspace already exists\n if (existsSync(workspacePath)) {\n result.success = false;\n result.errors.push(`Workspace already exists at ${workspacePath}`);\n return result;\n }\n\n if (dryRun) {\n result.steps.push('[DRY RUN] Would create workspace at: ' + workspacePath);\n return result;\n }\n\n // Create placeholders\n const placeholders = createPlaceholders(projectConfig, featureName, workspacePath);\n\n // Create workspace directory\n mkdirSync(workspacePath, { recursive: true });\n result.steps.push('Created workspace directory');\n\n // Handle polyrepo vs monorepo\n if (workspaceConfig.type === 'polyrepo' && workspaceConfig.repos) {\n // Create worktrees for each repo\n for (const repo of workspaceConfig.repos) {\n // Resolve symlinks to get the actual git repository path\n // (e.g., myn/frontend -> ../frontend needs to resolve to actual path)\n const rawRepoPath = join(projectConfig.path, repo.path);\n const repoPath = existsSync(rawRepoPath) ? realpathSync(rawRepoPath) : rawRepoPath;\n const targetPath = join(workspacePath, repo.name);\n const branchPrefix = repo.branch_prefix || 'feature/';\n const branchName = `${branchPrefix}${featureName}`;\n // Per-repo default_branch overrides workspace-level, falls back to 'main'\n const defaultBranch = repo.default_branch || workspaceConfig.default_branch || 'main';\n\n const worktreeResult = await createWorktree(repoPath, targetPath, branchName, defaultBranch);\n if (worktreeResult.success) {\n result.steps.push(`Created worktree for ${repo.name}: ${branchName} (from ${defaultBranch})`);\n } else {\n result.errors.push(`${repo.name}: ${worktreeResult.message}`);\n result.success = false; // Fail the entire workspace creation if any worktree fails\n }\n }\n } else {\n // Monorepo: create single worktree\n const branchName = `feature/${featureName}`;\n const defaultBranch = workspaceConfig.default_branch || 'main';\n const worktreeResult = await createWorktree(projectConfig.path, workspacePath, branchName, defaultBranch);\n if (worktreeResult.success) {\n result.steps.push(`Created worktree: ${branchName} (from ${defaultBranch})`);\n } else {\n result.errors.push(worktreeResult.message);\n result.success = false; // Fail the entire workspace creation if worktree fails\n }\n }\n\n // Sanitize any docker-compose files in the workspace to use platform-agnostic paths\n // This handles files inherited from worktrees that may have hardcoded home paths\n const devcontainerDir = join(workspacePath, '.devcontainer');\n if (existsSync(devcontainerDir)) {\n const composeFiles = readdirSync(devcontainerDir)\n .filter(f => f.includes('compose') && (f.endsWith('.yml') || f.endsWith('.yaml')));\n for (const composeFile of composeFiles) {\n sanitizeComposeFile(join(devcontainerDir, composeFile));\n }\n if (composeFiles.length > 0) {\n result.steps.push(`Sanitized ${composeFiles.length} compose file(s) for platform compatibility`);\n }\n }\n\n // Setup TLDR code analysis for workspace (after worktree creation to ensure directory is ready)\n try {\n // Check if python3 is available\n await execAsync('python3 --version');\n const venvPath = join(workspacePath, '.venv');\n const tldrBin = join(venvPath, 'bin', 'tldr');\n\n // Check if main branch already has a working venv with llm-tldr\n const mainVenvTldr = join(projectConfig.path, '.venv', 'bin', 'tldr');\n const mainVenvExists = existsSync(mainVenvTldr);\n\n if (mainVenvExists) {\n // Copy the entire venv from main — faster than pip install (seconds vs 30s+)\n await execAsync(`cp -a \"${join(projectConfig.path, '.venv')}\" \"${venvPath}\"`);\n result.steps.push('Copied Python venv from main branch');\n } else {\n // Create fresh venv and install llm-tldr\n await execAsync(`python3 -m venv \"${venvPath}\"`, { cwd: workspacePath });\n const pipPath = join(venvPath, 'bin', 'pip');\n await execAsync(`\"${pipPath}\" install llm-tldr`, { cwd: workspacePath, timeout: 120000 });\n result.steps.push('Created Python venv and installed llm-tldr');\n\n // Apply .tsx/.jsx support patch (upstream llm-tldr only checks .ts)\n const patchScript = join(projectConfig.path, 'scripts', 'patches', 'llm-tldr-tsx-support.py');\n if (existsSync(patchScript)) {\n await execAsync(`python3 \"${patchScript}\" \"${venvPath}\"`);\n result.steps.push('Applied llm-tldr .tsx/.jsx patch');\n }\n }\n\n // Verify tldr binary exists after setup\n if (!existsSync(tldrBin)) {\n result.steps.push('TLDR setup incomplete: tldr binary not found after venv creation');\n } else {\n // Copy .tldr index from main branch if it exists\n const mainTldrDir = join(projectConfig.path, '.tldr');\n const workspaceTldrDir = join(workspacePath, '.tldr');\n\n if (existsSync(mainTldrDir)) {\n await execAsync(`cp -r \"${mainTldrDir}\" \"${workspaceTldrDir}\"`);\n result.steps.push('Copied TLDR index from main branch');\n }\n\n // Start TLDR daemon for this workspace\n const { getTldrDaemonService } = await import('./tldr-daemon.js');\n const tldrService = getTldrDaemonService(workspacePath, venvPath);\n await tldrService.start(true);\n result.steps.push('Started TLDR daemon');\n\n // Warm the index in the background — ensures workspaces always have a working index\n // even when the main branch cache was empty (nothing to copy)\n try {\n await tldrService.warm(true); // background=true: non-blocking\n result.steps.push('TLDR index warm initiated (background)');\n } catch {\n // Non-fatal — daemon may not support warm yet\n }\n }\n } catch (error: any) {\n // TLDR setup is optional — don't fail workspace creation, but log clearly\n if (error.message?.includes('python3')) {\n result.steps.push('Skipped TLDR setup (python3 not available)');\n } else {\n console.warn(`⚠ TLDR setup failed: ${error.message}`);\n result.steps.push(`TLDR setup failed: ${error.message}`);\n }\n }\n\n // Configure DNS\n if (workspaceConfig.dns) {\n const dnsMethod = workspaceConfig.dns.sync_method || 'wsl2hosts';\n for (const entryPattern of workspaceConfig.dns.entries) {\n const hostname = replacePlaceholders(entryPattern, placeholders);\n\n if (addDnsEntry(dnsMethod, hostname)) {\n result.steps.push(`Added DNS entry: ${hostname} (${dnsMethod})`);\n }\n }\n\n // Sync to Windows if using wsl2hosts method\n if (dnsMethod === 'wsl2hosts') {\n const synced = await syncDnsToWindows();\n if (synced) {\n result.steps.push('Synced DNS to Windows hosts file');\n }\n }\n }\n\n // Assign ports\n if (workspaceConfig.ports) {\n for (const [portName, portConfig] of Object.entries(workspaceConfig.ports)) {\n const portFile = join(projectConfig.path, `.${portName}-ports`);\n try {\n const port = assignPort(portFile, featureFolder, portConfig.range);\n result.steps.push(`Assigned ${portName} port: ${port}`);\n // Add to placeholders for use in templates\n (placeholders as any)[`${portName.toUpperCase()}_PORT`] = String(port);\n } catch (error) {\n result.errors.push(`Failed to assign ${portName} port: ${error}`);\n }\n }\n }\n\n // Install base Panopticon skills/agents/rules from cache\n const mergeResult = mergeSkillsIntoWorkspace(workspacePath);\n const mergeTotal = mergeResult.added.length + mergeResult.updated.length;\n if (mergeTotal > 0) {\n result.steps.push(`Installed ${mergeTotal} Panopticon files (${mergeResult.added.length} new, ${mergeResult.updated.length} updated)`);\n }\n\n // Process agent templates (project template overlay — wins over Panopticon base)\n if (workspaceConfig.agent?.template_dir) {\n const templateDir = join(projectConfig.path, workspaceConfig.agent.template_dir);\n\n // Process template files\n const templateSteps = processTemplates(\n templateDir,\n workspacePath,\n placeholders,\n workspaceConfig.agent.templates\n );\n result.steps.push(...templateSteps);\n\n // Copy .claude/ directories from project template (copy_dirs replaces legacy symlinks)\n const dirsToSync = workspaceConfig.agent.copy_dirs || workspaceConfig.agent.symlinks;\n if (dirsToSync) {\n const copySteps = copyProjectTemplateDirs(templateDir, workspacePath, dirsToSync, placeholders);\n result.steps.push(...copySteps);\n }\n }\n\n // Generate .env file\n if (workspaceConfig.env?.template) {\n const envContent = replacePlaceholders(workspaceConfig.env.template, placeholders);\n writeFileSync(join(workspacePath, '.env'), envContent);\n result.steps.push('Created .env file');\n }\n\n // Process Docker compose templates\n if (workspaceConfig.docker?.compose_template) {\n const templateDir = join(projectConfig.path, workspaceConfig.docker.compose_template);\n const devcontainerDir = join(workspacePath, '.devcontainer');\n mkdirSync(devcontainerDir, { recursive: true });\n\n const templateSteps = processTemplates(templateDir, devcontainerDir, placeholders);\n result.steps.push(...templateSteps);\n\n // Copy non-template files (like Dockerfile)\n if (existsSync(templateDir)) {\n const files = readdirSync(templateDir);\n for (const file of files) {\n if (!file.endsWith('.template')) {\n const sourcePath = join(templateDir, file);\n const targetPath = join(devcontainerDir, file);\n copyFileSync(sourcePath, targetPath);\n }\n }\n }\n\n // Sanitize docker-compose files to use platform-agnostic paths\n // This fixes hardcoded /home/username or /Users/username paths\n const composeFiles = readdirSync(devcontainerDir)\n .filter(f => f.includes('compose') && (f.endsWith('.yml') || f.endsWith('.yaml')));\n for (const composeFile of composeFiles) {\n sanitizeComposeFile(join(devcontainerDir, composeFile));\n }\n if (composeFiles.length > 0) {\n result.steps.push(`Sanitized ${composeFiles.length} compose file(s) for platform compatibility`);\n }\n\n // Create ./dev symlink at workspace root pointing to .devcontainer/dev\n // Symlink keeps changes in sync - editing ./dev updates .devcontainer/dev\n const devScriptInContainer = join(devcontainerDir, 'dev');\n const devScriptAtRoot = join(workspacePath, 'dev');\n if (existsSync(devScriptInContainer) && !existsSync(devScriptAtRoot)) {\n try {\n symlinkSync('.devcontainer/dev', devScriptAtRoot);\n chmodSync(devScriptInContainer, 0o755); // Make executable\n result.steps.push('Created ./dev symlink');\n } catch (error) {\n result.errors.push(`Failed to create ./dev symlink: ${error}`);\n }\n }\n }\n\n // Note: Beads initialization is handled by the calling command (workspace.ts)\n // With beads v0.47.1+, worktrees use shared database with labels for isolation\n // The workspace.ts command creates a bead with workspace:issue-id label\n\n // Set up Cloudflare tunnel for external access (before Docker so containers can use tunnel URLs)\n if (workspaceConfig.tunnel) {\n const tunnelResult = await addTunnelIngress(workspaceConfig.tunnel, placeholders);\n result.steps.push(...tunnelResult.steps);\n if (!tunnelResult.success) {\n result.errors.push('Tunnel setup had failures (see steps for details)');\n }\n }\n\n // Create Hume EVI config and write env file for Docker (before Docker so containers pick up the config ID)\n if (workspaceConfig.hume) {\n const humeResult = await createHumeConfig(workspaceConfig.hume, placeholders);\n result.steps.push(...humeResult.steps);\n if (humeResult.configId) {\n writeFileSync(\n join(workspacePath, '.hume-config'),\n `HUME_CONFIG_ID=${humeResult.configId}\\nVITE_HUME_CONFIG_ID=${humeResult.configId}\\n`,\n );\n result.steps.push('Wrote .hume-config with Hume EVI config ID');\n }\n if (!humeResult.success) {\n result.errors.push('Hume EVI config setup had failures (see steps for details)');\n }\n }\n\n // Start Docker containers if requested\n if (startDocker) {\n // Check for Traefik\n if (workspaceConfig.docker?.traefik) {\n const traefikPath = join(projectConfig.path, workspaceConfig.docker.traefik);\n if (existsSync(traefikPath)) {\n try {\n await execAsync(`docker compose -f \"${traefikPath}\" up -d`, { cwd: projectConfig.path });\n result.steps.push('Started Traefik');\n } catch (error: any) {\n const msg = error?.message || String(error);\n if (msg.includes('port is already allocated') || msg.includes('address already in use')) {\n // Traefik (or another reverse proxy) is already running — not an error\n result.steps.push('Traefik already running (port in use)');\n } else {\n result.errors.push(`Failed to start Traefik: ${error}`);\n }\n }\n }\n }\n\n // Start workspace containers\n const composeLocations = [\n join(workspacePath, 'docker-compose.yml'),\n join(workspacePath, 'docker-compose.yaml'),\n join(workspacePath, '.devcontainer', 'docker-compose.yml'),\n join(workspacePath, '.devcontainer', 'docker-compose.devcontainer.yml'),\n ];\n\n for (const composePath of composeLocations) {\n if (existsSync(composePath)) {\n try {\n // Don't pass -p: the compose file's `name:` field is the authority.\n // Passing -p with a different value creates a second Docker project\n // on container restart, splitting services onto separate networks.\n await execAsync(`docker compose -f \"${composePath}\" up -d --build`, { cwd: dirname(composePath), timeout: 300000 });\n result.steps.push(`Started containers from ${basename(composePath)}`);\n } catch (error) {\n result.errors.push(`Failed to start containers: ${error}`);\n }\n break;\n }\n }\n }\n\n // Pre-trust workspace directory in Claude Code so agents don't get the trust prompt\n try {\n preTrustDirectory(workspacePath);\n result.steps.push('Pre-trusted workspace in Claude Code');\n } catch {\n // Non-fatal — agent can still work, user will just see trust prompt\n }\n\n result.success = result.errors.length === 0;\n return result;\n}\n\n/**\n * Pre-register a directory as trusted in Claude Code's ~/.claude.json.\n * This prevents the \"Quick safety check: Is this a project you created or one you trust?\" prompt\n * when agents are spawned in dynamically-created workspace directories.\n */\nexport function preTrustDirectory(dirPath: string): void {\n const claudeJsonPath = join(homedir(), '.claude.json');\n if (!existsSync(claudeJsonPath)) return;\n\n const data = JSON.parse(readFileSync(claudeJsonPath, 'utf8'));\n if (!data.projects) data.projects = {};\n\n // Only add if not already present\n if (data.projects[dirPath]) {\n if (!data.projects[dirPath].hasTrustDialogAccepted) {\n data.projects[dirPath].hasTrustDialogAccepted = true;\n writeFileSync(claudeJsonPath, JSON.stringify(data, null, 2), 'utf8');\n }\n return;\n }\n\n data.projects[dirPath] = {\n allowedTools: [],\n mcpContextUris: [],\n mcpServers: {},\n enabledMcpjsonServers: [],\n disabledMcpjsonServers: [],\n hasTrustDialogAccepted: true,\n projectOnboardingSeenCount: 0,\n hasClaudeMdExternalIncludesApproved: false,\n hasClaudeMdExternalIncludesWarningShown: false,\n };\n\n writeFileSync(claudeJsonPath, JSON.stringify(data, null, 2), 'utf8');\n}\n\nexport interface WorkspaceRemoveOptions {\n projectConfig: ProjectConfig;\n featureName: string;\n dryRun?: boolean;\n}\n\nexport interface WorkspaceRemoveResult {\n success: boolean;\n errors: string[];\n steps: string[];\n}\n\n/**\n * Result of Docker container cleanup for a workspace.\n */\nexport interface DockerCleanupResult {\n /** Whether compose files were found (containers may or may not have been running) */\n containersFound: boolean;\n /** Human-readable log of cleanup steps taken */\n steps: string[];\n}\n\n/**\n * Stop Docker containers and clean up Docker-created files for a workspace.\n *\n * Extracted as a standalone function so it can be used by:\n * - removeWorkspace() during normal workspace removal\n * - deep-wipe endpoint for complete issue cleanup\n * - workspace-migrate for pre-migration cleanup\n *\n * Failures are logged but never thrown — callers should not fail if Docker is unavailable.\n */\nexport async function stopWorkspaceDocker(\n workspacePath: string,\n projectName: string,\n featureName: string,\n): Promise<DockerCleanupResult> {\n const result: DockerCleanupResult = {\n containersFound: false,\n steps: [],\n };\n\n // Find all compose files in devcontainer directory (some projects use multiple)\n const devcontainerDir = join(workspacePath, '.devcontainer');\n const composeFiles: string[] = [];\n\n if (existsSync(devcontainerDir)) {\n const possibleFiles = [\n 'docker-compose.devcontainer.yml',\n 'docker-compose.yml',\n 'compose.yml',\n 'compose.infra.yml',\n 'compose.override.yml',\n ];\n for (const file of possibleFiles) {\n const fullPath = join(devcontainerDir, file);\n if (existsSync(fullPath)) {\n composeFiles.push(fullPath);\n }\n }\n }\n\n // Fallback: check for compose file in workspace root\n if (composeFiles.length === 0) {\n const rootCompose = join(workspacePath, 'docker-compose.yml');\n if (existsSync(rootCompose)) {\n composeFiles.push(rootCompose);\n }\n }\n\n if (composeFiles.length > 0) {\n result.containersFound = true;\n try {\n const fileFlags = composeFiles.map(f => `-f \"${f}\"`).join(' ');\n const cwd = existsSync(devcontainerDir) ? devcontainerDir : workspacePath;\n\n // Don't pass -p: let the compose file's `name:` field determine the project.\n // This must match what was used at startup to target the correct containers.\n await execAsync(`docker compose ${fileFlags} down -v --remove-orphans`, {\n cwd,\n timeout: 60000,\n });\n result.steps.push(`Stopped Docker containers (${composeFiles.length} compose files)`);\n } catch (error: any) {\n // Log but don't fail — containers might not be running\n result.steps.push(`Docker cleanup attempted (${error.message?.split('\\n')[0] || 'containers may not be running'})`);\n }\n }\n\n // Clean up Docker-created files (root-owned in containers)\n try {\n await execAsync(\n `docker run --rm -v \"${workspacePath}:/workspace\" alpine sh -c \"find /workspace -user root -delete 2>&1 | tail -100 || true\"`,\n { timeout: 30000, maxBuffer: 10 * 1024 * 1024 }\n );\n result.steps.push('Cleaned up Docker-created files');\n } catch {\n // Alpine container might not be available\n }\n\n return result;\n}\n\n/**\n * Remove a workspace\n */\nexport async function removeWorkspace(options: WorkspaceRemoveOptions): Promise<WorkspaceRemoveResult> {\n const { projectConfig, featureName, dryRun } = options;\n const result: WorkspaceRemoveResult = {\n success: true,\n errors: [],\n steps: [],\n };\n\n const workspaceConfig = projectConfig.workspace || getDefaultWorkspaceConfig();\n const workspacesDir = join(projectConfig.path, workspaceConfig.workspaces_dir || 'workspaces');\n const featureFolder = `feature-${featureName}`;\n const workspacePath = join(workspacesDir, featureFolder);\n\n if (!existsSync(workspacePath)) {\n result.success = false;\n result.errors.push(`Workspace not found at ${workspacePath}`);\n return result;\n }\n\n if (dryRun) {\n result.steps.push('[DRY RUN] Would remove workspace at: ' + workspacePath);\n return result;\n }\n\n // Stop TLDR daemon for workspace (if it exists)\n const venvPath = join(workspacePath, '.venv');\n if (existsSync(venvPath)) {\n try {\n const { getTldrDaemonService } = await import('./tldr-daemon.js');\n const tldrService = getTldrDaemonService(workspacePath, venvPath);\n await tldrService.stop();\n result.steps.push('Stopped TLDR daemon');\n } catch (error: any) {\n // Non-fatal - daemon may not be running\n console.warn(`⚠ Failed to stop TLDR daemon: ${error?.message}`);\n }\n }\n\n // Stop Docker containers and clean up Docker-created files\n const dockerResult = await stopWorkspaceDocker(workspacePath, projectConfig.name || 'workspace', featureName);\n result.steps.push(...dockerResult.steps);\n\n // Remove worktrees\n if (workspaceConfig.type === 'polyrepo' && workspaceConfig.repos) {\n for (const repo of workspaceConfig.repos) {\n const repoPath = join(projectConfig.path, repo.path);\n const targetPath = join(workspacePath, repo.name);\n const branchPrefix = repo.branch_prefix || 'feature/';\n const branchName = `${branchPrefix}${featureName}`;\n\n const worktreeResult = await removeWorktree(repoPath, targetPath, branchName);\n if (worktreeResult.success) {\n result.steps.push(`Removed worktree for ${repo.name}`);\n } else {\n result.errors.push(worktreeResult.message);\n }\n }\n } else {\n // Monorepo: remove single worktree\n const branchName = `feature/${featureName}`;\n const worktreeResult = await removeWorktree(projectConfig.path, workspacePath, branchName);\n if (worktreeResult.success) {\n result.steps.push('Removed worktree');\n } else {\n result.errors.push(worktreeResult.message);\n }\n }\n\n // Remove DNS entries\n if (workspaceConfig.dns) {\n const placeholders = createPlaceholders(projectConfig, featureName, workspacePath);\n\n const dnsMethod = workspaceConfig.dns.sync_method || 'wsl2hosts';\n for (const entryPattern of workspaceConfig.dns.entries) {\n const hostname = replacePlaceholders(entryPattern, placeholders);\n if (removeDnsEntry(dnsMethod, hostname)) {\n result.steps.push(`Removed DNS entry: ${hostname}`);\n }\n }\n }\n\n // Remove Cloudflare tunnel entries\n if (workspaceConfig.tunnel) {\n const placeholders = createPlaceholders(projectConfig, featureName, workspacePath);\n const tunnelResult = await removeTunnelIngress(workspaceConfig.tunnel, placeholders);\n result.steps.push(...tunnelResult.steps);\n }\n\n // Remove Hume EVI config\n if (workspaceConfig.hume) {\n const placeholders = createPlaceholders(projectConfig, featureName, workspacePath);\n const humeResult = await deleteHumeConfig(workspaceConfig.hume, placeholders);\n result.steps.push(...humeResult.steps);\n }\n\n // Release ports\n if (workspaceConfig.ports) {\n for (const [portName] of Object.entries(workspaceConfig.ports)) {\n const portFile = join(projectConfig.path, `.${portName}-ports`);\n if (releasePort(portFile, featureFolder)) {\n result.steps.push(`Released ${portName} port`);\n }\n }\n }\n\n // Remove workspace directory\n try {\n await execAsync(`rm -rf \"${workspacePath}\"`, { maxBuffer: 10 * 1024 * 1024 });\n result.steps.push('Removed workspace directory');\n } catch (error) {\n result.errors.push(`Failed to remove workspace directory: ${error}`);\n }\n\n result.success = result.errors.length === 0;\n return result;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,OAEK;AACP,SAAS,MAAM,UAAU,eAAe;AAuBxC,SAAS,SAAS,WAAmB,WAA6B;AAChE,QAAM,SAAmB,CAAC;AAC1B,MAAI,CAAC,WAAW,SAAS,EAAG,QAAO;AAEnC,WAAS,KAAK,KAAmB;AAC/B,UAAM,UAAU,YAAY,KAAK,EAAE,eAAe,KAAK,CAAC;AACxD,eAAW,SAAS,SAAS;AAC3B,YAAM,WAAW,KAAK,KAAK,MAAM,IAAI;AACrC,UAAI,MAAM,YAAY,GAAG;AACvB,aAAK,QAAQ;AAAA,MACf,WAAW,MAAM,OAAO,GAAG;AACzB,cAAM,MAAM,SAAS,WAAW,QAAQ;AACxC,cAAM,aAAa,KAAK,WAAW,GAAG;AACtC,kBAAU,QAAQ,UAAU,GAAG,EAAE,WAAW,KAAK,CAAC;AAClD,qBAAa,UAAU,UAAU;AACjC,eAAO,KAAK,GAAG;AAAA,MACjB;AAAA,IACF;AAAA,EACF;AAEA,OAAK,SAAS;AACd,SAAO;AACT;AAYO,SAAS,yBAAyB,eAAoC;AAC3E,QAAM,YAAY,KAAK,eAAe,SAAS;AAC/C,QAAM,eAAe,KAAK,WAAW,2BAA2B;AAChE,QAAM,WAAW,aAAa,YAAY;AAE1C,QAAM,SAAsB;AAAA,IAC1B,OAAO,CAAC;AAAA,IACR,SAAS,CAAC;AAAA,IACV,SAAS,CAAC;AAAA,IACV,WAAW,CAAC;AAAA,EACd;AAGA,YAAU,KAAK,WAAW,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AACxD,YAAU,KAAK,WAAW,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAGxD,QAAM,UAAgF;AAAA,IACpF,EAAE,UAAU,UAAU,WAAW,YAAY,cAAc,SAAS;AAAA,IACpE,EAAE,UAAU,UAAU,WAAW,kBAAkB,cAAc,SAAS;AAAA,IAC1E,EAAE,UAAU,SAAS,WAAW,iBAAiB,cAAc,QAAQ;AAAA,EACzE;AAEA,aAAW,EAAE,UAAU,WAAW,aAAa,KAAK,SAAS;AAC3D,QAAI,CAAC,WAAW,SAAS,EAAG;AAE5B,UAAM,SAAS,eAAe,GAAG,YAAY,MAAM;AACnD,UAAM,QAAQ,mBAAmB,WAAW,EAAE;AAE9C,eAAW,QAAQ,OAAO;AACxB,YAAM,eAAe,GAAG,MAAM,GAAG,KAAK,YAAY;AAClD,YAAM,aAAa,KAAK,WAAW,YAAY;AAC/C,YAAM,aAAa,SAAS,KAAK,YAAY;AAG7C,YAAM,SAAS,sBAAsB,YAAY,cAAc,QAAQ;AAEvE,cAAQ,OAAO,QAAQ;AAAA,QACrB,KAAK;AAEH,oBAAU,QAAQ,UAAU,GAAG,EAAE,WAAW,KAAK,CAAC;AAClD,uBAAa,KAAK,cAAc,UAAU;AAC1C,2BAAiB,UAAU,cAAc,YAAY,YAAY;AACjE,iBAAO,MAAM,KAAK,YAAY;AAC9B;AAAA,QAEF,KAAK;AAEH,uBAAa,KAAK,cAAc,UAAU;AAC1C,2BAAiB,UAAU,cAAc,YAAY,YAAY;AACjE,iBAAO,QAAQ,KAAK,YAAY;AAChC;AAAA,QAEF,KAAK;AAEH,iBAAO,QAAQ,KAAK,GAAG,YAAY,qBAAqB;AACxD;AAAA,QAEF,KAAK;AAEH,iBAAO,QAAQ,KAAK,GAAG,YAAY,eAAe;AAClD;AAAA,MACJ;AAAA,IACF;AAAA,EACF;AAGA,gBAAc,cAAc,QAAQ;AAEpC,SAAO;AACT;AAaO,SAAS,4BACd,eACA,aACA,WACU;AACV,QAAM,YAAY,KAAK,eAAe,SAAS;AAC/C,QAAM,eAAe,KAAK,WAAW,2BAA2B;AAChE,QAAM,WAAW,aAAa,YAAY;AAC1C,QAAM,YAAsB,CAAC;AAE7B,MAAI,CAAC,WAAW,WAAW,EAAG,QAAO;AAErC,MAAI,aAAa,UAAU,SAAS,GAAG;AAErC,eAAW,EAAE,QAAQ,OAAO,KAAK,WAAW;AAC1C,YAAM,aAAa,KAAK,aAAa,MAAM;AAC3C,UAAI,CAAC,WAAW,UAAU,EAAG;AAE7B,YAAM,aAAa,KAAK,eAAe,MAAM;AAC7C,gBAAU,QAAQ,UAAU,GAAG,EAAE,WAAW,KAAK,CAAC;AAGlD,UAAI,OAAO,SAAS,WAAW,GAAG;AAGhC;AAAA,MACF;AAEA,mBAAa,YAAY,UAAU;AAGnC,UAAI,OAAO,WAAW,UAAU,GAAG;AACjC,cAAM,eAAe,OAAO,MAAM,WAAW,MAAM;AACnD,cAAM,OAAO,SAAS,UAAU;AAChC,yBAAiB,UAAU,cAAc,MAAM,kBAAkB;AACjE,kBAAU,KAAK,YAAY;AAAA,MAC7B;AAAA,IACF;AAAA,EACF,OAAO;AAEL,UAAM,mBAAmB,KAAK,aAAa,SAAS;AACpD,QAAI,WAAW,gBAAgB,GAAG;AAChC,YAAM,SAAS,SAAS,kBAAkB,SAAS;AACnD,iBAAW,OAAO,QAAQ;AACxB,cAAM,aAAa,KAAK,WAAW,GAAG;AACtC,cAAM,OAAO,SAAS,UAAU;AAChC,yBAAiB,UAAU,KAAK,MAAM,kBAAkB;AACxD,kBAAU,KAAK,GAAG;AAAA,MACpB;AAAA,IACF;AAAA,EACF;AAGA,gBAAc,cAAc,QAAQ;AAEpC,SAAO;AACT;AA7MA;AAAA;AAAA;AAAA;AAUA;AACA;AAAA;AAAA;;;ACLA,SAAS,cAAAA,aAAY,aAAAC,YAAW,iBAAAC,gBAAe,gBAAAC,eAAc,eAAAC,cAAa,gBAAAC,eAAc,aAAa,WAAW,oBAAoB;AACpI,SAAS,QAAAC,OAAM,WAAAC,UAAS,UAAU,eAAe;AACjD,SAAS,eAAe;AACxB,SAAS,YAAY;AACrB,SAAS,iBAAiB;AAgC1B,SAAS,mBACP,eACA,aACA,eACsB;AACtB,QAAM,gBAAgB,WAAW,WAAW;AAC5C,QAAM,SAAS,cAAc,WAAW,KAAK,UAAU;AAEvD,SAAO;AAAA,IACL,cAAc;AAAA,IACd,gBAAgB;AAAA,IAChB,aAAa,WAAW,WAAW;AAAA,IACnC,iBAAiB,GAAG,SAAS,cAAc,IAAI,CAAC,IAAI,aAAa;AAAA,IACjE,QAAQ;AAAA,IACR,cAAc,SAAS,cAAc,IAAI;AAAA,IACzC,cAAc,cAAc;AAAA,IAC5B,cAAcA,SAAQ,cAAc,IAAI;AAAA,IACxC,gBAAgB;AAAA,IAChB,MAAM,QAAQ;AAAA,EAChB;AACF;AAMA,SAAS,oBAAoB,UAAwB;AACnD,MAAI,CAACP,YAAW,QAAQ,EAAG;AAE3B,MAAI,UAAUG,cAAa,UAAU,OAAO;AAC5C,QAAM,kBAAkB;AAIxB,QAAM,eAAe;AAAA,IACnB;AAAA;AAAA,IACA;AAAA;AAAA,EACF;AAEA,aAAW,WAAW,cAAc;AAClC,cAAU,QAAQ,QAAQ,SAAS,UAAU;AAAA,EAC/C;AAEA,MAAI,YAAY,iBAAiB;AAC/B,IAAAD,eAAc,UAAU,SAAS,OAAO;AAAA,EAC1C;AACF;AAKA,SAAS,oBAAoB,MAAuB;AAClD,SAAO,kBAAkB,KAAK,IAAI;AACpC;AASA,eAAe,eACb,UACA,YACA,YACA,gBAAwB,QACwB;AAChD,MAAI;AAEF,UAAM,UAAU,oBAAoB,EAAE,KAAK,SAAS,CAAC;AAGrD,UAAM,UAAU,sBAAsB,EAAE,KAAK,SAAS,CAAC;AAGvD,UAAM,EAAE,QAAQ,cAAc,IAAI,MAAM,UAAU,qBAAqB,EAAE,KAAK,SAAS,CAAC;AACxF,UAAM,EAAE,QAAQ,eAAe,IAAI,MAAM,UAAU,wBAAwB,EAAE,KAAK,SAAS,CAAC;AAE5F,UAAM,eACJ,cAAc,SAAS,UAAU,KACjC,eAAe,SAAS,UAAU,UAAU,EAAE;AAEhD,QAAI,cAAc;AAChB,YAAM,UAAU,qBAAqB,UAAU,MAAM,UAAU,KAAK,EAAE,KAAK,SAAS,CAAC;AAAA,IACvF,OAAO;AAEL,YAAM,UAAU,qBAAqB,UAAU,SAAS,UAAU,MAAM,aAAa,KAAK,EAAE,KAAK,SAAS,CAAC;AAAA,IAC7G;AAGA,UAAM,UAAU,+BAA+B,EAAE,KAAK,WAAW,CAAC,EAAE,MAAM,MAAM;AAAA,IAAC,CAAC;AAElF,WAAO,EAAE,SAAS,MAAM,SAAS,uBAAuB,UAAU,GAAG;AAAA,EACvE,SAAS,OAAO;AACd,WAAO,EAAE,SAAS,OAAO,SAAS,8BAA8B,KAAK,GAAG;AAAA,EAC1E;AACF;AAKA,eAAe,eACb,UACA,YACA,YACgD;AAChD,MAAI;AAEF,UAAM,UAAU,wBAAwB,UAAU,aAAa,EAAE,KAAK,SAAS,CAAC,EAAE,MAAM,MAAM;AAAA,IAAC,CAAC;AAGhG,UAAM,UAAU,kBAAkB,UAAU,KAAK,EAAE,KAAK,SAAS,CAAC,EAAE,MAAM,MAAM;AAAA,IAAC,CAAC;AAElF,WAAO,EAAE,SAAS,MAAM,SAAS,uBAAuB,UAAU,GAAG;AAAA,EACvE,SAAS,OAAO;AACd,WAAO,EAAE,SAAS,OAAO,SAAS,8BAA8B,KAAK,GAAG;AAAA,EAC1E;AACF;AAQA,SAAS,WACP,UACA,eACA,OACQ;AAER,MAAI,CAACF,YAAW,QAAQ,GAAG;AACzB,IAAAC,WAAUM,SAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAChD,IAAAL,eAAc,UAAU,EAAE;AAAA,EAC5B;AAEA,QAAM,UAAUC,cAAa,UAAU,OAAO;AAC9C,QAAM,QAAQ,QAAQ,MAAM,IAAI,EAAE,OAAO,OAAO;AAGhD,aAAW,QAAQ,OAAO;AACxB,UAAM,CAAC,QAAQ,IAAI,IAAI,KAAK,MAAM,GAAG;AACrC,QAAI,WAAW,eAAe;AAC5B,aAAO,SAAS,MAAM,EAAE;AAAA,IAC1B;AAAA,EACF;AAGA,QAAM,YAAY,IAAI,IAAI,MAAM,IAAI,OAAK,SAAS,EAAE,MAAM,GAAG,EAAE,CAAC,GAAG,EAAE,CAAC,CAAC;AACvE,WAAS,OAAO,MAAM,CAAC,GAAG,QAAQ,MAAM,CAAC,GAAG,QAAQ;AAClD,QAAI,CAAC,UAAU,IAAI,IAAI,GAAG;AACxB,MAAAD,eAAc,UAAU,WAAW,QAAQ,SAAS,IAAI,IAAI,KAAK,QAAQ,GAAG,aAAa,IAAI,IAAI;AAAA,CAAI;AACrG,aAAO;AAAA,IACT;AAAA,EACF;AAEA,QAAM,IAAI,MAAM,+BAA+B,MAAM,CAAC,CAAC,IAAI,MAAM,CAAC,CAAC,EAAE;AACvE;AAKA,SAAS,YAAY,UAAkB,eAAgC;AACrE,MAAI;AACF,QAAI,CAACF,YAAW,QAAQ,EAAG,QAAO;AAElC,QAAI,UAAUG,cAAa,UAAU,OAAO;AAC5C,UAAM,QAAQ,QAAQ,MAAM,IAAI,EAAE,OAAO,UAAQ,CAAC,KAAK,WAAW,GAAG,aAAa,GAAG,CAAC;AACtF,IAAAD,eAAc,UAAU,MAAM,KAAK,IAAI,CAAC;AACxC,WAAO;AAAA,EACT,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAKA,SAAS,iBACP,aACA,WACA,cACA,WACU;AACV,QAAM,QAAkB,CAAC;AAEzB,MAAI,CAACF,YAAW,WAAW,GAAG;AAC5B,WAAO;AAAA,EACT;AAGA,MAAI,aAAa,UAAU,SAAS,GAAG;AACrC,eAAW,EAAE,QAAQ,OAAO,KAAK,WAAW;AAC1C,YAAM,aAAaM,MAAK,aAAa,MAAM;AAC3C,YAAM,aAAaA,MAAK,WAAW,MAAM;AAEzC,UAAIN,YAAW,UAAU,GAAG;AAC1B,cAAM,UAAUG,cAAa,YAAY,OAAO;AAChD,cAAM,YAAY,oBAAoB,SAAS,YAAY;AAC3D,QAAAF,WAAUM,SAAQ,UAAU,GAAG,EAAE,WAAW,KAAK,CAAC;AAClD,QAAAL,eAAc,YAAY,SAAS;AACnC,cAAM,KAAK,uBAAuB,MAAM,OAAO,MAAM,EAAE;AAAA,MACzD;AAAA,IACF;AAAA,EACF,OAAO;AAEL,UAAM,QAAQE,aAAY,WAAW;AACrC,eAAW,QAAQ,OAAO;AACxB,UAAI,KAAK,SAAS,WAAW,GAAG;AAC9B,cAAM,aAAaE,MAAK,aAAa,IAAI;AACzC,cAAM,aAAaA,MAAK,WAAW,KAAK,QAAQ,aAAa,EAAE,CAAC;AAEhE,cAAM,UAAUH,cAAa,YAAY,OAAO;AAChD,cAAM,YAAY,oBAAoB,SAAS,YAAY;AAC3D,QAAAD,eAAc,YAAY,SAAS;AACnC,cAAM,KAAK,uBAAuB,IAAI,EAAE;AAAA,MAC1C;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AACT;AAsCA,SAAS,wBACP,WACA,WACA,MACA,cACU;AACV,QAAM,QAAkB,CAAC;AAEzB,aAAW,OAAO,MAAM;AAOtB,QAASM,WAAT,SAAiB,KAAa,MAAsB;AAClD,UAAIC,SAAQ;AACZ,MAAAR,WAAU,MAAM,EAAE,WAAW,KAAK,CAAC;AACnC,YAAM,UAAUG,aAAY,KAAK,EAAE,eAAe,KAAK,CAAC;AACxD,iBAAW,SAAS,SAAS;AAC3B,cAAM,WAAWE,MAAK,KAAK,MAAM,IAAI;AACrC,cAAM,YAAYA,MAAK,MAAM,MAAM,IAAI;AACvC,YAAI,MAAM,YAAY,GAAG;AACvB,UAAAG,UAASD,SAAQ,UAAU,SAAS;AAAA,QACtC,WAAW,MAAM,OAAO,GAAG;AACzB,gBAAM,MAAM,QAAQ,MAAM,IAAI,EAAE,YAAY;AAC5C,cAAI,gBAAgB,gBAAgB,IAAI,GAAG,GAAG;AAC5C,kBAAM,UAAUL,cAAa,UAAU,OAAO;AAC9C,YAAAD,eAAc,WAAW,oBAAoB,SAAS,YAAY,CAAC;AAAA,UACrE,OAAO;AACL,YAAAG,cAAa,UAAU,SAAS;AAAA,UAClC;AACA,UAAAI;AAAA,QACF;AAAA,MACF;AACA,aAAOA;AAAA,IACT;AArBS,kBAAAD;AANT,UAAM,aAAaF,MAAK,WAAW,GAAG;AACtC,UAAM,aAAaA,MAAK,WAAW,GAAG;AAEtC,QAAI,CAACN,YAAW,UAAU,EAAG;AA0B7B,UAAM,QAAQQ,SAAQ,YAAY,UAAU;AAC5C,UAAM,KAAK,UAAU,KAAK,iCAAiC,GAAG,EAAE;AAAA,EAClE;AAEA,SAAO;AACT;AAKA,eAAsB,gBAAgB,SAAiE;AACrG,QAAM,EAAE,eAAe,aAAa,aAAa,OAAO,IAAI;AAC5D,QAAM,SAAgC;AAAA,IACpC,SAAS;AAAA,IACT,eAAe;AAAA,IACf,QAAQ,CAAC;AAAA,IACT,OAAO,CAAC;AAAA,EACV;AAGA,MAAI,CAAC,oBAAoB,WAAW,GAAG;AACrC,WAAO,UAAU;AACjB,WAAO,OAAO,KAAK,0DAA0D;AAC7E,WAAO;AAAA,EACT;AAGA,MAAI,gBAAgB,QAAQ;AAC1B,WAAO,UAAU;AACjB,WAAO,OAAO,KAAK,8DAA8D;AACjF,WAAO;AAAA,EACT;AAEA,QAAM,kBAAkB,cAAc,aAAa,0BAA0B;AAC7E,QAAM,gBAAgBF,MAAK,cAAc,MAAM,gBAAgB,kBAAkB,YAAY;AAC7F,QAAM,gBAAgB,WAAW,WAAW;AAC5C,QAAM,gBAAgBA,MAAK,eAAe,aAAa;AACvD,SAAO,gBAAgB;AAGvB,MAAIN,YAAW,aAAa,GAAG;AAC7B,WAAO,UAAU;AACjB,WAAO,OAAO,KAAK,+BAA+B,aAAa,EAAE;AACjE,WAAO;AAAA,EACT;AAEA,MAAI,QAAQ;AACV,WAAO,MAAM,KAAK,0CAA0C,aAAa;AACzE,WAAO;AAAA,EACT;AAGA,QAAM,eAAe,mBAAmB,eAAe,aAAa,aAAa;AAGjF,EAAAC,WAAU,eAAe,EAAE,WAAW,KAAK,CAAC;AAC5C,SAAO,MAAM,KAAK,6BAA6B;AAG/C,MAAI,gBAAgB,SAAS,cAAc,gBAAgB,OAAO;AAEhE,eAAW,QAAQ,gBAAgB,OAAO;AAGxC,YAAM,cAAcK,MAAK,cAAc,MAAM,KAAK,IAAI;AACtD,YAAM,WAAWN,YAAW,WAAW,IAAI,aAAa,WAAW,IAAI;AACvE,YAAM,aAAaM,MAAK,eAAe,KAAK,IAAI;AAChD,YAAM,eAAe,KAAK,iBAAiB;AAC3C,YAAM,aAAa,GAAG,YAAY,GAAG,WAAW;AAEhD,YAAM,gBAAgB,KAAK,kBAAkB,gBAAgB,kBAAkB;AAE/E,YAAM,iBAAiB,MAAM,eAAe,UAAU,YAAY,YAAY,aAAa;AAC3F,UAAI,eAAe,SAAS;AAC1B,eAAO,MAAM,KAAK,wBAAwB,KAAK,IAAI,KAAK,UAAU,UAAU,aAAa,GAAG;AAAA,MAC9F,OAAO;AACL,eAAO,OAAO,KAAK,GAAG,KAAK,IAAI,KAAK,eAAe,OAAO,EAAE;AAC5D,eAAO,UAAU;AAAA,MACnB;AAAA,IACF;AAAA,EACF,OAAO;AAEL,UAAM,aAAa,WAAW,WAAW;AACzC,UAAM,gBAAgB,gBAAgB,kBAAkB;AACxD,UAAM,iBAAiB,MAAM,eAAe,cAAc,MAAM,eAAe,YAAY,aAAa;AACxG,QAAI,eAAe,SAAS;AAC1B,aAAO,MAAM,KAAK,qBAAqB,UAAU,UAAU,aAAa,GAAG;AAAA,IAC7E,OAAO;AACL,aAAO,OAAO,KAAK,eAAe,OAAO;AACzC,aAAO,UAAU;AAAA,IACnB;AAAA,EACF;AAIA,QAAM,kBAAkBA,MAAK,eAAe,eAAe;AAC3D,MAAIN,YAAW,eAAe,GAAG;AAC/B,UAAM,eAAeI,aAAY,eAAe,EAC7C,OAAO,OAAK,EAAE,SAAS,SAAS,MAAM,EAAE,SAAS,MAAM,KAAK,EAAE,SAAS,OAAO,EAAE;AACnF,eAAW,eAAe,cAAc;AACtC,0BAAoBE,MAAK,iBAAiB,WAAW,CAAC;AAAA,IACxD;AACA,QAAI,aAAa,SAAS,GAAG;AAC3B,aAAO,MAAM,KAAK,aAAa,aAAa,MAAM,6CAA6C;AAAA,IACjG;AAAA,EACF;AAGA,MAAI;AAEF,UAAM,UAAU,mBAAmB;AACnC,UAAM,WAAWA,MAAK,eAAe,OAAO;AAC5C,UAAM,UAAUA,MAAK,UAAU,OAAO,MAAM;AAG5C,UAAM,eAAeA,MAAK,cAAc,MAAM,SAAS,OAAO,MAAM;AACpE,UAAM,iBAAiBN,YAAW,YAAY;AAE9C,QAAI,gBAAgB;AAElB,YAAM,UAAU,UAAUM,MAAK,cAAc,MAAM,OAAO,CAAC,MAAM,QAAQ,GAAG;AAC5E,aAAO,MAAM,KAAK,qCAAqC;AAAA,IACzD,OAAO;AAEL,YAAM,UAAU,oBAAoB,QAAQ,KAAK,EAAE,KAAK,cAAc,CAAC;AACvE,YAAM,UAAUA,MAAK,UAAU,OAAO,KAAK;AAC3C,YAAM,UAAU,IAAI,OAAO,sBAAsB,EAAE,KAAK,eAAe,SAAS,KAAO,CAAC;AACxF,aAAO,MAAM,KAAK,4CAA4C;AAG9D,YAAM,cAAcA,MAAK,cAAc,MAAM,WAAW,WAAW,yBAAyB;AAC5F,UAAIN,YAAW,WAAW,GAAG;AAC3B,cAAM,UAAU,YAAY,WAAW,MAAM,QAAQ,GAAG;AACxD,eAAO,MAAM,KAAK,kCAAkC;AAAA,MACtD;AAAA,IACF;AAGA,QAAI,CAACA,YAAW,OAAO,GAAG;AACxB,aAAO,MAAM,KAAK,kEAAkE;AAAA,IACtF,OAAO;AAEL,YAAM,cAAcM,MAAK,cAAc,MAAM,OAAO;AACpD,YAAM,mBAAmBA,MAAK,eAAe,OAAO;AAEpD,UAAIN,YAAW,WAAW,GAAG;AAC3B,cAAM,UAAU,UAAU,WAAW,MAAM,gBAAgB,GAAG;AAC9D,eAAO,MAAM,KAAK,oCAAoC;AAAA,MACxD;AAGA,YAAM,EAAE,qBAAqB,IAAI,MAAM,OAAO,2BAAkB;AAChE,YAAM,cAAc,qBAAqB,eAAe,QAAQ;AAChE,YAAM,YAAY,MAAM,IAAI;AAC5B,aAAO,MAAM,KAAK,qBAAqB;AAIvC,UAAI;AACF,cAAM,YAAY,KAAK,IAAI;AAC3B,eAAO,MAAM,KAAK,wCAAwC;AAAA,MAC5D,QAAQ;AAAA,MAER;AAAA,IACF;AAAA,EACF,SAAS,OAAY;AAEnB,QAAI,MAAM,SAAS,SAAS,SAAS,GAAG;AACtC,aAAO,MAAM,KAAK,4CAA4C;AAAA,IAChE,OAAO;AACL,cAAQ,KAAK,6BAAwB,MAAM,OAAO,EAAE;AACpD,aAAO,MAAM,KAAK,sBAAsB,MAAM,OAAO,EAAE;AAAA,IACzD;AAAA,EACF;AAGA,MAAI,gBAAgB,KAAK;AACvB,UAAM,YAAY,gBAAgB,IAAI,eAAe;AACrD,eAAW,gBAAgB,gBAAgB,IAAI,SAAS;AACtD,YAAM,WAAW,oBAAoB,cAAc,YAAY;AAE/D,UAAI,YAAY,WAAW,QAAQ,GAAG;AACpC,eAAO,MAAM,KAAK,oBAAoB,QAAQ,KAAK,SAAS,GAAG;AAAA,MACjE;AAAA,IACF;AAGA,QAAI,cAAc,aAAa;AAC7B,YAAM,SAAS,MAAM,iBAAiB;AACtC,UAAI,QAAQ;AACV,eAAO,MAAM,KAAK,kCAAkC;AAAA,MACtD;AAAA,IACF;AAAA,EACF;AAGA,MAAI,gBAAgB,OAAO;AACzB,eAAW,CAAC,UAAU,UAAU,KAAK,OAAO,QAAQ,gBAAgB,KAAK,GAAG;AAC1E,YAAM,WAAWM,MAAK,cAAc,MAAM,IAAI,QAAQ,QAAQ;AAC9D,UAAI;AACF,cAAM,OAAO,WAAW,UAAU,eAAe,WAAW,KAAK;AACjE,eAAO,MAAM,KAAK,YAAY,QAAQ,UAAU,IAAI,EAAE;AAEtD,QAAC,aAAqB,GAAG,SAAS,YAAY,CAAC,OAAO,IAAI,OAAO,IAAI;AAAA,MACvE,SAAS,OAAO;AACd,eAAO,OAAO,KAAK,oBAAoB,QAAQ,UAAU,KAAK,EAAE;AAAA,MAClE;AAAA,IACF;AAAA,EACF;AAGA,QAAM,cAAc,yBAAyB,aAAa;AAC1D,QAAM,aAAa,YAAY,MAAM,SAAS,YAAY,QAAQ;AAClE,MAAI,aAAa,GAAG;AAClB,WAAO,MAAM,KAAK,aAAa,UAAU,sBAAsB,YAAY,MAAM,MAAM,SAAS,YAAY,QAAQ,MAAM,WAAW;AAAA,EACvI;AAGA,MAAI,gBAAgB,OAAO,cAAc;AACvC,UAAM,cAAcA,MAAK,cAAc,MAAM,gBAAgB,MAAM,YAAY;AAG/E,UAAM,gBAAgB;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA,gBAAgB,MAAM;AAAA,IACxB;AACA,WAAO,MAAM,KAAK,GAAG,aAAa;AAGlC,UAAM,aAAa,gBAAgB,MAAM,aAAa,gBAAgB,MAAM;AAC5E,QAAI,YAAY;AACd,YAAM,YAAY,wBAAwB,aAAa,eAAe,YAAY,YAAY;AAC9F,aAAO,MAAM,KAAK,GAAG,SAAS;AAAA,IAChC;AAAA,EACF;AAGA,MAAI,gBAAgB,KAAK,UAAU;AACjC,UAAM,aAAa,oBAAoB,gBAAgB,IAAI,UAAU,YAAY;AACjF,IAAAJ,eAAcI,MAAK,eAAe,MAAM,GAAG,UAAU;AACrD,WAAO,MAAM,KAAK,mBAAmB;AAAA,EACvC;AAGA,MAAI,gBAAgB,QAAQ,kBAAkB;AAC5C,UAAM,cAAcA,MAAK,cAAc,MAAM,gBAAgB,OAAO,gBAAgB;AACpF,UAAMI,mBAAkBJ,MAAK,eAAe,eAAe;AAC3D,IAAAL,WAAUS,kBAAiB,EAAE,WAAW,KAAK,CAAC;AAE9C,UAAM,gBAAgB,iBAAiB,aAAaA,kBAAiB,YAAY;AACjF,WAAO,MAAM,KAAK,GAAG,aAAa;AAGlC,QAAIV,YAAW,WAAW,GAAG;AAC3B,YAAM,QAAQI,aAAY,WAAW;AACrC,iBAAW,QAAQ,OAAO;AACxB,YAAI,CAAC,KAAK,SAAS,WAAW,GAAG;AAC/B,gBAAM,aAAaE,MAAK,aAAa,IAAI;AACzC,gBAAM,aAAaA,MAAKI,kBAAiB,IAAI;AAC7C,UAAAL,cAAa,YAAY,UAAU;AAAA,QACrC;AAAA,MACF;AAAA,IACF;AAIA,UAAM,eAAeD,aAAYM,gBAAe,EAC7C,OAAO,OAAK,EAAE,SAAS,SAAS,MAAM,EAAE,SAAS,MAAM,KAAK,EAAE,SAAS,OAAO,EAAE;AACnF,eAAW,eAAe,cAAc;AACtC,0BAAoBJ,MAAKI,kBAAiB,WAAW,CAAC;AAAA,IACxD;AACA,QAAI,aAAa,SAAS,GAAG;AAC3B,aAAO,MAAM,KAAK,aAAa,aAAa,MAAM,6CAA6C;AAAA,IACjG;AAIA,UAAM,uBAAuBJ,MAAKI,kBAAiB,KAAK;AACxD,UAAM,kBAAkBJ,MAAK,eAAe,KAAK;AACjD,QAAIN,YAAW,oBAAoB,KAAK,CAACA,YAAW,eAAe,GAAG;AACpE,UAAI;AACF,oBAAY,qBAAqB,eAAe;AAChD,kBAAU,sBAAsB,GAAK;AACrC,eAAO,MAAM,KAAK,uBAAuB;AAAA,MAC3C,SAAS,OAAO;AACd,eAAO,OAAO,KAAK,mCAAmC,KAAK,EAAE;AAAA,MAC/D;AAAA,IACF;AAAA,EACF;AAOA,MAAI,gBAAgB,QAAQ;AAC1B,UAAM,eAAe,MAAM,iBAAiB,gBAAgB,QAAQ,YAAY;AAChF,WAAO,MAAM,KAAK,GAAG,aAAa,KAAK;AACvC,QAAI,CAAC,aAAa,SAAS;AACzB,aAAO,OAAO,KAAK,mDAAmD;AAAA,IACxE;AAAA,EACF;AAGA,MAAI,gBAAgB,MAAM;AACxB,UAAM,aAAa,MAAM,iBAAiB,gBAAgB,MAAM,YAAY;AAC5E,WAAO,MAAM,KAAK,GAAG,WAAW,KAAK;AACrC,QAAI,WAAW,UAAU;AACvB,MAAAE;AAAA,QACEI,MAAK,eAAe,cAAc;AAAA,QAClC,kBAAkB,WAAW,QAAQ;AAAA,sBAAyB,WAAW,QAAQ;AAAA;AAAA,MACnF;AACA,aAAO,MAAM,KAAK,4CAA4C;AAAA,IAChE;AACA,QAAI,CAAC,WAAW,SAAS;AACvB,aAAO,OAAO,KAAK,4DAA4D;AAAA,IACjF;AAAA,EACF;AAGA,MAAI,aAAa;AAEf,QAAI,gBAAgB,QAAQ,SAAS;AACnC,YAAM,cAAcA,MAAK,cAAc,MAAM,gBAAgB,OAAO,OAAO;AAC3E,UAAIN,YAAW,WAAW,GAAG;AAC3B,YAAI;AACF,gBAAM,UAAU,sBAAsB,WAAW,WAAW,EAAE,KAAK,cAAc,KAAK,CAAC;AACvF,iBAAO,MAAM,KAAK,iBAAiB;AAAA,QACrC,SAAS,OAAY;AACnB,gBAAM,MAAM,OAAO,WAAW,OAAO,KAAK;AAC1C,cAAI,IAAI,SAAS,2BAA2B,KAAK,IAAI,SAAS,wBAAwB,GAAG;AAEvF,mBAAO,MAAM,KAAK,uCAAuC;AAAA,UAC3D,OAAO;AACL,mBAAO,OAAO,KAAK,4BAA4B,KAAK,EAAE;AAAA,UACxD;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAGA,UAAM,mBAAmB;AAAA,MACvBM,MAAK,eAAe,oBAAoB;AAAA,MACxCA,MAAK,eAAe,qBAAqB;AAAA,MACzCA,MAAK,eAAe,iBAAiB,oBAAoB;AAAA,MACzDA,MAAK,eAAe,iBAAiB,iCAAiC;AAAA,IACxE;AAEA,eAAW,eAAe,kBAAkB;AAC1C,UAAIN,YAAW,WAAW,GAAG;AAC3B,YAAI;AAIF,gBAAM,UAAU,sBAAsB,WAAW,mBAAmB,EAAE,KAAKO,SAAQ,WAAW,GAAG,SAAS,IAAO,CAAC;AAClH,iBAAO,MAAM,KAAK,2BAA2B,SAAS,WAAW,CAAC,EAAE;AAAA,QACtE,SAAS,OAAO;AACd,iBAAO,OAAO,KAAK,+BAA+B,KAAK,EAAE;AAAA,QAC3D;AACA;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAGA,MAAI;AACF,sBAAkB,aAAa;AAC/B,WAAO,MAAM,KAAK,sCAAsC;AAAA,EAC1D,QAAQ;AAAA,EAER;AAEA,SAAO,UAAU,OAAO,OAAO,WAAW;AAC1C,SAAO;AACT;AAOO,SAAS,kBAAkB,SAAuB;AACvD,QAAM,iBAAiBD,MAAK,QAAQ,GAAG,cAAc;AACrD,MAAI,CAACN,YAAW,cAAc,EAAG;AAEjC,QAAM,OAAO,KAAK,MAAMG,cAAa,gBAAgB,MAAM,CAAC;AAC5D,MAAI,CAAC,KAAK,SAAU,MAAK,WAAW,CAAC;AAGrC,MAAI,KAAK,SAAS,OAAO,GAAG;AAC1B,QAAI,CAAC,KAAK,SAAS,OAAO,EAAE,wBAAwB;AAClD,WAAK,SAAS,OAAO,EAAE,yBAAyB;AAChD,MAAAD,eAAc,gBAAgB,KAAK,UAAU,MAAM,MAAM,CAAC,GAAG,MAAM;AAAA,IACrE;AACA;AAAA,EACF;AAEA,OAAK,SAAS,OAAO,IAAI;AAAA,IACvB,cAAc,CAAC;AAAA,IACf,gBAAgB,CAAC;AAAA,IACjB,YAAY,CAAC;AAAA,IACb,uBAAuB,CAAC;AAAA,IACxB,wBAAwB,CAAC;AAAA,IACzB,wBAAwB;AAAA,IACxB,4BAA4B;AAAA,IAC5B,qCAAqC;AAAA,IACrC,yCAAyC;AAAA,EAC3C;AAEA,EAAAA,eAAc,gBAAgB,KAAK,UAAU,MAAM,MAAM,CAAC,GAAG,MAAM;AACrE;AAkCA,eAAsB,oBACpB,eACA,aACA,aAC8B;AAC9B,QAAM,SAA8B;AAAA,IAClC,iBAAiB;AAAA,IACjB,OAAO,CAAC;AAAA,EACV;AAGA,QAAM,kBAAkBI,MAAK,eAAe,eAAe;AAC3D,QAAM,eAAyB,CAAC;AAEhC,MAAIN,YAAW,eAAe,GAAG;AAC/B,UAAM,gBAAgB;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AACA,eAAW,QAAQ,eAAe;AAChC,YAAM,WAAWM,MAAK,iBAAiB,IAAI;AAC3C,UAAIN,YAAW,QAAQ,GAAG;AACxB,qBAAa,KAAK,QAAQ;AAAA,MAC5B;AAAA,IACF;AAAA,EACF;AAGA,MAAI,aAAa,WAAW,GAAG;AAC7B,UAAM,cAAcM,MAAK,eAAe,oBAAoB;AAC5D,QAAIN,YAAW,WAAW,GAAG;AAC3B,mBAAa,KAAK,WAAW;AAAA,IAC/B;AAAA,EACF;AAEA,MAAI,aAAa,SAAS,GAAG;AAC3B,WAAO,kBAAkB;AACzB,QAAI;AACF,YAAM,YAAY,aAAa,IAAI,OAAK,OAAO,CAAC,GAAG,EAAE,KAAK,GAAG;AAC7D,YAAM,MAAMA,YAAW,eAAe,IAAI,kBAAkB;AAI5D,YAAM,UAAU,kBAAkB,SAAS,6BAA6B;AAAA,QACtE;AAAA,QACA,SAAS;AAAA,MACX,CAAC;AACD,aAAO,MAAM,KAAK,8BAA8B,aAAa,MAAM,iBAAiB;AAAA,IACtF,SAAS,OAAY;AAEnB,aAAO,MAAM,KAAK,6BAA6B,MAAM,SAAS,MAAM,IAAI,EAAE,CAAC,KAAK,+BAA+B,GAAG;AAAA,IACpH;AAAA,EACF;AAGA,MAAI;AACF,UAAM;AAAA,MACJ,uBAAuB,aAAa;AAAA,MACpC,EAAE,SAAS,KAAO,WAAW,KAAK,OAAO,KAAK;AAAA,IAChD;AACA,WAAO,MAAM,KAAK,iCAAiC;AAAA,EACrD,QAAQ;AAAA,EAER;AAEA,SAAO;AACT;AAKA,eAAsB,gBAAgB,SAAiE;AACrG,QAAM,EAAE,eAAe,aAAa,OAAO,IAAI;AAC/C,QAAM,SAAgC;AAAA,IACpC,SAAS;AAAA,IACT,QAAQ,CAAC;AAAA,IACT,OAAO,CAAC;AAAA,EACV;AAEA,QAAM,kBAAkB,cAAc,aAAa,0BAA0B;AAC7E,QAAM,gBAAgBM,MAAK,cAAc,MAAM,gBAAgB,kBAAkB,YAAY;AAC7F,QAAM,gBAAgB,WAAW,WAAW;AAC5C,QAAM,gBAAgBA,MAAK,eAAe,aAAa;AAEvD,MAAI,CAACN,YAAW,aAAa,GAAG;AAC9B,WAAO,UAAU;AACjB,WAAO,OAAO,KAAK,0BAA0B,aAAa,EAAE;AAC5D,WAAO;AAAA,EACT;AAEA,MAAI,QAAQ;AACV,WAAO,MAAM,KAAK,0CAA0C,aAAa;AACzE,WAAO;AAAA,EACT;AAGA,QAAM,WAAWM,MAAK,eAAe,OAAO;AAC5C,MAAIN,YAAW,QAAQ,GAAG;AACxB,QAAI;AACF,YAAM,EAAE,qBAAqB,IAAI,MAAM,OAAO,2BAAkB;AAChE,YAAM,cAAc,qBAAqB,eAAe,QAAQ;AAChE,YAAM,YAAY,KAAK;AACvB,aAAO,MAAM,KAAK,qBAAqB;AAAA,IACzC,SAAS,OAAY;AAEnB,cAAQ,KAAK,sCAAiC,OAAO,OAAO,EAAE;AAAA,IAChE;AAAA,EACF;AAGA,QAAM,eAAe,MAAM,oBAAoB,eAAe,cAAc,QAAQ,aAAa,WAAW;AAC5G,SAAO,MAAM,KAAK,GAAG,aAAa,KAAK;AAGvC,MAAI,gBAAgB,SAAS,cAAc,gBAAgB,OAAO;AAChE,eAAW,QAAQ,gBAAgB,OAAO;AACxC,YAAM,WAAWM,MAAK,cAAc,MAAM,KAAK,IAAI;AACnD,YAAM,aAAaA,MAAK,eAAe,KAAK,IAAI;AAChD,YAAM,eAAe,KAAK,iBAAiB;AAC3C,YAAM,aAAa,GAAG,YAAY,GAAG,WAAW;AAEhD,YAAM,iBAAiB,MAAM,eAAe,UAAU,YAAY,UAAU;AAC5E,UAAI,eAAe,SAAS;AAC1B,eAAO,MAAM,KAAK,wBAAwB,KAAK,IAAI,EAAE;AAAA,MACvD,OAAO;AACL,eAAO,OAAO,KAAK,eAAe,OAAO;AAAA,MAC3C;AAAA,IACF;AAAA,EACF,OAAO;AAEL,UAAM,aAAa,WAAW,WAAW;AACzC,UAAM,iBAAiB,MAAM,eAAe,cAAc,MAAM,eAAe,UAAU;AACzF,QAAI,eAAe,SAAS;AAC1B,aAAO,MAAM,KAAK,kBAAkB;AAAA,IACtC,OAAO;AACL,aAAO,OAAO,KAAK,eAAe,OAAO;AAAA,IAC3C;AAAA,EACF;AAGA,MAAI,gBAAgB,KAAK;AACvB,UAAM,eAAe,mBAAmB,eAAe,aAAa,aAAa;AAEjF,UAAM,YAAY,gBAAgB,IAAI,eAAe;AACrD,eAAW,gBAAgB,gBAAgB,IAAI,SAAS;AACtD,YAAM,WAAW,oBAAoB,cAAc,YAAY;AAC/D,UAAI,eAAe,WAAW,QAAQ,GAAG;AACvC,eAAO,MAAM,KAAK,sBAAsB,QAAQ,EAAE;AAAA,MACpD;AAAA,IACF;AAAA,EACF;AAGA,MAAI,gBAAgB,QAAQ;AAC1B,UAAM,eAAe,mBAAmB,eAAe,aAAa,aAAa;AACjF,UAAM,eAAe,MAAM,oBAAoB,gBAAgB,QAAQ,YAAY;AACnF,WAAO,MAAM,KAAK,GAAG,aAAa,KAAK;AAAA,EACzC;AAGA,MAAI,gBAAgB,MAAM;AACxB,UAAM,eAAe,mBAAmB,eAAe,aAAa,aAAa;AACjF,UAAM,aAAa,MAAM,iBAAiB,gBAAgB,MAAM,YAAY;AAC5E,WAAO,MAAM,KAAK,GAAG,WAAW,KAAK;AAAA,EACvC;AAGA,MAAI,gBAAgB,OAAO;AACzB,eAAW,CAAC,QAAQ,KAAK,OAAO,QAAQ,gBAAgB,KAAK,GAAG;AAC9D,YAAM,WAAWA,MAAK,cAAc,MAAM,IAAI,QAAQ,QAAQ;AAC9D,UAAI,YAAY,UAAU,aAAa,GAAG;AACxC,eAAO,MAAM,KAAK,YAAY,QAAQ,OAAO;AAAA,MAC/C;AAAA,IACF;AAAA,EACF;AAGA,MAAI;AACF,UAAM,UAAU,WAAW,aAAa,KAAK,EAAE,WAAW,KAAK,OAAO,KAAK,CAAC;AAC5E,WAAO,MAAM,KAAK,6BAA6B;AAAA,EACjD,SAAS,OAAO;AACd,WAAO,OAAO,KAAK,yCAAyC,KAAK,EAAE;AAAA,EACrE;AAEA,SAAO,UAAU,OAAO,OAAO,WAAW;AAC1C,SAAO;AACT;AA/8BA,IAuBM,WAmRA;AA1SN;AAAA;AAAA;AAWA;AAOA;AACA;AACA;AACA;AAEA,IAAM,YAAY,UAAU,IAAI;AAmRhC,IAAM,kBAAkB,oBAAI,IAAI;AAAA,MAC9B;AAAA,MAAO;AAAA,MAAO;AAAA,MAAQ;AAAA,MAAS;AAAA,MAAS;AAAA,MAAO;AAAA,MAAO;AAAA,MAAQ;AAAA,MAAQ;AAAA,MAAS;AAAA,IACjF,CAAC;AAAA;AAAA;","names":["existsSync","mkdirSync","writeFileSync","readFileSync","readdirSync","copyFileSync","join","dirname","copyDir","count","devcontainerDir"]}
@@ -1,159 +0,0 @@
1
- import {
2
- createExeProvider,
3
- init_exe_provider
4
- } from "./chunk-ZDNQFWR5.js";
5
- import {
6
- __esm,
7
- init_esm_shims
8
- } from "./chunk-ZHC57RCV.js";
9
-
10
- // src/lib/remote/remote-agents.ts
11
- import { join } from "path";
12
- import { existsSync, readFileSync, writeFileSync, mkdirSync } from "fs";
13
- import { homedir } from "os";
14
- function getRemoteAgentStateFile(agentId) {
15
- return join(AGENTS_DIR, agentId, "remote-state.json");
16
- }
17
- function saveRemoteAgentState(state) {
18
- const dir = join(AGENTS_DIR, state.id);
19
- if (!existsSync(dir)) {
20
- mkdirSync(dir, { recursive: true });
21
- }
22
- writeFileSync(getRemoteAgentStateFile(state.id), JSON.stringify(state, null, 2));
23
- }
24
- function loadRemoteAgentState(agentId) {
25
- const file = getRemoteAgentStateFile(agentId);
26
- if (!existsSync(file)) return null;
27
- try {
28
- return JSON.parse(readFileSync(file, "utf-8"));
29
- } catch {
30
- return null;
31
- }
32
- }
33
- async function remoteSessionExists(exe, vmName, sessionName) {
34
- const result = await exe.ssh(vmName, `tmux has-session -t ${sessionName} 2>/dev/null && echo exists || echo not-found`);
35
- return result.stdout.trim() === "exists";
36
- }
37
- async function spawnRemoteAgent(options) {
38
- const { issueId, workspace, model = "claude-sonnet-4-6", prompt } = options;
39
- const agentId = `agent-${issueId.toLowerCase()}`;
40
- const vmName = workspace.vmName;
41
- const exe = createExeProvider({ infraVm: workspace.infraVm });
42
- const vmStatus = await exe.getStatus(vmName);
43
- if (vmStatus !== "running") {
44
- throw new Error(`VM ${vmName} is not running. Start it with: pan workspace start ${issueId}`);
45
- }
46
- if (await remoteSessionExists(exe, vmName, agentId)) {
47
- throw new Error(`Agent ${agentId} already running on ${vmName}. Use 'pan work tell' to message it.`);
48
- }
49
- const state = {
50
- id: agentId,
51
- issueId,
52
- vmName,
53
- model,
54
- status: "starting",
55
- startedAt: (/* @__PURE__ */ new Date()).toISOString(),
56
- location: "remote"
57
- };
58
- saveRemoteAgentState(state);
59
- let claudeCmd;
60
- if (prompt) {
61
- const promptFile = `/workspace/.panopticon/prompts/${agentId}.md`;
62
- await exe.ssh(vmName, `mkdir -p /workspace/.panopticon/prompts`);
63
- const promptBase64 = Buffer.from(prompt).toString("base64");
64
- await exe.ssh(vmName, `echo '${promptBase64}' | base64 -d > ${promptFile}`);
65
- const launcherScript = `/workspace/.panopticon/prompts/${agentId}-launcher.sh`;
66
- const launcherContent = `#!/bin/bash
67
- export PATH="/usr/local/bin:$PATH"
68
- prompt=$(cat "${promptFile}")
69
- exec claude --dangerously-skip-permissions --model ${model} "$prompt"
70
- `;
71
- const launcherBase64 = Buffer.from(launcherContent).toString("base64");
72
- await exe.ssh(vmName, `echo '${launcherBase64}' | base64 -d > ${launcherScript} && chmod +x ${launcherScript}`);
73
- claudeCmd = `bash ${launcherScript}`;
74
- } else {
75
- claudeCmd = `claude --dangerously-skip-permissions --model ${model}`;
76
- }
77
- const tmuxCmd = `tmux new-session -d -s ${agentId} -c /workspace '${claudeCmd}'`;
78
- const result = await exe.ssh(vmName, tmuxCmd);
79
- if (result.exitCode !== 0) {
80
- state.status = "error";
81
- saveRemoteAgentState(state);
82
- throw new Error(`Failed to start agent: ${result.stderr}`);
83
- }
84
- state.status = "running";
85
- saveRemoteAgentState(state);
86
- return state;
87
- }
88
- async function getRemoteAgentOutput(agentId, vmName, lines = 100) {
89
- const exe = createExeProvider();
90
- const result = await exe.ssh(vmName, `tmux capture-pane -t ${agentId} -p -S -${lines}`);
91
- return result.stdout;
92
- }
93
- async function sendToRemoteAgent(agentId, vmName, message) {
94
- const exe = createExeProvider();
95
- const escapedMessage = message.replace(/'/g, "'\\''");
96
- await exe.ssh(vmName, `tmux send-keys -t ${agentId} '${escapedMessage}'`);
97
- await exe.ssh(vmName, `tmux send-keys -t ${agentId} C-m`);
98
- }
99
- async function isRemoteAgentRunning(agentId, vmName) {
100
- const exe = createExeProvider();
101
- return remoteSessionExists(exe, vmName, agentId);
102
- }
103
- async function killRemoteAgent(agentId, vmName) {
104
- const exe = createExeProvider();
105
- await exe.ssh(vmName, `tmux kill-session -t ${agentId} 2>/dev/null || true`);
106
- const state = loadRemoteAgentState(agentId);
107
- if (state) {
108
- state.status = "stopped";
109
- saveRemoteAgentState(state);
110
- }
111
- }
112
- async function listRemoteAgents(vmName) {
113
- const exe = createExeProvider();
114
- const result = await exe.ssh(vmName, `tmux list-sessions -F "#{session_name}" 2>/dev/null | grep "^agent-" || true`);
115
- if (!result.stdout.trim()) {
116
- return [];
117
- }
118
- return result.stdout.trim().split("\n").filter(Boolean);
119
- }
120
- async function pollRemoteAgentStatus(agentId, vmName) {
121
- const exe = createExeProvider();
122
- const isRunning = await remoteSessionExists(exe, vmName, agentId);
123
- if (!isRunning) {
124
- return { isRunning: false, lastOutput: "", toolUses: [] };
125
- }
126
- const output = await getRemoteAgentOutput(agentId, vmName, 50);
127
- const toolUses = [];
128
- const toolPattern = /(?:Using|Calling|Running)\s+(\w+)\s+tool/gi;
129
- let match;
130
- while ((match = toolPattern.exec(output)) !== null) {
131
- toolUses.push(match[1]);
132
- }
133
- return {
134
- isRunning,
135
- lastOutput: output,
136
- toolUses
137
- };
138
- }
139
- var AGENTS_DIR;
140
- var init_remote_agents = __esm({
141
- "src/lib/remote/remote-agents.ts"() {
142
- init_esm_shims();
143
- init_exe_provider();
144
- AGENTS_DIR = join(homedir(), ".panopticon", "agents");
145
- }
146
- });
147
-
148
- export {
149
- loadRemoteAgentState,
150
- spawnRemoteAgent,
151
- getRemoteAgentOutput,
152
- sendToRemoteAgent,
153
- isRemoteAgentRunning,
154
- killRemoteAgent,
155
- listRemoteAgents,
156
- pollRemoteAgentStatus,
157
- init_remote_agents
158
- };
159
- //# sourceMappingURL=chunk-565HZ6VV.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/lib/remote/remote-agents.ts"],"sourcesContent":["/**\n * Remote Agent Management\n *\n * Spawn and manage Claude agents on remote exe.dev VMs.\n * Agents run in tmux sessions for persistence and monitoring.\n */\n\nimport { ExeProvider, createExeProvider } from './exe-provider.js';\nimport type { RemoteWorkspaceMetadata } from './interface.js';\nimport { join } from 'path';\nimport { existsSync, readFileSync, writeFileSync, mkdirSync } from 'fs';\nimport { homedir } from 'os';\n\nconst AGENTS_DIR = join(homedir(), '.panopticon', 'agents');\n\nexport interface RemoteAgentState {\n id: string;\n issueId: string;\n vmName: string;\n model: string;\n status: 'starting' | 'running' | 'stopped' | 'error';\n startedAt: string;\n lastActivity?: string;\n location: 'remote';\n}\n\n/**\n * Get agent state file path\n */\nfunction getRemoteAgentStateFile(agentId: string): string {\n return join(AGENTS_DIR, agentId, 'remote-state.json');\n}\n\n/**\n * Save remote agent state\n */\nfunction saveRemoteAgentState(state: RemoteAgentState): void {\n const dir = join(AGENTS_DIR, state.id);\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n writeFileSync(getRemoteAgentStateFile(state.id), JSON.stringify(state, null, 2));\n}\n\n/**\n * Load remote agent state\n */\nexport function loadRemoteAgentState(agentId: string): RemoteAgentState | null {\n const file = getRemoteAgentStateFile(agentId);\n if (!existsSync(file)) return null;\n\n try {\n return JSON.parse(readFileSync(file, 'utf-8'));\n } catch {\n return null;\n }\n}\n\n/**\n * Check if remote agent session exists\n */\nasync function remoteSessionExists(\n exe: ExeProvider,\n vmName: string,\n sessionName: string\n): Promise<boolean> {\n const result = await exe.ssh(vmName, `tmux has-session -t ${sessionName} 2>/dev/null && echo exists || echo not-found`);\n return result.stdout.trim() === 'exists';\n}\n\nexport interface SpawnRemoteAgentOptions {\n issueId: string;\n workspace: RemoteWorkspaceMetadata;\n model?: string;\n prompt?: string;\n phase?: string;\n}\n\n/**\n * Spawn a Claude agent on a remote VM\n */\nexport async function spawnRemoteAgent(options: SpawnRemoteAgentOptions): Promise<RemoteAgentState> {\n const { issueId, workspace, model = 'claude-sonnet-4-6', prompt } = options;\n\n const agentId = `agent-${issueId.toLowerCase()}`;\n const vmName = workspace.vmName;\n\n const exe = createExeProvider({ infraVm: workspace.infraVm });\n\n // Check if VM is running\n const vmStatus = await exe.getStatus(vmName);\n if (vmStatus !== 'running') {\n throw new Error(`VM ${vmName} is not running. Start it with: pan workspace start ${issueId}`);\n }\n\n // Check if agent already exists\n if (await remoteSessionExists(exe, vmName, agentId)) {\n throw new Error(`Agent ${agentId} already running on ${vmName}. Use 'pan work tell' to message it.`);\n }\n\n // Create agent state\n const state: RemoteAgentState = {\n id: agentId,\n issueId,\n vmName,\n model,\n status: 'starting',\n startedAt: new Date().toISOString(),\n location: 'remote',\n };\n\n saveRemoteAgentState(state);\n\n // Write prompt to file on remote VM if provided\n let claudeCmd: string;\n\n if (prompt) {\n // Write prompt to file on VM using base64 to avoid escaping issues\n const promptFile = `/workspace/.panopticon/prompts/${agentId}.md`;\n await exe.ssh(vmName, `mkdir -p /workspace/.panopticon/prompts`);\n const promptBase64 = Buffer.from(prompt).toString('base64');\n await exe.ssh(vmName, `echo '${promptBase64}' | base64 -d > ${promptFile}`);\n\n // Create launcher script using base64 to avoid shell interpretation\n const launcherScript = `/workspace/.panopticon/prompts/${agentId}-launcher.sh`;\n const launcherContent = `#!/bin/bash\nexport PATH=\"/usr/local/bin:\\$PATH\"\nprompt=\\$(cat \"${promptFile}\")\nexec claude --dangerously-skip-permissions --model ${model} \"\\$prompt\"\n`;\n const launcherBase64 = Buffer.from(launcherContent).toString('base64');\n await exe.ssh(vmName, `echo '${launcherBase64}' | base64 -d > ${launcherScript} && chmod +x ${launcherScript}`);\n\n claudeCmd = `bash ${launcherScript}`;\n } else {\n claudeCmd = `claude --dangerously-skip-permissions --model ${model}`;\n }\n\n // Create tmux session on remote VM\n const tmuxCmd = `tmux new-session -d -s ${agentId} -c /workspace '${claudeCmd}'`;\n const result = await exe.ssh(vmName, tmuxCmd);\n\n if (result.exitCode !== 0) {\n state.status = 'error';\n saveRemoteAgentState(state);\n throw new Error(`Failed to start agent: ${result.stderr}`);\n }\n\n // Update status\n state.status = 'running';\n saveRemoteAgentState(state);\n\n return state;\n}\n\n/**\n * Get remote agent output from tmux session\n */\nexport async function getRemoteAgentOutput(\n agentId: string,\n vmName: string,\n lines: number = 100\n): Promise<string> {\n const exe = createExeProvider();\n\n const result = await exe.ssh(vmName, `tmux capture-pane -t ${agentId} -p -S -${lines}`);\n return result.stdout;\n}\n\n/**\n * Send message to remote agent\n */\nexport async function sendToRemoteAgent(\n agentId: string,\n vmName: string,\n message: string\n): Promise<void> {\n const exe = createExeProvider();\n\n // Escape message for shell\n const escapedMessage = message.replace(/'/g, \"'\\\\''\");\n\n // Send keys to tmux session (send message then Enter)\n await exe.ssh(vmName, `tmux send-keys -t ${agentId} '${escapedMessage}'`);\n await exe.ssh(vmName, `tmux send-keys -t ${agentId} C-m`);\n}\n\n/**\n * Check if remote agent is still running\n */\nexport async function isRemoteAgentRunning(\n agentId: string,\n vmName: string\n): Promise<boolean> {\n const exe = createExeProvider();\n return remoteSessionExists(exe, vmName, agentId);\n}\n\n/**\n * Kill remote agent session\n */\nexport async function killRemoteAgent(\n agentId: string,\n vmName: string\n): Promise<void> {\n const exe = createExeProvider();\n await exe.ssh(vmName, `tmux kill-session -t ${agentId} 2>/dev/null || true`);\n\n // Update state\n const state = loadRemoteAgentState(agentId);\n if (state) {\n state.status = 'stopped';\n saveRemoteAgentState(state);\n }\n}\n\n/**\n * Get list of running remote agents on a VM\n */\nexport async function listRemoteAgents(vmName: string): Promise<string[]> {\n const exe = createExeProvider();\n\n const result = await exe.ssh(vmName, `tmux list-sessions -F \"#{session_name}\" 2>/dev/null | grep \"^agent-\" || true`);\n if (!result.stdout.trim()) {\n return [];\n }\n\n return result.stdout.trim().split('\\n').filter(Boolean);\n}\n\n/**\n * Poll remote agent for status updates\n * Returns parsed events from the agent output\n */\nexport async function pollRemoteAgentStatus(\n agentId: string,\n vmName: string\n): Promise<{\n isRunning: boolean;\n lastOutput: string;\n toolUses: string[];\n}> {\n const exe = createExeProvider();\n\n // Check if session exists\n const isRunning = await remoteSessionExists(exe, vmName, agentId);\n\n if (!isRunning) {\n return { isRunning: false, lastOutput: '', toolUses: [] };\n }\n\n // Get recent output\n const output = await getRemoteAgentOutput(agentId, vmName, 50);\n\n // Parse tool uses from output (simple pattern matching)\n const toolUses: string[] = [];\n const toolPattern = /(?:Using|Calling|Running)\\s+(\\w+)\\s+tool/gi;\n let match;\n while ((match = toolPattern.exec(output)) !== null) {\n toolUses.push(match[1]);\n }\n\n return {\n isRunning,\n lastOutput: output,\n toolUses,\n };\n}\n"],"mappings":";;;;;;;;;;AASA,SAAS,YAAY;AACrB,SAAS,YAAY,cAAc,eAAe,iBAAiB;AACnE,SAAS,eAAe;AAkBxB,SAAS,wBAAwB,SAAyB;AACxD,SAAO,KAAK,YAAY,SAAS,mBAAmB;AACtD;AAKA,SAAS,qBAAqB,OAA+B;AAC3D,QAAM,MAAM,KAAK,YAAY,MAAM,EAAE;AACrC,MAAI,CAAC,WAAW,GAAG,GAAG;AACpB,cAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,EACpC;AACA,gBAAc,wBAAwB,MAAM,EAAE,GAAG,KAAK,UAAU,OAAO,MAAM,CAAC,CAAC;AACjF;AAKO,SAAS,qBAAqB,SAA0C;AAC7E,QAAM,OAAO,wBAAwB,OAAO;AAC5C,MAAI,CAAC,WAAW,IAAI,EAAG,QAAO;AAE9B,MAAI;AACF,WAAO,KAAK,MAAM,aAAa,MAAM,OAAO,CAAC;AAAA,EAC/C,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAKA,eAAe,oBACb,KACA,QACA,aACkB;AAClB,QAAM,SAAS,MAAM,IAAI,IAAI,QAAQ,uBAAuB,WAAW,+CAA+C;AACtH,SAAO,OAAO,OAAO,KAAK,MAAM;AAClC;AAaA,eAAsB,iBAAiB,SAA6D;AAClG,QAAM,EAAE,SAAS,WAAW,QAAQ,qBAAqB,OAAO,IAAI;AAEpE,QAAM,UAAU,SAAS,QAAQ,YAAY,CAAC;AAC9C,QAAM,SAAS,UAAU;AAEzB,QAAM,MAAM,kBAAkB,EAAE,SAAS,UAAU,QAAQ,CAAC;AAG5D,QAAM,WAAW,MAAM,IAAI,UAAU,MAAM;AAC3C,MAAI,aAAa,WAAW;AAC1B,UAAM,IAAI,MAAM,MAAM,MAAM,uDAAuD,OAAO,EAAE;AAAA,EAC9F;AAGA,MAAI,MAAM,oBAAoB,KAAK,QAAQ,OAAO,GAAG;AACnD,UAAM,IAAI,MAAM,SAAS,OAAO,uBAAuB,MAAM,sCAAsC;AAAA,EACrG;AAGA,QAAM,QAA0B;AAAA,IAC9B,IAAI;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ;AAAA,IACR,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IAClC,UAAU;AAAA,EACZ;AAEA,uBAAqB,KAAK;AAG1B,MAAI;AAEJ,MAAI,QAAQ;AAEV,UAAM,aAAa,kCAAkC,OAAO;AAC5D,UAAM,IAAI,IAAI,QAAQ,yCAAyC;AAC/D,UAAM,eAAe,OAAO,KAAK,MAAM,EAAE,SAAS,QAAQ;AAC1D,UAAM,IAAI,IAAI,QAAQ,SAAS,YAAY,mBAAmB,UAAU,EAAE;AAG1E,UAAM,iBAAiB,kCAAkC,OAAO;AAChE,UAAM,kBAAkB;AAAA;AAAA,gBAEX,UAAU;AAAA,qDAC0B,KAAK;AAAA;AAEtD,UAAM,iBAAiB,OAAO,KAAK,eAAe,EAAE,SAAS,QAAQ;AACrE,UAAM,IAAI,IAAI,QAAQ,SAAS,cAAc,mBAAmB,cAAc,gBAAgB,cAAc,EAAE;AAE9G,gBAAY,QAAQ,cAAc;AAAA,EACpC,OAAO;AACL,gBAAY,iDAAiD,KAAK;AAAA,EACpE;AAGA,QAAM,UAAU,0BAA0B,OAAO,mBAAmB,SAAS;AAC7E,QAAM,SAAS,MAAM,IAAI,IAAI,QAAQ,OAAO;AAE5C,MAAI,OAAO,aAAa,GAAG;AACzB,UAAM,SAAS;AACf,yBAAqB,KAAK;AAC1B,UAAM,IAAI,MAAM,0BAA0B,OAAO,MAAM,EAAE;AAAA,EAC3D;AAGA,QAAM,SAAS;AACf,uBAAqB,KAAK;AAE1B,SAAO;AACT;AAKA,eAAsB,qBACpB,SACA,QACA,QAAgB,KACC;AACjB,QAAM,MAAM,kBAAkB;AAE9B,QAAM,SAAS,MAAM,IAAI,IAAI,QAAQ,wBAAwB,OAAO,WAAW,KAAK,EAAE;AACtF,SAAO,OAAO;AAChB;AAKA,eAAsB,kBACpB,SACA,QACA,SACe;AACf,QAAM,MAAM,kBAAkB;AAG9B,QAAM,iBAAiB,QAAQ,QAAQ,MAAM,OAAO;AAGpD,QAAM,IAAI,IAAI,QAAQ,qBAAqB,OAAO,KAAK,cAAc,GAAG;AACxE,QAAM,IAAI,IAAI,QAAQ,qBAAqB,OAAO,MAAM;AAC1D;AAKA,eAAsB,qBACpB,SACA,QACkB;AAClB,QAAM,MAAM,kBAAkB;AAC9B,SAAO,oBAAoB,KAAK,QAAQ,OAAO;AACjD;AAKA,eAAsB,gBACpB,SACA,QACe;AACf,QAAM,MAAM,kBAAkB;AAC9B,QAAM,IAAI,IAAI,QAAQ,wBAAwB,OAAO,sBAAsB;AAG3E,QAAM,QAAQ,qBAAqB,OAAO;AAC1C,MAAI,OAAO;AACT,UAAM,SAAS;AACf,yBAAqB,KAAK;AAAA,EAC5B;AACF;AAKA,eAAsB,iBAAiB,QAAmC;AACxE,QAAM,MAAM,kBAAkB;AAE9B,QAAM,SAAS,MAAM,IAAI,IAAI,QAAQ,8EAA8E;AACnH,MAAI,CAAC,OAAO,OAAO,KAAK,GAAG;AACzB,WAAO,CAAC;AAAA,EACV;AAEA,SAAO,OAAO,OAAO,KAAK,EAAE,MAAM,IAAI,EAAE,OAAO,OAAO;AACxD;AAMA,eAAsB,sBACpB,SACA,QAKC;AACD,QAAM,MAAM,kBAAkB;AAG9B,QAAM,YAAY,MAAM,oBAAoB,KAAK,QAAQ,OAAO;AAEhE,MAAI,CAAC,WAAW;AACd,WAAO,EAAE,WAAW,OAAO,YAAY,IAAI,UAAU,CAAC,EAAE;AAAA,EAC1D;AAGA,QAAM,SAAS,MAAM,qBAAqB,SAAS,QAAQ,EAAE;AAG7D,QAAM,WAAqB,CAAC;AAC5B,QAAM,cAAc;AACpB,MAAI;AACJ,UAAQ,QAAQ,YAAY,KAAK,MAAM,OAAO,MAAM;AAClD,aAAS,KAAK,MAAM,CAAC,CAAC;AAAA,EACxB;AAEA,SAAO;AAAA,IACL;AAAA,IACA,YAAY;AAAA,IACZ;AAAA,EACF;AACF;AA3QA,IAaM;AAbN;AAAA;AAAA;AAOA;AAMA,IAAM,aAAa,KAAK,QAAQ,GAAG,eAAe,QAAQ;AAAA;AAAA;","names":[]}
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/lib/database/schema.ts","../src/lib/database/index.ts","../src/lib/database/review-status-db.ts","../src/lib/review-status.ts"],"sourcesContent":["/**\n * Panopticon Database Schema\n *\n * Defines the unified schema for panopticon.db.\n * All persistent application state lives here.\n */\n\nimport type Database from 'better-sqlite3';\n\n// Schema version — increment when making breaking schema changes\nexport const SCHEMA_VERSION = 2;\n\n/**\n * Initialize the complete database schema.\n * Idempotent — uses CREATE TABLE IF NOT EXISTS throughout.\n */\nexport function initSchema(db: Database.Database): void {\n db.exec(`\n -- ===== Cost Events =====\n CREATE TABLE IF NOT EXISTS cost_events (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n ts TEXT NOT NULL,\n agent_id TEXT NOT NULL,\n issue_id TEXT NOT NULL,\n session_type TEXT NOT NULL DEFAULT 'unknown',\n provider TEXT NOT NULL DEFAULT 'anthropic',\n model TEXT NOT NULL,\n input INTEGER NOT NULL DEFAULT 0,\n output INTEGER NOT NULL DEFAULT 0,\n cache_read INTEGER NOT NULL DEFAULT 0,\n cache_write INTEGER NOT NULL DEFAULT 0,\n cost REAL NOT NULL DEFAULT 0,\n request_id TEXT,\n -- TLDR metrics\n tldr_interceptions INTEGER,\n tldr_bypasses INTEGER,\n tldr_tokens_saved INTEGER,\n tldr_bypass_reasons TEXT, -- JSON string\n -- WAL source tracking\n source_file TEXT -- path of WAL file this came from (for imports)\n );\n\n CREATE UNIQUE INDEX IF NOT EXISTS idx_cost_request_id\n ON cost_events(request_id) WHERE request_id IS NOT NULL;\n\n CREATE INDEX IF NOT EXISTS idx_cost_issue_id\n ON cost_events(issue_id, ts);\n\n CREATE INDEX IF NOT EXISTS idx_cost_agent_id\n ON cost_events(agent_id, ts);\n\n CREATE INDEX IF NOT EXISTS idx_cost_ts\n ON cost_events(ts);\n\n -- ===== Review Status =====\n CREATE TABLE IF NOT EXISTS review_status (\n issue_id TEXT PRIMARY KEY,\n review_status TEXT NOT NULL DEFAULT 'pending',\n test_status TEXT NOT NULL DEFAULT 'pending',\n merge_status TEXT,\n verification_status TEXT,\n verification_notes TEXT,\n verification_cycle_count INTEGER DEFAULT 0,\n verification_max_cycles INTEGER,\n review_notes TEXT,\n test_notes TEXT,\n merge_notes TEXT,\n updated_at TEXT NOT NULL,\n ready_for_merge INTEGER NOT NULL DEFAULT 0,\n auto_requeue_count INTEGER DEFAULT 0,\n pr_url TEXT\n );\n\n CREATE INDEX IF NOT EXISTS idx_review_status_updated\n ON review_status(updated_at);\n\n -- ===== Status History =====\n CREATE TABLE IF NOT EXISTS status_history (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n issue_id TEXT NOT NULL,\n type TEXT NOT NULL, -- 'review', 'test', 'merge'\n status TEXT NOT NULL,\n timestamp TEXT NOT NULL,\n notes TEXT,\n FOREIGN KEY (issue_id) REFERENCES review_status(issue_id) ON DELETE CASCADE\n );\n\n CREATE INDEX IF NOT EXISTS idx_status_history_issue\n ON status_history(issue_id, timestamp);\n\n -- UNIQUE constraint enables INSERT OR IGNORE deduplication in upsertReviewStatus\n CREATE UNIQUE INDEX IF NOT EXISTS idx_status_history_unique\n ON status_history(issue_id, type, status, timestamp);\n\n -- ===== Health Events =====\n CREATE TABLE IF NOT EXISTS health_events (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n agent_id TEXT NOT NULL,\n timestamp TEXT NOT NULL,\n state TEXT NOT NULL,\n previous_state TEXT,\n source TEXT,\n metadata TEXT -- JSON string\n );\n\n CREATE INDEX IF NOT EXISTS idx_health_agent_timestamp\n ON health_events(agent_id, timestamp);\n\n CREATE INDEX IF NOT EXISTS idx_health_timestamp\n ON health_events(timestamp);\n\n -- ===== Processed Sessions (for cost migration dedup) =====\n CREATE TABLE IF NOT EXISTS processed_sessions (\n session_id TEXT PRIMARY KEY,\n processed_at TEXT NOT NULL,\n event_count INTEGER NOT NULL DEFAULT 0\n );\n\n -- ===== API Cache =====\n CREATE TABLE IF NOT EXISTS api_cache (\n key TEXT PRIMARY KEY,\n value TEXT NOT NULL, -- JSON string\n expires_at TEXT,\n created_at TEXT NOT NULL\n );\n\n -- ===== Rate Limits =====\n CREATE TABLE IF NOT EXISTS rate_limits (\n service TEXT PRIMARY KEY,\n requests INTEGER NOT NULL DEFAULT 0,\n window_start TEXT NOT NULL,\n limit_per_window INTEGER NOT NULL DEFAULT 1000\n );\n `);\n\n // Record schema version\n db.pragma(`user_version = ${SCHEMA_VERSION}`);\n}\n\n/**\n * Run schema migrations if the database version is older than SCHEMA_VERSION.\n * This function handles upgrading from older schema versions.\n */\nexport function runMigrations(db: Database.Database): void {\n const currentVersion = db.pragma('user_version', { simple: true }) as number;\n\n if (currentVersion === SCHEMA_VERSION) {\n return; // Already at latest version\n }\n\n if (currentVersion === 0) {\n // Fresh database — just initialize the full schema\n initSchema(db);\n return;\n }\n\n // v1 → v2: add UNIQUE index on status_history for INSERT OR IGNORE dedup\n if (currentVersion < 2) {\n // Remove duplicate rows before adding the unique index (keep lowest id per unique key)\n db.exec(`\n DELETE FROM status_history\n WHERE id NOT IN (\n SELECT MIN(id)\n FROM status_history\n GROUP BY issue_id, type, status, timestamp\n );\n CREATE UNIQUE INDEX IF NOT EXISTS idx_status_history_unique\n ON status_history(issue_id, type, status, timestamp);\n `);\n }\n\n // After all migrations, set the version\n db.pragma(`user_version = ${SCHEMA_VERSION}`);\n}\n","/**\n * Panopticon Unified Database\n *\n * Single panopticon.db at ~/.panopticon/panopticon.db.\n * Singleton pattern — one connection shared across the process.\n *\n * IMPORTANT: This module is safe to import in both server and CLI contexts.\n * Never use execSync here — this is synchronous SQLite, not a subprocess.\n */\n\nimport Database from 'better-sqlite3';\nimport { join } from 'path';\nimport { existsSync, mkdirSync } from 'fs';\nimport { getPanopticonHome } from '../paths.js';\nimport { runMigrations } from './schema.js';\n\nlet _db: Database.Database | null = null;\n\n/**\n * Get the path to panopticon.db (dynamic, respects PANOPTICON_HOME override for tests)\n */\nexport function getDatabasePath(): string {\n return join(getPanopticonHome(), 'panopticon.db');\n}\n\n/**\n * Initialize and return the singleton database connection.\n * Safe to call multiple times — returns the existing connection after first call.\n */\nexport function getDatabase(): Database.Database {\n if (_db) {\n return _db;\n }\n\n const home = getPanopticonHome();\n if (!existsSync(home)) {\n mkdirSync(home, { recursive: true });\n }\n\n const dbPath = getDatabasePath();\n _db = new Database(dbPath);\n\n // Enable WAL mode for concurrent readers + single writer\n _db.pragma('journal_mode = WAL');\n // Enforce foreign keys\n _db.pragma('foreign_keys = ON');\n // Write-ahead log synchronization — NORMAL is safe and fast\n _db.pragma('synchronous = NORMAL');\n\n // Initialize or migrate schema\n runMigrations(_db);\n\n return _db;\n}\n\n/**\n * Close the database connection and release the singleton.\n * Primarily used in tests to get a fresh connection.\n */\nexport function closeDatabase(): void {\n if (_db) {\n _db.close();\n _db = null;\n }\n}\n\n/**\n * Force re-initialization of the database connection.\n * Used in tests after PANOPTICON_HOME changes.\n */\nexport function resetDatabase(): void {\n closeDatabase();\n}\n","/**\n * Review Status SQLite Storage\n *\n * Provides SQLite-backed CRUD for ReviewStatus, matching the interface in\n * src/lib/review-status.ts. Atomic single-transaction writes eliminate the\n * TOCTOU race in the JSON-backed implementation.\n */\n\nimport { getDatabase } from './index.js';\nimport type { ReviewStatus, StatusHistoryEntry } from '../review-status.js';\n\n// ============== Write operations ==============\n\n/**\n * Upsert a review status record atomically.\n * Replaces the JSON read-modify-write cycle with a single transaction.\n */\nexport function upsertReviewStatus(status: ReviewStatus): void {\n const db = getDatabase();\n\n const upsert = db.transaction((s: ReviewStatus) => {\n // Upsert main record\n db.prepare(`\n INSERT INTO review_status (\n issue_id, review_status, test_status, merge_status,\n verification_status, verification_notes,\n verification_cycle_count, verification_max_cycles,\n review_notes, test_notes, merge_notes,\n updated_at, ready_for_merge, auto_requeue_count, pr_url\n ) VALUES (\n ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?\n )\n ON CONFLICT(issue_id) DO UPDATE SET\n review_status = excluded.review_status,\n test_status = excluded.test_status,\n merge_status = excluded.merge_status,\n verification_status = excluded.verification_status,\n verification_notes = excluded.verification_notes,\n verification_cycle_count = excluded.verification_cycle_count,\n verification_max_cycles = excluded.verification_max_cycles,\n review_notes = excluded.review_notes,\n test_notes = excluded.test_notes,\n merge_notes = excluded.merge_notes,\n updated_at = excluded.updated_at,\n ready_for_merge = excluded.ready_for_merge,\n auto_requeue_count = excluded.auto_requeue_count,\n pr_url = excluded.pr_url\n `).run(\n s.issueId,\n s.reviewStatus,\n s.testStatus,\n s.mergeStatus ?? null,\n s.verificationStatus ?? null,\n s.verificationNotes ?? null,\n s.verificationCycleCount ?? null,\n s.verificationMaxCycles ?? null,\n s.reviewNotes ?? null,\n s.testNotes ?? null,\n s.mergeNotes ?? null,\n s.updatedAt,\n s.readyForMerge ? 1 : 0,\n s.autoRequeueCount ?? null,\n s.prUrl ?? null,\n );\n\n // Append new history entries (deduplicate by timestamp to avoid re-inserting)\n if (s.history && s.history.length > 0) {\n const insertHistory = db.prepare(`\n INSERT OR IGNORE INTO status_history (issue_id, type, status, timestamp, notes)\n VALUES (?, ?, ?, ?, ?)\n `);\n for (const entry of s.history) {\n insertHistory.run(s.issueId, entry.type, entry.status, entry.timestamp, entry.notes ?? null);\n }\n }\n });\n\n upsert(status);\n}\n\n/**\n * Delete a review status record and its history.\n */\nexport function deleteReviewStatus(issueId: string): void {\n const db = getDatabase();\n db.prepare('DELETE FROM review_status WHERE issue_id = ?').run(issueId);\n}\n\n// ============== Read operations ==============\n\n/**\n * Get a single review status by issue ID.\n */\nexport function getReviewStatusFromDb(issueId: string): ReviewStatus | null {\n const db = getDatabase();\n\n const row = db.prepare(`\n SELECT * FROM review_status WHERE issue_id = ?\n `).get(issueId) as DbReviewStatusRow | undefined;\n\n if (!row) return null;\n\n const history = getHistoryFromDb(issueId);\n return rowToReviewStatus(row, history);\n}\n\n/**\n * Get all review statuses.\n */\nexport function getAllReviewStatusesFromDb(): Record<string, ReviewStatus> {\n const db = getDatabase();\n\n const rows = db.prepare('SELECT * FROM review_status ORDER BY updated_at DESC').all() as DbReviewStatusRow[];\n const result: Record<string, ReviewStatus> = {};\n\n for (const row of rows) {\n const history = getHistoryFromDb(row.issue_id);\n result[row.issue_id] = rowToReviewStatus(row, history);\n }\n\n return result;\n}\n\n/**\n * Get history entries for an issue.\n */\nfunction getHistoryFromDb(issueId: string): StatusHistoryEntry[] {\n const db = getDatabase();\n const rows = db.prepare(`\n SELECT type, status, timestamp, notes\n FROM status_history\n WHERE issue_id = ?\n ORDER BY timestamp ASC\n `).all(issueId) as Array<{ type: string; status: string; timestamp: string; notes: string | null }>;\n\n return rows.map(r => ({\n type: r.type as 'review' | 'test' | 'merge',\n status: r.status,\n timestamp: r.timestamp,\n ...(r.notes ? { notes: r.notes } : {}),\n }));\n}\n\n// ============== Row mapping ==============\n\ninterface DbReviewStatusRow {\n issue_id: string;\n review_status: string;\n test_status: string;\n merge_status: string | null;\n verification_status: string | null;\n verification_notes: string | null;\n verification_cycle_count: number | null;\n verification_max_cycles: number | null;\n review_notes: string | null;\n test_notes: string | null;\n merge_notes: string | null;\n updated_at: string;\n ready_for_merge: number;\n auto_requeue_count: number | null;\n pr_url: string | null;\n}\n\nfunction rowToReviewStatus(row: DbReviewStatusRow, history: StatusHistoryEntry[]): ReviewStatus {\n return {\n issueId: row.issue_id,\n reviewStatus: row.review_status as ReviewStatus['reviewStatus'],\n testStatus: row.test_status as ReviewStatus['testStatus'],\n mergeStatus: row.merge_status as ReviewStatus['mergeStatus'] ?? undefined,\n verificationStatus: row.verification_status as ReviewStatus['verificationStatus'] ?? undefined,\n verificationNotes: row.verification_notes ?? undefined,\n verificationCycleCount: row.verification_cycle_count ?? undefined,\n verificationMaxCycles: row.verification_max_cycles ?? undefined,\n reviewNotes: row.review_notes ?? undefined,\n testNotes: row.test_notes ?? undefined,\n mergeNotes: row.merge_notes ?? undefined,\n updatedAt: row.updated_at,\n readyForMerge: row.ready_for_merge === 1,\n autoRequeueCount: row.auto_requeue_count ?? undefined,\n prUrl: row.pr_url ?? undefined,\n history: history.length > 0 ? history : undefined,\n };\n}\n","import { existsSync, readFileSync, writeFileSync, mkdirSync } from 'fs';\nimport { join, dirname } from 'path';\nimport { homedir } from 'os';\nimport { notifyPipeline } from './pipeline-notifier.js';\nimport {\n upsertReviewStatus as dbUpsert,\n deleteReviewStatus as dbDelete,\n getReviewStatusFromDb,\n getAllReviewStatusesFromDb,\n} from './database/review-status-db.js';\n\nexport interface StatusHistoryEntry {\n type: 'review' | 'test' | 'merge';\n status: string;\n timestamp: string;\n notes?: string;\n}\n\nexport interface ReviewStatus {\n issueId: string;\n reviewStatus: 'pending' | 'reviewing' | 'passed' | 'failed' | 'blocked';\n testStatus: 'pending' | 'testing' | 'passed' | 'failed' | 'skipped';\n mergeStatus?: 'pending' | 'merging' | 'merged' | 'failed';\n verificationStatus?: 'pending' | 'running' | 'passed' | 'failed' | 'skipped';\n verificationNotes?: string;\n verificationCycleCount?: number;\n verificationMaxCycles?: number;\n reviewNotes?: string;\n testNotes?: string;\n mergeNotes?: string;\n updatedAt: string;\n readyForMerge: boolean;\n autoRequeueCount?: number;\n prUrl?: string;\n history?: StatusHistoryEntry[];\n}\n\nconst DEFAULT_STATUS_FILE = join(homedir(), '.panopticon', 'review-status.json');\n\nexport function loadReviewStatuses(filePath = DEFAULT_STATUS_FILE): Record<string, ReviewStatus> {\n // Prefer SQLite when using the default path\n if (filePath === DEFAULT_STATUS_FILE) {\n try {\n return getAllReviewStatusesFromDb();\n } catch {\n // Fall through to JSON on DB error\n }\n }\n\n try {\n if (existsSync(filePath)) {\n return JSON.parse(readFileSync(filePath, 'utf-8'));\n }\n } catch (err) {\n console.error('Failed to load review statuses:', err);\n }\n return {};\n}\n\nexport function saveReviewStatuses(statuses: Record<string, ReviewStatus>, filePath = DEFAULT_STATUS_FILE): void {\n try {\n const dir = dirname(filePath);\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n writeFileSync(filePath, JSON.stringify(statuses, null, 2));\n } catch (err) {\n console.error('Failed to save review statuses:', err);\n }\n}\n\nexport function setReviewStatus(\n issueId: string,\n update: Partial<ReviewStatus>,\n filePath = DEFAULT_STATUS_FILE,\n): ReviewStatus {\n const statuses = loadReviewStatuses(filePath);\n const existing = statuses[issueId] || {\n issueId,\n reviewStatus: 'pending' as const,\n testStatus: 'pending' as const,\n updatedAt: new Date().toISOString(),\n readyForMerge: false,\n };\n\n // Guard: reject reviewStatus regression from 'passed' to 'reviewing' unless the caller\n // is explicitly resetting the merge lifecycle (update includes mergeStatus).\n // This is belt-and-suspenders — endpoint-level guards should catch this first.\n if (update.reviewStatus === 'reviewing' && existing.reviewStatus === 'passed' && update.mergeStatus === undefined) {\n console.warn(`[review-status] Rejecting reviewStatus regression from 'passed' to 'reviewing' for ${issueId} (mergeStatus not being reset)`);\n return existing as ReviewStatus;\n }\n\n const merged = { ...existing, ...update };\n\n // Track status transitions in history (last 10 entries)\n const history = [...(existing.history || [])];\n const now = new Date().toISOString();\n if (update.reviewStatus && update.reviewStatus !== existing.reviewStatus) {\n history.push({ type: 'review', status: update.reviewStatus, timestamp: now, notes: update.reviewNotes });\n }\n if (update.testStatus && update.testStatus !== existing.testStatus) {\n history.push({ type: 'test', status: update.testStatus, timestamp: now, notes: update.testNotes });\n }\n if (update.mergeStatus && update.mergeStatus !== existing.mergeStatus) {\n history.push({ type: 'merge', status: update.mergeStatus, timestamp: now });\n }\n while (history.length > 10) history.shift();\n\n const readyForMerge = update.readyForMerge !== undefined\n ? update.readyForMerge\n : (merged.reviewStatus === 'passed' && merged.testStatus === 'passed' && merged.mergeStatus !== 'merged');\n\n const updated: ReviewStatus = {\n ...merged,\n issueId,\n updatedAt: now,\n readyForMerge,\n history,\n };\n\n // SQLite first — it is the authoritative store (reads prefer SQLite)\n if (filePath === DEFAULT_STATUS_FILE) {\n try {\n dbUpsert(updated);\n } catch (err) {\n console.error('[review-status] SQLite write failed (continuing with JSON):', err);\n }\n }\n\n // JSON second — legacy fallback for tools that read review-status.json directly\n statuses[issueId] = updated;\n saveReviewStatuses(statuses, filePath);\n\n notifyPipeline({ type: 'status_changed', issueId, status: updated });\n\n return updated;\n}\n\nexport function getReviewStatus(issueId: string, filePath = DEFAULT_STATUS_FILE): ReviewStatus | null {\n // Prefer SQLite when using the default path\n if (filePath === DEFAULT_STATUS_FILE) {\n try {\n const fromDb = getReviewStatusFromDb(issueId);\n if (fromDb) return fromDb;\n } catch {\n // Fall through to JSON on DB error\n }\n }\n const statuses = loadReviewStatuses(filePath);\n return statuses[issueId] || null;\n}\n\nexport function clearReviewStatus(issueId: string, filePath = DEFAULT_STATUS_FILE): void {\n const statuses = loadReviewStatuses(filePath);\n delete statuses[issueId];\n saveReviewStatuses(statuses, filePath);\n\n // Dual-delete from SQLite when using the default path\n if (filePath === DEFAULT_STATUS_FILE) {\n try {\n dbDelete(issueId);\n } catch (err) {\n console.error('[review-status] SQLite delete failed (continuing with JSON):', err);\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;AAgBO,SAAS,WAAW,IAA6B;AACtD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAoHP;AAGD,KAAG,OAAO,kBAAkB,cAAc,EAAE;AAC9C;AAMO,SAAS,cAAc,IAA6B;AACzD,QAAM,iBAAiB,GAAG,OAAO,gBAAgB,EAAE,QAAQ,KAAK,CAAC;AAEjE,MAAI,mBAAmB,gBAAgB;AACrC;AAAA,EACF;AAEA,MAAI,mBAAmB,GAAG;AAExB,eAAW,EAAE;AACb;AAAA,EACF;AAGA,MAAI,iBAAiB,GAAG;AAEtB,OAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KASP;AAAA,EACH;AAGA,KAAG,OAAO,kBAAkB,cAAc,EAAE;AAC9C;AA7KA,IAUa;AAVb;AAAA;AAAA;AAAA;AAUO,IAAM,iBAAiB;AAAA;AAAA;;;ACA9B,OAAO,cAAc;AACrB,SAAS,YAAY;AACrB,SAAS,YAAY,iBAAiB;AAS/B,SAAS,kBAA0B;AACxC,SAAO,KAAK,kBAAkB,GAAG,eAAe;AAClD;AAMO,SAAS,cAAiC;AAC/C,MAAI,KAAK;AACP,WAAO;AAAA,EACT;AAEA,QAAM,OAAO,kBAAkB;AAC/B,MAAI,CAAC,WAAW,IAAI,GAAG;AACrB,cAAU,MAAM,EAAE,WAAW,KAAK,CAAC;AAAA,EACrC;AAEA,QAAM,SAAS,gBAAgB;AAC/B,QAAM,IAAI,SAAS,MAAM;AAGzB,MAAI,OAAO,oBAAoB;AAE/B,MAAI,OAAO,mBAAmB;AAE9B,MAAI,OAAO,sBAAsB;AAGjC,gBAAc,GAAG;AAEjB,SAAO;AACT;AAMO,SAAS,gBAAsB;AACpC,MAAI,KAAK;AACP,QAAI,MAAM;AACV,UAAM;AAAA,EACR;AACF;AAhEA,IAgBI;AAhBJ;AAAA;AAAA;AAAA;AAaA;AACA;AAEA,IAAI,MAAgC;AAAA;AAAA;;;ACC7B,SAAS,mBAAmB,QAA4B;AAC7D,QAAM,KAAK,YAAY;AAEvB,QAAM,SAAS,GAAG,YAAY,CAAC,MAAoB;AAEjD,OAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAyBV,EAAE;AAAA,MACD,EAAE;AAAA,MACF,EAAE;AAAA,MACF,EAAE;AAAA,MACF,EAAE,eAAe;AAAA,MACjB,EAAE,sBAAsB;AAAA,MACxB,EAAE,qBAAqB;AAAA,MACvB,EAAE,0BAA0B;AAAA,MAC5B,EAAE,yBAAyB;AAAA,MAC3B,EAAE,eAAe;AAAA,MACjB,EAAE,aAAa;AAAA,MACf,EAAE,cAAc;AAAA,MAChB,EAAE;AAAA,MACF,EAAE,gBAAgB,IAAI;AAAA,MACtB,EAAE,oBAAoB;AAAA,MACtB,EAAE,SAAS;AAAA,IACb;AAGA,QAAI,EAAE,WAAW,EAAE,QAAQ,SAAS,GAAG;AACrC,YAAM,gBAAgB,GAAG,QAAQ;AAAA;AAAA;AAAA,OAGhC;AACD,iBAAW,SAAS,EAAE,SAAS;AAC7B,sBAAc,IAAI,EAAE,SAAS,MAAM,MAAM,MAAM,QAAQ,MAAM,WAAW,MAAM,SAAS,IAAI;AAAA,MAC7F;AAAA,IACF;AAAA,EACF,CAAC;AAED,SAAO,MAAM;AACf;AAKO,SAAS,mBAAmB,SAAuB;AACxD,QAAM,KAAK,YAAY;AACvB,KAAG,QAAQ,8CAA8C,EAAE,IAAI,OAAO;AACxE;AAOO,SAAS,sBAAsB,SAAsC;AAC1E,QAAM,KAAK,YAAY;AAEvB,QAAM,MAAM,GAAG,QAAQ;AAAA;AAAA,GAEtB,EAAE,IAAI,OAAO;AAEd,MAAI,CAAC,IAAK,QAAO;AAEjB,QAAM,UAAU,iBAAiB,OAAO;AACxC,SAAO,kBAAkB,KAAK,OAAO;AACvC;AAKO,SAAS,6BAA2D;AACzE,QAAM,KAAK,YAAY;AAEvB,QAAM,OAAO,GAAG,QAAQ,sDAAsD,EAAE,IAAI;AACpF,QAAM,SAAuC,CAAC;AAE9C,aAAW,OAAO,MAAM;AACtB,UAAM,UAAU,iBAAiB,IAAI,QAAQ;AAC7C,WAAO,IAAI,QAAQ,IAAI,kBAAkB,KAAK,OAAO;AAAA,EACvD;AAEA,SAAO;AACT;AAKA,SAAS,iBAAiB,SAAuC;AAC/D,QAAM,KAAK,YAAY;AACvB,QAAM,OAAO,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,GAKvB,EAAE,IAAI,OAAO;AAEd,SAAO,KAAK,IAAI,QAAM;AAAA,IACpB,MAAM,EAAE;AAAA,IACR,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,GAAI,EAAE,QAAQ,EAAE,OAAO,EAAE,MAAM,IAAI,CAAC;AAAA,EACtC,EAAE;AACJ;AAsBA,SAAS,kBAAkB,KAAwB,SAA6C;AAC9F,SAAO;AAAA,IACL,SAAS,IAAI;AAAA,IACb,cAAc,IAAI;AAAA,IAClB,YAAY,IAAI;AAAA,IAChB,aAAa,IAAI,gBAA+C;AAAA,IAChE,oBAAoB,IAAI,uBAA6D;AAAA,IACrF,mBAAmB,IAAI,sBAAsB;AAAA,IAC7C,wBAAwB,IAAI,4BAA4B;AAAA,IACxD,uBAAuB,IAAI,2BAA2B;AAAA,IACtD,aAAa,IAAI,gBAAgB;AAAA,IACjC,WAAW,IAAI,cAAc;AAAA,IAC7B,YAAY,IAAI,eAAe;AAAA,IAC/B,WAAW,IAAI;AAAA,IACf,eAAe,IAAI,oBAAoB;AAAA,IACvC,kBAAkB,IAAI,sBAAsB;AAAA,IAC5C,OAAO,IAAI,UAAU;AAAA,IACrB,SAAS,QAAQ,SAAS,IAAI,UAAU;AAAA,EAC1C;AACF;AAtLA;AAAA;AAAA;AAAA;AAQA;AAAA;AAAA;;;ACRA,SAAS,cAAAA,aAAY,cAAc,eAAe,aAAAC,kBAAiB;AACnE,SAAS,QAAAC,OAAM,eAAe;AAC9B,SAAS,eAAe;AAqCjB,SAAS,mBAAmB,WAAW,qBAAmD;AAE/F,MAAI,aAAa,qBAAqB;AACpC,QAAI;AACF,aAAO,2BAA2B;AAAA,IACpC,QAAQ;AAAA,IAER;AAAA,EACF;AAEA,MAAI;AACF,QAAIF,YAAW,QAAQ,GAAG;AACxB,aAAO,KAAK,MAAM,aAAa,UAAU,OAAO,CAAC;AAAA,IACnD;AAAA,EACF,SAAS,KAAK;AACZ,YAAQ,MAAM,mCAAmC,GAAG;AAAA,EACtD;AACA,SAAO,CAAC;AACV;AAEO,SAAS,mBAAmB,UAAwC,WAAW,qBAA2B;AAC/G,MAAI;AACF,UAAM,MAAM,QAAQ,QAAQ;AAC5B,QAAI,CAACA,YAAW,GAAG,GAAG;AACpB,MAAAC,WAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,IACpC;AACA,kBAAc,UAAU,KAAK,UAAU,UAAU,MAAM,CAAC,CAAC;AAAA,EAC3D,SAAS,KAAK;AACZ,YAAQ,MAAM,mCAAmC,GAAG;AAAA,EACtD;AACF;AAEO,SAAS,gBACd,SACA,QACA,WAAW,qBACG;AACd,QAAM,WAAW,mBAAmB,QAAQ;AAC5C,QAAM,WAAW,SAAS,OAAO,KAAK;AAAA,IACpC;AAAA,IACA,cAAc;AAAA,IACd,YAAY;AAAA,IACZ,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IAClC,eAAe;AAAA,EACjB;AAKA,MAAI,OAAO,iBAAiB,eAAe,SAAS,iBAAiB,YAAY,OAAO,gBAAgB,QAAW;AACjH,YAAQ,KAAK,sFAAsF,OAAO,gCAAgC;AAC1I,WAAO;AAAA,EACT;AAEA,QAAM,SAAS,EAAE,GAAG,UAAU,GAAG,OAAO;AAGxC,QAAM,UAAU,CAAC,GAAI,SAAS,WAAW,CAAC,CAAE;AAC5C,QAAM,OAAM,oBAAI,KAAK,GAAE,YAAY;AACnC,MAAI,OAAO,gBAAgB,OAAO,iBAAiB,SAAS,cAAc;AACxE,YAAQ,KAAK,EAAE,MAAM,UAAU,QAAQ,OAAO,cAAc,WAAW,KAAK,OAAO,OAAO,YAAY,CAAC;AAAA,EACzG;AACA,MAAI,OAAO,cAAc,OAAO,eAAe,SAAS,YAAY;AAClE,YAAQ,KAAK,EAAE,MAAM,QAAQ,QAAQ,OAAO,YAAY,WAAW,KAAK,OAAO,OAAO,UAAU,CAAC;AAAA,EACnG;AACA,MAAI,OAAO,eAAe,OAAO,gBAAgB,SAAS,aAAa;AACrE,YAAQ,KAAK,EAAE,MAAM,SAAS,QAAQ,OAAO,aAAa,WAAW,IAAI,CAAC;AAAA,EAC5E;AACA,SAAO,QAAQ,SAAS,GAAI,SAAQ,MAAM;AAE1C,QAAM,gBAAgB,OAAO,kBAAkB,SAC3C,OAAO,gBACN,OAAO,iBAAiB,YAAY,OAAO,eAAe,YAAY,OAAO,gBAAgB;AAElG,QAAM,UAAwB;AAAA,IAC5B,GAAG;AAAA,IACH;AAAA,IACA,WAAW;AAAA,IACX;AAAA,IACA;AAAA,EACF;AAGA,MAAI,aAAa,qBAAqB;AACpC,QAAI;AACF,yBAAS,OAAO;AAAA,IAClB,SAAS,KAAK;AACZ,cAAQ,MAAM,+DAA+D,GAAG;AAAA,IAClF;AAAA,EACF;AAGA,WAAS,OAAO,IAAI;AACpB,qBAAmB,UAAU,QAAQ;AAErC,iBAAe,EAAE,MAAM,kBAAkB,SAAS,QAAQ,QAAQ,CAAC;AAEnE,SAAO;AACT;AAEO,SAAS,gBAAgB,SAAiB,WAAW,qBAA0C;AAEpG,MAAI,aAAa,qBAAqB;AACpC,QAAI;AACF,YAAM,SAAS,sBAAsB,OAAO;AAC5C,UAAI,OAAQ,QAAO;AAAA,IACrB,QAAQ;AAAA,IAER;AAAA,EACF;AACA,QAAM,WAAW,mBAAmB,QAAQ;AAC5C,SAAO,SAAS,OAAO,KAAK;AAC9B;AAEO,SAAS,kBAAkB,SAAiB,WAAW,qBAA2B;AACvF,QAAM,WAAW,mBAAmB,QAAQ;AAC5C,SAAO,SAAS,OAAO;AACvB,qBAAmB,UAAU,QAAQ;AAGrC,MAAI,aAAa,qBAAqB;AACpC,QAAI;AACF,yBAAS,OAAO;AAAA,IAClB,SAAS,KAAK;AACZ,cAAQ,MAAM,gEAAgE,GAAG;AAAA,IACnF;AAAA,EACF;AACF;AAtKA,IAqCM;AArCN;AAAA;AAAA;AAGA;AACA;AAiCA,IAAM,sBAAsBC,MAAK,QAAQ,GAAG,eAAe,oBAAoB;AAAA;AAAA;","names":["existsSync","mkdirSync","join"]}
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/lib/tracker/rally-api.ts","../src/lib/tracker/interface.ts","../src/lib/tracker/rally.ts"],"sourcesContent":["/**\n * Rally WSAPI REST Client\n *\n * Thin wrapper around native fetch for Rally Web Services API v2.0.\n * Provides typed methods for query, create, and update operations.\n */\n\nexport interface RallyQueryConfig {\n type: string;\n fetch?: string[];\n query?: string;\n limit?: number;\n workspace?: string;\n project?: string;\n projectScopeDown?: boolean;\n order?: string;\n}\n\nexport interface RallyQueryResult {\n QueryResult: {\n Results: any[];\n TotalResultCount: number;\n Errors: string[];\n Warnings: string[];\n };\n}\n\nexport interface RallyCreateConfig {\n type: string;\n data: any;\n fetch?: string[];\n}\n\nexport interface RallyCreateResult {\n CreateResult: {\n Object: any;\n Errors: string[];\n Warnings: string[];\n };\n}\n\nexport interface RallyUpdateConfig {\n type: string;\n ref: string;\n data: any;\n fetch?: string[];\n}\n\nexport interface RallyUpdateResult {\n OperationResult: {\n Object: any;\n Errors: string[];\n Warnings: string[];\n };\n}\n\nexport interface RallyApiConfig {\n apiKey: string;\n server?: string;\n requestOptions?: {\n headers?: Record<string, string>;\n };\n}\n\nexport class RallyRestApi {\n private apiKey: string;\n public server: string;\n private customHeaders: Record<string, string>;\n\n constructor(config: RallyApiConfig) {\n this.apiKey = config.apiKey;\n this.server = config.server || 'https://rally1.rallydev.com';\n this.customHeaders = config.requestOptions?.headers || {};\n }\n\n /**\n * Query Rally artifacts\n */\n async query(config: RallyQueryConfig): Promise<RallyQueryResult> {\n const params = new URLSearchParams();\n\n if (config.query) {\n params.set('query', config.query);\n }\n\n if (config.fetch && config.fetch.length > 0) {\n params.set('fetch', config.fetch.join(','));\n }\n\n if (config.limit !== undefined) {\n params.set('pagesize', String(config.limit));\n }\n\n if (config.workspace) {\n params.set('workspace', config.workspace);\n }\n\n if (config.project) {\n params.set('project', config.project);\n if (config.projectScopeDown) {\n params.set('projectScopeDown', 'true');\n }\n }\n\n if (config.order) {\n params.set('order', config.order);\n }\n\n const url = `${this.server}/slm/webservice/v2.0/${config.type}?${params.toString()}`;\n\n const response = await fetch(url, {\n method: 'GET',\n headers: {\n 'ZSESSIONID': this.apiKey,\n 'Content-Type': 'application/json',\n ...this.customHeaders,\n },\n });\n\n if (!response.ok) {\n if (response.status === 401) {\n throw new Error('Unauthorized: Invalid API key or insufficient permissions');\n }\n throw new Error(`Rally API query failed: ${response.status} ${response.statusText}`);\n }\n\n const result = await response.json() as RallyQueryResult;\n\n if (result.QueryResult.Errors && result.QueryResult.Errors.length > 0) {\n const errorDetail = result.QueryResult.Errors.join(', ');\n const queryDetail = config.query ? ` (Query: ${config.query})` : '';\n if (process.env.DEBUG?.includes('rally')) {\n console.error('[Rally WSAPI] Query failed:', { query: config.query, errors: result.QueryResult.Errors });\n }\n throw new Error(`Rally API query failed: ${errorDetail}${queryDetail}`);\n }\n\n return result;\n }\n\n /**\n * Create a Rally object\n */\n async create(config: RallyCreateConfig): Promise<RallyCreateResult> {\n const url = `${this.server}/slm/webservice/v2.0/${config.type}/create`;\n\n const body: any = {\n [config.type]: config.data,\n };\n\n const params = new URLSearchParams();\n if (config.fetch && config.fetch.length > 0) {\n params.set('fetch', config.fetch.join(','));\n }\n\n const finalUrl = params.toString() ? `${url}?${params.toString()}` : url;\n\n const response = await fetch(finalUrl, {\n method: 'POST',\n headers: {\n 'ZSESSIONID': this.apiKey,\n 'Content-Type': 'application/json',\n ...this.customHeaders,\n },\n body: JSON.stringify(body),\n });\n\n if (!response.ok) {\n throw new Error(`Rally API create failed: ${response.status} ${response.statusText}`);\n }\n\n const result = await response.json() as RallyCreateResult;\n\n if (result.CreateResult.Errors && result.CreateResult.Errors.length > 0) {\n throw new Error(`Rally API create failed: ${result.CreateResult.Errors.join(', ')}`);\n }\n\n return result;\n }\n\n /**\n * Update a Rally object\n */\n async update(config: RallyUpdateConfig): Promise<RallyUpdateResult> {\n // Extract ObjectID from ref (e.g., \"/hierarchicalrequirement/12345\" -> \"12345\")\n const objectId = config.ref.split('/').pop();\n const url = `${this.server}/slm/webservice/v2.0/${config.type}/${objectId}`;\n\n const body: any = {\n [config.type]: config.data,\n };\n\n const params = new URLSearchParams();\n if (config.fetch && config.fetch.length > 0) {\n params.set('fetch', config.fetch.join(','));\n }\n\n const finalUrl = params.toString() ? `${url}?${params.toString()}` : url;\n\n const response = await fetch(finalUrl, {\n method: 'POST',\n headers: {\n 'ZSESSIONID': this.apiKey,\n 'Content-Type': 'application/json',\n ...this.customHeaders,\n },\n body: JSON.stringify(body),\n });\n\n if (!response.ok) {\n throw new Error(`Rally API update failed: ${response.status} ${response.statusText}`);\n }\n\n const result = await response.json() as RallyUpdateResult;\n\n if (result.OperationResult.Errors && result.OperationResult.Errors.length > 0) {\n throw new Error(`Rally API update failed: ${result.OperationResult.Errors.join(', ')}`);\n }\n\n return result;\n }\n}\n","/**\n * Issue Tracker Abstraction Layer\n *\n * Provides a unified interface for different issue tracking systems\n * (Linear, GitHub Issues, GitLab Issues, etc.)\n */\n\n// Supported tracker types\nexport type TrackerType = 'linear' | 'github' | 'gitlab' | 'rally';\n\n// Normalized issue state (lowest common denominator)\nexport type IssueState = 'open' | 'in_progress' | 'closed';\n\n// Normalized issue format\nexport interface Issue {\n /** Tracker-specific unique ID */\n id: string;\n\n /** Human-readable reference (e.g., MIN-630, #42) */\n ref: string;\n\n /** Issue title */\n title: string;\n\n /** Issue description/body (markdown) */\n description: string;\n\n /** Normalized state */\n state: IssueState;\n\n /** Labels/tags */\n labels: string[];\n\n /** Assignee username/name */\n assignee?: string;\n\n /** Web URL to the issue */\n url: string;\n\n /** Which tracker this issue came from */\n tracker: TrackerType;\n\n /** Cross-tracker linked issue references */\n linkedIssues?: string[];\n\n /** Priority (1=urgent, 2=high, 3=normal, 4=low) */\n priority?: number;\n\n /** Due date (ISO string) */\n dueDate?: string;\n\n /** Creation timestamp (ISO string) */\n createdAt: string;\n\n /** Last update timestamp (ISO string) */\n updatedAt: string;\n\n /** Parent issue FormattedID (e.g., \"F1234\") for Rally hierarchy */\n parentRef?: string;\n\n /** Rally artifact type (e.g., \"HierarchicalRequirement\", \"PortfolioItem/Feature\") */\n artifactType?: string;\n\n /** Raw tracker state name before normalization (e.g., \"Discovering\", \"In-Progress\") */\n rawState?: string;\n}\n\n// Comment on an issue\nexport interface Comment {\n id: string;\n issueId: string;\n body: string;\n author: string;\n createdAt: string;\n updatedAt: string;\n}\n\n// Filters for listing issues\nexport interface IssueFilters {\n /** Filter by state */\n state?: IssueState;\n\n /** Filter by labels (AND logic) */\n labels?: string[];\n\n /** Filter by assignee */\n assignee?: string;\n\n /** Filter by team/project (tracker-specific) */\n team?: string;\n\n /** Search query for title/description */\n query?: string;\n\n /** Maximum number of results */\n limit?: number;\n\n /** Include closed issues (default: false) */\n includeClosed?: boolean;\n}\n\n// Data for creating a new issue\nexport interface NewIssue {\n title: string;\n description?: string;\n labels?: string[];\n assignee?: string;\n team?: string;\n priority?: number;\n dueDate?: string;\n}\n\n// Data for updating an issue\nexport interface IssueUpdate {\n title?: string;\n description?: string;\n state?: IssueState;\n labels?: string[];\n assignee?: string;\n priority?: number;\n dueDate?: string;\n}\n\n/**\n * Abstract interface for issue trackers.\n * Implementations must handle normalization to/from tracker-specific formats.\n */\nexport interface IssueTracker {\n /** Tracker type identifier */\n readonly name: TrackerType;\n\n /**\n * List issues matching filters\n */\n listIssues(filters?: IssueFilters): Promise<Issue[]>;\n\n /**\n * Get a single issue by ID or ref\n * @param id - Issue ID or human-readable ref (e.g., \"MIN-630\", \"#42\")\n */\n getIssue(id: string): Promise<Issue>;\n\n /**\n * Update an existing issue\n */\n updateIssue(id: string, update: IssueUpdate): Promise<Issue>;\n\n /**\n * Create a new issue\n */\n createIssue(issue: NewIssue): Promise<Issue>;\n\n /**\n * Get comments on an issue\n */\n getComments(issueId: string): Promise<Comment[]>;\n\n /**\n * Add a comment to an issue\n */\n addComment(issueId: string, body: string): Promise<Comment>;\n\n /**\n * Transition issue to a new state\n */\n transitionIssue(id: string, state: IssueState): Promise<void>;\n\n /**\n * Link a PR/MR to an issue\n */\n linkPR(issueId: string, prUrl: string): Promise<void>;\n}\n\n/**\n * Error thrown when a tracker feature is not implemented\n */\nexport class NotImplementedError extends Error {\n constructor(feature: string) {\n super(`Not implemented: ${feature}`);\n this.name = 'NotImplementedError';\n }\n}\n\n/**\n * Error thrown when an issue is not found\n */\nexport class IssueNotFoundError extends Error {\n constructor(id: string, tracker: TrackerType) {\n super(`Issue not found: ${id} (tracker: ${tracker})`);\n this.name = 'IssueNotFoundError';\n }\n}\n\n/**\n * Error thrown when tracker authentication fails\n */\nexport class TrackerAuthError extends Error {\n constructor(tracker: TrackerType, message: string) {\n super(`Authentication failed for ${tracker}: ${message}`);\n this.name = 'TrackerAuthError';\n }\n}\n","/**\n * Rally Tracker Adapter\n *\n * Implements IssueTracker interface for Broadcom Rally (formerly CA Agile Central).\n * Supports all Rally work item types: User Stories, Defects, Tasks, and Features.\n */\n\nimport { RallyRestApi } from './rally-api.js';\nimport type {\n Issue,\n IssueFilters,\n IssueState,\n IssueTracker,\n IssueUpdate,\n NewIssue,\n Comment,\n TrackerType,\n} from './interface.js';\nimport { IssueNotFoundError, TrackerAuthError, NotImplementedError } from './interface.js';\n\n// Map Rally ScheduleState/State to normalized IssueState.\n// Covers all standard states for User Stories, Defects, Tasks, and Features.\nconst STATE_MAP: Record<string, IssueState> = {\n // User Stories (ScheduleState)\n 'New': 'open',\n 'Idea': 'open',\n 'Defined': 'open',\n 'In-Progress': 'in_progress',\n 'Completed': 'closed',\n 'Accepted': 'closed',\n\n // Defects (State)\n 'Submitted': 'open',\n 'Open': 'in_progress', // \"Open\" defects are actively being worked\n 'Fixed': 'closed',\n 'Closed': 'closed',\n\n // Features / PortfolioItems (State)\n 'Discovering': 'open',\n 'Developing': 'in_progress',\n 'Done': 'closed',\n};\n\n// Rally artifact types we support\ntype RallyArtifactType = 'HierarchicalRequirement' | 'Defect' | 'Task' | 'PortfolioItem/Feature';\n\n/**\n * Type-specific query configuration.\n *\n * Rally WSAPI cannot filter generic Artifact by ScheduleState because not all\n * subtypes have that field. Instead, we query each type separately with its\n * own state field and merge the results. (PAN-168)\n */\ninterface ArtifactTypeQuery {\n /** WSAPI endpoint type (lowercase) */\n type: string;\n /** The state field used for filtering on this artifact type */\n stateField: 'ScheduleState' | 'State';\n /** State values that represent \"closed\" for this type */\n closedStates: string[];\n}\n\nconst QUERYABLE_TYPES: ArtifactTypeQuery[] = [\n { type: 'hierarchicalrequirement', stateField: 'ScheduleState', closedStates: ['Completed', 'Accepted'] },\n { type: 'defect', stateField: 'State', closedStates: ['Closed'] },\n { type: 'task', stateField: 'State', closedStates: ['Completed'] },\n { type: 'portfolioitem/feature', stateField: 'State', closedStates: ['Done'] },\n];\n\nconst FETCH_FIELDS = [\n 'ObjectID',\n 'FormattedID',\n 'Name',\n 'Description',\n 'ScheduleState',\n 'State',\n 'Tags',\n 'Owner',\n 'Priority',\n 'DueDate',\n 'CreationDate',\n 'LastUpdateDate',\n 'Parent',\n 'PortfolioItem',\n '_type',\n];\n\n// Rally priority strings to numbers\nconst PRIORITY_MAP: Record<string, number> = {\n 'Resolve Immediately': 0,\n High: 1,\n Normal: 2,\n Low: 3,\n};\n\n// Reverse priority mapping\nconst REVERSE_PRIORITY_MAP: Record<number, string> = {\n 0: 'Resolve Immediately',\n 1: 'High',\n 2: 'Normal',\n 3: 'Low',\n 4: 'Low',\n};\n\nexport interface RallyConfig {\n apiKey: string;\n server?: string; // Default: rally1.rallydev.com\n workspace?: string; // Rally workspace OID (e.g., \"/workspace/12345\")\n project?: string; // Rally project OID (e.g., \"/project/67890\")\n}\n\nexport class RallyTracker implements IssueTracker {\n readonly name: TrackerType = 'rally' as TrackerType;\n private restApi: RallyRestApi;\n private workspace?: string;\n private project?: string;\n\n constructor(config: RallyConfig) {\n if (!config.apiKey) {\n throw new TrackerAuthError('rally' as TrackerType, 'API key is required');\n }\n\n this.restApi = new RallyRestApi({\n apiKey: config.apiKey,\n server: config.server || 'https://rally1.rallydev.com',\n requestOptions: {\n headers: {\n 'X-RallyIntegrationType': 'Panopticon',\n 'X-RallyIntegrationName': 'Panopticon CLI',\n 'X-RallyIntegrationVendor': 'Mind Your Now',\n 'X-RallyIntegrationVersion': '0.2.0',\n },\n },\n });\n\n this.workspace = config.workspace;\n this.project = config.project;\n }\n\n /**\n * List issues by querying each artifact type separately and merging results.\n *\n * Rally WSAPI cannot apply ScheduleState filters across the generic Artifact\n * endpoint because not all subtypes have that field. We query each type with\n * its own state field, then merge and sort. (PAN-168)\n */\n async listIssues(filters?: IssueFilters): Promise<Issue[]> {\n if (process.env.DEBUG?.includes('rally')) {\n console.debug('[Rally] Query filters:', JSON.stringify(filters));\n }\n\n const limit = filters?.limit ?? 50;\n\n // Extract ObjectID from project ref for explicit query scoping\n // e.g., \"/project/822404704163\" → \"822404704163\"\n let projectObjectId: string | undefined;\n if (this.project) {\n const match = this.project.match(/\\/project\\/(\\d+)/);\n if (match) projectObjectId = match[1];\n }\n\n const queries = QUERYABLE_TYPES.map(async (artifactType) => {\n const queryString = this.buildQueryStringForType(filters, artifactType, projectObjectId);\n\n if (process.env.DEBUG?.includes('rally')) {\n console.debug(`[Rally] ${artifactType.type} query:`, queryString);\n }\n\n const query: any = {\n type: artifactType.type,\n fetch: FETCH_FIELDS,\n limit,\n query: queryString,\n };\n\n if (this.workspace) {\n query.workspace = this.workspace;\n }\n if (this.project) {\n query.project = this.project;\n query.projectScopeDown = true;\n }\n\n try {\n const result = await this.queryRally(query);\n return result.Results.map((artifact: any) => this.normalizeIssue(artifact));\n } catch (error: any) {\n if (error.message?.includes('Unauthorized') || error.message?.includes('401')) {\n throw new TrackerAuthError('rally' as TrackerType, 'Invalid API key or insufficient permissions');\n }\n // Log and skip individual type failures so other types still return\n if (process.env.DEBUG?.includes('rally')) {\n console.debug(`[Rally] Failed to query ${artifactType.type}:`, error.message);\n }\n return [];\n }\n });\n\n const results = await Promise.all(queries);\n const allIssues = results.flat();\n\n // Sort by most recently updated first, then apply overall limit\n allIssues.sort(\n (a, b) => new Date(b.updatedAt).getTime() - new Date(a.updatedAt).getTime()\n );\n\n return allIssues.slice(0, limit);\n }\n\n async getIssue(id: string): Promise<Issue> {\n try {\n // Rally FormattedIDs look like: US123, DE456, TA789, F012\n const query: any = {\n type: 'artifact',\n fetch: [\n 'FormattedID',\n 'Name',\n 'Description',\n 'ScheduleState',\n 'State',\n 'Tags',\n 'Owner',\n 'Priority',\n 'DueDate',\n 'CreationDate',\n 'LastUpdateDate',\n 'Parent',\n '_type',\n ],\n query: `(FormattedID = \"${id}\")`,\n };\n\n if (this.workspace) {\n query.workspace = this.workspace;\n }\n\n const result = await this.queryRally(query);\n\n if (!result.Results || result.Results.length === 0) {\n throw new IssueNotFoundError(id, 'rally' as TrackerType);\n }\n\n return this.normalizeIssue(result.Results[0]);\n } catch (error: any) {\n if (error instanceof IssueNotFoundError) throw error;\n throw new IssueNotFoundError(id, 'rally' as TrackerType);\n }\n }\n\n async updateIssue(id: string, update: IssueUpdate): Promise<Issue> {\n const issue = await this.getIssue(id);\n\n // Get the Rally object reference\n const query: any = {\n type: 'artifact',\n fetch: ['ObjectID', '_ref', '_type'],\n query: `(FormattedID = \"${id}\")`,\n };\n\n if (this.workspace) {\n query.workspace = this.workspace;\n }\n\n const result = await this.queryRally(query);\n if (!result.Results || result.Results.length === 0) {\n throw new IssueNotFoundError(id, 'rally' as TrackerType);\n }\n\n const artifact = result.Results[0];\n const updatePayload: Record<string, unknown> = {};\n\n if (update.title !== undefined) {\n updatePayload.Name = update.title;\n }\n if (update.description !== undefined) {\n updatePayload.Description = update.description;\n }\n if (update.state !== undefined) {\n const artifactType = (artifact._type || '').toLowerCase();\n const kind = artifactType.startsWith('portfolioitem') ? 'feature'\n : artifactType === 'defect' ? 'defect' : 'story';\n const rallyState = this.reverseMapState(update.state, kind);\n if (kind === 'story') {\n updatePayload.ScheduleState = rallyState;\n } else {\n updatePayload.State = rallyState;\n }\n }\n if (update.priority !== undefined) {\n updatePayload.Priority = REVERSE_PRIORITY_MAP[update.priority] || 'Normal';\n }\n if (update.dueDate !== undefined) {\n updatePayload.DueDate = update.dueDate;\n }\n\n if (Object.keys(updatePayload).length > 0) {\n await this.updateRally(artifact._type.toLowerCase(), artifact._ref, updatePayload);\n }\n\n return this.getIssue(id);\n }\n\n async createIssue(newIssue: NewIssue): Promise<Issue> {\n if (!this.project && !newIssue.team) {\n throw new Error('Project is required to create an issue. Set it in config or provide team field.');\n }\n\n const project = newIssue.team || this.project;\n\n // Default to HierarchicalRequirement (User Story) for new issues\n const createPayload: Record<string, unknown> = {\n Name: newIssue.title,\n Description: newIssue.description || '',\n Project: project,\n };\n\n if (newIssue.priority !== undefined) {\n createPayload.Priority = REVERSE_PRIORITY_MAP[newIssue.priority] || 'Normal';\n }\n if (newIssue.dueDate) {\n createPayload.DueDate = newIssue.dueDate;\n }\n if (this.workspace) {\n createPayload.Workspace = this.workspace;\n }\n\n const result = await this.createRally('hierarchicalrequirement', createPayload);\n\n // Fetch the created issue to return normalized format\n return this.getIssue(result.Object.FormattedID);\n }\n\n async getComments(issueId: string): Promise<Comment[]> {\n const issue = await this.getIssue(issueId);\n\n // Get the Rally object to find its Discussion\n const query: any = {\n type: 'artifact',\n fetch: ['ObjectID', '_ref', 'Discussion'],\n query: `(FormattedID = \"${issueId}\")`,\n };\n\n if (this.workspace) {\n query.workspace = this.workspace;\n }\n\n const result = await this.queryRally(query);\n if (!result.Results || result.Results.length === 0) {\n return [];\n }\n\n const artifact = result.Results[0];\n if (!artifact.Discussion) {\n return [];\n }\n\n // Query ConversationPosts for this Discussion\n const postsQuery: any = {\n type: 'conversationpost',\n fetch: ['ObjectID', 'Text', 'User', 'CreationDate', 'PostNumber'],\n query: `(Discussion = \"${artifact.Discussion._ref}\")`,\n order: 'PostNumber',\n };\n\n const postsResult = await this.queryRally(postsQuery);\n\n return (postsResult.Results || []).map((post: any) => ({\n id: post.ObjectID,\n issueId,\n body: post.Text || '',\n author: post.User?._refObjectName || 'Unknown',\n createdAt: post.CreationDate,\n updatedAt: post.CreationDate, // Rally doesn't track comment updates separately\n }));\n }\n\n async addComment(issueId: string, body: string): Promise<Comment> {\n // Get the Rally object to find its Discussion\n const query: any = {\n type: 'artifact',\n fetch: ['ObjectID', '_ref', 'Discussion'],\n query: `(FormattedID = \"${issueId}\")`,\n };\n\n if (this.workspace) {\n query.workspace = this.workspace;\n }\n\n const result = await this.queryRally(query);\n if (!result.Results || result.Results.length === 0) {\n throw new IssueNotFoundError(issueId, 'rally' as TrackerType);\n }\n\n const artifact = result.Results[0];\n\n // If no Discussion exists, create one\n let discussionRef = artifact.Discussion?._ref;\n if (!discussionRef) {\n const discussionResult = await this.createRally('conversationpost', {\n Artifact: artifact._ref,\n Text: body,\n });\n\n return {\n id: discussionResult.Object.ObjectID,\n issueId,\n body,\n author: 'Panopticon',\n createdAt: new Date().toISOString(),\n updatedAt: new Date().toISOString(),\n };\n }\n\n // Add a post to existing Discussion\n const postResult = await this.createRally('conversationpost', {\n Artifact: artifact._ref,\n Text: body,\n });\n\n return {\n id: postResult.Object.ObjectID,\n issueId,\n body,\n author: 'Panopticon',\n createdAt: new Date().toISOString(),\n updatedAt: new Date().toISOString(),\n };\n }\n\n async transitionIssue(id: string, state: IssueState): Promise<void> {\n await this.updateIssue(id, { state });\n }\n\n async linkPR(issueId: string, prUrl: string): Promise<void> {\n // Add a comment with the PR link\n await this.addComment(issueId, `Linked Pull Request: ${prUrl}`);\n }\n\n // Private helper methods\n\n /**\n * Build a Rally WSAPI query string for a specific artifact type.\n *\n * Each artifact type has its own state field:\n * - HierarchicalRequirement: ScheduleState (Defined, In-Progress, Completed, Accepted)\n * - Defect: State (Submitted, Open, Fixed, Closed)\n * - Task: State (Defined, In-Progress, Completed)\n *\n * Rally WSAPI v2.0 requires binary-nested AND/OR with outer parentheses.\n * (PAN-166, PAN-168)\n */\n private buildQueryStringForType(\n filters: IssueFilters | undefined,\n artifactType: ArtifactTypeQuery,\n projectObjectId?: string,\n ): string {\n const conditions: string[] = [];\n\n // Explicit project scoping — more reliable than WSAPI project param alone\n if (projectObjectId) {\n conditions.push(`(Project.ObjectID = \"${projectObjectId}\")`);\n }\n\n if (filters?.state && !filters.includeClosed) {\n const kind = artifactType.type.startsWith('portfolioitem') ? 'feature'\n : artifactType.type === 'defect' ? 'defect' : 'story';\n const rallyState = this.reverseMapState(filters.state, kind);\n conditions.push(`(${artifactType.stateField} = \"${rallyState}\")`);\n }\n\n if (!filters?.includeClosed) {\n // Exclude closed states for this specific artifact type\n const closedConditions = artifactType.closedStates.map(\n (state) => `(${artifactType.stateField} != \"${state}\")`\n );\n // Rally WSAPI only supports binary AND — nest into pairs\n const closedExpr = closedConditions.reduce(\n (acc, cond) => (acc ? `(${acc} AND ${cond})` : cond),\n '',\n );\n conditions.push(closedExpr);\n }\n\n if (filters?.assignee) {\n conditions.push(`(Owner.Name contains \"${filters.assignee}\")`);\n }\n\n if (filters?.labels && filters.labels.length > 0) {\n const labelConditions = filters.labels.map(\n (label) => `(Tags.Name contains \"${label}\")`\n );\n // Rally WSAPI only supports binary AND — nest into pairs\n const labelExpr = labelConditions.reduce((acc, cond) => acc ? `(${acc} AND ${cond})` : cond, '');\n conditions.push(labelExpr);\n }\n\n if (filters?.query) {\n conditions.push(`((Name contains \"${filters.query}\") OR (Description contains \"${filters.query}\"))`);\n }\n\n // Rally WSAPI only supports binary (expr AND expr) — reduce into nested pairs\n return conditions.reduce((acc, cond) => acc ? `(${acc} AND ${cond})` : cond, '');\n }\n\n private normalizeIssue(rallyArtifact: any): Issue {\n // Determine state from ScheduleState (User Stories, Tasks) or State (Defects, Features)\n // For PortfolioItem/Feature, State is a Rally ref object with Name/_refObjectName, not a string\n const rawStateValue = rallyArtifact.ScheduleState || rallyArtifact.State || 'Defined';\n const stateValue = typeof rawStateValue === 'object' && rawStateValue !== null\n ? (rawStateValue.Name || rawStateValue._refObjectName || 'Defined')\n : rawStateValue;\n const state = this.mapState(stateValue);\n\n // Extract tags — ensure all entries are strings\n const labels: string[] = [];\n if (rallyArtifact.Tags && rallyArtifact.Tags._tagsNameArray) {\n for (const tag of rallyArtifact.Tags._tagsNameArray) {\n if (typeof tag === 'string') {\n labels.push(tag);\n } else if (tag?.Name) {\n labels.push(tag.Name);\n }\n }\n }\n\n // Map priority\n const priority = rallyArtifact.Priority\n ? PRIORITY_MAP[rallyArtifact.Priority] ?? 2\n : undefined;\n\n // Use ObjectID if available, fall back to FormattedID\n const objectId = rallyArtifact.ObjectID || rallyArtifact.FormattedID;\n const artifactType = rallyArtifact._type || 'artifact';\n\n // Build URL - Rally's web UI detail path\n const baseUrl = this.restApi.server.replace('/slm/webservice/', '');\n const url = `${baseUrl}/#/detail/${artifactType.toLowerCase()}/${objectId}`;\n\n // Resolve parent reference.\n // For User Stories, PortfolioItem links to the parent Feature (F-prefixed),\n // while Parent links to a parent Story in the hierarchy. Prefer PortfolioItem\n // so that stories are correctly grouped under their Feature. (PAN-202)\n let parentRef: string | undefined;\n if (rallyArtifact.PortfolioItem) {\n if (rallyArtifact.PortfolioItem.FormattedID) {\n parentRef = rallyArtifact.PortfolioItem.FormattedID;\n } else if (rallyArtifact.PortfolioItem._refObjectName) {\n parentRef = rallyArtifact.PortfolioItem._refObjectName;\n }\n } else if (rallyArtifact.Parent) {\n if (rallyArtifact.Parent.FormattedID) {\n parentRef = rallyArtifact.Parent.FormattedID;\n } else if (rallyArtifact.Parent._refObjectName) {\n parentRef = rallyArtifact.Parent._refObjectName;\n }\n }\n\n return {\n id: String(objectId),\n ref: rallyArtifact.FormattedID,\n title: rallyArtifact.Name || '',\n description: rallyArtifact.Description || '',\n state,\n labels,\n assignee: rallyArtifact.Owner?._refObjectName,\n url,\n tracker: 'rally' as TrackerType,\n priority,\n dueDate: rallyArtifact.DueDate,\n createdAt: rallyArtifact.CreationDate,\n updatedAt: rallyArtifact.LastUpdateDate,\n parentRef,\n artifactType,\n rawState: stateValue,\n };\n }\n\n private mapState(rallyState: string): IssueState {\n return STATE_MAP[rallyState] ?? 'open';\n }\n\n private reverseMapState(state: IssueState, kind: 'story' | 'defect' | 'feature' = 'story'): string {\n if (kind === 'feature') {\n // Features / PortfolioItems use State: Discovering, Developing, Done\n switch (state) {\n case 'open': return 'Discovering';\n case 'in_progress': return 'Developing';\n case 'closed': return 'Done';\n default: return 'Discovering';\n }\n }\n if (kind === 'defect') {\n // Defects use State: Submitted, Open, Fixed, Closed\n switch (state) {\n case 'open': return 'Submitted';\n case 'in_progress': return 'Open';\n case 'closed': return 'Closed';\n default: return 'Submitted';\n }\n }\n // User Stories / Tasks use ScheduleState: Defined, In-Progress, Completed\n switch (state) {\n case 'open': return 'Defined';\n case 'in_progress': return 'In-Progress';\n case 'closed': return 'Completed';\n default: return 'Defined';\n }\n }\n\n // Rally API wrapper methods\n private async queryRally(queryConfig: any): Promise<any> {\n const result = await this.restApi.query(queryConfig);\n // Extract Results from WSAPI response format\n return {\n Results: result.QueryResult.Results,\n TotalResultCount: result.QueryResult.TotalResultCount,\n };\n }\n\n private async createRally(type: string, data: any): Promise<any> {\n const result = await this.restApi.create({\n type,\n data,\n fetch: ['FormattedID', 'ObjectID', '_ref'],\n });\n // Extract Object from WSAPI response format\n return {\n Object: result.CreateResult.Object,\n };\n }\n\n private async updateRally(type: string, ref: string, data: any): Promise<any> {\n const result = await this.restApi.update({\n type,\n ref,\n data,\n fetch: ['FormattedID', 'ObjectID'],\n });\n // Extract Object from WSAPI response format\n return {\n Object: result.OperationResult.Object,\n };\n }\n}\n"],"mappings":";;;;;;AAAA,IAgEa;AAhEb;AAAA;AAAA;AAAA;AAgEO,IAAM,eAAN,MAAmB;AAAA,MAChB;AAAA,MACD;AAAA,MACC;AAAA,MAER,YAAY,QAAwB;AAClC,aAAK,SAAS,OAAO;AACrB,aAAK,SAAS,OAAO,UAAU;AAC/B,aAAK,gBAAgB,OAAO,gBAAgB,WAAW,CAAC;AAAA,MAC1D;AAAA;AAAA;AAAA;AAAA,MAKA,MAAM,MAAM,QAAqD;AAC/D,cAAM,SAAS,IAAI,gBAAgB;AAEnC,YAAI,OAAO,OAAO;AAChB,iBAAO,IAAI,SAAS,OAAO,KAAK;AAAA,QAClC;AAEA,YAAI,OAAO,SAAS,OAAO,MAAM,SAAS,GAAG;AAC3C,iBAAO,IAAI,SAAS,OAAO,MAAM,KAAK,GAAG,CAAC;AAAA,QAC5C;AAEA,YAAI,OAAO,UAAU,QAAW;AAC9B,iBAAO,IAAI,YAAY,OAAO,OAAO,KAAK,CAAC;AAAA,QAC7C;AAEA,YAAI,OAAO,WAAW;AACpB,iBAAO,IAAI,aAAa,OAAO,SAAS;AAAA,QAC1C;AAEA,YAAI,OAAO,SAAS;AAClB,iBAAO,IAAI,WAAW,OAAO,OAAO;AACpC,cAAI,OAAO,kBAAkB;AAC3B,mBAAO,IAAI,oBAAoB,MAAM;AAAA,UACvC;AAAA,QACF;AAEA,YAAI,OAAO,OAAO;AAChB,iBAAO,IAAI,SAAS,OAAO,KAAK;AAAA,QAClC;AAEA,cAAM,MAAM,GAAG,KAAK,MAAM,wBAAwB,OAAO,IAAI,IAAI,OAAO,SAAS,CAAC;AAElF,cAAM,WAAW,MAAM,MAAM,KAAK;AAAA,UAChC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,cAAc,KAAK;AAAA,YACnB,gBAAgB;AAAA,YAChB,GAAG,KAAK;AAAA,UACV;AAAA,QACF,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,cAAI,SAAS,WAAW,KAAK;AAC3B,kBAAM,IAAI,MAAM,2DAA2D;AAAA,UAC7E;AACA,gBAAM,IAAI,MAAM,2BAA2B,SAAS,MAAM,IAAI,SAAS,UAAU,EAAE;AAAA,QACrF;AAEA,cAAM,SAAS,MAAM,SAAS,KAAK;AAEnC,YAAI,OAAO,YAAY,UAAU,OAAO,YAAY,OAAO,SAAS,GAAG;AACrE,gBAAM,cAAc,OAAO,YAAY,OAAO,KAAK,IAAI;AACvD,gBAAM,cAAc,OAAO,QAAQ,YAAY,OAAO,KAAK,MAAM;AACjE,cAAI,QAAQ,IAAI,OAAO,SAAS,OAAO,GAAG;AACxC,oBAAQ,MAAM,+BAA+B,EAAE,OAAO,OAAO,OAAO,QAAQ,OAAO,YAAY,OAAO,CAAC;AAAA,UACzG;AACA,gBAAM,IAAI,MAAM,2BAA2B,WAAW,GAAG,WAAW,EAAE;AAAA,QACxE;AAEA,eAAO;AAAA,MACT;AAAA;AAAA;AAAA;AAAA,MAKA,MAAM,OAAO,QAAuD;AAClE,cAAM,MAAM,GAAG,KAAK,MAAM,wBAAwB,OAAO,IAAI;AAE7D,cAAM,OAAY;AAAA,UAChB,CAAC,OAAO,IAAI,GAAG,OAAO;AAAA,QACxB;AAEA,cAAM,SAAS,IAAI,gBAAgB;AACnC,YAAI,OAAO,SAAS,OAAO,MAAM,SAAS,GAAG;AAC3C,iBAAO,IAAI,SAAS,OAAO,MAAM,KAAK,GAAG,CAAC;AAAA,QAC5C;AAEA,cAAM,WAAW,OAAO,SAAS,IAAI,GAAG,GAAG,IAAI,OAAO,SAAS,CAAC,KAAK;AAErE,cAAM,WAAW,MAAM,MAAM,UAAU;AAAA,UACrC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,cAAc,KAAK;AAAA,YACnB,gBAAgB;AAAA,YAChB,GAAG,KAAK;AAAA,UACV;AAAA,UACA,MAAM,KAAK,UAAU,IAAI;AAAA,QAC3B,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,gBAAM,IAAI,MAAM,4BAA4B,SAAS,MAAM,IAAI,SAAS,UAAU,EAAE;AAAA,QACtF;AAEA,cAAM,SAAS,MAAM,SAAS,KAAK;AAEnC,YAAI,OAAO,aAAa,UAAU,OAAO,aAAa,OAAO,SAAS,GAAG;AACvE,gBAAM,IAAI,MAAM,4BAA4B,OAAO,aAAa,OAAO,KAAK,IAAI,CAAC,EAAE;AAAA,QACrF;AAEA,eAAO;AAAA,MACT;AAAA;AAAA;AAAA;AAAA,MAKA,MAAM,OAAO,QAAuD;AAElE,cAAM,WAAW,OAAO,IAAI,MAAM,GAAG,EAAE,IAAI;AAC3C,cAAM,MAAM,GAAG,KAAK,MAAM,wBAAwB,OAAO,IAAI,IAAI,QAAQ;AAEzE,cAAM,OAAY;AAAA,UAChB,CAAC,OAAO,IAAI,GAAG,OAAO;AAAA,QACxB;AAEA,cAAM,SAAS,IAAI,gBAAgB;AACnC,YAAI,OAAO,SAAS,OAAO,MAAM,SAAS,GAAG;AAC3C,iBAAO,IAAI,SAAS,OAAO,MAAM,KAAK,GAAG,CAAC;AAAA,QAC5C;AAEA,cAAM,WAAW,OAAO,SAAS,IAAI,GAAG,GAAG,IAAI,OAAO,SAAS,CAAC,KAAK;AAErE,cAAM,WAAW,MAAM,MAAM,UAAU;AAAA,UACrC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,cAAc,KAAK;AAAA,YACnB,gBAAgB;AAAA,YAChB,GAAG,KAAK;AAAA,UACV;AAAA,UACA,MAAM,KAAK,UAAU,IAAI;AAAA,QAC3B,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,gBAAM,IAAI,MAAM,4BAA4B,SAAS,MAAM,IAAI,SAAS,UAAU,EAAE;AAAA,QACtF;AAEA,cAAM,SAAS,MAAM,SAAS,KAAK;AAEnC,YAAI,OAAO,gBAAgB,UAAU,OAAO,gBAAgB,OAAO,SAAS,GAAG;AAC7E,gBAAM,IAAI,MAAM,4BAA4B,OAAO,gBAAgB,OAAO,KAAK,IAAI,CAAC,EAAE;AAAA,QACxF;AAEA,eAAO;AAAA,MACT;AAAA,IACF;AAAA;AAAA;;;AC7NA,IAgLa,qBAUA,oBAUA;AApMb;AAAA;AAAA;AAAA;AAgLO,IAAM,sBAAN,cAAkC,MAAM;AAAA,MAC7C,YAAY,SAAiB;AAC3B,cAAM,oBAAoB,OAAO,EAAE;AACnC,aAAK,OAAO;AAAA,MACd;AAAA,IACF;AAKO,IAAM,qBAAN,cAAiC,MAAM;AAAA,MAC5C,YAAY,IAAY,SAAsB;AAC5C,cAAM,oBAAoB,EAAE,cAAc,OAAO,GAAG;AACpD,aAAK,OAAO;AAAA,MACd;AAAA,IACF;AAKO,IAAM,mBAAN,cAA+B,MAAM;AAAA,MAC1C,YAAY,SAAsB,SAAiB;AACjD,cAAM,6BAA6B,OAAO,KAAK,OAAO,EAAE;AACxD,aAAK,OAAO;AAAA,MACd;AAAA,IACF;AAAA;AAAA;;;ACzMA,IAsBM,WAwCA,iBAOA,cAmBA,cAQA,sBAeO;AA/Gb;AAAA;AAAA;AAOA;AAWA;AAIA,IAAM,YAAwC;AAAA;AAAA,MAE5C,OAAO;AAAA,MACP,QAAQ;AAAA,MACR,WAAW;AAAA,MACX,eAAe;AAAA,MACf,aAAa;AAAA,MACb,YAAY;AAAA;AAAA,MAGZ,aAAa;AAAA,MACb,QAAQ;AAAA;AAAA,MACR,SAAS;AAAA,MACT,UAAU;AAAA;AAAA,MAGV,eAAe;AAAA,MACf,cAAc;AAAA,MACd,QAAQ;AAAA,IACV;AAqBA,IAAM,kBAAuC;AAAA,MAC3C,EAAE,MAAM,2BAA2B,YAAY,iBAAiB,cAAc,CAAC,aAAa,UAAU,EAAE;AAAA,MACxG,EAAE,MAAM,UAAU,YAAY,SAAS,cAAc,CAAC,QAAQ,EAAE;AAAA,MAChE,EAAE,MAAM,QAAQ,YAAY,SAAS,cAAc,CAAC,WAAW,EAAE;AAAA,MACjE,EAAE,MAAM,yBAAyB,YAAY,SAAS,cAAc,CAAC,MAAM,EAAE;AAAA,IAC/E;AAEA,IAAM,eAAe;AAAA,MACnB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAGA,IAAM,eAAuC;AAAA,MAC3C,uBAAuB;AAAA,MACvB,MAAM;AAAA,MACN,QAAQ;AAAA,MACR,KAAK;AAAA,IACP;AAGA,IAAM,uBAA+C;AAAA,MACnD,GAAG;AAAA,MACH,GAAG;AAAA,MACH,GAAG;AAAA,MACH,GAAG;AAAA,MACH,GAAG;AAAA,IACL;AASO,IAAM,eAAN,MAA2C;AAAA,MACvC,OAAoB;AAAA,MACrB;AAAA,MACA;AAAA,MACA;AAAA,MAER,YAAY,QAAqB;AAC/B,YAAI,CAAC,OAAO,QAAQ;AAClB,gBAAM,IAAI,iBAAiB,SAAwB,qBAAqB;AAAA,QAC1E;AAEA,aAAK,UAAU,IAAI,aAAa;AAAA,UAC9B,QAAQ,OAAO;AAAA,UACf,QAAQ,OAAO,UAAU;AAAA,UACzB,gBAAgB;AAAA,YACd,SAAS;AAAA,cACP,0BAA0B;AAAA,cAC1B,0BAA0B;AAAA,cAC1B,4BAA4B;AAAA,cAC5B,6BAA6B;AAAA,YAC/B;AAAA,UACF;AAAA,QACF,CAAC;AAED,aAAK,YAAY,OAAO;AACxB,aAAK,UAAU,OAAO;AAAA,MACxB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MASA,MAAM,WAAW,SAA0C;AACzD,YAAI,QAAQ,IAAI,OAAO,SAAS,OAAO,GAAG;AACxC,kBAAQ,MAAM,0BAA0B,KAAK,UAAU,OAAO,CAAC;AAAA,QACjE;AAEA,cAAM,QAAQ,SAAS,SAAS;AAIhC,YAAI;AACJ,YAAI,KAAK,SAAS;AAChB,gBAAM,QAAQ,KAAK,QAAQ,MAAM,kBAAkB;AACnD,cAAI,MAAO,mBAAkB,MAAM,CAAC;AAAA,QACtC;AAEA,cAAM,UAAU,gBAAgB,IAAI,OAAO,iBAAiB;AAC1D,gBAAM,cAAc,KAAK,wBAAwB,SAAS,cAAc,eAAe;AAEvF,cAAI,QAAQ,IAAI,OAAO,SAAS,OAAO,GAAG;AACxC,oBAAQ,MAAM,WAAW,aAAa,IAAI,WAAW,WAAW;AAAA,UAClE;AAEA,gBAAM,QAAa;AAAA,YACjB,MAAM,aAAa;AAAA,YACnB,OAAO;AAAA,YACP;AAAA,YACA,OAAO;AAAA,UACT;AAEA,cAAI,KAAK,WAAW;AAClB,kBAAM,YAAY,KAAK;AAAA,UACzB;AACA,cAAI,KAAK,SAAS;AAChB,kBAAM,UAAU,KAAK;AACrB,kBAAM,mBAAmB;AAAA,UAC3B;AAEA,cAAI;AACF,kBAAM,SAAS,MAAM,KAAK,WAAW,KAAK;AAC1C,mBAAO,OAAO,QAAQ,IAAI,CAAC,aAAkB,KAAK,eAAe,QAAQ,CAAC;AAAA,UAC5E,SAAS,OAAY;AACnB,gBAAI,MAAM,SAAS,SAAS,cAAc,KAAK,MAAM,SAAS,SAAS,KAAK,GAAG;AAC7E,oBAAM,IAAI,iBAAiB,SAAwB,6CAA6C;AAAA,YAClG;AAEA,gBAAI,QAAQ,IAAI,OAAO,SAAS,OAAO,GAAG;AACxC,sBAAQ,MAAM,2BAA2B,aAAa,IAAI,KAAK,MAAM,OAAO;AAAA,YAC9E;AACA,mBAAO,CAAC;AAAA,UACV;AAAA,QACF,CAAC;AAED,cAAM,UAAU,MAAM,QAAQ,IAAI,OAAO;AACzC,cAAM,YAAY,QAAQ,KAAK;AAG/B,kBAAU;AAAA,UACR,CAAC,GAAG,MAAM,IAAI,KAAK,EAAE,SAAS,EAAE,QAAQ,IAAI,IAAI,KAAK,EAAE,SAAS,EAAE,QAAQ;AAAA,QAC5E;AAEA,eAAO,UAAU,MAAM,GAAG,KAAK;AAAA,MACjC;AAAA,MAEA,MAAM,SAAS,IAA4B;AACzC,YAAI;AAEF,gBAAM,QAAa;AAAA,YACjB,MAAM;AAAA,YACN,OAAO;AAAA,cACL;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,YACF;AAAA,YACA,OAAO,mBAAmB,EAAE;AAAA,UAC9B;AAEA,cAAI,KAAK,WAAW;AAClB,kBAAM,YAAY,KAAK;AAAA,UACzB;AAEA,gBAAM,SAAS,MAAM,KAAK,WAAW,KAAK;AAE1C,cAAI,CAAC,OAAO,WAAW,OAAO,QAAQ,WAAW,GAAG;AAClD,kBAAM,IAAI,mBAAmB,IAAI,OAAsB;AAAA,UACzD;AAEA,iBAAO,KAAK,eAAe,OAAO,QAAQ,CAAC,CAAC;AAAA,QAC9C,SAAS,OAAY;AACnB,cAAI,iBAAiB,mBAAoB,OAAM;AAC/C,gBAAM,IAAI,mBAAmB,IAAI,OAAsB;AAAA,QACzD;AAAA,MACF;AAAA,MAEA,MAAM,YAAY,IAAY,QAAqC;AACjE,cAAM,QAAQ,MAAM,KAAK,SAAS,EAAE;AAGpC,cAAM,QAAa;AAAA,UACjB,MAAM;AAAA,UACN,OAAO,CAAC,YAAY,QAAQ,OAAO;AAAA,UACnC,OAAO,mBAAmB,EAAE;AAAA,QAC9B;AAEA,YAAI,KAAK,WAAW;AAClB,gBAAM,YAAY,KAAK;AAAA,QACzB;AAEA,cAAM,SAAS,MAAM,KAAK,WAAW,KAAK;AAC1C,YAAI,CAAC,OAAO,WAAW,OAAO,QAAQ,WAAW,GAAG;AAClD,gBAAM,IAAI,mBAAmB,IAAI,OAAsB;AAAA,QACzD;AAEA,cAAM,WAAW,OAAO,QAAQ,CAAC;AACjC,cAAM,gBAAyC,CAAC;AAEhD,YAAI,OAAO,UAAU,QAAW;AAC9B,wBAAc,OAAO,OAAO;AAAA,QAC9B;AACA,YAAI,OAAO,gBAAgB,QAAW;AACpC,wBAAc,cAAc,OAAO;AAAA,QACrC;AACA,YAAI,OAAO,UAAU,QAAW;AAC9B,gBAAM,gBAAgB,SAAS,SAAS,IAAI,YAAY;AACxD,gBAAM,OAAO,aAAa,WAAW,eAAe,IAAI,YACpD,iBAAiB,WAAW,WAAW;AAC3C,gBAAM,aAAa,KAAK,gBAAgB,OAAO,OAAO,IAAI;AAC1D,cAAI,SAAS,SAAS;AACpB,0BAAc,gBAAgB;AAAA,UAChC,OAAO;AACL,0BAAc,QAAQ;AAAA,UACxB;AAAA,QACF;AACA,YAAI,OAAO,aAAa,QAAW;AACjC,wBAAc,WAAW,qBAAqB,OAAO,QAAQ,KAAK;AAAA,QACpE;AACA,YAAI,OAAO,YAAY,QAAW;AAChC,wBAAc,UAAU,OAAO;AAAA,QACjC;AAEA,YAAI,OAAO,KAAK,aAAa,EAAE,SAAS,GAAG;AACzC,gBAAM,KAAK,YAAY,SAAS,MAAM,YAAY,GAAG,SAAS,MAAM,aAAa;AAAA,QACnF;AAEA,eAAO,KAAK,SAAS,EAAE;AAAA,MACzB;AAAA,MAEA,MAAM,YAAY,UAAoC;AACpD,YAAI,CAAC,KAAK,WAAW,CAAC,SAAS,MAAM;AACnC,gBAAM,IAAI,MAAM,iFAAiF;AAAA,QACnG;AAEA,cAAM,UAAU,SAAS,QAAQ,KAAK;AAGtC,cAAM,gBAAyC;AAAA,UAC7C,MAAM,SAAS;AAAA,UACf,aAAa,SAAS,eAAe;AAAA,UACrC,SAAS;AAAA,QACX;AAEA,YAAI,SAAS,aAAa,QAAW;AACnC,wBAAc,WAAW,qBAAqB,SAAS,QAAQ,KAAK;AAAA,QACtE;AACA,YAAI,SAAS,SAAS;AACpB,wBAAc,UAAU,SAAS;AAAA,QACnC;AACA,YAAI,KAAK,WAAW;AAClB,wBAAc,YAAY,KAAK;AAAA,QACjC;AAEA,cAAM,SAAS,MAAM,KAAK,YAAY,2BAA2B,aAAa;AAG9E,eAAO,KAAK,SAAS,OAAO,OAAO,WAAW;AAAA,MAChD;AAAA,MAEA,MAAM,YAAY,SAAqC;AACrD,cAAM,QAAQ,MAAM,KAAK,SAAS,OAAO;AAGzC,cAAM,QAAa;AAAA,UACjB,MAAM;AAAA,UACN,OAAO,CAAC,YAAY,QAAQ,YAAY;AAAA,UACxC,OAAO,mBAAmB,OAAO;AAAA,QACnC;AAEA,YAAI,KAAK,WAAW;AAClB,gBAAM,YAAY,KAAK;AAAA,QACzB;AAEA,cAAM,SAAS,MAAM,KAAK,WAAW,KAAK;AAC1C,YAAI,CAAC,OAAO,WAAW,OAAO,QAAQ,WAAW,GAAG;AAClD,iBAAO,CAAC;AAAA,QACV;AAEA,cAAM,WAAW,OAAO,QAAQ,CAAC;AACjC,YAAI,CAAC,SAAS,YAAY;AACxB,iBAAO,CAAC;AAAA,QACV;AAGA,cAAM,aAAkB;AAAA,UACtB,MAAM;AAAA,UACN,OAAO,CAAC,YAAY,QAAQ,QAAQ,gBAAgB,YAAY;AAAA,UAChE,OAAO,kBAAkB,SAAS,WAAW,IAAI;AAAA,UACjD,OAAO;AAAA,QACT;AAEA,cAAM,cAAc,MAAM,KAAK,WAAW,UAAU;AAEpD,gBAAQ,YAAY,WAAW,CAAC,GAAG,IAAI,CAAC,UAAe;AAAA,UACrD,IAAI,KAAK;AAAA,UACT;AAAA,UACA,MAAM,KAAK,QAAQ;AAAA,UACnB,QAAQ,KAAK,MAAM,kBAAkB;AAAA,UACrC,WAAW,KAAK;AAAA,UAChB,WAAW,KAAK;AAAA;AAAA,QAClB,EAAE;AAAA,MACJ;AAAA,MAEA,MAAM,WAAW,SAAiB,MAAgC;AAEhE,cAAM,QAAa;AAAA,UACjB,MAAM;AAAA,UACN,OAAO,CAAC,YAAY,QAAQ,YAAY;AAAA,UACxC,OAAO,mBAAmB,OAAO;AAAA,QACnC;AAEA,YAAI,KAAK,WAAW;AAClB,gBAAM,YAAY,KAAK;AAAA,QACzB;AAEA,cAAM,SAAS,MAAM,KAAK,WAAW,KAAK;AAC1C,YAAI,CAAC,OAAO,WAAW,OAAO,QAAQ,WAAW,GAAG;AAClD,gBAAM,IAAI,mBAAmB,SAAS,OAAsB;AAAA,QAC9D;AAEA,cAAM,WAAW,OAAO,QAAQ,CAAC;AAGjC,YAAI,gBAAgB,SAAS,YAAY;AACzC,YAAI,CAAC,eAAe;AAClB,gBAAM,mBAAmB,MAAM,KAAK,YAAY,oBAAoB;AAAA,YAClE,UAAU,SAAS;AAAA,YACnB,MAAM;AAAA,UACR,CAAC;AAED,iBAAO;AAAA,YACL,IAAI,iBAAiB,OAAO;AAAA,YAC5B;AAAA,YACA;AAAA,YACA,QAAQ;AAAA,YACR,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,YAClC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,UACpC;AAAA,QACF;AAGA,cAAM,aAAa,MAAM,KAAK,YAAY,oBAAoB;AAAA,UAC5D,UAAU,SAAS;AAAA,UACnB,MAAM;AAAA,QACR,CAAC;AAED,eAAO;AAAA,UACL,IAAI,WAAW,OAAO;AAAA,UACtB;AAAA,UACA;AAAA,UACA,QAAQ;AAAA,UACR,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,UAClC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,QACpC;AAAA,MACF;AAAA,MAEA,MAAM,gBAAgB,IAAY,OAAkC;AAClE,cAAM,KAAK,YAAY,IAAI,EAAE,MAAM,CAAC;AAAA,MACtC;AAAA,MAEA,MAAM,OAAO,SAAiB,OAA8B;AAE1D,cAAM,KAAK,WAAW,SAAS,wBAAwB,KAAK,EAAE;AAAA,MAChE;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAeQ,wBACN,SACA,cACA,iBACQ;AACR,cAAM,aAAuB,CAAC;AAG9B,YAAI,iBAAiB;AACnB,qBAAW,KAAK,wBAAwB,eAAe,IAAI;AAAA,QAC7D;AAEA,YAAI,SAAS,SAAS,CAAC,QAAQ,eAAe;AAC5C,gBAAM,OAAO,aAAa,KAAK,WAAW,eAAe,IAAI,YACzD,aAAa,SAAS,WAAW,WAAW;AAChD,gBAAM,aAAa,KAAK,gBAAgB,QAAQ,OAAO,IAAI;AAC3D,qBAAW,KAAK,IAAI,aAAa,UAAU,OAAO,UAAU,IAAI;AAAA,QAClE;AAEA,YAAI,CAAC,SAAS,eAAe;AAE3B,gBAAM,mBAAmB,aAAa,aAAa;AAAA,YACjD,CAAC,UAAU,IAAI,aAAa,UAAU,QAAQ,KAAK;AAAA,UACrD;AAEA,gBAAM,aAAa,iBAAiB;AAAA,YAClC,CAAC,KAAK,SAAU,MAAM,IAAI,GAAG,QAAQ,IAAI,MAAM;AAAA,YAC/C;AAAA,UACF;AACA,qBAAW,KAAK,UAAU;AAAA,QAC5B;AAEA,YAAI,SAAS,UAAU;AACrB,qBAAW,KAAK,yBAAyB,QAAQ,QAAQ,IAAI;AAAA,QAC/D;AAEA,YAAI,SAAS,UAAU,QAAQ,OAAO,SAAS,GAAG;AAChD,gBAAM,kBAAkB,QAAQ,OAAO;AAAA,YACrC,CAAC,UAAU,wBAAwB,KAAK;AAAA,UAC1C;AAEA,gBAAM,YAAY,gBAAgB,OAAO,CAAC,KAAK,SAAS,MAAM,IAAI,GAAG,QAAQ,IAAI,MAAM,MAAM,EAAE;AAC/F,qBAAW,KAAK,SAAS;AAAA,QAC3B;AAEA,YAAI,SAAS,OAAO;AAClB,qBAAW,KAAK,oBAAoB,QAAQ,KAAK,gCAAgC,QAAQ,KAAK,KAAK;AAAA,QACrG;AAGA,eAAO,WAAW,OAAO,CAAC,KAAK,SAAS,MAAM,IAAI,GAAG,QAAQ,IAAI,MAAM,MAAM,EAAE;AAAA,MACjF;AAAA,MAEQ,eAAe,eAA2B;AAGhD,cAAM,gBAAgB,cAAc,iBAAiB,cAAc,SAAS;AAC5E,cAAM,aAAa,OAAO,kBAAkB,YAAY,kBAAkB,OACrE,cAAc,QAAQ,cAAc,kBAAkB,YACvD;AACJ,cAAM,QAAQ,KAAK,SAAS,UAAU;AAGtC,cAAM,SAAmB,CAAC;AAC1B,YAAI,cAAc,QAAQ,cAAc,KAAK,gBAAgB;AAC3D,qBAAW,OAAO,cAAc,KAAK,gBAAgB;AACnD,gBAAI,OAAO,QAAQ,UAAU;AAC3B,qBAAO,KAAK,GAAG;AAAA,YACjB,WAAW,KAAK,MAAM;AACpB,qBAAO,KAAK,IAAI,IAAI;AAAA,YACtB;AAAA,UACF;AAAA,QACF;AAGA,cAAM,WAAW,cAAc,WAC3B,aAAa,cAAc,QAAQ,KAAK,IACxC;AAGJ,cAAM,WAAW,cAAc,YAAY,cAAc;AACzD,cAAM,eAAe,cAAc,SAAS;AAG5C,cAAM,UAAU,KAAK,QAAQ,OAAO,QAAQ,oBAAoB,EAAE;AAClE,cAAM,MAAM,GAAG,OAAO,aAAa,aAAa,YAAY,CAAC,IAAI,QAAQ;AAMzE,YAAI;AACJ,YAAI,cAAc,eAAe;AAC/B,cAAI,cAAc,cAAc,aAAa;AAC3C,wBAAY,cAAc,cAAc;AAAA,UAC1C,WAAW,cAAc,cAAc,gBAAgB;AACrD,wBAAY,cAAc,cAAc;AAAA,UAC1C;AAAA,QACF,WAAW,cAAc,QAAQ;AAC/B,cAAI,cAAc,OAAO,aAAa;AACpC,wBAAY,cAAc,OAAO;AAAA,UACnC,WAAW,cAAc,OAAO,gBAAgB;AAC9C,wBAAY,cAAc,OAAO;AAAA,UACnC;AAAA,QACF;AAEA,eAAO;AAAA,UACL,IAAI,OAAO,QAAQ;AAAA,UACnB,KAAK,cAAc;AAAA,UACnB,OAAO,cAAc,QAAQ;AAAA,UAC7B,aAAa,cAAc,eAAe;AAAA,UAC1C;AAAA,UACA;AAAA,UACA,UAAU,cAAc,OAAO;AAAA,UAC/B;AAAA,UACA,SAAS;AAAA,UACT;AAAA,UACA,SAAS,cAAc;AAAA,UACvB,WAAW,cAAc;AAAA,UACzB,WAAW,cAAc;AAAA,UACzB;AAAA,UACA;AAAA,UACA,UAAU;AAAA,QACZ;AAAA,MACF;AAAA,MAEQ,SAAS,YAAgC;AAC/C,eAAO,UAAU,UAAU,KAAK;AAAA,MAClC;AAAA,MAEQ,gBAAgB,OAAmB,OAAuC,SAAiB;AACjG,YAAI,SAAS,WAAW;AAEtB,kBAAQ,OAAO;AAAA,YACb,KAAK;AAAQ,qBAAO;AAAA,YACpB,KAAK;AAAe,qBAAO;AAAA,YAC3B,KAAK;AAAU,qBAAO;AAAA,YACtB;AAAS,qBAAO;AAAA,UAClB;AAAA,QACF;AACA,YAAI,SAAS,UAAU;AAErB,kBAAQ,OAAO;AAAA,YACb,KAAK;AAAQ,qBAAO;AAAA,YACpB,KAAK;AAAe,qBAAO;AAAA,YAC3B,KAAK;AAAU,qBAAO;AAAA,YACtB;AAAS,qBAAO;AAAA,UAClB;AAAA,QACF;AAEA,gBAAQ,OAAO;AAAA,UACb,KAAK;AAAQ,mBAAO;AAAA,UACpB,KAAK;AAAe,mBAAO;AAAA,UAC3B,KAAK;AAAU,mBAAO;AAAA,UACtB;AAAS,mBAAO;AAAA,QAClB;AAAA,MACF;AAAA;AAAA,MAGA,MAAc,WAAW,aAAgC;AACvD,cAAM,SAAS,MAAM,KAAK,QAAQ,MAAM,WAAW;AAEnD,eAAO;AAAA,UACL,SAAS,OAAO,YAAY;AAAA,UAC5B,kBAAkB,OAAO,YAAY;AAAA,QACvC;AAAA,MACF;AAAA,MAEA,MAAc,YAAY,MAAc,MAAyB;AAC/D,cAAM,SAAS,MAAM,KAAK,QAAQ,OAAO;AAAA,UACvC;AAAA,UACA;AAAA,UACA,OAAO,CAAC,eAAe,YAAY,MAAM;AAAA,QAC3C,CAAC;AAED,eAAO;AAAA,UACL,QAAQ,OAAO,aAAa;AAAA,QAC9B;AAAA,MACF;AAAA,MAEA,MAAc,YAAY,MAAc,KAAa,MAAyB;AAC5E,cAAM,SAAS,MAAM,KAAK,QAAQ,OAAO;AAAA,UACvC;AAAA,UACA;AAAA,UACA;AAAA,UACA,OAAO,CAAC,eAAe,UAAU;AAAA,QACnC,CAAC;AAED,eAAO;AAAA,UACL,QAAQ,OAAO,gBAAgB;AAAA,QACjC;AAAA,MACF;AAAA,IACF;AAAA;AAAA;","names":[]}
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/lib/config.ts"],"sourcesContent":["import { readFileSync, writeFileSync, existsSync } from 'fs';\nimport { join, dirname, parse as parsePath } from 'path';\nimport { homedir } from 'os';\nimport { parse, stringify } from '@iarna/toml';\nimport { CONFIG_FILE } from './paths.js';\nimport type { TrackerType } from './tracker/interface.js';\n\n// Individual tracker configuration\nexport interface LinearConfig {\n type: 'linear';\n api_key_env?: string; // Env var name for API key (default: LINEAR_API_KEY)\n team?: string; // Default team prefix (e.g., 'MIN')\n}\n\nexport interface GitHubConfig {\n type: 'github';\n token_env?: string; // Env var name for token (default: GITHUB_TOKEN)\n owner: string; // Repository owner\n repo: string; // Repository name\n}\n\nexport interface GitLabConfig {\n type: 'gitlab';\n token_env?: string; // Env var name for token (default: GITLAB_TOKEN)\n project_id: string; // GitLab project ID\n}\n\nexport interface RallyConfig {\n type: 'rally';\n api_key_env?: string; // Env var name for API key (default: RALLY_API_KEY)\n server?: string; // Rally server URL (default: rally1.rallydev.com)\n workspace?: string; // Rally workspace OID (e.g., \"/workspace/12345\")\n project?: string; // Rally project OID (e.g., \"/project/67890\")\n}\n\nexport type TrackerConfigItem = LinearConfig | GitHubConfig | GitLabConfig | RallyConfig;\n\nexport interface TrackersConfig {\n primary: TrackerType;\n secondary?: TrackerType;\n linear?: LinearConfig;\n github?: GitHubConfig;\n gitlab?: GitLabConfig;\n rally?: RallyConfig;\n}\n\nexport interface RemoteExeConfig {\n /** Shared infrastructure VM for postgres/redis/traefik */\n infra_vm?: string;\n /** Postgres settings on infra VM */\n postgres_host?: string;\n postgres_port?: number;\n postgres_user?: string;\n postgres_password_env?: string;\n /** Redis settings on infra VM */\n redis_host?: string;\n redis_port?: number;\n}\n\nexport interface RemoteConfig {\n /** Enable remote workspace support */\n enabled: boolean;\n /** Remote provider type */\n provider?: 'exe';\n /** Default location for new workspaces */\n default_location?: 'local' | 'remote';\n /** Auto-hibernate idle workspaces after N minutes (0 = disabled) */\n auto_hibernate_minutes?: number;\n /** exe.dev specific configuration */\n exe?: RemoteExeConfig;\n}\n\nexport interface ShadowConfig {\n enabled: boolean;\n trackers: {\n linear: boolean;\n github: boolean;\n gitlab: boolean;\n rally: boolean;\n };\n}\n\nexport interface PanopticonConfig {\n panopticon: {\n version: string;\n };\n sync: {\n backup_before_sync: boolean;\n auto_sync?: boolean;\n strategy?: 'symlink' | 'copy';\n /** Parent directory where all projects live (e.g., ~/Projects).\n * Skills are placed at <devroot>/.claude/skills/ (project level).\n * Set to null or empty string to disable devroot skill placement. */\n devroot?: string | null;\n };\n trackers: TrackersConfig;\n dashboard: {\n port: number;\n api_port: number;\n };\n traefik: {\n enabled: boolean;\n dashboard_port?: number;\n domain?: string;\n dns_sync_method?: 'wsl2hosts' | 'hosts_file' | 'dnsmasq';\n };\n remote?: RemoteConfig;\n shadow: ShadowConfig;\n}\n\nconst DEFAULT_CONFIG: PanopticonConfig = {\n panopticon: {\n version: '1.0.0',\n },\n sync: {\n backup_before_sync: true,\n auto_sync: false,\n strategy: 'symlink',\n devroot: '~/Projects',\n },\n trackers: {\n primary: 'linear',\n linear: {\n type: 'linear',\n api_key_env: 'LINEAR_API_KEY',\n },\n },\n dashboard: {\n port: 3010,\n api_port: 3011,\n },\n traefik: {\n enabled: false,\n dashboard_port: 8080,\n domain: 'pan.localhost',\n },\n shadow: {\n enabled: false,\n trackers: {\n linear: false,\n github: false,\n gitlab: false,\n rally: false,\n },\n },\n};\n\n/**\n * Deep merge utility that recursively merges objects.\n * - Recursively merges nested objects\n * - Arrays in overrides replace defaults (not concatenated)\n * - User values take precedence over defaults\n */\nfunction deepMerge<T extends object>(defaults: T, overrides: Partial<T>): T {\n const result = { ...defaults };\n\n for (const key of Object.keys(overrides) as (keyof T)[]) {\n const defaultVal = defaults[key];\n const overrideVal = overrides[key];\n\n // Skip undefined values in overrides\n if (overrideVal === undefined) continue;\n\n // Deep merge if both values are non-array objects\n if (\n typeof defaultVal === 'object' &&\n defaultVal !== null &&\n !Array.isArray(defaultVal) &&\n typeof overrideVal === 'object' &&\n overrideVal !== null &&\n !Array.isArray(overrideVal)\n ) {\n result[key] = deepMerge(defaultVal, overrideVal as any);\n } else {\n // For primitives, arrays, or null - override wins\n result[key] = overrideVal as T[keyof T];\n }\n }\n\n return result;\n}\n\nexport function loadConfig(): PanopticonConfig {\n if (!existsSync(CONFIG_FILE)) {\n return DEFAULT_CONFIG;\n }\n\n try {\n const content = readFileSync(CONFIG_FILE, 'utf8');\n const parsed = parse(content) as unknown as Partial<PanopticonConfig>;\n return deepMerge(DEFAULT_CONFIG, parsed);\n } catch (error) {\n console.error('Warning: Failed to parse config, using defaults');\n return DEFAULT_CONFIG;\n }\n}\n\nexport function saveConfig(config: PanopticonConfig): void {\n const content = stringify(config as any);\n writeFileSync(CONFIG_FILE, content, 'utf8');\n}\n\nexport function getDefaultConfig(): PanopticonConfig {\n return JSON.parse(JSON.stringify(DEFAULT_CONFIG));\n}\n\n/**\n * Get the dashboard API base URL from config.\n * Reads from DASHBOARD_URL env var first, then config file, then defaults.\n */\nexport function getDashboardApiUrl(): string {\n if (process.env.DASHBOARD_URL) return process.env.DASHBOARD_URL;\n const config = loadConfig();\n const port = config.dashboard?.api_port || 3011;\n return `http://localhost:${port}`;\n}\n\n/**\n * Get the resolved devroot path from config.\n * Returns null if devroot is disabled (set to null or empty string).\n * Resolves ~ to home directory and validates the directory exists.\n */\nexport function getDevrootPath(): string | null {\n const config = loadConfig();\n const devroot = config.sync?.devroot;\n\n if (!devroot) return null;\n\n // Resolve ~ to home directory\n const resolved = devroot.startsWith('~/')\n ? join(homedir(), devroot.slice(2))\n : devroot;\n\n if (!existsSync(resolved)) return null;\n\n return resolved;\n}\n\n/**\n * Find the devroot for a given project path.\n * Tries config first, then walks up from projectPath looking for .claude/ directory.\n * Returns the project path itself as last resort.\n */\nexport function findDevrootForProject(projectPath: string): string {\n // 1. Explicit config takes priority\n const configured = getDevrootPath();\n if (configured) return configured;\n\n // 2. Walk up from project path to find nearest .claude/ directory\n let dir = projectPath;\n const root = parsePath(dir).root;\n while (dir !== root && dir !== homedir()) {\n const parent = dirname(dir);\n if (parent === dir) break;\n if (existsSync(join(parent, '.claude'))) {\n return parent;\n }\n dir = parent;\n }\n\n // 3. Fallback to project path itself\n return projectPath;\n}\n\n"],"mappings":";;;;;;;;;;AAAA,SAAS,cAAc,eAAe,kBAAkB;AACxD,SAAS,MAAM,SAAS,SAAS,iBAAiB;AAClD,SAAS,eAAe;AACxB,SAAS,OAAO,iBAAiB;AAsJjC,SAAS,UAA4B,UAAa,WAA0B;AAC1E,QAAM,SAAS,EAAE,GAAG,SAAS;AAE7B,aAAW,OAAO,OAAO,KAAK,SAAS,GAAkB;AACvD,UAAM,aAAa,SAAS,GAAG;AAC/B,UAAM,cAAc,UAAU,GAAG;AAGjC,QAAI,gBAAgB,OAAW;AAG/B,QACE,OAAO,eAAe,YACtB,eAAe,QACf,CAAC,MAAM,QAAQ,UAAU,KACzB,OAAO,gBAAgB,YACvB,gBAAgB,QAChB,CAAC,MAAM,QAAQ,WAAW,GAC1B;AACA,aAAO,GAAG,IAAI,UAAU,YAAY,WAAkB;AAAA,IACxD,OAAO;AAEL,aAAO,GAAG,IAAI;AAAA,IAChB;AAAA,EACF;AAEA,SAAO;AACT;AAEO,SAAS,aAA+B;AAC7C,MAAI,CAAC,WAAW,WAAW,GAAG;AAC5B,WAAO;AAAA,EACT;AAEA,MAAI;AACF,UAAM,UAAU,aAAa,aAAa,MAAM;AAChD,UAAM,SAAS,MAAM,OAAO;AAC5B,WAAO,UAAU,gBAAgB,MAAM;AAAA,EACzC,SAAS,OAAO;AACd,YAAQ,MAAM,iDAAiD;AAC/D,WAAO;AAAA,EACT;AACF;AAEO,SAAS,WAAW,QAAgC;AACzD,QAAM,UAAU,UAAU,MAAa;AACvC,gBAAc,aAAa,SAAS,MAAM;AAC5C;AAEO,SAAS,mBAAqC;AACnD,SAAO,KAAK,MAAM,KAAK,UAAU,cAAc,CAAC;AAClD;AAMO,SAAS,qBAA6B;AAC3C,MAAI,QAAQ,IAAI,cAAe,QAAO,QAAQ,IAAI;AAClD,QAAM,SAAS,WAAW;AAC1B,QAAM,OAAO,OAAO,WAAW,YAAY;AAC3C,SAAO,oBAAoB,IAAI;AACjC;AAOO,SAAS,iBAAgC;AAC9C,QAAM,SAAS,WAAW;AAC1B,QAAM,UAAU,OAAO,MAAM;AAE7B,MAAI,CAAC,QAAS,QAAO;AAGrB,QAAM,WAAW,QAAQ,WAAW,IAAI,IACpC,KAAK,QAAQ,GAAG,QAAQ,MAAM,CAAC,CAAC,IAChC;AAEJ,MAAI,CAAC,WAAW,QAAQ,EAAG,QAAO;AAElC,SAAO;AACT;AAOO,SAAS,sBAAsB,aAA6B;AAEjE,QAAM,aAAa,eAAe;AAClC,MAAI,WAAY,QAAO;AAGvB,MAAI,MAAM;AACV,QAAM,OAAO,UAAU,GAAG,EAAE;AAC5B,SAAO,QAAQ,QAAQ,QAAQ,QAAQ,GAAG;AACxC,UAAM,SAAS,QAAQ,GAAG;AAC1B,QAAI,WAAW,IAAK;AACpB,QAAI,WAAW,KAAK,QAAQ,SAAS,CAAC,GAAG;AACvC,aAAO;AAAA,IACT;AACA,UAAM;AAAA,EACR;AAGA,SAAO;AACT;AAtQA,IA8GM;AA9GN;AAAA;AAAA;AAIA;AA0GA,IAAM,iBAAmC;AAAA,MACvC,YAAY;AAAA,QACV,SAAS;AAAA,MACX;AAAA,MACA,MAAM;AAAA,QACJ,oBAAoB;AAAA,QACpB,WAAW;AAAA,QACX,UAAU;AAAA,QACV,SAAS;AAAA,MACX;AAAA,MACA,UAAU;AAAA,QACR,SAAS;AAAA,QACT,QAAQ;AAAA,UACN,MAAM;AAAA,UACN,aAAa;AAAA,QACf;AAAA,MACF;AAAA,MACA,WAAW;AAAA,QACT,MAAM;AAAA,QACN,UAAU;AAAA,MACZ;AAAA,MACA,SAAS;AAAA,QACP,SAAS;AAAA,QACT,gBAAgB;AAAA,QAChB,QAAQ;AAAA,MACV;AAAA,MACA,QAAQ;AAAA,QACN,SAAS;AAAA,QACT,UAAU;AAAA,UACR,QAAQ;AAAA,UACR,QAAQ;AAAA,UACR,QAAQ;AAAA,UACR,OAAO;AAAA,QACT;AAAA,MACF;AAAA,IACF;AAAA;AAAA;","names":[]}