@stackmemoryai/stackmemory 0.3.6 → 0.3.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agents/verifiers/base-verifier.js.map +2 -2
- package/dist/agents/verifiers/formatter-verifier.js.map +2 -2
- package/dist/agents/verifiers/llm-judge.js.map +2 -2
- package/dist/cli/claude-sm.js +24 -13
- package/dist/cli/claude-sm.js.map +2 -2
- package/dist/cli/codex-sm.js +24 -13
- package/dist/cli/codex-sm.js.map +2 -2
- package/dist/cli/commands/agent.js.map +2 -2
- package/dist/cli/commands/chromadb.js +217 -32
- package/dist/cli/commands/chromadb.js.map +2 -2
- package/dist/cli/commands/clear.js +12 -1
- package/dist/cli/commands/clear.js.map +2 -2
- package/dist/cli/commands/context.js +13 -2
- package/dist/cli/commands/context.js.map +2 -2
- package/dist/cli/commands/dashboard.js.map +2 -2
- package/dist/cli/commands/gc.js +202 -0
- package/dist/cli/commands/gc.js.map +7 -0
- package/dist/cli/commands/handoff.js +12 -1
- package/dist/cli/commands/handoff.js.map +2 -2
- package/dist/cli/commands/infinite-storage.js +32 -21
- package/dist/cli/commands/infinite-storage.js.map +2 -2
- package/dist/cli/commands/linear-create.js +13 -2
- package/dist/cli/commands/linear-create.js.map +2 -2
- package/dist/cli/commands/linear-list.js +12 -1
- package/dist/cli/commands/linear-list.js.map +2 -2
- package/dist/cli/commands/linear-migrate.js +12 -1
- package/dist/cli/commands/linear-migrate.js.map +2 -2
- package/dist/cli/commands/linear-test.js +12 -1
- package/dist/cli/commands/linear-test.js.map +2 -2
- package/dist/cli/commands/linear-unified.js +262 -0
- package/dist/cli/commands/linear-unified.js.map +7 -0
- package/dist/cli/commands/linear.js +17 -6
- package/dist/cli/commands/linear.js.map +2 -2
- package/dist/cli/commands/monitor.js.map +2 -2
- package/dist/cli/commands/onboard.js.map +2 -2
- package/dist/cli/commands/quality.js.map +2 -2
- package/dist/cli/commands/search.js.map +2 -2
- package/dist/cli/commands/session.js.map +2 -2
- package/dist/cli/commands/skills.js +12 -1
- package/dist/cli/commands/skills.js.map +2 -2
- package/dist/cli/commands/storage.js +18 -7
- package/dist/cli/commands/storage.js.map +2 -2
- package/dist/cli/commands/tasks.js.map +2 -2
- package/dist/cli/commands/tui.js +13 -2
- package/dist/cli/commands/tui.js.map +2 -2
- package/dist/cli/commands/webhook.js +14 -3
- package/dist/cli/commands/webhook.js.map +2 -2
- package/dist/cli/commands/workflow.js +14 -3
- package/dist/cli/commands/workflow.js.map +2 -2
- package/dist/cli/commands/worktree.js.map +2 -2
- package/dist/cli/index.js +18 -5
- package/dist/cli/index.js.map +2 -2
- package/dist/core/config/config-manager.js.map +2 -2
- package/dist/core/context/auto-context.js.map +2 -2
- package/dist/core/context/compaction-handler.js.map +2 -2
- package/dist/core/context/context-bridge.js.map +2 -2
- package/dist/core/context/dual-stack-manager.js.map +2 -2
- package/dist/core/context/frame-database.js.map +2 -2
- package/dist/core/context/frame-digest.js.map +2 -2
- package/dist/core/context/frame-handoff-manager.js.map +2 -2
- package/dist/core/context/frame-manager.js +12 -1
- package/dist/core/context/frame-manager.js.map +2 -2
- package/dist/core/context/frame-stack.js.map +2 -2
- package/dist/core/context/incremental-gc.js +279 -0
- package/dist/core/context/incremental-gc.js.map +7 -0
- package/dist/core/context/permission-manager.js +12 -1
- package/dist/core/context/permission-manager.js.map +2 -2
- package/dist/core/context/refactored-frame-manager.js.map +2 -2
- package/dist/core/context/shared-context-layer.js +12 -1
- package/dist/core/context/shared-context-layer.js.map +2 -2
- package/dist/core/context/stack-merge-resolver.js.map +2 -2
- package/dist/core/context/validation.js.map +2 -2
- package/dist/core/database/batch-operations.js.map +2 -2
- package/dist/core/database/connection-pool.js.map +2 -2
- package/dist/core/database/migration-manager.js.map +2 -2
- package/dist/core/database/paradedb-adapter.js.map +2 -2
- package/dist/core/database/query-cache.js.map +2 -2
- package/dist/core/database/query-router.js.map +2 -2
- package/dist/core/database/sqlite-adapter.js.map +2 -2
- package/dist/core/digest/enhanced-hybrid-digest.js.map +2 -2
- package/dist/core/errors/recovery.js.map +2 -2
- package/dist/core/merge/resolution-engine.js.map +2 -2
- package/dist/core/monitoring/error-handler.js.map +2 -2
- package/dist/core/monitoring/logger.js +14 -3
- package/dist/core/monitoring/logger.js.map +2 -2
- package/dist/core/monitoring/metrics.js +13 -2
- package/dist/core/monitoring/metrics.js.map +2 -2
- package/dist/core/monitoring/progress-tracker.js +12 -1
- package/dist/core/monitoring/progress-tracker.js.map +2 -2
- package/dist/core/monitoring/session-monitor.js.map +2 -2
- package/dist/core/performance/context-cache.js.map +2 -2
- package/dist/core/performance/lazy-context-loader.js.map +2 -2
- package/dist/core/performance/monitor.js.map +2 -2
- package/dist/core/performance/optimized-frame-context.js.map +2 -2
- package/dist/core/performance/performance-benchmark.js.map +2 -2
- package/dist/core/performance/performance-profiler.js +12 -1
- package/dist/core/performance/performance-profiler.js.map +2 -2
- package/dist/core/performance/streaming-jsonl-parser.js.map +2 -2
- package/dist/core/persistence/postgres-adapter.js.map +2 -2
- package/dist/core/projects/project-manager.js.map +2 -2
- package/dist/core/retrieval/context-retriever.js.map +2 -2
- package/dist/core/retrieval/graph-retrieval.js.map +2 -2
- package/dist/core/retrieval/llm-context-retrieval.js.map +2 -2
- package/dist/core/retrieval/retrieval-benchmarks.js.map +2 -2
- package/dist/core/retrieval/summary-generator.js.map +2 -2
- package/dist/core/session/clear-survival.js.map +2 -2
- package/dist/core/session/handoff-generator.js.map +2 -2
- package/dist/core/session/session-manager.js +16 -5
- package/dist/core/session/session-manager.js.map +2 -2
- package/dist/core/skills/skill-storage.js +13 -2
- package/dist/core/skills/skill-storage.js.map +2 -2
- package/dist/core/storage/chromadb-adapter.js.map +2 -2
- package/dist/core/storage/chromadb-simple.js.map +2 -2
- package/dist/core/storage/infinite-storage.js.map +2 -2
- package/dist/core/storage/railway-optimized-storage.js +19 -8
- package/dist/core/storage/railway-optimized-storage.js.map +2 -2
- package/dist/core/storage/remote-storage.js +12 -1
- package/dist/core/storage/remote-storage.js.map +2 -2
- package/dist/core/trace/cli-trace-wrapper.js +16 -5
- package/dist/core/trace/cli-trace-wrapper.js.map +2 -2
- package/dist/core/trace/db-trace-wrapper.js.map +2 -2
- package/dist/core/trace/debug-trace.js +21 -10
- package/dist/core/trace/debug-trace.js.map +2 -2
- package/dist/core/trace/index.js +46 -35
- package/dist/core/trace/index.js.map +2 -2
- package/dist/core/trace/trace-demo.js +12 -1
- package/dist/core/trace/trace-demo.js.map +2 -2
- package/dist/core/trace/trace-detector.js.map +2 -2
- package/dist/core/trace/trace-store.js.map +2 -2
- package/dist/core/utils/update-checker.js.map +2 -2
- package/dist/core/worktree/worktree-manager.js.map +2 -2
- package/dist/features/analytics/api/analytics-api.js.map +2 -2
- package/dist/features/analytics/core/analytics-service.js +12 -1
- package/dist/features/analytics/core/analytics-service.js.map +2 -2
- package/dist/features/analytics/queries/metrics-queries.js.map +2 -2
- package/dist/features/tasks/pebbles-task-store.js.map +2 -2
- package/dist/features/tui/components/analytics-panel.js.map +2 -2
- package/dist/features/tui/components/pr-tracker.js.map +2 -2
- package/dist/features/tui/components/session-monitor.js.map +2 -2
- package/dist/features/tui/components/subagent-fleet.js.map +2 -2
- package/dist/features/tui/components/task-board.js +650 -2
- package/dist/features/tui/components/task-board.js.map +2 -2
- package/dist/features/tui/index.js +16 -5
- package/dist/features/tui/index.js.map +2 -2
- package/dist/features/tui/services/data-service.js +25 -14
- package/dist/features/tui/services/data-service.js.map +2 -2
- package/dist/features/tui/services/linear-task-reader.js.map +2 -2
- package/dist/features/tui/services/websocket-client.js +13 -2
- package/dist/features/tui/services/websocket-client.js.map +2 -2
- package/dist/features/tui/terminal-compat.js +27 -16
- package/dist/features/tui/terminal-compat.js.map +2 -2
- package/dist/features/web/client/stores/task-store.js.map +2 -2
- package/dist/features/web/server/index.js +13 -2
- package/dist/features/web/server/index.js.map +2 -2
- package/dist/integrations/claude-code/enhanced-pre-clear-hooks.js.map +2 -2
- package/dist/integrations/claude-code/lifecycle-hooks.js.map +2 -2
- package/dist/integrations/claude-code/post-task-hooks.js.map +2 -2
- package/dist/integrations/linear/auth.js +17 -6
- package/dist/integrations/linear/auth.js.map +2 -2
- package/dist/integrations/linear/auto-sync.js.map +2 -2
- package/dist/integrations/linear/client.js.map +2 -2
- package/dist/integrations/linear/config.js.map +2 -2
- package/dist/integrations/linear/migration.js.map +2 -2
- package/dist/integrations/linear/oauth-server.js +13 -2
- package/dist/integrations/linear/oauth-server.js.map +2 -2
- package/dist/integrations/linear/rest-client.js.map +2 -2
- package/dist/integrations/linear/sync-enhanced.js +202 -0
- package/dist/integrations/linear/sync-enhanced.js.map +7 -0
- package/dist/integrations/linear/sync-manager.js.map +2 -2
- package/dist/integrations/linear/sync-service.js +12 -1
- package/dist/integrations/linear/sync-service.js.map +2 -2
- package/dist/integrations/linear/sync.js +34 -3
- package/dist/integrations/linear/sync.js.map +2 -2
- package/dist/integrations/linear/unified-sync.js +560 -0
- package/dist/integrations/linear/unified-sync.js.map +7 -0
- package/dist/integrations/linear/webhook-handler.js +12 -1
- package/dist/integrations/linear/webhook-handler.js.map +2 -2
- package/dist/integrations/linear/webhook-server.js +14 -3
- package/dist/integrations/linear/webhook-server.js.map +2 -2
- package/dist/integrations/linear/webhook.js +12 -1
- package/dist/integrations/linear/webhook.js.map +2 -2
- package/dist/integrations/mcp/handlers/context-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/linear-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/skill-handlers.js +13 -2
- package/dist/integrations/mcp/handlers/skill-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/task-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/trace-handlers.js.map +2 -2
- package/dist/integrations/mcp/middleware/tool-scoring.js.map +2 -2
- package/dist/integrations/mcp/refactored-server.js +15 -4
- package/dist/integrations/mcp/refactored-server.js.map +2 -2
- package/dist/integrations/mcp/server.js +12 -1
- package/dist/integrations/mcp/server.js.map +2 -2
- package/dist/integrations/mcp/tool-definitions.js.map +2 -2
- package/dist/integrations/pg-aiguide/embedding-provider.js +13 -2
- package/dist/integrations/pg-aiguide/embedding-provider.js.map +2 -2
- package/dist/integrations/pg-aiguide/semantic-search.js.map +2 -2
- package/dist/mcp/stackmemory-mcp-server.js +12 -1
- package/dist/mcp/stackmemory-mcp-server.js.map +2 -2
- package/dist/middleware/exponential-rate-limiter.js.map +2 -2
- package/dist/servers/production/auth-middleware.js +13 -2
- package/dist/servers/production/auth-middleware.js.map +2 -2
- package/dist/servers/railway/index.js +22 -11
- package/dist/servers/railway/index.js.map +2 -2
- package/dist/services/config-service.js.map +2 -2
- package/dist/services/context-service.js.map +2 -2
- package/dist/skills/claude-skills.js +105 -2
- package/dist/skills/claude-skills.js.map +2 -2
- package/dist/skills/dashboard-launcher.js.map +2 -2
- package/dist/skills/repo-ingestion-skill.js +561 -0
- package/dist/skills/repo-ingestion-skill.js.map +7 -0
- package/dist/utils/logger.js +12 -1
- package/dist/utils/logger.js.map +2 -2
- package/package.json +5 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/performance/streaming-jsonl-parser.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Optimized Streaming JSONL Parser\n * Memory-efficient parsing for large JSONL files with async streaming\n */\n\nimport { createReadStream } from 'fs';\nimport { createInterface } from 'readline';\nimport { Transform, pipeline } from 'stream';\nimport { promisify } from 'util';\nimport { logger } from '../monitoring/logger.js';\n\nconst pipelineAsync = promisify(pipeline);\n\nexport interface ParseOptions {\n maxLineLength?: number;\n batchSize?: number;\n filter?: (obj: any) => boolean;\n transform?: (obj: any) => any;\n onProgress?: (processed: number, total?: number) => void;\n}\n\nexport class StreamingJSONLParser {\n private readonly DEFAULT_BATCH_SIZE = 100;\n private readonly DEFAULT_MAX_LINE_LENGTH = 1024 * 1024; // 1MB per line\n\n /**\n * Stream-parse a JSONL file with batching and backpressure handling\n */\n async* parseStream<T = any>(\n filePath: string,\n options: ParseOptions = {}\n ): AsyncGenerator<T[], void, unknown> {\n const {\n batchSize = this.DEFAULT_BATCH_SIZE,\n maxLineLength = this.DEFAULT_MAX_LINE_LENGTH,\n filter,\n transform,\n onProgress,\n } = options;\n\n const stream = createReadStream(filePath, {\n encoding: 'utf8',\n highWaterMark: 64 * 1024, // 64KB chunks\n });\n\n const rl = createInterface({\n input: stream,\n crlfDelay: Infinity,\n historySize: 0, // Disable history for memory efficiency\n });\n\n let batch: T[] = [];\n let lineCount = 0;\n let processedCount = 0;\n let errorCount = 0;\n\n try {\n for await (const line of rl) {\n lineCount++;\n\n if (line.length > maxLineLength) {\n logger.warn('Skipping oversized line', {\n lineNumber: lineCount,\n length: line.length,\n maxLength: maxLineLength,\n });\n errorCount++;\n continue;\n }\n\n if (!line.trim()) continue;\n\n try {\n let obj = JSON.parse(line);\n\n if (filter && !filter(obj)) continue;\n if (transform) obj = transform(obj);\n\n batch.push(obj as T);\n processedCount++;\n\n if (batch.length >= batchSize) {\n yield batch;\n batch = [];\n onProgress?.(processedCount);\n }\n } catch (parseError) {\n errorCount++;\n logger.debug('Failed to parse JSONL line', {\n lineNumber: lineCount,\n error: parseError,\n preview: line.substring(0, 100),\n });\n }\n }\n\n // Yield remaining items\n if (batch.length > 0) {\n yield batch;\n onProgress?.(processedCount);\n }\n } finally {\n rl.close();\n stream.destroy();\n\n logger.debug('JSONL parsing complete', {\n filePath,\n totalLines: lineCount,\n processed: processedCount,\n errors: errorCount,\n });\n }\n }\n\n /**\n * Parse entire file into memory (use for smaller files)\n */\n async parseAll<T = any>(\n filePath: string,\n options: Omit<ParseOptions, 'batchSize'> = {}\n ): Promise<T[]> {\n const results: T[] = [];\n\n for await (const batch of this.parseStream<T>(filePath, {\n ...options,\n batchSize: Number.MAX_SAFE_INTEGER,\n })) {\n results.push(...batch);\n }\n\n return results;\n }\n\n /**\n * Process JSONL file with a custom processor function\n */\n async process<T = any, R = void>(\n filePath: string,\n processor: (items: T[]) => Promise<R>,\n options: ParseOptions = {}\n ): Promise<R[]> {\n const results: R[] = [];\n\n for await (const batch of this.parseStream<T>(filePath, options)) {\n const result = await processor(batch);\n results.push(result);\n }\n\n return results;\n }\n\n /**\n * Create a transform stream for JSONL parsing\n */\n createTransformStream<T = any>(\n options: ParseOptions = {}\n ): Transform {\n const { filter, transform, maxLineLength = this.DEFAULT_MAX_LINE_LENGTH } = options;\n let buffer = '';\n let lineCount = 0;\n\n return new Transform({\n objectMode: true,\n transform(chunk: Buffer | string, encoding, callback) {\n buffer += chunk.toString();\n const lines = buffer.split('\\n');\n \n // Keep incomplete line in buffer\n buffer = lines.pop() || '';\n\n for (const line of lines) {\n lineCount++;\n \n if (!line.trim()) continue;\n if (line.length > maxLineLength) {\n logger.warn('Skipping oversized line in transform', { lineCount });\n continue;\n }\n\n try {\n let obj = JSON.parse(line);\n \n if (filter && !filter(obj)) continue;\n if (transform) obj = transform(obj);\n \n this.push(obj);\n } catch (error) {\n logger.debug('Transform parse error', { lineCount, error });\n }\n }\n\n callback();\n },\n\n flush(callback) {\n // Process any remaining data\n if (buffer.trim()) {\n try {\n let obj = JSON.parse(buffer);\n if (!filter || filter(obj)) {\n if (transform) obj = transform(obj);\n this.push(obj);\n }\n } catch (error) {\n logger.debug('Flush parse error', { error });\n }\n }\n callback();\n }\n });\n }\n\n /**\n * Count lines in JSONL file without parsing\n */\n async countLines(filePath: string): Promise<number> {\n const stream = createReadStream(filePath, { encoding: 'utf8' });\n const rl = createInterface({ input: stream, historySize: 0 });\n \n let count = 0;\n for await (const _ of rl) {\n count++;\n }\n \n return count;\n }\n\n /**\n * Sample random lines from JSONL file\n */\n async* sampleLines<T = any>(\n filePath: string,\n sampleRate: number,\n options: Omit<ParseOptions, 'batchSize'> = {}\n ): AsyncGenerator<T, void, unknown> {\n if (sampleRate <= 0 || sampleRate > 1) {\n throw new Error('Sample rate must be between 0 and 1');\n }\n\n for await (const batch of this.parseStream<T>(filePath, options)) {\n for (const item of batch) {\n if (Math.random() < sampleRate) {\n yield item;\n }\n }\n }\n }\n}"],
|
|
5
|
-
"mappings": "AAKA,SAAS,wBAAwB;AACjC,SAAS,uBAAuB;AAChC,SAAS,WAAW,gBAAgB;AACpC,SAAS,iBAAiB;AAC1B,SAAS,cAAc;AAEvB,MAAM,gBAAgB,UAAU,QAAQ;AAUjC,MAAM,qBAAqB;AAAA,EACf,qBAAqB;AAAA,EACrB,0BAA0B,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAKlD,OAAO,YACL,UACA,UAAwB,CAAC,GACW;AACpC,UAAM;AAAA,MACJ,YAAY,KAAK;AAAA,MACjB,gBAAgB,KAAK;AAAA,MACrB;AAAA,MACA;AAAA,MACA;AAAA,IACF,IAAI;AAEJ,UAAM,SAAS,iBAAiB,UAAU;AAAA,MACxC,UAAU;AAAA,MACV,eAAe,KAAK;AAAA;AAAA,IACtB,CAAC;AAED,UAAM,KAAK,gBAAgB;AAAA,MACzB,OAAO;AAAA,MACP,WAAW;AAAA,MACX,aAAa;AAAA;AAAA,IACf,CAAC;AAED,QAAI,QAAa,CAAC;AAClB,QAAI,YAAY;AAChB,QAAI,iBAAiB;AACrB,QAAI,aAAa;AAEjB,QAAI;AACF,uBAAiB,QAAQ,IAAI;AAC3B;AAEA,YAAI,KAAK,SAAS,eAAe;AAC/B,iBAAO,KAAK,2BAA2B;AAAA,YACrC,YAAY;AAAA,YACZ,QAAQ,KAAK;AAAA,YACb,WAAW;AAAA,UACb,CAAC;AACD;AACA;AAAA,QACF;AAEA,YAAI,CAAC,KAAK,KAAK,EAAG;AAElB,YAAI;AACF,cAAI,MAAM,KAAK,MAAM,IAAI;AAEzB,cAAI,UAAU,CAAC,OAAO,GAAG,EAAG;AAC5B,cAAI,UAAW,OAAM,UAAU,GAAG;AAElC,gBAAM,KAAK,GAAQ;AACnB;AAEA,cAAI,MAAM,UAAU,WAAW;AAC7B,kBAAM;AACN,oBAAQ,CAAC;AACT,yBAAa,cAAc;AAAA,UAC7B;AAAA,QACF,SAAS,
|
|
4
|
+
"sourcesContent": ["/**\n * Optimized Streaming JSONL Parser\n * Memory-efficient parsing for large JSONL files with async streaming\n */\n\nimport { createReadStream } from 'fs';\nimport { createInterface } from 'readline';\nimport { Transform, pipeline } from 'stream';\nimport { promisify } from 'util';\nimport { logger } from '../monitoring/logger.js';\n\nconst pipelineAsync = promisify(pipeline);\n\nexport interface ParseOptions {\n maxLineLength?: number;\n batchSize?: number;\n filter?: (obj: any) => boolean;\n transform?: (obj: any) => any;\n onProgress?: (processed: number, total?: number) => void;\n}\n\nexport class StreamingJSONLParser {\n private readonly DEFAULT_BATCH_SIZE = 100;\n private readonly DEFAULT_MAX_LINE_LENGTH = 1024 * 1024; // 1MB per line\n\n /**\n * Stream-parse a JSONL file with batching and backpressure handling\n */\n async* parseStream<T = any>(\n filePath: string,\n options: ParseOptions = {}\n ): AsyncGenerator<T[], void, unknown> {\n const {\n batchSize = this.DEFAULT_BATCH_SIZE,\n maxLineLength = this.DEFAULT_MAX_LINE_LENGTH,\n filter,\n transform,\n onProgress,\n } = options;\n\n const stream = createReadStream(filePath, {\n encoding: 'utf8',\n highWaterMark: 64 * 1024, // 64KB chunks\n });\n\n const rl = createInterface({\n input: stream,\n crlfDelay: Infinity,\n historySize: 0, // Disable history for memory efficiency\n });\n\n let batch: T[] = [];\n let lineCount = 0;\n let processedCount = 0;\n let errorCount = 0;\n\n try {\n for await (const line of rl) {\n lineCount++;\n\n if (line.length > maxLineLength) {\n logger.warn('Skipping oversized line', {\n lineNumber: lineCount,\n length: line.length,\n maxLength: maxLineLength,\n });\n errorCount++;\n continue;\n }\n\n if (!line.trim()) continue;\n\n try {\n let obj = JSON.parse(line);\n\n if (filter && !filter(obj)) continue;\n if (transform) obj = transform(obj);\n\n batch.push(obj as T);\n processedCount++;\n\n if (batch.length >= batchSize) {\n yield batch;\n batch = [];\n onProgress?.(processedCount);\n }\n } catch (parseError: unknown) {\n errorCount++;\n logger.debug('Failed to parse JSONL line', {\n lineNumber: lineCount,\n error: parseError,\n preview: line.substring(0, 100),\n });\n }\n }\n\n // Yield remaining items\n if (batch.length > 0) {\n yield batch;\n onProgress?.(processedCount);\n }\n } finally {\n rl.close();\n stream.destroy();\n\n logger.debug('JSONL parsing complete', {\n filePath,\n totalLines: lineCount,\n processed: processedCount,\n errors: errorCount,\n });\n }\n }\n\n /**\n * Parse entire file into memory (use for smaller files)\n */\n async parseAll<T = any>(\n filePath: string,\n options: Omit<ParseOptions, 'batchSize'> = {}\n ): Promise<T[]> {\n const results: T[] = [];\n\n for await (const batch of this.parseStream<T>(filePath, {\n ...options,\n batchSize: Number.MAX_SAFE_INTEGER,\n })) {\n results.push(...batch);\n }\n\n return results;\n }\n\n /**\n * Process JSONL file with a custom processor function\n */\n async process<T = any, R = void>(\n filePath: string,\n processor: (items: T[]) => Promise<R>,\n options: ParseOptions = {}\n ): Promise<R[]> {\n const results: R[] = [];\n\n for await (const batch of this.parseStream<T>(filePath, options)) {\n const result = await processor(batch);\n results.push(result);\n }\n\n return results;\n }\n\n /**\n * Create a transform stream for JSONL parsing\n */\n createTransformStream<T = any>(\n options: ParseOptions = {}\n ): Transform {\n const { filter, transform, maxLineLength = this.DEFAULT_MAX_LINE_LENGTH } = options;\n let buffer = '';\n let lineCount = 0;\n\n return new Transform({\n objectMode: true,\n transform(chunk: Buffer | string, encoding, callback) {\n buffer += chunk.toString();\n const lines = buffer.split('\\n');\n \n // Keep incomplete line in buffer\n buffer = lines.pop() || '';\n\n for (const line of lines) {\n lineCount++;\n \n if (!line.trim()) continue;\n if (line.length > maxLineLength) {\n logger.warn('Skipping oversized line in transform', { lineCount });\n continue;\n }\n\n try {\n let obj = JSON.parse(line);\n \n if (filter && !filter(obj)) continue;\n if (transform) obj = transform(obj);\n \n this.push(obj);\n } catch (error: unknown) {\n logger.debug('Transform parse error', { lineCount, error });\n }\n }\n\n callback();\n },\n\n flush(callback) {\n // Process any remaining data\n if (buffer.trim()) {\n try {\n let obj = JSON.parse(buffer);\n if (!filter || filter(obj)) {\n if (transform) obj = transform(obj);\n this.push(obj);\n }\n } catch (error: unknown) {\n logger.debug('Flush parse error', { error });\n }\n }\n callback();\n }\n });\n }\n\n /**\n * Count lines in JSONL file without parsing\n */\n async countLines(filePath: string): Promise<number> {\n const stream = createReadStream(filePath, { encoding: 'utf8' });\n const rl = createInterface({ input: stream, historySize: 0 });\n \n let count = 0;\n for await (const _ of rl) {\n count++;\n }\n \n return count;\n }\n\n /**\n * Sample random lines from JSONL file\n */\n async* sampleLines<T = any>(\n filePath: string,\n sampleRate: number,\n options: Omit<ParseOptions, 'batchSize'> = {}\n ): AsyncGenerator<T, void, unknown> {\n if (sampleRate <= 0 || sampleRate > 1) {\n throw new Error('Sample rate must be between 0 and 1');\n }\n\n for await (const batch of this.parseStream<T>(filePath, options)) {\n for (const item of batch) {\n if (Math.random() < sampleRate) {\n yield item;\n }\n }\n }\n }\n}"],
|
|
5
|
+
"mappings": "AAKA,SAAS,wBAAwB;AACjC,SAAS,uBAAuB;AAChC,SAAS,WAAW,gBAAgB;AACpC,SAAS,iBAAiB;AAC1B,SAAS,cAAc;AAEvB,MAAM,gBAAgB,UAAU,QAAQ;AAUjC,MAAM,qBAAqB;AAAA,EACf,qBAAqB;AAAA,EACrB,0BAA0B,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAKlD,OAAO,YACL,UACA,UAAwB,CAAC,GACW;AACpC,UAAM;AAAA,MACJ,YAAY,KAAK;AAAA,MACjB,gBAAgB,KAAK;AAAA,MACrB;AAAA,MACA;AAAA,MACA;AAAA,IACF,IAAI;AAEJ,UAAM,SAAS,iBAAiB,UAAU;AAAA,MACxC,UAAU;AAAA,MACV,eAAe,KAAK;AAAA;AAAA,IACtB,CAAC;AAED,UAAM,KAAK,gBAAgB;AAAA,MACzB,OAAO;AAAA,MACP,WAAW;AAAA,MACX,aAAa;AAAA;AAAA,IACf,CAAC;AAED,QAAI,QAAa,CAAC;AAClB,QAAI,YAAY;AAChB,QAAI,iBAAiB;AACrB,QAAI,aAAa;AAEjB,QAAI;AACF,uBAAiB,QAAQ,IAAI;AAC3B;AAEA,YAAI,KAAK,SAAS,eAAe;AAC/B,iBAAO,KAAK,2BAA2B;AAAA,YACrC,YAAY;AAAA,YACZ,QAAQ,KAAK;AAAA,YACb,WAAW;AAAA,UACb,CAAC;AACD;AACA;AAAA,QACF;AAEA,YAAI,CAAC,KAAK,KAAK,EAAG;AAElB,YAAI;AACF,cAAI,MAAM,KAAK,MAAM,IAAI;AAEzB,cAAI,UAAU,CAAC,OAAO,GAAG,EAAG;AAC5B,cAAI,UAAW,OAAM,UAAU,GAAG;AAElC,gBAAM,KAAK,GAAQ;AACnB;AAEA,cAAI,MAAM,UAAU,WAAW;AAC7B,kBAAM;AACN,oBAAQ,CAAC;AACT,yBAAa,cAAc;AAAA,UAC7B;AAAA,QACF,SAAS,YAAqB;AAC5B;AACA,iBAAO,MAAM,8BAA8B;AAAA,YACzC,YAAY;AAAA,YACZ,OAAO;AAAA,YACP,SAAS,KAAK,UAAU,GAAG,GAAG;AAAA,UAChC,CAAC;AAAA,QACH;AAAA,MACF;AAGA,UAAI,MAAM,SAAS,GAAG;AACpB,cAAM;AACN,qBAAa,cAAc;AAAA,MAC7B;AAAA,IACF,UAAE;AACA,SAAG,MAAM;AACT,aAAO,QAAQ;AAEf,aAAO,MAAM,0BAA0B;AAAA,QACrC;AAAA,QACA,YAAY;AAAA,QACZ,WAAW;AAAA,QACX,QAAQ;AAAA,MACV,CAAC;AAAA,IACH;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,SACJ,UACA,UAA2C,CAAC,GAC9B;AACd,UAAM,UAAe,CAAC;AAEtB,qBAAiB,SAAS,KAAK,YAAe,UAAU;AAAA,MACtD,GAAG;AAAA,MACH,WAAW,OAAO;AAAA,IACpB,CAAC,GAAG;AACF,cAAQ,KAAK,GAAG,KAAK;AAAA,IACvB;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,QACJ,UACA,WACA,UAAwB,CAAC,GACX;AACd,UAAM,UAAe,CAAC;AAEtB,qBAAiB,SAAS,KAAK,YAAe,UAAU,OAAO,GAAG;AAChE,YAAM,SAAS,MAAM,UAAU,KAAK;AACpC,cAAQ,KAAK,MAAM;AAAA,IACrB;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,sBACE,UAAwB,CAAC,GACd;AACX,UAAM,EAAE,QAAQ,WAAW,gBAAgB,KAAK,wBAAwB,IAAI;AAC5E,QAAI,SAAS;AACb,QAAI,YAAY;AAEhB,WAAO,IAAI,UAAU;AAAA,MACnB,YAAY;AAAA,MACZ,UAAU,OAAwB,UAAU,UAAU;AACpD,kBAAU,MAAM,SAAS;AACzB,cAAM,QAAQ,OAAO,MAAM,IAAI;AAG/B,iBAAS,MAAM,IAAI,KAAK;AAExB,mBAAW,QAAQ,OAAO;AACxB;AAEA,cAAI,CAAC,KAAK,KAAK,EAAG;AAClB,cAAI,KAAK,SAAS,eAAe;AAC/B,mBAAO,KAAK,wCAAwC,EAAE,UAAU,CAAC;AACjE;AAAA,UACF;AAEA,cAAI;AACF,gBAAI,MAAM,KAAK,MAAM,IAAI;AAEzB,gBAAI,UAAU,CAAC,OAAO,GAAG,EAAG;AAC5B,gBAAI,UAAW,OAAM,UAAU,GAAG;AAElC,iBAAK,KAAK,GAAG;AAAA,UACf,SAAS,OAAgB;AACvB,mBAAO,MAAM,yBAAyB,EAAE,WAAW,MAAM,CAAC;AAAA,UAC5D;AAAA,QACF;AAEA,iBAAS;AAAA,MACX;AAAA,MAEA,MAAM,UAAU;AAEd,YAAI,OAAO,KAAK,GAAG;AACjB,cAAI;AACF,gBAAI,MAAM,KAAK,MAAM,MAAM;AAC3B,gBAAI,CAAC,UAAU,OAAO,GAAG,GAAG;AAC1B,kBAAI,UAAW,OAAM,UAAU,GAAG;AAClC,mBAAK,KAAK,GAAG;AAAA,YACf;AAAA,UACF,SAAS,OAAgB;AACvB,mBAAO,MAAM,qBAAqB,EAAE,MAAM,CAAC;AAAA,UAC7C;AAAA,QACF;AACA,iBAAS;AAAA,MACX;AAAA,IACF,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,WAAW,UAAmC;AAClD,UAAM,SAAS,iBAAiB,UAAU,EAAE,UAAU,OAAO,CAAC;AAC9D,UAAM,KAAK,gBAAgB,EAAE,OAAO,QAAQ,aAAa,EAAE,CAAC;AAE5D,QAAI,QAAQ;AACZ,qBAAiB,KAAK,IAAI;AACxB;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,OAAO,YACL,UACA,YACA,UAA2C,CAAC,GACV;AAClC,QAAI,cAAc,KAAK,aAAa,GAAG;AACrC,YAAM,IAAI,MAAM,qCAAqC;AAAA,IACvD;AAEA,qBAAiB,SAAS,KAAK,YAAe,UAAU,OAAO,GAAG;AAChE,iBAAW,QAAQ,OAAO;AACxB,YAAI,KAAK,OAAO,IAAI,YAAY;AAC9B,gBAAM;AAAA,QACR;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/persistence/postgres-adapter.ts"],
|
|
4
|
-
"sourcesContent": ["import { Pool, PoolConfig, QueryResult as PgQueryResult } from 'pg';\nimport { Database } from 'better-sqlite3';\nimport {\n PersistenceAdapter,\n QueryResult,\n TraceData,\n ContextData,\n} from '../types.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport interface PostgresConfig extends PoolConfig {\n enableTimescale?: boolean;\n enablePgvector?: boolean;\n vectorDimensions?: number;\n}\n\nexport class PostgresAdapter implements PersistenceAdapter {\n private pool: Pool;\n private config: PostgresConfig;\n private isInitialized = false;\n\n constructor(config: PostgresConfig) {\n this.config = {\n ...config,\n vectorDimensions: config.vectorDimensions || 1536, // OpenAI ada-002 dimensions\n };\n this.pool = new Pool(this.config);\n }\n\n async connect(): Promise<void> {\n try {\n await this.pool.connect();\n await this.initialize();\n this.isInitialized = true;\n logger.info('PostgreSQL connected successfully');\n } catch (error) {\n logger.error(\n 'Failed to connect to PostgreSQL',\n error instanceof Error ? error : undefined\n );\n throw error;\n }\n }\n\n async disconnect(): Promise<void> {\n await this.pool.end();\n this.isInitialized = false;\n logger.info('PostgreSQL disconnected');\n }\n\n async execute(query: string, params?: any[]): Promise<QueryResult> {\n try {\n const result: PgQueryResult = await this.pool.query(query, params);\n return {\n rows: result.rows,\n rowCount: result.rowCount || 0,\n fields: result.fields?.map((f) => ({\n name: f.name,\n type: f.dataTypeID.toString(),\n })),\n };\n } catch (error) {\n logger.error(\n 'Query execution failed',\n error instanceof Error ? error : new Error(String(error)),\n { query }\n );\n throw error;\n }\n }\n\n async beginTransaction(): Promise<void> {\n await this.execute('BEGIN');\n }\n\n async commit(): Promise<void> {\n await this.execute('COMMIT');\n }\n\n async rollback(): Promise<void> {\n await this.execute('ROLLBACK');\n }\n\n isConnected(): boolean {\n return this.isInitialized;\n }\n\n private async initialize(): Promise<void> {\n // Create base schema\n await this.createBaseSchema();\n\n // Enable extensions if configured\n if (this.config.enableTimescale) {\n await this.enableTimescale();\n }\n if (this.config.enablePgvector) {\n await this.enablePgvector();\n }\n }\n\n private async createBaseSchema(): Promise<void> {\n const queries = [\n // Projects table\n `CREATE TABLE IF NOT EXISTS projects (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n name VARCHAR(255) NOT NULL,\n path TEXT NOT NULL,\n created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n updated_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n metadata JSONB\n )`,\n\n // Sessions table\n `CREATE TABLE IF NOT EXISTS sessions (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n branch VARCHAR(255),\n started_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n ended_at TIMESTAMP,\n metadata JSONB\n )`,\n\n // Traces table\n `CREATE TABLE IF NOT EXISTS traces (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n session_id UUID REFERENCES sessions(id) ON DELETE CASCADE,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n type VARCHAR(100) NOT NULL,\n data JSONB NOT NULL,\n metadata JSONB\n )`,\n\n // Context frames table\n `CREATE TABLE IF NOT EXISTS context_frames (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n branch VARCHAR(255),\n content TEXT NOT NULL,\n summary TEXT,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n type VARCHAR(100) NOT NULL,\n metadata JSONB\n )`,\n\n // Decisions table\n `CREATE TABLE IF NOT EXISTS decisions (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n session_id UUID REFERENCES sessions(id) ON DELETE CASCADE,\n decision TEXT NOT NULL,\n rationale TEXT,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n metadata JSONB\n )`,\n ];\n\n for (const query of queries) {\n await this.execute(query);\n }\n\n // Create indexes separately (PostgreSQL doesn't support inline INDEX in CREATE TABLE)\n const indexes = [\n `CREATE INDEX IF NOT EXISTS idx_traces_session_timestamp \n ON traces(session_id, timestamp)`,\n `CREATE INDEX IF NOT EXISTS idx_context_project_branch \n ON context_frames(project_id, branch)`,\n `CREATE INDEX IF NOT EXISTS idx_traces_type \n ON traces(type)`,\n `CREATE INDEX IF NOT EXISTS idx_context_frames_timestamp \n ON context_frames(timestamp)`,\n `CREATE INDEX IF NOT EXISTS idx_decisions_project_session \n ON decisions(project_id, session_id)`,\n ];\n\n for (const index of indexes) {\n await this.execute(index);\n }\n }\n\n private async enableTimescale(): Promise<void> {\n try {\n await this.execute('CREATE EXTENSION IF NOT EXISTS timescaledb');\n\n // Convert traces to hypertable\n await this.execute(`\n SELECT create_hypertable('traces', 'timestamp',\n if_not_exists => TRUE,\n chunk_time_interval => INTERVAL '1 day'\n )\n `);\n\n // Convert context_frames to hypertable\n await this.execute(`\n SELECT create_hypertable('context_frames', 'timestamp',\n if_not_exists => TRUE,\n chunk_time_interval => INTERVAL '7 days'\n )\n `);\n\n logger.info('TimescaleDB extension enabled');\n } catch (error) {\n logger.warn(\n 'Failed to enable TimescaleDB',\n error instanceof Error ? error : undefined\n );\n }\n }\n\n private async enablePgvector(): Promise<void> {\n try {\n await this.execute('CREATE EXTENSION IF NOT EXISTS vector');\n\n // Add embedding columns\n await this.execute(`\n ALTER TABLE context_frames \n ADD COLUMN IF NOT EXISTS embedding vector(${this.config.vectorDimensions})\n `);\n\n await this.execute(`\n ALTER TABLE traces \n ADD COLUMN IF NOT EXISTS embedding vector(${this.config.vectorDimensions})\n `);\n\n // Create vector indexes\n await this.execute(`\n CREATE INDEX IF NOT EXISTS idx_context_embedding \n ON context_frames \n USING ivfflat (embedding vector_cosine_ops)\n WITH (lists = 100)\n `);\n\n logger.info('pgvector extension enabled');\n } catch (error) {\n logger.warn(\n 'Failed to enable pgvector',\n error instanceof Error ? error : undefined\n );\n }\n }\n\n // Data access methods\n async saveTrace(trace: TraceData): Promise<void> {\n await this.execute(\n `INSERT INTO traces (id, session_id, timestamp, type, data, metadata)\n VALUES ($1, $2, $3, $4, $5, $6)`,\n [\n trace.id,\n trace.sessionId,\n trace.timestamp,\n trace.type,\n JSON.stringify(trace.data),\n trace.metadata ? JSON.stringify(trace.metadata) : null,\n ]\n );\n }\n\n async saveContext(context: ContextData): Promise<void> {\n await this.execute(\n `INSERT INTO context_frames (id, project_id, branch, content, timestamp, type, metadata)\n VALUES ($1, $2, $3, $4, $5, $6, $7)`,\n [\n context.id,\n context.projectId,\n context.branch || null,\n context.content,\n context.timestamp,\n context.type,\n context.metadata ? JSON.stringify(context.metadata) : null,\n ]\n );\n }\n\n async getRecentTraces(sessionId: string, limit = 100): Promise<TraceData[]> {\n const result = await this.execute(\n `SELECT * FROM traces \n WHERE session_id = $1 \n ORDER BY timestamp DESC \n LIMIT $2`,\n [sessionId, limit]\n );\n\n return result.rows.map((row: any) => ({\n id: row.id,\n sessionId: row.session_id,\n timestamp: row.timestamp,\n type: row.type,\n data: row.data,\n metadata: row.metadata,\n }));\n }\n\n async getRecentContext(\n projectId: string,\n branch?: string,\n limit = 50\n ): Promise<ContextData[]> {\n const query = branch\n ? `SELECT * FROM context_frames \n WHERE project_id = $1 AND branch = $2 \n ORDER BY timestamp DESC \n LIMIT $3`\n : `SELECT * FROM context_frames \n WHERE project_id = $1 \n ORDER BY timestamp DESC \n LIMIT $2`;\n\n const params = branch ? [projectId, branch, limit] : [projectId, limit];\n const result = await this.execute(query, params);\n\n return result.rows.map((row: any) => ({\n id: row.id,\n projectId: row.project_id,\n branch: row.branch,\n content: row.content,\n timestamp: row.timestamp,\n type: row.type,\n metadata: row.metadata,\n }));\n }\n\n // Hybrid SQLite/PostgreSQL migration helper\n async migrateFromSQLite(sqliteDb: Database): Promise<void> {\n logger.info('Starting migration from SQLite to PostgreSQL');\n\n try {\n await this.beginTransaction();\n\n // Migrate projects\n const projects = sqliteDb\n .prepare('SELECT * FROM projects')\n .all() as any[];\n for (const project of projects) {\n await this.execute(\n 'INSERT INTO projects (id, name, path, created_at, updated_at, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n project.id,\n project.name,\n project.path,\n project.created_at,\n project.updated_at,\n project.metadata,\n ]\n );\n }\n\n // Migrate sessions\n const sessions = sqliteDb\n .prepare('SELECT * FROM sessions')\n .all() as any[];\n for (const session of sessions) {\n await this.execute(\n 'INSERT INTO sessions (id, project_id, branch, started_at, ended_at, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n session.id,\n session.project_id,\n session.branch,\n session.started_at,\n session.ended_at,\n session.metadata,\n ]\n );\n }\n\n // Migrate traces in batches\n const traceCount = sqliteDb\n .prepare('SELECT COUNT(*) as count FROM traces')\n .get() as { count: number };\n const batchSize = 1000;\n\n for (let offset = 0; offset < traceCount.count; offset += batchSize) {\n const traces = sqliteDb\n .prepare('SELECT * FROM traces LIMIT ? OFFSET ?')\n .all(batchSize, offset) as any[];\n\n for (const trace of traces) {\n await this.execute(\n 'INSERT INTO traces (id, session_id, timestamp, type, data, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n trace.id,\n trace.session_id,\n trace.timestamp,\n trace.type,\n trace.data,\n trace.metadata,\n ]\n );\n }\n\n logger.info(\n `Migrated ${offset + traces.length}/${traceCount.count} traces`\n );\n }\n\n // Migrate context frames\n const contexts = sqliteDb\n .prepare('SELECT * FROM context_frames')\n .all() as any[];\n for (const context of contexts) {\n await this.execute(\n 'INSERT INTO context_frames (id, project_id, branch, content, summary, timestamp, type, metadata) VALUES ($1, $2, $3, $4, $5, $6, $7, $8)',\n [\n context.id,\n context.project_id,\n context.branch,\n context.content,\n context.summary,\n context.timestamp,\n context.type,\n context.metadata,\n ]\n );\n }\n\n await this.commit();\n logger.info('Migration completed successfully');\n } catch (error) {\n await this.rollback();\n logger.error(\n 'Migration failed',\n error instanceof Error ? error : undefined\n );\n throw error;\n }\n }\n}\n"],
|
|
5
|
-
"mappings": "AAAA,SAAS,YAAsD;AAQ/D,SAAS,cAAc;AAQhB,MAAM,gBAA8C;AAAA,EACjD;AAAA,EACA;AAAA,EACA,gBAAgB;AAAA,EAExB,YAAY,QAAwB;AAClC,SAAK,SAAS;AAAA,MACZ,GAAG;AAAA,MACH,kBAAkB,OAAO,oBAAoB;AAAA;AAAA,IAC/C;AACA,SAAK,OAAO,IAAI,KAAK,KAAK,MAAM;AAAA,EAClC;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI;AACF,YAAM,KAAK,KAAK,QAAQ;AACxB,YAAM,KAAK,WAAW;AACtB,WAAK,gBAAgB;AACrB,aAAO,KAAK,mCAAmC;AAAA,IACjD,SAAS,
|
|
4
|
+
"sourcesContent": ["import { Pool, PoolConfig, QueryResult as PgQueryResult } from 'pg';\nimport { Database } from 'better-sqlite3';\nimport {\n PersistenceAdapter,\n QueryResult,\n TraceData,\n ContextData,\n} from '../types.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport interface PostgresConfig extends PoolConfig {\n enableTimescale?: boolean;\n enablePgvector?: boolean;\n vectorDimensions?: number;\n}\n\nexport class PostgresAdapter implements PersistenceAdapter {\n private pool: Pool;\n private config: PostgresConfig;\n private isInitialized = false;\n\n constructor(config: PostgresConfig) {\n this.config = {\n ...config,\n vectorDimensions: config.vectorDimensions || 1536, // OpenAI ada-002 dimensions\n };\n this.pool = new Pool(this.config);\n }\n\n async connect(): Promise<void> {\n try {\n await this.pool.connect();\n await this.initialize();\n this.isInitialized = true;\n logger.info('PostgreSQL connected successfully');\n } catch (error: unknown) {\n logger.error(\n 'Failed to connect to PostgreSQL',\n error instanceof Error ? error : undefined\n );\n throw error;\n }\n }\n\n async disconnect(): Promise<void> {\n await this.pool.end();\n this.isInitialized = false;\n logger.info('PostgreSQL disconnected');\n }\n\n async execute(query: string, params?: any[]): Promise<QueryResult> {\n try {\n const result: PgQueryResult = await this.pool.query(query, params);\n return {\n rows: result.rows,\n rowCount: result.rowCount || 0,\n fields: result.fields?.map((f) => ({\n name: f.name,\n type: f.dataTypeID.toString(),\n })),\n };\n } catch (error: unknown) {\n logger.error(\n 'Query execution failed',\n error instanceof Error ? error : new Error(String(error)),\n { query }\n );\n throw error;\n }\n }\n\n async beginTransaction(): Promise<void> {\n await this.execute('BEGIN');\n }\n\n async commit(): Promise<void> {\n await this.execute('COMMIT');\n }\n\n async rollback(): Promise<void> {\n await this.execute('ROLLBACK');\n }\n\n isConnected(): boolean {\n return this.isInitialized;\n }\n\n private async initialize(): Promise<void> {\n // Create base schema\n await this.createBaseSchema();\n\n // Enable extensions if configured\n if (this.config.enableTimescale) {\n await this.enableTimescale();\n }\n if (this.config.enablePgvector) {\n await this.enablePgvector();\n }\n }\n\n private async createBaseSchema(): Promise<void> {\n const queries = [\n // Projects table\n `CREATE TABLE IF NOT EXISTS projects (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n name VARCHAR(255) NOT NULL,\n path TEXT NOT NULL,\n created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n updated_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n metadata JSONB\n )`,\n\n // Sessions table\n `CREATE TABLE IF NOT EXISTS sessions (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n branch VARCHAR(255),\n started_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n ended_at TIMESTAMP,\n metadata JSONB\n )`,\n\n // Traces table\n `CREATE TABLE IF NOT EXISTS traces (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n session_id UUID REFERENCES sessions(id) ON DELETE CASCADE,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n type VARCHAR(100) NOT NULL,\n data JSONB NOT NULL,\n metadata JSONB\n )`,\n\n // Context frames table\n `CREATE TABLE IF NOT EXISTS context_frames (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n branch VARCHAR(255),\n content TEXT NOT NULL,\n summary TEXT,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n type VARCHAR(100) NOT NULL,\n metadata JSONB\n )`,\n\n // Decisions table\n `CREATE TABLE IF NOT EXISTS decisions (\n id UUID PRIMARY KEY DEFAULT gen_random_uuid(),\n project_id UUID REFERENCES projects(id) ON DELETE CASCADE,\n session_id UUID REFERENCES sessions(id) ON DELETE CASCADE,\n decision TEXT NOT NULL,\n rationale TEXT,\n timestamp TIMESTAMP DEFAULT CURRENT_TIMESTAMP,\n metadata JSONB\n )`,\n ];\n\n for (const query of queries) {\n await this.execute(query);\n }\n\n // Create indexes separately (PostgreSQL doesn't support inline INDEX in CREATE TABLE)\n const indexes = [\n `CREATE INDEX IF NOT EXISTS idx_traces_session_timestamp \n ON traces(session_id, timestamp)`,\n `CREATE INDEX IF NOT EXISTS idx_context_project_branch \n ON context_frames(project_id, branch)`,\n `CREATE INDEX IF NOT EXISTS idx_traces_type \n ON traces(type)`,\n `CREATE INDEX IF NOT EXISTS idx_context_frames_timestamp \n ON context_frames(timestamp)`,\n `CREATE INDEX IF NOT EXISTS idx_decisions_project_session \n ON decisions(project_id, session_id)`,\n ];\n\n for (const index of indexes) {\n await this.execute(index);\n }\n }\n\n private async enableTimescale(): Promise<void> {\n try {\n await this.execute('CREATE EXTENSION IF NOT EXISTS timescaledb');\n\n // Convert traces to hypertable\n await this.execute(`\n SELECT create_hypertable('traces', 'timestamp',\n if_not_exists => TRUE,\n chunk_time_interval => INTERVAL '1 day'\n )\n `);\n\n // Convert context_frames to hypertable\n await this.execute(`\n SELECT create_hypertable('context_frames', 'timestamp',\n if_not_exists => TRUE,\n chunk_time_interval => INTERVAL '7 days'\n )\n `);\n\n logger.info('TimescaleDB extension enabled');\n } catch (error: unknown) {\n logger.warn(\n 'Failed to enable TimescaleDB',\n error instanceof Error ? error : undefined\n );\n }\n }\n\n private async enablePgvector(): Promise<void> {\n try {\n await this.execute('CREATE EXTENSION IF NOT EXISTS vector');\n\n // Add embedding columns\n await this.execute(`\n ALTER TABLE context_frames \n ADD COLUMN IF NOT EXISTS embedding vector(${this.config.vectorDimensions})\n `);\n\n await this.execute(`\n ALTER TABLE traces \n ADD COLUMN IF NOT EXISTS embedding vector(${this.config.vectorDimensions})\n `);\n\n // Create vector indexes\n await this.execute(`\n CREATE INDEX IF NOT EXISTS idx_context_embedding \n ON context_frames \n USING ivfflat (embedding vector_cosine_ops)\n WITH (lists = 100)\n `);\n\n logger.info('pgvector extension enabled');\n } catch (error: unknown) {\n logger.warn(\n 'Failed to enable pgvector',\n error instanceof Error ? error : undefined\n );\n }\n }\n\n // Data access methods\n async saveTrace(trace: TraceData): Promise<void> {\n await this.execute(\n `INSERT INTO traces (id, session_id, timestamp, type, data, metadata)\n VALUES ($1, $2, $3, $4, $5, $6)`,\n [\n trace.id,\n trace.sessionId,\n trace.timestamp,\n trace.type,\n JSON.stringify(trace.data),\n trace.metadata ? JSON.stringify(trace.metadata) : null,\n ]\n );\n }\n\n async saveContext(context: ContextData): Promise<void> {\n await this.execute(\n `INSERT INTO context_frames (id, project_id, branch, content, timestamp, type, metadata)\n VALUES ($1, $2, $3, $4, $5, $6, $7)`,\n [\n context.id,\n context.projectId,\n context.branch || null,\n context.content,\n context.timestamp,\n context.type,\n context.metadata ? JSON.stringify(context.metadata) : null,\n ]\n );\n }\n\n async getRecentTraces(sessionId: string, limit = 100): Promise<TraceData[]> {\n const result = await this.execute(\n `SELECT * FROM traces \n WHERE session_id = $1 \n ORDER BY timestamp DESC \n LIMIT $2`,\n [sessionId, limit]\n );\n\n return result.rows.map((row: any) => ({\n id: row.id,\n sessionId: row.session_id,\n timestamp: row.timestamp,\n type: row.type,\n data: row.data,\n metadata: row.metadata,\n }));\n }\n\n async getRecentContext(\n projectId: string,\n branch?: string,\n limit = 50\n ): Promise<ContextData[]> {\n const query = branch\n ? `SELECT * FROM context_frames \n WHERE project_id = $1 AND branch = $2 \n ORDER BY timestamp DESC \n LIMIT $3`\n : `SELECT * FROM context_frames \n WHERE project_id = $1 \n ORDER BY timestamp DESC \n LIMIT $2`;\n\n const params = branch ? [projectId, branch, limit] : [projectId, limit];\n const result = await this.execute(query, params);\n\n return result.rows.map((row: any) => ({\n id: row.id,\n projectId: row.project_id,\n branch: row.branch,\n content: row.content,\n timestamp: row.timestamp,\n type: row.type,\n metadata: row.metadata,\n }));\n }\n\n // Hybrid SQLite/PostgreSQL migration helper\n async migrateFromSQLite(sqliteDb: Database): Promise<void> {\n logger.info('Starting migration from SQLite to PostgreSQL');\n\n try {\n await this.beginTransaction();\n\n // Migrate projects\n const projects = sqliteDb\n .prepare('SELECT * FROM projects')\n .all() as any[];\n for (const project of projects) {\n await this.execute(\n 'INSERT INTO projects (id, name, path, created_at, updated_at, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n project.id,\n project.name,\n project.path,\n project.created_at,\n project.updated_at,\n project.metadata,\n ]\n );\n }\n\n // Migrate sessions\n const sessions = sqliteDb\n .prepare('SELECT * FROM sessions')\n .all() as any[];\n for (const session of sessions) {\n await this.execute(\n 'INSERT INTO sessions (id, project_id, branch, started_at, ended_at, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n session.id,\n session.project_id,\n session.branch,\n session.started_at,\n session.ended_at,\n session.metadata,\n ]\n );\n }\n\n // Migrate traces in batches\n const traceCount = sqliteDb\n .prepare('SELECT COUNT(*) as count FROM traces')\n .get() as { count: number };\n const batchSize = 1000;\n\n for (let offset = 0; offset < traceCount.count; offset += batchSize) {\n const traces = sqliteDb\n .prepare('SELECT * FROM traces LIMIT ? OFFSET ?')\n .all(batchSize, offset) as any[];\n\n for (const trace of traces) {\n await this.execute(\n 'INSERT INTO traces (id, session_id, timestamp, type, data, metadata) VALUES ($1, $2, $3, $4, $5, $6)',\n [\n trace.id,\n trace.session_id,\n trace.timestamp,\n trace.type,\n trace.data,\n trace.metadata,\n ]\n );\n }\n\n logger.info(\n `Migrated ${offset + traces.length}/${traceCount.count} traces`\n );\n }\n\n // Migrate context frames\n const contexts = sqliteDb\n .prepare('SELECT * FROM context_frames')\n .all() as any[];\n for (const context of contexts) {\n await this.execute(\n 'INSERT INTO context_frames (id, project_id, branch, content, summary, timestamp, type, metadata) VALUES ($1, $2, $3, $4, $5, $6, $7, $8)',\n [\n context.id,\n context.project_id,\n context.branch,\n context.content,\n context.summary,\n context.timestamp,\n context.type,\n context.metadata,\n ]\n );\n }\n\n await this.commit();\n logger.info('Migration completed successfully');\n } catch (error: unknown) {\n await this.rollback();\n logger.error(\n 'Migration failed',\n error instanceof Error ? error : undefined\n );\n throw error;\n }\n }\n}\n"],
|
|
5
|
+
"mappings": "AAAA,SAAS,YAAsD;AAQ/D,SAAS,cAAc;AAQhB,MAAM,gBAA8C;AAAA,EACjD;AAAA,EACA;AAAA,EACA,gBAAgB;AAAA,EAExB,YAAY,QAAwB;AAClC,SAAK,SAAS;AAAA,MACZ,GAAG;AAAA,MACH,kBAAkB,OAAO,oBAAoB;AAAA;AAAA,IAC/C;AACA,SAAK,OAAO,IAAI,KAAK,KAAK,MAAM;AAAA,EAClC;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI;AACF,YAAM,KAAK,KAAK,QAAQ;AACxB,YAAM,KAAK,WAAW;AACtB,WAAK,gBAAgB;AACrB,aAAO,KAAK,mCAAmC;AAAA,IACjD,SAAS,OAAgB;AACvB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AACA,YAAM;AAAA,IACR;AAAA,EACF;AAAA,EAEA,MAAM,aAA4B;AAChC,UAAM,KAAK,KAAK,IAAI;AACpB,SAAK,gBAAgB;AACrB,WAAO,KAAK,yBAAyB;AAAA,EACvC;AAAA,EAEA,MAAM,QAAQ,OAAe,QAAsC;AACjE,QAAI;AACF,YAAM,SAAwB,MAAM,KAAK,KAAK,MAAM,OAAO,MAAM;AACjE,aAAO;AAAA,QACL,MAAM,OAAO;AAAA,QACb,UAAU,OAAO,YAAY;AAAA,QAC7B,QAAQ,OAAO,QAAQ,IAAI,CAAC,OAAO;AAAA,UACjC,MAAM,EAAE;AAAA,UACR,MAAM,EAAE,WAAW,SAAS;AAAA,QAC9B,EAAE;AAAA,MACJ;AAAA,IACF,SAAS,OAAgB;AACvB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QACxD,EAAE,MAAM;AAAA,MACV;AACA,YAAM;AAAA,IACR;AAAA,EACF;AAAA,EAEA,MAAM,mBAAkC;AACtC,UAAM,KAAK,QAAQ,OAAO;AAAA,EAC5B;AAAA,EAEA,MAAM,SAAwB;AAC5B,UAAM,KAAK,QAAQ,QAAQ;AAAA,EAC7B;AAAA,EAEA,MAAM,WAA0B;AAC9B,UAAM,KAAK,QAAQ,UAAU;AAAA,EAC/B;AAAA,EAEA,cAAuB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,MAAc,aAA4B;AAExC,UAAM,KAAK,iBAAiB;AAG5B,QAAI,KAAK,OAAO,iBAAiB;AAC/B,YAAM,KAAK,gBAAgB;AAAA,IAC7B;AACA,QAAI,KAAK,OAAO,gBAAgB;AAC9B,YAAM,KAAK,eAAe;AAAA,IAC5B;AAAA,EACF;AAAA,EAEA,MAAc,mBAAkC;AAC9C,UAAM,UAAU;AAAA;AAAA,MAEd;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAUA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAUA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAUA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAYA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IASF;AAEA,eAAW,SAAS,SAAS;AAC3B,YAAM,KAAK,QAAQ,KAAK;AAAA,IAC1B;AAGA,UAAM,UAAU;AAAA,MACd;AAAA;AAAA,MAEA;AAAA;AAAA,MAEA;AAAA;AAAA,MAEA;AAAA;AAAA,MAEA;AAAA;AAAA,IAEF;AAEA,eAAW,SAAS,SAAS;AAC3B,YAAM,KAAK,QAAQ,KAAK;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,MAAc,kBAAiC;AAC7C,QAAI;AACF,YAAM,KAAK,QAAQ,4CAA4C;AAG/D,YAAM,KAAK,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAKlB;AAGD,YAAM,KAAK,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAKlB;AAED,aAAO,KAAK,+BAA+B;AAAA,IAC7C,SAAS,OAAgB;AACvB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAc,iBAAgC;AAC5C,QAAI;AACF,YAAM,KAAK,QAAQ,uCAAuC;AAG1D,YAAM,KAAK,QAAQ;AAAA;AAAA,oDAE2B,KAAK,OAAO,gBAAgB;AAAA,OACzE;AAED,YAAM,KAAK,QAAQ;AAAA;AAAA,oDAE2B,KAAK,OAAO,gBAAgB;AAAA,OACzE;AAGD,YAAM,KAAK,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAKlB;AAED,aAAO,KAAK,4BAA4B;AAAA,IAC1C,SAAS,OAAgB;AACvB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA,EAGA,MAAM,UAAU,OAAiC;AAC/C,UAAM,KAAK;AAAA,MACT;AAAA;AAAA,MAEA;AAAA,QACE,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,KAAK,UAAU,MAAM,IAAI;AAAA,QACzB,MAAM,WAAW,KAAK,UAAU,MAAM,QAAQ,IAAI;AAAA,MACpD;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,YAAY,SAAqC;AACrD,UAAM,KAAK;AAAA,MACT;AAAA;AAAA,MAEA;AAAA,QACE,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,UAAU;AAAA,QAClB,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,WAAW,KAAK,UAAU,QAAQ,QAAQ,IAAI;AAAA,MACxD;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,gBAAgB,WAAmB,QAAQ,KAA2B;AAC1E,UAAM,SAAS,MAAM,KAAK;AAAA,MACxB;AAAA;AAAA;AAAA;AAAA,MAIA,CAAC,WAAW,KAAK;AAAA,IACnB;AAEA,WAAO,OAAO,KAAK,IAAI,CAAC,SAAc;AAAA,MACpC,IAAI,IAAI;AAAA,MACR,WAAW,IAAI;AAAA,MACf,WAAW,IAAI;AAAA,MACf,MAAM,IAAI;AAAA,MACV,MAAM,IAAI;AAAA,MACV,UAAU,IAAI;AAAA,IAChB,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,iBACJ,WACA,QACA,QAAQ,IACgB;AACxB,UAAM,QAAQ,SACV;AAAA;AAAA;AAAA,qBAIA;AAAA;AAAA;AAAA;AAKJ,UAAM,SAAS,SAAS,CAAC,WAAW,QAAQ,KAAK,IAAI,CAAC,WAAW,KAAK;AACtE,UAAM,SAAS,MAAM,KAAK,QAAQ,OAAO,MAAM;AAE/C,WAAO,OAAO,KAAK,IAAI,CAAC,SAAc;AAAA,MACpC,IAAI,IAAI;AAAA,MACR,WAAW,IAAI;AAAA,MACf,QAAQ,IAAI;AAAA,MACZ,SAAS,IAAI;AAAA,MACb,WAAW,IAAI;AAAA,MACf,MAAM,IAAI;AAAA,MACV,UAAU,IAAI;AAAA,IAChB,EAAE;AAAA,EACJ;AAAA;AAAA,EAGA,MAAM,kBAAkB,UAAmC;AACzD,WAAO,KAAK,8CAA8C;AAE1D,QAAI;AACF,YAAM,KAAK,iBAAiB;AAG5B,YAAM,WAAW,SACd,QAAQ,wBAAwB,EAChC,IAAI;AACP,iBAAW,WAAW,UAAU;AAC9B,cAAM,KAAK;AAAA,UACT;AAAA,UACA;AAAA,YACE,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV;AAAA,QACF;AAAA,MACF;AAGA,YAAM,WAAW,SACd,QAAQ,wBAAwB,EAChC,IAAI;AACP,iBAAW,WAAW,UAAU;AAC9B,cAAM,KAAK;AAAA,UACT;AAAA,UACA;AAAA,YACE,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV;AAAA,QACF;AAAA,MACF;AAGA,YAAM,aAAa,SAChB,QAAQ,sCAAsC,EAC9C,IAAI;AACP,YAAM,YAAY;AAElB,eAAS,SAAS,GAAG,SAAS,WAAW,OAAO,UAAU,WAAW;AACnE,cAAM,SAAS,SACZ,QAAQ,uCAAuC,EAC/C,IAAI,WAAW,MAAM;AAExB,mBAAW,SAAS,QAAQ;AAC1B,gBAAM,KAAK;AAAA,YACT;AAAA,YACA;AAAA,cACE,MAAM;AAAA,cACN,MAAM;AAAA,cACN,MAAM;AAAA,cACN,MAAM;AAAA,cACN,MAAM;AAAA,cACN,MAAM;AAAA,YACR;AAAA,UACF;AAAA,QACF;AAEA,eAAO;AAAA,UACL,YAAY,SAAS,OAAO,MAAM,IAAI,WAAW,KAAK;AAAA,QACxD;AAAA,MACF;AAGA,YAAM,WAAW,SACd,QAAQ,8BAA8B,EACtC,IAAI;AACP,iBAAW,WAAW,UAAU;AAC9B,cAAM,KAAK;AAAA,UACT;AAAA,UACA;AAAA,YACE,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV;AAAA,QACF;AAAA,MACF;AAEA,YAAM,KAAK,OAAO;AAClB,aAAO,KAAK,kCAAkC;AAAA,IAChD,SAAS,OAAgB;AACvB,YAAM,KAAK,SAAS;AACpB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AACA,YAAM;AAAA,IACR;AAAA,EACF;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/projects/project-manager.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Automatic Project Management for StackMemory\n * Auto-detects and organizes projects based on Git origins\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, mkdirSync, readFileSync, writeFileSync } from 'fs';\nimport { join, basename, dirname } from 'path';\nimport { homedir } from 'os';\nimport Database from 'better-sqlite3';\nimport { logger } from '../monitoring/logger.js';\nimport {\n DatabaseError,\n ProjectError,\n SystemError,\n ErrorCode,\n wrapError,\n createErrorHandler,\n} from '../errors/index.js';\nimport { retry, withTimeout } from '../errors/recovery.js';\n\nexport interface ProjectInfo {\n id: string;\n name: string;\n path: string;\n gitRemote?: string;\n organization?: string;\n accountType: 'personal' | 'work' | 'opensource' | 'client';\n isPrivate: boolean;\n primaryLanguage?: string;\n framework?: string;\n lastAccessed: Date;\n metadata: Record<string, any>;\n}\n\nexport interface OrganizationConfig {\n name: string;\n type: 'company' | 'personal' | 'opensource' | 'client';\n domains: string[];\n githubOrgs: string[];\n gitlabGroups?: string[];\n bitbucketTeams?: string[];\n accountType: 'personal' | 'work' | 'opensource' | 'client';\n autoPatterns?: string[];\n}\n\nexport class ProjectManager {\n private static instance: ProjectManager;\n private db!: Database.Database;\n private configPath: string;\n private organizations: Map<string, OrganizationConfig> = new Map();\n private projectCache: Map<string, ProjectInfo> = new Map();\n private currentProject?: ProjectInfo;\n\n private constructor() {\n this.configPath = join(homedir(), '.stackmemory');\n this.ensureDirectoryStructure();\n this.initializeDatabase();\n this.loadOrganizations();\n this.autoDiscoverOrganizations();\n }\n\n static getInstance(): ProjectManager {\n if (!ProjectManager.instance) {\n ProjectManager.instance = new ProjectManager();\n }\n return ProjectManager.instance;\n }\n\n /**\n * Auto-detect project from current directory\n */\n async detectProject(projectPath?: string): Promise<ProjectInfo> {\n const path = projectPath || process.cwd();\n const errorHandler = createErrorHandler({\n operation: 'detectProject',\n projectPath: path,\n });\n\n try {\n // Check cache first\n const cached = this.projectCache.get(path);\n if (cached && this.isCacheValid(cached)) {\n return cached;\n }\n\n const project = await this.analyzeProject(path);\n\n // Auto-categorize based on git origin\n if (project.gitRemote) {\n project.organization = this.extractOrganization(project.gitRemote);\n project.accountType = this.determineAccountType(\n project.gitRemote,\n project.organization\n );\n project.isPrivate = this.isPrivateRepo(project.gitRemote);\n }\n\n // Detect framework and language\n project.primaryLanguage = this.detectPrimaryLanguage(path);\n project.framework = this.detectFramework(path);\n\n // Store in database with retry logic\n await retry(\n () => Promise.resolve(this.saveProject(project)),\n {\n maxAttempts: 3,\n initialDelay: 100,\n onRetry: (attempt, error) => {\n logger.warn(`Retrying project save (attempt ${attempt})`, {\n projectId: project.id,\n error: error instanceof Error ? error.message : String(error),\n });\n },\n }\n );\n\n this.projectCache.set(path, project);\n this.currentProject = project;\n\n logger.info('Project auto-detected', {\n id: project.id,\n org: project.organization,\n type: project.accountType,\n });\n\n return project;\n } catch (error) {\n const wrappedError = errorHandler(error, {\n projectPath: path,\n operation: 'detectProject',\n });\n \n throw new ProjectError(\n `Failed to detect project at path: ${path}`,\n ErrorCode.PROJECT_INVALID_PATH,\n {\n projectPath: path,\n operation: 'detectProject',\n }\n );\n }\n }\n\n /**\n * Analyze project directory\n */\n private async analyzeProject(projectPath: string): Promise<ProjectInfo> {\n const gitInfo = this.getGitInfo(projectPath);\n const projectName = gitInfo.name || basename(projectPath);\n\n return {\n id: this.generateProjectId(gitInfo.remote || projectPath),\n name: projectName,\n path: projectPath,\n gitRemote: gitInfo.remote,\n organization: undefined,\n accountType: 'personal',\n isPrivate: false,\n lastAccessed: new Date(),\n metadata: {\n branch: gitInfo.branch,\n lastCommit: gitInfo.lastCommit,\n isDirty: gitInfo.isDirty,\n },\n };\n }\n\n /**\n * Extract Git information\n */\n private getGitInfo(projectPath: string): any {\n const info: any = {};\n const errorHandler = createErrorHandler({\n operation: 'getGitInfo',\n projectPath,\n });\n\n try {\n // Get remote origin\n info.remote = execSync('git config --get remote.origin.url', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000, // 5 second timeout\n }).trim();\n\n // Get current branch\n info.branch = execSync('git branch --show-current', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n }).trim();\n\n // Get last commit\n info.lastCommit = execSync('git log -1 --format=%H', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n }).trim();\n\n // Check if working tree is dirty\n const status = execSync('git status --porcelain', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n });\n info.isDirty = status.length > 0;\n\n // Extract project name from remote\n const match = info.remote.match(/\\/([^\\/]+?)(\\.git)?$/);\n info.name = match ? match[1] : basename(projectPath);\n } catch (error) {\n // Not a git repository or git not available\n logger.debug('Git info extraction failed, using directory name', {\n projectPath,\n error: error instanceof Error ? error.message : String(error),\n });\n info.name = basename(projectPath);\n }\n\n return info;\n }\n\n /**\n * Extract organization from Git remote\n */\n private extractOrganization(gitRemote: string): string {\n // GitHub: git@github.com:org/repo.git or https://github.com/org/repo\n const githubMatch = gitRemote.match(/github\\.com[:/]([^/]+)\\//);\n if (githubMatch) return githubMatch[1];\n\n // GitLab: git@gitlab.com:org/repo.git\n const gitlabMatch = gitRemote.match(/gitlab\\.com[:/]([^/]+)\\//);\n if (gitlabMatch) return gitlabMatch[1];\n\n // Bitbucket: git@bitbucket.org:org/repo.git\n const bitbucketMatch = gitRemote.match(/bitbucket\\.org[:/]([^/]+)\\//);\n if (bitbucketMatch) return bitbucketMatch[1];\n\n // Custom domain: git@git.company.com:team/repo.git\n const customMatch = gitRemote.match(/@([^:]+)[:/]([^/]+)\\//);\n if (customMatch) return customMatch[2];\n\n return 'unknown';\n }\n\n /**\n * Determine account type based on patterns\n */\n private determineAccountType(\n gitRemote: string,\n organization?: string\n ): 'personal' | 'work' | 'opensource' | 'client' {\n // Check against known organizations\n for (const [, org] of this.organizations) {\n if (org.githubOrgs.includes(organization || '')) {\n return org.accountType;\n }\n\n // Check if remote matches any known domain\n for (const domain of org.domains) {\n if (gitRemote.includes(domain)) {\n return org.accountType;\n }\n }\n }\n\n // Auto-detection heuristics\n if (organization) {\n // Common work patterns\n if (\n organization.includes('corp') ||\n organization.includes('company') ||\n organization.includes('team') ||\n organization.includes('work')\n ) {\n return 'work';\n }\n\n // Common opensource patterns\n if (\n organization.includes('apache') ||\n organization.includes('mozilla') ||\n organization.includes('foundation') ||\n gitRemote.includes('gitlab.freedesktop')\n ) {\n return 'opensource';\n }\n\n // Check if it's the user's own org\n const username = this.getCurrentGitUser();\n if (username && organization.toLowerCase() === username.toLowerCase()) {\n return 'personal';\n }\n }\n\n // Check if it's a private repo (likely work or personal)\n if (this.isPrivateRepo(gitRemote)) {\n // Use additional heuristics\n const currentPath = process.cwd();\n if (\n currentPath.includes('/work/') ||\n currentPath.includes('/Work/') ||\n currentPath.includes('/company/') ||\n currentPath.includes('/job/')\n ) {\n return 'work';\n }\n }\n\n return 'personal';\n }\n\n /**\n * Check if repository is private\n */\n private isPrivateRepo(gitRemote: string): boolean {\n // SSH URLs are typically private\n if (gitRemote.startsWith('git@')) {\n return true;\n }\n\n // HTTPS with credentials\n if (gitRemote.includes('@')) {\n return true;\n }\n\n // Try to check GitHub API (requires authentication for private repos)\n // This is a simplified check\n return false;\n }\n\n /**\n * Detect primary programming language\n */\n private detectPrimaryLanguage(projectPath: string): string | undefined {\n const checks = [\n { file: 'package.json', language: 'JavaScript/TypeScript' },\n { file: 'Cargo.toml', language: 'Rust' },\n { file: 'go.mod', language: 'Go' },\n { file: 'pom.xml', language: 'Java' },\n { file: 'requirements.txt', language: 'Python' },\n { file: 'Gemfile', language: 'Ruby' },\n { file: 'composer.json', language: 'PHP' },\n { file: '*.csproj', language: 'C#' },\n { file: 'Podfile', language: 'Swift/Objective-C' },\n ];\n\n for (const check of checks) {\n if (check.file.includes('*')) {\n // Glob pattern\n try {\n const files = execSync(\n `find ${projectPath} -maxdepth 2 -name \"${check.file}\" 2>/dev/null`,\n {\n encoding: 'utf-8',\n }\n );\n if (files.trim()) return check.language;\n } catch {}\n } else if (existsSync(join(projectPath, check.file))) {\n return check.language;\n }\n }\n\n return undefined;\n }\n\n /**\n * Detect framework\n */\n private detectFramework(projectPath: string): string | undefined {\n const packageJsonPath = join(projectPath, 'package.json');\n if (existsSync(packageJsonPath)) {\n try {\n const pkg = JSON.parse(readFileSync(packageJsonPath, 'utf-8'));\n const deps = { ...pkg.dependencies, ...pkg.devDependencies };\n\n // Check for common frameworks\n if (deps['next']) return 'Next.js';\n if (deps['react']) return 'React';\n if (deps['vue']) return 'Vue';\n if (deps['@angular/core']) return 'Angular';\n if (deps['express']) return 'Express';\n if (deps['fastify']) return 'Fastify';\n if (deps['@nestjs/core']) return 'NestJS';\n } catch {}\n }\n\n // Check for other framework indicators\n if (existsSync(join(projectPath, 'Cargo.toml'))) {\n const cargo = readFileSync(join(projectPath, 'Cargo.toml'), 'utf-8');\n if (cargo.includes('actix-web')) return 'Actix';\n if (cargo.includes('rocket')) return 'Rocket';\n }\n\n return undefined;\n }\n\n /**\n * Get current Git user\n */\n private getCurrentGitUser(): string | undefined {\n try {\n const email = execSync('git config --global user.email', {\n encoding: 'utf-8',\n }).trim();\n\n const username = email.split('@')[0];\n return username;\n } catch {\n return undefined;\n }\n }\n\n /**\n * Generate unique project ID\n */\n private generateProjectId(identifier: string): string {\n // Create a stable ID from the git remote or path\n const cleaned = identifier\n .replace(/\\.git$/, '')\n .replace(/[^a-zA-Z0-9-]/g, '-')\n .toLowerCase();\n\n return cleaned.substring(cleaned.length - 50); // Last 50 chars\n }\n\n /**\n * Initialize database\n */\n private initializeDatabase(): void {\n const dbPath = join(this.configPath, 'projects.db');\n const errorHandler = createErrorHandler({\n operation: 'initializeDatabase',\n dbPath,\n });\n\n try {\n this.db = new Database(dbPath);\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS projects (\n id TEXT PRIMARY KEY,\n name TEXT NOT NULL,\n path TEXT NOT NULL UNIQUE,\n git_remote TEXT,\n organization TEXT,\n account_type TEXT,\n is_private BOOLEAN,\n primary_language TEXT,\n framework TEXT,\n last_accessed DATETIME,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n updated_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS organizations (\n name TEXT PRIMARY KEY,\n type TEXT,\n account_type TEXT,\n config JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS project_contexts (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n project_id TEXT NOT NULL,\n context_type TEXT,\n content TEXT,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n FOREIGN KEY (project_id) REFERENCES projects(id)\n );\n\n CREATE INDEX IF NOT EXISTS idx_projects_org ON projects(organization);\n CREATE INDEX IF NOT EXISTS idx_projects_type ON projects(account_type);\n CREATE INDEX IF NOT EXISTS idx_contexts_project ON project_contexts(project_id);\n `);\n } catch (error) {\n const dbError = errorHandler(error, {\n dbPath,\n operation: 'initializeDatabase',\n });\n \n throw new DatabaseError(\n 'Failed to initialize projects database',\n ErrorCode.DB_MIGRATION_FAILED,\n {\n dbPath,\n configPath: this.configPath,\n operation: 'initializeDatabase',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Save project to database\n */\n private saveProject(project: ProjectInfo): void {\n try {\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO projects \n (id, name, path, git_remote, organization, account_type, is_private, \n primary_language, framework, last_accessed, metadata, updated_at)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, CURRENT_TIMESTAMP)\n `);\n\n stmt.run(\n project.id,\n project.name,\n project.path,\n project.gitRemote,\n project.organization,\n project.accountType,\n project.isPrivate ? 1 : 0,\n project.primaryLanguage,\n project.framework,\n project.lastAccessed.toISOString(),\n JSON.stringify(project.metadata)\n );\n } catch (error) {\n throw new DatabaseError(\n `Failed to save project: ${project.name}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n projectId: project.id,\n projectName: project.name,\n projectPath: project.path,\n operation: 'saveProject',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Load organizations configuration\n */\n private loadOrganizations(): void {\n const configFile = join(this.configPath, 'organizations.json');\n\n if (existsSync(configFile)) {\n try {\n const config = JSON.parse(readFileSync(configFile, 'utf-8'));\n for (const org of config.organizations || []) {\n this.organizations.set(org.name, org);\n }\n } catch (error) {\n logger.error(\n 'Failed to load organizations config',\n error instanceof Error ? error : undefined\n );\n }\n }\n }\n\n /**\n * Auto-discover organizations from existing projects\n */\n private autoDiscoverOrganizations(): void {\n const errorHandler = createErrorHandler({\n operation: 'autoDiscoverOrganizations',\n });\n\n try {\n const stmt = this.db.prepare(`\n SELECT DISTINCT organization, account_type, COUNT(*) as project_count\n FROM projects\n WHERE organization IS NOT NULL\n GROUP BY organization, account_type\n `);\n\n const orgs = stmt.all() as any[];\n\n for (const org of orgs) {\n if (!this.organizations.has(org.organization)) {\n // Auto-create organization config\n this.organizations.set(org.organization, {\n name: org.organization,\n type: org.account_type === 'work' ? 'company' : 'personal',\n domains: [],\n githubOrgs: [org.organization],\n accountType: org.account_type,\n autoPatterns: [],\n });\n }\n }\n } catch (error) {\n const wrappedError = errorHandler(error, {\n operation: 'autoDiscoverOrganizations',\n });\n \n logger.error(\n 'Failed to auto-discover organizations',\n error instanceof Error ? error : new Error(String(error)),\n {\n operation: 'autoDiscoverOrganizations',\n }\n );\n }\n }\n\n /**\n * Ensure directory structure exists\n */\n private ensureDirectoryStructure(): void {\n const dirs = [\n this.configPath,\n join(this.configPath, 'accounts'),\n join(this.configPath, 'accounts', 'personal'),\n join(this.configPath, 'accounts', 'work'),\n join(this.configPath, 'accounts', 'opensource'),\n join(this.configPath, 'accounts', 'client'),\n join(this.configPath, 'contexts'),\n join(this.configPath, 'patterns'),\n ];\n\n for (const dir of dirs) {\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n }\n }\n\n /**\n * Check if cache is still valid\n */\n private isCacheValid(project: ProjectInfo): boolean {\n const cacheAge = Date.now() - project.lastAccessed.getTime();\n return cacheAge < 5 * 60 * 1000; // 5 minutes\n }\n\n /**\n * Get all projects\n */\n getAllProjects(): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all() as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error) {\n throw new DatabaseError(\n 'Failed to get all projects',\n ErrorCode.DB_QUERY_FAILED,\n {\n operation: 'getAllProjects',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get projects by organization\n */\n getProjectsByOrganization(organization: string): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n WHERE organization = ?\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all(organization) as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error) {\n throw new DatabaseError(\n `Failed to get projects by organization: ${organization}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n organization,\n operation: 'getProjectsByOrganization',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get projects by account type\n */\n getProjectsByAccountType(accountType: string): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n WHERE account_type = ?\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all(accountType) as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error) {\n throw new DatabaseError(\n `Failed to get projects by account type: ${accountType}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n accountType,\n operation: 'getProjectsByAccountType',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get current project\n */\n getCurrentProject(): ProjectInfo | undefined {\n if (!this.currentProject) {\n this.detectProject();\n }\n return this.currentProject;\n }\n\n /**\n * Save organization config\n */\n saveOrganization(org: OrganizationConfig): void {\n const errorHandler = createErrorHandler({\n operation: 'saveOrganization',\n orgName: org.name,\n });\n\n try {\n this.organizations.set(org.name, org);\n\n // Save to file\n const configFile = join(this.configPath, 'organizations.json');\n const config = {\n organizations: Array.from(this.organizations.values()),\n };\n\n writeFileSync(configFile, JSON.stringify(config, null, 2));\n\n // Save to database\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO organizations (name, type, account_type, config)\n VALUES (?, ?, ?, ?)\n `);\n\n stmt.run(org.name, org.type, org.accountType, JSON.stringify(org));\n } catch (error) {\n const wrappedError = errorHandler(error, {\n orgName: org.name,\n operation: 'saveOrganization',\n });\n \n throw new DatabaseError(\n `Failed to save organization: ${org.name}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n orgName: org.name,\n operation: 'saveOrganization',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Auto-categorize all Git repositories in home directory\n */\n async scanAndCategorizeAllProjects(basePaths?: string[]): Promise<void> {\n const paths = basePaths || [\n join(homedir(), 'Dev'),\n join(homedir(), 'dev'),\n join(homedir(), 'Projects'),\n join(homedir(), 'projects'),\n join(homedir(), 'Work'),\n join(homedir(), 'work'),\n join(homedir(), 'Documents/GitHub'),\n join(homedir(), 'code'),\n ];\n\n logger.info('Scanning for Git repositories...');\n\n for (const basePath of paths) {\n if (!existsSync(basePath)) continue;\n\n try {\n // Find all .git directories with timeout\n const gitDirs = execSync(\n `find ${basePath} -type d -name .git -maxdepth 4 2>/dev/null`,\n { encoding: 'utf-8', timeout: 30000 } // 30 second timeout\n )\n .trim()\n .split('\\n')\n .filter(Boolean);\n\n for (const gitDir of gitDirs) {\n const projectPath = dirname(gitDir);\n\n try {\n await this.detectProject(projectPath);\n logger.info(`Discovered project: ${projectPath}`);\n } catch (error) {\n logger.warn(\n `Failed to analyze project: ${projectPath}`,\n {\n projectPath,\n error: error instanceof Error ? error.message : String(error),\n operation: 'scanAndCategorizeAllProjects',\n }\n );\n }\n }\n } catch (error) {\n logger.warn(\n `Failed to scan ${basePath}`,\n error instanceof Error ? { error } : undefined\n );\n }\n }\n\n logger.info(`Scan complete. Found ${this.projectCache.size} projects`);\n }\n\n /**\n * Generate summary report\n */\n generateReport(): string {\n const allProjects = this.getAllProjects();\n\n const report = {\n total: allProjects.length,\n byAccountType: {} as Record<string, number>,\n byOrganization: {} as Record<string, number>,\n byLanguage: {} as Record<string, number>,\n byFramework: {} as Record<string, number>,\n };\n\n for (const project of allProjects) {\n // Count by account type\n report.byAccountType[project.accountType] =\n (report.byAccountType[project.accountType] || 0) + 1;\n\n // Count by organization\n if (project.organization) {\n report.byOrganization[project.organization] =\n (report.byOrganization[project.organization] || 0) + 1;\n }\n\n // Count by language\n if (project.primaryLanguage) {\n report.byLanguage[project.primaryLanguage] =\n (report.byLanguage[project.primaryLanguage] || 0) + 1;\n }\n\n // Count by framework\n if (project.framework) {\n report.byFramework[project.framework] =\n (report.byFramework[project.framework] || 0) + 1;\n }\n }\n\n return JSON.stringify(report, null, 2);\n }\n}\n"],
|
|
5
|
-
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,WAAW,cAAc,qBAAqB;AACnE,SAAS,MAAM,UAAU,eAAe;AACxC,SAAS,eAAe;AACxB,OAAO,cAAc;AACrB,SAAS,cAAc;AACvB;AAAA,EACE;AAAA,EACA;AAAA,EAEA;AAAA,EAEA;AAAA,OACK;AACP,SAAS,aAA0B;AA2B5B,MAAM,eAAe;AAAA,EAC1B,OAAe;AAAA,EACP;AAAA,EACA;AAAA,EACA,gBAAiD,oBAAI,IAAI;AAAA,EACzD,eAAyC,oBAAI,IAAI;AAAA,EACjD;AAAA,EAEA,cAAc;AACpB,SAAK,aAAa,KAAK,QAAQ,GAAG,cAAc;AAChD,SAAK,yBAAyB;AAC9B,SAAK,mBAAmB;AACxB,SAAK,kBAAkB;AACvB,SAAK,0BAA0B;AAAA,EACjC;AAAA,EAEA,OAAO,cAA8B;AACnC,QAAI,CAAC,eAAe,UAAU;AAC5B,qBAAe,WAAW,IAAI,eAAe;AAAA,IAC/C;AACA,WAAO,eAAe;AAAA,EACxB;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,cAAc,aAA4C;AAC9D,UAAM,OAAO,eAAe,QAAQ,IAAI;AACxC,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX,aAAa;AAAA,IACf,CAAC;AAED,QAAI;AAEF,YAAM,SAAS,KAAK,aAAa,IAAI,IAAI;AACzC,UAAI,UAAU,KAAK,aAAa,MAAM,GAAG;AACvC,eAAO;AAAA,MACT;AAEA,YAAM,UAAU,MAAM,KAAK,eAAe,IAAI;AAG9C,UAAI,QAAQ,WAAW;AACrB,gBAAQ,eAAe,KAAK,oBAAoB,QAAQ,SAAS;AACjE,gBAAQ,cAAc,KAAK;AAAA,UACzB,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV;AACA,gBAAQ,YAAY,KAAK,cAAc,QAAQ,SAAS;AAAA,MAC1D;AAGA,cAAQ,kBAAkB,KAAK,sBAAsB,IAAI;AACzD,cAAQ,YAAY,KAAK,gBAAgB,IAAI;AAG7C,YAAM;AAAA,QACJ,MAAM,QAAQ,QAAQ,KAAK,YAAY,OAAO,CAAC;AAAA,QAC/C;AAAA,UACE,aAAa;AAAA,UACb,cAAc;AAAA,UACd,SAAS,CAAC,SAAS,UAAU;AAC3B,mBAAO,KAAK,kCAAkC,OAAO,KAAK;AAAA,cACxD,WAAW,QAAQ;AAAA,cACnB,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,YAC9D,CAAC;AAAA,UACH;AAAA,QACF;AAAA,MACF;AAEA,WAAK,aAAa,IAAI,MAAM,OAAO;AACnC,WAAK,iBAAiB;AAEtB,aAAO,KAAK,yBAAyB;AAAA,QACnC,IAAI,QAAQ;AAAA,QACZ,KAAK,QAAQ;AAAA,QACb,MAAM,QAAQ;AAAA,MAChB,CAAC;AAED,aAAO;AAAA,IACT,SAAS,OAAO;AACd,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,aAAa;AAAA,QACb,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR,qCAAqC,IAAI;AAAA,QACzC,UAAU;AAAA,QACV;AAAA,UACE,aAAa;AAAA,UACb,WAAW;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,eAAe,aAA2C;AACtE,UAAM,UAAU,KAAK,WAAW,WAAW;AAC3C,UAAM,cAAc,QAAQ,QAAQ,SAAS,WAAW;AAExD,WAAO;AAAA,MACL,IAAI,KAAK,kBAAkB,QAAQ,UAAU,WAAW;AAAA,MACxD,MAAM;AAAA,MACN,MAAM;AAAA,MACN,WAAW,QAAQ;AAAA,MACnB,cAAc;AAAA,MACd,aAAa;AAAA,MACb,WAAW;AAAA,MACX,cAAc,oBAAI,KAAK;AAAA,MACvB,UAAU;AAAA,QACR,QAAQ,QAAQ;AAAA,QAChB,YAAY,QAAQ;AAAA,QACpB,SAAS,QAAQ;AAAA,MACnB;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,WAAW,aAA0B;AAC3C,UAAM,OAAY,CAAC;AACnB,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX;AAAA,IACF,CAAC;AAED,QAAI;AAEF,WAAK,SAAS,SAAS,sCAAsC;AAAA,QAC3D,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,WAAK,SAAS,SAAS,6BAA6B;AAAA,QAClD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,WAAK,aAAa,SAAS,0BAA0B;AAAA,QACnD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,YAAM,SAAS,SAAS,0BAA0B;AAAA,QAChD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC;AACD,WAAK,UAAU,OAAO,SAAS;AAG/B,YAAM,QAAQ,KAAK,OAAO,MAAM,sBAAsB;AACtD,WAAK,OAAO,QAAQ,MAAM,CAAC,IAAI,SAAS,WAAW;AAAA,IACrD,SAAS,OAAO;AAEd,aAAO,MAAM,oDAAoD;AAAA,QAC/D;AAAA,QACA,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,CAAC;AACD,WAAK,OAAO,SAAS,WAAW;AAAA,IAClC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAoB,WAA2B;AAErD,UAAM,cAAc,UAAU,MAAM,0BAA0B;AAC9D,QAAI,YAAa,QAAO,YAAY,CAAC;AAGrC,UAAM,cAAc,UAAU,MAAM,0BAA0B;AAC9D,QAAI,YAAa,QAAO,YAAY,CAAC;AAGrC,UAAM,iBAAiB,UAAU,MAAM,6BAA6B;AACpE,QAAI,eAAgB,QAAO,eAAe,CAAC;AAG3C,UAAM,cAAc,UAAU,MAAM,uBAAuB;AAC3D,QAAI,YAAa,QAAO,YAAY,CAAC;AAErC,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,qBACN,WACA,cAC+C;AAE/C,eAAW,CAAC,EAAE,GAAG,KAAK,KAAK,eAAe;AACxC,UAAI,IAAI,WAAW,SAAS,gBAAgB,EAAE,GAAG;AAC/C,eAAO,IAAI;AAAA,MACb;AAGA,iBAAW,UAAU,IAAI,SAAS;AAChC,YAAI,UAAU,SAAS,MAAM,GAAG;AAC9B,iBAAO,IAAI;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAGA,QAAI,cAAc;AAEhB,UACE,aAAa,SAAS,MAAM,KAC5B,aAAa,SAAS,SAAS,KAC/B,aAAa,SAAS,MAAM,KAC5B,aAAa,SAAS,MAAM,GAC5B;AACA,eAAO;AAAA,MACT;AAGA,UACE,aAAa,SAAS,QAAQ,KAC9B,aAAa,SAAS,SAAS,KAC/B,aAAa,SAAS,YAAY,KAClC,UAAU,SAAS,oBAAoB,GACvC;AACA,eAAO;AAAA,MACT;AAGA,YAAM,WAAW,KAAK,kBAAkB;AACxC,UAAI,YAAY,aAAa,YAAY,MAAM,SAAS,YAAY,GAAG;AACrE,eAAO;AAAA,MACT;AAAA,IACF;AAGA,QAAI,KAAK,cAAc,SAAS,GAAG;AAEjC,YAAM,cAAc,QAAQ,IAAI;AAChC,UACE,YAAY,SAAS,QAAQ,KAC7B,YAAY,SAAS,QAAQ,KAC7B,YAAY,SAAS,WAAW,KAChC,YAAY,SAAS,OAAO,GAC5B;AACA,eAAO;AAAA,MACT;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,cAAc,WAA4B;AAEhD,QAAI,UAAU,WAAW,MAAM,GAAG;AAChC,aAAO;AAAA,IACT;AAGA,QAAI,UAAU,SAAS,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAIA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,sBAAsB,aAAyC;AACrE,UAAM,SAAS;AAAA,MACb,EAAE,MAAM,gBAAgB,UAAU,wBAAwB;AAAA,MAC1D,EAAE,MAAM,cAAc,UAAU,OAAO;AAAA,MACvC,EAAE,MAAM,UAAU,UAAU,KAAK;AAAA,MACjC,EAAE,MAAM,WAAW,UAAU,OAAO;AAAA,MACpC,EAAE,MAAM,oBAAoB,UAAU,SAAS;AAAA,MAC/C,EAAE,MAAM,WAAW,UAAU,OAAO;AAAA,MACpC,EAAE,MAAM,iBAAiB,UAAU,MAAM;AAAA,MACzC,EAAE,MAAM,YAAY,UAAU,KAAK;AAAA,MACnC,EAAE,MAAM,WAAW,UAAU,oBAAoB;AAAA,IACnD;AAEA,eAAW,SAAS,QAAQ;AAC1B,UAAI,MAAM,KAAK,SAAS,GAAG,GAAG;AAE5B,YAAI;AACF,gBAAM,QAAQ;AAAA,YACZ,QAAQ,WAAW,uBAAuB,MAAM,IAAI;AAAA,YACpD;AAAA,cACE,UAAU;AAAA,YACZ;AAAA,UACF;AACA,cAAI,MAAM,KAAK,EAAG,QAAO,MAAM;AAAA,QACjC,QAAQ;AAAA,QAAC;AAAA,MACX,WAAW,WAAW,KAAK,aAAa,MAAM,IAAI,CAAC,GAAG;AACpD,eAAO,MAAM;AAAA,MACf;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,gBAAgB,aAAyC;AAC/D,UAAM,kBAAkB,KAAK,aAAa,cAAc;AACxD,QAAI,WAAW,eAAe,GAAG;AAC/B,UAAI;AACF,cAAM,MAAM,KAAK,MAAM,aAAa,iBAAiB,OAAO,CAAC;AAC7D,cAAM,OAAO,EAAE,GAAG,IAAI,cAAc,GAAG,IAAI,gBAAgB;AAG3D,YAAI,KAAK,MAAM,EAAG,QAAO;AACzB,YAAI,KAAK,OAAO,EAAG,QAAO;AAC1B,YAAI,KAAK,KAAK,EAAG,QAAO;AACxB,YAAI,KAAK,eAAe,EAAG,QAAO;AAClC,YAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,YAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,YAAI,KAAK,cAAc,EAAG,QAAO;AAAA,MACnC,QAAQ;AAAA,MAAC;AAAA,IACX;AAGA,QAAI,WAAW,KAAK,aAAa,YAAY,CAAC,GAAG;AAC/C,YAAM,QAAQ,aAAa,KAAK,aAAa,YAAY,GAAG,OAAO;AACnE,UAAI,MAAM,SAAS,WAAW,EAAG,QAAO;AACxC,UAAI,MAAM,SAAS,QAAQ,EAAG,QAAO;AAAA,IACvC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAwC;AAC9C,QAAI;AACF,YAAM,QAAQ,SAAS,kCAAkC;AAAA,QACvD,UAAU;AAAA,MACZ,CAAC,EAAE,KAAK;AAER,YAAM,WAAW,MAAM,MAAM,GAAG,EAAE,CAAC;AACnC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,YAA4B;AAEpD,UAAM,UAAU,WACb,QAAQ,UAAU,EAAE,EACpB,QAAQ,kBAAkB,GAAG,EAC7B,YAAY;AAEf,WAAO,QAAQ,UAAU,QAAQ,SAAS,EAAE;AAAA,EAC9C;AAAA;AAAA;AAAA;AAAA,EAKQ,qBAA2B;AACjC,UAAM,SAAS,KAAK,KAAK,YAAY,aAAa;AAClD,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX;AAAA,IACF,CAAC;AAED,QAAI;AACF,WAAK,KAAK,IAAI,SAAS,MAAM;AAE7B,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAsCZ;AAAA,IACH,SAAS,OAAO;AACd,YAAM,UAAU,aAAa,OAAO;AAAA,QAClC;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR;AAAA,QACA,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,YAAY,KAAK;AAAA,UACjB,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,YAAY,SAA4B;AAC9C,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AAED,WAAK;AAAA,QACH,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,YAAY,IAAI;AAAA,QACxB,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,aAAa,YAAY;AAAA,QACjC,KAAK,UAAU,QAAQ,QAAQ;AAAA,MACjC;AAAA,IACF,SAAS,OAAO;AACd,YAAM,IAAI;AAAA,QACR,2BAA2B,QAAQ,IAAI;AAAA,QACvC,UAAU;AAAA,QACV;AAAA,UACE,WAAW,QAAQ;AAAA,UACnB,aAAa,QAAQ;AAAA,UACrB,aAAa,QAAQ;AAAA,UACrB,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAA0B;AAChC,UAAM,aAAa,KAAK,KAAK,YAAY,oBAAoB;AAE7D,QAAI,WAAW,UAAU,GAAG;AAC1B,UAAI;AACF,cAAM,SAAS,KAAK,MAAM,aAAa,YAAY,OAAO,CAAC;AAC3D,mBAAW,OAAO,OAAO,iBAAiB,CAAC,GAAG;AAC5C,eAAK,cAAc,IAAI,IAAI,MAAM,GAAG;AAAA,QACtC;AAAA,MACF,SAAS,OAAO;AACd,eAAO;AAAA,UACL;AAAA,UACA,iBAAiB,QAAQ,QAAQ;AAAA,QACnC;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,4BAAkC;AACxC,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,IACb,CAAC;AAED,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AAED,YAAM,OAAO,KAAK,IAAI;AAEtB,iBAAW,OAAO,MAAM;AACtB,YAAI,CAAC,KAAK,cAAc,IAAI,IAAI,YAAY,GAAG;AAE7C,eAAK,cAAc,IAAI,IAAI,cAAc;AAAA,YACvC,MAAM,IAAI;AAAA,YACV,MAAM,IAAI,iBAAiB,SAAS,YAAY;AAAA,YAChD,SAAS,CAAC;AAAA,YACV,YAAY,CAAC,IAAI,YAAY;AAAA,YAC7B,aAAa,IAAI;AAAA,YACjB,cAAc,CAAC;AAAA,UACjB,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,OAAO;AACd,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,WAAW;AAAA,MACb,CAAC;AAED,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QACxD;AAAA,UACE,WAAW;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,2BAAiC;AACvC,UAAM,OAAO;AAAA,MACX,KAAK;AAAA,MACL,KAAK,KAAK,YAAY,UAAU;AAAA,MAChC,KAAK,KAAK,YAAY,YAAY,UAAU;AAAA,MAC5C,KAAK,KAAK,YAAY,YAAY,MAAM;AAAA,MACxC,KAAK,KAAK,YAAY,YAAY,YAAY;AAAA,MAC9C,KAAK,KAAK,YAAY,YAAY,QAAQ;AAAA,MAC1C,KAAK,KAAK,YAAY,UAAU;AAAA,MAChC,KAAK,KAAK,YAAY,UAAU;AAAA,IAClC;AAEA,eAAW,OAAO,MAAM;AACtB,UAAI,CAAC,WAAW,GAAG,GAAG;AACpB,kBAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,MACpC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,aAAa,SAA+B;AAClD,UAAM,WAAW,KAAK,IAAI,IAAI,QAAQ,aAAa,QAAQ;AAC3D,WAAO,WAAW,IAAI,KAAK;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAgC;AAC9B,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,OAG5B;AAED,YAAM,WAAW,KAAK,IAAI;AAC1B,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAO;AACd,YAAM,IAAI;AAAA,QACR;AAAA,QACA,UAAU;AAAA,QACV;AAAA,UACE,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,0BAA0B,cAAqC;AAC7D,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAI5B;AAED,YAAM,WAAW,KAAK,IAAI,YAAY;AACtC,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAO;AACd,YAAM,IAAI;AAAA,QACR,2CAA2C,YAAY;AAAA,QACvD,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,yBAAyB,aAAoC;AAC3D,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAI5B;AAED,YAAM,WAAW,KAAK,IAAI,WAAW;AACrC,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAO;AACd,YAAM,IAAI;AAAA,QACR,2CAA2C,WAAW;AAAA,QACtD,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,oBAA6C;AAC3C,QAAI,CAAC,KAAK,gBAAgB;AACxB,WAAK,cAAc;AAAA,IACrB;AACA,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAiB,KAA+B;AAC9C,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX,SAAS,IAAI;AAAA,IACf,CAAC;AAED,QAAI;AACF,WAAK,cAAc,IAAI,IAAI,MAAM,GAAG;AAGpC,YAAM,aAAa,KAAK,KAAK,YAAY,oBAAoB;AAC7D,YAAM,SAAS;AAAA,QACb,eAAe,MAAM,KAAK,KAAK,cAAc,OAAO,CAAC;AAAA,MACvD;AAEA,oBAAc,YAAY,KAAK,UAAU,QAAQ,MAAM,CAAC,CAAC;AAGzD,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,OAG5B;AAED,WAAK,IAAI,IAAI,MAAM,IAAI,MAAM,IAAI,aAAa,KAAK,UAAU,GAAG,CAAC;AAAA,IACnE,SAAS,OAAO;AACd,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,SAAS,IAAI;AAAA,QACb,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR,gCAAgC,IAAI,IAAI;AAAA,QACxC,UAAU;AAAA,QACV;AAAA,UACE,SAAS,IAAI;AAAA,UACb,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,6BAA6B,WAAqC;AACtE,UAAM,QAAQ,aAAa;AAAA,MACzB,KAAK,QAAQ,GAAG,KAAK;AAAA,MACrB,KAAK,QAAQ,GAAG,KAAK;AAAA,MACrB,KAAK,QAAQ,GAAG,UAAU;AAAA,MAC1B,KAAK,QAAQ,GAAG,UAAU;AAAA,MAC1B,KAAK,QAAQ,GAAG,MAAM;AAAA,MACtB,KAAK,QAAQ,GAAG,MAAM;AAAA,MACtB,KAAK,QAAQ,GAAG,kBAAkB;AAAA,MAClC,KAAK,QAAQ,GAAG,MAAM;AAAA,IACxB;AAEA,WAAO,KAAK,kCAAkC;AAE9C,eAAW,YAAY,OAAO;AAC5B,UAAI,CAAC,WAAW,QAAQ,EAAG;AAE3B,UAAI;AAEF,cAAM,UAAU;AAAA,UACd,QAAQ,QAAQ;AAAA,UAChB,EAAE,UAAU,SAAS,SAAS,IAAM;AAAA;AAAA,QACtC,EACG,KAAK,EACL,MAAM,IAAI,EACV,OAAO,OAAO;AAEjB,mBAAW,UAAU,SAAS;AAC5B,gBAAM,cAAc,QAAQ,MAAM;AAElC,cAAI;AACF,kBAAM,KAAK,cAAc,WAAW;AACpC,mBAAO,KAAK,uBAAuB,WAAW,EAAE;AAAA,UAClD,SAAS,OAAO;AACd,mBAAO;AAAA,cACL,8BAA8B,WAAW;AAAA,cACzC;AAAA,gBACE;AAAA,gBACA,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,gBAC5D,WAAW;AAAA,cACb;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,eAAO;AAAA,UACL,kBAAkB,QAAQ;AAAA,UAC1B,iBAAiB,QAAQ,EAAE,MAAM,IAAI;AAAA,QACvC;AAAA,MACF;AAAA,IACF;AAEA,WAAO,KAAK,wBAAwB,KAAK,aAAa,IAAI,WAAW;AAAA,EACvE;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAyB;AACvB,UAAM,cAAc,KAAK,eAAe;AAExC,UAAM,SAAS;AAAA,MACb,OAAO,YAAY;AAAA,MACnB,eAAe,CAAC;AAAA,MAChB,gBAAgB,CAAC;AAAA,MACjB,YAAY,CAAC;AAAA,MACb,aAAa,CAAC;AAAA,IAChB;AAEA,eAAW,WAAW,aAAa;AAEjC,aAAO,cAAc,QAAQ,WAAW,KACrC,OAAO,cAAc,QAAQ,WAAW,KAAK,KAAK;AAGrD,UAAI,QAAQ,cAAc;AACxB,eAAO,eAAe,QAAQ,YAAY,KACvC,OAAO,eAAe,QAAQ,YAAY,KAAK,KAAK;AAAA,MACzD;AAGA,UAAI,QAAQ,iBAAiB;AAC3B,eAAO,WAAW,QAAQ,eAAe,KACtC,OAAO,WAAW,QAAQ,eAAe,KAAK,KAAK;AAAA,MACxD;AAGA,UAAI,QAAQ,WAAW;AACrB,eAAO,YAAY,QAAQ,SAAS,KACjC,OAAO,YAAY,QAAQ,SAAS,KAAK,KAAK;AAAA,MACnD;AAAA,IACF;AAEA,WAAO,KAAK,UAAU,QAAQ,MAAM,CAAC;AAAA,EACvC;AACF;",
|
|
4
|
+
"sourcesContent": ["/**\n * Automatic Project Management for StackMemory\n * Auto-detects and organizes projects based on Git origins\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, mkdirSync, readFileSync, writeFileSync } from 'fs';\nimport { join, basename, dirname } from 'path';\nimport { homedir } from 'os';\nimport Database from 'better-sqlite3';\nimport { logger } from '../monitoring/logger.js';\nimport {\n DatabaseError,\n ProjectError,\n SystemError,\n ErrorCode,\n wrapError,\n createErrorHandler,\n} from '../errors/index.js';\nimport { retry, withTimeout } from '../errors/recovery.js';\n\nexport interface ProjectInfo {\n id: string;\n name: string;\n path: string;\n gitRemote?: string;\n organization?: string;\n accountType: 'personal' | 'work' | 'opensource' | 'client';\n isPrivate: boolean;\n primaryLanguage?: string;\n framework?: string;\n lastAccessed: Date;\n metadata: Record<string, any>;\n}\n\nexport interface OrganizationConfig {\n name: string;\n type: 'company' | 'personal' | 'opensource' | 'client';\n domains: string[];\n githubOrgs: string[];\n gitlabGroups?: string[];\n bitbucketTeams?: string[];\n accountType: 'personal' | 'work' | 'opensource' | 'client';\n autoPatterns?: string[];\n}\n\nexport class ProjectManager {\n private static instance: ProjectManager;\n private db!: Database.Database;\n private configPath: string;\n private organizations: Map<string, OrganizationConfig> = new Map();\n private projectCache: Map<string, ProjectInfo> = new Map();\n private currentProject?: ProjectInfo;\n\n private constructor() {\n this.configPath = join(homedir(), '.stackmemory');\n this.ensureDirectoryStructure();\n this.initializeDatabase();\n this.loadOrganizations();\n this.autoDiscoverOrganizations();\n }\n\n static getInstance(): ProjectManager {\n if (!ProjectManager.instance) {\n ProjectManager.instance = new ProjectManager();\n }\n return ProjectManager.instance;\n }\n\n /**\n * Auto-detect project from current directory\n */\n async detectProject(projectPath?: string): Promise<ProjectInfo> {\n const path = projectPath || process.cwd();\n const errorHandler = createErrorHandler({\n operation: 'detectProject',\n projectPath: path,\n });\n\n try {\n // Check cache first\n const cached = this.projectCache.get(path);\n if (cached && this.isCacheValid(cached)) {\n return cached;\n }\n\n const project = await this.analyzeProject(path);\n\n // Auto-categorize based on git origin\n if (project.gitRemote) {\n project.organization = this.extractOrganization(project.gitRemote);\n project.accountType = this.determineAccountType(\n project.gitRemote,\n project.organization\n );\n project.isPrivate = this.isPrivateRepo(project.gitRemote);\n }\n\n // Detect framework and language\n project.primaryLanguage = this.detectPrimaryLanguage(path);\n project.framework = this.detectFramework(path);\n\n // Store in database with retry logic\n await retry(\n () => Promise.resolve(this.saveProject(project)),\n {\n maxAttempts: 3,\n initialDelay: 100,\n onRetry: (attempt, error) => {\n logger.warn(`Retrying project save (attempt ${attempt})`, {\n projectId: project.id,\n error: error instanceof Error ? error.message : String(error),\n });\n },\n }\n );\n\n this.projectCache.set(path, project);\n this.currentProject = project;\n\n logger.info('Project auto-detected', {\n id: project.id,\n org: project.organization,\n type: project.accountType,\n });\n\n return project;\n } catch (error: unknown) {\n const wrappedError = errorHandler(error, {\n projectPath: path,\n operation: 'detectProject',\n });\n \n throw new ProjectError(\n `Failed to detect project at path: ${path}`,\n ErrorCode.PROJECT_INVALID_PATH,\n {\n projectPath: path,\n operation: 'detectProject',\n }\n );\n }\n }\n\n /**\n * Analyze project directory\n */\n private async analyzeProject(projectPath: string): Promise<ProjectInfo> {\n const gitInfo = this.getGitInfo(projectPath);\n const projectName = gitInfo.name || basename(projectPath);\n\n return {\n id: this.generateProjectId(gitInfo.remote || projectPath),\n name: projectName,\n path: projectPath,\n gitRemote: gitInfo.remote,\n organization: undefined,\n accountType: 'personal',\n isPrivate: false,\n lastAccessed: new Date(),\n metadata: {\n branch: gitInfo.branch,\n lastCommit: gitInfo.lastCommit,\n isDirty: gitInfo.isDirty,\n },\n };\n }\n\n /**\n * Extract Git information\n */\n private getGitInfo(projectPath: string): any {\n const info: any = {};\n const errorHandler = createErrorHandler({\n operation: 'getGitInfo',\n projectPath,\n });\n\n try {\n // Get remote origin\n info.remote = execSync('git config --get remote.origin.url', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000, // 5 second timeout\n }).trim();\n\n // Get current branch\n info.branch = execSync('git branch --show-current', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n }).trim();\n\n // Get last commit\n info.lastCommit = execSync('git log -1 --format=%H', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n }).trim();\n\n // Check if working tree is dirty\n const status = execSync('git status --porcelain', {\n cwd: projectPath,\n encoding: 'utf-8',\n timeout: 5000,\n });\n info.isDirty = status.length > 0;\n\n // Extract project name from remote\n const match = info.remote.match(/\\/([^\\/]+?)(\\.git)?$/);\n info.name = match ? match[1] : basename(projectPath);\n } catch (error: unknown) {\n // Not a git repository or git not available\n logger.debug('Git info extraction failed, using directory name', {\n projectPath,\n error: error instanceof Error ? error.message : String(error),\n });\n info.name = basename(projectPath);\n }\n\n return info;\n }\n\n /**\n * Extract organization from Git remote\n */\n private extractOrganization(gitRemote: string): string {\n // GitHub: git@github.com:org/repo.git or https://github.com/org/repo\n const githubMatch = gitRemote.match(/github\\.com[:/]([^/]+)\\//);\n if (githubMatch) return githubMatch[1];\n\n // GitLab: git@gitlab.com:org/repo.git\n const gitlabMatch = gitRemote.match(/gitlab\\.com[:/]([^/]+)\\//);\n if (gitlabMatch) return gitlabMatch[1];\n\n // Bitbucket: git@bitbucket.org:org/repo.git\n const bitbucketMatch = gitRemote.match(/bitbucket\\.org[:/]([^/]+)\\//);\n if (bitbucketMatch) return bitbucketMatch[1];\n\n // Custom domain: git@git.company.com:team/repo.git\n const customMatch = gitRemote.match(/@([^:]+)[:/]([^/]+)\\//);\n if (customMatch) return customMatch[2];\n\n return 'unknown';\n }\n\n /**\n * Determine account type based on patterns\n */\n private determineAccountType(\n gitRemote: string,\n organization?: string\n ): 'personal' | 'work' | 'opensource' | 'client' {\n // Check against known organizations\n for (const [, org] of this.organizations) {\n if (org.githubOrgs.includes(organization || '')) {\n return org.accountType;\n }\n\n // Check if remote matches any known domain\n for (const domain of org.domains) {\n if (gitRemote.includes(domain)) {\n return org.accountType;\n }\n }\n }\n\n // Auto-detection heuristics\n if (organization) {\n // Common work patterns\n if (\n organization.includes('corp') ||\n organization.includes('company') ||\n organization.includes('team') ||\n organization.includes('work')\n ) {\n return 'work';\n }\n\n // Common opensource patterns\n if (\n organization.includes('apache') ||\n organization.includes('mozilla') ||\n organization.includes('foundation') ||\n gitRemote.includes('gitlab.freedesktop')\n ) {\n return 'opensource';\n }\n\n // Check if it's the user's own org\n const username = this.getCurrentGitUser();\n if (username && organization.toLowerCase() === username.toLowerCase()) {\n return 'personal';\n }\n }\n\n // Check if it's a private repo (likely work or personal)\n if (this.isPrivateRepo(gitRemote)) {\n // Use additional heuristics\n const currentPath = process.cwd();\n if (\n currentPath.includes('/work/') ||\n currentPath.includes('/Work/') ||\n currentPath.includes('/company/') ||\n currentPath.includes('/job/')\n ) {\n return 'work';\n }\n }\n\n return 'personal';\n }\n\n /**\n * Check if repository is private\n */\n private isPrivateRepo(gitRemote: string): boolean {\n // SSH URLs are typically private\n if (gitRemote.startsWith('git@')) {\n return true;\n }\n\n // HTTPS with credentials\n if (gitRemote.includes('@')) {\n return true;\n }\n\n // Try to check GitHub API (requires authentication for private repos)\n // This is a simplified check\n return false;\n }\n\n /**\n * Detect primary programming language\n */\n private detectPrimaryLanguage(projectPath: string): string | undefined {\n const checks = [\n { file: 'package.json', language: 'JavaScript/TypeScript' },\n { file: 'Cargo.toml', language: 'Rust' },\n { file: 'go.mod', language: 'Go' },\n { file: 'pom.xml', language: 'Java' },\n { file: 'requirements.txt', language: 'Python' },\n { file: 'Gemfile', language: 'Ruby' },\n { file: 'composer.json', language: 'PHP' },\n { file: '*.csproj', language: 'C#' },\n { file: 'Podfile', language: 'Swift/Objective-C' },\n ];\n\n for (const check of checks) {\n if (check.file.includes('*')) {\n // Glob pattern\n try {\n const files = execSync(\n `find ${projectPath} -maxdepth 2 -name \"${check.file}\" 2>/dev/null`,\n {\n encoding: 'utf-8',\n }\n );\n if (files.trim()) return check.language;\n } catch {}\n } else if (existsSync(join(projectPath, check.file))) {\n return check.language;\n }\n }\n\n return undefined;\n }\n\n /**\n * Detect framework\n */\n private detectFramework(projectPath: string): string | undefined {\n const packageJsonPath = join(projectPath, 'package.json');\n if (existsSync(packageJsonPath)) {\n try {\n const pkg = JSON.parse(readFileSync(packageJsonPath, 'utf-8'));\n const deps = { ...pkg.dependencies, ...pkg.devDependencies };\n\n // Check for common frameworks\n if (deps['next']) return 'Next.js';\n if (deps['react']) return 'React';\n if (deps['vue']) return 'Vue';\n if (deps['@angular/core']) return 'Angular';\n if (deps['express']) return 'Express';\n if (deps['fastify']) return 'Fastify';\n if (deps['@nestjs/core']) return 'NestJS';\n } catch {}\n }\n\n // Check for other framework indicators\n if (existsSync(join(projectPath, 'Cargo.toml'))) {\n const cargo = readFileSync(join(projectPath, 'Cargo.toml'), 'utf-8');\n if (cargo.includes('actix-web')) return 'Actix';\n if (cargo.includes('rocket')) return 'Rocket';\n }\n\n return undefined;\n }\n\n /**\n * Get current Git user\n */\n private getCurrentGitUser(): string | undefined {\n try {\n const email = execSync('git config --global user.email', {\n encoding: 'utf-8',\n }).trim();\n\n const username = email.split('@')[0];\n return username;\n } catch {\n return undefined;\n }\n }\n\n /**\n * Generate unique project ID\n */\n private generateProjectId(identifier: string): string {\n // Create a stable ID from the git remote or path\n const cleaned = identifier\n .replace(/\\.git$/, '')\n .replace(/[^a-zA-Z0-9-]/g, '-')\n .toLowerCase();\n\n return cleaned.substring(cleaned.length - 50); // Last 50 chars\n }\n\n /**\n * Initialize database\n */\n private initializeDatabase(): void {\n const dbPath = join(this.configPath, 'projects.db');\n const errorHandler = createErrorHandler({\n operation: 'initializeDatabase',\n dbPath,\n });\n\n try {\n this.db = new Database(dbPath);\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS projects (\n id TEXT PRIMARY KEY,\n name TEXT NOT NULL,\n path TEXT NOT NULL UNIQUE,\n git_remote TEXT,\n organization TEXT,\n account_type TEXT,\n is_private BOOLEAN,\n primary_language TEXT,\n framework TEXT,\n last_accessed DATETIME,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n updated_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS organizations (\n name TEXT PRIMARY KEY,\n type TEXT,\n account_type TEXT,\n config JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS project_contexts (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n project_id TEXT NOT NULL,\n context_type TEXT,\n content TEXT,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n FOREIGN KEY (project_id) REFERENCES projects(id)\n );\n\n CREATE INDEX IF NOT EXISTS idx_projects_org ON projects(organization);\n CREATE INDEX IF NOT EXISTS idx_projects_type ON projects(account_type);\n CREATE INDEX IF NOT EXISTS idx_contexts_project ON project_contexts(project_id);\n `);\n } catch (error: unknown) {\n const dbError = errorHandler(error, {\n dbPath,\n operation: 'initializeDatabase',\n });\n \n throw new DatabaseError(\n 'Failed to initialize projects database',\n ErrorCode.DB_MIGRATION_FAILED,\n {\n dbPath,\n configPath: this.configPath,\n operation: 'initializeDatabase',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Save project to database\n */\n private saveProject(project: ProjectInfo): void {\n try {\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO projects \n (id, name, path, git_remote, organization, account_type, is_private, \n primary_language, framework, last_accessed, metadata, updated_at)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, CURRENT_TIMESTAMP)\n `);\n\n stmt.run(\n project.id,\n project.name,\n project.path,\n project.gitRemote,\n project.organization,\n project.accountType,\n project.isPrivate ? 1 : 0,\n project.primaryLanguage,\n project.framework,\n project.lastAccessed.toISOString(),\n JSON.stringify(project.metadata)\n );\n } catch (error: unknown) {\n throw new DatabaseError(\n `Failed to save project: ${project.name}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n projectId: project.id,\n projectName: project.name,\n projectPath: project.path,\n operation: 'saveProject',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Load organizations configuration\n */\n private loadOrganizations(): void {\n const configFile = join(this.configPath, 'organizations.json');\n\n if (existsSync(configFile)) {\n try {\n const config = JSON.parse(readFileSync(configFile, 'utf-8'));\n for (const org of config.organizations || []) {\n this.organizations.set(org.name, org);\n }\n } catch (error: unknown) {\n logger.error(\n 'Failed to load organizations config',\n error instanceof Error ? error : undefined\n );\n }\n }\n }\n\n /**\n * Auto-discover organizations from existing projects\n */\n private autoDiscoverOrganizations(): void {\n const errorHandler = createErrorHandler({\n operation: 'autoDiscoverOrganizations',\n });\n\n try {\n const stmt = this.db.prepare(`\n SELECT DISTINCT organization, account_type, COUNT(*) as project_count\n FROM projects\n WHERE organization IS NOT NULL\n GROUP BY organization, account_type\n `);\n\n const orgs = stmt.all() as any[];\n\n for (const org of orgs) {\n if (!this.organizations.has(org.organization)) {\n // Auto-create organization config\n this.organizations.set(org.organization, {\n name: org.organization,\n type: org.account_type === 'work' ? 'company' : 'personal',\n domains: [],\n githubOrgs: [org.organization],\n accountType: org.account_type,\n autoPatterns: [],\n });\n }\n }\n } catch (error: unknown) {\n const wrappedError = errorHandler(error, {\n operation: 'autoDiscoverOrganizations',\n });\n \n logger.error(\n 'Failed to auto-discover organizations',\n error instanceof Error ? error : new Error(String(error)),\n {\n operation: 'autoDiscoverOrganizations',\n }\n );\n }\n }\n\n /**\n * Ensure directory structure exists\n */\n private ensureDirectoryStructure(): void {\n const dirs = [\n this.configPath,\n join(this.configPath, 'accounts'),\n join(this.configPath, 'accounts', 'personal'),\n join(this.configPath, 'accounts', 'work'),\n join(this.configPath, 'accounts', 'opensource'),\n join(this.configPath, 'accounts', 'client'),\n join(this.configPath, 'contexts'),\n join(this.configPath, 'patterns'),\n ];\n\n for (const dir of dirs) {\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n }\n }\n\n /**\n * Check if cache is still valid\n */\n private isCacheValid(project: ProjectInfo): boolean {\n const cacheAge = Date.now() - project.lastAccessed.getTime();\n return cacheAge < 5 * 60 * 1000; // 5 minutes\n }\n\n /**\n * Get all projects\n */\n getAllProjects(): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all() as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error: unknown) {\n throw new DatabaseError(\n 'Failed to get all projects',\n ErrorCode.DB_QUERY_FAILED,\n {\n operation: 'getAllProjects',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get projects by organization\n */\n getProjectsByOrganization(organization: string): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n WHERE organization = ?\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all(organization) as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error: unknown) {\n throw new DatabaseError(\n `Failed to get projects by organization: ${organization}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n organization,\n operation: 'getProjectsByOrganization',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get projects by account type\n */\n getProjectsByAccountType(accountType: string): ProjectInfo[] {\n try {\n const stmt = this.db.prepare(`\n SELECT * FROM projects\n WHERE account_type = ?\n ORDER BY last_accessed DESC\n `);\n\n const projects = stmt.all(accountType) as any[];\n return projects.map((p) => ({\n ...p,\n isPrivate: p.is_private === 1,\n lastAccessed: new Date(p.last_accessed),\n metadata: JSON.parse(p.metadata || '{}'),\n }));\n } catch (error: unknown) {\n throw new DatabaseError(\n `Failed to get projects by account type: ${accountType}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n accountType,\n operation: 'getProjectsByAccountType',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Get current project\n */\n getCurrentProject(): ProjectInfo | undefined {\n if (!this.currentProject) {\n this.detectProject();\n }\n return this.currentProject;\n }\n\n /**\n * Save organization config\n */\n saveOrganization(org: OrganizationConfig): void {\n const errorHandler = createErrorHandler({\n operation: 'saveOrganization',\n orgName: org.name,\n });\n\n try {\n this.organizations.set(org.name, org);\n\n // Save to file\n const configFile = join(this.configPath, 'organizations.json');\n const config = {\n organizations: Array.from(this.organizations.values()),\n };\n\n writeFileSync(configFile, JSON.stringify(config, null, 2));\n\n // Save to database\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO organizations (name, type, account_type, config)\n VALUES (?, ?, ?, ?)\n `);\n\n stmt.run(org.name, org.type, org.accountType, JSON.stringify(org));\n } catch (error: unknown) {\n const wrappedError = errorHandler(error, {\n orgName: org.name,\n operation: 'saveOrganization',\n });\n \n throw new DatabaseError(\n `Failed to save organization: ${org.name}`,\n ErrorCode.DB_QUERY_FAILED,\n {\n orgName: org.name,\n operation: 'saveOrganization',\n },\n error instanceof Error ? error : undefined\n );\n }\n }\n\n /**\n * Auto-categorize all Git repositories in home directory\n */\n async scanAndCategorizeAllProjects(basePaths?: string[]): Promise<void> {\n const paths = basePaths || [\n join(homedir(), 'Dev'),\n join(homedir(), 'dev'),\n join(homedir(), 'Projects'),\n join(homedir(), 'projects'),\n join(homedir(), 'Work'),\n join(homedir(), 'work'),\n join(homedir(), 'Documents/GitHub'),\n join(homedir(), 'code'),\n ];\n\n logger.info('Scanning for Git repositories...');\n\n for (const basePath of paths) {\n if (!existsSync(basePath)) continue;\n\n try {\n // Find all .git directories with timeout\n const gitDirs = execSync(\n `find ${basePath} -type d -name .git -maxdepth 4 2>/dev/null`,\n { encoding: 'utf-8', timeout: 30000 } // 30 second timeout\n )\n .trim()\n .split('\\n')\n .filter(Boolean);\n\n for (const gitDir of gitDirs) {\n const projectPath = dirname(gitDir);\n\n try {\n await this.detectProject(projectPath);\n logger.info(`Discovered project: ${projectPath}`);\n } catch (error: unknown) {\n logger.warn(\n `Failed to analyze project: ${projectPath}`,\n {\n projectPath,\n error: error instanceof Error ? error.message : String(error),\n operation: 'scanAndCategorizeAllProjects',\n }\n );\n }\n }\n } catch (error: unknown) {\n logger.warn(\n `Failed to scan ${basePath}`,\n error instanceof Error ? { error } : undefined\n );\n }\n }\n\n logger.info(`Scan complete. Found ${this.projectCache.size} projects`);\n }\n\n /**\n * Generate summary report\n */\n generateReport(): string {\n const allProjects = this.getAllProjects();\n\n const report = {\n total: allProjects.length,\n byAccountType: {} as Record<string, number>,\n byOrganization: {} as Record<string, number>,\n byLanguage: {} as Record<string, number>,\n byFramework: {} as Record<string, number>,\n };\n\n for (const project of allProjects) {\n // Count by account type\n report.byAccountType[project.accountType] =\n (report.byAccountType[project.accountType] || 0) + 1;\n\n // Count by organization\n if (project.organization) {\n report.byOrganization[project.organization] =\n (report.byOrganization[project.organization] || 0) + 1;\n }\n\n // Count by language\n if (project.primaryLanguage) {\n report.byLanguage[project.primaryLanguage] =\n (report.byLanguage[project.primaryLanguage] || 0) + 1;\n }\n\n // Count by framework\n if (project.framework) {\n report.byFramework[project.framework] =\n (report.byFramework[project.framework] || 0) + 1;\n }\n }\n\n return JSON.stringify(report, null, 2);\n }\n}\n"],
|
|
5
|
+
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,WAAW,cAAc,qBAAqB;AACnE,SAAS,MAAM,UAAU,eAAe;AACxC,SAAS,eAAe;AACxB,OAAO,cAAc;AACrB,SAAS,cAAc;AACvB;AAAA,EACE;AAAA,EACA;AAAA,EAEA;AAAA,EAEA;AAAA,OACK;AACP,SAAS,aAA0B;AA2B5B,MAAM,eAAe;AAAA,EAC1B,OAAe;AAAA,EACP;AAAA,EACA;AAAA,EACA,gBAAiD,oBAAI,IAAI;AAAA,EACzD,eAAyC,oBAAI,IAAI;AAAA,EACjD;AAAA,EAEA,cAAc;AACpB,SAAK,aAAa,KAAK,QAAQ,GAAG,cAAc;AAChD,SAAK,yBAAyB;AAC9B,SAAK,mBAAmB;AACxB,SAAK,kBAAkB;AACvB,SAAK,0BAA0B;AAAA,EACjC;AAAA,EAEA,OAAO,cAA8B;AACnC,QAAI,CAAC,eAAe,UAAU;AAC5B,qBAAe,WAAW,IAAI,eAAe;AAAA,IAC/C;AACA,WAAO,eAAe;AAAA,EACxB;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,cAAc,aAA4C;AAC9D,UAAM,OAAO,eAAe,QAAQ,IAAI;AACxC,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX,aAAa;AAAA,IACf,CAAC;AAED,QAAI;AAEF,YAAM,SAAS,KAAK,aAAa,IAAI,IAAI;AACzC,UAAI,UAAU,KAAK,aAAa,MAAM,GAAG;AACvC,eAAO;AAAA,MACT;AAEA,YAAM,UAAU,MAAM,KAAK,eAAe,IAAI;AAG9C,UAAI,QAAQ,WAAW;AACrB,gBAAQ,eAAe,KAAK,oBAAoB,QAAQ,SAAS;AACjE,gBAAQ,cAAc,KAAK;AAAA,UACzB,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV;AACA,gBAAQ,YAAY,KAAK,cAAc,QAAQ,SAAS;AAAA,MAC1D;AAGA,cAAQ,kBAAkB,KAAK,sBAAsB,IAAI;AACzD,cAAQ,YAAY,KAAK,gBAAgB,IAAI;AAG7C,YAAM;AAAA,QACJ,MAAM,QAAQ,QAAQ,KAAK,YAAY,OAAO,CAAC;AAAA,QAC/C;AAAA,UACE,aAAa;AAAA,UACb,cAAc;AAAA,UACd,SAAS,CAAC,SAAS,UAAU;AAC3B,mBAAO,KAAK,kCAAkC,OAAO,KAAK;AAAA,cACxD,WAAW,QAAQ;AAAA,cACnB,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,YAC9D,CAAC;AAAA,UACH;AAAA,QACF;AAAA,MACF;AAEA,WAAK,aAAa,IAAI,MAAM,OAAO;AACnC,WAAK,iBAAiB;AAEtB,aAAO,KAAK,yBAAyB;AAAA,QACnC,IAAI,QAAQ;AAAA,QACZ,KAAK,QAAQ;AAAA,QACb,MAAM,QAAQ;AAAA,MAChB,CAAC;AAED,aAAO;AAAA,IACT,SAAS,OAAgB;AACvB,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,aAAa;AAAA,QACb,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR,qCAAqC,IAAI;AAAA,QACzC,UAAU;AAAA,QACV;AAAA,UACE,aAAa;AAAA,UACb,WAAW;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,eAAe,aAA2C;AACtE,UAAM,UAAU,KAAK,WAAW,WAAW;AAC3C,UAAM,cAAc,QAAQ,QAAQ,SAAS,WAAW;AAExD,WAAO;AAAA,MACL,IAAI,KAAK,kBAAkB,QAAQ,UAAU,WAAW;AAAA,MACxD,MAAM;AAAA,MACN,MAAM;AAAA,MACN,WAAW,QAAQ;AAAA,MACnB,cAAc;AAAA,MACd,aAAa;AAAA,MACb,WAAW;AAAA,MACX,cAAc,oBAAI,KAAK;AAAA,MACvB,UAAU;AAAA,QACR,QAAQ,QAAQ;AAAA,QAChB,YAAY,QAAQ;AAAA,QACpB,SAAS,QAAQ;AAAA,MACnB;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,WAAW,aAA0B;AAC3C,UAAM,OAAY,CAAC;AACnB,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX;AAAA,IACF,CAAC;AAED,QAAI;AAEF,WAAK,SAAS,SAAS,sCAAsC;AAAA,QAC3D,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,WAAK,SAAS,SAAS,6BAA6B;AAAA,QAClD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,WAAK,aAAa,SAAS,0BAA0B;AAAA,QACnD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC,EAAE,KAAK;AAGR,YAAM,SAAS,SAAS,0BAA0B;AAAA,QAChD,KAAK;AAAA,QACL,UAAU;AAAA,QACV,SAAS;AAAA,MACX,CAAC;AACD,WAAK,UAAU,OAAO,SAAS;AAG/B,YAAM,QAAQ,KAAK,OAAO,MAAM,sBAAsB;AACtD,WAAK,OAAO,QAAQ,MAAM,CAAC,IAAI,SAAS,WAAW;AAAA,IACrD,SAAS,OAAgB;AAEvB,aAAO,MAAM,oDAAoD;AAAA,QAC/D;AAAA,QACA,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,CAAC;AACD,WAAK,OAAO,SAAS,WAAW;AAAA,IAClC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAoB,WAA2B;AAErD,UAAM,cAAc,UAAU,MAAM,0BAA0B;AAC9D,QAAI,YAAa,QAAO,YAAY,CAAC;AAGrC,UAAM,cAAc,UAAU,MAAM,0BAA0B;AAC9D,QAAI,YAAa,QAAO,YAAY,CAAC;AAGrC,UAAM,iBAAiB,UAAU,MAAM,6BAA6B;AACpE,QAAI,eAAgB,QAAO,eAAe,CAAC;AAG3C,UAAM,cAAc,UAAU,MAAM,uBAAuB;AAC3D,QAAI,YAAa,QAAO,YAAY,CAAC;AAErC,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,qBACN,WACA,cAC+C;AAE/C,eAAW,CAAC,EAAE,GAAG,KAAK,KAAK,eAAe;AACxC,UAAI,IAAI,WAAW,SAAS,gBAAgB,EAAE,GAAG;AAC/C,eAAO,IAAI;AAAA,MACb;AAGA,iBAAW,UAAU,IAAI,SAAS;AAChC,YAAI,UAAU,SAAS,MAAM,GAAG;AAC9B,iBAAO,IAAI;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAGA,QAAI,cAAc;AAEhB,UACE,aAAa,SAAS,MAAM,KAC5B,aAAa,SAAS,SAAS,KAC/B,aAAa,SAAS,MAAM,KAC5B,aAAa,SAAS,MAAM,GAC5B;AACA,eAAO;AAAA,MACT;AAGA,UACE,aAAa,SAAS,QAAQ,KAC9B,aAAa,SAAS,SAAS,KAC/B,aAAa,SAAS,YAAY,KAClC,UAAU,SAAS,oBAAoB,GACvC;AACA,eAAO;AAAA,MACT;AAGA,YAAM,WAAW,KAAK,kBAAkB;AACxC,UAAI,YAAY,aAAa,YAAY,MAAM,SAAS,YAAY,GAAG;AACrE,eAAO;AAAA,MACT;AAAA,IACF;AAGA,QAAI,KAAK,cAAc,SAAS,GAAG;AAEjC,YAAM,cAAc,QAAQ,IAAI;AAChC,UACE,YAAY,SAAS,QAAQ,KAC7B,YAAY,SAAS,QAAQ,KAC7B,YAAY,SAAS,WAAW,KAChC,YAAY,SAAS,OAAO,GAC5B;AACA,eAAO;AAAA,MACT;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,cAAc,WAA4B;AAEhD,QAAI,UAAU,WAAW,MAAM,GAAG;AAChC,aAAO;AAAA,IACT;AAGA,QAAI,UAAU,SAAS,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAIA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,sBAAsB,aAAyC;AACrE,UAAM,SAAS;AAAA,MACb,EAAE,MAAM,gBAAgB,UAAU,wBAAwB;AAAA,MAC1D,EAAE,MAAM,cAAc,UAAU,OAAO;AAAA,MACvC,EAAE,MAAM,UAAU,UAAU,KAAK;AAAA,MACjC,EAAE,MAAM,WAAW,UAAU,OAAO;AAAA,MACpC,EAAE,MAAM,oBAAoB,UAAU,SAAS;AAAA,MAC/C,EAAE,MAAM,WAAW,UAAU,OAAO;AAAA,MACpC,EAAE,MAAM,iBAAiB,UAAU,MAAM;AAAA,MACzC,EAAE,MAAM,YAAY,UAAU,KAAK;AAAA,MACnC,EAAE,MAAM,WAAW,UAAU,oBAAoB;AAAA,IACnD;AAEA,eAAW,SAAS,QAAQ;AAC1B,UAAI,MAAM,KAAK,SAAS,GAAG,GAAG;AAE5B,YAAI;AACF,gBAAM,QAAQ;AAAA,YACZ,QAAQ,WAAW,uBAAuB,MAAM,IAAI;AAAA,YACpD;AAAA,cACE,UAAU;AAAA,YACZ;AAAA,UACF;AACA,cAAI,MAAM,KAAK,EAAG,QAAO,MAAM;AAAA,QACjC,QAAQ;AAAA,QAAC;AAAA,MACX,WAAW,WAAW,KAAK,aAAa,MAAM,IAAI,CAAC,GAAG;AACpD,eAAO,MAAM;AAAA,MACf;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,gBAAgB,aAAyC;AAC/D,UAAM,kBAAkB,KAAK,aAAa,cAAc;AACxD,QAAI,WAAW,eAAe,GAAG;AAC/B,UAAI;AACF,cAAM,MAAM,KAAK,MAAM,aAAa,iBAAiB,OAAO,CAAC;AAC7D,cAAM,OAAO,EAAE,GAAG,IAAI,cAAc,GAAG,IAAI,gBAAgB;AAG3D,YAAI,KAAK,MAAM,EAAG,QAAO;AACzB,YAAI,KAAK,OAAO,EAAG,QAAO;AAC1B,YAAI,KAAK,KAAK,EAAG,QAAO;AACxB,YAAI,KAAK,eAAe,EAAG,QAAO;AAClC,YAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,YAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,YAAI,KAAK,cAAc,EAAG,QAAO;AAAA,MACnC,QAAQ;AAAA,MAAC;AAAA,IACX;AAGA,QAAI,WAAW,KAAK,aAAa,YAAY,CAAC,GAAG;AAC/C,YAAM,QAAQ,aAAa,KAAK,aAAa,YAAY,GAAG,OAAO;AACnE,UAAI,MAAM,SAAS,WAAW,EAAG,QAAO;AACxC,UAAI,MAAM,SAAS,QAAQ,EAAG,QAAO;AAAA,IACvC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAwC;AAC9C,QAAI;AACF,YAAM,QAAQ,SAAS,kCAAkC;AAAA,QACvD,UAAU;AAAA,MACZ,CAAC,EAAE,KAAK;AAER,YAAM,WAAW,MAAM,MAAM,GAAG,EAAE,CAAC;AACnC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,YAA4B;AAEpD,UAAM,UAAU,WACb,QAAQ,UAAU,EAAE,EACpB,QAAQ,kBAAkB,GAAG,EAC7B,YAAY;AAEf,WAAO,QAAQ,UAAU,QAAQ,SAAS,EAAE;AAAA,EAC9C;AAAA;AAAA;AAAA;AAAA,EAKQ,qBAA2B;AACjC,UAAM,SAAS,KAAK,KAAK,YAAY,aAAa;AAClD,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX;AAAA,IACF,CAAC;AAED,QAAI;AACF,WAAK,KAAK,IAAI,SAAS,MAAM;AAE7B,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAsCZ;AAAA,IACH,SAAS,OAAgB;AACvB,YAAM,UAAU,aAAa,OAAO;AAAA,QAClC;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR;AAAA,QACA,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,YAAY,KAAK;AAAA,UACjB,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,YAAY,SAA4B;AAC9C,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AAED,WAAK;AAAA,QACH,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,YAAY,IAAI;AAAA,QACxB,QAAQ;AAAA,QACR,QAAQ;AAAA,QACR,QAAQ,aAAa,YAAY;AAAA,QACjC,KAAK,UAAU,QAAQ,QAAQ;AAAA,MACjC;AAAA,IACF,SAAS,OAAgB;AACvB,YAAM,IAAI;AAAA,QACR,2BAA2B,QAAQ,IAAI;AAAA,QACvC,UAAU;AAAA,QACV;AAAA,UACE,WAAW,QAAQ;AAAA,UACnB,aAAa,QAAQ;AAAA,UACrB,aAAa,QAAQ;AAAA,UACrB,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAA0B;AAChC,UAAM,aAAa,KAAK,KAAK,YAAY,oBAAoB;AAE7D,QAAI,WAAW,UAAU,GAAG;AAC1B,UAAI;AACF,cAAM,SAAS,KAAK,MAAM,aAAa,YAAY,OAAO,CAAC;AAC3D,mBAAW,OAAO,OAAO,iBAAiB,CAAC,GAAG;AAC5C,eAAK,cAAc,IAAI,IAAI,MAAM,GAAG;AAAA,QACtC;AAAA,MACF,SAAS,OAAgB;AACvB,eAAO;AAAA,UACL;AAAA,UACA,iBAAiB,QAAQ,QAAQ;AAAA,QACnC;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,4BAAkC;AACxC,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,IACb,CAAC;AAED,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AAED,YAAM,OAAO,KAAK,IAAI;AAEtB,iBAAW,OAAO,MAAM;AACtB,YAAI,CAAC,KAAK,cAAc,IAAI,IAAI,YAAY,GAAG;AAE7C,eAAK,cAAc,IAAI,IAAI,cAAc;AAAA,YACvC,MAAM,IAAI;AAAA,YACV,MAAM,IAAI,iBAAiB,SAAS,YAAY;AAAA,YAChD,SAAS,CAAC;AAAA,YACV,YAAY,CAAC,IAAI,YAAY;AAAA,YAC7B,aAAa,IAAI;AAAA,YACjB,cAAc,CAAC;AAAA,UACjB,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,OAAgB;AACvB,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,WAAW;AAAA,MACb,CAAC;AAED,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QACxD;AAAA,UACE,WAAW;AAAA,QACb;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,2BAAiC;AACvC,UAAM,OAAO;AAAA,MACX,KAAK;AAAA,MACL,KAAK,KAAK,YAAY,UAAU;AAAA,MAChC,KAAK,KAAK,YAAY,YAAY,UAAU;AAAA,MAC5C,KAAK,KAAK,YAAY,YAAY,MAAM;AAAA,MACxC,KAAK,KAAK,YAAY,YAAY,YAAY;AAAA,MAC9C,KAAK,KAAK,YAAY,YAAY,QAAQ;AAAA,MAC1C,KAAK,KAAK,YAAY,UAAU;AAAA,MAChC,KAAK,KAAK,YAAY,UAAU;AAAA,IAClC;AAEA,eAAW,OAAO,MAAM;AACtB,UAAI,CAAC,WAAW,GAAG,GAAG;AACpB,kBAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,MACpC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,aAAa,SAA+B;AAClD,UAAM,WAAW,KAAK,IAAI,IAAI,QAAQ,aAAa,QAAQ;AAC3D,WAAO,WAAW,IAAI,KAAK;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAgC;AAC9B,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,OAG5B;AAED,YAAM,WAAW,KAAK,IAAI;AAC1B,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAgB;AACvB,YAAM,IAAI;AAAA,QACR;AAAA,QACA,UAAU;AAAA,QACV;AAAA,UACE,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,0BAA0B,cAAqC;AAC7D,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAI5B;AAED,YAAM,WAAW,KAAK,IAAI,YAAY;AACtC,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAgB;AACvB,YAAM,IAAI;AAAA,QACR,2CAA2C,YAAY;AAAA,QACvD,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,yBAAyB,aAAoC;AAC3D,QAAI;AACF,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAI5B;AAED,YAAM,WAAW,KAAK,IAAI,WAAW;AACrC,aAAO,SAAS,IAAI,CAAC,OAAO;AAAA,QAC1B,GAAG;AAAA,QACH,WAAW,EAAE,eAAe;AAAA,QAC5B,cAAc,IAAI,KAAK,EAAE,aAAa;AAAA,QACtC,UAAU,KAAK,MAAM,EAAE,YAAY,IAAI;AAAA,MACzC,EAAE;AAAA,IACJ,SAAS,OAAgB;AACvB,YAAM,IAAI;AAAA,QACR,2CAA2C,WAAW;AAAA,QACtD,UAAU;AAAA,QACV;AAAA,UACE;AAAA,UACA,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,oBAA6C;AAC3C,QAAI,CAAC,KAAK,gBAAgB;AACxB,WAAK,cAAc;AAAA,IACrB;AACA,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAiB,KAA+B;AAC9C,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX,SAAS,IAAI;AAAA,IACf,CAAC;AAED,QAAI;AACF,WAAK,cAAc,IAAI,IAAI,MAAM,GAAG;AAGpC,YAAM,aAAa,KAAK,KAAK,YAAY,oBAAoB;AAC7D,YAAM,SAAS;AAAA,QACb,eAAe,MAAM,KAAK,KAAK,cAAc,OAAO,CAAC;AAAA,MACvD;AAEA,oBAAc,YAAY,KAAK,UAAU,QAAQ,MAAM,CAAC,CAAC;AAGzD,YAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,OAG5B;AAED,WAAK,IAAI,IAAI,MAAM,IAAI,MAAM,IAAI,aAAa,KAAK,UAAU,GAAG,CAAC;AAAA,IACnE,SAAS,OAAgB;AACvB,YAAM,eAAe,aAAa,OAAO;AAAA,QACvC,SAAS,IAAI;AAAA,QACb,WAAW;AAAA,MACb,CAAC;AAED,YAAM,IAAI;AAAA,QACR,gCAAgC,IAAI,IAAI;AAAA,QACxC,UAAU;AAAA,QACV;AAAA,UACE,SAAS,IAAI;AAAA,UACb,WAAW;AAAA,QACb;AAAA,QACA,iBAAiB,QAAQ,QAAQ;AAAA,MACnC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,6BAA6B,WAAqC;AACtE,UAAM,QAAQ,aAAa;AAAA,MACzB,KAAK,QAAQ,GAAG,KAAK;AAAA,MACrB,KAAK,QAAQ,GAAG,KAAK;AAAA,MACrB,KAAK,QAAQ,GAAG,UAAU;AAAA,MAC1B,KAAK,QAAQ,GAAG,UAAU;AAAA,MAC1B,KAAK,QAAQ,GAAG,MAAM;AAAA,MACtB,KAAK,QAAQ,GAAG,MAAM;AAAA,MACtB,KAAK,QAAQ,GAAG,kBAAkB;AAAA,MAClC,KAAK,QAAQ,GAAG,MAAM;AAAA,IACxB;AAEA,WAAO,KAAK,kCAAkC;AAE9C,eAAW,YAAY,OAAO;AAC5B,UAAI,CAAC,WAAW,QAAQ,EAAG;AAE3B,UAAI;AAEF,cAAM,UAAU;AAAA,UACd,QAAQ,QAAQ;AAAA,UAChB,EAAE,UAAU,SAAS,SAAS,IAAM;AAAA;AAAA,QACtC,EACG,KAAK,EACL,MAAM,IAAI,EACV,OAAO,OAAO;AAEjB,mBAAW,UAAU,SAAS;AAC5B,gBAAM,cAAc,QAAQ,MAAM;AAElC,cAAI;AACF,kBAAM,KAAK,cAAc,WAAW;AACpC,mBAAO,KAAK,uBAAuB,WAAW,EAAE;AAAA,UAClD,SAAS,OAAgB;AACvB,mBAAO;AAAA,cACL,8BAA8B,WAAW;AAAA,cACzC;AAAA,gBACE;AAAA,gBACA,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,gBAC5D,WAAW;AAAA,cACb;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAgB;AACvB,eAAO;AAAA,UACL,kBAAkB,QAAQ;AAAA,UAC1B,iBAAiB,QAAQ,EAAE,MAAM,IAAI;AAAA,QACvC;AAAA,MACF;AAAA,IACF;AAEA,WAAO,KAAK,wBAAwB,KAAK,aAAa,IAAI,WAAW;AAAA,EACvE;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAyB;AACvB,UAAM,cAAc,KAAK,eAAe;AAExC,UAAM,SAAS;AAAA,MACb,OAAO,YAAY;AAAA,MACnB,eAAe,CAAC;AAAA,MAChB,gBAAgB,CAAC;AAAA,MACjB,YAAY,CAAC;AAAA,MACb,aAAa,CAAC;AAAA,IAChB;AAEA,eAAW,WAAW,aAAa;AAEjC,aAAO,cAAc,QAAQ,WAAW,KACrC,OAAO,cAAc,QAAQ,WAAW,KAAK,KAAK;AAGrD,UAAI,QAAQ,cAAc;AACxB,eAAO,eAAe,QAAQ,YAAY,KACvC,OAAO,eAAe,QAAQ,YAAY,KAAK,KAAK;AAAA,MACzD;AAGA,UAAI,QAAQ,iBAAiB;AAC3B,eAAO,WAAW,QAAQ,eAAe,KACtC,OAAO,WAAW,QAAQ,eAAe,KAAK,KAAK;AAAA,MACxD;AAGA,UAAI,QAAQ,WAAW;AACrB,eAAO,YAAY,QAAQ,SAAS,KACjC,OAAO,YAAY,QAAQ,SAAS,KAAK,KAAK;AAAA,MACnD;AAAA,IACF;AAEA,WAAO,KAAK,UAAU,QAAQ,MAAM,CAAC;AAAA,EACvC;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/retrieval/context-retriever.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * LLM-driven Context Retrieval System\n * Intelligently retrieves relevant context using ParadeDB search capabilities\n */\n\nimport {\n DatabaseAdapter,\n SearchOptions,\n} from '../database/database-adapter.js';\nimport { Frame } from '../context/frame-manager.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport interface ContextQuery {\n text: string;\n type?: 'semantic' | 'keyword' | 'hybrid';\n maxResults?: number;\n timeRange?: {\n start?: Date;\n end?: Date;\n };\n frameTypes?: string[];\n scoreThreshold?: number;\n includeDigests?: boolean;\n}\n\nexport interface RetrievedContext {\n frame: Frame;\n score: number;\n relevanceReason: string;\n retrievalMethod: 'bm25' | 'vector' | 'hybrid';\n matchedFields: string[];\n}\n\nexport interface ContextRetrievalResult {\n contexts: RetrievedContext[];\n totalMatches: number;\n retrievalTimeMs: number;\n strategy: string;\n queryAnalysis: {\n intent: string;\n concepts: string[];\n complexity: 'simple' | 'moderate' | 'complex';\n };\n}\n\nexport interface RetrievalStrategy {\n name: string;\n searchType: 'text' | 'vector' | 'hybrid';\n weights?: { text: number; vector: number };\n boost?: Record<string, number>;\n fallbackStrategy?: string;\n}\n\nexport class ContextRetriever {\n private readonly adapter: DatabaseAdapter;\n private readonly strategies: Map<string, RetrievalStrategy> = new Map();\n private queryCache = new Map<string, ContextRetrievalResult>();\n private cacheMaxSize = 100;\n private cacheExpiryMs = 300000; // 5 minutes\n\n constructor(adapter: DatabaseAdapter) {\n this.adapter = adapter;\n this.initializeStrategies();\n }\n\n private initializeStrategies(): void {\n // Keyword-based search for specific terms\n this.strategies.set('keyword', {\n name: 'Keyword Search',\n searchType: 'text',\n boost: {\n name: 2.0,\n digest_text: 1.5,\n inputs: 1.2,\n outputs: 1.2,\n },\n fallbackStrategy: 'semantic',\n });\n\n // Semantic search using vector embeddings\n this.strategies.set('semantic', {\n name: 'Semantic Search',\n searchType: 'vector',\n fallbackStrategy: 'hybrid',\n });\n\n // Hybrid approach combining text and vector search\n this.strategies.set('hybrid', {\n name: 'Hybrid Search',\n searchType: 'hybrid',\n weights: { text: 0.6, vector: 0.4 },\n boost: {\n name: 2.0,\n digest_text: 1.5,\n },\n fallbackStrategy: 'keyword',\n });\n\n // Recent activity search\n this.strategies.set('recent', {\n name: 'Recent Activity',\n searchType: 'text',\n boost: {\n created_at: 3.0,\n closed_at: 2.0,\n },\n fallbackStrategy: 'hybrid',\n });\n\n // Error and debugging context\n this.strategies.set('debug', {\n name: 'Debug Context',\n searchType: 'hybrid',\n weights: { text: 0.8, vector: 0.2 },\n boost: {\n type: 2.5, // Boost error frames\n digest_text: 2.0,\n outputs: 1.8,\n },\n fallbackStrategy: 'keyword',\n });\n }\n\n async retrieveContext(query: ContextQuery): Promise<ContextRetrievalResult> {\n const startTime = Date.now();\n \n // Handle empty query gracefully\n if (!query.text || query.text.trim().length === 0) {\n logger.debug('Empty query provided, returning empty result');\n return {\n contexts: [],\n totalMatches: 0,\n retrievalTimeMs: Date.now() - startTime,\n strategy: 'empty_query',\n queryAnalysis: {\n intent: 'general',\n concepts: [],\n complexity: 'simple',\n },\n };\n }\n \n const cacheKey = this.generateCacheKey(query);\n\n // Check cache first\n const cached = this.getCachedResult(cacheKey);\n if (cached) {\n logger.debug('Context retrieval cache hit');\n return cached;\n }\n\n try {\n logger.info('Starting LLM-driven context retrieval', {\n query: query.text,\n });\n\n // Analyze query to determine best strategy\n const queryAnalysis = await this.analyzeQuery(query);\n const strategy = this.selectStrategy(queryAnalysis, query);\n\n logger.debug('Selected retrieval strategy', {\n strategy: strategy.name,\n analysis: queryAnalysis,\n });\n\n // Execute retrieval with selected strategy\n const contexts = await this.executeRetrieval(\n query,\n strategy,\n queryAnalysis\n );\n\n // Post-process and rank results\n const rankedContexts = await this.rankAndFilter(\n contexts,\n query,\n queryAnalysis\n );\n\n const result: ContextRetrievalResult = {\n contexts: rankedContexts,\n totalMatches: contexts.length,\n retrievalTimeMs: Date.now() - startTime,\n strategy: strategy.name,\n queryAnalysis,\n };\n\n // Cache result\n this.cacheResult(cacheKey, result);\n\n logger.info('Context retrieval completed', {\n resultsCount: rankedContexts.length,\n timeMs: result.retrievalTimeMs,\n strategy: strategy.name,\n });\n\n return result;\n } catch (error) {\n logger.error('Context retrieval failed:', error);\n\n // Return fallback empty result\n return {\n contexts: [],\n totalMatches: 0,\n retrievalTimeMs: Date.now() - startTime,\n strategy: 'fallback',\n queryAnalysis: {\n intent: 'unknown',\n concepts: [],\n complexity: 'simple',\n },\n };\n }\n }\n\n private async analyzeQuery(query: ContextQuery): Promise<{\n intent: string;\n concepts: string[];\n complexity: 'simple' | 'moderate' | 'complex';\n }> {\n const text = query.text.toLowerCase().trim();\n const words = text.split(/\\s+/);\n\n // Determine intent based on keywords\n let intent = 'general';\n if (\n this.containsKeywords(text, [\n 'error',\n 'exception',\n 'fail',\n 'bug',\n 'issue',\n 'problem',\n 'debug',\n ])\n ) {\n intent = 'debug';\n } else if (\n this.containsKeywords(text, ['how', 'what', 'why', 'when', 'where'])\n ) {\n intent = 'explanation';\n } else if (\n this.containsKeywords(text, [\n 'implement',\n 'create',\n 'build',\n 'add',\n 'develop',\n ])\n ) {\n intent = 'implementation';\n } else if (\n this.containsKeywords(text, [\n 'recent',\n 'latest',\n 'last',\n 'current',\n 'happened',\n ])\n ) {\n intent = 'recent_activity';\n }\n\n // Extract concepts (simplified - in production would use NLP)\n const concepts = this.extractConcepts(text);\n\n // Determine complexity\n let complexity: 'simple' | 'moderate' | 'complex' = 'simple';\n if (words.length > 10 || concepts.length > 5) {\n complexity = 'complex';\n } else if (words.length > 5 || concepts.length > 2) {\n complexity = 'moderate';\n }\n\n return { intent, concepts, complexity };\n }\n\n private containsKeywords(text: string, keywords: string[]): boolean {\n return keywords.some((keyword) =>\n text.toLowerCase().includes(keyword.toLowerCase())\n );\n }\n\n private extractConcepts(text: string): string[] {\n // Simplified concept extraction - in production would use NLP/embeddings\n const technicalTerms = [\n 'database',\n 'sql',\n 'query',\n 'index',\n 'migration',\n 'adapter',\n 'frame',\n 'event',\n 'anchor',\n 'digest',\n 'context',\n 'search',\n 'vector',\n 'embedding',\n 'similarity',\n 'score',\n 'rank',\n 'performance',\n 'optimization',\n 'cache',\n 'pool',\n 'connection',\n 'error',\n 'exception',\n 'debug',\n 'trace',\n 'log',\n 'monitor',\n ];\n\n const concepts: string[] = [];\n const words = text.split(/\\W+/).map((w) => w.toLowerCase());\n\n for (const term of technicalTerms) {\n if (words.includes(term)) {\n concepts.push(term);\n }\n }\n\n // Add bigrams for common technical phrases\n const bigrams = this.extractBigrams(words);\n const technicalBigrams = [\n 'database adapter',\n 'query router',\n 'connection pool',\n 'vector search',\n ];\n\n for (const bigram of bigrams) {\n if (technicalBigrams.includes(bigram)) {\n concepts.push(bigram);\n }\n }\n\n return [...new Set(concepts)]; // Remove duplicates\n }\n\n private extractBigrams(words: string[]): string[] {\n const bigrams: string[] = [];\n for (let i = 0; i < words.length - 1; i++) {\n bigrams.push(`${words[i]} ${words[i + 1]}`);\n }\n return bigrams;\n }\n\n private selectStrategy(\n analysis: { intent: string; complexity: string },\n query: ContextQuery\n ): RetrievalStrategy {\n // Override with explicit query type\n if (query.type) {\n return (\n this.strategies.get(\n query.type === 'keyword'\n ? 'keyword'\n : query.type === 'semantic'\n ? 'semantic'\n : 'hybrid'\n ) || this.strategies.get('hybrid')!\n );\n }\n\n // Select based on intent and complexity\n switch (analysis.intent) {\n case 'debug':\n return this.strategies.get('debug')!;\n case 'recent_activity':\n return this.strategies.get('recent')!;\n case 'explanation':\n return analysis.complexity === 'simple'\n ? this.strategies.get('keyword')!\n : this.strategies.get('semantic')!;\n case 'implementation':\n return this.strategies.get('hybrid')!;\n default:\n return analysis.complexity === 'complex'\n ? this.strategies.get('semantic')!\n : this.strategies.get('keyword')!;\n }\n }\n\n private async executeRetrieval(\n query: ContextQuery,\n strategy: RetrievalStrategy,\n analysis: { intent: string; concepts: string[] }\n ): Promise<RetrievedContext[]> {\n const searchOptions: SearchOptions = {\n query: query.text,\n searchType: strategy.searchType,\n limit: query.maxResults || 20,\n scoreThreshold: query.scoreThreshold || 0.1,\n boost: strategy.boost,\n };\n\n // Add field filtering based on query type\n if (query.frameTypes) {\n searchOptions.fields = ['type', 'name', 'digest_text'];\n }\n\n let rawResults: Array<Frame & { score: number }> = [];\n\n try {\n if (strategy.searchType === 'hybrid' && strategy.weights) {\n // Use hybrid search with embeddings (placeholder - would need actual embeddings)\n const embedding = await this.generateEmbedding(query.text);\n rawResults = await this.adapter.searchHybrid(\n query.text,\n embedding,\n strategy.weights\n );\n } else {\n // Use text or vector search\n rawResults = await this.adapter.search(searchOptions);\n }\n } catch (error) {\n logger.warn(`Strategy ${strategy.name} failed, trying fallback:`, error);\n\n if (strategy.fallbackStrategy) {\n const fallbackStrategy = this.strategies.get(strategy.fallbackStrategy);\n if (fallbackStrategy) {\n return this.executeRetrieval(query, fallbackStrategy, analysis);\n }\n }\n\n // Return empty results instead of throwing to prevent cascading failures\n return [];\n }\n\n // Convert to RetrievedContext objects\n return rawResults.map((result) => ({\n frame: result,\n score: result.score,\n relevanceReason: this.generateRelevanceReason(result, query, analysis),\n retrievalMethod: strategy.searchType as 'bm25' | 'vector' | 'hybrid',\n matchedFields: this.identifyMatchedFields(result, query),\n }));\n }\n\n private async generateEmbedding(text: string): Promise<number[]> {\n // Placeholder - in production would use actual embedding service\n // For now, return a mock embedding\n const hash = this.simpleHash(text);\n return Array.from(\n { length: 384 },\n (_, i) => ((hash + i) % 100) / 100 - 0.5\n );\n }\n\n private simpleHash(str: string): number {\n let hash = 0;\n for (let i = 0; i < str.length; i++) {\n const char = str.charCodeAt(i);\n hash = (hash << 5) - hash + char;\n hash = hash & hash; // Convert to 32-bit integer\n }\n return Math.abs(hash);\n }\n\n private generateRelevanceReason(\n frame: Frame,\n query: ContextQuery,\n analysis: { intent: string; concepts: string[] }\n ): string {\n const reasons: string[] = [];\n\n // Check for direct matches\n if (frame.name.toLowerCase().includes(query.text.toLowerCase())) {\n reasons.push('Frame name matches query');\n }\n\n if (frame.digest_text?.toLowerCase().includes(query.text.toLowerCase())) {\n reasons.push('Content contains query terms');\n }\n\n // Check for concept matches\n for (const concept of analysis.concepts) {\n if (\n frame.digest_text?.toLowerCase().includes(concept.toLowerCase()) ||\n frame.name.toLowerCase().includes(concept.toLowerCase())\n ) {\n reasons.push(`Related to ${concept}`);\n }\n }\n\n // Frame type relevance\n if (analysis.intent === 'debug' && frame.type.includes('error')) {\n reasons.push('Error context for debugging');\n }\n\n return reasons.length > 0\n ? reasons.join('; ')\n : 'General semantic similarity';\n }\n\n private identifyMatchedFields(frame: Frame, query: ContextQuery): string[] {\n const matched: string[] = [];\n const queryLower = query.text.toLowerCase();\n\n if (frame.name.toLowerCase().includes(queryLower)) {\n matched.push('name');\n }\n\n if (frame.digest_text?.toLowerCase().includes(queryLower)) {\n matched.push('digest_text');\n }\n\n if (frame.type.toLowerCase().includes(queryLower)) {\n matched.push('type');\n }\n\n return matched;\n }\n\n private async rankAndFilter(\n contexts: RetrievedContext[],\n query: ContextQuery,\n analysis: { intent: string; complexity: string }\n ): Promise<RetrievedContext[]> {\n // Apply additional filtering\n let filtered = contexts;\n\n // Filter by time range\n if (query.timeRange) {\n filtered = filtered.filter((ctx) => {\n const frameTime = new Date(ctx.frame.created_at);\n const start = query.timeRange?.start;\n const end = query.timeRange?.end;\n\n return (!start || frameTime >= start) && (!end || frameTime <= end);\n });\n }\n\n // Filter by frame types\n if (query.frameTypes) {\n filtered = filtered.filter((ctx) =>\n query.frameTypes!.includes(ctx.frame.type)\n );\n }\n\n // Apply score threshold\n if (query.scoreThreshold) {\n filtered = filtered.filter((ctx) => ctx.score >= query.scoreThreshold!);\n }\n\n // Enhanced ranking based on multiple factors\n const ranked = filtered.map((ctx) => ({\n ...ctx,\n score: this.calculateEnhancedScore(ctx, query, analysis),\n }));\n\n // Sort by enhanced score\n ranked.sort((a, b) => b.score - a.score);\n\n // Limit results\n const maxResults = query.maxResults || 20;\n return ranked.slice(0, maxResults);\n }\n\n private calculateEnhancedScore(\n context: RetrievedContext,\n query: ContextQuery,\n analysis: { intent: string; concepts: string[] }\n ): number {\n let score = context.score;\n\n // Boost recent frames\n const ageHours = (Date.now() - context.frame.created_at) / (1000 * 60 * 60);\n if (ageHours < 24) {\n score *= 1.2; // 20% boost for frames from last 24 hours\n } else if (ageHours < 168) {\n // 1 week\n score *= 1.1; // 10% boost for frames from last week\n }\n\n // Boost based on frame completeness\n if (context.frame.closed_at) {\n score *= 1.1; // Completed frames are more valuable\n }\n\n // Boost based on intent matching\n if (analysis.intent === 'debug' && context.frame.type.includes('error')) {\n score *= 1.5;\n }\n\n // Boost based on matched fields\n if (context.matchedFields.includes('name')) {\n score *= 1.3; // Name matches are highly relevant\n }\n\n if (context.matchedFields.length > 1) {\n score *= 1.1; // Multiple field matches\n }\n\n // Penalize very old frames for recent queries\n if (analysis.intent === 'recent_activity' && ageHours > 168) {\n score *= 0.5;\n }\n\n return score;\n }\n\n private generateCacheKey(query: ContextQuery): string {\n return JSON.stringify({\n text: query.text,\n type: query.type,\n maxResults: query.maxResults,\n frameTypes: query.frameTypes,\n scoreThreshold: query.scoreThreshold,\n });\n }\n\n private getCachedResult(cacheKey: string): ContextRetrievalResult | null {\n const entry = this.queryCache.get(cacheKey);\n if (!entry) return null;\n\n // Check expiry (simplified - would include timestamp in real implementation)\n return entry;\n }\n\n private cacheResult(cacheKey: string, result: ContextRetrievalResult): void {\n // Implement LRU eviction if cache is full\n if (this.queryCache.size >= this.cacheMaxSize) {\n const firstKey = this.queryCache.keys().next().value;\n this.queryCache.delete(firstKey);\n }\n\n this.queryCache.set(cacheKey, result);\n }\n\n // Utility methods for integration\n async findSimilarFrames(\n frameId: string,\n limit = 10\n ): Promise<RetrievedContext[]> {\n const frame = await this.adapter.getFrame(frameId);\n if (!frame) {\n throw new Error(`Frame not found: ${frameId}`);\n }\n\n const query: ContextQuery = {\n text: frame.digest_text || frame.name,\n type: 'semantic',\n maxResults: limit,\n scoreThreshold: 0.3,\n };\n\n const result = await this.retrieveContext(query);\n\n // Filter out the original frame\n return result.contexts.filter((ctx) => ctx.frame.frame_id !== frameId);\n }\n\n async findContextForError(\n errorMessage: string,\n stackTrace?: string\n ): Promise<RetrievedContext[]> {\n const query: ContextQuery = {\n text: `${errorMessage} ${stackTrace || ''}`.trim(),\n type: 'hybrid',\n maxResults: 15,\n frameTypes: ['error', 'debug', 'function'],\n scoreThreshold: 0.2,\n };\n\n const result = await this.retrieveContext(query);\n return result.contexts;\n }\n\n async getRecentContext(\n hours = 24,\n frameTypes?: string[]\n ): Promise<RetrievedContext[]> {\n const query: ContextQuery = {\n text: 'recent activity context',\n type: 'keyword',\n maxResults: 50,\n timeRange: {\n start: new Date(Date.now() - hours * 60 * 60 * 1000),\n },\n frameTypes,\n scoreThreshold: 0.1,\n };\n\n const result = await this.retrieveContext(query);\n return result.contexts;\n }\n\n // Analytics and insights\n getRetrievalStats() {\n return {\n cacheSize: this.queryCache.size,\n strategiesCount: this.strategies.size,\n availableStrategies: Array.from(this.strategies.keys()),\n };\n }\n\n clearCache(): void {\n this.queryCache.clear();\n logger.info('Context retrieval cache cleared');\n }\n}\n"],
|
|
5
|
-
"mappings": "AAUA,SAAS,cAAc;AA2ChB,MAAM,iBAAiB;AAAA,EACX;AAAA,EACA,aAA6C,oBAAI,IAAI;AAAA,EAC9D,aAAa,oBAAI,IAAoC;AAAA,EACrD,eAAe;AAAA,EACf,gBAAgB;AAAA;AAAA,EAExB,YAAY,SAA0B;AACpC,SAAK,UAAU;AACf,SAAK,qBAAqB;AAAA,EAC5B;AAAA,EAEQ,uBAA6B;AAEnC,SAAK,WAAW,IAAI,WAAW;AAAA,MAC7B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,OAAO;AAAA,QACL,MAAM;AAAA,QACN,aAAa;AAAA,QACb,QAAQ;AAAA,QACR,SAAS;AAAA,MACX;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,YAAY;AAAA,MAC9B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,UAAU;AAAA,MAC5B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,SAAS,EAAE,MAAM,KAAK,QAAQ,IAAI;AAAA,MAClC,OAAO;AAAA,QACL,MAAM;AAAA,QACN,aAAa;AAAA,MACf;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,UAAU;AAAA,MAC5B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,OAAO;AAAA,QACL,YAAY;AAAA,QACZ,WAAW;AAAA,MACb;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,SAAS;AAAA,MAC3B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,SAAS,EAAE,MAAM,KAAK,QAAQ,IAAI;AAAA,MAClC,OAAO;AAAA,QACL,MAAM;AAAA;AAAA,QACN,aAAa;AAAA,QACb,SAAS;AAAA,MACX;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,gBAAgB,OAAsD;AAC1E,UAAM,YAAY,KAAK,IAAI;AAG3B,QAAI,CAAC,MAAM,QAAQ,MAAM,KAAK,KAAK,EAAE,WAAW,GAAG;AACjD,aAAO,MAAM,8CAA8C;AAC3D,aAAO;AAAA,QACL,UAAU,CAAC;AAAA,QACX,cAAc;AAAA,QACd,iBAAiB,KAAK,IAAI,IAAI;AAAA,QAC9B,UAAU;AAAA,QACV,eAAe;AAAA,UACb,QAAQ;AAAA,UACR,UAAU,CAAC;AAAA,UACX,YAAY;AAAA,QACd;AAAA,MACF;AAAA,IACF;AAEA,UAAM,WAAW,KAAK,iBAAiB,KAAK;AAG5C,UAAM,SAAS,KAAK,gBAAgB,QAAQ;AAC5C,QAAI,QAAQ;AACV,aAAO,MAAM,6BAA6B;AAC1C,aAAO;AAAA,IACT;AAEA,QAAI;AACF,aAAO,KAAK,yCAAyC;AAAA,QACnD,OAAO,MAAM;AAAA,MACf,CAAC;AAGD,YAAM,gBAAgB,MAAM,KAAK,aAAa,KAAK;AACnD,YAAM,WAAW,KAAK,eAAe,eAAe,KAAK;AAEzD,aAAO,MAAM,+BAA+B;AAAA,QAC1C,UAAU,SAAS;AAAA,QACnB,UAAU;AAAA,MACZ,CAAC;AAGD,YAAM,WAAW,MAAM,KAAK;AAAA,QAC1B;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAGA,YAAM,iBAAiB,MAAM,KAAK;AAAA,QAChC;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,YAAM,SAAiC;AAAA,QACrC,UAAU;AAAA,QACV,cAAc,SAAS;AAAA,QACvB,iBAAiB,KAAK,IAAI,IAAI;AAAA,QAC9B,UAAU,SAAS;AAAA,QACnB;AAAA,MACF;AAGA,WAAK,YAAY,UAAU,MAAM;AAEjC,aAAO,KAAK,+BAA+B;AAAA,QACzC,cAAc,eAAe;AAAA,QAC7B,QAAQ,OAAO;AAAA,QACf,UAAU,SAAS;AAAA,MACrB,CAAC;AAED,aAAO;AAAA,IACT,SAAS,
|
|
4
|
+
"sourcesContent": ["/**\n * LLM-driven Context Retrieval System\n * Intelligently retrieves relevant context using ParadeDB search capabilities\n */\n\nimport {\n DatabaseAdapter,\n SearchOptions,\n} from '../database/database-adapter.js';\nimport { Frame } from '../context/frame-manager.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport interface ContextQuery {\n text: string;\n type?: 'semantic' | 'keyword' | 'hybrid';\n maxResults?: number;\n timeRange?: {\n start?: Date;\n end?: Date;\n };\n frameTypes?: string[];\n scoreThreshold?: number;\n includeDigests?: boolean;\n}\n\nexport interface RetrievedContext {\n frame: Frame;\n score: number;\n relevanceReason: string;\n retrievalMethod: 'bm25' | 'vector' | 'hybrid';\n matchedFields: string[];\n}\n\nexport interface ContextRetrievalResult {\n contexts: RetrievedContext[];\n totalMatches: number;\n retrievalTimeMs: number;\n strategy: string;\n queryAnalysis: {\n intent: string;\n concepts: string[];\n complexity: 'simple' | 'moderate' | 'complex';\n };\n}\n\nexport interface RetrievalStrategy {\n name: string;\n searchType: 'text' | 'vector' | 'hybrid';\n weights?: { text: number; vector: number };\n boost?: Record<string, number>;\n fallbackStrategy?: string;\n}\n\nexport class ContextRetriever {\n private readonly adapter: DatabaseAdapter;\n private readonly strategies: Map<string, RetrievalStrategy> = new Map();\n private queryCache = new Map<string, ContextRetrievalResult>();\n private cacheMaxSize = 100;\n private cacheExpiryMs = 300000; // 5 minutes\n\n constructor(adapter: DatabaseAdapter) {\n this.adapter = adapter;\n this.initializeStrategies();\n }\n\n private initializeStrategies(): void {\n // Keyword-based search for specific terms\n this.strategies.set('keyword', {\n name: 'Keyword Search',\n searchType: 'text',\n boost: {\n name: 2.0,\n digest_text: 1.5,\n inputs: 1.2,\n outputs: 1.2,\n },\n fallbackStrategy: 'semantic',\n });\n\n // Semantic search using vector embeddings\n this.strategies.set('semantic', {\n name: 'Semantic Search',\n searchType: 'vector',\n fallbackStrategy: 'hybrid',\n });\n\n // Hybrid approach combining text and vector search\n this.strategies.set('hybrid', {\n name: 'Hybrid Search',\n searchType: 'hybrid',\n weights: { text: 0.6, vector: 0.4 },\n boost: {\n name: 2.0,\n digest_text: 1.5,\n },\n fallbackStrategy: 'keyword',\n });\n\n // Recent activity search\n this.strategies.set('recent', {\n name: 'Recent Activity',\n searchType: 'text',\n boost: {\n created_at: 3.0,\n closed_at: 2.0,\n },\n fallbackStrategy: 'hybrid',\n });\n\n // Error and debugging context\n this.strategies.set('debug', {\n name: 'Debug Context',\n searchType: 'hybrid',\n weights: { text: 0.8, vector: 0.2 },\n boost: {\n type: 2.5, // Boost error frames\n digest_text: 2.0,\n outputs: 1.8,\n },\n fallbackStrategy: 'keyword',\n });\n }\n\n async retrieveContext(query: ContextQuery): Promise<ContextRetrievalResult> {\n const startTime = Date.now();\n \n // Handle empty query gracefully\n if (!query.text || query.text.trim().length === 0) {\n logger.debug('Empty query provided, returning empty result');\n return {\n contexts: [],\n totalMatches: 0,\n retrievalTimeMs: Date.now() - startTime,\n strategy: 'empty_query',\n queryAnalysis: {\n intent: 'general',\n concepts: [],\n complexity: 'simple',\n },\n };\n }\n \n const cacheKey = this.generateCacheKey(query);\n\n // Check cache first\n const cached = this.getCachedResult(cacheKey);\n if (cached) {\n logger.debug('Context retrieval cache hit');\n return cached;\n }\n\n try {\n logger.info('Starting LLM-driven context retrieval', {\n query: query.text,\n });\n\n // Analyze query to determine best strategy\n const queryAnalysis = await this.analyzeQuery(query);\n const strategy = this.selectStrategy(queryAnalysis, query);\n\n logger.debug('Selected retrieval strategy', {\n strategy: strategy.name,\n analysis: queryAnalysis,\n });\n\n // Execute retrieval with selected strategy\n const contexts = await this.executeRetrieval(\n query,\n strategy,\n queryAnalysis\n );\n\n // Post-process and rank results\n const rankedContexts = await this.rankAndFilter(\n contexts,\n query,\n queryAnalysis\n );\n\n const result: ContextRetrievalResult = {\n contexts: rankedContexts,\n totalMatches: contexts.length,\n retrievalTimeMs: Date.now() - startTime,\n strategy: strategy.name,\n queryAnalysis,\n };\n\n // Cache result\n this.cacheResult(cacheKey, result);\n\n logger.info('Context retrieval completed', {\n resultsCount: rankedContexts.length,\n timeMs: result.retrievalTimeMs,\n strategy: strategy.name,\n });\n\n return result;\n } catch (error: unknown) {\n logger.error('Context retrieval failed:', error);\n\n // Return fallback empty result\n return {\n contexts: [],\n totalMatches: 0,\n retrievalTimeMs: Date.now() - startTime,\n strategy: 'fallback',\n queryAnalysis: {\n intent: 'unknown',\n concepts: [],\n complexity: 'simple',\n },\n };\n }\n }\n\n private async analyzeQuery(query: ContextQuery): Promise<{\n intent: string;\n concepts: string[];\n complexity: 'simple' | 'moderate' | 'complex';\n }> {\n const text = query.text.toLowerCase().trim();\n const words = text.split(/\\s+/);\n\n // Determine intent based on keywords\n let intent = 'general';\n if (\n this.containsKeywords(text, [\n 'error',\n 'exception',\n 'fail',\n 'bug',\n 'issue',\n 'problem',\n 'debug',\n ])\n ) {\n intent = 'debug';\n } else if (\n this.containsKeywords(text, ['how', 'what', 'why', 'when', 'where'])\n ) {\n intent = 'explanation';\n } else if (\n this.containsKeywords(text, [\n 'implement',\n 'create',\n 'build',\n 'add',\n 'develop',\n ])\n ) {\n intent = 'implementation';\n } else if (\n this.containsKeywords(text, [\n 'recent',\n 'latest',\n 'last',\n 'current',\n 'happened',\n ])\n ) {\n intent = 'recent_activity';\n }\n\n // Extract concepts (simplified - in production would use NLP)\n const concepts = this.extractConcepts(text);\n\n // Determine complexity\n let complexity: 'simple' | 'moderate' | 'complex' = 'simple';\n if (words.length > 10 || concepts.length > 5) {\n complexity = 'complex';\n } else if (words.length > 5 || concepts.length > 2) {\n complexity = 'moderate';\n }\n\n return { intent, concepts, complexity };\n }\n\n private containsKeywords(text: string, keywords: string[]): boolean {\n return keywords.some((keyword) =>\n text.toLowerCase().includes(keyword.toLowerCase())\n );\n }\n\n private extractConcepts(text: string): string[] {\n // Simplified concept extraction - in production would use NLP/embeddings\n const technicalTerms = [\n 'database',\n 'sql',\n 'query',\n 'index',\n 'migration',\n 'adapter',\n 'frame',\n 'event',\n 'anchor',\n 'digest',\n 'context',\n 'search',\n 'vector',\n 'embedding',\n 'similarity',\n 'score',\n 'rank',\n 'performance',\n 'optimization',\n 'cache',\n 'pool',\n 'connection',\n 'error',\n 'exception',\n 'debug',\n 'trace',\n 'log',\n 'monitor',\n ];\n\n const concepts: string[] = [];\n const words = text.split(/\\W+/).map((w) => w.toLowerCase());\n\n for (const term of technicalTerms) {\n if (words.includes(term)) {\n concepts.push(term);\n }\n }\n\n // Add bigrams for common technical phrases\n const bigrams = this.extractBigrams(words);\n const technicalBigrams = [\n 'database adapter',\n 'query router',\n 'connection pool',\n 'vector search',\n ];\n\n for (const bigram of bigrams) {\n if (technicalBigrams.includes(bigram)) {\n concepts.push(bigram);\n }\n }\n\n return [...new Set(concepts)]; // Remove duplicates\n }\n\n private extractBigrams(words: string[]): string[] {\n const bigrams: string[] = [];\n for (let i = 0; i < words.length - 1; i++) {\n bigrams.push(`${words[i]} ${words[i + 1]}`);\n }\n return bigrams;\n }\n\n private selectStrategy(\n analysis: { intent: string; complexity: string },\n query: ContextQuery\n ): RetrievalStrategy {\n // Override with explicit query type\n if (query.type) {\n return (\n this.strategies.get(\n query.type === 'keyword'\n ? 'keyword'\n : query.type === 'semantic'\n ? 'semantic'\n : 'hybrid'\n ) || this.strategies.get('hybrid')!\n );\n }\n\n // Select based on intent and complexity\n switch (analysis.intent) {\n case 'debug':\n return this.strategies.get('debug')!;\n case 'recent_activity':\n return this.strategies.get('recent')!;\n case 'explanation':\n return analysis.complexity === 'simple'\n ? this.strategies.get('keyword')!\n : this.strategies.get('semantic')!;\n case 'implementation':\n return this.strategies.get('hybrid')!;\n default:\n return analysis.complexity === 'complex'\n ? this.strategies.get('semantic')!\n : this.strategies.get('keyword')!;\n }\n }\n\n private async executeRetrieval(\n query: ContextQuery,\n strategy: RetrievalStrategy,\n analysis: { intent: string; concepts: string[] }\n ): Promise<RetrievedContext[]> {\n const searchOptions: SearchOptions = {\n query: query.text,\n searchType: strategy.searchType,\n limit: query.maxResults || 20,\n scoreThreshold: query.scoreThreshold || 0.1,\n boost: strategy.boost,\n };\n\n // Add field filtering based on query type\n if (query.frameTypes) {\n searchOptions.fields = ['type', 'name', 'digest_text'];\n }\n\n let rawResults: Array<Frame & { score: number }> = [];\n\n try {\n if (strategy.searchType === 'hybrid' && strategy.weights) {\n // Use hybrid search with embeddings (placeholder - would need actual embeddings)\n const embedding = await this.generateEmbedding(query.text);\n rawResults = await this.adapter.searchHybrid(\n query.text,\n embedding,\n strategy.weights\n );\n } else {\n // Use text or vector search\n rawResults = await this.adapter.search(searchOptions);\n }\n } catch (error: unknown) {\n logger.warn(`Strategy ${strategy.name} failed, trying fallback:`, error);\n\n if (strategy.fallbackStrategy) {\n const fallbackStrategy = this.strategies.get(strategy.fallbackStrategy);\n if (fallbackStrategy) {\n return this.executeRetrieval(query, fallbackStrategy, analysis);\n }\n }\n\n // Return empty results instead of throwing to prevent cascading failures\n return [];\n }\n\n // Convert to RetrievedContext objects\n return rawResults.map((result) => ({\n frame: result,\n score: result.score,\n relevanceReason: this.generateRelevanceReason(result, query, analysis),\n retrievalMethod: strategy.searchType as 'bm25' | 'vector' | 'hybrid',\n matchedFields: this.identifyMatchedFields(result, query),\n }));\n }\n\n private async generateEmbedding(text: string): Promise<number[]> {\n // Placeholder - in production would use actual embedding service\n // For now, return a mock embedding\n const hash = this.simpleHash(text);\n return Array.from(\n { length: 384 },\n (_, i) => ((hash + i) % 100) / 100 - 0.5\n );\n }\n\n private simpleHash(str: string): number {\n let hash = 0;\n for (let i = 0; i < str.length; i++) {\n const char = str.charCodeAt(i);\n hash = (hash << 5) - hash + char;\n hash = hash & hash; // Convert to 32-bit integer\n }\n return Math.abs(hash);\n }\n\n private generateRelevanceReason(\n frame: Frame,\n query: ContextQuery,\n analysis: { intent: string; concepts: string[] }\n ): string {\n const reasons: string[] = [];\n\n // Check for direct matches\n if (frame.name.toLowerCase().includes(query.text.toLowerCase())) {\n reasons.push('Frame name matches query');\n }\n\n if (frame.digest_text?.toLowerCase().includes(query.text.toLowerCase())) {\n reasons.push('Content contains query terms');\n }\n\n // Check for concept matches\n for (const concept of analysis.concepts) {\n if (\n frame.digest_text?.toLowerCase().includes(concept.toLowerCase()) ||\n frame.name.toLowerCase().includes(concept.toLowerCase())\n ) {\n reasons.push(`Related to ${concept}`);\n }\n }\n\n // Frame type relevance\n if (analysis.intent === 'debug' && frame.type.includes('error')) {\n reasons.push('Error context for debugging');\n }\n\n return reasons.length > 0\n ? reasons.join('; ')\n : 'General semantic similarity';\n }\n\n private identifyMatchedFields(frame: Frame, query: ContextQuery): string[] {\n const matched: string[] = [];\n const queryLower = query.text.toLowerCase();\n\n if (frame.name.toLowerCase().includes(queryLower)) {\n matched.push('name');\n }\n\n if (frame.digest_text?.toLowerCase().includes(queryLower)) {\n matched.push('digest_text');\n }\n\n if (frame.type.toLowerCase().includes(queryLower)) {\n matched.push('type');\n }\n\n return matched;\n }\n\n private async rankAndFilter(\n contexts: RetrievedContext[],\n query: ContextQuery,\n analysis: { intent: string; complexity: string }\n ): Promise<RetrievedContext[]> {\n // Apply additional filtering\n let filtered = contexts;\n\n // Filter by time range\n if (query.timeRange) {\n filtered = filtered.filter((ctx) => {\n const frameTime = new Date(ctx.frame.created_at);\n const start = query.timeRange?.start;\n const end = query.timeRange?.end;\n\n return (!start || frameTime >= start) && (!end || frameTime <= end);\n });\n }\n\n // Filter by frame types\n if (query.frameTypes) {\n filtered = filtered.filter((ctx) =>\n query.frameTypes!.includes(ctx.frame.type)\n );\n }\n\n // Apply score threshold\n if (query.scoreThreshold) {\n filtered = filtered.filter((ctx) => ctx.score >= query.scoreThreshold!);\n }\n\n // Enhanced ranking based on multiple factors\n const ranked = filtered.map((ctx) => ({\n ...ctx,\n score: this.calculateEnhancedScore(ctx, query, analysis),\n }));\n\n // Sort by enhanced score\n ranked.sort((a, b) => b.score - a.score);\n\n // Limit results\n const maxResults = query.maxResults || 20;\n return ranked.slice(0, maxResults);\n }\n\n private calculateEnhancedScore(\n context: RetrievedContext,\n query: ContextQuery,\n analysis: { intent: string; concepts: string[] }\n ): number {\n let score = context.score;\n\n // Boost recent frames\n const ageHours = (Date.now() - context.frame.created_at) / (1000 * 60 * 60);\n if (ageHours < 24) {\n score *= 1.2; // 20% boost for frames from last 24 hours\n } else if (ageHours < 168) {\n // 1 week\n score *= 1.1; // 10% boost for frames from last week\n }\n\n // Boost based on frame completeness\n if (context.frame.closed_at) {\n score *= 1.1; // Completed frames are more valuable\n }\n\n // Boost based on intent matching\n if (analysis.intent === 'debug' && context.frame.type.includes('error')) {\n score *= 1.5;\n }\n\n // Boost based on matched fields\n if (context.matchedFields.includes('name')) {\n score *= 1.3; // Name matches are highly relevant\n }\n\n if (context.matchedFields.length > 1) {\n score *= 1.1; // Multiple field matches\n }\n\n // Penalize very old frames for recent queries\n if (analysis.intent === 'recent_activity' && ageHours > 168) {\n score *= 0.5;\n }\n\n return score;\n }\n\n private generateCacheKey(query: ContextQuery): string {\n return JSON.stringify({\n text: query.text,\n type: query.type,\n maxResults: query.maxResults,\n frameTypes: query.frameTypes,\n scoreThreshold: query.scoreThreshold,\n });\n }\n\n private getCachedResult(cacheKey: string): ContextRetrievalResult | null {\n const entry = this.queryCache.get(cacheKey);\n if (!entry) return null;\n\n // Check expiry (simplified - would include timestamp in real implementation)\n return entry;\n }\n\n private cacheResult(cacheKey: string, result: ContextRetrievalResult): void {\n // Implement LRU eviction if cache is full\n if (this.queryCache.size >= this.cacheMaxSize) {\n const firstKey = this.queryCache.keys().next().value;\n this.queryCache.delete(firstKey);\n }\n\n this.queryCache.set(cacheKey, result);\n }\n\n // Utility methods for integration\n async findSimilarFrames(\n frameId: string,\n limit = 10\n ): Promise<RetrievedContext[]> {\n const frame = await this.adapter.getFrame(frameId);\n if (!frame) {\n throw new Error(`Frame not found: ${frameId}`);\n }\n\n const query: ContextQuery = {\n text: frame.digest_text || frame.name,\n type: 'semantic',\n maxResults: limit,\n scoreThreshold: 0.3,\n };\n\n const result = await this.retrieveContext(query);\n\n // Filter out the original frame\n return result.contexts.filter((ctx) => ctx.frame.frame_id !== frameId);\n }\n\n async findContextForError(\n errorMessage: string,\n stackTrace?: string\n ): Promise<RetrievedContext[]> {\n const query: ContextQuery = {\n text: `${errorMessage} ${stackTrace || ''}`.trim(),\n type: 'hybrid',\n maxResults: 15,\n frameTypes: ['error', 'debug', 'function'],\n scoreThreshold: 0.2,\n };\n\n const result = await this.retrieveContext(query);\n return result.contexts;\n }\n\n async getRecentContext(\n hours = 24,\n frameTypes?: string[]\n ): Promise<RetrievedContext[]> {\n const query: ContextQuery = {\n text: 'recent activity context',\n type: 'keyword',\n maxResults: 50,\n timeRange: {\n start: new Date(Date.now() - hours * 60 * 60 * 1000),\n },\n frameTypes,\n scoreThreshold: 0.1,\n };\n\n const result = await this.retrieveContext(query);\n return result.contexts;\n }\n\n // Analytics and insights\n getRetrievalStats() {\n return {\n cacheSize: this.queryCache.size,\n strategiesCount: this.strategies.size,\n availableStrategies: Array.from(this.strategies.keys()),\n };\n }\n\n clearCache(): void {\n this.queryCache.clear();\n logger.info('Context retrieval cache cleared');\n }\n}\n"],
|
|
5
|
+
"mappings": "AAUA,SAAS,cAAc;AA2ChB,MAAM,iBAAiB;AAAA,EACX;AAAA,EACA,aAA6C,oBAAI,IAAI;AAAA,EAC9D,aAAa,oBAAI,IAAoC;AAAA,EACrD,eAAe;AAAA,EACf,gBAAgB;AAAA;AAAA,EAExB,YAAY,SAA0B;AACpC,SAAK,UAAU;AACf,SAAK,qBAAqB;AAAA,EAC5B;AAAA,EAEQ,uBAA6B;AAEnC,SAAK,WAAW,IAAI,WAAW;AAAA,MAC7B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,OAAO;AAAA,QACL,MAAM;AAAA,QACN,aAAa;AAAA,QACb,QAAQ;AAAA,QACR,SAAS;AAAA,MACX;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,YAAY;AAAA,MAC9B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,UAAU;AAAA,MAC5B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,SAAS,EAAE,MAAM,KAAK,QAAQ,IAAI;AAAA,MAClC,OAAO;AAAA,QACL,MAAM;AAAA,QACN,aAAa;AAAA,MACf;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,UAAU;AAAA,MAC5B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,OAAO;AAAA,QACL,YAAY;AAAA,QACZ,WAAW;AAAA,MACb;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAGD,SAAK,WAAW,IAAI,SAAS;AAAA,MAC3B,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,SAAS,EAAE,MAAM,KAAK,QAAQ,IAAI;AAAA,MAClC,OAAO;AAAA,QACL,MAAM;AAAA;AAAA,QACN,aAAa;AAAA,QACb,SAAS;AAAA,MACX;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,gBAAgB,OAAsD;AAC1E,UAAM,YAAY,KAAK,IAAI;AAG3B,QAAI,CAAC,MAAM,QAAQ,MAAM,KAAK,KAAK,EAAE,WAAW,GAAG;AACjD,aAAO,MAAM,8CAA8C;AAC3D,aAAO;AAAA,QACL,UAAU,CAAC;AAAA,QACX,cAAc;AAAA,QACd,iBAAiB,KAAK,IAAI,IAAI;AAAA,QAC9B,UAAU;AAAA,QACV,eAAe;AAAA,UACb,QAAQ;AAAA,UACR,UAAU,CAAC;AAAA,UACX,YAAY;AAAA,QACd;AAAA,MACF;AAAA,IACF;AAEA,UAAM,WAAW,KAAK,iBAAiB,KAAK;AAG5C,UAAM,SAAS,KAAK,gBAAgB,QAAQ;AAC5C,QAAI,QAAQ;AACV,aAAO,MAAM,6BAA6B;AAC1C,aAAO;AAAA,IACT;AAEA,QAAI;AACF,aAAO,KAAK,yCAAyC;AAAA,QACnD,OAAO,MAAM;AAAA,MACf,CAAC;AAGD,YAAM,gBAAgB,MAAM,KAAK,aAAa,KAAK;AACnD,YAAM,WAAW,KAAK,eAAe,eAAe,KAAK;AAEzD,aAAO,MAAM,+BAA+B;AAAA,QAC1C,UAAU,SAAS;AAAA,QACnB,UAAU;AAAA,MACZ,CAAC;AAGD,YAAM,WAAW,MAAM,KAAK;AAAA,QAC1B;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAGA,YAAM,iBAAiB,MAAM,KAAK;AAAA,QAChC;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,YAAM,SAAiC;AAAA,QACrC,UAAU;AAAA,QACV,cAAc,SAAS;AAAA,QACvB,iBAAiB,KAAK,IAAI,IAAI;AAAA,QAC9B,UAAU,SAAS;AAAA,QACnB;AAAA,MACF;AAGA,WAAK,YAAY,UAAU,MAAM;AAEjC,aAAO,KAAK,+BAA+B;AAAA,QACzC,cAAc,eAAe;AAAA,QAC7B,QAAQ,OAAO;AAAA,QACf,UAAU,SAAS;AAAA,MACrB,CAAC;AAED,aAAO;AAAA,IACT,SAAS,OAAgB;AACvB,aAAO,MAAM,6BAA6B,KAAK;AAG/C,aAAO;AAAA,QACL,UAAU,CAAC;AAAA,QACX,cAAc;AAAA,QACd,iBAAiB,KAAK,IAAI,IAAI;AAAA,QAC9B,UAAU;AAAA,QACV,eAAe;AAAA,UACb,QAAQ;AAAA,UACR,UAAU,CAAC;AAAA,UACX,YAAY;AAAA,QACd;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAc,aAAa,OAIxB;AACD,UAAM,OAAO,MAAM,KAAK,YAAY,EAAE,KAAK;AAC3C,UAAM,QAAQ,KAAK,MAAM,KAAK;AAG9B,QAAI,SAAS;AACb,QACE,KAAK,iBAAiB,MAAM;AAAA,MAC1B;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,CAAC,GACD;AACA,eAAS;AAAA,IACX,WACE,KAAK,iBAAiB,MAAM,CAAC,OAAO,QAAQ,OAAO,QAAQ,OAAO,CAAC,GACnE;AACA,eAAS;AAAA,IACX,WACE,KAAK,iBAAiB,MAAM;AAAA,MAC1B;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,CAAC,GACD;AACA,eAAS;AAAA,IACX,WACE,KAAK,iBAAiB,MAAM;AAAA,MAC1B;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,CAAC,GACD;AACA,eAAS;AAAA,IACX;AAGA,UAAM,WAAW,KAAK,gBAAgB,IAAI;AAG1C,QAAI,aAAgD;AACpD,QAAI,MAAM,SAAS,MAAM,SAAS,SAAS,GAAG;AAC5C,mBAAa;AAAA,IACf,WAAW,MAAM,SAAS,KAAK,SAAS,SAAS,GAAG;AAClD,mBAAa;AAAA,IACf;AAEA,WAAO,EAAE,QAAQ,UAAU,WAAW;AAAA,EACxC;AAAA,EAEQ,iBAAiB,MAAc,UAA6B;AAClE,WAAO,SAAS;AAAA,MAAK,CAAC,YACpB,KAAK,YAAY,EAAE,SAAS,QAAQ,YAAY,CAAC;AAAA,IACnD;AAAA,EACF;AAAA,EAEQ,gBAAgB,MAAwB;AAE9C,UAAM,iBAAiB;AAAA,MACrB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAEA,UAAM,WAAqB,CAAC;AAC5B,UAAM,QAAQ,KAAK,MAAM,KAAK,EAAE,IAAI,CAAC,MAAM,EAAE,YAAY,CAAC;AAE1D,eAAW,QAAQ,gBAAgB;AACjC,UAAI,MAAM,SAAS,IAAI,GAAG;AACxB,iBAAS,KAAK,IAAI;AAAA,MACpB;AAAA,IACF;AAGA,UAAM,UAAU,KAAK,eAAe,KAAK;AACzC,UAAM,mBAAmB;AAAA,MACvB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAEA,eAAW,UAAU,SAAS;AAC5B,UAAI,iBAAiB,SAAS,MAAM,GAAG;AACrC,iBAAS,KAAK,MAAM;AAAA,MACtB;AAAA,IACF;AAEA,WAAO,CAAC,GAAG,IAAI,IAAI,QAAQ,CAAC;AAAA,EAC9B;AAAA,EAEQ,eAAe,OAA2B;AAChD,UAAM,UAAoB,CAAC;AAC3B,aAAS,IAAI,GAAG,IAAI,MAAM,SAAS,GAAG,KAAK;AACzC,cAAQ,KAAK,GAAG,MAAM,CAAC,CAAC,IAAI,MAAM,IAAI,CAAC,CAAC,EAAE;AAAA,IAC5C;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,eACN,UACA,OACmB;AAEnB,QAAI,MAAM,MAAM;AACd,aACE,KAAK,WAAW;AAAA,QACd,MAAM,SAAS,YACX,YACA,MAAM,SAAS,aACb,aACA;AAAA,MACR,KAAK,KAAK,WAAW,IAAI,QAAQ;AAAA,IAErC;AAGA,YAAQ,SAAS,QAAQ;AAAA,MACvB,KAAK;AACH,eAAO,KAAK,WAAW,IAAI,OAAO;AAAA,MACpC,KAAK;AACH,eAAO,KAAK,WAAW,IAAI,QAAQ;AAAA,MACrC,KAAK;AACH,eAAO,SAAS,eAAe,WAC3B,KAAK,WAAW,IAAI,SAAS,IAC7B,KAAK,WAAW,IAAI,UAAU;AAAA,MACpC,KAAK;AACH,eAAO,KAAK,WAAW,IAAI,QAAQ;AAAA,MACrC;AACE,eAAO,SAAS,eAAe,YAC3B,KAAK,WAAW,IAAI,UAAU,IAC9B,KAAK,WAAW,IAAI,SAAS;AAAA,IACrC;AAAA,EACF;AAAA,EAEA,MAAc,iBACZ,OACA,UACA,UAC6B;AAC7B,UAAM,gBAA+B;AAAA,MACnC,OAAO,MAAM;AAAA,MACb,YAAY,SAAS;AAAA,MACrB,OAAO,MAAM,cAAc;AAAA,MAC3B,gBAAgB,MAAM,kBAAkB;AAAA,MACxC,OAAO,SAAS;AAAA,IAClB;AAGA,QAAI,MAAM,YAAY;AACpB,oBAAc,SAAS,CAAC,QAAQ,QAAQ,aAAa;AAAA,IACvD;AAEA,QAAI,aAA+C,CAAC;AAEpD,QAAI;AACF,UAAI,SAAS,eAAe,YAAY,SAAS,SAAS;AAExD,cAAM,YAAY,MAAM,KAAK,kBAAkB,MAAM,IAAI;AACzD,qBAAa,MAAM,KAAK,QAAQ;AAAA,UAC9B,MAAM;AAAA,UACN;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF,OAAO;AAEL,qBAAa,MAAM,KAAK,QAAQ,OAAO,aAAa;AAAA,MACtD;AAAA,IACF,SAAS,OAAgB;AACvB,aAAO,KAAK,YAAY,SAAS,IAAI,6BAA6B,KAAK;AAEvE,UAAI,SAAS,kBAAkB;AAC7B,cAAM,mBAAmB,KAAK,WAAW,IAAI,SAAS,gBAAgB;AACtE,YAAI,kBAAkB;AACpB,iBAAO,KAAK,iBAAiB,OAAO,kBAAkB,QAAQ;AAAA,QAChE;AAAA,MACF;AAGA,aAAO,CAAC;AAAA,IACV;AAGA,WAAO,WAAW,IAAI,CAAC,YAAY;AAAA,MACjC,OAAO;AAAA,MACP,OAAO,OAAO;AAAA,MACd,iBAAiB,KAAK,wBAAwB,QAAQ,OAAO,QAAQ;AAAA,MACrE,iBAAiB,SAAS;AAAA,MAC1B,eAAe,KAAK,sBAAsB,QAAQ,KAAK;AAAA,IACzD,EAAE;AAAA,EACJ;AAAA,EAEA,MAAc,kBAAkB,MAAiC;AAG/D,UAAM,OAAO,KAAK,WAAW,IAAI;AACjC,WAAO,MAAM;AAAA,MACX,EAAE,QAAQ,IAAI;AAAA,MACd,CAAC,GAAG,OAAQ,OAAO,KAAK,MAAO,MAAM;AAAA,IACvC;AAAA,EACF;AAAA,EAEQ,WAAW,KAAqB;AACtC,QAAI,OAAO;AACX,aAAS,IAAI,GAAG,IAAI,IAAI,QAAQ,KAAK;AACnC,YAAM,OAAO,IAAI,WAAW,CAAC;AAC7B,cAAQ,QAAQ,KAAK,OAAO;AAC5B,aAAO,OAAO;AAAA,IAChB;AACA,WAAO,KAAK,IAAI,IAAI;AAAA,EACtB;AAAA,EAEQ,wBACN,OACA,OACA,UACQ;AACR,UAAM,UAAoB,CAAC;AAG3B,QAAI,MAAM,KAAK,YAAY,EAAE,SAAS,MAAM,KAAK,YAAY,CAAC,GAAG;AAC/D,cAAQ,KAAK,0BAA0B;AAAA,IACzC;AAEA,QAAI,MAAM,aAAa,YAAY,EAAE,SAAS,MAAM,KAAK,YAAY,CAAC,GAAG;AACvE,cAAQ,KAAK,8BAA8B;AAAA,IAC7C;AAGA,eAAW,WAAW,SAAS,UAAU;AACvC,UACE,MAAM,aAAa,YAAY,EAAE,SAAS,QAAQ,YAAY,CAAC,KAC/D,MAAM,KAAK,YAAY,EAAE,SAAS,QAAQ,YAAY,CAAC,GACvD;AACA,gBAAQ,KAAK,cAAc,OAAO,EAAE;AAAA,MACtC;AAAA,IACF;AAGA,QAAI,SAAS,WAAW,WAAW,MAAM,KAAK,SAAS,OAAO,GAAG;AAC/D,cAAQ,KAAK,6BAA6B;AAAA,IAC5C;AAEA,WAAO,QAAQ,SAAS,IACpB,QAAQ,KAAK,IAAI,IACjB;AAAA,EACN;AAAA,EAEQ,sBAAsB,OAAc,OAA+B;AACzE,UAAM,UAAoB,CAAC;AAC3B,UAAM,aAAa,MAAM,KAAK,YAAY;AAE1C,QAAI,MAAM,KAAK,YAAY,EAAE,SAAS,UAAU,GAAG;AACjD,cAAQ,KAAK,MAAM;AAAA,IACrB;AAEA,QAAI,MAAM,aAAa,YAAY,EAAE,SAAS,UAAU,GAAG;AACzD,cAAQ,KAAK,aAAa;AAAA,IAC5B;AAEA,QAAI,MAAM,KAAK,YAAY,EAAE,SAAS,UAAU,GAAG;AACjD,cAAQ,KAAK,MAAM;AAAA,IACrB;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAc,cACZ,UACA,OACA,UAC6B;AAE7B,QAAI,WAAW;AAGf,QAAI,MAAM,WAAW;AACnB,iBAAW,SAAS,OAAO,CAAC,QAAQ;AAClC,cAAM,YAAY,IAAI,KAAK,IAAI,MAAM,UAAU;AAC/C,cAAM,QAAQ,MAAM,WAAW;AAC/B,cAAM,MAAM,MAAM,WAAW;AAE7B,gBAAQ,CAAC,SAAS,aAAa,WAAW,CAAC,OAAO,aAAa;AAAA,MACjE,CAAC;AAAA,IACH;AAGA,QAAI,MAAM,YAAY;AACpB,iBAAW,SAAS;AAAA,QAAO,CAAC,QAC1B,MAAM,WAAY,SAAS,IAAI,MAAM,IAAI;AAAA,MAC3C;AAAA,IACF;AAGA,QAAI,MAAM,gBAAgB;AACxB,iBAAW,SAAS,OAAO,CAAC,QAAQ,IAAI,SAAS,MAAM,cAAe;AAAA,IACxE;AAGA,UAAM,SAAS,SAAS,IAAI,CAAC,SAAS;AAAA,MACpC,GAAG;AAAA,MACH,OAAO,KAAK,uBAAuB,KAAK,OAAO,QAAQ;AAAA,IACzD,EAAE;AAGF,WAAO,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAGvC,UAAM,aAAa,MAAM,cAAc;AACvC,WAAO,OAAO,MAAM,GAAG,UAAU;AAAA,EACnC;AAAA,EAEQ,uBACN,SACA,OACA,UACQ;AACR,QAAI,QAAQ,QAAQ;AAGpB,UAAM,YAAY,KAAK,IAAI,IAAI,QAAQ,MAAM,eAAe,MAAO,KAAK;AACxE,QAAI,WAAW,IAAI;AACjB,eAAS;AAAA,IACX,WAAW,WAAW,KAAK;AAEzB,eAAS;AAAA,IACX;AAGA,QAAI,QAAQ,MAAM,WAAW;AAC3B,eAAS;AAAA,IACX;AAGA,QAAI,SAAS,WAAW,WAAW,QAAQ,MAAM,KAAK,SAAS,OAAO,GAAG;AACvE,eAAS;AAAA,IACX;AAGA,QAAI,QAAQ,cAAc,SAAS,MAAM,GAAG;AAC1C,eAAS;AAAA,IACX;AAEA,QAAI,QAAQ,cAAc,SAAS,GAAG;AACpC,eAAS;AAAA,IACX;AAGA,QAAI,SAAS,WAAW,qBAAqB,WAAW,KAAK;AAC3D,eAAS;AAAA,IACX;AAEA,WAAO;AAAA,EACT;AAAA,EAEQ,iBAAiB,OAA6B;AACpD,WAAO,KAAK,UAAU;AAAA,MACpB,MAAM,MAAM;AAAA,MACZ,MAAM,MAAM;AAAA,MACZ,YAAY,MAAM;AAAA,MAClB,YAAY,MAAM;AAAA,MAClB,gBAAgB,MAAM;AAAA,IACxB,CAAC;AAAA,EACH;AAAA,EAEQ,gBAAgB,UAAiD;AACvE,UAAM,QAAQ,KAAK,WAAW,IAAI,QAAQ;AAC1C,QAAI,CAAC,MAAO,QAAO;AAGnB,WAAO;AAAA,EACT;AAAA,EAEQ,YAAY,UAAkB,QAAsC;AAE1E,QAAI,KAAK,WAAW,QAAQ,KAAK,cAAc;AAC7C,YAAM,WAAW,KAAK,WAAW,KAAK,EAAE,KAAK,EAAE;AAC/C,WAAK,WAAW,OAAO,QAAQ;AAAA,IACjC;AAEA,SAAK,WAAW,IAAI,UAAU,MAAM;AAAA,EACtC;AAAA;AAAA,EAGA,MAAM,kBACJ,SACA,QAAQ,IACqB;AAC7B,UAAM,QAAQ,MAAM,KAAK,QAAQ,SAAS,OAAO;AACjD,QAAI,CAAC,OAAO;AACV,YAAM,IAAI,MAAM,oBAAoB,OAAO,EAAE;AAAA,IAC/C;AAEA,UAAM,QAAsB;AAAA,MAC1B,MAAM,MAAM,eAAe,MAAM;AAAA,MACjC,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,gBAAgB;AAAA,IAClB;AAEA,UAAM,SAAS,MAAM,KAAK,gBAAgB,KAAK;AAG/C,WAAO,OAAO,SAAS,OAAO,CAAC,QAAQ,IAAI,MAAM,aAAa,OAAO;AAAA,EACvE;AAAA,EAEA,MAAM,oBACJ,cACA,YAC6B;AAC7B,UAAM,QAAsB;AAAA,MAC1B,MAAM,GAAG,YAAY,IAAI,cAAc,EAAE,GAAG,KAAK;AAAA,MACjD,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,YAAY,CAAC,SAAS,SAAS,UAAU;AAAA,MACzC,gBAAgB;AAAA,IAClB;AAEA,UAAM,SAAS,MAAM,KAAK,gBAAgB,KAAK;AAC/C,WAAO,OAAO;AAAA,EAChB;AAAA,EAEA,MAAM,iBACJ,QAAQ,IACR,YAC6B;AAC7B,UAAM,QAAsB;AAAA,MAC1B,MAAM;AAAA,MACN,MAAM;AAAA,MACN,YAAY;AAAA,MACZ,WAAW;AAAA,QACT,OAAO,IAAI,KAAK,KAAK,IAAI,IAAI,QAAQ,KAAK,KAAK,GAAI;AAAA,MACrD;AAAA,MACA;AAAA,MACA,gBAAgB;AAAA,IAClB;AAEA,UAAM,SAAS,MAAM,KAAK,gBAAgB,KAAK;AAC/C,WAAO,OAAO;AAAA,EAChB;AAAA;AAAA,EAGA,oBAAoB;AAClB,WAAO;AAAA,MACL,WAAW,KAAK,WAAW;AAAA,MAC3B,iBAAiB,KAAK,WAAW;AAAA,MACjC,qBAAqB,MAAM,KAAK,KAAK,WAAW,KAAK,CAAC;AAAA,IACxD;AAAA,EACF;AAAA,EAEA,aAAmB;AACjB,SAAK,WAAW,MAAM;AACtB,WAAO,KAAK,iCAAiC;AAAA,EAC/C;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|