@stackmemoryai/stackmemory 0.3.5 → 0.3.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agents/verifiers/base-verifier.js.map +2 -2
- package/dist/agents/verifiers/formatter-verifier.js.map +2 -2
- package/dist/agents/verifiers/llm-judge.js.map +2 -2
- package/dist/cli/claude-sm.js +24 -13
- package/dist/cli/claude-sm.js.map +2 -2
- package/dist/cli/codex-sm.js +24 -13
- package/dist/cli/codex-sm.js.map +2 -2
- package/dist/cli/commands/agent.js.map +2 -2
- package/dist/cli/commands/chromadb.js +217 -32
- package/dist/cli/commands/chromadb.js.map +2 -2
- package/dist/cli/commands/clear.js +12 -1
- package/dist/cli/commands/clear.js.map +2 -2
- package/dist/cli/commands/context.js +13 -2
- package/dist/cli/commands/context.js.map +2 -2
- package/dist/cli/commands/dashboard.js.map +2 -2
- package/dist/cli/commands/gc.js +202 -0
- package/dist/cli/commands/gc.js.map +7 -0
- package/dist/cli/commands/handoff.js +12 -1
- package/dist/cli/commands/handoff.js.map +2 -2
- package/dist/cli/commands/infinite-storage.js +251 -0
- package/dist/cli/commands/infinite-storage.js.map +7 -0
- package/dist/cli/commands/linear-create.js +13 -2
- package/dist/cli/commands/linear-create.js.map +2 -2
- package/dist/cli/commands/linear-list.js +12 -1
- package/dist/cli/commands/linear-list.js.map +2 -2
- package/dist/cli/commands/linear-migrate.js +12 -1
- package/dist/cli/commands/linear-migrate.js.map +2 -2
- package/dist/cli/commands/linear-test.js +12 -1
- package/dist/cli/commands/linear-test.js.map +2 -2
- package/dist/cli/commands/linear-unified.js +262 -0
- package/dist/cli/commands/linear-unified.js.map +7 -0
- package/dist/cli/commands/linear.js +17 -6
- package/dist/cli/commands/linear.js.map +2 -2
- package/dist/cli/commands/monitor.js.map +2 -2
- package/dist/cli/commands/onboard.js.map +2 -2
- package/dist/cli/commands/quality.js.map +2 -2
- package/dist/cli/commands/search.js.map +2 -2
- package/dist/cli/commands/session.js.map +2 -2
- package/dist/cli/commands/skills.js +12 -1
- package/dist/cli/commands/skills.js.map +2 -2
- package/dist/cli/commands/storage.js +18 -7
- package/dist/cli/commands/storage.js.map +2 -2
- package/dist/cli/commands/tasks.js.map +2 -2
- package/dist/cli/commands/tui.js +13 -2
- package/dist/cli/commands/tui.js.map +2 -2
- package/dist/cli/commands/webhook.js +14 -3
- package/dist/cli/commands/webhook.js.map +2 -2
- package/dist/cli/commands/workflow.js +14 -3
- package/dist/cli/commands/workflow.js.map +2 -2
- package/dist/cli/commands/worktree.js.map +2 -2
- package/dist/cli/index.js +20 -5
- package/dist/cli/index.js.map +2 -2
- package/dist/core/config/config-manager.js.map +2 -2
- package/dist/core/context/auto-context.js.map +2 -2
- package/dist/core/context/compaction-handler.js.map +2 -2
- package/dist/core/context/context-bridge.js.map +2 -2
- package/dist/core/context/dual-stack-manager.js.map +2 -2
- package/dist/core/context/frame-database.js.map +2 -2
- package/dist/core/context/frame-digest.js.map +2 -2
- package/dist/core/context/frame-handoff-manager.js.map +2 -2
- package/dist/core/context/frame-manager.js +12 -1
- package/dist/core/context/frame-manager.js.map +2 -2
- package/dist/core/context/frame-stack.js.map +2 -2
- package/dist/core/context/incremental-gc.js +279 -0
- package/dist/core/context/incremental-gc.js.map +7 -0
- package/dist/core/context/permission-manager.js +12 -1
- package/dist/core/context/permission-manager.js.map +2 -2
- package/dist/core/context/refactored-frame-manager.js.map +2 -2
- package/dist/core/context/shared-context-layer.js +12 -1
- package/dist/core/context/shared-context-layer.js.map +2 -2
- package/dist/core/context/stack-merge-resolver.js.map +2 -2
- package/dist/core/context/validation.js.map +2 -2
- package/dist/core/database/batch-operations.js.map +2 -2
- package/dist/core/database/connection-pool.js.map +2 -2
- package/dist/core/database/migration-manager.js.map +2 -2
- package/dist/core/database/paradedb-adapter.js.map +2 -2
- package/dist/core/database/query-cache.js.map +2 -2
- package/dist/core/database/query-router.js.map +2 -2
- package/dist/core/database/sqlite-adapter.js.map +2 -2
- package/dist/core/digest/enhanced-hybrid-digest.js.map +2 -2
- package/dist/core/errors/recovery.js.map +2 -2
- package/dist/core/merge/resolution-engine.js.map +2 -2
- package/dist/core/monitoring/error-handler.js.map +2 -2
- package/dist/core/monitoring/logger.js +14 -3
- package/dist/core/monitoring/logger.js.map +2 -2
- package/dist/core/monitoring/metrics.js +13 -2
- package/dist/core/monitoring/metrics.js.map +2 -2
- package/dist/core/monitoring/progress-tracker.js +12 -1
- package/dist/core/monitoring/progress-tracker.js.map +2 -2
- package/dist/core/monitoring/session-monitor.js.map +2 -2
- package/dist/core/performance/context-cache.js.map +2 -2
- package/dist/core/performance/lazy-context-loader.js.map +2 -2
- package/dist/core/performance/monitor.js.map +2 -2
- package/dist/core/performance/optimized-frame-context.js.map +2 -2
- package/dist/core/performance/performance-benchmark.js.map +2 -2
- package/dist/core/performance/performance-profiler.js +12 -1
- package/dist/core/performance/performance-profiler.js.map +2 -2
- package/dist/core/performance/streaming-jsonl-parser.js.map +2 -2
- package/dist/core/persistence/postgres-adapter.js.map +2 -2
- package/dist/core/projects/project-manager.js.map +2 -2
- package/dist/core/retrieval/context-retriever.js.map +2 -2
- package/dist/core/retrieval/graph-retrieval.js.map +2 -2
- package/dist/core/retrieval/llm-context-retrieval.js.map +2 -2
- package/dist/core/retrieval/retrieval-benchmarks.js.map +2 -2
- package/dist/core/retrieval/summary-generator.js.map +2 -2
- package/dist/core/session/clear-survival.js.map +2 -2
- package/dist/core/session/handoff-generator.js.map +2 -2
- package/dist/core/session/session-manager.js +16 -5
- package/dist/core/session/session-manager.js.map +2 -2
- package/dist/core/skills/skill-storage.js +13 -2
- package/dist/core/skills/skill-storage.js.map +2 -2
- package/dist/core/storage/chromadb-adapter.js.map +2 -2
- package/dist/core/storage/chromadb-simple.js +160 -0
- package/dist/core/storage/chromadb-simple.js.map +7 -0
- package/dist/core/storage/infinite-storage.js +443 -0
- package/dist/core/storage/infinite-storage.js.map +7 -0
- package/dist/core/storage/railway-optimized-storage.js +19 -8
- package/dist/core/storage/railway-optimized-storage.js.map +2 -2
- package/dist/core/storage/remote-storage.js +12 -1
- package/dist/core/storage/remote-storage.js.map +2 -2
- package/dist/core/trace/cli-trace-wrapper.js +16 -5
- package/dist/core/trace/cli-trace-wrapper.js.map +2 -2
- package/dist/core/trace/db-trace-wrapper.js.map +2 -2
- package/dist/core/trace/debug-trace.js +21 -10
- package/dist/core/trace/debug-trace.js.map +2 -2
- package/dist/core/trace/index.js +46 -35
- package/dist/core/trace/index.js.map +2 -2
- package/dist/core/trace/trace-demo.js +12 -1
- package/dist/core/trace/trace-demo.js.map +2 -2
- package/dist/core/trace/trace-detector.js.map +2 -2
- package/dist/core/trace/trace-store.js.map +2 -2
- package/dist/core/utils/compression.js +79 -0
- package/dist/core/utils/compression.js.map +7 -0
- package/dist/core/utils/update-checker.js.map +2 -2
- package/dist/core/worktree/worktree-manager.js.map +2 -2
- package/dist/features/analytics/api/analytics-api.js.map +2 -2
- package/dist/features/analytics/core/analytics-service.js +12 -1
- package/dist/features/analytics/core/analytics-service.js.map +2 -2
- package/dist/features/analytics/queries/metrics-queries.js.map +2 -2
- package/dist/features/tasks/pebbles-task-store.js.map +2 -2
- package/dist/features/tui/components/analytics-panel.js.map +2 -2
- package/dist/features/tui/components/pr-tracker.js.map +2 -2
- package/dist/features/tui/components/session-monitor.js.map +2 -2
- package/dist/features/tui/components/subagent-fleet.js.map +2 -2
- package/dist/features/tui/components/task-board.js +650 -2
- package/dist/features/tui/components/task-board.js.map +2 -2
- package/dist/features/tui/index.js +16 -5
- package/dist/features/tui/index.js.map +2 -2
- package/dist/features/tui/services/data-service.js +35 -52
- package/dist/features/tui/services/data-service.js.map +2 -2
- package/dist/features/tui/services/linear-task-reader.js +100 -0
- package/dist/features/tui/services/linear-task-reader.js.map +7 -0
- package/dist/features/tui/services/websocket-client.js +13 -2
- package/dist/features/tui/services/websocket-client.js.map +2 -2
- package/dist/features/tui/terminal-compat.js +27 -16
- package/dist/features/tui/terminal-compat.js.map +2 -2
- package/dist/features/web/client/stores/task-store.js +22 -0
- package/dist/features/web/client/stores/task-store.js.map +7 -0
- package/dist/features/web/server/index.js +182 -0
- package/dist/features/web/server/index.js.map +7 -0
- package/dist/integrations/claude-code/enhanced-pre-clear-hooks.js.map +2 -2
- package/dist/integrations/claude-code/lifecycle-hooks.js.map +2 -2
- package/dist/integrations/claude-code/post-task-hooks.js.map +2 -2
- package/dist/integrations/linear/auth.js +17 -6
- package/dist/integrations/linear/auth.js.map +2 -2
- package/dist/integrations/linear/auto-sync.js.map +2 -2
- package/dist/integrations/linear/client.js.map +2 -2
- package/dist/integrations/linear/config.js.map +2 -2
- package/dist/integrations/linear/migration.js.map +2 -2
- package/dist/integrations/linear/oauth-server.js +13 -2
- package/dist/integrations/linear/oauth-server.js.map +2 -2
- package/dist/integrations/linear/rest-client.js.map +2 -2
- package/dist/integrations/linear/sync-enhanced.js +202 -0
- package/dist/integrations/linear/sync-enhanced.js.map +7 -0
- package/dist/integrations/linear/sync-manager.js.map +2 -2
- package/dist/integrations/linear/sync-service.js +12 -1
- package/dist/integrations/linear/sync-service.js.map +2 -2
- package/dist/integrations/linear/sync.js +34 -3
- package/dist/integrations/linear/sync.js.map +2 -2
- package/dist/integrations/linear/unified-sync.js +560 -0
- package/dist/integrations/linear/unified-sync.js.map +7 -0
- package/dist/integrations/linear/webhook-handler.js +12 -1
- package/dist/integrations/linear/webhook-handler.js.map +2 -2
- package/dist/integrations/linear/webhook-server.js +14 -3
- package/dist/integrations/linear/webhook-server.js.map +2 -2
- package/dist/integrations/linear/webhook.js +12 -1
- package/dist/integrations/linear/webhook.js.map +2 -2
- package/dist/integrations/mcp/handlers/context-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/linear-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/skill-handlers.js +13 -2
- package/dist/integrations/mcp/handlers/skill-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/task-handlers.js.map +2 -2
- package/dist/integrations/mcp/handlers/trace-handlers.js.map +2 -2
- package/dist/integrations/mcp/middleware/tool-scoring.js.map +2 -2
- package/dist/integrations/mcp/refactored-server.js +15 -4
- package/dist/integrations/mcp/refactored-server.js.map +2 -2
- package/dist/integrations/mcp/server.js +12 -1
- package/dist/integrations/mcp/server.js.map +2 -2
- package/dist/integrations/mcp/tool-definitions.js.map +2 -2
- package/dist/integrations/pg-aiguide/embedding-provider.js +13 -2
- package/dist/integrations/pg-aiguide/embedding-provider.js.map +2 -2
- package/dist/integrations/pg-aiguide/semantic-search.js.map +2 -2
- package/dist/mcp/stackmemory-mcp-server.js +12 -1
- package/dist/mcp/stackmemory-mcp-server.js.map +2 -2
- package/dist/middleware/exponential-rate-limiter.js.map +2 -2
- package/dist/servers/production/auth-middleware.js +13 -2
- package/dist/servers/production/auth-middleware.js.map +2 -2
- package/dist/servers/railway/index.js +22 -11
- package/dist/servers/railway/index.js.map +2 -2
- package/dist/services/config-service.js.map +2 -2
- package/dist/services/context-service.js.map +2 -2
- package/dist/skills/claude-skills.js +150 -1
- package/dist/skills/claude-skills.js.map +2 -2
- package/dist/skills/dashboard-launcher.js +212 -0
- package/dist/skills/dashboard-launcher.js.map +7 -0
- package/dist/skills/repo-ingestion-skill.js +561 -0
- package/dist/skills/repo-ingestion-skill.js.map +7 -0
- package/dist/utils/logger.js +12 -1
- package/dist/utils/logger.js.map +2 -2
- package/package.json +7 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/trace/trace-store.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Trace Store - Database persistence for traces\n */\n\nimport Database from 'better-sqlite3';\nimport {\n Trace,\n ToolCall,\n TraceType,\n TraceMetadata,\n CompressedTrace,\n CompressionStrategy,\n} from './types.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport class TraceStore {\n private db: Database.Database;\n\n constructor(db: Database.Database) {\n this.db = db;\n this.initializeSchema();\n }\n\n /**\n * Initialize database schema for traces\n */\n private initializeSchema(): void {\n // Check if frames table exists (it may not in all contexts)\n const hasFramesTable = this.db\n .prepare(\n `\n SELECT name FROM sqlite_master \n WHERE type='table' AND name='frames'\n `\n )\n .get();\n\n // Create traces table with optional foreign key\n if (hasFramesTable) {\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS traces (\n id TEXT PRIMARY KEY,\n type TEXT NOT NULL,\n score REAL NOT NULL,\n summary TEXT NOT NULL,\n start_time INTEGER NOT NULL,\n end_time INTEGER NOT NULL,\n frame_id TEXT,\n user_id TEXT,\n files_modified TEXT,\n errors_encountered TEXT,\n decisions_recorded TEXT,\n causal_chain INTEGER,\n compressed_data TEXT,\n created_at INTEGER DEFAULT (unixepoch()),\n FOREIGN KEY (frame_id) REFERENCES frames(frame_id) ON DELETE SET NULL\n )\n `);\n } else {\n // Create without foreign key constraint\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS traces (\n id TEXT PRIMARY KEY,\n type TEXT NOT NULL,\n score REAL NOT NULL,\n summary TEXT NOT NULL,\n start_time INTEGER NOT NULL,\n end_time INTEGER NOT NULL,\n frame_id TEXT,\n user_id TEXT,\n files_modified TEXT,\n errors_encountered TEXT,\n decisions_recorded TEXT,\n causal_chain INTEGER,\n compressed_data TEXT,\n created_at INTEGER DEFAULT (unixepoch())\n )\n `);\n }\n\n // Create tool_calls table\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS tool_calls (\n id TEXT PRIMARY KEY,\n trace_id TEXT NOT NULL,\n tool TEXT NOT NULL,\n arguments TEXT,\n timestamp INTEGER NOT NULL,\n result TEXT,\n error TEXT,\n files_affected TEXT,\n duration INTEGER,\n sequence_number INTEGER NOT NULL,\n FOREIGN KEY (trace_id) REFERENCES traces(id) ON DELETE CASCADE\n )\n `);\n\n // Create indexes\n this.db.exec(`\n CREATE INDEX IF NOT EXISTS idx_traces_type ON traces(type);\n CREATE INDEX IF NOT EXISTS idx_traces_frame_id ON traces(frame_id);\n CREATE INDEX IF NOT EXISTS idx_traces_start_time ON traces(start_time);\n CREATE INDEX IF NOT EXISTS idx_traces_score ON traces(score);\n CREATE INDEX IF NOT EXISTS idx_tool_calls_trace_id ON tool_calls(trace_id);\n CREATE INDEX IF NOT EXISTS idx_tool_calls_timestamp ON tool_calls(timestamp);\n `);\n }\n\n /**\n * Save a trace to the database\n */\n saveTrace(trace: Trace): void {\n const traceStmt = this.db.prepare(`\n INSERT OR REPLACE INTO traces (\n id, type, score, summary, start_time, end_time,\n frame_id, user_id, files_modified, errors_encountered,\n decisions_recorded, causal_chain, compressed_data\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n const toolCallStmt = this.db.prepare(`\n INSERT OR REPLACE INTO tool_calls (\n id, trace_id, tool, arguments, timestamp, result,\n error, files_affected, duration, sequence_number\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n try {\n this.db.transaction(() => {\n // Save trace\n traceStmt.run(\n trace.id,\n trace.type,\n trace.score,\n trace.summary,\n trace.metadata.startTime,\n trace.metadata.endTime,\n trace.metadata.frameId || null,\n trace.metadata.userId || null,\n JSON.stringify(trace.metadata.filesModified),\n JSON.stringify(trace.metadata.errorsEncountered),\n JSON.stringify(trace.metadata.decisionsRecorded),\n trace.metadata.causalChain ? 1 : 0,\n trace.compressed ? JSON.stringify(trace.compressed) : null\n );\n\n // Save tool calls\n trace.tools.forEach((tool, index) => {\n toolCallStmt.run(\n tool.id,\n trace.id,\n tool.tool,\n tool.arguments ? JSON.stringify(tool.arguments) : null,\n tool.timestamp,\n tool.result ? JSON.stringify(tool.result) : null,\n tool.error || null,\n tool.filesAffected ? JSON.stringify(tool.filesAffected) : null,\n tool.duration || null,\n index\n );\n });\n })();\n\n logger.debug(\n `Saved trace ${trace.id} with ${trace.tools.length} tool calls`\n );\n } catch (error) {\n logger.error(`Failed to save trace ${trace.id}:`, error as Error);\n throw error;\n }\n }\n\n /**\n * Load a trace by ID\n */\n getTrace(id: string): Trace | null {\n const traceRow = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE id = ?\n `\n )\n .get(id) as any;\n\n if (!traceRow) {\n return null;\n }\n\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n }\n\n /**\n * Load all traces\n */\n getAllTraces(): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces ORDER BY start_time DESC\n `\n )\n .all() as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load traces by type\n */\n getTracesByType(type: TraceType): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE type = ? ORDER BY start_time DESC\n `\n )\n .all(type) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load traces by frame\n */\n getTracesByFrame(frameId: string): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE frame_id = ? ORDER BY start_time DESC\n `\n )\n .all(frameId) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load high-importance traces\n */\n getHighImportanceTraces(minScore: number = 0.7): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE score >= ? ORDER BY score DESC, start_time DESC\n `\n )\n .all(minScore) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load error traces\n */\n getErrorTraces(): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces \n WHERE type = ? OR errors_encountered != '[]'\n ORDER BY start_time DESC\n `\n )\n .all(TraceType.ERROR_RECOVERY) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Get trace statistics\n */\n getStatistics(): {\n totalTraces: number;\n tracesByType: Record<string, number>;\n averageScore: number;\n averageLength: number;\n errorRate: number;\n } {\n const stats = this.db\n .prepare(\n `\n SELECT \n COUNT(*) as total,\n AVG(score) as avg_score,\n AVG((\n SELECT COUNT(*) FROM tool_calls WHERE trace_id = traces.id\n )) as avg_length,\n SUM(CASE WHEN type = ? OR errors_encountered != '[]' THEN 1 ELSE 0 END) * 100.0 / COUNT(*) as error_rate\n FROM traces\n `\n )\n .get(TraceType.ERROR_RECOVERY) as any;\n\n const typeStats = this.db\n .prepare(\n `\n SELECT type, COUNT(*) as count\n FROM traces\n GROUP BY type\n `\n )\n .all() as any[];\n\n const tracesByType: Record<string, number> = {};\n typeStats.forEach((row) => {\n tracesByType[row.type] = row.count;\n });\n\n return {\n totalTraces: stats.total || 0,\n tracesByType,\n averageScore: stats.avg_score || 0,\n averageLength: stats.avg_length || 0,\n errorRate: stats.error_rate || 0,\n };\n }\n\n /**\n * Delete old traces\n */\n deleteOldTraces(olderThanMs: number): number {\n const cutoff = Date.now() - olderThanMs;\n const result = this.db\n .prepare(\n `\n DELETE FROM traces WHERE start_time < ?\n `\n )\n .run(cutoff);\n\n return result.changes;\n }\n\n /**\n * Convert database rows to Trace object\n */\n private rowsToTrace(traceRow: any, toolRows: any[]): Trace {\n const tools: ToolCall[] = toolRows.map((row) => ({\n id: row.id,\n tool: row.tool,\n arguments: row.arguments ? JSON.parse(row.arguments) : undefined,\n timestamp: row.timestamp,\n result: row.result ? JSON.parse(row.result) : undefined,\n error: row.error || undefined,\n filesAffected: row.files_affected\n ? JSON.parse(row.files_affected)\n : undefined,\n duration: row.duration || undefined,\n }));\n\n const metadata: TraceMetadata = {\n startTime: traceRow.start_time,\n endTime: traceRow.end_time,\n frameId: traceRow.frame_id || undefined,\n userId: traceRow.user_id || undefined,\n filesModified: JSON.parse(traceRow.files_modified || '[]'),\n errorsEncountered: JSON.parse(traceRow.errors_encountered || '[]'),\n decisionsRecorded: JSON.parse(traceRow.decisions_recorded || '[]'),\n causalChain: traceRow.causal_chain === 1,\n };\n\n const trace: Trace = {\n id: traceRow.id,\n type: traceRow.type as TraceType,\n tools,\n score: traceRow.score,\n summary: traceRow.summary,\n metadata,\n };\n\n if (traceRow.compressed_data) {\n trace.compressed = JSON.parse(traceRow.compressed_data);\n }\n\n return trace;\n }\n}\n"],
|
|
5
|
-
"mappings": "AAKA;AAAA,EAGE;AAAA,OAIK;AACP,SAAS,cAAc;AAEhB,MAAM,WAAW;AAAA,EACd;AAAA,EAER,YAAY,IAAuB;AACjC,SAAK,KAAK;AACV,SAAK,iBAAiB;AAAA,EACxB;AAAA;AAAA;AAAA;AAAA,EAKQ,mBAAyB;AAE/B,UAAM,iBAAiB,KAAK,GACzB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC,IAAI;AAGP,QAAI,gBAAgB;AAClB,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAkBZ;AAAA,IACH,OAAO;AAEL,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAiBZ;AAAA,IACH;AAGA,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAcZ;AAGD,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAOZ;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,UAAU,OAAoB;AAC5B,UAAM,YAAY,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAMjC;AAED,UAAM,eAAe,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,KAKpC;AAED,QAAI;AACF,WAAK,GAAG,YAAY,MAAM;AAExB,kBAAU;AAAA,UACR,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM,SAAS;AAAA,UACf,MAAM,SAAS;AAAA,UACf,MAAM,SAAS,WAAW;AAAA,UAC1B,MAAM,SAAS,UAAU;AAAA,UACzB,KAAK,UAAU,MAAM,SAAS,aAAa;AAAA,UAC3C,KAAK,UAAU,MAAM,SAAS,iBAAiB;AAAA,UAC/C,KAAK,UAAU,MAAM,SAAS,iBAAiB;AAAA,UAC/C,MAAM,SAAS,cAAc,IAAI;AAAA,UACjC,MAAM,aAAa,KAAK,UAAU,MAAM,UAAU,IAAI;AAAA,QACxD;AAGA,cAAM,MAAM,QAAQ,CAAC,MAAM,UAAU;AACnC,uBAAa;AAAA,YACX,KAAK;AAAA,YACL,MAAM;AAAA,YACN,KAAK;AAAA,YACL,KAAK,YAAY,KAAK,UAAU,KAAK,SAAS,IAAI;AAAA,YAClD,KAAK;AAAA,YACL,KAAK,SAAS,KAAK,UAAU,KAAK,MAAM,IAAI;AAAA,YAC5C,KAAK,SAAS;AAAA,YACd,KAAK,gBAAgB,KAAK,UAAU,KAAK,aAAa,IAAI;AAAA,YAC1D,KAAK,YAAY;AAAA,YACjB;AAAA,UACF;AAAA,QACF,CAAC;AAAA,MACH,CAAC,EAAE;AAEH,aAAO;AAAA,QACL,eAAe,MAAM,EAAE,SAAS,MAAM,MAAM,MAAM;AAAA,MACpD;AAAA,IACF,SAAS,
|
|
4
|
+
"sourcesContent": ["/**\n * Trace Store - Database persistence for traces\n */\n\nimport Database from 'better-sqlite3';\nimport {\n Trace,\n ToolCall,\n TraceType,\n TraceMetadata,\n CompressedTrace,\n CompressionStrategy,\n} from './types.js';\nimport { logger } from '../monitoring/logger.js';\n\nexport class TraceStore {\n private db: Database.Database;\n\n constructor(db: Database.Database) {\n this.db = db;\n this.initializeSchema();\n }\n\n /**\n * Initialize database schema for traces\n */\n private initializeSchema(): void {\n // Check if frames table exists (it may not in all contexts)\n const hasFramesTable = this.db\n .prepare(\n `\n SELECT name FROM sqlite_master \n WHERE type='table' AND name='frames'\n `\n )\n .get();\n\n // Create traces table with optional foreign key\n if (hasFramesTable) {\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS traces (\n id TEXT PRIMARY KEY,\n type TEXT NOT NULL,\n score REAL NOT NULL,\n summary TEXT NOT NULL,\n start_time INTEGER NOT NULL,\n end_time INTEGER NOT NULL,\n frame_id TEXT,\n user_id TEXT,\n files_modified TEXT,\n errors_encountered TEXT,\n decisions_recorded TEXT,\n causal_chain INTEGER,\n compressed_data TEXT,\n created_at INTEGER DEFAULT (unixepoch()),\n FOREIGN KEY (frame_id) REFERENCES frames(frame_id) ON DELETE SET NULL\n )\n `);\n } else {\n // Create without foreign key constraint\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS traces (\n id TEXT PRIMARY KEY,\n type TEXT NOT NULL,\n score REAL NOT NULL,\n summary TEXT NOT NULL,\n start_time INTEGER NOT NULL,\n end_time INTEGER NOT NULL,\n frame_id TEXT,\n user_id TEXT,\n files_modified TEXT,\n errors_encountered TEXT,\n decisions_recorded TEXT,\n causal_chain INTEGER,\n compressed_data TEXT,\n created_at INTEGER DEFAULT (unixepoch())\n )\n `);\n }\n\n // Create tool_calls table\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS tool_calls (\n id TEXT PRIMARY KEY,\n trace_id TEXT NOT NULL,\n tool TEXT NOT NULL,\n arguments TEXT,\n timestamp INTEGER NOT NULL,\n result TEXT,\n error TEXT,\n files_affected TEXT,\n duration INTEGER,\n sequence_number INTEGER NOT NULL,\n FOREIGN KEY (trace_id) REFERENCES traces(id) ON DELETE CASCADE\n )\n `);\n\n // Create indexes\n this.db.exec(`\n CREATE INDEX IF NOT EXISTS idx_traces_type ON traces(type);\n CREATE INDEX IF NOT EXISTS idx_traces_frame_id ON traces(frame_id);\n CREATE INDEX IF NOT EXISTS idx_traces_start_time ON traces(start_time);\n CREATE INDEX IF NOT EXISTS idx_traces_score ON traces(score);\n CREATE INDEX IF NOT EXISTS idx_tool_calls_trace_id ON tool_calls(trace_id);\n CREATE INDEX IF NOT EXISTS idx_tool_calls_timestamp ON tool_calls(timestamp);\n `);\n }\n\n /**\n * Save a trace to the database\n */\n saveTrace(trace: Trace): void {\n const traceStmt = this.db.prepare(`\n INSERT OR REPLACE INTO traces (\n id, type, score, summary, start_time, end_time,\n frame_id, user_id, files_modified, errors_encountered,\n decisions_recorded, causal_chain, compressed_data\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n const toolCallStmt = this.db.prepare(`\n INSERT OR REPLACE INTO tool_calls (\n id, trace_id, tool, arguments, timestamp, result,\n error, files_affected, duration, sequence_number\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n try {\n this.db.transaction(() => {\n // Save trace\n traceStmt.run(\n trace.id,\n trace.type,\n trace.score,\n trace.summary,\n trace.metadata.startTime,\n trace.metadata.endTime,\n trace.metadata.frameId || null,\n trace.metadata.userId || null,\n JSON.stringify(trace.metadata.filesModified),\n JSON.stringify(trace.metadata.errorsEncountered),\n JSON.stringify(trace.metadata.decisionsRecorded),\n trace.metadata.causalChain ? 1 : 0,\n trace.compressed ? JSON.stringify(trace.compressed) : null\n );\n\n // Save tool calls\n trace.tools.forEach((tool, index) => {\n toolCallStmt.run(\n tool.id,\n trace.id,\n tool.tool,\n tool.arguments ? JSON.stringify(tool.arguments) : null,\n tool.timestamp,\n tool.result ? JSON.stringify(tool.result) : null,\n tool.error || null,\n tool.filesAffected ? JSON.stringify(tool.filesAffected) : null,\n tool.duration || null,\n index\n );\n });\n })();\n\n logger.debug(\n `Saved trace ${trace.id} with ${trace.tools.length} tool calls`\n );\n } catch (error: unknown) {\n logger.error(`Failed to save trace ${trace.id}:`, error as Error);\n throw error;\n }\n }\n\n /**\n * Load a trace by ID\n */\n getTrace(id: string): Trace | null {\n const traceRow = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE id = ?\n `\n )\n .get(id) as any;\n\n if (!traceRow) {\n return null;\n }\n\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n }\n\n /**\n * Load all traces\n */\n getAllTraces(): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces ORDER BY start_time DESC\n `\n )\n .all() as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load traces by type\n */\n getTracesByType(type: TraceType): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE type = ? ORDER BY start_time DESC\n `\n )\n .all(type) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load traces by frame\n */\n getTracesByFrame(frameId: string): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE frame_id = ? ORDER BY start_time DESC\n `\n )\n .all(frameId) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load high-importance traces\n */\n getHighImportanceTraces(minScore: number = 0.7): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces WHERE score >= ? ORDER BY score DESC, start_time DESC\n `\n )\n .all(minScore) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Load error traces\n */\n getErrorTraces(): Trace[] {\n const traceRows = this.db\n .prepare(\n `\n SELECT * FROM traces \n WHERE type = ? OR errors_encountered != '[]'\n ORDER BY start_time DESC\n `\n )\n .all(TraceType.ERROR_RECOVERY) as any[];\n\n return traceRows.map((traceRow) => {\n const toolRows = this.db\n .prepare(\n `\n SELECT * FROM tool_calls WHERE trace_id = ? ORDER BY sequence_number\n `\n )\n .all(traceRow.id) as any[];\n\n return this.rowsToTrace(traceRow, toolRows);\n });\n }\n\n /**\n * Get trace statistics\n */\n getStatistics(): {\n totalTraces: number;\n tracesByType: Record<string, number>;\n averageScore: number;\n averageLength: number;\n errorRate: number;\n } {\n const stats = this.db\n .prepare(\n `\n SELECT \n COUNT(*) as total,\n AVG(score) as avg_score,\n AVG((\n SELECT COUNT(*) FROM tool_calls WHERE trace_id = traces.id\n )) as avg_length,\n SUM(CASE WHEN type = ? OR errors_encountered != '[]' THEN 1 ELSE 0 END) * 100.0 / COUNT(*) as error_rate\n FROM traces\n `\n )\n .get(TraceType.ERROR_RECOVERY) as any;\n\n const typeStats = this.db\n .prepare(\n `\n SELECT type, COUNT(*) as count\n FROM traces\n GROUP BY type\n `\n )\n .all() as any[];\n\n const tracesByType: Record<string, number> = {};\n typeStats.forEach((row) => {\n tracesByType[row.type] = row.count;\n });\n\n return {\n totalTraces: stats.total || 0,\n tracesByType,\n averageScore: stats.avg_score || 0,\n averageLength: stats.avg_length || 0,\n errorRate: stats.error_rate || 0,\n };\n }\n\n /**\n * Delete old traces\n */\n deleteOldTraces(olderThanMs: number): number {\n const cutoff = Date.now() - olderThanMs;\n const result = this.db\n .prepare(\n `\n DELETE FROM traces WHERE start_time < ?\n `\n )\n .run(cutoff);\n\n return result.changes;\n }\n\n /**\n * Convert database rows to Trace object\n */\n private rowsToTrace(traceRow: any, toolRows: any[]): Trace {\n const tools: ToolCall[] = toolRows.map((row) => ({\n id: row.id,\n tool: row.tool,\n arguments: row.arguments ? JSON.parse(row.arguments) : undefined,\n timestamp: row.timestamp,\n result: row.result ? JSON.parse(row.result) : undefined,\n error: row.error || undefined,\n filesAffected: row.files_affected\n ? JSON.parse(row.files_affected)\n : undefined,\n duration: row.duration || undefined,\n }));\n\n const metadata: TraceMetadata = {\n startTime: traceRow.start_time,\n endTime: traceRow.end_time,\n frameId: traceRow.frame_id || undefined,\n userId: traceRow.user_id || undefined,\n filesModified: JSON.parse(traceRow.files_modified || '[]'),\n errorsEncountered: JSON.parse(traceRow.errors_encountered || '[]'),\n decisionsRecorded: JSON.parse(traceRow.decisions_recorded || '[]'),\n causalChain: traceRow.causal_chain === 1,\n };\n\n const trace: Trace = {\n id: traceRow.id,\n type: traceRow.type as TraceType,\n tools,\n score: traceRow.score,\n summary: traceRow.summary,\n metadata,\n };\n\n if (traceRow.compressed_data) {\n trace.compressed = JSON.parse(traceRow.compressed_data);\n }\n\n return trace;\n }\n}\n"],
|
|
5
|
+
"mappings": "AAKA;AAAA,EAGE;AAAA,OAIK;AACP,SAAS,cAAc;AAEhB,MAAM,WAAW;AAAA,EACd;AAAA,EAER,YAAY,IAAuB;AACjC,SAAK,KAAK;AACV,SAAK,iBAAiB;AAAA,EACxB;AAAA;AAAA;AAAA;AAAA,EAKQ,mBAAyB;AAE/B,UAAM,iBAAiB,KAAK,GACzB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC,IAAI;AAGP,QAAI,gBAAgB;AAClB,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAkBZ;AAAA,IACH,OAAO;AAEL,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAiBZ;AAAA,IACH;AAGA,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAcZ;AAGD,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAOZ;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,UAAU,OAAoB;AAC5B,UAAM,YAAY,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAMjC;AAED,UAAM,eAAe,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,KAKpC;AAED,QAAI;AACF,WAAK,GAAG,YAAY,MAAM;AAExB,kBAAU;AAAA,UACR,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM,SAAS;AAAA,UACf,MAAM,SAAS;AAAA,UACf,MAAM,SAAS,WAAW;AAAA,UAC1B,MAAM,SAAS,UAAU;AAAA,UACzB,KAAK,UAAU,MAAM,SAAS,aAAa;AAAA,UAC3C,KAAK,UAAU,MAAM,SAAS,iBAAiB;AAAA,UAC/C,KAAK,UAAU,MAAM,SAAS,iBAAiB;AAAA,UAC/C,MAAM,SAAS,cAAc,IAAI;AAAA,UACjC,MAAM,aAAa,KAAK,UAAU,MAAM,UAAU,IAAI;AAAA,QACxD;AAGA,cAAM,MAAM,QAAQ,CAAC,MAAM,UAAU;AACnC,uBAAa;AAAA,YACX,KAAK;AAAA,YACL,MAAM;AAAA,YACN,KAAK;AAAA,YACL,KAAK,YAAY,KAAK,UAAU,KAAK,SAAS,IAAI;AAAA,YAClD,KAAK;AAAA,YACL,KAAK,SAAS,KAAK,UAAU,KAAK,MAAM,IAAI;AAAA,YAC5C,KAAK,SAAS;AAAA,YACd,KAAK,gBAAgB,KAAK,UAAU,KAAK,aAAa,IAAI;AAAA,YAC1D,KAAK,YAAY;AAAA,YACjB;AAAA,UACF;AAAA,QACF,CAAC;AAAA,MACH,CAAC,EAAE;AAEH,aAAO;AAAA,QACL,eAAe,MAAM,EAAE,SAAS,MAAM,MAAM,MAAM;AAAA,MACpD;AAAA,IACF,SAAS,OAAgB;AACvB,aAAO,MAAM,wBAAwB,MAAM,EAAE,KAAK,KAAc;AAChE,YAAM;AAAA,IACR;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,SAAS,IAA0B;AACjC,UAAM,WAAW,KAAK,GACnB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,EAAE;AAET,QAAI,CAAC,UAAU;AACb,aAAO;AAAA,IACT;AAEA,UAAM,WAAW,KAAK,GACnB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,EAAE;AAET,WAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,EAC5C;AAAA;AAAA;AAAA;AAAA,EAKA,eAAwB;AACtB,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI;AAEP,WAAO,UAAU,IAAI,CAAC,aAAa;AACjC,YAAM,WAAW,KAAK,GACnB;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,SAAS,EAAE;AAElB,aAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,IAC5C,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,gBAAgB,MAA0B;AACxC,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,IAAI;AAEX,WAAO,UAAU,IAAI,CAAC,aAAa;AACjC,YAAM,WAAW,KAAK,GACnB;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,SAAS,EAAE;AAElB,aAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,IAC5C,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAiB,SAA0B;AACzC,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,OAAO;AAEd,WAAO,UAAU,IAAI,CAAC,aAAa;AACjC,YAAM,WAAW,KAAK,GACnB;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,SAAS,EAAE;AAElB,aAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,IAC5C,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,wBAAwB,WAAmB,KAAc;AACvD,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,QAAQ;AAEf,WAAO,UAAU,IAAI,CAAC,aAAa;AACjC,YAAM,WAAW,KAAK,GACnB;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,SAAS,EAAE;AAElB,aAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,IAC5C,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,iBAA0B;AACxB,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EACC,IAAI,UAAU,cAAc;AAE/B,WAAO,UAAU,IAAI,CAAC,aAAa;AACjC,YAAM,WAAW,KAAK,GACnB;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,SAAS,EAAE;AAElB,aAAO,KAAK,YAAY,UAAU,QAAQ;AAAA,IAC5C,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,gBAME;AACA,UAAM,QAAQ,KAAK,GAChB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAUF,EACC,IAAI,UAAU,cAAc;AAE/B,UAAM,YAAY,KAAK,GACpB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EACC,IAAI;AAEP,UAAM,eAAuC,CAAC;AAC9C,cAAU,QAAQ,CAAC,QAAQ;AACzB,mBAAa,IAAI,IAAI,IAAI,IAAI;AAAA,IAC/B,CAAC;AAED,WAAO;AAAA,MACL,aAAa,MAAM,SAAS;AAAA,MAC5B;AAAA,MACA,cAAc,MAAM,aAAa;AAAA,MACjC,eAAe,MAAM,cAAc;AAAA,MACnC,WAAW,MAAM,cAAc;AAAA,IACjC;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,gBAAgB,aAA6B;AAC3C,UAAM,SAAS,KAAK,IAAI,IAAI;AAC5B,UAAM,SAAS,KAAK,GACjB;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,MAAM;AAEb,WAAO,OAAO;AAAA,EAChB;AAAA;AAAA;AAAA;AAAA,EAKQ,YAAY,UAAe,UAAwB;AACzD,UAAM,QAAoB,SAAS,IAAI,CAAC,SAAS;AAAA,MAC/C,IAAI,IAAI;AAAA,MACR,MAAM,IAAI;AAAA,MACV,WAAW,IAAI,YAAY,KAAK,MAAM,IAAI,SAAS,IAAI;AAAA,MACvD,WAAW,IAAI;AAAA,MACf,QAAQ,IAAI,SAAS,KAAK,MAAM,IAAI,MAAM,IAAI;AAAA,MAC9C,OAAO,IAAI,SAAS;AAAA,MACpB,eAAe,IAAI,iBACf,KAAK,MAAM,IAAI,cAAc,IAC7B;AAAA,MACJ,UAAU,IAAI,YAAY;AAAA,IAC5B,EAAE;AAEF,UAAM,WAA0B;AAAA,MAC9B,WAAW,SAAS;AAAA,MACpB,SAAS,SAAS;AAAA,MAClB,SAAS,SAAS,YAAY;AAAA,MAC9B,QAAQ,SAAS,WAAW;AAAA,MAC5B,eAAe,KAAK,MAAM,SAAS,kBAAkB,IAAI;AAAA,MACzD,mBAAmB,KAAK,MAAM,SAAS,sBAAsB,IAAI;AAAA,MACjE,mBAAmB,KAAK,MAAM,SAAS,sBAAsB,IAAI;AAAA,MACjE,aAAa,SAAS,iBAAiB;AAAA,IACzC;AAEA,UAAM,QAAe;AAAA,MACnB,IAAI,SAAS;AAAA,MACb,MAAM,SAAS;AAAA,MACf;AAAA,MACA,OAAO,SAAS;AAAA,MAChB,SAAS,SAAS;AAAA,MAClB;AAAA,IACF;AAEA,QAAI,SAAS,iBAAiB;AAC5B,YAAM,aAAa,KAAK,MAAM,SAAS,eAAe;AAAA,IACxD;AAEA,WAAO;AAAA,EACT;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -0,0 +1,79 @@
|
|
|
1
|
+
import zlib from "zlib";
|
|
2
|
+
import { promisify } from "util";
|
|
3
|
+
const gzipAsync = promisify(zlib.gzip);
|
|
4
|
+
const gunzipAsync = promisify(zlib.gunzip);
|
|
5
|
+
const brotliCompressAsync = promisify(zlib.brotliCompress);
|
|
6
|
+
const brotliDecompressAsync = promisify(zlib.brotliDecompress);
|
|
7
|
+
var CompressionType = /* @__PURE__ */ ((CompressionType2) => {
|
|
8
|
+
CompressionType2["NONE"] = "none";
|
|
9
|
+
CompressionType2["GZIP"] = "gzip";
|
|
10
|
+
CompressionType2["BROTLI"] = "brotli";
|
|
11
|
+
return CompressionType2;
|
|
12
|
+
})(CompressionType || {});
|
|
13
|
+
async function compress(data, options = {}) {
|
|
14
|
+
const { type = "gzip" /* GZIP */, level = 6 } = options;
|
|
15
|
+
const input = typeof data === "string" ? Buffer.from(data, "utf8") : data;
|
|
16
|
+
switch (type) {
|
|
17
|
+
case "none" /* NONE */:
|
|
18
|
+
return input;
|
|
19
|
+
case "gzip" /* GZIP */:
|
|
20
|
+
return gzipAsync(input, { level });
|
|
21
|
+
case "brotli" /* BROTLI */:
|
|
22
|
+
return brotliCompressAsync(input, {
|
|
23
|
+
params: {
|
|
24
|
+
[zlib.constants.BROTLI_PARAM_QUALITY]: level
|
|
25
|
+
}
|
|
26
|
+
});
|
|
27
|
+
default:
|
|
28
|
+
throw new Error(`Unknown compression type: ${type}`);
|
|
29
|
+
}
|
|
30
|
+
}
|
|
31
|
+
async function decompress(data, type = "gzip" /* GZIP */) {
|
|
32
|
+
let decompressed;
|
|
33
|
+
switch (type) {
|
|
34
|
+
case "none" /* NONE */:
|
|
35
|
+
decompressed = data;
|
|
36
|
+
break;
|
|
37
|
+
case "gzip" /* GZIP */:
|
|
38
|
+
decompressed = await gunzipAsync(data);
|
|
39
|
+
break;
|
|
40
|
+
case "brotli" /* BROTLI */:
|
|
41
|
+
decompressed = await brotliDecompressAsync(data);
|
|
42
|
+
break;
|
|
43
|
+
default:
|
|
44
|
+
throw new Error(`Unknown compression type: ${type}`);
|
|
45
|
+
}
|
|
46
|
+
return decompressed.toString("utf8");
|
|
47
|
+
}
|
|
48
|
+
function compressionRatio(original, compressed) {
|
|
49
|
+
if (original === 0) return 0;
|
|
50
|
+
return (1 - compressed / original) * 100;
|
|
51
|
+
}
|
|
52
|
+
function detectCompressionType(data) {
|
|
53
|
+
if (data.length >= 2 && data[0] === 31 && data[1] === 139) {
|
|
54
|
+
return "gzip" /* GZIP */;
|
|
55
|
+
}
|
|
56
|
+
if (data.length >= 4 && data[0] === 206 && data[1] === 178) {
|
|
57
|
+
return "brotli" /* BROTLI */;
|
|
58
|
+
}
|
|
59
|
+
return "none" /* NONE */;
|
|
60
|
+
}
|
|
61
|
+
function chooseOptimalCompression(data, speedPriority = false) {
|
|
62
|
+
const size = typeof data === "string" ? Buffer.byteLength(data) : data.length;
|
|
63
|
+
if (size < 1024) {
|
|
64
|
+
return "none" /* NONE */;
|
|
65
|
+
}
|
|
66
|
+
if (speedPriority || size < 100 * 1024) {
|
|
67
|
+
return "gzip" /* GZIP */;
|
|
68
|
+
}
|
|
69
|
+
return "brotli" /* BROTLI */;
|
|
70
|
+
}
|
|
71
|
+
export {
|
|
72
|
+
CompressionType,
|
|
73
|
+
chooseOptimalCompression,
|
|
74
|
+
compress,
|
|
75
|
+
compressionRatio,
|
|
76
|
+
decompress,
|
|
77
|
+
detectCompressionType
|
|
78
|
+
};
|
|
79
|
+
//# sourceMappingURL=compression.js.map
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
{
|
|
2
|
+
"version": 3,
|
|
3
|
+
"sources": ["../../../src/core/utils/compression.ts"],
|
|
4
|
+
"sourcesContent": ["/**\n * Compression utilities for storage optimization\n */\n\nimport zlib from 'zlib';\nimport { promisify } from 'util';\n\nconst gzipAsync = promisify(zlib.gzip);\nconst gunzipAsync = promisify(zlib.gunzip);\nconst brotliCompressAsync = promisify(zlib.brotliCompress);\nconst brotliDecompressAsync = promisify(zlib.brotliDecompress);\n\nexport enum CompressionType {\n NONE = 'none',\n GZIP = 'gzip',\n BROTLI = 'brotli',\n}\n\nexport interface CompressionOptions {\n type?: CompressionType;\n level?: number;\n}\n\n/**\n * Compress data using specified algorithm\n */\nexport async function compress(\n data: string | Buffer,\n options: CompressionOptions = {}\n): Promise<Buffer> {\n const { type = CompressionType.GZIP, level = 6 } = options;\n \n const input = typeof data === 'string' ? Buffer.from(data, 'utf8') : data;\n \n switch (type) {\n case CompressionType.NONE:\n return input;\n \n case CompressionType.GZIP:\n return gzipAsync(input, { level });\n \n case CompressionType.BROTLI:\n return brotliCompressAsync(input, {\n params: {\n [zlib.constants.BROTLI_PARAM_QUALITY]: level,\n },\n });\n \n default:\n throw new Error(`Unknown compression type: ${type}`);\n }\n}\n\n/**\n * Decompress data\n */\nexport async function decompress(\n data: Buffer,\n type: CompressionType = CompressionType.GZIP\n): Promise<string> {\n let decompressed: Buffer;\n \n switch (type) {\n case CompressionType.NONE:\n decompressed = data;\n break;\n \n case CompressionType.GZIP:\n decompressed = await gunzipAsync(data);\n break;\n \n case CompressionType.BROTLI:\n decompressed = await brotliDecompressAsync(data);\n break;\n \n default:\n throw new Error(`Unknown compression type: ${type}`);\n }\n \n return decompressed.toString('utf8');\n}\n\n/**\n * Calculate compression ratio\n */\nexport function compressionRatio(original: number, compressed: number): number {\n if (original === 0) return 0;\n return (1 - compressed / original) * 100;\n}\n\n/**\n * Auto-detect compression type from buffer\n */\nexport function detectCompressionType(data: Buffer): CompressionType {\n // Check for gzip magic number\n if (data.length >= 2 && data[0] === 0x1f && data[1] === 0x8b) {\n return CompressionType.GZIP;\n }\n \n // Check for brotli\n // Brotli doesn't have a consistent magic number, but we can try to decompress\n // This is a heuristic approach\n if (data.length >= 4 && data[0] === 0xce && data[1] === 0xb2) {\n return CompressionType.BROTLI;\n }\n \n return CompressionType.NONE;\n}\n\n/**\n * Choose optimal compression based on data characteristics\n */\nexport function chooseOptimalCompression(\n data: string | Buffer,\n speedPriority: boolean = false\n): CompressionType {\n const size = typeof data === 'string' ? Buffer.byteLength(data) : data.length;\n \n // Don't compress small data\n if (size < 1024) {\n return CompressionType.NONE;\n }\n \n // Use gzip for speed priority or medium data\n if (speedPriority || size < 100 * 1024) {\n return CompressionType.GZIP;\n }\n \n // Use brotli for large data and better compression\n return CompressionType.BROTLI;\n}"],
|
|
5
|
+
"mappings": "AAIA,OAAO,UAAU;AACjB,SAAS,iBAAiB;AAE1B,MAAM,YAAY,UAAU,KAAK,IAAI;AACrC,MAAM,cAAc,UAAU,KAAK,MAAM;AACzC,MAAM,sBAAsB,UAAU,KAAK,cAAc;AACzD,MAAM,wBAAwB,UAAU,KAAK,gBAAgB;AAEtD,IAAK,kBAAL,kBAAKA,qBAAL;AACL,EAAAA,iBAAA,UAAO;AACP,EAAAA,iBAAA,UAAO;AACP,EAAAA,iBAAA,YAAS;AAHC,SAAAA;AAAA,GAAA;AAcZ,eAAsB,SACpB,MACA,UAA8B,CAAC,GACd;AACjB,QAAM,EAAE,OAAO,mBAAsB,QAAQ,EAAE,IAAI;AAEnD,QAAM,QAAQ,OAAO,SAAS,WAAW,OAAO,KAAK,MAAM,MAAM,IAAI;AAErE,UAAQ,MAAM;AAAA,IACZ,KAAK;AACH,aAAO;AAAA,IAET,KAAK;AACH,aAAO,UAAU,OAAO,EAAE,MAAM,CAAC;AAAA,IAEnC,KAAK;AACH,aAAO,oBAAoB,OAAO;AAAA,QAChC,QAAQ;AAAA,UACN,CAAC,KAAK,UAAU,oBAAoB,GAAG;AAAA,QACzC;AAAA,MACF,CAAC;AAAA,IAEH;AACE,YAAM,IAAI,MAAM,6BAA6B,IAAI,EAAE;AAAA,EACvD;AACF;AAKA,eAAsB,WACpB,MACA,OAAwB,mBACP;AACjB,MAAI;AAEJ,UAAQ,MAAM;AAAA,IACZ,KAAK;AACH,qBAAe;AACf;AAAA,IAEF,KAAK;AACH,qBAAe,MAAM,YAAY,IAAI;AACrC;AAAA,IAEF,KAAK;AACH,qBAAe,MAAM,sBAAsB,IAAI;AAC/C;AAAA,IAEF;AACE,YAAM,IAAI,MAAM,6BAA6B,IAAI,EAAE;AAAA,EACvD;AAEA,SAAO,aAAa,SAAS,MAAM;AACrC;AAKO,SAAS,iBAAiB,UAAkB,YAA4B;AAC7E,MAAI,aAAa,EAAG,QAAO;AAC3B,UAAQ,IAAI,aAAa,YAAY;AACvC;AAKO,SAAS,sBAAsB,MAA+B;AAEnE,MAAI,KAAK,UAAU,KAAK,KAAK,CAAC,MAAM,MAAQ,KAAK,CAAC,MAAM,KAAM;AAC5D,WAAO;AAAA,EACT;AAKA,MAAI,KAAK,UAAU,KAAK,KAAK,CAAC,MAAM,OAAQ,KAAK,CAAC,MAAM,KAAM;AAC5D,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAKO,SAAS,yBACd,MACA,gBAAyB,OACR;AACjB,QAAM,OAAO,OAAO,SAAS,WAAW,OAAO,WAAW,IAAI,IAAI,KAAK;AAGvE,MAAI,OAAO,MAAM;AACf,WAAO;AAAA,EACT;AAGA,MAAI,iBAAiB,OAAO,MAAM,MAAM;AACtC,WAAO;AAAA,EACT;AAGA,SAAO;AACT;",
|
|
6
|
+
"names": ["CompressionType"]
|
|
7
|
+
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/utils/update-checker.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Update checker for StackMemory\n * Checks npm registry for newer versions\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, readFileSync, writeFileSync, mkdirSync } from 'fs';\nimport { join } from 'path';\nimport { homedir } from 'os';\nimport { logger } from '../monitoring/logger.js';\nimport { \n SystemError, \n ErrorCode, \n getErrorMessage,\n wrapError\n} from '../errors/index.js';\nimport { withTimeout, gracefulDegrade } from '../errors/recovery.js';\n\ninterface UpdateCache {\n lastChecked: number;\n latestVersion: string;\n currentVersion: string;\n}\n\nexport class UpdateChecker {\n private static CACHE_FILE = join(\n homedir(),\n '.stackmemory',\n 'update-check.json'\n );\n private static CHECK_INTERVAL = 24 * 60 * 60 * 1000; // 24 hours\n private static PACKAGE_NAME = '@stackmemoryai/stackmemory';\n\n /**\n * Check for updates and display notification if needed\n */\n static async checkForUpdates(\n currentVersion: string,\n silent = false\n ): Promise<void> {\n try {\n // Check cache first\n const cache = this.loadCache();\n const now = Date.now();\n\n // Skip check if we checked recently\n if (cache && now - cache.lastChecked < this.CHECK_INTERVAL) {\n if (\n !silent &&\n cache.latestVersion &&\n cache.latestVersion !== currentVersion\n ) {\n this.displayUpdateNotification(currentVersion, cache.latestVersion);\n }\n return;\n }\n\n // Fetch latest version from npm\n const latestVersion = await this.fetchLatestVersion();\n\n // Update cache\n this.saveCache({\n lastChecked: now,\n latestVersion,\n currentVersion,\n });\n\n // Display notification if update available\n if (\n !silent &&\n latestVersion &&\n this.isNewerVersion(currentVersion, latestVersion)\n ) {\n this.displayUpdateNotification(currentVersion, latestVersion);\n }\n } catch (error) {\n // Log the error with proper context but don't interrupt user workflow\n const wrappedError = wrapError(\n error,\n 'Update check failed',\n ErrorCode.INTERNAL_ERROR,\n { currentVersion, silent }\n );\n logger.debug('Update check failed:', { \n error: getErrorMessage(error),\n context: wrappedError.context \n });\n }\n }\n\n /**\n * Fetch latest version from npm registry\n */\n private static async fetchLatestVersion(): Promise<string> {\n try {\n // Use timeout to prevent hanging on slow network\n const fetchVersion = async () => {\n const output = execSync(`npm view ${this.PACKAGE_NAME} version`, {\n encoding: 'utf-8',\n stdio: ['pipe', 'pipe', 'ignore'],\n timeout: 5000, // 5 second timeout\n }).trim();\n return output;\n };\n\n // Wrap with timeout and graceful degradation\n return await gracefulDegrade(\n () => withTimeout(fetchVersion, 5000, 'npm registry timeout'),\n '',\n { operation: 'fetchLatestVersion', package: this.PACKAGE_NAME }\n );\n } catch (error) {\n const wrappedError = wrapError(\n error,\n 'Failed to fetch latest version from npm',\n ErrorCode.SERVICE_UNAVAILABLE,\n { package: this.PACKAGE_NAME }\n );\n logger.debug('Failed to fetch latest version:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n return '';\n }\n }\n\n /**\n * Compare version strings\n */\n private static isNewerVersion(current: string, latest: string): boolean {\n try {\n const currentParts = current.split('.').map(Number);\n const latestParts = latest.split('.').map(Number);\n\n // Handle malformed version strings\n if (currentParts.some(isNaN) || latestParts.some(isNaN)) {\n logger.debug('Invalid version format:', { current, latest });\n return false;\n }\n\n for (let i = 0; i < 3; i++) {\n const latestPart = latestParts[i] ?? 0;\n const currentPart = currentParts[i] ?? 0;\n if (latestPart > currentPart) return true;\n if (latestPart < currentPart) return false;\n }\n return false;\n } catch (error) {\n logger.debug('Version comparison failed:', {\n error: getErrorMessage(error),\n current,\n latest,\n });\n return false;\n }\n }\n\n /**\n * Display update notification\n */\n private static displayUpdateNotification(\n current: string,\n latest: string\n ): void {\n console.log('\\n' + '\u2500'.repeat(60));\n console.log('\uD83D\uDCE6 StackMemory Update Available!');\n console.log(` Current: v${current}`);\n console.log(` Latest: v${latest}`);\n console.log('\\n Update with:');\n console.log(' npm install -g @stackmemoryai/stackmemory@latest');\n console.log('\u2500'.repeat(60) + '\\n');\n }\n\n /**\n * Load update cache\n */\n private static loadCache(): UpdateCache | null {\n try {\n if (!existsSync(this.CACHE_FILE)) {\n return null;\n }\n\n const data = readFileSync(this.CACHE_FILE, 'utf-8');\n const cache = JSON.parse(data) as UpdateCache;\n\n // Validate cache structure\n if (\n typeof cache.lastChecked !== 'number' ||\n typeof cache.latestVersion !== 'string' ||\n typeof cache.currentVersion !== 'string'\n ) {\n logger.debug('Invalid cache format, ignoring');\n return null;\n }\n\n return cache;\n } catch (error) {\n // Cache errors should not interrupt operation\n const wrappedError = wrapError(\n error,\n 'Failed to load update cache',\n ErrorCode.INTERNAL_ERROR,\n { cacheFile: this.CACHE_FILE }\n );\n logger.debug('Failed to load update cache:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n return null;\n }\n }\n\n /**\n * Save update cache\n */\n private static saveCache(cache: UpdateCache): void {\n try {\n const dir = join(homedir(), '.stackmemory');\n \n // Create directory if it doesn't exist (safer than execSync)\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true, mode: 0o755 });\n }\n \n // Write cache with atomic operation (write to temp, then rename)\n const tempFile = `${this.CACHE_FILE}.tmp`;\n writeFileSync(tempFile, JSON.stringify(cache, null, 2), {\n mode: 0o644,\n });\n \n // Atomic rename\n if (existsSync(this.CACHE_FILE)) {\n writeFileSync(this.CACHE_FILE, JSON.stringify(cache, null, 2));\n } else {\n writeFileSync(this.CACHE_FILE, JSON.stringify(cache, null, 2));\n }\n } catch (error) {\n // Cache save errors should not interrupt operation\n const wrappedError = wrapError(\n error,\n 'Failed to save update cache',\n ErrorCode.INTERNAL_ERROR,\n { cacheFile: this.CACHE_FILE, cache }\n );\n logger.debug('Failed to save update cache:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n }\n }\n\n /**\n * Force check for updates (ignores cache)\n */\n static async forceCheck(currentVersion: string): Promise<void> {\n try {\n const latestVersion = await this.fetchLatestVersion();\n\n // Update cache\n this.saveCache({\n lastChecked: Date.now(),\n latestVersion,\n currentVersion,\n });\n\n if (latestVersion) {\n if (this.isNewerVersion(currentVersion, latestVersion)) {\n this.displayUpdateNotification(currentVersion, latestVersion);\n } else {\n console.log(`\u2705 StackMemory is up to date (v${currentVersion})`);\n }\n }\n } catch (error) {\n console.error('\u274C Update check failed:', (error as Error).message);\n }\n }\n}\n"],
|
|
5
|
-
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,cAAc,eAAe,iBAAiB;AACnE,SAAS,YAAY;AACrB,SAAS,eAAe;AACxB,SAAS,cAAc;AACvB;AAAA,EAEE;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,aAAa,uBAAuB;AAQtC,MAAM,cAAc;AAAA,EACzB,OAAe,aAAa;AAAA,IAC1B,QAAQ;AAAA,IACR;AAAA,IACA;AAAA,EACF;AAAA,EACA,OAAe,iBAAiB,KAAK,KAAK,KAAK;AAAA;AAAA,EAC/C,OAAe,eAAe;AAAA;AAAA;AAAA;AAAA,EAK9B,aAAa,gBACX,gBACA,SAAS,OACM;AACf,QAAI;AAEF,YAAM,QAAQ,KAAK,UAAU;AAC7B,YAAM,MAAM,KAAK,IAAI;AAGrB,UAAI,SAAS,MAAM,MAAM,cAAc,KAAK,gBAAgB;AAC1D,YACE,CAAC,UACD,MAAM,iBACN,MAAM,kBAAkB,gBACxB;AACA,eAAK,0BAA0B,gBAAgB,MAAM,aAAa;AAAA,QACpE;AACA;AAAA,MACF;AAGA,YAAM,gBAAgB,MAAM,KAAK,mBAAmB;AAGpD,WAAK,UAAU;AAAA,QACb,aAAa;AAAA,QACb;AAAA,QACA;AAAA,MACF,CAAC;AAGD,UACE,CAAC,UACD,iBACA,KAAK,eAAe,gBAAgB,aAAa,GACjD;AACA,aAAK,0BAA0B,gBAAgB,aAAa;AAAA,MAC9D;AAAA,IACF,SAAS,
|
|
4
|
+
"sourcesContent": ["/**\n * Update checker for StackMemory\n * Checks npm registry for newer versions\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, readFileSync, writeFileSync, mkdirSync } from 'fs';\nimport { join } from 'path';\nimport { homedir } from 'os';\nimport { logger } from '../monitoring/logger.js';\nimport { \n SystemError, \n ErrorCode, \n getErrorMessage,\n wrapError\n} from '../errors/index.js';\nimport { withTimeout, gracefulDegrade } from '../errors/recovery.js';\n\ninterface UpdateCache {\n lastChecked: number;\n latestVersion: string;\n currentVersion: string;\n}\n\nexport class UpdateChecker {\n private static CACHE_FILE = join(\n homedir(),\n '.stackmemory',\n 'update-check.json'\n );\n private static CHECK_INTERVAL = 24 * 60 * 60 * 1000; // 24 hours\n private static PACKAGE_NAME = '@stackmemoryai/stackmemory';\n\n /**\n * Check for updates and display notification if needed\n */\n static async checkForUpdates(\n currentVersion: string,\n silent = false\n ): Promise<void> {\n try {\n // Check cache first\n const cache = this.loadCache();\n const now = Date.now();\n\n // Skip check if we checked recently\n if (cache && now - cache.lastChecked < this.CHECK_INTERVAL) {\n if (\n !silent &&\n cache.latestVersion &&\n cache.latestVersion !== currentVersion\n ) {\n this.displayUpdateNotification(currentVersion, cache.latestVersion);\n }\n return;\n }\n\n // Fetch latest version from npm\n const latestVersion = await this.fetchLatestVersion();\n\n // Update cache\n this.saveCache({\n lastChecked: now,\n latestVersion,\n currentVersion,\n });\n\n // Display notification if update available\n if (\n !silent &&\n latestVersion &&\n this.isNewerVersion(currentVersion, latestVersion)\n ) {\n this.displayUpdateNotification(currentVersion, latestVersion);\n }\n } catch (error: unknown) {\n // Log the error with proper context but don't interrupt user workflow\n const wrappedError = wrapError(\n error,\n 'Update check failed',\n ErrorCode.INTERNAL_ERROR,\n { currentVersion, silent }\n );\n logger.debug('Update check failed:', { \n error: getErrorMessage(error),\n context: wrappedError.context \n });\n }\n }\n\n /**\n * Fetch latest version from npm registry\n */\n private static async fetchLatestVersion(): Promise<string> {\n try {\n // Use timeout to prevent hanging on slow network\n const fetchVersion = async () => {\n const output = execSync(`npm view ${this.PACKAGE_NAME} version`, {\n encoding: 'utf-8',\n stdio: ['pipe', 'pipe', 'ignore'],\n timeout: 5000, // 5 second timeout\n }).trim();\n return output;\n };\n\n // Wrap with timeout and graceful degradation\n return await gracefulDegrade(\n () => withTimeout(fetchVersion, 5000, 'npm registry timeout'),\n '',\n { operation: 'fetchLatestVersion', package: this.PACKAGE_NAME }\n );\n } catch (error: unknown) {\n const wrappedError = wrapError(\n error,\n 'Failed to fetch latest version from npm',\n ErrorCode.SERVICE_UNAVAILABLE,\n { package: this.PACKAGE_NAME }\n );\n logger.debug('Failed to fetch latest version:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n return '';\n }\n }\n\n /**\n * Compare version strings\n */\n private static isNewerVersion(current: string, latest: string): boolean {\n try {\n const currentParts = current.split('.').map(Number);\n const latestParts = latest.split('.').map(Number);\n\n // Handle malformed version strings\n if (currentParts.some(isNaN) || latestParts.some(isNaN)) {\n logger.debug('Invalid version format:', { current, latest });\n return false;\n }\n\n for (let i = 0; i < 3; i++) {\n const latestPart = latestParts[i] ?? 0;\n const currentPart = currentParts[i] ?? 0;\n if (latestPart > currentPart) return true;\n if (latestPart < currentPart) return false;\n }\n return false;\n } catch (error: unknown) {\n logger.debug('Version comparison failed:', {\n error: getErrorMessage(error),\n current,\n latest,\n });\n return false;\n }\n }\n\n /**\n * Display update notification\n */\n private static displayUpdateNotification(\n current: string,\n latest: string\n ): void {\n console.log('\\n' + '\u2500'.repeat(60));\n console.log('\uD83D\uDCE6 StackMemory Update Available!');\n console.log(` Current: v${current}`);\n console.log(` Latest: v${latest}`);\n console.log('\\n Update with:');\n console.log(' npm install -g @stackmemoryai/stackmemory@latest');\n console.log('\u2500'.repeat(60) + '\\n');\n }\n\n /**\n * Load update cache\n */\n private static loadCache(): UpdateCache | null {\n try {\n if (!existsSync(this.CACHE_FILE)) {\n return null;\n }\n\n const data = readFileSync(this.CACHE_FILE, 'utf-8');\n const cache = JSON.parse(data) as UpdateCache;\n\n // Validate cache structure\n if (\n typeof cache.lastChecked !== 'number' ||\n typeof cache.latestVersion !== 'string' ||\n typeof cache.currentVersion !== 'string'\n ) {\n logger.debug('Invalid cache format, ignoring');\n return null;\n }\n\n return cache;\n } catch (error: unknown) {\n // Cache errors should not interrupt operation\n const wrappedError = wrapError(\n error,\n 'Failed to load update cache',\n ErrorCode.INTERNAL_ERROR,\n { cacheFile: this.CACHE_FILE }\n );\n logger.debug('Failed to load update cache:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n return null;\n }\n }\n\n /**\n * Save update cache\n */\n private static saveCache(cache: UpdateCache): void {\n try {\n const dir = join(homedir(), '.stackmemory');\n \n // Create directory if it doesn't exist (safer than execSync)\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true, mode: 0o755 });\n }\n \n // Write cache with atomic operation (write to temp, then rename)\n const tempFile = `${this.CACHE_FILE}.tmp`;\n writeFileSync(tempFile, JSON.stringify(cache, null, 2), {\n mode: 0o644,\n });\n \n // Atomic rename\n if (existsSync(this.CACHE_FILE)) {\n writeFileSync(this.CACHE_FILE, JSON.stringify(cache, null, 2));\n } else {\n writeFileSync(this.CACHE_FILE, JSON.stringify(cache, null, 2));\n }\n } catch (error: unknown) {\n // Cache save errors should not interrupt operation\n const wrappedError = wrapError(\n error,\n 'Failed to save update cache',\n ErrorCode.INTERNAL_ERROR,\n { cacheFile: this.CACHE_FILE, cache }\n );\n logger.debug('Failed to save update cache:', {\n error: getErrorMessage(error),\n context: wrappedError.context,\n });\n }\n }\n\n /**\n * Force check for updates (ignores cache)\n */\n static async forceCheck(currentVersion: string): Promise<void> {\n try {\n const latestVersion = await this.fetchLatestVersion();\n\n // Update cache\n this.saveCache({\n lastChecked: Date.now(),\n latestVersion,\n currentVersion,\n });\n\n if (latestVersion) {\n if (this.isNewerVersion(currentVersion, latestVersion)) {\n this.displayUpdateNotification(currentVersion, latestVersion);\n } else {\n console.log(`\u2705 StackMemory is up to date (v${currentVersion})`);\n }\n }\n } catch (error: unknown) {\n console.error('\u274C Update check failed:', (error as Error).message);\n }\n }\n}\n"],
|
|
5
|
+
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,cAAc,eAAe,iBAAiB;AACnE,SAAS,YAAY;AACrB,SAAS,eAAe;AACxB,SAAS,cAAc;AACvB;AAAA,EAEE;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,aAAa,uBAAuB;AAQtC,MAAM,cAAc;AAAA,EACzB,OAAe,aAAa;AAAA,IAC1B,QAAQ;AAAA,IACR;AAAA,IACA;AAAA,EACF;AAAA,EACA,OAAe,iBAAiB,KAAK,KAAK,KAAK;AAAA;AAAA,EAC/C,OAAe,eAAe;AAAA;AAAA;AAAA;AAAA,EAK9B,aAAa,gBACX,gBACA,SAAS,OACM;AACf,QAAI;AAEF,YAAM,QAAQ,KAAK,UAAU;AAC7B,YAAM,MAAM,KAAK,IAAI;AAGrB,UAAI,SAAS,MAAM,MAAM,cAAc,KAAK,gBAAgB;AAC1D,YACE,CAAC,UACD,MAAM,iBACN,MAAM,kBAAkB,gBACxB;AACA,eAAK,0BAA0B,gBAAgB,MAAM,aAAa;AAAA,QACpE;AACA;AAAA,MACF;AAGA,YAAM,gBAAgB,MAAM,KAAK,mBAAmB;AAGpD,WAAK,UAAU;AAAA,QACb,aAAa;AAAA,QACb;AAAA,QACA;AAAA,MACF,CAAC;AAGD,UACE,CAAC,UACD,iBACA,KAAK,eAAe,gBAAgB,aAAa,GACjD;AACA,aAAK,0BAA0B,gBAAgB,aAAa;AAAA,MAC9D;AAAA,IACF,SAAS,OAAgB;AAEvB,YAAM,eAAe;AAAA,QACnB;AAAA,QACA;AAAA,QACA,UAAU;AAAA,QACV,EAAE,gBAAgB,OAAO;AAAA,MAC3B;AACA,aAAO,MAAM,wBAAwB;AAAA,QACnC,OAAO,gBAAgB,KAAK;AAAA,QAC5B,SAAS,aAAa;AAAA,MACxB,CAAC;AAAA,IACH;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,aAAqB,qBAAsC;AACzD,QAAI;AAEF,YAAM,eAAe,YAAY;AAC/B,cAAM,SAAS,SAAS,YAAY,KAAK,YAAY,YAAY;AAAA,UAC/D,UAAU;AAAA,UACV,OAAO,CAAC,QAAQ,QAAQ,QAAQ;AAAA,UAChC,SAAS;AAAA;AAAA,QACX,CAAC,EAAE,KAAK;AACR,eAAO;AAAA,MACT;AAGA,aAAO,MAAM;AAAA,QACX,MAAM,YAAY,cAAc,KAAM,sBAAsB;AAAA,QAC5D;AAAA,QACA,EAAE,WAAW,sBAAsB,SAAS,KAAK,aAAa;AAAA,MAChE;AAAA,IACF,SAAS,OAAgB;AACvB,YAAM,eAAe;AAAA,QACnB;AAAA,QACA;AAAA,QACA,UAAU;AAAA,QACV,EAAE,SAAS,KAAK,aAAa;AAAA,MAC/B;AACA,aAAO,MAAM,mCAAmC;AAAA,QAC9C,OAAO,gBAAgB,KAAK;AAAA,QAC5B,SAAS,aAAa;AAAA,MACxB,CAAC;AACD,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,OAAe,eAAe,SAAiB,QAAyB;AACtE,QAAI;AACF,YAAM,eAAe,QAAQ,MAAM,GAAG,EAAE,IAAI,MAAM;AAClD,YAAM,cAAc,OAAO,MAAM,GAAG,EAAE,IAAI,MAAM;AAGhD,UAAI,aAAa,KAAK,KAAK,KAAK,YAAY,KAAK,KAAK,GAAG;AACvD,eAAO,MAAM,2BAA2B,EAAE,SAAS,OAAO,CAAC;AAC3D,eAAO;AAAA,MACT;AAEA,eAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,cAAM,aAAa,YAAY,CAAC,KAAK;AACrC,cAAM,cAAc,aAAa,CAAC,KAAK;AACvC,YAAI,aAAa,YAAa,QAAO;AACrC,YAAI,aAAa,YAAa,QAAO;AAAA,MACvC;AACA,aAAO;AAAA,IACT,SAAS,OAAgB;AACvB,aAAO,MAAM,8BAA8B;AAAA,QACzC,OAAO,gBAAgB,KAAK;AAAA,QAC5B;AAAA,QACA;AAAA,MACF,CAAC;AACD,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,OAAe,0BACb,SACA,QACM;AACN,YAAQ,IAAI,OAAO,SAAI,OAAO,EAAE,CAAC;AACjC,YAAQ,IAAI,yCAAkC;AAC9C,YAAQ,IAAI,gBAAgB,OAAO,EAAE;AACrC,YAAQ,IAAI,gBAAgB,MAAM,EAAE;AACpC,YAAQ,IAAI,mBAAmB;AAC/B,YAAQ,IAAI,qDAAqD;AACjE,YAAQ,IAAI,SAAI,OAAO,EAAE,IAAI,IAAI;AAAA,EACnC;AAAA;AAAA;AAAA;AAAA,EAKA,OAAe,YAAgC;AAC7C,QAAI;AACF,UAAI,CAAC,WAAW,KAAK,UAAU,GAAG;AAChC,eAAO;AAAA,MACT;AAEA,YAAM,OAAO,aAAa,KAAK,YAAY,OAAO;AAClD,YAAM,QAAQ,KAAK,MAAM,IAAI;AAG7B,UACE,OAAO,MAAM,gBAAgB,YAC7B,OAAO,MAAM,kBAAkB,YAC/B,OAAO,MAAM,mBAAmB,UAChC;AACA,eAAO,MAAM,gCAAgC;AAC7C,eAAO;AAAA,MACT;AAEA,aAAO;AAAA,IACT,SAAS,OAAgB;AAEvB,YAAM,eAAe;AAAA,QACnB;AAAA,QACA;AAAA,QACA,UAAU;AAAA,QACV,EAAE,WAAW,KAAK,WAAW;AAAA,MAC/B;AACA,aAAO,MAAM,gCAAgC;AAAA,QAC3C,OAAO,gBAAgB,KAAK;AAAA,QAC5B,SAAS,aAAa;AAAA,MACxB,CAAC;AACD,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,OAAe,UAAU,OAA0B;AACjD,QAAI;AACF,YAAM,MAAM,KAAK,QAAQ,GAAG,cAAc;AAG1C,UAAI,CAAC,WAAW,GAAG,GAAG;AACpB,kBAAU,KAAK,EAAE,WAAW,MAAM,MAAM,IAAM,CAAC;AAAA,MACjD;AAGA,YAAM,WAAW,GAAG,KAAK,UAAU;AACnC,oBAAc,UAAU,KAAK,UAAU,OAAO,MAAM,CAAC,GAAG;AAAA,QACtD,MAAM;AAAA,MACR,CAAC;AAGD,UAAI,WAAW,KAAK,UAAU,GAAG;AAC/B,sBAAc,KAAK,YAAY,KAAK,UAAU,OAAO,MAAM,CAAC,CAAC;AAAA,MAC/D,OAAO;AACL,sBAAc,KAAK,YAAY,KAAK,UAAU,OAAO,MAAM,CAAC,CAAC;AAAA,MAC/D;AAAA,IACF,SAAS,OAAgB;AAEvB,YAAM,eAAe;AAAA,QACnB;AAAA,QACA;AAAA,QACA,UAAU;AAAA,QACV,EAAE,WAAW,KAAK,YAAY,MAAM;AAAA,MACtC;AACA,aAAO,MAAM,gCAAgC;AAAA,QAC3C,OAAO,gBAAgB,KAAK;AAAA,QAC5B,SAAS,aAAa;AAAA,MACxB,CAAC;AAAA,IACH;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,aAAa,WAAW,gBAAuC;AAC7D,QAAI;AACF,YAAM,gBAAgB,MAAM,KAAK,mBAAmB;AAGpD,WAAK,UAAU;AAAA,QACb,aAAa,KAAK,IAAI;AAAA,QACtB;AAAA,QACA;AAAA,MACF,CAAC;AAED,UAAI,eAAe;AACjB,YAAI,KAAK,eAAe,gBAAgB,aAAa,GAAG;AACtD,eAAK,0BAA0B,gBAAgB,aAAa;AAAA,QAC9D,OAAO;AACL,kBAAQ,IAAI,sCAAiC,cAAc,GAAG;AAAA,QAChE;AAAA,MACF;AAAA,IACF,SAAS,OAAgB;AACvB,cAAQ,MAAM,+BAA2B,MAAgB,OAAO;AAAA,IAClE;AAAA,EACF;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/worktree/worktree-manager.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Git Worktree Manager for StackMemory\n * Manages multiple instances across git worktrees with context isolation\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, mkdirSync, readFileSync, writeFileSync } from 'fs';\nimport { join, basename, dirname, resolve } from 'path';\nimport { homedir } from 'os';\nimport Database from 'better-sqlite3';\nimport { logger } from '../monitoring/logger.js';\nimport { ProjectManager } from '../projects/project-manager.js';\nimport {\n DatabaseError,\n SystemError,\n ErrorCode,\n createErrorHandler,\n} from '../errors/index.js';\nimport { retry } from '../errors/recovery.js';\n\nexport interface WorktreeInfo {\n path: string;\n branch: string;\n commit: string;\n isMainWorktree: boolean;\n isBare: boolean;\n isDetached: boolean;\n linkedPath?: string; // Path to main worktree\n contextId: string; // Unique context identifier\n}\n\nexport interface WorktreeConfig {\n enabled: boolean;\n autoDetect: boolean;\n isolateContexts: boolean;\n shareGlobalContext: boolean;\n syncInterval?: number; // Minutes between syncs\n maxWorktrees?: number;\n}\n\nexport interface WorktreeContext {\n worktreeId: string;\n projectId: string;\n branch: string;\n contextPath: string;\n dbPath: string;\n lastSynced: Date;\n metadata: Record<string, any>;\n}\n\nexport class WorktreeManager {\n private static instance: WorktreeManager;\n private config: WorktreeConfig;\n private configPath: string;\n private worktreeCache: Map<string, WorktreeInfo> = new Map();\n private contextMap: Map<string, WorktreeContext> = new Map();\n private db?: Database.Database;\n\n private constructor() {\n this.configPath = join(homedir(), '.stackmemory', 'worktree-config.json');\n this.config = this.loadConfig();\n \n try {\n if (this.config.enabled) {\n this.initialize();\n }\n } catch (error) {\n logger.error(\n 'Failed to initialize WorktreeManager',\n error instanceof Error ? error : new Error(String(error)),\n {\n configPath: this.configPath,\n }\n );\n // Don't throw here - allow the manager to be created without worktree support\n this.config.enabled = false;\n }\n }\n\n static getInstance(): WorktreeManager {\n if (!WorktreeManager.instance) {\n WorktreeManager.instance = new WorktreeManager();\n }\n return WorktreeManager.instance;\n }\n\n /**\n * Initialize worktree management\n */\n private initialize(): void {\n const dbPath = join(homedir(), '.stackmemory', 'worktrees.db');\n const errorHandler = createErrorHandler({\n operation: 'initialize',\n dbPath,\n });\n\n try {\n this.db = new Database(dbPath);\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS worktrees (\n id TEXT PRIMARY KEY,\n path TEXT NOT NULL UNIQUE,\n branch TEXT NOT NULL,\n commit TEXT,\n is_main BOOLEAN,\n is_bare BOOLEAN,\n is_detached BOOLEAN,\n linked_path TEXT,\n context_id TEXT UNIQUE,\n project_id TEXT,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n updated_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS worktree_contexts (\n context_id TEXT PRIMARY KEY,\n worktree_id TEXT NOT NULL,\n project_id TEXT,\n branch TEXT,\n context_path TEXT,\n db_path TEXT,\n last_synced DATETIME,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n FOREIGN KEY (worktree_id) REFERENCES worktrees(id)\n );\n\n CREATE TABLE IF NOT EXISTS context_sync (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n source_context TEXT,\n target_context TEXT,\n sync_type TEXT, -- 'push', 'pull', 'merge'\n data JSON,\n synced_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE INDEX IF NOT EXISTS idx_worktrees_project ON worktrees(project_id);\n CREATE INDEX IF NOT EXISTS idx_contexts_worktree ON worktree_contexts(worktree_id);\n `);\n } catch (error) {\n errorHandler(error);\n }\n }\n\n /**\n * Load configuration\n */\n private loadConfig(): WorktreeConfig {\n if (existsSync(this.configPath)) {\n try {\n return JSON.parse(readFileSync(this.configPath, 'utf-8'));\n } catch (error) {\n logger.error('Failed to load worktree config', error as Error);\n }\n }\n\n // Default config\n return {\n enabled: false,\n autoDetect: true,\n isolateContexts: true,\n shareGlobalContext: false,\n syncInterval: 15,\n maxWorktrees: 10,\n };\n }\n\n /**\n * Save configuration\n */\n saveConfig(config: Partial<WorktreeConfig>): void {\n this.config = { ...this.config, ...config };\n \n const configDir = dirname(this.configPath);\n if (!existsSync(configDir)) {\n mkdirSync(configDir, { recursive: true });\n }\n \n writeFileSync(this.configPath, JSON.stringify(this.config, null, 2));\n \n // Reinitialize if just enabled\n if (config.enabled && !this.db) {\n this.initialize();\n }\n \n logger.info('Worktree configuration updated', { config: this.config });\n }\n\n /**\n * Detect git worktrees in current repository\n */\n detectWorktrees(repoPath?: string): WorktreeInfo[] {\n const path = repoPath || process.cwd();\n \n try {\n // Get worktree list\n const output = execSync('git worktree list --porcelain', {\n cwd: path,\n encoding: 'utf-8',\n });\n\n const worktrees: WorktreeInfo[] = [];\n const lines = output.split('\\n');\n let currentWorktree: Partial<WorktreeInfo> = {};\n\n for (const line of lines) {\n if (line.startsWith('worktree ')) {\n if (currentWorktree.path) {\n worktrees.push(this.finalizeWorktreeInfo(currentWorktree));\n }\n currentWorktree = { path: line.substring(9) };\n } else if (line.startsWith('HEAD ')) {\n currentWorktree.commit = line.substring(5);\n } else if (line.startsWith('branch ')) {\n currentWorktree.branch = line.substring(7);\n } else if (line === 'bare') {\n currentWorktree.isBare = true;\n } else if (line === 'detached') {\n currentWorktree.isDetached = true;\n }\n }\n\n // Add last worktree\n if (currentWorktree.path) {\n worktrees.push(this.finalizeWorktreeInfo(currentWorktree));\n }\n\n // Determine main worktree\n if (worktrees.length > 0) {\n const mainPath = this.getMainWorktreePath(path);\n worktrees.forEach(wt => {\n wt.isMainWorktree = wt.path === mainPath;\n if (!wt.isMainWorktree) {\n wt.linkedPath = mainPath;\n }\n });\n }\n\n // Cache results\n worktrees.forEach(wt => {\n this.worktreeCache.set(wt.path, wt);\n if (this.config.enabled) {\n this.saveWorktree(wt);\n }\n });\n\n logger.info(`Detected ${worktrees.length} worktrees`, { \n count: worktrees.length,\n branches: worktrees.map(w => w.branch).filter(Boolean)\n });\n\n return worktrees;\n } catch (error) {\n logger.debug('Not a git repository or git worktree not available');\n return [];\n }\n }\n\n /**\n * Get main worktree path\n */\n private getMainWorktreePath(path: string): string {\n try {\n const gitDir = execSync('git rev-parse --git-common-dir', {\n cwd: path,\n encoding: 'utf-8',\n }).trim();\n\n // If it's a worktree, find the main repo\n if (gitDir.includes('/.git/worktrees/')) {\n const mainGitDir = gitDir.replace(/\\/\\.git\\/worktrees\\/.*$/, '');\n return mainGitDir;\n }\n\n // It's the main worktree\n return dirname(gitDir);\n } catch {\n return path;\n }\n }\n\n /**\n * Finalize worktree info with defaults\n */\n private finalizeWorktreeInfo(partial: Partial<WorktreeInfo>): WorktreeInfo {\n return {\n path: partial.path || '',\n branch: partial.branch || 'detached',\n commit: partial.commit || '',\n isMainWorktree: partial.isMainWorktree || false,\n isBare: partial.isBare || false,\n isDetached: partial.isDetached || false,\n linkedPath: partial.linkedPath,\n contextId: this.generateContextId(partial.path || '', partial.branch || ''),\n };\n }\n\n /**\n * Generate unique context ID for worktree\n */\n private generateContextId(path: string, branch: string): string {\n const repoName = basename(dirname(path));\n const sanitizedBranch = branch.replace(/[^a-zA-Z0-9-]/g, '_');\n return `${repoName}-${sanitizedBranch}-${Buffer.from(path).toString('base64').substring(0, 8)}`;\n }\n\n /**\n * Save worktree to database\n */\n private async saveWorktree(worktree: WorktreeInfo): Promise<void> {\n if (!this.db) return;\n\n const projectManager = ProjectManager.getInstance();\n const project = await projectManager.detectProject(worktree.path);\n\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO worktrees \n (id, path, branch, commit, is_main, is_bare, is_detached, linked_path, context_id, project_id, updated_at)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, CURRENT_TIMESTAMP)\n `);\n\n stmt.run(\n worktree.contextId,\n worktree.path,\n worktree.branch,\n worktree.commit,\n worktree.isMainWorktree ? 1 : 0,\n worktree.isBare ? 1 : 0,\n worktree.isDetached ? 1 : 0,\n worktree.linkedPath,\n worktree.contextId,\n project.id\n );\n }\n\n /**\n * Get or create context for worktree\n */\n getWorktreeContext(worktreePath: string): WorktreeContext {\n // Check cache\n const cached = this.contextMap.get(worktreePath);\n if (cached) {\n return cached;\n }\n\n const worktree = this.worktreeCache.get(worktreePath) || \n this.detectWorktrees(worktreePath).find(w => w.path === worktreePath);\n\n if (!worktree) {\n throw new Error(`No worktree found at path: ${worktreePath}`);\n }\n\n // Create isolated context path\n const contextBasePath = this.config.isolateContexts\n ? join(homedir(), '.stackmemory', 'worktrees', worktree.contextId)\n : join(worktreePath, '.stackmemory');\n\n // Ensure directory exists\n if (!existsSync(contextBasePath)) {\n mkdirSync(contextBasePath, { recursive: true });\n }\n\n const context: WorktreeContext = {\n worktreeId: worktree.contextId,\n projectId: '', // Will be filled by project manager\n branch: worktree.branch,\n contextPath: contextBasePath,\n dbPath: join(contextBasePath, 'context.db'),\n lastSynced: new Date(),\n metadata: {\n isMainWorktree: worktree.isMainWorktree,\n linkedPath: worktree.linkedPath,\n },\n };\n\n // Save to database\n if (this.db && this.config.enabled) {\n this.saveContext(context);\n }\n\n // Cache it\n this.contextMap.set(worktreePath, context);\n\n logger.info('Created worktree context', {\n worktree: worktree.branch,\n path: contextBasePath,\n });\n\n return context;\n }\n\n /**\n * Save context to database\n */\n private saveContext(context: WorktreeContext): void {\n if (!this.db) return;\n\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO worktree_contexts\n (context_id, worktree_id, project_id, branch, context_path, db_path, last_synced, metadata)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n stmt.run(\n context.worktreeId,\n context.worktreeId,\n context.projectId,\n context.branch,\n context.contextPath,\n context.dbPath,\n context.lastSynced.toISOString(),\n JSON.stringify(context.metadata)\n );\n }\n\n /**\n * Sync contexts between worktrees\n */\n async syncContexts(sourceWorktree: string, targetWorktree: string, syncType: 'push' | 'pull' | 'merge' = 'merge'): Promise<void> {\n const source = this.getWorktreeContext(sourceWorktree);\n const target = this.getWorktreeContext(targetWorktree);\n\n logger.info('Syncing contexts between worktrees', {\n source: source.branch,\n target: target.branch,\n type: syncType,\n });\n\n // Open both databases\n const sourceDb = new Database(source.dbPath);\n const targetDb = new Database(target.dbPath);\n\n try {\n // Get contexts from source\n const contexts = sourceDb.prepare(`\n SELECT * FROM contexts \n WHERE created_at > datetime('now', '-7 days')\n ORDER BY created_at DESC\n `).all();\n\n // Sync based on type\n if (syncType === 'push' || syncType === 'merge') {\n this.mergeContexts(contexts, targetDb, syncType === 'merge');\n }\n\n if (syncType === 'pull') {\n const targetContexts = targetDb.prepare(`\n SELECT * FROM contexts \n WHERE created_at > datetime('now', '-7 days')\n ORDER BY created_at DESC\n `).all();\n \n this.mergeContexts(targetContexts, sourceDb, false);\n }\n\n // Log sync operation\n if (this.db) {\n const stmt = this.db.prepare(`\n INSERT INTO context_sync (source_context, target_context, sync_type, data)\n VALUES (?, ?, ?, ?)\n `);\n\n stmt.run(\n source.worktreeId,\n target.worktreeId,\n syncType,\n JSON.stringify({ count: contexts.length })\n );\n }\n\n logger.info('Context sync completed', {\n synced: contexts.length,\n type: syncType,\n });\n } finally {\n sourceDb.close();\n targetDb.close();\n }\n }\n\n /**\n * Merge contexts into target database\n */\n private mergeContexts(contexts: any[], targetDb: Database.Database, bidirectional: boolean): void {\n const stmt = targetDb.prepare(`\n INSERT OR REPLACE INTO contexts (id, type, content, metadata, created_at)\n VALUES (?, ?, ?, ?, ?)\n `);\n\n for (const ctx of contexts) {\n try {\n stmt.run(ctx.id, ctx.type, ctx.content, ctx.metadata, ctx.created_at);\n } catch (error) {\n logger.warn('Failed to merge context', { id: ctx.id, error });\n }\n }\n }\n\n /**\n * List all active worktrees\n */\n listActiveWorktrees(): WorktreeInfo[] {\n if (!this.db) {\n return Array.from(this.worktreeCache.values());\n }\n\n const stmt = this.db.prepare(`\n SELECT * FROM worktrees\n ORDER BY is_main DESC, branch ASC\n `);\n\n const rows = stmt.all() as any[];\n \n return rows.map(row => ({\n path: row.path,\n branch: row.branch,\n commit: row.commit,\n isMainWorktree: row.is_main === 1,\n isBare: row.is_bare === 1,\n isDetached: row.is_detached === 1,\n linkedPath: row.linked_path,\n contextId: row.context_id,\n }));\n }\n\n /**\n * Clean up stale worktree contexts\n */\n cleanupStaleContexts(): void {\n if (!this.db) return;\n\n const activeWorktrees = this.detectWorktrees();\n const activePaths = new Set(activeWorktrees.map(w => w.path));\n\n // Get all stored worktrees\n const stmt = this.db.prepare('SELECT * FROM worktrees');\n const stored = stmt.all() as any[];\n\n // Remove stale entries\n const deleteStmt = this.db.prepare('DELETE FROM worktrees WHERE id = ?');\n const deleteContextStmt = this.db.prepare('DELETE FROM worktree_contexts WHERE worktree_id = ?');\n\n for (const worktree of stored) {\n if (!activePaths.has(worktree.path)) {\n deleteStmt.run(worktree.id);\n deleteContextStmt.run(worktree.id);\n \n logger.info('Cleaned up stale worktree context', {\n path: worktree.path,\n branch: worktree.branch,\n });\n }\n }\n }\n\n /**\n * Get configuration\n */\n getConfig(): WorktreeConfig {\n return { ...this.config };\n }\n\n /**\n * Check if worktree support is enabled\n */\n isEnabled(): boolean {\n return this.config.enabled;\n }\n\n /**\n * Enable or disable worktree support\n */\n setEnabled(enabled: boolean): void {\n this.saveConfig({ enabled });\n }\n}"],
|
|
5
|
-
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,WAAW,cAAc,qBAAqB;AACnE,SAAS,MAAM,UAAU,eAAwB;AACjD,SAAS,eAAe;AACxB,OAAO,cAAc;AACrB,SAAS,cAAc;AACvB,SAAS,sBAAsB;AAC/B;AAAA,EAIE;AAAA,OACK;AAiCA,MAAM,gBAAgB;AAAA,EAC3B,OAAe;AAAA,EACP;AAAA,EACA;AAAA,EACA,gBAA2C,oBAAI,IAAI;AAAA,EACnD,aAA2C,oBAAI,IAAI;AAAA,EACnD;AAAA,EAEA,cAAc;AACpB,SAAK,aAAa,KAAK,QAAQ,GAAG,gBAAgB,sBAAsB;AACxE,SAAK,SAAS,KAAK,WAAW;AAE9B,QAAI;AACF,UAAI,KAAK,OAAO,SAAS;AACvB,aAAK,WAAW;AAAA,MAClB;AAAA,IACF,SAAS,
|
|
4
|
+
"sourcesContent": ["/**\n * Git Worktree Manager for StackMemory\n * Manages multiple instances across git worktrees with context isolation\n */\n\nimport { execSync } from 'child_process';\nimport { existsSync, mkdirSync, readFileSync, writeFileSync } from 'fs';\nimport { join, basename, dirname, resolve } from 'path';\nimport { homedir } from 'os';\nimport Database from 'better-sqlite3';\nimport { logger } from '../monitoring/logger.js';\nimport { ProjectManager } from '../projects/project-manager.js';\nimport {\n DatabaseError,\n SystemError,\n ErrorCode,\n createErrorHandler,\n} from '../errors/index.js';\nimport { retry } from '../errors/recovery.js';\n\nexport interface WorktreeInfo {\n path: string;\n branch: string;\n commit: string;\n isMainWorktree: boolean;\n isBare: boolean;\n isDetached: boolean;\n linkedPath?: string; // Path to main worktree\n contextId: string; // Unique context identifier\n}\n\nexport interface WorktreeConfig {\n enabled: boolean;\n autoDetect: boolean;\n isolateContexts: boolean;\n shareGlobalContext: boolean;\n syncInterval?: number; // Minutes between syncs\n maxWorktrees?: number;\n}\n\nexport interface WorktreeContext {\n worktreeId: string;\n projectId: string;\n branch: string;\n contextPath: string;\n dbPath: string;\n lastSynced: Date;\n metadata: Record<string, any>;\n}\n\nexport class WorktreeManager {\n private static instance: WorktreeManager;\n private config: WorktreeConfig;\n private configPath: string;\n private worktreeCache: Map<string, WorktreeInfo> = new Map();\n private contextMap: Map<string, WorktreeContext> = new Map();\n private db?: Database.Database;\n\n private constructor() {\n this.configPath = join(homedir(), '.stackmemory', 'worktree-config.json');\n this.config = this.loadConfig();\n \n try {\n if (this.config.enabled) {\n this.initialize();\n }\n } catch (error: unknown) {\n logger.error(\n 'Failed to initialize WorktreeManager',\n error instanceof Error ? error : new Error(String(error)),\n {\n configPath: this.configPath,\n }\n );\n // Don't throw here - allow the manager to be created without worktree support\n this.config.enabled = false;\n }\n }\n\n static getInstance(): WorktreeManager {\n if (!WorktreeManager.instance) {\n WorktreeManager.instance = new WorktreeManager();\n }\n return WorktreeManager.instance;\n }\n\n /**\n * Initialize worktree management\n */\n private initialize(): void {\n const dbPath = join(homedir(), '.stackmemory', 'worktrees.db');\n const errorHandler = createErrorHandler({\n operation: 'initialize',\n dbPath,\n });\n\n try {\n this.db = new Database(dbPath);\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS worktrees (\n id TEXT PRIMARY KEY,\n path TEXT NOT NULL UNIQUE,\n branch TEXT NOT NULL,\n commit TEXT,\n is_main BOOLEAN,\n is_bare BOOLEAN,\n is_detached BOOLEAN,\n linked_path TEXT,\n context_id TEXT UNIQUE,\n project_id TEXT,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n updated_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE TABLE IF NOT EXISTS worktree_contexts (\n context_id TEXT PRIMARY KEY,\n worktree_id TEXT NOT NULL,\n project_id TEXT,\n branch TEXT,\n context_path TEXT,\n db_path TEXT,\n last_synced DATETIME,\n metadata JSON,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n FOREIGN KEY (worktree_id) REFERENCES worktrees(id)\n );\n\n CREATE TABLE IF NOT EXISTS context_sync (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n source_context TEXT,\n target_context TEXT,\n sync_type TEXT, -- 'push', 'pull', 'merge'\n data JSON,\n synced_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n\n CREATE INDEX IF NOT EXISTS idx_worktrees_project ON worktrees(project_id);\n CREATE INDEX IF NOT EXISTS idx_contexts_worktree ON worktree_contexts(worktree_id);\n `);\n } catch (error: unknown) {\n errorHandler(error);\n }\n }\n\n /**\n * Load configuration\n */\n private loadConfig(): WorktreeConfig {\n if (existsSync(this.configPath)) {\n try {\n return JSON.parse(readFileSync(this.configPath, 'utf-8'));\n } catch (error: unknown) {\n logger.error('Failed to load worktree config', error as Error);\n }\n }\n\n // Default config\n return {\n enabled: false,\n autoDetect: true,\n isolateContexts: true,\n shareGlobalContext: false,\n syncInterval: 15,\n maxWorktrees: 10,\n };\n }\n\n /**\n * Save configuration\n */\n saveConfig(config: Partial<WorktreeConfig>): void {\n this.config = { ...this.config, ...config };\n \n const configDir = dirname(this.configPath);\n if (!existsSync(configDir)) {\n mkdirSync(configDir, { recursive: true });\n }\n \n writeFileSync(this.configPath, JSON.stringify(this.config, null, 2));\n \n // Reinitialize if just enabled\n if (config.enabled && !this.db) {\n this.initialize();\n }\n \n logger.info('Worktree configuration updated', { config: this.config });\n }\n\n /**\n * Detect git worktrees in current repository\n */\n detectWorktrees(repoPath?: string): WorktreeInfo[] {\n const path = repoPath || process.cwd();\n \n try {\n // Get worktree list\n const output = execSync('git worktree list --porcelain', {\n cwd: path,\n encoding: 'utf-8',\n });\n\n const worktrees: WorktreeInfo[] = [];\n const lines = output.split('\\n');\n let currentWorktree: Partial<WorktreeInfo> = {};\n\n for (const line of lines) {\n if (line.startsWith('worktree ')) {\n if (currentWorktree.path) {\n worktrees.push(this.finalizeWorktreeInfo(currentWorktree));\n }\n currentWorktree = { path: line.substring(9) };\n } else if (line.startsWith('HEAD ')) {\n currentWorktree.commit = line.substring(5);\n } else if (line.startsWith('branch ')) {\n currentWorktree.branch = line.substring(7);\n } else if (line === 'bare') {\n currentWorktree.isBare = true;\n } else if (line === 'detached') {\n currentWorktree.isDetached = true;\n }\n }\n\n // Add last worktree\n if (currentWorktree.path) {\n worktrees.push(this.finalizeWorktreeInfo(currentWorktree));\n }\n\n // Determine main worktree\n if (worktrees.length > 0) {\n const mainPath = this.getMainWorktreePath(path);\n worktrees.forEach(wt => {\n wt.isMainWorktree = wt.path === mainPath;\n if (!wt.isMainWorktree) {\n wt.linkedPath = mainPath;\n }\n });\n }\n\n // Cache results\n worktrees.forEach(wt => {\n this.worktreeCache.set(wt.path, wt);\n if (this.config.enabled) {\n this.saveWorktree(wt);\n }\n });\n\n logger.info(`Detected ${worktrees.length} worktrees`, { \n count: worktrees.length,\n branches: worktrees.map((w: any) => w.branch).filter(Boolean)\n });\n\n return worktrees;\n } catch (error: unknown) {\n logger.debug('Not a git repository or git worktree not available');\n return [];\n }\n }\n\n /**\n * Get main worktree path\n */\n private getMainWorktreePath(path: string): string {\n try {\n const gitDir = execSync('git rev-parse --git-common-dir', {\n cwd: path,\n encoding: 'utf-8',\n }).trim();\n\n // If it's a worktree, find the main repo\n if (gitDir.includes('/.git/worktrees/')) {\n const mainGitDir = gitDir.replace(/\\/\\.git\\/worktrees\\/.*$/, '');\n return mainGitDir;\n }\n\n // It's the main worktree\n return dirname(gitDir);\n } catch {\n return path;\n }\n }\n\n /**\n * Finalize worktree info with defaults\n */\n private finalizeWorktreeInfo(partial: Partial<WorktreeInfo>): WorktreeInfo {\n return {\n path: partial.path || '',\n branch: partial.branch || 'detached',\n commit: partial.commit || '',\n isMainWorktree: partial.isMainWorktree || false,\n isBare: partial.isBare || false,\n isDetached: partial.isDetached || false,\n linkedPath: partial.linkedPath,\n contextId: this.generateContextId(partial.path || '', partial.branch || ''),\n };\n }\n\n /**\n * Generate unique context ID for worktree\n */\n private generateContextId(path: string, branch: string): string {\n const repoName = basename(dirname(path));\n const sanitizedBranch = branch.replace(/[^a-zA-Z0-9-]/g, '_');\n return `${repoName}-${sanitizedBranch}-${Buffer.from(path).toString('base64').substring(0, 8)}`;\n }\n\n /**\n * Save worktree to database\n */\n private async saveWorktree(worktree: WorktreeInfo): Promise<void> {\n if (!this.db) return;\n\n const projectManager = ProjectManager.getInstance();\n const project = await projectManager.detectProject(worktree.path);\n\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO worktrees \n (id, path, branch, commit, is_main, is_bare, is_detached, linked_path, context_id, project_id, updated_at)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, CURRENT_TIMESTAMP)\n `);\n\n stmt.run(\n worktree.contextId,\n worktree.path,\n worktree.branch,\n worktree.commit,\n worktree.isMainWorktree ? 1 : 0,\n worktree.isBare ? 1 : 0,\n worktree.isDetached ? 1 : 0,\n worktree.linkedPath,\n worktree.contextId,\n project.id\n );\n }\n\n /**\n * Get or create context for worktree\n */\n getWorktreeContext(worktreePath: string): WorktreeContext {\n // Check cache\n const cached = this.contextMap.get(worktreePath);\n if (cached) {\n return cached;\n }\n\n const worktree = this.worktreeCache.get(worktreePath) || \n this.detectWorktrees(worktreePath).find((w: any) => w.path === worktreePath);\n\n if (!worktree) {\n throw new Error(`No worktree found at path: ${worktreePath}`);\n }\n\n // Create isolated context path\n const contextBasePath = this.config.isolateContexts\n ? join(homedir(), '.stackmemory', 'worktrees', worktree.contextId)\n : join(worktreePath, '.stackmemory');\n\n // Ensure directory exists\n if (!existsSync(contextBasePath)) {\n mkdirSync(contextBasePath, { recursive: true });\n }\n\n const context: WorktreeContext = {\n worktreeId: worktree.contextId,\n projectId: '', // Will be filled by project manager\n branch: worktree.branch,\n contextPath: contextBasePath,\n dbPath: join(contextBasePath, 'context.db'),\n lastSynced: new Date(),\n metadata: {\n isMainWorktree: worktree.isMainWorktree,\n linkedPath: worktree.linkedPath,\n },\n };\n\n // Save to database\n if (this.db && this.config.enabled) {\n this.saveContext(context);\n }\n\n // Cache it\n this.contextMap.set(worktreePath, context);\n\n logger.info('Created worktree context', {\n worktree: worktree.branch,\n path: contextBasePath,\n });\n\n return context;\n }\n\n /**\n * Save context to database\n */\n private saveContext(context: WorktreeContext): void {\n if (!this.db) return;\n\n const stmt = this.db.prepare(`\n INSERT OR REPLACE INTO worktree_contexts\n (context_id, worktree_id, project_id, branch, context_path, db_path, last_synced, metadata)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n stmt.run(\n context.worktreeId,\n context.worktreeId,\n context.projectId,\n context.branch,\n context.contextPath,\n context.dbPath,\n context.lastSynced.toISOString(),\n JSON.stringify(context.metadata)\n );\n }\n\n /**\n * Sync contexts between worktrees\n */\n async syncContexts(sourceWorktree: string, targetWorktree: string, syncType: 'push' | 'pull' | 'merge' = 'merge'): Promise<void> {\n const source = this.getWorktreeContext(sourceWorktree);\n const target = this.getWorktreeContext(targetWorktree);\n\n logger.info('Syncing contexts between worktrees', {\n source: source.branch,\n target: target.branch,\n type: syncType,\n });\n\n // Open both databases\n const sourceDb = new Database(source.dbPath);\n const targetDb = new Database(target.dbPath);\n\n try {\n // Get contexts from source\n const contexts = sourceDb.prepare(`\n SELECT * FROM contexts \n WHERE created_at > datetime('now', '-7 days')\n ORDER BY created_at DESC\n `).all();\n\n // Sync based on type\n if (syncType === 'push' || syncType === 'merge') {\n this.mergeContexts(contexts, targetDb, syncType === 'merge');\n }\n\n if (syncType === 'pull') {\n const targetContexts = targetDb.prepare(`\n SELECT * FROM contexts \n WHERE created_at > datetime('now', '-7 days')\n ORDER BY created_at DESC\n `).all();\n \n this.mergeContexts(targetContexts, sourceDb, false);\n }\n\n // Log sync operation\n if (this.db) {\n const stmt = this.db.prepare(`\n INSERT INTO context_sync (source_context, target_context, sync_type, data)\n VALUES (?, ?, ?, ?)\n `);\n\n stmt.run(\n source.worktreeId,\n target.worktreeId,\n syncType,\n JSON.stringify({ count: contexts.length })\n );\n }\n\n logger.info('Context sync completed', {\n synced: contexts.length,\n type: syncType,\n });\n } finally {\n sourceDb.close();\n targetDb.close();\n }\n }\n\n /**\n * Merge contexts into target database\n */\n private mergeContexts(contexts: any[], targetDb: Database.Database, bidirectional: boolean): void {\n const stmt = targetDb.prepare(`\n INSERT OR REPLACE INTO contexts (id, type, content, metadata, created_at)\n VALUES (?, ?, ?, ?, ?)\n `);\n\n for (const ctx of contexts) {\n try {\n stmt.run(ctx.id, ctx.type, ctx.content, ctx.metadata, ctx.created_at);\n } catch (error: unknown) {\n logger.warn('Failed to merge context', { id: ctx.id, error });\n }\n }\n }\n\n /**\n * List all active worktrees\n */\n listActiveWorktrees(): WorktreeInfo[] {\n if (!this.db) {\n return Array.from(this.worktreeCache.values());\n }\n\n const stmt = this.db.prepare(`\n SELECT * FROM worktrees\n ORDER BY is_main DESC, branch ASC\n `);\n\n const rows = stmt.all() as any[];\n \n return rows.map((row: any) => ({\n path: row.path,\n branch: row.branch,\n commit: row.commit,\n isMainWorktree: row.is_main === 1,\n isBare: row.is_bare === 1,\n isDetached: row.is_detached === 1,\n linkedPath: row.linked_path,\n contextId: row.context_id,\n }));\n }\n\n /**\n * Clean up stale worktree contexts\n */\n cleanupStaleContexts(): void {\n if (!this.db) return;\n\n const activeWorktrees = this.detectWorktrees();\n const activePaths = new Set(activeWorktrees.map((w: any) => w.path));\n\n // Get all stored worktrees\n const stmt = this.db.prepare('SELECT * FROM worktrees');\n const stored = stmt.all() as any[];\n\n // Remove stale entries\n const deleteStmt = this.db.prepare('DELETE FROM worktrees WHERE id = ?');\n const deleteContextStmt = this.db.prepare('DELETE FROM worktree_contexts WHERE worktree_id = ?');\n\n for (const worktree of stored) {\n if (!activePaths.has(worktree.path)) {\n deleteStmt.run(worktree.id);\n deleteContextStmt.run(worktree.id);\n \n logger.info('Cleaned up stale worktree context', {\n path: worktree.path,\n branch: worktree.branch,\n });\n }\n }\n }\n\n /**\n * Get configuration\n */\n getConfig(): WorktreeConfig {\n return { ...this.config };\n }\n\n /**\n * Check if worktree support is enabled\n */\n isEnabled(): boolean {\n return this.config.enabled;\n }\n\n /**\n * Enable or disable worktree support\n */\n setEnabled(enabled: boolean): void {\n this.saveConfig({ enabled });\n }\n}"],
|
|
5
|
+
"mappings": "AAKA,SAAS,gBAAgB;AACzB,SAAS,YAAY,WAAW,cAAc,qBAAqB;AACnE,SAAS,MAAM,UAAU,eAAwB;AACjD,SAAS,eAAe;AACxB,OAAO,cAAc;AACrB,SAAS,cAAc;AACvB,SAAS,sBAAsB;AAC/B;AAAA,EAIE;AAAA,OACK;AAiCA,MAAM,gBAAgB;AAAA,EAC3B,OAAe;AAAA,EACP;AAAA,EACA;AAAA,EACA,gBAA2C,oBAAI,IAAI;AAAA,EACnD,aAA2C,oBAAI,IAAI;AAAA,EACnD;AAAA,EAEA,cAAc;AACpB,SAAK,aAAa,KAAK,QAAQ,GAAG,gBAAgB,sBAAsB;AACxE,SAAK,SAAS,KAAK,WAAW;AAE9B,QAAI;AACF,UAAI,KAAK,OAAO,SAAS;AACvB,aAAK,WAAW;AAAA,MAClB;AAAA,IACF,SAAS,OAAgB;AACvB,aAAO;AAAA,QACL;AAAA,QACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QACxD;AAAA,UACE,YAAY,KAAK;AAAA,QACnB;AAAA,MACF;AAEA,WAAK,OAAO,UAAU;AAAA,IACxB;AAAA,EACF;AAAA,EAEA,OAAO,cAA+B;AACpC,QAAI,CAAC,gBAAgB,UAAU;AAC7B,sBAAgB,WAAW,IAAI,gBAAgB;AAAA,IACjD;AACA,WAAO,gBAAgB;AAAA,EACzB;AAAA;AAAA;AAAA;AAAA,EAKQ,aAAmB;AACzB,UAAM,SAAS,KAAK,QAAQ,GAAG,gBAAgB,cAAc;AAC7D,UAAM,eAAe,mBAAmB;AAAA,MACtC,WAAW;AAAA,MACX;AAAA,IACF,CAAC;AAED,QAAI;AACF,WAAK,KAAK,IAAI,SAAS,MAAM;AAE7B,WAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAwCd;AAAA,IACD,SAAS,OAAgB;AACvB,mBAAa,KAAK;AAAA,IACpB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,aAA6B;AACnC,QAAI,WAAW,KAAK,UAAU,GAAG;AAC/B,UAAI;AACF,eAAO,KAAK,MAAM,aAAa,KAAK,YAAY,OAAO,CAAC;AAAA,MAC1D,SAAS,OAAgB;AACvB,eAAO,MAAM,kCAAkC,KAAc;AAAA,MAC/D;AAAA,IACF;AAGA,WAAO;AAAA,MACL,SAAS;AAAA,MACT,YAAY;AAAA,MACZ,iBAAiB;AAAA,MACjB,oBAAoB;AAAA,MACpB,cAAc;AAAA,MACd,cAAc;AAAA,IAChB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,WAAW,QAAuC;AAChD,SAAK,SAAS,EAAE,GAAG,KAAK,QAAQ,GAAG,OAAO;AAE1C,UAAM,YAAY,QAAQ,KAAK,UAAU;AACzC,QAAI,CAAC,WAAW,SAAS,GAAG;AAC1B,gBAAU,WAAW,EAAE,WAAW,KAAK,CAAC;AAAA,IAC1C;AAEA,kBAAc,KAAK,YAAY,KAAK,UAAU,KAAK,QAAQ,MAAM,CAAC,CAAC;AAGnE,QAAI,OAAO,WAAW,CAAC,KAAK,IAAI;AAC9B,WAAK,WAAW;AAAA,IAClB;AAEA,WAAO,KAAK,kCAAkC,EAAE,QAAQ,KAAK,OAAO,CAAC;AAAA,EACvE;AAAA;AAAA;AAAA;AAAA,EAKA,gBAAgB,UAAmC;AACjD,UAAM,OAAO,YAAY,QAAQ,IAAI;AAErC,QAAI;AAEF,YAAM,SAAS,SAAS,iCAAiC;AAAA,QACvD,KAAK;AAAA,QACL,UAAU;AAAA,MACZ,CAAC;AAED,YAAM,YAA4B,CAAC;AACnC,YAAM,QAAQ,OAAO,MAAM,IAAI;AAC/B,UAAI,kBAAyC,CAAC;AAE9C,iBAAW,QAAQ,OAAO;AACxB,YAAI,KAAK,WAAW,WAAW,GAAG;AAChC,cAAI,gBAAgB,MAAM;AACxB,sBAAU,KAAK,KAAK,qBAAqB,eAAe,CAAC;AAAA,UAC3D;AACA,4BAAkB,EAAE,MAAM,KAAK,UAAU,CAAC,EAAE;AAAA,QAC9C,WAAW,KAAK,WAAW,OAAO,GAAG;AACnC,0BAAgB,SAAS,KAAK,UAAU,CAAC;AAAA,QAC3C,WAAW,KAAK,WAAW,SAAS,GAAG;AACrC,0BAAgB,SAAS,KAAK,UAAU,CAAC;AAAA,QAC3C,WAAW,SAAS,QAAQ;AAC1B,0BAAgB,SAAS;AAAA,QAC3B,WAAW,SAAS,YAAY;AAC9B,0BAAgB,aAAa;AAAA,QAC/B;AAAA,MACF;AAGA,UAAI,gBAAgB,MAAM;AACxB,kBAAU,KAAK,KAAK,qBAAqB,eAAe,CAAC;AAAA,MAC3D;AAGA,UAAI,UAAU,SAAS,GAAG;AACxB,cAAM,WAAW,KAAK,oBAAoB,IAAI;AAC9C,kBAAU,QAAQ,QAAM;AACtB,aAAG,iBAAiB,GAAG,SAAS;AAChC,cAAI,CAAC,GAAG,gBAAgB;AACtB,eAAG,aAAa;AAAA,UAClB;AAAA,QACF,CAAC;AAAA,MACH;AAGA,gBAAU,QAAQ,QAAM;AACtB,aAAK,cAAc,IAAI,GAAG,MAAM,EAAE;AAClC,YAAI,KAAK,OAAO,SAAS;AACvB,eAAK,aAAa,EAAE;AAAA,QACtB;AAAA,MACF,CAAC;AAED,aAAO,KAAK,YAAY,UAAU,MAAM,cAAc;AAAA,QACpD,OAAO,UAAU;AAAA,QACjB,UAAU,UAAU,IAAI,CAAC,MAAW,EAAE,MAAM,EAAE,OAAO,OAAO;AAAA,MAC9D,CAAC;AAED,aAAO;AAAA,IACT,SAAS,OAAgB;AACvB,aAAO,MAAM,oDAAoD;AACjE,aAAO,CAAC;AAAA,IACV;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAoB,MAAsB;AAChD,QAAI;AACF,YAAM,SAAS,SAAS,kCAAkC;AAAA,QACxD,KAAK;AAAA,QACL,UAAU;AAAA,MACZ,CAAC,EAAE,KAAK;AAGR,UAAI,OAAO,SAAS,kBAAkB,GAAG;AACvC,cAAM,aAAa,OAAO,QAAQ,2BAA2B,EAAE;AAC/D,eAAO;AAAA,MACT;AAGA,aAAO,QAAQ,MAAM;AAAA,IACvB,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,qBAAqB,SAA8C;AACzE,WAAO;AAAA,MACL,MAAM,QAAQ,QAAQ;AAAA,MACtB,QAAQ,QAAQ,UAAU;AAAA,MAC1B,QAAQ,QAAQ,UAAU;AAAA,MAC1B,gBAAgB,QAAQ,kBAAkB;AAAA,MAC1C,QAAQ,QAAQ,UAAU;AAAA,MAC1B,YAAY,QAAQ,cAAc;AAAA,MAClC,YAAY,QAAQ;AAAA,MACpB,WAAW,KAAK,kBAAkB,QAAQ,QAAQ,IAAI,QAAQ,UAAU,EAAE;AAAA,IAC5E;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,MAAc,QAAwB;AAC9D,UAAM,WAAW,SAAS,QAAQ,IAAI,CAAC;AACvC,UAAM,kBAAkB,OAAO,QAAQ,kBAAkB,GAAG;AAC5D,WAAO,GAAG,QAAQ,IAAI,eAAe,IAAI,OAAO,KAAK,IAAI,EAAE,SAAS,QAAQ,EAAE,UAAU,GAAG,CAAC,CAAC;AAAA,EAC/F;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,aAAa,UAAuC;AAChE,QAAI,CAAC,KAAK,GAAI;AAEd,UAAM,iBAAiB,eAAe,YAAY;AAClD,UAAM,UAAU,MAAM,eAAe,cAAc,SAAS,IAAI;AAEhE,UAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,KAI5B;AAED,SAAK;AAAA,MACH,SAAS;AAAA,MACT,SAAS;AAAA,MACT,SAAS;AAAA,MACT,SAAS;AAAA,MACT,SAAS,iBAAiB,IAAI;AAAA,MAC9B,SAAS,SAAS,IAAI;AAAA,MACtB,SAAS,aAAa,IAAI;AAAA,MAC1B,SAAS;AAAA,MACT,SAAS;AAAA,MACT,QAAQ;AAAA,IACV;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,mBAAmB,cAAuC;AAExD,UAAM,SAAS,KAAK,WAAW,IAAI,YAAY;AAC/C,QAAI,QAAQ;AACV,aAAO;AAAA,IACT;AAEA,UAAM,WAAW,KAAK,cAAc,IAAI,YAAY,KACnC,KAAK,gBAAgB,YAAY,EAAE,KAAK,CAAC,MAAW,EAAE,SAAS,YAAY;AAE5F,QAAI,CAAC,UAAU;AACb,YAAM,IAAI,MAAM,8BAA8B,YAAY,EAAE;AAAA,IAC9D;AAGA,UAAM,kBAAkB,KAAK,OAAO,kBAChC,KAAK,QAAQ,GAAG,gBAAgB,aAAa,SAAS,SAAS,IAC/D,KAAK,cAAc,cAAc;AAGrC,QAAI,CAAC,WAAW,eAAe,GAAG;AAChC,gBAAU,iBAAiB,EAAE,WAAW,KAAK,CAAC;AAAA,IAChD;AAEA,UAAM,UAA2B;AAAA,MAC/B,YAAY,SAAS;AAAA,MACrB,WAAW;AAAA;AAAA,MACX,QAAQ,SAAS;AAAA,MACjB,aAAa;AAAA,MACb,QAAQ,KAAK,iBAAiB,YAAY;AAAA,MAC1C,YAAY,oBAAI,KAAK;AAAA,MACrB,UAAU;AAAA,QACR,gBAAgB,SAAS;AAAA,QACzB,YAAY,SAAS;AAAA,MACvB;AAAA,IACF;AAGA,QAAI,KAAK,MAAM,KAAK,OAAO,SAAS;AAClC,WAAK,YAAY,OAAO;AAAA,IAC1B;AAGA,SAAK,WAAW,IAAI,cAAc,OAAO;AAEzC,WAAO,KAAK,4BAA4B;AAAA,MACtC,UAAU,SAAS;AAAA,MACnB,MAAM;AAAA,IACR,CAAC;AAED,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,YAAY,SAAgC;AAClD,QAAI,CAAC,KAAK,GAAI;AAEd,UAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,KAI5B;AAED,SAAK;AAAA,MACH,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ,WAAW,YAAY;AAAA,MAC/B,KAAK,UAAU,QAAQ,QAAQ;AAAA,IACjC;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,aAAa,gBAAwB,gBAAwB,WAAsC,SAAwB;AAC/H,UAAM,SAAS,KAAK,mBAAmB,cAAc;AACrD,UAAM,SAAS,KAAK,mBAAmB,cAAc;AAErD,WAAO,KAAK,sCAAsC;AAAA,MAChD,QAAQ,OAAO;AAAA,MACf,QAAQ,OAAO;AAAA,MACf,MAAM;AAAA,IACR,CAAC;AAGD,UAAM,WAAW,IAAI,SAAS,OAAO,MAAM;AAC3C,UAAM,WAAW,IAAI,SAAS,OAAO,MAAM;AAE3C,QAAI;AAEF,YAAM,WAAW,SAAS,QAAQ;AAAA;AAAA;AAAA;AAAA,OAIjC,EAAE,IAAI;AAGP,UAAI,aAAa,UAAU,aAAa,SAAS;AAC/C,aAAK,cAAc,UAAU,UAAU,aAAa,OAAO;AAAA,MAC7D;AAEA,UAAI,aAAa,QAAQ;AACvB,cAAM,iBAAiB,SAAS,QAAQ;AAAA;AAAA;AAAA;AAAA,SAIvC,EAAE,IAAI;AAEP,aAAK,cAAc,gBAAgB,UAAU,KAAK;AAAA,MACpD;AAGA,UAAI,KAAK,IAAI;AACX,cAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,SAG5B;AAED,aAAK;AAAA,UACH,OAAO;AAAA,UACP,OAAO;AAAA,UACP;AAAA,UACA,KAAK,UAAU,EAAE,OAAO,SAAS,OAAO,CAAC;AAAA,QAC3C;AAAA,MACF;AAEA,aAAO,KAAK,0BAA0B;AAAA,QACpC,QAAQ,SAAS;AAAA,QACjB,MAAM;AAAA,MACR,CAAC;AAAA,IACH,UAAE;AACA,eAAS,MAAM;AACf,eAAS,MAAM;AAAA,IACjB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,cAAc,UAAiB,UAA6B,eAA8B;AAChG,UAAM,OAAO,SAAS,QAAQ;AAAA;AAAA;AAAA,KAG7B;AAED,eAAW,OAAO,UAAU;AAC1B,UAAI;AACF,aAAK,IAAI,IAAI,IAAI,IAAI,MAAM,IAAI,SAAS,IAAI,UAAU,IAAI,UAAU;AAAA,MACtE,SAAS,OAAgB;AACvB,eAAO,KAAK,2BAA2B,EAAE,IAAI,IAAI,IAAI,MAAM,CAAC;AAAA,MAC9D;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,sBAAsC;AACpC,QAAI,CAAC,KAAK,IAAI;AACZ,aAAO,MAAM,KAAK,KAAK,cAAc,OAAO,CAAC;AAAA,IAC/C;AAEA,UAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA,KAG5B;AAED,UAAM,OAAO,KAAK,IAAI;AAEtB,WAAO,KAAK,IAAI,CAAC,SAAc;AAAA,MAC7B,MAAM,IAAI;AAAA,MACV,QAAQ,IAAI;AAAA,MACZ,QAAQ,IAAI;AAAA,MACZ,gBAAgB,IAAI,YAAY;AAAA,MAChC,QAAQ,IAAI,YAAY;AAAA,MACxB,YAAY,IAAI,gBAAgB;AAAA,MAChC,YAAY,IAAI;AAAA,MAChB,WAAW,IAAI;AAAA,IACjB,EAAE;AAAA,EACJ;AAAA;AAAA;AAAA;AAAA,EAKA,uBAA6B;AAC3B,QAAI,CAAC,KAAK,GAAI;AAEd,UAAM,kBAAkB,KAAK,gBAAgB;AAC7C,UAAM,cAAc,IAAI,IAAI,gBAAgB,IAAI,CAAC,MAAW,EAAE,IAAI,CAAC;AAGnE,UAAM,OAAO,KAAK,GAAG,QAAQ,yBAAyB;AACtD,UAAM,SAAS,KAAK,IAAI;AAGxB,UAAM,aAAa,KAAK,GAAG,QAAQ,oCAAoC;AACvE,UAAM,oBAAoB,KAAK,GAAG,QAAQ,qDAAqD;AAE/F,eAAW,YAAY,QAAQ;AAC7B,UAAI,CAAC,YAAY,IAAI,SAAS,IAAI,GAAG;AACnC,mBAAW,IAAI,SAAS,EAAE;AAC1B,0BAAkB,IAAI,SAAS,EAAE;AAEjC,eAAO,KAAK,qCAAqC;AAAA,UAC/C,MAAM,SAAS;AAAA,UACf,QAAQ,SAAS;AAAA,QACnB,CAAC;AAAA,MACH;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,YAA4B;AAC1B,WAAO,EAAE,GAAG,KAAK,OAAO;AAAA,EAC1B;AAAA;AAAA;AAAA;AAAA,EAKA,YAAqB;AACnB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA;AAAA;AAAA;AAAA,EAKA,WAAW,SAAwB;AACjC,SAAK,WAAW,EAAE,QAAQ,CAAC;AAAA,EAC7B;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../../src/features/analytics/api/analytics-api.ts"],
|
|
4
|
-
"sourcesContent": ["import express, { Request, Response, Router } from 'express';\nimport { AnalyticsService } from '../core/analytics-service.js';\nimport { AnalyticsQuery, TimeRange } from '../types/metrics.js';\nimport { WebSocketServer, WebSocket } from 'ws';\nimport { Server } from 'http';\n\nexport class AnalyticsAPI {\n private router: Router;\n private analyticsService: AnalyticsService;\n private wss?: WebSocketServer;\n\n constructor(projectPath?: string) {\n this.router = express.Router();\n this.analyticsService = new AnalyticsService(projectPath);\n this.setupRoutes();\n }\n\n private setupRoutes(): void {\n this.router.use(express.json());\n\n this.router.get('/metrics', this.getMetrics.bind(this));\n this.router.get('/tasks', this.getTasks.bind(this));\n this.router.get('/team/:userId', this.getTeamMetrics.bind(this));\n this.router.post('/tasks', this.addTask.bind(this));\n this.router.put('/tasks/:taskId', this.updateTask.bind(this));\n this.router.post('/sync', this.syncLinear.bind(this));\n this.router.get('/export', this.exportMetrics.bind(this));\n }\n\n private async getMetrics(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n res.json({\n success: true,\n data: {\n metrics: dashboardState.metrics,\n lastUpdated: dashboardState.lastUpdated,\n },\n });\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async getTasks(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n res.json({\n success: true,\n data: {\n tasks: dashboardState.recentTasks,\n total: dashboardState.metrics.totalTasks,\n },\n });\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async getTeamMetrics(req: Request, res: Response): Promise<void> {\n try {\n const { userId } = req.params;\n const query = this.parseQuery(req.query);\n\n if (userId === 'all') {\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n res.json({\n success: true,\n data: dashboardState.teamMetrics,\n });\n } else {\n const dashboardState = await this.analyticsService.getDashboardState({\n ...query,\n userIds: [userId],\n });\n\n const userMetrics = dashboardState.teamMetrics.find(\n (m) => m.userId === userId\n );\n\n if (!userMetrics) {\n res.status(404).json({\n success: false,\n error: 'User metrics not found',\n });\n return;\n }\n\n res.json({\n success: true,\n data: userMetrics,\n });\n }\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async addTask(req: Request, res: Response): Promise<void> {\n try {\n const task = {\n ...req.body,\n createdAt: new Date(req.body.createdAt || Date.now()),\n completedAt: req.body.completedAt\n ? new Date(req.body.completedAt)\n : undefined,\n };\n\n await this.analyticsService.addTask(task);\n\n res.status(201).json({\n success: true,\n message: 'Task added successfully',\n });\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async updateTask(req: Request, res: Response): Promise<void> {\n try {\n const { taskId } = req.params;\n const updates = req.body;\n\n if (updates.completedAt) {\n updates.completedAt = new Date(updates.completedAt);\n }\n\n await this.analyticsService.updateTask(taskId, updates);\n\n res.json({\n success: true,\n message: 'Task updated successfully',\n });\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async syncLinear(req: Request, res: Response): Promise<void> {\n try {\n await this.analyticsService.syncLinearTasks();\n\n res.json({\n success: true,\n message: 'Linear tasks synced successfully',\n });\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private async exportMetrics(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const format = (req.query.format as 'json' | 'csv') || 'json';\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n if (format === 'csv') {\n const csv = this.convertToCSV(dashboardState);\n res.setHeader('Content-Type', 'text/csv');\n res.setHeader(\n 'Content-Disposition',\n 'attachment; filename=\"analytics-export.csv\"'\n );\n res.send(csv);\n } else {\n res.json({\n success: true,\n data: dashboardState,\n });\n }\n } catch (error) {\n this.handleError(res, error);\n }\n }\n\n private parseQuery(query: any): AnalyticsQuery {\n const result: AnalyticsQuery = {};\n\n if (query.start && query.end) {\n result.timeRange = {\n start: new Date(query.start),\n end: new Date(query.end),\n preset: query.preset,\n };\n } else if (query.preset) {\n result.timeRange = this.getPresetTimeRange(query.preset);\n }\n\n if (query.users) {\n result.userIds = Array.isArray(query.users) ? query.users : [query.users];\n }\n\n if (query.states) {\n result.states = Array.isArray(query.states)\n ? query.states\n : [query.states];\n }\n\n if (query.priorities) {\n result.priorities = Array.isArray(query.priorities)\n ? query.priorities\n : [query.priorities];\n }\n\n if (query.labels) {\n result.labels = Array.isArray(query.labels)\n ? query.labels\n : [query.labels];\n }\n\n if (query.limit) {\n result.limit = parseInt(query.limit);\n }\n\n if (query.offset) {\n result.offset = parseInt(query.offset);\n }\n\n return result;\n }\n\n private getPresetTimeRange(preset: string): TimeRange {\n const end = new Date();\n const start = new Date();\n\n switch (preset) {\n case 'today':\n start.setHours(0, 0, 0, 0);\n break;\n case '7d':\n start.setDate(start.getDate() - 7);\n break;\n case '30d':\n start.setDate(start.getDate() - 30);\n break;\n case '90d':\n start.setDate(start.getDate() - 90);\n break;\n default:\n start.setDate(start.getDate() - 7);\n }\n\n return { start, end, preset: preset as TimeRange['preset'] };\n }\n\n private convertToCSV(dashboardState: any): string {\n const tasks = dashboardState.recentTasks;\n if (!tasks || tasks.length === 0) return 'No data';\n\n const headers = Object.keys(tasks[0]).join(',');\n const rows = tasks.map((task: any) =>\n Object.values(task)\n .map((v) => (typeof v === 'object' ? JSON.stringify(v) : v))\n .join(',')\n );\n\n return [headers, ...rows].join('\\n');\n }\n\n private handleError(res: Response, error: any): void {\n console.error('Analytics API error:', error);\n res.status(500).json({\n success: false,\n error: error.message || 'Internal server error',\n });\n }\n\n setupWebSocket(server: Server): void {\n this.wss = new WebSocketServer({\n server,\n path: '/ws/analytics',\n });\n\n this.wss.on('connection', (ws: WebSocket) => {\n console.log('WebSocket client connected to analytics');\n\n const unsubscribe = this.analyticsService.subscribeToUpdates((state) => {\n if (ws.readyState === WebSocket.OPEN) {\n ws.send(\n JSON.stringify({\n type: 'update',\n data: state,\n })\n );\n }\n });\n\n ws.on('message', async (message: string) => {\n try {\n const data = JSON.parse(message);\n\n if (data.type === 'subscribe') {\n const query = this.parseQuery(data.query || {});\n const state = await this.analyticsService.getDashboardState(query);\n\n ws.send(\n JSON.stringify({\n type: 'initial',\n data: state,\n })\n );\n }\n } catch (error) {\n ws.send(\n JSON.stringify({\n type: 'error',\n error: 'Invalid message format',\n })\n );\n }\n });\n\n ws.on('close', () => {\n unsubscribe();\n console.log('WebSocket client disconnected');\n });\n\n ws.on('error', (error) => {\n console.error('WebSocket error:', error);\n unsubscribe();\n });\n });\n }\n\n getRouter(): Router {\n return this.router;\n }\n\n close(): void {\n this.analyticsService.close();\n if (this.wss) {\n this.wss.close();\n }\n }\n}\n"],
|
|
5
|
-
"mappings": "AAAA,OAAO,aAA4C;AACnD,SAAS,wBAAwB;AAEjC,SAAS,iBAAiB,iBAAiB;AAGpC,MAAM,aAAa;AAAA,EAChB;AAAA,EACA;AAAA,EACA;AAAA,EAER,YAAY,aAAsB;AAChC,SAAK,SAAS,QAAQ,OAAO;AAC7B,SAAK,mBAAmB,IAAI,iBAAiB,WAAW;AACxD,SAAK,YAAY;AAAA,EACnB;AAAA,EAEQ,cAAoB;AAC1B,SAAK,OAAO,IAAI,QAAQ,KAAK,CAAC;AAE9B,SAAK,OAAO,IAAI,YAAY,KAAK,WAAW,KAAK,IAAI,CAAC;AACtD,SAAK,OAAO,IAAI,UAAU,KAAK,SAAS,KAAK,IAAI,CAAC;AAClD,SAAK,OAAO,IAAI,iBAAiB,KAAK,eAAe,KAAK,IAAI,CAAC;AAC/D,SAAK,OAAO,KAAK,UAAU,KAAK,QAAQ,KAAK,IAAI,CAAC;AAClD,SAAK,OAAO,IAAI,kBAAkB,KAAK,WAAW,KAAK,IAAI,CAAC;AAC5D,SAAK,OAAO,KAAK,SAAS,KAAK,WAAW,KAAK,IAAI,CAAC;AACpD,SAAK,OAAO,IAAI,WAAW,KAAK,cAAc,KAAK,IAAI,CAAC;AAAA,EAC1D;AAAA,EAEA,MAAc,WAAW,KAAc,KAA8B;AACnE,QAAI;AACF,YAAM,QAAQ,KAAK,WAAW,IAAI,KAAK;AACvC,YAAM,iBACJ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AAErD,UAAI,KAAK;AAAA,QACP,SAAS;AAAA,QACT,MAAM;AAAA,UACJ,SAAS,eAAe;AAAA,UACxB,aAAa,eAAe;AAAA,QAC9B;AAAA,MACF,CAAC;AAAA,IACH,SAAS,
|
|
4
|
+
"sourcesContent": ["import express, { Request, Response, Router } from 'express';\nimport { AnalyticsService } from '../core/analytics-service.js';\nimport { AnalyticsQuery, TimeRange } from '../types/metrics.js';\nimport { WebSocketServer, WebSocket } from 'ws';\nimport { Server } from 'http';\n\nexport class AnalyticsAPI {\n private router: Router;\n private analyticsService: AnalyticsService;\n private wss?: WebSocketServer;\n\n constructor(projectPath?: string) {\n this.router = express.Router();\n this.analyticsService = new AnalyticsService(projectPath);\n this.setupRoutes();\n }\n\n private setupRoutes(): void {\n this.router.use(express.json());\n\n this.router.get('/metrics', this.getMetrics.bind(this));\n this.router.get('/tasks', this.getTasks.bind(this));\n this.router.get('/team/:userId', this.getTeamMetrics.bind(this));\n this.router.post('/tasks', this.addTask.bind(this));\n this.router.put('/tasks/:taskId', this.updateTask.bind(this));\n this.router.post('/sync', this.syncLinear.bind(this));\n this.router.get('/export', this.exportMetrics.bind(this));\n }\n\n private async getMetrics(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n res.json({\n success: true,\n data: {\n metrics: dashboardState.metrics,\n lastUpdated: dashboardState.lastUpdated,\n },\n });\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async getTasks(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n res.json({\n success: true,\n data: {\n tasks: dashboardState.recentTasks,\n total: dashboardState.metrics.totalTasks,\n },\n });\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async getTeamMetrics(req: Request, res: Response): Promise<void> {\n try {\n const { userId } = req.params;\n const query = this.parseQuery(req.query);\n\n if (userId === 'all') {\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n res.json({\n success: true,\n data: dashboardState.teamMetrics,\n });\n } else {\n const dashboardState = await this.analyticsService.getDashboardState({\n ...query,\n userIds: [userId],\n });\n\n const userMetrics = dashboardState.teamMetrics.find(\n (m) => m.userId === userId\n );\n\n if (!userMetrics) {\n res.status(404).json({\n success: false,\n error: 'User metrics not found',\n });\n return;\n }\n\n res.json({\n success: true,\n data: userMetrics,\n });\n }\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async addTask(req: Request, res: Response): Promise<void> {\n try {\n const task = {\n ...req.body,\n createdAt: new Date(req.body.createdAt || Date.now()),\n completedAt: req.body.completedAt\n ? new Date(req.body.completedAt)\n : undefined,\n };\n\n await this.analyticsService.addTask(task);\n\n res.status(201).json({\n success: true,\n message: 'Task added successfully',\n });\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async updateTask(req: Request, res: Response): Promise<void> {\n try {\n const { taskId } = req.params;\n const updates = req.body;\n\n if (updates.completedAt) {\n updates.completedAt = new Date(updates.completedAt);\n }\n\n await this.analyticsService.updateTask(taskId, updates);\n\n res.json({\n success: true,\n message: 'Task updated successfully',\n });\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async syncLinear(req: Request, res: Response): Promise<void> {\n try {\n await this.analyticsService.syncLinearTasks();\n\n res.json({\n success: true,\n message: 'Linear tasks synced successfully',\n });\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private async exportMetrics(req: Request, res: Response): Promise<void> {\n try {\n const query = this.parseQuery(req.query);\n const format = (req.query.format as 'json' | 'csv') || 'json';\n const dashboardState =\n await this.analyticsService.getDashboardState(query);\n\n if (format === 'csv') {\n const csv = this.convertToCSV(dashboardState);\n res.setHeader('Content-Type', 'text/csv');\n res.setHeader(\n 'Content-Disposition',\n 'attachment; filename=\"analytics-export.csv\"'\n );\n res.send(csv);\n } else {\n res.json({\n success: true,\n data: dashboardState,\n });\n }\n } catch (error: unknown) {\n this.handleError(res, error);\n }\n }\n\n private parseQuery(query: any): AnalyticsQuery {\n const result: AnalyticsQuery = {};\n\n if (query.start && query.end) {\n result.timeRange = {\n start: new Date(query.start),\n end: new Date(query.end),\n preset: query.preset,\n };\n } else if (query.preset) {\n result.timeRange = this.getPresetTimeRange(query.preset);\n }\n\n if (query.users) {\n result.userIds = Array.isArray(query.users) ? query.users : [query.users];\n }\n\n if (query.states) {\n result.states = Array.isArray(query.states)\n ? query.states\n : [query.states];\n }\n\n if (query.priorities) {\n result.priorities = Array.isArray(query.priorities)\n ? query.priorities\n : [query.priorities];\n }\n\n if (query.labels) {\n result.labels = Array.isArray(query.labels)\n ? query.labels\n : [query.labels];\n }\n\n if (query.limit) {\n result.limit = parseInt(query.limit);\n }\n\n if (query.offset) {\n result.offset = parseInt(query.offset);\n }\n\n return result;\n }\n\n private getPresetTimeRange(preset: string): TimeRange {\n const end = new Date();\n const start = new Date();\n\n switch (preset) {\n case 'today':\n start.setHours(0, 0, 0, 0);\n break;\n case '7d':\n start.setDate(start.getDate() - 7);\n break;\n case '30d':\n start.setDate(start.getDate() - 30);\n break;\n case '90d':\n start.setDate(start.getDate() - 90);\n break;\n default:\n start.setDate(start.getDate() - 7);\n }\n\n return { start, end, preset: preset as TimeRange['preset'] };\n }\n\n private convertToCSV(dashboardState: any): string {\n const tasks = dashboardState.recentTasks;\n if (!tasks || tasks.length === 0) return 'No data';\n\n const headers = Object.keys(tasks[0]).join(',');\n const rows = tasks.map((task: any) =>\n Object.values(task)\n .map((v) => (typeof v === 'object' ? JSON.stringify(v) : v))\n .join(',')\n );\n\n return [headers, ...rows].join('\\n');\n }\n\n private handleError(res: Response, error: any): void {\n console.error('Analytics API error:', error);\n res.status(500).json({\n success: false,\n error: error.message || 'Internal server error',\n });\n }\n\n setupWebSocket(server: Server): void {\n this.wss = new WebSocketServer({\n server,\n path: '/ws/analytics',\n });\n\n this.wss.on('connection', (ws: WebSocket) => {\n console.log('WebSocket client connected to analytics');\n\n const unsubscribe = this.analyticsService.subscribeToUpdates((state) => {\n if (ws.readyState === WebSocket.OPEN) {\n ws.send(\n JSON.stringify({\n type: 'update',\n data: state,\n })\n );\n }\n });\n\n ws.on('message', async (message: string) => {\n try {\n const data = JSON.parse(message);\n\n if (data.type === 'subscribe') {\n const query = this.parseQuery(data.query || {});\n const state = await this.analyticsService.getDashboardState(query);\n\n ws.send(\n JSON.stringify({\n type: 'initial',\n data: state,\n })\n );\n }\n } catch (error: unknown) {\n ws.send(\n JSON.stringify({\n type: 'error',\n error: 'Invalid message format',\n })\n );\n }\n });\n\n ws.on('close', () => {\n unsubscribe();\n console.log('WebSocket client disconnected');\n });\n\n ws.on('error', (error) => {\n console.error('WebSocket error:', error);\n unsubscribe();\n });\n });\n }\n\n getRouter(): Router {\n return this.router;\n }\n\n close(): void {\n this.analyticsService.close();\n if (this.wss) {\n this.wss.close();\n }\n }\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,aAA4C;AACnD,SAAS,wBAAwB;AAEjC,SAAS,iBAAiB,iBAAiB;AAGpC,MAAM,aAAa;AAAA,EAChB;AAAA,EACA;AAAA,EACA;AAAA,EAER,YAAY,aAAsB;AAChC,SAAK,SAAS,QAAQ,OAAO;AAC7B,SAAK,mBAAmB,IAAI,iBAAiB,WAAW;AACxD,SAAK,YAAY;AAAA,EACnB;AAAA,EAEQ,cAAoB;AAC1B,SAAK,OAAO,IAAI,QAAQ,KAAK,CAAC;AAE9B,SAAK,OAAO,IAAI,YAAY,KAAK,WAAW,KAAK,IAAI,CAAC;AACtD,SAAK,OAAO,IAAI,UAAU,KAAK,SAAS,KAAK,IAAI,CAAC;AAClD,SAAK,OAAO,IAAI,iBAAiB,KAAK,eAAe,KAAK,IAAI,CAAC;AAC/D,SAAK,OAAO,KAAK,UAAU,KAAK,QAAQ,KAAK,IAAI,CAAC;AAClD,SAAK,OAAO,IAAI,kBAAkB,KAAK,WAAW,KAAK,IAAI,CAAC;AAC5D,SAAK,OAAO,KAAK,SAAS,KAAK,WAAW,KAAK,IAAI,CAAC;AACpD,SAAK,OAAO,IAAI,WAAW,KAAK,cAAc,KAAK,IAAI,CAAC;AAAA,EAC1D;AAAA,EAEA,MAAc,WAAW,KAAc,KAA8B;AACnE,QAAI;AACF,YAAM,QAAQ,KAAK,WAAW,IAAI,KAAK;AACvC,YAAM,iBACJ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AAErD,UAAI,KAAK;AAAA,QACP,SAAS;AAAA,QACT,MAAM;AAAA,UACJ,SAAS,eAAe;AAAA,UACxB,aAAa,eAAe;AAAA,QAC9B;AAAA,MACF,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,SAAS,KAAc,KAA8B;AACjE,QAAI;AACF,YAAM,QAAQ,KAAK,WAAW,IAAI,KAAK;AACvC,YAAM,iBACJ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AAErD,UAAI,KAAK;AAAA,QACP,SAAS;AAAA,QACT,MAAM;AAAA,UACJ,OAAO,eAAe;AAAA,UACtB,OAAO,eAAe,QAAQ;AAAA,QAChC;AAAA,MACF,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,eAAe,KAAc,KAA8B;AACvE,QAAI;AACF,YAAM,EAAE,OAAO,IAAI,IAAI;AACvB,YAAM,QAAQ,KAAK,WAAW,IAAI,KAAK;AAEvC,UAAI,WAAW,OAAO;AACpB,cAAM,iBACJ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AACrD,YAAI,KAAK;AAAA,UACP,SAAS;AAAA,UACT,MAAM,eAAe;AAAA,QACvB,CAAC;AAAA,MACH,OAAO;AACL,cAAM,iBAAiB,MAAM,KAAK,iBAAiB,kBAAkB;AAAA,UACnE,GAAG;AAAA,UACH,SAAS,CAAC,MAAM;AAAA,QAClB,CAAC;AAED,cAAM,cAAc,eAAe,YAAY;AAAA,UAC7C,CAAC,MAAM,EAAE,WAAW;AAAA,QACtB;AAEA,YAAI,CAAC,aAAa;AAChB,cAAI,OAAO,GAAG,EAAE,KAAK;AAAA,YACnB,SAAS;AAAA,YACT,OAAO;AAAA,UACT,CAAC;AACD;AAAA,QACF;AAEA,YAAI,KAAK;AAAA,UACP,SAAS;AAAA,UACT,MAAM;AAAA,QACR,CAAC;AAAA,MACH;AAAA,IACF,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,QAAQ,KAAc,KAA8B;AAChE,QAAI;AACF,YAAM,OAAO;AAAA,QACX,GAAG,IAAI;AAAA,QACP,WAAW,IAAI,KAAK,IAAI,KAAK,aAAa,KAAK,IAAI,CAAC;AAAA,QACpD,aAAa,IAAI,KAAK,cAClB,IAAI,KAAK,IAAI,KAAK,WAAW,IAC7B;AAAA,MACN;AAEA,YAAM,KAAK,iBAAiB,QAAQ,IAAI;AAExC,UAAI,OAAO,GAAG,EAAE,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,SAAS;AAAA,MACX,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,WAAW,KAAc,KAA8B;AACnE,QAAI;AACF,YAAM,EAAE,OAAO,IAAI,IAAI;AACvB,YAAM,UAAU,IAAI;AAEpB,UAAI,QAAQ,aAAa;AACvB,gBAAQ,cAAc,IAAI,KAAK,QAAQ,WAAW;AAAA,MACpD;AAEA,YAAM,KAAK,iBAAiB,WAAW,QAAQ,OAAO;AAEtD,UAAI,KAAK;AAAA,QACP,SAAS;AAAA,QACT,SAAS;AAAA,MACX,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,WAAW,KAAc,KAA8B;AACnE,QAAI;AACF,YAAM,KAAK,iBAAiB,gBAAgB;AAE5C,UAAI,KAAK;AAAA,QACP,SAAS;AAAA,QACT,SAAS;AAAA,MACX,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEA,MAAc,cAAc,KAAc,KAA8B;AACtE,QAAI;AACF,YAAM,QAAQ,KAAK,WAAW,IAAI,KAAK;AACvC,YAAM,SAAU,IAAI,MAAM,UAA6B;AACvD,YAAM,iBACJ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AAErD,UAAI,WAAW,OAAO;AACpB,cAAM,MAAM,KAAK,aAAa,cAAc;AAC5C,YAAI,UAAU,gBAAgB,UAAU;AACxC,YAAI;AAAA,UACF;AAAA,UACA;AAAA,QACF;AACA,YAAI,KAAK,GAAG;AAAA,MACd,OAAO;AACL,YAAI,KAAK;AAAA,UACP,SAAS;AAAA,UACT,MAAM;AAAA,QACR,CAAC;AAAA,MACH;AAAA,IACF,SAAS,OAAgB;AACvB,WAAK,YAAY,KAAK,KAAK;AAAA,IAC7B;AAAA,EACF;AAAA,EAEQ,WAAW,OAA4B;AAC7C,UAAM,SAAyB,CAAC;AAEhC,QAAI,MAAM,SAAS,MAAM,KAAK;AAC5B,aAAO,YAAY;AAAA,QACjB,OAAO,IAAI,KAAK,MAAM,KAAK;AAAA,QAC3B,KAAK,IAAI,KAAK,MAAM,GAAG;AAAA,QACvB,QAAQ,MAAM;AAAA,MAChB;AAAA,IACF,WAAW,MAAM,QAAQ;AACvB,aAAO,YAAY,KAAK,mBAAmB,MAAM,MAAM;AAAA,IACzD;AAEA,QAAI,MAAM,OAAO;AACf,aAAO,UAAU,MAAM,QAAQ,MAAM,KAAK,IAAI,MAAM,QAAQ,CAAC,MAAM,KAAK;AAAA,IAC1E;AAEA,QAAI,MAAM,QAAQ;AAChB,aAAO,SAAS,MAAM,QAAQ,MAAM,MAAM,IACtC,MAAM,SACN,CAAC,MAAM,MAAM;AAAA,IACnB;AAEA,QAAI,MAAM,YAAY;AACpB,aAAO,aAAa,MAAM,QAAQ,MAAM,UAAU,IAC9C,MAAM,aACN,CAAC,MAAM,UAAU;AAAA,IACvB;AAEA,QAAI,MAAM,QAAQ;AAChB,aAAO,SAAS,MAAM,QAAQ,MAAM,MAAM,IACtC,MAAM,SACN,CAAC,MAAM,MAAM;AAAA,IACnB;AAEA,QAAI,MAAM,OAAO;AACf,aAAO,QAAQ,SAAS,MAAM,KAAK;AAAA,IACrC;AAEA,QAAI,MAAM,QAAQ;AAChB,aAAO,SAAS,SAAS,MAAM,MAAM;AAAA,IACvC;AAEA,WAAO;AAAA,EACT;AAAA,EAEQ,mBAAmB,QAA2B;AACpD,UAAM,MAAM,oBAAI,KAAK;AACrB,UAAM,QAAQ,oBAAI,KAAK;AAEvB,YAAQ,QAAQ;AAAA,MACd,KAAK;AACH,cAAM,SAAS,GAAG,GAAG,GAAG,CAAC;AACzB;AAAA,MACF,KAAK;AACH,cAAM,QAAQ,MAAM,QAAQ,IAAI,CAAC;AACjC;AAAA,MACF,KAAK;AACH,cAAM,QAAQ,MAAM,QAAQ,IAAI,EAAE;AAClC;AAAA,MACF,KAAK;AACH,cAAM,QAAQ,MAAM,QAAQ,IAAI,EAAE;AAClC;AAAA,MACF;AACE,cAAM,QAAQ,MAAM,QAAQ,IAAI,CAAC;AAAA,IACrC;AAEA,WAAO,EAAE,OAAO,KAAK,OAAsC;AAAA,EAC7D;AAAA,EAEQ,aAAa,gBAA6B;AAChD,UAAM,QAAQ,eAAe;AAC7B,QAAI,CAAC,SAAS,MAAM,WAAW,EAAG,QAAO;AAEzC,UAAM,UAAU,OAAO,KAAK,MAAM,CAAC,CAAC,EAAE,KAAK,GAAG;AAC9C,UAAM,OAAO,MAAM;AAAA,MAAI,CAAC,SACtB,OAAO,OAAO,IAAI,EACf,IAAI,CAAC,MAAO,OAAO,MAAM,WAAW,KAAK,UAAU,CAAC,IAAI,CAAE,EAC1D,KAAK,GAAG;AAAA,IACb;AAEA,WAAO,CAAC,SAAS,GAAG,IAAI,EAAE,KAAK,IAAI;AAAA,EACrC;AAAA,EAEQ,YAAY,KAAe,OAAkB;AACnD,YAAQ,MAAM,wBAAwB,KAAK;AAC3C,QAAI,OAAO,GAAG,EAAE,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,OAAO,MAAM,WAAW;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,eAAe,QAAsB;AACnC,SAAK,MAAM,IAAI,gBAAgB;AAAA,MAC7B;AAAA,MACA,MAAM;AAAA,IACR,CAAC;AAED,SAAK,IAAI,GAAG,cAAc,CAAC,OAAkB;AAC3C,cAAQ,IAAI,yCAAyC;AAErD,YAAM,cAAc,KAAK,iBAAiB,mBAAmB,CAAC,UAAU;AACtE,YAAI,GAAG,eAAe,UAAU,MAAM;AACpC,aAAG;AAAA,YACD,KAAK,UAAU;AAAA,cACb,MAAM;AAAA,cACN,MAAM;AAAA,YACR,CAAC;AAAA,UACH;AAAA,QACF;AAAA,MACF,CAAC;AAED,SAAG,GAAG,WAAW,OAAO,YAAoB;AAC1C,YAAI;AACF,gBAAM,OAAO,KAAK,MAAM,OAAO;AAE/B,cAAI,KAAK,SAAS,aAAa;AAC7B,kBAAM,QAAQ,KAAK,WAAW,KAAK,SAAS,CAAC,CAAC;AAC9C,kBAAM,QAAQ,MAAM,KAAK,iBAAiB,kBAAkB,KAAK;AAEjE,eAAG;AAAA,cACD,KAAK,UAAU;AAAA,gBACb,MAAM;AAAA,gBACN,MAAM;AAAA,cACR,CAAC;AAAA,YACH;AAAA,UACF;AAAA,QACF,SAAS,OAAgB;AACvB,aAAG;AAAA,YACD,KAAK,UAAU;AAAA,cACb,MAAM;AAAA,cACN,OAAO;AAAA,YACT,CAAC;AAAA,UACH;AAAA,QACF;AAAA,MACF,CAAC;AAED,SAAG,GAAG,SAAS,MAAM;AACnB,oBAAY;AACZ,gBAAQ,IAAI,+BAA+B;AAAA,MAC7C,CAAC;AAED,SAAG,GAAG,SAAS,CAAC,UAAU;AACxB,gBAAQ,MAAM,oBAAoB,KAAK;AACvC,oBAAY;AAAA,MACd,CAAC;AAAA,IACH,CAAC;AAAA,EACH;AAAA,EAEA,YAAoB;AAClB,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,QAAc;AACZ,SAAK,iBAAiB,MAAM;AAC5B,QAAI,KAAK,KAAK;AACZ,WAAK,IAAI,MAAM;AAAA,IACjB;AAAA,EACF;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -5,6 +5,17 @@ import Database from "better-sqlite3";
|
|
|
5
5
|
import path from "path";
|
|
6
6
|
import fs from "fs";
|
|
7
7
|
import os from "os";
|
|
8
|
+
function getEnv(key, defaultValue) {
|
|
9
|
+
const value = process.env[key];
|
|
10
|
+
if (value === void 0) {
|
|
11
|
+
if (defaultValue !== void 0) return defaultValue;
|
|
12
|
+
throw new Error(`Environment variable ${key} is required`);
|
|
13
|
+
}
|
|
14
|
+
return value;
|
|
15
|
+
}
|
|
16
|
+
function getOptionalEnv(key) {
|
|
17
|
+
return process.env[key];
|
|
18
|
+
}
|
|
8
19
|
class AnalyticsService {
|
|
9
20
|
metricsQueries;
|
|
10
21
|
linearClient;
|
|
@@ -18,7 +29,7 @@ class AnalyticsService {
|
|
|
18
29
|
this.ensureDirectoryExists();
|
|
19
30
|
this.metricsQueries = new MetricsQueries(this.dbPath);
|
|
20
31
|
this.initializeTaskStore();
|
|
21
|
-
if (process.env
|
|
32
|
+
if (process.env["LINEAR_API_KEY"]) {
|
|
22
33
|
this.initializeLinearIntegration();
|
|
23
34
|
}
|
|
24
35
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../../src/features/analytics/core/analytics-service.ts"],
|
|
4
|
-
"sourcesContent": ["import { MetricsQueries } from '../queries/metrics-queries.js';\nimport { LinearClient } from '../../../integrations/linear/client.js';\nimport { PebblesTaskStore } from '../../tasks/pebbles-task-store.js';\nimport Database from 'better-sqlite3';\nimport {\n TaskMetrics,\n TeamMetrics,\n TaskAnalytics,\n DashboardState,\n TimeRange,\n AnalyticsQuery,\n} from '../types/metrics.js';\nimport path from 'path';\nimport fs from 'fs';\nimport os from 'os';\n\nexport class AnalyticsService {\n private metricsQueries: MetricsQueries;\n private linearClient?: LinearClient;\n private taskStore?: PebblesTaskStore;\n private dbPath: string;\n private projectPath: string;\n private updateCallbacks: Set<(state: DashboardState) => void> = new Set();\n\n constructor(projectPath?: string) {\n this.projectPath = projectPath || process.cwd();\n this.dbPath = path.join(this.projectPath, '.stackmemory', 'analytics.db');\n\n this.ensureDirectoryExists();\n this.metricsQueries = new MetricsQueries(this.dbPath);\n\n // Initialize task store for syncing\n this.initializeTaskStore();\n\n if (process.env.LINEAR_API_KEY) {\n this.initializeLinearIntegration();\n }\n }\n\n private initializeTaskStore(): void {\n try {\n const contextDbPath = path.join(\n this.projectPath,\n '.stackmemory',\n 'context.db'\n );\n if (fs.existsSync(contextDbPath)) {\n const db = new Database(contextDbPath);\n this.taskStore = new PebblesTaskStore(this.projectPath, db);\n }\n } catch (error) {\n console.error('Failed to initialize task store:', error);\n }\n }\n\n private ensureDirectoryExists(): void {\n const dir = path.dirname(this.dbPath);\n if (!fs.existsSync(dir)) {\n fs.mkdirSync(dir, { recursive: true });\n }\n }\n\n private async initializeLinearIntegration(): Promise<void> {\n try {\n const configPath = path.join(\n os.homedir(),\n '.stackmemory',\n 'linear-config.json'\n );\n if (fs.existsSync(configPath)) {\n const config = JSON.parse(fs.readFileSync(configPath, 'utf-8'));\n this.linearClient = new LinearClient(config);\n await this.syncLinearTasks();\n }\n } catch (error) {\n console.error('Failed to initialize Linear integration:', error);\n }\n }\n\n async syncLinearTasks(): Promise<void> {\n // First sync from task store (which includes Linear-synced tasks)\n await this.syncFromTaskStore();\n\n // Then try direct Linear sync if client available\n if (this.linearClient) {\n try {\n const issues = await this.linearClient.getIssues({ limit: 100 });\n for (const issue of issues) {\n const task: TaskAnalytics = {\n id: issue.id,\n title: issue.title,\n state: this.mapLinearState(issue.state.type),\n createdAt: new Date(issue.createdAt),\n completedAt:\n issue.state.type === 'completed'\n ? new Date(issue.updatedAt)\n : undefined,\n estimatedEffort: issue.estimate ? issue.estimate * 60 : undefined,\n assigneeId: issue.assignee?.id,\n priority: this.mapLinearPriority(issue.priority),\n labels: Array.isArray(issue.labels)\n ? issue.labels.map((l: any) => l.name)\n : (issue.labels as any)?.nodes?.map((l: any) => l.name) || [],\n blockingIssues: [],\n };\n this.metricsQueries.upsertTask(task);\n }\n } catch (error) {\n console.error('Failed to sync from Linear API:', error);\n }\n }\n\n await this.notifyUpdate();\n }\n\n async syncFromTaskStore(): Promise<number> {\n if (!this.taskStore) return 0;\n\n try {\n // Get all tasks including completed ones\n const allTasks = this.getAllTasksFromStore();\n let synced = 0;\n\n for (const task of allTasks) {\n const analyticsTask: TaskAnalytics = {\n id: task.id,\n title: task.title,\n state: this.mapTaskStatus(task.status),\n createdAt: new Date(task.created_at * 1000),\n completedAt: task.completed_at\n ? new Date(task.completed_at * 1000)\n : undefined,\n estimatedEffort: task.estimated_effort,\n actualEffort: task.actual_effort,\n assigneeId: task.assignee,\n priority: task.priority as TaskAnalytics['priority'],\n labels: task.tags || [],\n blockingIssues: task.depends_on || [],\n };\n this.metricsQueries.upsertTask(analyticsTask);\n synced++;\n }\n\n return synced;\n } catch (error) {\n console.error('Failed to sync from task store:', error);\n return 0;\n }\n }\n\n private getAllTasksFromStore(): any[] {\n if (!this.taskStore) return [];\n\n try {\n // Access the db directly to get ALL tasks including completed\n const contextDbPath = path.join(\n this.projectPath,\n '.stackmemory',\n 'context.db'\n );\n const db = new Database(contextDbPath);\n\n const rows = db\n .prepare(\n `\n SELECT * FROM task_cache \n ORDER BY created_at DESC\n `\n )\n .all() as any[];\n\n db.close();\n\n // Hydrate the rows\n return rows.map((row) => ({\n id: row.id,\n title: row.title,\n description: row.description,\n status: row.status,\n priority: row.priority,\n created_at: row.created_at,\n completed_at: row.completed_at,\n estimated_effort: row.estimated_effort,\n actual_effort: row.actual_effort,\n assignee: row.assignee,\n tags: JSON.parse(row.tags || '[]'),\n depends_on: JSON.parse(row.depends_on || '[]'),\n }));\n } catch (error) {\n console.error('Failed to get all tasks:', error);\n return [];\n }\n }\n\n private mapTaskStatus(status: string): TaskAnalytics['state'] {\n const statusMap: Record<string, TaskAnalytics['state']> = {\n pending: 'todo',\n in_progress: 'in_progress',\n completed: 'completed',\n blocked: 'blocked',\n cancelled: 'blocked',\n };\n return statusMap[status] || 'todo';\n }\n\n private mapLinearState(linearState: string): TaskAnalytics['state'] {\n const stateMap: Record<string, TaskAnalytics['state']> = {\n backlog: 'todo',\n unstarted: 'todo',\n started: 'in_progress',\n completed: 'completed',\n done: 'completed',\n canceled: 'blocked',\n };\n return stateMap[linearState.toLowerCase()] || 'todo';\n }\n\n private mapLinearPriority(priority: number): TaskAnalytics['priority'] {\n if (priority === 1) return 'urgent';\n if (priority === 2) return 'high';\n if (priority === 3) return 'medium';\n return 'low';\n }\n\n async getDashboardState(query: AnalyticsQuery = {}): Promise<DashboardState> {\n const timeRange = query.timeRange || this.getDefaultTimeRange();\n\n const metrics = this.metricsQueries.getTaskMetrics({\n ...query,\n timeRange,\n });\n\n const recentTasks = this.metricsQueries.getRecentTasks({\n ...query,\n limit: 20,\n });\n\n const teamMetrics = await this.getTeamMetrics(query);\n\n return {\n metrics,\n teamMetrics,\n recentTasks,\n timeRange,\n teamFilter: query.userIds || [],\n isLive: this.updateCallbacks.size > 0,\n lastUpdated: new Date(),\n };\n }\n\n private async getTeamMetrics(query: AnalyticsQuery): Promise<TeamMetrics[]> {\n const uniqueUserIds = new Set<string>();\n const tasks = this.metricsQueries.getRecentTasks({ limit: 1000 });\n\n tasks.forEach((task) => {\n if (task.assigneeId) {\n uniqueUserIds.add(task.assigneeId);\n }\n });\n\n const teamMetrics: TeamMetrics[] = [];\n const totalCompleted = tasks.filter((t) => t.state === 'completed').length;\n\n for (const userId of uniqueUserIds) {\n const userQuery = { ...query, userIds: [userId] };\n const individualMetrics = this.metricsQueries.getTaskMetrics(userQuery);\n\n teamMetrics.push({\n userId,\n userName: await this.getUserName(userId),\n individualMetrics,\n contributionPercentage:\n totalCompleted > 0\n ? (individualMetrics.completedTasks / totalCompleted) * 100\n : 0,\n lastActive: new Date(),\n });\n }\n\n return teamMetrics.sort(\n (a, b) => b.contributionPercentage - a.contributionPercentage\n );\n }\n\n private async getUserName(userId: string): Promise<string> {\n // Stub for now - would need LinearClient to expose user query method\n return userId;\n }\n\n private getDefaultTimeRange(): TimeRange {\n const end = new Date();\n const start = new Date();\n start.setDate(start.getDate() - 7);\n\n return {\n start,\n end,\n preset: '7d',\n };\n }\n\n subscribeToUpdates(callback: (state: DashboardState) => void): () => void {\n this.updateCallbacks.add(callback);\n\n return () => {\n this.updateCallbacks.delete(callback);\n };\n }\n\n private async notifyUpdate(): Promise<void> {\n const state = await this.getDashboardState();\n this.updateCallbacks.forEach((callback) => callback(state));\n }\n\n async addTask(task: TaskAnalytics): Promise<void> {\n this.metricsQueries.upsertTask(task);\n await this.notifyUpdate();\n }\n\n async updateTask(\n taskId: string,\n updates: Partial<TaskAnalytics>\n ): Promise<void> {\n const tasks = this.metricsQueries.getRecentTasks({ limit: 1 });\n const existingTask = tasks.find((t) => t.id === taskId);\n\n if (existingTask) {\n const updatedTask = { ...existingTask, ...updates };\n this.metricsQueries.upsertTask(updatedTask);\n await this.notifyUpdate();\n }\n }\n\n close(): void {\n this.metricsQueries.close();\n }\n}\n"],
|
|
5
|
-
"mappings": "AAAA,SAAS,sBAAsB;AAC/B,SAAS,oBAAoB;AAC7B,SAAS,wBAAwB;AACjC,OAAO,cAAc;AASrB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,OAAO,QAAQ;
|
|
4
|
+
"sourcesContent": ["import { MetricsQueries } from '../queries/metrics-queries.js';\nimport { LinearClient } from '../../../integrations/linear/client.js';\nimport { PebblesTaskStore } from '../../tasks/pebbles-task-store.js';\nimport Database from 'better-sqlite3';\nimport {\n TaskMetrics,\n TeamMetrics,\n TaskAnalytics,\n DashboardState,\n TimeRange,\n AnalyticsQuery,\n} from '../types/metrics.js';\nimport path from 'path';\nimport fs from 'fs';\nimport os from 'os';\n// Type-safe environment variable access\nfunction getEnv(key: string, defaultValue?: string): string {\n const value = process.env[key];\n if (value === undefined) {\n if (defaultValue !== undefined) return defaultValue;\n throw new Error(`Environment variable ${key} is required`);\n }\n return value;\n}\n\nfunction getOptionalEnv(key: string): string | undefined {\n return process.env[key];\n}\n\n\nexport class AnalyticsService {\n private metricsQueries: MetricsQueries;\n private linearClient?: LinearClient;\n private taskStore?: PebblesTaskStore;\n private dbPath: string;\n private projectPath: string;\n private updateCallbacks: Set<(state: DashboardState) => void> = new Set();\n\n constructor(projectPath?: string) {\n this.projectPath = projectPath || process.cwd();\n this.dbPath = path.join(this.projectPath, '.stackmemory', 'analytics.db');\n\n this.ensureDirectoryExists();\n this.metricsQueries = new MetricsQueries(this.dbPath);\n\n // Initialize task store for syncing\n this.initializeTaskStore();\n\n if (process.env['LINEAR_API_KEY']) {\n this.initializeLinearIntegration();\n }\n }\n\n private initializeTaskStore(): void {\n try {\n const contextDbPath = path.join(\n this.projectPath,\n '.stackmemory',\n 'context.db'\n );\n if (fs.existsSync(contextDbPath)) {\n const db = new Database(contextDbPath);\n this.taskStore = new PebblesTaskStore(this.projectPath, db);\n }\n } catch (error: unknown) {\n console.error('Failed to initialize task store:', error);\n }\n }\n\n private ensureDirectoryExists(): void {\n const dir = path.dirname(this.dbPath);\n if (!fs.existsSync(dir)) {\n fs.mkdirSync(dir, { recursive: true });\n }\n }\n\n private async initializeLinearIntegration(): Promise<void> {\n try {\n const configPath = path.join(\n os.homedir(),\n '.stackmemory',\n 'linear-config.json'\n );\n if (fs.existsSync(configPath)) {\n const config = JSON.parse(fs.readFileSync(configPath, 'utf-8'));\n this.linearClient = new LinearClient(config);\n await this.syncLinearTasks();\n }\n } catch (error: unknown) {\n console.error('Failed to initialize Linear integration:', error);\n }\n }\n\n async syncLinearTasks(): Promise<void> {\n // First sync from task store (which includes Linear-synced tasks)\n await this.syncFromTaskStore();\n\n // Then try direct Linear sync if client available\n if (this.linearClient) {\n try {\n const issues = await this.linearClient.getIssues({ limit: 100 });\n for (const issue of issues) {\n const task: TaskAnalytics = {\n id: issue.id,\n title: issue.title,\n state: this.mapLinearState(issue.state.type),\n createdAt: new Date(issue.createdAt),\n completedAt:\n issue.state.type === 'completed'\n ? new Date(issue.updatedAt)\n : undefined,\n estimatedEffort: issue.estimate ? issue.estimate * 60 : undefined,\n assigneeId: issue.assignee?.id,\n priority: this.mapLinearPriority(issue.priority),\n labels: Array.isArray(issue.labels)\n ? issue.labels.map((l: any) => l.name)\n : (issue.labels as any)?.nodes?.map((l: any) => l.name) || [],\n blockingIssues: [],\n };\n this.metricsQueries.upsertTask(task);\n }\n } catch (error: unknown) {\n console.error('Failed to sync from Linear API:', error);\n }\n }\n\n await this.notifyUpdate();\n }\n\n async syncFromTaskStore(): Promise<number> {\n if (!this.taskStore) return 0;\n\n try {\n // Get all tasks including completed ones\n const allTasks = this.getAllTasksFromStore();\n let synced = 0;\n\n for (const task of allTasks) {\n const analyticsTask: TaskAnalytics = {\n id: task.id,\n title: task.title,\n state: this.mapTaskStatus(task.status),\n createdAt: new Date(task.created_at * 1000),\n completedAt: task.completed_at\n ? new Date(task.completed_at * 1000)\n : undefined,\n estimatedEffort: task.estimated_effort,\n actualEffort: task.actual_effort,\n assigneeId: task.assignee,\n priority: task.priority as TaskAnalytics['priority'],\n labels: task.tags || [],\n blockingIssues: task.depends_on || [],\n };\n this.metricsQueries.upsertTask(analyticsTask);\n synced++;\n }\n\n return synced;\n } catch (error: unknown) {\n console.error('Failed to sync from task store:', error);\n return 0;\n }\n }\n\n private getAllTasksFromStore(): any[] {\n if (!this.taskStore) return [];\n\n try {\n // Access the db directly to get ALL tasks including completed\n const contextDbPath = path.join(\n this.projectPath,\n '.stackmemory',\n 'context.db'\n );\n const db = new Database(contextDbPath);\n\n const rows = db\n .prepare(\n `\n SELECT * FROM task_cache \n ORDER BY created_at DESC\n `\n )\n .all() as any[];\n\n db.close();\n\n // Hydrate the rows\n return rows.map((row) => ({\n id: row.id,\n title: row.title,\n description: row.description,\n status: row.status,\n priority: row.priority,\n created_at: row.created_at,\n completed_at: row.completed_at,\n estimated_effort: row.estimated_effort,\n actual_effort: row.actual_effort,\n assignee: row.assignee,\n tags: JSON.parse(row.tags || '[]'),\n depends_on: JSON.parse(row.depends_on || '[]'),\n }));\n } catch (error: unknown) {\n console.error('Failed to get all tasks:', error);\n return [];\n }\n }\n\n private mapTaskStatus(status: string): TaskAnalytics['state'] {\n const statusMap: Record<string, TaskAnalytics['state']> = {\n pending: 'todo',\n in_progress: 'in_progress',\n completed: 'completed',\n blocked: 'blocked',\n cancelled: 'blocked',\n };\n return statusMap[status] || 'todo';\n }\n\n private mapLinearState(linearState: string): TaskAnalytics['state'] {\n const stateMap: Record<string, TaskAnalytics['state']> = {\n backlog: 'todo',\n unstarted: 'todo',\n started: 'in_progress',\n completed: 'completed',\n done: 'completed',\n canceled: 'blocked',\n };\n return stateMap[linearState.toLowerCase()] || 'todo';\n }\n\n private mapLinearPriority(priority: number): TaskAnalytics['priority'] {\n if (priority === 1) return 'urgent';\n if (priority === 2) return 'high';\n if (priority === 3) return 'medium';\n return 'low';\n }\n\n async getDashboardState(query: AnalyticsQuery = {}): Promise<DashboardState> {\n const timeRange = query.timeRange || this.getDefaultTimeRange();\n\n const metrics = this.metricsQueries.getTaskMetrics({\n ...query,\n timeRange,\n });\n\n const recentTasks = this.metricsQueries.getRecentTasks({\n ...query,\n limit: 20,\n });\n\n const teamMetrics = await this.getTeamMetrics(query);\n\n return {\n metrics,\n teamMetrics,\n recentTasks,\n timeRange,\n teamFilter: query.userIds || [],\n isLive: this.updateCallbacks.size > 0,\n lastUpdated: new Date(),\n };\n }\n\n private async getTeamMetrics(query: AnalyticsQuery): Promise<TeamMetrics[]> {\n const uniqueUserIds = new Set<string>();\n const tasks = this.metricsQueries.getRecentTasks({ limit: 1000 });\n\n tasks.forEach((task) => {\n if (task.assigneeId) {\n uniqueUserIds.add(task.assigneeId);\n }\n });\n\n const teamMetrics: TeamMetrics[] = [];\n const totalCompleted = tasks.filter((t) => t.state === 'completed').length;\n\n for (const userId of uniqueUserIds) {\n const userQuery = { ...query, userIds: [userId] };\n const individualMetrics = this.metricsQueries.getTaskMetrics(userQuery);\n\n teamMetrics.push({\n userId,\n userName: await this.getUserName(userId),\n individualMetrics,\n contributionPercentage:\n totalCompleted > 0\n ? (individualMetrics.completedTasks / totalCompleted) * 100\n : 0,\n lastActive: new Date(),\n });\n }\n\n return teamMetrics.sort(\n (a, b) => b.contributionPercentage - a.contributionPercentage\n );\n }\n\n private async getUserName(userId: string): Promise<string> {\n // Stub for now - would need LinearClient to expose user query method\n return userId;\n }\n\n private getDefaultTimeRange(): TimeRange {\n const end = new Date();\n const start = new Date();\n start.setDate(start.getDate() - 7);\n\n return {\n start,\n end,\n preset: '7d',\n };\n }\n\n subscribeToUpdates(callback: (state: DashboardState) => void): () => void {\n this.updateCallbacks.add(callback);\n\n return () => {\n this.updateCallbacks.delete(callback);\n };\n }\n\n private async notifyUpdate(): Promise<void> {\n const state = await this.getDashboardState();\n this.updateCallbacks.forEach((callback) => callback(state));\n }\n\n async addTask(task: TaskAnalytics): Promise<void> {\n this.metricsQueries.upsertTask(task);\n await this.notifyUpdate();\n }\n\n async updateTask(\n taskId: string,\n updates: Partial<TaskAnalytics>\n ): Promise<void> {\n const tasks = this.metricsQueries.getRecentTasks({ limit: 1 });\n const existingTask = tasks.find((t) => t.id === taskId);\n\n if (existingTask) {\n const updatedTask = { ...existingTask, ...updates };\n this.metricsQueries.upsertTask(updatedTask);\n await this.notifyUpdate();\n }\n }\n\n close(): void {\n this.metricsQueries.close();\n }\n}\n"],
|
|
5
|
+
"mappings": "AAAA,SAAS,sBAAsB;AAC/B,SAAS,oBAAoB;AAC7B,SAAS,wBAAwB;AACjC,OAAO,cAAc;AASrB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,OAAO,QAAQ;AAEf,SAAS,OAAO,KAAa,cAA+B;AAC1D,QAAM,QAAQ,QAAQ,IAAI,GAAG;AAC7B,MAAI,UAAU,QAAW;AACvB,QAAI,iBAAiB,OAAW,QAAO;AACvC,UAAM,IAAI,MAAM,wBAAwB,GAAG,cAAc;AAAA,EAC3D;AACA,SAAO;AACT;AAEA,SAAS,eAAe,KAAiC;AACvD,SAAO,QAAQ,IAAI,GAAG;AACxB;AAGO,MAAM,iBAAiB;AAAA,EACpB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA,kBAAwD,oBAAI,IAAI;AAAA,EAExE,YAAY,aAAsB;AAChC,SAAK,cAAc,eAAe,QAAQ,IAAI;AAC9C,SAAK,SAAS,KAAK,KAAK,KAAK,aAAa,gBAAgB,cAAc;AAExE,SAAK,sBAAsB;AAC3B,SAAK,iBAAiB,IAAI,eAAe,KAAK,MAAM;AAGpD,SAAK,oBAAoB;AAEzB,QAAI,QAAQ,IAAI,gBAAgB,GAAG;AACjC,WAAK,4BAA4B;AAAA,IACnC;AAAA,EACF;AAAA,EAEQ,sBAA4B;AAClC,QAAI;AACF,YAAM,gBAAgB,KAAK;AAAA,QACzB,KAAK;AAAA,QACL;AAAA,QACA;AAAA,MACF;AACA,UAAI,GAAG,WAAW,aAAa,GAAG;AAChC,cAAM,KAAK,IAAI,SAAS,aAAa;AACrC,aAAK,YAAY,IAAI,iBAAiB,KAAK,aAAa,EAAE;AAAA,MAC5D;AAAA,IACF,SAAS,OAAgB;AACvB,cAAQ,MAAM,oCAAoC,KAAK;AAAA,IACzD;AAAA,EACF;AAAA,EAEQ,wBAA8B;AACpC,UAAM,MAAM,KAAK,QAAQ,KAAK,MAAM;AACpC,QAAI,CAAC,GAAG,WAAW,GAAG,GAAG;AACvB,SAAG,UAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,IACvC;AAAA,EACF;AAAA,EAEA,MAAc,8BAA6C;AACzD,QAAI;AACF,YAAM,aAAa,KAAK;AAAA,QACtB,GAAG,QAAQ;AAAA,QACX;AAAA,QACA;AAAA,MACF;AACA,UAAI,GAAG,WAAW,UAAU,GAAG;AAC7B,cAAM,SAAS,KAAK,MAAM,GAAG,aAAa,YAAY,OAAO,CAAC;AAC9D,aAAK,eAAe,IAAI,aAAa,MAAM;AAC3C,cAAM,KAAK,gBAAgB;AAAA,MAC7B;AAAA,IACF,SAAS,OAAgB;AACvB,cAAQ,MAAM,4CAA4C,KAAK;AAAA,IACjE;AAAA,EACF;AAAA,EAEA,MAAM,kBAAiC;AAErC,UAAM,KAAK,kBAAkB;AAG7B,QAAI,KAAK,cAAc;AACrB,UAAI;AACF,cAAM,SAAS,MAAM,KAAK,aAAa,UAAU,EAAE,OAAO,IAAI,CAAC;AAC/D,mBAAW,SAAS,QAAQ;AAC1B,gBAAM,OAAsB;AAAA,YAC1B,IAAI,MAAM;AAAA,YACV,OAAO,MAAM;AAAA,YACb,OAAO,KAAK,eAAe,MAAM,MAAM,IAAI;AAAA,YAC3C,WAAW,IAAI,KAAK,MAAM,SAAS;AAAA,YACnC,aACE,MAAM,MAAM,SAAS,cACjB,IAAI,KAAK,MAAM,SAAS,IACxB;AAAA,YACN,iBAAiB,MAAM,WAAW,MAAM,WAAW,KAAK;AAAA,YACxD,YAAY,MAAM,UAAU;AAAA,YAC5B,UAAU,KAAK,kBAAkB,MAAM,QAAQ;AAAA,YAC/C,QAAQ,MAAM,QAAQ,MAAM,MAAM,IAC9B,MAAM,OAAO,IAAI,CAAC,MAAW,EAAE,IAAI,IAClC,MAAM,QAAgB,OAAO,IAAI,CAAC,MAAW,EAAE,IAAI,KAAK,CAAC;AAAA,YAC9D,gBAAgB,CAAC;AAAA,UACnB;AACA,eAAK,eAAe,WAAW,IAAI;AAAA,QACrC;AAAA,MACF,SAAS,OAAgB;AACvB,gBAAQ,MAAM,mCAAmC,KAAK;AAAA,MACxD;AAAA,IACF;AAEA,UAAM,KAAK,aAAa;AAAA,EAC1B;AAAA,EAEA,MAAM,oBAAqC;AACzC,QAAI,CAAC,KAAK,UAAW,QAAO;AAE5B,QAAI;AAEF,YAAM,WAAW,KAAK,qBAAqB;AAC3C,UAAI,SAAS;AAEb,iBAAW,QAAQ,UAAU;AAC3B,cAAM,gBAA+B;AAAA,UACnC,IAAI,KAAK;AAAA,UACT,OAAO,KAAK;AAAA,UACZ,OAAO,KAAK,cAAc,KAAK,MAAM;AAAA,UACrC,WAAW,IAAI,KAAK,KAAK,aAAa,GAAI;AAAA,UAC1C,aAAa,KAAK,eACd,IAAI,KAAK,KAAK,eAAe,GAAI,IACjC;AAAA,UACJ,iBAAiB,KAAK;AAAA,UACtB,cAAc,KAAK;AAAA,UACnB,YAAY,KAAK;AAAA,UACjB,UAAU,KAAK;AAAA,UACf,QAAQ,KAAK,QAAQ,CAAC;AAAA,UACtB,gBAAgB,KAAK,cAAc,CAAC;AAAA,QACtC;AACA,aAAK,eAAe,WAAW,aAAa;AAC5C;AAAA,MACF;AAEA,aAAO;AAAA,IACT,SAAS,OAAgB;AACvB,cAAQ,MAAM,mCAAmC,KAAK;AACtD,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEQ,uBAA8B;AACpC,QAAI,CAAC,KAAK,UAAW,QAAO,CAAC;AAE7B,QAAI;AAEF,YAAM,gBAAgB,KAAK;AAAA,QACzB,KAAK;AAAA,QACL;AAAA,QACA;AAAA,MACF;AACA,YAAM,KAAK,IAAI,SAAS,aAAa;AAErC,YAAM,OAAO,GACV;AAAA,QACC;AAAA;AAAA;AAAA;AAAA,MAIF,EACC,IAAI;AAEP,SAAG,MAAM;AAGT,aAAO,KAAK,IAAI,CAAC,SAAS;AAAA,QACxB,IAAI,IAAI;AAAA,QACR,OAAO,IAAI;AAAA,QACX,aAAa,IAAI;AAAA,QACjB,QAAQ,IAAI;AAAA,QACZ,UAAU,IAAI;AAAA,QACd,YAAY,IAAI;AAAA,QAChB,cAAc,IAAI;AAAA,QAClB,kBAAkB,IAAI;AAAA,QACtB,eAAe,IAAI;AAAA,QACnB,UAAU,IAAI;AAAA,QACd,MAAM,KAAK,MAAM,IAAI,QAAQ,IAAI;AAAA,QACjC,YAAY,KAAK,MAAM,IAAI,cAAc,IAAI;AAAA,MAC/C,EAAE;AAAA,IACJ,SAAS,OAAgB;AACvB,cAAQ,MAAM,4BAA4B,KAAK;AAC/C,aAAO,CAAC;AAAA,IACV;AAAA,EACF;AAAA,EAEQ,cAAc,QAAwC;AAC5D,UAAM,YAAoD;AAAA,MACxD,SAAS;AAAA,MACT,aAAa;AAAA,MACb,WAAW;AAAA,MACX,SAAS;AAAA,MACT,WAAW;AAAA,IACb;AACA,WAAO,UAAU,MAAM,KAAK;AAAA,EAC9B;AAAA,EAEQ,eAAe,aAA6C;AAClE,UAAM,WAAmD;AAAA,MACvD,SAAS;AAAA,MACT,WAAW;AAAA,MACX,SAAS;AAAA,MACT,WAAW;AAAA,MACX,MAAM;AAAA,MACN,UAAU;AAAA,IACZ;AACA,WAAO,SAAS,YAAY,YAAY,CAAC,KAAK;AAAA,EAChD;AAAA,EAEQ,kBAAkB,UAA6C;AACrE,QAAI,aAAa,EAAG,QAAO;AAC3B,QAAI,aAAa,EAAG,QAAO;AAC3B,QAAI,aAAa,EAAG,QAAO;AAC3B,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,kBAAkB,QAAwB,CAAC,GAA4B;AAC3E,UAAM,YAAY,MAAM,aAAa,KAAK,oBAAoB;AAE9D,UAAM,UAAU,KAAK,eAAe,eAAe;AAAA,MACjD,GAAG;AAAA,MACH;AAAA,IACF,CAAC;AAED,UAAM,cAAc,KAAK,eAAe,eAAe;AAAA,MACrD,GAAG;AAAA,MACH,OAAO;AAAA,IACT,CAAC;AAED,UAAM,cAAc,MAAM,KAAK,eAAe,KAAK;AAEnD,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,YAAY,MAAM,WAAW,CAAC;AAAA,MAC9B,QAAQ,KAAK,gBAAgB,OAAO;AAAA,MACpC,aAAa,oBAAI,KAAK;AAAA,IACxB;AAAA,EACF;AAAA,EAEA,MAAc,eAAe,OAA+C;AAC1E,UAAM,gBAAgB,oBAAI,IAAY;AACtC,UAAM,QAAQ,KAAK,eAAe,eAAe,EAAE,OAAO,IAAK,CAAC;AAEhE,UAAM,QAAQ,CAAC,SAAS;AACtB,UAAI,KAAK,YAAY;AACnB,sBAAc,IAAI,KAAK,UAAU;AAAA,MACnC;AAAA,IACF,CAAC;AAED,UAAM,cAA6B,CAAC;AACpC,UAAM,iBAAiB,MAAM,OAAO,CAAC,MAAM,EAAE,UAAU,WAAW,EAAE;AAEpE,eAAW,UAAU,eAAe;AAClC,YAAM,YAAY,EAAE,GAAG,OAAO,SAAS,CAAC,MAAM,EAAE;AAChD,YAAM,oBAAoB,KAAK,eAAe,eAAe,SAAS;AAEtE,kBAAY,KAAK;AAAA,QACf;AAAA,QACA,UAAU,MAAM,KAAK,YAAY,MAAM;AAAA,QACvC;AAAA,QACA,wBACE,iBAAiB,IACZ,kBAAkB,iBAAiB,iBAAkB,MACtD;AAAA,QACN,YAAY,oBAAI,KAAK;AAAA,MACvB,CAAC;AAAA,IACH;AAEA,WAAO,YAAY;AAAA,MACjB,CAAC,GAAG,MAAM,EAAE,yBAAyB,EAAE;AAAA,IACzC;AAAA,EACF;AAAA,EAEA,MAAc,YAAY,QAAiC;AAEzD,WAAO;AAAA,EACT;AAAA,EAEQ,sBAAiC;AACvC,UAAM,MAAM,oBAAI,KAAK;AACrB,UAAM,QAAQ,oBAAI,KAAK;AACvB,UAAM,QAAQ,MAAM,QAAQ,IAAI,CAAC;AAEjC,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,QAAQ;AAAA,IACV;AAAA,EACF;AAAA,EAEA,mBAAmB,UAAuD;AACxE,SAAK,gBAAgB,IAAI,QAAQ;AAEjC,WAAO,MAAM;AACX,WAAK,gBAAgB,OAAO,QAAQ;AAAA,IACtC;AAAA,EACF;AAAA,EAEA,MAAc,eAA8B;AAC1C,UAAM,QAAQ,MAAM,KAAK,kBAAkB;AAC3C,SAAK,gBAAgB,QAAQ,CAAC,aAAa,SAAS,KAAK,CAAC;AAAA,EAC5D;AAAA,EAEA,MAAM,QAAQ,MAAoC;AAChD,SAAK,eAAe,WAAW,IAAI;AACnC,UAAM,KAAK,aAAa;AAAA,EAC1B;AAAA,EAEA,MAAM,WACJ,QACA,SACe;AACf,UAAM,QAAQ,KAAK,eAAe,eAAe,EAAE,OAAO,EAAE,CAAC;AAC7D,UAAM,eAAe,MAAM,KAAK,CAAC,MAAM,EAAE,OAAO,MAAM;AAEtD,QAAI,cAAc;AAChB,YAAM,cAAc,EAAE,GAAG,cAAc,GAAG,QAAQ;AAClD,WAAK,eAAe,WAAW,WAAW;AAC1C,YAAM,KAAK,aAAa;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,QAAc;AACZ,SAAK,eAAe,MAAM;AAAA,EAC5B;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|