@stackmemoryai/stackmemory 0.3.7 → 0.3.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agents/core/agent-task-manager.js +5 -5
- package/dist/agents/core/agent-task-manager.js.map +2 -2
- package/dist/agents/verifiers/base-verifier.js +2 -2
- package/dist/agents/verifiers/base-verifier.js.map +2 -2
- package/dist/cli/claude-sm.js +0 -11
- package/dist/cli/claude-sm.js.map +2 -2
- package/dist/cli/codex-sm.js +0 -11
- package/dist/cli/codex-sm.js.map +2 -2
- package/dist/cli/commands/chromadb.js +64 -34
- package/dist/cli/commands/chromadb.js.map +2 -2
- package/dist/cli/commands/clear.js +9 -13
- package/dist/cli/commands/clear.js.map +2 -2
- package/dist/cli/commands/config.js +43 -33
- package/dist/cli/commands/config.js.map +2 -2
- package/dist/cli/commands/context.js.map +2 -2
- package/dist/cli/commands/dashboard.js +41 -13
- package/dist/cli/commands/dashboard.js.map +2 -2
- package/dist/cli/commands/gc.js +69 -20
- package/dist/cli/commands/gc.js.map +2 -2
- package/dist/cli/commands/handoff.js.map +2 -2
- package/dist/cli/commands/infinite-storage.js +60 -19
- package/dist/cli/commands/infinite-storage.js.map +2 -2
- package/dist/cli/commands/linear-create.js +36 -8
- package/dist/cli/commands/linear-create.js.map +2 -2
- package/dist/cli/commands/linear-list.js +33 -10
- package/dist/cli/commands/linear-list.js.map +2 -2
- package/dist/cli/commands/linear-migrate.js +17 -4
- package/dist/cli/commands/linear-migrate.js.map +2 -2
- package/dist/cli/commands/linear-test.js +14 -6
- package/dist/cli/commands/linear-test.js.map +2 -2
- package/dist/cli/commands/linear-unified.js +123 -35
- package/dist/cli/commands/linear-unified.js.map +2 -2
- package/dist/cli/commands/linear.js.map +2 -2
- package/dist/cli/commands/monitor.js.map +2 -2
- package/dist/cli/commands/onboard.js +35 -8
- package/dist/cli/commands/onboard.js.map +2 -2
- package/dist/cli/commands/quality.js +2 -7
- package/dist/cli/commands/quality.js.map +2 -2
- package/dist/cli/commands/session.js +23 -6
- package/dist/cli/commands/session.js.map +2 -2
- package/dist/cli/commands/skills.js +72 -27
- package/dist/cli/commands/skills.js.map +2 -2
- package/dist/cli/commands/storage.js +108 -38
- package/dist/cli/commands/storage.js.map +2 -2
- package/dist/cli/commands/tui.js.map +2 -2
- package/dist/cli/commands/webhook.js +57 -18
- package/dist/cli/commands/webhook.js.map +2 -2
- package/dist/cli/commands/workflow.js +8 -15
- package/dist/cli/commands/workflow.js.map +2 -2
- package/dist/cli/commands/worktree.js +34 -13
- package/dist/cli/commands/worktree.js.map +2 -2
- package/dist/cli/index.js +0 -11
- package/dist/cli/index.js.map +2 -2
- package/dist/core/config/types.js.map +1 -1
- package/dist/core/context/auto-context.js +10 -6
- package/dist/core/context/auto-context.js.map +2 -2
- package/dist/core/context/context-bridge.js.map +2 -2
- package/dist/core/context/frame-database.js +13 -3
- package/dist/core/context/frame-database.js.map +2 -2
- package/dist/core/context/frame-digest.js +7 -5
- package/dist/core/context/frame-digest.js.map +2 -2
- package/dist/core/context/frame-manager.js.map +2 -2
- package/dist/core/context/frame-stack.js +16 -5
- package/dist/core/context/frame-stack.js.map +2 -2
- package/dist/core/context/incremental-gc.js +10 -3
- package/dist/core/context/incremental-gc.js.map +2 -2
- package/dist/core/context/index.js.map +1 -1
- package/dist/core/context/permission-manager.js.map +2 -2
- package/dist/core/context/recursive-context-manager.js +582 -0
- package/dist/core/context/recursive-context-manager.js.map +7 -0
- package/dist/core/context/refactored-frame-manager.js +12 -3
- package/dist/core/context/refactored-frame-manager.js.map +2 -2
- package/dist/core/context/shared-context-layer.js +4 -2
- package/dist/core/context/shared-context-layer.js.map +2 -2
- package/dist/core/database/batch-operations.js +112 -86
- package/dist/core/database/batch-operations.js.map +2 -2
- package/dist/core/database/query-cache.js +19 -9
- package/dist/core/database/query-cache.js.map +2 -2
- package/dist/core/database/sqlite-adapter.js +1 -1
- package/dist/core/database/sqlite-adapter.js.map +2 -2
- package/dist/core/digest/enhanced-hybrid-digest.js +8 -2
- package/dist/core/digest/enhanced-hybrid-digest.js.map +2 -2
- package/dist/core/errors/recovery.js +9 -2
- package/dist/core/errors/recovery.js.map +2 -2
- package/dist/core/execution/parallel-executor.js +254 -0
- package/dist/core/execution/parallel-executor.js.map +7 -0
- package/dist/core/frame/workflow-templates-stub.js.map +1 -1
- package/dist/core/frame/workflow-templates.js +40 -1
- package/dist/core/frame/workflow-templates.js.map +2 -2
- package/dist/core/monitoring/logger.js +6 -1
- package/dist/core/monitoring/logger.js.map +2 -2
- package/dist/core/monitoring/metrics.js.map +2 -2
- package/dist/core/monitoring/progress-tracker.js.map +2 -2
- package/dist/core/performance/context-cache.js.map +2 -2
- package/dist/core/performance/lazy-context-loader.js +24 -20
- package/dist/core/performance/lazy-context-loader.js.map +2 -2
- package/dist/core/performance/optimized-frame-context.js +27 -12
- package/dist/core/performance/optimized-frame-context.js.map +2 -2
- package/dist/core/performance/performance-benchmark.js +10 -6
- package/dist/core/performance/performance-benchmark.js.map +2 -2
- package/dist/core/performance/performance-profiler.js +51 -14
- package/dist/core/performance/performance-profiler.js.map +2 -2
- package/dist/core/performance/streaming-jsonl-parser.js +5 -1
- package/dist/core/performance/streaming-jsonl-parser.js.map +2 -2
- package/dist/core/projects/project-manager.js +14 -20
- package/dist/core/projects/project-manager.js.map +2 -2
- package/dist/core/retrieval/context-retriever.js.map +1 -1
- package/dist/core/retrieval/llm-context-retrieval.js.map +2 -2
- package/dist/core/session/clear-survival-stub.js +5 -1
- package/dist/core/session/clear-survival-stub.js.map +2 -2
- package/dist/core/session/clear-survival.js +35 -0
- package/dist/core/session/clear-survival.js.map +2 -2
- package/dist/core/session/index.js.map +1 -1
- package/dist/core/session/session-manager.js.map +2 -2
- package/dist/core/storage/chromadb-adapter.js +6 -2
- package/dist/core/storage/chromadb-adapter.js.map +2 -2
- package/dist/core/storage/chromadb-simple.js +17 -5
- package/dist/core/storage/chromadb-simple.js.map +2 -2
- package/dist/core/storage/infinite-storage.js +109 -46
- package/dist/core/storage/infinite-storage.js.map +2 -2
- package/dist/core/storage/railway-optimized-storage.js +48 -22
- package/dist/core/storage/railway-optimized-storage.js.map +2 -2
- package/dist/core/storage/remote-storage.js +41 -23
- package/dist/core/storage/remote-storage.js.map +2 -2
- package/dist/core/trace/cli-trace-wrapper.js +9 -2
- package/dist/core/trace/cli-trace-wrapper.js.map +2 -2
- package/dist/core/trace/db-trace-wrapper.js +96 -68
- package/dist/core/trace/db-trace-wrapper.js.map +2 -2
- package/dist/core/trace/debug-trace.js +25 -8
- package/dist/core/trace/debug-trace.js.map +2 -2
- package/dist/core/trace/index.js +6 -2
- package/dist/core/trace/index.js.map +2 -2
- package/dist/core/trace/linear-api-wrapper.js +10 -5
- package/dist/core/trace/linear-api-wrapper.js.map +2 -2
- package/dist/core/trace/trace-demo.js +14 -10
- package/dist/core/trace/trace-demo.js.map +2 -2
- package/dist/core/trace/trace-detector.js +9 -2
- package/dist/core/trace/trace-detector.js.map +2 -2
- package/dist/core/trace/types.js.map +1 -1
- package/dist/core/utils/compression.js.map +1 -1
- package/dist/core/utils/update-checker.js.map +1 -1
- package/dist/core/worktree/worktree-manager.js +18 -7
- package/dist/core/worktree/worktree-manager.js.map +2 -2
- package/dist/features/analytics/core/analytics-service.js.map +2 -2
- package/dist/features/analytics/queries/metrics-queries.js +1 -1
- package/dist/features/analytics/queries/metrics-queries.js.map +2 -2
- package/dist/features/tasks/pebbles-task-store.js.map +1 -1
- package/dist/features/tui/components/analytics-panel.js +36 -15
- package/dist/features/tui/components/analytics-panel.js.map +2 -2
- package/dist/features/tui/components/pr-tracker.js +19 -7
- package/dist/features/tui/components/pr-tracker.js.map +2 -2
- package/dist/features/tui/components/session-monitor.js +22 -9
- package/dist/features/tui/components/session-monitor.js.map +2 -2
- package/dist/features/tui/components/subagent-fleet.js +20 -13
- package/dist/features/tui/components/subagent-fleet.js.map +2 -2
- package/dist/features/tui/components/task-board.js +26 -10
- package/dist/features/tui/components/task-board.js.map +2 -2
- package/dist/features/tui/index.js.map +2 -2
- package/dist/features/tui/services/data-service.js +6 -2
- package/dist/features/tui/services/data-service.js.map +2 -2
- package/dist/features/tui/services/linear-task-reader.js +3 -1
- package/dist/features/tui/services/linear-task-reader.js.map +2 -2
- package/dist/features/tui/services/websocket-client.js +3 -1
- package/dist/features/tui/services/websocket-client.js.map +2 -2
- package/dist/features/tui/terminal-compat.js +6 -2
- package/dist/features/tui/terminal-compat.js.map +2 -2
- package/dist/features/web/client/stores/task-store.js.map +2 -2
- package/dist/features/web/server/index.js +18 -10
- package/dist/features/web/server/index.js.map +2 -2
- package/dist/integrations/anthropic/client.js +259 -0
- package/dist/integrations/anthropic/client.js.map +7 -0
- package/dist/integrations/claude-code/subagent-client.js +404 -0
- package/dist/integrations/claude-code/subagent-client.js.map +7 -0
- package/dist/integrations/linear/sync-service.js +12 -13
- package/dist/integrations/linear/sync-service.js.map +2 -2
- package/dist/integrations/linear/sync.js +174 -12
- package/dist/integrations/linear/sync.js.map +2 -2
- package/dist/integrations/linear/unified-sync.js +1 -1
- package/dist/integrations/linear/unified-sync.js.map +1 -1
- package/dist/integrations/linear/webhook-server.js +15 -16
- package/dist/integrations/linear/webhook-server.js.map +2 -2
- package/dist/mcp/stackmemory-mcp-server.js +0 -11
- package/dist/mcp/stackmemory-mcp-server.js.map +2 -2
- package/dist/servers/production/auth-middleware.js.map +2 -2
- package/dist/servers/railway/index.js.map +2 -2
- package/dist/services/config-service.js +6 -7
- package/dist/services/config-service.js.map +2 -2
- package/dist/services/context-service.js +11 -12
- package/dist/services/context-service.js.map +2 -2
- package/dist/skills/claude-skills.js +101 -2
- package/dist/skills/claude-skills.js.map +2 -2
- package/dist/skills/dashboard-launcher.js.map +2 -2
- package/dist/skills/recursive-agent-orchestrator.js +559 -0
- package/dist/skills/recursive-agent-orchestrator.js.map +7 -0
- package/dist/skills/repo-ingestion-skill.js.map +2 -2
- package/dist/skills/security-secrets-scanner.js +265 -0
- package/dist/skills/security-secrets-scanner.js.map +7 -0
- package/dist/utils/env.js +46 -0
- package/dist/utils/env.js.map +7 -0
- package/dist/utils/logger.js +0 -11
- package/dist/utils/logger.js.map +2 -2
- package/package.json +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/database/batch-operations.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Batch Database Operations\n * High-performance bulk operations with transaction management\n */\n\nimport Database from 'better-sqlite3';\nimport { getConnectionPool } from './connection-pool.js';\nimport { logger } from '../monitoring/logger.js';\nimport { trace } from '../trace/index.js';\n\nexport interface BatchOperation {\n table: string;\n operation: 'insert' | 'update' | 'delete';\n data: Record<string, any>[];\n onConflict?: 'ignore' | 'replace' | 'update';\n}\n\nexport interface BulkInsertOptions {\n batchSize?: number;\n onConflict?: 'ignore' | 'replace' | 'update';\n enableTransactions?: boolean;\n parallelTables?: boolean;\n}\n\nexport interface BatchStats {\n totalRecords: number;\n batchesProcessed: number;\n successfulInserts: number;\n failedInserts: number;\n totalTimeMs: number;\n avgBatchTimeMs: number;\n}\n\n/**\n * High-performance batch operations manager\n */\nexport class BatchOperationsManager {\n private db: Database.Database;\n private preparedStatements = new Map<string, Database.Statement>();\n private batchQueue: BatchOperation[] = [];\n private isProcessing = false;\n\n constructor(db?: Database.Database) {\n if (db) {\n this.db = db;\n this.initializePreparedStatements();\n } else {\n // Will be initialized when used with getConnectionPool().withConnection()\n this.db = undefined as any;\n }\n }\n\n /**\n * Add events in bulk with optimized batching\n */\n async bulkInsertEvents(\n events: Array<{\n frame_id: string;\n run_id: string;\n seq: number;\n event_type: string;\n payload: any;\n ts: number;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n const {\n batchSize = 100,\n onConflict = 'ignore',\n enableTransactions = true,\n } = options;\n\n return this.performBulkInsert('events', events, {\n batchSize,\n onConflict,\n enableTransactions,\n preprocessor: (event) => ({\n ...event,\n event_id: `evt_${event.frame_id}_${event.seq}_${Date.now()}`,\n payload: JSON.stringify(event.payload),\n }),\n });\n }\n\n /**\n * Add anchors in bulk\n */\n async bulkInsertAnchors(\n anchors: Array<{\n frame_id: string;\n type: string;\n text: string;\n priority: number;\n metadata: any;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n return this.performBulkInsert('anchors', anchors, {\n ...options,\n preprocessor: (anchor) => ({\n ...anchor,\n anchor_id: `anc_${anchor.frame_id}_${Date.now()}_${Math.random().toString(36).substr(2, 9)}`,\n metadata: JSON.stringify(anchor.metadata),\n created_at: Date.now(),\n }),\n });\n }\n\n /**\n * Bulk update frame digests\n */\n async bulkUpdateFrameDigests(\n updates: Array<{\n frame_id: string;\n digest_text: string;\n digest_json: any;\n closed_at?: number;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n const {\n batchSize = 50,\n enableTransactions = true,\n } = options;\n\n return trace.traceAsync('function', 'bulkUpdateFrameDigests', { count: updates.length }, async () => {\n const startTime = performance.now();\n const stats: BatchStats = {\n totalRecords: updates.length,\n batchesProcessed: 0,\n successfulInserts: 0,\n failedInserts: 0,\n totalTimeMs: 0,\n avgBatchTimeMs: 0,\n };\n\n if (updates.length === 0) return stats;\n\n const stmt = this.db.prepare(`\n UPDATE frames \n SET digest_text = ?, \n digest_json = ?, \n closed_at = COALESCE(?, closed_at),\n state = CASE WHEN ? IS NOT NULL THEN 'closed' ELSE state END\n WHERE frame_id = ?\n `);\n\n const updateFn = (batch: typeof updates) => {\n for (const update of batch) {\n try {\n const result = stmt.run(\n update.digest_text,\n JSON.stringify(update.digest_json),\n update.closed_at,\n update.closed_at,\n update.frame_id\n );\n stats.successfulInserts += result.changes;\n } catch (error: unknown) {\n stats.failedInserts++;\n logger.warn('Failed to update frame digest', {\n frameId: update.frame_id,\n error: (error as Error).message,\n });\n }\n }\n };\n\n if (enableTransactions) {\n const transaction = this.db.transaction(updateFn);\n await this.processBatches(updates, batchSize, transaction, stats);\n } else {\n await this.processBatches(updates, batchSize, updateFn, stats);\n }\n\n stats.totalTimeMs = performance.now() - startTime;\n stats.avgBatchTimeMs = stats.batchesProcessed > 0 \n ? stats.totalTimeMs / stats.batchesProcessed \n : 0;\n\n logger.info('Bulk frame digest update completed', stats as unknown as Record<string, unknown>);\n return stats;\n });\n }\n\n /**\n * Generic bulk insert with preprocessing\n */\n private async performBulkInsert<T extends Record<string, any>>(\n table: string,\n records: T[],\n options: BulkInsertOptions & {\n preprocessor?: (record: T) => Record<string, any>;\n } = {}\n ): Promise<BatchStats> {\n const {\n batchSize = 100,\n onConflict = 'ignore',\n enableTransactions = true,\n preprocessor,\n } = options;\n\n return trace.traceAsync('function', `bulkInsert${table}`, { count: records.length }, async () => {\n const startTime = performance.now();\n const stats: BatchStats = {\n totalRecords: records.length,\n batchesProcessed: 0,\n successfulInserts: 0,\n failedInserts: 0,\n totalTimeMs: 0,\n avgBatchTimeMs: 0,\n };\n\n if (records.length === 0) return stats;\n\n // Preprocess records if needed\n const processedRecords = preprocessor \n ? records.map(preprocessor)\n : records;\n\n // Build dynamic insert statement\n const firstRecord = processedRecords[0];\n const columns = Object.keys(firstRecord);\n const placeholders = columns.map(() => '?').join(', ');\n const conflictClause = this.getConflictClause(onConflict);\n \n const insertSql = `INSERT ${conflictClause} INTO ${table} (${columns.join(', ')}) VALUES (${placeholders})`;\n const stmt = this.db.prepare(insertSql);\n\n const insertFn = (batch: typeof processedRecords) => {\n for (const record of batch) {\n try {\n const values = columns.map((col: any) => record[col]);\n const result = stmt.run(...values);\n stats.successfulInserts += result.changes;\n } catch (error: unknown) {\n stats.failedInserts++;\n logger.warn(`Failed to insert ${table} record`, {\n record,\n error: (error as Error).message,\n });\n }\n }\n };\n\n if (enableTransactions) {\n const transaction = this.db.transaction(insertFn);\n await this.processBatches(processedRecords, batchSize, transaction, stats);\n } else {\n await this.processBatches(processedRecords, batchSize, insertFn, stats);\n }\n\n stats.totalTimeMs = performance.now() - startTime;\n stats.avgBatchTimeMs = stats.batchesProcessed > 0 \n ? stats.totalTimeMs / stats.batchesProcessed \n : 0;\n\n logger.info(`Bulk ${table} insert completed`, stats as unknown as Record<string, unknown>);\n return stats;\n });\n }\n\n /**\n * Process records in batches\n */\n private async processBatches<T>(\n records: T[],\n batchSize: number,\n processFn: (batch: T[]) => void,\n stats: BatchStats\n ): Promise<void> {\n for (let i = 0; i < records.length; i += batchSize) {\n const batch = records.slice(i, i + batchSize);\n const batchStart = performance.now();\n \n try {\n processFn(batch);\n stats.batchesProcessed++;\n \n const batchTime = performance.now() - batchStart;\n logger.debug('Batch processed', {\n batchNumber: stats.batchesProcessed,\n records: batch.length,\n timeMs: batchTime.toFixed(2),\n });\n\n // Yield control to prevent blocking\n if (stats.batchesProcessed % 10 === 0) {\n await new Promise(resolve => setImmediate(resolve));\n }\n\n } catch (error: unknown) {\n stats.failedInserts += batch.length;\n logger.error('Batch processing failed', error as Error, {\n batchNumber: stats.batchesProcessed + 1,\n batchSize: batch.length,\n });\n }\n }\n }\n\n /**\n * Queue batch operation for later processing\n */\n queueBatchOperation(operation: BatchOperation): void {\n this.batchQueue.push(operation);\n \n if (this.batchQueue.length >= 10 && !this.isProcessing) {\n setImmediate(() => this.processBatchQueue());\n }\n }\n\n /**\n * Process queued batch operations\n */\n async processBatchQueue(): Promise<void> {\n if (this.isProcessing || this.batchQueue.length === 0) {\n return;\n }\n\n this.isProcessing = true;\n const operations = [...this.batchQueue];\n this.batchQueue = [];\n\n try {\n const groupedOps = this.groupOperationsByTable(operations);\n \n for (const [table, tableOps] of groupedOps) {\n await this.processTableOperations(table, tableOps);\n }\n\n logger.info('Batch queue processed', {\n operations: operations.length,\n tables: groupedOps.size,\n });\n\n } catch (error: unknown) {\n logger.error('Batch queue processing failed', error as Error);\n } finally {\n this.isProcessing = false;\n }\n }\n\n /**\n * Flush any remaining queued operations\n */\n async flush(): Promise<void> {\n if (this.batchQueue.length > 0) {\n await this.processBatchQueue();\n }\n }\n\n /**\n * Get SQL conflict clause\n */\n private getConflictClause(onConflict: string): string {\n switch (onConflict) {\n case 'ignore':\n return 'OR IGNORE';\n case 'replace':\n return 'OR REPLACE';\n case 'update':\n return 'ON CONFLICT DO UPDATE SET';\n default:\n return '';\n }\n }\n\n /**\n * Group operations by table for efficient processing\n */\n private groupOperationsByTable(operations: BatchOperation[]): Map<string, BatchOperation[]> {\n const grouped = new Map<string, BatchOperation[]>();\n \n for (const op of operations) {\n if (!grouped.has(op.table)) {\n grouped.set(op.table, []);\n }\n grouped.get(op.table)!.push(op);\n }\n \n return grouped;\n }\n\n /**\n * Process all operations for a specific table\n */\n private async processTableOperations(table: string, operations: BatchOperation[]): Promise<void> {\n for (const op of operations) {\n switch (op.operation) {\n case 'insert':\n await this.performBulkInsert(table, op.data, {\n onConflict: op.onConflict,\n });\n break;\n // Add update and delete operations as needed\n default:\n logger.warn('Unsupported batch operation', { table, operation: op.operation });\n }\n }\n }\n\n /**\n * Initialize commonly used prepared statements\n */\n private initializePreparedStatements(): void {\n // Event insertion\n this.preparedStatements.set('insert_event', \n this.db.prepare(`\n INSERT OR IGNORE INTO events \n (event_id, frame_id, run_id, seq, event_type, payload, ts) \n VALUES (?, ?, ?, ?, ?, ?, ?)\n `)\n );\n\n // Anchor insertion\n this.preparedStatements.set('insert_anchor',\n this.db.prepare(`\n INSERT OR IGNORE INTO anchors \n (anchor_id, frame_id, type, text, priority, metadata, created_at) \n VALUES (?, ?, ?, ?, ?, ?, ?)\n `)\n );\n\n logger.info('Batch operations prepared statements initialized');\n }\n\n /**\n * Cleanup resources\n */\n cleanup(): void {\n // Modern better-sqlite3 automatically handles cleanup\n this.preparedStatements.clear();\n }\n}\n\n// Global batch operations manager\nlet globalBatchManager: BatchOperationsManager | null = null;\n\n/**\n * Get or create global batch operations manager\n */\nexport function getBatchManager(db?: Database.Database): BatchOperationsManager {\n if (!globalBatchManager) {\n globalBatchManager = new BatchOperationsManager(db);\n }\n return globalBatchManager;\n}\n\n/**\n * Convenience function for bulk event insertion\n */\nexport async function bulkInsertEvents(\n events: any[],\n options?: BulkInsertOptions\n): Promise<BatchStats> {\n const manager = getBatchManager();\n return manager.bulkInsertEvents(events, options);\n}\n\n/**\n * Convenience function for bulk anchor insertion\n */\nexport async function bulkInsertAnchors(\n anchors: any[],\n options?: BulkInsertOptions\n): Promise<BatchStats> {\n const manager = getBatchManager();\n return manager.bulkInsertAnchors(anchors, options);\n}"],
|
|
5
|
-
"mappings": "AAOA,SAAS,cAAc;AACvB,SAAS,aAAa;AA4Bf,MAAM,uBAAuB;AAAA,EAC1B;AAAA,EACA,qBAAqB,oBAAI,IAAgC;AAAA,EACzD,aAA+B,CAAC;AAAA,EAChC,eAAe;AAAA,EAEvB,YAAY,IAAwB;AAClC,QAAI,IAAI;AACN,WAAK,KAAK;AACV,WAAK,6BAA6B;AAAA,IACpC,OAAO;AAEL,WAAK,KAAK;AAAA,IACZ;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,iBACJ,QAQA,UAA6B,CAAC,GACT;AACrB,UAAM;AAAA,MACJ,YAAY;AAAA,MACZ,aAAa;AAAA,MACb,qBAAqB;AAAA,IACvB,IAAI;AAEJ,WAAO,KAAK,kBAAkB,UAAU,QAAQ;AAAA,MAC9C;AAAA,MACA;AAAA,MACA;AAAA,MACA,cAAc,CAAC,WAAW;AAAA,QACxB,GAAG;AAAA,QACH,UAAU,OAAO,MAAM,QAAQ,IAAI,MAAM,GAAG,IAAI,KAAK,IAAI,CAAC;AAAA,QAC1D,SAAS,KAAK,UAAU,MAAM,OAAO;AAAA,MACvC;AAAA,IACF,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,kBACJ,SAOA,UAA6B,CAAC,GACT;AACrB,WAAO,KAAK,kBAAkB,WAAW,SAAS;AAAA,MAChD,GAAG;AAAA,MACH,cAAc,CAAC,YAAY;AAAA,QACzB,GAAG;AAAA,QACH,WAAW,OAAO,OAAO,QAAQ,IAAI,KAAK,IAAI,CAAC,IAAI,KAAK,OAAO,EAAE,SAAS,EAAE,EAAE,OAAO,GAAG,CAAC,CAAC;AAAA,QAC1F,UAAU,KAAK,UAAU,OAAO,QAAQ;AAAA,QACxC,YAAY,KAAK,IAAI;AAAA,MACvB;AAAA,IACF,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,uBACJ,SAMA,UAA6B,CAAC,GACT;AACrB,UAAM
|
|
4
|
+
"sourcesContent": ["/**\n * Batch Database Operations\n * High-performance bulk operations with transaction management\n */\n\nimport Database from 'better-sqlite3';\nimport { getConnectionPool } from './connection-pool.js';\nimport { logger } from '../monitoring/logger.js';\nimport { trace } from '../trace/index.js';\n\nexport interface BatchOperation {\n table: string;\n operation: 'insert' | 'update' | 'delete';\n data: Record<string, any>[];\n onConflict?: 'ignore' | 'replace' | 'update';\n}\n\nexport interface BulkInsertOptions {\n batchSize?: number;\n onConflict?: 'ignore' | 'replace' | 'update';\n enableTransactions?: boolean;\n parallelTables?: boolean;\n}\n\nexport interface BatchStats {\n totalRecords: number;\n batchesProcessed: number;\n successfulInserts: number;\n failedInserts: number;\n totalTimeMs: number;\n avgBatchTimeMs: number;\n}\n\n/**\n * High-performance batch operations manager\n */\nexport class BatchOperationsManager {\n private db: Database.Database;\n private preparedStatements = new Map<string, Database.Statement>();\n private batchQueue: BatchOperation[] = [];\n private isProcessing = false;\n\n constructor(db?: Database.Database) {\n if (db) {\n this.db = db;\n this.initializePreparedStatements();\n } else {\n // Will be initialized when used with getConnectionPool().withConnection()\n this.db = undefined as any;\n }\n }\n\n /**\n * Add events in bulk with optimized batching\n */\n async bulkInsertEvents(\n events: Array<{\n frame_id: string;\n run_id: string;\n seq: number;\n event_type: string;\n payload: any;\n ts: number;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n const {\n batchSize = 100,\n onConflict = 'ignore',\n enableTransactions = true,\n } = options;\n\n return this.performBulkInsert('events', events, {\n batchSize,\n onConflict,\n enableTransactions,\n preprocessor: (event) => ({\n ...event,\n event_id: `evt_${event.frame_id}_${event.seq}_${Date.now()}`,\n payload: JSON.stringify(event.payload),\n }),\n });\n }\n\n /**\n * Add anchors in bulk\n */\n async bulkInsertAnchors(\n anchors: Array<{\n frame_id: string;\n type: string;\n text: string;\n priority: number;\n metadata: any;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n return this.performBulkInsert('anchors', anchors, {\n ...options,\n preprocessor: (anchor) => ({\n ...anchor,\n anchor_id: `anc_${anchor.frame_id}_${Date.now()}_${Math.random().toString(36).substr(2, 9)}`,\n metadata: JSON.stringify(anchor.metadata),\n created_at: Date.now(),\n }),\n });\n }\n\n /**\n * Bulk update frame digests\n */\n async bulkUpdateFrameDigests(\n updates: Array<{\n frame_id: string;\n digest_text: string;\n digest_json: any;\n closed_at?: number;\n }>,\n options: BulkInsertOptions = {}\n ): Promise<BatchStats> {\n const { batchSize = 50, enableTransactions = true } = options;\n\n return trace.traceAsync(\n 'function',\n 'bulkUpdateFrameDigests',\n { count: updates.length },\n async () => {\n const startTime = performance.now();\n const stats: BatchStats = {\n totalRecords: updates.length,\n batchesProcessed: 0,\n successfulInserts: 0,\n failedInserts: 0,\n totalTimeMs: 0,\n avgBatchTimeMs: 0,\n };\n\n if (updates.length === 0) return stats;\n\n const stmt = this.db.prepare(`\n UPDATE frames \n SET digest_text = ?, \n digest_json = ?, \n closed_at = COALESCE(?, closed_at),\n state = CASE WHEN ? IS NOT NULL THEN 'closed' ELSE state END\n WHERE frame_id = ?\n `);\n\n const updateFn = (batch: typeof updates) => {\n for (const update of batch) {\n try {\n const result = stmt.run(\n update.digest_text,\n JSON.stringify(update.digest_json),\n update.closed_at,\n update.closed_at,\n update.frame_id\n );\n stats.successfulInserts += result.changes;\n } catch (error: unknown) {\n stats.failedInserts++;\n logger.warn('Failed to update frame digest', {\n frameId: update.frame_id,\n error: (error as Error).message,\n });\n }\n }\n };\n\n if (enableTransactions) {\n const transaction = this.db.transaction(updateFn);\n await this.processBatches(updates, batchSize, transaction, stats);\n } else {\n await this.processBatches(updates, batchSize, updateFn, stats);\n }\n\n stats.totalTimeMs = performance.now() - startTime;\n stats.avgBatchTimeMs =\n stats.batchesProcessed > 0\n ? stats.totalTimeMs / stats.batchesProcessed\n : 0;\n\n logger.info(\n 'Bulk frame digest update completed',\n stats as unknown as Record<string, unknown>\n );\n return stats;\n }\n );\n }\n\n /**\n * Generic bulk insert with preprocessing\n */\n private async performBulkInsert<T extends Record<string, any>>(\n table: string,\n records: T[],\n options: BulkInsertOptions & {\n preprocessor?: (record: T) => Record<string, any>;\n } = {}\n ): Promise<BatchStats> {\n const {\n batchSize = 100,\n onConflict = 'ignore',\n enableTransactions = true,\n preprocessor,\n } = options;\n\n return trace.traceAsync(\n 'function',\n `bulkInsert${table}`,\n { count: records.length },\n async () => {\n const startTime = performance.now();\n const stats: BatchStats = {\n totalRecords: records.length,\n batchesProcessed: 0,\n successfulInserts: 0,\n failedInserts: 0,\n totalTimeMs: 0,\n avgBatchTimeMs: 0,\n };\n\n if (records.length === 0) return stats;\n\n // Preprocess records if needed\n const processedRecords = preprocessor\n ? records.map(preprocessor)\n : records;\n\n // Build dynamic insert statement\n const firstRecord = processedRecords[0];\n const columns = Object.keys(firstRecord);\n const placeholders = columns.map(() => '?').join(', ');\n const conflictClause = this.getConflictClause(onConflict);\n\n const insertSql = `INSERT ${conflictClause} INTO ${table} (${columns.join(', ')}) VALUES (${placeholders})`;\n const stmt = this.db.prepare(insertSql);\n\n const insertFn = (batch: typeof processedRecords) => {\n for (const record of batch) {\n try {\n const values = columns.map((col: any) => record[col]);\n const result = stmt.run(...values);\n stats.successfulInserts += result.changes;\n } catch (error: unknown) {\n stats.failedInserts++;\n logger.warn(`Failed to insert ${table} record`, {\n record,\n error: (error as Error).message,\n });\n }\n }\n };\n\n if (enableTransactions) {\n const transaction = this.db.transaction(insertFn);\n await this.processBatches(\n processedRecords,\n batchSize,\n transaction,\n stats\n );\n } else {\n await this.processBatches(\n processedRecords,\n batchSize,\n insertFn,\n stats\n );\n }\n\n stats.totalTimeMs = performance.now() - startTime;\n stats.avgBatchTimeMs =\n stats.batchesProcessed > 0\n ? stats.totalTimeMs / stats.batchesProcessed\n : 0;\n\n logger.info(\n `Bulk ${table} insert completed`,\n stats as unknown as Record<string, unknown>\n );\n return stats;\n }\n );\n }\n\n /**\n * Process records in batches\n */\n private async processBatches<T>(\n records: T[],\n batchSize: number,\n processFn: (batch: T[]) => void,\n stats: BatchStats\n ): Promise<void> {\n for (let i = 0; i < records.length; i += batchSize) {\n const batch = records.slice(i, i + batchSize);\n const batchStart = performance.now();\n\n try {\n processFn(batch);\n stats.batchesProcessed++;\n\n const batchTime = performance.now() - batchStart;\n logger.debug('Batch processed', {\n batchNumber: stats.batchesProcessed,\n records: batch.length,\n timeMs: batchTime.toFixed(2),\n });\n\n // Yield control to prevent blocking\n if (stats.batchesProcessed % 10 === 0) {\n await new Promise((resolve) => setImmediate(resolve));\n }\n } catch (error: unknown) {\n stats.failedInserts += batch.length;\n logger.error('Batch processing failed', error as Error, {\n batchNumber: stats.batchesProcessed + 1,\n batchSize: batch.length,\n });\n }\n }\n }\n\n /**\n * Queue batch operation for later processing\n */\n queueBatchOperation(operation: BatchOperation): void {\n this.batchQueue.push(operation);\n\n if (this.batchQueue.length >= 10 && !this.isProcessing) {\n setImmediate(() => this.processBatchQueue());\n }\n }\n\n /**\n * Process queued batch operations\n */\n async processBatchQueue(): Promise<void> {\n if (this.isProcessing || this.batchQueue.length === 0) {\n return;\n }\n\n this.isProcessing = true;\n const operations = [...this.batchQueue];\n this.batchQueue = [];\n\n try {\n const groupedOps = this.groupOperationsByTable(operations);\n\n for (const [table, tableOps] of groupedOps) {\n await this.processTableOperations(table, tableOps);\n }\n\n logger.info('Batch queue processed', {\n operations: operations.length,\n tables: groupedOps.size,\n });\n } catch (error: unknown) {\n logger.error('Batch queue processing failed', error as Error);\n } finally {\n this.isProcessing = false;\n }\n }\n\n /**\n * Flush any remaining queued operations\n */\n async flush(): Promise<void> {\n if (this.batchQueue.length > 0) {\n await this.processBatchQueue();\n }\n }\n\n /**\n * Get SQL conflict clause\n */\n private getConflictClause(onConflict: string): string {\n switch (onConflict) {\n case 'ignore':\n return 'OR IGNORE';\n case 'replace':\n return 'OR REPLACE';\n case 'update':\n return 'ON CONFLICT DO UPDATE SET';\n default:\n return '';\n }\n }\n\n /**\n * Group operations by table for efficient processing\n */\n private groupOperationsByTable(\n operations: BatchOperation[]\n ): Map<string, BatchOperation[]> {\n const grouped = new Map<string, BatchOperation[]>();\n\n for (const op of operations) {\n if (!grouped.has(op.table)) {\n grouped.set(op.table, []);\n }\n grouped.get(op.table)!.push(op);\n }\n\n return grouped;\n }\n\n /**\n * Process all operations for a specific table\n */\n private async processTableOperations(\n table: string,\n operations: BatchOperation[]\n ): Promise<void> {\n for (const op of operations) {\n switch (op.operation) {\n case 'insert':\n await this.performBulkInsert(table, op.data, {\n onConflict: op.onConflict,\n });\n break;\n // Add update and delete operations as needed\n default:\n logger.warn('Unsupported batch operation', {\n table,\n operation: op.operation,\n });\n }\n }\n }\n\n /**\n * Initialize commonly used prepared statements\n */\n private initializePreparedStatements(): void {\n // Event insertion\n this.preparedStatements.set(\n 'insert_event',\n this.db.prepare(`\n INSERT OR IGNORE INTO events \n (event_id, frame_id, run_id, seq, event_type, payload, ts) \n VALUES (?, ?, ?, ?, ?, ?, ?)\n `)\n );\n\n // Anchor insertion\n this.preparedStatements.set(\n 'insert_anchor',\n this.db.prepare(`\n INSERT OR IGNORE INTO anchors \n (anchor_id, frame_id, type, text, priority, metadata, created_at) \n VALUES (?, ?, ?, ?, ?, ?, ?)\n `)\n );\n\n logger.info('Batch operations prepared statements initialized');\n }\n\n /**\n * Cleanup resources\n */\n cleanup(): void {\n // Modern better-sqlite3 automatically handles cleanup\n this.preparedStatements.clear();\n }\n}\n\n// Global batch operations manager\nlet globalBatchManager: BatchOperationsManager | null = null;\n\n/**\n * Get or create global batch operations manager\n */\nexport function getBatchManager(\n db?: Database.Database\n): BatchOperationsManager {\n if (!globalBatchManager) {\n globalBatchManager = new BatchOperationsManager(db);\n }\n return globalBatchManager;\n}\n\n/**\n * Convenience function for bulk event insertion\n */\nexport async function bulkInsertEvents(\n events: any[],\n options?: BulkInsertOptions\n): Promise<BatchStats> {\n const manager = getBatchManager();\n return manager.bulkInsertEvents(events, options);\n}\n\n/**\n * Convenience function for bulk anchor insertion\n */\nexport async function bulkInsertAnchors(\n anchors: any[],\n options?: BulkInsertOptions\n): Promise<BatchStats> {\n const manager = getBatchManager();\n return manager.bulkInsertAnchors(anchors, options);\n}\n"],
|
|
5
|
+
"mappings": "AAOA,SAAS,cAAc;AACvB,SAAS,aAAa;AA4Bf,MAAM,uBAAuB;AAAA,EAC1B;AAAA,EACA,qBAAqB,oBAAI,IAAgC;AAAA,EACzD,aAA+B,CAAC;AAAA,EAChC,eAAe;AAAA,EAEvB,YAAY,IAAwB;AAClC,QAAI,IAAI;AACN,WAAK,KAAK;AACV,WAAK,6BAA6B;AAAA,IACpC,OAAO;AAEL,WAAK,KAAK;AAAA,IACZ;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,iBACJ,QAQA,UAA6B,CAAC,GACT;AACrB,UAAM;AAAA,MACJ,YAAY;AAAA,MACZ,aAAa;AAAA,MACb,qBAAqB;AAAA,IACvB,IAAI;AAEJ,WAAO,KAAK,kBAAkB,UAAU,QAAQ;AAAA,MAC9C;AAAA,MACA;AAAA,MACA;AAAA,MACA,cAAc,CAAC,WAAW;AAAA,QACxB,GAAG;AAAA,QACH,UAAU,OAAO,MAAM,QAAQ,IAAI,MAAM,GAAG,IAAI,KAAK,IAAI,CAAC;AAAA,QAC1D,SAAS,KAAK,UAAU,MAAM,OAAO;AAAA,MACvC;AAAA,IACF,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,kBACJ,SAOA,UAA6B,CAAC,GACT;AACrB,WAAO,KAAK,kBAAkB,WAAW,SAAS;AAAA,MAChD,GAAG;AAAA,MACH,cAAc,CAAC,YAAY;AAAA,QACzB,GAAG;AAAA,QACH,WAAW,OAAO,OAAO,QAAQ,IAAI,KAAK,IAAI,CAAC,IAAI,KAAK,OAAO,EAAE,SAAS,EAAE,EAAE,OAAO,GAAG,CAAC,CAAC;AAAA,QAC1F,UAAU,KAAK,UAAU,OAAO,QAAQ;AAAA,QACxC,YAAY,KAAK,IAAI;AAAA,MACvB;AAAA,IACF,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,uBACJ,SAMA,UAA6B,CAAC,GACT;AACrB,UAAM,EAAE,YAAY,IAAI,qBAAqB,KAAK,IAAI;AAEtD,WAAO,MAAM;AAAA,MACX;AAAA,MACA;AAAA,MACA,EAAE,OAAO,QAAQ,OAAO;AAAA,MACxB,YAAY;AACV,cAAM,YAAY,YAAY,IAAI;AAClC,cAAM,QAAoB;AAAA,UACxB,cAAc,QAAQ;AAAA,UACtB,kBAAkB;AAAA,UAClB,mBAAmB;AAAA,UACnB,eAAe;AAAA,UACf,aAAa;AAAA,UACb,gBAAgB;AAAA,QAClB;AAEA,YAAI,QAAQ,WAAW,EAAG,QAAO;AAEjC,cAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO9B;AAEC,cAAM,WAAW,CAAC,UAA0B;AAC1C,qBAAW,UAAU,OAAO;AAC1B,gBAAI;AACF,oBAAM,SAAS,KAAK;AAAA,gBAClB,OAAO;AAAA,gBACP,KAAK,UAAU,OAAO,WAAW;AAAA,gBACjC,OAAO;AAAA,gBACP,OAAO;AAAA,gBACP,OAAO;AAAA,cACT;AACA,oBAAM,qBAAqB,OAAO;AAAA,YACpC,SAAS,OAAgB;AACvB,oBAAM;AACN,qBAAO,KAAK,iCAAiC;AAAA,gBAC3C,SAAS,OAAO;AAAA,gBAChB,OAAQ,MAAgB;AAAA,cAC1B,CAAC;AAAA,YACH;AAAA,UACF;AAAA,QACF;AAEA,YAAI,oBAAoB;AACtB,gBAAM,cAAc,KAAK,GAAG,YAAY,QAAQ;AAChD,gBAAM,KAAK,eAAe,SAAS,WAAW,aAAa,KAAK;AAAA,QAClE,OAAO;AACL,gBAAM,KAAK,eAAe,SAAS,WAAW,UAAU,KAAK;AAAA,QAC/D;AAEA,cAAM,cAAc,YAAY,IAAI,IAAI;AACxC,cAAM,iBACJ,MAAM,mBAAmB,IACrB,MAAM,cAAc,MAAM,mBAC1B;AAEN,eAAO;AAAA,UACL;AAAA,UACA;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,kBACZ,OACA,SACA,UAEI,CAAC,GACgB;AACrB,UAAM;AAAA,MACJ,YAAY;AAAA,MACZ,aAAa;AAAA,MACb,qBAAqB;AAAA,MACrB;AAAA,IACF,IAAI;AAEJ,WAAO,MAAM;AAAA,MACX;AAAA,MACA,aAAa,KAAK;AAAA,MAClB,EAAE,OAAO,QAAQ,OAAO;AAAA,MACxB,YAAY;AACV,cAAM,YAAY,YAAY,IAAI;AAClC,cAAM,QAAoB;AAAA,UACxB,cAAc,QAAQ;AAAA,UACtB,kBAAkB;AAAA,UAClB,mBAAmB;AAAA,UACnB,eAAe;AAAA,UACf,aAAa;AAAA,UACb,gBAAgB;AAAA,QAClB;AAEA,YAAI,QAAQ,WAAW,EAAG,QAAO;AAGjC,cAAM,mBAAmB,eACrB,QAAQ,IAAI,YAAY,IACxB;AAGJ,cAAM,cAAc,iBAAiB,CAAC;AACtC,cAAM,UAAU,OAAO,KAAK,WAAW;AACvC,cAAM,eAAe,QAAQ,IAAI,MAAM,GAAG,EAAE,KAAK,IAAI;AACrD,cAAM,iBAAiB,KAAK,kBAAkB,UAAU;AAExD,cAAM,YAAY,UAAU,cAAc,SAAS,KAAK,KAAK,QAAQ,KAAK,IAAI,CAAC,aAAa,YAAY;AACxG,cAAM,OAAO,KAAK,GAAG,QAAQ,SAAS;AAEtC,cAAM,WAAW,CAAC,UAAmC;AACnD,qBAAW,UAAU,OAAO;AAC1B,gBAAI;AACF,oBAAM,SAAS,QAAQ,IAAI,CAAC,QAAa,OAAO,GAAG,CAAC;AACpD,oBAAM,SAAS,KAAK,IAAI,GAAG,MAAM;AACjC,oBAAM,qBAAqB,OAAO;AAAA,YACpC,SAAS,OAAgB;AACvB,oBAAM;AACN,qBAAO,KAAK,oBAAoB,KAAK,WAAW;AAAA,gBAC9C;AAAA,gBACA,OAAQ,MAAgB;AAAA,cAC1B,CAAC;AAAA,YACH;AAAA,UACF;AAAA,QACF;AAEA,YAAI,oBAAoB;AACtB,gBAAM,cAAc,KAAK,GAAG,YAAY,QAAQ;AAChD,gBAAM,KAAK;AAAA,YACT;AAAA,YACA;AAAA,YACA;AAAA,YACA;AAAA,UACF;AAAA,QACF,OAAO;AACL,gBAAM,KAAK;AAAA,YACT;AAAA,YACA;AAAA,YACA;AAAA,YACA;AAAA,UACF;AAAA,QACF;AAEA,cAAM,cAAc,YAAY,IAAI,IAAI;AACxC,cAAM,iBACJ,MAAM,mBAAmB,IACrB,MAAM,cAAc,MAAM,mBAC1B;AAEN,eAAO;AAAA,UACL,QAAQ,KAAK;AAAA,UACb;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,eACZ,SACA,WACA,WACA,OACe;AACf,aAAS,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK,WAAW;AAClD,YAAM,QAAQ,QAAQ,MAAM,GAAG,IAAI,SAAS;AAC5C,YAAM,aAAa,YAAY,IAAI;AAEnC,UAAI;AACF,kBAAU,KAAK;AACf,cAAM;AAEN,cAAM,YAAY,YAAY,IAAI,IAAI;AACtC,eAAO,MAAM,mBAAmB;AAAA,UAC9B,aAAa,MAAM;AAAA,UACnB,SAAS,MAAM;AAAA,UACf,QAAQ,UAAU,QAAQ,CAAC;AAAA,QAC7B,CAAC;AAGD,YAAI,MAAM,mBAAmB,OAAO,GAAG;AACrC,gBAAM,IAAI,QAAQ,CAAC,YAAY,aAAa,OAAO,CAAC;AAAA,QACtD;AAAA,MACF,SAAS,OAAgB;AACvB,cAAM,iBAAiB,MAAM;AAC7B,eAAO,MAAM,2BAA2B,OAAgB;AAAA,UACtD,aAAa,MAAM,mBAAmB;AAAA,UACtC,WAAW,MAAM;AAAA,QACnB,CAAC;AAAA,MACH;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,oBAAoB,WAAiC;AACnD,SAAK,WAAW,KAAK,SAAS;AAE9B,QAAI,KAAK,WAAW,UAAU,MAAM,CAAC,KAAK,cAAc;AACtD,mBAAa,MAAM,KAAK,kBAAkB,CAAC;AAAA,IAC7C;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,oBAAmC;AACvC,QAAI,KAAK,gBAAgB,KAAK,WAAW,WAAW,GAAG;AACrD;AAAA,IACF;AAEA,SAAK,eAAe;AACpB,UAAM,aAAa,CAAC,GAAG,KAAK,UAAU;AACtC,SAAK,aAAa,CAAC;AAEnB,QAAI;AACF,YAAM,aAAa,KAAK,uBAAuB,UAAU;AAEzD,iBAAW,CAAC,OAAO,QAAQ,KAAK,YAAY;AAC1C,cAAM,KAAK,uBAAuB,OAAO,QAAQ;AAAA,MACnD;AAEA,aAAO,KAAK,yBAAyB;AAAA,QACnC,YAAY,WAAW;AAAA,QACvB,QAAQ,WAAW;AAAA,MACrB,CAAC;AAAA,IACH,SAAS,OAAgB;AACvB,aAAO,MAAM,iCAAiC,KAAc;AAAA,IAC9D,UAAE;AACA,WAAK,eAAe;AAAA,IACtB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,QAAuB;AAC3B,QAAI,KAAK,WAAW,SAAS,GAAG;AAC9B,YAAM,KAAK,kBAAkB;AAAA,IAC/B;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,YAA4B;AACpD,YAAQ,YAAY;AAAA,MAClB,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT;AACE,eAAO;AAAA,IACX;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,uBACN,YAC+B;AAC/B,UAAM,UAAU,oBAAI,IAA8B;AAElD,eAAW,MAAM,YAAY;AAC3B,UAAI,CAAC,QAAQ,IAAI,GAAG,KAAK,GAAG;AAC1B,gBAAQ,IAAI,GAAG,OAAO,CAAC,CAAC;AAAA,MAC1B;AACA,cAAQ,IAAI,GAAG,KAAK,EAAG,KAAK,EAAE;AAAA,IAChC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,uBACZ,OACA,YACe;AACf,eAAW,MAAM,YAAY;AAC3B,cAAQ,GAAG,WAAW;AAAA,QACpB,KAAK;AACH,gBAAM,KAAK,kBAAkB,OAAO,GAAG,MAAM;AAAA,YAC3C,YAAY,GAAG;AAAA,UACjB,CAAC;AACD;AAAA;AAAA,QAEF;AACE,iBAAO,KAAK,+BAA+B;AAAA,YACzC;AAAA,YACA,WAAW,GAAG;AAAA,UAChB,CAAC;AAAA,MACL;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,+BAAqC;AAE3C,SAAK,mBAAmB;AAAA,MACtB;AAAA,MACA,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAIf;AAAA,IACH;AAGA,SAAK,mBAAmB;AAAA,MACtB;AAAA,MACA,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAIf;AAAA,IACH;AAEA,WAAO,KAAK,kDAAkD;AAAA,EAChE;AAAA;AAAA;AAAA;AAAA,EAKA,UAAgB;AAEd,SAAK,mBAAmB,MAAM;AAAA,EAChC;AACF;AAGA,IAAI,qBAAoD;AAKjD,SAAS,gBACd,IACwB;AACxB,MAAI,CAAC,oBAAoB;AACvB,yBAAqB,IAAI,uBAAuB,EAAE;AAAA,EACpD;AACA,SAAO;AACT;AAKA,eAAsB,iBACpB,QACA,SACqB;AACrB,QAAM,UAAU,gBAAgB;AAChC,SAAO,QAAQ,iBAAiB,QAAQ,OAAO;AACjD;AAKA,eAAsB,kBACpB,SACA,SACqB;AACrB,QAAM,UAAU,gBAAgB;AAChC,SAAO,QAAQ,kBAAkB,SAAS,OAAO;AACnD;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -113,7 +113,10 @@ class LRUQueryCache {
|
|
|
113
113
|
count++;
|
|
114
114
|
}
|
|
115
115
|
}
|
|
116
|
-
logger.info("Pattern invalidation", {
|
|
116
|
+
logger.info("Pattern invalidation", {
|
|
117
|
+
pattern: pattern.source,
|
|
118
|
+
invalidated: count
|
|
119
|
+
});
|
|
117
120
|
return count;
|
|
118
121
|
}
|
|
119
122
|
/**
|
|
@@ -131,7 +134,10 @@ class LRUQueryCache {
|
|
|
131
134
|
* Get cache contents for debugging
|
|
132
135
|
*/
|
|
133
136
|
debug() {
|
|
134
|
-
return Array.from(this.cache.entries()).map(([key, entry]) => ({
|
|
137
|
+
return Array.from(this.cache.entries()).map(([key, entry]) => ({
|
|
138
|
+
key,
|
|
139
|
+
entry
|
|
140
|
+
}));
|
|
135
141
|
}
|
|
136
142
|
/**
|
|
137
143
|
* Cleanup expired entries
|
|
@@ -149,7 +155,10 @@ class LRUQueryCache {
|
|
|
149
155
|
}
|
|
150
156
|
}
|
|
151
157
|
if (removed > 0) {
|
|
152
|
-
logger.info("Cache cleanup completed", {
|
|
158
|
+
logger.info("Cache cleanup completed", {
|
|
159
|
+
removed,
|
|
160
|
+
remaining: this.cache.size
|
|
161
|
+
});
|
|
153
162
|
}
|
|
154
163
|
return removed;
|
|
155
164
|
}
|
|
@@ -158,7 +167,7 @@ class StackMemoryQueryCache {
|
|
|
158
167
|
frameCache = new LRUQueryCache({ maxSize: 500, ttlMs: 3e5 });
|
|
159
168
|
// 5 min
|
|
160
169
|
eventCache = new LRUQueryCache({ maxSize: 1e3, ttlMs: 18e4 });
|
|
161
|
-
// 3 min
|
|
170
|
+
// 3 min
|
|
162
171
|
anchorCache = new LRUQueryCache({ maxSize: 200, ttlMs: 6e5 });
|
|
163
172
|
// 10 min
|
|
164
173
|
digestCache = new LRUQueryCache({ maxSize: 100, ttlMs: 9e5 });
|
|
@@ -220,7 +229,7 @@ class StackMemoryQueryCache {
|
|
|
220
229
|
logger.info("Invalidated frame caches", { frameId });
|
|
221
230
|
}
|
|
222
231
|
/**
|
|
223
|
-
* Invalidate all caches for a project
|
|
232
|
+
* Invalidate all caches for a project
|
|
224
233
|
*/
|
|
225
234
|
invalidateProject(projectId) {
|
|
226
235
|
const pattern = new RegExp(`^(frame|context|events|anchors|digest):.+`);
|
|
@@ -229,7 +238,10 @@ class StackMemoryQueryCache {
|
|
|
229
238
|
total += this.eventCache.invalidatePattern(pattern);
|
|
230
239
|
total += this.anchorCache.invalidatePattern(pattern);
|
|
231
240
|
total += this.digestCache.invalidatePattern(pattern);
|
|
232
|
-
logger.info("Invalidated project caches", {
|
|
241
|
+
logger.info("Invalidated project caches", {
|
|
242
|
+
projectId,
|
|
243
|
+
totalInvalidated: total
|
|
244
|
+
});
|
|
233
245
|
}
|
|
234
246
|
/**
|
|
235
247
|
* Get comprehensive cache metrics
|
|
@@ -270,9 +282,7 @@ function getQueryCache() {
|
|
|
270
282
|
return globalCache;
|
|
271
283
|
}
|
|
272
284
|
function createCacheKey(queryName, params) {
|
|
273
|
-
const paramsStr = params.map(
|
|
274
|
-
(p) => typeof p === "object" ? JSON.stringify(p) : String(p)
|
|
275
|
-
).join(":");
|
|
285
|
+
const paramsStr = params.map((p) => typeof p === "object" ? JSON.stringify(p) : String(p)).join(":");
|
|
276
286
|
return `${queryName}:${paramsStr}`;
|
|
277
287
|
}
|
|
278
288
|
export {
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/database/query-cache.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Query Result Cache\n * LRU cache for frequently accessed database query results\n */\n\nimport { logger } from '../monitoring/logger.js';\n\nexport interface CacheOptions {\n maxSize?: number;\n ttlMs?: number;\n enableMetrics?: boolean;\n}\n\nexport interface CacheEntry<T> {\n value: T;\n createdAt: number;\n accessCount: number;\n lastAccessed: number;\n}\n\nexport interface CacheMetrics {\n hits: number;\n misses: number;\n evictions: number;\n totalQueries: number;\n hitRate: number;\n size: number;\n maxSize: number;\n}\n\nexport class LRUQueryCache<T = any> {\n private cache = new Map<string, CacheEntry<T>>();\n private maxSize: number;\n private ttlMs: number;\n private enableMetrics: boolean;\n
|
|
5
|
-
"mappings": "AAKA,SAAS,cAAc;AAyBhB,MAAM,cAAuB;AAAA,EAC1B,QAAQ,oBAAI,IAA2B;AAAA,EACvC;AAAA,EACA;AAAA,EACA;AAAA;AAAA,EAGA,UAA8D;AAAA,IACpE,MAAM;AAAA,IACN,QAAQ;AAAA,IACR,WAAW;AAAA,IACX,cAAc;AAAA,EAChB;AAAA,EAEA,YAAY,UAAwB,CAAC,GAAG;AACtC,SAAK,UAAU,QAAQ,WAAW;AAClC,SAAK,QAAQ,QAAQ,SAAS;AAC9B,SAAK,gBAAgB,QAAQ,iBAAiB;AAE9C,WAAO,KAAK,2BAA2B;AAAA,MACrC,SAAS,KAAK;AAAA,MACd,OAAO,KAAK;AAAA,MACZ,eAAe,KAAK;AAAA,IACtB,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,IAAI,KAA4B;AAC9B,QAAI,KAAK,eAAe;AACtB,WAAK,QAAQ;AAAA,IACf;AAEA,UAAM,QAAQ,KAAK,MAAM,IAAI,GAAG;AAEhC,QAAI,CAAC,OAAO;AACV,UAAI,KAAK,eAAe;AACtB,aAAK,QAAQ;AAAA,MACf;AACA,aAAO;AAAA,IACT;AAGA,UAAM,MAAM,KAAK,IAAI;AACrB,QAAI,MAAM,MAAM,YAAY,KAAK,OAAO;AACtC,WAAK,MAAM,OAAO,GAAG;AACrB,UAAI,KAAK,eAAe;AACtB,aAAK,QAAQ;AACb,aAAK,QAAQ;AAAA,MACf;AACA,aAAO,MAAM,uBAAuB,EAAE,KAAK,KAAK,MAAM,MAAM,UAAU,CAAC;AACvE,aAAO;AAAA,IACT;AAGA,UAAM;AACN,UAAM,eAAe;AACrB,SAAK,MAAM,OAAO,GAAG;AACrB,SAAK,MAAM,IAAI,KAAK,KAAK;AAEzB,QAAI,KAAK,eAAe;AACtB,WAAK,QAAQ;AAAA,IACf;AAEA,WAAO,MAAM,aAAa,EAAE,KAAK,aAAa,MAAM,YAAY,CAAC;AACjE,WAAO,MAAM;AAAA,EACf;AAAA;AAAA;AAAA;AAAA,EAKA,IAAI,KAAa,OAAgB;AAC/B,UAAM,MAAM,KAAK,IAAI;AAGrB,QAAI,KAAK,MAAM,IAAI,GAAG,GAAG;AACvB,WAAK,MAAM,OAAO,GAAG;AAAA,IACvB;AAGA,WAAO,KAAK,MAAM,QAAQ,KAAK,SAAS;AACtC,YAAM,WAAW,KAAK,MAAM,KAAK,EAAE,KAAK,EAAE;AAC1C,UAAI,aAAa,QAAW;AAC1B,aAAK,MAAM,OAAO,QAAQ;AAC1B,YAAI,KAAK,eAAe;AACtB,eAAK,QAAQ;AAAA,QACf;AACA,eAAO,MAAM,qBAAqB,EAAE,KAAK,SAAS,CAAC;AAAA,MACrD,OAAO;AACL;AAAA,MACF;AAAA,IACF;AAGA,UAAM,QAAuB;AAAA,MAC3B;AAAA,MACA,WAAW;AAAA,MACX,aAAa;AAAA,MACb,cAAc;AAAA,IAChB;AAEA,SAAK,MAAM,IAAI,KAAK,KAAK;AACzB,WAAO,MAAM,aAAa,EAAE,KAAK,MAAM,KAAK,MAAM,KAAK,CAAC;AAAA,EAC1D;AAAA;AAAA;AAAA;AAAA,EAKA,OAAO,KAAsB;AAC3B,UAAM,UAAU,KAAK,MAAM,OAAO,GAAG;AACrC,QAAI,SAAS;AACX,aAAO,MAAM,gBAAgB,EAAE,IAAI,CAAC;AAAA,IACtC;AACA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,QAAc;AACZ,UAAM,OAAO,KAAK,MAAM;AACxB,SAAK,MAAM,MAAM;AACjB,WAAO,KAAK,iBAAiB,EAAE,cAAc,KAAK,CAAC;AAAA,EACrD;AAAA;AAAA;AAAA;AAAA,EAKA,kBAAkB,SAAyB;AACzC,QAAI,QAAQ;AACZ,eAAW,OAAO,KAAK,MAAM,KAAK,GAAG;AACnC,UAAI,QAAQ,KAAK,GAAG,GAAG;AACrB,aAAK,MAAM,OAAO,GAAG;AACrB;AAAA,MACF;AAAA,IACF;AACA,WAAO,KAAK,wBAAwB,
|
|
4
|
+
"sourcesContent": ["/**\n * Query Result Cache\n * LRU cache for frequently accessed database query results\n */\n\nimport { logger } from '../monitoring/logger.js';\n\nexport interface CacheOptions {\n maxSize?: number;\n ttlMs?: number;\n enableMetrics?: boolean;\n}\n\nexport interface CacheEntry<T> {\n value: T;\n createdAt: number;\n accessCount: number;\n lastAccessed: number;\n}\n\nexport interface CacheMetrics {\n hits: number;\n misses: number;\n evictions: number;\n totalQueries: number;\n hitRate: number;\n size: number;\n maxSize: number;\n}\n\nexport class LRUQueryCache<T = any> {\n private cache = new Map<string, CacheEntry<T>>();\n private maxSize: number;\n private ttlMs: number;\n private enableMetrics: boolean;\n\n // Metrics\n private metrics: Omit<CacheMetrics, 'hitRate' | 'size' | 'maxSize'> = {\n hits: 0,\n misses: 0,\n evictions: 0,\n totalQueries: 0,\n };\n\n constructor(options: CacheOptions = {}) {\n this.maxSize = options.maxSize ?? 1000;\n this.ttlMs = options.ttlMs ?? 300000; // 5 minutes default\n this.enableMetrics = options.enableMetrics ?? true;\n\n logger.info('Query cache initialized', {\n maxSize: this.maxSize,\n ttlMs: this.ttlMs,\n enableMetrics: this.enableMetrics,\n });\n }\n\n /**\n * Get a value from cache\n */\n get(key: string): T | undefined {\n if (this.enableMetrics) {\n this.metrics.totalQueries++;\n }\n\n const entry = this.cache.get(key);\n\n if (!entry) {\n if (this.enableMetrics) {\n this.metrics.misses++;\n }\n return undefined;\n }\n\n // Check TTL\n const now = Date.now();\n if (now - entry.createdAt > this.ttlMs) {\n this.cache.delete(key);\n if (this.enableMetrics) {\n this.metrics.misses++;\n this.metrics.evictions++;\n }\n logger.debug('Cache entry expired', { key, age: now - entry.createdAt });\n return undefined;\n }\n\n // Update access stats and move to end (most recently used)\n entry.accessCount++;\n entry.lastAccessed = now;\n this.cache.delete(key);\n this.cache.set(key, entry);\n\n if (this.enableMetrics) {\n this.metrics.hits++;\n }\n\n logger.debug('Cache hit', { key, accessCount: entry.accessCount });\n return entry.value;\n }\n\n /**\n * Set a value in cache\n */\n set(key: string, value: T): void {\n const now = Date.now();\n\n // Remove existing entry if present\n if (this.cache.has(key)) {\n this.cache.delete(key);\n }\n\n // Evict least recently used entries if at capacity\n while (this.cache.size >= this.maxSize) {\n const firstKey = this.cache.keys().next().value;\n if (firstKey !== undefined) {\n this.cache.delete(firstKey);\n if (this.enableMetrics) {\n this.metrics.evictions++;\n }\n logger.debug('Evicted LRU entry', { key: firstKey });\n } else {\n break;\n }\n }\n\n // Add new entry\n const entry: CacheEntry<T> = {\n value,\n createdAt: now,\n accessCount: 0,\n lastAccessed: now,\n };\n\n this.cache.set(key, entry);\n logger.debug('Cache set', { key, size: this.cache.size });\n }\n\n /**\n * Delete a specific key\n */\n delete(key: string): boolean {\n const deleted = this.cache.delete(key);\n if (deleted) {\n logger.debug('Cache delete', { key });\n }\n return deleted;\n }\n\n /**\n * Clear all cached entries\n */\n clear(): void {\n const size = this.cache.size;\n this.cache.clear();\n logger.info('Cache cleared', { previousSize: size });\n }\n\n /**\n * Invalidate entries matching a pattern\n */\n invalidatePattern(pattern: RegExp): number {\n let count = 0;\n for (const key of this.cache.keys()) {\n if (pattern.test(key)) {\n this.cache.delete(key);\n count++;\n }\n }\n logger.info('Pattern invalidation', {\n pattern: pattern.source,\n invalidated: count,\n });\n return count;\n }\n\n /**\n * Get cache metrics\n */\n getMetrics(): CacheMetrics {\n return {\n ...this.metrics,\n hitRate:\n this.metrics.totalQueries > 0\n ? this.metrics.hits / this.metrics.totalQueries\n : 0,\n size: this.cache.size,\n maxSize: this.maxSize,\n };\n }\n\n /**\n * Get cache contents for debugging\n */\n debug(): Array<{ key: string; entry: CacheEntry<T> }> {\n return Array.from(this.cache.entries()).map(([key, entry]) => ({\n key,\n entry,\n }));\n }\n\n /**\n * Cleanup expired entries\n */\n cleanup(): number {\n const now = Date.now();\n let removed = 0;\n\n for (const [key, entry] of this.cache.entries()) {\n if (now - entry.createdAt > this.ttlMs) {\n this.cache.delete(key);\n removed++;\n if (this.enableMetrics) {\n this.metrics.evictions++;\n }\n }\n }\n\n if (removed > 0) {\n logger.info('Cache cleanup completed', {\n removed,\n remaining: this.cache.size,\n });\n }\n\n return removed;\n }\n}\n\n/**\n * Query cache specifically designed for StackMemory operations\n */\nexport class StackMemoryQueryCache {\n private frameCache = new LRUQueryCache<any>({ maxSize: 500, ttlMs: 300000 }); // 5 min\n private eventCache = new LRUQueryCache<any>({ maxSize: 1000, ttlMs: 180000 }); // 3 min\n private anchorCache = new LRUQueryCache<any>({ maxSize: 200, ttlMs: 600000 }); // 10 min\n private digestCache = new LRUQueryCache<any>({ maxSize: 100, ttlMs: 900000 }); // 15 min\n\n /**\n * Cache frame data\n */\n cacheFrame(frameId: string, data: any): void {\n this.frameCache.set(`frame:${frameId}`, data);\n }\n\n getFrame(frameId: string): any {\n return this.frameCache.get(`frame:${frameId}`);\n }\n\n /**\n * Cache frame context assemblies (expensive operations)\n */\n cacheFrameContext(frameId: string, context: any): void {\n this.frameCache.set(`context:${frameId}`, context);\n }\n\n getFrameContext(frameId: string): any {\n return this.frameCache.get(`context:${frameId}`);\n }\n\n /**\n * Cache events for a frame\n */\n cacheFrameEvents(frameId: string, events: any[]): void {\n this.eventCache.set(`events:${frameId}`, events);\n }\n\n getFrameEvents(frameId: string): any[] {\n return this.eventCache.get(`events:${frameId}`);\n }\n\n /**\n * Cache anchors\n */\n cacheAnchors(frameId: string, anchors: any[]): void {\n this.anchorCache.set(`anchors:${frameId}`, anchors);\n }\n\n getAnchors(frameId: string): any[] {\n return this.anchorCache.get(`anchors:${frameId}`);\n }\n\n /**\n * Cache digest data\n */\n cacheDigest(frameId: string, digest: any): void {\n this.digestCache.set(`digest:${frameId}`, digest);\n }\n\n getDigest(frameId: string): any {\n return this.digestCache.get(`digest:${frameId}`);\n }\n\n /**\n * Invalidate caches for a specific frame\n */\n invalidateFrame(frameId: string): void {\n this.frameCache.delete(`frame:${frameId}`);\n this.frameCache.delete(`context:${frameId}`);\n this.eventCache.delete(`events:${frameId}`);\n this.anchorCache.delete(`anchors:${frameId}`);\n this.digestCache.delete(`digest:${frameId}`);\n\n logger.info('Invalidated frame caches', { frameId });\n }\n\n /**\n * Invalidate all caches for a project\n */\n invalidateProject(projectId: string): void {\n const pattern = new RegExp(`^(frame|context|events|anchors|digest):.+`);\n\n let total = 0;\n total += this.frameCache.invalidatePattern(pattern);\n total += this.eventCache.invalidatePattern(pattern);\n total += this.anchorCache.invalidatePattern(pattern);\n total += this.digestCache.invalidatePattern(pattern);\n\n logger.info('Invalidated project caches', {\n projectId,\n totalInvalidated: total,\n });\n }\n\n /**\n * Get comprehensive cache metrics\n */\n getMetrics() {\n return {\n frame: this.frameCache.getMetrics(),\n event: this.eventCache.getMetrics(),\n anchor: this.anchorCache.getMetrics(),\n digest: this.digestCache.getMetrics(),\n };\n }\n\n /**\n * Cleanup all caches\n */\n cleanup(): void {\n this.frameCache.cleanup();\n this.eventCache.cleanup();\n this.anchorCache.cleanup();\n this.digestCache.cleanup();\n }\n\n /**\n * Clear all caches\n */\n clear(): void {\n this.frameCache.clear();\n this.eventCache.clear();\n this.anchorCache.clear();\n this.digestCache.clear();\n\n logger.info('All StackMemory caches cleared');\n }\n}\n\n// Global cache instance\nlet globalCache: StackMemoryQueryCache | null = null;\n\n/**\n * Get or create global query cache\n */\nexport function getQueryCache(): StackMemoryQueryCache {\n if (!globalCache) {\n globalCache = new StackMemoryQueryCache();\n }\n return globalCache;\n}\n\n/**\n * Create a cache key from query parameters\n */\nexport function createCacheKey(queryName: string, params: any[]): string {\n const paramsStr = params\n .map((p: any) => (typeof p === 'object' ? JSON.stringify(p) : String(p)))\n .join(':');\n\n return `${queryName}:${paramsStr}`;\n}\n"],
|
|
5
|
+
"mappings": "AAKA,SAAS,cAAc;AAyBhB,MAAM,cAAuB;AAAA,EAC1B,QAAQ,oBAAI,IAA2B;AAAA,EACvC;AAAA,EACA;AAAA,EACA;AAAA;AAAA,EAGA,UAA8D;AAAA,IACpE,MAAM;AAAA,IACN,QAAQ;AAAA,IACR,WAAW;AAAA,IACX,cAAc;AAAA,EAChB;AAAA,EAEA,YAAY,UAAwB,CAAC,GAAG;AACtC,SAAK,UAAU,QAAQ,WAAW;AAClC,SAAK,QAAQ,QAAQ,SAAS;AAC9B,SAAK,gBAAgB,QAAQ,iBAAiB;AAE9C,WAAO,KAAK,2BAA2B;AAAA,MACrC,SAAS,KAAK;AAAA,MACd,OAAO,KAAK;AAAA,MACZ,eAAe,KAAK;AAAA,IACtB,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,IAAI,KAA4B;AAC9B,QAAI,KAAK,eAAe;AACtB,WAAK,QAAQ;AAAA,IACf;AAEA,UAAM,QAAQ,KAAK,MAAM,IAAI,GAAG;AAEhC,QAAI,CAAC,OAAO;AACV,UAAI,KAAK,eAAe;AACtB,aAAK,QAAQ;AAAA,MACf;AACA,aAAO;AAAA,IACT;AAGA,UAAM,MAAM,KAAK,IAAI;AACrB,QAAI,MAAM,MAAM,YAAY,KAAK,OAAO;AACtC,WAAK,MAAM,OAAO,GAAG;AACrB,UAAI,KAAK,eAAe;AACtB,aAAK,QAAQ;AACb,aAAK,QAAQ;AAAA,MACf;AACA,aAAO,MAAM,uBAAuB,EAAE,KAAK,KAAK,MAAM,MAAM,UAAU,CAAC;AACvE,aAAO;AAAA,IACT;AAGA,UAAM;AACN,UAAM,eAAe;AACrB,SAAK,MAAM,OAAO,GAAG;AACrB,SAAK,MAAM,IAAI,KAAK,KAAK;AAEzB,QAAI,KAAK,eAAe;AACtB,WAAK,QAAQ;AAAA,IACf;AAEA,WAAO,MAAM,aAAa,EAAE,KAAK,aAAa,MAAM,YAAY,CAAC;AACjE,WAAO,MAAM;AAAA,EACf;AAAA;AAAA;AAAA;AAAA,EAKA,IAAI,KAAa,OAAgB;AAC/B,UAAM,MAAM,KAAK,IAAI;AAGrB,QAAI,KAAK,MAAM,IAAI,GAAG,GAAG;AACvB,WAAK,MAAM,OAAO,GAAG;AAAA,IACvB;AAGA,WAAO,KAAK,MAAM,QAAQ,KAAK,SAAS;AACtC,YAAM,WAAW,KAAK,MAAM,KAAK,EAAE,KAAK,EAAE;AAC1C,UAAI,aAAa,QAAW;AAC1B,aAAK,MAAM,OAAO,QAAQ;AAC1B,YAAI,KAAK,eAAe;AACtB,eAAK,QAAQ;AAAA,QACf;AACA,eAAO,MAAM,qBAAqB,EAAE,KAAK,SAAS,CAAC;AAAA,MACrD,OAAO;AACL;AAAA,MACF;AAAA,IACF;AAGA,UAAM,QAAuB;AAAA,MAC3B;AAAA,MACA,WAAW;AAAA,MACX,aAAa;AAAA,MACb,cAAc;AAAA,IAChB;AAEA,SAAK,MAAM,IAAI,KAAK,KAAK;AACzB,WAAO,MAAM,aAAa,EAAE,KAAK,MAAM,KAAK,MAAM,KAAK,CAAC;AAAA,EAC1D;AAAA;AAAA;AAAA;AAAA,EAKA,OAAO,KAAsB;AAC3B,UAAM,UAAU,KAAK,MAAM,OAAO,GAAG;AACrC,QAAI,SAAS;AACX,aAAO,MAAM,gBAAgB,EAAE,IAAI,CAAC;AAAA,IACtC;AACA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,QAAc;AACZ,UAAM,OAAO,KAAK,MAAM;AACxB,SAAK,MAAM,MAAM;AACjB,WAAO,KAAK,iBAAiB,EAAE,cAAc,KAAK,CAAC;AAAA,EACrD;AAAA;AAAA;AAAA;AAAA,EAKA,kBAAkB,SAAyB;AACzC,QAAI,QAAQ;AACZ,eAAW,OAAO,KAAK,MAAM,KAAK,GAAG;AACnC,UAAI,QAAQ,KAAK,GAAG,GAAG;AACrB,aAAK,MAAM,OAAO,GAAG;AACrB;AAAA,MACF;AAAA,IACF;AACA,WAAO,KAAK,wBAAwB;AAAA,MAClC,SAAS,QAAQ;AAAA,MACjB,aAAa;AAAA,IACf,CAAC;AACD,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,aAA2B;AACzB,WAAO;AAAA,MACL,GAAG,KAAK;AAAA,MACR,SACE,KAAK,QAAQ,eAAe,IACxB,KAAK,QAAQ,OAAO,KAAK,QAAQ,eACjC;AAAA,MACN,MAAM,KAAK,MAAM;AAAA,MACjB,SAAS,KAAK;AAAA,IAChB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,QAAsD;AACpD,WAAO,MAAM,KAAK,KAAK,MAAM,QAAQ,CAAC,EAAE,IAAI,CAAC,CAAC,KAAK,KAAK,OAAO;AAAA,MAC7D;AAAA,MACA;AAAA,IACF,EAAE;AAAA,EACJ;AAAA;AAAA;AAAA;AAAA,EAKA,UAAkB;AAChB,UAAM,MAAM,KAAK,IAAI;AACrB,QAAI,UAAU;AAEd,eAAW,CAAC,KAAK,KAAK,KAAK,KAAK,MAAM,QAAQ,GAAG;AAC/C,UAAI,MAAM,MAAM,YAAY,KAAK,OAAO;AACtC,aAAK,MAAM,OAAO,GAAG;AACrB;AACA,YAAI,KAAK,eAAe;AACtB,eAAK,QAAQ;AAAA,QACf;AAAA,MACF;AAAA,IACF;AAEA,QAAI,UAAU,GAAG;AACf,aAAO,KAAK,2BAA2B;AAAA,QACrC;AAAA,QACA,WAAW,KAAK,MAAM;AAAA,MACxB,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AACF;AAKO,MAAM,sBAAsB;AAAA,EACzB,aAAa,IAAI,cAAmB,EAAE,SAAS,KAAK,OAAO,IAAO,CAAC;AAAA;AAAA,EACnE,aAAa,IAAI,cAAmB,EAAE,SAAS,KAAM,OAAO,KAAO,CAAC;AAAA;AAAA,EACpE,cAAc,IAAI,cAAmB,EAAE,SAAS,KAAK,OAAO,IAAO,CAAC;AAAA;AAAA,EACpE,cAAc,IAAI,cAAmB,EAAE,SAAS,KAAK,OAAO,IAAO,CAAC;AAAA;AAAA;AAAA;AAAA;AAAA,EAK5E,WAAW,SAAiB,MAAiB;AAC3C,SAAK,WAAW,IAAI,SAAS,OAAO,IAAI,IAAI;AAAA,EAC9C;AAAA,EAEA,SAAS,SAAsB;AAC7B,WAAO,KAAK,WAAW,IAAI,SAAS,OAAO,EAAE;AAAA,EAC/C;AAAA;AAAA;AAAA;AAAA,EAKA,kBAAkB,SAAiB,SAAoB;AACrD,SAAK,WAAW,IAAI,WAAW,OAAO,IAAI,OAAO;AAAA,EACnD;AAAA,EAEA,gBAAgB,SAAsB;AACpC,WAAO,KAAK,WAAW,IAAI,WAAW,OAAO,EAAE;AAAA,EACjD;AAAA;AAAA;AAAA;AAAA,EAKA,iBAAiB,SAAiB,QAAqB;AACrD,SAAK,WAAW,IAAI,UAAU,OAAO,IAAI,MAAM;AAAA,EACjD;AAAA,EAEA,eAAe,SAAwB;AACrC,WAAO,KAAK,WAAW,IAAI,UAAU,OAAO,EAAE;AAAA,EAChD;AAAA;AAAA;AAAA;AAAA,EAKA,aAAa,SAAiB,SAAsB;AAClD,SAAK,YAAY,IAAI,WAAW,OAAO,IAAI,OAAO;AAAA,EACpD;AAAA,EAEA,WAAW,SAAwB;AACjC,WAAO,KAAK,YAAY,IAAI,WAAW,OAAO,EAAE;AAAA,EAClD;AAAA;AAAA;AAAA;AAAA,EAKA,YAAY,SAAiB,QAAmB;AAC9C,SAAK,YAAY,IAAI,UAAU,OAAO,IAAI,MAAM;AAAA,EAClD;AAAA,EAEA,UAAU,SAAsB;AAC9B,WAAO,KAAK,YAAY,IAAI,UAAU,OAAO,EAAE;AAAA,EACjD;AAAA;AAAA;AAAA;AAAA,EAKA,gBAAgB,SAAuB;AACrC,SAAK,WAAW,OAAO,SAAS,OAAO,EAAE;AACzC,SAAK,WAAW,OAAO,WAAW,OAAO,EAAE;AAC3C,SAAK,WAAW,OAAO,UAAU,OAAO,EAAE;AAC1C,SAAK,YAAY,OAAO,WAAW,OAAO,EAAE;AAC5C,SAAK,YAAY,OAAO,UAAU,OAAO,EAAE;AAE3C,WAAO,KAAK,4BAA4B,EAAE,QAAQ,CAAC;AAAA,EACrD;AAAA;AAAA;AAAA;AAAA,EAKA,kBAAkB,WAAyB;AACzC,UAAM,UAAU,IAAI,OAAO,2CAA2C;AAEtE,QAAI,QAAQ;AACZ,aAAS,KAAK,WAAW,kBAAkB,OAAO;AAClD,aAAS,KAAK,WAAW,kBAAkB,OAAO;AAClD,aAAS,KAAK,YAAY,kBAAkB,OAAO;AACnD,aAAS,KAAK,YAAY,kBAAkB,OAAO;AAEnD,WAAO,KAAK,8BAA8B;AAAA,MACxC;AAAA,MACA,kBAAkB;AAAA,IACpB,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,aAAa;AACX,WAAO;AAAA,MACL,OAAO,KAAK,WAAW,WAAW;AAAA,MAClC,OAAO,KAAK,WAAW,WAAW;AAAA,MAClC,QAAQ,KAAK,YAAY,WAAW;AAAA,MACpC,QAAQ,KAAK,YAAY,WAAW;AAAA,IACtC;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,UAAgB;AACd,SAAK,WAAW,QAAQ;AACxB,SAAK,WAAW,QAAQ;AACxB,SAAK,YAAY,QAAQ;AACzB,SAAK,YAAY,QAAQ;AAAA,EAC3B;AAAA;AAAA;AAAA;AAAA,EAKA,QAAc;AACZ,SAAK,WAAW,MAAM;AACtB,SAAK,WAAW,MAAM;AACtB,SAAK,YAAY,MAAM;AACvB,SAAK,YAAY,MAAM;AAEvB,WAAO,KAAK,gCAAgC;AAAA,EAC9C;AACF;AAGA,IAAI,cAA4C;AAKzC,SAAS,gBAAuC;AACrD,MAAI,CAAC,aAAa;AAChB,kBAAc,IAAI,sBAAsB;AAAA,EAC1C;AACA,SAAO;AACT;AAKO,SAAS,eAAe,WAAmB,QAAuB;AACvE,QAAM,YAAY,OACf,IAAI,CAAC,MAAY,OAAO,MAAM,WAAW,KAAK,UAAU,CAAC,IAAI,OAAO,CAAC,CAAE,EACvE,KAAK,GAAG;AAEX,SAAO,GAAG,SAAS,IAAI,SAAS;AAClC;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -334,7 +334,7 @@ class SQLiteAdapter extends FeatureAwareDatabaseAdapter {
|
|
|
334
334
|
// Limited search (basic LIKE queries)
|
|
335
335
|
async search(options) {
|
|
336
336
|
if (!this.db) throw new Error("Database not connected");
|
|
337
|
-
|
|
337
|
+
const sql = `
|
|
338
338
|
SELECT *,
|
|
339
339
|
CASE
|
|
340
340
|
WHEN name LIKE ? THEN 1.0
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/database/sqlite-adapter.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * SQLite Database Adapter\n * Maintains backward compatibility with existing SQLite implementation\n */\n\nimport Database from 'better-sqlite3';\nimport {\n FeatureAwareDatabaseAdapter,\n DatabaseFeatures,\n SearchOptions,\n QueryOptions,\n AggregationOptions,\n BulkOperation,\n DatabaseStats,\n CountResult,\n VersionResult,\n FrameRow,\n EventRow,\n AnchorRow,\n} from './database-adapter.js';\nimport type { Frame, Event, Anchor } from '../context/frame-manager.js';\nimport { logger } from '../monitoring/logger.js';\nimport * as fs from 'fs/promises';\nimport * as path from 'path';\n\nexport interface SQLiteConfig {\n dbPath: string;\n walMode?: boolean;\n busyTimeout?: number;\n cacheSize?: number;\n synchronous?: 'OFF' | 'NORMAL' | 'FULL' | 'EXTRA';\n}\n\nexport class SQLiteAdapter extends FeatureAwareDatabaseAdapter {\n private db: Database.Database | null = null;\n private readonly dbPath: string;\n private inTransactionFlag = false;\n\n constructor(projectId: string, config: SQLiteConfig) {\n super(projectId, config);\n this.dbPath = config.dbPath;\n }\n\n getFeatures(): DatabaseFeatures {\n return {\n supportsFullTextSearch: false, // Could enable with FTS5\n supportsVectorSearch: false,\n supportsPartitioning: false,\n supportsAnalytics: false,\n supportsCompression: false,\n supportsMaterializedViews: false,\n supportsParallelQueries: false,\n };\n }\n\n async connect(): Promise<void> {\n if (this.db) return;\n\n const config = this.config as SQLiteConfig;\n\n // Ensure directory exists\n const dir = path.dirname(this.dbPath);\n await fs.mkdir(dir, { recursive: true });\n\n this.db = new Database(this.dbPath);\n\n // Configure SQLite for better performance\n if (config.walMode !== false) {\n this.db.pragma('journal_mode = WAL');\n }\n\n if (config.busyTimeout) {\n this.db.pragma(`busy_timeout = ${config.busyTimeout}`);\n }\n\n if (config.cacheSize) {\n this.db.pragma(`cache_size = ${config.cacheSize}`);\n }\n\n if (config.synchronous) {\n this.db.pragma(`synchronous = ${config.synchronous}`);\n }\n\n logger.info('SQLite database connected', { dbPath: this.dbPath });\n }\n\n async disconnect(): Promise<void> {\n if (!this.db) return;\n\n this.db.close();\n this.db = null;\n logger.info('SQLite database disconnected');\n }\n\n /**\n * Get raw database handle for testing purposes\n * @internal\n */\n getRawDatabase(): Database.Database | null {\n return this.db;\n }\n\n isConnected(): boolean {\n return this.db !== null && this.db.open;\n }\n\n async ping(): Promise<boolean> {\n if (!this.db) return false;\n\n try {\n this.db.prepare('SELECT 1').get();\n return true;\n } catch {\n return false;\n }\n }\n\n async initializeSchema(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS frames (\n frame_id TEXT PRIMARY KEY,\n run_id TEXT NOT NULL,\n project_id TEXT NOT NULL,\n parent_frame_id TEXT REFERENCES frames(frame_id),\n depth INTEGER NOT NULL DEFAULT 0,\n type TEXT NOT NULL,\n name TEXT NOT NULL,\n state TEXT DEFAULT 'active',\n inputs TEXT DEFAULT '{}',\n outputs TEXT DEFAULT '{}',\n digest_text TEXT,\n digest_json TEXT DEFAULT '{}',\n created_at INTEGER DEFAULT (unixepoch()),\n closed_at INTEGER\n );\n\n CREATE TABLE IF NOT EXISTS events (\n event_id TEXT PRIMARY KEY,\n run_id TEXT NOT NULL,\n frame_id TEXT NOT NULL,\n seq INTEGER NOT NULL,\n event_type TEXT NOT NULL,\n payload TEXT NOT NULL,\n ts INTEGER DEFAULT (unixepoch()),\n FOREIGN KEY(frame_id) REFERENCES frames(frame_id)\n );\n\n CREATE TABLE IF NOT EXISTS anchors (\n anchor_id TEXT PRIMARY KEY,\n frame_id TEXT NOT NULL,\n project_id TEXT NOT NULL,\n type TEXT NOT NULL,\n text TEXT NOT NULL,\n priority INTEGER DEFAULT 0,\n created_at INTEGER DEFAULT (unixepoch()),\n metadata TEXT DEFAULT '{}',\n FOREIGN KEY(frame_id) REFERENCES frames(frame_id)\n );\n\n CREATE TABLE IF NOT EXISTS schema_version (\n version INTEGER PRIMARY KEY,\n applied_at INTEGER DEFAULT (unixepoch())\n );\n\n -- Indexes for performance\n CREATE INDEX IF NOT EXISTS idx_frames_run ON frames(run_id);\n CREATE INDEX IF NOT EXISTS idx_frames_project ON frames(project_id);\n CREATE INDEX IF NOT EXISTS idx_frames_parent ON frames(parent_frame_id);\n CREATE INDEX IF NOT EXISTS idx_frames_state ON frames(state);\n CREATE INDEX IF NOT EXISTS idx_frames_created ON frames(created_at DESC);\n CREATE INDEX IF NOT EXISTS idx_events_frame ON events(frame_id);\n CREATE INDEX IF NOT EXISTS idx_events_seq ON events(frame_id, seq);\n CREATE INDEX IF NOT EXISTS idx_anchors_frame ON anchors(frame_id);\n\n -- Set initial schema version if not exists\n INSERT OR IGNORE INTO schema_version (version) VALUES (1);\n `);\n }\n\n async migrateSchema(targetVersion: number): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n const currentVersion = await this.getSchemaVersion();\n\n if (currentVersion >= targetVersion) {\n logger.info('Schema already at target version', {\n currentVersion,\n targetVersion,\n });\n return;\n }\n\n // Apply migrations sequentially\n for (let v = currentVersion + 1; v <= targetVersion; v++) {\n logger.info(`Applying migration to version ${v}`);\n // Migration logic would go here\n this.db.prepare('UPDATE schema_version SET version = ?').run(v);\n }\n }\n\n async getSchemaVersion(): Promise<number> {\n if (!this.db) throw new Error('Database not connected');\n\n try {\n const result = this.db\n .prepare('SELECT MAX(version) as version FROM schema_version')\n .get() as VersionResult;\n return result?.version || 0;\n } catch {\n return 0;\n }\n }\n\n // Frame operations\n async createFrame(frame: Partial<Frame>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const frameId = frame.frame_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO frames (\n frame_id, run_id, project_id, parent_frame_id, depth,\n type, name, state, inputs, outputs, digest_text, digest_json\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n frameId,\n frame.run_id,\n frame.project_id || this.projectId,\n frame.parent_frame_id || null,\n frame.depth || 0,\n frame.type,\n frame.name,\n frame.state || 'active',\n JSON.stringify(frame.inputs || {}),\n JSON.stringify(frame.outputs || {}),\n frame.digest_text || null,\n JSON.stringify(frame.digest_json || {})\n );\n\n return frameId;\n }\n\n async getFrame(frameId: string): Promise<Frame | null> {\n if (!this.db) throw new Error('Database not connected');\n\n const row = this.db\n .prepare('SELECT * FROM frames WHERE frame_id = ?')\n .get(frameId) as FrameRow | undefined;\n\n if (!row) return null;\n\n return {\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n };\n }\n\n async updateFrame(frameId: string, updates: Partial<Frame>): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n const fields = [];\n const values = [];\n\n if (updates.state !== undefined) {\n fields.push('state = ?');\n values.push(updates.state);\n }\n\n if (updates.outputs !== undefined) {\n fields.push('outputs = ?');\n values.push(JSON.stringify(updates.outputs));\n }\n\n if (updates.digest_text !== undefined) {\n fields.push('digest_text = ?');\n values.push(updates.digest_text);\n }\n\n if (updates.digest_json !== undefined) {\n fields.push('digest_json = ?');\n values.push(JSON.stringify(updates.digest_json));\n }\n\n if (updates.closed_at !== undefined) {\n fields.push('closed_at = ?');\n values.push(updates.closed_at);\n }\n\n if (fields.length === 0) return;\n\n values.push(frameId);\n\n this.db\n .prepare(\n `\n UPDATE frames SET ${fields.join(', ')} WHERE frame_id = ?\n `\n )\n .run(...values);\n }\n\n async deleteFrame(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n // Delete in order due to foreign keys\n await this.deleteFrameAnchors(frameId);\n await this.deleteFrameEvents(frameId);\n\n this.db.prepare('DELETE FROM frames WHERE frame_id = ?').run(frameId);\n }\n\n async getActiveFrames(runId?: string): Promise<Frame[]> {\n if (!this.db) throw new Error('Database not connected');\n\n let query = \"SELECT * FROM frames WHERE state = 'active'\";\n const params = [];\n\n if (runId) {\n query += ' AND run_id = ?';\n params.push(runId);\n }\n\n query += ' ORDER BY depth ASC, created_at ASC';\n\n const rows = this.db.prepare(query).all(...params) as any[];\n\n return rows.map((row) => ({\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n }));\n }\n\n async closeFrame(frameId: string, outputs?: any): Promise<void> {\n await this.updateFrame(frameId, {\n state: 'closed',\n outputs,\n closed_at: Date.now(),\n });\n }\n\n // Event operations\n async createEvent(event: Partial<Event>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const eventId = event.event_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO events (event_id, run_id, frame_id, seq, event_type, payload, ts)\n VALUES (?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n eventId,\n event.run_id,\n event.frame_id,\n event.seq || 0,\n event.event_type,\n JSON.stringify(event.payload || {}),\n event.ts || Date.now()\n );\n\n return eventId;\n }\n\n async getFrameEvents(\n frameId: string,\n options?: QueryOptions\n ): Promise<Event[]> {\n if (!this.db) throw new Error('Database not connected');\n\n let query = 'SELECT * FROM events WHERE frame_id = ?';\n query += this.buildOrderByClause(\n options?.orderBy || 'seq',\n options?.orderDirection\n );\n query += this.buildLimitClause(options?.limit, options?.offset);\n\n const rows = this.db.prepare(query).all(frameId) as any[];\n\n return rows.map((row) => ({\n ...row,\n payload: JSON.parse(row.payload || '{}'),\n }));\n }\n\n async deleteFrameEvents(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('DELETE FROM events WHERE frame_id = ?').run(frameId);\n }\n\n // Anchor operations\n async createAnchor(anchor: Partial<Anchor>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const anchorId = anchor.anchor_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO anchors (anchor_id, frame_id, project_id, type, text, priority, metadata)\n VALUES (?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n anchorId,\n anchor.frame_id,\n anchor.project_id || this.projectId,\n anchor.type,\n anchor.text,\n anchor.priority || 0,\n JSON.stringify(anchor.metadata || {})\n );\n\n return anchorId;\n }\n\n async getFrameAnchors(frameId: string): Promise<Anchor[]> {\n if (!this.db) throw new Error('Database not connected');\n\n const rows = this.db\n .prepare(\n `\n SELECT * FROM anchors WHERE frame_id = ? \n ORDER BY priority DESC, created_at ASC\n `\n )\n .all(frameId) as any[];\n\n return rows.map((row) => ({\n ...row,\n metadata: JSON.parse(row.metadata || '{}'),\n }));\n }\n\n async deleteFrameAnchors(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('DELETE FROM anchors WHERE frame_id = ?').run(frameId);\n }\n\n // Limited search (basic LIKE queries)\n async search(\n options: SearchOptions\n ): Promise<Array<Frame & { score: number }>> {\n if (!this.db) throw new Error('Database not connected');\n\n // SQLite doesn't support HAVING on non-aggregate queries, so we filter in application\n let sql = `\n SELECT *, \n CASE \n WHEN name LIKE ? THEN 1.0\n WHEN digest_text LIKE ? THEN 0.8\n WHEN inputs LIKE ? THEN 0.6\n ELSE 0.5\n END as score\n FROM frames\n WHERE name LIKE ? OR digest_text LIKE ? OR inputs LIKE ?\n ORDER BY score DESC\n `;\n\n const params = Array(6).fill(`%${options.query}%`);\n\n let rows = this.db.prepare(sql).all(...params) as any[];\n\n // Apply score threshold in application layer\n if (options.scoreThreshold) {\n rows = rows.filter((row) => row.score >= options.scoreThreshold);\n }\n\n // Apply limit and offset in application layer if threshold is used\n if (options.limit || options.offset) {\n const start = options.offset || 0;\n const end = options.limit ? start + options.limit : rows.length;\n rows = rows.slice(start, end);\n }\n\n return rows.map((row) => ({\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n }));\n }\n\n async searchByVector(\n embedding: number[],\n options?: QueryOptions\n ): Promise<Array<Frame & { similarity: number }>> {\n // Not supported in SQLite\n logger.warn('Vector search not supported in SQLite adapter');\n return [];\n }\n\n async searchHybrid(\n textQuery: string,\n embedding: number[],\n weights?: { text: number; vector: number }\n ): Promise<Array<Frame & { score: number }>> {\n // Fall back to text search only\n return this.search({ query: textQuery, ...weights });\n }\n\n // Basic aggregation\n async aggregate(\n table: string,\n options: AggregationOptions\n ): Promise<Record<string, any>[]> {\n if (!this.db) throw new Error('Database not connected');\n\n const metrics = options.metrics\n .map(\n (m) =>\n `${m.operation}(${m.field}) AS ${m.alias || `${m.operation}_${m.field}`}`\n )\n .join(', ');\n\n let sql = `SELECT ${options.groupBy.join(', ')}, ${metrics} FROM ${table}`;\n sql += ` GROUP BY ${options.groupBy.join(', ')}`;\n\n if (options.having) {\n const havingClauses = Object.entries(options.having).map(\n ([key, value]) =>\n `${key} ${typeof value === 'object' ? value.op : '='} ?`\n );\n sql += ` HAVING ${havingClauses.join(' AND ')}`;\n }\n\n return this.db\n .prepare(sql)\n .all(...Object.values(options.having || {})) as any[];\n }\n\n // Pattern detection (basic)\n async detectPatterns(timeRange?: { start: Date; end: Date }): Promise<\n Array<{\n pattern: string;\n type: string;\n frequency: number;\n lastSeen: Date;\n }>\n > {\n if (!this.db) throw new Error('Database not connected');\n\n let sql = `\n SELECT type as pattern, type, COUNT(*) as frequency, MAX(created_at) as last_seen\n FROM frames\n `;\n\n const params = [];\n if (timeRange) {\n sql += ' WHERE created_at >= ? AND created_at <= ?';\n params.push(\n Math.floor(timeRange.start.getTime() / 1000),\n Math.floor(timeRange.end.getTime() / 1000)\n );\n }\n\n sql += ' GROUP BY type HAVING COUNT(*) > 1 ORDER BY frequency DESC';\n\n const rows = this.db.prepare(sql).all(...params) as any[];\n\n return rows.map((row) => ({\n pattern: row.pattern,\n type: row.type,\n frequency: row.frequency,\n lastSeen: new Date(row.last_seen * 1000),\n }));\n }\n\n // Bulk operations\n async executeBulk(operations: BulkOperation[]): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n await this.inTransaction(async () => {\n for (const op of operations) {\n switch (op.type) {\n case 'insert':\n // Build insert dynamically based on data\n const insertCols = Object.keys(op.data);\n const insertPlaceholders = insertCols.map(() => '?').join(',');\n this.db!.prepare(\n `INSERT INTO ${op.table} (${insertCols.join(',')}) VALUES (${insertPlaceholders})`\n ).run(...Object.values(op.data));\n break;\n\n case 'update':\n const updateSets = Object.keys(op.data)\n .map((k) => `${k} = ?`)\n .join(',');\n const whereClause = this.buildWhereClause(op.where || {});\n this.db!.prepare(\n `UPDATE ${op.table} SET ${updateSets} ${whereClause}`\n ).run(...Object.values(op.data), ...Object.values(op.where || {}));\n break;\n\n case 'delete':\n const deleteWhere = this.buildWhereClause(op.where || {});\n this.db!.prepare(`DELETE FROM ${op.table} ${deleteWhere}`).run(\n ...Object.values(op.where || {})\n );\n break;\n }\n }\n });\n }\n\n async vacuum(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.pragma('vacuum');\n logger.info('SQLite database vacuumed');\n }\n\n async analyze(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.pragma('analyze');\n logger.info('SQLite database analyzed');\n }\n\n // Statistics\n async getStats(): Promise<DatabaseStats> {\n if (!this.db) throw new Error('Database not connected');\n\n const stats = {\n totalFrames: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM frames')\n .get() as CountResult\n ).count,\n activeFrames: (\n this.db\n .prepare(\n \"SELECT COUNT(*) as count FROM frames WHERE state = 'active'\"\n )\n .get() as CountResult\n ).count,\n totalEvents: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM events')\n .get() as CountResult\n ).count,\n totalAnchors: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM anchors')\n .get() as CountResult\n ).count,\n diskUsage: 0,\n };\n\n // Get file size\n try {\n const fileStats = await fs.stat(this.dbPath);\n stats.diskUsage = fileStats.size;\n } catch {}\n\n return stats;\n }\n\n async getQueryStats(): Promise<\n Array<{\n query: string;\n calls: number;\n meanTime: number;\n totalTime: number;\n }>\n > {\n // SQLite doesn't have built-in query stats\n logger.warn('Query stats not available for SQLite');\n return [];\n }\n\n // Transaction support\n async beginTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('BEGIN').run();\n this.inTransactionFlag = true;\n }\n\n async commitTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('COMMIT').run();\n this.inTransactionFlag = false;\n }\n\n async rollbackTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('ROLLBACK').run();\n this.inTransactionFlag = false;\n }\n\n async inTransaction(\n callback: (adapter: DatabaseAdapter) => Promise<void>\n ): Promise<void> {\n await this.beginTransaction();\n\n try {\n await callback(this);\n await this.commitTransaction();\n } catch (error: unknown) {\n await this.rollbackTransaction();\n throw error;\n }\n }\n\n // Export/Import\n async exportData(\n tables: string[],\n format: 'json' | 'parquet' | 'csv'\n ): Promise<Buffer> {\n if (!this.db) throw new Error('Database not connected');\n\n if (format !== 'json') {\n throw new Error(`Format ${format} not supported for SQLite export`);\n }\n\n const data: Record<string, any[]> = {};\n\n for (const table of tables) {\n data[table] = this.db.prepare(`SELECT * FROM ${table}`).all();\n }\n\n return Buffer.from(JSON.stringify(data, null, 2));\n }\n\n async importData(\n data: Buffer,\n format: 'json' | 'parquet' | 'csv',\n options?: { truncate?: boolean; upsert?: boolean }\n ): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n if (format !== 'json') {\n throw new Error(`Format ${format} not supported for SQLite import`);\n }\n\n const parsed = JSON.parse(data.toString());\n\n await this.inTransaction(async () => {\n for (const [table, rows] of Object.entries(parsed)) {\n if (options?.truncate) {\n this.db!.prepare(`DELETE FROM ${table}`).run();\n }\n\n for (const row of rows as any[]) {\n const cols = Object.keys(row);\n const placeholders = cols.map(() => '?').join(',');\n\n if (options?.upsert) {\n const updates = cols.map((c) => `${c} = excluded.${c}`).join(',');\n this.db!.prepare(\n `INSERT INTO ${table} (${cols.join(',')}) VALUES (${placeholders})\n ON CONFLICT DO UPDATE SET ${updates}`\n ).run(...Object.values(row));\n } else {\n this.db!.prepare(\n `INSERT INTO ${table} (${cols.join(',')}) VALUES (${placeholders})`\n ).run(...Object.values(row));\n }\n }\n }\n });\n }\n}\n"],
|
|
5
|
-
"mappings": "AAKA,OAAO,cAAc;AACrB;AAAA,EACE;AAAA,OAYK;AAEP,SAAS,cAAc;AACvB,YAAY,QAAQ;AACpB,YAAY,UAAU;AAUf,MAAM,sBAAsB,4BAA4B;AAAA,EACrD,KAA+B;AAAA,EACtB;AAAA,EACT,oBAAoB;AAAA,EAE5B,YAAY,WAAmB,QAAsB;AACnD,UAAM,WAAW,MAAM;AACvB,SAAK,SAAS,OAAO;AAAA,EACvB;AAAA,EAEA,cAAgC;AAC9B,WAAO;AAAA,MACL,wBAAwB;AAAA;AAAA,MACxB,sBAAsB;AAAA,MACtB,sBAAsB;AAAA,MACtB,mBAAmB;AAAA,MACnB,qBAAqB;AAAA,MACrB,2BAA2B;AAAA,MAC3B,yBAAyB;AAAA,IAC3B;AAAA,EACF;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI,KAAK,GAAI;AAEb,UAAM,SAAS,KAAK;AAGpB,UAAM,MAAM,KAAK,QAAQ,KAAK,MAAM;AACpC,UAAM,GAAG,MAAM,KAAK,EAAE,WAAW,KAAK,CAAC;AAEvC,SAAK,KAAK,IAAI,SAAS,KAAK,MAAM;AAGlC,QAAI,OAAO,YAAY,OAAO;AAC5B,WAAK,GAAG,OAAO,oBAAoB;AAAA,IACrC;AAEA,QAAI,OAAO,aAAa;AACtB,WAAK,GAAG,OAAO,kBAAkB,OAAO,WAAW,EAAE;AAAA,IACvD;AAEA,QAAI,OAAO,WAAW;AACpB,WAAK,GAAG,OAAO,gBAAgB,OAAO,SAAS,EAAE;AAAA,IACnD;AAEA,QAAI,OAAO,aAAa;AACtB,WAAK,GAAG,OAAO,iBAAiB,OAAO,WAAW,EAAE;AAAA,IACtD;AAEA,WAAO,KAAK,6BAA6B,EAAE,QAAQ,KAAK,OAAO,CAAC;AAAA,EAClE;AAAA,EAEA,MAAM,aAA4B;AAChC,QAAI,CAAC,KAAK,GAAI;AAEd,SAAK,GAAG,MAAM;AACd,SAAK,KAAK;AACV,WAAO,KAAK,8BAA8B;AAAA,EAC5C;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,iBAA2C;AACzC,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,cAAuB;AACrB,WAAO,KAAK,OAAO,QAAQ,KAAK,GAAG;AAAA,EACrC;AAAA,EAEA,MAAM,OAAyB;AAC7B,QAAI,CAAC,KAAK,GAAI,QAAO;AAErB,QAAI;AACF,WAAK,GAAG,QAAQ,UAAU,EAAE,IAAI;AAChC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,mBAAkC;AACtC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KA0DZ;AAAA,EACH;AAAA,EAEA,MAAM,cAAc,eAAsC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,iBAAiB,MAAM,KAAK,iBAAiB;AAEnD,QAAI,kBAAkB,eAAe;AACnC,aAAO,KAAK,oCAAoC;AAAA,QAC9C;AAAA,QACA;AAAA,MACF,CAAC;AACD;AAAA,IACF;AAGA,aAAS,IAAI,iBAAiB,GAAG,KAAK,eAAe,KAAK;AACxD,aAAO,KAAK,iCAAiC,CAAC,EAAE;AAEhD,WAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,CAAC;AAAA,IAChE;AAAA,EACF;AAAA,EAEA,MAAM,mBAAoC;AACxC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI;AACF,YAAM,SAAS,KAAK,GACjB,QAAQ,oDAAoD,EAC5D,IAAI;AACP,aAAO,QAAQ,WAAW;AAAA,IAC5B,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA,EAGA,MAAM,YAAY,OAAwC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,MAAM,YAAY,KAAK,WAAW;AAElD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAMF,EACC;AAAA,MACC;AAAA,MACA,MAAM;AAAA,MACN,MAAM,cAAc,KAAK;AAAA,MACzB,MAAM,mBAAmB;AAAA,MACzB,MAAM,SAAS;AAAA,MACf,MAAM;AAAA,MACN,MAAM;AAAA,MACN,MAAM,SAAS;AAAA,MACf,KAAK,UAAU,MAAM,UAAU,CAAC,CAAC;AAAA,MACjC,KAAK,UAAU,MAAM,WAAW,CAAC,CAAC;AAAA,MAClC,MAAM,eAAe;AAAA,MACrB,KAAK,UAAU,MAAM,eAAe,CAAC,CAAC;AAAA,IACxC;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,SAAS,SAAwC;AACrD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,MAAM,KAAK,GACd,QAAQ,yCAAyC,EACjD,IAAI,OAAO;AAEd,QAAI,CAAC,IAAK,QAAO;AAEjB,WAAO;AAAA,MACL,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD;AAAA,EACF;AAAA,EAEA,MAAM,YAAY,SAAiB,SAAwC;AACzE,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,SAAS,CAAC;AAChB,UAAM,SAAS,CAAC;AAEhB,QAAI,QAAQ,UAAU,QAAW;AAC/B,aAAO,KAAK,WAAW;AACvB,aAAO,KAAK,QAAQ,KAAK;AAAA,IAC3B;AAEA,QAAI,QAAQ,YAAY,QAAW;AACjC,aAAO,KAAK,aAAa;AACzB,aAAO,KAAK,KAAK,UAAU,QAAQ,OAAO,CAAC;AAAA,IAC7C;AAEA,QAAI,QAAQ,gBAAgB,QAAW;AACrC,aAAO,KAAK,iBAAiB;AAC7B,aAAO,KAAK,QAAQ,WAAW;AAAA,IACjC;AAEA,QAAI,QAAQ,gBAAgB,QAAW;AACrC,aAAO,KAAK,iBAAiB;AAC7B,aAAO,KAAK,KAAK,UAAU,QAAQ,WAAW,CAAC;AAAA,IACjD;AAEA,QAAI,QAAQ,cAAc,QAAW;AACnC,aAAO,KAAK,eAAe;AAC3B,aAAO,KAAK,QAAQ,SAAS;AAAA,IAC/B;AAEA,QAAI,OAAO,WAAW,EAAG;AAEzB,WAAO,KAAK,OAAO;AAEnB,SAAK,GACF;AAAA,MACC;AAAA,0BACkB,OAAO,KAAK,IAAI,CAAC;AAAA;AAAA,IAErC,EACC,IAAI,GAAG,MAAM;AAAA,EAClB;AAAA,EAEA,MAAM,YAAY,SAAgC;AAChD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAGtD,UAAM,KAAK,mBAAmB,OAAO;AACrC,UAAM,KAAK,kBAAkB,OAAO;AAEpC,SAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,OAAO;AAAA,EACtE;AAAA,EAEA,MAAM,gBAAgB,OAAkC;AACtD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,QAAQ;AACZ,UAAM,SAAS,CAAC;AAEhB,QAAI,OAAO;AACT,eAAS;AACT,aAAO,KAAK,KAAK;AAAA,IACnB;AAEA,aAAS;AAET,UAAM,OAAO,KAAK,GAAG,QAAQ,KAAK,EAAE,IAAI,GAAG,MAAM;AAEjD,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,WAAW,SAAiB,SAA8B;AAC9D,UAAM,KAAK,YAAY,SAAS;AAAA,MAC9B,OAAO;AAAA,MACP;AAAA,MACA,WAAW,KAAK,IAAI;AAAA,IACtB,CAAC;AAAA,EACH;AAAA;AAAA,EAGA,MAAM,YAAY,OAAwC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,MAAM,YAAY,KAAK,WAAW;AAElD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC;AAAA,MACC;AAAA,MACA,MAAM;AAAA,MACN,MAAM;AAAA,MACN,MAAM,OAAO;AAAA,MACb,MAAM;AAAA,MACN,KAAK,UAAU,MAAM,WAAW,CAAC,CAAC;AAAA,MAClC,MAAM,MAAM,KAAK,IAAI;AAAA,IACvB;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,eACJ,SACA,SACkB;AAClB,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,QAAQ;AACZ,aAAS,KAAK;AAAA,MACZ,SAAS,WAAW;AAAA,MACpB,SAAS;AAAA,IACX;AACA,aAAS,KAAK,iBAAiB,SAAS,OAAO,SAAS,MAAM;AAE9D,UAAM,OAAO,KAAK,GAAG,QAAQ,KAAK,EAAE,IAAI,OAAO;AAE/C,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,IACzC,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,kBAAkB,SAAgC;AACtD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,OAAO;AAAA,EACtE;AAAA;AAAA,EAGA,MAAM,aAAa,QAA0C;AAC3D,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,WAAW,OAAO,aAAa,KAAK,WAAW;AAErD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC;AAAA,MACC;AAAA,MACA,OAAO;AAAA,MACP,OAAO,cAAc,KAAK;AAAA,MAC1B,OAAO;AAAA,MACP,OAAO;AAAA,MACP,OAAO,YAAY;AAAA,MACnB,KAAK,UAAU,OAAO,YAAY,CAAC,CAAC;AAAA,IACtC;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,OAAO,KAAK,GACf;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC,IAAI,OAAO;AAEd,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,UAAU,KAAK,MAAM,IAAI,YAAY,IAAI;AAAA,IAC3C,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,mBAAmB,SAAgC;AACvD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,wCAAwC,EAAE,IAAI,OAAO;AAAA,EACvE;AAAA;AAAA,EAGA,MAAM,OACJ,SAC2C;AAC3C,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAGtD,QAAI,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAaV,UAAM,SAAS,MAAM,CAAC,EAAE,KAAK,IAAI,QAAQ,KAAK,GAAG;AAEjD,QAAI,OAAO,KAAK,GAAG,QAAQ,GAAG,EAAE,IAAI,GAAG,MAAM;AAG7C,QAAI,QAAQ,gBAAgB;AAC1B,aAAO,KAAK,OAAO,CAAC,QAAQ,IAAI,SAAS,QAAQ,cAAc;AAAA,IACjE;AAGA,QAAI,QAAQ,SAAS,QAAQ,QAAQ;AACnC,YAAM,QAAQ,QAAQ,UAAU;AAChC,YAAM,MAAM,QAAQ,QAAQ,QAAQ,QAAQ,QAAQ,KAAK;AACzD,aAAO,KAAK,MAAM,OAAO,GAAG;AAAA,IAC9B;AAEA,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,eACJ,WACA,SACgD;AAEhD,WAAO,KAAK,+CAA+C;AAC3D,WAAO,CAAC;AAAA,EACV;AAAA,EAEA,MAAM,aACJ,WACA,WACA,SAC2C;AAE3C,WAAO,KAAK,OAAO,EAAE,OAAO,WAAW,GAAG,QAAQ,CAAC;AAAA,EACrD;AAAA;AAAA,EAGA,MAAM,UACJ,OACA,SACgC;AAChC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,QAAQ,QACrB;AAAA,MACC,CAAC,MACC,GAAG,EAAE,SAAS,IAAI,EAAE,KAAK,QAAQ,EAAE,SAAS,GAAG,EAAE,SAAS,IAAI,EAAE,KAAK,EAAE;AAAA,IAC3E,EACC,KAAK,IAAI;AAEZ,QAAI,MAAM,UAAU,QAAQ,QAAQ,KAAK,IAAI,CAAC,KAAK,OAAO,SAAS,KAAK;AACxE,WAAO,aAAa,QAAQ,QAAQ,KAAK,IAAI,CAAC;AAE9C,QAAI,QAAQ,QAAQ;AAClB,YAAM,gBAAgB,OAAO,QAAQ,QAAQ,MAAM,EAAE;AAAA,QACnD,CAAC,CAAC,KAAK,KAAK,MACV,GAAG,GAAG,IAAI,OAAO,UAAU,WAAW,MAAM,KAAK,GAAG;AAAA,MACxD;AACA,aAAO,WAAW,cAAc,KAAK,OAAO,CAAC;AAAA,IAC/C;AAEA,WAAO,KAAK,GACT,QAAQ,GAAG,EACX,IAAI,GAAG,OAAO,OAAO,QAAQ,UAAU,CAAC,CAAC,CAAC;AAAA,EAC/C;AAAA;AAAA,EAGA,MAAM,eAAe,WAOnB;AACA,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,MAAM;AAAA;AAAA;AAAA;AAKV,UAAM,SAAS,CAAC;AAChB,QAAI,WAAW;AACb,aAAO;AACP,aAAO;AAAA,QACL,KAAK,MAAM,UAAU,MAAM,QAAQ,IAAI,GAAI;AAAA,QAC3C,KAAK,MAAM,UAAU,IAAI,QAAQ,IAAI,GAAI;AAAA,MAC3C;AAAA,IACF;AAEA,WAAO;AAEP,UAAM,OAAO,KAAK,GAAG,QAAQ,GAAG,EAAE,IAAI,GAAG,MAAM;AAE/C,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,SAAS,IAAI;AAAA,MACb,MAAM,IAAI;AAAA,MACV,WAAW,IAAI;AAAA,MACf,UAAU,IAAI,KAAK,IAAI,YAAY,GAAI;AAAA,IACzC,EAAE;AAAA,EACJ;AAAA;AAAA,EAGA,MAAM,YAAY,YAA4C;AAC5D,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,KAAK,cAAc,YAAY;AACnC,iBAAW,MAAM,YAAY;AAC3B,gBAAQ,GAAG,MAAM;AAAA,UACf,KAAK;AAEH,kBAAM,aAAa,OAAO,KAAK,GAAG,IAAI;AACtC,kBAAM,qBAAqB,WAAW,IAAI,MAAM,GAAG,EAAE,KAAK,GAAG;AAC7D,iBAAK,GAAI;AAAA,cACP,eAAe,GAAG,KAAK,KAAK,WAAW,KAAK,GAAG,CAAC,aAAa,kBAAkB;AAAA,YACjF,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,IAAI,CAAC;AAC/B;AAAA,UAEF,KAAK;AACH,kBAAM,aAAa,OAAO,KAAK,GAAG,IAAI,EACnC,IAAI,CAAC,MAAM,GAAG,CAAC,MAAM,EACrB,KAAK,GAAG;AACX,kBAAM,cAAc,KAAK,iBAAiB,GAAG,SAAS,CAAC,CAAC;AACxD,iBAAK,GAAI;AAAA,cACP,UAAU,GAAG,KAAK,QAAQ,UAAU,IAAI,WAAW;AAAA,YACrD,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,IAAI,GAAG,GAAG,OAAO,OAAO,GAAG,SAAS,CAAC,CAAC,CAAC;AACjE;AAAA,UAEF,KAAK;AACH,kBAAM,cAAc,KAAK,iBAAiB,GAAG,SAAS,CAAC,CAAC;AACxD,iBAAK,GAAI,QAAQ,eAAe,GAAG,KAAK,IAAI,WAAW,EAAE,EAAE;AAAA,cACzD,GAAG,OAAO,OAAO,GAAG,SAAS,CAAC,CAAC;AAAA,YACjC;AACA;AAAA,QACJ;AAAA,MACF;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,SAAwB;AAC5B,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,OAAO,QAAQ;AACvB,WAAO,KAAK,0BAA0B;AAAA,EACxC;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,OAAO,SAAS;AACxB,WAAO,KAAK,0BAA0B;AAAA,EACxC;AAAA;AAAA,EAGA,MAAM,WAAmC;AACvC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,QAAQ;AAAA,MACZ,aACE,KAAK,GACF,QAAQ,sCAAsC,EAC9C,IAAI,EACP;AAAA,MACF,cACE,KAAK,GACF;AAAA,QACC;AAAA,MACF,EACC,IAAI,EACP;AAAA,MACF,aACE,KAAK,GACF,QAAQ,sCAAsC,EAC9C,IAAI,EACP;AAAA,MACF,cACE,KAAK,GACF,QAAQ,uCAAuC,EAC/C,IAAI,EACP;AAAA,MACF,WAAW;AAAA,IACb;AAGA,QAAI;AACF,YAAM,YAAY,MAAM,GAAG,KAAK,KAAK,MAAM;AAC3C,YAAM,YAAY,UAAU;AAAA,IAC9B,QAAQ;AAAA,IAAC;AAET,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,gBAOJ;AAEA,WAAO,KAAK,sCAAsC;AAClD,WAAO,CAAC;AAAA,EACV;AAAA;AAAA,EAGA,MAAM,mBAAkC;AACtC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,OAAO,EAAE,IAAI;AAC7B,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,oBAAmC;AACvC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,QAAQ,EAAE,IAAI;AAC9B,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,sBAAqC;AACzC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,UAAU,EAAE,IAAI;AAChC,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,cACJ,UACe;AACf,UAAM,KAAK,iBAAiB;AAE5B,QAAI;AACF,YAAM,SAAS,IAAI;AACnB,YAAM,KAAK,kBAAkB;AAAA,IAC/B,SAAS,OAAgB;AACvB,YAAM,KAAK,oBAAoB;AAC/B,YAAM;AAAA,IACR;AAAA,EACF;AAAA;AAAA,EAGA,MAAM,WACJ,QACA,QACiB;AACjB,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,WAAW,QAAQ;AACrB,YAAM,IAAI,MAAM,UAAU,MAAM,kCAAkC;AAAA,IACpE;AAEA,UAAM,OAA8B,CAAC;AAErC,eAAW,SAAS,QAAQ;AAC1B,WAAK,KAAK,IAAI,KAAK,GAAG,QAAQ,iBAAiB,KAAK,EAAE,EAAE,IAAI;AAAA,IAC9D;AAEA,WAAO,OAAO,KAAK,KAAK,UAAU,MAAM,MAAM,CAAC,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WACJ,MACA,QACA,SACe;AACf,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,WAAW,QAAQ;AACrB,YAAM,IAAI,MAAM,UAAU,MAAM,kCAAkC;AAAA,IACpE;AAEA,UAAM,SAAS,KAAK,MAAM,KAAK,SAAS,CAAC;AAEzC,UAAM,KAAK,cAAc,YAAY;AACnC,iBAAW,CAAC,OAAO,IAAI,KAAK,OAAO,QAAQ,MAAM,GAAG;AAClD,YAAI,SAAS,UAAU;AACrB,eAAK,GAAI,QAAQ,eAAe,KAAK,EAAE,EAAE,IAAI;AAAA,QAC/C;AAEA,mBAAW,OAAO,MAAe;AAC/B,gBAAM,OAAO,OAAO,KAAK,GAAG;AAC5B,gBAAM,eAAe,KAAK,IAAI,MAAM,GAAG,EAAE,KAAK,GAAG;AAEjD,cAAI,SAAS,QAAQ;AACnB,kBAAM,UAAU,KAAK,IAAI,CAAC,MAAM,GAAG,CAAC,eAAe,CAAC,EAAE,EAAE,KAAK,GAAG;AAChE,iBAAK,GAAI;AAAA,cACP,eAAe,KAAK,KAAK,KAAK,KAAK,GAAG,CAAC,aAAa,YAAY;AAAA,2CACnC,OAAO;AAAA,YACtC,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,CAAC;AAAA,UAC7B,OAAO;AACL,iBAAK,GAAI;AAAA,cACP,eAAe,KAAK,KAAK,KAAK,KAAK,GAAG,CAAC,aAAa,YAAY;AAAA,YAClE,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,CAAC;AAAA,UAC7B;AAAA,QACF;AAAA,MACF;AAAA,IACF,CAAC;AAAA,EACH;AACF;",
|
|
4
|
+
"sourcesContent": ["/**\n * SQLite Database Adapter\n * Maintains backward compatibility with existing SQLite implementation\n */\n\nimport Database from 'better-sqlite3';\nimport {\n FeatureAwareDatabaseAdapter,\n DatabaseFeatures,\n SearchOptions,\n QueryOptions,\n AggregationOptions,\n BulkOperation,\n DatabaseStats,\n CountResult,\n VersionResult,\n FrameRow,\n EventRow,\n AnchorRow,\n} from './database-adapter.js';\nimport type { Frame, Event, Anchor } from '../context/frame-manager.js';\nimport { logger } from '../monitoring/logger.js';\nimport * as fs from 'fs/promises';\nimport * as path from 'path';\n\nexport interface SQLiteConfig {\n dbPath: string;\n walMode?: boolean;\n busyTimeout?: number;\n cacheSize?: number;\n synchronous?: 'OFF' | 'NORMAL' | 'FULL' | 'EXTRA';\n}\n\nexport class SQLiteAdapter extends FeatureAwareDatabaseAdapter {\n private db: Database.Database | null = null;\n private readonly dbPath: string;\n private inTransactionFlag = false;\n\n constructor(projectId: string, config: SQLiteConfig) {\n super(projectId, config);\n this.dbPath = config.dbPath;\n }\n\n getFeatures(): DatabaseFeatures {\n return {\n supportsFullTextSearch: false, // Could enable with FTS5\n supportsVectorSearch: false,\n supportsPartitioning: false,\n supportsAnalytics: false,\n supportsCompression: false,\n supportsMaterializedViews: false,\n supportsParallelQueries: false,\n };\n }\n\n async connect(): Promise<void> {\n if (this.db) return;\n\n const config = this.config as SQLiteConfig;\n\n // Ensure directory exists\n const dir = path.dirname(this.dbPath);\n await fs.mkdir(dir, { recursive: true });\n\n this.db = new Database(this.dbPath);\n\n // Configure SQLite for better performance\n if (config.walMode !== false) {\n this.db.pragma('journal_mode = WAL');\n }\n\n if (config.busyTimeout) {\n this.db.pragma(`busy_timeout = ${config.busyTimeout}`);\n }\n\n if (config.cacheSize) {\n this.db.pragma(`cache_size = ${config.cacheSize}`);\n }\n\n if (config.synchronous) {\n this.db.pragma(`synchronous = ${config.synchronous}`);\n }\n\n logger.info('SQLite database connected', { dbPath: this.dbPath });\n }\n\n async disconnect(): Promise<void> {\n if (!this.db) return;\n\n this.db.close();\n this.db = null;\n logger.info('SQLite database disconnected');\n }\n\n /**\n * Get raw database handle for testing purposes\n * @internal\n */\n getRawDatabase(): Database.Database | null {\n return this.db;\n }\n\n isConnected(): boolean {\n return this.db !== null && this.db.open;\n }\n\n async ping(): Promise<boolean> {\n if (!this.db) return false;\n\n try {\n this.db.prepare('SELECT 1').get();\n return true;\n } catch {\n return false;\n }\n }\n\n async initializeSchema(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS frames (\n frame_id TEXT PRIMARY KEY,\n run_id TEXT NOT NULL,\n project_id TEXT NOT NULL,\n parent_frame_id TEXT REFERENCES frames(frame_id),\n depth INTEGER NOT NULL DEFAULT 0,\n type TEXT NOT NULL,\n name TEXT NOT NULL,\n state TEXT DEFAULT 'active',\n inputs TEXT DEFAULT '{}',\n outputs TEXT DEFAULT '{}',\n digest_text TEXT,\n digest_json TEXT DEFAULT '{}',\n created_at INTEGER DEFAULT (unixepoch()),\n closed_at INTEGER\n );\n\n CREATE TABLE IF NOT EXISTS events (\n event_id TEXT PRIMARY KEY,\n run_id TEXT NOT NULL,\n frame_id TEXT NOT NULL,\n seq INTEGER NOT NULL,\n event_type TEXT NOT NULL,\n payload TEXT NOT NULL,\n ts INTEGER DEFAULT (unixepoch()),\n FOREIGN KEY(frame_id) REFERENCES frames(frame_id)\n );\n\n CREATE TABLE IF NOT EXISTS anchors (\n anchor_id TEXT PRIMARY KEY,\n frame_id TEXT NOT NULL,\n project_id TEXT NOT NULL,\n type TEXT NOT NULL,\n text TEXT NOT NULL,\n priority INTEGER DEFAULT 0,\n created_at INTEGER DEFAULT (unixepoch()),\n metadata TEXT DEFAULT '{}',\n FOREIGN KEY(frame_id) REFERENCES frames(frame_id)\n );\n\n CREATE TABLE IF NOT EXISTS schema_version (\n version INTEGER PRIMARY KEY,\n applied_at INTEGER DEFAULT (unixepoch())\n );\n\n -- Indexes for performance\n CREATE INDEX IF NOT EXISTS idx_frames_run ON frames(run_id);\n CREATE INDEX IF NOT EXISTS idx_frames_project ON frames(project_id);\n CREATE INDEX IF NOT EXISTS idx_frames_parent ON frames(parent_frame_id);\n CREATE INDEX IF NOT EXISTS idx_frames_state ON frames(state);\n CREATE INDEX IF NOT EXISTS idx_frames_created ON frames(created_at DESC);\n CREATE INDEX IF NOT EXISTS idx_events_frame ON events(frame_id);\n CREATE INDEX IF NOT EXISTS idx_events_seq ON events(frame_id, seq);\n CREATE INDEX IF NOT EXISTS idx_anchors_frame ON anchors(frame_id);\n\n -- Set initial schema version if not exists\n INSERT OR IGNORE INTO schema_version (version) VALUES (1);\n `);\n }\n\n async migrateSchema(targetVersion: number): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n const currentVersion = await this.getSchemaVersion();\n\n if (currentVersion >= targetVersion) {\n logger.info('Schema already at target version', {\n currentVersion,\n targetVersion,\n });\n return;\n }\n\n // Apply migrations sequentially\n for (let v = currentVersion + 1; v <= targetVersion; v++) {\n logger.info(`Applying migration to version ${v}`);\n // Migration logic would go here\n this.db.prepare('UPDATE schema_version SET version = ?').run(v);\n }\n }\n\n async getSchemaVersion(): Promise<number> {\n if (!this.db) throw new Error('Database not connected');\n\n try {\n const result = this.db\n .prepare('SELECT MAX(version) as version FROM schema_version')\n .get() as VersionResult;\n return result?.version || 0;\n } catch {\n return 0;\n }\n }\n\n // Frame operations\n async createFrame(frame: Partial<Frame>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const frameId = frame.frame_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO frames (\n frame_id, run_id, project_id, parent_frame_id, depth,\n type, name, state, inputs, outputs, digest_text, digest_json\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n frameId,\n frame.run_id,\n frame.project_id || this.projectId,\n frame.parent_frame_id || null,\n frame.depth || 0,\n frame.type,\n frame.name,\n frame.state || 'active',\n JSON.stringify(frame.inputs || {}),\n JSON.stringify(frame.outputs || {}),\n frame.digest_text || null,\n JSON.stringify(frame.digest_json || {})\n );\n\n return frameId;\n }\n\n async getFrame(frameId: string): Promise<Frame | null> {\n if (!this.db) throw new Error('Database not connected');\n\n const row = this.db\n .prepare('SELECT * FROM frames WHERE frame_id = ?')\n .get(frameId) as FrameRow | undefined;\n\n if (!row) return null;\n\n return {\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n };\n }\n\n async updateFrame(frameId: string, updates: Partial<Frame>): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n const fields = [];\n const values = [];\n\n if (updates.state !== undefined) {\n fields.push('state = ?');\n values.push(updates.state);\n }\n\n if (updates.outputs !== undefined) {\n fields.push('outputs = ?');\n values.push(JSON.stringify(updates.outputs));\n }\n\n if (updates.digest_text !== undefined) {\n fields.push('digest_text = ?');\n values.push(updates.digest_text);\n }\n\n if (updates.digest_json !== undefined) {\n fields.push('digest_json = ?');\n values.push(JSON.stringify(updates.digest_json));\n }\n\n if (updates.closed_at !== undefined) {\n fields.push('closed_at = ?');\n values.push(updates.closed_at);\n }\n\n if (fields.length === 0) return;\n\n values.push(frameId);\n\n this.db\n .prepare(\n `\n UPDATE frames SET ${fields.join(', ')} WHERE frame_id = ?\n `\n )\n .run(...values);\n }\n\n async deleteFrame(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n // Delete in order due to foreign keys\n await this.deleteFrameAnchors(frameId);\n await this.deleteFrameEvents(frameId);\n\n this.db.prepare('DELETE FROM frames WHERE frame_id = ?').run(frameId);\n }\n\n async getActiveFrames(runId?: string): Promise<Frame[]> {\n if (!this.db) throw new Error('Database not connected');\n\n let query = \"SELECT * FROM frames WHERE state = 'active'\";\n const params = [];\n\n if (runId) {\n query += ' AND run_id = ?';\n params.push(runId);\n }\n\n query += ' ORDER BY depth ASC, created_at ASC';\n\n const rows = this.db.prepare(query).all(...params) as any[];\n\n return rows.map((row) => ({\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n }));\n }\n\n async closeFrame(frameId: string, outputs?: any): Promise<void> {\n await this.updateFrame(frameId, {\n state: 'closed',\n outputs,\n closed_at: Date.now(),\n });\n }\n\n // Event operations\n async createEvent(event: Partial<Event>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const eventId = event.event_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO events (event_id, run_id, frame_id, seq, event_type, payload, ts)\n VALUES (?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n eventId,\n event.run_id,\n event.frame_id,\n event.seq || 0,\n event.event_type,\n JSON.stringify(event.payload || {}),\n event.ts || Date.now()\n );\n\n return eventId;\n }\n\n async getFrameEvents(\n frameId: string,\n options?: QueryOptions\n ): Promise<Event[]> {\n if (!this.db) throw new Error('Database not connected');\n\n let query = 'SELECT * FROM events WHERE frame_id = ?';\n query += this.buildOrderByClause(\n options?.orderBy || 'seq',\n options?.orderDirection\n );\n query += this.buildLimitClause(options?.limit, options?.offset);\n\n const rows = this.db.prepare(query).all(frameId) as any[];\n\n return rows.map((row) => ({\n ...row,\n payload: JSON.parse(row.payload || '{}'),\n }));\n }\n\n async deleteFrameEvents(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('DELETE FROM events WHERE frame_id = ?').run(frameId);\n }\n\n // Anchor operations\n async createAnchor(anchor: Partial<Anchor>): Promise<string> {\n if (!this.db) throw new Error('Database not connected');\n\n const anchorId = anchor.anchor_id || this.generateId();\n\n this.db\n .prepare(\n `\n INSERT INTO anchors (anchor_id, frame_id, project_id, type, text, priority, metadata)\n VALUES (?, ?, ?, ?, ?, ?, ?)\n `\n )\n .run(\n anchorId,\n anchor.frame_id,\n anchor.project_id || this.projectId,\n anchor.type,\n anchor.text,\n anchor.priority || 0,\n JSON.stringify(anchor.metadata || {})\n );\n\n return anchorId;\n }\n\n async getFrameAnchors(frameId: string): Promise<Anchor[]> {\n if (!this.db) throw new Error('Database not connected');\n\n const rows = this.db\n .prepare(\n `\n SELECT * FROM anchors WHERE frame_id = ? \n ORDER BY priority DESC, created_at ASC\n `\n )\n .all(frameId) as any[];\n\n return rows.map((row) => ({\n ...row,\n metadata: JSON.parse(row.metadata || '{}'),\n }));\n }\n\n async deleteFrameAnchors(frameId: string): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('DELETE FROM anchors WHERE frame_id = ?').run(frameId);\n }\n\n // Limited search (basic LIKE queries)\n async search(\n options: SearchOptions\n ): Promise<Array<Frame & { score: number }>> {\n if (!this.db) throw new Error('Database not connected');\n\n // SQLite doesn't support HAVING on non-aggregate queries, so we filter in application\n const sql = `\n SELECT *, \n CASE \n WHEN name LIKE ? THEN 1.0\n WHEN digest_text LIKE ? THEN 0.8\n WHEN inputs LIKE ? THEN 0.6\n ELSE 0.5\n END as score\n FROM frames\n WHERE name LIKE ? OR digest_text LIKE ? OR inputs LIKE ?\n ORDER BY score DESC\n `;\n\n const params = Array(6).fill(`%${options.query}%`);\n\n let rows = this.db.prepare(sql).all(...params) as any[];\n\n // Apply score threshold in application layer\n if (options.scoreThreshold) {\n rows = rows.filter((row) => row.score >= options.scoreThreshold);\n }\n\n // Apply limit and offset in application layer if threshold is used\n if (options.limit || options.offset) {\n const start = options.offset || 0;\n const end = options.limit ? start + options.limit : rows.length;\n rows = rows.slice(start, end);\n }\n\n return rows.map((row) => ({\n ...row,\n inputs: JSON.parse(row.inputs || '{}'),\n outputs: JSON.parse(row.outputs || '{}'),\n digest_json: JSON.parse(row.digest_json || '{}'),\n }));\n }\n\n async searchByVector(\n embedding: number[],\n options?: QueryOptions\n ): Promise<Array<Frame & { similarity: number }>> {\n // Not supported in SQLite\n logger.warn('Vector search not supported in SQLite adapter');\n return [];\n }\n\n async searchHybrid(\n textQuery: string,\n embedding: number[],\n weights?: { text: number; vector: number }\n ): Promise<Array<Frame & { score: number }>> {\n // Fall back to text search only\n return this.search({ query: textQuery, ...weights });\n }\n\n // Basic aggregation\n async aggregate(\n table: string,\n options: AggregationOptions\n ): Promise<Record<string, any>[]> {\n if (!this.db) throw new Error('Database not connected');\n\n const metrics = options.metrics\n .map(\n (m) =>\n `${m.operation}(${m.field}) AS ${m.alias || `${m.operation}_${m.field}`}`\n )\n .join(', ');\n\n let sql = `SELECT ${options.groupBy.join(', ')}, ${metrics} FROM ${table}`;\n sql += ` GROUP BY ${options.groupBy.join(', ')}`;\n\n if (options.having) {\n const havingClauses = Object.entries(options.having).map(\n ([key, value]) =>\n `${key} ${typeof value === 'object' ? value.op : '='} ?`\n );\n sql += ` HAVING ${havingClauses.join(' AND ')}`;\n }\n\n return this.db\n .prepare(sql)\n .all(...Object.values(options.having || {})) as any[];\n }\n\n // Pattern detection (basic)\n async detectPatterns(timeRange?: { start: Date; end: Date }): Promise<\n Array<{\n pattern: string;\n type: string;\n frequency: number;\n lastSeen: Date;\n }>\n > {\n if (!this.db) throw new Error('Database not connected');\n\n let sql = `\n SELECT type as pattern, type, COUNT(*) as frequency, MAX(created_at) as last_seen\n FROM frames\n `;\n\n const params = [];\n if (timeRange) {\n sql += ' WHERE created_at >= ? AND created_at <= ?';\n params.push(\n Math.floor(timeRange.start.getTime() / 1000),\n Math.floor(timeRange.end.getTime() / 1000)\n );\n }\n\n sql += ' GROUP BY type HAVING COUNT(*) > 1 ORDER BY frequency DESC';\n\n const rows = this.db.prepare(sql).all(...params) as any[];\n\n return rows.map((row) => ({\n pattern: row.pattern,\n type: row.type,\n frequency: row.frequency,\n lastSeen: new Date(row.last_seen * 1000),\n }));\n }\n\n // Bulk operations\n async executeBulk(operations: BulkOperation[]): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n await this.inTransaction(async () => {\n for (const op of operations) {\n switch (op.type) {\n case 'insert':\n // Build insert dynamically based on data\n const insertCols = Object.keys(op.data);\n const insertPlaceholders = insertCols.map(() => '?').join(',');\n this.db!.prepare(\n `INSERT INTO ${op.table} (${insertCols.join(',')}) VALUES (${insertPlaceholders})`\n ).run(...Object.values(op.data));\n break;\n\n case 'update':\n const updateSets = Object.keys(op.data)\n .map((k) => `${k} = ?`)\n .join(',');\n const whereClause = this.buildWhereClause(op.where || {});\n this.db!.prepare(\n `UPDATE ${op.table} SET ${updateSets} ${whereClause}`\n ).run(...Object.values(op.data), ...Object.values(op.where || {}));\n break;\n\n case 'delete':\n const deleteWhere = this.buildWhereClause(op.where || {});\n this.db!.prepare(`DELETE FROM ${op.table} ${deleteWhere}`).run(\n ...Object.values(op.where || {})\n );\n break;\n }\n }\n });\n }\n\n async vacuum(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.pragma('vacuum');\n logger.info('SQLite database vacuumed');\n }\n\n async analyze(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.pragma('analyze');\n logger.info('SQLite database analyzed');\n }\n\n // Statistics\n async getStats(): Promise<DatabaseStats> {\n if (!this.db) throw new Error('Database not connected');\n\n const stats = {\n totalFrames: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM frames')\n .get() as CountResult\n ).count,\n activeFrames: (\n this.db\n .prepare(\n \"SELECT COUNT(*) as count FROM frames WHERE state = 'active'\"\n )\n .get() as CountResult\n ).count,\n totalEvents: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM events')\n .get() as CountResult\n ).count,\n totalAnchors: (\n this.db\n .prepare('SELECT COUNT(*) as count FROM anchors')\n .get() as CountResult\n ).count,\n diskUsage: 0,\n };\n\n // Get file size\n try {\n const fileStats = await fs.stat(this.dbPath);\n stats.diskUsage = fileStats.size;\n } catch {}\n\n return stats;\n }\n\n async getQueryStats(): Promise<\n Array<{\n query: string;\n calls: number;\n meanTime: number;\n totalTime: number;\n }>\n > {\n // SQLite doesn't have built-in query stats\n logger.warn('Query stats not available for SQLite');\n return [];\n }\n\n // Transaction support\n async beginTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('BEGIN').run();\n this.inTransactionFlag = true;\n }\n\n async commitTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('COMMIT').run();\n this.inTransactionFlag = false;\n }\n\n async rollbackTransaction(): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n this.db.prepare('ROLLBACK').run();\n this.inTransactionFlag = false;\n }\n\n async inTransaction(\n callback: (adapter: DatabaseAdapter) => Promise<void>\n ): Promise<void> {\n await this.beginTransaction();\n\n try {\n await callback(this);\n await this.commitTransaction();\n } catch (error: unknown) {\n await this.rollbackTransaction();\n throw error;\n }\n }\n\n // Export/Import\n async exportData(\n tables: string[],\n format: 'json' | 'parquet' | 'csv'\n ): Promise<Buffer> {\n if (!this.db) throw new Error('Database not connected');\n\n if (format !== 'json') {\n throw new Error(`Format ${format} not supported for SQLite export`);\n }\n\n const data: Record<string, any[]> = {};\n\n for (const table of tables) {\n data[table] = this.db.prepare(`SELECT * FROM ${table}`).all();\n }\n\n return Buffer.from(JSON.stringify(data, null, 2));\n }\n\n async importData(\n data: Buffer,\n format: 'json' | 'parquet' | 'csv',\n options?: { truncate?: boolean; upsert?: boolean }\n ): Promise<void> {\n if (!this.db) throw new Error('Database not connected');\n\n if (format !== 'json') {\n throw new Error(`Format ${format} not supported for SQLite import`);\n }\n\n const parsed = JSON.parse(data.toString());\n\n await this.inTransaction(async () => {\n for (const [table, rows] of Object.entries(parsed)) {\n if (options?.truncate) {\n this.db!.prepare(`DELETE FROM ${table}`).run();\n }\n\n for (const row of rows as any[]) {\n const cols = Object.keys(row);\n const placeholders = cols.map(() => '?').join(',');\n\n if (options?.upsert) {\n const updates = cols.map((c) => `${c} = excluded.${c}`).join(',');\n this.db!.prepare(\n `INSERT INTO ${table} (${cols.join(',')}) VALUES (${placeholders})\n ON CONFLICT DO UPDATE SET ${updates}`\n ).run(...Object.values(row));\n } else {\n this.db!.prepare(\n `INSERT INTO ${table} (${cols.join(',')}) VALUES (${placeholders})`\n ).run(...Object.values(row));\n }\n }\n }\n });\n }\n}\n"],
|
|
5
|
+
"mappings": "AAKA,OAAO,cAAc;AACrB;AAAA,EACE;AAAA,OAYK;AAEP,SAAS,cAAc;AACvB,YAAY,QAAQ;AACpB,YAAY,UAAU;AAUf,MAAM,sBAAsB,4BAA4B;AAAA,EACrD,KAA+B;AAAA,EACtB;AAAA,EACT,oBAAoB;AAAA,EAE5B,YAAY,WAAmB,QAAsB;AACnD,UAAM,WAAW,MAAM;AACvB,SAAK,SAAS,OAAO;AAAA,EACvB;AAAA,EAEA,cAAgC;AAC9B,WAAO;AAAA,MACL,wBAAwB;AAAA;AAAA,MACxB,sBAAsB;AAAA,MACtB,sBAAsB;AAAA,MACtB,mBAAmB;AAAA,MACnB,qBAAqB;AAAA,MACrB,2BAA2B;AAAA,MAC3B,yBAAyB;AAAA,IAC3B;AAAA,EACF;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI,KAAK,GAAI;AAEb,UAAM,SAAS,KAAK;AAGpB,UAAM,MAAM,KAAK,QAAQ,KAAK,MAAM;AACpC,UAAM,GAAG,MAAM,KAAK,EAAE,WAAW,KAAK,CAAC;AAEvC,SAAK,KAAK,IAAI,SAAS,KAAK,MAAM;AAGlC,QAAI,OAAO,YAAY,OAAO;AAC5B,WAAK,GAAG,OAAO,oBAAoB;AAAA,IACrC;AAEA,QAAI,OAAO,aAAa;AACtB,WAAK,GAAG,OAAO,kBAAkB,OAAO,WAAW,EAAE;AAAA,IACvD;AAEA,QAAI,OAAO,WAAW;AACpB,WAAK,GAAG,OAAO,gBAAgB,OAAO,SAAS,EAAE;AAAA,IACnD;AAEA,QAAI,OAAO,aAAa;AACtB,WAAK,GAAG,OAAO,iBAAiB,OAAO,WAAW,EAAE;AAAA,IACtD;AAEA,WAAO,KAAK,6BAA6B,EAAE,QAAQ,KAAK,OAAO,CAAC;AAAA,EAClE;AAAA,EAEA,MAAM,aAA4B;AAChC,QAAI,CAAC,KAAK,GAAI;AAEd,SAAK,GAAG,MAAM;AACd,SAAK,KAAK;AACV,WAAO,KAAK,8BAA8B;AAAA,EAC5C;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,iBAA2C;AACzC,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,cAAuB;AACrB,WAAO,KAAK,OAAO,QAAQ,KAAK,GAAG;AAAA,EACrC;AAAA,EAEA,MAAM,OAAyB;AAC7B,QAAI,CAAC,KAAK,GAAI,QAAO;AAErB,QAAI;AACF,WAAK,GAAG,QAAQ,UAAU,EAAE,IAAI;AAChC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,mBAAkC;AACtC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KA0DZ;AAAA,EACH;AAAA,EAEA,MAAM,cAAc,eAAsC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,iBAAiB,MAAM,KAAK,iBAAiB;AAEnD,QAAI,kBAAkB,eAAe;AACnC,aAAO,KAAK,oCAAoC;AAAA,QAC9C;AAAA,QACA;AAAA,MACF,CAAC;AACD;AAAA,IACF;AAGA,aAAS,IAAI,iBAAiB,GAAG,KAAK,eAAe,KAAK;AACxD,aAAO,KAAK,iCAAiC,CAAC,EAAE;AAEhD,WAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,CAAC;AAAA,IAChE;AAAA,EACF;AAAA,EAEA,MAAM,mBAAoC;AACxC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI;AACF,YAAM,SAAS,KAAK,GACjB,QAAQ,oDAAoD,EAC5D,IAAI;AACP,aAAO,QAAQ,WAAW;AAAA,IAC5B,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA,EAGA,MAAM,YAAY,OAAwC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,MAAM,YAAY,KAAK,WAAW;AAElD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAMF,EACC;AAAA,MACC;AAAA,MACA,MAAM;AAAA,MACN,MAAM,cAAc,KAAK;AAAA,MACzB,MAAM,mBAAmB;AAAA,MACzB,MAAM,SAAS;AAAA,MACf,MAAM;AAAA,MACN,MAAM;AAAA,MACN,MAAM,SAAS;AAAA,MACf,KAAK,UAAU,MAAM,UAAU,CAAC,CAAC;AAAA,MACjC,KAAK,UAAU,MAAM,WAAW,CAAC,CAAC;AAAA,MAClC,MAAM,eAAe;AAAA,MACrB,KAAK,UAAU,MAAM,eAAe,CAAC,CAAC;AAAA,IACxC;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,SAAS,SAAwC;AACrD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,MAAM,KAAK,GACd,QAAQ,yCAAyC,EACjD,IAAI,OAAO;AAEd,QAAI,CAAC,IAAK,QAAO;AAEjB,WAAO;AAAA,MACL,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD;AAAA,EACF;AAAA,EAEA,MAAM,YAAY,SAAiB,SAAwC;AACzE,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,SAAS,CAAC;AAChB,UAAM,SAAS,CAAC;AAEhB,QAAI,QAAQ,UAAU,QAAW;AAC/B,aAAO,KAAK,WAAW;AACvB,aAAO,KAAK,QAAQ,KAAK;AAAA,IAC3B;AAEA,QAAI,QAAQ,YAAY,QAAW;AACjC,aAAO,KAAK,aAAa;AACzB,aAAO,KAAK,KAAK,UAAU,QAAQ,OAAO,CAAC;AAAA,IAC7C;AAEA,QAAI,QAAQ,gBAAgB,QAAW;AACrC,aAAO,KAAK,iBAAiB;AAC7B,aAAO,KAAK,QAAQ,WAAW;AAAA,IACjC;AAEA,QAAI,QAAQ,gBAAgB,QAAW;AACrC,aAAO,KAAK,iBAAiB;AAC7B,aAAO,KAAK,KAAK,UAAU,QAAQ,WAAW,CAAC;AAAA,IACjD;AAEA,QAAI,QAAQ,cAAc,QAAW;AACnC,aAAO,KAAK,eAAe;AAC3B,aAAO,KAAK,QAAQ,SAAS;AAAA,IAC/B;AAEA,QAAI,OAAO,WAAW,EAAG;AAEzB,WAAO,KAAK,OAAO;AAEnB,SAAK,GACF;AAAA,MACC;AAAA,0BACkB,OAAO,KAAK,IAAI,CAAC;AAAA;AAAA,IAErC,EACC,IAAI,GAAG,MAAM;AAAA,EAClB;AAAA,EAEA,MAAM,YAAY,SAAgC;AAChD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAGtD,UAAM,KAAK,mBAAmB,OAAO;AACrC,UAAM,KAAK,kBAAkB,OAAO;AAEpC,SAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,OAAO;AAAA,EACtE;AAAA,EAEA,MAAM,gBAAgB,OAAkC;AACtD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,QAAQ;AACZ,UAAM,SAAS,CAAC;AAEhB,QAAI,OAAO;AACT,eAAS;AACT,aAAO,KAAK,KAAK;AAAA,IACnB;AAEA,aAAS;AAET,UAAM,OAAO,KAAK,GAAG,QAAQ,KAAK,EAAE,IAAI,GAAG,MAAM;AAEjD,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,WAAW,SAAiB,SAA8B;AAC9D,UAAM,KAAK,YAAY,SAAS;AAAA,MAC9B,OAAO;AAAA,MACP;AAAA,MACA,WAAW,KAAK,IAAI;AAAA,IACtB,CAAC;AAAA,EACH;AAAA;AAAA,EAGA,MAAM,YAAY,OAAwC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,MAAM,YAAY,KAAK,WAAW;AAElD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC;AAAA,MACC;AAAA,MACA,MAAM;AAAA,MACN,MAAM;AAAA,MACN,MAAM,OAAO;AAAA,MACb,MAAM;AAAA,MACN,KAAK,UAAU,MAAM,WAAW,CAAC,CAAC;AAAA,MAClC,MAAM,MAAM,KAAK,IAAI;AAAA,IACvB;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,eACJ,SACA,SACkB;AAClB,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,QAAQ;AACZ,aAAS,KAAK;AAAA,MACZ,SAAS,WAAW;AAAA,MACpB,SAAS;AAAA,IACX;AACA,aAAS,KAAK,iBAAiB,SAAS,OAAO,SAAS,MAAM;AAE9D,UAAM,OAAO,KAAK,GAAG,QAAQ,KAAK,EAAE,IAAI,OAAO;AAE/C,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,IACzC,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,kBAAkB,SAAgC;AACtD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,uCAAuC,EAAE,IAAI,OAAO;AAAA,EACtE;AAAA;AAAA,EAGA,MAAM,aAAa,QAA0C;AAC3D,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,WAAW,OAAO,aAAa,KAAK,WAAW;AAErD,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC;AAAA,MACC;AAAA,MACA,OAAO;AAAA,MACP,OAAO,cAAc,KAAK;AAAA,MAC1B,OAAO;AAAA,MACP,OAAO;AAAA,MACP,OAAO,YAAY;AAAA,MACnB,KAAK,UAAU,OAAO,YAAY,CAAC,CAAC;AAAA,IACtC;AAEF,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,OAAO,KAAK,GACf;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC,IAAI,OAAO;AAEd,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,UAAU,KAAK,MAAM,IAAI,YAAY,IAAI;AAAA,IAC3C,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,mBAAmB,SAAgC;AACvD,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,wCAAwC,EAAE,IAAI,OAAO;AAAA,EACvE;AAAA;AAAA,EAGA,MAAM,OACJ,SAC2C;AAC3C,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAGtD,UAAM,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAaZ,UAAM,SAAS,MAAM,CAAC,EAAE,KAAK,IAAI,QAAQ,KAAK,GAAG;AAEjD,QAAI,OAAO,KAAK,GAAG,QAAQ,GAAG,EAAE,IAAI,GAAG,MAAM;AAG7C,QAAI,QAAQ,gBAAgB;AAC1B,aAAO,KAAK,OAAO,CAAC,QAAQ,IAAI,SAAS,QAAQ,cAAc;AAAA,IACjE;AAGA,QAAI,QAAQ,SAAS,QAAQ,QAAQ;AACnC,YAAM,QAAQ,QAAQ,UAAU;AAChC,YAAM,MAAM,QAAQ,QAAQ,QAAQ,QAAQ,QAAQ,KAAK;AACzD,aAAO,KAAK,MAAM,OAAO,GAAG;AAAA,IAC9B;AAEA,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,GAAG;AAAA,MACH,QAAQ,KAAK,MAAM,IAAI,UAAU,IAAI;AAAA,MACrC,SAAS,KAAK,MAAM,IAAI,WAAW,IAAI;AAAA,MACvC,aAAa,KAAK,MAAM,IAAI,eAAe,IAAI;AAAA,IACjD,EAAE;AAAA,EACJ;AAAA,EAEA,MAAM,eACJ,WACA,SACgD;AAEhD,WAAO,KAAK,+CAA+C;AAC3D,WAAO,CAAC;AAAA,EACV;AAAA,EAEA,MAAM,aACJ,WACA,WACA,SAC2C;AAE3C,WAAO,KAAK,OAAO,EAAE,OAAO,WAAW,GAAG,QAAQ,CAAC;AAAA,EACrD;AAAA;AAAA,EAGA,MAAM,UACJ,OACA,SACgC;AAChC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,UAAU,QAAQ,QACrB;AAAA,MACC,CAAC,MACC,GAAG,EAAE,SAAS,IAAI,EAAE,KAAK,QAAQ,EAAE,SAAS,GAAG,EAAE,SAAS,IAAI,EAAE,KAAK,EAAE;AAAA,IAC3E,EACC,KAAK,IAAI;AAEZ,QAAI,MAAM,UAAU,QAAQ,QAAQ,KAAK,IAAI,CAAC,KAAK,OAAO,SAAS,KAAK;AACxE,WAAO,aAAa,QAAQ,QAAQ,KAAK,IAAI,CAAC;AAE9C,QAAI,QAAQ,QAAQ;AAClB,YAAM,gBAAgB,OAAO,QAAQ,QAAQ,MAAM,EAAE;AAAA,QACnD,CAAC,CAAC,KAAK,KAAK,MACV,GAAG,GAAG,IAAI,OAAO,UAAU,WAAW,MAAM,KAAK,GAAG;AAAA,MACxD;AACA,aAAO,WAAW,cAAc,KAAK,OAAO,CAAC;AAAA,IAC/C;AAEA,WAAO,KAAK,GACT,QAAQ,GAAG,EACX,IAAI,GAAG,OAAO,OAAO,QAAQ,UAAU,CAAC,CAAC,CAAC;AAAA,EAC/C;AAAA;AAAA,EAGA,MAAM,eAAe,WAOnB;AACA,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,MAAM;AAAA;AAAA;AAAA;AAKV,UAAM,SAAS,CAAC;AAChB,QAAI,WAAW;AACb,aAAO;AACP,aAAO;AAAA,QACL,KAAK,MAAM,UAAU,MAAM,QAAQ,IAAI,GAAI;AAAA,QAC3C,KAAK,MAAM,UAAU,IAAI,QAAQ,IAAI,GAAI;AAAA,MAC3C;AAAA,IACF;AAEA,WAAO;AAEP,UAAM,OAAO,KAAK,GAAG,QAAQ,GAAG,EAAE,IAAI,GAAG,MAAM;AAE/C,WAAO,KAAK,IAAI,CAAC,SAAS;AAAA,MACxB,SAAS,IAAI;AAAA,MACb,MAAM,IAAI;AAAA,MACV,WAAW,IAAI;AAAA,MACf,UAAU,IAAI,KAAK,IAAI,YAAY,GAAI;AAAA,IACzC,EAAE;AAAA,EACJ;AAAA;AAAA,EAGA,MAAM,YAAY,YAA4C;AAC5D,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,KAAK,cAAc,YAAY;AACnC,iBAAW,MAAM,YAAY;AAC3B,gBAAQ,GAAG,MAAM;AAAA,UACf,KAAK;AAEH,kBAAM,aAAa,OAAO,KAAK,GAAG,IAAI;AACtC,kBAAM,qBAAqB,WAAW,IAAI,MAAM,GAAG,EAAE,KAAK,GAAG;AAC7D,iBAAK,GAAI;AAAA,cACP,eAAe,GAAG,KAAK,KAAK,WAAW,KAAK,GAAG,CAAC,aAAa,kBAAkB;AAAA,YACjF,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,IAAI,CAAC;AAC/B;AAAA,UAEF,KAAK;AACH,kBAAM,aAAa,OAAO,KAAK,GAAG,IAAI,EACnC,IAAI,CAAC,MAAM,GAAG,CAAC,MAAM,EACrB,KAAK,GAAG;AACX,kBAAM,cAAc,KAAK,iBAAiB,GAAG,SAAS,CAAC,CAAC;AACxD,iBAAK,GAAI;AAAA,cACP,UAAU,GAAG,KAAK,QAAQ,UAAU,IAAI,WAAW;AAAA,YACrD,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,IAAI,GAAG,GAAG,OAAO,OAAO,GAAG,SAAS,CAAC,CAAC,CAAC;AACjE;AAAA,UAEF,KAAK;AACH,kBAAM,cAAc,KAAK,iBAAiB,GAAG,SAAS,CAAC,CAAC;AACxD,iBAAK,GAAI,QAAQ,eAAe,GAAG,KAAK,IAAI,WAAW,EAAE,EAAE;AAAA,cACzD,GAAG,OAAO,OAAO,GAAG,SAAS,CAAC,CAAC;AAAA,YACjC;AACA;AAAA,QACJ;AAAA,MACF;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,SAAwB;AAC5B,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,OAAO,QAAQ;AACvB,WAAO,KAAK,0BAA0B;AAAA,EACxC;AAAA,EAEA,MAAM,UAAyB;AAC7B,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,OAAO,SAAS;AACxB,WAAO,KAAK,0BAA0B;AAAA,EACxC;AAAA;AAAA,EAGA,MAAM,WAAmC;AACvC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,UAAM,QAAQ;AAAA,MACZ,aACE,KAAK,GACF,QAAQ,sCAAsC,EAC9C,IAAI,EACP;AAAA,MACF,cACE,KAAK,GACF;AAAA,QACC;AAAA,MACF,EACC,IAAI,EACP;AAAA,MACF,aACE,KAAK,GACF,QAAQ,sCAAsC,EAC9C,IAAI,EACP;AAAA,MACF,cACE,KAAK,GACF,QAAQ,uCAAuC,EAC/C,IAAI,EACP;AAAA,MACF,WAAW;AAAA,IACb;AAGA,QAAI;AACF,YAAM,YAAY,MAAM,GAAG,KAAK,KAAK,MAAM;AAC3C,YAAM,YAAY,UAAU;AAAA,IAC9B,QAAQ;AAAA,IAAC;AAET,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,gBAOJ;AAEA,WAAO,KAAK,sCAAsC;AAClD,WAAO,CAAC;AAAA,EACV;AAAA;AAAA,EAGA,MAAM,mBAAkC;AACtC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,OAAO,EAAE,IAAI;AAC7B,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,oBAAmC;AACvC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,QAAQ,EAAE,IAAI;AAC9B,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,sBAAqC;AACzC,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,SAAK,GAAG,QAAQ,UAAU,EAAE,IAAI;AAChC,SAAK,oBAAoB;AAAA,EAC3B;AAAA,EAEA,MAAM,cACJ,UACe;AACf,UAAM,KAAK,iBAAiB;AAE5B,QAAI;AACF,YAAM,SAAS,IAAI;AACnB,YAAM,KAAK,kBAAkB;AAAA,IAC/B,SAAS,OAAgB;AACvB,YAAM,KAAK,oBAAoB;AAC/B,YAAM;AAAA,IACR;AAAA,EACF;AAAA;AAAA,EAGA,MAAM,WACJ,QACA,QACiB;AACjB,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,WAAW,QAAQ;AACrB,YAAM,IAAI,MAAM,UAAU,MAAM,kCAAkC;AAAA,IACpE;AAEA,UAAM,OAA8B,CAAC;AAErC,eAAW,SAAS,QAAQ;AAC1B,WAAK,KAAK,IAAI,KAAK,GAAG,QAAQ,iBAAiB,KAAK,EAAE,EAAE,IAAI;AAAA,IAC9D;AAEA,WAAO,OAAO,KAAK,KAAK,UAAU,MAAM,MAAM,CAAC,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WACJ,MACA,QACA,SACe;AACf,QAAI,CAAC,KAAK,GAAI,OAAM,IAAI,MAAM,wBAAwB;AAEtD,QAAI,WAAW,QAAQ;AACrB,YAAM,IAAI,MAAM,UAAU,MAAM,kCAAkC;AAAA,IACpE;AAEA,UAAM,SAAS,KAAK,MAAM,KAAK,SAAS,CAAC;AAEzC,UAAM,KAAK,cAAc,YAAY;AACnC,iBAAW,CAAC,OAAO,IAAI,KAAK,OAAO,QAAQ,MAAM,GAAG;AAClD,YAAI,SAAS,UAAU;AACrB,eAAK,GAAI,QAAQ,eAAe,KAAK,EAAE,EAAE,IAAI;AAAA,QAC/C;AAEA,mBAAW,OAAO,MAAe;AAC/B,gBAAM,OAAO,OAAO,KAAK,GAAG;AAC5B,gBAAM,eAAe,KAAK,IAAI,MAAM,GAAG,EAAE,KAAK,GAAG;AAEjD,cAAI,SAAS,QAAQ;AACnB,kBAAM,UAAU,KAAK,IAAI,CAAC,MAAM,GAAG,CAAC,eAAe,CAAC,EAAE,EAAE,KAAK,GAAG;AAChE,iBAAK,GAAI;AAAA,cACP,eAAe,KAAK,KAAK,KAAK,KAAK,GAAG,CAAC,aAAa,YAAY;AAAA,2CACnC,OAAO;AAAA,YACtC,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,CAAC;AAAA,UAC7B,OAAO;AACL,iBAAK,GAAI;AAAA,cACP,eAAe,KAAK,KAAK,KAAK,KAAK,GAAG,CAAC,aAAa,YAAY;AAAA,YAClE,EAAE,IAAI,GAAG,OAAO,OAAO,GAAG,CAAC;AAAA,UAC7B;AAAA,QACF;AAAA,MACF;AAAA,IACF,CAAC;AAAA,EACH;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -49,7 +49,10 @@ class EnhancedHybridDigestGenerator extends HybridDigestGenerator {
|
|
|
49
49
|
timeSinceLastInput: now - this.lastInputTime
|
|
50
50
|
});
|
|
51
51
|
this.processQueue().catch((error) => {
|
|
52
|
-
logger.error(
|
|
52
|
+
logger.error(
|
|
53
|
+
"Error processing digest queue during idle",
|
|
54
|
+
error instanceof Error ? error : new Error(String(error))
|
|
55
|
+
);
|
|
53
56
|
});
|
|
54
57
|
}
|
|
55
58
|
}
|
|
@@ -76,7 +79,10 @@ class EnhancedHybridDigestGenerator extends HybridDigestGenerator {
|
|
|
76
79
|
});
|
|
77
80
|
this.prioritizeFrame(frameId);
|
|
78
81
|
this.processQueue().catch((error) => {
|
|
79
|
-
logger.error(
|
|
82
|
+
logger.error(
|
|
83
|
+
"Error processing digest on frame close",
|
|
84
|
+
error instanceof Error ? error : new Error(String(error))
|
|
85
|
+
);
|
|
80
86
|
});
|
|
81
87
|
}
|
|
82
88
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/core/digest/enhanced-hybrid-digest.ts"],
|
|
4
|
-
"sourcesContent": ["/**\n * Enhanced Hybrid Digest Generator for STA-96\n * Implements 60/40 split: deterministic extraction + AI-generated insights\n */\n\nimport Database from 'better-sqlite3';\nimport {\n HybridDigest,\n DeterministicDigest,\n AIGeneratedDigest,\n DigestConfig,\n DigestInput,\n DigestGenerationRequest,\n DigestQueueStats,\n DigestLLMProvider,\n DigestStatus,\n DEFAULT_DIGEST_CONFIG,\n} from './types.js';\nimport { HybridDigestGenerator } from './hybrid-digest-generator.js';\nimport { logger } from '../monitoring/logger.js';\n\n/**\n * Enhanced AI digest with 40% content\n */\nexport interface EnhancedAIDigest extends AIGeneratedDigest {\n /** Key decisions made and their reasoning */\n keyDecisions?: string[];\n /** Important insights discovered */\n insights?: string[];\n /** Suggested next steps */\n nextSteps?: string[];\n /** Patterns detected */\n patterns?: string[];\n /** Technical debt or improvements identified */\n technicalDebt?: string[];\n}\n\n/**\n * Idle detection configuration\n */\nexport interface IdleDetectionConfig {\n /** No tool calls threshold (ms) */\n noToolCallThreshold: number;\n /** No user input threshold (ms) */\n noInputThreshold: number;\n /** Frame closed immediately triggers processing */\n processOnFrameClose: boolean;\n /** Check interval (ms) */\n checkInterval: number;\n}\n\nconst DEFAULT_IDLE_CONFIG: IdleDetectionConfig = {\n noToolCallThreshold: 30000, // 30 seconds\n noInputThreshold: 60000, // 60 seconds\n processOnFrameClose: true,\n checkInterval: 10000, // Check every 10 seconds\n};\n\n/**\n * Enhanced Hybrid Digest Generator\n * Implements 60% deterministic + 40% AI insights\n */\nexport class EnhancedHybridDigestGenerator extends HybridDigestGenerator {\n private idleConfig: IdleDetectionConfig;\n private lastToolCallTime: number = Date.now();\n private lastInputTime: number = Date.now();\n private idleCheckInterval?: NodeJS.Timeout;\n private activeFrames = new Set<string>();\n\n constructor(\n db: Database.Database,\n config: Partial<DigestConfig> = {},\n llmProvider?: DigestLLMProvider,\n idleConfig: Partial<IdleDetectionConfig> = {}\n ) {\n // Update config to reflect 60/40 split\n const enhancedConfig = {\n ...config,\n maxTokens: config.maxTokens || 200, // Keep under 200 tokens as per requirement\n enableAIGeneration: config.enableAIGeneration ?? true,\n };\n\n super(db, enhancedConfig, llmProvider);\n this.idleConfig = { ...DEFAULT_IDLE_CONFIG, ...idleConfig };\n this.startIdleDetection();\n }\n\n /**\n * Start idle detection monitoring\n */\n private startIdleDetection(): void {\n this.idleCheckInterval = setInterval(() => {\n this.checkIdleState();\n }, this.idleConfig.checkInterval);\n }\n\n /**\n * Check if system is idle and trigger processing\n */\n private checkIdleState(): void {\n const now = Date.now();\n\n // Check for idle based on tool calls\n const toolCallIdle =\n now - this.lastToolCallTime > this.idleConfig.noToolCallThreshold;\n\n // Check for idle based on user input\n const inputIdle =\n now - this.lastInputTime > this.idleConfig.noInputThreshold;\n\n if (toolCallIdle || inputIdle) {\n logger.debug('Idle state detected, triggering digest processing', {\n toolCallIdle,\n inputIdle,\n timeSinceLastToolCall: now - this.lastToolCallTime,\n timeSinceLastInput: now - this.lastInputTime,\n });\n\n this.processQueue().catch((error) => {\n logger.error('Error processing digest queue during idle', error instanceof Error ? error : new Error(String(error)));\n });\n }\n }\n\n /**\n * Record tool call activity\n */\n public recordToolCall(): void {\n this.lastToolCallTime = Date.now();\n }\n\n /**\n * Record user input activity\n */\n public recordUserInput(): void {\n this.lastInputTime = Date.now();\n }\n\n /**\n * Handle frame closure - immediately trigger digest if configured\n */\n public onFrameClosed(frameId: string): void {\n this.activeFrames.delete(frameId);\n\n if (this.idleConfig.processOnFrameClose) {\n logger.info('Frame closed, triggering immediate digest processing', {\n frameId,\n });\n\n // Process this specific frame with high priority\n this.prioritizeFrame(frameId);\n this.processQueue().catch((error) => {\n logger.error('Error processing digest on frame close', error instanceof Error ? error : new Error(String(error)));\n });\n }\n }\n\n /**\n * Handle frame opened\n */\n public onFrameOpened(frameId: string): void {\n this.activeFrames.add(frameId);\n }\n\n /**\n * Prioritize a specific frame for processing\n */\n private prioritizeFrame(frameId: string): void {\n try {\n this.db\n .prepare(\n `\n UPDATE digest_queue \n SET priority = 'high', updated_at = unixepoch()\n WHERE frame_id = ? AND status = 'pending'\n `\n )\n .run(frameId);\n } catch (error: unknown) {\n logger.error('Failed to prioritize frame', error as Error);\n }\n }\n\n /**\n * Enhanced AI generation with 40% content\n */\n protected async generateEnhancedAI(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): Promise<EnhancedAIDigest> {\n if (!this.llmProvider) {\n throw new Error('No LLM provider configured');\n }\n\n // Build enhanced prompt for 40% AI content\n const prompt = this.buildEnhancedPrompt(input, deterministic);\n\n // Generate with LLM\n const response = await this.llmProvider.generateSummary(\n input,\n deterministic,\n this.config.maxTokens\n );\n\n // Parse and structure the enhanced response\n const enhanced: EnhancedAIDigest = {\n ...response,\n keyDecisions: this.extractKeyDecisions(response),\n insights: this.extractInsights(response),\n nextSteps: this.extractNextSteps(response),\n patterns: this.detectPatterns(input, deterministic),\n technicalDebt: this.identifyTechnicalDebt(input, deterministic),\n };\n\n return enhanced;\n }\n\n /**\n * Build enhanced prompt for AI generation\n */\n private buildEnhancedPrompt(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string {\n const parts: string[] = [\n `Analyze this development frame and provide insights (max ${this.config.maxTokens} tokens):`,\n '',\n `Frame: ${input.frame.name} (${input.frame.type})`,\n `Duration: ${deterministic.durationSeconds}s`,\n `Files Modified: ${deterministic.filesModified.length}`,\n `Tool Calls: ${deterministic.toolCallCount}`,\n `Errors: ${deterministic.errorsEncountered.length}`,\n '',\n 'Provide:',\n '1. Key decisions made and why (2-3 items)',\n '2. Important insights or learnings (1-2 items)',\n '3. Suggested next steps (2-3 items)',\n '4. Any patterns or anti-patterns observed',\n '5. Technical debt or improvements needed',\n '',\n 'Be concise and actionable. Focus on value, not description.',\n ];\n\n return parts.join('\\n');\n }\n\n /**\n * Extract key decisions from AI response\n */\n private extractKeyDecisions(response: AIGeneratedDigest): string[] {\n // This would parse the AI response for decision patterns\n // For now, return placeholder\n return [];\n }\n\n /**\n * Extract insights from AI response\n */\n private extractInsights(response: AIGeneratedDigest): string[] {\n const insights: string[] = [];\n if (response.insight) {\n insights.push(response.insight);\n }\n return insights;\n }\n\n /**\n * Extract next steps from AI response\n */\n private extractNextSteps(response: AIGeneratedDigest): string[] {\n // Parse AI response for next steps\n return [];\n }\n\n /**\n * Detect patterns in the frame activity\n */\n private detectPatterns(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string[] {\n const patterns: string[] = [];\n\n // Detect retry patterns\n if (deterministic.errorsEncountered.some((e) => e.count > 2)) {\n patterns.push('Multiple retry attempts detected');\n }\n\n // Detect test-driven development\n const hasTests = deterministic.testsRun.length > 0;\n const hasCodeChanges = deterministic.filesModified.some(\n (f) => f.operation === 'modify' && !f.path.includes('test')\n );\n if (hasTests && hasCodeChanges) {\n patterns.push('Test-driven development pattern observed');\n }\n\n // Detect refactoring patterns\n const manyFileChanges = deterministic.filesModified.length > 5;\n const noNewFiles = !deterministic.filesModified.some(\n (f) => f.operation === 'create'\n );\n if (manyFileChanges && noNewFiles) {\n patterns.push('Refactoring pattern detected');\n }\n\n return patterns;\n }\n\n /**\n * Identify technical debt\n */\n private identifyTechnicalDebt(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string[] {\n const debt: string[] = [];\n\n // Check for missing tests\n if (\n deterministic.filesModified.length > 3 &&\n deterministic.testsRun.length === 0\n ) {\n debt.push('Code changes without corresponding tests');\n }\n\n // Check for unresolved errors\n const unresolvedErrors = deterministic.errorsEncountered.filter(\n (e) => !e.resolved\n );\n if (unresolvedErrors.length > 0) {\n debt.push(`${unresolvedErrors.length} unresolved errors remain`);\n }\n\n // Check for TODOs in decisions\n if (deterministic.decisions.some((d) => d.toLowerCase().includes('todo'))) {\n debt.push('TODOs added to codebase');\n }\n\n return debt;\n }\n\n /**\n * Generate digest with 60/40 split\n */\n public generateDigest(input: DigestInput): HybridDigest {\n // Record activity\n this.recordToolCall();\n\n // Generate base digest (60% deterministic)\n const digest = super.generateDigest(input);\n\n // Ensure AI generation is queued for 40% content\n if (this.config.enableAIGeneration && this.llmProvider) {\n digest.status = 'ai_pending';\n }\n\n return digest;\n }\n\n /**\n * Handle interruption gracefully\n */\n public handleInterruption(): void {\n logger.info('User activity detected, pausing digest processing');\n\n // Update timestamps\n this.recordUserInput();\n this.recordToolCall();\n\n // Don't stop processing entirely, just deprioritize\n this.db\n .prepare(\n `\n UPDATE digest_queue \n SET priority = 'low'\n WHERE status = 'processing'\n `\n )\n .run();\n }\n\n /**\n * Get idle status\n */\n public getIdleStatus(): {\n isIdle: boolean;\n timeSinceLastToolCall: number;\n timeSinceLastInput: number;\n activeFrames: number;\n } {\n const now = Date.now();\n return {\n isIdle:\n now - this.lastToolCallTime > this.idleConfig.noToolCallThreshold ||\n now - this.lastInputTime > this.idleConfig.noInputThreshold,\n timeSinceLastToolCall: now - this.lastToolCallTime,\n timeSinceLastInput: now - this.lastInputTime,\n activeFrames: this.activeFrames.size,\n };\n }\n\n /**\n * Cleanup on shutdown\n */\n public shutdown(): void {\n if (this.idleCheckInterval) {\n clearInterval(this.idleCheckInterval);\n }\n }\n}\n"],
|
|
5
|
-
"mappings": "AAkBA,SAAS,6BAA6B;AACtC,SAAS,cAAc;AAgCvB,MAAM,sBAA2C;AAAA,EAC/C,qBAAqB;AAAA;AAAA,EACrB,kBAAkB;AAAA;AAAA,EAClB,qBAAqB;AAAA,EACrB,eAAe;AAAA;AACjB;AAMO,MAAM,sCAAsC,sBAAsB;AAAA,EAC/D;AAAA,EACA,mBAA2B,KAAK,IAAI;AAAA,EACpC,gBAAwB,KAAK,IAAI;AAAA,EACjC;AAAA,EACA,eAAe,oBAAI,IAAY;AAAA,EAEvC,YACE,IACA,SAAgC,CAAC,GACjC,aACA,aAA2C,CAAC,GAC5C;AAEA,UAAM,iBAAiB;AAAA,MACrB,GAAG;AAAA,MACH,WAAW,OAAO,aAAa;AAAA;AAAA,MAC/B,oBAAoB,OAAO,sBAAsB;AAAA,IACnD;AAEA,UAAM,IAAI,gBAAgB,WAAW;AACrC,SAAK,aAAa,EAAE,GAAG,qBAAqB,GAAG,WAAW;AAC1D,SAAK,mBAAmB;AAAA,EAC1B;AAAA;AAAA;AAAA;AAAA,EAKQ,qBAA2B;AACjC,SAAK,oBAAoB,YAAY,MAAM;AACzC,WAAK,eAAe;AAAA,IACtB,GAAG,KAAK,WAAW,aAAa;AAAA,EAClC;AAAA;AAAA;AAAA;AAAA,EAKQ,iBAAuB;AAC7B,UAAM,MAAM,KAAK,IAAI;AAGrB,UAAM,eACJ,MAAM,KAAK,mBAAmB,KAAK,WAAW;AAGhD,UAAM,YACJ,MAAM,KAAK,gBAAgB,KAAK,WAAW;AAE7C,QAAI,gBAAgB,WAAW;AAC7B,aAAO,MAAM,qDAAqD;AAAA,QAChE;AAAA,QACA;AAAA,QACA,uBAAuB,MAAM,KAAK;AAAA,QAClC,oBAAoB,MAAM,KAAK;AAAA,MACjC,CAAC;AAED,WAAK,aAAa,EAAE,MAAM,CAAC,UAAU;AACnC,eAAO,
|
|
4
|
+
"sourcesContent": ["/**\n * Enhanced Hybrid Digest Generator for STA-96\n * Implements 60/40 split: deterministic extraction + AI-generated insights\n */\n\nimport Database from 'better-sqlite3';\nimport {\n HybridDigest,\n DeterministicDigest,\n AIGeneratedDigest,\n DigestConfig,\n DigestInput,\n DigestGenerationRequest,\n DigestQueueStats,\n DigestLLMProvider,\n DigestStatus,\n DEFAULT_DIGEST_CONFIG,\n} from './types.js';\nimport { HybridDigestGenerator } from './hybrid-digest-generator.js';\nimport { logger } from '../monitoring/logger.js';\n\n/**\n * Enhanced AI digest with 40% content\n */\nexport interface EnhancedAIDigest extends AIGeneratedDigest {\n /** Key decisions made and their reasoning */\n keyDecisions?: string[];\n /** Important insights discovered */\n insights?: string[];\n /** Suggested next steps */\n nextSteps?: string[];\n /** Patterns detected */\n patterns?: string[];\n /** Technical debt or improvements identified */\n technicalDebt?: string[];\n}\n\n/**\n * Idle detection configuration\n */\nexport interface IdleDetectionConfig {\n /** No tool calls threshold (ms) */\n noToolCallThreshold: number;\n /** No user input threshold (ms) */\n noInputThreshold: number;\n /** Frame closed immediately triggers processing */\n processOnFrameClose: boolean;\n /** Check interval (ms) */\n checkInterval: number;\n}\n\nconst DEFAULT_IDLE_CONFIG: IdleDetectionConfig = {\n noToolCallThreshold: 30000, // 30 seconds\n noInputThreshold: 60000, // 60 seconds\n processOnFrameClose: true,\n checkInterval: 10000, // Check every 10 seconds\n};\n\n/**\n * Enhanced Hybrid Digest Generator\n * Implements 60% deterministic + 40% AI insights\n */\nexport class EnhancedHybridDigestGenerator extends HybridDigestGenerator {\n private idleConfig: IdleDetectionConfig;\n private lastToolCallTime: number = Date.now();\n private lastInputTime: number = Date.now();\n private idleCheckInterval?: NodeJS.Timeout;\n private activeFrames = new Set<string>();\n\n constructor(\n db: Database.Database,\n config: Partial<DigestConfig> = {},\n llmProvider?: DigestLLMProvider,\n idleConfig: Partial<IdleDetectionConfig> = {}\n ) {\n // Update config to reflect 60/40 split\n const enhancedConfig = {\n ...config,\n maxTokens: config.maxTokens || 200, // Keep under 200 tokens as per requirement\n enableAIGeneration: config.enableAIGeneration ?? true,\n };\n\n super(db, enhancedConfig, llmProvider);\n this.idleConfig = { ...DEFAULT_IDLE_CONFIG, ...idleConfig };\n this.startIdleDetection();\n }\n\n /**\n * Start idle detection monitoring\n */\n private startIdleDetection(): void {\n this.idleCheckInterval = setInterval(() => {\n this.checkIdleState();\n }, this.idleConfig.checkInterval);\n }\n\n /**\n * Check if system is idle and trigger processing\n */\n private checkIdleState(): void {\n const now = Date.now();\n\n // Check for idle based on tool calls\n const toolCallIdle =\n now - this.lastToolCallTime > this.idleConfig.noToolCallThreshold;\n\n // Check for idle based on user input\n const inputIdle =\n now - this.lastInputTime > this.idleConfig.noInputThreshold;\n\n if (toolCallIdle || inputIdle) {\n logger.debug('Idle state detected, triggering digest processing', {\n toolCallIdle,\n inputIdle,\n timeSinceLastToolCall: now - this.lastToolCallTime,\n timeSinceLastInput: now - this.lastInputTime,\n });\n\n this.processQueue().catch((error) => {\n logger.error(\n 'Error processing digest queue during idle',\n error instanceof Error ? error : new Error(String(error))\n );\n });\n }\n }\n\n /**\n * Record tool call activity\n */\n public recordToolCall(): void {\n this.lastToolCallTime = Date.now();\n }\n\n /**\n * Record user input activity\n */\n public recordUserInput(): void {\n this.lastInputTime = Date.now();\n }\n\n /**\n * Handle frame closure - immediately trigger digest if configured\n */\n public onFrameClosed(frameId: string): void {\n this.activeFrames.delete(frameId);\n\n if (this.idleConfig.processOnFrameClose) {\n logger.info('Frame closed, triggering immediate digest processing', {\n frameId,\n });\n\n // Process this specific frame with high priority\n this.prioritizeFrame(frameId);\n this.processQueue().catch((error) => {\n logger.error(\n 'Error processing digest on frame close',\n error instanceof Error ? error : new Error(String(error))\n );\n });\n }\n }\n\n /**\n * Handle frame opened\n */\n public onFrameOpened(frameId: string): void {\n this.activeFrames.add(frameId);\n }\n\n /**\n * Prioritize a specific frame for processing\n */\n private prioritizeFrame(frameId: string): void {\n try {\n this.db\n .prepare(\n `\n UPDATE digest_queue \n SET priority = 'high', updated_at = unixepoch()\n WHERE frame_id = ? AND status = 'pending'\n `\n )\n .run(frameId);\n } catch (error: unknown) {\n logger.error('Failed to prioritize frame', error as Error);\n }\n }\n\n /**\n * Enhanced AI generation with 40% content\n */\n protected async generateEnhancedAI(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): Promise<EnhancedAIDigest> {\n if (!this.llmProvider) {\n throw new Error('No LLM provider configured');\n }\n\n // Build enhanced prompt for 40% AI content\n const prompt = this.buildEnhancedPrompt(input, deterministic);\n\n // Generate with LLM\n const response = await this.llmProvider.generateSummary(\n input,\n deterministic,\n this.config.maxTokens\n );\n\n // Parse and structure the enhanced response\n const enhanced: EnhancedAIDigest = {\n ...response,\n keyDecisions: this.extractKeyDecisions(response),\n insights: this.extractInsights(response),\n nextSteps: this.extractNextSteps(response),\n patterns: this.detectPatterns(input, deterministic),\n technicalDebt: this.identifyTechnicalDebt(input, deterministic),\n };\n\n return enhanced;\n }\n\n /**\n * Build enhanced prompt for AI generation\n */\n private buildEnhancedPrompt(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string {\n const parts: string[] = [\n `Analyze this development frame and provide insights (max ${this.config.maxTokens} tokens):`,\n '',\n `Frame: ${input.frame.name} (${input.frame.type})`,\n `Duration: ${deterministic.durationSeconds}s`,\n `Files Modified: ${deterministic.filesModified.length}`,\n `Tool Calls: ${deterministic.toolCallCount}`,\n `Errors: ${deterministic.errorsEncountered.length}`,\n '',\n 'Provide:',\n '1. Key decisions made and why (2-3 items)',\n '2. Important insights or learnings (1-2 items)',\n '3. Suggested next steps (2-3 items)',\n '4. Any patterns or anti-patterns observed',\n '5. Technical debt or improvements needed',\n '',\n 'Be concise and actionable. Focus on value, not description.',\n ];\n\n return parts.join('\\n');\n }\n\n /**\n * Extract key decisions from AI response\n */\n private extractKeyDecisions(response: AIGeneratedDigest): string[] {\n // This would parse the AI response for decision patterns\n // For now, return placeholder\n return [];\n }\n\n /**\n * Extract insights from AI response\n */\n private extractInsights(response: AIGeneratedDigest): string[] {\n const insights: string[] = [];\n if (response.insight) {\n insights.push(response.insight);\n }\n return insights;\n }\n\n /**\n * Extract next steps from AI response\n */\n private extractNextSteps(response: AIGeneratedDigest): string[] {\n // Parse AI response for next steps\n return [];\n }\n\n /**\n * Detect patterns in the frame activity\n */\n private detectPatterns(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string[] {\n const patterns: string[] = [];\n\n // Detect retry patterns\n if (deterministic.errorsEncountered.some((e) => e.count > 2)) {\n patterns.push('Multiple retry attempts detected');\n }\n\n // Detect test-driven development\n const hasTests = deterministic.testsRun.length > 0;\n const hasCodeChanges = deterministic.filesModified.some(\n (f) => f.operation === 'modify' && !f.path.includes('test')\n );\n if (hasTests && hasCodeChanges) {\n patterns.push('Test-driven development pattern observed');\n }\n\n // Detect refactoring patterns\n const manyFileChanges = deterministic.filesModified.length > 5;\n const noNewFiles = !deterministic.filesModified.some(\n (f) => f.operation === 'create'\n );\n if (manyFileChanges && noNewFiles) {\n patterns.push('Refactoring pattern detected');\n }\n\n return patterns;\n }\n\n /**\n * Identify technical debt\n */\n private identifyTechnicalDebt(\n input: DigestInput,\n deterministic: DeterministicDigest\n ): string[] {\n const debt: string[] = [];\n\n // Check for missing tests\n if (\n deterministic.filesModified.length > 3 &&\n deterministic.testsRun.length === 0\n ) {\n debt.push('Code changes without corresponding tests');\n }\n\n // Check for unresolved errors\n const unresolvedErrors = deterministic.errorsEncountered.filter(\n (e) => !e.resolved\n );\n if (unresolvedErrors.length > 0) {\n debt.push(`${unresolvedErrors.length} unresolved errors remain`);\n }\n\n // Check for TODOs in decisions\n if (deterministic.decisions.some((d) => d.toLowerCase().includes('todo'))) {\n debt.push('TODOs added to codebase');\n }\n\n return debt;\n }\n\n /**\n * Generate digest with 60/40 split\n */\n public generateDigest(input: DigestInput): HybridDigest {\n // Record activity\n this.recordToolCall();\n\n // Generate base digest (60% deterministic)\n const digest = super.generateDigest(input);\n\n // Ensure AI generation is queued for 40% content\n if (this.config.enableAIGeneration && this.llmProvider) {\n digest.status = 'ai_pending';\n }\n\n return digest;\n }\n\n /**\n * Handle interruption gracefully\n */\n public handleInterruption(): void {\n logger.info('User activity detected, pausing digest processing');\n\n // Update timestamps\n this.recordUserInput();\n this.recordToolCall();\n\n // Don't stop processing entirely, just deprioritize\n this.db\n .prepare(\n `\n UPDATE digest_queue \n SET priority = 'low'\n WHERE status = 'processing'\n `\n )\n .run();\n }\n\n /**\n * Get idle status\n */\n public getIdleStatus(): {\n isIdle: boolean;\n timeSinceLastToolCall: number;\n timeSinceLastInput: number;\n activeFrames: number;\n } {\n const now = Date.now();\n return {\n isIdle:\n now - this.lastToolCallTime > this.idleConfig.noToolCallThreshold ||\n now - this.lastInputTime > this.idleConfig.noInputThreshold,\n timeSinceLastToolCall: now - this.lastToolCallTime,\n timeSinceLastInput: now - this.lastInputTime,\n activeFrames: this.activeFrames.size,\n };\n }\n\n /**\n * Cleanup on shutdown\n */\n public shutdown(): void {\n if (this.idleCheckInterval) {\n clearInterval(this.idleCheckInterval);\n }\n }\n}\n"],
|
|
5
|
+
"mappings": "AAkBA,SAAS,6BAA6B;AACtC,SAAS,cAAc;AAgCvB,MAAM,sBAA2C;AAAA,EAC/C,qBAAqB;AAAA;AAAA,EACrB,kBAAkB;AAAA;AAAA,EAClB,qBAAqB;AAAA,EACrB,eAAe;AAAA;AACjB;AAMO,MAAM,sCAAsC,sBAAsB;AAAA,EAC/D;AAAA,EACA,mBAA2B,KAAK,IAAI;AAAA,EACpC,gBAAwB,KAAK,IAAI;AAAA,EACjC;AAAA,EACA,eAAe,oBAAI,IAAY;AAAA,EAEvC,YACE,IACA,SAAgC,CAAC,GACjC,aACA,aAA2C,CAAC,GAC5C;AAEA,UAAM,iBAAiB;AAAA,MACrB,GAAG;AAAA,MACH,WAAW,OAAO,aAAa;AAAA;AAAA,MAC/B,oBAAoB,OAAO,sBAAsB;AAAA,IACnD;AAEA,UAAM,IAAI,gBAAgB,WAAW;AACrC,SAAK,aAAa,EAAE,GAAG,qBAAqB,GAAG,WAAW;AAC1D,SAAK,mBAAmB;AAAA,EAC1B;AAAA;AAAA;AAAA;AAAA,EAKQ,qBAA2B;AACjC,SAAK,oBAAoB,YAAY,MAAM;AACzC,WAAK,eAAe;AAAA,IACtB,GAAG,KAAK,WAAW,aAAa;AAAA,EAClC;AAAA;AAAA;AAAA;AAAA,EAKQ,iBAAuB;AAC7B,UAAM,MAAM,KAAK,IAAI;AAGrB,UAAM,eACJ,MAAM,KAAK,mBAAmB,KAAK,WAAW;AAGhD,UAAM,YACJ,MAAM,KAAK,gBAAgB,KAAK,WAAW;AAE7C,QAAI,gBAAgB,WAAW;AAC7B,aAAO,MAAM,qDAAqD;AAAA,QAChE;AAAA,QACA;AAAA,QACA,uBAAuB,MAAM,KAAK;AAAA,QAClC,oBAAoB,MAAM,KAAK;AAAA,MACjC,CAAC;AAED,WAAK,aAAa,EAAE,MAAM,CAAC,UAAU;AACnC,eAAO;AAAA,UACL;AAAA,UACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QAC1D;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKO,iBAAuB;AAC5B,SAAK,mBAAmB,KAAK,IAAI;AAAA,EACnC;AAAA;AAAA;AAAA;AAAA,EAKO,kBAAwB;AAC7B,SAAK,gBAAgB,KAAK,IAAI;AAAA,EAChC;AAAA;AAAA;AAAA;AAAA,EAKO,cAAc,SAAuB;AAC1C,SAAK,aAAa,OAAO,OAAO;AAEhC,QAAI,KAAK,WAAW,qBAAqB;AACvC,aAAO,KAAK,wDAAwD;AAAA,QAClE;AAAA,MACF,CAAC;AAGD,WAAK,gBAAgB,OAAO;AAC5B,WAAK,aAAa,EAAE,MAAM,CAAC,UAAU;AACnC,eAAO;AAAA,UACL;AAAA,UACA,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,QAC1D;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKO,cAAc,SAAuB;AAC1C,SAAK,aAAa,IAAI,OAAO;AAAA,EAC/B;AAAA;AAAA;AAAA;AAAA,EAKQ,gBAAgB,SAAuB;AAC7C,QAAI;AACF,WAAK,GACF;AAAA,QACC;AAAA;AAAA;AAAA;AAAA;AAAA,MAKF,EACC,IAAI,OAAO;AAAA,IAChB,SAAS,OAAgB;AACvB,aAAO,MAAM,8BAA8B,KAAc;AAAA,IAC3D;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAgB,mBACd,OACA,eAC2B;AAC3B,QAAI,CAAC,KAAK,aAAa;AACrB,YAAM,IAAI,MAAM,4BAA4B;AAAA,IAC9C;AAGA,UAAM,SAAS,KAAK,oBAAoB,OAAO,aAAa;AAG5D,UAAM,WAAW,MAAM,KAAK,YAAY;AAAA,MACtC;AAAA,MACA;AAAA,MACA,KAAK,OAAO;AAAA,IACd;AAGA,UAAM,WAA6B;AAAA,MACjC,GAAG;AAAA,MACH,cAAc,KAAK,oBAAoB,QAAQ;AAAA,MAC/C,UAAU,KAAK,gBAAgB,QAAQ;AAAA,MACvC,WAAW,KAAK,iBAAiB,QAAQ;AAAA,MACzC,UAAU,KAAK,eAAe,OAAO,aAAa;AAAA,MAClD,eAAe,KAAK,sBAAsB,OAAO,aAAa;AAAA,IAChE;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBACN,OACA,eACQ;AACR,UAAM,QAAkB;AAAA,MACtB,4DAA4D,KAAK,OAAO,SAAS;AAAA,MACjF;AAAA,MACA,UAAU,MAAM,MAAM,IAAI,KAAK,MAAM,MAAM,IAAI;AAAA,MAC/C,aAAa,cAAc,eAAe;AAAA,MAC1C,mBAAmB,cAAc,cAAc,MAAM;AAAA,MACrD,eAAe,cAAc,aAAa;AAAA,MAC1C,WAAW,cAAc,kBAAkB,MAAM;AAAA,MACjD;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAEA,WAAO,MAAM,KAAK,IAAI;AAAA,EACxB;AAAA;AAAA;AAAA;AAAA,EAKQ,oBAAoB,UAAuC;AAGjE,WAAO,CAAC;AAAA,EACV;AAAA;AAAA;AAAA;AAAA,EAKQ,gBAAgB,UAAuC;AAC7D,UAAM,WAAqB,CAAC;AAC5B,QAAI,SAAS,SAAS;AACpB,eAAS,KAAK,SAAS,OAAO;AAAA,IAChC;AACA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,iBAAiB,UAAuC;AAE9D,WAAO,CAAC;AAAA,EACV;AAAA;AAAA;AAAA;AAAA,EAKQ,eACN,OACA,eACU;AACV,UAAM,WAAqB,CAAC;AAG5B,QAAI,cAAc,kBAAkB,KAAK,CAAC,MAAM,EAAE,QAAQ,CAAC,GAAG;AAC5D,eAAS,KAAK,kCAAkC;AAAA,IAClD;AAGA,UAAM,WAAW,cAAc,SAAS,SAAS;AACjD,UAAM,iBAAiB,cAAc,cAAc;AAAA,MACjD,CAAC,MAAM,EAAE,cAAc,YAAY,CAAC,EAAE,KAAK,SAAS,MAAM;AAAA,IAC5D;AACA,QAAI,YAAY,gBAAgB;AAC9B,eAAS,KAAK,0CAA0C;AAAA,IAC1D;AAGA,UAAM,kBAAkB,cAAc,cAAc,SAAS;AAC7D,UAAM,aAAa,CAAC,cAAc,cAAc;AAAA,MAC9C,CAAC,MAAM,EAAE,cAAc;AAAA,IACzB;AACA,QAAI,mBAAmB,YAAY;AACjC,eAAS,KAAK,8BAA8B;AAAA,IAC9C;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,sBACN,OACA,eACU;AACV,UAAM,OAAiB,CAAC;AAGxB,QACE,cAAc,cAAc,SAAS,KACrC,cAAc,SAAS,WAAW,GAClC;AACA,WAAK,KAAK,0CAA0C;AAAA,IACtD;AAGA,UAAM,mBAAmB,cAAc,kBAAkB;AAAA,MACvD,CAAC,MAAM,CAAC,EAAE;AAAA,IACZ;AACA,QAAI,iBAAiB,SAAS,GAAG;AAC/B,WAAK,KAAK,GAAG,iBAAiB,MAAM,2BAA2B;AAAA,IACjE;AAGA,QAAI,cAAc,UAAU,KAAK,CAAC,MAAM,EAAE,YAAY,EAAE,SAAS,MAAM,CAAC,GAAG;AACzE,WAAK,KAAK,yBAAyB;AAAA,IACrC;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKO,eAAe,OAAkC;AAEtD,SAAK,eAAe;AAGpB,UAAM,SAAS,MAAM,eAAe,KAAK;AAGzC,QAAI,KAAK,OAAO,sBAAsB,KAAK,aAAa;AACtD,aAAO,SAAS;AAAA,IAClB;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKO,qBAA2B;AAChC,WAAO,KAAK,mDAAmD;AAG/D,SAAK,gBAAgB;AACrB,SAAK,eAAe;AAGpB,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EACC,IAAI;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKO,gBAKL;AACA,UAAM,MAAM,KAAK,IAAI;AACrB,WAAO;AAAA,MACL,QACE,MAAM,KAAK,mBAAmB,KAAK,WAAW,uBAC9C,MAAM,KAAK,gBAAgB,KAAK,WAAW;AAAA,MAC7C,uBAAuB,MAAM,KAAK;AAAA,MAClC,oBAAoB,MAAM,KAAK;AAAA,MAC/B,cAAc,KAAK,aAAa;AAAA,IAClC;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKO,WAAiB;AACtB,QAAI,KAAK,mBAAmB;AAC1B,oBAAc,KAAK,iBAAiB;AAAA,IACtC;AAAA,EACF;AACF;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -46,7 +46,12 @@ async function retry(fn, options = {}) {
|
|
|
46
46
|
if (!isRetryableError(error) || attempt === maxAttempts) {
|
|
47
47
|
throw error;
|
|
48
48
|
}
|
|
49
|
-
const delay = calculateBackoff(
|
|
49
|
+
const delay = calculateBackoff(
|
|
50
|
+
attempt,
|
|
51
|
+
initialDelay,
|
|
52
|
+
maxDelay,
|
|
53
|
+
backoffFactor
|
|
54
|
+
);
|
|
50
55
|
logger.warn(`Retry attempt ${attempt}/${maxAttempts} after ${delay}ms`, {
|
|
51
56
|
error: getErrorMessage(error),
|
|
52
57
|
attempt,
|
|
@@ -199,7 +204,9 @@ async function withTimeout(fn, timeoutMs, timeoutMessage) {
|
|
|
199
204
|
new Promise(
|
|
200
205
|
(_, reject) => setTimeout(
|
|
201
206
|
() => reject(
|
|
202
|
-
new Error(
|
|
207
|
+
new Error(
|
|
208
|
+
timeoutMessage ?? `Operation timed out after ${timeoutMs}ms`
|
|
209
|
+
)
|
|
203
210
|
),
|
|
204
211
|
timeoutMs
|
|
205
212
|
)
|