@fml-inc/panopticon 0.1.1 → 0.1.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (63) hide show
  1. package/dist/api/client.d.ts +2 -2
  2. package/dist/api/client.js +1 -1
  3. package/dist/{chunk-KLXRBD4N.js → chunk-2PMFNSGA.js} +6 -5
  4. package/dist/{chunk-KLXRBD4N.js.map → chunk-2PMFNSGA.js.map} +1 -1
  5. package/dist/{chunk-3BUJ7URA.js → chunk-3ILOOWUF.js} +66 -2
  6. package/dist/chunk-3ILOOWUF.js.map +1 -0
  7. package/dist/{chunk-NXH7AONS.js → chunk-5FRYHDPJ.js} +8 -6
  8. package/dist/chunk-5FRYHDPJ.js.map +1 -0
  9. package/dist/{chunk-TCKL7E4K.js → chunk-5IPC5XXZ.js} +4 -4
  10. package/dist/{chunk-YZBYEULL.js → chunk-75IXBXBO.js} +5 -5
  11. package/dist/{chunk-HRCEIYKU.js → chunk-7NBZ4NMJ.js} +2 -2
  12. package/dist/{chunk-4SM2H22C.js → chunk-HO443ZQM.js} +1 -1
  13. package/dist/{chunk-4SM2H22C.js.map → chunk-HO443ZQM.js.map} +1 -1
  14. package/dist/{chunk-3TZAKV3M.js → chunk-OSWBZKK5.js} +2 -2
  15. package/dist/{chunk-SEXU2WYG.js → chunk-P5KAIE3O.js} +5 -4
  16. package/dist/chunk-P5KAIE3O.js.map +1 -0
  17. package/dist/{chunk-XO5NQRTD.js → chunk-RNICEX6N.js} +2 -2
  18. package/dist/{chunk-L7G27XWF.js → chunk-RRHBTXHB.js} +3 -3
  19. package/dist/{chunk-BVOE7A2Z.js → chunk-UD6H6EP6.js} +8 -6
  20. package/dist/chunk-UD6H6EP6.js.map +1 -0
  21. package/dist/{chunk-LWXF7YRG.js → chunk-UDUBRSIG.js} +2 -2
  22. package/dist/{chunk-SUGSQ4YI.js → chunk-UEEWG7FD.js} +4 -4
  23. package/dist/{chunk-DZ5HJFB4.js → chunk-UM5U6FPH.js} +60 -2
  24. package/dist/chunk-UM5U6FPH.js.map +1 -0
  25. package/dist/{chunk-VXQ33OYT.js → chunk-YV2AN73J.js} +47 -19
  26. package/dist/chunk-YV2AN73J.js.map +1 -0
  27. package/dist/cli.js +6 -6
  28. package/dist/cli.js.map +1 -1
  29. package/dist/db.js +1 -1
  30. package/dist/doctor.js +4 -4
  31. package/dist/hooks/handler.js +4 -4
  32. package/dist/index.d.ts +2 -2
  33. package/dist/index.js +15 -15
  34. package/dist/mcp/server.js +1 -1
  35. package/dist/otlp/server.js +5 -5
  36. package/dist/pricing.js +2 -2
  37. package/dist/proxy/server.js +5 -5
  38. package/dist/prune.js +2 -2
  39. package/dist/query.js +2 -2
  40. package/dist/{reparse-636YZCE3.js → reparse-TFYD67WP.js} +5 -5
  41. package/dist/scanner.d.ts +7 -1
  42. package/dist/scanner.js +1 -1
  43. package/dist/server.js +13 -13
  44. package/dist/setup.js +3 -3
  45. package/dist/sync/index.d.ts +2 -2
  46. package/dist/sync/index.js +4 -4
  47. package/dist/{types-D-MYCBol.d.ts → types-DrhrWbWe.d.ts} +1 -0
  48. package/package.json +1 -1
  49. package/dist/chunk-3BUJ7URA.js.map +0 -1
  50. package/dist/chunk-BVOE7A2Z.js.map +0 -1
  51. package/dist/chunk-DZ5HJFB4.js.map +0 -1
  52. package/dist/chunk-NXH7AONS.js.map +0 -1
  53. package/dist/chunk-SEXU2WYG.js.map +0 -1
  54. package/dist/chunk-VXQ33OYT.js.map +0 -1
  55. /package/dist/{chunk-TCKL7E4K.js.map → chunk-5IPC5XXZ.js.map} +0 -0
  56. /package/dist/{chunk-YZBYEULL.js.map → chunk-75IXBXBO.js.map} +0 -0
  57. /package/dist/{chunk-HRCEIYKU.js.map → chunk-7NBZ4NMJ.js.map} +0 -0
  58. /package/dist/{chunk-3TZAKV3M.js.map → chunk-OSWBZKK5.js.map} +0 -0
  59. /package/dist/{chunk-XO5NQRTD.js.map → chunk-RNICEX6N.js.map} +0 -0
  60. /package/dist/{chunk-L7G27XWF.js.map → chunk-RRHBTXHB.js.map} +0 -0
  61. /package/dist/{chunk-LWXF7YRG.js.map → chunk-UDUBRSIG.js.map} +0 -0
  62. /package/dist/{chunk-SUGSQ4YI.js.map → chunk-UEEWG7FD.js.map} +0 -0
  63. /package/dist/{reparse-636YZCE3.js.map → reparse-TFYD67WP.js.map} +0 -0
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/sync/reader.ts","../src/sync/registry.ts","../src/sync/watermark.ts"],"sourcesContent":["import { getDb } from \"../db/schema.js\";\nimport type {\n HookEventRecord,\n MessageSyncRecord,\n MetricRow,\n OtelLogRecord,\n OtelSpanRecord,\n RepoConfigSnapshotRecord,\n ScannerEventRecord,\n ScannerTurnRecord,\n SessionSyncRecord,\n ToolCallSyncRecord,\n UserConfigSnapshotRecord,\n} from \"./types.js\";\n\nfunction parseJson(raw: string | null): Record<string, unknown> | null {\n if (!raw) return null;\n try {\n const parsed = JSON.parse(raw);\n return typeof parsed === \"object\" && parsed !== null ? parsed : null;\n } catch {\n return null;\n }\n}\n\n// ── Hook events ──────────────────────────────────────────────────────────────\n\nconst HOOK_EVENTS_SQL = `\n SELECT h.id, h.session_id, h.sync_id, h.event_type, h.timestamp_ms, h.cwd, h.repository,\n h.tool_name, decompress(h.payload) as payload,\n h.user_prompt, h.file_path, h.command, h.tool_result,\n s.target\n FROM hook_events h\n LEFT JOIN sessions s ON s.session_id = h.session_id\n WHERE h.id > ?\n ORDER BY h.id\n LIMIT ?\n`;\n\nexport function readHookEvents(\n afterId: number,\n limit: number,\n): { rows: HookEventRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(HOOK_EVENTS_SQL).all(afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n event_type: string;\n timestamp_ms: number;\n cwd: string | null;\n repository: string | null;\n tool_name: string | null;\n payload: string | null;\n user_prompt: string | null;\n file_path: string | null;\n command: string | null;\n tool_result: string | null;\n target: string | null;\n }>;\n\n const rows: HookEventRecord[] = rawRows.map((r) => ({\n hookId: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n eventType: r.event_type,\n timestampMs: r.timestamp_ms,\n cwd: r.cwd,\n repository: r.repository,\n toolName: r.tool_name,\n payload: parseJson(r.payload),\n userPrompt: r.user_prompt,\n filePath: r.file_path,\n command: r.command,\n toolResult: r.tool_result,\n target: r.target,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].hookId : afterId;\n return { rows, maxId };\n}\n\n// ── OTLP logs ────────────────────────────────────────────────────────────────\n\n/** Body types that hooks already cover — filtered out when hooks are installed. */\nconst HOOK_COVERED_BODIES = [\n // Claude Code\n \"claude_code.user_prompt\",\n \"claude_code.tool_decision\",\n \"claude_code.tool_result\",\n // Gemini CLI\n \"gemini_cli.user_prompt\",\n \"gemini_cli.tool_call\",\n \"gemini_cli.hook_call\",\n];\n\nconst ALL_LOGS_SQL = `\n SELECT id, sync_id, timestamp_ns, body, attributes, resource_attributes,\n severity_text, session_id, prompt_id, trace_id, span_id\n FROM otel_logs\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\ninterface RawOtelLogRow {\n id: number;\n sync_id: string | null;\n timestamp_ns: number;\n body: string | null;\n attributes: string | null;\n resource_attributes: string | null;\n severity_text: string | null;\n session_id: string | null;\n prompt_id: string | null;\n trace_id: string | null;\n span_id: string | null;\n}\n\nfunction mapOtelRows(rawRows: RawOtelLogRow[]): OtelLogRecord[] {\n return rawRows.map((r) => ({\n id: r.id,\n syncId: r.sync_id,\n timestampNs: r.timestamp_ns,\n body: r.body,\n attributes: parseJson(r.attributes),\n resourceAttributes: parseJson(r.resource_attributes),\n severityText: r.severity_text,\n sessionId: r.session_id,\n promptId: r.prompt_id,\n traceId: r.trace_id,\n spanId: r.span_id,\n }));\n}\n\n/**\n * Read OTLP logs in batches.\n * When hooksInstalled is true, filters out body types that hooks cover\n * (tool_decision, tool_result, user_prompt) to avoid double-counting.\n */\n/**\n * Read OTLP logs in batches.\n * When hooksInstalled is true, filters out body types that hooks cover\n * (tool_decision, tool_result, user_prompt) to avoid double-counting.\n *\n * maxId advances to the highest id scanned (not just the highest returned),\n * so the watermark skips past blocks of filtered rows without stalling.\n */\nexport function readOtelLogs(\n afterId: number,\n limit: number,\n hooksInstalled: boolean,\n): { rows: OtelLogRecord[]; maxId: number } {\n const db = getDb();\n\n if (!hooksInstalled) {\n const rawRows = db\n .prepare(ALL_LOGS_SQL)\n .all(afterId, limit) as RawOtelLogRow[];\n const rows = mapOtelRows(rawRows);\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n }\n\n // When filtering, we need the max id from the unfiltered range so the\n // watermark advances past blocks of hook-covered rows.\n const scanMaxId = (\n db\n .prepare(\n \"SELECT MAX(id) as m FROM (SELECT id FROM otel_logs WHERE id > ? ORDER BY id LIMIT ?)\",\n )\n .get(afterId, limit) as { m: number | null }\n ).m;\n\n if (scanMaxId == null) return { rows: [], maxId: afterId };\n\n // Only return filtered rows within the scanned range (id <= scanMaxId),\n // not beyond it — otherwise we'd skip ahead of the scan window.\n const rawRows = db\n .prepare(\n `SELECT id, sync_id, timestamp_ns, body, attributes, resource_attributes,\n severity_text, session_id, prompt_id, trace_id, span_id\n FROM otel_logs\n WHERE id > ? AND id <= ?\n AND body NOT IN (${HOOK_COVERED_BODIES.map(() => \"?\").join(\", \")})\n ORDER BY id`,\n )\n .all(afterId, scanMaxId, ...HOOK_COVERED_BODIES) as RawOtelLogRow[];\n\n const rows = mapOtelRows(rawRows);\n return { rows, maxId: scanMaxId };\n}\n\n// ── Metrics ──────────────────────────────────────────────────────────────────\n\nconst METRICS_SQL = `\n SELECT id, sync_id, timestamp_ns, name, value, metric_type, unit,\n attributes, resource_attributes, session_id\n FROM otel_metrics\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readMetrics(\n afterId: number,\n limit: number,\n): { rows: MetricRow[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(METRICS_SQL).all(afterId, limit) as Array<{\n id: number;\n sync_id: string | null;\n timestamp_ns: number;\n name: string;\n value: number;\n metric_type: string | null;\n unit: string | null;\n attributes: string | null;\n resource_attributes: string | null;\n session_id: string | null;\n }>;\n\n const rows: MetricRow[] = rawRows.map((r) => ({\n id: r.id,\n syncId: r.sync_id,\n timestampNs: r.timestamp_ns,\n name: r.name,\n value: r.value,\n metricType: r.metric_type,\n unit: r.unit,\n attributes: parseJson(r.attributes),\n resourceAttributes: parseJson(r.resource_attributes),\n sessionId: r.session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Scanner turns ───────────────────────────────────────────────────────────\n\nconst SCANNER_TURNS_SQL = `\n SELECT t.id, t.session_id, t.sync_id, t.source, t.turn_index, t.timestamp_ms,\n t.model, t.role, t.content_preview,\n t.input_tokens, t.output_tokens, t.cache_read_tokens,\n t.cache_creation_tokens, t.reasoning_tokens,\n s.cli_version\n FROM scanner_turns t\n LEFT JOIN sessions s ON s.session_id = t.session_id\n WHERE t.id > ?\n ORDER BY t.id\n LIMIT ?\n`;\n\nexport function readScannerTurns(\n afterId: number,\n limit: number,\n): { rows: ScannerTurnRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(SCANNER_TURNS_SQL).all(afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n source: string;\n turn_index: number;\n timestamp_ms: number;\n model: string | null;\n role: string | null;\n content_preview: string | null;\n input_tokens: number;\n output_tokens: number;\n cache_read_tokens: number;\n cache_creation_tokens: number;\n reasoning_tokens: number;\n cli_version: string | null;\n }>;\n\n const rows: ScannerTurnRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n source: r.source,\n turnIndex: r.turn_index,\n timestampMs: r.timestamp_ms,\n model: r.model,\n role: r.role,\n contentPreview: r.content_preview,\n inputTokens: r.input_tokens,\n outputTokens: r.output_tokens,\n cacheReadTokens: r.cache_read_tokens,\n cacheCreationTokens: r.cache_creation_tokens,\n reasoningTokens: r.reasoning_tokens,\n cliVersion: r.cli_version,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Scanner events ──────────────────────────────────────────────────────────\n\nconst SCANNER_EVENTS_SQL = `\n SELECT id, session_id, sync_id, source, event_type, timestamp_ms,\n tool_name, tool_input, tool_output, content, metadata\n FROM scanner_events\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readScannerEvents(\n afterId: number,\n limit: number,\n): { rows: ScannerEventRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(SCANNER_EVENTS_SQL).all(afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n source: string;\n event_type: string;\n timestamp_ms: number;\n tool_name: string | null;\n tool_input: string | null;\n tool_output: string | null;\n content: string | null;\n metadata: string | null;\n }>;\n\n const rows: ScannerEventRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n source: r.source,\n eventType: r.event_type,\n timestampMs: r.timestamp_ms,\n toolName: r.tool_name,\n toolInput: r.tool_input,\n toolOutput: r.tool_output,\n content: r.content,\n metadata: parseJson(r.metadata),\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── OTLP spans ─────────────────────────────────────────────────────────────\n\nconst SPANS_SQL = `\n SELECT id, trace_id, span_id, parent_span_id, name, kind,\n start_time_ns, end_time_ns, status_code, status_message,\n attributes, resource_attributes, session_id\n FROM otel_spans\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readOtelSpans(\n afterId: number,\n limit: number,\n): { rows: OtelSpanRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(SPANS_SQL).all(afterId, limit) as Array<{\n id: number;\n trace_id: string;\n span_id: string;\n parent_span_id: string | null;\n name: string;\n kind: number | null;\n start_time_ns: number;\n end_time_ns: number;\n status_code: number | null;\n status_message: string | null;\n attributes: string | null;\n resource_attributes: string | null;\n session_id: string | null;\n }>;\n\n const rows: OtelSpanRecord[] = rawRows.map((r) => ({\n id: r.id,\n traceId: r.trace_id,\n spanId: r.span_id,\n parentSpanId: r.parent_span_id,\n name: r.name,\n kind: r.kind,\n startTimeNs: r.start_time_ns,\n endTimeNs: r.end_time_ns,\n statusCode: r.status_code,\n statusMessage: r.status_message,\n attributes: parseJson(r.attributes),\n resourceAttributes: parseJson(r.resource_attributes),\n sessionId: r.session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── JSON parse helpers for config snapshots ─────────────────────────────────\n\nfunction parseJsonArray(raw: string | null): unknown[] {\n if (!raw) return [];\n try {\n const parsed = JSON.parse(raw);\n return Array.isArray(parsed) ? parsed : [];\n } catch {\n return [];\n }\n}\n\nfunction parseJsonObject(raw: string | null): Record<string, unknown> {\n if (!raw) return {};\n try {\n const parsed = JSON.parse(raw);\n return typeof parsed === \"object\" &&\n parsed !== null &&\n !Array.isArray(parsed)\n ? parsed\n : {};\n } catch {\n return {};\n }\n}\n\n// ── User config snapshots ──────────────────────────────────────────────────\n\nconst USER_CONFIG_SQL = `\n SELECT id, device_name, snapshot_at_ms, content_hash,\n permissions, enabled_plugins, hooks, commands, rules, skills, plugin_hooks\n FROM user_config_snapshots\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readUserConfigSnapshots(\n afterId: number,\n limit: number,\n): { rows: UserConfigSnapshotRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(USER_CONFIG_SQL).all(afterId, limit) as Array<{\n id: number;\n device_name: string;\n snapshot_at_ms: number;\n content_hash: string;\n permissions: string | null;\n enabled_plugins: string | null;\n hooks: string | null;\n commands: string | null;\n rules: string | null;\n skills: string | null;\n plugin_hooks: string | null;\n }>;\n\n const rows: UserConfigSnapshotRecord[] = rawRows.map((r) => ({\n id: r.id,\n deviceName: r.device_name,\n snapshotAtMs: r.snapshot_at_ms,\n contentHash: r.content_hash,\n permissions: parseJsonObject(r.permissions),\n enabledPlugins: parseJsonArray(r.enabled_plugins),\n hooks: parseJsonArray(r.hooks),\n commands: parseJsonArray(r.commands),\n rules: parseJsonArray(r.rules),\n skills: parseJsonArray(r.skills),\n pluginHooks: parseJsonArray(r.plugin_hooks),\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Repo config snapshots ──────────────────────────────────────────────────\n\nconst REPO_CONFIG_SQL = `\n SELECT id, repository, cwd, session_id, snapshot_at_ms, content_hash,\n hooks, mcp_servers, commands, agents, rules,\n local_hooks, local_mcp_servers, local_permissions,\n local_is_gitignored, instructions\n FROM repo_config_snapshots\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readRepoConfigSnapshots(\n afterId: number,\n limit: number,\n): { rows: RepoConfigSnapshotRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(REPO_CONFIG_SQL).all(afterId, limit) as Array<{\n id: number;\n repository: string;\n cwd: string;\n session_id: string | null;\n snapshot_at_ms: number;\n content_hash: string;\n hooks: string | null;\n mcp_servers: string | null;\n commands: string | null;\n agents: string | null;\n rules: string | null;\n local_hooks: string | null;\n local_mcp_servers: string | null;\n local_permissions: string | null;\n local_is_gitignored: number;\n instructions: string | null;\n }>;\n\n const rows: RepoConfigSnapshotRecord[] = rawRows.map((r) => ({\n id: r.id,\n repository: r.repository,\n cwd: r.cwd,\n sessionId: r.session_id,\n snapshotAtMs: r.snapshot_at_ms,\n contentHash: r.content_hash,\n hooks: parseJsonArray(r.hooks),\n mcpServers: parseJsonArray(r.mcp_servers),\n commands: parseJsonArray(r.commands),\n agents: parseJsonArray(r.agents),\n rules: parseJsonArray(r.rules),\n localHooks: parseJsonArray(r.local_hooks),\n localMcpServers: parseJsonArray(r.local_mcp_servers),\n localPermissions: parseJsonObject(r.local_permissions),\n localIsGitignored: r.local_is_gitignored === 1,\n instructions: parseJsonArray(r.instructions),\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Messages ────────────────────────────────────────────────────────────────\n\nconst MESSAGES_SQL = `\n SELECT id, session_id, ordinal, role, content, timestamp_ms,\n has_thinking, has_tool_use, content_length, is_system,\n model, token_usage, context_tokens, output_tokens,\n has_context_tokens, has_output_tokens\n FROM messages\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readMessages(\n afterId: number,\n limit: number,\n): { rows: MessageSyncRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(MESSAGES_SQL).all(afterId, limit) as Array<{\n id: number;\n session_id: string;\n ordinal: number;\n role: string;\n content: string;\n timestamp_ms: number | null;\n has_thinking: number;\n has_tool_use: number;\n content_length: number;\n is_system: number;\n model: string;\n token_usage: string;\n context_tokens: number;\n output_tokens: number;\n has_context_tokens: number;\n has_output_tokens: number;\n }>;\n\n const rows: MessageSyncRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n ordinal: r.ordinal,\n role: r.role,\n content: r.content,\n timestampMs: r.timestamp_ms,\n hasThinking: r.has_thinking === 1,\n hasToolUse: r.has_tool_use === 1,\n contentLength: r.content_length,\n isSystem: r.is_system === 1,\n model: r.model,\n tokenUsage: r.token_usage,\n contextTokens: r.context_tokens,\n outputTokens: r.output_tokens,\n hasContextTokens: r.has_context_tokens === 1,\n hasOutputTokens: r.has_output_tokens === 1,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Tool calls ──────────────────────────────────────────────────────────────\n\nconst TOOL_CALLS_SQL = `\n SELECT id, message_id, session_id, sync_id, tool_name, category, tool_use_id,\n input_json, skill_name, result_content_length, result_content,\n subagent_session_id\n FROM tool_calls\n WHERE id > ?\n ORDER BY id\n LIMIT ?\n`;\n\nexport function readToolCalls(\n afterId: number,\n limit: number,\n): { rows: ToolCallSyncRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(TOOL_CALLS_SQL).all(afterId, limit) as Array<{\n id: number;\n message_id: number;\n session_id: string;\n sync_id: string | null;\n tool_name: string;\n category: string;\n tool_use_id: string | null;\n input_json: string | null;\n skill_name: string | null;\n result_content_length: number | null;\n result_content: string | null;\n subagent_session_id: string | null;\n }>;\n\n const rows: ToolCallSyncRecord[] = rawRows.map((r) => ({\n id: r.id,\n messageId: r.message_id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n toolName: r.tool_name,\n category: r.category,\n toolUseId: r.tool_use_id,\n inputJson: r.input_json,\n skillName: r.skill_name,\n resultContentLength: r.result_content_length,\n resultContent: r.result_content,\n subagentSessionId: r.subagent_session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n// ── Sessions (watermark on sync_seq) ────────────────────────────────────────\n\nconst SESSIONS_SQL = `\n SELECT session_id, target, started_at_ms, ended_at_ms, cwd, first_prompt,\n permission_mode, agent_version,\n total_input_tokens, total_output_tokens, total_cache_read_tokens,\n total_cache_creation_tokens, total_reasoning_tokens, turn_count,\n models, summary, tool_counts, hook_tool_counts, event_type_counts, hook_event_type_counts, sync_seq,\n project, machine, message_count, user_message_count,\n parent_session_id, relationship_type, is_automated, created_at\n FROM sessions\n WHERE sync_seq > ?\n ORDER BY sync_seq\n LIMIT ?\n`;\n\nexport function readSessions(\n afterSeq: number,\n limit: number,\n): { rows: SessionSyncRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db.prepare(SESSIONS_SQL).all(afterSeq, limit) as Array<{\n session_id: string;\n target: string | null;\n started_at_ms: number | null;\n ended_at_ms: number | null;\n cwd: string | null;\n first_prompt: string | null;\n permission_mode: string | null;\n agent_version: string | null;\n total_input_tokens: number | null;\n total_output_tokens: number | null;\n total_cache_read_tokens: number | null;\n total_cache_creation_tokens: number | null;\n total_reasoning_tokens: number | null;\n turn_count: number | null;\n models: string | null;\n summary: string | null;\n tool_counts: string | null;\n hook_tool_counts: string | null;\n event_type_counts: string | null;\n hook_event_type_counts: string | null;\n sync_seq: number;\n project: string | null;\n machine: string;\n message_count: number;\n user_message_count: number;\n parent_session_id: string | null;\n relationship_type: string;\n is_automated: number;\n created_at: number | null;\n }>;\n\n if (rawRows.length === 0) return { rows: [], maxId: afterSeq };\n\n const sessionIds = rawRows.map((r) => r.session_id);\n const placeholders = sessionIds.map(() => \"?\").join(\", \");\n\n const repoRows = db\n .prepare(\n `SELECT session_id, repository, first_seen_ms, git_user_name, git_user_email, branch\n FROM session_repositories\n WHERE session_id IN (${placeholders})`,\n )\n .all(...sessionIds) as Array<{\n session_id: string;\n repository: string;\n first_seen_ms: number;\n git_user_name: string | null;\n git_user_email: string | null;\n branch: string | null;\n }>;\n\n const cwdRows = db\n .prepare(\n `SELECT session_id, cwd, first_seen_ms\n FROM session_cwds\n WHERE session_id IN (${placeholders})`,\n )\n .all(...sessionIds) as Array<{\n session_id: string;\n cwd: string;\n first_seen_ms: number;\n }>;\n\n const reposBySession = new Map<string, SessionSyncRecord[\"repositories\"]>();\n for (const r of repoRows) {\n const list = reposBySession.get(r.session_id) ?? [];\n list.push({\n repository: r.repository,\n firstSeenMs: r.first_seen_ms,\n gitUserName: r.git_user_name,\n gitUserEmail: r.git_user_email,\n branch: r.branch,\n });\n reposBySession.set(r.session_id, list);\n }\n\n const cwdsBySession = new Map<string, SessionSyncRecord[\"cwds\"]>();\n for (const r of cwdRows) {\n const list = cwdsBySession.get(r.session_id) ?? [];\n list.push({ cwd: r.cwd, firstSeenMs: r.first_seen_ms });\n cwdsBySession.set(r.session_id, list);\n }\n\n const rows: SessionSyncRecord[] = rawRows.map((r) => ({\n sessionId: r.session_id,\n target: r.target,\n startedAtMs: r.started_at_ms,\n endedAtMs: r.ended_at_ms,\n cwd: r.cwd,\n firstPrompt: r.first_prompt,\n permissionMode: r.permission_mode,\n agentVersion: r.agent_version,\n totalInputTokens: r.total_input_tokens,\n totalOutputTokens: r.total_output_tokens,\n totalCacheReadTokens: r.total_cache_read_tokens,\n totalCacheCreationTokens: r.total_cache_creation_tokens,\n totalReasoningTokens: r.total_reasoning_tokens,\n turnCount: r.turn_count,\n models: r.models,\n summary: r.summary,\n toolCounts: parseJsonObject(r.tool_counts) as Record<string, number>,\n hookToolCounts: parseJsonObject(r.hook_tool_counts) as Record<\n string,\n number\n >,\n eventTypeCounts: parseJsonObject(r.event_type_counts) as Record<\n string,\n number\n >,\n hookEventTypeCounts: parseJsonObject(r.hook_event_type_counts) as Record<\n string,\n number\n >,\n project: r.project,\n machine: r.machine,\n messageCount: r.message_count,\n userMessageCount: r.user_message_count,\n parentSessionId: r.parent_session_id,\n relationshipType: r.relationship_type,\n isAutomated: r.is_automated === 1,\n createdAt: r.created_at,\n repositories: reposBySession.get(r.session_id) ?? [],\n cwds: cwdsBySession.get(r.session_id) ?? [],\n }));\n\n const maxId = rawRows[rawRows.length - 1].sync_seq;\n return { rows, maxId };\n}\n\n// ── Per-session readers (for gated sync) ──────────────────────────────────\n\nexport function readSessionMessages(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: MessageSyncRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, session_id, ordinal, role, content, timestamp_ms,\n has_thinking, has_tool_use, content_length, is_system,\n model, token_usage, context_tokens, output_tokens,\n has_context_tokens, has_output_tokens\n FROM messages\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n session_id: string;\n ordinal: number;\n role: string;\n content: string;\n timestamp_ms: number | null;\n has_thinking: number;\n has_tool_use: number;\n content_length: number;\n is_system: number;\n model: string;\n token_usage: string;\n context_tokens: number;\n output_tokens: number;\n has_context_tokens: number;\n has_output_tokens: number;\n }>;\n\n const rows: MessageSyncRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n ordinal: r.ordinal,\n role: r.role,\n content: r.content,\n timestampMs: r.timestamp_ms,\n hasThinking: r.has_thinking === 1,\n hasToolUse: r.has_tool_use === 1,\n contentLength: r.content_length,\n isSystem: r.is_system === 1,\n model: r.model,\n tokenUsage: r.token_usage,\n contextTokens: r.context_tokens,\n outputTokens: r.output_tokens,\n hasContextTokens: r.has_context_tokens === 1,\n hasOutputTokens: r.has_output_tokens === 1,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionToolCalls(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: ToolCallSyncRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, message_id, session_id, sync_id, tool_name, category, tool_use_id,\n input_json, skill_name, result_content_length, result_content,\n subagent_session_id\n FROM tool_calls\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n message_id: number;\n session_id: string;\n sync_id: string | null;\n tool_name: string;\n category: string;\n tool_use_id: string | null;\n input_json: string | null;\n skill_name: string | null;\n result_content_length: number | null;\n result_content: string | null;\n subagent_session_id: string | null;\n }>;\n\n const rows: ToolCallSyncRecord[] = rawRows.map((r) => ({\n id: r.id,\n messageId: r.message_id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n toolName: r.tool_name,\n category: r.category,\n toolUseId: r.tool_use_id,\n inputJson: r.input_json,\n skillName: r.skill_name,\n resultContentLength: r.result_content_length,\n resultContent: r.result_content,\n subagentSessionId: r.subagent_session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionScannerTurns(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: ScannerTurnRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT t.id, t.session_id, t.sync_id, t.source, t.turn_index, t.timestamp_ms,\n t.model, t.role, t.content_preview,\n t.input_tokens, t.output_tokens, t.cache_read_tokens,\n t.cache_creation_tokens, t.reasoning_tokens,\n s.cli_version\n FROM scanner_turns t\n LEFT JOIN sessions s ON s.session_id = t.session_id\n WHERE t.session_id = ? AND t.id > ?\n ORDER BY t.id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n source: string;\n turn_index: number;\n timestamp_ms: number;\n model: string | null;\n role: string | null;\n content_preview: string | null;\n input_tokens: number;\n output_tokens: number;\n cache_read_tokens: number;\n cache_creation_tokens: number;\n reasoning_tokens: number;\n cli_version: string | null;\n }>;\n\n const rows: ScannerTurnRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n source: r.source,\n turnIndex: r.turn_index,\n timestampMs: r.timestamp_ms,\n model: r.model,\n role: r.role,\n contentPreview: r.content_preview,\n inputTokens: r.input_tokens,\n outputTokens: r.output_tokens,\n cacheReadTokens: r.cache_read_tokens,\n cacheCreationTokens: r.cache_creation_tokens,\n reasoningTokens: r.reasoning_tokens,\n cliVersion: r.cli_version,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionScannerEvents(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: ScannerEventRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, session_id, sync_id, source, event_type, timestamp_ms,\n tool_name, tool_input, tool_output, content, metadata\n FROM scanner_events\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n source: string;\n event_type: string;\n timestamp_ms: number;\n tool_name: string | null;\n tool_input: string | null;\n tool_output: string | null;\n content: string | null;\n metadata: string | null;\n }>;\n\n const rows: ScannerEventRecord[] = rawRows.map((r) => ({\n id: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n source: r.source,\n eventType: r.event_type,\n timestampMs: r.timestamp_ms,\n toolName: r.tool_name,\n toolInput: r.tool_input,\n toolOutput: r.tool_output,\n content: r.content,\n metadata: parseJson(r.metadata),\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionHookEvents(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: HookEventRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT h.id, h.session_id, h.sync_id, h.event_type, h.timestamp_ms, h.cwd, h.repository,\n h.tool_name, decompress(h.payload) as payload,\n h.user_prompt, h.file_path, h.command, h.tool_result,\n s.target\n FROM hook_events h\n LEFT JOIN sessions s ON s.session_id = h.session_id\n WHERE h.session_id = ? AND h.id > ?\n ORDER BY h.id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n session_id: string;\n sync_id: string | null;\n event_type: string;\n timestamp_ms: number;\n cwd: string | null;\n repository: string | null;\n tool_name: string | null;\n payload: string | null;\n user_prompt: string | null;\n file_path: string | null;\n command: string | null;\n tool_result: string | null;\n target: string | null;\n }>;\n\n const rows: HookEventRecord[] = rawRows.map((r) => ({\n hookId: r.id,\n sessionId: r.session_id,\n syncId: r.sync_id,\n eventType: r.event_type,\n timestampMs: r.timestamp_ms,\n cwd: r.cwd,\n repository: r.repository,\n toolName: r.tool_name,\n payload: parseJson(r.payload),\n userPrompt: r.user_prompt,\n filePath: r.file_path,\n command: r.command,\n toolResult: r.tool_result,\n target: r.target,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].hookId : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionOtelLogs(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: OtelLogRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, sync_id, timestamp_ns, body, attributes, resource_attributes,\n severity_text, session_id, prompt_id, trace_id, span_id\n FROM otel_logs\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as RawOtelLogRow[];\n\n const rows = mapOtelRows(rawRows);\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionOtelMetrics(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: MetricRow[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, sync_id, timestamp_ns, name, value, metric_type, unit,\n attributes, resource_attributes, session_id\n FROM otel_metrics\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n sync_id: string | null;\n timestamp_ns: number;\n name: string;\n value: number;\n metric_type: string | null;\n unit: string | null;\n attributes: string | null;\n resource_attributes: string | null;\n session_id: string | null;\n }>;\n\n const rows: MetricRow[] = rawRows.map((r) => ({\n id: r.id,\n syncId: r.sync_id,\n timestampNs: r.timestamp_ns,\n name: r.name,\n value: r.value,\n metricType: r.metric_type,\n unit: r.unit,\n attributes: parseJson(r.attributes),\n resourceAttributes: parseJson(r.resource_attributes),\n sessionId: r.session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\nexport function readSessionOtelSpans(\n sessionId: string,\n afterId: number,\n limit: number,\n): { rows: OtelSpanRecord[]; maxId: number } {\n const db = getDb();\n const rawRows = db\n .prepare(\n `SELECT id, trace_id, span_id, parent_span_id, name, kind,\n start_time_ns, end_time_ns, status_code, status_message,\n attributes, resource_attributes, session_id\n FROM otel_spans\n WHERE session_id = ? AND id > ?\n ORDER BY id\n LIMIT ?`,\n )\n .all(sessionId, afterId, limit) as Array<{\n id: number;\n trace_id: string;\n span_id: string;\n parent_span_id: string | null;\n name: string;\n kind: number | null;\n start_time_ns: number;\n end_time_ns: number;\n status_code: number | null;\n status_message: string | null;\n attributes: string | null;\n resource_attributes: string | null;\n session_id: string | null;\n }>;\n\n const rows: OtelSpanRecord[] = rawRows.map((r) => ({\n id: r.id,\n traceId: r.trace_id,\n spanId: r.span_id,\n parentSpanId: r.parent_span_id,\n name: r.name,\n kind: r.kind,\n startTimeNs: r.start_time_ns,\n endTimeNs: r.end_time_ns,\n statusCode: r.status_code,\n statusMessage: r.status_message,\n attributes: parseJson(r.attributes),\n resourceAttributes: parseJson(r.resource_attributes),\n sessionId: r.session_id,\n }));\n\n const maxId = rows.length > 0 ? rows[rows.length - 1].id : afterId;\n return { rows, maxId };\n}\n\n/** Map of table name → per-session reader for gated sync. */\nexport const SESSION_READERS: Record<\n string,\n (\n sessionId: string,\n afterId: number,\n limit: number,\n ) => { rows: unknown[]; maxId: number }\n> = {\n messages: readSessionMessages,\n tool_calls: readSessionToolCalls,\n scanner_turns: readSessionScannerTurns,\n scanner_events: readSessionScannerEvents,\n hook_events: readSessionHookEvents,\n otel_logs: readSessionOtelLogs,\n otel_metrics: readSessionOtelMetrics,\n otel_spans: readSessionOtelSpans,\n};\n","import {\n readHookEvents,\n readMessages,\n readMetrics,\n readOtelLogs,\n readOtelSpans,\n readRepoConfigSnapshots,\n readScannerEvents,\n readScannerTurns,\n readSessions,\n readToolCalls,\n readUserConfigSnapshots,\n} from \"./reader.js\";\nimport type { TableSyncDescriptor } from \"./types.js\";\n\n/**\n * Ordered list of table sync descriptors. The order matches the\n * round-robin execution order in the sync loop.\n *\n * All tables sync via POST /v1/sync with {table, rows} payload.\n * Session-linked tables are filtered by repo attribution in the sync loop.\n */\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const TABLE_SYNC_REGISTRY: TableSyncDescriptor<any>[] = [\n // ── Session-linked tables (filtered by repo attribution) ─────────────────\n\n {\n table: \"sessions\",\n logNoun: \"sessions\",\n read: (afterId, limit) => readSessions(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"messages\",\n logNoun: \"messages\",\n read: (afterId, limit) => readMessages(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"tool_calls\",\n logNoun: \"tool calls\",\n read: (afterId, limit) => readToolCalls(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"scanner_turns\",\n logNoun: \"turns\",\n read: (afterId, limit) => readScannerTurns(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"scanner_events\",\n logNoun: \"events\",\n read: (afterId, limit) => readScannerEvents(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"hook_events\",\n logNoun: \"events\",\n read: (afterId, limit) => readHookEvents(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"otel_logs\",\n logNoun: \"logs\",\n read: (afterId, limit) => readOtelLogs(afterId, limit, false),\n sessionLinked: true,\n },\n\n {\n table: \"otel_metrics\",\n logNoun: \"metrics\",\n read: (afterId, limit) => readMetrics(afterId, limit),\n sessionLinked: true,\n },\n\n {\n table: \"otel_spans\",\n logNoun: \"spans\",\n read: (afterId, limit) => readOtelSpans(afterId, limit),\n sessionLinked: true,\n },\n\n // ── Non-session tables (always synced) ────────────────────────────────────\n\n {\n table: \"user_config_snapshots\",\n logNoun: \"snapshots\",\n read: (afterId, limit) => readUserConfigSnapshots(afterId, limit),\n sessionLinked: false,\n },\n\n {\n table: \"repo_config_snapshots\",\n logNoun: \"snapshots\",\n read: (afterId, limit) => readRepoConfigSnapshots(afterId, limit),\n sessionLinked: false,\n },\n];\n","import { getDb } from \"../db/schema.js\";\nimport { TABLE_SYNC_REGISTRY } from \"./registry.js\";\n\nexport function watermarkKey(table: string, targetName: string): string {\n return `${table}:${targetName}`;\n}\n\nexport function readWatermark(key: string): number {\n const db = getDb();\n const row = db\n .prepare(\"SELECT value FROM watermarks WHERE key = ?\")\n .get(key) as { value: number } | undefined;\n return row?.value ?? 0;\n}\n\nexport function writeWatermark(key: string, value: number): void {\n const db = getDb();\n db.prepare(\n \"INSERT INTO watermarks (key, value) VALUES (?, ?) ON CONFLICT(key) DO UPDATE SET value = excluded.value\",\n ).run(key, value);\n}\n\nexport function resetWatermarks(targetName?: string): void {\n const db = getDb();\n if (targetName) {\n const stmt = db.prepare(\"DELETE FROM watermarks WHERE key = ?\");\n for (const desc of TABLE_SYNC_REGISTRY) {\n stmt.run(watermarkKey(desc.table, targetName));\n }\n db.prepare(\"DELETE FROM target_session_sync WHERE target = ?\").run(\n targetName,\n );\n } else {\n db.prepare(\"DELETE FROM watermarks\").run();\n db.prepare(\"DELETE FROM target_session_sync\").run();\n }\n}\n"],"mappings":";;;;;AAeA,SAAS,UAAU,KAAoD;AACrE,MAAI,CAAC,IAAK,QAAO;AACjB,MAAI;AACF,UAAM,SAAS,KAAK,MAAM,GAAG;AAC7B,WAAO,OAAO,WAAW,YAAY,WAAW,OAAO,SAAS;AAAA,EAClE,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAIA,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAYjB,SAAS,eACd,SACA,OAC4C;AAC5C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,eAAe,EAAE,IAAI,SAAS,KAAK;AAiB9D,QAAM,OAA0B,QAAQ,IAAI,CAAC,OAAO;AAAA,IAClD,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,KAAK,EAAE;AAAA,IACP,YAAY,EAAE;AAAA,IACd,UAAU,EAAE;AAAA,IACZ,SAAS,UAAU,EAAE,OAAO;AAAA,IAC5B,YAAY,EAAE;AAAA,IACd,UAAU,EAAE;AAAA,IACZ,SAAS,EAAE;AAAA,IACX,YAAY,EAAE;AAAA,IACd,QAAQ,EAAE;AAAA,EACZ,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,SAAS;AAC/D,SAAO,EAAE,MAAM,MAAM;AACvB;AAKA,IAAM,sBAAsB;AAAA;AAAA,EAE1B;AAAA,EACA;AAAA,EACA;AAAA;AAAA,EAEA;AAAA,EACA;AAAA,EACA;AACF;AAEA,IAAM,eAAe;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAuBrB,SAAS,YAAY,SAA2C;AAC9D,SAAO,QAAQ,IAAI,CAAC,OAAO;AAAA,IACzB,IAAI,EAAE;AAAA,IACN,QAAQ,EAAE;AAAA,IACV,aAAa,EAAE;AAAA,IACf,MAAM,EAAE;AAAA,IACR,YAAY,UAAU,EAAE,UAAU;AAAA,IAClC,oBAAoB,UAAU,EAAE,mBAAmB;AAAA,IACnD,cAAc,EAAE;AAAA,IAChB,WAAW,EAAE;AAAA,IACb,UAAU,EAAE;AAAA,IACZ,SAAS,EAAE;AAAA,IACX,QAAQ,EAAE;AAAA,EACZ,EAAE;AACJ;AAeO,SAAS,aACd,SACA,OACA,gBAC0C;AAC1C,QAAM,KAAK,MAAM;AAEjB,MAAI,CAAC,gBAAgB;AACnB,UAAMA,WAAU,GACb,QAAQ,YAAY,EACpB,IAAI,SAAS,KAAK;AACrB,UAAMC,QAAO,YAAYD,QAAO;AAChC,UAAM,QAAQC,MAAK,SAAS,IAAIA,MAAKA,MAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,WAAO,EAAE,MAAAA,OAAM,MAAM;AAAA,EACvB;AAIA,QAAM,YACJ,GACG;AAAA,IACC;AAAA,EACF,EACC,IAAI,SAAS,KAAK,EACrB;AAEF,MAAI,aAAa,KAAM,QAAO,EAAE,MAAM,CAAC,GAAG,OAAO,QAAQ;AAIzD,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA,4BAIsB,oBAAoB,IAAI,MAAM,GAAG,EAAE,KAAK,IAAI,CAAC;AAAA;AAAA,EAErE,EACC,IAAI,SAAS,WAAW,GAAG,mBAAmB;AAEjD,QAAM,OAAO,YAAY,OAAO;AAChC,SAAO,EAAE,MAAM,OAAO,UAAU;AAClC;AAIA,IAAM,cAAc;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AASb,SAAS,YACd,SACA,OACsC;AACtC,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,WAAW,EAAE,IAAI,SAAS,KAAK;AAa1D,QAAM,OAAoB,QAAQ,IAAI,CAAC,OAAO;AAAA,IAC5C,IAAI,EAAE;AAAA,IACN,QAAQ,EAAE;AAAA,IACV,aAAa,EAAE;AAAA,IACf,MAAM,EAAE;AAAA,IACR,OAAO,EAAE;AAAA,IACT,YAAY,EAAE;AAAA,IACd,MAAM,EAAE;AAAA,IACR,YAAY,UAAU,EAAE,UAAU;AAAA,IAClC,oBAAoB,UAAU,EAAE,mBAAmB;AAAA,IACnD,WAAW,EAAE;AAAA,EACf,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,oBAAoB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAanB,SAAS,iBACd,SACA,OAC8C;AAC9C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,iBAAiB,EAAE,IAAI,SAAS,KAAK;AAkBhE,QAAM,OAA4B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACpD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,OAAO,EAAE;AAAA,IACT,MAAM,EAAE;AAAA,IACR,gBAAgB,EAAE;AAAA,IAClB,aAAa,EAAE;AAAA,IACf,cAAc,EAAE;AAAA,IAChB,iBAAiB,EAAE;AAAA,IACnB,qBAAqB,EAAE;AAAA,IACvB,iBAAiB,EAAE;AAAA,IACnB,YAAY,EAAE;AAAA,EAChB,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,qBAAqB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AASpB,SAAS,kBACd,SACA,OAC+C;AAC/C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,kBAAkB,EAAE,IAAI,SAAS,KAAK;AAcjE,QAAM,OAA6B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACrD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,UAAU,EAAE;AAAA,IACZ,WAAW,EAAE;AAAA,IACb,YAAY,EAAE;AAAA,IACd,SAAS,EAAE;AAAA,IACX,UAAU,UAAU,EAAE,QAAQ;AAAA,EAChC,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,YAAY;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAUX,SAAS,cACd,SACA,OAC2C;AAC3C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,SAAS,EAAE,IAAI,SAAS,KAAK;AAgBxD,QAAM,OAAyB,QAAQ,IAAI,CAAC,OAAO;AAAA,IACjD,IAAI,EAAE;AAAA,IACN,SAAS,EAAE;AAAA,IACX,QAAQ,EAAE;AAAA,IACV,cAAc,EAAE;AAAA,IAChB,MAAM,EAAE;AAAA,IACR,MAAM,EAAE;AAAA,IACR,aAAa,EAAE;AAAA,IACf,WAAW,EAAE;AAAA,IACb,YAAY,EAAE;AAAA,IACd,eAAe,EAAE;AAAA,IACjB,YAAY,UAAU,EAAE,UAAU;AAAA,IAClC,oBAAoB,UAAU,EAAE,mBAAmB;AAAA,IACnD,WAAW,EAAE;AAAA,EACf,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,SAAS,eAAe,KAA+B;AACrD,MAAI,CAAC,IAAK,QAAO,CAAC;AAClB,MAAI;AACF,UAAM,SAAS,KAAK,MAAM,GAAG;AAC7B,WAAO,MAAM,QAAQ,MAAM,IAAI,SAAS,CAAC;AAAA,EAC3C,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAEA,SAAS,gBAAgB,KAA6C;AACpE,MAAI,CAAC,IAAK,QAAO,CAAC;AAClB,MAAI;AACF,UAAM,SAAS,KAAK,MAAM,GAAG;AAC7B,WAAO,OAAO,WAAW,YACvB,WAAW,QACX,CAAC,MAAM,QAAQ,MAAM,IACnB,SACA,CAAC;AAAA,EACP,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAIA,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AASjB,SAAS,wBACd,SACA,OACqD;AACrD,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,eAAe,EAAE,IAAI,SAAS,KAAK;AAc9D,QAAM,OAAmC,QAAQ,IAAI,CAAC,OAAO;AAAA,IAC3D,IAAI,EAAE;AAAA,IACN,YAAY,EAAE;AAAA,IACd,cAAc,EAAE;AAAA,IAChB,aAAa,EAAE;AAAA,IACf,aAAa,gBAAgB,EAAE,WAAW;AAAA,IAC1C,gBAAgB,eAAe,EAAE,eAAe;AAAA,IAChD,OAAO,eAAe,EAAE,KAAK;AAAA,IAC7B,UAAU,eAAe,EAAE,QAAQ;AAAA,IACnC,OAAO,eAAe,EAAE,KAAK;AAAA,IAC7B,QAAQ,eAAe,EAAE,MAAM;AAAA,IAC/B,aAAa,eAAe,EAAE,YAAY;AAAA,EAC5C,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAWjB,SAAS,wBACd,SACA,OACqD;AACrD,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,eAAe,EAAE,IAAI,SAAS,KAAK;AAmB9D,QAAM,OAAmC,QAAQ,IAAI,CAAC,OAAO;AAAA,IAC3D,IAAI,EAAE;AAAA,IACN,YAAY,EAAE;AAAA,IACd,KAAK,EAAE;AAAA,IACP,WAAW,EAAE;AAAA,IACb,cAAc,EAAE;AAAA,IAChB,aAAa,EAAE;AAAA,IACf,OAAO,eAAe,EAAE,KAAK;AAAA,IAC7B,YAAY,eAAe,EAAE,WAAW;AAAA,IACxC,UAAU,eAAe,EAAE,QAAQ;AAAA,IACnC,QAAQ,eAAe,EAAE,MAAM;AAAA,IAC/B,OAAO,eAAe,EAAE,KAAK;AAAA,IAC7B,YAAY,eAAe,EAAE,WAAW;AAAA,IACxC,iBAAiB,eAAe,EAAE,iBAAiB;AAAA,IACnD,kBAAkB,gBAAgB,EAAE,iBAAiB;AAAA,IACrD,mBAAmB,EAAE,wBAAwB;AAAA,IAC7C,cAAc,eAAe,EAAE,YAAY;AAAA,EAC7C,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,eAAe;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAWd,SAAS,aACd,SACA,OAC8C;AAC9C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,YAAY,EAAE,IAAI,SAAS,KAAK;AAmB3D,QAAM,OAA4B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACpD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,SAAS,EAAE;AAAA,IACX,MAAM,EAAE;AAAA,IACR,SAAS,EAAE;AAAA,IACX,aAAa,EAAE;AAAA,IACf,aAAa,EAAE,iBAAiB;AAAA,IAChC,YAAY,EAAE,iBAAiB;AAAA,IAC/B,eAAe,EAAE;AAAA,IACjB,UAAU,EAAE,cAAc;AAAA,IAC1B,OAAO,EAAE;AAAA,IACT,YAAY,EAAE;AAAA,IACd,eAAe,EAAE;AAAA,IACjB,cAAc,EAAE;AAAA,IAChB,kBAAkB,EAAE,uBAAuB;AAAA,IAC3C,iBAAiB,EAAE,sBAAsB;AAAA,EAC3C,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,iBAAiB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAUhB,SAAS,cACd,SACA,OAC+C;AAC/C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,cAAc,EAAE,IAAI,SAAS,KAAK;AAe7D,QAAM,OAA6B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACrD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,UAAU,EAAE;AAAA,IACZ,UAAU,EAAE;AAAA,IACZ,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,qBAAqB,EAAE;AAAA,IACvB,eAAe,EAAE;AAAA,IACjB,mBAAmB,EAAE;AAAA,EACvB,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAIA,IAAM,eAAe;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAcd,SAAS,aACd,UACA,OAC8C;AAC9C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GAAG,QAAQ,YAAY,EAAE,IAAI,UAAU,KAAK;AAgC5D,MAAI,QAAQ,WAAW,EAAG,QAAO,EAAE,MAAM,CAAC,GAAG,OAAO,SAAS;AAE7D,QAAM,aAAa,QAAQ,IAAI,CAAC,MAAM,EAAE,UAAU;AAClD,QAAM,eAAe,WAAW,IAAI,MAAM,GAAG,EAAE,KAAK,IAAI;AAExD,QAAM,WAAW,GACd;AAAA,IACC;AAAA;AAAA,8BAEwB,YAAY;AAAA,EACtC,EACC,IAAI,GAAG,UAAU;AASpB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA,8BAEwB,YAAY;AAAA,EACtC,EACC,IAAI,GAAG,UAAU;AAMpB,QAAM,iBAAiB,oBAAI,IAA+C;AAC1E,aAAW,KAAK,UAAU;AACxB,UAAM,OAAO,eAAe,IAAI,EAAE,UAAU,KAAK,CAAC;AAClD,SAAK,KAAK;AAAA,MACR,YAAY,EAAE;AAAA,MACd,aAAa,EAAE;AAAA,MACf,aAAa,EAAE;AAAA,MACf,cAAc,EAAE;AAAA,MAChB,QAAQ,EAAE;AAAA,IACZ,CAAC;AACD,mBAAe,IAAI,EAAE,YAAY,IAAI;AAAA,EACvC;AAEA,QAAM,gBAAgB,oBAAI,IAAuC;AACjE,aAAW,KAAK,SAAS;AACvB,UAAM,OAAO,cAAc,IAAI,EAAE,UAAU,KAAK,CAAC;AACjD,SAAK,KAAK,EAAE,KAAK,EAAE,KAAK,aAAa,EAAE,cAAc,CAAC;AACtD,kBAAc,IAAI,EAAE,YAAY,IAAI;AAAA,EACtC;AAEA,QAAM,OAA4B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACpD,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,aAAa,EAAE;AAAA,IACf,WAAW,EAAE;AAAA,IACb,KAAK,EAAE;AAAA,IACP,aAAa,EAAE;AAAA,IACf,gBAAgB,EAAE;AAAA,IAClB,cAAc,EAAE;AAAA,IAChB,kBAAkB,EAAE;AAAA,IACpB,mBAAmB,EAAE;AAAA,IACrB,sBAAsB,EAAE;AAAA,IACxB,0BAA0B,EAAE;AAAA,IAC5B,sBAAsB,EAAE;AAAA,IACxB,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,SAAS,EAAE;AAAA,IACX,YAAY,gBAAgB,EAAE,WAAW;AAAA,IACzC,gBAAgB,gBAAgB,EAAE,gBAAgB;AAAA,IAIlD,iBAAiB,gBAAgB,EAAE,iBAAiB;AAAA,IAIpD,qBAAqB,gBAAgB,EAAE,sBAAsB;AAAA,IAI7D,SAAS,EAAE;AAAA,IACX,SAAS,EAAE;AAAA,IACX,cAAc,EAAE;AAAA,IAChB,kBAAkB,EAAE;AAAA,IACpB,iBAAiB,EAAE;AAAA,IACnB,kBAAkB,EAAE;AAAA,IACpB,aAAa,EAAE,iBAAiB;AAAA,IAChC,WAAW,EAAE;AAAA,IACb,cAAc,eAAe,IAAI,EAAE,UAAU,KAAK,CAAC;AAAA,IACnD,MAAM,cAAc,IAAI,EAAE,UAAU,KAAK,CAAC;AAAA,EAC5C,EAAE;AAEF,QAAM,QAAQ,QAAQ,QAAQ,SAAS,CAAC,EAAE;AAC1C,SAAO,EAAE,MAAM,MAAM;AACvB;AAIO,SAAS,oBACd,WACA,SACA,OAC8C;AAC9C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAQF,EACC,IAAI,WAAW,SAAS,KAAK;AAmBhC,QAAM,OAA4B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACpD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,SAAS,EAAE;AAAA,IACX,MAAM,EAAE;AAAA,IACR,SAAS,EAAE;AAAA,IACX,aAAa,EAAE;AAAA,IACf,aAAa,EAAE,iBAAiB;AAAA,IAChC,YAAY,EAAE,iBAAiB;AAAA,IAC/B,eAAe,EAAE;AAAA,IACjB,UAAU,EAAE,cAAc;AAAA,IAC1B,OAAO,EAAE;AAAA,IACT,YAAY,EAAE;AAAA,IACd,eAAe,EAAE;AAAA,IACjB,cAAc,EAAE;AAAA,IAChB,kBAAkB,EAAE,uBAAuB;AAAA,IAC3C,iBAAiB,EAAE,sBAAsB;AAAA,EAC3C,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,qBACd,WACA,SACA,OAC+C;AAC/C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOF,EACC,IAAI,WAAW,SAAS,KAAK;AAehC,QAAM,OAA6B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACrD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,UAAU,EAAE;AAAA,IACZ,UAAU,EAAE;AAAA,IACZ,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,WAAW,EAAE;AAAA,IACb,qBAAqB,EAAE;AAAA,IACvB,eAAe,EAAE;AAAA,IACjB,mBAAmB,EAAE;AAAA,EACvB,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,wBACd,WACA,SACA,OAC8C;AAC9C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAUF,EACC,IAAI,WAAW,SAAS,KAAK;AAkBhC,QAAM,OAA4B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACpD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,OAAO,EAAE;AAAA,IACT,MAAM,EAAE;AAAA,IACR,gBAAgB,EAAE;AAAA,IAClB,aAAa,EAAE;AAAA,IACf,cAAc,EAAE;AAAA,IAChB,iBAAiB,EAAE;AAAA,IACnB,qBAAqB,EAAE;AAAA,IACvB,iBAAiB,EAAE;AAAA,IACnB,YAAY,EAAE;AAAA,EAChB,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,yBACd,WACA,SACA,OAC+C;AAC/C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EACC,IAAI,WAAW,SAAS,KAAK;AAchC,QAAM,OAA6B,QAAQ,IAAI,CAAC,OAAO;AAAA,IACrD,IAAI,EAAE;AAAA,IACN,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,UAAU,EAAE;AAAA,IACZ,WAAW,EAAE;AAAA,IACb,YAAY,EAAE;AAAA,IACd,SAAS,EAAE;AAAA,IACX,UAAU,UAAU,EAAE,QAAQ;AAAA,EAChC,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,sBACd,WACA,SACA,OAC4C;AAC5C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EASF,EACC,IAAI,WAAW,SAAS,KAAK;AAiBhC,QAAM,OAA0B,QAAQ,IAAI,CAAC,OAAO;AAAA,IAClD,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,QAAQ,EAAE;AAAA,IACV,WAAW,EAAE;AAAA,IACb,aAAa,EAAE;AAAA,IACf,KAAK,EAAE;AAAA,IACP,YAAY,EAAE;AAAA,IACd,UAAU,EAAE;AAAA,IACZ,SAAS,UAAU,EAAE,OAAO;AAAA,IAC5B,YAAY,EAAE;AAAA,IACd,UAAU,EAAE;AAAA,IACZ,SAAS,EAAE;AAAA,IACX,YAAY,EAAE;AAAA,IACd,QAAQ,EAAE;AAAA,EACZ,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,SAAS;AAC/D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,oBACd,WACA,SACA,OAC0C;AAC1C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EACC,IAAI,WAAW,SAAS,KAAK;AAEhC,QAAM,OAAO,YAAY,OAAO;AAChC,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,uBACd,WACA,SACA,OACsC;AACtC,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EACC,IAAI,WAAW,SAAS,KAAK;AAahC,QAAM,OAAoB,QAAQ,IAAI,CAAC,OAAO;AAAA,IAC5C,IAAI,EAAE;AAAA,IACN,QAAQ,EAAE;AAAA,IACV,aAAa,EAAE;AAAA,IACf,MAAM,EAAE;AAAA,IACR,OAAO,EAAE;AAAA,IACT,YAAY,EAAE;AAAA,IACd,MAAM,EAAE;AAAA,IACR,YAAY,UAAU,EAAE,UAAU;AAAA,IAClC,oBAAoB,UAAU,EAAE,mBAAmB;AAAA,IACnD,WAAW,EAAE;AAAA,EACf,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAEO,SAAS,qBACd,WACA,SACA,OAC2C;AAC3C,QAAM,KAAK,MAAM;AACjB,QAAM,UAAU,GACb;AAAA,IACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOF,EACC,IAAI,WAAW,SAAS,KAAK;AAgBhC,QAAM,OAAyB,QAAQ,IAAI,CAAC,OAAO;AAAA,IACjD,IAAI,EAAE;AAAA,IACN,SAAS,EAAE;AAAA,IACX,QAAQ,EAAE;AAAA,IACV,cAAc,EAAE;AAAA,IAChB,MAAM,EAAE;AAAA,IACR,MAAM,EAAE;AAAA,IACR,aAAa,EAAE;AAAA,IACf,WAAW,EAAE;AAAA,IACb,YAAY,EAAE;AAAA,IACd,eAAe,EAAE;AAAA,IACjB,YAAY,UAAU,EAAE,UAAU;AAAA,IAClC,oBAAoB,UAAU,EAAE,mBAAmB;AAAA,IACnD,WAAW,EAAE;AAAA,EACf,EAAE;AAEF,QAAM,QAAQ,KAAK,SAAS,IAAI,KAAK,KAAK,SAAS,CAAC,EAAE,KAAK;AAC3D,SAAO,EAAE,MAAM,MAAM;AACvB;AAGO,IAAM,kBAOT;AAAA,EACF,UAAU;AAAA,EACV,YAAY;AAAA,EACZ,eAAe;AAAA,EACf,gBAAgB;AAAA,EAChB,aAAa;AAAA,EACb,WAAW;AAAA,EACX,cAAc;AAAA,EACd,YAAY;AACd;;;AC9pCO,IAAM,sBAAkD;AAAA;AAAA,EAG7D;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,aAAa,SAAS,KAAK;AAAA,IACrD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,aAAa,SAAS,KAAK;AAAA,IACrD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,cAAc,SAAS,KAAK;AAAA,IACtD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,iBAAiB,SAAS,KAAK;AAAA,IACzD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,kBAAkB,SAAS,KAAK;AAAA,IAC1D,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,eAAe,SAAS,KAAK;AAAA,IACvD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,aAAa,SAAS,OAAO,KAAK;AAAA,IAC5D,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,YAAY,SAAS,KAAK;AAAA,IACpD,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,cAAc,SAAS,KAAK;AAAA,IACtD,eAAe;AAAA,EACjB;AAAA;AAAA,EAIA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,wBAAwB,SAAS,KAAK;AAAA,IAChE,eAAe;AAAA,EACjB;AAAA,EAEA;AAAA,IACE,OAAO;AAAA,IACP,SAAS;AAAA,IACT,MAAM,CAAC,SAAS,UAAU,wBAAwB,SAAS,KAAK;AAAA,IAChE,eAAe;AAAA,EACjB;AACF;;;ACrGO,SAAS,aAAa,OAAe,YAA4B;AACtE,SAAO,GAAG,KAAK,IAAI,UAAU;AAC/B;AAEO,SAAS,cAAc,KAAqB;AACjD,QAAM,KAAK,MAAM;AACjB,QAAM,MAAM,GACT,QAAQ,4CAA4C,EACpD,IAAI,GAAG;AACV,SAAO,KAAK,SAAS;AACvB;AAEO,SAAS,eAAe,KAAa,OAAqB;AAC/D,QAAM,KAAK,MAAM;AACjB,KAAG;AAAA,IACD;AAAA,EACF,EAAE,IAAI,KAAK,KAAK;AAClB;AAEO,SAAS,gBAAgB,YAA2B;AACzD,QAAM,KAAK,MAAM;AACjB,MAAI,YAAY;AACd,UAAM,OAAO,GAAG,QAAQ,sCAAsC;AAC9D,eAAW,QAAQ,qBAAqB;AACtC,WAAK,IAAI,aAAa,KAAK,OAAO,UAAU,CAAC;AAAA,IAC/C;AACA,OAAG,QAAQ,kDAAkD,EAAE;AAAA,MAC7D;AAAA,IACF;AAAA,EACF,OAAO;AACL,OAAG,QAAQ,wBAAwB,EAAE,IAAI;AACzC,OAAG,QAAQ,iCAAiC,EAAE,IAAI;AAAA,EACpD;AACF;","names":["rawRows","rows"]}
@@ -19,7 +19,7 @@ import {
19
19
  } from "@sentry/core";
20
20
  var initialized = false;
21
21
  function getVersion() {
22
- return true ? "0.1.1+29c570b" : "dev";
22
+ return true ? "0.1.3+1deea73" : "dev";
23
23
  }
24
24
  var SCRUBBED_BREADCRUMB_FIELDS = [
25
25
  "prompt",
@@ -167,4 +167,4 @@ export {
167
167
  setTag,
168
168
  flushSentry
169
169
  };
170
- //# sourceMappingURL=chunk-XO5NQRTD.js.map
170
+ //# sourceMappingURL=chunk-RNICEX6N.js.map
@@ -1,13 +1,13 @@
1
1
  import {
2
2
  refreshPricing
3
- } from "./chunk-3TZAKV3M.js";
3
+ } from "./chunk-OSWBZKK5.js";
4
4
  import {
5
5
  allTargets
6
6
  } from "./chunk-QVK6VGCV.js";
7
7
  import {
8
8
  closeDb,
9
9
  getDb
10
- } from "./chunk-DZ5HJFB4.js";
10
+ } from "./chunk-UM5U6FPH.js";
11
11
  import {
12
12
  config,
13
13
  ensureDataDir
@@ -127,4 +127,4 @@ export {
127
127
  fetchPricing,
128
128
  configureShellEnv
129
129
  };
130
- //# sourceMappingURL=chunk-L7G27XWF.js.map
130
+ //# sourceMappingURL=chunk-RRHBTXHB.js.map
@@ -1,6 +1,6 @@
1
1
  import {
2
2
  getDb
3
- } from "./chunk-DZ5HJFB4.js";
3
+ } from "./chunk-UM5U6FPH.js";
4
4
 
5
5
  // src/db/store.ts
6
6
  import { createHash } from "crypto";
@@ -131,7 +131,8 @@ function insertUserConfigSnapshot(snap) {
131
131
  hooks: snap.hooks,
132
132
  commands: snap.commands,
133
133
  rules: snap.rules,
134
- skills: snap.skills
134
+ skills: snap.skills,
135
+ pluginHooks: snap.pluginHooks
135
136
  });
136
137
  const existing = db.prepare(
137
138
  "SELECT content_hash FROM user_config_snapshots WHERE device_name = ? ORDER BY snapshot_at_ms DESC LIMIT 1"
@@ -139,8 +140,8 @@ function insertUserConfigSnapshot(snap) {
139
140
  if (existing?.content_hash === hash) return false;
140
141
  db.prepare(
141
142
  `INSERT INTO user_config_snapshots
142
- (device_name, snapshot_at_ms, content_hash, permissions, enabled_plugins, hooks, commands, rules, skills)
143
- VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)`
143
+ (device_name, snapshot_at_ms, content_hash, permissions, enabled_plugins, hooks, commands, rules, skills, plugin_hooks)
144
+ VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`
144
145
  ).run(
145
146
  snap.deviceName,
146
147
  Date.now(),
@@ -150,7 +151,8 @@ function insertUserConfigSnapshot(snap) {
150
151
  JSON.stringify(snap.hooks),
151
152
  JSON.stringify(snap.commands),
152
153
  JSON.stringify(snap.rules),
153
- JSON.stringify(snap.skills)
154
+ JSON.stringify(snap.skills),
155
+ JSON.stringify(snap.pluginHooks)
154
156
  );
155
157
  return true;
156
158
  }
@@ -409,4 +411,4 @@ export {
409
411
  incrementEventTypeCount,
410
412
  insertHookEvent
411
413
  };
412
- //# sourceMappingURL=chunk-BVOE7A2Z.js.map
414
+ //# sourceMappingURL=chunk-UD6H6EP6.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/db/store.ts"],"sourcesContent":["import { createHash } from \"node:crypto\";\nimport { gzipSync } from \"node:zlib\";\n\nimport { getDb } from \"./schema.js\";\n\nexport interface OtelLogRow {\n timestamp_ns: number;\n observed_timestamp_ns?: number;\n severity_number?: number;\n severity_text?: string;\n body?: string;\n attributes?: Record<string, unknown>;\n resource_attributes?: Record<string, unknown>;\n session_id?: string;\n prompt_id?: string;\n trace_id?: string;\n span_id?: string;\n}\n\nexport interface OtelMetricRow {\n timestamp_ns: number;\n name: string;\n value: number;\n metric_type?: string;\n unit?: string;\n attributes?: Record<string, unknown>;\n resource_attributes?: Record<string, unknown>;\n session_id?: string;\n}\n\nexport interface OtelSpanRow {\n trace_id: string;\n span_id: string;\n parent_span_id?: string;\n name: string;\n kind?: number;\n start_time_ns: number;\n end_time_ns: number;\n status_code?: number;\n status_message?: string;\n attributes?: Record<string, unknown>;\n resource_attributes?: Record<string, unknown>;\n session_id?: string;\n}\n\nexport interface HookEventRow {\n session_id: string;\n event_type: string;\n timestamp_ms: number;\n cwd?: string;\n repository?: string;\n tool_name?: string;\n target?: string;\n user_prompt?: string;\n file_path?: string;\n command?: string;\n plan?: string;\n allowed_prompts?: string;\n payload: unknown;\n}\n\nconst INSERT_LOG_SQL = `\n INSERT INTO otel_logs (timestamp_ns, observed_timestamp_ns, severity_number, severity_text, body, attributes, resource_attributes, session_id, prompt_id, trace_id, span_id)\n VALUES (@timestamp_ns, @observed_timestamp_ns, @severity_number, @severity_text, @body, @attributes, @resource_attributes, @session_id, @prompt_id, @trace_id, @span_id)\n`;\n\nconst INSERT_METRIC_SQL = `\n INSERT INTO otel_metrics (timestamp_ns, name, value, metric_type, unit, attributes, resource_attributes, session_id)\n VALUES (@timestamp_ns, @name, @value, @metric_type, @unit, @attributes, @resource_attributes, @session_id)\n`;\n\nconst INSERT_SPAN_SQL = `\n INSERT OR IGNORE INTO otel_spans (trace_id, span_id, parent_span_id, name, kind, start_time_ns, end_time_ns, status_code, status_message, attributes, resource_attributes, session_id)\n VALUES (@trace_id, @span_id, @parent_span_id, @name, @kind, @start_time_ns, @end_time_ns, @status_code, @status_message, @attributes, @resource_attributes, @session_id)\n`;\n\nconst INSERT_HOOK_SQL = `\n INSERT INTO hook_events (session_id, event_type, timestamp_ms, cwd, repository, tool_name,\n target, user_prompt, file_path, command, tool_result, plan, allowed_prompts, payload)\n VALUES (@session_id, @event_type, @timestamp_ms, @cwd, @repository, @tool_name,\n @target, @user_prompt, @file_path, @command, @tool_result, @plan, @allowed_prompts, @payload)\n`;\n\nexport function insertOtelLogs(rows: OtelLogRow[]): void {\n const db = getDb();\n const stmt = db.prepare(INSERT_LOG_SQL);\n const insertMany = db.transaction((rows: OtelLogRow[]) => {\n for (const row of rows) {\n stmt.run({\n timestamp_ns: row.timestamp_ns,\n observed_timestamp_ns: row.observed_timestamp_ns ?? null,\n severity_number: row.severity_number ?? null,\n severity_text: row.severity_text ?? null,\n body: row.body ?? null,\n attributes: row.attributes ? JSON.stringify(row.attributes) : null,\n resource_attributes: row.resource_attributes\n ? JSON.stringify(row.resource_attributes)\n : null,\n session_id: row.session_id ?? null,\n prompt_id: row.prompt_id ?? null,\n trace_id: row.trace_id ?? null,\n span_id: row.span_id ?? null,\n });\n }\n });\n insertMany(rows);\n}\n\nexport function insertOtelMetrics(rows: OtelMetricRow[]): void {\n const db = getDb();\n const stmt = db.prepare(INSERT_METRIC_SQL);\n\n const insertMany = db.transaction((rows: OtelMetricRow[]) => {\n for (const row of rows) {\n const sessionId = row.session_id ?? null;\n\n stmt.run({\n timestamp_ns: row.timestamp_ns,\n name: row.name,\n value: row.value,\n metric_type: row.metric_type ?? null,\n unit: row.unit ?? null,\n attributes: row.attributes ? JSON.stringify(row.attributes) : null,\n resource_attributes: row.resource_attributes\n ? JSON.stringify(row.resource_attributes)\n : null,\n session_id: sessionId,\n });\n }\n });\n insertMany(rows);\n}\n\nexport function insertOtelSpans(rows: OtelSpanRow[]): void {\n const db = getDb();\n const stmt = db.prepare(INSERT_SPAN_SQL);\n\n const insertMany = db.transaction((rows: OtelSpanRow[]) => {\n for (const row of rows) {\n stmt.run({\n trace_id: row.trace_id,\n span_id: row.span_id,\n parent_span_id: row.parent_span_id ?? null,\n name: row.name,\n kind: row.kind ?? null,\n start_time_ns: row.start_time_ns,\n end_time_ns: row.end_time_ns,\n status_code: row.status_code ?? null,\n status_message: row.status_message ?? null,\n attributes: row.attributes ? JSON.stringify(row.attributes) : null,\n resource_attributes: row.resource_attributes\n ? JSON.stringify(row.resource_attributes)\n : null,\n session_id: row.session_id ?? null,\n });\n }\n });\n insertMany(rows);\n}\n\nexport function upsertSessionRepository(\n sessionId: string,\n repository: string,\n timestampMs: number,\n gitIdentity?: { name: string | null; email: string | null },\n branch?: string | null,\n): void {\n const db = getDb();\n const existing = db\n .prepare(\n `SELECT 1 FROM session_repositories WHERE session_id = ? AND repository = ?`,\n )\n .get(sessionId, repository);\n\n db.prepare(\n `INSERT INTO session_repositories (session_id, repository, first_seen_ms, git_user_name, git_user_email, branch)\n VALUES (?, ?, ?, ?, ?, ?)\n ON CONFLICT(session_id, repository) DO UPDATE SET\n git_user_name = COALESCE(session_repositories.git_user_name, excluded.git_user_name),\n git_user_email = COALESCE(session_repositories.git_user_email, excluded.git_user_email),\n branch = COALESCE(excluded.branch, session_repositories.branch)`,\n ).run(\n sessionId,\n repository,\n timestampMs,\n gitIdentity?.name ?? null,\n gitIdentity?.email ?? null,\n branch ?? null,\n );\n\n // When a NEW repo is associated, bump sync_seq so the session re-syncs\n // with updated repository info (avoids backend repo-filter rejections)\n if (!existing) {\n db.prepare(\n `UPDATE sessions SET sync_seq = COALESCE(sync_seq, 0) + 1 WHERE session_id = ?`,\n ).run(sessionId);\n }\n}\n\nexport function upsertSessionCwd(\n sessionId: string,\n cwd: string,\n timestampMs: number,\n): void {\n const db = getDb();\n db.prepare(\n \"INSERT INTO session_cwds (session_id, cwd, first_seen_ms) VALUES (?, ?, ?) ON CONFLICT DO NOTHING\",\n ).run(sessionId, cwd, timestampMs);\n}\n\n// ---------------------------------------------------------------------------\n// Config snapshots\n// ---------------------------------------------------------------------------\n\nfunction contentHash(obj: Record<string, unknown>): string {\n return createHash(\"sha256\")\n .update(JSON.stringify(obj, Object.keys(obj).sort()))\n .digest(\"hex\");\n}\n\nexport interface UserConfigSnapshot {\n deviceName: string;\n permissions: unknown;\n enabledPlugins: unknown;\n hooks: unknown;\n commands: unknown;\n rules: unknown;\n skills: unknown;\n pluginHooks: unknown;\n}\n\n/**\n * Insert a user config snapshot if the content has changed since the last one\n * for this device. Returns true if a new row was inserted.\n */\nexport function insertUserConfigSnapshot(snap: UserConfigSnapshot): boolean {\n const db = getDb();\n const hash = contentHash({\n permissions: snap.permissions,\n enabledPlugins: snap.enabledPlugins,\n hooks: snap.hooks,\n commands: snap.commands,\n rules: snap.rules,\n skills: snap.skills,\n pluginHooks: snap.pluginHooks,\n });\n\n // Check if latest snapshot for this device has the same hash\n const existing = db\n .prepare(\n \"SELECT content_hash FROM user_config_snapshots WHERE device_name = ? ORDER BY snapshot_at_ms DESC LIMIT 1\",\n )\n .get(snap.deviceName) as { content_hash: string } | undefined;\n\n if (existing?.content_hash === hash) return false;\n\n db.prepare(\n `INSERT INTO user_config_snapshots\n (device_name, snapshot_at_ms, content_hash, permissions, enabled_plugins, hooks, commands, rules, skills, plugin_hooks)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`,\n ).run(\n snap.deviceName,\n Date.now(),\n hash,\n JSON.stringify(snap.permissions),\n JSON.stringify(snap.enabledPlugins),\n JSON.stringify(snap.hooks),\n JSON.stringify(snap.commands),\n JSON.stringify(snap.rules),\n JSON.stringify(snap.skills),\n JSON.stringify(snap.pluginHooks),\n );\n return true;\n}\n\nexport interface RepoConfigSnapshot {\n repository: string;\n cwd: string;\n sessionId?: string;\n // project layer\n hooks: unknown;\n mcpServers: unknown;\n commands: unknown;\n agents: unknown;\n rules: unknown;\n // project local layer\n localHooks: unknown;\n localMcpServers: unknown;\n localPermissions: unknown;\n localIsGitignored: boolean;\n // instructions\n instructions: unknown;\n}\n\n/**\n * Insert a repo config snapshot if the content has changed since the last one\n * for this repository. Returns true if a new row was inserted.\n */\nexport function insertRepoConfigSnapshot(snap: RepoConfigSnapshot): boolean {\n const db = getDb();\n const hash = contentHash({\n hooks: snap.hooks,\n mcpServers: snap.mcpServers,\n commands: snap.commands,\n agents: snap.agents,\n rules: snap.rules,\n localHooks: snap.localHooks,\n localMcpServers: snap.localMcpServers,\n localPermissions: snap.localPermissions,\n localIsGitignored: snap.localIsGitignored,\n instructions: snap.instructions,\n });\n\n const existing = db\n .prepare(\n \"SELECT content_hash FROM repo_config_snapshots WHERE repository = ? ORDER BY snapshot_at_ms DESC LIMIT 1\",\n )\n .get(snap.repository) as { content_hash: string } | undefined;\n\n if (existing?.content_hash === hash) return false;\n\n db.prepare(\n `INSERT INTO repo_config_snapshots\n (repository, cwd, session_id, snapshot_at_ms, content_hash,\n hooks, mcp_servers, commands, agents, rules,\n local_hooks, local_mcp_servers, local_permissions, local_is_gitignored,\n instructions)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)`,\n ).run(\n snap.repository,\n snap.cwd,\n snap.sessionId ?? null,\n Date.now(),\n hash,\n JSON.stringify(snap.hooks),\n JSON.stringify(snap.mcpServers),\n JSON.stringify(snap.commands),\n JSON.stringify(snap.agents),\n JSON.stringify(snap.rules),\n JSON.stringify(snap.localHooks),\n JSON.stringify(snap.localMcpServers),\n JSON.stringify(snap.localPermissions),\n snap.localIsGitignored ? 1 : 0,\n JSON.stringify(snap.instructions),\n );\n return true;\n}\n\nexport interface SessionUpsert {\n session_id: string;\n target?: string;\n started_at_ms?: number;\n ended_at_ms?: number;\n first_prompt?: string;\n permission_mode?: string;\n agent_version?: string;\n // Scanner-sourced fields\n model?: string;\n cli_version?: string;\n scanner_file_path?: string;\n total_input_tokens?: number;\n total_output_tokens?: number;\n total_cache_read_tokens?: number;\n total_cache_creation_tokens?: number;\n total_reasoning_tokens?: number;\n turn_count?: number;\n // OTLP-sourced tokens\n otel_input_tokens?: number;\n otel_output_tokens?: number;\n otel_cache_read_tokens?: number;\n otel_cache_creation_tokens?: number;\n // Metadata\n project?: string;\n created_at?: number;\n has_hooks?: number;\n has_otel?: number;\n has_scanner?: number;\n parent_session_id?: string;\n relationship_type?: string;\n is_automated?: number;\n}\n\nexport function upsertSession(row: SessionUpsert): void {\n const db = getDb();\n db.prepare(\n `INSERT INTO sessions (session_id, target, started_at_ms, ended_at_ms, first_prompt,\n permission_mode, agent_version, model, cli_version, scanner_file_path,\n total_input_tokens, total_output_tokens, total_cache_read_tokens,\n total_cache_creation_tokens, total_reasoning_tokens, turn_count,\n otel_input_tokens, otel_output_tokens, otel_cache_read_tokens, otel_cache_creation_tokens,\n models, project, created_at, parent_session_id, relationship_type, is_automated,\n has_hooks, has_otel, has_scanner)\n VALUES (@session_id, @target, @started_at_ms, @ended_at_ms, @first_prompt,\n @permission_mode, @agent_version, @model, @cli_version, @scanner_file_path,\n @total_input_tokens, @total_output_tokens, @total_cache_read_tokens,\n @total_cache_creation_tokens, @total_reasoning_tokens, @turn_count,\n @otel_input_tokens, @otel_output_tokens, @otel_cache_read_tokens, @otel_cache_creation_tokens,\n @model, @project, @created_at, @parent_session_id, @relationship_type, @is_automated,\n @has_hooks, @has_otel, @has_scanner)\n ON CONFLICT(session_id) DO UPDATE SET\n target = COALESCE(excluded.target, sessions.target),\n started_at_ms = COALESCE(excluded.started_at_ms, sessions.started_at_ms),\n ended_at_ms = COALESCE(excluded.ended_at_ms, sessions.ended_at_ms),\n first_prompt = COALESCE(sessions.first_prompt, excluded.first_prompt),\n permission_mode = COALESCE(excluded.permission_mode, sessions.permission_mode),\n agent_version = COALESCE(excluded.agent_version, sessions.agent_version),\n model = COALESCE(excluded.model, sessions.model),\n cli_version = COALESCE(excluded.cli_version, sessions.cli_version),\n scanner_file_path = COALESCE(excluded.scanner_file_path, sessions.scanner_file_path),\n total_input_tokens = COALESCE(excluded.total_input_tokens, sessions.total_input_tokens),\n total_output_tokens = COALESCE(excluded.total_output_tokens, sessions.total_output_tokens),\n total_cache_read_tokens = COALESCE(excluded.total_cache_read_tokens, sessions.total_cache_read_tokens),\n total_cache_creation_tokens = COALESCE(excluded.total_cache_creation_tokens, sessions.total_cache_creation_tokens),\n total_reasoning_tokens = COALESCE(excluded.total_reasoning_tokens, sessions.total_reasoning_tokens),\n turn_count = COALESCE(excluded.turn_count, sessions.turn_count),\n otel_input_tokens = COALESCE(excluded.otel_input_tokens, sessions.otel_input_tokens),\n otel_output_tokens = COALESCE(excluded.otel_output_tokens, sessions.otel_output_tokens),\n otel_cache_read_tokens = COALESCE(excluded.otel_cache_read_tokens, sessions.otel_cache_read_tokens),\n otel_cache_creation_tokens = COALESCE(excluded.otel_cache_creation_tokens, sessions.otel_cache_creation_tokens),\n models = CASE\n WHEN excluded.model IS NULL THEN sessions.models\n WHEN sessions.models IS NULL THEN excluded.model\n WHEN sessions.models LIKE '%' || excluded.model || '%' THEN sessions.models\n ELSE sessions.models || ',' || excluded.model\n END,\n project = COALESCE(sessions.project, excluded.project),\n parent_session_id = COALESCE(excluded.parent_session_id, sessions.parent_session_id),\n relationship_type = COALESCE(excluded.relationship_type, sessions.relationship_type),\n is_automated = COALESCE(excluded.is_automated, sessions.is_automated),\n has_hooks = MAX(COALESCE(sessions.has_hooks, 0), COALESCE(excluded.has_hooks, 0)),\n has_otel = MAX(COALESCE(sessions.has_otel, 0), COALESCE(excluded.has_otel, 0)),\n has_scanner = MAX(COALESCE(sessions.has_scanner, 0), COALESCE(excluded.has_scanner, 0)),\n sync_dirty = 1,\n sync_seq = COALESCE(sessions.sync_seq, 0) + 1`,\n ).run({\n session_id: row.session_id,\n target: row.target ?? null,\n started_at_ms: row.started_at_ms ?? null,\n ended_at_ms: row.ended_at_ms ?? null,\n first_prompt: row.first_prompt ?? null,\n permission_mode: row.permission_mode ?? null,\n agent_version: row.agent_version ?? null,\n model: row.model ?? null,\n cli_version: row.cli_version ?? null,\n scanner_file_path: row.scanner_file_path ?? null,\n total_input_tokens: row.total_input_tokens ?? null,\n total_output_tokens: row.total_output_tokens ?? null,\n total_cache_read_tokens: row.total_cache_read_tokens ?? null,\n total_cache_creation_tokens: row.total_cache_creation_tokens ?? null,\n total_reasoning_tokens: row.total_reasoning_tokens ?? null,\n turn_count: row.turn_count ?? null,\n otel_input_tokens: row.otel_input_tokens ?? null,\n otel_output_tokens: row.otel_output_tokens ?? null,\n otel_cache_read_tokens: row.otel_cache_read_tokens ?? null,\n otel_cache_creation_tokens: row.otel_cache_creation_tokens ?? null,\n project: row.project ?? null,\n created_at: row.created_at ?? null,\n parent_session_id: row.parent_session_id ?? null,\n relationship_type: row.relationship_type ?? null,\n is_automated: row.is_automated ?? null,\n has_hooks: row.has_hooks ?? null,\n has_otel: row.has_otel ?? null,\n has_scanner: row.has_scanner ?? null,\n });\n}\n\n/** Prefixes/substrings that identify automated (non-interactive) sessions. */\nconst AUTOMATED_PREFIXES = [\n \"You are a code reviewer.\",\n \"You are a security code reviewer.\",\n \"You are a design reviewer.\",\n \"You are a code assistant. Your task is to address\",\n \"You are a code review insights analyst.\",\n \"You are reviewing whether an implementation matches\",\n \"You are a plan document reviewer.\",\n \"You are a spec document reviewer.\",\n \"You are summarizing a day of AI agent activity.\",\n \"You are analyzing AI agent sessions.\",\n \"## Analysis Request\",\n \"# Fix Request\",\n];\nconst AUTOMATED_SUBSTRINGS = [\"invoked by roborev to perform this review\"];\n\nfunction isAutomatedPrompt(firstPrompt: string): boolean {\n for (const p of AUTOMATED_PREFIXES) {\n if (firstPrompt.startsWith(p)) return true;\n }\n for (const s of AUTOMATED_SUBSTRINGS) {\n if (firstPrompt.includes(s)) return true;\n }\n return false;\n}\n\n/**\n * Recompute message_count, user_message_count, and is_automated\n * from the messages table. is_automated is set when user_message_count <= 1\n * and first_prompt matches a known automated pattern.\n */\nexport function updateSessionMessageCounts(sessionId: string): void {\n const db = getDb();\n\n // Count non-system user messages\n const counts = db\n .prepare(\n `SELECT\n (SELECT COUNT(*) FROM messages WHERE session_id = ?) as msg_count,\n (SELECT COUNT(*) FROM messages WHERE session_id = ? AND role = 'user' AND is_system = 0) as user_count,\n (SELECT first_prompt FROM sessions WHERE session_id = ?) as first_prompt`,\n )\n .get(sessionId, sessionId, sessionId) as {\n msg_count: number;\n user_count: number;\n first_prompt: string | null;\n };\n\n const isAutomated =\n counts.user_count <= 1 &&\n counts.first_prompt != null &&\n isAutomatedPrompt(counts.first_prompt)\n ? 1\n : 0;\n\n db.prepare(\n `UPDATE sessions SET\n message_count = ?,\n user_message_count = ?,\n is_automated = CASE WHEN ? > 1 THEN 0 ELSE ? END,\n sync_seq = COALESCE(sync_seq, 0) + 1\n WHERE session_id = ?`,\n ).run(\n counts.msg_count,\n counts.user_count,\n counts.user_count,\n isAutomated,\n sessionId,\n );\n}\n\n/**\n * Increment a hook tool count for a session. Uses JSON_SET to atomically\n * update the hook_tool_counts JSON object.\n * Does NOT bump sync_seq — the scanner drives sync_seq via updateSessionTotals.\n * Hook data syncs via Phase 2 (hook_events table), not the session row.\n */\nexport function incrementToolCount(sessionId: string, toolName: string): void {\n const db = getDb();\n db.prepare(\n `UPDATE sessions\n SET hook_tool_counts = JSON_SET(\n COALESCE(hook_tool_counts, '{}'),\n '$.' || @tool,\n COALESCE(JSON_EXTRACT(hook_tool_counts, '$.' || @tool), 0) + 1\n )\n WHERE session_id = @session_id`,\n ).run({ session_id: sessionId, tool: toolName });\n}\n\n/**\n * Increment a hook event type count for a session.\n * Does NOT bump sync_seq — same rationale as incrementToolCount.\n */\nexport function incrementEventTypeCount(\n sessionId: string,\n eventType: string,\n): void {\n const db = getDb();\n db.prepare(\n `UPDATE sessions\n SET hook_event_type_counts = JSON_SET(\n COALESCE(hook_event_type_counts, '{}'),\n '$.' || @event_type,\n COALESCE(JSON_EXTRACT(hook_event_type_counts, '$.' || @event_type), 0) + 1\n )\n WHERE session_id = @session_id`,\n ).run({ session_id: sessionId, event_type: eventType });\n}\n\nfunction extractStr(\n obj: Record<string, unknown> | undefined,\n key: string,\n): string | undefined {\n const v = obj?.[key];\n return typeof v === \"string\" ? v : undefined;\n}\n\nexport function insertHookEvent(row: HookEventRow): void {\n const db = getDb();\n const data = row.payload as Record<string, unknown>;\n const toolInput = data.tool_input as Record<string, unknown> | undefined;\n\n // Extract high-value fields into columns for indexed queries\n const userPrompt =\n extractStr(data, \"prompt\") ?? extractStr(data, \"user_prompt\");\n const filePath = extractStr(toolInput, \"file_path\");\n const command = extractStr(toolInput, \"command\");\n const plan = extractStr(toolInput, \"plan\");\n const toolResultRaw = data.tool_result ?? data.tool_response;\n const toolResult = toolResultRaw\n ? typeof toolResultRaw === \"string\"\n ? toolResultRaw\n : JSON.stringify(toolResultRaw)\n : undefined;\n const allowedPrompts = toolInput?.allowedPrompts\n ? JSON.stringify(toolInput.allowedPrompts)\n : undefined;\n\n const fullJson = JSON.stringify(data);\n\n const insertWithFts = db.transaction(() => {\n db.prepare(INSERT_HOOK_SQL).run({\n session_id: row.session_id,\n event_type: row.event_type,\n timestamp_ms: row.timestamp_ms,\n cwd: row.cwd ?? null,\n repository: row.repository ?? null,\n tool_name: row.tool_name ?? null,\n target: row.target ?? null,\n user_prompt: userPrompt ?? null,\n file_path: filePath ?? null,\n command: command ?? null,\n tool_result: toolResult ?? null,\n plan: plan ?? null,\n allowed_prompts: allowedPrompts ?? null,\n payload: gzipSync(Buffer.from(fullJson)),\n });\n const { id } = db.prepare(\"SELECT last_insert_rowid() as id\").get() as {\n id: number;\n };\n db.prepare(\"INSERT INTO hook_events_fts(rowid, payload) VALUES (?, ?)\").run(\n id,\n fullJson,\n );\n });\n insertWithFts();\n}\n"],"mappings":";;;;;AAAA,SAAS,kBAAkB;AAC3B,SAAS,gBAAgB;AA4DzB,IAAM,iBAAiB;AAAA;AAAA;AAAA;AAKvB,IAAM,oBAAoB;AAAA;AAAA;AAAA;AAK1B,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAKxB,IAAM,kBAAkB;AAAA;AAAA;AAAA;AAAA;AAAA;AAOjB,SAAS,eAAe,MAA0B;AACvD,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,GAAG,QAAQ,cAAc;AACtC,QAAM,aAAa,GAAG,YAAY,CAACA,UAAuB;AACxD,eAAW,OAAOA,OAAM;AACtB,WAAK,IAAI;AAAA,QACP,cAAc,IAAI;AAAA,QAClB,uBAAuB,IAAI,yBAAyB;AAAA,QACpD,iBAAiB,IAAI,mBAAmB;AAAA,QACxC,eAAe,IAAI,iBAAiB;AAAA,QACpC,MAAM,IAAI,QAAQ;AAAA,QAClB,YAAY,IAAI,aAAa,KAAK,UAAU,IAAI,UAAU,IAAI;AAAA,QAC9D,qBAAqB,IAAI,sBACrB,KAAK,UAAU,IAAI,mBAAmB,IACtC;AAAA,QACJ,YAAY,IAAI,cAAc;AAAA,QAC9B,WAAW,IAAI,aAAa;AAAA,QAC5B,UAAU,IAAI,YAAY;AAAA,QAC1B,SAAS,IAAI,WAAW;AAAA,MAC1B,CAAC;AAAA,IACH;AAAA,EACF,CAAC;AACD,aAAW,IAAI;AACjB;AAEO,SAAS,kBAAkB,MAA6B;AAC7D,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,GAAG,QAAQ,iBAAiB;AAEzC,QAAM,aAAa,GAAG,YAAY,CAACA,UAA0B;AAC3D,eAAW,OAAOA,OAAM;AACtB,YAAM,YAAY,IAAI,cAAc;AAEpC,WAAK,IAAI;AAAA,QACP,cAAc,IAAI;AAAA,QAClB,MAAM,IAAI;AAAA,QACV,OAAO,IAAI;AAAA,QACX,aAAa,IAAI,eAAe;AAAA,QAChC,MAAM,IAAI,QAAQ;AAAA,QAClB,YAAY,IAAI,aAAa,KAAK,UAAU,IAAI,UAAU,IAAI;AAAA,QAC9D,qBAAqB,IAAI,sBACrB,KAAK,UAAU,IAAI,mBAAmB,IACtC;AAAA,QACJ,YAAY;AAAA,MACd,CAAC;AAAA,IACH;AAAA,EACF,CAAC;AACD,aAAW,IAAI;AACjB;AAEO,SAAS,gBAAgB,MAA2B;AACzD,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,GAAG,QAAQ,eAAe;AAEvC,QAAM,aAAa,GAAG,YAAY,CAACA,UAAwB;AACzD,eAAW,OAAOA,OAAM;AACtB,WAAK,IAAI;AAAA,QACP,UAAU,IAAI;AAAA,QACd,SAAS,IAAI;AAAA,QACb,gBAAgB,IAAI,kBAAkB;AAAA,QACtC,MAAM,IAAI;AAAA,QACV,MAAM,IAAI,QAAQ;AAAA,QAClB,eAAe,IAAI;AAAA,QACnB,aAAa,IAAI;AAAA,QACjB,aAAa,IAAI,eAAe;AAAA,QAChC,gBAAgB,IAAI,kBAAkB;AAAA,QACtC,YAAY,IAAI,aAAa,KAAK,UAAU,IAAI,UAAU,IAAI;AAAA,QAC9D,qBAAqB,IAAI,sBACrB,KAAK,UAAU,IAAI,mBAAmB,IACtC;AAAA,QACJ,YAAY,IAAI,cAAc;AAAA,MAChC,CAAC;AAAA,IACH;AAAA,EACF,CAAC;AACD,aAAW,IAAI;AACjB;AAEO,SAAS,wBACd,WACA,YACA,aACA,aACA,QACM;AACN,QAAM,KAAK,MAAM;AACjB,QAAM,WAAW,GACd;AAAA,IACC;AAAA,EACF,EACC,IAAI,WAAW,UAAU;AAE5B,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EAAE;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,aAAa,QAAQ;AAAA,IACrB,aAAa,SAAS;AAAA,IACtB,UAAU;AAAA,EACZ;AAIA,MAAI,CAAC,UAAU;AACb,OAAG;AAAA,MACD;AAAA,IACF,EAAE,IAAI,SAAS;AAAA,EACjB;AACF;AAEO,SAAS,iBACd,WACA,KACA,aACM;AACN,QAAM,KAAK,MAAM;AACjB,KAAG;AAAA,IACD;AAAA,EACF,EAAE,IAAI,WAAW,KAAK,WAAW;AACnC;AAMA,SAAS,YAAY,KAAsC;AACzD,SAAO,WAAW,QAAQ,EACvB,OAAO,KAAK,UAAU,KAAK,OAAO,KAAK,GAAG,EAAE,KAAK,CAAC,CAAC,EACnD,OAAO,KAAK;AACjB;AAiBO,SAAS,yBAAyB,MAAmC;AAC1E,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,YAAY;AAAA,IACvB,aAAa,KAAK;AAAA,IAClB,gBAAgB,KAAK;AAAA,IACrB,OAAO,KAAK;AAAA,IACZ,UAAU,KAAK;AAAA,IACf,OAAO,KAAK;AAAA,IACZ,QAAQ,KAAK;AAAA,IACb,aAAa,KAAK;AAAA,EACpB,CAAC;AAGD,QAAM,WAAW,GACd;AAAA,IACC;AAAA,EACF,EACC,IAAI,KAAK,UAAU;AAEtB,MAAI,UAAU,iBAAiB,KAAM,QAAO;AAE5C,KAAG;AAAA,IACD;AAAA;AAAA;AAAA,EAGF,EAAE;AAAA,IACA,KAAK;AAAA,IACL,KAAK,IAAI;AAAA,IACT;AAAA,IACA,KAAK,UAAU,KAAK,WAAW;AAAA,IAC/B,KAAK,UAAU,KAAK,cAAc;AAAA,IAClC,KAAK,UAAU,KAAK,KAAK;AAAA,IACzB,KAAK,UAAU,KAAK,QAAQ;AAAA,IAC5B,KAAK,UAAU,KAAK,KAAK;AAAA,IACzB,KAAK,UAAU,KAAK,MAAM;AAAA,IAC1B,KAAK,UAAU,KAAK,WAAW;AAAA,EACjC;AACA,SAAO;AACT;AAyBO,SAAS,yBAAyB,MAAmC;AAC1E,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,YAAY;AAAA,IACvB,OAAO,KAAK;AAAA,IACZ,YAAY,KAAK;AAAA,IACjB,UAAU,KAAK;AAAA,IACf,QAAQ,KAAK;AAAA,IACb,OAAO,KAAK;AAAA,IACZ,YAAY,KAAK;AAAA,IACjB,iBAAiB,KAAK;AAAA,IACtB,kBAAkB,KAAK;AAAA,IACvB,mBAAmB,KAAK;AAAA,IACxB,cAAc,KAAK;AAAA,EACrB,CAAC;AAED,QAAM,WAAW,GACd;AAAA,IACC;AAAA,EACF,EACC,IAAI,KAAK,UAAU;AAEtB,MAAI,UAAU,iBAAiB,KAAM,QAAO;AAE5C,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EAAE;AAAA,IACA,KAAK;AAAA,IACL,KAAK;AAAA,IACL,KAAK,aAAa;AAAA,IAClB,KAAK,IAAI;AAAA,IACT;AAAA,IACA,KAAK,UAAU,KAAK,KAAK;AAAA,IACzB,KAAK,UAAU,KAAK,UAAU;AAAA,IAC9B,KAAK,UAAU,KAAK,QAAQ;AAAA,IAC5B,KAAK,UAAU,KAAK,MAAM;AAAA,IAC1B,KAAK,UAAU,KAAK,KAAK;AAAA,IACzB,KAAK,UAAU,KAAK,UAAU;AAAA,IAC9B,KAAK,UAAU,KAAK,eAAe;AAAA,IACnC,KAAK,UAAU,KAAK,gBAAgB;AAAA,IACpC,KAAK,oBAAoB,IAAI;AAAA,IAC7B,KAAK,UAAU,KAAK,YAAY;AAAA,EAClC;AACA,SAAO;AACT;AAoCO,SAAS,cAAc,KAA0B;AACtD,QAAM,KAAK,MAAM;AACjB,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAiDF,EAAE,IAAI;AAAA,IACJ,YAAY,IAAI;AAAA,IAChB,QAAQ,IAAI,UAAU;AAAA,IACtB,eAAe,IAAI,iBAAiB;AAAA,IACpC,aAAa,IAAI,eAAe;AAAA,IAChC,cAAc,IAAI,gBAAgB;AAAA,IAClC,iBAAiB,IAAI,mBAAmB;AAAA,IACxC,eAAe,IAAI,iBAAiB;AAAA,IACpC,OAAO,IAAI,SAAS;AAAA,IACpB,aAAa,IAAI,eAAe;AAAA,IAChC,mBAAmB,IAAI,qBAAqB;AAAA,IAC5C,oBAAoB,IAAI,sBAAsB;AAAA,IAC9C,qBAAqB,IAAI,uBAAuB;AAAA,IAChD,yBAAyB,IAAI,2BAA2B;AAAA,IACxD,6BAA6B,IAAI,+BAA+B;AAAA,IAChE,wBAAwB,IAAI,0BAA0B;AAAA,IACtD,YAAY,IAAI,cAAc;AAAA,IAC9B,mBAAmB,IAAI,qBAAqB;AAAA,IAC5C,oBAAoB,IAAI,sBAAsB;AAAA,IAC9C,wBAAwB,IAAI,0BAA0B;AAAA,IACtD,4BAA4B,IAAI,8BAA8B;AAAA,IAC9D,SAAS,IAAI,WAAW;AAAA,IACxB,YAAY,IAAI,cAAc;AAAA,IAC9B,mBAAmB,IAAI,qBAAqB;AAAA,IAC5C,mBAAmB,IAAI,qBAAqB;AAAA,IAC5C,cAAc,IAAI,gBAAgB;AAAA,IAClC,WAAW,IAAI,aAAa;AAAA,IAC5B,UAAU,IAAI,YAAY;AAAA,IAC1B,aAAa,IAAI,eAAe;AAAA,EAClC,CAAC;AACH;AAGA,IAAM,qBAAqB;AAAA,EACzB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AACA,IAAM,uBAAuB,CAAC,2CAA2C;AAEzE,SAAS,kBAAkB,aAA8B;AACvD,aAAW,KAAK,oBAAoB;AAClC,QAAI,YAAY,WAAW,CAAC,EAAG,QAAO;AAAA,EACxC;AACA,aAAW,KAAK,sBAAsB;AACpC,QAAI,YAAY,SAAS,CAAC,EAAG,QAAO;AAAA,EACtC;AACA,SAAO;AACT;AAOO,SAAS,2BAA2B,WAAyB;AAClE,QAAM,KAAK,MAAM;AAGjB,QAAM,SAAS,GACZ;AAAA,IACC;AAAA;AAAA;AAAA;AAAA,EAIF,EACC,IAAI,WAAW,WAAW,SAAS;AAMtC,QAAM,cACJ,OAAO,cAAc,KACrB,OAAO,gBAAgB,QACvB,kBAAkB,OAAO,YAAY,IACjC,IACA;AAEN,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAMF,EAAE;AAAA,IACA,OAAO;AAAA,IACP,OAAO;AAAA,IACP,OAAO;AAAA,IACP;AAAA,IACA;AAAA,EACF;AACF;AAQO,SAAS,mBAAmB,WAAmB,UAAwB;AAC5E,QAAM,KAAK,MAAM;AACjB,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOF,EAAE,IAAI,EAAE,YAAY,WAAW,MAAM,SAAS,CAAC;AACjD;AAMO,SAAS,wBACd,WACA,WACM;AACN,QAAM,KAAK,MAAM;AACjB,KAAG;AAAA,IACD;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOF,EAAE,IAAI,EAAE,YAAY,WAAW,YAAY,UAAU,CAAC;AACxD;AAEA,SAAS,WACP,KACA,KACoB;AACpB,QAAM,IAAI,MAAM,GAAG;AACnB,SAAO,OAAO,MAAM,WAAW,IAAI;AACrC;AAEO,SAAS,gBAAgB,KAAyB;AACvD,QAAM,KAAK,MAAM;AACjB,QAAM,OAAO,IAAI;AACjB,QAAM,YAAY,KAAK;AAGvB,QAAM,aACJ,WAAW,MAAM,QAAQ,KAAK,WAAW,MAAM,aAAa;AAC9D,QAAM,WAAW,WAAW,WAAW,WAAW;AAClD,QAAM,UAAU,WAAW,WAAW,SAAS;AAC/C,QAAM,OAAO,WAAW,WAAW,MAAM;AACzC,QAAM,gBAAgB,KAAK,eAAe,KAAK;AAC/C,QAAM,aAAa,gBACf,OAAO,kBAAkB,WACvB,gBACA,KAAK,UAAU,aAAa,IAC9B;AACJ,QAAM,iBAAiB,WAAW,iBAC9B,KAAK,UAAU,UAAU,cAAc,IACvC;AAEJ,QAAM,WAAW,KAAK,UAAU,IAAI;AAEpC,QAAM,gBAAgB,GAAG,YAAY,MAAM;AACzC,OAAG,QAAQ,eAAe,EAAE,IAAI;AAAA,MAC9B,YAAY,IAAI;AAAA,MAChB,YAAY,IAAI;AAAA,MAChB,cAAc,IAAI;AAAA,MAClB,KAAK,IAAI,OAAO;AAAA,MAChB,YAAY,IAAI,cAAc;AAAA,MAC9B,WAAW,IAAI,aAAa;AAAA,MAC5B,QAAQ,IAAI,UAAU;AAAA,MACtB,aAAa,cAAc;AAAA,MAC3B,WAAW,YAAY;AAAA,MACvB,SAAS,WAAW;AAAA,MACpB,aAAa,cAAc;AAAA,MAC3B,MAAM,QAAQ;AAAA,MACd,iBAAiB,kBAAkB;AAAA,MACnC,SAAS,SAAS,OAAO,KAAK,QAAQ,CAAC;AAAA,IACzC,CAAC;AACD,UAAM,EAAE,GAAG,IAAI,GAAG,QAAQ,kCAAkC,EAAE,IAAI;AAGlE,OAAG,QAAQ,2DAA2D,EAAE;AAAA,MACtE;AAAA,MACA;AAAA,IACF;AAAA,EACF,CAAC;AACD,gBAAc;AAChB;","names":["rows"]}
@@ -3,7 +3,7 @@ import {
3
3
  } from "./chunk-QVK6VGCV.js";
4
4
  import {
5
5
  getDb
6
- } from "./chunk-DZ5HJFB4.js";
6
+ } from "./chunk-UM5U6FPH.js";
7
7
 
8
8
  // src/db/query.ts
9
9
  function parseSince(since) {
@@ -623,4 +623,4 @@ export {
623
623
  rawQuery,
624
624
  dbStats
625
625
  };
626
- //# sourceMappingURL=chunk-LWXF7YRG.js.map
626
+ //# sourceMappingURL=chunk-UDUBRSIG.js.map
@@ -1,20 +1,20 @@
1
1
  import {
2
2
  readWatermark,
3
3
  watermarkKey
4
- } from "./chunk-SEXU2WYG.js";
4
+ } from "./chunk-P5KAIE3O.js";
5
5
  import {
6
6
  loadUnifiedConfig
7
7
  } from "./chunk-QK5442ZP.js";
8
8
  import {
9
9
  dbStats
10
- } from "./chunk-LWXF7YRG.js";
10
+ } from "./chunk-UDUBRSIG.js";
11
11
  import {
12
12
  allTargets
13
13
  } from "./chunk-QVK6VGCV.js";
14
14
  import {
15
15
  closeDb,
16
16
  getDb
17
- } from "./chunk-DZ5HJFB4.js";
17
+ } from "./chunk-UM5U6FPH.js";
18
18
  import {
19
19
  config
20
20
  } from "./chunk-K7YUPLES.js";
@@ -261,4 +261,4 @@ async function doctor() {
261
261
  export {
262
262
  doctor
263
263
  };
264
- //# sourceMappingURL=chunk-SUGSQ4YI.js.map
264
+ //# sourceMappingURL=chunk-UEEWG7FD.js.map
@@ -2,6 +2,61 @@ import {
2
2
  config
3
3
  } from "./chunk-K7YUPLES.js";
4
4
 
5
+ // src/db/migrations.ts
6
+ var MIGRATIONS = [
7
+ {
8
+ id: 1,
9
+ name: "add_plugin_hooks_to_user_config",
10
+ sql: "ALTER TABLE user_config_snapshots ADD COLUMN plugin_hooks JSON NOT NULL DEFAULT '[]'"
11
+ }
12
+ ];
13
+ function runMigrations(db, migrations = MIGRATIONS) {
14
+ const trackingExists = db.prepare(
15
+ "SELECT 1 FROM sqlite_master WHERE type='table' AND name='schema_migrations'"
16
+ ).get();
17
+ db.exec(`
18
+ CREATE TABLE IF NOT EXISTS schema_migrations (
19
+ id INTEGER PRIMARY KEY,
20
+ name TEXT NOT NULL,
21
+ applied_at TEXT NOT NULL DEFAULT (datetime('now'))
22
+ )
23
+ `);
24
+ if (!trackingExists) {
25
+ let hasData = false;
26
+ try {
27
+ hasData = !!db.prepare("SELECT 1 FROM sessions LIMIT 1").get();
28
+ } catch {
29
+ }
30
+ if (!hasData) {
31
+ const stamp = db.prepare(
32
+ "INSERT INTO schema_migrations (id, name) VALUES (?, ?)"
33
+ );
34
+ for (const m of migrations) {
35
+ stamp.run(m.id, m.name);
36
+ }
37
+ return;
38
+ }
39
+ }
40
+ const applied = new Set(
41
+ db.prepare("SELECT id FROM schema_migrations").all().map((r) => r.id)
42
+ );
43
+ for (const migration of migrations) {
44
+ if (applied.has(migration.id)) continue;
45
+ const run = db.transaction(() => {
46
+ if (migration.sql) {
47
+ db.exec(migration.sql);
48
+ } else if (migration.up) {
49
+ migration.up(db);
50
+ }
51
+ db.prepare("INSERT INTO schema_migrations (id, name) VALUES (?, ?)").run(
52
+ migration.id,
53
+ migration.name
54
+ );
55
+ });
56
+ run();
57
+ }
58
+ }
59
+
5
60
  // src/db/schema.ts
6
61
  import fs from "fs";
7
62
  import { gunzipSync } from "zlib";
@@ -276,7 +331,8 @@ CREATE TABLE IF NOT EXISTS user_config_snapshots (
276
331
  hooks JSON NOT NULL DEFAULT '[]',
277
332
  commands JSON NOT NULL DEFAULT '[]',
278
333
  rules JSON NOT NULL DEFAULT '[]',
279
- skills JSON NOT NULL DEFAULT '[]'
334
+ skills JSON NOT NULL DEFAULT '[]',
335
+ plugin_hooks JSON NOT NULL DEFAULT '[]'
280
336
  );
281
337
 
282
338
  CREATE TABLE IF NOT EXISTS repo_config_snapshots (
@@ -433,6 +489,7 @@ function getDb() {
433
489
  _db.pragma("busy_timeout = 5000");
434
490
  registerCompressionFunctions(_db);
435
491
  _db.exec(SCHEMA_SQL);
492
+ runMigrations(_db);
436
493
  const currentVersion = _db.pragma("user_version", { simple: true }) ?? 0;
437
494
  if (currentVersion < SCANNER_DATA_VERSION) {
438
495
  _needsResync = true;
@@ -457,6 +514,7 @@ function closeDb() {
457
514
  }
458
515
 
459
516
  export {
517
+ runMigrations,
460
518
  SCHEMA_SQL,
461
519
  SCANNER_DATA_VERSION,
462
520
  getDb,
@@ -464,4 +522,4 @@ export {
464
522
  markResyncComplete,
465
523
  closeDb
466
524
  };
467
- //# sourceMappingURL=chunk-DZ5HJFB4.js.map
525
+ //# sourceMappingURL=chunk-UM5U6FPH.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/db/migrations.ts","../src/db/schema.ts"],"sourcesContent":["/**\n * Schema migration system for panopticon's embedded SQLite database.\n *\n * ## Conventions\n *\n * 1. SCHEMA_SQL in schema.ts is ALWAYS the latest desired schema.\n * It uses CREATE TABLE IF NOT EXISTS, making it idempotent.\n *\n * 2. When adding a column: update the CREATE TABLE in SCHEMA_SQL\n * AND add a migration here with ALTER TABLE ADD COLUMN.\n * Both must exist. SCHEMA_SQL handles fresh DBs; migrations\n * handle existing DBs.\n *\n * 3. When adding a new table: add it to SCHEMA_SQL. No migration\n * needed (CREATE TABLE IF NOT EXISTS handles it).\n *\n * 4. When adding an index: add it to SCHEMA_SQL. No migration\n * needed (CREATE INDEX IF NOT EXISTS handles it).\n *\n * 5. For complex changes (data backfill, column rename via rebuild,\n * virtual table recreation): add an `up` function migration.\n * Update SCHEMA_SQL to reflect the final state.\n *\n * 6. Never reorder or remove migrations. Only append.\n *\n * 7. Migration IDs are sequential integers starting from 1.\n *\n * 8. No down migrations. This is an embedded app — users always\n * upgrade forward. Rolling back means reinstalling.\n */\n\nimport type Database from \"better-sqlite3\";\n\nexport interface Migration {\n id: number;\n name: string;\n /** Simple migrations: single SQL statement. */\n sql?: string;\n /** Complex migrations: function that receives the db handle. */\n up?: (db: Database.Database) => void;\n}\n\n// ---------------------------------------------------------------------------\n// Migration registry — append only, never reorder or remove\n// ---------------------------------------------------------------------------\n\nexport const MIGRATIONS: Migration[] = [\n {\n id: 1,\n name: \"add_plugin_hooks_to_user_config\",\n sql: \"ALTER TABLE user_config_snapshots ADD COLUMN plugin_hooks JSON NOT NULL DEFAULT '[]'\",\n },\n];\n\n// ---------------------------------------------------------------------------\n// Migration runner\n// ---------------------------------------------------------------------------\n\n/**\n * Apply pending schema migrations to the database.\n *\n * On a fresh database (where SCHEMA_SQL just created all tables with all\n * columns), the `schema_migrations` table won't exist yet. In that case\n * we stamp all migrations as applied without executing them — SCHEMA_SQL\n * already reflects the final state.\n *\n * On an existing database, unapplied migrations run sequentially inside\n * transactions.\n */\nexport function runMigrations(\n db: Database.Database,\n migrations: Migration[] = MIGRATIONS,\n): void {\n const trackingExists = db\n .prepare(\n \"SELECT 1 FROM sqlite_master WHERE type='table' AND name='schema_migrations'\",\n )\n .get();\n\n db.exec(`\n CREATE TABLE IF NOT EXISTS schema_migrations (\n id INTEGER PRIMARY KEY,\n name TEXT NOT NULL,\n applied_at TEXT NOT NULL DEFAULT (datetime('now'))\n )\n `);\n\n if (!trackingExists) {\n // No tracking table could mean:\n // a) Truly fresh DB — SCHEMA_SQL just created all tables with all columns\n // b) Pre-migration-system DB — tables existed before migrations were added,\n // and CREATE TABLE IF NOT EXISTS didn't add new columns\n //\n // Distinguish by checking for existing data. A fresh DB has no rows yet.\n let hasData = false;\n try {\n hasData = !!db.prepare(\"SELECT 1 FROM sessions LIMIT 1\").get();\n } catch {\n // sessions table doesn't exist — definitely a fresh/test DB\n }\n\n if (!hasData) {\n // Fresh database: SCHEMA_SQL already created everything.\n // Stamp all migrations as applied without executing them.\n const stamp = db.prepare(\n \"INSERT INTO schema_migrations (id, name) VALUES (?, ?)\",\n );\n for (const m of migrations) {\n stamp.run(m.id, m.name);\n }\n return;\n }\n // Pre-migration-system DB: fall through to run migrations normally\n }\n\n // Existing database: run unapplied migrations sequentially\n const applied = new Set(\n (\n db.prepare(\"SELECT id FROM schema_migrations\").all() as Array<{\n id: number;\n }>\n ).map((r) => r.id),\n );\n\n for (const migration of migrations) {\n if (applied.has(migration.id)) continue;\n const run = db.transaction(() => {\n if (migration.sql) {\n db.exec(migration.sql);\n } else if (migration.up) {\n migration.up(db);\n }\n db.prepare(\"INSERT INTO schema_migrations (id, name) VALUES (?, ?)\").run(\n migration.id,\n migration.name,\n );\n });\n run();\n }\n}\n","import fs from \"node:fs\";\nimport { gunzipSync } from \"node:zlib\";\nimport Database from \"better-sqlite3\";\nimport { config } from \"../config.js\";\nimport { runMigrations } from \"./migrations.js\";\n\nexport { runMigrations } from \"./migrations.js\";\n\nexport const SCHEMA_SQL = `\n\n-- ── OTLP tables ─────────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS otel_logs (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp_ns INTEGER NOT NULL,\n observed_timestamp_ns INTEGER,\n severity_number INTEGER,\n severity_text TEXT,\n body TEXT,\n attributes JSON,\n resource_attributes JSON,\n session_id TEXT,\n prompt_id TEXT,\n trace_id TEXT,\n span_id TEXT,\n sync_id TEXT DEFAULT (hex(randomblob(8)))\n);\n\nCREATE TABLE IF NOT EXISTS otel_metrics (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp_ns INTEGER NOT NULL,\n name TEXT NOT NULL,\n value REAL NOT NULL,\n metric_type TEXT,\n unit TEXT,\n attributes JSON,\n resource_attributes JSON,\n session_id TEXT,\n sync_id TEXT DEFAULT (hex(randomblob(8)))\n);\n\nCREATE TABLE IF NOT EXISTS otel_spans (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n trace_id TEXT NOT NULL,\n span_id TEXT NOT NULL,\n parent_span_id TEXT,\n name TEXT NOT NULL,\n kind INTEGER,\n start_time_ns INTEGER NOT NULL,\n end_time_ns INTEGER NOT NULL,\n status_code INTEGER,\n status_message TEXT,\n attributes JSON,\n resource_attributes JSON,\n session_id TEXT,\n UNIQUE(trace_id, span_id)\n);\n\n-- ── Hook events ─────────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS hook_events (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n session_id TEXT NOT NULL,\n event_type TEXT NOT NULL,\n timestamp_ms INTEGER NOT NULL,\n cwd TEXT,\n repository TEXT,\n tool_name TEXT,\n payload BLOB NOT NULL,\n user_prompt TEXT,\n file_path TEXT,\n command TEXT,\n plan TEXT,\n allowed_prompts TEXT,\n tool_result TEXT,\n target TEXT,\n sync_id TEXT DEFAULT (hex(randomblob(8)))\n);\n\nCREATE VIRTUAL TABLE IF NOT EXISTS hook_events_fts USING fts5(\n payload,\n content='',\n contentless_delete=1,\n tokenize='trigram'\n);\n\n-- ── Sessions ────────────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS sessions (\n session_id TEXT PRIMARY KEY,\n target TEXT,\n started_at_ms INTEGER,\n ended_at_ms INTEGER,\n cwd TEXT,\n first_prompt TEXT,\n permission_mode TEXT,\n agent_version TEXT,\n model TEXT,\n cli_version TEXT,\n scanner_file_path TEXT,\n total_input_tokens INTEGER DEFAULT 0,\n total_output_tokens INTEGER DEFAULT 0,\n total_cache_read_tokens INTEGER DEFAULT 0,\n total_cache_creation_tokens INTEGER DEFAULT 0,\n total_reasoning_tokens INTEGER DEFAULT 0,\n turn_count INTEGER DEFAULT 0,\n otel_input_tokens INTEGER DEFAULT 0,\n otel_output_tokens INTEGER DEFAULT 0,\n otel_cache_read_tokens INTEGER DEFAULT 0,\n otel_cache_creation_tokens INTEGER DEFAULT 0,\n models TEXT,\n has_hooks INTEGER DEFAULT 0,\n has_otel INTEGER DEFAULT 0,\n has_scanner INTEGER DEFAULT 0,\n summary TEXT,\n summary_version INTEGER DEFAULT 0,\n sync_dirty INTEGER DEFAULT 0,\n sync_seq INTEGER DEFAULT 0,\n tool_counts JSON DEFAULT '{}',\n hook_tool_counts JSON DEFAULT '{}',\n event_type_counts JSON DEFAULT '{}',\n hook_event_type_counts JSON DEFAULT '{}',\n project TEXT,\n machine TEXT NOT NULL DEFAULT 'local',\n message_count INTEGER DEFAULT 0,\n user_message_count INTEGER DEFAULT 0,\n parent_session_id TEXT,\n relationship_type TEXT DEFAULT '',\n is_automated INTEGER DEFAULT 0,\n created_at INTEGER\n);\n\nCREATE TABLE IF NOT EXISTS session_repositories (\n session_id TEXT NOT NULL,\n repository TEXT NOT NULL,\n first_seen_ms INTEGER NOT NULL,\n git_user_name TEXT,\n git_user_email TEXT,\n branch TEXT,\n UNIQUE(session_id, repository)\n);\n\nCREATE TABLE IF NOT EXISTS session_cwds (\n session_id TEXT NOT NULL,\n cwd TEXT NOT NULL,\n first_seen_ms INTEGER NOT NULL,\n UNIQUE(session_id, cwd)\n);\n\n-- ── Messages ────────────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS messages (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n session_id TEXT NOT NULL,\n ordinal INTEGER NOT NULL,\n role TEXT NOT NULL,\n content TEXT NOT NULL,\n timestamp_ms INTEGER,\n has_thinking INTEGER NOT NULL DEFAULT 0,\n has_tool_use INTEGER NOT NULL DEFAULT 0,\n content_length INTEGER NOT NULL DEFAULT 0,\n is_system INTEGER NOT NULL DEFAULT 0,\n model TEXT NOT NULL DEFAULT '',\n token_usage TEXT NOT NULL DEFAULT '',\n context_tokens INTEGER NOT NULL DEFAULT 0,\n output_tokens INTEGER NOT NULL DEFAULT 0,\n has_context_tokens INTEGER NOT NULL DEFAULT 0,\n has_output_tokens INTEGER NOT NULL DEFAULT 0,\n uuid TEXT,\n parent_uuid TEXT,\n UNIQUE(session_id, ordinal)\n);\n\nCREATE VIRTUAL TABLE IF NOT EXISTS messages_fts USING fts5(\n content,\n content='',\n contentless_delete=1,\n tokenize='trigram'\n);\n\n-- ── Tool calls ──────────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS tool_calls (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n message_id INTEGER NOT NULL,\n session_id TEXT NOT NULL,\n tool_name TEXT NOT NULL,\n category TEXT NOT NULL,\n tool_use_id TEXT,\n input_json TEXT,\n skill_name TEXT,\n result_content_length INTEGER,\n result_content TEXT,\n duration_ms INTEGER,\n subagent_session_id TEXT,\n sync_id TEXT DEFAULT (hex(randomblob(8)))\n);\n\n-- ── Scanner tables ──────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS scanner_turns (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n session_id TEXT NOT NULL,\n source TEXT NOT NULL,\n turn_index INTEGER NOT NULL,\n timestamp_ms INTEGER NOT NULL,\n model TEXT,\n role TEXT,\n content_preview TEXT,\n input_tokens INTEGER DEFAULT 0,\n output_tokens INTEGER DEFAULT 0,\n cache_read_tokens INTEGER DEFAULT 0,\n cache_creation_tokens INTEGER DEFAULT 0,\n reasoning_tokens INTEGER DEFAULT 0,\n sync_id TEXT DEFAULT (hex(randomblob(8))),\n UNIQUE(session_id, source, turn_index)\n);\n\nCREATE TABLE IF NOT EXISTS scanner_events (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n session_id TEXT NOT NULL,\n source TEXT NOT NULL,\n event_type TEXT NOT NULL,\n timestamp_ms INTEGER NOT NULL,\n tool_name TEXT,\n tool_input TEXT,\n tool_output TEXT,\n content TEXT,\n metadata JSON,\n sync_id TEXT DEFAULT (hex(randomblob(8))),\n UNIQUE(session_id, source, event_type, timestamp_ms, tool_name)\n);\n\nCREATE TABLE IF NOT EXISTS scanner_file_watermarks (\n file_path TEXT PRIMARY KEY,\n byte_offset INTEGER NOT NULL DEFAULT 0,\n last_scanned_ms INTEGER NOT NULL,\n archived_size INTEGER DEFAULT 0\n);\n\n-- ── Session summaries ───────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS session_summary_deltas (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n session_id TEXT NOT NULL,\n delta_index INTEGER NOT NULL,\n created_at_ms INTEGER NOT NULL,\n from_turn INTEGER NOT NULL,\n to_turn INTEGER NOT NULL,\n content TEXT NOT NULL,\n method TEXT NOT NULL DEFAULT 'deterministic',\n UNIQUE(session_id, delta_index)\n);\n\n-- ── Model pricing ───────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS model_pricing (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n model_id TEXT NOT NULL,\n input_per_m REAL NOT NULL,\n output_per_m REAL NOT NULL,\n cache_read_per_m REAL NOT NULL DEFAULT 0,\n cache_write_per_m REAL NOT NULL DEFAULT 0,\n updated_ms INTEGER NOT NULL\n);\n\n-- ── Config snapshots ────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS user_config_snapshots (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n device_name TEXT NOT NULL,\n snapshot_at_ms INTEGER NOT NULL,\n content_hash TEXT NOT NULL,\n permissions JSON NOT NULL DEFAULT '{}',\n enabled_plugins JSON NOT NULL DEFAULT '[]',\n hooks JSON NOT NULL DEFAULT '[]',\n commands JSON NOT NULL DEFAULT '[]',\n rules JSON NOT NULL DEFAULT '[]',\n skills JSON NOT NULL DEFAULT '[]',\n plugin_hooks JSON NOT NULL DEFAULT '[]'\n);\n\nCREATE TABLE IF NOT EXISTS repo_config_snapshots (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n repository TEXT NOT NULL,\n cwd TEXT NOT NULL,\n session_id TEXT,\n snapshot_at_ms INTEGER NOT NULL,\n content_hash TEXT NOT NULL,\n hooks JSON NOT NULL DEFAULT '[]',\n mcp_servers JSON NOT NULL DEFAULT '[]',\n commands JSON NOT NULL DEFAULT '[]',\n agents JSON NOT NULL DEFAULT '[]',\n rules JSON NOT NULL DEFAULT '[]',\n local_hooks JSON NOT NULL DEFAULT '[]',\n local_mcp_servers JSON NOT NULL DEFAULT '[]',\n local_permissions JSON NOT NULL DEFAULT '{}',\n local_is_gitignored INTEGER NOT NULL DEFAULT 1,\n instructions JSON NOT NULL DEFAULT '[]'\n);\n\n-- ── Sync watermarks ─────────────────────────────────────────────────────────\n\nCREATE TABLE IF NOT EXISTS watermarks (\n key TEXT PRIMARY KEY,\n value INTEGER NOT NULL\n);\n\n-- ── Per-session sync state ─────────────────────────────────────────────────\n\nDROP TABLE IF EXISTS pending_session_sync;\n\nCREATE TABLE IF NOT EXISTS target_session_sync (\n session_id TEXT NOT NULL,\n target TEXT NOT NULL,\n confirmed INTEGER DEFAULT 0,\n sync_seq INTEGER DEFAULT 0,\n synced_seq INTEGER DEFAULT 0,\n wm_messages INTEGER DEFAULT 0,\n wm_tool_calls INTEGER DEFAULT 0,\n wm_scanner_turns INTEGER DEFAULT 0,\n wm_scanner_events INTEGER DEFAULT 0,\n wm_hook_events INTEGER DEFAULT 0,\n wm_otel_logs INTEGER DEFAULT 0,\n wm_otel_metrics INTEGER DEFAULT 0,\n wm_otel_spans INTEGER DEFAULT 0,\n PRIMARY KEY (session_id, target)\n);\n\n-- ── Indexes ─────────────────────────────────────────────────────────────────\n\n-- otel_logs\nCREATE INDEX IF NOT EXISTS idx_logs_session ON otel_logs(session_id);\nCREATE INDEX IF NOT EXISTS idx_logs_body ON otel_logs(body);\nCREATE INDEX IF NOT EXISTS idx_logs_ts ON otel_logs(timestamp_ns);\nCREATE INDEX IF NOT EXISTS idx_logs_prompt ON otel_logs(prompt_id);\n\n-- otel_metrics\nCREATE INDEX IF NOT EXISTS idx_metrics_session ON otel_metrics(session_id);\nCREATE INDEX IF NOT EXISTS idx_metrics_name ON otel_metrics(name);\nCREATE INDEX IF NOT EXISTS idx_metrics_ts ON otel_metrics(timestamp_ns);\n\n-- otel_spans\nCREATE INDEX IF NOT EXISTS idx_spans_session ON otel_spans(session_id);\nCREATE INDEX IF NOT EXISTS idx_spans_trace ON otel_spans(trace_id);\nCREATE INDEX IF NOT EXISTS idx_spans_start ON otel_spans(start_time_ns);\n\n-- hook_events\nCREATE INDEX IF NOT EXISTS idx_hooks_session ON hook_events(session_id);\nCREATE INDEX IF NOT EXISTS idx_hooks_type ON hook_events(event_type);\nCREATE INDEX IF NOT EXISTS idx_hooks_tool ON hook_events(tool_name);\nCREATE INDEX IF NOT EXISTS idx_hooks_ts ON hook_events(timestamp_ms);\nCREATE INDEX IF NOT EXISTS idx_hooks_file_path ON hook_events(file_path);\nCREATE INDEX IF NOT EXISTS idx_hooks_target ON hook_events(target);\n\n-- sessions\nCREATE INDEX IF NOT EXISTS idx_sessions_target ON sessions(target);\nCREATE INDEX IF NOT EXISTS idx_sessions_started ON sessions(started_at_ms);\nCREATE INDEX IF NOT EXISTS idx_sessions_sync_seq ON sessions(sync_seq);\nCREATE INDEX IF NOT EXISTS idx_sessions_project ON sessions(project);\nCREATE INDEX IF NOT EXISTS idx_sessions_machine ON sessions(machine);\nCREATE INDEX IF NOT EXISTS idx_sessions_parent ON sessions(parent_session_id)\n WHERE parent_session_id IS NOT NULL;\n\n-- session_repositories\nCREATE INDEX IF NOT EXISTS idx_session_repos_session ON session_repositories(session_id);\nCREATE INDEX IF NOT EXISTS idx_session_repos_repo ON session_repositories(repository);\n\n-- session_cwds\nCREATE INDEX IF NOT EXISTS idx_session_cwds_session ON session_cwds(session_id);\n\n-- messages\nCREATE INDEX IF NOT EXISTS idx_messages_session_ordinal ON messages(session_id, ordinal);\nCREATE INDEX IF NOT EXISTS idx_messages_session_role ON messages(session_id, role);\n\n-- tool_calls\nCREATE INDEX IF NOT EXISTS idx_tool_calls_session ON tool_calls(session_id);\nCREATE INDEX IF NOT EXISTS idx_tool_calls_category ON tool_calls(category);\nCREATE INDEX IF NOT EXISTS idx_tool_calls_skill ON tool_calls(skill_name)\n WHERE skill_name IS NOT NULL;\nCREATE INDEX IF NOT EXISTS idx_tool_calls_subagent ON tool_calls(subagent_session_id)\n WHERE subagent_session_id IS NOT NULL;\n\n-- scanner_turns\nCREATE INDEX IF NOT EXISTS idx_scanner_turns_session ON scanner_turns(session_id);\nCREATE INDEX IF NOT EXISTS idx_scanner_turns_ts ON scanner_turns(timestamp_ms);\n\n-- scanner_events\nCREATE INDEX IF NOT EXISTS idx_scanner_events_session ON scanner_events(session_id);\nCREATE INDEX IF NOT EXISTS idx_scanner_events_type ON scanner_events(event_type);\n\n-- session_summary_deltas\nCREATE INDEX IF NOT EXISTS idx_summary_deltas_session ON session_summary_deltas(session_id);\n\n-- model_pricing\nCREATE INDEX IF NOT EXISTS idx_model_pricing_model ON model_pricing(model_id, updated_ms);\n\n-- user_config_snapshots\nCREATE INDEX IF NOT EXISTS idx_user_config_ts ON user_config_snapshots(snapshot_at_ms);\nCREATE INDEX IF NOT EXISTS idx_user_config_device_hash ON user_config_snapshots(device_name, content_hash);\n\n-- repo_config_snapshots\nCREATE INDEX IF NOT EXISTS idx_repo_config_repo_ts ON repo_config_snapshots(repository, snapshot_at_ms);\nCREATE INDEX IF NOT EXISTS idx_repo_config_session ON repo_config_snapshots(session_id);\nCREATE INDEX IF NOT EXISTS idx_repo_config_repo_hash ON repo_config_snapshots(repository, content_hash);\n\n`;\n\n/**\n * Scanner data version. Increment when parser logic changes in ways that\n * affect stored data (new fields extracted, content formatting changes,\n * fork detection improvements, etc.). On startup, if the DB's user_version\n * is lower than this, a full resync is triggered automatically.\n */\nexport const SCANNER_DATA_VERSION = 1;\n\n// ---------------------------------------------------------------------------\n// Database initialization\n// ---------------------------------------------------------------------------\n\nlet _db: Database.Database | null = null;\nlet _needsResync = false;\n\nfunction registerCompressionFunctions(db: Database.Database): void {\n db.function(\"decompress\", (blob: Buffer | null) =>\n blob ? gunzipSync(blob).toString() : null,\n );\n}\n\nexport function getDb(): Database.Database {\n // If the db file was deleted (e.g. uninstall --purge) while this process\n // still holds a stale connection, drop it so we don't serve old data.\n if (_db && !fs.existsSync(config.dbPath)) {\n try {\n _db.close();\n } catch {}\n _db = null;\n }\n if (_db) return _db;\n\n // Don't auto-create the data directory — callers that need to bootstrap\n // the DB (install, initDb, hook handler) call ensureDataDir() first.\n if (!fs.existsSync(config.dataDir)) {\n throw new Error(\n `Panopticon data directory not found: ${config.dataDir}. Run \"panopticon install\" to set up.`,\n );\n }\n\n _db = new Database(config.dbPath);\n _db.pragma(\"auto_vacuum = INCREMENTAL\");\n _db.pragma(\"journal_mode = WAL\");\n _db.pragma(\"busy_timeout = 5000\");\n\n registerCompressionFunctions(_db);\n _db.exec(SCHEMA_SQL);\n runMigrations(_db);\n\n // Check data version for resync\n const currentVersion = (_db.pragma(\"user_version\", { simple: true }) ??\n 0) as number;\n if (currentVersion < SCANNER_DATA_VERSION) {\n _needsResync = true;\n } else {\n _needsResync = false;\n }\n\n return _db;\n}\n\n/** True when the DB was opened with a stale data version. */\nexport function needsResync(): boolean {\n return _needsResync;\n}\n\n/** Mark resync as complete and stamp the current data version. */\nexport function markResyncComplete(): void {\n _needsResync = false;\n const db = getDb();\n db.pragma(`user_version = ${SCANNER_DATA_VERSION}`);\n}\n\nexport function closeDb(): void {\n if (_db) {\n _db.close();\n _db = null;\n }\n}\n"],"mappings":";;;;;AA8CO,IAAM,aAA0B;AAAA,EACrC;AAAA,IACE,IAAI;AAAA,IACJ,MAAM;AAAA,IACN,KAAK;AAAA,EACP;AACF;AAiBO,SAAS,cACd,IACA,aAA0B,YACpB;AACN,QAAM,iBAAiB,GACpB;AAAA,IACC;AAAA,EACF,EACC,IAAI;AAEP,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAMP;AAED,MAAI,CAAC,gBAAgB;AAOnB,QAAI,UAAU;AACd,QAAI;AACF,gBAAU,CAAC,CAAC,GAAG,QAAQ,gCAAgC,EAAE,IAAI;AAAA,IAC/D,QAAQ;AAAA,IAER;AAEA,QAAI,CAAC,SAAS;AAGZ,YAAM,QAAQ,GAAG;AAAA,QACf;AAAA,MACF;AACA,iBAAW,KAAK,YAAY;AAC1B,cAAM,IAAI,EAAE,IAAI,EAAE,IAAI;AAAA,MACxB;AACA;AAAA,IACF;AAAA,EAEF;AAGA,QAAM,UAAU,IAAI;AAAA,IAEhB,GAAG,QAAQ,kCAAkC,EAAE,IAAI,EAGnD,IAAI,CAAC,MAAM,EAAE,EAAE;AAAA,EACnB;AAEA,aAAW,aAAa,YAAY;AAClC,QAAI,QAAQ,IAAI,UAAU,EAAE,EAAG;AAC/B,UAAM,MAAM,GAAG,YAAY,MAAM;AAC/B,UAAI,UAAU,KAAK;AACjB,WAAG,KAAK,UAAU,GAAG;AAAA,MACvB,WAAW,UAAU,IAAI;AACvB,kBAAU,GAAG,EAAE;AAAA,MACjB;AACA,SAAG,QAAQ,wDAAwD,EAAE;AAAA,QACnE,UAAU;AAAA,QACV,UAAU;AAAA,MACZ;AAAA,IACF,CAAC;AACD,QAAI;AAAA,EACN;AACF;;;AC3IA,OAAO,QAAQ;AACf,SAAS,kBAAkB;AAC3B,OAAO,cAAc;AAMd,IAAM,aAAa;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAsZnB,IAAM,uBAAuB;AAMpC,IAAI,MAAgC;AACpC,IAAI,eAAe;AAEnB,SAAS,6BAA6B,IAA6B;AACjE,KAAG;AAAA,IAAS;AAAA,IAAc,CAAC,SACzB,OAAO,WAAW,IAAI,EAAE,SAAS,IAAI;AAAA,EACvC;AACF;AAEO,SAAS,QAA2B;AAGzC,MAAI,OAAO,CAAC,GAAG,WAAW,OAAO,MAAM,GAAG;AACxC,QAAI;AACF,UAAI,MAAM;AAAA,IACZ,QAAQ;AAAA,IAAC;AACT,UAAM;AAAA,EACR;AACA,MAAI,IAAK,QAAO;AAIhB,MAAI,CAAC,GAAG,WAAW,OAAO,OAAO,GAAG;AAClC,UAAM,IAAI;AAAA,MACR,wCAAwC,OAAO,OAAO;AAAA,IACxD;AAAA,EACF;AAEA,QAAM,IAAI,SAAS,OAAO,MAAM;AAChC,MAAI,OAAO,2BAA2B;AACtC,MAAI,OAAO,oBAAoB;AAC/B,MAAI,OAAO,qBAAqB;AAEhC,+BAA6B,GAAG;AAChC,MAAI,KAAK,UAAU;AACnB,gBAAc,GAAG;AAGjB,QAAM,iBAAkB,IAAI,OAAO,gBAAgB,EAAE,QAAQ,KAAK,CAAC,KACjE;AACF,MAAI,iBAAiB,sBAAsB;AACzC,mBAAe;AAAA,EACjB,OAAO;AACL,mBAAe;AAAA,EACjB;AAEA,SAAO;AACT;AAGO,SAAS,cAAuB;AACrC,SAAO;AACT;AAGO,SAAS,qBAA2B;AACzC,iBAAe;AACf,QAAM,KAAK,MAAM;AACjB,KAAG,OAAO,kBAAkB,oBAAoB,EAAE;AACpD;AAEO,SAAS,UAAgB;AAC9B,MAAI,KAAK;AACP,QAAI,MAAM;AACV,UAAM;AAAA,EACR;AACF;","names":[]}
@@ -1,45 +1,45 @@
1
1
  import {
2
2
  handleOtlpRequest
3
- } from "./chunk-TCKL7E4K.js";
3
+ } from "./chunk-5IPC5XXZ.js";
4
4
  import {
5
5
  handleProxyRequest,
6
6
  processHookEvent,
7
7
  tunnelWebSocket
8
- } from "./chunk-KLXRBD4N.js";
8
+ } from "./chunk-2PMFNSGA.js";
9
9
  import {
10
10
  addTarget,
11
11
  createSyncLoop,
12
12
  listTargets,
13
13
  removeTarget
14
- } from "./chunk-YZBYEULL.js";
14
+ } from "./chunk-75IXBXBO.js";
15
15
  import {
16
16
  createScannerLoop
17
- } from "./chunk-NXH7AONS.js";
17
+ } from "./chunk-5FRYHDPJ.js";
18
18
  import {
19
19
  autoPrune,
20
20
  pruneEstimate,
21
21
  pruneExecute
22
- } from "./chunk-HRCEIYKU.js";
22
+ } from "./chunk-7NBZ4NMJ.js";
23
23
  import {
24
24
  addBreadcrumb,
25
25
  captureException,
26
26
  flushSentry,
27
27
  initSentry,
28
28
  setTag
29
- } from "./chunk-XO5NQRTD.js";
29
+ } from "./chunk-RNICEX6N.js";
30
30
  import {
31
31
  log
32
32
  } from "./chunk-7Q3BJMLG.js";
33
33
  import {
34
34
  refreshPricing
35
- } from "./chunk-3TZAKV3M.js";
35
+ } from "./chunk-OSWBZKK5.js";
36
36
  import {
37
37
  TABLE_SYNC_REGISTRY,
38
38
  readWatermark,
39
39
  resetWatermarks,
40
40
  watermarkKey,
41
41
  writeWatermark
42
- } from "./chunk-SEXU2WYG.js";
42
+ } from "./chunk-P5KAIE3O.js";
43
43
  import {
44
44
  loadUnifiedConfig
45
45
  } from "./chunk-QK5442ZP.js";
@@ -53,10 +53,10 @@ import {
53
53
  rawQuery,
54
54
  search,
55
55
  sessionTimeline
56
- } from "./chunk-LWXF7YRG.js";
56
+ } from "./chunk-UDUBRSIG.js";
57
57
  import {
58
58
  getDb
59
- } from "./chunk-DZ5HJFB4.js";
59
+ } from "./chunk-UM5U6FPH.js";
60
60
  import {
61
61
  config
62
62
  } from "./chunk-K7YUPLES.js";
@@ -132,16 +132,44 @@ var EXEC = {
132
132
  const target = p.target;
133
133
  if (!target) throw new Error("target is required");
134
134
  const db = getDb();
135
+ const WM_COLUMNS = {
136
+ messages: "wm_messages",
137
+ tool_calls: "wm_tool_calls",
138
+ scanner_turns: "wm_scanner_turns",
139
+ scanner_events: "wm_scanner_events",
140
+ hook_events: "wm_hook_events",
141
+ otel_logs: "wm_otel_logs",
142
+ otel_metrics: "wm_otel_metrics",
143
+ otel_spans: "wm_otel_spans"
144
+ };
135
145
  const pending = {};
136
146
  for (const desc of TABLE_SYNC_REGISTRY) {
137
- const key = watermarkKey(desc.table, target);
138
- const wm = readWatermark(key);
139
- const maxId = db.prepare(
140
- `SELECT MAX(${desc.table === "sessions" ? "sync_seq" : "id"}) as m FROM ${desc.table}`
141
- ).get()?.m ?? 0;
142
- const count = Math.max(0, maxId - wm);
143
- if (count > 0) {
144
- pending[desc.table] = { maxId, watermark: wm, pending: count };
147
+ const wmCol = WM_COLUMNS[desc.table];
148
+ if (desc.sessionLinked && wmCol) {
149
+ const total = db.prepare(`SELECT COUNT(*) as c FROM ${desc.table}`).get()?.c ?? 0;
150
+ const pendingCount = db.prepare(
151
+ `SELECT COUNT(*) as c FROM ${desc.table} t
152
+ LEFT JOIN target_session_sync tss
153
+ ON tss.session_id = t.session_id AND tss.target = ?
154
+ WHERE tss.${wmCol} IS NULL OR t.id > tss.${wmCol}`
155
+ ).get(target)?.c ?? 0;
156
+ if (pendingCount > 0) {
157
+ pending[desc.table] = {
158
+ total,
159
+ synced: total - pendingCount,
160
+ pending: pendingCount
161
+ };
162
+ }
163
+ } else {
164
+ const key = watermarkKey(desc.table, target);
165
+ const wm = readWatermark(key);
166
+ const maxId = db.prepare(
167
+ `SELECT MAX(${desc.table === "sessions" ? "sync_seq" : "id"}) as m FROM ${desc.table}`
168
+ ).get()?.m ?? 0;
169
+ const count = Math.max(0, maxId - wm);
170
+ if (count > 0) {
171
+ pending[desc.table] = { total: maxId, synced: wm, pending: count };
172
+ }
145
173
  }
146
174
  }
147
175
  const totalPending = Object.values(pending).reduce(
@@ -444,4 +472,4 @@ export {
444
472
  syncAwarePrune,
445
473
  createUnifiedServer
446
474
  };
447
- //# sourceMappingURL=chunk-VXQ33OYT.js.map
475
+ //# sourceMappingURL=chunk-YV2AN73J.js.map