@indiekitai/pg-dash 0.4.4 → 0.4.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cli.js +50 -31
- package/dist/cli.js.map +1 -1
- package/dist/mcp.js +25 -11
- package/dist/mcp.js.map +1 -1
- package/package.json +1 -1
package/dist/mcp.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/mcp.ts","../src/server/queries/overview.ts","../src/server/queries/tables.ts","../src/server/queries/schema.ts","../src/server/queries/activity.ts","../src/server/advisor.ts","../src/server/queries/slow-queries.ts","../src/server/snapshot.ts","../src/server/env-differ.ts","../src/server/schema-diff.ts","../src/server/schema-tracker.ts","../src/server/query-analyzer.ts","../src/server/migration-checker.ts","../src/server/unused-indexes.ts","../src/server/bloat.ts","../src/server/autovacuum.ts","../src/server/locks.ts","../src/server/config-checker.ts"],"sourcesContent":["// MCP Server for pg-dash — exposes PostgreSQL monitoring tools\n\nimport { McpServer } from \"@modelcontextprotocol/sdk/server/mcp.js\";\nimport { StdioServerTransport } from \"@modelcontextprotocol/sdk/server/stdio.js\";\nimport { Pool } from \"pg\";\nimport { z } from \"zod\";\nimport { getOverview } from \"./server/queries/overview.js\";\nimport { getTables } from \"./server/queries/tables.js\";\nimport { getSchemaTableDetail } from \"./server/queries/schema.js\";\nimport { getActivity } from \"./server/queries/activity.js\";\nimport { getAdvisorReport, isSafeFix } from \"./server/advisor.js\";\nimport { getSlowQueries } from \"./server/queries/slow-queries.js\";\nimport { saveSnapshot, loadSnapshot, diffSnapshots } from \"./server/snapshot.js\";\nimport { diffEnvironments } from \"./server/env-differ.js\";\nimport { analyzeExplainPlan, detectQueryRegressions } from \"./server/query-analyzer.js\";\nimport { analyzeMigration } from \"./server/migration-checker.js\";\nimport { getUnusedIndexes } from \"./server/unused-indexes.js\";\nimport { getBloatReport } from \"./server/bloat.js\";\nimport { getAutovacuumReport } from \"./server/autovacuum.js\";\nimport { getLockReport } from \"./server/locks.js\";\nimport { getConfigReport } from \"./server/config-checker.js\";\nimport Database from \"better-sqlite3\";\nimport path from \"node:path\";\nimport os from \"node:os\";\nimport fs, { readFileSync } from \"node:fs\";\n\nconst pkg = JSON.parse(readFileSync(new URL(\"../package.json\", import.meta.url), \"utf-8\"));\n\nconst connString = process.argv[2] || process.env.PG_DASH_CONNECTION_STRING;\nif (!connString) {\n console.error(\"Usage: pg-dash-mcp <connection-string>\");\n console.error(\" or set PG_DASH_CONNECTION_STRING env var\");\n process.exit(1);\n}\n\nconst pool = new Pool({ connectionString: connString, connectionTimeoutMillis: 10000 });\nconst longQueryThreshold = parseInt(process.env.PG_DASH_LONG_QUERY_THRESHOLD || \"5\", 10);\nconst dataDir = process.env.PG_DASH_DATA_DIR || path.join(os.homedir(), \".pg-dash\");\nfs.mkdirSync(dataDir, { recursive: true });\n\n// Open schema and alerts DBs (read-only for history queries)\nlet schemaDb: Database.Database | null = null;\nlet alertsDb: Database.Database | null = null;\ntry {\n const schemaPath = path.join(dataDir, \"schema.db\");\n if (fs.existsSync(schemaPath)) schemaDb = new Database(schemaPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\ntry {\n const alertsPath = path.join(dataDir, \"alerts.db\");\n if (fs.existsSync(alertsPath)) alertsDb = new Database(alertsPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\n\nlet queryStatsDb: Database.Database | null = null;\ntry {\n const queryStatsPath = path.join(dataDir, \"query-stats.db\");\n if (fs.existsSync(queryStatsPath)) queryStatsDb = new Database(queryStatsPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\n\nconst server = new McpServer({ name: \"pg-dash\", version: pkg.version });\n\nserver.tool(\"pg_dash_overview\", \"Get database overview (version, uptime, size, connections)\", {}, async () => {\n try {\n const data = await getOverview(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_health\", \"Get health advisor report with score, grade, and issues\", {}, async () => {\n try {\n const data = await getAdvisorReport(pool, longQueryThreshold);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_tables\", \"List all tables with sizes and row counts\", {}, async () => {\n try {\n const data = await getTables(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_table_detail\", \"Get detailed info about a specific table\", { table: z.string().describe(\"Table name (e.g. 'users' or 'public.users')\") }, async ({ table }) => {\n try {\n const data = await getSchemaTableDetail(pool, table);\n if (!data) return { content: [{ type: \"text\", text: \"Table not found\" }], isError: true };\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_activity\", \"Get current database activity (active queries, connections)\", {}, async () => {\n try {\n const data = await getActivity(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_schema_changes\", \"Get recent schema changes\", {}, async () => {\n try {\n if (!schemaDb) return { content: [{ type: \"text\", text: \"No schema tracking data available. Run pg-dash server first.\" }] };\n const changes = schemaDb.prepare(\"SELECT * FROM schema_changes ORDER BY timestamp DESC LIMIT 50\").all();\n return { content: [{ type: \"text\", text: JSON.stringify(changes, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_fix\", \"Execute a safe fix (VACUUM, ANALYZE, REINDEX, etc.)\", { sql: z.string().describe(\"SQL to execute (must be a safe operation)\") }, async ({ sql }) => {\n try {\n if (!isSafeFix(sql)) return { content: [{ type: \"text\", text: \"Operation not allowed. Only VACUUM, ANALYZE, REINDEX, CREATE/DROP INDEX CONCURRENTLY, pg_terminate_backend, pg_cancel_backend, and EXPLAIN ANALYZE are permitted.\" }], isError: true };\n const client = await pool.connect();\n try {\n const start = Date.now();\n const result = await client.query(sql);\n return { content: [{ type: \"text\", text: JSON.stringify({ ok: true, duration: Date.now() - start, rowCount: result.rowCount, rows: result.rows || [] }, null, 2) }] };\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_alerts\", \"Get alert history\", {}, async () => {\n try {\n if (!alertsDb) return { content: [{ type: \"text\", text: \"No alerts data available. Run pg-dash server first.\" }] };\n const history = alertsDb.prepare(\"SELECT * FROM alert_history ORDER BY timestamp DESC LIMIT 50\").all();\n return { content: [{ type: \"text\", text: JSON.stringify(history, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\n// --- New tools ---\n\nserver.tool(\"pg_dash_explain\", \"Run EXPLAIN ANALYZE on a SELECT query (read-only, wrapped in BEGIN/ROLLBACK)\", { query: z.string().describe(\"SELECT query to explain\") }, async ({ query }) => {\n try {\n if (!/^\\s*SELECT\\b/i.test(query)) return { content: [{ type: \"text\", text: \"Error: Only SELECT queries are allowed\" }], isError: true };\n const client = await pool.connect();\n try {\n await client.query(\"SET statement_timeout = '30s'\");\n await client.query(\"BEGIN\");\n try {\n const r = await client.query(`EXPLAIN (ANALYZE, BUFFERS, FORMAT JSON) ${query}`);\n await client.query(\"ROLLBACK\");\n await client.query(\"RESET statement_timeout\");\n return { content: [{ type: \"text\", text: JSON.stringify(r.rows[0][\"QUERY PLAN\"], null, 2) }] };\n } catch (err: any) {\n await client.query(\"ROLLBACK\").catch(() => {});\n await client.query(\"RESET statement_timeout\").catch(() => {});\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_batch_fix\", \"Get batch fix SQL for issues (optionally filtered by category)\", { category: z.string().optional().describe(\"Filter by issue type prefix, e.g. 'schema-missing-fk-index'\") }, async ({ category }) => {\n try {\n const report = await getAdvisorReport(pool, longQueryThreshold);\n let fixes = report.batchFixes;\n if (category) fixes = fixes.filter((f) => f.type.startsWith(category));\n if (fixes.length === 0) return { content: [{ type: \"text\", text: \"No batch fixes found\" + (category ? ` for category '${category}'` : \"\") }] };\n const combined = fixes.map((f) => `-- ${f.title}\\n${f.sql}`).join(\"\\n\\n\");\n return { content: [{ type: \"text\", text: combined }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_slow_queries\", \"Get top slow queries from pg_stat_statements\", {\n limit: z.number().optional().default(20).describe(\"Max queries to return (default 20)\"),\n orderBy: z.enum([\"total_time\", \"mean_time\", \"calls\"]).optional().default(\"total_time\").describe(\"Sort order\"),\n}, async ({ limit, orderBy }) => {\n try {\n const all = await getSlowQueries(pool);\n if (all.length === 0) return { content: [{ type: \"text\", text: \"No slow query data available. pg_stat_statements may not be installed.\" }] };\n const sorted = [...all].sort((a, b) => (b as any)[orderBy] - (a as any)[orderBy]);\n return { content: [{ type: \"text\", text: JSON.stringify(sorted.slice(0, limit), null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_table_sizes\", \"Get table sizes with data/index breakdown (top 30)\", {}, async () => {\n try {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT schemaname, relname,\n pg_total_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as total_size,\n pg_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as table_size,\n pg_indexes_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as index_size\n FROM pg_stat_user_tables\n ORDER BY pg_total_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) DESC\n LIMIT 30\n `);\n const tables = r.rows.map((row: any) => ({\n schema: row.schemaname,\n name: row.relname,\n totalSize: parseInt(row.total_size),\n tableSize: parseInt(row.table_size),\n indexSize: parseInt(row.index_size),\n }));\n return { content: [{ type: \"text\", text: JSON.stringify(tables, null, 2) }] };\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_export\", \"Export full health report\", { format: z.enum([\"json\", \"md\"]).optional().default(\"json\").describe(\"Output format: json or md\") }, async ({ format }) => {\n try {\n const [overview, advisor] = await Promise.all([\n getOverview(pool),\n getAdvisorReport(pool, longQueryThreshold),\n ]);\n if (format === \"md\") {\n const lines: string[] = [];\n lines.push(`# pg-dash Health Report`);\n lines.push(`\\nGenerated: ${new Date().toISOString()}\\n`);\n lines.push(`## Overview\\n`);\n lines.push(`- **PostgreSQL**: ${overview.version}`);\n lines.push(`- **Database Size**: ${overview.dbSize}`);\n lines.push(`- **Connections**: ${overview.connections.active} active / ${overview.connections.idle} idle / ${overview.connections.max} max`);\n lines.push(`\\n## Health Score: ${advisor.score}/100 (Grade: ${advisor.grade})\\n`);\n lines.push(`| Category | Grade | Score | Issues |`);\n lines.push(`|----------|-------|-------|--------|`);\n for (const [cat, b] of Object.entries(advisor.breakdown)) {\n lines.push(`| ${cat} | ${b.grade} | ${b.score}/100 | ${b.count} |`);\n }\n if (advisor.issues.length > 0) {\n lines.push(`\\n### Issues (${advisor.issues.length})\\n`);\n for (const issue of advisor.issues) {\n const icon = issue.severity === \"critical\" ? \"🔴\" : issue.severity === \"warning\" ? \"🟡\" : \"🔵\";\n lines.push(`- ${icon} [${issue.severity}] ${issue.title}`);\n }\n }\n if (advisor.batchFixes.length > 0) {\n lines.push(`\\n### 🔧 Batch Fixes\\n`);\n for (const fix of advisor.batchFixes) {\n lines.push(`\\`\\`\\`sql\\n${fix.sql}\\n\\`\\`\\`\\n`);\n }\n }\n return { content: [{ type: \"text\", text: lines.join(\"\\n\") }] };\n }\n return { content: [{ type: \"text\", text: JSON.stringify({ overview, advisor, exportedAt: new Date().toISOString() }, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_diff\", \"Compare current health with last saved snapshot\", {}, async () => {\n try {\n const snapshotPath = path.join(dataDir, \"last-check.json\");\n const prev = loadSnapshot(snapshotPath);\n const current = await getAdvisorReport(pool, longQueryThreshold);\n if (!prev) {\n saveSnapshot(snapshotPath, current);\n return { content: [{ type: \"text\", text: JSON.stringify({ message: \"No previous snapshot found. Current result saved as baseline.\", score: current.score, grade: current.grade, issues: current.issues.length }, null, 2) }] };\n }\n const diff = diffSnapshots(prev.result, current);\n saveSnapshot(snapshotPath, current);\n return { content: [{ type: \"text\", text: JSON.stringify({ ...diff, previousTimestamp: prev.timestamp }, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\n \"pg_dash_check_migration\",\n \"Analyze migration SQL for safety risks (lock tables, missing tables, destructive ops)\",\n {\n sql: z.string().describe(\"Migration SQL content to analyze\"),\n },\n async ({ sql }) => {\n try {\n const result = await analyzeMigration(sql, pool);\n return { content: [{ type: \"text\", text: JSON.stringify(result, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_analyze_query\",\n \"Deep analysis of a SQL query: runs EXPLAIN ANALYZE, detects missing indexes, and provides specific optimization recommendations\",\n {\n sql: z.string().describe(\"SELECT query to analyze\"),\n },\n async ({ sql }) => {\n try {\n if (!/^\\s*SELECT\\b/i.test(sql)) {\n return { content: [{ type: \"text\", text: \"Error: Only SELECT queries are allowed\" }], isError: true };\n }\n const client = await pool.connect();\n try {\n await client.query(\"SET statement_timeout = '30s'\");\n await client.query(\"BEGIN\");\n try {\n const r = await client.query(`EXPLAIN (ANALYZE, BUFFERS, FORMAT JSON) ${sql}`);\n await client.query(\"ROLLBACK\");\n await client.query(\"RESET statement_timeout\");\n\n const plan = r.rows[0][\"QUERY PLAN\"];\n const analysis = await analyzeExplainPlan(plan, pool);\n\n return {\n content: [{\n type: \"text\",\n text: JSON.stringify({ plan, analysis }, null, 2),\n }],\n };\n } catch (err: any) {\n await client.query(\"ROLLBACK\").catch(() => {});\n await client.query(\"RESET statement_timeout\").catch(() => {});\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_query_regressions\",\n \"Detect queries that have gotten significantly slower (>50% degradation) compared to historical baselines\",\n {\n windowHours: z.number().optional().describe(\"Hours to look back (default: 24)\"),\n },\n async ({ windowHours }) => {\n try {\n const regressions = await detectQueryRegressions(pool, queryStatsDb, windowHours ?? 24);\n if (regressions.length === 0) {\n return { content: [{ type: \"text\", text: \"No query regressions detected in the specified window.\" }] };\n }\n return { content: [{ type: \"text\", text: JSON.stringify(regressions, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_compare_env\",\n \"Compare schema and health between two PostgreSQL environments. Detects missing tables, columns, indexes.\",\n {\n targetUrl: z.string().describe(\"Target database connection string to compare against\"),\n includeHealth: z.boolean().optional().describe(\"Also compare health scores and issues\"),\n },\n async ({ targetUrl, includeHealth }) => {\n try {\n // source uses the existing pool's connection string; target gets its own pool (created inside diffEnvironments)\n // We pass connString so the source pool is created fresh with a 10s timeout; existing pool is unaffected\n const result = await diffEnvironments(connString, targetUrl, { includeHealth: includeHealth ?? false });\n return { content: [{ type: \"text\", text: JSON.stringify(result, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\"pg_dash_unused_indexes\", \"Find unused indexes that waste space and slow down writes\", {}, async () => {\n try {\n const report = await getUnusedIndexes(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_bloat\", \"Detect table bloat (dead tuples) that slow down queries\", {}, async () => {\n try {\n const report = await getBloatReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_autovacuum\", \"Check autovacuum health — which tables are stale or never vacuumed\", {}, async () => {\n try {\n const report = await getAutovacuumReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_locks\", \"Show active lock waits and long-running queries blocking the database\", {}, async () => {\n try {\n const report = await getLockReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_config_check\", \"Audit PostgreSQL configuration settings and get tuning recommendations\", {}, async () => {\n try {\n const report = await getConfigReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nconst transport = new StdioServerTransport();\nawait server.connect(transport);\n","import type { Pool } from \"pg\";\n\nexport async function getOverview(pool: Pool) {\n const client = await pool.connect();\n try {\n const version = await client.query(\"SHOW server_version\");\n const uptime = await client.query(\n \"SELECT to_char(now() - pg_postmaster_start_time(), 'DD \\\"d\\\" HH24 \\\"h\\\" MI \\\"m\\\"') AS uptime\"\n );\n const dbSize = await client.query(\n \"SELECT pg_size_pretty(pg_database_size(current_database())) AS size\"\n );\n const dbCount = await client.query(\n \"SELECT count(*)::int AS count FROM pg_database WHERE NOT datistemplate\"\n );\n const connections = await client.query(`\n SELECT\n (SELECT count(*)::int FROM pg_stat_activity WHERE state = 'active') AS active,\n (SELECT count(*)::int FROM pg_stat_activity WHERE state = 'idle') AS idle,\n (SELECT setting::int FROM pg_settings WHERE name = 'max_connections') AS max\n `);\n\n return {\n version: version.rows[0].server_version,\n uptime: uptime.rows[0].uptime,\n dbSize: dbSize.rows[0].size,\n databaseCount: dbCount.rows[0].count,\n connections: connections.rows[0],\n };\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport async function getTables(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n schemaname AS schema,\n relname AS name,\n pg_size_pretty(pg_total_relation_size(relid)) AS total_size,\n pg_total_relation_size(relid) AS size_bytes,\n n_live_tup AS rows,\n n_dead_tup AS dead_tuples,\n CASE WHEN n_live_tup > 0 \n THEN round(n_dead_tup::numeric / n_live_tup * 100, 1) \n ELSE 0 END AS dead_pct\n FROM pg_stat_user_tables\n ORDER BY pg_total_relation_size(relid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport async function getSchemaTables(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n c.relname AS name,\n n.nspname AS schema,\n pg_size_pretty(pg_total_relation_size(c.oid)) AS total_size,\n pg_total_relation_size(c.oid) AS total_size_bytes,\n pg_size_pretty(pg_relation_size(c.oid)) AS table_size,\n pg_size_pretty(pg_total_relation_size(c.oid) - pg_relation_size(c.oid)) AS index_size,\n s.n_live_tup AS row_count,\n obj_description(c.oid) AS description\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n LEFT JOIN pg_stat_user_tables s ON s.relid = c.oid\n WHERE c.relkind = 'r' AND n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY pg_total_relation_size(c.oid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaTableDetail(pool: Pool, tableName: string) {\n const client = await pool.connect();\n try {\n // Parse schema.table or default to public\n const parts = tableName.split(\".\");\n const schema = parts.length > 1 ? parts[0] : \"public\";\n const name = parts.length > 1 ? parts[1] : parts[0];\n\n // Table info\n const tableInfo = await client.query(`\n SELECT\n c.relname AS name, n.nspname AS schema,\n pg_size_pretty(pg_total_relation_size(c.oid)) AS total_size,\n pg_size_pretty(pg_relation_size(c.oid)) AS table_size,\n pg_size_pretty(pg_total_relation_size(c.oid) - pg_relation_size(c.oid)) AS index_size,\n pg_size_pretty(pg_relation_size(c.reltoastrelid)) AS toast_size,\n s.n_live_tup AS row_count, s.n_dead_tup AS dead_tuples,\n s.last_vacuum, s.last_autovacuum, s.last_analyze, s.last_autoanalyze,\n s.seq_scan, s.idx_scan\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n LEFT JOIN pg_stat_user_tables s ON s.relid = c.oid\n WHERE c.relname = $1 AND n.nspname = $2 AND c.relkind = 'r'\n `, [name, schema]);\n\n if (tableInfo.rows.length === 0) return null;\n\n // Columns\n const columns = await client.query(`\n SELECT\n a.attname AS name,\n pg_catalog.format_type(a.atttypid, a.atttypmod) AS type,\n NOT a.attnotnull AS nullable,\n pg_get_expr(d.adbin, d.adrelid) AS default_value,\n col_description(a.attrelid, a.attnum) AS description\n FROM pg_attribute a\n LEFT JOIN pg_attrdef d ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = (SELECT c.oid FROM pg_class c JOIN pg_namespace n ON c.relnamespace = n.oid WHERE c.relname = $1 AND n.nspname = $2)\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n `, [name, schema]);\n\n // Indexes\n const indexes = await client.query(`\n SELECT\n i.relname AS name,\n am.amname AS type,\n pg_size_pretty(pg_relation_size(i.oid)) AS size,\n pg_get_indexdef(idx.indexrelid) AS definition,\n idx.indisunique AS is_unique,\n idx.indisprimary AS is_primary,\n s.idx_scan, s.idx_tup_read, s.idx_tup_fetch\n FROM pg_index idx\n JOIN pg_class i ON idx.indexrelid = i.oid\n JOIN pg_class t ON idx.indrelid = t.oid\n JOIN pg_namespace n ON t.relnamespace = n.oid\n JOIN pg_am am ON i.relam = am.oid\n LEFT JOIN pg_stat_user_indexes s ON s.indexrelid = i.oid\n WHERE t.relname = $1 AND n.nspname = $2\n ORDER BY i.relname\n `, [name, schema]);\n\n // Constraints\n const constraints = await client.query(`\n SELECT\n conname AS name,\n CASE contype WHEN 'p' THEN 'PRIMARY KEY' WHEN 'f' THEN 'FOREIGN KEY'\n WHEN 'u' THEN 'UNIQUE' WHEN 'c' THEN 'CHECK' WHEN 'x' THEN 'EXCLUDE' END AS type,\n pg_get_constraintdef(oid) AS definition\n FROM pg_constraint\n WHERE conrelid = (SELECT c.oid FROM pg_class c JOIN pg_namespace n ON c.relnamespace = n.oid WHERE c.relname = $1 AND n.nspname = $2)\n ORDER BY\n CASE contype WHEN 'p' THEN 1 WHEN 'u' THEN 2 WHEN 'f' THEN 3 WHEN 'c' THEN 4 ELSE 5 END\n `, [name, schema]);\n\n // Foreign keys (outgoing)\n const foreignKeys = await client.query(`\n SELECT\n conname AS name,\n a.attname AS column_name,\n confrelid::regclass::text AS referenced_table,\n af.attname AS referenced_column\n FROM pg_constraint c\n JOIN pg_attribute a ON a.attrelid = c.conrelid AND a.attnum = ANY(c.conkey)\n JOIN pg_attribute af ON af.attrelid = c.confrelid AND af.attnum = ANY(c.confkey)\n WHERE c.contype = 'f'\n AND c.conrelid = (SELECT cl.oid FROM pg_class cl JOIN pg_namespace n ON cl.relnamespace = n.oid WHERE cl.relname = $1 AND n.nspname = $2)\n `, [name, schema]);\n\n // Sample data (first 10 rows)\n let sampleData: any[] = [];\n try {\n const sample = await client.query(\n `SELECT * FROM ${client.escapeIdentifier(schema)}.${client.escapeIdentifier(name)} LIMIT 10`\n );\n sampleData = sample.rows;\n } catch (err) { console.error(\"[schema] Error:\", (err as Error).message); }\n\n return {\n ...tableInfo.rows[0],\n columns: columns.rows,\n indexes: indexes.rows,\n constraints: constraints.rows,\n foreignKeys: foreignKeys.rows,\n sampleData,\n };\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaIndexes(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n n.nspname AS schema,\n t.relname AS table_name,\n i.relname AS name,\n am.amname AS type,\n pg_size_pretty(pg_relation_size(i.oid)) AS size,\n pg_relation_size(i.oid) AS size_bytes,\n pg_get_indexdef(idx.indexrelid) AS definition,\n idx.indisunique AS is_unique,\n idx.indisprimary AS is_primary,\n s.idx_scan, s.idx_tup_read, s.idx_tup_fetch\n FROM pg_index idx\n JOIN pg_class i ON idx.indexrelid = i.oid\n JOIN pg_class t ON idx.indrelid = t.oid\n JOIN pg_namespace n ON t.relnamespace = n.oid\n JOIN pg_am am ON i.relam = am.oid\n LEFT JOIN pg_stat_user_indexes s ON s.indexrelid = i.oid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY pg_relation_size(i.oid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaFunctions(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n n.nspname AS schema,\n p.proname AS name,\n pg_get_function_result(p.oid) AS return_type,\n pg_get_function_arguments(p.oid) AS arguments,\n l.lanname AS language,\n p.prosrc AS source,\n CASE p.prokind WHEN 'f' THEN 'function' WHEN 'p' THEN 'procedure' WHEN 'a' THEN 'aggregate' WHEN 'w' THEN 'window' END AS kind\n FROM pg_proc p\n JOIN pg_namespace n ON p.pronamespace = n.oid\n JOIN pg_language l ON p.prolang = l.oid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY n.nspname, p.proname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaExtensions(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT extname AS name, extversion AS installed_version,\n n.nspname AS schema, obj_description(e.oid) AS description\n FROM pg_extension e\n JOIN pg_namespace n ON e.extnamespace = n.oid\n ORDER BY extname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaEnums(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n t.typname AS name,\n n.nspname AS schema,\n array_agg(e.enumlabel ORDER BY e.enumsortorder) AS values\n FROM pg_type t\n JOIN pg_namespace n ON t.typnamespace = n.oid\n JOIN pg_enum e ON t.oid = e.enumtypid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n GROUP BY t.typname, n.nspname\n ORDER BY t.typname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport interface Activity {\n pid: number;\n query: string;\n state: string;\n wait_event: string | null;\n wait_event_type: string | null;\n duration: string | null;\n client_addr: string | null;\n application_name: string;\n backend_start: string;\n}\n\nexport async function getActivity(pool: Pool): Promise<Activity[]> {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n pid,\n COALESCE(query, '') AS query,\n COALESCE(state, 'unknown') AS state,\n wait_event,\n wait_event_type,\n CASE WHEN state = 'active' THEN (now() - query_start)::text\n WHEN state = 'idle in transaction' THEN (now() - state_change)::text\n ELSE NULL END AS duration,\n client_addr::text,\n COALESCE(application_name, '') AS application_name,\n backend_start::text\n FROM pg_stat_activity\n WHERE pid != pg_backend_pid()\n AND state IS NOT NULL\n ORDER BY\n CASE state\n WHEN 'active' THEN 1\n WHEN 'idle in transaction' THEN 2\n ELSE 3\n END,\n query_start ASC NULLS LAST\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport interface AdvisorIssue {\n id: string;\n severity: \"critical\" | \"warning\" | \"info\";\n category: \"performance\" | \"maintenance\" | \"schema\" | \"security\";\n title: string;\n description: string;\n fix: string;\n impact: string;\n effort: \"quick\" | \"moderate\" | \"involved\";\n}\n\nexport interface BatchFix {\n type: string;\n title: string;\n count: number;\n sql: string;\n}\n\nexport interface AdvisorResult {\n score: number;\n grade: string;\n issues: AdvisorIssue[];\n breakdown: Record<string, { score: number; grade: string; count: number }>;\n skipped: string[];\n ignoredCount: number;\n batchFixes: BatchFix[];\n}\n\nconst SEVERITY_WEIGHT = { critical: 15, warning: 5, info: 1 } as const;\nconst MAX_DEDUCTION = { critical: 60, warning: 30, info: 10 } as const;\n\nexport function computeAdvisorScore(issues: AdvisorIssue[]): number {\n let score = 100;\n const deductions = { critical: 0, warning: 0, info: 0 };\n const counts = { critical: 0, warning: 0, info: 0 };\n for (const issue of issues) {\n counts[issue.severity]++;\n const n = counts[issue.severity];\n const weight = SEVERITY_WEIGHT[issue.severity];\n // Diminishing penalty: full for first 3, half for 4-10, quarter for 11+\n let penalty: number;\n if (n <= 3) penalty = weight;\n else if (n <= 10) penalty = weight * 0.5;\n else penalty = weight * 0.25;\n deductions[issue.severity] += penalty;\n }\n // Cap deductions per severity\n for (const sev of [\"critical\", \"warning\", \"info\"] as const) {\n score -= Math.min(deductions[sev], MAX_DEDUCTION[sev]);\n }\n return Math.max(0, Math.min(100, Math.round(score)));\n}\n\nexport function gradeFromScore(score: number): string {\n if (score >= 90) return \"A\";\n if (score >= 80) return \"B\";\n if (score >= 70) return \"C\";\n if (score >= 50) return \"D\";\n return \"F\";\n}\n\nfunction computeBreakdown(issues: AdvisorIssue[]): Record<string, { score: number; grade: string; count: number }> {\n const categories = [\"performance\", \"maintenance\", \"schema\", \"security\"] as const;\n const result: Record<string, { score: number; grade: string; count: number }> = {};\n for (const cat of categories) {\n const catIssues = issues.filter((i) => i.category === cat);\n const score = computeAdvisorScore(catIssues);\n result[cat] = { score, grade: gradeFromScore(score), count: catIssues.length };\n }\n return result;\n}\n\nexport async function getAdvisorReport(pool: Pool, longQueryThreshold = 5): Promise<AdvisorResult> {\n const client = await pool.connect();\n const issues: AdvisorIssue[] = [];\n const skipped: string[] = [];\n\n try {\n // Detect PG version for compatibility\n const versionResult = await client.query(\"SHOW server_version_num\");\n const pgVersion = parseInt(versionResult.rows[0].server_version_num);\n\n // ── Performance Advisors ───────────────────────────────────────\n\n // Missing indexes (high seq scans on large tables)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, seq_scan, seq_tup_read, n_live_tup,\n pg_size_pretty(pg_total_relation_size(relid)) AS size\n FROM pg_stat_user_tables\n WHERE n_live_tup > 10000 AND seq_scan > 100\n ORDER BY seq_tup_read DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-seq-scan-${row.schemaname}-${row.relname}`,\n severity: row.seq_scan > 1000 ? \"warning\" : \"info\",\n category: \"performance\",\n title: `High sequential scans on ${row.relname}`,\n description: `Table ${row.schemaname}.${row.relname} (${row.n_live_tup} rows, ${row.size}) has ${row.seq_scan} sequential scans reading ${Number(row.seq_tup_read).toLocaleString()} tuples. Consider adding indexes on frequently filtered columns.`,\n fix: `-- Identify commonly filtered columns and add indexes:\\n-- EXPLAIN ANALYZE SELECT * FROM ${row.schemaname}.${row.relname} WHERE <your_condition>;\\nCREATE INDEX CONCURRENTLY idx_${row.relname}_<column> ON ${row.schemaname}.${row.relname} (<column>);`,\n impact: \"Queries will continue to do full table scans, degrading performance as the table grows.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking seq scans:\", (err as Error).message); skipped.push(\"seq scans: \" + (err as Error).message);\n }\n\n // Bloated indexes (index size > 3x table size)\n try {\n const r = await client.query(`\n SELECT\n schemaname, relname, indexrelname,\n pg_relation_size(indexrelid) AS idx_size,\n pg_relation_size(relid) AS tbl_size,\n pg_size_pretty(pg_relation_size(indexrelid)) AS idx_size_pretty,\n pg_size_pretty(pg_relation_size(relid)) AS tbl_size_pretty\n FROM pg_stat_user_indexes\n WHERE pg_relation_size(indexrelid) > 1048576\n AND pg_relation_size(indexrelid) > pg_relation_size(relid) * 3\n ORDER BY pg_relation_size(indexrelid) DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-bloated-idx-${row.indexrelname}`,\n severity: \"warning\",\n category: \"performance\",\n title: `Bloated index ${row.indexrelname}`,\n description: `Index ${row.indexrelname} on ${row.relname} is ${row.idx_size_pretty} but the table is only ${row.tbl_size_pretty}. The index may need rebuilding.`,\n fix: `REINDEX INDEX CONCURRENTLY ${row.schemaname}.${row.indexrelname};`,\n impact: \"Bloated indexes waste disk space and slow down queries that use them.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking bloated indexes:\", (err as Error).message); skipped.push(\"bloated indexes: \" + (err as Error).message);\n }\n\n // Table bloat (dead tuples > 10%)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, n_dead_tup, n_live_tup,\n CASE WHEN n_live_tup > 0 THEN round(n_dead_tup::numeric / n_live_tup * 100, 1) ELSE 0 END AS dead_pct,\n pg_size_pretty(pg_total_relation_size(relid)) AS size\n FROM pg_stat_user_tables\n WHERE n_live_tup > 1000 AND n_dead_tup::float / GREATEST(n_live_tup, 1) > 0.1\n ORDER BY n_dead_tup DESC LIMIT 10\n `);\n for (const row of r.rows) {\n const pct = parseFloat(row.dead_pct);\n issues.push({\n id: `perf-bloat-${row.schemaname}-${row.relname}`,\n severity: pct > 30 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Table bloat on ${row.relname} (${row.dead_pct}% dead)`,\n description: `${row.schemaname}.${row.relname} has ${Number(row.n_dead_tup).toLocaleString()} dead tuples (${row.dead_pct}% of ${Number(row.n_live_tup).toLocaleString()} live rows). Size: ${row.size}.`,\n fix: `VACUUM FULL ${row.schemaname}.${row.relname};`,\n impact: \"Dead tuples waste storage and degrade scan performance.\",\n effort: pct > 30 ? \"moderate\" : \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking table bloat:\", (err as Error).message); skipped.push(\"table bloat: \" + (err as Error).message);\n }\n\n // Cache efficiency per table\n try {\n const r = await client.query(`\n SELECT schemaname, relname,\n heap_blks_hit, heap_blks_read,\n CASE WHEN (heap_blks_hit + heap_blks_read) = 0 THEN 1\n ELSE heap_blks_hit::float / (heap_blks_hit + heap_blks_read) END AS ratio\n FROM pg_statio_user_tables\n WHERE (heap_blks_hit + heap_blks_read) > 100\n ORDER BY ratio ASC LIMIT 5\n `);\n for (const row of r.rows) {\n const ratio = parseFloat(row.ratio);\n if (ratio < 0.9) {\n issues.push({\n id: `perf-cache-${row.schemaname}-${row.relname}`,\n severity: ratio < 0.5 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Poor cache hit ratio on ${row.relname}`,\n description: `Table ${row.schemaname}.${row.relname} has a cache hit ratio of ${(ratio * 100).toFixed(1)}%. Most reads are going to disk.`,\n fix: `-- Consider increasing shared_buffers or reducing working set:\\nSHOW shared_buffers;`,\n impact: \"Disk reads are orders of magnitude slower than memory reads.\",\n effort: \"involved\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking cache efficiency:\", (err as Error).message); skipped.push(\"cache efficiency: \" + (err as Error).message);\n }\n\n // Slow queries from pg_stat_statements\n try {\n const extCheck = await client.query(\"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\");\n if (extCheck.rows.length > 0) {\n const r = await client.query(`\n SELECT query, calls, mean_exec_time, total_exec_time,\n round(mean_exec_time::numeric, 2) AS mean_ms,\n round(total_exec_time::numeric / 1000, 2) AS total_sec\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%' AND query NOT LIKE '%pg_catalog%'\n AND mean_exec_time > 100\n ORDER BY mean_exec_time DESC LIMIT 5\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-slow-${row.query.slice(0, 30).replace(/\\W/g, \"_\")}`,\n severity: parseFloat(row.mean_ms) > 1000 ? \"warning\" : \"info\",\n category: \"performance\",\n title: `Slow query (avg ${row.mean_ms}ms)`,\n description: `Query averaging ${row.mean_ms}ms over ${row.calls} calls (total: ${row.total_sec}s): ${row.query.slice(0, 200)}`,\n fix: `EXPLAIN ANALYZE ${row.query.slice(0, 500)};`,\n impact: \"Slow queries degrade overall database responsiveness.\",\n effort: \"moderate\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking slow queries:\", (err as Error).message); skipped.push(\"slow queries: \" + (err as Error).message);\n }\n\n // ── Maintenance Advisors ───────────────────────────────────────\n\n // VACUUM overdue\n try {\n const r = await client.query(`\n SELECT schemaname, relname, last_vacuum, last_autovacuum, n_dead_tup\n FROM pg_stat_user_tables\n WHERE n_live_tup > 100\n AND (last_vacuum IS NULL AND last_autovacuum IS NULL\n OR GREATEST(last_vacuum, last_autovacuum) < now() - interval '7 days')\n ORDER BY n_dead_tup DESC LIMIT 15\n `);\n for (const row of r.rows) {\n const never = !row.last_vacuum && !row.last_autovacuum;\n issues.push({\n id: `maint-vacuum-${row.schemaname}-${row.relname}`,\n severity: never ? \"warning\" : \"info\",\n category: \"maintenance\",\n title: `VACUUM ${never ? \"never run\" : \"overdue\"} on ${row.relname}`,\n description: `${row.schemaname}.${row.relname} ${never ? \"has never been vacuumed\" : \"was last vacuumed over 7 days ago\"}. Dead tuples: ${Number(row.n_dead_tup).toLocaleString()}.`,\n fix: `VACUUM ANALYZE ${row.schemaname}.${row.relname};`,\n impact: \"Dead tuples accumulate, increasing table size and degrading query performance.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking vacuum overdue:\", (err as Error).message); skipped.push(\"vacuum overdue: \" + (err as Error).message);\n }\n\n // ANALYZE overdue\n try {\n const r = await client.query(`\n SELECT schemaname, relname\n FROM pg_stat_user_tables\n WHERE n_live_tup > 100\n AND last_analyze IS NULL AND last_autoanalyze IS NULL\n AND NOT EXISTS (\n SELECT 1 FROM pg_stat_user_tables t2\n WHERE t2.relname = pg_stat_user_tables.relname\n AND (t2.last_vacuum IS NULL AND t2.last_autovacuum IS NULL)\n )\n LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `maint-analyze-${row.schemaname}-${row.relname}`,\n severity: \"info\",\n category: \"maintenance\",\n title: `ANALYZE never run on ${row.relname}`,\n description: `${row.schemaname}.${row.relname} has never been analyzed. The query planner may choose suboptimal plans.`,\n fix: `ANALYZE ${row.schemaname}.${row.relname};`,\n impact: \"Without statistics, the query planner makes poor estimates leading to slow queries.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking analyze overdue:\", (err as Error).message); skipped.push(\"analyze overdue: \" + (err as Error).message);\n }\n\n // Transaction ID wraparound risk\n try {\n const r = await client.query(`\n SELECT datname, age(datfrozenxid) AS xid_age\n FROM pg_database\n WHERE datname = current_database()\n `);\n for (const row of r.rows) {\n const age = parseInt(row.xid_age);\n if (age > 1_000_000_000) {\n issues.push({\n id: `maint-xid-wraparound`,\n severity: \"critical\",\n category: \"maintenance\",\n title: `Transaction ID wraparound risk`,\n description: `Database ${row.datname} has datfrozenxid age of ${age.toLocaleString()}. Wraparound occurs at ~2 billion.`,\n fix: `VACUUM FREEZE;`,\n impact: \"If wraparound occurs, PostgreSQL will shut down to prevent data loss.\",\n effort: \"involved\",\n });\n } else if (age > 500_000_000) {\n issues.push({\n id: `maint-xid-warning`,\n severity: \"warning\",\n category: \"maintenance\",\n title: `Transaction ID age is high`,\n description: `Database ${row.datname} has datfrozenxid age of ${age.toLocaleString()}.`,\n fix: `VACUUM FREEZE;`,\n impact: \"Approaching transaction ID wraparound threshold.\",\n effort: \"moderate\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking xid wraparound:\", (err as Error).message); skipped.push(\"xid wraparound: \" + (err as Error).message);\n }\n\n // Idle connections > 10 min\n try {\n const r = await client.query(`\n SELECT pid, state, now() - state_change AS idle_duration,\n client_addr::text, application_name,\n extract(epoch from now() - state_change)::int AS idle_seconds\n FROM pg_stat_activity\n WHERE state IN ('idle', 'idle in transaction')\n AND now() - state_change > $1 * interval '1 minute'\n AND pid != pg_backend_pid()\n `, [longQueryThreshold]);\n for (const row of r.rows) {\n const isIdleTx = row.state === \"idle in transaction\";\n issues.push({\n id: `maint-idle-${row.pid}`,\n severity: isIdleTx ? \"warning\" : \"info\",\n category: \"maintenance\",\n title: `${isIdleTx ? \"Idle in transaction\" : \"Idle connection\"} (PID ${row.pid})`,\n description: `PID ${row.pid} from ${row.client_addr || \"local\"} (${row.application_name || \"unknown\"}) has been ${row.state} for ${Math.round(row.idle_seconds / 60)} minutes.`,\n fix: `SELECT pg_terminate_backend(${row.pid});`,\n impact: isIdleTx ? \"Idle-in-transaction connections hold locks and prevent VACUUM.\" : \"Idle connections consume connection slots.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking idle connections:\", (err as Error).message); skipped.push(\"idle connections: \" + (err as Error).message);\n }\n\n // ── Schema Advisors ────────────────────────────────────────────\n\n // Missing primary keys\n try {\n const r = await client.query(`\n SELECT c.relname AS table_name, n.nspname AS schema\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n WHERE c.relkind = 'r' AND n.nspname = 'public'\n AND NOT EXISTS (\n SELECT 1 FROM pg_constraint con WHERE con.conrelid = c.oid AND con.contype = 'p'\n )\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-no-pk-${row.schema}-${row.table_name}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Missing primary key on ${row.table_name}`,\n description: `Table ${row.schema}.${row.table_name} has no primary key. This can cause replication issues and makes row identification unreliable.`,\n fix: `ALTER TABLE ${row.schema}.${row.table_name} ADD PRIMARY KEY (<column>);`,\n impact: \"No primary key means no unique row identity, problematic for replication and ORMs.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking missing primary keys:\", (err as Error).message); skipped.push(\"missing primary keys: \" + (err as Error).message);\n }\n\n // Unused indexes (idx_scan = 0, size > 1MB)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, indexrelname, idx_scan,\n pg_size_pretty(pg_relation_size(indexrelid)) AS idx_size,\n pg_relation_size(indexrelid) AS idx_bytes\n FROM pg_stat_user_indexes\n WHERE idx_scan = 0\n AND indexrelname NOT LIKE '%_pkey'\n AND pg_relation_size(indexrelid) > 1048576\n ORDER BY pg_relation_size(indexrelid) DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-unused-idx-${row.indexrelname}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Unused index ${row.indexrelname} (${row.idx_size})`,\n description: `Index ${row.indexrelname} on ${row.relname} has never been used (0 scans) and takes ${row.idx_size}.`,\n fix: `DROP INDEX CONCURRENTLY ${row.schemaname}.${row.indexrelname};`,\n impact: \"Unused indexes waste disk space and slow down writes.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking unused indexes:\", (err as Error).message); skipped.push(\"unused indexes: \" + (err as Error).message);\n }\n\n // Duplicate indexes\n try {\n const r = await client.query(`\n SELECT array_agg(idx.indexrelid::regclass::text) AS indexes,\n idx.indrelid::regclass::text AS table_name,\n pg_size_pretty(sum(pg_relation_size(idx.indexrelid))) AS total_size\n FROM pg_index idx\n GROUP BY idx.indrelid, idx.indkey\n HAVING count(*) > 1\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-dup-idx-${row.table_name}-${row.indexes[0]}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Duplicate indexes on ${row.table_name}`,\n description: `These indexes cover the same columns on ${row.table_name}: ${row.indexes.join(\", \")}. Total wasted space: ${row.total_size}.`,\n fix: `-- Keep one, drop the rest:\\nDROP INDEX CONCURRENTLY ${row.indexes.slice(1).join(\";\\nDROP INDEX CONCURRENTLY \")};`,\n impact: \"Duplicate indexes double the write overhead and waste disk space.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking duplicate indexes:\", (err as Error).message); skipped.push(\"duplicate indexes: \" + (err as Error).message);\n }\n\n // Missing foreign key indexes\n try {\n const r = await client.query(`\n SELECT\n conrelid::regclass::text AS table_name,\n a.attname AS column_name,\n confrelid::regclass::text AS referenced_table\n FROM pg_constraint c\n JOIN pg_attribute a ON a.attrelid = c.conrelid AND a.attnum = ANY(c.conkey)\n WHERE c.contype = 'f'\n AND NOT EXISTS (\n SELECT 1 FROM pg_index i\n WHERE i.indrelid = c.conrelid\n AND a.attnum = ANY(i.indkey)\n )\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-fk-no-idx-${row.table_name}-${row.column_name}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Missing index on FK column ${row.table_name}.${row.column_name}`,\n description: `Foreign key column ${row.column_name} on ${row.table_name} (references ${row.referenced_table}) has no index. This causes slow JOINs and cascading deletes.`,\n fix: `CREATE INDEX CONCURRENTLY idx_${row.table_name.replace(/\\./g, \"_\")}_${row.column_name} ON ${row.table_name} (${row.column_name});`,\n impact: \"JOINs and cascading deletes on this FK will require full table scans.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking missing FK indexes:\", (err as Error).message); skipped.push(\"missing FK indexes: \" + (err as Error).message);\n }\n\n // ── Infrastructure Advisors ──────────────────────────────────────\n\n // Lock detection\n try {\n const r = await client.query(`\n SELECT blocked_locks.pid AS blocked_pid,\n blocking_locks.pid AS blocking_pid,\n blocked_activity.query AS blocked_query\n FROM pg_catalog.pg_locks blocked_locks\n JOIN pg_catalog.pg_locks blocking_locks ON blocking_locks.locktype = blocked_locks.locktype\n AND blocking_locks.database IS NOT DISTINCT FROM blocked_locks.database\n AND blocking_locks.relation IS NOT DISTINCT FROM blocked_locks.relation\n AND blocking_locks.page IS NOT DISTINCT FROM blocked_locks.page\n AND blocking_locks.tuple IS NOT DISTINCT FROM blocked_locks.tuple\n AND blocking_locks.virtualxid IS NOT DISTINCT FROM blocked_locks.virtualxid\n AND blocking_locks.transactionid IS NOT DISTINCT FROM blocked_locks.transactionid\n AND blocking_locks.classid IS NOT DISTINCT FROM blocked_locks.classid\n AND blocking_locks.objid IS NOT DISTINCT FROM blocked_locks.objid\n AND blocking_locks.objsubid IS NOT DISTINCT FROM blocked_locks.objsubid\n AND blocking_locks.pid != blocked_locks.pid\n JOIN pg_catalog.pg_stat_activity blocked_activity ON blocked_activity.pid = blocked_locks.pid\n WHERE NOT blocked_locks.granted\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-lock-blocked-${row.blocked_pid}`,\n severity: \"warning\",\n category: \"performance\",\n title: `Blocked query (PID ${row.blocked_pid} blocked by PID ${row.blocking_pid})`,\n description: `PID ${row.blocked_pid} is waiting for a lock held by PID ${row.blocking_pid}. Query: ${(row.blocked_query || \"\").slice(0, 200)}`,\n fix: `SELECT pg_cancel_backend(${row.blocking_pid});`,\n impact: \"Blocked queries cause cascading delays and potential timeouts.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking locks:\", (err as Error).message); skipped.push(\"locks: \" + (err as Error).message);\n }\n\n // WAL/replication lag\n try {\n const r = await client.query(`\n SELECT CASE WHEN pg_is_in_recovery()\n THEN pg_wal_lsn_diff(pg_last_wal_receive_lsn(), pg_last_wal_replay_lsn())\n ELSE 0 END AS lag_bytes\n `);\n const lagBytes = parseInt(r.rows[0]?.lag_bytes ?? \"0\");\n if (lagBytes > 1048576) { // > 1MB\n issues.push({\n id: `perf-replication-lag`,\n severity: lagBytes > 104857600 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Replication lag: ${(lagBytes / 1048576).toFixed(1)} MB`,\n description: `WAL replay is lagging by ${(lagBytes / 1048576).toFixed(1)} MB. This indicates the replica is falling behind.`,\n fix: `-- Check replication status:\\nSELECT * FROM pg_stat_replication;`,\n impact: \"High replication lag means the replica has stale data and failover may lose transactions.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking replication lag:\", (err as Error).message); skipped.push(\"replication lag: \" + (err as Error).message);\n }\n\n // Checkpoint frequency\n try {\n const checkpointView = pgVersion >= 170000 ? 'pg_stat_checkpointer' : 'pg_stat_bgwriter';\n const r = await client.query(`\n SELECT checkpoints_req, checkpoints_timed,\n CASE WHEN (checkpoints_req + checkpoints_timed) = 0 THEN 0\n ELSE round(checkpoints_req::numeric / (checkpoints_req + checkpoints_timed) * 100, 1) END AS req_pct\n FROM ${checkpointView}\n `);\n const reqPct = parseFloat(r.rows[0]?.req_pct ?? \"0\");\n if (reqPct > 50) {\n issues.push({\n id: `maint-checkpoint-frequency`,\n severity: reqPct > 80 ? \"warning\" : \"info\",\n category: \"maintenance\",\n title: `${reqPct}% of checkpoints are requested (not timed)`,\n description: `${r.rows[0]?.checkpoints_req} requested vs ${r.rows[0]?.checkpoints_timed} timed checkpoints. High requested checkpoints indicate checkpoint_completion_target or max_wal_size may need tuning.`,\n fix: `-- Increase max_wal_size:\\nALTER SYSTEM SET max_wal_size = '2GB';\\nSELECT pg_reload_conf();`,\n impact: \"Frequent requested checkpoints cause I/O spikes and degrade performance.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking checkpoint frequency:\", (err as Error).message); skipped.push(\"checkpoint frequency: \" + (err as Error).message);\n }\n\n // AutoVACUUM config check\n try {\n const r = await client.query(`SELECT setting FROM pg_settings WHERE name = 'autovacuum'`);\n if (r.rows[0]?.setting === \"off\") {\n issues.push({\n id: `maint-autovacuum-disabled`,\n severity: \"critical\",\n category: \"maintenance\",\n title: `Autovacuum is disabled`,\n description: `Autovacuum is turned off. Dead tuples will accumulate and transaction ID wraparound becomes a risk.`,\n fix: `ALTER SYSTEM SET autovacuum = on;\\nSELECT pg_reload_conf();`,\n impact: \"Without autovacuum, tables bloat indefinitely and risk transaction ID wraparound shutdown.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking autovacuum:\", (err as Error).message); skipped.push(\"autovacuum: \" + (err as Error).message);\n }\n\n // shared_buffers / work_mem check\n try {\n const sbRes = await client.query(`SELECT setting, unit FROM pg_settings WHERE name = 'shared_buffers'`);\n const memRes = await client.query(`\n SELECT (SELECT setting::bigint FROM pg_settings WHERE name = 'shared_buffers') *\n (SELECT setting::bigint FROM pg_settings WHERE name = 'block_size') AS shared_bytes\n `);\n const sharedBytes = parseInt(memRes.rows[0]?.shared_bytes ?? \"0\");\n // Get total RAM from OS via a simple query (pg doesn't expose this directly, but we can estimate)\n // We'll compare against a reasonable minimum: if shared_buffers < 128MB, warn\n if (sharedBytes > 0 && sharedBytes < 128 * 1024 * 1024) {\n issues.push({\n id: `perf-shared-buffers-low`,\n severity: \"warning\",\n category: \"performance\",\n title: `shared_buffers is only ${(sharedBytes / 1048576).toFixed(0)} MB`,\n description: `shared_buffers is set to ${sbRes.rows[0]?.setting}${sbRes.rows[0]?.unit || \"\"}. Recommended: ~25% of system RAM, typically at least 256MB for production.`,\n fix: `ALTER SYSTEM SET shared_buffers = '256MB';\\n-- Requires restart`,\n impact: \"Low shared_buffers means more disk I/O and poor cache hit ratios.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking shared_buffers:\", (err as Error).message); skipped.push(\"shared_buffers: \" + (err as Error).message);\n }\n\n try {\n const r = await client.query(`SELECT setting, unit FROM pg_settings WHERE name = 'work_mem'`);\n const workMemKB = parseInt(r.rows[0]?.setting ?? \"0\");\n if (workMemKB > 0 && workMemKB < 4096) { // < 4MB\n issues.push({\n id: `perf-work-mem-low`,\n severity: \"info\",\n category: \"performance\",\n title: `work_mem is only ${workMemKB < 1024 ? workMemKB + \"kB\" : (workMemKB / 1024).toFixed(0) + \"MB\"}`,\n description: `work_mem is ${r.rows[0]?.setting}${r.rows[0]?.unit || \"\"}. Low work_mem causes sorts and hash operations to spill to disk.`,\n fix: `ALTER SYSTEM SET work_mem = '16MB';\\nSELECT pg_reload_conf();`,\n impact: \"Operations that exceed work_mem use temporary disk files, which is much slower.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking work_mem:\", (err as Error).message); skipped.push(\"work_mem: \" + (err as Error).message);\n }\n\n // ── Security Advisors ──────────────────────────────────────────\n\n // Superuser connections from non-localhost\n try {\n const r = await client.query(`\n SELECT pid, usename, client_addr::text\n FROM pg_stat_activity\n WHERE usename IN (SELECT rolname FROM pg_roles WHERE rolsuper)\n AND client_addr IS NOT NULL\n AND client_addr::text NOT IN ('127.0.0.1', '::1')\n AND pid != pg_backend_pid()\n `);\n for (const row of r.rows) {\n issues.push({\n id: `sec-superuser-remote-${row.pid}`,\n severity: \"critical\",\n category: \"security\",\n title: `Superuser ${row.usename} connected from ${row.client_addr}`,\n description: `Superuser ${row.usename} has an active connection from non-localhost address ${row.client_addr}. This is a security risk.`,\n fix: `-- Restrict superuser access in pg_hba.conf to localhost only.\\n-- Then: SELECT pg_reload_conf();`,\n impact: \"Remote superuser access is a significant security vulnerability.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking superuser connections:\", (err as Error).message); skipped.push(\"superuser connections: \" + (err as Error).message);\n }\n\n // SSL disabled\n try {\n const r = await client.query(`SELECT setting FROM pg_settings WHERE name = 'ssl'`);\n if (r.rows[0]?.setting === \"off\") {\n issues.push({\n id: `sec-ssl-off`,\n severity: \"warning\",\n category: \"security\",\n title: `SSL is disabled`,\n description: `SSL is turned off. Database connections are not encrypted.`,\n fix: `-- Enable SSL in postgresql.conf:\\n-- ssl = on\\n-- ssl_cert_file = 'server.crt'\\n-- ssl_key_file = 'server.key'\\nSELECT pg_reload_conf();`,\n impact: \"Database traffic can be intercepted and read in transit.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking SSL check:\", (err as Error).message); skipped.push(\"SSL check: \" + (err as Error).message);\n }\n\n // Password authentication check (PG 15+)\n try {\n const r = await client.query(`\n SELECT type, database, user_name, auth_method\n FROM pg_hba_file_rules\n WHERE auth_method = 'trust' AND type != 'local'\n LIMIT 5\n `);\n for (const row of r.rows) {\n issues.push({\n id: `sec-trust-auth-${row.database}-${row.user_name}`,\n severity: \"critical\",\n category: \"security\",\n title: `Trust authentication for ${row.user_name}@${row.database}`,\n description: `HBA rule allows trust (no password) authentication for ${row.type} connections to ${row.database} as ${row.user_name}.`,\n fix: `-- Change auth_method from 'trust' to 'scram-sha-256' in pg_hba.conf\\n-- Then: SELECT pg_reload_conf();`,\n impact: \"Anyone can connect without a password.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking trust auth:\", (err as Error).message); skipped.push(\"trust auth: \" + (err as Error).message);\n } // pg_hba_file_rules not available pre-PG15\n\n // Filter out ignored issues\n const ignoredIds = getIgnoredIssues();\n const ignoredSet = new Set(ignoredIds);\n const activeIssues = issues.filter(i => !ignoredSet.has(i.id));\n const ignoredCount = issues.length - activeIssues.length;\n\n // Generate batch fixes for groups of same-type issues\n const batchFixes: BatchFix[] = [];\n const groups = new Map<string, AdvisorIssue[]>();\n for (const issue of activeIssues) {\n // Group by id prefix (everything before the last dash-separated segment with variable data)\n const prefix = issue.id.replace(/-[^-]+$/, \"\");\n if (!groups.has(prefix)) groups.set(prefix, []);\n groups.get(prefix)!.push(issue);\n }\n const BATCH_TITLES: Record<string, string> = {\n \"schema-fk-no-idx\": \"Create all missing FK indexes\",\n \"schema-unused-idx\": \"Drop all unused indexes\",\n \"schema-no-pk\": \"Fix all tables missing primary keys\",\n \"maint-vacuum\": \"VACUUM all overdue tables\",\n \"maint-analyze\": \"ANALYZE all tables missing statistics\",\n \"perf-bloated-idx\": \"REINDEX all bloated indexes\",\n \"perf-bloat\": \"VACUUM FULL all bloated tables\",\n };\n for (const [prefix, group] of groups) {\n if (group.length <= 1) continue;\n const title = BATCH_TITLES[prefix] || `Fix all ${group.length} ${prefix} issues`;\n const sql = group.map(i => i.fix.split(\"\\n\").filter(l => !l.trim().startsWith(\"--\")).join(\"\\n\").trim()).filter(Boolean).join(\";\\n\") + \";\";\n batchFixes.push({ type: prefix, title: `${title} (${group.length})`, count: group.length, sql });\n }\n\n const score = computeAdvisorScore(activeIssues);\n return {\n score,\n grade: gradeFromScore(score),\n issues: activeIssues,\n breakdown: computeBreakdown(activeIssues),\n skipped,\n ignoredCount,\n batchFixes,\n };\n } finally {\n client.release();\n }\n}\n\n// ── Ignored Issues Management ──────────────────────────────────\n\nimport Database from \"better-sqlite3\";\nimport path from \"node:path\";\nimport os from \"node:os\";\nimport fs from \"node:fs\";\n\nlet _ignoredDb: ReturnType<typeof Database> | null = null;\n\nfunction getIgnoredDb(): ReturnType<typeof Database> {\n if (_ignoredDb) return _ignoredDb;\n const dataDir = process.env.PG_DASH_DATA_DIR || path.join(os.homedir(), \".pg-dash\");\n fs.mkdirSync(dataDir, { recursive: true });\n const dbPath = path.join(dataDir, \"alerts.db\");\n _ignoredDb = new Database(dbPath);\n _ignoredDb.pragma(\"journal_mode = WAL\");\n _ignoredDb.exec(\"CREATE TABLE IF NOT EXISTS ignored_issues (issue_id TEXT PRIMARY KEY, ignored_at INTEGER)\");\n return _ignoredDb;\n}\n\nexport function getIgnoredIssues(): string[] {\n try {\n const db = getIgnoredDb();\n return db.prepare(\"SELECT issue_id FROM ignored_issues\").all().map((r: any) => r.issue_id);\n } catch {\n return [];\n }\n}\n\nexport function ignoreIssue(issueId: string): void {\n const db = getIgnoredDb();\n db.prepare(\"INSERT OR REPLACE INTO ignored_issues (issue_id, ignored_at) VALUES (?, ?)\").run(issueId, Date.now());\n}\n\nexport function unignoreIssue(issueId: string): void {\n const db = getIgnoredDb();\n db.prepare(\"DELETE FROM ignored_issues WHERE issue_id = ?\").run(issueId);\n}\n\n// Allowed SQL operations for the fix endpoint\n\nexport function isSafeFix(sql: string): boolean {\n const trimmed = sql.trim();\n if (!trimmed) return false;\n\n // Reject multi-statement SQL (split on semicolons, ignore trailing)\n const statements = trimmed.replace(/;\\s*$/, \"\").split(\";\").map(s => s.trim()).filter(Boolean);\n if (statements.length !== 1) return false;\n\n const upper = statements[0].toUpperCase();\n\n // EXPLAIN ANALYZE — only allow if followed by SELECT\n if (upper.startsWith(\"EXPLAIN ANALYZE\")) {\n const afterExplain = upper.replace(/^EXPLAIN\\s+ANALYZE\\s+/, \"\").trimStart();\n return afterExplain.startsWith(\"SELECT\");\n }\n\n // Simple prefix allowlist for single statements\n const ALLOWED_PREFIXES = [\n \"VACUUM\",\n \"ANALYZE\",\n \"REINDEX\",\n \"CREATE INDEX CONCURRENTLY\",\n \"DROP INDEX CONCURRENTLY\",\n \"SELECT PG_TERMINATE_BACKEND(\",\n \"SELECT PG_CANCEL_BACKEND(\",\n ];\n\n return ALLOWED_PREFIXES.some((p) => upper.startsWith(p));\n}\n","import type { Pool } from \"pg\";\n\nexport interface SlowQuery {\n queryid: string;\n query: string;\n calls: number;\n total_time: number;\n mean_time: number;\n rows: number;\n total_time_pretty: string;\n mean_time_pretty: string;\n}\n\nexport async function getSlowQueries(pool: Pool): Promise<SlowQuery[]> {\n const client = await pool.connect();\n try {\n // Check if pg_stat_statements is available\n const extCheck = await client.query(\n \"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\"\n );\n if (extCheck.rows.length === 0) {\n return [];\n }\n\n const r = await client.query(`\n SELECT\n queryid::text,\n query,\n calls::int,\n total_exec_time AS total_time,\n mean_exec_time AS mean_time,\n rows::int,\n round(total_exec_time::numeric / 1000, 2)::text || 's' AS total_time_pretty,\n round(mean_exec_time::numeric, 2)::text || 'ms' AS mean_time_pretty\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%'\n AND query NOT LIKE '%pg_catalog%'\n ORDER BY total_exec_time DESC\n LIMIT 50\n `);\n return r.rows;\n } catch {\n // pg_stat_statements might not be accessible\n return [];\n } finally {\n client.release();\n }\n}\n","import fs from \"node:fs\";\nimport path from \"node:path\";\nimport type { AdvisorResult, AdvisorIssue } from \"./advisor.js\";\n\nexport interface Snapshot {\n timestamp: string;\n result: AdvisorResult;\n}\n\nexport interface SnapshotDiff {\n scoreDelta: number;\n previousScore: number;\n currentScore: number;\n previousGrade: string;\n currentGrade: string;\n newIssues: AdvisorIssue[];\n resolvedIssues: AdvisorIssue[];\n unchanged: AdvisorIssue[];\n}\n\n/**\n * Normalize a dynamic issue ID for stable comparison.\n * Strips trailing -<number> suffixes so IDs like `maint-idle-12345`\n * (where 12345 is a PID that changes every run) don't produce false noise.\n */\nfunction normalizeIssueId(id: string): string {\n return id.replace(/-\\d+$/, \"\");\n}\n\n/**\n * Save a health-check snapshot to a specific file path.\n * The parent directory is created automatically.\n *\n * @param snapshotPath Full path to the JSON file (e.g. ~/.pg-dash/last-check.json)\n */\nexport function saveSnapshot(snapshotPath: string, result: AdvisorResult): void {\n fs.mkdirSync(path.dirname(snapshotPath), { recursive: true });\n const snapshot: Snapshot = { timestamp: new Date().toISOString(), result };\n fs.writeFileSync(snapshotPath, JSON.stringify(snapshot, null, 2));\n}\n\n/**\n * Load a previously saved snapshot from a specific file path.\n * Returns null if the file doesn't exist or cannot be parsed.\n *\n * @param snapshotPath Full path to the JSON file\n */\nexport function loadSnapshot(snapshotPath: string): Snapshot | null {\n if (!fs.existsSync(snapshotPath)) return null;\n try {\n return JSON.parse(fs.readFileSync(snapshotPath, \"utf-8\"));\n } catch {\n return null;\n }\n}\n\nexport function diffSnapshots(prev: AdvisorResult, current: AdvisorResult): SnapshotDiff {\n // Use normalized IDs for comparison to avoid noise from dynamic suffixes\n // (e.g. maint-idle-12345 where 12345 is a PID that changes every run).\n const prevNormIds = new Set(prev.issues.map((i) => normalizeIssueId(i.id)));\n const currNormIds = new Set(current.issues.map((i) => normalizeIssueId(i.id)));\n\n const newIssues = current.issues.filter((i) => !prevNormIds.has(normalizeIssueId(i.id)));\n const resolvedIssues = prev.issues.filter((i) => !currNormIds.has(normalizeIssueId(i.id)));\n const unchanged = current.issues.filter((i) => prevNormIds.has(normalizeIssueId(i.id)));\n\n return {\n scoreDelta: current.score - prev.score,\n previousScore: prev.score,\n currentScore: current.score,\n previousGrade: prev.grade,\n currentGrade: current.grade,\n newIssues,\n resolvedIssues,\n unchanged,\n };\n}\n","import { Pool } from \"pg\";\nimport { getAdvisorReport } from \"./advisor.js\";\nimport { buildLiveSnapshot } from \"./schema-tracker.js\";\nimport { diffSchemaSnapshots } from \"./schema-diff.js\";\n\nexport interface ColumnInfo {\n name: string;\n type: string;\n nullable: boolean;\n default?: string;\n}\n\nexport interface ColumnTypeDiff {\n column: string;\n sourceType: string;\n targetType: string;\n}\n\nexport interface ColumnNullableDiff {\n column: string;\n sourceNullable: boolean;\n targetNullable: boolean;\n}\n\nexport interface ColumnDefaultDiff {\n column: string;\n sourceDefault: string | null;\n targetDefault: string | null;\n}\n\nexport interface ColumnDiff {\n table: string;\n missingColumns: ColumnInfo[]; // source has, target doesn't\n extraColumns: ColumnInfo[]; // target has, source doesn't\n typeDiffs: ColumnTypeDiff[]; // same name, different type\n nullableDiffs: ColumnNullableDiff[]; // same name, different nullable\n defaultDiffs: ColumnDefaultDiff[]; // same name, different default\n}\n\nexport interface IndexDefDiff {\n name: string;\n sourceDef: string;\n targetDef: string;\n}\n\nexport interface IndexDiff {\n table: string;\n missingIndexes: string[]; // source has, target doesn't\n extraIndexes: string[]; // target has, source doesn't\n modifiedIndexes: IndexDefDiff[]; // same name, different definition\n}\n\nexport interface ConstraintDiff {\n table: string | null;\n type: \"missing\" | \"extra\" | \"modified\";\n name: string;\n detail: string;\n}\n\nexport interface EnumDiff {\n type: \"missing\" | \"extra\" | \"modified\";\n name: string;\n detail: string;\n}\n\nexport interface SchemaDiff {\n missingTables: string[];\n extraTables: string[];\n columnDiffs: ColumnDiff[];\n indexDiffs: IndexDiff[];\n constraintDiffs: ConstraintDiff[];\n enumDiffs: EnumDiff[];\n}\n\nexport interface HealthDiff {\n source: { score: number; grade: string; url: string };\n target: { score: number; grade: string; url: string };\n sourceOnlyIssues: string[];\n targetOnlyIssues: string[];\n}\n\nexport interface EnvDiffResult {\n schema: SchemaDiff;\n health?: HealthDiff;\n checkedAt: string;\n summary: {\n schemaDrifts: number;\n identical: boolean;\n };\n}\n\n// ----- internal types -----\n\ninterface RawColumn {\n table_name: string;\n column_name: string;\n data_type: string;\n is_nullable: string;\n column_default: string | null;\n}\n\ninterface RawIndex {\n tablename: string;\n indexname: string;\n indexdef: string;\n}\n\n// ----- query helpers -----\n\nasync function fetchTables(pool: Pool): Promise<string[]> {\n const res = await pool.query<{ table_name: string }>(`\n SELECT table_name\n FROM information_schema.tables\n WHERE table_schema = 'public' AND table_type = 'BASE TABLE'\n ORDER BY table_name\n `);\n return res.rows.map((r) => r.table_name);\n}\n\nasync function fetchColumns(pool: Pool): Promise<RawColumn[]> {\n const res = await pool.query<RawColumn>(`\n SELECT table_name, column_name, data_type, is_nullable, column_default\n FROM information_schema.columns\n WHERE table_schema = 'public'\n ORDER BY table_name, ordinal_position\n `);\n return res.rows;\n}\n\nasync function fetchIndexes(pool: Pool): Promise<RawIndex[]> {\n const res = await pool.query<RawIndex>(`\n SELECT tablename, indexname, indexdef\n FROM pg_indexes\n WHERE schemaname = 'public' AND indexname NOT LIKE '%_pkey'\n ORDER BY tablename, indexname\n `);\n return res.rows;\n}\n\n// ----- diff logic -----\n\nfunction diffTables(sourceTables: string[], targetTables: string[]): { missingTables: string[]; extraTables: string[] } {\n const sourceSet = new Set(sourceTables);\n const targetSet = new Set(targetTables);\n return {\n missingTables: sourceTables.filter((t) => !targetSet.has(t)),\n extraTables: targetTables.filter((t) => !sourceSet.has(t)),\n };\n}\n\nfunction groupColumnsByTable(columns: RawColumn[]): Map<string, Map<string, ColumnInfo>> {\n const map = new Map<string, Map<string, ColumnInfo>>();\n for (const col of columns) {\n if (!map.has(col.table_name)) map.set(col.table_name, new Map());\n const info: ColumnInfo = {\n name: col.column_name,\n type: col.data_type,\n nullable: col.is_nullable === \"YES\",\n };\n if (col.column_default !== null && col.column_default !== undefined) {\n info.default = col.column_default;\n }\n map.get(col.table_name)!.set(col.column_name, info);\n }\n return map;\n}\n\nfunction diffColumns(\n sourceCols: RawColumn[],\n targetCols: RawColumn[],\n commonTables: string[]\n): ColumnDiff[] {\n const sourceByTable = groupColumnsByTable(sourceCols);\n const targetByTable = groupColumnsByTable(targetCols);\n const diffs: ColumnDiff[] = [];\n\n for (const table of commonTables) {\n const srcMap = sourceByTable.get(table) ?? new Map<string, ColumnInfo>();\n const tgtMap = targetByTable.get(table) ?? new Map<string, ColumnInfo>();\n\n const missingColumns: ColumnInfo[] = [];\n const extraColumns: ColumnInfo[] = [];\n const typeDiffs: ColumnTypeDiff[] = [];\n const nullableDiffs: ColumnNullableDiff[] = [];\n const defaultDiffs: ColumnDefaultDiff[] = [];\n\n for (const [colName, srcInfo] of srcMap) {\n if (!tgtMap.has(colName)) {\n missingColumns.push(srcInfo);\n } else {\n const tgtInfo = tgtMap.get(colName)!;\n if (srcInfo.type !== tgtInfo.type) {\n typeDiffs.push({ column: colName, sourceType: srcInfo.type, targetType: tgtInfo.type });\n }\n if (srcInfo.nullable !== tgtInfo.nullable) {\n nullableDiffs.push({ column: colName, sourceNullable: srcInfo.nullable, targetNullable: tgtInfo.nullable });\n }\n if ((srcInfo.default ?? null) !== (tgtInfo.default ?? null)) {\n defaultDiffs.push({ column: colName, sourceDefault: srcInfo.default ?? null, targetDefault: tgtInfo.default ?? null });\n }\n }\n }\n\n for (const [colName, tgtInfo] of tgtMap) {\n if (!srcMap.has(colName)) {\n extraColumns.push(tgtInfo);\n }\n }\n\n if (missingColumns.length > 0 || extraColumns.length > 0 || typeDiffs.length > 0 ||\n nullableDiffs.length > 0 || defaultDiffs.length > 0) {\n diffs.push({ table, missingColumns, extraColumns, typeDiffs, nullableDiffs, defaultDiffs });\n }\n }\n\n return diffs;\n}\n\nfunction groupIndexesByTable(indexes: RawIndex[]): Map<string, Map<string, string>> {\n const map = new Map<string, Map<string, string>>();\n for (const idx of indexes) {\n if (!map.has(idx.tablename)) map.set(idx.tablename, new Map());\n map.get(idx.tablename)!.set(idx.indexname, idx.indexdef);\n }\n return map;\n}\n\nfunction diffIndexes(\n sourceIdxs: RawIndex[],\n targetIdxs: RawIndex[],\n commonTables: string[]\n): IndexDiff[] {\n const srcByTable = groupIndexesByTable(sourceIdxs);\n const tgtByTable = groupIndexesByTable(targetIdxs);\n const diffs: IndexDiff[] = [];\n\n // All tables that have any indexes in source or target\n const allTables = new Set([\n ...sourceIdxs.map((i) => i.tablename),\n ...targetIdxs.map((i) => i.tablename),\n ]);\n\n for (const table of allTables) {\n // Only diff tables that exist in both environments (common tables + tables not in either missingTables/extraTables)\n if (!commonTables.includes(table)) continue;\n\n const srcMap = srcByTable.get(table) ?? new Map<string, string>();\n const tgtMap = tgtByTable.get(table) ?? new Map<string, string>();\n\n const missingIndexes = [...srcMap.keys()].filter((i) => !tgtMap.has(i));\n const extraIndexes = [...tgtMap.keys()].filter((i) => !srcMap.has(i));\n const modifiedIndexes: IndexDefDiff[] = [];\n\n for (const [name, srcDef] of srcMap) {\n if (tgtMap.has(name)) {\n const tgtDef = tgtMap.get(name)!;\n if (srcDef !== tgtDef) {\n modifiedIndexes.push({ name, sourceDef: srcDef, targetDef: tgtDef });\n }\n }\n }\n\n if (missingIndexes.length > 0 || extraIndexes.length > 0 || modifiedIndexes.length > 0) {\n diffs.push({ table, missingIndexes, extraIndexes, modifiedIndexes });\n }\n }\n\n return diffs;\n}\n\nfunction countSchemaDrifts(schema: SchemaDiff): number {\n let n = schema.missingTables.length + schema.extraTables.length;\n for (const cd of schema.columnDiffs) {\n n += cd.missingColumns.length + cd.extraColumns.length + cd.typeDiffs.length +\n cd.nullableDiffs.length + cd.defaultDiffs.length;\n }\n for (const id of schema.indexDiffs) {\n n += id.missingIndexes.length + id.extraIndexes.length + id.modifiedIndexes.length;\n }\n n += (schema.constraintDiffs ?? []).length;\n n += (schema.enumDiffs ?? []).length;\n return n;\n}\n\n// ----- public API -----\n\nexport async function diffEnvironments(\n sourceConn: string,\n targetConn: string,\n options?: { includeHealth?: boolean }\n): Promise<EnvDiffResult> {\n const sourcePool = new Pool({ connectionString: sourceConn, connectionTimeoutMillis: 10000 });\n const targetPool = new Pool({ connectionString: targetConn, connectionTimeoutMillis: 10000 });\n\n try {\n // Run all schema queries in parallel (basic + deep snapshots for constraints/enums)\n const [\n sourceTables,\n targetTables,\n sourceCols,\n targetCols,\n sourceIdxs,\n targetIdxs,\n sourceSnap,\n targetSnap,\n ] = await Promise.all([\n fetchTables(sourcePool),\n fetchTables(targetPool),\n fetchColumns(sourcePool),\n fetchColumns(targetPool),\n fetchIndexes(sourcePool),\n fetchIndexes(targetPool),\n buildLiveSnapshot(sourcePool).catch(() => null),\n buildLiveSnapshot(targetPool).catch(() => null),\n ]);\n\n const { missingTables, extraTables } = diffTables(sourceTables, targetTables);\n const targetSet = new Set(targetTables);\n const commonTables = sourceTables.filter((t) => targetSet.has(t));\n\n const columnDiffs = diffColumns(sourceCols, targetCols, commonTables);\n const indexDiffs = diffIndexes(sourceIdxs, targetIdxs, commonTables);\n\n // Constraint + enum diffs via snapshot comparison\n const constraintDiffs: ConstraintDiff[] = [];\n const enumDiffs: EnumDiff[] = [];\n\n if (sourceSnap && targetSnap) {\n // diffSnapshots treats source as \"old\" and target as \"new\":\n // added = target has, source doesn't (extra in target)\n // removed = source has, target doesn't (missing in target)\n const snapChanges = diffSchemaSnapshots(sourceSnap, targetSnap);\n\n for (const c of snapChanges) {\n if (c.object_type === \"constraint\") {\n constraintDiffs.push({\n table: c.table_name ?? null,\n type: c.change_type === \"added\" ? \"extra\" : c.change_type === \"removed\" ? \"missing\" : \"modified\",\n name: c.detail.split(\" \")[1] ?? c.detail,\n detail: c.detail,\n });\n } else if (c.object_type === \"enum\") {\n enumDiffs.push({\n type: c.change_type === \"added\" ? \"extra\" : c.change_type === \"removed\" ? \"missing\" : \"modified\",\n name: c.detail.split(\" \")[1] ?? c.detail,\n detail: c.detail,\n });\n }\n }\n }\n\n const schema: SchemaDiff = { missingTables, extraTables, columnDiffs, indexDiffs, constraintDiffs, enumDiffs };\n const schemaDrifts = countSchemaDrifts(schema);\n\n let health: HealthDiff | undefined;\n\n if (options?.includeHealth) {\n const longQueryThreshold = 5;\n const [srcReport, tgtReport] = await Promise.all([\n getAdvisorReport(sourcePool, longQueryThreshold),\n getAdvisorReport(targetPool, longQueryThreshold),\n ]);\n\n const srcIssueKeys = new Set(srcReport.issues.map((i) => i.title));\n const tgtIssueKeys = new Set(tgtReport.issues.map((i) => i.title));\n\n const sourceOnlyIssues = srcReport.issues\n .filter((i) => !tgtIssueKeys.has(i.title))\n .map((i) => `${i.severity}: ${i.title}`);\n\n const targetOnlyIssues = tgtReport.issues\n .filter((i) => !srcIssueKeys.has(i.title))\n .map((i) => `${i.severity}: ${i.title}`);\n\n health = {\n source: { score: srcReport.score, grade: srcReport.grade, url: maskConnectionString(sourceConn) },\n target: { score: tgtReport.score, grade: tgtReport.grade, url: maskConnectionString(targetConn) },\n sourceOnlyIssues,\n targetOnlyIssues,\n };\n }\n\n return {\n schema,\n health,\n checkedAt: new Date().toISOString(),\n summary: {\n schemaDrifts,\n identical: schemaDrifts === 0,\n },\n };\n } finally {\n await Promise.allSettled([sourcePool.end(), targetPool.end()]);\n }\n}\n\n/** Mask password in a connection string to avoid leaking credentials */\nfunction maskConnectionString(connStr: string): string {\n try {\n const url = new URL(connStr);\n if (url.password) url.password = \"***\";\n return url.toString();\n } catch {\n return \"<redacted>\";\n }\n}\n\n// ----- formatters -----\n\nexport function formatTextDiff(result: EnvDiffResult): string {\n const lines: string[] = [];\n const sep = \"══════════════════════════════════════\";\n\n lines.push(`Environment Diff`);\n lines.push(sep);\n lines.push(``);\n lines.push(`Schema Drift:`);\n\n const { schema } = result;\n\n if (schema.missingTables.length > 0) {\n lines.push(` ✗ target missing tables: ${schema.missingTables.join(\", \")}`);\n }\n if (schema.extraTables.length > 0) {\n lines.push(` ⚠ target has extra tables: ${schema.extraTables.join(\", \")}`);\n }\n\n const missingCols: string[] = [];\n const extraCols: string[] = [];\n const typeChanges: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const col of cd.missingColumns) {\n missingCols.push(` ${cd.table}: ${col.name} (${col.type})`);\n }\n for (const col of cd.extraColumns) {\n extraCols.push(` ${cd.table}: ${col.name} (${col.type})`);\n }\n for (const td of cd.typeDiffs) {\n typeChanges.push(` ${cd.table}.${td.column}: ${td.sourceType} → ${td.targetType}`);\n }\n }\n\n if (missingCols.length > 0) {\n lines.push(` ✗ target missing columns:`);\n lines.push(...missingCols);\n }\n if (extraCols.length > 0) {\n lines.push(` ⚠ target has extra columns:`);\n lines.push(...extraCols);\n }\n if (typeChanges.length > 0) {\n lines.push(` ~ column type differences:`);\n lines.push(...typeChanges);\n }\n\n const nullableChanges: string[] = [];\n const defaultChanges: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const nd of cd.nullableDiffs) {\n const src = nd.sourceNullable ? \"nullable\" : \"NOT NULL\";\n const tgt = nd.targetNullable ? \"nullable\" : \"NOT NULL\";\n nullableChanges.push(` ${cd.table}.${nd.column}: source=${src} → target=${tgt}`);\n }\n for (const dd of cd.defaultDiffs) {\n const src = dd.sourceDefault ?? \"(none)\";\n const tgt = dd.targetDefault ?? \"(none)\";\n defaultChanges.push(` ${cd.table}.${dd.column}: source=${src} → target=${tgt}`);\n }\n }\n\n if (nullableChanges.length > 0) {\n lines.push(` ~ nullable differences:`);\n lines.push(...nullableChanges);\n }\n if (defaultChanges.length > 0) {\n lines.push(` ~ default differences:`);\n lines.push(...defaultChanges);\n }\n\n const missingIdxs: string[] = [];\n const extraIdxs: string[] = [];\n const modifiedIdxs: string[] = [];\n\n for (const id of schema.indexDiffs) {\n for (const idx of id.missingIndexes) {\n missingIdxs.push(` ${id.table}: ${idx}`);\n }\n for (const idx of id.extraIndexes) {\n extraIdxs.push(` ${id.table}: ${idx}`);\n }\n for (const mi of id.modifiedIndexes) {\n modifiedIdxs.push(` ${id.table}: ${mi.name} source=\"${mi.sourceDef}\" → target=\"${mi.targetDef}\"`);\n }\n }\n\n if (missingIdxs.length > 0) {\n lines.push(` ✗ target missing indexes:`);\n lines.push(...missingIdxs);\n }\n if (extraIdxs.length > 0) {\n lines.push(` ⚠ target has extra indexes:`);\n lines.push(...extraIdxs);\n }\n if (modifiedIdxs.length > 0) {\n lines.push(` ~ index definition differences:`);\n lines.push(...modifiedIdxs);\n }\n\n // Constraint diffs\n const missingConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"missing\");\n const extraConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"extra\");\n const modifiedConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"modified\");\n\n if (missingConstraints.length > 0) {\n lines.push(` ✗ target missing constraints:`);\n for (const c of missingConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n if (extraConstraints.length > 0) {\n lines.push(` ⚠ target has extra constraints:`);\n for (const c of extraConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n if (modifiedConstraints.length > 0) {\n lines.push(` ~ constraint differences:`);\n for (const c of modifiedConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n\n // Enum diffs\n const missingEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"missing\");\n const extraEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"extra\");\n const modifiedEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"modified\");\n\n if (missingEnums.length > 0) {\n lines.push(` ✗ target missing enums:`);\n for (const e of missingEnums) lines.push(` ${e.detail}`);\n }\n if (extraEnums.length > 0) {\n lines.push(` ⚠ target has extra enums:`);\n for (const e of extraEnums) lines.push(` ${e.detail}`);\n }\n if (modifiedEnums.length > 0) {\n lines.push(` ~ enum differences:`);\n for (const e of modifiedEnums) lines.push(` ${e.detail}`);\n }\n\n const noSchemaChanges = schema.missingTables.length === 0 && schema.extraTables.length === 0 &&\n schema.columnDiffs.length === 0 && schema.indexDiffs.length === 0 &&\n (schema.constraintDiffs ?? []).length === 0 && (schema.enumDiffs ?? []).length === 0 &&\n nullableChanges.length === 0 && defaultChanges.length === 0 && modifiedIdxs.length === 0;\n if (noSchemaChanges) {\n lines.push(` ✓ Schemas are identical`);\n }\n\n if (result.health) {\n const h = result.health;\n lines.push(``);\n lines.push(`Health Comparison:`);\n lines.push(` Source: ${h.source.score}/100 (${h.source.grade}) | Target: ${h.target.score}/100 (${h.target.grade})`);\n lines.push(` Source-only issues: ${h.sourceOnlyIssues.length === 0 ? \"(none)\" : \"\"}`);\n for (const iss of h.sourceOnlyIssues) lines.push(` - ${iss}`);\n lines.push(` Target-only issues: ${h.targetOnlyIssues.length === 0 ? \"(none)\" : \"\"}`);\n for (const iss of h.targetOnlyIssues) lines.push(` - ${iss}`);\n }\n\n lines.push(``);\n lines.push(sep);\n const { schemaDrifts, identical } = result.summary;\n lines.push(`Total: ${schemaDrifts} schema drift${schemaDrifts !== 1 ? \"s\" : \"\"} | Environments are ${identical ? \"in sync ✓\" : \"NOT in sync ✗\"}`);\n\n return lines.join(\"\\n\");\n}\n\nexport function formatMdDiff(result: EnvDiffResult): string {\n const lines: string[] = [];\n lines.push(`## 🔄 Environment Diff`);\n lines.push(``);\n lines.push(`### Schema Drift`);\n lines.push(``);\n\n const { schema } = result;\n const rows: Array<[string, string]> = [];\n\n if (schema.missingTables.length > 0) {\n rows.push([`❌ Missing tables`, schema.missingTables.map((t) => `\\`${t}\\``).join(\", \")]);\n }\n if (schema.extraTables.length > 0) {\n rows.push([`⚠️ Extra tables`, schema.extraTables.map((t) => `\\`${t}\\``).join(\", \")]);\n }\n\n const missingColItems: string[] = [];\n const extraColItems: string[] = [];\n const typeItems: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const col of cd.missingColumns) {\n missingColItems.push(`\\`${cd.table}.${col.name}\\``);\n }\n for (const col of cd.extraColumns) {\n extraColItems.push(`\\`${cd.table}.${col.name}\\``);\n }\n for (const td of cd.typeDiffs) {\n typeItems.push(`\\`${cd.table}.${td.column}\\` (${td.sourceType}→${td.targetType})`);\n }\n }\n\n if (missingColItems.length > 0) rows.push([`❌ Missing columns`, missingColItems.join(\", \")]);\n if (extraColItems.length > 0) rows.push([`⚠️ Extra columns`, extraColItems.join(\", \")]);\n if (typeItems.length > 0) rows.push([`~ Type differences`, typeItems.join(\", \")]);\n\n const nullableItems: string[] = [];\n const defaultItems: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const nd of cd.nullableDiffs) {\n const src = nd.sourceNullable ? \"nullable\" : \"NOT NULL\";\n const tgt = nd.targetNullable ? \"nullable\" : \"NOT NULL\";\n nullableItems.push(`\\`${cd.table}.${nd.column}\\` (${src}→${tgt})`);\n }\n for (const dd of cd.defaultDiffs) {\n const src = dd.sourceDefault ?? \"(none)\";\n const tgt = dd.targetDefault ?? \"(none)\";\n defaultItems.push(`\\`${cd.table}.${dd.column}\\` (${src}→${tgt})`);\n }\n }\n\n if (nullableItems.length > 0) rows.push([`~ Nullable differences`, nullableItems.join(\", \")]);\n if (defaultItems.length > 0) rows.push([`~ Default differences`, defaultItems.join(\", \")]);\n\n const missingIdxItems: string[] = [];\n const extraIdxItems: string[] = [];\n const modifiedIdxItems: string[] = [];\n\n for (const id of schema.indexDiffs) {\n for (const idx of id.missingIndexes) missingIdxItems.push(`\\`${id.table}.${idx}\\``);\n for (const idx of id.extraIndexes) extraIdxItems.push(`\\`${id.table}.${idx}\\``);\n for (const mi of id.modifiedIndexes) modifiedIdxItems.push(`\\`${id.table}.${mi.name}\\``);\n }\n\n if (missingIdxItems.length > 0) rows.push([`❌ Missing indexes`, missingIdxItems.join(\", \")]);\n if (extraIdxItems.length > 0) rows.push([`⚠️ Extra indexes`, extraIdxItems.join(\", \")]);\n if (modifiedIdxItems.length > 0) rows.push([`~ Modified indexes`, modifiedIdxItems.join(\", \")]);\n\n // Constraints\n const missingConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"missing\").map((c) => c.detail);\n const extraConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"extra\").map((c) => c.detail);\n const modConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"modified\").map((c) => c.detail);\n if (missingConItems.length > 0) rows.push([`❌ Missing constraints`, missingConItems.join(\"; \")]);\n if (extraConItems.length > 0) rows.push([`⚠️ Extra constraints`, extraConItems.join(\"; \")]);\n if (modConItems.length > 0) rows.push([`~ Modified constraints`, modConItems.join(\"; \")]);\n\n // Enums\n const missingEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"missing\").map((e) => e.detail);\n const extraEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"extra\").map((e) => e.detail);\n const modEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"modified\").map((e) => e.detail);\n if (missingEnumItems.length > 0) rows.push([`❌ Missing enums`, missingEnumItems.join(\"; \")]);\n if (extraEnumItems.length > 0) rows.push([`⚠️ Extra enums`, extraEnumItems.join(\"; \")]);\n if (modEnumItems.length > 0) rows.push([`~ Modified enums`, modEnumItems.join(\"; \")]);\n\n if (rows.length > 0) {\n lines.push(`| Type | Details |`);\n lines.push(`|------|---------|`);\n for (const [type, details] of rows) {\n lines.push(`| ${type} | ${details} |`);\n }\n } else {\n lines.push(`✅ Schemas are identical`);\n }\n\n if (result.health) {\n const h = result.health;\n lines.push(``);\n lines.push(`### Health Comparison`);\n lines.push(``);\n lines.push(`| | Score | Grade |`);\n lines.push(`|--|-------|-------|`);\n lines.push(`| Source | ${h.source.score}/100 | ${h.source.grade} |`);\n lines.push(`| Target | ${h.target.score}/100 | ${h.target.grade} |`);\n\n if (h.targetOnlyIssues.length > 0) {\n lines.push(``);\n lines.push(`**Target-only issues:**`);\n for (const iss of h.targetOnlyIssues) lines.push(`- ${iss}`);\n }\n if (h.sourceOnlyIssues.length > 0) {\n lines.push(``);\n lines.push(`**Source-only issues:**`);\n for (const iss of h.sourceOnlyIssues) lines.push(`- ${iss}`);\n }\n }\n\n lines.push(``);\n const { schemaDrifts, identical } = result.summary;\n lines.push(`**Result: ${schemaDrifts} drift${schemaDrifts !== 1 ? \"s\" : \"\"} — environments are ${identical ? \"in sync ✓\" : \"NOT in sync\"}**`);\n\n return lines.join(\"\\n\");\n}\n","// Schema Diff — compares two schema snapshots and produces a change list\n\nexport interface SchemaSnapshot {\n tables: SnapshotTable[];\n enums: SnapshotEnum[];\n}\n\nexport interface SnapshotTable {\n name: string;\n schema: string;\n columns: SnapshotColumn[];\n indexes: SnapshotIndex[];\n constraints: SnapshotConstraint[];\n}\n\nexport interface SnapshotColumn {\n name: string;\n type: string;\n nullable: boolean;\n default_value: string | null;\n}\n\nexport interface SnapshotIndex {\n name: string;\n definition: string;\n is_unique: boolean;\n is_primary: boolean;\n}\n\nexport interface SnapshotConstraint {\n name: string;\n type: string;\n definition: string;\n}\n\nexport interface SnapshotEnum {\n name: string;\n schema: string;\n values: string[];\n}\n\nexport interface SchemaChange {\n change_type: \"added\" | \"removed\" | \"modified\";\n object_type: \"table\" | \"column\" | \"index\" | \"constraint\" | \"enum\";\n table_name: string | null;\n detail: string;\n}\n\nexport function diffSchemaSnapshots(oldSnap: SchemaSnapshot, newSnap: SchemaSnapshot): SchemaChange[] {\n const changes: SchemaChange[] = [];\n\n const oldTableMap = new Map(oldSnap.tables.map((t) => [`${t.schema}.${t.name}`, t]));\n const newTableMap = new Map(newSnap.tables.map((t) => [`${t.schema}.${t.name}`, t]));\n\n // Tables added/removed\n for (const [key, t] of newTableMap) {\n if (!oldTableMap.has(key)) {\n changes.push({ change_type: \"added\", object_type: \"table\", table_name: key, detail: `Table ${key} added` });\n }\n }\n for (const [key] of oldTableMap) {\n if (!newTableMap.has(key)) {\n changes.push({ change_type: \"removed\", object_type: \"table\", table_name: key, detail: `Table ${key} removed` });\n }\n }\n\n // Compare matching tables\n for (const [key, newTable] of newTableMap) {\n const oldTable = oldTableMap.get(key);\n if (!oldTable) continue;\n\n // Columns\n const oldCols = new Map(oldTable.columns.map((c) => [c.name, c]));\n const newCols = new Map(newTable.columns.map((c) => [c.name, c]));\n\n for (const [name, col] of newCols) {\n const oldCol = oldCols.get(name);\n if (!oldCol) {\n changes.push({ change_type: \"added\", object_type: \"column\", table_name: key, detail: `Column ${name} added (${col.type})` });\n } else {\n if (oldCol.type !== col.type) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} type changed: ${oldCol.type} → ${col.type}` });\n }\n if (oldCol.nullable !== col.nullable) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} nullable changed: ${oldCol.nullable} → ${col.nullable}` });\n }\n if (oldCol.default_value !== col.default_value) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} default changed: ${oldCol.default_value ?? \"NULL\"} → ${col.default_value ?? \"NULL\"}` });\n }\n }\n }\n for (const name of oldCols.keys()) {\n if (!newCols.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"column\", table_name: key, detail: `Column ${name} removed` });\n }\n }\n\n // Indexes\n const oldIdx = new Map(oldTable.indexes.map((i) => [i.name, i]));\n const newIdx = new Map(newTable.indexes.map((i) => [i.name, i]));\n for (const [name, idx] of newIdx) {\n if (!oldIdx.has(name)) {\n changes.push({ change_type: \"added\", object_type: \"index\", table_name: key, detail: `Index ${name} added` });\n } else if (oldIdx.get(name)!.definition !== idx.definition) {\n changes.push({ change_type: \"modified\", object_type: \"index\", table_name: key, detail: `Index ${name} definition changed` });\n }\n }\n for (const name of oldIdx.keys()) {\n if (!newIdx.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"index\", table_name: key, detail: `Index ${name} removed` });\n }\n }\n\n // Constraints\n const oldCon = new Map(oldTable.constraints.map((c) => [c.name, c]));\n const newCon = new Map(newTable.constraints.map((c) => [c.name, c]));\n for (const [name, con] of newCon) {\n if (!oldCon.has(name)) {\n changes.push({ change_type: \"added\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} added (${con.type})` });\n } else if (oldCon.get(name)!.definition !== con.definition) {\n changes.push({ change_type: \"modified\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} definition changed` });\n }\n }\n for (const name of oldCon.keys()) {\n if (!newCon.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} removed` });\n }\n }\n }\n\n // Enums\n const oldEnums = new Map((oldSnap.enums || []).map((e) => [`${e.schema}.${e.name}`, e]));\n const newEnums = new Map((newSnap.enums || []).map((e) => [`${e.schema}.${e.name}`, e]));\n for (const [key, en] of newEnums) {\n const oldEn = oldEnums.get(key);\n if (!oldEn) {\n changes.push({ change_type: \"added\", object_type: \"enum\", table_name: null, detail: `Enum ${key} added (${en.values.join(\", \")})` });\n } else {\n const added = en.values.filter((v) => !oldEn.values.includes(v));\n const removed = oldEn.values.filter((v) => !en.values.includes(v));\n for (const v of added) {\n changes.push({ change_type: \"modified\", object_type: \"enum\", table_name: null, detail: `Enum ${key}: value '${v}' added` });\n }\n for (const v of removed) {\n changes.push({ change_type: \"modified\", object_type: \"enum\", table_name: null, detail: `Enum ${key}: value '${v}' removed` });\n }\n }\n }\n for (const key of oldEnums.keys()) {\n if (!newEnums.has(key)) {\n changes.push({ change_type: \"removed\", object_type: \"enum\", table_name: null, detail: `Enum ${key} removed` });\n }\n }\n\n return changes;\n}\n","// Schema Tracker — takes schema snapshots, stores in SQLite, detects changes\n\nconst SNAPSHOT_RETENTION = 50;\n\nimport type { Pool } from \"pg\";\nimport type Database from \"better-sqlite3\";\nimport { getSchemaTables, getSchemaTableDetail, getSchemaEnums } from \"./queries/schema.js\";\nimport { diffSchemaSnapshots, type SchemaSnapshot, type SchemaChange } from \"./schema-diff.js\";\n\n/** Build a full schema snapshot from a live pool — reusable for env comparison */\nexport async function buildLiveSnapshot(pool: Pool): Promise<SchemaSnapshot> {\n const tables = await getSchemaTables(pool);\n const enums = await getSchemaEnums(pool);\n\n const detailedTables = await Promise.all(\n tables.map(async (t: any) => {\n const detail = await getSchemaTableDetail(pool, `${t.schema}.${t.name}`);\n if (!detail) return null;\n return {\n name: detail.name,\n schema: detail.schema,\n columns: detail.columns.map((c: any) => ({\n name: c.name,\n type: c.type,\n nullable: c.nullable,\n default_value: c.default_value,\n })),\n indexes: detail.indexes.map((i: any) => ({\n name: i.name,\n definition: i.definition,\n is_unique: i.is_unique,\n is_primary: i.is_primary,\n })),\n constraints: detail.constraints.map((c: any) => ({\n name: c.name,\n type: c.type,\n definition: c.definition,\n })),\n };\n })\n );\n\n return {\n tables: detailedTables.filter(Boolean) as SchemaSnapshot[\"tables\"],\n enums: enums.map((e: any) => ({ name: e.name, schema: e.schema, values: e.values })),\n };\n}\n\nexport class SchemaTracker {\n private db: Database.Database;\n private pool: Pool;\n private intervalMs: number;\n private timer: ReturnType<typeof setInterval> | null = null;\n\n constructor(db: Database.Database, pool: Pool, intervalMs = 6 * 60 * 60 * 1000) {\n this.db = db;\n this.pool = pool;\n this.intervalMs = intervalMs;\n this.initTables();\n }\n\n private initTables() {\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS schema_snapshots (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp INTEGER NOT NULL,\n snapshot TEXT NOT NULL\n );\n CREATE TABLE IF NOT EXISTS schema_changes (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n snapshot_id INTEGER NOT NULL,\n timestamp INTEGER NOT NULL,\n change_type TEXT NOT NULL,\n object_type TEXT NOT NULL,\n table_name TEXT,\n detail TEXT NOT NULL,\n FOREIGN KEY (snapshot_id) REFERENCES schema_snapshots(id)\n );\n `);\n }\n\n async takeSnapshot(): Promise<{ snapshotId: number; changes: SchemaChange[] }> {\n const snapshot = await this.buildSnapshot();\n const now = Date.now();\n const json = JSON.stringify(snapshot);\n\n const info = this.db.prepare(\"INSERT INTO schema_snapshots (timestamp, snapshot) VALUES (?, ?)\").run(now, json);\n const snapshotId = Number(info.lastInsertRowid);\n\n // Prune old snapshots, keeping only the most recent SNAPSHOT_RETENTION\n this.db.prepare(`\n DELETE FROM schema_snapshots\n WHERE id NOT IN (\n SELECT id FROM schema_snapshots\n ORDER BY timestamp DESC\n LIMIT ?\n )\n `).run(SNAPSHOT_RETENTION);\n\n // Diff against previous\n const prev = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id < ? ORDER BY id DESC LIMIT 1\").get(snapshotId) as { snapshot: string } | undefined;\n let changes: SchemaChange[] = [];\n if (prev) {\n const oldSnap: SchemaSnapshot = JSON.parse(prev.snapshot);\n changes = diffSchemaSnapshots(oldSnap, snapshot);\n if (changes.length > 0) {\n const insert = this.db.prepare(\"INSERT INTO schema_changes (snapshot_id, timestamp, change_type, object_type, table_name, detail) VALUES (?, ?, ?, ?, ?, ?)\");\n const tx = this.db.transaction((chs: SchemaChange[]) => {\n for (const c of chs) {\n insert.run(snapshotId, now, c.change_type, c.object_type, c.table_name, c.detail);\n }\n });\n tx(changes);\n }\n }\n\n return { snapshotId, changes };\n }\n\n private async buildSnapshot(): Promise<SchemaSnapshot> {\n return buildLiveSnapshot(this.pool);\n }\n\n start() {\n // Take initial snapshot\n this.takeSnapshot().catch((err) => console.error(\"Schema snapshot error:\", err.message));\n this.timer = setInterval(() => {\n this.takeSnapshot().catch((err) => console.error(\"Schema snapshot error:\", err.message));\n }, this.intervalMs);\n }\n\n stop() {\n if (this.timer) {\n clearInterval(this.timer);\n this.timer = null;\n }\n }\n\n // API helpers\n getHistory(limit = 30) {\n return this.db.prepare(\"SELECT id, timestamp FROM schema_snapshots ORDER BY id DESC LIMIT ?\").all(limit);\n }\n\n getChanges(since?: number) {\n if (since) {\n return this.db.prepare(\"SELECT * FROM schema_changes WHERE timestamp >= ? ORDER BY timestamp DESC\").all(since);\n }\n return this.db.prepare(\"SELECT * FROM schema_changes ORDER BY timestamp DESC LIMIT 100\").all();\n }\n\n getLatestChanges() {\n const latest = this.db.prepare(\"SELECT id FROM schema_snapshots ORDER BY id DESC LIMIT 1\").get() as { id: number } | undefined;\n if (!latest) return [];\n return this.db.prepare(\"SELECT * FROM schema_changes WHERE snapshot_id = ? ORDER BY id\").all(latest.id);\n }\n\n getDiff(fromId: number, toId: number) {\n const from = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id = ?\").get(fromId) as { snapshot: string } | undefined;\n const to = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id = ?\").get(toId) as { snapshot: string } | undefined;\n if (!from || !to) return null;\n return diffSchemaSnapshots(JSON.parse(from.snapshot), JSON.parse(to.snapshot));\n }\n}\n","// query-analyzer.ts — deep EXPLAIN plan analysis with auto index suggestions\n\nimport type { Pool } from \"pg\";\n\n// ─── Types ───────────────────────────────────────────────────────────────────\n\nexport interface PlanNodeSummary {\n nodeType: string;\n table?: string;\n totalCost: number;\n actualRows?: number;\n actualTime?: number; // ms\n filter?: string;\n}\n\nexport interface SeqScanInfo {\n table: string;\n rowCount: number;\n filter?: string; // filter condition from explain output\n suggestion?: string;\n}\n\nexport interface IndexSuggestion {\n table: string;\n columns: string[];\n reason: string;\n sql: string; // CREATE INDEX CONCURRENTLY …\n estimatedBenefit: \"high\" | \"medium\" | \"low\";\n}\n\nexport interface ExplainAnalysis {\n planNodes: PlanNodeSummary[];\n seqScans: SeqScanInfo[];\n missingIndexes: IndexSuggestion[];\n costEstimate: {\n totalCost: number;\n actualTime?: number;\n planningTime?: number;\n };\n recommendations: string[];\n}\n\nexport interface QueryRegressionInfo {\n queryId: string; // queryid from pg_stat_statements\n currentMeanMs: number;\n previousMeanMs: number;\n changePercent: number;\n degradedAt?: string; // approximate timestamp\n}\n\n// ─── Helpers ─────────────────────────────────────────────────────────────────\n\n/**\n * Recursively walk a plan tree (EXPLAIN FORMAT JSON) and collect every node.\n * Each node looks like { \"Node Type\": \"...\", \"Plans\": [...], ... }\n */\nfunction collectNodes(node: any, acc: any[] = []): any[] {\n if (!node || typeof node !== \"object\") return acc;\n acc.push(node);\n const plans = node[\"Plans\"] ?? node[\"plans\"];\n if (Array.isArray(plans)) {\n for (const child of plans) collectNodes(child, acc);\n }\n return acc;\n}\n\n/**\n * Extract simple column names from a Postgres filter expression.\n * Handles patterns like:\n * (col = $1)\n * (col > $1)\n * (col IS NULL)\n * (col IS NOT NULL)\n * (col ~~ '%foo%') -- LIKE\n */\nfunction extractColumnsFromFilter(filter: string): string[] {\n // Match identifiers that appear before comparison operators\n const colPattern = /\\(?\"?([a-z_][a-z0-9_]*)\"?\\s*(?:=|<|>|<=|>=|<>|!=|IS\\s+(?:NOT\\s+)?NULL|~~|!~~)/gi;\n const found = new Set<string>();\n let m: RegExpExecArray | null;\n while ((m = colPattern.exec(filter)) !== null) {\n const col = m[1].toLowerCase();\n // Skip Postgres internal names\n if (![\"and\", \"or\", \"not\", \"true\", \"false\", \"null\"].includes(col)) {\n found.add(col);\n }\n }\n return Array.from(found);\n}\n\n/**\n * Fetch the list of indexed column sets for a given table from pg_indexes.\n * Returns an array of column name arrays (one per index).\n */\nasync function getExistingIndexColumns(pool: Pool, tableName: string): Promise<string[][]> {\n try {\n // Query pg_indexes to get index definitions\n const r = await pool.query(\n `SELECT indexdef FROM pg_indexes WHERE tablename = $1`,\n [tableName]\n );\n return r.rows.map((row: any) => {\n // Parse column list from: ... ON table (col1, col2, ...)\n const m = /\\(([^)]+)\\)/.exec(row.indexdef);\n if (!m) return [] as string[];\n return m[1]\n .split(\",\")\n .map((c: string) => c.trim().replace(/^\"|\"$/g, \"\").toLowerCase());\n });\n } catch {\n return [];\n }\n}\n\n/**\n * Benefit rating based on estimated row count.\n */\nfunction rateBenefit(rowCount: number): \"high\" | \"medium\" | \"low\" {\n if (rowCount > 100_000) return \"high\";\n if (rowCount >= 10_000) return \"medium\";\n return \"low\";\n}\n\n/**\n * Format a large number as human-readable (1.2M, 50K, etc.)\n */\nfunction fmtRows(n: number): string {\n if (n >= 1_000_000) return `${(n / 1_000_000).toFixed(1)}M`;\n if (n >= 1_000) return `${(n / 1_000).toFixed(0)}K`;\n return String(n);\n}\n\n// ─── Core analysis ────────────────────────────────────────────────────────────\n\n/**\n * Analyse a EXPLAIN (FORMAT JSON) result and return rich diagnostics.\n *\n * @param explainJson - The value of `r.rows[0][\"QUERY PLAN\"]` (an array with one plan object)\n * @param pool - Optional PG pool; without it only static analysis is performed\n */\nexport async function analyzeExplainPlan(\n explainJson: any,\n pool?: Pool | null\n): Promise<ExplainAnalysis> {\n const result: ExplainAnalysis = {\n planNodes: [],\n seqScans: [],\n missingIndexes: [],\n costEstimate: { totalCost: 0 },\n recommendations: [],\n };\n\n if (!explainJson || !Array.isArray(explainJson) || explainJson.length === 0) {\n return result;\n }\n\n const topLevel = explainJson[0];\n const planRoot = topLevel?.[\"Plan\"] ?? topLevel?.[\"plan\"];\n\n // Planning / execution times from top-level\n const planningTime: number | undefined = topLevel?.[\"Planning Time\"] ?? undefined;\n const executionTime: number | undefined = topLevel?.[\"Execution Time\"] ?? undefined;\n\n if (!planRoot) return result;\n\n // Collect all nodes\n const allNodes = collectNodes(planRoot);\n\n // Build planNodes summary\n result.planNodes = allNodes.map((n: any) => {\n const s: PlanNodeSummary = {\n nodeType: n[\"Node Type\"] ?? \"Unknown\",\n totalCost: n[\"Total Cost\"] ?? 0,\n };\n if (n[\"Relation Name\"]) s.table = n[\"Relation Name\"];\n if (n[\"Actual Rows\"] !== undefined) s.actualRows = n[\"Actual Rows\"];\n if (n[\"Actual Total Time\"] !== undefined) s.actualTime = n[\"Actual Total Time\"];\n if (n[\"Filter\"]) s.filter = n[\"Filter\"];\n return s;\n });\n\n // Cost estimate from root node\n result.costEstimate = {\n totalCost: planRoot[\"Total Cost\"] ?? 0,\n actualTime: executionTime,\n planningTime,\n };\n\n // ── Seq Scan analysis ──────────────────────────────────────────────────────\n const seqScanNodes = allNodes.filter((n: any) => n[\"Node Type\"] === \"Seq Scan\");\n\n for (const node of seqScanNodes) {\n const table: string = node[\"Relation Name\"] ?? \"unknown\";\n const rowCount: number = node[\"Plan Rows\"] ?? node[\"Actual Rows\"] ?? 0;\n const filter: string | undefined = node[\"Filter\"];\n\n const info: SeqScanInfo = { table, rowCount, filter };\n\n if (rowCount > 10_000) {\n info.suggestion = filter\n ? `Consider adding an index to support the filter on ${table}`\n : `Full table scan on large table ${table} — review query`;\n }\n\n result.seqScans.push(info);\n }\n\n // ── Missing index inference ────────────────────────────────────────────────\n for (const scan of result.seqScans) {\n if (!scan.filter) continue;\n\n const cols = extractColumnsFromFilter(scan.filter);\n if (cols.length === 0) continue;\n\n // Check existing indexes (needs DB)\n let existingIndexCols: string[][] = [];\n if (pool) {\n existingIndexCols = await getExistingIndexColumns(pool, scan.table);\n }\n\n // Filter out columns already covered as the leading column of an existing index\n const uncoveredCols = cols.filter(\n (col) => !existingIndexCols.some((idxCols) => idxCols.length > 0 && idxCols[0] === col)\n );\n\n if (uncoveredCols.length === 0) continue;\n\n const benefit = rateBenefit(scan.rowCount);\n\n if (uncoveredCols.length >= 2) {\n // Suggest a composite index\n const idxName = `idx_${scan.table}_${uncoveredCols.join(\"_\")}`;\n const sql = `CREATE INDEX CONCURRENTLY ${idxName} ON ${scan.table} (${uncoveredCols.join(\", \")})`;\n result.missingIndexes.push({\n table: scan.table,\n columns: uncoveredCols,\n reason: `Seq Scan with multi-column filter (${uncoveredCols.join(\", \")}) on ${fmtRows(scan.rowCount)} rows — composite index preferred`,\n sql,\n estimatedBenefit: benefit,\n });\n } else {\n // Single column\n const col = uncoveredCols[0];\n const idxName = `idx_${scan.table}_${col}`;\n const sql = `CREATE INDEX CONCURRENTLY ${idxName} ON ${scan.table} (${col})`;\n result.missingIndexes.push({\n table: scan.table,\n columns: [col],\n reason: `Seq Scan with Filter on ${col} (${fmtRows(scan.rowCount)} rows)`,\n sql,\n estimatedBenefit: benefit,\n });\n }\n }\n\n // ── Recommendations ────────────────────────────────────────────────────────\n for (const scan of result.seqScans) {\n if (scan.rowCount > 10_000) {\n const filterPart = scan.filter\n ? ` — consider adding index on ${extractColumnsFromFilter(scan.filter).join(\", \") || \"filter columns\"}`\n : \" — no filter; full scan may be intentional\";\n result.recommendations.push(\n `Seq Scan on ${scan.table} (${fmtRows(scan.rowCount)} rows)${filterPart}`\n );\n }\n }\n\n if (planningTime !== undefined) {\n const label = planningTime > 10 ? \"high — check statistics\" : \"normal\";\n result.recommendations.push(`Planning time ${planningTime.toFixed(1)}ms — ${label}`);\n }\n\n if (result.missingIndexes.length === 0 && result.seqScans.length === 0) {\n result.recommendations.push(\"No obvious sequential scans detected — query looks efficient\");\n }\n\n return result;\n}\n\n// ─── Regression detection ─────────────────────────────────────────────────────\n\n/**\n * Detect queries whose mean execution time has increased by more than 50%\n * compared to the earliest snapshot in the query_stats store for the given window.\n *\n * This is a best-effort function; it silently returns [] if pg_stat_statements\n * is unavailable or the query_stats store doesn't have enough history.\n *\n * @param pool - PG pool (used to read pg_stat_statements)\n * @param statsDb - Optional better-sqlite3 Database with query_stats table\n * @param windowHours - How far back to compare (default 24 h)\n */\nexport async function detectQueryRegressions(\n pool: Pool,\n statsDb?: any | null,\n windowHours = 24\n): Promise<QueryRegressionInfo[]> {\n try {\n // ── 1. Check pg_stat_statements is available ───────────────────────────\n const extCheck = await pool.query(\n \"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\"\n );\n if (extCheck.rows.length === 0) return [];\n\n // ── 2. Get current snapshot from pg_stat_statements ───────────────────\n const current = await pool.query(`\n SELECT queryid::text AS queryid, mean_exec_time\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%'\n AND queryid IS NOT NULL\n `);\n\n const currentMap = new Map<string, number>();\n for (const row of current.rows) {\n currentMap.set(row.queryid, parseFloat(row.mean_exec_time));\n }\n\n if (!statsDb) return [];\n\n // ── 3. Fetch historical baselines from SQLite query_stats ─────────────\n const windowMs = windowHours * 60 * 60 * 1000;\n const since = Date.now() - windowMs;\n\n let historical: { queryid: string; mean_exec_time: number; timestamp: number }[];\n try {\n historical = statsDb\n .prepare(\n `SELECT queryid, mean_exec_time, timestamp\n FROM query_stats\n WHERE timestamp >= ?\n ORDER BY queryid, timestamp ASC`\n )\n .all(since) as any[];\n } catch {\n return [];\n }\n\n // Keep only the *earliest* record per queryid in the window\n const baselineMap = new Map<string, { meanMs: number; timestamp: number }>();\n for (const row of historical) {\n if (!baselineMap.has(row.queryid)) {\n baselineMap.set(row.queryid, {\n meanMs: row.mean_exec_time,\n timestamp: row.timestamp,\n });\n }\n }\n\n // ── 4. Detect regressions > 50% ────────────────────────────────────────\n const regressions: QueryRegressionInfo[] = [];\n\n for (const [queryId, baseline] of baselineMap) {\n const currentMean = currentMap.get(queryId);\n if (currentMean === undefined || baseline.meanMs === 0) continue;\n\n const changePercent =\n ((currentMean - baseline.meanMs) / baseline.meanMs) * 100;\n\n if (changePercent > 50) {\n regressions.push({\n queryId,\n currentMeanMs: currentMean,\n previousMeanMs: baseline.meanMs,\n changePercent: Math.round(changePercent),\n degradedAt: new Date(baseline.timestamp).toISOString(),\n });\n }\n }\n\n return regressions.sort((a, b) => b.changePercent - a.changePercent);\n } catch {\n return [];\n }\n}\n","// Migration safety checker — static + dynamic analysis of SQL migration files\n\nimport type { Pool } from \"pg\";\n\nexport interface MigrationIssue {\n severity: \"error\" | \"warning\" | \"info\";\n code: string;\n message: string;\n suggestion?: string;\n lineNumber?: number;\n tableName?: string;\n estimatedRows?: number;\n estimatedLockSeconds?: number;\n}\n\nexport interface MigrationCheckResult {\n safe: boolean;\n issues: MigrationIssue[];\n summary: {\n errors: number;\n warnings: number;\n infos: number;\n };\n checkedAt: string;\n}\n\n// Strip SQL comments while preserving line numbers (replace with spaces)\nfunction stripComments(sql: string): string {\n // Replace /* ... */ block comments (preserve newlines for line number tracking)\n let stripped = sql.replace(/\\/\\*[\\s\\S]*?\\*\\//g, (match) =>\n match.replace(/[^\\n]/g, \" \")\n );\n // Replace -- single-line comments (preserve the newline)\n stripped = stripped.replace(/--[^\\n]*/g, (match) => \" \".repeat(match.length));\n return stripped;\n}\n\n// Helper: find line number of a match in the original SQL\nfunction findLineNumber(sql: string, matchIndex: number): number {\n const before = sql.slice(0, matchIndex);\n return before.split(\"\\n\").length;\n}\n\n// Extract bare table name from possibly-quoted or schema-qualified identifier\nfunction bareTable(name: string): string {\n return name\n .replace(/^public\\./i, \"\")\n .replace(/\"/g, \"\")\n .toLowerCase()\n .trim();\n}\n\n// Parse all table names operated on by this migration\nfunction extractOperatedTables(sql: string): {\n indexTables: string[]; // CREATE INDEX ON <table>\n alterTables: string[]; // ALTER TABLE <table>\n dropTables: string[]; // DROP TABLE <table>\n refTables: string[]; // REFERENCES <table>\n} {\n sql = stripComments(sql);\n const indexTables: string[] = [];\n const alterTables: string[] = [];\n const dropTables: string[] = [];\n const refTables: string[] = [];\n\n // CREATE INDEX ... ON table\n const idxRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+(?:CONCURRENTLY\\s+)?(?:IF\\s+NOT\\s+EXISTS\\s+)?(?:\\w+\\s+)?ON\\s+([\\w.\"]+)/gi;\n let m: RegExpExecArray | null;\n while ((m = idxRe.exec(sql)) !== null) indexTables.push(bareTable(m[1]));\n\n // ALTER TABLE table\n const altRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n while ((m = altRe.exec(sql)) !== null) alterTables.push(bareTable(m[1]));\n\n // DROP TABLE\n const dropRe = /\\bDROP\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n while ((m = dropRe.exec(sql)) !== null) dropTables.push(bareTable(m[1]));\n\n // REFERENCES table\n const refRe = /\\bREFERENCES\\s+([\\w.\"]+)/gi;\n while ((m = refRe.exec(sql)) !== null) refTables.push(bareTable(m[1]));\n\n return { indexTables, alterTables, dropTables, refTables };\n}\n\n// Static analysis — no DB needed\nfunction staticCheck(sql: string): MigrationIssue[] {\n const issues: MigrationIssue[] = [];\n // Strip comments before analysis to avoid false positives from commented-out SQL\n sql = stripComments(sql);\n\n // Determine tables created IN THIS MIGRATION (so we know they're brand-new)\n const createdTablesRe = /\\bCREATE\\s+TABLE\\s+(?:IF\\s+NOT\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n const createdTables = new Set<string>();\n let m: RegExpExecArray | null;\n while ((m = createdTablesRe.exec(sql)) !== null) createdTables.add(bareTable(m[1]));\n\n // 1. CREATE INDEX without CONCURRENTLY (on tables NOT created in this migration)\n const idxRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+(?!CONCURRENTLY)((?:IF\\s+NOT\\s+EXISTS\\s+)?(?:\\w+\\s+)?ON\\s+([\\w.\"]+))/gi;\n while ((m = idxRe.exec(sql)) !== null) {\n const table = bareTable(m[2]);\n const lineNumber = findLineNumber(sql, m.index);\n if (!createdTables.has(table)) {\n issues.push({\n severity: \"warning\",\n code: \"INDEX_WITHOUT_CONCURRENTLY\",\n message: `CREATE INDEX on existing table will lock writes. Use CREATE INDEX CONCURRENTLY to avoid downtime.`,\n suggestion: \"Replace CREATE INDEX with CREATE INDEX CONCURRENTLY\",\n lineNumber,\n tableName: table,\n });\n }\n }\n\n // 2. CREATE INDEX CONCURRENTLY → info\n const idxConcRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+CONCURRENTLY\\b/gi;\n while ((m = idxConcRe.exec(sql)) !== null) {\n issues.push({\n severity: \"info\",\n code: \"INDEX_CONCURRENTLY_OK\",\n message: \"CREATE INDEX CONCURRENTLY — safe, no write lock\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 3 & 4. ALTER TABLE ... ADD COLUMN ... NOT NULL (with/without DEFAULT)\n // Match: ALTER TABLE <t> ADD COLUMN <col> <type> [DEFAULT <val>] [NOT NULL | NULL]\n const addColRe =\n /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ADD\\s+(?:COLUMN\\s+)?(?:IF\\s+NOT\\s+EXISTS\\s+)?[\\w\"]+\\s+[\\w\\s()\"',.[\\]]+?(?=;|$)/gi;\n while ((m = addColRe.exec(sql)) !== null) {\n const fragment = m[0];\n const table = bareTable(m[1]);\n const lineNumber = findLineNumber(sql, m.index);\n const fragUpper = fragment.toUpperCase();\n\n const hasNotNull = /\\bNOT\\s+NULL\\b/.test(fragUpper);\n const hasDefault = /\\bDEFAULT\\b/.test(fragUpper);\n\n if (hasNotNull && !hasDefault) {\n issues.push({\n severity: \"error\",\n code: \"ADD_COLUMN_NOT_NULL_NO_DEFAULT\",\n message: \"ADD COLUMN NOT NULL without DEFAULT will fail if table has existing rows\",\n suggestion: \"Add a DEFAULT value, then remove it after migration\",\n lineNumber,\n tableName: table,\n });\n } else if (hasNotNull && hasDefault) {\n issues.push({\n severity: \"warning\",\n code: \"ADD_COLUMN_REWRITES_TABLE\",\n message: \"ADD COLUMN with NOT NULL DEFAULT may rewrite table on PostgreSQL < 11\",\n suggestion: \"On PostgreSQL 11+ with a constant default this is safe. For older versions, add column nullable first.\",\n lineNumber,\n tableName: table,\n });\n }\n }\n\n // 5. DROP TABLE\n const dropRe = /\\bDROP\\s+TABLE\\b/gi;\n while ((m = dropRe.exec(sql)) !== null) {\n issues.push({\n severity: \"warning\",\n code: \"DROP_TABLE\",\n message: \"DROP TABLE is destructive. Ensure this is intentional and data is backed up.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 5b. ALTER COLUMN TYPE — rewrites the entire table and locks it\n const alterTypeRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ALTER\\s+(?:COLUMN\\s+)?[\\w\"]+\\s+TYPE\\b/gi;\n while ((m = alterTypeRe.exec(sql)) !== null) {\n const table = bareTable(m[1]);\n issues.push({\n severity: \"warning\",\n code: \"ALTER_COLUMN_TYPE\",\n message: \"ALTER COLUMN TYPE rewrites the entire table and acquires an exclusive lock.\",\n suggestion: \"Consider using a new column + backfill + rename strategy to avoid downtime.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5c. DROP COLUMN — safe in PostgreSQL 9.0+ (marks invisible, no rewrite), but breaks app code\n const dropColRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+DROP\\s+(?:COLUMN\\s+)(?:IF\\s+EXISTS\\s+)?[\\w\"]+\\b/gi;\n while ((m = dropColRe.exec(sql)) !== null) {\n const table = bareTable(m[1]);\n issues.push({\n severity: \"info\",\n code: \"DROP_COLUMN\",\n message: \"DROP COLUMN is safe in PostgreSQL (no table rewrite), but may break application code referencing that column.\",\n suggestion: \"Ensure no application code references this column before dropping it.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5d-i. RENAME TABLE\n const renameTableRe = /ALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?(\\w+)\\s+RENAME\\s+TO\\s+(\\w+)/gi;\n while ((m = renameTableRe.exec(sql)) !== null) {\n const oldName = m[1];\n const newName = m[2];\n issues.push({\n severity: \"warning\",\n code: \"RENAME_TABLE\",\n message: `Renaming table \"${oldName}\" to \"${newName}\" breaks application code referencing the old name`,\n suggestion: \"Deploy application code that handles both names before renaming, or use a view with the old name after renaming.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: oldName,\n });\n }\n\n // 5d-ii. RENAME COLUMN\n const renameColumnRe = /ALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?(\\w+)\\s+RENAME\\s+COLUMN\\s+(\\w+)\\s+TO\\s+(\\w+)/gi;\n while ((m = renameColumnRe.exec(sql)) !== null) {\n const table = m[1];\n const oldCol = m[2];\n const newCol = m[3];\n issues.push({\n severity: \"warning\",\n code: \"RENAME_COLUMN\",\n message: `Renaming column \"${oldCol}\" to \"${newCol}\" on table \"${table}\" breaks application code referencing the old column name`,\n suggestion: \"Add new column, backfill data, update application to use new column, then drop old column (expand/contract pattern).\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5e. ADD CONSTRAINT without NOT VALID — performs a full table scan to validate\n const addConRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ADD\\s+CONSTRAINT\\b[^;]*(;|$)/gi;\n while ((m = addConRe.exec(sql)) !== null) {\n const fragment = m[0];\n const table = bareTable(m[1]);\n const fragUpper = fragment.toUpperCase();\n // Skip if NOT VALID is already present\n if (!/\\bNOT\\s+VALID\\b/.test(fragUpper)) {\n issues.push({\n severity: \"warning\",\n code: \"ADD_CONSTRAINT_SCANS_TABLE\",\n message: \"ADD CONSTRAINT validates all existing rows and holds an exclusive lock during the scan.\",\n suggestion: \"Use ADD CONSTRAINT ... NOT VALID to skip validation, then VALIDATE CONSTRAINT in a separate transaction.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n }\n\n // 5e. CREATE INDEX CONCURRENTLY inside transaction (BEGIN/COMMIT)\n const hasTransaction = /\\bBEGIN\\b/i.test(sql) || /\\bSTART\\s+TRANSACTION\\b/i.test(sql);\n const hasConcurrently = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+CONCURRENTLY\\b/i.test(sql);\n if (hasTransaction && hasConcurrently) {\n issues.push({\n severity: \"error\",\n code: \"CONCURRENTLY_IN_TRANSACTION\",\n message: \"CREATE INDEX CONCURRENTLY cannot run inside a transaction block. It will fail at runtime.\",\n suggestion: \"Remove the BEGIN/COMMIT wrapper, or use a migration tool that runs CONCURRENTLY outside transactions.\",\n });\n }\n\n // 6. TRUNCATE\n const truncRe = /\\bTRUNCATE\\b/gi;\n while ((m = truncRe.exec(sql)) !== null) {\n issues.push({\n severity: \"warning\",\n code: \"TRUNCATE_TABLE\",\n message: \"TRUNCATE will delete all rows. Ensure this is intentional.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 7. DELETE FROM without WHERE\n const delRe = /\\bDELETE\\s+FROM\\s+[\\w.\"]+\\s*(?:;|$)/gi;\n while ((m = delRe.exec(sql)) !== null) {\n // If there's no WHERE clause in this statement\n const stmt = m[0];\n if (!/\\bWHERE\\b/i.test(stmt)) {\n issues.push({\n severity: \"warning\",\n code: \"DELETE_WITHOUT_WHERE\",\n message: \"DELETE without WHERE clause will remove all rows.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n }\n\n // 8. UPDATE ... SET without WHERE\n const updRe = /\\bUPDATE\\s+[\\w.\"]+\\s+SET\\b[^;]*(;|$)/gi;\n while ((m = updRe.exec(sql)) !== null) {\n const stmt = m[0];\n if (!/\\bWHERE\\b/i.test(stmt)) {\n issues.push({\n severity: \"warning\",\n code: \"UPDATE_WITHOUT_WHERE\",\n message: \"UPDATE without WHERE clause will modify all rows.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n }\n\n return issues;\n}\n\n// Dynamic analysis — requires a running PG pool\nasync function dynamicCheck(sql: string, pool: Pool, staticIssues: MigrationIssue[]): Promise<MigrationIssue[]> {\n const issues: MigrationIssue[] = [];\n const { indexTables, alterTables, dropTables, refTables } = extractOperatedTables(sql);\n\n // All tables we need to look up\n const allTables = [...new Set([...indexTables, ...alterTables, ...dropTables])];\n\n // Query row counts for all tables at once\n const tableStats = new Map<string, { rowCount: number; totalSize: number }>();\n if (allTables.length > 0) {\n try {\n const res = await pool.query<{ tablename: string; n_live_tup: string; total_size: string }>(\n `SELECT tablename,\n n_live_tup,\n pg_total_relation_size(schemaname||'.'||tablename) AS total_size\n FROM pg_stat_user_tables\n WHERE tablename = ANY($1)`,\n [allTables]\n );\n for (const row of res.rows) {\n tableStats.set(row.tablename, {\n rowCount: parseInt(row.n_live_tup ?? \"0\", 10),\n totalSize: parseInt(row.total_size ?? \"0\", 10),\n });\n }\n } catch (_) {\n // Ignore DB errors in dynamic check\n }\n }\n\n // Upgrade CREATE INDEX (non-CONCURRENTLY) issues based on actual row counts\n for (const issue of staticIssues) {\n if (issue.code === \"INDEX_WITHOUT_CONCURRENTLY\" && issue.tableName) {\n const stats = tableStats.get(issue.tableName);\n if (stats) {\n const { rowCount } = stats;\n const lockSecs = Math.round(rowCount / 50000);\n issue.estimatedRows = rowCount;\n issue.estimatedLockSeconds = lockSecs;\n\n if (rowCount > 1_000_000) {\n issue.severity = \"error\";\n issue.message = `CREATE INDEX on '${issue.tableName}' will lock writes for ~${lockSecs}s (${(rowCount / 1e6).toFixed(1)}M rows). CRITICAL — use CREATE INDEX CONCURRENTLY.`;\n } else if (rowCount > 100_000) {\n issue.message = `CREATE INDEX on '${issue.tableName}' will lock writes for ~${lockSecs}s (${(rowCount / 1000).toFixed(0)}k rows).`;\n }\n }\n }\n }\n\n // Validate REFERENCES tables exist\n const uniqueRefTables = [...new Set(refTables)];\n for (const table of uniqueRefTables) {\n try {\n const res = await pool.query<{ tablename: string }>(\n `SELECT tablename FROM pg_tables WHERE schemaname = 'public' AND tablename = $1`,\n [table]\n );\n if (res.rows.length === 0) {\n issues.push({\n severity: \"error\",\n code: \"MISSING_TABLE\",\n message: `Table '${table}' referenced in migration does not exist`,\n tableName: table,\n });\n }\n } catch (_) {\n // Ignore\n }\n }\n\n return issues;\n}\n\nexport async function analyzeMigration(sql: string, pool?: Pool): Promise<MigrationCheckResult> {\n const trimmed = sql.trim();\n\n if (!trimmed) {\n return {\n safe: true,\n issues: [],\n summary: { errors: 0, warnings: 0, infos: 0 },\n checkedAt: new Date().toISOString(),\n };\n }\n\n // Static checks first (mutates issue severity if dynamic info is available)\n const issues = staticCheck(trimmed);\n\n // Dynamic checks (augments existing issues + adds new ones like MISSING_TABLE)\n if (pool) {\n const dynamicIssues = await dynamicCheck(trimmed, pool, issues);\n issues.push(...dynamicIssues);\n }\n\n const errors = issues.filter((i) => i.severity === \"error\").length;\n const warnings = issues.filter((i) => i.severity === \"warning\").length;\n const infos = issues.filter((i) => i.severity === \"info\").length;\n\n return {\n safe: errors === 0,\n issues,\n summary: { errors, warnings, infos },\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface UnusedIndex {\n schema: string;\n table: string;\n index: string;\n indexSize: string; // human-readable e.g. \"2.4 MB\"\n indexSizeBytes: number;\n scans: number; // idx_scan from pg_stat_user_indexes\n lastUsed: string | null; // timestamp of last stats reset — index has had 0 scans since this date\n suggestion: string;\n}\n\nexport interface UnusedIndexReport {\n indexes: UnusedIndex[];\n totalWastedBytes: number;\n totalWasted: string; // human-readable\n checkedAt: string;\n}\n\nexport function formatBytes(bytes: number): string {\n if (bytes < 1024) return \"< 1 KB\";\n if (bytes < 1024 * 1024) return `${Math.round(bytes / 1024)} KB`;\n if (bytes < 1024 * 1024 * 1024) return `${(bytes / (1024 * 1024)).toFixed(1)} MB`;\n if (bytes < 1024 ** 4) return `${(bytes / 1024 ** 3).toFixed(1)} GB`;\n return `${(bytes / 1024 ** 4).toFixed(1)} TB`;\n}\n\nexport async function getUnusedIndexes(pool: Pool): Promise<UnusedIndexReport> {\n const [indexResult, bgwriterResult] = await Promise.all([\n pool.query(`\n SELECT\n s.schemaname,\n s.relname AS table_name,\n s.indexrelname AS index_name,\n pg_relation_size(s.indexrelid) AS index_size_bytes,\n s.idx_scan,\n i.indexdef\n FROM pg_stat_user_indexes s\n JOIN pg_indexes i ON s.schemaname = i.schemaname\n AND s.relname = i.tablename\n AND s.indexrelname = i.indexname\n WHERE s.schemaname = 'public'\n AND s.idx_scan = 0\n AND i.indexdef NOT LIKE '%UNIQUE%'\n AND s.indexrelname NOT LIKE '%_pkey'\n ORDER BY pg_relation_size(s.indexrelid) DESC\n `),\n pool.query(`SELECT stats_reset FROM pg_stat_bgwriter`),\n ]);\n\n const statsReset = bgwriterResult.rows[0]?.stats_reset\n ? new Date(bgwriterResult.rows[0].stats_reset).toISOString()\n : null;\n\n const filteredRows = indexResult.rows.filter((row: any) => {\n const def: string = row.indexdef ?? \"\";\n // Exclude partial indexes (have WHERE clause)\n if (def.includes(\" WHERE \")) return false;\n // Exclude expression indexes: column list has nested parens e.g. lower(email)\n // Extract the column portion between the first ( and last )\n const colStart = def.indexOf(\"(\");\n const colEnd = def.lastIndexOf(\")\");\n if (colStart !== -1 && colEnd !== -1) {\n const cols = def.slice(colStart + 1, colEnd);\n if (cols.includes(\"(\")) return false; // expression index\n }\n return true;\n });\n\n const indexes: UnusedIndex[] = filteredRows.map((row: any) => {\n const sizeBytes = parseInt(row.index_size_bytes, 10) || 0;\n const index = row.index_name as string;\n const table = row.table_name as string;\n return {\n schema: row.schemaname as string,\n table,\n index,\n indexSize: formatBytes(sizeBytes),\n indexSizeBytes: sizeBytes,\n scans: parseInt(row.idx_scan, 10) || 0,\n lastUsed: statsReset,\n suggestion: `Index ${index} on ${table} has never been used (0 scans). Consider dropping it: DROP INDEX CONCURRENTLY \"${index.replace(/\"/g, '\"\"')}\"`,\n };\n });\n\n const totalWastedBytes = indexes.reduce((sum, idx) => sum + idx.indexSizeBytes, 0);\n\n return {\n indexes,\n totalWastedBytes,\n totalWasted: formatBytes(totalWastedBytes),\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface TableBloat {\n schema: string;\n table: string;\n liveRows: number;\n deadRows: number;\n bloatPercent: number; // dead / (live + dead) * 100, rounded 1dp\n lastAutoVacuum: string | null;\n lastVacuum: string | null;\n suggestion: string;\n}\n\nexport interface BloatReport {\n tables: TableBloat[]; // sorted by bloatPercent DESC, only tables with bloatPercent >= 10\n checkedAt: string;\n}\n\nfunction getSuggestion(table: string, bloatPercent: number): string {\n if (bloatPercent >= 50) {\n return `HIGH bloat on ${table} (${bloatPercent}% dead rows). Run: VACUUM ANALYZE ${table}`;\n } else if (bloatPercent >= 20) {\n return `Moderate bloat on ${table} (${bloatPercent}% dead rows). Consider VACUUM ANALYZE ${table}`;\n } else {\n return `Minor bloat on ${table} (${bloatPercent}% dead rows). Autovacuum should handle this.`;\n }\n}\n\nexport async function getBloatReport(pool: Pool): Promise<BloatReport> {\n const result = await pool.query(`\n SELECT\n schemaname,\n relname AS table_name,\n n_live_tup,\n n_dead_tup,\n last_autovacuum,\n last_vacuum\n FROM pg_stat_user_tables\n WHERE schemaname = 'public'\n AND (n_live_tup + n_dead_tup) > 0\n ORDER BY (n_dead_tup::float / (n_live_tup + n_dead_tup)) DESC\n `);\n\n const tables: TableBloat[] = [];\n\n for (const row of result.rows) {\n const live = parseInt(row.n_live_tup, 10) || 0;\n const dead = parseInt(row.n_dead_tup, 10) || 0;\n const total = live + dead;\n if (total === 0) continue;\n\n const bloatPercent = Math.round((dead / total) * 1000) / 10; // 1dp\n if (bloatPercent < 10) continue;\n\n const table = row.table_name as string;\n tables.push({\n schema: row.schemaname as string,\n table,\n liveRows: live,\n deadRows: dead,\n bloatPercent,\n lastAutoVacuum: row.last_autovacuum ? new Date(row.last_autovacuum).toISOString() : null,\n lastVacuum: row.last_vacuum ? new Date(row.last_vacuum).toISOString() : null,\n suggestion: getSuggestion(table, bloatPercent),\n });\n }\n\n // Sort by bloatPercent DESC (DB query orders by dead ratio, but re-sort after filtering)\n tables.sort((a, b) => b.bloatPercent - a.bloatPercent);\n\n return {\n tables,\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface AutovacuumTableStatus {\n schema: string;\n table: string;\n lastAutoVacuum: string | null;\n lastAutoAnalyze: string | null;\n deadTuples: number;\n liveTuples: number;\n vacuumCount: number;\n analyzeCount: number;\n status: \"ok\" | \"stale\" | \"never\" | \"overdue\";\n suggestion: string | null;\n}\n\nexport interface AutovacuumReport {\n tables: AutovacuumTableStatus[];\n settings: {\n autovacuumEnabled: boolean;\n vacuumCostDelay: string;\n autovacuumMaxWorkers: number;\n autovacuumNaptime: string;\n };\n checkedAt: string;\n}\n\nfunction classifyStatus(\n lastAutoVacuum: Date | null,\n deadTuples: number,\n vacuumCount: number\n): AutovacuumTableStatus[\"status\"] {\n if (lastAutoVacuum === null) return \"never\"; // covers all null cases\n\n const daysSince = (Date.now() - lastAutoVacuum.getTime()) / (1000 * 60 * 60 * 24);\n\n if (daysSince > 7 && deadTuples > 10_000) return \"overdue\";\n if (daysSince > 3) return \"stale\";\n return \"ok\";\n}\n\nfunction getSuggestion(status: \"ok\" | \"stale\" | \"never\" | \"overdue\", table: string): string | null {\n switch (status) {\n case \"never\":\n return `Table ${table} has never been autovacuumed. Check if autovacuum is enabled and the table has enough churn.`;\n case \"overdue\":\n return `Table ${table} is overdue for vacuum and has many dead tuples. Run: VACUUM ANALYZE ${table}`;\n case \"stale\":\n return `Table ${table} hasn't been vacuumed in over 3 days. Monitor for bloat.`;\n case \"ok\":\n return null;\n }\n}\n\nexport async function getAutovacuumReport(pool: Pool): Promise<AutovacuumReport> {\n const [tableResult, settingsResult] = await Promise.all([\n pool.query(`\n SELECT\n schemaname, relname,\n last_autovacuum, last_autoanalyze,\n n_dead_tup, n_live_tup,\n autovacuum_count, autoanalyze_count\n FROM pg_stat_user_tables\n WHERE schemaname = 'public'\n ORDER BY n_dead_tup DESC\n `),\n pool.query(`\n SELECT name, setting\n FROM pg_settings\n WHERE name IN ('autovacuum', 'autovacuum_vacuum_cost_delay', 'autovacuum_max_workers', 'autovacuum_naptime')\n `),\n ]);\n\n const tables: AutovacuumTableStatus[] = tableResult.rows.map((row: any) => {\n const lastAutoVacuumDate = row.last_autovacuum ? new Date(row.last_autovacuum) : null;\n const deadTuples = parseInt(row.n_dead_tup, 10) || 0;\n const liveTuples = parseInt(row.n_live_tup, 10) || 0;\n const vacuumCount = parseInt(row.autovacuum_count, 10) || 0;\n const analyzeCount = parseInt(row.autoanalyze_count, 10) || 0;\n const status = classifyStatus(lastAutoVacuumDate, deadTuples, vacuumCount);\n const table = row.relname as string;\n\n return {\n schema: row.schemaname as string,\n table,\n lastAutoVacuum: lastAutoVacuumDate ? lastAutoVacuumDate.toISOString() : null,\n lastAutoAnalyze: row.last_autoanalyze ? new Date(row.last_autoanalyze).toISOString() : null,\n deadTuples,\n liveTuples,\n vacuumCount,\n analyzeCount,\n status,\n suggestion: getSuggestion(status, table),\n };\n });\n\n const settingsMap = new Map<string, string>();\n for (const row of settingsResult.rows) {\n settingsMap.set(row.name, row.setting);\n }\n\n return {\n tables,\n settings: {\n autovacuumEnabled: settingsMap.get(\"autovacuum\") !== \"off\",\n vacuumCostDelay: `${settingsMap.get(\"autovacuum_vacuum_cost_delay\") ?? \"2\"}ms`,\n autovacuumMaxWorkers: parseInt(settingsMap.get(\"autovacuum_max_workers\") ?? \"3\", 10),\n autovacuumNaptime: `${settingsMap.get(\"autovacuum_naptime\") ?? \"60\"}s`,\n },\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface LockWait {\n blockedPid: number;\n blockedQuery: string;\n blockedDuration: string; // e.g. \"00:00:45\"\n blockingPid: number;\n blockingQuery: string;\n blockingDuration: string;\n table: string | null;\n lockType: string;\n}\n\nexport interface LockReport {\n waitingLocks: LockWait[];\n longRunningQueries: Array<{\n pid: number;\n duration: string;\n query: string;\n state: string;\n waitEventType: string | null;\n }>;\n checkedAt: string;\n}\n\nexport function formatDurationSecs(secs: number): string {\n const h = Math.floor(secs / 3600);\n const m = Math.floor((secs % 3600) / 60);\n const s = secs % 60;\n return [\n String(h).padStart(2, \"0\"),\n String(m).padStart(2, \"0\"),\n String(s).padStart(2, \"0\"),\n ].join(\":\");\n}\n\nexport async function getLockReport(pool: Pool): Promise<LockReport> {\n const [locksResult, longResult] = await Promise.all([\n pool.query(`\n SELECT\n blocked.pid AS blocked_pid,\n blocked.query AS blocked_query,\n EXTRACT(EPOCH FROM (NOW() - blocked.query_start))::int AS blocked_secs,\n blocking.pid AS blocking_pid,\n blocking.query AS blocking_query,\n EXTRACT(EPOCH FROM (NOW() - blocking.query_start))::int AS blocking_secs,\n blocked_locks.relation::regclass::text AS table_name,\n blocked_locks.locktype\n FROM pg_catalog.pg_locks blocked_locks\n JOIN pg_catalog.pg_stat_activity blocked ON blocked.pid = blocked_locks.pid\n JOIN pg_catalog.pg_locks blocking_locks\n ON blocking_locks.locktype = blocked_locks.locktype\n AND blocking_locks.relation IS NOT DISTINCT FROM blocked_locks.relation\n AND blocking_locks.pid != blocked_locks.pid\n AND blocking_locks.granted = true\n JOIN pg_catalog.pg_stat_activity blocking ON blocking.pid = blocking_locks.pid\n WHERE NOT blocked_locks.granted\n `),\n pool.query(`\n SELECT\n pid,\n EXTRACT(EPOCH FROM (NOW() - query_start))::int AS duration_secs,\n query,\n state,\n wait_event_type\n FROM pg_stat_activity\n WHERE state != 'idle'\n AND query_start IS NOT NULL\n AND EXTRACT(EPOCH FROM (NOW() - query_start)) > 5\n AND query NOT LIKE '%pg_stat_activity%'\n ORDER BY duration_secs DESC\n LIMIT 20\n `),\n ]);\n\n // Deduplicate by (blockedPid, blockingPid) — same pair may appear multiple times\n // for different lock types; keep only the first occurrence.\n const seen = new Set<string>();\n const waitingLocks: LockWait[] = [];\n for (const row of locksResult.rows) {\n const key = `${row.blocked_pid}:${row.blocking_pid}`;\n if (!seen.has(key)) {\n seen.add(key);\n waitingLocks.push({\n blockedPid: parseInt(row.blocked_pid, 10),\n blockedQuery: row.blocked_query as string,\n blockedDuration: formatDurationSecs(parseInt(row.blocked_secs, 10) || 0),\n blockingPid: parseInt(row.blocking_pid, 10),\n blockingQuery: row.blocking_query as string,\n blockingDuration: formatDurationSecs(parseInt(row.blocking_secs, 10) || 0),\n table: row.table_name ?? null,\n lockType: row.locktype as string,\n });\n }\n }\n\n const longRunningQueries = longResult.rows.map((row: any) => ({\n pid: parseInt(row.pid, 10),\n duration: formatDurationSecs(parseInt(row.duration_secs, 10) || 0),\n query: row.query as string,\n state: row.state as string,\n waitEventType: row.wait_event_type ?? null,\n }));\n\n return {\n waitingLocks,\n longRunningQueries,\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface ConfigRecommendation {\n setting: string;\n currentValue: string;\n recommendedValue: string;\n reason: string;\n severity: \"error\" | \"warning\" | \"info\";\n docs?: string;\n}\n\nexport interface ConfigReport {\n recommendations: ConfigRecommendation[];\n serverInfo: {\n maxConnections: number;\n sharedBuffers: string;\n workMem: string;\n effectiveCacheSize: string;\n maintenanceWorkMem: string;\n walBuffers: string;\n checkpointCompletionTarget: string;\n randomPageCost: string;\n autovacuumVacuumScaleFactor: string;\n };\n checkedAt: string;\n}\n\n// Convert a setting value + unit to bytes for comparison\n// In pg_settings, 'setting' is in the displayed 'unit'\nfunction settingToBytes(value: string, unit: string | undefined): number {\n const v = parseFloat(value);\n if (!unit) return v;\n switch (unit.toLowerCase()) {\n case \"b\": return v;\n case \"kb\": return v * 1024;\n case \"8kb\": return v * 8 * 1024; // shared_buffers, effective_cache_size\n case \"mb\": return v * 1024 * 1024;\n case \"gb\": return v * 1024 * 1024 * 1024;\n default: return v;\n }\n}\n\nfunction settingToMb(value: string, unit: string | undefined): number {\n return settingToBytes(value, unit) / (1024 * 1024);\n}\n\n// Format a memory setting to a human-readable string with units\nfunction formatMemSetting(rawValue: string | null | undefined, unit?: string): string {\n if (!rawValue) return \"unknown\";\n const bytes = settingToBytes(rawValue, unit ?? \"\");\n if (bytes <= 0 || isNaN(bytes)) return rawValue; // fallback for special values like -1 (auto)\n if (bytes >= 1024 ** 3) return `${(bytes / 1024 ** 3).toFixed(1)}GB`;\n if (bytes >= 1024 ** 2) return `${Math.round(bytes / 1024 ** 2)}MB`;\n if (bytes >= 1024) return `${Math.round(bytes / 1024)}KB`;\n return `${bytes}B`;\n}\n\nexport async function getConfigReport(pool: Pool): Promise<ConfigReport> {\n const result = await pool.query(`\n SELECT name, setting, unit\n FROM pg_settings\n WHERE name IN (\n 'max_connections', 'shared_buffers', 'work_mem',\n 'effective_cache_size', 'maintenance_work_mem', 'wal_buffers',\n 'checkpoint_completion_target', 'random_page_cost',\n 'autovacuum_vacuum_scale_factor', 'autovacuum_analyze_scale_factor',\n 'log_min_duration_statement', 'idle_in_transaction_session_timeout',\n 'effective_io_concurrency'\n )\n `);\n\n const settings: Record<string, { setting: string; unit: string | undefined }> = {};\n for (const row of result.rows) {\n settings[row.name] = { setting: row.setting, unit: row.unit ?? undefined };\n }\n\n const recommendations: ConfigRecommendation[] = [];\n\n const get = (name: string) => settings[name]?.setting ?? null;\n const getUnit = (name: string) => settings[name]?.unit;\n\n // 1. shared_buffers: < 128MB → warning\n const sharedBuffersSetting = get(\"shared_buffers\");\n if (sharedBuffersSetting !== null) {\n const mb = settingToMb(sharedBuffersSetting, getUnit(\"shared_buffers\"));\n if (mb < 128) {\n recommendations.push({\n setting: \"shared_buffers\",\n currentValue: `${Math.round(mb)}MB`,\n recommendedValue: \"256MB\",\n reason: \"shared_buffers should be at least 25% of RAM; typical starting point is 256MB–1GB\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-SHARED-BUFFERS\",\n });\n }\n }\n\n // 2. work_mem: <= 4MB → info\n const workMemSetting = get(\"work_mem\");\n if (workMemSetting !== null) {\n const mb = settingToMb(workMemSetting, getUnit(\"work_mem\"));\n if (mb <= 4) {\n recommendations.push({\n setting: \"work_mem\",\n currentValue: `${mb % 1 === 0 ? mb : mb.toFixed(1)}MB`,\n recommendedValue: \"16MB\",\n reason: \"work_mem of 4MB is conservative; consider 16MB–64MB for analytical queries (but multiply by max_connections for total)\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-WORK-MEM\",\n });\n }\n }\n\n // 3. checkpoint_completion_target: < 0.9 → warning\n const cctSetting = get(\"checkpoint_completion_target\");\n if (cctSetting !== null) {\n const v = parseFloat(cctSetting);\n if (v < 0.9) {\n recommendations.push({\n setting: \"checkpoint_completion_target\",\n currentValue: cctSetting,\n recommendedValue: \"0.9\",\n reason: \"Set to 0.9 to spread checkpoint I/O over 90% of checkpoint interval\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-wal.html#GUC-CHECKPOINT-COMPLETION-TARGET\",\n });\n }\n }\n\n // 4. random_page_cost: > 2.0 → info\n const rpcSetting = get(\"random_page_cost\");\n if (rpcSetting !== null) {\n const v = parseFloat(rpcSetting);\n if (v > 2.0) {\n recommendations.push({\n setting: \"random_page_cost\",\n currentValue: rpcSetting,\n recommendedValue: \"1.1\",\n reason: \"If using SSDs, set random_page_cost=1.1 (default 4.0 is tuned for spinning disks)\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-query.html#GUC-RANDOM-PAGE-COST\",\n });\n }\n }\n\n // 5. autovacuum_vacuum_scale_factor: >= 0.2 → info\n const avsfSetting = get(\"autovacuum_vacuum_scale_factor\");\n if (avsfSetting !== null) {\n const v = parseFloat(avsfSetting);\n if (v >= 0.2) {\n recommendations.push({\n setting: \"autovacuum_vacuum_scale_factor\",\n currentValue: avsfSetting,\n recommendedValue: \"0.05\",\n reason: \"Consider lowering to 0.05–0.1 for large tables to vacuum more frequently\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-autovacuum.html#GUC-AUTOVACUUM-VACUUM-SCALE-FACTOR\",\n });\n }\n }\n\n // 6. log_min_duration_statement: = -1 → info\n const lmdsSetting = get(\"log_min_duration_statement\");\n if (lmdsSetting !== null && parseInt(lmdsSetting, 10) === -1) {\n recommendations.push({\n setting: \"log_min_duration_statement\",\n currentValue: \"-1\",\n recommendedValue: \"1000\",\n reason: \"Consider setting to 1000 (log queries > 1s) for performance monitoring\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-logging.html#GUC-LOG-MIN-DURATION-STATEMENT\",\n });\n }\n\n // 7. idle_in_transaction_session_timeout: = 0 → warning\n const iitsSetting = get(\"idle_in_transaction_session_timeout\");\n if (iitsSetting !== null && parseInt(iitsSetting, 10) === 0) {\n recommendations.push({\n setting: \"idle_in_transaction_session_timeout\",\n currentValue: \"0\",\n recommendedValue: \"60000\",\n reason: \"Set idle_in_transaction_session_timeout=60000 (60s) to prevent stuck transactions from holding locks\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-client.html#GUC-IDLE-IN-TRANSACTION-SESSION-TIMEOUT\",\n });\n }\n\n // 8. effective_io_concurrency: = 1 → info\n const eicSetting = get(\"effective_io_concurrency\");\n if (eicSetting !== null && parseInt(eicSetting, 10) === 1) {\n recommendations.push({\n setting: \"effective_io_concurrency\",\n currentValue: \"1\",\n recommendedValue: \"200\",\n reason: \"If using SSDs, set effective_io_concurrency=200 for better parallel I/O\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-EFFECTIVE-IO-CONCURRENCY\",\n });\n }\n\n // 9. wal_buffers: skip if -1 (auto)\n\n // 10. maintenance_work_mem: <= 64MB → info\n const mwmSetting = get(\"maintenance_work_mem\");\n if (mwmSetting !== null) {\n const mb = settingToMb(mwmSetting, getUnit(\"maintenance_work_mem\"));\n if (mb <= 64) {\n recommendations.push({\n setting: \"maintenance_work_mem\",\n currentValue: `${mb % 1 === 0 ? mb : mb.toFixed(1)}MB`,\n recommendedValue: \"256MB\",\n reason: \"Consider 256MB for faster VACUUM and index builds\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-MAINTENANCE-WORK-MEM\",\n });\n }\n }\n\n const maxConnSetting = get(\"max_connections\");\n\n // max_connections > 200 without connection pooler is a common perf trap\n if (maxConnSetting !== null) {\n const maxConn = parseInt(maxConnSetting, 10);\n if (maxConn > 200) {\n recommendations.push({\n setting: \"max_connections\",\n currentValue: String(maxConn),\n recommendedValue: \"100\",\n reason: `max_connections=${maxConn} is high. Each connection uses ~5–10MB RAM. Without a connection pooler (PgBouncer), this leads to memory pressure and context-switch overhead. Consider lowering to 100 and using a pooler.`,\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-connection.html#GUC-MAX-CONNECTIONS\",\n });\n }\n }\n\n const serverInfo = {\n maxConnections: maxConnSetting !== null ? parseInt(maxConnSetting, 10) : 0,\n sharedBuffers: formatMemSetting(sharedBuffersSetting, getUnit(\"shared_buffers\")),\n workMem: formatMemSetting(workMemSetting, getUnit(\"work_mem\")),\n effectiveCacheSize: formatMemSetting(get(\"effective_cache_size\"), getUnit(\"effective_cache_size\")),\n maintenanceWorkMem: formatMemSetting(mwmSetting, getUnit(\"maintenance_work_mem\")),\n walBuffers: get(\"wal_buffers\") ?? \"\",\n checkpointCompletionTarget: cctSetting ?? \"\",\n randomPageCost: rpcSetting ?? \"\",\n autovacuumVacuumScaleFactor: avsfSetting ?? \"\",\n };\n\n return {\n recommendations,\n serverInfo,\n checkedAt: new Date().toISOString(),\n };\n}\n"],"mappings":";;;AAEA,SAAS,iBAAiB;AAC1B,SAAS,4BAA4B;AACrC,SAAS,QAAAA,aAAY;AACrB,SAAS,SAAS;;;ACHlB,eAAsB,YAAYC,OAAY;AAC5C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,UAAU,MAAM,OAAO,MAAM,qBAAqB;AACxD,UAAM,SAAS,MAAM,OAAO;AAAA,MAC1B;AAAA,IACF;AACA,UAAM,SAAS,MAAM,OAAO;AAAA,MAC1B;AAAA,IACF;AACA,UAAM,UAAU,MAAM,OAAO;AAAA,MAC3B;AAAA,IACF;AACA,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,KAKtC;AAED,WAAO;AAAA,MACL,SAAS,QAAQ,KAAK,CAAC,EAAE;AAAA,MACzB,QAAQ,OAAO,KAAK,CAAC,EAAE;AAAA,MACvB,QAAQ,OAAO,KAAK,CAAC,EAAE;AAAA,MACvB,eAAe,QAAQ,KAAK,CAAC,EAAE;AAAA,MAC/B,aAAa,YAAY,KAAK,CAAC;AAAA,IACjC;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;AC9BA,eAAsB,UAAUC,OAAY;AAC1C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAa5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACrBA,eAAsB,gBAAgBC,OAAY;AAChD,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAe5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AAEA,eAAsB,qBAAqBA,OAAY,WAAmB;AACxE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AAEF,UAAM,QAAQ,UAAU,MAAM,GAAG;AACjC,UAAM,SAAS,MAAM,SAAS,IAAI,MAAM,CAAC,IAAI;AAC7C,UAAM,OAAO,MAAM,SAAS,IAAI,MAAM,CAAC,IAAI,MAAM,CAAC;AAGlD,UAAM,YAAY,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAclC,CAAC,MAAM,MAAM,CAAC;AAEjB,QAAI,UAAU,KAAK,WAAW,EAAG,QAAO;AAGxC,UAAM,UAAU,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAYhC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,UAAU,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAiBhC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAUpC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAWpC,CAAC,MAAM,MAAM,CAAC;AAGjB,QAAI,aAAoB,CAAC;AACzB,QAAI;AACF,YAAM,SAAS,MAAM,OAAO;AAAA,QAC1B,iBAAiB,OAAO,iBAAiB,MAAM,CAAC,IAAI,OAAO,iBAAiB,IAAI,CAAC;AAAA,MACnF;AACA,mBAAa,OAAO;AAAA,IACtB,SAAS,KAAK;AAAE,cAAQ,MAAM,mBAAoB,IAAc,OAAO;AAAA,IAAG;AAE1E,WAAO;AAAA,MACL,GAAG,UAAU,KAAK,CAAC;AAAA,MACnB,SAAS,QAAQ;AAAA,MACjB,SAAS,QAAQ;AAAA,MACjB,aAAa,YAAY;AAAA,MACzB,aAAa,YAAY;AAAA,MACzB;AAAA,IACF;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AAwEA,eAAsB,eAAeC,OAAY;AAC/C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAW5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACrNA,eAAsB,YAAYC,OAAiC;AACjE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAuB5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACsrBA,OAAO,cAAc;AACrB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,OAAO,QAAQ;AAxsBf,IAAM,kBAAkB,EAAE,UAAU,IAAI,SAAS,GAAG,MAAM,EAAE;AAC5D,IAAM,gBAAgB,EAAE,UAAU,IAAI,SAAS,IAAI,MAAM,GAAG;AAErD,SAAS,oBAAoB,QAAgC;AAClE,MAAI,QAAQ;AACZ,QAAM,aAAa,EAAE,UAAU,GAAG,SAAS,GAAG,MAAM,EAAE;AACtD,QAAM,SAAS,EAAE,UAAU,GAAG,SAAS,GAAG,MAAM,EAAE;AAClD,aAAW,SAAS,QAAQ;AAC1B,WAAO,MAAM,QAAQ;AACrB,UAAM,IAAI,OAAO,MAAM,QAAQ;AAC/B,UAAM,SAAS,gBAAgB,MAAM,QAAQ;AAE7C,QAAI;AACJ,QAAI,KAAK,EAAG,WAAU;AAAA,aACb,KAAK,GAAI,WAAU,SAAS;AAAA,QAChC,WAAU,SAAS;AACxB,eAAW,MAAM,QAAQ,KAAK;AAAA,EAChC;AAEA,aAAW,OAAO,CAAC,YAAY,WAAW,MAAM,GAAY;AAC1D,aAAS,KAAK,IAAI,WAAW,GAAG,GAAG,cAAc,GAAG,CAAC;AAAA,EACvD;AACA,SAAO,KAAK,IAAI,GAAG,KAAK,IAAI,KAAK,KAAK,MAAM,KAAK,CAAC,CAAC;AACrD;AAEO,SAAS,eAAe,OAAuB;AACpD,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,SAAO;AACT;AAEA,SAAS,iBAAiB,QAAyF;AACjH,QAAM,aAAa,CAAC,eAAe,eAAe,UAAU,UAAU;AACtE,QAAM,SAA0E,CAAC;AACjF,aAAW,OAAO,YAAY;AAC5B,UAAM,YAAY,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,GAAG;AACzD,UAAM,QAAQ,oBAAoB,SAAS;AAC3C,WAAO,GAAG,IAAI,EAAE,OAAO,OAAO,eAAe,KAAK,GAAG,OAAO,UAAU,OAAO;AAAA,EAC/E;AACA,SAAO;AACT;AAEA,eAAsB,iBAAiBC,OAAYC,sBAAqB,GAA2B;AACjG,QAAM,SAAS,MAAMD,MAAK,QAAQ;AAClC,QAAM,SAAyB,CAAC;AAChC,QAAM,UAAoB,CAAC;AAE3B,MAAI;AAEF,UAAM,gBAAgB,MAAM,OAAO,MAAM,yBAAyB;AAClE,UAAM,YAAY,SAAS,cAAc,KAAK,CAAC,EAAE,kBAAkB;AAKnE,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAM5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,iBAAiB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAClD,UAAU,IAAI,WAAW,MAAO,YAAY;AAAA,UAC5C,UAAU;AAAA,UACV,OAAO,4BAA4B,IAAI,OAAO;AAAA,UAC9C,aAAa,SAAS,IAAI,UAAU,IAAI,IAAI,OAAO,KAAK,IAAI,UAAU,UAAU,IAAI,IAAI,SAAS,IAAI,QAAQ,6BAA6B,OAAO,IAAI,YAAY,EAAE,eAAe,CAAC;AAAA,UACnL,KAAK;AAAA,mCAA4F,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,gCAA2D,IAAI,OAAO,gBAAgB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjP,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,uCAAwC,IAAc,OAAO;AAAG,cAAQ,KAAK,gBAAiB,IAAc,OAAO;AAAA,IACnI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAW5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,oBAAoB,IAAI,YAAY;AAAA,UACxC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,iBAAiB,IAAI,YAAY;AAAA,UACxC,aAAa,SAAS,IAAI,YAAY,OAAO,IAAI,OAAO,OAAO,IAAI,eAAe,0BAA0B,IAAI,eAAe;AAAA,UAC/H,KAAK,8BAA8B,IAAI,UAAU,IAAI,IAAI,YAAY;AAAA,UACrE,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,MAAM,WAAW,IAAI,QAAQ;AACnC,eAAO,KAAK;AAAA,UACV,IAAI,cAAc,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC/C,UAAU,MAAM,KAAK,aAAa;AAAA,UAClC,UAAU;AAAA,UACV,OAAO,kBAAkB,IAAI,OAAO,KAAK,IAAI,QAAQ;AAAA,UACrD,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO,QAAQ,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC,iBAAiB,IAAI,QAAQ,QAAQ,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC,sBAAsB,IAAI,IAAI;AAAA,UACtM,KAAK,eAAe,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjD,QAAQ;AAAA,UACR,QAAQ,MAAM,KAAK,aAAa;AAAA,QAClC,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,yCAA0C,IAAc,OAAO;AAAG,cAAQ,KAAK,kBAAmB,IAAc,OAAO;AAAA,IACvI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,QAAQ,WAAW,IAAI,KAAK;AAClC,YAAI,QAAQ,KAAK;AACf,iBAAO,KAAK;AAAA,YACV,IAAI,cAAc,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,YAC/C,UAAU,QAAQ,MAAM,aAAa;AAAA,YACrC,UAAU;AAAA,YACV,OAAO,2BAA2B,IAAI,OAAO;AAAA,YAC7C,aAAa,SAAS,IAAI,UAAU,IAAI,IAAI,OAAO,8BAA8B,QAAQ,KAAK,QAAQ,CAAC,CAAC;AAAA,YACxG,KAAK;AAAA;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,8CAA+C,IAAc,OAAO;AAAG,cAAQ,KAAK,uBAAwB,IAAc,OAAO;AAAA,IACjJ;AAGA,QAAI;AACF,YAAM,WAAW,MAAM,OAAO,MAAM,iEAAiE;AACrG,UAAI,SAAS,KAAK,SAAS,GAAG;AAC5B,cAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,SAQ5B;AACD,mBAAW,OAAO,EAAE,MAAM;AACxB,iBAAO,KAAK;AAAA,YACV,IAAI,aAAa,IAAI,MAAM,MAAM,GAAG,EAAE,EAAE,QAAQ,OAAO,GAAG,CAAC;AAAA,YAC3D,UAAU,WAAW,IAAI,OAAO,IAAI,MAAO,YAAY;AAAA,YACvD,UAAU;AAAA,YACV,OAAO,mBAAmB,IAAI,OAAO;AAAA,YACrC,aAAa,mBAAmB,IAAI,OAAO,WAAW,IAAI,KAAK,kBAAkB,IAAI,SAAS,OAAO,IAAI,MAAM,MAAM,GAAG,GAAG,CAAC;AAAA,YAC5H,KAAK,mBAAmB,IAAI,MAAM,MAAM,GAAG,GAAG,CAAC;AAAA,YAC/C,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,0CAA2C,IAAc,OAAO;AAAG,cAAQ,KAAK,mBAAoB,IAAc,OAAO;AAAA,IACzI;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,QAAQ,CAAC,IAAI,eAAe,CAAC,IAAI;AACvC,eAAO,KAAK;AAAA,UACV,IAAI,gBAAgB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjD,UAAU,QAAQ,YAAY;AAAA,UAC9B,UAAU;AAAA,UACV,OAAO,UAAU,QAAQ,cAAc,SAAS,OAAO,IAAI,OAAO;AAAA,UAClE,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO,IAAI,QAAQ,4BAA4B,mCAAmC,kBAAkB,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC;AAAA,UACjL,KAAK,kBAAkB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACpD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAW5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,iBAAiB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAClD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,wBAAwB,IAAI,OAAO;AAAA,UAC1C,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC7C,KAAK,WAAW,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC7C,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,OAI5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,MAAM,SAAS,IAAI,OAAO;AAChC,YAAI,MAAM,KAAe;AACvB,iBAAO,KAAK;AAAA,YACV,IAAI;AAAA,YACJ,UAAU;AAAA,YACV,UAAU;AAAA,YACV,OAAO;AAAA,YACP,aAAa,YAAY,IAAI,OAAO,4BAA4B,IAAI,eAAe,CAAC;AAAA,YACpF,KAAK;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH,WAAW,MAAM,KAAa;AAC5B,iBAAO,KAAK;AAAA,YACV,IAAI;AAAA,YACJ,UAAU;AAAA,YACV,UAAU;AAAA,YACV,OAAO;AAAA,YACP,aAAa,YAAY,IAAI,OAAO,4BAA4B,IAAI,eAAe,CAAC;AAAA,YACpF,KAAK;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,SAQ1B,CAACC,mBAAkB,CAAC;AACvB,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,WAAW,IAAI,UAAU;AAC/B,eAAO,KAAK;AAAA,UACV,IAAI,cAAc,IAAI,GAAG;AAAA,UACzB,UAAU,WAAW,YAAY;AAAA,UACjC,UAAU;AAAA,UACV,OAAO,GAAG,WAAW,wBAAwB,iBAAiB,SAAS,IAAI,GAAG;AAAA,UAC9E,aAAa,OAAO,IAAI,GAAG,SAAS,IAAI,eAAe,OAAO,KAAK,IAAI,oBAAoB,SAAS,cAAc,IAAI,KAAK,QAAQ,KAAK,MAAM,IAAI,eAAe,EAAE,CAAC;AAAA,UACpK,KAAK,+BAA+B,IAAI,GAAG;AAAA,UAC3C,QAAQ,WAAW,mEAAmE;AAAA,UACtF,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,8CAA+C,IAAc,OAAO;AAAG,cAAQ,KAAK,uBAAwB,IAAc,OAAO;AAAA,IACjJ;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,gBAAgB,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAChD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,0BAA0B,IAAI,UAAU;AAAA,UAC/C,aAAa,SAAS,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAClD,KAAK,eAAe,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAChD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,kDAAmD,IAAc,OAAO;AAAG,cAAQ,KAAK,2BAA4B,IAAc,OAAO;AAAA,IACzJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAS5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,qBAAqB,IAAI,YAAY;AAAA,UACzC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,gBAAgB,IAAI,YAAY,KAAK,IAAI,QAAQ;AAAA,UACxD,aAAa,SAAS,IAAI,YAAY,OAAO,IAAI,OAAO,4CAA4C,IAAI,QAAQ;AAAA,UAChH,KAAK,2BAA2B,IAAI,UAAU,IAAI,IAAI,YAAY;AAAA,UAClE,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,kBAAkB,IAAI,UAAU,IAAI,IAAI,QAAQ,CAAC,CAAC;AAAA,UACtD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,wBAAwB,IAAI,UAAU;AAAA,UAC7C,aAAa,2CAA2C,IAAI,UAAU,KAAK,IAAI,QAAQ,KAAK,IAAI,CAAC,yBAAyB,IAAI,UAAU;AAAA,UACxI,KAAK;AAAA,0BAAwD,IAAI,QAAQ,MAAM,CAAC,EAAE,KAAK,6BAA6B,CAAC;AAAA,UACrH,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,+CAAgD,IAAc,OAAO;AAAG,cAAQ,KAAK,wBAAyB,IAAc,OAAO;AAAA,IACnJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAa5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,oBAAoB,IAAI,UAAU,IAAI,IAAI,WAAW;AAAA,UACzD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,8BAA8B,IAAI,UAAU,IAAI,IAAI,WAAW;AAAA,UACtE,aAAa,sBAAsB,IAAI,WAAW,OAAO,IAAI,UAAU,gBAAgB,IAAI,gBAAgB;AAAA,UAC3G,KAAK,iCAAiC,IAAI,WAAW,QAAQ,OAAO,GAAG,CAAC,IAAI,IAAI,WAAW,OAAO,IAAI,UAAU,KAAK,IAAI,WAAW;AAAA,UACpI,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,gDAAiD,IAAc,OAAO;AAAG,cAAQ,KAAK,yBAA0B,IAAc,OAAO;AAAA,IACrJ;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAkB5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,qBAAqB,IAAI,WAAW;AAAA,UACxC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,sBAAsB,IAAI,WAAW,mBAAmB,IAAI,YAAY;AAAA,UAC/E,aAAa,OAAO,IAAI,WAAW,sCAAsC,IAAI,YAAY,aAAa,IAAI,iBAAiB,IAAI,MAAM,GAAG,GAAG,CAAC;AAAA,UAC5I,KAAK,4BAA4B,IAAI,YAAY;AAAA,UACjD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,mCAAoC,IAAc,OAAO;AAAG,cAAQ,KAAK,YAAa,IAAc,OAAO;AAAA,IAC3H;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,OAI5B;AACD,YAAM,WAAW,SAAS,EAAE,KAAK,CAAC,GAAG,aAAa,GAAG;AACrD,UAAI,WAAW,SAAS;AACtB,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU,WAAW,YAAY,aAAa;AAAA,UAC9C,UAAU;AAAA,UACV,OAAO,qBAAqB,WAAW,SAAS,QAAQ,CAAC,CAAC;AAAA,UAC1D,aAAa,6BAA6B,WAAW,SAAS,QAAQ,CAAC,CAAC;AAAA,UACxE,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,iBAAiB,aAAa,OAAS,yBAAyB;AACtE,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,eAIpB,cAAc;AAAA,OACtB;AACD,YAAM,SAAS,WAAW,EAAE,KAAK,CAAC,GAAG,WAAW,GAAG;AACnD,UAAI,SAAS,IAAI;AACf,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU,SAAS,KAAK,YAAY;AAAA,UACpC,UAAU;AAAA,UACV,OAAO,GAAG,MAAM;AAAA,UAChB,aAAa,GAAG,EAAE,KAAK,CAAC,GAAG,eAAe,iBAAiB,EAAE,KAAK,CAAC,GAAG,iBAAiB;AAAA,UACvF,KAAK;AAAA;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,kDAAmD,IAAc,OAAO;AAAG,cAAQ,KAAK,2BAA4B,IAAc,OAAO;AAAA,IACzJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,2DAA2D;AACxF,UAAI,EAAE,KAAK,CAAC,GAAG,YAAY,OAAO;AAChC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO;AAAA,UACP,aAAa;AAAA,UACb,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,wCAAyC,IAAc,OAAO;AAAG,cAAQ,KAAK,iBAAkB,IAAc,OAAO;AAAA,IACrI;AAGA,QAAI;AACF,YAAM,QAAQ,MAAM,OAAO,MAAM,qEAAqE;AACtG,YAAM,SAAS,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA,OAGjC;AACD,YAAM,cAAc,SAAS,OAAO,KAAK,CAAC,GAAG,gBAAgB,GAAG;AAGhE,UAAI,cAAc,KAAK,cAAc,MAAM,OAAO,MAAM;AACtD,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,2BAA2B,cAAc,SAAS,QAAQ,CAAC,CAAC;AAAA,UACnE,aAAa,4BAA4B,MAAM,KAAK,CAAC,GAAG,OAAO,GAAG,MAAM,KAAK,CAAC,GAAG,QAAQ,EAAE;AAAA,UAC3F,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAEA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,+DAA+D;AAC5F,YAAM,YAAY,SAAS,EAAE,KAAK,CAAC,GAAG,WAAW,GAAG;AACpD,UAAI,YAAY,KAAK,YAAY,MAAM;AACrC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,oBAAoB,YAAY,OAAO,YAAY,QAAQ,YAAY,MAAM,QAAQ,CAAC,IAAI,IAAI;AAAA,UACrG,aAAa,eAAe,EAAE,KAAK,CAAC,GAAG,OAAO,GAAG,EAAE,KAAK,CAAC,GAAG,QAAQ,EAAE;AAAA,UACtE,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,sCAAuC,IAAc,OAAO;AAAG,cAAQ,KAAK,eAAgB,IAAc,OAAO;AAAA,IACjI;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,wBAAwB,IAAI,GAAG;AAAA,UACnC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,aAAa,IAAI,OAAO,mBAAmB,IAAI,WAAW;AAAA,UACjE,aAAa,aAAa,IAAI,OAAO,wDAAwD,IAAI,WAAW;AAAA,UAC5G,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,mDAAoD,IAAc,OAAO;AAAG,cAAQ,KAAK,4BAA6B,IAAc,OAAO;AAAA,IAC3J;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,oDAAoD;AACjF,UAAI,EAAE,KAAK,CAAC,GAAG,YAAY,OAAO;AAChC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO;AAAA,UACP,aAAa;AAAA,UACb,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,uCAAwC,IAAc,OAAO;AAAG,cAAQ,KAAK,gBAAiB,IAAc,OAAO;AAAA,IACnI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,kBAAkB,IAAI,QAAQ,IAAI,IAAI,SAAS;AAAA,UACnD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,4BAA4B,IAAI,SAAS,IAAI,IAAI,QAAQ;AAAA,UAChE,aAAa,0DAA0D,IAAI,IAAI,mBAAmB,IAAI,QAAQ,OAAO,IAAI,SAAS;AAAA,UAClI,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,wCAAyC,IAAc,OAAO;AAAG,cAAQ,KAAK,iBAAkB,IAAc,OAAO;AAAA,IACrI;AAGA,UAAM,aAAa,iBAAiB;AACpC,UAAM,aAAa,IAAI,IAAI,UAAU;AACrC,UAAM,eAAe,OAAO,OAAO,OAAK,CAAC,WAAW,IAAI,EAAE,EAAE,CAAC;AAC7D,UAAM,eAAe,OAAO,SAAS,aAAa;AAGlD,UAAM,aAAyB,CAAC;AAChC,UAAM,SAAS,oBAAI,IAA4B;AAC/C,eAAW,SAAS,cAAc;AAEhC,YAAM,SAAS,MAAM,GAAG,QAAQ,WAAW,EAAE;AAC7C,UAAI,CAAC,OAAO,IAAI,MAAM,EAAG,QAAO,IAAI,QAAQ,CAAC,CAAC;AAC9C,aAAO,IAAI,MAAM,EAAG,KAAK,KAAK;AAAA,IAChC;AACA,UAAM,eAAuC;AAAA,MAC3C,oBAAoB;AAAA,MACpB,qBAAqB;AAAA,MACrB,gBAAgB;AAAA,MAChB,gBAAgB;AAAA,MAChB,iBAAiB;AAAA,MACjB,oBAAoB;AAAA,MACpB,cAAc;AAAA,IAChB;AACA,eAAW,CAAC,QAAQ,KAAK,KAAK,QAAQ;AACpC,UAAI,MAAM,UAAU,EAAG;AACvB,YAAM,QAAQ,aAAa,MAAM,KAAK,WAAW,MAAM,MAAM,IAAI,MAAM;AACvE,YAAM,MAAM,MAAM,IAAI,OAAK,EAAE,IAAI,MAAM,IAAI,EAAE,OAAO,OAAK,CAAC,EAAE,KAAK,EAAE,WAAW,IAAI,CAAC,EAAE,KAAK,IAAI,EAAE,KAAK,CAAC,EAAE,OAAO,OAAO,EAAE,KAAK,KAAK,IAAI;AACtI,iBAAW,KAAK,EAAE,MAAM,QAAQ,OAAO,GAAG,KAAK,KAAK,MAAM,MAAM,KAAK,OAAO,MAAM,QAAQ,IAAI,CAAC;AAAA,IACjG;AAEA,UAAM,QAAQ,oBAAoB,YAAY;AAC9C,WAAO;AAAA,MACL;AAAA,MACA,OAAO,eAAe,KAAK;AAAA,MAC3B,QAAQ;AAAA,MACR,WAAW,iBAAiB,YAAY;AAAA,MACxC;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AASA,IAAI,aAAiD;AAErD,SAAS,eAA4C;AACnD,MAAI,WAAY,QAAO;AACvB,QAAMC,WAAU,QAAQ,IAAI,oBAAoB,KAAK,KAAK,GAAG,QAAQ,GAAG,UAAU;AAClF,KAAG,UAAUA,UAAS,EAAE,WAAW,KAAK,CAAC;AACzC,QAAM,SAAS,KAAK,KAAKA,UAAS,WAAW;AAC7C,eAAa,IAAI,SAAS,MAAM;AAChC,aAAW,OAAO,oBAAoB;AACtC,aAAW,KAAK,2FAA2F;AAC3G,SAAO;AACT;AAEO,SAAS,mBAA6B;AAC3C,MAAI;AACF,UAAM,KAAK,aAAa;AACxB,WAAO,GAAG,QAAQ,qCAAqC,EAAE,IAAI,EAAE,IAAI,CAAC,MAAW,EAAE,QAAQ;AAAA,EAC3F,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAcO,SAAS,UAAU,KAAsB;AAC9C,QAAM,UAAU,IAAI,KAAK;AACzB,MAAI,CAAC,QAAS,QAAO;AAGrB,QAAM,aAAa,QAAQ,QAAQ,SAAS,EAAE,EAAE,MAAM,GAAG,EAAE,IAAI,OAAK,EAAE,KAAK,CAAC,EAAE,OAAO,OAAO;AAC5F,MAAI,WAAW,WAAW,EAAG,QAAO;AAEpC,QAAM,QAAQ,WAAW,CAAC,EAAE,YAAY;AAGxC,MAAI,MAAM,WAAW,iBAAiB,GAAG;AACvC,UAAM,eAAe,MAAM,QAAQ,yBAAyB,EAAE,EAAE,UAAU;AAC1E,WAAO,aAAa,WAAW,QAAQ;AAAA,EACzC;AAGA,QAAM,mBAAmB;AAAA,IACvB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAEA,SAAO,iBAAiB,KAAK,CAAC,MAAM,MAAM,WAAW,CAAC,CAAC;AACzD;;;ACzxBA,eAAsB,eAAeC,OAAkC;AACrE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AAEF,UAAM,WAAW,MAAM,OAAO;AAAA,MAC5B;AAAA,IACF;AACA,QAAI,SAAS,KAAK,WAAW,GAAG;AAC9B,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAe5B;AACD,WAAO,EAAE;AAAA,EACX,QAAQ;AAEN,WAAO,CAAC;AAAA,EACV,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;AC/CA,OAAOC,SAAQ;AACf,OAAOC,WAAU;AAwBjB,SAAS,iBAAiB,IAAoB;AAC5C,SAAO,GAAG,QAAQ,SAAS,EAAE;AAC/B;AAQO,SAAS,aAAa,cAAsB,QAA6B;AAC9E,EAAAD,IAAG,UAAUC,MAAK,QAAQ,YAAY,GAAG,EAAE,WAAW,KAAK,CAAC;AAC5D,QAAM,WAAqB,EAAE,YAAW,oBAAI,KAAK,GAAE,YAAY,GAAG,OAAO;AACzE,EAAAD,IAAG,cAAc,cAAc,KAAK,UAAU,UAAU,MAAM,CAAC,CAAC;AAClE;AAQO,SAAS,aAAa,cAAuC;AAClE,MAAI,CAACA,IAAG,WAAW,YAAY,EAAG,QAAO;AACzC,MAAI;AACF,WAAO,KAAK,MAAMA,IAAG,aAAa,cAAc,OAAO,CAAC;AAAA,EAC1D,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAEO,SAAS,cAAc,MAAqB,SAAsC;AAGvF,QAAM,cAAc,IAAI,IAAI,KAAK,OAAO,IAAI,CAAC,MAAM,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAC1E,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAE7E,QAAM,YAAY,QAAQ,OAAO,OAAO,CAAC,MAAM,CAAC,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AACvF,QAAM,iBAAiB,KAAK,OAAO,OAAO,CAAC,MAAM,CAAC,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AACzF,QAAM,YAAY,QAAQ,OAAO,OAAO,CAAC,MAAM,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAEtF,SAAO;AAAA,IACL,YAAY,QAAQ,QAAQ,KAAK;AAAA,IACjC,eAAe,KAAK;AAAA,IACpB,cAAc,QAAQ;AAAA,IACtB,eAAe,KAAK;AAAA,IACpB,cAAc,QAAQ;AAAA,IACtB;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5EA,SAAS,YAAY;;;ACgDd,SAAS,oBAAoB,SAAyB,SAAyC;AACpG,QAAM,UAA0B,CAAC;AAEjC,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACnF,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AAGnF,aAAW,CAAC,KAAK,CAAC,KAAK,aAAa;AAClC,QAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,GAAG,SAAS,CAAC;AAAA,IAC5G;AAAA,EACF;AACA,aAAW,CAAC,GAAG,KAAK,aAAa;AAC/B,QAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,GAAG,WAAW,CAAC;AAAA,IAChH;AAAA,EACF;AAGA,aAAW,CAAC,KAAK,QAAQ,KAAK,aAAa;AACzC,UAAM,WAAW,YAAY,IAAI,GAAG;AACpC,QAAI,CAAC,SAAU;AAGf,UAAM,UAAU,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAChE,UAAM,UAAU,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAEhE,eAAW,CAAC,MAAM,GAAG,KAAK,SAAS;AACjC,YAAM,SAAS,QAAQ,IAAI,IAAI;AAC/B,UAAI,CAAC,QAAQ;AACX,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,WAAW,IAAI,IAAI,IAAI,CAAC;AAAA,MAC7H,OAAO;AACL,YAAI,OAAO,SAAS,IAAI,MAAM;AAC5B,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,kBAAkB,OAAO,IAAI,WAAM,IAAI,IAAI,GAAG,CAAC;AAAA,QACvJ;AACA,YAAI,OAAO,aAAa,IAAI,UAAU;AACpC,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,sBAAsB,OAAO,QAAQ,WAAM,IAAI,QAAQ,GAAG,CAAC;AAAA,QACnK;AACA,YAAI,OAAO,kBAAkB,IAAI,eAAe;AAC9C,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,qBAAqB,OAAO,iBAAiB,MAAM,WAAM,IAAI,iBAAiB,MAAM,GAAG,CAAC;AAAA,QAChM;AAAA,MACF;AAAA,IACF;AACA,eAAW,QAAQ,QAAQ,KAAK,GAAG;AACjC,UAAI,CAAC,QAAQ,IAAI,IAAI,GAAG;AACtB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,WAAW,CAAC;AAAA,MACnH;AAAA,IACF;AAGA,UAAM,SAAS,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAC/D,UAAM,SAAS,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAC/D,eAAW,CAAC,MAAM,GAAG,KAAK,QAAQ;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,SAAS,CAAC;AAAA,MAC7G,WAAW,OAAO,IAAI,IAAI,EAAG,eAAe,IAAI,YAAY;AAC1D,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,sBAAsB,CAAC;AAAA,MAC7H;AAAA,IACF;AACA,eAAW,QAAQ,OAAO,KAAK,GAAG;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,WAAW,CAAC;AAAA,MACjH;AAAA,IACF;AAGA,UAAM,SAAS,IAAI,IAAI,SAAS,YAAY,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AACnE,UAAM,SAAS,IAAI,IAAI,SAAS,YAAY,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AACnE,eAAW,CAAC,MAAM,GAAG,KAAK,QAAQ;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,WAAW,IAAI,IAAI,IAAI,CAAC;AAAA,MACrI,WAAW,OAAO,IAAI,IAAI,EAAG,eAAe,IAAI,YAAY;AAC1D,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,sBAAsB,CAAC;AAAA,MACvI;AAAA,IACF;AACA,eAAW,QAAQ,OAAO,KAAK,GAAG;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,WAAW,CAAC;AAAA,MAC3H;AAAA,IACF;AAAA,EACF;AAGA,QAAM,WAAW,IAAI,KAAK,QAAQ,SAAS,CAAC,GAAG,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACvF,QAAM,WAAW,IAAI,KAAK,QAAQ,SAAS,CAAC,GAAG,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACvF,aAAW,CAAC,KAAK,EAAE,KAAK,UAAU;AAChC,UAAM,QAAQ,SAAS,IAAI,GAAG;AAC9B,QAAI,CAAC,OAAO;AACV,cAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,WAAW,GAAG,OAAO,KAAK,IAAI,CAAC,IAAI,CAAC;AAAA,IACrI,OAAO;AACL,YAAM,QAAQ,GAAG,OAAO,OAAO,CAAC,MAAM,CAAC,MAAM,OAAO,SAAS,CAAC,CAAC;AAC/D,YAAM,UAAU,MAAM,OAAO,OAAO,CAAC,MAAM,CAAC,GAAG,OAAO,SAAS,CAAC,CAAC;AACjE,iBAAW,KAAK,OAAO;AACrB,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,YAAY,CAAC,UAAU,CAAC;AAAA,MAC5H;AACA,iBAAW,KAAK,SAAS;AACvB,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,YAAY,CAAC,YAAY,CAAC;AAAA,MAC9H;AAAA,IACF;AAAA,EACF;AACA,aAAW,OAAO,SAAS,KAAK,GAAG;AACjC,QAAI,CAAC,SAAS,IAAI,GAAG,GAAG;AACtB,cAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,WAAW,CAAC;AAAA,IAC/G;AAAA,EACF;AAEA,SAAO;AACT;;;ACjJA,eAAsB,kBAAkBE,OAAqC;AAC3E,QAAM,SAAS,MAAM,gBAAgBA,KAAI;AACzC,QAAM,QAAQ,MAAM,eAAeA,KAAI;AAEvC,QAAM,iBAAiB,MAAM,QAAQ;AAAA,IACnC,OAAO,IAAI,OAAO,MAAW;AAC3B,YAAM,SAAS,MAAM,qBAAqBA,OAAM,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,EAAE;AACvE,UAAI,CAAC,OAAQ,QAAO;AACpB,aAAO;AAAA,QACL,MAAM,OAAO;AAAA,QACb,QAAQ,OAAO;AAAA,QACf,SAAS,OAAO,QAAQ,IAAI,CAAC,OAAY;AAAA,UACvC,MAAM,EAAE;AAAA,UACR,MAAM,EAAE;AAAA,UACR,UAAU,EAAE;AAAA,UACZ,eAAe,EAAE;AAAA,QACnB,EAAE;AAAA,QACF,SAAS,OAAO,QAAQ,IAAI,CAAC,OAAY;AAAA,UACvC,MAAM,EAAE;AAAA,UACR,YAAY,EAAE;AAAA,UACd,WAAW,EAAE;AAAA,UACb,YAAY,EAAE;AAAA,QAChB,EAAE;AAAA,QACF,aAAa,OAAO,YAAY,IAAI,CAAC,OAAY;AAAA,UAC/C,MAAM,EAAE;AAAA,UACR,MAAM,EAAE;AAAA,UACR,YAAY,EAAE;AAAA,QAChB,EAAE;AAAA,MACJ;AAAA,IACF,CAAC;AAAA,EACH;AAEA,SAAO;AAAA,IACL,QAAQ,eAAe,OAAO,OAAO;AAAA,IACrC,OAAO,MAAM,IAAI,CAAC,OAAY,EAAE,MAAM,EAAE,MAAM,QAAQ,EAAE,QAAQ,QAAQ,EAAE,OAAO,EAAE;AAAA,EACrF;AACF;;;AF+DA,eAAe,YAAYC,OAA+B;AACxD,QAAM,MAAM,MAAMA,MAAK,MAA8B;AAAA;AAAA;AAAA;AAAA;AAAA,GAKpD;AACD,SAAO,IAAI,KAAK,IAAI,CAAC,MAAM,EAAE,UAAU;AACzC;AAEA,eAAe,aAAaA,OAAkC;AAC5D,QAAM,MAAM,MAAMA,MAAK,MAAiB;AAAA;AAAA;AAAA;AAAA;AAAA,GAKvC;AACD,SAAO,IAAI;AACb;AAEA,eAAe,aAAaA,OAAiC;AAC3D,QAAM,MAAM,MAAMA,MAAK,MAAgB;AAAA;AAAA;AAAA;AAAA;AAAA,GAKtC;AACD,SAAO,IAAI;AACb;AAIA,SAAS,WAAW,cAAwB,cAA4E;AACtH,QAAM,YAAY,IAAI,IAAI,YAAY;AACtC,QAAM,YAAY,IAAI,IAAI,YAAY;AACtC,SAAO;AAAA,IACL,eAAe,aAAa,OAAO,CAAC,MAAM,CAAC,UAAU,IAAI,CAAC,CAAC;AAAA,IAC3D,aAAa,aAAa,OAAO,CAAC,MAAM,CAAC,UAAU,IAAI,CAAC,CAAC;AAAA,EAC3D;AACF;AAEA,SAAS,oBAAoB,SAA4D;AACvF,QAAM,MAAM,oBAAI,IAAqC;AACrD,aAAW,OAAO,SAAS;AACzB,QAAI,CAAC,IAAI,IAAI,IAAI,UAAU,EAAG,KAAI,IAAI,IAAI,YAAY,oBAAI,IAAI,CAAC;AAC/D,UAAM,OAAmB;AAAA,MACvB,MAAM,IAAI;AAAA,MACV,MAAM,IAAI;AAAA,MACV,UAAU,IAAI,gBAAgB;AAAA,IAChC;AACA,QAAI,IAAI,mBAAmB,QAAQ,IAAI,mBAAmB,QAAW;AACnE,WAAK,UAAU,IAAI;AAAA,IACrB;AACA,QAAI,IAAI,IAAI,UAAU,EAAG,IAAI,IAAI,aAAa,IAAI;AAAA,EACpD;AACA,SAAO;AACT;AAEA,SAAS,YACP,YACA,YACA,cACc;AACd,QAAM,gBAAgB,oBAAoB,UAAU;AACpD,QAAM,gBAAgB,oBAAoB,UAAU;AACpD,QAAM,QAAsB,CAAC;AAE7B,aAAW,SAAS,cAAc;AAChC,UAAM,SAAS,cAAc,IAAI,KAAK,KAAK,oBAAI,IAAwB;AACvE,UAAM,SAAS,cAAc,IAAI,KAAK,KAAK,oBAAI,IAAwB;AAEvE,UAAM,iBAA+B,CAAC;AACtC,UAAM,eAA6B,CAAC;AACpC,UAAM,YAA8B,CAAC;AACrC,UAAM,gBAAsC,CAAC;AAC7C,UAAM,eAAoC,CAAC;AAE3C,eAAW,CAAC,SAAS,OAAO,KAAK,QAAQ;AACvC,UAAI,CAAC,OAAO,IAAI,OAAO,GAAG;AACxB,uBAAe,KAAK,OAAO;AAAA,MAC7B,OAAO;AACL,cAAM,UAAU,OAAO,IAAI,OAAO;AAClC,YAAI,QAAQ,SAAS,QAAQ,MAAM;AACjC,oBAAU,KAAK,EAAE,QAAQ,SAAS,YAAY,QAAQ,MAAM,YAAY,QAAQ,KAAK,CAAC;AAAA,QACxF;AACA,YAAI,QAAQ,aAAa,QAAQ,UAAU;AACzC,wBAAc,KAAK,EAAE,QAAQ,SAAS,gBAAgB,QAAQ,UAAU,gBAAgB,QAAQ,SAAS,CAAC;AAAA,QAC5G;AACA,aAAK,QAAQ,WAAW,WAAW,QAAQ,WAAW,OAAO;AAC3D,uBAAa,KAAK,EAAE,QAAQ,SAAS,eAAe,QAAQ,WAAW,MAAM,eAAe,QAAQ,WAAW,KAAK,CAAC;AAAA,QACvH;AAAA,MACF;AAAA,IACF;AAEA,eAAW,CAAC,SAAS,OAAO,KAAK,QAAQ;AACvC,UAAI,CAAC,OAAO,IAAI,OAAO,GAAG;AACxB,qBAAa,KAAK,OAAO;AAAA,MAC3B;AAAA,IACF;AAEA,QAAI,eAAe,SAAS,KAAK,aAAa,SAAS,KAAK,UAAU,SAAS,KAC3E,cAAc,SAAS,KAAK,aAAa,SAAS,GAAG;AACvD,YAAM,KAAK,EAAE,OAAO,gBAAgB,cAAc,WAAW,eAAe,aAAa,CAAC;AAAA,IAC5F;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,oBAAoB,SAAuD;AAClF,QAAM,MAAM,oBAAI,IAAiC;AACjD,aAAW,OAAO,SAAS;AACzB,QAAI,CAAC,IAAI,IAAI,IAAI,SAAS,EAAG,KAAI,IAAI,IAAI,WAAW,oBAAI,IAAI,CAAC;AAC7D,QAAI,IAAI,IAAI,SAAS,EAAG,IAAI,IAAI,WAAW,IAAI,QAAQ;AAAA,EACzD;AACA,SAAO;AACT;AAEA,SAAS,YACP,YACA,YACA,cACa;AACb,QAAM,aAAa,oBAAoB,UAAU;AACjD,QAAM,aAAa,oBAAoB,UAAU;AACjD,QAAM,QAAqB,CAAC;AAG5B,QAAM,YAAY,oBAAI,IAAI;AAAA,IACxB,GAAG,WAAW,IAAI,CAAC,MAAM,EAAE,SAAS;AAAA,IACpC,GAAG,WAAW,IAAI,CAAC,MAAM,EAAE,SAAS;AAAA,EACtC,CAAC;AAED,aAAW,SAAS,WAAW;AAE7B,QAAI,CAAC,aAAa,SAAS,KAAK,EAAG;AAEnC,UAAM,SAAS,WAAW,IAAI,KAAK,KAAK,oBAAI,IAAoB;AAChE,UAAM,SAAS,WAAW,IAAI,KAAK,KAAK,oBAAI,IAAoB;AAEhE,UAAM,iBAAiB,CAAC,GAAG,OAAO,KAAK,CAAC,EAAE,OAAO,CAAC,MAAM,CAAC,OAAO,IAAI,CAAC,CAAC;AACtE,UAAM,eAAe,CAAC,GAAG,OAAO,KAAK,CAAC,EAAE,OAAO,CAAC,MAAM,CAAC,OAAO,IAAI,CAAC,CAAC;AACpE,UAAM,kBAAkC,CAAC;AAEzC,eAAW,CAAC,MAAM,MAAM,KAAK,QAAQ;AACnC,UAAI,OAAO,IAAI,IAAI,GAAG;AACpB,cAAM,SAAS,OAAO,IAAI,IAAI;AAC9B,YAAI,WAAW,QAAQ;AACrB,0BAAgB,KAAK,EAAE,MAAM,WAAW,QAAQ,WAAW,OAAO,CAAC;AAAA,QACrE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,eAAe,SAAS,KAAK,aAAa,SAAS,KAAK,gBAAgB,SAAS,GAAG;AACtF,YAAM,KAAK,EAAE,OAAO,gBAAgB,cAAc,gBAAgB,CAAC;AAAA,IACrE;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,kBAAkB,QAA4B;AACrD,MAAI,IAAI,OAAO,cAAc,SAAS,OAAO,YAAY;AACzD,aAAW,MAAM,OAAO,aAAa;AACnC,SAAK,GAAG,eAAe,SAAS,GAAG,aAAa,SAAS,GAAG,UAAU,SACjE,GAAG,cAAc,SAAS,GAAG,aAAa;AAAA,EACjD;AACA,aAAW,MAAM,OAAO,YAAY;AAClC,SAAK,GAAG,eAAe,SAAS,GAAG,aAAa,SAAS,GAAG,gBAAgB;AAAA,EAC9E;AACA,QAAM,OAAO,mBAAmB,CAAC,GAAG;AACpC,QAAM,OAAO,aAAa,CAAC,GAAG;AAC9B,SAAO;AACT;AAIA,eAAsB,iBACpB,YACA,YACA,SACwB;AACxB,QAAM,aAAa,IAAI,KAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AAC5F,QAAM,aAAa,IAAI,KAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AAE5F,MAAI;AAEF,UAAM;AAAA,MACJ;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,IAAI,MAAM,QAAQ,IAAI;AAAA,MACpB,YAAY,UAAU;AAAA,MACtB,YAAY,UAAU;AAAA,MACtB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,kBAAkB,UAAU,EAAE,MAAM,MAAM,IAAI;AAAA,MAC9C,kBAAkB,UAAU,EAAE,MAAM,MAAM,IAAI;AAAA,IAChD,CAAC;AAED,UAAM,EAAE,eAAe,YAAY,IAAI,WAAW,cAAc,YAAY;AAC5E,UAAM,YAAY,IAAI,IAAI,YAAY;AACtC,UAAM,eAAe,aAAa,OAAO,CAAC,MAAM,UAAU,IAAI,CAAC,CAAC;AAEhE,UAAM,cAAc,YAAY,YAAY,YAAY,YAAY;AACpE,UAAM,aAAa,YAAY,YAAY,YAAY,YAAY;AAGnE,UAAM,kBAAoC,CAAC;AAC3C,UAAM,YAAwB,CAAC;AAE/B,QAAI,cAAc,YAAY;AAI5B,YAAM,cAAc,oBAAoB,YAAY,UAAU;AAE9D,iBAAW,KAAK,aAAa;AAC3B,YAAI,EAAE,gBAAgB,cAAc;AAClC,0BAAgB,KAAK;AAAA,YACnB,OAAO,EAAE,cAAc;AAAA,YACvB,MAAM,EAAE,gBAAgB,UAAU,UAAU,EAAE,gBAAgB,YAAY,YAAY;AAAA,YACtF,MAAM,EAAE,OAAO,MAAM,GAAG,EAAE,CAAC,KAAK,EAAE;AAAA,YAClC,QAAQ,EAAE;AAAA,UACZ,CAAC;AAAA,QACH,WAAW,EAAE,gBAAgB,QAAQ;AACnC,oBAAU,KAAK;AAAA,YACb,MAAM,EAAE,gBAAgB,UAAU,UAAU,EAAE,gBAAgB,YAAY,YAAY;AAAA,YACtF,MAAM,EAAE,OAAO,MAAM,GAAG,EAAE,CAAC,KAAK,EAAE;AAAA,YAClC,QAAQ,EAAE;AAAA,UACZ,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF;AAEA,UAAM,SAAqB,EAAE,eAAe,aAAa,aAAa,YAAY,iBAAiB,UAAU;AAC7G,UAAM,eAAe,kBAAkB,MAAM;AAE7C,QAAI;AAEJ,QAAI,SAAS,eAAe;AAC1B,YAAMC,sBAAqB;AAC3B,YAAM,CAAC,WAAW,SAAS,IAAI,MAAM,QAAQ,IAAI;AAAA,QAC/C,iBAAiB,YAAYA,mBAAkB;AAAA,QAC/C,iBAAiB,YAAYA,mBAAkB;AAAA,MACjD,CAAC;AAED,YAAM,eAAe,IAAI,IAAI,UAAU,OAAO,IAAI,CAAC,MAAM,EAAE,KAAK,CAAC;AACjE,YAAM,eAAe,IAAI,IAAI,UAAU,OAAO,IAAI,CAAC,MAAM,EAAE,KAAK,CAAC;AAEjE,YAAM,mBAAmB,UAAU,OAChC,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,KAAK,CAAC,EACxC,IAAI,CAAC,MAAM,GAAG,EAAE,QAAQ,KAAK,EAAE,KAAK,EAAE;AAEzC,YAAM,mBAAmB,UAAU,OAChC,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,KAAK,CAAC,EACxC,IAAI,CAAC,MAAM,GAAG,EAAE,QAAQ,KAAK,EAAE,KAAK,EAAE;AAEzC,eAAS;AAAA,QACP,QAAQ,EAAE,OAAO,UAAU,OAAO,OAAO,UAAU,OAAO,KAAK,qBAAqB,UAAU,EAAE;AAAA,QAChG,QAAQ,EAAE,OAAO,UAAU,OAAO,OAAO,UAAU,OAAO,KAAK,qBAAqB,UAAU,EAAE;AAAA,QAChG;AAAA,QACA;AAAA,MACF;AAAA,IACF;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,MAClC,SAAS;AAAA,QACP;AAAA,QACA,WAAW,iBAAiB;AAAA,MAC9B;AAAA,IACF;AAAA,EACF,UAAE;AACA,UAAM,QAAQ,WAAW,CAAC,WAAW,IAAI,GAAG,WAAW,IAAI,CAAC,CAAC;AAAA,EAC/D;AACF;AAGA,SAAS,qBAAqB,SAAyB;AACrD,MAAI;AACF,UAAM,MAAM,IAAI,IAAI,OAAO;AAC3B,QAAI,IAAI,SAAU,KAAI,WAAW;AACjC,WAAO,IAAI,SAAS;AAAA,EACtB,QAAQ;AACN,WAAO;AAAA,EACT;AACF;;;AG7VA,SAAS,aAAa,MAAW,MAAa,CAAC,GAAU;AACvD,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,MAAI,KAAK,IAAI;AACb,QAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,OAAO;AAC3C,MAAI,MAAM,QAAQ,KAAK,GAAG;AACxB,eAAW,SAAS,MAAO,cAAa,OAAO,GAAG;AAAA,EACpD;AACA,SAAO;AACT;AAWA,SAAS,yBAAyB,QAA0B;AAE1D,QAAM,aAAa;AACnB,QAAM,QAAQ,oBAAI,IAAY;AAC9B,MAAI;AACJ,UAAQ,IAAI,WAAW,KAAK,MAAM,OAAO,MAAM;AAC7C,UAAM,MAAM,EAAE,CAAC,EAAE,YAAY;AAE7B,QAAI,CAAC,CAAC,OAAO,MAAM,OAAO,QAAQ,SAAS,MAAM,EAAE,SAAS,GAAG,GAAG;AAChE,YAAM,IAAI,GAAG;AAAA,IACf;AAAA,EACF;AACA,SAAO,MAAM,KAAK,KAAK;AACzB;AAMA,eAAe,wBAAwBC,OAAY,WAAwC;AACzF,MAAI;AAEF,UAAM,IAAI,MAAMA,MAAK;AAAA,MACnB;AAAA,MACA,CAAC,SAAS;AAAA,IACZ;AACA,WAAO,EAAE,KAAK,IAAI,CAAC,QAAa;AAE9B,YAAM,IAAI,cAAc,KAAK,IAAI,QAAQ;AACzC,UAAI,CAAC,EAAG,QAAO,CAAC;AAChB,aAAO,EAAE,CAAC,EACP,MAAM,GAAG,EACT,IAAI,CAAC,MAAc,EAAE,KAAK,EAAE,QAAQ,UAAU,EAAE,EAAE,YAAY,CAAC;AAAA,IACpE,CAAC;AAAA,EACH,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAKA,SAAS,YAAY,UAA6C;AAChE,MAAI,WAAW,IAAS,QAAO;AAC/B,MAAI,YAAY,IAAQ,QAAO;AAC/B,SAAO;AACT;AAKA,SAAS,QAAQ,GAAmB;AAClC,MAAI,KAAK,IAAW,QAAO,IAAI,IAAI,KAAW,QAAQ,CAAC,CAAC;AACxD,MAAI,KAAK,IAAO,QAAO,IAAI,IAAI,KAAO,QAAQ,CAAC,CAAC;AAChD,SAAO,OAAO,CAAC;AACjB;AAUA,eAAsB,mBACpB,aACAA,OAC0B;AAC1B,QAAM,SAA0B;AAAA,IAC9B,WAAW,CAAC;AAAA,IACZ,UAAU,CAAC;AAAA,IACX,gBAAgB,CAAC;AAAA,IACjB,cAAc,EAAE,WAAW,EAAE;AAAA,IAC7B,iBAAiB,CAAC;AAAA,EACpB;AAEA,MAAI,CAAC,eAAe,CAAC,MAAM,QAAQ,WAAW,KAAK,YAAY,WAAW,GAAG;AAC3E,WAAO;AAAA,EACT;AAEA,QAAM,WAAW,YAAY,CAAC;AAC9B,QAAM,WAAW,WAAW,MAAM,KAAK,WAAW,MAAM;AAGxD,QAAM,eAAmC,WAAW,eAAe,KAAK;AACxE,QAAM,gBAAoC,WAAW,gBAAgB,KAAK;AAE1E,MAAI,CAAC,SAAU,QAAO;AAGtB,QAAM,WAAW,aAAa,QAAQ;AAGtC,SAAO,YAAY,SAAS,IAAI,CAAC,MAAW;AAC1C,UAAM,IAAqB;AAAA,MACzB,UAAU,EAAE,WAAW,KAAK;AAAA,MAC5B,WAAW,EAAE,YAAY,KAAK;AAAA,IAChC;AACA,QAAI,EAAE,eAAe,EAAG,GAAE,QAAQ,EAAE,eAAe;AACnD,QAAI,EAAE,aAAa,MAAM,OAAW,GAAE,aAAa,EAAE,aAAa;AAClE,QAAI,EAAE,mBAAmB,MAAM,OAAW,GAAE,aAAa,EAAE,mBAAmB;AAC9E,QAAI,EAAE,QAAQ,EAAG,GAAE,SAAS,EAAE,QAAQ;AACtC,WAAO;AAAA,EACT,CAAC;AAGD,SAAO,eAAe;AAAA,IACpB,WAAW,SAAS,YAAY,KAAK;AAAA,IACrC,YAAY;AAAA,IACZ;AAAA,EACF;AAGA,QAAM,eAAe,SAAS,OAAO,CAAC,MAAW,EAAE,WAAW,MAAM,UAAU;AAE9E,aAAW,QAAQ,cAAc;AAC/B,UAAM,QAAgB,KAAK,eAAe,KAAK;AAC/C,UAAM,WAAmB,KAAK,WAAW,KAAK,KAAK,aAAa,KAAK;AACrE,UAAM,SAA6B,KAAK,QAAQ;AAEhD,UAAM,OAAoB,EAAE,OAAO,UAAU,OAAO;AAEpD,QAAI,WAAW,KAAQ;AACrB,WAAK,aAAa,SACd,qDAAqD,KAAK,KAC1D,kCAAkC,KAAK;AAAA,IAC7C;AAEA,WAAO,SAAS,KAAK,IAAI;AAAA,EAC3B;AAGA,aAAW,QAAQ,OAAO,UAAU;AAClC,QAAI,CAAC,KAAK,OAAQ;AAElB,UAAM,OAAO,yBAAyB,KAAK,MAAM;AACjD,QAAI,KAAK,WAAW,EAAG;AAGvB,QAAI,oBAAgC,CAAC;AACrC,QAAIA,OAAM;AACR,0BAAoB,MAAM,wBAAwBA,OAAM,KAAK,KAAK;AAAA,IACpE;AAGA,UAAM,gBAAgB,KAAK;AAAA,MACzB,CAAC,QAAQ,CAAC,kBAAkB,KAAK,CAAC,YAAY,QAAQ,SAAS,KAAK,QAAQ,CAAC,MAAM,GAAG;AAAA,IACxF;AAEA,QAAI,cAAc,WAAW,EAAG;AAEhC,UAAM,UAAU,YAAY,KAAK,QAAQ;AAEzC,QAAI,cAAc,UAAU,GAAG;AAE7B,YAAM,UAAU,OAAO,KAAK,KAAK,IAAI,cAAc,KAAK,GAAG,CAAC;AAC5D,YAAM,MAAM,6BAA6B,OAAO,OAAO,KAAK,KAAK,KAAK,cAAc,KAAK,IAAI,CAAC;AAC9F,aAAO,eAAe,KAAK;AAAA,QACzB,OAAO,KAAK;AAAA,QACZ,SAAS;AAAA,QACT,QAAQ,sCAAsC,cAAc,KAAK,IAAI,CAAC,QAAQ,QAAQ,KAAK,QAAQ,CAAC;AAAA,QACpG;AAAA,QACA,kBAAkB;AAAA,MACpB,CAAC;AAAA,IACH,OAAO;AAEL,YAAM,MAAM,cAAc,CAAC;AAC3B,YAAM,UAAU,OAAO,KAAK,KAAK,IAAI,GAAG;AACxC,YAAM,MAAM,6BAA6B,OAAO,OAAO,KAAK,KAAK,KAAK,GAAG;AACzE,aAAO,eAAe,KAAK;AAAA,QACzB,OAAO,KAAK;AAAA,QACZ,SAAS,CAAC,GAAG;AAAA,QACb,QAAQ,2BAA2B,GAAG,KAAK,QAAQ,KAAK,QAAQ,CAAC;AAAA,QACjE;AAAA,QACA,kBAAkB;AAAA,MACpB,CAAC;AAAA,IACH;AAAA,EACF;AAGA,aAAW,QAAQ,OAAO,UAAU;AAClC,QAAI,KAAK,WAAW,KAAQ;AAC1B,YAAM,aAAa,KAAK,SACpB,oCAA+B,yBAAyB,KAAK,MAAM,EAAE,KAAK,IAAI,KAAK,gBAAgB,KACnG;AACJ,aAAO,gBAAgB;AAAA,QACrB,eAAe,KAAK,KAAK,KAAK,QAAQ,KAAK,QAAQ,CAAC,SAAS,UAAU;AAAA,MACzE;AAAA,IACF;AAAA,EACF;AAEA,MAAI,iBAAiB,QAAW;AAC9B,UAAM,QAAQ,eAAe,KAAK,iCAA4B;AAC9D,WAAO,gBAAgB,KAAK,iBAAiB,aAAa,QAAQ,CAAC,CAAC,aAAQ,KAAK,EAAE;AAAA,EACrF;AAEA,MAAI,OAAO,eAAe,WAAW,KAAK,OAAO,SAAS,WAAW,GAAG;AACtE,WAAO,gBAAgB,KAAK,mEAA8D;AAAA,EAC5F;AAEA,SAAO;AACT;AAeA,eAAsB,uBACpBA,OACA,SACA,cAAc,IACkB;AAChC,MAAI;AAEF,UAAM,WAAW,MAAMA,MAAK;AAAA,MAC1B;AAAA,IACF;AACA,QAAI,SAAS,KAAK,WAAW,EAAG,QAAO,CAAC;AAGxC,UAAM,UAAU,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,KAKhC;AAED,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,OAAO,QAAQ,MAAM;AAC9B,iBAAW,IAAI,IAAI,SAAS,WAAW,IAAI,cAAc,CAAC;AAAA,IAC5D;AAEA,QAAI,CAAC,QAAS,QAAO,CAAC;AAGtB,UAAM,WAAW,cAAc,KAAK,KAAK;AACzC,UAAM,QAAQ,KAAK,IAAI,IAAI;AAE3B,QAAI;AACJ,QAAI;AACF,mBAAa,QACV;AAAA,QACC;AAAA;AAAA;AAAA;AAAA,MAIF,EACC,IAAI,KAAK;AAAA,IACd,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAGA,UAAM,cAAc,oBAAI,IAAmD;AAC3E,eAAW,OAAO,YAAY;AAC5B,UAAI,CAAC,YAAY,IAAI,IAAI,OAAO,GAAG;AACjC,oBAAY,IAAI,IAAI,SAAS;AAAA,UAC3B,QAAQ,IAAI;AAAA,UACZ,WAAW,IAAI;AAAA,QACjB,CAAC;AAAA,MACH;AAAA,IACF;AAGA,UAAM,cAAqC,CAAC;AAE5C,eAAW,CAAC,SAAS,QAAQ,KAAK,aAAa;AAC7C,YAAM,cAAc,WAAW,IAAI,OAAO;AAC1C,UAAI,gBAAgB,UAAa,SAAS,WAAW,EAAG;AAExD,YAAM,iBACF,cAAc,SAAS,UAAU,SAAS,SAAU;AAExD,UAAI,gBAAgB,IAAI;AACtB,oBAAY,KAAK;AAAA,UACf;AAAA,UACA,eAAe;AAAA,UACf,gBAAgB,SAAS;AAAA,UACzB,eAAe,KAAK,MAAM,aAAa;AAAA,UACvC,YAAY,IAAI,KAAK,SAAS,SAAS,EAAE,YAAY;AAAA,QACvD,CAAC;AAAA,MACH;AAAA,IACF;AAEA,WAAO,YAAY,KAAK,CAAC,GAAG,MAAM,EAAE,gBAAgB,EAAE,aAAa;AAAA,EACrE,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;;;AC1VA,SAAS,cAAc,KAAqB;AAE1C,MAAI,WAAW,IAAI;AAAA,IAAQ;AAAA,IAAqB,CAAC,UAC/C,MAAM,QAAQ,UAAU,GAAG;AAAA,EAC7B;AAEA,aAAW,SAAS,QAAQ,aAAa,CAAC,UAAU,IAAI,OAAO,MAAM,MAAM,CAAC;AAC5E,SAAO;AACT;AAGA,SAAS,eAAe,KAAa,YAA4B;AAC/D,QAAM,SAAS,IAAI,MAAM,GAAG,UAAU;AACtC,SAAO,OAAO,MAAM,IAAI,EAAE;AAC5B;AAGA,SAAS,UAAU,MAAsB;AACvC,SAAO,KACJ,QAAQ,cAAc,EAAE,EACxB,QAAQ,MAAM,EAAE,EAChB,YAAY,EACZ,KAAK;AACV;AAGA,SAAS,sBAAsB,KAK7B;AACA,QAAM,cAAc,GAAG;AACvB,QAAM,cAAwB,CAAC;AAC/B,QAAM,cAAwB,CAAC;AAC/B,QAAM,aAAuB,CAAC;AAC9B,QAAM,YAAsB,CAAC;AAG7B,QAAM,QAAQ;AACd,MAAI;AACJ,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,aAAY,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,aAAY,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,SAAS;AACf,UAAQ,IAAI,OAAO,KAAK,GAAG,OAAO,KAAM,YAAW,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,WAAU,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAErE,SAAO,EAAE,aAAa,aAAa,YAAY,UAAU;AAC3D;AAGA,SAAS,YAAY,KAA+B;AAClD,QAAM,SAA2B,CAAC;AAElC,QAAM,cAAc,GAAG;AAGvB,QAAM,kBAAkB;AACxB,QAAM,gBAAgB,oBAAI,IAAY;AACtC,MAAI;AACJ,UAAQ,IAAI,gBAAgB,KAAK,GAAG,OAAO,KAAM,eAAc,IAAI,UAAU,EAAE,CAAC,CAAC,CAAC;AAGlF,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AACrC,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,aAAa,eAAe,KAAK,EAAE,KAAK;AAC9C,QAAI,CAAC,cAAc,IAAI,KAAK,GAAG;AAC7B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,YAAY;AAClB,UAAQ,IAAI,UAAU,KAAK,GAAG,OAAO,MAAM;AACzC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAIA,QAAM,WACJ;AACF,UAAQ,IAAI,SAAS,KAAK,GAAG,OAAO,MAAM;AACxC,UAAM,WAAW,EAAE,CAAC;AACpB,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,aAAa,eAAe,KAAK,EAAE,KAAK;AAC9C,UAAM,YAAY,SAAS,YAAY;AAEvC,UAAM,aAAa,iBAAiB,KAAK,SAAS;AAClD,UAAM,aAAa,cAAc,KAAK,SAAS;AAE/C,QAAI,cAAc,CAAC,YAAY;AAC7B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH,WAAW,cAAc,YAAY;AACnC,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,SAAS;AACf,UAAQ,IAAI,OAAO,KAAK,GAAG,OAAO,MAAM;AACtC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAGA,QAAM,cAAc;AACpB,UAAQ,IAAI,YAAY,KAAK,GAAG,OAAO,MAAM;AAC3C,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,YAAY;AAClB,UAAQ,IAAI,UAAU,KAAK,GAAG,OAAO,MAAM;AACzC,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,gBAAgB;AACtB,UAAQ,IAAI,cAAc,KAAK,GAAG,OAAO,MAAM;AAC7C,UAAM,UAAU,EAAE,CAAC;AACnB,UAAM,UAAU,EAAE,CAAC;AACnB,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS,mBAAmB,OAAO,SAAS,OAAO;AAAA,MACnD,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,iBAAiB;AACvB,UAAQ,IAAI,eAAe,KAAK,GAAG,OAAO,MAAM;AAC9C,UAAM,QAAQ,EAAE,CAAC;AACjB,UAAM,SAAS,EAAE,CAAC;AAClB,UAAM,SAAS,EAAE,CAAC;AAClB,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS,oBAAoB,MAAM,SAAS,MAAM,eAAe,KAAK;AAAA,MACtE,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,WAAW;AACjB,UAAQ,IAAI,SAAS,KAAK,GAAG,OAAO,MAAM;AACxC,UAAM,WAAW,EAAE,CAAC;AACpB,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,YAAY,SAAS,YAAY;AAEvC,QAAI,CAAC,kBAAkB,KAAK,SAAS,GAAG;AACtC,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,QACvC,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,iBAAiB,aAAa,KAAK,GAAG,KAAK,2BAA2B,KAAK,GAAG;AACpF,QAAM,kBAAkB,mDAAmD,KAAK,GAAG;AACnF,MAAI,kBAAkB,iBAAiB;AACrC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,IACd,CAAC;AAAA,EACH;AAGA,QAAM,UAAU;AAChB,UAAQ,IAAI,QAAQ,KAAK,GAAG,OAAO,MAAM;AACvC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAGA,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AAErC,UAAM,OAAO,EAAE,CAAC;AAChB,QAAI,CAAC,aAAa,KAAK,IAAI,GAAG;AAC5B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACzC,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AACrC,UAAM,OAAO,EAAE,CAAC;AAChB,QAAI,CAAC,aAAa,KAAK,IAAI,GAAG;AAC5B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACzC,CAAC;AAAA,IACH;AAAA,EACF;AAEA,SAAO;AACT;AAGA,eAAe,aAAa,KAAaC,OAAY,cAA2D;AAC9G,QAAM,SAA2B,CAAC;AAClC,QAAM,EAAE,aAAa,aAAa,YAAY,UAAU,IAAI,sBAAsB,GAAG;AAGrF,QAAM,YAAY,CAAC,GAAG,oBAAI,IAAI,CAAC,GAAG,aAAa,GAAG,aAAa,GAAG,UAAU,CAAC,CAAC;AAG9E,QAAM,aAAa,oBAAI,IAAqD;AAC5E,MAAI,UAAU,SAAS,GAAG;AACxB,QAAI;AACF,YAAM,MAAM,MAAMA,MAAK;AAAA,QACrB;AAAA;AAAA;AAAA;AAAA;AAAA,QAKA,CAAC,SAAS;AAAA,MACZ;AACA,iBAAW,OAAO,IAAI,MAAM;AAC1B,mBAAW,IAAI,IAAI,WAAW;AAAA,UAC5B,UAAU,SAAS,IAAI,cAAc,KAAK,EAAE;AAAA,UAC5C,WAAW,SAAS,IAAI,cAAc,KAAK,EAAE;AAAA,QAC/C,CAAC;AAAA,MACH;AAAA,IACF,SAAS,GAAG;AAAA,IAEZ;AAAA,EACF;AAGA,aAAW,SAAS,cAAc;AAChC,QAAI,MAAM,SAAS,gCAAgC,MAAM,WAAW;AAClE,YAAM,QAAQ,WAAW,IAAI,MAAM,SAAS;AAC5C,UAAI,OAAO;AACT,cAAM,EAAE,SAAS,IAAI;AACrB,cAAM,WAAW,KAAK,MAAM,WAAW,GAAK;AAC5C,cAAM,gBAAgB;AACtB,cAAM,uBAAuB;AAE7B,YAAI,WAAW,KAAW;AACxB,gBAAM,WAAW;AACjB,gBAAM,UAAU,oBAAoB,MAAM,SAAS,2BAA2B,QAAQ,OAAO,WAAW,KAAK,QAAQ,CAAC,CAAC;AAAA,QACzH,WAAW,WAAW,KAAS;AAC7B,gBAAM,UAAU,oBAAoB,MAAM,SAAS,2BAA2B,QAAQ,OAAO,WAAW,KAAM,QAAQ,CAAC,CAAC;AAAA,QAC1H;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAGA,QAAM,kBAAkB,CAAC,GAAG,IAAI,IAAI,SAAS,CAAC;AAC9C,aAAW,SAAS,iBAAiB;AACnC,QAAI;AACF,YAAM,MAAM,MAAMA,MAAK;AAAA,QACrB;AAAA,QACA,CAAC,KAAK;AAAA,MACR;AACA,UAAI,IAAI,KAAK,WAAW,GAAG;AACzB,eAAO,KAAK;AAAA,UACV,UAAU;AAAA,UACV,MAAM;AAAA,UACN,SAAS,UAAU,KAAK;AAAA,UACxB,WAAW;AAAA,QACb,CAAC;AAAA,MACH;AAAA,IACF,SAAS,GAAG;AAAA,IAEZ;AAAA,EACF;AAEA,SAAO;AACT;AAEA,eAAsB,iBAAiB,KAAaA,OAA4C;AAC9F,QAAM,UAAU,IAAI,KAAK;AAEzB,MAAI,CAAC,SAAS;AACZ,WAAO;AAAA,MACL,MAAM;AAAA,MACN,QAAQ,CAAC;AAAA,MACT,SAAS,EAAE,QAAQ,GAAG,UAAU,GAAG,OAAO,EAAE;AAAA,MAC5C,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IACpC;AAAA,EACF;AAGA,QAAM,SAAS,YAAY,OAAO;AAGlC,MAAIA,OAAM;AACR,UAAM,gBAAgB,MAAM,aAAa,SAASA,OAAM,MAAM;AAC9D,WAAO,KAAK,GAAG,aAAa;AAAA,EAC9B;AAEA,QAAM,SAAS,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,OAAO,EAAE;AAC5D,QAAM,WAAW,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,SAAS,EAAE;AAChE,QAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,MAAM,EAAE;AAE1D,SAAO;AAAA,IACL,MAAM,WAAW;AAAA,IACjB;AAAA,IACA,SAAS,EAAE,QAAQ,UAAU,MAAM;AAAA,IACnC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;ACrYO,SAAS,YAAY,OAAuB;AACjD,MAAI,QAAQ,KAAM,QAAO;AACzB,MAAI,QAAQ,OAAO,KAAM,QAAO,GAAG,KAAK,MAAM,QAAQ,IAAI,CAAC;AAC3D,MAAI,QAAQ,OAAO,OAAO,KAAM,QAAO,IAAI,SAAS,OAAO,OAAO,QAAQ,CAAC,CAAC;AAC5E,MAAI,QAAQ,QAAQ,EAAG,QAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAC/D,SAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAC1C;AAEA,eAAsB,iBAAiBC,OAAwC;AAC7E,QAAM,CAAC,aAAa,cAAc,IAAI,MAAM,QAAQ,IAAI;AAAA,IACtDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAiBV;AAAA,IACDA,MAAK,MAAM,0CAA0C;AAAA,EACvD,CAAC;AAED,QAAM,aAAa,eAAe,KAAK,CAAC,GAAG,cACvC,IAAI,KAAK,eAAe,KAAK,CAAC,EAAE,WAAW,EAAE,YAAY,IACzD;AAEJ,QAAM,eAAe,YAAY,KAAK,OAAO,CAAC,QAAa;AACzD,UAAM,MAAc,IAAI,YAAY;AAEpC,QAAI,IAAI,SAAS,SAAS,EAAG,QAAO;AAGpC,UAAM,WAAW,IAAI,QAAQ,GAAG;AAChC,UAAM,SAAS,IAAI,YAAY,GAAG;AAClC,QAAI,aAAa,MAAM,WAAW,IAAI;AACpC,YAAM,OAAO,IAAI,MAAM,WAAW,GAAG,MAAM;AAC3C,UAAI,KAAK,SAAS,GAAG,EAAG,QAAO;AAAA,IACjC;AACA,WAAO;AAAA,EACT,CAAC;AAED,QAAM,UAAyB,aAAa,IAAI,CAAC,QAAa;AAC5D,UAAM,YAAY,SAAS,IAAI,kBAAkB,EAAE,KAAK;AACxD,UAAM,QAAQ,IAAI;AAClB,UAAM,QAAQ,IAAI;AAClB,WAAO;AAAA,MACL,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA;AAAA,MACA,WAAW,YAAY,SAAS;AAAA,MAChC,gBAAgB;AAAA,MAChB,OAAO,SAAS,IAAI,UAAU,EAAE,KAAK;AAAA,MACrC,UAAU;AAAA,MACV,YAAY,SAAS,KAAK,OAAO,KAAK,kFAAkF,MAAM,QAAQ,MAAM,IAAI,CAAC;AAAA,IACnJ;AAAA,EACF,CAAC;AAED,QAAM,mBAAmB,QAAQ,OAAO,CAAC,KAAK,QAAQ,MAAM,IAAI,gBAAgB,CAAC;AAEjF,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,aAAa,YAAY,gBAAgB;AAAA,IACzC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AC5EA,SAAS,cAAc,OAAe,cAA8B;AAClE,MAAI,gBAAgB,IAAI;AACtB,WAAO,iBAAiB,KAAK,KAAK,YAAY,qCAAqC,KAAK;AAAA,EAC1F,WAAW,gBAAgB,IAAI;AAC7B,WAAO,qBAAqB,KAAK,KAAK,YAAY,yCAAyC,KAAK;AAAA,EAClG,OAAO;AACL,WAAO,kBAAkB,KAAK,KAAK,YAAY;AAAA,EACjD;AACF;AAEA,eAAsB,eAAeC,OAAkC;AACrE,QAAM,SAAS,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAY/B;AAED,QAAM,SAAuB,CAAC;AAE9B,aAAW,OAAO,OAAO,MAAM;AAC7B,UAAM,OAAO,SAAS,IAAI,YAAY,EAAE,KAAK;AAC7C,UAAM,OAAO,SAAS,IAAI,YAAY,EAAE,KAAK;AAC7C,UAAM,QAAQ,OAAO;AACrB,QAAI,UAAU,EAAG;AAEjB,UAAM,eAAe,KAAK,MAAO,OAAO,QAAS,GAAI,IAAI;AACzD,QAAI,eAAe,GAAI;AAEvB,UAAM,QAAQ,IAAI;AAClB,WAAO,KAAK;AAAA,MACV,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA,UAAU;AAAA,MACV,UAAU;AAAA,MACV;AAAA,MACA,gBAAgB,IAAI,kBAAkB,IAAI,KAAK,IAAI,eAAe,EAAE,YAAY,IAAI;AAAA,MACpF,YAAY,IAAI,cAAc,IAAI,KAAK,IAAI,WAAW,EAAE,YAAY,IAAI;AAAA,MACxE,YAAY,cAAc,OAAO,YAAY;AAAA,IAC/C,CAAC;AAAA,EACH;AAGA,SAAO,KAAK,CAAC,GAAG,MAAM,EAAE,eAAe,EAAE,YAAY;AAErD,SAAO;AAAA,IACL;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AChDA,SAAS,eACP,gBACA,YACA,aACiC;AACjC,MAAI,mBAAmB,KAAM,QAAO;AAEpC,QAAM,aAAa,KAAK,IAAI,IAAI,eAAe,QAAQ,MAAM,MAAO,KAAK,KAAK;AAE9E,MAAI,YAAY,KAAK,aAAa,IAAQ,QAAO;AACjD,MAAI,YAAY,EAAG,QAAO;AAC1B,SAAO;AACT;AAEA,SAASC,eAAc,QAA8C,OAA8B;AACjG,UAAQ,QAAQ;AAAA,IACd,KAAK;AACH,aAAO,SAAS,KAAK;AAAA,IACvB,KAAK;AACH,aAAO,SAAS,KAAK,wEAAwE,KAAK;AAAA,IACpG,KAAK;AACH,aAAO,SAAS,KAAK;AAAA,IACvB,KAAK;AACH,aAAO;AAAA,EACX;AACF;AAEA,eAAsB,oBAAoBC,OAAuC;AAC/E,QAAM,CAAC,aAAa,cAAc,IAAI,MAAM,QAAQ,IAAI;AAAA,IACtDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KASV;AAAA,IACDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA,KAIV;AAAA,EACH,CAAC;AAED,QAAM,SAAkC,YAAY,KAAK,IAAI,CAAC,QAAa;AACzE,UAAM,qBAAqB,IAAI,kBAAkB,IAAI,KAAK,IAAI,eAAe,IAAI;AACjF,UAAM,aAAa,SAAS,IAAI,YAAY,EAAE,KAAK;AACnD,UAAM,aAAa,SAAS,IAAI,YAAY,EAAE,KAAK;AACnD,UAAM,cAAc,SAAS,IAAI,kBAAkB,EAAE,KAAK;AAC1D,UAAM,eAAe,SAAS,IAAI,mBAAmB,EAAE,KAAK;AAC5D,UAAM,SAAS,eAAe,oBAAoB,YAAY,WAAW;AACzE,UAAM,QAAQ,IAAI;AAElB,WAAO;AAAA,MACL,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA,gBAAgB,qBAAqB,mBAAmB,YAAY,IAAI;AAAA,MACxE,iBAAiB,IAAI,mBAAmB,IAAI,KAAK,IAAI,gBAAgB,EAAE,YAAY,IAAI;AAAA,MACvF;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,YAAYD,eAAc,QAAQ,KAAK;AAAA,IACzC;AAAA,EACF,CAAC;AAED,QAAM,cAAc,oBAAI,IAAoB;AAC5C,aAAW,OAAO,eAAe,MAAM;AACrC,gBAAY,IAAI,IAAI,MAAM,IAAI,OAAO;AAAA,EACvC;AAEA,SAAO;AAAA,IACL;AAAA,IACA,UAAU;AAAA,MACR,mBAAmB,YAAY,IAAI,YAAY,MAAM;AAAA,MACrD,iBAAiB,GAAG,YAAY,IAAI,8BAA8B,KAAK,GAAG;AAAA,MAC1E,sBAAsB,SAAS,YAAY,IAAI,wBAAwB,KAAK,KAAK,EAAE;AAAA,MACnF,mBAAmB,GAAG,YAAY,IAAI,oBAAoB,KAAK,IAAI;AAAA,IACrE;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;ACrFO,SAAS,mBAAmB,MAAsB;AACvD,QAAM,IAAI,KAAK,MAAM,OAAO,IAAI;AAChC,QAAM,IAAI,KAAK,MAAO,OAAO,OAAQ,EAAE;AACvC,QAAM,IAAI,OAAO;AACjB,SAAO;AAAA,IACL,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,IACzB,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,IACzB,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,EAC3B,EAAE,KAAK,GAAG;AACZ;AAEA,eAAsB,cAAcE,OAAiC;AACnE,QAAM,CAAC,aAAa,UAAU,IAAI,MAAM,QAAQ,IAAI;AAAA,IAClDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAmBV;AAAA,IACDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAcV;AAAA,EACH,CAAC;AAID,QAAM,OAAO,oBAAI,IAAY;AAC7B,QAAM,eAA2B,CAAC;AAClC,aAAW,OAAO,YAAY,MAAM;AAClC,UAAM,MAAM,GAAG,IAAI,WAAW,IAAI,IAAI,YAAY;AAClD,QAAI,CAAC,KAAK,IAAI,GAAG,GAAG;AAClB,WAAK,IAAI,GAAG;AACZ,mBAAa,KAAK;AAAA,QAChB,YAAY,SAAS,IAAI,aAAa,EAAE;AAAA,QACxC,cAAc,IAAI;AAAA,QAClB,iBAAiB,mBAAmB,SAAS,IAAI,cAAc,EAAE,KAAK,CAAC;AAAA,QACvE,aAAa,SAAS,IAAI,cAAc,EAAE;AAAA,QAC1C,eAAe,IAAI;AAAA,QACnB,kBAAkB,mBAAmB,SAAS,IAAI,eAAe,EAAE,KAAK,CAAC;AAAA,QACzE,OAAO,IAAI,cAAc;AAAA,QACzB,UAAU,IAAI;AAAA,MAChB,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,qBAAqB,WAAW,KAAK,IAAI,CAAC,SAAc;AAAA,IAC5D,KAAK,SAAS,IAAI,KAAK,EAAE;AAAA,IACzB,UAAU,mBAAmB,SAAS,IAAI,eAAe,EAAE,KAAK,CAAC;AAAA,IACjE,OAAO,IAAI;AAAA,IACX,OAAO,IAAI;AAAA,IACX,eAAe,IAAI,mBAAmB;AAAA,EACxC,EAAE;AAEF,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AChFA,SAAS,eAAe,OAAe,MAAkC;AACvE,QAAM,IAAI,WAAW,KAAK;AAC1B,MAAI,CAAC,KAAM,QAAO;AAClB,UAAQ,KAAK,YAAY,GAAG;AAAA,IAC1B,KAAK;AAAO,aAAO;AAAA,IACnB,KAAK;AAAO,aAAO,IAAI;AAAA,IACvB,KAAK;AAAO,aAAO,IAAI,IAAI;AAAA;AAAA,IAC3B,KAAK;AAAO,aAAO,IAAI,OAAO;AAAA,IAC9B,KAAK;AAAO,aAAO,IAAI,OAAO,OAAO;AAAA,IACrC;AAAY,aAAO;AAAA,EACrB;AACF;AAEA,SAAS,YAAY,OAAe,MAAkC;AACpE,SAAO,eAAe,OAAO,IAAI,KAAK,OAAO;AAC/C;AAGA,SAAS,iBAAiB,UAAqC,MAAuB;AACpF,MAAI,CAAC,SAAU,QAAO;AACtB,QAAM,QAAQ,eAAe,UAAU,QAAQ,EAAE;AACjD,MAAI,SAAS,KAAK,MAAM,KAAK,EAAG,QAAO;AACvC,MAAI,SAAS,QAAQ,EAAG,QAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAChE,MAAI,SAAS,QAAQ,EAAG,QAAO,GAAG,KAAK,MAAM,QAAQ,QAAQ,CAAC,CAAC;AAC/D,MAAI,SAAS,KAAM,QAAO,GAAG,KAAK,MAAM,QAAQ,IAAI,CAAC;AACrD,SAAO,GAAG,KAAK;AACjB;AAEA,eAAsB,gBAAgBC,OAAmC;AACvE,QAAM,SAAS,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAW/B;AAED,QAAM,WAA0E,CAAC;AACjF,aAAW,OAAO,OAAO,MAAM;AAC7B,aAAS,IAAI,IAAI,IAAI,EAAE,SAAS,IAAI,SAAS,MAAM,IAAI,QAAQ,OAAU;AAAA,EAC3E;AAEA,QAAM,kBAA0C,CAAC;AAEjD,QAAM,MAAM,CAAC,SAAiB,SAAS,IAAI,GAAG,WAAW;AACzD,QAAM,UAAU,CAAC,SAAiB,SAAS,IAAI,GAAG;AAGlD,QAAM,uBAAuB,IAAI,gBAAgB;AACjD,MAAI,yBAAyB,MAAM;AACjC,UAAM,KAAK,YAAY,sBAAsB,QAAQ,gBAAgB,CAAC;AACtE,QAAI,KAAK,KAAK;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,EAAE,CAAC;AAAA,QAC/B,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,iBAAiB,IAAI,UAAU;AACrC,MAAI,mBAAmB,MAAM;AAC3B,UAAM,KAAK,YAAY,gBAAgB,QAAQ,UAAU,CAAC;AAC1D,QAAI,MAAM,GAAG;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,IAAI,KAAK,GAAG,QAAQ,CAAC,CAAC;AAAA,QAClD,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,aAAa,IAAI,8BAA8B;AACrD,MAAI,eAAe,MAAM;AACvB,UAAM,IAAI,WAAW,UAAU;AAC/B,QAAI,IAAI,KAAK;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,aAAa,IAAI,kBAAkB;AACzC,MAAI,eAAe,MAAM;AACvB,UAAM,IAAI,WAAW,UAAU;AAC/B,QAAI,IAAI,GAAK;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,cAAc,IAAI,gCAAgC;AACxD,MAAI,gBAAgB,MAAM;AACxB,UAAM,IAAI,WAAW,WAAW;AAChC,QAAI,KAAK,KAAK;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,cAAc,IAAI,4BAA4B;AACpD,MAAI,gBAAgB,QAAQ,SAAS,aAAa,EAAE,MAAM,IAAI;AAC5D,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAGA,QAAM,cAAc,IAAI,qCAAqC;AAC7D,MAAI,gBAAgB,QAAQ,SAAS,aAAa,EAAE,MAAM,GAAG;AAC3D,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAGA,QAAM,aAAa,IAAI,0BAA0B;AACjD,MAAI,eAAe,QAAQ,SAAS,YAAY,EAAE,MAAM,GAAG;AACzD,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAKA,QAAM,aAAa,IAAI,sBAAsB;AAC7C,MAAI,eAAe,MAAM;AACvB,UAAM,KAAK,YAAY,YAAY,QAAQ,sBAAsB,CAAC;AAClE,QAAI,MAAM,IAAI;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,IAAI,KAAK,GAAG,QAAQ,CAAC,CAAC;AAAA,QAClD,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,iBAAiB,IAAI,iBAAiB;AAG5C,MAAI,mBAAmB,MAAM;AAC3B,UAAM,UAAU,SAAS,gBAAgB,EAAE;AAC3C,QAAI,UAAU,KAAK;AACjB,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,OAAO,OAAO;AAAA,QAC5B,kBAAkB;AAAA,QAClB,QAAQ,mBAAmB,OAAO;AAAA,QAClC,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,aAAa;AAAA,IACjB,gBAAgB,mBAAmB,OAAO,SAAS,gBAAgB,EAAE,IAAI;AAAA,IACzE,eAAe,iBAAiB,sBAAsB,QAAQ,gBAAgB,CAAC;AAAA,IAC/E,SAAS,iBAAiB,gBAAgB,QAAQ,UAAU,CAAC;AAAA,IAC7D,oBAAoB,iBAAiB,IAAI,sBAAsB,GAAG,QAAQ,sBAAsB,CAAC;AAAA,IACjG,oBAAoB,iBAAiB,YAAY,QAAQ,sBAAsB,CAAC;AAAA,IAChF,YAAY,IAAI,aAAa,KAAK;AAAA,IAClC,4BAA4B,cAAc;AAAA,IAC1C,gBAAgB,cAAc;AAAA,IAC9B,6BAA6B,eAAe;AAAA,EAC9C;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AjBvOA,OAAOC,eAAc;AACrB,OAAOC,WAAU;AACjB,OAAOC,SAAQ;AACf,OAAOC,OAAM,oBAAoB;AAEjC,IAAM,MAAM,KAAK,MAAM,aAAa,IAAI,IAAI,mBAAmB,YAAY,GAAG,GAAG,OAAO,CAAC;AAEzF,IAAM,aAAa,QAAQ,KAAK,CAAC,KAAK,QAAQ,IAAI;AAClD,IAAI,CAAC,YAAY;AACf,UAAQ,MAAM,wCAAwC;AACtD,UAAQ,MAAM,4CAA4C;AAC1D,UAAQ,KAAK,CAAC;AAChB;AAEA,IAAM,OAAO,IAAIC,MAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AACtF,IAAM,qBAAqB,SAAS,QAAQ,IAAI,gCAAgC,KAAK,EAAE;AACvF,IAAM,UAAU,QAAQ,IAAI,oBAAoBH,MAAK,KAAKC,IAAG,QAAQ,GAAG,UAAU;AAClFC,IAAG,UAAU,SAAS,EAAE,WAAW,KAAK,CAAC;AAGzC,IAAI,WAAqC;AACzC,IAAI,WAAqC;AACzC,IAAI;AACF,QAAM,aAAaF,MAAK,KAAK,SAAS,WAAW;AACjD,MAAIE,IAAG,WAAW,UAAU,EAAG,YAAW,IAAIH,UAAS,YAAY,EAAE,UAAU,KAAK,CAAC;AACvF,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AACvE,IAAI;AACF,QAAM,aAAaC,MAAK,KAAK,SAAS,WAAW;AACjD,MAAIE,IAAG,WAAW,UAAU,EAAG,YAAW,IAAIH,UAAS,YAAY,EAAE,UAAU,KAAK,CAAC;AACvF,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AAEvE,IAAI,eAAyC;AAC7C,IAAI;AACF,QAAM,iBAAiBC,MAAK,KAAK,SAAS,gBAAgB;AAC1D,MAAIE,IAAG,WAAW,cAAc,EAAG,gBAAe,IAAIH,UAAS,gBAAgB,EAAE,UAAU,KAAK,CAAC;AACnG,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AAEvE,IAAM,SAAS,IAAI,UAAU,EAAE,MAAM,WAAW,SAAS,IAAI,QAAQ,CAAC;AAEtE,OAAO,KAAK,oBAAoB,8DAA8D,CAAC,GAAG,YAAY;AAC5G,MAAI;AACF,UAAM,OAAO,MAAM,YAAY,IAAI;AACnC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,2DAA2D,CAAC,GAAG,YAAY;AACvG,MAAI;AACF,UAAM,OAAO,MAAM,iBAAiB,MAAM,kBAAkB;AAC5D,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,6CAA6C,CAAC,GAAG,YAAY;AACzF,MAAI;AACF,UAAM,OAAO,MAAM,UAAU,IAAI;AACjC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,4CAA4C,EAAE,OAAO,EAAE,OAAO,EAAE,SAAS,6CAA6C,EAAE,GAAG,OAAO,EAAE,MAAM,MAAM;AAClL,MAAI;AACF,UAAM,OAAO,MAAM,qBAAqB,MAAM,KAAK;AACnD,QAAI,CAAC,KAAM,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,kBAAkB,CAAC,GAAG,SAAS,KAAK;AACxF,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,oBAAoB,+DAA+D,CAAC,GAAG,YAAY;AAC7G,MAAI;AACF,UAAM,OAAO,MAAM,YAAY,IAAI;AACnC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,0BAA0B,6BAA6B,CAAC,GAAG,YAAY;AACjF,MAAI;AACF,QAAI,CAAC,SAAU,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,+DAA+D,CAAC,EAAE;AAC1H,UAAM,UAAU,SAAS,QAAQ,+DAA+D,EAAE,IAAI;AACtG,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,SAAS,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC/E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,eAAe,uDAAuD,EAAE,KAAK,EAAE,OAAO,EAAE,SAAS,2CAA2C,EAAE,GAAG,OAAO,EAAE,IAAI,MAAM;AAC9K,MAAI;AACF,QAAI,CAAC,UAAU,GAAG,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,oKAAoK,CAAC,GAAG,SAAS,KAAK;AACpP,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,QAAQ,KAAK,IAAI;AACvB,YAAM,SAAS,MAAM,OAAO,MAAM,GAAG;AACrC,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,IAAI,MAAM,UAAU,KAAK,IAAI,IAAI,OAAO,UAAU,OAAO,UAAU,MAAM,OAAO,QAAQ,CAAC,EAAE,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IACtK,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,qBAAqB,CAAC,GAAG,YAAY;AACjE,MAAI;AACF,QAAI,CAAC,SAAU,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,sDAAsD,CAAC,EAAE;AACjH,UAAM,UAAU,SAAS,QAAQ,8DAA8D,EAAE,IAAI;AACrG,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,SAAS,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC/E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAID,OAAO,KAAK,mBAAmB,gFAAgF,EAAE,OAAO,EAAE,OAAO,EAAE,SAAS,yBAAyB,EAAE,GAAG,OAAO,EAAE,MAAM,MAAM;AAC7L,MAAI;AACF,QAAI,CAAC,gBAAgB,KAAK,KAAK,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yCAAyC,CAAC,GAAG,SAAS,KAAK;AACtI,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,OAAO,MAAM,+BAA+B;AAClD,YAAM,OAAO,MAAM,OAAO;AAC1B,UAAI;AACF,cAAM,IAAI,MAAM,OAAO,MAAM,2CAA2C,KAAK,EAAE;AAC/E,cAAM,OAAO,MAAM,UAAU;AAC7B,cAAM,OAAO,MAAM,yBAAyB;AAC5C,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,KAAK,CAAC,EAAE,YAAY,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,MAC/F,SAAS,KAAU;AACjB,cAAM,OAAO,MAAM,UAAU,EAAE,MAAM,MAAM;AAAA,QAAC,CAAC;AAC7C,cAAM,OAAO,MAAM,yBAAyB,EAAE,MAAM,MAAM;AAAA,QAAC,CAAC;AAC5D,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,MACrF;AAAA,IACF,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,qBAAqB,kEAAkE,EAAE,UAAU,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,6DAA6D,EAAE,GAAG,OAAO,EAAE,SAAS,MAAM;AACtO,MAAI;AACF,UAAM,SAAS,MAAM,iBAAiB,MAAM,kBAAkB;AAC9D,QAAI,QAAQ,OAAO;AACnB,QAAI,SAAU,SAAQ,MAAM,OAAO,CAAC,MAAM,EAAE,KAAK,WAAW,QAAQ,CAAC;AACrE,QAAI,MAAM,WAAW,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,0BAA0B,WAAW,kBAAkB,QAAQ,MAAM,IAAI,CAAC,EAAE;AAC7I,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,MAAM,EAAE,KAAK;AAAA,EAAK,EAAE,GAAG,EAAE,EAAE,KAAK,MAAM;AACxE,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,SAAS,CAAC,EAAE;AAAA,EACvD,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,gDAAgD;AAAA,EAClF,OAAO,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE,EAAE,SAAS,oCAAoC;AAAA,EACtF,SAAS,EAAE,KAAK,CAAC,cAAc,aAAa,OAAO,CAAC,EAAE,SAAS,EAAE,QAAQ,YAAY,EAAE,SAAS,YAAY;AAC9G,GAAG,OAAO,EAAE,OAAO,QAAQ,MAAM;AAC/B,MAAI;AACF,UAAM,MAAM,MAAM,eAAe,IAAI;AACrC,QAAI,IAAI,WAAW,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yEAAyE,CAAC,EAAE;AAC3I,UAAM,SAAS,CAAC,GAAG,GAAG,EAAE,KAAK,CAAC,GAAG,MAAO,EAAU,OAAO,IAAK,EAAU,OAAO,CAAC;AAChF,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,OAAO,MAAM,GAAG,KAAK,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9F,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,uBAAuB,sDAAsD,CAAC,GAAG,YAAY;AACvG,MAAI;AACF,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,YAAM,SAAS,EAAE,KAAK,IAAI,CAAC,SAAc;AAAA,QACvC,QAAQ,IAAI;AAAA,QACZ,MAAM,IAAI;AAAA,QACV,WAAW,SAAS,IAAI,UAAU;AAAA,QAClC,WAAW,SAAS,IAAI,UAAU;AAAA,QAClC,WAAW,SAAS,IAAI,UAAU;AAAA,MACpC,EAAE;AACF,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,6BAA6B,EAAE,QAAQ,EAAE,KAAK,CAAC,QAAQ,IAAI,CAAC,EAAE,SAAS,EAAE,QAAQ,MAAM,EAAE,SAAS,2BAA2B,EAAE,GAAG,OAAO,EAAE,OAAO,MAAM;AACpL,MAAI;AACF,UAAM,CAAC,UAAU,OAAO,IAAI,MAAM,QAAQ,IAAI;AAAA,MAC5C,YAAY,IAAI;AAAA,MAChB,iBAAiB,MAAM,kBAAkB;AAAA,IAC3C,CAAC;AACD,QAAI,WAAW,MAAM;AACnB,YAAM,QAAkB,CAAC;AACzB,YAAM,KAAK,yBAAyB;AACpC,YAAM,KAAK;AAAA,cAAgB,oBAAI,KAAK,GAAE,YAAY,CAAC;AAAA,CAAI;AACvD,YAAM,KAAK;AAAA,CAAe;AAC1B,YAAM,KAAK,qBAAqB,SAAS,OAAO,EAAE;AAClD,YAAM,KAAK,wBAAwB,SAAS,MAAM,EAAE;AACpD,YAAM,KAAK,sBAAsB,SAAS,YAAY,MAAM,aAAa,SAAS,YAAY,IAAI,WAAW,SAAS,YAAY,GAAG,MAAM;AAC3I,YAAM,KAAK;AAAA,mBAAsB,QAAQ,KAAK,gBAAgB,QAAQ,KAAK;AAAA,CAAK;AAChF,YAAM,KAAK,uCAAuC;AAClD,YAAM,KAAK,uCAAuC;AAClD,iBAAW,CAAC,KAAK,CAAC,KAAK,OAAO,QAAQ,QAAQ,SAAS,GAAG;AACxD,cAAM,KAAK,KAAK,GAAG,MAAM,EAAE,KAAK,MAAM,EAAE,KAAK,UAAU,EAAE,KAAK,IAAI;AAAA,MACpE;AACA,UAAI,QAAQ,OAAO,SAAS,GAAG;AAC7B,cAAM,KAAK;AAAA,cAAiB,QAAQ,OAAO,MAAM;AAAA,CAAK;AACtD,mBAAW,SAAS,QAAQ,QAAQ;AAClC,gBAAM,OAAO,MAAM,aAAa,aAAa,cAAO,MAAM,aAAa,YAAY,cAAO;AAC1F,gBAAM,KAAK,KAAK,IAAI,KAAK,MAAM,QAAQ,KAAK,MAAM,KAAK,EAAE;AAAA,QAC3D;AAAA,MACF;AACA,UAAI,QAAQ,WAAW,SAAS,GAAG;AACjC,cAAM,KAAK;AAAA;AAAA,CAAwB;AACnC,mBAAW,OAAO,QAAQ,YAAY;AACpC,gBAAM,KAAK;AAAA,EAAc,IAAI,GAAG;AAAA;AAAA,CAAY;AAAA,QAC9C;AAAA,MACF;AACA,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,MAAM,KAAK,IAAI,EAAE,CAAC,EAAE;AAAA,IAC/D;AACA,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,UAAU,SAAS,aAAY,oBAAI,KAAK,GAAE,YAAY,EAAE,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EACnI,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,gBAAgB,mDAAmD,CAAC,GAAG,YAAY;AAC7F,MAAI;AACF,UAAM,eAAeC,MAAK,KAAK,SAAS,iBAAiB;AACzD,UAAM,OAAO,aAAa,YAAY;AACtC,UAAM,UAAU,MAAM,iBAAiB,MAAM,kBAAkB;AAC/D,QAAI,CAAC,MAAM;AACT,mBAAa,cAAc,OAAO;AAClC,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,SAAS,iEAAiE,OAAO,QAAQ,OAAO,OAAO,QAAQ,OAAO,QAAQ,QAAQ,OAAO,OAAO,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC/N;AACA,UAAM,OAAO,cAAc,KAAK,QAAQ,OAAO;AAC/C,iBAAa,cAAc,OAAO;AAClC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,GAAG,MAAM,mBAAmB,KAAK,UAAU,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EACtH,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,KAAK,EAAE,OAAO,EAAE,SAAS,kCAAkC;AAAA,EAC7D;AAAA,EACA,OAAO,EAAE,IAAI,MAAM;AACjB,QAAI;AACF,YAAM,SAAS,MAAM,iBAAiB,KAAK,IAAI;AAC/C,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,KAAK,EAAE,OAAO,EAAE,SAAS,yBAAyB;AAAA,EACpD;AAAA,EACA,OAAO,EAAE,IAAI,MAAM;AACjB,QAAI;AACF,UAAI,CAAC,gBAAgB,KAAK,GAAG,GAAG;AAC9B,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yCAAyC,CAAC,GAAG,SAAS,KAAK;AAAA,MACtG;AACA,YAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,UAAI;AACF,cAAM,OAAO,MAAM,+BAA+B;AAClD,cAAM,OAAO,MAAM,OAAO;AAC1B,YAAI;AACF,gBAAM,IAAI,MAAM,OAAO,MAAM,2CAA2C,GAAG,EAAE;AAC7E,gBAAM,OAAO,MAAM,UAAU;AAC7B,gBAAM,OAAO,MAAM,yBAAyB;AAE5C,gBAAM,OAAO,EAAE,KAAK,CAAC,EAAE,YAAY;AACnC,gBAAM,WAAW,MAAM,mBAAmB,MAAM,IAAI;AAEpD,iBAAO;AAAA,YACL,SAAS,CAAC;AAAA,cACR,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,MAAM,SAAS,GAAG,MAAM,CAAC;AAAA,YAClD,CAAC;AAAA,UACH;AAAA,QACF,SAAS,KAAU;AACjB,gBAAM,OAAO,MAAM,UAAU,EAAE,MAAM,MAAM;AAAA,UAAC,CAAC;AAC7C,gBAAM,OAAO,MAAM,yBAAyB,EAAE,MAAM,MAAM;AAAA,UAAC,CAAC;AAC5D,iBAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,QACrF;AAAA,MACF,UAAE;AACA,eAAO,QAAQ;AAAA,MACjB;AAAA,IACF,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,aAAa,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,kCAAkC;AAAA,EAChF;AAAA,EACA,OAAO,EAAE,YAAY,MAAM;AACzB,QAAI;AACF,YAAM,cAAc,MAAM,uBAAuB,MAAM,cAAc,eAAe,EAAE;AACtF,UAAI,YAAY,WAAW,GAAG;AAC5B,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yDAAyD,CAAC,EAAE;AAAA,MACvG;AACA,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,aAAa,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IACnF,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,WAAW,EAAE,OAAO,EAAE,SAAS,sDAAsD;AAAA,IACrF,eAAe,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS,uCAAuC;AAAA,EACxF;AAAA,EACA,OAAO,EAAE,WAAW,cAAc,MAAM;AACtC,QAAI;AAGF,YAAM,SAAS,MAAM,iBAAiB,YAAY,WAAW,EAAE,eAAe,iBAAiB,MAAM,CAAC;AACtG,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO,KAAK,0BAA0B,6DAA6D,CAAC,GAAG,YAAY;AACjH,MAAI;AACF,UAAM,SAAS,MAAM,iBAAiB,IAAI;AAC1C,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,iBAAiB,2DAA2D,CAAC,GAAG,YAAY;AACtG,MAAI;AACF,UAAM,SAAS,MAAM,eAAe,IAAI;AACxC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,sBAAsB,2EAAsE,CAAC,GAAG,YAAY;AACtH,MAAI;AACF,UAAM,SAAS,MAAM,oBAAoB,IAAI;AAC7C,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,iBAAiB,yEAAyE,CAAC,GAAG,YAAY;AACpH,MAAI;AACF,UAAM,SAAS,MAAM,cAAc,IAAI;AACvC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,0EAA0E,CAAC,GAAG,YAAY;AAC5H,MAAI;AACF,UAAM,SAAS,MAAM,gBAAgB,IAAI;AACzC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,IAAM,YAAY,IAAI,qBAAqB;AAC3C,MAAM,OAAO,QAAQ,SAAS;","names":["Pool","pool","pool","pool","pool","pool","pool","longQueryThreshold","dataDir","pool","fs","path","pool","pool","longQueryThreshold","pool","pool","pool","pool","getSuggestion","pool","pool","pool","Database","path","os","fs","Pool"]}
|
|
1
|
+
{"version":3,"sources":["../src/mcp.ts","../src/server/queries/overview.ts","../src/server/queries/tables.ts","../src/server/queries/schema.ts","../src/server/queries/activity.ts","../src/server/advisor.ts","../src/server/queries/slow-queries.ts","../src/server/snapshot.ts","../src/server/env-differ.ts","../src/server/schema-diff.ts","../src/server/schema-tracker.ts","../src/server/query-analyzer.ts","../src/server/migration-checker.ts","../src/server/unused-indexes.ts","../src/server/bloat.ts","../src/server/autovacuum.ts","../src/server/locks.ts","../src/server/config-checker.ts"],"sourcesContent":["// MCP Server for pg-dash — exposes PostgreSQL monitoring tools\n\nimport { McpServer } from \"@modelcontextprotocol/sdk/server/mcp.js\";\nimport { StdioServerTransport } from \"@modelcontextprotocol/sdk/server/stdio.js\";\nimport { Pool } from \"pg\";\nimport { z } from \"zod\";\nimport { getOverview } from \"./server/queries/overview.js\";\nimport { getTables } from \"./server/queries/tables.js\";\nimport { getSchemaTableDetail } from \"./server/queries/schema.js\";\nimport { getActivity } from \"./server/queries/activity.js\";\nimport { getAdvisorReport, isSafeFix } from \"./server/advisor.js\";\nimport { getSlowQueries } from \"./server/queries/slow-queries.js\";\nimport { saveSnapshot, loadSnapshot, diffSnapshots } from \"./server/snapshot.js\";\nimport { diffEnvironments } from \"./server/env-differ.js\";\nimport { analyzeExplainPlan, detectQueryRegressions } from \"./server/query-analyzer.js\";\nimport { analyzeMigration } from \"./server/migration-checker.js\";\nimport { getUnusedIndexes } from \"./server/unused-indexes.js\";\nimport { getBloatReport } from \"./server/bloat.js\";\nimport { getAutovacuumReport } from \"./server/autovacuum.js\";\nimport { getLockReport } from \"./server/locks.js\";\nimport { getConfigReport } from \"./server/config-checker.js\";\nimport Database from \"better-sqlite3\";\nimport path from \"node:path\";\nimport os from \"node:os\";\nimport fs, { readFileSync } from \"node:fs\";\n\nconst pkg = JSON.parse(readFileSync(new URL(\"../package.json\", import.meta.url), \"utf-8\"));\n\nconst connString = process.argv[2] || process.env.PG_DASH_CONNECTION_STRING;\nif (!connString) {\n console.error(\"Usage: pg-dash-mcp <connection-string>\");\n console.error(\" or set PG_DASH_CONNECTION_STRING env var\");\n process.exit(1);\n}\n\nconst pool = new Pool({ connectionString: connString, connectionTimeoutMillis: 10000 });\nconst longQueryThreshold = parseInt(process.env.PG_DASH_LONG_QUERY_THRESHOLD || \"5\", 10);\nconst dataDir = process.env.PG_DASH_DATA_DIR || path.join(os.homedir(), \".pg-dash\");\nfs.mkdirSync(dataDir, { recursive: true });\n\n// Open schema and alerts DBs (read-only for history queries)\nlet schemaDb: Database.Database | null = null;\nlet alertsDb: Database.Database | null = null;\ntry {\n const schemaPath = path.join(dataDir, \"schema.db\");\n if (fs.existsSync(schemaPath)) schemaDb = new Database(schemaPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\ntry {\n const alertsPath = path.join(dataDir, \"alerts.db\");\n if (fs.existsSync(alertsPath)) alertsDb = new Database(alertsPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\n\nlet queryStatsDb: Database.Database | null = null;\ntry {\n const queryStatsPath = path.join(dataDir, \"query-stats.db\");\n if (fs.existsSync(queryStatsPath)) queryStatsDb = new Database(queryStatsPath, { readonly: true });\n} catch (err) { console.error(\"[mcp] Error:\", (err as Error).message); }\n\nconst server = new McpServer({ name: \"pg-dash\", version: pkg.version });\n\nserver.tool(\"pg_dash_overview\", \"Get database overview (version, uptime, size, connections)\", {}, async () => {\n try {\n const data = await getOverview(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_health\", \"Get health advisor report with score, grade, and issues\", {}, async () => {\n try {\n const data = await getAdvisorReport(pool, longQueryThreshold);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_tables\", \"List all tables with sizes and row counts\", {}, async () => {\n try {\n const data = await getTables(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_table_detail\", \"Get detailed info about a specific table\", { table: z.string().describe(\"Table name (e.g. 'users' or 'public.users')\") }, async ({ table }) => {\n try {\n const data = await getSchemaTableDetail(pool, table);\n if (!data) return { content: [{ type: \"text\", text: \"Table not found\" }], isError: true };\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_activity\", \"Get current database activity (active queries, connections)\", {}, async () => {\n try {\n const data = await getActivity(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(data, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_schema_changes\", \"Get recent schema changes\", {}, async () => {\n try {\n if (!schemaDb) return { content: [{ type: \"text\", text: \"No schema tracking data available. Run pg-dash server first.\" }] };\n const changes = schemaDb.prepare(\"SELECT * FROM schema_changes ORDER BY timestamp DESC LIMIT 50\").all();\n return { content: [{ type: \"text\", text: JSON.stringify(changes, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_fix\", \"Execute a safe fix (VACUUM, ANALYZE, REINDEX, etc.)\", { sql: z.string().describe(\"SQL to execute (must be a safe operation)\") }, async ({ sql }) => {\n try {\n if (!isSafeFix(sql)) return { content: [{ type: \"text\", text: \"Operation not allowed. Only VACUUM, ANALYZE, REINDEX, CREATE/DROP INDEX CONCURRENTLY, pg_terminate_backend, pg_cancel_backend, and EXPLAIN ANALYZE are permitted.\" }], isError: true };\n const client = await pool.connect();\n try {\n const start = Date.now();\n const result = await client.query(sql);\n return { content: [{ type: \"text\", text: JSON.stringify({ ok: true, duration: Date.now() - start, rowCount: result.rowCount, rows: result.rows || [] }, null, 2) }] };\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_alerts\", \"Get alert history\", {}, async () => {\n try {\n if (!alertsDb) return { content: [{ type: \"text\", text: \"No alerts data available. Run pg-dash server first.\" }] };\n const history = alertsDb.prepare(\"SELECT * FROM alert_history ORDER BY timestamp DESC LIMIT 50\").all();\n return { content: [{ type: \"text\", text: JSON.stringify(history, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\n// --- New tools ---\n\nserver.tool(\"pg_dash_explain\", \"Run EXPLAIN ANALYZE on a SELECT query (read-only, wrapped in BEGIN/ROLLBACK)\", { query: z.string().describe(\"SELECT query to explain\") }, async ({ query }) => {\n try {\n if (!/^\\s*SELECT\\b/i.test(query)) return { content: [{ type: \"text\", text: \"Error: Only SELECT queries are allowed\" }], isError: true };\n const client = await pool.connect();\n try {\n await client.query(\"SET statement_timeout = '30s'\");\n await client.query(\"BEGIN\");\n try {\n const r = await client.query(`EXPLAIN (ANALYZE, BUFFERS, FORMAT JSON) ${query}`);\n await client.query(\"ROLLBACK\");\n await client.query(\"RESET statement_timeout\");\n return { content: [{ type: \"text\", text: JSON.stringify(r.rows[0][\"QUERY PLAN\"], null, 2) }] };\n } catch (err: any) {\n await client.query(\"ROLLBACK\").catch(() => {});\n await client.query(\"RESET statement_timeout\").catch(() => {});\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_batch_fix\", \"Get batch fix SQL for issues (optionally filtered by category)\", { category: z.string().optional().describe(\"Filter by issue type prefix, e.g. 'schema-missing-fk-index'\") }, async ({ category }) => {\n try {\n const report = await getAdvisorReport(pool, longQueryThreshold);\n let fixes = report.batchFixes;\n if (category) fixes = fixes.filter((f) => f.type.startsWith(category));\n if (fixes.length === 0) return { content: [{ type: \"text\", text: \"No batch fixes found\" + (category ? ` for category '${category}'` : \"\") }] };\n const combined = fixes.map((f) => `-- ${f.title}\\n${f.sql}`).join(\"\\n\\n\");\n return { content: [{ type: \"text\", text: combined }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_slow_queries\", \"Get top slow queries from pg_stat_statements\", {\n limit: z.number().optional().default(20).describe(\"Max queries to return (default 20)\"),\n orderBy: z.enum([\"total_time\", \"mean_time\", \"calls\"]).optional().default(\"total_time\").describe(\"Sort order\"),\n}, async ({ limit, orderBy }) => {\n try {\n const all = await getSlowQueries(pool);\n if (all.length === 0) return { content: [{ type: \"text\", text: \"No slow query data available. pg_stat_statements may not be installed.\" }] };\n const sorted = [...all].sort((a, b) => (b as any)[orderBy] - (a as any)[orderBy]);\n return { content: [{ type: \"text\", text: JSON.stringify(sorted.slice(0, limit), null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_table_sizes\", \"Get table sizes with data/index breakdown (top 30)\", {}, async () => {\n try {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT schemaname, relname,\n pg_total_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as total_size,\n pg_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as table_size,\n pg_indexes_size(quote_ident(schemaname) || '.' || quote_ident(relname)) as index_size\n FROM pg_stat_user_tables\n ORDER BY pg_total_relation_size(quote_ident(schemaname) || '.' || quote_ident(relname)) DESC\n LIMIT 30\n `);\n const tables = r.rows.map((row: any) => ({\n schema: row.schemaname,\n name: row.relname,\n totalSize: parseInt(row.total_size),\n tableSize: parseInt(row.table_size),\n indexSize: parseInt(row.index_size),\n }));\n return { content: [{ type: \"text\", text: JSON.stringify(tables, null, 2) }] };\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_export\", \"Export full health report\", { format: z.enum([\"json\", \"md\"]).optional().default(\"json\").describe(\"Output format: json or md\") }, async ({ format }) => {\n try {\n const [overview, advisor] = await Promise.all([\n getOverview(pool),\n getAdvisorReport(pool, longQueryThreshold),\n ]);\n if (format === \"md\") {\n const lines: string[] = [];\n lines.push(`# pg-dash Health Report`);\n lines.push(`\\nGenerated: ${new Date().toISOString()}\\n`);\n lines.push(`## Overview\\n`);\n lines.push(`- **PostgreSQL**: ${overview.version}`);\n lines.push(`- **Database Size**: ${overview.dbSize}`);\n lines.push(`- **Connections**: ${overview.connections.active} active / ${overview.connections.idle} idle / ${overview.connections.max} max`);\n lines.push(`\\n## Health Score: ${advisor.score}/100 (Grade: ${advisor.grade})\\n`);\n lines.push(`| Category | Grade | Score | Issues |`);\n lines.push(`|----------|-------|-------|--------|`);\n for (const [cat, b] of Object.entries(advisor.breakdown)) {\n lines.push(`| ${cat} | ${b.grade} | ${b.score}/100 | ${b.count} |`);\n }\n if (advisor.issues.length > 0) {\n lines.push(`\\n### Issues (${advisor.issues.length})\\n`);\n for (const issue of advisor.issues) {\n const icon = issue.severity === \"critical\" ? \"🔴\" : issue.severity === \"warning\" ? \"🟡\" : \"🔵\";\n lines.push(`- ${icon} [${issue.severity}] ${issue.title}`);\n }\n }\n if (advisor.batchFixes.length > 0) {\n lines.push(`\\n### 🔧 Batch Fixes\\n`);\n for (const fix of advisor.batchFixes) {\n lines.push(`\\`\\`\\`sql\\n${fix.sql}\\n\\`\\`\\`\\n`);\n }\n }\n return { content: [{ type: \"text\", text: lines.join(\"\\n\") }] };\n }\n return { content: [{ type: \"text\", text: JSON.stringify({ overview, advisor, exportedAt: new Date().toISOString() }, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_diff\", \"Compare current health with last saved snapshot\", {}, async () => {\n try {\n const snapshotPath = path.join(dataDir, \"last-check.json\");\n const prev = loadSnapshot(snapshotPath);\n const current = await getAdvisorReport(pool, longQueryThreshold);\n if (!prev) {\n saveSnapshot(snapshotPath, current);\n return { content: [{ type: \"text\", text: JSON.stringify({ message: \"No previous snapshot found. Current result saved as baseline.\", score: current.score, grade: current.grade, issues: current.issues.length }, null, 2) }] };\n }\n const diff = diffSnapshots(prev.result, current);\n saveSnapshot(snapshotPath, current);\n return { content: [{ type: \"text\", text: JSON.stringify({ ...diff, previousTimestamp: prev.timestamp }, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\n \"pg_dash_check_migration\",\n \"Analyze migration SQL for safety risks (lock tables, missing tables, destructive ops)\",\n {\n sql: z.string().describe(\"Migration SQL content to analyze\"),\n },\n async ({ sql }) => {\n try {\n const result = await analyzeMigration(sql, pool);\n return { content: [{ type: \"text\", text: JSON.stringify(result, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_analyze_query\",\n \"Deep analysis of a SQL query: runs EXPLAIN ANALYZE, detects missing indexes, and provides specific optimization recommendations\",\n {\n sql: z.string().describe(\"SELECT query to analyze\"),\n },\n async ({ sql }) => {\n try {\n if (!/^\\s*SELECT\\b/i.test(sql)) {\n return { content: [{ type: \"text\", text: \"Error: Only SELECT queries are allowed\" }], isError: true };\n }\n const client = await pool.connect();\n try {\n await client.query(\"SET statement_timeout = '30s'\");\n await client.query(\"BEGIN\");\n try {\n const r = await client.query(`EXPLAIN (ANALYZE, BUFFERS, FORMAT JSON) ${sql}`);\n await client.query(\"ROLLBACK\");\n await client.query(\"RESET statement_timeout\");\n\n const plan = r.rows[0][\"QUERY PLAN\"];\n const analysis = await analyzeExplainPlan(plan, pool);\n\n return {\n content: [{\n type: \"text\",\n text: JSON.stringify({ plan, analysis }, null, 2),\n }],\n };\n } catch (err: any) {\n await client.query(\"ROLLBACK\").catch(() => {});\n await client.query(\"RESET statement_timeout\").catch(() => {});\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n } finally {\n client.release();\n }\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_query_regressions\",\n \"Detect queries that have gotten significantly slower (>50% degradation) compared to historical baselines\",\n {\n windowHours: z.number().optional().describe(\"Hours to look back (default: 24)\"),\n },\n async ({ windowHours }) => {\n try {\n const regressions = await detectQueryRegressions(pool, queryStatsDb, windowHours ?? 24);\n if (regressions.length === 0) {\n return { content: [{ type: \"text\", text: \"No query regressions detected in the specified window.\" }] };\n }\n return { content: [{ type: \"text\", text: JSON.stringify(regressions, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\n \"pg_dash_compare_env\",\n \"Compare schema and health between two PostgreSQL environments. Detects missing tables, columns, indexes.\",\n {\n targetUrl: z.string().describe(\"Target database connection string to compare against\"),\n includeHealth: z.boolean().optional().describe(\"Also compare health scores and issues\"),\n },\n async ({ targetUrl, includeHealth }) => {\n try {\n // source uses the existing pool's connection string; target gets its own pool (created inside diffEnvironments)\n // We pass connString so the source pool is created fresh with a 10s timeout; existing pool is unaffected\n const result = await diffEnvironments(connString, targetUrl, { includeHealth: includeHealth ?? false });\n return { content: [{ type: \"text\", text: JSON.stringify(result, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n }\n);\n\nserver.tool(\"pg_dash_unused_indexes\", \"Find unused indexes that waste space and slow down writes\", {}, async () => {\n try {\n const report = await getUnusedIndexes(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_bloat\", \"Detect table bloat (dead tuples) that slow down queries\", {}, async () => {\n try {\n const report = await getBloatReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_autovacuum\", \"Check autovacuum health — which tables are stale or never vacuumed\", {}, async () => {\n try {\n const report = await getAutovacuumReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_locks\", \"Show active lock waits and long-running queries blocking the database\", {}, async () => {\n try {\n const report = await getLockReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nserver.tool(\"pg_dash_config_check\", \"Audit PostgreSQL configuration settings and get tuning recommendations\", {}, async () => {\n try {\n const report = await getConfigReport(pool);\n return { content: [{ type: \"text\", text: JSON.stringify(report, null, 2) }] };\n } catch (err: any) {\n return { content: [{ type: \"text\", text: `Error: ${err.message}` }], isError: true };\n }\n});\n\nconst transport = new StdioServerTransport();\nawait server.connect(transport);\n","import type { Pool } from \"pg\";\n\nexport async function getOverview(pool: Pool) {\n const client = await pool.connect();\n try {\n const version = await client.query(\"SHOW server_version\");\n const uptime = await client.query(\n \"SELECT to_char(now() - pg_postmaster_start_time(), 'DD \\\"d\\\" HH24 \\\"h\\\" MI \\\"m\\\"') AS uptime\"\n );\n const dbSize = await client.query(\n \"SELECT pg_size_pretty(pg_database_size(current_database())) AS size\"\n );\n const dbCount = await client.query(\n \"SELECT count(*)::int AS count FROM pg_database WHERE NOT datistemplate\"\n );\n const connections = await client.query(`\n SELECT\n (SELECT count(*)::int FROM pg_stat_activity WHERE state = 'active') AS active,\n (SELECT count(*)::int FROM pg_stat_activity WHERE state = 'idle') AS idle,\n (SELECT setting::int FROM pg_settings WHERE name = 'max_connections') AS max\n `);\n\n return {\n version: version.rows[0].server_version,\n uptime: uptime.rows[0].uptime,\n dbSize: dbSize.rows[0].size,\n databaseCount: dbCount.rows[0].count,\n connections: connections.rows[0],\n };\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport async function getTables(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n schemaname AS schema,\n relname AS name,\n pg_size_pretty(pg_total_relation_size(relid)) AS total_size,\n pg_total_relation_size(relid) AS size_bytes,\n n_live_tup AS rows,\n n_dead_tup AS dead_tuples,\n CASE WHEN n_live_tup > 0 \n THEN round(n_dead_tup::numeric / n_live_tup * 100, 1) \n ELSE 0 END AS dead_pct\n FROM pg_stat_user_tables\n ORDER BY pg_total_relation_size(relid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport async function getSchemaTables(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n c.relname AS name,\n n.nspname AS schema,\n pg_size_pretty(pg_total_relation_size(c.oid)) AS total_size,\n pg_total_relation_size(c.oid) AS total_size_bytes,\n pg_size_pretty(pg_relation_size(c.oid)) AS table_size,\n pg_size_pretty(pg_total_relation_size(c.oid) - pg_relation_size(c.oid)) AS index_size,\n s.n_live_tup AS row_count,\n obj_description(c.oid) AS description\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n LEFT JOIN pg_stat_user_tables s ON s.relid = c.oid\n WHERE c.relkind = 'r' AND n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY pg_total_relation_size(c.oid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaTableDetail(pool: Pool, tableName: string) {\n const client = await pool.connect();\n try {\n // Parse schema.table or default to public\n const parts = tableName.split(\".\");\n const schema = parts.length > 1 ? parts[0] : \"public\";\n const name = parts.length > 1 ? parts[1] : parts[0];\n\n // Table info\n const tableInfo = await client.query(`\n SELECT\n c.relname AS name, n.nspname AS schema,\n pg_size_pretty(pg_total_relation_size(c.oid)) AS total_size,\n pg_size_pretty(pg_relation_size(c.oid)) AS table_size,\n pg_size_pretty(pg_total_relation_size(c.oid) - pg_relation_size(c.oid)) AS index_size,\n pg_size_pretty(pg_relation_size(c.reltoastrelid)) AS toast_size,\n s.n_live_tup AS row_count, s.n_dead_tup AS dead_tuples,\n s.last_vacuum, s.last_autovacuum, s.last_analyze, s.last_autoanalyze,\n s.seq_scan, s.idx_scan\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n LEFT JOIN pg_stat_user_tables s ON s.relid = c.oid\n WHERE c.relname = $1 AND n.nspname = $2 AND c.relkind = 'r'\n `, [name, schema]);\n\n if (tableInfo.rows.length === 0) return null;\n\n // Columns\n const columns = await client.query(`\n SELECT\n a.attname AS name,\n pg_catalog.format_type(a.atttypid, a.atttypmod) AS type,\n NOT a.attnotnull AS nullable,\n pg_get_expr(d.adbin, d.adrelid) AS default_value,\n col_description(a.attrelid, a.attnum) AS description\n FROM pg_attribute a\n LEFT JOIN pg_attrdef d ON a.attrelid = d.adrelid AND a.attnum = d.adnum\n WHERE a.attrelid = (SELECT c.oid FROM pg_class c JOIN pg_namespace n ON c.relnamespace = n.oid WHERE c.relname = $1 AND n.nspname = $2)\n AND a.attnum > 0 AND NOT a.attisdropped\n ORDER BY a.attnum\n `, [name, schema]);\n\n // Indexes\n const indexes = await client.query(`\n SELECT\n i.relname AS name,\n am.amname AS type,\n pg_size_pretty(pg_relation_size(i.oid)) AS size,\n pg_get_indexdef(idx.indexrelid) AS definition,\n idx.indisunique AS is_unique,\n idx.indisprimary AS is_primary,\n s.idx_scan, s.idx_tup_read, s.idx_tup_fetch\n FROM pg_index idx\n JOIN pg_class i ON idx.indexrelid = i.oid\n JOIN pg_class t ON idx.indrelid = t.oid\n JOIN pg_namespace n ON t.relnamespace = n.oid\n JOIN pg_am am ON i.relam = am.oid\n LEFT JOIN pg_stat_user_indexes s ON s.indexrelid = i.oid\n WHERE t.relname = $1 AND n.nspname = $2\n ORDER BY i.relname\n `, [name, schema]);\n\n // Constraints\n const constraints = await client.query(`\n SELECT\n conname AS name,\n CASE contype WHEN 'p' THEN 'PRIMARY KEY' WHEN 'f' THEN 'FOREIGN KEY'\n WHEN 'u' THEN 'UNIQUE' WHEN 'c' THEN 'CHECK' WHEN 'x' THEN 'EXCLUDE' END AS type,\n pg_get_constraintdef(oid) AS definition\n FROM pg_constraint\n WHERE conrelid = (SELECT c.oid FROM pg_class c JOIN pg_namespace n ON c.relnamespace = n.oid WHERE c.relname = $1 AND n.nspname = $2)\n ORDER BY\n CASE contype WHEN 'p' THEN 1 WHEN 'u' THEN 2 WHEN 'f' THEN 3 WHEN 'c' THEN 4 ELSE 5 END\n `, [name, schema]);\n\n // Foreign keys (outgoing)\n const foreignKeys = await client.query(`\n SELECT\n conname AS name,\n a.attname AS column_name,\n confrelid::regclass::text AS referenced_table,\n af.attname AS referenced_column\n FROM pg_constraint c\n JOIN pg_attribute a ON a.attrelid = c.conrelid AND a.attnum = ANY(c.conkey)\n JOIN pg_attribute af ON af.attrelid = c.confrelid AND af.attnum = ANY(c.confkey)\n WHERE c.contype = 'f'\n AND c.conrelid = (SELECT cl.oid FROM pg_class cl JOIN pg_namespace n ON cl.relnamespace = n.oid WHERE cl.relname = $1 AND n.nspname = $2)\n `, [name, schema]);\n\n // Sample data (first 10 rows)\n let sampleData: any[] = [];\n try {\n const sample = await client.query(\n `SELECT * FROM ${client.escapeIdentifier(schema)}.${client.escapeIdentifier(name)} LIMIT 10`\n );\n sampleData = sample.rows;\n } catch (err) { console.error(\"[schema] Error:\", (err as Error).message); }\n\n return {\n ...tableInfo.rows[0],\n columns: columns.rows,\n indexes: indexes.rows,\n constraints: constraints.rows,\n foreignKeys: foreignKeys.rows,\n sampleData,\n };\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaIndexes(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n n.nspname AS schema,\n t.relname AS table_name,\n i.relname AS name,\n am.amname AS type,\n pg_size_pretty(pg_relation_size(i.oid)) AS size,\n pg_relation_size(i.oid) AS size_bytes,\n pg_get_indexdef(idx.indexrelid) AS definition,\n idx.indisunique AS is_unique,\n idx.indisprimary AS is_primary,\n s.idx_scan, s.idx_tup_read, s.idx_tup_fetch\n FROM pg_index idx\n JOIN pg_class i ON idx.indexrelid = i.oid\n JOIN pg_class t ON idx.indrelid = t.oid\n JOIN pg_namespace n ON t.relnamespace = n.oid\n JOIN pg_am am ON i.relam = am.oid\n LEFT JOIN pg_stat_user_indexes s ON s.indexrelid = i.oid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY pg_relation_size(i.oid) DESC\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaFunctions(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n n.nspname AS schema,\n p.proname AS name,\n pg_get_function_result(p.oid) AS return_type,\n pg_get_function_arguments(p.oid) AS arguments,\n l.lanname AS language,\n p.prosrc AS source,\n CASE p.prokind WHEN 'f' THEN 'function' WHEN 'p' THEN 'procedure' WHEN 'a' THEN 'aggregate' WHEN 'w' THEN 'window' END AS kind\n FROM pg_proc p\n JOIN pg_namespace n ON p.pronamespace = n.oid\n JOIN pg_language l ON p.prolang = l.oid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n ORDER BY n.nspname, p.proname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaExtensions(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT extname AS name, extversion AS installed_version,\n n.nspname AS schema, obj_description(e.oid) AS description\n FROM pg_extension e\n JOIN pg_namespace n ON e.extnamespace = n.oid\n ORDER BY extname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n\nexport async function getSchemaEnums(pool: Pool) {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n t.typname AS name,\n n.nspname AS schema,\n array_agg(e.enumlabel ORDER BY e.enumsortorder) AS values\n FROM pg_type t\n JOIN pg_namespace n ON t.typnamespace = n.oid\n JOIN pg_enum e ON t.oid = e.enumtypid\n WHERE n.nspname NOT IN ('pg_catalog', 'information_schema')\n GROUP BY t.typname, n.nspname\n ORDER BY t.typname\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport interface Activity {\n pid: number;\n query: string;\n state: string;\n wait_event: string | null;\n wait_event_type: string | null;\n duration: string | null;\n client_addr: string | null;\n application_name: string;\n backend_start: string;\n}\n\nexport async function getActivity(pool: Pool): Promise<Activity[]> {\n const client = await pool.connect();\n try {\n const r = await client.query(`\n SELECT\n pid,\n COALESCE(query, '') AS query,\n COALESCE(state, 'unknown') AS state,\n wait_event,\n wait_event_type,\n CASE WHEN state = 'active' THEN (now() - query_start)::text\n WHEN state = 'idle in transaction' THEN (now() - state_change)::text\n ELSE NULL END AS duration,\n client_addr::text,\n COALESCE(application_name, '') AS application_name,\n backend_start::text\n FROM pg_stat_activity\n WHERE pid != pg_backend_pid()\n AND state IS NOT NULL\n ORDER BY\n CASE state\n WHEN 'active' THEN 1\n WHEN 'idle in transaction' THEN 2\n ELSE 3\n END,\n query_start ASC NULLS LAST\n `);\n return r.rows;\n } finally {\n client.release();\n }\n}\n","import type { Pool } from \"pg\";\n\nexport interface AdvisorIssue {\n id: string;\n severity: \"critical\" | \"warning\" | \"info\";\n category: \"performance\" | \"maintenance\" | \"schema\" | \"security\";\n title: string;\n description: string;\n fix: string;\n impact: string;\n effort: \"quick\" | \"moderate\" | \"involved\";\n}\n\nexport interface BatchFix {\n type: string;\n title: string;\n count: number;\n sql: string;\n}\n\nexport interface AdvisorResult {\n score: number;\n grade: string;\n issues: AdvisorIssue[];\n breakdown: Record<string, { score: number; grade: string; count: number }>;\n skipped: string[];\n ignoredCount: number;\n batchFixes: BatchFix[];\n}\n\nconst SEVERITY_WEIGHT = { critical: 15, warning: 5, info: 1 } as const;\nconst MAX_DEDUCTION = { critical: 60, warning: 30, info: 10 } as const;\n\nexport function computeAdvisorScore(issues: AdvisorIssue[]): number {\n let score = 100;\n const deductions = { critical: 0, warning: 0, info: 0 };\n const counts = { critical: 0, warning: 0, info: 0 };\n for (const issue of issues) {\n counts[issue.severity]++;\n const n = counts[issue.severity];\n const weight = SEVERITY_WEIGHT[issue.severity];\n // Diminishing penalty: full for first 3, half for 4-10, quarter for 11+\n let penalty: number;\n if (n <= 3) penalty = weight;\n else if (n <= 10) penalty = weight * 0.5;\n else penalty = weight * 0.25;\n deductions[issue.severity] += penalty;\n }\n // Cap deductions per severity\n for (const sev of [\"critical\", \"warning\", \"info\"] as const) {\n score -= Math.min(deductions[sev], MAX_DEDUCTION[sev]);\n }\n return Math.max(0, Math.min(100, Math.round(score)));\n}\n\nexport function gradeFromScore(score: number): string {\n if (score >= 90) return \"A\";\n if (score >= 80) return \"B\";\n if (score >= 70) return \"C\";\n if (score >= 50) return \"D\";\n return \"F\";\n}\n\nfunction computeBreakdown(issues: AdvisorIssue[]): Record<string, { score: number; grade: string; count: number }> {\n const categories = [\"performance\", \"maintenance\", \"schema\", \"security\"] as const;\n const result: Record<string, { score: number; grade: string; count: number }> = {};\n for (const cat of categories) {\n const catIssues = issues.filter((i) => i.category === cat);\n const score = computeAdvisorScore(catIssues);\n result[cat] = { score, grade: gradeFromScore(score), count: catIssues.length };\n }\n return result;\n}\n\nexport async function getAdvisorReport(pool: Pool, longQueryThreshold = 5): Promise<AdvisorResult> {\n const client = await pool.connect();\n const issues: AdvisorIssue[] = [];\n const skipped: string[] = [];\n\n try {\n // Detect PG version for compatibility\n const versionResult = await client.query(\"SHOW server_version_num\");\n const pgVersion = parseInt(versionResult.rows[0].server_version_num);\n\n // ── Performance Advisors ───────────────────────────────────────\n\n // Missing indexes (high seq scans on large tables)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, seq_scan, seq_tup_read, n_live_tup,\n pg_size_pretty(pg_total_relation_size(relid)) AS size\n FROM pg_stat_user_tables\n WHERE n_live_tup > 10000 AND seq_scan > 100\n ORDER BY seq_tup_read DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-seq-scan-${row.schemaname}-${row.relname}`,\n severity: row.seq_scan > 1000 ? \"warning\" : \"info\",\n category: \"performance\",\n title: `High sequential scans on ${row.relname}`,\n description: `Table ${row.schemaname}.${row.relname} (${row.n_live_tup} rows, ${row.size}) has ${row.seq_scan} sequential scans reading ${Number(row.seq_tup_read).toLocaleString()} tuples. Consider adding indexes on frequently filtered columns.`,\n fix: `-- Identify commonly filtered columns and add indexes:\\n-- EXPLAIN ANALYZE SELECT * FROM ${row.schemaname}.${row.relname} WHERE <your_condition>;\\nCREATE INDEX CONCURRENTLY idx_${row.relname}_<column> ON ${row.schemaname}.${row.relname} (<column>);`,\n impact: \"Queries will continue to do full table scans, degrading performance as the table grows.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking seq scans:\", (err as Error).message); skipped.push(\"seq scans: \" + (err as Error).message);\n }\n\n // Bloated indexes (index size > 3x table size)\n try {\n const r = await client.query(`\n SELECT\n schemaname, relname, indexrelname,\n pg_relation_size(indexrelid) AS idx_size,\n pg_relation_size(relid) AS tbl_size,\n pg_size_pretty(pg_relation_size(indexrelid)) AS idx_size_pretty,\n pg_size_pretty(pg_relation_size(relid)) AS tbl_size_pretty\n FROM pg_stat_user_indexes\n WHERE pg_relation_size(indexrelid) > 1048576\n AND pg_relation_size(indexrelid) > pg_relation_size(relid) * 3\n ORDER BY pg_relation_size(indexrelid) DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-bloated-idx-${row.indexrelname}`,\n severity: \"warning\",\n category: \"performance\",\n title: `Bloated index ${row.indexrelname}`,\n description: `Index ${row.indexrelname} on ${row.relname} is ${row.idx_size_pretty} but the table is only ${row.tbl_size_pretty}. The index may need rebuilding.`,\n fix: `REINDEX INDEX CONCURRENTLY ${row.schemaname}.${row.indexrelname};`,\n impact: \"Bloated indexes waste disk space and slow down queries that use them.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking bloated indexes:\", (err as Error).message); skipped.push(\"bloated indexes: \" + (err as Error).message);\n }\n\n // Table bloat (dead tuples > 10%)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, n_dead_tup, n_live_tup,\n CASE WHEN n_live_tup > 0 THEN round(n_dead_tup::numeric / n_live_tup * 100, 1) ELSE 0 END AS dead_pct,\n pg_size_pretty(pg_total_relation_size(relid)) AS size\n FROM pg_stat_user_tables\n WHERE n_live_tup > 1000 AND n_dead_tup::float / GREATEST(n_live_tup, 1) > 0.1\n ORDER BY n_dead_tup DESC LIMIT 10\n `);\n for (const row of r.rows) {\n const pct = parseFloat(row.dead_pct);\n issues.push({\n id: `perf-bloat-${row.schemaname}-${row.relname}`,\n severity: pct > 30 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Table bloat on ${row.relname} (${row.dead_pct}% dead)`,\n description: `${row.schemaname}.${row.relname} has ${Number(row.n_dead_tup).toLocaleString()} dead tuples (${row.dead_pct}% of ${Number(row.n_live_tup).toLocaleString()} live rows). Size: ${row.size}.`,\n fix: `VACUUM FULL ${row.schemaname}.${row.relname};`,\n impact: \"Dead tuples waste storage and degrade scan performance.\",\n effort: pct > 30 ? \"moderate\" : \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking table bloat:\", (err as Error).message); skipped.push(\"table bloat: \" + (err as Error).message);\n }\n\n // Cache efficiency per table\n try {\n const r = await client.query(`\n SELECT schemaname, relname,\n heap_blks_hit, heap_blks_read,\n CASE WHEN (heap_blks_hit + heap_blks_read) = 0 THEN 1\n ELSE heap_blks_hit::float / (heap_blks_hit + heap_blks_read) END AS ratio\n FROM pg_statio_user_tables\n WHERE (heap_blks_hit + heap_blks_read) > 100\n ORDER BY ratio ASC LIMIT 5\n `);\n for (const row of r.rows) {\n const ratio = parseFloat(row.ratio);\n if (ratio < 0.9) {\n issues.push({\n id: `perf-cache-${row.schemaname}-${row.relname}`,\n severity: ratio < 0.5 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Poor cache hit ratio on ${row.relname}`,\n description: `Table ${row.schemaname}.${row.relname} has a cache hit ratio of ${(ratio * 100).toFixed(1)}%. Most reads are going to disk.`,\n fix: `-- Consider increasing shared_buffers or reducing working set:\\nSHOW shared_buffers;`,\n impact: \"Disk reads are orders of magnitude slower than memory reads.\",\n effort: \"involved\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking cache efficiency:\", (err as Error).message); skipped.push(\"cache efficiency: \" + (err as Error).message);\n }\n\n // Slow queries from pg_stat_statements\n try {\n const extCheck = await client.query(\"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\");\n if (extCheck.rows.length > 0) {\n const r = await client.query(`\n SELECT query, calls, mean_exec_time, total_exec_time,\n round(mean_exec_time::numeric, 2) AS mean_ms,\n round(total_exec_time::numeric / 1000, 2) AS total_sec\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%' AND query NOT LIKE '%pg_catalog%'\n AND mean_exec_time > 100\n ORDER BY mean_exec_time DESC LIMIT 5\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-slow-${row.query.slice(0, 30).replace(/\\W/g, \"_\")}`,\n severity: parseFloat(row.mean_ms) > 1000 ? \"warning\" : \"info\",\n category: \"performance\",\n title: `Slow query (avg ${row.mean_ms}ms)`,\n description: `Query averaging ${row.mean_ms}ms over ${row.calls} calls (total: ${row.total_sec}s): ${row.query.slice(0, 200)}`,\n fix: `EXPLAIN ANALYZE ${row.query.slice(0, 500)};`,\n impact: \"Slow queries degrade overall database responsiveness.\",\n effort: \"moderate\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking slow queries:\", (err as Error).message); skipped.push(\"slow queries: \" + (err as Error).message);\n }\n\n // ── Maintenance Advisors ───────────────────────────────────────\n\n // VACUUM overdue\n try {\n const r = await client.query(`\n SELECT schemaname, relname, last_vacuum, last_autovacuum, n_dead_tup\n FROM pg_stat_user_tables\n WHERE n_live_tup > 100\n AND (last_vacuum IS NULL AND last_autovacuum IS NULL\n OR GREATEST(last_vacuum, last_autovacuum) < now() - interval '7 days')\n ORDER BY n_dead_tup DESC LIMIT 15\n `);\n for (const row of r.rows) {\n const never = !row.last_vacuum && !row.last_autovacuum;\n issues.push({\n id: `maint-vacuum-${row.schemaname}-${row.relname}`,\n severity: never ? \"warning\" : \"info\",\n category: \"maintenance\",\n title: `VACUUM ${never ? \"never run\" : \"overdue\"} on ${row.relname}`,\n description: `${row.schemaname}.${row.relname} ${never ? \"has never been vacuumed\" : \"was last vacuumed over 7 days ago\"}. Dead tuples: ${Number(row.n_dead_tup).toLocaleString()}.`,\n fix: `VACUUM ANALYZE ${row.schemaname}.${row.relname};`,\n impact: \"Dead tuples accumulate, increasing table size and degrading query performance.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking vacuum overdue:\", (err as Error).message); skipped.push(\"vacuum overdue: \" + (err as Error).message);\n }\n\n // ANALYZE overdue\n try {\n const r = await client.query(`\n SELECT schemaname, relname\n FROM pg_stat_user_tables\n WHERE n_live_tup > 100\n AND last_analyze IS NULL AND last_autoanalyze IS NULL\n AND NOT EXISTS (\n SELECT 1 FROM pg_stat_user_tables t2\n WHERE t2.relname = pg_stat_user_tables.relname\n AND (t2.last_vacuum IS NULL AND t2.last_autovacuum IS NULL)\n )\n LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `maint-analyze-${row.schemaname}-${row.relname}`,\n severity: \"info\",\n category: \"maintenance\",\n title: `ANALYZE never run on ${row.relname}`,\n description: `${row.schemaname}.${row.relname} has never been analyzed. The query planner may choose suboptimal plans.`,\n fix: `ANALYZE ${row.schemaname}.${row.relname};`,\n impact: \"Without statistics, the query planner makes poor estimates leading to slow queries.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking analyze overdue:\", (err as Error).message); skipped.push(\"analyze overdue: \" + (err as Error).message);\n }\n\n // Transaction ID wraparound risk\n try {\n const r = await client.query(`\n SELECT datname, age(datfrozenxid) AS xid_age\n FROM pg_database\n WHERE datname = current_database()\n `);\n for (const row of r.rows) {\n const age = parseInt(row.xid_age);\n if (age > 1_000_000_000) {\n issues.push({\n id: `maint-xid-wraparound`,\n severity: \"critical\",\n category: \"maintenance\",\n title: `Transaction ID wraparound risk`,\n description: `Database ${row.datname} has datfrozenxid age of ${age.toLocaleString()}. Wraparound occurs at ~2 billion.`,\n fix: `VACUUM FREEZE;`,\n impact: \"If wraparound occurs, PostgreSQL will shut down to prevent data loss.\",\n effort: \"involved\",\n });\n } else if (age > 500_000_000) {\n issues.push({\n id: `maint-xid-warning`,\n severity: \"warning\",\n category: \"maintenance\",\n title: `Transaction ID age is high`,\n description: `Database ${row.datname} has datfrozenxid age of ${age.toLocaleString()}.`,\n fix: `VACUUM FREEZE;`,\n impact: \"Approaching transaction ID wraparound threshold.\",\n effort: \"moderate\",\n });\n }\n }\n } catch (err) {\n console.error(\"[advisor] Error checking xid wraparound:\", (err as Error).message); skipped.push(\"xid wraparound: \" + (err as Error).message);\n }\n\n // Idle-in-transaction connections (genuine problem: hold locks, block VACUUM)\n // Note: plain 'idle' connections are normal connection pool behavior and are NOT flagged.\n try {\n const r = await client.query(`\n SELECT pid,\n client_addr::text, application_name,\n extract(epoch from now() - state_change)::int AS idle_seconds\n FROM pg_stat_activity\n WHERE state = 'idle in transaction'\n AND now() - state_change > $1 * interval '1 minute'\n AND pid != pg_backend_pid()\n ORDER BY idle_seconds DESC\n `, [longQueryThreshold]);\n if (r.rows.length === 1) {\n const row = r.rows[0];\n issues.push({\n id: `maint-idle-tx-${row.pid}`,\n severity: \"warning\",\n category: \"maintenance\",\n title: `Idle-in-transaction connection (PID ${row.pid})`,\n description: `PID ${row.pid} from ${row.client_addr || \"local\"} (${row.application_name || \"unknown\"}) has been idle in transaction for ${Math.round(row.idle_seconds / 60)} minutes. This holds locks and blocks VACUUM.`,\n fix: `SELECT pg_terminate_backend(${row.pid});`,\n impact: \"Idle-in-transaction connections hold locks and prevent VACUUM.\",\n effort: \"quick\",\n });\n } else if (r.rows.length > 1) {\n const pids = r.rows.map((row: any) => row.pid);\n const maxMin = Math.round(r.rows[0].idle_seconds / 60);\n issues.push({\n id: `maint-idle-tx-multi`,\n severity: \"warning\",\n category: \"maintenance\",\n title: `${r.rows.length} idle-in-transaction connections (longest: ${maxMin}m)`,\n description: `${r.rows.length} connections have been idle in transaction for over ${longQueryThreshold} minutes. These hold locks and prevent VACUUM from running.`,\n fix: `SELECT pg_terminate_backend(pid) FROM pg_stat_activity WHERE state = 'idle in transaction' AND now() - state_change > interval '${longQueryThreshold} minutes';`,\n impact: \"Idle-in-transaction connections hold locks and prevent VACUUM.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking idle-in-transaction connections:\", (err as Error).message); skipped.push(\"idle-in-transaction: \" + (err as Error).message);\n }\n\n // ── Schema Advisors ────────────────────────────────────────────\n\n // Missing primary keys\n try {\n const r = await client.query(`\n SELECT c.relname AS table_name, n.nspname AS schema\n FROM pg_class c\n JOIN pg_namespace n ON c.relnamespace = n.oid\n WHERE c.relkind = 'r' AND n.nspname = 'public'\n AND NOT EXISTS (\n SELECT 1 FROM pg_constraint con WHERE con.conrelid = c.oid AND con.contype = 'p'\n )\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-no-pk-${row.schema}-${row.table_name}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Missing primary key on ${row.table_name}`,\n description: `Table ${row.schema}.${row.table_name} has no primary key. This can cause replication issues and makes row identification unreliable.`,\n fix: `ALTER TABLE ${row.schema}.${row.table_name} ADD PRIMARY KEY (<column>);`,\n impact: \"No primary key means no unique row identity, problematic for replication and ORMs.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking missing primary keys:\", (err as Error).message); skipped.push(\"missing primary keys: \" + (err as Error).message);\n }\n\n // Unused indexes (idx_scan = 0, size > 1MB)\n try {\n const r = await client.query(`\n SELECT schemaname, relname, indexrelname, idx_scan,\n pg_size_pretty(pg_relation_size(indexrelid)) AS idx_size,\n pg_relation_size(indexrelid) AS idx_bytes\n FROM pg_stat_user_indexes\n WHERE idx_scan = 0\n AND indexrelname NOT LIKE '%_pkey'\n AND pg_relation_size(indexrelid) > 1048576\n ORDER BY pg_relation_size(indexrelid) DESC LIMIT 10\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-unused-idx-${row.indexrelname}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Unused index ${row.indexrelname} (${row.idx_size})`,\n description: `Index ${row.indexrelname} on ${row.relname} has never been used (0 scans) and takes ${row.idx_size}.`,\n fix: `DROP INDEX CONCURRENTLY ${row.schemaname}.${row.indexrelname};`,\n impact: \"Unused indexes waste disk space and slow down writes.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking unused indexes:\", (err as Error).message); skipped.push(\"unused indexes: \" + (err as Error).message);\n }\n\n // Duplicate indexes\n try {\n const r = await client.query(`\n SELECT array_agg(idx.indexrelid::regclass::text) AS indexes,\n idx.indrelid::regclass::text AS table_name,\n pg_size_pretty(sum(pg_relation_size(idx.indexrelid))) AS total_size\n FROM pg_index idx\n GROUP BY idx.indrelid, idx.indkey\n HAVING count(*) > 1\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-dup-idx-${row.table_name}-${row.indexes[0]}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Duplicate indexes on ${row.table_name}`,\n description: `These indexes cover the same columns on ${row.table_name}: ${row.indexes.join(\", \")}. Total wasted space: ${row.total_size}.`,\n fix: `-- Keep one, drop the rest:\\nDROP INDEX CONCURRENTLY ${row.indexes.slice(1).join(\";\\nDROP INDEX CONCURRENTLY \")};`,\n impact: \"Duplicate indexes double the write overhead and waste disk space.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking duplicate indexes:\", (err as Error).message); skipped.push(\"duplicate indexes: \" + (err as Error).message);\n }\n\n // Missing foreign key indexes\n try {\n const r = await client.query(`\n SELECT\n conrelid::regclass::text AS table_name,\n a.attname AS column_name,\n confrelid::regclass::text AS referenced_table\n FROM pg_constraint c\n JOIN pg_attribute a ON a.attrelid = c.conrelid AND a.attnum = ANY(c.conkey)\n WHERE c.contype = 'f'\n AND NOT EXISTS (\n SELECT 1 FROM pg_index i\n WHERE i.indrelid = c.conrelid\n AND a.attnum = ANY(i.indkey)\n )\n `);\n for (const row of r.rows) {\n issues.push({\n id: `schema-fk-no-idx-${row.table_name}-${row.column_name}`,\n severity: \"warning\",\n category: \"schema\",\n title: `Missing index on FK column ${row.table_name}.${row.column_name}`,\n description: `Foreign key column ${row.column_name} on ${row.table_name} (references ${row.referenced_table}) has no index. This causes slow JOINs and cascading deletes.`,\n fix: `CREATE INDEX CONCURRENTLY idx_${row.table_name.replace(/\\./g, \"_\")}_${row.column_name} ON ${row.table_name} (${row.column_name});`,\n impact: \"JOINs and cascading deletes on this FK will require full table scans.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking missing FK indexes:\", (err as Error).message); skipped.push(\"missing FK indexes: \" + (err as Error).message);\n }\n\n // ── Infrastructure Advisors ──────────────────────────────────────\n\n // Lock detection\n try {\n const r = await client.query(`\n SELECT blocked_locks.pid AS blocked_pid,\n blocking_locks.pid AS blocking_pid,\n blocked_activity.query AS blocked_query\n FROM pg_catalog.pg_locks blocked_locks\n JOIN pg_catalog.pg_locks blocking_locks ON blocking_locks.locktype = blocked_locks.locktype\n AND blocking_locks.database IS NOT DISTINCT FROM blocked_locks.database\n AND blocking_locks.relation IS NOT DISTINCT FROM blocked_locks.relation\n AND blocking_locks.page IS NOT DISTINCT FROM blocked_locks.page\n AND blocking_locks.tuple IS NOT DISTINCT FROM blocked_locks.tuple\n AND blocking_locks.virtualxid IS NOT DISTINCT FROM blocked_locks.virtualxid\n AND blocking_locks.transactionid IS NOT DISTINCT FROM blocked_locks.transactionid\n AND blocking_locks.classid IS NOT DISTINCT FROM blocked_locks.classid\n AND blocking_locks.objid IS NOT DISTINCT FROM blocked_locks.objid\n AND blocking_locks.objsubid IS NOT DISTINCT FROM blocked_locks.objsubid\n AND blocking_locks.pid != blocked_locks.pid\n JOIN pg_catalog.pg_stat_activity blocked_activity ON blocked_activity.pid = blocked_locks.pid\n WHERE NOT blocked_locks.granted\n `);\n for (const row of r.rows) {\n issues.push({\n id: `perf-lock-blocked-${row.blocked_pid}`,\n severity: \"warning\",\n category: \"performance\",\n title: `Blocked query (PID ${row.blocked_pid} blocked by PID ${row.blocking_pid})`,\n description: `PID ${row.blocked_pid} is waiting for a lock held by PID ${row.blocking_pid}. Query: ${(row.blocked_query || \"\").slice(0, 200)}`,\n fix: `SELECT pg_cancel_backend(${row.blocking_pid});`,\n impact: \"Blocked queries cause cascading delays and potential timeouts.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking locks:\", (err as Error).message); skipped.push(\"locks: \" + (err as Error).message);\n }\n\n // WAL/replication lag\n try {\n const r = await client.query(`\n SELECT CASE WHEN pg_is_in_recovery()\n THEN pg_wal_lsn_diff(pg_last_wal_receive_lsn(), pg_last_wal_replay_lsn())\n ELSE 0 END AS lag_bytes\n `);\n const lagBytes = parseInt(r.rows[0]?.lag_bytes ?? \"0\");\n if (lagBytes > 1048576) { // > 1MB\n issues.push({\n id: `perf-replication-lag`,\n severity: lagBytes > 104857600 ? \"critical\" : \"warning\",\n category: \"performance\",\n title: `Replication lag: ${(lagBytes / 1048576).toFixed(1)} MB`,\n description: `WAL replay is lagging by ${(lagBytes / 1048576).toFixed(1)} MB. This indicates the replica is falling behind.`,\n fix: `-- Check replication status:\\nSELECT * FROM pg_stat_replication;`,\n impact: \"High replication lag means the replica has stale data and failover may lose transactions.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking replication lag:\", (err as Error).message); skipped.push(\"replication lag: \" + (err as Error).message);\n }\n\n // Checkpoint frequency\n try {\n const checkpointView = pgVersion >= 170000 ? 'pg_stat_checkpointer' : 'pg_stat_bgwriter';\n const r = await client.query(`\n SELECT checkpoints_req, checkpoints_timed,\n CASE WHEN (checkpoints_req + checkpoints_timed) = 0 THEN 0\n ELSE round(checkpoints_req::numeric / (checkpoints_req + checkpoints_timed) * 100, 1) END AS req_pct\n FROM ${checkpointView}\n `);\n const reqPct = parseFloat(r.rows[0]?.req_pct ?? \"0\");\n if (reqPct > 50) {\n issues.push({\n id: `maint-checkpoint-frequency`,\n severity: reqPct > 80 ? \"warning\" : \"info\",\n category: \"maintenance\",\n title: `${reqPct}% of checkpoints are requested (not timed)`,\n description: `${r.rows[0]?.checkpoints_req} requested vs ${r.rows[0]?.checkpoints_timed} timed checkpoints. High requested checkpoints indicate checkpoint_completion_target or max_wal_size may need tuning.`,\n fix: `-- Increase max_wal_size:\\nALTER SYSTEM SET max_wal_size = '2GB';\\nSELECT pg_reload_conf();`,\n impact: \"Frequent requested checkpoints cause I/O spikes and degrade performance.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking checkpoint frequency:\", (err as Error).message); skipped.push(\"checkpoint frequency: \" + (err as Error).message);\n }\n\n // AutoVACUUM config check\n try {\n const r = await client.query(`SELECT setting FROM pg_settings WHERE name = 'autovacuum'`);\n if (r.rows[0]?.setting === \"off\") {\n issues.push({\n id: `maint-autovacuum-disabled`,\n severity: \"critical\",\n category: \"maintenance\",\n title: `Autovacuum is disabled`,\n description: `Autovacuum is turned off. Dead tuples will accumulate and transaction ID wraparound becomes a risk.`,\n fix: `ALTER SYSTEM SET autovacuum = on;\\nSELECT pg_reload_conf();`,\n impact: \"Without autovacuum, tables bloat indefinitely and risk transaction ID wraparound shutdown.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking autovacuum:\", (err as Error).message); skipped.push(\"autovacuum: \" + (err as Error).message);\n }\n\n // shared_buffers / work_mem check\n try {\n const sbRes = await client.query(`SELECT setting, unit FROM pg_settings WHERE name = 'shared_buffers'`);\n const memRes = await client.query(`\n SELECT (SELECT setting::bigint FROM pg_settings WHERE name = 'shared_buffers') *\n (SELECT setting::bigint FROM pg_settings WHERE name = 'block_size') AS shared_bytes\n `);\n const sharedBytes = parseInt(memRes.rows[0]?.shared_bytes ?? \"0\");\n // Get total RAM from OS via a simple query (pg doesn't expose this directly, but we can estimate)\n // We'll compare against a reasonable minimum: if shared_buffers < 128MB, warn\n if (sharedBytes > 0 && sharedBytes < 128 * 1024 * 1024) {\n issues.push({\n id: `perf-shared-buffers-low`,\n severity: \"warning\",\n category: \"performance\",\n title: `shared_buffers is only ${(sharedBytes / 1048576).toFixed(0)} MB`,\n description: `shared_buffers is set to ${sbRes.rows[0]?.setting}${sbRes.rows[0]?.unit || \"\"}. Recommended: ~25% of system RAM, typically at least 256MB for production.`,\n fix: `ALTER SYSTEM SET shared_buffers = '256MB';\\n-- Requires restart`,\n impact: \"Low shared_buffers means more disk I/O and poor cache hit ratios.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking shared_buffers:\", (err as Error).message); skipped.push(\"shared_buffers: \" + (err as Error).message);\n }\n\n try {\n const r = await client.query(`SELECT setting, unit FROM pg_settings WHERE name = 'work_mem'`);\n const workMemKB = parseInt(r.rows[0]?.setting ?? \"0\");\n if (workMemKB > 0 && workMemKB < 4096) { // < 4MB\n issues.push({\n id: `perf-work-mem-low`,\n severity: \"info\",\n category: \"performance\",\n title: `work_mem is only ${workMemKB < 1024 ? workMemKB + \"kB\" : (workMemKB / 1024).toFixed(0) + \"MB\"}`,\n description: `work_mem is ${r.rows[0]?.setting}${r.rows[0]?.unit || \"\"}. Low work_mem causes sorts and hash operations to spill to disk.`,\n fix: `ALTER SYSTEM SET work_mem = '16MB';\\nSELECT pg_reload_conf();`,\n impact: \"Operations that exceed work_mem use temporary disk files, which is much slower.\",\n effort: \"quick\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking work_mem:\", (err as Error).message); skipped.push(\"work_mem: \" + (err as Error).message);\n }\n\n // ── Security Advisors ──────────────────────────────────────────\n\n // Superuser connections from non-localhost\n try {\n const r = await client.query(`\n SELECT pid, usename, client_addr::text\n FROM pg_stat_activity\n WHERE usename IN (SELECT rolname FROM pg_roles WHERE rolsuper)\n AND client_addr IS NOT NULL\n AND client_addr::text NOT IN ('127.0.0.1', '::1')\n AND pid != pg_backend_pid()\n `);\n for (const row of r.rows) {\n issues.push({\n id: `sec-superuser-remote-${row.pid}`,\n severity: \"critical\",\n category: \"security\",\n title: `Superuser ${row.usename} connected from ${row.client_addr}`,\n description: `Superuser ${row.usename} has an active connection from non-localhost address ${row.client_addr}. This is a security risk.`,\n fix: `-- Restrict superuser access in pg_hba.conf to localhost only.\\n-- Then: SELECT pg_reload_conf();`,\n impact: \"Remote superuser access is a significant security vulnerability.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking superuser connections:\", (err as Error).message); skipped.push(\"superuser connections: \" + (err as Error).message);\n }\n\n // SSL disabled\n try {\n const r = await client.query(`SELECT setting FROM pg_settings WHERE name = 'ssl'`);\n if (r.rows[0]?.setting === \"off\") {\n issues.push({\n id: `sec-ssl-off`,\n severity: \"warning\",\n category: \"security\",\n title: `SSL is disabled`,\n description: `SSL is turned off. Database connections are not encrypted.`,\n fix: `-- Enable SSL in postgresql.conf:\\n-- ssl = on\\n-- ssl_cert_file = 'server.crt'\\n-- ssl_key_file = 'server.key'\\nSELECT pg_reload_conf();`,\n impact: \"Database traffic can be intercepted and read in transit.\",\n effort: \"involved\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking SSL check:\", (err as Error).message); skipped.push(\"SSL check: \" + (err as Error).message);\n }\n\n // Password authentication check (PG 15+)\n try {\n const r = await client.query(`\n SELECT type, database, user_name, auth_method\n FROM pg_hba_file_rules\n WHERE auth_method = 'trust' AND type != 'local'\n LIMIT 5\n `);\n for (const row of r.rows) {\n issues.push({\n id: `sec-trust-auth-${row.database}-${row.user_name}`,\n severity: \"critical\",\n category: \"security\",\n title: `Trust authentication for ${row.user_name}@${row.database}`,\n description: `HBA rule allows trust (no password) authentication for ${row.type} connections to ${row.database} as ${row.user_name}.`,\n fix: `-- Change auth_method from 'trust' to 'scram-sha-256' in pg_hba.conf\\n-- Then: SELECT pg_reload_conf();`,\n impact: \"Anyone can connect without a password.\",\n effort: \"moderate\",\n });\n }\n } catch (err) {\n console.error(\"[advisor] Error checking trust auth:\", (err as Error).message); skipped.push(\"trust auth: \" + (err as Error).message);\n } // pg_hba_file_rules not available pre-PG15\n\n // Filter out ignored issues\n const ignoredIds = getIgnoredIssues();\n const ignoredSet = new Set(ignoredIds);\n const activeIssues = issues.filter(i => !ignoredSet.has(i.id));\n const ignoredCount = issues.length - activeIssues.length;\n\n // Generate batch fixes for groups of same-type issues\n const batchFixes: BatchFix[] = [];\n const groups = new Map<string, AdvisorIssue[]>();\n for (const issue of activeIssues) {\n // Group by id prefix (everything before the last dash-separated segment with variable data)\n const prefix = issue.id.replace(/-[^-]+$/, \"\");\n if (!groups.has(prefix)) groups.set(prefix, []);\n groups.get(prefix)!.push(issue);\n }\n const BATCH_TITLES: Record<string, string> = {\n \"schema-fk-no-idx\": \"Create all missing FK indexes\",\n \"schema-unused-idx\": \"Drop all unused indexes\",\n \"schema-no-pk\": \"Fix all tables missing primary keys\",\n \"maint-vacuum\": \"VACUUM all overdue tables\",\n \"maint-analyze\": \"ANALYZE all tables missing statistics\",\n \"perf-bloated-idx\": \"REINDEX all bloated indexes\",\n \"perf-bloat\": \"VACUUM FULL all bloated tables\",\n };\n for (const [prefix, group] of groups) {\n if (group.length <= 1) continue;\n const title = BATCH_TITLES[prefix] || `Fix all ${group.length} ${prefix} issues`;\n const sql = group.map(i => i.fix.split(\"\\n\").filter(l => !l.trim().startsWith(\"--\")).join(\"\\n\").trim()).filter(Boolean).join(\";\\n\") + \";\";\n batchFixes.push({ type: prefix, title: `${title} (${group.length})`, count: group.length, sql });\n }\n\n const score = computeAdvisorScore(activeIssues);\n return {\n score,\n grade: gradeFromScore(score),\n issues: activeIssues,\n breakdown: computeBreakdown(activeIssues),\n skipped,\n ignoredCount,\n batchFixes,\n };\n } finally {\n client.release();\n }\n}\n\n// ── Ignored Issues Management ──────────────────────────────────\n\nimport Database from \"better-sqlite3\";\nimport path from \"node:path\";\nimport os from \"node:os\";\nimport fs from \"node:fs\";\n\nlet _ignoredDb: ReturnType<typeof Database> | null = null;\n\nfunction getIgnoredDb(): ReturnType<typeof Database> {\n if (_ignoredDb) return _ignoredDb;\n const dataDir = process.env.PG_DASH_DATA_DIR || path.join(os.homedir(), \".pg-dash\");\n fs.mkdirSync(dataDir, { recursive: true });\n const dbPath = path.join(dataDir, \"alerts.db\");\n _ignoredDb = new Database(dbPath);\n _ignoredDb.pragma(\"journal_mode = WAL\");\n _ignoredDb.exec(\"CREATE TABLE IF NOT EXISTS ignored_issues (issue_id TEXT PRIMARY KEY, ignored_at INTEGER)\");\n return _ignoredDb;\n}\n\nexport function getIgnoredIssues(): string[] {\n try {\n const db = getIgnoredDb();\n return db.prepare(\"SELECT issue_id FROM ignored_issues\").all().map((r: any) => r.issue_id);\n } catch {\n return [];\n }\n}\n\nexport function ignoreIssue(issueId: string): void {\n const db = getIgnoredDb();\n db.prepare(\"INSERT OR REPLACE INTO ignored_issues (issue_id, ignored_at) VALUES (?, ?)\").run(issueId, Date.now());\n}\n\nexport function unignoreIssue(issueId: string): void {\n const db = getIgnoredDb();\n db.prepare(\"DELETE FROM ignored_issues WHERE issue_id = ?\").run(issueId);\n}\n\n// Allowed SQL operations for the fix endpoint\n\nexport function isSafeFix(sql: string): boolean {\n const trimmed = sql.trim();\n if (!trimmed) return false;\n\n // Reject multi-statement SQL (split on semicolons, ignore trailing)\n const statements = trimmed.replace(/;\\s*$/, \"\").split(\";\").map(s => s.trim()).filter(Boolean);\n if (statements.length !== 1) return false;\n\n const upper = statements[0].toUpperCase();\n\n // EXPLAIN ANALYZE — only allow if followed by SELECT\n if (upper.startsWith(\"EXPLAIN ANALYZE\")) {\n const afterExplain = upper.replace(/^EXPLAIN\\s+ANALYZE\\s+/, \"\").trimStart();\n return afterExplain.startsWith(\"SELECT\");\n }\n\n // Simple prefix allowlist for single statements\n const ALLOWED_PREFIXES = [\n \"VACUUM\",\n \"ANALYZE\",\n \"REINDEX\",\n \"CREATE INDEX CONCURRENTLY\",\n \"DROP INDEX CONCURRENTLY\",\n \"SELECT PG_TERMINATE_BACKEND(\",\n \"SELECT PG_CANCEL_BACKEND(\",\n ];\n\n return ALLOWED_PREFIXES.some((p) => upper.startsWith(p));\n}\n","import type { Pool } from \"pg\";\n\nexport interface SlowQuery {\n queryid: string;\n query: string;\n calls: number;\n total_time: number;\n mean_time: number;\n rows: number;\n total_time_pretty: string;\n mean_time_pretty: string;\n}\n\nexport async function getSlowQueries(pool: Pool): Promise<SlowQuery[]> {\n const client = await pool.connect();\n try {\n // Check if pg_stat_statements is available\n const extCheck = await client.query(\n \"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\"\n );\n if (extCheck.rows.length === 0) {\n return [];\n }\n\n const r = await client.query(`\n SELECT\n queryid::text,\n query,\n calls::int,\n total_exec_time AS total_time,\n mean_exec_time AS mean_time,\n rows::int,\n round(total_exec_time::numeric / 1000, 2)::text || 's' AS total_time_pretty,\n round(mean_exec_time::numeric, 2)::text || 'ms' AS mean_time_pretty\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%'\n AND query NOT LIKE '%pg_catalog%'\n ORDER BY total_exec_time DESC\n LIMIT 50\n `);\n return r.rows;\n } catch {\n // pg_stat_statements might not be accessible\n return [];\n } finally {\n client.release();\n }\n}\n","import fs from \"node:fs\";\nimport path from \"node:path\";\nimport type { AdvisorResult, AdvisorIssue } from \"./advisor.js\";\n\nexport interface Snapshot {\n timestamp: string;\n result: AdvisorResult;\n}\n\nexport interface SnapshotDiff {\n scoreDelta: number;\n previousScore: number;\n currentScore: number;\n previousGrade: string;\n currentGrade: string;\n newIssues: AdvisorIssue[];\n resolvedIssues: AdvisorIssue[];\n unchanged: AdvisorIssue[];\n}\n\n/**\n * Normalize a dynamic issue ID for stable comparison.\n * Strips trailing -<number> suffixes so IDs like `maint-idle-12345`\n * (where 12345 is a PID that changes every run) don't produce false noise.\n */\nfunction normalizeIssueId(id: string): string {\n return id.replace(/-\\d+$/, \"\");\n}\n\n/**\n * Save a health-check snapshot to a specific file path.\n * The parent directory is created automatically.\n *\n * @param snapshotPath Full path to the JSON file (e.g. ~/.pg-dash/last-check.json)\n */\nexport function saveSnapshot(snapshotPath: string, result: AdvisorResult): void {\n fs.mkdirSync(path.dirname(snapshotPath), { recursive: true });\n const snapshot: Snapshot = { timestamp: new Date().toISOString(), result };\n fs.writeFileSync(snapshotPath, JSON.stringify(snapshot, null, 2));\n}\n\n/**\n * Load a previously saved snapshot from a specific file path.\n * Returns null if the file doesn't exist or cannot be parsed.\n *\n * @param snapshotPath Full path to the JSON file\n */\nexport function loadSnapshot(snapshotPath: string): Snapshot | null {\n if (!fs.existsSync(snapshotPath)) return null;\n try {\n return JSON.parse(fs.readFileSync(snapshotPath, \"utf-8\"));\n } catch {\n return null;\n }\n}\n\nexport function diffSnapshots(prev: AdvisorResult, current: AdvisorResult): SnapshotDiff {\n // Use normalized IDs for comparison to avoid noise from dynamic suffixes\n // (e.g. maint-idle-12345 where 12345 is a PID that changes every run).\n const prevNormIds = new Set(prev.issues.map((i) => normalizeIssueId(i.id)));\n const currNormIds = new Set(current.issues.map((i) => normalizeIssueId(i.id)));\n\n const newIssues = current.issues.filter((i) => !prevNormIds.has(normalizeIssueId(i.id)));\n const resolvedIssues = prev.issues.filter((i) => !currNormIds.has(normalizeIssueId(i.id)));\n const unchanged = current.issues.filter((i) => prevNormIds.has(normalizeIssueId(i.id)));\n\n return {\n scoreDelta: current.score - prev.score,\n previousScore: prev.score,\n currentScore: current.score,\n previousGrade: prev.grade,\n currentGrade: current.grade,\n newIssues,\n resolvedIssues,\n unchanged,\n };\n}\n","import { Pool } from \"pg\";\nimport { getAdvisorReport } from \"./advisor.js\";\nimport { buildLiveSnapshot } from \"./schema-tracker.js\";\nimport { diffSchemaSnapshots } from \"./schema-diff.js\";\n\nexport interface ColumnInfo {\n name: string;\n type: string;\n nullable: boolean;\n default?: string;\n}\n\nexport interface ColumnTypeDiff {\n column: string;\n sourceType: string;\n targetType: string;\n}\n\nexport interface ColumnNullableDiff {\n column: string;\n sourceNullable: boolean;\n targetNullable: boolean;\n}\n\nexport interface ColumnDefaultDiff {\n column: string;\n sourceDefault: string | null;\n targetDefault: string | null;\n}\n\nexport interface ColumnDiff {\n table: string;\n missingColumns: ColumnInfo[]; // source has, target doesn't\n extraColumns: ColumnInfo[]; // target has, source doesn't\n typeDiffs: ColumnTypeDiff[]; // same name, different type\n nullableDiffs: ColumnNullableDiff[]; // same name, different nullable\n defaultDiffs: ColumnDefaultDiff[]; // same name, different default\n}\n\nexport interface IndexDefDiff {\n name: string;\n sourceDef: string;\n targetDef: string;\n}\n\nexport interface IndexDiff {\n table: string;\n missingIndexes: string[]; // source has, target doesn't\n extraIndexes: string[]; // target has, source doesn't\n modifiedIndexes: IndexDefDiff[]; // same name, different definition\n}\n\nexport interface ConstraintDiff {\n table: string | null;\n type: \"missing\" | \"extra\" | \"modified\";\n name: string;\n detail: string;\n}\n\nexport interface EnumDiff {\n type: \"missing\" | \"extra\" | \"modified\";\n name: string;\n detail: string;\n}\n\nexport interface SchemaDiff {\n missingTables: string[];\n extraTables: string[];\n columnDiffs: ColumnDiff[];\n indexDiffs: IndexDiff[];\n constraintDiffs: ConstraintDiff[];\n enumDiffs: EnumDiff[];\n}\n\nexport interface HealthDiff {\n source: { score: number; grade: string; url: string };\n target: { score: number; grade: string; url: string };\n sourceOnlyIssues: string[];\n targetOnlyIssues: string[];\n}\n\nexport interface EnvDiffResult {\n schema: SchemaDiff;\n health?: HealthDiff;\n checkedAt: string;\n summary: {\n schemaDrifts: number;\n identical: boolean;\n };\n}\n\n// ----- internal types -----\n\ninterface RawColumn {\n table_name: string;\n column_name: string;\n data_type: string;\n is_nullable: string;\n column_default: string | null;\n}\n\ninterface RawIndex {\n tablename: string;\n indexname: string;\n indexdef: string;\n}\n\n// ----- query helpers -----\n\nasync function fetchTables(pool: Pool): Promise<string[]> {\n const res = await pool.query<{ table_name: string }>(`\n SELECT table_name\n FROM information_schema.tables\n WHERE table_schema = 'public' AND table_type = 'BASE TABLE'\n ORDER BY table_name\n `);\n return res.rows.map((r) => r.table_name);\n}\n\nasync function fetchColumns(pool: Pool): Promise<RawColumn[]> {\n const res = await pool.query<RawColumn>(`\n SELECT table_name, column_name, data_type, is_nullable, column_default\n FROM information_schema.columns\n WHERE table_schema = 'public'\n ORDER BY table_name, ordinal_position\n `);\n return res.rows;\n}\n\nasync function fetchIndexes(pool: Pool): Promise<RawIndex[]> {\n const res = await pool.query<RawIndex>(`\n SELECT tablename, indexname, indexdef\n FROM pg_indexes\n WHERE schemaname = 'public' AND indexname NOT LIKE '%_pkey'\n ORDER BY tablename, indexname\n `);\n return res.rows;\n}\n\n// ----- diff logic -----\n\nfunction diffTables(sourceTables: string[], targetTables: string[]): { missingTables: string[]; extraTables: string[] } {\n const sourceSet = new Set(sourceTables);\n const targetSet = new Set(targetTables);\n return {\n missingTables: sourceTables.filter((t) => !targetSet.has(t)),\n extraTables: targetTables.filter((t) => !sourceSet.has(t)),\n };\n}\n\nfunction groupColumnsByTable(columns: RawColumn[]): Map<string, Map<string, ColumnInfo>> {\n const map = new Map<string, Map<string, ColumnInfo>>();\n for (const col of columns) {\n if (!map.has(col.table_name)) map.set(col.table_name, new Map());\n const info: ColumnInfo = {\n name: col.column_name,\n type: col.data_type,\n nullable: col.is_nullable === \"YES\",\n };\n if (col.column_default !== null && col.column_default !== undefined) {\n info.default = col.column_default;\n }\n map.get(col.table_name)!.set(col.column_name, info);\n }\n return map;\n}\n\nfunction diffColumns(\n sourceCols: RawColumn[],\n targetCols: RawColumn[],\n commonTables: string[]\n): ColumnDiff[] {\n const sourceByTable = groupColumnsByTable(sourceCols);\n const targetByTable = groupColumnsByTable(targetCols);\n const diffs: ColumnDiff[] = [];\n\n for (const table of commonTables) {\n const srcMap = sourceByTable.get(table) ?? new Map<string, ColumnInfo>();\n const tgtMap = targetByTable.get(table) ?? new Map<string, ColumnInfo>();\n\n const missingColumns: ColumnInfo[] = [];\n const extraColumns: ColumnInfo[] = [];\n const typeDiffs: ColumnTypeDiff[] = [];\n const nullableDiffs: ColumnNullableDiff[] = [];\n const defaultDiffs: ColumnDefaultDiff[] = [];\n\n for (const [colName, srcInfo] of srcMap) {\n if (!tgtMap.has(colName)) {\n missingColumns.push(srcInfo);\n } else {\n const tgtInfo = tgtMap.get(colName)!;\n if (srcInfo.type !== tgtInfo.type) {\n typeDiffs.push({ column: colName, sourceType: srcInfo.type, targetType: tgtInfo.type });\n }\n if (srcInfo.nullable !== tgtInfo.nullable) {\n nullableDiffs.push({ column: colName, sourceNullable: srcInfo.nullable, targetNullable: tgtInfo.nullable });\n }\n if ((srcInfo.default ?? null) !== (tgtInfo.default ?? null)) {\n defaultDiffs.push({ column: colName, sourceDefault: srcInfo.default ?? null, targetDefault: tgtInfo.default ?? null });\n }\n }\n }\n\n for (const [colName, tgtInfo] of tgtMap) {\n if (!srcMap.has(colName)) {\n extraColumns.push(tgtInfo);\n }\n }\n\n if (missingColumns.length > 0 || extraColumns.length > 0 || typeDiffs.length > 0 ||\n nullableDiffs.length > 0 || defaultDiffs.length > 0) {\n diffs.push({ table, missingColumns, extraColumns, typeDiffs, nullableDiffs, defaultDiffs });\n }\n }\n\n return diffs;\n}\n\nfunction groupIndexesByTable(indexes: RawIndex[]): Map<string, Map<string, string>> {\n const map = new Map<string, Map<string, string>>();\n for (const idx of indexes) {\n if (!map.has(idx.tablename)) map.set(idx.tablename, new Map());\n map.get(idx.tablename)!.set(idx.indexname, idx.indexdef);\n }\n return map;\n}\n\nfunction diffIndexes(\n sourceIdxs: RawIndex[],\n targetIdxs: RawIndex[],\n commonTables: string[]\n): IndexDiff[] {\n const srcByTable = groupIndexesByTable(sourceIdxs);\n const tgtByTable = groupIndexesByTable(targetIdxs);\n const diffs: IndexDiff[] = [];\n\n // All tables that have any indexes in source or target\n const allTables = new Set([\n ...sourceIdxs.map((i) => i.tablename),\n ...targetIdxs.map((i) => i.tablename),\n ]);\n\n for (const table of allTables) {\n // Only diff tables that exist in both environments (common tables + tables not in either missingTables/extraTables)\n if (!commonTables.includes(table)) continue;\n\n const srcMap = srcByTable.get(table) ?? new Map<string, string>();\n const tgtMap = tgtByTable.get(table) ?? new Map<string, string>();\n\n const missingIndexes = [...srcMap.keys()].filter((i) => !tgtMap.has(i));\n const extraIndexes = [...tgtMap.keys()].filter((i) => !srcMap.has(i));\n const modifiedIndexes: IndexDefDiff[] = [];\n\n for (const [name, srcDef] of srcMap) {\n if (tgtMap.has(name)) {\n const tgtDef = tgtMap.get(name)!;\n if (srcDef !== tgtDef) {\n modifiedIndexes.push({ name, sourceDef: srcDef, targetDef: tgtDef });\n }\n }\n }\n\n if (missingIndexes.length > 0 || extraIndexes.length > 0 || modifiedIndexes.length > 0) {\n diffs.push({ table, missingIndexes, extraIndexes, modifiedIndexes });\n }\n }\n\n return diffs;\n}\n\nfunction countSchemaDrifts(schema: SchemaDiff): number {\n let n = schema.missingTables.length + schema.extraTables.length;\n for (const cd of schema.columnDiffs) {\n n += cd.missingColumns.length + cd.extraColumns.length + cd.typeDiffs.length +\n cd.nullableDiffs.length + cd.defaultDiffs.length;\n }\n for (const id of schema.indexDiffs) {\n n += id.missingIndexes.length + id.extraIndexes.length + id.modifiedIndexes.length;\n }\n n += (schema.constraintDiffs ?? []).length;\n n += (schema.enumDiffs ?? []).length;\n return n;\n}\n\n// ----- public API -----\n\nexport async function diffEnvironments(\n sourceConn: string,\n targetConn: string,\n options?: { includeHealth?: boolean }\n): Promise<EnvDiffResult> {\n const sourcePool = new Pool({ connectionString: sourceConn, connectionTimeoutMillis: 10000 });\n const targetPool = new Pool({ connectionString: targetConn, connectionTimeoutMillis: 10000 });\n\n try {\n // Run all schema queries in parallel (basic + deep snapshots for constraints/enums)\n const [\n sourceTables,\n targetTables,\n sourceCols,\n targetCols,\n sourceIdxs,\n targetIdxs,\n sourceSnap,\n targetSnap,\n ] = await Promise.all([\n fetchTables(sourcePool),\n fetchTables(targetPool),\n fetchColumns(sourcePool),\n fetchColumns(targetPool),\n fetchIndexes(sourcePool),\n fetchIndexes(targetPool),\n buildLiveSnapshot(sourcePool).catch(() => null),\n buildLiveSnapshot(targetPool).catch(() => null),\n ]);\n\n const { missingTables, extraTables } = diffTables(sourceTables, targetTables);\n const targetSet = new Set(targetTables);\n const commonTables = sourceTables.filter((t) => targetSet.has(t));\n\n const columnDiffs = diffColumns(sourceCols, targetCols, commonTables);\n const indexDiffs = diffIndexes(sourceIdxs, targetIdxs, commonTables);\n\n // Constraint + enum diffs via snapshot comparison\n const constraintDiffs: ConstraintDiff[] = [];\n const enumDiffs: EnumDiff[] = [];\n\n if (sourceSnap && targetSnap) {\n // diffSnapshots treats source as \"old\" and target as \"new\":\n // added = target has, source doesn't (extra in target)\n // removed = source has, target doesn't (missing in target)\n const snapChanges = diffSchemaSnapshots(sourceSnap, targetSnap);\n\n for (const c of snapChanges) {\n if (c.object_type === \"constraint\") {\n constraintDiffs.push({\n table: c.table_name ?? null,\n type: c.change_type === \"added\" ? \"extra\" : c.change_type === \"removed\" ? \"missing\" : \"modified\",\n name: c.detail.split(\" \")[1] ?? c.detail,\n detail: c.detail,\n });\n } else if (c.object_type === \"enum\") {\n enumDiffs.push({\n type: c.change_type === \"added\" ? \"extra\" : c.change_type === \"removed\" ? \"missing\" : \"modified\",\n name: c.detail.split(\" \")[1] ?? c.detail,\n detail: c.detail,\n });\n }\n }\n }\n\n const schema: SchemaDiff = { missingTables, extraTables, columnDiffs, indexDiffs, constraintDiffs, enumDiffs };\n const schemaDrifts = countSchemaDrifts(schema);\n\n let health: HealthDiff | undefined;\n\n if (options?.includeHealth) {\n const longQueryThreshold = 5;\n const [srcReport, tgtReport] = await Promise.all([\n getAdvisorReport(sourcePool, longQueryThreshold),\n getAdvisorReport(targetPool, longQueryThreshold),\n ]);\n\n const srcIssueKeys = new Set(srcReport.issues.map((i) => i.title));\n const tgtIssueKeys = new Set(tgtReport.issues.map((i) => i.title));\n\n const sourceOnlyIssues = srcReport.issues\n .filter((i) => !tgtIssueKeys.has(i.title))\n .map((i) => `${i.severity}: ${i.title}`);\n\n const targetOnlyIssues = tgtReport.issues\n .filter((i) => !srcIssueKeys.has(i.title))\n .map((i) => `${i.severity}: ${i.title}`);\n\n health = {\n source: { score: srcReport.score, grade: srcReport.grade, url: maskConnectionString(sourceConn) },\n target: { score: tgtReport.score, grade: tgtReport.grade, url: maskConnectionString(targetConn) },\n sourceOnlyIssues,\n targetOnlyIssues,\n };\n }\n\n return {\n schema,\n health,\n checkedAt: new Date().toISOString(),\n summary: {\n schemaDrifts,\n identical: schemaDrifts === 0,\n },\n };\n } finally {\n await Promise.allSettled([sourcePool.end(), targetPool.end()]);\n }\n}\n\n/** Mask password in a connection string to avoid leaking credentials */\nfunction maskConnectionString(connStr: string): string {\n try {\n const url = new URL(connStr);\n if (url.password) url.password = \"***\";\n return url.toString();\n } catch {\n return \"<redacted>\";\n }\n}\n\n// ----- formatters -----\n\nexport function formatTextDiff(result: EnvDiffResult): string {\n const lines: string[] = [];\n const sep = \"══════════════════════════════════════\";\n\n lines.push(`Environment Diff`);\n lines.push(sep);\n lines.push(``);\n lines.push(`Schema Drift:`);\n\n const { schema } = result;\n\n if (schema.missingTables.length > 0) {\n lines.push(` ✗ target missing tables: ${schema.missingTables.join(\", \")}`);\n }\n if (schema.extraTables.length > 0) {\n lines.push(` ⚠ target has extra tables: ${schema.extraTables.join(\", \")}`);\n }\n\n const missingCols: string[] = [];\n const extraCols: string[] = [];\n const typeChanges: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const col of cd.missingColumns) {\n missingCols.push(` ${cd.table}: ${col.name} (${col.type})`);\n }\n for (const col of cd.extraColumns) {\n extraCols.push(` ${cd.table}: ${col.name} (${col.type})`);\n }\n for (const td of cd.typeDiffs) {\n typeChanges.push(` ${cd.table}.${td.column}: ${td.sourceType} → ${td.targetType}`);\n }\n }\n\n if (missingCols.length > 0) {\n lines.push(` ✗ target missing columns:`);\n lines.push(...missingCols);\n }\n if (extraCols.length > 0) {\n lines.push(` ⚠ target has extra columns:`);\n lines.push(...extraCols);\n }\n if (typeChanges.length > 0) {\n lines.push(` ~ column type differences:`);\n lines.push(...typeChanges);\n }\n\n const nullableChanges: string[] = [];\n const defaultChanges: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const nd of cd.nullableDiffs) {\n const src = nd.sourceNullable ? \"nullable\" : \"NOT NULL\";\n const tgt = nd.targetNullable ? \"nullable\" : \"NOT NULL\";\n nullableChanges.push(` ${cd.table}.${nd.column}: source=${src} → target=${tgt}`);\n }\n for (const dd of cd.defaultDiffs) {\n const src = dd.sourceDefault ?? \"(none)\";\n const tgt = dd.targetDefault ?? \"(none)\";\n defaultChanges.push(` ${cd.table}.${dd.column}: source=${src} → target=${tgt}`);\n }\n }\n\n if (nullableChanges.length > 0) {\n lines.push(` ~ nullable differences:`);\n lines.push(...nullableChanges);\n }\n if (defaultChanges.length > 0) {\n lines.push(` ~ default differences:`);\n lines.push(...defaultChanges);\n }\n\n const missingIdxs: string[] = [];\n const extraIdxs: string[] = [];\n const modifiedIdxs: string[] = [];\n\n for (const id of schema.indexDiffs) {\n for (const idx of id.missingIndexes) {\n missingIdxs.push(` ${id.table}: ${idx}`);\n }\n for (const idx of id.extraIndexes) {\n extraIdxs.push(` ${id.table}: ${idx}`);\n }\n for (const mi of id.modifiedIndexes) {\n modifiedIdxs.push(` ${id.table}: ${mi.name} source=\"${mi.sourceDef}\" → target=\"${mi.targetDef}\"`);\n }\n }\n\n if (missingIdxs.length > 0) {\n lines.push(` ✗ target missing indexes:`);\n lines.push(...missingIdxs);\n }\n if (extraIdxs.length > 0) {\n lines.push(` ⚠ target has extra indexes:`);\n lines.push(...extraIdxs);\n }\n if (modifiedIdxs.length > 0) {\n lines.push(` ~ index definition differences:`);\n lines.push(...modifiedIdxs);\n }\n\n // Constraint diffs\n const missingConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"missing\");\n const extraConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"extra\");\n const modifiedConstraints = (schema.constraintDiffs ?? []).filter((c) => c.type === \"modified\");\n\n if (missingConstraints.length > 0) {\n lines.push(` ✗ target missing constraints:`);\n for (const c of missingConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n if (extraConstraints.length > 0) {\n lines.push(` ⚠ target has extra constraints:`);\n for (const c of extraConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n if (modifiedConstraints.length > 0) {\n lines.push(` ~ constraint differences:`);\n for (const c of modifiedConstraints) {\n lines.push(` ${c.table ? c.table + \": \" : \"\"}${c.detail}`);\n }\n }\n\n // Enum diffs\n const missingEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"missing\");\n const extraEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"extra\");\n const modifiedEnums = (schema.enumDiffs ?? []).filter((e) => e.type === \"modified\");\n\n if (missingEnums.length > 0) {\n lines.push(` ✗ target missing enums:`);\n for (const e of missingEnums) lines.push(` ${e.detail}`);\n }\n if (extraEnums.length > 0) {\n lines.push(` ⚠ target has extra enums:`);\n for (const e of extraEnums) lines.push(` ${e.detail}`);\n }\n if (modifiedEnums.length > 0) {\n lines.push(` ~ enum differences:`);\n for (const e of modifiedEnums) lines.push(` ${e.detail}`);\n }\n\n const noSchemaChanges = schema.missingTables.length === 0 && schema.extraTables.length === 0 &&\n schema.columnDiffs.length === 0 && schema.indexDiffs.length === 0 &&\n (schema.constraintDiffs ?? []).length === 0 && (schema.enumDiffs ?? []).length === 0 &&\n nullableChanges.length === 0 && defaultChanges.length === 0 && modifiedIdxs.length === 0;\n if (noSchemaChanges) {\n lines.push(` ✓ Schemas are identical`);\n }\n\n if (result.health) {\n const h = result.health;\n lines.push(``);\n lines.push(`Health Comparison:`);\n lines.push(` Source: ${h.source.score}/100 (${h.source.grade}) | Target: ${h.target.score}/100 (${h.target.grade})`);\n lines.push(` Source-only issues: ${h.sourceOnlyIssues.length === 0 ? \"(none)\" : \"\"}`);\n for (const iss of h.sourceOnlyIssues) lines.push(` - ${iss}`);\n lines.push(` Target-only issues: ${h.targetOnlyIssues.length === 0 ? \"(none)\" : \"\"}`);\n for (const iss of h.targetOnlyIssues) lines.push(` - ${iss}`);\n }\n\n lines.push(``);\n lines.push(sep);\n const { schemaDrifts, identical } = result.summary;\n lines.push(`Total: ${schemaDrifts} schema drift${schemaDrifts !== 1 ? \"s\" : \"\"} | Environments are ${identical ? \"in sync ✓\" : \"NOT in sync ✗\"}`);\n\n return lines.join(\"\\n\");\n}\n\nexport function formatMdDiff(result: EnvDiffResult): string {\n const lines: string[] = [];\n lines.push(`## 🔄 Environment Diff`);\n lines.push(``);\n lines.push(`### Schema Drift`);\n lines.push(``);\n\n const { schema } = result;\n const rows: Array<[string, string]> = [];\n\n if (schema.missingTables.length > 0) {\n rows.push([`❌ Missing tables`, schema.missingTables.map((t) => `\\`${t}\\``).join(\", \")]);\n }\n if (schema.extraTables.length > 0) {\n rows.push([`⚠️ Extra tables`, schema.extraTables.map((t) => `\\`${t}\\``).join(\", \")]);\n }\n\n const missingColItems: string[] = [];\n const extraColItems: string[] = [];\n const typeItems: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const col of cd.missingColumns) {\n missingColItems.push(`\\`${cd.table}.${col.name}\\``);\n }\n for (const col of cd.extraColumns) {\n extraColItems.push(`\\`${cd.table}.${col.name}\\``);\n }\n for (const td of cd.typeDiffs) {\n typeItems.push(`\\`${cd.table}.${td.column}\\` (${td.sourceType}→${td.targetType})`);\n }\n }\n\n if (missingColItems.length > 0) rows.push([`❌ Missing columns`, missingColItems.join(\", \")]);\n if (extraColItems.length > 0) rows.push([`⚠️ Extra columns`, extraColItems.join(\", \")]);\n if (typeItems.length > 0) rows.push([`~ Type differences`, typeItems.join(\", \")]);\n\n const nullableItems: string[] = [];\n const defaultItems: string[] = [];\n\n for (const cd of schema.columnDiffs) {\n for (const nd of cd.nullableDiffs) {\n const src = nd.sourceNullable ? \"nullable\" : \"NOT NULL\";\n const tgt = nd.targetNullable ? \"nullable\" : \"NOT NULL\";\n nullableItems.push(`\\`${cd.table}.${nd.column}\\` (${src}→${tgt})`);\n }\n for (const dd of cd.defaultDiffs) {\n const src = dd.sourceDefault ?? \"(none)\";\n const tgt = dd.targetDefault ?? \"(none)\";\n defaultItems.push(`\\`${cd.table}.${dd.column}\\` (${src}→${tgt})`);\n }\n }\n\n if (nullableItems.length > 0) rows.push([`~ Nullable differences`, nullableItems.join(\", \")]);\n if (defaultItems.length > 0) rows.push([`~ Default differences`, defaultItems.join(\", \")]);\n\n const missingIdxItems: string[] = [];\n const extraIdxItems: string[] = [];\n const modifiedIdxItems: string[] = [];\n\n for (const id of schema.indexDiffs) {\n for (const idx of id.missingIndexes) missingIdxItems.push(`\\`${id.table}.${idx}\\``);\n for (const idx of id.extraIndexes) extraIdxItems.push(`\\`${id.table}.${idx}\\``);\n for (const mi of id.modifiedIndexes) modifiedIdxItems.push(`\\`${id.table}.${mi.name}\\``);\n }\n\n if (missingIdxItems.length > 0) rows.push([`❌ Missing indexes`, missingIdxItems.join(\", \")]);\n if (extraIdxItems.length > 0) rows.push([`⚠️ Extra indexes`, extraIdxItems.join(\", \")]);\n if (modifiedIdxItems.length > 0) rows.push([`~ Modified indexes`, modifiedIdxItems.join(\", \")]);\n\n // Constraints\n const missingConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"missing\").map((c) => c.detail);\n const extraConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"extra\").map((c) => c.detail);\n const modConItems = (schema.constraintDiffs ?? []).filter((c) => c.type === \"modified\").map((c) => c.detail);\n if (missingConItems.length > 0) rows.push([`❌ Missing constraints`, missingConItems.join(\"; \")]);\n if (extraConItems.length > 0) rows.push([`⚠️ Extra constraints`, extraConItems.join(\"; \")]);\n if (modConItems.length > 0) rows.push([`~ Modified constraints`, modConItems.join(\"; \")]);\n\n // Enums\n const missingEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"missing\").map((e) => e.detail);\n const extraEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"extra\").map((e) => e.detail);\n const modEnumItems = (schema.enumDiffs ?? []).filter((e) => e.type === \"modified\").map((e) => e.detail);\n if (missingEnumItems.length > 0) rows.push([`❌ Missing enums`, missingEnumItems.join(\"; \")]);\n if (extraEnumItems.length > 0) rows.push([`⚠️ Extra enums`, extraEnumItems.join(\"; \")]);\n if (modEnumItems.length > 0) rows.push([`~ Modified enums`, modEnumItems.join(\"; \")]);\n\n if (rows.length > 0) {\n lines.push(`| Type | Details |`);\n lines.push(`|------|---------|`);\n for (const [type, details] of rows) {\n lines.push(`| ${type} | ${details} |`);\n }\n } else {\n lines.push(`✅ Schemas are identical`);\n }\n\n if (result.health) {\n const h = result.health;\n lines.push(``);\n lines.push(`### Health Comparison`);\n lines.push(``);\n lines.push(`| | Score | Grade |`);\n lines.push(`|--|-------|-------|`);\n lines.push(`| Source | ${h.source.score}/100 | ${h.source.grade} |`);\n lines.push(`| Target | ${h.target.score}/100 | ${h.target.grade} |`);\n\n if (h.targetOnlyIssues.length > 0) {\n lines.push(``);\n lines.push(`**Target-only issues:**`);\n for (const iss of h.targetOnlyIssues) lines.push(`- ${iss}`);\n }\n if (h.sourceOnlyIssues.length > 0) {\n lines.push(``);\n lines.push(`**Source-only issues:**`);\n for (const iss of h.sourceOnlyIssues) lines.push(`- ${iss}`);\n }\n }\n\n lines.push(``);\n const { schemaDrifts, identical } = result.summary;\n lines.push(`**Result: ${schemaDrifts} drift${schemaDrifts !== 1 ? \"s\" : \"\"} — environments are ${identical ? \"in sync ✓\" : \"NOT in sync\"}**`);\n\n return lines.join(\"\\n\");\n}\n","// Schema Diff — compares two schema snapshots and produces a change list\n\nexport interface SchemaSnapshot {\n tables: SnapshotTable[];\n enums: SnapshotEnum[];\n}\n\nexport interface SnapshotTable {\n name: string;\n schema: string;\n columns: SnapshotColumn[];\n indexes: SnapshotIndex[];\n constraints: SnapshotConstraint[];\n}\n\nexport interface SnapshotColumn {\n name: string;\n type: string;\n nullable: boolean;\n default_value: string | null;\n}\n\nexport interface SnapshotIndex {\n name: string;\n definition: string;\n is_unique: boolean;\n is_primary: boolean;\n}\n\nexport interface SnapshotConstraint {\n name: string;\n type: string;\n definition: string;\n}\n\nexport interface SnapshotEnum {\n name: string;\n schema: string;\n values: string[];\n}\n\nexport interface SchemaChange {\n change_type: \"added\" | \"removed\" | \"modified\";\n object_type: \"table\" | \"column\" | \"index\" | \"constraint\" | \"enum\";\n table_name: string | null;\n detail: string;\n}\n\nexport function diffSchemaSnapshots(oldSnap: SchemaSnapshot, newSnap: SchemaSnapshot): SchemaChange[] {\n const changes: SchemaChange[] = [];\n\n const oldTableMap = new Map(oldSnap.tables.map((t) => [`${t.schema}.${t.name}`, t]));\n const newTableMap = new Map(newSnap.tables.map((t) => [`${t.schema}.${t.name}`, t]));\n\n // Tables added/removed\n for (const [key, t] of newTableMap) {\n if (!oldTableMap.has(key)) {\n changes.push({ change_type: \"added\", object_type: \"table\", table_name: key, detail: `Table ${key} added` });\n }\n }\n for (const [key] of oldTableMap) {\n if (!newTableMap.has(key)) {\n changes.push({ change_type: \"removed\", object_type: \"table\", table_name: key, detail: `Table ${key} removed` });\n }\n }\n\n // Compare matching tables\n for (const [key, newTable] of newTableMap) {\n const oldTable = oldTableMap.get(key);\n if (!oldTable) continue;\n\n // Columns\n const oldCols = new Map(oldTable.columns.map((c) => [c.name, c]));\n const newCols = new Map(newTable.columns.map((c) => [c.name, c]));\n\n for (const [name, col] of newCols) {\n const oldCol = oldCols.get(name);\n if (!oldCol) {\n changes.push({ change_type: \"added\", object_type: \"column\", table_name: key, detail: `Column ${name} added (${col.type})` });\n } else {\n if (oldCol.type !== col.type) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} type changed: ${oldCol.type} → ${col.type}` });\n }\n if (oldCol.nullable !== col.nullable) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} nullable changed: ${oldCol.nullable} → ${col.nullable}` });\n }\n if (oldCol.default_value !== col.default_value) {\n changes.push({ change_type: \"modified\", object_type: \"column\", table_name: key, detail: `Column ${name} default changed: ${oldCol.default_value ?? \"NULL\"} → ${col.default_value ?? \"NULL\"}` });\n }\n }\n }\n for (const name of oldCols.keys()) {\n if (!newCols.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"column\", table_name: key, detail: `Column ${name} removed` });\n }\n }\n\n // Indexes\n const oldIdx = new Map(oldTable.indexes.map((i) => [i.name, i]));\n const newIdx = new Map(newTable.indexes.map((i) => [i.name, i]));\n for (const [name, idx] of newIdx) {\n if (!oldIdx.has(name)) {\n changes.push({ change_type: \"added\", object_type: \"index\", table_name: key, detail: `Index ${name} added` });\n } else if (oldIdx.get(name)!.definition !== idx.definition) {\n changes.push({ change_type: \"modified\", object_type: \"index\", table_name: key, detail: `Index ${name} definition changed` });\n }\n }\n for (const name of oldIdx.keys()) {\n if (!newIdx.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"index\", table_name: key, detail: `Index ${name} removed` });\n }\n }\n\n // Constraints\n const oldCon = new Map(oldTable.constraints.map((c) => [c.name, c]));\n const newCon = new Map(newTable.constraints.map((c) => [c.name, c]));\n for (const [name, con] of newCon) {\n if (!oldCon.has(name)) {\n changes.push({ change_type: \"added\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} added (${con.type})` });\n } else if (oldCon.get(name)!.definition !== con.definition) {\n changes.push({ change_type: \"modified\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} definition changed` });\n }\n }\n for (const name of oldCon.keys()) {\n if (!newCon.has(name)) {\n changes.push({ change_type: \"removed\", object_type: \"constraint\", table_name: key, detail: `Constraint ${name} removed` });\n }\n }\n }\n\n // Enums\n const oldEnums = new Map((oldSnap.enums || []).map((e) => [`${e.schema}.${e.name}`, e]));\n const newEnums = new Map((newSnap.enums || []).map((e) => [`${e.schema}.${e.name}`, e]));\n for (const [key, en] of newEnums) {\n const oldEn = oldEnums.get(key);\n if (!oldEn) {\n changes.push({ change_type: \"added\", object_type: \"enum\", table_name: null, detail: `Enum ${key} added (${en.values.join(\", \")})` });\n } else {\n const added = en.values.filter((v) => !oldEn.values.includes(v));\n const removed = oldEn.values.filter((v) => !en.values.includes(v));\n for (const v of added) {\n changes.push({ change_type: \"modified\", object_type: \"enum\", table_name: null, detail: `Enum ${key}: value '${v}' added` });\n }\n for (const v of removed) {\n changes.push({ change_type: \"modified\", object_type: \"enum\", table_name: null, detail: `Enum ${key}: value '${v}' removed` });\n }\n }\n }\n for (const key of oldEnums.keys()) {\n if (!newEnums.has(key)) {\n changes.push({ change_type: \"removed\", object_type: \"enum\", table_name: null, detail: `Enum ${key} removed` });\n }\n }\n\n return changes;\n}\n","// Schema Tracker — takes schema snapshots, stores in SQLite, detects changes\n\nconst SNAPSHOT_RETENTION = 50;\n\nimport type { Pool } from \"pg\";\nimport type Database from \"better-sqlite3\";\nimport { getSchemaTables, getSchemaTableDetail, getSchemaEnums } from \"./queries/schema.js\";\nimport { diffSchemaSnapshots, type SchemaSnapshot, type SchemaChange } from \"./schema-diff.js\";\n\n/** Build a full schema snapshot from a live pool — reusable for env comparison */\nexport async function buildLiveSnapshot(pool: Pool): Promise<SchemaSnapshot> {\n const tables = await getSchemaTables(pool);\n const enums = await getSchemaEnums(pool);\n\n const detailedTables = await Promise.all(\n tables.map(async (t: any) => {\n const detail = await getSchemaTableDetail(pool, `${t.schema}.${t.name}`);\n if (!detail) return null;\n return {\n name: detail.name,\n schema: detail.schema,\n columns: detail.columns.map((c: any) => ({\n name: c.name,\n type: c.type,\n nullable: c.nullable,\n default_value: c.default_value,\n })),\n indexes: detail.indexes.map((i: any) => ({\n name: i.name,\n definition: i.definition,\n is_unique: i.is_unique,\n is_primary: i.is_primary,\n })),\n constraints: detail.constraints.map((c: any) => ({\n name: c.name,\n type: c.type,\n definition: c.definition,\n })),\n };\n })\n );\n\n return {\n tables: detailedTables.filter(Boolean) as SchemaSnapshot[\"tables\"],\n enums: enums.map((e: any) => ({ name: e.name, schema: e.schema, values: e.values })),\n };\n}\n\nexport class SchemaTracker {\n private db: Database.Database;\n private pool: Pool;\n private intervalMs: number;\n private timer: ReturnType<typeof setInterval> | null = null;\n\n constructor(db: Database.Database, pool: Pool, intervalMs = 6 * 60 * 60 * 1000) {\n this.db = db;\n this.pool = pool;\n this.intervalMs = intervalMs;\n this.initTables();\n }\n\n private initTables() {\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS schema_snapshots (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp INTEGER NOT NULL,\n snapshot TEXT NOT NULL\n );\n CREATE TABLE IF NOT EXISTS schema_changes (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n snapshot_id INTEGER NOT NULL,\n timestamp INTEGER NOT NULL,\n change_type TEXT NOT NULL,\n object_type TEXT NOT NULL,\n table_name TEXT,\n detail TEXT NOT NULL,\n FOREIGN KEY (snapshot_id) REFERENCES schema_snapshots(id)\n );\n `);\n }\n\n async takeSnapshot(): Promise<{ snapshotId: number; changes: SchemaChange[] }> {\n const snapshot = await this.buildSnapshot();\n const now = Date.now();\n const json = JSON.stringify(snapshot);\n\n const info = this.db.prepare(\"INSERT INTO schema_snapshots (timestamp, snapshot) VALUES (?, ?)\").run(now, json);\n const snapshotId = Number(info.lastInsertRowid);\n\n // Prune old snapshots, keeping only the most recent SNAPSHOT_RETENTION\n this.db.prepare(`\n DELETE FROM schema_snapshots\n WHERE id NOT IN (\n SELECT id FROM schema_snapshots\n ORDER BY timestamp DESC\n LIMIT ?\n )\n `).run(SNAPSHOT_RETENTION);\n\n // Diff against previous\n const prev = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id < ? ORDER BY id DESC LIMIT 1\").get(snapshotId) as { snapshot: string } | undefined;\n let changes: SchemaChange[] = [];\n if (prev) {\n const oldSnap: SchemaSnapshot = JSON.parse(prev.snapshot);\n changes = diffSchemaSnapshots(oldSnap, snapshot);\n if (changes.length > 0) {\n const insert = this.db.prepare(\"INSERT INTO schema_changes (snapshot_id, timestamp, change_type, object_type, table_name, detail) VALUES (?, ?, ?, ?, ?, ?)\");\n const tx = this.db.transaction((chs: SchemaChange[]) => {\n for (const c of chs) {\n insert.run(snapshotId, now, c.change_type, c.object_type, c.table_name, c.detail);\n }\n });\n tx(changes);\n }\n }\n\n return { snapshotId, changes };\n }\n\n private async buildSnapshot(): Promise<SchemaSnapshot> {\n return buildLiveSnapshot(this.pool);\n }\n\n start() {\n // Take initial snapshot\n this.takeSnapshot().catch((err) => console.error(\"Schema snapshot error:\", err.message));\n this.timer = setInterval(() => {\n this.takeSnapshot().catch((err) => console.error(\"Schema snapshot error:\", err.message));\n }, this.intervalMs);\n }\n\n stop() {\n if (this.timer) {\n clearInterval(this.timer);\n this.timer = null;\n }\n }\n\n // API helpers\n getHistory(limit = 30) {\n return this.db.prepare(\"SELECT id, timestamp FROM schema_snapshots ORDER BY id DESC LIMIT ?\").all(limit);\n }\n\n getChanges(since?: number) {\n if (since) {\n return this.db.prepare(\"SELECT * FROM schema_changes WHERE timestamp >= ? ORDER BY timestamp DESC\").all(since);\n }\n return this.db.prepare(\"SELECT * FROM schema_changes ORDER BY timestamp DESC LIMIT 100\").all();\n }\n\n getLatestChanges() {\n const latest = this.db.prepare(\"SELECT id FROM schema_snapshots ORDER BY id DESC LIMIT 1\").get() as { id: number } | undefined;\n if (!latest) return [];\n return this.db.prepare(\"SELECT * FROM schema_changes WHERE snapshot_id = ? ORDER BY id\").all(latest.id);\n }\n\n getDiff(fromId: number, toId: number) {\n const from = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id = ?\").get(fromId) as { snapshot: string } | undefined;\n const to = this.db.prepare(\"SELECT snapshot FROM schema_snapshots WHERE id = ?\").get(toId) as { snapshot: string } | undefined;\n if (!from || !to) return null;\n return diffSchemaSnapshots(JSON.parse(from.snapshot), JSON.parse(to.snapshot));\n }\n}\n","// query-analyzer.ts — deep EXPLAIN plan analysis with auto index suggestions\n\nimport type { Pool } from \"pg\";\n\n// ─── Types ───────────────────────────────────────────────────────────────────\n\nexport interface PlanNodeSummary {\n nodeType: string;\n table?: string;\n totalCost: number;\n actualRows?: number;\n actualTime?: number; // ms\n filter?: string;\n}\n\nexport interface SeqScanInfo {\n table: string;\n rowCount: number;\n filter?: string; // filter condition from explain output\n suggestion?: string;\n}\n\nexport interface IndexSuggestion {\n table: string;\n columns: string[];\n reason: string;\n sql: string; // CREATE INDEX CONCURRENTLY …\n estimatedBenefit: \"high\" | \"medium\" | \"low\";\n}\n\nexport interface ExplainAnalysis {\n planNodes: PlanNodeSummary[];\n seqScans: SeqScanInfo[];\n missingIndexes: IndexSuggestion[];\n costEstimate: {\n totalCost: number;\n actualTime?: number;\n planningTime?: number;\n };\n recommendations: string[];\n}\n\nexport interface QueryRegressionInfo {\n queryId: string; // queryid from pg_stat_statements\n currentMeanMs: number;\n previousMeanMs: number;\n changePercent: number;\n degradedAt?: string; // approximate timestamp\n}\n\n// ─── Helpers ─────────────────────────────────────────────────────────────────\n\n/**\n * Recursively walk a plan tree (EXPLAIN FORMAT JSON) and collect every node.\n * Each node looks like { \"Node Type\": \"...\", \"Plans\": [...], ... }\n */\nfunction collectNodes(node: any, acc: any[] = []): any[] {\n if (!node || typeof node !== \"object\") return acc;\n acc.push(node);\n const plans = node[\"Plans\"] ?? node[\"plans\"];\n if (Array.isArray(plans)) {\n for (const child of plans) collectNodes(child, acc);\n }\n return acc;\n}\n\n/**\n * Extract simple column names from a Postgres filter expression.\n * Handles patterns like:\n * (col = $1)\n * (col > $1)\n * (col IS NULL)\n * (col IS NOT NULL)\n * (col ~~ '%foo%') -- LIKE\n */\nfunction extractColumnsFromFilter(filter: string): string[] {\n // Match identifiers that appear before comparison operators\n const colPattern = /\\(?\"?([a-z_][a-z0-9_]*)\"?\\s*(?:=|<|>|<=|>=|<>|!=|IS\\s+(?:NOT\\s+)?NULL|~~|!~~)/gi;\n const found = new Set<string>();\n let m: RegExpExecArray | null;\n while ((m = colPattern.exec(filter)) !== null) {\n const col = m[1].toLowerCase();\n // Skip Postgres internal names\n if (![\"and\", \"or\", \"not\", \"true\", \"false\", \"null\"].includes(col)) {\n found.add(col);\n }\n }\n return Array.from(found);\n}\n\n/**\n * Fetch the list of indexed column sets for a given table from pg_indexes.\n * Returns an array of column name arrays (one per index).\n */\nasync function getExistingIndexColumns(pool: Pool, tableName: string): Promise<string[][]> {\n try {\n // Query pg_indexes to get index definitions\n const r = await pool.query(\n `SELECT indexdef FROM pg_indexes WHERE tablename = $1`,\n [tableName]\n );\n return r.rows.map((row: any) => {\n // Parse column list from: ... ON table (col1, col2, ...)\n const m = /\\(([^)]+)\\)/.exec(row.indexdef);\n if (!m) return [] as string[];\n return m[1]\n .split(\",\")\n .map((c: string) => c.trim().replace(/^\"|\"$/g, \"\").toLowerCase());\n });\n } catch {\n return [];\n }\n}\n\n/**\n * Benefit rating based on estimated row count.\n */\nfunction rateBenefit(rowCount: number): \"high\" | \"medium\" | \"low\" {\n if (rowCount > 100_000) return \"high\";\n if (rowCount >= 10_000) return \"medium\";\n return \"low\";\n}\n\n/**\n * Format a large number as human-readable (1.2M, 50K, etc.)\n */\nfunction fmtRows(n: number): string {\n if (n >= 1_000_000) return `${(n / 1_000_000).toFixed(1)}M`;\n if (n >= 1_000) return `${(n / 1_000).toFixed(0)}K`;\n return String(n);\n}\n\n// ─── Core analysis ────────────────────────────────────────────────────────────\n\n/**\n * Analyse a EXPLAIN (FORMAT JSON) result and return rich diagnostics.\n *\n * @param explainJson - The value of `r.rows[0][\"QUERY PLAN\"]` (an array with one plan object)\n * @param pool - Optional PG pool; without it only static analysis is performed\n */\nexport async function analyzeExplainPlan(\n explainJson: any,\n pool?: Pool | null\n): Promise<ExplainAnalysis> {\n const result: ExplainAnalysis = {\n planNodes: [],\n seqScans: [],\n missingIndexes: [],\n costEstimate: { totalCost: 0 },\n recommendations: [],\n };\n\n if (!explainJson || !Array.isArray(explainJson) || explainJson.length === 0) {\n return result;\n }\n\n const topLevel = explainJson[0];\n const planRoot = topLevel?.[\"Plan\"] ?? topLevel?.[\"plan\"];\n\n // Planning / execution times from top-level\n const planningTime: number | undefined = topLevel?.[\"Planning Time\"] ?? undefined;\n const executionTime: number | undefined = topLevel?.[\"Execution Time\"] ?? undefined;\n\n if (!planRoot) return result;\n\n // Collect all nodes\n const allNodes = collectNodes(planRoot);\n\n // Build planNodes summary\n result.planNodes = allNodes.map((n: any) => {\n const s: PlanNodeSummary = {\n nodeType: n[\"Node Type\"] ?? \"Unknown\",\n totalCost: n[\"Total Cost\"] ?? 0,\n };\n if (n[\"Relation Name\"]) s.table = n[\"Relation Name\"];\n if (n[\"Actual Rows\"] !== undefined) s.actualRows = n[\"Actual Rows\"];\n if (n[\"Actual Total Time\"] !== undefined) s.actualTime = n[\"Actual Total Time\"];\n if (n[\"Filter\"]) s.filter = n[\"Filter\"];\n return s;\n });\n\n // Cost estimate from root node\n result.costEstimate = {\n totalCost: planRoot[\"Total Cost\"] ?? 0,\n actualTime: executionTime,\n planningTime,\n };\n\n // ── Seq Scan analysis ──────────────────────────────────────────────────────\n const seqScanNodes = allNodes.filter((n: any) => n[\"Node Type\"] === \"Seq Scan\");\n\n for (const node of seqScanNodes) {\n const table: string = node[\"Relation Name\"] ?? \"unknown\";\n const rowCount: number = node[\"Plan Rows\"] ?? node[\"Actual Rows\"] ?? 0;\n const filter: string | undefined = node[\"Filter\"];\n\n const info: SeqScanInfo = { table, rowCount, filter };\n\n if (rowCount > 10_000) {\n info.suggestion = filter\n ? `Consider adding an index to support the filter on ${table}`\n : `Full table scan on large table ${table} — review query`;\n }\n\n result.seqScans.push(info);\n }\n\n // ── Missing index inference ────────────────────────────────────────────────\n for (const scan of result.seqScans) {\n if (!scan.filter) continue;\n\n const cols = extractColumnsFromFilter(scan.filter);\n if (cols.length === 0) continue;\n\n // Check existing indexes (needs DB)\n let existingIndexCols: string[][] = [];\n if (pool) {\n existingIndexCols = await getExistingIndexColumns(pool, scan.table);\n }\n\n // Filter out columns already covered as the leading column of an existing index\n const uncoveredCols = cols.filter(\n (col) => !existingIndexCols.some((idxCols) => idxCols.length > 0 && idxCols[0] === col)\n );\n\n if (uncoveredCols.length === 0) continue;\n\n const benefit = rateBenefit(scan.rowCount);\n\n if (uncoveredCols.length >= 2) {\n // Suggest a composite index\n const idxName = `idx_${scan.table}_${uncoveredCols.join(\"_\")}`;\n const sql = `CREATE INDEX CONCURRENTLY ${idxName} ON ${scan.table} (${uncoveredCols.join(\", \")})`;\n result.missingIndexes.push({\n table: scan.table,\n columns: uncoveredCols,\n reason: `Seq Scan with multi-column filter (${uncoveredCols.join(\", \")}) on ${fmtRows(scan.rowCount)} rows — composite index preferred`,\n sql,\n estimatedBenefit: benefit,\n });\n } else {\n // Single column\n const col = uncoveredCols[0];\n const idxName = `idx_${scan.table}_${col}`;\n const sql = `CREATE INDEX CONCURRENTLY ${idxName} ON ${scan.table} (${col})`;\n result.missingIndexes.push({\n table: scan.table,\n columns: [col],\n reason: `Seq Scan with Filter on ${col} (${fmtRows(scan.rowCount)} rows)`,\n sql,\n estimatedBenefit: benefit,\n });\n }\n }\n\n // ── Recommendations ────────────────────────────────────────────────────────\n for (const scan of result.seqScans) {\n if (scan.rowCount > 10_000) {\n const filterPart = scan.filter\n ? ` — consider adding index on ${extractColumnsFromFilter(scan.filter).join(\", \") || \"filter columns\"}`\n : \" — no filter; full scan may be intentional\";\n result.recommendations.push(\n `Seq Scan on ${scan.table} (${fmtRows(scan.rowCount)} rows)${filterPart}`\n );\n }\n }\n\n if (planningTime !== undefined) {\n const label = planningTime > 10 ? \"high — check statistics\" : \"normal\";\n result.recommendations.push(`Planning time ${planningTime.toFixed(1)}ms — ${label}`);\n }\n\n if (result.missingIndexes.length === 0 && result.seqScans.length === 0) {\n result.recommendations.push(\"No obvious sequential scans detected — query looks efficient\");\n }\n\n return result;\n}\n\n// ─── Regression detection ─────────────────────────────────────────────────────\n\n/**\n * Detect queries whose mean execution time has increased by more than 50%\n * compared to the earliest snapshot in the query_stats store for the given window.\n *\n * This is a best-effort function; it silently returns [] if pg_stat_statements\n * is unavailable or the query_stats store doesn't have enough history.\n *\n * @param pool - PG pool (used to read pg_stat_statements)\n * @param statsDb - Optional better-sqlite3 Database with query_stats table\n * @param windowHours - How far back to compare (default 24 h)\n */\nexport async function detectQueryRegressions(\n pool: Pool,\n statsDb?: any | null,\n windowHours = 24\n): Promise<QueryRegressionInfo[]> {\n try {\n // ── 1. Check pg_stat_statements is available ───────────────────────────\n const extCheck = await pool.query(\n \"SELECT 1 FROM pg_extension WHERE extname = 'pg_stat_statements'\"\n );\n if (extCheck.rows.length === 0) return [];\n\n // ── 2. Get current snapshot from pg_stat_statements ───────────────────\n const current = await pool.query(`\n SELECT queryid::text AS queryid, mean_exec_time\n FROM pg_stat_statements\n WHERE query NOT LIKE '%pg_stat%'\n AND queryid IS NOT NULL\n `);\n\n const currentMap = new Map<string, number>();\n for (const row of current.rows) {\n currentMap.set(row.queryid, parseFloat(row.mean_exec_time));\n }\n\n if (!statsDb) return [];\n\n // ── 3. Fetch historical baselines from SQLite query_stats ─────────────\n const windowMs = windowHours * 60 * 60 * 1000;\n const since = Date.now() - windowMs;\n\n let historical: { queryid: string; mean_exec_time: number; timestamp: number }[];\n try {\n historical = statsDb\n .prepare(\n `SELECT queryid, mean_exec_time, timestamp\n FROM query_stats\n WHERE timestamp >= ?\n ORDER BY queryid, timestamp ASC`\n )\n .all(since) as any[];\n } catch {\n return [];\n }\n\n // Keep only the *earliest* record per queryid in the window\n const baselineMap = new Map<string, { meanMs: number; timestamp: number }>();\n for (const row of historical) {\n if (!baselineMap.has(row.queryid)) {\n baselineMap.set(row.queryid, {\n meanMs: row.mean_exec_time,\n timestamp: row.timestamp,\n });\n }\n }\n\n // ── 4. Detect regressions > 50% ────────────────────────────────────────\n const regressions: QueryRegressionInfo[] = [];\n\n for (const [queryId, baseline] of baselineMap) {\n const currentMean = currentMap.get(queryId);\n if (currentMean === undefined || baseline.meanMs === 0) continue;\n\n const changePercent =\n ((currentMean - baseline.meanMs) / baseline.meanMs) * 100;\n\n if (changePercent > 50) {\n regressions.push({\n queryId,\n currentMeanMs: currentMean,\n previousMeanMs: baseline.meanMs,\n changePercent: Math.round(changePercent),\n degradedAt: new Date(baseline.timestamp).toISOString(),\n });\n }\n }\n\n return regressions.sort((a, b) => b.changePercent - a.changePercent);\n } catch {\n return [];\n }\n}\n","// Migration safety checker — static + dynamic analysis of SQL migration files\n\nimport type { Pool } from \"pg\";\n\nexport interface MigrationIssue {\n severity: \"error\" | \"warning\" | \"info\";\n code: string;\n message: string;\n suggestion?: string;\n lineNumber?: number;\n tableName?: string;\n estimatedRows?: number;\n estimatedLockSeconds?: number;\n}\n\nexport interface MigrationCheckResult {\n safe: boolean;\n issues: MigrationIssue[];\n summary: {\n errors: number;\n warnings: number;\n infos: number;\n };\n checkedAt: string;\n}\n\n// Strip SQL comments while preserving line numbers (replace with spaces)\nfunction stripComments(sql: string): string {\n // Replace /* ... */ block comments (preserve newlines for line number tracking)\n let stripped = sql.replace(/\\/\\*[\\s\\S]*?\\*\\//g, (match) =>\n match.replace(/[^\\n]/g, \" \")\n );\n // Replace -- single-line comments (preserve the newline)\n stripped = stripped.replace(/--[^\\n]*/g, (match) => \" \".repeat(match.length));\n return stripped;\n}\n\n// Helper: find line number of a match in the original SQL\nfunction findLineNumber(sql: string, matchIndex: number): number {\n const before = sql.slice(0, matchIndex);\n return before.split(\"\\n\").length;\n}\n\n// Extract bare table name from possibly-quoted or schema-qualified identifier\nfunction bareTable(name: string): string {\n return name\n .replace(/^public\\./i, \"\")\n .replace(/\"/g, \"\")\n .toLowerCase()\n .trim();\n}\n\n// Parse all table names operated on by this migration\nfunction extractOperatedTables(sql: string): {\n indexTables: string[]; // CREATE INDEX ON <table>\n alterTables: string[]; // ALTER TABLE <table>\n dropTables: string[]; // DROP TABLE <table>\n refTables: string[]; // REFERENCES <table>\n} {\n sql = stripComments(sql);\n const indexTables: string[] = [];\n const alterTables: string[] = [];\n const dropTables: string[] = [];\n const refTables: string[] = [];\n\n // CREATE INDEX ... ON table\n const idxRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+(?:CONCURRENTLY\\s+)?(?:IF\\s+NOT\\s+EXISTS\\s+)?(?:\\w+\\s+)?ON\\s+([\\w.\"]+)/gi;\n let m: RegExpExecArray | null;\n while ((m = idxRe.exec(sql)) !== null) indexTables.push(bareTable(m[1]));\n\n // ALTER TABLE table\n const altRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n while ((m = altRe.exec(sql)) !== null) alterTables.push(bareTable(m[1]));\n\n // DROP TABLE\n const dropRe = /\\bDROP\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n while ((m = dropRe.exec(sql)) !== null) dropTables.push(bareTable(m[1]));\n\n // REFERENCES table\n const refRe = /\\bREFERENCES\\s+([\\w.\"]+)/gi;\n while ((m = refRe.exec(sql)) !== null) refTables.push(bareTable(m[1]));\n\n return { indexTables, alterTables, dropTables, refTables };\n}\n\n// Static analysis — no DB needed\nfunction staticCheck(sql: string): MigrationIssue[] {\n const issues: MigrationIssue[] = [];\n // Strip comments before analysis to avoid false positives from commented-out SQL\n sql = stripComments(sql);\n\n // Determine tables created IN THIS MIGRATION (so we know they're brand-new)\n const createdTablesRe = /\\bCREATE\\s+TABLE\\s+(?:IF\\s+NOT\\s+EXISTS\\s+)?([\\w.\"]+)/gi;\n const createdTables = new Set<string>();\n let m: RegExpExecArray | null;\n while ((m = createdTablesRe.exec(sql)) !== null) createdTables.add(bareTable(m[1]));\n\n // 1. CREATE INDEX without CONCURRENTLY (on tables NOT created in this migration)\n const idxRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+(?!CONCURRENTLY)((?:IF\\s+NOT\\s+EXISTS\\s+)?(?:\\w+\\s+)?ON\\s+([\\w.\"]+))/gi;\n while ((m = idxRe.exec(sql)) !== null) {\n const table = bareTable(m[2]);\n const lineNumber = findLineNumber(sql, m.index);\n if (!createdTables.has(table)) {\n issues.push({\n severity: \"warning\",\n code: \"INDEX_WITHOUT_CONCURRENTLY\",\n message: `CREATE INDEX on existing table will lock writes. Use CREATE INDEX CONCURRENTLY to avoid downtime.`,\n suggestion: \"Replace CREATE INDEX with CREATE INDEX CONCURRENTLY\",\n lineNumber,\n tableName: table,\n });\n }\n }\n\n // 2. CREATE INDEX CONCURRENTLY → info\n const idxConcRe = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+CONCURRENTLY\\b/gi;\n while ((m = idxConcRe.exec(sql)) !== null) {\n issues.push({\n severity: \"info\",\n code: \"INDEX_CONCURRENTLY_OK\",\n message: \"CREATE INDEX CONCURRENTLY — safe, no write lock\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 3 & 4. ALTER TABLE ... ADD COLUMN ... NOT NULL (with/without DEFAULT)\n // Match: ALTER TABLE <t> ADD COLUMN <col> <type> [DEFAULT <val>] [NOT NULL | NULL]\n const addColRe =\n /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ADD\\s+(?:COLUMN\\s+)?(?:IF\\s+NOT\\s+EXISTS\\s+)?[\\w\"]+\\s+[\\w\\s()\"',.[\\]]+?(?=;|$)/gi;\n while ((m = addColRe.exec(sql)) !== null) {\n const fragment = m[0];\n const table = bareTable(m[1]);\n const lineNumber = findLineNumber(sql, m.index);\n const fragUpper = fragment.toUpperCase();\n\n const hasNotNull = /\\bNOT\\s+NULL\\b/.test(fragUpper);\n const hasDefault = /\\bDEFAULT\\b/.test(fragUpper);\n\n if (hasNotNull && !hasDefault) {\n issues.push({\n severity: \"error\",\n code: \"ADD_COLUMN_NOT_NULL_NO_DEFAULT\",\n message: \"ADD COLUMN NOT NULL without DEFAULT will fail if table has existing rows\",\n suggestion: \"Add a DEFAULT value, then remove it after migration\",\n lineNumber,\n tableName: table,\n });\n } else if (hasNotNull && hasDefault) {\n issues.push({\n severity: \"warning\",\n code: \"ADD_COLUMN_REWRITES_TABLE\",\n message: \"ADD COLUMN with NOT NULL DEFAULT may rewrite table on PostgreSQL < 11\",\n suggestion: \"On PostgreSQL 11+ with a constant default this is safe. For older versions, add column nullable first.\",\n lineNumber,\n tableName: table,\n });\n }\n }\n\n // 5. DROP TABLE\n const dropRe = /\\bDROP\\s+TABLE\\b/gi;\n while ((m = dropRe.exec(sql)) !== null) {\n issues.push({\n severity: \"warning\",\n code: \"DROP_TABLE\",\n message: \"DROP TABLE is destructive. Ensure this is intentional and data is backed up.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 5b. ALTER COLUMN TYPE — rewrites the entire table and locks it\n const alterTypeRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ALTER\\s+(?:COLUMN\\s+)?[\\w\"]+\\s+TYPE\\b/gi;\n while ((m = alterTypeRe.exec(sql)) !== null) {\n const table = bareTable(m[1]);\n issues.push({\n severity: \"warning\",\n code: \"ALTER_COLUMN_TYPE\",\n message: \"ALTER COLUMN TYPE rewrites the entire table and acquires an exclusive lock.\",\n suggestion: \"Consider using a new column + backfill + rename strategy to avoid downtime.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5c. DROP COLUMN — safe in PostgreSQL 9.0+ (marks invisible, no rewrite), but breaks app code\n const dropColRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+DROP\\s+(?:COLUMN\\s+)(?:IF\\s+EXISTS\\s+)?[\\w\"]+\\b/gi;\n while ((m = dropColRe.exec(sql)) !== null) {\n const table = bareTable(m[1]);\n issues.push({\n severity: \"info\",\n code: \"DROP_COLUMN\",\n message: \"DROP COLUMN is safe in PostgreSQL (no table rewrite), but may break application code referencing that column.\",\n suggestion: \"Ensure no application code references this column before dropping it.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5d-i. RENAME TABLE\n const renameTableRe = /ALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?(\\w+)\\s+RENAME\\s+TO\\s+(\\w+)/gi;\n while ((m = renameTableRe.exec(sql)) !== null) {\n const oldName = m[1];\n const newName = m[2];\n issues.push({\n severity: \"warning\",\n code: \"RENAME_TABLE\",\n message: `Renaming table \"${oldName}\" to \"${newName}\" breaks application code referencing the old name`,\n suggestion: \"Deploy application code that handles both names before renaming, or use a view with the old name after renaming.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: oldName,\n });\n }\n\n // 5d-ii. RENAME COLUMN\n const renameColumnRe = /ALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?(\\w+)\\s+RENAME\\s+COLUMN\\s+(\\w+)\\s+TO\\s+(\\w+)/gi;\n while ((m = renameColumnRe.exec(sql)) !== null) {\n const table = m[1];\n const oldCol = m[2];\n const newCol = m[3];\n issues.push({\n severity: \"warning\",\n code: \"RENAME_COLUMN\",\n message: `Renaming column \"${oldCol}\" to \"${newCol}\" on table \"${table}\" breaks application code referencing the old column name`,\n suggestion: \"Add new column, backfill data, update application to use new column, then drop old column (expand/contract pattern).\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n\n // 5e. ADD CONSTRAINT without NOT VALID — performs a full table scan to validate\n const addConRe = /\\bALTER\\s+TABLE\\s+(?:IF\\s+EXISTS\\s+)?([\\w.\"]+)\\s+ADD\\s+CONSTRAINT\\b[^;]*(;|$)/gi;\n while ((m = addConRe.exec(sql)) !== null) {\n const fragment = m[0];\n const table = bareTable(m[1]);\n const fragUpper = fragment.toUpperCase();\n // Skip if NOT VALID is already present\n if (!/\\bNOT\\s+VALID\\b/.test(fragUpper)) {\n issues.push({\n severity: \"warning\",\n code: \"ADD_CONSTRAINT_SCANS_TABLE\",\n message: \"ADD CONSTRAINT validates all existing rows and holds an exclusive lock during the scan.\",\n suggestion: \"Use ADD CONSTRAINT ... NOT VALID to skip validation, then VALIDATE CONSTRAINT in a separate transaction.\",\n lineNumber: findLineNumber(sql, m.index),\n tableName: table,\n });\n }\n }\n\n // 5e. CREATE INDEX CONCURRENTLY inside transaction (BEGIN/COMMIT)\n const hasTransaction = /\\bBEGIN\\b/i.test(sql) || /\\bSTART\\s+TRANSACTION\\b/i.test(sql);\n const hasConcurrently = /\\bCREATE\\s+(?:UNIQUE\\s+)?INDEX\\s+CONCURRENTLY\\b/i.test(sql);\n if (hasTransaction && hasConcurrently) {\n issues.push({\n severity: \"error\",\n code: \"CONCURRENTLY_IN_TRANSACTION\",\n message: \"CREATE INDEX CONCURRENTLY cannot run inside a transaction block. It will fail at runtime.\",\n suggestion: \"Remove the BEGIN/COMMIT wrapper, or use a migration tool that runs CONCURRENTLY outside transactions.\",\n });\n }\n\n // 6. TRUNCATE\n const truncRe = /\\bTRUNCATE\\b/gi;\n while ((m = truncRe.exec(sql)) !== null) {\n issues.push({\n severity: \"warning\",\n code: \"TRUNCATE_TABLE\",\n message: \"TRUNCATE will delete all rows. Ensure this is intentional.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n\n // 7. DELETE FROM without WHERE\n const delRe = /\\bDELETE\\s+FROM\\s+[\\w.\"]+\\s*(?:;|$)/gi;\n while ((m = delRe.exec(sql)) !== null) {\n // If there's no WHERE clause in this statement\n const stmt = m[0];\n if (!/\\bWHERE\\b/i.test(stmt)) {\n issues.push({\n severity: \"warning\",\n code: \"DELETE_WITHOUT_WHERE\",\n message: \"DELETE without WHERE clause will remove all rows.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n }\n\n // 8. UPDATE ... SET without WHERE\n const updRe = /\\bUPDATE\\s+[\\w.\"]+\\s+SET\\b[^;]*(;|$)/gi;\n while ((m = updRe.exec(sql)) !== null) {\n const stmt = m[0];\n if (!/\\bWHERE\\b/i.test(stmt)) {\n issues.push({\n severity: \"warning\",\n code: \"UPDATE_WITHOUT_WHERE\",\n message: \"UPDATE without WHERE clause will modify all rows.\",\n lineNumber: findLineNumber(sql, m.index),\n });\n }\n }\n\n return issues;\n}\n\n// Dynamic analysis — requires a running PG pool\nasync function dynamicCheck(sql: string, pool: Pool, staticIssues: MigrationIssue[]): Promise<MigrationIssue[]> {\n const issues: MigrationIssue[] = [];\n const { indexTables, alterTables, dropTables, refTables } = extractOperatedTables(sql);\n\n // All tables we need to look up\n const allTables = [...new Set([...indexTables, ...alterTables, ...dropTables])];\n\n // Query row counts for all tables at once\n const tableStats = new Map<string, { rowCount: number; totalSize: number }>();\n if (allTables.length > 0) {\n try {\n const res = await pool.query<{ tablename: string; n_live_tup: string; total_size: string }>(\n `SELECT tablename,\n n_live_tup,\n pg_total_relation_size(schemaname||'.'||tablename) AS total_size\n FROM pg_stat_user_tables\n WHERE tablename = ANY($1)`,\n [allTables]\n );\n for (const row of res.rows) {\n tableStats.set(row.tablename, {\n rowCount: parseInt(row.n_live_tup ?? \"0\", 10),\n totalSize: parseInt(row.total_size ?? \"0\", 10),\n });\n }\n } catch (_) {\n // Ignore DB errors in dynamic check\n }\n }\n\n // Upgrade CREATE INDEX (non-CONCURRENTLY) issues based on actual row counts\n for (const issue of staticIssues) {\n if (issue.code === \"INDEX_WITHOUT_CONCURRENTLY\" && issue.tableName) {\n const stats = tableStats.get(issue.tableName);\n if (stats) {\n const { rowCount } = stats;\n const lockSecs = Math.round(rowCount / 50000);\n issue.estimatedRows = rowCount;\n issue.estimatedLockSeconds = lockSecs;\n\n if (rowCount > 1_000_000) {\n issue.severity = \"error\";\n issue.message = `CREATE INDEX on '${issue.tableName}' will lock writes for ~${lockSecs}s (${(rowCount / 1e6).toFixed(1)}M rows). CRITICAL — use CREATE INDEX CONCURRENTLY.`;\n } else if (rowCount > 100_000) {\n issue.message = `CREATE INDEX on '${issue.tableName}' will lock writes for ~${lockSecs}s (${(rowCount / 1000).toFixed(0)}k rows).`;\n }\n }\n }\n }\n\n // Validate REFERENCES tables exist\n const uniqueRefTables = [...new Set(refTables)];\n for (const table of uniqueRefTables) {\n try {\n const res = await pool.query<{ tablename: string }>(\n `SELECT tablename FROM pg_tables WHERE schemaname = 'public' AND tablename = $1`,\n [table]\n );\n if (res.rows.length === 0) {\n issues.push({\n severity: \"error\",\n code: \"MISSING_TABLE\",\n message: `Table '${table}' referenced in migration does not exist`,\n tableName: table,\n });\n }\n } catch (_) {\n // Ignore\n }\n }\n\n return issues;\n}\n\nexport async function analyzeMigration(sql: string, pool?: Pool): Promise<MigrationCheckResult> {\n const trimmed = sql.trim();\n\n if (!trimmed) {\n return {\n safe: true,\n issues: [],\n summary: { errors: 0, warnings: 0, infos: 0 },\n checkedAt: new Date().toISOString(),\n };\n }\n\n // Static checks first (mutates issue severity if dynamic info is available)\n const issues = staticCheck(trimmed);\n\n // Dynamic checks (augments existing issues + adds new ones like MISSING_TABLE)\n if (pool) {\n const dynamicIssues = await dynamicCheck(trimmed, pool, issues);\n issues.push(...dynamicIssues);\n }\n\n const errors = issues.filter((i) => i.severity === \"error\").length;\n const warnings = issues.filter((i) => i.severity === \"warning\").length;\n const infos = issues.filter((i) => i.severity === \"info\").length;\n\n return {\n safe: errors === 0,\n issues,\n summary: { errors, warnings, infos },\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface UnusedIndex {\n schema: string;\n table: string;\n index: string;\n indexSize: string; // human-readable e.g. \"2.4 MB\"\n indexSizeBytes: number;\n scans: number; // idx_scan from pg_stat_user_indexes\n lastUsed: string | null; // timestamp of last stats reset — index has had 0 scans since this date\n suggestion: string;\n}\n\nexport interface UnusedIndexReport {\n indexes: UnusedIndex[];\n totalWastedBytes: number;\n totalWasted: string; // human-readable\n checkedAt: string;\n}\n\nexport function formatBytes(bytes: number): string {\n if (bytes < 1024) return \"< 1 KB\";\n if (bytes < 1024 * 1024) return `${Math.round(bytes / 1024)} KB`;\n if (bytes < 1024 * 1024 * 1024) return `${(bytes / (1024 * 1024)).toFixed(1)} MB`;\n if (bytes < 1024 ** 4) return `${(bytes / 1024 ** 3).toFixed(1)} GB`;\n return `${(bytes / 1024 ** 4).toFixed(1)} TB`;\n}\n\nexport async function getUnusedIndexes(pool: Pool): Promise<UnusedIndexReport> {\n const [indexResult, bgwriterResult] = await Promise.all([\n pool.query(`\n SELECT\n s.schemaname,\n s.relname AS table_name,\n s.indexrelname AS index_name,\n pg_relation_size(s.indexrelid) AS index_size_bytes,\n s.idx_scan,\n i.indexdef\n FROM pg_stat_user_indexes s\n JOIN pg_indexes i ON s.schemaname = i.schemaname\n AND s.relname = i.tablename\n AND s.indexrelname = i.indexname\n WHERE s.schemaname = 'public'\n AND s.idx_scan = 0\n AND i.indexdef NOT LIKE '%UNIQUE%'\n AND s.indexrelname NOT LIKE '%_pkey'\n ORDER BY pg_relation_size(s.indexrelid) DESC\n `),\n pool.query(`SELECT stats_reset FROM pg_stat_bgwriter`),\n ]);\n\n const statsReset = bgwriterResult.rows[0]?.stats_reset\n ? new Date(bgwriterResult.rows[0].stats_reset).toISOString()\n : null;\n\n const filteredRows = indexResult.rows.filter((row: any) => {\n const def: string = row.indexdef ?? \"\";\n // Exclude partial indexes (have WHERE clause)\n if (def.includes(\" WHERE \")) return false;\n // Exclude expression indexes: column list has nested parens e.g. lower(email)\n // Extract the column portion between the first ( and last )\n const colStart = def.indexOf(\"(\");\n const colEnd = def.lastIndexOf(\")\");\n if (colStart !== -1 && colEnd !== -1) {\n const cols = def.slice(colStart + 1, colEnd);\n if (cols.includes(\"(\")) return false; // expression index\n }\n return true;\n });\n\n const indexes: UnusedIndex[] = filteredRows.map((row: any) => {\n const sizeBytes = parseInt(row.index_size_bytes, 10) || 0;\n const index = row.index_name as string;\n const table = row.table_name as string;\n return {\n schema: row.schemaname as string,\n table,\n index,\n indexSize: formatBytes(sizeBytes),\n indexSizeBytes: sizeBytes,\n scans: parseInt(row.idx_scan, 10) || 0,\n lastUsed: statsReset,\n suggestion: `Index ${index} on ${table} has never been used (0 scans). Consider dropping it: DROP INDEX CONCURRENTLY \"${index.replace(/\"/g, '\"\"')}\"`,\n };\n });\n\n const totalWastedBytes = indexes.reduce((sum, idx) => sum + idx.indexSizeBytes, 0);\n\n return {\n indexes,\n totalWastedBytes,\n totalWasted: formatBytes(totalWastedBytes),\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface TableBloat {\n schema: string;\n table: string;\n liveRows: number;\n deadRows: number;\n bloatPercent: number; // dead / (live + dead) * 100, rounded 1dp\n lastAutoVacuum: string | null;\n lastVacuum: string | null;\n suggestion: string;\n}\n\nexport interface BloatReport {\n tables: TableBloat[]; // sorted by bloatPercent DESC, only tables with bloatPercent >= 10\n checkedAt: string;\n}\n\nfunction getSuggestion(table: string, bloatPercent: number): string {\n if (bloatPercent >= 50) {\n return `HIGH bloat on ${table} (${bloatPercent}% dead rows). Run: VACUUM ANALYZE ${table}`;\n } else if (bloatPercent >= 20) {\n return `Moderate bloat on ${table} (${bloatPercent}% dead rows). Consider VACUUM ANALYZE ${table}`;\n } else {\n return `Minor bloat on ${table} (${bloatPercent}% dead rows). Autovacuum should handle this.`;\n }\n}\n\nexport async function getBloatReport(pool: Pool): Promise<BloatReport> {\n const result = await pool.query(`\n SELECT\n schemaname,\n relname AS table_name,\n n_live_tup,\n n_dead_tup,\n last_autovacuum,\n last_vacuum\n FROM pg_stat_user_tables\n WHERE schemaname = 'public'\n AND (n_live_tup + n_dead_tup) > 0\n ORDER BY (n_dead_tup::float / (n_live_tup + n_dead_tup)) DESC\n `);\n\n const tables: TableBloat[] = [];\n\n for (const row of result.rows) {\n const live = parseInt(row.n_live_tup, 10) || 0;\n const dead = parseInt(row.n_dead_tup, 10) || 0;\n const total = live + dead;\n if (total === 0) continue;\n\n const bloatPercent = Math.round((dead / total) * 1000) / 10; // 1dp\n if (bloatPercent < 10) continue;\n\n const table = row.table_name as string;\n tables.push({\n schema: row.schemaname as string,\n table,\n liveRows: live,\n deadRows: dead,\n bloatPercent,\n lastAutoVacuum: row.last_autovacuum ? new Date(row.last_autovacuum).toISOString() : null,\n lastVacuum: row.last_vacuum ? new Date(row.last_vacuum).toISOString() : null,\n suggestion: getSuggestion(table, bloatPercent),\n });\n }\n\n // Sort by bloatPercent DESC (DB query orders by dead ratio, but re-sort after filtering)\n tables.sort((a, b) => b.bloatPercent - a.bloatPercent);\n\n return {\n tables,\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface AutovacuumTableStatus {\n schema: string;\n table: string;\n lastAutoVacuum: string | null;\n lastAutoAnalyze: string | null;\n deadTuples: number;\n liveTuples: number;\n vacuumCount: number;\n analyzeCount: number;\n status: \"ok\" | \"stale\" | \"never\" | \"overdue\";\n suggestion: string | null;\n}\n\nexport interface AutovacuumReport {\n tables: AutovacuumTableStatus[];\n settings: {\n autovacuumEnabled: boolean;\n vacuumCostDelay: string;\n autovacuumMaxWorkers: number;\n autovacuumNaptime: string;\n };\n checkedAt: string;\n}\n\nfunction classifyStatus(\n lastAutoVacuum: Date | null,\n deadTuples: number,\n vacuumCount: number\n): AutovacuumTableStatus[\"status\"] {\n if (lastAutoVacuum === null) return \"never\"; // covers all null cases\n\n const daysSince = (Date.now() - lastAutoVacuum.getTime()) / (1000 * 60 * 60 * 24);\n\n if (daysSince > 7 && deadTuples > 10_000) return \"overdue\";\n if (daysSince > 3) return \"stale\";\n return \"ok\";\n}\n\nfunction getSuggestion(status: \"ok\" | \"stale\" | \"never\" | \"overdue\", table: string): string | null {\n switch (status) {\n case \"never\":\n return `Table ${table} has never been autovacuumed. Check if autovacuum is enabled and the table has enough churn.`;\n case \"overdue\":\n return `Table ${table} is overdue for vacuum and has many dead tuples. Run: VACUUM ANALYZE ${table}`;\n case \"stale\":\n return `Table ${table} hasn't been vacuumed in over 3 days. Monitor for bloat.`;\n case \"ok\":\n return null;\n }\n}\n\nexport async function getAutovacuumReport(pool: Pool): Promise<AutovacuumReport> {\n const [tableResult, settingsResult] = await Promise.all([\n pool.query(`\n SELECT\n schemaname, relname,\n last_autovacuum, last_autoanalyze,\n n_dead_tup, n_live_tup,\n autovacuum_count, autoanalyze_count\n FROM pg_stat_user_tables\n WHERE schemaname = 'public'\n ORDER BY n_dead_tup DESC\n `),\n pool.query(`\n SELECT name, setting\n FROM pg_settings\n WHERE name IN ('autovacuum', 'autovacuum_vacuum_cost_delay', 'autovacuum_max_workers', 'autovacuum_naptime')\n `),\n ]);\n\n const tables: AutovacuumTableStatus[] = tableResult.rows.map((row: any) => {\n const lastAutoVacuumDate = row.last_autovacuum ? new Date(row.last_autovacuum) : null;\n const deadTuples = parseInt(row.n_dead_tup, 10) || 0;\n const liveTuples = parseInt(row.n_live_tup, 10) || 0;\n const vacuumCount = parseInt(row.autovacuum_count, 10) || 0;\n const analyzeCount = parseInt(row.autoanalyze_count, 10) || 0;\n const status = classifyStatus(lastAutoVacuumDate, deadTuples, vacuumCount);\n const table = row.relname as string;\n\n return {\n schema: row.schemaname as string,\n table,\n lastAutoVacuum: lastAutoVacuumDate ? lastAutoVacuumDate.toISOString() : null,\n lastAutoAnalyze: row.last_autoanalyze ? new Date(row.last_autoanalyze).toISOString() : null,\n deadTuples,\n liveTuples,\n vacuumCount,\n analyzeCount,\n status,\n suggestion: getSuggestion(status, table),\n };\n });\n\n const settingsMap = new Map<string, string>();\n for (const row of settingsResult.rows) {\n settingsMap.set(row.name, row.setting);\n }\n\n return {\n tables,\n settings: {\n autovacuumEnabled: settingsMap.get(\"autovacuum\") !== \"off\",\n vacuumCostDelay: `${settingsMap.get(\"autovacuum_vacuum_cost_delay\") ?? \"2\"}ms`,\n autovacuumMaxWorkers: parseInt(settingsMap.get(\"autovacuum_max_workers\") ?? \"3\", 10),\n autovacuumNaptime: `${settingsMap.get(\"autovacuum_naptime\") ?? \"60\"}s`,\n },\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface LockWait {\n blockedPid: number;\n blockedQuery: string;\n blockedDuration: string; // e.g. \"00:00:45\"\n blockingPid: number;\n blockingQuery: string;\n blockingDuration: string;\n table: string | null;\n lockType: string;\n}\n\nexport interface LockReport {\n waitingLocks: LockWait[];\n longRunningQueries: Array<{\n pid: number;\n duration: string;\n query: string;\n state: string;\n waitEventType: string | null;\n }>;\n checkedAt: string;\n}\n\nexport function formatDurationSecs(secs: number): string {\n const h = Math.floor(secs / 3600);\n const m = Math.floor((secs % 3600) / 60);\n const s = secs % 60;\n return [\n String(h).padStart(2, \"0\"),\n String(m).padStart(2, \"0\"),\n String(s).padStart(2, \"0\"),\n ].join(\":\");\n}\n\nexport async function getLockReport(pool: Pool): Promise<LockReport> {\n const [locksResult, longResult] = await Promise.all([\n pool.query(`\n SELECT\n blocked.pid AS blocked_pid,\n blocked.query AS blocked_query,\n EXTRACT(EPOCH FROM (NOW() - blocked.query_start))::int AS blocked_secs,\n blocking.pid AS blocking_pid,\n blocking.query AS blocking_query,\n EXTRACT(EPOCH FROM (NOW() - blocking.query_start))::int AS blocking_secs,\n blocked_locks.relation::regclass::text AS table_name,\n blocked_locks.locktype\n FROM pg_catalog.pg_locks blocked_locks\n JOIN pg_catalog.pg_stat_activity blocked ON blocked.pid = blocked_locks.pid\n JOIN pg_catalog.pg_locks blocking_locks\n ON blocking_locks.locktype = blocked_locks.locktype\n AND blocking_locks.relation IS NOT DISTINCT FROM blocked_locks.relation\n AND blocking_locks.pid != blocked_locks.pid\n AND blocking_locks.granted = true\n JOIN pg_catalog.pg_stat_activity blocking ON blocking.pid = blocking_locks.pid\n WHERE NOT blocked_locks.granted\n `),\n pool.query(`\n SELECT\n pid,\n EXTRACT(EPOCH FROM (NOW() - query_start))::int AS duration_secs,\n query,\n state,\n wait_event_type\n FROM pg_stat_activity\n WHERE state != 'idle'\n AND query_start IS NOT NULL\n AND EXTRACT(EPOCH FROM (NOW() - query_start)) > 5\n AND query NOT LIKE '%pg_stat_activity%'\n ORDER BY duration_secs DESC\n LIMIT 20\n `),\n ]);\n\n // Deduplicate by (blockedPid, blockingPid) — same pair may appear multiple times\n // for different lock types; keep only the first occurrence.\n const seen = new Set<string>();\n const waitingLocks: LockWait[] = [];\n for (const row of locksResult.rows) {\n const key = `${row.blocked_pid}:${row.blocking_pid}`;\n if (!seen.has(key)) {\n seen.add(key);\n waitingLocks.push({\n blockedPid: parseInt(row.blocked_pid, 10),\n blockedQuery: row.blocked_query as string,\n blockedDuration: formatDurationSecs(parseInt(row.blocked_secs, 10) || 0),\n blockingPid: parseInt(row.blocking_pid, 10),\n blockingQuery: row.blocking_query as string,\n blockingDuration: formatDurationSecs(parseInt(row.blocking_secs, 10) || 0),\n table: row.table_name ?? null,\n lockType: row.locktype as string,\n });\n }\n }\n\n const longRunningQueries = longResult.rows.map((row: any) => ({\n pid: parseInt(row.pid, 10),\n duration: formatDurationSecs(parseInt(row.duration_secs, 10) || 0),\n query: row.query as string,\n state: row.state as string,\n waitEventType: row.wait_event_type ?? null,\n }));\n\n return {\n waitingLocks,\n longRunningQueries,\n checkedAt: new Date().toISOString(),\n };\n}\n","import type { Pool } from \"pg\";\n\nexport interface ConfigRecommendation {\n setting: string;\n currentValue: string;\n recommendedValue: string;\n reason: string;\n severity: \"error\" | \"warning\" | \"info\";\n docs?: string;\n}\n\nexport interface ConfigReport {\n recommendations: ConfigRecommendation[];\n serverInfo: {\n maxConnections: number;\n sharedBuffers: string;\n workMem: string;\n effectiveCacheSize: string;\n maintenanceWorkMem: string;\n walBuffers: string;\n checkpointCompletionTarget: string;\n randomPageCost: string;\n autovacuumVacuumScaleFactor: string;\n };\n checkedAt: string;\n}\n\n// Convert a setting value + unit to bytes for comparison\n// In pg_settings, 'setting' is in the displayed 'unit'\nfunction settingToBytes(value: string, unit: string | undefined): number {\n const v = parseFloat(value);\n if (!unit) return v;\n switch (unit.toLowerCase()) {\n case \"b\": return v;\n case \"kb\": return v * 1024;\n case \"8kb\": return v * 8 * 1024; // shared_buffers, effective_cache_size\n case \"mb\": return v * 1024 * 1024;\n case \"gb\": return v * 1024 * 1024 * 1024;\n default: return v;\n }\n}\n\nfunction settingToMb(value: string, unit: string | undefined): number {\n return settingToBytes(value, unit) / (1024 * 1024);\n}\n\n// Format a memory setting to a human-readable string with units\nfunction formatMemSetting(rawValue: string | null | undefined, unit?: string): string {\n if (!rawValue) return \"unknown\";\n const bytes = settingToBytes(rawValue, unit ?? \"\");\n if (bytes <= 0 || isNaN(bytes)) return rawValue; // fallback for special values like -1 (auto)\n if (bytes >= 1024 ** 3) return `${(bytes / 1024 ** 3).toFixed(1)}GB`;\n if (bytes >= 1024 ** 2) return `${Math.round(bytes / 1024 ** 2)}MB`;\n if (bytes >= 1024) return `${Math.round(bytes / 1024)}KB`;\n return `${bytes}B`;\n}\n\nexport async function getConfigReport(pool: Pool): Promise<ConfigReport> {\n const result = await pool.query(`\n SELECT name, setting, unit\n FROM pg_settings\n WHERE name IN (\n 'max_connections', 'shared_buffers', 'work_mem',\n 'effective_cache_size', 'maintenance_work_mem', 'wal_buffers',\n 'checkpoint_completion_target', 'random_page_cost',\n 'autovacuum_vacuum_scale_factor', 'autovacuum_analyze_scale_factor',\n 'log_min_duration_statement', 'idle_in_transaction_session_timeout',\n 'effective_io_concurrency'\n )\n `);\n\n const settings: Record<string, { setting: string; unit: string | undefined }> = {};\n for (const row of result.rows) {\n settings[row.name] = { setting: row.setting, unit: row.unit ?? undefined };\n }\n\n const recommendations: ConfigRecommendation[] = [];\n\n const get = (name: string) => settings[name]?.setting ?? null;\n const getUnit = (name: string) => settings[name]?.unit;\n\n // 1. shared_buffers: < 128MB → warning\n const sharedBuffersSetting = get(\"shared_buffers\");\n if (sharedBuffersSetting !== null) {\n const mb = settingToMb(sharedBuffersSetting, getUnit(\"shared_buffers\"));\n if (mb < 128) {\n recommendations.push({\n setting: \"shared_buffers\",\n currentValue: `${Math.round(mb)}MB`,\n recommendedValue: \"256MB\",\n reason: \"shared_buffers should be at least 25% of RAM; typical starting point is 256MB–1GB\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-SHARED-BUFFERS\",\n });\n }\n }\n\n // 2. work_mem: <= 4MB → info\n const workMemSetting = get(\"work_mem\");\n if (workMemSetting !== null) {\n const mb = settingToMb(workMemSetting, getUnit(\"work_mem\"));\n if (mb <= 4) {\n recommendations.push({\n setting: \"work_mem\",\n currentValue: `${mb % 1 === 0 ? mb : mb.toFixed(1)}MB`,\n recommendedValue: \"16MB\",\n reason: \"work_mem of 4MB is conservative; consider 16MB–64MB for analytical queries (but multiply by max_connections for total)\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-WORK-MEM\",\n });\n }\n }\n\n // 3. checkpoint_completion_target: < 0.9 → warning\n const cctSetting = get(\"checkpoint_completion_target\");\n if (cctSetting !== null) {\n const v = parseFloat(cctSetting);\n if (v < 0.9) {\n recommendations.push({\n setting: \"checkpoint_completion_target\",\n currentValue: cctSetting,\n recommendedValue: \"0.9\",\n reason: \"Set to 0.9 to spread checkpoint I/O over 90% of checkpoint interval\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-wal.html#GUC-CHECKPOINT-COMPLETION-TARGET\",\n });\n }\n }\n\n // 4. random_page_cost: > 2.0 → info\n const rpcSetting = get(\"random_page_cost\");\n if (rpcSetting !== null) {\n const v = parseFloat(rpcSetting);\n if (v > 2.0) {\n recommendations.push({\n setting: \"random_page_cost\",\n currentValue: rpcSetting,\n recommendedValue: \"1.1\",\n reason: \"If using SSDs, set random_page_cost=1.1 (default 4.0 is tuned for spinning disks)\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-query.html#GUC-RANDOM-PAGE-COST\",\n });\n }\n }\n\n // 5. autovacuum_vacuum_scale_factor: >= 0.2 → info\n const avsfSetting = get(\"autovacuum_vacuum_scale_factor\");\n if (avsfSetting !== null) {\n const v = parseFloat(avsfSetting);\n if (v >= 0.2) {\n recommendations.push({\n setting: \"autovacuum_vacuum_scale_factor\",\n currentValue: avsfSetting,\n recommendedValue: \"0.05\",\n reason: \"Consider lowering to 0.05–0.1 for large tables to vacuum more frequently\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-autovacuum.html#GUC-AUTOVACUUM-VACUUM-SCALE-FACTOR\",\n });\n }\n }\n\n // 6. log_min_duration_statement: = -1 → info\n const lmdsSetting = get(\"log_min_duration_statement\");\n if (lmdsSetting !== null && parseInt(lmdsSetting, 10) === -1) {\n recommendations.push({\n setting: \"log_min_duration_statement\",\n currentValue: \"-1\",\n recommendedValue: \"1000\",\n reason: \"Consider setting to 1000 (log queries > 1s) for performance monitoring\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-logging.html#GUC-LOG-MIN-DURATION-STATEMENT\",\n });\n }\n\n // 7. idle_in_transaction_session_timeout: = 0 → warning\n const iitsSetting = get(\"idle_in_transaction_session_timeout\");\n if (iitsSetting !== null && parseInt(iitsSetting, 10) === 0) {\n recommendations.push({\n setting: \"idle_in_transaction_session_timeout\",\n currentValue: \"0\",\n recommendedValue: \"60000\",\n reason: \"Set idle_in_transaction_session_timeout=60000 (60s) to prevent stuck transactions from holding locks\",\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-client.html#GUC-IDLE-IN-TRANSACTION-SESSION-TIMEOUT\",\n });\n }\n\n // 8. effective_io_concurrency: = 1 → info\n const eicSetting = get(\"effective_io_concurrency\");\n if (eicSetting !== null && parseInt(eicSetting, 10) === 1) {\n recommendations.push({\n setting: \"effective_io_concurrency\",\n currentValue: \"1\",\n recommendedValue: \"200\",\n reason: \"If using SSDs, set effective_io_concurrency=200 for better parallel I/O\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-EFFECTIVE-IO-CONCURRENCY\",\n });\n }\n\n // 9. wal_buffers: skip if -1 (auto)\n\n // 10. maintenance_work_mem: <= 64MB → info\n const mwmSetting = get(\"maintenance_work_mem\");\n if (mwmSetting !== null) {\n const mb = settingToMb(mwmSetting, getUnit(\"maintenance_work_mem\"));\n if (mb <= 64) {\n recommendations.push({\n setting: \"maintenance_work_mem\",\n currentValue: `${mb % 1 === 0 ? mb : mb.toFixed(1)}MB`,\n recommendedValue: \"256MB\",\n reason: \"Consider 256MB for faster VACUUM and index builds\",\n severity: \"info\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-resource.html#GUC-MAINTENANCE-WORK-MEM\",\n });\n }\n }\n\n const maxConnSetting = get(\"max_connections\");\n\n // max_connections > 200 without connection pooler is a common perf trap\n if (maxConnSetting !== null) {\n const maxConn = parseInt(maxConnSetting, 10);\n if (maxConn > 200) {\n recommendations.push({\n setting: \"max_connections\",\n currentValue: String(maxConn),\n recommendedValue: \"100\",\n reason: `max_connections=${maxConn} is high. Each connection uses ~5–10MB RAM. Without a connection pooler (PgBouncer), this leads to memory pressure and context-switch overhead. Consider lowering to 100 and using a pooler.`,\n severity: \"warning\",\n docs: \"https://www.postgresql.org/docs/current/runtime-config-connection.html#GUC-MAX-CONNECTIONS\",\n });\n }\n }\n\n const serverInfo = {\n maxConnections: maxConnSetting !== null ? parseInt(maxConnSetting, 10) : 0,\n sharedBuffers: formatMemSetting(sharedBuffersSetting, getUnit(\"shared_buffers\")),\n workMem: formatMemSetting(workMemSetting, getUnit(\"work_mem\")),\n effectiveCacheSize: formatMemSetting(get(\"effective_cache_size\"), getUnit(\"effective_cache_size\")),\n maintenanceWorkMem: formatMemSetting(mwmSetting, getUnit(\"maintenance_work_mem\")),\n walBuffers: get(\"wal_buffers\") ?? \"\",\n checkpointCompletionTarget: cctSetting ?? \"\",\n randomPageCost: rpcSetting ?? \"\",\n autovacuumVacuumScaleFactor: avsfSetting ?? \"\",\n };\n\n return {\n recommendations,\n serverInfo,\n checkedAt: new Date().toISOString(),\n };\n}\n"],"mappings":";;;AAEA,SAAS,iBAAiB;AAC1B,SAAS,4BAA4B;AACrC,SAAS,QAAAA,aAAY;AACrB,SAAS,SAAS;;;ACHlB,eAAsB,YAAYC,OAAY;AAC5C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,UAAU,MAAM,OAAO,MAAM,qBAAqB;AACxD,UAAM,SAAS,MAAM,OAAO;AAAA,MAC1B;AAAA,IACF;AACA,UAAM,SAAS,MAAM,OAAO;AAAA,MAC1B;AAAA,IACF;AACA,UAAM,UAAU,MAAM,OAAO;AAAA,MAC3B;AAAA,IACF;AACA,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,KAKtC;AAED,WAAO;AAAA,MACL,SAAS,QAAQ,KAAK,CAAC,EAAE;AAAA,MACzB,QAAQ,OAAO,KAAK,CAAC,EAAE;AAAA,MACvB,QAAQ,OAAO,KAAK,CAAC,EAAE;AAAA,MACvB,eAAe,QAAQ,KAAK,CAAC,EAAE;AAAA,MAC/B,aAAa,YAAY,KAAK,CAAC;AAAA,IACjC;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;AC9BA,eAAsB,UAAUC,OAAY;AAC1C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAa5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACrBA,eAAsB,gBAAgBC,OAAY;AAChD,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAe5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AAEA,eAAsB,qBAAqBA,OAAY,WAAmB;AACxE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AAEF,UAAM,QAAQ,UAAU,MAAM,GAAG;AACjC,UAAM,SAAS,MAAM,SAAS,IAAI,MAAM,CAAC,IAAI;AAC7C,UAAM,OAAO,MAAM,SAAS,IAAI,MAAM,CAAC,IAAI,MAAM,CAAC;AAGlD,UAAM,YAAY,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAclC,CAAC,MAAM,MAAM,CAAC;AAEjB,QAAI,UAAU,KAAK,WAAW,EAAG,QAAO;AAGxC,UAAM,UAAU,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAYhC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,UAAU,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAiBhC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAUpC,CAAC,MAAM,MAAM,CAAC;AAGjB,UAAM,cAAc,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAWpC,CAAC,MAAM,MAAM,CAAC;AAGjB,QAAI,aAAoB,CAAC;AACzB,QAAI;AACF,YAAM,SAAS,MAAM,OAAO;AAAA,QAC1B,iBAAiB,OAAO,iBAAiB,MAAM,CAAC,IAAI,OAAO,iBAAiB,IAAI,CAAC;AAAA,MACnF;AACA,mBAAa,OAAO;AAAA,IACtB,SAAS,KAAK;AAAE,cAAQ,MAAM,mBAAoB,IAAc,OAAO;AAAA,IAAG;AAE1E,WAAO;AAAA,MACL,GAAG,UAAU,KAAK,CAAC;AAAA,MACnB,SAAS,QAAQ;AAAA,MACjB,SAAS,QAAQ;AAAA,MACjB,aAAa,YAAY;AAAA,MACzB,aAAa,YAAY;AAAA,MACzB;AAAA,IACF;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AAwEA,eAAsB,eAAeC,OAAY;AAC/C,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAW5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACrNA,eAAsB,YAAYC,OAAiC;AACjE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AACF,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAuB5B;AACD,WAAO,EAAE;AAAA,EACX,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;ACqsBA,OAAO,cAAc;AACrB,OAAO,UAAU;AACjB,OAAO,QAAQ;AACf,OAAO,QAAQ;AAvtBf,IAAM,kBAAkB,EAAE,UAAU,IAAI,SAAS,GAAG,MAAM,EAAE;AAC5D,IAAM,gBAAgB,EAAE,UAAU,IAAI,SAAS,IAAI,MAAM,GAAG;AAErD,SAAS,oBAAoB,QAAgC;AAClE,MAAI,QAAQ;AACZ,QAAM,aAAa,EAAE,UAAU,GAAG,SAAS,GAAG,MAAM,EAAE;AACtD,QAAM,SAAS,EAAE,UAAU,GAAG,SAAS,GAAG,MAAM,EAAE;AAClD,aAAW,SAAS,QAAQ;AAC1B,WAAO,MAAM,QAAQ;AACrB,UAAM,IAAI,OAAO,MAAM,QAAQ;AAC/B,UAAM,SAAS,gBAAgB,MAAM,QAAQ;AAE7C,QAAI;AACJ,QAAI,KAAK,EAAG,WAAU;AAAA,aACb,KAAK,GAAI,WAAU,SAAS;AAAA,QAChC,WAAU,SAAS;AACxB,eAAW,MAAM,QAAQ,KAAK;AAAA,EAChC;AAEA,aAAW,OAAO,CAAC,YAAY,WAAW,MAAM,GAAY;AAC1D,aAAS,KAAK,IAAI,WAAW,GAAG,GAAG,cAAc,GAAG,CAAC;AAAA,EACvD;AACA,SAAO,KAAK,IAAI,GAAG,KAAK,IAAI,KAAK,KAAK,MAAM,KAAK,CAAC,CAAC;AACrD;AAEO,SAAS,eAAe,OAAuB;AACpD,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,MAAI,SAAS,GAAI,QAAO;AACxB,SAAO;AACT;AAEA,SAAS,iBAAiB,QAAyF;AACjH,QAAM,aAAa,CAAC,eAAe,eAAe,UAAU,UAAU;AACtE,QAAM,SAA0E,CAAC;AACjF,aAAW,OAAO,YAAY;AAC5B,UAAM,YAAY,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,GAAG;AACzD,UAAM,QAAQ,oBAAoB,SAAS;AAC3C,WAAO,GAAG,IAAI,EAAE,OAAO,OAAO,eAAe,KAAK,GAAG,OAAO,UAAU,OAAO;AAAA,EAC/E;AACA,SAAO;AACT;AAEA,eAAsB,iBAAiBC,OAAYC,sBAAqB,GAA2B;AACjG,QAAM,SAAS,MAAMD,MAAK,QAAQ;AAClC,QAAM,SAAyB,CAAC;AAChC,QAAM,UAAoB,CAAC;AAE3B,MAAI;AAEF,UAAM,gBAAgB,MAAM,OAAO,MAAM,yBAAyB;AAClE,UAAM,YAAY,SAAS,cAAc,KAAK,CAAC,EAAE,kBAAkB;AAKnE,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAM5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,iBAAiB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAClD,UAAU,IAAI,WAAW,MAAO,YAAY;AAAA,UAC5C,UAAU;AAAA,UACV,OAAO,4BAA4B,IAAI,OAAO;AAAA,UAC9C,aAAa,SAAS,IAAI,UAAU,IAAI,IAAI,OAAO,KAAK,IAAI,UAAU,UAAU,IAAI,IAAI,SAAS,IAAI,QAAQ,6BAA6B,OAAO,IAAI,YAAY,EAAE,eAAe,CAAC;AAAA,UACnL,KAAK;AAAA,mCAA4F,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,gCAA2D,IAAI,OAAO,gBAAgB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjP,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,uCAAwC,IAAc,OAAO;AAAG,cAAQ,KAAK,gBAAiB,IAAc,OAAO;AAAA,IACnI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAW5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,oBAAoB,IAAI,YAAY;AAAA,UACxC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,iBAAiB,IAAI,YAAY;AAAA,UACxC,aAAa,SAAS,IAAI,YAAY,OAAO,IAAI,OAAO,OAAO,IAAI,eAAe,0BAA0B,IAAI,eAAe;AAAA,UAC/H,KAAK,8BAA8B,IAAI,UAAU,IAAI,IAAI,YAAY;AAAA,UACrE,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,MAAM,WAAW,IAAI,QAAQ;AACnC,eAAO,KAAK;AAAA,UACV,IAAI,cAAc,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC/C,UAAU,MAAM,KAAK,aAAa;AAAA,UAClC,UAAU;AAAA,UACV,OAAO,kBAAkB,IAAI,OAAO,KAAK,IAAI,QAAQ;AAAA,UACrD,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO,QAAQ,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC,iBAAiB,IAAI,QAAQ,QAAQ,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC,sBAAsB,IAAI,IAAI;AAAA,UACtM,KAAK,eAAe,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjD,QAAQ;AAAA,UACR,QAAQ,MAAM,KAAK,aAAa;AAAA,QAClC,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,yCAA0C,IAAc,OAAO;AAAG,cAAQ,KAAK,kBAAmB,IAAc,OAAO;AAAA,IACvI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,QAAQ,WAAW,IAAI,KAAK;AAClC,YAAI,QAAQ,KAAK;AACf,iBAAO,KAAK;AAAA,YACV,IAAI,cAAc,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,YAC/C,UAAU,QAAQ,MAAM,aAAa;AAAA,YACrC,UAAU;AAAA,YACV,OAAO,2BAA2B,IAAI,OAAO;AAAA,YAC7C,aAAa,SAAS,IAAI,UAAU,IAAI,IAAI,OAAO,8BAA8B,QAAQ,KAAK,QAAQ,CAAC,CAAC;AAAA,YACxG,KAAK;AAAA;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,8CAA+C,IAAc,OAAO;AAAG,cAAQ,KAAK,uBAAwB,IAAc,OAAO;AAAA,IACjJ;AAGA,QAAI;AACF,YAAM,WAAW,MAAM,OAAO,MAAM,iEAAiE;AACrG,UAAI,SAAS,KAAK,SAAS,GAAG;AAC5B,cAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,SAQ5B;AACD,mBAAW,OAAO,EAAE,MAAM;AACxB,iBAAO,KAAK;AAAA,YACV,IAAI,aAAa,IAAI,MAAM,MAAM,GAAG,EAAE,EAAE,QAAQ,OAAO,GAAG,CAAC;AAAA,YAC3D,UAAU,WAAW,IAAI,OAAO,IAAI,MAAO,YAAY;AAAA,YACvD,UAAU;AAAA,YACV,OAAO,mBAAmB,IAAI,OAAO;AAAA,YACrC,aAAa,mBAAmB,IAAI,OAAO,WAAW,IAAI,KAAK,kBAAkB,IAAI,SAAS,OAAO,IAAI,MAAM,MAAM,GAAG,GAAG,CAAC;AAAA,YAC5H,KAAK,mBAAmB,IAAI,MAAM,MAAM,GAAG,GAAG,CAAC;AAAA,YAC/C,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,0CAA2C,IAAc,OAAO;AAAG,cAAQ,KAAK,mBAAoB,IAAc,OAAO;AAAA,IACzI;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,QAAQ,CAAC,IAAI,eAAe,CAAC,IAAI;AACvC,eAAO,KAAK;AAAA,UACV,IAAI,gBAAgB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACjD,UAAU,QAAQ,YAAY;AAAA,UAC9B,UAAU;AAAA,UACV,OAAO,UAAU,QAAQ,cAAc,SAAS,OAAO,IAAI,OAAO;AAAA,UAClE,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO,IAAI,QAAQ,4BAA4B,mCAAmC,kBAAkB,OAAO,IAAI,UAAU,EAAE,eAAe,CAAC;AAAA,UACjL,KAAK,kBAAkB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UACpD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAW5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,iBAAiB,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAClD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,wBAAwB,IAAI,OAAO;AAAA,UAC1C,aAAa,GAAG,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC7C,KAAK,WAAW,IAAI,UAAU,IAAI,IAAI,OAAO;AAAA,UAC7C,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,OAI5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,cAAM,MAAM,SAAS,IAAI,OAAO;AAChC,YAAI,MAAM,KAAe;AACvB,iBAAO,KAAK;AAAA,YACV,IAAI;AAAA,YACJ,UAAU;AAAA,YACV,UAAU;AAAA,YACV,OAAO;AAAA,YACP,aAAa,YAAY,IAAI,OAAO,4BAA4B,IAAI,eAAe,CAAC;AAAA,YACpF,KAAK;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH,WAAW,MAAM,KAAa;AAC5B,iBAAO,KAAK;AAAA,YACV,IAAI;AAAA,YACJ,UAAU;AAAA,YACV,UAAU;AAAA,YACV,OAAO;AAAA,YACP,aAAa,YAAY,IAAI,OAAO,4BAA4B,IAAI,eAAe,CAAC;AAAA,YACpF,KAAK;AAAA,YACL,QAAQ;AAAA,YACR,QAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAIA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,SAS1B,CAACC,mBAAkB,CAAC;AACvB,UAAI,EAAE,KAAK,WAAW,GAAG;AACvB,cAAM,MAAM,EAAE,KAAK,CAAC;AACpB,eAAO,KAAK;AAAA,UACV,IAAI,iBAAiB,IAAI,GAAG;AAAA,UAC5B,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,uCAAuC,IAAI,GAAG;AAAA,UACrD,aAAa,OAAO,IAAI,GAAG,SAAS,IAAI,eAAe,OAAO,KAAK,IAAI,oBAAoB,SAAS,sCAAsC,KAAK,MAAM,IAAI,eAAe,EAAE,CAAC;AAAA,UAC3K,KAAK,+BAA+B,IAAI,GAAG;AAAA,UAC3C,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH,WAAW,EAAE,KAAK,SAAS,GAAG;AAC5B,cAAM,OAAO,EAAE,KAAK,IAAI,CAAC,QAAa,IAAI,GAAG;AAC7C,cAAM,SAAS,KAAK,MAAM,EAAE,KAAK,CAAC,EAAE,eAAe,EAAE;AACrD,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,GAAG,EAAE,KAAK,MAAM,8CAA8C,MAAM;AAAA,UAC3E,aAAa,GAAG,EAAE,KAAK,MAAM,uDAAuDA,mBAAkB;AAAA,UACtG,KAAK,mIAAmIA,mBAAkB;AAAA,UAC1J,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6DAA8D,IAAc,OAAO;AAAG,cAAQ,KAAK,0BAA2B,IAAc,OAAO;AAAA,IACnK;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,gBAAgB,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAChD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,0BAA0B,IAAI,UAAU;AAAA,UAC/C,aAAa,SAAS,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAClD,KAAK,eAAe,IAAI,MAAM,IAAI,IAAI,UAAU;AAAA,UAChD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,kDAAmD,IAAc,OAAO;AAAG,cAAQ,KAAK,2BAA4B,IAAc,OAAO;AAAA,IACzJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAS5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,qBAAqB,IAAI,YAAY;AAAA,UACzC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,gBAAgB,IAAI,YAAY,KAAK,IAAI,QAAQ;AAAA,UACxD,aAAa,SAAS,IAAI,YAAY,OAAO,IAAI,OAAO,4CAA4C,IAAI,QAAQ;AAAA,UAChH,KAAK,2BAA2B,IAAI,UAAU,IAAI,IAAI,YAAY;AAAA,UAClE,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,kBAAkB,IAAI,UAAU,IAAI,IAAI,QAAQ,CAAC,CAAC;AAAA,UACtD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,wBAAwB,IAAI,UAAU;AAAA,UAC7C,aAAa,2CAA2C,IAAI,UAAU,KAAK,IAAI,QAAQ,KAAK,IAAI,CAAC,yBAAyB,IAAI,UAAU;AAAA,UACxI,KAAK;AAAA,0BAAwD,IAAI,QAAQ,MAAM,CAAC,EAAE,KAAK,6BAA6B,CAAC;AAAA,UACrH,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,+CAAgD,IAAc,OAAO;AAAG,cAAQ,KAAK,wBAAyB,IAAc,OAAO;AAAA,IACnJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAa5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,oBAAoB,IAAI,UAAU,IAAI,IAAI,WAAW;AAAA,UACzD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,8BAA8B,IAAI,UAAU,IAAI,IAAI,WAAW;AAAA,UACtE,aAAa,sBAAsB,IAAI,WAAW,OAAO,IAAI,UAAU,gBAAgB,IAAI,gBAAgB;AAAA,UAC3G,KAAK,iCAAiC,IAAI,WAAW,QAAQ,OAAO,GAAG,CAAC,IAAI,IAAI,WAAW,OAAO,IAAI,UAAU,KAAK,IAAI,WAAW;AAAA,UACpI,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,gDAAiD,IAAc,OAAO;AAAG,cAAQ,KAAK,yBAA0B,IAAc,OAAO;AAAA,IACrJ;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAkB5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,qBAAqB,IAAI,WAAW;AAAA,UACxC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,sBAAsB,IAAI,WAAW,mBAAmB,IAAI,YAAY;AAAA,UAC/E,aAAa,OAAO,IAAI,WAAW,sCAAsC,IAAI,YAAY,aAAa,IAAI,iBAAiB,IAAI,MAAM,GAAG,GAAG,CAAC;AAAA,UAC5I,KAAK,4BAA4B,IAAI,YAAY;AAAA,UACjD,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,mCAAoC,IAAc,OAAO;AAAG,cAAQ,KAAK,YAAa,IAAc,OAAO;AAAA,IAC3H;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,OAI5B;AACD,YAAM,WAAW,SAAS,EAAE,KAAK,CAAC,GAAG,aAAa,GAAG;AACrD,UAAI,WAAW,SAAS;AACtB,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU,WAAW,YAAY,aAAa;AAAA,UAC9C,UAAU;AAAA,UACV,OAAO,qBAAqB,WAAW,SAAS,QAAQ,CAAC,CAAC;AAAA,UAC1D,aAAa,6BAA6B,WAAW,SAAS,QAAQ,CAAC,CAAC;AAAA,UACxE,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,6CAA8C,IAAc,OAAO;AAAG,cAAQ,KAAK,sBAAuB,IAAc,OAAO;AAAA,IAC/I;AAGA,QAAI;AACF,YAAM,iBAAiB,aAAa,OAAS,yBAAyB;AACtE,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA,eAIpB,cAAc;AAAA,OACtB;AACD,YAAM,SAAS,WAAW,EAAE,KAAK,CAAC,GAAG,WAAW,GAAG;AACnD,UAAI,SAAS,IAAI;AACf,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU,SAAS,KAAK,YAAY;AAAA,UACpC,UAAU;AAAA,UACV,OAAO,GAAG,MAAM;AAAA,UAChB,aAAa,GAAG,EAAE,KAAK,CAAC,GAAG,eAAe,iBAAiB,EAAE,KAAK,CAAC,GAAG,iBAAiB;AAAA,UACvF,KAAK;AAAA;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,kDAAmD,IAAc,OAAO;AAAG,cAAQ,KAAK,2BAA4B,IAAc,OAAO;AAAA,IACzJ;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,2DAA2D;AACxF,UAAI,EAAE,KAAK,CAAC,GAAG,YAAY,OAAO;AAChC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO;AAAA,UACP,aAAa;AAAA,UACb,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,wCAAyC,IAAc,OAAO;AAAG,cAAQ,KAAK,iBAAkB,IAAc,OAAO;AAAA,IACrI;AAGA,QAAI;AACF,YAAM,QAAQ,MAAM,OAAO,MAAM,qEAAqE;AACtG,YAAM,SAAS,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA,OAGjC;AACD,YAAM,cAAc,SAAS,OAAO,KAAK,CAAC,GAAG,gBAAgB,GAAG;AAGhE,UAAI,cAAc,KAAK,cAAc,MAAM,OAAO,MAAM;AACtD,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,2BAA2B,cAAc,SAAS,QAAQ,CAAC,CAAC;AAAA,UACnE,aAAa,4BAA4B,MAAM,KAAK,CAAC,GAAG,OAAO,GAAG,MAAM,KAAK,CAAC,GAAG,QAAQ,EAAE;AAAA,UAC3F,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,4CAA6C,IAAc,OAAO;AAAG,cAAQ,KAAK,qBAAsB,IAAc,OAAO;AAAA,IAC7I;AAEA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,+DAA+D;AAC5F,YAAM,YAAY,SAAS,EAAE,KAAK,CAAC,GAAG,WAAW,GAAG;AACpD,UAAI,YAAY,KAAK,YAAY,MAAM;AACrC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,oBAAoB,YAAY,OAAO,YAAY,QAAQ,YAAY,MAAM,QAAQ,CAAC,IAAI,IAAI;AAAA,UACrG,aAAa,eAAe,EAAE,KAAK,CAAC,GAAG,OAAO,GAAG,EAAE,KAAK,CAAC,GAAG,QAAQ,EAAE;AAAA,UACtE,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,sCAAuC,IAAc,OAAO;AAAG,cAAQ,KAAK,eAAgB,IAAc,OAAO;AAAA,IACjI;AAKA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAO5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,wBAAwB,IAAI,GAAG;AAAA,UACnC,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,aAAa,IAAI,OAAO,mBAAmB,IAAI,WAAW;AAAA,UACjE,aAAa,aAAa,IAAI,OAAO,wDAAwD,IAAI,WAAW;AAAA,UAC5G,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,mDAAoD,IAAc,OAAO;AAAG,cAAQ,KAAK,4BAA6B,IAAc,OAAO;AAAA,IAC3J;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM,oDAAoD;AACjF,UAAI,EAAE,KAAK,CAAC,GAAG,YAAY,OAAO;AAChC,eAAO,KAAK;AAAA,UACV,IAAI;AAAA,UACJ,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO;AAAA,UACP,aAAa;AAAA,UACb,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,uCAAwC,IAAc,OAAO;AAAG,cAAQ,KAAK,gBAAiB,IAAc,OAAO;AAAA,IACnI;AAGA,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,OAK5B;AACD,iBAAW,OAAO,EAAE,MAAM;AACxB,eAAO,KAAK;AAAA,UACV,IAAI,kBAAkB,IAAI,QAAQ,IAAI,IAAI,SAAS;AAAA,UACnD,UAAU;AAAA,UACV,UAAU;AAAA,UACV,OAAO,4BAA4B,IAAI,SAAS,IAAI,IAAI,QAAQ;AAAA,UAChE,aAAa,0DAA0D,IAAI,IAAI,mBAAmB,IAAI,QAAQ,OAAO,IAAI,SAAS;AAAA,UAClI,KAAK;AAAA;AAAA,UACL,QAAQ;AAAA,UACR,QAAQ;AAAA,QACV,CAAC;AAAA,MACH;AAAA,IACF,SAAS,KAAK;AACZ,cAAQ,MAAM,wCAAyC,IAAc,OAAO;AAAG,cAAQ,KAAK,iBAAkB,IAAc,OAAO;AAAA,IACrI;AAGA,UAAM,aAAa,iBAAiB;AACpC,UAAM,aAAa,IAAI,IAAI,UAAU;AACrC,UAAM,eAAe,OAAO,OAAO,OAAK,CAAC,WAAW,IAAI,EAAE,EAAE,CAAC;AAC7D,UAAM,eAAe,OAAO,SAAS,aAAa;AAGlD,UAAM,aAAyB,CAAC;AAChC,UAAM,SAAS,oBAAI,IAA4B;AAC/C,eAAW,SAAS,cAAc;AAEhC,YAAM,SAAS,MAAM,GAAG,QAAQ,WAAW,EAAE;AAC7C,UAAI,CAAC,OAAO,IAAI,MAAM,EAAG,QAAO,IAAI,QAAQ,CAAC,CAAC;AAC9C,aAAO,IAAI,MAAM,EAAG,KAAK,KAAK;AAAA,IAChC;AACA,UAAM,eAAuC;AAAA,MAC3C,oBAAoB;AAAA,MACpB,qBAAqB;AAAA,MACrB,gBAAgB;AAAA,MAChB,gBAAgB;AAAA,MAChB,iBAAiB;AAAA,MACjB,oBAAoB;AAAA,MACpB,cAAc;AAAA,IAChB;AACA,eAAW,CAAC,QAAQ,KAAK,KAAK,QAAQ;AACpC,UAAI,MAAM,UAAU,EAAG;AACvB,YAAM,QAAQ,aAAa,MAAM,KAAK,WAAW,MAAM,MAAM,IAAI,MAAM;AACvE,YAAM,MAAM,MAAM,IAAI,OAAK,EAAE,IAAI,MAAM,IAAI,EAAE,OAAO,OAAK,CAAC,EAAE,KAAK,EAAE,WAAW,IAAI,CAAC,EAAE,KAAK,IAAI,EAAE,KAAK,CAAC,EAAE,OAAO,OAAO,EAAE,KAAK,KAAK,IAAI;AACtI,iBAAW,KAAK,EAAE,MAAM,QAAQ,OAAO,GAAG,KAAK,KAAK,MAAM,MAAM,KAAK,OAAO,MAAM,QAAQ,IAAI,CAAC;AAAA,IACjG;AAEA,UAAM,QAAQ,oBAAoB,YAAY;AAC9C,WAAO;AAAA,MACL;AAAA,MACA,OAAO,eAAe,KAAK;AAAA,MAC3B,QAAQ;AAAA,MACR,WAAW,iBAAiB,YAAY;AAAA,MACxC;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;AASA,IAAI,aAAiD;AAErD,SAAS,eAA4C;AACnD,MAAI,WAAY,QAAO;AACvB,QAAMC,WAAU,QAAQ,IAAI,oBAAoB,KAAK,KAAK,GAAG,QAAQ,GAAG,UAAU;AAClF,KAAG,UAAUA,UAAS,EAAE,WAAW,KAAK,CAAC;AACzC,QAAM,SAAS,KAAK,KAAKA,UAAS,WAAW;AAC7C,eAAa,IAAI,SAAS,MAAM;AAChC,aAAW,OAAO,oBAAoB;AACtC,aAAW,KAAK,2FAA2F;AAC3G,SAAO;AACT;AAEO,SAAS,mBAA6B;AAC3C,MAAI;AACF,UAAM,KAAK,aAAa;AACxB,WAAO,GAAG,QAAQ,qCAAqC,EAAE,IAAI,EAAE,IAAI,CAAC,MAAW,EAAE,QAAQ;AAAA,EAC3F,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAcO,SAAS,UAAU,KAAsB;AAC9C,QAAM,UAAU,IAAI,KAAK;AACzB,MAAI,CAAC,QAAS,QAAO;AAGrB,QAAM,aAAa,QAAQ,QAAQ,SAAS,EAAE,EAAE,MAAM,GAAG,EAAE,IAAI,OAAK,EAAE,KAAK,CAAC,EAAE,OAAO,OAAO;AAC5F,MAAI,WAAW,WAAW,EAAG,QAAO;AAEpC,QAAM,QAAQ,WAAW,CAAC,EAAE,YAAY;AAGxC,MAAI,MAAM,WAAW,iBAAiB,GAAG;AACvC,UAAM,eAAe,MAAM,QAAQ,yBAAyB,EAAE,EAAE,UAAU;AAC1E,WAAO,aAAa,WAAW,QAAQ;AAAA,EACzC;AAGA,QAAM,mBAAmB;AAAA,IACvB;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAEA,SAAO,iBAAiB,KAAK,CAAC,MAAM,MAAM,WAAW,CAAC,CAAC;AACzD;;;ACxyBA,eAAsB,eAAeC,OAAkC;AACrE,QAAM,SAAS,MAAMA,MAAK,QAAQ;AAClC,MAAI;AAEF,UAAM,WAAW,MAAM,OAAO;AAAA,MAC5B;AAAA,IACF;AACA,QAAI,SAAS,KAAK,WAAW,GAAG;AAC9B,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAe5B;AACD,WAAO,EAAE;AAAA,EACX,QAAQ;AAEN,WAAO,CAAC;AAAA,EACV,UAAE;AACA,WAAO,QAAQ;AAAA,EACjB;AACF;;;AC/CA,OAAOC,SAAQ;AACf,OAAOC,WAAU;AAwBjB,SAAS,iBAAiB,IAAoB;AAC5C,SAAO,GAAG,QAAQ,SAAS,EAAE;AAC/B;AAQO,SAAS,aAAa,cAAsB,QAA6B;AAC9E,EAAAD,IAAG,UAAUC,MAAK,QAAQ,YAAY,GAAG,EAAE,WAAW,KAAK,CAAC;AAC5D,QAAM,WAAqB,EAAE,YAAW,oBAAI,KAAK,GAAE,YAAY,GAAG,OAAO;AACzE,EAAAD,IAAG,cAAc,cAAc,KAAK,UAAU,UAAU,MAAM,CAAC,CAAC;AAClE;AAQO,SAAS,aAAa,cAAuC;AAClE,MAAI,CAACA,IAAG,WAAW,YAAY,EAAG,QAAO;AACzC,MAAI;AACF,WAAO,KAAK,MAAMA,IAAG,aAAa,cAAc,OAAO,CAAC;AAAA,EAC1D,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAEO,SAAS,cAAc,MAAqB,SAAsC;AAGvF,QAAM,cAAc,IAAI,IAAI,KAAK,OAAO,IAAI,CAAC,MAAM,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAC1E,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAE7E,QAAM,YAAY,QAAQ,OAAO,OAAO,CAAC,MAAM,CAAC,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AACvF,QAAM,iBAAiB,KAAK,OAAO,OAAO,CAAC,MAAM,CAAC,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AACzF,QAAM,YAAY,QAAQ,OAAO,OAAO,CAAC,MAAM,YAAY,IAAI,iBAAiB,EAAE,EAAE,CAAC,CAAC;AAEtF,SAAO;AAAA,IACL,YAAY,QAAQ,QAAQ,KAAK;AAAA,IACjC,eAAe,KAAK;AAAA,IACpB,cAAc,QAAQ;AAAA,IACtB,eAAe,KAAK;AAAA,IACpB,cAAc,QAAQ;AAAA,IACtB;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5EA,SAAS,YAAY;;;ACgDd,SAAS,oBAAoB,SAAyB,SAAyC;AACpG,QAAM,UAA0B,CAAC;AAEjC,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACnF,QAAM,cAAc,IAAI,IAAI,QAAQ,OAAO,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AAGnF,aAAW,CAAC,KAAK,CAAC,KAAK,aAAa;AAClC,QAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,GAAG,SAAS,CAAC;AAAA,IAC5G;AAAA,EACF;AACA,aAAW,CAAC,GAAG,KAAK,aAAa;AAC/B,QAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,GAAG,WAAW,CAAC;AAAA,IAChH;AAAA,EACF;AAGA,aAAW,CAAC,KAAK,QAAQ,KAAK,aAAa;AACzC,UAAM,WAAW,YAAY,IAAI,GAAG;AACpC,QAAI,CAAC,SAAU;AAGf,UAAM,UAAU,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAChE,UAAM,UAAU,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAEhE,eAAW,CAAC,MAAM,GAAG,KAAK,SAAS;AACjC,YAAM,SAAS,QAAQ,IAAI,IAAI;AAC/B,UAAI,CAAC,QAAQ;AACX,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,WAAW,IAAI,IAAI,IAAI,CAAC;AAAA,MAC7H,OAAO;AACL,YAAI,OAAO,SAAS,IAAI,MAAM;AAC5B,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,kBAAkB,OAAO,IAAI,WAAM,IAAI,IAAI,GAAG,CAAC;AAAA,QACvJ;AACA,YAAI,OAAO,aAAa,IAAI,UAAU;AACpC,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,sBAAsB,OAAO,QAAQ,WAAM,IAAI,QAAQ,GAAG,CAAC;AAAA,QACnK;AACA,YAAI,OAAO,kBAAkB,IAAI,eAAe;AAC9C,kBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,qBAAqB,OAAO,iBAAiB,MAAM,WAAM,IAAI,iBAAiB,MAAM,GAAG,CAAC;AAAA,QAChM;AAAA,MACF;AAAA,IACF;AACA,eAAW,QAAQ,QAAQ,KAAK,GAAG;AACjC,UAAI,CAAC,QAAQ,IAAI,IAAI,GAAG;AACtB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,UAAU,YAAY,KAAK,QAAQ,UAAU,IAAI,WAAW,CAAC;AAAA,MACnH;AAAA,IACF;AAGA,UAAM,SAAS,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAC/D,UAAM,SAAS,IAAI,IAAI,SAAS,QAAQ,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAC/D,eAAW,CAAC,MAAM,GAAG,KAAK,QAAQ;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,SAAS,CAAC;AAAA,MAC7G,WAAW,OAAO,IAAI,IAAI,EAAG,eAAe,IAAI,YAAY;AAC1D,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,sBAAsB,CAAC;AAAA,MAC7H;AAAA,IACF;AACA,eAAW,QAAQ,OAAO,KAAK,GAAG;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,SAAS,YAAY,KAAK,QAAQ,SAAS,IAAI,WAAW,CAAC;AAAA,MACjH;AAAA,IACF;AAGA,UAAM,SAAS,IAAI,IAAI,SAAS,YAAY,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AACnE,UAAM,SAAS,IAAI,IAAI,SAAS,YAAY,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AACnE,eAAW,CAAC,MAAM,GAAG,KAAK,QAAQ;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,WAAW,IAAI,IAAI,IAAI,CAAC;AAAA,MACrI,WAAW,OAAO,IAAI,IAAI,EAAG,eAAe,IAAI,YAAY;AAC1D,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,sBAAsB,CAAC;AAAA,MACvI;AAAA,IACF;AACA,eAAW,QAAQ,OAAO,KAAK,GAAG;AAChC,UAAI,CAAC,OAAO,IAAI,IAAI,GAAG;AACrB,gBAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,cAAc,YAAY,KAAK,QAAQ,cAAc,IAAI,WAAW,CAAC;AAAA,MAC3H;AAAA,IACF;AAAA,EACF;AAGA,QAAM,WAAW,IAAI,KAAK,QAAQ,SAAS,CAAC,GAAG,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACvF,QAAM,WAAW,IAAI,KAAK,QAAQ,SAAS,CAAC,GAAG,IAAI,CAAC,MAAM,CAAC,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,IAAI,CAAC,CAAC,CAAC;AACvF,aAAW,CAAC,KAAK,EAAE,KAAK,UAAU;AAChC,UAAM,QAAQ,SAAS,IAAI,GAAG;AAC9B,QAAI,CAAC,OAAO;AACV,cAAQ,KAAK,EAAE,aAAa,SAAS,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,WAAW,GAAG,OAAO,KAAK,IAAI,CAAC,IAAI,CAAC;AAAA,IACrI,OAAO;AACL,YAAM,QAAQ,GAAG,OAAO,OAAO,CAAC,MAAM,CAAC,MAAM,OAAO,SAAS,CAAC,CAAC;AAC/D,YAAM,UAAU,MAAM,OAAO,OAAO,CAAC,MAAM,CAAC,GAAG,OAAO,SAAS,CAAC,CAAC;AACjE,iBAAW,KAAK,OAAO;AACrB,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,YAAY,CAAC,UAAU,CAAC;AAAA,MAC5H;AACA,iBAAW,KAAK,SAAS;AACvB,gBAAQ,KAAK,EAAE,aAAa,YAAY,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,YAAY,CAAC,YAAY,CAAC;AAAA,MAC9H;AAAA,IACF;AAAA,EACF;AACA,aAAW,OAAO,SAAS,KAAK,GAAG;AACjC,QAAI,CAAC,SAAS,IAAI,GAAG,GAAG;AACtB,cAAQ,KAAK,EAAE,aAAa,WAAW,aAAa,QAAQ,YAAY,MAAM,QAAQ,QAAQ,GAAG,WAAW,CAAC;AAAA,IAC/G;AAAA,EACF;AAEA,SAAO;AACT;;;ACjJA,eAAsB,kBAAkBE,OAAqC;AAC3E,QAAM,SAAS,MAAM,gBAAgBA,KAAI;AACzC,QAAM,QAAQ,MAAM,eAAeA,KAAI;AAEvC,QAAM,iBAAiB,MAAM,QAAQ;AAAA,IACnC,OAAO,IAAI,OAAO,MAAW;AAC3B,YAAM,SAAS,MAAM,qBAAqBA,OAAM,GAAG,EAAE,MAAM,IAAI,EAAE,IAAI,EAAE;AACvE,UAAI,CAAC,OAAQ,QAAO;AACpB,aAAO;AAAA,QACL,MAAM,OAAO;AAAA,QACb,QAAQ,OAAO;AAAA,QACf,SAAS,OAAO,QAAQ,IAAI,CAAC,OAAY;AAAA,UACvC,MAAM,EAAE;AAAA,UACR,MAAM,EAAE;AAAA,UACR,UAAU,EAAE;AAAA,UACZ,eAAe,EAAE;AAAA,QACnB,EAAE;AAAA,QACF,SAAS,OAAO,QAAQ,IAAI,CAAC,OAAY;AAAA,UACvC,MAAM,EAAE;AAAA,UACR,YAAY,EAAE;AAAA,UACd,WAAW,EAAE;AAAA,UACb,YAAY,EAAE;AAAA,QAChB,EAAE;AAAA,QACF,aAAa,OAAO,YAAY,IAAI,CAAC,OAAY;AAAA,UAC/C,MAAM,EAAE;AAAA,UACR,MAAM,EAAE;AAAA,UACR,YAAY,EAAE;AAAA,QAChB,EAAE;AAAA,MACJ;AAAA,IACF,CAAC;AAAA,EACH;AAEA,SAAO;AAAA,IACL,QAAQ,eAAe,OAAO,OAAO;AAAA,IACrC,OAAO,MAAM,IAAI,CAAC,OAAY,EAAE,MAAM,EAAE,MAAM,QAAQ,EAAE,QAAQ,QAAQ,EAAE,OAAO,EAAE;AAAA,EACrF;AACF;;;AF+DA,eAAe,YAAYC,OAA+B;AACxD,QAAM,MAAM,MAAMA,MAAK,MAA8B;AAAA;AAAA;AAAA;AAAA;AAAA,GAKpD;AACD,SAAO,IAAI,KAAK,IAAI,CAAC,MAAM,EAAE,UAAU;AACzC;AAEA,eAAe,aAAaA,OAAkC;AAC5D,QAAM,MAAM,MAAMA,MAAK,MAAiB;AAAA;AAAA;AAAA;AAAA;AAAA,GAKvC;AACD,SAAO,IAAI;AACb;AAEA,eAAe,aAAaA,OAAiC;AAC3D,QAAM,MAAM,MAAMA,MAAK,MAAgB;AAAA;AAAA;AAAA;AAAA;AAAA,GAKtC;AACD,SAAO,IAAI;AACb;AAIA,SAAS,WAAW,cAAwB,cAA4E;AACtH,QAAM,YAAY,IAAI,IAAI,YAAY;AACtC,QAAM,YAAY,IAAI,IAAI,YAAY;AACtC,SAAO;AAAA,IACL,eAAe,aAAa,OAAO,CAAC,MAAM,CAAC,UAAU,IAAI,CAAC,CAAC;AAAA,IAC3D,aAAa,aAAa,OAAO,CAAC,MAAM,CAAC,UAAU,IAAI,CAAC,CAAC;AAAA,EAC3D;AACF;AAEA,SAAS,oBAAoB,SAA4D;AACvF,QAAM,MAAM,oBAAI,IAAqC;AACrD,aAAW,OAAO,SAAS;AACzB,QAAI,CAAC,IAAI,IAAI,IAAI,UAAU,EAAG,KAAI,IAAI,IAAI,YAAY,oBAAI,IAAI,CAAC;AAC/D,UAAM,OAAmB;AAAA,MACvB,MAAM,IAAI;AAAA,MACV,MAAM,IAAI;AAAA,MACV,UAAU,IAAI,gBAAgB;AAAA,IAChC;AACA,QAAI,IAAI,mBAAmB,QAAQ,IAAI,mBAAmB,QAAW;AACnE,WAAK,UAAU,IAAI;AAAA,IACrB;AACA,QAAI,IAAI,IAAI,UAAU,EAAG,IAAI,IAAI,aAAa,IAAI;AAAA,EACpD;AACA,SAAO;AACT;AAEA,SAAS,YACP,YACA,YACA,cACc;AACd,QAAM,gBAAgB,oBAAoB,UAAU;AACpD,QAAM,gBAAgB,oBAAoB,UAAU;AACpD,QAAM,QAAsB,CAAC;AAE7B,aAAW,SAAS,cAAc;AAChC,UAAM,SAAS,cAAc,IAAI,KAAK,KAAK,oBAAI,IAAwB;AACvE,UAAM,SAAS,cAAc,IAAI,KAAK,KAAK,oBAAI,IAAwB;AAEvE,UAAM,iBAA+B,CAAC;AACtC,UAAM,eAA6B,CAAC;AACpC,UAAM,YAA8B,CAAC;AACrC,UAAM,gBAAsC,CAAC;AAC7C,UAAM,eAAoC,CAAC;AAE3C,eAAW,CAAC,SAAS,OAAO,KAAK,QAAQ;AACvC,UAAI,CAAC,OAAO,IAAI,OAAO,GAAG;AACxB,uBAAe,KAAK,OAAO;AAAA,MAC7B,OAAO;AACL,cAAM,UAAU,OAAO,IAAI,OAAO;AAClC,YAAI,QAAQ,SAAS,QAAQ,MAAM;AACjC,oBAAU,KAAK,EAAE,QAAQ,SAAS,YAAY,QAAQ,MAAM,YAAY,QAAQ,KAAK,CAAC;AAAA,QACxF;AACA,YAAI,QAAQ,aAAa,QAAQ,UAAU;AACzC,wBAAc,KAAK,EAAE,QAAQ,SAAS,gBAAgB,QAAQ,UAAU,gBAAgB,QAAQ,SAAS,CAAC;AAAA,QAC5G;AACA,aAAK,QAAQ,WAAW,WAAW,QAAQ,WAAW,OAAO;AAC3D,uBAAa,KAAK,EAAE,QAAQ,SAAS,eAAe,QAAQ,WAAW,MAAM,eAAe,QAAQ,WAAW,KAAK,CAAC;AAAA,QACvH;AAAA,MACF;AAAA,IACF;AAEA,eAAW,CAAC,SAAS,OAAO,KAAK,QAAQ;AACvC,UAAI,CAAC,OAAO,IAAI,OAAO,GAAG;AACxB,qBAAa,KAAK,OAAO;AAAA,MAC3B;AAAA,IACF;AAEA,QAAI,eAAe,SAAS,KAAK,aAAa,SAAS,KAAK,UAAU,SAAS,KAC3E,cAAc,SAAS,KAAK,aAAa,SAAS,GAAG;AACvD,YAAM,KAAK,EAAE,OAAO,gBAAgB,cAAc,WAAW,eAAe,aAAa,CAAC;AAAA,IAC5F;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,oBAAoB,SAAuD;AAClF,QAAM,MAAM,oBAAI,IAAiC;AACjD,aAAW,OAAO,SAAS;AACzB,QAAI,CAAC,IAAI,IAAI,IAAI,SAAS,EAAG,KAAI,IAAI,IAAI,WAAW,oBAAI,IAAI,CAAC;AAC7D,QAAI,IAAI,IAAI,SAAS,EAAG,IAAI,IAAI,WAAW,IAAI,QAAQ;AAAA,EACzD;AACA,SAAO;AACT;AAEA,SAAS,YACP,YACA,YACA,cACa;AACb,QAAM,aAAa,oBAAoB,UAAU;AACjD,QAAM,aAAa,oBAAoB,UAAU;AACjD,QAAM,QAAqB,CAAC;AAG5B,QAAM,YAAY,oBAAI,IAAI;AAAA,IACxB,GAAG,WAAW,IAAI,CAAC,MAAM,EAAE,SAAS;AAAA,IACpC,GAAG,WAAW,IAAI,CAAC,MAAM,EAAE,SAAS;AAAA,EACtC,CAAC;AAED,aAAW,SAAS,WAAW;AAE7B,QAAI,CAAC,aAAa,SAAS,KAAK,EAAG;AAEnC,UAAM,SAAS,WAAW,IAAI,KAAK,KAAK,oBAAI,IAAoB;AAChE,UAAM,SAAS,WAAW,IAAI,KAAK,KAAK,oBAAI,IAAoB;AAEhE,UAAM,iBAAiB,CAAC,GAAG,OAAO,KAAK,CAAC,EAAE,OAAO,CAAC,MAAM,CAAC,OAAO,IAAI,CAAC,CAAC;AACtE,UAAM,eAAe,CAAC,GAAG,OAAO,KAAK,CAAC,EAAE,OAAO,CAAC,MAAM,CAAC,OAAO,IAAI,CAAC,CAAC;AACpE,UAAM,kBAAkC,CAAC;AAEzC,eAAW,CAAC,MAAM,MAAM,KAAK,QAAQ;AACnC,UAAI,OAAO,IAAI,IAAI,GAAG;AACpB,cAAM,SAAS,OAAO,IAAI,IAAI;AAC9B,YAAI,WAAW,QAAQ;AACrB,0BAAgB,KAAK,EAAE,MAAM,WAAW,QAAQ,WAAW,OAAO,CAAC;AAAA,QACrE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,eAAe,SAAS,KAAK,aAAa,SAAS,KAAK,gBAAgB,SAAS,GAAG;AACtF,YAAM,KAAK,EAAE,OAAO,gBAAgB,cAAc,gBAAgB,CAAC;AAAA,IACrE;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,kBAAkB,QAA4B;AACrD,MAAI,IAAI,OAAO,cAAc,SAAS,OAAO,YAAY;AACzD,aAAW,MAAM,OAAO,aAAa;AACnC,SAAK,GAAG,eAAe,SAAS,GAAG,aAAa,SAAS,GAAG,UAAU,SACjE,GAAG,cAAc,SAAS,GAAG,aAAa;AAAA,EACjD;AACA,aAAW,MAAM,OAAO,YAAY;AAClC,SAAK,GAAG,eAAe,SAAS,GAAG,aAAa,SAAS,GAAG,gBAAgB;AAAA,EAC9E;AACA,QAAM,OAAO,mBAAmB,CAAC,GAAG;AACpC,QAAM,OAAO,aAAa,CAAC,GAAG;AAC9B,SAAO;AACT;AAIA,eAAsB,iBACpB,YACA,YACA,SACwB;AACxB,QAAM,aAAa,IAAI,KAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AAC5F,QAAM,aAAa,IAAI,KAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AAE5F,MAAI;AAEF,UAAM;AAAA,MACJ;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,IAAI,MAAM,QAAQ,IAAI;AAAA,MACpB,YAAY,UAAU;AAAA,MACtB,YAAY,UAAU;AAAA,MACtB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,aAAa,UAAU;AAAA,MACvB,kBAAkB,UAAU,EAAE,MAAM,MAAM,IAAI;AAAA,MAC9C,kBAAkB,UAAU,EAAE,MAAM,MAAM,IAAI;AAAA,IAChD,CAAC;AAED,UAAM,EAAE,eAAe,YAAY,IAAI,WAAW,cAAc,YAAY;AAC5E,UAAM,YAAY,IAAI,IAAI,YAAY;AACtC,UAAM,eAAe,aAAa,OAAO,CAAC,MAAM,UAAU,IAAI,CAAC,CAAC;AAEhE,UAAM,cAAc,YAAY,YAAY,YAAY,YAAY;AACpE,UAAM,aAAa,YAAY,YAAY,YAAY,YAAY;AAGnE,UAAM,kBAAoC,CAAC;AAC3C,UAAM,YAAwB,CAAC;AAE/B,QAAI,cAAc,YAAY;AAI5B,YAAM,cAAc,oBAAoB,YAAY,UAAU;AAE9D,iBAAW,KAAK,aAAa;AAC3B,YAAI,EAAE,gBAAgB,cAAc;AAClC,0BAAgB,KAAK;AAAA,YACnB,OAAO,EAAE,cAAc;AAAA,YACvB,MAAM,EAAE,gBAAgB,UAAU,UAAU,EAAE,gBAAgB,YAAY,YAAY;AAAA,YACtF,MAAM,EAAE,OAAO,MAAM,GAAG,EAAE,CAAC,KAAK,EAAE;AAAA,YAClC,QAAQ,EAAE;AAAA,UACZ,CAAC;AAAA,QACH,WAAW,EAAE,gBAAgB,QAAQ;AACnC,oBAAU,KAAK;AAAA,YACb,MAAM,EAAE,gBAAgB,UAAU,UAAU,EAAE,gBAAgB,YAAY,YAAY;AAAA,YACtF,MAAM,EAAE,OAAO,MAAM,GAAG,EAAE,CAAC,KAAK,EAAE;AAAA,YAClC,QAAQ,EAAE;AAAA,UACZ,CAAC;AAAA,QACH;AAAA,MACF;AAAA,IACF;AAEA,UAAM,SAAqB,EAAE,eAAe,aAAa,aAAa,YAAY,iBAAiB,UAAU;AAC7G,UAAM,eAAe,kBAAkB,MAAM;AAE7C,QAAI;AAEJ,QAAI,SAAS,eAAe;AAC1B,YAAMC,sBAAqB;AAC3B,YAAM,CAAC,WAAW,SAAS,IAAI,MAAM,QAAQ,IAAI;AAAA,QAC/C,iBAAiB,YAAYA,mBAAkB;AAAA,QAC/C,iBAAiB,YAAYA,mBAAkB;AAAA,MACjD,CAAC;AAED,YAAM,eAAe,IAAI,IAAI,UAAU,OAAO,IAAI,CAAC,MAAM,EAAE,KAAK,CAAC;AACjE,YAAM,eAAe,IAAI,IAAI,UAAU,OAAO,IAAI,CAAC,MAAM,EAAE,KAAK,CAAC;AAEjE,YAAM,mBAAmB,UAAU,OAChC,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,KAAK,CAAC,EACxC,IAAI,CAAC,MAAM,GAAG,EAAE,QAAQ,KAAK,EAAE,KAAK,EAAE;AAEzC,YAAM,mBAAmB,UAAU,OAChC,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,KAAK,CAAC,EACxC,IAAI,CAAC,MAAM,GAAG,EAAE,QAAQ,KAAK,EAAE,KAAK,EAAE;AAEzC,eAAS;AAAA,QACP,QAAQ,EAAE,OAAO,UAAU,OAAO,OAAO,UAAU,OAAO,KAAK,qBAAqB,UAAU,EAAE;AAAA,QAChG,QAAQ,EAAE,OAAO,UAAU,OAAO,OAAO,UAAU,OAAO,KAAK,qBAAqB,UAAU,EAAE;AAAA,QAChG;AAAA,QACA;AAAA,MACF;AAAA,IACF;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,MAClC,SAAS;AAAA,QACP;AAAA,QACA,WAAW,iBAAiB;AAAA,MAC9B;AAAA,IACF;AAAA,EACF,UAAE;AACA,UAAM,QAAQ,WAAW,CAAC,WAAW,IAAI,GAAG,WAAW,IAAI,CAAC,CAAC;AAAA,EAC/D;AACF;AAGA,SAAS,qBAAqB,SAAyB;AACrD,MAAI;AACF,UAAM,MAAM,IAAI,IAAI,OAAO;AAC3B,QAAI,IAAI,SAAU,KAAI,WAAW;AACjC,WAAO,IAAI,SAAS;AAAA,EACtB,QAAQ;AACN,WAAO;AAAA,EACT;AACF;;;AG7VA,SAAS,aAAa,MAAW,MAAa,CAAC,GAAU;AACvD,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,MAAI,KAAK,IAAI;AACb,QAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,OAAO;AAC3C,MAAI,MAAM,QAAQ,KAAK,GAAG;AACxB,eAAW,SAAS,MAAO,cAAa,OAAO,GAAG;AAAA,EACpD;AACA,SAAO;AACT;AAWA,SAAS,yBAAyB,QAA0B;AAE1D,QAAM,aAAa;AACnB,QAAM,QAAQ,oBAAI,IAAY;AAC9B,MAAI;AACJ,UAAQ,IAAI,WAAW,KAAK,MAAM,OAAO,MAAM;AAC7C,UAAM,MAAM,EAAE,CAAC,EAAE,YAAY;AAE7B,QAAI,CAAC,CAAC,OAAO,MAAM,OAAO,QAAQ,SAAS,MAAM,EAAE,SAAS,GAAG,GAAG;AAChE,YAAM,IAAI,GAAG;AAAA,IACf;AAAA,EACF;AACA,SAAO,MAAM,KAAK,KAAK;AACzB;AAMA,eAAe,wBAAwBC,OAAY,WAAwC;AACzF,MAAI;AAEF,UAAM,IAAI,MAAMA,MAAK;AAAA,MACnB;AAAA,MACA,CAAC,SAAS;AAAA,IACZ;AACA,WAAO,EAAE,KAAK,IAAI,CAAC,QAAa;AAE9B,YAAM,IAAI,cAAc,KAAK,IAAI,QAAQ;AACzC,UAAI,CAAC,EAAG,QAAO,CAAC;AAChB,aAAO,EAAE,CAAC,EACP,MAAM,GAAG,EACT,IAAI,CAAC,MAAc,EAAE,KAAK,EAAE,QAAQ,UAAU,EAAE,EAAE,YAAY,CAAC;AAAA,IACpE,CAAC;AAAA,EACH,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;AAKA,SAAS,YAAY,UAA6C;AAChE,MAAI,WAAW,IAAS,QAAO;AAC/B,MAAI,YAAY,IAAQ,QAAO;AAC/B,SAAO;AACT;AAKA,SAAS,QAAQ,GAAmB;AAClC,MAAI,KAAK,IAAW,QAAO,IAAI,IAAI,KAAW,QAAQ,CAAC,CAAC;AACxD,MAAI,KAAK,IAAO,QAAO,IAAI,IAAI,KAAO,QAAQ,CAAC,CAAC;AAChD,SAAO,OAAO,CAAC;AACjB;AAUA,eAAsB,mBACpB,aACAA,OAC0B;AAC1B,QAAM,SAA0B;AAAA,IAC9B,WAAW,CAAC;AAAA,IACZ,UAAU,CAAC;AAAA,IACX,gBAAgB,CAAC;AAAA,IACjB,cAAc,EAAE,WAAW,EAAE;AAAA,IAC7B,iBAAiB,CAAC;AAAA,EACpB;AAEA,MAAI,CAAC,eAAe,CAAC,MAAM,QAAQ,WAAW,KAAK,YAAY,WAAW,GAAG;AAC3E,WAAO;AAAA,EACT;AAEA,QAAM,WAAW,YAAY,CAAC;AAC9B,QAAM,WAAW,WAAW,MAAM,KAAK,WAAW,MAAM;AAGxD,QAAM,eAAmC,WAAW,eAAe,KAAK;AACxE,QAAM,gBAAoC,WAAW,gBAAgB,KAAK;AAE1E,MAAI,CAAC,SAAU,QAAO;AAGtB,QAAM,WAAW,aAAa,QAAQ;AAGtC,SAAO,YAAY,SAAS,IAAI,CAAC,MAAW;AAC1C,UAAM,IAAqB;AAAA,MACzB,UAAU,EAAE,WAAW,KAAK;AAAA,MAC5B,WAAW,EAAE,YAAY,KAAK;AAAA,IAChC;AACA,QAAI,EAAE,eAAe,EAAG,GAAE,QAAQ,EAAE,eAAe;AACnD,QAAI,EAAE,aAAa,MAAM,OAAW,GAAE,aAAa,EAAE,aAAa;AAClE,QAAI,EAAE,mBAAmB,MAAM,OAAW,GAAE,aAAa,EAAE,mBAAmB;AAC9E,QAAI,EAAE,QAAQ,EAAG,GAAE,SAAS,EAAE,QAAQ;AACtC,WAAO;AAAA,EACT,CAAC;AAGD,SAAO,eAAe;AAAA,IACpB,WAAW,SAAS,YAAY,KAAK;AAAA,IACrC,YAAY;AAAA,IACZ;AAAA,EACF;AAGA,QAAM,eAAe,SAAS,OAAO,CAAC,MAAW,EAAE,WAAW,MAAM,UAAU;AAE9E,aAAW,QAAQ,cAAc;AAC/B,UAAM,QAAgB,KAAK,eAAe,KAAK;AAC/C,UAAM,WAAmB,KAAK,WAAW,KAAK,KAAK,aAAa,KAAK;AACrE,UAAM,SAA6B,KAAK,QAAQ;AAEhD,UAAM,OAAoB,EAAE,OAAO,UAAU,OAAO;AAEpD,QAAI,WAAW,KAAQ;AACrB,WAAK,aAAa,SACd,qDAAqD,KAAK,KAC1D,kCAAkC,KAAK;AAAA,IAC7C;AAEA,WAAO,SAAS,KAAK,IAAI;AAAA,EAC3B;AAGA,aAAW,QAAQ,OAAO,UAAU;AAClC,QAAI,CAAC,KAAK,OAAQ;AAElB,UAAM,OAAO,yBAAyB,KAAK,MAAM;AACjD,QAAI,KAAK,WAAW,EAAG;AAGvB,QAAI,oBAAgC,CAAC;AACrC,QAAIA,OAAM;AACR,0BAAoB,MAAM,wBAAwBA,OAAM,KAAK,KAAK;AAAA,IACpE;AAGA,UAAM,gBAAgB,KAAK;AAAA,MACzB,CAAC,QAAQ,CAAC,kBAAkB,KAAK,CAAC,YAAY,QAAQ,SAAS,KAAK,QAAQ,CAAC,MAAM,GAAG;AAAA,IACxF;AAEA,QAAI,cAAc,WAAW,EAAG;AAEhC,UAAM,UAAU,YAAY,KAAK,QAAQ;AAEzC,QAAI,cAAc,UAAU,GAAG;AAE7B,YAAM,UAAU,OAAO,KAAK,KAAK,IAAI,cAAc,KAAK,GAAG,CAAC;AAC5D,YAAM,MAAM,6BAA6B,OAAO,OAAO,KAAK,KAAK,KAAK,cAAc,KAAK,IAAI,CAAC;AAC9F,aAAO,eAAe,KAAK;AAAA,QACzB,OAAO,KAAK;AAAA,QACZ,SAAS;AAAA,QACT,QAAQ,sCAAsC,cAAc,KAAK,IAAI,CAAC,QAAQ,QAAQ,KAAK,QAAQ,CAAC;AAAA,QACpG;AAAA,QACA,kBAAkB;AAAA,MACpB,CAAC;AAAA,IACH,OAAO;AAEL,YAAM,MAAM,cAAc,CAAC;AAC3B,YAAM,UAAU,OAAO,KAAK,KAAK,IAAI,GAAG;AACxC,YAAM,MAAM,6BAA6B,OAAO,OAAO,KAAK,KAAK,KAAK,GAAG;AACzE,aAAO,eAAe,KAAK;AAAA,QACzB,OAAO,KAAK;AAAA,QACZ,SAAS,CAAC,GAAG;AAAA,QACb,QAAQ,2BAA2B,GAAG,KAAK,QAAQ,KAAK,QAAQ,CAAC;AAAA,QACjE;AAAA,QACA,kBAAkB;AAAA,MACpB,CAAC;AAAA,IACH;AAAA,EACF;AAGA,aAAW,QAAQ,OAAO,UAAU;AAClC,QAAI,KAAK,WAAW,KAAQ;AAC1B,YAAM,aAAa,KAAK,SACpB,oCAA+B,yBAAyB,KAAK,MAAM,EAAE,KAAK,IAAI,KAAK,gBAAgB,KACnG;AACJ,aAAO,gBAAgB;AAAA,QACrB,eAAe,KAAK,KAAK,KAAK,QAAQ,KAAK,QAAQ,CAAC,SAAS,UAAU;AAAA,MACzE;AAAA,IACF;AAAA,EACF;AAEA,MAAI,iBAAiB,QAAW;AAC9B,UAAM,QAAQ,eAAe,KAAK,iCAA4B;AAC9D,WAAO,gBAAgB,KAAK,iBAAiB,aAAa,QAAQ,CAAC,CAAC,aAAQ,KAAK,EAAE;AAAA,EACrF;AAEA,MAAI,OAAO,eAAe,WAAW,KAAK,OAAO,SAAS,WAAW,GAAG;AACtE,WAAO,gBAAgB,KAAK,mEAA8D;AAAA,EAC5F;AAEA,SAAO;AACT;AAeA,eAAsB,uBACpBA,OACA,SACA,cAAc,IACkB;AAChC,MAAI;AAEF,UAAM,WAAW,MAAMA,MAAK;AAAA,MAC1B;AAAA,IACF;AACA,QAAI,SAAS,KAAK,WAAW,EAAG,QAAO,CAAC;AAGxC,UAAM,UAAU,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA,KAKhC;AAED,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,OAAO,QAAQ,MAAM;AAC9B,iBAAW,IAAI,IAAI,SAAS,WAAW,IAAI,cAAc,CAAC;AAAA,IAC5D;AAEA,QAAI,CAAC,QAAS,QAAO,CAAC;AAGtB,UAAM,WAAW,cAAc,KAAK,KAAK;AACzC,UAAM,QAAQ,KAAK,IAAI,IAAI;AAE3B,QAAI;AACJ,QAAI;AACF,mBAAa,QACV;AAAA,QACC;AAAA;AAAA;AAAA;AAAA,MAIF,EACC,IAAI,KAAK;AAAA,IACd,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAGA,UAAM,cAAc,oBAAI,IAAmD;AAC3E,eAAW,OAAO,YAAY;AAC5B,UAAI,CAAC,YAAY,IAAI,IAAI,OAAO,GAAG;AACjC,oBAAY,IAAI,IAAI,SAAS;AAAA,UAC3B,QAAQ,IAAI;AAAA,UACZ,WAAW,IAAI;AAAA,QACjB,CAAC;AAAA,MACH;AAAA,IACF;AAGA,UAAM,cAAqC,CAAC;AAE5C,eAAW,CAAC,SAAS,QAAQ,KAAK,aAAa;AAC7C,YAAM,cAAc,WAAW,IAAI,OAAO;AAC1C,UAAI,gBAAgB,UAAa,SAAS,WAAW,EAAG;AAExD,YAAM,iBACF,cAAc,SAAS,UAAU,SAAS,SAAU;AAExD,UAAI,gBAAgB,IAAI;AACtB,oBAAY,KAAK;AAAA,UACf;AAAA,UACA,eAAe;AAAA,UACf,gBAAgB,SAAS;AAAA,UACzB,eAAe,KAAK,MAAM,aAAa;AAAA,UACvC,YAAY,IAAI,KAAK,SAAS,SAAS,EAAE,YAAY;AAAA,QACvD,CAAC;AAAA,MACH;AAAA,IACF;AAEA,WAAO,YAAY,KAAK,CAAC,GAAG,MAAM,EAAE,gBAAgB,EAAE,aAAa;AAAA,EACrE,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;;;AC1VA,SAAS,cAAc,KAAqB;AAE1C,MAAI,WAAW,IAAI;AAAA,IAAQ;AAAA,IAAqB,CAAC,UAC/C,MAAM,QAAQ,UAAU,GAAG;AAAA,EAC7B;AAEA,aAAW,SAAS,QAAQ,aAAa,CAAC,UAAU,IAAI,OAAO,MAAM,MAAM,CAAC;AAC5E,SAAO;AACT;AAGA,SAAS,eAAe,KAAa,YAA4B;AAC/D,QAAM,SAAS,IAAI,MAAM,GAAG,UAAU;AACtC,SAAO,OAAO,MAAM,IAAI,EAAE;AAC5B;AAGA,SAAS,UAAU,MAAsB;AACvC,SAAO,KACJ,QAAQ,cAAc,EAAE,EACxB,QAAQ,MAAM,EAAE,EAChB,YAAY,EACZ,KAAK;AACV;AAGA,SAAS,sBAAsB,KAK7B;AACA,QAAM,cAAc,GAAG;AACvB,QAAM,cAAwB,CAAC;AAC/B,QAAM,cAAwB,CAAC;AAC/B,QAAM,aAAuB,CAAC;AAC9B,QAAM,YAAsB,CAAC;AAG7B,QAAM,QAAQ;AACd,MAAI;AACJ,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,aAAY,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,aAAY,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,SAAS;AACf,UAAQ,IAAI,OAAO,KAAK,GAAG,OAAO,KAAM,YAAW,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAGvE,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,KAAM,WAAU,KAAK,UAAU,EAAE,CAAC,CAAC,CAAC;AAErE,SAAO,EAAE,aAAa,aAAa,YAAY,UAAU;AAC3D;AAGA,SAAS,YAAY,KAA+B;AAClD,QAAM,SAA2B,CAAC;AAElC,QAAM,cAAc,GAAG;AAGvB,QAAM,kBAAkB;AACxB,QAAM,gBAAgB,oBAAI,IAAY;AACtC,MAAI;AACJ,UAAQ,IAAI,gBAAgB,KAAK,GAAG,OAAO,KAAM,eAAc,IAAI,UAAU,EAAE,CAAC,CAAC,CAAC;AAGlF,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AACrC,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,aAAa,eAAe,KAAK,EAAE,KAAK;AAC9C,QAAI,CAAC,cAAc,IAAI,KAAK,GAAG;AAC7B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,YAAY;AAClB,UAAQ,IAAI,UAAU,KAAK,GAAG,OAAO,MAAM;AACzC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAIA,QAAM,WACJ;AACF,UAAQ,IAAI,SAAS,KAAK,GAAG,OAAO,MAAM;AACxC,UAAM,WAAW,EAAE,CAAC;AACpB,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,aAAa,eAAe,KAAK,EAAE,KAAK;AAC9C,UAAM,YAAY,SAAS,YAAY;AAEvC,UAAM,aAAa,iBAAiB,KAAK,SAAS;AAClD,UAAM,aAAa,cAAc,KAAK,SAAS;AAE/C,QAAI,cAAc,CAAC,YAAY;AAC7B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH,WAAW,cAAc,YAAY;AACnC,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ;AAAA,QACA,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,SAAS;AACf,UAAQ,IAAI,OAAO,KAAK,GAAG,OAAO,MAAM;AACtC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAGA,QAAM,cAAc;AACpB,UAAQ,IAAI,YAAY,KAAK,GAAG,OAAO,MAAM;AAC3C,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,YAAY;AAClB,UAAQ,IAAI,UAAU,KAAK,GAAG,OAAO,MAAM;AACzC,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,gBAAgB;AACtB,UAAQ,IAAI,cAAc,KAAK,GAAG,OAAO,MAAM;AAC7C,UAAM,UAAU,EAAE,CAAC;AACnB,UAAM,UAAU,EAAE,CAAC;AACnB,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS,mBAAmB,OAAO,SAAS,OAAO;AAAA,MACnD,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,iBAAiB;AACvB,UAAQ,IAAI,eAAe,KAAK,GAAG,OAAO,MAAM;AAC9C,UAAM,QAAQ,EAAE,CAAC;AACjB,UAAM,SAAS,EAAE,CAAC;AAClB,UAAM,SAAS,EAAE,CAAC;AAClB,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS,oBAAoB,MAAM,SAAS,MAAM,eAAe,KAAK;AAAA,MACtE,YAAY;AAAA,MACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACvC,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,QAAM,WAAW;AACjB,UAAQ,IAAI,SAAS,KAAK,GAAG,OAAO,MAAM;AACxC,UAAM,WAAW,EAAE,CAAC;AACpB,UAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AAC5B,UAAM,YAAY,SAAS,YAAY;AAEvC,QAAI,CAAC,kBAAkB,KAAK,SAAS,GAAG;AACtC,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY;AAAA,QACZ,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,QACvC,WAAW;AAAA,MACb,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,iBAAiB,aAAa,KAAK,GAAG,KAAK,2BAA2B,KAAK,GAAG;AACpF,QAAM,kBAAkB,mDAAmD,KAAK,GAAG;AACnF,MAAI,kBAAkB,iBAAiB;AACrC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY;AAAA,IACd,CAAC;AAAA,EACH;AAGA,QAAM,UAAU;AAChB,UAAQ,IAAI,QAAQ,KAAK,GAAG,OAAO,MAAM;AACvC,WAAO,KAAK;AAAA,MACV,UAAU;AAAA,MACV,MAAM;AAAA,MACN,SAAS;AAAA,MACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,IACzC,CAAC;AAAA,EACH;AAGA,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AAErC,UAAM,OAAO,EAAE,CAAC;AAChB,QAAI,CAAC,aAAa,KAAK,IAAI,GAAG;AAC5B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACzC,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,QAAQ;AACd,UAAQ,IAAI,MAAM,KAAK,GAAG,OAAO,MAAM;AACrC,UAAM,OAAO,EAAE,CAAC;AAChB,QAAI,CAAC,aAAa,KAAK,IAAI,GAAG;AAC5B,aAAO,KAAK;AAAA,QACV,UAAU;AAAA,QACV,MAAM;AAAA,QACN,SAAS;AAAA,QACT,YAAY,eAAe,KAAK,EAAE,KAAK;AAAA,MACzC,CAAC;AAAA,IACH;AAAA,EACF;AAEA,SAAO;AACT;AAGA,eAAe,aAAa,KAAaC,OAAY,cAA2D;AAC9G,QAAM,SAA2B,CAAC;AAClC,QAAM,EAAE,aAAa,aAAa,YAAY,UAAU,IAAI,sBAAsB,GAAG;AAGrF,QAAM,YAAY,CAAC,GAAG,oBAAI,IAAI,CAAC,GAAG,aAAa,GAAG,aAAa,GAAG,UAAU,CAAC,CAAC;AAG9E,QAAM,aAAa,oBAAI,IAAqD;AAC5E,MAAI,UAAU,SAAS,GAAG;AACxB,QAAI;AACF,YAAM,MAAM,MAAMA,MAAK;AAAA,QACrB;AAAA;AAAA;AAAA;AAAA;AAAA,QAKA,CAAC,SAAS;AAAA,MACZ;AACA,iBAAW,OAAO,IAAI,MAAM;AAC1B,mBAAW,IAAI,IAAI,WAAW;AAAA,UAC5B,UAAU,SAAS,IAAI,cAAc,KAAK,EAAE;AAAA,UAC5C,WAAW,SAAS,IAAI,cAAc,KAAK,EAAE;AAAA,QAC/C,CAAC;AAAA,MACH;AAAA,IACF,SAAS,GAAG;AAAA,IAEZ;AAAA,EACF;AAGA,aAAW,SAAS,cAAc;AAChC,QAAI,MAAM,SAAS,gCAAgC,MAAM,WAAW;AAClE,YAAM,QAAQ,WAAW,IAAI,MAAM,SAAS;AAC5C,UAAI,OAAO;AACT,cAAM,EAAE,SAAS,IAAI;AACrB,cAAM,WAAW,KAAK,MAAM,WAAW,GAAK;AAC5C,cAAM,gBAAgB;AACtB,cAAM,uBAAuB;AAE7B,YAAI,WAAW,KAAW;AACxB,gBAAM,WAAW;AACjB,gBAAM,UAAU,oBAAoB,MAAM,SAAS,2BAA2B,QAAQ,OAAO,WAAW,KAAK,QAAQ,CAAC,CAAC;AAAA,QACzH,WAAW,WAAW,KAAS;AAC7B,gBAAM,UAAU,oBAAoB,MAAM,SAAS,2BAA2B,QAAQ,OAAO,WAAW,KAAM,QAAQ,CAAC,CAAC;AAAA,QAC1H;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAGA,QAAM,kBAAkB,CAAC,GAAG,IAAI,IAAI,SAAS,CAAC;AAC9C,aAAW,SAAS,iBAAiB;AACnC,QAAI;AACF,YAAM,MAAM,MAAMA,MAAK;AAAA,QACrB;AAAA,QACA,CAAC,KAAK;AAAA,MACR;AACA,UAAI,IAAI,KAAK,WAAW,GAAG;AACzB,eAAO,KAAK;AAAA,UACV,UAAU;AAAA,UACV,MAAM;AAAA,UACN,SAAS,UAAU,KAAK;AAAA,UACxB,WAAW;AAAA,QACb,CAAC;AAAA,MACH;AAAA,IACF,SAAS,GAAG;AAAA,IAEZ;AAAA,EACF;AAEA,SAAO;AACT;AAEA,eAAsB,iBAAiB,KAAaA,OAA4C;AAC9F,QAAM,UAAU,IAAI,KAAK;AAEzB,MAAI,CAAC,SAAS;AACZ,WAAO;AAAA,MACL,MAAM;AAAA,MACN,QAAQ,CAAC;AAAA,MACT,SAAS,EAAE,QAAQ,GAAG,UAAU,GAAG,OAAO,EAAE;AAAA,MAC5C,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,IACpC;AAAA,EACF;AAGA,QAAM,SAAS,YAAY,OAAO;AAGlC,MAAIA,OAAM;AACR,UAAM,gBAAgB,MAAM,aAAa,SAASA,OAAM,MAAM;AAC9D,WAAO,KAAK,GAAG,aAAa;AAAA,EAC9B;AAEA,QAAM,SAAS,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,OAAO,EAAE;AAC5D,QAAM,WAAW,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,SAAS,EAAE;AAChE,QAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,EAAE,aAAa,MAAM,EAAE;AAE1D,SAAO;AAAA,IACL,MAAM,WAAW;AAAA,IACjB;AAAA,IACA,SAAS,EAAE,QAAQ,UAAU,MAAM;AAAA,IACnC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;ACrYO,SAAS,YAAY,OAAuB;AACjD,MAAI,QAAQ,KAAM,QAAO;AACzB,MAAI,QAAQ,OAAO,KAAM,QAAO,GAAG,KAAK,MAAM,QAAQ,IAAI,CAAC;AAC3D,MAAI,QAAQ,OAAO,OAAO,KAAM,QAAO,IAAI,SAAS,OAAO,OAAO,QAAQ,CAAC,CAAC;AAC5E,MAAI,QAAQ,QAAQ,EAAG,QAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAC/D,SAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAC1C;AAEA,eAAsB,iBAAiBC,OAAwC;AAC7E,QAAM,CAAC,aAAa,cAAc,IAAI,MAAM,QAAQ,IAAI;AAAA,IACtDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAiBV;AAAA,IACDA,MAAK,MAAM,0CAA0C;AAAA,EACvD,CAAC;AAED,QAAM,aAAa,eAAe,KAAK,CAAC,GAAG,cACvC,IAAI,KAAK,eAAe,KAAK,CAAC,EAAE,WAAW,EAAE,YAAY,IACzD;AAEJ,QAAM,eAAe,YAAY,KAAK,OAAO,CAAC,QAAa;AACzD,UAAM,MAAc,IAAI,YAAY;AAEpC,QAAI,IAAI,SAAS,SAAS,EAAG,QAAO;AAGpC,UAAM,WAAW,IAAI,QAAQ,GAAG;AAChC,UAAM,SAAS,IAAI,YAAY,GAAG;AAClC,QAAI,aAAa,MAAM,WAAW,IAAI;AACpC,YAAM,OAAO,IAAI,MAAM,WAAW,GAAG,MAAM;AAC3C,UAAI,KAAK,SAAS,GAAG,EAAG,QAAO;AAAA,IACjC;AACA,WAAO;AAAA,EACT,CAAC;AAED,QAAM,UAAyB,aAAa,IAAI,CAAC,QAAa;AAC5D,UAAM,YAAY,SAAS,IAAI,kBAAkB,EAAE,KAAK;AACxD,UAAM,QAAQ,IAAI;AAClB,UAAM,QAAQ,IAAI;AAClB,WAAO;AAAA,MACL,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA;AAAA,MACA,WAAW,YAAY,SAAS;AAAA,MAChC,gBAAgB;AAAA,MAChB,OAAO,SAAS,IAAI,UAAU,EAAE,KAAK;AAAA,MACrC,UAAU;AAAA,MACV,YAAY,SAAS,KAAK,OAAO,KAAK,kFAAkF,MAAM,QAAQ,MAAM,IAAI,CAAC;AAAA,IACnJ;AAAA,EACF,CAAC;AAED,QAAM,mBAAmB,QAAQ,OAAO,CAAC,KAAK,QAAQ,MAAM,IAAI,gBAAgB,CAAC;AAEjF,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,aAAa,YAAY,gBAAgB;AAAA,IACzC,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AC5EA,SAAS,cAAc,OAAe,cAA8B;AAClE,MAAI,gBAAgB,IAAI;AACtB,WAAO,iBAAiB,KAAK,KAAK,YAAY,qCAAqC,KAAK;AAAA,EAC1F,WAAW,gBAAgB,IAAI;AAC7B,WAAO,qBAAqB,KAAK,KAAK,YAAY,yCAAyC,KAAK;AAAA,EAClG,OAAO;AACL,WAAO,kBAAkB,KAAK,KAAK,YAAY;AAAA,EACjD;AACF;AAEA,eAAsB,eAAeC,OAAkC;AACrE,QAAM,SAAS,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAY/B;AAED,QAAM,SAAuB,CAAC;AAE9B,aAAW,OAAO,OAAO,MAAM;AAC7B,UAAM,OAAO,SAAS,IAAI,YAAY,EAAE,KAAK;AAC7C,UAAM,OAAO,SAAS,IAAI,YAAY,EAAE,KAAK;AAC7C,UAAM,QAAQ,OAAO;AACrB,QAAI,UAAU,EAAG;AAEjB,UAAM,eAAe,KAAK,MAAO,OAAO,QAAS,GAAI,IAAI;AACzD,QAAI,eAAe,GAAI;AAEvB,UAAM,QAAQ,IAAI;AAClB,WAAO,KAAK;AAAA,MACV,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA,UAAU;AAAA,MACV,UAAU;AAAA,MACV;AAAA,MACA,gBAAgB,IAAI,kBAAkB,IAAI,KAAK,IAAI,eAAe,EAAE,YAAY,IAAI;AAAA,MACpF,YAAY,IAAI,cAAc,IAAI,KAAK,IAAI,WAAW,EAAE,YAAY,IAAI;AAAA,MACxE,YAAY,cAAc,OAAO,YAAY;AAAA,IAC/C,CAAC;AAAA,EACH;AAGA,SAAO,KAAK,CAAC,GAAG,MAAM,EAAE,eAAe,EAAE,YAAY;AAErD,SAAO;AAAA,IACL;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AChDA,SAAS,eACP,gBACA,YACA,aACiC;AACjC,MAAI,mBAAmB,KAAM,QAAO;AAEpC,QAAM,aAAa,KAAK,IAAI,IAAI,eAAe,QAAQ,MAAM,MAAO,KAAK,KAAK;AAE9E,MAAI,YAAY,KAAK,aAAa,IAAQ,QAAO;AACjD,MAAI,YAAY,EAAG,QAAO;AAC1B,SAAO;AACT;AAEA,SAASC,eAAc,QAA8C,OAA8B;AACjG,UAAQ,QAAQ;AAAA,IACd,KAAK;AACH,aAAO,SAAS,KAAK;AAAA,IACvB,KAAK;AACH,aAAO,SAAS,KAAK,wEAAwE,KAAK;AAAA,IACpG,KAAK;AACH,aAAO,SAAS,KAAK;AAAA,IACvB,KAAK;AACH,aAAO;AAAA,EACX;AACF;AAEA,eAAsB,oBAAoBC,OAAuC;AAC/E,QAAM,CAAC,aAAa,cAAc,IAAI,MAAM,QAAQ,IAAI;AAAA,IACtDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KASV;AAAA,IACDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA,KAIV;AAAA,EACH,CAAC;AAED,QAAM,SAAkC,YAAY,KAAK,IAAI,CAAC,QAAa;AACzE,UAAM,qBAAqB,IAAI,kBAAkB,IAAI,KAAK,IAAI,eAAe,IAAI;AACjF,UAAM,aAAa,SAAS,IAAI,YAAY,EAAE,KAAK;AACnD,UAAM,aAAa,SAAS,IAAI,YAAY,EAAE,KAAK;AACnD,UAAM,cAAc,SAAS,IAAI,kBAAkB,EAAE,KAAK;AAC1D,UAAM,eAAe,SAAS,IAAI,mBAAmB,EAAE,KAAK;AAC5D,UAAM,SAAS,eAAe,oBAAoB,YAAY,WAAW;AACzE,UAAM,QAAQ,IAAI;AAElB,WAAO;AAAA,MACL,QAAQ,IAAI;AAAA,MACZ;AAAA,MACA,gBAAgB,qBAAqB,mBAAmB,YAAY,IAAI;AAAA,MACxE,iBAAiB,IAAI,mBAAmB,IAAI,KAAK,IAAI,gBAAgB,EAAE,YAAY,IAAI;AAAA,MACvF;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,YAAYD,eAAc,QAAQ,KAAK;AAAA,IACzC;AAAA,EACF,CAAC;AAED,QAAM,cAAc,oBAAI,IAAoB;AAC5C,aAAW,OAAO,eAAe,MAAM;AACrC,gBAAY,IAAI,IAAI,MAAM,IAAI,OAAO;AAAA,EACvC;AAEA,SAAO;AAAA,IACL;AAAA,IACA,UAAU;AAAA,MACR,mBAAmB,YAAY,IAAI,YAAY,MAAM;AAAA,MACrD,iBAAiB,GAAG,YAAY,IAAI,8BAA8B,KAAK,GAAG;AAAA,MAC1E,sBAAsB,SAAS,YAAY,IAAI,wBAAwB,KAAK,KAAK,EAAE;AAAA,MACnF,mBAAmB,GAAG,YAAY,IAAI,oBAAoB,KAAK,IAAI;AAAA,IACrE;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;ACrFO,SAAS,mBAAmB,MAAsB;AACvD,QAAM,IAAI,KAAK,MAAM,OAAO,IAAI;AAChC,QAAM,IAAI,KAAK,MAAO,OAAO,OAAQ,EAAE;AACvC,QAAM,IAAI,OAAO;AACjB,SAAO;AAAA,IACL,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,IACzB,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,IACzB,OAAO,CAAC,EAAE,SAAS,GAAG,GAAG;AAAA,EAC3B,EAAE,KAAK,GAAG;AACZ;AAEA,eAAsB,cAAcE,OAAiC;AACnE,QAAM,CAAC,aAAa,UAAU,IAAI,MAAM,QAAQ,IAAI;AAAA,IAClDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAmBV;AAAA,IACDA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAcV;AAAA,EACH,CAAC;AAID,QAAM,OAAO,oBAAI,IAAY;AAC7B,QAAM,eAA2B,CAAC;AAClC,aAAW,OAAO,YAAY,MAAM;AAClC,UAAM,MAAM,GAAG,IAAI,WAAW,IAAI,IAAI,YAAY;AAClD,QAAI,CAAC,KAAK,IAAI,GAAG,GAAG;AAClB,WAAK,IAAI,GAAG;AACZ,mBAAa,KAAK;AAAA,QAChB,YAAY,SAAS,IAAI,aAAa,EAAE;AAAA,QACxC,cAAc,IAAI;AAAA,QAClB,iBAAiB,mBAAmB,SAAS,IAAI,cAAc,EAAE,KAAK,CAAC;AAAA,QACvE,aAAa,SAAS,IAAI,cAAc,EAAE;AAAA,QAC1C,eAAe,IAAI;AAAA,QACnB,kBAAkB,mBAAmB,SAAS,IAAI,eAAe,EAAE,KAAK,CAAC;AAAA,QACzE,OAAO,IAAI,cAAc;AAAA,QACzB,UAAU,IAAI;AAAA,MAChB,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,qBAAqB,WAAW,KAAK,IAAI,CAAC,SAAc;AAAA,IAC5D,KAAK,SAAS,IAAI,KAAK,EAAE;AAAA,IACzB,UAAU,mBAAmB,SAAS,IAAI,eAAe,EAAE,KAAK,CAAC;AAAA,IACjE,OAAO,IAAI;AAAA,IACX,OAAO,IAAI;AAAA,IACX,eAAe,IAAI,mBAAmB;AAAA,EACxC,EAAE;AAEF,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AChFA,SAAS,eAAe,OAAe,MAAkC;AACvE,QAAM,IAAI,WAAW,KAAK;AAC1B,MAAI,CAAC,KAAM,QAAO;AAClB,UAAQ,KAAK,YAAY,GAAG;AAAA,IAC1B,KAAK;AAAO,aAAO;AAAA,IACnB,KAAK;AAAO,aAAO,IAAI;AAAA,IACvB,KAAK;AAAO,aAAO,IAAI,IAAI;AAAA;AAAA,IAC3B,KAAK;AAAO,aAAO,IAAI,OAAO;AAAA,IAC9B,KAAK;AAAO,aAAO,IAAI,OAAO,OAAO;AAAA,IACrC;AAAY,aAAO;AAAA,EACrB;AACF;AAEA,SAAS,YAAY,OAAe,MAAkC;AACpE,SAAO,eAAe,OAAO,IAAI,KAAK,OAAO;AAC/C;AAGA,SAAS,iBAAiB,UAAqC,MAAuB;AACpF,MAAI,CAAC,SAAU,QAAO;AACtB,QAAM,QAAQ,eAAe,UAAU,QAAQ,EAAE;AACjD,MAAI,SAAS,KAAK,MAAM,KAAK,EAAG,QAAO;AACvC,MAAI,SAAS,QAAQ,EAAG,QAAO,IAAI,QAAQ,QAAQ,GAAG,QAAQ,CAAC,CAAC;AAChE,MAAI,SAAS,QAAQ,EAAG,QAAO,GAAG,KAAK,MAAM,QAAQ,QAAQ,CAAC,CAAC;AAC/D,MAAI,SAAS,KAAM,QAAO,GAAG,KAAK,MAAM,QAAQ,IAAI,CAAC;AACrD,SAAO,GAAG,KAAK;AACjB;AAEA,eAAsB,gBAAgBC,OAAmC;AACvE,QAAM,SAAS,MAAMA,MAAK,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAW/B;AAED,QAAM,WAA0E,CAAC;AACjF,aAAW,OAAO,OAAO,MAAM;AAC7B,aAAS,IAAI,IAAI,IAAI,EAAE,SAAS,IAAI,SAAS,MAAM,IAAI,QAAQ,OAAU;AAAA,EAC3E;AAEA,QAAM,kBAA0C,CAAC;AAEjD,QAAM,MAAM,CAAC,SAAiB,SAAS,IAAI,GAAG,WAAW;AACzD,QAAM,UAAU,CAAC,SAAiB,SAAS,IAAI,GAAG;AAGlD,QAAM,uBAAuB,IAAI,gBAAgB;AACjD,MAAI,yBAAyB,MAAM;AACjC,UAAM,KAAK,YAAY,sBAAsB,QAAQ,gBAAgB,CAAC;AACtE,QAAI,KAAK,KAAK;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,EAAE,CAAC;AAAA,QAC/B,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,iBAAiB,IAAI,UAAU;AACrC,MAAI,mBAAmB,MAAM;AAC3B,UAAM,KAAK,YAAY,gBAAgB,QAAQ,UAAU,CAAC;AAC1D,QAAI,MAAM,GAAG;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,IAAI,KAAK,GAAG,QAAQ,CAAC,CAAC;AAAA,QAClD,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,aAAa,IAAI,8BAA8B;AACrD,MAAI,eAAe,MAAM;AACvB,UAAM,IAAI,WAAW,UAAU;AAC/B,QAAI,IAAI,KAAK;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,aAAa,IAAI,kBAAkB;AACzC,MAAI,eAAe,MAAM;AACvB,UAAM,IAAI,WAAW,UAAU;AAC/B,QAAI,IAAI,GAAK;AACX,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,cAAc,IAAI,gCAAgC;AACxD,MAAI,gBAAgB,MAAM;AACxB,UAAM,IAAI,WAAW,WAAW;AAChC,QAAI,KAAK,KAAK;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc;AAAA,QACd,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAGA,QAAM,cAAc,IAAI,4BAA4B;AACpD,MAAI,gBAAgB,QAAQ,SAAS,aAAa,EAAE,MAAM,IAAI;AAC5D,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAGA,QAAM,cAAc,IAAI,qCAAqC;AAC7D,MAAI,gBAAgB,QAAQ,SAAS,aAAa,EAAE,MAAM,GAAG;AAC3D,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAGA,QAAM,aAAa,IAAI,0BAA0B;AACjD,MAAI,eAAe,QAAQ,SAAS,YAAY,EAAE,MAAM,GAAG;AACzD,oBAAgB,KAAK;AAAA,MACnB,SAAS;AAAA,MACT,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,QAAQ;AAAA,MACR,UAAU;AAAA,MACV,MAAM;AAAA,IACR,CAAC;AAAA,EACH;AAKA,QAAM,aAAa,IAAI,sBAAsB;AAC7C,MAAI,eAAe,MAAM;AACvB,UAAM,KAAK,YAAY,YAAY,QAAQ,sBAAsB,CAAC;AAClE,QAAI,MAAM,IAAI;AACZ,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,GAAG,KAAK,MAAM,IAAI,KAAK,GAAG,QAAQ,CAAC,CAAC;AAAA,QAClD,kBAAkB;AAAA,QAClB,QAAQ;AAAA,QACR,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,iBAAiB,IAAI,iBAAiB;AAG5C,MAAI,mBAAmB,MAAM;AAC3B,UAAM,UAAU,SAAS,gBAAgB,EAAE;AAC3C,QAAI,UAAU,KAAK;AACjB,sBAAgB,KAAK;AAAA,QACnB,SAAS;AAAA,QACT,cAAc,OAAO,OAAO;AAAA,QAC5B,kBAAkB;AAAA,QAClB,QAAQ,mBAAmB,OAAO;AAAA,QAClC,UAAU;AAAA,QACV,MAAM;AAAA,MACR,CAAC;AAAA,IACH;AAAA,EACF;AAEA,QAAM,aAAa;AAAA,IACjB,gBAAgB,mBAAmB,OAAO,SAAS,gBAAgB,EAAE,IAAI;AAAA,IACzE,eAAe,iBAAiB,sBAAsB,QAAQ,gBAAgB,CAAC;AAAA,IAC/E,SAAS,iBAAiB,gBAAgB,QAAQ,UAAU,CAAC;AAAA,IAC7D,oBAAoB,iBAAiB,IAAI,sBAAsB,GAAG,QAAQ,sBAAsB,CAAC;AAAA,IACjG,oBAAoB,iBAAiB,YAAY,QAAQ,sBAAsB,CAAC;AAAA,IAChF,YAAY,IAAI,aAAa,KAAK;AAAA,IAClC,4BAA4B,cAAc;AAAA,IAC1C,gBAAgB,cAAc;AAAA,IAC9B,6BAA6B,eAAe;AAAA,EAC9C;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,EACpC;AACF;;;AjBvOA,OAAOC,eAAc;AACrB,OAAOC,WAAU;AACjB,OAAOC,SAAQ;AACf,OAAOC,OAAM,oBAAoB;AAEjC,IAAM,MAAM,KAAK,MAAM,aAAa,IAAI,IAAI,mBAAmB,YAAY,GAAG,GAAG,OAAO,CAAC;AAEzF,IAAM,aAAa,QAAQ,KAAK,CAAC,KAAK,QAAQ,IAAI;AAClD,IAAI,CAAC,YAAY;AACf,UAAQ,MAAM,wCAAwC;AACtD,UAAQ,MAAM,4CAA4C;AAC1D,UAAQ,KAAK,CAAC;AAChB;AAEA,IAAM,OAAO,IAAIC,MAAK,EAAE,kBAAkB,YAAY,yBAAyB,IAAM,CAAC;AACtF,IAAM,qBAAqB,SAAS,QAAQ,IAAI,gCAAgC,KAAK,EAAE;AACvF,IAAM,UAAU,QAAQ,IAAI,oBAAoBH,MAAK,KAAKC,IAAG,QAAQ,GAAG,UAAU;AAClFC,IAAG,UAAU,SAAS,EAAE,WAAW,KAAK,CAAC;AAGzC,IAAI,WAAqC;AACzC,IAAI,WAAqC;AACzC,IAAI;AACF,QAAM,aAAaF,MAAK,KAAK,SAAS,WAAW;AACjD,MAAIE,IAAG,WAAW,UAAU,EAAG,YAAW,IAAIH,UAAS,YAAY,EAAE,UAAU,KAAK,CAAC;AACvF,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AACvE,IAAI;AACF,QAAM,aAAaC,MAAK,KAAK,SAAS,WAAW;AACjD,MAAIE,IAAG,WAAW,UAAU,EAAG,YAAW,IAAIH,UAAS,YAAY,EAAE,UAAU,KAAK,CAAC;AACvF,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AAEvE,IAAI,eAAyC;AAC7C,IAAI;AACF,QAAM,iBAAiBC,MAAK,KAAK,SAAS,gBAAgB;AAC1D,MAAIE,IAAG,WAAW,cAAc,EAAG,gBAAe,IAAIH,UAAS,gBAAgB,EAAE,UAAU,KAAK,CAAC;AACnG,SAAS,KAAK;AAAE,UAAQ,MAAM,gBAAiB,IAAc,OAAO;AAAG;AAEvE,IAAM,SAAS,IAAI,UAAU,EAAE,MAAM,WAAW,SAAS,IAAI,QAAQ,CAAC;AAEtE,OAAO,KAAK,oBAAoB,8DAA8D,CAAC,GAAG,YAAY;AAC5G,MAAI;AACF,UAAM,OAAO,MAAM,YAAY,IAAI;AACnC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,2DAA2D,CAAC,GAAG,YAAY;AACvG,MAAI;AACF,UAAM,OAAO,MAAM,iBAAiB,MAAM,kBAAkB;AAC5D,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,6CAA6C,CAAC,GAAG,YAAY;AACzF,MAAI;AACF,UAAM,OAAO,MAAM,UAAU,IAAI;AACjC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,4CAA4C,EAAE,OAAO,EAAE,OAAO,EAAE,SAAS,6CAA6C,EAAE,GAAG,OAAO,EAAE,MAAM,MAAM;AAClL,MAAI;AACF,UAAM,OAAO,MAAM,qBAAqB,MAAM,KAAK;AACnD,QAAI,CAAC,KAAM,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,kBAAkB,CAAC,GAAG,SAAS,KAAK;AACxF,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,oBAAoB,+DAA+D,CAAC,GAAG,YAAY;AAC7G,MAAI;AACF,UAAM,OAAO,MAAM,YAAY,IAAI;AACnC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,MAAM,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC5E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,0BAA0B,6BAA6B,CAAC,GAAG,YAAY;AACjF,MAAI;AACF,QAAI,CAAC,SAAU,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,+DAA+D,CAAC,EAAE;AAC1H,UAAM,UAAU,SAAS,QAAQ,+DAA+D,EAAE,IAAI;AACtG,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,SAAS,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC/E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,eAAe,uDAAuD,EAAE,KAAK,EAAE,OAAO,EAAE,SAAS,2CAA2C,EAAE,GAAG,OAAO,EAAE,IAAI,MAAM;AAC9K,MAAI;AACF,QAAI,CAAC,UAAU,GAAG,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,oKAAoK,CAAC,GAAG,SAAS,KAAK;AACpP,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,QAAQ,KAAK,IAAI;AACvB,YAAM,SAAS,MAAM,OAAO,MAAM,GAAG;AACrC,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,IAAI,MAAM,UAAU,KAAK,IAAI,IAAI,OAAO,UAAU,OAAO,UAAU,MAAM,OAAO,QAAQ,CAAC,EAAE,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IACtK,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,qBAAqB,CAAC,GAAG,YAAY;AACjE,MAAI;AACF,QAAI,CAAC,SAAU,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,sDAAsD,CAAC,EAAE;AACjH,UAAM,UAAU,SAAS,QAAQ,8DAA8D,EAAE,IAAI;AACrG,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,SAAS,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC/E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAID,OAAO,KAAK,mBAAmB,gFAAgF,EAAE,OAAO,EAAE,OAAO,EAAE,SAAS,yBAAyB,EAAE,GAAG,OAAO,EAAE,MAAM,MAAM;AAC7L,MAAI;AACF,QAAI,CAAC,gBAAgB,KAAK,KAAK,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yCAAyC,CAAC,GAAG,SAAS,KAAK;AACtI,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,OAAO,MAAM,+BAA+B;AAClD,YAAM,OAAO,MAAM,OAAO;AAC1B,UAAI;AACF,cAAM,IAAI,MAAM,OAAO,MAAM,2CAA2C,KAAK,EAAE;AAC/E,cAAM,OAAO,MAAM,UAAU;AAC7B,cAAM,OAAO,MAAM,yBAAyB;AAC5C,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,KAAK,CAAC,EAAE,YAAY,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,MAC/F,SAAS,KAAU;AACjB,cAAM,OAAO,MAAM,UAAU,EAAE,MAAM,MAAM;AAAA,QAAC,CAAC;AAC7C,cAAM,OAAO,MAAM,yBAAyB,EAAE,MAAM,MAAM;AAAA,QAAC,CAAC;AAC5D,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,MACrF;AAAA,IACF,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,qBAAqB,kEAAkE,EAAE,UAAU,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,6DAA6D,EAAE,GAAG,OAAO,EAAE,SAAS,MAAM;AACtO,MAAI;AACF,UAAM,SAAS,MAAM,iBAAiB,MAAM,kBAAkB;AAC9D,QAAI,QAAQ,OAAO;AACnB,QAAI,SAAU,SAAQ,MAAM,OAAO,CAAC,MAAM,EAAE,KAAK,WAAW,QAAQ,CAAC;AACrE,QAAI,MAAM,WAAW,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,0BAA0B,WAAW,kBAAkB,QAAQ,MAAM,IAAI,CAAC,EAAE;AAC7I,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,MAAM,EAAE,KAAK;AAAA,EAAK,EAAE,GAAG,EAAE,EAAE,KAAK,MAAM;AACxE,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,SAAS,CAAC,EAAE;AAAA,EACvD,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,gDAAgD;AAAA,EAClF,OAAO,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE,EAAE,SAAS,oCAAoC;AAAA,EACtF,SAAS,EAAE,KAAK,CAAC,cAAc,aAAa,OAAO,CAAC,EAAE,SAAS,EAAE,QAAQ,YAAY,EAAE,SAAS,YAAY;AAC9G,GAAG,OAAO,EAAE,OAAO,QAAQ,MAAM;AAC/B,MAAI;AACF,UAAM,MAAM,MAAM,eAAe,IAAI;AACrC,QAAI,IAAI,WAAW,EAAG,QAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yEAAyE,CAAC,EAAE;AAC3I,UAAM,SAAS,CAAC,GAAG,GAAG,EAAE,KAAK,CAAC,GAAG,MAAO,EAAU,OAAO,IAAK,EAAU,OAAO,CAAC;AAChF,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,OAAO,MAAM,GAAG,KAAK,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9F,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,uBAAuB,sDAAsD,CAAC,GAAG,YAAY;AACvG,MAAI;AACF,UAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,QAAI;AACF,YAAM,IAAI,MAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAQ5B;AACD,YAAM,SAAS,EAAE,KAAK,IAAI,CAAC,SAAc;AAAA,QACvC,QAAQ,IAAI;AAAA,QACZ,MAAM,IAAI;AAAA,QACV,WAAW,SAAS,IAAI,UAAU;AAAA,QAClC,WAAW,SAAS,IAAI,UAAU;AAAA,QAClC,WAAW,SAAS,IAAI,UAAU;AAAA,MACpC,EAAE;AACF,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,kBAAkB,6BAA6B,EAAE,QAAQ,EAAE,KAAK,CAAC,QAAQ,IAAI,CAAC,EAAE,SAAS,EAAE,QAAQ,MAAM,EAAE,SAAS,2BAA2B,EAAE,GAAG,OAAO,EAAE,OAAO,MAAM;AACpL,MAAI;AACF,UAAM,CAAC,UAAU,OAAO,IAAI,MAAM,QAAQ,IAAI;AAAA,MAC5C,YAAY,IAAI;AAAA,MAChB,iBAAiB,MAAM,kBAAkB;AAAA,IAC3C,CAAC;AACD,QAAI,WAAW,MAAM;AACnB,YAAM,QAAkB,CAAC;AACzB,YAAM,KAAK,yBAAyB;AACpC,YAAM,KAAK;AAAA,cAAgB,oBAAI,KAAK,GAAE,YAAY,CAAC;AAAA,CAAI;AACvD,YAAM,KAAK;AAAA,CAAe;AAC1B,YAAM,KAAK,qBAAqB,SAAS,OAAO,EAAE;AAClD,YAAM,KAAK,wBAAwB,SAAS,MAAM,EAAE;AACpD,YAAM,KAAK,sBAAsB,SAAS,YAAY,MAAM,aAAa,SAAS,YAAY,IAAI,WAAW,SAAS,YAAY,GAAG,MAAM;AAC3I,YAAM,KAAK;AAAA,mBAAsB,QAAQ,KAAK,gBAAgB,QAAQ,KAAK;AAAA,CAAK;AAChF,YAAM,KAAK,uCAAuC;AAClD,YAAM,KAAK,uCAAuC;AAClD,iBAAW,CAAC,KAAK,CAAC,KAAK,OAAO,QAAQ,QAAQ,SAAS,GAAG;AACxD,cAAM,KAAK,KAAK,GAAG,MAAM,EAAE,KAAK,MAAM,EAAE,KAAK,UAAU,EAAE,KAAK,IAAI;AAAA,MACpE;AACA,UAAI,QAAQ,OAAO,SAAS,GAAG;AAC7B,cAAM,KAAK;AAAA,cAAiB,QAAQ,OAAO,MAAM;AAAA,CAAK;AACtD,mBAAW,SAAS,QAAQ,QAAQ;AAClC,gBAAM,OAAO,MAAM,aAAa,aAAa,cAAO,MAAM,aAAa,YAAY,cAAO;AAC1F,gBAAM,KAAK,KAAK,IAAI,KAAK,MAAM,QAAQ,KAAK,MAAM,KAAK,EAAE;AAAA,QAC3D;AAAA,MACF;AACA,UAAI,QAAQ,WAAW,SAAS,GAAG;AACjC,cAAM,KAAK;AAAA;AAAA,CAAwB;AACnC,mBAAW,OAAO,QAAQ,YAAY;AACpC,gBAAM,KAAK;AAAA,EAAc,IAAI,GAAG;AAAA;AAAA,CAAY;AAAA,QAC9C;AAAA,MACF;AACA,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,MAAM,KAAK,IAAI,EAAE,CAAC,EAAE;AAAA,IAC/D;AACA,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,UAAU,SAAS,aAAY,oBAAI,KAAK,GAAE,YAAY,EAAE,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EACnI,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,gBAAgB,mDAAmD,CAAC,GAAG,YAAY;AAC7F,MAAI;AACF,UAAM,eAAeC,MAAK,KAAK,SAAS,iBAAiB;AACzD,UAAM,OAAO,aAAa,YAAY;AACtC,UAAM,UAAU,MAAM,iBAAiB,MAAM,kBAAkB;AAC/D,QAAI,CAAC,MAAM;AACT,mBAAa,cAAc,OAAO;AAClC,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,SAAS,iEAAiE,OAAO,QAAQ,OAAO,OAAO,QAAQ,OAAO,QAAQ,QAAQ,OAAO,OAAO,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC/N;AACA,UAAM,OAAO,cAAc,KAAK,QAAQ,OAAO;AAC/C,iBAAa,cAAc,OAAO;AAClC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,EAAE,GAAG,MAAM,mBAAmB,KAAK,UAAU,GAAG,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EACtH,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,KAAK,EAAE,OAAO,EAAE,SAAS,kCAAkC;AAAA,EAC7D;AAAA,EACA,OAAO,EAAE,IAAI,MAAM;AACjB,QAAI;AACF,YAAM,SAAS,MAAM,iBAAiB,KAAK,IAAI;AAC/C,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,KAAK,EAAE,OAAO,EAAE,SAAS,yBAAyB;AAAA,EACpD;AAAA,EACA,OAAO,EAAE,IAAI,MAAM;AACjB,QAAI;AACF,UAAI,CAAC,gBAAgB,KAAK,GAAG,GAAG;AAC9B,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yCAAyC,CAAC,GAAG,SAAS,KAAK;AAAA,MACtG;AACA,YAAM,SAAS,MAAM,KAAK,QAAQ;AAClC,UAAI;AACF,cAAM,OAAO,MAAM,+BAA+B;AAClD,cAAM,OAAO,MAAM,OAAO;AAC1B,YAAI;AACF,gBAAM,IAAI,MAAM,OAAO,MAAM,2CAA2C,GAAG,EAAE;AAC7E,gBAAM,OAAO,MAAM,UAAU;AAC7B,gBAAM,OAAO,MAAM,yBAAyB;AAE5C,gBAAM,OAAO,EAAE,KAAK,CAAC,EAAE,YAAY;AACnC,gBAAM,WAAW,MAAM,mBAAmB,MAAM,IAAI;AAEpD,iBAAO;AAAA,YACL,SAAS,CAAC;AAAA,cACR,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,MAAM,SAAS,GAAG,MAAM,CAAC;AAAA,YAClD,CAAC;AAAA,UACH;AAAA,QACF,SAAS,KAAU;AACjB,gBAAM,OAAO,MAAM,UAAU,EAAE,MAAM,MAAM;AAAA,UAAC,CAAC;AAC7C,gBAAM,OAAO,MAAM,yBAAyB,EAAE,MAAM,MAAM;AAAA,UAAC,CAAC;AAC5D,iBAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,QACrF;AAAA,MACF,UAAE;AACA,eAAO,QAAQ;AAAA,MACjB;AAAA,IACF,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,aAAa,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,kCAAkC;AAAA,EAChF;AAAA,EACA,OAAO,EAAE,YAAY,MAAM;AACzB,QAAI;AACF,YAAM,cAAc,MAAM,uBAAuB,MAAM,cAAc,eAAe,EAAE;AACtF,UAAI,YAAY,WAAW,GAAG;AAC5B,eAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,yDAAyD,CAAC,EAAE;AAAA,MACvG;AACA,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,aAAa,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IACnF,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO;AAAA,EACL;AAAA,EACA;AAAA,EACA;AAAA,IACE,WAAW,EAAE,OAAO,EAAE,SAAS,sDAAsD;AAAA,IACrF,eAAe,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS,uCAAuC;AAAA,EACxF;AAAA,EACA,OAAO,EAAE,WAAW,cAAc,MAAM;AACtC,QAAI;AAGF,YAAM,SAAS,MAAM,iBAAiB,YAAY,WAAW,EAAE,eAAe,iBAAiB,MAAM,CAAC;AACtG,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,IAC9E,SAAS,KAAU;AACjB,aAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,IACrF;AAAA,EACF;AACF;AAEA,OAAO,KAAK,0BAA0B,6DAA6D,CAAC,GAAG,YAAY;AACjH,MAAI;AACF,UAAM,SAAS,MAAM,iBAAiB,IAAI;AAC1C,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,iBAAiB,2DAA2D,CAAC,GAAG,YAAY;AACtG,MAAI;AACF,UAAM,SAAS,MAAM,eAAe,IAAI;AACxC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,sBAAsB,2EAAsE,CAAC,GAAG,YAAY;AACtH,MAAI;AACF,UAAM,SAAS,MAAM,oBAAoB,IAAI;AAC7C,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,iBAAiB,yEAAyE,CAAC,GAAG,YAAY;AACpH,MAAI;AACF,UAAM,SAAS,MAAM,cAAc,IAAI;AACvC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,OAAO,KAAK,wBAAwB,0EAA0E,CAAC,GAAG,YAAY;AAC5H,MAAI;AACF,UAAM,SAAS,MAAM,gBAAgB,IAAI;AACzC,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,KAAK,UAAU,QAAQ,MAAM,CAAC,EAAE,CAAC,EAAE;AAAA,EAC9E,SAAS,KAAU;AACjB,WAAO,EAAE,SAAS,CAAC,EAAE,MAAM,QAAQ,MAAM,UAAU,IAAI,OAAO,GAAG,CAAC,GAAG,SAAS,KAAK;AAAA,EACrF;AACF,CAAC;AAED,IAAM,YAAY,IAAI,qBAAqB;AAC3C,MAAM,OAAO,QAAQ,SAAS;","names":["Pool","pool","pool","pool","pool","pool","pool","longQueryThreshold","dataDir","pool","fs","path","pool","pool","longQueryThreshold","pool","pool","pool","pool","getSuggestion","pool","pool","pool","Database","path","os","fs","Pool"]}
|