@neverinfamous/postgres-mcp 1.2.0 → 1.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +68 -62
- package/dist/adapters/postgresql/PostgresAdapter.d.ts.map +1 -1
- package/dist/adapters/postgresql/PostgresAdapter.js +53 -3
- package/dist/adapters/postgresql/PostgresAdapter.js.map +1 -1
- package/dist/adapters/postgresql/prompts/ltree.js +2 -2
- package/dist/adapters/postgresql/prompts/ltree.js.map +1 -1
- package/dist/adapters/postgresql/schemas/admin.d.ts +10 -5
- package/dist/adapters/postgresql/schemas/admin.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/admin.js +10 -5
- package/dist/adapters/postgresql/schemas/admin.js.map +1 -1
- package/dist/adapters/postgresql/schemas/backup.d.ts +8 -4
- package/dist/adapters/postgresql/schemas/backup.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/backup.js +11 -4
- package/dist/adapters/postgresql/schemas/backup.js.map +1 -1
- package/dist/adapters/postgresql/schemas/core.d.ts +53 -19
- package/dist/adapters/postgresql/schemas/core.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/core.js +61 -17
- package/dist/adapters/postgresql/schemas/core.js.map +1 -1
- package/dist/adapters/postgresql/schemas/cron.d.ts +51 -32
- package/dist/adapters/postgresql/schemas/cron.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/cron.js +64 -44
- package/dist/adapters/postgresql/schemas/cron.js.map +1 -1
- package/dist/adapters/postgresql/schemas/extensions.d.ts +168 -73
- package/dist/adapters/postgresql/schemas/extensions.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/extensions.js +177 -60
- package/dist/adapters/postgresql/schemas/extensions.js.map +1 -1
- package/dist/adapters/postgresql/schemas/index.d.ts +5 -5
- package/dist/adapters/postgresql/schemas/index.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/index.js +9 -7
- package/dist/adapters/postgresql/schemas/index.js.map +1 -1
- package/dist/adapters/postgresql/schemas/jsonb.d.ts +94 -42
- package/dist/adapters/postgresql/schemas/jsonb.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/jsonb.js +101 -30
- package/dist/adapters/postgresql/schemas/jsonb.js.map +1 -1
- package/dist/adapters/postgresql/schemas/monitoring.d.ts +28 -11
- package/dist/adapters/postgresql/schemas/monitoring.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/monitoring.js +49 -24
- package/dist/adapters/postgresql/schemas/monitoring.js.map +1 -1
- package/dist/adapters/postgresql/schemas/partitioning.d.ts +5 -4
- package/dist/adapters/postgresql/schemas/partitioning.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/partitioning.js +5 -4
- package/dist/adapters/postgresql/schemas/partitioning.js.map +1 -1
- package/dist/adapters/postgresql/schemas/performance.d.ts +60 -30
- package/dist/adapters/postgresql/schemas/performance.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/performance.js +85 -22
- package/dist/adapters/postgresql/schemas/performance.js.map +1 -1
- package/dist/adapters/postgresql/schemas/postgis.d.ts +20 -0
- package/dist/adapters/postgresql/schemas/postgis.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/postgis.js +20 -0
- package/dist/adapters/postgresql/schemas/postgis.js.map +1 -1
- package/dist/adapters/postgresql/schemas/schema-mgmt.d.ts +35 -23
- package/dist/adapters/postgresql/schemas/schema-mgmt.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/schema-mgmt.js +69 -26
- package/dist/adapters/postgresql/schemas/schema-mgmt.js.map +1 -1
- package/dist/adapters/postgresql/schemas/stats.d.ts +33 -20
- package/dist/adapters/postgresql/schemas/stats.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/stats.js +36 -20
- package/dist/adapters/postgresql/schemas/stats.js.map +1 -1
- package/dist/adapters/postgresql/schemas/text-search.d.ts +8 -5
- package/dist/adapters/postgresql/schemas/text-search.d.ts.map +1 -1
- package/dist/adapters/postgresql/schemas/text-search.js +11 -4
- package/dist/adapters/postgresql/schemas/text-search.js.map +1 -1
- package/dist/adapters/postgresql/tools/admin.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/admin.js +211 -140
- package/dist/adapters/postgresql/tools/admin.js.map +1 -1
- package/dist/adapters/postgresql/tools/backup/dump.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/backup/dump.js +360 -337
- package/dist/adapters/postgresql/tools/backup/dump.js.map +1 -1
- package/dist/adapters/postgresql/tools/citext.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/citext.js +221 -163
- package/dist/adapters/postgresql/tools/citext.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/convenience.d.ts +9 -1
- package/dist/adapters/postgresql/tools/core/convenience.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/convenience.js +78 -11
- package/dist/adapters/postgresql/tools/core/convenience.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/error-helpers.d.ts +48 -0
- package/dist/adapters/postgresql/tools/core/error-helpers.d.ts.map +1 -0
- package/dist/adapters/postgresql/tools/core/error-helpers.js +256 -0
- package/dist/adapters/postgresql/tools/core/error-helpers.js.map +1 -0
- package/dist/adapters/postgresql/tools/core/health.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/health.js +18 -4
- package/dist/adapters/postgresql/tools/core/health.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/indexes.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/indexes.js +45 -4
- package/dist/adapters/postgresql/tools/core/indexes.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/objects.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/objects.js +104 -85
- package/dist/adapters/postgresql/tools/core/objects.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/query.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/query.js +100 -42
- package/dist/adapters/postgresql/tools/core/query.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/schemas.d.ts +51 -25
- package/dist/adapters/postgresql/tools/core/schemas.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/schemas.js +51 -25
- package/dist/adapters/postgresql/tools/core/schemas.js.map +1 -1
- package/dist/adapters/postgresql/tools/core/tables.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/core/tables.js +68 -28
- package/dist/adapters/postgresql/tools/core/tables.js.map +1 -1
- package/dist/adapters/postgresql/tools/cron.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/cron.js +274 -179
- package/dist/adapters/postgresql/tools/cron.js.map +1 -1
- package/dist/adapters/postgresql/tools/jsonb/advanced.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/jsonb/advanced.js +372 -284
- package/dist/adapters/postgresql/tools/jsonb/advanced.js.map +1 -1
- package/dist/adapters/postgresql/tools/jsonb/basic.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/jsonb/basic.js +617 -398
- package/dist/adapters/postgresql/tools/jsonb/basic.js.map +1 -1
- package/dist/adapters/postgresql/tools/kcache.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/kcache.js +278 -246
- package/dist/adapters/postgresql/tools/kcache.js.map +1 -1
- package/dist/adapters/postgresql/tools/ltree.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/ltree.js +121 -35
- package/dist/adapters/postgresql/tools/ltree.js.map +1 -1
- package/dist/adapters/postgresql/tools/monitoring.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/monitoring.js +54 -34
- package/dist/adapters/postgresql/tools/monitoring.js.map +1 -1
- package/dist/adapters/postgresql/tools/partitioning.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/partitioning.js +79 -15
- package/dist/adapters/postgresql/tools/partitioning.js.map +1 -1
- package/dist/adapters/postgresql/tools/partman/management.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/partman/management.js +11 -4
- package/dist/adapters/postgresql/tools/partman/management.js.map +1 -1
- package/dist/adapters/postgresql/tools/partman/operations.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/partman/operations.js +132 -19
- package/dist/adapters/postgresql/tools/partman/operations.js.map +1 -1
- package/dist/adapters/postgresql/tools/performance/analysis.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/performance/analysis.js +264 -160
- package/dist/adapters/postgresql/tools/performance/analysis.js.map +1 -1
- package/dist/adapters/postgresql/tools/performance/explain.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/performance/explain.js +61 -21
- package/dist/adapters/postgresql/tools/performance/explain.js.map +1 -1
- package/dist/adapters/postgresql/tools/performance/monitoring.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/performance/monitoring.js +44 -7
- package/dist/adapters/postgresql/tools/performance/monitoring.js.map +1 -1
- package/dist/adapters/postgresql/tools/performance/optimization.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/performance/optimization.js +92 -81
- package/dist/adapters/postgresql/tools/performance/optimization.js.map +1 -1
- package/dist/adapters/postgresql/tools/performance/stats.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/performance/stats.js +124 -36
- package/dist/adapters/postgresql/tools/performance/stats.js.map +1 -1
- package/dist/adapters/postgresql/tools/pgcrypto.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/pgcrypto.js +244 -89
- package/dist/adapters/postgresql/tools/pgcrypto.js.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/advanced.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/advanced.js +285 -222
- package/dist/adapters/postgresql/tools/postgis/advanced.js.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/basic.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/basic.js +359 -249
- package/dist/adapters/postgresql/tools/postgis/basic.js.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/standalone.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/postgis/standalone.js +135 -51
- package/dist/adapters/postgresql/tools/postgis/standalone.js.map +1 -1
- package/dist/adapters/postgresql/tools/schema.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/schema.js +504 -231
- package/dist/adapters/postgresql/tools/schema.js.map +1 -1
- package/dist/adapters/postgresql/tools/stats/advanced.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/stats/advanced.js +515 -476
- package/dist/adapters/postgresql/tools/stats/advanced.js.map +1 -1
- package/dist/adapters/postgresql/tools/stats/basic.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/stats/basic.js +302 -293
- package/dist/adapters/postgresql/tools/stats/basic.js.map +1 -1
- package/dist/adapters/postgresql/tools/text.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/text.js +392 -218
- package/dist/adapters/postgresql/tools/text.js.map +1 -1
- package/dist/adapters/postgresql/tools/transactions.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/transactions.js +157 -50
- package/dist/adapters/postgresql/tools/transactions.js.map +1 -1
- package/dist/adapters/postgresql/tools/vector/advanced.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/vector/advanced.js +18 -0
- package/dist/adapters/postgresql/tools/vector/advanced.js.map +1 -1
- package/dist/adapters/postgresql/tools/vector/basic.d.ts.map +1 -1
- package/dist/adapters/postgresql/tools/vector/basic.js +100 -53
- package/dist/adapters/postgresql/tools/vector/basic.js.map +1 -1
- package/dist/codemode/api.js +1 -1
- package/dist/codemode/api.js.map +1 -1
- package/dist/constants/ServerInstructions.d.ts +1 -1
- package/dist/constants/ServerInstructions.d.ts.map +1 -1
- package/dist/constants/ServerInstructions.js +45 -7
- package/dist/constants/ServerInstructions.js.map +1 -1
- package/package.json +9 -5
|
@@ -10,6 +10,7 @@ import { buildProgressContext, sendProgress, } from "../../../../utils/progress-
|
|
|
10
10
|
import { CopyExportSchema, CopyExportSchemaBase, DumpSchemaSchema,
|
|
11
11
|
// Output schemas
|
|
12
12
|
DumpTableOutputSchema, DumpSchemaOutputSchema, CopyExportOutputSchema, CopyImportOutputSchema, } from "../../schemas/index.js";
|
|
13
|
+
import { formatPostgresError } from "../core/error-helpers.js";
|
|
13
14
|
export function createDumpTableTool(adapter) {
|
|
14
15
|
return {
|
|
15
16
|
name: "pg_dump_table",
|
|
@@ -33,35 +34,36 @@ export function createDumpTableTool(adapter) {
|
|
|
33
34
|
annotations: readOnly("Dump Table"),
|
|
34
35
|
icons: getToolIcons("backup", readOnly("Dump Table")),
|
|
35
36
|
handler: async (params, _context) => {
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
// Parse schema.table format (e.g., 'public.users' -> schema='public', table='users')
|
|
42
|
-
// If table contains a dot, always parse it as schema.table (embedded schema takes priority)
|
|
43
|
-
let tableName = parsed.table;
|
|
44
|
-
let schemaName = parsed.schema ?? "public";
|
|
45
|
-
if (parsed.table.includes(".")) {
|
|
46
|
-
const parts = parsed.table.split(".");
|
|
47
|
-
if (parts.length === 2 && parts[0] && parts[1]) {
|
|
48
|
-
schemaName = parts[0];
|
|
49
|
-
tableName = parts[1];
|
|
37
|
+
try {
|
|
38
|
+
const parsed = params;
|
|
39
|
+
// Validate required table parameter
|
|
40
|
+
if (!parsed.table || parsed.table.trim() === "") {
|
|
41
|
+
throw new Error("table parameter is required");
|
|
50
42
|
}
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
43
|
+
// Parse schema.table format (e.g., 'public.users' -> schema='public', table='users')
|
|
44
|
+
// If table contains a dot, always parse it as schema.table (embedded schema takes priority)
|
|
45
|
+
let tableName = parsed.table;
|
|
46
|
+
let schemaName = parsed.schema ?? "public";
|
|
47
|
+
if (parsed.table.includes(".")) {
|
|
48
|
+
const parts = parsed.table.split(".");
|
|
49
|
+
if (parts.length === 2 && parts[0] && parts[1]) {
|
|
50
|
+
schemaName = parts[0];
|
|
51
|
+
tableName = parts[1];
|
|
52
|
+
}
|
|
53
|
+
}
|
|
54
|
+
// Check if it's a sequence by querying pg_class
|
|
55
|
+
const relkindResult = await adapter.executeQuery(`
|
|
54
56
|
SELECT relkind FROM pg_class c
|
|
55
57
|
JOIN pg_namespace n ON c.relnamespace = n.oid
|
|
56
58
|
WHERE n.nspname = '${schemaName}' AND c.relname = '${tableName}'
|
|
57
59
|
`);
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
60
|
+
const relkind = relkindResult.rows?.[0]?.["relkind"];
|
|
61
|
+
// relkind 'S' = sequence
|
|
62
|
+
if (relkind === "S") {
|
|
63
|
+
// Use pg_sequence system catalog (works in all PostgreSQL versions 10+)
|
|
64
|
+
// Fallback to basic DDL if query fails
|
|
65
|
+
try {
|
|
66
|
+
const seqInfo = await adapter.executeQuery(`
|
|
65
67
|
SELECT s.seqstart as start_value, s.seqincrement as increment_by,
|
|
66
68
|
s.seqmin as min_value, s.seqmax as max_value, s.seqcycle as cycle
|
|
67
69
|
FROM pg_sequence s
|
|
@@ -69,91 +71,91 @@ export function createDumpTableTool(adapter) {
|
|
|
69
71
|
JOIN pg_namespace n ON c.relnamespace = n.oid
|
|
70
72
|
WHERE n.nspname = '${schemaName}' AND c.relname = '${tableName}'
|
|
71
73
|
`);
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
|
|
95
|
-
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
74
|
+
const seq = seqInfo.rows?.[0];
|
|
75
|
+
if (seq !== undefined) {
|
|
76
|
+
const startVal = typeof seq["start_value"] === "number" ||
|
|
77
|
+
typeof seq["start_value"] === "bigint"
|
|
78
|
+
? String(seq["start_value"])
|
|
79
|
+
: null;
|
|
80
|
+
const incrVal = typeof seq["increment_by"] === "number" ||
|
|
81
|
+
typeof seq["increment_by"] === "bigint"
|
|
82
|
+
? Number(seq["increment_by"])
|
|
83
|
+
: null;
|
|
84
|
+
const minVal = typeof seq["min_value"] === "number" ||
|
|
85
|
+
typeof seq["min_value"] === "bigint"
|
|
86
|
+
? String(seq["min_value"])
|
|
87
|
+
: null;
|
|
88
|
+
const maxVal = typeof seq["max_value"] === "number" ||
|
|
89
|
+
typeof seq["max_value"] === "bigint"
|
|
90
|
+
? String(seq["max_value"])
|
|
91
|
+
: null;
|
|
92
|
+
const startValue = startVal !== null ? ` START ${startVal}` : "";
|
|
93
|
+
const increment = incrVal !== null && incrVal !== 1
|
|
94
|
+
? ` INCREMENT ${String(incrVal)}`
|
|
95
|
+
: "";
|
|
96
|
+
const minValue = minVal !== null ? ` MINVALUE ${minVal}` : "";
|
|
97
|
+
const maxValue = maxVal !== null ? ` MAXVALUE ${maxVal}` : "";
|
|
98
|
+
const cycle = seq["cycle"] === true ? " CYCLE" : "";
|
|
99
|
+
const ddl = `CREATE SEQUENCE "${schemaName}"."${tableName}"${startValue}${increment}${minValue}${maxValue}${cycle};`;
|
|
100
|
+
return {
|
|
101
|
+
ddl,
|
|
102
|
+
type: "sequence",
|
|
103
|
+
note: "Use pg_list_sequences to see all sequences.",
|
|
104
|
+
...(parsed.includeData === true && {
|
|
105
|
+
warning: "includeData is ignored for sequences - sequences have no row data to export",
|
|
106
|
+
}),
|
|
107
|
+
};
|
|
108
|
+
}
|
|
106
109
|
}
|
|
110
|
+
catch {
|
|
111
|
+
// Query failed, use basic DDL
|
|
112
|
+
}
|
|
113
|
+
// Fallback if pg_sequence query fails
|
|
114
|
+
return {
|
|
115
|
+
ddl: `CREATE SEQUENCE "${schemaName}"."${tableName}";`,
|
|
116
|
+
type: "sequence",
|
|
117
|
+
note: "Basic CREATE SEQUENCE. Use pg_list_sequences for details.",
|
|
118
|
+
...(parsed.includeData === true && {
|
|
119
|
+
warning: "includeData is ignored for sequences - sequences have no row data to export",
|
|
120
|
+
}),
|
|
121
|
+
};
|
|
107
122
|
}
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
111
|
-
|
|
112
|
-
return {
|
|
113
|
-
ddl: `CREATE SEQUENCE "${schemaName}"."${tableName}";`,
|
|
114
|
-
type: "sequence",
|
|
115
|
-
note: "Basic CREATE SEQUENCE. Use pg_list_sequences for details.",
|
|
116
|
-
...(parsed.includeData === true && {
|
|
117
|
-
warning: "includeData is ignored for sequences - sequences have no row data to export",
|
|
118
|
-
}),
|
|
119
|
-
};
|
|
120
|
-
}
|
|
121
|
-
// relkind 'v' = view, 'm' = materialized view
|
|
122
|
-
if (relkind === "v" || relkind === "m") {
|
|
123
|
-
try {
|
|
124
|
-
const viewDefResult = await adapter.executeQuery(`
|
|
123
|
+
// relkind 'v' = view, 'm' = materialized view
|
|
124
|
+
if (relkind === "v" || relkind === "m") {
|
|
125
|
+
try {
|
|
126
|
+
const viewDefResult = await adapter.executeQuery(`
|
|
125
127
|
SELECT definition FROM pg_views
|
|
126
128
|
WHERE schemaname = '${schemaName}' AND viewname = '${tableName}'
|
|
127
129
|
`);
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
136
|
-
|
|
130
|
+
const definition = viewDefResult.rows?.[0]?.["definition"];
|
|
131
|
+
if (typeof definition === "string") {
|
|
132
|
+
const createType = relkind === "m" ? "MATERIALIZED VIEW" : "VIEW";
|
|
133
|
+
const ddl = `CREATE ${createType} "${schemaName}"."${tableName}" AS\n${definition.trim()}`;
|
|
134
|
+
return {
|
|
135
|
+
ddl,
|
|
136
|
+
type: relkind === "m" ? "materialized_view" : "view",
|
|
137
|
+
note: `Use pg_list_views to see all views.`,
|
|
138
|
+
};
|
|
139
|
+
}
|
|
137
140
|
}
|
|
141
|
+
catch {
|
|
142
|
+
// Query failed, use basic DDL
|
|
143
|
+
}
|
|
144
|
+
// Fallback for views
|
|
145
|
+
const createType = relkind === "m" ? "MATERIALIZED VIEW" : "VIEW";
|
|
146
|
+
return {
|
|
147
|
+
ddl: `-- Unable to retrieve ${createType.toLowerCase()} definition\nCREATE ${createType} "${schemaName}"."${tableName}" AS SELECT ...;`,
|
|
148
|
+
type: relkind === "m" ? "materialized_view" : "view",
|
|
149
|
+
note: "View definition could not be retrieved. Use pg_list_views for details.",
|
|
150
|
+
};
|
|
138
151
|
}
|
|
139
|
-
|
|
140
|
-
|
|
141
|
-
|
|
142
|
-
|
|
143
|
-
|
|
144
|
-
|
|
145
|
-
|
|
146
|
-
type: relkind === "m" ? "materialized_view" : "view",
|
|
147
|
-
note: "View definition could not be retrieved. Use pg_list_views for details.",
|
|
148
|
-
};
|
|
149
|
-
}
|
|
150
|
-
// Check if it's a partitioned table (relkind 'p') and get partition info
|
|
151
|
-
let partitionClause = "";
|
|
152
|
-
const isPartitionedTable = relkind === "p";
|
|
153
|
-
if (isPartitionedTable) {
|
|
154
|
-
try {
|
|
155
|
-
// Query pg_partitioned_table to get partition strategy and key columns
|
|
156
|
-
const partInfo = await adapter.executeQuery(`
|
|
152
|
+
// Check if it's a partitioned table (relkind 'p') and get partition info
|
|
153
|
+
let partitionClause = "";
|
|
154
|
+
const isPartitionedTable = relkind === "p";
|
|
155
|
+
if (isPartitionedTable) {
|
|
156
|
+
try {
|
|
157
|
+
// Query pg_partitioned_table to get partition strategy and key columns
|
|
158
|
+
const partInfo = await adapter.executeQuery(`
|
|
157
159
|
SELECT pt.partstrat,
|
|
158
160
|
array_agg(a.attname ORDER BY partattrs.ord) as partition_columns
|
|
159
161
|
FROM pg_partitioned_table pt
|
|
@@ -164,128 +166,137 @@ export function createDumpTableTool(adapter) {
|
|
|
164
166
|
WHERE n.nspname = '${schemaName}' AND c.relname = '${tableName}'
|
|
165
167
|
GROUP BY pt.partstrat
|
|
166
168
|
`);
|
|
167
|
-
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
|
|
171
|
-
|
|
172
|
-
|
|
173
|
-
|
|
174
|
-
|
|
175
|
-
|
|
176
|
-
|
|
177
|
-
|
|
178
|
-
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
|
|
192
|
-
|
|
193
|
-
|
|
194
|
-
|
|
195
|
-
|
|
169
|
+
const partRow = partInfo.rows?.[0];
|
|
170
|
+
if (partRow) {
|
|
171
|
+
const strategy = partRow["partstrat"];
|
|
172
|
+
const columns = partRow["partition_columns"];
|
|
173
|
+
// Map strategy code to keyword
|
|
174
|
+
const strategyMap = {
|
|
175
|
+
r: "RANGE",
|
|
176
|
+
l: "LIST",
|
|
177
|
+
h: "HASH",
|
|
178
|
+
};
|
|
179
|
+
const strategyKeyword = typeof strategy === "string"
|
|
180
|
+
? (strategyMap[strategy] ?? "RANGE")
|
|
181
|
+
: "RANGE";
|
|
182
|
+
// Build column list - PostgreSQL returns array_agg as string like "{col1,col2}"
|
|
183
|
+
let columnList = "";
|
|
184
|
+
if (Array.isArray(columns)) {
|
|
185
|
+
columnList = columns.map((c) => `"${String(c)}"`).join(", ");
|
|
186
|
+
}
|
|
187
|
+
else if (typeof columns === "string") {
|
|
188
|
+
// Parse PostgreSQL array literal format: "{col1,col2}" -> ["col1", "col2"]
|
|
189
|
+
const parsed = columns
|
|
190
|
+
.replace(/^\{/, "")
|
|
191
|
+
.replace(/\}$/, "")
|
|
192
|
+
.split(",")
|
|
193
|
+
.filter((c) => c.length > 0);
|
|
194
|
+
columnList = parsed.map((c) => `"${c.trim()}"`).join(", ");
|
|
195
|
+
}
|
|
196
|
+
if (columnList) {
|
|
197
|
+
partitionClause = ` PARTITION BY ${strategyKeyword} (${columnList})`;
|
|
198
|
+
}
|
|
196
199
|
}
|
|
197
200
|
}
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
// Partition info query failed, continue without partition clause
|
|
201
|
-
}
|
|
202
|
-
}
|
|
203
|
-
const tableInfo = await adapter.describeTable(tableName, schemaName);
|
|
204
|
-
const columns = tableInfo.columns
|
|
205
|
-
?.map((col) => {
|
|
206
|
-
let def = ` "${col.name}" ${col.type}`;
|
|
207
|
-
if (col.defaultValue !== undefined && col.defaultValue !== null) {
|
|
208
|
-
let defaultStr;
|
|
209
|
-
if (typeof col.defaultValue === "object") {
|
|
210
|
-
defaultStr = JSON.stringify(col.defaultValue);
|
|
201
|
+
catch {
|
|
202
|
+
// Partition info query failed, continue without partition clause
|
|
211
203
|
}
|
|
212
|
-
else if (typeof col.defaultValue === "string" ||
|
|
213
|
-
typeof col.defaultValue === "number" ||
|
|
214
|
-
typeof col.defaultValue === "boolean") {
|
|
215
|
-
defaultStr = String(col.defaultValue);
|
|
216
|
-
}
|
|
217
|
-
else {
|
|
218
|
-
defaultStr = JSON.stringify(col.defaultValue);
|
|
219
|
-
}
|
|
220
|
-
def += ` DEFAULT ${defaultStr}`;
|
|
221
204
|
}
|
|
222
|
-
|
|
223
|
-
|
|
224
|
-
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
|
|
230
|
-
|
|
231
|
-
|
|
232
|
-
|
|
233
|
-
|
|
234
|
-
|
|
235
|
-
|
|
236
|
-
|
|
237
|
-
|
|
238
|
-
|
|
239
|
-
|
|
240
|
-
|
|
241
|
-
|
|
242
|
-
|
|
243
|
-
|
|
244
|
-
|
|
245
|
-
|
|
246
|
-
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
|
|
250
|
-
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
254
|
-
|
|
255
|
-
|
|
256
|
-
|
|
257
|
-
|
|
258
|
-
|
|
259
|
-
|
|
260
|
-
|
|
261
|
-
|
|
262
|
-
|
|
263
|
-
|
|
264
|
-
|
|
265
|
-
|
|
266
|
-
|
|
267
|
-
|
|
268
|
-
|
|
205
|
+
const tableInfo = await adapter.describeTable(tableName, schemaName);
|
|
206
|
+
const columns = tableInfo.columns
|
|
207
|
+
?.map((col) => {
|
|
208
|
+
let def = ` "${col.name}" ${col.type}`;
|
|
209
|
+
if (col.defaultValue !== undefined && col.defaultValue !== null) {
|
|
210
|
+
let defaultStr;
|
|
211
|
+
if (typeof col.defaultValue === "object") {
|
|
212
|
+
defaultStr = JSON.stringify(col.defaultValue);
|
|
213
|
+
}
|
|
214
|
+
else if (typeof col.defaultValue === "string" ||
|
|
215
|
+
typeof col.defaultValue === "number" ||
|
|
216
|
+
typeof col.defaultValue === "boolean") {
|
|
217
|
+
defaultStr = String(col.defaultValue);
|
|
218
|
+
}
|
|
219
|
+
else {
|
|
220
|
+
defaultStr = JSON.stringify(col.defaultValue);
|
|
221
|
+
}
|
|
222
|
+
def += ` DEFAULT ${defaultStr}`;
|
|
223
|
+
}
|
|
224
|
+
if (!col.nullable)
|
|
225
|
+
def += " NOT NULL";
|
|
226
|
+
return def;
|
|
227
|
+
})
|
|
228
|
+
.join(",\n") ?? "";
|
|
229
|
+
const createTable = `CREATE TABLE "${schemaName}"."${tableName}" (\n${columns}\n)${partitionClause};`;
|
|
230
|
+
const result = {
|
|
231
|
+
ddl: createTable,
|
|
232
|
+
type: isPartitionedTable ? "partitioned_table" : "table",
|
|
233
|
+
note: isPartitionedTable
|
|
234
|
+
? "For partition children use pg_list_partitions, for indexes use pg_get_indexes, for constraints use pg_get_constraints."
|
|
235
|
+
: "Basic CREATE TABLE only. For indexes use pg_get_indexes, for constraints use pg_get_constraints.",
|
|
236
|
+
};
|
|
237
|
+
if (parsed.includeData) {
|
|
238
|
+
// Default limit is 500 to prevent large payloads, 0 means no limit
|
|
239
|
+
const effectiveLimit = parsed.limit === 0 ? null : (parsed.limit ?? 500);
|
|
240
|
+
const limitClause = effectiveLimit !== null ? ` LIMIT ${String(effectiveLimit)}` : "";
|
|
241
|
+
const dataResult = await adapter.executeQuery(`SELECT * FROM "${schemaName}"."${tableName}"${limitClause}`);
|
|
242
|
+
if (dataResult.rows !== undefined && dataResult.rows.length > 0) {
|
|
243
|
+
const firstRow = dataResult.rows[0];
|
|
244
|
+
if (firstRow === undefined)
|
|
245
|
+
return result;
|
|
246
|
+
const cols = Object.keys(firstRow)
|
|
247
|
+
.map((c) => `"${c}"`)
|
|
248
|
+
.join(", ");
|
|
249
|
+
const inserts = dataResult.rows
|
|
250
|
+
.map((row) => {
|
|
251
|
+
const vals = Object.entries(row)
|
|
252
|
+
.map(([, value]) => {
|
|
253
|
+
if (value === null)
|
|
254
|
+
return "NULL";
|
|
255
|
+
// Handle Date objects - format as PostgreSQL timestamp
|
|
256
|
+
if (value instanceof Date) {
|
|
257
|
+
const iso = value.toISOString();
|
|
258
|
+
// Convert ISO 8601 to PostgreSQL format: 'YYYY-MM-DD HH:MM:SS.mmm'
|
|
259
|
+
const pgTimestamp = iso
|
|
269
260
|
.replace("T", " ")
|
|
270
|
-
.replace("Z", "")
|
|
271
|
-
|
|
272
|
-
return `'${pgTimestamp.replace(/\\/g, "\\\\").replace(/'/g, "''")}'`;
|
|
261
|
+
.replace("Z", "");
|
|
262
|
+
return `'${pgTimestamp}'`;
|
|
273
263
|
}
|
|
274
|
-
|
|
275
|
-
|
|
276
|
-
|
|
277
|
-
|
|
278
|
-
|
|
279
|
-
|
|
264
|
+
if (typeof value === "string") {
|
|
265
|
+
// Escape backslashes first, then single quotes (PostgreSQL string literal escaping)
|
|
266
|
+
const escaped = value
|
|
267
|
+
.replace(/\\/g, "\\\\")
|
|
268
|
+
.replace(/'/g, "''");
|
|
269
|
+
// Check if string looks like an ISO timestamp
|
|
270
|
+
if (/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}/.test(value)) {
|
|
271
|
+
// Convert ISO format to PostgreSQL format
|
|
272
|
+
const pgTimestamp = value
|
|
273
|
+
.replace("T", " ")
|
|
274
|
+
.replace("Z", "")
|
|
275
|
+
.replace(/\.\d+$/, "");
|
|
276
|
+
return `'${pgTimestamp.replace(/\\/g, "\\\\").replace(/'/g, "''")}'`;
|
|
277
|
+
}
|
|
278
|
+
return `'${escaped}'`;
|
|
279
|
+
}
|
|
280
|
+
if (typeof value === "number" || typeof value === "boolean")
|
|
281
|
+
return String(value);
|
|
282
|
+
// For objects (JSONB, arrays), use PostgreSQL JSONB literal
|
|
283
|
+
return `'${JSON.stringify(value).replace(/\\/g, "\\\\").replace(/'/g, "''")}'::jsonb`;
|
|
284
|
+
})
|
|
285
|
+
.join(", ");
|
|
286
|
+
return `INSERT INTO "${schemaName}"."${tableName}" (${cols}) VALUES (${vals});`;
|
|
280
287
|
})
|
|
281
|
-
.join("
|
|
282
|
-
|
|
283
|
-
}
|
|
284
|
-
.join("\n");
|
|
285
|
-
result.insertStatements = inserts;
|
|
288
|
+
.join("\n");
|
|
289
|
+
result.insertStatements = inserts;
|
|
290
|
+
}
|
|
286
291
|
}
|
|
292
|
+
return result;
|
|
293
|
+
}
|
|
294
|
+
catch (error) {
|
|
295
|
+
return {
|
|
296
|
+
success: false,
|
|
297
|
+
error: formatPostgresError(error, { tool: "pg_dump_table" }),
|
|
298
|
+
};
|
|
287
299
|
}
|
|
288
|
-
return result;
|
|
289
300
|
},
|
|
290
301
|
};
|
|
291
302
|
}
|
|
@@ -343,144 +354,156 @@ export function createCopyExportTool(adapter) {
|
|
|
343
354
|
annotations: readOnly("Copy Export"),
|
|
344
355
|
icons: getToolIcons("backup", readOnly("Copy Export")),
|
|
345
356
|
handler: async (params, context) => {
|
|
346
|
-
|
|
347
|
-
|
|
348
|
-
|
|
349
|
-
|
|
350
|
-
|
|
351
|
-
|
|
352
|
-
|
|
353
|
-
|
|
354
|
-
|
|
355
|
-
|
|
356
|
-
|
|
357
|
-
|
|
358
|
-
|
|
359
|
-
|
|
360
|
-
|
|
361
|
-
if (
|
|
362
|
-
|
|
363
|
-
|
|
364
|
-
|
|
365
|
-
|
|
366
|
-
|
|
367
|
-
|
|
368
|
-
|
|
369
|
-
|
|
370
|
-
|
|
371
|
-
|
|
372
|
-
|
|
357
|
+
try {
|
|
358
|
+
const progress = buildProgressContext(context);
|
|
359
|
+
await sendProgress(progress, 1, 3, "Preparing COPY export...");
|
|
360
|
+
const { query, format, header, delimiter, conflictWarning, effectiveLimit, } = CopyExportSchema.parse(params); // Use transform for validation
|
|
361
|
+
const options = [];
|
|
362
|
+
options.push(`FORMAT ${format ?? "csv"}`);
|
|
363
|
+
if (header !== false)
|
|
364
|
+
options.push("HEADER");
|
|
365
|
+
if (delimiter)
|
|
366
|
+
options.push(`DELIMITER '${delimiter}'`);
|
|
367
|
+
const copyCommand = `COPY (${query}) TO STDOUT WITH (${options.join(", ")})`;
|
|
368
|
+
void copyCommand;
|
|
369
|
+
await sendProgress(progress, 2, 3, "Executing query...");
|
|
370
|
+
const result = await adapter.executeQuery(query);
|
|
371
|
+
// Handle CSV format (default)
|
|
372
|
+
if (format === "csv" || format === undefined) {
|
|
373
|
+
if (result.rows === undefined || result.rows.length === 0) {
|
|
374
|
+
return {
|
|
375
|
+
data: "",
|
|
376
|
+
rowCount: 0,
|
|
377
|
+
note: "Query returned no rows. Headers omitted for empty results.",
|
|
378
|
+
...(conflictWarning !== undefined
|
|
379
|
+
? { warning: conflictWarning }
|
|
380
|
+
: {}),
|
|
381
|
+
};
|
|
382
|
+
}
|
|
383
|
+
const firstRowData = result.rows[0];
|
|
384
|
+
if (firstRowData === undefined) {
|
|
385
|
+
return {
|
|
386
|
+
data: "",
|
|
387
|
+
rowCount: 0,
|
|
388
|
+
note: "Query returned no rows. Headers omitted for empty results.",
|
|
389
|
+
...(conflictWarning !== undefined
|
|
390
|
+
? { warning: conflictWarning }
|
|
391
|
+
: {}),
|
|
392
|
+
};
|
|
393
|
+
}
|
|
394
|
+
const headers = Object.keys(firstRowData);
|
|
395
|
+
const delim = delimiter ?? ",";
|
|
396
|
+
const lines = [];
|
|
397
|
+
if (header !== false) {
|
|
398
|
+
lines.push(headers.join(delim));
|
|
399
|
+
}
|
|
400
|
+
for (const row of result.rows) {
|
|
401
|
+
lines.push(headers
|
|
402
|
+
.map((h) => {
|
|
403
|
+
const v = row[h];
|
|
404
|
+
if (v === null)
|
|
405
|
+
return "";
|
|
406
|
+
if (typeof v === "object")
|
|
407
|
+
return JSON.stringify(v);
|
|
408
|
+
if (typeof v !== "string" &&
|
|
409
|
+
typeof v !== "number" &&
|
|
410
|
+
typeof v !== "boolean") {
|
|
411
|
+
return JSON.stringify(v);
|
|
412
|
+
}
|
|
413
|
+
const s = String(v);
|
|
414
|
+
return s.includes(delim) ||
|
|
415
|
+
s.includes('"') ||
|
|
416
|
+
s.includes("\n")
|
|
417
|
+
? `"${s.replace(/"/g, '""')}"`
|
|
418
|
+
: s;
|
|
419
|
+
})
|
|
420
|
+
.join(delim));
|
|
421
|
+
}
|
|
422
|
+
// Mark as truncated if any limit was applied AND rows returned equals that limit
|
|
423
|
+
// This indicates there are likely more rows available
|
|
424
|
+
const isTruncated = effectiveLimit !== undefined &&
|
|
425
|
+
result.rows.length === effectiveLimit;
|
|
426
|
+
await sendProgress(progress, 3, 3, "Export complete");
|
|
373
427
|
return {
|
|
374
|
-
data: "",
|
|
375
|
-
rowCount:
|
|
376
|
-
|
|
428
|
+
data: lines.join("\n"),
|
|
429
|
+
rowCount: result.rows.length,
|
|
430
|
+
...(isTruncated ? { truncated: true, limit: effectiveLimit } : {}),
|
|
377
431
|
...(conflictWarning !== undefined
|
|
378
432
|
? { warning: conflictWarning }
|
|
379
433
|
: {}),
|
|
380
434
|
};
|
|
381
435
|
}
|
|
382
|
-
|
|
383
|
-
|
|
384
|
-
|
|
385
|
-
|
|
386
|
-
|
|
387
|
-
|
|
388
|
-
|
|
389
|
-
|
|
390
|
-
|
|
391
|
-
|
|
392
|
-
|
|
393
|
-
|
|
394
|
-
|
|
395
|
-
|
|
396
|
-
|
|
397
|
-
|
|
398
|
-
|
|
436
|
+
// Handle TEXT format - tab-delimited with \N for NULLs
|
|
437
|
+
if (format === "text") {
|
|
438
|
+
if (result.rows === undefined || result.rows.length === 0) {
|
|
439
|
+
return {
|
|
440
|
+
data: "",
|
|
441
|
+
rowCount: 0,
|
|
442
|
+
note: "Query returned no rows. Headers omitted for empty results.",
|
|
443
|
+
...(conflictWarning !== undefined
|
|
444
|
+
? { warning: conflictWarning }
|
|
445
|
+
: {}),
|
|
446
|
+
};
|
|
447
|
+
}
|
|
448
|
+
const firstRowData = result.rows[0];
|
|
449
|
+
if (firstRowData === undefined) {
|
|
450
|
+
return {
|
|
451
|
+
data: "",
|
|
452
|
+
rowCount: 0,
|
|
453
|
+
note: "Query returned no rows. Headers omitted for empty results.",
|
|
454
|
+
...(conflictWarning !== undefined
|
|
455
|
+
? { warning: conflictWarning }
|
|
456
|
+
: {}),
|
|
457
|
+
};
|
|
458
|
+
}
|
|
459
|
+
const headers = Object.keys(firstRowData);
|
|
460
|
+
const delim = delimiter ?? "\t";
|
|
461
|
+
const lines = [];
|
|
462
|
+
if (header !== false) {
|
|
463
|
+
lines.push(headers.join(delim));
|
|
464
|
+
}
|
|
465
|
+
for (const row of result.rows) {
|
|
466
|
+
lines.push(headers
|
|
467
|
+
.map((h) => {
|
|
468
|
+
const v = row[h];
|
|
469
|
+
if (v === null)
|
|
470
|
+
return "\\N"; // PostgreSQL NULL representation in text format
|
|
471
|
+
if (typeof v === "object")
|
|
472
|
+
return JSON.stringify(v);
|
|
473
|
+
if (typeof v === "string" ||
|
|
474
|
+
typeof v === "number" ||
|
|
475
|
+
typeof v === "boolean") {
|
|
476
|
+
return String(v);
|
|
477
|
+
}
|
|
478
|
+
// Fallback for any other type
|
|
399
479
|
return JSON.stringify(v);
|
|
400
|
-
}
|
|
401
|
-
|
|
402
|
-
|
|
403
|
-
|
|
404
|
-
|
|
405
|
-
|
|
406
|
-
.
|
|
407
|
-
|
|
408
|
-
// Mark as truncated if any limit was applied AND rows returned equals that limit
|
|
409
|
-
// This indicates there are likely more rows available
|
|
410
|
-
const isTruncated = effectiveLimit !== undefined && result.rows.length === effectiveLimit;
|
|
411
|
-
await sendProgress(progress, 3, 3, "Export complete");
|
|
412
|
-
return {
|
|
413
|
-
data: lines.join("\n"),
|
|
414
|
-
rowCount: result.rows.length,
|
|
415
|
-
...(isTruncated ? { truncated: true, limit: effectiveLimit } : {}),
|
|
416
|
-
...(conflictWarning !== undefined
|
|
417
|
-
? { warning: conflictWarning }
|
|
418
|
-
: {}),
|
|
419
|
-
};
|
|
420
|
-
}
|
|
421
|
-
// Handle TEXT format - tab-delimited with \N for NULLs
|
|
422
|
-
if (format === "text") {
|
|
423
|
-
if (result.rows === undefined || result.rows.length === 0) {
|
|
424
|
-
return {
|
|
425
|
-
data: "",
|
|
426
|
-
rowCount: 0,
|
|
427
|
-
note: "Query returned no rows. Headers omitted for empty results.",
|
|
428
|
-
...(conflictWarning !== undefined
|
|
429
|
-
? { warning: conflictWarning }
|
|
430
|
-
: {}),
|
|
431
|
-
};
|
|
432
|
-
}
|
|
433
|
-
const firstRowData = result.rows[0];
|
|
434
|
-
if (firstRowData === undefined) {
|
|
480
|
+
})
|
|
481
|
+
.join(delim));
|
|
482
|
+
}
|
|
483
|
+
// Mark as truncated if any limit was applied AND rows returned equals that limit
|
|
484
|
+
// This indicates there are likely more rows available
|
|
485
|
+
const isTruncated = effectiveLimit !== undefined &&
|
|
486
|
+
result.rows.length === effectiveLimit;
|
|
487
|
+
await sendProgress(progress, 3, 3, "Export complete");
|
|
435
488
|
return {
|
|
436
|
-
data: "",
|
|
437
|
-
rowCount:
|
|
438
|
-
|
|
489
|
+
data: lines.join("\n"),
|
|
490
|
+
rowCount: result.rows.length,
|
|
491
|
+
...(isTruncated ? { truncated: true, limit: effectiveLimit } : {}),
|
|
439
492
|
...(conflictWarning !== undefined
|
|
440
493
|
? { warning: conflictWarning }
|
|
441
494
|
: {}),
|
|
442
495
|
};
|
|
443
496
|
}
|
|
444
|
-
|
|
445
|
-
|
|
446
|
-
|
|
447
|
-
|
|
448
|
-
|
|
449
|
-
}
|
|
450
|
-
for (const row of result.rows) {
|
|
451
|
-
lines.push(headers
|
|
452
|
-
.map((h) => {
|
|
453
|
-
const v = row[h];
|
|
454
|
-
if (v === null)
|
|
455
|
-
return "\\N"; // PostgreSQL NULL representation in text format
|
|
456
|
-
if (typeof v === "object")
|
|
457
|
-
return JSON.stringify(v);
|
|
458
|
-
if (typeof v === "string" ||
|
|
459
|
-
typeof v === "number" ||
|
|
460
|
-
typeof v === "boolean") {
|
|
461
|
-
return String(v);
|
|
462
|
-
}
|
|
463
|
-
// Fallback for any other type
|
|
464
|
-
return JSON.stringify(v);
|
|
465
|
-
})
|
|
466
|
-
.join(delim));
|
|
467
|
-
}
|
|
468
|
-
// Mark as truncated if any limit was applied AND rows returned equals that limit
|
|
469
|
-
// This indicates there are likely more rows available
|
|
470
|
-
const isTruncated = effectiveLimit !== undefined && result.rows.length === effectiveLimit;
|
|
471
|
-
await sendProgress(progress, 3, 3, "Export complete");
|
|
497
|
+
// Handle BINARY format - not supported via MCP protocol
|
|
498
|
+
// Binary data cannot be safely serialized to JSON without corruption
|
|
499
|
+
throw new Error('Binary format is not supported via MCP protocol. Use format: "csv" or "text" instead. For binary export, use pg_dump_schema to generate a pg_dump command.');
|
|
500
|
+
}
|
|
501
|
+
catch (error) {
|
|
472
502
|
return {
|
|
473
|
-
|
|
474
|
-
|
|
475
|
-
...(isTruncated ? { truncated: true, limit: effectiveLimit } : {}),
|
|
476
|
-
...(conflictWarning !== undefined
|
|
477
|
-
? { warning: conflictWarning }
|
|
478
|
-
: {}),
|
|
503
|
+
success: false,
|
|
504
|
+
error: formatPostgresError(error, { tool: "pg_copy_export" }),
|
|
479
505
|
};
|
|
480
506
|
}
|
|
481
|
-
// Handle BINARY format - not supported via MCP protocol
|
|
482
|
-
// Binary data cannot be safely serialized to JSON without corruption
|
|
483
|
-
throw new Error('Binary format is not supported via MCP protocol. Use format: "csv" or "text" instead. For binary export, use pg_dump_schema to generate a pg_dump command.');
|
|
484
507
|
},
|
|
485
508
|
};
|
|
486
509
|
}
|