@event-driven-io/dumbo 0.13.0-beta.2 → 0.13.0-beta.21
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/chunk-65DKXURG.js +481 -0
- package/dist/chunk-65DKXURG.js.map +1 -0
- package/dist/chunk-7WEBHXZD.cjs +481 -0
- package/dist/chunk-7WEBHXZD.cjs.map +1 -0
- package/dist/chunk-AMACBMAZ.cjs +556 -0
- package/dist/chunk-AMACBMAZ.cjs.map +1 -0
- package/dist/chunk-FC4JR2R3.js +83 -0
- package/dist/chunk-FC4JR2R3.js.map +1 -0
- package/dist/chunk-H7ZKIJHB.cjs +83 -0
- package/dist/chunk-H7ZKIJHB.cjs.map +1 -0
- package/dist/chunk-IVJ3SNPJ.js +34 -0
- package/dist/chunk-IVJ3SNPJ.js.map +1 -0
- package/dist/chunk-R7P7CNMK.cjs +34 -0
- package/dist/chunk-R7P7CNMK.cjs.map +1 -0
- package/dist/{chunk-XVV3OOQX.js → chunk-T4A6BQRA.js} +209 -58
- package/dist/chunk-T4A6BQRA.js.map +1 -0
- package/dist/{chunk-N7RWT46K.js → chunk-XO5T5N5S.js} +2135 -1613
- package/dist/chunk-XO5T5N5S.js.map +1 -0
- package/dist/{chunk-OJ34O3Q2.cjs → chunk-XOQHZABK.cjs} +2161 -1639
- package/dist/chunk-XOQHZABK.cjs.map +1 -0
- package/dist/cloudflare.cjs +451 -0
- package/dist/cloudflare.cjs.map +1 -0
- package/dist/cloudflare.d.cts +92 -0
- package/dist/cloudflare.d.ts +92 -0
- package/dist/cloudflare.js +451 -0
- package/dist/cloudflare.js.map +1 -0
- package/dist/{columnProcessors-DMPpTPqM.d.ts → columnProcessors-C7fRa54L.d.ts} +1 -1
- package/dist/{columnProcessors-BX-sH7ah.d.cts → columnProcessors-Difrrk4c.d.cts} +1 -1
- package/dist/{connectionString-B1wm0TFc.d.cts → connectionString-oeg1LD5V.d.cts} +352 -137
- package/dist/{connectionString-B1wm0TFc.d.ts → connectionString-oeg1LD5V.d.ts} +352 -137
- package/dist/index.cjs +96 -4
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +23 -18
- package/dist/index.d.ts +23 -18
- package/dist/index.js +101 -9
- package/dist/index.js.map +1 -1
- package/dist/pg.cjs +382 -9
- package/dist/pg.cjs.map +1 -1
- package/dist/pg.d.cts +81 -105
- package/dist/pg.d.ts +81 -105
- package/dist/pg.js +434 -61
- package/dist/pg.js.map +1 -1
- package/dist/postgresql.cjs +50 -0
- package/dist/postgresql.cjs.map +1 -0
- package/dist/postgresql.d.cts +61 -0
- package/dist/postgresql.d.ts +61 -0
- package/dist/postgresql.js +50 -0
- package/dist/sqlite.cjs +56 -0
- package/dist/sqlite.cjs.map +1 -0
- package/dist/{index-C0h0c380.d.cts → sqlite.d.cts} +31 -10
- package/dist/{index-C2z_XBn6.d.ts → sqlite.d.ts} +31 -10
- package/dist/sqlite.js +56 -0
- package/dist/sqlite3.cjs +240 -11
- package/dist/sqlite3.cjs.map +1 -1
- package/dist/sqlite3.d.cts +39 -19
- package/dist/sqlite3.d.ts +39 -19
- package/dist/sqlite3.js +251 -22
- package/dist/sqlite3.js.map +1 -1
- package/package.json +69 -29
- package/dist/chunk-A7TC7IOP.cjs +0 -55
- package/dist/chunk-A7TC7IOP.cjs.map +0 -1
- package/dist/chunk-F7JQ7BX7.js +0 -240
- package/dist/chunk-F7JQ7BX7.js.map +0 -1
- package/dist/chunk-I34X53VL.js +0 -688
- package/dist/chunk-I34X53VL.js.map +0 -1
- package/dist/chunk-ISNF6USX.cjs +0 -688
- package/dist/chunk-ISNF6USX.cjs.map +0 -1
- package/dist/chunk-IYEHOE4S.cjs +0 -405
- package/dist/chunk-IYEHOE4S.cjs.map +0 -1
- package/dist/chunk-L2YZQAG3.cjs +0 -240
- package/dist/chunk-L2YZQAG3.cjs.map +0 -1
- package/dist/chunk-N7RWT46K.js.map +0 -1
- package/dist/chunk-OJ34O3Q2.cjs.map +0 -1
- package/dist/chunk-TXSETOGH.js +0 -55
- package/dist/chunk-TXSETOGH.js.map +0 -1
- package/dist/chunk-XVV3OOQX.js.map +0 -1
- package/dist/d1.cjs +0 -277
- package/dist/d1.cjs.map +0 -1
- package/dist/d1.d.cts +0 -72
- package/dist/d1.d.ts +0 -72
- package/dist/d1.js +0 -277
- package/dist/d1.js.map +0 -1
- package/dist/pg-3ACXFMU4.cjs +0 -59
- package/dist/pg-3ACXFMU4.cjs.map +0 -1
- package/dist/pg-GHOW3XSG.js +0 -59
- package/dist/sqlite3-EEIKQCJR.js +0 -25
- package/dist/sqlite3-SE4DDYZE.cjs +0 -25
- package/dist/sqlite3-SE4DDYZE.cjs.map +0 -1
- /package/dist/{pg-GHOW3XSG.js.map → postgresql.js.map} +0 -0
- /package/dist/{sqlite3-EEIKQCJR.js.map → sqlite.js.map} +0 -0
package/dist/pg.cjs
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true});
|
|
1
|
+
"use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; } function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }
|
|
2
2
|
|
|
3
3
|
|
|
4
4
|
|
|
@@ -11,6 +11,7 @@
|
|
|
11
11
|
|
|
12
12
|
|
|
13
13
|
|
|
14
|
+
var _chunk7WEBHXZDcjs = require('./chunk-7WEBHXZD.cjs');
|
|
14
15
|
|
|
15
16
|
|
|
16
17
|
|
|
@@ -21,6 +22,7 @@
|
|
|
21
22
|
|
|
22
23
|
|
|
23
24
|
|
|
25
|
+
var _chunkH7ZKIJHBcjs = require('./chunk-H7ZKIJHB.cjs');
|
|
24
26
|
|
|
25
27
|
|
|
26
28
|
|
|
@@ -32,18 +34,394 @@
|
|
|
32
34
|
|
|
33
35
|
|
|
34
36
|
|
|
37
|
+
var _chunkXOQHZABKcjs = require('./chunk-XOQHZABK.cjs');
|
|
35
38
|
|
|
39
|
+
// src/storage/postgresql/pg/connections/connection.ts
|
|
40
|
+
var _pg = require('pg'); var _pg2 = _interopRequireDefault(_pg);
|
|
36
41
|
|
|
42
|
+
// src/storage/postgresql/pg/execute/execute.ts
|
|
37
43
|
|
|
44
|
+
var isPgNativePool = (poolOrClient) => {
|
|
45
|
+
return poolOrClient instanceof _pg2.default.Pool;
|
|
46
|
+
};
|
|
47
|
+
var isPgClient = (poolOrClient) => poolOrClient instanceof _pg2.default.Client;
|
|
48
|
+
var isPgPoolClient = (poolOrClient) => "release" in poolOrClient && typeof poolOrClient.release === "function";
|
|
49
|
+
var pgExecute = async (poolOrClient, handle) => {
|
|
50
|
+
const client = isPgNativePool(poolOrClient) ? await poolOrClient.connect() : poolOrClient;
|
|
51
|
+
try {
|
|
52
|
+
return await handle(client);
|
|
53
|
+
} finally {
|
|
54
|
+
if (isPgNativePool(poolOrClient) && isPgPoolClient(client))
|
|
55
|
+
client.release();
|
|
56
|
+
}
|
|
57
|
+
};
|
|
58
|
+
var pgSQLExecutor = ({
|
|
59
|
+
serializer
|
|
60
|
+
}) => ({
|
|
61
|
+
driverType: PgDriverType,
|
|
62
|
+
query: async (client, sql, options) => {
|
|
63
|
+
const results = await batchQuery(
|
|
64
|
+
client,
|
|
65
|
+
[sql],
|
|
66
|
+
serializer,
|
|
67
|
+
options
|
|
68
|
+
);
|
|
69
|
+
return results[0];
|
|
70
|
+
},
|
|
71
|
+
batchQuery: (client, sqls, options) => batchQuery(client, sqls, serializer, options),
|
|
72
|
+
command: async (client, sql, options) => {
|
|
73
|
+
const results = await batchCommand(
|
|
74
|
+
client,
|
|
75
|
+
[sql],
|
|
76
|
+
serializer,
|
|
77
|
+
options
|
|
78
|
+
);
|
|
79
|
+
return results[0];
|
|
80
|
+
},
|
|
81
|
+
batchCommand: (client, sqls, options) => batchCommand(client, sqls, serializer, options),
|
|
82
|
+
formatter: _chunk7WEBHXZDcjs.pgFormatter
|
|
83
|
+
});
|
|
84
|
+
async function batchQuery(client, sqls, serializer, options) {
|
|
85
|
+
const results = Array(
|
|
86
|
+
sqls.length
|
|
87
|
+
);
|
|
88
|
+
if (_optionalChain([options, 'optionalAccess', _ => _.timeoutMs])) {
|
|
89
|
+
await client.query(`SET statement_timeout = ${options.timeoutMs}`);
|
|
90
|
+
}
|
|
91
|
+
for (let i = 0; i < sqls.length; i++) {
|
|
92
|
+
const { query, params } = _chunk7WEBHXZDcjs.pgFormatter.format(sqls[i], { serializer });
|
|
93
|
+
_chunkXOQHZABKcjs.tracer.info("db:sql:query", {
|
|
94
|
+
query,
|
|
95
|
+
params,
|
|
96
|
+
debugSQL: _chunk7WEBHXZDcjs.pgFormatter.describe(sqls[i], { serializer })
|
|
97
|
+
});
|
|
98
|
+
try {
|
|
99
|
+
let result = params.length > 0 ? await client.query(query, params) : await client.query(query);
|
|
100
|
+
if (_optionalChain([options, 'optionalAccess', _2 => _2.mapping])) {
|
|
101
|
+
result = {
|
|
102
|
+
...result,
|
|
103
|
+
rows: result.rows.map(
|
|
104
|
+
(row) => _chunkXOQHZABKcjs.mapSQLQueryResult.call(void 0, row, options.mapping)
|
|
105
|
+
)
|
|
106
|
+
};
|
|
107
|
+
}
|
|
108
|
+
results[i] = { rowCount: result.rowCount, rows: result.rows };
|
|
109
|
+
} catch (error) {
|
|
110
|
+
_chunkXOQHZABKcjs.tracer.error("db:sql:batch_query:execute:error", { error });
|
|
111
|
+
throw _chunk7WEBHXZDcjs.mapPostgresError.call(void 0, error);
|
|
112
|
+
}
|
|
113
|
+
}
|
|
114
|
+
return results;
|
|
115
|
+
}
|
|
116
|
+
async function batchCommand(client, sqls, serializer, options) {
|
|
117
|
+
const results = Array(
|
|
118
|
+
sqls.length
|
|
119
|
+
);
|
|
120
|
+
if (_optionalChain([options, 'optionalAccess', _3 => _3.timeoutMs])) {
|
|
121
|
+
await client.query(`SET statement_timeout = ${options.timeoutMs}`);
|
|
122
|
+
}
|
|
123
|
+
for (let i = 0; i < sqls.length; i++) {
|
|
124
|
+
const { query, params } = _chunk7WEBHXZDcjs.pgFormatter.format(sqls[i], { serializer });
|
|
125
|
+
_chunkXOQHZABKcjs.tracer.info("db:sql:command", {
|
|
126
|
+
query,
|
|
127
|
+
params,
|
|
128
|
+
debugSQL: _chunk7WEBHXZDcjs.pgFormatter.describe(sqls[i], { serializer })
|
|
129
|
+
});
|
|
130
|
+
try {
|
|
131
|
+
let result = params.length > 0 ? await client.query(query, params) : await client.query(query);
|
|
132
|
+
if (_optionalChain([options, 'optionalAccess', _4 => _4.mapping])) {
|
|
133
|
+
result = {
|
|
134
|
+
...result,
|
|
135
|
+
rows: result.rows.map(
|
|
136
|
+
(row) => _chunkXOQHZABKcjs.mapSQLQueryResult.call(void 0, row, options.mapping)
|
|
137
|
+
)
|
|
138
|
+
};
|
|
139
|
+
}
|
|
140
|
+
results[i] = { rowCount: result.rowCount, rows: result.rows };
|
|
141
|
+
if (_optionalChain([options, 'optionalAccess', _5 => _5.assertChanges]) && (_nullishCoalesce(results[i].rowCount, () => ( 0))) === 0) {
|
|
142
|
+
throw new (0, _chunkXOQHZABKcjs.BatchCommandNoChangesError)(i);
|
|
143
|
+
}
|
|
144
|
+
} catch (error) {
|
|
145
|
+
_chunkXOQHZABKcjs.tracer.error("db:sql:batch_command:execute:error", { error });
|
|
146
|
+
throw _chunk7WEBHXZDcjs.mapPostgresError.call(void 0, error);
|
|
147
|
+
}
|
|
148
|
+
}
|
|
149
|
+
return results;
|
|
150
|
+
}
|
|
151
|
+
|
|
152
|
+
// src/storage/postgresql/pg/connections/transaction.ts
|
|
153
|
+
var pgTransaction = (connection, serializer) => (getClient, options) => ({
|
|
154
|
+
connection: connection(),
|
|
155
|
+
driverType: PgDriverType,
|
|
156
|
+
begin: async () => {
|
|
157
|
+
const client = await getClient;
|
|
158
|
+
const parts = ["BEGIN"];
|
|
159
|
+
if (_optionalChain([options, 'optionalAccess', _6 => _6.isolationLevel])) {
|
|
160
|
+
parts.push(`ISOLATION LEVEL ${options.isolationLevel}`);
|
|
161
|
+
}
|
|
162
|
+
if (_optionalChain([options, 'optionalAccess', _7 => _7.readonly])) {
|
|
163
|
+
parts.push("READ ONLY");
|
|
164
|
+
}
|
|
165
|
+
await client.query(parts.join(" "));
|
|
166
|
+
},
|
|
167
|
+
commit: async () => {
|
|
168
|
+
const client = await getClient;
|
|
169
|
+
try {
|
|
170
|
+
await client.query("COMMIT");
|
|
171
|
+
} finally {
|
|
172
|
+
if (_optionalChain([options, 'optionalAccess', _8 => _8.close])) await _optionalChain([options, 'optionalAccess', _9 => _9.close, 'call', _10 => _10(client)]);
|
|
173
|
+
}
|
|
174
|
+
},
|
|
175
|
+
rollback: async (error) => {
|
|
176
|
+
const client = await getClient;
|
|
177
|
+
try {
|
|
178
|
+
await client.query("ROLLBACK");
|
|
179
|
+
} finally {
|
|
180
|
+
if (_optionalChain([options, 'optionalAccess', _11 => _11.close])) await _optionalChain([options, 'optionalAccess', _12 => _12.close, 'call', _13 => _13(client, error)]);
|
|
181
|
+
}
|
|
182
|
+
},
|
|
183
|
+
execute: _chunkXOQHZABKcjs.sqlExecutor.call(void 0, pgSQLExecutor({ serializer }), {
|
|
184
|
+
connect: () => getClient
|
|
185
|
+
})
|
|
186
|
+
});
|
|
187
|
+
|
|
188
|
+
// src/storage/postgresql/pg/connections/connection.ts
|
|
189
|
+
var PgDriverType = "PostgreSQL:pg";
|
|
190
|
+
var pgClientConnection = (options) => {
|
|
191
|
+
const { connect, close } = options;
|
|
192
|
+
return _chunkXOQHZABKcjs.createConnection.call(void 0, {
|
|
193
|
+
driverType: PgDriverType,
|
|
194
|
+
connect,
|
|
195
|
+
close,
|
|
196
|
+
initTransaction: (connection) => pgTransaction(connection, options.serializer),
|
|
197
|
+
executor: pgSQLExecutor,
|
|
198
|
+
serializer: options.serializer
|
|
199
|
+
});
|
|
200
|
+
};
|
|
201
|
+
var pgPoolClientConnection = (options) => {
|
|
202
|
+
const { connect, close } = options;
|
|
203
|
+
return _chunkXOQHZABKcjs.createConnection.call(void 0, {
|
|
204
|
+
driverType: PgDriverType,
|
|
205
|
+
connect,
|
|
206
|
+
close,
|
|
207
|
+
initTransaction: (connection) => pgTransaction(connection, options.serializer),
|
|
208
|
+
executor: pgSQLExecutor,
|
|
209
|
+
serializer: options.serializer
|
|
210
|
+
});
|
|
211
|
+
};
|
|
212
|
+
function pgConnection(options) {
|
|
213
|
+
return options.type === "Client" ? pgClientConnection(options) : pgPoolClientConnection(options);
|
|
214
|
+
}
|
|
215
|
+
var checkConnection = async (connectionString) => {
|
|
216
|
+
const client = new _pg2.default.Client({
|
|
217
|
+
connectionString
|
|
218
|
+
});
|
|
219
|
+
try {
|
|
220
|
+
await client.connect();
|
|
221
|
+
return { successful: true };
|
|
222
|
+
} catch (error) {
|
|
223
|
+
const code = error instanceof Error && "code" in error && typeof error.code === "string" ? error.code : void 0;
|
|
224
|
+
return {
|
|
225
|
+
successful: false,
|
|
226
|
+
errorType: code === "ECONNREFUSED" ? "ConnectionRefused" : code === "28P01" ? "Authentication" : "Unknown",
|
|
227
|
+
code,
|
|
228
|
+
error
|
|
229
|
+
};
|
|
230
|
+
} finally {
|
|
231
|
+
await client.end();
|
|
232
|
+
}
|
|
233
|
+
};
|
|
234
|
+
|
|
235
|
+
// src/storage/postgresql/pg/connections/pool.ts
|
|
236
|
+
|
|
237
|
+
|
|
238
|
+
// src/storage/postgresql/pg/serialization/index.ts
|
|
239
|
+
var setPgTypeParser = (client, options) => {
|
|
240
|
+
if (_optionalChain([options, 'optionalAccess', _14 => _14.parseBigInts]) === true)
|
|
241
|
+
client.setTypeParser(20, (val) => BigInt(val));
|
|
242
|
+
if (_optionalChain([options, 'optionalAccess', _15 => _15.serializer])) {
|
|
243
|
+
client.setTypeParser(3802, (val) => options.serializer.deserialize(val));
|
|
244
|
+
client.setTypeParser(114, (val) => options.serializer.deserialize(val));
|
|
245
|
+
}
|
|
246
|
+
};
|
|
247
|
+
|
|
248
|
+
// src/storage/postgresql/pg/connections/pool.ts
|
|
249
|
+
var pgNativePool = (options) => {
|
|
250
|
+
const { connectionString, database } = options;
|
|
251
|
+
const pool = getPgPool({ connectionString, database });
|
|
252
|
+
const getConnection = () => pgConnection({
|
|
253
|
+
type: "PoolClient",
|
|
254
|
+
connect: async () => {
|
|
255
|
+
const client = await pool.connect();
|
|
256
|
+
setPgTypeParser(client, {
|
|
257
|
+
parseBigInts: true,
|
|
258
|
+
serializer: options.serializer
|
|
259
|
+
});
|
|
260
|
+
return client;
|
|
261
|
+
},
|
|
262
|
+
close: (client) => Promise.resolve(client.release()),
|
|
263
|
+
serializer: options.serializer
|
|
264
|
+
});
|
|
265
|
+
const open = () => Promise.resolve(getConnection());
|
|
266
|
+
const close = () => endPgPool({ connectionString, database });
|
|
267
|
+
return _chunkXOQHZABKcjs.createConnectionPool.call(void 0, {
|
|
268
|
+
driverType: PgDriverType,
|
|
269
|
+
connection: open,
|
|
270
|
+
close,
|
|
271
|
+
getConnection
|
|
272
|
+
});
|
|
273
|
+
};
|
|
274
|
+
var pgAmbientNativePool = (options) => {
|
|
275
|
+
const { pool } = options;
|
|
276
|
+
return _chunkXOQHZABKcjs.createConnectionPool.call(void 0, {
|
|
277
|
+
driverType: PgDriverType,
|
|
278
|
+
getConnection: () => pgConnection({
|
|
279
|
+
type: "PoolClient",
|
|
280
|
+
connect: () => pool.connect(),
|
|
281
|
+
close: (client) => Promise.resolve(client.release()),
|
|
282
|
+
serializer: options.serializer
|
|
283
|
+
})
|
|
284
|
+
});
|
|
285
|
+
};
|
|
286
|
+
var pgAmbientConnectionPool = (options) => {
|
|
287
|
+
const { connection } = options;
|
|
288
|
+
return _chunkXOQHZABKcjs.createAmbientConnectionPool.call(void 0, {
|
|
289
|
+
driverType: PgDriverType,
|
|
290
|
+
connection
|
|
291
|
+
});
|
|
292
|
+
};
|
|
293
|
+
var pgClientPool = (options) => {
|
|
294
|
+
const { connectionString, database } = options;
|
|
295
|
+
return _chunkXOQHZABKcjs.createConnectionPool.call(void 0, {
|
|
296
|
+
driverType: PgDriverType,
|
|
297
|
+
getConnection: () => {
|
|
298
|
+
const connect = async () => {
|
|
299
|
+
const client = new _pg2.default.Client({ connectionString, database });
|
|
300
|
+
setPgTypeParser(client, {
|
|
301
|
+
parseBigInts: true,
|
|
302
|
+
serializer: options.serializer
|
|
303
|
+
});
|
|
304
|
+
await client.connect();
|
|
305
|
+
return client;
|
|
306
|
+
};
|
|
307
|
+
return pgConnection({
|
|
308
|
+
type: "Client",
|
|
309
|
+
connect,
|
|
310
|
+
close: (client) => client.end(),
|
|
311
|
+
serializer: options.serializer
|
|
312
|
+
});
|
|
313
|
+
}
|
|
314
|
+
});
|
|
315
|
+
};
|
|
316
|
+
var pgAmbientClientPool = (options) => {
|
|
317
|
+
const { client } = options;
|
|
318
|
+
const getConnection = () => {
|
|
319
|
+
const connect = () => Promise.resolve(client);
|
|
320
|
+
return pgConnection({
|
|
321
|
+
type: "Client",
|
|
322
|
+
connect,
|
|
323
|
+
close: () => Promise.resolve(),
|
|
324
|
+
serializer: options.serializer
|
|
325
|
+
});
|
|
326
|
+
};
|
|
327
|
+
const open = () => Promise.resolve(getConnection());
|
|
328
|
+
const close = () => Promise.resolve();
|
|
329
|
+
return _chunkXOQHZABKcjs.createConnectionPool.call(void 0, {
|
|
330
|
+
driverType: PgDriverType,
|
|
331
|
+
connection: open,
|
|
332
|
+
close,
|
|
333
|
+
getConnection
|
|
334
|
+
});
|
|
335
|
+
};
|
|
336
|
+
function pgPool(options) {
|
|
337
|
+
const { connectionString, database } = options;
|
|
338
|
+
const serializer = _nullishCoalesce(_optionalChain([options, 'access', _16 => _16.serialization, 'optionalAccess', _17 => _17.serializer]), () => ( _chunkXOQHZABKcjs.JSONSerializer));
|
|
339
|
+
if ("client" in options && options.client)
|
|
340
|
+
return pgAmbientClientPool({ client: options.client, serializer });
|
|
341
|
+
if ("connection" in options && options.connection)
|
|
342
|
+
return pgAmbientConnectionPool({
|
|
343
|
+
connection: options.connection
|
|
344
|
+
});
|
|
345
|
+
if ("pooled" in options && options.pooled === false)
|
|
346
|
+
return pgClientPool({ connectionString, database, serializer });
|
|
347
|
+
if ("pool" in options && options.pool)
|
|
348
|
+
return pgAmbientNativePool({ pool: options.pool, serializer });
|
|
349
|
+
return pgNativePool({
|
|
350
|
+
connectionString,
|
|
351
|
+
database,
|
|
352
|
+
serializer
|
|
353
|
+
});
|
|
354
|
+
}
|
|
355
|
+
var pools = /* @__PURE__ */ new Map();
|
|
356
|
+
var usageCounter = /* @__PURE__ */ new Map();
|
|
357
|
+
var getPgPool = (connectionStringOrOptions) => {
|
|
358
|
+
const connectionString = typeof connectionStringOrOptions === "string" ? connectionStringOrOptions : connectionStringOrOptions.connectionString;
|
|
359
|
+
const poolOptions = typeof connectionStringOrOptions === "string" ? { connectionString } : connectionStringOrOptions;
|
|
360
|
+
const database = _nullishCoalesce(poolOptions.database, () => ( (poolOptions.connectionString ? _nullishCoalesce(_chunkH7ZKIJHBcjs.parseDatabaseName.call(void 0, poolOptions.connectionString), () => ( _chunkH7ZKIJHBcjs.defaultPostgreSqlDatabase)) : void 0)));
|
|
361
|
+
const lookupKey = key(connectionString, database);
|
|
362
|
+
updatePoolUsageCounter(lookupKey, 1);
|
|
363
|
+
return _nullishCoalesce(pools.get(lookupKey), () => ( pools.set(lookupKey, new _pg2.default.Pool(poolOptions)).get(lookupKey)));
|
|
364
|
+
};
|
|
365
|
+
var endPgPool = async ({
|
|
366
|
+
connectionString,
|
|
367
|
+
database,
|
|
368
|
+
force
|
|
369
|
+
}) => {
|
|
370
|
+
database = _nullishCoalesce(_nullishCoalesce(database, () => ( _chunkH7ZKIJHBcjs.parseDatabaseName.call(void 0, connectionString))), () => ( void 0));
|
|
371
|
+
const lookupKey = key(connectionString, database);
|
|
372
|
+
const pool = pools.get(lookupKey);
|
|
373
|
+
if (pool && (updatePoolUsageCounter(lookupKey, -1) <= 0 || force === true)) {
|
|
374
|
+
await onEndPool(lookupKey, pool);
|
|
375
|
+
}
|
|
376
|
+
};
|
|
377
|
+
var onEndPool = async (lookupKey, pool) => {
|
|
378
|
+
try {
|
|
379
|
+
await pool.end();
|
|
380
|
+
} catch (error) {
|
|
381
|
+
_chunkXOQHZABKcjs.tracer.error("connection-closing-error", { lookupKey, error });
|
|
382
|
+
}
|
|
383
|
+
pools.delete(lookupKey);
|
|
384
|
+
};
|
|
385
|
+
var endAllPgPools = () => Promise.all(
|
|
386
|
+
[...pools.entries()].map(([lookupKey, pool]) => onEndPool(lookupKey, pool))
|
|
387
|
+
);
|
|
388
|
+
var key = (connectionString, database) => `${connectionString}|${_nullishCoalesce(database, () => ( _chunkH7ZKIJHBcjs.defaultPostgreSqlDatabase))}`;
|
|
389
|
+
var updatePoolUsageCounter = (lookupKey, by) => {
|
|
390
|
+
const currentCounter = _nullishCoalesce(usageCounter.get(lookupKey), () => ( 0));
|
|
391
|
+
const newCounter = currentCounter + by;
|
|
392
|
+
usageCounter.set(lookupKey, currentCounter + by);
|
|
393
|
+
return newCounter;
|
|
394
|
+
};
|
|
38
395
|
|
|
396
|
+
// src/storage/postgresql/pg/index.ts
|
|
397
|
+
var tryParseConnectionString = (connectionString) => {
|
|
398
|
+
try {
|
|
399
|
+
return _chunkH7ZKIJHBcjs.PostgreSQLConnectionString.call(void 0, connectionString);
|
|
400
|
+
} catch (e) {
|
|
401
|
+
return null;
|
|
402
|
+
}
|
|
403
|
+
};
|
|
404
|
+
var pgDumboDriver = {
|
|
405
|
+
driverType: PgDriverType,
|
|
406
|
+
createPool: (options) => pgPool(options),
|
|
407
|
+
sqlFormatter: _chunk7WEBHXZDcjs.pgFormatter,
|
|
408
|
+
defaultMigratorOptions: _chunk7WEBHXZDcjs.DefaultPostgreSQLMigratorOptions,
|
|
409
|
+
canHandle: _chunkXOQHZABKcjs.canHandleDriverWithConnectionString.call(void 0,
|
|
410
|
+
PgDriverType,
|
|
411
|
+
tryParseConnectionString
|
|
412
|
+
),
|
|
413
|
+
databaseMetadata: _chunkH7ZKIJHBcjs.postgreSQLMetadata
|
|
414
|
+
};
|
|
415
|
+
var usePgDumboDriver = () => {
|
|
416
|
+
_chunkXOQHZABKcjs.dumboDatabaseDriverRegistry.register(PgDriverType, pgDumboDriver);
|
|
417
|
+
};
|
|
418
|
+
usePgDumboDriver();
|
|
39
419
|
|
|
40
420
|
|
|
41
421
|
|
|
42
422
|
|
|
43
423
|
|
|
44
424
|
|
|
45
|
-
var _chunkISNF6USXcjs = require('./chunk-ISNF6USX.cjs');
|
|
46
|
-
require('./chunk-OJ34O3Q2.cjs');
|
|
47
425
|
|
|
48
426
|
|
|
49
427
|
|
|
@@ -84,10 +462,5 @@ require('./chunk-OJ34O3Q2.cjs');
|
|
|
84
462
|
|
|
85
463
|
|
|
86
464
|
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
exports.AdvisoryLock = _chunkISNF6USXcjs.AdvisoryLock; exports.DefaultPostgreSQLMigratorOptions = _chunkISNF6USXcjs.DefaultPostgreSQLMigratorOptions; exports.NodePostgresDriverType = _chunkISNF6USXcjs.NodePostgresDriverType; exports.PostgreSQLConnectionString = _chunkISNF6USXcjs.PostgreSQLConnectionString; exports.PostgreSQLDatabaseName = _chunkISNF6USXcjs.PostgreSQLDatabaseName; exports.acquireAdvisoryLock = _chunkISNF6USXcjs.acquireAdvisoryLock; exports.advisoryLock = _chunkISNF6USXcjs.advisoryLock; exports.checkConnection = _chunkISNF6USXcjs.checkConnection; exports.connectionPool = _chunkISNF6USXcjs.connectionPool; exports.databaseDriver = _chunkISNF6USXcjs.pgDatabaseDriver; exports.defaultPostgreSQLConnectionString = _chunkISNF6USXcjs.defaultPostgreSQLConnectionString; exports.defaultPostgreSqlDatabase = _chunkISNF6USXcjs.defaultPostgreSqlDatabase; exports.endAllPools = _chunkISNF6USXcjs.endAllPools; exports.endPool = _chunkISNF6USXcjs.endPool; exports.functionExists = _chunkISNF6USXcjs.functionExists; exports.functionExistsSQL = _chunkISNF6USXcjs.functionExistsSQL; exports.getDatabaseNameOrDefault = _chunkISNF6USXcjs.getDatabaseNameOrDefault; exports.getPool = _chunkISNF6USXcjs.getPool; exports.isNodePostgresClient = _chunkISNF6USXcjs.isNodePostgresClient; exports.isNodePostgresNativePool = _chunkISNF6USXcjs.isNodePostgresNativePool; exports.isNodePostgresPoolClient = _chunkISNF6USXcjs.isNodePostgresPoolClient; exports.nodePostgresAmbientClientPool = _chunkISNF6USXcjs.nodePostgresAmbientClientPool; exports.nodePostgresAmbientConnectionPool = _chunkISNF6USXcjs.nodePostgresAmbientConnectionPool; exports.nodePostgresAmbientNativePool = _chunkISNF6USXcjs.nodePostgresAmbientNativePool; exports.nodePostgresClientConnection = _chunkISNF6USXcjs.nodePostgresClientConnection; exports.nodePostgresClientPool = _chunkISNF6USXcjs.nodePostgresClientPool; exports.nodePostgresConnection = _chunkISNF6USXcjs.nodePostgresConnection; exports.nodePostgresExecute = _chunkISNF6USXcjs.nodePostgresExecute; exports.nodePostgresNativePool = _chunkISNF6USXcjs.nodePostgresNativePool; exports.nodePostgresPool = _chunkISNF6USXcjs.nodePostgresPool; exports.nodePostgresPoolClientConnection = _chunkISNF6USXcjs.nodePostgresPoolClientConnection; exports.nodePostgresSQLExecutor = _chunkISNF6USXcjs.nodePostgresSQLExecutor; exports.nodePostgresTransaction = _chunkISNF6USXcjs.nodePostgresTransaction; exports.onEndPool = _chunkISNF6USXcjs.onEndPool; exports.pgDatabaseDriver = _chunkISNF6USXcjs.pgDatabaseDriver; exports.pgFormatter = _chunkISNF6USXcjs.pgFormatter; exports.postgreSQLColumnProcessors = _chunkISNF6USXcjs.postgreSQLColumnProcessors; exports.postgresPool = _chunkISNF6USXcjs.postgresPool; exports.releaseAdvisoryLock = _chunkISNF6USXcjs.releaseAdvisoryLock; exports.setNodePostgresTypeParser = _chunkISNF6USXcjs.setNodePostgresTypeParser; exports.tableExists = _chunkISNF6USXcjs.tableExists; exports.tableExistsSQL = _chunkISNF6USXcjs.tableExistsSQL; exports.tryAcquireAdvisoryLock = _chunkISNF6USXcjs.tryAcquireAdvisoryLock; exports.usePgDatabaseDriver = _chunkISNF6USXcjs.usePgDatabaseDriver;
|
|
465
|
+
exports.AdvisoryLock = _chunk7WEBHXZDcjs.AdvisoryLock; exports.DefaultPostgreSQLMigratorOptions = _chunk7WEBHXZDcjs.DefaultPostgreSQLMigratorOptions; exports.PgDriverType = PgDriverType; exports.PostgreSQLArrayProcessor = _chunk7WEBHXZDcjs.PostgreSQLArrayProcessor; exports.PostgreSQLConnectionString = _chunkH7ZKIJHBcjs.PostgreSQLConnectionString; exports.PostgreSQLDatabaseName = _chunk7WEBHXZDcjs.PostgreSQLDatabaseName; exports.PostgreSQLExpandSQLInProcessor = _chunk7WEBHXZDcjs.PostgreSQLExpandSQLInProcessor; exports.acquireAdvisoryLock = _chunk7WEBHXZDcjs.acquireAdvisoryLock; exports.advisoryLock = _chunk7WEBHXZDcjs.advisoryLock; exports.checkConnection = checkConnection; exports.defaultPostgreSQLConnectionString = _chunkH7ZKIJHBcjs.defaultPostgreSQLConnectionString; exports.defaultPostgreSqlDatabase = _chunkH7ZKIJHBcjs.defaultPostgreSqlDatabase; exports.endAllPgPools = endAllPgPools; exports.endPgPool = endPgPool; exports.functionExists = _chunkH7ZKIJHBcjs.functionExists; exports.functionExistsSQL = _chunkH7ZKIJHBcjs.functionExistsSQL; exports.getPgPool = getPgPool; exports.isPgClient = isPgClient; exports.isPgNativePool = isPgNativePool; exports.isPgPoolClient = isPgPoolClient; exports.mapPostgresError = _chunk7WEBHXZDcjs.mapPostgresError; exports.onEndPool = onEndPool; exports.parseDatabaseName = _chunkH7ZKIJHBcjs.parseDatabaseName; exports.pgAmbientClientPool = pgAmbientClientPool; exports.pgAmbientConnectionPool = pgAmbientConnectionPool; exports.pgAmbientNativePool = pgAmbientNativePool; exports.pgClientConnection = pgClientConnection; exports.pgClientPool = pgClientPool; exports.pgConnection = pgConnection; exports.pgDumboDriver = pgDumboDriver; exports.pgExecute = pgExecute; exports.pgFormatter = _chunk7WEBHXZDcjs.pgFormatter; exports.pgNativePool = pgNativePool; exports.pgPool = pgPool; exports.pgPoolClientConnection = pgPoolClientConnection; exports.pgSQLExecutor = pgSQLExecutor; exports.pgTransaction = pgTransaction; exports.postgreSQLColumnProcessors = _chunk7WEBHXZDcjs.postgreSQLColumnProcessors; exports.postgreSQLMetadata = _chunkH7ZKIJHBcjs.postgreSQLMetadata; exports.releaseAdvisoryLock = _chunk7WEBHXZDcjs.releaseAdvisoryLock; exports.setPgTypeParser = setPgTypeParser; exports.tableExists = _chunkH7ZKIJHBcjs.tableExists; exports.tableExistsSQL = _chunkH7ZKIJHBcjs.tableExistsSQL; exports.tryAcquireAdvisoryLock = _chunk7WEBHXZDcjs.tryAcquireAdvisoryLock; exports.usePgDumboDriver = usePgDumboDriver;
|
|
93
466
|
//# sourceMappingURL=pg.cjs.map
|
package/dist/pg.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["/home/runner/work/Pongo/Pongo/src/packages/dumbo/dist/pg.cjs"],"names":[],"mappings":"AAAA;AACE;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACF,wDAA6B;AAC7B,gCAA6B;AAC7B;AACE;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACF,6iGAAC","file":"/home/runner/work/Pongo/Pongo/src/packages/dumbo/dist/pg.cjs"}
|
|
1
|
+
{"version":3,"sources":["/home/runner/work/Pongo/Pongo/src/packages/dumbo/dist/pg.cjs","../src/storage/postgresql/pg/connections/connection.ts","../src/storage/postgresql/pg/execute/execute.ts","../src/storage/postgresql/pg/connections/transaction.ts","../src/storage/postgresql/pg/connections/pool.ts","../src/storage/postgresql/pg/serialization/index.ts","../src/storage/postgresql/pg/index.ts"],"names":[],"mappings":"AAAA;AACE;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACF,wDAA6B;AAC7B;AACE;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACF,wDAA6B;AAC7B;AACE;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACF,wDAA6B;AAC7B;AACA;ACtCA,gEAAe;ADwCf;AACA;AEzCA;AAiBO,IAAM,eAAA,EAAiB,CAC5B,YAAA,EAAA,GAC4B;AAC5B,EAAA,OAAO,aAAA,WAAwB,YAAA,CAAG,IAAA;AACpC,CAAA;AAEO,IAAM,WAAA,EAAa,CACxB,YAAA,EAAA,GAC8B,aAAA,WAAwB,YAAA,CAAG,MAAA;AAEpD,IAAM,eAAA,EAAiB,CAC5B,YAAA,EAAA,GAEA,UAAA,GAAa,aAAA,GAAgB,OAAO,YAAA,CAAa,QAAA,IAAY,UAAA;AAExD,IAAM,UAAA,EAAY,MAAA,CACvB,YAAA,EACA,MAAA,EAAA,GACG;AACH,EAAA,MAAM,OAAA,EAAS,cAAA,CAAe,YAAY,EAAA,EACtC,MAAM,YAAA,CAAa,OAAA,CAAQ,EAAA,EAC3B,YAAA;AAEJ,EAAA,IAAI;AACF,IAAA,OAAO,MAAM,MAAA,CAAO,MAAM,CAAA;AAAA,EAC5B,EAAA,QAAE;AAEA,IAAA,GAAA,CAAI,cAAA,CAAe,YAAY,EAAA,GAAK,cAAA,CAAe,MAAM,CAAA;AACvD,MAAA,MAAA,CAAO,OAAA,CAAQ,CAAA;AAAA,EACnB;AACF,CAAA;AAIO,IAAM,cAAA,EAAgB,CAAC;AAAA,EAC5B;AACF,CAAA,EAAA,GAAA,CAEsB;AAAA,EACpB,UAAA,EAAY,YAAA;AAAA,EACZ,KAAA,EAAO,MAAA,CACL,MAAA,EACA,GAAA,EACA,OAAA,EAAA,GACG;AACH,IAAA,MAAM,QAAA,EAAU,MAAM,UAAA;AAAA,MACpB,MAAA;AAAA,MACA,CAAC,GAAG,CAAA;AAAA,MACJ,UAAA;AAAA,MACA;AAAA,IACF,CAAA;AACA,IAAA,OAAO,OAAA,CAAQ,CAAC,CAAA;AAAA,EAClB,CAAA;AAAA,EACA,UAAA,EAAY,CACV,MAAA,EACA,IAAA,EACA,OAAA,EAAA,GACG,UAAA,CAAmB,MAAA,EAAQ,IAAA,EAAM,UAAA,EAAY,OAAO,CAAA;AAAA,EACzD,OAAA,EAAS,MAAA,CACP,MAAA,EACA,GAAA,EACA,OAAA,EAAA,GACG;AACH,IAAA,MAAM,QAAA,EAAU,MAAM,YAAA;AAAA,MACpB,MAAA;AAAA,MACA,CAAC,GAAG,CAAA;AAAA,MACJ,UAAA;AAAA,MACA;AAAA,IACF,CAAA;AACA,IAAA,OAAO,OAAA,CAAQ,CAAC,CAAA;AAAA,EAClB,CAAA;AAAA,EACA,YAAA,EAAc,CACZ,MAAA,EACA,IAAA,EACA,OAAA,EAAA,GACG,YAAA,CAAqB,MAAA,EAAQ,IAAA,EAAM,UAAA,EAAY,OAAO,CAAA;AAAA,EAC3D,SAAA,EAAW;AACb,CAAA,CAAA;AAEA,MAAA,SAAe,UAAA,CACb,MAAA,EACA,IAAA,EACA,UAAA,EACA,OAAA,EACgC;AAChC,EAAA,MAAM,QAAA,EAAiC,KAAA;AAAA,IACrC,IAAA,CAAK;AAAA,EACP,CAAA;AAEA,EAAA,GAAA,iBAAI,OAAA,2BAAS,WAAA,EAAW;AACtB,IAAA,MAAM,MAAA,CAAO,KAAA,CAAM,CAAA,wBAAA,EAA2B,OAAA,CAAQ,SAAS,CAAA,CAAA;AACjE,EAAA;AAGsC,EAAA;AACqB,IAAA;AAC7B,IAAA;AAC1B,MAAA;AACA,MAAA;AACuD,MAAA;AACxD,IAAA;AACG,IAAA;AAGsC,MAAA;AAGlB,MAAA;AACX,QAAA;AACJ,UAAA;AACe,UAAA;AACuB,YAAA;AACzC,UAAA;AACF,QAAA;AACF,MAAA;AAE4D,MAAA;AAC9C,IAAA;AAC4C,MAAA;AAC9B,MAAA;AAC9B,IAAA;AACF,EAAA;AAEO,EAAA;AACT;AAOkC;AACO,EAAA;AAChC,IAAA;AACP,EAAA;AAEwB,EAAA;AACyC,IAAA;AACjE,EAAA;AAGsC,EAAA;AACqB,IAAA;AAC3B,IAAA;AAC5B,MAAA;AACA,MAAA;AACuD,MAAA;AACxD,IAAA;AACG,IAAA;AAGsC,MAAA;AAGlB,MAAA;AACX,QAAA;AACJ,UAAA;AACe,UAAA;AACuB,YAAA;AACzC,UAAA;AACF,QAAA;AACF,MAAA;AAE4D,MAAA;AAEE,MAAA;AACtB,QAAA;AACxC,MAAA;AACc,IAAA;AAC8C,MAAA;AAChC,MAAA;AAC9B,IAAA;AACF,EAAA;AAEO,EAAA;AACT;AF1CoE;AACA;AGtHhE;AAIuB,EAAA;AACX,EAAA;AACO,EAAA;AACI,IAAA;AACC,IAAA;AACO,IAAA;AAC2B,MAAA;AACxD,IAAA;AACuB,IAAA;AACC,MAAA;AACxB,IAAA;AACkC,IAAA;AACpC,EAAA;AACoB,EAAA;AACG,IAAA;AAEjB,IAAA;AACyB,MAAA;AAC3B,IAAA;AAC+C,MAAA;AACjD,IAAA;AACF,EAAA;AACqC,EAAA;AACd,IAAA;AACjB,IAAA;AAC2B,MAAA;AAC7B,IAAA;AACsD,MAAA;AACxD,IAAA;AACF,EAAA;AACoD,EAAA;AACnC,IAAA;AAChB,EAAA;AACH;AHoHkE;AACA;AC/K1B;AAiDjB;AACI,EAAA;AAEH,EAAA;AACV,IAAA;AACZ,IAAA;AACA,IAAA;AAE4B,IAAA;AAClB,IAAA;AACU,IAAA;AACrB,EAAA;AACH;AAI6B;AACA,EAAA;AAEH,EAAA;AACV,IAAA;AACZ,IAAA;AACA,IAAA;AAE4B,IAAA;AAClB,IAAA;AACU,IAAA;AACrB,EAAA;AACH;AAc+C;AAGzC,EAAA;AACN;AAaqC;AACN,EAAA;AAC3B,IAAA;AACD,EAAA;AAEG,EAAA;AACmB,IAAA;AACK,IAAA;AACZ,EAAA;AAIZ,IAAA;AAIK,IAAA;AACO,MAAA;AAIN,MAAA;AAGN,MAAA;AACA,MAAA;AACF,IAAA;AACA,EAAA;AAEiB,IAAA;AACnB,EAAA;AACF;ADkFoE;AACA;AI1OrD;AJ4OqD;AACA;AKpO/D;AAE2B,EAAA;AACiB,IAAA;AAEtB,EAAA;AAGgC,IAAA;AAID,IAAA;AACxD,EAAA;AACF;AL+NoE;AACA;AI1MhD;AACqB,EAAA;AACc,EAAA;AAGtC,EAAA;AACL,IAAA;AACe,IAAA;AACe,MAAA;AAEV,MAAA;AACR,QAAA;AACM,QAAA;AACrB,MAAA;AAEM,MAAA;AACT,IAAA;AACmD,IAAA;AAC/B,IAAA;AACrB,EAAA;AAE+C,EAAA;AACU,EAAA;AAEhC,EAAA;AACd,IAAA;AACA,IAAA;AACZ,IAAA;AACA,IAAA;AACD,EAAA;AACH;AAKoB;AACD,EAAA;AAEW,EAAA;AACd,IAAA;AAEG,IAAA;AACL,MAAA;AACsB,MAAA;AACuB,MAAA;AAC/B,MAAA;AACrB,IAAA;AACJ,EAAA;AACH;AAI+B;AACN,EAAA;AAEY,EAAA;AACrB,IAAA;AACZ,IAAA;AACD,EAAA;AACH;AAM2B;AACc,EAAA;AAEX,EAAA;AACd,IAAA;AACS,IAAA;AACS,MAAA;AACiC,QAAA;AAEnC,QAAA;AACR,UAAA;AACM,UAAA;AACrB,QAAA;AAEoB,QAAA;AACd,QAAA;AACT,MAAA;AAEoB,MAAA;AACZ,QAAA;AACN,QAAA;AAC8B,QAAA;AACV,QAAA;AACrB,MAAA;AACH,IAAA;AACD,EAAA;AACH;AAK2B;AACN,EAAA;AAES,EAAA;AACkB,IAAA;AAExB,IAAA;AACZ,MAAA;AACN,MAAA;AAC6B,MAAA;AACT,MAAA;AACrB,IAAA;AACH,EAAA;AAEkD,EAAA;AACd,EAAA;AAER,EAAA;AACd,IAAA;AACA,IAAA;AACZ,IAAA;AACA,IAAA;AACD,EAAA;AACH;AAuDgE;AACvB,EAAA;AAEiB,EAAA;AAErB,EAAA;AAC+B,IAAA;AAE3B,EAAA;AACN,IAAA;AACT,MAAA;AACrB,IAAA;AAE2C,EAAA;AACkB,IAAA;AAE/B,EAAA;AAC8B,IAAA;AAE3C,EAAA;AAClB,IAAA;AACA,IAAA;AACA,IAAA;AACD,EAAA;AACH;AAE4C;AACM;AAIpC;AAE2B,EAAA;AAKA,EAAA;AAMxB,EAAA;AAKiC,EAAA;AAEb,EAAA;AAIA,EAAA;AAErC;AAEgC;AAC9B,EAAA;AACA,EAAA;AACA,EAAA;AAKmB;AAC2C,EAAA;AACd,EAAA;AAEhB,EAAA;AAC2B,EAAA;AAC1B,IAAA;AACjC,EAAA;AACF;AAEqE;AAC/D,EAAA;AACa,IAAA;AACD,EAAA;AAC+C,IAAA;AAC/D,EAAA;AACsB,EAAA;AACxB;AAGU;AACoD,EAAA;AAC5D;AAGuB;AAEiD;AAClB,EAAA;AAClB,EAAA;AAEW,EAAA;AAExC,EAAA;AACT;AJ0EoE;AACA;AMxXL;AACzD,EAAA;AACgD,IAAA;AAC5C,EAAA;AACC,IAAA;AACT,EAAA;AACF;AAMI;AACU,EAAA;AAC4C,EAAA;AAC1C,EAAA;AACU,EAAA;AACb,EAAA;AACT,IAAA;AACA,IAAA;AACF,EAAA;AACkB,EAAA;AACpB;AAEsC;AAC4B,EAAA;AAClE;AAEiB;ANmXmD;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA","file":"/home/runner/work/Pongo/Pongo/src/packages/dumbo/dist/pg.cjs","sourcesContent":[null,"import pg from 'pg';\nimport type { JSONSerializer } from '../../../../core';\nimport {\n createConnection,\n type Connection,\n type DatabaseTransaction,\n} from '../../../../core';\nimport type { PostgreSQLDriverType } from '../../core';\nimport { pgSQLExecutor } from '../execute';\nimport { pgTransaction, type PgTransactionOptions } from './transaction';\n\nexport type PgDriverType = PostgreSQLDriverType<'pg'>;\nexport const PgDriverType: PgDriverType = 'PostgreSQL:pg';\n\nexport type PgPoolClient = pg.PoolClient;\nexport type PgClient = pg.Client;\n\nexport type PgClientOrPoolClient = PgPoolClient | PgClient;\n\nexport type PgPoolOrClient = pg.Pool | PgPoolClient | PgClient;\n\nexport type PgClientConnection = Connection<\n PgClientConnection,\n PgDriverType,\n PgClient,\n DatabaseTransaction<PgClientConnection>,\n PgTransactionOptions\n>;\n\nexport type PgPoolClientConnection = Connection<\n PgPoolClientConnection,\n PgDriverType,\n PgPoolClient,\n DatabaseTransaction<PgPoolClientConnection>,\n PgTransactionOptions\n>;\n\nexport type PgConnection = PgPoolClientConnection | PgClientConnection;\n\nexport type PgPoolClientOptions = {\n type: 'PoolClient';\n connect: () => Promise<PgPoolClient>;\n close: (client: PgPoolClient) => Promise<void>;\n};\n\nexport type PgClientOptions = {\n type: 'Client';\n connect: () => Promise<PgClient>;\n close: (client: PgClient) => Promise<void>;\n};\n\nexport type PgClientConnectionOptions = PgClientOptions & {\n serializer: JSONSerializer;\n};\n\nexport type PgPoolClientConnectionOptions = PgPoolClientOptions & {\n serializer: JSONSerializer;\n};\n\nexport const pgClientConnection = (\n options: PgClientConnectionOptions,\n): PgClientConnection => {\n const { connect, close } = options;\n\n return createConnection({\n driverType: PgDriverType,\n connect,\n close,\n initTransaction: (connection) =>\n pgTransaction(connection, options.serializer),\n executor: pgSQLExecutor,\n serializer: options.serializer,\n });\n};\n\nexport const pgPoolClientConnection = (\n options: PgPoolClientConnectionOptions,\n): PgPoolClientConnection => {\n const { connect, close } = options;\n\n return createConnection({\n driverType: PgDriverType,\n connect,\n close,\n initTransaction: (connection) =>\n pgTransaction(connection, options.serializer),\n executor: pgSQLExecutor,\n serializer: options.serializer,\n });\n};\n\nexport type PgConnectionOptions =\n | PgPoolClientConnectionOptions\n | PgClientConnectionOptions;\n\nexport function pgConnection(\n options: PgPoolClientConnectionOptions,\n): PgPoolClientConnection;\nexport function pgConnection(\n options: PgClientConnectionOptions,\n): PgClientConnection;\nexport function pgConnection(\n options: PgPoolClientConnectionOptions | PgClientConnectionOptions,\n): PgPoolClientConnection | PgClientConnection {\n return options.type === 'Client'\n ? pgClientConnection(options)\n : pgPoolClientConnection(options);\n}\n\nexport type ConnectionCheckResult =\n | { successful: true }\n | {\n successful: false;\n code: string | undefined;\n errorType: 'ConnectionRefused' | 'Authentication' | 'Unknown';\n error: unknown;\n };\n\nexport const checkConnection = async (\n connectionString: string,\n): Promise<ConnectionCheckResult> => {\n const client = new pg.Client({\n connectionString,\n });\n\n try {\n await client.connect();\n return { successful: true };\n } catch (error) {\n const code =\n error instanceof Error &&\n 'code' in error &&\n typeof error.code === 'string'\n ? error.code\n : undefined;\n\n return {\n successful: false,\n errorType:\n code === 'ECONNREFUSED'\n ? 'ConnectionRefused'\n : code === '28P01'\n ? 'Authentication'\n : 'Unknown',\n code,\n error,\n };\n } finally {\n // Ensure the client is closed properly if connected\n await client.end();\n }\n};\n","import pg from 'pg';\nimport type { JSONSerializer } from '../../../../core';\nimport {\n BatchCommandNoChangesError,\n mapSQLQueryResult,\n tracer,\n type BatchSQLCommandOptions,\n type DbSQLExecutor,\n type QueryResult,\n type QueryResultRow,\n type SQL,\n type SQLQueryOptions,\n} from '../../../../core';\nimport { pgFormatter } from '../../core';\nimport { mapPostgresError } from '../../core/errors/errorMapper';\nimport { PgDriverType, type PgClientOrPoolClient } from '../connections';\n\nexport const isPgNativePool = (\n poolOrClient: pg.Pool | pg.PoolClient | pg.Client,\n): poolOrClient is pg.Pool => {\n return poolOrClient instanceof pg.Pool;\n};\n\nexport const isPgClient = (\n poolOrClient: pg.Pool | pg.PoolClient | pg.Client,\n): poolOrClient is pg.Client => poolOrClient instanceof pg.Client;\n\nexport const isPgPoolClient = (\n poolOrClient: pg.Pool | pg.PoolClient | pg.Client,\n): poolOrClient is pg.PoolClient =>\n 'release' in poolOrClient && typeof poolOrClient.release === 'function';\n\nexport const pgExecute = async <Result = void>(\n poolOrClient: pg.Pool | pg.PoolClient | pg.Client,\n handle: (client: pg.PoolClient | pg.Client) => Promise<Result>,\n) => {\n const client = isPgNativePool(poolOrClient)\n ? await poolOrClient.connect()\n : poolOrClient;\n\n try {\n return await handle(client);\n } finally {\n // release only if client wasn't injected externally\n if (isPgNativePool(poolOrClient) && isPgPoolClient(client))\n client.release();\n }\n};\n\nexport type PgSQLExecutor = DbSQLExecutor<PgDriverType, PgClientOrPoolClient>;\n\nexport const pgSQLExecutor = ({\n serializer,\n}: {\n serializer: JSONSerializer;\n}): PgSQLExecutor => ({\n driverType: PgDriverType,\n query: async <Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sql: SQL,\n options: SQLQueryOptions | undefined,\n ) => {\n const results = await batchQuery<Result>(\n client,\n [sql],\n serializer,\n options,\n );\n return results[0]!;\n },\n batchQuery: <Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sqls: SQL[],\n options: SQLQueryOptions | undefined,\n ) => batchQuery<Result>(client, sqls, serializer, options),\n command: async <Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sql: SQL,\n options: BatchSQLCommandOptions | undefined,\n ) => {\n const results = await batchCommand<Result>(\n client,\n [sql],\n serializer,\n options,\n );\n return results[0]!;\n },\n batchCommand: <Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sqls: SQL[],\n options: BatchSQLCommandOptions | undefined,\n ) => batchCommand<Result>(client, sqls, serializer, options),\n formatter: pgFormatter,\n});\n\nasync function batchQuery<Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sqls: SQL[],\n serializer: JSONSerializer,\n options?: SQLQueryOptions,\n): Promise<QueryResult<Result>[]> {\n const results: QueryResult<Result>[] = Array<QueryResult<Result>>(\n sqls.length,\n );\n\n if (options?.timeoutMs) {\n await client.query(`SET statement_timeout = ${options.timeoutMs}`);\n }\n\n //TODO: make it smarter at some point\n for (let i = 0; i < sqls.length; i++) {\n const { query, params } = pgFormatter.format(sqls[i]!, { serializer });\n tracer.info('db:sql:query', {\n query,\n params,\n debugSQL: pgFormatter.describe(sqls[i]!, { serializer }),\n });\n try {\n let result =\n params.length > 0\n ? await client.query<Result>(query, params)\n : await client.query<Result>(query);\n\n if (options?.mapping) {\n result = {\n ...result,\n rows: result.rows.map((row) =>\n mapSQLQueryResult(row, options.mapping!),\n ),\n };\n }\n\n results[i] = { rowCount: result.rowCount, rows: result.rows };\n } catch (error) {\n tracer.error('db:sql:batch_query:execute:error', { error });\n throw mapPostgresError(error);\n }\n }\n\n return results;\n}\n\nasync function batchCommand<Result extends QueryResultRow = QueryResultRow>(\n client: PgClientOrPoolClient,\n sqls: SQL[],\n serializer: JSONSerializer,\n options?: BatchSQLCommandOptions,\n): Promise<QueryResult<Result>[]> {\n const results: QueryResult<Result>[] = Array<QueryResult<Result>>(\n sqls.length,\n );\n\n if (options?.timeoutMs) {\n await client.query(`SET statement_timeout = ${options.timeoutMs}`);\n }\n\n //TODO: make it smarter at some point\n for (let i = 0; i < sqls.length; i++) {\n const { query, params } = pgFormatter.format(sqls[i]!, { serializer });\n tracer.info('db:sql:command', {\n query,\n params,\n debugSQL: pgFormatter.describe(sqls[i]!, { serializer }),\n });\n try {\n let result =\n params.length > 0\n ? await client.query<Result>(query, params)\n : await client.query<Result>(query);\n\n if (options?.mapping) {\n result = {\n ...result,\n rows: result.rows.map((row) =>\n mapSQLQueryResult(row, options.mapping!),\n ),\n };\n }\n\n results[i] = { rowCount: result.rowCount, rows: result.rows };\n\n if (options?.assertChanges && (results[i]!.rowCount ?? 0) === 0) {\n throw new BatchCommandNoChangesError(i);\n }\n } catch (error) {\n tracer.error('db:sql:batch_command:execute:error', { error });\n throw mapPostgresError(error);\n }\n }\n\n return results;\n}\n","import type { JSONSerializer } from '../../../../core';\nimport {\n sqlExecutor,\n type AnyConnection,\n type DatabaseTransaction,\n type DatabaseTransactionOptions,\n} from '../../../../core';\nimport { pgSQLExecutor } from '../execute';\nimport {\n PgDriverType,\n type PgConnection,\n type PgPoolOrClient,\n} from './connection';\n\nexport type PgTransaction = DatabaseTransaction<PgConnection>;\n\nexport type PgIsolationLevel =\n | 'READ UNCOMMITTED'\n | 'READ COMMITTED'\n | 'REPEATABLE READ'\n | 'SERIALIZABLE';\n\nexport type PgTransactionOptions = DatabaseTransactionOptions & {\n isolationLevel?: PgIsolationLevel;\n};\n\nexport const pgTransaction =\n <ConnectionType extends AnyConnection = AnyConnection>(\n connection: () => ConnectionType,\n serializer: JSONSerializer,\n ) =>\n <DbClient extends PgPoolOrClient = PgPoolOrClient>(\n getClient: Promise<DbClient>,\n options?: {\n close: (client: DbClient, error?: unknown) => Promise<void>;\n } & PgTransactionOptions,\n ): DatabaseTransaction<ConnectionType> => ({\n connection: connection(),\n driverType: PgDriverType,\n begin: async () => {\n const client = await getClient;\n const parts = ['BEGIN'];\n if (options?.isolationLevel) {\n parts.push(`ISOLATION LEVEL ${options.isolationLevel}`);\n }\n if (options?.readonly) {\n parts.push('READ ONLY');\n }\n await client.query(parts.join(' '));\n },\n commit: async () => {\n const client = await getClient;\n\n try {\n await client.query('COMMIT');\n } finally {\n if (options?.close) await options?.close(client);\n }\n },\n rollback: async (error?: unknown) => {\n const client = await getClient;\n try {\n await client.query('ROLLBACK');\n } finally {\n if (options?.close) await options?.close(client, error);\n }\n },\n execute: sqlExecutor(pgSQLExecutor({ serializer }), {\n connect: () => getClient,\n }),\n });\n","import pg from 'pg';\nimport {\n createAmbientConnectionPool,\n createConnectionPool,\n JSONSerializer,\n tracer,\n type ConnectionPool,\n type InferTransactionFromConnection,\n type InferTransactionOptionsFromConnection,\n type JSONSerializationOptions,\n} from '../../../../core';\nimport { defaultPostgreSqlDatabase, parseDatabaseName } from '../../core';\nimport { setPgTypeParser } from '../serialization';\nimport {\n pgConnection,\n PgDriverType,\n type PgClientConnection,\n type PgPoolClientConnection,\n} from './connection';\n\nexport type PgNativePool = ConnectionPool<\n PgPoolClientConnection,\n InferTransactionFromConnection<PgPoolClientConnection>,\n InferTransactionOptionsFromConnection<PgPoolClientConnection>\n>;\n\nexport type PgAmbientClientPool = ConnectionPool<\n PgClientConnection,\n InferTransactionFromConnection<PgClientConnection>,\n InferTransactionOptionsFromConnection<PgClientConnection>\n>;\n\nexport type PgAmbientConnectionPool = ConnectionPool<\n PgPoolClientConnection | PgClientConnection\n>;\n\nexport type PgPool =\n | PgNativePool\n | PgAmbientClientPool\n | PgAmbientConnectionPool;\n\nexport const pgNativePool = (options: {\n connectionString: string;\n database?: string | undefined;\n serializer: JSONSerializer;\n}): PgNativePool => {\n const { connectionString, database } = options;\n const pool = getPgPool({ connectionString, database });\n\n const getConnection = () =>\n pgConnection({\n type: 'PoolClient',\n connect: async () => {\n const client = await pool.connect();\n\n setPgTypeParser(client, {\n parseBigInts: true,\n serializer: options.serializer,\n });\n\n return client;\n },\n close: (client) => Promise.resolve(client.release()),\n serializer: options.serializer,\n });\n\n const open = () => Promise.resolve(getConnection());\n const close = () => endPgPool({ connectionString, database });\n\n return createConnectionPool({\n driverType: PgDriverType,\n connection: open,\n close,\n getConnection,\n });\n};\n\nexport const pgAmbientNativePool = (options: {\n pool: pg.Pool;\n serializer: JSONSerializer;\n}): PgNativePool => {\n const { pool } = options;\n\n return createConnectionPool({\n driverType: PgDriverType,\n getConnection: () =>\n pgConnection({\n type: 'PoolClient',\n connect: () => pool.connect(),\n close: (client) => Promise.resolve(client.release()),\n serializer: options.serializer,\n }),\n });\n};\n\nexport const pgAmbientConnectionPool = (options: {\n connection: PgPoolClientConnection | PgClientConnection;\n}): PgAmbientConnectionPool => {\n const { connection } = options;\n\n return createAmbientConnectionPool({\n driverType: PgDriverType,\n connection,\n });\n};\n\nexport const pgClientPool = (options: {\n connectionString: string;\n database?: string | undefined;\n serializer: JSONSerializer;\n}): PgAmbientClientPool => {\n const { connectionString, database } = options;\n\n return createConnectionPool({\n driverType: PgDriverType,\n getConnection: () => {\n const connect = async () => {\n const client = new pg.Client({ connectionString, database });\n\n setPgTypeParser(client, {\n parseBigInts: true,\n serializer: options.serializer,\n });\n\n await client.connect();\n return client;\n };\n\n return pgConnection({\n type: 'Client',\n connect,\n close: (client) => client.end(),\n serializer: options.serializer,\n });\n },\n });\n};\n\nexport const pgAmbientClientPool = (options: {\n client: pg.Client;\n serializer: JSONSerializer;\n}): PgAmbientClientPool => {\n const { client } = options;\n\n const getConnection = () => {\n const connect = () => Promise.resolve(client);\n\n return pgConnection({\n type: 'Client',\n connect,\n close: () => Promise.resolve(),\n serializer: options.serializer,\n });\n };\n\n const open = () => Promise.resolve(getConnection());\n const close = () => Promise.resolve();\n\n return createConnectionPool({\n driverType: PgDriverType,\n connection: open,\n close,\n getConnection,\n });\n};\n\nexport type PgPoolPooledOptions =\n | {\n connectionString: string;\n database?: string;\n pooled: true;\n pool: pg.Pool;\n }\n | {\n connectionString: string;\n database?: string;\n pool: pg.Pool;\n }\n | {\n connectionString: string;\n database?: string;\n pooled: true;\n }\n | {\n connectionString: string;\n database?: string;\n };\n\nexport type PgPoolNotPooledOptions =\n | {\n connectionString: string;\n database?: string;\n pooled: false;\n client: pg.Client;\n }\n | {\n connectionString: string;\n database?: string;\n client: pg.Client;\n }\n | {\n connectionString: string;\n database?: string;\n pooled: false;\n }\n | {\n connectionString: string;\n database?: string;\n connection: PgPoolClientConnection | PgClientConnection;\n pooled?: false;\n };\n\nexport type PgPoolOptions = (PgPoolPooledOptions | PgPoolNotPooledOptions) &\n JSONSerializationOptions;\n\nexport function pgPool(options: PgPoolPooledOptions): PgNativePool;\nexport function pgPool(options: PgPoolNotPooledOptions): PgAmbientClientPool;\nexport function pgPool(\n options: PgPoolOptions,\n): PgNativePool | PgAmbientClientPool | PgAmbientConnectionPool {\n const { connectionString, database } = options;\n\n const serializer = options.serialization?.serializer ?? JSONSerializer;\n\n if ('client' in options && options.client)\n return pgAmbientClientPool({ client: options.client, serializer });\n\n if ('connection' in options && options.connection)\n return pgAmbientConnectionPool({\n connection: options.connection,\n });\n\n if ('pooled' in options && options.pooled === false)\n return pgClientPool({ connectionString, database, serializer });\n\n if ('pool' in options && options.pool)\n return pgAmbientNativePool({ pool: options.pool, serializer });\n\n return pgNativePool({\n connectionString,\n database,\n serializer,\n });\n}\n\nconst pools: Map<string, pg.Pool> = new Map();\nconst usageCounter: Map<string, number> = new Map();\n\nexport const getPgPool = (\n connectionStringOrOptions: string | pg.PoolConfig,\n): pg.Pool => {\n const connectionString =\n typeof connectionStringOrOptions === 'string'\n ? connectionStringOrOptions\n : connectionStringOrOptions.connectionString!;\n\n const poolOptions =\n typeof connectionStringOrOptions === 'string'\n ? { connectionString }\n : connectionStringOrOptions;\n\n const database =\n poolOptions.database ??\n (poolOptions.connectionString\n ? (parseDatabaseName(poolOptions.connectionString) ??\n defaultPostgreSqlDatabase)\n : undefined);\n\n const lookupKey = key(connectionString, database);\n\n updatePoolUsageCounter(lookupKey, 1);\n\n return (\n pools.get(lookupKey) ??\n pools.set(lookupKey, new pg.Pool(poolOptions)).get(lookupKey)!\n );\n};\n\nexport const endPgPool = async ({\n connectionString,\n database,\n force,\n}: {\n connectionString: string;\n database?: string | undefined;\n force?: boolean;\n}): Promise<void> => {\n database = database ?? parseDatabaseName(connectionString) ?? undefined;\n const lookupKey = key(connectionString, database);\n\n const pool = pools.get(lookupKey);\n if (pool && (updatePoolUsageCounter(lookupKey, -1) <= 0 || force === true)) {\n await onEndPool(lookupKey, pool);\n }\n};\n\nexport const onEndPool = async (lookupKey: string, pool: pg.Pool) => {\n try {\n await pool.end();\n } catch (error) {\n tracer.error('connection-closing-error', { lookupKey, error });\n }\n pools.delete(lookupKey);\n};\n\nexport const endAllPgPools = () =>\n Promise.all(\n [...pools.entries()].map(([lookupKey, pool]) => onEndPool(lookupKey, pool)),\n );\n\nconst key = (connectionString: string, database: string | undefined) =>\n `${connectionString}|${database ?? defaultPostgreSqlDatabase}`;\n\nconst updatePoolUsageCounter = (lookupKey: string, by: 1 | -1): number => {\n const currentCounter = usageCounter.get(lookupKey) ?? 0;\n const newCounter = currentCounter + by;\n\n usageCounter.set(lookupKey, currentCounter + by);\n\n return newCounter;\n};\n","import type pg from 'pg';\nimport type { JSONSerializer } from '../../../../core/serializer';\n\nexport const setPgTypeParser = (\n client: pg.Client | pg.PoolClient,\n options?: {\n parseBigInts?: boolean;\n serializer: JSONSerializer;\n },\n) => {\n // BigInt\n if (options?.parseBigInts === true)\n client.setTypeParser(20, (val) => BigInt(val));\n\n if (options?.serializer) {\n // JSONB\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n client.setTypeParser(3802, (val) => options.serializer.deserialize(val));\n\n // JSON\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n client.setTypeParser(114, (val) => options.serializer.deserialize(val));\n }\n};\n","import {\n canHandleDriverWithConnectionString,\n type DumboDatabaseDriver,\n dumboDatabaseDriverRegistry,\n} from '../../../core';\nimport {\n DefaultPostgreSQLMigratorOptions,\n pgFormatter,\n PostgreSQLConnectionString,\n postgreSQLMetadata,\n} from '../core';\nimport {\n type PgConnection,\n PgDriverType,\n pgPool,\n type PgPool,\n type PgPoolOptions,\n} from './connections';\n\nconst tryParseConnectionString = (connectionString: string) => {\n try {\n return PostgreSQLConnectionString(connectionString);\n } catch {\n return null;\n }\n};\n\nexport const pgDumboDriver: DumboDatabaseDriver<\n PgConnection,\n PgPoolOptions,\n PgPool\n> = {\n driverType: PgDriverType,\n createPool: (options) => pgPool(options as PgPoolOptions),\n sqlFormatter: pgFormatter,\n defaultMigratorOptions: DefaultPostgreSQLMigratorOptions,\n canHandle: canHandleDriverWithConnectionString(\n PgDriverType,\n tryParseConnectionString,\n ),\n databaseMetadata: postgreSQLMetadata,\n};\n\nexport const usePgDumboDriver = () => {\n dumboDatabaseDriverRegistry.register(PgDriverType, pgDumboDriver);\n};\n\nusePgDumboDriver();\n\nexport type PgDumboOptions = PgPoolOptions;\nexport * from './connections';\nexport * from './execute';\nexport * from './serialization';\n"]}
|