@stonyx/orm 0.2.1-beta.80 → 0.2.1-beta.82

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -19,6 +19,13 @@ const {
19
19
  MYSQL_DATABASE,
20
20
  MYSQL_CONNECTION_LIMIT,
21
21
  MYSQL_MIGRATIONS_DIR,
22
+ PG_HOST,
23
+ PG_PORT,
24
+ PG_USER,
25
+ PG_PASSWORD,
26
+ PG_DATABASE,
27
+ PG_CONNECTION_LIMIT,
28
+ PG_MIGRATIONS_DIR,
22
29
  } = process.env;
23
30
 
24
31
  export default {
@@ -50,6 +57,16 @@ export default {
50
57
  migrationsDir: MYSQL_MIGRATIONS_DIR ?? 'migrations',
51
58
  migrationsTable: '__migrations',
52
59
  } : undefined,
60
+ postgres: PG_HOST ? {
61
+ host: PG_HOST ?? 'localhost',
62
+ port: parseInt(PG_PORT ?? '5432'),
63
+ user: PG_USER ?? 'postgres',
64
+ password: PG_PASSWORD ?? '',
65
+ database: PG_DATABASE ?? 'stonyx',
66
+ connectionLimit: parseInt(PG_CONNECTION_LIMIT ?? '10'),
67
+ migrationsDir: PG_MIGRATIONS_DIR ?? 'migrations',
68
+ migrationsTable: '__migrations',
69
+ } : undefined,
53
70
  restServer: {
54
71
  enabled: ORM_USE_REST_SERVER ?? 'true', // Whether to load restServer for automatic route setup or
55
72
  route: ORM_REST_ROUTE ?? '/',
package/package.json CHANGED
@@ -4,7 +4,7 @@
4
4
  "stonyx-async",
5
5
  "stonyx-module"
6
6
  ],
7
- "version": "0.2.1-beta.80",
7
+ "version": "0.2.1-beta.82",
8
8
  "description": "",
9
9
  "main": "src/main.js",
10
10
  "type": "module",
@@ -48,12 +48,16 @@
48
48
  },
49
49
  "peerDependencies": {
50
50
  "@stonyx/rest-server": ">=0.2.1-beta.14",
51
- "mysql2": "^3.0.0"
51
+ "mysql2": "^3.0.0",
52
+ "pg": "^8.0.0"
52
53
  },
53
54
  "peerDependenciesMeta": {
54
55
  "mysql2": {
55
56
  "optional": true
56
57
  },
58
+ "pg": {
59
+ "optional": true
60
+ },
57
61
  "@stonyx/rest-server": {
58
62
  "optional": true
59
63
  }
@@ -62,6 +66,7 @@
62
66
  "@stonyx/rest-server": "0.2.1-beta.30",
63
67
  "@stonyx/utils": "0.2.3-beta.7",
64
68
  "mysql2": "^3.20.0",
69
+ "pg": "^8.20.0",
65
70
  "qunit": "^2.24.1",
66
71
  "sinon": "^21.0.0"
67
72
  },
package/src/main.js CHANGED
@@ -109,11 +109,16 @@ export default class Orm {
109
109
 
110
110
  setup(eventNames);
111
111
 
112
- if (config.orm.mysql) {
112
+ if (config.orm.postgres) {
113
+ const { default: PostgresDB } = await import('./postgres/postgres-db.js');
114
+ this.sqlDb = new PostgresDB();
115
+ this.db = this.sqlDb;
116
+ promises.push(this.sqlDb.init());
117
+ } else if (config.orm.mysql) {
113
118
  const { default: MysqlDB } = await import('./mysql/mysql-db.js');
114
- this.mysqlDb = new MysqlDB();
115
- this.db = this.mysqlDb;
116
- promises.push(this.mysqlDb.init());
119
+ this.sqlDb = new MysqlDB();
120
+ this.db = this.sqlDb;
121
+ promises.push(this.sqlDb.init());
117
122
  } else if (this.options.dbType !== 'none') {
118
123
  const db = new DB();
119
124
  this.db = db;
@@ -131,9 +136,9 @@ export default class Orm {
131
136
  return modelClass?.memory === true;
132
137
  };
133
138
 
134
- // Wire up MySQL reference for on-demand queries from store.find()/findAll()
135
- if (this.mysqlDb) {
136
- Orm.store._mysqlDb = this.mysqlDb;
139
+ // Wire up SQL adapter reference for on-demand queries from store.find()/findAll()
140
+ if (this.sqlDb) {
141
+ Orm.store._sqlDb = this.sqlDb;
137
142
  }
138
143
 
139
144
  Orm.ready = await Promise.all(promises);
@@ -141,11 +146,11 @@ export default class Orm {
141
146
  }
142
147
 
143
148
  async startup() {
144
- if (this.mysqlDb) await this.mysqlDb.startup();
149
+ if (this.sqlDb) await this.sqlDb.startup();
145
150
  }
146
151
 
147
152
  async shutdown() {
148
- if (this.mysqlDb) await this.mysqlDb.shutdown();
153
+ if (this.sqlDb) await this.sqlDb.shutdown();
149
154
  }
150
155
 
151
156
  static get db() {
@@ -110,10 +110,10 @@ export function updateRecord(record, rawData, userOptions={}) {
110
110
  function assignRecordId(modelName, rawData) {
111
111
  if (rawData.id) return;
112
112
 
113
- // In MySQL mode with numeric IDs, defer to MySQL auto-increment
114
- if (Orm.instance?.mysqlDb && !isStringIdModel(modelName)) {
113
+ // In SQL mode with numeric IDs, defer to database auto-increment
114
+ if (Orm.instance?.sqlDb && !isStringIdModel(modelName)) {
115
115
  rawData.id = `__pending_${Date.now()}_${Math.random()}`;
116
- rawData.__pendingMysqlId = true;
116
+ rawData.__pendingSqlId = true;
117
117
  return;
118
118
  }
119
119
 
@@ -352,7 +352,7 @@ export default class MysqlDB {
352
352
  const insertData = this._recordToRow(record, schema);
353
353
 
354
354
  // For auto-increment models, remove the pending ID
355
- const isPendingId = record.__data.__pendingMysqlId;
355
+ const isPendingId = record.__data.__pendingSqlId;
356
356
 
357
357
  if (isPendingId) {
358
358
  delete insertData.id;
@@ -380,7 +380,7 @@ export default class MysqlDB {
380
380
  response.data.id = realId;
381
381
  }
382
382
 
383
- delete record.__data.__pendingMysqlId;
383
+ delete record.__data.__pendingSqlId;
384
384
  }
385
385
  }
386
386
 
@@ -385,9 +385,9 @@ export default class OrmRequest extends Request {
385
385
  // Execute main handler
386
386
  const response = await handler(request, state);
387
387
 
388
- // Persist to MySQL for write operations
389
- if (Orm.instance.mysqlDb && WRITE_OPERATIONS.has(operation)) {
390
- await Orm.instance.mysqlDb.persist(operation, this.model, context, response);
388
+ // Persist to SQL database for write operations
389
+ if (Orm.instance.sqlDb && WRITE_OPERATIONS.has(operation)) {
390
+ await Orm.instance.sqlDb.persist(operation, this.model, context, response);
391
391
  }
392
392
 
393
393
  // Add response and relevant records to context
@@ -0,0 +1,30 @@
1
+ let pool = null;
2
+
3
+ export async function getPool(pgConfig) {
4
+ if (pool) return pool;
5
+
6
+ const { default: pg } = await import('pg');
7
+
8
+ pool = new pg.Pool({
9
+ host: pgConfig.host,
10
+ port: pgConfig.port,
11
+ user: pgConfig.user,
12
+ password: pgConfig.password,
13
+ database: pgConfig.database,
14
+ max: pgConfig.connectionLimit,
15
+ idleTimeoutMillis: 30000,
16
+ connectionTimeoutMillis: 10000,
17
+ });
18
+
19
+ // Enable pgvector extension
20
+ await pool.query('CREATE EXTENSION IF NOT EXISTS vector');
21
+
22
+ return pool;
23
+ }
24
+
25
+ export async function closePool() {
26
+ if (!pool) return;
27
+
28
+ await pool.end();
29
+ pool = null;
30
+ }
@@ -0,0 +1,302 @@
1
+ import { introspectModels, introspectViews, buildTableDDL, buildViewDDL, buildVectorIndexDDL, schemasToSnapshot, viewSchemasToSnapshot, getTopologicalOrder } from './schema-introspector.js';
2
+ import { readFile, createFile, createDirectory, fileExists } from '@stonyx/utils/file';
3
+ import path from 'path';
4
+ import config from 'stonyx/config';
5
+ import log from 'stonyx/log';
6
+
7
+ export async function generateMigration(description = 'migration') {
8
+ const { migrationsDir } = config.orm.postgres;
9
+ const rootPath = config.rootPath;
10
+ const migrationsPath = path.resolve(rootPath, migrationsDir);
11
+
12
+ await createDirectory(migrationsPath);
13
+
14
+ const schemas = introspectModels();
15
+ const currentSnapshot = schemasToSnapshot(schemas);
16
+ const previousSnapshot = await loadLatestSnapshot(migrationsPath);
17
+ const diff = diffSnapshots(previousSnapshot, currentSnapshot);
18
+
19
+ // Don't return early — check view changes too before deciding
20
+ if (!diff.hasChanges) {
21
+ const viewSchemasPrelim = introspectViews();
22
+ const currentViewSnapshotPrelim = viewSchemasToSnapshot(viewSchemasPrelim);
23
+ const previousViewSnapshotPrelim = extractViewsFromSnapshot(previousSnapshot);
24
+ const viewDiffPrelim = diffViewSnapshots(previousViewSnapshotPrelim, currentViewSnapshotPrelim);
25
+
26
+ if (!viewDiffPrelim.hasChanges) {
27
+ log.db('No schema changes detected.');
28
+ return null;
29
+ }
30
+ }
31
+
32
+ const upStatements = [];
33
+ const downStatements = [];
34
+
35
+ // pgvector extension (include in initial migration)
36
+ if (Object.keys(previousSnapshot).length === 0) {
37
+ upStatements.push('CREATE EXTENSION IF NOT EXISTS vector;');
38
+ }
39
+
40
+ // New tables — in topological order (parents before children)
41
+ const allOrder = getTopologicalOrder(schemas);
42
+ const addedOrdered = allOrder.filter(name => diff.addedModels.includes(name));
43
+
44
+ for (const name of addedOrdered) {
45
+ upStatements.push(buildTableDDL(name, schemas[name], schemas) + ';');
46
+
47
+ // HNSW indexes for vector columns
48
+ const indexStatements = buildVectorIndexDDL(name, schemas[name]);
49
+ for (const stmt of indexStatements) {
50
+ upStatements.push(stmt + ';');
51
+ }
52
+
53
+ downStatements.unshift(`DROP TABLE IF EXISTS "${schemas[name].table}" CASCADE;`);
54
+ }
55
+
56
+ // Removed tables (warn only, commented out)
57
+ for (const name of diff.removedModels) {
58
+ upStatements.push(`-- WARNING: Model '${name}' was removed. Uncomment to drop table:`);
59
+ upStatements.push(`-- DROP TABLE IF EXISTS "${previousSnapshot[name].table}" CASCADE;`);
60
+ downStatements.push(`-- Recreate table for removed model '${name}' manually if needed`);
61
+ }
62
+
63
+ // Added columns
64
+ for (const { model, column, type } of diff.addedColumns) {
65
+ const table = currentSnapshot[model].table;
66
+ upStatements.push(`ALTER TABLE "${table}" ADD COLUMN "${column}" ${type};`);
67
+ downStatements.push(`ALTER TABLE "${table}" DROP COLUMN "${column}";`);
68
+
69
+ // Add HNSW index if it's a vector column
70
+ if (type.startsWith('vector(')) {
71
+ upStatements.push(`CREATE INDEX IF NOT EXISTS "idx_${table}_${column}_hnsw" ON "${table}" USING hnsw ("${column}" vector_cosine_ops) WITH (m = 16, ef_construction = 200);`);
72
+ }
73
+ }
74
+
75
+ // Removed columns
76
+ for (const { model, column, type } of diff.removedColumns) {
77
+ const table = previousSnapshot[model].table;
78
+ upStatements.push(`ALTER TABLE "${table}" DROP COLUMN "${column}";`);
79
+ downStatements.push(`ALTER TABLE "${table}" ADD COLUMN "${column}" ${type};`);
80
+ }
81
+
82
+ // Changed column types
83
+ for (const { model, column, from, to } of diff.changedColumns) {
84
+ const table = currentSnapshot[model].table;
85
+ upStatements.push(`ALTER TABLE "${table}" ALTER COLUMN "${column}" TYPE ${to};`);
86
+ downStatements.push(`ALTER TABLE "${table}" ALTER COLUMN "${column}" TYPE ${from};`);
87
+ }
88
+
89
+ // Added foreign keys
90
+ for (const { model, column, references } of diff.addedForeignKeys) {
91
+ const table = currentSnapshot[model].table;
92
+ const refModel = Object.entries(currentSnapshot).find(([, s]) => s.table === references.references);
93
+ const fkType = refModel && refModel[1].idType === 'string' ? 'VARCHAR(255)' : 'INTEGER';
94
+ const constraintName = `fk_${table}_${column}`;
95
+ upStatements.push(`ALTER TABLE "${table}" ADD COLUMN "${column}" ${fkType};`);
96
+ upStatements.push(`ALTER TABLE "${table}" ADD CONSTRAINT "${constraintName}" FOREIGN KEY ("${column}") REFERENCES "${references.references}"("${references.column}") ON DELETE SET NULL;`);
97
+ downStatements.push(`ALTER TABLE "${table}" DROP CONSTRAINT "${constraintName}";`);
98
+ downStatements.push(`ALTER TABLE "${table}" DROP COLUMN "${column}";`);
99
+ }
100
+
101
+ // Removed foreign keys
102
+ for (const { model, column, references } of diff.removedForeignKeys) {
103
+ const table = previousSnapshot[model].table;
104
+ const refModel = Object.entries(previousSnapshot).find(([, s]) => s.table === references.references);
105
+ const fkType = refModel && refModel[1].idType === 'string' ? 'VARCHAR(255)' : 'INTEGER';
106
+ const constraintName = `fk_${table}_${column}`;
107
+ upStatements.push(`ALTER TABLE "${table}" DROP CONSTRAINT "${constraintName}";`);
108
+ upStatements.push(`ALTER TABLE "${table}" DROP COLUMN "${column}";`);
109
+ downStatements.push(`ALTER TABLE "${table}" ADD COLUMN "${column}" ${fkType};`);
110
+ downStatements.push(`ALTER TABLE "${table}" ADD CONSTRAINT "${constraintName}" FOREIGN KEY ("${column}") REFERENCES "${references.references}"("${references.column}") ON DELETE SET NULL;`);
111
+ }
112
+
113
+ // View migrations — views are created AFTER tables (dependency order)
114
+ const viewSchemas = introspectViews();
115
+ const currentViewSnapshot = viewSchemasToSnapshot(viewSchemas);
116
+ const previousViewSnapshot = extractViewsFromSnapshot(previousSnapshot);
117
+ const viewDiff = diffViewSnapshots(previousViewSnapshot, currentViewSnapshot);
118
+
119
+ if (viewDiff.hasChanges) {
120
+ upStatements.push('');
121
+ upStatements.push('-- Views');
122
+ downStatements.push('');
123
+ downStatements.push('-- Views');
124
+
125
+ // Added views
126
+ for (const name of viewDiff.addedViews) {
127
+ try {
128
+ const ddl = buildViewDDL(name, viewSchemas[name], schemas);
129
+ upStatements.push(ddl + ';');
130
+ downStatements.unshift(`DROP VIEW IF EXISTS "${viewSchemas[name].viewName}";`);
131
+ } catch (error) {
132
+ upStatements.push(`-- WARNING: Could not generate DDL for view '${name}': ${error.message}`);
133
+ }
134
+ }
135
+
136
+ // Removed views
137
+ for (const name of viewDiff.removedViews) {
138
+ upStatements.push(`-- WARNING: View '${name}' was removed. Uncomment to drop view:`);
139
+ upStatements.push(`-- DROP VIEW IF EXISTS "${previousViewSnapshot[name].viewName}";`);
140
+ downStatements.push(`-- Recreate view for removed view '${name}' manually if needed`);
141
+ }
142
+
143
+ // Changed views (source or aggregates changed)
144
+ for (const name of viewDiff.changedViews) {
145
+ try {
146
+ const ddl = buildViewDDL(name, viewSchemas[name], schemas);
147
+ upStatements.push(ddl + ';');
148
+ } catch (error) {
149
+ upStatements.push(`-- WARNING: Could not generate DDL for changed view '${name}': ${error.message}`);
150
+ }
151
+ }
152
+ }
153
+
154
+ const combinedHasChanges = diff.hasChanges || viewDiff.hasChanges;
155
+
156
+ if (!combinedHasChanges) {
157
+ log.db('No schema changes detected.');
158
+ return null;
159
+ }
160
+
161
+ // Merge view snapshot into the main snapshot
162
+ const combinedSnapshot = { ...currentSnapshot };
163
+ for (const [name, viewSnap] of Object.entries(currentViewSnapshot)) {
164
+ combinedSnapshot[name] = viewSnap;
165
+ }
166
+
167
+ const sanitizedDescription = description.replace(/\s+/g, '_').replace(/[^a-zA-Z0-9_]/g, '');
168
+ const timestamp = Math.floor(Date.now() / 1000);
169
+ const filename = `${timestamp}_${sanitizedDescription}.sql`;
170
+ const content = `-- UP\n${upStatements.join('\n')}\n\n-- DOWN\n${downStatements.join('\n')}\n`;
171
+
172
+ await createFile(path.join(migrationsPath, filename), content);
173
+ await createFile(path.join(migrationsPath, '.snapshot.json'), JSON.stringify(combinedSnapshot, null, 2));
174
+
175
+ log.db(`Migration generated: ${filename}`);
176
+
177
+ return { filename, content, snapshot: combinedSnapshot };
178
+ }
179
+
180
+ export async function loadLatestSnapshot(migrationsPath) {
181
+ const snapshotPath = path.join(migrationsPath, '.snapshot.json');
182
+ const exists = await fileExists(snapshotPath);
183
+
184
+ if (!exists) return {};
185
+
186
+ return readFile(snapshotPath, { json: true });
187
+ }
188
+
189
+ export function diffSnapshots(previous, current) {
190
+ const addedModels = [];
191
+ const removedModels = [];
192
+ const addedColumns = [];
193
+ const removedColumns = [];
194
+ const changedColumns = [];
195
+ const addedForeignKeys = [];
196
+ const removedForeignKeys = [];
197
+
198
+ // Find added models
199
+ for (const name of Object.keys(current)) {
200
+ if (!previous[name]) addedModels.push(name);
201
+ }
202
+
203
+ // Find removed models
204
+ for (const name of Object.keys(previous)) {
205
+ if (!current[name]) removedModels.push(name);
206
+ }
207
+
208
+ // Find column changes in existing models
209
+ for (const name of Object.keys(current)) {
210
+ if (!previous[name]) continue;
211
+
212
+ const { columns: prevCols = {} } = previous[name];
213
+ const { columns: currCols = {} } = current[name];
214
+
215
+ // Added columns
216
+ for (const [col, type] of Object.entries(currCols)) {
217
+ if (!prevCols[col]) {
218
+ addedColumns.push({ model: name, column: col, type });
219
+ } else if (prevCols[col] !== type) {
220
+ changedColumns.push({ model: name, column: col, from: prevCols[col], to: type });
221
+ }
222
+ }
223
+
224
+ // Removed columns
225
+ for (const [col, type] of Object.entries(prevCols)) {
226
+ if (!currCols[col]) {
227
+ removedColumns.push({ model: name, column: col, type });
228
+ }
229
+ }
230
+
231
+ // Foreign key changes
232
+ const prevFKs = previous[name].foreignKeys || {};
233
+ const currFKs = current[name].foreignKeys || {};
234
+
235
+ for (const [col, refs] of Object.entries(currFKs)) {
236
+ if (!prevFKs[col]) {
237
+ addedForeignKeys.push({ model: name, column: col, references: refs });
238
+ }
239
+ }
240
+
241
+ for (const [col, refs] of Object.entries(prevFKs)) {
242
+ if (!currFKs[col]) {
243
+ removedForeignKeys.push({ model: name, column: col, references: refs });
244
+ }
245
+ }
246
+ }
247
+
248
+ const hasChanges = addedModels.length > 0 || removedModels.length > 0 ||
249
+ addedColumns.length > 0 || removedColumns.length > 0 ||
250
+ changedColumns.length > 0 || addedForeignKeys.length > 0 || removedForeignKeys.length > 0;
251
+
252
+ return {
253
+ hasChanges,
254
+ addedModels,
255
+ removedModels,
256
+ addedColumns,
257
+ removedColumns,
258
+ changedColumns,
259
+ addedForeignKeys,
260
+ removedForeignKeys,
261
+ };
262
+ }
263
+
264
+ export function detectSchemaDrift(schemas, snapshot) {
265
+ const current = schemasToSnapshot(schemas);
266
+ return diffSnapshots(snapshot, current);
267
+ }
268
+
269
+ export function extractViewsFromSnapshot(snapshot) {
270
+ const views = {};
271
+ for (const [name, entry] of Object.entries(snapshot)) {
272
+ if (entry.isView) views[name] = entry;
273
+ }
274
+ return views;
275
+ }
276
+
277
+ export function diffViewSnapshots(previous, current) {
278
+ const addedViews = [];
279
+ const removedViews = [];
280
+ const changedViews = [];
281
+
282
+ for (const name of Object.keys(current)) {
283
+ if (!previous[name]) {
284
+ addedViews.push(name);
285
+ } else if (
286
+ current[name].viewQuery !== previous[name].viewQuery ||
287
+ current[name].source !== previous[name].source
288
+ ) {
289
+ changedViews.push(name);
290
+ }
291
+ }
292
+
293
+ for (const name of Object.keys(previous)) {
294
+ if (!current[name]) {
295
+ removedViews.push(name);
296
+ }
297
+ }
298
+
299
+ const hasChanges = addedViews.length > 0 || removedViews.length > 0 || changedViews.length > 0;
300
+
301
+ return { hasChanges, addedViews, removedViews, changedViews };
302
+ }
@@ -0,0 +1,109 @@
1
+ import { readFile, fileExists } from '@stonyx/utils/file';
2
+ import path from 'path';
3
+ import fs from 'fs/promises';
4
+
5
+ export async function ensureMigrationsTable(pool, tableName = '__migrations') {
6
+ await pool.query(`
7
+ CREATE TABLE IF NOT EXISTS "${tableName}" (
8
+ id INTEGER GENERATED ALWAYS AS IDENTITY PRIMARY KEY,
9
+ filename VARCHAR(255) NOT NULL UNIQUE,
10
+ applied_at TIMESTAMPTZ DEFAULT NOW()
11
+ )
12
+ `);
13
+ }
14
+
15
+ export async function getAppliedMigrations(pool, tableName = '__migrations') {
16
+ const result = await pool.query(
17
+ `SELECT filename FROM "${tableName}" ORDER BY id ASC`
18
+ );
19
+
20
+ return result.rows.map(row => row.filename);
21
+ }
22
+
23
+ export async function getMigrationFiles(migrationsDir) {
24
+ const exists = await fileExists(migrationsDir);
25
+ if (!exists) return [];
26
+
27
+ const entries = await fs.readdir(migrationsDir);
28
+
29
+ return entries
30
+ .filter(f => f.endsWith('.sql'))
31
+ .sort();
32
+ }
33
+
34
+ export function parseMigrationFile(content) {
35
+ const upMarker = '-- UP';
36
+ const downMarker = '-- DOWN';
37
+ const upIndex = content.indexOf(upMarker);
38
+ const downIndex = content.indexOf(downMarker);
39
+
40
+ if (upIndex === -1) {
41
+ return { up: content.trim(), down: '' };
42
+ }
43
+
44
+ const upStart = upIndex + upMarker.length;
45
+ const upEnd = downIndex !== -1 ? downIndex : content.length;
46
+ const up = content.slice(upStart, upEnd).trim();
47
+ const down = downIndex !== -1 ? content.slice(downIndex + downMarker.length).trim() : '';
48
+
49
+ return { up, down };
50
+ }
51
+
52
+ export async function applyMigration(pool, filename, upSql, tableName = '__migrations') {
53
+ const client = await pool.connect();
54
+
55
+ try {
56
+ await client.query('BEGIN');
57
+
58
+ const statements = splitStatements(upSql);
59
+
60
+ for (const stmt of statements) {
61
+ await client.query(stmt);
62
+ }
63
+
64
+ await client.query(
65
+ `INSERT INTO "${tableName}" (filename) VALUES ($1)`,
66
+ [filename]
67
+ );
68
+
69
+ await client.query('COMMIT');
70
+ } catch (error) {
71
+ await client.query('ROLLBACK');
72
+ throw error;
73
+ } finally {
74
+ client.release();
75
+ }
76
+ }
77
+
78
+ export async function rollbackMigration(pool, filename, downSql, tableName = '__migrations') {
79
+ const client = await pool.connect();
80
+
81
+ try {
82
+ await client.query('BEGIN');
83
+
84
+ const statements = splitStatements(downSql);
85
+
86
+ for (const stmt of statements) {
87
+ await client.query(stmt);
88
+ }
89
+
90
+ await client.query(
91
+ `DELETE FROM "${tableName}" WHERE filename = $1`,
92
+ [filename]
93
+ );
94
+
95
+ await client.query('COMMIT');
96
+ } catch (error) {
97
+ await client.query('ROLLBACK');
98
+ throw error;
99
+ } finally {
100
+ client.release();
101
+ }
102
+ }
103
+
104
+ function splitStatements(sql) {
105
+ return sql
106
+ .split(';')
107
+ .map(s => s.trim())
108
+ .filter(s => s.length > 0 && !s.startsWith('--'));
109
+ }