driftsql 2.0.0-beta.1 → 2.0.0-beta.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,4 @@
1
+ import { n as init_postgres, t as PostgresGenerator } from "./postgres-CjYq8GvL.js";
2
+
3
+ init_postgres();
4
+ export { PostgresGenerator };
@@ -0,0 +1,95 @@
1
+ //#region rolldown:runtime
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __esm = (fn, res) => function() {
7
+ return fn && (res = (0, fn[__getOwnPropNames(fn)[0]])(fn = 0)), res;
8
+ };
9
+ var __export = (all) => {
10
+ let target = {};
11
+ for (var name in all) __defProp(target, name, {
12
+ get: all[name],
13
+ enumerable: true
14
+ });
15
+ return target;
16
+ };
17
+ var __copyProps = (to, from, except, desc) => {
18
+ if (from && typeof from === "object" || typeof from === "function") for (var keys = __getOwnPropNames(from), i = 0, n = keys.length, key; i < n; i++) {
19
+ key = keys[i];
20
+ if (!__hasOwnProp.call(to, key) && key !== except) __defProp(to, key, {
21
+ get: ((k) => from[k]).bind(null, key),
22
+ enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable
23
+ });
24
+ }
25
+ return to;
26
+ };
27
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
28
+
29
+ //#endregion
30
+ //#region src/schema/generators/postgres.ts
31
+ var postgres_exports = /* @__PURE__ */ __export({ PostgresGenerator: () => PostgresGenerator });
32
+ var PostgresGenerator;
33
+ var init_postgres = __esm({ "src/schema/generators/postgres.ts": (() => {
34
+ PostgresGenerator = class {
35
+ generateCreateTable(table) {
36
+ const lines = [];
37
+ lines.push(`CREATE TABLE "${table.name}" (`);
38
+ const columnDefs = [];
39
+ for (const column of table.columns) columnDefs.push(" " + this.generateColumnDefinition(column));
40
+ if (table.primaryKey && table.primaryKey.length > 0) {
41
+ const pkColumns = table.primaryKey.map((col) => `"${col}"`).join(", ");
42
+ columnDefs.push(` PRIMARY KEY (${pkColumns})`);
43
+ }
44
+ if (table.checks && table.checks.length > 0) for (const check of table.checks) columnDefs.push(` CONSTRAINT "${check.name}" CHECK (${check.expression})`);
45
+ lines.push(columnDefs.join(",\n"));
46
+ lines.push(");");
47
+ const sql = [lines.join("\n")];
48
+ if (table.indexes && table.indexes.length > 0) for (const index of table.indexes) sql.push(this.generateIndex(table.name, index));
49
+ return sql.join("\n\n");
50
+ }
51
+ generateDropTable(tableName) {
52
+ return `DROP TABLE IF EXISTS "${tableName}" CASCADE;`;
53
+ }
54
+ generateColumnDefinition(column) {
55
+ const parts = [`"${column.name}"`];
56
+ parts.push(this.getColumnType(column));
57
+ if (column.primaryKey && !column.type.includes("serial")) parts.push("PRIMARY KEY");
58
+ if (column.notNull && !column.primaryKey && !column.type.includes("serial")) parts.push("NOT NULL");
59
+ if (column.unique) parts.push("UNIQUE");
60
+ if (column.default !== void 0) parts.push(`DEFAULT ${this.formatDefault(column.default)}`);
61
+ if (column.references) {
62
+ const ref = column.references;
63
+ parts.push(`REFERENCES "${ref.table}"("${ref.column}")`);
64
+ if (ref.onDelete) parts.push(`ON DELETE ${ref.onDelete}`);
65
+ if (ref.onUpdate) parts.push(`ON UPDATE ${ref.onUpdate}`);
66
+ }
67
+ if (column.check) parts.push(`CHECK (${column.check})`);
68
+ return parts.join(" ");
69
+ }
70
+ getColumnType(column) {
71
+ let type = column.type.toUpperCase();
72
+ if (column.length) type += `(${column.length})`;
73
+ else if (column.precision) if (column.scale !== void 0) type += `(${column.precision}, ${column.scale})`;
74
+ else type += `(${column.precision})`;
75
+ return type;
76
+ }
77
+ formatDefault(value) {
78
+ if (typeof value === "string") {
79
+ if (value.toUpperCase() === "NOW()" || value.toUpperCase() === "CURRENT_TIMESTAMP") return value.toUpperCase();
80
+ return `'${value}'`;
81
+ }
82
+ if (typeof value === "boolean") return value ? "TRUE" : "FALSE";
83
+ return String(value);
84
+ }
85
+ generateIndex(tableName, index) {
86
+ const unique = index.unique ? "UNIQUE " : "";
87
+ const type = index.type ? ` USING ${index.type.toUpperCase()}` : "";
88
+ const columns = index.columns.map((col) => `"${col}"`).join(", ");
89
+ return `CREATE ${unique}INDEX "${index.name}" ON "${tableName}"${type} (${columns});`;
90
+ }
91
+ };
92
+ }) });
93
+
94
+ //#endregion
95
+ export { __export as a, __esm as i, init_postgres as n, __toCommonJS as o, postgres_exports as r, PostgresGenerator as t };
@@ -0,0 +1,88 @@
1
+ import fs from "node:fs/promises";
2
+
3
+ //#region src/schema/type-generator.ts
4
+ var TypeGenerator = class {
5
+ generateTypes(tables) {
6
+ let output = "// This file was automatically generated - DO NOT EDIT\n\n";
7
+ for (const table of tables) {
8
+ output += this.generateTableInterface(table);
9
+ output += "\n\n";
10
+ }
11
+ output += this.generateDatabaseInterface(tables);
12
+ return output;
13
+ }
14
+ generateTableInterface(table) {
15
+ let lines = [`export interface ${this.pascalCase(table.name)} {`];
16
+ for (const column of table.columns) {
17
+ const tsType = this.columnToTypeScript(column);
18
+ const optional = !column.notNull && !column.primaryKey ? "?" : "";
19
+ lines.push(` ${column.name}${optional}: ${tsType}`);
20
+ }
21
+ lines.push("}");
22
+ return lines.join("\n");
23
+ }
24
+ generateDatabaseInterface(tables) {
25
+ let lines = ["export interface Database {"];
26
+ for (const table of tables) {
27
+ const interfaceName = this.pascalCase(table.name);
28
+ lines.push(` ${table.name}: ${interfaceName}`);
29
+ }
30
+ lines.push("}");
31
+ return lines.join("\n");
32
+ }
33
+ columnToTypeScript(column) {
34
+ const isNullable = !column.notNull && !column.primaryKey;
35
+ let baseType;
36
+ switch (column.type.toLowerCase()) {
37
+ case "serial":
38
+ case "bigserial":
39
+ case "integer":
40
+ case "int":
41
+ case "bigint":
42
+ case "smallint":
43
+ case "numeric":
44
+ case "decimal":
45
+ case "real":
46
+ case "double precision":
47
+ baseType = "number";
48
+ break;
49
+ case "text":
50
+ case "varchar":
51
+ case "char":
52
+ case "uuid":
53
+ baseType = "string";
54
+ break;
55
+ case "boolean":
56
+ baseType = "boolean";
57
+ break;
58
+ case "timestamp":
59
+ case "timestamptz":
60
+ case "date":
61
+ case "time":
62
+ baseType = "Date | string";
63
+ break;
64
+ case "json":
65
+ case "jsonb":
66
+ baseType = "any";
67
+ break;
68
+ case "bytea":
69
+ baseType = "Buffer";
70
+ break;
71
+ default: baseType = "any";
72
+ }
73
+ return isNullable ? `${baseType} | null` : baseType;
74
+ }
75
+ pascalCase(str) {
76
+ return str.split("_").map((word) => word.charAt(0).toUpperCase() + word.slice(1).toLowerCase()).join("");
77
+ }
78
+ async writeToFile(tables, filePath) {
79
+ const content = this.generateTypes(tables);
80
+ await fs.writeFile(filePath, content, "utf8");
81
+ }
82
+ };
83
+ async function generateTypesFromSchema(tables, outputPath) {
84
+ await new TypeGenerator().writeToFile(tables, outputPath);
85
+ }
86
+
87
+ //#endregion
88
+ export { generateTypesFromSchema as n, TypeGenerator as t };
@@ -0,0 +1,3 @@
1
+ import { n as generateTypesFromSchema, t as TypeGenerator } from "./type-generator-Ba8bgnMm.js";
2
+
3
+ export { TypeGenerator, generateTypesFromSchema };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "driftsql",
3
- "version": "2.0.0-beta.1",
3
+ "version": "2.0.0-beta.3",
4
4
  "description": "A modern, type-safe SQL client with built-in schema builder and migration system for TypeScript.",
5
5
  "type": "module",
6
6
  "license": "MIT",
@@ -20,8 +20,7 @@
20
20
  "module": "./dist/index.js",
21
21
  "types": "./dist/index.d.ts",
22
22
  "bin": {
23
- "driftsql": "./dist/cli/index.js",
24
- "sql": "./dist/index.js"
23
+ "driftsql": "./dist/index.js"
25
24
  },
26
25
  "exports": {
27
26
  ".": "./dist/index.js",
@@ -1 +0,0 @@
1
- export { };
package/dist/cli/index.js DELETED
@@ -1,57 +0,0 @@
1
- #!/usr/bin/env node
2
- import"../postgres-9C7eE0wB.js";import{a as e,o as t,r as n,s as r}from"../src-B1L6LdRV.js";import"../type-generator-CpqI7vBb.js";import i from"consola";import a from"node:fs/promises";import o from"node:path";import{Command as s}from"commander";const c=new s;c.name(`driftsql`).description(`DriftSQL CLI - Database migrations and schema management`).version(`0.0.1`),c.command(`migrate:up`).description(`Run all pending migrations`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).option(`-c, --config <path>`,`Config file path`,`./driftsql.config.ts`).action(async e=>{try{let{client:t,migrations:n}=await u(e.dir,e.config);await new r(t).upAll(n),await t.close(),i.success(`All migrations applied successfully`)}catch(e){i.error(`Migration failed:`,e),process.exit(1)}}),c.command(`migrate:down`).description(`Rollback the last migration`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).option(`-c, --config <path>`,`Config file path`,`./driftsql.config.ts`).action(async e=>{try{let{client:t,migrations:n}=await u(e.dir,e.config),a=new r(t),o=await a.getAppliedMigrations(),s=n.find(e=>e.version===o[o.length-1]);s?await a.down(s):i.info(`No migrations to rollback`),await t.close()}catch(e){i.error(`Rollback failed:`,e),process.exit(1)}}),c.command(`migrate:reset`).description(`Reset all migrations (down then up)`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).option(`-c, --config <path>`,`Config file path`,`./driftsql.config.ts`).action(async e=>{try{let{client:t,migrations:n}=await u(e.dir,e.config);await new r(t).reset(n),await t.close(),i.success(`All migrations reset successfully`)}catch(e){i.error(`Reset failed:`,e),process.exit(1)}}),c.command(`migrate:status`).description(`Show migration status`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).option(`-c, --config <path>`,`Config file path`,`./driftsql.config.ts`).action(async e=>{try{let{client:t,migrations:n}=await u(e.dir,e.config),a=await new r(t).getAppliedMigrations();i.info(`Migration Status:
3
- `);for(let e of n){let t=a.includes(e.version)?`✓ Applied`:`✗ Pending`;i.log(`${t} - ${e.version} (${e.name})`)}await t.close()}catch(e){i.error(`Failed to get status:`,e),process.exit(1)}}),c.command(`migrate:create <name>`).description(`Create a new migration file`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).action(async(e,t)=>{try{let n=new Date().toISOString().replace(/[-:T.]/g,``).slice(0,14),r=`${n}_${e}.ts`,s=o.join(t.dir,r);await a.mkdir(t.dir,{recursive:!0});let c=`import { createMigration } from 'driftsql'
4
-
5
- export const migration = createMigration('postgres')
6
- .raw('-- Add your migration SQL here')
7
- .build('${n}', '${e}')
8
- `;await a.writeFile(s,c,`utf8`),i.success(`Created migration: ${r}`)}catch(e){i.error(`Failed to create migration:`,e),process.exit(1)}}),c.command(`migrate:generate [name]`).description(`Automatically generate migration from schema changes`).option(`-s, --schema <path>`,`Schema file path`,`./schema.ts`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).action(async(r,s)=>{try{let c=new n,l=o.resolve(process.cwd(),s.schema);i.start(`Loading schema...`);let u=await import(l),d=u.default||u.schema;(!d||!Array.isArray(d))&&(i.error(`Schema file must export default or named "schema" as an array of table definitions`),i.info(`Example: export default [usersTable.getDefinition(), postsTable.getDefinition()]`),process.exit(1));let f=d,p=await c.load();if(!p){i.info(`No previous schema snapshot found, creating initial migration...`);let e=new Date().toISOString().replace(/[-:T.]/g,``).slice(0,14),t=r||`initial`,n=`${e}_${t}.ts`,l=o.join(s.dir,n);await a.mkdir(s.dir,{recursive:!0});let{PostgresGenerator:u}=await import(`../postgres-DfhzvfQU.js`),d=new u,p=f.map(e=>d.generateCreateTable(e)).map(e=>` .raw(\`${e.replace(/`/g,"\\`")}\`)`).join(`
9
- `),m=`import { createMigration } from 'driftsql'
10
-
11
- // Initial migration - automatically generated
12
- // Generated at: ${new Date().toISOString()}
13
-
14
- export const migration = createMigration('postgres')
15
- ${p}
16
- .build('${e}', '${t}')
17
- `;await a.writeFile(l,m,`utf8`),await c.save(f);let{generateTypesFromSchema:h}=await import(`../type-generator-DdCKALf8.js`);await h(f,o.join(process.cwd(),`db-types.ts`)),i.success(`\nCreated initial migration: ${n}`),i.success(`Generated TypeScript types: db-types.ts`),i.info(`Snapshot saved. Run "driftsql migrate:up" to apply changes.`);return}i.info(`Detecting changes...`);let m=e(p.tables,f,`postgres`);if(m.length===0){i.success(`No schema changes detected!`);return}i.info(`Found ${m.length} change(s):`),m.forEach((e,t)=>{i.log(` ${t+1}. ${e.type} - ${e.table}`)});let h=new Date().toISOString().replace(/[-:T.]/g,``).slice(0,14),g=r||`auto_migration`,_=`${h}_${g}.ts`,v=o.join(s.dir,_);await a.mkdir(s.dir,{recursive:!0});let y=t(m,h,g).build(h,g).up.map(e=>` .raw(\`${e.replace(/`/g,"\\`")}\``).join(`
18
- `),b=`import { createMigration } from 'driftsql'
19
-
20
- // This migration was automatically generated
21
- // Generated at: ${new Date().toISOString()}
22
- // Changes: ${m.length}
23
-
24
- export const migration = createMigration('postgres')
25
- ${y}
26
- .build('${h}', '${g}')
27
- `;await a.writeFile(v,b,`utf8`),await c.save(f);let{generateTypesFromSchema:x}=await import(`../type-generator-DdCKALf8.js`);await x(f,o.join(process.cwd(),`db-types.ts`)),i.success(`\nGenerated migration: ${_}`),i.success(`Updated TypeScript types: db-types.ts`),i.info(`Snapshot updated. Run "driftsql migrate:up" to apply changes.`)}catch(e){i.error(`Failed to generate migration:`,e),process.exit(1)}}),c.command(`db:inspect`).description(`Inspect database and generate TypeScript types`).option(`-c, --config <path>`,`Config file path`,`./driftsql.config.ts`).option(`-o, --output <path>`,`Output file path`,`./db-types.ts`).action(async e=>{try{let t=await l(e.config);await t.inspectDB({driver:t.getDriver(),outputFile:e.output}),await t.close()}catch(e){i.error(`Inspection failed:`,e),process.exit(1)}}),c.command(`generate:types`).description(`Generate TypeScript types from schema file`).option(`-s, --schema <path>`,`Schema file path`,`./schema.ts`).option(`-o, --output <path>`,`Output file path`,`./db-types.ts`).action(async e=>{try{let t=o.resolve(process.cwd(),e.schema),n=o.resolve(process.cwd(),e.output);i.start(`Loading schema...`);let r=await import(t),a=r.default||r.schema;(!a||!Array.isArray(a))&&(i.error(`Schema file must export default or named "schema" as an array of table definitions`),process.exit(1));let{generateTypesFromSchema:s}=await import(`../type-generator-DdCKALf8.js`);await s(a,n),i.success(`TypeScript types generated: ${e.output}`)}catch(e){i.error(`Type generation failed:`,e),process.exit(1)}}),c.command(`init`).description(`Initialize DriftSQL with a basic notes schema`).option(`-d, --dir <directory>`,`Migrations directory`,`./migrations`).action(async e=>{try{await a.mkdir(e.dir,{recursive:!0});let t=`.driftsql/
28
- .env
29
- `,n=o.join(process.cwd(),`driftsql.config.ts`),r=o.join(process.cwd(),`schema.ts`),s=o.join(process.cwd(),`.env.example`),c=o.join(process.cwd(),`.gitignore`),l=[{path:n,content:`import { SQLClient, PostgresDriver } from 'driftsql'
30
-
31
- export default new SQLClient({
32
- driver: new PostgresDriver({
33
- connectionString: process.env.DATABASE_URL!,
34
- }),
35
- })
36
- `,name:`driftsql.config.ts`},{path:r,content:`import { createTable, serial, varchar, text, timestamp, boolean } from 'driftsql'
37
-
38
- const notesTable = createTable('notes', (table) => {
39
- table
40
- .column(serial('id').primaryKey())
41
- .column(varchar('title', 255).notNull())
42
- .column(text('content'))
43
- .column(boolean('is_archived').default(false).notNull())
44
- .column(timestamp('created_at').default('CURRENT_TIMESTAMP').notNull())
45
- .column(timestamp('updated_at').default('CURRENT_TIMESTAMP').notNull())
46
- .index('idx_notes_archived', ['is_archived'])
47
- .index('idx_notes_created_at', ['created_at'])
48
- })
49
-
50
- // Export array of table definitions
51
- export default [notesTable.getDefinition()]
52
- `,name:`schema.ts`},{path:s,content:`DATABASE_URL=postgresql://user:password@localhost:5432/mydb
53
- `,name:`.env.example`}];for(let e of l)try{await a.access(e.path),i.warn(`${e.name} already exists, skipping...`)}catch{await a.writeFile(e.path,e.content,`utf8`),i.success(`Created ${e.name}`)}try{(await a.readFile(c,`utf8`)).includes(`.driftsql/`)||(await a.appendFile(c,`
54
- `+t),i.success(`Updated .gitignore`))}catch{await a.writeFile(c,t,`utf8`),i.success(`Created .gitignore`)}i.success(`
55
- ✨ DriftSQL initialized successfully!
56
- `),i.info(`Next steps:`),i.info(`1. Copy .env.example to .env and update DATABASE_URL`),i.info(`2. Run: driftsql migrate:generate initial (generates migration + types)`),i.info(`3. Run: driftsql migrate:up`),i.info(`4. Import Database type from db-types.ts for type safety`),i.info(`5. Edit schema.ts and run migrate:generate to auto-detect changes!
57
- `)}catch(e){i.error(`Initialization failed:`,e),process.exit(1)}});async function l(e){try{let t=await import(o.resolve(process.cwd(),e));return t.default||t.client}catch(t){throw i.error(`Failed to load config from ${e}`),t}}async function u(e,t){let n=await l(t),r=[];try{let t=o.resolve(process.cwd(),e),i=(await a.readdir(t)).filter(e=>e.endsWith(`.ts`)||e.endsWith(`.js`)).sort();for(let e of i){let n=await import(o.join(t,e)),i=n.migration||n.default;i&&r.push(i)}return{client:n,migrations:r}}catch(t){throw i.error(`Failed to load migrations from ${e}`),t}}c.parse();export{};
@@ -1,5 +0,0 @@
1
- var e=Object.defineProperty,t=Object.getOwnPropertyDescriptor,n=Object.getOwnPropertyNames,r=Object.prototype.hasOwnProperty,i=(e,t)=>()=>(e&&(t=e(e=0)),t),a=t=>{let n={};for(var r in t)e(n,r,{get:t[r],enumerable:!0});return n},o=(i,a,o,s)=>{if(a&&typeof a==`object`||typeof a==`function`)for(var c=n(a),l=0,u=c.length,d;l<u;l++)d=c[l],!r.call(i,d)&&d!==o&&e(i,d,{get:(e=>a[e]).bind(null,d),enumerable:!(s=t(a,d))||s.enumerable});return i},s=t=>o(e({},`__esModule`,{value:!0}),t),c=a({PostgresGenerator:()=>l}),l,u=i((()=>{l=class{generateCreateTable(e){let t=[];t.push(`CREATE TABLE "${e.name}" (`);let n=[];for(let t of e.columns)n.push(` `+this.generateColumnDefinition(t));if(e.primaryKey&&e.primaryKey.length>0){let t=e.primaryKey.map(e=>`"${e}"`).join(`, `);n.push(` PRIMARY KEY (${t})`)}if(e.checks&&e.checks.length>0)for(let t of e.checks)n.push(` CONSTRAINT "${t.name}" CHECK (${t.expression})`);t.push(n.join(`,
2
- `)),t.push(`);`);let r=[t.join(`
3
- `)];if(e.indexes&&e.indexes.length>0)for(let t of e.indexes)r.push(this.generateIndex(e.name,t));return r.join(`
4
-
5
- `)}generateDropTable(e){return`DROP TABLE IF EXISTS "${e}" CASCADE;`}generateColumnDefinition(e){let t=[`"${e.name}"`];if(t.push(this.getColumnType(e)),e.primaryKey&&!e.type.includes(`serial`)&&t.push(`PRIMARY KEY`),e.notNull&&!e.primaryKey&&!e.type.includes(`serial`)&&t.push(`NOT NULL`),e.unique&&t.push(`UNIQUE`),e.default!==void 0&&t.push(`DEFAULT ${this.formatDefault(e.default)}`),e.references){let n=e.references;t.push(`REFERENCES "${n.table}"("${n.column}")`),n.onDelete&&t.push(`ON DELETE ${n.onDelete}`),n.onUpdate&&t.push(`ON UPDATE ${n.onUpdate}`)}return e.check&&t.push(`CHECK (${e.check})`),t.join(` `)}getColumnType(e){let t=e.type.toUpperCase();return e.length?t+=`(${e.length})`:e.precision&&(e.scale===void 0?t+=`(${e.precision})`:t+=`(${e.precision}, ${e.scale})`),t}formatDefault(e){return typeof e==`string`?e.toUpperCase()===`NOW()`||e.toUpperCase()===`CURRENT_TIMESTAMP`?e.toUpperCase():`'${e}'`:typeof e==`boolean`?e?`TRUE`:`FALSE`:String(e)}generateIndex(e,t){let n=t.unique?`UNIQUE `:``,r=t.type?` USING ${t.type.toUpperCase()}`:``,i=t.columns.map(e=>`"${e}"`).join(`, `);return`CREATE ${n}INDEX "${t.name}" ON "${e}"${r} (${i});`}}}));export{a,i,u as n,s as o,c as r,l as t};
@@ -1 +0,0 @@
1
- import{n as e,t}from"./postgres-9C7eE0wB.js";e();export{t as PostgresGenerator};
@@ -1,59 +0,0 @@
1
- import{a as e,i as t,n,o as r,r as i,t as a}from"./postgres-9C7eE0wB.js";import o from"consola";import s from"chalk";import c from"node:fs/promises";import l from"postgres";import u from"node:path";function d(e){return`transaction`in e&&typeof e.transaction==`function`}function f(e){return`prepare`in e&&typeof e.prepare==`function`}var p=class extends Error{constructor(e,t,n){super(e),this.driverType=t,this.originalError=n,this.name=`DatabaseError`}},m=class extends p{constructor(e,t,n){super(`Query failed: ${t}`,e,n),this.name=`QueryError`}};const h=(e,t=3e4)=>Promise.race([e,new Promise((e,n)=>setTimeout(()=>n(Error(`Query timeout after ${t}ms`)),t))]),g=async(e,t=3,n=1e3)=>{for(let r=1;r<=t;r++)try{return await e()}catch(e){if(r===t)throw e;let i=n*2**(r-1);o.warn(`Query attempt ${r} failed, retrying in ${i}ms...`,e),await new Promise(e=>setTimeout(e,i))}throw Error(`Max retries exceeded`)},_=(e,t=!1,n=`postgres`)=>{let r=t?` | null`:``;switch(e.toLowerCase()){case`uuid`:return`string${r}`;case`character varying`:case`varchar`:case`text`:case`char`:case`character`:case`longtext`:case`mediumtext`:case`tinytext`:return`string${r}`;case`integer`:case`int`:case`int4`:case`smallint`:case`int2`:case`bigint`:case`int8`:case`serial`:case`bigserial`:case`numeric`:case`decimal`:case`real`:case`float4`:case`double precision`:case`float8`:case`tinyint`:case`mediumint`:case`float`:case`double`:return`number${r}`;case`boolean`:case`bool`:case`bit`:return`boolean${r}`;case`timestamp`:case`timestamp with time zone`:case`timestamp without time zone`:case`timestamptz`:case`date`:case`time`:case`time with time zone`:case`time without time zone`:case`timetz`:case`interval`:case`datetime`:case`year`:return`Date${r}`;case`json`:case`jsonb`:return`any${r}`;case`array`:return`any[]${r}`;case`bytea`:case`binary`:case`varbinary`:case`blob`:case`longblob`:case`mediumblob`:case`tinyblob`:return`Buffer${r}`;case`enum`:case`set`:return`string${r}`;default:return o.warn(`Unknown ${n} type: ${e}, defaulting to 'any'`),`any${r}`}},v=e=>`${e.constructor.name||`unknown driver`}`;var y=async e=>{o.warn(`inspectDB is experimental and may make mistakes when inspecting your database. However it will not destroy your database.`);let{driver:t,outputFile:n=`db-types.ts`}=e,r=v(t);o.start(`Inspecting database using ${r} driver`);let i=new $({driver:t}),a=``;try{let e,t;if(r===`mysql`){let n=(await h(g(()=>i.query("SELECT DATABASE() as `database`",[])),1e4)).rows[0]?.database;if(!n)throw Error(`Could not determine current MySQL database name`);o.success(`Using MySQL database: ${n}`),e=`SELECT TABLE_NAME as table_name
2
- FROM information_schema.tables
3
- WHERE TABLE_SCHEMA = ?
4
- AND TABLE_TYPE = 'BASE TABLE'
5
- ORDER BY TABLE_NAME`,t=n}else if(r===`postgres`||r===`neon`)e=`SELECT table_name
6
- FROM information_schema.tables
7
- WHERE table_schema = $1
8
- AND table_type = 'BASE TABLE'
9
- ORDER BY table_name`,t=`public`;else if(r===`libsql`||r===`sqlite`||r===`sqlitecloud`)e=`SELECT name as table_name
10
- FROM sqlite_master
11
- WHERE type = 'table'
12
- ORDER BY name`,t=void 0;else throw Error(`Unsupported driver type: ${r}`);let l=await h(g(()=>i.query(e,t?[t]:[])),3e4);o.info(`Tables in the database:`,l.rows.map(e=>e.table_name).join(`, `));let u=0,d=l.rows.length;for(let e of l.rows){let n=e.table_name;if(!(n.startsWith(`sqlite_sequence`)||n.startsWith(`_prisma_migrations`))){u++,o.info(`[${u}/${d}] Inspecting table: ${n}`);try{let e,c;r===`mysql`?(e=`
13
- SELECT
14
- COLUMN_NAME as column_name,
15
- DATA_TYPE as data_type,
16
- IS_NULLABLE as is_nullable,
17
- COLUMN_DEFAULT as column_default
18
- FROM information_schema.columns
19
- WHERE TABLE_NAME = ?
20
- AND TABLE_SCHEMA = ?
21
- ORDER BY ORDINAL_POSITION
22
- `,c=[n,t]):r===`postgres`||r===`neon`?(e=`
23
- SELECT
24
- column_name,
25
- data_type,
26
- is_nullable,
27
- column_default
28
- FROM information_schema.columns
29
- WHERE table_name = $1
30
- AND table_schema = $2
31
- ORDER BY ordinal_position
32
- `,c=[n,t]):(e=`
33
- SELECT
34
- name as column_name,
35
- type as data_type,
36
- CASE WHEN "notnull" = 0 THEN 'YES' ELSE 'NO' END as is_nullable,
37
- dflt_value as column_default
38
- FROM pragma_table_info(?)
39
- ORDER BY cid
40
- `,c=[n]);let l=await h(g(()=>i.query(e,c)),15e3);if(l.rows.length===0){o.info(`No columns found for table: ${n}`);continue}o.info(`Columns in '${n}'`),l.rows.forEach(e=>{let t=e.data_type+(e.is_nullable===`YES`?` (nullable)`:``);o.info(` > ${s.bold.yellow(t)} > ${e.column_name}`)});let u=new Map;l.rows.forEach(e=>{u.has(e.column_name)||u.set(e.column_name,e)}),a+=`export interface ${n.charAt(0).toUpperCase()+n.slice(1)} {\n`;for(let e of u.values()){let t=_(e.data_type,e.is_nullable===`YES`,r);a+=` ${e.column_name}: ${t};\n`}a+=`}
41
-
42
- `}catch(e){o.error(`Failed to process table ${n}:`,e),o.info(`Skipping table ${n} and continuing...`);continue}}}a+=`export interface Database {
43
- `;for(let e of l.rows){if(e.table_name.startsWith(`sqlite_sequence`)||e.table_name.startsWith(`_prisma_migrations`))continue;let t=e.table_name.charAt(0).toUpperCase()+e.table_name.slice(1);a+=` ${e.table_name}: ${t};\n`}a+=`}
44
-
45
- `,await c.writeFile(n,a,`utf8`),o.success(`TypeScript types written to ${n}`),o.success(`Successfully processed ${u} tables`)}catch(e){throw o.error(`Fatal error during database inspection:`,e),e}finally{await i.close().catch(e=>o.error(`Error closing client:`,e))}};function b(e){return{async findFirst(t,n){let r=[],i=[],a=1;if(n)for(let[e,t]of Object.entries(n))r.push(`"${e}" = $${a}`),i.push(t),a++;let o=`SELECT * FROM "${t}" ${r.length>0?`WHERE ${r.join(` AND `)}`:``} LIMIT 1`;return await e.query(o,i)},async findMany(t,n){let r=[],i=[],a=1;if(n?.where)for(let[e,t]of Object.entries(n.where))r.push(`"${e}" = $${a}`),i.push(t),a++;let o=r.length>0?`WHERE ${r.join(` AND `)}`:``,s=n?.limit?`LIMIT $${a++}`:``,c=n?.offset?`OFFSET $${a++}`:``;n?.limit&&i.push(n.limit),n?.offset&&i.push(n.offset);let l=`SELECT * FROM "${t}" ${o} ${s} ${c}`.trim();return await e.query(l,i)},async insert(t,n){let r=Object.keys(n),i=Object.values(n),a=`INSERT INTO "${t}" (${r.map(e=>`"${e}"`).join(`, `)}) VALUES (${r.map((e,t)=>`$${t+1}`).join(`, `)}) RETURNING *`;return await e.query(a,i)},async update(t,n,r){let i=Object.entries(n),a=Object.entries(r),o=[],s=1,c=i.map(([e,t])=>(o.push(t),`"${e}" = $${s++}`)),l=a.map(([e,t])=>(o.push(t),`"${e}" = $${s++}`)),u=`UPDATE "${t}" SET ${c.join(`, `)} WHERE ${l.join(` AND `)} RETURNING *`;return await e.query(u,o)},async delete(t,n){let r=[],i=[],a=1;for(let[e,t]of Object.entries(n))r.push(`"${e}" = $${a}`),i.push(t),a++;let o=`DELETE FROM "${t}" WHERE ${r.join(` AND `)}`;return(await e.query(o,i)).rowCount}}}var x=class{client;findFirst;findMany;insert;update;delete;constructor(e){this.client=l(e.connectionString,{max:e.max??10,idle_timeout:e.idle_timeout??30,connect_timeout:e.connect_timeout??10,prepare:!1,transform:{undefined:null}});let t=b(this);this.findFirst=t.findFirst,this.findMany=t.findMany,this.insert=t.insert,this.update=t.update,this.delete=t.delete}async query(e,t){try{let n=await this.client.unsafe(e,t||[]),r=Array.isArray(n)?[...n]:[];return{rows:r,rowCount:r.length,command:n.command}}catch(t){throw new m(`postgres`,e,t)}}async close(){await this.client.end()}};function S(e){return{async findFirst(t,n){let r=[],i=[];if(n)for(let[e,t]of Object.entries(n))r.push(`\`${e}\` = ?`),i.push(t);let a=`SELECT * FROM \`${t}\` ${r.length>0?`WHERE ${r.join(` AND `)}`:``} LIMIT 1`;return await e.query(a,i)},async findMany(t,n){let r=[],i=[];if(n?.where)for(let[e,t]of Object.entries(n.where))r.push(`\`${e}\` = ?`),i.push(t);let a=r.length>0?`WHERE ${r.join(` AND `)}`:``,o=n?.limit?`LIMIT ?`:``,s=n?.offset?`OFFSET ?`:``;n?.limit&&i.push(n.limit),n?.offset&&i.push(n.offset);let c=`SELECT * FROM \`${t}\` ${a} ${o} ${s}`.trim();return await e.query(c,i)},async insert(t,n){let r=Object.keys(n),i=Object.values(n),a=`INSERT INTO \`${t}\` (${r.map(e=>`\`${e}\``).join(`, `)}) VALUES (${r.map(()=>`?`).join(`, `)})`;await e.query(a,i);let o=`SELECT * FROM \`${t}\` WHERE id = LAST_INSERT_ID()`;return await e.query(o)},async update(t,n,r){let i=Object.entries(n),a=Object.entries(r),o=[],s=i.map(([e,t])=>(o.push(t),`\`${e}\` = ?`)),c=a.map(([e,t])=>(o.push(t),`\`${e}\` = ?`)),l=`UPDATE \`${t}\` SET ${s.join(`, `)} WHERE ${c.join(` AND `)}`;await e.query(l,o);let u=Object.values(r),d=`SELECT * FROM \`${t}\` WHERE ${Object.keys(r).map(e=>`\`${e}\` = ?`).join(` AND `)}`;return await e.query(d,u)},async delete(t,n){let r=[],i=[];for(let[e,t]of Object.entries(n))r.push(`\`${e}\` = ?`),i.push(t);let a=`DELETE FROM \`${t}\` WHERE ${r.join(` AND `)}`;return(await e.query(a,i)).rowCount}}}function C(e){return{async findFirst(t,n){let r=[],i=[];if(n)for(let[e,t]of Object.entries(n))r.push(`"${e}" = ?`),i.push(t);let a=`SELECT * FROM "${t}" ${r.length>0?`WHERE ${r.join(` AND `)}`:``} LIMIT 1`;return await e.query(a,i)},async findMany(t,n){let r=[],i=[];if(n?.where)for(let[e,t]of Object.entries(n.where))r.push(`"${e}" = ?`),i.push(t);let a=r.length>0?`WHERE ${r.join(` AND `)}`:``,o=n?.limit?`LIMIT ?`:``,s=n?.offset?`OFFSET ?`:``;n?.limit&&i.push(n.limit),n?.offset&&i.push(n.offset);let c=`SELECT * FROM "${t}" ${a} ${o} ${s}`.trim();return await e.query(c,i)},async insert(t,n){let r=Object.keys(n),i=Object.values(n),a=`INSERT INTO "${t}" (${r.map(e=>`"${e}"`).join(`, `)}) VALUES (${r.map(()=>`?`).join(`, `)}) RETURNING *`;return await e.query(a,i)},async update(t,n,r){let i=Object.entries(n),a=Object.entries(r),o=[],s=i.map(([e,t])=>(o.push(t),`"${e}" = ?`)),c=a.map(([e,t])=>(o.push(t),`"${e}" = ?`)),l=`UPDATE "${t}" SET ${s.join(`, `)} WHERE ${c.join(` AND `)} RETURNING *`;return await e.query(l,o)},async delete(t,n){let r=[],i=[];for(let[e,t]of Object.entries(n))r.push(`"${e}" = ?`),i.push(t);let a=`DELETE FROM "${t}" WHERE ${r.join(` AND `)}`;return(await e.query(a,i)).rowCount}}}var w=class{definition;constructor(e,t){this.definition={name:e,type:t}}length(e){return this.definition.length=e,this}precision(e,t){return this.definition.precision=e,t!==void 0&&(this.definition.scale=t),this}primaryKey(){return this.definition.primaryKey=!0,this.definition.notNull=!0,this}notNull(){return this.definition.notNull=!0,this}unique(){return this.definition.unique=!0,this}default(e){return this.definition.default=e,this}references(e,t=`id`){return this.definition.references={table:e,column:t},this}onDelete(e){return this.definition.references&&(this.definition.references.onDelete=e),this}onUpdate(e){return this.definition.references&&(this.definition.references.onUpdate=e),this}check(e){return this.definition.check=e,this}getDefinition(){return this.definition}};function ee(e){return new w(e,`serial`)}function te(e){return new w(e,`bigserial`)}function T(e){return new w(e,`integer`)}function E(e){return new w(e,`bigint`)}function D(e){return new w(e,`smallint`)}function O(e){return new w(e,`text`)}function k(e,t){let n=new w(e,`varchar`);return t&&n.length(t),n}function A(e,t){let n=new w(e,`char`);return t&&n.length(t),n}function j(e){return new w(e,`boolean`)}function M(e){return new w(e,`timestamp`)}function N(e){return new w(e,`timestamptz`)}function P(e){return new w(e,`date`)}function F(e){return new w(e,`time`)}function I(e){return new w(e,`json`)}function L(e){return new w(e,`jsonb`)}function R(e){return new w(e,`uuid`)}function z(e,t,n){let r=new w(e,`decimal`);return t&&r.precision(t,n),r}function B(e,t,n){let r=new w(e,`numeric`);return t&&r.precision(t,n),r}function ne(e){return new w(e,`real`)}function V(e){return new w(e,`double precision`)}function H(e){return new w(e,`bytea`)}var U=class{definition;constructor(e){this.definition={name:e,columns:[],indexes:[],checks:[]}}column(e){return this.definition.columns.push(e.getDefinition()),this}primaryKey(...e){return this.definition.primaryKey=e,this}index(e,t,n){let r={name:e,columns:t,unique:n?.unique,type:n?.type};return this.definition.indexes?.push(r),this}unique(e,...t){return this.index(e,t,{unique:!0})}check(e,t){return this.definition.checks?.push({name:e,expression:t}),this}getDefinition(){return this.definition}};function W(e,t){let n=new U(e);return t(n),n}var G=e({MySQLGenerator:()=>K}),K,q=t((()=>{K=class{generateCreateTable(e){let t=[];t.push(`CREATE TABLE \`${e.name}\` (`);let n=[];for(let t of e.columns)n.push(` `+this.generateColumnDefinition(t));if(e.primaryKey&&e.primaryKey.length>0){let t=e.primaryKey.map(e=>`\`${e}\``).join(`, `);n.push(` PRIMARY KEY (${t})`)}if(e.checks&&e.checks.length>0)for(let t of e.checks)n.push(` CONSTRAINT \`${t.name}\` CHECK (${t.expression})`);t.push(n.join(`,
46
- `)),t.push(`) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_unicode_ci;`);let r=[t.join(`
47
- `)];if(e.indexes&&e.indexes.length>0)for(let t of e.indexes)r.push(this.generateIndex(e.name,t));return r.join(`
48
-
49
- `)}generateDropTable(e){return`DROP TABLE IF EXISTS \`${e}\`;`}generateColumnDefinition(e){let t=[`\`${e.name}\``];if(t.push(this.getColumnType(e)),e.notNull&&!e.primaryKey&&e.type!==`serial`&&e.type!==`bigserial`&&t.push(`NOT NULL`),e.unique&&t.push(`UNIQUE`),(e.type===`serial`||e.type===`bigserial`)&&t.push(`AUTO_INCREMENT`),e.default!==void 0&&t.push(`DEFAULT ${this.formatDefault(e.default)}`),e.references){let n=e.references;t.push(`REFERENCES \`${n.table}\`(\`${n.column}\`)`),n.onDelete&&t.push(`ON DELETE ${n.onDelete}`),n.onUpdate&&t.push(`ON UPDATE ${n.onUpdate}`)}return e.check&&t.push(`CHECK (${e.check})`),t.join(` `)}getColumnType(e){let t=e.type;if(t===`serial`)return`INT`;if(t===`bigserial`)return`BIGINT`;if(t===`timestamptz`)return`TIMESTAMP`;if(t===`jsonb`)return`JSON`;if(t===`bytea`)return`BLOB`;if(t===`double precision`)return`DOUBLE`;let n=t.toUpperCase();return e.length?n+=`(${e.length})`:e.precision&&(e.scale===void 0?n+=`(${e.precision})`:n+=`(${e.precision}, ${e.scale})`),n}formatDefault(e){return typeof e==`string`?e.toUpperCase()===`NOW()`||e.toUpperCase()===`CURRENT_TIMESTAMP`?`CURRENT_TIMESTAMP`:`'${e}'`:typeof e==`boolean`?e?`1`:`0`:String(e)}generateIndex(e,t){let n=t.unique?`UNIQUE `:``,r=t.type&&t.type!==`btree`?` USING ${t.type.toUpperCase()}`:``,i=t.columns.map(e=>`\`${e}\``).join(`, `);return`CREATE ${n}INDEX \`${t.name}\` ON \`${e}\`${r} (${i});`}}})),J=e({SQLiteGenerator:()=>Y}),Y,X=t((()=>{Y=class{generateCreateTable(e){let t=[];t.push(`CREATE TABLE "${e.name}" (`);let n=[];for(let t of e.columns)n.push(` `+this.generateColumnDefinition(t));if(e.primaryKey&&e.primaryKey.length>0){let t=e.primaryKey.map(e=>`"${e}"`).join(`, `);n.push(` PRIMARY KEY (${t})`)}if(e.checks&&e.checks.length>0)for(let t of e.checks)n.push(` CONSTRAINT "${t.name}" CHECK (${t.expression})`);t.push(n.join(`,
50
- `)),t.push(`);`);let r=[t.join(`
51
- `)];if(e.indexes&&e.indexes.length>0)for(let t of e.indexes)r.push(this.generateIndex(e.name,t));return r.join(`
52
-
53
- `)}generateDropTable(e){return`DROP TABLE IF EXISTS "${e}";`}generateColumnDefinition(e){let t=[`"${e.name}"`];if(t.push(this.getColumnType(e)),e.primaryKey&&(t.push(`PRIMARY KEY`),(e.type===`serial`||e.type===`bigserial`)&&t.push(`AUTOINCREMENT`)),e.notNull&&!e.primaryKey&&t.push(`NOT NULL`),e.unique&&t.push(`UNIQUE`),e.default!==void 0&&t.push(`DEFAULT ${this.formatDefault(e.default)}`),e.references){let n=e.references;t.push(`REFERENCES "${n.table}"("${n.column}")`),n.onDelete&&t.push(`ON DELETE ${n.onDelete}`),n.onUpdate&&t.push(`ON UPDATE ${n.onUpdate}`)}return e.check&&t.push(`CHECK (${e.check})`),t.join(` `)}getColumnType(e){let t=e.type;return t===`serial`||t===`bigserial`||t===`bigint`||t===`smallint`?`INTEGER`:t===`varchar`||t===`char`||t===`timestamptz`||t===`timestamp`||t===`date`||t===`time`||t===`uuid`||t===`json`||t===`jsonb`?`TEXT`:t===`bytea`?`BLOB`:t===`decimal`||t===`numeric`||t===`real`||t===`double precision`?`REAL`:t===`boolean`?`INTEGER`:t.toUpperCase()}formatDefault(e){return typeof e==`string`?e.toUpperCase()===`NOW()`||e.toUpperCase()===`CURRENT_TIMESTAMP`?`CURRENT_TIMESTAMP`:`'${e}'`:typeof e==`boolean`?e?`1`:`0`:String(e)}generateIndex(e,t){let n=t.unique?`UNIQUE `:``,r=t.columns.map(e=>`"${e}"`).join(`, `);return`CREATE ${n}INDEX "${t.name}" ON "${e}" (${r});`}}}));n(),q(),X();var Z=class{upStatements=[];downStatements=[];dialect;constructor(e=`postgres`){this.dialect=e}createTable(e){let t=e.getDefinition(),n=this.getGenerator();return this.upStatements.push(n.generateCreateTable(t)),this.downStatements.unshift(n.generateDropTable(t.name)),this}dropTable(e){let t=this.getGenerator();return this.upStatements.push(t.generateDropTable(e)),this}addColumn(e,t){return this.upStatements.push(`ALTER TABLE "${e}" ADD COLUMN ${t};`),this}dropColumn(e,t){return this.upStatements.push(`ALTER TABLE "${e}" DROP COLUMN "${t}";`),this}renameTable(e,t){return this.dialect===`postgres`?this.upStatements.push(`ALTER TABLE "${e}" RENAME TO "${t}";`):this.dialect===`mysql`?this.upStatements.push(`RENAME TABLE \`${e}\` TO \`${t}\`;`):this.upStatements.push(`ALTER TABLE "${e}" RENAME TO "${t}";`),this}raw(e,t){return this.upStatements.push(e),t&&this.downStatements.unshift(t),this}build(e,t){return{version:e,name:t,up:this.upStatements,down:this.downStatements}}getGenerator(){switch(this.dialect){case`postgres`:return new a;case`mysql`:return new K;case`sqlite`:return new Y;default:return new a}}};function re(e=`postgres`){return new Z(e)}var ie=class{client;constructor(e){this.client=e}async ensureMigrationsTable(){await this.client.query(`
54
- CREATE TABLE IF NOT EXISTS _migrations (
55
- version VARCHAR(255) PRIMARY KEY,
56
- name VARCHAR(255) NOT NULL,
57
- applied_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
58
- );
59
- `)}async getAppliedMigrations(){return await this.ensureMigrationsTable(),(await this.client.query(`SELECT version FROM _migrations ORDER BY version`)).rows.map(e=>e.version)}async up(e){if((await this.getAppliedMigrations()).includes(e.version)){o.info(`Migration ${e.version} (${e.name}) already applied`);return}o.start(`Running migration ${e.version} (${e.name})`);for(let t of e.up)await this.client.query(t);await this.client.query(`INSERT INTO _migrations (version, name) VALUES ($1, $2)`,[e.version,e.name]),o.success(`Migration ${e.version} (${e.name}) applied successfully`)}async down(e){if(!(await this.getAppliedMigrations()).includes(e.version)){o.info(`Migration ${e.version} (${e.name}) not applied`);return}o.start(`Rolling back migration ${e.version} (${e.name})`);for(let t of e.down)await this.client.query(t);await this.client.query(`DELETE FROM _migrations WHERE version = $1`,[e.version]),o.success(`Migration ${e.version} (${e.name}) rolled back successfully`)}async upAll(e){let t=await this.getAppliedMigrations(),n=e.filter(e=>!t.includes(e.version));if(n.length===0){o.info(`No pending migrations`);return}o.info(`Found ${n.length} pending migrations`);for(let e of n)await this.up(e)}async downAll(e){let t=await this.getAppliedMigrations(),n=e.filter(e=>t.includes(e.version)).reverse();if(n.length===0){o.info(`No migrations to rollback`);return}o.info(`Rolling back ${n.length} migrations`);for(let e of n)await this.down(e)}async reset(e){await this.downAll(e),await this.upAll(e)}},Q=class{dialect;constructor(e=`postgres`){this.dialect=e}detectChanges(e,t){let n=[],r=new Map(e.map(e=>[e.name,e])),i=new Map(t.map(e=>[e.name,e]));for(let[e,t]of i){let i=r.get(e);i?(n.push(...this.detectColumnChanges(i,t)),n.push(...this.detectIndexChanges(i,t))):n.push({type:`create_table`,table:e,details:t})}for(let[e]of r)i.has(e)||n.push({type:`drop_table`,table:e});return n}detectColumnChanges(e,t){let n=[],r=new Map(e.columns.map(e=>[e.name,e])),i=new Map(t.columns.map(e=>[e.name,e]));for(let[e,a]of i){let i=r.get(e);i?this.hasColumnChanged(i,a)&&n.push({type:`modify_column`,table:t.name,details:{old:i,new:a}}):n.push({type:`add_column`,table:t.name,details:a})}for(let[e]of r)i.has(e)||n.push({type:`drop_column`,table:t.name,details:{name:e}});return n}detectIndexChanges(e,t){let n=[],r=new Map((e.indexes||[]).map(e=>[e.name,e])),i=new Map((t.indexes||[]).map(e=>[e.name,e]));for(let[e,a]of i)r.has(e)||n.push({type:`create_index`,table:t.name,details:a});for(let[e]of r)i.has(e)||n.push({type:`drop_index`,table:t.name,details:{name:e}});return n}hasColumnChanged(e,t){return e.type!==t.type||e.length!==t.length||e.precision!==t.precision||e.scale!==t.scale||e.notNull!==t.notNull||e.unique!==t.unique||e.default!==t.default||JSON.stringify(e.references)!==JSON.stringify(t.references)}generateMigration(e,t,n){let r=new Z(this.dialect);for(let t of e)switch(t.type){case`create_table`:{let e=t.details,n=this.getGenerator(),i=n.generateCreateTable(e);r.raw(i,n.generateDropTable(e.name));break}case`drop_table`:{let e=this.getGenerator();r.raw(e.generateDropTable(t.table));break}case`add_column`:{let e=t.details,n=this.generateColumnSQL(e);r.raw(`ALTER TABLE "${t.table}" ADD COLUMN ${n};`,`ALTER TABLE "${t.table}" DROP COLUMN "${e.name}";`);break}case`drop_column`:{let e=t.details.name;r.raw(`ALTER TABLE "${t.table}" DROP COLUMN "${e}";`);break}case`create_index`:{let e=t.details,n=this.generateIndexSQL(t.table,e);r.raw(n,`DROP INDEX "${e.name}";`);break}case`drop_index`:{let e=t.details.name;r.raw(`DROP INDEX "${e}";`);break}}return r}generateColumnSQL(e){let t=[`"${e.name}"`];return t.push(e.type.toUpperCase()),e.length&&(t[1]+=`(${e.length})`),e.notNull&&t.push(`NOT NULL`),e.unique&&t.push(`UNIQUE`),e.default!==void 0&&t.push(`DEFAULT ${this.formatDefault(e.default)}`),t.join(` `)}generateIndexSQL(e,t){let n=t.unique?`UNIQUE `:``,r=t.columns.map(e=>`"${e}"`).join(`, `);return`CREATE ${n}INDEX "${t.name}" ON "${e}" (${r});`}formatDefault(e){return typeof e==`string`?e.toUpperCase()===`NOW()`||e.toUpperCase()===`CURRENT_TIMESTAMP`?e.toUpperCase():`'${e}'`:typeof e==`boolean`?e?`TRUE`:`FALSE`:String(e)}getGenerator(){let{PostgresGenerator:e}=(n(),r(i)),{MySQLGenerator:t}=(q(),r(G)),{SQLiteGenerator:a}=(X(),r(J));switch(this.dialect){case`postgres`:return new e;case`mysql`:return new t;case`sqlite`:return new a;default:return new e}}};function ae(e,t,n=`postgres`){return new Q(n).detectChanges(e,t)}function oe(e,t,n,r=`postgres`){return new Q(r).generateMigration(e,t,n)}var se=class{snapshotPath;constructor(e=`./.driftsql/snapshot.json`){this.snapshotPath=e}async save(e){let t={version:`1`,timestamp:Date.now(),tables:e},n=u.dirname(this.snapshotPath);await c.mkdir(n,{recursive:!0}),await c.writeFile(this.snapshotPath,JSON.stringify(t,null,2),`utf8`)}async load(){try{let e=await c.readFile(this.snapshotPath,`utf8`);return JSON.parse(e)}catch{return null}}async exists(){try{return await c.access(this.snapshotPath),!0}catch{return!1}}};n(),q(),X();var $=class e{primaryDriver;fallbackDrivers;queryCache;cacheTTL;maxCacheSize=100;constructor(e){this.primaryDriver=e.driver,this.fallbackDrivers=e.fallbackDrivers||[],this.queryCache=new Map,this.cacheTTL=e.cacheTTL??5e3}isCacheValid(e){return Date.now()-e.timestamp<this.cacheTTL}getCacheKey(e,t){return JSON.stringify({sql:e,params:t})}maintainCacheSize(){if(this.queryCache.size>this.maxCacheSize){let e=this.queryCache.keys().next().value;e!==void 0&&this.queryCache.delete(e)}}cacheResult(e,t){this.queryCache.set(e,{result:t,timestamp:Date.now()}),this.maintainCacheSize()}async tryFallbackDrivers(e,t){o.warn(`Query failed with ${this.primaryDriver.constructor.name}, trying fallbacks`);let n;for(let r of this.fallbackDrivers)try{return await r.query(e,t)}catch(e){n=e}throw n||new p(`All drivers failed to execute query`,`unknown`)}async query(e,t){let n=this.getCacheKey(e,t),r=this.queryCache.get(n);if(r&&this.isCacheValid(r))return r.result;try{let r=await this.primaryDriver.query(e,t);return this.cacheResult(n,r),r}catch(n){if(this.fallbackDrivers.length===0)throw n;return await this.tryFallbackDrivers(e,t)}}async transaction(t){if(!d(this.primaryDriver))throw new p(`Primary driver does not support transactions`,this.primaryDriver.constructor.name);return await this.primaryDriver.transaction(async n=>await t(new e({driver:n,fallbackDrivers:[]})))}async prepare(e){if(!f(this.primaryDriver))throw new p(`Primary driver does not support prepared statements`,this.primaryDriver.constructor.name);return await this.primaryDriver.prepare(e)}async findFirst(e,t){if(!this.primaryDriver.findFirst)throw new p(`Primary driver does not support findFirst`,this.primaryDriver.constructor.name);return(await this.primaryDriver.findFirst(e,t))?.rows[0]||null}async findMany(e,t){if(!this.primaryDriver.findMany)throw new p(`Primary driver does not support findMany`,this.primaryDriver.constructor.name);return(await this.primaryDriver.findMany(e,t)).rows}async insert(e,t){if(!this.primaryDriver.insert)throw new p(`Primary driver does not support insert`,this.primaryDriver.constructor.name);return(await this.primaryDriver.insert(e,t)).rows[0]}async update(e,t,n){if(!this.primaryDriver.update)throw new p(`Primary driver does not support update`,this.primaryDriver.constructor.name);return(await this.primaryDriver.update(e,t,n)).rows[0]||null}async delete(e,t){if(!this.primaryDriver.delete)throw new p(`Primary driver does not support delete`,this.primaryDriver.constructor.name);return await this.primaryDriver.delete(e,t)}getDriver(){return this.primaryDriver}async inspectDB(e){return await y({driver:this.getDriver(),...e})}supportsTransactions(){return d(this.primaryDriver)}supportsPreparedStatements(){return f(this.primaryDriver)}clearCache(){this.queryCache.clear()}getCacheStats(){return{size:this.queryCache.size,ttl:this.cacheTTL}}async close(){this.queryCache.clear();let e=[this.primaryDriver,...this.fallbackDrivers].map(e=>e.close().catch(t=>o.warn(`Error closing ${e.constructor.name}:`,t)));await Promise.all(e)}};const ce=$;export{O as A,T as C,ne as D,B as E,k as F,C as I,S as L,M,N,ee as O,R as P,x as R,V as S,L as T,j as _,ae as a,P as b,Z as c,K as d,U as f,te as g,E as h,Q as i,F as j,D as k,re as l,w as m,$ as n,oe as o,W as p,se as r,ie as s,ce as t,Y as u,H as v,I as w,z as x,A as y,b as z};
@@ -1,7 +0,0 @@
1
- import e from"node:fs/promises";var t=class{generateTypes(e){let t=`// This file was automatically generated - DO NOT EDIT
2
-
3
- `;for(let n of e)t+=this.generateTableInterface(n),t+=`
4
-
5
- `;return t+=this.generateDatabaseInterface(e),t}generateTableInterface(e){let t=[`export interface ${this.pascalCase(e.name)} {`];for(let n of e.columns){let e=this.columnToTypeScript(n),r=!n.notNull&&!n.primaryKey?`?`:``;t.push(` ${n.name}${r}: ${e}`)}return t.push(`}`),t.join(`
6
- `)}generateDatabaseInterface(e){let t=[`export interface Database {`];for(let n of e){let e=this.pascalCase(n.name);t.push(` ${n.name}: ${e}`)}return t.push(`}`),t.join(`
7
- `)}columnToTypeScript(e){let t=!e.notNull&&!e.primaryKey,n;switch(e.type.toLowerCase()){case`serial`:case`bigserial`:case`integer`:case`int`:case`bigint`:case`smallint`:case`numeric`:case`decimal`:case`real`:case`double precision`:n=`number`;break;case`text`:case`varchar`:case`char`:case`uuid`:n=`string`;break;case`boolean`:n=`boolean`;break;case`timestamp`:case`timestamptz`:case`date`:case`time`:n=`Date | string`;break;case`json`:case`jsonb`:n=`any`;break;case`bytea`:n=`Buffer`;break;default:n=`any`}return t?`${n} | null`:n}pascalCase(e){return e.split(`_`).map(e=>e.charAt(0).toUpperCase()+e.slice(1).toLowerCase()).join(``)}async writeToFile(t,n){let r=this.generateTypes(t);await e.writeFile(n,r,`utf8`)}};async function n(e,n){await new t().writeToFile(e,n)}export{n,t};
@@ -1 +0,0 @@
1
- import{n as e,t}from"./type-generator-CpqI7vBb.js";export{t as TypeGenerator,e as generateTypesFromSchema};