@hedystia/db 2.0.7 → 2.0.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -8,13 +8,14 @@ let path = require("path");
8
8
  * Create a new migration file
9
9
  * @param {string} name - Migration name
10
10
  * @param {string} [path] - Output directory
11
+ * @param {boolean} [noId] - Omit timestamp from variable name and migration id
11
12
  */
12
- function createMigration(name, path$1) {
13
+ function createMigration(name, path$1, noId) {
13
14
  const dir = path$1 ?? "src/database/migrations";
14
15
  require_fs.ensureDir(dir);
15
16
  const timestamp = require_naming.generateTimestamp();
16
17
  const filePath = (0, path.join)(dir, `${timestamp}_${name}.ts`);
17
- require_fs.writeFileSafe(filePath, require_templates.generateMigrationTemplate(`${timestamp}_${name}`));
18
+ require_fs.writeFileSafe(filePath, require_templates.generateMigrationTemplate(noId ? name : `${timestamp}_${name}`, noId ? require_naming.toCamelCase(name) : `${require_naming.toCamelCase(name)}${timestamp}`));
18
19
  console.log(`Created migration: ${filePath}`);
19
20
  }
20
21
  //#endregion
@@ -1 +1 @@
1
- {"version":3,"file":"migration.cjs","names":["path","generateTimestamp","generateMigrationTemplate"],"sources":["../../../src/cli/commands/migration.ts"],"sourcesContent":["import { join } from \"path\";\nimport { DEFAULT_MIGRATIONS_PATH } from \"../../constants\";\nimport { generateMigrationTemplate } from \"../../migrations\";\nimport { ensureDir, writeFileSafe } from \"../../utils\";\nimport { generateTimestamp } from \"../../utils/naming\";\n\n/**\n * Create a new migration file\n * @param {string} name - Migration name\n * @param {string} [path] - Output directory\n */\nexport function createMigration(name: string, path?: string): void {\n const dir = path ?? DEFAULT_MIGRATIONS_PATH;\n ensureDir(dir);\n\n const timestamp = generateTimestamp();\n const fileName = `${timestamp}_${name}.ts`;\n const filePath = join(dir, fileName);\n const migrationId = `${timestamp}_${name}`;\n const content = generateMigrationTemplate(migrationId);\n\n writeFileSafe(filePath, content);\n console.log(`Created migration: ${filePath}`);\n}\n"],"mappings":";;;;;;;;;;;AAWA,SAAgB,gBAAgB,MAAc,QAAqB;CACjE,MAAM,MAAMA,UAAAA;AACZ,YAAA,UAAU,IAAI;CAEd,MAAM,YAAYC,eAAAA,mBAAmB;CAErC,MAAM,YAAA,GAAA,KAAA,MAAgB,KADL,GAAG,UAAU,GAAG,KAAK,KACF;AAIpC,YAAA,cAAc,UAFEC,kBAAAA,0BADI,GAAG,UAAU,GAAG,OACkB,CAEtB;AAChC,SAAQ,IAAI,sBAAsB,WAAW"}
1
+ {"version":3,"file":"migration.cjs","names":["path","generateTimestamp","generateMigrationTemplate","toCamelCase"],"sources":["../../../src/cli/commands/migration.ts"],"sourcesContent":["import { join } from \"path\";\nimport { DEFAULT_MIGRATIONS_PATH } from \"../../constants\";\nimport { generateMigrationTemplate } from \"../../migrations\";\nimport { ensureDir, writeFileSafe } from \"../../utils\";\nimport { generateTimestamp, toCamelCase } from \"../../utils/naming\";\n\n/**\n * Create a new migration file\n * @param {string} name - Migration name\n * @param {string} [path] - Output directory\n * @param {boolean} [noId] - Omit timestamp from variable name and migration id\n */\nexport function createMigration(name: string, path?: string, noId?: boolean): void {\n const dir = path ?? DEFAULT_MIGRATIONS_PATH;\n ensureDir(dir);\n\n const timestamp = generateTimestamp();\n const fileName = `${timestamp}_${name}.ts`;\n const filePath = join(dir, fileName);\n const migrationId = noId ? name : `${timestamp}_${name}`;\n const varName = noId ? toCamelCase(name) : `${toCamelCase(name)}${timestamp}`;\n const content = generateMigrationTemplate(migrationId, varName);\n\n writeFileSafe(filePath, content);\n console.log(`Created migration: ${filePath}`);\n}\n"],"mappings":";;;;;;;;;;;;AAYA,SAAgB,gBAAgB,MAAc,QAAe,MAAsB;CACjF,MAAM,MAAMA,UAAAA;AACZ,YAAA,UAAU,IAAI;CAEd,MAAM,YAAYC,eAAAA,mBAAmB;CAErC,MAAM,YAAA,GAAA,KAAA,MAAgB,KADL,GAAG,UAAU,GAAG,KAAK,KACF;AAKpC,YAAA,cAAc,UAFEC,kBAAAA,0BAFI,OAAO,OAAO,GAAG,UAAU,GAAG,QAClC,OAAOC,eAAAA,YAAY,KAAK,GAAG,GAAGA,eAAAA,YAAY,KAAK,GAAG,YACH,CAE/B;AAChC,SAAQ,IAAI,sBAAsB,WAAW"}
@@ -1,7 +1,7 @@
1
1
  import { __esmMin } from "../../_virtual/_rolldown/runtime.mjs";
2
2
  import { init_constants } from "../../constants.mjs";
3
3
  import { ensureDir, writeFileSafe } from "../../utils/fs.mjs";
4
- import { generateTimestamp, init_naming } from "../../utils/naming.mjs";
4
+ import { generateTimestamp, init_naming, toCamelCase } from "../../utils/naming.mjs";
5
5
  import { init_utils } from "../../utils/index.mjs";
6
6
  import { generateMigrationTemplate } from "../../migrations/templates.mjs";
7
7
  import { init_migrations } from "../../migrations/index.mjs";
@@ -11,13 +11,14 @@ import { join } from "path";
11
11
  * Create a new migration file
12
12
  * @param {string} name - Migration name
13
13
  * @param {string} [path] - Output directory
14
+ * @param {boolean} [noId] - Omit timestamp from variable name and migration id
14
15
  */
15
- function createMigration(name, path) {
16
+ function createMigration(name, path, noId) {
16
17
  const dir = path ?? "src/database/migrations";
17
18
  ensureDir(dir);
18
19
  const timestamp = generateTimestamp();
19
20
  const filePath = join(dir, `${timestamp}_${name}.ts`);
20
- writeFileSafe(filePath, generateMigrationTemplate(`${timestamp}_${name}`));
21
+ writeFileSafe(filePath, generateMigrationTemplate(noId ? name : `${timestamp}_${name}`, noId ? toCamelCase(name) : `${toCamelCase(name)}${timestamp}`));
21
22
  console.log(`Created migration: ${filePath}`);
22
23
  }
23
24
  var init_migration = __esmMin((() => {
@@ -1 +1 @@
1
- {"version":3,"file":"migration.mjs","names":[],"sources":["../../../src/cli/commands/migration.ts"],"sourcesContent":["import { join } from \"path\";\nimport { DEFAULT_MIGRATIONS_PATH } from \"../../constants\";\nimport { generateMigrationTemplate } from \"../../migrations\";\nimport { ensureDir, writeFileSafe } from \"../../utils\";\nimport { generateTimestamp } from \"../../utils/naming\";\n\n/**\n * Create a new migration file\n * @param {string} name - Migration name\n * @param {string} [path] - Output directory\n */\nexport function createMigration(name: string, path?: string): void {\n const dir = path ?? DEFAULT_MIGRATIONS_PATH;\n ensureDir(dir);\n\n const timestamp = generateTimestamp();\n const fileName = `${timestamp}_${name}.ts`;\n const filePath = join(dir, fileName);\n const migrationId = `${timestamp}_${name}`;\n const content = generateMigrationTemplate(migrationId);\n\n writeFileSafe(filePath, content);\n console.log(`Created migration: ${filePath}`);\n}\n"],"mappings":";;;;;;;;;;;;;;AAWA,SAAgB,gBAAgB,MAAc,MAAqB;CACjE,MAAM,MAAM,QAAA;AACZ,WAAU,IAAI;CAEd,MAAM,YAAY,mBAAmB;CAErC,MAAM,WAAW,KAAK,KADL,GAAG,UAAU,GAAG,KAAK,KACF;AAIpC,eAAc,UAFE,0BADI,GAAG,UAAU,GAAG,OACkB,CAEtB;AAChC,SAAQ,IAAI,sBAAsB,WAAW;;;iBArBW;kBACG;aACN;cACA"}
1
+ {"version":3,"file":"migration.mjs","names":[],"sources":["../../../src/cli/commands/migration.ts"],"sourcesContent":["import { join } from \"path\";\nimport { DEFAULT_MIGRATIONS_PATH } from \"../../constants\";\nimport { generateMigrationTemplate } from \"../../migrations\";\nimport { ensureDir, writeFileSafe } from \"../../utils\";\nimport { generateTimestamp, toCamelCase } from \"../../utils/naming\";\n\n/**\n * Create a new migration file\n * @param {string} name - Migration name\n * @param {string} [path] - Output directory\n * @param {boolean} [noId] - Omit timestamp from variable name and migration id\n */\nexport function createMigration(name: string, path?: string, noId?: boolean): void {\n const dir = path ?? DEFAULT_MIGRATIONS_PATH;\n ensureDir(dir);\n\n const timestamp = generateTimestamp();\n const fileName = `${timestamp}_${name}.ts`;\n const filePath = join(dir, fileName);\n const migrationId = noId ? name : `${timestamp}_${name}`;\n const varName = noId ? toCamelCase(name) : `${toCamelCase(name)}${timestamp}`;\n const content = generateMigrationTemplate(migrationId, varName);\n\n writeFileSafe(filePath, content);\n console.log(`Created migration: ${filePath}`);\n}\n"],"mappings":";;;;;;;;;;;;;;;AAYA,SAAgB,gBAAgB,MAAc,MAAe,MAAsB;CACjF,MAAM,MAAM,QAAA;AACZ,WAAU,IAAI;CAEd,MAAM,YAAY,mBAAmB;CAErC,MAAM,WAAW,KAAK,KADL,GAAG,UAAU,GAAG,KAAK,KACF;AAKpC,eAAc,UAFE,0BAFI,OAAO,OAAO,GAAG,UAAU,GAAG,QAClC,OAAO,YAAY,KAAK,GAAG,GAAG,YAAY,KAAK,GAAG,YACH,CAE/B;AAChC,SAAQ,IAAI,sBAAsB,WAAW;;;iBAvBW;kBACG;aACN;cACa"}
package/dist/cli.cjs CHANGED
@@ -13,6 +13,9 @@ function getFlag(flag) {
13
13
  function getName() {
14
14
  return getFlag("--name") ?? args[2];
15
15
  }
16
+ function hasFlag(flag) {
17
+ return args.includes(flag);
18
+ }
16
19
  function getMigrateOptions() {
17
20
  return {
18
21
  migrationsPath: getFlag("--migrations"),
@@ -27,11 +30,11 @@ switch (command) {
27
30
  if ((subCommand === "create" ? "create" : "create") === "create") {
28
31
  const name = subCommand === "create" ? getName() : subCommand;
29
32
  if (!name) {
30
- console.log("Usage: @hedystia/db migration create <name> [--path <path>]");
31
- console.log(" @hedystia/db migration <name> [--path <path>]");
33
+ console.log("Usage: @hedystia/db migration create <name> [--path <path>] [--no-id]");
34
+ console.log(" @hedystia/db migration <name> [--path <path>] [--no-id]");
32
35
  process.exit(1);
33
36
  }
34
- require_migration.createMigration(name, getFlag("--path"));
37
+ require_migration.createMigration(name, getFlag("--path"), hasFlag("--no-id"));
35
38
  }
36
39
  break;
37
40
  case "schema":
@@ -73,8 +76,8 @@ switch (command) {
73
76
  console.log("@hedystia/db CLI");
74
77
  console.log("");
75
78
  console.log("Commands:");
76
- console.log(" migration create <name> [--path <path>] Create a new migration file");
77
- console.log(" migration <name> [--path <path>] Create a new migration file (shorthand)");
79
+ console.log(" migration create <name> [--path <path>] [--no-id] Create a new migration file");
80
+ console.log(" migration <name> [--path <path>] [--no-id] Create a new migration file (shorthand)");
78
81
  console.log(" schema create <name> [--path <path>] Create a new schema file");
79
82
  console.log(" schema <name> [--path <path>] Create a new schema file (shorthand)");
80
83
  console.log(" migrate up [options] Run pending migrations");
package/dist/cli.cjs.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"cli.cjs","names":[],"sources":["../src/cli.ts"],"sourcesContent":["#!/usr/bin/env bun\nimport { migrateDown, migrateUp } from \"./cli/commands/migrate\";\nimport { createMigration } from \"./cli/commands/migration\";\nimport { createSchema } from \"./cli/commands/schema\";\n\nconst args = process.argv.slice(2);\nconst command = args[0];\nconst subCommand = args[1];\n\nfunction getFlag(flag: string): string | undefined {\n const index = args.indexOf(flag);\n if (index !== -1 && args[index + 1]) {\n return args[index + 1];\n }\n return undefined;\n}\n\nfunction getName(): string | undefined {\n return getFlag(\"--name\") ?? args[2];\n}\n\nfunction getMigrateOptions() {\n return {\n migrationsPath: getFlag(\"--migrations\"),\n schemasPath: getFlag(\"--schemas\"),\n database: getFlag(\"--database\"),\n connection: getFlag(\"--connection\"),\n steps: getFlag(\"--steps\") ? Number(getFlag(\"--steps\")) : undefined,\n };\n}\n\nswitch (command) {\n case \"migration\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db migration create <name> [--path <path>]\");\n console.log(\" @hedystia/db migration <name> [--path <path>]\");\n process.exit(1);\n }\n createMigration(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"schema\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db schema create <name> [--path <path>]\");\n console.log(\" @hedystia/db schema <name> [--path <path>]\");\n process.exit(1);\n }\n createSchema(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"migrate\": {\n const options = getMigrateOptions();\n if (subCommand === \"up\") {\n migrateUp(options).catch((err) => {\n console.error(\"Migration failed:\", err.message);\n process.exit(1);\n });\n } else if (subCommand === \"down\") {\n migrateDown(options).catch((err) => {\n console.error(\"Rollback failed:\", err.message);\n process.exit(1);\n });\n } else {\n console.log(\"Usage: @hedystia/db migrate up [options]\");\n console.log(\" @hedystia/db migrate down [options]\");\n console.log(\"\");\n console.log(\"Options:\");\n console.log(\" --migrations <path> Path to migrations directory\");\n console.log(\" --schemas <path> Path to schemas directory\");\n console.log(\" --database <type> Database type (sqlite, mysql, mariadb)\");\n console.log(\" --connection <config> Connection string or JSON config\");\n console.log(\n \" --steps <n> Number of migrations to rollback (down only, default: 1)\",\n );\n process.exit(1);\n }\n break;\n }\n default: {\n console.log(\"@hedystia/db CLI\");\n console.log(\"\");\n console.log(\"Commands:\");\n console.log(\" migration create <name> [--path <path>] Create a new migration file\");\n console.log(\n \" migration <name> [--path <path>] Create a new migration file (shorthand)\",\n );\n console.log(\" schema create <name> [--path <path>] Create a new schema file\");\n console.log(\" schema <name> [--path <path>] Create a new schema file (shorthand)\");\n console.log(\" migrate up [options] Run pending migrations\");\n console.log(\" migrate down [options] Rollback migrations\");\n break;\n }\n}\n"],"mappings":";;;;;AAKA,MAAM,OAAO,QAAQ,KAAK,MAAM,EAAE;AAClC,MAAM,UAAU,KAAK;AACrB,MAAM,aAAa,KAAK;AAExB,SAAS,QAAQ,MAAkC;CACjD,MAAM,QAAQ,KAAK,QAAQ,KAAK;AAChC,KAAI,UAAU,MAAM,KAAK,QAAQ,GAC/B,QAAO,KAAK,QAAQ;;AAKxB,SAAS,UAA8B;AACrC,QAAO,QAAQ,SAAS,IAAI,KAAK;;AAGnC,SAAS,oBAAoB;AAC3B,QAAO;EACL,gBAAgB,QAAQ,eAAe;EACvC,aAAa,QAAQ,YAAY;EACjC,UAAU,QAAQ,aAAa;EAC/B,YAAY,QAAQ,eAAe;EACnC,OAAO,QAAQ,UAAU,GAAG,OAAO,QAAQ,UAAU,CAAC,GAAG,KAAA;EAC1D;;AAGH,QAAQ,SAAR;CACE,KAAK;AAEH,OADe,eAAe,WAAW,WAAW,cACrC,UAAU;GACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,OAAI,CAAC,MAAM;AACT,YAAQ,IAAI,8DAA8D;AAC1E,YAAQ,IAAI,uDAAuD;AACnE,YAAQ,KAAK,EAAE;;AAEjB,qBAAA,gBAAgB,MAAM,QAAQ,SAAS,CAAC;;AAE1C;CAEF,KAAK;AAEH,OADe,eAAe,WAAW,WAAW,cACrC,UAAU;GACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,OAAI,CAAC,MAAM;AACT,YAAQ,IAAI,2DAA2D;AACvE,YAAQ,IAAI,oDAAoD;AAChE,YAAQ,KAAK,EAAE;;AAEjB,kBAAA,aAAa,MAAM,QAAQ,SAAS,CAAC;;AAEvC;CAEF,KAAK,WAAW;EACd,MAAM,UAAU,mBAAmB;AACnC,MAAI,eAAe,KACjB,iBAAA,UAAU,QAAQ,CAAC,OAAO,QAAQ;AAChC,WAAQ,MAAM,qBAAqB,IAAI,QAAQ;AAC/C,WAAQ,KAAK,EAAE;IACf;WACO,eAAe,OACxB,iBAAA,YAAY,QAAQ,CAAC,OAAO,QAAQ;AAClC,WAAQ,MAAM,oBAAoB,IAAI,QAAQ;AAC9C,WAAQ,KAAK,EAAE;IACf;OACG;AACL,WAAQ,IAAI,2CAA2C;AACvD,WAAQ,IAAI,6CAA6C;AACzD,WAAQ,IAAI,GAAG;AACf,WAAQ,IAAI,WAAW;AACvB,WAAQ,IAAI,wDAAwD;AACpE,WAAQ,IAAI,qDAAqD;AACjE,WAAQ,IAAI,kEAAkE;AAC9E,WAAQ,IAAI,4DAA4D;AACxE,WAAQ,IACN,oFACD;AACD,WAAQ,KAAK,EAAE;;AAEjB;;CAEF;AACE,UAAQ,IAAI,mBAAmB;AAC/B,UAAQ,IAAI,GAAG;AACf,UAAQ,IAAI,YAAY;AACxB,UAAQ,IAAI,yEAAyE;AACrF,UAAQ,IACN,qFACD;AACD,UAAQ,IAAI,sEAAsE;AAClF,UAAQ,IAAI,kFAAkF;AAC9F,UAAQ,IAAI,oEAAoE;AAChF,UAAQ,IAAI,iEAAiE;AAC7E"}
1
+ {"version":3,"file":"cli.cjs","names":[],"sources":["../src/cli.ts"],"sourcesContent":["#!/usr/bin/env bun\nimport { migrateDown, migrateUp } from \"./cli/commands/migrate\";\nimport { createMigration } from \"./cli/commands/migration\";\nimport { createSchema } from \"./cli/commands/schema\";\n\nconst args = process.argv.slice(2);\nconst command = args[0];\nconst subCommand = args[1];\n\nfunction getFlag(flag: string): string | undefined {\n const index = args.indexOf(flag);\n if (index !== -1 && args[index + 1]) {\n return args[index + 1];\n }\n return undefined;\n}\n\nfunction getName(): string | undefined {\n return getFlag(\"--name\") ?? args[2];\n}\n\nfunction hasFlag(flag: string): boolean {\n return args.includes(flag);\n}\n\nfunction getMigrateOptions() {\n return {\n migrationsPath: getFlag(\"--migrations\"),\n schemasPath: getFlag(\"--schemas\"),\n database: getFlag(\"--database\"),\n connection: getFlag(\"--connection\"),\n steps: getFlag(\"--steps\") ? Number(getFlag(\"--steps\")) : undefined,\n };\n}\n\nswitch (command) {\n case \"migration\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db migration create <name> [--path <path>] [--no-id]\");\n console.log(\" @hedystia/db migration <name> [--path <path>] [--no-id]\");\n process.exit(1);\n }\n createMigration(name, getFlag(\"--path\"), hasFlag(\"--no-id\"));\n }\n break;\n }\n case \"schema\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db schema create <name> [--path <path>]\");\n console.log(\" @hedystia/db schema <name> [--path <path>]\");\n process.exit(1);\n }\n createSchema(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"migrate\": {\n const options = getMigrateOptions();\n if (subCommand === \"up\") {\n migrateUp(options).catch((err) => {\n console.error(\"Migration failed:\", err.message);\n process.exit(1);\n });\n } else if (subCommand === \"down\") {\n migrateDown(options).catch((err) => {\n console.error(\"Rollback failed:\", err.message);\n process.exit(1);\n });\n } else {\n console.log(\"Usage: @hedystia/db migrate up [options]\");\n console.log(\" @hedystia/db migrate down [options]\");\n console.log(\"\");\n console.log(\"Options:\");\n console.log(\" --migrations <path> Path to migrations directory\");\n console.log(\" --schemas <path> Path to schemas directory\");\n console.log(\" --database <type> Database type (sqlite, mysql, mariadb)\");\n console.log(\" --connection <config> Connection string or JSON config\");\n console.log(\n \" --steps <n> Number of migrations to rollback (down only, default: 1)\",\n );\n process.exit(1);\n }\n break;\n }\n default: {\n console.log(\"@hedystia/db CLI\");\n console.log(\"\");\n console.log(\"Commands:\");\n console.log(\n \" migration create <name> [--path <path>] [--no-id] Create a new migration file\",\n );\n console.log(\n \" migration <name> [--path <path>] [--no-id] Create a new migration file (shorthand)\",\n );\n console.log(\" schema create <name> [--path <path>] Create a new schema file\");\n console.log(\" schema <name> [--path <path>] Create a new schema file (shorthand)\");\n console.log(\" migrate up [options] Run pending migrations\");\n console.log(\" migrate down [options] Rollback migrations\");\n break;\n }\n}\n"],"mappings":";;;;;AAKA,MAAM,OAAO,QAAQ,KAAK,MAAM,EAAE;AAClC,MAAM,UAAU,KAAK;AACrB,MAAM,aAAa,KAAK;AAExB,SAAS,QAAQ,MAAkC;CACjD,MAAM,QAAQ,KAAK,QAAQ,KAAK;AAChC,KAAI,UAAU,MAAM,KAAK,QAAQ,GAC/B,QAAO,KAAK,QAAQ;;AAKxB,SAAS,UAA8B;AACrC,QAAO,QAAQ,SAAS,IAAI,KAAK;;AAGnC,SAAS,QAAQ,MAAuB;AACtC,QAAO,KAAK,SAAS,KAAK;;AAG5B,SAAS,oBAAoB;AAC3B,QAAO;EACL,gBAAgB,QAAQ,eAAe;EACvC,aAAa,QAAQ,YAAY;EACjC,UAAU,QAAQ,aAAa;EAC/B,YAAY,QAAQ,eAAe;EACnC,OAAO,QAAQ,UAAU,GAAG,OAAO,QAAQ,UAAU,CAAC,GAAG,KAAA;EAC1D;;AAGH,QAAQ,SAAR;CACE,KAAK;AAEH,OADe,eAAe,WAAW,WAAW,cACrC,UAAU;GACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,OAAI,CAAC,MAAM;AACT,YAAQ,IAAI,wEAAwE;AACpF,YAAQ,IAAI,iEAAiE;AAC7E,YAAQ,KAAK,EAAE;;AAEjB,qBAAA,gBAAgB,MAAM,QAAQ,SAAS,EAAE,QAAQ,UAAU,CAAC;;AAE9D;CAEF,KAAK;AAEH,OADe,eAAe,WAAW,WAAW,cACrC,UAAU;GACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,OAAI,CAAC,MAAM;AACT,YAAQ,IAAI,2DAA2D;AACvE,YAAQ,IAAI,oDAAoD;AAChE,YAAQ,KAAK,EAAE;;AAEjB,kBAAA,aAAa,MAAM,QAAQ,SAAS,CAAC;;AAEvC;CAEF,KAAK,WAAW;EACd,MAAM,UAAU,mBAAmB;AACnC,MAAI,eAAe,KACjB,iBAAA,UAAU,QAAQ,CAAC,OAAO,QAAQ;AAChC,WAAQ,MAAM,qBAAqB,IAAI,QAAQ;AAC/C,WAAQ,KAAK,EAAE;IACf;WACO,eAAe,OACxB,iBAAA,YAAY,QAAQ,CAAC,OAAO,QAAQ;AAClC,WAAQ,MAAM,oBAAoB,IAAI,QAAQ;AAC9C,WAAQ,KAAK,EAAE;IACf;OACG;AACL,WAAQ,IAAI,2CAA2C;AACvD,WAAQ,IAAI,6CAA6C;AACzD,WAAQ,IAAI,GAAG;AACf,WAAQ,IAAI,WAAW;AACvB,WAAQ,IAAI,wDAAwD;AACpE,WAAQ,IAAI,qDAAqD;AACjE,WAAQ,IAAI,kEAAkE;AAC9E,WAAQ,IAAI,4DAA4D;AACxE,WAAQ,IACN,oFACD;AACD,WAAQ,KAAK,EAAE;;AAEjB;;CAEF;AACE,UAAQ,IAAI,mBAAmB;AAC/B,UAAQ,IAAI,GAAG;AACf,UAAQ,IAAI,YAAY;AACxB,UAAQ,IACN,mFACD;AACD,UAAQ,IACN,+FACD;AACD,UAAQ,IAAI,sEAAsE;AAClF,UAAQ,IAAI,kFAAkF;AAC9F,UAAQ,IAAI,oEAAoE;AAChF,UAAQ,IAAI,iEAAiE;AAC7E"}
package/dist/cli.mjs CHANGED
@@ -18,6 +18,9 @@ var require_cli = /* @__PURE__ */ __commonJSMin((() => {
18
18
  function getName() {
19
19
  return getFlag("--name") ?? args[2];
20
20
  }
21
+ function hasFlag(flag) {
22
+ return args.includes(flag);
23
+ }
21
24
  function getMigrateOptions() {
22
25
  return {
23
26
  migrationsPath: getFlag("--migrations"),
@@ -32,11 +35,11 @@ var require_cli = /* @__PURE__ */ __commonJSMin((() => {
32
35
  if ((subCommand === "create" ? "create" : "create") === "create") {
33
36
  const name = subCommand === "create" ? getName() : subCommand;
34
37
  if (!name) {
35
- console.log("Usage: @hedystia/db migration create <name> [--path <path>]");
36
- console.log(" @hedystia/db migration <name> [--path <path>]");
38
+ console.log("Usage: @hedystia/db migration create <name> [--path <path>] [--no-id]");
39
+ console.log(" @hedystia/db migration <name> [--path <path>] [--no-id]");
37
40
  process.exit(1);
38
41
  }
39
- createMigration(name, getFlag("--path"));
42
+ createMigration(name, getFlag("--path"), hasFlag("--no-id"));
40
43
  }
41
44
  break;
42
45
  case "schema":
@@ -78,8 +81,8 @@ var require_cli = /* @__PURE__ */ __commonJSMin((() => {
78
81
  console.log("@hedystia/db CLI");
79
82
  console.log("");
80
83
  console.log("Commands:");
81
- console.log(" migration create <name> [--path <path>] Create a new migration file");
82
- console.log(" migration <name> [--path <path>] Create a new migration file (shorthand)");
84
+ console.log(" migration create <name> [--path <path>] [--no-id] Create a new migration file");
85
+ console.log(" migration <name> [--path <path>] [--no-id] Create a new migration file (shorthand)");
83
86
  console.log(" schema create <name> [--path <path>] Create a new schema file");
84
87
  console.log(" schema <name> [--path <path>] Create a new schema file (shorthand)");
85
88
  console.log(" migrate up [options] Run pending migrations");
package/dist/cli.mjs.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"cli.mjs","names":[],"sources":["../src/cli.ts"],"sourcesContent":["#!/usr/bin/env bun\nimport { migrateDown, migrateUp } from \"./cli/commands/migrate\";\nimport { createMigration } from \"./cli/commands/migration\";\nimport { createSchema } from \"./cli/commands/schema\";\n\nconst args = process.argv.slice(2);\nconst command = args[0];\nconst subCommand = args[1];\n\nfunction getFlag(flag: string): string | undefined {\n const index = args.indexOf(flag);\n if (index !== -1 && args[index + 1]) {\n return args[index + 1];\n }\n return undefined;\n}\n\nfunction getName(): string | undefined {\n return getFlag(\"--name\") ?? args[2];\n}\n\nfunction getMigrateOptions() {\n return {\n migrationsPath: getFlag(\"--migrations\"),\n schemasPath: getFlag(\"--schemas\"),\n database: getFlag(\"--database\"),\n connection: getFlag(\"--connection\"),\n steps: getFlag(\"--steps\") ? Number(getFlag(\"--steps\")) : undefined,\n };\n}\n\nswitch (command) {\n case \"migration\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db migration create <name> [--path <path>]\");\n console.log(\" @hedystia/db migration <name> [--path <path>]\");\n process.exit(1);\n }\n createMigration(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"schema\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db schema create <name> [--path <path>]\");\n console.log(\" @hedystia/db schema <name> [--path <path>]\");\n process.exit(1);\n }\n createSchema(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"migrate\": {\n const options = getMigrateOptions();\n if (subCommand === \"up\") {\n migrateUp(options).catch((err) => {\n console.error(\"Migration failed:\", err.message);\n process.exit(1);\n });\n } else if (subCommand === \"down\") {\n migrateDown(options).catch((err) => {\n console.error(\"Rollback failed:\", err.message);\n process.exit(1);\n });\n } else {\n console.log(\"Usage: @hedystia/db migrate up [options]\");\n console.log(\" @hedystia/db migrate down [options]\");\n console.log(\"\");\n console.log(\"Options:\");\n console.log(\" --migrations <path> Path to migrations directory\");\n console.log(\" --schemas <path> Path to schemas directory\");\n console.log(\" --database <type> Database type (sqlite, mysql, mariadb)\");\n console.log(\" --connection <config> Connection string or JSON config\");\n console.log(\n \" --steps <n> Number of migrations to rollback (down only, default: 1)\",\n );\n process.exit(1);\n }\n break;\n }\n default: {\n console.log(\"@hedystia/db CLI\");\n console.log(\"\");\n console.log(\"Commands:\");\n console.log(\" migration create <name> [--path <path>] Create a new migration file\");\n console.log(\n \" migration <name> [--path <path>] Create a new migration file (shorthand)\",\n );\n console.log(\" schema create <name> [--path <path>] Create a new schema file\");\n console.log(\" schema <name> [--path <path>] Create a new schema file (shorthand)\");\n console.log(\" migrate up [options] Run pending migrations\");\n console.log(\" migrate down [options] Rollback migrations\");\n break;\n }\n}\n"],"mappings":";;;;;;;eACgE;iBACL;cACN;CAErD,MAAM,OAAO,QAAQ,KAAK,MAAM,EAAE;CAClC,MAAM,UAAU,KAAK;CACrB,MAAM,aAAa,KAAK;CAExB,SAAS,QAAQ,MAAkC;EACjD,MAAM,QAAQ,KAAK,QAAQ,KAAK;AAChC,MAAI,UAAU,MAAM,KAAK,QAAQ,GAC/B,QAAO,KAAK,QAAQ;;CAKxB,SAAS,UAA8B;AACrC,SAAO,QAAQ,SAAS,IAAI,KAAK;;CAGnC,SAAS,oBAAoB;AAC3B,SAAO;GACL,gBAAgB,QAAQ,eAAe;GACvC,aAAa,QAAQ,YAAY;GACjC,UAAU,QAAQ,aAAa;GAC/B,YAAY,QAAQ,eAAe;GACnC,OAAO,QAAQ,UAAU,GAAG,OAAO,QAAQ,UAAU,CAAC,GAAG,KAAA;GAC1D;;AAGH,SAAQ,SAAR;EACE,KAAK;AAEH,QADe,eAAe,WAAW,WAAW,cACrC,UAAU;IACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,QAAI,CAAC,MAAM;AACT,aAAQ,IAAI,8DAA8D;AAC1E,aAAQ,IAAI,uDAAuD;AACnE,aAAQ,KAAK,EAAE;;AAEjB,oBAAgB,MAAM,QAAQ,SAAS,CAAC;;AAE1C;EAEF,KAAK;AAEH,QADe,eAAe,WAAW,WAAW,cACrC,UAAU;IACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,QAAI,CAAC,MAAM;AACT,aAAQ,IAAI,2DAA2D;AACvE,aAAQ,IAAI,oDAAoD;AAChE,aAAQ,KAAK,EAAE;;AAEjB,iBAAa,MAAM,QAAQ,SAAS,CAAC;;AAEvC;EAEF,KAAK,WAAW;GACd,MAAM,UAAU,mBAAmB;AACnC,OAAI,eAAe,KACjB,WAAU,QAAQ,CAAC,OAAO,QAAQ;AAChC,YAAQ,MAAM,qBAAqB,IAAI,QAAQ;AAC/C,YAAQ,KAAK,EAAE;KACf;YACO,eAAe,OACxB,aAAY,QAAQ,CAAC,OAAO,QAAQ;AAClC,YAAQ,MAAM,oBAAoB,IAAI,QAAQ;AAC9C,YAAQ,KAAK,EAAE;KACf;QACG;AACL,YAAQ,IAAI,2CAA2C;AACvD,YAAQ,IAAI,6CAA6C;AACzD,YAAQ,IAAI,GAAG;AACf,YAAQ,IAAI,WAAW;AACvB,YAAQ,IAAI,wDAAwD;AACpE,YAAQ,IAAI,qDAAqD;AACjE,YAAQ,IAAI,kEAAkE;AAC9E,YAAQ,IAAI,4DAA4D;AACxE,YAAQ,IACN,oFACD;AACD,YAAQ,KAAK,EAAE;;AAEjB;;EAEF;AACE,WAAQ,IAAI,mBAAmB;AAC/B,WAAQ,IAAI,GAAG;AACf,WAAQ,IAAI,YAAY;AACxB,WAAQ,IAAI,yEAAyE;AACrF,WAAQ,IACN,qFACD;AACD,WAAQ,IAAI,sEAAsE;AAClF,WAAQ,IAAI,kFAAkF;AAC9F,WAAQ,IAAI,oEAAoE;AAChF,WAAQ,IAAI,iEAAiE;AAC7E"}
1
+ {"version":3,"file":"cli.mjs","names":[],"sources":["../src/cli.ts"],"sourcesContent":["#!/usr/bin/env bun\nimport { migrateDown, migrateUp } from \"./cli/commands/migrate\";\nimport { createMigration } from \"./cli/commands/migration\";\nimport { createSchema } from \"./cli/commands/schema\";\n\nconst args = process.argv.slice(2);\nconst command = args[0];\nconst subCommand = args[1];\n\nfunction getFlag(flag: string): string | undefined {\n const index = args.indexOf(flag);\n if (index !== -1 && args[index + 1]) {\n return args[index + 1];\n }\n return undefined;\n}\n\nfunction getName(): string | undefined {\n return getFlag(\"--name\") ?? args[2];\n}\n\nfunction hasFlag(flag: string): boolean {\n return args.includes(flag);\n}\n\nfunction getMigrateOptions() {\n return {\n migrationsPath: getFlag(\"--migrations\"),\n schemasPath: getFlag(\"--schemas\"),\n database: getFlag(\"--database\"),\n connection: getFlag(\"--connection\"),\n steps: getFlag(\"--steps\") ? Number(getFlag(\"--steps\")) : undefined,\n };\n}\n\nswitch (command) {\n case \"migration\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db migration create <name> [--path <path>] [--no-id]\");\n console.log(\" @hedystia/db migration <name> [--path <path>] [--no-id]\");\n process.exit(1);\n }\n createMigration(name, getFlag(\"--path\"), hasFlag(\"--no-id\"));\n }\n break;\n }\n case \"schema\": {\n const action = subCommand === \"create\" ? \"create\" : \"create\";\n if (action === \"create\") {\n const name = subCommand === \"create\" ? getName() : subCommand;\n if (!name) {\n console.log(\"Usage: @hedystia/db schema create <name> [--path <path>]\");\n console.log(\" @hedystia/db schema <name> [--path <path>]\");\n process.exit(1);\n }\n createSchema(name, getFlag(\"--path\"));\n }\n break;\n }\n case \"migrate\": {\n const options = getMigrateOptions();\n if (subCommand === \"up\") {\n migrateUp(options).catch((err) => {\n console.error(\"Migration failed:\", err.message);\n process.exit(1);\n });\n } else if (subCommand === \"down\") {\n migrateDown(options).catch((err) => {\n console.error(\"Rollback failed:\", err.message);\n process.exit(1);\n });\n } else {\n console.log(\"Usage: @hedystia/db migrate up [options]\");\n console.log(\" @hedystia/db migrate down [options]\");\n console.log(\"\");\n console.log(\"Options:\");\n console.log(\" --migrations <path> Path to migrations directory\");\n console.log(\" --schemas <path> Path to schemas directory\");\n console.log(\" --database <type> Database type (sqlite, mysql, mariadb)\");\n console.log(\" --connection <config> Connection string or JSON config\");\n console.log(\n \" --steps <n> Number of migrations to rollback (down only, default: 1)\",\n );\n process.exit(1);\n }\n break;\n }\n default: {\n console.log(\"@hedystia/db CLI\");\n console.log(\"\");\n console.log(\"Commands:\");\n console.log(\n \" migration create <name> [--path <path>] [--no-id] Create a new migration file\",\n );\n console.log(\n \" migration <name> [--path <path>] [--no-id] Create a new migration file (shorthand)\",\n );\n console.log(\" schema create <name> [--path <path>] Create a new schema file\");\n console.log(\" schema <name> [--path <path>] Create a new schema file (shorthand)\");\n console.log(\" migrate up [options] Run pending migrations\");\n console.log(\" migrate down [options] Rollback migrations\");\n break;\n }\n}\n"],"mappings":";;;;;;;eACgE;iBACL;cACN;CAErD,MAAM,OAAO,QAAQ,KAAK,MAAM,EAAE;CAClC,MAAM,UAAU,KAAK;CACrB,MAAM,aAAa,KAAK;CAExB,SAAS,QAAQ,MAAkC;EACjD,MAAM,QAAQ,KAAK,QAAQ,KAAK;AAChC,MAAI,UAAU,MAAM,KAAK,QAAQ,GAC/B,QAAO,KAAK,QAAQ;;CAKxB,SAAS,UAA8B;AACrC,SAAO,QAAQ,SAAS,IAAI,KAAK;;CAGnC,SAAS,QAAQ,MAAuB;AACtC,SAAO,KAAK,SAAS,KAAK;;CAG5B,SAAS,oBAAoB;AAC3B,SAAO;GACL,gBAAgB,QAAQ,eAAe;GACvC,aAAa,QAAQ,YAAY;GACjC,UAAU,QAAQ,aAAa;GAC/B,YAAY,QAAQ,eAAe;GACnC,OAAO,QAAQ,UAAU,GAAG,OAAO,QAAQ,UAAU,CAAC,GAAG,KAAA;GAC1D;;AAGH,SAAQ,SAAR;EACE,KAAK;AAEH,QADe,eAAe,WAAW,WAAW,cACrC,UAAU;IACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,QAAI,CAAC,MAAM;AACT,aAAQ,IAAI,wEAAwE;AACpF,aAAQ,IAAI,iEAAiE;AAC7E,aAAQ,KAAK,EAAE;;AAEjB,oBAAgB,MAAM,QAAQ,SAAS,EAAE,QAAQ,UAAU,CAAC;;AAE9D;EAEF,KAAK;AAEH,QADe,eAAe,WAAW,WAAW,cACrC,UAAU;IACvB,MAAM,OAAO,eAAe,WAAW,SAAS,GAAG;AACnD,QAAI,CAAC,MAAM;AACT,aAAQ,IAAI,2DAA2D;AACvE,aAAQ,IAAI,oDAAoD;AAChE,aAAQ,KAAK,EAAE;;AAEjB,iBAAa,MAAM,QAAQ,SAAS,CAAC;;AAEvC;EAEF,KAAK,WAAW;GACd,MAAM,UAAU,mBAAmB;AACnC,OAAI,eAAe,KACjB,WAAU,QAAQ,CAAC,OAAO,QAAQ;AAChC,YAAQ,MAAM,qBAAqB,IAAI,QAAQ;AAC/C,YAAQ,KAAK,EAAE;KACf;YACO,eAAe,OACxB,aAAY,QAAQ,CAAC,OAAO,QAAQ;AAClC,YAAQ,MAAM,oBAAoB,IAAI,QAAQ;AAC9C,YAAQ,KAAK,EAAE;KACf;QACG;AACL,YAAQ,IAAI,2CAA2C;AACvD,YAAQ,IAAI,6CAA6C;AACzD,YAAQ,IAAI,GAAG;AACf,YAAQ,IAAI,WAAW;AACvB,YAAQ,IAAI,wDAAwD;AACpE,YAAQ,IAAI,qDAAqD;AACjE,YAAQ,IAAI,kEAAkE;AAC9E,YAAQ,IAAI,4DAA4D;AACxE,YAAQ,IACN,oFACD;AACD,YAAQ,KAAK,EAAE;;AAEjB;;EAEF;AACE,WAAQ,IAAI,mBAAmB;AAC/B,WAAQ,IAAI,GAAG;AACf,WAAQ,IAAI,YAAY;AACxB,WAAQ,IACN,mFACD;AACD,WAAQ,IACN,+FACD;AACD,WAAQ,IAAI,sEAAsE;AAClF,WAAQ,IAAI,kFAAkF;AAC9F,WAAQ,IAAI,oEAAoE;AAChF,WAAQ,IAAI,iEAAiE;AAC7E"}
@@ -18,7 +18,14 @@ function normalizeSchemas(schemas) {
18
18
  if (Array.isArray(schemas)) return schemas;
19
19
  return Object.values(schemas).filter((v) => v != null && typeof v === "object" && v.__table === true);
20
20
  }
21
+ function buildSchemaKeyMap(rawSchemas) {
22
+ const map = /* @__PURE__ */ new Map();
23
+ if (Array.isArray(rawSchemas)) return map;
24
+ for (const [key, value] of Object.entries(rawSchemas)) if (value != null && typeof value === "object" && value.__table === true) map.set(value.__name, key);
25
+ return map;
26
+ }
21
27
  function database(config) {
28
+ const schemaKeyMap = buildSchemaKeyMap(config.schemas);
22
29
  const schemas = normalizeSchemas(config.schemas);
23
30
  const registry = new require_registry.SchemaRegistry();
24
31
  registry.register(schemas);
@@ -111,7 +118,8 @@ function database(config) {
111
118
  };
112
119
  return original;
113
120
  } });
114
- instance[schema.__name] = proxy;
121
+ const accessorKey = schemaKeyMap.get(schema.__name) ?? schema.__name;
122
+ instance[accessorKey] = proxy;
115
123
  }
116
124
  return instance;
117
125
  }
@@ -154,8 +162,11 @@ function createMigrationContext(driver, registry) {
154
162
  return {
155
163
  schema: {
156
164
  createTable: async (tableDef) => {
157
- const meta = registry.getTable(tableDef.__name);
158
- if (meta) await driver.createTable(meta);
165
+ const meta = registry.getTable(tableDef.__name) ?? {
166
+ name: tableDef.__name,
167
+ columns: [...tableDef.__columns]
168
+ };
169
+ await driver.createTable(meta);
159
170
  },
160
171
  dropTable: async (name) => {
161
172
  await driver.dropTable(name);
@@ -1 +1 @@
1
- {"version":3,"file":"database.cjs","names":["SchemaRegistry","DatabaseError","createDriver","CacheManager","TableRepository","MIGRATIONS_TABLE"],"sources":["../../src/core/database.ts"],"sourcesContent":["import { CacheManager } from \"../cache\";\nimport { MIGRATIONS_TABLE } from \"../constants\";\nimport { createDriver } from \"../drivers\";\nimport { DatabaseError } from \"../errors\";\nimport { SchemaRegistry } from \"../schema\";\nimport type {\n AnyTableDef,\n ConnectionConfig,\n DatabaseConfig,\n DatabaseDriver,\n DeleteOptions,\n InferRow,\n InferSchemas,\n MigrationDefinition,\n QueryOptions,\n RelationQueryMap,\n ResolveResult,\n TableMetadata,\n UpdateOptions,\n WhereClause,\n} from \"../types\";\nimport { TableRepository } from \"./repository\";\n\ntype TypedTableRepository<S extends readonly AnyTableDef[], T extends AnyTableDef> = {\n /**\n * Find all rows matching the given options\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n find<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find all rows matching the given options (alias for {@link find})\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n findMany<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find the first row matching the given options\n * @param options - Filter, sort, and eagerly load relations\n * @returns The first matching row, or `null` if none found\n */\n findFirst<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O> | null>;\n\n /**\n * Insert one or more rows into the table\n * @param data - Row data (or array of row data) to insert\n * @returns The inserted row\n */\n insert(data: Partial<InferRow<T>> | Partial<InferRow<T>>[]): Promise<InferRow<T>>;\n\n /**\n * Insert multiple rows into the table\n * @param data - Array of row data to insert\n * @returns Array of inserted rows\n */\n insertMany(data: Partial<InferRow<T>>[]): Promise<InferRow<T>[]>;\n\n /**\n * Update rows matching the where clause\n * @param options - Where clause and partial data to apply\n * @returns Array of updated rows\n */\n update(options: UpdateOptions<InferRow<T>>): Promise<InferRow<T>[]>;\n\n /**\n * Delete rows matching the where clause\n * @param options - Where clause to select rows for deletion\n * @returns Number of deleted rows\n */\n delete(options: DeleteOptions<InferRow<T>>): Promise<number>;\n\n /**\n * Count rows matching the where clause\n * @param options - Optional where clause to filter rows\n * @returns Number of matching rows\n */\n count(options?: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<number>;\n\n /**\n * Check whether at least one row matches the where clause\n * @param options - Where clause to check\n * @returns `true` if a matching row exists, `false` otherwise\n */\n exists(options: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<boolean>;\n\n /**\n * Insert a row if it doesn't exist, or update it if it does\n * @param options - Where clause to check, data to create, and data to update\n * @returns The created or updated row\n */\n upsert(options: {\n where: WhereClause<InferRow<T>>;\n create: Partial<InferRow<T>>;\n update: Partial<InferRow<T>>;\n }): Promise<InferRow<T>>;\n\n /**\n * Remove all rows from the table\n * @returns Resolves when the table has been truncated\n */\n truncate(): Promise<void>;\n};\n\ntype ExtractRepos<S> = S extends readonly AnyTableDef[]\n ? { [T in S[number] as T[\"__name\"]]: TypedTableRepository<S, T> }\n : S extends Record<string, any>\n ? {\n [K in keyof S as S[K] extends AnyTableDef ? K : never]: TypedTableRepository<\n InferSchemas<S>,\n Extract<S[K], AnyTableDef>\n >;\n }\n : never;\n\ntype DatabaseInstance<S> = ExtractRepos<S> & {\n /**\n * Initialize the database connection, create tables and run migrations\n * @returns {Promise<void>}\n */\n initialize(): Promise<void>;\n /**\n * Close the database connection\n * @returns {Promise<void>}\n */\n close(): Promise<void>;\n /**\n * Get the underlying database driver\n * @returns {DatabaseDriver} The database driver\n */\n getDriver(): DatabaseDriver;\n /**\n * Get the schema registry\n * @returns {SchemaRegistry} The schema registry\n */\n getRegistry(): SchemaRegistry;\n /**\n * Get the cache manager\n * @returns {CacheManager} The cache manager\n */\n getCache(): CacheManager;\n /**\n * Execute a raw SQL query\n * @param {string} sql - SQL query\n * @param {unknown[]} [params] - Query parameters\n * @returns {Promise<any[]>} Query results\n */\n raw(sql: string, params?: unknown[]): Promise<any[]>;\n /**\n * Execute within a transaction\n * @param {() => Promise<T>} fn - Function to execute\n * @returns {Promise<T>} Result\n */\n transaction<T>(fn: () => Promise<T>): Promise<T>;\n /**\n * Run pending migrations manually\n * @returns {Promise<void>}\n */\n migrateUp(): Promise<void>;\n /**\n * Rollback migrations\n * @param {number} [steps=1] - Number of migrations to rollback\n * @returns {Promise<string[]>} Names of rolled back migrations\n */\n migrateDown(steps?: number): Promise<string[]>;\n};\n\n/**\n * Create a database instance with typed repositories for each schema\n * @param {DatabaseConfig} config - Database configuration\n * @returns {DatabaseInstance<S>} Database instance with table repositories\n */\nfunction normalizeMigrations(\n migrations: MigrationDefinition[] | Record<string, unknown>,\n): MigrationDefinition[] {\n if (Array.isArray(migrations)) {\n return migrations;\n }\n return Object.values(migrations).filter(\n (v): v is MigrationDefinition =>\n v != null && typeof v === \"object\" && \"name\" in v && \"up\" in v && \"down\" in v,\n );\n}\n\nfunction normalizeSchemas<S extends readonly AnyTableDef[]>(schemas: any): S {\n if (Array.isArray(schemas)) {\n return schemas as any as S;\n }\n return Object.values(schemas).filter(\n (v): v is AnyTableDef => v != null && typeof v === \"object\" && (v as any).__table === true,\n ) as any;\n}\n\nexport function database<const S extends readonly AnyTableDef[] | Record<string, any>>(\n config: DatabaseConfig & { schemas: S },\n): DatabaseInstance<S> {\n const schemas = normalizeSchemas(config.schemas) as any;\n const registry = new SchemaRegistry();\n registry.register(schemas);\n\n const dbName = typeof config.database === \"string\" ? config.database : config.database.name;\n\n let connectionConfig: ConnectionConfig | undefined;\n if (Array.isArray(config.connection)) {\n if (dbName === \"sqlite\") {\n connectionConfig = config.connection.find((c) => \"filename\" in c) ?? config.connection[0];\n } else if (dbName === \"mysql\" || dbName === \"mariadb\") {\n connectionConfig = config.connection.find((c) => \"host\" in c) ?? config.connection[0];\n } else if (dbName === \"file\") {\n connectionConfig = config.connection.find((c) => \"directory\" in c) ?? config.connection[0];\n } else {\n connectionConfig = config.connection[0];\n }\n } else {\n connectionConfig = config.connection;\n }\n\n if (!connectionConfig) {\n throw new DatabaseError(\"Connection config is required\");\n }\n\n const driver = createDriver(config.database, connectionConfig as ConnectionConfig);\n const cache = new CacheManager(config.cache);\n\n let initialized = false;\n let initPromise: Promise<void> | null = null;\n\n const ensureInitialized = async () => {\n if (initialized) {\n return;\n }\n if (initPromise) {\n return initPromise;\n }\n initPromise = doInit();\n await initPromise;\n initialized = true;\n };\n\n const doInit = async () => {\n await driver.connect();\n\n if (config.syncSchemas) {\n const allMetadata = driver.getAllTableColumns ? await driver.getAllTableColumns() : null;\n\n const syncPromises = Array.from(registry.getAllTables()).map(async ([, tableMeta]) => {\n const existingCols = allMetadata ? allMetadata[tableMeta.name] : null;\n const exists = allMetadata ? !!existingCols : await driver.tableExists(tableMeta.name);\n\n if (!exists) {\n await driver.createTable(tableMeta);\n } else {\n const cols = existingCols || (await driver.getTableColumns(tableMeta.name));\n const existingNames = new Set(cols.map((c) => c.name));\n const addColumnPromises = tableMeta.columns\n .filter((colMeta) => !existingNames.has(colMeta.name))\n .map((colMeta) => driver.addColumn(tableMeta.name, colMeta));\n await Promise.all(addColumnPromises);\n }\n });\n await Promise.all(syncPromises);\n }\n\n if (config.runMigrations && config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n };\n\n const repos = new Map<string, TableRepository<any>>();\n for (const schema of schemas) {\n const repo = new TableRepository(schema.__name, driver, cache, registry, schema.__cache);\n repos.set(schema.__name, repo);\n }\n\n const instance: any = {\n initialize: async () => {\n await ensureInitialized();\n },\n close: async () => {\n await driver.disconnect();\n cache.clear();\n initialized = false;\n initPromise = null;\n },\n getDriver: () => driver,\n getRegistry: () => registry,\n getCache: () => cache,\n raw: async (sql: string, params?: unknown[]) => {\n await ensureInitialized();\n return driver.query(sql, params);\n },\n transaction: async <T>(fn: () => Promise<T>) => {\n await ensureInitialized();\n return driver.transaction(fn);\n },\n migrateUp: async () => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n },\n migrateDown: async (steps = 1) => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n return rollbackMigrations(driver, registry, migrations, steps);\n }\n return [];\n },\n };\n\n for (const schema of schemas) {\n const repo = repos.get(schema.__name)!;\n const proxy = new Proxy(repo, {\n get(target, prop, receiver) {\n const original = Reflect.get(target, prop, receiver);\n if (typeof original === \"function\") {\n return async (...args: any[]) => {\n await ensureInitialized();\n return original.apply(target, args);\n };\n }\n return original;\n },\n });\n instance[schema.__name] = proxy;\n }\n\n return instance as any as DatabaseInstance<S>;\n}\n\nfunction getMigrationsTableMeta(): TableMetadata {\n return {\n name: MIGRATIONS_TABLE,\n columns: [\n {\n name: \"id\",\n type: \"integer\",\n primaryKey: true,\n autoIncrement: true,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n },\n {\n name: \"name\",\n type: \"varchar\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n length: 255,\n },\n {\n name: \"executed_at\",\n type: \"datetime\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: false,\n defaultValue: undefined,\n },\n ],\n };\n}\n\nfunction createMigrationContext(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n): MigrationDefinition[\"up\"] extends (ctx: infer C) => any ? C : never {\n return {\n schema: {\n createTable: async (tableDef: AnyTableDef) => {\n const meta = registry.getTable(tableDef.__name);\n if (meta) {\n await driver.createTable(meta);\n }\n },\n dropTable: async (name: string) => {\n await driver.dropTable(name);\n },\n addColumn: async (table: string, _name: string, column: any) => {\n await driver.addColumn(table, column);\n },\n dropColumn: async (table: string, name: string) => {\n await driver.dropColumn(table, name);\n },\n renameColumn: async (table: string, oldName: string, newName: string) => {\n await driver.renameColumn(table, oldName, newName);\n },\n addIndex: async () => {},\n dropIndex: async () => {},\n },\n sql: async (query: string, params?: unknown[]) => {\n return driver.execute(query, params);\n },\n };\n}\n\nasync function ensureMigrationsTable(driver: DatabaseDriver): Promise<void> {\n const exists = await driver.tableExists(MIGRATIONS_TABLE);\n if (!exists) {\n await driver.createTable(getMigrationsTableMeta());\n }\n}\n\nasync function runMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n): Promise<void> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\``);\n const executedNames = new Set(executed.map((r: any) => r.name));\n\n const ctx = createMigrationContext(driver, registry);\n\n for (const migration of migrations) {\n if (executedNames.has(migration.name)) {\n continue;\n }\n\n await migration.up(ctx);\n\n await driver.execute(\n `INSERT INTO \\`${MIGRATIONS_TABLE}\\` (\\`name\\`, \\`executed_at\\`) VALUES (?, ?)`,\n [migration.name, new Date()],\n );\n }\n}\n\nasync function rollbackMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n steps = 1,\n): Promise<string[]> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\` ORDER BY id DESC`);\n const executedNames = executed.map((r: any) => r.name as string);\n\n const ctx = createMigrationContext(driver, registry);\n const migrationMap = new Map(migrations.map((m) => [m.name, m]));\n const rolledBack: string[] = [];\n\n const toRollback = executedNames.slice(0, steps);\n\n for (const name of toRollback) {\n const migration = migrationMap.get(name);\n if (!migration) {\n continue;\n }\n\n await migration.down(ctx);\n\n await driver.execute(`DELETE FROM \\`${MIGRATIONS_TABLE}\\` WHERE \\`name\\` = ?`, [name]);\n rolledBack.push(name);\n }\n\n return rolledBack;\n}\n"],"mappings":";;;;;;;;;;;;AAmLA,SAAS,oBACP,YACuB;AACvB,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,QAAO,OAAO,OAAO,WAAW,CAAC,QAC9B,MACC,KAAK,QAAQ,OAAO,MAAM,YAAY,UAAU,KAAK,QAAQ,KAAK,UAAU,EAC/E;;AAGH,SAAS,iBAAmD,SAAiB;AAC3E,KAAI,MAAM,QAAQ,QAAQ,CACxB,QAAO;AAET,QAAO,OAAO,OAAO,QAAQ,CAAC,QAC3B,MAAwB,KAAK,QAAQ,OAAO,MAAM,YAAa,EAAU,YAAY,KACvF;;AAGH,SAAgB,SACd,QACqB;CACrB,MAAM,UAAU,iBAAiB,OAAO,QAAQ;CAChD,MAAM,WAAW,IAAIA,iBAAAA,gBAAgB;AACrC,UAAS,SAAS,QAAQ;CAE1B,MAAM,SAAS,OAAO,OAAO,aAAa,WAAW,OAAO,WAAW,OAAO,SAAS;CAEvF,IAAI;AACJ,KAAI,MAAM,QAAQ,OAAO,WAAW,CAClC,KAAI,WAAW,SACb,oBAAmB,OAAO,WAAW,MAAM,MAAM,cAAc,EAAE,IAAI,OAAO,WAAW;UAC9E,WAAW,WAAW,WAAW,UAC1C,oBAAmB,OAAO,WAAW,MAAM,MAAM,UAAU,EAAE,IAAI,OAAO,WAAW;UAC1E,WAAW,OACpB,oBAAmB,OAAO,WAAW,MAAM,MAAM,eAAe,EAAE,IAAI,OAAO,WAAW;KAExF,oBAAmB,OAAO,WAAW;KAGvC,oBAAmB,OAAO;AAG5B,KAAI,CAAC,iBACH,OAAM,IAAIC,eAAAA,cAAc,gCAAgC;CAG1D,MAAM,SAASC,cAAAA,aAAa,OAAO,UAAU,iBAAqC;CAClF,MAAM,QAAQ,IAAIC,gBAAAA,aAAa,OAAO,MAAM;CAE5C,IAAI,cAAc;CAClB,IAAI,cAAoC;CAExC,MAAM,oBAAoB,YAAY;AACpC,MAAI,YACF;AAEF,MAAI,YACF,QAAO;AAET,gBAAc,QAAQ;AACtB,QAAM;AACN,gBAAc;;CAGhB,MAAM,SAAS,YAAY;AACzB,QAAM,OAAO,SAAS;AAEtB,MAAI,OAAO,aAAa;GACtB,MAAM,cAAc,OAAO,qBAAqB,MAAM,OAAO,oBAAoB,GAAG;GAEpF,MAAM,eAAe,MAAM,KAAK,SAAS,cAAc,CAAC,CAAC,IAAI,OAAO,GAAG,eAAe;IACpF,MAAM,eAAe,cAAc,YAAY,UAAU,QAAQ;AAGjE,QAAI,EAFW,cAAc,CAAC,CAAC,eAAe,MAAM,OAAO,YAAY,UAAU,KAAK,EAGpF,OAAM,OAAO,YAAY,UAAU;SAC9B;KACL,MAAM,OAAO,gBAAiB,MAAM,OAAO,gBAAgB,UAAU,KAAK;KAC1E,MAAM,gBAAgB,IAAI,IAAI,KAAK,KAAK,MAAM,EAAE,KAAK,CAAC;KACtD,MAAM,oBAAoB,UAAU,QACjC,QAAQ,YAAY,CAAC,cAAc,IAAI,QAAQ,KAAK,CAAC,CACrD,KAAK,YAAY,OAAO,UAAU,UAAU,MAAM,QAAQ,CAAC;AAC9D,WAAM,QAAQ,IAAI,kBAAkB;;KAEtC;AACF,SAAM,QAAQ,IAAI,aAAa;;AAGjC,MAAI,OAAO,iBAAiB,OAAO,YAAY;GAC7C,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,OAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;CAKvD,MAAM,wBAAQ,IAAI,KAAmC;AACrD,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,IAAIC,mBAAAA,gBAAgB,OAAO,QAAQ,QAAQ,OAAO,UAAU,OAAO,QAAQ;AACxF,QAAM,IAAI,OAAO,QAAQ,KAAK;;CAGhC,MAAM,WAAgB;EACpB,YAAY,YAAY;AACtB,SAAM,mBAAmB;;EAE3B,OAAO,YAAY;AACjB,SAAM,OAAO,YAAY;AACzB,SAAM,OAAO;AACb,iBAAc;AACd,iBAAc;;EAEhB,iBAAiB;EACjB,mBAAmB;EACnB,gBAAgB;EAChB,KAAK,OAAO,KAAa,WAAuB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,MAAM,KAAK,OAAO;;EAElC,aAAa,OAAU,OAAyB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,YAAY,GAAG;;EAE/B,WAAW,YAAY;AACrB,SAAM,mBAAmB;AACzB,OAAI,OAAO,YAAY;IACrB,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,QAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;EAIvD,aAAa,OAAO,QAAQ,MAAM;AAChC,SAAM,mBAAmB;AACzB,OAAI,OAAO,WAET,QAAO,mBAAmB,QAAQ,UADf,oBAAoB,OAAO,WAAW,EACD,MAAM;AAEhE,UAAO,EAAE;;EAEZ;AAED,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,MAAM,IAAI,OAAO,OAAO;EACrC,MAAM,QAAQ,IAAI,MAAM,MAAM,EAC5B,IAAI,QAAQ,MAAM,UAAU;GAC1B,MAAM,WAAW,QAAQ,IAAI,QAAQ,MAAM,SAAS;AACpD,OAAI,OAAO,aAAa,WACtB,QAAO,OAAO,GAAG,SAAgB;AAC/B,UAAM,mBAAmB;AACzB,WAAO,SAAS,MAAM,QAAQ,KAAK;;AAGvC,UAAO;KAEV,CAAC;AACF,WAAS,OAAO,UAAU;;AAG5B,QAAO;;AAGT,SAAS,yBAAwC;AAC/C,QAAO;EACL,MAAMC,kBAAAA;EACN,SAAS;GACP;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACd,QAAQ;IACT;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACF;EACF;;AAGH,SAAS,uBACP,QACA,UACqE;AACrE,QAAO;EACL,QAAQ;GACN,aAAa,OAAO,aAA0B;IAC5C,MAAM,OAAO,SAAS,SAAS,SAAS,OAAO;AAC/C,QAAI,KACF,OAAM,OAAO,YAAY,KAAK;;GAGlC,WAAW,OAAO,SAAiB;AACjC,UAAM,OAAO,UAAU,KAAK;;GAE9B,WAAW,OAAO,OAAe,OAAe,WAAgB;AAC9D,UAAM,OAAO,UAAU,OAAO,OAAO;;GAEvC,YAAY,OAAO,OAAe,SAAiB;AACjD,UAAM,OAAO,WAAW,OAAO,KAAK;;GAEtC,cAAc,OAAO,OAAe,SAAiB,YAAoB;AACvE,UAAM,OAAO,aAAa,OAAO,SAAS,QAAQ;;GAEpD,UAAU,YAAY;GACtB,WAAW,YAAY;GACxB;EACD,KAAK,OAAO,OAAe,WAAuB;AAChD,UAAO,OAAO,QAAQ,OAAO,OAAO;;EAEvC;;AAGH,eAAe,sBAAsB,QAAuC;AAE1E,KAAI,CADW,MAAM,OAAO,YAAA,wBAA6B,CAEvD,OAAM,OAAO,YAAY,wBAAwB,CAAC;;AAItD,eAAe,cACb,QACA,UACA,YACe;AACf,OAAM,sBAAsB,OAAO;CAEnC,MAAM,WAAW,MAAM,OAAO,MAAM,sBAAsBA,kBAAAA,iBAAiB,IAAI;CAC/E,MAAM,gBAAgB,IAAI,IAAI,SAAS,KAAK,MAAW,EAAE,KAAK,CAAC;CAE/D,MAAM,MAAM,uBAAuB,QAAQ,SAAS;AAEpD,MAAK,MAAM,aAAa,YAAY;AAClC,MAAI,cAAc,IAAI,UAAU,KAAK,CACnC;AAGF,QAAM,UAAU,GAAG,IAAI;AAEvB,QAAM,OAAO,QACX,iBAAiBA,kBAAAA,iBAAiB,+CAClC,CAAC,UAAU,sBAAM,IAAI,MAAM,CAAC,CAC7B;;;AAIL,eAAe,mBACb,QACA,UACA,YACA,QAAQ,GACW;AACnB,OAAM,sBAAsB,OAAO;CAGnC,MAAM,iBADW,MAAM,OAAO,MAAM,sBAAsBA,kBAAAA,iBAAiB,qBAAqB,EACjE,KAAK,MAAW,EAAE,KAAe;CAEhE,MAAM,MAAM,uBAAuB,QAAQ,SAAS;CACpD,MAAM,eAAe,IAAI,IAAI,WAAW,KAAK,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;CAChE,MAAM,aAAuB,EAAE;CAE/B,MAAM,aAAa,cAAc,MAAM,GAAG,MAAM;AAEhD,MAAK,MAAM,QAAQ,YAAY;EAC7B,MAAM,YAAY,aAAa,IAAI,KAAK;AACxC,MAAI,CAAC,UACH;AAGF,QAAM,UAAU,KAAK,IAAI;AAEzB,QAAM,OAAO,QAAQ,iBAAiBA,kBAAAA,iBAAiB,wBAAwB,CAAC,KAAK,CAAC;AACtF,aAAW,KAAK,KAAK;;AAGvB,QAAO"}
1
+ {"version":3,"file":"database.cjs","names":["SchemaRegistry","DatabaseError","createDriver","CacheManager","TableRepository","MIGRATIONS_TABLE"],"sources":["../../src/core/database.ts"],"sourcesContent":["import { CacheManager } from \"../cache\";\nimport { MIGRATIONS_TABLE } from \"../constants\";\nimport { createDriver } from \"../drivers\";\nimport { DatabaseError } from \"../errors\";\nimport { SchemaRegistry } from \"../schema\";\nimport type {\n AnyTableDef,\n ConnectionConfig,\n DatabaseConfig,\n DatabaseDriver,\n DeleteOptions,\n InferRow,\n InferSchemas,\n MigrationDefinition,\n QueryOptions,\n RelationQueryMap,\n ResolveResult,\n TableMetadata,\n UpdateOptions,\n WhereClause,\n} from \"../types\";\nimport { TableRepository } from \"./repository\";\n\ntype TypedTableRepository<S extends readonly AnyTableDef[], T extends AnyTableDef> = {\n /**\n * Find all rows matching the given options\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n find<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find all rows matching the given options (alias for {@link find})\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n findMany<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find the first row matching the given options\n * @param options - Filter, sort, and eagerly load relations\n * @returns The first matching row, or `null` if none found\n */\n findFirst<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O> | null>;\n\n /**\n * Insert one or more rows into the table\n * @param data - Row data (or array of row data) to insert\n * @returns The inserted row\n */\n insert(data: Partial<InferRow<T>> | Partial<InferRow<T>>[]): Promise<InferRow<T>>;\n\n /**\n * Insert multiple rows into the table\n * @param data - Array of row data to insert\n * @returns Array of inserted rows\n */\n insertMany(data: Partial<InferRow<T>>[]): Promise<InferRow<T>[]>;\n\n /**\n * Update rows matching the where clause\n * @param options - Where clause and partial data to apply\n * @returns Array of updated rows\n */\n update(options: UpdateOptions<InferRow<T>>): Promise<InferRow<T>[]>;\n\n /**\n * Delete rows matching the where clause\n * @param options - Where clause to select rows for deletion\n * @returns Number of deleted rows\n */\n delete(options: DeleteOptions<InferRow<T>>): Promise<number>;\n\n /**\n * Count rows matching the where clause\n * @param options - Optional where clause to filter rows\n * @returns Number of matching rows\n */\n count(options?: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<number>;\n\n /**\n * Check whether at least one row matches the where clause\n * @param options - Where clause to check\n * @returns `true` if a matching row exists, `false` otherwise\n */\n exists(options: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<boolean>;\n\n /**\n * Insert a row if it doesn't exist, or update it if it does\n * @param options - Where clause to check, data to create, and data to update\n * @returns The created or updated row\n */\n upsert(options: {\n where: WhereClause<InferRow<T>>;\n create: Partial<InferRow<T>>;\n update: Partial<InferRow<T>>;\n }): Promise<InferRow<T>>;\n\n /**\n * Remove all rows from the table\n * @returns Resolves when the table has been truncated\n */\n truncate(): Promise<void>;\n};\n\ntype ExtractRepos<S> = S extends readonly AnyTableDef[]\n ? { [T in S[number] as T[\"__name\"]]: TypedTableRepository<S, T> }\n : S extends Record<string, any>\n ? {\n [K in keyof S as S[K] extends AnyTableDef ? K : never]: TypedTableRepository<\n InferSchemas<S>,\n Extract<S[K], AnyTableDef>\n >;\n }\n : never;\n\ntype DatabaseInstance<S> = ExtractRepos<S> & {\n /**\n * Initialize the database connection, create tables and run migrations\n * @returns {Promise<void>}\n */\n initialize(): Promise<void>;\n /**\n * Close the database connection\n * @returns {Promise<void>}\n */\n close(): Promise<void>;\n /**\n * Get the underlying database driver\n * @returns {DatabaseDriver} The database driver\n */\n getDriver(): DatabaseDriver;\n /**\n * Get the schema registry\n * @returns {SchemaRegistry} The schema registry\n */\n getRegistry(): SchemaRegistry;\n /**\n * Get the cache manager\n * @returns {CacheManager} The cache manager\n */\n getCache(): CacheManager;\n /**\n * Execute a raw SQL query\n * @param {string} sql - SQL query\n * @param {unknown[]} [params] - Query parameters\n * @returns {Promise<any[]>} Query results\n */\n raw(sql: string, params?: unknown[]): Promise<any[]>;\n /**\n * Execute within a transaction\n * @param {() => Promise<T>} fn - Function to execute\n * @returns {Promise<T>} Result\n */\n transaction<T>(fn: () => Promise<T>): Promise<T>;\n /**\n * Run pending migrations manually\n * @returns {Promise<void>}\n */\n migrateUp(): Promise<void>;\n /**\n * Rollback migrations\n * @param {number} [steps=1] - Number of migrations to rollback\n * @returns {Promise<string[]>} Names of rolled back migrations\n */\n migrateDown(steps?: number): Promise<string[]>;\n};\n\n/**\n * Create a database instance with typed repositories for each schema\n * @param {DatabaseConfig} config - Database configuration\n * @returns {DatabaseInstance<S>} Database instance with table repositories\n */\nfunction normalizeMigrations(\n migrations: MigrationDefinition[] | Record<string, unknown>,\n): MigrationDefinition[] {\n if (Array.isArray(migrations)) {\n return migrations;\n }\n return Object.values(migrations).filter(\n (v): v is MigrationDefinition =>\n v != null && typeof v === \"object\" && \"name\" in v && \"up\" in v && \"down\" in v,\n );\n}\n\nfunction normalizeSchemas<S extends readonly AnyTableDef[]>(schemas: any): S {\n if (Array.isArray(schemas)) {\n return schemas as any as S;\n }\n return Object.values(schemas).filter(\n (v): v is AnyTableDef => v != null && typeof v === \"object\" && (v as any).__table === true,\n ) as any;\n}\n\nfunction buildSchemaKeyMap(rawSchemas: any): Map<string, string> {\n const map = new Map<string, string>();\n if (Array.isArray(rawSchemas)) {\n return map;\n }\n for (const [key, value] of Object.entries(rawSchemas)) {\n if (value != null && typeof value === \"object\" && (value as any).__table === true) {\n map.set((value as any).__name, key);\n }\n }\n return map;\n}\n\nexport function database<const S extends readonly AnyTableDef[] | Record<string, any>>(\n config: DatabaseConfig & { schemas: S },\n): DatabaseInstance<S> {\n const schemaKeyMap = buildSchemaKeyMap(config.schemas);\n const schemas = normalizeSchemas(config.schemas) as any;\n const registry = new SchemaRegistry();\n registry.register(schemas);\n\n const dbName = typeof config.database === \"string\" ? config.database : config.database.name;\n\n let connectionConfig: ConnectionConfig | undefined;\n if (Array.isArray(config.connection)) {\n if (dbName === \"sqlite\") {\n connectionConfig = config.connection.find((c) => \"filename\" in c) ?? config.connection[0];\n } else if (dbName === \"mysql\" || dbName === \"mariadb\") {\n connectionConfig = config.connection.find((c) => \"host\" in c) ?? config.connection[0];\n } else if (dbName === \"file\") {\n connectionConfig = config.connection.find((c) => \"directory\" in c) ?? config.connection[0];\n } else {\n connectionConfig = config.connection[0];\n }\n } else {\n connectionConfig = config.connection;\n }\n\n if (!connectionConfig) {\n throw new DatabaseError(\"Connection config is required\");\n }\n\n const driver = createDriver(config.database, connectionConfig as ConnectionConfig);\n const cache = new CacheManager(config.cache);\n\n let initialized = false;\n let initPromise: Promise<void> | null = null;\n\n const ensureInitialized = async () => {\n if (initialized) {\n return;\n }\n if (initPromise) {\n return initPromise;\n }\n initPromise = doInit();\n await initPromise;\n initialized = true;\n };\n\n const doInit = async () => {\n await driver.connect();\n\n if (config.syncSchemas) {\n const allMetadata = driver.getAllTableColumns ? await driver.getAllTableColumns() : null;\n\n const syncPromises = Array.from(registry.getAllTables()).map(async ([, tableMeta]) => {\n const existingCols = allMetadata ? allMetadata[tableMeta.name] : null;\n const exists = allMetadata ? !!existingCols : await driver.tableExists(tableMeta.name);\n\n if (!exists) {\n await driver.createTable(tableMeta);\n } else {\n const cols = existingCols || (await driver.getTableColumns(tableMeta.name));\n const existingNames = new Set(cols.map((c) => c.name));\n const addColumnPromises = tableMeta.columns\n .filter((colMeta) => !existingNames.has(colMeta.name))\n .map((colMeta) => driver.addColumn(tableMeta.name, colMeta));\n await Promise.all(addColumnPromises);\n }\n });\n await Promise.all(syncPromises);\n }\n\n if (config.runMigrations && config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n };\n\n const repos = new Map<string, TableRepository<any>>();\n for (const schema of schemas) {\n const repo = new TableRepository(schema.__name, driver, cache, registry, schema.__cache);\n repos.set(schema.__name, repo);\n }\n\n const instance: any = {\n initialize: async () => {\n await ensureInitialized();\n },\n close: async () => {\n await driver.disconnect();\n cache.clear();\n initialized = false;\n initPromise = null;\n },\n getDriver: () => driver,\n getRegistry: () => registry,\n getCache: () => cache,\n raw: async (sql: string, params?: unknown[]) => {\n await ensureInitialized();\n return driver.query(sql, params);\n },\n transaction: async <T>(fn: () => Promise<T>) => {\n await ensureInitialized();\n return driver.transaction(fn);\n },\n migrateUp: async () => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n },\n migrateDown: async (steps = 1) => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n return rollbackMigrations(driver, registry, migrations, steps);\n }\n return [];\n },\n };\n\n for (const schema of schemas) {\n const repo = repos.get(schema.__name)!;\n const proxy = new Proxy(repo, {\n get(target, prop, receiver) {\n const original = Reflect.get(target, prop, receiver);\n if (typeof original === \"function\") {\n return async (...args: any[]) => {\n await ensureInitialized();\n return original.apply(target, args);\n };\n }\n return original;\n },\n });\n const accessorKey = schemaKeyMap.get(schema.__name) ?? schema.__name;\n instance[accessorKey] = proxy;\n }\n\n return instance as any as DatabaseInstance<S>;\n}\n\nfunction getMigrationsTableMeta(): TableMetadata {\n return {\n name: MIGRATIONS_TABLE,\n columns: [\n {\n name: \"id\",\n type: \"integer\",\n primaryKey: true,\n autoIncrement: true,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n },\n {\n name: \"name\",\n type: \"varchar\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n length: 255,\n },\n {\n name: \"executed_at\",\n type: \"datetime\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: false,\n defaultValue: undefined,\n },\n ],\n };\n}\n\nfunction createMigrationContext(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n): MigrationDefinition[\"up\"] extends (ctx: infer C) => any ? C : never {\n return {\n schema: {\n createTable: async (tableDef: AnyTableDef) => {\n const meta = registry.getTable(tableDef.__name) ?? {\n name: tableDef.__name,\n columns: [...tableDef.__columns],\n };\n await driver.createTable(meta);\n },\n dropTable: async (name: string) => {\n await driver.dropTable(name);\n },\n addColumn: async (table: string, _name: string, column: any) => {\n await driver.addColumn(table, column);\n },\n dropColumn: async (table: string, name: string) => {\n await driver.dropColumn(table, name);\n },\n renameColumn: async (table: string, oldName: string, newName: string) => {\n await driver.renameColumn(table, oldName, newName);\n },\n addIndex: async () => {},\n dropIndex: async () => {},\n },\n sql: async (query: string, params?: unknown[]) => {\n return driver.execute(query, params);\n },\n };\n}\n\nasync function ensureMigrationsTable(driver: DatabaseDriver): Promise<void> {\n const exists = await driver.tableExists(MIGRATIONS_TABLE);\n if (!exists) {\n await driver.createTable(getMigrationsTableMeta());\n }\n}\n\nasync function runMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n): Promise<void> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\``);\n const executedNames = new Set(executed.map((r: any) => r.name));\n\n const ctx = createMigrationContext(driver, registry);\n\n for (const migration of migrations) {\n if (executedNames.has(migration.name)) {\n continue;\n }\n\n await migration.up(ctx);\n\n await driver.execute(\n `INSERT INTO \\`${MIGRATIONS_TABLE}\\` (\\`name\\`, \\`executed_at\\`) VALUES (?, ?)`,\n [migration.name, new Date()],\n );\n }\n}\n\nasync function rollbackMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n steps = 1,\n): Promise<string[]> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\` ORDER BY id DESC`);\n const executedNames = executed.map((r: any) => r.name as string);\n\n const ctx = createMigrationContext(driver, registry);\n const migrationMap = new Map(migrations.map((m) => [m.name, m]));\n const rolledBack: string[] = [];\n\n const toRollback = executedNames.slice(0, steps);\n\n for (const name of toRollback) {\n const migration = migrationMap.get(name);\n if (!migration) {\n continue;\n }\n\n await migration.down(ctx);\n\n await driver.execute(`DELETE FROM \\`${MIGRATIONS_TABLE}\\` WHERE \\`name\\` = ?`, [name]);\n rolledBack.push(name);\n }\n\n return rolledBack;\n}\n"],"mappings":";;;;;;;;;;;;AAmLA,SAAS,oBACP,YACuB;AACvB,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,QAAO,OAAO,OAAO,WAAW,CAAC,QAC9B,MACC,KAAK,QAAQ,OAAO,MAAM,YAAY,UAAU,KAAK,QAAQ,KAAK,UAAU,EAC/E;;AAGH,SAAS,iBAAmD,SAAiB;AAC3E,KAAI,MAAM,QAAQ,QAAQ,CACxB,QAAO;AAET,QAAO,OAAO,OAAO,QAAQ,CAAC,QAC3B,MAAwB,KAAK,QAAQ,OAAO,MAAM,YAAa,EAAU,YAAY,KACvF;;AAGH,SAAS,kBAAkB,YAAsC;CAC/D,MAAM,sBAAM,IAAI,KAAqB;AACrC,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,MAAK,MAAM,CAAC,KAAK,UAAU,OAAO,QAAQ,WAAW,CACnD,KAAI,SAAS,QAAQ,OAAO,UAAU,YAAa,MAAc,YAAY,KAC3E,KAAI,IAAK,MAAc,QAAQ,IAAI;AAGvC,QAAO;;AAGT,SAAgB,SACd,QACqB;CACrB,MAAM,eAAe,kBAAkB,OAAO,QAAQ;CACtD,MAAM,UAAU,iBAAiB,OAAO,QAAQ;CAChD,MAAM,WAAW,IAAIA,iBAAAA,gBAAgB;AACrC,UAAS,SAAS,QAAQ;CAE1B,MAAM,SAAS,OAAO,OAAO,aAAa,WAAW,OAAO,WAAW,OAAO,SAAS;CAEvF,IAAI;AACJ,KAAI,MAAM,QAAQ,OAAO,WAAW,CAClC,KAAI,WAAW,SACb,oBAAmB,OAAO,WAAW,MAAM,MAAM,cAAc,EAAE,IAAI,OAAO,WAAW;UAC9E,WAAW,WAAW,WAAW,UAC1C,oBAAmB,OAAO,WAAW,MAAM,MAAM,UAAU,EAAE,IAAI,OAAO,WAAW;UAC1E,WAAW,OACpB,oBAAmB,OAAO,WAAW,MAAM,MAAM,eAAe,EAAE,IAAI,OAAO,WAAW;KAExF,oBAAmB,OAAO,WAAW;KAGvC,oBAAmB,OAAO;AAG5B,KAAI,CAAC,iBACH,OAAM,IAAIC,eAAAA,cAAc,gCAAgC;CAG1D,MAAM,SAASC,cAAAA,aAAa,OAAO,UAAU,iBAAqC;CAClF,MAAM,QAAQ,IAAIC,gBAAAA,aAAa,OAAO,MAAM;CAE5C,IAAI,cAAc;CAClB,IAAI,cAAoC;CAExC,MAAM,oBAAoB,YAAY;AACpC,MAAI,YACF;AAEF,MAAI,YACF,QAAO;AAET,gBAAc,QAAQ;AACtB,QAAM;AACN,gBAAc;;CAGhB,MAAM,SAAS,YAAY;AACzB,QAAM,OAAO,SAAS;AAEtB,MAAI,OAAO,aAAa;GACtB,MAAM,cAAc,OAAO,qBAAqB,MAAM,OAAO,oBAAoB,GAAG;GAEpF,MAAM,eAAe,MAAM,KAAK,SAAS,cAAc,CAAC,CAAC,IAAI,OAAO,GAAG,eAAe;IACpF,MAAM,eAAe,cAAc,YAAY,UAAU,QAAQ;AAGjE,QAAI,EAFW,cAAc,CAAC,CAAC,eAAe,MAAM,OAAO,YAAY,UAAU,KAAK,EAGpF,OAAM,OAAO,YAAY,UAAU;SAC9B;KACL,MAAM,OAAO,gBAAiB,MAAM,OAAO,gBAAgB,UAAU,KAAK;KAC1E,MAAM,gBAAgB,IAAI,IAAI,KAAK,KAAK,MAAM,EAAE,KAAK,CAAC;KACtD,MAAM,oBAAoB,UAAU,QACjC,QAAQ,YAAY,CAAC,cAAc,IAAI,QAAQ,KAAK,CAAC,CACrD,KAAK,YAAY,OAAO,UAAU,UAAU,MAAM,QAAQ,CAAC;AAC9D,WAAM,QAAQ,IAAI,kBAAkB;;KAEtC;AACF,SAAM,QAAQ,IAAI,aAAa;;AAGjC,MAAI,OAAO,iBAAiB,OAAO,YAAY;GAC7C,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,OAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;CAKvD,MAAM,wBAAQ,IAAI,KAAmC;AACrD,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,IAAIC,mBAAAA,gBAAgB,OAAO,QAAQ,QAAQ,OAAO,UAAU,OAAO,QAAQ;AACxF,QAAM,IAAI,OAAO,QAAQ,KAAK;;CAGhC,MAAM,WAAgB;EACpB,YAAY,YAAY;AACtB,SAAM,mBAAmB;;EAE3B,OAAO,YAAY;AACjB,SAAM,OAAO,YAAY;AACzB,SAAM,OAAO;AACb,iBAAc;AACd,iBAAc;;EAEhB,iBAAiB;EACjB,mBAAmB;EACnB,gBAAgB;EAChB,KAAK,OAAO,KAAa,WAAuB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,MAAM,KAAK,OAAO;;EAElC,aAAa,OAAU,OAAyB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,YAAY,GAAG;;EAE/B,WAAW,YAAY;AACrB,SAAM,mBAAmB;AACzB,OAAI,OAAO,YAAY;IACrB,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,QAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;EAIvD,aAAa,OAAO,QAAQ,MAAM;AAChC,SAAM,mBAAmB;AACzB,OAAI,OAAO,WAET,QAAO,mBAAmB,QAAQ,UADf,oBAAoB,OAAO,WAAW,EACD,MAAM;AAEhE,UAAO,EAAE;;EAEZ;AAED,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,MAAM,IAAI,OAAO,OAAO;EACrC,MAAM,QAAQ,IAAI,MAAM,MAAM,EAC5B,IAAI,QAAQ,MAAM,UAAU;GAC1B,MAAM,WAAW,QAAQ,IAAI,QAAQ,MAAM,SAAS;AACpD,OAAI,OAAO,aAAa,WACtB,QAAO,OAAO,GAAG,SAAgB;AAC/B,UAAM,mBAAmB;AACzB,WAAO,SAAS,MAAM,QAAQ,KAAK;;AAGvC,UAAO;KAEV,CAAC;EACF,MAAM,cAAc,aAAa,IAAI,OAAO,OAAO,IAAI,OAAO;AAC9D,WAAS,eAAe;;AAG1B,QAAO;;AAGT,SAAS,yBAAwC;AAC/C,QAAO;EACL,MAAMC,kBAAAA;EACN,SAAS;GACP;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACd,QAAQ;IACT;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACF;EACF;;AAGH,SAAS,uBACP,QACA,UACqE;AACrE,QAAO;EACL,QAAQ;GACN,aAAa,OAAO,aAA0B;IAC5C,MAAM,OAAO,SAAS,SAAS,SAAS,OAAO,IAAI;KACjD,MAAM,SAAS;KACf,SAAS,CAAC,GAAG,SAAS,UAAU;KACjC;AACD,UAAM,OAAO,YAAY,KAAK;;GAEhC,WAAW,OAAO,SAAiB;AACjC,UAAM,OAAO,UAAU,KAAK;;GAE9B,WAAW,OAAO,OAAe,OAAe,WAAgB;AAC9D,UAAM,OAAO,UAAU,OAAO,OAAO;;GAEvC,YAAY,OAAO,OAAe,SAAiB;AACjD,UAAM,OAAO,WAAW,OAAO,KAAK;;GAEtC,cAAc,OAAO,OAAe,SAAiB,YAAoB;AACvE,UAAM,OAAO,aAAa,OAAO,SAAS,QAAQ;;GAEpD,UAAU,YAAY;GACtB,WAAW,YAAY;GACxB;EACD,KAAK,OAAO,OAAe,WAAuB;AAChD,UAAO,OAAO,QAAQ,OAAO,OAAO;;EAEvC;;AAGH,eAAe,sBAAsB,QAAuC;AAE1E,KAAI,CADW,MAAM,OAAO,YAAA,wBAA6B,CAEvD,OAAM,OAAO,YAAY,wBAAwB,CAAC;;AAItD,eAAe,cACb,QACA,UACA,YACe;AACf,OAAM,sBAAsB,OAAO;CAEnC,MAAM,WAAW,MAAM,OAAO,MAAM,sBAAsBA,kBAAAA,iBAAiB,IAAI;CAC/E,MAAM,gBAAgB,IAAI,IAAI,SAAS,KAAK,MAAW,EAAE,KAAK,CAAC;CAE/D,MAAM,MAAM,uBAAuB,QAAQ,SAAS;AAEpD,MAAK,MAAM,aAAa,YAAY;AAClC,MAAI,cAAc,IAAI,UAAU,KAAK,CACnC;AAGF,QAAM,UAAU,GAAG,IAAI;AAEvB,QAAM,OAAO,QACX,iBAAiBA,kBAAAA,iBAAiB,+CAClC,CAAC,UAAU,sBAAM,IAAI,MAAM,CAAC,CAC7B;;;AAIL,eAAe,mBACb,QACA,UACA,YACA,QAAQ,GACW;AACnB,OAAM,sBAAsB,OAAO;CAGnC,MAAM,iBADW,MAAM,OAAO,MAAM,sBAAsBA,kBAAAA,iBAAiB,qBAAqB,EACjE,KAAK,MAAW,EAAE,KAAe;CAEhE,MAAM,MAAM,uBAAuB,QAAQ,SAAS;CACpD,MAAM,eAAe,IAAI,IAAI,WAAW,KAAK,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;CAChE,MAAM,aAAuB,EAAE;CAE/B,MAAM,aAAa,cAAc,MAAM,GAAG,MAAM;AAEhD,MAAK,MAAM,QAAQ,YAAY;EAC7B,MAAM,YAAY,aAAa,IAAI,KAAK;AACxC,MAAI,CAAC,UACH;AAGF,QAAM,UAAU,KAAK,IAAI;AAEzB,QAAM,OAAO,QAAQ,iBAAiBA,kBAAAA,iBAAiB,wBAAwB,CAAC,KAAK,CAAC;AACtF,aAAW,KAAK,KAAK;;AAGvB,QAAO"}
@@ -21,7 +21,14 @@ function normalizeSchemas(schemas) {
21
21
  if (Array.isArray(schemas)) return schemas;
22
22
  return Object.values(schemas).filter((v) => v != null && typeof v === "object" && v.__table === true);
23
23
  }
24
+ function buildSchemaKeyMap(rawSchemas) {
25
+ const map = /* @__PURE__ */ new Map();
26
+ if (Array.isArray(rawSchemas)) return map;
27
+ for (const [key, value] of Object.entries(rawSchemas)) if (value != null && typeof value === "object" && value.__table === true) map.set(value.__name, key);
28
+ return map;
29
+ }
24
30
  function database(config) {
31
+ const schemaKeyMap = buildSchemaKeyMap(config.schemas);
25
32
  const schemas = normalizeSchemas(config.schemas);
26
33
  const registry = new SchemaRegistry();
27
34
  registry.register(schemas);
@@ -114,7 +121,8 @@ function database(config) {
114
121
  };
115
122
  return original;
116
123
  } });
117
- instance[schema.__name] = proxy;
124
+ const accessorKey = schemaKeyMap.get(schema.__name) ?? schema.__name;
125
+ instance[accessorKey] = proxy;
118
126
  }
119
127
  return instance;
120
128
  }
@@ -157,8 +165,11 @@ function createMigrationContext(driver, registry) {
157
165
  return {
158
166
  schema: {
159
167
  createTable: async (tableDef) => {
160
- const meta = registry.getTable(tableDef.__name);
161
- if (meta) await driver.createTable(meta);
168
+ const meta = registry.getTable(tableDef.__name) ?? {
169
+ name: tableDef.__name,
170
+ columns: [...tableDef.__columns]
171
+ };
172
+ await driver.createTable(meta);
162
173
  },
163
174
  dropTable: async (name) => {
164
175
  await driver.dropTable(name);
@@ -1 +1 @@
1
- {"version":3,"file":"database.mjs","names":[],"sources":["../../src/core/database.ts"],"sourcesContent":["import { CacheManager } from \"../cache\";\nimport { MIGRATIONS_TABLE } from \"../constants\";\nimport { createDriver } from \"../drivers\";\nimport { DatabaseError } from \"../errors\";\nimport { SchemaRegistry } from \"../schema\";\nimport type {\n AnyTableDef,\n ConnectionConfig,\n DatabaseConfig,\n DatabaseDriver,\n DeleteOptions,\n InferRow,\n InferSchemas,\n MigrationDefinition,\n QueryOptions,\n RelationQueryMap,\n ResolveResult,\n TableMetadata,\n UpdateOptions,\n WhereClause,\n} from \"../types\";\nimport { TableRepository } from \"./repository\";\n\ntype TypedTableRepository<S extends readonly AnyTableDef[], T extends AnyTableDef> = {\n /**\n * Find all rows matching the given options\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n find<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find all rows matching the given options (alias for {@link find})\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n findMany<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find the first row matching the given options\n * @param options - Filter, sort, and eagerly load relations\n * @returns The first matching row, or `null` if none found\n */\n findFirst<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O> | null>;\n\n /**\n * Insert one or more rows into the table\n * @param data - Row data (or array of row data) to insert\n * @returns The inserted row\n */\n insert(data: Partial<InferRow<T>> | Partial<InferRow<T>>[]): Promise<InferRow<T>>;\n\n /**\n * Insert multiple rows into the table\n * @param data - Array of row data to insert\n * @returns Array of inserted rows\n */\n insertMany(data: Partial<InferRow<T>>[]): Promise<InferRow<T>[]>;\n\n /**\n * Update rows matching the where clause\n * @param options - Where clause and partial data to apply\n * @returns Array of updated rows\n */\n update(options: UpdateOptions<InferRow<T>>): Promise<InferRow<T>[]>;\n\n /**\n * Delete rows matching the where clause\n * @param options - Where clause to select rows for deletion\n * @returns Number of deleted rows\n */\n delete(options: DeleteOptions<InferRow<T>>): Promise<number>;\n\n /**\n * Count rows matching the where clause\n * @param options - Optional where clause to filter rows\n * @returns Number of matching rows\n */\n count(options?: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<number>;\n\n /**\n * Check whether at least one row matches the where clause\n * @param options - Where clause to check\n * @returns `true` if a matching row exists, `false` otherwise\n */\n exists(options: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<boolean>;\n\n /**\n * Insert a row if it doesn't exist, or update it if it does\n * @param options - Where clause to check, data to create, and data to update\n * @returns The created or updated row\n */\n upsert(options: {\n where: WhereClause<InferRow<T>>;\n create: Partial<InferRow<T>>;\n update: Partial<InferRow<T>>;\n }): Promise<InferRow<T>>;\n\n /**\n * Remove all rows from the table\n * @returns Resolves when the table has been truncated\n */\n truncate(): Promise<void>;\n};\n\ntype ExtractRepos<S> = S extends readonly AnyTableDef[]\n ? { [T in S[number] as T[\"__name\"]]: TypedTableRepository<S, T> }\n : S extends Record<string, any>\n ? {\n [K in keyof S as S[K] extends AnyTableDef ? K : never]: TypedTableRepository<\n InferSchemas<S>,\n Extract<S[K], AnyTableDef>\n >;\n }\n : never;\n\ntype DatabaseInstance<S> = ExtractRepos<S> & {\n /**\n * Initialize the database connection, create tables and run migrations\n * @returns {Promise<void>}\n */\n initialize(): Promise<void>;\n /**\n * Close the database connection\n * @returns {Promise<void>}\n */\n close(): Promise<void>;\n /**\n * Get the underlying database driver\n * @returns {DatabaseDriver} The database driver\n */\n getDriver(): DatabaseDriver;\n /**\n * Get the schema registry\n * @returns {SchemaRegistry} The schema registry\n */\n getRegistry(): SchemaRegistry;\n /**\n * Get the cache manager\n * @returns {CacheManager} The cache manager\n */\n getCache(): CacheManager;\n /**\n * Execute a raw SQL query\n * @param {string} sql - SQL query\n * @param {unknown[]} [params] - Query parameters\n * @returns {Promise<any[]>} Query results\n */\n raw(sql: string, params?: unknown[]): Promise<any[]>;\n /**\n * Execute within a transaction\n * @param {() => Promise<T>} fn - Function to execute\n * @returns {Promise<T>} Result\n */\n transaction<T>(fn: () => Promise<T>): Promise<T>;\n /**\n * Run pending migrations manually\n * @returns {Promise<void>}\n */\n migrateUp(): Promise<void>;\n /**\n * Rollback migrations\n * @param {number} [steps=1] - Number of migrations to rollback\n * @returns {Promise<string[]>} Names of rolled back migrations\n */\n migrateDown(steps?: number): Promise<string[]>;\n};\n\n/**\n * Create a database instance with typed repositories for each schema\n * @param {DatabaseConfig} config - Database configuration\n * @returns {DatabaseInstance<S>} Database instance with table repositories\n */\nfunction normalizeMigrations(\n migrations: MigrationDefinition[] | Record<string, unknown>,\n): MigrationDefinition[] {\n if (Array.isArray(migrations)) {\n return migrations;\n }\n return Object.values(migrations).filter(\n (v): v is MigrationDefinition =>\n v != null && typeof v === \"object\" && \"name\" in v && \"up\" in v && \"down\" in v,\n );\n}\n\nfunction normalizeSchemas<S extends readonly AnyTableDef[]>(schemas: any): S {\n if (Array.isArray(schemas)) {\n return schemas as any as S;\n }\n return Object.values(schemas).filter(\n (v): v is AnyTableDef => v != null && typeof v === \"object\" && (v as any).__table === true,\n ) as any;\n}\n\nexport function database<const S extends readonly AnyTableDef[] | Record<string, any>>(\n config: DatabaseConfig & { schemas: S },\n): DatabaseInstance<S> {\n const schemas = normalizeSchemas(config.schemas) as any;\n const registry = new SchemaRegistry();\n registry.register(schemas);\n\n const dbName = typeof config.database === \"string\" ? config.database : config.database.name;\n\n let connectionConfig: ConnectionConfig | undefined;\n if (Array.isArray(config.connection)) {\n if (dbName === \"sqlite\") {\n connectionConfig = config.connection.find((c) => \"filename\" in c) ?? config.connection[0];\n } else if (dbName === \"mysql\" || dbName === \"mariadb\") {\n connectionConfig = config.connection.find((c) => \"host\" in c) ?? config.connection[0];\n } else if (dbName === \"file\") {\n connectionConfig = config.connection.find((c) => \"directory\" in c) ?? config.connection[0];\n } else {\n connectionConfig = config.connection[0];\n }\n } else {\n connectionConfig = config.connection;\n }\n\n if (!connectionConfig) {\n throw new DatabaseError(\"Connection config is required\");\n }\n\n const driver = createDriver(config.database, connectionConfig as ConnectionConfig);\n const cache = new CacheManager(config.cache);\n\n let initialized = false;\n let initPromise: Promise<void> | null = null;\n\n const ensureInitialized = async () => {\n if (initialized) {\n return;\n }\n if (initPromise) {\n return initPromise;\n }\n initPromise = doInit();\n await initPromise;\n initialized = true;\n };\n\n const doInit = async () => {\n await driver.connect();\n\n if (config.syncSchemas) {\n const allMetadata = driver.getAllTableColumns ? await driver.getAllTableColumns() : null;\n\n const syncPromises = Array.from(registry.getAllTables()).map(async ([, tableMeta]) => {\n const existingCols = allMetadata ? allMetadata[tableMeta.name] : null;\n const exists = allMetadata ? !!existingCols : await driver.tableExists(tableMeta.name);\n\n if (!exists) {\n await driver.createTable(tableMeta);\n } else {\n const cols = existingCols || (await driver.getTableColumns(tableMeta.name));\n const existingNames = new Set(cols.map((c) => c.name));\n const addColumnPromises = tableMeta.columns\n .filter((colMeta) => !existingNames.has(colMeta.name))\n .map((colMeta) => driver.addColumn(tableMeta.name, colMeta));\n await Promise.all(addColumnPromises);\n }\n });\n await Promise.all(syncPromises);\n }\n\n if (config.runMigrations && config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n };\n\n const repos = new Map<string, TableRepository<any>>();\n for (const schema of schemas) {\n const repo = new TableRepository(schema.__name, driver, cache, registry, schema.__cache);\n repos.set(schema.__name, repo);\n }\n\n const instance: any = {\n initialize: async () => {\n await ensureInitialized();\n },\n close: async () => {\n await driver.disconnect();\n cache.clear();\n initialized = false;\n initPromise = null;\n },\n getDriver: () => driver,\n getRegistry: () => registry,\n getCache: () => cache,\n raw: async (sql: string, params?: unknown[]) => {\n await ensureInitialized();\n return driver.query(sql, params);\n },\n transaction: async <T>(fn: () => Promise<T>) => {\n await ensureInitialized();\n return driver.transaction(fn);\n },\n migrateUp: async () => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n },\n migrateDown: async (steps = 1) => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n return rollbackMigrations(driver, registry, migrations, steps);\n }\n return [];\n },\n };\n\n for (const schema of schemas) {\n const repo = repos.get(schema.__name)!;\n const proxy = new Proxy(repo, {\n get(target, prop, receiver) {\n const original = Reflect.get(target, prop, receiver);\n if (typeof original === \"function\") {\n return async (...args: any[]) => {\n await ensureInitialized();\n return original.apply(target, args);\n };\n }\n return original;\n },\n });\n instance[schema.__name] = proxy;\n }\n\n return instance as any as DatabaseInstance<S>;\n}\n\nfunction getMigrationsTableMeta(): TableMetadata {\n return {\n name: MIGRATIONS_TABLE,\n columns: [\n {\n name: \"id\",\n type: \"integer\",\n primaryKey: true,\n autoIncrement: true,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n },\n {\n name: \"name\",\n type: \"varchar\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n length: 255,\n },\n {\n name: \"executed_at\",\n type: \"datetime\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: false,\n defaultValue: undefined,\n },\n ],\n };\n}\n\nfunction createMigrationContext(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n): MigrationDefinition[\"up\"] extends (ctx: infer C) => any ? C : never {\n return {\n schema: {\n createTable: async (tableDef: AnyTableDef) => {\n const meta = registry.getTable(tableDef.__name);\n if (meta) {\n await driver.createTable(meta);\n }\n },\n dropTable: async (name: string) => {\n await driver.dropTable(name);\n },\n addColumn: async (table: string, _name: string, column: any) => {\n await driver.addColumn(table, column);\n },\n dropColumn: async (table: string, name: string) => {\n await driver.dropColumn(table, name);\n },\n renameColumn: async (table: string, oldName: string, newName: string) => {\n await driver.renameColumn(table, oldName, newName);\n },\n addIndex: async () => {},\n dropIndex: async () => {},\n },\n sql: async (query: string, params?: unknown[]) => {\n return driver.execute(query, params);\n },\n };\n}\n\nasync function ensureMigrationsTable(driver: DatabaseDriver): Promise<void> {\n const exists = await driver.tableExists(MIGRATIONS_TABLE);\n if (!exists) {\n await driver.createTable(getMigrationsTableMeta());\n }\n}\n\nasync function runMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n): Promise<void> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\``);\n const executedNames = new Set(executed.map((r: any) => r.name));\n\n const ctx = createMigrationContext(driver, registry);\n\n for (const migration of migrations) {\n if (executedNames.has(migration.name)) {\n continue;\n }\n\n await migration.up(ctx);\n\n await driver.execute(\n `INSERT INTO \\`${MIGRATIONS_TABLE}\\` (\\`name\\`, \\`executed_at\\`) VALUES (?, ?)`,\n [migration.name, new Date()],\n );\n }\n}\n\nasync function rollbackMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n steps = 1,\n): Promise<string[]> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\` ORDER BY id DESC`);\n const executedNames = executed.map((r: any) => r.name as string);\n\n const ctx = createMigrationContext(driver, registry);\n const migrationMap = new Map(migrations.map((m) => [m.name, m]));\n const rolledBack: string[] = [];\n\n const toRollback = executedNames.slice(0, steps);\n\n for (const name of toRollback) {\n const migration = migrationMap.get(name);\n if (!migration) {\n continue;\n }\n\n await migration.down(ctx);\n\n await driver.execute(`DELETE FROM \\`${MIGRATIONS_TABLE}\\` WHERE \\`name\\` = ?`, [name]);\n rolledBack.push(name);\n }\n\n return rolledBack;\n}\n"],"mappings":";;;;;;;;;;;;;;;AAmLA,SAAS,oBACP,YACuB;AACvB,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,QAAO,OAAO,OAAO,WAAW,CAAC,QAC9B,MACC,KAAK,QAAQ,OAAO,MAAM,YAAY,UAAU,KAAK,QAAQ,KAAK,UAAU,EAC/E;;AAGH,SAAS,iBAAmD,SAAiB;AAC3E,KAAI,MAAM,QAAQ,QAAQ,CACxB,QAAO;AAET,QAAO,OAAO,OAAO,QAAQ,CAAC,QAC3B,MAAwB,KAAK,QAAQ,OAAO,MAAM,YAAa,EAAU,YAAY,KACvF;;AAGH,SAAgB,SACd,QACqB;CACrB,MAAM,UAAU,iBAAiB,OAAO,QAAQ;CAChD,MAAM,WAAW,IAAI,gBAAgB;AACrC,UAAS,SAAS,QAAQ;CAE1B,MAAM,SAAS,OAAO,OAAO,aAAa,WAAW,OAAO,WAAW,OAAO,SAAS;CAEvF,IAAI;AACJ,KAAI,MAAM,QAAQ,OAAO,WAAW,CAClC,KAAI,WAAW,SACb,oBAAmB,OAAO,WAAW,MAAM,MAAM,cAAc,EAAE,IAAI,OAAO,WAAW;UAC9E,WAAW,WAAW,WAAW,UAC1C,oBAAmB,OAAO,WAAW,MAAM,MAAM,UAAU,EAAE,IAAI,OAAO,WAAW;UAC1E,WAAW,OACpB,oBAAmB,OAAO,WAAW,MAAM,MAAM,eAAe,EAAE,IAAI,OAAO,WAAW;KAExF,oBAAmB,OAAO,WAAW;KAGvC,oBAAmB,OAAO;AAG5B,KAAI,CAAC,iBACH,OAAM,IAAI,cAAc,gCAAgC;CAG1D,MAAM,SAAS,aAAa,OAAO,UAAU,iBAAqC;CAClF,MAAM,QAAQ,IAAI,aAAa,OAAO,MAAM;CAE5C,IAAI,cAAc;CAClB,IAAI,cAAoC;CAExC,MAAM,oBAAoB,YAAY;AACpC,MAAI,YACF;AAEF,MAAI,YACF,QAAO;AAET,gBAAc,QAAQ;AACtB,QAAM;AACN,gBAAc;;CAGhB,MAAM,SAAS,YAAY;AACzB,QAAM,OAAO,SAAS;AAEtB,MAAI,OAAO,aAAa;GACtB,MAAM,cAAc,OAAO,qBAAqB,MAAM,OAAO,oBAAoB,GAAG;GAEpF,MAAM,eAAe,MAAM,KAAK,SAAS,cAAc,CAAC,CAAC,IAAI,OAAO,GAAG,eAAe;IACpF,MAAM,eAAe,cAAc,YAAY,UAAU,QAAQ;AAGjE,QAAI,EAFW,cAAc,CAAC,CAAC,eAAe,MAAM,OAAO,YAAY,UAAU,KAAK,EAGpF,OAAM,OAAO,YAAY,UAAU;SAC9B;KACL,MAAM,OAAO,gBAAiB,MAAM,OAAO,gBAAgB,UAAU,KAAK;KAC1E,MAAM,gBAAgB,IAAI,IAAI,KAAK,KAAK,MAAM,EAAE,KAAK,CAAC;KACtD,MAAM,oBAAoB,UAAU,QACjC,QAAQ,YAAY,CAAC,cAAc,IAAI,QAAQ,KAAK,CAAC,CACrD,KAAK,YAAY,OAAO,UAAU,UAAU,MAAM,QAAQ,CAAC;AAC9D,WAAM,QAAQ,IAAI,kBAAkB;;KAEtC;AACF,SAAM,QAAQ,IAAI,aAAa;;AAGjC,MAAI,OAAO,iBAAiB,OAAO,YAAY;GAC7C,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,OAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;CAKvD,MAAM,wBAAQ,IAAI,KAAmC;AACrD,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,IAAI,gBAAgB,OAAO,QAAQ,QAAQ,OAAO,UAAU,OAAO,QAAQ;AACxF,QAAM,IAAI,OAAO,QAAQ,KAAK;;CAGhC,MAAM,WAAgB;EACpB,YAAY,YAAY;AACtB,SAAM,mBAAmB;;EAE3B,OAAO,YAAY;AACjB,SAAM,OAAO,YAAY;AACzB,SAAM,OAAO;AACb,iBAAc;AACd,iBAAc;;EAEhB,iBAAiB;EACjB,mBAAmB;EACnB,gBAAgB;EAChB,KAAK,OAAO,KAAa,WAAuB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,MAAM,KAAK,OAAO;;EAElC,aAAa,OAAU,OAAyB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,YAAY,GAAG;;EAE/B,WAAW,YAAY;AACrB,SAAM,mBAAmB;AACzB,OAAI,OAAO,YAAY;IACrB,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,QAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;EAIvD,aAAa,OAAO,QAAQ,MAAM;AAChC,SAAM,mBAAmB;AACzB,OAAI,OAAO,WAET,QAAO,mBAAmB,QAAQ,UADf,oBAAoB,OAAO,WAAW,EACD,MAAM;AAEhE,UAAO,EAAE;;EAEZ;AAED,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,MAAM,IAAI,OAAO,OAAO;EACrC,MAAM,QAAQ,IAAI,MAAM,MAAM,EAC5B,IAAI,QAAQ,MAAM,UAAU;GAC1B,MAAM,WAAW,QAAQ,IAAI,QAAQ,MAAM,SAAS;AACpD,OAAI,OAAO,aAAa,WACtB,QAAO,OAAO,GAAG,SAAgB;AAC/B,UAAM,mBAAmB;AACzB,WAAO,SAAS,MAAM,QAAQ,KAAK;;AAGvC,UAAO;KAEV,CAAC;AACF,WAAS,OAAO,UAAU;;AAG5B,QAAO;;AAGT,SAAS,yBAAwC;AAC/C,QAAO;EACL,MAAM;EACN,SAAS;GACP;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACd,QAAQ;IACT;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACF;EACF;;AAGH,SAAS,uBACP,QACA,UACqE;AACrE,QAAO;EACL,QAAQ;GACN,aAAa,OAAO,aAA0B;IAC5C,MAAM,OAAO,SAAS,SAAS,SAAS,OAAO;AAC/C,QAAI,KACF,OAAM,OAAO,YAAY,KAAK;;GAGlC,WAAW,OAAO,SAAiB;AACjC,UAAM,OAAO,UAAU,KAAK;;GAE9B,WAAW,OAAO,OAAe,OAAe,WAAgB;AAC9D,UAAM,OAAO,UAAU,OAAO,OAAO;;GAEvC,YAAY,OAAO,OAAe,SAAiB;AACjD,UAAM,OAAO,WAAW,OAAO,KAAK;;GAEtC,cAAc,OAAO,OAAe,SAAiB,YAAoB;AACvE,UAAM,OAAO,aAAa,OAAO,SAAS,QAAQ;;GAEpD,UAAU,YAAY;GACtB,WAAW,YAAY;GACxB;EACD,KAAK,OAAO,OAAe,WAAuB;AAChD,UAAO,OAAO,QAAQ,OAAO,OAAO;;EAEvC;;AAGH,eAAe,sBAAsB,QAAuC;AAE1E,KAAI,CADW,MAAM,OAAO,YAAA,wBAA6B,CAEvD,OAAM,OAAO,YAAY,wBAAwB,CAAC;;AAItD,eAAe,cACb,QACA,UACA,YACe;AACf,OAAM,sBAAsB,OAAO;CAEnC,MAAM,WAAW,MAAM,OAAO,MAAM,sBAAsB,iBAAiB,IAAI;CAC/E,MAAM,gBAAgB,IAAI,IAAI,SAAS,KAAK,MAAW,EAAE,KAAK,CAAC;CAE/D,MAAM,MAAM,uBAAuB,QAAQ,SAAS;AAEpD,MAAK,MAAM,aAAa,YAAY;AAClC,MAAI,cAAc,IAAI,UAAU,KAAK,CACnC;AAGF,QAAM,UAAU,GAAG,IAAI;AAEvB,QAAM,OAAO,QACX,iBAAiB,iBAAiB,+CAClC,CAAC,UAAU,sBAAM,IAAI,MAAM,CAAC,CAC7B;;;AAIL,eAAe,mBACb,QACA,UACA,YACA,QAAQ,GACW;AACnB,OAAM,sBAAsB,OAAO;CAGnC,MAAM,iBADW,MAAM,OAAO,MAAM,sBAAsB,iBAAiB,qBAAqB,EACjE,KAAK,MAAW,EAAE,KAAe;CAEhE,MAAM,MAAM,uBAAuB,QAAQ,SAAS;CACpD,MAAM,eAAe,IAAI,IAAI,WAAW,KAAK,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;CAChE,MAAM,aAAuB,EAAE;CAE/B,MAAM,aAAa,cAAc,MAAM,GAAG,MAAM;AAEhD,MAAK,MAAM,QAAQ,YAAY;EAC7B,MAAM,YAAY,aAAa,IAAI,KAAK;AACxC,MAAI,CAAC,UACH;AAGF,QAAM,UAAU,KAAK,IAAI;AAEzB,QAAM,OAAO,QAAQ,iBAAiB,iBAAiB,wBAAwB,CAAC,KAAK,CAAC;AACtF,aAAW,KAAK,KAAK;;AAGvB,QAAO;;;aA3d+B;iBACQ;eACN;cACA;cACC;kBAiBI"}
1
+ {"version":3,"file":"database.mjs","names":[],"sources":["../../src/core/database.ts"],"sourcesContent":["import { CacheManager } from \"../cache\";\nimport { MIGRATIONS_TABLE } from \"../constants\";\nimport { createDriver } from \"../drivers\";\nimport { DatabaseError } from \"../errors\";\nimport { SchemaRegistry } from \"../schema\";\nimport type {\n AnyTableDef,\n ConnectionConfig,\n DatabaseConfig,\n DatabaseDriver,\n DeleteOptions,\n InferRow,\n InferSchemas,\n MigrationDefinition,\n QueryOptions,\n RelationQueryMap,\n ResolveResult,\n TableMetadata,\n UpdateOptions,\n WhereClause,\n} from \"../types\";\nimport { TableRepository } from \"./repository\";\n\ntype TypedTableRepository<S extends readonly AnyTableDef[], T extends AnyTableDef> = {\n /**\n * Find all rows matching the given options\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n find<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find all rows matching the given options (alias for {@link find})\n * @param options - Filter, sort, paginate, and eagerly load relations\n * @returns Array of matching rows\n */\n findMany<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O>[]>;\n\n /**\n * Find the first row matching the given options\n * @param options - Filter, sort, and eagerly load relations\n * @returns The first matching row, or `null` if none found\n */\n findFirst<O extends QueryOptions<InferRow<T>, RelationQueryMap<S, T>> = {}>(\n options?: O,\n ): Promise<ResolveResult<S, T, O> | null>;\n\n /**\n * Insert one or more rows into the table\n * @param data - Row data (or array of row data) to insert\n * @returns The inserted row\n */\n insert(data: Partial<InferRow<T>> | Partial<InferRow<T>>[]): Promise<InferRow<T>>;\n\n /**\n * Insert multiple rows into the table\n * @param data - Array of row data to insert\n * @returns Array of inserted rows\n */\n insertMany(data: Partial<InferRow<T>>[]): Promise<InferRow<T>[]>;\n\n /**\n * Update rows matching the where clause\n * @param options - Where clause and partial data to apply\n * @returns Array of updated rows\n */\n update(options: UpdateOptions<InferRow<T>>): Promise<InferRow<T>[]>;\n\n /**\n * Delete rows matching the where clause\n * @param options - Where clause to select rows for deletion\n * @returns Number of deleted rows\n */\n delete(options: DeleteOptions<InferRow<T>>): Promise<number>;\n\n /**\n * Count rows matching the where clause\n * @param options - Optional where clause to filter rows\n * @returns Number of matching rows\n */\n count(options?: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<number>;\n\n /**\n * Check whether at least one row matches the where clause\n * @param options - Where clause to check\n * @returns `true` if a matching row exists, `false` otherwise\n */\n exists(options: Pick<QueryOptions<InferRow<T>>, \"where\">): Promise<boolean>;\n\n /**\n * Insert a row if it doesn't exist, or update it if it does\n * @param options - Where clause to check, data to create, and data to update\n * @returns The created or updated row\n */\n upsert(options: {\n where: WhereClause<InferRow<T>>;\n create: Partial<InferRow<T>>;\n update: Partial<InferRow<T>>;\n }): Promise<InferRow<T>>;\n\n /**\n * Remove all rows from the table\n * @returns Resolves when the table has been truncated\n */\n truncate(): Promise<void>;\n};\n\ntype ExtractRepos<S> = S extends readonly AnyTableDef[]\n ? { [T in S[number] as T[\"__name\"]]: TypedTableRepository<S, T> }\n : S extends Record<string, any>\n ? {\n [K in keyof S as S[K] extends AnyTableDef ? K : never]: TypedTableRepository<\n InferSchemas<S>,\n Extract<S[K], AnyTableDef>\n >;\n }\n : never;\n\ntype DatabaseInstance<S> = ExtractRepos<S> & {\n /**\n * Initialize the database connection, create tables and run migrations\n * @returns {Promise<void>}\n */\n initialize(): Promise<void>;\n /**\n * Close the database connection\n * @returns {Promise<void>}\n */\n close(): Promise<void>;\n /**\n * Get the underlying database driver\n * @returns {DatabaseDriver} The database driver\n */\n getDriver(): DatabaseDriver;\n /**\n * Get the schema registry\n * @returns {SchemaRegistry} The schema registry\n */\n getRegistry(): SchemaRegistry;\n /**\n * Get the cache manager\n * @returns {CacheManager} The cache manager\n */\n getCache(): CacheManager;\n /**\n * Execute a raw SQL query\n * @param {string} sql - SQL query\n * @param {unknown[]} [params] - Query parameters\n * @returns {Promise<any[]>} Query results\n */\n raw(sql: string, params?: unknown[]): Promise<any[]>;\n /**\n * Execute within a transaction\n * @param {() => Promise<T>} fn - Function to execute\n * @returns {Promise<T>} Result\n */\n transaction<T>(fn: () => Promise<T>): Promise<T>;\n /**\n * Run pending migrations manually\n * @returns {Promise<void>}\n */\n migrateUp(): Promise<void>;\n /**\n * Rollback migrations\n * @param {number} [steps=1] - Number of migrations to rollback\n * @returns {Promise<string[]>} Names of rolled back migrations\n */\n migrateDown(steps?: number): Promise<string[]>;\n};\n\n/**\n * Create a database instance with typed repositories for each schema\n * @param {DatabaseConfig} config - Database configuration\n * @returns {DatabaseInstance<S>} Database instance with table repositories\n */\nfunction normalizeMigrations(\n migrations: MigrationDefinition[] | Record<string, unknown>,\n): MigrationDefinition[] {\n if (Array.isArray(migrations)) {\n return migrations;\n }\n return Object.values(migrations).filter(\n (v): v is MigrationDefinition =>\n v != null && typeof v === \"object\" && \"name\" in v && \"up\" in v && \"down\" in v,\n );\n}\n\nfunction normalizeSchemas<S extends readonly AnyTableDef[]>(schemas: any): S {\n if (Array.isArray(schemas)) {\n return schemas as any as S;\n }\n return Object.values(schemas).filter(\n (v): v is AnyTableDef => v != null && typeof v === \"object\" && (v as any).__table === true,\n ) as any;\n}\n\nfunction buildSchemaKeyMap(rawSchemas: any): Map<string, string> {\n const map = new Map<string, string>();\n if (Array.isArray(rawSchemas)) {\n return map;\n }\n for (const [key, value] of Object.entries(rawSchemas)) {\n if (value != null && typeof value === \"object\" && (value as any).__table === true) {\n map.set((value as any).__name, key);\n }\n }\n return map;\n}\n\nexport function database<const S extends readonly AnyTableDef[] | Record<string, any>>(\n config: DatabaseConfig & { schemas: S },\n): DatabaseInstance<S> {\n const schemaKeyMap = buildSchemaKeyMap(config.schemas);\n const schemas = normalizeSchemas(config.schemas) as any;\n const registry = new SchemaRegistry();\n registry.register(schemas);\n\n const dbName = typeof config.database === \"string\" ? config.database : config.database.name;\n\n let connectionConfig: ConnectionConfig | undefined;\n if (Array.isArray(config.connection)) {\n if (dbName === \"sqlite\") {\n connectionConfig = config.connection.find((c) => \"filename\" in c) ?? config.connection[0];\n } else if (dbName === \"mysql\" || dbName === \"mariadb\") {\n connectionConfig = config.connection.find((c) => \"host\" in c) ?? config.connection[0];\n } else if (dbName === \"file\") {\n connectionConfig = config.connection.find((c) => \"directory\" in c) ?? config.connection[0];\n } else {\n connectionConfig = config.connection[0];\n }\n } else {\n connectionConfig = config.connection;\n }\n\n if (!connectionConfig) {\n throw new DatabaseError(\"Connection config is required\");\n }\n\n const driver = createDriver(config.database, connectionConfig as ConnectionConfig);\n const cache = new CacheManager(config.cache);\n\n let initialized = false;\n let initPromise: Promise<void> | null = null;\n\n const ensureInitialized = async () => {\n if (initialized) {\n return;\n }\n if (initPromise) {\n return initPromise;\n }\n initPromise = doInit();\n await initPromise;\n initialized = true;\n };\n\n const doInit = async () => {\n await driver.connect();\n\n if (config.syncSchemas) {\n const allMetadata = driver.getAllTableColumns ? await driver.getAllTableColumns() : null;\n\n const syncPromises = Array.from(registry.getAllTables()).map(async ([, tableMeta]) => {\n const existingCols = allMetadata ? allMetadata[tableMeta.name] : null;\n const exists = allMetadata ? !!existingCols : await driver.tableExists(tableMeta.name);\n\n if (!exists) {\n await driver.createTable(tableMeta);\n } else {\n const cols = existingCols || (await driver.getTableColumns(tableMeta.name));\n const existingNames = new Set(cols.map((c) => c.name));\n const addColumnPromises = tableMeta.columns\n .filter((colMeta) => !existingNames.has(colMeta.name))\n .map((colMeta) => driver.addColumn(tableMeta.name, colMeta));\n await Promise.all(addColumnPromises);\n }\n });\n await Promise.all(syncPromises);\n }\n\n if (config.runMigrations && config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n };\n\n const repos = new Map<string, TableRepository<any>>();\n for (const schema of schemas) {\n const repo = new TableRepository(schema.__name, driver, cache, registry, schema.__cache);\n repos.set(schema.__name, repo);\n }\n\n const instance: any = {\n initialize: async () => {\n await ensureInitialized();\n },\n close: async () => {\n await driver.disconnect();\n cache.clear();\n initialized = false;\n initPromise = null;\n },\n getDriver: () => driver,\n getRegistry: () => registry,\n getCache: () => cache,\n raw: async (sql: string, params?: unknown[]) => {\n await ensureInitialized();\n return driver.query(sql, params);\n },\n transaction: async <T>(fn: () => Promise<T>) => {\n await ensureInitialized();\n return driver.transaction(fn);\n },\n migrateUp: async () => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n if (migrations.length > 0) {\n await runMigrations(driver, registry, migrations);\n }\n }\n },\n migrateDown: async (steps = 1) => {\n await ensureInitialized();\n if (config.migrations) {\n const migrations = normalizeMigrations(config.migrations);\n return rollbackMigrations(driver, registry, migrations, steps);\n }\n return [];\n },\n };\n\n for (const schema of schemas) {\n const repo = repos.get(schema.__name)!;\n const proxy = new Proxy(repo, {\n get(target, prop, receiver) {\n const original = Reflect.get(target, prop, receiver);\n if (typeof original === \"function\") {\n return async (...args: any[]) => {\n await ensureInitialized();\n return original.apply(target, args);\n };\n }\n return original;\n },\n });\n const accessorKey = schemaKeyMap.get(schema.__name) ?? schema.__name;\n instance[accessorKey] = proxy;\n }\n\n return instance as any as DatabaseInstance<S>;\n}\n\nfunction getMigrationsTableMeta(): TableMetadata {\n return {\n name: MIGRATIONS_TABLE,\n columns: [\n {\n name: \"id\",\n type: \"integer\",\n primaryKey: true,\n autoIncrement: true,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n },\n {\n name: \"name\",\n type: \"varchar\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: true,\n defaultValue: undefined,\n length: 255,\n },\n {\n name: \"executed_at\",\n type: \"datetime\",\n primaryKey: false,\n autoIncrement: false,\n notNull: true,\n unique: false,\n defaultValue: undefined,\n },\n ],\n };\n}\n\nfunction createMigrationContext(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n): MigrationDefinition[\"up\"] extends (ctx: infer C) => any ? C : never {\n return {\n schema: {\n createTable: async (tableDef: AnyTableDef) => {\n const meta = registry.getTable(tableDef.__name) ?? {\n name: tableDef.__name,\n columns: [...tableDef.__columns],\n };\n await driver.createTable(meta);\n },\n dropTable: async (name: string) => {\n await driver.dropTable(name);\n },\n addColumn: async (table: string, _name: string, column: any) => {\n await driver.addColumn(table, column);\n },\n dropColumn: async (table: string, name: string) => {\n await driver.dropColumn(table, name);\n },\n renameColumn: async (table: string, oldName: string, newName: string) => {\n await driver.renameColumn(table, oldName, newName);\n },\n addIndex: async () => {},\n dropIndex: async () => {},\n },\n sql: async (query: string, params?: unknown[]) => {\n return driver.execute(query, params);\n },\n };\n}\n\nasync function ensureMigrationsTable(driver: DatabaseDriver): Promise<void> {\n const exists = await driver.tableExists(MIGRATIONS_TABLE);\n if (!exists) {\n await driver.createTable(getMigrationsTableMeta());\n }\n}\n\nasync function runMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n): Promise<void> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\``);\n const executedNames = new Set(executed.map((r: any) => r.name));\n\n const ctx = createMigrationContext(driver, registry);\n\n for (const migration of migrations) {\n if (executedNames.has(migration.name)) {\n continue;\n }\n\n await migration.up(ctx);\n\n await driver.execute(\n `INSERT INTO \\`${MIGRATIONS_TABLE}\\` (\\`name\\`, \\`executed_at\\`) VALUES (?, ?)`,\n [migration.name, new Date()],\n );\n }\n}\n\nasync function rollbackMigrations(\n driver: DatabaseDriver,\n registry: SchemaRegistry,\n migrations: MigrationDefinition[],\n steps = 1,\n): Promise<string[]> {\n await ensureMigrationsTable(driver);\n\n const executed = await driver.query(`SELECT name FROM \\`${MIGRATIONS_TABLE}\\` ORDER BY id DESC`);\n const executedNames = executed.map((r: any) => r.name as string);\n\n const ctx = createMigrationContext(driver, registry);\n const migrationMap = new Map(migrations.map((m) => [m.name, m]));\n const rolledBack: string[] = [];\n\n const toRollback = executedNames.slice(0, steps);\n\n for (const name of toRollback) {\n const migration = migrationMap.get(name);\n if (!migration) {\n continue;\n }\n\n await migration.down(ctx);\n\n await driver.execute(`DELETE FROM \\`${MIGRATIONS_TABLE}\\` WHERE \\`name\\` = ?`, [name]);\n rolledBack.push(name);\n }\n\n return rolledBack;\n}\n"],"mappings":";;;;;;;;;;;;;;;AAmLA,SAAS,oBACP,YACuB;AACvB,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,QAAO,OAAO,OAAO,WAAW,CAAC,QAC9B,MACC,KAAK,QAAQ,OAAO,MAAM,YAAY,UAAU,KAAK,QAAQ,KAAK,UAAU,EAC/E;;AAGH,SAAS,iBAAmD,SAAiB;AAC3E,KAAI,MAAM,QAAQ,QAAQ,CACxB,QAAO;AAET,QAAO,OAAO,OAAO,QAAQ,CAAC,QAC3B,MAAwB,KAAK,QAAQ,OAAO,MAAM,YAAa,EAAU,YAAY,KACvF;;AAGH,SAAS,kBAAkB,YAAsC;CAC/D,MAAM,sBAAM,IAAI,KAAqB;AACrC,KAAI,MAAM,QAAQ,WAAW,CAC3B,QAAO;AAET,MAAK,MAAM,CAAC,KAAK,UAAU,OAAO,QAAQ,WAAW,CACnD,KAAI,SAAS,QAAQ,OAAO,UAAU,YAAa,MAAc,YAAY,KAC3E,KAAI,IAAK,MAAc,QAAQ,IAAI;AAGvC,QAAO;;AAGT,SAAgB,SACd,QACqB;CACrB,MAAM,eAAe,kBAAkB,OAAO,QAAQ;CACtD,MAAM,UAAU,iBAAiB,OAAO,QAAQ;CAChD,MAAM,WAAW,IAAI,gBAAgB;AACrC,UAAS,SAAS,QAAQ;CAE1B,MAAM,SAAS,OAAO,OAAO,aAAa,WAAW,OAAO,WAAW,OAAO,SAAS;CAEvF,IAAI;AACJ,KAAI,MAAM,QAAQ,OAAO,WAAW,CAClC,KAAI,WAAW,SACb,oBAAmB,OAAO,WAAW,MAAM,MAAM,cAAc,EAAE,IAAI,OAAO,WAAW;UAC9E,WAAW,WAAW,WAAW,UAC1C,oBAAmB,OAAO,WAAW,MAAM,MAAM,UAAU,EAAE,IAAI,OAAO,WAAW;UAC1E,WAAW,OACpB,oBAAmB,OAAO,WAAW,MAAM,MAAM,eAAe,EAAE,IAAI,OAAO,WAAW;KAExF,oBAAmB,OAAO,WAAW;KAGvC,oBAAmB,OAAO;AAG5B,KAAI,CAAC,iBACH,OAAM,IAAI,cAAc,gCAAgC;CAG1D,MAAM,SAAS,aAAa,OAAO,UAAU,iBAAqC;CAClF,MAAM,QAAQ,IAAI,aAAa,OAAO,MAAM;CAE5C,IAAI,cAAc;CAClB,IAAI,cAAoC;CAExC,MAAM,oBAAoB,YAAY;AACpC,MAAI,YACF;AAEF,MAAI,YACF,QAAO;AAET,gBAAc,QAAQ;AACtB,QAAM;AACN,gBAAc;;CAGhB,MAAM,SAAS,YAAY;AACzB,QAAM,OAAO,SAAS;AAEtB,MAAI,OAAO,aAAa;GACtB,MAAM,cAAc,OAAO,qBAAqB,MAAM,OAAO,oBAAoB,GAAG;GAEpF,MAAM,eAAe,MAAM,KAAK,SAAS,cAAc,CAAC,CAAC,IAAI,OAAO,GAAG,eAAe;IACpF,MAAM,eAAe,cAAc,YAAY,UAAU,QAAQ;AAGjE,QAAI,EAFW,cAAc,CAAC,CAAC,eAAe,MAAM,OAAO,YAAY,UAAU,KAAK,EAGpF,OAAM,OAAO,YAAY,UAAU;SAC9B;KACL,MAAM,OAAO,gBAAiB,MAAM,OAAO,gBAAgB,UAAU,KAAK;KAC1E,MAAM,gBAAgB,IAAI,IAAI,KAAK,KAAK,MAAM,EAAE,KAAK,CAAC;KACtD,MAAM,oBAAoB,UAAU,QACjC,QAAQ,YAAY,CAAC,cAAc,IAAI,QAAQ,KAAK,CAAC,CACrD,KAAK,YAAY,OAAO,UAAU,UAAU,MAAM,QAAQ,CAAC;AAC9D,WAAM,QAAQ,IAAI,kBAAkB;;KAEtC;AACF,SAAM,QAAQ,IAAI,aAAa;;AAGjC,MAAI,OAAO,iBAAiB,OAAO,YAAY;GAC7C,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,OAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;CAKvD,MAAM,wBAAQ,IAAI,KAAmC;AACrD,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,IAAI,gBAAgB,OAAO,QAAQ,QAAQ,OAAO,UAAU,OAAO,QAAQ;AACxF,QAAM,IAAI,OAAO,QAAQ,KAAK;;CAGhC,MAAM,WAAgB;EACpB,YAAY,YAAY;AACtB,SAAM,mBAAmB;;EAE3B,OAAO,YAAY;AACjB,SAAM,OAAO,YAAY;AACzB,SAAM,OAAO;AACb,iBAAc;AACd,iBAAc;;EAEhB,iBAAiB;EACjB,mBAAmB;EACnB,gBAAgB;EAChB,KAAK,OAAO,KAAa,WAAuB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,MAAM,KAAK,OAAO;;EAElC,aAAa,OAAU,OAAyB;AAC9C,SAAM,mBAAmB;AACzB,UAAO,OAAO,YAAY,GAAG;;EAE/B,WAAW,YAAY;AACrB,SAAM,mBAAmB;AACzB,OAAI,OAAO,YAAY;IACrB,MAAM,aAAa,oBAAoB,OAAO,WAAW;AACzD,QAAI,WAAW,SAAS,EACtB,OAAM,cAAc,QAAQ,UAAU,WAAW;;;EAIvD,aAAa,OAAO,QAAQ,MAAM;AAChC,SAAM,mBAAmB;AACzB,OAAI,OAAO,WAET,QAAO,mBAAmB,QAAQ,UADf,oBAAoB,OAAO,WAAW,EACD,MAAM;AAEhE,UAAO,EAAE;;EAEZ;AAED,MAAK,MAAM,UAAU,SAAS;EAC5B,MAAM,OAAO,MAAM,IAAI,OAAO,OAAO;EACrC,MAAM,QAAQ,IAAI,MAAM,MAAM,EAC5B,IAAI,QAAQ,MAAM,UAAU;GAC1B,MAAM,WAAW,QAAQ,IAAI,QAAQ,MAAM,SAAS;AACpD,OAAI,OAAO,aAAa,WACtB,QAAO,OAAO,GAAG,SAAgB;AAC/B,UAAM,mBAAmB;AACzB,WAAO,SAAS,MAAM,QAAQ,KAAK;;AAGvC,UAAO;KAEV,CAAC;EACF,MAAM,cAAc,aAAa,IAAI,OAAO,OAAO,IAAI,OAAO;AAC9D,WAAS,eAAe;;AAG1B,QAAO;;AAGT,SAAS,yBAAwC;AAC/C,QAAO;EACL,MAAM;EACN,SAAS;GACP;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACd,QAAQ;IACT;GACD;IACE,MAAM;IACN,MAAM;IACN,YAAY;IACZ,eAAe;IACf,SAAS;IACT,QAAQ;IACR,cAAc,KAAA;IACf;GACF;EACF;;AAGH,SAAS,uBACP,QACA,UACqE;AACrE,QAAO;EACL,QAAQ;GACN,aAAa,OAAO,aAA0B;IAC5C,MAAM,OAAO,SAAS,SAAS,SAAS,OAAO,IAAI;KACjD,MAAM,SAAS;KACf,SAAS,CAAC,GAAG,SAAS,UAAU;KACjC;AACD,UAAM,OAAO,YAAY,KAAK;;GAEhC,WAAW,OAAO,SAAiB;AACjC,UAAM,OAAO,UAAU,KAAK;;GAE9B,WAAW,OAAO,OAAe,OAAe,WAAgB;AAC9D,UAAM,OAAO,UAAU,OAAO,OAAO;;GAEvC,YAAY,OAAO,OAAe,SAAiB;AACjD,UAAM,OAAO,WAAW,OAAO,KAAK;;GAEtC,cAAc,OAAO,OAAe,SAAiB,YAAoB;AACvE,UAAM,OAAO,aAAa,OAAO,SAAS,QAAQ;;GAEpD,UAAU,YAAY;GACtB,WAAW,YAAY;GACxB;EACD,KAAK,OAAO,OAAe,WAAuB;AAChD,UAAO,OAAO,QAAQ,OAAO,OAAO;;EAEvC;;AAGH,eAAe,sBAAsB,QAAuC;AAE1E,KAAI,CADW,MAAM,OAAO,YAAA,wBAA6B,CAEvD,OAAM,OAAO,YAAY,wBAAwB,CAAC;;AAItD,eAAe,cACb,QACA,UACA,YACe;AACf,OAAM,sBAAsB,OAAO;CAEnC,MAAM,WAAW,MAAM,OAAO,MAAM,sBAAsB,iBAAiB,IAAI;CAC/E,MAAM,gBAAgB,IAAI,IAAI,SAAS,KAAK,MAAW,EAAE,KAAK,CAAC;CAE/D,MAAM,MAAM,uBAAuB,QAAQ,SAAS;AAEpD,MAAK,MAAM,aAAa,YAAY;AAClC,MAAI,cAAc,IAAI,UAAU,KAAK,CACnC;AAGF,QAAM,UAAU,GAAG,IAAI;AAEvB,QAAM,OAAO,QACX,iBAAiB,iBAAiB,+CAClC,CAAC,UAAU,sBAAM,IAAI,MAAM,CAAC,CAC7B;;;AAIL,eAAe,mBACb,QACA,UACA,YACA,QAAQ,GACW;AACnB,OAAM,sBAAsB,OAAO;CAGnC,MAAM,iBADW,MAAM,OAAO,MAAM,sBAAsB,iBAAiB,qBAAqB,EACjE,KAAK,MAAW,EAAE,KAAe;CAEhE,MAAM,MAAM,uBAAuB,QAAQ,SAAS;CACpD,MAAM,eAAe,IAAI,IAAI,WAAW,KAAK,MAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;CAChE,MAAM,aAAuB,EAAE;CAE/B,MAAM,aAAa,cAAc,MAAM,GAAG,MAAM;AAEhD,MAAK,MAAM,QAAQ,YAAY;EAC7B,MAAM,YAAY,aAAa,IAAI,KAAK;AACxC,MAAI,CAAC,UACH;AAGF,QAAM,UAAU,KAAK,IAAI;AAEzB,QAAM,OAAO,QAAQ,iBAAiB,iBAAiB,wBAAwB,CAAC,KAAK,CAAC;AACtF,aAAW,KAAK,KAAK;;AAGvB,QAAO;;;aA3e+B;iBACQ;eACN;cACA;cACC;kBAiBI"}
@@ -1,13 +1,14 @@
1
1
  //#region src/migrations/templates.ts
2
2
  /**
3
3
  * Generate a migration file template
4
- * @param {string} name - Migration name
4
+ * @param {string} id - Migration id (e.g. "20260320095958_sessions")
5
+ * @param {string} varName - Variable name for the export (camelCase)
5
6
  * @returns {string} Migration file content
6
7
  */
7
- function generateMigrationTemplate(name) {
8
+ function generateMigrationTemplate(id, varName) {
8
9
  return `import { migration } from "@hedystia/db";
9
10
 
10
- export default migration("${name}", {
11
+ export const ${varName} = migration("${id}", {
11
12
  async up({ schema, sql }) {
12
13
  // Add your migration logic here
13
14
  },
@@ -1 +1 @@
1
- {"version":3,"file":"templates.cjs","names":[],"sources":["../../src/migrations/templates.ts"],"sourcesContent":["/**\n * Generate a migration file template\n * @param {string} name - Migration name\n * @returns {string} Migration file content\n */\nexport function generateMigrationTemplate(name: string): string {\n return `import { migration } from \"@hedystia/db\";\n\nexport default migration(\"${name}\", {\n async up({ schema, sql }) {\n // Add your migration logic here\n },\n async down({ schema, sql }) {\n // Add your rollback logic here\n },\n});\n`;\n}\n\n/**\n * Generate a schema file template\n * @param {string} name - Table name\n * @returns {string} Schema file content\n */\nexport function generateSchemaTemplate(name: string): string {\n return `import { table, integer, datetime } from \"@hedystia/db\";\n\nexport const ${name} = table(\"${name}\", {\n id: integer().primaryKey().autoIncrement(),\n createdAt: datetime().default(new Date()),\n updatedAt: datetime().default(new Date()),\n});\n`;\n}\n"],"mappings":";;;;;;AAKA,SAAgB,0BAA0B,MAAsB;AAC9D,QAAO;;4BAEmB,KAAK;;;;;;;;;;;;;;;AAgBjC,SAAgB,uBAAuB,MAAsB;AAC3D,QAAO;;eAEM,KAAK,YAAY,KAAK"}
1
+ {"version":3,"file":"templates.cjs","names":[],"sources":["../../src/migrations/templates.ts"],"sourcesContent":["/**\n * Generate a migration file template\n * @param {string} id - Migration id (e.g. \"20260320095958_sessions\")\n * @param {string} varName - Variable name for the export (camelCase)\n * @returns {string} Migration file content\n */\nexport function generateMigrationTemplate(id: string, varName: string): string {\n return `import { migration } from \"@hedystia/db\";\n\nexport const ${varName} = migration(\"${id}\", {\n async up({ schema, sql }) {\n // Add your migration logic here\n },\n async down({ schema, sql }) {\n // Add your rollback logic here\n },\n});\n`;\n}\n\n/**\n * Generate a schema file template\n * @param {string} name - Table name\n * @returns {string} Schema file content\n */\nexport function generateSchemaTemplate(name: string): string {\n return `import { table, integer, datetime } from \"@hedystia/db\";\n\nexport const ${name} = table(\"${name}\", {\n id: integer().primaryKey().autoIncrement(),\n createdAt: datetime().default(new Date()),\n updatedAt: datetime().default(new Date()),\n});\n`;\n}\n"],"mappings":";;;;;;;AAMA,SAAgB,0BAA0B,IAAY,SAAyB;AAC7E,QAAO;;eAEM,QAAQ,gBAAgB,GAAG;;;;;;;;;;;;;;;AAgB1C,SAAgB,uBAAuB,MAAsB;AAC3D,QAAO;;eAEM,KAAK,YAAY,KAAK"}
@@ -1,10 +1,11 @@
1
1
  //#region src/migrations/templates.d.ts
2
2
  /**
3
3
  * Generate a migration file template
4
- * @param {string} name - Migration name
4
+ * @param {string} id - Migration id (e.g. "20260320095958_sessions")
5
+ * @param {string} varName - Variable name for the export (camelCase)
5
6
  * @returns {string} Migration file content
6
7
  */
7
- declare function generateMigrationTemplate(name: string): string;
8
+ declare function generateMigrationTemplate(id: string, varName: string): string;
8
9
  /**
9
10
  * Generate a schema file template
10
11
  * @param {string} name - Table name
@@ -1,10 +1,11 @@
1
1
  //#region src/migrations/templates.d.ts
2
2
  /**
3
3
  * Generate a migration file template
4
- * @param {string} name - Migration name
4
+ * @param {string} id - Migration id (e.g. "20260320095958_sessions")
5
+ * @param {string} varName - Variable name for the export (camelCase)
5
6
  * @returns {string} Migration file content
6
7
  */
7
- declare function generateMigrationTemplate(name: string): string;
8
+ declare function generateMigrationTemplate(id: string, varName: string): string;
8
9
  /**
9
10
  * Generate a schema file template
10
11
  * @param {string} name - Table name
@@ -2,13 +2,14 @@ import { __esmMin } from "../_virtual/_rolldown/runtime.mjs";
2
2
  //#region src/migrations/templates.ts
3
3
  /**
4
4
  * Generate a migration file template
5
- * @param {string} name - Migration name
5
+ * @param {string} id - Migration id (e.g. "20260320095958_sessions")
6
+ * @param {string} varName - Variable name for the export (camelCase)
6
7
  * @returns {string} Migration file content
7
8
  */
8
- function generateMigrationTemplate(name) {
9
+ function generateMigrationTemplate(id, varName) {
9
10
  return `import { migration } from "@hedystia/db";
10
11
 
11
- export default migration("${name}", {
12
+ export const ${varName} = migration("${id}", {
12
13
  async up({ schema, sql }) {
13
14
  // Add your migration logic here
14
15
  },
@@ -1 +1 @@
1
- {"version":3,"file":"templates.mjs","names":[],"sources":["../../src/migrations/templates.ts"],"sourcesContent":["/**\n * Generate a migration file template\n * @param {string} name - Migration name\n * @returns {string} Migration file content\n */\nexport function generateMigrationTemplate(name: string): string {\n return `import { migration } from \"@hedystia/db\";\n\nexport default migration(\"${name}\", {\n async up({ schema, sql }) {\n // Add your migration logic here\n },\n async down({ schema, sql }) {\n // Add your rollback logic here\n },\n});\n`;\n}\n\n/**\n * Generate a schema file template\n * @param {string} name - Table name\n * @returns {string} Schema file content\n */\nexport function generateSchemaTemplate(name: string): string {\n return `import { table, integer, datetime } from \"@hedystia/db\";\n\nexport const ${name} = table(\"${name}\", {\n id: integer().primaryKey().autoIncrement(),\n createdAt: datetime().default(new Date()),\n updatedAt: datetime().default(new Date()),\n});\n`;\n}\n"],"mappings":";;;;;;;AAKA,SAAgB,0BAA0B,MAAsB;AAC9D,QAAO;;4BAEmB,KAAK;;;;;;;;;;;;;;;AAgBjC,SAAgB,uBAAuB,MAAsB;AAC3D,QAAO;;eAEM,KAAK,YAAY,KAAK"}
1
+ {"version":3,"file":"templates.mjs","names":[],"sources":["../../src/migrations/templates.ts"],"sourcesContent":["/**\n * Generate a migration file template\n * @param {string} id - Migration id (e.g. \"20260320095958_sessions\")\n * @param {string} varName - Variable name for the export (camelCase)\n * @returns {string} Migration file content\n */\nexport function generateMigrationTemplate(id: string, varName: string): string {\n return `import { migration } from \"@hedystia/db\";\n\nexport const ${varName} = migration(\"${id}\", {\n async up({ schema, sql }) {\n // Add your migration logic here\n },\n async down({ schema, sql }) {\n // Add your rollback logic here\n },\n});\n`;\n}\n\n/**\n * Generate a schema file template\n * @param {string} name - Table name\n * @returns {string} Schema file content\n */\nexport function generateSchemaTemplate(name: string): string {\n return `import { table, integer, datetime } from \"@hedystia/db\";\n\nexport const ${name} = table(\"${name}\", {\n id: integer().primaryKey().autoIncrement(),\n createdAt: datetime().default(new Date()),\n updatedAt: datetime().default(new Date()),\n});\n`;\n}\n"],"mappings":";;;;;;;;AAMA,SAAgB,0BAA0B,IAAY,SAAyB;AAC7E,QAAO;;eAEM,QAAQ,gBAAgB,GAAG;;;;;;;;;;;;;;;AAgB1C,SAAgB,uBAAuB,MAAsB;AAC3D,QAAO;;eAEM,KAAK,YAAY,KAAK"}
@@ -7,7 +7,16 @@ function generateTimestamp() {
7
7
  const now = /* @__PURE__ */ new Date();
8
8
  return `${now.getFullYear()}${String(now.getMonth() + 1).padStart(2, "0")}${String(now.getDate()).padStart(2, "0")}${String(now.getHours()).padStart(2, "0")}${String(now.getMinutes()).padStart(2, "0")}${String(now.getSeconds()).padStart(2, "0")}`;
9
9
  }
10
+ /**
11
+ * Convert a snake_case string to camelCase
12
+ * @param {string} str - Input string
13
+ * @returns {string} Camel case string
14
+ */
15
+ function toCamelCase(str) {
16
+ return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());
17
+ }
10
18
  //#endregion
11
19
  exports.generateTimestamp = generateTimestamp;
20
+ exports.toCamelCase = toCamelCase;
12
21
 
13
22
  //# sourceMappingURL=naming.cjs.map
@@ -1 +1 @@
1
- {"version":3,"file":"naming.cjs","names":[],"sources":["../../src/utils/naming.ts"],"sourcesContent":["/**\n * Generate a timestamp string for migration file naming\n * @returns {string} Formatted timestamp\n */\nexport function generateTimestamp(): string {\n const now = new Date();\n const y = now.getFullYear();\n const m = String(now.getMonth() + 1).padStart(2, \"0\");\n const day = String(now.getDate()).padStart(2, \"0\");\n const h = String(now.getHours()).padStart(2, \"0\");\n const min = String(now.getMinutes()).padStart(2, \"0\");\n const s = String(now.getSeconds()).padStart(2, \"0\");\n return `${y}${m}${day}${h}${min}${s}`;\n}\n\n/**\n * Convert a camelCase or PascalCase string to snake_case\n * @param {string} str - Input string\n * @returns {string} Snake case string\n */\nexport function toSnakeCase(str: string): string {\n return str\n .replace(/([A-Z])/g, \"_$1\")\n .toLowerCase()\n .replace(/^_/, \"\");\n}\n\n/**\n * Convert a snake_case string to camelCase\n * @param {string} str - Input string\n * @returns {string} Camel case string\n */\nexport function toCamelCase(str: string): string {\n return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());\n}\n"],"mappings":";;;;;AAIA,SAAgB,oBAA4B;CAC1C,MAAM,sBAAM,IAAI,MAAM;AAOtB,QAAO,GANG,IAAI,aAAa,GACjB,OAAO,IAAI,UAAU,GAAG,EAAE,CAAC,SAAS,GAAG,IAAI,GACzC,OAAO,IAAI,SAAS,CAAC,CAAC,SAAS,GAAG,IAAI,GACxC,OAAO,IAAI,UAAU,CAAC,CAAC,SAAS,GAAG,IAAI,GACrC,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI,GAC3C,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI"}
1
+ {"version":3,"file":"naming.cjs","names":[],"sources":["../../src/utils/naming.ts"],"sourcesContent":["/**\n * Generate a timestamp string for migration file naming\n * @returns {string} Formatted timestamp\n */\nexport function generateTimestamp(): string {\n const now = new Date();\n const y = now.getFullYear();\n const m = String(now.getMonth() + 1).padStart(2, \"0\");\n const day = String(now.getDate()).padStart(2, \"0\");\n const h = String(now.getHours()).padStart(2, \"0\");\n const min = String(now.getMinutes()).padStart(2, \"0\");\n const s = String(now.getSeconds()).padStart(2, \"0\");\n return `${y}${m}${day}${h}${min}${s}`;\n}\n\n/**\n * Convert a camelCase or PascalCase string to snake_case\n * @param {string} str - Input string\n * @returns {string} Snake case string\n */\nexport function toSnakeCase(str: string): string {\n return str\n .replace(/([A-Z])/g, \"_$1\")\n .toLowerCase()\n .replace(/^_/, \"\");\n}\n\n/**\n * Convert a snake_case string to camelCase\n * @param {string} str - Input string\n * @returns {string} Camel case string\n */\nexport function toCamelCase(str: string): string {\n return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());\n}\n"],"mappings":";;;;;AAIA,SAAgB,oBAA4B;CAC1C,MAAM,sBAAM,IAAI,MAAM;AAOtB,QAAO,GANG,IAAI,aAAa,GACjB,OAAO,IAAI,UAAU,GAAG,EAAE,CAAC,SAAS,GAAG,IAAI,GACzC,OAAO,IAAI,SAAS,CAAC,CAAC,SAAS,GAAG,IAAI,GACxC,OAAO,IAAI,UAAU,CAAC,CAAC,SAAS,GAAG,IAAI,GACrC,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI,GAC3C,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI;;;;;;;AAqBrD,SAAgB,YAAY,KAAqB;AAC/C,QAAO,IAAI,QAAQ,cAAc,GAAG,MAAM,EAAE,aAAa,CAAC"}
@@ -8,9 +8,17 @@ function generateTimestamp() {
8
8
  const now = /* @__PURE__ */ new Date();
9
9
  return `${now.getFullYear()}${String(now.getMonth() + 1).padStart(2, "0")}${String(now.getDate()).padStart(2, "0")}${String(now.getHours()).padStart(2, "0")}${String(now.getMinutes()).padStart(2, "0")}${String(now.getSeconds()).padStart(2, "0")}`;
10
10
  }
11
+ /**
12
+ * Convert a snake_case string to camelCase
13
+ * @param {string} str - Input string
14
+ * @returns {string} Camel case string
15
+ */
16
+ function toCamelCase(str) {
17
+ return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());
18
+ }
11
19
  var init_naming = __esmMin((() => {}));
12
20
  //#endregion
13
21
  init_naming();
14
- export { generateTimestamp, init_naming };
22
+ export { generateTimestamp, init_naming, toCamelCase };
15
23
 
16
24
  //# sourceMappingURL=naming.mjs.map
@@ -1 +1 @@
1
- {"version":3,"file":"naming.mjs","names":[],"sources":["../../src/utils/naming.ts"],"sourcesContent":["/**\n * Generate a timestamp string for migration file naming\n * @returns {string} Formatted timestamp\n */\nexport function generateTimestamp(): string {\n const now = new Date();\n const y = now.getFullYear();\n const m = String(now.getMonth() + 1).padStart(2, \"0\");\n const day = String(now.getDate()).padStart(2, \"0\");\n const h = String(now.getHours()).padStart(2, \"0\");\n const min = String(now.getMinutes()).padStart(2, \"0\");\n const s = String(now.getSeconds()).padStart(2, \"0\");\n return `${y}${m}${day}${h}${min}${s}`;\n}\n\n/**\n * Convert a camelCase or PascalCase string to snake_case\n * @param {string} str - Input string\n * @returns {string} Snake case string\n */\nexport function toSnakeCase(str: string): string {\n return str\n .replace(/([A-Z])/g, \"_$1\")\n .toLowerCase()\n .replace(/^_/, \"\");\n}\n\n/**\n * Convert a snake_case string to camelCase\n * @param {string} str - Input string\n * @returns {string} Camel case string\n */\nexport function toCamelCase(str: string): string {\n return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());\n}\n"],"mappings":";;;;;;AAIA,SAAgB,oBAA4B;CAC1C,MAAM,sBAAM,IAAI,MAAM;AAOtB,QAAO,GANG,IAAI,aAAa,GACjB,OAAO,IAAI,UAAU,GAAG,EAAE,CAAC,SAAS,GAAG,IAAI,GACzC,OAAO,IAAI,SAAS,CAAC,CAAC,SAAS,GAAG,IAAI,GACxC,OAAO,IAAI,UAAU,CAAC,CAAC,SAAS,GAAG,IAAI,GACrC,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI,GAC3C,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI"}
1
+ {"version":3,"file":"naming.mjs","names":[],"sources":["../../src/utils/naming.ts"],"sourcesContent":["/**\n * Generate a timestamp string for migration file naming\n * @returns {string} Formatted timestamp\n */\nexport function generateTimestamp(): string {\n const now = new Date();\n const y = now.getFullYear();\n const m = String(now.getMonth() + 1).padStart(2, \"0\");\n const day = String(now.getDate()).padStart(2, \"0\");\n const h = String(now.getHours()).padStart(2, \"0\");\n const min = String(now.getMinutes()).padStart(2, \"0\");\n const s = String(now.getSeconds()).padStart(2, \"0\");\n return `${y}${m}${day}${h}${min}${s}`;\n}\n\n/**\n * Convert a camelCase or PascalCase string to snake_case\n * @param {string} str - Input string\n * @returns {string} Snake case string\n */\nexport function toSnakeCase(str: string): string {\n return str\n .replace(/([A-Z])/g, \"_$1\")\n .toLowerCase()\n .replace(/^_/, \"\");\n}\n\n/**\n * Convert a snake_case string to camelCase\n * @param {string} str - Input string\n * @returns {string} Camel case string\n */\nexport function toCamelCase(str: string): string {\n return str.replace(/_([a-z])/g, (_, c) => c.toUpperCase());\n}\n"],"mappings":";;;;;;AAIA,SAAgB,oBAA4B;CAC1C,MAAM,sBAAM,IAAI,MAAM;AAOtB,QAAO,GANG,IAAI,aAAa,GACjB,OAAO,IAAI,UAAU,GAAG,EAAE,CAAC,SAAS,GAAG,IAAI,GACzC,OAAO,IAAI,SAAS,CAAC,CAAC,SAAS,GAAG,IAAI,GACxC,OAAO,IAAI,UAAU,CAAC,CAAC,SAAS,GAAG,IAAI,GACrC,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI,GAC3C,OAAO,IAAI,YAAY,CAAC,CAAC,SAAS,GAAG,IAAI;;;;;;;AAqBrD,SAAgB,YAAY,KAAqB;AAC/C,QAAO,IAAI,QAAQ,cAAc,GAAG,MAAM,EAAE,aAAa,CAAC"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@hedystia/db",
3
- "version": "2.0.7",
3
+ "version": "2.0.9",
4
4
  "description": "Next-gen TypeScript ORM for building type-safe database layers at lightspeed",
5
5
  "homepage": "https://docs.hedystia.com",
6
6
  "devDependencies": {