@conorroberts/utils 0.0.10 → 0.0.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cache.cjs ADDED
@@ -0,0 +1,56 @@
1
+ "use strict";
2
+ var __create = Object.create;
3
+ var __defProp = Object.defineProperty;
4
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
+ var __getOwnPropNames = Object.getOwnPropertyNames;
6
+ var __getProtoOf = Object.getPrototypeOf;
7
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
8
+ var __export = (target, all) => {
9
+ for (var name in all)
10
+ __defProp(target, name, { get: all[name], enumerable: true });
11
+ };
12
+ var __copyProps = (to, from, except, desc) => {
13
+ if (from && typeof from === "object" || typeof from === "function") {
14
+ for (let key of __getOwnPropNames(from))
15
+ if (!__hasOwnProp.call(to, key) && key !== except)
16
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
17
+ }
18
+ return to;
19
+ };
20
+ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
+ // If the importer is in node compatibility mode or this is not an ESM
22
+ // file that has been converted to a CommonJS file using a Babel-
23
+ // compatible transform (i.e. "__esModule" has not been set), then set
24
+ // "default" to the CommonJS "module.exports" for node compatibility.
25
+ isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
+ mod
27
+ ));
28
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
+
30
+ // src/cache.ts
31
+ var cache_exports = {};
32
+ __export(cache_exports, {
33
+ Cache: () => Cache
34
+ });
35
+ module.exports = __toCommonJS(cache_exports);
36
+ var import_unstorage = require("unstorage");
37
+ var import_redis = __toESM(require("unstorage/drivers/redis"), 1);
38
+ var Cache = class {
39
+ _cache;
40
+ constructor(args) {
41
+ this._cache = (0, import_unstorage.createStorage)({
42
+ driver: (0, import_redis.default)(args)
43
+ });
44
+ }
45
+ get cache() {
46
+ return this._cache;
47
+ }
48
+ ttl(date) {
49
+ return Math.floor((date.getTime() - Date.now()) / 1e3);
50
+ }
51
+ };
52
+ // Annotate the CommonJS export names for ESM import in node:
53
+ 0 && (module.exports = {
54
+ Cache
55
+ });
56
+ //# sourceMappingURL=cache.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/cache.ts"],"sourcesContent":["import {\r\n createStorage as createUnstorage,\r\n type Storage,\r\n type StorageValue,\r\n} from \"unstorage\";\r\nimport redisDriver, { type RedisOptions } from \"unstorage/drivers/redis\";\r\n\r\nexport class Cache {\r\n private _cache: Storage<StorageValue>;\r\n\r\n constructor(args: RedisOptions) {\r\n this._cache = createUnstorage({\r\n driver: redisDriver(args),\r\n });\r\n }\r\n\r\n get cache() {\r\n return this._cache;\r\n }\r\n\r\n public ttl(date: Date) {\r\n return Math.floor((date.getTime() - Date.now()) / 1000);\r\n }\r\n}\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,uBAIO;AACP,mBAA+C;AAExC,IAAM,QAAN,MAAY;AAAA,EACT;AAAA,EAER,YAAY,MAAoB;AAC9B,SAAK,aAAS,iBAAAA,eAAgB;AAAA,MAC5B,YAAQ,aAAAC,SAAY,IAAI;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,IAAI,QAAQ;AACV,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,MAAY;AACrB,WAAO,KAAK,OAAO,KAAK,QAAQ,IAAI,KAAK,IAAI,KAAK,GAAI;AAAA,EACxD;AACF;","names":["createUnstorage","redisDriver"]}
@@ -0,0 +1,11 @@
1
+ import { Storage, StorageValue } from 'unstorage';
2
+ import { RedisOptions } from 'unstorage/drivers/redis';
3
+
4
+ declare class Cache {
5
+ private _cache;
6
+ constructor(args: RedisOptions);
7
+ get cache(): Storage<StorageValue>;
8
+ ttl(date: Date): number;
9
+ }
10
+
11
+ export { Cache };
package/dist/cache.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/cache.ts"],"sourcesContent":["import {\n createStorage as createUnstorage,\n type Storage,\n type StorageValue,\n} from \"unstorage\";\nimport redisDriver, { type RedisOptions } from \"unstorage/drivers/redis\";\n\nexport class Cache {\n private _cache: Storage<StorageValue>;\n\n constructor(args: RedisOptions) {\n this._cache = createUnstorage({\n driver: redisDriver(args),\n });\n }\n\n get cache() {\n return this._cache;\n }\n\n public ttl(date: Date) {\n return Math.floor((date.getTime() - Date.now()) / 1000);\n }\n}\n"],"mappings":";AAAA;AAAA,EACE,iBAAiB;AAAA,OAGZ;AACP,OAAO,iBAAwC;AAExC,IAAM,QAAN,MAAY;AAAA,EACT;AAAA,EAER,YAAY,MAAoB;AAC9B,SAAK,SAAS,gBAAgB;AAAA,MAC5B,QAAQ,YAAY,IAAI;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,IAAI,QAAQ;AACV,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,MAAY;AACrB,WAAO,KAAK,OAAO,KAAK,QAAQ,IAAI,KAAK,IAAI,KAAK,GAAI;AAAA,EACxD;AACF;","names":[]}
1
+ {"version":3,"sources":["../src/cache.ts"],"sourcesContent":["import {\r\n createStorage as createUnstorage,\r\n type Storage,\r\n type StorageValue,\r\n} from \"unstorage\";\r\nimport redisDriver, { type RedisOptions } from \"unstorage/drivers/redis\";\r\n\r\nexport class Cache {\r\n private _cache: Storage<StorageValue>;\r\n\r\n constructor(args: RedisOptions) {\r\n this._cache = createUnstorage({\r\n driver: redisDriver(args),\r\n });\r\n }\r\n\r\n get cache() {\r\n return this._cache;\r\n }\r\n\r\n public ttl(date: Date) {\r\n return Math.floor((date.getTime() - Date.now()) / 1000);\r\n }\r\n}\r\n"],"mappings":";AAAA;AAAA,EACE,iBAAiB;AAAA,OAGZ;AACP,OAAO,iBAAwC;AAExC,IAAM,QAAN,MAAY;AAAA,EACT;AAAA,EAER,YAAY,MAAoB;AAC9B,SAAK,SAAS,gBAAgB;AAAA,MAC5B,QAAQ,YAAY,IAAI;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,IAAI,QAAQ;AACV,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,MAAY;AACrB,WAAO,KAAK,OAAO,KAAK,QAAQ,IAAI,KAAK,IAAI,KAAK,GAAI;AAAA,EACxD;AACF;","names":[]}
package/dist/db.cjs ADDED
@@ -0,0 +1,45 @@
1
+ "use strict";
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __export = (target, all) => {
7
+ for (var name in all)
8
+ __defProp(target, name, { get: all[name], enumerable: true });
9
+ };
10
+ var __copyProps = (to, from, except, desc) => {
11
+ if (from && typeof from === "object" || typeof from === "function") {
12
+ for (let key of __getOwnPropNames(from))
13
+ if (!__hasOwnProp.call(to, key) && key !== except)
14
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
+ }
16
+ return to;
17
+ };
18
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
+
20
+ // src/db.ts
21
+ var db_exports = {};
22
+ __export(db_exports, {
23
+ createDbClient: () => createDbClient,
24
+ createLibsqlClient: () => createLibsqlClient
25
+ });
26
+ module.exports = __toCommonJS(db_exports);
27
+ var import_client = require("@libsql/client");
28
+ var import_libsql = require("drizzle-orm/libsql");
29
+ var createLibsqlClient = (args) => {
30
+ return (0, import_client.createClient)(args);
31
+ };
32
+ var createDbClient = (schema, args) => {
33
+ const client = createLibsqlClient(args);
34
+ const db = (0, import_libsql.drizzle)(client, {
35
+ schema,
36
+ logger: false
37
+ });
38
+ return db;
39
+ };
40
+ // Annotate the CommonJS export names for ESM import in node:
41
+ 0 && (module.exports = {
42
+ createDbClient,
43
+ createLibsqlClient
44
+ });
45
+ //# sourceMappingURL=db.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/db.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\r\nimport { LibSQLDatabase, drizzle } from \"drizzle-orm/libsql\";\r\n\r\nexport const createLibsqlClient = (args: {\r\n url: string;\r\n authToken?: string;\r\n}) => {\r\n return createClient(args);\r\n};\r\n\r\nexport const createDbClient = <TSchema extends Record<string, unknown>>(\r\n schema: TSchema,\r\n args: { url: string; authToken?: string }\r\n) => {\r\n const client = createLibsqlClient(args);\r\n const db = drizzle(client, {\r\n schema,\r\n logger: false,\r\n });\r\n\r\n return db;\r\n};\r\n\r\nexport type DatabaseClient<TSchema extends Record<string, unknown>> =\r\n LibSQLDatabase<TSchema>;\r\nexport type DatabaseClientTransactionContext<\r\n TSchema extends Record<string, unknown>\r\n> = Parameters<Parameters<DatabaseClient<TSchema>[\"transaction\"]>[0]>[0];\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAA6B;AAC7B,oBAAwC;AAEjC,IAAM,qBAAqB,CAAC,SAG7B;AACJ,aAAO,4BAAa,IAAI;AAC1B;AAEO,IAAM,iBAAiB,CAC5B,QACA,SACG;AACH,QAAM,SAAS,mBAAmB,IAAI;AACtC,QAAM,SAAK,uBAAQ,QAAQ;AAAA,IACzB;AAAA,IACA,QAAQ;AAAA,EACV,CAAC;AAED,SAAO;AACT;","names":[]}
package/dist/db.d.cts ADDED
@@ -0,0 +1,15 @@
1
+ import * as _libsql_client from '@libsql/client';
2
+ import { LibSQLDatabase } from 'drizzle-orm/libsql';
3
+
4
+ declare const createLibsqlClient: (args: {
5
+ url: string;
6
+ authToken?: string;
7
+ }) => _libsql_client.Client;
8
+ declare const createDbClient: <TSchema extends Record<string, unknown>>(schema: TSchema, args: {
9
+ url: string;
10
+ authToken?: string;
11
+ }) => LibSQLDatabase<TSchema>;
12
+ type DatabaseClient<TSchema extends Record<string, unknown>> = LibSQLDatabase<TSchema>;
13
+ type DatabaseClientTransactionContext<TSchema extends Record<string, unknown>> = Parameters<Parameters<DatabaseClient<TSchema>["transaction"]>[0]>[0];
14
+
15
+ export { type DatabaseClient, type DatabaseClientTransactionContext, createDbClient, createLibsqlClient };
package/dist/db.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/db.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\nimport { LibSQLDatabase, drizzle } from \"drizzle-orm/libsql\";\n\nexport const createLibsqlClient = (args: {\n url: string;\n authToken?: string;\n}) => {\n return createClient(args);\n};\n\nexport const createDbClient = <TSchema extends Record<string, unknown>>(\n schema: TSchema,\n args: { url: string; authToken?: string }\n) => {\n const client = createLibsqlClient(args);\n const db = drizzle(client, {\n schema,\n logger: false,\n });\n\n return db;\n};\n\nexport type DatabaseClient<TSchema extends Record<string, unknown>> =\n LibSQLDatabase<TSchema>;\nexport type DatabaseClientTransactionContext<\n TSchema extends Record<string, unknown>\n> = Parameters<Parameters<DatabaseClient<TSchema>[\"transaction\"]>[0]>[0];\n"],"mappings":";AAAA,SAAS,oBAAoB;AAC7B,SAAyB,eAAe;AAEjC,IAAM,qBAAqB,CAAC,SAG7B;AACJ,SAAO,aAAa,IAAI;AAC1B;AAEO,IAAM,iBAAiB,CAC5B,QACA,SACG;AACH,QAAM,SAAS,mBAAmB,IAAI;AACtC,QAAM,KAAK,QAAQ,QAAQ;AAAA,IACzB;AAAA,IACA,QAAQ;AAAA,EACV,CAAC;AAED,SAAO;AACT;","names":[]}
1
+ {"version":3,"sources":["../src/db.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\r\nimport { LibSQLDatabase, drizzle } from \"drizzle-orm/libsql\";\r\n\r\nexport const createLibsqlClient = (args: {\r\n url: string;\r\n authToken?: string;\r\n}) => {\r\n return createClient(args);\r\n};\r\n\r\nexport const createDbClient = <TSchema extends Record<string, unknown>>(\r\n schema: TSchema,\r\n args: { url: string; authToken?: string }\r\n) => {\r\n const client = createLibsqlClient(args);\r\n const db = drizzle(client, {\r\n schema,\r\n logger: false,\r\n });\r\n\r\n return db;\r\n};\r\n\r\nexport type DatabaseClient<TSchema extends Record<string, unknown>> =\r\n LibSQLDatabase<TSchema>;\r\nexport type DatabaseClientTransactionContext<\r\n TSchema extends Record<string, unknown>\r\n> = Parameters<Parameters<DatabaseClient<TSchema>[\"transaction\"]>[0]>[0];\r\n"],"mappings":";AAAA,SAAS,oBAAoB;AAC7B,SAAyB,eAAe;AAEjC,IAAM,qBAAqB,CAAC,SAG7B;AACJ,SAAO,aAAa,IAAI;AAC1B;AAEO,IAAM,iBAAiB,CAC5B,QACA,SACG;AACH,QAAM,SAAS,mBAAmB,IAAI;AACtC,QAAM,KAAK,QAAQ,QAAQ;AAAA,IACzB;AAAA,IACA,QAAQ;AAAA,EACV,CAAC;AAED,SAAO;AACT;","names":[]}
package/dist/env.cjs ADDED
@@ -0,0 +1,65 @@
1
+ "use strict";
2
+ var __create = Object.create;
3
+ var __defProp = Object.defineProperty;
4
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
+ var __getOwnPropNames = Object.getOwnPropertyNames;
6
+ var __getProtoOf = Object.getPrototypeOf;
7
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
8
+ var __export = (target, all) => {
9
+ for (var name in all)
10
+ __defProp(target, name, { get: all[name], enumerable: true });
11
+ };
12
+ var __copyProps = (to, from, except, desc) => {
13
+ if (from && typeof from === "object" || typeof from === "function") {
14
+ for (let key of __getOwnPropNames(from))
15
+ if (!__hasOwnProp.call(to, key) && key !== except)
16
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
17
+ }
18
+ return to;
19
+ };
20
+ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
+ // If the importer is in node compatibility mode or this is not an ESM
22
+ // file that has been converted to a CommonJS file using a Babel-
23
+ // compatible transform (i.e. "__esModule" has not been set), then set
24
+ // "default" to the CommonJS "module.exports" for node compatibility.
25
+ isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
+ mod
27
+ ));
28
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
+
30
+ // src/env.ts
31
+ var env_exports = {};
32
+ __export(env_exports, {
33
+ createEnv: () => createEnv
34
+ });
35
+ module.exports = __toCommonJS(env_exports);
36
+ var import_remeda = require("remeda");
37
+ var v = __toESM(require("valibot"), 1);
38
+ var PUBLIC_ENV_PREFIX = "PUBLIC_";
39
+ var createEnv = (args) => {
40
+ const pairs = Object.entries(args.schema);
41
+ const serverEnv = /* @__PURE__ */ new Map();
42
+ for (const [key, value] of pairs) {
43
+ const result = v.safeParse(value, args.env[key] ?? null);
44
+ if (!result.success) {
45
+ console.error(`Environment variable "${key}" is invalid`);
46
+ process.exit(1);
47
+ }
48
+ serverEnv.set(key, result.output);
49
+ }
50
+ const clientEnv = (0, import_remeda.pipe)(
51
+ serverEnv,
52
+ (obj) => Array.from(obj.entries()),
53
+ (pairs2) => pairs2.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),
54
+ (pairs2) => Object.fromEntries(pairs2)
55
+ );
56
+ return {
57
+ client: clientEnv,
58
+ server: Object.fromEntries(serverEnv.entries())
59
+ };
60
+ };
61
+ // Annotate the CommonJS export names for ESM import in node:
62
+ 0 && (module.exports = {
63
+ createEnv
64
+ });
65
+ //# sourceMappingURL=env.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/env.ts"],"sourcesContent":["import { pipe } from \"remeda\";\r\nimport * as v from \"valibot\";\r\n\r\nconst PUBLIC_ENV_PREFIX = \"PUBLIC_\" as const;\r\n\r\n/**\r\n * Validates your environment variables against the given Valibot schema;\r\n * @param args\r\n * @returns An object containing client environment variables and another containing server environment variables\r\n */\r\nexport const createEnv = <\r\n Schema extends Record<string, v.GenericSchema>,\r\n Env = {\r\n [K in keyof Schema]: v.InferOutput<Schema[K]>;\r\n }\r\n>(args: {\r\n schema: Schema;\r\n env: any;\r\n}) => {\r\n const pairs = Object.entries(args.schema);\r\n const serverEnv = new Map();\r\n\r\n for (const [key, value] of pairs) {\r\n const result = v.safeParse(value, args.env[key] ?? null);\r\n\r\n if (!result.success) {\r\n console.error(`Environment variable \"${key}\" is invalid`);\r\n process.exit(1);\r\n }\r\n\r\n serverEnv.set(key, result.output);\r\n }\r\n\r\n type ClientEnvKeys = Exclude<\r\n {\r\n [K in keyof Env]: K extends `${typeof PUBLIC_ENV_PREFIX}${string}`\r\n ? K\r\n : never;\r\n }[keyof Env],\r\n undefined\r\n >;\r\n\r\n type ClientEnv = {\r\n [B in ClientEnvKeys]: Env[B];\r\n };\r\n\r\n const clientEnv = pipe(\r\n serverEnv,\r\n (obj) => Array.from(obj.entries()),\r\n (pairs) => pairs.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),\r\n (pairs) => Object.fromEntries(pairs)\r\n ) as ClientEnv;\r\n\r\n return {\r\n client: clientEnv,\r\n server: Object.fromEntries(serverEnv.entries()) as Env,\r\n };\r\n};\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAqB;AACrB,QAAmB;AAEnB,IAAM,oBAAoB;AAOnB,IAAM,YAAY,CAKvB,SAGI;AACJ,QAAM,QAAQ,OAAO,QAAQ,KAAK,MAAM;AACxC,QAAM,YAAY,oBAAI,IAAI;AAE1B,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO;AAChC,UAAM,SAAW,YAAU,OAAO,KAAK,IAAI,GAAG,KAAK,IAAI;AAEvD,QAAI,CAAC,OAAO,SAAS;AACnB,cAAQ,MAAM,yBAAyB,GAAG,cAAc;AACxD,cAAQ,KAAK,CAAC;AAAA,IAChB;AAEA,cAAU,IAAI,KAAK,OAAO,MAAM;AAAA,EAClC;AAeA,QAAM,gBAAY;AAAA,IAChB;AAAA,IACA,CAAC,QAAQ,MAAM,KAAK,IAAI,QAAQ,CAAC;AAAA,IACjC,CAACA,WAAUA,OAAM,OAAO,CAAC,CAAC,CAAC,MAAM,EAAE,WAAW,iBAAiB,CAAC;AAAA,IAChE,CAACA,WAAU,OAAO,YAAYA,MAAK;AAAA,EACrC;AAEA,SAAO;AAAA,IACL,QAAQ;AAAA,IACR,QAAQ,OAAO,YAAY,UAAU,QAAQ,CAAC;AAAA,EAChD;AACF;","names":["pairs"]}
package/dist/env.d.cts ADDED
@@ -0,0 +1,16 @@
1
+ import * as v from 'valibot';
2
+
3
+ /**
4
+ * Validates your environment variables against the given Valibot schema;
5
+ * @param args
6
+ * @returns An object containing client environment variables and another containing server environment variables
7
+ */
8
+ declare const createEnv: <Schema extends Record<string, v.GenericSchema>, Env = { [K in keyof Schema]: v.InferOutput<Schema[K]>; }>(args: {
9
+ schema: Schema;
10
+ env: any;
11
+ }) => {
12
+ client: { [B in Exclude<{ [K_1 in keyof Env]: K_1 extends `PUBLIC_${string}` ? K_1 : never; }[keyof Env], undefined>]: Env[B]; };
13
+ server: Env;
14
+ };
15
+
16
+ export { createEnv };
package/dist/env.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/env.ts"],"sourcesContent":["import { pipe } from \"remeda\";\nimport * as v from \"valibot\";\n\nconst PUBLIC_ENV_PREFIX = \"PUBLIC_\" as const;\n\n/**\n * Validates your environment variables against the given Valibot schema;\n * @param args\n * @returns An object containing client environment variables and another containing server environment variables\n */\nexport const createEnv = <\n Schema extends Record<string, v.GenericSchema>,\n Env = {\n [K in keyof Schema]: v.InferOutput<Schema[K]>;\n }\n>(args: {\n schema: Schema;\n env: any;\n}) => {\n const pairs = Object.entries(args.schema);\n const serverEnv = new Map();\n\n for (const [key, value] of pairs) {\n const result = v.safeParse(value, args.env[key] ?? null);\n\n if (!result.success) {\n console.error(`Environment variable \"${key}\" is invalid`);\n process.exit(1);\n }\n\n serverEnv.set(key, result.output);\n }\n\n type ClientEnvKeys = Exclude<\n {\n [K in keyof Env]: K extends `${typeof PUBLIC_ENV_PREFIX}${string}`\n ? K\n : never;\n }[keyof Env],\n undefined\n >;\n\n type ClientEnv = {\n [B in ClientEnvKeys]: Env[B];\n };\n\n const clientEnv = pipe(\n serverEnv,\n (obj) => Array.from(obj.entries()),\n (pairs) => pairs.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),\n (pairs) => Object.fromEntries(pairs)\n ) as ClientEnv;\n\n return {\n client: clientEnv,\n server: Object.fromEntries(serverEnv.entries()) as Env,\n };\n};\n"],"mappings":";AAAA,SAAS,YAAY;AACrB,YAAY,OAAO;AAEnB,IAAM,oBAAoB;AAOnB,IAAM,YAAY,CAKvB,SAGI;AACJ,QAAM,QAAQ,OAAO,QAAQ,KAAK,MAAM;AACxC,QAAM,YAAY,oBAAI,IAAI;AAE1B,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO;AAChC,UAAM,SAAW,YAAU,OAAO,KAAK,IAAI,GAAG,KAAK,IAAI;AAEvD,QAAI,CAAC,OAAO,SAAS;AACnB,cAAQ,MAAM,yBAAyB,GAAG,cAAc;AACxD,cAAQ,KAAK,CAAC;AAAA,IAChB;AAEA,cAAU,IAAI,KAAK,OAAO,MAAM;AAAA,EAClC;AAeA,QAAM,YAAY;AAAA,IAChB;AAAA,IACA,CAAC,QAAQ,MAAM,KAAK,IAAI,QAAQ,CAAC;AAAA,IACjC,CAACA,WAAUA,OAAM,OAAO,CAAC,CAAC,CAAC,MAAM,EAAE,WAAW,iBAAiB,CAAC;AAAA,IAChE,CAACA,WAAU,OAAO,YAAYA,MAAK;AAAA,EACrC;AAEA,SAAO;AAAA,IACL,QAAQ;AAAA,IACR,QAAQ,OAAO,YAAY,UAAU,QAAQ,CAAC;AAAA,EAChD;AACF;","names":["pairs"]}
1
+ {"version":3,"sources":["../src/env.ts"],"sourcesContent":["import { pipe } from \"remeda\";\r\nimport * as v from \"valibot\";\r\n\r\nconst PUBLIC_ENV_PREFIX = \"PUBLIC_\" as const;\r\n\r\n/**\r\n * Validates your environment variables against the given Valibot schema;\r\n * @param args\r\n * @returns An object containing client environment variables and another containing server environment variables\r\n */\r\nexport const createEnv = <\r\n Schema extends Record<string, v.GenericSchema>,\r\n Env = {\r\n [K in keyof Schema]: v.InferOutput<Schema[K]>;\r\n }\r\n>(args: {\r\n schema: Schema;\r\n env: any;\r\n}) => {\r\n const pairs = Object.entries(args.schema);\r\n const serverEnv = new Map();\r\n\r\n for (const [key, value] of pairs) {\r\n const result = v.safeParse(value, args.env[key] ?? null);\r\n\r\n if (!result.success) {\r\n console.error(`Environment variable \"${key}\" is invalid`);\r\n process.exit(1);\r\n }\r\n\r\n serverEnv.set(key, result.output);\r\n }\r\n\r\n type ClientEnvKeys = Exclude<\r\n {\r\n [K in keyof Env]: K extends `${typeof PUBLIC_ENV_PREFIX}${string}`\r\n ? K\r\n : never;\r\n }[keyof Env],\r\n undefined\r\n >;\r\n\r\n type ClientEnv = {\r\n [B in ClientEnvKeys]: Env[B];\r\n };\r\n\r\n const clientEnv = pipe(\r\n serverEnv,\r\n (obj) => Array.from(obj.entries()),\r\n (pairs) => pairs.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),\r\n (pairs) => Object.fromEntries(pairs)\r\n ) as ClientEnv;\r\n\r\n return {\r\n client: clientEnv,\r\n server: Object.fromEntries(serverEnv.entries()) as Env,\r\n };\r\n};\r\n"],"mappings":";AAAA,SAAS,YAAY;AACrB,YAAY,OAAO;AAEnB,IAAM,oBAAoB;AAOnB,IAAM,YAAY,CAKvB,SAGI;AACJ,QAAM,QAAQ,OAAO,QAAQ,KAAK,MAAM;AACxC,QAAM,YAAY,oBAAI,IAAI;AAE1B,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO;AAChC,UAAM,SAAW,YAAU,OAAO,KAAK,IAAI,GAAG,KAAK,IAAI;AAEvD,QAAI,CAAC,OAAO,SAAS;AACnB,cAAQ,MAAM,yBAAyB,GAAG,cAAc;AACxD,cAAQ,KAAK,CAAC;AAAA,IAChB;AAEA,cAAU,IAAI,KAAK,OAAO,MAAM;AAAA,EAClC;AAeA,QAAM,YAAY;AAAA,IAChB;AAAA,IACA,CAAC,QAAQ,MAAM,KAAK,IAAI,QAAQ,CAAC;AAAA,IACjC,CAACA,WAAUA,OAAM,OAAO,CAAC,CAAC,CAAC,MAAM,EAAE,WAAW,iBAAiB,CAAC;AAAA,IAChE,CAACA,WAAU,OAAO,YAAYA,MAAK;AAAA,EACrC;AAEA,SAAO;AAAA,IACL,QAAQ;AAAA,IACR,QAAQ,OAAO,YAAY,UAAU,QAAQ,CAAC;AAAA,EAChD;AACF;","names":["pairs"]}
@@ -0,0 +1,175 @@
1
+ "use strict";
2
+ var __create = Object.create;
3
+ var __defProp = Object.defineProperty;
4
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
+ var __getOwnPropNames = Object.getOwnPropertyNames;
6
+ var __getProtoOf = Object.getPrototypeOf;
7
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
8
+ var __export = (target, all) => {
9
+ for (var name in all)
10
+ __defProp(target, name, { get: all[name], enumerable: true });
11
+ };
12
+ var __copyProps = (to, from, except, desc) => {
13
+ if (from && typeof from === "object" || typeof from === "function") {
14
+ for (let key of __getOwnPropNames(from))
15
+ if (!__hasOwnProp.call(to, key) && key !== except)
16
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
17
+ }
18
+ return to;
19
+ };
20
+ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
+ // If the importer is in node compatibility mode or this is not an ESM
22
+ // file that has been converted to a CommonJS file using a Babel-
23
+ // compatible transform (i.e. "__esModule" has not been set), then set
24
+ // "default" to the CommonJS "module.exports" for node compatibility.
25
+ isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
+ mod
27
+ ));
28
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
+
30
+ // src/images.ts
31
+ var images_exports = {};
32
+ __export(images_exports, {
33
+ ImageUtils: () => ImageUtils
34
+ });
35
+ module.exports = __toCommonJS(images_exports);
36
+ var import_cuid2 = require("@paralleldrive/cuid2");
37
+ var import_dayjs = __toESM(require("dayjs"), 1);
38
+ var import_ofetch = require("ofetch");
39
+ var ImageUtils = class {
40
+ blacklist = ["img.clerk.com"];
41
+ account;
42
+ _imageIds;
43
+ constructor(args) {
44
+ this.account = args.accountId;
45
+ this._imageIds = args.imageIds;
46
+ if (args.blacklist) {
47
+ this.blacklist.push(...args.blacklist);
48
+ }
49
+ }
50
+ get imageIds() {
51
+ if (!this._imageIds) {
52
+ throw new Error("imageIds was not supplied in constructor");
53
+ }
54
+ return this._imageIds;
55
+ }
56
+ url(id) {
57
+ return `https://imagedelivery.net/${this.account}/${id}/public`;
58
+ }
59
+ isBlacklisted(url) {
60
+ return this.blacklist.some((u) => url.includes(u));
61
+ }
62
+ isProtected(id) {
63
+ if (!this._imageIds) {
64
+ return false;
65
+ }
66
+ return Object.values(this._imageIds).some((e) => e === id);
67
+ }
68
+ /**
69
+ * Will only operate on images that have been uploaded via cloudflare images
70
+ */
71
+ optimizeUrl(url, options) {
72
+ if (this.isBlacklisted(url)) {
73
+ return url;
74
+ }
75
+ return url.replace("public", this.createImageOptionsString(options));
76
+ }
77
+ optimizeId(id, options) {
78
+ return this.optimizeUrl(this.url(id), options);
79
+ }
80
+ createOptionsSearchParams(options) {
81
+ const params = new URLSearchParams();
82
+ const pairs = Object.entries(options);
83
+ for (const [key, val] of pairs) {
84
+ if (val === void 0) {
85
+ continue;
86
+ }
87
+ params.set(key, val.toString());
88
+ }
89
+ return params;
90
+ }
91
+ createImageOptionsString(options) {
92
+ const params = this.createOptionsSearchParams(options);
93
+ return Array.from(params.entries()).map(([key, val]) => `${key}=${val}`).join(",");
94
+ }
95
+ async createUploadUrls(count, args) {
96
+ if (count === 0) {
97
+ return [];
98
+ }
99
+ const headers = new Headers();
100
+ headers.set("Authorization", `Bearer ${args.apiKey}`);
101
+ const urls = await Promise.all(
102
+ Array.from({ length: count }).map(async () => {
103
+ try {
104
+ const form = new FormData();
105
+ const id = (0, import_cuid2.createId)();
106
+ form.append("id", id);
107
+ form.append("expiry", (0, import_dayjs.default)().add(5, "minute").toISOString());
108
+ const img = await (0, import_ofetch.ofetch)(
109
+ `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,
110
+ { method: "POST", headers, body: form }
111
+ );
112
+ if (!img.success) {
113
+ throw new Error("Error uploading image");
114
+ }
115
+ return { url: img.result.uploadURL, id };
116
+ } catch (e) {
117
+ console.error("Error uploading image");
118
+ throw e;
119
+ }
120
+ })
121
+ );
122
+ return urls;
123
+ }
124
+ async upload(url, body) {
125
+ const fetchResponse = await (0, import_ofetch.ofetch)(url, {
126
+ method: "POST",
127
+ body
128
+ });
129
+ if (!fetchResponse.success) {
130
+ throw new Error("Failed to upload image");
131
+ }
132
+ const downloadUrl = fetchResponse.result.variants[0];
133
+ if (!downloadUrl) {
134
+ throw new Error("Could not find download URL");
135
+ }
136
+ return downloadUrl;
137
+ }
138
+ async delete(id, args) {
139
+ if (this.isProtected(id)) {
140
+ return { success: true };
141
+ }
142
+ try {
143
+ const headers = new Headers();
144
+ headers.set("Authorization", `Bearer ${args.apiKey}`);
145
+ await (0, import_ofetch.ofetch)(
146
+ `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,
147
+ {
148
+ method: "POST",
149
+ headers
150
+ }
151
+ );
152
+ return { success: true };
153
+ } catch (_e) {
154
+ return { success: false };
155
+ }
156
+ }
157
+ async batchUpload(files) {
158
+ return await Promise.all(
159
+ files.map(async (e) => {
160
+ const formData = new FormData();
161
+ formData.append("file", e.file);
162
+ const downloadUrl = await this.upload(e.url.value, formData);
163
+ return {
164
+ url: downloadUrl,
165
+ id: e.url.id
166
+ };
167
+ })
168
+ );
169
+ }
170
+ };
171
+ // Annotate the CommonJS export names for ESM import in node:
172
+ 0 && (module.exports = {
173
+ ImageUtils
174
+ });
175
+ //# sourceMappingURL=images.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/images.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport dayjs from \"dayjs\";\r\nimport { ofetch } from \"ofetch\";\r\n\r\nexport interface OptimizedImageOptions {\r\n anim?: boolean;\r\n background?: string;\r\n blur?: number;\r\n brightness?: number;\r\n compression?: \"fast\"; // faster compression = larger file size\r\n contrast?: number;\r\n dpr?: number;\r\n fit?: \"scale-down\" | \"contain\" | \"cover\" | \"crop\" | \"pad\";\r\n format?: \"webp\" | \"avif\" | \"json\";\r\n gamma?: number;\r\n width?: number;\r\n height?: number;\r\n metadata?: \"keep\" | \"copyright\" | \"none\";\r\n quality?: number;\r\n rotate?: number;\r\n sharpen?: number;\r\n}\r\n\r\nexport interface CreateImageUrlResponse {\r\n result: {\r\n id: string;\r\n uploadURL: string;\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\ninterface UploadImageResponse {\r\n result: {\r\n id: string;\r\n filename: string;\r\n uploaded: string;\r\n requireSignedURLs: boolean;\r\n variants: string[];\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\nexport class ImageUtils<ImageIds extends Record<string, any>> {\r\n private blacklist: string[] = [\"img.clerk.com\"];\r\n private account: string;\r\n private _imageIds: ImageIds | undefined;\r\n\r\n constructor(args: {\r\n accountId: string;\r\n blacklist?: string[];\r\n imageIds?: ImageIds;\r\n }) {\r\n this.account = args.accountId;\r\n\r\n this._imageIds = args.imageIds;\r\n\r\n if (args.blacklist) {\r\n this.blacklist.push(...args.blacklist);\r\n }\r\n }\r\n\r\n get imageIds() {\r\n if (!this._imageIds) {\r\n throw new Error(\"imageIds was not supplied in constructor\");\r\n }\r\n\r\n return this._imageIds;\r\n }\r\n\r\n public url(id: string) {\r\n return `https://imagedelivery.net/${this.account}/${id}/public`;\r\n }\r\n\r\n private isBlacklisted(url: string) {\r\n return this.blacklist.some((u) => url.includes(u));\r\n }\r\n\r\n private isProtected(id: string) {\r\n if (!this._imageIds) {\r\n return false;\r\n }\r\n\r\n return Object.values(this._imageIds).some((e) => e === id);\r\n }\r\n\r\n /**\r\n * Will only operate on images that have been uploaded via cloudflare images\r\n */\r\n public optimizeUrl(url: string, options: OptimizedImageOptions) {\r\n if (this.isBlacklisted(url)) {\r\n return url;\r\n }\r\n\r\n // Final format should look similar to: https://imagedelivery.net/<ACCOUNT_HASH>/<IMAGE_ID>/w=400,sharpen=3\r\n return url.replace(\"public\", this.createImageOptionsString(options));\r\n }\r\n\r\n public optimizeId(id: string, options: OptimizedImageOptions) {\r\n return this.optimizeUrl(this.url(id), options);\r\n }\r\n\r\n public createOptionsSearchParams(options: OptimizedImageOptions) {\r\n const params = new URLSearchParams();\r\n\r\n const pairs = Object.entries(options);\r\n\r\n for (const [key, val] of pairs) {\r\n if (val === undefined) {\r\n continue;\r\n }\r\n\r\n params.set(key, val.toString());\r\n }\r\n\r\n return params;\r\n }\r\n\r\n public createImageOptionsString(options: OptimizedImageOptions) {\r\n const params = this.createOptionsSearchParams(options);\r\n\r\n return Array.from(params.entries())\r\n .map(([key, val]) => `${key}=${val}`)\r\n .join(\",\");\r\n }\r\n\r\n public async createUploadUrls(count: number, args: { apiKey: string }) {\r\n if (count === 0) {\r\n return [];\r\n }\r\n\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n const urls = await Promise.all(\r\n Array.from({ length: count }).map(async () => {\r\n try {\r\n const form = new FormData();\r\n const id = createId();\r\n form.append(\"id\", id);\r\n form.append(\"expiry\", dayjs().add(5, \"minute\").toISOString());\r\n\r\n const img = await ofetch<CreateImageUrlResponse>(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,\r\n { method: \"POST\", headers, body: form }\r\n );\r\n\r\n if (!img.success) {\r\n throw new Error(\"Error uploading image\");\r\n }\r\n\r\n return { url: img.result.uploadURL, id };\r\n } catch (e) {\r\n console.error(\"Error uploading image\");\r\n throw e;\r\n }\r\n })\r\n );\r\n\r\n return urls;\r\n }\r\n\r\n public async upload(url: string, body: FormData) {\r\n const fetchResponse = await ofetch<UploadImageResponse>(url, {\r\n method: \"POST\",\r\n body,\r\n });\r\n\r\n if (!fetchResponse.success) {\r\n throw new Error(\"Failed to upload image\");\r\n }\r\n\r\n const downloadUrl = fetchResponse.result.variants[0];\r\n\r\n if (!downloadUrl) {\r\n throw new Error(\"Could not find download URL\");\r\n }\r\n\r\n return downloadUrl;\r\n }\r\n\r\n public async delete(id: string, args: { apiKey: string }) {\r\n if (this.isProtected(id)) {\r\n return { success: true };\r\n }\r\n\r\n try {\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n await ofetch(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,\r\n {\r\n method: \"POST\",\r\n headers,\r\n }\r\n );\r\n return { success: true };\r\n } catch (_e) {\r\n return { success: false };\r\n }\r\n }\r\n\r\n public async batchUpload(\r\n files: { file: File; url: { id: string; value: string } }[]\r\n ) {\r\n return await Promise.all(\r\n files.map(async (e) => {\r\n const formData = new FormData();\r\n formData.append(\"file\", e.file);\r\n\r\n const downloadUrl = await this.upload(e.url.value, formData);\r\n\r\n return {\r\n url: downloadUrl,\r\n id: e.url.id,\r\n };\r\n })\r\n );\r\n }\r\n}\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mBAAyB;AACzB,mBAAkB;AAClB,oBAAuB;AA4ChB,IAAM,aAAN,MAAuD;AAAA,EACpD,YAAsB,CAAC,eAAe;AAAA,EACtC;AAAA,EACA;AAAA,EAER,YAAY,MAIT;AACD,SAAK,UAAU,KAAK;AAEpB,SAAK,YAAY,KAAK;AAEtB,QAAI,KAAK,WAAW;AAClB,WAAK,UAAU,KAAK,GAAG,KAAK,SAAS;AAAA,IACvC;AAAA,EACF;AAAA,EAEA,IAAI,WAAW;AACb,QAAI,CAAC,KAAK,WAAW;AACnB,YAAM,IAAI,MAAM,0CAA0C;AAAA,IAC5D;AAEA,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,IAAY;AACrB,WAAO,6BAA6B,KAAK,OAAO,IAAI,EAAE;AAAA,EACxD;AAAA,EAEQ,cAAc,KAAa;AACjC,WAAO,KAAK,UAAU,KAAK,CAAC,MAAM,IAAI,SAAS,CAAC,CAAC;AAAA,EACnD;AAAA,EAEQ,YAAY,IAAY;AAC9B,QAAI,CAAC,KAAK,WAAW;AACnB,aAAO;AAAA,IACT;AAEA,WAAO,OAAO,OAAO,KAAK,SAAS,EAAE,KAAK,CAAC,MAAM,MAAM,EAAE;AAAA,EAC3D;AAAA;AAAA;AAAA;AAAA,EAKO,YAAY,KAAa,SAAgC;AAC9D,QAAI,KAAK,cAAc,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAGA,WAAO,IAAI,QAAQ,UAAU,KAAK,yBAAyB,OAAO,CAAC;AAAA,EACrE;AAAA,EAEO,WAAW,IAAY,SAAgC;AAC5D,WAAO,KAAK,YAAY,KAAK,IAAI,EAAE,GAAG,OAAO;AAAA,EAC/C;AAAA,EAEO,0BAA0B,SAAgC;AAC/D,UAAM,SAAS,IAAI,gBAAgB;AAEnC,UAAM,QAAQ,OAAO,QAAQ,OAAO;AAEpC,eAAW,CAAC,KAAK,GAAG,KAAK,OAAO;AAC9B,UAAI,QAAQ,QAAW;AACrB;AAAA,MACF;AAEA,aAAO,IAAI,KAAK,IAAI,SAAS,CAAC;AAAA,IAChC;AAEA,WAAO;AAAA,EACT;AAAA,EAEO,yBAAyB,SAAgC;AAC9D,UAAM,SAAS,KAAK,0BAA0B,OAAO;AAErD,WAAO,MAAM,KAAK,OAAO,QAAQ,CAAC,EAC/B,IAAI,CAAC,CAAC,KAAK,GAAG,MAAM,GAAG,GAAG,IAAI,GAAG,EAAE,EACnC,KAAK,GAAG;AAAA,EACb;AAAA,EAEA,MAAa,iBAAiB,OAAe,MAA0B;AACrE,QAAI,UAAU,GAAG;AACf,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAU,IAAI,QAAQ;AAC5B,YAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,UAAM,OAAO,MAAM,QAAQ;AAAA,MACzB,MAAM,KAAK,EAAE,QAAQ,MAAM,CAAC,EAAE,IAAI,YAAY;AAC5C,YAAI;AACF,gBAAM,OAAO,IAAI,SAAS;AAC1B,gBAAM,SAAK,uBAAS;AACpB,eAAK,OAAO,MAAM,EAAE;AACpB,eAAK,OAAO,cAAU,aAAAA,SAAM,EAAE,IAAI,GAAG,QAAQ,EAAE,YAAY,CAAC;AAE5D,gBAAM,MAAM,UAAM;AAAA,YAChB,iDAAiD,KAAK,OAAO;AAAA,YAC7D,EAAE,QAAQ,QAAQ,SAAS,MAAM,KAAK;AAAA,UACxC;AAEA,cAAI,CAAC,IAAI,SAAS;AAChB,kBAAM,IAAI,MAAM,uBAAuB;AAAA,UACzC;AAEA,iBAAO,EAAE,KAAK,IAAI,OAAO,WAAW,GAAG;AAAA,QACzC,SAAS,GAAG;AACV,kBAAQ,MAAM,uBAAuB;AACrC,gBAAM;AAAA,QACR;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,KAAa,MAAgB;AAC/C,UAAM,gBAAgB,UAAM,sBAA4B,KAAK;AAAA,MAC3D,QAAQ;AAAA,MACR;AAAA,IACF,CAAC;AAED,QAAI,CAAC,cAAc,SAAS;AAC1B,YAAM,IAAI,MAAM,wBAAwB;AAAA,IAC1C;AAEA,UAAM,cAAc,cAAc,OAAO,SAAS,CAAC;AAEnD,QAAI,CAAC,aAAa;AAChB,YAAM,IAAI,MAAM,6BAA6B;AAAA,IAC/C;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,IAAY,MAA0B;AACxD,QAAI,KAAK,YAAY,EAAE,GAAG;AACxB,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB;AAEA,QAAI;AACF,YAAM,UAAU,IAAI,QAAQ;AAC5B,cAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,gBAAM;AAAA,QACJ,iDAAiD,KAAK,OAAO,cAAc,EAAE;AAAA,QAC7E;AAAA,UACE,QAAQ;AAAA,UACR;AAAA,QACF;AAAA,MACF;AACA,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB,SAAS,IAAI;AACX,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,MAAa,YACX,OACA;AACA,WAAO,MAAM,QAAQ;AAAA,MACnB,MAAM,IAAI,OAAO,MAAM;AACrB,cAAM,WAAW,IAAI,SAAS;AAC9B,iBAAS,OAAO,QAAQ,EAAE,IAAI;AAE9B,cAAM,cAAc,MAAM,KAAK,OAAO,EAAE,IAAI,OAAO,QAAQ;AAE3D,eAAO;AAAA,UACL,KAAK;AAAA,UACL,IAAI,EAAE,IAAI;AAAA,QACZ;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AACF;","names":["dayjs"]}
@@ -0,0 +1,72 @@
1
+ interface OptimizedImageOptions {
2
+ anim?: boolean;
3
+ background?: string;
4
+ blur?: number;
5
+ brightness?: number;
6
+ compression?: "fast";
7
+ contrast?: number;
8
+ dpr?: number;
9
+ fit?: "scale-down" | "contain" | "cover" | "crop" | "pad";
10
+ format?: "webp" | "avif" | "json";
11
+ gamma?: number;
12
+ width?: number;
13
+ height?: number;
14
+ metadata?: "keep" | "copyright" | "none";
15
+ quality?: number;
16
+ rotate?: number;
17
+ sharpen?: number;
18
+ }
19
+ interface CreateImageUrlResponse {
20
+ result: {
21
+ id: string;
22
+ uploadURL: string;
23
+ };
24
+ success: boolean;
25
+ errors: unknown[];
26
+ messages: unknown[];
27
+ }
28
+ declare class ImageUtils<ImageIds extends Record<string, any>> {
29
+ private blacklist;
30
+ private account;
31
+ private _imageIds;
32
+ constructor(args: {
33
+ accountId: string;
34
+ blacklist?: string[];
35
+ imageIds?: ImageIds;
36
+ });
37
+ get imageIds(): ImageIds;
38
+ url(id: string): string;
39
+ private isBlacklisted;
40
+ private isProtected;
41
+ /**
42
+ * Will only operate on images that have been uploaded via cloudflare images
43
+ */
44
+ optimizeUrl(url: string, options: OptimizedImageOptions): string;
45
+ optimizeId(id: string, options: OptimizedImageOptions): string;
46
+ createOptionsSearchParams(options: OptimizedImageOptions): URLSearchParams;
47
+ createImageOptionsString(options: OptimizedImageOptions): string;
48
+ createUploadUrls(count: number, args: {
49
+ apiKey: string;
50
+ }): Promise<{
51
+ url: string;
52
+ id: string;
53
+ }[]>;
54
+ upload(url: string, body: FormData): Promise<string>;
55
+ delete(id: string, args: {
56
+ apiKey: string;
57
+ }): Promise<{
58
+ success: boolean;
59
+ }>;
60
+ batchUpload(files: {
61
+ file: File;
62
+ url: {
63
+ id: string;
64
+ value: string;
65
+ };
66
+ }[]): Promise<{
67
+ url: string;
68
+ id: string;
69
+ }[]>;
70
+ }
71
+
72
+ export { type CreateImageUrlResponse, ImageUtils, type OptimizedImageOptions };
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/images.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\nimport dayjs from \"dayjs\";\nimport { ofetch } from \"ofetch\";\n\nexport interface OptimizedImageOptions {\n anim?: boolean;\n background?: string;\n blur?: number;\n brightness?: number;\n compression?: \"fast\"; // faster compression = larger file size\n contrast?: number;\n dpr?: number;\n fit?: \"scale-down\" | \"contain\" | \"cover\" | \"crop\" | \"pad\";\n format?: \"webp\" | \"avif\" | \"json\";\n gamma?: number;\n width?: number;\n height?: number;\n metadata?: \"keep\" | \"copyright\" | \"none\";\n quality?: number;\n rotate?: number;\n sharpen?: number;\n}\n\nexport interface CreateImageUrlResponse {\n result: {\n id: string;\n uploadURL: string;\n };\n success: boolean;\n errors: unknown[];\n messages: unknown[];\n}\n\ninterface UploadImageResponse {\n result: {\n id: string;\n filename: string;\n uploaded: string;\n requireSignedURLs: boolean;\n variants: string[];\n };\n success: boolean;\n errors: unknown[];\n messages: unknown[];\n}\n\nexport class ImageUtils<ImageIds extends Record<string, any>> {\n private blacklist: string[] = [\"img.clerk.com\"];\n private account: string;\n private _imageIds: ImageIds | undefined;\n\n constructor(args: {\n accountId: string;\n blacklist?: string[];\n imageIds?: ImageIds;\n }) {\n this.account = args.accountId;\n\n this._imageIds = args.imageIds;\n\n if (args.blacklist) {\n this.blacklist.push(...args.blacklist);\n }\n }\n\n get imageIds() {\n if (!this._imageIds) {\n throw new Error(\"imageIds was not supplied in constructor\");\n }\n\n return this._imageIds;\n }\n\n public url(id: string) {\n return `https://imagedelivery.net/${this.account}/${id}/public`;\n }\n\n private isBlacklisted(url: string) {\n return this.blacklist.some((u) => url.includes(u));\n }\n\n private isProtected(id: string) {\n if (!this._imageIds) {\n return false;\n }\n\n return Object.values(this._imageIds).some((e) => e === id);\n }\n\n /**\n * Will only operate on images that have been uploaded via cloudflare images\n */\n public optimizeUrl(url: string, options: OptimizedImageOptions) {\n if (this.isBlacklisted(url)) {\n return url;\n }\n\n // Final format should look similar to: https://imagedelivery.net/<ACCOUNT_HASH>/<IMAGE_ID>/w=400,sharpen=3\n return url.replace(\"public\", this.createImageOptionsString(options));\n }\n\n public optimizeId(id: string, options: OptimizedImageOptions) {\n return this.optimizeUrl(this.url(id), options);\n }\n\n public createOptionsSearchParams(options: OptimizedImageOptions) {\n const params = new URLSearchParams();\n\n const pairs = Object.entries(options);\n\n for (const [key, val] of pairs) {\n if (val === undefined) {\n continue;\n }\n\n params.set(key, val.toString());\n }\n\n return params;\n }\n\n public createImageOptionsString(options: OptimizedImageOptions) {\n const params = this.createOptionsSearchParams(options);\n\n return Array.from(params.entries())\n .map(([key, val]) => `${key}=${val}`)\n .join(\",\");\n }\n\n public async createUploadUrls(count: number, args: { apiKey: string }) {\n if (count === 0) {\n return [];\n }\n\n const headers = new Headers();\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\n\n const urls = await Promise.all(\n Array.from({ length: count }).map(async () => {\n try {\n const form = new FormData();\n const id = createId();\n form.append(\"id\", id);\n form.append(\"expiry\", dayjs().add(5, \"minute\").toISOString());\n\n const img = await ofetch<CreateImageUrlResponse>(\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,\n { method: \"POST\", headers, body: form }\n );\n\n if (!img.success) {\n throw new Error(\"Error uploading image\");\n }\n\n return { url: img.result.uploadURL, id };\n } catch (e) {\n console.error(\"Error uploading image\");\n throw e;\n }\n })\n );\n\n return urls;\n }\n\n public async upload(url: string, body: FormData) {\n const fetchResponse = await ofetch<UploadImageResponse>(url, {\n method: \"POST\",\n body,\n });\n\n if (!fetchResponse.success) {\n throw new Error(\"Failed to upload image\");\n }\n\n const downloadUrl = fetchResponse.result.variants[0];\n\n if (!downloadUrl) {\n throw new Error(\"Could not find download URL\");\n }\n\n return downloadUrl;\n }\n\n public async delete(id: string, args: { apiKey: string }) {\n if (this.isProtected(id)) {\n return { success: true };\n }\n\n try {\n const headers = new Headers();\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\n\n await ofetch(\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,\n {\n method: \"POST\",\n headers,\n }\n );\n return { success: true };\n } catch (_e) {\n return { success: false };\n }\n }\n\n public async batchUpload(\n files: { file: File; url: { id: string; value: string } }[]\n ) {\n return await Promise.all(\n files.map(async (e) => {\n const formData = new FormData();\n formData.append(\"file\", e.file);\n\n const downloadUrl = await this.upload(e.url.value, formData);\n\n return {\n url: downloadUrl,\n id: e.url.id,\n };\n })\n );\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,OAAO,WAAW;AAClB,SAAS,cAAc;AA4ChB,IAAM,aAAN,MAAuD;AAAA,EACpD,YAAsB,CAAC,eAAe;AAAA,EACtC;AAAA,EACA;AAAA,EAER,YAAY,MAIT;AACD,SAAK,UAAU,KAAK;AAEpB,SAAK,YAAY,KAAK;AAEtB,QAAI,KAAK,WAAW;AAClB,WAAK,UAAU,KAAK,GAAG,KAAK,SAAS;AAAA,IACvC;AAAA,EACF;AAAA,EAEA,IAAI,WAAW;AACb,QAAI,CAAC,KAAK,WAAW;AACnB,YAAM,IAAI,MAAM,0CAA0C;AAAA,IAC5D;AAEA,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,IAAY;AACrB,WAAO,6BAA6B,KAAK,OAAO,IAAI,EAAE;AAAA,EACxD;AAAA,EAEQ,cAAc,KAAa;AACjC,WAAO,KAAK,UAAU,KAAK,CAAC,MAAM,IAAI,SAAS,CAAC,CAAC;AAAA,EACnD;AAAA,EAEQ,YAAY,IAAY;AAC9B,QAAI,CAAC,KAAK,WAAW;AACnB,aAAO;AAAA,IACT;AAEA,WAAO,OAAO,OAAO,KAAK,SAAS,EAAE,KAAK,CAAC,MAAM,MAAM,EAAE;AAAA,EAC3D;AAAA;AAAA;AAAA;AAAA,EAKO,YAAY,KAAa,SAAgC;AAC9D,QAAI,KAAK,cAAc,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAGA,WAAO,IAAI,QAAQ,UAAU,KAAK,yBAAyB,OAAO,CAAC;AAAA,EACrE;AAAA,EAEO,WAAW,IAAY,SAAgC;AAC5D,WAAO,KAAK,YAAY,KAAK,IAAI,EAAE,GAAG,OAAO;AAAA,EAC/C;AAAA,EAEO,0BAA0B,SAAgC;AAC/D,UAAM,SAAS,IAAI,gBAAgB;AAEnC,UAAM,QAAQ,OAAO,QAAQ,OAAO;AAEpC,eAAW,CAAC,KAAK,GAAG,KAAK,OAAO;AAC9B,UAAI,QAAQ,QAAW;AACrB;AAAA,MACF;AAEA,aAAO,IAAI,KAAK,IAAI,SAAS,CAAC;AAAA,IAChC;AAEA,WAAO;AAAA,EACT;AAAA,EAEO,yBAAyB,SAAgC;AAC9D,UAAM,SAAS,KAAK,0BAA0B,OAAO;AAErD,WAAO,MAAM,KAAK,OAAO,QAAQ,CAAC,EAC/B,IAAI,CAAC,CAAC,KAAK,GAAG,MAAM,GAAG,GAAG,IAAI,GAAG,EAAE,EACnC,KAAK,GAAG;AAAA,EACb;AAAA,EAEA,MAAa,iBAAiB,OAAe,MAA0B;AACrE,QAAI,UAAU,GAAG;AACf,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAU,IAAI,QAAQ;AAC5B,YAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,UAAM,OAAO,MAAM,QAAQ;AAAA,MACzB,MAAM,KAAK,EAAE,QAAQ,MAAM,CAAC,EAAE,IAAI,YAAY;AAC5C,YAAI;AACF,gBAAM,OAAO,IAAI,SAAS;AAC1B,gBAAM,KAAK,SAAS;AACpB,eAAK,OAAO,MAAM,EAAE;AACpB,eAAK,OAAO,UAAU,MAAM,EAAE,IAAI,GAAG,QAAQ,EAAE,YAAY,CAAC;AAE5D,gBAAM,MAAM,MAAM;AAAA,YAChB,iDAAiD,KAAK,OAAO;AAAA,YAC7D,EAAE,QAAQ,QAAQ,SAAS,MAAM,KAAK;AAAA,UACxC;AAEA,cAAI,CAAC,IAAI,SAAS;AAChB,kBAAM,IAAI,MAAM,uBAAuB;AAAA,UACzC;AAEA,iBAAO,EAAE,KAAK,IAAI,OAAO,WAAW,GAAG;AAAA,QACzC,SAAS,GAAG;AACV,kBAAQ,MAAM,uBAAuB;AACrC,gBAAM;AAAA,QACR;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,KAAa,MAAgB;AAC/C,UAAM,gBAAgB,MAAM,OAA4B,KAAK;AAAA,MAC3D,QAAQ;AAAA,MACR;AAAA,IACF,CAAC;AAED,QAAI,CAAC,cAAc,SAAS;AAC1B,YAAM,IAAI,MAAM,wBAAwB;AAAA,IAC1C;AAEA,UAAM,cAAc,cAAc,OAAO,SAAS,CAAC;AAEnD,QAAI,CAAC,aAAa;AAChB,YAAM,IAAI,MAAM,6BAA6B;AAAA,IAC/C;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,IAAY,MAA0B;AACxD,QAAI,KAAK,YAAY,EAAE,GAAG;AACxB,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB;AAEA,QAAI;AACF,YAAM,UAAU,IAAI,QAAQ;AAC5B,cAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,YAAM;AAAA,QACJ,iDAAiD,KAAK,OAAO,cAAc,EAAE;AAAA,QAC7E;AAAA,UACE,QAAQ;AAAA,UACR;AAAA,QACF;AAAA,MACF;AACA,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB,SAAS,IAAI;AACX,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,MAAa,YACX,OACA;AACA,WAAO,MAAM,QAAQ;AAAA,MACnB,MAAM,IAAI,OAAO,MAAM;AACrB,cAAM,WAAW,IAAI,SAAS;AAC9B,iBAAS,OAAO,QAAQ,EAAE,IAAI;AAE9B,cAAM,cAAc,MAAM,KAAK,OAAO,EAAE,IAAI,OAAO,QAAQ;AAE3D,eAAO;AAAA,UACL,KAAK;AAAA,UACL,IAAI,EAAE,IAAI;AAAA,QACZ;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AACF;","names":[]}
1
+ {"version":3,"sources":["../src/images.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport dayjs from \"dayjs\";\r\nimport { ofetch } from \"ofetch\";\r\n\r\nexport interface OptimizedImageOptions {\r\n anim?: boolean;\r\n background?: string;\r\n blur?: number;\r\n brightness?: number;\r\n compression?: \"fast\"; // faster compression = larger file size\r\n contrast?: number;\r\n dpr?: number;\r\n fit?: \"scale-down\" | \"contain\" | \"cover\" | \"crop\" | \"pad\";\r\n format?: \"webp\" | \"avif\" | \"json\";\r\n gamma?: number;\r\n width?: number;\r\n height?: number;\r\n metadata?: \"keep\" | \"copyright\" | \"none\";\r\n quality?: number;\r\n rotate?: number;\r\n sharpen?: number;\r\n}\r\n\r\nexport interface CreateImageUrlResponse {\r\n result: {\r\n id: string;\r\n uploadURL: string;\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\ninterface UploadImageResponse {\r\n result: {\r\n id: string;\r\n filename: string;\r\n uploaded: string;\r\n requireSignedURLs: boolean;\r\n variants: string[];\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\nexport class ImageUtils<ImageIds extends Record<string, any>> {\r\n private blacklist: string[] = [\"img.clerk.com\"];\r\n private account: string;\r\n private _imageIds: ImageIds | undefined;\r\n\r\n constructor(args: {\r\n accountId: string;\r\n blacklist?: string[];\r\n imageIds?: ImageIds;\r\n }) {\r\n this.account = args.accountId;\r\n\r\n this._imageIds = args.imageIds;\r\n\r\n if (args.blacklist) {\r\n this.blacklist.push(...args.blacklist);\r\n }\r\n }\r\n\r\n get imageIds() {\r\n if (!this._imageIds) {\r\n throw new Error(\"imageIds was not supplied in constructor\");\r\n }\r\n\r\n return this._imageIds;\r\n }\r\n\r\n public url(id: string) {\r\n return `https://imagedelivery.net/${this.account}/${id}/public`;\r\n }\r\n\r\n private isBlacklisted(url: string) {\r\n return this.blacklist.some((u) => url.includes(u));\r\n }\r\n\r\n private isProtected(id: string) {\r\n if (!this._imageIds) {\r\n return false;\r\n }\r\n\r\n return Object.values(this._imageIds).some((e) => e === id);\r\n }\r\n\r\n /**\r\n * Will only operate on images that have been uploaded via cloudflare images\r\n */\r\n public optimizeUrl(url: string, options: OptimizedImageOptions) {\r\n if (this.isBlacklisted(url)) {\r\n return url;\r\n }\r\n\r\n // Final format should look similar to: https://imagedelivery.net/<ACCOUNT_HASH>/<IMAGE_ID>/w=400,sharpen=3\r\n return url.replace(\"public\", this.createImageOptionsString(options));\r\n }\r\n\r\n public optimizeId(id: string, options: OptimizedImageOptions) {\r\n return this.optimizeUrl(this.url(id), options);\r\n }\r\n\r\n public createOptionsSearchParams(options: OptimizedImageOptions) {\r\n const params = new URLSearchParams();\r\n\r\n const pairs = Object.entries(options);\r\n\r\n for (const [key, val] of pairs) {\r\n if (val === undefined) {\r\n continue;\r\n }\r\n\r\n params.set(key, val.toString());\r\n }\r\n\r\n return params;\r\n }\r\n\r\n public createImageOptionsString(options: OptimizedImageOptions) {\r\n const params = this.createOptionsSearchParams(options);\r\n\r\n return Array.from(params.entries())\r\n .map(([key, val]) => `${key}=${val}`)\r\n .join(\",\");\r\n }\r\n\r\n public async createUploadUrls(count: number, args: { apiKey: string }) {\r\n if (count === 0) {\r\n return [];\r\n }\r\n\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n const urls = await Promise.all(\r\n Array.from({ length: count }).map(async () => {\r\n try {\r\n const form = new FormData();\r\n const id = createId();\r\n form.append(\"id\", id);\r\n form.append(\"expiry\", dayjs().add(5, \"minute\").toISOString());\r\n\r\n const img = await ofetch<CreateImageUrlResponse>(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,\r\n { method: \"POST\", headers, body: form }\r\n );\r\n\r\n if (!img.success) {\r\n throw new Error(\"Error uploading image\");\r\n }\r\n\r\n return { url: img.result.uploadURL, id };\r\n } catch (e) {\r\n console.error(\"Error uploading image\");\r\n throw e;\r\n }\r\n })\r\n );\r\n\r\n return urls;\r\n }\r\n\r\n public async upload(url: string, body: FormData) {\r\n const fetchResponse = await ofetch<UploadImageResponse>(url, {\r\n method: \"POST\",\r\n body,\r\n });\r\n\r\n if (!fetchResponse.success) {\r\n throw new Error(\"Failed to upload image\");\r\n }\r\n\r\n const downloadUrl = fetchResponse.result.variants[0];\r\n\r\n if (!downloadUrl) {\r\n throw new Error(\"Could not find download URL\");\r\n }\r\n\r\n return downloadUrl;\r\n }\r\n\r\n public async delete(id: string, args: { apiKey: string }) {\r\n if (this.isProtected(id)) {\r\n return { success: true };\r\n }\r\n\r\n try {\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n await ofetch(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,\r\n {\r\n method: \"POST\",\r\n headers,\r\n }\r\n );\r\n return { success: true };\r\n } catch (_e) {\r\n return { success: false };\r\n }\r\n }\r\n\r\n public async batchUpload(\r\n files: { file: File; url: { id: string; value: string } }[]\r\n ) {\r\n return await Promise.all(\r\n files.map(async (e) => {\r\n const formData = new FormData();\r\n formData.append(\"file\", e.file);\r\n\r\n const downloadUrl = await this.upload(e.url.value, formData);\r\n\r\n return {\r\n url: downloadUrl,\r\n id: e.url.id,\r\n };\r\n })\r\n );\r\n }\r\n}\r\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,OAAO,WAAW;AAClB,SAAS,cAAc;AA4ChB,IAAM,aAAN,MAAuD;AAAA,EACpD,YAAsB,CAAC,eAAe;AAAA,EACtC;AAAA,EACA;AAAA,EAER,YAAY,MAIT;AACD,SAAK,UAAU,KAAK;AAEpB,SAAK,YAAY,KAAK;AAEtB,QAAI,KAAK,WAAW;AAClB,WAAK,UAAU,KAAK,GAAG,KAAK,SAAS;AAAA,IACvC;AAAA,EACF;AAAA,EAEA,IAAI,WAAW;AACb,QAAI,CAAC,KAAK,WAAW;AACnB,YAAM,IAAI,MAAM,0CAA0C;AAAA,IAC5D;AAEA,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,IAAY;AACrB,WAAO,6BAA6B,KAAK,OAAO,IAAI,EAAE;AAAA,EACxD;AAAA,EAEQ,cAAc,KAAa;AACjC,WAAO,KAAK,UAAU,KAAK,CAAC,MAAM,IAAI,SAAS,CAAC,CAAC;AAAA,EACnD;AAAA,EAEQ,YAAY,IAAY;AAC9B,QAAI,CAAC,KAAK,WAAW;AACnB,aAAO;AAAA,IACT;AAEA,WAAO,OAAO,OAAO,KAAK,SAAS,EAAE,KAAK,CAAC,MAAM,MAAM,EAAE;AAAA,EAC3D;AAAA;AAAA;AAAA;AAAA,EAKO,YAAY,KAAa,SAAgC;AAC9D,QAAI,KAAK,cAAc,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAGA,WAAO,IAAI,QAAQ,UAAU,KAAK,yBAAyB,OAAO,CAAC;AAAA,EACrE;AAAA,EAEO,WAAW,IAAY,SAAgC;AAC5D,WAAO,KAAK,YAAY,KAAK,IAAI,EAAE,GAAG,OAAO;AAAA,EAC/C;AAAA,EAEO,0BAA0B,SAAgC;AAC/D,UAAM,SAAS,IAAI,gBAAgB;AAEnC,UAAM,QAAQ,OAAO,QAAQ,OAAO;AAEpC,eAAW,CAAC,KAAK,GAAG,KAAK,OAAO;AAC9B,UAAI,QAAQ,QAAW;AACrB;AAAA,MACF;AAEA,aAAO,IAAI,KAAK,IAAI,SAAS,CAAC;AAAA,IAChC;AAEA,WAAO;AAAA,EACT;AAAA,EAEO,yBAAyB,SAAgC;AAC9D,UAAM,SAAS,KAAK,0BAA0B,OAAO;AAErD,WAAO,MAAM,KAAK,OAAO,QAAQ,CAAC,EAC/B,IAAI,CAAC,CAAC,KAAK,GAAG,MAAM,GAAG,GAAG,IAAI,GAAG,EAAE,EACnC,KAAK,GAAG;AAAA,EACb;AAAA,EAEA,MAAa,iBAAiB,OAAe,MAA0B;AACrE,QAAI,UAAU,GAAG;AACf,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAU,IAAI,QAAQ;AAC5B,YAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,UAAM,OAAO,MAAM,QAAQ;AAAA,MACzB,MAAM,KAAK,EAAE,QAAQ,MAAM,CAAC,EAAE,IAAI,YAAY;AAC5C,YAAI;AACF,gBAAM,OAAO,IAAI,SAAS;AAC1B,gBAAM,KAAK,SAAS;AACpB,eAAK,OAAO,MAAM,EAAE;AACpB,eAAK,OAAO,UAAU,MAAM,EAAE,IAAI,GAAG,QAAQ,EAAE,YAAY,CAAC;AAE5D,gBAAM,MAAM,MAAM;AAAA,YAChB,iDAAiD,KAAK,OAAO;AAAA,YAC7D,EAAE,QAAQ,QAAQ,SAAS,MAAM,KAAK;AAAA,UACxC;AAEA,cAAI,CAAC,IAAI,SAAS;AAChB,kBAAM,IAAI,MAAM,uBAAuB;AAAA,UACzC;AAEA,iBAAO,EAAE,KAAK,IAAI,OAAO,WAAW,GAAG;AAAA,QACzC,SAAS,GAAG;AACV,kBAAQ,MAAM,uBAAuB;AACrC,gBAAM;AAAA,QACR;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,KAAa,MAAgB;AAC/C,UAAM,gBAAgB,MAAM,OAA4B,KAAK;AAAA,MAC3D,QAAQ;AAAA,MACR;AAAA,IACF,CAAC;AAED,QAAI,CAAC,cAAc,SAAS;AAC1B,YAAM,IAAI,MAAM,wBAAwB;AAAA,IAC1C;AAEA,UAAM,cAAc,cAAc,OAAO,SAAS,CAAC;AAEnD,QAAI,CAAC,aAAa;AAChB,YAAM,IAAI,MAAM,6BAA6B;AAAA,IAC/C;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,IAAY,MAA0B;AACxD,QAAI,KAAK,YAAY,EAAE,GAAG;AACxB,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB;AAEA,QAAI;AACF,YAAM,UAAU,IAAI,QAAQ;AAC5B,cAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,YAAM;AAAA,QACJ,iDAAiD,KAAK,OAAO,cAAc,EAAE;AAAA,QAC7E;AAAA,UACE,QAAQ;AAAA,UACR;AAAA,QACF;AAAA,MACF;AACA,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB,SAAS,IAAI;AACX,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,MAAa,YACX,OACA;AACA,WAAO,MAAM,QAAQ;AAAA,MACnB,MAAM,IAAI,OAAO,MAAM;AACrB,cAAM,WAAW,IAAI,SAAS;AAC9B,iBAAS,OAAO,QAAQ,EAAE,IAAI;AAE9B,cAAM,cAAc,MAAM,KAAK,OAAO,EAAE,IAAI,OAAO,QAAQ;AAE3D,eAAO;AAAA,UACL,KAAK;AAAA,UACL,IAAI,EAAE,IAAI;AAAA,QACZ;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AACF;","names":[]}
@@ -0,0 +1,48 @@
1
+ "use strict";
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __export = (target, all) => {
7
+ for (var name in all)
8
+ __defProp(target, name, { get: all[name], enumerable: true });
9
+ };
10
+ var __copyProps = (to, from, except, desc) => {
11
+ if (from && typeof from === "object" || typeof from === "function") {
12
+ for (let key of __getOwnPropNames(from))
13
+ if (!__hasOwnProp.call(to, key) && key !== except)
14
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
+ }
16
+ return to;
17
+ };
18
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
+
20
+ // src/logger.ts
21
+ var logger_exports = {};
22
+ __export(logger_exports, {
23
+ createLogger: () => createLogger
24
+ });
25
+ module.exports = __toCommonJS(logger_exports);
26
+ var import_pino = require("pino");
27
+ var createLogger = (args) => {
28
+ const l = (0, import_pino.pino)(
29
+ {
30
+ level: "info",
31
+ redact: [],
32
+ transport: args.pretty ? {
33
+ target: "pino-pretty"
34
+ } : void 0
35
+ },
36
+ args.token ? import_pino.pino.transport({
37
+ target: "@logtail/pino",
38
+ options: { sourceToken: args.token }
39
+ }) : void 0
40
+ );
41
+ l.child({ service: args.service });
42
+ return l;
43
+ };
44
+ // Annotate the CommonJS export names for ESM import in node:
45
+ 0 && (module.exports = {
46
+ createLogger
47
+ });
48
+ //# sourceMappingURL=logger.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/logger.ts"],"sourcesContent":["import { pino } from \"pino\";\r\n\r\nexport const createLogger = (args: {\r\n token?: string | undefined | null;\r\n pretty?: boolean;\r\n service: string;\r\n}) => {\r\n const l = pino(\r\n {\r\n level: \"info\",\r\n redact: [],\r\n transport: args.pretty\r\n ? {\r\n target: \"pino-pretty\",\r\n }\r\n : undefined,\r\n },\r\n\r\n args.token\r\n ? pino.transport({\r\n target: \"@logtail/pino\",\r\n options: { sourceToken: args.token },\r\n })\r\n : undefined,\r\n );\r\n\r\n l.child({ service: args.service });\r\n\r\n return l;\r\n};\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,kBAAqB;AAEd,IAAM,eAAe,CAAC,SAIvB;AACJ,QAAM,QAAI;AAAA,IACR;AAAA,MACE,OAAO;AAAA,MACP,QAAQ,CAAC;AAAA,MACT,WAAW,KAAK,SACZ;AAAA,QACE,QAAQ;AAAA,MACV,IACA;AAAA,IACN;AAAA,IAEA,KAAK,QACD,iBAAK,UAAU;AAAA,MACb,QAAQ;AAAA,MACR,SAAS,EAAE,aAAa,KAAK,MAAM;AAAA,IACrC,CAAC,IACD;AAAA,EACN;AAEA,IAAE,MAAM,EAAE,SAAS,KAAK,QAAQ,CAAC;AAEjC,SAAO;AACT;","names":[]}
@@ -0,0 +1,9 @@
1
+ import * as pino from 'pino';
2
+
3
+ declare const createLogger: (args: {
4
+ token?: string | undefined | null;
5
+ pretty?: boolean;
6
+ service: string;
7
+ }) => pino.Logger<never>;
8
+
9
+ export { createLogger };
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/logger.ts"],"sourcesContent":["import { pino } from \"pino\";\n\nexport const createLogger = (args: {\n token?: string | undefined | null;\n pretty?: boolean;\n service: string;\n}) => {\n const l = pino(\n {\n level: \"info\",\n redact: [],\n transport: args.pretty\n ? {\n target: \"pino-pretty\",\n }\n : undefined,\n },\n\n args.token\n ? pino.transport({\n target: \"@logtail/pino\",\n options: { sourceToken: args.token },\n })\n : undefined,\n );\n\n l.child({ service: args.service });\n\n return l;\n};\n"],"mappings":";AAAA,SAAS,YAAY;AAEd,IAAM,eAAe,CAAC,SAIvB;AACJ,QAAM,IAAI;AAAA,IACR;AAAA,MACE,OAAO;AAAA,MACP,QAAQ,CAAC;AAAA,MACT,WAAW,KAAK,SACZ;AAAA,QACE,QAAQ;AAAA,MACV,IACA;AAAA,IACN;AAAA,IAEA,KAAK,QACD,KAAK,UAAU;AAAA,MACb,QAAQ;AAAA,MACR,SAAS,EAAE,aAAa,KAAK,MAAM;AAAA,IACrC,CAAC,IACD;AAAA,EACN;AAEA,IAAE,MAAM,EAAE,SAAS,KAAK,QAAQ,CAAC;AAEjC,SAAO;AACT;","names":[]}
1
+ {"version":3,"sources":["../src/logger.ts"],"sourcesContent":["import { pino } from \"pino\";\r\n\r\nexport const createLogger = (args: {\r\n token?: string | undefined | null;\r\n pretty?: boolean;\r\n service: string;\r\n}) => {\r\n const l = pino(\r\n {\r\n level: \"info\",\r\n redact: [],\r\n transport: args.pretty\r\n ? {\r\n target: \"pino-pretty\",\r\n }\r\n : undefined,\r\n },\r\n\r\n args.token\r\n ? pino.transport({\r\n target: \"@logtail/pino\",\r\n options: { sourceToken: args.token },\r\n })\r\n : undefined,\r\n );\r\n\r\n l.child({ service: args.service });\r\n\r\n return l;\r\n};\r\n"],"mappings":";AAAA,SAAS,YAAY;AAEd,IAAM,eAAe,CAAC,SAIvB;AACJ,QAAM,IAAI;AAAA,IACR;AAAA,MACE,OAAO;AAAA,MACP,QAAQ,CAAC;AAAA,MACT,WAAW,KAAK,SACZ;AAAA,QACE,QAAQ;AAAA,MACV,IACA;AAAA,IACN;AAAA,IAEA,KAAK,QACD,KAAK,UAAU;AAAA,MACb,QAAQ;AAAA,MACR,SAAS,EAAE,aAAa,KAAK,MAAM;AAAA,IACrC,CAAC,IACD;AAAA,EACN;AAEA,IAAE,MAAM,EAAE,SAAS,KAAK,QAAQ,CAAC;AAEjC,SAAO;AACT;","names":[]}
@@ -0,0 +1,55 @@
1
+ "use strict";
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __export = (target, all) => {
7
+ for (var name in all)
8
+ __defProp(target, name, { get: all[name], enumerable: true });
9
+ };
10
+ var __copyProps = (to, from, except, desc) => {
11
+ if (from && typeof from === "object" || typeof from === "function") {
12
+ for (let key of __getOwnPropNames(from))
13
+ if (!__hasOwnProp.call(to, key) && key !== except)
14
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
+ }
16
+ return to;
17
+ };
18
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
+
20
+ // src/migrate.ts
21
+ var migrate_exports = {};
22
+ __export(migrate_exports, {
23
+ migrate: () => migrate
24
+ });
25
+ module.exports = __toCommonJS(migrate_exports);
26
+ var import_client = require("@libsql/client");
27
+ var import_libsql = require("drizzle-orm/libsql");
28
+ var import_migrator = require("drizzle-orm/libsql/migrator");
29
+ var migrate = async (schema, args) => {
30
+ let url = args.url;
31
+ if (url.startsWith("http")) {
32
+ url = url.replace(/http(s)?/, "libsql");
33
+ }
34
+ const db = (0, import_libsql.drizzle)(
35
+ (0, import_client.createClient)(
36
+ // Auth token must be either 1) present and not undefined or 2) not present
37
+ args.token ? {
38
+ url,
39
+ authToken: args.token
40
+ } : { url }
41
+ ),
42
+ { schema }
43
+ );
44
+ console.info("Running migrations");
45
+ await (0, import_migrator.migrate)(db, {
46
+ migrationsFolder: args.migrationsFolder
47
+ });
48
+ console.info("Migrations applied");
49
+ process.exit(0);
50
+ };
51
+ // Annotate the CommonJS export names for ESM import in node:
52
+ 0 && (module.exports = {
53
+ migrate
54
+ });
55
+ //# sourceMappingURL=migrate.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/migrate.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\r\nimport { drizzle } from \"drizzle-orm/libsql\";\r\nimport { migrate as runDrizzleMigrate } from \"drizzle-orm/libsql/migrator\";\r\n\r\nexport const migrate = async <TSchema extends Record<string, any>>(\r\n schema: TSchema,\r\n args: {\r\n url: string;\r\n token?: string;\r\n migrationsFolder: string;\r\n }\r\n) => {\r\n let url = args.url;\r\n\r\n // Migrations are only supported via the libsql protocol\r\n if (url.startsWith(\"http\")) {\r\n url = url.replace(/http(s)?/, \"libsql\");\r\n }\r\n\r\n const db = drizzle(\r\n createClient(\r\n // Auth token must be either 1) present and not undefined or 2) not present\r\n args.token\r\n ? {\r\n url,\r\n authToken: args.token,\r\n }\r\n : { url }\r\n ),\r\n { schema }\r\n );\r\n\r\n console.info(\"Running migrations\");\r\n\r\n await runDrizzleMigrate(db, {\r\n migrationsFolder: args.migrationsFolder,\r\n });\r\n\r\n console.info(\"Migrations applied\");\r\n process.exit(0);\r\n};\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAA6B;AAC7B,oBAAwB;AACxB,sBAA6C;AAEtC,IAAM,UAAU,OACrB,QACA,SAKG;AACH,MAAI,MAAM,KAAK;AAGf,MAAI,IAAI,WAAW,MAAM,GAAG;AAC1B,UAAM,IAAI,QAAQ,YAAY,QAAQ;AAAA,EACxC;AAEA,QAAM,SAAK;AAAA,QACT;AAAA;AAAA,MAEE,KAAK,QACD;AAAA,QACE;AAAA,QACA,WAAW,KAAK;AAAA,MAClB,IACA,EAAE,IAAI;AAAA,IACZ;AAAA,IACA,EAAE,OAAO;AAAA,EACX;AAEA,UAAQ,KAAK,oBAAoB;AAEjC,YAAM,gBAAAA,SAAkB,IAAI;AAAA,IAC1B,kBAAkB,KAAK;AAAA,EACzB,CAAC;AAED,UAAQ,KAAK,oBAAoB;AACjC,UAAQ,KAAK,CAAC;AAChB;","names":["runDrizzleMigrate"]}
@@ -0,0 +1,7 @@
1
+ declare const migrate: <TSchema extends Record<string, any>>(schema: TSchema, args: {
2
+ url: string;
3
+ token?: string;
4
+ migrationsFolder: string;
5
+ }) => Promise<never>;
6
+
7
+ export { migrate };
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/migrate.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\nimport { drizzle } from \"drizzle-orm/libsql\";\nimport { migrate as runDrizzleMigrate } from \"drizzle-orm/libsql/migrator\";\n\nexport const migrate = async <TSchema extends Record<string, any>>(\n schema: TSchema,\n args: {\n url: string;\n token?: string;\n migrationsFolder: string;\n }\n) => {\n let url = args.url;\n\n // Migrations are only supported via the libsql protocol\n if (url.startsWith(\"http\")) {\n url = url.replace(/http(s)?/, \"libsql\");\n }\n\n const db = drizzle(\n createClient(\n // Auth token must be either 1) present and not undefined or 2) not present\n args.token\n ? {\n url,\n authToken: args.token,\n }\n : { url }\n ),\n { schema }\n );\n\n console.info(\"Running migrations\");\n\n await runDrizzleMigrate(db, {\n migrationsFolder: args.migrationsFolder,\n });\n\n console.info(\"Migrations applied\");\n process.exit(0);\n};\n"],"mappings":";AAAA,SAAS,oBAAoB;AAC7B,SAAS,eAAe;AACxB,SAAS,WAAW,yBAAyB;AAEtC,IAAM,UAAU,OACrB,QACA,SAKG;AACH,MAAI,MAAM,KAAK;AAGf,MAAI,IAAI,WAAW,MAAM,GAAG;AAC1B,UAAM,IAAI,QAAQ,YAAY,QAAQ;AAAA,EACxC;AAEA,QAAM,KAAK;AAAA,IACT;AAAA;AAAA,MAEE,KAAK,QACD;AAAA,QACE;AAAA,QACA,WAAW,KAAK;AAAA,MAClB,IACA,EAAE,IAAI;AAAA,IACZ;AAAA,IACA,EAAE,OAAO;AAAA,EACX;AAEA,UAAQ,KAAK,oBAAoB;AAEjC,QAAM,kBAAkB,IAAI;AAAA,IAC1B,kBAAkB,KAAK;AAAA,EACzB,CAAC;AAED,UAAQ,KAAK,oBAAoB;AACjC,UAAQ,KAAK,CAAC;AAChB;","names":[]}
1
+ {"version":3,"sources":["../src/migrate.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\r\nimport { drizzle } from \"drizzle-orm/libsql\";\r\nimport { migrate as runDrizzleMigrate } from \"drizzle-orm/libsql/migrator\";\r\n\r\nexport const migrate = async <TSchema extends Record<string, any>>(\r\n schema: TSchema,\r\n args: {\r\n url: string;\r\n token?: string;\r\n migrationsFolder: string;\r\n }\r\n) => {\r\n let url = args.url;\r\n\r\n // Migrations are only supported via the libsql protocol\r\n if (url.startsWith(\"http\")) {\r\n url = url.replace(/http(s)?/, \"libsql\");\r\n }\r\n\r\n const db = drizzle(\r\n createClient(\r\n // Auth token must be either 1) present and not undefined or 2) not present\r\n args.token\r\n ? {\r\n url,\r\n authToken: args.token,\r\n }\r\n : { url }\r\n ),\r\n { schema }\r\n );\r\n\r\n console.info(\"Running migrations\");\r\n\r\n await runDrizzleMigrate(db, {\r\n migrationsFolder: args.migrationsFolder,\r\n });\r\n\r\n console.info(\"Migrations applied\");\r\n process.exit(0);\r\n};\r\n"],"mappings":";AAAA,SAAS,oBAAoB;AAC7B,SAAS,eAAe;AACxB,SAAS,WAAW,yBAAyB;AAEtC,IAAM,UAAU,OACrB,QACA,SAKG;AACH,MAAI,MAAM,KAAK;AAGf,MAAI,IAAI,WAAW,MAAM,GAAG;AAC1B,UAAM,IAAI,QAAQ,YAAY,QAAQ;AAAA,EACxC;AAEA,QAAM,KAAK;AAAA,IACT;AAAA;AAAA,MAEE,KAAK,QACD;AAAA,QACE;AAAA,QACA,WAAW,KAAK;AAAA,MAClB,IACA,EAAE,IAAI;AAAA,IACZ;AAAA,IACA,EAAE,OAAO;AAAA,EACX;AAEA,UAAQ,KAAK,oBAAoB;AAEjC,QAAM,kBAAkB,IAAI;AAAA,IAC1B,kBAAkB,KAAK;AAAA,EACzB,CAAC;AAED,UAAQ,KAAK,oBAAoB;AACjC,UAAQ,KAAK,CAAC;AAChB;","names":[]}
@@ -0,0 +1,44 @@
1
+ "use strict";
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __export = (target, all) => {
7
+ for (var name in all)
8
+ __defProp(target, name, { get: all[name], enumerable: true });
9
+ };
10
+ var __copyProps = (to, from, except, desc) => {
11
+ if (from && typeof from === "object" || typeof from === "function") {
12
+ for (let key of __getOwnPropNames(from))
13
+ if (!__hasOwnProp.call(to, key) && key !== except)
14
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
+ }
16
+ return to;
17
+ };
18
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
+
20
+ // src/schema.ts
21
+ var schema_exports = {};
22
+ __export(schema_exports, {
23
+ columns: () => columns
24
+ });
25
+ module.exports = __toCommonJS(schema_exports);
26
+ var import_cuid2 = require("@paralleldrive/cuid2");
27
+ var import_sqlite_core = require("drizzle-orm/sqlite-core");
28
+ var timeColumns = {
29
+ createdAt: (0, import_sqlite_core.int)("created_at", { mode: "timestamp_ms" }).notNull().$default(() => /* @__PURE__ */ new Date()),
30
+ updatedAt: (0, import_sqlite_core.int)("updated_at", { mode: "timestamp_ms" }).notNull().$default(() => /* @__PURE__ */ new Date()).$onUpdate(() => /* @__PURE__ */ new Date())
31
+ };
32
+ var commonColumns = {
33
+ id: (0, import_sqlite_core.text)("id").primaryKey().$defaultFn(() => (0, import_cuid2.createId)()),
34
+ ...timeColumns
35
+ };
36
+ var columns = {
37
+ time: timeColumns,
38
+ common: commonColumns
39
+ };
40
+ // Annotate the CommonJS export names for ESM import in node:
41
+ 0 && (module.exports = {
42
+ columns
43
+ });
44
+ //# sourceMappingURL=schema.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/schema.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport { int, text } from \"drizzle-orm/sqlite-core\";\r\n\r\nconst timeColumns = {\r\n createdAt: int(\"created_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$default(() => new Date()),\r\n updatedAt: int(\"updated_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$default(() => new Date())\r\n .$onUpdate(() => new Date()),\r\n};\r\n\r\nconst commonColumns = {\r\n id: text(\"id\")\r\n .primaryKey()\r\n .$defaultFn(() => createId()),\r\n ...timeColumns,\r\n};\r\n\r\nexport const columns = {\r\n time: timeColumns,\r\n common: commonColumns,\r\n};\r\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mBAAyB;AACzB,yBAA0B;AAE1B,IAAM,cAAc;AAAA,EAClB,eAAW,wBAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,SAAS,MAAM,oBAAI,KAAK,CAAC;AAAA,EAC5B,eAAW,wBAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,SAAS,MAAM,oBAAI,KAAK,CAAC,EACzB,UAAU,MAAM,oBAAI,KAAK,CAAC;AAC/B;AAEA,IAAM,gBAAgB;AAAA,EACpB,QAAI,yBAAK,IAAI,EACV,WAAW,EACX,WAAW,UAAM,uBAAS,CAAC;AAAA,EAC9B,GAAG;AACL;AAEO,IAAM,UAAU;AAAA,EACrB,MAAM;AAAA,EACN,QAAQ;AACV;","names":[]}
@@ -0,0 +1,16 @@
1
+ import * as drizzle_orm from 'drizzle-orm';
2
+ import * as drizzle_orm_sqlite_core from 'drizzle-orm/sqlite-core';
3
+
4
+ declare const columns: {
5
+ time: {
6
+ createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
7
+ updatedAt: drizzle_orm.HasDefault<drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>>;
8
+ };
9
+ common: {
10
+ createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
11
+ updatedAt: drizzle_orm.HasDefault<drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>>;
12
+ id: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTextBuilderInitial<"id", [string, ...string[]]>>>;
13
+ };
14
+ };
15
+
16
+ export { columns };
package/dist/schema.d.ts CHANGED
@@ -4,11 +4,11 @@ import * as drizzle_orm_sqlite_core from 'drizzle-orm/sqlite-core';
4
4
  declare const columns: {
5
5
  time: {
6
6
  createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
7
- updatedAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>;
7
+ updatedAt: drizzle_orm.HasDefault<drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>>;
8
8
  };
9
9
  common: {
10
10
  createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
11
- updatedAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>;
11
+ updatedAt: drizzle_orm.HasDefault<drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>>;
12
12
  id: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTextBuilderInitial<"id", [string, ...string[]]>>>;
13
13
  };
14
14
  };
package/dist/schema.js CHANGED
@@ -2,8 +2,8 @@
2
2
  import { createId } from "@paralleldrive/cuid2";
3
3
  import { int, text } from "drizzle-orm/sqlite-core";
4
4
  var timeColumns = {
5
- createdAt: int("created_at", { mode: "timestamp_ms" }).notNull().$defaultFn(() => /* @__PURE__ */ new Date()),
6
- updatedAt: int("updated_at", { mode: "timestamp_ms" }).notNull().$defaultFn(() => /* @__PURE__ */ new Date())
5
+ createdAt: int("created_at", { mode: "timestamp_ms" }).notNull().$default(() => /* @__PURE__ */ new Date()),
6
+ updatedAt: int("updated_at", { mode: "timestamp_ms" }).notNull().$default(() => /* @__PURE__ */ new Date()).$onUpdate(() => /* @__PURE__ */ new Date())
7
7
  };
8
8
  var commonColumns = {
9
9
  id: text("id").primaryKey().$defaultFn(() => createId()),
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/schema.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\nimport { int, text } from \"drizzle-orm/sqlite-core\";\n\nconst timeColumns = {\n createdAt: int(\"created_at\", { mode: \"timestamp_ms\" })\n .notNull()\n .$defaultFn(() => new Date()),\n updatedAt: int(\"updated_at\", { mode: \"timestamp_ms\" })\n .notNull()\n .$defaultFn(() => new Date()),\n};\n\nconst commonColumns = {\n id: text(\"id\")\n .primaryKey()\n .$defaultFn(() => createId()),\n ...timeColumns,\n};\n\nexport const columns = {\n time: timeColumns,\n common: commonColumns,\n};\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,KAAK,YAAY;AAE1B,IAAM,cAAc;AAAA,EAClB,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,WAAW,MAAM,oBAAI,KAAK,CAAC;AAAA,EAC9B,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,WAAW,MAAM,oBAAI,KAAK,CAAC;AAChC;AAEA,IAAM,gBAAgB;AAAA,EACpB,IAAI,KAAK,IAAI,EACV,WAAW,EACX,WAAW,MAAM,SAAS,CAAC;AAAA,EAC9B,GAAG;AACL;AAEO,IAAM,UAAU;AAAA,EACrB,MAAM;AAAA,EACN,QAAQ;AACV;","names":[]}
1
+ {"version":3,"sources":["../src/schema.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport { int, text } from \"drizzle-orm/sqlite-core\";\r\n\r\nconst timeColumns = {\r\n createdAt: int(\"created_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$default(() => new Date()),\r\n updatedAt: int(\"updated_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$default(() => new Date())\r\n .$onUpdate(() => new Date()),\r\n};\r\n\r\nconst commonColumns = {\r\n id: text(\"id\")\r\n .primaryKey()\r\n .$defaultFn(() => createId()),\r\n ...timeColumns,\r\n};\r\n\r\nexport const columns = {\r\n time: timeColumns,\r\n common: commonColumns,\r\n};\r\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,KAAK,YAAY;AAE1B,IAAM,cAAc;AAAA,EAClB,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,SAAS,MAAM,oBAAI,KAAK,CAAC;AAAA,EAC5B,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,SAAS,MAAM,oBAAI,KAAK,CAAC,EACzB,UAAU,MAAM,oBAAI,KAAK,CAAC;AAC/B;AAEA,IAAM,gBAAgB;AAAA,EACpB,IAAI,KAAK,IAAI,EACV,WAAW,EACX,WAAW,MAAM,SAAS,CAAC;AAAA,EAC9B,GAAG;AACL;AAEO,IAAM,UAAU;AAAA,EACrB,MAAM;AAAA,EACN,QAAQ;AACV;","names":[]}
package/package.json CHANGED
@@ -9,31 +9,38 @@
9
9
  "exports": {
10
10
  "./db": {
11
11
  "types": "./dist/db.d.ts",
12
- "default": "./dist/db.js"
12
+ "default": "./dist/db.js",
13
+ "require": "./dist/db.cjs"
13
14
  },
14
15
  "./schema": {
15
16
  "types": "./dist/schema.d.ts",
16
- "default": "./dist/schema.js"
17
+ "default": "./dist/schema.js",
18
+ "require": "./dist/schema.cjs"
17
19
  },
18
20
  "./logger": {
19
21
  "types": "./dist/logger.d.ts",
20
- "default": "./dist/logger.js"
22
+ "default": "./dist/logger.js",
23
+ "require": "./dist/logger.cjs"
21
24
  },
22
25
  "./cache": {
23
26
  "types": "./dist/cache.d.ts",
24
- "default": "./dist/cache.js"
27
+ "default": "./dist/cache.js",
28
+ "require": "./dist/cache.cjs"
25
29
  },
26
30
  "./env": {
27
31
  "types": "./dist/env.d.ts",
28
- "default": "./dist/env.js"
32
+ "default": "./dist/env.js",
33
+ "require": "./dist/env.cjs"
29
34
  },
30
35
  "./migrate": {
31
36
  "types": "./dist/migrate.d.ts",
32
- "default": "./dist/migrate.js"
37
+ "default": "./dist/migrate.js",
38
+ "require": "./dist/migrate.cjs"
33
39
  },
34
40
  "./images": {
35
41
  "types": "./dist/images.d.ts",
36
- "default": "./dist/images.js"
42
+ "default": "./dist/images.js",
43
+ "require": "./dist/images.cjs"
37
44
  }
38
45
  },
39
46
  "dependencies": {
@@ -58,7 +65,7 @@
58
65
  "tsup": "^8.0.1",
59
66
  "typescript": "^5.4.5"
60
67
  },
61
- "version": "0.0.10",
68
+ "version": "0.0.12",
62
69
  "scripts": {
63
70
  "dev": "tsup --watch",
64
71
  "build": "tsc --noEmit && tsup",