@conorroberts/utils 0.0.4 → 0.0.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cache.d.ts +11 -0
- package/dist/cache.js +23 -0
- package/dist/cache.js.map +1 -0
- package/dist/db.d.ts +15 -0
- package/dist/db.js +19 -0
- package/dist/db.js.map +1 -0
- package/dist/env.d.ts +16 -0
- package/dist/env.js +27 -0
- package/dist/env.js.map +1 -0
- package/dist/images.d.ts +72 -0
- package/{src/images.ts → dist/images.js} +139 -223
- package/dist/images.js.map +1 -0
- package/dist/logger.d.ts +9 -0
- package/dist/logger.js +23 -0
- package/dist/logger.js.map +1 -0
- package/dist/schema.d.ts +16 -0
- package/dist/schema.js +19 -0
- package/dist/schema.js.map +1 -0
- package/package.json +5 -2
- package/.gitattributes +0 -2
- package/biome.json +0 -53
- package/publish.sh +0 -2
- package/src/cache.ts +0 -24
- package/src/db.ts +0 -28
- package/src/env.ts +0 -52
- package/src/logger.ts +0 -30
- package/src/migrate.ts +0 -41
- package/src/schema.ts +0 -23
- package/src/utils.ts +0 -3
- package/tsconfig.json +0 -19
- package/tsup.config.ts +0 -15
package/dist/cache.d.ts
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
import { Storage, StorageValue } from 'unstorage';
|
|
2
|
+
import { RedisOptions } from 'unstorage/drivers/redis';
|
|
3
|
+
|
|
4
|
+
declare class Cache {
|
|
5
|
+
private _cache;
|
|
6
|
+
constructor(args: RedisOptions);
|
|
7
|
+
get cache(): Storage<StorageValue>;
|
|
8
|
+
ttl(date: Date): number;
|
|
9
|
+
}
|
|
10
|
+
|
|
11
|
+
export { Cache };
|
package/dist/cache.js
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
1
|
+
// src/cache.ts
|
|
2
|
+
import {
|
|
3
|
+
createStorage as createUnstorage
|
|
4
|
+
} from "unstorage";
|
|
5
|
+
import redisDriver from "unstorage/drivers/redis";
|
|
6
|
+
var Cache = class {
|
|
7
|
+
_cache;
|
|
8
|
+
constructor(args) {
|
|
9
|
+
this._cache = createUnstorage({
|
|
10
|
+
driver: redisDriver(args)
|
|
11
|
+
});
|
|
12
|
+
}
|
|
13
|
+
get cache() {
|
|
14
|
+
return this._cache;
|
|
15
|
+
}
|
|
16
|
+
ttl(date) {
|
|
17
|
+
return Math.floor((date.getTime() - Date.now()) / 1e3);
|
|
18
|
+
}
|
|
19
|
+
};
|
|
20
|
+
export {
|
|
21
|
+
Cache
|
|
22
|
+
};
|
|
23
|
+
//# sourceMappingURL=cache.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/cache.ts"],"sourcesContent":["import {\r\n createStorage as createUnstorage,\r\n type Storage,\r\n type StorageValue,\r\n} from \"unstorage\";\r\nimport redisDriver, { type RedisOptions } from \"unstorage/drivers/redis\";\r\n\r\nexport class Cache {\r\n private _cache: Storage<StorageValue>;\r\n\r\n constructor(args: RedisOptions) {\r\n this._cache = createUnstorage({\r\n driver: redisDriver(args),\r\n });\r\n }\r\n\r\n get cache() {\r\n return this._cache;\r\n }\r\n\r\n public ttl(date: Date) {\r\n return Math.floor((date.getTime() - Date.now()) / 1000);\r\n }\r\n}\r\n"],"mappings":";AAAA;AAAA,EACE,iBAAiB;AAAA,OAGZ;AACP,OAAO,iBAAwC;AAExC,IAAM,QAAN,MAAY;AAAA,EACT;AAAA,EAER,YAAY,MAAoB;AAC9B,SAAK,SAAS,gBAAgB;AAAA,MAC5B,QAAQ,YAAY,IAAI;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,IAAI,QAAQ;AACV,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,MAAY;AACrB,WAAO,KAAK,OAAO,KAAK,QAAQ,IAAI,KAAK,IAAI,KAAK,GAAI;AAAA,EACxD;AACF;","names":[]}
|
package/dist/db.d.ts
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
import * as _libsql_client from '@libsql/client';
|
|
2
|
+
import { LibSQLDatabase } from 'drizzle-orm/libsql';
|
|
3
|
+
|
|
4
|
+
declare const createLibsqlClient: (args: {
|
|
5
|
+
url: string;
|
|
6
|
+
authToken?: string;
|
|
7
|
+
}) => _libsql_client.Client;
|
|
8
|
+
declare const createDbClient: <TSchema extends Record<string, unknown>>(schema: TSchema, args: {
|
|
9
|
+
url: string;
|
|
10
|
+
authToken?: string;
|
|
11
|
+
}) => LibSQLDatabase<TSchema>;
|
|
12
|
+
type DatabaseClient<TSchema extends Record<string, unknown>> = LibSQLDatabase<TSchema>;
|
|
13
|
+
type DatabaseClientTransactionContext<TSchema extends Record<string, unknown>> = Parameters<Parameters<DatabaseClient<TSchema>["transaction"]>[0]>[0];
|
|
14
|
+
|
|
15
|
+
export { type DatabaseClient, type DatabaseClientTransactionContext, createDbClient, createLibsqlClient };
|
package/dist/db.js
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
// src/db.ts
|
|
2
|
+
import { createClient } from "@libsql/client";
|
|
3
|
+
import { drizzle } from "drizzle-orm/libsql";
|
|
4
|
+
var createLibsqlClient = (args) => {
|
|
5
|
+
return createClient(args);
|
|
6
|
+
};
|
|
7
|
+
var createDbClient = (schema, args) => {
|
|
8
|
+
const client = createLibsqlClient(args);
|
|
9
|
+
const db = drizzle(client, {
|
|
10
|
+
schema,
|
|
11
|
+
logger: false
|
|
12
|
+
});
|
|
13
|
+
return db;
|
|
14
|
+
};
|
|
15
|
+
export {
|
|
16
|
+
createDbClient,
|
|
17
|
+
createLibsqlClient
|
|
18
|
+
};
|
|
19
|
+
//# sourceMappingURL=db.js.map
|
package/dist/db.js.map
ADDED
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/db.ts"],"sourcesContent":["import { createClient } from \"@libsql/client\";\r\nimport { LibSQLDatabase, drizzle } from \"drizzle-orm/libsql\";\r\n\r\nexport const createLibsqlClient = (args: {\r\n url: string;\r\n authToken?: string;\r\n}) => {\r\n return createClient(args);\r\n};\r\n\r\nexport const createDbClient = <TSchema extends Record<string, unknown>>(\r\n schema: TSchema,\r\n args: { url: string; authToken?: string }\r\n) => {\r\n const client = createLibsqlClient(args);\r\n const db = drizzle(client, {\r\n schema,\r\n logger: false,\r\n });\r\n\r\n return db;\r\n};\r\n\r\nexport type DatabaseClient<TSchema extends Record<string, unknown>> =\r\n LibSQLDatabase<TSchema>;\r\nexport type DatabaseClientTransactionContext<\r\n TSchema extends Record<string, unknown>\r\n> = Parameters<Parameters<DatabaseClient<TSchema>[\"transaction\"]>[0]>[0];\r\n"],"mappings":";AAAA,SAAS,oBAAoB;AAC7B,SAAyB,eAAe;AAEjC,IAAM,qBAAqB,CAAC,SAG7B;AACJ,SAAO,aAAa,IAAI;AAC1B;AAEO,IAAM,iBAAiB,CAC5B,QACA,SACG;AACH,QAAM,SAAS,mBAAmB,IAAI;AACtC,QAAM,KAAK,QAAQ,QAAQ;AAAA,IACzB;AAAA,IACA,QAAQ;AAAA,EACV,CAAC;AAED,SAAO;AACT;","names":[]}
|
package/dist/env.d.ts
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
import { BaseSchema, Output } from 'valibot';
|
|
2
|
+
|
|
3
|
+
/**
|
|
4
|
+
* Validates your environment variables against the given Valibot schema;
|
|
5
|
+
* @param args
|
|
6
|
+
* @returns An object containing client environment variables and another containing server environment variables
|
|
7
|
+
*/
|
|
8
|
+
declare const createEnv: <Schema extends Record<string, BaseSchema>, Env = { [K in keyof Schema]: Output<Schema[K]>; }>(args: {
|
|
9
|
+
schema: Schema;
|
|
10
|
+
env: any;
|
|
11
|
+
}) => {
|
|
12
|
+
client: { [B in Exclude<{ [K_1 in keyof Env]: K_1 extends `PUBLIC_${string}` ? K_1 : never; }[keyof Env], undefined>]: Env[B]; };
|
|
13
|
+
server: Env;
|
|
14
|
+
};
|
|
15
|
+
|
|
16
|
+
export { createEnv };
|
package/dist/env.js
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
1
|
+
// src/env.ts
|
|
2
|
+
import { pipe } from "remeda";
|
|
3
|
+
import * as v from "valibot";
|
|
4
|
+
var PUBLIC_ENV_PREFIX = "PUBLIC_";
|
|
5
|
+
var createEnv = (args) => {
|
|
6
|
+
const pairs = Object.entries(args.schema);
|
|
7
|
+
const serverEnv = /* @__PURE__ */ new Map();
|
|
8
|
+
for (const [key, value] of pairs) {
|
|
9
|
+
const result = v.safeParse(value, args.env[key] ?? null);
|
|
10
|
+
if (!result.success) {
|
|
11
|
+
console.error(`Environment variable "${key}" is invalid`);
|
|
12
|
+
process.exit(1);
|
|
13
|
+
}
|
|
14
|
+
serverEnv.set(key, result.output);
|
|
15
|
+
}
|
|
16
|
+
const clientEnv = pipe(
|
|
17
|
+
serverEnv,
|
|
18
|
+
(obj) => Array.from(obj.entries()),
|
|
19
|
+
(pairs2) => pairs2.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),
|
|
20
|
+
(pairs2) => Object.fromEntries(pairs2)
|
|
21
|
+
);
|
|
22
|
+
return { client: clientEnv, server: Object.fromEntries(serverEnv.entries()) };
|
|
23
|
+
};
|
|
24
|
+
export {
|
|
25
|
+
createEnv
|
|
26
|
+
};
|
|
27
|
+
//# sourceMappingURL=env.js.map
|
package/dist/env.js.map
ADDED
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/env.ts"],"sourcesContent":["import { pipe } from \"remeda\";\r\nimport type { BaseSchema, Output } from \"valibot\";\r\nimport * as v from \"valibot\";\r\n\r\nconst PUBLIC_ENV_PREFIX = \"PUBLIC_\" as const;\r\n\r\n/**\r\n * Validates your environment variables against the given Valibot schema;\r\n * @param args\r\n * @returns An object containing client environment variables and another containing server environment variables\r\n */\r\nexport const createEnv = <\r\n Schema extends Record<string, BaseSchema>,\r\n Env = {\r\n [K in keyof Schema]: Output<Schema[K]>;\r\n },\r\n>(args: {\r\n schema: Schema;\r\n env: any;\r\n}) => {\r\n const pairs = Object.entries(args.schema);\r\n const serverEnv = new Map();\r\n\r\n for (const [key, value] of pairs) {\r\n const result = v.safeParse(value, args.env[key] ?? null);\r\n\r\n if (!result.success) {\r\n console.error(`Environment variable \"${key}\" is invalid`);\r\n process.exit(1);\r\n }\r\n\r\n serverEnv.set(key, result.output);\r\n }\r\n\r\n type ClientEnvKeys = Exclude<\r\n { [K in keyof Env]: K extends `${typeof PUBLIC_ENV_PREFIX}${string}` ? K : never }[keyof Env],\r\n undefined\r\n >;\r\n\r\n type ClientEnv = {\r\n [B in ClientEnvKeys]: Env[B];\r\n };\r\n\r\n const clientEnv = pipe(\r\n serverEnv,\r\n (obj) => Array.from(obj.entries()),\r\n (pairs) => pairs.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),\r\n (pairs) => Object.fromEntries(pairs),\r\n ) as ClientEnv;\r\n\r\n return { client: clientEnv, server: Object.fromEntries(serverEnv.entries()) as Env };\r\n};\r\n"],"mappings":";AAAA,SAAS,YAAY;AAErB,YAAY,OAAO;AAEnB,IAAM,oBAAoB;AAOnB,IAAM,YAAY,CAKvB,SAGI;AACJ,QAAM,QAAQ,OAAO,QAAQ,KAAK,MAAM;AACxC,QAAM,YAAY,oBAAI,IAAI;AAE1B,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO;AAChC,UAAM,SAAW,YAAU,OAAO,KAAK,IAAI,GAAG,KAAK,IAAI;AAEvD,QAAI,CAAC,OAAO,SAAS;AACnB,cAAQ,MAAM,yBAAyB,GAAG,cAAc;AACxD,cAAQ,KAAK,CAAC;AAAA,IAChB;AAEA,cAAU,IAAI,KAAK,OAAO,MAAM;AAAA,EAClC;AAWA,QAAM,YAAY;AAAA,IAChB;AAAA,IACA,CAAC,QAAQ,MAAM,KAAK,IAAI,QAAQ,CAAC;AAAA,IACjC,CAACA,WAAUA,OAAM,OAAO,CAAC,CAAC,CAAC,MAAM,EAAE,WAAW,iBAAiB,CAAC;AAAA,IAChE,CAACA,WAAU,OAAO,YAAYA,MAAK;AAAA,EACrC;AAEA,SAAO,EAAE,QAAQ,WAAW,QAAQ,OAAO,YAAY,UAAU,QAAQ,CAAC,EAAS;AACrF;","names":["pairs"]}
|
package/dist/images.d.ts
ADDED
|
@@ -0,0 +1,72 @@
|
|
|
1
|
+
interface OptimizedImageOptions {
|
|
2
|
+
anim?: boolean;
|
|
3
|
+
background?: string;
|
|
4
|
+
blur?: number;
|
|
5
|
+
brightness?: number;
|
|
6
|
+
compression?: "fast";
|
|
7
|
+
contrast?: number;
|
|
8
|
+
dpr?: number;
|
|
9
|
+
fit?: "scale-down" | "contain" | "cover" | "crop" | "pad";
|
|
10
|
+
format?: "webp" | "avif" | "json";
|
|
11
|
+
gamma?: number;
|
|
12
|
+
width?: number;
|
|
13
|
+
height?: number;
|
|
14
|
+
metadata?: "keep" | "copyright" | "none";
|
|
15
|
+
quality?: number;
|
|
16
|
+
rotate?: number;
|
|
17
|
+
sharpen?: number;
|
|
18
|
+
}
|
|
19
|
+
interface CreateImageUrlResponse {
|
|
20
|
+
result: {
|
|
21
|
+
id: string;
|
|
22
|
+
uploadURL: string;
|
|
23
|
+
};
|
|
24
|
+
success: boolean;
|
|
25
|
+
errors: unknown[];
|
|
26
|
+
messages: unknown[];
|
|
27
|
+
}
|
|
28
|
+
declare class ImageUtils<ImageIds extends Record<string, string>> {
|
|
29
|
+
private blacklist;
|
|
30
|
+
private account;
|
|
31
|
+
private _imageIds;
|
|
32
|
+
constructor(args: {
|
|
33
|
+
accountId: string;
|
|
34
|
+
blacklist?: string[];
|
|
35
|
+
imageIds?: ImageIds;
|
|
36
|
+
});
|
|
37
|
+
get imageIds(): ImageIds;
|
|
38
|
+
url(id: string): string;
|
|
39
|
+
private isBlacklisted;
|
|
40
|
+
private isProtected;
|
|
41
|
+
/**
|
|
42
|
+
* Will only operate on images that have been uploaded via cloudflare images
|
|
43
|
+
*/
|
|
44
|
+
optimizeUrl(url: string, options: OptimizedImageOptions): string;
|
|
45
|
+
optimizeId(id: string, options: OptimizedImageOptions): string;
|
|
46
|
+
createOptionsSearchParams(options: OptimizedImageOptions): URLSearchParams;
|
|
47
|
+
createImageOptionsString(options: OptimizedImageOptions): string;
|
|
48
|
+
createUploadUrls(count: number, args: {
|
|
49
|
+
apiKey: string;
|
|
50
|
+
}): Promise<{
|
|
51
|
+
url: string;
|
|
52
|
+
id: string;
|
|
53
|
+
}[]>;
|
|
54
|
+
upload(url: string, body: FormData): Promise<string>;
|
|
55
|
+
delete(id: string, args: {
|
|
56
|
+
apiKey: string;
|
|
57
|
+
}): Promise<{
|
|
58
|
+
success: boolean;
|
|
59
|
+
}>;
|
|
60
|
+
batchUpload(files: {
|
|
61
|
+
file: File;
|
|
62
|
+
url: {
|
|
63
|
+
id: string;
|
|
64
|
+
value: string;
|
|
65
|
+
};
|
|
66
|
+
}[]): Promise<{
|
|
67
|
+
url: string;
|
|
68
|
+
id: string;
|
|
69
|
+
}[]>;
|
|
70
|
+
}
|
|
71
|
+
|
|
72
|
+
export { type CreateImageUrlResponse, ImageUtils, type OptimizedImageOptions };
|
|
@@ -1,223 +1,139 @@
|
|
|
1
|
-
|
|
2
|
-
import
|
|
3
|
-
import
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
}
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
}
|
|
57
|
-
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
return
|
|
88
|
-
}
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
if (
|
|
95
|
-
|
|
96
|
-
}
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
111
|
-
|
|
112
|
-
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
|
|
127
|
-
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
136
|
-
|
|
137
|
-
|
|
138
|
-
|
|
139
|
-
|
|
140
|
-
try {
|
|
141
|
-
const form = new FormData();
|
|
142
|
-
const id = createId();
|
|
143
|
-
form.append("id", id);
|
|
144
|
-
form.append("expiry", dayjs().add(5, "minute").toISOString());
|
|
145
|
-
|
|
146
|
-
const img = await ofetch<CreateImageUrlResponse>(
|
|
147
|
-
`https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,
|
|
148
|
-
{ method: "POST", headers, body: form }
|
|
149
|
-
);
|
|
150
|
-
|
|
151
|
-
if (!img.success) {
|
|
152
|
-
throw new Error("Error uploading image");
|
|
153
|
-
}
|
|
154
|
-
|
|
155
|
-
return { url: img.result.uploadURL, id };
|
|
156
|
-
} catch (e) {
|
|
157
|
-
throw e;
|
|
158
|
-
}
|
|
159
|
-
})
|
|
160
|
-
);
|
|
161
|
-
|
|
162
|
-
return urls;
|
|
163
|
-
}
|
|
164
|
-
|
|
165
|
-
public async upload(url: string, body: FormData) {
|
|
166
|
-
const fetchResponse = await ofetch<UploadImageResponse>(url, {
|
|
167
|
-
method: "POST",
|
|
168
|
-
body,
|
|
169
|
-
});
|
|
170
|
-
|
|
171
|
-
if (!fetchResponse.success) {
|
|
172
|
-
throw new Error("Failed to upload image");
|
|
173
|
-
}
|
|
174
|
-
|
|
175
|
-
const downloadUrl = fetchResponse.result.variants[0];
|
|
176
|
-
|
|
177
|
-
if (!downloadUrl) {
|
|
178
|
-
throw new Error("Could not find download URL");
|
|
179
|
-
}
|
|
180
|
-
|
|
181
|
-
return downloadUrl;
|
|
182
|
-
}
|
|
183
|
-
|
|
184
|
-
public async delete(id: string, args: { apiKey: string }) {
|
|
185
|
-
if (this.isProtected(id)) {
|
|
186
|
-
return { success: true };
|
|
187
|
-
}
|
|
188
|
-
|
|
189
|
-
try {
|
|
190
|
-
const headers = new Headers();
|
|
191
|
-
headers.set("Authorization", `Bearer ${args.apiKey}`);
|
|
192
|
-
|
|
193
|
-
await ofetch(
|
|
194
|
-
`https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,
|
|
195
|
-
{
|
|
196
|
-
method: "POST",
|
|
197
|
-
headers,
|
|
198
|
-
}
|
|
199
|
-
);
|
|
200
|
-
return { success: true };
|
|
201
|
-
} catch (_e) {
|
|
202
|
-
return { success: false };
|
|
203
|
-
}
|
|
204
|
-
}
|
|
205
|
-
|
|
206
|
-
public async batchUpload(
|
|
207
|
-
files: { file: File; url: { id: string; value: string } }[]
|
|
208
|
-
) {
|
|
209
|
-
return await Promise.all(
|
|
210
|
-
files.map(async (e) => {
|
|
211
|
-
const formData = new FormData();
|
|
212
|
-
formData.append("file", e.file);
|
|
213
|
-
|
|
214
|
-
const downloadUrl = await this.upload(e.url.value, formData);
|
|
215
|
-
|
|
216
|
-
return {
|
|
217
|
-
url: downloadUrl,
|
|
218
|
-
id: e.url.id,
|
|
219
|
-
};
|
|
220
|
-
})
|
|
221
|
-
);
|
|
222
|
-
}
|
|
223
|
-
}
|
|
1
|
+
// src/images.ts
|
|
2
|
+
import { createId } from "@paralleldrive/cuid2";
|
|
3
|
+
import dayjs from "dayjs";
|
|
4
|
+
import { ofetch } from "ofetch";
|
|
5
|
+
var ImageUtils = class {
|
|
6
|
+
blacklist = ["img.clerk.com"];
|
|
7
|
+
account;
|
|
8
|
+
_imageIds;
|
|
9
|
+
constructor(args) {
|
|
10
|
+
this.account = args.accountId;
|
|
11
|
+
this._imageIds = args.imageIds;
|
|
12
|
+
if (args.blacklist) {
|
|
13
|
+
this.blacklist.push(...args.blacklist);
|
|
14
|
+
}
|
|
15
|
+
}
|
|
16
|
+
get imageIds() {
|
|
17
|
+
if (!this._imageIds) {
|
|
18
|
+
throw new Error(`imageIds was not supplied in constructor`);
|
|
19
|
+
}
|
|
20
|
+
return this._imageIds;
|
|
21
|
+
}
|
|
22
|
+
url(id) {
|
|
23
|
+
return `https://imagedelivery.net/${this.account}/${id}/public`;
|
|
24
|
+
}
|
|
25
|
+
isBlacklisted(url) {
|
|
26
|
+
return this.blacklist.some((u) => url.includes(u));
|
|
27
|
+
}
|
|
28
|
+
isProtected(id) {
|
|
29
|
+
if (!this._imageIds) {
|
|
30
|
+
return false;
|
|
31
|
+
}
|
|
32
|
+
return Object.values(this._imageIds).some((e) => e === id);
|
|
33
|
+
}
|
|
34
|
+
/**
|
|
35
|
+
* Will only operate on images that have been uploaded via cloudflare images
|
|
36
|
+
*/
|
|
37
|
+
optimizeUrl(url, options) {
|
|
38
|
+
if (this.isBlacklisted(url)) {
|
|
39
|
+
return url;
|
|
40
|
+
}
|
|
41
|
+
return url.replace("public", this.createImageOptionsString(options));
|
|
42
|
+
}
|
|
43
|
+
optimizeId(id, options) {
|
|
44
|
+
return this.optimizeUrl(this.url(id), options);
|
|
45
|
+
}
|
|
46
|
+
createOptionsSearchParams(options) {
|
|
47
|
+
const params = new URLSearchParams();
|
|
48
|
+
const pairs = Object.entries(options);
|
|
49
|
+
for (const [key, val] of pairs) {
|
|
50
|
+
if (val === void 0) {
|
|
51
|
+
continue;
|
|
52
|
+
}
|
|
53
|
+
params.set(key, val.toString());
|
|
54
|
+
}
|
|
55
|
+
return params;
|
|
56
|
+
}
|
|
57
|
+
createImageOptionsString(options) {
|
|
58
|
+
const params = this.createOptionsSearchParams(options);
|
|
59
|
+
return Array.from(params.entries()).map(([key, val]) => `${key}=${val}`).join(",");
|
|
60
|
+
}
|
|
61
|
+
async createUploadUrls(count, args) {
|
|
62
|
+
if (count === 0) {
|
|
63
|
+
return [];
|
|
64
|
+
}
|
|
65
|
+
const headers = new Headers();
|
|
66
|
+
headers.set("Authorization", `Bearer ${args.apiKey}`);
|
|
67
|
+
const urls = await Promise.all(
|
|
68
|
+
Array.from({ length: count }).map(async () => {
|
|
69
|
+
try {
|
|
70
|
+
const form = new FormData();
|
|
71
|
+
const id = createId();
|
|
72
|
+
form.append("id", id);
|
|
73
|
+
form.append("expiry", dayjs().add(5, "minute").toISOString());
|
|
74
|
+
const img = await ofetch(
|
|
75
|
+
`https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,
|
|
76
|
+
{ method: "POST", headers, body: form }
|
|
77
|
+
);
|
|
78
|
+
if (!img.success) {
|
|
79
|
+
throw new Error("Error uploading image");
|
|
80
|
+
}
|
|
81
|
+
return { url: img.result.uploadURL, id };
|
|
82
|
+
} catch (e) {
|
|
83
|
+
throw e;
|
|
84
|
+
}
|
|
85
|
+
})
|
|
86
|
+
);
|
|
87
|
+
return urls;
|
|
88
|
+
}
|
|
89
|
+
async upload(url, body) {
|
|
90
|
+
const fetchResponse = await ofetch(url, {
|
|
91
|
+
method: "POST",
|
|
92
|
+
body
|
|
93
|
+
});
|
|
94
|
+
if (!fetchResponse.success) {
|
|
95
|
+
throw new Error("Failed to upload image");
|
|
96
|
+
}
|
|
97
|
+
const downloadUrl = fetchResponse.result.variants[0];
|
|
98
|
+
if (!downloadUrl) {
|
|
99
|
+
throw new Error("Could not find download URL");
|
|
100
|
+
}
|
|
101
|
+
return downloadUrl;
|
|
102
|
+
}
|
|
103
|
+
async delete(id, args) {
|
|
104
|
+
if (this.isProtected(id)) {
|
|
105
|
+
return { success: true };
|
|
106
|
+
}
|
|
107
|
+
try {
|
|
108
|
+
const headers = new Headers();
|
|
109
|
+
headers.set("Authorization", `Bearer ${args.apiKey}`);
|
|
110
|
+
await ofetch(
|
|
111
|
+
`https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,
|
|
112
|
+
{
|
|
113
|
+
method: "POST",
|
|
114
|
+
headers
|
|
115
|
+
}
|
|
116
|
+
);
|
|
117
|
+
return { success: true };
|
|
118
|
+
} catch (_e) {
|
|
119
|
+
return { success: false };
|
|
120
|
+
}
|
|
121
|
+
}
|
|
122
|
+
async batchUpload(files) {
|
|
123
|
+
return await Promise.all(
|
|
124
|
+
files.map(async (e) => {
|
|
125
|
+
const formData = new FormData();
|
|
126
|
+
formData.append("file", e.file);
|
|
127
|
+
const downloadUrl = await this.upload(e.url.value, formData);
|
|
128
|
+
return {
|
|
129
|
+
url: downloadUrl,
|
|
130
|
+
id: e.url.id
|
|
131
|
+
};
|
|
132
|
+
})
|
|
133
|
+
);
|
|
134
|
+
}
|
|
135
|
+
};
|
|
136
|
+
export {
|
|
137
|
+
ImageUtils
|
|
138
|
+
};
|
|
139
|
+
//# sourceMappingURL=images.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/images.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport dayjs from \"dayjs\";\r\nimport { ofetch } from \"ofetch\";\r\n\r\nexport interface OptimizedImageOptions {\r\n anim?: boolean;\r\n background?: string;\r\n blur?: number;\r\n brightness?: number;\r\n compression?: \"fast\"; // faster compression = larger file size\r\n contrast?: number;\r\n dpr?: number;\r\n fit?: \"scale-down\" | \"contain\" | \"cover\" | \"crop\" | \"pad\";\r\n format?: \"webp\" | \"avif\" | \"json\";\r\n gamma?: number;\r\n width?: number;\r\n height?: number;\r\n metadata?: \"keep\" | \"copyright\" | \"none\";\r\n quality?: number;\r\n rotate?: number;\r\n sharpen?: number;\r\n}\r\n\r\nexport interface CreateImageUrlResponse {\r\n result: {\r\n id: string;\r\n uploadURL: string;\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\ninterface UploadImageResponse {\r\n result: {\r\n id: string;\r\n filename: string;\r\n uploaded: string;\r\n requireSignedURLs: boolean;\r\n variants: string[];\r\n };\r\n success: boolean;\r\n errors: unknown[];\r\n messages: unknown[];\r\n}\r\n\r\nexport class ImageUtils<ImageIds extends Record<string, string>> {\r\n private blacklist: string[] = [\"img.clerk.com\"];\r\n private account: string;\r\n private _imageIds: ImageIds | undefined;\r\n\r\n constructor(args: {\r\n accountId: string;\r\n blacklist?: string[];\r\n imageIds?: ImageIds;\r\n }) {\r\n this.account = args.accountId;\r\n\r\n this._imageIds = args.imageIds;\r\n\r\n if (args.blacklist) {\r\n this.blacklist.push(...args.blacklist);\r\n }\r\n }\r\n\r\n get imageIds() {\r\n if (!this._imageIds) {\r\n throw new Error(`imageIds was not supplied in constructor`);\r\n }\r\n\r\n return this._imageIds;\r\n }\r\n\r\n public url(id: string) {\r\n return `https://imagedelivery.net/${this.account}/${id}/public`;\r\n }\r\n\r\n private isBlacklisted(url: string) {\r\n return this.blacklist.some((u) => url.includes(u));\r\n }\r\n\r\n private isProtected(id: string) {\r\n if (!this._imageIds) {\r\n return false;\r\n }\r\n\r\n return Object.values(this._imageIds).some((e) => e === id);\r\n }\r\n\r\n /**\r\n * Will only operate on images that have been uploaded via cloudflare images\r\n */\r\n public optimizeUrl(url: string, options: OptimizedImageOptions) {\r\n if (this.isBlacklisted(url)) {\r\n return url;\r\n }\r\n\r\n // Final format should look similar to: https://imagedelivery.net/<ACCOUNT_HASH>/<IMAGE_ID>/w=400,sharpen=3\r\n return url.replace(\"public\", this.createImageOptionsString(options));\r\n }\r\n\r\n public optimizeId(id: string, options: OptimizedImageOptions) {\r\n return this.optimizeUrl(this.url(id), options);\r\n }\r\n\r\n public createOptionsSearchParams(options: OptimizedImageOptions) {\r\n const params = new URLSearchParams();\r\n\r\n const pairs = Object.entries(options);\r\n\r\n for (const [key, val] of pairs) {\r\n if (val === undefined) {\r\n continue;\r\n }\r\n\r\n params.set(key, val.toString());\r\n }\r\n\r\n return params;\r\n }\r\n\r\n public createImageOptionsString(options: OptimizedImageOptions) {\r\n const params = this.createOptionsSearchParams(options);\r\n\r\n return Array.from(params.entries())\r\n .map(([key, val]) => `${key}=${val}`)\r\n .join(\",\");\r\n }\r\n\r\n public async createUploadUrls(count: number, args: { apiKey: string }) {\r\n if (count === 0) {\r\n return [];\r\n }\r\n\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n const urls = await Promise.all(\r\n Array.from({ length: count }).map(async () => {\r\n try {\r\n const form = new FormData();\r\n const id = createId();\r\n form.append(\"id\", id);\r\n form.append(\"expiry\", dayjs().add(5, \"minute\").toISOString());\r\n\r\n const img = await ofetch<CreateImageUrlResponse>(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v2/direct_upload`,\r\n { method: \"POST\", headers, body: form }\r\n );\r\n\r\n if (!img.success) {\r\n throw new Error(\"Error uploading image\");\r\n }\r\n\r\n return { url: img.result.uploadURL, id };\r\n } catch (e) {\r\n throw e;\r\n }\r\n })\r\n );\r\n\r\n return urls;\r\n }\r\n\r\n public async upload(url: string, body: FormData) {\r\n const fetchResponse = await ofetch<UploadImageResponse>(url, {\r\n method: \"POST\",\r\n body,\r\n });\r\n\r\n if (!fetchResponse.success) {\r\n throw new Error(\"Failed to upload image\");\r\n }\r\n\r\n const downloadUrl = fetchResponse.result.variants[0];\r\n\r\n if (!downloadUrl) {\r\n throw new Error(\"Could not find download URL\");\r\n }\r\n\r\n return downloadUrl;\r\n }\r\n\r\n public async delete(id: string, args: { apiKey: string }) {\r\n if (this.isProtected(id)) {\r\n return { success: true };\r\n }\r\n\r\n try {\r\n const headers = new Headers();\r\n headers.set(\"Authorization\", `Bearer ${args.apiKey}`);\r\n\r\n await ofetch(\r\n `https://api.cloudflare.com/client/v4/accounts/${this.account}/images/v1/${id}`,\r\n {\r\n method: \"POST\",\r\n headers,\r\n }\r\n );\r\n return { success: true };\r\n } catch (_e) {\r\n return { success: false };\r\n }\r\n }\r\n\r\n public async batchUpload(\r\n files: { file: File; url: { id: string; value: string } }[]\r\n ) {\r\n return await Promise.all(\r\n files.map(async (e) => {\r\n const formData = new FormData();\r\n formData.append(\"file\", e.file);\r\n\r\n const downloadUrl = await this.upload(e.url.value, formData);\r\n\r\n return {\r\n url: downloadUrl,\r\n id: e.url.id,\r\n };\r\n })\r\n );\r\n }\r\n}\r\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,OAAO,WAAW;AAClB,SAAS,cAAc;AA4ChB,IAAM,aAAN,MAA0D;AAAA,EACvD,YAAsB,CAAC,eAAe;AAAA,EACtC;AAAA,EACA;AAAA,EAER,YAAY,MAIT;AACD,SAAK,UAAU,KAAK;AAEpB,SAAK,YAAY,KAAK;AAEtB,QAAI,KAAK,WAAW;AAClB,WAAK,UAAU,KAAK,GAAG,KAAK,SAAS;AAAA,IACvC;AAAA,EACF;AAAA,EAEA,IAAI,WAAW;AACb,QAAI,CAAC,KAAK,WAAW;AACnB,YAAM,IAAI,MAAM,0CAA0C;AAAA,IAC5D;AAEA,WAAO,KAAK;AAAA,EACd;AAAA,EAEO,IAAI,IAAY;AACrB,WAAO,6BAA6B,KAAK,OAAO,IAAI,EAAE;AAAA,EACxD;AAAA,EAEQ,cAAc,KAAa;AACjC,WAAO,KAAK,UAAU,KAAK,CAAC,MAAM,IAAI,SAAS,CAAC,CAAC;AAAA,EACnD;AAAA,EAEQ,YAAY,IAAY;AAC9B,QAAI,CAAC,KAAK,WAAW;AACnB,aAAO;AAAA,IACT;AAEA,WAAO,OAAO,OAAO,KAAK,SAAS,EAAE,KAAK,CAAC,MAAM,MAAM,EAAE;AAAA,EAC3D;AAAA;AAAA;AAAA;AAAA,EAKO,YAAY,KAAa,SAAgC;AAC9D,QAAI,KAAK,cAAc,GAAG,GAAG;AAC3B,aAAO;AAAA,IACT;AAGA,WAAO,IAAI,QAAQ,UAAU,KAAK,yBAAyB,OAAO,CAAC;AAAA,EACrE;AAAA,EAEO,WAAW,IAAY,SAAgC;AAC5D,WAAO,KAAK,YAAY,KAAK,IAAI,EAAE,GAAG,OAAO;AAAA,EAC/C;AAAA,EAEO,0BAA0B,SAAgC;AAC/D,UAAM,SAAS,IAAI,gBAAgB;AAEnC,UAAM,QAAQ,OAAO,QAAQ,OAAO;AAEpC,eAAW,CAAC,KAAK,GAAG,KAAK,OAAO;AAC9B,UAAI,QAAQ,QAAW;AACrB;AAAA,MACF;AAEA,aAAO,IAAI,KAAK,IAAI,SAAS,CAAC;AAAA,IAChC;AAEA,WAAO;AAAA,EACT;AAAA,EAEO,yBAAyB,SAAgC;AAC9D,UAAM,SAAS,KAAK,0BAA0B,OAAO;AAErD,WAAO,MAAM,KAAK,OAAO,QAAQ,CAAC,EAC/B,IAAI,CAAC,CAAC,KAAK,GAAG,MAAM,GAAG,GAAG,IAAI,GAAG,EAAE,EACnC,KAAK,GAAG;AAAA,EACb;AAAA,EAEA,MAAa,iBAAiB,OAAe,MAA0B;AACrE,QAAI,UAAU,GAAG;AACf,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAU,IAAI,QAAQ;AAC5B,YAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,UAAM,OAAO,MAAM,QAAQ;AAAA,MACzB,MAAM,KAAK,EAAE,QAAQ,MAAM,CAAC,EAAE,IAAI,YAAY;AAC5C,YAAI;AACF,gBAAM,OAAO,IAAI,SAAS;AAC1B,gBAAM,KAAK,SAAS;AACpB,eAAK,OAAO,MAAM,EAAE;AACpB,eAAK,OAAO,UAAU,MAAM,EAAE,IAAI,GAAG,QAAQ,EAAE,YAAY,CAAC;AAE5D,gBAAM,MAAM,MAAM;AAAA,YAChB,iDAAiD,KAAK,OAAO;AAAA,YAC7D,EAAE,QAAQ,QAAQ,SAAS,MAAM,KAAK;AAAA,UACxC;AAEA,cAAI,CAAC,IAAI,SAAS;AAChB,kBAAM,IAAI,MAAM,uBAAuB;AAAA,UACzC;AAEA,iBAAO,EAAE,KAAK,IAAI,OAAO,WAAW,GAAG;AAAA,QACzC,SAAS,GAAG;AACV,gBAAM;AAAA,QACR;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,KAAa,MAAgB;AAC/C,UAAM,gBAAgB,MAAM,OAA4B,KAAK;AAAA,MAC3D,QAAQ;AAAA,MACR;AAAA,IACF,CAAC;AAED,QAAI,CAAC,cAAc,SAAS;AAC1B,YAAM,IAAI,MAAM,wBAAwB;AAAA,IAC1C;AAEA,UAAM,cAAc,cAAc,OAAO,SAAS,CAAC;AAEnD,QAAI,CAAC,aAAa;AAChB,YAAM,IAAI,MAAM,6BAA6B;AAAA,IAC/C;AAEA,WAAO;AAAA,EACT;AAAA,EAEA,MAAa,OAAO,IAAY,MAA0B;AACxD,QAAI,KAAK,YAAY,EAAE,GAAG;AACxB,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB;AAEA,QAAI;AACF,YAAM,UAAU,IAAI,QAAQ;AAC5B,cAAQ,IAAI,iBAAiB,UAAU,KAAK,MAAM,EAAE;AAEpD,YAAM;AAAA,QACJ,iDAAiD,KAAK,OAAO,cAAc,EAAE;AAAA,QAC7E;AAAA,UACE,QAAQ;AAAA,UACR;AAAA,QACF;AAAA,MACF;AACA,aAAO,EAAE,SAAS,KAAK;AAAA,IACzB,SAAS,IAAI;AACX,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAAA,EAEA,MAAa,YACX,OACA;AACA,WAAO,MAAM,QAAQ;AAAA,MACnB,MAAM,IAAI,OAAO,MAAM;AACrB,cAAM,WAAW,IAAI,SAAS;AAC9B,iBAAS,OAAO,QAAQ,EAAE,IAAI;AAE9B,cAAM,cAAc,MAAM,KAAK,OAAO,EAAE,IAAI,OAAO,QAAQ;AAE3D,eAAO;AAAA,UACL,KAAK;AAAA,UACL,IAAI,EAAE,IAAI;AAAA,QACZ;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AACF;","names":[]}
|
package/dist/logger.d.ts
ADDED
package/dist/logger.js
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
1
|
+
// src/logger.ts
|
|
2
|
+
import { pino } from "pino";
|
|
3
|
+
var createLogger = (args) => {
|
|
4
|
+
const l = pino(
|
|
5
|
+
{
|
|
6
|
+
level: "info",
|
|
7
|
+
redact: [],
|
|
8
|
+
transport: args.pretty ? {
|
|
9
|
+
target: "pino-pretty"
|
|
10
|
+
} : void 0
|
|
11
|
+
},
|
|
12
|
+
args.token ? pino.transport({
|
|
13
|
+
target: "@logtail/pino",
|
|
14
|
+
options: { sourceToken: args.token }
|
|
15
|
+
}) : void 0
|
|
16
|
+
);
|
|
17
|
+
l.child({ service: args.service });
|
|
18
|
+
return l;
|
|
19
|
+
};
|
|
20
|
+
export {
|
|
21
|
+
createLogger
|
|
22
|
+
};
|
|
23
|
+
//# sourceMappingURL=logger.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/logger.ts"],"sourcesContent":["import { pino } from \"pino\";\r\n\r\nexport const createLogger = (args: {\r\n token?: string | undefined | null;\r\n pretty?: boolean;\r\n service: string;\r\n}) => {\r\n const l = pino(\r\n {\r\n level: \"info\",\r\n redact: [],\r\n transport: args.pretty\r\n ? {\r\n target: \"pino-pretty\",\r\n }\r\n : undefined,\r\n },\r\n\r\n args.token\r\n ? pino.transport({\r\n target: \"@logtail/pino\",\r\n options: { sourceToken: args.token },\r\n })\r\n : undefined,\r\n );\r\n\r\n l.child({ service: args.service });\r\n\r\n return l;\r\n};\r\n"],"mappings":";AAAA,SAAS,YAAY;AAEd,IAAM,eAAe,CAAC,SAIvB;AACJ,QAAM,IAAI;AAAA,IACR;AAAA,MACE,OAAO;AAAA,MACP,QAAQ,CAAC;AAAA,MACT,WAAW,KAAK,SACZ;AAAA,QACE,QAAQ;AAAA,MACV,IACA;AAAA,IACN;AAAA,IAEA,KAAK,QACD,KAAK,UAAU;AAAA,MACb,QAAQ;AAAA,MACR,SAAS,EAAE,aAAa,KAAK,MAAM;AAAA,IACrC,CAAC,IACD;AAAA,EACN;AAEA,IAAE,MAAM,EAAE,SAAS,KAAK,QAAQ,CAAC;AAEjC,SAAO;AACT;","names":[]}
|
package/dist/schema.d.ts
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
import * as drizzle_orm from 'drizzle-orm';
|
|
2
|
+
import * as drizzle_orm_sqlite_core from 'drizzle-orm/sqlite-core';
|
|
3
|
+
|
|
4
|
+
declare const columns: {
|
|
5
|
+
time: {
|
|
6
|
+
createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
|
|
7
|
+
updatedAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>;
|
|
8
|
+
};
|
|
9
|
+
common: {
|
|
10
|
+
createdAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"created_at">>>;
|
|
11
|
+
updatedAt: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTimestampBuilderInitial<"updated_at">>>;
|
|
12
|
+
id: drizzle_orm.HasDefault<drizzle_orm.NotNull<drizzle_orm_sqlite_core.SQLiteTextBuilderInitial<"id", [string, ...string[]]>>>;
|
|
13
|
+
};
|
|
14
|
+
};
|
|
15
|
+
|
|
16
|
+
export { columns };
|
package/dist/schema.js
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
// src/schema.ts
|
|
2
|
+
import { createId } from "@paralleldrive/cuid2";
|
|
3
|
+
import { int, text } from "drizzle-orm/sqlite-core";
|
|
4
|
+
var timeColumns = {
|
|
5
|
+
createdAt: int("created_at", { mode: "timestamp_ms" }).notNull().$defaultFn(() => /* @__PURE__ */ new Date()),
|
|
6
|
+
updatedAt: int("updated_at", { mode: "timestamp_ms" }).notNull().$defaultFn(() => /* @__PURE__ */ new Date())
|
|
7
|
+
};
|
|
8
|
+
var commonColumns = {
|
|
9
|
+
id: text("id").primaryKey().$defaultFn(() => createId()),
|
|
10
|
+
...timeColumns
|
|
11
|
+
};
|
|
12
|
+
var columns = {
|
|
13
|
+
time: timeColumns,
|
|
14
|
+
common: commonColumns
|
|
15
|
+
};
|
|
16
|
+
export {
|
|
17
|
+
columns
|
|
18
|
+
};
|
|
19
|
+
//# sourceMappingURL=schema.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/schema.ts"],"sourcesContent":["import { createId } from \"@paralleldrive/cuid2\";\r\nimport { int, text } from \"drizzle-orm/sqlite-core\";\r\n\r\nconst timeColumns = {\r\n createdAt: int(\"created_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$defaultFn(() => new Date()),\r\n updatedAt: int(\"updated_at\", { mode: \"timestamp_ms\" })\r\n .notNull()\r\n .$defaultFn(() => new Date()),\r\n};\r\n\r\nconst commonColumns = {\r\n id: text(\"id\")\r\n .primaryKey()\r\n .$defaultFn(() => createId()),\r\n ...timeColumns,\r\n};\r\n\r\nexport const columns = {\r\n time: timeColumns,\r\n common: commonColumns,\r\n};\r\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,KAAK,YAAY;AAE1B,IAAM,cAAc;AAAA,EAClB,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,WAAW,MAAM,oBAAI,KAAK,CAAC;AAAA,EAC9B,WAAW,IAAI,cAAc,EAAE,MAAM,eAAe,CAAC,EAClD,QAAQ,EACR,WAAW,MAAM,oBAAI,KAAK,CAAC;AAChC;AAEA,IAAM,gBAAgB;AAAA,EACpB,IAAI,KAAK,IAAI,EACV,WAAW,EACX,WAAW,MAAM,SAAS,CAAC;AAAA,EAC9B,GAAG;AACL;AAEO,IAAM,UAAU;AAAA,EACrB,MAAM;AAAA,EACN,QAAQ;AACV;","names":[]}
|
package/package.json
CHANGED
|
@@ -3,6 +3,9 @@
|
|
|
3
3
|
"description": "",
|
|
4
4
|
"private": false,
|
|
5
5
|
"type": "module",
|
|
6
|
+
"files": [
|
|
7
|
+
"dist/**"
|
|
8
|
+
],
|
|
6
9
|
"exports": {
|
|
7
10
|
"./db": {
|
|
8
11
|
"types": "./dist/db.d.ts",
|
|
@@ -41,7 +44,7 @@
|
|
|
41
44
|
"pino-pretty": "^10.3.1",
|
|
42
45
|
"remeda": "^2.0.10",
|
|
43
46
|
"unstorage": "^1.10.2",
|
|
44
|
-
"valibot": "0.
|
|
47
|
+
"valibot": "0.35.0"
|
|
45
48
|
},
|
|
46
49
|
"keywords": [],
|
|
47
50
|
"author": "",
|
|
@@ -51,7 +54,7 @@
|
|
|
51
54
|
"tsup": "^8.0.1",
|
|
52
55
|
"typescript": "^5.4.5"
|
|
53
56
|
},
|
|
54
|
-
"version": "0.0.
|
|
57
|
+
"version": "0.0.6",
|
|
55
58
|
"scripts": {
|
|
56
59
|
"dev": "tsup --watch",
|
|
57
60
|
"build": "tsc --noEmit && tsup",
|
package/.gitattributes
DELETED
package/biome.json
DELETED
|
@@ -1,53 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"$schema": "https://biomejs.dev/schemas/1.8.3/schema.json",
|
|
3
|
-
"organizeImports": {
|
|
4
|
-
"enabled": false
|
|
5
|
-
},
|
|
6
|
-
"linter": {
|
|
7
|
-
"enabled": true,
|
|
8
|
-
"rules": {
|
|
9
|
-
"recommended": true,
|
|
10
|
-
"correctness": {
|
|
11
|
-
"noUnusedImports": "error",
|
|
12
|
-
"noUnusedVariables": "error",
|
|
13
|
-
"noUnusedLabels": "error",
|
|
14
|
-
"useHookAtTopLevel": "error",
|
|
15
|
-
"noUndeclaredVariables": "error",
|
|
16
|
-
"useJsxKeyInIterable": "error"
|
|
17
|
-
},
|
|
18
|
-
"suspicious": {
|
|
19
|
-
"noExplicitAny": "off"
|
|
20
|
-
},
|
|
21
|
-
"a11y": {
|
|
22
|
-
"useMediaCaption": "off",
|
|
23
|
-
"useAltText": "off",
|
|
24
|
-
"useKeyWithClickEvents": "off",
|
|
25
|
-
"useKeyWithMouseEvents": "off",
|
|
26
|
-
"noSvgWithoutTitle": "off"
|
|
27
|
-
},
|
|
28
|
-
"style": {
|
|
29
|
-
"noParameterAssign": "off",
|
|
30
|
-
"noUselessElse": "off"
|
|
31
|
-
},
|
|
32
|
-
"complexity": {
|
|
33
|
-
"noForEach": "off"
|
|
34
|
-
}
|
|
35
|
-
},
|
|
36
|
-
"include": [
|
|
37
|
-
"**/*.ts",
|
|
38
|
-
"**/*.tsx",
|
|
39
|
-
"**/*.json"
|
|
40
|
-
]
|
|
41
|
-
},
|
|
42
|
-
"javascript": {
|
|
43
|
-
"formatter": {
|
|
44
|
-
"arrowParentheses": "always",
|
|
45
|
-
"lineWidth": 120,
|
|
46
|
-
"indentWidth": 2,
|
|
47
|
-
"quoteStyle": "double",
|
|
48
|
-
"trailingComma": "all",
|
|
49
|
-
"lineEnding": "crlf",
|
|
50
|
-
"indentStyle": "space"
|
|
51
|
-
}
|
|
52
|
-
}
|
|
53
|
-
}
|
package/publish.sh
DELETED
package/src/cache.ts
DELETED
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
import {
|
|
2
|
-
createStorage as createUnstorage,
|
|
3
|
-
type Storage,
|
|
4
|
-
type StorageValue,
|
|
5
|
-
} from "unstorage";
|
|
6
|
-
import redisDriver, { type RedisOptions } from "unstorage/drivers/redis";
|
|
7
|
-
|
|
8
|
-
export class Cache {
|
|
9
|
-
private _cache: Storage<StorageValue>;
|
|
10
|
-
|
|
11
|
-
constructor(args: RedisOptions) {
|
|
12
|
-
this._cache = createUnstorage({
|
|
13
|
-
driver: redisDriver(args),
|
|
14
|
-
});
|
|
15
|
-
}
|
|
16
|
-
|
|
17
|
-
get cache() {
|
|
18
|
-
return this._cache;
|
|
19
|
-
}
|
|
20
|
-
|
|
21
|
-
public ttl(date: Date) {
|
|
22
|
-
return Math.floor((date.getTime() - Date.now()) / 1000);
|
|
23
|
-
}
|
|
24
|
-
}
|
package/src/db.ts
DELETED
|
@@ -1,28 +0,0 @@
|
|
|
1
|
-
import { createClient } from "@libsql/client";
|
|
2
|
-
import { LibSQLDatabase, drizzle } from "drizzle-orm/libsql";
|
|
3
|
-
|
|
4
|
-
export const createLibsqlClient = (args: {
|
|
5
|
-
url: string;
|
|
6
|
-
authToken?: string;
|
|
7
|
-
}) => {
|
|
8
|
-
return createClient(args);
|
|
9
|
-
};
|
|
10
|
-
|
|
11
|
-
export const createDbClient = <TSchema extends Record<string, unknown>>(
|
|
12
|
-
schema: TSchema,
|
|
13
|
-
args: { url: string; authToken?: string }
|
|
14
|
-
) => {
|
|
15
|
-
const client = createLibsqlClient(args);
|
|
16
|
-
const db = drizzle(client, {
|
|
17
|
-
schema,
|
|
18
|
-
logger: false,
|
|
19
|
-
});
|
|
20
|
-
|
|
21
|
-
return db;
|
|
22
|
-
};
|
|
23
|
-
|
|
24
|
-
export type DatabaseClient<TSchema extends Record<string, unknown>> =
|
|
25
|
-
LibSQLDatabase<TSchema>;
|
|
26
|
-
export type DatabaseClientTransactionContext<
|
|
27
|
-
TSchema extends Record<string, unknown>
|
|
28
|
-
> = Parameters<Parameters<DatabaseClient<TSchema>["transaction"]>[0]>[0];
|
package/src/env.ts
DELETED
|
@@ -1,52 +0,0 @@
|
|
|
1
|
-
import { pipe } from "remeda";
|
|
2
|
-
import type { BaseSchema, Output } from "valibot";
|
|
3
|
-
import * as v from "valibot";
|
|
4
|
-
|
|
5
|
-
const PUBLIC_ENV_PREFIX = "PUBLIC_" as const;
|
|
6
|
-
|
|
7
|
-
/**
|
|
8
|
-
* Validates your environment variables against the given Valibot schema;
|
|
9
|
-
* @param args
|
|
10
|
-
* @returns An object containing client environment variables and another containing server environment variables
|
|
11
|
-
*/
|
|
12
|
-
export const createEnv = <
|
|
13
|
-
Schema extends Record<string, BaseSchema>,
|
|
14
|
-
Env = {
|
|
15
|
-
[K in keyof Schema]: Output<Schema[K]>;
|
|
16
|
-
},
|
|
17
|
-
>(args: {
|
|
18
|
-
schema: Schema;
|
|
19
|
-
env: any;
|
|
20
|
-
}) => {
|
|
21
|
-
const pairs = Object.entries(args.schema);
|
|
22
|
-
const serverEnv = new Map();
|
|
23
|
-
|
|
24
|
-
for (const [key, value] of pairs) {
|
|
25
|
-
const result = v.safeParse(value, args.env[key] ?? null);
|
|
26
|
-
|
|
27
|
-
if (!result.success) {
|
|
28
|
-
console.error(`Environment variable "${key}" is invalid`);
|
|
29
|
-
process.exit(1);
|
|
30
|
-
}
|
|
31
|
-
|
|
32
|
-
serverEnv.set(key, result.output);
|
|
33
|
-
}
|
|
34
|
-
|
|
35
|
-
type ClientEnvKeys = Exclude<
|
|
36
|
-
{ [K in keyof Env]: K extends `${typeof PUBLIC_ENV_PREFIX}${string}` ? K : never }[keyof Env],
|
|
37
|
-
undefined
|
|
38
|
-
>;
|
|
39
|
-
|
|
40
|
-
type ClientEnv = {
|
|
41
|
-
[B in ClientEnvKeys]: Env[B];
|
|
42
|
-
};
|
|
43
|
-
|
|
44
|
-
const clientEnv = pipe(
|
|
45
|
-
serverEnv,
|
|
46
|
-
(obj) => Array.from(obj.entries()),
|
|
47
|
-
(pairs) => pairs.filter(([k]) => k.startsWith(PUBLIC_ENV_PREFIX)),
|
|
48
|
-
(pairs) => Object.fromEntries(pairs),
|
|
49
|
-
) as ClientEnv;
|
|
50
|
-
|
|
51
|
-
return { client: clientEnv, server: Object.fromEntries(serverEnv.entries()) as Env };
|
|
52
|
-
};
|
package/src/logger.ts
DELETED
|
@@ -1,30 +0,0 @@
|
|
|
1
|
-
import { pino } from "pino";
|
|
2
|
-
|
|
3
|
-
export const createLogger = (args: {
|
|
4
|
-
token?: string | undefined | null;
|
|
5
|
-
pretty?: boolean;
|
|
6
|
-
service: string;
|
|
7
|
-
}) => {
|
|
8
|
-
const l = pino(
|
|
9
|
-
{
|
|
10
|
-
level: "info",
|
|
11
|
-
redact: [],
|
|
12
|
-
transport: args.pretty
|
|
13
|
-
? {
|
|
14
|
-
target: "pino-pretty",
|
|
15
|
-
}
|
|
16
|
-
: undefined,
|
|
17
|
-
},
|
|
18
|
-
|
|
19
|
-
args.token
|
|
20
|
-
? pino.transport({
|
|
21
|
-
target: "@logtail/pino",
|
|
22
|
-
options: { sourceToken: args.token },
|
|
23
|
-
})
|
|
24
|
-
: undefined,
|
|
25
|
-
);
|
|
26
|
-
|
|
27
|
-
l.child({ service: args.service });
|
|
28
|
-
|
|
29
|
-
return l;
|
|
30
|
-
};
|
package/src/migrate.ts
DELETED
|
@@ -1,41 +0,0 @@
|
|
|
1
|
-
import { createClient } from "@libsql/client";
|
|
2
|
-
import { drizzle } from "drizzle-orm/libsql";
|
|
3
|
-
import { migrate as runDrizzleMigrate } from "drizzle-orm/libsql/migrator";
|
|
4
|
-
|
|
5
|
-
export const migrate = async <TSchema extends Record<string, any>>(
|
|
6
|
-
schema: TSchema,
|
|
7
|
-
args: {
|
|
8
|
-
url: string;
|
|
9
|
-
token?: string;
|
|
10
|
-
migrationsFolder: string;
|
|
11
|
-
}
|
|
12
|
-
) => {
|
|
13
|
-
let url = args.url;
|
|
14
|
-
|
|
15
|
-
// Migrations are only supported via the libsql protocol
|
|
16
|
-
if (url.startsWith("http")) {
|
|
17
|
-
url = url.replace(/http(s)?/, "libsql");
|
|
18
|
-
}
|
|
19
|
-
|
|
20
|
-
const db = drizzle(
|
|
21
|
-
createClient(
|
|
22
|
-
// Auth token must be either 1) present and not undefined or 2) not present
|
|
23
|
-
args.token
|
|
24
|
-
? {
|
|
25
|
-
url,
|
|
26
|
-
authToken: args.token,
|
|
27
|
-
}
|
|
28
|
-
: { url }
|
|
29
|
-
),
|
|
30
|
-
{ schema }
|
|
31
|
-
);
|
|
32
|
-
|
|
33
|
-
console.info("Running migrations");
|
|
34
|
-
|
|
35
|
-
await runDrizzleMigrate(db, {
|
|
36
|
-
migrationsFolder: args.migrationsFolder,
|
|
37
|
-
});
|
|
38
|
-
|
|
39
|
-
console.info("Migrations applied");
|
|
40
|
-
process.exit(0);
|
|
41
|
-
};
|
package/src/schema.ts
DELETED
|
@@ -1,23 +0,0 @@
|
|
|
1
|
-
import { createId } from "@paralleldrive/cuid2";
|
|
2
|
-
import { int, text } from "drizzle-orm/sqlite-core";
|
|
3
|
-
|
|
4
|
-
const timeColumns = {
|
|
5
|
-
createdAt: int("created_at", { mode: "timestamp_ms" })
|
|
6
|
-
.notNull()
|
|
7
|
-
.$defaultFn(() => new Date()),
|
|
8
|
-
updatedAt: int("updated_at", { mode: "timestamp_ms" })
|
|
9
|
-
.notNull()
|
|
10
|
-
.$defaultFn(() => new Date()),
|
|
11
|
-
};
|
|
12
|
-
|
|
13
|
-
const commonColumns = {
|
|
14
|
-
id: text("id")
|
|
15
|
-
.primaryKey()
|
|
16
|
-
.$defaultFn(() => createId()),
|
|
17
|
-
...timeColumns,
|
|
18
|
-
};
|
|
19
|
-
|
|
20
|
-
export const columns = {
|
|
21
|
-
time: timeColumns,
|
|
22
|
-
common: commonColumns,
|
|
23
|
-
};
|
package/src/utils.ts
DELETED
package/tsconfig.json
DELETED
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"compilerOptions": {
|
|
3
|
-
"esModuleInterop": true,
|
|
4
|
-
"skipLibCheck": true,
|
|
5
|
-
"target": "es2022",
|
|
6
|
-
"verbatimModuleSyntax": true,
|
|
7
|
-
"allowJs": true,
|
|
8
|
-
"resolveJsonModule": true,
|
|
9
|
-
"moduleDetection": "force",
|
|
10
|
-
"strict": true,
|
|
11
|
-
"noUncheckedIndexedAccess": true,
|
|
12
|
-
"moduleResolution": "Bundler",
|
|
13
|
-
"module": "ESNext",
|
|
14
|
-
"noEmit": true,
|
|
15
|
-
"lib": ["DOM", "DOM.Iterable", "ES2022"],
|
|
16
|
-
"baseUrl": "."
|
|
17
|
-
},
|
|
18
|
-
"include": ["src/**/*", "package.json"]
|
|
19
|
-
}
|
package/tsup.config.ts
DELETED
|
@@ -1,15 +0,0 @@
|
|
|
1
|
-
import { defineConfig } from "tsup";
|
|
2
|
-
|
|
3
|
-
export default defineConfig({
|
|
4
|
-
format: ["esm"],
|
|
5
|
-
sourcemap: true,
|
|
6
|
-
dts: true,
|
|
7
|
-
entry: {
|
|
8
|
-
schema: "src/schema.ts",
|
|
9
|
-
db: "src/db.ts",
|
|
10
|
-
cache: "src/cache.ts",
|
|
11
|
-
logger: "src/logger.ts",
|
|
12
|
-
env: "src/env.ts",
|
|
13
|
-
images: "src/images.ts",
|
|
14
|
-
},
|
|
15
|
-
});
|