@hatk/hatk 0.0.1-alpha.4 → 0.0.1-alpha.41
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapter.d.ts +19 -0
- package/dist/adapter.d.ts.map +1 -0
- package/dist/adapter.js +107 -0
- package/dist/backfill.d.ts +60 -1
- package/dist/backfill.d.ts.map +1 -1
- package/dist/backfill.js +167 -33
- package/dist/car.d.ts +59 -1
- package/dist/car.d.ts.map +1 -1
- package/dist/car.js +179 -7
- package/dist/cbor.d.ts +37 -0
- package/dist/cbor.d.ts.map +1 -1
- package/dist/cbor.js +36 -3
- package/dist/cid.d.ts +37 -0
- package/dist/cid.d.ts.map +1 -1
- package/dist/cid.js +38 -3
- package/dist/cli.js +417 -133
- package/dist/cloudflare/container.d.ts +73 -0
- package/dist/cloudflare/container.d.ts.map +1 -0
- package/dist/cloudflare/container.js +232 -0
- package/dist/cloudflare/hooks.d.ts +33 -0
- package/dist/cloudflare/hooks.d.ts.map +1 -0
- package/dist/cloudflare/hooks.js +40 -0
- package/dist/cloudflare/init.d.ts +27 -0
- package/dist/cloudflare/init.d.ts.map +1 -0
- package/dist/cloudflare/init.js +103 -0
- package/dist/cloudflare/worker.d.ts +27 -0
- package/dist/cloudflare/worker.d.ts.map +1 -0
- package/dist/cloudflare/worker.js +54 -0
- package/dist/config.d.ts +12 -1
- package/dist/config.d.ts.map +1 -1
- package/dist/config.js +36 -9
- package/dist/database/adapter-factory.d.ts +6 -0
- package/dist/database/adapter-factory.d.ts.map +1 -0
- package/dist/database/adapter-factory.js +20 -0
- package/dist/database/adapters/d1.d.ts +56 -0
- package/dist/database/adapters/d1.d.ts.map +1 -0
- package/dist/database/adapters/d1.js +108 -0
- package/dist/database/adapters/duckdb-search.d.ts +12 -0
- package/dist/database/adapters/duckdb-search.d.ts.map +1 -0
- package/dist/database/adapters/duckdb-search.js +27 -0
- package/dist/database/adapters/duckdb.d.ts +25 -0
- package/dist/database/adapters/duckdb.d.ts.map +1 -0
- package/dist/database/adapters/duckdb.js +161 -0
- package/dist/database/adapters/sqlite-search.d.ts +23 -0
- package/dist/database/adapters/sqlite-search.d.ts.map +1 -0
- package/dist/database/adapters/sqlite-search.js +74 -0
- package/dist/database/adapters/sqlite.d.ts +18 -0
- package/dist/database/adapters/sqlite.d.ts.map +1 -0
- package/dist/database/adapters/sqlite.js +87 -0
- package/dist/database/db.d.ts +159 -0
- package/dist/database/db.d.ts.map +1 -0
- package/dist/database/db.js +1445 -0
- package/dist/database/dialect.d.ts +45 -0
- package/dist/database/dialect.d.ts.map +1 -0
- package/dist/database/dialect.js +72 -0
- package/dist/database/fts.d.ts +27 -0
- package/dist/database/fts.d.ts.map +1 -0
- package/dist/database/fts.js +846 -0
- package/dist/database/index.d.ts +7 -0
- package/dist/database/index.d.ts.map +1 -0
- package/dist/database/index.js +6 -0
- package/dist/database/ports.d.ts +50 -0
- package/dist/database/ports.d.ts.map +1 -0
- package/dist/database/ports.js +1 -0
- package/dist/database/schema.d.ts +61 -0
- package/dist/database/schema.d.ts.map +1 -0
- package/dist/database/schema.js +394 -0
- package/dist/db.d.ts +1 -1
- package/dist/db.d.ts.map +1 -1
- package/dist/db.js +4 -38
- package/dist/dev-entry.d.ts +8 -0
- package/dist/dev-entry.d.ts.map +1 -0
- package/dist/dev-entry.js +111 -0
- package/dist/feeds.d.ts +12 -8
- package/dist/feeds.d.ts.map +1 -1
- package/dist/feeds.js +45 -6
- package/dist/fts.d.ts.map +1 -1
- package/dist/fts.js +5 -0
- package/dist/hooks.d.ts +43 -0
- package/dist/hooks.d.ts.map +1 -0
- package/dist/hooks.js +102 -0
- package/dist/hydrate.d.ts +6 -5
- package/dist/hydrate.d.ts.map +1 -1
- package/dist/hydrate.js +4 -16
- package/dist/indexer.d.ts +22 -0
- package/dist/indexer.d.ts.map +1 -1
- package/dist/indexer.js +70 -7
- package/dist/labels.d.ts +34 -0
- package/dist/labels.d.ts.map +1 -1
- package/dist/labels.js +66 -6
- package/dist/logger.d.ts +29 -0
- package/dist/logger.d.ts.map +1 -1
- package/dist/logger.js +29 -0
- package/dist/main.js +135 -67
- package/dist/mst.d.ts +18 -1
- package/dist/mst.d.ts.map +1 -1
- package/dist/mst.js +19 -8
- package/dist/oauth/db.d.ts.map +1 -1
- package/dist/oauth/db.js +43 -17
- package/dist/oauth/server.d.ts +2 -0
- package/dist/oauth/server.d.ts.map +1 -1
- package/dist/oauth/server.js +103 -8
- package/dist/oauth/session.d.ts +11 -0
- package/dist/oauth/session.d.ts.map +1 -0
- package/dist/oauth/session.js +65 -0
- package/dist/opengraph.d.ts +10 -0
- package/dist/opengraph.d.ts.map +1 -1
- package/dist/opengraph.js +73 -39
- package/dist/pds-proxy.d.ts +42 -0
- package/dist/pds-proxy.d.ts.map +1 -0
- package/dist/pds-proxy.js +189 -0
- package/dist/renderer.d.ts +27 -0
- package/dist/renderer.d.ts.map +1 -0
- package/dist/renderer.js +46 -0
- package/dist/resolve-hatk.d.ts +6 -0
- package/dist/resolve-hatk.d.ts.map +1 -0
- package/dist/resolve-hatk.js +20 -0
- package/dist/response.d.ts +16 -0
- package/dist/response.d.ts.map +1 -0
- package/dist/response.js +69 -0
- package/dist/scanner.d.ts +21 -0
- package/dist/scanner.d.ts.map +1 -0
- package/dist/scanner.js +88 -0
- package/dist/schema.d.ts +8 -0
- package/dist/schema.d.ts.map +1 -1
- package/dist/schema.js +29 -0
- package/dist/seed.d.ts +19 -0
- package/dist/seed.d.ts.map +1 -1
- package/dist/seed.js +43 -4
- package/dist/server-init.d.ts +8 -0
- package/dist/server-init.d.ts.map +1 -0
- package/dist/server-init.js +61 -0
- package/dist/server.d.ts +26 -3
- package/dist/server.d.ts.map +1 -1
- package/dist/server.js +528 -635
- package/dist/setup.d.ts +28 -1
- package/dist/setup.d.ts.map +1 -1
- package/dist/setup.js +50 -3
- package/dist/test.d.ts +1 -1
- package/dist/test.d.ts.map +1 -1
- package/dist/test.js +38 -32
- package/dist/views.js +1 -1
- package/dist/vite-plugin.d.ts +1 -1
- package/dist/vite-plugin.d.ts.map +1 -1
- package/dist/vite-plugin.js +254 -66
- package/dist/xrpc.d.ts +60 -10
- package/dist/xrpc.d.ts.map +1 -1
- package/dist/xrpc.js +155 -39
- package/package.json +13 -6
- package/public/admin.html +0 -54
package/dist/main.js
CHANGED
|
@@ -1,28 +1,49 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
|
|
2
|
+
var __rewriteRelativeImportExtension = (this && this.__rewriteRelativeImportExtension) || function (path, preserveJsx) {
|
|
3
|
+
if (typeof path === "string" && /^\.\.?\//.test(path)) {
|
|
4
|
+
return path.replace(/\.(tsx)$|((?:\.d)?)((?:\.[^./]+?)?)\.([cm]?)ts$/i, function (m, tsx, d, ext, cm) {
|
|
5
|
+
return tsx ? preserveJsx ? ".jsx" : ".js" : d && (!ext || !cm) ? m : (d + ext + "." + cm.toLowerCase() + "js");
|
|
6
|
+
});
|
|
7
|
+
}
|
|
8
|
+
return path;
|
|
9
|
+
};
|
|
10
|
+
import { mkdirSync, writeFileSync, existsSync } from 'node:fs';
|
|
3
11
|
import { dirname, resolve } from 'node:path';
|
|
12
|
+
import { registerHatkResolveHook } from "./resolve-hatk.js";
|
|
4
13
|
import { log } from "./logger.js";
|
|
5
14
|
import { loadConfig } from "./config.js";
|
|
6
|
-
import { loadLexicons, storeLexicons, discoverCollections,
|
|
15
|
+
import { loadLexicons, storeLexicons, discoverCollections, buildSchemas } from "./database/schema.js";
|
|
7
16
|
import { discoverViews } from "./views.js";
|
|
8
|
-
import { initDatabase, getCursor, querySQL,
|
|
17
|
+
import { initDatabase, getCursor, querySQL, getSqlDialect, getSchemaDump, migrateSchema } from "./database/db.js";
|
|
18
|
+
import { createAdapter } from "./database/adapter-factory.js";
|
|
19
|
+
import { getDialect } from "./database/dialect.js";
|
|
20
|
+
import { setSearchPort } from "./database/fts.js";
|
|
9
21
|
import { initFeeds, listFeeds } from "./feeds.js";
|
|
10
|
-
import { initXrpc, listXrpc, configureRelay } from "./xrpc.js";
|
|
22
|
+
import { initXrpc, listXrpc, configureRelay, configureOAuth, callXrpc } from "./xrpc.js";
|
|
11
23
|
import { initOpengraph } from "./opengraph.js";
|
|
12
24
|
import { initLabels, getLabelDefinitions } from "./labels.js";
|
|
13
25
|
import { startIndexer } from "./indexer.js";
|
|
14
|
-
import { rebuildAllIndexes } from "./fts.js";
|
|
15
|
-
import {
|
|
26
|
+
import { rebuildAllIndexes } from "./database/fts.js";
|
|
27
|
+
import { createHandler, registerCoreHandlers } from "./server.js";
|
|
28
|
+
import { serve } from "./adapter.js";
|
|
16
29
|
import { validateLexicons } from '@bigmoves/lexicon';
|
|
17
30
|
import { relayHttpUrl } from "./config.js";
|
|
18
31
|
import { runBackfill } from "./backfill.js";
|
|
19
32
|
import { initOAuth } from "./oauth/server.js";
|
|
20
|
-
import {
|
|
33
|
+
import { parseSessionCookie, getSessionCookieName } from "./oauth/session.js";
|
|
34
|
+
import { loadOnLoginHook } from "./hooks.js";
|
|
21
35
|
import { initSetup } from "./setup.js";
|
|
22
|
-
|
|
36
|
+
import { initServer } from "./server-init.js";
|
|
37
|
+
function logMemory(phase) {
|
|
38
|
+
const mem = process.memoryUsage();
|
|
39
|
+
log(`[mem] ${phase}: heap=${Math.round(mem.heapUsed / 1024 / 1024)}MB rss=${Math.round(mem.rss / 1024 / 1024)}MB external=${Math.round(mem.external / 1024 / 1024)}MB arrayBuffers=${Math.round(mem.arrayBuffers / 1024 / 1024)}MB`);
|
|
40
|
+
}
|
|
41
|
+
const configPath = process.argv[2] || 'hatk.config.ts';
|
|
23
42
|
const configDir = dirname(resolve(configPath));
|
|
43
|
+
registerHatkResolveHook();
|
|
44
|
+
logMemory('startup');
|
|
24
45
|
// 1. Load config
|
|
25
|
-
const config = loadConfig(configPath);
|
|
46
|
+
const config = await loadConfig(configPath);
|
|
26
47
|
configureRelay(config.relay);
|
|
27
48
|
// 2. Load lexicons, validate schemas, and discover collections
|
|
28
49
|
const lexicons = loadLexicons(resolve(configDir, 'lexicons'));
|
|
@@ -44,44 +65,66 @@ if (collections.length === 0) {
|
|
|
44
65
|
log(`[main] Loaded config: ${collections.length} collections`);
|
|
45
66
|
// Discover view defs from lexicons
|
|
46
67
|
discoverViews();
|
|
47
|
-
|
|
48
|
-
const schemas =
|
|
49
|
-
const
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
uri TEXT PRIMARY KEY,
|
|
56
|
-
cid TEXT,
|
|
57
|
-
did TEXT NOT NULL,
|
|
58
|
-
indexed_at TIMESTAMP NOT NULL,
|
|
59
|
-
data JSON
|
|
60
|
-
);
|
|
61
|
-
CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_indexed ON "${nsid}"(indexed_at DESC);
|
|
62
|
-
CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_author ON "${nsid}"(did);`;
|
|
63
|
-
schemas.push({ collection: nsid, tableName: `"${nsid}"`, columns: [], refColumns: [], children: [], unions: [] });
|
|
64
|
-
ddlStatements.push(genericDDL);
|
|
65
|
-
continue;
|
|
68
|
+
const engineDialect = getDialect(config.databaseEngine);
|
|
69
|
+
const { schemas, ddlStatements } = buildSchemas(lexicons, collections, engineDialect);
|
|
70
|
+
for (const s of schemas) {
|
|
71
|
+
if (s.columns.length === 0) {
|
|
72
|
+
log(`[main] No lexicon found for ${s.collection}, using generic JSON storage`);
|
|
73
|
+
}
|
|
74
|
+
else {
|
|
75
|
+
log(`[main] Schema for ${s.collection}: ${s.columns.length} columns, ${s.unions.length} unions`);
|
|
66
76
|
}
|
|
67
|
-
const schema = generateTableSchema(nsid, lexicon, lexicons);
|
|
68
|
-
schemas.push(schema);
|
|
69
|
-
ddlStatements.push(generateCreateTableSQL(schema));
|
|
70
|
-
log(`[main] Schema for ${nsid}: ${schema.columns.length} columns, ${schema.unions.length} unions`);
|
|
71
77
|
}
|
|
72
|
-
// 3. Ensure data directory exists and initialize
|
|
78
|
+
// 3. Ensure data directory exists and initialize database
|
|
73
79
|
if (config.database !== ':memory:') {
|
|
74
80
|
mkdirSync(dirname(config.database), { recursive: true });
|
|
75
81
|
}
|
|
76
|
-
await
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
+
const { adapter, searchPort } = await createAdapter(config.databaseEngine);
|
|
83
|
+
setSearchPort(searchPort);
|
|
84
|
+
await initDatabase(adapter, config.database, schemas, ddlStatements);
|
|
85
|
+
logMemory('after-db-init');
|
|
86
|
+
log(`[main] Database initialized (${config.databaseEngine}, ${config.database === ':memory:' ? 'in-memory' : config.database})`);
|
|
87
|
+
// Auto-migrate schema if lexicons changed
|
|
88
|
+
const migrationChanges = await migrateSchema(schemas);
|
|
89
|
+
if (migrationChanges.length > 0) {
|
|
90
|
+
log(`[main] Applied ${migrationChanges.length} schema migration(s)`);
|
|
91
|
+
}
|
|
92
|
+
// 3b. Run setup hooks, feeds, xrpc, og, labels
|
|
93
|
+
const serverDir = resolve(configDir, 'server');
|
|
94
|
+
if (existsSync(serverDir)) {
|
|
95
|
+
// New: single server/ directory
|
|
96
|
+
await initServer(serverDir);
|
|
97
|
+
}
|
|
98
|
+
else {
|
|
99
|
+
// Legacy: separate directories
|
|
100
|
+
await initSetup(resolve(configDir, 'setup'));
|
|
101
|
+
await loadOnLoginHook(resolve(configDir, 'hooks'));
|
|
102
|
+
await initFeeds(resolve(configDir, 'feeds'));
|
|
103
|
+
log(`[main] Feeds initialized: ${listFeeds()
|
|
104
|
+
.map((f) => f.name)
|
|
105
|
+
.join(', ') || 'none'}`);
|
|
106
|
+
await initXrpc(resolve(configDir, 'xrpc'));
|
|
107
|
+
log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
|
|
108
|
+
await initOpengraph(resolve(configDir, 'og'));
|
|
109
|
+
log(`[main] OpenGraph initialized`);
|
|
110
|
+
await initLabels(resolve(configDir, 'labels'));
|
|
111
|
+
log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
|
|
112
|
+
}
|
|
113
|
+
// Register built-in dev.hatk.* handlers so callXrpc() can find them
|
|
114
|
+
registerCoreHandlers(collections, config.oauth);
|
|
115
|
+
configureOAuth(config.oauth);
|
|
116
|
+
// Write db/schema.sql (after setup, so setup-created tables are included)
|
|
117
|
+
try {
|
|
118
|
+
const schemaDir = resolve(configDir, 'db');
|
|
119
|
+
mkdirSync(schemaDir, { recursive: true });
|
|
120
|
+
const schemaDump = await getSchemaDump();
|
|
121
|
+
writeFileSync(resolve(schemaDir, 'schema.sql'), `-- This file is auto-generated by hatk on startup. Do not edit.\n-- Database engine: ${config.databaseEngine}\n\n${schemaDump}\n`);
|
|
122
|
+
log(`[main] Schema written to db/schema.sql`);
|
|
123
|
+
}
|
|
124
|
+
catch { }
|
|
82
125
|
// Detect orphaned tables
|
|
83
126
|
try {
|
|
84
|
-
const existingTables = await querySQL(
|
|
127
|
+
const existingTables = (await querySQL(getSqlDialect().listTablesQuery));
|
|
85
128
|
for (const row of existingTables) {
|
|
86
129
|
const tableName = row.table_name;
|
|
87
130
|
const isChildTable = collections.some((c) => tableName.startsWith(c + '__'));
|
|
@@ -91,24 +134,57 @@ try {
|
|
|
91
134
|
}
|
|
92
135
|
}
|
|
93
136
|
catch { }
|
|
94
|
-
// 4. Initialize feeds, xrpc handlers, og, labels from directories
|
|
95
|
-
await initFeeds(resolve(configDir, 'feeds'));
|
|
96
|
-
log(`[main] Feeds initialized: ${listFeeds()
|
|
97
|
-
.map((f) => f.name)
|
|
98
|
-
.join(', ') || 'none'}`);
|
|
99
|
-
await initXrpc(resolve(configDir, 'xrpc'));
|
|
100
|
-
log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
|
|
101
|
-
await initOpengraph(resolve(configDir, 'og'));
|
|
102
|
-
log(`[main] OpenGraph initialized`);
|
|
103
|
-
await initLabels(resolve(configDir, 'labels'));
|
|
104
|
-
log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
|
|
105
137
|
if (config.oauth) {
|
|
106
138
|
await initOAuth(config.oauth, config.plc, config.relay);
|
|
107
139
|
log(`[main] OAuth initialized (issuer: ${config.oauth.issuer})`);
|
|
108
140
|
}
|
|
141
|
+
logMemory('before-server');
|
|
109
142
|
// 5. Start server immediately (don't wait for backfill)
|
|
110
143
|
const collectionSet = new Set(collections);
|
|
111
|
-
|
|
144
|
+
const backfillOpts = {
|
|
145
|
+
pdsUrl: relayHttpUrl(config.relay),
|
|
146
|
+
plcUrl: config.plc,
|
|
147
|
+
collections: collectionSet,
|
|
148
|
+
config: config.backfill,
|
|
149
|
+
};
|
|
150
|
+
function runBackfillAndRestart() {
|
|
151
|
+
runBackfill(backfillOpts)
|
|
152
|
+
.then(async (recordCount) => {
|
|
153
|
+
log('[main] Backfill complete, building FTS indexes...');
|
|
154
|
+
await rebuildAllIndexes(collections);
|
|
155
|
+
log('[main] FTS indexes ready');
|
|
156
|
+
return recordCount;
|
|
157
|
+
})
|
|
158
|
+
.then((recordCount) => {
|
|
159
|
+
if (recordCount > 0 && !process.env.DEV_MODE) {
|
|
160
|
+
logMemory('after-backfill');
|
|
161
|
+
log('[main] Restarting to reclaim memory...');
|
|
162
|
+
process.exit(1);
|
|
163
|
+
}
|
|
164
|
+
})
|
|
165
|
+
.catch((err) => {
|
|
166
|
+
console.error('[main] Backfill error:', err.message);
|
|
167
|
+
});
|
|
168
|
+
}
|
|
169
|
+
const handler = createHandler({
|
|
170
|
+
collections,
|
|
171
|
+
publicDir: config.publicDir,
|
|
172
|
+
oauth: config.oauth,
|
|
173
|
+
admins: config.admins,
|
|
174
|
+
onResync: runBackfillAndRestart,
|
|
175
|
+
});
|
|
176
|
+
globalThis.__hatk_callXrpc = callXrpc;
|
|
177
|
+
globalThis.__hatk_parseSessionCookie = parseSessionCookie;
|
|
178
|
+
globalThis.__hatk_sessionCookieName = getSessionCookieName();
|
|
179
|
+
// Detect SvelteKit build output and use it as fallback handler
|
|
180
|
+
let fallback = undefined;
|
|
181
|
+
const sveltekitHandler = resolve(configDir, 'build', 'handler.js');
|
|
182
|
+
if (existsSync(sveltekitHandler)) {
|
|
183
|
+
const sk = await import(__rewriteRelativeImportExtension(/* @vite-ignore */ sveltekitHandler));
|
|
184
|
+
fallback = sk.handler;
|
|
185
|
+
log(`[main] SvelteKit handler loaded from build/handler.js`);
|
|
186
|
+
}
|
|
187
|
+
serve(handler, config.port, undefined, fallback);
|
|
112
188
|
log(`\nhatk running:`);
|
|
113
189
|
log(` Relay: ${config.relay}`);
|
|
114
190
|
log(` Database: ${config.database}`);
|
|
@@ -117,6 +193,7 @@ log(` Collections: ${collections.join(', ')}`);
|
|
|
117
193
|
log(` Feeds: ${listFeeds()
|
|
118
194
|
.map((f) => f.name)
|
|
119
195
|
.join(', ')}`);
|
|
196
|
+
logMemory('after-server');
|
|
120
197
|
// 6. Start indexer with cursor
|
|
121
198
|
const cursor = await getCursor('relay');
|
|
122
199
|
startIndexer({
|
|
@@ -127,22 +204,13 @@ startIndexer({
|
|
|
127
204
|
cursor,
|
|
128
205
|
fetchTimeout: config.backfill.fetchTimeout,
|
|
129
206
|
maxRetries: config.backfill.maxRetries,
|
|
207
|
+
parallelism: config.backfill.parallelism,
|
|
130
208
|
ftsRebuildInterval: config.ftsRebuildInterval,
|
|
131
209
|
});
|
|
132
210
|
// 7. Run backfill in background
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
136
|
-
|
|
137
|
-
|
|
138
|
-
})
|
|
139
|
-
.then(() => {
|
|
140
|
-
log('[main] Backfill complete, rebuilding FTS indexes...');
|
|
141
|
-
return rebuildAllIndexes(collections);
|
|
142
|
-
})
|
|
143
|
-
.then(() => {
|
|
144
|
-
log('[main] FTS indexes ready');
|
|
145
|
-
})
|
|
146
|
-
.catch((err) => {
|
|
147
|
-
console.error('[main] Backfill error:', err.message);
|
|
211
|
+
runBackfillAndRestart();
|
|
212
|
+
// Graceful shutdown
|
|
213
|
+
process.on('SIGTERM', () => {
|
|
214
|
+
log('[main] Received SIGTERM, shutting down...');
|
|
215
|
+
process.exit(0);
|
|
148
216
|
});
|
package/dist/mst.d.ts
CHANGED
|
@@ -1,6 +1,23 @@
|
|
|
1
|
+
/** A single entry from a Merkle Search Tree — a record path paired with its content CID. */
|
|
1
2
|
export interface MstEntry {
|
|
3
|
+
/** Record path, e.g. "xyz.marketplace.listing/3mfniulnr7c2g" */
|
|
2
4
|
path: string;
|
|
5
|
+
/** CID of the record's CBOR block */
|
|
3
6
|
cid: string;
|
|
4
7
|
}
|
|
5
|
-
|
|
8
|
+
/**
|
|
9
|
+
* Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
|
|
10
|
+
*
|
|
11
|
+
* The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
|
|
12
|
+
* record paths to CIDs. Each node contains a left subtree pointer, an array of entries
|
|
13
|
+
* (each with a prefix length, key suffix, value CID, and right subtree pointer), and
|
|
14
|
+
* keys are reconstructed by combining the prefix of the previous key with the suffix.
|
|
15
|
+
*
|
|
16
|
+
* @param blocks - Block store that resolves CIDs to raw CBOR bytes
|
|
17
|
+
* @param rootCid - CID of the MST root node
|
|
18
|
+
* @yields {MstEntry} Record entries in lexicographic key order
|
|
19
|
+
*/
|
|
20
|
+
export declare function walkMst(blocks: {
|
|
21
|
+
get(cid: string): Uint8Array | undefined;
|
|
22
|
+
}, rootCid: string): Generator<MstEntry>;
|
|
6
23
|
//# sourceMappingURL=mst.d.ts.map
|
package/dist/mst.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,MAAM,WAAW,QAAQ;IACvB,IAAI,EAAE,MAAM,CAAA;IACZ,GAAG,EAAE,MAAM,CAAA;CACZ;AAED,
|
|
1
|
+
{"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,4FAA4F;AAC5F,MAAM,WAAW,QAAQ;IACvB,gEAAgE;IAChE,IAAI,EAAE,MAAM,CAAA;IACZ,qCAAqC;IACrC,GAAG,EAAE,MAAM,CAAA;CACZ;AAED;;;;;;;;;;;GAWG;AACH,wBAAiB,OAAO,CAAC,MAAM,EAAE;IAAE,GAAG,CAAC,GAAG,EAAE,MAAM,GAAG,UAAU,GAAG,SAAS,CAAA;CAAE,EAAE,OAAO,EAAE,MAAM,GAAG,SAAS,CAAC,QAAQ,CAAC,CA+BnH"}
|
package/dist/mst.js
CHANGED
|
@@ -1,14 +1,26 @@
|
|
|
1
1
|
import { cborDecode } from "./cbor.js";
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
2
|
+
/**
|
|
3
|
+
* Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
|
|
4
|
+
*
|
|
5
|
+
* The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
|
|
6
|
+
* record paths to CIDs. Each node contains a left subtree pointer, an array of entries
|
|
7
|
+
* (each with a prefix length, key suffix, value CID, and right subtree pointer), and
|
|
8
|
+
* keys are reconstructed by combining the prefix of the previous key with the suffix.
|
|
9
|
+
*
|
|
10
|
+
* @param blocks - Block store that resolves CIDs to raw CBOR bytes
|
|
11
|
+
* @param rootCid - CID of the MST root node
|
|
12
|
+
* @yields {MstEntry} Record entries in lexicographic key order
|
|
13
|
+
*/
|
|
14
|
+
export function* walkMst(blocks, rootCid) {
|
|
15
|
+
/** Recursively visit an MST node, reconstructing full keys from prefix-compressed entries. */
|
|
16
|
+
function* visit(cid, prefix) {
|
|
5
17
|
const data = blocks.get(cid);
|
|
6
18
|
if (!data)
|
|
7
19
|
return prefix;
|
|
8
20
|
const { value: node } = cborDecode(data);
|
|
9
21
|
// Visit left subtree
|
|
10
22
|
if (node.l?.$link)
|
|
11
|
-
visit(node.l.$link, prefix);
|
|
23
|
+
yield* visit(node.l.$link, prefix);
|
|
12
24
|
let lastKey = prefix;
|
|
13
25
|
for (const entry of node.e || []) {
|
|
14
26
|
const keySuffix = entry.k instanceof Uint8Array ? new TextDecoder().decode(entry.k) : entry.k;
|
|
@@ -16,15 +28,14 @@ export function walkMst(blocks, rootCid) {
|
|
|
16
28
|
const fullKey = lastKey.substring(0, prefixLen) + keySuffix;
|
|
17
29
|
lastKey = fullKey;
|
|
18
30
|
if (entry.v?.$link) {
|
|
19
|
-
|
|
31
|
+
yield { path: fullKey, cid: entry.v.$link };
|
|
20
32
|
}
|
|
21
33
|
// Visit right subtree
|
|
22
34
|
if (entry.t?.$link) {
|
|
23
|
-
visit(entry.t.$link, lastKey);
|
|
35
|
+
yield* visit(entry.t.$link, lastKey);
|
|
24
36
|
}
|
|
25
37
|
}
|
|
26
38
|
return lastKey;
|
|
27
39
|
}
|
|
28
|
-
visit(rootCid, '');
|
|
29
|
-
return entries;
|
|
40
|
+
yield* visit(rootCid, '');
|
|
30
41
|
}
|
package/dist/oauth/db.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,
|
|
1
|
+
{"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMtG;AAID,wBAAsB,iBAAiB,CACrC,UAAU,EAAE,MAAM,EAClB,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,WAAW,EAAE,MAAM,CAAA;IACnB,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,aAAa,EAAE,MAAM,CAAA;IACrB,mBAAmB,CAAC,EAAE,MAAM,CAAA;IAC5B,OAAO,EAAE,MAAM,CAAA;IACf,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,QAAQ,CAAC,EAAE,MAAM,CAAA;IACjB,GAAG,CAAC,EAAE,MAAM,CAAA;IACZ,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,SAAS,EAAE,MAAM,CAAA;CAClB,GACA,OAAO,CAAC,IAAI,CAAC,CAsBf;AAED,wBAAsB,eAAe,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAM7E;AAED,wBAAsB,kBAAkB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE1E;AAID,wBAAsB,aAAa,CAAC,IAAI,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMnF;AAED,wBAAsB,eAAe,CAAC,IAAI,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAK1E;AAID,wBAAsB,YAAY,CAChC,GAAG,EAAE,MAAM,EACX,IAAI,EAAE;IACJ,WAAW,EAAE,MAAM,CAAA;IACnB,WAAW,EAAE,MAAM,CAAA;IACnB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,OAAO,EAAE,MAAM,CAAA;IACf,cAAc,CAAC,EAAE,MAAM,CAAA;CACxB,GACA,OAAO,CAAC,IAAI,CAAC,CAMf;AAED,wBAAsB,UAAU,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGjE;AAED,wBAAsB,aAAa,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE9D;AAID,wBAAsB,iBAAiB,CACrC,KAAK,EAAE,MAAM,EACb,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;IACX,OAAO,EAAE,MAAM,CAAA;IACf,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,SAAS,CAAC,EAAE,MAAM,CAAA;CACnB,GACA,OAAO,CAAC,IAAI,CAAC,CAQf;AAED,wBAAsB,eAAe,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGxE;AAED,wBAAsB,kBAAkB,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAErE;AAID,wBAAsB,oBAAoB,CAAC,GAAG,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,OAAO,CAAC,CAK3F;AAED,wBAAsB,mBAAmB,IAAI,OAAO,CAAC,IAAI,CAAC,CAQzD"}
|
package/dist/oauth/db.js
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
// packages/hatk/src/oauth/db.ts
|
|
2
|
-
import { querySQL, runSQL } from "../db.js";
|
|
2
|
+
import { querySQL, runSQL } from "../database/db.js";
|
|
3
3
|
// --- DDL ---
|
|
4
4
|
export const OAUTH_DDL = `
|
|
5
5
|
CREATE TABLE IF NOT EXISTS _oauth_keys (
|
|
@@ -62,18 +62,38 @@ CREATE TABLE IF NOT EXISTS _oauth_dpop_jtis (
|
|
|
62
62
|
`;
|
|
63
63
|
// --- Key Management ---
|
|
64
64
|
export async function getServerKey(kid) {
|
|
65
|
-
const rows = await querySQL('SELECT private_key, public_key FROM _oauth_keys WHERE kid = $1', [kid]);
|
|
65
|
+
const rows = (await querySQL('SELECT private_key, public_key FROM _oauth_keys WHERE kid = $1', [kid]));
|
|
66
66
|
if (rows.length === 0)
|
|
67
67
|
return null;
|
|
68
68
|
return { privateKey: rows[0].private_key, publicKey: rows[0].public_key };
|
|
69
69
|
}
|
|
70
70
|
export async function storeServerKey(kid, privateKey, publicKey) {
|
|
71
|
-
await runSQL('INSERT OR REPLACE INTO _oauth_keys (kid, private_key, public_key) VALUES ($1, $2, $3)',
|
|
71
|
+
await runSQL('INSERT OR REPLACE INTO _oauth_keys (kid, private_key, public_key) VALUES ($1, $2, $3)', [
|
|
72
|
+
kid,
|
|
73
|
+
privateKey,
|
|
74
|
+
publicKey,
|
|
75
|
+
]);
|
|
72
76
|
}
|
|
73
77
|
// --- OAuth Request Storage ---
|
|
74
78
|
export async function storeOAuthRequest(requestUri, data) {
|
|
75
79
|
await runSQL(`INSERT INTO _oauth_requests (request_uri, client_id, redirect_uri, scope, state, code_challenge, code_challenge_method, dpop_jkt, pds_request_uri, pds_auth_server, pds_code_verifier, pds_state, did, login_hint, expires_at)
|
|
76
|
-
VALUES ($1,$2,$3,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13,$14,$15)`,
|
|
80
|
+
VALUES ($1,$2,$3,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13,$14,$15)`, [
|
|
81
|
+
requestUri,
|
|
82
|
+
data.clientId,
|
|
83
|
+
data.redirectUri,
|
|
84
|
+
data.scope || null,
|
|
85
|
+
data.state || null,
|
|
86
|
+
data.codeChallenge,
|
|
87
|
+
data.codeChallengeMethod || 'S256',
|
|
88
|
+
data.dpopJkt,
|
|
89
|
+
data.pdsRequestUri || null,
|
|
90
|
+
data.pdsAuthServer || null,
|
|
91
|
+
data.pdsCodeVerifier || null,
|
|
92
|
+
data.pdsState || null,
|
|
93
|
+
data.did || null,
|
|
94
|
+
data.loginHint || null,
|
|
95
|
+
data.expiresAt,
|
|
96
|
+
]);
|
|
77
97
|
}
|
|
78
98
|
export async function getOAuthRequest(requestUri) {
|
|
79
99
|
const rows = await querySQL('SELECT * FROM _oauth_requests WHERE request_uri = $1 AND expires_at > $2', [
|
|
@@ -83,57 +103,63 @@ export async function getOAuthRequest(requestUri) {
|
|
|
83
103
|
return rows.length > 0 ? rows[0] : null;
|
|
84
104
|
}
|
|
85
105
|
export async function deleteOAuthRequest(requestUri) {
|
|
86
|
-
await runSQL('DELETE FROM _oauth_requests WHERE request_uri = $1', requestUri);
|
|
106
|
+
await runSQL('DELETE FROM _oauth_requests WHERE request_uri = $1', [requestUri]);
|
|
87
107
|
}
|
|
88
108
|
// --- Authorization Codes ---
|
|
89
109
|
export async function storeAuthCode(code, requestUri) {
|
|
90
|
-
await runSQL('INSERT INTO _oauth_codes (code, request_uri, created_at) VALUES ($1, $2, $3)',
|
|
110
|
+
await runSQL('INSERT INTO _oauth_codes (code, request_uri, created_at) VALUES ($1, $2, $3)', [
|
|
111
|
+
code,
|
|
112
|
+
requestUri,
|
|
113
|
+
Math.floor(Date.now() / 1000),
|
|
114
|
+
]);
|
|
91
115
|
}
|
|
92
116
|
export async function consumeAuthCode(code) {
|
|
93
|
-
const rows = await querySQL('SELECT request_uri FROM _oauth_codes WHERE code = $1', [code]);
|
|
117
|
+
const rows = (await querySQL('SELECT request_uri FROM _oauth_codes WHERE code = $1', [code]));
|
|
94
118
|
if (rows.length === 0)
|
|
95
119
|
return null;
|
|
96
|
-
await runSQL('DELETE FROM _oauth_codes WHERE code = $1', code);
|
|
120
|
+
await runSQL('DELETE FROM _oauth_codes WHERE code = $1', [code]);
|
|
97
121
|
return rows[0].request_uri;
|
|
98
122
|
}
|
|
99
123
|
// --- Sessions ---
|
|
100
124
|
export async function storeSession(did, data) {
|
|
101
125
|
await runSQL(`INSERT OR REPLACE INTO _oauth_sessions (did, pds_endpoint, access_token, refresh_token, dpop_jkt, token_expires_at, updated_at)
|
|
102
|
-
VALUES ($1,$2,$3,$4,$5,$6,CURRENT_TIMESTAMP)`, did, data.pdsEndpoint, data.accessToken, data.refreshToken || null, data.dpopJkt, data.tokenExpiresAt || null);
|
|
126
|
+
VALUES ($1,$2,$3,$4,$5,$6,CURRENT_TIMESTAMP)`, [did, data.pdsEndpoint, data.accessToken, data.refreshToken || null, data.dpopJkt, data.tokenExpiresAt || null]);
|
|
103
127
|
}
|
|
104
128
|
export async function getSession(did) {
|
|
105
129
|
const rows = await querySQL('SELECT * FROM _oauth_sessions WHERE did = $1', [did]);
|
|
106
130
|
return rows.length > 0 ? rows[0] : null;
|
|
107
131
|
}
|
|
108
132
|
export async function deleteSession(did) {
|
|
109
|
-
await runSQL('DELETE FROM _oauth_sessions WHERE did = $1', did);
|
|
133
|
+
await runSQL('DELETE FROM _oauth_sessions WHERE did = $1', [did]);
|
|
110
134
|
}
|
|
111
135
|
// --- Refresh Tokens ---
|
|
112
136
|
export async function storeRefreshToken(token, data) {
|
|
113
137
|
const now = Math.floor(Date.now() / 1000);
|
|
114
138
|
const expiresAt = data.expiresAt ?? now + 14 * 86400; // 14 days default
|
|
115
139
|
await runSQL(`INSERT INTO _oauth_refresh_tokens (token, client_id, did, dpop_jkt, scope, created_at, expires_at)
|
|
116
|
-
VALUES ($1,$2,$3,$4,$5,$6,$7)`, token, data.clientId, data.did, data.dpopJkt, data.scope || null, now, expiresAt);
|
|
140
|
+
VALUES ($1,$2,$3,$4,$5,$6,$7)`, [token, data.clientId, data.did, data.dpopJkt, data.scope || null, now, expiresAt]);
|
|
117
141
|
}
|
|
118
142
|
export async function getRefreshToken(token) {
|
|
119
143
|
const rows = await querySQL('SELECT * FROM _oauth_refresh_tokens WHERE token = $1', [token]);
|
|
120
144
|
return rows.length > 0 ? rows[0] : null;
|
|
121
145
|
}
|
|
122
146
|
export async function revokeRefreshToken(token) {
|
|
123
|
-
await runSQL('UPDATE _oauth_refresh_tokens SET revoked = 1 WHERE token = $1', token);
|
|
147
|
+
await runSQL('UPDATE _oauth_refresh_tokens SET revoked = 1 WHERE token = $1', [token]);
|
|
124
148
|
}
|
|
125
149
|
// --- DPoP JTI Replay Protection ---
|
|
126
150
|
export async function checkAndStoreDpopJti(jti, expiresAt) {
|
|
127
151
|
const rows = await querySQL('SELECT 1 FROM _oauth_dpop_jtis WHERE jti = $1', [jti]);
|
|
128
152
|
if (rows.length > 0)
|
|
129
153
|
return false;
|
|
130
|
-
await runSQL('INSERT INTO _oauth_dpop_jtis (jti, expires_at) VALUES ($1, $2)', jti, expiresAt);
|
|
154
|
+
await runSQL('INSERT INTO _oauth_dpop_jtis (jti, expires_at) VALUES ($1, $2)', [jti, expiresAt]);
|
|
131
155
|
return true;
|
|
132
156
|
}
|
|
133
157
|
export async function cleanupExpiredOAuth() {
|
|
134
158
|
const now = Math.floor(Date.now() / 1000);
|
|
135
|
-
await runSQL('DELETE FROM _oauth_dpop_jtis WHERE expires_at < $1', now);
|
|
136
|
-
await runSQL('DELETE FROM _oauth_requests WHERE expires_at < $1', now);
|
|
137
|
-
await runSQL('DELETE FROM _oauth_codes WHERE created_at < $1', now - 600);
|
|
138
|
-
await runSQL('DELETE FROM _oauth_refresh_tokens WHERE revoked = 1 OR (expires_at IS NOT NULL AND expires_at < $1)',
|
|
159
|
+
await runSQL('DELETE FROM _oauth_dpop_jtis WHERE expires_at < $1', [now]);
|
|
160
|
+
await runSQL('DELETE FROM _oauth_requests WHERE expires_at < $1', [now]);
|
|
161
|
+
await runSQL('DELETE FROM _oauth_codes WHERE created_at < $1', [now - 600]);
|
|
162
|
+
await runSQL('DELETE FROM _oauth_refresh_tokens WHERE revoked = 1 OR (expires_at IS NOT NULL AND expires_at < $1)', [
|
|
163
|
+
now,
|
|
164
|
+
]);
|
|
139
165
|
}
|
package/dist/oauth/server.d.ts
CHANGED
|
@@ -64,10 +64,12 @@ export declare function handlePar(config: OAuthConfig, body: Record<string, stri
|
|
|
64
64
|
expires_in: number;
|
|
65
65
|
}>;
|
|
66
66
|
export declare function buildAuthorizeRedirect(config: OAuthConfig, request: any): string;
|
|
67
|
+
export declare function serverLogin(config: OAuthConfig, handle: string): Promise<string>;
|
|
67
68
|
export declare function handleCallback(config: OAuthConfig, code: string, state: string | null, iss: string | null): Promise<{
|
|
68
69
|
requestUri: string;
|
|
69
70
|
clientRedirectUri: string;
|
|
70
71
|
clientState: string | null;
|
|
72
|
+
did: string;
|
|
71
73
|
}>;
|
|
72
74
|
export declare function handleToken(config: OAuthConfig, body: Record<string, string>, dpopHeader: string, requestUrl: string): Promise<any>;
|
|
73
75
|
export declare function refreshPdsSession(config: OAuthConfig, session: {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"server.d.ts","sourceRoot":"","sources":["../../src/oauth/server.ts"],"names":[],"mappings":"AAEA,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,cAAc,CAAA;
|
|
1
|
+
{"version":3,"file":"server.d.ts","sourceRoot":"","sources":["../../src/oauth/server.ts"],"names":[],"mappings":"AAEA,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,cAAc,CAAA;AA2E/C,wBAAsB,SAAS,CAAC,OAAO,EAAE,WAAW,EAAE,MAAM,EAAE,MAAM,EAAE,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAsBrG;AAID,wBAAgB,qBAAqB,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,WAAW;;;;;;;;;;;;;;;;;;;EAqBxE;AAED,wBAAgB,4BAA4B,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,WAAW;;;;;EAO/E;AAED,wBAAgB,OAAO;;;;;;;;;;;;;;;;;;;;;;EAWtB;AAED,wBAAgB,iBAAiB,CAAC,MAAM,EAAE,MAAM,EAAE,MAAM,EAAE,WAAW;;;;;;;;;EAcpE;AAID,wBAAsB,SAAS,CAC7B,MAAM,EAAE,WAAW,EACnB,IAAI,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,EAC5B,UAAU,EAAE,MAAM,EAClB,UAAU,EAAE,MAAM,GACjB,OAAO,CAAC;IAAE,WAAW,EAAE,MAAM,CAAC;IAAC,UAAU,EAAE,MAAM,CAAA;CAAE,CAAC,CA2ItD;AAID,wBAAgB,sBAAsB,CAAC,MAAM,EAAE,WAAW,EAAE,OAAO,EAAE,GAAG,GAAG,MAAM,CAShF;AAID,wBAAsB,WAAW,CAAC,MAAM,EAAE,WAAW,EAAE,MAAM,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,CAAC,CAoGtF;AAID,wBAAsB,cAAc,CAClC,MAAM,EAAE,WAAW,EACnB,IAAI,EAAE,MAAM,EACZ,KAAK,EAAE,MAAM,GAAG,IAAI,EACpB,GAAG,EAAE,MAAM,GAAG,IAAI,GACjB,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,iBAAiB,EAAE,MAAM,CAAC;IAAC,WAAW,EAAE,MAAM,GAAG,IAAI,CAAC;IAAC,GAAG,EAAE,MAAM,CAAA;CAAE,CAAC,CAyHrG;AAID,wBAAsB,WAAW,CAC/B,MAAM,EAAE,WAAW,EACnB,IAAI,EAAE,MAAM,CAAC,MAAM,EAAE,MAAM,CAAC,EAC5B,UAAU,EAAE,MAAM,EAClB,UAAU,EAAE,MAAM,GACjB,OAAO,CAAC,GAAG,CAAC,CAUd;AA0JD,wBAAsB,iBAAiB,CACrC,MAAM,EAAE,WAAW,EACnB,OAAO,EAAE;IAAE,GAAG,EAAE,MAAM,CAAC;IAAC,YAAY,EAAE,MAAM,CAAC;IAAC,aAAa,EAAE,MAAM,CAAC;IAAC,QAAQ,EAAE,MAAM,CAAA;CAAE,GACtF,OAAO,CAAC;IAAE,WAAW,EAAE,MAAM,CAAC;IAAC,YAAY,CAAC,EAAE,MAAM,CAAC;IAAC,SAAS,CAAC,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAmEpF;AAID,wBAAsB,YAAY,CAChC,UAAU,EAAE,MAAM,GAAG,IAAI,EACzB,UAAU,EAAE,MAAM,GAAG,IAAI,EACzB,MAAM,EAAE,MAAM,EACd,GAAG,EAAE,MAAM,GACV,OAAO,CAAC;IAAE,GAAG,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CA0BjC"}
|