@hatk/hatk 0.0.1-alpha.4 → 0.0.1-alpha.40
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapter.d.ts +19 -0
- package/dist/adapter.d.ts.map +1 -0
- package/dist/adapter.js +107 -0
- package/dist/backfill.d.ts +60 -1
- package/dist/backfill.d.ts.map +1 -1
- package/dist/backfill.js +167 -33
- package/dist/car.d.ts +59 -1
- package/dist/car.d.ts.map +1 -1
- package/dist/car.js +179 -7
- package/dist/cbor.d.ts +37 -0
- package/dist/cbor.d.ts.map +1 -1
- package/dist/cbor.js +36 -3
- package/dist/cid.d.ts +37 -0
- package/dist/cid.d.ts.map +1 -1
- package/dist/cid.js +38 -3
- package/dist/cli.js +417 -133
- package/dist/cloudflare/container.d.ts +73 -0
- package/dist/cloudflare/container.d.ts.map +1 -0
- package/dist/cloudflare/container.js +232 -0
- package/dist/cloudflare/hooks.d.ts +33 -0
- package/dist/cloudflare/hooks.d.ts.map +1 -0
- package/dist/cloudflare/hooks.js +40 -0
- package/dist/cloudflare/init.d.ts +27 -0
- package/dist/cloudflare/init.d.ts.map +1 -0
- package/dist/cloudflare/init.js +103 -0
- package/dist/cloudflare/worker.d.ts +27 -0
- package/dist/cloudflare/worker.d.ts.map +1 -0
- package/dist/cloudflare/worker.js +54 -0
- package/dist/config.d.ts +12 -1
- package/dist/config.d.ts.map +1 -1
- package/dist/config.js +36 -9
- package/dist/database/adapter-factory.d.ts +6 -0
- package/dist/database/adapter-factory.d.ts.map +1 -0
- package/dist/database/adapter-factory.js +20 -0
- package/dist/database/adapters/d1.d.ts +56 -0
- package/dist/database/adapters/d1.d.ts.map +1 -0
- package/dist/database/adapters/d1.js +108 -0
- package/dist/database/adapters/duckdb-search.d.ts +12 -0
- package/dist/database/adapters/duckdb-search.d.ts.map +1 -0
- package/dist/database/adapters/duckdb-search.js +27 -0
- package/dist/database/adapters/duckdb.d.ts +25 -0
- package/dist/database/adapters/duckdb.d.ts.map +1 -0
- package/dist/database/adapters/duckdb.js +161 -0
- package/dist/database/adapters/sqlite-search.d.ts +23 -0
- package/dist/database/adapters/sqlite-search.d.ts.map +1 -0
- package/dist/database/adapters/sqlite-search.js +74 -0
- package/dist/database/adapters/sqlite.d.ts +18 -0
- package/dist/database/adapters/sqlite.d.ts.map +1 -0
- package/dist/database/adapters/sqlite.js +87 -0
- package/dist/database/db.d.ts +159 -0
- package/dist/database/db.d.ts.map +1 -0
- package/dist/database/db.js +1445 -0
- package/dist/database/dialect.d.ts +45 -0
- package/dist/database/dialect.d.ts.map +1 -0
- package/dist/database/dialect.js +72 -0
- package/dist/database/fts.d.ts +27 -0
- package/dist/database/fts.d.ts.map +1 -0
- package/dist/database/fts.js +846 -0
- package/dist/database/index.d.ts +7 -0
- package/dist/database/index.d.ts.map +1 -0
- package/dist/database/index.js +6 -0
- package/dist/database/ports.d.ts +50 -0
- package/dist/database/ports.d.ts.map +1 -0
- package/dist/database/ports.js +1 -0
- package/dist/database/schema.d.ts +61 -0
- package/dist/database/schema.d.ts.map +1 -0
- package/dist/database/schema.js +394 -0
- package/dist/db.d.ts +1 -1
- package/dist/db.d.ts.map +1 -1
- package/dist/db.js +4 -38
- package/dist/dev-entry.d.ts +8 -0
- package/dist/dev-entry.d.ts.map +1 -0
- package/dist/dev-entry.js +110 -0
- package/dist/feeds.d.ts +12 -8
- package/dist/feeds.d.ts.map +1 -1
- package/dist/feeds.js +45 -6
- package/dist/fts.d.ts.map +1 -1
- package/dist/fts.js +5 -0
- package/dist/hooks.d.ts +22 -0
- package/dist/hooks.d.ts.map +1 -0
- package/dist/hooks.js +75 -0
- package/dist/hydrate.d.ts +6 -5
- package/dist/hydrate.d.ts.map +1 -1
- package/dist/hydrate.js +4 -16
- package/dist/indexer.d.ts +20 -0
- package/dist/indexer.d.ts.map +1 -1
- package/dist/indexer.js +53 -7
- package/dist/labels.d.ts +34 -0
- package/dist/labels.d.ts.map +1 -1
- package/dist/labels.js +66 -6
- package/dist/logger.d.ts +29 -0
- package/dist/logger.d.ts.map +1 -1
- package/dist/logger.js +29 -0
- package/dist/main.js +134 -67
- package/dist/mst.d.ts +18 -1
- package/dist/mst.d.ts.map +1 -1
- package/dist/mst.js +19 -8
- package/dist/oauth/db.d.ts.map +1 -1
- package/dist/oauth/db.js +43 -17
- package/dist/oauth/server.d.ts +2 -0
- package/dist/oauth/server.d.ts.map +1 -1
- package/dist/oauth/server.js +102 -7
- package/dist/oauth/session.d.ts +11 -0
- package/dist/oauth/session.d.ts.map +1 -0
- package/dist/oauth/session.js +65 -0
- package/dist/opengraph.d.ts +10 -0
- package/dist/opengraph.d.ts.map +1 -1
- package/dist/opengraph.js +73 -39
- package/dist/pds-proxy.d.ts +42 -0
- package/dist/pds-proxy.d.ts.map +1 -0
- package/dist/pds-proxy.js +189 -0
- package/dist/renderer.d.ts +27 -0
- package/dist/renderer.d.ts.map +1 -0
- package/dist/renderer.js +46 -0
- package/dist/resolve-hatk.d.ts +6 -0
- package/dist/resolve-hatk.d.ts.map +1 -0
- package/dist/resolve-hatk.js +20 -0
- package/dist/response.d.ts +16 -0
- package/dist/response.d.ts.map +1 -0
- package/dist/response.js +69 -0
- package/dist/scanner.d.ts +21 -0
- package/dist/scanner.d.ts.map +1 -0
- package/dist/scanner.js +88 -0
- package/dist/schema.d.ts +8 -0
- package/dist/schema.d.ts.map +1 -1
- package/dist/schema.js +29 -0
- package/dist/seed.d.ts +19 -0
- package/dist/seed.d.ts.map +1 -1
- package/dist/seed.js +43 -4
- package/dist/server-init.d.ts +8 -0
- package/dist/server-init.d.ts.map +1 -0
- package/dist/server-init.js +61 -0
- package/dist/server.d.ts +26 -3
- package/dist/server.d.ts.map +1 -1
- package/dist/server.js +528 -635
- package/dist/setup.d.ts +28 -1
- package/dist/setup.d.ts.map +1 -1
- package/dist/setup.js +50 -3
- package/dist/test.d.ts +1 -1
- package/dist/test.d.ts.map +1 -1
- package/dist/test.js +38 -32
- package/dist/views.js +1 -1
- package/dist/vite-plugin.d.ts +1 -1
- package/dist/vite-plugin.d.ts.map +1 -1
- package/dist/vite-plugin.js +254 -66
- package/dist/xrpc.d.ts +46 -10
- package/dist/xrpc.d.ts.map +1 -1
- package/dist/xrpc.js +128 -39
- package/package.json +13 -6
- package/public/admin.html +0 -54
package/dist/labels.d.ts
CHANGED
|
@@ -13,7 +13,34 @@ export interface LabelRuleContext {
|
|
|
13
13
|
value: Record<string, any>;
|
|
14
14
|
};
|
|
15
15
|
}
|
|
16
|
+
export interface LabelModule {
|
|
17
|
+
definition?: LabelDefinition;
|
|
18
|
+
evaluate?: (ctx: LabelRuleContext) => Promise<string[]>;
|
|
19
|
+
}
|
|
20
|
+
export declare function defineLabel(module: LabelModule): {
|
|
21
|
+
definition?: LabelDefinition;
|
|
22
|
+
evaluate?: (ctx: LabelRuleContext) => Promise<string[]>;
|
|
23
|
+
__type: "labels";
|
|
24
|
+
};
|
|
25
|
+
/**
|
|
26
|
+
* Discover and load label rule modules from the `labels/` directory.
|
|
27
|
+
*
|
|
28
|
+
* Each module should default-export an object with an optional `definition`
|
|
29
|
+
* (label metadata like severity and blur behavior) and an optional `evaluate`
|
|
30
|
+
* function that returns label values to apply to a record.
|
|
31
|
+
*
|
|
32
|
+
* @param labelsDir - Absolute path to the `labels/` directory
|
|
33
|
+
*/
|
|
16
34
|
export declare function initLabels(labelsDir: string): Promise<void>;
|
|
35
|
+
/** Register a single label module from a scanned server/ module. */
|
|
36
|
+
export declare function registerLabelModule(name: string, labelMod: {
|
|
37
|
+
definition?: LabelDefinition;
|
|
38
|
+
evaluate?: (ctx: LabelRuleContext) => Promise<string[]>;
|
|
39
|
+
}): void;
|
|
40
|
+
/**
|
|
41
|
+
* Evaluate all loaded label rules against a record and persist any resulting labels.
|
|
42
|
+
* Called after each record is indexed. Rule errors are logged but never block indexing.
|
|
43
|
+
*/
|
|
17
44
|
export declare function runLabelRules(record: {
|
|
18
45
|
uri: string;
|
|
19
46
|
cid: string;
|
|
@@ -21,9 +48,16 @@ export declare function runLabelRules(record: {
|
|
|
21
48
|
collection: string;
|
|
22
49
|
value: Record<string, any>;
|
|
23
50
|
}): Promise<void>;
|
|
51
|
+
/**
|
|
52
|
+
* Re-evaluate all label rules against every existing record in the given collections.
|
|
53
|
+
* Used by `/admin/rescan-labels` to apply new or updated rules retroactively.
|
|
54
|
+
*
|
|
55
|
+
* @returns Count of records scanned and new labels applied
|
|
56
|
+
*/
|
|
24
57
|
export declare function rescanLabels(collections: string[]): Promise<{
|
|
25
58
|
scanned: number;
|
|
26
59
|
labeled: number;
|
|
27
60
|
}>;
|
|
61
|
+
/** Return all label definitions discovered during {@link initLabels}. */
|
|
28
62
|
export declare function getLabelDefinitions(): LabelDefinition[];
|
|
29
63
|
//# sourceMappingURL=labels.d.ts.map
|
package/dist/labels.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"labels.d.ts","sourceRoot":"","sources":["../src/labels.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"labels.d.ts","sourceRoot":"","sources":["../src/labels.ts"],"names":[],"mappings":"AA8BA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,aAAa,CAAA;AAIlD,wDAAwD;AACxD,MAAM,WAAW,gBAAgB;IAC/B,EAAE,EAAE;QACF,KAAK,EAAE,CAAC,GAAG,EAAE,MAAM,EAAE,MAAM,CAAC,EAAE,GAAG,EAAE,KAAK,OAAO,CAAC,GAAG,EAAE,CAAC,CAAA;QACtD,GAAG,EAAE,CAAC,GAAG,EAAE,MAAM,EAAE,GAAG,MAAM,EAAE,GAAG,EAAE,KAAK,OAAO,CAAC,IAAI,CAAC,CAAA;KACtD,CAAA;IACD,MAAM,EAAE;QACN,GAAG,EAAE,MAAM,CAAA;QACX,GAAG,EAAE,MAAM,CAAA;QACX,GAAG,EAAE,MAAM,CAAA;QACX,UAAU,EAAE,MAAM,CAAA;QAClB,KAAK,EAAE,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAA;KAC3B,CAAA;CACF;AAED,MAAM,WAAW,WAAW;IAC1B,UAAU,CAAC,EAAE,eAAe,CAAA;IAC5B,QAAQ,CAAC,EAAE,CAAC,GAAG,EAAE,gBAAgB,KAAK,OAAO,CAAC,MAAM,EAAE,CAAC,CAAA;CACxD;AAED,wBAAgB,WAAW,CAAC,MAAM,EAAE,WAAW;iBAJhC,eAAe;eACjB,CAAC,GAAG,EAAE,gBAAgB,KAAK,OAAO,CAAC,MAAM,EAAE,CAAC;;EAKxD;AAYD;;;;;;;;GAQG;AACH,wBAAsB,UAAU,CAAC,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAmCjE;AAED,oEAAoE;AACpE,wBAAgB,mBAAmB,CACjC,IAAI,EAAE,MAAM,EACZ,QAAQ,EAAE;IAAE,UAAU,CAAC,EAAE,eAAe,CAAC;IAAC,QAAQ,CAAC,EAAE,CAAC,GAAG,EAAE,gBAAgB,KAAK,OAAO,CAAC,MAAM,EAAE,CAAC,CAAA;CAAE,GAClG,IAAI,CAON;AAED;;;GAGG;AACH,wBAAsB,aAAa,CAAC,MAAM,EAAE;IAC1C,GAAG,EAAE,MAAM,CAAA;IACX,GAAG,EAAE,MAAM,CAAA;IACX,GAAG,EAAE,MAAM,CAAA;IACX,UAAU,EAAE,MAAM,CAAA;IAClB,KAAK,EAAE,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAA;CAC3B,GAAG,OAAO,CAAC,IAAI,CAAC,CAyBhB;AAED;;;;;GAKG;AACH,wBAAsB,YAAY,CAAC,WAAW,EAAE,MAAM,EAAE,GAAG,OAAO,CAAC;IAAE,OAAO,EAAE,MAAM,CAAC;IAAC,OAAO,EAAE,MAAM,CAAA;CAAE,CAAC,CAuCvG;AAED,yEAAyE;AACzE,wBAAgB,mBAAmB,IAAI,eAAe,EAAE,CAEvD"}
|
package/dist/labels.js
CHANGED
|
@@ -6,13 +6,53 @@ var __rewriteRelativeImportExtension = (this && this.__rewriteRelativeImportExte
|
|
|
6
6
|
}
|
|
7
7
|
return path;
|
|
8
8
|
};
|
|
9
|
+
/**
|
|
10
|
+
* Label system for applying moderation labels to records as they are indexed.
|
|
11
|
+
*
|
|
12
|
+
* Place label modules in the `labels/` directory. Each module default-exports
|
|
13
|
+
* an object with a `definition` (label metadata) and/or an `evaluate` function
|
|
14
|
+
* (rule that returns label values for a given record).
|
|
15
|
+
*
|
|
16
|
+
* @example
|
|
17
|
+
* ```ts
|
|
18
|
+
* // labels/nsfw.ts
|
|
19
|
+
* import type { LabelRuleContext } from '@hatk/hatk/labels'
|
|
20
|
+
*
|
|
21
|
+
* export default {
|
|
22
|
+
* definition: {
|
|
23
|
+
* identifier: 'nsfw',
|
|
24
|
+
* severity: 'alert',
|
|
25
|
+
* blurs: 'media',
|
|
26
|
+
* defaultSetting: 'warn',
|
|
27
|
+
* locales: [{ lang: 'en', name: 'NSFW', description: 'Not safe for work' }],
|
|
28
|
+
* },
|
|
29
|
+
*
|
|
30
|
+
* async evaluate(ctx: LabelRuleContext): Promise<string[]> {
|
|
31
|
+
* if (ctx.record.value.nsfw === true) return ['nsfw']
|
|
32
|
+
* return []
|
|
33
|
+
* },
|
|
34
|
+
* }
|
|
35
|
+
* ```
|
|
36
|
+
*/
|
|
9
37
|
import { resolve } from 'node:path';
|
|
10
38
|
import { readdirSync } from 'node:fs';
|
|
11
|
-
import { querySQL, runSQL, insertLabels, getSchema } from "./db.js";
|
|
39
|
+
import { querySQL, runSQL, insertLabels, getSchema } from "./database/db.js";
|
|
12
40
|
import { log, emit } from "./logger.js";
|
|
41
|
+
export function defineLabel(module) {
|
|
42
|
+
return { __type: 'labels', ...module };
|
|
43
|
+
}
|
|
13
44
|
const rules = [];
|
|
14
45
|
let labelDefs = [];
|
|
15
46
|
let labelSrc = 'self';
|
|
47
|
+
/**
|
|
48
|
+
* Discover and load label rule modules from the `labels/` directory.
|
|
49
|
+
*
|
|
50
|
+
* Each module should default-export an object with an optional `definition`
|
|
51
|
+
* (label metadata like severity and blur behavior) and an optional `evaluate`
|
|
52
|
+
* function that returns label values to apply to a record.
|
|
53
|
+
*
|
|
54
|
+
* @param labelsDir - Absolute path to the `labels/` directory
|
|
55
|
+
*/
|
|
16
56
|
export async function initLabels(labelsDir) {
|
|
17
57
|
let files;
|
|
18
58
|
try {
|
|
@@ -26,7 +66,7 @@ export async function initLabels(labelsDir) {
|
|
|
26
66
|
for (const file of files) {
|
|
27
67
|
const name = file.replace(/\.(ts|js)$/, '');
|
|
28
68
|
const scriptPath = resolve(labelsDir, file);
|
|
29
|
-
const mod = await import(__rewriteRelativeImportExtension(scriptPath));
|
|
69
|
+
const mod = await import(__rewriteRelativeImportExtension(/* @vite-ignore */ `${scriptPath}?t=${Date.now()}`));
|
|
30
70
|
const handler = mod.default;
|
|
31
71
|
if (handler.definition) {
|
|
32
72
|
labelDefs.push(handler.definition);
|
|
@@ -45,6 +85,19 @@ export async function initLabels(labelsDir) {
|
|
|
45
85
|
log(`[labels] ${labelDefs.length} label definitions loaded`);
|
|
46
86
|
}
|
|
47
87
|
}
|
|
88
|
+
/** Register a single label module from a scanned server/ module. */
|
|
89
|
+
export function registerLabelModule(name, labelMod) {
|
|
90
|
+
if (labelMod.definition) {
|
|
91
|
+
labelDefs.push(labelMod.definition);
|
|
92
|
+
}
|
|
93
|
+
if (labelMod.evaluate) {
|
|
94
|
+
rules.push({ name, evaluate: labelMod.evaluate });
|
|
95
|
+
}
|
|
96
|
+
}
|
|
97
|
+
/**
|
|
98
|
+
* Evaluate all loaded label rules against a record and persist any resulting labels.
|
|
99
|
+
* Called after each record is indexed. Rule errors are logged but never block indexing.
|
|
100
|
+
*/
|
|
48
101
|
export async function runLabelRules(record) {
|
|
49
102
|
if (rules.length === 0)
|
|
50
103
|
return;
|
|
@@ -69,15 +122,21 @@ export async function runLabelRules(record) {
|
|
|
69
122
|
emit('labels', 'applied', { count: allLabels.length, uri: record.uri, vals: allLabels.map((l) => l.val) });
|
|
70
123
|
}
|
|
71
124
|
}
|
|
125
|
+
/**
|
|
126
|
+
* Re-evaluate all label rules against every existing record in the given collections.
|
|
127
|
+
* Used by `/admin/rescan-labels` to apply new or updated rules retroactively.
|
|
128
|
+
*
|
|
129
|
+
* @returns Count of records scanned and new labels applied
|
|
130
|
+
*/
|
|
72
131
|
export async function rescanLabels(collections) {
|
|
73
|
-
const beforeRows = await querySQL(`SELECT COUNT(*) as count FROM _labels`);
|
|
132
|
+
const beforeRows = (await querySQL(`SELECT COUNT(*) as count FROM _labels`));
|
|
74
133
|
const beforeCount = Number(beforeRows[0]?.count || 0);
|
|
75
134
|
let scanned = 0;
|
|
76
135
|
for (const collection of collections) {
|
|
77
136
|
const schema = getSchema(collection);
|
|
78
137
|
if (!schema)
|
|
79
138
|
continue;
|
|
80
|
-
const rows = await querySQL(`SELECT * FROM ${schema.tableName}`);
|
|
139
|
+
const rows = (await querySQL(`SELECT * FROM ${schema.tableName}`));
|
|
81
140
|
for (const row of rows) {
|
|
82
141
|
scanned++;
|
|
83
142
|
const value = {};
|
|
@@ -85,7 +144,7 @@ export async function rescanLabels(collections) {
|
|
|
85
144
|
let v = row[col.name];
|
|
86
145
|
if (v === null || v === undefined)
|
|
87
146
|
continue;
|
|
88
|
-
if (col.
|
|
147
|
+
if (col.isJson && typeof v === 'string') {
|
|
89
148
|
try {
|
|
90
149
|
v = JSON.parse(v);
|
|
91
150
|
}
|
|
@@ -102,10 +161,11 @@ export async function rescanLabels(collections) {
|
|
|
102
161
|
});
|
|
103
162
|
}
|
|
104
163
|
}
|
|
105
|
-
const afterRows = await querySQL(`SELECT COUNT(*) as count FROM _labels`);
|
|
164
|
+
const afterRows = (await querySQL(`SELECT COUNT(*) as count FROM _labels`));
|
|
106
165
|
const afterCount = Number(afterRows[0]?.count || 0);
|
|
107
166
|
return { scanned, labeled: afterCount - beforeCount };
|
|
108
167
|
}
|
|
168
|
+
/** Return all label definitions discovered during {@link initLabels}. */
|
|
109
169
|
export function getLabelDefinitions() {
|
|
110
170
|
return labelDefs;
|
|
111
171
|
}
|
package/dist/logger.d.ts
CHANGED
|
@@ -1,4 +1,33 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Unstructured debug log — use sparingly for human-readable dev output.
|
|
3
|
+
* Prefer {@link emit} for anything that should be queryable in production.
|
|
4
|
+
* Disabled when `DEBUG=0`.
|
|
5
|
+
*/
|
|
1
6
|
export declare function log(...args: unknown[]): void;
|
|
7
|
+
/**
|
|
8
|
+
* Emit a structured wide event as a single JSON line to stdout.
|
|
9
|
+
*
|
|
10
|
+
* Each call produces one canonical log line with a timestamp, module, operation,
|
|
11
|
+
* and arbitrary key-value fields — designed for columnar search and aggregation,
|
|
12
|
+
* not string grep. Pack as much context as possible into `fields` (request IDs,
|
|
13
|
+
* durations, status codes, user DIDs, counts) so a single event tells the full
|
|
14
|
+
* story. See https://loggingsucks.com for the philosophy behind this approach.
|
|
15
|
+
*
|
|
16
|
+
* Disabled when `DEBUG=0`.
|
|
17
|
+
*
|
|
18
|
+
* @param module - Subsystem emitting the event (e.g. "server", "indexer", "backfill")
|
|
19
|
+
* @param op - Operation name (e.g. "request", "commit", "memory")
|
|
20
|
+
* @param fields - High-cardinality key-value context — include everything relevant
|
|
21
|
+
*/
|
|
2
22
|
export declare function emit(module: string, op: string, fields: Record<string, unknown>): void;
|
|
23
|
+
/**
|
|
24
|
+
* Start a millisecond timer. Call the returned function to get elapsed ms.
|
|
25
|
+
* Use with {@link emit} to add `duration_ms` to wide events.
|
|
26
|
+
*
|
|
27
|
+
* @example
|
|
28
|
+
* const elapsed = timer()
|
|
29
|
+
* await doWork()
|
|
30
|
+
* emit('server', 'request', { path, status_code, duration_ms: elapsed() })
|
|
31
|
+
*/
|
|
3
32
|
export declare function timer(): () => number;
|
|
4
33
|
//# sourceMappingURL=logger.d.ts.map
|
package/dist/logger.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"logger.d.ts","sourceRoot":"","sources":["../src/logger.ts"],"names":[],"mappings":"AAAA,wBAAgB,GAAG,CAAC,GAAG,IAAI,EAAE,OAAO,EAAE,GAAG,IAAI,CAG5C;AAED,wBAAgB,IAAI,CAAC,MAAM,EAAE,MAAM,EAAE,EAAE,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI,CAWtF;AAED,wBAAgB,KAAK,IAAI,MAAM,MAAM,CAGpC"}
|
|
1
|
+
{"version":3,"file":"logger.d.ts","sourceRoot":"","sources":["../src/logger.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,wBAAgB,GAAG,CAAC,GAAG,IAAI,EAAE,OAAO,EAAE,GAAG,IAAI,CAG5C;AAED;;;;;;;;;;;;;;GAcG;AACH,wBAAgB,IAAI,CAAC,MAAM,EAAE,MAAM,EAAE,EAAE,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI,CAWtF;AAED;;;;;;;;GAQG;AACH,wBAAgB,KAAK,IAAI,MAAM,MAAM,CAGpC"}
|
package/dist/logger.js
CHANGED
|
@@ -1,8 +1,28 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Unstructured debug log — use sparingly for human-readable dev output.
|
|
3
|
+
* Prefer {@link emit} for anything that should be queryable in production.
|
|
4
|
+
* Disabled when `DEBUG=0`.
|
|
5
|
+
*/
|
|
1
6
|
export function log(...args) {
|
|
2
7
|
if (process.env.DEBUG === '0')
|
|
3
8
|
return;
|
|
4
9
|
console.log(...args);
|
|
5
10
|
}
|
|
11
|
+
/**
|
|
12
|
+
* Emit a structured wide event as a single JSON line to stdout.
|
|
13
|
+
*
|
|
14
|
+
* Each call produces one canonical log line with a timestamp, module, operation,
|
|
15
|
+
* and arbitrary key-value fields — designed for columnar search and aggregation,
|
|
16
|
+
* not string grep. Pack as much context as possible into `fields` (request IDs,
|
|
17
|
+
* durations, status codes, user DIDs, counts) so a single event tells the full
|
|
18
|
+
* story. See https://loggingsucks.com for the philosophy behind this approach.
|
|
19
|
+
*
|
|
20
|
+
* Disabled when `DEBUG=0`.
|
|
21
|
+
*
|
|
22
|
+
* @param module - Subsystem emitting the event (e.g. "server", "indexer", "backfill")
|
|
23
|
+
* @param op - Operation name (e.g. "request", "commit", "memory")
|
|
24
|
+
* @param fields - High-cardinality key-value context — include everything relevant
|
|
25
|
+
*/
|
|
6
26
|
export function emit(module, op, fields) {
|
|
7
27
|
if (process.env.DEBUG === '0')
|
|
8
28
|
return;
|
|
@@ -17,6 +37,15 @@ export function emit(module, op, fields) {
|
|
|
17
37
|
}
|
|
18
38
|
process.stdout.write(JSON.stringify(entry) + '\n');
|
|
19
39
|
}
|
|
40
|
+
/**
|
|
41
|
+
* Start a millisecond timer. Call the returned function to get elapsed ms.
|
|
42
|
+
* Use with {@link emit} to add `duration_ms` to wide events.
|
|
43
|
+
*
|
|
44
|
+
* @example
|
|
45
|
+
* const elapsed = timer()
|
|
46
|
+
* await doWork()
|
|
47
|
+
* emit('server', 'request', { path, status_code, duration_ms: elapsed() })
|
|
48
|
+
*/
|
|
20
49
|
export function timer() {
|
|
21
50
|
const start = performance.now();
|
|
22
51
|
return () => Math.round(performance.now() - start);
|
package/dist/main.js
CHANGED
|
@@ -1,28 +1,49 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
|
|
2
|
+
var __rewriteRelativeImportExtension = (this && this.__rewriteRelativeImportExtension) || function (path, preserveJsx) {
|
|
3
|
+
if (typeof path === "string" && /^\.\.?\//.test(path)) {
|
|
4
|
+
return path.replace(/\.(tsx)$|((?:\.d)?)((?:\.[^./]+?)?)\.([cm]?)ts$/i, function (m, tsx, d, ext, cm) {
|
|
5
|
+
return tsx ? preserveJsx ? ".jsx" : ".js" : d && (!ext || !cm) ? m : (d + ext + "." + cm.toLowerCase() + "js");
|
|
6
|
+
});
|
|
7
|
+
}
|
|
8
|
+
return path;
|
|
9
|
+
};
|
|
10
|
+
import { mkdirSync, writeFileSync, existsSync } from 'node:fs';
|
|
3
11
|
import { dirname, resolve } from 'node:path';
|
|
12
|
+
import { registerHatkResolveHook } from "./resolve-hatk.js";
|
|
4
13
|
import { log } from "./logger.js";
|
|
5
14
|
import { loadConfig } from "./config.js";
|
|
6
|
-
import { loadLexicons, storeLexicons, discoverCollections,
|
|
15
|
+
import { loadLexicons, storeLexicons, discoverCollections, buildSchemas } from "./database/schema.js";
|
|
7
16
|
import { discoverViews } from "./views.js";
|
|
8
|
-
import { initDatabase, getCursor, querySQL,
|
|
17
|
+
import { initDatabase, getCursor, querySQL, getSqlDialect, getSchemaDump, migrateSchema } from "./database/db.js";
|
|
18
|
+
import { createAdapter } from "./database/adapter-factory.js";
|
|
19
|
+
import { getDialect } from "./database/dialect.js";
|
|
20
|
+
import { setSearchPort } from "./database/fts.js";
|
|
9
21
|
import { initFeeds, listFeeds } from "./feeds.js";
|
|
10
|
-
import { initXrpc, listXrpc, configureRelay } from "./xrpc.js";
|
|
22
|
+
import { initXrpc, listXrpc, configureRelay, callXrpc } from "./xrpc.js";
|
|
11
23
|
import { initOpengraph } from "./opengraph.js";
|
|
12
24
|
import { initLabels, getLabelDefinitions } from "./labels.js";
|
|
13
25
|
import { startIndexer } from "./indexer.js";
|
|
14
|
-
import { rebuildAllIndexes } from "./fts.js";
|
|
15
|
-
import {
|
|
26
|
+
import { rebuildAllIndexes } from "./database/fts.js";
|
|
27
|
+
import { createHandler, registerCoreHandlers } from "./server.js";
|
|
28
|
+
import { serve } from "./adapter.js";
|
|
16
29
|
import { validateLexicons } from '@bigmoves/lexicon';
|
|
17
30
|
import { relayHttpUrl } from "./config.js";
|
|
18
31
|
import { runBackfill } from "./backfill.js";
|
|
19
32
|
import { initOAuth } from "./oauth/server.js";
|
|
20
|
-
import {
|
|
33
|
+
import { parseSessionCookie, getSessionCookieName } from "./oauth/session.js";
|
|
34
|
+
import { loadOnLoginHook } from "./hooks.js";
|
|
21
35
|
import { initSetup } from "./setup.js";
|
|
22
|
-
|
|
36
|
+
import { initServer } from "./server-init.js";
|
|
37
|
+
function logMemory(phase) {
|
|
38
|
+
const mem = process.memoryUsage();
|
|
39
|
+
log(`[mem] ${phase}: heap=${Math.round(mem.heapUsed / 1024 / 1024)}MB rss=${Math.round(mem.rss / 1024 / 1024)}MB external=${Math.round(mem.external / 1024 / 1024)}MB arrayBuffers=${Math.round(mem.arrayBuffers / 1024 / 1024)}MB`);
|
|
40
|
+
}
|
|
41
|
+
const configPath = process.argv[2] || 'hatk.config.ts';
|
|
23
42
|
const configDir = dirname(resolve(configPath));
|
|
43
|
+
registerHatkResolveHook();
|
|
44
|
+
logMemory('startup');
|
|
24
45
|
// 1. Load config
|
|
25
|
-
const config = loadConfig(configPath);
|
|
46
|
+
const config = await loadConfig(configPath);
|
|
26
47
|
configureRelay(config.relay);
|
|
27
48
|
// 2. Load lexicons, validate schemas, and discover collections
|
|
28
49
|
const lexicons = loadLexicons(resolve(configDir, 'lexicons'));
|
|
@@ -44,44 +65,65 @@ if (collections.length === 0) {
|
|
|
44
65
|
log(`[main] Loaded config: ${collections.length} collections`);
|
|
45
66
|
// Discover view defs from lexicons
|
|
46
67
|
discoverViews();
|
|
47
|
-
|
|
48
|
-
const schemas =
|
|
49
|
-
const
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
uri TEXT PRIMARY KEY,
|
|
56
|
-
cid TEXT,
|
|
57
|
-
did TEXT NOT NULL,
|
|
58
|
-
indexed_at TIMESTAMP NOT NULL,
|
|
59
|
-
data JSON
|
|
60
|
-
);
|
|
61
|
-
CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_indexed ON "${nsid}"(indexed_at DESC);
|
|
62
|
-
CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_author ON "${nsid}"(did);`;
|
|
63
|
-
schemas.push({ collection: nsid, tableName: `"${nsid}"`, columns: [], refColumns: [], children: [], unions: [] });
|
|
64
|
-
ddlStatements.push(genericDDL);
|
|
65
|
-
continue;
|
|
68
|
+
const engineDialect = getDialect(config.databaseEngine);
|
|
69
|
+
const { schemas, ddlStatements } = buildSchemas(lexicons, collections, engineDialect);
|
|
70
|
+
for (const s of schemas) {
|
|
71
|
+
if (s.columns.length === 0) {
|
|
72
|
+
log(`[main] No lexicon found for ${s.collection}, using generic JSON storage`);
|
|
73
|
+
}
|
|
74
|
+
else {
|
|
75
|
+
log(`[main] Schema for ${s.collection}: ${s.columns.length} columns, ${s.unions.length} unions`);
|
|
66
76
|
}
|
|
67
|
-
const schema = generateTableSchema(nsid, lexicon, lexicons);
|
|
68
|
-
schemas.push(schema);
|
|
69
|
-
ddlStatements.push(generateCreateTableSQL(schema));
|
|
70
|
-
log(`[main] Schema for ${nsid}: ${schema.columns.length} columns, ${schema.unions.length} unions`);
|
|
71
77
|
}
|
|
72
|
-
// 3. Ensure data directory exists and initialize
|
|
78
|
+
// 3. Ensure data directory exists and initialize database
|
|
73
79
|
if (config.database !== ':memory:') {
|
|
74
80
|
mkdirSync(dirname(config.database), { recursive: true });
|
|
75
81
|
}
|
|
76
|
-
await
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
+
const { adapter, searchPort } = await createAdapter(config.databaseEngine);
|
|
83
|
+
setSearchPort(searchPort);
|
|
84
|
+
await initDatabase(adapter, config.database, schemas, ddlStatements);
|
|
85
|
+
logMemory('after-db-init');
|
|
86
|
+
log(`[main] Database initialized (${config.databaseEngine}, ${config.database === ':memory:' ? 'in-memory' : config.database})`);
|
|
87
|
+
// Auto-migrate schema if lexicons changed
|
|
88
|
+
const migrationChanges = await migrateSchema(schemas);
|
|
89
|
+
if (migrationChanges.length > 0) {
|
|
90
|
+
log(`[main] Applied ${migrationChanges.length} schema migration(s)`);
|
|
91
|
+
}
|
|
92
|
+
// 3b. Run setup hooks, feeds, xrpc, og, labels
|
|
93
|
+
const serverDir = resolve(configDir, 'server');
|
|
94
|
+
if (existsSync(serverDir)) {
|
|
95
|
+
// New: single server/ directory
|
|
96
|
+
await initServer(serverDir);
|
|
97
|
+
}
|
|
98
|
+
else {
|
|
99
|
+
// Legacy: separate directories
|
|
100
|
+
await initSetup(resolve(configDir, 'setup'));
|
|
101
|
+
await loadOnLoginHook(resolve(configDir, 'hooks'));
|
|
102
|
+
await initFeeds(resolve(configDir, 'feeds'));
|
|
103
|
+
log(`[main] Feeds initialized: ${listFeeds()
|
|
104
|
+
.map((f) => f.name)
|
|
105
|
+
.join(', ') || 'none'}`);
|
|
106
|
+
await initXrpc(resolve(configDir, 'xrpc'));
|
|
107
|
+
log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
|
|
108
|
+
await initOpengraph(resolve(configDir, 'og'));
|
|
109
|
+
log(`[main] OpenGraph initialized`);
|
|
110
|
+
await initLabels(resolve(configDir, 'labels'));
|
|
111
|
+
log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
|
|
112
|
+
}
|
|
113
|
+
// Register built-in dev.hatk.* handlers so callXrpc() can find them
|
|
114
|
+
registerCoreHandlers(collections, config.oauth);
|
|
115
|
+
// Write db/schema.sql (after setup, so setup-created tables are included)
|
|
116
|
+
try {
|
|
117
|
+
const schemaDir = resolve(configDir, 'db');
|
|
118
|
+
mkdirSync(schemaDir, { recursive: true });
|
|
119
|
+
const schemaDump = await getSchemaDump();
|
|
120
|
+
writeFileSync(resolve(schemaDir, 'schema.sql'), `-- This file is auto-generated by hatk on startup. Do not edit.\n-- Database engine: ${config.databaseEngine}\n\n${schemaDump}\n`);
|
|
121
|
+
log(`[main] Schema written to db/schema.sql`);
|
|
122
|
+
}
|
|
123
|
+
catch { }
|
|
82
124
|
// Detect orphaned tables
|
|
83
125
|
try {
|
|
84
|
-
const existingTables = await querySQL(
|
|
126
|
+
const existingTables = (await querySQL(getSqlDialect().listTablesQuery));
|
|
85
127
|
for (const row of existingTables) {
|
|
86
128
|
const tableName = row.table_name;
|
|
87
129
|
const isChildTable = collections.some((c) => tableName.startsWith(c + '__'));
|
|
@@ -91,24 +133,57 @@ try {
|
|
|
91
133
|
}
|
|
92
134
|
}
|
|
93
135
|
catch { }
|
|
94
|
-
// 4. Initialize feeds, xrpc handlers, og, labels from directories
|
|
95
|
-
await initFeeds(resolve(configDir, 'feeds'));
|
|
96
|
-
log(`[main] Feeds initialized: ${listFeeds()
|
|
97
|
-
.map((f) => f.name)
|
|
98
|
-
.join(', ') || 'none'}`);
|
|
99
|
-
await initXrpc(resolve(configDir, 'xrpc'));
|
|
100
|
-
log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
|
|
101
|
-
await initOpengraph(resolve(configDir, 'og'));
|
|
102
|
-
log(`[main] OpenGraph initialized`);
|
|
103
|
-
await initLabels(resolve(configDir, 'labels'));
|
|
104
|
-
log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
|
|
105
136
|
if (config.oauth) {
|
|
106
137
|
await initOAuth(config.oauth, config.plc, config.relay);
|
|
107
138
|
log(`[main] OAuth initialized (issuer: ${config.oauth.issuer})`);
|
|
108
139
|
}
|
|
140
|
+
logMemory('before-server');
|
|
109
141
|
// 5. Start server immediately (don't wait for backfill)
|
|
110
142
|
const collectionSet = new Set(collections);
|
|
111
|
-
|
|
143
|
+
const backfillOpts = {
|
|
144
|
+
pdsUrl: relayHttpUrl(config.relay),
|
|
145
|
+
plcUrl: config.plc,
|
|
146
|
+
collections: collectionSet,
|
|
147
|
+
config: config.backfill,
|
|
148
|
+
};
|
|
149
|
+
function runBackfillAndRestart() {
|
|
150
|
+
runBackfill(backfillOpts)
|
|
151
|
+
.then(async (recordCount) => {
|
|
152
|
+
log('[main] Backfill complete, building FTS indexes...');
|
|
153
|
+
await rebuildAllIndexes(collections);
|
|
154
|
+
log('[main] FTS indexes ready');
|
|
155
|
+
return recordCount;
|
|
156
|
+
})
|
|
157
|
+
.then((recordCount) => {
|
|
158
|
+
if (recordCount > 0 && !process.env.DEV_MODE) {
|
|
159
|
+
logMemory('after-backfill');
|
|
160
|
+
log('[main] Restarting to reclaim memory...');
|
|
161
|
+
process.exit(1);
|
|
162
|
+
}
|
|
163
|
+
})
|
|
164
|
+
.catch((err) => {
|
|
165
|
+
console.error('[main] Backfill error:', err.message);
|
|
166
|
+
});
|
|
167
|
+
}
|
|
168
|
+
const handler = createHandler({
|
|
169
|
+
collections,
|
|
170
|
+
publicDir: config.publicDir,
|
|
171
|
+
oauth: config.oauth,
|
|
172
|
+
admins: config.admins,
|
|
173
|
+
onResync: runBackfillAndRestart,
|
|
174
|
+
});
|
|
175
|
+
globalThis.__hatk_callXrpc = callXrpc;
|
|
176
|
+
globalThis.__hatk_parseSessionCookie = parseSessionCookie;
|
|
177
|
+
globalThis.__hatk_sessionCookieName = getSessionCookieName();
|
|
178
|
+
// Detect SvelteKit build output and use it as fallback handler
|
|
179
|
+
let fallback = undefined;
|
|
180
|
+
const sveltekitHandler = resolve(configDir, 'build', 'handler.js');
|
|
181
|
+
if (existsSync(sveltekitHandler)) {
|
|
182
|
+
const sk = await import(__rewriteRelativeImportExtension(/* @vite-ignore */ sveltekitHandler));
|
|
183
|
+
fallback = sk.handler;
|
|
184
|
+
log(`[main] SvelteKit handler loaded from build/handler.js`);
|
|
185
|
+
}
|
|
186
|
+
serve(handler, config.port, undefined, fallback);
|
|
112
187
|
log(`\nhatk running:`);
|
|
113
188
|
log(` Relay: ${config.relay}`);
|
|
114
189
|
log(` Database: ${config.database}`);
|
|
@@ -117,6 +192,7 @@ log(` Collections: ${collections.join(', ')}`);
|
|
|
117
192
|
log(` Feeds: ${listFeeds()
|
|
118
193
|
.map((f) => f.name)
|
|
119
194
|
.join(', ')}`);
|
|
195
|
+
logMemory('after-server');
|
|
120
196
|
// 6. Start indexer with cursor
|
|
121
197
|
const cursor = await getCursor('relay');
|
|
122
198
|
startIndexer({
|
|
@@ -127,22 +203,13 @@ startIndexer({
|
|
|
127
203
|
cursor,
|
|
128
204
|
fetchTimeout: config.backfill.fetchTimeout,
|
|
129
205
|
maxRetries: config.backfill.maxRetries,
|
|
206
|
+
parallelism: config.backfill.parallelism,
|
|
130
207
|
ftsRebuildInterval: config.ftsRebuildInterval,
|
|
131
208
|
});
|
|
132
209
|
// 7. Run backfill in background
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
136
|
-
|
|
137
|
-
|
|
138
|
-
})
|
|
139
|
-
.then(() => {
|
|
140
|
-
log('[main] Backfill complete, rebuilding FTS indexes...');
|
|
141
|
-
return rebuildAllIndexes(collections);
|
|
142
|
-
})
|
|
143
|
-
.then(() => {
|
|
144
|
-
log('[main] FTS indexes ready');
|
|
145
|
-
})
|
|
146
|
-
.catch((err) => {
|
|
147
|
-
console.error('[main] Backfill error:', err.message);
|
|
210
|
+
runBackfillAndRestart();
|
|
211
|
+
// Graceful shutdown
|
|
212
|
+
process.on('SIGTERM', () => {
|
|
213
|
+
log('[main] Received SIGTERM, shutting down...');
|
|
214
|
+
process.exit(0);
|
|
148
215
|
});
|
package/dist/mst.d.ts
CHANGED
|
@@ -1,6 +1,23 @@
|
|
|
1
|
+
/** A single entry from a Merkle Search Tree — a record path paired with its content CID. */
|
|
1
2
|
export interface MstEntry {
|
|
3
|
+
/** Record path, e.g. "xyz.marketplace.listing/3mfniulnr7c2g" */
|
|
2
4
|
path: string;
|
|
5
|
+
/** CID of the record's CBOR block */
|
|
3
6
|
cid: string;
|
|
4
7
|
}
|
|
5
|
-
|
|
8
|
+
/**
|
|
9
|
+
* Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
|
|
10
|
+
*
|
|
11
|
+
* The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
|
|
12
|
+
* record paths to CIDs. Each node contains a left subtree pointer, an array of entries
|
|
13
|
+
* (each with a prefix length, key suffix, value CID, and right subtree pointer), and
|
|
14
|
+
* keys are reconstructed by combining the prefix of the previous key with the suffix.
|
|
15
|
+
*
|
|
16
|
+
* @param blocks - Block store that resolves CIDs to raw CBOR bytes
|
|
17
|
+
* @param rootCid - CID of the MST root node
|
|
18
|
+
* @yields {MstEntry} Record entries in lexicographic key order
|
|
19
|
+
*/
|
|
20
|
+
export declare function walkMst(blocks: {
|
|
21
|
+
get(cid: string): Uint8Array | undefined;
|
|
22
|
+
}, rootCid: string): Generator<MstEntry>;
|
|
6
23
|
//# sourceMappingURL=mst.d.ts.map
|
package/dist/mst.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,MAAM,WAAW,QAAQ;IACvB,IAAI,EAAE,MAAM,CAAA;IACZ,GAAG,EAAE,MAAM,CAAA;CACZ;AAED,
|
|
1
|
+
{"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,4FAA4F;AAC5F,MAAM,WAAW,QAAQ;IACvB,gEAAgE;IAChE,IAAI,EAAE,MAAM,CAAA;IACZ,qCAAqC;IACrC,GAAG,EAAE,MAAM,CAAA;CACZ;AAED;;;;;;;;;;;GAWG;AACH,wBAAiB,OAAO,CAAC,MAAM,EAAE;IAAE,GAAG,CAAC,GAAG,EAAE,MAAM,GAAG,UAAU,GAAG,SAAS,CAAA;CAAE,EAAE,OAAO,EAAE,MAAM,GAAG,SAAS,CAAC,QAAQ,CAAC,CA+BnH"}
|
package/dist/mst.js
CHANGED
|
@@ -1,14 +1,26 @@
|
|
|
1
1
|
import { cborDecode } from "./cbor.js";
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
2
|
+
/**
|
|
3
|
+
* Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
|
|
4
|
+
*
|
|
5
|
+
* The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
|
|
6
|
+
* record paths to CIDs. Each node contains a left subtree pointer, an array of entries
|
|
7
|
+
* (each with a prefix length, key suffix, value CID, and right subtree pointer), and
|
|
8
|
+
* keys are reconstructed by combining the prefix of the previous key with the suffix.
|
|
9
|
+
*
|
|
10
|
+
* @param blocks - Block store that resolves CIDs to raw CBOR bytes
|
|
11
|
+
* @param rootCid - CID of the MST root node
|
|
12
|
+
* @yields {MstEntry} Record entries in lexicographic key order
|
|
13
|
+
*/
|
|
14
|
+
export function* walkMst(blocks, rootCid) {
|
|
15
|
+
/** Recursively visit an MST node, reconstructing full keys from prefix-compressed entries. */
|
|
16
|
+
function* visit(cid, prefix) {
|
|
5
17
|
const data = blocks.get(cid);
|
|
6
18
|
if (!data)
|
|
7
19
|
return prefix;
|
|
8
20
|
const { value: node } = cborDecode(data);
|
|
9
21
|
// Visit left subtree
|
|
10
22
|
if (node.l?.$link)
|
|
11
|
-
visit(node.l.$link, prefix);
|
|
23
|
+
yield* visit(node.l.$link, prefix);
|
|
12
24
|
let lastKey = prefix;
|
|
13
25
|
for (const entry of node.e || []) {
|
|
14
26
|
const keySuffix = entry.k instanceof Uint8Array ? new TextDecoder().decode(entry.k) : entry.k;
|
|
@@ -16,15 +28,14 @@ export function walkMst(blocks, rootCid) {
|
|
|
16
28
|
const fullKey = lastKey.substring(0, prefixLen) + keySuffix;
|
|
17
29
|
lastKey = fullKey;
|
|
18
30
|
if (entry.v?.$link) {
|
|
19
|
-
|
|
31
|
+
yield { path: fullKey, cid: entry.v.$link };
|
|
20
32
|
}
|
|
21
33
|
// Visit right subtree
|
|
22
34
|
if (entry.t?.$link) {
|
|
23
|
-
visit(entry.t.$link, lastKey);
|
|
35
|
+
yield* visit(entry.t.$link, lastKey);
|
|
24
36
|
}
|
|
25
37
|
}
|
|
26
38
|
return lastKey;
|
|
27
39
|
}
|
|
28
|
-
visit(rootCid, '');
|
|
29
|
-
return entries;
|
|
40
|
+
yield* visit(rootCid, '');
|
|
30
41
|
}
|
package/dist/oauth/db.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,
|
|
1
|
+
{"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMtG;AAID,wBAAsB,iBAAiB,CACrC,UAAU,EAAE,MAAM,EAClB,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,WAAW,EAAE,MAAM,CAAA;IACnB,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,aAAa,EAAE,MAAM,CAAA;IACrB,mBAAmB,CAAC,EAAE,MAAM,CAAA;IAC5B,OAAO,EAAE,MAAM,CAAA;IACf,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,QAAQ,CAAC,EAAE,MAAM,CAAA;IACjB,GAAG,CAAC,EAAE,MAAM,CAAA;IACZ,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,SAAS,EAAE,MAAM,CAAA;CAClB,GACA,OAAO,CAAC,IAAI,CAAC,CAsBf;AAED,wBAAsB,eAAe,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAM7E;AAED,wBAAsB,kBAAkB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE1E;AAID,wBAAsB,aAAa,CAAC,IAAI,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMnF;AAED,wBAAsB,eAAe,CAAC,IAAI,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAK1E;AAID,wBAAsB,YAAY,CAChC,GAAG,EAAE,MAAM,EACX,IAAI,EAAE;IACJ,WAAW,EAAE,MAAM,CAAA;IACnB,WAAW,EAAE,MAAM,CAAA;IACnB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,OAAO,EAAE,MAAM,CAAA;IACf,cAAc,CAAC,EAAE,MAAM,CAAA;CACxB,GACA,OAAO,CAAC,IAAI,CAAC,CAMf;AAED,wBAAsB,UAAU,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGjE;AAED,wBAAsB,aAAa,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE9D;AAID,wBAAsB,iBAAiB,CACrC,KAAK,EAAE,MAAM,EACb,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;IACX,OAAO,EAAE,MAAM,CAAA;IACf,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,SAAS,CAAC,EAAE,MAAM,CAAA;CACnB,GACA,OAAO,CAAC,IAAI,CAAC,CAQf;AAED,wBAAsB,eAAe,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGxE;AAED,wBAAsB,kBAAkB,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAErE;AAID,wBAAsB,oBAAoB,CAAC,GAAG,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,OAAO,CAAC,CAK3F;AAED,wBAAsB,mBAAmB,IAAI,OAAO,CAAC,IAAI,CAAC,CAQzD"}
|