@hatk/hatk 0.0.1-alpha.3 → 0.0.1-alpha.30

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (132) hide show
  1. package/dist/adapter.d.ts +19 -0
  2. package/dist/adapter.d.ts.map +1 -0
  3. package/dist/adapter.js +94 -0
  4. package/dist/backfill.d.ts +60 -1
  5. package/dist/backfill.d.ts.map +1 -1
  6. package/dist/backfill.js +166 -32
  7. package/dist/car.d.ts +59 -1
  8. package/dist/car.d.ts.map +1 -1
  9. package/dist/car.js +179 -7
  10. package/dist/cbor.d.ts +37 -0
  11. package/dist/cbor.d.ts.map +1 -1
  12. package/dist/cbor.js +36 -3
  13. package/dist/cid.d.ts +37 -0
  14. package/dist/cid.d.ts.map +1 -1
  15. package/dist/cid.js +38 -3
  16. package/dist/cli.js +356 -123
  17. package/dist/config.d.ts +12 -1
  18. package/dist/config.d.ts.map +1 -1
  19. package/dist/config.js +36 -9
  20. package/dist/database/adapter-factory.d.ts +6 -0
  21. package/dist/database/adapter-factory.d.ts.map +1 -0
  22. package/dist/database/adapter-factory.js +20 -0
  23. package/dist/database/adapters/duckdb-search.d.ts +12 -0
  24. package/dist/database/adapters/duckdb-search.d.ts.map +1 -0
  25. package/dist/database/adapters/duckdb-search.js +27 -0
  26. package/dist/database/adapters/duckdb.d.ts +25 -0
  27. package/dist/database/adapters/duckdb.d.ts.map +1 -0
  28. package/dist/database/adapters/duckdb.js +161 -0
  29. package/dist/database/adapters/sqlite-search.d.ts +18 -0
  30. package/dist/database/adapters/sqlite-search.d.ts.map +1 -0
  31. package/dist/database/adapters/sqlite-search.js +38 -0
  32. package/dist/database/adapters/sqlite.d.ts +18 -0
  33. package/dist/database/adapters/sqlite.d.ts.map +1 -0
  34. package/dist/database/adapters/sqlite.js +87 -0
  35. package/dist/database/db.d.ts +149 -0
  36. package/dist/database/db.d.ts.map +1 -0
  37. package/dist/database/db.js +1460 -0
  38. package/dist/database/dialect.d.ts +45 -0
  39. package/dist/database/dialect.d.ts.map +1 -0
  40. package/dist/database/dialect.js +72 -0
  41. package/dist/database/fts.d.ts +24 -0
  42. package/dist/database/fts.d.ts.map +1 -0
  43. package/dist/database/fts.js +777 -0
  44. package/dist/database/index.d.ts +7 -0
  45. package/dist/database/index.d.ts.map +1 -0
  46. package/dist/database/index.js +6 -0
  47. package/dist/database/ports.d.ts +44 -0
  48. package/dist/database/ports.d.ts.map +1 -0
  49. package/dist/database/ports.js +1 -0
  50. package/dist/database/schema.d.ts +60 -0
  51. package/dist/database/schema.d.ts.map +1 -0
  52. package/dist/database/schema.js +388 -0
  53. package/dist/db.d.ts +1 -1
  54. package/dist/db.d.ts.map +1 -1
  55. package/dist/db.js +4 -38
  56. package/dist/dev-entry.d.ts +8 -0
  57. package/dist/dev-entry.d.ts.map +1 -0
  58. package/dist/dev-entry.js +109 -0
  59. package/dist/feeds.d.ts +4 -0
  60. package/dist/feeds.d.ts.map +1 -1
  61. package/dist/feeds.js +42 -3
  62. package/dist/fts.d.ts.map +1 -1
  63. package/dist/fts.js +5 -0
  64. package/dist/hooks.d.ts +22 -0
  65. package/dist/hooks.d.ts.map +1 -0
  66. package/dist/hooks.js +75 -0
  67. package/dist/hydrate.js +1 -1
  68. package/dist/indexer.d.ts +20 -0
  69. package/dist/indexer.d.ts.map +1 -1
  70. package/dist/indexer.js +48 -6
  71. package/dist/labels.d.ts +34 -0
  72. package/dist/labels.d.ts.map +1 -1
  73. package/dist/labels.js +63 -3
  74. package/dist/logger.d.ts +29 -0
  75. package/dist/logger.d.ts.map +1 -1
  76. package/dist/logger.js +29 -0
  77. package/dist/main.js +131 -67
  78. package/dist/mst.d.ts +18 -1
  79. package/dist/mst.d.ts.map +1 -1
  80. package/dist/mst.js +19 -8
  81. package/dist/oauth/db.d.ts.map +1 -1
  82. package/dist/oauth/db.js +41 -15
  83. package/dist/oauth/server.d.ts +2 -0
  84. package/dist/oauth/server.d.ts.map +1 -1
  85. package/dist/oauth/server.js +102 -7
  86. package/dist/oauth/session.d.ts +9 -0
  87. package/dist/oauth/session.d.ts.map +1 -0
  88. package/dist/oauth/session.js +65 -0
  89. package/dist/opengraph.d.ts +10 -0
  90. package/dist/opengraph.d.ts.map +1 -1
  91. package/dist/opengraph.js +103 -5
  92. package/dist/pds-proxy.d.ts +39 -0
  93. package/dist/pds-proxy.d.ts.map +1 -0
  94. package/dist/pds-proxy.js +173 -0
  95. package/dist/renderer.d.ts +27 -0
  96. package/dist/renderer.d.ts.map +1 -0
  97. package/dist/renderer.js +46 -0
  98. package/dist/resolve-hatk.d.ts +6 -0
  99. package/dist/resolve-hatk.d.ts.map +1 -0
  100. package/dist/resolve-hatk.js +20 -0
  101. package/dist/response.d.ts +16 -0
  102. package/dist/response.d.ts.map +1 -0
  103. package/dist/response.js +69 -0
  104. package/dist/scanner.d.ts +21 -0
  105. package/dist/scanner.d.ts.map +1 -0
  106. package/dist/scanner.js +88 -0
  107. package/dist/schema.d.ts +8 -0
  108. package/dist/schema.d.ts.map +1 -1
  109. package/dist/schema.js +29 -0
  110. package/dist/seed.d.ts +19 -0
  111. package/dist/seed.d.ts.map +1 -1
  112. package/dist/seed.js +43 -4
  113. package/dist/server-init.d.ts +8 -0
  114. package/dist/server-init.d.ts.map +1 -0
  115. package/dist/server-init.js +59 -0
  116. package/dist/server.d.ts +26 -3
  117. package/dist/server.d.ts.map +1 -1
  118. package/dist/server.js +487 -616
  119. package/dist/setup.d.ts +28 -1
  120. package/dist/setup.d.ts.map +1 -1
  121. package/dist/setup.js +50 -3
  122. package/dist/test.d.ts +1 -1
  123. package/dist/test.d.ts.map +1 -1
  124. package/dist/test.js +38 -32
  125. package/dist/views.js +1 -1
  126. package/dist/vite-plugin.d.ts +1 -1
  127. package/dist/vite-plugin.d.ts.map +1 -1
  128. package/dist/vite-plugin.js +252 -66
  129. package/dist/xrpc.d.ts +36 -0
  130. package/dist/xrpc.d.ts.map +1 -1
  131. package/dist/xrpc.js +124 -3
  132. package/package.json +12 -5
package/dist/labels.js CHANGED
@@ -6,13 +6,53 @@ var __rewriteRelativeImportExtension = (this && this.__rewriteRelativeImportExte
6
6
  }
7
7
  return path;
8
8
  };
9
+ /**
10
+ * Label system for applying moderation labels to records as they are indexed.
11
+ *
12
+ * Place label modules in the `labels/` directory. Each module default-exports
13
+ * an object with a `definition` (label metadata) and/or an `evaluate` function
14
+ * (rule that returns label values for a given record).
15
+ *
16
+ * @example
17
+ * ```ts
18
+ * // labels/nsfw.ts
19
+ * import type { LabelRuleContext } from '@hatk/hatk/labels'
20
+ *
21
+ * export default {
22
+ * definition: {
23
+ * identifier: 'nsfw',
24
+ * severity: 'alert',
25
+ * blurs: 'media',
26
+ * defaultSetting: 'warn',
27
+ * locales: [{ lang: 'en', name: 'NSFW', description: 'Not safe for work' }],
28
+ * },
29
+ *
30
+ * async evaluate(ctx: LabelRuleContext): Promise<string[]> {
31
+ * if (ctx.record.value.nsfw === true) return ['nsfw']
32
+ * return []
33
+ * },
34
+ * }
35
+ * ```
36
+ */
9
37
  import { resolve } from 'node:path';
10
38
  import { readdirSync } from 'node:fs';
11
- import { querySQL, runSQL, insertLabels, getSchema } from "./db.js";
39
+ import { querySQL, runSQL, insertLabels, getSchema } from "./database/db.js";
12
40
  import { log, emit } from "./logger.js";
41
+ export function defineLabels(module) {
42
+ return { __type: 'labels', ...module };
43
+ }
13
44
  const rules = [];
14
45
  let labelDefs = [];
15
46
  let labelSrc = 'self';
47
+ /**
48
+ * Discover and load label rule modules from the `labels/` directory.
49
+ *
50
+ * Each module should default-export an object with an optional `definition`
51
+ * (label metadata like severity and blur behavior) and an optional `evaluate`
52
+ * function that returns label values to apply to a record.
53
+ *
54
+ * @param labelsDir - Absolute path to the `labels/` directory
55
+ */
16
56
  export async function initLabels(labelsDir) {
17
57
  let files;
18
58
  try {
@@ -26,7 +66,7 @@ export async function initLabels(labelsDir) {
26
66
  for (const file of files) {
27
67
  const name = file.replace(/\.(ts|js)$/, '');
28
68
  const scriptPath = resolve(labelsDir, file);
29
- const mod = await import(__rewriteRelativeImportExtension(scriptPath));
69
+ const mod = await import(__rewriteRelativeImportExtension(/* @vite-ignore */ `${scriptPath}?t=${Date.now()}`));
30
70
  const handler = mod.default;
31
71
  if (handler.definition) {
32
72
  labelDefs.push(handler.definition);
@@ -45,6 +85,19 @@ export async function initLabels(labelsDir) {
45
85
  log(`[labels] ${labelDefs.length} label definitions loaded`);
46
86
  }
47
87
  }
88
+ /** Register a single label module from a scanned server/ module. */
89
+ export function registerLabelModule(name, labelMod) {
90
+ if (labelMod.definition) {
91
+ labelDefs.push(labelMod.definition);
92
+ }
93
+ if (labelMod.evaluate) {
94
+ rules.push({ name, evaluate: labelMod.evaluate });
95
+ }
96
+ }
97
+ /**
98
+ * Evaluate all loaded label rules against a record and persist any resulting labels.
99
+ * Called after each record is indexed. Rule errors are logged but never block indexing.
100
+ */
48
101
  export async function runLabelRules(record) {
49
102
  if (rules.length === 0)
50
103
  return;
@@ -69,6 +122,12 @@ export async function runLabelRules(record) {
69
122
  emit('labels', 'applied', { count: allLabels.length, uri: record.uri, vals: allLabels.map((l) => l.val) });
70
123
  }
71
124
  }
125
+ /**
126
+ * Re-evaluate all label rules against every existing record in the given collections.
127
+ * Used by `/admin/rescan-labels` to apply new or updated rules retroactively.
128
+ *
129
+ * @returns Count of records scanned and new labels applied
130
+ */
72
131
  export async function rescanLabels(collections) {
73
132
  const beforeRows = await querySQL(`SELECT COUNT(*) as count FROM _labels`);
74
133
  const beforeCount = Number(beforeRows[0]?.count || 0);
@@ -85,7 +144,7 @@ export async function rescanLabels(collections) {
85
144
  let v = row[col.name];
86
145
  if (v === null || v === undefined)
87
146
  continue;
88
- if (col.duckdbType === 'JSON' && typeof v === 'string') {
147
+ if (col.sqlType === 'JSON' && typeof v === 'string') {
89
148
  try {
90
149
  v = JSON.parse(v);
91
150
  }
@@ -106,6 +165,7 @@ export async function rescanLabels(collections) {
106
165
  const afterCount = Number(afterRows[0]?.count || 0);
107
166
  return { scanned, labeled: afterCount - beforeCount };
108
167
  }
168
+ /** Return all label definitions discovered during {@link initLabels}. */
109
169
  export function getLabelDefinitions() {
110
170
  return labelDefs;
111
171
  }
package/dist/logger.d.ts CHANGED
@@ -1,4 +1,33 @@
1
+ /**
2
+ * Unstructured debug log — use sparingly for human-readable dev output.
3
+ * Prefer {@link emit} for anything that should be queryable in production.
4
+ * Disabled when `DEBUG=0`.
5
+ */
1
6
  export declare function log(...args: unknown[]): void;
7
+ /**
8
+ * Emit a structured wide event as a single JSON line to stdout.
9
+ *
10
+ * Each call produces one canonical log line with a timestamp, module, operation,
11
+ * and arbitrary key-value fields — designed for columnar search and aggregation,
12
+ * not string grep. Pack as much context as possible into `fields` (request IDs,
13
+ * durations, status codes, user DIDs, counts) so a single event tells the full
14
+ * story. See https://loggingsucks.com for the philosophy behind this approach.
15
+ *
16
+ * Disabled when `DEBUG=0`.
17
+ *
18
+ * @param module - Subsystem emitting the event (e.g. "server", "indexer", "backfill")
19
+ * @param op - Operation name (e.g. "request", "commit", "memory")
20
+ * @param fields - High-cardinality key-value context — include everything relevant
21
+ */
2
22
  export declare function emit(module: string, op: string, fields: Record<string, unknown>): void;
23
+ /**
24
+ * Start a millisecond timer. Call the returned function to get elapsed ms.
25
+ * Use with {@link emit} to add `duration_ms` to wide events.
26
+ *
27
+ * @example
28
+ * const elapsed = timer()
29
+ * await doWork()
30
+ * emit('server', 'request', { path, status_code, duration_ms: elapsed() })
31
+ */
3
32
  export declare function timer(): () => number;
4
33
  //# sourceMappingURL=logger.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"logger.d.ts","sourceRoot":"","sources":["../src/logger.ts"],"names":[],"mappings":"AAAA,wBAAgB,GAAG,CAAC,GAAG,IAAI,EAAE,OAAO,EAAE,GAAG,IAAI,CAG5C;AAED,wBAAgB,IAAI,CAAC,MAAM,EAAE,MAAM,EAAE,EAAE,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI,CAWtF;AAED,wBAAgB,KAAK,IAAI,MAAM,MAAM,CAGpC"}
1
+ {"version":3,"file":"logger.d.ts","sourceRoot":"","sources":["../src/logger.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AACH,wBAAgB,GAAG,CAAC,GAAG,IAAI,EAAE,OAAO,EAAE,GAAG,IAAI,CAG5C;AAED;;;;;;;;;;;;;;GAcG;AACH,wBAAgB,IAAI,CAAC,MAAM,EAAE,MAAM,EAAE,EAAE,EAAE,MAAM,EAAE,MAAM,EAAE,MAAM,CAAC,MAAM,EAAE,OAAO,CAAC,GAAG,IAAI,CAWtF;AAED;;;;;;;;GAQG;AACH,wBAAgB,KAAK,IAAI,MAAM,MAAM,CAGpC"}
package/dist/logger.js CHANGED
@@ -1,8 +1,28 @@
1
+ /**
2
+ * Unstructured debug log — use sparingly for human-readable dev output.
3
+ * Prefer {@link emit} for anything that should be queryable in production.
4
+ * Disabled when `DEBUG=0`.
5
+ */
1
6
  export function log(...args) {
2
7
  if (process.env.DEBUG === '0')
3
8
  return;
4
9
  console.log(...args);
5
10
  }
11
+ /**
12
+ * Emit a structured wide event as a single JSON line to stdout.
13
+ *
14
+ * Each call produces one canonical log line with a timestamp, module, operation,
15
+ * and arbitrary key-value fields — designed for columnar search and aggregation,
16
+ * not string grep. Pack as much context as possible into `fields` (request IDs,
17
+ * durations, status codes, user DIDs, counts) so a single event tells the full
18
+ * story. See https://loggingsucks.com for the philosophy behind this approach.
19
+ *
20
+ * Disabled when `DEBUG=0`.
21
+ *
22
+ * @param module - Subsystem emitting the event (e.g. "server", "indexer", "backfill")
23
+ * @param op - Operation name (e.g. "request", "commit", "memory")
24
+ * @param fields - High-cardinality key-value context — include everything relevant
25
+ */
6
26
  export function emit(module, op, fields) {
7
27
  if (process.env.DEBUG === '0')
8
28
  return;
@@ -17,6 +37,15 @@ export function emit(module, op, fields) {
17
37
  }
18
38
  process.stdout.write(JSON.stringify(entry) + '\n');
19
39
  }
40
+ /**
41
+ * Start a millisecond timer. Call the returned function to get elapsed ms.
42
+ * Use with {@link emit} to add `duration_ms` to wide events.
43
+ *
44
+ * @example
45
+ * const elapsed = timer()
46
+ * await doWork()
47
+ * emit('server', 'request', { path, status_code, duration_ms: elapsed() })
48
+ */
20
49
  export function timer() {
21
50
  const start = performance.now();
22
51
  return () => Math.round(performance.now() - start);
package/dist/main.js CHANGED
@@ -1,28 +1,49 @@
1
1
  #!/usr/bin/env node
2
- import { mkdirSync } from 'node:fs';
2
+ var __rewriteRelativeImportExtension = (this && this.__rewriteRelativeImportExtension) || function (path, preserveJsx) {
3
+ if (typeof path === "string" && /^\.\.?\//.test(path)) {
4
+ return path.replace(/\.(tsx)$|((?:\.d)?)((?:\.[^./]+?)?)\.([cm]?)ts$/i, function (m, tsx, d, ext, cm) {
5
+ return tsx ? preserveJsx ? ".jsx" : ".js" : d && (!ext || !cm) ? m : (d + ext + "." + cm.toLowerCase() + "js");
6
+ });
7
+ }
8
+ return path;
9
+ };
10
+ import { mkdirSync, writeFileSync, existsSync } from 'node:fs';
3
11
  import { dirname, resolve } from 'node:path';
12
+ import { registerHatkResolveHook } from "./resolve-hatk.js";
4
13
  import { log } from "./logger.js";
5
14
  import { loadConfig } from "./config.js";
6
- import { loadLexicons, storeLexicons, discoverCollections, generateTableSchema, generateCreateTableSQL, } from "./schema.js";
15
+ import { loadLexicons, storeLexicons, discoverCollections, buildSchemas } from "./database/schema.js";
7
16
  import { discoverViews } from "./views.js";
8
- import { initDatabase, getCursor, querySQL, backfillChildTables } from "./db.js";
17
+ import { initDatabase, getCursor, querySQL, getSqlDialect, getSchemaDump, migrateSchema } from "./database/db.js";
18
+ import { createAdapter } from "./database/adapter-factory.js";
19
+ import { getDialect } from "./database/dialect.js";
20
+ import { setSearchPort } from "./database/fts.js";
9
21
  import { initFeeds, listFeeds } from "./feeds.js";
10
- import { initXrpc, listXrpc, configureRelay } from "./xrpc.js";
22
+ import { initXrpc, listXrpc, configureRelay, callXrpc } from "./xrpc.js";
11
23
  import { initOpengraph } from "./opengraph.js";
12
24
  import { initLabels, getLabelDefinitions } from "./labels.js";
13
25
  import { startIndexer } from "./indexer.js";
14
- import { rebuildAllIndexes } from "./fts.js";
15
- import { startServer } from "./server.js";
26
+ import { rebuildAllIndexes } from "./database/fts.js";
27
+ import { createHandler, registerCoreHandlers } from "./server.js";
28
+ import { serve } from "./adapter.js";
16
29
  import { validateLexicons } from '@bigmoves/lexicon';
17
30
  import { relayHttpUrl } from "./config.js";
18
31
  import { runBackfill } from "./backfill.js";
19
32
  import { initOAuth } from "./oauth/server.js";
20
- import { loadOnLoginHook } from "./oauth/hooks.js";
33
+ import { parseSessionCookie, getSessionCookieName } from "./oauth/session.js";
34
+ import { loadOnLoginHook } from "./hooks.js";
21
35
  import { initSetup } from "./setup.js";
22
- const configPath = process.argv[2] || 'config.yaml';
36
+ import { initServer } from "./server-init.js";
37
+ function logMemory(phase) {
38
+ const mem = process.memoryUsage();
39
+ log(`[mem] ${phase}: heap=${Math.round(mem.heapUsed / 1024 / 1024)}MB rss=${Math.round(mem.rss / 1024 / 1024)}MB external=${Math.round(mem.external / 1024 / 1024)}MB arrayBuffers=${Math.round(mem.arrayBuffers / 1024 / 1024)}MB`);
40
+ }
41
+ const configPath = process.argv[2] || 'hatk.config.ts';
23
42
  const configDir = dirname(resolve(configPath));
43
+ registerHatkResolveHook();
44
+ logMemory('startup');
24
45
  // 1. Load config
25
- const config = loadConfig(configPath);
46
+ const config = await loadConfig(configPath);
26
47
  configureRelay(config.relay);
27
48
  // 2. Load lexicons, validate schemas, and discover collections
28
49
  const lexicons = loadLexicons(resolve(configDir, 'lexicons'));
@@ -44,44 +65,63 @@ if (collections.length === 0) {
44
65
  log(`[main] Loaded config: ${collections.length} collections`);
45
66
  // Discover view defs from lexicons
46
67
  discoverViews();
47
- await loadOnLoginHook(resolve(configDir, 'hooks'));
48
- const schemas = [];
49
- const ddlStatements = [];
50
- for (const nsid of collections) {
51
- const lexicon = lexicons.get(nsid);
52
- if (!lexicon) {
53
- log(`[main] No lexicon found for ${nsid}, using generic JSON storage`);
54
- const genericDDL = `CREATE TABLE IF NOT EXISTS "${nsid}" (
55
- uri TEXT PRIMARY KEY,
56
- cid TEXT,
57
- did TEXT NOT NULL,
58
- indexed_at TIMESTAMP NOT NULL,
59
- data JSON
60
- );
61
- CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_indexed ON "${nsid}"(indexed_at DESC);
62
- CREATE INDEX IF NOT EXISTS idx_${nsid.replace(/\./g, '_')}_author ON "${nsid}"(did);`;
63
- schemas.push({ collection: nsid, tableName: `"${nsid}"`, columns: [], refColumns: [], children: [], unions: [] });
64
- ddlStatements.push(genericDDL);
65
- continue;
68
+ const engineDialect = getDialect(config.databaseEngine);
69
+ const { schemas, ddlStatements } = buildSchemas(lexicons, collections, engineDialect);
70
+ for (const s of schemas) {
71
+ if (s.columns.length === 0) {
72
+ log(`[main] No lexicon found for ${s.collection}, using generic JSON storage`);
73
+ }
74
+ else {
75
+ log(`[main] Schema for ${s.collection}: ${s.columns.length} columns, ${s.unions.length} unions`);
66
76
  }
67
- const schema = generateTableSchema(nsid, lexicon, lexicons);
68
- schemas.push(schema);
69
- ddlStatements.push(generateCreateTableSQL(schema));
70
- log(`[main] Schema for ${nsid}: ${schema.columns.length} columns, ${schema.unions.length} unions`);
71
77
  }
72
- // 3. Ensure data directory exists and initialize DuckDB
78
+ // 3. Ensure data directory exists and initialize database
73
79
  if (config.database !== ':memory:') {
74
80
  mkdirSync(dirname(config.database), { recursive: true });
75
81
  }
76
- await initDatabase(config.database, schemas, ddlStatements);
77
- log(`[main] DuckDB initialized (${config.database === ':memory:' ? 'in-memory' : config.database})`);
78
- // 3a. Backfill child tables for decomposed arrays (one-time migration)
79
- await backfillChildTables();
80
- // 3b. Run setup hooks (after DB init, before server)
81
- await initSetup(resolve(configDir, 'setup'));
82
+ const { adapter, searchPort } = await createAdapter(config.databaseEngine);
83
+ setSearchPort(searchPort);
84
+ await initDatabase(adapter, config.database, schemas, ddlStatements);
85
+ logMemory('after-db-init');
86
+ log(`[main] Database initialized (${config.databaseEngine}, ${config.database === ':memory:' ? 'in-memory' : config.database})`);
87
+ // Auto-migrate schema if lexicons changed
88
+ const migrationChanges = await migrateSchema(schemas);
89
+ if (migrationChanges.length > 0) {
90
+ log(`[main] Applied ${migrationChanges.length} schema migration(s)`);
91
+ }
92
+ // 3b. Run setup hooks, feeds, xrpc, og, labels
93
+ const serverDir = resolve(configDir, 'server');
94
+ if (existsSync(serverDir)) {
95
+ // New: single server/ directory
96
+ await initServer(serverDir);
97
+ }
98
+ else {
99
+ // Legacy: separate directories
100
+ await initSetup(resolve(configDir, 'setup'));
101
+ await loadOnLoginHook(resolve(configDir, 'hooks'));
102
+ await initFeeds(resolve(configDir, 'feeds'));
103
+ log(`[main] Feeds initialized: ${listFeeds().map((f) => f.name).join(', ') || 'none'}`);
104
+ await initXrpc(resolve(configDir, 'xrpc'));
105
+ log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
106
+ await initOpengraph(resolve(configDir, 'og'));
107
+ log(`[main] OpenGraph initialized`);
108
+ await initLabels(resolve(configDir, 'labels'));
109
+ log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
110
+ }
111
+ // Register built-in dev.hatk.* handlers so callXrpc() can find them
112
+ registerCoreHandlers(collections, config.oauth);
113
+ // Write db/schema.sql (after setup, so setup-created tables are included)
114
+ try {
115
+ const schemaDir = resolve(configDir, 'db');
116
+ mkdirSync(schemaDir, { recursive: true });
117
+ const schemaDump = await getSchemaDump();
118
+ writeFileSync(resolve(schemaDir, 'schema.sql'), `-- This file is auto-generated by hatk on startup. Do not edit.\n-- Database engine: ${config.databaseEngine}\n\n${schemaDump}\n`);
119
+ log(`[main] Schema written to db/schema.sql`);
120
+ }
121
+ catch { }
82
122
  // Detect orphaned tables
83
123
  try {
84
- const existingTables = await querySQL(`SELECT table_name FROM information_schema.tables WHERE table_schema = 'main' AND table_name NOT LIKE '\\_%' ESCAPE '\\'`);
124
+ const existingTables = await querySQL(getSqlDialect().listTablesQuery);
85
125
  for (const row of existingTables) {
86
126
  const tableName = row.table_name;
87
127
  const isChildTable = collections.some((c) => tableName.startsWith(c + '__'));
@@ -91,24 +131,56 @@ try {
91
131
  }
92
132
  }
93
133
  catch { }
94
- // 4. Initialize feeds, xrpc handlers, og, labels from directories
95
- await initFeeds(resolve(configDir, 'feeds'));
96
- log(`[main] Feeds initialized: ${listFeeds()
97
- .map((f) => f.name)
98
- .join(', ') || 'none'}`);
99
- await initXrpc(resolve(configDir, 'xrpc'));
100
- log(`[main] XRPC handlers initialized: ${listXrpc().join(', ') || 'none'}`);
101
- await initOpengraph(resolve(configDir, 'og'));
102
- log(`[main] OpenGraph initialized`);
103
- await initLabels(resolve(configDir, 'labels'));
104
- log(`[main] Labels initialized: ${getLabelDefinitions().length} definitions`);
105
134
  if (config.oauth) {
106
135
  await initOAuth(config.oauth, config.plc, config.relay);
107
136
  log(`[main] OAuth initialized (issuer: ${config.oauth.issuer})`);
108
137
  }
138
+ logMemory('before-server');
109
139
  // 5. Start server immediately (don't wait for backfill)
110
140
  const collectionSet = new Set(collections);
111
- startServer(config.port, collections, config.publicDir, config.oauth, config.admins);
141
+ const backfillOpts = {
142
+ pdsUrl: relayHttpUrl(config.relay),
143
+ plcUrl: config.plc,
144
+ collections: collectionSet,
145
+ config: config.backfill,
146
+ };
147
+ function runBackfillAndRestart() {
148
+ runBackfill(backfillOpts)
149
+ .then((recordCount) => {
150
+ log('[main] Backfill complete, rebuilding FTS indexes...');
151
+ return rebuildAllIndexes(collections).then(() => recordCount);
152
+ })
153
+ .then((recordCount) => {
154
+ log('[main] FTS indexes ready');
155
+ if (recordCount > 0 && !process.env.DEV_MODE) {
156
+ logMemory('after-backfill');
157
+ log('[main] Restarting to reclaim memory...');
158
+ process.exit(1);
159
+ }
160
+ })
161
+ .catch((err) => {
162
+ console.error('[main] Backfill error:', err.message);
163
+ });
164
+ }
165
+ const handler = createHandler({
166
+ collections,
167
+ publicDir: config.publicDir,
168
+ oauth: config.oauth,
169
+ admins: config.admins,
170
+ onResync: runBackfillAndRestart,
171
+ });
172
+ globalThis.__hatk_callXrpc = callXrpc;
173
+ globalThis.__hatk_parseSessionCookie = parseSessionCookie;
174
+ globalThis.__hatk_sessionCookieName = getSessionCookieName();
175
+ // Detect SvelteKit build output and use it as fallback handler
176
+ let fallback = undefined;
177
+ const sveltekitHandler = resolve(configDir, 'build', 'handler.js');
178
+ if (existsSync(sveltekitHandler)) {
179
+ const sk = await import(__rewriteRelativeImportExtension(/* @vite-ignore */ sveltekitHandler));
180
+ fallback = sk.handler;
181
+ log(`[main] SvelteKit handler loaded from build/handler.js`);
182
+ }
183
+ serve(handler, config.port, undefined, fallback);
112
184
  log(`\nhatk running:`);
113
185
  log(` Relay: ${config.relay}`);
114
186
  log(` Database: ${config.database}`);
@@ -117,6 +189,7 @@ log(` Collections: ${collections.join(', ')}`);
117
189
  log(` Feeds: ${listFeeds()
118
190
  .map((f) => f.name)
119
191
  .join(', ')}`);
192
+ logMemory('after-server');
120
193
  // 6. Start indexer with cursor
121
194
  const cursor = await getCursor('relay');
122
195
  startIndexer({
@@ -127,22 +200,13 @@ startIndexer({
127
200
  cursor,
128
201
  fetchTimeout: config.backfill.fetchTimeout,
129
202
  maxRetries: config.backfill.maxRetries,
203
+ parallelism: config.backfill.parallelism,
130
204
  ftsRebuildInterval: config.ftsRebuildInterval,
131
205
  });
132
206
  // 7. Run backfill in background
133
- runBackfill({
134
- pdsUrl: relayHttpUrl(config.relay),
135
- plcUrl: config.plc,
136
- collections: collectionSet,
137
- config: config.backfill,
138
- })
139
- .then(() => {
140
- log('[main] Backfill complete, rebuilding FTS indexes...');
141
- return rebuildAllIndexes(collections);
142
- })
143
- .then(() => {
144
- log('[main] FTS indexes ready');
145
- })
146
- .catch((err) => {
147
- console.error('[main] Backfill error:', err.message);
207
+ runBackfillAndRestart();
208
+ // Graceful shutdown
209
+ process.on('SIGTERM', () => {
210
+ log('[main] Received SIGTERM, shutting down...');
211
+ process.exit(0);
148
212
  });
package/dist/mst.d.ts CHANGED
@@ -1,6 +1,23 @@
1
+ /** A single entry from a Merkle Search Tree — a record path paired with its content CID. */
1
2
  export interface MstEntry {
3
+ /** Record path, e.g. "xyz.marketplace.listing/3mfniulnr7c2g" */
2
4
  path: string;
5
+ /** CID of the record's CBOR block */
3
6
  cid: string;
4
7
  }
5
- export declare function walkMst(blocks: Map<string, Uint8Array>, rootCid: string): MstEntry[];
8
+ /**
9
+ * Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
10
+ *
11
+ * The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
12
+ * record paths to CIDs. Each node contains a left subtree pointer, an array of entries
13
+ * (each with a prefix length, key suffix, value CID, and right subtree pointer), and
14
+ * keys are reconstructed by combining the prefix of the previous key with the suffix.
15
+ *
16
+ * @param blocks - Block store that resolves CIDs to raw CBOR bytes
17
+ * @param rootCid - CID of the MST root node
18
+ * @yields {MstEntry} Record entries in lexicographic key order
19
+ */
20
+ export declare function walkMst(blocks: {
21
+ get(cid: string): Uint8Array | undefined;
22
+ }, rootCid: string): Generator<MstEntry>;
6
23
  //# sourceMappingURL=mst.d.ts.map
package/dist/mst.d.ts.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,MAAM,WAAW,QAAQ;IACvB,IAAI,EAAE,MAAM,CAAA;IACZ,GAAG,EAAE,MAAM,CAAA;CACZ;AAED,wBAAgB,OAAO,CAAC,MAAM,EAAE,GAAG,CAAC,MAAM,EAAE,UAAU,CAAC,EAAE,OAAO,EAAE,MAAM,GAAG,QAAQ,EAAE,CAiCpF"}
1
+ {"version":3,"file":"mst.d.ts","sourceRoot":"","sources":["../src/mst.ts"],"names":[],"mappings":"AAEA,4FAA4F;AAC5F,MAAM,WAAW,QAAQ;IACvB,gEAAgE;IAChE,IAAI,EAAE,MAAM,CAAA;IACZ,qCAAqC;IACrC,GAAG,EAAE,MAAM,CAAA;CACZ;AAED;;;;;;;;;;;GAWG;AACH,wBAAiB,OAAO,CAAC,MAAM,EAAE;IAAE,GAAG,CAAC,GAAG,EAAE,MAAM,GAAG,UAAU,GAAG,SAAS,CAAA;CAAE,EAAE,OAAO,EAAE,MAAM,GAAG,SAAS,CAAC,QAAQ,CAAC,CA+BnH"}
package/dist/mst.js CHANGED
@@ -1,14 +1,26 @@
1
1
  import { cborDecode } from "./cbor.js";
2
- export function walkMst(blocks, rootCid) {
3
- const entries = [];
4
- function visit(cid, prefix) {
2
+ /**
3
+ * Walk an AT Protocol Merkle Search Tree (MST) in key order, yielding every record entry.
4
+ *
5
+ * The MST is a prefix-compressed B+ tree used by AT Protocol repositories to map
6
+ * record paths to CIDs. Each node contains a left subtree pointer, an array of entries
7
+ * (each with a prefix length, key suffix, value CID, and right subtree pointer), and
8
+ * keys are reconstructed by combining the prefix of the previous key with the suffix.
9
+ *
10
+ * @param blocks - Block store that resolves CIDs to raw CBOR bytes
11
+ * @param rootCid - CID of the MST root node
12
+ * @yields {MstEntry} Record entries in lexicographic key order
13
+ */
14
+ export function* walkMst(blocks, rootCid) {
15
+ /** Recursively visit an MST node, reconstructing full keys from prefix-compressed entries. */
16
+ function* visit(cid, prefix) {
5
17
  const data = blocks.get(cid);
6
18
  if (!data)
7
19
  return prefix;
8
20
  const { value: node } = cborDecode(data);
9
21
  // Visit left subtree
10
22
  if (node.l?.$link)
11
- visit(node.l.$link, prefix);
23
+ yield* visit(node.l.$link, prefix);
12
24
  let lastKey = prefix;
13
25
  for (const entry of node.e || []) {
14
26
  const keySuffix = entry.k instanceof Uint8Array ? new TextDecoder().decode(entry.k) : entry.k;
@@ -16,15 +28,14 @@ export function walkMst(blocks, rootCid) {
16
28
  const fullKey = lastKey.substring(0, prefixLen) + keySuffix;
17
29
  lastKey = fullKey;
18
30
  if (entry.v?.$link) {
19
- entries.push({ path: fullKey, cid: entry.v.$link });
31
+ yield { path: fullKey, cid: entry.v.$link };
20
32
  }
21
33
  // Visit right subtree
22
34
  if (entry.t?.$link) {
23
- visit(entry.t.$link, lastKey);
35
+ yield* visit(entry.t.$link, lastKey);
24
36
  }
25
37
  }
26
38
  return lastKey;
27
39
  }
28
- visit(rootCid, '');
29
- return entries;
40
+ yield* visit(rootCid, '');
30
41
  }
@@ -1 +1 @@
1
- {"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAOtG;AAID,wBAAsB,iBAAiB,CACrC,UAAU,EAAE,MAAM,EAClB,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,WAAW,EAAE,MAAM,CAAA;IACnB,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,aAAa,EAAE,MAAM,CAAA;IACrB,mBAAmB,CAAC,EAAE,MAAM,CAAA;IAC5B,OAAO,EAAE,MAAM,CAAA;IACf,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,QAAQ,CAAC,EAAE,MAAM,CAAA;IACjB,GAAG,CAAC,EAAE,MAAM,CAAA;IACZ,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,SAAS,EAAE,MAAM,CAAA;CAClB,GACA,OAAO,CAAC,IAAI,CAAC,CAoBf;AAED,wBAAsB,eAAe,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAM7E;AAED,wBAAsB,kBAAkB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE1E;AAID,wBAAsB,aAAa,CAAC,IAAI,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAOnF;AAED,wBAAsB,eAAe,CAAC,IAAI,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAK1E;AAID,wBAAsB,YAAY,CAChC,GAAG,EAAE,MAAM,EACX,IAAI,EAAE;IACJ,WAAW,EAAE,MAAM,CAAA;IACnB,WAAW,EAAE,MAAM,CAAA;IACnB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,OAAO,EAAE,MAAM,CAAA;IACf,cAAc,CAAC,EAAE,MAAM,CAAA;CACxB,GACA,OAAO,CAAC,IAAI,CAAC,CAWf;AAED,wBAAsB,UAAU,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGjE;AAED,wBAAsB,aAAa,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE9D;AAID,wBAAsB,iBAAiB,CACrC,KAAK,EAAE,MAAM,EACb,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;IACX,OAAO,EAAE,MAAM,CAAA;IACf,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,SAAS,CAAC,EAAE,MAAM,CAAA;CACnB,GACA,OAAO,CAAC,IAAI,CAAC,CAcf;AAED,wBAAsB,eAAe,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGxE;AAED,wBAAsB,kBAAkB,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAErE;AAID,wBAAsB,oBAAoB,CAAC,GAAG,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,OAAO,CAAC,CAK3F;AAED,wBAAsB,mBAAmB,IAAI,OAAO,CAAC,IAAI,CAAC,CASzD"}
1
+ {"version":3,"file":"db.d.ts","sourceRoot":"","sources":["../../src/oauth/db.ts"],"names":[],"mappings":"AAMA,eAAO,MAAM,SAAS,87CA0DrB,CAAA;AAID,wBAAsB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;IAAE,UAAU,EAAE,MAAM,CAAC;IAAC,SAAS,EAAE,MAAM,CAAA;CAAE,GAAG,IAAI,CAAC,CAIzG;AAED,wBAAsB,cAAc,CAAC,GAAG,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMtG;AAID,wBAAsB,iBAAiB,CACrC,UAAU,EAAE,MAAM,EAClB,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,WAAW,EAAE,MAAM,CAAA;IACnB,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,aAAa,EAAE,MAAM,CAAA;IACrB,mBAAmB,CAAC,EAAE,MAAM,CAAA;IAC5B,OAAO,EAAE,MAAM,CAAA;IACf,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,aAAa,CAAC,EAAE,MAAM,CAAA;IACtB,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,QAAQ,CAAC,EAAE,MAAM,CAAA;IACjB,GAAG,CAAC,EAAE,MAAM,CAAA;IACZ,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,SAAS,EAAE,MAAM,CAAA;CAClB,GACA,OAAO,CAAC,IAAI,CAAC,CAsBf;AAED,wBAAsB,eAAe,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAM7E;AAED,wBAAsB,kBAAkB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE1E;AAID,wBAAsB,aAAa,CAAC,IAAI,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAMnF;AAED,wBAAsB,eAAe,CAAC,IAAI,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,CAK1E;AAID,wBAAsB,YAAY,CAChC,GAAG,EAAE,MAAM,EACX,IAAI,EAAE;IACJ,WAAW,EAAE,MAAM,CAAA;IACnB,WAAW,EAAE,MAAM,CAAA;IACnB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,OAAO,EAAE,MAAM,CAAA;IACf,cAAc,CAAC,EAAE,MAAM,CAAA;CACxB,GACA,OAAO,CAAC,IAAI,CAAC,CAMf;AAED,wBAAsB,UAAU,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGjE;AAED,wBAAsB,aAAa,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAE9D;AAID,wBAAsB,iBAAiB,CACrC,KAAK,EAAE,MAAM,EACb,IAAI,EAAE;IACJ,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;IACX,OAAO,EAAE,MAAM,CAAA;IACf,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,SAAS,CAAC,EAAE,MAAM,CAAA;CACnB,GACA,OAAO,CAAC,IAAI,CAAC,CAQf;AAED,wBAAsB,eAAe,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAGxE;AAED,wBAAsB,kBAAkB,CAAC,KAAK,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAErE;AAID,wBAAsB,oBAAoB,CAAC,GAAG,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,OAAO,CAAC,OAAO,CAAC,CAK3F;AAED,wBAAsB,mBAAmB,IAAI,OAAO,CAAC,IAAI,CAAC,CAQzD"}