@syncular/server-dialect-postgres 0.0.1-100

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/src/index.ts ADDED
@@ -0,0 +1,688 @@
1
+ /**
2
+ * @syncular/server-dialect-postgres - PostgreSQL Server Sync Dialect
3
+ *
4
+ * Driver-agnostic PostgreSQL dialect for sync. Works with any Postgres-compatible
5
+ * Kysely dialect (pg, pglite, neon, etc.).
6
+ *
7
+ * Tables:
8
+ * - sync_commits: commit log (idempotency + ordering)
9
+ * - sync_table_commits: commit routing index (fast pull by table)
10
+ * - sync_changes: change log (JSONB scopes for filtering)
11
+ * - sync_client_cursors: per-client cursor tracking (pruning/observability)
12
+ */
13
+
14
+ import type { ScopeValues, StoredScopes, SyncOp } from '@syncular/core';
15
+ import type { DbExecutor } from '@syncular/server';
16
+ import {
17
+ BaseServerSyncDialect,
18
+ coerceIsoString,
19
+ coerceNumber,
20
+ type IncrementalPullRow,
21
+ type IncrementalPullRowsArgs,
22
+ parseScopes,
23
+ } from '@syncular/server';
24
+ import type { SyncChangeRow, SyncCoreDb } from '@syncular/server/schema';
25
+ import type { Kysely, RawBuilder, Transaction } from 'kysely';
26
+ import { sql } from 'kysely';
27
+
28
+ function isActiveTransaction<DB extends SyncCoreDb>(
29
+ db: Kysely<DB>
30
+ ): db is Kysely<DB> & Transaction<DB> {
31
+ return (db as { isTransaction?: boolean }).isTransaction === true;
32
+ }
33
+
34
+ function createSavepointName(): string {
35
+ const randomPart = Math.floor(Math.random() * 1_000_000_000).toString(36);
36
+ return `syncular_sp_${Date.now().toString(36)}_${randomPart}`;
37
+ }
38
+
39
+ export class PostgresServerSyncDialect extends BaseServerSyncDialect {
40
+ readonly name = 'postgres' as const;
41
+ readonly supportsForUpdate = true;
42
+ readonly supportsSavepoints = true;
43
+
44
+ // ===========================================================================
45
+ // SQL Fragment Hooks
46
+ // ===========================================================================
47
+
48
+ protected buildNumberListFilter(values: number[]): RawBuilder<unknown> {
49
+ return sql`= ANY(${values}::bigint[])`;
50
+ }
51
+
52
+ protected buildStringListFilter(values: string[]): RawBuilder<unknown> {
53
+ return sql`= ANY(${values}::text[])`;
54
+ }
55
+
56
+ // ===========================================================================
57
+ // Schema Setup
58
+ // ===========================================================================
59
+
60
+ async ensureSyncSchema<DB extends SyncCoreDb>(db: Kysely<DB>): Promise<void> {
61
+ await db.schema
62
+ .createTable('sync_commits')
63
+ .ifNotExists()
64
+ .addColumn('commit_seq', 'bigserial', (col) => col.primaryKey())
65
+ .addColumn('partition_id', 'text', (col) =>
66
+ col.notNull().defaultTo('default')
67
+ )
68
+ .addColumn('actor_id', 'text', (col) => col.notNull())
69
+ .addColumn('client_id', 'text', (col) => col.notNull())
70
+ .addColumn('client_commit_id', 'text', (col) => col.notNull())
71
+ .addColumn('created_at', 'timestamptz', (col) =>
72
+ col.notNull().defaultTo(sql`now()`)
73
+ )
74
+ .addColumn('meta', 'jsonb')
75
+ .addColumn('result_json', 'jsonb')
76
+ .addColumn('change_count', 'integer', (col) => col.notNull().defaultTo(0))
77
+ .addColumn('affected_tables', sql`text[]`, (col) =>
78
+ col.notNull().defaultTo(sql`ARRAY[]::text[]`)
79
+ )
80
+ .execute();
81
+
82
+ // Ensure new columns exist for dev environments that already created the table.
83
+ await sql`ALTER TABLE sync_commits
84
+ ADD COLUMN IF NOT EXISTS change_count integer NOT NULL DEFAULT 0`.execute(
85
+ db
86
+ );
87
+ await sql`ALTER TABLE sync_commits
88
+ ADD COLUMN IF NOT EXISTS affected_tables text[] NOT NULL DEFAULT ARRAY[]::text[]`.execute(
89
+ db
90
+ );
91
+ await sql`ALTER TABLE sync_commits
92
+ ADD COLUMN IF NOT EXISTS partition_id text NOT NULL DEFAULT 'default'`.execute(
93
+ db
94
+ );
95
+
96
+ await sql`DROP INDEX IF EXISTS idx_sync_commits_client_commit`.execute(db);
97
+ await db.schema
98
+ .createIndex('idx_sync_commits_client_commit')
99
+ .ifNotExists()
100
+ .on('sync_commits')
101
+ .columns(['partition_id', 'client_id', 'client_commit_id'])
102
+ .unique()
103
+ .execute();
104
+
105
+ // Table-based commit routing index
106
+ await db.schema
107
+ .createTable('sync_table_commits')
108
+ .ifNotExists()
109
+ .addColumn('partition_id', 'text', (col) =>
110
+ col.notNull().defaultTo('default')
111
+ )
112
+ .addColumn('table', 'text', (col) => col.notNull())
113
+ .addColumn('commit_seq', 'bigint', (col) =>
114
+ col.notNull().references('sync_commits.commit_seq').onDelete('cascade')
115
+ )
116
+ .addPrimaryKeyConstraint('sync_table_commits_pk', [
117
+ 'partition_id',
118
+ 'table',
119
+ 'commit_seq',
120
+ ])
121
+ .execute();
122
+
123
+ await sql`ALTER TABLE sync_table_commits
124
+ ADD COLUMN IF NOT EXISTS partition_id text NOT NULL DEFAULT 'default'`.execute(
125
+ db
126
+ );
127
+
128
+ await db.schema
129
+ .createIndex('idx_sync_table_commits_commit_seq')
130
+ .ifNotExists()
131
+ .on('sync_table_commits')
132
+ .columns(['partition_id', 'commit_seq'])
133
+ .execute();
134
+
135
+ // Changes table with JSONB scopes
136
+ await db.schema
137
+ .createTable('sync_changes')
138
+ .ifNotExists()
139
+ .addColumn('change_id', 'bigserial', (col) => col.primaryKey())
140
+ .addColumn('partition_id', 'text', (col) =>
141
+ col.notNull().defaultTo('default')
142
+ )
143
+ .addColumn('commit_seq', 'bigint', (col) =>
144
+ col.notNull().references('sync_commits.commit_seq').onDelete('cascade')
145
+ )
146
+ .addColumn('table', 'text', (col) => col.notNull())
147
+ .addColumn('row_id', 'text', (col) => col.notNull())
148
+ .addColumn('op', 'text', (col) => col.notNull())
149
+ .addColumn('row_json', 'jsonb')
150
+ .addColumn('row_version', 'bigint')
151
+ .addColumn('scopes', 'jsonb', (col) => col.notNull())
152
+ .execute();
153
+
154
+ await sql`ALTER TABLE sync_changes
155
+ ADD COLUMN IF NOT EXISTS partition_id text NOT NULL DEFAULT 'default'`.execute(
156
+ db
157
+ );
158
+
159
+ await db.schema
160
+ .createIndex('idx_sync_changes_commit_seq')
161
+ .ifNotExists()
162
+ .on('sync_changes')
163
+ .columns(['partition_id', 'commit_seq'])
164
+ .execute();
165
+
166
+ await db.schema
167
+ .createIndex('idx_sync_changes_table')
168
+ .ifNotExists()
169
+ .on('sync_changes')
170
+ .columns(['partition_id', 'table'])
171
+ .execute();
172
+
173
+ await this.ensureIndex(
174
+ db,
175
+ 'idx_sync_changes_scopes',
176
+ 'CREATE INDEX idx_sync_changes_scopes ON sync_changes USING GIN (scopes)'
177
+ );
178
+
179
+ await db.schema
180
+ .createTable('sync_client_cursors')
181
+ .ifNotExists()
182
+ .addColumn('partition_id', 'text', (col) =>
183
+ col.notNull().defaultTo('default')
184
+ )
185
+ .addColumn('client_id', 'text', (col) => col.notNull())
186
+ .addColumn('actor_id', 'text', (col) => col.notNull())
187
+ .addColumn('cursor', 'bigint', (col) => col.notNull().defaultTo(0))
188
+ .addColumn('effective_scopes', 'jsonb', (col) =>
189
+ col.notNull().defaultTo(sql`'{}'::jsonb`)
190
+ )
191
+ .addColumn('updated_at', 'timestamptz', (col) =>
192
+ col.notNull().defaultTo(sql`now()`)
193
+ )
194
+ .addPrimaryKeyConstraint('sync_client_cursors_pk', [
195
+ 'partition_id',
196
+ 'client_id',
197
+ ])
198
+ .execute();
199
+
200
+ await sql`ALTER TABLE sync_client_cursors
201
+ ADD COLUMN IF NOT EXISTS partition_id text NOT NULL DEFAULT 'default'`.execute(
202
+ db
203
+ );
204
+
205
+ await db.schema
206
+ .createIndex('idx_sync_client_cursors_updated_at')
207
+ .ifNotExists()
208
+ .on('sync_client_cursors')
209
+ .columns(['updated_at'])
210
+ .execute();
211
+
212
+ await db.schema
213
+ .createTable('sync_snapshot_chunks')
214
+ .ifNotExists()
215
+ .addColumn('chunk_id', 'text', (col) => col.primaryKey())
216
+ .addColumn('partition_id', 'text', (col) =>
217
+ col.notNull().defaultTo('default')
218
+ )
219
+ .addColumn('scope_key', 'text', (col) => col.notNull())
220
+ .addColumn('scope', 'text', (col) => col.notNull())
221
+ .addColumn('as_of_commit_seq', 'bigint', (col) => col.notNull())
222
+ .addColumn('row_cursor', 'text', (col) => col.notNull().defaultTo(''))
223
+ .addColumn('row_limit', 'integer', (col) => col.notNull())
224
+ .addColumn('encoding', 'text', (col) => col.notNull())
225
+ .addColumn('compression', 'text', (col) => col.notNull())
226
+ .addColumn('sha256', 'text', (col) => col.notNull())
227
+ .addColumn('byte_length', 'integer', (col) => col.notNull())
228
+ .addColumn('blob_hash', 'text', (col) => col.notNull().defaultTo(''))
229
+ .addColumn('body', 'bytea') // Deprecated: use blob storage
230
+ .addColumn('created_at', 'timestamptz', (col) =>
231
+ col.notNull().defaultTo(sql`now()`)
232
+ )
233
+ .addColumn('expires_at', 'timestamptz', (col) => col.notNull())
234
+ .execute();
235
+
236
+ await sql`ALTER TABLE sync_snapshot_chunks
237
+ ADD COLUMN IF NOT EXISTS partition_id text NOT NULL DEFAULT 'default'`.execute(
238
+ db
239
+ );
240
+
241
+ await db.schema
242
+ .createIndex('idx_sync_snapshot_chunks_expires_at')
243
+ .ifNotExists()
244
+ .on('sync_snapshot_chunks')
245
+ .columns(['expires_at'])
246
+ .execute();
247
+
248
+ await db.schema
249
+ .createIndex('idx_sync_snapshot_chunks_page_key')
250
+ .ifNotExists()
251
+ .on('sync_snapshot_chunks')
252
+ .columns([
253
+ 'partition_id',
254
+ 'scope_key',
255
+ 'scope',
256
+ 'as_of_commit_seq',
257
+ 'row_cursor',
258
+ 'row_limit',
259
+ 'encoding',
260
+ 'compression',
261
+ ])
262
+ .unique()
263
+ .execute();
264
+ }
265
+
266
+ // ===========================================================================
267
+ // Transaction Control
268
+ // ===========================================================================
269
+
270
+ async executeInTransaction<DB extends SyncCoreDb, T>(
271
+ db: Kysely<DB>,
272
+ fn: (executor: DbExecutor<DB>) => Promise<T>
273
+ ): Promise<T> {
274
+ if (isActiveTransaction(db)) {
275
+ const savepoint = createSavepointName();
276
+ await sql.raw(`SAVEPOINT ${savepoint}`).execute(db);
277
+ try {
278
+ const result = await fn(db);
279
+ await sql.raw(`RELEASE SAVEPOINT ${savepoint}`).execute(db);
280
+ return result;
281
+ } catch (error) {
282
+ await sql.raw(`ROLLBACK TO SAVEPOINT ${savepoint}`).execute(db);
283
+ await sql.raw(`RELEASE SAVEPOINT ${savepoint}`).execute(db);
284
+ throw error;
285
+ }
286
+ }
287
+ return db.transaction().execute(fn);
288
+ }
289
+
290
+ async setRepeatableRead<DB extends SyncCoreDb>(
291
+ trx: DbExecutor<DB>
292
+ ): Promise<void> {
293
+ await sql`SET TRANSACTION ISOLATION LEVEL REPEATABLE READ`.execute(trx);
294
+ }
295
+
296
+ // ===========================================================================
297
+ // Overrides (dialect-specific optimizations / casts)
298
+ // ===========================================================================
299
+
300
+ async readCommitSeqsForPull<DB extends SyncCoreDb>(
301
+ db: Kysely<DB> | Transaction<DB>,
302
+ args: {
303
+ cursor: number;
304
+ limitCommits: number;
305
+ tables: string[];
306
+ partitionId?: string;
307
+ }
308
+ ): Promise<number[]> {
309
+ const partitionId = args.partitionId ?? 'default';
310
+ if (args.tables.length === 0) return [];
311
+
312
+ // Single-table fast path: skip DISTINCT since (partition_id, table, commit_seq) is PK
313
+ if (args.tables.length === 1) {
314
+ const res = await sql<{ commit_seq: unknown }>`
315
+ SELECT commit_seq
316
+ FROM sync_table_commits
317
+ WHERE partition_id = ${partitionId}
318
+ AND "table" = ${args.tables[0]}
319
+ AND commit_seq > ${args.cursor}
320
+ ORDER BY commit_seq ASC
321
+ LIMIT ${args.limitCommits}
322
+ `.execute(db);
323
+
324
+ return res.rows
325
+ .map((r) => coerceNumber(r.commit_seq))
326
+ .filter(
327
+ (n): n is number =>
328
+ typeof n === 'number' && Number.isFinite(n) && n > args.cursor
329
+ );
330
+ }
331
+
332
+ // Multi-table: use ANY() with DISTINCT
333
+ return super.readCommitSeqsForPull(db, args);
334
+ }
335
+
336
+ async recordClientCursor<DB extends SyncCoreDb>(
337
+ db: Kysely<DB> | Transaction<DB>,
338
+ args: {
339
+ partitionId?: string;
340
+ clientId: string;
341
+ actorId: string;
342
+ cursor: number;
343
+ effectiveScopes: ScopeValues;
344
+ }
345
+ ): Promise<void> {
346
+ const partitionId = args.partitionId ?? 'default';
347
+ const now = new Date().toISOString();
348
+ const scopesJson = JSON.stringify(args.effectiveScopes);
349
+
350
+ await sql`
351
+ INSERT INTO sync_client_cursors (partition_id, client_id, actor_id, cursor, effective_scopes, updated_at)
352
+ VALUES (${partitionId}, ${args.clientId}, ${args.actorId}, ${args.cursor}, ${scopesJson}::jsonb, ${now})
353
+ ON CONFLICT(partition_id, client_id) DO UPDATE SET
354
+ actor_id = ${args.actorId},
355
+ cursor = ${args.cursor},
356
+ effective_scopes = ${scopesJson}::jsonb,
357
+ updated_at = ${now}
358
+ `.execute(db);
359
+ }
360
+
361
+ // ===========================================================================
362
+ // Commit/Change Log Queries (dialect-specific)
363
+ // ===========================================================================
364
+
365
+ async readChangesForCommits<DB extends SyncCoreDb>(
366
+ db: DbExecutor<DB>,
367
+ args: {
368
+ commitSeqs: number[];
369
+ table: string;
370
+ scopes: ScopeValues;
371
+ partitionId?: string;
372
+ }
373
+ ): Promise<SyncChangeRow[]> {
374
+ const partitionId = args.partitionId ?? 'default';
375
+ if (args.commitSeqs.length === 0) return [];
376
+
377
+ // Build JSONB containment conditions for scope filtering
378
+ const scopeConditions: ReturnType<typeof sql>[] = [];
379
+ for (const [key, value] of Object.entries(args.scopes)) {
380
+ if (Array.isArray(value)) {
381
+ scopeConditions.push(sql`scopes->>${key} = ANY(${value}::text[])`);
382
+ } else {
383
+ scopeConditions.push(sql`scopes->>${key} = ${value}`);
384
+ }
385
+ }
386
+
387
+ let query = sql<{
388
+ commit_seq: unknown;
389
+ table: string;
390
+ row_id: string;
391
+ op: string;
392
+ row_json: unknown | null;
393
+ row_version: unknown | null;
394
+ scopes: unknown;
395
+ }>`
396
+ SELECT commit_seq, "table", row_id, op, row_json, row_version, scopes
397
+ FROM sync_changes
398
+ WHERE commit_seq = ANY(${args.commitSeqs}::bigint[])
399
+ AND partition_id = ${partitionId}
400
+ AND "table" = ${args.table}
401
+ `;
402
+
403
+ if (scopeConditions.length > 0) {
404
+ const scopeFilter = sql.join(scopeConditions, sql` AND `);
405
+ query = sql<{
406
+ commit_seq: unknown;
407
+ table: string;
408
+ row_id: string;
409
+ op: string;
410
+ row_json: unknown | null;
411
+ row_version: unknown | null;
412
+ scopes: unknown;
413
+ }>`
414
+ SELECT commit_seq, "table", row_id, op, row_json, row_version, scopes
415
+ FROM sync_changes
416
+ WHERE commit_seq = ANY(${args.commitSeqs}::bigint[])
417
+ AND partition_id = ${partitionId}
418
+ AND "table" = ${args.table}
419
+ AND (${scopeFilter})
420
+ ORDER BY commit_seq ASC, change_id ASC
421
+ `;
422
+ }
423
+
424
+ const res = await query.execute(db);
425
+
426
+ return res.rows.map((row) => ({
427
+ commit_seq: coerceNumber(row.commit_seq) ?? 0,
428
+ table: row.table,
429
+ row_id: row.row_id,
430
+ op: row.op as SyncOp,
431
+ row_json: row.row_json ?? null,
432
+ row_version: coerceNumber(row.row_version),
433
+ scopes: parseScopes(row.scopes),
434
+ }));
435
+ }
436
+
437
+ protected override async readIncrementalPullRowsBatch<DB extends SyncCoreDb>(
438
+ db: DbExecutor<DB>,
439
+ args: Omit<IncrementalPullRowsArgs, 'batchSize'>
440
+ ): Promise<IncrementalPullRow[]> {
441
+ const partitionId = args.partitionId ?? 'default';
442
+ const limitCommits = Math.max(1, Math.min(500, args.limitCommits));
443
+
444
+ // Build scope filter conditions
445
+ const scopeConditions: ReturnType<typeof sql>[] = [];
446
+ for (const [key, value] of Object.entries(args.scopes)) {
447
+ if (Array.isArray(value)) {
448
+ scopeConditions.push(sql`c.scopes->>${key} = ANY(${value}::text[])`);
449
+ } else {
450
+ scopeConditions.push(sql`c.scopes->>${key} = ${value}`);
451
+ }
452
+ }
453
+
454
+ const scopeFilter =
455
+ scopeConditions.length > 0
456
+ ? sql.join(scopeConditions, sql` AND `)
457
+ : sql`TRUE`;
458
+
459
+ const res = await sql<{
460
+ commit_seq: unknown;
461
+ actor_id: string;
462
+ created_at: unknown;
463
+ change_id: unknown;
464
+ table: string;
465
+ row_id: string;
466
+ op: string;
467
+ row_json: unknown | null;
468
+ row_version: unknown | null;
469
+ scopes: unknown;
470
+ }>`
471
+ WITH commit_seqs AS (
472
+ SELECT DISTINCT tc.commit_seq
473
+ FROM sync_table_commits tc
474
+ JOIN sync_commits cm ON cm.commit_seq = tc.commit_seq
475
+ WHERE tc.partition_id = ${partitionId}
476
+ AND tc."table" = ${args.table}
477
+ AND cm.partition_id = ${partitionId}
478
+ AND tc.commit_seq > ${args.cursor}
479
+ AND EXISTS (
480
+ SELECT 1
481
+ FROM sync_changes c
482
+ WHERE c.commit_seq = tc.commit_seq
483
+ AND c.partition_id = ${partitionId}
484
+ AND c."table" = ${args.table}
485
+ AND (${scopeFilter})
486
+ )
487
+ ORDER BY tc.commit_seq ASC
488
+ LIMIT ${limitCommits}
489
+ )
490
+ SELECT
491
+ cm.commit_seq,
492
+ cm.actor_id,
493
+ cm.created_at,
494
+ c.change_id,
495
+ c."table",
496
+ c.row_id,
497
+ c.op,
498
+ c.row_json,
499
+ c.row_version,
500
+ c.scopes
501
+ FROM commit_seqs cs
502
+ JOIN sync_commits cm ON cm.commit_seq = cs.commit_seq
503
+ JOIN sync_changes c ON c.commit_seq = cs.commit_seq
504
+ WHERE cm.partition_id = ${partitionId}
505
+ AND c.partition_id = ${partitionId}
506
+ AND c."table" = ${args.table}
507
+ AND (${scopeFilter})
508
+ ORDER BY cm.commit_seq ASC, c.change_id ASC
509
+ `.execute(db);
510
+
511
+ return res.rows.map((row) => ({
512
+ commit_seq: coerceNumber(row.commit_seq) ?? 0,
513
+ actor_id: row.actor_id,
514
+ created_at: coerceIsoString(row.created_at),
515
+ change_id: coerceNumber(row.change_id) ?? 0,
516
+ table: row.table,
517
+ row_id: row.row_id,
518
+ op: row.op as SyncOp,
519
+ row_json: row.row_json ?? null,
520
+ row_version: coerceNumber(row.row_version),
521
+ scopes: parseScopes(row.scopes),
522
+ }));
523
+ }
524
+
525
+ async compactChanges<DB extends SyncCoreDb>(
526
+ db: DbExecutor<DB>,
527
+ args: { fullHistoryHours: number }
528
+ ): Promise<number> {
529
+ const cutoffIso = new Date(
530
+ Date.now() - args.fullHistoryHours * 60 * 60 * 1000
531
+ ).toISOString();
532
+
533
+ const res = await sql`
534
+ WITH ranked AS (
535
+ SELECT
536
+ c.change_id,
537
+ row_number() OVER (
538
+ PARTITION BY c.partition_id, c."table", c.row_id, c.scopes
539
+ ORDER BY c.commit_seq DESC, c.change_id DESC
540
+ ) AS rn
541
+ FROM sync_changes c
542
+ JOIN sync_commits cm ON cm.commit_seq = c.commit_seq
543
+ WHERE cm.created_at < ${cutoffIso}
544
+ )
545
+ DELETE FROM sync_changes
546
+ WHERE change_id IN (SELECT change_id FROM ranked WHERE rn > 1)
547
+ `.execute(db);
548
+
549
+ const deletedChanges = Number(res.numAffectedRows ?? 0);
550
+
551
+ // Remove routing index entries that no longer have any remaining changes
552
+ await sql`
553
+ DELETE FROM sync_table_commits tc
554
+ USING sync_commits cm
555
+ WHERE cm.commit_seq = tc.commit_seq
556
+ AND cm.partition_id = tc.partition_id
557
+ AND cm.created_at < ${cutoffIso}
558
+ AND NOT EXISTS (
559
+ SELECT 1
560
+ FROM sync_changes c
561
+ WHERE c.commit_seq = tc.commit_seq
562
+ AND c.partition_id = tc.partition_id
563
+ AND c."table" = tc."table"
564
+ )
565
+ `.execute(db);
566
+
567
+ return deletedChanges;
568
+ }
569
+
570
+ // ===========================================================================
571
+ // Scope Conversion Helpers
572
+ // ===========================================================================
573
+
574
+ scopesToDb(scopes: StoredScopes): StoredScopes {
575
+ return scopes;
576
+ }
577
+
578
+ dbToArray(value: unknown): string[] {
579
+ if (Array.isArray(value)) {
580
+ return value.filter((k: unknown): k is string => typeof k === 'string');
581
+ }
582
+ return [];
583
+ }
584
+
585
+ arrayToDb(values: string[]): string[] {
586
+ return values.filter((v) => v.length > 0);
587
+ }
588
+
589
+ // ===========================================================================
590
+ // Console Schema (Request Events)
591
+ // ===========================================================================
592
+
593
+ async ensureConsoleSchema<DB extends SyncCoreDb>(
594
+ db: Kysely<DB>
595
+ ): Promise<void> {
596
+ await sql`
597
+ CREATE TABLE IF NOT EXISTS sync_request_events (
598
+ event_id BIGSERIAL PRIMARY KEY,
599
+ event_type TEXT NOT NULL,
600
+ actor_id TEXT NOT NULL,
601
+ client_id TEXT NOT NULL,
602
+ transport_path TEXT NOT NULL DEFAULT 'direct',
603
+ status_code INTEGER NOT NULL,
604
+ outcome TEXT NOT NULL,
605
+ duration_ms INTEGER NOT NULL,
606
+ commit_seq BIGINT,
607
+ operation_count INTEGER,
608
+ row_count INTEGER,
609
+ tables TEXT[] NOT NULL DEFAULT ARRAY[]::TEXT[],
610
+ error_message TEXT,
611
+ created_at TIMESTAMPTZ NOT NULL DEFAULT NOW()
612
+ )
613
+ `.execute(db);
614
+ await sql`
615
+ ALTER TABLE sync_request_events
616
+ ADD COLUMN IF NOT EXISTS transport_path TEXT NOT NULL DEFAULT 'direct'
617
+ `.execute(db);
618
+
619
+ await this.ensureIndex(
620
+ db,
621
+ 'idx_sync_request_events_created_at',
622
+ 'CREATE INDEX idx_sync_request_events_created_at ON sync_request_events(created_at DESC)'
623
+ );
624
+ await this.ensureIndex(
625
+ db,
626
+ 'idx_sync_request_events_event_type',
627
+ 'CREATE INDEX idx_sync_request_events_event_type ON sync_request_events(event_type)'
628
+ );
629
+ await this.ensureIndex(
630
+ db,
631
+ 'idx_sync_request_events_client_id',
632
+ 'CREATE INDEX idx_sync_request_events_client_id ON sync_request_events(client_id)'
633
+ );
634
+
635
+ // API Keys table
636
+ await sql`
637
+ CREATE TABLE IF NOT EXISTS sync_api_keys (
638
+ key_id TEXT PRIMARY KEY,
639
+ key_hash TEXT NOT NULL,
640
+ key_prefix TEXT NOT NULL,
641
+ name TEXT NOT NULL,
642
+ key_type TEXT NOT NULL,
643
+ scope_keys TEXT[] DEFAULT ARRAY[]::TEXT[],
644
+ actor_id TEXT,
645
+ created_at TIMESTAMPTZ NOT NULL DEFAULT NOW(),
646
+ expires_at TIMESTAMPTZ,
647
+ last_used_at TIMESTAMPTZ,
648
+ revoked_at TIMESTAMPTZ
649
+ )
650
+ `.execute(db);
651
+
652
+ await this.ensureIndex(
653
+ db,
654
+ 'idx_sync_api_keys_key_hash',
655
+ 'CREATE INDEX idx_sync_api_keys_key_hash ON sync_api_keys(key_hash)'
656
+ );
657
+ await this.ensureIndex(
658
+ db,
659
+ 'idx_sync_api_keys_key_type',
660
+ 'CREATE INDEX idx_sync_api_keys_key_type ON sync_api_keys(key_type)'
661
+ );
662
+ }
663
+
664
+ // ===========================================================================
665
+ // Private Helpers
666
+ // ===========================================================================
667
+
668
+ private async ensureIndex<DB extends SyncCoreDb>(
669
+ db: Kysely<DB>,
670
+ indexName: string,
671
+ createSql: string
672
+ ): Promise<void> {
673
+ const exists = await sql<{ ok: 1 }>`
674
+ SELECT 1 as ok
675
+ FROM pg_indexes
676
+ WHERE schemaname = 'public'
677
+ AND indexname = ${indexName}
678
+ LIMIT 1
679
+ `.execute(db);
680
+
681
+ if (exists.rows.length > 0) return;
682
+ await sql.raw(createSql).execute(db);
683
+ }
684
+ }
685
+
686
+ export function createPostgresServerDialect(): PostgresServerSyncDialect {
687
+ return new PostgresServerSyncDialect();
688
+ }