@rocicorp/zero 1.4.0-canary.1 → 1.4.0-canary.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (175) hide show
  1. package/out/analyze-query/src/analyze-cli.d.ts +0 -1
  2. package/out/analyze-query/src/analyze-cli.d.ts.map +1 -1
  3. package/out/analyze-query/src/analyze-cli.js +0 -1
  4. package/out/analyze-query/src/analyze-cli.js.map +1 -1
  5. package/out/analyze-query/src/bin-analyze.js +11 -10
  6. package/out/analyze-query/src/bin-analyze.js.map +1 -1
  7. package/out/analyze-query/src/bin-transform.js +1 -1
  8. package/out/analyze-query/src/bin-transform.js.map +1 -1
  9. package/out/replicache/src/btree/node.d.ts +1 -1
  10. package/out/replicache/src/btree/node.d.ts.map +1 -1
  11. package/out/replicache/src/btree/node.js +34 -21
  12. package/out/replicache/src/btree/node.js.map +1 -1
  13. package/out/replicache/src/btree/write.js +1 -2
  14. package/out/replicache/src/btree/write.js.map +1 -1
  15. package/out/replicache/src/kv/sqlite-store.d.ts.map +1 -1
  16. package/out/replicache/src/kv/sqlite-store.js +7 -1
  17. package/out/replicache/src/kv/sqlite-store.js.map +1 -1
  18. package/out/replicache/src/with-transactions.d.ts.map +1 -1
  19. package/out/replicache/src/with-transactions.js +16 -2
  20. package/out/replicache/src/with-transactions.js.map +1 -1
  21. package/out/shared/src/btree-set.d.ts +6 -0
  22. package/out/shared/src/btree-set.d.ts.map +1 -1
  23. package/out/shared/src/btree-set.js +34 -0
  24. package/out/shared/src/btree-set.js.map +1 -1
  25. package/out/zero/package.js +8 -2
  26. package/out/zero/package.js.map +1 -1
  27. package/out/zero/src/adapters/kysely.d.ts +2 -0
  28. package/out/zero/src/adapters/kysely.d.ts.map +1 -0
  29. package/out/zero/src/adapters/kysely.js +2 -0
  30. package/out/zero/src/zero.js +2 -1
  31. package/out/zero-cache/src/auth/write-authorizer.d.ts.map +1 -1
  32. package/out/zero-cache/src/auth/write-authorizer.js +14 -1
  33. package/out/zero-cache/src/auth/write-authorizer.js.map +1 -1
  34. package/out/zero-cache/src/config/zero-config.d.ts +18 -0
  35. package/out/zero-cache/src/config/zero-config.d.ts.map +1 -1
  36. package/out/zero-cache/src/config/zero-config.js +35 -3
  37. package/out/zero-cache/src/config/zero-config.js.map +1 -1
  38. package/out/zero-cache/src/db/migration-lite.js +8 -1
  39. package/out/zero-cache/src/db/migration-lite.js.map +1 -1
  40. package/out/zero-cache/src/db/pg-to-lite.d.ts +1 -1
  41. package/out/zero-cache/src/db/pg-to-lite.d.ts.map +1 -1
  42. package/out/zero-cache/src/db/pg-to-lite.js +13 -13
  43. package/out/zero-cache/src/db/pg-to-lite.js.map +1 -1
  44. package/out/zero-cache/src/observability/metrics.d.ts +36 -6
  45. package/out/zero-cache/src/observability/metrics.d.ts.map +1 -1
  46. package/out/zero-cache/src/observability/metrics.js +55 -10
  47. package/out/zero-cache/src/observability/metrics.js.map +1 -1
  48. package/out/zero-cache/src/scripts/decommission.d.ts.map +1 -1
  49. package/out/zero-cache/src/scripts/decommission.js +3 -3
  50. package/out/zero-cache/src/scripts/decommission.js.map +1 -1
  51. package/out/zero-cache/src/scripts/deploy-permissions.js +1 -1
  52. package/out/zero-cache/src/scripts/deploy-permissions.js.map +1 -1
  53. package/out/zero-cache/src/server/change-streamer.d.ts.map +1 -1
  54. package/out/zero-cache/src/server/change-streamer.js +4 -5
  55. package/out/zero-cache/src/server/change-streamer.js.map +1 -1
  56. package/out/zero-cache/src/server/main.d.ts.map +1 -1
  57. package/out/zero-cache/src/server/main.js +6 -1
  58. package/out/zero-cache/src/server/main.js.map +1 -1
  59. package/out/zero-cache/src/server/reaper.d.ts.map +1 -1
  60. package/out/zero-cache/src/server/reaper.js +1 -4
  61. package/out/zero-cache/src/server/reaper.js.map +1 -1
  62. package/out/zero-cache/src/server/shadow-syncer.js +35 -0
  63. package/out/zero-cache/src/server/shadow-syncer.js.map +1 -0
  64. package/out/zero-cache/src/server/syncer.d.ts.map +1 -1
  65. package/out/zero-cache/src/server/syncer.js +2 -8
  66. package/out/zero-cache/src/server/syncer.js.map +1 -1
  67. package/out/zero-cache/src/server/worker-urls.d.ts +1 -0
  68. package/out/zero-cache/src/server/worker-urls.d.ts.map +1 -1
  69. package/out/zero-cache/src/server/worker-urls.js +2 -1
  70. package/out/zero-cache/src/server/worker-urls.js.map +1 -1
  71. package/out/zero-cache/src/services/analyze.d.ts.map +1 -1
  72. package/out/zero-cache/src/services/analyze.js +1 -1
  73. package/out/zero-cache/src/services/analyze.js.map +1 -1
  74. package/out/zero-cache/src/services/change-source/pg/backfill-stream.d.ts +8 -1
  75. package/out/zero-cache/src/services/change-source/pg/backfill-stream.d.ts.map +1 -1
  76. package/out/zero-cache/src/services/change-source/pg/backfill-stream.js +31 -18
  77. package/out/zero-cache/src/services/change-source/pg/backfill-stream.js.map +1 -1
  78. package/out/zero-cache/src/services/change-source/pg/change-source.d.ts.map +1 -1
  79. package/out/zero-cache/src/services/change-source/pg/change-source.js +48 -47
  80. package/out/zero-cache/src/services/change-source/pg/change-source.js.map +1 -1
  81. package/out/zero-cache/src/services/change-source/pg/initial-sync.d.ts +6 -1
  82. package/out/zero-cache/src/services/change-source/pg/initial-sync.d.ts.map +1 -1
  83. package/out/zero-cache/src/services/change-source/pg/initial-sync.js +64 -22
  84. package/out/zero-cache/src/services/change-source/pg/initial-sync.js.map +1 -1
  85. package/out/zero-cache/src/services/change-streamer/change-streamer-http.d.ts.map +1 -1
  86. package/out/zero-cache/src/services/change-streamer/change-streamer-http.js +2 -3
  87. package/out/zero-cache/src/services/change-streamer/change-streamer-http.js.map +1 -1
  88. package/out/zero-cache/src/services/change-streamer/schema/tables.js +1 -1
  89. package/out/zero-cache/src/services/replicator/change-processor.d.ts.map +1 -1
  90. package/out/zero-cache/src/services/replicator/change-processor.js +10 -3
  91. package/out/zero-cache/src/services/replicator/change-processor.js.map +1 -1
  92. package/out/zero-cache/src/services/shadow-sync/shadow-sync-service.js +49 -0
  93. package/out/zero-cache/src/services/shadow-sync/shadow-sync-service.js.map +1 -0
  94. package/out/zero-cache/src/services/statz.js +3 -3
  95. package/out/zero-cache/src/services/statz.js.map +1 -1
  96. package/out/zero-cache/src/services/view-syncer/client-handler.js +3 -6
  97. package/out/zero-cache/src/services/view-syncer/client-handler.js.map +1 -1
  98. package/out/zero-cache/src/services/view-syncer/cvr-store.d.ts +1 -0
  99. package/out/zero-cache/src/services/view-syncer/cvr-store.d.ts.map +1 -1
  100. package/out/zero-cache/src/services/view-syncer/cvr-store.js +34 -11
  101. package/out/zero-cache/src/services/view-syncer/cvr-store.js.map +1 -1
  102. package/out/zero-cache/src/services/view-syncer/cvr.d.ts +16 -1
  103. package/out/zero-cache/src/services/view-syncer/cvr.d.ts.map +1 -1
  104. package/out/zero-cache/src/services/view-syncer/cvr.js +19 -1
  105. package/out/zero-cache/src/services/view-syncer/cvr.js.map +1 -1
  106. package/out/zero-cache/src/services/view-syncer/inspect-handler.js +1 -1
  107. package/out/zero-cache/src/services/view-syncer/inspect-handler.js.map +1 -1
  108. package/out/zero-cache/src/services/view-syncer/pipeline-driver.d.ts +8 -2
  109. package/out/zero-cache/src/services/view-syncer/pipeline-driver.d.ts.map +1 -1
  110. package/out/zero-cache/src/services/view-syncer/pipeline-driver.js +50 -10
  111. package/out/zero-cache/src/services/view-syncer/pipeline-driver.js.map +1 -1
  112. package/out/zero-cache/src/services/view-syncer/row-record-cache.js +4 -7
  113. package/out/zero-cache/src/services/view-syncer/row-record-cache.js.map +1 -1
  114. package/out/zero-cache/src/services/view-syncer/row-set-signature.d.ts +17 -0
  115. package/out/zero-cache/src/services/view-syncer/row-set-signature.d.ts.map +1 -0
  116. package/out/zero-cache/src/services/view-syncer/row-set-signature.js +29 -0
  117. package/out/zero-cache/src/services/view-syncer/row-set-signature.js.map +1 -0
  118. package/out/zero-cache/src/services/view-syncer/schema/cvr.d.ts +1 -0
  119. package/out/zero-cache/src/services/view-syncer/schema/cvr.d.ts.map +1 -1
  120. package/out/zero-cache/src/services/view-syncer/schema/cvr.js +1 -0
  121. package/out/zero-cache/src/services/view-syncer/schema/cvr.js.map +1 -1
  122. package/out/zero-cache/src/services/view-syncer/schema/init.d.ts.map +1 -1
  123. package/out/zero-cache/src/services/view-syncer/schema/init.js +5 -1
  124. package/out/zero-cache/src/services/view-syncer/schema/init.js.map +1 -1
  125. package/out/zero-cache/src/services/view-syncer/schema/types.d.ts +105 -0
  126. package/out/zero-cache/src/services/view-syncer/schema/types.d.ts.map +1 -1
  127. package/out/zero-cache/src/services/view-syncer/schema/types.js +8 -4
  128. package/out/zero-cache/src/services/view-syncer/schema/types.js.map +1 -1
  129. package/out/zero-cache/src/services/view-syncer/view-syncer.d.ts.map +1 -1
  130. package/out/zero-cache/src/services/view-syncer/view-syncer.js +18 -28
  131. package/out/zero-cache/src/services/view-syncer/view-syncer.js.map +1 -1
  132. package/out/zero-cache/src/types/pg.d.ts +1 -1
  133. package/out/zero-cache/src/types/pg.d.ts.map +1 -1
  134. package/out/zero-cache/src/types/pg.js +8 -2
  135. package/out/zero-cache/src/types/pg.js.map +1 -1
  136. package/out/zero-cache/src/types/timeout.d.ts +11 -0
  137. package/out/zero-cache/src/types/timeout.d.ts.map +1 -0
  138. package/out/zero-cache/src/types/timeout.js +26 -0
  139. package/out/zero-cache/src/types/timeout.js.map +1 -0
  140. package/out/zero-cache/src/workers/connection.js +3 -3
  141. package/out/zero-cache/src/workers/connection.js.map +1 -1
  142. package/out/zero-client/src/client/version.js +1 -1
  143. package/out/zero-client/src/mod.d.ts +1 -0
  144. package/out/zero-client/src/mod.d.ts.map +1 -1
  145. package/out/zero-client/src/mod.js +1 -0
  146. package/out/zero-react/src/zero.js +1 -0
  147. package/out/zero-server/src/adapters/kysely.d.ts +69 -0
  148. package/out/zero-server/src/adapters/kysely.d.ts.map +1 -0
  149. package/out/zero-server/src/adapters/kysely.js +82 -0
  150. package/out/zero-server/src/adapters/kysely.js.map +1 -0
  151. package/out/zero-server/src/adapters/postgresjs.d.ts.map +1 -1
  152. package/out/zero-server/src/adapters/postgresjs.js +1 -1
  153. package/out/zero-server/src/adapters/postgresjs.js.map +1 -1
  154. package/out/zero-solid/src/zero.js +1 -0
  155. package/out/zql/src/ivm/memory-source.d.ts.map +1 -1
  156. package/out/zql/src/ivm/memory-source.js +3 -3
  157. package/out/zql/src/ivm/memory-source.js.map +1 -1
  158. package/out/zql/src/query/query-internals.d.ts.map +1 -1
  159. package/out/zql/src/query/query-internals.js +1 -1
  160. package/out/zql/src/query/query-internals.js.map +1 -1
  161. package/out/zql/src/query/validate-input.d.ts +8 -0
  162. package/out/zql/src/query/validate-input.d.ts.map +1 -1
  163. package/out/zql/src/query/validate-input.js +15 -2
  164. package/out/zql/src/query/validate-input.js.map +1 -1
  165. package/out/zqlite/src/query-builder.js +19 -7
  166. package/out/zqlite/src/query-builder.js.map +1 -1
  167. package/package.json +10 -2
  168. package/out/analyze-query/src/explain-queries.d.ts +0 -4
  169. package/out/analyze-query/src/explain-queries.d.ts.map +0 -1
  170. package/out/analyze-query/src/explain-queries.js +0 -13
  171. package/out/analyze-query/src/explain-queries.js.map +0 -1
  172. package/out/otel/src/test-log-config.d.ts +0 -8
  173. package/out/otel/src/test-log-config.d.ts.map +0 -1
  174. package/out/otel/src/test-log-config.js +0 -12
  175. package/out/otel/src/test-log-config.js.map +0 -1
@@ -1 +1 @@
1
- {"version":3,"file":"change-processor.js","names":["#db","#changeLog","#tableMetadata","#mode","#failService","#tableSpecs","#failure","#currentTx","#fail","#processMessage","#beginTransaction","#lc","#startMs","#version","#jsonFormat","#columnMetadata","#reloadTableSpecs","#tableSpec","#upsert","#getKey","#logSetOp","#logDeleteOp","#delete","#logTruncateOp","#logResetOp","#bumpVersions","#pos","#numChangeLogEntries","#schemaChanged","#completedBackfill"],"sources":["../../../../../../zero-cache/src/services/replicator/change-processor.ts"],"sourcesContent":["import type {LogContext} from '@rocicorp/logger';\nimport {SqliteError} from '@rocicorp/zero-sqlite3';\nimport {AbortError} from '../../../../shared/src/abort-error.ts';\nimport {assert, unreachable} from '../../../../shared/src/asserts.ts';\nimport {stringify} from '../../../../shared/src/bigint-json.ts';\nimport {must} from '../../../../shared/src/must.ts';\nimport type {DownloadStatus} from '../../../../zero-events/src/status.ts';\nimport {\n createLiteIndexStatement,\n createLiteTableStatement,\n liteColumnDef,\n} from '../../db/create.ts';\nimport {\n computeZqlSpecs,\n listIndexes,\n listTables,\n type LiteTableSpecWithReplicationStatus,\n} from '../../db/lite-tables.ts';\nimport {\n mapPostgresToLite,\n mapPostgresToLiteColumn,\n mapPostgresToLiteIndex,\n} from '../../db/pg-to-lite.ts';\nimport type {StatementRunner} from '../../db/statements.ts';\nimport type {LexiVersion} from '../../types/lexi-version.ts';\nimport {\n JSON_PARSED,\n liteRow,\n type JSONFormat,\n type LiteRow,\n type LiteRowKey,\n type LiteValueType,\n} from '../../types/lite.ts';\nimport {liteTableName} from '../../types/names.ts';\nimport {id} from '../../types/sql.ts';\nimport type {\n BackfillCompleted,\n Change,\n ColumnAdd,\n ColumnDrop,\n ColumnUpdate,\n Identifier,\n IndexCreate,\n IndexDrop,\n MessageBackfill,\n MessageCommit,\n MessageDelete,\n MessageInsert,\n MessageRelation,\n MessageTruncate,\n MessageUpdate,\n TableCreate,\n TableDrop,\n TableRename,\n TableUpdateMetadata,\n} from '../change-source/protocol/current/data.ts';\nimport type {ChangeStreamData} from '../change-source/protocol/current/downstream.ts';\nimport type {ReplicatorMode} from './replicator.ts';\nimport {ChangeLog, DEL_OP, SET_OP} from './schema/change-log.ts';\nimport {ColumnMetadataStore} from './schema/column-metadata.ts';\nimport {\n ZERO_VERSION_COLUMN_NAME,\n updateReplicationWatermark,\n} from './schema/replication-state.ts';\nimport {TableMetadataTracker} from './schema/table-metadata.ts';\n\nexport type ChangeProcessorMode = ReplicatorMode | 'initial-sync';\n\nexport type CommitResult = {\n watermark: string;\n completedBackfill: DownloadStatus | undefined;\n schemaUpdated: boolean;\n changeLogUpdated: boolean;\n};\n\n/**\n * The ChangeProcessor partitions the stream of messages into transactions\n * by creating a {@link TransactionProcessor} when a transaction begins, and dispatching\n * messages to it until the commit is received.\n *\n * From https://www.postgresql.org/docs/current/protocol-logical-replication.html#PROTOCOL-LOGICAL-MESSAGES-FLOW :\n *\n * \"The logical replication protocol sends individual transactions one by one.\n * This means that all messages between a pair of Begin and Commit messages\n * belong to the same transaction.\"\n */\nexport class ChangeProcessor {\n readonly #db: StatementRunner;\n readonly #changeLog: ChangeLog;\n readonly #tableMetadata: TableMetadataTracker;\n readonly #mode: ChangeProcessorMode;\n readonly #failService: (lc: LogContext, err: unknown) => void;\n\n // The TransactionProcessor lazily loads table specs into this Map,\n // and reloads them after a schema change. It is cached here to avoid\n // reading them from the DB on every transaction.\n readonly #tableSpecs = new Map<string, LiteTableSpecWithReplicationStatus>();\n\n #currentTx: TransactionProcessor | null = null;\n\n #failure: Error | undefined;\n\n constructor(\n db: StatementRunner,\n mode: ChangeProcessorMode,\n failService: (lc: LogContext, err: unknown) => void,\n ) {\n this.#db = db;\n this.#changeLog = new ChangeLog(db.db);\n this.#tableMetadata = new TableMetadataTracker(db.db);\n this.#mode = mode;\n this.#failService = failService;\n }\n\n #fail(lc: LogContext, err: unknown) {\n if (!this.#failure) {\n this.#currentTx?.abort(lc); // roll back any pending transaction.\n\n this.#failure = ensureError(err);\n\n if (!(err instanceof AbortError)) {\n // Propagate the failure up to the service.\n lc.error?.('Message Processing failed:', this.#failure);\n this.#failService(lc, this.#failure);\n }\n }\n }\n\n abort(lc: LogContext) {\n this.#fail(lc, new AbortError());\n }\n\n /** @return If a transaction was committed. */\n processMessage(\n lc: LogContext,\n downstream: ChangeStreamData,\n ): CommitResult | null {\n const [type, message] = downstream;\n if (this.#failure) {\n lc.debug?.(`Dropping ${message.tag}`);\n return null;\n }\n try {\n const watermark =\n type === 'begin'\n ? downstream[2].commitWatermark\n : type === 'commit'\n ? downstream[2].watermark\n : undefined;\n return this.#processMessage(lc, message, watermark);\n } catch (e) {\n this.#fail(lc, e);\n }\n return null;\n }\n\n #beginTransaction(\n lc: LogContext,\n commitVersion: string,\n jsonFormat: JSONFormat,\n ): TransactionProcessor {\n const start = Date.now();\n\n // litestream can technically hold the lock for an arbitrary amount of time\n // when checkpointing a large commit. Crashing on the busy-timeout in this\n // scenario will either produce a corrupt backup or otherwise prevent\n // replication from proceeding.\n //\n // Instead, retry the lock acquisition indefinitely. If this masks\n // an unknown deadlock situation, manual intervention will be necessary.\n for (let i = 0; ; i++) {\n try {\n return new TransactionProcessor(\n lc,\n this.#db,\n this.#mode,\n this.#changeLog,\n this.#tableMetadata,\n this.#tableSpecs,\n commitVersion,\n jsonFormat,\n );\n } catch (e) {\n if (e instanceof SqliteError && e.code === 'SQLITE_BUSY') {\n lc.warn?.(\n `SQLITE_BUSY for ${Date.now() - start} ms (attempt ${i + 1}). ` +\n `This is only expected if litestream is performing a large ` +\n `checkpoint.`,\n e,\n );\n continue;\n }\n throw e;\n }\n }\n }\n\n /** @return If a transaction was committed. */\n #processMessage(\n lc: LogContext,\n msg: Change,\n watermark: string | undefined,\n ): CommitResult | null {\n if (msg.tag === 'begin') {\n if (this.#currentTx) {\n throw new Error(`Already in a transaction ${stringify(msg)}`);\n }\n this.#currentTx = this.#beginTransaction(\n lc,\n must(watermark),\n msg.json ?? JSON_PARSED,\n );\n return null;\n }\n\n // For non-begin messages, there should be a #currentTx set.\n const tx = this.#currentTx;\n if (!tx) {\n throw new Error(\n `Received message outside of transaction: ${stringify(msg)}`,\n );\n }\n\n if (msg.tag === 'commit') {\n // Undef this.#currentTx to allow the assembly of the next transaction.\n this.#currentTx = null;\n\n assert(watermark, 'watermark is required for commit messages');\n return tx.processCommit(msg, watermark);\n }\n\n if (msg.tag === 'rollback') {\n this.#currentTx?.abort(lc);\n this.#currentTx = null;\n return null;\n }\n\n switch (msg.tag) {\n case 'insert':\n tx.processInsert(msg);\n break;\n case 'update':\n tx.processUpdate(msg);\n break;\n case 'delete':\n tx.processDelete(msg);\n break;\n case 'truncate':\n tx.processTruncate(msg);\n break;\n case 'create-table':\n tx.processCreateTable(msg);\n break;\n case 'rename-table':\n tx.processRenameTable(msg);\n break;\n case 'update-table-metadata':\n tx.processTableMetadata(msg);\n break;\n case 'add-column':\n tx.processAddColumn(msg);\n break;\n case 'update-column':\n tx.processUpdateColumn(msg);\n break;\n case 'drop-column':\n tx.processDropColumn(msg);\n break;\n case 'drop-table':\n tx.processDropTable(msg);\n break;\n case 'create-index':\n tx.processCreateIndex(msg);\n break;\n case 'drop-index':\n tx.processDropIndex(msg);\n break;\n case 'backfill':\n tx.processBackfill(msg);\n break;\n case 'backfill-completed':\n tx.processBackfillCompleted(msg);\n break;\n default:\n unreachable(msg);\n }\n\n return null;\n }\n}\n\n/**\n * The {@link TransactionProcessor} handles the sequence of messages from\n * upstream, from `BEGIN` to `COMMIT` and executes the corresponding mutations\n * on the {@link postgres.TransactionSql} on the replica.\n *\n * When applying row contents to the replica, the `_0_version` column is added / updated,\n * and a corresponding entry in the `ChangeLog` is added. The version value is derived\n * from the watermark of the preceding transaction (stored as the `nextStateVersion` in the\n * `ReplicationState` table).\n *\n * Side note: For non-streaming Postgres transactions, the commitEndLsn (and thus\n * commit watermark) is available in the `begin` message, so it could theoretically\n * be used for the row version of changes within the transaction. However, the\n * commitEndLsn is not available in the streaming (in-progress) transaction\n * protocol, and may not be available for CDC streams of other upstream types.\n * Therefore, the zero replication protocol is designed to not require the commit\n * watermark when a transaction begins.\n *\n * Also of interest is the fact that all INSERT Messages are logically applied as\n * UPSERTs. See {@link processInsert} for the underlying motivation.\n */\nclass TransactionProcessor {\n readonly #lc: LogContext;\n readonly #startMs: number;\n readonly #db: StatementRunner;\n readonly #mode: ChangeProcessorMode;\n readonly #version: LexiVersion;\n readonly #changeLog: ChangeLog;\n readonly #tableMetadata: TableMetadataTracker;\n readonly #tableSpecs: Map<string, LiteTableSpecWithReplicationStatus>;\n readonly #jsonFormat: JSONFormat;\n readonly #columnMetadata: ColumnMetadataStore;\n\n #pos = 0;\n #schemaChanged = false;\n #numChangeLogEntries = 0;\n\n constructor(\n lc: LogContext,\n db: StatementRunner,\n mode: ChangeProcessorMode,\n changeLog: ChangeLog,\n tableMetadata: TableMetadataTracker,\n tableSpecs: Map<string, LiteTableSpecWithReplicationStatus>,\n commitVersion: LexiVersion,\n jsonFormat: JSONFormat,\n ) {\n this.#startMs = Date.now();\n this.#mode = mode;\n this.#jsonFormat = jsonFormat;\n\n switch (mode) {\n case 'serving':\n // Although the Replicator / Incremental Syncer is the only writer of the replica,\n // a `BEGIN CONCURRENT` transaction is used to allow View Syncers to simulate\n // (i.e. and `ROLLBACK`) changes on historic snapshots of the database for the\n // purpose of IVM).\n //\n // This TransactionProcessor is the only logic that will actually\n // `COMMIT` any transactions to the replica.\n db.beginConcurrent();\n break;\n case 'backup':\n // For the backup-replicator (i.e. replication-manager), there are no View Syncers\n // and thus BEGIN CONCURRENT is not necessary. In fact, BEGIN CONCURRENT can cause\n // deadlocks with forced wal-checkpoints (which `litestream replicate` performs),\n // so it is important to use vanilla transactions in this configuration.\n db.beginImmediate();\n break;\n case 'initial-sync':\n // When the ChangeProcessor is used for initial-sync, the calling code\n // handles the transaction boundaries.\n break;\n default:\n unreachable();\n }\n this.#db = db;\n this.#version = commitVersion;\n this.#lc = lc.withContext('version', commitVersion);\n this.#changeLog = changeLog;\n this.#tableMetadata = tableMetadata;\n this.#tableSpecs = tableSpecs;\n // The column_metadata table is guaranteed to exist since the\n // replica-schema.ts migration to v8.\n this.#columnMetadata = must(ColumnMetadataStore.getInstance(db.db));\n\n if (this.#tableSpecs.size === 0) {\n this.#reloadTableSpecs();\n }\n }\n\n #reloadTableSpecs() {\n this.#tableSpecs.clear();\n // zqlSpecs include the primary key derived from unique indexes\n const zqlSpecs = computeZqlSpecs(this.#lc, this.#db.db, {\n includeBackfillingColumns: true,\n });\n for (let spec of listTables(this.#db.db)) {\n if (!spec.primaryKey) {\n spec = {\n ...spec,\n primaryKey: [\n ...(zqlSpecs.get(spec.name)?.tableSpec.primaryKey ?? []),\n ],\n };\n }\n this.#tableSpecs.set(spec.name, spec);\n }\n }\n\n #tableSpec(name: string) {\n return must(this.#tableSpecs.get(name), `Unknown table ${name}`);\n }\n\n #getKey(\n {row, numCols}: {row: LiteRow; numCols: number},\n {relation}: {relation: MessageRelation},\n ): LiteRowKey {\n const keyColumns =\n relation.rowKey.type !== 'full'\n ? relation.rowKey.columns // already a suitable key\n : this.#tableSpec(liteTableName(relation)).primaryKey;\n if (!keyColumns?.length) {\n throw new Error(\n `Cannot replicate table \"${relation.name}\" without a PRIMARY KEY or UNIQUE INDEX`,\n );\n }\n // For the common case (replica identity default), the row is already the\n // key for deletes and updates, in which case a new object can be avoided.\n if (numCols === keyColumns.length) {\n return row;\n }\n const key: Record<string, LiteValueType> = {};\n for (const col of keyColumns) {\n key[col] = row[col];\n }\n return key;\n }\n\n processInsert(insert: MessageInsert) {\n const table = liteTableName(insert.relation);\n const tableSpec = this.#tableSpec(table);\n const newRow = liteRow(insert.new, tableSpec, this.#jsonFormat);\n\n this.#upsert(table, {\n ...newRow.row,\n [ZERO_VERSION_COLUMN_NAME]: this.#version,\n });\n\n if (insert.relation.rowKey.columns.length === 0) {\n // INSERTs can be replicated for rows without a PRIMARY KEY or a\n // UNIQUE INDEX. These are written to the replica but not recorded\n // in the changeLog, because these rows cannot participate in IVM.\n //\n // (Once the table schema has been corrected to include a key, the\n // associated schema change will reset pipelines and data can be\n // loaded via hydration.)\n return;\n }\n const key = this.#getKey(newRow, insert);\n this.#logSetOp(table, key, getBackfilledColumns(newRow.row, tableSpec));\n }\n\n #upsert(table: string, row: LiteRow) {\n const columns = Object.keys(row).map(c => id(c));\n this.#db.run(\n `\n INSERT OR REPLACE INTO ${id(table)} (${columns.join(',')})\n VALUES (${Array.from({length: columns.length}).fill('?').join(',')})\n `,\n Object.values(row),\n );\n }\n\n // Updates by default are applied as UPDATE commands to support partial\n // row specifications from the change source. In particular, this is needed\n // to handle updates for which unchanged TOASTed values are not sent:\n //\n // https://www.postgresql.org/docs/current/protocol-logicalrep-message-formats.html#PROTOCOL-LOGICALREP-MESSAGE-FORMATS-TUPLEDATA\n //\n // However, in certain cases an UPDATE may be received for a row that\n // was not initially synced, such as when, an existing table is added\n // to the app's publication.\n //\n // In order to facilitate \"resumptive\" replication, the logic falls back to\n // an INSERT if the update did not change any rows.\n processUpdate(update: MessageUpdate) {\n const table = liteTableName(update.relation);\n const tableSpec = this.#tableSpec(table);\n const newRow = liteRow(update.new, tableSpec, this.#jsonFormat);\n const row = {...newRow.row, [ZERO_VERSION_COLUMN_NAME]: this.#version};\n\n // update.key is set with the old values if the key has changed.\n const oldKey = update.key\n ? this.#getKey(\n liteRow(update.key, this.#tableSpec(table), this.#jsonFormat),\n update,\n )\n : null;\n const newKey = this.#getKey(newRow, update);\n\n if (oldKey) {\n this.#logDeleteOp(table, oldKey, tableSpec.backfilling);\n }\n this.#logSetOp(table, newKey, getBackfilledColumns(newRow.row, tableSpec));\n\n const currKey = oldKey ?? newKey;\n const conds = Object.keys(currKey).map(col => `${id(col)}=?`);\n const setExprs = Object.keys(row).map(col => `${id(col)}=?`);\n\n const {changes} = this.#db.run(\n `\n UPDATE ${id(table)}\n SET ${setExprs.join(',')}\n WHERE ${conds.join(' AND ')}\n `,\n [...Object.values(row), ...Object.values(currKey)],\n );\n\n // If the UPDATE did not affect any rows, perform an UPSERT of the\n // new row for resumptive replication.\n if (changes === 0) {\n this.#upsert(table, row);\n }\n }\n\n processDelete(del: MessageDelete) {\n const table = liteTableName(del.relation);\n const tableSpec = this.#tableSpec(table);\n const rowKey = this.#getKey(\n liteRow(del.key, tableSpec, this.#jsonFormat),\n del,\n );\n\n this.#delete(table, rowKey);\n this.#logDeleteOp(table, rowKey, tableSpec.backfilling);\n }\n\n #delete(table: string, rowKey: LiteRowKey) {\n const conds = Object.keys(rowKey).map(col => `${id(col)}=?`);\n this.#db.run(\n `DELETE FROM ${id(table)} WHERE ${conds.join(' AND ')}`,\n Object.values(rowKey),\n );\n }\n\n processTruncate(truncate: MessageTruncate) {\n for (const relation of truncate.relations) {\n const table = liteTableName(relation);\n // Update replica data.\n this.#db.run(`DELETE FROM ${id(table)}`);\n\n // Update change log.\n this.#logTruncateOp(table);\n }\n }\n\n processCreateTable(create: TableCreate) {\n if (create.metadata) {\n this.#tableMetadata.setUpstreamMetadata(create.spec, create.metadata);\n }\n const table = mapPostgresToLite(create.spec);\n this.#db.db.exec(createLiteTableStatement(table));\n\n // Write to metadata table\n for (const [colName, colSpec] of Object.entries(create.spec.columns)) {\n this.#columnMetadata.insert(\n table.name,\n colName,\n colSpec,\n create.backfill?.[colName],\n );\n }\n\n if (\n Object.keys(create.backfill ?? {}).length ===\n Object.keys(create.spec.columns).length\n ) {\n this.#reloadTableSpecs();\n } else {\n // Make the table visible immediately unless all of the columns are\n // being backfilled. In the backfill case, the version bump will happen\n // with the backfill is complete.\n this.#logResetOp(table.name);\n }\n this.#lc.info?.(create.tag, table.name);\n }\n\n processTableMetadata(msg: TableUpdateMetadata) {\n this.#tableMetadata.setUpstreamMetadata(msg.table, msg.new);\n }\n\n processRenameTable(rename: TableRename) {\n this.#tableMetadata.rename(rename.old, rename.new);\n\n const oldName = liteTableName(rename.old);\n const newName = liteTableName(rename.new);\n this.#db.db.exec(`ALTER TABLE ${id(oldName)} RENAME TO ${id(newName)}`);\n\n // Rename in metadata table\n this.#columnMetadata.renameTable(oldName, newName);\n\n this.#bumpVersions(rename.new);\n this.#logResetOp(oldName);\n this.#lc.info?.(rename.tag, oldName, newName);\n }\n\n processAddColumn(msg: ColumnAdd) {\n if (msg.tableMetadata) {\n this.#tableMetadata.setUpstreamMetadata(msg.table, msg.tableMetadata);\n }\n const table = liteTableName(msg.table);\n const {name} = msg.column;\n const spec = mapPostgresToLiteColumn(table, msg.column);\n this.#db.db.exec(\n `ALTER TABLE ${id(table)} ADD ${id(name)} ${liteColumnDef(spec)}`,\n );\n\n // Write to metadata table\n this.#columnMetadata.insert(table, name, msg.column.spec, msg.backfill);\n\n if (msg.backfill) {\n this.#reloadTableSpecs();\n } else {\n // Make the new column visible immediately if it's not being backfilled.\n // Otherwise, the version bump will happen with the backfill is complete.\n this.#bumpVersions(msg.table);\n }\n this.#lc.info?.(msg.tag, table, msg.column);\n }\n\n processUpdateColumn(msg: ColumnUpdate) {\n const table = liteTableName(msg.table);\n let oldName = msg.old.name;\n const newName = msg.new.name;\n\n // update-column can ignore defaults because it does not change the values\n // in existing rows.\n //\n // https://www.postgresql.org/docs/current/sql-altertable.html#SQL-ALTERTABLE-DESC-SET-DROP-DEFAULT\n //\n // \"The new default value will only apply in subsequent INSERT or UPDATE\n // commands; it does not cause rows already in the table to change.\"\n //\n // This allows support for _changing_ column defaults to any expression,\n // since it does not affect what the replica needs to do.\n const oldSpec = mapPostgresToLiteColumn(table, msg.old, 'ignore-default');\n const newSpec = mapPostgresToLiteColumn(table, msg.new, 'ignore-default');\n\n // The only updates that are relevant are the column name and the data type.\n if (oldName === newName && oldSpec.dataType === newSpec.dataType) {\n this.#lc.info?.(msg.tag, 'no thing to update', oldSpec, newSpec);\n return;\n }\n // If the data type changes, we have to make a new column with the new data type\n // and copy the values over.\n if (oldSpec.dataType !== newSpec.dataType) {\n // Remember (and drop) the indexes that reference the column.\n const indexes = listIndexes(this.#db.db).filter(\n idx => idx.tableName === table && oldName in idx.columns,\n );\n const stmts = indexes.map(idx => `DROP INDEX IF EXISTS ${id(idx.name)};`);\n const tmpName = `tmp.${newName}`;\n stmts.push(`\n ALTER TABLE ${id(table)} ADD ${id(tmpName)} ${liteColumnDef(newSpec)};\n UPDATE ${id(table)} SET ${id(tmpName)} = ${id(oldName)};\n ALTER TABLE ${id(table)} DROP ${id(oldName)};\n `);\n for (const idx of indexes) {\n // Re-create the indexes to reference the new column.\n idx.columns[tmpName] = idx.columns[oldName];\n delete idx.columns[oldName];\n stmts.push(createLiteIndexStatement(idx));\n }\n this.#db.db.exec(stmts.join(''));\n oldName = tmpName;\n }\n if (oldName !== newName) {\n this.#db.db.exec(\n `ALTER TABLE ${id(table)} RENAME ${id(oldName)} TO ${id(newName)}`,\n );\n }\n\n // Update metadata table\n this.#columnMetadata.update(\n table,\n msg.old.name,\n msg.new.name,\n msg.new.spec,\n );\n\n this.#bumpVersions(msg.table);\n this.#lc.info?.(msg.tag, table, msg.new);\n }\n\n processDropColumn(msg: ColumnDrop) {\n const table = liteTableName(msg.table);\n const {column} = msg;\n this.#db.db.exec(`ALTER TABLE ${id(table)} DROP ${id(column)}`);\n\n // Delete from metadata table\n this.#columnMetadata.deleteColumn(table, column);\n\n this.#bumpVersions(msg.table);\n this.#lc.info?.(msg.tag, table, column);\n }\n\n processDropTable(drop: TableDrop) {\n this.#tableMetadata.drop(drop.id);\n\n const name = liteTableName(drop.id);\n this.#db.db.exec(`DROP TABLE IF EXISTS ${id(name)}`);\n\n // Delete from metadata table\n this.#columnMetadata.deleteTable(name);\n\n this.#logResetOp(name);\n this.#lc.info?.(drop.tag, name);\n }\n\n processCreateIndex(create: IndexCreate) {\n const index = mapPostgresToLiteIndex(create.spec);\n this.#db.db.exec(createLiteIndexStatement(index));\n\n // indexes affect tables visibility (e.g. sync-ability is gated on\n // having a unique index), so reset pipelines to refresh table schemas.\n // However, the reset is not necessary if the index is for a table\n // that is not yet visible due to backfilling.\n const tableSpec = must(this.#tableSpecs.get(index.tableName));\n if (\n (tableSpec.backfilling ?? []).length ===\n Object.entries(tableSpec.columns).length - 1 // don't count _0_version\n ) {\n this.#reloadTableSpecs();\n } else {\n this.#logResetOp(index.tableName);\n }\n this.#lc.info?.(create.tag, index.name);\n }\n\n processDropIndex(drop: IndexDrop) {\n const name = liteTableName(drop.id);\n this.#db.db.exec(`DROP INDEX IF EXISTS ${id(name)}`);\n this.#lc.info?.(drop.tag, name);\n }\n\n #bumpVersions(table: Identifier) {\n this.#tableMetadata.setMinRowVersion(table, this.#version);\n this.#logResetOp(liteTableName(table));\n }\n\n /**\n * @param backfilledColumns `backfilling` columns for which values were set\n */\n #logSetOp(\n table: string,\n key: LiteRowKey,\n backfilledColumns: string[] | undefined,\n ) {\n // The \"serving\" replicator always writes to the change-log (for IVM).\n // The \"backup\" replicator only needs to write to the change log\n // when writing columns that are being backfilled.\n if (this.#mode === 'serving' || backfilledColumns !== undefined) {\n this.#changeLog.logSetOp(\n this.#version,\n this.#pos++,\n table,\n key,\n backfilledColumns,\n );\n this.#numChangeLogEntries++;\n }\n }\n\n #logDeleteOp(table: string, key: LiteRowKey, backfilling?: string[]) {\n // The \"serving\" replicator always writes to the change-log (for IVM).\n // The \"backup\" replicator only needs to write to the change log\n // when writing columns that are being backfilled.\n if (this.#mode === 'serving' || backfilling?.length) {\n this.#changeLog.logDeleteOp(this.#version, this.#pos++, table, key);\n this.#numChangeLogEntries++;\n }\n }\n\n #logTruncateOp(table: string) {\n if (this.#mode === 'serving') {\n this.#changeLog.logTruncateOp(this.#version, table);\n this.#numChangeLogEntries++;\n }\n }\n\n #logResetOp(table: string) {\n this.#schemaChanged = true;\n if (this.#mode === 'serving') {\n this.#changeLog.logResetOp(this.#version, table);\n this.#numChangeLogEntries++;\n }\n this.#reloadTableSpecs();\n }\n\n processBackfill({relation, watermark, columns, rowValues}: MessageBackfill) {\n const tableName = liteTableName(relation);\n const tableSpec = must(this.#tableSpecs.get(tableName));\n const rowKeyCols = relation.rowKey.columns;\n const cols = [...rowKeyCols, ...columns];\n\n // Common parts of the INSERT sql statement.\n const insertColsStr = [...cols, ZERO_VERSION_COLUMN_NAME].map(id).join(',');\n const qMarks = Array.from({length: cols.length + 1})\n .fill('?')\n .join(',');\n const rowKeyColsStr = rowKeyCols.map(id).join(',');\n\n let backfilled = 0;\n let skipped = 0;\n for (const v of rowValues) {\n const row = liteRow(\n Object.fromEntries(cols.map((c, i) => [c, v[i]])),\n tableSpec,\n this.#jsonFormat,\n );\n const rowKey = this.#getKey(row, {relation});\n const rowOp = this.#changeLog.getLatestRowOp(tableName, rowKey);\n if (rowOp?.op === DEL_OP && rowOp.stateVersion > watermark) {\n skipped++;\n continue; // the row was deleted after the backfill snapshot\n }\n const updates =\n rowOp?.op === SET_OP\n ? cols.filter(\n c => (rowOp.backfillingColumnVersions[c] ?? '') <= watermark,\n )\n : cols;\n if (updates.length === 0) {\n // row already has newer values for all backfilling columns.\n skipped++;\n continue;\n }\n const updateStmts = updates.map(col => `${id(col)}=excluded.${id(col)}`);\n this.#db.run(\n /*sql*/ `\n INSERT INTO ${id(tableName)} (${insertColsStr}) VALUES (${qMarks})\n ON CONFLICT (${rowKeyColsStr})\n DO UPDATE SET ${updateStmts.join(',')};\n `,\n ...Object.values(row.row),\n watermark, // the _0_version for new rows (i.e. table backfill)\n );\n backfilled++;\n }\n\n this.#lc.debug?.(\n `backfilled ${backfilled} rows (skipped ${skipped}) into ${tableName}`,\n );\n }\n\n #completedBackfill: DownloadStatus | undefined;\n\n processBackfillCompleted({relation, columns, status}: BackfillCompleted) {\n const tableName = liteTableName(relation);\n const rowKeyCols = relation.rowKey.columns;\n const cols = [...rowKeyCols, ...columns];\n\n const columnMetadata = must(ColumnMetadataStore.getInstance(this.#db.db));\n for (const col of cols) {\n columnMetadata.clearBackfilling(tableName, col);\n }\n // Given that new columns are being exposed for every row in the table, bump the\n // row version for all rows.\n this.#bumpVersions(relation);\n if (status) {\n this.#completedBackfill = {table: tableName, columns: cols, ...status};\n }\n this.#lc.info?.(`finished backfilling ${tableName}`);\n\n // Note that there is no need to clear the backfillingColumnVersions values\n // in the changeLog. It could theoretically be done for clarity but:\n // (1) it could be non-trivial in terms of latency introduced and\n // (2) the data must be preserved if _other_ columns are in the process\n // of being backfilled\n //\n // Thus, for speed and simplicity, the values are left as is. (Note that\n // subsequent replicated changes to those rows will clear the values if\n // no backfills are in progress).\n }\n\n processCommit(commit: MessageCommit, watermark: string): CommitResult {\n if (watermark !== this.#version) {\n throw new Error(\n `'commit' version ${watermark} does not match 'begin' version ${\n this.#version\n }: ${stringify(commit)}`,\n );\n }\n updateReplicationWatermark(this.#db, watermark);\n\n if (this.#schemaChanged) {\n const start = Date.now();\n this.#db.db.pragma('optimize');\n this.#lc.info?.(\n `PRAGMA optimized after schema change (${Date.now() - start} ms)`,\n );\n }\n\n if (this.#mode !== 'initial-sync') {\n this.#db.commit();\n }\n\n const elapsedMs = Date.now() - this.#startMs;\n this.#lc.debug?.(`Committed tx@${this.#version} (${elapsedMs} ms)`);\n\n return {\n watermark,\n completedBackfill: this.#completedBackfill,\n schemaUpdated: this.#schemaChanged,\n changeLogUpdated: this.#numChangeLogEntries > 0,\n };\n }\n\n abort(lc: LogContext) {\n lc.info?.(`aborting transaction ${this.#version}`);\n this.#db.rollback();\n }\n}\n\nfunction getBackfilledColumns(\n row: LiteRow,\n {backfilling}: LiteTableSpecWithReplicationStatus,\n): string[] | undefined {\n if (!backfilling?.length) {\n return undefined; // common case\n }\n return backfilling.filter(col => col in row);\n}\n\nfunction ensureError(err: unknown): Error {\n if (err instanceof Error) {\n return err;\n }\n const error = new Error();\n error.cause = err;\n return error;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsFA,IAAa,kBAAb,MAA6B;CAC3B;CACA;CACA;CACA;CACA;CAKA,8BAAuB,IAAI,KAAiD;CAE5E,aAA0C;CAE1C;CAEA,YACE,IACA,MACA,aACA;AACA,QAAA,KAAW;AACX,QAAA,YAAkB,IAAI,UAAU,GAAG,GAAG;AACtC,QAAA,gBAAsB,IAAI,qBAAqB,GAAG,GAAG;AACrD,QAAA,OAAa;AACb,QAAA,cAAoB;;CAGtB,MAAM,IAAgB,KAAc;AAClC,MAAI,CAAC,MAAA,SAAe;AAClB,SAAA,WAAiB,MAAM,GAAG;AAE1B,SAAA,UAAgB,YAAY,IAAI;AAEhC,OAAI,EAAE,eAAe,aAAa;AAEhC,OAAG,QAAQ,8BAA8B,MAAA,QAAc;AACvD,UAAA,YAAkB,IAAI,MAAA,QAAc;;;;CAK1C,MAAM,IAAgB;AACpB,QAAA,KAAW,IAAI,IAAI,YAAY,CAAC;;;CAIlC,eACE,IACA,YACqB;EACrB,MAAM,CAAC,MAAM,WAAW;AACxB,MAAI,MAAA,SAAe;AACjB,MAAG,QAAQ,YAAY,QAAQ,MAAM;AACrC,UAAO;;AAET,MAAI;GACF,MAAM,YACJ,SAAS,UACL,WAAW,GAAG,kBACd,SAAS,WACP,WAAW,GAAG,YACd,KAAA;AACR,UAAO,MAAA,eAAqB,IAAI,SAAS,UAAU;WAC5C,GAAG;AACV,SAAA,KAAW,IAAI,EAAE;;AAEnB,SAAO;;CAGT,kBACE,IACA,eACA,YACsB;EACtB,MAAM,QAAQ,KAAK,KAAK;AASxB,OAAK,IAAI,IAAI,IAAK,IAChB,KAAI;AACF,UAAO,IAAI,qBACT,IACA,MAAA,IACA,MAAA,MACA,MAAA,WACA,MAAA,eACA,MAAA,YACA,eACA,WACD;WACM,GAAG;AACV,OAAI,aAAa,eAAe,EAAE,SAAS,eAAe;AACxD,OAAG,OACD,mBAAmB,KAAK,KAAK,GAAG,MAAM,eAAe,IAAI,EAAE,2EAG3D,EACD;AACD;;AAEF,SAAM;;;;CAMZ,gBACE,IACA,KACA,WACqB;AACrB,MAAI,IAAI,QAAQ,SAAS;AACvB,OAAI,MAAA,UACF,OAAM,IAAI,MAAM,4BAA4B,UAAU,IAAI,GAAG;AAE/D,SAAA,YAAkB,MAAA,iBAChB,IACA,KAAK,UAAU,EACf,IAAI,QAAA,IACL;AACD,UAAO;;EAIT,MAAM,KAAK,MAAA;AACX,MAAI,CAAC,GACH,OAAM,IAAI,MACR,4CAA4C,UAAU,IAAI,GAC3D;AAGH,MAAI,IAAI,QAAQ,UAAU;AAExB,SAAA,YAAkB;AAElB,UAAO,WAAW,4CAA4C;AAC9D,UAAO,GAAG,cAAc,KAAK,UAAU;;AAGzC,MAAI,IAAI,QAAQ,YAAY;AAC1B,SAAA,WAAiB,MAAM,GAAG;AAC1B,SAAA,YAAkB;AAClB,UAAO;;AAGT,UAAQ,IAAI,KAAZ;GACE,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,gBAAgB,IAAI;AACvB;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,qBAAqB,IAAI;AAC5B;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,oBAAoB,IAAI;AAC3B;GACF,KAAK;AACH,OAAG,kBAAkB,IAAI;AACzB;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,gBAAgB,IAAI;AACvB;GACF,KAAK;AACH,OAAG,yBAAyB,IAAI;AAChC;GACF,QACE,aAAY,IAAI;;AAGpB,SAAO;;;;;;;;;;;;;;;;;;;;;;;;AAyBX,IAAM,uBAAN,MAA2B;CACzB;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CAEA,OAAO;CACP,iBAAiB;CACjB,uBAAuB;CAEvB,YACE,IACA,IACA,MACA,WACA,eACA,YACA,eACA,YACA;AACA,QAAA,UAAgB,KAAK,KAAK;AAC1B,QAAA,OAAa;AACb,QAAA,aAAmB;AAEnB,UAAQ,MAAR;GACE,KAAK;AAQH,OAAG,iBAAiB;AACpB;GACF,KAAK;AAKH,OAAG,gBAAgB;AACnB;GACF,KAAK,eAGH;GACF,QACE,cAAa;;AAEjB,QAAA,KAAW;AACX,QAAA,UAAgB;AAChB,QAAA,KAAW,GAAG,YAAY,WAAW,cAAc;AACnD,QAAA,YAAkB;AAClB,QAAA,gBAAsB;AACtB,QAAA,aAAmB;AAGnB,QAAA,iBAAuB,KAAK,oBAAoB,YAAY,GAAG,GAAG,CAAC;AAEnE,MAAI,MAAA,WAAiB,SAAS,EAC5B,OAAA,kBAAwB;;CAI5B,oBAAoB;AAClB,QAAA,WAAiB,OAAO;EAExB,MAAM,WAAW,gBAAgB,MAAA,IAAU,MAAA,GAAS,IAAI,EACtD,2BAA2B,MAC5B,CAAC;AACF,OAAK,IAAI,QAAQ,WAAW,MAAA,GAAS,GAAG,EAAE;AACxC,OAAI,CAAC,KAAK,WACR,QAAO;IACL,GAAG;IACH,YAAY,CACV,GAAI,SAAS,IAAI,KAAK,KAAK,EAAE,UAAU,cAAc,EAAE,CACxD;IACF;AAEH,SAAA,WAAiB,IAAI,KAAK,MAAM,KAAK;;;CAIzC,WAAW,MAAc;AACvB,SAAO,KAAK,MAAA,WAAiB,IAAI,KAAK,EAAE,iBAAiB,OAAO;;CAGlE,QACE,EAAC,KAAK,WACN,EAAC,YACW;EACZ,MAAM,aACJ,SAAS,OAAO,SAAS,SACrB,SAAS,OAAO,UAChB,MAAA,UAAgB,cAAc,SAAS,CAAC,CAAC;AAC/C,MAAI,CAAC,YAAY,OACf,OAAM,IAAI,MACR,2BAA2B,SAAS,KAAK,yCAC1C;AAIH,MAAI,YAAY,WAAW,OACzB,QAAO;EAET,MAAM,MAAqC,EAAE;AAC7C,OAAK,MAAM,OAAO,WAChB,KAAI,OAAO,IAAI;AAEjB,SAAO;;CAGT,cAAc,QAAuB;EACnC,MAAM,QAAQ,cAAc,OAAO,SAAS;EAC5C,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,QAAQ,OAAO,KAAK,WAAW,MAAA,WAAiB;AAE/D,QAAA,OAAa,OAAO;GAClB,GAAG,OAAO;IACT,2BAA2B,MAAA;GAC7B,CAAC;AAEF,MAAI,OAAO,SAAS,OAAO,QAAQ,WAAW,EAQ5C;EAEF,MAAM,MAAM,MAAA,OAAa,QAAQ,OAAO;AACxC,QAAA,SAAe,OAAO,KAAK,qBAAqB,OAAO,KAAK,UAAU,CAAC;;CAGzE,QAAQ,OAAe,KAAc;EACnC,MAAM,UAAU,OAAO,KAAK,IAAI,CAAC,KAAI,MAAK,GAAG,EAAE,CAAC;AAChD,QAAA,GAAS,IACP;+BACyB,GAAG,MAAM,CAAC,IAAI,QAAQ,KAAK,IAAI,CAAC;kBAC7C,MAAM,KAAK,EAAC,QAAQ,QAAQ,QAAO,CAAC,CAAC,KAAK,IAAI,CAAC,KAAK,IAAI,CAAC;SAErE,OAAO,OAAO,IAAI,CACnB;;CAeH,cAAc,QAAuB;EACnC,MAAM,QAAQ,cAAc,OAAO,SAAS;EAC5C,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,QAAQ,OAAO,KAAK,WAAW,MAAA,WAAiB;EAC/D,MAAM,MAAM;GAAC,GAAG,OAAO;IAAM,2BAA2B,MAAA;GAAc;EAGtE,MAAM,SAAS,OAAO,MAClB,MAAA,OACE,QAAQ,OAAO,KAAK,MAAA,UAAgB,MAAM,EAAE,MAAA,WAAiB,EAC7D,OACD,GACD;EACJ,MAAM,SAAS,MAAA,OAAa,QAAQ,OAAO;AAE3C,MAAI,OACF,OAAA,YAAkB,OAAO,QAAQ,UAAU,YAAY;AAEzD,QAAA,SAAe,OAAO,QAAQ,qBAAqB,OAAO,KAAK,UAAU,CAAC;EAE1E,MAAM,UAAU,UAAU;EAC1B,MAAM,QAAQ,OAAO,KAAK,QAAQ,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;EAC7D,MAAM,WAAW,OAAO,KAAK,IAAI,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;EAE5D,MAAM,EAAC,YAAW,MAAA,GAAS,IACzB;eACS,GAAG,MAAM,CAAC;cACX,SAAS,KAAK,IAAI,CAAC;gBACjB,MAAM,KAAK,QAAQ,CAAC;SAE9B,CAAC,GAAG,OAAO,OAAO,IAAI,EAAE,GAAG,OAAO,OAAO,QAAQ,CAAC,CACnD;AAID,MAAI,YAAY,EACd,OAAA,OAAa,OAAO,IAAI;;CAI5B,cAAc,KAAoB;EAChC,MAAM,QAAQ,cAAc,IAAI,SAAS;EACzC,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,MAAA,OACb,QAAQ,IAAI,KAAK,WAAW,MAAA,WAAiB,EAC7C,IACD;AAED,QAAA,OAAa,OAAO,OAAO;AAC3B,QAAA,YAAkB,OAAO,QAAQ,UAAU,YAAY;;CAGzD,QAAQ,OAAe,QAAoB;EACzC,MAAM,QAAQ,OAAO,KAAK,OAAO,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;AAC5D,QAAA,GAAS,IACP,eAAe,GAAG,MAAM,CAAC,SAAS,MAAM,KAAK,QAAQ,IACrD,OAAO,OAAO,OAAO,CACtB;;CAGH,gBAAgB,UAA2B;AACzC,OAAK,MAAM,YAAY,SAAS,WAAW;GACzC,MAAM,QAAQ,cAAc,SAAS;AAErC,SAAA,GAAS,IAAI,eAAe,GAAG,MAAM,GAAG;AAGxC,SAAA,cAAoB,MAAM;;;CAI9B,mBAAmB,QAAqB;AACtC,MAAI,OAAO,SACT,OAAA,cAAoB,oBAAoB,OAAO,MAAM,OAAO,SAAS;EAEvE,MAAM,QAAQ,kBAAkB,OAAO,KAAK;AAC5C,QAAA,GAAS,GAAG,KAAK,yBAAyB,MAAM,CAAC;AAGjD,OAAK,MAAM,CAAC,SAAS,YAAY,OAAO,QAAQ,OAAO,KAAK,QAAQ,CAClE,OAAA,eAAqB,OACnB,MAAM,MACN,SACA,SACA,OAAO,WAAW,SACnB;AAGH,MACE,OAAO,KAAK,OAAO,YAAY,EAAE,CAAC,CAAC,WACnC,OAAO,KAAK,OAAO,KAAK,QAAQ,CAAC,OAEjC,OAAA,kBAAwB;MAKxB,OAAA,WAAiB,MAAM,KAAK;AAE9B,QAAA,GAAS,OAAO,OAAO,KAAK,MAAM,KAAK;;CAGzC,qBAAqB,KAA0B;AAC7C,QAAA,cAAoB,oBAAoB,IAAI,OAAO,IAAI,IAAI;;CAG7D,mBAAmB,QAAqB;AACtC,QAAA,cAAoB,OAAO,OAAO,KAAK,OAAO,IAAI;EAElD,MAAM,UAAU,cAAc,OAAO,IAAI;EACzC,MAAM,UAAU,cAAc,OAAO,IAAI;AACzC,QAAA,GAAS,GAAG,KAAK,eAAe,GAAG,QAAQ,CAAC,aAAa,GAAG,QAAQ,GAAG;AAGvE,QAAA,eAAqB,YAAY,SAAS,QAAQ;AAElD,QAAA,aAAmB,OAAO,IAAI;AAC9B,QAAA,WAAiB,QAAQ;AACzB,QAAA,GAAS,OAAO,OAAO,KAAK,SAAS,QAAQ;;CAG/C,iBAAiB,KAAgB;AAC/B,MAAI,IAAI,cACN,OAAA,cAAoB,oBAAoB,IAAI,OAAO,IAAI,cAAc;EAEvE,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,MAAM,EAAC,SAAQ,IAAI;EACnB,MAAM,OAAO,wBAAwB,OAAO,IAAI,OAAO;AACvD,QAAA,GAAS,GAAG,KACV,eAAe,GAAG,MAAM,CAAC,OAAO,GAAG,KAAK,CAAC,GAAG,cAAc,KAAK,GAChE;AAGD,QAAA,eAAqB,OAAO,OAAO,MAAM,IAAI,OAAO,MAAM,IAAI,SAAS;AAEvE,MAAI,IAAI,SACN,OAAA,kBAAwB;MAIxB,OAAA,aAAmB,IAAI,MAAM;AAE/B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,IAAI,OAAO;;CAG7C,oBAAoB,KAAmB;EACrC,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,IAAI,UAAU,IAAI,IAAI;EACtB,MAAM,UAAU,IAAI,IAAI;EAYxB,MAAM,UAAU,wBAAwB,OAAO,IAAI,KAAK,iBAAiB;EACzE,MAAM,UAAU,wBAAwB,OAAO,IAAI,KAAK,iBAAiB;AAGzE,MAAI,YAAY,WAAW,QAAQ,aAAa,QAAQ,UAAU;AAChE,SAAA,GAAS,OAAO,IAAI,KAAK,sBAAsB,SAAS,QAAQ;AAChE;;AAIF,MAAI,QAAQ,aAAa,QAAQ,UAAU;GAEzC,MAAM,UAAU,YAAY,MAAA,GAAS,GAAG,CAAC,QACvC,QAAO,IAAI,cAAc,SAAS,WAAW,IAAI,QAClD;GACD,MAAM,QAAQ,QAAQ,KAAI,QAAO,wBAAwB,GAAG,IAAI,KAAK,CAAC,GAAG;GACzE,MAAM,UAAU,OAAO;AACvB,SAAM,KAAK;sBACK,GAAG,MAAM,CAAC,OAAO,GAAG,QAAQ,CAAC,GAAG,cAAc,QAAQ,CAAC;iBAC5D,GAAG,MAAM,CAAC,OAAO,GAAG,QAAQ,CAAC,KAAK,GAAG,QAAQ,CAAC;sBACzC,GAAG,MAAM,CAAC,QAAQ,GAAG,QAAQ,CAAC;UAC1C;AACJ,QAAK,MAAM,OAAO,SAAS;AAEzB,QAAI,QAAQ,WAAW,IAAI,QAAQ;AACnC,WAAO,IAAI,QAAQ;AACnB,UAAM,KAAK,yBAAyB,IAAI,CAAC;;AAE3C,SAAA,GAAS,GAAG,KAAK,MAAM,KAAK,GAAG,CAAC;AAChC,aAAU;;AAEZ,MAAI,YAAY,QACd,OAAA,GAAS,GAAG,KACV,eAAe,GAAG,MAAM,CAAC,UAAU,GAAG,QAAQ,CAAC,MAAM,GAAG,QAAQ,GACjE;AAIH,QAAA,eAAqB,OACnB,OACA,IAAI,IAAI,MACR,IAAI,IAAI,MACR,IAAI,IAAI,KACT;AAED,QAAA,aAAmB,IAAI,MAAM;AAC7B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,IAAI,IAAI;;CAG1C,kBAAkB,KAAiB;EACjC,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,MAAM,EAAC,WAAU;AACjB,QAAA,GAAS,GAAG,KAAK,eAAe,GAAG,MAAM,CAAC,QAAQ,GAAG,OAAO,GAAG;AAG/D,QAAA,eAAqB,aAAa,OAAO,OAAO;AAEhD,QAAA,aAAmB,IAAI,MAAM;AAC7B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,OAAO;;CAGzC,iBAAiB,MAAiB;AAChC,QAAA,cAAoB,KAAK,KAAK,GAAG;EAEjC,MAAM,OAAO,cAAc,KAAK,GAAG;AACnC,QAAA,GAAS,GAAG,KAAK,wBAAwB,GAAG,KAAK,GAAG;AAGpD,QAAA,eAAqB,YAAY,KAAK;AAEtC,QAAA,WAAiB,KAAK;AACtB,QAAA,GAAS,OAAO,KAAK,KAAK,KAAK;;CAGjC,mBAAmB,QAAqB;EACtC,MAAM,QAAQ,uBAAuB,OAAO,KAAK;AACjD,QAAA,GAAS,GAAG,KAAK,yBAAyB,MAAM,CAAC;EAMjD,MAAM,YAAY,KAAK,MAAA,WAAiB,IAAI,MAAM,UAAU,CAAC;AAC7D,OACG,UAAU,eAAe,EAAE,EAAE,WAC9B,OAAO,QAAQ,UAAU,QAAQ,CAAC,SAAS,EAE3C,OAAA,kBAAwB;MAExB,OAAA,WAAiB,MAAM,UAAU;AAEnC,QAAA,GAAS,OAAO,OAAO,KAAK,MAAM,KAAK;;CAGzC,iBAAiB,MAAiB;EAChC,MAAM,OAAO,cAAc,KAAK,GAAG;AACnC,QAAA,GAAS,GAAG,KAAK,wBAAwB,GAAG,KAAK,GAAG;AACpD,QAAA,GAAS,OAAO,KAAK,KAAK,KAAK;;CAGjC,cAAc,OAAmB;AAC/B,QAAA,cAAoB,iBAAiB,OAAO,MAAA,QAAc;AAC1D,QAAA,WAAiB,cAAc,MAAM,CAAC;;;;;CAMxC,UACE,OACA,KACA,mBACA;AAIA,MAAI,MAAA,SAAe,aAAa,sBAAsB,KAAA,GAAW;AAC/D,SAAA,UAAgB,SACd,MAAA,SACA,MAAA,OACA,OACA,KACA,kBACD;AACD,SAAA;;;CAIJ,aAAa,OAAe,KAAiB,aAAwB;AAInE,MAAI,MAAA,SAAe,aAAa,aAAa,QAAQ;AACnD,SAAA,UAAgB,YAAY,MAAA,SAAe,MAAA,OAAa,OAAO,IAAI;AACnE,SAAA;;;CAIJ,eAAe,OAAe;AAC5B,MAAI,MAAA,SAAe,WAAW;AAC5B,SAAA,UAAgB,cAAc,MAAA,SAAe,MAAM;AACnD,SAAA;;;CAIJ,YAAY,OAAe;AACzB,QAAA,gBAAsB;AACtB,MAAI,MAAA,SAAe,WAAW;AAC5B,SAAA,UAAgB,WAAW,MAAA,SAAe,MAAM;AAChD,SAAA;;AAEF,QAAA,kBAAwB;;CAG1B,gBAAgB,EAAC,UAAU,WAAW,SAAS,aAA6B;EAC1E,MAAM,YAAY,cAAc,SAAS;EACzC,MAAM,YAAY,KAAK,MAAA,WAAiB,IAAI,UAAU,CAAC;EACvD,MAAM,aAAa,SAAS,OAAO;EACnC,MAAM,OAAO,CAAC,GAAG,YAAY,GAAG,QAAQ;EAGxC,MAAM,gBAAgB,CAAC,GAAG,MAAM,yBAAyB,CAAC,IAAI,GAAG,CAAC,KAAK,IAAI;EAC3E,MAAM,SAAS,MAAM,KAAK,EAAC,QAAQ,KAAK,SAAS,GAAE,CAAC,CACjD,KAAK,IAAI,CACT,KAAK,IAAI;EACZ,MAAM,gBAAgB,WAAW,IAAI,GAAG,CAAC,KAAK,IAAI;EAElD,IAAI,aAAa;EACjB,IAAI,UAAU;AACd,OAAK,MAAM,KAAK,WAAW;GACzB,MAAM,MAAM,QACV,OAAO,YAAY,KAAK,KAAK,GAAG,MAAM,CAAC,GAAG,EAAE,GAAG,CAAC,CAAC,EACjD,WACA,MAAA,WACD;GACD,MAAM,SAAS,MAAA,OAAa,KAAK,EAAC,UAAS,CAAC;GAC5C,MAAM,QAAQ,MAAA,UAAgB,eAAe,WAAW,OAAO;AAC/D,OAAI,OAAO,OAAA,OAAiB,MAAM,eAAe,WAAW;AAC1D;AACA;;GAEF,MAAM,UACJ,OAAO,OAAA,MACH,KAAK,QACH,OAAM,MAAM,0BAA0B,MAAM,OAAO,UACpD,GACD;AACN,OAAI,QAAQ,WAAW,GAAG;AAExB;AACA;;GAEF,MAAM,cAAc,QAAQ,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,YAAY,GAAG,IAAI,GAAG;AACxE,SAAA,GAAS,IACC;sBACM,GAAG,UAAU,CAAC,IAAI,cAAc,YAAY,OAAO;yBAChD,cAAc;0BACb,YAAY,KAAK,IAAI,CAAC;SAExC,GAAG,OAAO,OAAO,IAAI,IAAI,EACzB,UACD;AACD;;AAGF,QAAA,GAAS,QACP,cAAc,WAAW,iBAAiB,QAAQ,SAAS,YAC5D;;CAGH;CAEA,yBAAyB,EAAC,UAAU,SAAS,UAA4B;EACvE,MAAM,YAAY,cAAc,SAAS;EAEzC,MAAM,OAAO,CAAC,GADK,SAAS,OAAO,SACN,GAAG,QAAQ;EAExC,MAAM,iBAAiB,KAAK,oBAAoB,YAAY,MAAA,GAAS,GAAG,CAAC;AACzE,OAAK,MAAM,OAAO,KAChB,gBAAe,iBAAiB,WAAW,IAAI;AAIjD,QAAA,aAAmB,SAAS;AAC5B,MAAI,OACF,OAAA,oBAA0B;GAAC,OAAO;GAAW,SAAS;GAAM,GAAG;GAAO;AAExE,QAAA,GAAS,OAAO,wBAAwB,YAAY;;CAatD,cAAc,QAAuB,WAAiC;AACpE,MAAI,cAAc,MAAA,QAChB,OAAM,IAAI,MACR,oBAAoB,UAAU,kCAC5B,MAAA,QACD,IAAI,UAAU,OAAO,GACvB;AAEH,6BAA2B,MAAA,IAAU,UAAU;AAE/C,MAAI,MAAA,eAAqB;GACvB,MAAM,QAAQ,KAAK,KAAK;AACxB,SAAA,GAAS,GAAG,OAAO,WAAW;AAC9B,SAAA,GAAS,OACP,yCAAyC,KAAK,KAAK,GAAG,MAAM,MAC7D;;AAGH,MAAI,MAAA,SAAe,eACjB,OAAA,GAAS,QAAQ;EAGnB,MAAM,YAAY,KAAK,KAAK,GAAG,MAAA;AAC/B,QAAA,GAAS,QAAQ,gBAAgB,MAAA,QAAc,IAAI,UAAU,MAAM;AAEnE,SAAO;GACL;GACA,mBAAmB,MAAA;GACnB,eAAe,MAAA;GACf,kBAAkB,MAAA,sBAA4B;GAC/C;;CAGH,MAAM,IAAgB;AACpB,KAAG,OAAO,wBAAwB,MAAA,UAAgB;AAClD,QAAA,GAAS,UAAU;;;AAIvB,SAAS,qBACP,KACA,EAAC,eACqB;AACtB,KAAI,CAAC,aAAa,OAChB;AAEF,QAAO,YAAY,QAAO,QAAO,OAAO,IAAI;;AAG9C,SAAS,YAAY,KAAqB;AACxC,KAAI,eAAe,MACjB,QAAO;CAET,MAAM,wBAAQ,IAAI,OAAO;AACzB,OAAM,QAAQ;AACd,QAAO"}
1
+ {"version":3,"file":"change-processor.js","names":["#db","#changeLog","#tableMetadata","#mode","#failService","#tableSpecs","#failure","#currentTx","#fail","#processMessage","#beginTransaction","#lc","#startMs","#version","#jsonFormat","#columnMetadata","#reloadTableSpecs","#tableSpec","#upsert","#getKey","#logSetOp","#logDeleteOp","#delete","#logTruncateOp","#logResetOp","#bumpVersions","#pos","#numChangeLogEntries","#schemaChanged","#completedBackfill"],"sources":["../../../../../../zero-cache/src/services/replicator/change-processor.ts"],"sourcesContent":["import type {LogContext} from '@rocicorp/logger';\nimport {SqliteError} from '@rocicorp/zero-sqlite3';\nimport {AbortError} from '../../../../shared/src/abort-error.ts';\nimport {assert, unreachable} from '../../../../shared/src/asserts.ts';\nimport {stringify} from '../../../../shared/src/bigint-json.ts';\nimport {must} from '../../../../shared/src/must.ts';\nimport type {DownloadStatus} from '../../../../zero-events/src/status.ts';\nimport {\n createLiteIndexStatement,\n createLiteTableStatement,\n liteColumnDef,\n} from '../../db/create.ts';\nimport {\n computeZqlSpecs,\n listIndexes,\n listTables,\n type LiteTableSpecWithReplicationStatus,\n} from '../../db/lite-tables.ts';\nimport {\n mapPostgresToLite,\n mapPostgresToLiteColumn,\n mapPostgresToLiteIndex,\n} from '../../db/pg-to-lite.ts';\nimport type {StatementRunner} from '../../db/statements.ts';\nimport type {LexiVersion} from '../../types/lexi-version.ts';\nimport {\n JSON_PARSED,\n liteRow,\n type JSONFormat,\n type LiteRow,\n type LiteRowKey,\n type LiteValueType,\n} from '../../types/lite.ts';\nimport {liteTableName} from '../../types/names.ts';\nimport {id} from '../../types/sql.ts';\nimport type {\n BackfillCompleted,\n Change,\n ColumnAdd,\n ColumnDrop,\n ColumnUpdate,\n Identifier,\n IndexCreate,\n IndexDrop,\n MessageBackfill,\n MessageCommit,\n MessageDelete,\n MessageInsert,\n MessageRelation,\n MessageTruncate,\n MessageUpdate,\n TableCreate,\n TableDrop,\n TableRename,\n TableUpdateMetadata,\n} from '../change-source/protocol/current/data.ts';\nimport type {ChangeStreamData} from '../change-source/protocol/current/downstream.ts';\nimport type {ReplicatorMode} from './replicator.ts';\nimport {ChangeLog, DEL_OP, SET_OP} from './schema/change-log.ts';\nimport {ColumnMetadataStore} from './schema/column-metadata.ts';\nimport {\n ZERO_VERSION_COLUMN_NAME,\n updateReplicationWatermark,\n} from './schema/replication-state.ts';\nimport {TableMetadataTracker} from './schema/table-metadata.ts';\n\nexport type ChangeProcessorMode = ReplicatorMode | 'initial-sync';\n\nexport type CommitResult = {\n watermark: string;\n completedBackfill: DownloadStatus | undefined;\n schemaUpdated: boolean;\n changeLogUpdated: boolean;\n};\n\n/**\n * The ChangeProcessor partitions the stream of messages into transactions\n * by creating a {@link TransactionProcessor} when a transaction begins, and dispatching\n * messages to it until the commit is received.\n *\n * From https://www.postgresql.org/docs/current/protocol-logical-replication.html#PROTOCOL-LOGICAL-MESSAGES-FLOW :\n *\n * \"The logical replication protocol sends individual transactions one by one.\n * This means that all messages between a pair of Begin and Commit messages\n * belong to the same transaction.\"\n */\nexport class ChangeProcessor {\n readonly #db: StatementRunner;\n readonly #changeLog: ChangeLog;\n readonly #tableMetadata: TableMetadataTracker;\n readonly #mode: ChangeProcessorMode;\n readonly #failService: (lc: LogContext, err: unknown) => void;\n\n // The TransactionProcessor lazily loads table specs into this Map,\n // and reloads them after a schema change. It is cached here to avoid\n // reading them from the DB on every transaction.\n readonly #tableSpecs = new Map<string, LiteTableSpecWithReplicationStatus>();\n\n #currentTx: TransactionProcessor | null = null;\n\n #failure: Error | undefined;\n\n constructor(\n db: StatementRunner,\n mode: ChangeProcessorMode,\n failService: (lc: LogContext, err: unknown) => void,\n ) {\n this.#db = db;\n this.#changeLog = new ChangeLog(db.db);\n this.#tableMetadata = new TableMetadataTracker(db.db);\n this.#mode = mode;\n this.#failService = failService;\n }\n\n #fail(lc: LogContext, err: unknown) {\n if (!this.#failure) {\n let failureError = err;\n try {\n this.#currentTx?.abort(lc); // roll back any pending transaction.\n } catch (rollbackError) {\n const combinedError = new Error(\n `Message processing failed and rollback also failed: operation error = ${String(err)}; rollback error = ${String(rollbackError)}`,\n );\n combinedError.cause = err;\n failureError = combinedError;\n }\n\n this.#failure = ensureError(failureError);\n\n if (!(this.#failure instanceof AbortError)) {\n // Propagate the failure up to the service.\n lc.error?.('Message Processing failed:', this.#failure);\n this.#failService(lc, this.#failure);\n }\n }\n }\n\n abort(lc: LogContext) {\n this.#fail(lc, new AbortError());\n }\n\n /** @return If a transaction was committed. */\n processMessage(\n lc: LogContext,\n downstream: ChangeStreamData,\n ): CommitResult | null {\n const [type, message] = downstream;\n if (this.#failure) {\n lc.debug?.(`Dropping ${message.tag}`);\n return null;\n }\n try {\n const watermark =\n type === 'begin'\n ? downstream[2].commitWatermark\n : type === 'commit'\n ? downstream[2].watermark\n : undefined;\n return this.#processMessage(lc, message, watermark);\n } catch (e) {\n this.#fail(lc, e);\n }\n return null;\n }\n\n #beginTransaction(\n lc: LogContext,\n commitVersion: string,\n jsonFormat: JSONFormat,\n ): TransactionProcessor {\n const start = Date.now();\n\n // litestream can technically hold the lock for an arbitrary amount of time\n // when checkpointing a large commit. Crashing on the busy-timeout in this\n // scenario will either produce a corrupt backup or otherwise prevent\n // replication from proceeding.\n //\n // Instead, retry the lock acquisition indefinitely. If this masks\n // an unknown deadlock situation, manual intervention will be necessary.\n for (let i = 0; ; i++) {\n try {\n return new TransactionProcessor(\n lc,\n this.#db,\n this.#mode,\n this.#changeLog,\n this.#tableMetadata,\n this.#tableSpecs,\n commitVersion,\n jsonFormat,\n );\n } catch (e) {\n if (e instanceof SqliteError && e.code === 'SQLITE_BUSY') {\n lc.warn?.(\n `SQLITE_BUSY for ${Date.now() - start} ms (attempt ${i + 1}). ` +\n `This is only expected if litestream is performing a large ` +\n `checkpoint.`,\n e,\n );\n continue;\n }\n throw e;\n }\n }\n }\n\n /** @return If a transaction was committed. */\n #processMessage(\n lc: LogContext,\n msg: Change,\n watermark: string | undefined,\n ): CommitResult | null {\n if (msg.tag === 'begin') {\n if (this.#currentTx) {\n throw new Error(`Already in a transaction ${stringify(msg)}`);\n }\n this.#currentTx = this.#beginTransaction(\n lc,\n must(watermark),\n msg.json ?? JSON_PARSED,\n );\n return null;\n }\n\n // For non-begin messages, there should be a #currentTx set.\n const tx = this.#currentTx;\n if (!tx) {\n throw new Error(\n `Received message outside of transaction: ${stringify(msg)}`,\n );\n }\n\n if (msg.tag === 'commit') {\n // Undef this.#currentTx to allow the assembly of the next transaction.\n this.#currentTx = null;\n\n assert(watermark, 'watermark is required for commit messages');\n return tx.processCommit(msg, watermark);\n }\n\n if (msg.tag === 'rollback') {\n this.#currentTx?.abort(lc);\n this.#currentTx = null;\n return null;\n }\n\n switch (msg.tag) {\n case 'insert':\n tx.processInsert(msg);\n break;\n case 'update':\n tx.processUpdate(msg);\n break;\n case 'delete':\n tx.processDelete(msg);\n break;\n case 'truncate':\n tx.processTruncate(msg);\n break;\n case 'create-table':\n tx.processCreateTable(msg);\n break;\n case 'rename-table':\n tx.processRenameTable(msg);\n break;\n case 'update-table-metadata':\n tx.processTableMetadata(msg);\n break;\n case 'add-column':\n tx.processAddColumn(msg);\n break;\n case 'update-column':\n tx.processUpdateColumn(msg);\n break;\n case 'drop-column':\n tx.processDropColumn(msg);\n break;\n case 'drop-table':\n tx.processDropTable(msg);\n break;\n case 'create-index':\n tx.processCreateIndex(msg);\n break;\n case 'drop-index':\n tx.processDropIndex(msg);\n break;\n case 'backfill':\n tx.processBackfill(msg);\n break;\n case 'backfill-completed':\n tx.processBackfillCompleted(msg);\n break;\n default:\n unreachable(msg);\n }\n\n return null;\n }\n}\n\n/**\n * The {@link TransactionProcessor} handles the sequence of messages from\n * upstream, from `BEGIN` to `COMMIT` and executes the corresponding mutations\n * on the {@link postgres.TransactionSql} on the replica.\n *\n * When applying row contents to the replica, the `_0_version` column is added / updated,\n * and a corresponding entry in the `ChangeLog` is added. The version value is derived\n * from the watermark of the preceding transaction (stored as the `nextStateVersion` in the\n * `ReplicationState` table).\n *\n * Side note: For non-streaming Postgres transactions, the commitEndLsn (and thus\n * commit watermark) is available in the `begin` message, so it could theoretically\n * be used for the row version of changes within the transaction. However, the\n * commitEndLsn is not available in the streaming (in-progress) transaction\n * protocol, and may not be available for CDC streams of other upstream types.\n * Therefore, the zero replication protocol is designed to not require the commit\n * watermark when a transaction begins.\n *\n * Also of interest is the fact that all INSERT Messages are logically applied as\n * UPSERTs. See {@link processInsert} for the underlying motivation.\n */\nclass TransactionProcessor {\n readonly #lc: LogContext;\n readonly #startMs: number;\n readonly #db: StatementRunner;\n readonly #mode: ChangeProcessorMode;\n readonly #version: LexiVersion;\n readonly #changeLog: ChangeLog;\n readonly #tableMetadata: TableMetadataTracker;\n readonly #tableSpecs: Map<string, LiteTableSpecWithReplicationStatus>;\n readonly #jsonFormat: JSONFormat;\n readonly #columnMetadata: ColumnMetadataStore;\n\n #pos = 0;\n #schemaChanged = false;\n #numChangeLogEntries = 0;\n\n constructor(\n lc: LogContext,\n db: StatementRunner,\n mode: ChangeProcessorMode,\n changeLog: ChangeLog,\n tableMetadata: TableMetadataTracker,\n tableSpecs: Map<string, LiteTableSpecWithReplicationStatus>,\n commitVersion: LexiVersion,\n jsonFormat: JSONFormat,\n ) {\n this.#startMs = Date.now();\n this.#mode = mode;\n this.#jsonFormat = jsonFormat;\n\n switch (mode) {\n case 'serving':\n // Although the Replicator / Incremental Syncer is the only writer of the replica,\n // a `BEGIN CONCURRENT` transaction is used to allow View Syncers to simulate\n // (i.e. and `ROLLBACK`) changes on historic snapshots of the database for the\n // purpose of IVM).\n //\n // This TransactionProcessor is the only logic that will actually\n // `COMMIT` any transactions to the replica.\n db.beginConcurrent();\n break;\n case 'backup':\n // For the backup-replicator (i.e. replication-manager), there are no View Syncers\n // and thus BEGIN CONCURRENT is not necessary. In fact, BEGIN CONCURRENT can cause\n // deadlocks with forced wal-checkpoints (which `litestream replicate` performs),\n // so it is important to use vanilla transactions in this configuration.\n db.beginImmediate();\n break;\n case 'initial-sync':\n // When the ChangeProcessor is used for initial-sync, the calling code\n // handles the transaction boundaries.\n break;\n default:\n unreachable();\n }\n this.#db = db;\n this.#version = commitVersion;\n this.#lc = lc.withContext('version', commitVersion);\n this.#changeLog = changeLog;\n this.#tableMetadata = tableMetadata;\n this.#tableSpecs = tableSpecs;\n // The column_metadata table is guaranteed to exist since the\n // replica-schema.ts migration to v8.\n this.#columnMetadata = must(ColumnMetadataStore.getInstance(db.db));\n\n if (this.#tableSpecs.size === 0) {\n this.#reloadTableSpecs();\n }\n }\n\n #reloadTableSpecs() {\n this.#tableSpecs.clear();\n // zqlSpecs include the primary key derived from unique indexes\n const zqlSpecs = computeZqlSpecs(this.#lc, this.#db.db, {\n includeBackfillingColumns: true,\n });\n for (let spec of listTables(this.#db.db)) {\n if (!spec.primaryKey) {\n spec = {\n ...spec,\n primaryKey: [\n ...(zqlSpecs.get(spec.name)?.tableSpec.primaryKey ?? []),\n ],\n };\n }\n this.#tableSpecs.set(spec.name, spec);\n }\n }\n\n #tableSpec(name: string) {\n return must(this.#tableSpecs.get(name), `Unknown table ${name}`);\n }\n\n #getKey(\n {row, numCols}: {row: LiteRow; numCols: number},\n {relation}: {relation: MessageRelation},\n ): LiteRowKey {\n const keyColumns =\n relation.rowKey.type !== 'full'\n ? relation.rowKey.columns // already a suitable key\n : this.#tableSpec(liteTableName(relation)).primaryKey;\n if (!keyColumns?.length) {\n throw new Error(\n `Cannot replicate table \"${relation.name}\" without a PRIMARY KEY or UNIQUE INDEX`,\n );\n }\n // For the common case (replica identity default), the row is already the\n // key for deletes and updates, in which case a new object can be avoided.\n if (numCols === keyColumns.length) {\n return row;\n }\n const key: Record<string, LiteValueType> = {};\n for (const col of keyColumns) {\n key[col] = row[col];\n }\n return key;\n }\n\n processInsert(insert: MessageInsert) {\n const table = liteTableName(insert.relation);\n const tableSpec = this.#tableSpec(table);\n const newRow = liteRow(insert.new, tableSpec, this.#jsonFormat);\n\n this.#upsert(table, {\n ...newRow.row,\n [ZERO_VERSION_COLUMN_NAME]: this.#version,\n });\n\n if (insert.relation.rowKey.columns.length === 0) {\n // INSERTs can be replicated for rows without a PRIMARY KEY or a\n // UNIQUE INDEX. These are written to the replica but not recorded\n // in the changeLog, because these rows cannot participate in IVM.\n //\n // (Once the table schema has been corrected to include a key, the\n // associated schema change will reset pipelines and data can be\n // loaded via hydration.)\n return;\n }\n const key = this.#getKey(newRow, insert);\n this.#logSetOp(table, key, getBackfilledColumns(newRow.row, tableSpec));\n }\n\n #upsert(table: string, row: LiteRow) {\n const columns = Object.keys(row).map(c => id(c));\n this.#db.run(\n `\n INSERT OR REPLACE INTO ${id(table)} (${columns.join(',')})\n VALUES (${Array.from({length: columns.length}).fill('?').join(',')})\n `,\n Object.values(row),\n );\n }\n\n // Updates by default are applied as UPDATE commands to support partial\n // row specifications from the change source. In particular, this is needed\n // to handle updates for which unchanged TOASTed values are not sent:\n //\n // https://www.postgresql.org/docs/current/protocol-logicalrep-message-formats.html#PROTOCOL-LOGICALREP-MESSAGE-FORMATS-TUPLEDATA\n //\n // However, in certain cases an UPDATE may be received for a row that\n // was not initially synced, such as when, an existing table is added\n // to the app's publication.\n //\n // In order to facilitate \"resumptive\" replication, the logic falls back to\n // an INSERT if the update did not change any rows.\n processUpdate(update: MessageUpdate) {\n const table = liteTableName(update.relation);\n const tableSpec = this.#tableSpec(table);\n const newRow = liteRow(update.new, tableSpec, this.#jsonFormat);\n const row = {...newRow.row, [ZERO_VERSION_COLUMN_NAME]: this.#version};\n\n // update.key is set with the old values if the key has changed.\n const oldKey = update.key\n ? this.#getKey(\n liteRow(update.key, this.#tableSpec(table), this.#jsonFormat),\n update,\n )\n : null;\n const newKey = this.#getKey(newRow, update);\n\n if (oldKey) {\n this.#logDeleteOp(table, oldKey, tableSpec.backfilling);\n }\n this.#logSetOp(table, newKey, getBackfilledColumns(newRow.row, tableSpec));\n\n const currKey = oldKey ?? newKey;\n const conds = Object.keys(currKey).map(col => `${id(col)}=?`);\n const setExprs = Object.keys(row).map(col => `${id(col)}=?`);\n\n const {changes} = this.#db.run(\n `\n UPDATE ${id(table)}\n SET ${setExprs.join(',')}\n WHERE ${conds.join(' AND ')}\n `,\n [...Object.values(row), ...Object.values(currKey)],\n );\n\n // If the UPDATE did not affect any rows, perform an UPSERT of the\n // new row for resumptive replication.\n if (changes === 0) {\n this.#upsert(table, row);\n }\n }\n\n processDelete(del: MessageDelete) {\n const table = liteTableName(del.relation);\n const tableSpec = this.#tableSpec(table);\n const rowKey = this.#getKey(\n liteRow(del.key, tableSpec, this.#jsonFormat),\n del,\n );\n\n this.#delete(table, rowKey);\n this.#logDeleteOp(table, rowKey, tableSpec.backfilling);\n }\n\n #delete(table: string, rowKey: LiteRowKey) {\n const conds = Object.keys(rowKey).map(col => `${id(col)}=?`);\n this.#db.run(\n `DELETE FROM ${id(table)} WHERE ${conds.join(' AND ')}`,\n Object.values(rowKey),\n );\n }\n\n processTruncate(truncate: MessageTruncate) {\n for (const relation of truncate.relations) {\n const table = liteTableName(relation);\n // Update replica data.\n this.#db.run(`DELETE FROM ${id(table)}`);\n\n // Update change log.\n this.#logTruncateOp(table);\n }\n }\n\n processCreateTable(create: TableCreate) {\n if (create.metadata) {\n this.#tableMetadata.setUpstreamMetadata(create.spec, create.metadata);\n }\n const table = mapPostgresToLite(create.spec);\n this.#db.db.exec(createLiteTableStatement(table));\n\n // Write to metadata table\n for (const [colName, colSpec] of Object.entries(create.spec.columns)) {\n this.#columnMetadata.insert(\n table.name,\n colName,\n colSpec,\n create.backfill?.[colName],\n );\n }\n\n if (\n Object.keys(create.backfill ?? {}).length ===\n Object.keys(create.spec.columns).length\n ) {\n this.#reloadTableSpecs();\n } else {\n // Make the table visible immediately unless all of the columns are\n // being backfilled. In the backfill case, the version bump will happen\n // with the backfill is complete.\n this.#logResetOp(table.name);\n }\n this.#lc.info?.(create.tag, table.name);\n }\n\n processTableMetadata(msg: TableUpdateMetadata) {\n this.#tableMetadata.setUpstreamMetadata(msg.table, msg.new);\n }\n\n processRenameTable(rename: TableRename) {\n this.#tableMetadata.rename(rename.old, rename.new);\n\n const oldName = liteTableName(rename.old);\n const newName = liteTableName(rename.new);\n this.#db.db.exec(`ALTER TABLE ${id(oldName)} RENAME TO ${id(newName)}`);\n\n // Rename in metadata table\n this.#columnMetadata.renameTable(oldName, newName);\n\n this.#bumpVersions(rename.new);\n this.#logResetOp(oldName);\n this.#lc.info?.(rename.tag, oldName, newName);\n }\n\n processAddColumn(msg: ColumnAdd) {\n if (msg.tableMetadata) {\n this.#tableMetadata.setUpstreamMetadata(msg.table, msg.tableMetadata);\n }\n const table = liteTableName(msg.table);\n const {name} = msg.column;\n const spec = mapPostgresToLiteColumn(table, msg.column);\n this.#db.db.exec(\n `ALTER TABLE ${id(table)} ADD ${id(name)} ${liteColumnDef(spec)}`,\n );\n\n // Write to metadata table\n this.#columnMetadata.insert(table, name, msg.column.spec, msg.backfill);\n\n if (msg.backfill) {\n this.#reloadTableSpecs();\n } else {\n // Make the new column visible immediately if it's not being backfilled.\n // Otherwise, the version bump will happen with the backfill is complete.\n this.#bumpVersions(msg.table);\n }\n this.#lc.info?.(msg.tag, table, msg.column);\n }\n\n processUpdateColumn(msg: ColumnUpdate) {\n const table = liteTableName(msg.table);\n let oldName = msg.old.name;\n const newName = msg.new.name;\n\n // update-column can ignore defaults because it does not change the values\n // in existing rows.\n //\n // https://www.postgresql.org/docs/current/sql-altertable.html#SQL-ALTERTABLE-DESC-SET-DROP-DEFAULT\n //\n // \"The new default value will only apply in subsequent INSERT or UPDATE\n // commands; it does not cause rows already in the table to change.\"\n //\n // This allows support for _changing_ column defaults to any expression,\n // since it does not affect what the replica needs to do.\n const oldSpec = mapPostgresToLiteColumn(table, msg.old, 'ignore-default');\n const newSpec = mapPostgresToLiteColumn(table, msg.new, 'ignore-default');\n\n // The only updates that are relevant are the column name and the data type.\n if (oldName === newName && oldSpec.dataType === newSpec.dataType) {\n this.#lc.info?.(msg.tag, 'no thing to update', oldSpec, newSpec);\n return;\n }\n // If the data type changes, we have to make a new column with the new data type\n // and copy the values over.\n if (oldSpec.dataType !== newSpec.dataType) {\n // Remember (and drop) the indexes that reference the column.\n const indexes = listIndexes(this.#db.db).filter(\n idx => idx.tableName === table && oldName in idx.columns,\n );\n const stmts = indexes.map(idx => `DROP INDEX IF EXISTS ${id(idx.name)};`);\n const tmpName = `tmp.${newName}`;\n stmts.push(`\n ALTER TABLE ${id(table)} ADD ${id(tmpName)} ${liteColumnDef(newSpec)};\n UPDATE ${id(table)} SET ${id(tmpName)} = ${id(oldName)};\n ALTER TABLE ${id(table)} DROP ${id(oldName)};\n `);\n for (const idx of indexes) {\n // Re-create the indexes to reference the new column.\n idx.columns[tmpName] = idx.columns[oldName];\n delete idx.columns[oldName];\n stmts.push(createLiteIndexStatement(idx));\n }\n this.#db.db.exec(stmts.join(''));\n oldName = tmpName;\n }\n if (oldName !== newName) {\n this.#db.db.exec(\n `ALTER TABLE ${id(table)} RENAME ${id(oldName)} TO ${id(newName)}`,\n );\n }\n\n // Update metadata table\n this.#columnMetadata.update(\n table,\n msg.old.name,\n msg.new.name,\n msg.new.spec,\n );\n\n this.#bumpVersions(msg.table);\n this.#lc.info?.(msg.tag, table, msg.new);\n }\n\n processDropColumn(msg: ColumnDrop) {\n const table = liteTableName(msg.table);\n const {column} = msg;\n this.#db.db.exec(`ALTER TABLE ${id(table)} DROP ${id(column)}`);\n\n // Delete from metadata table\n this.#columnMetadata.deleteColumn(table, column);\n\n this.#bumpVersions(msg.table);\n this.#lc.info?.(msg.tag, table, column);\n }\n\n processDropTable(drop: TableDrop) {\n this.#tableMetadata.drop(drop.id);\n\n const name = liteTableName(drop.id);\n this.#db.db.exec(`DROP TABLE IF EXISTS ${id(name)}`);\n\n // Delete from metadata table\n this.#columnMetadata.deleteTable(name);\n\n this.#logResetOp(name);\n this.#lc.info?.(drop.tag, name);\n }\n\n processCreateIndex(create: IndexCreate) {\n const index = mapPostgresToLiteIndex(create.spec);\n this.#db.db.exec(createLiteIndexStatement(index));\n\n // indexes affect tables visibility (e.g. sync-ability is gated on\n // having a unique index), so reset pipelines to refresh table schemas.\n // However, the reset is not necessary if the index is for a table\n // that is not yet visible due to backfilling.\n const tableSpec = must(this.#tableSpecs.get(index.tableName));\n if (\n (tableSpec.backfilling ?? []).length ===\n Object.entries(tableSpec.columns).length - 1 // don't count _0_version\n ) {\n this.#reloadTableSpecs();\n } else {\n this.#logResetOp(index.tableName);\n }\n this.#lc.info?.(create.tag, index.name);\n }\n\n processDropIndex(drop: IndexDrop) {\n const name = liteTableName(drop.id);\n this.#db.db.exec(`DROP INDEX IF EXISTS ${id(name)}`);\n this.#lc.info?.(drop.tag, name);\n }\n\n #bumpVersions(table: Identifier) {\n this.#tableMetadata.setMinRowVersion(table, this.#version);\n this.#logResetOp(liteTableName(table));\n }\n\n /**\n * @param backfilledColumns `backfilling` columns for which values were set\n */\n #logSetOp(\n table: string,\n key: LiteRowKey,\n backfilledColumns: string[] | undefined,\n ) {\n // The \"serving\" replicator always writes to the change-log (for IVM).\n // The \"backup\" replicator only needs to write to the change log\n // when writing columns that are being backfilled.\n if (this.#mode === 'serving' || backfilledColumns !== undefined) {\n this.#changeLog.logSetOp(\n this.#version,\n this.#pos++,\n table,\n key,\n backfilledColumns,\n );\n this.#numChangeLogEntries++;\n }\n }\n\n #logDeleteOp(table: string, key: LiteRowKey, backfilling?: string[]) {\n // The \"serving\" replicator always writes to the change-log (for IVM).\n // The \"backup\" replicator only needs to write to the change log\n // when writing columns that are being backfilled.\n if (this.#mode === 'serving' || backfilling?.length) {\n this.#changeLog.logDeleteOp(this.#version, this.#pos++, table, key);\n this.#numChangeLogEntries++;\n }\n }\n\n #logTruncateOp(table: string) {\n if (this.#mode === 'serving') {\n this.#changeLog.logTruncateOp(this.#version, table);\n this.#numChangeLogEntries++;\n }\n }\n\n #logResetOp(table: string) {\n this.#schemaChanged = true;\n if (this.#mode === 'serving') {\n this.#changeLog.logResetOp(this.#version, table);\n this.#numChangeLogEntries++;\n }\n this.#reloadTableSpecs();\n }\n\n processBackfill({relation, watermark, columns, rowValues}: MessageBackfill) {\n const tableName = liteTableName(relation);\n const tableSpec = must(this.#tableSpecs.get(tableName));\n const rowKeyCols = relation.rowKey.columns;\n const cols = [...rowKeyCols, ...columns];\n\n // Common parts of the INSERT sql statement.\n const insertColsStr = [...cols, ZERO_VERSION_COLUMN_NAME].map(id).join(',');\n const qMarks = Array.from({length: cols.length + 1})\n .fill('?')\n .join(',');\n const rowKeyColsStr = rowKeyCols.map(id).join(',');\n\n let backfilled = 0;\n let skipped = 0;\n for (const v of rowValues) {\n const row = liteRow(\n Object.fromEntries(cols.map((c, i) => [c, v[i]])),\n tableSpec,\n this.#jsonFormat,\n );\n const rowKey = this.#getKey(row, {relation});\n const rowOp = this.#changeLog.getLatestRowOp(tableName, rowKey);\n if (rowOp?.op === DEL_OP && rowOp.stateVersion > watermark) {\n skipped++;\n continue; // the row was deleted after the backfill snapshot\n }\n const updates =\n rowOp?.op === SET_OP\n ? cols.filter(\n c => (rowOp.backfillingColumnVersions[c] ?? '') <= watermark,\n )\n : cols;\n if (updates.length === 0) {\n // row already has newer values for all backfilling columns.\n skipped++;\n continue;\n }\n const updateStmts = updates.map(col => `${id(col)}=excluded.${id(col)}`);\n this.#db.run(\n /*sql*/ `\n INSERT INTO ${id(tableName)} (${insertColsStr}) VALUES (${qMarks})\n ON CONFLICT (${rowKeyColsStr})\n DO UPDATE SET ${updateStmts.join(',')};\n `,\n ...Object.values(row.row),\n watermark, // the _0_version for new rows (i.e. table backfill)\n );\n backfilled++;\n }\n\n this.#lc.debug?.(\n `backfilled ${backfilled} rows (skipped ${skipped}) into ${tableName}`,\n );\n }\n\n #completedBackfill: DownloadStatus | undefined;\n\n processBackfillCompleted({relation, columns, status}: BackfillCompleted) {\n const tableName = liteTableName(relation);\n const rowKeyCols = relation.rowKey.columns;\n const cols = [...rowKeyCols, ...columns];\n\n const columnMetadata = must(ColumnMetadataStore.getInstance(this.#db.db));\n for (const col of cols) {\n columnMetadata.clearBackfilling(tableName, col);\n }\n // Given that new columns are being exposed for every row in the table, bump the\n // row version for all rows.\n this.#bumpVersions(relation);\n if (status) {\n this.#completedBackfill = {table: tableName, columns: cols, ...status};\n }\n this.#lc.info?.(`finished backfilling ${tableName}`);\n\n // Note that there is no need to clear the backfillingColumnVersions values\n // in the changeLog. It could theoretically be done for clarity but:\n // (1) it could be non-trivial in terms of latency introduced and\n // (2) the data must be preserved if _other_ columns are in the process\n // of being backfilled\n //\n // Thus, for speed and simplicity, the values are left as is. (Note that\n // subsequent replicated changes to those rows will clear the values if\n // no backfills are in progress).\n }\n\n processCommit(commit: MessageCommit, watermark: string): CommitResult {\n if (watermark !== this.#version) {\n throw new Error(\n `'commit' version ${watermark} does not match 'begin' version ${\n this.#version\n }: ${stringify(commit)}`,\n );\n }\n updateReplicationWatermark(this.#db, watermark);\n\n if (this.#schemaChanged) {\n const start = Date.now();\n this.#db.db.pragma('optimize');\n this.#lc.info?.(\n `PRAGMA optimized after schema change (${Date.now() - start} ms)`,\n );\n }\n\n if (this.#mode !== 'initial-sync') {\n this.#db.commit();\n }\n\n const elapsedMs = Date.now() - this.#startMs;\n this.#lc.debug?.(`Committed tx@${this.#version} (${elapsedMs} ms)`);\n\n return {\n watermark,\n completedBackfill: this.#completedBackfill,\n schemaUpdated: this.#schemaChanged,\n changeLogUpdated: this.#numChangeLogEntries > 0,\n };\n }\n\n abort(lc: LogContext) {\n lc.info?.(`aborting transaction ${this.#version}`);\n this.#db.rollback();\n }\n}\n\nfunction getBackfilledColumns(\n row: LiteRow,\n {backfilling}: LiteTableSpecWithReplicationStatus,\n): string[] | undefined {\n if (!backfilling?.length) {\n return undefined; // common case\n }\n return backfilling.filter(col => col in row);\n}\n\nfunction ensureError(err: unknown): Error {\n if (err instanceof Error) {\n return err;\n }\n const error = new Error();\n error.cause = err;\n return error;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsFA,IAAa,kBAAb,MAA6B;CAC3B;CACA;CACA;CACA;CACA;CAKA,8BAAuB,IAAI,KAAiD;CAE5E,aAA0C;CAE1C;CAEA,YACE,IACA,MACA,aACA;AACA,QAAA,KAAW;AACX,QAAA,YAAkB,IAAI,UAAU,GAAG,GAAG;AACtC,QAAA,gBAAsB,IAAI,qBAAqB,GAAG,GAAG;AACrD,QAAA,OAAa;AACb,QAAA,cAAoB;;CAGtB,MAAM,IAAgB,KAAc;AAClC,MAAI,CAAC,MAAA,SAAe;GAClB,IAAI,eAAe;AACnB,OAAI;AACF,UAAA,WAAiB,MAAM,GAAG;YACnB,eAAe;IACtB,MAAM,gCAAgB,IAAI,MACxB,yEAAyE,OAAO,IAAI,CAAC,qBAAqB,OAAO,cAAc,GAChI;AACD,kBAAc,QAAQ;AACtB,mBAAe;;AAGjB,SAAA,UAAgB,YAAY,aAAa;AAEzC,OAAI,EAAE,MAAA,mBAAyB,aAAa;AAE1C,OAAG,QAAQ,8BAA8B,MAAA,QAAc;AACvD,UAAA,YAAkB,IAAI,MAAA,QAAc;;;;CAK1C,MAAM,IAAgB;AACpB,QAAA,KAAW,IAAI,IAAI,YAAY,CAAC;;;CAIlC,eACE,IACA,YACqB;EACrB,MAAM,CAAC,MAAM,WAAW;AACxB,MAAI,MAAA,SAAe;AACjB,MAAG,QAAQ,YAAY,QAAQ,MAAM;AACrC,UAAO;;AAET,MAAI;GACF,MAAM,YACJ,SAAS,UACL,WAAW,GAAG,kBACd,SAAS,WACP,WAAW,GAAG,YACd,KAAA;AACR,UAAO,MAAA,eAAqB,IAAI,SAAS,UAAU;WAC5C,GAAG;AACV,SAAA,KAAW,IAAI,EAAE;;AAEnB,SAAO;;CAGT,kBACE,IACA,eACA,YACsB;EACtB,MAAM,QAAQ,KAAK,KAAK;AASxB,OAAK,IAAI,IAAI,IAAK,IAChB,KAAI;AACF,UAAO,IAAI,qBACT,IACA,MAAA,IACA,MAAA,MACA,MAAA,WACA,MAAA,eACA,MAAA,YACA,eACA,WACD;WACM,GAAG;AACV,OAAI,aAAa,eAAe,EAAE,SAAS,eAAe;AACxD,OAAG,OACD,mBAAmB,KAAK,KAAK,GAAG,MAAM,eAAe,IAAI,EAAE,2EAG3D,EACD;AACD;;AAEF,SAAM;;;;CAMZ,gBACE,IACA,KACA,WACqB;AACrB,MAAI,IAAI,QAAQ,SAAS;AACvB,OAAI,MAAA,UACF,OAAM,IAAI,MAAM,4BAA4B,UAAU,IAAI,GAAG;AAE/D,SAAA,YAAkB,MAAA,iBAChB,IACA,KAAK,UAAU,EACf,IAAI,QAAA,IACL;AACD,UAAO;;EAIT,MAAM,KAAK,MAAA;AACX,MAAI,CAAC,GACH,OAAM,IAAI,MACR,4CAA4C,UAAU,IAAI,GAC3D;AAGH,MAAI,IAAI,QAAQ,UAAU;AAExB,SAAA,YAAkB;AAElB,UAAO,WAAW,4CAA4C;AAC9D,UAAO,GAAG,cAAc,KAAK,UAAU;;AAGzC,MAAI,IAAI,QAAQ,YAAY;AAC1B,SAAA,WAAiB,MAAM,GAAG;AAC1B,SAAA,YAAkB;AAClB,UAAO;;AAGT,UAAQ,IAAI,KAAZ;GACE,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,cAAc,IAAI;AACrB;GACF,KAAK;AACH,OAAG,gBAAgB,IAAI;AACvB;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,qBAAqB,IAAI;AAC5B;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,oBAAoB,IAAI;AAC3B;GACF,KAAK;AACH,OAAG,kBAAkB,IAAI;AACzB;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,mBAAmB,IAAI;AAC1B;GACF,KAAK;AACH,OAAG,iBAAiB,IAAI;AACxB;GACF,KAAK;AACH,OAAG,gBAAgB,IAAI;AACvB;GACF,KAAK;AACH,OAAG,yBAAyB,IAAI;AAChC;GACF,QACE,aAAY,IAAI;;AAGpB,SAAO;;;;;;;;;;;;;;;;;;;;;;;;AAyBX,IAAM,uBAAN,MAA2B;CACzB;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CAEA,OAAO;CACP,iBAAiB;CACjB,uBAAuB;CAEvB,YACE,IACA,IACA,MACA,WACA,eACA,YACA,eACA,YACA;AACA,QAAA,UAAgB,KAAK,KAAK;AAC1B,QAAA,OAAa;AACb,QAAA,aAAmB;AAEnB,UAAQ,MAAR;GACE,KAAK;AAQH,OAAG,iBAAiB;AACpB;GACF,KAAK;AAKH,OAAG,gBAAgB;AACnB;GACF,KAAK,eAGH;GACF,QACE,cAAa;;AAEjB,QAAA,KAAW;AACX,QAAA,UAAgB;AAChB,QAAA,KAAW,GAAG,YAAY,WAAW,cAAc;AACnD,QAAA,YAAkB;AAClB,QAAA,gBAAsB;AACtB,QAAA,aAAmB;AAGnB,QAAA,iBAAuB,KAAK,oBAAoB,YAAY,GAAG,GAAG,CAAC;AAEnE,MAAI,MAAA,WAAiB,SAAS,EAC5B,OAAA,kBAAwB;;CAI5B,oBAAoB;AAClB,QAAA,WAAiB,OAAO;EAExB,MAAM,WAAW,gBAAgB,MAAA,IAAU,MAAA,GAAS,IAAI,EACtD,2BAA2B,MAC5B,CAAC;AACF,OAAK,IAAI,QAAQ,WAAW,MAAA,GAAS,GAAG,EAAE;AACxC,OAAI,CAAC,KAAK,WACR,QAAO;IACL,GAAG;IACH,YAAY,CACV,GAAI,SAAS,IAAI,KAAK,KAAK,EAAE,UAAU,cAAc,EAAE,CACxD;IACF;AAEH,SAAA,WAAiB,IAAI,KAAK,MAAM,KAAK;;;CAIzC,WAAW,MAAc;AACvB,SAAO,KAAK,MAAA,WAAiB,IAAI,KAAK,EAAE,iBAAiB,OAAO;;CAGlE,QACE,EAAC,KAAK,WACN,EAAC,YACW;EACZ,MAAM,aACJ,SAAS,OAAO,SAAS,SACrB,SAAS,OAAO,UAChB,MAAA,UAAgB,cAAc,SAAS,CAAC,CAAC;AAC/C,MAAI,CAAC,YAAY,OACf,OAAM,IAAI,MACR,2BAA2B,SAAS,KAAK,yCAC1C;AAIH,MAAI,YAAY,WAAW,OACzB,QAAO;EAET,MAAM,MAAqC,EAAE;AAC7C,OAAK,MAAM,OAAO,WAChB,KAAI,OAAO,IAAI;AAEjB,SAAO;;CAGT,cAAc,QAAuB;EACnC,MAAM,QAAQ,cAAc,OAAO,SAAS;EAC5C,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,QAAQ,OAAO,KAAK,WAAW,MAAA,WAAiB;AAE/D,QAAA,OAAa,OAAO;GAClB,GAAG,OAAO;IACT,2BAA2B,MAAA;GAC7B,CAAC;AAEF,MAAI,OAAO,SAAS,OAAO,QAAQ,WAAW,EAQ5C;EAEF,MAAM,MAAM,MAAA,OAAa,QAAQ,OAAO;AACxC,QAAA,SAAe,OAAO,KAAK,qBAAqB,OAAO,KAAK,UAAU,CAAC;;CAGzE,QAAQ,OAAe,KAAc;EACnC,MAAM,UAAU,OAAO,KAAK,IAAI,CAAC,KAAI,MAAK,GAAG,EAAE,CAAC;AAChD,QAAA,GAAS,IACP;+BACyB,GAAG,MAAM,CAAC,IAAI,QAAQ,KAAK,IAAI,CAAC;kBAC7C,MAAM,KAAK,EAAC,QAAQ,QAAQ,QAAO,CAAC,CAAC,KAAK,IAAI,CAAC,KAAK,IAAI,CAAC;SAErE,OAAO,OAAO,IAAI,CACnB;;CAeH,cAAc,QAAuB;EACnC,MAAM,QAAQ,cAAc,OAAO,SAAS;EAC5C,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,QAAQ,OAAO,KAAK,WAAW,MAAA,WAAiB;EAC/D,MAAM,MAAM;GAAC,GAAG,OAAO;IAAM,2BAA2B,MAAA;GAAc;EAGtE,MAAM,SAAS,OAAO,MAClB,MAAA,OACE,QAAQ,OAAO,KAAK,MAAA,UAAgB,MAAM,EAAE,MAAA,WAAiB,EAC7D,OACD,GACD;EACJ,MAAM,SAAS,MAAA,OAAa,QAAQ,OAAO;AAE3C,MAAI,OACF,OAAA,YAAkB,OAAO,QAAQ,UAAU,YAAY;AAEzD,QAAA,SAAe,OAAO,QAAQ,qBAAqB,OAAO,KAAK,UAAU,CAAC;EAE1E,MAAM,UAAU,UAAU;EAC1B,MAAM,QAAQ,OAAO,KAAK,QAAQ,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;EAC7D,MAAM,WAAW,OAAO,KAAK,IAAI,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;EAE5D,MAAM,EAAC,YAAW,MAAA,GAAS,IACzB;eACS,GAAG,MAAM,CAAC;cACX,SAAS,KAAK,IAAI,CAAC;gBACjB,MAAM,KAAK,QAAQ,CAAC;SAE9B,CAAC,GAAG,OAAO,OAAO,IAAI,EAAE,GAAG,OAAO,OAAO,QAAQ,CAAC,CACnD;AAID,MAAI,YAAY,EACd,OAAA,OAAa,OAAO,IAAI;;CAI5B,cAAc,KAAoB;EAChC,MAAM,QAAQ,cAAc,IAAI,SAAS;EACzC,MAAM,YAAY,MAAA,UAAgB,MAAM;EACxC,MAAM,SAAS,MAAA,OACb,QAAQ,IAAI,KAAK,WAAW,MAAA,WAAiB,EAC7C,IACD;AAED,QAAA,OAAa,OAAO,OAAO;AAC3B,QAAA,YAAkB,OAAO,QAAQ,UAAU,YAAY;;CAGzD,QAAQ,OAAe,QAAoB;EACzC,MAAM,QAAQ,OAAO,KAAK,OAAO,CAAC,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,IAAI;AAC5D,QAAA,GAAS,IACP,eAAe,GAAG,MAAM,CAAC,SAAS,MAAM,KAAK,QAAQ,IACrD,OAAO,OAAO,OAAO,CACtB;;CAGH,gBAAgB,UAA2B;AACzC,OAAK,MAAM,YAAY,SAAS,WAAW;GACzC,MAAM,QAAQ,cAAc,SAAS;AAErC,SAAA,GAAS,IAAI,eAAe,GAAG,MAAM,GAAG;AAGxC,SAAA,cAAoB,MAAM;;;CAI9B,mBAAmB,QAAqB;AACtC,MAAI,OAAO,SACT,OAAA,cAAoB,oBAAoB,OAAO,MAAM,OAAO,SAAS;EAEvE,MAAM,QAAQ,kBAAkB,OAAO,KAAK;AAC5C,QAAA,GAAS,GAAG,KAAK,yBAAyB,MAAM,CAAC;AAGjD,OAAK,MAAM,CAAC,SAAS,YAAY,OAAO,QAAQ,OAAO,KAAK,QAAQ,CAClE,OAAA,eAAqB,OACnB,MAAM,MACN,SACA,SACA,OAAO,WAAW,SACnB;AAGH,MACE,OAAO,KAAK,OAAO,YAAY,EAAE,CAAC,CAAC,WACnC,OAAO,KAAK,OAAO,KAAK,QAAQ,CAAC,OAEjC,OAAA,kBAAwB;MAKxB,OAAA,WAAiB,MAAM,KAAK;AAE9B,QAAA,GAAS,OAAO,OAAO,KAAK,MAAM,KAAK;;CAGzC,qBAAqB,KAA0B;AAC7C,QAAA,cAAoB,oBAAoB,IAAI,OAAO,IAAI,IAAI;;CAG7D,mBAAmB,QAAqB;AACtC,QAAA,cAAoB,OAAO,OAAO,KAAK,OAAO,IAAI;EAElD,MAAM,UAAU,cAAc,OAAO,IAAI;EACzC,MAAM,UAAU,cAAc,OAAO,IAAI;AACzC,QAAA,GAAS,GAAG,KAAK,eAAe,GAAG,QAAQ,CAAC,aAAa,GAAG,QAAQ,GAAG;AAGvE,QAAA,eAAqB,YAAY,SAAS,QAAQ;AAElD,QAAA,aAAmB,OAAO,IAAI;AAC9B,QAAA,WAAiB,QAAQ;AACzB,QAAA,GAAS,OAAO,OAAO,KAAK,SAAS,QAAQ;;CAG/C,iBAAiB,KAAgB;AAC/B,MAAI,IAAI,cACN,OAAA,cAAoB,oBAAoB,IAAI,OAAO,IAAI,cAAc;EAEvE,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,MAAM,EAAC,SAAQ,IAAI;EACnB,MAAM,OAAO,wBAAwB,OAAO,IAAI,OAAO;AACvD,QAAA,GAAS,GAAG,KACV,eAAe,GAAG,MAAM,CAAC,OAAO,GAAG,KAAK,CAAC,GAAG,cAAc,KAAK,GAChE;AAGD,QAAA,eAAqB,OAAO,OAAO,MAAM,IAAI,OAAO,MAAM,IAAI,SAAS;AAEvE,MAAI,IAAI,SACN,OAAA,kBAAwB;MAIxB,OAAA,aAAmB,IAAI,MAAM;AAE/B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,IAAI,OAAO;;CAG7C,oBAAoB,KAAmB;EACrC,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,IAAI,UAAU,IAAI,IAAI;EACtB,MAAM,UAAU,IAAI,IAAI;EAYxB,MAAM,UAAU,wBAAwB,OAAO,IAAI,KAAK,iBAAiB;EACzE,MAAM,UAAU,wBAAwB,OAAO,IAAI,KAAK,iBAAiB;AAGzE,MAAI,YAAY,WAAW,QAAQ,aAAa,QAAQ,UAAU;AAChE,SAAA,GAAS,OAAO,IAAI,KAAK,sBAAsB,SAAS,QAAQ;AAChE;;AAIF,MAAI,QAAQ,aAAa,QAAQ,UAAU;GAEzC,MAAM,UAAU,YAAY,MAAA,GAAS,GAAG,CAAC,QACvC,QAAO,IAAI,cAAc,SAAS,WAAW,IAAI,QAClD;GACD,MAAM,QAAQ,QAAQ,KAAI,QAAO,wBAAwB,GAAG,IAAI,KAAK,CAAC,GAAG;GACzE,MAAM,UAAU,OAAO;AACvB,SAAM,KAAK;sBACK,GAAG,MAAM,CAAC,OAAO,GAAG,QAAQ,CAAC,GAAG,cAAc,QAAQ,CAAC;iBAC5D,GAAG,MAAM,CAAC,OAAO,GAAG,QAAQ,CAAC,KAAK,GAAG,QAAQ,CAAC;sBACzC,GAAG,MAAM,CAAC,QAAQ,GAAG,QAAQ,CAAC;UAC1C;AACJ,QAAK,MAAM,OAAO,SAAS;AAEzB,QAAI,QAAQ,WAAW,IAAI,QAAQ;AACnC,WAAO,IAAI,QAAQ;AACnB,UAAM,KAAK,yBAAyB,IAAI,CAAC;;AAE3C,SAAA,GAAS,GAAG,KAAK,MAAM,KAAK,GAAG,CAAC;AAChC,aAAU;;AAEZ,MAAI,YAAY,QACd,OAAA,GAAS,GAAG,KACV,eAAe,GAAG,MAAM,CAAC,UAAU,GAAG,QAAQ,CAAC,MAAM,GAAG,QAAQ,GACjE;AAIH,QAAA,eAAqB,OACnB,OACA,IAAI,IAAI,MACR,IAAI,IAAI,MACR,IAAI,IAAI,KACT;AAED,QAAA,aAAmB,IAAI,MAAM;AAC7B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,IAAI,IAAI;;CAG1C,kBAAkB,KAAiB;EACjC,MAAM,QAAQ,cAAc,IAAI,MAAM;EACtC,MAAM,EAAC,WAAU;AACjB,QAAA,GAAS,GAAG,KAAK,eAAe,GAAG,MAAM,CAAC,QAAQ,GAAG,OAAO,GAAG;AAG/D,QAAA,eAAqB,aAAa,OAAO,OAAO;AAEhD,QAAA,aAAmB,IAAI,MAAM;AAC7B,QAAA,GAAS,OAAO,IAAI,KAAK,OAAO,OAAO;;CAGzC,iBAAiB,MAAiB;AAChC,QAAA,cAAoB,KAAK,KAAK,GAAG;EAEjC,MAAM,OAAO,cAAc,KAAK,GAAG;AACnC,QAAA,GAAS,GAAG,KAAK,wBAAwB,GAAG,KAAK,GAAG;AAGpD,QAAA,eAAqB,YAAY,KAAK;AAEtC,QAAA,WAAiB,KAAK;AACtB,QAAA,GAAS,OAAO,KAAK,KAAK,KAAK;;CAGjC,mBAAmB,QAAqB;EACtC,MAAM,QAAQ,uBAAuB,OAAO,KAAK;AACjD,QAAA,GAAS,GAAG,KAAK,yBAAyB,MAAM,CAAC;EAMjD,MAAM,YAAY,KAAK,MAAA,WAAiB,IAAI,MAAM,UAAU,CAAC;AAC7D,OACG,UAAU,eAAe,EAAE,EAAE,WAC9B,OAAO,QAAQ,UAAU,QAAQ,CAAC,SAAS,EAE3C,OAAA,kBAAwB;MAExB,OAAA,WAAiB,MAAM,UAAU;AAEnC,QAAA,GAAS,OAAO,OAAO,KAAK,MAAM,KAAK;;CAGzC,iBAAiB,MAAiB;EAChC,MAAM,OAAO,cAAc,KAAK,GAAG;AACnC,QAAA,GAAS,GAAG,KAAK,wBAAwB,GAAG,KAAK,GAAG;AACpD,QAAA,GAAS,OAAO,KAAK,KAAK,KAAK;;CAGjC,cAAc,OAAmB;AAC/B,QAAA,cAAoB,iBAAiB,OAAO,MAAA,QAAc;AAC1D,QAAA,WAAiB,cAAc,MAAM,CAAC;;;;;CAMxC,UACE,OACA,KACA,mBACA;AAIA,MAAI,MAAA,SAAe,aAAa,sBAAsB,KAAA,GAAW;AAC/D,SAAA,UAAgB,SACd,MAAA,SACA,MAAA,OACA,OACA,KACA,kBACD;AACD,SAAA;;;CAIJ,aAAa,OAAe,KAAiB,aAAwB;AAInE,MAAI,MAAA,SAAe,aAAa,aAAa,QAAQ;AACnD,SAAA,UAAgB,YAAY,MAAA,SAAe,MAAA,OAAa,OAAO,IAAI;AACnE,SAAA;;;CAIJ,eAAe,OAAe;AAC5B,MAAI,MAAA,SAAe,WAAW;AAC5B,SAAA,UAAgB,cAAc,MAAA,SAAe,MAAM;AACnD,SAAA;;;CAIJ,YAAY,OAAe;AACzB,QAAA,gBAAsB;AACtB,MAAI,MAAA,SAAe,WAAW;AAC5B,SAAA,UAAgB,WAAW,MAAA,SAAe,MAAM;AAChD,SAAA;;AAEF,QAAA,kBAAwB;;CAG1B,gBAAgB,EAAC,UAAU,WAAW,SAAS,aAA6B;EAC1E,MAAM,YAAY,cAAc,SAAS;EACzC,MAAM,YAAY,KAAK,MAAA,WAAiB,IAAI,UAAU,CAAC;EACvD,MAAM,aAAa,SAAS,OAAO;EACnC,MAAM,OAAO,CAAC,GAAG,YAAY,GAAG,QAAQ;EAGxC,MAAM,gBAAgB,CAAC,GAAG,MAAM,yBAAyB,CAAC,IAAI,GAAG,CAAC,KAAK,IAAI;EAC3E,MAAM,SAAS,MAAM,KAAK,EAAC,QAAQ,KAAK,SAAS,GAAE,CAAC,CACjD,KAAK,IAAI,CACT,KAAK,IAAI;EACZ,MAAM,gBAAgB,WAAW,IAAI,GAAG,CAAC,KAAK,IAAI;EAElD,IAAI,aAAa;EACjB,IAAI,UAAU;AACd,OAAK,MAAM,KAAK,WAAW;GACzB,MAAM,MAAM,QACV,OAAO,YAAY,KAAK,KAAK,GAAG,MAAM,CAAC,GAAG,EAAE,GAAG,CAAC,CAAC,EACjD,WACA,MAAA,WACD;GACD,MAAM,SAAS,MAAA,OAAa,KAAK,EAAC,UAAS,CAAC;GAC5C,MAAM,QAAQ,MAAA,UAAgB,eAAe,WAAW,OAAO;AAC/D,OAAI,OAAO,OAAA,OAAiB,MAAM,eAAe,WAAW;AAC1D;AACA;;GAEF,MAAM,UACJ,OAAO,OAAA,MACH,KAAK,QACH,OAAM,MAAM,0BAA0B,MAAM,OAAO,UACpD,GACD;AACN,OAAI,QAAQ,WAAW,GAAG;AAExB;AACA;;GAEF,MAAM,cAAc,QAAQ,KAAI,QAAO,GAAG,GAAG,IAAI,CAAC,YAAY,GAAG,IAAI,GAAG;AACxE,SAAA,GAAS,IACC;sBACM,GAAG,UAAU,CAAC,IAAI,cAAc,YAAY,OAAO;yBAChD,cAAc;0BACb,YAAY,KAAK,IAAI,CAAC;SAExC,GAAG,OAAO,OAAO,IAAI,IAAI,EACzB,UACD;AACD;;AAGF,QAAA,GAAS,QACP,cAAc,WAAW,iBAAiB,QAAQ,SAAS,YAC5D;;CAGH;CAEA,yBAAyB,EAAC,UAAU,SAAS,UAA4B;EACvE,MAAM,YAAY,cAAc,SAAS;EAEzC,MAAM,OAAO,CAAC,GADK,SAAS,OAAO,SACN,GAAG,QAAQ;EAExC,MAAM,iBAAiB,KAAK,oBAAoB,YAAY,MAAA,GAAS,GAAG,CAAC;AACzE,OAAK,MAAM,OAAO,KAChB,gBAAe,iBAAiB,WAAW,IAAI;AAIjD,QAAA,aAAmB,SAAS;AAC5B,MAAI,OACF,OAAA,oBAA0B;GAAC,OAAO;GAAW,SAAS;GAAM,GAAG;GAAO;AAExE,QAAA,GAAS,OAAO,wBAAwB,YAAY;;CAatD,cAAc,QAAuB,WAAiC;AACpE,MAAI,cAAc,MAAA,QAChB,OAAM,IAAI,MACR,oBAAoB,UAAU,kCAC5B,MAAA,QACD,IAAI,UAAU,OAAO,GACvB;AAEH,6BAA2B,MAAA,IAAU,UAAU;AAE/C,MAAI,MAAA,eAAqB;GACvB,MAAM,QAAQ,KAAK,KAAK;AACxB,SAAA,GAAS,GAAG,OAAO,WAAW;AAC9B,SAAA,GAAS,OACP,yCAAyC,KAAK,KAAK,GAAG,MAAM,MAC7D;;AAGH,MAAI,MAAA,SAAe,eACjB,OAAA,GAAS,QAAQ;EAGnB,MAAM,YAAY,KAAK,KAAK,GAAG,MAAA;AAC/B,QAAA,GAAS,QAAQ,gBAAgB,MAAA,QAAc,IAAI,UAAU,MAAM;AAEnE,SAAO;GACL;GACA,mBAAmB,MAAA;GACnB,eAAe,MAAA;GACf,kBAAkB,MAAA,sBAA4B;GAC/C;;CAGH,MAAM,IAAgB;AACpB,KAAG,OAAO,wBAAwB,MAAA,UAAgB;AAClD,QAAA,GAAS,UAAU;;;AAIvB,SAAS,qBACP,KACA,EAAC,eACqB;AACtB,KAAI,CAAC,aAAa,OAChB;AAEF,QAAO,YAAY,QAAO,QAAO,OAAO,IAAI;;AAG9C,SAAS,YAAY,KAAqB;AACxC,KAAI,eAAe,MACjB,QAAO;CAET,MAAM,wBAAQ,IAAI,OAAO;AACzB,OAAM,QAAQ;AACd,QAAO"}
@@ -0,0 +1,49 @@
1
+ import { promiseVoid } from "../../../../shared/src/resolved-promises.js";
2
+ import { RunningState } from "../running-state.js";
3
+ import { shadowInitialSync } from "../change-source/pg/initial-sync.js";
4
+ //#region ../zero-cache/src/services/shadow-sync/shadow-sync-service.ts
5
+ var ShadowSyncService = class {
6
+ id = "shadow-syncer";
7
+ #lc;
8
+ #shard;
9
+ #upstreamURI;
10
+ #context;
11
+ #options;
12
+ #state = new RunningState("shadow-syncer");
13
+ constructor(lc, shard, upstreamURI, context, options) {
14
+ this.#lc = lc;
15
+ this.#shard = shard;
16
+ this.#upstreamURI = upstreamURI;
17
+ this.#context = context;
18
+ this.#options = options;
19
+ }
20
+ async run() {
21
+ const { intervalMs, sampleRate, maxRowsPerTable, textCopy } = this.#options;
22
+ const firstRunDelay = intervalMs + Math.floor(Math.random() * intervalMs);
23
+ this.#lc.info?.(`shadow-syncer started; first run in ${firstRunDelay} ms, then every ${intervalMs} ms`);
24
+ await this.#state.sleep(firstRunDelay);
25
+ while (this.#state.shouldRun()) {
26
+ const start = performance.now();
27
+ try {
28
+ await shadowInitialSync(this.#lc, this.#shard, this.#upstreamURI, {
29
+ sampleRate,
30
+ maxRowsPerTable
31
+ }, this.#context, textCopy !== void 0 ? { textCopy } : void 0);
32
+ const elapsed = performance.now() - start;
33
+ this.#lc.info?.(`shadow initial-sync completed (${elapsed.toFixed(0)} ms)`);
34
+ } catch (e) {
35
+ const elapsed = performance.now() - start;
36
+ this.#lc.error?.(`shadow initial-sync failed after ${elapsed.toFixed(0)} ms`, e);
37
+ }
38
+ await this.#state.sleep(intervalMs);
39
+ }
40
+ }
41
+ stop() {
42
+ this.#state.stop(this.#lc);
43
+ return promiseVoid;
44
+ }
45
+ };
46
+ //#endregion
47
+ export { ShadowSyncService };
48
+
49
+ //# sourceMappingURL=shadow-sync-service.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"shadow-sync-service.js","names":["#lc","#shard","#upstreamURI","#context","#options","#state"],"sources":["../../../../../../zero-cache/src/services/shadow-sync/shadow-sync-service.ts"],"sourcesContent":["import type {LogContext} from '@rocicorp/logger';\nimport {promiseVoid} from '../../../../shared/src/resolved-promises.ts';\nimport type {ShardConfig} from '../../types/shards.ts';\nimport type {ServerContext} from '../change-source/pg/initial-sync.ts';\nimport {shadowInitialSync} from '../change-source/pg/initial-sync.ts';\nimport {RunningState} from '../running-state.ts';\nimport type {Service} from '../service.ts';\n\nexport type ShadowSyncOptions = {\n intervalMs: number;\n sampleRate: number;\n maxRowsPerTable: number;\n textCopy?: boolean | undefined;\n};\n\nexport class ShadowSyncService implements Service {\n readonly id = 'shadow-syncer';\n\n readonly #lc: LogContext;\n readonly #shard: ShardConfig;\n readonly #upstreamURI: string;\n readonly #context: ServerContext;\n readonly #options: ShadowSyncOptions;\n readonly #state = new RunningState('shadow-syncer');\n\n constructor(\n lc: LogContext,\n shard: ShardConfig,\n upstreamURI: string,\n context: ServerContext,\n options: ShadowSyncOptions,\n ) {\n this.#lc = lc;\n this.#shard = shard;\n this.#upstreamURI = upstreamURI;\n this.#context = context;\n this.#options = options;\n }\n\n async run() {\n const {intervalMs, sampleRate, maxRowsPerTable, textCopy} = this.#options;\n\n // Why: wait at least one full interval before the first run so shadow\n // sync never fires immediately on task startup, and add a random\n // fraction of the interval on top so a fleet-wide restart does not\n // cause every task to canary simultaneously.\n const firstRunDelay = intervalMs + Math.floor(Math.random() * intervalMs);\n this.#lc.info?.(\n `shadow-syncer started; first run in ${firstRunDelay} ms, then every ${intervalMs} ms`,\n );\n await this.#state.sleep(firstRunDelay);\n\n while (this.#state.shouldRun()) {\n const start = performance.now();\n try {\n await shadowInitialSync(\n this.#lc,\n this.#shard,\n this.#upstreamURI,\n {sampleRate, maxRowsPerTable},\n this.#context,\n textCopy !== undefined ? {textCopy} : undefined,\n );\n const elapsed = performance.now() - start;\n this.#lc.info?.(\n `shadow initial-sync completed (${elapsed.toFixed(0)} ms)`,\n );\n } catch (e) {\n const elapsed = performance.now() - start;\n this.#lc.error?.(\n `shadow initial-sync failed after ${elapsed.toFixed(0)} ms`,\n e,\n );\n }\n await this.#state.sleep(intervalMs);\n }\n }\n\n stop(): Promise<void> {\n this.#state.stop(this.#lc);\n return promiseVoid;\n }\n}\n"],"mappings":";;;;AAeA,IAAa,oBAAb,MAAkD;CAChD,KAAc;CAEd;CACA;CACA;CACA;CACA;CACA,SAAkB,IAAI,aAAa,gBAAgB;CAEnD,YACE,IACA,OACA,aACA,SACA,SACA;AACA,QAAA,KAAW;AACX,QAAA,QAAc;AACd,QAAA,cAAoB;AACpB,QAAA,UAAgB;AAChB,QAAA,UAAgB;;CAGlB,MAAM,MAAM;EACV,MAAM,EAAC,YAAY,YAAY,iBAAiB,aAAY,MAAA;EAM5D,MAAM,gBAAgB,aAAa,KAAK,MAAM,KAAK,QAAQ,GAAG,WAAW;AACzE,QAAA,GAAS,OACP,uCAAuC,cAAc,kBAAkB,WAAW,KACnF;AACD,QAAM,MAAA,MAAY,MAAM,cAAc;AAEtC,SAAO,MAAA,MAAY,WAAW,EAAE;GAC9B,MAAM,QAAQ,YAAY,KAAK;AAC/B,OAAI;AACF,UAAM,kBACJ,MAAA,IACA,MAAA,OACA,MAAA,aACA;KAAC;KAAY;KAAgB,EAC7B,MAAA,SACA,aAAa,KAAA,IAAY,EAAC,UAAS,GAAG,KAAA,EACvC;IACD,MAAM,UAAU,YAAY,KAAK,GAAG;AACpC,UAAA,GAAS,OACP,kCAAkC,QAAQ,QAAQ,EAAE,CAAC,MACtD;YACM,GAAG;IACV,MAAM,UAAU,YAAY,KAAK,GAAG;AACpC,UAAA,GAAS,QACP,oCAAoC,QAAQ,QAAQ,EAAE,CAAC,MACvD,EACD;;AAEH,SAAM,MAAA,MAAY,MAAM,WAAW;;;CAIvC,OAAsB;AACpB,QAAA,MAAY,KAAK,MAAA,GAAS;AAC1B,SAAO"}
@@ -12,7 +12,7 @@ import auth from "basic-auth";
12
12
  //#region ../zero-cache/src/services/statz.ts
13
13
  async function upstreamStats(lc, config) {
14
14
  const schema = upstreamSchema(getShardID(config));
15
- const sql = pgClient(lc, config.upstream.db);
15
+ const sql = pgClient(lc, config.upstream.db, "statz-upstream");
16
16
  try {
17
17
  return await getPgStats([
18
18
  ["numReplicas", sql`SELECT COUNT(*) as "c" FROM ${sql(schema)}."replicas"`],
@@ -25,7 +25,7 @@ async function upstreamStats(lc, config) {
25
25
  }
26
26
  async function cvrStats(lc, config) {
27
27
  const schema = upstreamSchema(getShardID(config)) + "/cvr";
28
- const sql = pgClient(lc, config.cvr.db);
28
+ const sql = pgClient(lc, config.cvr.db, "statz-cvr");
29
29
  function numQueriesPerClientGroup(active) {
30
30
  const filter = active ? sql`WHERE "inactivatedAt" IS NULL AND deleted = false` : sql`WHERE "inactivatedAt" IS NOT NULL AND ("inactivatedAt" + "ttl") > NOW()`;
31
31
  return sql`WITH
@@ -136,7 +136,7 @@ async function cvrStats(lc, config) {
136
136
  }
137
137
  async function changeLogStats(lc, config) {
138
138
  const schema = upstreamSchema(getShardID(config)) + "/cdc";
139
- const sql = pgClient(lc, config.change.db);
139
+ const sql = pgClient(lc, config.change.db, "statz-change");
140
140
  try {
141
141
  return await getPgStats([["changeLogSize", sql`SELECT COUNT(*) as "change_log_size" FROM ${sql(schema)}."changeLog"`]]);
142
142
  } finally {
@@ -1 +1 @@
1
- {"version":3,"file":"statz.js","names":[],"sources":["../../../../../zero-cache/src/services/statz.ts"],"sourcesContent":["import fs from 'fs';\nimport os from 'os';\nimport type {LogContext} from '@rocicorp/logger';\nimport auth from 'basic-auth';\nimport type {FastifyReply, FastifyRequest} from 'fastify';\nimport {BigIntJSON} from '../../../shared/src/bigint-json.ts';\nimport {Database} from '../../../zqlite/src/db.ts';\nimport type {NormalizedZeroConfig as ZeroConfig} from '../config/normalize.ts';\nimport {isAdminPasswordValid} from '../config/zero-config.ts';\nimport {StatementRunner} from '../db/statements.ts';\nimport {pgClient} from '../types/pg.ts';\nimport {getShardID, upstreamSchema} from '../types/shards.ts';\nimport {fromStateVersionString} from './change-source/pg/lsn.ts';\nimport {getReplicationState} from './replicator/schema/replication-state.ts';\n\nasync function upstreamStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config));\n const sql = pgClient(lc, config.upstream.db);\n try {\n return await getPgStats([\n [\n 'numReplicas',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"replicas\"`,\n ],\n [\n 'numClientsWithMutations',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"clients\"`,\n ],\n [\n 'numMutationsProcessed',\n sql`SELECT SUM(\"lastMutationID\") as \"c\" FROM ${sql(schema)}.\"clients\"`,\n ],\n ]);\n } finally {\n await sql.end();\n }\n}\n\nasync function cvrStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config)) + '/cvr';\n const sql = pgClient(lc, config.cvr.db);\n\n function numQueriesPerClientGroup(\n active: boolean,\n ): ReturnType<ReturnType<typeof pgClient>> {\n const filter = active\n ? sql`WHERE \"inactivatedAt\" IS NULL AND deleted = false`\n : sql`WHERE \"inactivatedAt\" IS NOT NULL AND (\"inactivatedAt\" + \"ttl\") > NOW()`;\n return sql`WITH\n group_counts AS (\n SELECT\n \"clientGroupID\",\n COUNT(*) AS num_queries\n FROM ${sql(schema)}.\"desires\"\n ${filter}\n GROUP BY \"clientGroupID\"\n ),\n -- Count distinct clientIDs per clientGroupID\n client_per_group_counts AS (\n SELECT\n \"clientGroupID\",\n COUNT(DISTINCT \"clientID\") AS num_clients\n FROM ${sql(schema)}.\"desires\"\n ${filter}\n GROUP BY \"clientGroupID\"\n )\n -- Combine all the information\n SELECT\n g.\"clientGroupID\",\n cpg.num_clients,\n g.num_queries\n FROM group_counts g\n JOIN client_per_group_counts cpg ON g.\"clientGroupID\" = cpg.\"clientGroupID\"\n ORDER BY g.num_queries DESC;`;\n }\n\n try {\n return await getPgStats([\n [\n 'totalNumQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\"`,\n ],\n [\n 'numUniqueQueryHashes',\n sql`SELECT COUNT(DISTINCT \"queryHash\") as \"c\" FROM ${sql(\n schema,\n )}.\"desires\"`,\n ],\n [\n 'numActiveQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"inactivatedAt\" IS NULL AND \"deleted\" = false`,\n ],\n [\n 'numInactiveQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"inactivatedAt\" IS NOT NULL AND (\"inactivatedAt\" + \"ttl\") > NOW()`,\n ],\n [\n 'numDeletedQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"deleted\" = true`,\n ],\n [\n 'freshQueriesPercentiles',\n sql`WITH client_group_counts AS (\n -- Count inactive desires per clientGroupID\n SELECT\n \"clientGroupID\",\n COUNT(*) AS fresh_count\n FROM ${sql(schema)}.\"desires\"\n WHERE\n (\"inactivatedAt\" IS NOT NULL\n AND (\"inactivatedAt\" + \"ttl\") > NOW()) OR (\"inactivatedAt\" IS NULL\n AND deleted = false)\n GROUP BY \"clientGroupID\"\n )\n\n SELECT\n percentile_cont(0.50) WITHIN GROUP (ORDER BY fresh_count) AS \"p50\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY fresh_count) AS \"p75\",\n percentile_cont(0.90) WITHIN GROUP (ORDER BY fresh_count) AS \"p90\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY fresh_count) AS \"p95\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY fresh_count) AS \"p99\",\n MIN(fresh_count) AS \"min\",\n MAX(fresh_count) AS \"max\",\n AVG(fresh_count) AS \"avg\"\n FROM client_group_counts;`,\n ],\n [\n 'rowsPerClientGroupPercentiles',\n sql`WITH client_group_counts AS (\n -- Count inactive desires per clientGroupID\n SELECT\n \"clientGroupID\",\n COUNT(*) AS row_count\n FROM ${sql(schema)}.\"rows\"\n GROUP BY \"clientGroupID\"\n )\n SELECT\n percentile_cont(0.50) WITHIN GROUP (ORDER BY row_count) AS \"p50\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY row_count) AS \"p75\",\n percentile_cont(0.90) WITHIN GROUP (ORDER BY row_count) AS \"p90\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY row_count) AS \"p95\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY row_count) AS \"p99\",\n MIN(row_count) AS \"min\",\n MAX(row_count) AS \"max\",\n AVG(row_count) AS \"avg\"\n FROM client_group_counts;`,\n ],\n [\n // check for AST blowup due to DNF conversion.\n 'astSizes',\n sql`SELECT\n percentile_cont(0.25) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"25th_percentile\",\n percentile_cont(0.5) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"50th_percentile\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"75th_percentile\",\n percentile_cont(0.9) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"90th_percentile\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"95th_percentile\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"99th_percentile\",\n MIN(length(\"clientAST\"::text)) AS \"minimum_length\",\n MAX(length(\"clientAST\"::text)) AS \"maximum_length\",\n AVG(length(\"clientAST\"::text))::integer AS \"average_length\",\n COUNT(*) AS \"total_records\"\n FROM ${sql(schema)}.\"queries\";`,\n ],\n [\n // output the hash of the largest AST\n 'biggestAstHash',\n sql`SELECT \"queryHash\", length(\"clientAST\"::text) AS \"ast_length\"\n FROM ${sql(schema)}.\"queries\"\n ORDER BY length(\"clientAST\"::text) DESC\n LIMIT 1;`,\n ],\n [\n 'totalActiveQueriesPerClientAndClientGroup',\n numQueriesPerClientGroup(true),\n ],\n [\n 'totalInactiveQueriesPerClientAndClientGroup',\n numQueriesPerClientGroup(false),\n ],\n [\n 'totalRowsPerClientGroup',\n sql`SELECT \"clientGroupID\", COUNT(*) as \"c\" FROM ${sql(\n schema,\n )}.\"rows\" GROUP BY \"clientGroupID\" ORDER BY \"c\" DESC`,\n ],\n [\n 'numRowsPerQuery',\n sql`SELECT\n k.key AS \"queryHash\",\n COUNT(*) AS row_count\n FROM ${sql(schema)}.\"rows\" r,\n LATERAL jsonb_each(r.\"refCounts\") k\n GROUP BY k.key\n ORDER BY row_count DESC;`,\n ],\n ] satisfies [\n name: string,\n query: ReturnType<ReturnType<typeof pgClient>>,\n ][]);\n } finally {\n await sql.end();\n }\n}\n\nasync function changeLogStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config)) + '/cdc';\n const sql = pgClient(lc, config.change.db);\n\n try {\n return await getPgStats([\n [\n 'changeLogSize',\n sql`SELECT COUNT(*) as \"change_log_size\" FROM ${sql(schema)}.\"changeLog\"`,\n ],\n ]);\n } finally {\n await sql.end();\n }\n}\n\nfunction replicaStats(lc: LogContext, config: ZeroConfig) {\n const db = new Database(lc, config.replica.file);\n try {\n return Object.fromEntries([\n ['wal checkpoint', pick(first(db.pragma('WAL_CHECKPOINT')))],\n ['page count', pick(first(db.pragma('PAGE_COUNT')))],\n ['page size', pick(first(db.pragma('PAGE_SIZE')))],\n ['journal mode', pick(first(db.pragma('JOURNAL_MODE')))],\n ['synchronous', pick(first(db.pragma('SYNCHRONOUS')))],\n ['cache size', pick(first(db.pragma('CACHE_SIZE')))],\n ['auto vacuum', pick(first(db.pragma('AUTO_VACUUM')))],\n ['freelist count', pick(first(db.pragma('FREELIST_COUNT')))],\n ['wal autocheckpoint', pick(first(db.pragma('WAL_AUTOCHECKPOINT')))],\n ['db file stats', fs.statSync(config.replica.file)],\n ] as const);\n } finally {\n db.close();\n }\n}\n\nfunction replicationStats(lc: LogContext, config: ZeroConfig) {\n const db = new Database(lc, config.replica.file);\n try {\n return getReplicationStats(db);\n } finally {\n db.close();\n }\n}\n\nfunction getReplicationStats(db: Database) {\n const {stateVersion} = getReplicationState(new StatementRunner(db));\n const lsn = fromStateVersionString(stateVersion);\n return {lsn};\n}\n\nfunction osStats() {\n return Object.fromEntries([\n ['load avg', os.loadavg()],\n ['uptime', os.uptime()],\n ['total mem', os.totalmem()],\n ['free mem', os.freemem()],\n ['cpus', os.cpus().length],\n ['available parallelism', os.availableParallelism()],\n ['platform', os.platform()],\n ['arch', os.arch()],\n ['release', os.release()],\n ] as const);\n}\n\nasync function getPgStats(\n pendingQueries: [\n name: string,\n query: ReturnType<ReturnType<typeof pgClient>>,\n ][],\n) {\n const results = await Promise.all(\n pendingQueries.map(async ([name, query]) => [name, await query] as const),\n );\n return Object.fromEntries(results);\n}\n\ntype StatsObject = Record<string, unknown>;\n\nfunction printStats(group: string, statsObject: StatsObject): string {\n const lines: string[] = ['\\n' + header(group)];\n for (const [name, result] of Object.entries(statsObject)) {\n lines.push('\\n' + name + ': ' + BigIntJSON.stringify(result, null, 2));\n }\n lines.push('\\n');\n return lines.join('');\n}\n\n/**\n * HTTP query parameters:\n * * `group`: restricts the groups for which stats are computed\n * * `format=json`: returns the stats as a JSON object\n * * `pretty`: formats the JSON object with indentation\n */\nexport async function handleStatzRequest(\n lc: LogContext,\n config: ZeroConfig,\n req: FastifyRequest,\n res: FastifyReply,\n) {\n const credentials = auth(req);\n if (!isAdminPasswordValid(lc, config, credentials?.pass)) {\n void res\n .code(401)\n .header('WWW-Authenticate', 'Basic realm=\"Statz Protected Area\"')\n .send('Unauthorized');\n return;\n }\n\n const statsFns: Record<string, () => Promise<StatsObject> | StatsObject> = {\n upstream: () => upstreamStats(lc, config),\n cvr: () => cvrStats(lc, config),\n changeLog: () => changeLogStats(lc, config),\n replica: () => replicaStats(lc, config),\n replication: () => replicationStats(lc, config),\n os: () => osStats(),\n };\n\n async function computeStats(group: string): Promise<[string, StatsObject]> {\n try {\n return [group, await statsFns[group]()];\n } catch (e) {\n lc.error?.(`error computing ${group} stats`, e);\n return [group, {error: String(e)}];\n }\n }\n\n const query = req.query as Record<string, unknown>;\n const groups =\n typeof query.group === 'string'\n ? query.group.split(',')\n : Array.isArray(query.group)\n ? query.group\n : undefined;\n\n const stats = await Promise.all(\n groups\n ? groups.filter(g => g in statsFns).map(computeStats)\n : Object.keys(statsFns).map(computeStats),\n );\n\n if (query.format === 'json') {\n const indent = query.pretty !== undefined ? 2 : undefined;\n await res\n .header('Content-Type', 'application/json')\n .send(BigIntJSON.stringify(Object.fromEntries(stats), null, indent));\n return;\n } else {\n const body = stats\n .map(([group, statsObject]) => printStats(group, statsObject))\n .join('');\n await res.header('Content-Type', 'text/plain; charset=utf-8').send(body);\n }\n}\n\nfunction first(x: object[]): object {\n return x[0];\n}\n\nfunction pick(x: object): unknown {\n return Object.values(x)[0];\n}\n\nfunction header(name: string): string {\n return `=== ${name} ===\\n`;\n}\n"],"mappings":";;;;;;;;;;;;AAeA,eAAe,cAAc,IAAgB,QAAoB;CAC/D,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC;CACjD,MAAM,MAAM,SAAS,IAAI,OAAO,SAAS,GAAG;AAC5C,KAAI;AACF,SAAO,MAAM,WAAW;GACtB,CACE,eACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,aAC/C;GACD,CACE,2BACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,YAC/C;GACD,CACE,yBACA,GAAG,4CAA4C,IAAI,OAAO,CAAC,YAC5D;GACF,CAAC;WACM;AACR,QAAM,IAAI,KAAK;;;AAInB,eAAe,SAAS,IAAgB,QAAoB;CAC1D,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC,GAAG;CACpD,MAAM,MAAM,SAAS,IAAI,OAAO,IAAI,GAAG;CAEvC,SAAS,yBACP,QACyC;EACzC,MAAM,SAAS,SACX,GAAG,sDACH,GAAG;AACP,SAAO,GAAG;;;;;aAKD,IAAI,OAAO,CAAC;QACjB,OAAO;;;;;;;;aAQF,IAAI,OAAO,CAAC;QACjB,OAAO;;;;;;;;;;;;AAab,KAAI;AACF,SAAO,MAAM,WAAW;GACtB,CACE,mBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,YAC/C;GACD,CACE,wBACA,GAAG,kDAAkD,IACnD,OACD,CAAC,YACH;GACD,CACE,oBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,gEAC/C;GACD,CACE,sBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,oFAC/C;GACD,CACE,qBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,mCAC/C;GACD,CACE,2BACA,GAAG;;;;;eAKI,IAAI,OAAO,CAAC;;;;;;;;;;;;;;;;;iCAkBpB;GACD,CACE,iCACA,GAAG;;;;;eAKI,IAAI,OAAO,CAAC;;;;;;;;;;;;iCAapB;GACD,CAEE,YACA,GAAG;;;;;;;;;;;aAWE,IAAI,OAAO,CAAC,aAClB;GACD,CAEE,kBACA,GAAG;aACE,IAAI,OAAO,CAAC;;gBAGlB;GACD,CACE,6CACA,yBAAyB,KAAK,CAC/B;GACD,CACE,+CACA,yBAAyB,MAAM,CAChC;GACD,CACE,2BACA,GAAG,gDAAgD,IACjD,OACD,CAAC,oDACH;GACD,CACE,mBACA,GAAG;;;aAGE,IAAI,OAAO,CAAC;;;gCAIlB;GACF,CAGG;WACI;AACR,QAAM,IAAI,KAAK;;;AAInB,eAAe,eAAe,IAAgB,QAAoB;CAChE,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC,GAAG;CACpD,MAAM,MAAM,SAAS,IAAI,OAAO,OAAO,GAAG;AAE1C,KAAI;AACF,SAAO,MAAM,WAAW,CACtB,CACE,iBACA,GAAG,6CAA6C,IAAI,OAAO,CAAC,cAC7D,CACF,CAAC;WACM;AACR,QAAM,IAAI,KAAK;;;AAInB,SAAS,aAAa,IAAgB,QAAoB;CACxD,MAAM,KAAK,IAAI,SAAS,IAAI,OAAO,QAAQ,KAAK;AAChD,KAAI;AACF,SAAO,OAAO,YAAY;GACxB,CAAC,kBAAkB,KAAK,MAAM,GAAG,OAAO,iBAAiB,CAAC,CAAC,CAAC;GAC5D,CAAC,cAAc,KAAK,MAAM,GAAG,OAAO,aAAa,CAAC,CAAC,CAAC;GACpD,CAAC,aAAa,KAAK,MAAM,GAAG,OAAO,YAAY,CAAC,CAAC,CAAC;GAClD,CAAC,gBAAgB,KAAK,MAAM,GAAG,OAAO,eAAe,CAAC,CAAC,CAAC;GACxD,CAAC,eAAe,KAAK,MAAM,GAAG,OAAO,cAAc,CAAC,CAAC,CAAC;GACtD,CAAC,cAAc,KAAK,MAAM,GAAG,OAAO,aAAa,CAAC,CAAC,CAAC;GACpD,CAAC,eAAe,KAAK,MAAM,GAAG,OAAO,cAAc,CAAC,CAAC,CAAC;GACtD,CAAC,kBAAkB,KAAK,MAAM,GAAG,OAAO,iBAAiB,CAAC,CAAC,CAAC;GAC5D,CAAC,sBAAsB,KAAK,MAAM,GAAG,OAAO,qBAAqB,CAAC,CAAC,CAAC;GACpE,CAAC,iBAAiB,GAAG,SAAS,OAAO,QAAQ,KAAK,CAAC;GACpD,CAAU;WACH;AACR,KAAG,OAAO;;;AAId,SAAS,iBAAiB,IAAgB,QAAoB;CAC5D,MAAM,KAAK,IAAI,SAAS,IAAI,OAAO,QAAQ,KAAK;AAChD,KAAI;AACF,SAAO,oBAAoB,GAAG;WACtB;AACR,KAAG,OAAO;;;AAId,SAAS,oBAAoB,IAAc;CACzC,MAAM,EAAC,iBAAgB,oBAAoB,IAAI,gBAAgB,GAAG,CAAC;AAEnE,QAAO,EAAC,KADI,uBAAuB,aAAa,EACpC;;AAGd,SAAS,UAAU;AACjB,QAAO,OAAO,YAAY;EACxB,CAAC,YAAY,GAAG,SAAS,CAAC;EAC1B,CAAC,UAAU,GAAG,QAAQ,CAAC;EACvB,CAAC,aAAa,GAAG,UAAU,CAAC;EAC5B,CAAC,YAAY,GAAG,SAAS,CAAC;EAC1B,CAAC,QAAQ,GAAG,MAAM,CAAC,OAAO;EAC1B,CAAC,yBAAyB,GAAG,sBAAsB,CAAC;EACpD,CAAC,YAAY,GAAG,UAAU,CAAC;EAC3B,CAAC,QAAQ,GAAG,MAAM,CAAC;EACnB,CAAC,WAAW,GAAG,SAAS,CAAC;EAC1B,CAAU;;AAGb,eAAe,WACb,gBAIA;CACA,MAAM,UAAU,MAAM,QAAQ,IAC5B,eAAe,IAAI,OAAO,CAAC,MAAM,WAAW,CAAC,MAAM,MAAM,MAAM,CAAU,CAC1E;AACD,QAAO,OAAO,YAAY,QAAQ;;AAKpC,SAAS,WAAW,OAAe,aAAkC;CACnE,MAAM,QAAkB,CAAC,OAAO,OAAO,MAAM,CAAC;AAC9C,MAAK,MAAM,CAAC,MAAM,WAAW,OAAO,QAAQ,YAAY,CACtD,OAAM,KAAK,OAAO,OAAO,OAAO,WAAW,UAAU,QAAQ,MAAM,EAAE,CAAC;AAExE,OAAM,KAAK,KAAK;AAChB,QAAO,MAAM,KAAK,GAAG;;;;;;;;AASvB,eAAsB,mBACpB,IACA,QACA,KACA,KACA;AAEA,KAAI,CAAC,qBAAqB,IAAI,QADV,KAAK,IAAI,EACsB,KAAK,EAAE;AACnD,MACF,KAAK,IAAI,CACT,OAAO,oBAAoB,uCAAqC,CAChE,KAAK,eAAe;AACvB;;CAGF,MAAM,WAAqE;EACzE,gBAAgB,cAAc,IAAI,OAAO;EACzC,WAAW,SAAS,IAAI,OAAO;EAC/B,iBAAiB,eAAe,IAAI,OAAO;EAC3C,eAAe,aAAa,IAAI,OAAO;EACvC,mBAAmB,iBAAiB,IAAI,OAAO;EAC/C,UAAU,SAAS;EACpB;CAED,eAAe,aAAa,OAA+C;AACzE,MAAI;AACF,UAAO,CAAC,OAAO,MAAM,SAAS,QAAQ,CAAC;WAChC,GAAG;AACV,MAAG,QAAQ,mBAAmB,MAAM,SAAS,EAAE;AAC/C,UAAO,CAAC,OAAO,EAAC,OAAO,OAAO,EAAE,EAAC,CAAC;;;CAItC,MAAM,QAAQ,IAAI;CAClB,MAAM,SACJ,OAAO,MAAM,UAAU,WACnB,MAAM,MAAM,MAAM,IAAI,GACtB,MAAM,QAAQ,MAAM,MAAM,GACxB,MAAM,QACN,KAAA;CAER,MAAM,QAAQ,MAAM,QAAQ,IAC1B,SACI,OAAO,QAAO,MAAK,KAAK,SAAS,CAAC,IAAI,aAAa,GACnD,OAAO,KAAK,SAAS,CAAC,IAAI,aAAa,CAC5C;AAED,KAAI,MAAM,WAAW,QAAQ;EAC3B,MAAM,SAAS,MAAM,WAAW,KAAA,IAAY,IAAI,KAAA;AAChD,QAAM,IACH,OAAO,gBAAgB,mBAAmB,CAC1C,KAAK,WAAW,UAAU,OAAO,YAAY,MAAM,EAAE,MAAM,OAAO,CAAC;AACtE;QACK;EACL,MAAM,OAAO,MACV,KAAK,CAAC,OAAO,iBAAiB,WAAW,OAAO,YAAY,CAAC,CAC7D,KAAK,GAAG;AACX,QAAM,IAAI,OAAO,gBAAgB,4BAA4B,CAAC,KAAK,KAAK;;;AAI5E,SAAS,MAAM,GAAqB;AAClC,QAAO,EAAE;;AAGX,SAAS,KAAK,GAAoB;AAChC,QAAO,OAAO,OAAO,EAAE,CAAC;;AAG1B,SAAS,OAAO,MAAsB;AACpC,QAAO,OAAO,KAAK"}
1
+ {"version":3,"file":"statz.js","names":[],"sources":["../../../../../zero-cache/src/services/statz.ts"],"sourcesContent":["import fs from 'fs';\nimport os from 'os';\nimport type {LogContext} from '@rocicorp/logger';\nimport auth from 'basic-auth';\nimport type {FastifyReply, FastifyRequest} from 'fastify';\nimport {BigIntJSON} from '../../../shared/src/bigint-json.ts';\nimport {Database} from '../../../zqlite/src/db.ts';\nimport type {NormalizedZeroConfig as ZeroConfig} from '../config/normalize.ts';\nimport {isAdminPasswordValid} from '../config/zero-config.ts';\nimport {StatementRunner} from '../db/statements.ts';\nimport {pgClient} from '../types/pg.ts';\nimport {getShardID, upstreamSchema} from '../types/shards.ts';\nimport {fromStateVersionString} from './change-source/pg/lsn.ts';\nimport {getReplicationState} from './replicator/schema/replication-state.ts';\n\nasync function upstreamStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config));\n const sql = pgClient(lc, config.upstream.db, 'statz-upstream');\n try {\n return await getPgStats([\n [\n 'numReplicas',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"replicas\"`,\n ],\n [\n 'numClientsWithMutations',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"clients\"`,\n ],\n [\n 'numMutationsProcessed',\n sql`SELECT SUM(\"lastMutationID\") as \"c\" FROM ${sql(schema)}.\"clients\"`,\n ],\n ]);\n } finally {\n await sql.end();\n }\n}\n\nasync function cvrStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config)) + '/cvr';\n const sql = pgClient(lc, config.cvr.db, 'statz-cvr');\n\n function numQueriesPerClientGroup(\n active: boolean,\n ): ReturnType<ReturnType<typeof pgClient>> {\n const filter = active\n ? sql`WHERE \"inactivatedAt\" IS NULL AND deleted = false`\n : sql`WHERE \"inactivatedAt\" IS NOT NULL AND (\"inactivatedAt\" + \"ttl\") > NOW()`;\n return sql`WITH\n group_counts AS (\n SELECT\n \"clientGroupID\",\n COUNT(*) AS num_queries\n FROM ${sql(schema)}.\"desires\"\n ${filter}\n GROUP BY \"clientGroupID\"\n ),\n -- Count distinct clientIDs per clientGroupID\n client_per_group_counts AS (\n SELECT\n \"clientGroupID\",\n COUNT(DISTINCT \"clientID\") AS num_clients\n FROM ${sql(schema)}.\"desires\"\n ${filter}\n GROUP BY \"clientGroupID\"\n )\n -- Combine all the information\n SELECT\n g.\"clientGroupID\",\n cpg.num_clients,\n g.num_queries\n FROM group_counts g\n JOIN client_per_group_counts cpg ON g.\"clientGroupID\" = cpg.\"clientGroupID\"\n ORDER BY g.num_queries DESC;`;\n }\n\n try {\n return await getPgStats([\n [\n 'totalNumQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\"`,\n ],\n [\n 'numUniqueQueryHashes',\n sql`SELECT COUNT(DISTINCT \"queryHash\") as \"c\" FROM ${sql(\n schema,\n )}.\"desires\"`,\n ],\n [\n 'numActiveQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"inactivatedAt\" IS NULL AND \"deleted\" = false`,\n ],\n [\n 'numInactiveQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"inactivatedAt\" IS NOT NULL AND (\"inactivatedAt\" + \"ttl\") > NOW()`,\n ],\n [\n 'numDeletedQueries',\n sql`SELECT COUNT(*) as \"c\" FROM ${sql(schema)}.\"desires\" WHERE \"deleted\" = true`,\n ],\n [\n 'freshQueriesPercentiles',\n sql`WITH client_group_counts AS (\n -- Count inactive desires per clientGroupID\n SELECT\n \"clientGroupID\",\n COUNT(*) AS fresh_count\n FROM ${sql(schema)}.\"desires\"\n WHERE\n (\"inactivatedAt\" IS NOT NULL\n AND (\"inactivatedAt\" + \"ttl\") > NOW()) OR (\"inactivatedAt\" IS NULL\n AND deleted = false)\n GROUP BY \"clientGroupID\"\n )\n\n SELECT\n percentile_cont(0.50) WITHIN GROUP (ORDER BY fresh_count) AS \"p50\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY fresh_count) AS \"p75\",\n percentile_cont(0.90) WITHIN GROUP (ORDER BY fresh_count) AS \"p90\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY fresh_count) AS \"p95\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY fresh_count) AS \"p99\",\n MIN(fresh_count) AS \"min\",\n MAX(fresh_count) AS \"max\",\n AVG(fresh_count) AS \"avg\"\n FROM client_group_counts;`,\n ],\n [\n 'rowsPerClientGroupPercentiles',\n sql`WITH client_group_counts AS (\n -- Count inactive desires per clientGroupID\n SELECT\n \"clientGroupID\",\n COUNT(*) AS row_count\n FROM ${sql(schema)}.\"rows\"\n GROUP BY \"clientGroupID\"\n )\n SELECT\n percentile_cont(0.50) WITHIN GROUP (ORDER BY row_count) AS \"p50\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY row_count) AS \"p75\",\n percentile_cont(0.90) WITHIN GROUP (ORDER BY row_count) AS \"p90\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY row_count) AS \"p95\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY row_count) AS \"p99\",\n MIN(row_count) AS \"min\",\n MAX(row_count) AS \"max\",\n AVG(row_count) AS \"avg\"\n FROM client_group_counts;`,\n ],\n [\n // check for AST blowup due to DNF conversion.\n 'astSizes',\n sql`SELECT\n percentile_cont(0.25) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"25th_percentile\",\n percentile_cont(0.5) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"50th_percentile\",\n percentile_cont(0.75) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"75th_percentile\",\n percentile_cont(0.9) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"90th_percentile\",\n percentile_cont(0.95) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"95th_percentile\",\n percentile_cont(0.99) WITHIN GROUP (ORDER BY length(\"clientAST\"::text)) AS \"99th_percentile\",\n MIN(length(\"clientAST\"::text)) AS \"minimum_length\",\n MAX(length(\"clientAST\"::text)) AS \"maximum_length\",\n AVG(length(\"clientAST\"::text))::integer AS \"average_length\",\n COUNT(*) AS \"total_records\"\n FROM ${sql(schema)}.\"queries\";`,\n ],\n [\n // output the hash of the largest AST\n 'biggestAstHash',\n sql`SELECT \"queryHash\", length(\"clientAST\"::text) AS \"ast_length\"\n FROM ${sql(schema)}.\"queries\"\n ORDER BY length(\"clientAST\"::text) DESC\n LIMIT 1;`,\n ],\n [\n 'totalActiveQueriesPerClientAndClientGroup',\n numQueriesPerClientGroup(true),\n ],\n [\n 'totalInactiveQueriesPerClientAndClientGroup',\n numQueriesPerClientGroup(false),\n ],\n [\n 'totalRowsPerClientGroup',\n sql`SELECT \"clientGroupID\", COUNT(*) as \"c\" FROM ${sql(\n schema,\n )}.\"rows\" GROUP BY \"clientGroupID\" ORDER BY \"c\" DESC`,\n ],\n [\n 'numRowsPerQuery',\n sql`SELECT\n k.key AS \"queryHash\",\n COUNT(*) AS row_count\n FROM ${sql(schema)}.\"rows\" r,\n LATERAL jsonb_each(r.\"refCounts\") k\n GROUP BY k.key\n ORDER BY row_count DESC;`,\n ],\n ] satisfies [\n name: string,\n query: ReturnType<ReturnType<typeof pgClient>>,\n ][]);\n } finally {\n await sql.end();\n }\n}\n\nasync function changeLogStats(lc: LogContext, config: ZeroConfig) {\n const schema = upstreamSchema(getShardID(config)) + '/cdc';\n const sql = pgClient(lc, config.change.db, 'statz-change');\n\n try {\n return await getPgStats([\n [\n 'changeLogSize',\n sql`SELECT COUNT(*) as \"change_log_size\" FROM ${sql(schema)}.\"changeLog\"`,\n ],\n ]);\n } finally {\n await sql.end();\n }\n}\n\nfunction replicaStats(lc: LogContext, config: ZeroConfig) {\n const db = new Database(lc, config.replica.file);\n try {\n return Object.fromEntries([\n ['wal checkpoint', pick(first(db.pragma('WAL_CHECKPOINT')))],\n ['page count', pick(first(db.pragma('PAGE_COUNT')))],\n ['page size', pick(first(db.pragma('PAGE_SIZE')))],\n ['journal mode', pick(first(db.pragma('JOURNAL_MODE')))],\n ['synchronous', pick(first(db.pragma('SYNCHRONOUS')))],\n ['cache size', pick(first(db.pragma('CACHE_SIZE')))],\n ['auto vacuum', pick(first(db.pragma('AUTO_VACUUM')))],\n ['freelist count', pick(first(db.pragma('FREELIST_COUNT')))],\n ['wal autocheckpoint', pick(first(db.pragma('WAL_AUTOCHECKPOINT')))],\n ['db file stats', fs.statSync(config.replica.file)],\n ] as const);\n } finally {\n db.close();\n }\n}\n\nfunction replicationStats(lc: LogContext, config: ZeroConfig) {\n const db = new Database(lc, config.replica.file);\n try {\n return getReplicationStats(db);\n } finally {\n db.close();\n }\n}\n\nfunction getReplicationStats(db: Database) {\n const {stateVersion} = getReplicationState(new StatementRunner(db));\n const lsn = fromStateVersionString(stateVersion);\n return {lsn};\n}\n\nfunction osStats() {\n return Object.fromEntries([\n ['load avg', os.loadavg()],\n ['uptime', os.uptime()],\n ['total mem', os.totalmem()],\n ['free mem', os.freemem()],\n ['cpus', os.cpus().length],\n ['available parallelism', os.availableParallelism()],\n ['platform', os.platform()],\n ['arch', os.arch()],\n ['release', os.release()],\n ] as const);\n}\n\nasync function getPgStats(\n pendingQueries: [\n name: string,\n query: ReturnType<ReturnType<typeof pgClient>>,\n ][],\n) {\n const results = await Promise.all(\n pendingQueries.map(async ([name, query]) => [name, await query] as const),\n );\n return Object.fromEntries(results);\n}\n\ntype StatsObject = Record<string, unknown>;\n\nfunction printStats(group: string, statsObject: StatsObject): string {\n const lines: string[] = ['\\n' + header(group)];\n for (const [name, result] of Object.entries(statsObject)) {\n lines.push('\\n' + name + ': ' + BigIntJSON.stringify(result, null, 2));\n }\n lines.push('\\n');\n return lines.join('');\n}\n\n/**\n * HTTP query parameters:\n * * `group`: restricts the groups for which stats are computed\n * * `format=json`: returns the stats as a JSON object\n * * `pretty`: formats the JSON object with indentation\n */\nexport async function handleStatzRequest(\n lc: LogContext,\n config: ZeroConfig,\n req: FastifyRequest,\n res: FastifyReply,\n) {\n const credentials = auth(req);\n if (!isAdminPasswordValid(lc, config, credentials?.pass)) {\n void res\n .code(401)\n .header('WWW-Authenticate', 'Basic realm=\"Statz Protected Area\"')\n .send('Unauthorized');\n return;\n }\n\n const statsFns: Record<string, () => Promise<StatsObject> | StatsObject> = {\n upstream: () => upstreamStats(lc, config),\n cvr: () => cvrStats(lc, config),\n changeLog: () => changeLogStats(lc, config),\n replica: () => replicaStats(lc, config),\n replication: () => replicationStats(lc, config),\n os: () => osStats(),\n };\n\n async function computeStats(group: string): Promise<[string, StatsObject]> {\n try {\n return [group, await statsFns[group]()];\n } catch (e) {\n lc.error?.(`error computing ${group} stats`, e);\n return [group, {error: String(e)}];\n }\n }\n\n const query = req.query as Record<string, unknown>;\n const groups =\n typeof query.group === 'string'\n ? query.group.split(',')\n : Array.isArray(query.group)\n ? query.group\n : undefined;\n\n const stats = await Promise.all(\n groups\n ? groups.filter(g => g in statsFns).map(computeStats)\n : Object.keys(statsFns).map(computeStats),\n );\n\n if (query.format === 'json') {\n const indent = query.pretty !== undefined ? 2 : undefined;\n await res\n .header('Content-Type', 'application/json')\n .send(BigIntJSON.stringify(Object.fromEntries(stats), null, indent));\n return;\n } else {\n const body = stats\n .map(([group, statsObject]) => printStats(group, statsObject))\n .join('');\n await res.header('Content-Type', 'text/plain; charset=utf-8').send(body);\n }\n}\n\nfunction first(x: object[]): object {\n return x[0];\n}\n\nfunction pick(x: object): unknown {\n return Object.values(x)[0];\n}\n\nfunction header(name: string): string {\n return `=== ${name} ===\\n`;\n}\n"],"mappings":";;;;;;;;;;;;AAeA,eAAe,cAAc,IAAgB,QAAoB;CAC/D,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC;CACjD,MAAM,MAAM,SAAS,IAAI,OAAO,SAAS,IAAI,iBAAiB;AAC9D,KAAI;AACF,SAAO,MAAM,WAAW;GACtB,CACE,eACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,aAC/C;GACD,CACE,2BACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,YAC/C;GACD,CACE,yBACA,GAAG,4CAA4C,IAAI,OAAO,CAAC,YAC5D;GACF,CAAC;WACM;AACR,QAAM,IAAI,KAAK;;;AAInB,eAAe,SAAS,IAAgB,QAAoB;CAC1D,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC,GAAG;CACpD,MAAM,MAAM,SAAS,IAAI,OAAO,IAAI,IAAI,YAAY;CAEpD,SAAS,yBACP,QACyC;EACzC,MAAM,SAAS,SACX,GAAG,sDACH,GAAG;AACP,SAAO,GAAG;;;;;aAKD,IAAI,OAAO,CAAC;QACjB,OAAO;;;;;;;;aAQF,IAAI,OAAO,CAAC;QACjB,OAAO;;;;;;;;;;;;AAab,KAAI;AACF,SAAO,MAAM,WAAW;GACtB,CACE,mBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,YAC/C;GACD,CACE,wBACA,GAAG,kDAAkD,IACnD,OACD,CAAC,YACH;GACD,CACE,oBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,gEAC/C;GACD,CACE,sBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,oFAC/C;GACD,CACE,qBACA,GAAG,+BAA+B,IAAI,OAAO,CAAC,mCAC/C;GACD,CACE,2BACA,GAAG;;;;;eAKI,IAAI,OAAO,CAAC;;;;;;;;;;;;;;;;;iCAkBpB;GACD,CACE,iCACA,GAAG;;;;;eAKI,IAAI,OAAO,CAAC;;;;;;;;;;;;iCAapB;GACD,CAEE,YACA,GAAG;;;;;;;;;;;aAWE,IAAI,OAAO,CAAC,aAClB;GACD,CAEE,kBACA,GAAG;aACE,IAAI,OAAO,CAAC;;gBAGlB;GACD,CACE,6CACA,yBAAyB,KAAK,CAC/B;GACD,CACE,+CACA,yBAAyB,MAAM,CAChC;GACD,CACE,2BACA,GAAG,gDAAgD,IACjD,OACD,CAAC,oDACH;GACD,CACE,mBACA,GAAG;;;aAGE,IAAI,OAAO,CAAC;;;gCAIlB;GACF,CAGG;WACI;AACR,QAAM,IAAI,KAAK;;;AAInB,eAAe,eAAe,IAAgB,QAAoB;CAChE,MAAM,SAAS,eAAe,WAAW,OAAO,CAAC,GAAG;CACpD,MAAM,MAAM,SAAS,IAAI,OAAO,OAAO,IAAI,eAAe;AAE1D,KAAI;AACF,SAAO,MAAM,WAAW,CACtB,CACE,iBACA,GAAG,6CAA6C,IAAI,OAAO,CAAC,cAC7D,CACF,CAAC;WACM;AACR,QAAM,IAAI,KAAK;;;AAInB,SAAS,aAAa,IAAgB,QAAoB;CACxD,MAAM,KAAK,IAAI,SAAS,IAAI,OAAO,QAAQ,KAAK;AAChD,KAAI;AACF,SAAO,OAAO,YAAY;GACxB,CAAC,kBAAkB,KAAK,MAAM,GAAG,OAAO,iBAAiB,CAAC,CAAC,CAAC;GAC5D,CAAC,cAAc,KAAK,MAAM,GAAG,OAAO,aAAa,CAAC,CAAC,CAAC;GACpD,CAAC,aAAa,KAAK,MAAM,GAAG,OAAO,YAAY,CAAC,CAAC,CAAC;GAClD,CAAC,gBAAgB,KAAK,MAAM,GAAG,OAAO,eAAe,CAAC,CAAC,CAAC;GACxD,CAAC,eAAe,KAAK,MAAM,GAAG,OAAO,cAAc,CAAC,CAAC,CAAC;GACtD,CAAC,cAAc,KAAK,MAAM,GAAG,OAAO,aAAa,CAAC,CAAC,CAAC;GACpD,CAAC,eAAe,KAAK,MAAM,GAAG,OAAO,cAAc,CAAC,CAAC,CAAC;GACtD,CAAC,kBAAkB,KAAK,MAAM,GAAG,OAAO,iBAAiB,CAAC,CAAC,CAAC;GAC5D,CAAC,sBAAsB,KAAK,MAAM,GAAG,OAAO,qBAAqB,CAAC,CAAC,CAAC;GACpE,CAAC,iBAAiB,GAAG,SAAS,OAAO,QAAQ,KAAK,CAAC;GACpD,CAAU;WACH;AACR,KAAG,OAAO;;;AAId,SAAS,iBAAiB,IAAgB,QAAoB;CAC5D,MAAM,KAAK,IAAI,SAAS,IAAI,OAAO,QAAQ,KAAK;AAChD,KAAI;AACF,SAAO,oBAAoB,GAAG;WACtB;AACR,KAAG,OAAO;;;AAId,SAAS,oBAAoB,IAAc;CACzC,MAAM,EAAC,iBAAgB,oBAAoB,IAAI,gBAAgB,GAAG,CAAC;AAEnE,QAAO,EAAC,KADI,uBAAuB,aAAa,EACpC;;AAGd,SAAS,UAAU;AACjB,QAAO,OAAO,YAAY;EACxB,CAAC,YAAY,GAAG,SAAS,CAAC;EAC1B,CAAC,UAAU,GAAG,QAAQ,CAAC;EACvB,CAAC,aAAa,GAAG,UAAU,CAAC;EAC5B,CAAC,YAAY,GAAG,SAAS,CAAC;EAC1B,CAAC,QAAQ,GAAG,MAAM,CAAC,OAAO;EAC1B,CAAC,yBAAyB,GAAG,sBAAsB,CAAC;EACpD,CAAC,YAAY,GAAG,UAAU,CAAC;EAC3B,CAAC,QAAQ,GAAG,MAAM,CAAC;EACnB,CAAC,WAAW,GAAG,SAAS,CAAC;EAC1B,CAAU;;AAGb,eAAe,WACb,gBAIA;CACA,MAAM,UAAU,MAAM,QAAQ,IAC5B,eAAe,IAAI,OAAO,CAAC,MAAM,WAAW,CAAC,MAAM,MAAM,MAAM,CAAU,CAC1E;AACD,QAAO,OAAO,YAAY,QAAQ;;AAKpC,SAAS,WAAW,OAAe,aAAkC;CACnE,MAAM,QAAkB,CAAC,OAAO,OAAO,MAAM,CAAC;AAC9C,MAAK,MAAM,CAAC,MAAM,WAAW,OAAO,QAAQ,YAAY,CACtD,OAAM,KAAK,OAAO,OAAO,OAAO,WAAW,UAAU,QAAQ,MAAM,EAAE,CAAC;AAExE,OAAM,KAAK,KAAK;AAChB,QAAO,MAAM,KAAK,GAAG;;;;;;;;AASvB,eAAsB,mBACpB,IACA,QACA,KACA,KACA;AAEA,KAAI,CAAC,qBAAqB,IAAI,QADV,KAAK,IAAI,EACsB,KAAK,EAAE;AACnD,MACF,KAAK,IAAI,CACT,OAAO,oBAAoB,uCAAqC,CAChE,KAAK,eAAe;AACvB;;CAGF,MAAM,WAAqE;EACzE,gBAAgB,cAAc,IAAI,OAAO;EACzC,WAAW,SAAS,IAAI,OAAO;EAC/B,iBAAiB,eAAe,IAAI,OAAO;EAC3C,eAAe,aAAa,IAAI,OAAO;EACvC,mBAAmB,iBAAiB,IAAI,OAAO;EAC/C,UAAU,SAAS;EACpB;CAED,eAAe,aAAa,OAA+C;AACzE,MAAI;AACF,UAAO,CAAC,OAAO,MAAM,SAAS,QAAQ,CAAC;WAChC,GAAG;AACV,MAAG,QAAQ,mBAAmB,MAAM,SAAS,EAAE;AAC/C,UAAO,CAAC,OAAO,EAAC,OAAO,OAAO,EAAE,EAAC,CAAC;;;CAItC,MAAM,QAAQ,IAAI;CAClB,MAAM,SACJ,OAAO,MAAM,UAAU,WACnB,MAAM,MAAM,MAAM,IAAI,GACtB,MAAM,QAAQ,MAAM,MAAM,GACxB,MAAM,QACN,KAAA;CAER,MAAM,QAAQ,MAAM,QAAQ,IAC1B,SACI,OAAO,QAAO,MAAK,KAAK,SAAS,CAAC,IAAI,aAAa,GACnD,OAAO,KAAK,SAAS,CAAC,IAAI,aAAa,CAC5C;AAED,KAAI,MAAM,WAAW,QAAQ;EAC3B,MAAM,SAAS,MAAM,WAAW,KAAA,IAAY,IAAI,KAAA;AAChD,QAAM,IACH,OAAO,gBAAgB,mBAAmB,CAC1C,KAAK,WAAW,UAAU,OAAO,YAAY,MAAM,EAAE,MAAM,OAAO,CAAC;AACtE;QACK;EACL,MAAM,OAAO,MACV,KAAK,CAAC,OAAO,iBAAiB,WAAW,OAAO,YAAY,CAAC,CAC7D,KAAK,GAAG;AACX,QAAM,IAAI,OAAO,gBAAgB,4BAA4B,CAAC,KAAK,KAAK;;;AAI5E,SAAS,MAAM,GAAqB;AAClC,QAAO,EAAE;;AAGX,SAAS,KAAK,GAAoB;AAChC,QAAO,OAAO,OAAO,EAAE,CAAC;;AAG1B,SAAS,OAAO,MAAsB;AACpC,QAAO,OAAO,KAAK"}
@@ -7,7 +7,7 @@ import { ProtocolError } from "../../../../zero-protocol/src/error.js";
7
7
  import { primaryKeyValueRecordSchema } from "../../../../zero-protocol/src/primary-key.js";
8
8
  import { mutationResultSchema } from "../../../../zero-protocol/src/push.js";
9
9
  import { upstreamSchema } from "../../types/shards.js";
10
- import { getOrCreateCounter, getOrCreateHistogram } from "../../observability/metrics.js";
10
+ import { getOrCreateCounter, getOrCreateLatencyHistogram } from "../../observability/metrics.js";
11
11
  import { cmpVersions, cookieToVersion, versionToCookie, versionToNullableCookie } from "./schema/types.js";
12
12
  import { getLogLevel, wrapWithProtocolError } from "../../types/error-with-level.js";
13
13
  //#region ../zero-cache/src/services/view-syncer/client-handler.ts
@@ -44,10 +44,7 @@ var ClientHandler = class {
44
44
  #lc;
45
45
  #downstream;
46
46
  #baseVersion;
47
- #pokeTime = getOrCreateHistogram("sync", "poke.time", {
48
- description: "Time elapsed for each poke transaction. Canceled / noop pokes are excluded.",
49
- unit: "s"
50
- });
47
+ #pokeTime = getOrCreateLatencyHistogram("sync", "poke.time", "Time elapsed for each poke transaction. Canceled / noop pokes are excluded.");
51
48
  #pokeTransactions = getOrCreateCounter("sync", "poke.transactions", "Count of poke transactions.");
52
49
  #pokedRows = getOrCreateCounter("sync", "poke.rows", "Count of poked rows.");
53
50
  constructor(lc, clientGroupID, clientID, wsID, shard, baseCookie, downstream) {
@@ -191,7 +188,7 @@ var ClientHandler = class {
191
188
  this.#baseVersion = finalVersion;
192
189
  const elapsed = performance.now() - start;
193
190
  this.#pokeTransactions.add(1);
194
- this.#pokeTime.record(elapsed / 1e3);
191
+ this.#pokeTime.recordMs(elapsed);
195
192
  }
196
193
  };
197
194
  }
@@ -1 +1 @@
1
- {"version":3,"file":"client-handler.js","names":["#clientGroupID","#zeroClientsTable","#zeroMutationsTable","#lc","#downstream","#pokeTime","#pokeTransactions","#pokedRows","#baseVersion","#push","#updateLMIDs"],"sources":["../../../../../../zero-cache/src/services/view-syncer/client-handler.ts"],"sourcesContent":["import type {LogContext} from '@rocicorp/logger';\nimport {assert, unreachable} from '../../../../shared/src/asserts.ts';\nimport type {JSONObject} from '../../../../shared/src/bigint-json.ts';\nimport {\n assertJSONValue,\n type JSONObject as SafeJSONObject,\n} from '../../../../shared/src/json.ts';\nimport {promiseVoid} from '../../../../shared/src/resolved-promises.ts';\nimport * as v from '../../../../shared/src/valita.ts';\nimport type {Writable} from '../../../../shared/src/writable.ts';\nimport type {ErroredQuery} from '../../../../zero-protocol/src/custom-queries.ts';\nimport {rowSchema} from '../../../../zero-protocol/src/data.ts';\nimport type {DeleteClientsBody} from '../../../../zero-protocol/src/delete-clients.ts';\nimport type {Downstream} from '../../../../zero-protocol/src/down.ts';\nimport {\n ProtocolError,\n type TransformFailedBody,\n} from '../../../../zero-protocol/src/error.ts';\nimport type {InspectDownBody} from '../../../../zero-protocol/src/inspect-down.ts';\nimport type {\n PokePartBody,\n PokeStartBody,\n} from '../../../../zero-protocol/src/poke.ts';\nimport {primaryKeyValueRecordSchema} from '../../../../zero-protocol/src/primary-key.ts';\nimport {mutationResultSchema} from '../../../../zero-protocol/src/push.ts';\nimport type {RowPatchOp} from '../../../../zero-protocol/src/row-patch.ts';\nimport {\n getOrCreateCounter,\n getOrCreateHistogram,\n} from '../../observability/metrics.ts';\nimport {\n getLogLevel,\n wrapWithProtocolError,\n} from '../../types/error-with-level.ts';\nimport {upstreamSchema, type ShardID} from '../../types/shards.ts';\nimport type {Subscription} from '../../types/subscription.ts';\nimport {\n cmpVersions,\n cookieToVersion,\n versionToCookie,\n versionToNullableCookie,\n type CVRVersion,\n type DelQueryPatch,\n type NullableCVRVersion,\n type PutQueryPatch,\n type RowID,\n} from './schema/types.ts';\n\nexport type PutRowPatch = {\n type: 'row';\n op: 'put';\n id: RowID;\n contents: JSONObject;\n};\n\nexport type DeleteRowPatch = {\n type: 'row';\n op: 'del';\n id: RowID;\n};\n\nexport type RowPatch = PutRowPatch | DeleteRowPatch;\nexport type ConfigPatch = DelQueryPatch | PutQueryPatch;\n\nexport type Patch = ConfigPatch | RowPatch;\n\nexport type PatchToVersion = {\n patch: Patch;\n toVersion: CVRVersion;\n};\n\nexport interface PokeHandler {\n addPatch(patch: PatchToVersion): Promise<void>;\n cancel(): Promise<void>;\n end(finalVersion: CVRVersion): Promise<void>;\n}\n\nconst NOOP: PokeHandler = {\n addPatch: () => promiseVoid,\n cancel: () => promiseVoid,\n end: () => promiseVoid,\n};\n\n/** Wraps PokeHandlers for multiple clients in a single PokeHandler. */\nexport function startPoke(\n clients: ClientHandler[],\n tentativeVersion: CVRVersion,\n): PokeHandler {\n const pokers = clients.map(c => c.startPoke(tentativeVersion));\n\n // Promise.allSettled() ensures that a failed (e.g. disconnected) client\n // does not prevent other clients from receiving the pokes. However, the\n // rate (per client group) will be limited by the slowest connection.\n return {\n addPatch: async patch => {\n await Promise.allSettled(pokers.map(poker => poker.addPatch(patch)));\n },\n cancel: async () => {\n await Promise.allSettled(pokers.map(poker => poker.cancel()));\n },\n end: async finalVersion => {\n await Promise.allSettled(pokers.map(poker => poker.end(finalVersion)));\n },\n };\n}\n\n// Semi-arbitrary threshold at which poke body parts are flushed.\n// When row size is being computed, that should be used as a threshold instead.\nconst PART_COUNT_FLUSH_THRESHOLD = 100;\n\n/**\n * Handles a single `ViewSyncer` connection.\n */\nexport class ClientHandler {\n readonly #clientGroupID: string;\n readonly clientID: string;\n readonly wsID: string;\n readonly #zeroClientsTable: string;\n readonly #zeroMutationsTable: string;\n readonly #lc: LogContext;\n readonly #downstream: Subscription<Downstream>;\n #baseVersion: NullableCVRVersion;\n\n readonly #pokeTime = getOrCreateHistogram('sync', 'poke.time', {\n description:\n 'Time elapsed for each poke transaction. Canceled / noop pokes are excluded.',\n unit: 's',\n });\n\n readonly #pokeTransactions = getOrCreateCounter(\n 'sync',\n 'poke.transactions',\n 'Count of poke transactions.',\n );\n\n readonly #pokedRows = getOrCreateCounter(\n 'sync',\n 'poke.rows',\n 'Count of poked rows.',\n );\n\n constructor(\n lc: LogContext,\n clientGroupID: string,\n clientID: string,\n wsID: string,\n shard: ShardID,\n baseCookie: string | null,\n downstream: Subscription<Downstream>,\n ) {\n lc.debug?.('new client handler');\n this.#clientGroupID = clientGroupID;\n this.clientID = clientID;\n this.wsID = wsID;\n this.#zeroClientsTable = `${upstreamSchema(shard)}.clients`;\n this.#zeroMutationsTable = `${upstreamSchema(shard)}.mutations`;\n this.#lc = lc;\n this.#downstream = downstream;\n this.#baseVersion = cookieToVersion(baseCookie);\n }\n\n version(): NullableCVRVersion {\n return this.#baseVersion;\n }\n\n async #push(msg: Downstream): Promise<void> {\n const {result} = this.#downstream.push(msg);\n await result;\n }\n\n fail(e: unknown) {\n this.#lc[getLogLevel(e)]?.(\n `view-syncer closing connection with error: ${String(e)}`,\n e,\n );\n this.#downstream.fail(wrapWithProtocolError(e));\n }\n\n close(reason: string) {\n this.#lc.debug?.(`view-syncer closing connection: ${reason}`);\n this.#downstream.cancel();\n }\n\n startPoke(tentativeVersion: CVRVersion): PokeHandler {\n const pokeID = versionToCookie(tentativeVersion);\n const lc = this.#lc.withContext('pokeID', pokeID);\n\n if (cmpVersions(this.#baseVersion, tentativeVersion) >= 0) {\n lc.info?.(`already caught up, not sending poke.`);\n return NOOP;\n }\n\n const baseCookie = versionToNullableCookie(this.#baseVersion);\n const cookie = versionToCookie(tentativeVersion);\n lc.info?.(`starting poke from ${baseCookie} to ${cookie}`);\n\n const start = performance.now();\n\n const pokeStart: PokeStartBody = {pokeID, baseCookie};\n\n let pokeStarted = false;\n let body: PokePartBody | undefined;\n let partCount = 0;\n const ensureBody = async () => {\n if (!pokeStarted) {\n await this.#push(['pokeStart', pokeStart]);\n pokeStarted = true;\n }\n return (body ??= {pokeID});\n };\n const flushBody = async () => {\n if (body) {\n await this.#push(['pokePart', body]);\n body = undefined;\n partCount = 0;\n }\n };\n\n const addPatch = async (patchToVersion: PatchToVersion) => {\n const {patch, toVersion} = patchToVersion;\n if (cmpVersions(toVersion, this.#baseVersion) <= 0) {\n return;\n }\n const body = await ensureBody();\n\n const {type, op} = patch;\n switch (type) {\n case 'query': {\n const patches = patch.clientID\n ? ((body.desiredQueriesPatches ??= {})[patch.clientID] ??= [])\n : (body.gotQueriesPatch ??= []);\n if (op === 'put') {\n patches.push({op, hash: patch.id});\n } else {\n patches.push({op, hash: patch.id});\n }\n break;\n }\n case 'row':\n if (patch.id.table === this.#zeroClientsTable) {\n this.#updateLMIDs((body.lastMutationIDChanges ??= {}), patch);\n } else if (patch.id.table === this.#zeroMutationsTable) {\n const patches = (body.mutationsPatch ??= []);\n if (op === 'put') {\n const row = v.parse(\n ensureSafeJSON(patch.contents),\n mutationRowSchema,\n 'passthrough',\n );\n patches.push({\n op: 'put',\n mutation: {\n id: {\n clientID: row.clientID,\n id: row.mutationID,\n },\n result: row.result,\n },\n });\n } else {\n const {clientID, mutationID} = patch.id.rowKey;\n assert(\n typeof clientID === 'string',\n 'client id must be a string',\n );\n const id = Number(mutationID);\n assert(\n !Number.isNaN(id) && Number.isFinite(id) && id >= 0,\n 'mutation id must be a finite number',\n );\n patches.push({\n op: 'del',\n id: {\n clientID,\n id,\n },\n });\n }\n } else {\n (body.rowsPatch ??= []).push(makeRowPatch(patch));\n }\n break;\n default:\n unreachable(patch);\n }\n\n if (++partCount >= PART_COUNT_FLUSH_THRESHOLD) {\n await flushBody();\n }\n };\n\n return {\n addPatch: async (patchToVersion: PatchToVersion) => {\n try {\n await addPatch(patchToVersion);\n if (patchToVersion.patch.type === 'row') {\n this.#pokedRows.add(1);\n }\n } catch (e) {\n this.#downstream.fail(wrapWithProtocolError(e));\n }\n },\n\n cancel: async () => {\n if (pokeStarted) {\n await this.#push(['pokeEnd', {pokeID, cookie: '', cancel: true}]);\n }\n },\n\n end: async (finalVersion: CVRVersion) => {\n const cookie = versionToCookie(finalVersion);\n if (!pokeStarted) {\n if (cmpVersions(this.#baseVersion, finalVersion) === 0) {\n return; // Nothing changed and nothing was sent.\n }\n await this.#push(['pokeStart', pokeStart]);\n } else if (cmpVersions(this.#baseVersion, finalVersion) >= 0) {\n // Sanity check: If the poke was started, the finalVersion\n // must be > #baseVersion.\n throw new Error(\n `Patches were sent but finalVersion ${finalVersion} is ` +\n `not greater than baseVersion ${this.#baseVersion}`,\n );\n }\n await flushBody();\n await this.#push(['pokeEnd', {pokeID, cookie}]);\n this.#baseVersion = finalVersion;\n\n const elapsed = performance.now() - start;\n this.#pokeTransactions.add(1);\n this.#pokeTime.record(elapsed / 1000);\n },\n };\n }\n\n async sendDeleteClients(\n lc: LogContext,\n deletedClientIDs: string[],\n deletedClientGroupIDs: string[],\n ) {\n const deleteClientsBody: Writable<DeleteClientsBody> = {};\n if (deletedClientIDs.length > 0) {\n deleteClientsBody.clientIDs = deletedClientIDs;\n }\n if (deletedClientGroupIDs.length > 0) {\n deleteClientsBody.clientGroupIDs = deletedClientGroupIDs;\n }\n lc.debug?.('sending deleteClients', deleteClientsBody);\n await this.#push(['deleteClients', deleteClientsBody]);\n }\n\n sendQueryTransformApplicationErrors(errors: ErroredQuery[]) {\n void this.#push(['transformError', errors]);\n }\n\n sendQueryTransformFailedError(error: TransformFailedBody) {\n this.fail(new ProtocolError(error));\n }\n\n sendInspectResponse(lc: LogContext, response: InspectDownBody): void {\n lc.debug?.('sending inspect response', response);\n this.#downstream.push(['inspect', response]);\n }\n\n #updateLMIDs(lmids: Record<string, number>, patch: RowPatch) {\n if (patch.op === 'put') {\n const row = ensureSafeJSON(patch.contents);\n const {clientGroupID, clientID, lastMutationID} = v.parse(\n row,\n lmidRowSchema,\n 'passthrough',\n );\n if (clientGroupID !== this.#clientGroupID) {\n this.#lc.error?.(\n `Received clients row for wrong clientGroupID. Ignoring.`,\n clientGroupID,\n );\n } else {\n lmids[clientID] = lastMutationID;\n }\n } else {\n // The 'constrain' and 'del' ops for clients can be ignored.\n patch.op satisfies 'constrain' | 'del';\n }\n }\n}\n\n// Note: The {APP_ID}_{SHARD_ID}.clients table is set up in replicator/initial-sync.ts.\nconst lmidRowSchema = v.object({\n clientGroupID: v.string(),\n clientID: v.string(),\n lastMutationID: v.number(), // Actually returned as a bigint, but converted by ensureSafeJSON().\n});\n\nconst mutationRowSchema = v.object({\n clientGroupID: v.string(),\n clientID: v.string(),\n mutationID: v.number(),\n result: mutationResultSchema,\n});\n\nfunction makeRowPatch(patch: RowPatch): RowPatchOp {\n const {\n op,\n id: {table: tableName, rowKey: id},\n } = patch;\n\n switch (op) {\n case 'put':\n return {\n op: 'put',\n tableName,\n value: v.parse(ensureSafeJSON(patch.contents), rowSchema),\n };\n\n case 'del':\n return {\n op,\n tableName,\n id: v.parse(id, primaryKeyValueRecordSchema),\n };\n\n default:\n unreachable(op);\n }\n}\n\n/**\n * Column values of type INT8 are returned as the `bigint` from the\n * Postgres library. These are converted to `number` if they are within\n * the safe Number range, allowing the protocol to support numbers larger\n * than 32-bits. Values outside of the safe number range (e.g. > 2^53) will\n * result in an Error.\n */\nexport function ensureSafeJSON(row: JSONObject): SafeJSONObject {\n const modified = Object.entries(row)\n .filter(([k, v]) => {\n if (typeof v === 'bigint') {\n if (v >= Number.MIN_SAFE_INTEGER && v <= Number.MAX_SAFE_INTEGER) {\n return true; // send this entry onto the next map() step.\n }\n throw new Error(`Value of \"${k}\" exceeds safe Number range (${v})`);\n } else if (typeof v === 'object') {\n assertJSONValue(v);\n }\n return false;\n })\n .map(([k, v]) => [k, Number(v)]);\n\n return modified.length\n ? {...row, ...Object.fromEntries(modified)}\n : (row as SafeJSONObject);\n}\n"],"mappings":";;;;;;;;;;;;;AA6EA,IAAM,OAAoB;CACxB,gBAAgB;CAChB,cAAc;CACd,WAAW;CACZ;;AAGD,SAAgB,UACd,SACA,kBACa;CACb,MAAM,SAAS,QAAQ,KAAI,MAAK,EAAE,UAAU,iBAAiB,CAAC;AAK9D,QAAO;EACL,UAAU,OAAM,UAAS;AACvB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,SAAS,MAAM,CAAC,CAAC;;EAEtE,QAAQ,YAAY;AAClB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,QAAQ,CAAC,CAAC;;EAE/D,KAAK,OAAM,iBAAgB;AACzB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,IAAI,aAAa,CAAC,CAAC;;EAEzE;;AAKH,IAAM,6BAA6B;;;;AAKnC,IAAa,gBAAb,MAA2B;CACzB;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CAEA,YAAqB,qBAAqB,QAAQ,aAAa;EAC7D,aACE;EACF,MAAM;EACP,CAAC;CAEF,oBAA6B,mBAC3B,QACA,qBACA,8BACD;CAED,aAAsB,mBACpB,QACA,aACA,uBACD;CAED,YACE,IACA,eACA,UACA,MACA,OACA,YACA,YACA;AACA,KAAG,QAAQ,qBAAqB;AAChC,QAAA,gBAAsB;AACtB,OAAK,WAAW;AAChB,OAAK,OAAO;AACZ,QAAA,mBAAyB,GAAG,eAAe,MAAM,CAAC;AAClD,QAAA,qBAA2B,GAAG,eAAe,MAAM,CAAC;AACpD,QAAA,KAAW;AACX,QAAA,aAAmB;AACnB,QAAA,cAAoB,gBAAgB,WAAW;;CAGjD,UAA8B;AAC5B,SAAO,MAAA;;CAGT,OAAA,KAAY,KAAgC;EAC1C,MAAM,EAAC,WAAU,MAAA,WAAiB,KAAK,IAAI;AAC3C,QAAM;;CAGR,KAAK,GAAY;AACf,QAAA,GAAS,YAAY,EAAE,IACrB,8CAA8C,OAAO,EAAE,IACvD,EACD;AACD,QAAA,WAAiB,KAAK,sBAAsB,EAAE,CAAC;;CAGjD,MAAM,QAAgB;AACpB,QAAA,GAAS,QAAQ,mCAAmC,SAAS;AAC7D,QAAA,WAAiB,QAAQ;;CAG3B,UAAU,kBAA2C;EACnD,MAAM,SAAS,gBAAgB,iBAAiB;EAChD,MAAM,KAAK,MAAA,GAAS,YAAY,UAAU,OAAO;AAEjD,MAAI,YAAY,MAAA,aAAmB,iBAAiB,IAAI,GAAG;AACzD,MAAG,OAAO,uCAAuC;AACjD,UAAO;;EAGT,MAAM,aAAa,wBAAwB,MAAA,YAAkB;EAC7D,MAAM,SAAS,gBAAgB,iBAAiB;AAChD,KAAG,OAAO,sBAAsB,WAAW,MAAM,SAAS;EAE1D,MAAM,QAAQ,YAAY,KAAK;EAE/B,MAAM,YAA2B;GAAC;GAAQ;GAAW;EAErD,IAAI,cAAc;EAClB,IAAI;EACJ,IAAI,YAAY;EAChB,MAAM,aAAa,YAAY;AAC7B,OAAI,CAAC,aAAa;AAChB,UAAM,MAAA,KAAW,CAAC,aAAa,UAAU,CAAC;AAC1C,kBAAc;;AAEhB,UAAQ,SAAS,EAAC,QAAO;;EAE3B,MAAM,YAAY,YAAY;AAC5B,OAAI,MAAM;AACR,UAAM,MAAA,KAAW,CAAC,YAAY,KAAK,CAAC;AACpC,WAAO,KAAA;AACP,gBAAY;;;EAIhB,MAAM,WAAW,OAAO,mBAAmC;GACzD,MAAM,EAAC,OAAO,cAAa;AAC3B,OAAI,YAAY,WAAW,MAAA,YAAkB,IAAI,EAC/C;GAEF,MAAM,OAAO,MAAM,YAAY;GAE/B,MAAM,EAAC,MAAM,OAAM;AACnB,WAAQ,MAAR;IACE,KAAK,SAAS;KACZ,MAAM,UAAU,MAAM,WACjB,CAAC,KAAK,0BAA0B,EAAE,EAAE,MAAM,cAAc,EAAE,GAC1D,KAAK,oBAAoB,EAAE;AAChC,SAAI,OAAO,MACT,SAAQ,KAAK;MAAC;MAAI,MAAM,MAAM;MAAG,CAAC;SAElC,SAAQ,KAAK;MAAC;MAAI,MAAM,MAAM;MAAG,CAAC;AAEpC;;IAEF,KAAK;AACH,SAAI,MAAM,GAAG,UAAU,MAAA,iBACrB,OAAA,YAAmB,KAAK,0BAA0B,EAAE,EAAG,MAAM;cACpD,MAAM,GAAG,UAAU,MAAA,oBAA0B;MACtD,MAAM,UAAW,KAAK,mBAAmB,EAAE;AAC3C,UAAI,OAAO,OAAO;OAChB,MAAM,MAAM,MACV,eAAe,MAAM,SAAS,EAC9B,mBACA,cACD;AACD,eAAQ,KAAK;QACX,IAAI;QACJ,UAAU;SACR,IAAI;UACF,UAAU,IAAI;UACd,IAAI,IAAI;UACT;SACD,QAAQ,IAAI;SACb;QACF,CAAC;aACG;OACL,MAAM,EAAC,UAAU,eAAc,MAAM,GAAG;AACxC,cACE,OAAO,aAAa,UACpB,6BACD;OACD,MAAM,KAAK,OAAO,WAAW;AAC7B,cACE,CAAC,OAAO,MAAM,GAAG,IAAI,OAAO,SAAS,GAAG,IAAI,MAAM,GAClD,sCACD;AACD,eAAQ,KAAK;QACX,IAAI;QACJ,IAAI;SACF;SACA;SACD;QACF,CAAC;;WAGJ,EAAC,KAAK,cAAc,EAAE,EAAE,KAAK,aAAa,MAAM,CAAC;AAEnD;IACF,QACE,aAAY,MAAM;;AAGtB,OAAI,EAAE,aAAa,2BACjB,OAAM,WAAW;;AAIrB,SAAO;GACL,UAAU,OAAO,mBAAmC;AAClD,QAAI;AACF,WAAM,SAAS,eAAe;AAC9B,SAAI,eAAe,MAAM,SAAS,MAChC,OAAA,UAAgB,IAAI,EAAE;aAEjB,GAAG;AACV,WAAA,WAAiB,KAAK,sBAAsB,EAAE,CAAC;;;GAInD,QAAQ,YAAY;AAClB,QAAI,YACF,OAAM,MAAA,KAAW,CAAC,WAAW;KAAC;KAAQ,QAAQ;KAAI,QAAQ;KAAK,CAAC,CAAC;;GAIrE,KAAK,OAAO,iBAA6B;IACvC,MAAM,SAAS,gBAAgB,aAAa;AAC5C,QAAI,CAAC,aAAa;AAChB,SAAI,YAAY,MAAA,aAAmB,aAAa,KAAK,EACnD;AAEF,WAAM,MAAA,KAAW,CAAC,aAAa,UAAU,CAAC;eACjC,YAAY,MAAA,aAAmB,aAAa,IAAI,EAGzD,OAAM,IAAI,MACR,sCAAsC,aAAa,mCACjB,MAAA,cACnC;AAEH,UAAM,WAAW;AACjB,UAAM,MAAA,KAAW,CAAC,WAAW;KAAC;KAAQ;KAAO,CAAC,CAAC;AAC/C,UAAA,cAAoB;IAEpB,MAAM,UAAU,YAAY,KAAK,GAAG;AACpC,UAAA,iBAAuB,IAAI,EAAE;AAC7B,UAAA,SAAe,OAAO,UAAU,IAAK;;GAExC;;CAGH,MAAM,kBACJ,IACA,kBACA,uBACA;EACA,MAAM,oBAAiD,EAAE;AACzD,MAAI,iBAAiB,SAAS,EAC5B,mBAAkB,YAAY;AAEhC,MAAI,sBAAsB,SAAS,EACjC,mBAAkB,iBAAiB;AAErC,KAAG,QAAQ,yBAAyB,kBAAkB;AACtD,QAAM,MAAA,KAAW,CAAC,iBAAiB,kBAAkB,CAAC;;CAGxD,oCAAoC,QAAwB;AACrD,QAAA,KAAW,CAAC,kBAAkB,OAAO,CAAC;;CAG7C,8BAA8B,OAA4B;AACxD,OAAK,KAAK,IAAI,cAAc,MAAM,CAAC;;CAGrC,oBAAoB,IAAgB,UAAiC;AACnE,KAAG,QAAQ,4BAA4B,SAAS;AAChD,QAAA,WAAiB,KAAK,CAAC,WAAW,SAAS,CAAC;;CAG9C,aAAa,OAA+B,OAAiB;AAC3D,MAAI,MAAM,OAAO,OAAO;GAEtB,MAAM,EAAC,eAAe,UAAU,mBAAkB,MADtC,eAAe,MAAM,SAAS,EAGxC,eACA,cACD;AACD,OAAI,kBAAkB,MAAA,cACpB,OAAA,GAAS,QACP,2DACA,cACD;OAED,OAAM,YAAY;QAIpB,OAAM;;;AAMZ,IAAM,gBAAgB,eAAE,OAAO;CAC7B,eAAe,eAAE,QAAQ;CACzB,UAAU,eAAE,QAAQ;CACpB,gBAAgB,eAAE,QAAQ;CAC3B,CAAC;AAEF,IAAM,oBAAoB,eAAE,OAAO;CACjC,eAAe,eAAE,QAAQ;CACzB,UAAU,eAAE,QAAQ;CACpB,YAAY,eAAE,QAAQ;CACtB,QAAQ;CACT,CAAC;AAEF,SAAS,aAAa,OAA6B;CACjD,MAAM,EACJ,IACA,IAAI,EAAC,OAAO,WAAW,QAAQ,SAC7B;AAEJ,SAAQ,IAAR;EACE,KAAK,MACH,QAAO;GACL,IAAI;GACJ;GACA,OAAO,MAAQ,eAAe,MAAM,SAAS,EAAE,UAAU;GAC1D;EAEH,KAAK,MACH,QAAO;GACL;GACA;GACA,IAAI,MAAQ,IAAI,4BAA4B;GAC7C;EAEH,QACE,aAAY,GAAG;;;;;;;;;;AAWrB,SAAgB,eAAe,KAAiC;CAC9D,MAAM,WAAW,OAAO,QAAQ,IAAI,CACjC,QAAQ,CAAC,GAAG,OAAO;AAClB,MAAI,OAAO,MAAM,UAAU;AACzB,OAAI,KAAK,OAAO,oBAAoB,KAAK,OAAO,iBAC9C,QAAO;AAET,SAAM,IAAI,MAAM,aAAa,EAAE,+BAA+B,EAAE,GAAG;aAC1D,OAAO,MAAM,SACtB,iBAAgB,EAAE;AAEpB,SAAO;GACP,CACD,KAAK,CAAC,GAAG,OAAO,CAAC,GAAG,OAAO,EAAE,CAAC,CAAC;AAElC,QAAO,SAAS,SACZ;EAAC,GAAG;EAAK,GAAG,OAAO,YAAY,SAAS;EAAC,GACxC"}
1
+ {"version":3,"file":"client-handler.js","names":["#clientGroupID","#zeroClientsTable","#zeroMutationsTable","#lc","#downstream","#pokeTime","#pokeTransactions","#pokedRows","#baseVersion","#push","#updateLMIDs"],"sources":["../../../../../../zero-cache/src/services/view-syncer/client-handler.ts"],"sourcesContent":["import type {LogContext} from '@rocicorp/logger';\nimport {assert, unreachable} from '../../../../shared/src/asserts.ts';\nimport type {JSONObject} from '../../../../shared/src/bigint-json.ts';\nimport {\n assertJSONValue,\n type JSONObject as SafeJSONObject,\n} from '../../../../shared/src/json.ts';\nimport {promiseVoid} from '../../../../shared/src/resolved-promises.ts';\nimport * as v from '../../../../shared/src/valita.ts';\nimport type {Writable} from '../../../../shared/src/writable.ts';\nimport type {ErroredQuery} from '../../../../zero-protocol/src/custom-queries.ts';\nimport {rowSchema} from '../../../../zero-protocol/src/data.ts';\nimport type {DeleteClientsBody} from '../../../../zero-protocol/src/delete-clients.ts';\nimport type {Downstream} from '../../../../zero-protocol/src/down.ts';\nimport {\n ProtocolError,\n type TransformFailedBody,\n} from '../../../../zero-protocol/src/error.ts';\nimport type {InspectDownBody} from '../../../../zero-protocol/src/inspect-down.ts';\nimport type {\n PokePartBody,\n PokeStartBody,\n} from '../../../../zero-protocol/src/poke.ts';\nimport {primaryKeyValueRecordSchema} from '../../../../zero-protocol/src/primary-key.ts';\nimport {mutationResultSchema} from '../../../../zero-protocol/src/push.ts';\nimport type {RowPatchOp} from '../../../../zero-protocol/src/row-patch.ts';\nimport {\n getOrCreateCounter,\n getOrCreateLatencyHistogram,\n} from '../../observability/metrics.ts';\nimport {\n getLogLevel,\n wrapWithProtocolError,\n} from '../../types/error-with-level.ts';\nimport {upstreamSchema, type ShardID} from '../../types/shards.ts';\nimport type {Subscription} from '../../types/subscription.ts';\nimport {\n cmpVersions,\n cookieToVersion,\n versionToCookie,\n versionToNullableCookie,\n type CVRVersion,\n type DelQueryPatch,\n type NullableCVRVersion,\n type PutQueryPatch,\n type RowID,\n} from './schema/types.ts';\n\nexport type PutRowPatch = {\n type: 'row';\n op: 'put';\n id: RowID;\n contents: JSONObject;\n};\n\nexport type DeleteRowPatch = {\n type: 'row';\n op: 'del';\n id: RowID;\n};\n\nexport type RowPatch = PutRowPatch | DeleteRowPatch;\nexport type ConfigPatch = DelQueryPatch | PutQueryPatch;\n\nexport type Patch = ConfigPatch | RowPatch;\n\nexport type PatchToVersion = {\n patch: Patch;\n toVersion: CVRVersion;\n};\n\nexport interface PokeHandler {\n addPatch(patch: PatchToVersion): Promise<void>;\n cancel(): Promise<void>;\n end(finalVersion: CVRVersion): Promise<void>;\n}\n\nconst NOOP: PokeHandler = {\n addPatch: () => promiseVoid,\n cancel: () => promiseVoid,\n end: () => promiseVoid,\n};\n\n/** Wraps PokeHandlers for multiple clients in a single PokeHandler. */\nexport function startPoke(\n clients: ClientHandler[],\n tentativeVersion: CVRVersion,\n): PokeHandler {\n const pokers = clients.map(c => c.startPoke(tentativeVersion));\n\n // Promise.allSettled() ensures that a failed (e.g. disconnected) client\n // does not prevent other clients from receiving the pokes. However, the\n // rate (per client group) will be limited by the slowest connection.\n return {\n addPatch: async patch => {\n await Promise.allSettled(pokers.map(poker => poker.addPatch(patch)));\n },\n cancel: async () => {\n await Promise.allSettled(pokers.map(poker => poker.cancel()));\n },\n end: async finalVersion => {\n await Promise.allSettled(pokers.map(poker => poker.end(finalVersion)));\n },\n };\n}\n\n// Semi-arbitrary threshold at which poke body parts are flushed.\n// When row size is being computed, that should be used as a threshold instead.\nconst PART_COUNT_FLUSH_THRESHOLD = 100;\n\n/**\n * Handles a single `ViewSyncer` connection.\n */\nexport class ClientHandler {\n readonly #clientGroupID: string;\n readonly clientID: string;\n readonly wsID: string;\n readonly #zeroClientsTable: string;\n readonly #zeroMutationsTable: string;\n readonly #lc: LogContext;\n readonly #downstream: Subscription<Downstream>;\n #baseVersion: NullableCVRVersion;\n\n readonly #pokeTime = getOrCreateLatencyHistogram(\n 'sync',\n 'poke.time',\n 'Time elapsed for each poke transaction. Canceled / noop pokes are excluded.',\n );\n\n readonly #pokeTransactions = getOrCreateCounter(\n 'sync',\n 'poke.transactions',\n 'Count of poke transactions.',\n );\n\n readonly #pokedRows = getOrCreateCounter(\n 'sync',\n 'poke.rows',\n 'Count of poked rows.',\n );\n\n constructor(\n lc: LogContext,\n clientGroupID: string,\n clientID: string,\n wsID: string,\n shard: ShardID,\n baseCookie: string | null,\n downstream: Subscription<Downstream>,\n ) {\n lc.debug?.('new client handler');\n this.#clientGroupID = clientGroupID;\n this.clientID = clientID;\n this.wsID = wsID;\n this.#zeroClientsTable = `${upstreamSchema(shard)}.clients`;\n this.#zeroMutationsTable = `${upstreamSchema(shard)}.mutations`;\n this.#lc = lc;\n this.#downstream = downstream;\n this.#baseVersion = cookieToVersion(baseCookie);\n }\n\n version(): NullableCVRVersion {\n return this.#baseVersion;\n }\n\n async #push(msg: Downstream): Promise<void> {\n const {result} = this.#downstream.push(msg);\n await result;\n }\n\n fail(e: unknown) {\n this.#lc[getLogLevel(e)]?.(\n `view-syncer closing connection with error: ${String(e)}`,\n e,\n );\n this.#downstream.fail(wrapWithProtocolError(e));\n }\n\n close(reason: string) {\n this.#lc.debug?.(`view-syncer closing connection: ${reason}`);\n this.#downstream.cancel();\n }\n\n startPoke(tentativeVersion: CVRVersion): PokeHandler {\n const pokeID = versionToCookie(tentativeVersion);\n const lc = this.#lc.withContext('pokeID', pokeID);\n\n if (cmpVersions(this.#baseVersion, tentativeVersion) >= 0) {\n lc.info?.(`already caught up, not sending poke.`);\n return NOOP;\n }\n\n const baseCookie = versionToNullableCookie(this.#baseVersion);\n const cookie = versionToCookie(tentativeVersion);\n lc.info?.(`starting poke from ${baseCookie} to ${cookie}`);\n\n const start = performance.now();\n\n const pokeStart: PokeStartBody = {pokeID, baseCookie};\n\n let pokeStarted = false;\n let body: PokePartBody | undefined;\n let partCount = 0;\n const ensureBody = async () => {\n if (!pokeStarted) {\n await this.#push(['pokeStart', pokeStart]);\n pokeStarted = true;\n }\n return (body ??= {pokeID});\n };\n const flushBody = async () => {\n if (body) {\n await this.#push(['pokePart', body]);\n body = undefined;\n partCount = 0;\n }\n };\n\n const addPatch = async (patchToVersion: PatchToVersion) => {\n const {patch, toVersion} = patchToVersion;\n if (cmpVersions(toVersion, this.#baseVersion) <= 0) {\n return;\n }\n const body = await ensureBody();\n\n const {type, op} = patch;\n switch (type) {\n case 'query': {\n const patches = patch.clientID\n ? ((body.desiredQueriesPatches ??= {})[patch.clientID] ??= [])\n : (body.gotQueriesPatch ??= []);\n if (op === 'put') {\n patches.push({op, hash: patch.id});\n } else {\n patches.push({op, hash: patch.id});\n }\n break;\n }\n case 'row':\n if (patch.id.table === this.#zeroClientsTable) {\n this.#updateLMIDs((body.lastMutationIDChanges ??= {}), patch);\n } else if (patch.id.table === this.#zeroMutationsTable) {\n const patches = (body.mutationsPatch ??= []);\n if (op === 'put') {\n const row = v.parse(\n ensureSafeJSON(patch.contents),\n mutationRowSchema,\n 'passthrough',\n );\n patches.push({\n op: 'put',\n mutation: {\n id: {\n clientID: row.clientID,\n id: row.mutationID,\n },\n result: row.result,\n },\n });\n } else {\n const {clientID, mutationID} = patch.id.rowKey;\n assert(\n typeof clientID === 'string',\n 'client id must be a string',\n );\n const id = Number(mutationID);\n assert(\n !Number.isNaN(id) && Number.isFinite(id) && id >= 0,\n 'mutation id must be a finite number',\n );\n patches.push({\n op: 'del',\n id: {\n clientID,\n id,\n },\n });\n }\n } else {\n (body.rowsPatch ??= []).push(makeRowPatch(patch));\n }\n break;\n default:\n unreachable(patch);\n }\n\n if (++partCount >= PART_COUNT_FLUSH_THRESHOLD) {\n await flushBody();\n }\n };\n\n return {\n addPatch: async (patchToVersion: PatchToVersion) => {\n try {\n await addPatch(patchToVersion);\n if (patchToVersion.patch.type === 'row') {\n this.#pokedRows.add(1);\n }\n } catch (e) {\n this.#downstream.fail(wrapWithProtocolError(e));\n }\n },\n\n cancel: async () => {\n if (pokeStarted) {\n await this.#push(['pokeEnd', {pokeID, cookie: '', cancel: true}]);\n }\n },\n\n end: async (finalVersion: CVRVersion) => {\n const cookie = versionToCookie(finalVersion);\n if (!pokeStarted) {\n if (cmpVersions(this.#baseVersion, finalVersion) === 0) {\n return; // Nothing changed and nothing was sent.\n }\n await this.#push(['pokeStart', pokeStart]);\n } else if (cmpVersions(this.#baseVersion, finalVersion) >= 0) {\n // Sanity check: If the poke was started, the finalVersion\n // must be > #baseVersion.\n throw new Error(\n `Patches were sent but finalVersion ${finalVersion} is ` +\n `not greater than baseVersion ${this.#baseVersion}`,\n );\n }\n await flushBody();\n await this.#push(['pokeEnd', {pokeID, cookie}]);\n this.#baseVersion = finalVersion;\n\n const elapsed = performance.now() - start;\n this.#pokeTransactions.add(1);\n this.#pokeTime.recordMs(elapsed);\n },\n };\n }\n\n async sendDeleteClients(\n lc: LogContext,\n deletedClientIDs: string[],\n deletedClientGroupIDs: string[],\n ) {\n const deleteClientsBody: Writable<DeleteClientsBody> = {};\n if (deletedClientIDs.length > 0) {\n deleteClientsBody.clientIDs = deletedClientIDs;\n }\n if (deletedClientGroupIDs.length > 0) {\n deleteClientsBody.clientGroupIDs = deletedClientGroupIDs;\n }\n lc.debug?.('sending deleteClients', deleteClientsBody);\n await this.#push(['deleteClients', deleteClientsBody]);\n }\n\n sendQueryTransformApplicationErrors(errors: ErroredQuery[]) {\n void this.#push(['transformError', errors]);\n }\n\n sendQueryTransformFailedError(error: TransformFailedBody) {\n this.fail(new ProtocolError(error));\n }\n\n sendInspectResponse(lc: LogContext, response: InspectDownBody): void {\n lc.debug?.('sending inspect response', response);\n this.#downstream.push(['inspect', response]);\n }\n\n #updateLMIDs(lmids: Record<string, number>, patch: RowPatch) {\n if (patch.op === 'put') {\n const row = ensureSafeJSON(patch.contents);\n const {clientGroupID, clientID, lastMutationID} = v.parse(\n row,\n lmidRowSchema,\n 'passthrough',\n );\n if (clientGroupID !== this.#clientGroupID) {\n this.#lc.error?.(\n `Received clients row for wrong clientGroupID. Ignoring.`,\n clientGroupID,\n );\n } else {\n lmids[clientID] = lastMutationID;\n }\n } else {\n // The 'constrain' and 'del' ops for clients can be ignored.\n patch.op satisfies 'constrain' | 'del';\n }\n }\n}\n\n// Note: The {APP_ID}_{SHARD_ID}.clients table is set up in replicator/initial-sync.ts.\nconst lmidRowSchema = v.object({\n clientGroupID: v.string(),\n clientID: v.string(),\n lastMutationID: v.number(), // Actually returned as a bigint, but converted by ensureSafeJSON().\n});\n\nconst mutationRowSchema = v.object({\n clientGroupID: v.string(),\n clientID: v.string(),\n mutationID: v.number(),\n result: mutationResultSchema,\n});\n\nfunction makeRowPatch(patch: RowPatch): RowPatchOp {\n const {\n op,\n id: {table: tableName, rowKey: id},\n } = patch;\n\n switch (op) {\n case 'put':\n return {\n op: 'put',\n tableName,\n value: v.parse(ensureSafeJSON(patch.contents), rowSchema),\n };\n\n case 'del':\n return {\n op,\n tableName,\n id: v.parse(id, primaryKeyValueRecordSchema),\n };\n\n default:\n unreachable(op);\n }\n}\n\n/**\n * Column values of type INT8 are returned as the `bigint` from the\n * Postgres library. These are converted to `number` if they are within\n * the safe Number range, allowing the protocol to support numbers larger\n * than 32-bits. Values outside of the safe number range (e.g. > 2^53) will\n * result in an Error.\n */\nexport function ensureSafeJSON(row: JSONObject): SafeJSONObject {\n const modified = Object.entries(row)\n .filter(([k, v]) => {\n if (typeof v === 'bigint') {\n if (v >= Number.MIN_SAFE_INTEGER && v <= Number.MAX_SAFE_INTEGER) {\n return true; // send this entry onto the next map() step.\n }\n throw new Error(`Value of \"${k}\" exceeds safe Number range (${v})`);\n } else if (typeof v === 'object') {\n assertJSONValue(v);\n }\n return false;\n })\n .map(([k, v]) => [k, Number(v)]);\n\n return modified.length\n ? {...row, ...Object.fromEntries(modified)}\n : (row as SafeJSONObject);\n}\n"],"mappings":";;;;;;;;;;;;;AA6EA,IAAM,OAAoB;CACxB,gBAAgB;CAChB,cAAc;CACd,WAAW;CACZ;;AAGD,SAAgB,UACd,SACA,kBACa;CACb,MAAM,SAAS,QAAQ,KAAI,MAAK,EAAE,UAAU,iBAAiB,CAAC;AAK9D,QAAO;EACL,UAAU,OAAM,UAAS;AACvB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,SAAS,MAAM,CAAC,CAAC;;EAEtE,QAAQ,YAAY;AAClB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,QAAQ,CAAC,CAAC;;EAE/D,KAAK,OAAM,iBAAgB;AACzB,SAAM,QAAQ,WAAW,OAAO,KAAI,UAAS,MAAM,IAAI,aAAa,CAAC,CAAC;;EAEzE;;AAKH,IAAM,6BAA6B;;;;AAKnC,IAAa,gBAAb,MAA2B;CACzB;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CAEA,YAAqB,4BACnB,QACA,aACA,8EACD;CAED,oBAA6B,mBAC3B,QACA,qBACA,8BACD;CAED,aAAsB,mBACpB,QACA,aACA,uBACD;CAED,YACE,IACA,eACA,UACA,MACA,OACA,YACA,YACA;AACA,KAAG,QAAQ,qBAAqB;AAChC,QAAA,gBAAsB;AACtB,OAAK,WAAW;AAChB,OAAK,OAAO;AACZ,QAAA,mBAAyB,GAAG,eAAe,MAAM,CAAC;AAClD,QAAA,qBAA2B,GAAG,eAAe,MAAM,CAAC;AACpD,QAAA,KAAW;AACX,QAAA,aAAmB;AACnB,QAAA,cAAoB,gBAAgB,WAAW;;CAGjD,UAA8B;AAC5B,SAAO,MAAA;;CAGT,OAAA,KAAY,KAAgC;EAC1C,MAAM,EAAC,WAAU,MAAA,WAAiB,KAAK,IAAI;AAC3C,QAAM;;CAGR,KAAK,GAAY;AACf,QAAA,GAAS,YAAY,EAAE,IACrB,8CAA8C,OAAO,EAAE,IACvD,EACD;AACD,QAAA,WAAiB,KAAK,sBAAsB,EAAE,CAAC;;CAGjD,MAAM,QAAgB;AACpB,QAAA,GAAS,QAAQ,mCAAmC,SAAS;AAC7D,QAAA,WAAiB,QAAQ;;CAG3B,UAAU,kBAA2C;EACnD,MAAM,SAAS,gBAAgB,iBAAiB;EAChD,MAAM,KAAK,MAAA,GAAS,YAAY,UAAU,OAAO;AAEjD,MAAI,YAAY,MAAA,aAAmB,iBAAiB,IAAI,GAAG;AACzD,MAAG,OAAO,uCAAuC;AACjD,UAAO;;EAGT,MAAM,aAAa,wBAAwB,MAAA,YAAkB;EAC7D,MAAM,SAAS,gBAAgB,iBAAiB;AAChD,KAAG,OAAO,sBAAsB,WAAW,MAAM,SAAS;EAE1D,MAAM,QAAQ,YAAY,KAAK;EAE/B,MAAM,YAA2B;GAAC;GAAQ;GAAW;EAErD,IAAI,cAAc;EAClB,IAAI;EACJ,IAAI,YAAY;EAChB,MAAM,aAAa,YAAY;AAC7B,OAAI,CAAC,aAAa;AAChB,UAAM,MAAA,KAAW,CAAC,aAAa,UAAU,CAAC;AAC1C,kBAAc;;AAEhB,UAAQ,SAAS,EAAC,QAAO;;EAE3B,MAAM,YAAY,YAAY;AAC5B,OAAI,MAAM;AACR,UAAM,MAAA,KAAW,CAAC,YAAY,KAAK,CAAC;AACpC,WAAO,KAAA;AACP,gBAAY;;;EAIhB,MAAM,WAAW,OAAO,mBAAmC;GACzD,MAAM,EAAC,OAAO,cAAa;AAC3B,OAAI,YAAY,WAAW,MAAA,YAAkB,IAAI,EAC/C;GAEF,MAAM,OAAO,MAAM,YAAY;GAE/B,MAAM,EAAC,MAAM,OAAM;AACnB,WAAQ,MAAR;IACE,KAAK,SAAS;KACZ,MAAM,UAAU,MAAM,WACjB,CAAC,KAAK,0BAA0B,EAAE,EAAE,MAAM,cAAc,EAAE,GAC1D,KAAK,oBAAoB,EAAE;AAChC,SAAI,OAAO,MACT,SAAQ,KAAK;MAAC;MAAI,MAAM,MAAM;MAAG,CAAC;SAElC,SAAQ,KAAK;MAAC;MAAI,MAAM,MAAM;MAAG,CAAC;AAEpC;;IAEF,KAAK;AACH,SAAI,MAAM,GAAG,UAAU,MAAA,iBACrB,OAAA,YAAmB,KAAK,0BAA0B,EAAE,EAAG,MAAM;cACpD,MAAM,GAAG,UAAU,MAAA,oBAA0B;MACtD,MAAM,UAAW,KAAK,mBAAmB,EAAE;AAC3C,UAAI,OAAO,OAAO;OAChB,MAAM,MAAM,MACV,eAAe,MAAM,SAAS,EAC9B,mBACA,cACD;AACD,eAAQ,KAAK;QACX,IAAI;QACJ,UAAU;SACR,IAAI;UACF,UAAU,IAAI;UACd,IAAI,IAAI;UACT;SACD,QAAQ,IAAI;SACb;QACF,CAAC;aACG;OACL,MAAM,EAAC,UAAU,eAAc,MAAM,GAAG;AACxC,cACE,OAAO,aAAa,UACpB,6BACD;OACD,MAAM,KAAK,OAAO,WAAW;AAC7B,cACE,CAAC,OAAO,MAAM,GAAG,IAAI,OAAO,SAAS,GAAG,IAAI,MAAM,GAClD,sCACD;AACD,eAAQ,KAAK;QACX,IAAI;QACJ,IAAI;SACF;SACA;SACD;QACF,CAAC;;WAGJ,EAAC,KAAK,cAAc,EAAE,EAAE,KAAK,aAAa,MAAM,CAAC;AAEnD;IACF,QACE,aAAY,MAAM;;AAGtB,OAAI,EAAE,aAAa,2BACjB,OAAM,WAAW;;AAIrB,SAAO;GACL,UAAU,OAAO,mBAAmC;AAClD,QAAI;AACF,WAAM,SAAS,eAAe;AAC9B,SAAI,eAAe,MAAM,SAAS,MAChC,OAAA,UAAgB,IAAI,EAAE;aAEjB,GAAG;AACV,WAAA,WAAiB,KAAK,sBAAsB,EAAE,CAAC;;;GAInD,QAAQ,YAAY;AAClB,QAAI,YACF,OAAM,MAAA,KAAW,CAAC,WAAW;KAAC;KAAQ,QAAQ;KAAI,QAAQ;KAAK,CAAC,CAAC;;GAIrE,KAAK,OAAO,iBAA6B;IACvC,MAAM,SAAS,gBAAgB,aAAa;AAC5C,QAAI,CAAC,aAAa;AAChB,SAAI,YAAY,MAAA,aAAmB,aAAa,KAAK,EACnD;AAEF,WAAM,MAAA,KAAW,CAAC,aAAa,UAAU,CAAC;eACjC,YAAY,MAAA,aAAmB,aAAa,IAAI,EAGzD,OAAM,IAAI,MACR,sCAAsC,aAAa,mCACjB,MAAA,cACnC;AAEH,UAAM,WAAW;AACjB,UAAM,MAAA,KAAW,CAAC,WAAW;KAAC;KAAQ;KAAO,CAAC,CAAC;AAC/C,UAAA,cAAoB;IAEpB,MAAM,UAAU,YAAY,KAAK,GAAG;AACpC,UAAA,iBAAuB,IAAI,EAAE;AAC7B,UAAA,SAAe,SAAS,QAAQ;;GAEnC;;CAGH,MAAM,kBACJ,IACA,kBACA,uBACA;EACA,MAAM,oBAAiD,EAAE;AACzD,MAAI,iBAAiB,SAAS,EAC5B,mBAAkB,YAAY;AAEhC,MAAI,sBAAsB,SAAS,EACjC,mBAAkB,iBAAiB;AAErC,KAAG,QAAQ,yBAAyB,kBAAkB;AACtD,QAAM,MAAA,KAAW,CAAC,iBAAiB,kBAAkB,CAAC;;CAGxD,oCAAoC,QAAwB;AACrD,QAAA,KAAW,CAAC,kBAAkB,OAAO,CAAC;;CAG7C,8BAA8B,OAA4B;AACxD,OAAK,KAAK,IAAI,cAAc,MAAM,CAAC;;CAGrC,oBAAoB,IAAgB,UAAiC;AACnE,KAAG,QAAQ,4BAA4B,SAAS;AAChD,QAAA,WAAiB,KAAK,CAAC,WAAW,SAAS,CAAC;;CAG9C,aAAa,OAA+B,OAAiB;AAC3D,MAAI,MAAM,OAAO,OAAO;GAEtB,MAAM,EAAC,eAAe,UAAU,mBAAkB,MADtC,eAAe,MAAM,SAAS,EAGxC,eACA,cACD;AACD,OAAI,kBAAkB,MAAA,cACpB,OAAA,GAAS,QACP,2DACA,cACD;OAED,OAAM,YAAY;QAIpB,OAAM;;;AAMZ,IAAM,gBAAgB,eAAE,OAAO;CAC7B,eAAe,eAAE,QAAQ;CACzB,UAAU,eAAE,QAAQ;CACpB,gBAAgB,eAAE,QAAQ;CAC3B,CAAC;AAEF,IAAM,oBAAoB,eAAE,OAAO;CACjC,eAAe,eAAE,QAAQ;CACzB,UAAU,eAAE,QAAQ;CACpB,YAAY,eAAE,QAAQ;CACtB,QAAQ;CACT,CAAC;AAEF,SAAS,aAAa,OAA6B;CACjD,MAAM,EACJ,IACA,IAAI,EAAC,OAAO,WAAW,QAAQ,SAC7B;AAEJ,SAAQ,IAAR;EACE,KAAK,MACH,QAAO;GACL,IAAI;GACJ;GACA,OAAO,MAAQ,eAAe,MAAM,SAAS,EAAE,UAAU;GAC1D;EAEH,KAAK,MACH,QAAO;GACL;GACA;GACA,IAAI,MAAQ,IAAI,4BAA4B;GAC7C;EAEH,QACE,aAAY,GAAG;;;;;;;;;;AAWrB,SAAgB,eAAe,KAAiC;CAC9D,MAAM,WAAW,OAAO,QAAQ,IAAI,CACjC,QAAQ,CAAC,GAAG,OAAO;AAClB,MAAI,OAAO,MAAM,UAAU;AACzB,OAAI,KAAK,OAAO,oBAAoB,KAAK,OAAO,iBAC9C,QAAO;AAET,SAAM,IAAI,MAAM,aAAa,EAAE,+BAA+B,EAAE,GAAG;aAC1D,OAAO,MAAM,SACtB,iBAAgB,EAAE;AAEpB,SAAO;GACP,CACD,KAAK,CAAC,GAAG,OAAO,CAAC,GAAG,OAAO,EAAE,CAAC,CAAC;AAElC,QAAO,SAAS,SACZ;EAAC,GAAG;EAAK,GAAG,OAAO,YAAY,SAAS;EAAC,GACxC"}
@@ -56,6 +56,7 @@ export declare class CVRStore {
56
56
  markQueryAsDeleted(version: CVRVersion, queryPatch: QueryPatch): void;
57
57
  putQuery(query: QueryRecord): void;
58
58
  updateQuery(query: QueryRecord): void;
59
+ updateRowSetSignature(queryHash: string, signature: string): void;
59
60
  insertClient(client: ClientRecord): void;
60
61
  deleteClient(clientID: string): void;
61
62
  putDesiredQuery(newVersion: CVRVersion, query: {
@@ -1 +1 @@
1
- {"version":3,"file":"cvr-store.d.ts","sourceRoot":"","sources":["../../../../../../zero-cache/src/services/view-syncer/cvr-store.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAC,UAAU,EAAC,MAAM,kBAAkB,CAAC;AAiBjD,OAAO,KAAK,EAAC,eAAe,EAAC,MAAM,+CAA+C,CAAC;AAMnF,OAAO,EAAC,sBAAsB,EAAC,MAAM,iCAAiC,CAAC;AACvE,OAAO,KAAK,EAAC,UAAU,EAAE,mBAAmB,EAAC,MAAM,mBAAmB,CAAC;AAEvE,OAAO,EAAY,KAAK,OAAO,EAAC,MAAM,uBAAuB,CAAC;AAC9D,OAAO,KAAK,EAAQ,cAAc,EAAC,MAAM,qBAAqB,CAAC;AAC/D,OAAO,KAAK,EAAC,GAAG,EAAE,WAAW,EAAC,MAAM,UAAU,CAAC;AAE/C,OAAO,EAKL,KAAK,OAAO,EACb,MAAM,iBAAiB,CAAC;AACzB,OAAO,EAEL,KAAK,YAAY,EAGjB,KAAK,UAAU,EAGf,KAAK,kBAAkB,EACvB,KAAK,UAAU,EACf,KAAK,WAAW,EAEhB,KAAK,KAAK,EACV,KAAK,SAAS,EAGf,MAAM,mBAAmB,CAAC;AAC3B,OAAO,EACL,KAAK,QAAQ,EAGd,MAAM,gBAAgB,CAAC;AAExB,MAAM,MAAM,aAAa,GAAG;IAC1B,SAAS,EAAE,MAAM,CAAC;IAClB,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,YAAY,EAAE,MAAM,CAAC;IACrB,UAAU,EAAE,MAAM,CAAC;CACpB,CAAC;AA4FF,qBAAa,QAAQ;;gBAmCjB,EAAE,EAAE,UAAU,EACd,KAAK,EAAE,UAAU,EACjB,KAAK,EAAE,OAAO,EACd,MAAM,EAAE,MAAM,EACd,KAAK,EAAE,MAAM,EACb,WAAW,EAAE,CAAC,CAAC,EAAE,OAAO,KAAK,IAAI,EACjC,qBAAqB,SAA2B,EAChD,eAAe,SAAoB,EACnC,yBAAyB,SAAM,EAAE,qBAAqB;IACtD,YAAY,oBAAa;IA6B3B,IAAI,CAAC,EAAE,EAAE,UAAU,EAAE,eAAe,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,CAAC;IAyN3D,aAAa,IAAI,OAAO,CAAC,WAAW,CAAC,KAAK,EAAE,SAAS,CAAC,CAAC;IAIvD,YAAY,CAAC,GAAG,EAAE,SAAS,GAAG,IAAI;IAIlC;;;;;;;OAOG;IACH,YAAY,CAAC,EAAE,EAAE,KAAK,GAAG,IAAI;IAI7B;;;;OAIG;IACH,YAAY,CAAC,GAAG,GAAG,EAAE,KAAK,EAAE;IAM5B;;;;OAIG;IACG,cAAc,CAAC,QAAQ,EAAE,QAAQ,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC;IAO3E;;;;;OAKG;IACG,WAAW,IAAI,OAAO,CAAC,QAAQ,GAAG,SAAS,CAAC;IAelD,WAAW,CAAC,EACV,OAAO,EACP,cAAc,EACd,UAAU,EACV,YAAY,EACZ,SAAS,EACT,QAAQ,GACT,EAAE,IAAI,CACL,WAAW,EACT,SAAS,GACT,gBAAgB,GAChB,YAAY,GACZ,cAAc,GACd,WAAW,GACX,UAAU,CACb,GAAG,IAAI;IAqBR,kBAAkB,CAAC,OAAO,EAAE,UAAU,EAAE,UAAU,EAAE,UAAU,GAAG,IAAI;IASrE,QAAQ,CAAC,KAAK,EAAE,WAAW,GAAG,IAAI;IAelC,WAAW,CAAC,KAAK,EAAE,WAAW;IAc9B,YAAY,CAAC,MAAM,EAAE,YAAY,GAAG,IAAI;IAYxC,YAAY,CAAC,QAAQ,EAAE,MAAM;IAU7B,eAAe,CACb,UAAU,EAAE,UAAU,EACtB,KAAK,EAAE;QAAC,EAAE,EAAE,MAAM,CAAA;KAAC,EACnB,MAAM,EAAE;QAAC,EAAE,EAAE,MAAM,CAAA;KAAC,EACpB,OAAO,EAAE,OAAO,EAChB,aAAa,EAAE,QAAQ,GAAG,SAAS,EACnC,GAAG,EAAE,MAAM,GACV,IAAI;IAkBP,iBAAiB,CACf,EAAE,EAAE,UAAU,EACd,YAAY,EAAE,kBAAkB,EAChC,OAAO,EAAE,WAAW,EACpB,OAAO,EAAE,UAAU,EACnB,kBAAkB,GAAE,MAAM,EAAO,GAChC,cAAc,CAAC,OAAO,EAAE,EAAE,IAAI,EAAE,SAAS,CAAC;IAUvC,oBAAoB,CACxB,EAAE,EAAE,UAAU,EACd,YAAY,EAAE,kBAAkB,EAChC,OAAO,EAAE,WAAW,EACpB,OAAO,EAAE,UAAU,GAClB,OAAO,CAAC,cAAc,EAAE,CAAC;IA4d5B,IAAI,QAAQ,IAAI,MAAM,CAErB;IAEK,KAAK,CACT,EAAE,EAAE,UAAU,EACd,sBAAsB,EAAE,UAAU,EAClC,GAAG,EAAE,WAAW,EAChB,eAAe,EAAE,MAAM,GACtB,OAAO,CAAC,aAAa,GAAG,IAAI,CAAC;IAkChC,iBAAiB,IAAI,OAAO;IAI5B,qDAAqD;IACrD,OAAO,CAAC,EAAE,EAAE,UAAU,GAAG,OAAO,CAAC,IAAI,CAAC;IAIhC,cAAc,CAClB,EAAE,EAAE,UAAU,EACd,QAAQ,EAAE,QAAQ,EAClB,QAAQ,CAAC,EAAE,MAAM,GAChB,OAAO,CAAC,eAAe,EAAE,CAAC;CAsC9B;AAED;;;;GAIG;AACH,wBAAsB,YAAY,CAChC,EAAE,EAAE,mBAAmB,EACvB,MAAM,EAAE,MAAM,EACd,aAAa,EAAE,MAAM,EACrB,sBAAsB,EAAE,UAAU,GACjC,OAAO,CAAC,IAAI,CAAC,CAUf;AAED,qBAAa,mBAAoB,SAAQ,sBAAsB;gBACjD,OAAO,EAAE,MAAM;CAU5B;AAED,qBAAa,+BAAgC,SAAQ,sBAAsB;IACzE,QAAQ,CAAC,IAAI,qCAAqC;gBAEtC,eAAe,EAAE,MAAM,EAAE,aAAa,EAAE,MAAM;CAU3D;AAED,qBAAa,cAAe,SAAQ,sBAAsB;IACxD,QAAQ,CAAC,IAAI,oBAAoB;gBAG/B,KAAK,EAAE,MAAM,GAAG,IAAI,EACpB,SAAS,EAAE,MAAM,GAAG,IAAI,EACxB,eAAe,EAAE,MAAM;CAe1B;AAED,qBAAa,wBAAyB,SAAQ,sBAAsB;IAClE,QAAQ,CAAC,IAAI,8BAA8B;gBAE/B,KAAK,EAAE,OAAO;CAW3B;AAED,qBAAa,sBAAuB,SAAQ,KAAK;IAC/C,QAAQ,CAAC,IAAI,4BAA4B;IACzC,QAAQ,CAAC,UAAU,EAAE,MAAM,CAAC;IAC5B,QAAQ,CAAC,WAAW,EAAE,MAAM,GAAG,IAAI,CAAC;gBAExB,UAAU,EAAE,MAAM,EAAE,WAAW,EAAE,MAAM,GAAG,IAAI;CAK3D"}
1
+ {"version":3,"file":"cvr-store.d.ts","sourceRoot":"","sources":["../../../../../../zero-cache/src/services/view-syncer/cvr-store.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAC,UAAU,EAAC,MAAM,kBAAkB,CAAC;AAiBjD,OAAO,KAAK,EAAC,eAAe,EAAC,MAAM,+CAA+C,CAAC;AAMnF,OAAO,EAAC,sBAAsB,EAAC,MAAM,iCAAiC,CAAC;AACvE,OAAO,KAAK,EAAC,UAAU,EAAE,mBAAmB,EAAC,MAAM,mBAAmB,CAAC;AAEvE,OAAO,EAAY,KAAK,OAAO,EAAC,MAAM,uBAAuB,CAAC;AAC9D,OAAO,KAAK,EAAQ,cAAc,EAAC,MAAM,qBAAqB,CAAC;AAC/D,OAAO,KAAK,EAAC,GAAG,EAAE,WAAW,EAAC,MAAM,UAAU,CAAC;AAE/C,OAAO,EAKL,KAAK,OAAO,EACb,MAAM,iBAAiB,CAAC;AACzB,OAAO,EAEL,KAAK,YAAY,EAGjB,KAAK,UAAU,EAGf,KAAK,kBAAkB,EACvB,KAAK,UAAU,EACf,KAAK,WAAW,EAEhB,KAAK,KAAK,EACV,KAAK,SAAS,EAGf,MAAM,mBAAmB,CAAC;AAC3B,OAAO,EACL,KAAK,QAAQ,EAGd,MAAM,gBAAgB,CAAC;AAExB,MAAM,MAAM,aAAa,GAAG;IAC1B,SAAS,EAAE,MAAM,CAAC;IAClB,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,MAAM,CAAC;IAChB,OAAO,EAAE,MAAM,CAAC;IAChB,IAAI,EAAE,MAAM,CAAC;IACb,YAAY,EAAE,MAAM,CAAC;IACrB,UAAU,EAAE,MAAM,CAAC;CACpB,CAAC;AAqGF,qBAAa,QAAQ;;gBAmCjB,EAAE,EAAE,UAAU,EACd,KAAK,EAAE,UAAU,EACjB,KAAK,EAAE,OAAO,EACd,MAAM,EAAE,MAAM,EACd,KAAK,EAAE,MAAM,EACb,WAAW,EAAE,CAAC,CAAC,EAAE,OAAO,KAAK,IAAI,EACjC,qBAAqB,SAA2B,EAChD,eAAe,SAAoB,EACnC,yBAAyB,SAAM,EAAE,qBAAqB;IACtD,YAAY,oBAAa;IAmC3B,IAAI,CAAC,EAAE,EAAE,UAAU,EAAE,eAAe,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,CAAC;IAyN3D,aAAa,IAAI,OAAO,CAAC,WAAW,CAAC,KAAK,EAAE,SAAS,CAAC,CAAC;IAIvD,YAAY,CAAC,GAAG,EAAE,SAAS,GAAG,IAAI;IAIlC;;;;;;;OAOG;IACH,YAAY,CAAC,EAAE,EAAE,KAAK,GAAG,IAAI;IAI7B;;;;OAIG;IACH,YAAY,CAAC,GAAG,GAAG,EAAE,KAAK,EAAE;IAM5B;;;;OAIG;IACG,cAAc,CAAC,QAAQ,EAAE,QAAQ,EAAE,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC;IAO3E;;;;;OAKG;IACG,WAAW,IAAI,OAAO,CAAC,QAAQ,GAAG,SAAS,CAAC;IAelD,WAAW,CAAC,EACV,OAAO,EACP,cAAc,EACd,UAAU,EACV,YAAY,EACZ,SAAS,EACT,QAAQ,GACT,EAAE,IAAI,CACL,WAAW,EACT,SAAS,GACT,gBAAgB,GAChB,YAAY,GACZ,cAAc,GACd,WAAW,GACX,UAAU,CACb,GAAG,IAAI;IAqBR,kBAAkB,CAAC,OAAO,EAAE,UAAU,EAAE,UAAU,EAAE,UAAU,GAAG,IAAI;IASrE,QAAQ,CAAC,KAAK,EAAE,WAAW,GAAG,IAAI;IAelC,WAAW,CAAC,KAAK,EAAE,WAAW;IAc9B,qBAAqB,CAAC,SAAS,EAAE,MAAM,EAAE,SAAS,EAAE,MAAM,GAAG,IAAI;IAIjE,YAAY,CAAC,MAAM,EAAE,YAAY,GAAG,IAAI;IAYxC,YAAY,CAAC,QAAQ,EAAE,MAAM;IAU7B,eAAe,CACb,UAAU,EAAE,UAAU,EACtB,KAAK,EAAE;QAAC,EAAE,EAAE,MAAM,CAAA;KAAC,EACnB,MAAM,EAAE;QAAC,EAAE,EAAE,MAAM,CAAA;KAAC,EACpB,OAAO,EAAE,OAAO,EAChB,aAAa,EAAE,QAAQ,GAAG,SAAS,EACnC,GAAG,EAAE,MAAM,GACV,IAAI;IAkBP,iBAAiB,CACf,EAAE,EAAE,UAAU,EACd,YAAY,EAAE,kBAAkB,EAChC,OAAO,EAAE,WAAW,EACpB,OAAO,EAAE,UAAU,EACnB,kBAAkB,GAAE,MAAM,EAAO,GAChC,cAAc,CAAC,OAAO,EAAE,EAAE,IAAI,EAAE,SAAS,CAAC;IAUvC,oBAAoB,CACxB,EAAE,EAAE,UAAU,EACd,YAAY,EAAE,kBAAkB,EAChC,OAAO,EAAE,WAAW,EACpB,OAAO,EAAE,UAAU,GAClB,OAAO,CAAC,cAAc,EAAE,CAAC;IAwe5B,IAAI,QAAQ,IAAI,MAAM,CAErB;IAEK,KAAK,CACT,EAAE,EAAE,UAAU,EACd,sBAAsB,EAAE,UAAU,EAClC,GAAG,EAAE,WAAW,EAChB,eAAe,EAAE,MAAM,GACtB,OAAO,CAAC,aAAa,GAAG,IAAI,CAAC;IAkChC,iBAAiB,IAAI,OAAO;IAI5B,qDAAqD;IACrD,OAAO,CAAC,EAAE,EAAE,UAAU,GAAG,OAAO,CAAC,IAAI,CAAC;IAIhC,cAAc,CAClB,EAAE,EAAE,UAAU,EACd,QAAQ,EAAE,QAAQ,EAClB,QAAQ,CAAC,EAAE,MAAM,GAChB,OAAO,CAAC,eAAe,EAAE,CAAC;CAsC9B;AAED;;;;GAIG;AACH,wBAAsB,YAAY,CAChC,EAAE,EAAE,mBAAmB,EACvB,MAAM,EAAE,MAAM,EACd,aAAa,EAAE,MAAM,EACrB,sBAAsB,EAAE,UAAU,GACjC,OAAO,CAAC,IAAI,CAAC,CAUf;AAED,qBAAa,mBAAoB,SAAQ,sBAAsB;gBACjD,OAAO,EAAE,MAAM;CAU5B;AAED,qBAAa,+BAAgC,SAAQ,sBAAsB;IACzE,QAAQ,CAAC,IAAI,qCAAqC;gBAEtC,eAAe,EAAE,MAAM,EAAE,aAAa,EAAE,MAAM;CAU3D;AAED,qBAAa,cAAe,SAAQ,sBAAsB;IACxD,QAAQ,CAAC,IAAI,oBAAoB;gBAG/B,KAAK,EAAE,MAAM,GAAG,IAAI,EACpB,SAAS,EAAE,MAAM,GAAG,IAAI,EACxB,eAAe,EAAE,MAAM;CAe1B;AAED,qBAAa,wBAAyB,SAAQ,sBAAsB;IAClE,QAAQ,CAAC,IAAI,8BAA8B;gBAE/B,KAAK,EAAE,OAAO;CAW3B;AAED,qBAAa,sBAAuB,SAAQ,KAAK;IAC/C,QAAQ,CAAC,IAAI,4BAA4B;IACzC,QAAQ,CAAC,UAAU,EAAE,MAAM,CAAC;IAC5B,QAAQ,CAAC,WAAW,EAAE,MAAM,GAAG,IAAI,CAAC;gBAExB,UAAU,EAAE,MAAM,EAAE,WAAW,EAAE,MAAM,GAAG,IAAI;CAK3D"}