@event-driven-io/emmett-postgresql 0.25.0 → 0.26.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.js.map +1 -1
- package/package.json +3 -3
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/postgreSQLEventStore.ts","../../emmett/src/config/plugins/index.ts","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/validation/dates.ts","../../emmett/src/typing/command.ts","../../emmett/src/typing/event.ts","../../emmett/src/typing/workflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/eventStore/afterCommit/forwardToMessageBus.ts","../../emmett/src/eventStore/events/index.ts","../../emmett/src/eventStore/eventStore.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/eventStore/subscriptions/caughtUpTransformStream.ts","../../emmett/src/eventStore/subscriptions/streamingCoordinator.ts","../../emmett/src/streaming/transformations/notifyAboutNoActiveReaders.ts","../../emmett/src/streaming/writers/writeToStream.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/utils/iterators.ts","../../emmett/src/taskProcessing/taskProcessor.ts","../../emmett/src/utils/locking/index.ts","../../emmett/src/utils/retry.ts","../../emmett/src/commandHandling/handleCommand.ts","../../emmett/src/commandHandling/handleCommandWithDecider.ts","../../emmett/src/messageBus/index.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/projections/index.ts","../../emmett/src/streaming/binaryArrays.ts","../../emmett/src/streaming/collectors/collect.ts","../../emmett/src/streaming/decoders/binary.ts","../../emmett/src/streaming/decoders/string.ts","../../emmett/src/streaming/decoders/json.ts","../../emmett/src/streaming/decoders/object.ts","../../emmett/src/streaming/decoders/composite.ts","../../emmett/src/streaming/generators/fromArray.ts","../../emmett/src/streaming/generators/index.ts","../../emmett/src/streaming/restream.ts","../../emmett/src/streaming/transformations/filter.ts","../../emmett/src/streaming/transformations/map.ts","../../emmett/src/streaming/transformations/reduce.ts","../../emmett/src/streaming/transformations/retry.ts","../../emmett/src/streaming/transformations/skip.ts","../../emmett/src/streaming/transformations/stopAfter.ts","../../emmett/src/streaming/transformations/stopOn.ts","../../emmett/src/streaming/transformations/take.ts","../../emmett/src/streaming/transformations/waitAtMost.ts","../../emmett/src/streaming/transformations/index.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/testing/deciderSpecification.ts","../../emmett/src/testing/wrapEventStore.ts","../src/eventStore/projections/index.ts","../src/eventStore/projections/pongo/pongoProjectionSpec.ts","../src/eventStore/projections/pongo/projections.ts","../src/eventStore/projections/postgresProjectionSpec.ts","../src/eventStore/schema/index.ts","../src/eventStore/schema/appendToStream.ts","../src/eventStore/schema/typing.ts","../src/eventStore/schema/storeSubscriptionCheckpoint.ts","../src/eventStore/schema/tables.ts","../src/eventStore/schema/readLastMessageGlobalPosition.ts","../src/eventStore/schema/readMessagesBatch.ts","../src/eventStore/schema/readStream.ts","../src/eventStore/schema/readSubscriptionCheckpoint.ts","../src/eventStore/subscriptions/messageBatchProcessing/index.ts","../src/eventStore/subscriptions/postgreSQLEventStoreConsumer.ts","../src/eventStore/subscriptions/postgreSQLEventStoreSubscription.ts"],"sourcesContent":["import {\n dumbo,\n type MigrationStyle,\n type NodePostgresClientConnection,\n type NodePostgresConnector,\n type NodePostgresPool,\n type NodePostgresPoolClientConnection,\n} from '@event-driven-io/dumbo';\nimport {\n assertExpectedVersionMatchesCurrent,\n ExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResultWithGlobalPosition,\n type Event,\n type EventStore,\n type EventStoreSession,\n type EventStoreSessionFactory,\n type ProjectionRegistration,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport pg from 'pg';\nimport {\n handleProjections,\n type PostgreSQLProjectionHandlerContext,\n} from './projections';\nimport {\n appendToStream,\n createEventStoreSchema,\n readStream,\n schemaSQL,\n type AppendToStreamPreCommitHook,\n} from './schema';\n\nexport interface PostgresEventStore\n extends EventStore<PostgresReadEventMetadata>,\n EventStoreSessionFactory<PostgresEventStore> {\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition>;\n close(): Promise<void>;\n schema: {\n sql(): string;\n print(): void;\n migrate(): Promise<void>;\n };\n}\n\nexport type PostgresReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type PostgresReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n PostgresReadEventMetadata\n>;\n\ntype PostgresEventStorePooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n };\n\ntype PostgresEventStoreNotPooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n connection:\n | NodePostgresPoolClientConnection\n | NodePostgresClientConnection;\n pooled?: false;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n dumbo: NodePostgresPool;\n pooled?: false;\n };\n\nexport type PostgresEventStoreConnectionOptions =\n | PostgresEventStorePooledOptions\n | PostgresEventStoreNotPooledOptions;\n\nexport type PostgresEventStoreOptions = {\n projections?: ProjectionRegistration<\n 'inline',\n PostgresReadEventMetadata,\n PostgreSQLProjectionHandlerContext\n >[];\n schema?: { autoMigration?: MigrationStyle };\n connectionOptions?: PostgresEventStoreConnectionOptions;\n};\n\nexport const defaultPostgreSQLOptions: PostgresEventStoreOptions = {\n projections: [],\n schema: { autoMigration: 'CreateOrUpdate' },\n};\n\nexport const PostgreSQLEventStoreDefaultStreamVersion = 0n;\n\nexport const getPostgreSQLEventStore = (\n connectionString: string,\n options: PostgresEventStoreOptions = defaultPostgreSQLOptions,\n): PostgresEventStore => {\n const poolOptions = {\n connectionString,\n ...(options.connectionOptions ? options.connectionOptions : {}),\n };\n const pool = 'dumbo' in poolOptions ? poolOptions.dumbo : dumbo(poolOptions);\n let migrateSchema: Promise<void>;\n\n const autoGenerateSchema =\n options.schema?.autoMigration === undefined ||\n options.schema?.autoMigration !== 'None';\n\n const ensureSchemaExists = () => {\n if (!autoGenerateSchema) return Promise.resolve();\n\n if (!migrateSchema) {\n migrateSchema = createEventStoreSchema(pool);\n }\n return migrateSchema;\n };\n\n const inlineProjections = (options.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n const preCommitHook: AppendToStreamPreCommitHook | undefined =\n inlineProjections.length > 0\n ? (events, { transaction }) =>\n handleProjections({\n projections: inlineProjections,\n connection: {\n connectionString,\n transaction,\n },\n // TODO: Add proper handling of global data\n // Currently it's not available as append doesn't return array of global position but just the last one\n events: events as ReadEvent<Event, PostgresReadEventMetadata>[],\n })\n : undefined;\n\n return {\n schema: {\n sql: () => schemaSQL.join(''),\n print: () => console.log(schemaSQL.join('')),\n migrate: async () => {\n await (migrateSchema = createEventStoreSchema(pool));\n },\n },\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n PostgresReadEventMetadata\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const expectedStreamVersion = read?.expectedStreamVersion;\n\n let state = initialState();\n\n const result = await this.readStream<EventType>(streamName, options.read);\n const currentStreamVersion = result.currentStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n PostgreSQLEventStoreDefaultStreamVersion,\n );\n\n for (const event of result.events) {\n if (!event) continue;\n\n state = evolve(state, event);\n }\n\n return {\n currentStreamVersion: currentStreamVersion,\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: async <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions,\n ): Promise<ReadStreamResult<EventType, PostgresReadEventMetadata>> => {\n await ensureSchemaExists();\n return readStream<EventType>(pool.execute, streamName, options);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition> => {\n await ensureSchemaExists();\n // TODO: This has to be smarter when we introduce urn-based resolution\n const [firstPart, ...rest] = streamName.split('-');\n\n const streamType =\n firstPart && rest.length > 0 ? firstPart : 'emt:unknown';\n\n const appendResult = await appendToStream(\n pool,\n streamName,\n streamType,\n events,\n {\n ...options,\n preCommitHook,\n },\n );\n\n if (!appendResult.success)\n throw new ExpectedVersionConflictError<bigint>(\n -1n, //TODO: Return actual version in case of error\n options?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n );\n\n return {\n nextExpectedStreamVersion: appendResult.nextStreamPosition,\n lastEventGlobalPosition: appendResult.lastGlobalPosition,\n createdNewStream:\n appendResult.nextStreamPosition >= BigInt(events.length),\n };\n },\n close: () => pool.close(),\n\n async withSession<T = unknown>(\n callback: (session: EventStoreSession<PostgresEventStore>) => Promise<T>,\n ): Promise<T> {\n return await pool.withConnection(async (connection) => {\n const storeOptions: PostgresEventStoreOptions = {\n ...options,\n connectionOptions: {\n connection,\n },\n };\n\n const eventStore = getPostgreSQLEventStore(\n connectionString,\n storeOptions,\n );\n\n return callback({\n eventStore,\n close: () => Promise.resolve(),\n });\n });\n },\n };\n};\n","export type EmmettPluginConfig =\n | {\n name: string;\n register: EmmettPluginRegistration[];\n }\n | string;\n\nexport type EmmettPluginType = 'cli';\n\nexport type EmmettCliPluginRegistration = { pluginType: 'cli'; path?: string };\n\nexport type EmmettPluginRegistration = EmmettCliPluginRegistration;\n\nexport type EmmettCliCommand = {\n addCommand<CliCommand>(command: CliCommand): CliCommand;\n};\n\nexport type EmmettCliPlugin = {\n pluginType: 'cli';\n name: string;\n registerCommands: (program: EmmettCliCommand) => Promise<void> | void;\n};\n\nexport type EmmettPlugin = EmmettCliPlugin;\n\nexport const isPluginConfig = (\n plugin: Partial<EmmettPluginConfig> | string | undefined,\n): plugin is EmmettPluginConfig =>\n plugin !== undefined &&\n (typeof plugin === 'string' ||\n ('name' in plugin &&\n plugin.name !== undefined &&\n typeof plugin.name === 'string'));\n","import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : 500;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: 412,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 400,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 403,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: 404,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { ValidationError } from '../errors';\n\nexport const formatDateToUtcYYYYMMDD = (date: Date) => {\n // Use the 'en-CA' locale which formats as 'yyyy-mm-dd'\n const formatter = new Intl.DateTimeFormat('en-CA', {\n timeZone: 'UTC',\n year: 'numeric',\n month: '2-digit',\n day: '2-digit',\n });\n\n // Format the date\n return formatter.format(date);\n};\n\n// Function to validate 'yyyy-mm-dd' format\nexport const isValidYYYYMMDD = (dateString: string) => {\n const regex = /^\\d{4}-\\d{2}-\\d{2}$/;\n return regex.test(dateString);\n};\n\nexport const parseDateFromUtcYYYYMMDD = (dateString: string) => {\n const date = new Date(dateString + 'T00:00:00Z');\n\n if (!isValidYYYYMMDD(dateString)) {\n throw new ValidationError('Invalid date format, must be yyyy-mm-dd');\n }\n\n if (isNaN(date.getTime())) {\n throw new ValidationError('Invalid date format');\n }\n\n return date;\n};\n","import type { DefaultRecord, Flavour } from './';\n\nexport type Command<\n CommandType extends string = string,\n CommandData extends DefaultRecord = DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Flavour<\n Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: Readonly<CommandData>;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n >,\n 'Command'\n>;\n\nexport type CommandTypeOf<T extends Command> = T['type'];\nexport type CommandDataOf<T extends Command> = T['data'];\nexport type CommandMetaDataOf<T extends Command> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CreateCommandType<\n CommandType extends string,\n CommandData extends DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: CommandData;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const command = <CommandType extends Command<string, any, any>>(\n ...args: CommandMetaDataOf<CommandType> extends undefined\n ? [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata?: DefaultCommandMetadata | undefined,\n ]\n : [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata: CommandMetaDataOf<CommandType>,\n ]\n): CommandType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata } as CommandType)\n : ({ type, data } as CommandType);\n};\n\nexport type DefaultCommandMetadata = { now: Date };\n","import type { DefaultRecord, Flavour } from './';\n\nexport type BigIntStreamPosition = bigint;\nexport type BigIntGlobalPosition = bigint;\n\nexport type Event<\n EventType extends string = string,\n EventData extends DefaultRecord = DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Flavour<\n Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n >,\n 'Event'\n>;\n\nexport type EventTypeOf<T extends Event> = T['type'];\nexport type EventDataOf<T extends Event> = T['data'];\nexport type EventMetaDataOf<T extends Event> = T extends { metadata: infer M }\n ? M\n : undefined;\n\nexport type CanHandle<T extends Event> = EventTypeOf<T>[];\n\nexport type CreateEventType<\n EventType extends string,\n EventData extends DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const event = <EventType extends Event<string, any, any>>(\n ...args: EventMetaDataOf<EventType> extends undefined\n ? [type: EventTypeOf<EventType>, data: EventDataOf<EventType>]\n : [\n type: EventTypeOf<EventType>,\n data: EventDataOf<EventType>,\n metadata: EventMetaDataOf<EventType>,\n ]\n): EventType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata } as EventType)\n : ({ type, data } as EventType);\n};\n\nexport type CombinedReadEventMetadata<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n EventMetaDataOf<EventType> extends undefined\n ? EventMetaDataType\n : EventMetaDataOf<EventType> & EventMetaDataType;\n\nexport type ReadEvent<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = EventType & {\n metadata: CombinedReadEventMetadata<EventType, EventMetaDataType>;\n};\n\nexport type ReadEventMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = Readonly<{\n eventId: string;\n streamPosition: StreamPosition;\n streamName: string;\n}> &\n (GlobalPosition extends undefined\n ? object\n : { globalPosition: GlobalPosition });\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyReadEventMetadata = ReadEventMetadata<any, any>;\n\nexport type ReadEventMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = ReadEventMetadata<GlobalPosition>;\n\nexport type ReadEventMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = ReadEventMetadata<undefined, StreamPosition>;\n\nexport type GlobalPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ReadEventMetadataType extends ReadEventMetadata<infer GP, any> ? GP : never;\n\nexport type StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ReadEventMetadataType extends ReadEventMetadata<any, infer SV> ? SV : never;\n","import type { Command } from './command';\nimport type { Event } from './event';\n\n/// Inspired by https://blog.bittacklr.be/the-workflow-pattern.html\n\nexport type Workflow<\n Input extends Event | Command,\n State,\n Output extends Event | Command,\n> = {\n decide: (command: Input, state: State) => WorkflowOutput<Output>[];\n evolve: (currentState: State, event: WorkflowEvent<Output>) => State;\n initialState: () => State;\n};\n\nexport type WorkflowEvent<Output extends Command | Event> = Extract<\n Output,\n { __brand?: 'Event' }\n>;\n\nexport type WorkflowCommand<Output extends Command | Event> = Extract<\n Output,\n { __brand?: 'Command' }\n>;\n\nexport type WorkflowOutput<TOutput extends Command | Event> =\n | { kind: 'Reply'; message: TOutput }\n | { kind: 'Send'; message: WorkflowCommand<TOutput> }\n | { kind: 'Publish'; message: WorkflowEvent<TOutput> }\n | {\n kind: 'Schedule';\n message: TOutput;\n when: { afterInMs: number } | { at: Date };\n }\n | { kind: 'Complete' }\n | { kind: 'Accept' }\n | { kind: 'Ignore'; reason: string }\n | { kind: 'Error'; reason: string };\n\nexport const reply = <TOutput extends Command | Event>(\n message: TOutput,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Reply',\n message,\n };\n};\n\nexport const send = <TOutput extends Command | Event>(\n message: WorkflowCommand<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Send',\n message,\n };\n};\n\nexport const publish = <TOutput extends Command | Event>(\n message: WorkflowEvent<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Publish',\n message,\n };\n};\n\nexport const schedule = <TOutput extends Command | Event>(\n message: TOutput,\n when: { afterInMs: number } | { at: Date },\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Schedule',\n message,\n when,\n };\n};\n\nexport const complete = <\n TOutput extends Command | Event,\n>(): WorkflowOutput<TOutput> => {\n return {\n kind: 'Complete',\n };\n};\n\nexport const ignore = <TOutput extends Command | Event>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Ignore',\n reason,\n };\n};\n\nexport const error = <TOutput extends Command | Event>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Error',\n reason,\n };\n};\n\nexport const accept = <\n TOutput extends Command | Event,\n>(): WorkflowOutput<TOutput> => {\n return { kind: 'Accept' };\n};\n","import { type Event, type ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\ntype AfterEventStoreCommitHandlerWithoutContext<Store extends EventStore> = (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n) => Promise<void> | void;\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext = never,\n> = [HandlerContext] extends [never]\n ? AfterEventStoreCommitHandlerWithoutContext<Store>\n : (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n context: HandlerContext,\n ) => Promise<void> | void;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext = never,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext = never,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","import type { EventsPublisher } from '../../messageBus';\nimport type { Event, ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\nimport type { AfterEventStoreCommitHandler } from './afterEventStoreCommitHandler';\n\nexport const forwardToMessageBus =\n <Store extends EventStore, HandlerContext = never>(\n eventPublisher: EventsPublisher,\n ): AfterEventStoreCommitHandler<Store, HandlerContext> =>\n async (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n ): Promise<void> => {\n for (const message of messages) {\n await eventPublisher.publish(message);\n }\n };\n","import {\n event,\n type Event,\n type EventDataOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\n\nexport const GlobalStreamCaughtUpType = '__emt:GlobalStreamCaughtUp';\n\nexport type GlobalStreamCaughtUp = Event<\n '__emt:GlobalStreamCaughtUp',\n { globalPosition: bigint },\n { globalPosition: bigint }\n>;\n\nexport const isGlobalStreamCaughtUp = (\n event: Event,\n): event is GlobalStreamCaughtUp => event.type === GlobalStreamCaughtUpType;\n\nexport const caughtUpEventFrom =\n (position: bigint) =>\n (\n event: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n ): event is ReadEvent<\n GlobalStreamCaughtUp,\n ReadEventMetadataWithGlobalPosition\n > =>\n event.type === GlobalStreamCaughtUpType &&\n event.metadata?.globalPosition >= position;\n\nexport const globalStreamCaughtUp = (\n data: EventDataOf<GlobalStreamCaughtUp>,\n): GlobalStreamCaughtUp =>\n event<GlobalStreamCaughtUp>(GlobalStreamCaughtUpType, data, {\n globalPosition: data.globalPosition,\n });\n\nexport const isSubscriptionEvent = (\n event: Event,\n): event is GlobalSubscriptionEvent => isGlobalStreamCaughtUp(event);\n\nexport const isNotInternalEvent = (event: Event): boolean =>\n !isGlobalStreamCaughtUp(event);\n\nexport type GlobalSubscriptionEvent = GlobalStreamCaughtUp;\n","//import type { ReadableStream } from 'web-streams-polyfill';\nimport type {\n AnyReadEventMetadata,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n Event,\n GlobalPositionTypeOfReadEventMetadata,\n ReadEvent,\n ReadEventMetadata,\n StreamPositionTypeOfReadEventMetadata,\n} from '../typing';\nimport type { AfterEventStoreCommitHandler } from './afterCommit';\n//import type { GlobalSubscriptionEvent } from './events';\nimport type { ExpectedStreamVersion } from './expectedVersion';\n\n// #region event-store\nexport interface EventStore<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> {\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType, ReadEventMetadataType>,\n ): Promise<\n AggregateStreamResult<\n State,\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<ReadStreamResult<EventType, ReadEventMetadataType>>;\n\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<\n AppendToStreamResult<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n // streamEvents(): ReadableStream<\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // >;\n}\n\nexport type EventStoreReadEventMetadata<Store extends EventStore> =\n Store extends EventStore<infer ReadEventMetadataType>\n ? ReadEventMetadataType extends ReadEventMetadata<infer GV, infer SV>\n ? ReadEventMetadata<GV, SV> & ReadEventMetadataType\n : never\n : never;\n\nexport type GlobalPositionTypeOfEventStore<Store extends EventStore> =\n GlobalPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type StreamPositionTypeOfEventStore<Store extends EventStore> =\n StreamPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type EventStoreSession<EventStoreType extends EventStore> = {\n eventStore: EventStoreType;\n close: () => Promise<void>;\n};\n\nexport interface EventStoreSessionFactory<EventStoreType extends EventStore> {\n withSession<T = unknown>(\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n ): Promise<T>;\n}\n// #endregion event-store\n\nexport const canCreateEventStoreSession = <Store extends EventStore>(\n eventStore: Store | EventStoreSessionFactory<Store>,\n): eventStore is EventStoreSessionFactory<Store> => 'withSession' in eventStore;\n\nexport const nulloSessionFactory = <EventStoreType extends EventStore>(\n eventStore: EventStoreType,\n): EventStoreSessionFactory<EventStoreType> => ({\n withSession: (callback) => {\n const nulloSession: EventStoreSession<EventStoreType> = {\n eventStore,\n close: () => Promise.resolve(),\n };\n\n return callback(nulloSession);\n },\n});\n\n////////////////////////////////////////////////////////////////////\n/// ReadStream types\n////////////////////////////////////////////////////////////////////\n\nexport type ReadStreamOptions<StreamVersion = BigIntStreamPosition> = (\n | {\n from: StreamVersion;\n }\n | { to: StreamVersion }\n | { from: StreamVersion; maxCount?: bigint }\n | {\n expectedStreamVersion: ExpectedStreamVersion<StreamVersion>;\n }\n) & {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type ReadStreamResult<\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n currentStreamVersion: StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>;\n events: ReadEvent<EventType, ReadEventMetadataType>[];\n streamExists: boolean;\n};\n\n////////////////////////////////////////////////////////////////////\n/// AggregateStream types\n////////////////////////////////////////////////////////////////////\n\ntype Evolve<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n | ((currentState: State, event: EventType) => State)\n | ((\n currentState: State,\n event: ReadEvent<EventType, ReadEventMetadataType>,\n ) => State)\n | ((currentState: State, event: ReadEvent<EventType>) => State);\n\nexport type AggregateStreamOptions<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n evolve: Evolve<State, EventType, ReadEventMetadataType>;\n initialState: () => State;\n read?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >;\n};\n\nexport type AggregateStreamResult<\n State,\n StreamPosition = BigIntStreamPosition,\n> = {\n currentStreamVersion: StreamPosition;\n state: State;\n streamExists: boolean;\n};\n\nexport type AggregateStreamResultWithGlobalPosition<\n State,\n StreamPosition = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> =\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: true;\n lastEventGlobalPosition: GlobalPosition;\n })\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: false;\n });\n\nexport type AggregateStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['aggregateStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// AppendToStream types\n////////////////////////////////////////////////////////////////////\n\nexport type AppendToStreamOptions<StreamVersion = BigIntStreamPosition> = {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type AppendToStreamResult<StreamVersion = BigIntStreamPosition> = {\n nextExpectedStreamVersion: StreamVersion;\n createdNewStream: boolean;\n};\n\nexport type AppendToStreamResultWithGlobalPosition<\n StreamVersion = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> = AppendToStreamResult<StreamVersion> & {\n lastEventGlobalPosition: GlobalPosition;\n};\n\nexport type AppendStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['appendToStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// DefaultEventStoreOptions\n////////////////////////////////////////////////////////////////////\n\nexport type DefaultEventStoreOptions<\n Store extends EventStore,\n HandlerContext = never,\n> = {\n /**\n * Pluggable set of hooks informing about the event store internal behaviour.\n */\n hooks?: {\n /**\n * This hook will be called **AFTER** events were stored in the event store.\n * It's designed to handle scenarios where delivery and ordering guarantees do not matter much.\n *\n * **WARNINGS:**\n *\n * 1. It will be called **EXACTLY ONCE** if append succeded.\n * 2. If the hook fails, its append **will still silently succeed**, and no error will be thrown.\n * 3. Wen process crashes after events were committed, but before the hook was called, delivery won't be retried.\n * That can lead to state inconsistencies.\n * 4. In the case of high concurrent traffic, **race conditions may cause ordering issues**.\n * For instance, where the second hook takes longer to process than the first one, ordering won't be guaranteed.\n *\n * @type {AfterEventStoreCommitHandler<Store, HandlerContext>}\n */\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n };\n};\n","import { ConcurrencyError } from '../errors';\nimport type { BigIntStreamPosition, Flavour } from '../typing';\n\nexport type ExpectedStreamVersion<VersionType = BigIntStreamPosition> =\n | ExpectedStreamVersionWithValue<VersionType>\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue<VersionType = BigIntStreamPosition> =\n Flavour<VersionType, 'StreamVersion'>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = <StreamVersion = BigIntStreamPosition>(\n current: StreamVersion | undefined,\n expected: ExpectedStreamVersion<StreamVersion>,\n defaultVersion: StreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = <\n StreamVersion = BigIntStreamPosition,\n>(\n current: StreamVersion,\n expected: ExpectedStreamVersion<StreamVersion> | undefined,\n defaultVersion: StreamVersion,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError<\n VersionType = BigIntStreamPosition,\n> extends ConcurrencyError {\n constructor(\n current: VersionType,\n expected: ExpectedStreamVersion<VersionType>,\n ) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError;\n","import { v4 as uuid } from 'uuid';\nimport type {\n BigIntStreamPosition,\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport {\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultEventStoreOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { StreamingCoordinator } from './subscriptions';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition>;\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore>;\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n const streamingCoordinator = StreamingCoordinator();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType>(streamName, read);\n\n const events = result?.events ?? [];\n\n return {\n currentStreamVersion: BigInt(events.length),\n state: events.reduce(evolve, initialState()),\n streamExists: result.streamExists,\n };\n },\n\n readStream: <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options && 'from' in options ? options.from : 0);\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : (events?.length ?? 1),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? events\n .map(\n (e) =>\n e as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n )\n .slice(from, to)\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition: BigInt(getAllEventsCount() + index + 1),\n };\n return {\n ...event,\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [...currentEvents, ...newEvents]);\n await streamingCoordinator.notify(newEvents);\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit(newEvents, eventStoreOptions?.hooks);\n\n return result;\n },\n\n //streamEvents: streamingCoordinator.stream,\n };\n};\n","import { TransformStream } from 'web-streams-polyfill';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport { globalStreamCaughtUp, type GlobalSubscriptionEvent } from '../events';\n\nexport const streamTrackingGlobalPosition = (\n currentEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n) => new CaughtUpTransformStream(currentEvents);\n\nexport class CaughtUpTransformStream extends TransformStream<\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n | ReadEvent<Event, ReadEventMetadataWithGlobalPosition>\n | GlobalSubscriptionEvent\n> {\n private _currentPosition: bigint;\n private _logPosition: bigint;\n\n constructor(events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]) {\n super({\n start: (controller) => {\n let globalPosition = 0n;\n for (const event of events) {\n controller.enqueue(event);\n globalPosition = event.metadata.globalPosition;\n }\n controller.enqueue(globalStreamCaughtUp({ globalPosition }));\n },\n transform: (event, controller) => {\n this._currentPosition = event.metadata.globalPosition;\n controller.enqueue(event);\n\n if (this._currentPosition < this._logPosition) return;\n\n controller.enqueue(\n globalStreamCaughtUp({ globalPosition: this._currentPosition }),\n );\n },\n });\n\n this._currentPosition = this._logPosition =\n events.length > 0\n ? events[events.length - 1]!.metadata.globalPosition\n : 0n;\n }\n\n public set logPosition(value: bigint) {\n this._logPosition = value;\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport { notifyAboutNoActiveReadersStream } from '../../streaming/transformations/notifyAboutNoActiveReaders';\nimport { writeToStream } from '../../streaming/writers';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport {\n CaughtUpTransformStream,\n streamTrackingGlobalPosition,\n} from './caughtUpTransformStream';\n\nexport const StreamingCoordinator = () => {\n const allEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[] = [];\n const listeners = new Map<string, CaughtUpTransformStream>();\n\n return {\n notify: async (\n events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n ) => {\n if (events.length === 0) return;\n\n allEvents.push(...events);\n\n for (const listener of listeners.values()) {\n listener.logPosition =\n events[events.length - 1]!.metadata.globalPosition;\n\n await writeToStream(listener, events);\n }\n },\n\n stream: () => {\n const streamId = uuid();\n const transformStream = streamTrackingGlobalPosition(allEvents);\n\n listeners.set(streamId, transformStream);\n return transformStream.readable.pipeThrough(\n notifyAboutNoActiveReadersStream(\n (stream) => {\n if (listeners.has(stream.streamId))\n listeners.delete(stream.streamId);\n },\n { streamId },\n ),\n );\n },\n };\n};\n","import { v4 as uuid } from 'uuid';\nimport { TransformStream } from 'web-streams-polyfill';\n\nexport const notifyAboutNoActiveReadersStream = <Item>(\n onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);\n\nexport class NotifyAboutNoActiveReadersStream<Item> extends TransformStream<\n Item,\n Item\n> {\n private checkInterval: NodeJS.Timeout | null = null;\n public readonly streamId: string;\n private _isStopped: boolean = false;\n public get hasActiveSubscribers() {\n return !this._isStopped;\n }\n\n constructor(\n private onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n ) {\n super({\n cancel: (reason) => {\n console.log('Stream was canceled. Reason:', reason);\n this.stopChecking();\n },\n });\n this.streamId = options?.streamId ?? uuid();\n\n this.onNoActiveReaderCallback = onNoActiveReaderCallback;\n\n this.startChecking(options?.intervalCheckInMs ?? 20);\n }\n\n private startChecking(interval: number) {\n this.checkInterval = setInterval(() => {\n this.checkNoActiveReader();\n }, interval);\n }\n\n private stopChecking() {\n if (!this.checkInterval) return;\n\n clearInterval(this.checkInterval);\n this.checkInterval = null;\n this._isStopped = true;\n this.onNoActiveReaderCallback(this);\n }\n\n private checkNoActiveReader() {\n if (!this.readable.locked && !this._isStopped) {\n this.stopChecking();\n }\n }\n}\n","import type { ReadableStream, TransformStream } from 'web-streams-polyfill';\nimport { noMoreWritingOn } from './noMoreWritingOn';\n\nexport const writeToStream = async <In, Out = In>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<boolean> => {\n if (stream.writable.locked) return false;\n\n const writer = stream.writable.getWriter();\n await writer.ready;\n\n if (!stream.readable.locked) return false;\n\n try {\n for (const item of items) {\n await writer.write(item);\n }\n } catch (error) {\n console.log(error);\n } finally {\n await writer.close();\n }\n return true;\n};\n\nexport const writeToStreamAndStop = async <In, Out>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<ReadableStream<Out>> => {\n await writeToStream(stream, items);\n return await noMoreWritingOn(stream);\n};\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","export const deepEquals = <T>(left: T, right: T): boolean => {\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n if (Array.isArray(left)) {\n return (\n Array.isArray(right) &&\n left.length === right.length &&\n left.every((val, index) => deepEquals(val, right[index]))\n );\n }\n\n if (\n typeof left !== 'object' ||\n typeof right !== 'object' ||\n left === null ||\n right === null\n ) {\n return left === right;\n }\n\n if (Array.isArray(right)) return false;\n\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (\n keys1.length !== keys2.length ||\n !keys1.every((key) => keys2.includes(key))\n )\n return false;\n\n for (const key in left) {\n if (left[key] instanceof Function && right[key] instanceof Function)\n continue;\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export const sum = (\n iterator: Iterator<number, number, number> | Iterator<number>,\n) => {\n let value,\n done: boolean | undefined,\n sum = 0;\n do {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n ({ value, done } = iterator.next());\n sum += value || 0;\n } while (!done);\n return sum;\n};\n","import { EmmettError } from '../errors';\n\nexport type TaskQueue = TaskQueueItem[];\n\nexport type TaskQueueItem = {\n task: () => Promise<void>;\n options?: EnqueueTaskOptions;\n};\n\nexport type TaskProcessorOptions = {\n maxActiveTasks: number;\n maxQueueSize: number;\n maxTaskIdleTime?: number;\n};\n\nexport type Task<T> = (context: TaskContext) => Promise<T>;\n\nexport type TaskContext = {\n ack: () => void;\n};\n\nexport type EnqueueTaskOptions = { taskGroupId?: string };\n\nexport class TaskProcessor {\n private queue: TaskQueue = [];\n private isProcessing = false;\n private activeTasks = 0;\n private activeGroups: Set<string> = new Set();\n\n constructor(private options: TaskProcessorOptions) {}\n\n enqueue<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n if (this.queue.length >= this.options.maxQueueSize) {\n return Promise.reject(\n new EmmettError(\n 'Too many pending connections. Please try again later.',\n ),\n );\n }\n\n return this.schedule(task, options);\n }\n\n waitForEndOfProcessing(): Promise<void> {\n return this.schedule(({ ack }) => Promise.resolve(ack()));\n }\n\n private schedule<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n return promiseWithDeadline(\n (resolve, reject) => {\n const taskWithContext = () => {\n return new Promise<void>((resolveTask, failTask) => {\n const taskPromise = task({\n ack: resolveTask,\n });\n\n taskPromise.then(resolve).catch((err) => {\n // eslint-disable-next-line @typescript-eslint/prefer-promise-reject-errors\n failTask(err);\n reject(err);\n });\n });\n };\n\n this.queue.push({ task: taskWithContext, options });\n if (!this.isProcessing) {\n this.ensureProcessing();\n }\n },\n { deadline: this.options.maxTaskIdleTime },\n );\n }\n\n private ensureProcessing(): void {\n if (this.isProcessing) return;\n this.isProcessing = true;\n this.processQueue();\n }\n\n private processQueue(): void {\n try {\n while (\n this.activeTasks < this.options.maxActiveTasks &&\n this.queue.length > 0\n ) {\n const item = this.takeFirstAvailableItem();\n\n if (item === null) return;\n\n const groupId = item.options?.taskGroupId;\n\n if (groupId) {\n // Mark the group as active\n this.activeGroups.add(groupId);\n }\n\n this.activeTasks++;\n void this.executeItem(item);\n }\n } catch (error) {\n console.error(error);\n throw error;\n } finally {\n this.isProcessing = false;\n if (\n this.hasItemsToProcess() &&\n this.activeTasks < this.options.maxActiveTasks\n ) {\n this.ensureProcessing();\n }\n }\n }\n\n private async executeItem({ task, options }: TaskQueueItem): Promise<void> {\n try {\n await task();\n } finally {\n this.activeTasks--;\n\n // Mark the group as inactive after task completion\n if (options && options.taskGroupId) {\n this.activeGroups.delete(options.taskGroupId);\n }\n\n this.ensureProcessing();\n }\n }\n\n private takeFirstAvailableItem = (): TaskQueueItem | null => {\n const taskIndex = this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n );\n\n if (taskIndex === -1) {\n // All remaining tasks are blocked by active groups\n return null;\n }\n\n // Remove the task from the queue\n const [item] = this.queue.splice(taskIndex, 1);\n\n return item ?? null;\n };\n\n private hasItemsToProcess = (): boolean =>\n this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n ) !== -1;\n}\n\nconst DEFAULT_PROMISE_DEADLINE = 2147483647;\n\nconst promiseWithDeadline = <T>(\n executor: (\n resolve: (value: T | PromiseLike<T>) => void,\n reject: (reason?: unknown) => void,\n ) => void,\n options: { deadline?: number },\n) => {\n return new Promise<T>((resolve, reject) => {\n let taskStarted = false;\n\n const maxWaitingTime = options.deadline || DEFAULT_PROMISE_DEADLINE;\n\n let timeoutId: NodeJS.Timeout | null = setTimeout(() => {\n if (!taskStarted) {\n reject(\n new Error('Task was not started within the maximum waiting time'),\n );\n }\n }, maxWaitingTime);\n\n executor((value) => {\n taskStarted = true;\n if (timeoutId) {\n clearTimeout(timeoutId);\n }\n timeoutId = null;\n resolve(value);\n }, reject);\n });\n};\n","import { TaskProcessor } from '../../taskProcessing';\n\nexport type LockOptions = { lockId: number };\n\nexport type AcquireLockOptions = { lockId: string };\nexport type ReleaseLockOptions = { lockId: string };\n\nexport type Lock = {\n acquire(options: AcquireLockOptions): Promise<void>;\n tryAcquire(options: AcquireLockOptions): Promise<boolean>;\n release(options: ReleaseLockOptions): Promise<boolean>;\n withAcquire: <Result = unknown>(\n handle: () => Promise<Result>,\n options: AcquireLockOptions,\n ) => Promise<Result>;\n};\n\nexport const InProcessLock = (): Lock => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: Number.MAX_VALUE,\n maxQueueSize: Number.MAX_VALUE,\n });\n\n // Map to store ack functions of currently held locks: lockId -> ack()\n const locks = new Map<string, () => void>();\n\n return {\n async acquire({ lockId }: AcquireLockOptions): Promise<void> {\n // If the lock is already held, we just queue up another task in the same group.\n // TaskProcessor ensures tasks in the same group run one at a time.\n await new Promise<void>((resolve, reject) => {\n taskProcessor\n .enqueue(\n ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n // We do NOT call ack() here. We hold onto the lock.\n resolve();\n return Promise.resolve();\n },\n { taskGroupId: lockId },\n )\n .catch(reject);\n });\n },\n\n async tryAcquire({ lockId }: AcquireLockOptions): Promise<boolean> {\n // If lock is already held, fail immediately\n if (locks.has(lockId)) {\n return false;\n }\n\n // TODO: Check pending queue\n await this.acquire({ lockId });\n\n return true;\n },\n\n release({ lockId }: ReleaseLockOptions): Promise<boolean> {\n const ack = locks.get(lockId);\n if (ack === undefined) {\n return Promise.resolve(true);\n }\n locks.delete(lockId);\n ack();\n return Promise.resolve(true);\n },\n\n async withAcquire<Result = unknown>(\n handle: () => Promise<Result>,\n { lockId }: AcquireLockOptions,\n ): Promise<Result> {\n return taskProcessor.enqueue(\n async ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n\n // We do NOT call ack() here. We hold onto the lock.\n try {\n return await handle();\n } finally {\n locks.delete(lockId);\n ack();\n }\n },\n { taskGroupId: lockId },\n );\n },\n };\n};\n","import retry from 'async-retry';\n\nexport type AsyncRetryOptions = retry.Options & {\n shouldRetryError?: (error: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n return await fn();\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","import {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport type { Event } from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\n\nexport const CommandHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type CommandHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromCommandHandlerRetryOptions = (\n retryOptions: CommandHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return CommandHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...CommandHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region command-handler\nexport type CommandHandlerResult<\n State,\n StreamEvent extends Event,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newState: State;\n newEvents: StreamEvent[];\n};\n\nexport type CommandHandlerOptions<State, StreamEvent extends Event> = {\n evolve: (state: State, event: StreamEvent) => State;\n initialState: () => State;\n mapToStreamId?: (id: string) => string;\n retry?: CommandHandlerRetryOptions;\n};\n\nexport type HandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] &\n (\n | {\n expectedStreamVersion?: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >;\n }\n | {\n retry?: CommandHandlerRetryOptions;\n }\n );\n\nexport const CommandHandler =\n <State, StreamEvent extends Event>(\n options: CommandHandlerOptions<State, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n store: Store,\n id: string,\n handle: (\n state: State,\n ) =>\n | StreamEvent\n | StreamEvent[]\n | Promise<StreamEvent>\n | Promise<StreamEvent[]>,\n handleOptions?: HandleOptions<Store>,\n ): Promise<CommandHandlerResult<State, StreamEvent, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n CommandHandlerResult<\n State,\n StreamEvent,\n StreamPositionTypeOfEventStore<Store>\n >\n >(store, async ({ eventStore }) => {\n const { evolve, initialState } = options;\n const mapToStreamId = options.mapToStreamId ?? ((id) => id);\n\n const streamName = mapToStreamId(id);\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n State,\n StreamEvent\n >(streamName, {\n evolve,\n initialState,\n read: {\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const {\n state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n currentStreamVersion,\n streamExists: _streamExists,\n ...restOfAggregationResult\n } = aggregationResult;\n\n // 3. Run business logic\n const result = await handle(state);\n\n const newEvents = Array.isArray(result) ? result : [result];\n\n if (newEvents.length === 0) {\n return {\n ...restOfAggregationResult,\n newEvents: [],\n newState: state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n nextExpectedStreamVersion: currentStreamVersion,\n createdNewStream: false,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const expectedStreamVersion: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n > =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? (currentStreamVersion as ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >)\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n newEvents,\n {\n ...handleOptions,\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with updated state\n return {\n ...appendResult,\n newEvents,\n newState: newEvents.reduce(evolve, state),\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n });\n\n return result;\n },\n fromCommandHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion command-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type { EventStore } from '../eventStore';\nimport type { Command, Event } from '../typing';\nimport type { Decider } from '../typing/decider';\nimport {\n CommandHandler,\n type CommandHandlerOptions,\n type HandleOptions,\n} from './handleCommand';\n\n// #region command-handler\n\nexport type DeciderCommandHandlerOptions<\n State,\n CommandType extends Command,\n StreamEvent extends Event,\n> = CommandHandlerOptions<State, StreamEvent> &\n Decider<State, CommandType, StreamEvent>;\n\nexport const DeciderCommandHandler =\n <State, CommandType extends Command, StreamEvent extends Event>(\n options: DeciderCommandHandlerOptions<State, CommandType, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n eventStore: Store,\n id: string,\n command: CommandType,\n handleOptions?: HandleOptions<Store>,\n ) => {\n const { decide, ...rest } = options;\n\n return CommandHandler<State, StreamEvent>(rest)(\n eventStore,\n id,\n (state) => decide(command, state),\n handleOptions,\n );\n };\n// #endregion command-handler\n","import { EmmettError } from '../errors';\nimport {\n type Command,\n type CommandTypeOf,\n type Event,\n type EventTypeOf,\n} from '../typing';\n\nexport interface CommandSender {\n send<CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void>;\n}\n\nexport interface EventsPublisher {\n publish<EventType extends Event = Event>(event: EventType): Promise<void>;\n}\n\nexport type ScheduleOptions = { afterInMs: number } | { at: Date };\n\nexport interface MessageScheduler<CommandOrEvent extends Command | Event> {\n schedule<MessageType extends CommandOrEvent>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandBus extends CommandSender, MessageScheduler<Command> {}\n\nexport interface EventBus extends EventsPublisher, MessageScheduler<Event> {}\n\nexport interface MessageBus extends CommandBus, EventBus {\n schedule<MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\ntype CommandHandler<CommandType extends Command = Command> = (\n command: CommandType,\n) => Promise<void> | void;\n\nexport interface CommandProcessor {\n handle<CommandType extends Command>(\n commandHandler: CommandHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void;\n}\n\ntype EventHandler<EventType extends Event = Event> = (\n event: EventType,\n) => Promise<void> | void;\n\nexport interface EventProcessor {\n subscribe<EventType extends Event>(\n eventHandler: EventHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void;\n}\n\nexport type ScheduledMessage = {\n message: Event | Command;\n options?: ScheduleOptions;\n};\n\nexport interface ScheduledMessageProcessor {\n dequeue(): ScheduledMessage[];\n}\n\nexport type MessageHandler = CommandHandler | EventHandler;\n\nexport type MessageProcessor = EventProcessor | CommandProcessor;\n\nexport const getInMemoryMessageBus = (): MessageBus &\n EventProcessor &\n CommandProcessor &\n ScheduledMessageProcessor => {\n const allHandlers = new Map<string, MessageHandler[]>();\n let pendingMessages: ScheduledMessage[] = [];\n\n return {\n send: async <CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void> => {\n const handlers = allHandlers.get(command.type);\n\n if (handlers === undefined || handlers.length === 0)\n throw new EmmettError(\n `No handler registered for command ${command.type}!`,\n );\n\n const commandHandler = handlers[0] as CommandHandler<CommandType>;\n\n await commandHandler(command);\n },\n\n publish: async <EventType extends Event = Event>(\n event: EventType,\n ): Promise<void> => {\n const handlers = allHandlers.get(event.type) ?? [];\n\n for (const handler of handlers) {\n const eventHandler = handler as EventHandler<EventType>;\n\n await eventHandler(event);\n }\n },\n\n schedule: <MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void => {\n pendingMessages = [...pendingMessages, { message, options: when }];\n },\n\n handle: <CommandType extends Command>(\n commandHandler: CommandHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void => {\n const alreadyRegistered = [...allHandlers.keys()].filter((registered) =>\n commandTypes.includes(registered),\n );\n\n if (alreadyRegistered.length > 0)\n throw new EmmettError(\n `Cannot register handler for commands ${alreadyRegistered.join(', ')} as they're already registered!`,\n );\n for (const commandType of commandTypes) {\n allHandlers.set(commandType, [commandHandler as MessageHandler]);\n }\n },\n\n subscribe<EventType extends Event>(\n eventHandler: EventHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void {\n for (const eventType of eventTypes) {\n if (!allHandlers.has(eventType)) allHandlers.set(eventType, []);\n\n allHandlers.set(eventType, [\n ...(allHandlers.get(eventType) ?? []),\n eventHandler as MessageHandler,\n ]);\n }\n },\n\n dequeue: (): ScheduledMessage[] => {\n const pending = pendingMessages;\n pendingMessages = [];\n return pending;\n },\n };\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","import { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyReadEventMetadata,\n CanHandle,\n DefaultRecord,\n Event,\n ReadEvent,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n events: ReadEvent<EventType, EventMetaDataType>[],\n context: ProjectionHandlerContext,\n) => Promise<void> | void;\n\nexport interface ProjectionDefinition<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<Event>;\n handle: ProjectionHandler<\n Event,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n}\n\nexport interface TypedProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n > = TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >,\n>(\n definition: ProjectionDefintionType,\n): ProjectionDefintionType => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n > = ProjectionDefinition<ReadEventMetadataType, ProjectionHandlerContext>,\n>(\n definitions: ProjectionDefintionType[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] => definitions.map((projection) => ({ type: 'inline', projection }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n > = ProjectionDefinition<ReadEventMetadataType, ProjectionHandlerContext>,\n>(\n definitions: ProjectionDefintionType[],\n): ProjectionRegistration<\n 'async',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] => definitions.map((projection) => ({ type: 'async', projection }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","export const concatUint8Arrays = (chunks: Uint8Array[]): Uint8Array => {\n const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const chunk of chunks) {\n result.set(chunk, offset);\n offset += chunk.length;\n }\n return result;\n};\n","import type { ReadableStream } from 'web-streams-polyfill';\n\nexport const collect = async <T>(stream: ReadableStream<T>): Promise<T[]> => {\n const results: T[] = [];\n\n for await (const value of stream) {\n results.push(value as T);\n }\n\n return results;\n};\n","import type { Decoder } from '.';\nimport { concatUint8Arrays } from '../binaryArrays';\n\nexport class BinaryJsonDecoder<Decoded>\n implements Decoder<Uint8Array, Decoded>\n{\n private buffer: Uint8Array[] = [];\n\n addToBuffer(data: Uint8Array): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n return text.includes('\\n');\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage()) {\n return null;\n }\n\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n const delimiterIndex = text.indexOf('\\n');\n\n if (delimiterIndex === -1) {\n return null;\n }\n\n const jsonString = text.slice(0, delimiterIndex);\n const remaining = new Uint8Array(combined.buffer, delimiterIndex + 1);\n this.buffer = remaining.byteLength > 0 ? [remaining] : [];\n\n return JSON.parse(jsonString) as Decoded;\n }\n}\n","import type { Decoder } from '.';\n\nexport class StringDecoder<Decoded> implements Decoder<string, Decoded> {\n protected buffer: string[] = [];\n\n constructor(private transform: (input: string) => Decoded) {\n this.transform = transform;\n }\n\n addToBuffer(data: string): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer.some((chunk) => chunk.includes('\\n'));\n }\n\n decode(): Decoded | null {\n const completeString = this.buffer.join('');\n\n if (!this.hasCompleteMessage()) {\n if (completeString.trim().length > 0) {\n throw new Error('Unterminated string in JSON at position');\n }\n return null;\n }\n\n const delimiterIndex = completeString.indexOf('\\n');\n const message = completeString.slice(0, delimiterIndex).trim();\n this.buffer = [completeString.slice(delimiterIndex + 1)];\n\n return this.transform(message);\n }\n}\n","import { StringDecoder } from './string';\n\nexport class JsonDecoder<Decoded> extends StringDecoder<Decoded> {\n constructor() {\n super((jsonString) => JSON.parse(jsonString) as Decoded);\n }\n}\n","import type { Decoder } from '.';\n\nexport class ObjectDecoder<Decoded> implements Decoder<Decoded, Decoded> {\n private buffer: Decoded | null = null;\n\n addToBuffer(data: Decoded): void {\n this.buffer = data;\n }\n\n clearBuffer(): void {\n this.buffer = null;\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer !== null;\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage() || !this.buffer) {\n return null;\n }\n\n const data = this.buffer;\n this.clearBuffer();\n return data as Decoded;\n }\n}\n","import type { Decoder } from '.';\nimport { BinaryJsonDecoder } from './binary';\nimport { JsonDecoder } from './json';\nimport { ObjectDecoder } from './object';\n\nexport class CompositeDecoder<Decoded> implements Decoder<unknown, Decoded> {\n constructor(\n private decoders: [(data: unknown) => boolean, Decoder<unknown, Decoded>][],\n ) {}\n\n private decoderFor(data: unknown): Decoder<unknown, unknown> | null {\n const decoder = this.decoders.find((d) => d[0](data));\n\n if (!decoder) return null;\n\n return decoder[1];\n }\n\n addToBuffer(data: unknown): void {\n this.decoderFor(data)?.addToBuffer(data);\n }\n\n clearBuffer(): void {\n for (const decoder of this.decoders.map((d) => d[1])) {\n decoder.clearBuffer();\n }\n }\n\n hasCompleteMessage(): boolean {\n return this.decoders.some((d) => d[1].hasCompleteMessage());\n }\n\n decode(): Decoded | null {\n const decoder = this.decoders\n .map((d) => d[1])\n .find((d) => d.hasCompleteMessage());\n\n return decoder?.decode() ?? null;\n }\n}\n\nexport class DefaultDecoder<Decoded> extends CompositeDecoder<Decoded> {\n constructor() {\n super([\n [(data) => typeof data === 'string', new JsonDecoder<Decoded>()],\n [(data) => data instanceof Uint8Array, new BinaryJsonDecoder<Decoded>()],\n [(data) => typeof data === 'object', new ObjectDecoder<Decoded>()],\n ]);\n }\n}\n","import { ReadableStream } from 'web-streams-polyfill';\n\nexport const fromArray = <T>(chunks: T[]) =>\n new ReadableStream<T>({\n start(controller) {\n for (const chunk of chunks) controller.enqueue(chunk);\n controller.close();\n },\n });\n","import { fromArray } from './fromArray';\nexport const streamGenerators = { fromArray };\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport type { AsyncRetryOptions } from '../utils';\nimport type { Decoder } from './decoders';\nimport { DefaultDecoder } from './decoders/composite';\nimport { streamTransformations } from './transformations';\n\nconst { retry } = streamTransformations;\n\nexport const restream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n): ReadableStream<Transformed> =>\n retry(createSourceStream, handleChunk(transform, decoder), retryOptions)\n .readable;\n\nconst handleChunk =\n <Source = unknown, Transformed = Source, StreamType = Source>(\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n ) =>\n (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ): void => {\n const { done: isDone, value } = readResult;\n\n if (value) decoder.addToBuffer(value);\n\n if (!isDone && !decoder.hasCompleteMessage()) return;\n\n decodeAndTransform(decoder, transform, controller);\n };\n\nconst decodeAndTransform = <StreamType, Source, Transformed = Source>(\n decoder: Decoder<StreamType, Source>,\n transform: (input: Source) => Transformed,\n controller: TransformStreamDefaultController<Transformed>,\n) => {\n try {\n const decoded = decoder.decode();\n if (!decoded) return; // TODO: Add a proper handling of decode errors\n\n const transformed = transform(decoded);\n controller.enqueue(transformed);\n } catch (error) {\n controller.error(new Error(`Decoding error: ${error?.toString()}`));\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const filter = <Item>(filter: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n if (filter(chunk)) {\n controller.enqueue(chunk);\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const map = <From, To>(map: (item: From) => To) =>\n new TransformStream<From, To>({\n transform(chunk, controller) {\n controller.enqueue(map(chunk));\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const reduce = <I, O>(\n reducer: (accumulator: O, chunk: I) => O,\n initialValue: O,\n) => new ReduceTransformStream<I, O>(reducer, initialValue);\n\nexport class ReduceTransformStream<I, O> extends TransformStream<I, O> {\n private accumulator: O;\n private reducer: (accumulator: O, chunk: I) => O;\n\n constructor(reducer: (accumulator: O, chunk: I) => O, initialValue: O) {\n super({\n transform: (chunk) => {\n this.accumulator = this.reducer(this.accumulator, chunk);\n },\n flush: (controller) => {\n controller.enqueue(this.accumulator);\n controller.terminate();\n },\n });\n\n this.accumulator = initialValue;\n this.reducer = reducer;\n }\n}\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n TransformStream,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport { type AsyncRetryOptions, asyncRetry } from '../../utils';\n\nexport const retryStream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n): TransformStream<Source, Transformed> =>\n new TransformStream<Source, Transformed>({\n start(controller) {\n asyncRetry(\n () => onRestream(createSourceStream, handleChunk, controller),\n retryOptions,\n ).catch((error) => {\n controller.error(error);\n });\n },\n });\n\nconst onRestream = async <StreamType, Source, Transformed = Source>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n controller: TransformStreamDefaultController<Transformed>,\n): Promise<void> => {\n const sourceStream = createSourceStream();\n const reader = sourceStream.getReader();\n\n try {\n let done: boolean;\n\n do {\n const result = await reader.read();\n done = result.done;\n\n await handleChunk(result, controller);\n\n if (done) {\n controller.terminate();\n }\n } while (!done);\n } finally {\n reader.releaseLock();\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const skip = <T>(limit: number) => new SkipTransformStream<T>(limit);\n\nexport class SkipTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private skip: number;\n\n constructor(skip: number) {\n super({\n transform: (chunk, controller) => {\n this.count++;\n if (this.count > this.skip) {\n controller.enqueue(chunk);\n }\n },\n });\n\n this.skip = skip;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopAfter = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n controller.enqueue(chunk);\n\n if (stopCondition(chunk)) {\n controller.terminate();\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopOn = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n async transform(chunk, controller) {\n if (!stopCondition(chunk)) {\n controller.enqueue(chunk);\n return;\n }\n await Promise.resolve();\n controller.terminate();\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const take = <T>(limit: number) => new TakeTransformStream<T>(limit);\n\nexport class TakeTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private limit: number;\n\n constructor(limit: number) {\n super({\n transform: (chunk, controller) => {\n if (this.count < this.limit) {\n this.count++;\n controller.enqueue(chunk);\n } else {\n controller.terminate();\n }\n },\n });\n\n this.limit = limit;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const waitAtMost = <Item>(waitTimeInMs: number) =>\n new TransformStream<Item, Item>({\n start(controller) {\n const timeoutId = setTimeout(() => {\n controller.terminate();\n }, waitTimeInMs);\n\n const originalTerminate = controller.terminate.bind(controller);\n\n // Clear the timeout if the stream is terminated early\n controller.terminate = () => {\n clearTimeout(timeoutId);\n originalTerminate();\n };\n },\n transform(chunk, controller) {\n controller.enqueue(chunk);\n },\n });\n","import { filter } from './filter';\nimport { map } from './map';\nimport {\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n} from './notifyAboutNoActiveReaders';\nimport { reduce, ReduceTransformStream } from './reduce';\nimport { retryStream } from './retry';\nimport { skip, SkipTransformStream } from './skip';\nimport { stopAfter } from './stopAfter';\nimport { stopOn } from './stopOn';\nimport { take, TakeTransformStream } from './take';\nimport { waitAtMost } from './waitAtMost';\n\nexport const streamTransformations = {\n filter,\n take,\n TakeTransformStream,\n skip,\n SkipTransformStream,\n map,\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n reduce,\n ReduceTransformStream,\n retry: retryStream,\n stopAfter,\n stopOn,\n waitAtMost,\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<void>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import { isErrorConstructor, type ErrorConstructor } from '../errors';\nimport { AssertionError, assertThatArray, assertTrue } from './assertions';\n\ntype ErrorCheck<ErrorType> = (error: ErrorType) => boolean;\n\nexport type ThenThrows<ErrorType extends Error> =\n | (() => void)\n | ((errorConstructor: ErrorConstructor<ErrorType>) => void)\n | ((errorCheck: ErrorCheck<ErrorType>) => void)\n | ((\n errorConstructor: ErrorConstructor<ErrorType>,\n errorCheck?: ErrorCheck<ErrorType>,\n ) => void);\n\nexport type DeciderSpecfication<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => void;\n thenNothingHappened: () => void;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => void;\n };\n};\n\nexport const DeciderSpecification = {\n for: <Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Event | Event[];\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n }): DeciderSpecfication<Command, Event> => {\n {\n return (givenEvents: Event | Event[]) => {\n return {\n when: (command: Command) => {\n const handle = () => {\n const existingEvents = Array.isArray(givenEvents)\n ? givenEvents\n : [givenEvents];\n\n const currentState = existingEvents.reduce<State>(\n decider.evolve,\n decider.initialState(),\n );\n\n return decider.decide(command, currentState);\n };\n\n return {\n then: (expectedEvents: Event | Event[]): void => {\n const resultEvents = handle();\n\n const resultEventsArray = Array.isArray(resultEvents)\n ? resultEvents\n : [resultEvents];\n\n const expectedEventsArray = Array.isArray(expectedEvents)\n ? expectedEvents\n : [expectedEvents];\n\n assertThatArray(resultEventsArray).containsOnlyElementsMatching(\n expectedEventsArray,\n );\n },\n thenNothingHappened: (): void => {\n const resultEvents = handle();\n\n const resultEventsArray = Array.isArray(resultEvents)\n ? resultEvents\n : [resultEvents];\n\n assertThatArray(resultEventsArray).isEmpty();\n },\n thenThrows: <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): void => {\n try {\n handle();\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n }\n },\n};\n","import type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n EventStore,\n EventStoreReadEventMetadata,\n ReadStreamOptions,\n ReadStreamResult,\n StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport { type Event, type EventMetaDataOf } from '../typing';\n\nexport type TestEventStream<EventType extends Event = Event> = [\n string,\n EventType[],\n];\n\nexport type EventStoreWrapper<Store extends EventStore> = Store & {\n appendedEvents: Map<string, TestEventStream>;\n setup<EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>>;\n};\n\nexport const WrapEventStore = <Store extends EventStore>(\n eventStore: Store,\n): EventStoreWrapper<Store> => {\n const appendedEvents = new Map<string, TestEventStream>();\n\n const wrapped = {\n ...eventStore,\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<\n AggregateStreamResult<State, StreamPositionTypeOfEventStore<Store>>\n > {\n return eventStore.aggregateStream(streamName, options);\n },\n\n async readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<\n ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >\n > {\n return (await eventStore.readStream(\n streamName,\n options,\n )) as ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >;\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n const result = await eventStore.appendToStream(\n streamName,\n events,\n options,\n );\n\n const currentStream = appendedEvents.get(streamName) ?? [streamName, []];\n\n appendedEvents.set(streamName, [\n streamName,\n [...currentStream[1], ...events],\n ]);\n\n return result;\n },\n\n appendedEvents,\n\n setup: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n return eventStore.appendToStream(streamName, events);\n },\n\n // streamEvents: (): ReadableStream<\n // // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // > => {\n // return eventStore.streamEvents();\n // },\n };\n\n return wrapped as EventStoreWrapper<Store>;\n};\n","import {\n type NodePostgresClient,\n type NodePostgresTransaction,\n type SQL,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n projection,\n type CanHandle,\n type Event,\n type ProjectionHandler,\n type ReadEvent,\n type TypedProjectionDefinition,\n} from '@event-driven-io/emmett';\nimport type { PostgresReadEventMetadata } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionHandlerContext = {\n connectionString: string;\n client: NodePostgresClient;\n execute: SQLExecutor;\n transaction: NodePostgresTransaction;\n};\n\nexport type PostgreSQLProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext\n>;\n\nexport type PostgreSQLProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext\n>;\n\nexport type ProjectionHandlerOptions<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n events: ReadEvent<EventType, EventMetaDataType>[];\n projections: PostgreSQLProjectionDefinition<EventType, EventMetaDataType>[];\n connection: {\n connectionString: string;\n transaction: NodePostgresTransaction;\n };\n};\n\nexport const handleProjections = async <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: ProjectionHandlerOptions<EventType, EventMetaDataType>,\n): Promise<void> => {\n const {\n projections: allProjections,\n events,\n connection: { transaction, connectionString },\n } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n const client = (await transaction.connection.open()) as NodePostgresClient;\n\n for (const projection of projections) {\n await projection.handle(events, {\n connectionString,\n client,\n transaction,\n execute: transaction.execute,\n });\n }\n};\n\nexport const postgreSQLProjection = <\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n definition: PostgreSQLProjectionDefinition<EventType, EventMetaDataType>,\n): PostgreSQLProjectionDefinition =>\n projection<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext,\n PostgreSQLProjectionDefinition<EventType, EventMetaDataType>\n >(definition) as PostgreSQLProjectionDefinition;\n\nexport const postgreSQLRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL[]> | SQL[],\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: SQL[] = await handle(events, context);\n\n await context.execute.batchCommand(sqls);\n },\n });\n\nexport const postgreSQLRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL> | SQL,\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: SQL[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n\nexport * from './pongo';\nexport * from './postgresProjectionSpec';\n","import { type Dumbo } from '@event-driven-io/dumbo';\nimport {\n assertDeepEqual,\n assertEqual,\n assertIsNotNull,\n assertThatArray,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoCollection,\n type PongoDocument,\n type PongoFilter,\n type WithId,\n} from '@event-driven-io/pongo';\nimport { type PostgreSQLProjectionAssert } from '..';\n\nexport type PongoAssertOptions = {\n inCollection: string;\n inDatabase?: string;\n};\n\nconst withCollection = (\n handle: (collection: PongoCollection<PongoDocument>) => Promise<void>,\n options: {\n pool: Dumbo;\n connectionString: string;\n } & PongoAssertOptions,\n) => {\n const { pool, connectionString, inDatabase, inCollection } = options;\n\n return pool.withConnection(async (connection) => {\n const pongo = pongoClient(connectionString, {\n connectionOptions: { connection },\n });\n try {\n const collection = pongo.db(inDatabase).collection(inCollection);\n\n return handle(collection);\n } finally {\n await pongo.close();\n }\n });\n};\n\nconst withoutIdAndVersion = <Doc extends PongoDocument | WithId<PongoDocument>>(\n doc: Doc,\n) => {\n const { _id, _version, ...without } = doc;\n\n return without;\n};\n\nconst assertDocumentsEqual = <\n Doc extends PongoDocument | WithId<PongoDocument>,\n>(\n actual: PongoDocument,\n expected: Doc,\n) => {\n if ('_id' in expected)\n assertEqual(\n expected._id,\n actual._id,\n // eslint-disable-next-line @typescript-eslint/restrict-template-expressions\n `Document ids are not matching! Expected: ${expected._id}, Actual: ${actual._id}`,\n );\n\n return assertDeepEqual(\n withoutIdAndVersion(actual),\n withoutIdAndVersion(expected),\n );\n};\n\ntype FilterOrId<Doc extends PongoDocument | WithId<PongoDocument>> =\n | { withId: string }\n | {\n matchingFilter: PongoFilter<Doc>;\n };\n\nexport const documentExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n\n assertDocumentsEqual(result, document);\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsAreTheSame =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n documents: Doc[],\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n documents.length,\n result.length,\n 'Different Documents Count than expected',\n );\n\n for (let i = 0; i < documents.length; i++) {\n assertThatArray(result as Doc[]).contains(documents[i]!);\n }\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsMatchingHaveCount =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n expectedCount: number,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n expectedCount,\n result.length,\n 'Different Documents Count than expected',\n );\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentMatchingExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertThatArray(result).isNotEmpty();\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentDoesNotExist =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n },\n { ...options, ...assertOptions },\n );\n\nexport const expectPongoDocuments = {\n fromCollection: <Doc extends PongoDocument | WithId<PongoDocument>>(\n collectionName: string,\n ) => {\n return {\n withId: (id: string) => {\n return {\n toBeEqual: (document: Doc) =>\n documentExists(document, {\n withId: id,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n withId: id,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n withId: id,\n inCollection: collectionName,\n }),\n };\n },\n matching: <Doc extends PongoDocument | WithId<PongoDocument>>(\n filter: PongoFilter<Doc>,\n ) => {\n return {\n toBeTheSame: (documents: Doc[]) =>\n documentsAreTheSame<Doc>(documents, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toHaveCount: (expectedCount: number) =>\n documentsMatchingHaveCount(expectedCount, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n };\n },\n };\n },\n};\n","import {\n type CanHandle,\n type Event,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoClient,\n type PongoDocument,\n} from '@event-driven-io/pongo';\nimport {\n postgreSQLProjection,\n type PostgreSQLProjectionDefinition,\n type PostgreSQLProjectionHandlerContext,\n} from '..';\nimport type { PostgresReadEventMetadata } from '../../postgreSQLEventStore';\n\nexport type PongoProjectionHandlerContext =\n PostgreSQLProjectionHandlerContext & {\n pongo: PongoClient;\n };\n\nexport type PongoWithNotNullDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | ((\n document: Document,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Document | null)\n | ((\n document: Document,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoWithNullableDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | ((\n document: Document | null,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Document | null)\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | PongoWithNotNullDocumentEvolve<Document, EventType, EventMetaDataType>\n | PongoWithNullableDocumentEvolve<Document, EventType, EventMetaDataType>;\n\nexport type PongoProjectionOptions<\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n handle: (\n events: ReadEvent<EventType, EventMetaDataType>[],\n context: PongoProjectionHandlerContext,\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoProjection = <\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>({\n handle,\n canHandle,\n}: PongoProjectionOptions<\n EventType,\n EventMetaDataType\n>): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType, EventMetaDataType>({\n canHandle,\n handle: async (events, context) => {\n const { connectionString, client } = context;\n const pongo = pongoClient(connectionString, {\n connectionOptions: { client },\n });\n await handle(events, {\n ...context,\n pongo,\n });\n },\n });\n\nexport type PongoMultiStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n canHandle: CanHandle<EventType>;\n\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n} & (\n | {\n evolve: PongoWithNullableDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: PongoWithNotNullDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Document;\n }\n);\n\nexport const pongoMultiStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: PongoMultiStreamProjectionOptions<\n Document,\n EventType,\n EventMetaDataType\n >,\n): PostgreSQLProjectionDefinition => {\n const { collectionName, getDocumentId, canHandle } = options;\n\n return pongoProjection({\n handle: async (events, { pongo }) => {\n const collection = pongo.db().collection<Document>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), async (document) => {\n return 'initialState' in options\n ? await options.evolve(\n document ?? options.initialState(),\n event as ReadEvent<EventType, EventMetaDataType>,\n )\n : await options.evolve(\n document,\n event as ReadEvent<EventType, EventMetaDataType>,\n );\n });\n }\n },\n canHandle,\n });\n};\n\nexport type PongoSingleStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n canHandle: CanHandle<EventType>;\n\n collectionName: string;\n} & (\n | {\n evolve: PongoWithNullableDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: PongoWithNotNullDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Document;\n }\n);\n\nexport const pongoSingleStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: PongoSingleStreamProjectionOptions<\n Document,\n EventType,\n EventMetaDataType\n >,\n): PostgreSQLProjectionDefinition => {\n return pongoMultiStreamProjection<Document, EventType, EventMetaDataType>({\n ...options,\n getDocumentId: (event) => event.metadata.streamName,\n });\n};\n","import {\n dumbo,\n type Dumbo,\n type DumboOptions,\n type QueryResultRow,\n type SQL,\n} from '@event-driven-io/dumbo';\nimport {\n assertFails,\n AssertionError,\n assertThatArray,\n assertTrue,\n isErrorConstructor,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { handleProjections, type PostgreSQLProjectionDefinition } from '.';\nimport type { PostgresReadEventMetadata } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type PostgreSQLProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type PostgreSQLProjectionSpec<EventType extends Event> = (\n givenEvents: PostgreSQLProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n options?: PostgreSQLProjectionSpecWhenOptions,\n ) => {\n then: (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type PostgreSQLProjectionAssert = (options: {\n pool: Dumbo;\n connectionString: string;\n}) => Promise<void | boolean>;\n\nexport type PostgreSQLProjectionSpecOptions = {\n projection: PostgreSQLProjectionDefinition;\n} & DumboOptions;\n\nexport const PostgreSQLProjectionSpec = {\n for: <EventType extends Event>(\n options: PostgreSQLProjectionSpecOptions,\n ): PostgreSQLProjectionSpec<EventType> => {\n {\n const { projection, ...dumoOptions } = options;\n const { connectionString } = dumoOptions;\n\n return (givenEvents: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n options?: PostgreSQLProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, PostgresReadEventMetadata>[] =\n [];\n\n const run = async (pool: Dumbo) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: PostgresReadEventMetadata = {\n globalPosition: ++globalPosition,\n streamPosition: globalPosition,\n streamName: `test-${uuid()}`,\n eventId: uuid(),\n };\n\n allEvents.push({\n ...event,\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n PostgresReadEventMetadata\n >,\n });\n }\n\n await pool.withTransaction((transaction) =>\n handleProjections({\n events: allEvents,\n projections: [projection],\n connection: {\n connectionString,\n transaction,\n },\n }),\n );\n };\n\n return {\n then: async (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n\n const succeeded = await assert({ pool, connectionString });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await pool.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await pool.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n streamName: string,\n event: PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>,\n): PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n streamName: string,\n events: PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>[],\n): PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\nexport const assertSQLQueryResultMatches =\n <T extends QueryResultRow>(sql: SQL, rows: T[]): PostgreSQLProjectionAssert =>\n async ({ pool: { execute } }) => {\n const result = await execute.query<T>(sql);\n\n assertThatArray(rows).containsExactlyInAnyOrder(result.rows);\n };\n\nexport const expectSQL = {\n query: (sql: SQL) => ({\n resultRows: {\n toBeTheSame: <T extends QueryResultRow>(rows: T[]) =>\n assertSQLQueryResultMatches(sql, rows),\n },\n }),\n};\n","import { type NodePostgresPool, type SQL } from '@event-driven-io/dumbo';\nimport { appendEventsSQL } from './appendToStream';\nimport { storeSubscriptionCheckpointSQL } from './storeSubscriptionCheckpoint';\nimport {\n addDefaultPartition,\n addEventsPartitions,\n addModuleForAllTenantsSQL,\n addModuleSQL,\n addTablePartitions,\n addTenantForAllModulesSQL,\n addTenantSQL,\n eventsTableSQL,\n sanitizeNameSQL,\n streamsTableSQL,\n subscriptionsTableSQL,\n} from './tables';\n\nexport * from './appendToStream';\nexport * from './readLastMessageGlobalPosition';\nexport * from './readMessagesBatch';\nexport * from './readStream';\nexport * from './readSubscriptionCheckpoint';\nexport * from './storeSubscriptionCheckpoint';\nexport * from './tables';\nexport * from './typing';\n\nexport const schemaSQL: SQL[] = [\n streamsTableSQL,\n eventsTableSQL,\n subscriptionsTableSQL,\n sanitizeNameSQL,\n addTablePartitions,\n addEventsPartitions,\n addModuleSQL,\n addTenantSQL,\n addModuleForAllTenantsSQL,\n addTenantForAllModulesSQL,\n appendEventsSQL,\n addDefaultPartition,\n storeSubscriptionCheckpointSQL,\n];\n\nexport const createEventStoreSchema = async (\n pool: NodePostgresPool,\n): Promise<void> => {\n await pool.withTransaction(({ execute }) => execute.batchCommand(schemaSQL));\n};\n","import {\n rawSql,\n single,\n sql,\n type NodePostgresPool,\n type NodePostgresTransaction,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n JSONParser,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AppendToStreamOptions,\n type Event,\n type ExpectedStreamVersion,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { defaultTag, eventsTable, streamsTable } from './typing';\n\nexport const appendEventsSQL = rawSql(\n `CREATE OR REPLACE FUNCTION emt_append_event(\n v_event_ids text[],\n v_events_data jsonb[],\n v_events_metadata jsonb[],\n v_event_schema_versions text[],\n v_event_types text[],\n v_stream_id text,\n v_stream_type text,\n v_expected_stream_position bigint DEFAULT NULL,\n v_partition text DEFAULT emt_sanitize_name('default_partition')\n ) RETURNS TABLE (\n success boolean,\n next_stream_position bigint,\n last_global_position bigint,\n transaction_id xid8\n ) LANGUAGE plpgsql\n AS $$\n DECLARE\n v_next_stream_position bigint;\n v_position bigint;\n v_updated_rows int;\n v_transaction_id xid8;\n v_last_global_position bigint;\n BEGIN\n IF v_expected_stream_position IS NULL THEN\n SELECT COALESCE(max(stream_position), 0) INTO v_expected_stream_position\n FROM ${streamsTable.name}\n WHERE stream_id = v_stream_id AND partition = v_partition;\n END IF;\n\n v_next_stream_position := v_expected_stream_position + array_upper(v_events_data, 1);\n v_transaction_id := pg_current_xact_id();\n\n WITH ev AS (\n SELECT row_number() OVER () + v_expected_stream_position AS stream_position, \n event_data, \n event_metadata, \n schema_version, \n event_id, \n event_type\n FROM (\n SELECT *\n FROM \n unnest(v_event_ids, v_events_data, v_events_metadata, v_event_schema_versions, v_event_types) \n AS event(event_id, event_data, event_metadata, schema_version, event_type)\n ) AS event\n ),\n all_events_insert AS (\n INSERT INTO ${eventsTable.name}\n (stream_id, stream_position, partition, event_data, event_metadata, event_schema_version, event_type, event_id, transaction_id)\n SELECT \n v_stream_id, ev.stream_position, v_partition, ev.event_data, ev.event_metadata, ev.schema_version, ev.event_type, ev.event_id, v_transaction_id\n FROM ev\n RETURNING global_position\n )\n SELECT \n max(global_position) INTO v_last_global_position \n FROM \n all_events_insert;\n\n\n IF v_expected_stream_position = 0 THEN\n INSERT INTO ${streamsTable.name}\n (stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)\n VALUES\n (v_stream_id, v_next_stream_position, v_partition, v_stream_type, '{}', FALSE);\n ELSE\n UPDATE ${streamsTable.name} as s \n SET stream_position = v_next_stream_position\n WHERE stream_id = v_stream_id AND stream_position = v_expected_stream_position AND partition = v_partition AND is_archived = FALSE;\n\n get diagnostics v_updated_rows = row_count;\n\n IF v_updated_rows = 0 THEN\n RETURN QUERY SELECT FALSE, NULL::bigint, NULL::bigint, NULL::xid8;\n RETURN;\n END IF;\n END IF;\n\n RETURN QUERY SELECT TRUE, v_next_stream_position, v_last_global_position, v_transaction_id;\n END;\n $$;\n `,\n);\n\ntype AppendEventResult =\n | {\n success: true;\n nextStreamPosition: bigint;\n lastGlobalPosition: bigint;\n transactionId: string;\n }\n | { success: false };\n\nexport type AppendToStreamPreCommitHook = (\n events: ReadEvent[],\n context: {\n transaction: NodePostgresTransaction;\n },\n) => Promise<void>;\n\nexport const appendToStream = (\n pool: NodePostgresPool,\n streamName: string,\n streamType: string,\n events: Event[],\n options?: AppendToStreamOptions & {\n partition?: string;\n preCommitHook?: AppendToStreamPreCommitHook;\n },\n): Promise<AppendEventResult> =>\n pool.withTransaction<AppendEventResult>(async (transaction) => {\n const { execute } = transaction;\n\n if (events.length === 0)\n return { success: false, result: { success: false } };\n\n let appendResult: AppendEventSqlResult;\n\n try {\n const expectedStreamVersion = toExpectedVersion(\n options?.expectedStreamVersion,\n );\n\n const eventsToAppend: ReadEvent[] = events.map((e, i) => ({\n ...e,\n metadata: {\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(i),\n ...('metadata' in e ? (e.metadata ?? {}) : {}),\n },\n }));\n\n // TODO: return global positions from append raw and other generated data\n appendResult = await appendEventsRaw(\n execute,\n streamName,\n streamType,\n eventsToAppend,\n {\n expectedStreamVersion,\n },\n );\n\n if (options?.preCommitHook)\n await options.preCommitHook(eventsToAppend, { transaction });\n } catch (error) {\n if (!isOptimisticConcurrencyError(error)) throw error;\n\n appendResult = {\n success: false,\n last_global_position: null,\n next_stream_position: null,\n transaction_id: null,\n };\n }\n\n const {\n success,\n next_stream_position,\n last_global_position,\n transaction_id,\n } = appendResult;\n\n return {\n success,\n result:\n success &&\n next_stream_position &&\n last_global_position &&\n transaction_id\n ? {\n success: true,\n nextStreamPosition: BigInt(next_stream_position),\n lastGlobalPosition: BigInt(last_global_position),\n transactionId: transaction_id,\n }\n : { success: false },\n };\n });\n\nconst toExpectedVersion = (\n expected: ExpectedStreamVersion | undefined,\n): bigint | null => {\n if (expected === undefined) return null;\n\n if (expected === NO_CONCURRENCY_CHECK) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_DOES_NOT_EXIST) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_EXISTS) return null;\n\n return expected as bigint;\n};\n\nconst isOptimisticConcurrencyError = (error: unknown): boolean =>\n error instanceof Error && 'code' in error && error.code === '23505';\n\ntype AppendEventSqlResult = {\n success: boolean;\n next_stream_position: string | null;\n last_global_position: string | null;\n transaction_id: string | null | undefined;\n};\n\nconst appendEventsRaw = (\n execute: SQLExecutor,\n streamId: string,\n streamType: string,\n events: ReadEvent[],\n options?: {\n expectedStreamVersion: bigint | null;\n partition?: string;\n },\n): Promise<AppendEventSqlResult> =>\n single(\n execute.command<AppendEventSqlResult>(\n sql(\n `SELECT * FROM emt_append_event(\n ARRAY[%s]::text[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::text[],\n ARRAY[%s]::text[],\n %L::text,\n %L::text,\n %s::bigint,\n %L::text\n )`,\n events.map((e) => sql('%L', e.metadata.eventId)).join(','),\n events.map((e) => sql('%L', JSONParser.stringify(e.data))).join(','),\n events\n .map((e) => sql('%L', JSONParser.stringify(e.metadata ?? {})))\n .join(','),\n events.map(() => `'1'`).join(','),\n events.map((e) => sql('%L', e.type)).join(','),\n streamId,\n streamType,\n options?.expectedStreamVersion ?? 'NULL',\n options?.partition ?? defaultTag,\n ),\n ),\n );\n","export const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = 'emt:default';\n\nexport const globalNames = {\n module: `${emmettPrefix}:module:${globalTag}`,\n tenant: `${emmettPrefix}:tenant:${globalTag}`,\n};\n\nconst columns = {\n partition: {\n name: 'partition',\n },\n isArchived: { name: 'is_archived' },\n};\n\nexport const streamsTable = {\n name: `${emmettPrefix}_streams`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const eventsTable = {\n name: `${emmettPrefix}_events`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const subscriptionsTable = {\n name: `${emmettPrefix}_subscriptions`,\n};\n","import { single, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, subscriptionsTable } from './typing';\n\nexport const storeSubscriptionCheckpointSQL = sql(`\nCREATE OR REPLACE FUNCTION store_subscription_checkpoint(\n p_subscription_id VARCHAR(100),\n p_version BIGINT,\n p_position BIGINT,\n p_check_position BIGINT,\n p_transaction_id xid8,\n p_partition TEXT DEFAULT '${defaultTag}'\n) RETURNS INT AS $$\nDECLARE\n current_position BIGINT;\nBEGIN\n -- Handle the case when p_check_position is provided\n IF p_check_position IS NOT NULL THEN\n -- Try to update if the position matches p_check_position\n UPDATE \"${subscriptionsTable.name}\"\n SET \n \"last_processed_position\" = p_position, \n \"last_processed_transaction_id\" = p_transaction_id\n WHERE \"subscription_id\" = p_subscription_id AND \"last_processed_position\" = p_check_position AND \"partition\" = p_partition;\n\n IF FOUND THEN\n RETURN 1; -- Successfully updated\n END IF;\n\n -- Retrieve the current position\n SELECT \"last_processed_position\" INTO current_position\n FROM \"${subscriptionsTable.name}\"\n WHERE \"subscription_id\" = p_subscription_id AND \"partition\" = p_partition;\n\n -- Return appropriate codes based on current position\n IF current_position = p_position THEN\n RETURN 0; -- Idempotent check: position already set\n ELSIF current_position > p_check_position THEN\n RETURN 2; -- Failure: current position is greater\n ELSE\n RETURN 2; -- Default failure case for mismatched positions\n END IF;\n END IF;\n\n -- Handle the case when p_check_position is NULL: Insert if not exists\n BEGIN\n INSERT INTO \"${subscriptionsTable.name}\"(\"subscription_id\", \"version\", \"last_processed_position\", \"partition\", \"last_processed_transaction_id\")\n VALUES (p_subscription_id, p_version, p_position, p_partition, p_transaction_id);\n RETURN 1; -- Successfully inserted\n EXCEPTION WHEN unique_violation THEN\n -- If insertion failed, it means the row already exists\n SELECT \"last_processed_position\" INTO current_position\n FROM \"${subscriptionsTable.name}\"\n WHERE \"subscription_id\" = p_subscription_id AND \"partition\" = p_partition;\n\n IF current_position = p_position THEN\n RETURN 0; -- Idempotent check: position already set\n ELSE\n RETURN 2; -- Insertion failed, row already exists with different position\n END IF;\n END;\nEND;\n$$ LANGUAGE plpgsql;\n`);\n\nexport type StoreLastProcessedSubscriptionPositionResult<\n Position extends bigint | null = bigint,\n> =\n | {\n success: true;\n newPosition: Position;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint | null>>;\nexport function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint>>;\nexport async function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint | null>> {\n try {\n const { result } = await single(\n execute.command<{ result: 0 | 1 | 2 }>(\n sql(\n `SELECT store_subscription_checkpoint(%L, %s, %L, %L, pg_current_xact_id(), %L) as result;`,\n options.subscriptionId,\n options.version ?? 1,\n options.newPosition,\n options.lastProcessedPosition,\n options.partition ?? defaultTag,\n ),\n ),\n );\n\n return result === 1\n ? { success: true, newPosition: options.newPosition }\n : { success: false, reason: result === 0 ? 'IGNORED' : 'MISMATCH' };\n } catch (error) {\n console.log(error);\n throw error;\n }\n}\n","import { rawSql } from '@event-driven-io/dumbo';\nimport {\n defaultTag,\n eventsTable,\n globalTag,\n streamsTable,\n subscriptionsTable,\n} from './typing';\n\nexport const streamsTableSQL = rawSql(\n `CREATE TABLE IF NOT EXISTS ${streamsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}__${globalTag}',\n stream_type TEXT NOT NULL,\n stream_metadata JSONB NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n PRIMARY KEY (stream_id, stream_position, partition, is_archived),\n UNIQUE (stream_id, partition, is_archived)\n ) PARTITION BY LIST (partition);`,\n);\n\nexport const eventsTableSQL = rawSql(\n `\n CREATE SEQUENCE IF NOT EXISTS emt_global_event_position;\n\n CREATE TABLE IF NOT EXISTS ${eventsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n event_data JSONB NOT NULL,\n event_metadata JSONB NOT NULL,\n event_schema_version TEXT NOT NULL,\n event_type TEXT NOT NULL,\n event_id TEXT NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n global_position BIGINT DEFAULT nextval('emt_global_event_position'),\n transaction_id XID8 NOT NULL,\n created TIMESTAMPTZ NOT NULL DEFAULT now(),\n PRIMARY KEY (stream_id, stream_position, partition, is_archived)\n ) PARTITION BY LIST (partition);`,\n);\n\nexport const subscriptionsTableSQL = rawSql(\n `\n CREATE TABLE IF NOT EXISTS ${subscriptionsTable.name}(\n subscription_id TEXT NOT NULL,\n version INT NOT NULL DEFAULT 1,\n partition TEXT NOT NULL DEFAULT '${globalTag}__${globalTag}',\n last_processed_position BIGINT NOT NULL,\n last_processed_transaction_id XID8 NOT NULL,\n PRIMARY KEY (subscription_id, partition, version)\n ) PARTITION BY LIST (partition);\n`,\n);\n\nexport const sanitizeNameSQL = rawSql(\n `CREATE OR REPLACE FUNCTION emt_sanitize_name(input_name TEXT) RETURNS TEXT AS $$\n BEGIN\n RETURN REGEXP_REPLACE(input_name, '[^a-zA-Z0-9_]', '_', 'g');\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addTablePartitions = rawSql(\n `\n CREATE OR REPLACE FUNCTION emt_add_table_partition(tableName TEXT, partition_name TEXT) RETURNS void AS $$\n DECLARE\n v_main_partiton_name TEXT;\n v_active_partiton_name TEXT;\n v_archived_partiton_name TEXT;\n BEGIN \n v_main_partiton_name := emt_sanitize_name(tableName || '_' || partition_name);\n v_active_partiton_name := emt_sanitize_name(v_main_partiton_name || '_active');\n v_archived_partiton_name := emt_sanitize_name(v_main_partiton_name || '_archived');\n\n\n -- create default partition\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (%L) PARTITION BY LIST (is_archived);',\n v_main_partiton_name, tableName, partition_name\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (FALSE);',\n v_active_partiton_name, v_main_partiton_name\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (TRUE);',\n v_archived_partiton_name, v_main_partiton_name\n );\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addEventsPartitions = rawSql(\n `\n CREATE OR REPLACE FUNCTION emt_add_partition(partition_name TEXT) RETURNS void AS $$\n BEGIN \n PERFORM emt_add_table_partition('${eventsTable.name}', partition_name);\n PERFORM emt_add_table_partition('${streamsTable.name}', partition_name);\n\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (%L);',\n emt_sanitize_name('${subscriptionsTable.name}' || '_' || partition_name), '${subscriptionsTable.name}', partition_name\n );\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addModuleSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_module(new_module TEXT) RETURNS void AS $$\n BEGIN\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}'), '${eventsTable.name}', new_module, '${globalTag}'\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}' || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}' || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}'), '${streamsTable.name}', new_module, '${globalTag}'\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}' || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}' || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addTenantSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_tenant(new_module TEXT, new_tenant TEXT) RETURNS void AS $$\n BEGIN\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant), '${eventsTable.name}', new_module, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant), '${streamsTable.name}', new_module, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant)\n );\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addModuleForAllTenantsSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_module_for_all_tenants(new_module TEXT) RETURNS void AS $$\n DECLARE\n tenant_record RECORD;\n BEGIN\n PERFORM add_module(new_module);\n \n FOR tenant_record IN SELECT DISTINCT tenant FROM ${eventsTable.name}\n LOOP\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant), '${eventsTable.name}', new_module, tenant_record.tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant), '${streamsTable.name}', new_module, tenant_record.tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n END LOOP;\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addTenantForAllModulesSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_tenant_for_all_modules(new_tenant TEXT) RETURNS void AS $$\n DECLARE\n module_record RECORD;\n BEGIN\n FOR module_record IN SELECT DISTINCT partitionname FROM pg_partman.part_config WHERE parent_table = '${eventsTable.name}'\n LOOP\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant), '${eventsTable.name}', module_record.partitionname, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant), '${streamsTable.name}', module_record.partitionname, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n END LOOP;\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addDefaultPartition = rawSql(\n `SELECT emt_add_partition('${defaultTag}');`,\n);\n","import { singleOrNull, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadLastMessageGlobalPositionSqlResult = {\n global_position: string;\n};\n\nexport type ReadLastMessageGlobalPositionResult = {\n currentGlobalPosition: bigint | null;\n};\n\nexport const readLastMessageGlobalPosition = async (\n execute: SQLExecutor,\n options?: { partition?: string },\n): Promise<ReadLastMessageGlobalPositionResult> => {\n const result = await singleOrNull(\n execute.query<ReadLastMessageGlobalPositionSqlResult>(\n sql(\n `SELECT global_position\n FROM ${eventsTable.name}\n WHERE partition = %L AND is_archived = FALSE AND transaction_id < pg_snapshot_xmin(pg_current_snapshot())\n ORDER BY transaction_id, global_position\n LIMIT 1`,\n options?.partition ?? defaultTag,\n ),\n ),\n );\n\n return {\n currentGlobalPosition:\n result !== null ? BigInt(result.global_position) : null,\n };\n};\n","import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n type CombinedReadEventMetadata,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadata,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadMessagesBatchSqlResult<EventType extends Event> = {\n stream_position: string;\n stream_id: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport type ReadMessagesBatchOptions =\n | {\n after: bigint;\n batchSize: number;\n }\n | {\n from: bigint;\n batchSize: number;\n }\n | { to: bigint; batchSize: number }\n | { from: bigint; to: bigint };\n\nexport type ReadMessagesBatchResult<\n EventType extends Event,\n ReadEventMetadataType extends ReadEventMetadata = ReadEventMetadata,\n> = {\n currentGlobalPosition: bigint;\n messages: ReadEvent<EventType, ReadEventMetadataType>[];\n areEventsLeft: boolean;\n};\n\nexport const readMessagesBatch = async <\n MessageType extends Event,\n ReadEventMetadataType extends\n ReadEventMetadataWithGlobalPosition = ReadEventMetadataWithGlobalPosition,\n>(\n execute: SQLExecutor,\n options: ReadMessagesBatchOptions & { partition?: string },\n): Promise<ReadMessagesBatchResult<MessageType, ReadEventMetadataType>> => {\n const from =\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after + 1n\n : 0n;\n const batchSize =\n options && 'batchSize' in options\n ? options.batchSize\n : options.to - options.from;\n\n const fromCondition: string =\n from !== -0n ? `AND global_position >= ${from}` : '';\n\n const toCondition =\n 'to' in options ? `AND global_position <= ${options.to}` : '';\n\n const limitCondition =\n 'batchSize' in options ? `LIMIT ${options.batchSize}` : '';\n\n const events: ReadEvent<MessageType, ReadEventMetadataType>[] = await mapRows(\n execute.query<ReadMessagesBatchSqlResult<MessageType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE partition = %L AND is_archived = FALSE AND transaction_id < pg_snapshot_xmin(pg_current_snapshot()) ${fromCondition} ${toCondition}\n ORDER BY transaction_id, global_position\n ${limitCondition}`,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = {\n type: row.event_type,\n data: row.event_data,\n metadata: row.event_metadata,\n } as unknown as MessageType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n eventId: row.event_id,\n streamName: row.stream_id,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n metadata: metadata as CombinedReadEventMetadata<\n MessageType,\n ReadEventMetadataType\n >,\n };\n },\n );\n\n return events.length > 0\n ? {\n currentGlobalPosition:\n events[events.length - 1]!.metadata.globalPosition,\n messages: events,\n areEventsLeft: events.length === batchSize,\n }\n : {\n currentGlobalPosition:\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after\n : 0n,\n messages: [],\n areEventsLeft: false,\n };\n};\n","import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n type CombinedReadEventMetadata,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { PostgreSQLEventStoreDefaultStreamVersion } from '../postgreSQLEventStore';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadStreamSqlResult<EventType extends Event> = {\n stream_position: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n execute: SQLExecutor,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const events: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n await mapRows(\n execute.query<ReadStreamSqlResult<EventType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${fromCondition} ${toCondition}`,\n streamId,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = {\n type: row.event_type,\n data: row.event_data,\n metadata: row.event_metadata,\n } as unknown as EventType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n eventId: row.event_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n metadata: metadata as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n },\n );\n\n return events.length > 0\n ? {\n currentStreamVersion:\n events[events.length - 1]!.metadata.streamPosition,\n events,\n streamExists: true,\n }\n : {\n currentStreamVersion: PostgreSQLEventStoreDefaultStreamVersion,\n events: [],\n streamExists: false,\n };\n};\n","import { singleOrNull, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, subscriptionsTable } from './typing';\n\ntype ReadSubscriptionCheckpointSqlResult = {\n last_processed_position: string;\n};\n\nexport type ReadSubscriptionCheckpointResult = {\n lastProcessedPosition: bigint | null;\n};\n\nexport const readSubscriptionCheckpoint = async (\n execute: SQLExecutor,\n options: { subscriptionId: string; partition?: string },\n): Promise<ReadSubscriptionCheckpointResult> => {\n const result = await singleOrNull(\n execute.query<ReadSubscriptionCheckpointSqlResult>(\n sql(\n `SELECT last_processed_position\n FROM ${subscriptionsTable.name}\n WHERE partition = %L AND subscription_id = %L\n LIMIT 1`,\n options?.partition ?? defaultTag,\n options.subscriptionId,\n ),\n ),\n );\n\n return {\n lastProcessedPosition:\n result !== null ? BigInt(result.last_processed_position) : null,\n };\n};\n","import { type SQLExecutor } from '@event-driven-io/dumbo';\nimport type {\n EmmettError,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { readLastMessageGlobalPosition } from '../../schema/readLastMessageGlobalPosition';\nimport {\n readMessagesBatch,\n type ReadMessagesBatchOptions,\n} from '../../schema/readMessagesBatch';\n\nexport const DefaultPostgreSQLEventStoreSubscriptionBatchSize = 100;\nexport const DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs = 50;\n\nexport type PostgreSQLEventStoreMessagesBatch<EventType extends Event = Event> =\n {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n };\n\nexport type PostgreSQLEventStoreMessagesBatchHandlerResult = void | {\n type: 'STOP';\n reason?: string;\n error?: EmmettError;\n};\n\nexport type PostgreSQLEventStoreMessagesBatchHandler<\n EventType extends Event = Event,\n> = (\n messagesBatch: PostgreSQLEventStoreMessagesBatch<EventType>,\n) =>\n | Promise<PostgreSQLEventStoreMessagesBatchHandlerResult>\n | PostgreSQLEventStoreMessagesBatchHandlerResult;\n\nexport type PostgreSQLEventStoreMessageBatchPullerOptions<\n EventType extends Event = Event,\n> = {\n executor: SQLExecutor;\n pullingFrequencyInMs: number;\n batchSize: number;\n eachBatch: PostgreSQLEventStoreMessagesBatchHandler<EventType>;\n};\n\nexport type PostgreSQLEventStoreMessageBatchPullerStartFrom =\n | { globalPosition: bigint }\n | 'BEGINNING'\n | 'END';\n\nexport type PostgreSQLEventStoreMessageBatchPullerStartOptions = {\n startFrom: PostgreSQLEventStoreMessageBatchPullerStartFrom;\n};\n\nexport type PostgreSQLEventStoreMessageBatchPuller = {\n isRunning: boolean;\n start(\n options: PostgreSQLEventStoreMessageBatchPullerStartOptions,\n ): Promise<void>;\n stop(): Promise<void>;\n};\n\nexport const postgreSQLEventStoreMessageBatchPuller = <\n EventType extends Event = Event,\n>({\n executor,\n batchSize,\n eachBatch,\n pullingFrequencyInMs,\n}: PostgreSQLEventStoreMessageBatchPullerOptions<EventType>): PostgreSQLEventStoreMessageBatchPuller => {\n let isRunning = false;\n\n let start: Promise<void>;\n\n const pullMessages = async (\n options: PostgreSQLEventStoreMessageBatchPullerStartOptions,\n ) => {\n const after =\n options.startFrom === 'BEGINNING'\n ? 0n\n : options.startFrom === 'END'\n ? ((await readLastMessageGlobalPosition(executor))\n .currentGlobalPosition ?? 0n)\n : options.startFrom.globalPosition;\n\n const readMessagesOptions: ReadMessagesBatchOptions = {\n after,\n batchSize,\n };\n\n let waitTime = 100;\n\n do {\n const { messages, currentGlobalPosition, areEventsLeft } =\n await readMessagesBatch<EventType>(executor, readMessagesOptions);\n\n if (messages.length > 0) {\n const result = await eachBatch({ messages });\n\n if (result && result.type === 'STOP') {\n isRunning = false;\n break;\n }\n }\n\n readMessagesOptions.after = currentGlobalPosition;\n\n await new Promise((resolve) => setTimeout(resolve, waitTime));\n\n if (!areEventsLeft) {\n waitTime = Math.min(waitTime * 2, 1000);\n } else {\n waitTime = pullingFrequencyInMs;\n }\n } while (isRunning);\n };\n\n return {\n get isRunning() {\n return isRunning;\n },\n start: (options) => {\n if (isRunning) return start;\n\n start = (async () => {\n isRunning = true;\n\n return pullMessages(options);\n })();\n\n return start;\n },\n stop: async () => {\n if (!isRunning) return;\n isRunning = false;\n await start;\n },\n };\n};\n\nexport const zipPostgreSQLEventStoreMessageBatchPullerStartFrom = (\n options: (PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined)[],\n): PostgreSQLEventStoreMessageBatchPullerStartFrom => {\n if (\n options.length === 0 ||\n options.some((o) => o === undefined || o === 'BEGINNING')\n )\n return 'BEGINNING';\n\n if (options.every((o) => o === 'END')) return 'END';\n\n return options\n .filter((o) => o !== undefined && o !== 'BEGINNING' && o !== 'END')\n .sort((a, b) => (a > b ? 1 : -1))[0]!;\n};\n","import { dumbo } from '@event-driven-io/dumbo';\nimport { EmmettError, type Event } from '@event-driven-io/emmett';\nimport {\n DefaultPostgreSQLEventStoreSubscriptionBatchSize,\n DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs,\n postgreSQLEventStoreMessageBatchPuller,\n zipPostgreSQLEventStoreMessageBatchPullerStartFrom,\n type PostgreSQLEventStoreMessageBatchPuller,\n type PostgreSQLEventStoreMessagesBatchHandler,\n} from './messageBatchProcessing';\nimport {\n postgreSQLEventStoreSubscription,\n type PostgreSQLEventStoreSubscription,\n type PostgreSQLEventStoreSubscriptionOptions,\n} from './postgreSQLEventStoreSubscription';\n\nexport type PostgreSQLEventStoreConsumerOptions = {\n connectionString: string;\n subscriptions?: PostgreSQLEventStoreSubscription[];\n pooling?: {\n batchSize?: number;\n pullingFrequencyInMs?: number;\n };\n};\n\nexport type PostgreSQLEventStoreConsumer = Readonly<{\n connectionString: string;\n isRunning: boolean;\n subscriptions: PostgreSQLEventStoreSubscription[];\n subscribe: <EventType extends Event = Event>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n ) => PostgreSQLEventStoreSubscription<EventType>;\n start: () => Promise<void>;\n stop: () => Promise<void>;\n close: () => Promise<void>;\n}>;\n\nexport const postgreSQLEventStoreConsumer = (\n options: PostgreSQLEventStoreConsumerOptions,\n): PostgreSQLEventStoreConsumer => {\n let isRunning = false;\n const { connectionString, pooling } = options;\n const subscriptions = options.subscriptions ?? [];\n\n let start: Promise<void>;\n\n let currentMessagePooler: PostgreSQLEventStoreMessageBatchPuller | undefined;\n\n const pool = dumbo({ connectionString });\n\n const eachBatch: PostgreSQLEventStoreMessagesBatchHandler = async (\n messagesBatch,\n ) => {\n const activeSubscriptions = subscriptions.filter((s) => s.isActive);\n\n if (activeSubscriptions.length === 0)\n return {\n type: 'STOP',\n reason: 'No active subscriptions',\n };\n\n const result = await Promise.allSettled(\n activeSubscriptions.map((s) => {\n // TODO: Add here filtering to only pass messages that can be handled by subscription\n return s.handle(messagesBatch, { pool });\n }),\n );\n\n return result.some(\n (r) => r.status === 'fulfilled' && r.value?.type !== 'STOP',\n )\n ? undefined\n : {\n type: 'STOP',\n };\n };\n\n const messagePooler = (currentMessagePooler =\n postgreSQLEventStoreMessageBatchPuller({\n executor: pool.execute,\n eachBatch,\n batchSize:\n pooling?.batchSize ?? DefaultPostgreSQLEventStoreSubscriptionBatchSize,\n pullingFrequencyInMs:\n pooling?.pullingFrequencyInMs ??\n DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs,\n }));\n\n const stop = async () => {\n if (!isRunning) return;\n isRunning = false;\n if (currentMessagePooler) {\n await currentMessagePooler.stop();\n currentMessagePooler = undefined;\n }\n await start;\n };\n\n return {\n connectionString,\n subscriptions,\n get isRunning() {\n return isRunning;\n },\n subscribe: <EventType extends Event = Event>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n ): PostgreSQLEventStoreSubscription<EventType> => {\n const subscription = postgreSQLEventStoreSubscription<EventType>(options);\n\n subscriptions.push(subscription);\n\n return subscription;\n },\n start: () => {\n if (isRunning) return start;\n\n start = (async () => {\n if (subscriptions.length === 0)\n return Promise.reject(\n new EmmettError(\n 'Cannot start consumer without at least a single subscription',\n ),\n );\n\n isRunning = true;\n\n const startFrom = zipPostgreSQLEventStoreMessageBatchPullerStartFrom(\n await Promise.all(subscriptions.map((o) => o.start(pool.execute))),\n );\n\n return messagePooler.start({ startFrom });\n })();\n\n return start;\n },\n stop,\n close: async () => {\n await stop();\n await pool.close();\n },\n };\n};\n","import { type Dumbo, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n EmmettError,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport {\n readSubscriptionCheckpoint,\n storeSubscriptionCheckpoint,\n} from '../schema';\nimport type { PostgreSQLEventStoreMessageBatchPullerStartFrom } from './messageBatchProcessing';\n\nexport type PostgreSQLEventStoreSubscriptionEventsBatch<\n EventType extends Event = Event,\n> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type PostgreSQLEventStoreSubscription<EventType extends Event = Event> =\n {\n id: string;\n start: (\n execute: SQLExecutor,\n ) => Promise<PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined>;\n isActive: boolean;\n handle: (\n messagesBatch: PostgreSQLEventStoreSubscriptionEventsBatch<EventType>,\n context: { pool: Dumbo },\n ) => Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult>;\n };\n\nexport const PostgreSQLEventStoreSubscription = {\n result: {\n skip: (options?: {\n reason?: string;\n }): PostgreSQLEventStoreSubscriptionMessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): PostgreSQLEventStoreSubscriptionMessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type PostgreSQLEventStoreSubscriptionMessageHandlerResult =\n | void\n | { type: 'SKIP'; reason?: string }\n | { type: 'STOP'; reason?: string; error?: EmmettError };\n\nexport type PostgreSQLEventStoreSubscriptionEachMessageHandler<\n EventType extends Event = Event,\n> = (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n) =>\n | Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult>\n | PostgreSQLEventStoreSubscriptionMessageHandlerResult;\n\nexport type PostgreSQLEventStoreSubscriptionStartFrom =\n | PostgreSQLEventStoreMessageBatchPullerStartFrom\n | 'CURRENT';\n\nexport type PostgreSQLEventStoreSubscriptionOptions<\n EventType extends Event = Event,\n> = {\n subscriptionId: string;\n version?: number;\n partition?: string;\n startFrom?: PostgreSQLEventStoreSubscriptionStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n eachMessage: PostgreSQLEventStoreSubscriptionEachMessageHandler<EventType>;\n};\n\nexport const postgreSQLEventStoreSubscription = <\n EventType extends Event = Event,\n>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n): PostgreSQLEventStoreSubscription => {\n const { eachMessage } = options;\n let isActive = true;\n //let lastProcessedPosition: bigint | null = null;\n\n return {\n id: options.subscriptionId,\n start: async (\n execute: SQLExecutor,\n ): Promise<PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined> => {\n isActive = true;\n if (options.startFrom !== 'CURRENT') return options.startFrom;\n\n const { lastProcessedPosition } = await readSubscriptionCheckpoint(\n execute,\n {\n subscriptionId: options.subscriptionId,\n partition: options.partition,\n },\n );\n\n if (lastProcessedPosition === null) return 'BEGINNING';\n\n return { globalPosition: lastProcessedPosition };\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n { messages },\n { pool },\n ): Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult> => {\n if (!isActive) return;\n\n return pool.withTransaction(async (tx) => {\n let result:\n | PostgreSQLEventStoreSubscriptionMessageHandlerResult\n | undefined = undefined;\n\n let lastProcessedPosition: bigint | null = null;\n\n for (const message of messages) {\n const typedMessage = message as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >;\n\n const messageProcessingResult = await eachMessage(typedMessage);\n\n // TODO: Add correct handling of the storing checkpoint\n await storeSubscriptionCheckpoint(tx.execute, {\n subscriptionId: options.subscriptionId,\n version: options.version,\n lastProcessedPosition,\n newPosition: typedMessage.metadata.globalPosition,\n partition: options.partition,\n });\n\n lastProcessedPosition = typedMessage.metadata.globalPosition;\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(typedMessage)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n });\n },\n };\n};\n"],"mappings":";AAAA;AAAA,EACE,SAAAA;AAAA,OAMK;;;AECA,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAE9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ;ACLV,IAAM,qBAAqB,CAEhC,WAC0C;AAE1C,SACE,OAAO,WAAW,cAClB,OAAO;EAEP,OAAO,UAAU,gBAAgB;AAErC;AAEO,IAAM,cAAN,MAAM,qBAAoB,MAAM;EAC9B;EAEP,YACE,SACA;AACA,UAAM,YACJ,WAAW,OAAO,YAAY,YAAY,eAAe,UACrD,QAAQ,YACR,SAAS,OAAO,IACd,UACA;AACR,UAAM,UACJ,WAAW,OAAO,YAAY,YAAY,aAAa,UACnD,QAAQ,UACR,SAAS,OAAO,IACd,UACA,2BAA2B,SAAS;AAE5C,UAAM,OAAO;AACb,SAAK,YAAY;AAGjB,WAAO,eAAe,MAAM,aAAY,SAAS;EACnD;AACF;AAEO,IAAM,mBAAN,MAAM,0BAAyB,YAAY;EAChD,YACS,SACA,UACP,SACA;AACA,UAAM;MACJ,WAAW;MACX,SACE,WACA,oBAAoB,SAAS,SAAS,CAAC,2BAA2B,SAAS,SAAS,CAAC;IACzF,CAAC;AATM,SAAA,UAAA;AACA,SAAA,WAAA;AAWP,WAAO,eAAe,MAAM,kBAAiB,SAAS;EACxD;AACF;;;AU/DA,SAAS,MAAMC,aAAY;ACA3B,SAAS,uBAAuB;ACAhC,SAAS,MAAMA,aAAY;ACA3B,SAAS,MAAM,YAAY;AAC3B,SAAS,mBAAAC,wBAAuB;ASDhC,OAAO,WAAW;AaAlB,SAAS,sBAAsB;AEA/B,OAIO;ACJP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC;EAGE,mBAAAA;OAEK;ACLP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ArCezB,IAAM,gBAAgB;AACtB,IAAM,wBACX;AACK,IAAM,uBACX;AAEK,IAAM,yBAAyB,CACpC,SACA,UACA,mBACY;AACZ,MAAI,aAAa,qBAAsB,QAAO;AAE9C,MAAI,YAAY,sBAAuB,QAAO,YAAY;AAE1D,MAAI,YAAY,cAAe,QAAO,YAAY;AAElD,SAAO,YAAY;AACrB;AAEO,IAAM,sCAAsC,CAGjD,SACA,UACA,mBACS;AACT,eAAa;AAEb,MAAI,CAAC,uBAAuB,SAAS,UAAU,cAAc;AAC3D,UAAM,IAAI,6BAA6B,SAAS,QAAQ;AAC5D;AAEO,IAAM,+BAAN,MAAM,sCAEH,iBAAiB;EACzB,YACE,SACA,UACA;AACA,UAAM,SAAS,SAAS,GAAG,UAAU,SAAS,CAAC;AAG/C,WAAO,eAAe,MAAM,8BAA6B,SAAS;EACpE;AACF;AIzDO,IAAM,mCAAmC,CAC9C,0BAGA,UAA6D,CAAC,MAC3D,IAAI,iCAAiC,0BAA0B,OAAO;AAEpE,IAAM,mCAAN,cAAqDC,iBAG1D;EAQA,YACU,0BAGR,UAA6D,CAAC,GAC9D;AACA,UAAM;MACJ,QAAQ,CAAC,WAAW;AAClB,gBAAQ,IAAI,gCAAgC,MAAM;AAClD,aAAK,aAAa;MACpB;IACF,CAAC;AAVO,SAAA,2BAAA;AAWR,SAAK,WAAW,SAAS,YAAY,KAAK;AAE1C,SAAK,2BAA2B;AAEhC,SAAK,cAAc,SAAS,qBAAqB,EAAE;EACrD;EAxBQ,gBAAuC;EAC/B;EACR,aAAsB;EAC9B,IAAW,uBAAuB;AAChC,WAAO,CAAC,KAAK;EACf;EAqBQ,cAAc,UAAkB;AACtC,SAAK,gBAAgB,YAAY,MAAM;AACrC,WAAK,oBAAoB;IAC3B,GAAG,QAAQ;EACb;EAEQ,eAAe;AACrB,QAAI,CAAC,KAAK,cAAe;AAEzB,kBAAc,KAAK,aAAa;AAChC,SAAK,gBAAgB;AACrB,SAAK,aAAa;AAClB,SAAK,yBAAyB,IAAI;EACpC;EAEQ,sBAAsB;AAC5B,QAAI,CAAC,KAAK,SAAS,UAAU,CAAC,KAAK,YAAY;AAC7C,WAAK,aAAa;IACpB;EACF;AACF;AK5DO,IAAM,aAAa,CAAI,MAAS,UAAsB;AAC3D,MAAI,YAAY,IAAI,GAAG;AACrB,WAAO,KAAK,OAAO,KAAK;EAC1B;AAEA,MAAI,MAAM,QAAQ,IAAI,GAAG;AACvB,WACE,MAAM,QAAQ,KAAK,KACnB,KAAK,WAAW,MAAM,UACtB,KAAK,MAAM,CAAC,KAAK,UAAU,WAAW,KAAK,MAAM,KAAK,CAAC,CAAC;EAE5D;AAEA,MACE,OAAO,SAAS,YAChB,OAAO,UAAU,YACjB,SAAS,QACT,UAAU,MACV;AACA,WAAO,SAAS;EAClB;AAEA,MAAI,MAAM,QAAQ,KAAK,EAAG,QAAO;AAEjC,QAAM,QAAQ,OAAO,KAAK,IAAI;AAC9B,QAAM,QAAQ,OAAO,KAAK,KAAK;AAE/B,MACE,MAAM,WAAW,MAAM,UACvB,CAAC,MAAM,MAAM,CAAC,QAAQ,MAAM,SAAS,GAAG,CAAC;AAEzC,WAAO;AAET,aAAW,OAAO,MAAM;AACtB,QAAI,KAAK,GAAG,aAAa,YAAY,MAAM,GAAG,aAAa;AACzD;AAEF,UAAM,UAAU,WAAW,KAAK,GAAG,GAAG,MAAM,GAAG,CAAC;AAChD,QAAI,CAAC,SAAS;AACZ,aAAO;IACT;EACF;AAEA,SAAO;AACT;AAIO,IAAM,cAAc,CAAI,SAAkC;AAC/D,SACE,QACA,OAAO,SAAS,YAChB,YAAY,QACZ,OAAO,KAAK,QAAQ,MAAM;AAE9B;AI/CO,IAAM,aAAa,OACxB,IACA,SACe;AACf,MAAI,SAAS,UAAa,KAAK,YAAY,EAAG,QAAO,GAAG;AAExD,SAAO;IACL,OAAO,SAAS;AACd,UAAI;AACF,eAAO,MAAM,GAAG;MAClB,SAASC,QAAO;AACd,YAAI,MAAM,oBAAoB,CAAC,KAAK,iBAAiBA,MAAK,GAAG;AAC3D,eAAKA,MAAc;QACrB;AACA,cAAMA;MACR;IACF;IACA,QAAQ,EAAE,SAAS,EAAE;EACvB;AACF;AI3BO,IAAM,aAAN,cAAyB,MAAM;EACpC,YAAY,MAAc;AACxB,UAAM,iBAAiB,IAAI,EAAE;EAC/B;AACF;AA0BO,IAAM,aAAa;EACxB,WAAW,CACT,OACA,YACG;AACH,WAAO,KAAK;MACV,SAAS,MAAM,QAAQ,IAAI,KAA6B,IAAI;;;MAG5D,CAAC,GAAG,MAAO,OAAO,MAAM,WAAW,EAAE,SAAS,IAAI;IACpD;EACF;EACA,OAAO,CACL,MACA,YACmB;AACnB,UAAM,SAAkB,KAAK,MAAM,MAAM,SAAS,OAAO;AAEzD,QAAI,SAAS,aAAa,CAAC,SAAS,UAAc,MAAM;AACtD,YAAM,IAAI,WAAW,IAAI;AAE3B,WAAO,SAAS,MACZ,QAAQ,IAAI,MAA8B,IACzC;EACP;AACF;ACoCO,IAAM,aAAa,CAcxB,eAC4B;AWxGvB,IAAM,SAAS,CAAOC,YAC3B,IAAIC,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,QAAID,QAAO,KAAK,GAAG;AACjB,iBAAW,QAAQ,KAAK;IAC1B;EACF;AACF,CAAC;ACPI,IAAM,MAAM,CAAWE,SAC5B,IAAID,iBAA0B;EAC5B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQC,KAAI,KAAK,CAAC;EAC/B;AACF,CAAC;ACLI,IAAM,SAAS,CACpB,SACA,iBACG,IAAI,sBAA4B,SAAS,YAAY;AAEnD,IAAM,wBAAN,cAA0CD,iBAAsB;EAC7D;EACA;EAER,YAAY,SAA0C,cAAiB;AACrE,UAAM;MACJ,WAAW,CAAC,UAAU;AACpB,aAAK,cAAc,KAAK,QAAQ,KAAK,aAAa,KAAK;MACzD;MACA,OAAO,CAAC,eAAe;AACrB,mBAAW,QAAQ,KAAK,WAAW;AACnC,mBAAW,UAAU;MACvB;IACF,CAAC;AAED,SAAK,cAAc;AACnB,SAAK,UAAU;EACjB;AACF;ACjBO,IAAM,cAAc,CAKzB,oBACAE,cAIA,eAAkC,EAAE,SAAS,MAAM,YAAY,GAAG,MAElE,IAAIF,iBAAqC;EACvC,MAAM,YAAY;AAChB;MACE,MAAM,WAAW,oBAAoBE,cAAa,UAAU;MAC5D;IACF,EAAE,MAAM,CAACC,WAAU;AACjB,iBAAW,MAAMA,MAAK;IACxB,CAAC;EACH;AACF,CAAC;AAEH,IAAM,aAAa,OACjB,oBACAD,cAIA,eACkB;AAClB,QAAM,eAAe,mBAAmB;AACxC,QAAM,SAAS,aAAa,UAAU;AAEtC,MAAI;AACF,QAAI;AAEJ,OAAG;AACD,YAAM,SAAS,MAAM,OAAO,KAAK;AACjC,aAAO,OAAO;AAEd,YAAMA,aAAY,QAAQ,UAAU;AAEpC,UAAI,MAAM;AACR,mBAAW,UAAU;MACvB;IACF,SAAS,CAAC;EACZ,UAAA;AACE,WAAO,YAAY;EACrB;AACF;ACxDO,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCF,iBAAsB;EACxD,QAAQ;EACR;EAER,YAAYI,OAAc;AACxB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,aAAK;AACL,YAAI,KAAK,QAAQ,KAAK,MAAM;AAC1B,qBAAW,QAAQ,KAAK;QAC1B;MACF;IACF,CAAC;AAED,SAAK,OAAOA;EACd;AACF;AClBO,IAAM,YAAY,CAAO,kBAC9B,IAAIJ,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;AAExB,QAAI,cAAc,KAAK,GAAG;AACxB,iBAAW,UAAU;IACvB;EACF;AACF,CAAC;ACTI,IAAM,SAAS,CAAO,kBAC3B,IAAIA,iBAA4B;EAC9B,MAAM,UAAU,OAAO,YAAY;AACjC,QAAI,CAAC,cAAc,KAAK,GAAG;AACzB,iBAAW,QAAQ,KAAK;AACxB;IACF;AACA,UAAM,QAAQ,QAAQ;AACtB,eAAW,UAAU;EACvB;AACF,CAAC;ACVI,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCA,kBAAsB;EACxD,QAAQ;EACR;EAER,YAAY,OAAe;AACzB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,YAAI,KAAK,QAAQ,KAAK,OAAO;AAC3B,eAAK;AACL,qBAAW,QAAQ,KAAK;QAC1B,OAAO;AACL,qBAAW,UAAU;QACvB;MACF;IACF,CAAC;AAED,SAAK,QAAQ;EACf;AACF;ACpBO,IAAM,aAAa,CAAO,iBAC/B,IAAIA,kBAA4B;EAC9B,MAAM,YAAY;AAChB,UAAM,YAAY,WAAW,MAAM;AACjC,iBAAW,UAAU;IACvB,GAAG,YAAY;AAEf,UAAM,oBAAoB,WAAW,UAAU,KAAK,UAAU;AAG9D,eAAW,YAAY,MAAM;AAC3B,mBAAa,SAAS;AACtB,wBAAkB;IACpB;EACF;EACA,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;EAC1B;AACF,CAAC;ACNI,IAAM,wBAAwB;EACnC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA,OAAO;EACP;EACA;EACA;AACF;AVnBA,IAAM,EAAE,OAAAK,OAAM,IAAI;AWNX,IAAM,iBAAN,cAA6B,MAAM;EACxC,YAAY,SAAiB;AAC3B,UAAM,OAAO;EACf;AACF;AAEO,IAAM,WAAW,CAAC,UAAmB,WAA6B;AACvE,QAAM,MAAM;AACZ,QAAM,MAAM;AAEZ,WAAS,GAAG;AACZ,WAAS,GAAG;AAEZ,SAAO,OAAO,KAAK,GAAG,EAAE,MAAM,CAAC,QAAgB;AAC7C,QAAI,OAAO,IAAI,GAAG,KAAK,UAAU;AAC/B,aAAO,SAAS,IAAI,GAAG,GAAG,IAAI,GAAG,CAAC;IACpC;AACA,WAAO,IAAI,GAAG,MAAM,IAAI,GAAG;EAC7B,CAAC;AACH;AAEO,IAAM,cAAc,CAAC,YAAqB;AAC/C,QAAM,IAAI,eAAe,WAAW,uCAAuC;AAC7E;AAwGO,IAAM,kBAAkB,CAC7B,QACA,UACA,YACG;AACH,MAAI,CAAC,WAAW,QAAQ,QAAQ;AAC9B,UAAM,IAAI;MACR,WACE;EAAY,WAAW,UAAU,QAAQ,CAAC;;EAAsB,WAAW,UAAU,MAAM,CAAC;IAChG;AACJ;AA4BO,SAAS,WACd,WACA,SAC2B;AAC3B,MAAI,cAAc;AAChB,UAAM,IAAI,eAAe,WAAW,oBAAoB;AAC5D;AAEO,SAAS,SACd,KACA,SACkB;AAClB,MAAI,CAAC,IAAK,OAAM,IAAI,eAAe,WAAW,yBAAyB;AACzE;AAEO,SAAS,YACd,UACA,QACA,SACM;AACN,MAAI,aAAa;AACf,UAAM,IAAI;MACR,GAAG,WAAW,uBAAuB;YAAgB,WAAW,UAAU,QAAQ,CAAC;UAAa,WAAW,UAAU,MAAM,CAAC;IAC9H;AACJ;AAEO,SAAS,eACd,KACA,OACA,SACM;AACN,MAAI,QAAQ;AACV,UAAM,IAAI;MACR,WAAW,sBAAsB,WAAW,UAAU,GAAG,CAAC;IAC5D;AACJ;AAEO,SAAS,gBACd,QACqB;AACrB,iBAAe,QAAQ,IAAI;AAC3B,WAAS,MAAM;AACjB;AAsFO,IAAM,kBAAkB,CAAI,UAAe;AAChD,SAAO;IACL,SAAS,MACP;MACE,MAAM;MACN;MACA,sBAAsB,WAAW,UAAU,KAAK,CAAC;IACnD;IACF,YAAY,MAAM,eAAe,MAAM,QAAQ,GAAG,gBAAgB;IAClE,SAAS,CAAC,WAAmB,YAAY,MAAM,QAAQ,MAAM;IAC7D,kBAAkB,CAAC,UAAe;AAChC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,0BAA0B,CAAC,UAAe;AACxC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,SAAS,GAAG,EAAE,CAAC,CAAC,CAAC;IACpE;IACA,8BAA8B,CAAC,UAAe;AAC5C,kBAAY,MAAM,QAAQ,MAAM,QAAQ,4BAA4B;AACpE,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,SAAS,GAAG,EAAE,CAAC,CAAC,CAAC;IACpE;IACA,2BAA2B,CAAC,UAAe;AACzC,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,qCAAqC,CAAC,UAAe;AACnD,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,2BAA2B,CAAC,UAAe;AACzC,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,eAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;AACrC,mBAAW,WAAW,MAAM,CAAC,GAAG,MAAM,CAAC,CAAC,CAAC;MAC3C;IACF;IACA,iBAAiB,CAAC,SAAY;AAC5B,kBAAY,MAAM,QAAQ,CAAC;AAC3B,iBAAW,WAAW,MAAM,CAAC,GAAG,IAAI,CAAC;IACvC;IACA,UAAU,CAAC,SAAY;AACrB,iBAAW,MAAM,KAAK,CAAC,MAAM,WAAW,GAAG,IAAI,CAAC,CAAC;IACnD;IACA,4BAA4B,CAAC,UAAe;AAC1C;QACE,MACG,IAAI,CAAC,MAAM,MAAM,OAAO,CAAC,MAAM,WAAW,GAAG,CAAC,CAAC,EAAE,MAAM,EACvD,OAAO,CAAC,MAAM,MAAM,CAAC,EAAE,WAAW,MAAM;MAC7C;IACF;IACA,eAAe,CAAC,UAAe;AAC7B,iBAAW,MAAM,KAAK,CAAC,MAAM,MAAM,KAAK,CAAC,MAAM,WAAW,GAAG,CAAC,CAAC,CAAC,CAAC;IACnE;IACA,UAAU,CAAC,YAAkC;AAC3C,iBAAW,MAAM,MAAM,OAAO,CAAC;IACjC;IACA,YAAY,CAAC,YAAkC;AAC7C,iBAAW,MAAM,KAAK,OAAO,CAAC;IAChC;IACA,eAAe,OACb,YACkB;AAClB,iBAAW,QAAQ,OAAO;AACxB,mBAAW,MAAM,QAAQ,IAAI,CAAC;MAChC;IACF;EACF;AACF;;;AnDhVA,OAAe;;;AsD1Bf,OAKO;;;ACLP,OAA2B;AAO3B;AAAA,EACE;AAAA,OAKK;AAQP,IAAM,iBAAiB,CACrB,QACA,YAIG;AACH,QAAM,EAAE,MAAM,kBAAkB,YAAY,aAAa,IAAI;AAE7D,SAAO,KAAK,eAAe,OAAO,eAAe;AAC/C,UAAM,QAAQ,YAAY,kBAAkB;AAAA,MAC1C,mBAAmB,EAAE,WAAW;AAAA,IAClC,CAAC;AACD,QAAI;AACF,YAAM,aAAa,MAAM,GAAG,UAAU,EAAE,WAAW,YAAY;AAE/D,aAAO,OAAO,UAAU;AAAA,IAC1B,UAAE;AACA,YAAM,MAAM,MAAM;AAAA,IACpB;AAAA,EACF,CAAC;AACH;AAEA,IAAM,sBAAsB,CAC1B,QACG;AACH,QAAM,EAAE,KAAK,UAAU,GAAG,QAAQ,IAAI;AAEtC,SAAO;AACT;AAEA,IAAM,uBAAuB,CAG3B,QACA,aACG;AACH,MAAI,SAAS;AACX;AAAA,MACE,SAAS;AAAA,MACT,OAAO;AAAA;AAAA,MAEP,4CAA4C,SAAS,GAAG,aAAa,OAAO,GAAG;AAAA,IACjF;AAEF,SAAO;AAAA,IACL,oBAAoB,MAAM;AAAA,IAC1B,oBAAoB,QAAQ;AAAA,EAC9B;AACF;AAQO,IAAM,iBACX,CACE,UACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM;AAEtB,yBAAqB,QAAQ,QAAQ;AAAA,EACvC;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,sBACX,CACE,WACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA;AAAA,MACE,UAAU;AAAA,MACV,OAAO;AAAA,MACP;AAAA,IACF;AAEA,aAAS,IAAI,GAAG,IAAI,UAAU,QAAQ,KAAK;AACzC,sBAAgB,MAAe,EAAE,SAAS,UAAU,CAAC,CAAE;AAAA,IACzD;AAAA,EACF;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,6BACX,CACE,eACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA;AAAA,MACE;AAAA,MACA,OAAO;AAAA,MACP;AAAA,IACF;AAAA,EACF;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,yBACX,CACE,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM,EAAE,WAAW;AAAA,EACrC;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,uBACX,CACE,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM;AAAA,EACxB;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,uBAAuB;AAAA,EAClC,gBAAgB,CACd,mBACG;AACH,WAAO;AAAA,MACL,QAAQ,CAAC,OAAe;AACtB,eAAO;AAAA,UACL,WAAW,CAAC,aACV,eAAe,UAAU;AAAA,YACvB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,SAAS,MACP,uBAAuB;AAAA,YACrB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,YAAY,MACV,qBAAqB;AAAA,YACnB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,QACL;AAAA,MACF;AAAA,MACA,UAAU,CACRC,YACG;AACH,eAAO;AAAA,UACL,aAAa,CAAC,cACZ,oBAAyB,WAAW;AAAA,YAClC,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,aAAa,CAAC,kBACZ,2BAA2B,eAAe;AAAA,YACxC,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,SAAS,MACP,uBAAuB;AAAA,YACrB,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,YAAY,MACV,qBAAqB;AAAA,YACnB,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,QACL;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;;;ACxOA;AAAA,EACE,eAAAC;AAAA,OAGK;AAgEA,IAAM,kBAAkB,CAI7B;AAAA,EACA;AAAA,EACA;AACF,MAIE,qBAAmD;AAAA,EACjD;AAAA,EACA,QAAQ,OAAO,QAAQ,YAAY;AACjC,UAAM,EAAE,kBAAkB,OAAO,IAAI;AACrC,UAAM,QAAQC,aAAY,kBAAkB;AAAA,MAC1C,mBAAmB,EAAE,OAAO;AAAA,IAC9B,CAAC;AACD,UAAM,OAAO,QAAQ;AAAA,MACnB,GAAG;AAAA,MACH;AAAA,IACF,CAAC;AAAA,EACH;AACF,CAAC;AA8BI,IAAM,6BAA6B,CAMxC,YAKmC;AACnC,QAAM,EAAE,gBAAgB,eAAe,UAAU,IAAI;AAErD,SAAO,gBAAgB;AAAA,IACrB,QAAQ,OAAO,QAAQ,EAAE,MAAM,MAAM;AACnC,YAAM,aAAa,MAAM,GAAG,EAAE,WAAqB,cAAc;AAEjE,iBAAW,SAAS,QAAQ;AAC1B,cAAM,WAAW,OAAO,cAAc,KAAK,GAAG,OAAO,aAAa;AAChE,iBAAO,kBAAkB,UACrB,MAAM,QAAQ;AAAA,YACZ,YAAY,QAAQ,aAAa;AAAA,YACjC;AAAA,UACF,IACA,MAAM,QAAQ;AAAA,YACZ;AAAA,YACA;AAAA,UACF;AAAA,QACN,CAAC;AAAA,MACH;AAAA,IACF;AAAA,IACA;AAAA,EACF,CAAC;AACH;AA6BO,IAAM,8BAA8B,CAMzC,YAKmC;AACnC,SAAO,2BAAmE;AAAA,IACxE,GAAG;AAAA,IACH,eAAe,CAAC,UAAU,MAAM,SAAS;AAAA,EAC3C,CAAC;AACH;;;AC7MA;AAAA,EACE;AAAA,OAKK;AAYP,SAAS,MAAMC,aAAY;AAwCpB,IAAM,2BAA2B;AAAA,EACtC,KAAK,CACH,YACwC;AACxC;AACE,YAAM,EAAE,YAAAC,aAAY,GAAG,YAAY,IAAI;AACvC,YAAM,EAAE,iBAAiB,IAAI;AAE7B,aAAO,CAAC,gBAA4D;AAClE,eAAO;AAAA,UACL,MAAM,CACJ,QACAC,aACG;AACH,kBAAM,YACJ,CAAC;AAEH,kBAAM,MAAM,OAAO,SAAgB;AACjC,kBAAI,iBAAiB;AACrB,oBAAM,gBAAgBA,UAAS,iBAAiB;AAEhD,yBAAW,SAAS;AAAA,gBAClB,GAAG;AAAA,gBACH,GAAG,MAAM,KAAK,EAAE,QAAQ,cAAc,CAAC,EAAE,QAAQ,MAAM,MAAM;AAAA,cAC/D,GAAG;AACD,sBAAM,WAAsC;AAAA,kBAC1C,gBAAgB,EAAE;AAAA,kBAClB,gBAAgB;AAAA,kBAChB,YAAY,QAAQC,MAAK,CAAC;AAAA,kBAC1B,SAASA,MAAK;AAAA,gBAChB;AAEA,0BAAU,KAAK;AAAA,kBACb,GAAG;AAAA,kBACH,UAAU;AAAA,oBACR,GAAG;AAAA,oBACH,GAAI,cAAc,QAAS,MAAM,YAAY,CAAC,IAAK,CAAC;AAAA,kBACtD;AAAA,gBAIF,CAAC;AAAA,cACH;AAEA,oBAAM,KAAK;AAAA,gBAAgB,CAAC,gBAC1B,kBAAkB;AAAA,kBAChB,QAAQ;AAAA,kBACR,aAAa,CAACF,WAAU;AAAA,kBACxB,YAAY;AAAA,oBACV;AAAA,oBACA;AAAA,kBACF;AAAA,gBACF,CAAC;AAAA,cACH;AAAA,YACF;AAEA,mBAAO;AAAA,cACL,MAAM,OACJ,QACA,YACkB;AAClB,sBAAM,OAAO,MAAM,WAAW;AAC9B,oBAAI;AACF,wBAAM,IAAI,IAAI;AAEd,wBAAM,YAAY,MAAM,OAAO,EAAE,MAAM,iBAAiB,CAAC;AAEzD,sBAAI,cAAc,UAAa,cAAc;AAC3C;AAAA,sBACE,WACE;AAAA,oBACJ;AAAA,gBACJ,UAAE;AACA,wBAAM,KAAK,MAAM;AAAA,gBACnB;AAAA,cACF;AAAA,cACA,YAAY,UACP,SACe;AAClB,sBAAM,OAAO,MAAM,WAAW;AAC9B,oBAAI;AACF,wBAAM,IAAI,IAAI;AACd,wBAAM,IAAI,eAAe,kCAAkC;AAAA,gBAC7D,SAAS,OAAO;AACd,sBAAI,iBAAiB,eAAgB,OAAM;AAE3C,sBAAI,KAAK,WAAW,EAAG;AAEvB,sBAAI,CAAC,mBAAmB,KAAK,CAAC,CAAC,GAAG;AAChC;AAAA,sBACE,KAAK,CAAC,EAAE,KAAkB;AAAA,sBAC1B,2CAA2C,OAAO,SAAS,CAAC;AAAA,oBAC9D;AACA;AAAA,kBACF;AAEA;AAAA,oBACE,iBAAiB,KAAK,CAAC;AAAA,oBACvB,yDAAyD,OAAO,SAAS,CAAC;AAAA,kBAC5E;AAEA,sBAAI,KAAK,CAAC,GAAG;AACX;AAAA,sBACE,KAAK,CAAC,EAAE,KAAkB;AAAA,sBAC1B,2CAA2C,OAAO,SAAS,CAAC;AAAA,oBAC9D;AAAA,kBACF;AAAA,gBACF,UAAE;AACA,wBAAM,KAAK,MAAM;AAAA,gBACnB;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAEO,IAAM,gBAAgB,CAK3B,YACA,UACgE;AAChE,SAAO;AAAA,IACL,GAAG;AAAA,IACH,UAAU;AAAA,MACR,GAAI,MAAM,YAAY,CAAC;AAAA,MACvB,YAAY,MAAM,UAAU,cAAc;AAAA,IAC5C;AAAA,EACF;AACF;AAEO,IAAM,iBAAiB,CAK5B,YACA,WACkE;AAClE,SAAO,OAAO,IAAI,CAAC,MAAM,cAAc,YAAY,CAAC,CAAC;AACvD;AAEO,IAAM,oBAAoB;AAE1B,IAAM,8BACX,CAA2BG,MAAU,SACrC,OAAO,EAAE,MAAM,EAAE,QAAQ,EAAE,MAAM;AAC/B,QAAM,SAAS,MAAM,QAAQ,MAASA,IAAG;AAEzC,kBAAgB,IAAI,EAAE,0BAA0B,OAAO,IAAI;AAC7D;AAEK,IAAM,YAAY;AAAA,EACvB,OAAO,CAACA,UAAc;AAAA,IACpB,YAAY;AAAA,MACV,aAAa,CAA2B,SACtC,4BAA4BA,MAAK,IAAI;AAAA,IACzC;AAAA,EACF;AACF;;;AHtKO,IAAM,oBAAoB,OAK/B,YACkB;AAClB,QAAM;AAAA,IACJ,aAAa;AAAA,IACb;AAAA,IACA,YAAY,EAAE,aAAa,iBAAiB;AAAA,EAC9C,IAAI;AAEJ,QAAM,aAAa,OAAO,IAAI,CAAC,MAAM,EAAE,IAAI;AAE3C,QAAM,cAAc,eAAe;AAAA,IAAO,CAAC,MACzC,EAAE,UAAU,KAAK,CAAC,SAAS,WAAW,SAAS,IAAI,CAAC;AAAA,EACtD;AAEA,QAAM,SAAU,MAAM,YAAY,WAAW,KAAK;AAElD,aAAWC,eAAc,aAAa;AACpC,UAAMA,YAAW,OAAO,QAAQ;AAAA,MAC9B;AAAA,MACA;AAAA,MACA;AAAA,MACA,SAAS,YAAY;AAAA,IACvB,CAAC;AAAA,EACH;AACF;AAEO,IAAM,uBAAuB,CAKlC,eAEA,WAKE,UAAU;AAEP,IAAM,kCAAkC,CAC7C,WAIG,cAEH,qBAAgC;AAAA,EAC9B;AAAA,EACA,QAAQ,OAAO,QAAQ,YAAY;AACjC,UAAM,OAAc,MAAM,OAAO,QAAQ,OAAO;AAEhD,UAAM,QAAQ,QAAQ,aAAa,IAAI;AAAA,EACzC;AACF,CAAC;AAEI,IAAM,6BAA6B,CACxC,WAIG,cAEH;AAAA,EACE,OAAO,QAAQ,YAAY;AACzB,UAAM,OAAc,CAAC;AAErB,eAAW,SAAS,QAAQ;AAC1B,WAAK,KAAK,MAAM,OAAO,OAAO,OAAO,CAAC;AAAA,IACxC;AACA,WAAO;AAAA,EACT;AAAA,EACA,GAAG;AACL;;;AItIF,OAAgD;;;ACAhD;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,OAIK;AAWP,SAAS,MAAMC,aAAY;;;AClBpB,IAAM,eAAe;AAErB,IAAM,YAAY;AAClB,IAAM,aAAa;AAEnB,IAAM,cAAc;AAAA,EACzB,QAAQ,GAAG,YAAY,WAAW,SAAS;AAAA,EAC3C,QAAQ,GAAG,YAAY,WAAW,SAAS;AAC7C;AAEA,IAAM,UAAU;AAAA,EACd,WAAW;AAAA,IACT,MAAM;AAAA,EACR;AAAA,EACA,YAAY,EAAE,MAAM,cAAc;AACpC;AAEO,IAAM,eAAe;AAAA,EAC1B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,cAAc;AAAA,EACzB,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,qBAAqB;AAAA,EAChC,MAAM,GAAG,YAAY;AACvB;;;ADdO,IAAM,kBAAkB;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iBA0Be,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAsBV,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAchB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,mBAKtB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgBpC;AAkBO,IAAM,iBAAiB,CAC5B,MACA,YACA,YACA,QACA,YAKA,KAAK,gBAAmC,OAAO,gBAAgB;AAC7D,QAAM,EAAE,QAAQ,IAAI;AAEpB,MAAI,OAAO,WAAW;AACpB,WAAO,EAAE,SAAS,OAAO,QAAQ,EAAE,SAAS,MAAM,EAAE;AAEtD,MAAI;AAEJ,MAAI;AACF,UAAM,wBAAwB;AAAA,MAC5B,SAAS;AAAA,IACX;AAEA,UAAM,iBAA8B,OAAO,IAAI,CAAC,GAAG,OAAO;AAAA,MACxD,GAAG;AAAA,MACH,UAAU;AAAA,QACR;AAAA,QACA,SAASC,MAAK;AAAA,QACd,gBAAgB,OAAO,CAAC;AAAA,QACxB,GAAI,cAAc,IAAK,EAAE,YAAY,CAAC,IAAK,CAAC;AAAA,MAC9C;AAAA,IACF,EAAE;AAGF,mBAAe,MAAM;AAAA,MACnB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,QACE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,SAAS;AACX,YAAM,QAAQ,cAAc,gBAAgB,EAAE,YAAY,CAAC;AAAA,EAC/D,SAAS,OAAO;AACd,QAAI,CAAC,6BAA6B,KAAK,EAAG,OAAM;AAEhD,mBAAe;AAAA,MACb,SAAS;AAAA,MACT,sBAAsB;AAAA,MACtB,sBAAsB;AAAA,MACtB,gBAAgB;AAAA,IAClB;AAAA,EACF;AAEA,QAAM;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,IAAI;AAEJ,SAAO;AAAA,IACL;AAAA,IACA,QACE,WACA,wBACA,wBACA,iBACI;AAAA,MACE,SAAS;AAAA,MACT,oBAAoB,OAAO,oBAAoB;AAAA,MAC/C,oBAAoB,OAAO,oBAAoB;AAAA,MAC/C,eAAe;AAAA,IACjB,IACA,EAAE,SAAS,MAAM;AAAA,EACzB;AACF,CAAC;AAEH,IAAM,oBAAoB,CACxB,aACkB;AAClB,MAAI,aAAa,OAAW,QAAO;AAEnC,MAAI,aAAa,qBAAsB,QAAO;AAG9C,MAAI,YAAY,sBAAuB,QAAO;AAG9C,MAAI,YAAY,cAAe,QAAO;AAEtC,SAAO;AACT;AAEA,IAAM,+BAA+B,CAAC,UACpC,iBAAiB,SAAS,UAAU,SAAS,MAAM,SAAS;AAS9D,IAAM,kBAAkB,CACtB,SACA,UACA,YACA,QACA,YAKA;AAAA,EACE,QAAQ;AAAA,IACN;AAAA,MACE;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAWA,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,EAAE,SAAS,OAAO,CAAC,EAAE,KAAK,GAAG;AAAA,MACzD,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,WAAW,UAAU,EAAE,IAAI,CAAC,CAAC,EAAE,KAAK,GAAG;AAAA,MACnE,OACG,IAAI,CAAC,MAAM,IAAI,MAAM,WAAW,UAAU,EAAE,YAAY,CAAC,CAAC,CAAC,CAAC,EAC5D,KAAK,GAAG;AAAA,MACX,OAAO,IAAI,MAAM,KAAK,EAAE,KAAK,GAAG;AAAA,MAChC,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,EAAE,IAAI,CAAC,EAAE,KAAK,GAAG;AAAA,MAC7C;AAAA,MACA;AAAA,MACA,SAAS,yBAAyB;AAAA,MAClC,SAAS,aAAa;AAAA,IACxB;AAAA,EACF;AACF;;;AE3QF,SAAS,UAAAC,SAAQ,OAAAC,YAA6B;AAGvC,IAAM,iCAAiCC,KAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,8BAOpB,UAAU;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,gBAQxB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,cAYzB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qBAehB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,cAM9B,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,CAWpC;AA+BD,eAAsB,4BACpB,SACA,SAOsE;AACtE,MAAI;AACF,UAAM,EAAE,OAAO,IAAI,MAAMC;AAAA,MACvB,QAAQ;AAAA,QACND;AAAA,UACE;AAAA,UACA,QAAQ;AAAA,UACR,QAAQ,WAAW;AAAA,UACnB,QAAQ;AAAA,UACR,QAAQ;AAAA,UACR,QAAQ,aAAa;AAAA,QACvB;AAAA,MACF;AAAA,IACF;AAEA,WAAO,WAAW,IACd,EAAE,SAAS,MAAM,aAAa,QAAQ,YAAY,IAClD,EAAE,SAAS,OAAO,QAAQ,WAAW,IAAI,YAAY,WAAW;AAAA,EACtE,SAAS,OAAO;AACd,YAAQ,IAAI,KAAK;AACjB,UAAM;AAAA,EACR;AACF;;;AC5HA,SAAS,UAAAE,eAAc;AAShB,IAAM,kBAAkBC;AAAA,EAC7B,8BAA8B,aAAa,IAAI;AAAA;AAAA;AAAA,sEAGqB,SAAS,KAAK,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAO7F;AAEO,IAAM,iBAAiBA;AAAA,EAC5B;AAAA;AAAA;AAAA,+BAG6B,YAAY,IAAI;AAAA;AAAA;AAAA,2EAG4B,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAYpF;AAEO,IAAM,wBAAwBA;AAAA,EACnC;AAAA,+BAC6B,mBAAmB,IAAI;AAAA;AAAA;AAAA,iFAG2B,SAAS,KAAK,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAMxG;AAEO,IAAM,kBAAkBA;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA;AAKF;AAEO,IAAM,qBAAqBA;AAAA,EAChC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgCF;AAEO,IAAM,sBAAsBA;AAAA,EACjC;AAAA;AAAA;AAAA,yCAGuC,YAAY,IAAI;AAAA,yCAChB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,+BAK3B,mBAAmB,IAAI,iCAAiC,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAI9G;AAEO,IAAM,eAAeA;AAAA,EAC1B;AAAA;AAAA;AAAA,mBAGiB,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,mCAIA,YAAY,IAAI,gCAAgC,SAAS,QAAQ,YAAY,IAAI,mBAAmB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM7G,YAAY,IAAI,gCAAgC,SAAS,uCAAuC,YAAY,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAMzJ,YAAY,IAAI,gCAAgC,SAAS,yCAAyC,YAAY,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA,mBAG3K,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,mCAID,aAAa,IAAI,gCAAgC,SAAS,QAAQ,aAAa,IAAI,mBAAmB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM/G,aAAa,IAAI,gCAAgC,SAAS,uCAAuC,aAAa,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM3J,aAAa,IAAI,gCAAgC,SAAS,yCAAyC,aAAa,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAKhM;AAEO,IAAM,eAAeA;AAAA,EAC1B;AAAA;AAAA;AAAA,iBAGe,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,iCAIA,YAAY,IAAI,6CAA6C,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM7E,YAAY,IAAI,4EAA4E,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM5G,YAAY,IAAI,8EAA8E,YAAY,IAAI;AAAA;AAAA;AAAA,iBAG9H,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,iCAID,aAAa,IAAI,6CAA6C,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM/E,aAAa,IAAI,4EAA4E,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM9G,aAAa,IAAI,8EAA8E,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAKjJ;AAEO,IAAM,4BAA4BA;AAAA,EACvC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,2DAOyD,YAAY,IAAI;AAAA;AAAA,qBAEtD,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,qCAIA,YAAY,IAAI,uDAAuD,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMvF,YAAY,IAAI,sFAAsF,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMtH,YAAY,IAAI,wFAAwF,YAAY,IAAI;AAAA;AAAA;AAAA,qBAGxI,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,qCAID,aAAa,IAAI,uDAAuD,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMzF,aAAa,IAAI,sFAAsF,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMxH,aAAa,IAAI,wFAAwF,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAM/J;AAEO,IAAM,4BAA4BA;AAAA,EACvC;AAAA;AAAA;AAAA;AAAA;AAAA,+GAK6G,YAAY,IAAI;AAAA;AAAA,qBAE1G,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,qCAIA,YAAY,IAAI,8DAA8D,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM9F,YAAY,IAAI,6FAA6F,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM7H,YAAY,IAAI,+FAA+F,YAAY,IAAI;AAAA;AAAA;AAAA,qBAG/I,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,qCAID,aAAa,IAAI,8DAA8D,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMhG,aAAa,IAAI,6FAA6F,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM/H,aAAa,IAAI,+FAA+F,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAMtK;AAEO,IAAM,sBAAsBA;AAAA,EACjC,6BAA6B,UAAU;AACzC;;;ACzTA,SAAS,cAAc,OAAAC,YAA6B;AAW7C,IAAM,gCAAgC,OAC3C,SACA,YACiD;AACjD,QAAM,SAAS,MAAM;AAAA,IACnB,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,QAI1B,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,eAAe,IAAI;AAAA,EACvD;AACF;;;AChCA,SAAS,SAAS,OAAAC,YAA6B;AA+CxC,IAAM,oBAAoB,OAK/B,SACA,YACyE;AACzE,QAAM,OACJ,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QAAQ,KAChB;AACR,QAAM,YACJ,WAAW,eAAe,UACtB,QAAQ,YACR,QAAQ,KAAK,QAAQ;AAE3B,QAAM,gBACJ,SAAS,CAAC,KAAK,0BAA0B,IAAI,KAAK;AAEpD,QAAM,cACJ,QAAQ,UAAU,0BAA0B,QAAQ,EAAE,KAAK;AAE7D,QAAM,iBACJ,eAAe,UAAU,SAAS,QAAQ,SAAS,KAAK;AAE1D,QAAM,SAA0D,MAAM;AAAA,IACpE,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,YAAY,IAAI;AAAA,uHACqF,aAAa,IAAI,WAAW;AAAA;AAAA,aAEtI,cAAc;AAAA,QACnB,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,IACA,CAAC,QAAQ;AACP,YAAM,WAAW;AAAA,QACf,MAAM,IAAI;AAAA,QACV,MAAM,IAAI;AAAA,QACV,UAAU,IAAI;AAAA,MAChB;AAEA,YAAM,WAAgD;AAAA,QACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,QAC1D,SAAS,IAAI;AAAA,QACb,YAAY,IAAI;AAAA,QAChB,gBAAgB,OAAO,IAAI,eAAe;AAAA,QAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC5C;AAEA,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MAIF;AAAA,IACF;AAAA,EACF;AAEA,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,uBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC,UAAU;AAAA,IACV,eAAe,OAAO,WAAW;AAAA,EACnC,IACA;AAAA,IACE,uBACE,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QACR;AAAA,IACR,UAAU,CAAC;AAAA,IACX,eAAe;AAAA,EACjB;AACN;;;AChIA,SAAS,WAAAC,UAAS,OAAAC,YAA6B;AA2BxC,IAAM,aAAa,OACxB,SACA,UACA,YAGG;AACH,QAAM,gBACJ,WAAW,UAAU,UACjB,0BAA0B,QAAQ,IAAI,KACtC;AAEN,QAAM,KAAK;AAAA,IACT,WAAW,QAAQ,UACf,QAAQ,KACR,WAAW,cAAc,WAAW,QAAQ,WAC1C,QAAQ,OAAO,QAAQ,WACvB;AAAA,EACR;AAEA,QAAM,cAAc,CAAC,MAAM,EAAE,IAAI,0BAA0B,EAAE,KAAK;AAElE,QAAM,SACJ,MAAMC;AAAA,IACJ,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACQ,YAAY,IAAI;AAAA,6EAC2C,aAAa,IAAI,WAAW;AAAA,QAC/F;AAAA,QACA,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,IACA,CAAC,QAAQ;AACP,YAAM,WAAW;AAAA,QACf,MAAM,IAAI;AAAA,QACV,MAAM,IAAI;AAAA,QACV,UAAU,IAAI;AAAA,MAChB;AAEA,YAAM,WAAgD;AAAA,QACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,QAC1D,SAAS,IAAI;AAAA,QACb,YAAY;AAAA,QACZ,gBAAgB,OAAO,IAAI,eAAe;AAAA,QAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC5C;AAEA,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MAIF;AAAA,IACF;AAAA,EACF;AAEF,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,sBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC;AAAA,IACA,cAAc;AAAA,EAChB,IACA;AAAA,IACE,sBAAsB;AAAA,IACtB,QAAQ,CAAC;AAAA,IACT,cAAc;AAAA,EAChB;AACN;;;ACjGA,SAAS,gBAAAC,eAAc,OAAAC,YAA6B;AAW7C,IAAM,6BAA6B,OACxC,SACA,YAC8C;AAC9C,QAAM,SAAS,MAAMC;AAAA,IACnB,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,mBAAmB,IAAI;AAAA;AAAA;AAAA,QAGjC,SAAS,aAAa;AAAA,QACtB,QAAQ;AAAA,MACV;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,uBAAuB,IAAI;AAAA,EAC/D;AACF;;;ARNO,IAAM,YAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,yBAAyB,OACpC,SACkB;AAClB,QAAM,KAAK,gBAAgB,CAAC,EAAE,QAAQ,MAAM,QAAQ,aAAa,SAAS,CAAC;AAC7E;;;A1D6FO,IAAM,2BAAsD;AAAA,EACjE,aAAa,CAAC;AAAA,EACd,QAAQ,EAAE,eAAe,iBAAiB;AAC5C;AAEO,IAAM,2CAA2C;AAEjD,IAAM,0BAA0B,CACrC,kBACA,UAAqC,6BACd;AACvB,QAAM,cAAc;AAAA,IAClB;AAAA,IACA,GAAI,QAAQ,oBAAoB,QAAQ,oBAAoB,CAAC;AAAA,EAC/D;AACA,QAAM,OAAO,WAAW,cAAc,YAAY,QAAQC,OAAM,WAAW;AAC3E,MAAI;AAEJ,QAAM,qBACJ,QAAQ,QAAQ,kBAAkB,UAClC,QAAQ,QAAQ,kBAAkB;AAEpC,QAAM,qBAAqB,MAAM;AAC/B,QAAI,CAAC,mBAAoB,QAAO,QAAQ,QAAQ;AAEhD,QAAI,CAAC,eAAe;AAClB,sBAAgB,uBAAuB,IAAI;AAAA,IAC7C;AACA,WAAO;AAAA,EACT;AAEA,QAAM,qBAAqB,QAAQ,eAAe,CAAC,GAChD,OAAO,CAAC,EAAE,KAAK,MAAM,SAAS,QAAQ,EACtC,IAAI,CAAC,EAAE,YAAAC,YAAW,MAAMA,WAAU;AAErC,QAAM,gBACJ,kBAAkB,SAAS,IACvB,CAAC,QAAQ,EAAE,YAAY,MACrB,kBAAkB;AAAA,IAChB,aAAa;AAAA,IACb,YAAY;AAAA,MACV;AAAA,MACA;AAAA,IACF;AAAA;AAAA;AAAA,IAGA;AAAA,EACF,CAAC,IACH;AAEN,SAAO;AAAA,IACL,QAAQ;AAAA,MACN,KAAK,MAAM,UAAU,KAAK,EAAE;AAAA,MAC5B,OAAO,MAAM,QAAQ,IAAI,UAAU,KAAK,EAAE,CAAC;AAAA,MAC3C,SAAS,YAAY;AACnB,eAAO,gBAAgB,uBAAuB,IAAI;AAAA,MACpD;AAAA,IACF;AAAA,IACA,MAAM,gBACJ,YACAC,UAKuC;AACvC,YAAM,EAAE,QAAQ,cAAc,KAAK,IAAIA;AAEvC,YAAM,wBAAwB,MAAM;AAEpC,UAAI,QAAQ,aAAa;AAEzB,YAAM,SAAS,MAAM,KAAK,WAAsB,YAAYA,SAAQ,IAAI;AACxE,YAAM,uBAAuB,OAAO;AAEpC;AAAA,QACE;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,iBAAW,SAAS,OAAO,QAAQ;AACjC,YAAI,CAAC,MAAO;AAEZ,gBAAQ,OAAO,OAAO,KAAK;AAAA,MAC7B;AAEA,aAAO;AAAA,QACL;AAAA,QACA;AAAA,QACA,cAAc,OAAO;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,YAAY,OACV,YACAA,aACoE;AACpE,YAAM,mBAAmB;AACzB,aAAO,WAAsB,KAAK,SAAS,YAAYA,QAAO;AAAA,IAChE;AAAA,IAEA,gBAAgB,OACd,YACA,QACAA,aACoD;AACpD,YAAM,mBAAmB;AAEzB,YAAM,CAAC,WAAW,GAAG,IAAI,IAAI,WAAW,MAAM,GAAG;AAEjD,YAAM,aACJ,aAAa,KAAK,SAAS,IAAI,YAAY;AAE7C,YAAM,eAAe,MAAM;AAAA,QACzB;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,UACE,GAAGA;AAAA,UACH;AAAA,QACF;AAAA,MACF;AAEA,UAAI,CAAC,aAAa;AAChB,cAAM,IAAI;AAAA,UACR,CAAC;AAAA;AAAA,UACDA,UAAS,yBAAyB;AAAA,QACpC;AAEF,aAAO;AAAA,QACL,2BAA2B,aAAa;AAAA,QACxC,yBAAyB,aAAa;AAAA,QACtC,kBACE,aAAa,sBAAsB,OAAO,OAAO,MAAM;AAAA,MAC3D;AAAA,IACF;AAAA,IACA,OAAO,MAAM,KAAK,MAAM;AAAA,IAExB,MAAM,YACJ,UACY;AACZ,aAAO,MAAM,KAAK,eAAe,OAAO,eAAe;AACrD,cAAM,eAA0C;AAAA,UAC9C,GAAG;AAAA,UACH,mBAAmB;AAAA,YACjB;AAAA,UACF;AAAA,QACF;AAEA,cAAM,aAAa;AAAA,UACjB;AAAA,UACA;AAAA,QACF;AAEA,eAAO,SAAS;AAAA,UACd;AAAA,UACA,OAAO,MAAM,QAAQ,QAAQ;AAAA,QAC/B,CAAC;AAAA,MACH,CAAC;AAAA,IACH;AAAA,EACF;AACF;;;AmE9SA,OAAiC;AAa1B,IAAM,mDAAmD;AACzD,IAAM,8DAA8D;AA+CpE,IAAM,yCAAyC,CAEpD;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF,MAAwG;AACtG,MAAI,YAAY;AAEhB,MAAI;AAEJ,QAAM,eAAe,OACnB,YACG;AACH,UAAM,QACJ,QAAQ,cAAc,cAClB,KACA,QAAQ,cAAc,SAClB,MAAM,8BAA8B,QAAQ,GAC3C,yBAAyB,KAC5B,QAAQ,UAAU;AAE1B,UAAM,sBAAgD;AAAA,MACpD;AAAA,MACA;AAAA,IACF;AAEA,QAAI,WAAW;AAEf,OAAG;AACD,YAAM,EAAE,UAAU,uBAAuB,cAAc,IACrD,MAAM,kBAA6B,UAAU,mBAAmB;AAElE,UAAI,SAAS,SAAS,GAAG;AACvB,cAAM,SAAS,MAAM,UAAU,EAAE,SAAS,CAAC;AAE3C,YAAI,UAAU,OAAO,SAAS,QAAQ;AACpC,sBAAY;AACZ;AAAA,QACF;AAAA,MACF;AAEA,0BAAoB,QAAQ;AAE5B,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAE5D,UAAI,CAAC,eAAe;AAClB,mBAAW,KAAK,IAAI,WAAW,GAAG,GAAI;AAAA,MACxC,OAAO;AACL,mBAAW;AAAA,MACb;AAAA,IACF,SAAS;AAAA,EACX;AAEA,SAAO;AAAA,IACL,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,OAAO,CAAC,YAAY;AAClB,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,oBAAY;AAEZ,eAAO,aAAa,OAAO;AAAA,MAC7B,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA,MAAM,YAAY;AAChB,UAAI,CAAC,UAAW;AAChB,kBAAY;AACZ,YAAM;AAAA,IACR;AAAA,EACF;AACF;AAEO,IAAM,qDAAqD,CAChE,YACoD;AACpD,MACE,QAAQ,WAAW,KACnB,QAAQ,KAAK,CAAC,MAAM,MAAM,UAAa,MAAM,WAAW;AAExD,WAAO;AAET,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,KAAK,EAAG,QAAO;AAE9C,SAAO,QACJ,OAAO,CAAC,MAAM,MAAM,UAAa,MAAM,eAAe,MAAM,KAAK,EACjE,KAAK,CAAC,GAAG,MAAO,IAAI,IAAI,IAAI,EAAG,EAAE,CAAC;AACvC;;;ACzJA,SAAS,SAAAC,cAAa;;;ACAtB,OAA6C;AAgCtC,IAAM,mCAAmC;AAAA,EAC9C,QAAQ;AAAA,IACN,MAAM,CAAC,aAEsD;AAAA,MAC3D,MAAM;AAAA,MACN,GAAI,WAAW,CAAC;AAAA,IAClB;AAAA,IACA,MAAM,CAAC,aAGsD;AAAA,MAC3D,MAAM;AAAA,MACN,GAAI,WAAW,CAAC;AAAA,IAClB;AAAA,EACF;AACF;AAgCO,IAAM,mCAAmC,CAG9C,YACqC;AACrC,QAAM,EAAE,YAAY,IAAI;AACxB,MAAI,WAAW;AAGf,SAAO;AAAA,IACL,IAAI,QAAQ;AAAA,IACZ,OAAO,OACL,YACyE;AACzE,iBAAW;AACX,UAAI,QAAQ,cAAc,UAAW,QAAO,QAAQ;AAEpD,YAAM,EAAE,sBAAsB,IAAI,MAAM;AAAA,QACtC;AAAA,QACA;AAAA,UACE,gBAAgB,QAAQ;AAAA,UACxB,WAAW,QAAQ;AAAA,QACrB;AAAA,MACF;AAEA,UAAI,0BAA0B,KAAM,QAAO;AAE3C,aAAO,EAAE,gBAAgB,sBAAsB;AAAA,IACjD;AAAA,IACA,IAAI,WAAW;AACb,aAAO;AAAA,IACT;AAAA,IACA,QAAQ,OACN,EAAE,SAAS,GACX,EAAE,KAAK,MAC2D;AAClE,UAAI,CAAC,SAAU;AAEf,aAAO,KAAK,gBAAgB,OAAO,OAAO;AACxC,YAAI,SAEY;AAEhB,YAAI,wBAAuC;AAE3C,mBAAW,WAAW,UAAU;AAC9B,gBAAM,eAAe;AAKrB,gBAAM,0BAA0B,MAAM,YAAY,YAAY;AAG9D,gBAAM,4BAA4B,GAAG,SAAS;AAAA,YAC5C,gBAAgB,QAAQ;AAAA,YACxB,SAAS,QAAQ;AAAA,YACjB;AAAA,YACA,aAAa,aAAa,SAAS;AAAA,YACnC,WAAW,QAAQ;AAAA,UACrB,CAAC;AAED,kCAAwB,aAAa,SAAS;AAE9C,cACE,2BACA,wBAAwB,SAAS,QACjC;AACA,uBAAW;AACX,qBAAS;AACT;AAAA,UACF;AAEA,cAAI,QAAQ,aAAa,QAAQ,UAAU,YAAY,GAAG;AACxD,uBAAW;AACX,qBAAS,EAAE,MAAM,QAAQ,QAAQ,yBAAyB;AAC1D;AAAA,UACF;AAEA,cACE,2BACA,wBAAwB,SAAS;AAEjC;AAAA,QACJ;AAEA,eAAO;AAAA,MACT,CAAC;AAAA,IACH;AAAA,EACF;AACF;;;ADrIO,IAAM,+BAA+B,CAC1C,YACiC;AACjC,MAAI,YAAY;AAChB,QAAM,EAAE,kBAAkB,QAAQ,IAAI;AACtC,QAAM,gBAAgB,QAAQ,iBAAiB,CAAC;AAEhD,MAAI;AAEJ,MAAI;AAEJ,QAAM,OAAOC,OAAM,EAAE,iBAAiB,CAAC;AAEvC,QAAM,YAAsD,OAC1D,kBACG;AACH,UAAM,sBAAsB,cAAc,OAAO,CAAC,MAAM,EAAE,QAAQ;AAElE,QAAI,oBAAoB,WAAW;AACjC,aAAO;AAAA,QACL,MAAM;AAAA,QACN,QAAQ;AAAA,MACV;AAEF,UAAM,SAAS,MAAM,QAAQ;AAAA,MAC3B,oBAAoB,IAAI,CAAC,MAAM;AAE7B,eAAO,EAAE,OAAO,eAAe,EAAE,KAAK,CAAC;AAAA,MACzC,CAAC;AAAA,IACH;AAEA,WAAO,OAAO;AAAA,MACZ,CAAC,MAAM,EAAE,WAAW,eAAe,EAAE,OAAO,SAAS;AAAA,IACvD,IACI,SACA;AAAA,MACE,MAAM;AAAA,IACR;AAAA,EACN;AAEA,QAAM,gBAAiB,uBACrB,uCAAuC;AAAA,IACrC,UAAU,KAAK;AAAA,IACf;AAAA,IACA,WACE,SAAS,aAAa;AAAA,IACxB,sBACE,SAAS,wBACT;AAAA,EACJ,CAAC;AAEH,QAAM,OAAO,YAAY;AACvB,QAAI,CAAC,UAAW;AAChB,gBAAY;AACZ,QAAI,sBAAsB;AACxB,YAAM,qBAAqB,KAAK;AAChC,6BAAuB;AAAA,IACzB;AACA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,WAAW,CACTC,aACgD;AAChD,YAAM,eAAe,iCAA4CA,QAAO;AAExE,oBAAc,KAAK,YAAY;AAE/B,aAAO;AAAA,IACT;AAAA,IACA,OAAO,MAAM;AACX,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,YAAI,cAAc,WAAW;AAC3B,iBAAO,QAAQ;AAAA,YACb,IAAI;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAEF,oBAAY;AAEZ,cAAM,YAAY;AAAA,UAChB,MAAM,QAAQ,IAAI,cAAc,IAAI,CAAC,MAAM,EAAE,MAAM,KAAK,OAAO,CAAC,CAAC;AAAA,QACnE;AAEA,eAAO,cAAc,MAAM,EAAE,UAAU,CAAC;AAAA,MAC1C,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA;AAAA,IACA,OAAO,YAAY;AACjB,YAAM,KAAK;AACX,YAAM,KAAK,MAAM;AAAA,IACnB;AAAA,EACF;AACF;","names":["dumbo","uuid","TransformStream","TransformStream","error","filter","TransformStream","map","handleChunk","error","skip","retry","filter","pongoClient","pongoClient","uuid","projection","options","uuid","sql","projection","uuid","uuid","single","sql","sql","single","rawSql","rawSql","sql","sql","sql","sql","mapRows","sql","mapRows","sql","singleOrNull","sql","singleOrNull","sql","dumbo","projection","options","dumbo","dumbo","options"]}
|
|
1
|
+
{"version":3,"sources":["../src/eventStore/postgreSQLEventStore.ts","../../emmett/src/config/plugins/index.ts","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/validation/dates.ts","../../emmett/src/typing/command.ts","../../emmett/src/typing/event.ts","../../emmett/src/typing/workflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/eventStore/afterCommit/forwardToMessageBus.ts","../../emmett/src/eventStore/events/index.ts","../../emmett/src/eventStore/eventStore.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/eventStore/subscriptions/caughtUpTransformStream.ts","../../emmett/src/eventStore/subscriptions/streamingCoordinator.ts","../../emmett/src/streaming/transformations/notifyAboutNoActiveReaders.ts","../../emmett/src/streaming/writers/writeToStream.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/utils/iterators.ts","../../emmett/src/taskProcessing/taskProcessor.ts","../../emmett/src/utils/locking/index.ts","../../emmett/src/utils/retry.ts","../../emmett/src/commandHandling/handleCommand.ts","../../emmett/src/commandHandling/handleCommandWithDecider.ts","../../emmett/src/messageBus/index.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/projections/index.ts","../../emmett/src/streaming/binaryArrays.ts","../../emmett/src/streaming/collectors/collect.ts","../../emmett/src/streaming/decoders/binary.ts","../../emmett/src/streaming/decoders/string.ts","../../emmett/src/streaming/decoders/json.ts","../../emmett/src/streaming/decoders/object.ts","../../emmett/src/streaming/decoders/composite.ts","../../emmett/src/streaming/generators/fromArray.ts","../../emmett/src/streaming/generators/index.ts","../../emmett/src/streaming/restream.ts","../../emmett/src/streaming/transformations/filter.ts","../../emmett/src/streaming/transformations/map.ts","../../emmett/src/streaming/transformations/reduce.ts","../../emmett/src/streaming/transformations/retry.ts","../../emmett/src/streaming/transformations/skip.ts","../../emmett/src/streaming/transformations/stopAfter.ts","../../emmett/src/streaming/transformations/stopOn.ts","../../emmett/src/streaming/transformations/take.ts","../../emmett/src/streaming/transformations/waitAtMost.ts","../../emmett/src/streaming/transformations/index.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/testing/deciderSpecification.ts","../../emmett/src/testing/wrapEventStore.ts","../src/eventStore/projections/index.ts","../src/eventStore/projections/pongo/pongoProjectionSpec.ts","../src/eventStore/projections/pongo/projections.ts","../src/eventStore/projections/postgresProjectionSpec.ts","../src/eventStore/schema/index.ts","../src/eventStore/schema/appendToStream.ts","../src/eventStore/schema/typing.ts","../src/eventStore/schema/storeSubscriptionCheckpoint.ts","../src/eventStore/schema/tables.ts","../src/eventStore/schema/readLastMessageGlobalPosition.ts","../src/eventStore/schema/readMessagesBatch.ts","../src/eventStore/schema/readStream.ts","../src/eventStore/schema/readSubscriptionCheckpoint.ts","../src/eventStore/subscriptions/messageBatchProcessing/index.ts","../src/eventStore/subscriptions/postgreSQLEventStoreConsumer.ts","../src/eventStore/subscriptions/postgreSQLEventStoreSubscription.ts"],"sourcesContent":["import {\n dumbo,\n type MigrationStyle,\n type NodePostgresClientConnection,\n type NodePostgresConnector,\n type NodePostgresPool,\n type NodePostgresPoolClientConnection,\n} from '@event-driven-io/dumbo';\nimport {\n assertExpectedVersionMatchesCurrent,\n ExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResultWithGlobalPosition,\n type Event,\n type EventStore,\n type EventStoreSession,\n type EventStoreSessionFactory,\n type ProjectionRegistration,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport pg from 'pg';\nimport {\n handleProjections,\n type PostgreSQLProjectionHandlerContext,\n} from './projections';\nimport {\n appendToStream,\n createEventStoreSchema,\n readStream,\n schemaSQL,\n type AppendToStreamPreCommitHook,\n} from './schema';\n\nexport interface PostgresEventStore\n extends EventStore<PostgresReadEventMetadata>,\n EventStoreSessionFactory<PostgresEventStore> {\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition>;\n close(): Promise<void>;\n schema: {\n sql(): string;\n print(): void;\n migrate(): Promise<void>;\n };\n}\n\nexport type PostgresReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type PostgresReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n PostgresReadEventMetadata\n>;\n\ntype PostgresEventStorePooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n };\n\ntype PostgresEventStoreNotPooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n connection:\n | NodePostgresPoolClientConnection\n | NodePostgresClientConnection;\n pooled?: false;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n dumbo: NodePostgresPool;\n pooled?: false;\n };\n\nexport type PostgresEventStoreConnectionOptions =\n | PostgresEventStorePooledOptions\n | PostgresEventStoreNotPooledOptions;\n\nexport type PostgresEventStoreOptions = {\n projections?: ProjectionRegistration<\n 'inline',\n PostgresReadEventMetadata,\n PostgreSQLProjectionHandlerContext\n >[];\n schema?: { autoMigration?: MigrationStyle };\n connectionOptions?: PostgresEventStoreConnectionOptions;\n};\n\nexport const defaultPostgreSQLOptions: PostgresEventStoreOptions = {\n projections: [],\n schema: { autoMigration: 'CreateOrUpdate' },\n};\n\nexport const PostgreSQLEventStoreDefaultStreamVersion = 0n;\n\nexport const getPostgreSQLEventStore = (\n connectionString: string,\n options: PostgresEventStoreOptions = defaultPostgreSQLOptions,\n): PostgresEventStore => {\n const poolOptions = {\n connectionString,\n ...(options.connectionOptions ? options.connectionOptions : {}),\n };\n const pool = 'dumbo' in poolOptions ? poolOptions.dumbo : dumbo(poolOptions);\n let migrateSchema: Promise<void>;\n\n const autoGenerateSchema =\n options.schema?.autoMigration === undefined ||\n options.schema?.autoMigration !== 'None';\n\n const ensureSchemaExists = () => {\n if (!autoGenerateSchema) return Promise.resolve();\n\n if (!migrateSchema) {\n migrateSchema = createEventStoreSchema(pool);\n }\n return migrateSchema;\n };\n\n const inlineProjections = (options.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n const preCommitHook: AppendToStreamPreCommitHook | undefined =\n inlineProjections.length > 0\n ? (events, { transaction }) =>\n handleProjections({\n projections: inlineProjections,\n connection: {\n connectionString,\n transaction,\n },\n // TODO: Add proper handling of global data\n // Currently it's not available as append doesn't return array of global position but just the last one\n events: events as ReadEvent<Event, PostgresReadEventMetadata>[],\n })\n : undefined;\n\n return {\n schema: {\n sql: () => schemaSQL.join(''),\n print: () => console.log(schemaSQL.join('')),\n migrate: async () => {\n await (migrateSchema = createEventStoreSchema(pool));\n },\n },\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n PostgresReadEventMetadata\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const expectedStreamVersion = read?.expectedStreamVersion;\n\n let state = initialState();\n\n const result = await this.readStream<EventType>(streamName, options.read);\n const currentStreamVersion = result.currentStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n PostgreSQLEventStoreDefaultStreamVersion,\n );\n\n for (const event of result.events) {\n if (!event) continue;\n\n state = evolve(state, event);\n }\n\n return {\n currentStreamVersion: currentStreamVersion,\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: async <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions,\n ): Promise<ReadStreamResult<EventType, PostgresReadEventMetadata>> => {\n await ensureSchemaExists();\n return readStream<EventType>(pool.execute, streamName, options);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition> => {\n await ensureSchemaExists();\n // TODO: This has to be smarter when we introduce urn-based resolution\n const [firstPart, ...rest] = streamName.split('-');\n\n const streamType =\n firstPart && rest.length > 0 ? firstPart : 'emt:unknown';\n\n const appendResult = await appendToStream(\n pool,\n streamName,\n streamType,\n events,\n {\n ...options,\n preCommitHook,\n },\n );\n\n if (!appendResult.success)\n throw new ExpectedVersionConflictError<bigint>(\n -1n, //TODO: Return actual version in case of error\n options?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n );\n\n return {\n nextExpectedStreamVersion: appendResult.nextStreamPosition,\n lastEventGlobalPosition: appendResult.lastGlobalPosition,\n createdNewStream:\n appendResult.nextStreamPosition >= BigInt(events.length),\n };\n },\n close: () => pool.close(),\n\n async withSession<T = unknown>(\n callback: (session: EventStoreSession<PostgresEventStore>) => Promise<T>,\n ): Promise<T> {\n return await pool.withConnection(async (connection) => {\n const storeOptions: PostgresEventStoreOptions = {\n ...options,\n connectionOptions: {\n connection,\n },\n };\n\n const eventStore = getPostgreSQLEventStore(\n connectionString,\n storeOptions,\n );\n\n return callback({\n eventStore,\n close: () => Promise.resolve(),\n });\n });\n },\n };\n};\n","export type EmmettPluginConfig =\n | {\n name: string;\n register: EmmettPluginRegistration[];\n }\n | string;\n\nexport type EmmettPluginType = 'cli';\n\nexport type EmmettCliPluginRegistration = { pluginType: 'cli'; path?: string };\n\nexport type EmmettPluginRegistration = EmmettCliPluginRegistration;\n\nexport type EmmettCliCommand = {\n addCommand<CliCommand>(command: CliCommand): CliCommand;\n};\n\nexport type EmmettCliPlugin = {\n pluginType: 'cli';\n name: string;\n registerCommands: (program: EmmettCliCommand) => Promise<void> | void;\n};\n\nexport type EmmettPlugin = EmmettCliPlugin;\n\nexport const isPluginConfig = (\n plugin: Partial<EmmettPluginConfig> | string | undefined,\n): plugin is EmmettPluginConfig =>\n plugin !== undefined &&\n (typeof plugin === 'string' ||\n ('name' in plugin &&\n plugin.name !== undefined &&\n typeof plugin.name === 'string'));\n","import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : 500;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: 412,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 400,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 403,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: 404,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { ValidationError } from '../errors';\n\nexport const formatDateToUtcYYYYMMDD = (date: Date) => {\n // Use the 'en-CA' locale which formats as 'yyyy-mm-dd'\n const formatter = new Intl.DateTimeFormat('en-CA', {\n timeZone: 'UTC',\n year: 'numeric',\n month: '2-digit',\n day: '2-digit',\n });\n\n // Format the date\n return formatter.format(date);\n};\n\n// Function to validate 'yyyy-mm-dd' format\nexport const isValidYYYYMMDD = (dateString: string) => {\n const regex = /^\\d{4}-\\d{2}-\\d{2}$/;\n return regex.test(dateString);\n};\n\nexport const parseDateFromUtcYYYYMMDD = (dateString: string) => {\n const date = new Date(dateString + 'T00:00:00Z');\n\n if (!isValidYYYYMMDD(dateString)) {\n throw new ValidationError('Invalid date format, must be yyyy-mm-dd');\n }\n\n if (isNaN(date.getTime())) {\n throw new ValidationError('Invalid date format');\n }\n\n return date;\n};\n","import type { DefaultRecord, Flavour } from './';\n\nexport type Command<\n CommandType extends string = string,\n CommandData extends DefaultRecord = DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Flavour<\n Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: Readonly<CommandData>;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n >,\n 'Command'\n>;\n\nexport type CommandTypeOf<T extends Command> = T['type'];\nexport type CommandDataOf<T extends Command> = T['data'];\nexport type CommandMetaDataOf<T extends Command> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CreateCommandType<\n CommandType extends string,\n CommandData extends DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: CommandData;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const command = <CommandType extends Command<string, any, any>>(\n ...args: CommandMetaDataOf<CommandType> extends undefined\n ? [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata?: DefaultCommandMetadata | undefined,\n ]\n : [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata: CommandMetaDataOf<CommandType>,\n ]\n): CommandType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata } as CommandType)\n : ({ type, data } as CommandType);\n};\n\nexport type DefaultCommandMetadata = { now: Date };\n","import type { DefaultRecord, Flavour } from './';\n\nexport type BigIntStreamPosition = bigint;\nexport type BigIntGlobalPosition = bigint;\n\nexport type Event<\n EventType extends string = string,\n EventData extends DefaultRecord = DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Flavour<\n Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n >,\n 'Event'\n>;\n\nexport type EventTypeOf<T extends Event> = T['type'];\nexport type EventDataOf<T extends Event> = T['data'];\nexport type EventMetaDataOf<T extends Event> = T extends { metadata: infer M }\n ? M\n : undefined;\n\nexport type CanHandle<T extends Event> = EventTypeOf<T>[];\n\nexport type CreateEventType<\n EventType extends string,\n EventData extends DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const event = <EventType extends Event<string, any, any>>(\n ...args: EventMetaDataOf<EventType> extends undefined\n ? [type: EventTypeOf<EventType>, data: EventDataOf<EventType>]\n : [\n type: EventTypeOf<EventType>,\n data: EventDataOf<EventType>,\n metadata: EventMetaDataOf<EventType>,\n ]\n): EventType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata } as EventType)\n : ({ type, data } as EventType);\n};\n\nexport type CombinedReadEventMetadata<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n EventMetaDataOf<EventType> extends undefined\n ? EventMetaDataType\n : EventMetaDataOf<EventType> & EventMetaDataType;\n\nexport type ReadEvent<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = EventType & {\n metadata: CombinedReadEventMetadata<EventType, EventMetaDataType>;\n};\n\nexport type ReadEventMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = Readonly<{\n eventId: string;\n streamPosition: StreamPosition;\n streamName: string;\n}> &\n (GlobalPosition extends undefined\n ? object\n : { globalPosition: GlobalPosition });\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyReadEventMetadata = ReadEventMetadata<any, any>;\n\nexport type ReadEventMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = ReadEventMetadata<GlobalPosition>;\n\nexport type ReadEventMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = ReadEventMetadata<undefined, StreamPosition>;\n\nexport type GlobalPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ReadEventMetadataType extends ReadEventMetadata<infer GP, any> ? GP : never;\n\nexport type StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ReadEventMetadataType extends ReadEventMetadata<any, infer SV> ? SV : never;\n","import type { Command } from './command';\nimport type { Event } from './event';\n\n/// Inspired by https://blog.bittacklr.be/the-workflow-pattern.html\n\nexport type Workflow<\n Input extends Event | Command,\n State,\n Output extends Event | Command,\n> = {\n decide: (command: Input, state: State) => WorkflowOutput<Output>[];\n evolve: (currentState: State, event: WorkflowEvent<Output>) => State;\n initialState: () => State;\n};\n\nexport type WorkflowEvent<Output extends Command | Event> = Extract<\n Output,\n { __brand?: 'Event' }\n>;\n\nexport type WorkflowCommand<Output extends Command | Event> = Extract<\n Output,\n { __brand?: 'Command' }\n>;\n\nexport type WorkflowOutput<TOutput extends Command | Event> =\n | { kind: 'Reply'; message: TOutput }\n | { kind: 'Send'; message: WorkflowCommand<TOutput> }\n | { kind: 'Publish'; message: WorkflowEvent<TOutput> }\n | {\n kind: 'Schedule';\n message: TOutput;\n when: { afterInMs: number } | { at: Date };\n }\n | { kind: 'Complete' }\n | { kind: 'Accept' }\n | { kind: 'Ignore'; reason: string }\n | { kind: 'Error'; reason: string };\n\nexport const reply = <TOutput extends Command | Event>(\n message: TOutput,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Reply',\n message,\n };\n};\n\nexport const send = <TOutput extends Command | Event>(\n message: WorkflowCommand<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Send',\n message,\n };\n};\n\nexport const publish = <TOutput extends Command | Event>(\n message: WorkflowEvent<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Publish',\n message,\n };\n};\n\nexport const schedule = <TOutput extends Command | Event>(\n message: TOutput,\n when: { afterInMs: number } | { at: Date },\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Schedule',\n message,\n when,\n };\n};\n\nexport const complete = <\n TOutput extends Command | Event,\n>(): WorkflowOutput<TOutput> => {\n return {\n kind: 'Complete',\n };\n};\n\nexport const ignore = <TOutput extends Command | Event>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Ignore',\n reason,\n };\n};\n\nexport const error = <TOutput extends Command | Event>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n kind: 'Error',\n reason,\n };\n};\n\nexport const accept = <\n TOutput extends Command | Event,\n>(): WorkflowOutput<TOutput> => {\n return { kind: 'Accept' };\n};\n","import { type Event, type ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\ntype AfterEventStoreCommitHandlerWithoutContext<Store extends EventStore> = (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n) => Promise<void> | void;\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext = never,\n> = [HandlerContext] extends [never]\n ? AfterEventStoreCommitHandlerWithoutContext<Store>\n : (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n context: HandlerContext,\n ) => Promise<void> | void;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext = never,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext = never,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","import type { EventsPublisher } from '../../messageBus';\nimport type { Event, ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\nimport type { AfterEventStoreCommitHandler } from './afterEventStoreCommitHandler';\n\nexport const forwardToMessageBus =\n <Store extends EventStore, HandlerContext = never>(\n eventPublisher: EventsPublisher,\n ): AfterEventStoreCommitHandler<Store, HandlerContext> =>\n async (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n ): Promise<void> => {\n for (const message of messages) {\n await eventPublisher.publish(message);\n }\n };\n","import {\n event,\n type Event,\n type EventDataOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\n\nexport const GlobalStreamCaughtUpType = '__emt:GlobalStreamCaughtUp';\n\nexport type GlobalStreamCaughtUp = Event<\n '__emt:GlobalStreamCaughtUp',\n { globalPosition: bigint },\n { globalPosition: bigint }\n>;\n\nexport const isGlobalStreamCaughtUp = (\n event: Event,\n): event is GlobalStreamCaughtUp => event.type === GlobalStreamCaughtUpType;\n\nexport const caughtUpEventFrom =\n (position: bigint) =>\n (\n event: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n ): event is ReadEvent<\n GlobalStreamCaughtUp,\n ReadEventMetadataWithGlobalPosition\n > =>\n event.type === GlobalStreamCaughtUpType &&\n event.metadata?.globalPosition >= position;\n\nexport const globalStreamCaughtUp = (\n data: EventDataOf<GlobalStreamCaughtUp>,\n): GlobalStreamCaughtUp =>\n event<GlobalStreamCaughtUp>(GlobalStreamCaughtUpType, data, {\n globalPosition: data.globalPosition,\n });\n\nexport const isSubscriptionEvent = (\n event: Event,\n): event is GlobalSubscriptionEvent => isGlobalStreamCaughtUp(event);\n\nexport const isNotInternalEvent = (event: Event): boolean =>\n !isGlobalStreamCaughtUp(event);\n\nexport type GlobalSubscriptionEvent = GlobalStreamCaughtUp;\n","//import type { ReadableStream } from 'web-streams-polyfill';\nimport type {\n AnyReadEventMetadata,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n Event,\n GlobalPositionTypeOfReadEventMetadata,\n ReadEvent,\n ReadEventMetadata,\n StreamPositionTypeOfReadEventMetadata,\n} from '../typing';\nimport type { AfterEventStoreCommitHandler } from './afterCommit';\n//import type { GlobalSubscriptionEvent } from './events';\nimport type { ExpectedStreamVersion } from './expectedVersion';\n\n// #region event-store\nexport interface EventStore<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> {\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType, ReadEventMetadataType>,\n ): Promise<\n AggregateStreamResult<\n State,\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<ReadStreamResult<EventType, ReadEventMetadataType>>;\n\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<\n AppendToStreamResult<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n // streamEvents(): ReadableStream<\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // >;\n}\n\nexport type EventStoreReadEventMetadata<Store extends EventStore> =\n Store extends EventStore<infer ReadEventMetadataType>\n ? ReadEventMetadataType extends ReadEventMetadata<infer GV, infer SV>\n ? ReadEventMetadata<GV, SV> & ReadEventMetadataType\n : never\n : never;\n\nexport type GlobalPositionTypeOfEventStore<Store extends EventStore> =\n GlobalPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type StreamPositionTypeOfEventStore<Store extends EventStore> =\n StreamPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type EventStoreSession<EventStoreType extends EventStore> = {\n eventStore: EventStoreType;\n close: () => Promise<void>;\n};\n\nexport interface EventStoreSessionFactory<EventStoreType extends EventStore> {\n withSession<T = unknown>(\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n ): Promise<T>;\n}\n// #endregion event-store\n\nexport const canCreateEventStoreSession = <Store extends EventStore>(\n eventStore: Store | EventStoreSessionFactory<Store>,\n): eventStore is EventStoreSessionFactory<Store> => 'withSession' in eventStore;\n\nexport const nulloSessionFactory = <EventStoreType extends EventStore>(\n eventStore: EventStoreType,\n): EventStoreSessionFactory<EventStoreType> => ({\n withSession: (callback) => {\n const nulloSession: EventStoreSession<EventStoreType> = {\n eventStore,\n close: () => Promise.resolve(),\n };\n\n return callback(nulloSession);\n },\n});\n\n////////////////////////////////////////////////////////////////////\n/// ReadStream types\n////////////////////////////////////////////////////////////////////\n\nexport type ReadStreamOptions<StreamVersion = BigIntStreamPosition> = (\n | {\n from: StreamVersion;\n }\n | { to: StreamVersion }\n | { from: StreamVersion; maxCount?: bigint }\n | {\n expectedStreamVersion: ExpectedStreamVersion<StreamVersion>;\n }\n) & {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type ReadStreamResult<\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n currentStreamVersion: StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>;\n events: ReadEvent<EventType, ReadEventMetadataType>[];\n streamExists: boolean;\n};\n\n////////////////////////////////////////////////////////////////////\n/// AggregateStream types\n////////////////////////////////////////////////////////////////////\n\ntype Evolve<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n | ((currentState: State, event: EventType) => State)\n | ((\n currentState: State,\n event: ReadEvent<EventType, ReadEventMetadataType>,\n ) => State)\n | ((currentState: State, event: ReadEvent<EventType>) => State);\n\nexport type AggregateStreamOptions<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n evolve: Evolve<State, EventType, ReadEventMetadataType>;\n initialState: () => State;\n read?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >;\n};\n\nexport type AggregateStreamResult<\n State,\n StreamPosition = BigIntStreamPosition,\n> = {\n currentStreamVersion: StreamPosition;\n state: State;\n streamExists: boolean;\n};\n\nexport type AggregateStreamResultWithGlobalPosition<\n State,\n StreamPosition = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> =\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: true;\n lastEventGlobalPosition: GlobalPosition;\n })\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: false;\n });\n\nexport type AggregateStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['aggregateStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// AppendToStream types\n////////////////////////////////////////////////////////////////////\n\nexport type AppendToStreamOptions<StreamVersion = BigIntStreamPosition> = {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type AppendToStreamResult<StreamVersion = BigIntStreamPosition> = {\n nextExpectedStreamVersion: StreamVersion;\n createdNewStream: boolean;\n};\n\nexport type AppendToStreamResultWithGlobalPosition<\n StreamVersion = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> = AppendToStreamResult<StreamVersion> & {\n lastEventGlobalPosition: GlobalPosition;\n};\n\nexport type AppendStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['appendToStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// DefaultEventStoreOptions\n////////////////////////////////////////////////////////////////////\n\nexport type DefaultEventStoreOptions<\n Store extends EventStore,\n HandlerContext = never,\n> = {\n /**\n * Pluggable set of hooks informing about the event store internal behaviour.\n */\n hooks?: {\n /**\n * This hook will be called **AFTER** events were stored in the event store.\n * It's designed to handle scenarios where delivery and ordering guarantees do not matter much.\n *\n * **WARNINGS:**\n *\n * 1. It will be called **EXACTLY ONCE** if append succeded.\n * 2. If the hook fails, its append **will still silently succeed**, and no error will be thrown.\n * 3. Wen process crashes after events were committed, but before the hook was called, delivery won't be retried.\n * That can lead to state inconsistencies.\n * 4. In the case of high concurrent traffic, **race conditions may cause ordering issues**.\n * For instance, where the second hook takes longer to process than the first one, ordering won't be guaranteed.\n *\n * @type {AfterEventStoreCommitHandler<Store, HandlerContext>}\n */\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n };\n};\n","import { ConcurrencyError } from '../errors';\nimport type { BigIntStreamPosition, Flavour } from '../typing';\n\nexport type ExpectedStreamVersion<VersionType = BigIntStreamPosition> =\n | ExpectedStreamVersionWithValue<VersionType>\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue<VersionType = BigIntStreamPosition> =\n Flavour<VersionType, 'StreamVersion'>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = <StreamVersion = BigIntStreamPosition>(\n current: StreamVersion | undefined,\n expected: ExpectedStreamVersion<StreamVersion>,\n defaultVersion: StreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = <\n StreamVersion = BigIntStreamPosition,\n>(\n current: StreamVersion,\n expected: ExpectedStreamVersion<StreamVersion> | undefined,\n defaultVersion: StreamVersion,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError<\n VersionType = BigIntStreamPosition,\n> extends ConcurrencyError {\n constructor(\n current: VersionType,\n expected: ExpectedStreamVersion<VersionType>,\n ) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError;\n","import { v4 as uuid } from 'uuid';\nimport type {\n BigIntStreamPosition,\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport {\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultEventStoreOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { StreamingCoordinator } from './subscriptions';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition>;\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore>;\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n const streamingCoordinator = StreamingCoordinator();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType>(streamName, read);\n\n const events = result?.events ?? [];\n\n return {\n currentStreamVersion: BigInt(events.length),\n state: events.reduce(evolve, initialState()),\n streamExists: result.streamExists,\n };\n },\n\n readStream: <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options && 'from' in options ? options.from : 0);\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : (events?.length ?? 1),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? events\n .map(\n (e) =>\n e as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n )\n .slice(from, to)\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition: BigInt(getAllEventsCount() + index + 1),\n };\n return {\n ...event,\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [...currentEvents, ...newEvents]);\n await streamingCoordinator.notify(newEvents);\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit(newEvents, eventStoreOptions?.hooks);\n\n return result;\n },\n\n //streamEvents: streamingCoordinator.stream,\n };\n};\n","import { TransformStream } from 'web-streams-polyfill';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport { globalStreamCaughtUp, type GlobalSubscriptionEvent } from '../events';\n\nexport const streamTrackingGlobalPosition = (\n currentEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n) => new CaughtUpTransformStream(currentEvents);\n\nexport class CaughtUpTransformStream extends TransformStream<\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n | ReadEvent<Event, ReadEventMetadataWithGlobalPosition>\n | GlobalSubscriptionEvent\n> {\n private _currentPosition: bigint;\n private _logPosition: bigint;\n\n constructor(events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]) {\n super({\n start: (controller) => {\n let globalPosition = 0n;\n for (const event of events) {\n controller.enqueue(event);\n globalPosition = event.metadata.globalPosition;\n }\n controller.enqueue(globalStreamCaughtUp({ globalPosition }));\n },\n transform: (event, controller) => {\n this._currentPosition = event.metadata.globalPosition;\n controller.enqueue(event);\n\n if (this._currentPosition < this._logPosition) return;\n\n controller.enqueue(\n globalStreamCaughtUp({ globalPosition: this._currentPosition }),\n );\n },\n });\n\n this._currentPosition = this._logPosition =\n events.length > 0\n ? events[events.length - 1]!.metadata.globalPosition\n : 0n;\n }\n\n public set logPosition(value: bigint) {\n this._logPosition = value;\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport { notifyAboutNoActiveReadersStream } from '../../streaming/transformations/notifyAboutNoActiveReaders';\nimport { writeToStream } from '../../streaming/writers';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport {\n CaughtUpTransformStream,\n streamTrackingGlobalPosition,\n} from './caughtUpTransformStream';\n\nexport const StreamingCoordinator = () => {\n const allEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[] = [];\n const listeners = new Map<string, CaughtUpTransformStream>();\n\n return {\n notify: async (\n events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n ) => {\n if (events.length === 0) return;\n\n allEvents.push(...events);\n\n for (const listener of listeners.values()) {\n listener.logPosition =\n events[events.length - 1]!.metadata.globalPosition;\n\n await writeToStream(listener, events);\n }\n },\n\n stream: () => {\n const streamId = uuid();\n const transformStream = streamTrackingGlobalPosition(allEvents);\n\n listeners.set(streamId, transformStream);\n return transformStream.readable.pipeThrough(\n notifyAboutNoActiveReadersStream(\n (stream) => {\n if (listeners.has(stream.streamId))\n listeners.delete(stream.streamId);\n },\n { streamId },\n ),\n );\n },\n };\n};\n","import { v4 as uuid } from 'uuid';\nimport { TransformStream } from 'web-streams-polyfill';\n\nexport const notifyAboutNoActiveReadersStream = <Item>(\n onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);\n\nexport class NotifyAboutNoActiveReadersStream<Item> extends TransformStream<\n Item,\n Item\n> {\n private checkInterval: NodeJS.Timeout | null = null;\n public readonly streamId: string;\n private _isStopped: boolean = false;\n public get hasActiveSubscribers() {\n return !this._isStopped;\n }\n\n constructor(\n private onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n ) {\n super({\n cancel: (reason) => {\n console.log('Stream was canceled. Reason:', reason);\n this.stopChecking();\n },\n });\n this.streamId = options?.streamId ?? uuid();\n\n this.onNoActiveReaderCallback = onNoActiveReaderCallback;\n\n this.startChecking(options?.intervalCheckInMs ?? 20);\n }\n\n private startChecking(interval: number) {\n this.checkInterval = setInterval(() => {\n this.checkNoActiveReader();\n }, interval);\n }\n\n private stopChecking() {\n if (!this.checkInterval) return;\n\n clearInterval(this.checkInterval);\n this.checkInterval = null;\n this._isStopped = true;\n this.onNoActiveReaderCallback(this);\n }\n\n private checkNoActiveReader() {\n if (!this.readable.locked && !this._isStopped) {\n this.stopChecking();\n }\n }\n}\n","import type { ReadableStream, TransformStream } from 'web-streams-polyfill';\nimport { noMoreWritingOn } from './noMoreWritingOn';\n\nexport const writeToStream = async <In, Out = In>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<boolean> => {\n if (stream.writable.locked) return false;\n\n const writer = stream.writable.getWriter();\n await writer.ready;\n\n if (!stream.readable.locked) return false;\n\n try {\n for (const item of items) {\n await writer.write(item);\n }\n } catch (error) {\n console.log(error);\n } finally {\n await writer.close();\n }\n return true;\n};\n\nexport const writeToStreamAndStop = async <In, Out>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<ReadableStream<Out>> => {\n await writeToStream(stream, items);\n return await noMoreWritingOn(stream);\n};\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","export const deepEquals = <T>(left: T, right: T): boolean => {\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n if (Array.isArray(left)) {\n return (\n Array.isArray(right) &&\n left.length === right.length &&\n left.every((val, index) => deepEquals(val, right[index]))\n );\n }\n\n if (\n typeof left !== 'object' ||\n typeof right !== 'object' ||\n left === null ||\n right === null\n ) {\n return left === right;\n }\n\n if (Array.isArray(right)) return false;\n\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (\n keys1.length !== keys2.length ||\n !keys1.every((key) => keys2.includes(key))\n )\n return false;\n\n for (const key in left) {\n if (left[key] instanceof Function && right[key] instanceof Function)\n continue;\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export const sum = (\n iterator: Iterator<number, number, number> | Iterator<number>,\n) => {\n let value,\n done: boolean | undefined,\n sum = 0;\n do {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n ({ value, done } = iterator.next());\n sum += value || 0;\n } while (!done);\n return sum;\n};\n","import { EmmettError } from '../errors';\n\nexport type TaskQueue = TaskQueueItem[];\n\nexport type TaskQueueItem = {\n task: () => Promise<void>;\n options?: EnqueueTaskOptions;\n};\n\nexport type TaskProcessorOptions = {\n maxActiveTasks: number;\n maxQueueSize: number;\n maxTaskIdleTime?: number;\n};\n\nexport type Task<T> = (context: TaskContext) => Promise<T>;\n\nexport type TaskContext = {\n ack: () => void;\n};\n\nexport type EnqueueTaskOptions = { taskGroupId?: string };\n\nexport class TaskProcessor {\n private queue: TaskQueue = [];\n private isProcessing = false;\n private activeTasks = 0;\n private activeGroups: Set<string> = new Set();\n\n constructor(private options: TaskProcessorOptions) {}\n\n enqueue<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n if (this.queue.length >= this.options.maxQueueSize) {\n return Promise.reject(\n new EmmettError(\n 'Too many pending connections. Please try again later.',\n ),\n );\n }\n\n return this.schedule(task, options);\n }\n\n waitForEndOfProcessing(): Promise<void> {\n return this.schedule(({ ack }) => Promise.resolve(ack()));\n }\n\n private schedule<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n return promiseWithDeadline(\n (resolve, reject) => {\n const taskWithContext = () => {\n return new Promise<void>((resolveTask, failTask) => {\n const taskPromise = task({\n ack: resolveTask,\n });\n\n taskPromise.then(resolve).catch((err) => {\n // eslint-disable-next-line @typescript-eslint/prefer-promise-reject-errors\n failTask(err);\n reject(err);\n });\n });\n };\n\n this.queue.push({ task: taskWithContext, options });\n if (!this.isProcessing) {\n this.ensureProcessing();\n }\n },\n { deadline: this.options.maxTaskIdleTime },\n );\n }\n\n private ensureProcessing(): void {\n if (this.isProcessing) return;\n this.isProcessing = true;\n this.processQueue();\n }\n\n private processQueue(): void {\n try {\n while (\n this.activeTasks < this.options.maxActiveTasks &&\n this.queue.length > 0\n ) {\n const item = this.takeFirstAvailableItem();\n\n if (item === null) return;\n\n const groupId = item.options?.taskGroupId;\n\n if (groupId) {\n // Mark the group as active\n this.activeGroups.add(groupId);\n }\n\n this.activeTasks++;\n void this.executeItem(item);\n }\n } catch (error) {\n console.error(error);\n throw error;\n } finally {\n this.isProcessing = false;\n if (\n this.hasItemsToProcess() &&\n this.activeTasks < this.options.maxActiveTasks\n ) {\n this.ensureProcessing();\n }\n }\n }\n\n private async executeItem({ task, options }: TaskQueueItem): Promise<void> {\n try {\n await task();\n } finally {\n this.activeTasks--;\n\n // Mark the group as inactive after task completion\n if (options && options.taskGroupId) {\n this.activeGroups.delete(options.taskGroupId);\n }\n\n this.ensureProcessing();\n }\n }\n\n private takeFirstAvailableItem = (): TaskQueueItem | null => {\n const taskIndex = this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n );\n\n if (taskIndex === -1) {\n // All remaining tasks are blocked by active groups\n return null;\n }\n\n // Remove the task from the queue\n const [item] = this.queue.splice(taskIndex, 1);\n\n return item ?? null;\n };\n\n private hasItemsToProcess = (): boolean =>\n this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n ) !== -1;\n}\n\nconst DEFAULT_PROMISE_DEADLINE = 2147483647;\n\nconst promiseWithDeadline = <T>(\n executor: (\n resolve: (value: T | PromiseLike<T>) => void,\n reject: (reason?: unknown) => void,\n ) => void,\n options: { deadline?: number },\n) => {\n return new Promise<T>((resolve, reject) => {\n let taskStarted = false;\n\n const maxWaitingTime = options.deadline || DEFAULT_PROMISE_DEADLINE;\n\n let timeoutId: NodeJS.Timeout | null = setTimeout(() => {\n if (!taskStarted) {\n reject(\n new Error('Task was not started within the maximum waiting time'),\n );\n }\n }, maxWaitingTime);\n\n executor((value) => {\n taskStarted = true;\n if (timeoutId) {\n clearTimeout(timeoutId);\n }\n timeoutId = null;\n resolve(value);\n }, reject);\n });\n};\n","import { TaskProcessor } from '../../taskProcessing';\n\nexport type LockOptions = { lockId: number };\n\nexport type AcquireLockOptions = { lockId: string };\nexport type ReleaseLockOptions = { lockId: string };\n\nexport type Lock = {\n acquire(options: AcquireLockOptions): Promise<void>;\n tryAcquire(options: AcquireLockOptions): Promise<boolean>;\n release(options: ReleaseLockOptions): Promise<boolean>;\n withAcquire: <Result = unknown>(\n handle: () => Promise<Result>,\n options: AcquireLockOptions,\n ) => Promise<Result>;\n};\n\nexport const InProcessLock = (): Lock => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: Number.MAX_VALUE,\n maxQueueSize: Number.MAX_VALUE,\n });\n\n // Map to store ack functions of currently held locks: lockId -> ack()\n const locks = new Map<string, () => void>();\n\n return {\n async acquire({ lockId }: AcquireLockOptions): Promise<void> {\n // If the lock is already held, we just queue up another task in the same group.\n // TaskProcessor ensures tasks in the same group run one at a time.\n await new Promise<void>((resolve, reject) => {\n taskProcessor\n .enqueue(\n ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n // We do NOT call ack() here. We hold onto the lock.\n resolve();\n return Promise.resolve();\n },\n { taskGroupId: lockId },\n )\n .catch(reject);\n });\n },\n\n async tryAcquire({ lockId }: AcquireLockOptions): Promise<boolean> {\n // If lock is already held, fail immediately\n if (locks.has(lockId)) {\n return false;\n }\n\n // TODO: Check pending queue\n await this.acquire({ lockId });\n\n return true;\n },\n\n release({ lockId }: ReleaseLockOptions): Promise<boolean> {\n const ack = locks.get(lockId);\n if (ack === undefined) {\n return Promise.resolve(true);\n }\n locks.delete(lockId);\n ack();\n return Promise.resolve(true);\n },\n\n async withAcquire<Result = unknown>(\n handle: () => Promise<Result>,\n { lockId }: AcquireLockOptions,\n ): Promise<Result> {\n return taskProcessor.enqueue(\n async ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n\n // We do NOT call ack() here. We hold onto the lock.\n try {\n return await handle();\n } finally {\n locks.delete(lockId);\n ack();\n }\n },\n { taskGroupId: lockId },\n );\n },\n };\n};\n","import retry from 'async-retry';\n\nexport type AsyncRetryOptions = retry.Options & {\n shouldRetryError?: (error: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n return await fn();\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","import {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport type { Event } from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\n\nexport const CommandHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type CommandHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromCommandHandlerRetryOptions = (\n retryOptions: CommandHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return CommandHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...CommandHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region command-handler\nexport type CommandHandlerResult<\n State,\n StreamEvent extends Event,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newState: State;\n newEvents: StreamEvent[];\n};\n\nexport type CommandHandlerOptions<State, StreamEvent extends Event> = {\n evolve: (state: State, event: StreamEvent) => State;\n initialState: () => State;\n mapToStreamId?: (id: string) => string;\n retry?: CommandHandlerRetryOptions;\n};\n\nexport type HandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] &\n (\n | {\n expectedStreamVersion?: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >;\n }\n | {\n retry?: CommandHandlerRetryOptions;\n }\n );\n\nexport const CommandHandler =\n <State, StreamEvent extends Event>(\n options: CommandHandlerOptions<State, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n store: Store,\n id: string,\n handle: (\n state: State,\n ) => StreamEvent | StreamEvent[] | Promise<StreamEvent | StreamEvent[]>,\n handleOptions?: HandleOptions<Store>,\n ): Promise<CommandHandlerResult<State, StreamEvent, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n CommandHandlerResult<\n State,\n StreamEvent,\n StreamPositionTypeOfEventStore<Store>\n >\n >(store, async ({ eventStore }) => {\n const { evolve, initialState } = options;\n const mapToStreamId = options.mapToStreamId ?? ((id) => id);\n\n const streamName = mapToStreamId(id);\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n State,\n StreamEvent\n >(streamName, {\n evolve,\n initialState,\n read: {\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const {\n state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n currentStreamVersion,\n streamExists: _streamExists,\n ...restOfAggregationResult\n } = aggregationResult;\n\n // 3. Run business logic\n const result = await handle(state);\n\n const newEvents = Array.isArray(result) ? result : [result];\n\n if (newEvents.length === 0) {\n return {\n ...restOfAggregationResult,\n newEvents: [],\n newState: state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n nextExpectedStreamVersion: currentStreamVersion,\n createdNewStream: false,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const expectedStreamVersion: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n > =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? (currentStreamVersion as ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >)\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n newEvents,\n {\n ...handleOptions,\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with updated state\n return {\n ...appendResult,\n newEvents,\n newState: newEvents.reduce(evolve, state),\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n });\n\n return result;\n },\n fromCommandHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion command-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type { EventStore } from '../eventStore';\nimport type { Command, Event } from '../typing';\nimport type { Decider } from '../typing/decider';\nimport {\n CommandHandler,\n type CommandHandlerOptions,\n type HandleOptions,\n} from './handleCommand';\n\n// #region command-handler\n\nexport type DeciderCommandHandlerOptions<\n State,\n CommandType extends Command,\n StreamEvent extends Event,\n> = CommandHandlerOptions<State, StreamEvent> &\n Decider<State, CommandType, StreamEvent>;\n\nexport const DeciderCommandHandler =\n <State, CommandType extends Command, StreamEvent extends Event>(\n options: DeciderCommandHandlerOptions<State, CommandType, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n eventStore: Store,\n id: string,\n command: CommandType,\n handleOptions?: HandleOptions<Store>,\n ) => {\n const { decide, ...rest } = options;\n\n return CommandHandler<State, StreamEvent>(rest)(\n eventStore,\n id,\n (state) => decide(command, state),\n handleOptions,\n );\n };\n// #endregion command-handler\n","import { EmmettError } from '../errors';\nimport {\n type Command,\n type CommandTypeOf,\n type Event,\n type EventTypeOf,\n} from '../typing';\n\nexport interface CommandSender {\n send<CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void>;\n}\n\nexport interface EventsPublisher {\n publish<EventType extends Event = Event>(event: EventType): Promise<void>;\n}\n\nexport type ScheduleOptions = { afterInMs: number } | { at: Date };\n\nexport interface MessageScheduler<CommandOrEvent extends Command | Event> {\n schedule<MessageType extends CommandOrEvent>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandBus extends CommandSender, MessageScheduler<Command> {}\n\nexport interface EventBus extends EventsPublisher, MessageScheduler<Event> {}\n\nexport interface MessageBus extends CommandBus, EventBus {\n schedule<MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\ntype CommandHandler<CommandType extends Command = Command> = (\n command: CommandType,\n) => Promise<void> | void;\n\nexport interface CommandProcessor {\n handle<CommandType extends Command>(\n commandHandler: CommandHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void;\n}\n\ntype EventHandler<EventType extends Event = Event> = (\n event: EventType,\n) => Promise<void> | void;\n\nexport interface EventProcessor {\n subscribe<EventType extends Event>(\n eventHandler: EventHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void;\n}\n\nexport type ScheduledMessage = {\n message: Event | Command;\n options?: ScheduleOptions;\n};\n\nexport interface ScheduledMessageProcessor {\n dequeue(): ScheduledMessage[];\n}\n\nexport type MessageHandler = CommandHandler | EventHandler;\n\nexport type MessageProcessor = EventProcessor | CommandProcessor;\n\nexport const getInMemoryMessageBus = (): MessageBus &\n EventProcessor &\n CommandProcessor &\n ScheduledMessageProcessor => {\n const allHandlers = new Map<string, MessageHandler[]>();\n let pendingMessages: ScheduledMessage[] = [];\n\n return {\n send: async <CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void> => {\n const handlers = allHandlers.get(command.type);\n\n if (handlers === undefined || handlers.length === 0)\n throw new EmmettError(\n `No handler registered for command ${command.type}!`,\n );\n\n const commandHandler = handlers[0] as CommandHandler<CommandType>;\n\n await commandHandler(command);\n },\n\n publish: async <EventType extends Event = Event>(\n event: EventType,\n ): Promise<void> => {\n const handlers = allHandlers.get(event.type) ?? [];\n\n for (const handler of handlers) {\n const eventHandler = handler as EventHandler<EventType>;\n\n await eventHandler(event);\n }\n },\n\n schedule: <MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void => {\n pendingMessages = [...pendingMessages, { message, options: when }];\n },\n\n handle: <CommandType extends Command>(\n commandHandler: CommandHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void => {\n const alreadyRegistered = [...allHandlers.keys()].filter((registered) =>\n commandTypes.includes(registered),\n );\n\n if (alreadyRegistered.length > 0)\n throw new EmmettError(\n `Cannot register handler for commands ${alreadyRegistered.join(', ')} as they're already registered!`,\n );\n for (const commandType of commandTypes) {\n allHandlers.set(commandType, [commandHandler as MessageHandler]);\n }\n },\n\n subscribe<EventType extends Event>(\n eventHandler: EventHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void {\n for (const eventType of eventTypes) {\n if (!allHandlers.has(eventType)) allHandlers.set(eventType, []);\n\n allHandlers.set(eventType, [\n ...(allHandlers.get(eventType) ?? []),\n eventHandler as MessageHandler,\n ]);\n }\n },\n\n dequeue: (): ScheduledMessage[] => {\n const pending = pendingMessages;\n pendingMessages = [];\n return pending;\n },\n };\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","import { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyReadEventMetadata,\n CanHandle,\n DefaultRecord,\n Event,\n ReadEvent,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n events: ReadEvent<EventType, EventMetaDataType>[],\n context: ProjectionHandlerContext,\n) => Promise<void> | void;\n\nexport interface ProjectionDefinition<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<Event>;\n handle: ProjectionHandler<\n Event,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n}\n\nexport interface TypedProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n > = TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >,\n>(\n definition: ProjectionDefintionType,\n): ProjectionDefintionType => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n > = ProjectionDefinition<ReadEventMetadataType, ProjectionHandlerContext>,\n>(\n definitions: ProjectionDefintionType[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] => definitions.map((projection) => ({ type: 'inline', projection }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n ProjectionDefintionType extends ProjectionDefinition<\n ReadEventMetadataType,\n ProjectionHandlerContext\n > = ProjectionDefinition<ReadEventMetadataType, ProjectionHandlerContext>,\n>(\n definitions: ProjectionDefintionType[],\n): ProjectionRegistration<\n 'async',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] => definitions.map((projection) => ({ type: 'async', projection }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","export const concatUint8Arrays = (chunks: Uint8Array[]): Uint8Array => {\n const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const chunk of chunks) {\n result.set(chunk, offset);\n offset += chunk.length;\n }\n return result;\n};\n","import type { ReadableStream } from 'web-streams-polyfill';\n\nexport const collect = async <T>(stream: ReadableStream<T>): Promise<T[]> => {\n const results: T[] = [];\n\n for await (const value of stream) {\n results.push(value as T);\n }\n\n return results;\n};\n","import type { Decoder } from '.';\nimport { concatUint8Arrays } from '../binaryArrays';\n\nexport class BinaryJsonDecoder<Decoded>\n implements Decoder<Uint8Array, Decoded>\n{\n private buffer: Uint8Array[] = [];\n\n addToBuffer(data: Uint8Array): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n return text.includes('\\n');\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage()) {\n return null;\n }\n\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n const delimiterIndex = text.indexOf('\\n');\n\n if (delimiterIndex === -1) {\n return null;\n }\n\n const jsonString = text.slice(0, delimiterIndex);\n const remaining = new Uint8Array(combined.buffer, delimiterIndex + 1);\n this.buffer = remaining.byteLength > 0 ? [remaining] : [];\n\n return JSON.parse(jsonString) as Decoded;\n }\n}\n","import type { Decoder } from '.';\n\nexport class StringDecoder<Decoded> implements Decoder<string, Decoded> {\n protected buffer: string[] = [];\n\n constructor(private transform: (input: string) => Decoded) {\n this.transform = transform;\n }\n\n addToBuffer(data: string): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer.some((chunk) => chunk.includes('\\n'));\n }\n\n decode(): Decoded | null {\n const completeString = this.buffer.join('');\n\n if (!this.hasCompleteMessage()) {\n if (completeString.trim().length > 0) {\n throw new Error('Unterminated string in JSON at position');\n }\n return null;\n }\n\n const delimiterIndex = completeString.indexOf('\\n');\n const message = completeString.slice(0, delimiterIndex).trim();\n this.buffer = [completeString.slice(delimiterIndex + 1)];\n\n return this.transform(message);\n }\n}\n","import { StringDecoder } from './string';\n\nexport class JsonDecoder<Decoded> extends StringDecoder<Decoded> {\n constructor() {\n super((jsonString) => JSON.parse(jsonString) as Decoded);\n }\n}\n","import type { Decoder } from '.';\n\nexport class ObjectDecoder<Decoded> implements Decoder<Decoded, Decoded> {\n private buffer: Decoded | null = null;\n\n addToBuffer(data: Decoded): void {\n this.buffer = data;\n }\n\n clearBuffer(): void {\n this.buffer = null;\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer !== null;\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage() || !this.buffer) {\n return null;\n }\n\n const data = this.buffer;\n this.clearBuffer();\n return data as Decoded;\n }\n}\n","import type { Decoder } from '.';\nimport { BinaryJsonDecoder } from './binary';\nimport { JsonDecoder } from './json';\nimport { ObjectDecoder } from './object';\n\nexport class CompositeDecoder<Decoded> implements Decoder<unknown, Decoded> {\n constructor(\n private decoders: [(data: unknown) => boolean, Decoder<unknown, Decoded>][],\n ) {}\n\n private decoderFor(data: unknown): Decoder<unknown, unknown> | null {\n const decoder = this.decoders.find((d) => d[0](data));\n\n if (!decoder) return null;\n\n return decoder[1];\n }\n\n addToBuffer(data: unknown): void {\n this.decoderFor(data)?.addToBuffer(data);\n }\n\n clearBuffer(): void {\n for (const decoder of this.decoders.map((d) => d[1])) {\n decoder.clearBuffer();\n }\n }\n\n hasCompleteMessage(): boolean {\n return this.decoders.some((d) => d[1].hasCompleteMessage());\n }\n\n decode(): Decoded | null {\n const decoder = this.decoders\n .map((d) => d[1])\n .find((d) => d.hasCompleteMessage());\n\n return decoder?.decode() ?? null;\n }\n}\n\nexport class DefaultDecoder<Decoded> extends CompositeDecoder<Decoded> {\n constructor() {\n super([\n [(data) => typeof data === 'string', new JsonDecoder<Decoded>()],\n [(data) => data instanceof Uint8Array, new BinaryJsonDecoder<Decoded>()],\n [(data) => typeof data === 'object', new ObjectDecoder<Decoded>()],\n ]);\n }\n}\n","import { ReadableStream } from 'web-streams-polyfill';\n\nexport const fromArray = <T>(chunks: T[]) =>\n new ReadableStream<T>({\n start(controller) {\n for (const chunk of chunks) controller.enqueue(chunk);\n controller.close();\n },\n });\n","import { fromArray } from './fromArray';\nexport const streamGenerators = { fromArray };\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport type { AsyncRetryOptions } from '../utils';\nimport type { Decoder } from './decoders';\nimport { DefaultDecoder } from './decoders/composite';\nimport { streamTransformations } from './transformations';\n\nconst { retry } = streamTransformations;\n\nexport const restream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n): ReadableStream<Transformed> =>\n retry(createSourceStream, handleChunk(transform, decoder), retryOptions)\n .readable;\n\nconst handleChunk =\n <Source = unknown, Transformed = Source, StreamType = Source>(\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n ) =>\n (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ): void => {\n const { done: isDone, value } = readResult;\n\n if (value) decoder.addToBuffer(value);\n\n if (!isDone && !decoder.hasCompleteMessage()) return;\n\n decodeAndTransform(decoder, transform, controller);\n };\n\nconst decodeAndTransform = <StreamType, Source, Transformed = Source>(\n decoder: Decoder<StreamType, Source>,\n transform: (input: Source) => Transformed,\n controller: TransformStreamDefaultController<Transformed>,\n) => {\n try {\n const decoded = decoder.decode();\n if (!decoded) return; // TODO: Add a proper handling of decode errors\n\n const transformed = transform(decoded);\n controller.enqueue(transformed);\n } catch (error) {\n controller.error(new Error(`Decoding error: ${error?.toString()}`));\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const filter = <Item>(filter: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n if (filter(chunk)) {\n controller.enqueue(chunk);\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const map = <From, To>(map: (item: From) => To) =>\n new TransformStream<From, To>({\n transform(chunk, controller) {\n controller.enqueue(map(chunk));\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const reduce = <I, O>(\n reducer: (accumulator: O, chunk: I) => O,\n initialValue: O,\n) => new ReduceTransformStream<I, O>(reducer, initialValue);\n\nexport class ReduceTransformStream<I, O> extends TransformStream<I, O> {\n private accumulator: O;\n private reducer: (accumulator: O, chunk: I) => O;\n\n constructor(reducer: (accumulator: O, chunk: I) => O, initialValue: O) {\n super({\n transform: (chunk) => {\n this.accumulator = this.reducer(this.accumulator, chunk);\n },\n flush: (controller) => {\n controller.enqueue(this.accumulator);\n controller.terminate();\n },\n });\n\n this.accumulator = initialValue;\n this.reducer = reducer;\n }\n}\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n TransformStream,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport { type AsyncRetryOptions, asyncRetry } from '../../utils';\n\nexport const retryStream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n): TransformStream<Source, Transformed> =>\n new TransformStream<Source, Transformed>({\n start(controller) {\n asyncRetry(\n () => onRestream(createSourceStream, handleChunk, controller),\n retryOptions,\n ).catch((error) => {\n controller.error(error);\n });\n },\n });\n\nconst onRestream = async <StreamType, Source, Transformed = Source>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n controller: TransformStreamDefaultController<Transformed>,\n): Promise<void> => {\n const sourceStream = createSourceStream();\n const reader = sourceStream.getReader();\n\n try {\n let done: boolean;\n\n do {\n const result = await reader.read();\n done = result.done;\n\n await handleChunk(result, controller);\n\n if (done) {\n controller.terminate();\n }\n } while (!done);\n } finally {\n reader.releaseLock();\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const skip = <T>(limit: number) => new SkipTransformStream<T>(limit);\n\nexport class SkipTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private skip: number;\n\n constructor(skip: number) {\n super({\n transform: (chunk, controller) => {\n this.count++;\n if (this.count > this.skip) {\n controller.enqueue(chunk);\n }\n },\n });\n\n this.skip = skip;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopAfter = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n controller.enqueue(chunk);\n\n if (stopCondition(chunk)) {\n controller.terminate();\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopOn = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n async transform(chunk, controller) {\n if (!stopCondition(chunk)) {\n controller.enqueue(chunk);\n return;\n }\n await Promise.resolve();\n controller.terminate();\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const take = <T>(limit: number) => new TakeTransformStream<T>(limit);\n\nexport class TakeTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private limit: number;\n\n constructor(limit: number) {\n super({\n transform: (chunk, controller) => {\n if (this.count < this.limit) {\n this.count++;\n controller.enqueue(chunk);\n } else {\n controller.terminate();\n }\n },\n });\n\n this.limit = limit;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const waitAtMost = <Item>(waitTimeInMs: number) =>\n new TransformStream<Item, Item>({\n start(controller) {\n const timeoutId = setTimeout(() => {\n controller.terminate();\n }, waitTimeInMs);\n\n const originalTerminate = controller.terminate.bind(controller);\n\n // Clear the timeout if the stream is terminated early\n controller.terminate = () => {\n clearTimeout(timeoutId);\n originalTerminate();\n };\n },\n transform(chunk, controller) {\n controller.enqueue(chunk);\n },\n });\n","import { filter } from './filter';\nimport { map } from './map';\nimport {\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n} from './notifyAboutNoActiveReaders';\nimport { reduce, ReduceTransformStream } from './reduce';\nimport { retryStream } from './retry';\nimport { skip, SkipTransformStream } from './skip';\nimport { stopAfter } from './stopAfter';\nimport { stopOn } from './stopOn';\nimport { take, TakeTransformStream } from './take';\nimport { waitAtMost } from './waitAtMost';\n\nexport const streamTransformations = {\n filter,\n take,\n TakeTransformStream,\n skip,\n SkipTransformStream,\n map,\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n reduce,\n ReduceTransformStream,\n retry: retryStream,\n stopAfter,\n stopOn,\n waitAtMost,\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<void>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import { isErrorConstructor, type ErrorConstructor } from '../errors';\nimport { AssertionError, assertThatArray, assertTrue } from './assertions';\n\ntype ErrorCheck<ErrorType> = (error: ErrorType) => boolean;\n\nexport type ThenThrows<ErrorType extends Error> =\n | (() => void)\n | ((errorConstructor: ErrorConstructor<ErrorType>) => void)\n | ((errorCheck: ErrorCheck<ErrorType>) => void)\n | ((\n errorConstructor: ErrorConstructor<ErrorType>,\n errorCheck?: ErrorCheck<ErrorType>,\n ) => void);\n\nexport type DeciderSpecfication<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => void;\n thenNothingHappened: () => void;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => void;\n };\n};\n\nexport const DeciderSpecification = {\n for: <Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Event | Event[];\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n }): DeciderSpecfication<Command, Event> => {\n {\n return (givenEvents: Event | Event[]) => {\n return {\n when: (command: Command) => {\n const handle = () => {\n const existingEvents = Array.isArray(givenEvents)\n ? givenEvents\n : [givenEvents];\n\n const currentState = existingEvents.reduce<State>(\n decider.evolve,\n decider.initialState(),\n );\n\n return decider.decide(command, currentState);\n };\n\n return {\n then: (expectedEvents: Event | Event[]): void => {\n const resultEvents = handle();\n\n const resultEventsArray = Array.isArray(resultEvents)\n ? resultEvents\n : [resultEvents];\n\n const expectedEventsArray = Array.isArray(expectedEvents)\n ? expectedEvents\n : [expectedEvents];\n\n assertThatArray(resultEventsArray).containsOnlyElementsMatching(\n expectedEventsArray,\n );\n },\n thenNothingHappened: (): void => {\n const resultEvents = handle();\n\n const resultEventsArray = Array.isArray(resultEvents)\n ? resultEvents\n : [resultEvents];\n\n assertThatArray(resultEventsArray).isEmpty();\n },\n thenThrows: <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): void => {\n try {\n handle();\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n }\n },\n};\n","import type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n EventStore,\n EventStoreReadEventMetadata,\n ReadStreamOptions,\n ReadStreamResult,\n StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport { type Event, type EventMetaDataOf } from '../typing';\n\nexport type TestEventStream<EventType extends Event = Event> = [\n string,\n EventType[],\n];\n\nexport type EventStoreWrapper<Store extends EventStore> = Store & {\n appendedEvents: Map<string, TestEventStream>;\n setup<EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>>;\n};\n\nexport const WrapEventStore = <Store extends EventStore>(\n eventStore: Store,\n): EventStoreWrapper<Store> => {\n const appendedEvents = new Map<string, TestEventStream>();\n\n const wrapped = {\n ...eventStore,\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<\n AggregateStreamResult<State, StreamPositionTypeOfEventStore<Store>>\n > {\n return eventStore.aggregateStream(streamName, options);\n },\n\n async readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<\n ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >\n > {\n return (await eventStore.readStream(\n streamName,\n options,\n )) as ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >;\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n const result = await eventStore.appendToStream(\n streamName,\n events,\n options,\n );\n\n const currentStream = appendedEvents.get(streamName) ?? [streamName, []];\n\n appendedEvents.set(streamName, [\n streamName,\n [...currentStream[1], ...events],\n ]);\n\n return result;\n },\n\n appendedEvents,\n\n setup: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n return eventStore.appendToStream(streamName, events);\n },\n\n // streamEvents: (): ReadableStream<\n // // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // > => {\n // return eventStore.streamEvents();\n // },\n };\n\n return wrapped as EventStoreWrapper<Store>;\n};\n","import {\n type NodePostgresClient,\n type NodePostgresTransaction,\n type SQL,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n projection,\n type CanHandle,\n type Event,\n type ProjectionHandler,\n type ReadEvent,\n type TypedProjectionDefinition,\n} from '@event-driven-io/emmett';\nimport type { PostgresReadEventMetadata } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionHandlerContext = {\n connectionString: string;\n client: NodePostgresClient;\n execute: SQLExecutor;\n transaction: NodePostgresTransaction;\n};\n\nexport type PostgreSQLProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext\n>;\n\nexport type PostgreSQLProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = TypedProjectionDefinition<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext\n>;\n\nexport type ProjectionHandlerOptions<\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n events: ReadEvent<EventType, EventMetaDataType>[];\n projections: PostgreSQLProjectionDefinition<EventType, EventMetaDataType>[];\n connection: {\n connectionString: string;\n transaction: NodePostgresTransaction;\n };\n};\n\nexport const handleProjections = async <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: ProjectionHandlerOptions<EventType, EventMetaDataType>,\n): Promise<void> => {\n const {\n projections: allProjections,\n events,\n connection: { transaction, connectionString },\n } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n const client = (await transaction.connection.open()) as NodePostgresClient;\n\n for (const projection of projections) {\n await projection.handle(events, {\n connectionString,\n client,\n transaction,\n execute: transaction.execute,\n });\n }\n};\n\nexport const postgreSQLProjection = <\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n definition: PostgreSQLProjectionDefinition<EventType, EventMetaDataType>,\n): PostgreSQLProjectionDefinition =>\n projection<\n EventType,\n EventMetaDataType,\n PostgreSQLProjectionHandlerContext,\n PostgreSQLProjectionDefinition<EventType, EventMetaDataType>\n >(definition) as PostgreSQLProjectionDefinition;\n\nexport const postgreSQLRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL[]> | SQL[],\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: SQL[] = await handle(events, context);\n\n await context.execute.batchCommand(sqls);\n },\n });\n\nexport const postgreSQLRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL> | SQL,\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: SQL[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n\nexport * from './pongo';\nexport * from './postgresProjectionSpec';\n","import { type Dumbo } from '@event-driven-io/dumbo';\nimport {\n assertDeepEqual,\n assertEqual,\n assertIsNotNull,\n assertThatArray,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoCollection,\n type PongoDocument,\n type PongoFilter,\n type WithId,\n} from '@event-driven-io/pongo';\nimport { type PostgreSQLProjectionAssert } from '..';\n\nexport type PongoAssertOptions = {\n inCollection: string;\n inDatabase?: string;\n};\n\nconst withCollection = (\n handle: (collection: PongoCollection<PongoDocument>) => Promise<void>,\n options: {\n pool: Dumbo;\n connectionString: string;\n } & PongoAssertOptions,\n) => {\n const { pool, connectionString, inDatabase, inCollection } = options;\n\n return pool.withConnection(async (connection) => {\n const pongo = pongoClient(connectionString, {\n connectionOptions: { connection },\n });\n try {\n const collection = pongo.db(inDatabase).collection(inCollection);\n\n return handle(collection);\n } finally {\n await pongo.close();\n }\n });\n};\n\nconst withoutIdAndVersion = <Doc extends PongoDocument | WithId<PongoDocument>>(\n doc: Doc,\n) => {\n const { _id, _version, ...without } = doc;\n\n return without;\n};\n\nconst assertDocumentsEqual = <\n Doc extends PongoDocument | WithId<PongoDocument>,\n>(\n actual: PongoDocument,\n expected: Doc,\n) => {\n if ('_id' in expected)\n assertEqual(\n expected._id,\n actual._id,\n // eslint-disable-next-line @typescript-eslint/restrict-template-expressions\n `Document ids are not matching! Expected: ${expected._id}, Actual: ${actual._id}`,\n );\n\n return assertDeepEqual(\n withoutIdAndVersion(actual),\n withoutIdAndVersion(expected),\n );\n};\n\ntype FilterOrId<Doc extends PongoDocument | WithId<PongoDocument>> =\n | { withId: string }\n | {\n matchingFilter: PongoFilter<Doc>;\n };\n\nexport const documentExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n\n assertDocumentsEqual(result, document);\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsAreTheSame =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n documents: Doc[],\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n documents.length,\n result.length,\n 'Different Documents Count than expected',\n );\n\n for (let i = 0; i < documents.length; i++) {\n assertThatArray(result as Doc[]).contains(documents[i]!);\n }\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsMatchingHaveCount =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n expectedCount: number,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n expectedCount,\n result.length,\n 'Different Documents Count than expected',\n );\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentMatchingExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertThatArray(result).isNotEmpty();\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentDoesNotExist =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n },\n { ...options, ...assertOptions },\n );\n\nexport const expectPongoDocuments = {\n fromCollection: <Doc extends PongoDocument | WithId<PongoDocument>>(\n collectionName: string,\n ) => {\n return {\n withId: (id: string) => {\n return {\n toBeEqual: (document: Doc) =>\n documentExists(document, {\n withId: id,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n withId: id,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n withId: id,\n inCollection: collectionName,\n }),\n };\n },\n matching: <Doc extends PongoDocument | WithId<PongoDocument>>(\n filter: PongoFilter<Doc>,\n ) => {\n return {\n toBeTheSame: (documents: Doc[]) =>\n documentsAreTheSame<Doc>(documents, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toHaveCount: (expectedCount: number) =>\n documentsMatchingHaveCount(expectedCount, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n };\n },\n };\n },\n};\n","import {\n type CanHandle,\n type Event,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoClient,\n type PongoDocument,\n} from '@event-driven-io/pongo';\nimport {\n postgreSQLProjection,\n type PostgreSQLProjectionDefinition,\n type PostgreSQLProjectionHandlerContext,\n} from '..';\nimport type { PostgresReadEventMetadata } from '../../postgreSQLEventStore';\n\nexport type PongoProjectionHandlerContext =\n PostgreSQLProjectionHandlerContext & {\n pongo: PongoClient;\n };\n\nexport type PongoWithNotNullDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | ((\n document: Document,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Document | null)\n | ((\n document: Document,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoWithNullableDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | ((\n document: Document | null,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Document | null)\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> =\n | PongoWithNotNullDocumentEvolve<Document, EventType, EventMetaDataType>\n | PongoWithNullableDocumentEvolve<Document, EventType, EventMetaDataType>;\n\nexport type PongoProjectionOptions<\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n handle: (\n events: ReadEvent<EventType, EventMetaDataType>[],\n context: PongoProjectionHandlerContext,\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoProjection = <\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>({\n handle,\n canHandle,\n}: PongoProjectionOptions<\n EventType,\n EventMetaDataType\n>): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType, EventMetaDataType>({\n canHandle,\n handle: async (events, context) => {\n const { connectionString, client } = context;\n const pongo = pongoClient(connectionString, {\n connectionOptions: { client },\n });\n await handle(events, {\n ...context,\n pongo,\n });\n },\n });\n\nexport type PongoMultiStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n canHandle: CanHandle<EventType>;\n\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n} & (\n | {\n evolve: PongoWithNullableDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: PongoWithNotNullDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Document;\n }\n);\n\nexport const pongoMultiStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: PongoMultiStreamProjectionOptions<\n Document,\n EventType,\n EventMetaDataType\n >,\n): PostgreSQLProjectionDefinition => {\n const { collectionName, getDocumentId, canHandle } = options;\n\n return pongoProjection({\n handle: async (events, { pongo }) => {\n const collection = pongo.db().collection<Document>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), async (document) => {\n return 'initialState' in options\n ? await options.evolve(\n document ?? options.initialState(),\n event as ReadEvent<EventType, EventMetaDataType>,\n )\n : await options.evolve(\n document,\n event as ReadEvent<EventType, EventMetaDataType>,\n );\n });\n }\n },\n canHandle,\n });\n};\n\nexport type PongoSingleStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = {\n canHandle: CanHandle<EventType>;\n\n collectionName: string;\n} & (\n | {\n evolve: PongoWithNullableDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: PongoWithNotNullDocumentEvolve<\n Document,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Document;\n }\n);\n\nexport const pongoSingleStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n options: PongoSingleStreamProjectionOptions<\n Document,\n EventType,\n EventMetaDataType\n >,\n): PostgreSQLProjectionDefinition => {\n return pongoMultiStreamProjection<Document, EventType, EventMetaDataType>({\n ...options,\n getDocumentId: (event) => event.metadata.streamName,\n });\n};\n","import {\n dumbo,\n type Dumbo,\n type DumboOptions,\n type QueryResultRow,\n type SQL,\n} from '@event-driven-io/dumbo';\nimport {\n assertFails,\n AssertionError,\n assertThatArray,\n assertTrue,\n isErrorConstructor,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { handleProjections, type PostgreSQLProjectionDefinition } from '.';\nimport type { PostgresReadEventMetadata } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type PostgreSQLProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type PostgreSQLProjectionSpec<EventType extends Event> = (\n givenEvents: PostgreSQLProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n options?: PostgreSQLProjectionSpecWhenOptions,\n ) => {\n then: (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type PostgreSQLProjectionAssert = (options: {\n pool: Dumbo;\n connectionString: string;\n}) => Promise<void | boolean>;\n\nexport type PostgreSQLProjectionSpecOptions = {\n projection: PostgreSQLProjectionDefinition;\n} & DumboOptions;\n\nexport const PostgreSQLProjectionSpec = {\n for: <EventType extends Event>(\n options: PostgreSQLProjectionSpecOptions,\n ): PostgreSQLProjectionSpec<EventType> => {\n {\n const { projection, ...dumoOptions } = options;\n const { connectionString } = dumoOptions;\n\n return (givenEvents: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n options?: PostgreSQLProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, PostgresReadEventMetadata>[] =\n [];\n\n const run = async (pool: Dumbo) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: PostgresReadEventMetadata = {\n globalPosition: ++globalPosition,\n streamPosition: globalPosition,\n streamName: `test-${uuid()}`,\n eventId: uuid(),\n };\n\n allEvents.push({\n ...event,\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n PostgresReadEventMetadata\n >,\n });\n }\n\n await pool.withTransaction((transaction) =>\n handleProjections({\n events: allEvents,\n projections: [projection],\n connection: {\n connectionString,\n transaction,\n },\n }),\n );\n };\n\n return {\n then: async (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n\n const succeeded = await assert({ pool, connectionString });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await pool.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await pool.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n streamName: string,\n event: PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>,\n): PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n PostgresReadEventMetadata = PostgresReadEventMetadata,\n>(\n streamName: string,\n events: PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>[],\n): PostgreSQLProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\nexport const assertSQLQueryResultMatches =\n <T extends QueryResultRow>(sql: SQL, rows: T[]): PostgreSQLProjectionAssert =>\n async ({ pool: { execute } }) => {\n const result = await execute.query<T>(sql);\n\n assertThatArray(rows).containsExactlyInAnyOrder(result.rows);\n };\n\nexport const expectSQL = {\n query: (sql: SQL) => ({\n resultRows: {\n toBeTheSame: <T extends QueryResultRow>(rows: T[]) =>\n assertSQLQueryResultMatches(sql, rows),\n },\n }),\n};\n","import { type NodePostgresPool, type SQL } from '@event-driven-io/dumbo';\nimport { appendEventsSQL } from './appendToStream';\nimport { storeSubscriptionCheckpointSQL } from './storeSubscriptionCheckpoint';\nimport {\n addDefaultPartition,\n addEventsPartitions,\n addModuleForAllTenantsSQL,\n addModuleSQL,\n addTablePartitions,\n addTenantForAllModulesSQL,\n addTenantSQL,\n eventsTableSQL,\n sanitizeNameSQL,\n streamsTableSQL,\n subscriptionsTableSQL,\n} from './tables';\n\nexport * from './appendToStream';\nexport * from './readLastMessageGlobalPosition';\nexport * from './readMessagesBatch';\nexport * from './readStream';\nexport * from './readSubscriptionCheckpoint';\nexport * from './storeSubscriptionCheckpoint';\nexport * from './tables';\nexport * from './typing';\n\nexport const schemaSQL: SQL[] = [\n streamsTableSQL,\n eventsTableSQL,\n subscriptionsTableSQL,\n sanitizeNameSQL,\n addTablePartitions,\n addEventsPartitions,\n addModuleSQL,\n addTenantSQL,\n addModuleForAllTenantsSQL,\n addTenantForAllModulesSQL,\n appendEventsSQL,\n addDefaultPartition,\n storeSubscriptionCheckpointSQL,\n];\n\nexport const createEventStoreSchema = async (\n pool: NodePostgresPool,\n): Promise<void> => {\n await pool.withTransaction(({ execute }) => execute.batchCommand(schemaSQL));\n};\n","import {\n rawSql,\n single,\n sql,\n type NodePostgresPool,\n type NodePostgresTransaction,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n JSONParser,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AppendToStreamOptions,\n type Event,\n type ExpectedStreamVersion,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { defaultTag, eventsTable, streamsTable } from './typing';\n\nexport const appendEventsSQL = rawSql(\n `CREATE OR REPLACE FUNCTION emt_append_event(\n v_event_ids text[],\n v_events_data jsonb[],\n v_events_metadata jsonb[],\n v_event_schema_versions text[],\n v_event_types text[],\n v_stream_id text,\n v_stream_type text,\n v_expected_stream_position bigint DEFAULT NULL,\n v_partition text DEFAULT emt_sanitize_name('default_partition')\n ) RETURNS TABLE (\n success boolean,\n next_stream_position bigint,\n last_global_position bigint,\n transaction_id xid8\n ) LANGUAGE plpgsql\n AS $$\n DECLARE\n v_next_stream_position bigint;\n v_position bigint;\n v_updated_rows int;\n v_transaction_id xid8;\n v_last_global_position bigint;\n BEGIN\n IF v_expected_stream_position IS NULL THEN\n SELECT COALESCE(max(stream_position), 0) INTO v_expected_stream_position\n FROM ${streamsTable.name}\n WHERE stream_id = v_stream_id AND partition = v_partition;\n END IF;\n\n v_next_stream_position := v_expected_stream_position + array_upper(v_events_data, 1);\n v_transaction_id := pg_current_xact_id();\n\n WITH ev AS (\n SELECT row_number() OVER () + v_expected_stream_position AS stream_position, \n event_data, \n event_metadata, \n schema_version, \n event_id, \n event_type\n FROM (\n SELECT *\n FROM \n unnest(v_event_ids, v_events_data, v_events_metadata, v_event_schema_versions, v_event_types) \n AS event(event_id, event_data, event_metadata, schema_version, event_type)\n ) AS event\n ),\n all_events_insert AS (\n INSERT INTO ${eventsTable.name}\n (stream_id, stream_position, partition, event_data, event_metadata, event_schema_version, event_type, event_id, transaction_id)\n SELECT \n v_stream_id, ev.stream_position, v_partition, ev.event_data, ev.event_metadata, ev.schema_version, ev.event_type, ev.event_id, v_transaction_id\n FROM ev\n RETURNING global_position\n )\n SELECT \n max(global_position) INTO v_last_global_position \n FROM \n all_events_insert;\n\n\n IF v_expected_stream_position = 0 THEN\n INSERT INTO ${streamsTable.name}\n (stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)\n VALUES\n (v_stream_id, v_next_stream_position, v_partition, v_stream_type, '{}', FALSE);\n ELSE\n UPDATE ${streamsTable.name} as s \n SET stream_position = v_next_stream_position\n WHERE stream_id = v_stream_id AND stream_position = v_expected_stream_position AND partition = v_partition AND is_archived = FALSE;\n\n get diagnostics v_updated_rows = row_count;\n\n IF v_updated_rows = 0 THEN\n RETURN QUERY SELECT FALSE, NULL::bigint, NULL::bigint, NULL::xid8;\n RETURN;\n END IF;\n END IF;\n\n RETURN QUERY SELECT TRUE, v_next_stream_position, v_last_global_position, v_transaction_id;\n END;\n $$;\n `,\n);\n\ntype AppendEventResult =\n | {\n success: true;\n nextStreamPosition: bigint;\n lastGlobalPosition: bigint;\n transactionId: string;\n }\n | { success: false };\n\nexport type AppendToStreamPreCommitHook = (\n events: ReadEvent[],\n context: {\n transaction: NodePostgresTransaction;\n },\n) => Promise<void>;\n\nexport const appendToStream = (\n pool: NodePostgresPool,\n streamName: string,\n streamType: string,\n events: Event[],\n options?: AppendToStreamOptions & {\n partition?: string;\n preCommitHook?: AppendToStreamPreCommitHook;\n },\n): Promise<AppendEventResult> =>\n pool.withTransaction<AppendEventResult>(async (transaction) => {\n const { execute } = transaction;\n\n if (events.length === 0)\n return { success: false, result: { success: false } };\n\n let appendResult: AppendEventSqlResult;\n\n try {\n const expectedStreamVersion = toExpectedVersion(\n options?.expectedStreamVersion,\n );\n\n const eventsToAppend: ReadEvent[] = events.map((e, i) => ({\n ...e,\n metadata: {\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(i),\n ...('metadata' in e ? (e.metadata ?? {}) : {}),\n },\n }));\n\n // TODO: return global positions from append raw and other generated data\n appendResult = await appendEventsRaw(\n execute,\n streamName,\n streamType,\n eventsToAppend,\n {\n expectedStreamVersion,\n },\n );\n\n if (options?.preCommitHook)\n await options.preCommitHook(eventsToAppend, { transaction });\n } catch (error) {\n if (!isOptimisticConcurrencyError(error)) throw error;\n\n appendResult = {\n success: false,\n last_global_position: null,\n next_stream_position: null,\n transaction_id: null,\n };\n }\n\n const {\n success,\n next_stream_position,\n last_global_position,\n transaction_id,\n } = appendResult;\n\n return {\n success,\n result:\n success &&\n next_stream_position &&\n last_global_position &&\n transaction_id\n ? {\n success: true,\n nextStreamPosition: BigInt(next_stream_position),\n lastGlobalPosition: BigInt(last_global_position),\n transactionId: transaction_id,\n }\n : { success: false },\n };\n });\n\nconst toExpectedVersion = (\n expected: ExpectedStreamVersion | undefined,\n): bigint | null => {\n if (expected === undefined) return null;\n\n if (expected === NO_CONCURRENCY_CHECK) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_DOES_NOT_EXIST) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_EXISTS) return null;\n\n return expected as bigint;\n};\n\nconst isOptimisticConcurrencyError = (error: unknown): boolean =>\n error instanceof Error && 'code' in error && error.code === '23505';\n\ntype AppendEventSqlResult = {\n success: boolean;\n next_stream_position: string | null;\n last_global_position: string | null;\n transaction_id: string | null | undefined;\n};\n\nconst appendEventsRaw = (\n execute: SQLExecutor,\n streamId: string,\n streamType: string,\n events: ReadEvent[],\n options?: {\n expectedStreamVersion: bigint | null;\n partition?: string;\n },\n): Promise<AppendEventSqlResult> =>\n single(\n execute.command<AppendEventSqlResult>(\n sql(\n `SELECT * FROM emt_append_event(\n ARRAY[%s]::text[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::text[],\n ARRAY[%s]::text[],\n %L::text,\n %L::text,\n %s::bigint,\n %L::text\n )`,\n events.map((e) => sql('%L', e.metadata.eventId)).join(','),\n events.map((e) => sql('%L', JSONParser.stringify(e.data))).join(','),\n events\n .map((e) => sql('%L', JSONParser.stringify(e.metadata ?? {})))\n .join(','),\n events.map(() => `'1'`).join(','),\n events.map((e) => sql('%L', e.type)).join(','),\n streamId,\n streamType,\n options?.expectedStreamVersion ?? 'NULL',\n options?.partition ?? defaultTag,\n ),\n ),\n );\n","export const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = 'emt:default';\n\nexport const globalNames = {\n module: `${emmettPrefix}:module:${globalTag}`,\n tenant: `${emmettPrefix}:tenant:${globalTag}`,\n};\n\nconst columns = {\n partition: {\n name: 'partition',\n },\n isArchived: { name: 'is_archived' },\n};\n\nexport const streamsTable = {\n name: `${emmettPrefix}_streams`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const eventsTable = {\n name: `${emmettPrefix}_events`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const subscriptionsTable = {\n name: `${emmettPrefix}_subscriptions`,\n};\n","import { single, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, subscriptionsTable } from './typing';\n\nexport const storeSubscriptionCheckpointSQL = sql(`\nCREATE OR REPLACE FUNCTION store_subscription_checkpoint(\n p_subscription_id VARCHAR(100),\n p_version BIGINT,\n p_position BIGINT,\n p_check_position BIGINT,\n p_transaction_id xid8,\n p_partition TEXT DEFAULT '${defaultTag}'\n) RETURNS INT AS $$\nDECLARE\n current_position BIGINT;\nBEGIN\n -- Handle the case when p_check_position is provided\n IF p_check_position IS NOT NULL THEN\n -- Try to update if the position matches p_check_position\n UPDATE \"${subscriptionsTable.name}\"\n SET \n \"last_processed_position\" = p_position, \n \"last_processed_transaction_id\" = p_transaction_id\n WHERE \"subscription_id\" = p_subscription_id AND \"last_processed_position\" = p_check_position AND \"partition\" = p_partition;\n\n IF FOUND THEN\n RETURN 1; -- Successfully updated\n END IF;\n\n -- Retrieve the current position\n SELECT \"last_processed_position\" INTO current_position\n FROM \"${subscriptionsTable.name}\"\n WHERE \"subscription_id\" = p_subscription_id AND \"partition\" = p_partition;\n\n -- Return appropriate codes based on current position\n IF current_position = p_position THEN\n RETURN 0; -- Idempotent check: position already set\n ELSIF current_position > p_check_position THEN\n RETURN 2; -- Failure: current position is greater\n ELSE\n RETURN 2; -- Default failure case for mismatched positions\n END IF;\n END IF;\n\n -- Handle the case when p_check_position is NULL: Insert if not exists\n BEGIN\n INSERT INTO \"${subscriptionsTable.name}\"(\"subscription_id\", \"version\", \"last_processed_position\", \"partition\", \"last_processed_transaction_id\")\n VALUES (p_subscription_id, p_version, p_position, p_partition, p_transaction_id);\n RETURN 1; -- Successfully inserted\n EXCEPTION WHEN unique_violation THEN\n -- If insertion failed, it means the row already exists\n SELECT \"last_processed_position\" INTO current_position\n FROM \"${subscriptionsTable.name}\"\n WHERE \"subscription_id\" = p_subscription_id AND \"partition\" = p_partition;\n\n IF current_position = p_position THEN\n RETURN 0; -- Idempotent check: position already set\n ELSE\n RETURN 2; -- Insertion failed, row already exists with different position\n END IF;\n END;\nEND;\n$$ LANGUAGE plpgsql;\n`);\n\nexport type StoreLastProcessedSubscriptionPositionResult<\n Position extends bigint | null = bigint,\n> =\n | {\n success: true;\n newPosition: Position;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint | null>>;\nexport function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint>>;\nexport async function storeSubscriptionCheckpoint(\n execute: SQLExecutor,\n options: {\n subscriptionId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedSubscriptionPositionResult<bigint | null>> {\n try {\n const { result } = await single(\n execute.command<{ result: 0 | 1 | 2 }>(\n sql(\n `SELECT store_subscription_checkpoint(%L, %s, %L, %L, pg_current_xact_id(), %L) as result;`,\n options.subscriptionId,\n options.version ?? 1,\n options.newPosition,\n options.lastProcessedPosition,\n options.partition ?? defaultTag,\n ),\n ),\n );\n\n return result === 1\n ? { success: true, newPosition: options.newPosition }\n : { success: false, reason: result === 0 ? 'IGNORED' : 'MISMATCH' };\n } catch (error) {\n console.log(error);\n throw error;\n }\n}\n","import { rawSql } from '@event-driven-io/dumbo';\nimport {\n defaultTag,\n eventsTable,\n globalTag,\n streamsTable,\n subscriptionsTable,\n} from './typing';\n\nexport const streamsTableSQL = rawSql(\n `CREATE TABLE IF NOT EXISTS ${streamsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}__${globalTag}',\n stream_type TEXT NOT NULL,\n stream_metadata JSONB NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n PRIMARY KEY (stream_id, stream_position, partition, is_archived),\n UNIQUE (stream_id, partition, is_archived)\n ) PARTITION BY LIST (partition);`,\n);\n\nexport const eventsTableSQL = rawSql(\n `\n CREATE SEQUENCE IF NOT EXISTS emt_global_event_position;\n\n CREATE TABLE IF NOT EXISTS ${eventsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n event_data JSONB NOT NULL,\n event_metadata JSONB NOT NULL,\n event_schema_version TEXT NOT NULL,\n event_type TEXT NOT NULL,\n event_id TEXT NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n global_position BIGINT DEFAULT nextval('emt_global_event_position'),\n transaction_id XID8 NOT NULL,\n created TIMESTAMPTZ NOT NULL DEFAULT now(),\n PRIMARY KEY (stream_id, stream_position, partition, is_archived)\n ) PARTITION BY LIST (partition);`,\n);\n\nexport const subscriptionsTableSQL = rawSql(\n `\n CREATE TABLE IF NOT EXISTS ${subscriptionsTable.name}(\n subscription_id TEXT NOT NULL,\n version INT NOT NULL DEFAULT 1,\n partition TEXT NOT NULL DEFAULT '${globalTag}__${globalTag}',\n last_processed_position BIGINT NOT NULL,\n last_processed_transaction_id XID8 NOT NULL,\n PRIMARY KEY (subscription_id, partition, version)\n ) PARTITION BY LIST (partition);\n`,\n);\n\nexport const sanitizeNameSQL = rawSql(\n `CREATE OR REPLACE FUNCTION emt_sanitize_name(input_name TEXT) RETURNS TEXT AS $$\n BEGIN\n RETURN REGEXP_REPLACE(input_name, '[^a-zA-Z0-9_]', '_', 'g');\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addTablePartitions = rawSql(\n `\n CREATE OR REPLACE FUNCTION emt_add_table_partition(tableName TEXT, partition_name TEXT) RETURNS void AS $$\n DECLARE\n v_main_partiton_name TEXT;\n v_active_partiton_name TEXT;\n v_archived_partiton_name TEXT;\n BEGIN \n v_main_partiton_name := emt_sanitize_name(tableName || '_' || partition_name);\n v_active_partiton_name := emt_sanitize_name(v_main_partiton_name || '_active');\n v_archived_partiton_name := emt_sanitize_name(v_main_partiton_name || '_archived');\n\n\n -- create default partition\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (%L) PARTITION BY LIST (is_archived);',\n v_main_partiton_name, tableName, partition_name\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (FALSE);',\n v_active_partiton_name, v_main_partiton_name\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (TRUE);',\n v_archived_partiton_name, v_main_partiton_name\n );\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addEventsPartitions = rawSql(\n `\n CREATE OR REPLACE FUNCTION emt_add_partition(partition_name TEXT) RETURNS void AS $$\n BEGIN \n PERFORM emt_add_table_partition('${eventsTable.name}', partition_name);\n PERFORM emt_add_table_partition('${streamsTable.name}', partition_name);\n\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (%L);',\n emt_sanitize_name('${subscriptionsTable.name}' || '_' || partition_name), '${subscriptionsTable.name}', partition_name\n );\n END;\n $$ LANGUAGE plpgsql;`,\n);\n\nexport const addModuleSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_module(new_module TEXT) RETURNS void AS $$\n BEGIN\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}'), '${eventsTable.name}', new_module, '${globalTag}'\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}' || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}' || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}'), '${streamsTable.name}', new_module, '${globalTag}'\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}' || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}' || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || '${globalTag}')\n );\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addTenantSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_tenant(new_module TEXT, new_tenant TEXT) RETURNS void AS $$\n BEGIN\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant), '${eventsTable.name}', new_module, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant), '${streamsTable.name}', new_module, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || new_tenant)\n );\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addModuleForAllTenantsSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_module_for_all_tenants(new_module TEXT) RETURNS void AS $$\n DECLARE\n tenant_record RECORD;\n BEGIN\n PERFORM add_module(new_module);\n \n FOR tenant_record IN SELECT DISTINCT tenant FROM ${eventsTable.name}\n LOOP\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant), '${eventsTable.name}', new_module, tenant_record.tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant), '${streamsTable.name}', new_module, tenant_record.tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || new_module || '__' || tenant_record.tenant)\n );\n END LOOP;\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addTenantForAllModulesSQL = rawSql(\n `\n CREATE OR REPLACE FUNCTION add_tenant_for_all_modules(new_tenant TEXT) RETURNS void AS $$\n DECLARE\n module_record RECORD;\n BEGIN\n FOR module_record IN SELECT DISTINCT partitionname FROM pg_partman.part_config WHERE parent_table = '${eventsTable.name}'\n LOOP\n -- For ${eventsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant), '${eventsTable.name}', module_record.partitionname, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${eventsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n -- For ${streamsTable.name} table\n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I PARTITION OF %I\n FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant), '${streamsTable.name}', module_record.partitionname, new_tenant\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I\n FOR VALUES IN (FALSE);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n \n EXECUTE format('\n CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I\n FOR VALUES IN (TRUE);',\n emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${streamsTable.name}_' || module_record.partitionname || '__' || new_tenant)\n );\n END LOOP;\n END;\n $$ LANGUAGE plpgsql;\n `,\n);\n\nexport const addDefaultPartition = rawSql(\n `SELECT emt_add_partition('${defaultTag}');`,\n);\n","import { singleOrNull, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadLastMessageGlobalPositionSqlResult = {\n global_position: string;\n};\n\nexport type ReadLastMessageGlobalPositionResult = {\n currentGlobalPosition: bigint | null;\n};\n\nexport const readLastMessageGlobalPosition = async (\n execute: SQLExecutor,\n options?: { partition?: string },\n): Promise<ReadLastMessageGlobalPositionResult> => {\n const result = await singleOrNull(\n execute.query<ReadLastMessageGlobalPositionSqlResult>(\n sql(\n `SELECT global_position\n FROM ${eventsTable.name}\n WHERE partition = %L AND is_archived = FALSE AND transaction_id < pg_snapshot_xmin(pg_current_snapshot())\n ORDER BY transaction_id, global_position\n LIMIT 1`,\n options?.partition ?? defaultTag,\n ),\n ),\n );\n\n return {\n currentGlobalPosition:\n result !== null ? BigInt(result.global_position) : null,\n };\n};\n","import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n type CombinedReadEventMetadata,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadata,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadMessagesBatchSqlResult<EventType extends Event> = {\n stream_position: string;\n stream_id: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport type ReadMessagesBatchOptions =\n | {\n after: bigint;\n batchSize: number;\n }\n | {\n from: bigint;\n batchSize: number;\n }\n | { to: bigint; batchSize: number }\n | { from: bigint; to: bigint };\n\nexport type ReadMessagesBatchResult<\n EventType extends Event,\n ReadEventMetadataType extends ReadEventMetadata = ReadEventMetadata,\n> = {\n currentGlobalPosition: bigint;\n messages: ReadEvent<EventType, ReadEventMetadataType>[];\n areEventsLeft: boolean;\n};\n\nexport const readMessagesBatch = async <\n MessageType extends Event,\n ReadEventMetadataType extends\n ReadEventMetadataWithGlobalPosition = ReadEventMetadataWithGlobalPosition,\n>(\n execute: SQLExecutor,\n options: ReadMessagesBatchOptions & { partition?: string },\n): Promise<ReadMessagesBatchResult<MessageType, ReadEventMetadataType>> => {\n const from =\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after + 1n\n : 0n;\n const batchSize =\n options && 'batchSize' in options\n ? options.batchSize\n : options.to - options.from;\n\n const fromCondition: string =\n from !== -0n ? `AND global_position >= ${from}` : '';\n\n const toCondition =\n 'to' in options ? `AND global_position <= ${options.to}` : '';\n\n const limitCondition =\n 'batchSize' in options ? `LIMIT ${options.batchSize}` : '';\n\n const events: ReadEvent<MessageType, ReadEventMetadataType>[] = await mapRows(\n execute.query<ReadMessagesBatchSqlResult<MessageType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE partition = %L AND is_archived = FALSE AND transaction_id < pg_snapshot_xmin(pg_current_snapshot()) ${fromCondition} ${toCondition}\n ORDER BY transaction_id, global_position\n ${limitCondition}`,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = {\n type: row.event_type,\n data: row.event_data,\n metadata: row.event_metadata,\n } as unknown as MessageType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n eventId: row.event_id,\n streamName: row.stream_id,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n metadata: metadata as CombinedReadEventMetadata<\n MessageType,\n ReadEventMetadataType\n >,\n };\n },\n );\n\n return events.length > 0\n ? {\n currentGlobalPosition:\n events[events.length - 1]!.metadata.globalPosition,\n messages: events,\n areEventsLeft: events.length === batchSize,\n }\n : {\n currentGlobalPosition:\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after\n : 0n,\n messages: [],\n areEventsLeft: false,\n };\n};\n","import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n type CombinedReadEventMetadata,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { PostgreSQLEventStoreDefaultStreamVersion } from '../postgreSQLEventStore';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadStreamSqlResult<EventType extends Event> = {\n stream_position: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n execute: SQLExecutor,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const events: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n await mapRows(\n execute.query<ReadStreamSqlResult<EventType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${fromCondition} ${toCondition}`,\n streamId,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = {\n type: row.event_type,\n data: row.event_data,\n metadata: row.event_metadata,\n } as unknown as EventType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n eventId: row.event_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n metadata: metadata as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n },\n );\n\n return events.length > 0\n ? {\n currentStreamVersion:\n events[events.length - 1]!.metadata.streamPosition,\n events,\n streamExists: true,\n }\n : {\n currentStreamVersion: PostgreSQLEventStoreDefaultStreamVersion,\n events: [],\n streamExists: false,\n };\n};\n","import { singleOrNull, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport { defaultTag, subscriptionsTable } from './typing';\n\ntype ReadSubscriptionCheckpointSqlResult = {\n last_processed_position: string;\n};\n\nexport type ReadSubscriptionCheckpointResult = {\n lastProcessedPosition: bigint | null;\n};\n\nexport const readSubscriptionCheckpoint = async (\n execute: SQLExecutor,\n options: { subscriptionId: string; partition?: string },\n): Promise<ReadSubscriptionCheckpointResult> => {\n const result = await singleOrNull(\n execute.query<ReadSubscriptionCheckpointSqlResult>(\n sql(\n `SELECT last_processed_position\n FROM ${subscriptionsTable.name}\n WHERE partition = %L AND subscription_id = %L\n LIMIT 1`,\n options?.partition ?? defaultTag,\n options.subscriptionId,\n ),\n ),\n );\n\n return {\n lastProcessedPosition:\n result !== null ? BigInt(result.last_processed_position) : null,\n };\n};\n","import { type SQLExecutor } from '@event-driven-io/dumbo';\nimport type {\n EmmettError,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { readLastMessageGlobalPosition } from '../../schema/readLastMessageGlobalPosition';\nimport {\n readMessagesBatch,\n type ReadMessagesBatchOptions,\n} from '../../schema/readMessagesBatch';\n\nexport const DefaultPostgreSQLEventStoreSubscriptionBatchSize = 100;\nexport const DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs = 50;\n\nexport type PostgreSQLEventStoreMessagesBatch<EventType extends Event = Event> =\n {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n };\n\nexport type PostgreSQLEventStoreMessagesBatchHandlerResult = void | {\n type: 'STOP';\n reason?: string;\n error?: EmmettError;\n};\n\nexport type PostgreSQLEventStoreMessagesBatchHandler<\n EventType extends Event = Event,\n> = (\n messagesBatch: PostgreSQLEventStoreMessagesBatch<EventType>,\n) =>\n | Promise<PostgreSQLEventStoreMessagesBatchHandlerResult>\n | PostgreSQLEventStoreMessagesBatchHandlerResult;\n\nexport type PostgreSQLEventStoreMessageBatchPullerOptions<\n EventType extends Event = Event,\n> = {\n executor: SQLExecutor;\n pullingFrequencyInMs: number;\n batchSize: number;\n eachBatch: PostgreSQLEventStoreMessagesBatchHandler<EventType>;\n};\n\nexport type PostgreSQLEventStoreMessageBatchPullerStartFrom =\n | { globalPosition: bigint }\n | 'BEGINNING'\n | 'END';\n\nexport type PostgreSQLEventStoreMessageBatchPullerStartOptions = {\n startFrom: PostgreSQLEventStoreMessageBatchPullerStartFrom;\n};\n\nexport type PostgreSQLEventStoreMessageBatchPuller = {\n isRunning: boolean;\n start(\n options: PostgreSQLEventStoreMessageBatchPullerStartOptions,\n ): Promise<void>;\n stop(): Promise<void>;\n};\n\nexport const postgreSQLEventStoreMessageBatchPuller = <\n EventType extends Event = Event,\n>({\n executor,\n batchSize,\n eachBatch,\n pullingFrequencyInMs,\n}: PostgreSQLEventStoreMessageBatchPullerOptions<EventType>): PostgreSQLEventStoreMessageBatchPuller => {\n let isRunning = false;\n\n let start: Promise<void>;\n\n const pullMessages = async (\n options: PostgreSQLEventStoreMessageBatchPullerStartOptions,\n ) => {\n const after =\n options.startFrom === 'BEGINNING'\n ? 0n\n : options.startFrom === 'END'\n ? ((await readLastMessageGlobalPosition(executor))\n .currentGlobalPosition ?? 0n)\n : options.startFrom.globalPosition;\n\n const readMessagesOptions: ReadMessagesBatchOptions = {\n after,\n batchSize,\n };\n\n let waitTime = 100;\n\n do {\n const { messages, currentGlobalPosition, areEventsLeft } =\n await readMessagesBatch<EventType>(executor, readMessagesOptions);\n\n if (messages.length > 0) {\n const result = await eachBatch({ messages });\n\n if (result && result.type === 'STOP') {\n isRunning = false;\n break;\n }\n }\n\n readMessagesOptions.after = currentGlobalPosition;\n\n await new Promise((resolve) => setTimeout(resolve, waitTime));\n\n if (!areEventsLeft) {\n waitTime = Math.min(waitTime * 2, 1000);\n } else {\n waitTime = pullingFrequencyInMs;\n }\n } while (isRunning);\n };\n\n return {\n get isRunning() {\n return isRunning;\n },\n start: (options) => {\n if (isRunning) return start;\n\n start = (async () => {\n isRunning = true;\n\n return pullMessages(options);\n })();\n\n return start;\n },\n stop: async () => {\n if (!isRunning) return;\n isRunning = false;\n await start;\n },\n };\n};\n\nexport const zipPostgreSQLEventStoreMessageBatchPullerStartFrom = (\n options: (PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined)[],\n): PostgreSQLEventStoreMessageBatchPullerStartFrom => {\n if (\n options.length === 0 ||\n options.some((o) => o === undefined || o === 'BEGINNING')\n )\n return 'BEGINNING';\n\n if (options.every((o) => o === 'END')) return 'END';\n\n return options\n .filter((o) => o !== undefined && o !== 'BEGINNING' && o !== 'END')\n .sort((a, b) => (a > b ? 1 : -1))[0]!;\n};\n","import { dumbo } from '@event-driven-io/dumbo';\nimport { EmmettError, type Event } from '@event-driven-io/emmett';\nimport {\n DefaultPostgreSQLEventStoreSubscriptionBatchSize,\n DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs,\n postgreSQLEventStoreMessageBatchPuller,\n zipPostgreSQLEventStoreMessageBatchPullerStartFrom,\n type PostgreSQLEventStoreMessageBatchPuller,\n type PostgreSQLEventStoreMessagesBatchHandler,\n} from './messageBatchProcessing';\nimport {\n postgreSQLEventStoreSubscription,\n type PostgreSQLEventStoreSubscription,\n type PostgreSQLEventStoreSubscriptionOptions,\n} from './postgreSQLEventStoreSubscription';\n\nexport type PostgreSQLEventStoreConsumerOptions = {\n connectionString: string;\n subscriptions?: PostgreSQLEventStoreSubscription[];\n pooling?: {\n batchSize?: number;\n pullingFrequencyInMs?: number;\n };\n};\n\nexport type PostgreSQLEventStoreConsumer = Readonly<{\n connectionString: string;\n isRunning: boolean;\n subscriptions: PostgreSQLEventStoreSubscription[];\n subscribe: <EventType extends Event = Event>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n ) => PostgreSQLEventStoreSubscription<EventType>;\n start: () => Promise<void>;\n stop: () => Promise<void>;\n close: () => Promise<void>;\n}>;\n\nexport const postgreSQLEventStoreConsumer = (\n options: PostgreSQLEventStoreConsumerOptions,\n): PostgreSQLEventStoreConsumer => {\n let isRunning = false;\n const { connectionString, pooling } = options;\n const subscriptions = options.subscriptions ?? [];\n\n let start: Promise<void>;\n\n let currentMessagePooler: PostgreSQLEventStoreMessageBatchPuller | undefined;\n\n const pool = dumbo({ connectionString });\n\n const eachBatch: PostgreSQLEventStoreMessagesBatchHandler = async (\n messagesBatch,\n ) => {\n const activeSubscriptions = subscriptions.filter((s) => s.isActive);\n\n if (activeSubscriptions.length === 0)\n return {\n type: 'STOP',\n reason: 'No active subscriptions',\n };\n\n const result = await Promise.allSettled(\n activeSubscriptions.map((s) => {\n // TODO: Add here filtering to only pass messages that can be handled by subscription\n return s.handle(messagesBatch, { pool });\n }),\n );\n\n return result.some(\n (r) => r.status === 'fulfilled' && r.value?.type !== 'STOP',\n )\n ? undefined\n : {\n type: 'STOP',\n };\n };\n\n const messagePooler = (currentMessagePooler =\n postgreSQLEventStoreMessageBatchPuller({\n executor: pool.execute,\n eachBatch,\n batchSize:\n pooling?.batchSize ?? DefaultPostgreSQLEventStoreSubscriptionBatchSize,\n pullingFrequencyInMs:\n pooling?.pullingFrequencyInMs ??\n DefaultPostgreSQLEventStoreSubscriptionPullingFrequencyInMs,\n }));\n\n const stop = async () => {\n if (!isRunning) return;\n isRunning = false;\n if (currentMessagePooler) {\n await currentMessagePooler.stop();\n currentMessagePooler = undefined;\n }\n await start;\n };\n\n return {\n connectionString,\n subscriptions,\n get isRunning() {\n return isRunning;\n },\n subscribe: <EventType extends Event = Event>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n ): PostgreSQLEventStoreSubscription<EventType> => {\n const subscription = postgreSQLEventStoreSubscription<EventType>(options);\n\n subscriptions.push(subscription);\n\n return subscription;\n },\n start: () => {\n if (isRunning) return start;\n\n start = (async () => {\n if (subscriptions.length === 0)\n return Promise.reject(\n new EmmettError(\n 'Cannot start consumer without at least a single subscription',\n ),\n );\n\n isRunning = true;\n\n const startFrom = zipPostgreSQLEventStoreMessageBatchPullerStartFrom(\n await Promise.all(subscriptions.map((o) => o.start(pool.execute))),\n );\n\n return messagePooler.start({ startFrom });\n })();\n\n return start;\n },\n stop,\n close: async () => {\n await stop();\n await pool.close();\n },\n };\n};\n","import { type Dumbo, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n EmmettError,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport {\n readSubscriptionCheckpoint,\n storeSubscriptionCheckpoint,\n} from '../schema';\nimport type { PostgreSQLEventStoreMessageBatchPullerStartFrom } from './messageBatchProcessing';\n\nexport type PostgreSQLEventStoreSubscriptionEventsBatch<\n EventType extends Event = Event,\n> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type PostgreSQLEventStoreSubscription<EventType extends Event = Event> =\n {\n id: string;\n start: (\n execute: SQLExecutor,\n ) => Promise<PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined>;\n isActive: boolean;\n handle: (\n messagesBatch: PostgreSQLEventStoreSubscriptionEventsBatch<EventType>,\n context: { pool: Dumbo },\n ) => Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult>;\n };\n\nexport const PostgreSQLEventStoreSubscription = {\n result: {\n skip: (options?: {\n reason?: string;\n }): PostgreSQLEventStoreSubscriptionMessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): PostgreSQLEventStoreSubscriptionMessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type PostgreSQLEventStoreSubscriptionMessageHandlerResult =\n | void\n | { type: 'SKIP'; reason?: string }\n | { type: 'STOP'; reason?: string; error?: EmmettError };\n\nexport type PostgreSQLEventStoreSubscriptionEachMessageHandler<\n EventType extends Event = Event,\n> = (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n) =>\n | Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult>\n | PostgreSQLEventStoreSubscriptionMessageHandlerResult;\n\nexport type PostgreSQLEventStoreSubscriptionStartFrom =\n | PostgreSQLEventStoreMessageBatchPullerStartFrom\n | 'CURRENT';\n\nexport type PostgreSQLEventStoreSubscriptionOptions<\n EventType extends Event = Event,\n> = {\n subscriptionId: string;\n version?: number;\n partition?: string;\n startFrom?: PostgreSQLEventStoreSubscriptionStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n eachMessage: PostgreSQLEventStoreSubscriptionEachMessageHandler<EventType>;\n};\n\nexport const postgreSQLEventStoreSubscription = <\n EventType extends Event = Event,\n>(\n options: PostgreSQLEventStoreSubscriptionOptions<EventType>,\n): PostgreSQLEventStoreSubscription => {\n const { eachMessage } = options;\n let isActive = true;\n //let lastProcessedPosition: bigint | null = null;\n\n return {\n id: options.subscriptionId,\n start: async (\n execute: SQLExecutor,\n ): Promise<PostgreSQLEventStoreMessageBatchPullerStartFrom | undefined> => {\n isActive = true;\n if (options.startFrom !== 'CURRENT') return options.startFrom;\n\n const { lastProcessedPosition } = await readSubscriptionCheckpoint(\n execute,\n {\n subscriptionId: options.subscriptionId,\n partition: options.partition,\n },\n );\n\n if (lastProcessedPosition === null) return 'BEGINNING';\n\n return { globalPosition: lastProcessedPosition };\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n { messages },\n { pool },\n ): Promise<PostgreSQLEventStoreSubscriptionMessageHandlerResult> => {\n if (!isActive) return;\n\n return pool.withTransaction(async (tx) => {\n let result:\n | PostgreSQLEventStoreSubscriptionMessageHandlerResult\n | undefined = undefined;\n\n let lastProcessedPosition: bigint | null = null;\n\n for (const message of messages) {\n const typedMessage = message as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >;\n\n const messageProcessingResult = await eachMessage(typedMessage);\n\n // TODO: Add correct handling of the storing checkpoint\n await storeSubscriptionCheckpoint(tx.execute, {\n subscriptionId: options.subscriptionId,\n version: options.version,\n lastProcessedPosition,\n newPosition: typedMessage.metadata.globalPosition,\n partition: options.partition,\n });\n\n lastProcessedPosition = typedMessage.metadata.globalPosition;\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(typedMessage)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n });\n },\n };\n};\n"],"mappings":";AAAA;AAAA,EACE,SAAAA;AAAA,OAMK;;;AECA,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAE9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ;ACLV,IAAM,qBAAqB,CAEhC,WAC0C;AAE1C,SACE,OAAO,WAAW,cAClB,OAAO;EAEP,OAAO,UAAU,gBAAgB;AAErC;AAEO,IAAM,cAAN,MAAM,qBAAoB,MAAM;EAC9B;EAEP,YACE,SACA;AACA,UAAM,YACJ,WAAW,OAAO,YAAY,YAAY,eAAe,UACrD,QAAQ,YACR,SAAS,OAAO,IACd,UACA;AACR,UAAM,UACJ,WAAW,OAAO,YAAY,YAAY,aAAa,UACnD,QAAQ,UACR,SAAS,OAAO,IACd,UACA,2BAA2B,SAAS;AAE5C,UAAM,OAAO;AACb,SAAK,YAAY;AAGjB,WAAO,eAAe,MAAM,aAAY,SAAS;EACnD;AACF;AAEO,IAAM,mBAAN,MAAM,0BAAyB,YAAY;EAChD,YACS,SACA,UACP,SACA;AACA,UAAM;MACJ,WAAW;MACX,SACE,WACA,oBAAoB,SAAS,SAAS,CAAC,2BAA2B,SAAS,SAAS,CAAC;IACzF,CAAC;AATM,SAAA,UAAA;AACA,SAAA,WAAA;AAWP,WAAO,eAAe,MAAM,kBAAiB,SAAS;EACxD;AACF;;;AU/DA,SAAS,MAAMC,aAAY;ACA3B,SAAS,uBAAuB;ACAhC,SAAS,MAAMA,aAAY;ACA3B,SAAS,MAAM,YAAY;AAC3B,SAAS,mBAAAC,wBAAuB;ASDhC,OAAO,WAAW;AaAlB,SAAS,sBAAsB;AEA/B,OAIO;ACJP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC;EAGE,mBAAAA;OAEK;ACLP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ArCezB,IAAM,gBAAgB;AACtB,IAAM,wBACX;AACK,IAAM,uBACX;AAEK,IAAM,yBAAyB,CACpC,SACA,UACA,mBACY;AACZ,MAAI,aAAa,qBAAsB,QAAO;AAE9C,MAAI,YAAY,sBAAuB,QAAO,YAAY;AAE1D,MAAI,YAAY,cAAe,QAAO,YAAY;AAElD,SAAO,YAAY;AACrB;AAEO,IAAM,sCAAsC,CAGjD,SACA,UACA,mBACS;AACT,eAAa;AAEb,MAAI,CAAC,uBAAuB,SAAS,UAAU,cAAc;AAC3D,UAAM,IAAI,6BAA6B,SAAS,QAAQ;AAC5D;AAEO,IAAM,+BAAN,MAAM,sCAEH,iBAAiB;EACzB,YACE,SACA,UACA;AACA,UAAM,SAAS,SAAS,GAAG,UAAU,SAAS,CAAC;AAG/C,WAAO,eAAe,MAAM,8BAA6B,SAAS;EACpE;AACF;AIzDO,IAAM,mCAAmC,CAC9C,0BAGA,UAA6D,CAAC,MAC3D,IAAI,iCAAiC,0BAA0B,OAAO;AAEpE,IAAM,mCAAN,cAAqDC,iBAG1D;EAQA,YACU,0BAGR,UAA6D,CAAC,GAC9D;AACA,UAAM;MACJ,QAAQ,CAAC,WAAW;AAClB,gBAAQ,IAAI,gCAAgC,MAAM;AAClD,aAAK,aAAa;MACpB;IACF,CAAC;AAVO,SAAA,2BAAA;AAWR,SAAK,WAAW,SAAS,YAAY,KAAK;AAE1C,SAAK,2BAA2B;AAEhC,SAAK,cAAc,SAAS,qBAAqB,EAAE;EACrD;EAxBQ,gBAAuC;EAC/B;EACR,aAAsB;EAC9B,IAAW,uBAAuB;AAChC,WAAO,CAAC,KAAK;EACf;EAqBQ,cAAc,UAAkB;AACtC,SAAK,gBAAgB,YAAY,MAAM;AACrC,WAAK,oBAAoB;IAC3B,GAAG,QAAQ;EACb;EAEQ,eAAe;AACrB,QAAI,CAAC,KAAK,cAAe;AAEzB,kBAAc,KAAK,aAAa;AAChC,SAAK,gBAAgB;AACrB,SAAK,aAAa;AAClB,SAAK,yBAAyB,IAAI;EACpC;EAEQ,sBAAsB;AAC5B,QAAI,CAAC,KAAK,SAAS,UAAU,CAAC,KAAK,YAAY;AAC7C,WAAK,aAAa;IACpB;EACF;AACF;AK5DO,IAAM,aAAa,CAAI,MAAS,UAAsB;AAC3D,MAAI,YAAY,IAAI,GAAG;AACrB,WAAO,KAAK,OAAO,KAAK;EAC1B;AAEA,MAAI,MAAM,QAAQ,IAAI,GAAG;AACvB,WACE,MAAM,QAAQ,KAAK,KACnB,KAAK,WAAW,MAAM,UACtB,KAAK,MAAM,CAAC,KAAK,UAAU,WAAW,KAAK,MAAM,KAAK,CAAC,CAAC;EAE5D;AAEA,MACE,OAAO,SAAS,YAChB,OAAO,UAAU,YACjB,SAAS,QACT,UAAU,MACV;AACA,WAAO,SAAS;EAClB;AAEA,MAAI,MAAM,QAAQ,KAAK,EAAG,QAAO;AAEjC,QAAM,QAAQ,OAAO,KAAK,IAAI;AAC9B,QAAM,QAAQ,OAAO,KAAK,KAAK;AAE/B,MACE,MAAM,WAAW,MAAM,UACvB,CAAC,MAAM,MAAM,CAAC,QAAQ,MAAM,SAAS,GAAG,CAAC;AAEzC,WAAO;AAET,aAAW,OAAO,MAAM;AACtB,QAAI,KAAK,GAAG,aAAa,YAAY,MAAM,GAAG,aAAa;AACzD;AAEF,UAAM,UAAU,WAAW,KAAK,GAAG,GAAG,MAAM,GAAG,CAAC;AAChD,QAAI,CAAC,SAAS;AACZ,aAAO;IACT;EACF;AAEA,SAAO;AACT;AAIO,IAAM,cAAc,CAAI,SAAkC;AAC/D,SACE,QACA,OAAO,SAAS,YAChB,YAAY,QACZ,OAAO,KAAK,QAAQ,MAAM;AAE9B;AI/CO,IAAM,aAAa,OACxB,IACA,SACe;AACf,MAAI,SAAS,UAAa,KAAK,YAAY,EAAG,QAAO,GAAG;AAExD,SAAO;IACL,OAAO,SAAS;AACd,UAAI;AACF,eAAO,MAAM,GAAG;MAClB,SAASC,QAAO;AACd,YAAI,MAAM,oBAAoB,CAAC,KAAK,iBAAiBA,MAAK,GAAG;AAC3D,eAAKA,MAAc;QACrB;AACA,cAAMA;MACR;IACF;IACA,QAAQ,EAAE,SAAS,EAAE;EACvB;AACF;AI3BO,IAAM,aAAN,cAAyB,MAAM;EACpC,YAAY,MAAc;AACxB,UAAM,iBAAiB,IAAI,EAAE;EAC/B;AACF;AA0BO,IAAM,aAAa;EACxB,WAAW,CACT,OACA,YACG;AACH,WAAO,KAAK;MACV,SAAS,MAAM,QAAQ,IAAI,KAA6B,IAAI;;;MAG5D,CAAC,GAAG,MAAO,OAAO,MAAM,WAAW,EAAE,SAAS,IAAI;IACpD;EACF;EACA,OAAO,CACL,MACA,YACmB;AACnB,UAAM,SAAkB,KAAK,MAAM,MAAM,SAAS,OAAO;AAEzD,QAAI,SAAS,aAAa,CAAC,SAAS,UAAc,MAAM;AACtD,YAAM,IAAI,WAAW,IAAI;AAE3B,WAAO,SAAS,MACZ,QAAQ,IAAI,MAA8B,IACzC;EACP;AACF;ACoCO,IAAM,aAAa,CAcxB,eAC4B;AWxGvB,IAAM,SAAS,CAAOC,YAC3B,IAAIC,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,QAAID,QAAO,KAAK,GAAG;AACjB,iBAAW,QAAQ,KAAK;IAC1B;EACF;AACF,CAAC;ACPI,IAAM,MAAM,CAAWE,SAC5B,IAAID,iBAA0B;EAC5B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQC,KAAI,KAAK,CAAC;EAC/B;AACF,CAAC;ACLI,IAAM,SAAS,CACpB,SACA,iBACG,IAAI,sBAA4B,SAAS,YAAY;AAEnD,IAAM,wBAAN,cAA0CD,iBAAsB;EAC7D;EACA;EAER,YAAY,SAA0C,cAAiB;AACrE,UAAM;MACJ,WAAW,CAAC,UAAU;AACpB,aAAK,cAAc,KAAK,QAAQ,KAAK,aAAa,KAAK;MACzD;MACA,OAAO,CAAC,eAAe;AACrB,mBAAW,QAAQ,KAAK,WAAW;AACnC,mBAAW,UAAU;MACvB;IACF,CAAC;AAED,SAAK,cAAc;AACnB,SAAK,UAAU;EACjB;AACF;ACjBO,IAAM,cAAc,CAKzB,oBACAE,cAIA,eAAkC,EAAE,SAAS,MAAM,YAAY,GAAG,MAElE,IAAIF,iBAAqC;EACvC,MAAM,YAAY;AAChB;MACE,MAAM,WAAW,oBAAoBE,cAAa,UAAU;MAC5D;IACF,EAAE,MAAM,CAACC,WAAU;AACjB,iBAAW,MAAMA,MAAK;IACxB,CAAC;EACH;AACF,CAAC;AAEH,IAAM,aAAa,OACjB,oBACAD,cAIA,eACkB;AAClB,QAAM,eAAe,mBAAmB;AACxC,QAAM,SAAS,aAAa,UAAU;AAEtC,MAAI;AACF,QAAI;AAEJ,OAAG;AACD,YAAM,SAAS,MAAM,OAAO,KAAK;AACjC,aAAO,OAAO;AAEd,YAAMA,aAAY,QAAQ,UAAU;AAEpC,UAAI,MAAM;AACR,mBAAW,UAAU;MACvB;IACF,SAAS,CAAC;EACZ,UAAA;AACE,WAAO,YAAY;EACrB;AACF;ACxDO,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCF,iBAAsB;EACxD,QAAQ;EACR;EAER,YAAYI,OAAc;AACxB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,aAAK;AACL,YAAI,KAAK,QAAQ,KAAK,MAAM;AAC1B,qBAAW,QAAQ,KAAK;QAC1B;MACF;IACF,CAAC;AAED,SAAK,OAAOA;EACd;AACF;AClBO,IAAM,YAAY,CAAO,kBAC9B,IAAIJ,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;AAExB,QAAI,cAAc,KAAK,GAAG;AACxB,iBAAW,UAAU;IACvB;EACF;AACF,CAAC;ACTI,IAAM,SAAS,CAAO,kBAC3B,IAAIA,iBAA4B;EAC9B,MAAM,UAAU,OAAO,YAAY;AACjC,QAAI,CAAC,cAAc,KAAK,GAAG;AACzB,iBAAW,QAAQ,KAAK;AACxB;IACF;AACA,UAAM,QAAQ,QAAQ;AACtB,eAAW,UAAU;EACvB;AACF,CAAC;ACVI,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCA,kBAAsB;EACxD,QAAQ;EACR;EAER,YAAY,OAAe;AACzB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,YAAI,KAAK,QAAQ,KAAK,OAAO;AAC3B,eAAK;AACL,qBAAW,QAAQ,KAAK;QAC1B,OAAO;AACL,qBAAW,UAAU;QACvB;MACF;IACF,CAAC;AAED,SAAK,QAAQ;EACf;AACF;ACpBO,IAAM,aAAa,CAAO,iBAC/B,IAAIA,kBAA4B;EAC9B,MAAM,YAAY;AAChB,UAAM,YAAY,WAAW,MAAM;AACjC,iBAAW,UAAU;IACvB,GAAG,YAAY;AAEf,UAAM,oBAAoB,WAAW,UAAU,KAAK,UAAU;AAG9D,eAAW,YAAY,MAAM;AAC3B,mBAAa,SAAS;AACtB,wBAAkB;IACpB;EACF;EACA,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;EAC1B;AACF,CAAC;ACNI,IAAM,wBAAwB;EACnC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA,OAAO;EACP;EACA;EACA;AACF;AVnBA,IAAM,EAAE,OAAAK,OAAM,IAAI;AWNX,IAAM,iBAAN,cAA6B,MAAM;EACxC,YAAY,SAAiB;AAC3B,UAAM,OAAO;EACf;AACF;AAEO,IAAM,WAAW,CAAC,UAAmB,WAA6B;AACvE,QAAM,MAAM;AACZ,QAAM,MAAM;AAEZ,WAAS,GAAG;AACZ,WAAS,GAAG;AAEZ,SAAO,OAAO,KAAK,GAAG,EAAE,MAAM,CAAC,QAAgB;AAC7C,QAAI,OAAO,IAAI,GAAG,KAAK,UAAU;AAC/B,aAAO,SAAS,IAAI,GAAG,GAAG,IAAI,GAAG,CAAC;IACpC;AACA,WAAO,IAAI,GAAG,MAAM,IAAI,GAAG;EAC7B,CAAC;AACH;AAEO,IAAM,cAAc,CAAC,YAAqB;AAC/C,QAAM,IAAI,eAAe,WAAW,uCAAuC;AAC7E;AAwGO,IAAM,kBAAkB,CAC7B,QACA,UACA,YACG;AACH,MAAI,CAAC,WAAW,QAAQ,QAAQ;AAC9B,UAAM,IAAI;MACR,WACE;EAAY,WAAW,UAAU,QAAQ,CAAC;;EAAsB,WAAW,UAAU,MAAM,CAAC;IAChG;AACJ;AA4BO,SAAS,WACd,WACA,SAC2B;AAC3B,MAAI,cAAc;AAChB,UAAM,IAAI,eAAe,WAAW,oBAAoB;AAC5D;AAEO,SAAS,SACd,KACA,SACkB;AAClB,MAAI,CAAC,IAAK,OAAM,IAAI,eAAe,WAAW,yBAAyB;AACzE;AAEO,SAAS,YACd,UACA,QACA,SACM;AACN,MAAI,aAAa;AACf,UAAM,IAAI;MACR,GAAG,WAAW,uBAAuB;YAAgB,WAAW,UAAU,QAAQ,CAAC;UAAa,WAAW,UAAU,MAAM,CAAC;IAC9H;AACJ;AAEO,SAAS,eACd,KACA,OACA,SACM;AACN,MAAI,QAAQ;AACV,UAAM,IAAI;MACR,WAAW,sBAAsB,WAAW,UAAU,GAAG,CAAC;IAC5D;AACJ;AAEO,SAAS,gBACd,QACqB;AACrB,iBAAe,QAAQ,IAAI;AAC3B,WAAS,MAAM;AACjB;AAsFO,IAAM,kBAAkB,CAAI,UAAe;AAChD,SAAO;IACL,SAAS,MACP;MACE,MAAM;MACN;MACA,sBAAsB,WAAW,UAAU,KAAK,CAAC;IACnD;IACF,YAAY,MAAM,eAAe,MAAM,QAAQ,GAAG,gBAAgB;IAClE,SAAS,CAAC,WAAmB,YAAY,MAAM,QAAQ,MAAM;IAC7D,kBAAkB,CAAC,UAAe;AAChC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,0BAA0B,CAAC,UAAe;AACxC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,SAAS,GAAG,EAAE,CAAC,CAAC,CAAC;IACpE;IACA,8BAA8B,CAAC,UAAe;AAC5C,kBAAY,MAAM,QAAQ,MAAM,QAAQ,4BAA4B;AACpE,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,SAAS,GAAG,EAAE,CAAC,CAAC,CAAC;IACpE;IACA,2BAA2B,CAAC,UAAe;AACzC,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,qCAAqC,CAAC,UAAe;AACnD,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,iBAAW,MAAM,MAAM,CAAC,OAAO,MAAM,KAAK,CAAC,MAAM,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;IACtE;IACA,2BAA2B,CAAC,UAAe;AACzC,kBAAY,MAAM,QAAQ,MAAM,MAAM;AACtC,eAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;AACrC,mBAAW,WAAW,MAAM,CAAC,GAAG,MAAM,CAAC,CAAC,CAAC;MAC3C;IACF;IACA,iBAAiB,CAAC,SAAY;AAC5B,kBAAY,MAAM,QAAQ,CAAC;AAC3B,iBAAW,WAAW,MAAM,CAAC,GAAG,IAAI,CAAC;IACvC;IACA,UAAU,CAAC,SAAY;AACrB,iBAAW,MAAM,KAAK,CAAC,MAAM,WAAW,GAAG,IAAI,CAAC,CAAC;IACnD;IACA,4BAA4B,CAAC,UAAe;AAC1C;QACE,MACG,IAAI,CAAC,MAAM,MAAM,OAAO,CAAC,MAAM,WAAW,GAAG,CAAC,CAAC,EAAE,MAAM,EACvD,OAAO,CAAC,MAAM,MAAM,CAAC,EAAE,WAAW,MAAM;MAC7C;IACF;IACA,eAAe,CAAC,UAAe;AAC7B,iBAAW,MAAM,KAAK,CAAC,MAAM,MAAM,KAAK,CAAC,MAAM,WAAW,GAAG,CAAC,CAAC,CAAC,CAAC;IACnE;IACA,UAAU,CAAC,YAAkC;AAC3C,iBAAW,MAAM,MAAM,OAAO,CAAC;IACjC;IACA,YAAY,CAAC,YAAkC;AAC7C,iBAAW,MAAM,KAAK,OAAO,CAAC;IAChC;IACA,eAAe,OACb,YACkB;AAClB,iBAAW,QAAQ,OAAO;AACxB,mBAAW,MAAM,QAAQ,IAAI,CAAC;MAChC;IACF;EACF;AACF;;;AnDhVA,OAAe;;;AsD1Bf,OAKO;;;ACLP,OAA2B;AAO3B;AAAA,EACE;AAAA,OAKK;AAQP,IAAM,iBAAiB,CACrB,QACA,YAIG;AACH,QAAM,EAAE,MAAM,kBAAkB,YAAY,aAAa,IAAI;AAE7D,SAAO,KAAK,eAAe,OAAO,eAAe;AAC/C,UAAM,QAAQ,YAAY,kBAAkB;AAAA,MAC1C,mBAAmB,EAAE,WAAW;AAAA,IAClC,CAAC;AACD,QAAI;AACF,YAAM,aAAa,MAAM,GAAG,UAAU,EAAE,WAAW,YAAY;AAE/D,aAAO,OAAO,UAAU;AAAA,IAC1B,UAAE;AACA,YAAM,MAAM,MAAM;AAAA,IACpB;AAAA,EACF,CAAC;AACH;AAEA,IAAM,sBAAsB,CAC1B,QACG;AACH,QAAM,EAAE,KAAK,UAAU,GAAG,QAAQ,IAAI;AAEtC,SAAO;AACT;AAEA,IAAM,uBAAuB,CAG3B,QACA,aACG;AACH,MAAI,SAAS;AACX;AAAA,MACE,SAAS;AAAA,MACT,OAAO;AAAA;AAAA,MAEP,4CAA4C,SAAS,GAAG,aAAa,OAAO,GAAG;AAAA,IACjF;AAEF,SAAO;AAAA,IACL,oBAAoB,MAAM;AAAA,IAC1B,oBAAoB,QAAQ;AAAA,EAC9B;AACF;AAQO,IAAM,iBACX,CACE,UACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM;AAEtB,yBAAqB,QAAQ,QAAQ;AAAA,EACvC;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,sBACX,CACE,WACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA;AAAA,MACE,UAAU;AAAA,MACV,OAAO;AAAA,MACP;AAAA,IACF;AAEA,aAAS,IAAI,GAAG,IAAI,UAAU,QAAQ,KAAK;AACzC,sBAAgB,MAAe,EAAE,SAAS,UAAU,CAAC,CAAE;AAAA,IACzD;AAAA,EACF;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,6BACX,CACE,eACA,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA;AAAA,MACE;AAAA,MACA,OAAO;AAAA,MACP;AAAA,IACF;AAAA,EACF;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,yBACX,CACE,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM,EAAE,WAAW;AAAA,EACrC;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,uBACX,CACE,YAEF,CAAC,kBACC;AAAA,EACE,OAAO,eAAe;AACpB,UAAM,SAAS,MAAM,WAAW;AAAA,MAC9B,YAAY,UACR,EAAE,KAAK,QAAQ,OAAO,IACtB,QAAQ;AAAA,IACd;AAEA,oBAAgB,MAAM;AAAA,EACxB;AAAA,EACA,EAAE,GAAG,SAAS,GAAG,cAAc;AACjC;AAEG,IAAM,uBAAuB;AAAA,EAClC,gBAAgB,CACd,mBACG;AACH,WAAO;AAAA,MACL,QAAQ,CAAC,OAAe;AACtB,eAAO;AAAA,UACL,WAAW,CAAC,aACV,eAAe,UAAU;AAAA,YACvB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,SAAS,MACP,uBAAuB;AAAA,YACrB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,YAAY,MACV,qBAAqB;AAAA,YACnB,QAAQ;AAAA,YACR,cAAc;AAAA,UAChB,CAAC;AAAA,QACL;AAAA,MACF;AAAA,MACA,UAAU,CACRC,YACG;AACH,eAAO;AAAA,UACL,aAAa,CAAC,cACZ,oBAAyB,WAAW;AAAA,YAClC,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,aAAa,CAAC,kBACZ,2BAA2B,eAAe;AAAA,YACxC,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,SAAS,MACP,uBAAuB;AAAA,YACrB,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,UACH,YAAY,MACV,qBAAqB;AAAA,YACnB,gBAAgBA;AAAA,YAChB,cAAc;AAAA,UAChB,CAAC;AAAA,QACL;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;;;ACxOA;AAAA,EACE,eAAAC;AAAA,OAGK;AAgEA,IAAM,kBAAkB,CAI7B;AAAA,EACA;AAAA,EACA;AACF,MAIE,qBAAmD;AAAA,EACjD;AAAA,EACA,QAAQ,OAAO,QAAQ,YAAY;AACjC,UAAM,EAAE,kBAAkB,OAAO,IAAI;AACrC,UAAM,QAAQC,aAAY,kBAAkB;AAAA,MAC1C,mBAAmB,EAAE,OAAO;AAAA,IAC9B,CAAC;AACD,UAAM,OAAO,QAAQ;AAAA,MACnB,GAAG;AAAA,MACH;AAAA,IACF,CAAC;AAAA,EACH;AACF,CAAC;AA8BI,IAAM,6BAA6B,CAMxC,YAKmC;AACnC,QAAM,EAAE,gBAAgB,eAAe,UAAU,IAAI;AAErD,SAAO,gBAAgB;AAAA,IACrB,QAAQ,OAAO,QAAQ,EAAE,MAAM,MAAM;AACnC,YAAM,aAAa,MAAM,GAAG,EAAE,WAAqB,cAAc;AAEjE,iBAAW,SAAS,QAAQ;AAC1B,cAAM,WAAW,OAAO,cAAc,KAAK,GAAG,OAAO,aAAa;AAChE,iBAAO,kBAAkB,UACrB,MAAM,QAAQ;AAAA,YACZ,YAAY,QAAQ,aAAa;AAAA,YACjC;AAAA,UACF,IACA,MAAM,QAAQ;AAAA,YACZ;AAAA,YACA;AAAA,UACF;AAAA,QACN,CAAC;AAAA,MACH;AAAA,IACF;AAAA,IACA;AAAA,EACF,CAAC;AACH;AA6BO,IAAM,8BAA8B,CAMzC,YAKmC;AACnC,SAAO,2BAAmE;AAAA,IACxE,GAAG;AAAA,IACH,eAAe,CAAC,UAAU,MAAM,SAAS;AAAA,EAC3C,CAAC;AACH;;;AC7MA;AAAA,EACE;AAAA,OAKK;AAYP,SAAS,MAAMC,aAAY;AAwCpB,IAAM,2BAA2B;AAAA,EACtC,KAAK,CACH,YACwC;AACxC;AACE,YAAM,EAAE,YAAAC,aAAY,GAAG,YAAY,IAAI;AACvC,YAAM,EAAE,iBAAiB,IAAI;AAE7B,aAAO,CAAC,gBAA4D;AAClE,eAAO;AAAA,UACL,MAAM,CACJ,QACAC,aACG;AACH,kBAAM,YACJ,CAAC;AAEH,kBAAM,MAAM,OAAO,SAAgB;AACjC,kBAAI,iBAAiB;AACrB,oBAAM,gBAAgBA,UAAS,iBAAiB;AAEhD,yBAAW,SAAS;AAAA,gBAClB,GAAG;AAAA,gBACH,GAAG,MAAM,KAAK,EAAE,QAAQ,cAAc,CAAC,EAAE,QAAQ,MAAM,MAAM;AAAA,cAC/D,GAAG;AACD,sBAAM,WAAsC;AAAA,kBAC1C,gBAAgB,EAAE;AAAA,kBAClB,gBAAgB;AAAA,kBAChB,YAAY,QAAQC,MAAK,CAAC;AAAA,kBAC1B,SAASA,MAAK;AAAA,gBAChB;AAEA,0BAAU,KAAK;AAAA,kBACb,GAAG;AAAA,kBACH,UAAU;AAAA,oBACR,GAAG;AAAA,oBACH,GAAI,cAAc,QAAS,MAAM,YAAY,CAAC,IAAK,CAAC;AAAA,kBACtD;AAAA,gBAIF,CAAC;AAAA,cACH;AAEA,oBAAM,KAAK;AAAA,gBAAgB,CAAC,gBAC1B,kBAAkB;AAAA,kBAChB,QAAQ;AAAA,kBACR,aAAa,CAACF,WAAU;AAAA,kBACxB,YAAY;AAAA,oBACV;AAAA,oBACA;AAAA,kBACF;AAAA,gBACF,CAAC;AAAA,cACH;AAAA,YACF;AAEA,mBAAO;AAAA,cACL,MAAM,OACJ,QACA,YACkB;AAClB,sBAAM,OAAO,MAAM,WAAW;AAC9B,oBAAI;AACF,wBAAM,IAAI,IAAI;AAEd,wBAAM,YAAY,MAAM,OAAO,EAAE,MAAM,iBAAiB,CAAC;AAEzD,sBAAI,cAAc,UAAa,cAAc;AAC3C;AAAA,sBACE,WACE;AAAA,oBACJ;AAAA,gBACJ,UAAE;AACA,wBAAM,KAAK,MAAM;AAAA,gBACnB;AAAA,cACF;AAAA,cACA,YAAY,UACP,SACe;AAClB,sBAAM,OAAO,MAAM,WAAW;AAC9B,oBAAI;AACF,wBAAM,IAAI,IAAI;AACd,wBAAM,IAAI,eAAe,kCAAkC;AAAA,gBAC7D,SAAS,OAAO;AACd,sBAAI,iBAAiB,eAAgB,OAAM;AAE3C,sBAAI,KAAK,WAAW,EAAG;AAEvB,sBAAI,CAAC,mBAAmB,KAAK,CAAC,CAAC,GAAG;AAChC;AAAA,sBACE,KAAK,CAAC,EAAE,KAAkB;AAAA,sBAC1B,2CAA2C,OAAO,SAAS,CAAC;AAAA,oBAC9D;AACA;AAAA,kBACF;AAEA;AAAA,oBACE,iBAAiB,KAAK,CAAC;AAAA,oBACvB,yDAAyD,OAAO,SAAS,CAAC;AAAA,kBAC5E;AAEA,sBAAI,KAAK,CAAC,GAAG;AACX;AAAA,sBACE,KAAK,CAAC,EAAE,KAAkB;AAAA,sBAC1B,2CAA2C,OAAO,SAAS,CAAC;AAAA,oBAC9D;AAAA,kBACF;AAAA,gBACF,UAAE;AACA,wBAAM,KAAK,MAAM;AAAA,gBACnB;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAEO,IAAM,gBAAgB,CAK3B,YACA,UACgE;AAChE,SAAO;AAAA,IACL,GAAG;AAAA,IACH,UAAU;AAAA,MACR,GAAI,MAAM,YAAY,CAAC;AAAA,MACvB,YAAY,MAAM,UAAU,cAAc;AAAA,IAC5C;AAAA,EACF;AACF;AAEO,IAAM,iBAAiB,CAK5B,YACA,WACkE;AAClE,SAAO,OAAO,IAAI,CAAC,MAAM,cAAc,YAAY,CAAC,CAAC;AACvD;AAEO,IAAM,oBAAoB;AAE1B,IAAM,8BACX,CAA2BG,MAAU,SACrC,OAAO,EAAE,MAAM,EAAE,QAAQ,EAAE,MAAM;AAC/B,QAAM,SAAS,MAAM,QAAQ,MAASA,IAAG;AAEzC,kBAAgB,IAAI,EAAE,0BAA0B,OAAO,IAAI;AAC7D;AAEK,IAAM,YAAY;AAAA,EACvB,OAAO,CAACA,UAAc;AAAA,IACpB,YAAY;AAAA,MACV,aAAa,CAA2B,SACtC,4BAA4BA,MAAK,IAAI;AAAA,IACzC;AAAA,EACF;AACF;;;AHtKO,IAAM,oBAAoB,OAK/B,YACkB;AAClB,QAAM;AAAA,IACJ,aAAa;AAAA,IACb;AAAA,IACA,YAAY,EAAE,aAAa,iBAAiB;AAAA,EAC9C,IAAI;AAEJ,QAAM,aAAa,OAAO,IAAI,CAAC,MAAM,EAAE,IAAI;AAE3C,QAAM,cAAc,eAAe;AAAA,IAAO,CAAC,MACzC,EAAE,UAAU,KAAK,CAAC,SAAS,WAAW,SAAS,IAAI,CAAC;AAAA,EACtD;AAEA,QAAM,SAAU,MAAM,YAAY,WAAW,KAAK;AAElD,aAAWC,eAAc,aAAa;AACpC,UAAMA,YAAW,OAAO,QAAQ;AAAA,MAC9B;AAAA,MACA;AAAA,MACA;AAAA,MACA,SAAS,YAAY;AAAA,IACvB,CAAC;AAAA,EACH;AACF;AAEO,IAAM,uBAAuB,CAKlC,eAEA,WAKE,UAAU;AAEP,IAAM,kCAAkC,CAC7C,WAIG,cAEH,qBAAgC;AAAA,EAC9B;AAAA,EACA,QAAQ,OAAO,QAAQ,YAAY;AACjC,UAAM,OAAc,MAAM,OAAO,QAAQ,OAAO;AAEhD,UAAM,QAAQ,QAAQ,aAAa,IAAI;AAAA,EACzC;AACF,CAAC;AAEI,IAAM,6BAA6B,CACxC,WAIG,cAEH;AAAA,EACE,OAAO,QAAQ,YAAY;AACzB,UAAM,OAAc,CAAC;AAErB,eAAW,SAAS,QAAQ;AAC1B,WAAK,KAAK,MAAM,OAAO,OAAO,OAAO,CAAC;AAAA,IACxC;AACA,WAAO;AAAA,EACT;AAAA,EACA,GAAG;AACL;;;AItIF,OAAgD;;;ACAhD;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,OAIK;AAWP,SAAS,MAAMC,aAAY;;;AClBpB,IAAM,eAAe;AAErB,IAAM,YAAY;AAClB,IAAM,aAAa;AAEnB,IAAM,cAAc;AAAA,EACzB,QAAQ,GAAG,YAAY,WAAW,SAAS;AAAA,EAC3C,QAAQ,GAAG,YAAY,WAAW,SAAS;AAC7C;AAEA,IAAM,UAAU;AAAA,EACd,WAAW;AAAA,IACT,MAAM;AAAA,EACR;AAAA,EACA,YAAY,EAAE,MAAM,cAAc;AACpC;AAEO,IAAM,eAAe;AAAA,EAC1B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,cAAc;AAAA,EACzB,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,qBAAqB;AAAA,EAChC,MAAM,GAAG,YAAY;AACvB;;;ADdO,IAAM,kBAAkB;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iBA0Be,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAsBV,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAchB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,mBAKtB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgBpC;AAkBO,IAAM,iBAAiB,CAC5B,MACA,YACA,YACA,QACA,YAKA,KAAK,gBAAmC,OAAO,gBAAgB;AAC7D,QAAM,EAAE,QAAQ,IAAI;AAEpB,MAAI,OAAO,WAAW;AACpB,WAAO,EAAE,SAAS,OAAO,QAAQ,EAAE,SAAS,MAAM,EAAE;AAEtD,MAAI;AAEJ,MAAI;AACF,UAAM,wBAAwB;AAAA,MAC5B,SAAS;AAAA,IACX;AAEA,UAAM,iBAA8B,OAAO,IAAI,CAAC,GAAG,OAAO;AAAA,MACxD,GAAG;AAAA,MACH,UAAU;AAAA,QACR;AAAA,QACA,SAASC,MAAK;AAAA,QACd,gBAAgB,OAAO,CAAC;AAAA,QACxB,GAAI,cAAc,IAAK,EAAE,YAAY,CAAC,IAAK,CAAC;AAAA,MAC9C;AAAA,IACF,EAAE;AAGF,mBAAe,MAAM;AAAA,MACnB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,QACE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,SAAS;AACX,YAAM,QAAQ,cAAc,gBAAgB,EAAE,YAAY,CAAC;AAAA,EAC/D,SAAS,OAAO;AACd,QAAI,CAAC,6BAA6B,KAAK,EAAG,OAAM;AAEhD,mBAAe;AAAA,MACb,SAAS;AAAA,MACT,sBAAsB;AAAA,MACtB,sBAAsB;AAAA,MACtB,gBAAgB;AAAA,IAClB;AAAA,EACF;AAEA,QAAM;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,IAAI;AAEJ,SAAO;AAAA,IACL;AAAA,IACA,QACE,WACA,wBACA,wBACA,iBACI;AAAA,MACE,SAAS;AAAA,MACT,oBAAoB,OAAO,oBAAoB;AAAA,MAC/C,oBAAoB,OAAO,oBAAoB;AAAA,MAC/C,eAAe;AAAA,IACjB,IACA,EAAE,SAAS,MAAM;AAAA,EACzB;AACF,CAAC;AAEH,IAAM,oBAAoB,CACxB,aACkB;AAClB,MAAI,aAAa,OAAW,QAAO;AAEnC,MAAI,aAAa,qBAAsB,QAAO;AAG9C,MAAI,YAAY,sBAAuB,QAAO;AAG9C,MAAI,YAAY,cAAe,QAAO;AAEtC,SAAO;AACT;AAEA,IAAM,+BAA+B,CAAC,UACpC,iBAAiB,SAAS,UAAU,SAAS,MAAM,SAAS;AAS9D,IAAM,kBAAkB,CACtB,SACA,UACA,YACA,QACA,YAKA;AAAA,EACE,QAAQ;AAAA,IACN;AAAA,MACE;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAWA,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,EAAE,SAAS,OAAO,CAAC,EAAE,KAAK,GAAG;AAAA,MACzD,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,WAAW,UAAU,EAAE,IAAI,CAAC,CAAC,EAAE,KAAK,GAAG;AAAA,MACnE,OACG,IAAI,CAAC,MAAM,IAAI,MAAM,WAAW,UAAU,EAAE,YAAY,CAAC,CAAC,CAAC,CAAC,EAC5D,KAAK,GAAG;AAAA,MACX,OAAO,IAAI,MAAM,KAAK,EAAE,KAAK,GAAG;AAAA,MAChC,OAAO,IAAI,CAAC,MAAM,IAAI,MAAM,EAAE,IAAI,CAAC,EAAE,KAAK,GAAG;AAAA,MAC7C;AAAA,MACA;AAAA,MACA,SAAS,yBAAyB;AAAA,MAClC,SAAS,aAAa;AAAA,IACxB;AAAA,EACF;AACF;;;AE3QF,SAAS,UAAAC,SAAQ,OAAAC,YAA6B;AAGvC,IAAM,iCAAiCC,KAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,8BAOpB,UAAU;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,gBAQxB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,cAYzB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qBAehB,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,cAM9B,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,CAWpC;AA+BD,eAAsB,4BACpB,SACA,SAOsE;AACtE,MAAI;AACF,UAAM,EAAE,OAAO,IAAI,MAAMC;AAAA,MACvB,QAAQ;AAAA,QACND;AAAA,UACE;AAAA,UACA,QAAQ;AAAA,UACR,QAAQ,WAAW;AAAA,UACnB,QAAQ;AAAA,UACR,QAAQ;AAAA,UACR,QAAQ,aAAa;AAAA,QACvB;AAAA,MACF;AAAA,IACF;AAEA,WAAO,WAAW,IACd,EAAE,SAAS,MAAM,aAAa,QAAQ,YAAY,IAClD,EAAE,SAAS,OAAO,QAAQ,WAAW,IAAI,YAAY,WAAW;AAAA,EACtE,SAAS,OAAO;AACd,YAAQ,IAAI,KAAK;AACjB,UAAM;AAAA,EACR;AACF;;;AC5HA,SAAS,UAAAE,eAAc;AAShB,IAAM,kBAAkBC;AAAA,EAC7B,8BAA8B,aAAa,IAAI;AAAA;AAAA;AAAA,sEAGqB,SAAS,KAAK,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAO7F;AAEO,IAAM,iBAAiBA;AAAA,EAC5B;AAAA;AAAA;AAAA,+BAG6B,YAAY,IAAI;AAAA;AAAA;AAAA,2EAG4B,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAYpF;AAEO,IAAM,wBAAwBA;AAAA,EACnC;AAAA,+BAC6B,mBAAmB,IAAI;AAAA;AAAA;AAAA,iFAG2B,SAAS,KAAK,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAMxG;AAEO,IAAM,kBAAkBA;AAAA,EAC7B;AAAA;AAAA;AAAA;AAAA;AAKF;AAEO,IAAM,qBAAqBA;AAAA,EAChC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAgCF;AAEO,IAAM,sBAAsBA;AAAA,EACjC;AAAA;AAAA;AAAA,yCAGuC,YAAY,IAAI;AAAA,yCAChB,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,+BAK3B,mBAAmB,IAAI,iCAAiC,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAI9G;AAEO,IAAM,eAAeA;AAAA,EAC1B;AAAA;AAAA;AAAA,mBAGiB,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,mCAIA,YAAY,IAAI,gCAAgC,SAAS,QAAQ,YAAY,IAAI,mBAAmB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM7G,YAAY,IAAI,gCAAgC,SAAS,uCAAuC,YAAY,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAMzJ,YAAY,IAAI,gCAAgC,SAAS,yCAAyC,YAAY,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA,mBAG3K,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,mCAID,aAAa,IAAI,gCAAgC,SAAS,QAAQ,aAAa,IAAI,mBAAmB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM/G,aAAa,IAAI,gCAAgC,SAAS,uCAAuC,aAAa,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,mCAM3J,aAAa,IAAI,gCAAgC,SAAS,yCAAyC,aAAa,IAAI,gCAAgC,SAAS;AAAA;AAAA;AAAA;AAAA;AAKhM;AAEO,IAAM,eAAeA;AAAA,EAC1B;AAAA;AAAA;AAAA,iBAGe,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,iCAIA,YAAY,IAAI,6CAA6C,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM7E,YAAY,IAAI,4EAA4E,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM5G,YAAY,IAAI,8EAA8E,YAAY,IAAI;AAAA;AAAA;AAAA,iBAG9H,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,iCAID,aAAa,IAAI,6CAA6C,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM/E,aAAa,IAAI,4EAA4E,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iCAM9G,aAAa,IAAI,8EAA8E,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAKjJ;AAEO,IAAM,4BAA4BA;AAAA,EACvC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,2DAOyD,YAAY,IAAI;AAAA;AAAA,qBAEtD,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,qCAIA,YAAY,IAAI,uDAAuD,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMvF,YAAY,IAAI,sFAAsF,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMtH,YAAY,IAAI,wFAAwF,YAAY,IAAI;AAAA;AAAA;AAAA,qBAGxI,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,qCAID,aAAa,IAAI,uDAAuD,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMzF,aAAa,IAAI,sFAAsF,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMxH,aAAa,IAAI,wFAAwF,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAM/J;AAEO,IAAM,4BAA4BA;AAAA,EACvC;AAAA;AAAA;AAAA;AAAA;AAAA,+GAK6G,YAAY,IAAI;AAAA;AAAA,qBAE1G,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,qCAIA,YAAY,IAAI,8DAA8D,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM9F,YAAY,IAAI,6FAA6F,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM7H,YAAY,IAAI,+FAA+F,YAAY,IAAI;AAAA;AAAA;AAAA,qBAG/I,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA,qCAID,aAAa,IAAI,8DAA8D,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAMhG,aAAa,IAAI,6FAA6F,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,qCAM/H,aAAa,IAAI,+FAA+F,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAMtK;AAEO,IAAM,sBAAsBA;AAAA,EACjC,6BAA6B,UAAU;AACzC;;;ACzTA,SAAS,cAAc,OAAAC,YAA6B;AAW7C,IAAM,gCAAgC,OAC3C,SACA,YACiD;AACjD,QAAM,SAAS,MAAM;AAAA,IACnB,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,YAAY,IAAI;AAAA;AAAA;AAAA;AAAA,QAI1B,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,eAAe,IAAI;AAAA,EACvD;AACF;;;AChCA,SAAS,SAAS,OAAAC,YAA6B;AA+CxC,IAAM,oBAAoB,OAK/B,SACA,YACyE;AACzE,QAAM,OACJ,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QAAQ,KAChB;AACR,QAAM,YACJ,WAAW,eAAe,UACtB,QAAQ,YACR,QAAQ,KAAK,QAAQ;AAE3B,QAAM,gBACJ,SAAS,CAAC,KAAK,0BAA0B,IAAI,KAAK;AAEpD,QAAM,cACJ,QAAQ,UAAU,0BAA0B,QAAQ,EAAE,KAAK;AAE7D,QAAM,iBACJ,eAAe,UAAU,SAAS,QAAQ,SAAS,KAAK;AAE1D,QAAM,SAA0D,MAAM;AAAA,IACpE,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,YAAY,IAAI;AAAA,uHACqF,aAAa,IAAI,WAAW;AAAA;AAAA,aAEtI,cAAc;AAAA,QACnB,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,IACA,CAAC,QAAQ;AACP,YAAM,WAAW;AAAA,QACf,MAAM,IAAI;AAAA,QACV,MAAM,IAAI;AAAA,QACV,UAAU,IAAI;AAAA,MAChB;AAEA,YAAM,WAAgD;AAAA,QACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,QAC1D,SAAS,IAAI;AAAA,QACb,YAAY,IAAI;AAAA,QAChB,gBAAgB,OAAO,IAAI,eAAe;AAAA,QAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC5C;AAEA,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MAIF;AAAA,IACF;AAAA,EACF;AAEA,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,uBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC,UAAU;AAAA,IACV,eAAe,OAAO,WAAW;AAAA,EACnC,IACA;AAAA,IACE,uBACE,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QACR;AAAA,IACR,UAAU,CAAC;AAAA,IACX,eAAe;AAAA,EACjB;AACN;;;AChIA,SAAS,WAAAC,UAAS,OAAAC,YAA6B;AA2BxC,IAAM,aAAa,OACxB,SACA,UACA,YAGG;AACH,QAAM,gBACJ,WAAW,UAAU,UACjB,0BAA0B,QAAQ,IAAI,KACtC;AAEN,QAAM,KAAK;AAAA,IACT,WAAW,QAAQ,UACf,QAAQ,KACR,WAAW,cAAc,WAAW,QAAQ,WAC1C,QAAQ,OAAO,QAAQ,WACvB;AAAA,EACR;AAEA,QAAM,cAAc,CAAC,MAAM,EAAE,IAAI,0BAA0B,EAAE,KAAK;AAElE,QAAM,SACJ,MAAMC;AAAA,IACJ,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACQ,YAAY,IAAI;AAAA,6EAC2C,aAAa,IAAI,WAAW;AAAA,QAC/F;AAAA,QACA,SAAS,aAAa;AAAA,MACxB;AAAA,IACF;AAAA,IACA,CAAC,QAAQ;AACP,YAAM,WAAW;AAAA,QACf,MAAM,IAAI;AAAA,QACV,MAAM,IAAI;AAAA,QACV,UAAU,IAAI;AAAA,MAChB;AAEA,YAAM,WAAgD;AAAA,QACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,QAC1D,SAAS,IAAI;AAAA,QACb,YAAY;AAAA,QACZ,gBAAgB,OAAO,IAAI,eAAe;AAAA,QAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC5C;AAEA,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MAIF;AAAA,IACF;AAAA,EACF;AAEF,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,sBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC;AAAA,IACA,cAAc;AAAA,EAChB,IACA;AAAA,IACE,sBAAsB;AAAA,IACtB,QAAQ,CAAC;AAAA,IACT,cAAc;AAAA,EAChB;AACN;;;ACjGA,SAAS,gBAAAC,eAAc,OAAAC,YAA6B;AAW7C,IAAM,6BAA6B,OACxC,SACA,YAC8C;AAC9C,QAAM,SAAS,MAAMC;AAAA,IACnB,QAAQ;AAAA,MACNC;AAAA,QACE;AAAA,kBACU,mBAAmB,IAAI;AAAA;AAAA;AAAA,QAGjC,SAAS,aAAa;AAAA,QACtB,QAAQ;AAAA,MACV;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,uBAAuB,IAAI;AAAA,EAC/D;AACF;;;ARNO,IAAM,YAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,yBAAyB,OACpC,SACkB;AAClB,QAAM,KAAK,gBAAgB,CAAC,EAAE,QAAQ,MAAM,QAAQ,aAAa,SAAS,CAAC;AAC7E;;;A1D6FO,IAAM,2BAAsD;AAAA,EACjE,aAAa,CAAC;AAAA,EACd,QAAQ,EAAE,eAAe,iBAAiB;AAC5C;AAEO,IAAM,2CAA2C;AAEjD,IAAM,0BAA0B,CACrC,kBACA,UAAqC,6BACd;AACvB,QAAM,cAAc;AAAA,IAClB;AAAA,IACA,GAAI,QAAQ,oBAAoB,QAAQ,oBAAoB,CAAC;AAAA,EAC/D;AACA,QAAM,OAAO,WAAW,cAAc,YAAY,QAAQC,OAAM,WAAW;AAC3E,MAAI;AAEJ,QAAM,qBACJ,QAAQ,QAAQ,kBAAkB,UAClC,QAAQ,QAAQ,kBAAkB;AAEpC,QAAM,qBAAqB,MAAM;AAC/B,QAAI,CAAC,mBAAoB,QAAO,QAAQ,QAAQ;AAEhD,QAAI,CAAC,eAAe;AAClB,sBAAgB,uBAAuB,IAAI;AAAA,IAC7C;AACA,WAAO;AAAA,EACT;AAEA,QAAM,qBAAqB,QAAQ,eAAe,CAAC,GAChD,OAAO,CAAC,EAAE,KAAK,MAAM,SAAS,QAAQ,EACtC,IAAI,CAAC,EAAE,YAAAC,YAAW,MAAMA,WAAU;AAErC,QAAM,gBACJ,kBAAkB,SAAS,IACvB,CAAC,QAAQ,EAAE,YAAY,MACrB,kBAAkB;AAAA,IAChB,aAAa;AAAA,IACb,YAAY;AAAA,MACV;AAAA,MACA;AAAA,IACF;AAAA;AAAA;AAAA,IAGA;AAAA,EACF,CAAC,IACH;AAEN,SAAO;AAAA,IACL,QAAQ;AAAA,MACN,KAAK,MAAM,UAAU,KAAK,EAAE;AAAA,MAC5B,OAAO,MAAM,QAAQ,IAAI,UAAU,KAAK,EAAE,CAAC;AAAA,MAC3C,SAAS,YAAY;AACnB,eAAO,gBAAgB,uBAAuB,IAAI;AAAA,MACpD;AAAA,IACF;AAAA,IACA,MAAM,gBACJ,YACAC,UAKuC;AACvC,YAAM,EAAE,QAAQ,cAAc,KAAK,IAAIA;AAEvC,YAAM,wBAAwB,MAAM;AAEpC,UAAI,QAAQ,aAAa;AAEzB,YAAM,SAAS,MAAM,KAAK,WAAsB,YAAYA,SAAQ,IAAI;AACxE,YAAM,uBAAuB,OAAO;AAEpC;AAAA,QACE;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,iBAAW,SAAS,OAAO,QAAQ;AACjC,YAAI,CAAC,MAAO;AAEZ,gBAAQ,OAAO,OAAO,KAAK;AAAA,MAC7B;AAEA,aAAO;AAAA,QACL;AAAA,QACA;AAAA,QACA,cAAc,OAAO;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,YAAY,OACV,YACAA,aACoE;AACpE,YAAM,mBAAmB;AACzB,aAAO,WAAsB,KAAK,SAAS,YAAYA,QAAO;AAAA,IAChE;AAAA,IAEA,gBAAgB,OACd,YACA,QACAA,aACoD;AACpD,YAAM,mBAAmB;AAEzB,YAAM,CAAC,WAAW,GAAG,IAAI,IAAI,WAAW,MAAM,GAAG;AAEjD,YAAM,aACJ,aAAa,KAAK,SAAS,IAAI,YAAY;AAE7C,YAAM,eAAe,MAAM;AAAA,QACzB;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,UACE,GAAGA;AAAA,UACH;AAAA,QACF;AAAA,MACF;AAEA,UAAI,CAAC,aAAa;AAChB,cAAM,IAAI;AAAA,UACR,CAAC;AAAA;AAAA,UACDA,UAAS,yBAAyB;AAAA,QACpC;AAEF,aAAO;AAAA,QACL,2BAA2B,aAAa;AAAA,QACxC,yBAAyB,aAAa;AAAA,QACtC,kBACE,aAAa,sBAAsB,OAAO,OAAO,MAAM;AAAA,MAC3D;AAAA,IACF;AAAA,IACA,OAAO,MAAM,KAAK,MAAM;AAAA,IAExB,MAAM,YACJ,UACY;AACZ,aAAO,MAAM,KAAK,eAAe,OAAO,eAAe;AACrD,cAAM,eAA0C;AAAA,UAC9C,GAAG;AAAA,UACH,mBAAmB;AAAA,YACjB;AAAA,UACF;AAAA,QACF;AAEA,cAAM,aAAa;AAAA,UACjB;AAAA,UACA;AAAA,QACF;AAEA,eAAO,SAAS;AAAA,UACd;AAAA,UACA,OAAO,MAAM,QAAQ,QAAQ;AAAA,QAC/B,CAAC;AAAA,MACH,CAAC;AAAA,IACH;AAAA,EACF;AACF;;;AmE9SA,OAAiC;AAa1B,IAAM,mDAAmD;AACzD,IAAM,8DAA8D;AA+CpE,IAAM,yCAAyC,CAEpD;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF,MAAwG;AACtG,MAAI,YAAY;AAEhB,MAAI;AAEJ,QAAM,eAAe,OACnB,YACG;AACH,UAAM,QACJ,QAAQ,cAAc,cAClB,KACA,QAAQ,cAAc,SAClB,MAAM,8BAA8B,QAAQ,GAC3C,yBAAyB,KAC5B,QAAQ,UAAU;AAE1B,UAAM,sBAAgD;AAAA,MACpD;AAAA,MACA;AAAA,IACF;AAEA,QAAI,WAAW;AAEf,OAAG;AACD,YAAM,EAAE,UAAU,uBAAuB,cAAc,IACrD,MAAM,kBAA6B,UAAU,mBAAmB;AAElE,UAAI,SAAS,SAAS,GAAG;AACvB,cAAM,SAAS,MAAM,UAAU,EAAE,SAAS,CAAC;AAE3C,YAAI,UAAU,OAAO,SAAS,QAAQ;AACpC,sBAAY;AACZ;AAAA,QACF;AAAA,MACF;AAEA,0BAAoB,QAAQ;AAE5B,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAE5D,UAAI,CAAC,eAAe;AAClB,mBAAW,KAAK,IAAI,WAAW,GAAG,GAAI;AAAA,MACxC,OAAO;AACL,mBAAW;AAAA,MACb;AAAA,IACF,SAAS;AAAA,EACX;AAEA,SAAO;AAAA,IACL,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,OAAO,CAAC,YAAY;AAClB,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,oBAAY;AAEZ,eAAO,aAAa,OAAO;AAAA,MAC7B,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA,MAAM,YAAY;AAChB,UAAI,CAAC,UAAW;AAChB,kBAAY;AACZ,YAAM;AAAA,IACR;AAAA,EACF;AACF;AAEO,IAAM,qDAAqD,CAChE,YACoD;AACpD,MACE,QAAQ,WAAW,KACnB,QAAQ,KAAK,CAAC,MAAM,MAAM,UAAa,MAAM,WAAW;AAExD,WAAO;AAET,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,KAAK,EAAG,QAAO;AAE9C,SAAO,QACJ,OAAO,CAAC,MAAM,MAAM,UAAa,MAAM,eAAe,MAAM,KAAK,EACjE,KAAK,CAAC,GAAG,MAAO,IAAI,IAAI,IAAI,EAAG,EAAE,CAAC;AACvC;;;ACzJA,SAAS,SAAAC,cAAa;;;ACAtB,OAA6C;AAgCtC,IAAM,mCAAmC;AAAA,EAC9C,QAAQ;AAAA,IACN,MAAM,CAAC,aAEsD;AAAA,MAC3D,MAAM;AAAA,MACN,GAAI,WAAW,CAAC;AAAA,IAClB;AAAA,IACA,MAAM,CAAC,aAGsD;AAAA,MAC3D,MAAM;AAAA,MACN,GAAI,WAAW,CAAC;AAAA,IAClB;AAAA,EACF;AACF;AAgCO,IAAM,mCAAmC,CAG9C,YACqC;AACrC,QAAM,EAAE,YAAY,IAAI;AACxB,MAAI,WAAW;AAGf,SAAO;AAAA,IACL,IAAI,QAAQ;AAAA,IACZ,OAAO,OACL,YACyE;AACzE,iBAAW;AACX,UAAI,QAAQ,cAAc,UAAW,QAAO,QAAQ;AAEpD,YAAM,EAAE,sBAAsB,IAAI,MAAM;AAAA,QACtC;AAAA,QACA;AAAA,UACE,gBAAgB,QAAQ;AAAA,UACxB,WAAW,QAAQ;AAAA,QACrB;AAAA,MACF;AAEA,UAAI,0BAA0B,KAAM,QAAO;AAE3C,aAAO,EAAE,gBAAgB,sBAAsB;AAAA,IACjD;AAAA,IACA,IAAI,WAAW;AACb,aAAO;AAAA,IACT;AAAA,IACA,QAAQ,OACN,EAAE,SAAS,GACX,EAAE,KAAK,MAC2D;AAClE,UAAI,CAAC,SAAU;AAEf,aAAO,KAAK,gBAAgB,OAAO,OAAO;AACxC,YAAI,SAEY;AAEhB,YAAI,wBAAuC;AAE3C,mBAAW,WAAW,UAAU;AAC9B,gBAAM,eAAe;AAKrB,gBAAM,0BAA0B,MAAM,YAAY,YAAY;AAG9D,gBAAM,4BAA4B,GAAG,SAAS;AAAA,YAC5C,gBAAgB,QAAQ;AAAA,YACxB,SAAS,QAAQ;AAAA,YACjB;AAAA,YACA,aAAa,aAAa,SAAS;AAAA,YACnC,WAAW,QAAQ;AAAA,UACrB,CAAC;AAED,kCAAwB,aAAa,SAAS;AAE9C,cACE,2BACA,wBAAwB,SAAS,QACjC;AACA,uBAAW;AACX,qBAAS;AACT;AAAA,UACF;AAEA,cAAI,QAAQ,aAAa,QAAQ,UAAU,YAAY,GAAG;AACxD,uBAAW;AACX,qBAAS,EAAE,MAAM,QAAQ,QAAQ,yBAAyB;AAC1D;AAAA,UACF;AAEA,cACE,2BACA,wBAAwB,SAAS;AAEjC;AAAA,QACJ;AAEA,eAAO;AAAA,MACT,CAAC;AAAA,IACH;AAAA,EACF;AACF;;;ADrIO,IAAM,+BAA+B,CAC1C,YACiC;AACjC,MAAI,YAAY;AAChB,QAAM,EAAE,kBAAkB,QAAQ,IAAI;AACtC,QAAM,gBAAgB,QAAQ,iBAAiB,CAAC;AAEhD,MAAI;AAEJ,MAAI;AAEJ,QAAM,OAAOC,OAAM,EAAE,iBAAiB,CAAC;AAEvC,QAAM,YAAsD,OAC1D,kBACG;AACH,UAAM,sBAAsB,cAAc,OAAO,CAAC,MAAM,EAAE,QAAQ;AAElE,QAAI,oBAAoB,WAAW;AACjC,aAAO;AAAA,QACL,MAAM;AAAA,QACN,QAAQ;AAAA,MACV;AAEF,UAAM,SAAS,MAAM,QAAQ;AAAA,MAC3B,oBAAoB,IAAI,CAAC,MAAM;AAE7B,eAAO,EAAE,OAAO,eAAe,EAAE,KAAK,CAAC;AAAA,MACzC,CAAC;AAAA,IACH;AAEA,WAAO,OAAO;AAAA,MACZ,CAAC,MAAM,EAAE,WAAW,eAAe,EAAE,OAAO,SAAS;AAAA,IACvD,IACI,SACA;AAAA,MACE,MAAM;AAAA,IACR;AAAA,EACN;AAEA,QAAM,gBAAiB,uBACrB,uCAAuC;AAAA,IACrC,UAAU,KAAK;AAAA,IACf;AAAA,IACA,WACE,SAAS,aAAa;AAAA,IACxB,sBACE,SAAS,wBACT;AAAA,EACJ,CAAC;AAEH,QAAM,OAAO,YAAY;AACvB,QAAI,CAAC,UAAW;AAChB,gBAAY;AACZ,QAAI,sBAAsB;AACxB,YAAM,qBAAqB,KAAK;AAChC,6BAAuB;AAAA,IACzB;AACA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,WAAW,CACTC,aACgD;AAChD,YAAM,eAAe,iCAA4CA,QAAO;AAExE,oBAAc,KAAK,YAAY;AAE/B,aAAO;AAAA,IACT;AAAA,IACA,OAAO,MAAM;AACX,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,YAAI,cAAc,WAAW;AAC3B,iBAAO,QAAQ;AAAA,YACb,IAAI;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAEF,oBAAY;AAEZ,cAAM,YAAY;AAAA,UAChB,MAAM,QAAQ,IAAI,cAAc,IAAI,CAAC,MAAM,EAAE,MAAM,KAAK,OAAO,CAAC,CAAC;AAAA,QACnE;AAEA,eAAO,cAAc,MAAM,EAAE,UAAU,CAAC;AAAA,MAC1C,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA;AAAA,IACA,OAAO,YAAY;AACjB,YAAM,KAAK;AACX,YAAM,KAAK,MAAM;AAAA,IACnB;AAAA,EACF;AACF;","names":["dumbo","uuid","TransformStream","TransformStream","error","filter","TransformStream","map","handleChunk","error","skip","retry","filter","pongoClient","pongoClient","uuid","projection","options","uuid","sql","projection","uuid","uuid","single","sql","sql","single","rawSql","rawSql","sql","sql","sql","sql","mapRows","sql","mapRows","sql","singleOrNull","sql","singleOrNull","sql","dumbo","projection","options","dumbo","dumbo","options"]}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@event-driven-io/emmett-postgresql",
|
|
3
|
-
"version": "0.
|
|
3
|
+
"version": "0.26.0",
|
|
4
4
|
"type": "module",
|
|
5
5
|
"description": "Emmett - PostgreSQL - Event Sourcing development made simple",
|
|
6
6
|
"scripts": {
|
|
@@ -70,10 +70,10 @@
|
|
|
70
70
|
],
|
|
71
71
|
"devDependencies": {
|
|
72
72
|
"@testcontainers/postgresql": "^10.12.0",
|
|
73
|
-
"@event-driven-io/emmett-testcontainers": "0.
|
|
73
|
+
"@event-driven-io/emmett-testcontainers": "0.26.0"
|
|
74
74
|
},
|
|
75
75
|
"peerDependencies": {
|
|
76
|
-
"@event-driven-io/emmett": "0.
|
|
76
|
+
"@event-driven-io/emmett": "0.26.0",
|
|
77
77
|
"@event-driven-io/pongo": "0.16.4"
|
|
78
78
|
}
|
|
79
79
|
}
|