@event-driven-io/emmett-sqlite 0.38.4 → 0.38.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/connection/sqliteConnection.ts","../../emmett/src/config/plugins/index.ts","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/validation/dates.ts","../../emmett/src/typing/command.ts","../../emmett/src/typing/event.ts","../../emmett/src/typing/message.ts","../../emmett/src/typing/workflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/eventStore/afterCommit/forwardToMessageBus.ts","../../emmett/src/eventStore/events/index.ts","../../emmett/src/eventStore/eventStore.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/database/inMemoryDatabase.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/utils/iterators.ts","../../emmett/src/taskProcessing/taskProcessor.ts","../../emmett/src/utils/locking/index.ts","../../emmett/src/utils/promises.ts","../../emmett/src/utils/retry.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/database/utils.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjection.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjectionSpec.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/testing/deciderSpecification.ts","../../emmett/src/testing/wrapEventStore.ts","../../emmett/src/eventStore/subscriptions/caughtUpTransformStream.ts","../../emmett/src/eventStore/subscriptions/streamingCoordinator.ts","../../emmett/src/streaming/transformations/notifyAboutNoActiveReaders.ts","../../emmett/src/streaming/writers/writeToStream.ts","../../emmett/src/commandHandling/handleCommand.ts","../../emmett/src/commandHandling/handleCommandWithDecider.ts","../../emmett/src/messageBus/index.ts","../../emmett/src/processors/processors.ts","../../emmett/src/processors/inMemoryProcessors.ts","../../emmett/src/projections/index.ts","../../emmett/src/streaming/binaryArrays.ts","../../emmett/src/streaming/collectors/collect.ts","../../emmett/src/streaming/decoders/binary.ts","../../emmett/src/streaming/decoders/string.ts","../../emmett/src/streaming/decoders/json.ts","../../emmett/src/streaming/decoders/object.ts","../../emmett/src/streaming/decoders/composite.ts","../../emmett/src/streaming/generators/fromArray.ts","../../emmett/src/streaming/generators/index.ts","../../emmett/src/streaming/restream.ts","../../emmett/src/streaming/transformations/filter.ts","../../emmett/src/streaming/transformations/map.ts","../../emmett/src/streaming/transformations/reduce.ts","../../emmett/src/streaming/transformations/retry.ts","../../emmett/src/streaming/transformations/skip.ts","../../emmett/src/streaming/transformations/stopAfter.ts","../../emmett/src/streaming/transformations/stopOn.ts","../../emmett/src/streaming/transformations/take.ts","../../emmett/src/streaming/transformations/waitAtMost.ts","../../emmett/src/streaming/transformations/index.ts","../src/eventStore/schema/appendToStream.ts","../src/eventStore/schema/typing.ts","../src/eventStore/schema/tables.ts","../src/eventStore/schema/utils.ts","../src/eventStore/schema/readLastMessageGlobalPosition.ts","../src/eventStore/schema/readMessagesBatch.ts","../src/eventStore/schema/readProcessorCheckpoint.ts","../src/eventStore/consumers/messageBatchProcessing/index.ts","../src/eventStore/consumers/sqliteProcessor.ts","../src/eventStore/consumers/sqliteEventStoreConsumer.ts","../src/eventStore/projections/index.ts","../src/eventStore/SQLiteEventStore.ts","../src/eventStore/schema/readStream.ts","../src/eventStore/schema/storeProcessorCheckpoint.ts"],"sourcesContent":["import sqlite3 from 'sqlite3';\n\nexport type Parameters = object | string | bigint | number | boolean | null;\n\nexport type SQLiteConnection = {\n close: () => void;\n command: (sql: string, values?: Parameters[]) => Promise<sqlite3.RunResult>;\n query: <T>(sql: string, values?: Parameters[]) => Promise<T[]>;\n querySingle: <T>(sql: string, values?: Parameters[]) => Promise<T | null>;\n withTransaction: <T>(fn: () => Promise<T>) => Promise<T>;\n};\n\nexport interface SQLiteError extends Error {\n errno: number;\n}\n\nexport const isSQLiteError = (error: unknown): error is SQLiteError => {\n if (error instanceof Error && 'code' in error) {\n return true;\n }\n\n return false;\n};\n\nexport type InMemorySharedCacheSQLiteDatabase = 'file::memory:?cache=shared';\nexport const InMemorySharedCacheSQLiteDatabase = 'file::memory:?cache=shared';\nexport type InMemorySQLiteDatabase = ':memory:';\nexport const InMemorySQLiteDatabase = ':memory:';\n\ntype SQLiteConnectionOptions = {\n // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n fileName: InMemorySQLiteDatabase | string | undefined;\n};\n\nexport const sqliteConnection = (\n options: SQLiteConnectionOptions,\n): SQLiteConnection => {\n const fileName = options.fileName ?? InMemorySQLiteDatabase;\n let db: sqlite3.Database;\n\n if (fileName.startsWith('file:')) {\n db = new sqlite3.Database(\n fileName,\n sqlite3.OPEN_URI | sqlite3.OPEN_READWRITE | sqlite3.OPEN_CREATE,\n );\n } else {\n db = new sqlite3.Database(fileName);\n }\n db.run('PRAGMA journal_mode = WAL;');\n let transactionNesting = 0;\n\n return {\n close: (): void => db.close(),\n command: (sql: string, params?: Parameters[]) =>\n new Promise<sqlite3.RunResult>((resolve, reject) => {\n db.run(\n sql,\n params ?? [],\n function (this: sqlite3.RunResult, err: Error | null) {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(this);\n },\n );\n }),\n query: <T>(sql: string, params?: Parameters[]): Promise<T[]> =>\n new Promise((resolve, reject) => {\n db.all(sql, params ?? [], (err: Error | null, result: T[]) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(result);\n });\n }),\n querySingle: <T>(sql: string, params?: Parameters[]): Promise<T | null> =>\n new Promise((resolve, reject) => {\n db.get(sql, params ?? [], (err: Error | null, result: T | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(result);\n });\n }),\n withTransaction: async <T>(fn: () => Promise<T>) => {\n try {\n if (transactionNesting++ == 0) {\n await beginTransaction(db);\n }\n const result = await fn();\n\n if (transactionNesting === 1) await commitTransaction(db);\n transactionNesting--;\n\n return result;\n } catch (err) {\n console.log(err);\n\n if (--transactionNesting === 0) await rollbackTransaction(db);\n\n throw err;\n }\n },\n };\n};\n\nconst beginTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('BEGIN IMMEDIATE TRANSACTION', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n\nconst commitTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('COMMIT', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n\nconst rollbackTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('ROLLBACK', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n","export type EmmettPluginConfig =\n | {\n name: string;\n register: EmmettPluginRegistration[];\n }\n | string;\n\nexport type EmmettPluginType = 'cli';\n\nexport type EmmettCliPluginRegistration = { pluginType: 'cli'; path?: string };\n\nexport type EmmettPluginRegistration = EmmettCliPluginRegistration;\n\nexport type EmmettCliCommand = {\n addCommand<CliCommand>(command: CliCommand): CliCommand;\n};\n\nexport type EmmettCliPlugin = {\n pluginType: 'cli';\n name: string;\n registerCommands: (program: EmmettCliCommand) => Promise<void> | void;\n};\n\nexport type EmmettPlugin = EmmettCliPlugin;\n\nexport const isPluginConfig = (\n plugin: Partial<EmmettPluginConfig> | string | undefined,\n): plugin is EmmettPluginConfig =>\n plugin !== undefined &&\n (typeof plugin === 'string' ||\n ('name' in plugin &&\n plugin.name !== undefined &&\n typeof plugin.name === 'string'));\n","import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const isBigint = (val: any): val is bigint =>\n typeof val === 'bigint' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : 500;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n\n public static mapFrom(\n error: Error | { message?: string; errorCode?: number },\n ): EmmettError {\n if (error instanceof EmmettError) {\n return error;\n }\n\n return new EmmettError({\n errorCode:\n 'errorCode' in error &&\n error.errorCode !== undefined &&\n error.errorCode !== null\n ? error.errorCode\n : 500,\n message: error.message ?? 'An unknown error occurred',\n });\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: 412,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\nexport class ConcurrencyInMemoryDatabaseError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 412,\n message: message ?? `Expected document state does not match current one!`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyInMemoryDatabaseError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 400,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 403,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: 404,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { ValidationError } from '../errors';\n\nexport const formatDateToUtcYYYYMMDD = (date: Date) => {\n // Use the 'en-CA' locale which formats as 'yyyy-mm-dd'\n const formatter = new Intl.DateTimeFormat('en-CA', {\n timeZone: 'UTC',\n year: 'numeric',\n month: '2-digit',\n day: '2-digit',\n });\n\n // Format the date\n return formatter.format(date);\n};\n\n// Function to validate 'yyyy-mm-dd' format\nexport const isValidYYYYMMDD = (dateString: string) => {\n const regex = /^\\d{4}-\\d{2}-\\d{2}$/;\n return regex.test(dateString);\n};\n\nexport const parseDateFromUtcYYYYMMDD = (dateString: string) => {\n const date = new Date(dateString + 'T00:00:00Z');\n\n if (!isValidYYYYMMDD(dateString)) {\n throw new ValidationError('Invalid date format, must be yyyy-mm-dd');\n }\n\n if (isNaN(date.getTime())) {\n throw new ValidationError('Invalid date format');\n }\n\n return date;\n};\n","import type { DefaultRecord } from './';\n\nexport type Command<\n CommandType extends string = string,\n CommandData extends DefaultRecord = DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: Readonly<CommandData>;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n> & { readonly kind?: 'Command' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyCommand = Command<any, any, any>;\n\nexport type CommandTypeOf<T extends Command> = T['type'];\nexport type CommandDataOf<T extends Command> = T['data'];\nexport type CommandMetaDataOf<T extends Command> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CreateCommandType<\n CommandType extends string,\n CommandData extends DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: CommandData;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n> & { readonly kind?: 'Command' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const command = <CommandType extends Command<string, any, any>>(\n ...args: CommandMetaDataOf<CommandType> extends undefined\n ? [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata?: DefaultCommandMetadata | undefined,\n ]\n : [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata: CommandMetaDataOf<CommandType>,\n ]\n): CommandType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind: 'Command' } as CommandType)\n : ({ type, data, kind: 'Command' } as CommandType);\n};\n\nexport type DefaultCommandMetadata = { now: Date };\n","import type { DefaultRecord } from './';\nimport type {\n AnyRecordedMessageMetadata,\n CombinedMessageMetadata,\n CommonRecordedMessageMetadata,\n GlobalPositionTypeOfRecordedMessageMetadata,\n RecordedMessage,\n RecordedMessageMetadata,\n RecordedMessageMetadataWithGlobalPosition,\n RecordedMessageMetadataWithoutGlobalPosition,\n StreamPositionTypeOfRecordedMessageMetadata,\n} from './message';\n\nexport type BigIntStreamPosition = bigint;\nexport type BigIntGlobalPosition = bigint;\n\nexport type Event<\n EventType extends string = string,\n EventData extends DefaultRecord = DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n> & { readonly kind?: 'Event' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyEvent = Event<any, any, any>;\n\nexport type EventTypeOf<T extends Event> = T['type'];\nexport type EventDataOf<T extends Event> = T['data'];\nexport type EventMetaDataOf<T extends Event> = T extends { metadata: infer M }\n ? M\n : undefined;\n\nexport type CreateEventType<\n EventType extends string,\n EventData extends DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n> & { readonly kind?: 'Event' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const event = <EventType extends Event<string, any, any>>(\n ...args: EventMetaDataOf<EventType> extends undefined\n ? [type: EventTypeOf<EventType>, data: EventDataOf<EventType>]\n : [\n type: EventTypeOf<EventType>,\n data: EventDataOf<EventType>,\n metadata: EventMetaDataOf<EventType>,\n ]\n): EventType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind: 'Event' } as EventType)\n : ({ type, data, kind: 'Event' } as EventType);\n};\n\nexport type CombinedReadEventMetadata<\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = CombinedMessageMetadata<EventType, EventMetaDataType>;\n\nexport type ReadEvent<\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = RecordedMessage<EventType, EventMetaDataType>;\n\nexport type AnyReadEvent<\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = ReadEvent<AnyEvent, EventMetaDataType>;\n\nexport type CommonReadEventMetadata<StreamPosition = BigIntStreamPosition> =\n CommonRecordedMessageMetadata<StreamPosition>;\n\nexport type ReadEventMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadata<GlobalPosition, StreamPosition>;\n\nexport type AnyReadEventMetadata = AnyRecordedMessageMetadata;\n\nexport type ReadEventMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = RecordedMessageMetadataWithGlobalPosition<GlobalPosition>;\n\nexport type ReadEventMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadataWithoutGlobalPosition<StreamPosition>;\n\nexport type GlobalPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n GlobalPositionTypeOfRecordedMessageMetadata<ReadEventMetadataType>;\n\nexport type StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n StreamPositionTypeOfRecordedMessageMetadata<ReadEventMetadataType>;\n","import type {\n AnyCommand,\n AnyEvent,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n Command,\n DefaultRecord,\n Event,\n} from '.';\n\nexport type Message<\n Type extends string = string,\n Data extends DefaultRecord = DefaultRecord,\n MetaData extends DefaultRecord | undefined = undefined,\n> = Command<Type, Data, MetaData> | Event<Type, Data, MetaData>;\n\nexport type AnyMessage = AnyEvent | AnyCommand;\n\nexport type MessageKindOf<T extends Message> = T['kind'];\nexport type MessageTypeOf<T extends Message> = T['type'];\nexport type MessageDataOf<T extends Message> = T['data'];\nexport type MessageMetaDataOf<T extends Message> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CanHandle<T extends Message> = MessageTypeOf<T>[];\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const message = <MessageType extends Message<string, any, any>>(\n ...args: MessageMetaDataOf<MessageType> extends undefined\n ? [\n kind: MessageKindOf<MessageType>,\n type: MessageTypeOf<MessageType>,\n data: MessageDataOf<MessageType>,\n ]\n : [\n kind: MessageKindOf<MessageType>,\n type: MessageTypeOf<MessageType>,\n data: MessageDataOf<MessageType>,\n metadata: MessageMetaDataOf<MessageType>,\n ]\n): MessageType => {\n const [kind, type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind } as MessageType)\n : ({ type, data, kind } as MessageType);\n};\n\nexport type CombinedMessageMetadata<\n MessageType extends Message = Message,\n MessageMetaDataType extends DefaultRecord = DefaultRecord,\n> =\n MessageMetaDataOf<MessageType> extends undefined\n ? MessageMetaDataType\n : MessageMetaDataOf<MessageType> & MessageMetaDataType;\n\nexport type CombineMetadata<\n MessageType extends Message = Message,\n MessageMetaDataType extends DefaultRecord = DefaultRecord,\n> = MessageType & {\n metadata: CombinedMessageMetadata<MessageType, MessageMetaDataType>;\n};\n\nexport type RecordedMessage<\n MessageType extends Message = Message,\n MessageMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = CombineMetadata<MessageType, MessageMetaDataType> & {\n kind: NonNullable<MessageKindOf<Message>>;\n};\n\nexport type CommonRecordedMessageMetadata<\n StreamPosition = BigIntStreamPosition,\n> = Readonly<{\n messageId: string;\n streamPosition: StreamPosition;\n streamName: string;\n}>;\n\nexport type WithGlobalPosition<GlobalPosition> = Readonly<{\n globalPosition: GlobalPosition;\n}>;\n\nexport type RecordedMessageMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = CommonRecordedMessageMetadata<StreamPosition> &\n // eslint-disable-next-line @typescript-eslint/no-empty-object-type\n (GlobalPosition extends undefined ? {} : WithGlobalPosition<GlobalPosition>);\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyRecordedMessageMetadata = RecordedMessageMetadata<any, any>;\n\nexport type RecordedMessageMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = RecordedMessageMetadata<GlobalPosition>;\n\nexport type RecordedMessageMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadata<undefined, StreamPosition>;\n\nexport type GlobalPositionTypeOfRecordedMessageMetadata<\n RecordedMessageMetadataType,\n> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RecordedMessageMetadataType extends RecordedMessageMetadata<infer GP, any>\n ? GP\n : never;\n\nexport type StreamPositionTypeOfRecordedMessageMetadata<\n RecordedMessageMetadataType,\n> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RecordedMessageMetadataType extends RecordedMessageMetadata<any, infer SV>\n ? SV\n : never;\n","import type { AnyCommand } from './command';\nimport type { AnyEvent } from './event';\n\n/// Inspired by https://blog.bittacklr.be/the-workflow-pattern.html\n\nexport type Workflow<\n Input extends AnyEvent | AnyCommand,\n State,\n Output extends AnyEvent | AnyCommand,\n> = {\n decide: (command: Input, state: State) => WorkflowOutput<Output>[];\n evolve: (currentState: State, event: WorkflowEvent<Output>) => State;\n initialState: () => State;\n};\n\nexport type WorkflowEvent<Output extends AnyEvent | AnyCommand> = Extract<\n Output,\n { kind?: 'Event' }\n>;\n\nexport type WorkflowCommand<Output extends AnyEvent | AnyCommand> = Extract<\n Output,\n { kind?: 'Command' }\n>;\n\nexport type WorkflowOutput<TOutput extends AnyEvent | AnyCommand> =\n | { action: 'Reply'; message: TOutput }\n | { action: 'Send'; message: WorkflowCommand<TOutput> }\n | { action: 'Publish'; message: WorkflowEvent<TOutput> }\n | {\n action: 'Schedule';\n message: TOutput;\n when: { afterInMs: number } | { at: Date };\n }\n | { action: 'Complete' }\n | { action: 'Accept' }\n | { action: 'Ignore'; reason: string }\n | { action: 'Error'; reason: string };\n\nexport const reply = <TOutput extends AnyEvent | AnyCommand>(\n message: TOutput,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Reply',\n message,\n };\n};\n\nexport const send = <TOutput extends AnyEvent | AnyCommand>(\n message: WorkflowCommand<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Send',\n message,\n };\n};\n\nexport const publish = <TOutput extends AnyEvent | AnyCommand>(\n message: WorkflowEvent<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Publish',\n message,\n };\n};\n\nexport const schedule = <TOutput extends AnyEvent | AnyCommand>(\n message: TOutput,\n when: { afterInMs: number } | { at: Date },\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Schedule',\n message,\n when,\n };\n};\n\nexport const complete = <\n TOutput extends AnyEvent | AnyCommand,\n>(): WorkflowOutput<TOutput> => {\n return {\n action: 'Complete',\n };\n};\n\nexport const ignore = <TOutput extends AnyEvent | AnyCommand>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Ignore',\n reason,\n };\n};\n\nexport const error = <TOutput extends AnyEvent | AnyCommand>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Error',\n reason,\n };\n};\n\nexport const accept = <\n TOutput extends AnyEvent | AnyCommand,\n>(): WorkflowOutput<TOutput> => {\n return { action: 'Accept' };\n};\n","import {\n type BatchRecordedMessageHandlerWithContext,\n type BatchRecordedMessageHandlerWithoutContext,\n type DefaultRecord,\n type Event,\n type ReadEvent,\n} from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\nexport type BeforeEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","import type { EventsPublisher } from '../../messageBus';\nimport type { DefaultRecord, Event, ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\nimport type { AfterEventStoreCommitHandler } from './afterEventStoreCommitHandler';\n\nexport const forwardToMessageBus = <\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n eventPublisher: EventsPublisher,\n): AfterEventStoreCommitHandler<Store, HandlerContext> =>\n (async (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n ): Promise<void> => {\n for (const message of messages) {\n await eventPublisher.publish(message);\n }\n }) as AfterEventStoreCommitHandler<Store, HandlerContext>;\n","import {\n event,\n type Event,\n type EventDataOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\n\nexport const GlobalStreamCaughtUpType = '__emt:GlobalStreamCaughtUp';\n\nexport type GlobalStreamCaughtUp = Event<\n '__emt:GlobalStreamCaughtUp',\n { globalPosition: bigint },\n { globalPosition: bigint }\n>;\n\nexport const isGlobalStreamCaughtUp = (\n event: Event,\n): event is GlobalStreamCaughtUp => event.type === GlobalStreamCaughtUpType;\n\nexport const caughtUpEventFrom =\n (position: bigint) =>\n (\n event: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n ): event is ReadEvent<\n GlobalStreamCaughtUp,\n ReadEventMetadataWithGlobalPosition\n > =>\n event.type === GlobalStreamCaughtUpType &&\n event.metadata?.globalPosition >= position;\n\nexport const globalStreamCaughtUp = (\n data: EventDataOf<GlobalStreamCaughtUp>,\n): GlobalStreamCaughtUp =>\n event<GlobalStreamCaughtUp>(GlobalStreamCaughtUpType, data, {\n globalPosition: data.globalPosition,\n });\n\nexport const isSubscriptionEvent = (\n event: Event,\n): event is GlobalSubscriptionEvent => isGlobalStreamCaughtUp(event);\n\nexport const isNotInternalEvent = (event: Event): boolean =>\n !isGlobalStreamCaughtUp(event);\n\nexport type GlobalSubscriptionEvent = GlobalStreamCaughtUp;\n","//import type { ReadableStream } from 'web-streams-polyfill';\nimport type {\n AnyReadEventMetadata,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n CommonReadEventMetadata,\n DefaultRecord,\n Event,\n GlobalPositionTypeOfReadEventMetadata,\n ReadEvent,\n ReadEventMetadata,\n StreamPositionTypeOfReadEventMetadata,\n WithGlobalPosition,\n} from '../typing';\nimport type { AfterEventStoreCommitHandler } from './afterCommit';\n//import type { GlobalSubscriptionEvent } from './events';\nimport type { ExpectedStreamVersion } from './expectedVersion';\n\n// #region event-store\nexport interface EventStore<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> {\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType, ReadEventMetadataType>,\n ): Promise<\n AggregateStreamResult<\n State,\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<ReadStreamResult<EventType, ReadEventMetadataType>>;\n\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<\n AppendToStreamResult<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n // streamEvents(): ReadableStream<\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // >;\n}\n\nexport type EventStoreReadEventMetadata<Store extends EventStore> =\n Store extends EventStore<infer T>\n ? T extends CommonReadEventMetadata<infer SP>\n ? T extends WithGlobalPosition<infer GP>\n ? ReadEventMetadata<GP, SP> & T\n : ReadEventMetadata<undefined, SP> & T\n : never\n : never;\n\nexport type GlobalPositionTypeOfEventStore<Store extends EventStore> =\n GlobalPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type StreamPositionTypeOfEventStore<Store extends EventStore> =\n StreamPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type EventStoreSession<EventStoreType extends EventStore> = {\n eventStore: EventStoreType;\n close: () => Promise<void>;\n};\n\nexport interface EventStoreSessionFactory<EventStoreType extends EventStore> {\n withSession<T = unknown>(\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n ): Promise<T>;\n}\n// #endregion event-store\n\nexport const canCreateEventStoreSession = <Store extends EventStore>(\n eventStore: Store | EventStoreSessionFactory<Store>,\n): eventStore is EventStoreSessionFactory<Store> => 'withSession' in eventStore;\n\nexport const nulloSessionFactory = <EventStoreType extends EventStore>(\n eventStore: EventStoreType,\n): EventStoreSessionFactory<EventStoreType> => ({\n withSession: (callback) => {\n const nulloSession: EventStoreSession<EventStoreType> = {\n eventStore,\n close: () => Promise.resolve(),\n };\n\n return callback(nulloSession);\n },\n});\n\n////////////////////////////////////////////////////////////////////\n/// ReadStream types\n////////////////////////////////////////////////////////////////////\n\nexport type ReadStreamOptions<StreamVersion = BigIntStreamPosition> = (\n | {\n from: StreamVersion;\n }\n | { to: StreamVersion }\n | { from: StreamVersion; maxCount?: bigint }\n | {\n expectedStreamVersion: ExpectedStreamVersion<StreamVersion>;\n }\n) & {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type ReadStreamResult<\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n currentStreamVersion: StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>;\n events: ReadEvent<EventType, ReadEventMetadataType>[];\n streamExists: boolean;\n};\n\n////////////////////////////////////////////////////////////////////\n/// AggregateStream types\n////////////////////////////////////////////////////////////////////\n\ntype Evolve<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n | ((currentState: State, event: EventType) => State)\n | ((\n currentState: State,\n event: ReadEvent<EventType, ReadEventMetadataType>,\n ) => State)\n | ((currentState: State, event: ReadEvent<EventType>) => State);\n\nexport type AggregateStreamOptions<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n evolve: Evolve<State, EventType, ReadEventMetadataType>;\n initialState: () => State;\n read?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >;\n};\n\nexport type AggregateStreamResult<\n State,\n StreamPosition = BigIntStreamPosition,\n> = {\n currentStreamVersion: StreamPosition;\n state: State;\n streamExists: boolean;\n};\n\nexport type AggregateStreamResultWithGlobalPosition<\n State,\n StreamPosition = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> =\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: true;\n lastEventGlobalPosition: GlobalPosition;\n })\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: false;\n });\n\nexport type AggregateStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['aggregateStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// AppendToStream types\n////////////////////////////////////////////////////////////////////\n\nexport type AppendToStreamOptions<StreamVersion = BigIntStreamPosition> = {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type AppendToStreamResult<StreamVersion = BigIntStreamPosition> = {\n nextExpectedStreamVersion: StreamVersion;\n createdNewStream: boolean;\n};\n\nexport type AppendToStreamResultWithGlobalPosition<\n StreamVersion = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> = AppendToStreamResult<StreamVersion> & {\n lastEventGlobalPosition: GlobalPosition;\n};\n\nexport type AppendStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['appendToStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// DefaultEventStoreOptions\n////////////////////////////////////////////////////////////////////\n\nexport type DefaultEventStoreOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n /**\n * Pluggable set of hooks informing about the event store internal behaviour.\n */\n hooks?: {\n /**\n * This hook will be called **AFTER** events were stored in the event store.\n * It's designed to handle scenarios where delivery and ordering guarantees do not matter much.\n *\n * **WARNINGS:**\n *\n * 1. It will be called **EXACTLY ONCE** if append succeded.\n * 2. If the hook fails, its append **will still silently succeed**, and no error will be thrown.\n * 3. Wen process crashes after events were committed, but before the hook was called, delivery won't be retried.\n * That can lead to state inconsistencies.\n * 4. In the case of high concurrent traffic, **race conditions may cause ordering issues**.\n * For instance, where the second hook takes longer to process than the first one, ordering won't be guaranteed.\n *\n * @type {AfterEventStoreCommitHandler<Store, HandlerContext>}\n */\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n };\n};\n","import { ConcurrencyError } from '../errors';\nimport type { BigIntStreamPosition, Flavour } from '../typing';\n\nexport type ExpectedStreamVersion<VersionType = BigIntStreamPosition> =\n | ExpectedStreamVersionWithValue<VersionType>\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue<VersionType = BigIntStreamPosition> =\n Flavour<VersionType, 'StreamVersion'>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = <StreamVersion = BigIntStreamPosition>(\n current: StreamVersion | undefined,\n expected: ExpectedStreamVersion<StreamVersion>,\n defaultVersion: StreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = <\n StreamVersion = BigIntStreamPosition,\n>(\n current: StreamVersion,\n expected: ExpectedStreamVersion<StreamVersion> | undefined,\n defaultVersion: StreamVersion,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError<\n VersionType = BigIntStreamPosition,\n> extends ConcurrencyError {\n constructor(\n current: VersionType,\n expected: ExpectedStreamVersion<VersionType>,\n ) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError;\n","import { v4 as uuid } from 'uuid';\nimport {\n getInMemoryDatabase,\n type InMemoryDatabase,\n} from '../database/inMemoryDatabase';\nimport type { ProjectionRegistration } from '../projections';\nimport type {\n BigIntStreamPosition,\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport {\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultEventStoreOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { handleInMemoryProjections } from './projections/inMemory';\nimport { StreamingCoordinator } from './subscriptions';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition> & {\n database: InMemoryDatabase;\n };\n\nexport type InMemoryReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type InMemoryProjectionHandlerContext = {\n eventStore?: InMemoryEventStore;\n database?: InMemoryDatabase;\n};\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore> & {\n projections?: ProjectionRegistration<\n 'inline',\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >[];\n database?: InMemoryDatabase;\n };\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n const streamingCoordinator = StreamingCoordinator();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n // Get the database instance to be used for projections\n const database = eventStoreOptions?.database || getInMemoryDatabase();\n\n // Extract inline projections from options\n const inlineProjections = (eventStoreOptions?.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n // Create the event store object\n const eventStore: InMemoryEventStore = {\n database,\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType>(streamName, read);\n\n const events = result?.events ?? [];\n\n return {\n currentStreamVersion: BigInt(events.length),\n state: events.reduce(evolve, initialState()),\n streamExists: result.streamExists,\n };\n },\n\n readStream: <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options && 'from' in options ? options.from : 0);\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : (events?.length ?? 1),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? events\n .map(\n (e) =>\n e as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n )\n .slice(from, to)\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition: BigInt(getAllEventsCount() + index + 1),\n };\n return {\n ...event,\n kind: event.kind ?? 'Event',\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [...currentEvents, ...newEvents]);\n await streamingCoordinator.notify(newEvents);\n\n // Process projections if there are any registered\n if (inlineProjections.length > 0) {\n await handleInMemoryProjections({\n projections: inlineProjections,\n events: newEvents,\n database: eventStore.database,\n eventStore,\n });\n }\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit<InMemoryEventStore>(\n newEvents,\n eventStoreOptions?.hooks,\n );\n\n return result;\n },\n\n //streamEvents: streamingCoordinator.stream,\n };\n\n return eventStore;\n};\n","import { v7 as uuid } from 'uuid';\nimport { deepEquals } from '../utils';\nimport {\n type DatabaseHandleOptionErrors,\n type DatabaseHandleOptions,\n type DatabaseHandleResult,\n type DeleteResult,\n type Document,\n type DocumentHandler,\n type InsertOneResult,\n type OptionalUnlessRequiredIdAndVersion,\n type ReplaceOneOptions,\n type UpdateResult,\n type WithIdAndVersion,\n type WithoutId,\n} from './types';\nimport { expectedVersionValue, operationResult } from './utils';\n\nexport interface InMemoryDocumentsCollection<T extends Document> {\n handle: (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ) => Promise<DatabaseHandleResult<T>>;\n findOne: (predicate?: Predicate<T>) => Promise<T | null>;\n find: (predicate?: Predicate<T>) => Promise<T[]>;\n insertOne: (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ) => Promise<InsertOneResult>;\n deleteOne: (predicate?: Predicate<T>) => Promise<DeleteResult>;\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ) => Promise<UpdateResult>;\n}\n\nexport interface InMemoryDatabase {\n collection: <T extends Document>(\n name: string,\n ) => InMemoryDocumentsCollection<T>;\n}\n\ntype Predicate<T> = (item: T) => boolean;\ntype CollectionName = string;\n\nexport const getInMemoryDatabase = (): InMemoryDatabase => {\n const storage = new Map<CollectionName, WithIdAndVersion<Document>[]>();\n\n return {\n collection: <T extends Document, CollectionName extends string>(\n collectionName: CollectionName,\n collectionOptions: {\n errors?: DatabaseHandleOptionErrors;\n } = {},\n ): InMemoryDocumentsCollection<T> => {\n const ensureCollectionCreated = () => {\n if (!storage.has(collectionName)) storage.set(collectionName, []);\n };\n\n const errors = collectionOptions.errors;\n\n const collection = {\n collectionName,\n insertOne: async (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ): Promise<InsertOneResult> => {\n ensureCollectionCreated();\n\n const _id = (document._id as string | undefined | null) ?? uuid();\n const _version = document._version ?? 1n;\n\n const existing = await collection.findOne((c) => c._id === _id);\n\n if (existing) {\n return operationResult<InsertOneResult>(\n {\n successful: false,\n insertedId: null,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n }\n\n const documentsInCollection = storage.get(collectionName)!;\n const newDocument = { ...document, _id, _version };\n const newCollection = [...documentsInCollection, newDocument];\n storage.set(collectionName, newCollection);\n\n return operationResult<InsertOneResult>(\n {\n successful: true,\n insertedId: _id,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n },\n findOne: (predicate?: Predicate<T>): Promise<T | null> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n const firstOne = filteredDocuments?.[0] ?? null;\n\n return Promise.resolve(firstOne as T | null);\n },\n find: (predicate?: Predicate<T>): Promise<T[]> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n return Promise.resolve(filteredDocuments as T[]);\n },\n deleteOne: (predicate?: Predicate<T>): Promise<DeleteResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n if (predicate) {\n const foundIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (foundIndex === -1) {\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: false,\n matchedCount: 0,\n deletedCount: 0,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n } else {\n const newCollection = documentsInCollection.toSpliced(\n foundIndex,\n 1,\n );\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n }\n }\n\n const newCollection = documentsInCollection.slice(1);\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n },\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ): Promise<UpdateResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n const firstIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (firstIndex === undefined || firstIndex === -1) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 0,\n modifiedCount: 0,\n nextExpectedVersion: 0n,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const existing = documentsInCollection[firstIndex]!;\n\n if (\n typeof options?.expectedVersion === 'bigint' &&\n existing._version !== options.expectedVersion\n ) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 1,\n modifiedCount: 0,\n nextExpectedVersion: existing._version,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const newVersion = existing._version + 1n;\n\n const newCollection = documentsInCollection.with(firstIndex, {\n _id: existing._id,\n ...document,\n _version: newVersion,\n });\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: true,\n modifiedCount: 1,\n matchedCount: firstIndex,\n nextExpectedVersion: newVersion,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n },\n handle: async (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ): Promise<DatabaseHandleResult<T>> => {\n const { expectedVersion: version, ...operationOptions } =\n options ?? {};\n ensureCollectionCreated();\n const existing = await collection.findOne(({ _id }) => _id === id);\n\n const expectedVersion = expectedVersionValue(version);\n\n if (\n (existing == null && version === 'DOCUMENT_EXISTS') ||\n (existing == null && expectedVersion != null) ||\n (existing != null && version === 'DOCUMENT_DOES_NOT_EXIST') ||\n (existing != null &&\n expectedVersion !== null &&\n existing._version !== expectedVersion)\n ) {\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: false,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n }\n\n const result = handle(existing !== null ? { ...existing } : null);\n\n if (deepEquals(existing, result))\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n\n if (!existing && result) {\n const newDoc = { ...result, _id: id };\n const insertResult = await collection.insertOne({\n ...newDoc,\n _id: id,\n } as OptionalUnlessRequiredIdAndVersion<T>);\n return {\n ...insertResult,\n document: {\n ...newDoc,\n _version: insertResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n if (existing && !result) {\n const deleteResult = await collection.deleteOne(\n ({ _id }) => id === _id,\n );\n return { ...deleteResult, document: null };\n }\n\n if (existing && result) {\n const replaceResult = await collection.replaceOne(\n ({ _id }) => id === _id,\n result,\n {\n ...operationOptions,\n expectedVersion: expectedVersion ?? 'DOCUMENT_EXISTS',\n },\n );\n return {\n ...replaceResult,\n document: {\n ...result,\n _version: replaceResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n },\n };\n\n return collection;\n },\n };\n};\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","export const deepEquals = <T>(left: T, right: T): boolean => {\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n if (Array.isArray(left)) {\n return (\n Array.isArray(right) &&\n left.length === right.length &&\n left.every((val, index) => deepEquals(val, right[index]))\n );\n }\n\n if (\n typeof left !== 'object' ||\n typeof right !== 'object' ||\n left === null ||\n right === null\n ) {\n return left === right;\n }\n\n if (Array.isArray(right)) return false;\n\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (\n keys1.length !== keys2.length ||\n !keys1.every((key) => keys2.includes(key))\n )\n return false;\n\n for (const key in left) {\n if (left[key] instanceof Function && right[key] instanceof Function)\n continue;\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export const sum = (\n iterator: Iterator<number, number, number> | Iterator<number>,\n) => {\n let value,\n done: boolean | undefined,\n sum = 0;\n do {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n ({ value, done } = iterator.next());\n sum += value || 0;\n } while (!done);\n return sum;\n};\n","import { EmmettError } from '../errors';\n\nexport type TaskQueue = TaskQueueItem[];\n\nexport type TaskQueueItem = {\n task: () => Promise<void>;\n options?: EnqueueTaskOptions;\n};\n\nexport type TaskProcessorOptions = {\n maxActiveTasks: number;\n maxQueueSize: number;\n maxTaskIdleTime?: number;\n};\n\nexport type Task<T> = (context: TaskContext) => Promise<T>;\n\nexport type TaskContext = {\n ack: () => void;\n};\n\nexport type EnqueueTaskOptions = { taskGroupId?: string };\n\nexport class TaskProcessor {\n private queue: TaskQueue = [];\n private isProcessing = false;\n private activeTasks = 0;\n private activeGroups: Set<string> = new Set();\n\n constructor(private options: TaskProcessorOptions) {}\n\n enqueue<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n if (this.queue.length >= this.options.maxQueueSize) {\n return Promise.reject(\n new EmmettError(\n 'Too many pending connections. Please try again later.',\n ),\n );\n }\n\n return this.schedule(task, options);\n }\n\n waitForEndOfProcessing(): Promise<void> {\n return this.schedule(({ ack }) => Promise.resolve(ack()));\n }\n\n private schedule<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n return promiseWithDeadline(\n (resolve, reject) => {\n const taskWithContext = () => {\n return new Promise<void>((resolveTask, failTask) => {\n const taskPromise = task({\n ack: resolveTask,\n });\n\n taskPromise.then(resolve).catch((err) => {\n // eslint-disable-next-line @typescript-eslint/prefer-promise-reject-errors\n failTask(err);\n reject(err);\n });\n });\n };\n\n this.queue.push({ task: taskWithContext, options });\n if (!this.isProcessing) {\n this.ensureProcessing();\n }\n },\n { deadline: this.options.maxTaskIdleTime },\n );\n }\n\n private ensureProcessing(): void {\n if (this.isProcessing) return;\n this.isProcessing = true;\n this.processQueue();\n }\n\n private processQueue(): void {\n try {\n while (\n this.activeTasks < this.options.maxActiveTasks &&\n this.queue.length > 0\n ) {\n const item = this.takeFirstAvailableItem();\n\n if (item === null) return;\n\n const groupId = item.options?.taskGroupId;\n\n if (groupId) {\n // Mark the group as active\n this.activeGroups.add(groupId);\n }\n\n this.activeTasks++;\n void this.executeItem(item);\n }\n } catch (error) {\n console.error(error);\n throw error;\n } finally {\n this.isProcessing = false;\n if (\n this.hasItemsToProcess() &&\n this.activeTasks < this.options.maxActiveTasks\n ) {\n this.ensureProcessing();\n }\n }\n }\n\n private async executeItem({ task, options }: TaskQueueItem): Promise<void> {\n try {\n await task();\n } finally {\n this.activeTasks--;\n\n // Mark the group as inactive after task completion\n if (options && options.taskGroupId) {\n this.activeGroups.delete(options.taskGroupId);\n }\n\n this.ensureProcessing();\n }\n }\n\n private takeFirstAvailableItem = (): TaskQueueItem | null => {\n const taskIndex = this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n );\n\n if (taskIndex === -1) {\n // All remaining tasks are blocked by active groups\n return null;\n }\n\n // Remove the task from the queue\n const [item] = this.queue.splice(taskIndex, 1);\n\n return item ?? null;\n };\n\n private hasItemsToProcess = (): boolean =>\n this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n ) !== -1;\n}\n\nconst DEFAULT_PROMISE_DEADLINE = 2147483647;\n\nconst promiseWithDeadline = <T>(\n executor: (\n resolve: (value: T | PromiseLike<T>) => void,\n reject: (reason?: unknown) => void,\n ) => void,\n options: { deadline?: number },\n) => {\n return new Promise<T>((resolve, reject) => {\n let taskStarted = false;\n\n const maxWaitingTime = options.deadline || DEFAULT_PROMISE_DEADLINE;\n\n let timeoutId: NodeJS.Timeout | null = setTimeout(() => {\n if (!taskStarted) {\n reject(\n new Error('Task was not started within the maximum waiting time'),\n );\n }\n }, maxWaitingTime);\n\n executor((value) => {\n taskStarted = true;\n if (timeoutId) {\n clearTimeout(timeoutId);\n }\n timeoutId = null;\n resolve(value);\n }, reject);\n });\n};\n","import { TaskProcessor } from '../../taskProcessing';\n\nexport type LockOptions = { lockId: number };\n\nexport type AcquireLockOptions = { lockId: string };\nexport type ReleaseLockOptions = { lockId: string };\n\nexport type Lock = {\n acquire(options: AcquireLockOptions): Promise<void>;\n tryAcquire(options: AcquireLockOptions): Promise<boolean>;\n release(options: ReleaseLockOptions): Promise<boolean>;\n withAcquire: <Result = unknown>(\n handle: () => Promise<Result>,\n options: AcquireLockOptions,\n ) => Promise<Result>;\n};\n\nexport const InProcessLock = (): Lock => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: Number.MAX_VALUE,\n maxQueueSize: Number.MAX_VALUE,\n });\n\n // Map to store ack functions of currently held locks: lockId -> ack()\n const locks = new Map<string, () => void>();\n\n return {\n async acquire({ lockId }: AcquireLockOptions): Promise<void> {\n // If the lock is already held, we just queue up another task in the same group.\n // TaskProcessor ensures tasks in the same group run one at a time.\n await new Promise<void>((resolve, reject) => {\n taskProcessor\n .enqueue(\n ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n // We do NOT call ack() here. We hold onto the lock.\n resolve();\n return Promise.resolve();\n },\n { taskGroupId: lockId },\n )\n .catch(reject);\n });\n },\n\n async tryAcquire({ lockId }: AcquireLockOptions): Promise<boolean> {\n // If lock is already held, fail immediately\n if (locks.has(lockId)) {\n return false;\n }\n\n // TODO: Check pending queue\n await this.acquire({ lockId });\n\n return true;\n },\n\n release({ lockId }: ReleaseLockOptions): Promise<boolean> {\n const ack = locks.get(lockId);\n if (ack === undefined) {\n return Promise.resolve(true);\n }\n locks.delete(lockId);\n ack();\n return Promise.resolve(true);\n },\n\n async withAcquire<Result = unknown>(\n handle: () => Promise<Result>,\n { lockId }: AcquireLockOptions,\n ): Promise<Result> {\n return taskProcessor.enqueue(\n async ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n\n // We do NOT call ack() here. We hold onto the lock.\n try {\n return await handle();\n } finally {\n locks.delete(lockId);\n ack();\n }\n },\n { taskGroupId: lockId },\n );\n },\n };\n};\n","export const delay = (ms: number): Promise<void> => {\n return new Promise((resolve) => setTimeout(resolve, ms));\n};\n\nexport type AsyncAwaiter<T = void> = {\n wait: Promise<T>;\n resolve: (value: T | PromiseLike<T>) => void;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n reject: (reason?: any) => void;\n reset: () => void;\n};\n\n// TODO: Remove this after migrating to Node 22\nexport const asyncAwaiter = <T = void>(): AsyncAwaiter<T> => {\n const result: AsyncAwaiter<T> = {} as AsyncAwaiter<T>;\n\n (result.reset = () => {\n result.wait = new Promise<T>((res, rej) => {\n result.resolve = res;\n result.reject = rej;\n });\n })();\n\n return result;\n};\n","import retry from 'async-retry';\nimport { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\n\nexport type AsyncRetryOptions<T = unknown> = retry.Options & {\n shouldRetryResult?: (result: T) => boolean;\n shouldRetryError?: (error?: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions<T>,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n const result = await fn();\n\n if (opts?.shouldRetryResult && opts.shouldRetryResult(result)) {\n throw new EmmettError(\n `Retrying because of result: ${JSONParser.stringify(result)}`,\n );\n }\n return result;\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n return undefined as unknown as T;\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","import { ConcurrencyInMemoryDatabaseError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n DatabaseHandleOptionErrors,\n ExpectedDocumentVersion,\n ExpectedDocumentVersionGeneral,\n ExpectedDocumentVersionValue,\n OperationResult,\n} from './types';\n\nexport const isGeneralExpectedDocumentVersion = (\n version: ExpectedDocumentVersion | undefined,\n): version is ExpectedDocumentVersionGeneral => {\n return (\n version === 'DOCUMENT_DOES_NOT_EXIST' ||\n version === 'DOCUMENT_EXISTS' ||\n version === 'NO_CONCURRENCY_CHECK'\n );\n};\n\nexport const expectedVersionValue = (\n version: ExpectedDocumentVersion | undefined,\n): ExpectedDocumentVersionValue | null =>\n version === undefined || isGeneralExpectedDocumentVersion(version)\n ? null\n : version;\n\nexport const operationResult = <T extends OperationResult>(\n result: Omit<T, 'assertSuccess' | 'acknowledged' | 'assertSuccessful'>,\n options: {\n operationName: string;\n collectionName: string;\n errors?: DatabaseHandleOptionErrors;\n },\n): T => {\n const operationResult: T = {\n ...result,\n acknowledged: true,\n successful: result.successful,\n assertSuccessful: (errorMessage?: string) => {\n const { successful } = result;\n const { operationName, collectionName } = options;\n\n if (!successful)\n throw new ConcurrencyInMemoryDatabaseError(\n errorMessage ??\n `${operationName} on ${collectionName} failed. Expected document state does not match current one! Result: ${JSONParser.stringify(result)}!`,\n );\n },\n } as T;\n\n if (options.errors?.throwOnOperationFailures)\n operationResult.assertSuccessful();\n\n return operationResult;\n};\n","import type { InMemoryDatabase } from '../../../database/inMemoryDatabase';\nimport type {\n ProjectionDefinition,\n TruncateProjection,\n} from '../../../projections';\nimport type { CanHandle, Event, ReadEvent } from '../../../typing';\nimport {\n type InMemoryProjectionHandlerContext,\n type InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\nexport const DATABASE_REQUIRED_ERROR_MESSAGE =\n 'Database is required in context for InMemory projections';\n\nexport type InMemoryProjectionDefinition<EventType extends Event> =\n ProjectionDefinition<\n EventType,\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >;\n\nexport type InMemoryProjectionHandlerOptions<EventType extends Event = Event> =\n {\n projections: InMemoryProjectionDefinition<EventType>[];\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[];\n database: InMemoryDatabase;\n eventStore?: InMemoryProjectionHandlerContext['eventStore'];\n };\n\n/**\n * Handles projections for the InMemoryEventStore\n * Similar to the PostgreSQL implementation, this processes events through projections\n */\nexport const handleInMemoryProjections = async <\n EventType extends Event = Event,\n>(\n options: InMemoryProjectionHandlerOptions<EventType>,\n): Promise<void> => {\n const { projections, events, database, eventStore } = options;\n\n // Get all event types from the events batch to filter projections\n const eventTypes = events.map((e) => e.type);\n\n // Filter projections that can handle these event types\n const relevantProjections = projections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n // Process each projection\n for (const projection of relevantProjections) {\n await projection.handle(events, {\n eventStore,\n database,\n });\n }\n};\n\nexport type InMemoryWithNotNullDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = (\n document: DocumentType,\n event: ReadEvent<EventType, InMemoryReadEventMetadata>,\n) => DocumentType | null;\n\nexport type InMemoryWithNullableDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = (\n document: DocumentType | null,\n event: ReadEvent<EventType, InMemoryReadEventMetadata>,\n) => DocumentType | null;\n\nexport type InMemoryDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> =\n | InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>\n | InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n\nexport type InMemoryProjectionOptions<EventType extends Event> = {\n handle: (\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[],\n context: InMemoryProjectionHandlerContext & { database: InMemoryDatabase },\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n truncate?: TruncateProjection<\n InMemoryProjectionHandlerContext & { database: InMemoryDatabase }\n >;\n};\n\n/**\n * Creates an InMemory projection\n */\nexport const inMemoryProjection = <EventType extends Event>({\n truncate,\n handle,\n canHandle,\n}: InMemoryProjectionOptions<EventType>): InMemoryProjectionDefinition<EventType> => ({\n canHandle,\n handle: async (events, context) => {\n if (!context.database) {\n throw new Error(DATABASE_REQUIRED_ERROR_MESSAGE);\n }\n await handle(events, {\n ...context,\n database: context.database,\n });\n },\n truncate: truncate\n ? (context) => {\n if (!context.database) {\n throw new Error(DATABASE_REQUIRED_ERROR_MESSAGE);\n }\n return truncate({\n ...context,\n database: context.database,\n });\n }\n : undefined,\n});\n\n/**\n * Creates a multi-stream projection for InMemoryDatabase\n */\nexport type InMemoryMultiStreamProjectionOptions<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = {\n canHandle: CanHandle<EventType>;\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n} & (\n | {\n evolve: InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n }\n | {\n evolve: InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>;\n initialState: () => DocumentType;\n }\n);\n\n/**\n * Creates a projection that handles events across multiple streams\n */\nexport const inMemoryMultiStreamProjection = <\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n>(\n options: InMemoryMultiStreamProjectionOptions<DocumentType, EventType>,\n): InMemoryProjectionDefinition<EventType> => {\n const { collectionName, getDocumentId, canHandle } = options;\n\n return inMemoryProjection({\n handle: async (\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[],\n { database },\n ) => {\n const collection = database.collection<DocumentType>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), (document) => {\n if ('initialState' in options) {\n return options.evolve(document ?? options.initialState(), event);\n } else {\n return options.evolve(document, event);\n }\n });\n }\n },\n canHandle,\n truncate: async ({\n database,\n }: InMemoryProjectionHandlerContext & { database: InMemoryDatabase }) => {\n // For InMemory database, we can't directly truncate a collection\n // So we'll delete all documents from the collection\n const collection = database.collection<DocumentType>(collectionName);\n const documents = await collection.find();\n\n for (const doc of documents) {\n if (doc && '_id' in doc) {\n const id = doc._id;\n await collection.deleteOne((d) => d._id === id);\n }\n }\n },\n });\n};\n\n/**\n * Creates a single-stream projection for InMemoryDatabase\n */\nexport type InMemorySingleStreamProjectionOptions<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = {\n canHandle: CanHandle<EventType>;\n getDocumentId?: (event: ReadEvent<EventType>) => string;\n collectionName: string;\n} & (\n | {\n evolve: InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n }\n | {\n evolve: InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>;\n initialState: () => DocumentType;\n }\n);\n\n/**\n * Creates a projection that handles events from a single stream\n */\nexport const inMemorySingleStreamProjection = <\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n>(\n options: InMemorySingleStreamProjectionOptions<DocumentType, EventType>,\n): InMemoryProjectionDefinition<EventType> => {\n return inMemoryMultiStreamProjection<DocumentType, EventType>({\n ...options,\n getDocumentId:\n options.getDocumentId ?? ((event) => event.metadata.streamName),\n });\n};\n","import { v4 as uuid } from 'uuid';\nimport {\n handleInMemoryProjections,\n type InMemoryProjectionDefinition,\n} from '.';\nimport {\n getInMemoryDatabase,\n type Document,\n type InMemoryDatabase,\n} from '../../../database';\nimport { isErrorConstructor } from '../../../errors';\nimport { JSONParser } from '../../../serialization';\nimport {\n assertFails,\n AssertionError,\n assertTrue,\n type ThenThrows,\n} from '../../../testing';\nimport type { CombinedReadEventMetadata, ReadEvent } from '../../../typing';\nimport { type Event } from '../../../typing';\nimport type {\n InMemoryEventStore,\n InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\n// Define a more specific type for T that extends Document\ntype DocumentWithId = Document & { _id?: string | number };\n\nexport type InMemoryProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type InMemoryProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type InMemoryProjectionSpec<EventType extends Event> = (\n givenEvents: InMemoryProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n then: (assert: InMemoryProjectionAssert, message?: string) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type InMemoryProjectionAssert = (options: {\n database: InMemoryDatabase;\n}) => Promise<void | boolean>;\n\nexport type InMemoryProjectionSpecOptions<EventType extends Event> = {\n projection: InMemoryProjectionDefinition<EventType>;\n};\n\nexport const InMemoryProjectionSpec = {\n for: <EventType extends Event>(\n options: InMemoryProjectionSpecOptions<EventType>,\n ): InMemoryProjectionSpec<EventType> => {\n const { projection } = options;\n\n return (givenEvents: InMemoryProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, InMemoryReadEventMetadata>[] =\n [];\n\n const run = async (database: InMemoryDatabase) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: InMemoryReadEventMetadata = {\n globalPosition: ++globalPosition,\n streamPosition: globalPosition,\n streamName: event.metadata?.streamName ?? `test-${uuid()}`,\n messageId: uuid(),\n };\n\n allEvents.push({\n ...event,\n kind: 'Event',\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n InMemoryReadEventMetadata\n >,\n });\n }\n\n // Create a minimal mock EventStore implementation\n const mockEventStore = {\n database,\n aggregateStream: async () => {\n return Promise.resolve({\n state: {},\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n readStream: async () => {\n return Promise.resolve({\n events: [],\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n appendToStream: async () => {\n return Promise.resolve({\n nextExpectedStreamVersion: 0n,\n createdNewStream: false,\n });\n },\n } as InMemoryEventStore;\n\n await handleInMemoryProjections({\n events: allEvents,\n projections: [projection],\n database,\n eventStore: mockEventStore,\n });\n };\n\n return {\n then: async (\n assertFn: InMemoryProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n await run(database);\n\n const succeeded = await assertFn({ database });\n\n if (succeeded !== undefined && succeeded === false) {\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n }\n },\n thenThrows: async <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n try {\n await run(database);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n },\n};\n\n// Helper functions for creating events in stream\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n>(\n streamName: string,\n event: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>,\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n>(\n streamName: string,\n events: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[],\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\n// Assertion helpers for checking documents\nexport function documentExists<T extends DocumentWithId>(\n expected: Partial<T>,\n options: { inCollection: string; withId: string | number },\n): InMemoryProjectionAssert {\n return async ({ database }) => {\n const collection = database.collection<T>(options.inCollection);\n\n const document = await collection.findOne((doc) => {\n // Handle both string IDs and numeric IDs in a type-safe way\n const docId = '_id' in doc ? doc._id : undefined;\n return docId === options.withId;\n });\n\n if (!document) {\n assertFails(\n `Document with ID ${options.withId} does not exist in collection ${options.inCollection}`,\n );\n return Promise.resolve(false);\n }\n\n // Check that all expected properties exist with expected values\n for (const [key, value] of Object.entries(expected)) {\n const propKey = key as keyof typeof document;\n if (\n !(key in document) ||\n JSONParser.stringify(document[propKey]) !== JSONParser.stringify(value)\n ) {\n assertFails(`Property ${key} doesn't match the expected value`);\n return Promise.resolve(false);\n }\n }\n\n return Promise.resolve(true);\n };\n}\n\n// Helper for checking document contents\nexport const expectInMemoryDocuments = {\n fromCollection: <T extends DocumentWithId>(collectionName: string) => ({\n withId: (id: string | number) => ({\n toBeEqual: (expected: Partial<T>): InMemoryProjectionAssert =>\n documentExists<T>(expected, {\n inCollection: collectionName,\n withId: id,\n }),\n }),\n }),\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<void>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import { isErrorConstructor, type ErrorConstructor } from '../errors';\nimport { AssertionError, assertThatArray, assertTrue } from './assertions';\n\ntype ErrorCheck<ErrorType> = (error: ErrorType) => boolean;\n\nexport type ThenThrows<ErrorType extends Error> =\n | (() => void)\n | ((errorConstructor: ErrorConstructor<ErrorType>) => void)\n | ((errorCheck: ErrorCheck<ErrorType>) => void)\n | ((\n errorConstructor: ErrorConstructor<ErrorType>,\n errorCheck?: ErrorCheck<ErrorType>,\n ) => void);\n\nexport type DeciderSpecification<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => void;\n thenNothingHappened: () => void;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => void;\n };\n};\nexport type AsyncDeciderSpecification<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => Promise<void>;\n thenNothingHappened: () => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport const DeciderSpecification = {\n for: deciderSpecificationFor,\n};\n\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Event | Event[];\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}): DeciderSpecification<Command, Event>;\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Promise<Event | Event[]>;\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}): AsyncDeciderSpecification<Command, Event>;\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (\n command: Command,\n state: State,\n ) => Event | Event[] | Promise<Event | Event[]>;\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}):\n | DeciderSpecification<Command, Event>\n | AsyncDeciderSpecification<Command, Event> {\n {\n return (givenEvents: Event | Event[]) => {\n return {\n when: (command: Command) => {\n const handle = () => {\n const existingEvents = Array.isArray(givenEvents)\n ? givenEvents\n : [givenEvents];\n\n const currentState = existingEvents.reduce<State>(\n decider.evolve,\n decider.initialState(),\n );\n\n return decider.decide(command, currentState);\n };\n\n return {\n then: (expectedEvents: Event | Event[]): void | Promise<void> => {\n const resultEvents = handle();\n\n if (resultEvents instanceof Promise) {\n return resultEvents.then((events) => {\n thenHandler(events, expectedEvents);\n });\n }\n\n thenHandler(resultEvents, expectedEvents);\n },\n thenNothingHappened: (): void | Promise<void> => {\n const resultEvents = handle();\n\n if (resultEvents instanceof Promise) {\n return resultEvents.then((events) => {\n thenNothingHappensHandler(events);\n });\n }\n\n thenNothingHappensHandler(resultEvents);\n },\n thenThrows: <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): void | Promise<void> => {\n try {\n const result = handle();\n if (result instanceof Promise) {\n return result\n .then(() => {\n throw new AssertionError(\n 'Handler did not fail as expected',\n );\n })\n .catch((error) => {\n thenThrowsErrorHandler(error, args);\n });\n }\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n thenThrowsErrorHandler(error, args);\n }\n },\n };\n },\n };\n };\n }\n}\n\nfunction thenHandler<Event>(\n events: Event | Event[],\n expectedEvents: Event | Event[],\n): void {\n const resultEventsArray = Array.isArray(events) ? events : [events];\n\n const expectedEventsArray = Array.isArray(expectedEvents)\n ? expectedEvents\n : [expectedEvents];\n\n assertThatArray(resultEventsArray).containsOnlyElementsMatching(\n expectedEventsArray,\n );\n}\n\nfunction thenNothingHappensHandler<Event>(events: Event | Event[]): void {\n const resultEventsArray = Array.isArray(events) ? events : [events];\n assertThatArray(resultEventsArray).isEmpty();\n}\n\nfunction thenThrowsErrorHandler<ErrorType extends Error>(\n error: unknown,\n args: Parameters<ThenThrows<ErrorType>>,\n): void {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n}\n","import type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n EventStore,\n EventStoreReadEventMetadata,\n ReadStreamOptions,\n ReadStreamResult,\n StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport { type Event, type EventMetaDataOf } from '../typing';\n\nexport type TestEventStream<EventType extends Event = Event> = [\n string,\n EventType[],\n];\n\nexport type EventStoreWrapper<Store extends EventStore> = Store & {\n appendedEvents: Map<string, TestEventStream>;\n setup<EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>>;\n};\n\nexport const WrapEventStore = <Store extends EventStore>(\n eventStore: Store,\n): EventStoreWrapper<Store> => {\n const appendedEvents = new Map<string, TestEventStream>();\n\n const wrapped = {\n ...eventStore,\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<\n AggregateStreamResult<State, StreamPositionTypeOfEventStore<Store>>\n > {\n return eventStore.aggregateStream(streamName, options);\n },\n\n async readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<\n ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >\n > {\n return (await eventStore.readStream(\n streamName,\n options,\n )) as ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >;\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n const result = await eventStore.appendToStream(\n streamName,\n events,\n options,\n );\n\n const currentStream = appendedEvents.get(streamName) ?? [streamName, []];\n\n appendedEvents.set(streamName, [\n streamName,\n [...currentStream[1], ...events],\n ]);\n\n return result;\n },\n\n appendedEvents,\n\n setup: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n return eventStore.appendToStream(streamName, events);\n },\n\n // streamEvents: (): ReadableStream<\n // // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // > => {\n // return eventStore.streamEvents();\n // },\n };\n\n return wrapped as EventStoreWrapper<Store>;\n};\n","import { TransformStream } from 'web-streams-polyfill';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport { globalStreamCaughtUp, type GlobalSubscriptionEvent } from '../events';\n\nexport const streamTrackingGlobalPosition = (\n currentEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n) => new CaughtUpTransformStream(currentEvents);\n\nexport class CaughtUpTransformStream extends TransformStream<\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n | ReadEvent<Event, ReadEventMetadataWithGlobalPosition>\n | GlobalSubscriptionEvent\n> {\n private _currentPosition: bigint;\n private _logPosition: bigint;\n\n constructor(events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]) {\n super({\n start: (controller) => {\n let globalPosition = 0n;\n for (const event of events) {\n controller.enqueue(event);\n globalPosition = event.metadata.globalPosition;\n }\n controller.enqueue(globalStreamCaughtUp({ globalPosition }));\n },\n transform: (event, controller) => {\n this._currentPosition = event.metadata.globalPosition;\n controller.enqueue(event);\n\n if (this._currentPosition < this._logPosition) return;\n\n controller.enqueue(\n globalStreamCaughtUp({ globalPosition: this._currentPosition }),\n );\n },\n });\n\n this._currentPosition = this._logPosition =\n events.length > 0\n ? events[events.length - 1]!.metadata.globalPosition\n : 0n;\n }\n\n public set logPosition(value: bigint) {\n this._logPosition = value;\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport { notifyAboutNoActiveReadersStream } from '../../streaming/transformations/notifyAboutNoActiveReaders';\nimport { writeToStream } from '../../streaming/writers';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport {\n CaughtUpTransformStream,\n streamTrackingGlobalPosition,\n} from './caughtUpTransformStream';\n\nexport const StreamingCoordinator = () => {\n const allEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[] = [];\n const listeners = new Map<string, CaughtUpTransformStream>();\n\n return {\n notify: async (\n events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n ) => {\n if (events.length === 0) return;\n\n allEvents.push(...events);\n\n for (const listener of listeners.values()) {\n listener.logPosition =\n events[events.length - 1]!.metadata.globalPosition;\n\n await writeToStream(listener, events);\n }\n },\n\n stream: () => {\n const streamId = uuid();\n const transformStream = streamTrackingGlobalPosition(allEvents);\n\n listeners.set(streamId, transformStream);\n return transformStream.readable.pipeThrough(\n notifyAboutNoActiveReadersStream(\n (stream) => {\n if (listeners.has(stream.streamId))\n listeners.delete(stream.streamId);\n },\n { streamId },\n ),\n );\n },\n };\n};\n","import { v4 as uuid } from 'uuid';\nimport { TransformStream } from 'web-streams-polyfill';\n\nexport const notifyAboutNoActiveReadersStream = <Item>(\n onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);\n\nexport class NotifyAboutNoActiveReadersStream<Item> extends TransformStream<\n Item,\n Item\n> {\n private checkInterval: NodeJS.Timeout | null = null;\n public readonly streamId: string;\n private _isStopped: boolean = false;\n public get hasActiveSubscribers() {\n return !this._isStopped;\n }\n\n constructor(\n private onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n ) {\n super({\n cancel: (reason) => {\n console.log('Stream was canceled. Reason:', reason);\n this.stopChecking();\n },\n });\n this.streamId = options?.streamId ?? uuid();\n\n this.onNoActiveReaderCallback = onNoActiveReaderCallback;\n\n this.startChecking(options?.intervalCheckInMs ?? 20);\n }\n\n private startChecking(interval: number) {\n this.checkInterval = setInterval(() => {\n this.checkNoActiveReader();\n }, interval);\n }\n\n private stopChecking() {\n if (!this.checkInterval) return;\n\n clearInterval(this.checkInterval);\n this.checkInterval = null;\n this._isStopped = true;\n this.onNoActiveReaderCallback(this);\n }\n\n private checkNoActiveReader() {\n if (!this.readable.locked && !this._isStopped) {\n this.stopChecking();\n }\n }\n}\n","import type { ReadableStream, TransformStream } from 'web-streams-polyfill';\nimport { noMoreWritingOn } from './noMoreWritingOn';\n\nexport const writeToStream = async <In, Out = In>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<boolean> => {\n if (stream.writable.locked) return false;\n\n const writer = stream.writable.getWriter();\n await writer.ready;\n\n if (!stream.readable.locked) return false;\n\n try {\n for (const item of items) {\n await writer.write(item);\n }\n } catch (error) {\n console.log(error);\n } finally {\n await writer.close();\n }\n return true;\n};\n\nexport const writeToStreamAndStop = async <In, Out>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<ReadableStream<Out>> => {\n await writeToStream(stream, items);\n return await noMoreWritingOn(stream);\n};\n","import {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport type { Event } from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\n\nexport const CommandHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type CommandHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromCommandHandlerRetryOptions = (\n retryOptions: CommandHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return CommandHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...CommandHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region command-handler\nexport type CommandHandlerResult<\n State,\n StreamEvent extends Event,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newState: State;\n newEvents: StreamEvent[];\n};\n\nexport type CommandHandlerOptions<State, StreamEvent extends Event> = {\n evolve: (state: State, event: StreamEvent) => State;\n initialState: () => State;\n mapToStreamId?: (id: string) => string;\n retry?: CommandHandlerRetryOptions;\n};\n\nexport type HandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] &\n (\n | {\n expectedStreamVersion?: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >;\n }\n | {\n retry?: CommandHandlerRetryOptions;\n }\n );\n\ntype CommandHandlerFunction<State, StreamEvent extends Event> = (\n state: State,\n) => StreamEvent | StreamEvent[] | Promise<StreamEvent | StreamEvent[]>;\n\nexport const CommandHandler =\n <State, StreamEvent extends Event>(\n options: CommandHandlerOptions<State, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n store: Store,\n id: string,\n handle:\n | CommandHandlerFunction<State, StreamEvent>\n | CommandHandlerFunction<State, StreamEvent>[],\n handleOptions?: HandleOptions<Store>,\n ): Promise<CommandHandlerResult<State, StreamEvent, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n CommandHandlerResult<\n State,\n StreamEvent,\n StreamPositionTypeOfEventStore<Store>\n >\n >(store, async ({ eventStore }) => {\n const { evolve, initialState } = options;\n const mapToStreamId = options.mapToStreamId ?? ((id) => id);\n\n const streamName = mapToStreamId(id);\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n State,\n StreamEvent\n >(streamName, {\n evolve,\n initialState,\n read: {\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n currentStreamVersion,\n streamExists: _streamExists,\n ...restOfAggregationResult\n } = aggregationResult;\n\n let state = aggregationResult.state;\n\n const handlers = Array.isArray(handle) ? handle : [handle];\n let eventsToAppend: StreamEvent[] = [];\n\n // 3. Run business logic\n for (const handler of handlers) {\n const result = await handler(state);\n\n const newEvents = Array.isArray(result) ? result : [result];\n\n if (newEvents.length > 0) {\n state = newEvents.reduce(evolve, state);\n }\n\n eventsToAppend = [...eventsToAppend, ...newEvents];\n }\n\n //const newEvents = Array.isArray(result) ? result : [result];\n\n if (eventsToAppend.length === 0) {\n return {\n ...restOfAggregationResult,\n newEvents: [],\n newState: state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n nextExpectedStreamVersion: currentStreamVersion,\n createdNewStream: false,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const expectedStreamVersion: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n > =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? (currentStreamVersion as ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >)\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n eventsToAppend,\n {\n ...handleOptions,\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with updated state\n return {\n ...appendResult,\n newEvents: eventsToAppend,\n newState: state,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n });\n\n return result;\n },\n fromCommandHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion command-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type { EventStore } from '../eventStore';\nimport { type Command, type Event } from '../typing';\nimport type { Decider } from '../typing/decider';\nimport {\n CommandHandler,\n type CommandHandlerOptions,\n type HandleOptions,\n} from './handleCommand';\n\n// #region command-handler\n\nexport type DeciderCommandHandlerOptions<\n State,\n CommandType extends Command,\n StreamEvent extends Event,\n> = CommandHandlerOptions<State, StreamEvent> &\n Decider<State, CommandType, StreamEvent>;\n\nexport const DeciderCommandHandler =\n <State, CommandType extends Command, StreamEvent extends Event>(\n options: DeciderCommandHandlerOptions<State, CommandType, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n eventStore: Store,\n id: string,\n commands: CommandType | CommandType[],\n handleOptions?: HandleOptions<Store>,\n ) => {\n const { decide, ...rest } = options;\n\n const deciders = (Array.isArray(commands) ? commands : [commands]).map(\n (command) => (state: State) => decide(command, state),\n );\n\n return CommandHandler<State, StreamEvent>(rest)(\n eventStore,\n id,\n deciders,\n handleOptions,\n );\n };\n// #endregion command-handler\n","import { EmmettError } from '../errors';\nimport {\n type AnyCommand,\n type AnyMessage,\n type Command,\n type CommandTypeOf,\n type Event,\n type EventTypeOf,\n type Message,\n type SingleMessageHandler,\n type SingleRawMessageHandlerWithoutContext,\n} from '../typing';\n\nexport interface CommandSender {\n send<CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void>;\n}\n\nexport interface EventsPublisher {\n publish<EventType extends Event = Event>(event: EventType): Promise<void>;\n}\n\nexport type ScheduleOptions = { afterInMs: number } | { at: Date };\n\nexport interface MessageScheduler<CommandOrEvent extends Command | Event> {\n schedule<MessageType extends CommandOrEvent>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandBus extends CommandSender, MessageScheduler<Command> {}\n\nexport interface EventBus extends EventsPublisher, MessageScheduler<Event> {}\n\nexport interface MessageBus extends CommandBus, EventBus {\n schedule<MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandProcessor {\n handle<CommandType extends Command>(\n commandHandler: SingleMessageHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void;\n}\nexport interface EventSubscription {\n subscribe<EventType extends Event>(\n eventHandler: SingleMessageHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void;\n}\n\nexport type ScheduledMessage = {\n message: Message;\n options?: ScheduleOptions;\n};\n\nexport interface ScheduledMessageProcessor {\n dequeue(): ScheduledMessage[];\n}\n\nexport type MessageSubscription = EventSubscription | CommandProcessor;\n\nexport const getInMemoryMessageBus = (): MessageBus &\n EventSubscription &\n CommandProcessor &\n ScheduledMessageProcessor => {\n const allHandlers = new Map<\n string,\n SingleRawMessageHandlerWithoutContext<AnyMessage>[]\n >();\n let pendingMessages: ScheduledMessage[] = [];\n\n return {\n send: async <CommandType extends Command = AnyCommand>(\n command: CommandType,\n ): Promise<void> => {\n const handlers = allHandlers.get(command.type);\n\n if (handlers === undefined || handlers.length === 0)\n throw new EmmettError(\n `No handler registered for command ${command.type}!`,\n );\n\n const commandHandler = handlers[0]!;\n\n await commandHandler(command);\n },\n\n publish: async <EventType extends Event = Event>(\n event: EventType,\n ): Promise<void> => {\n const handlers = allHandlers.get(event.type) ?? [];\n\n for (const handler of handlers) {\n const eventHandler = handler;\n\n await eventHandler(event);\n }\n },\n\n schedule: <MessageType extends Message>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void => {\n pendingMessages = [...pendingMessages, { message, options: when }];\n },\n\n handle: <CommandType extends Command>(\n commandHandler: SingleMessageHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void => {\n const alreadyRegistered = [...allHandlers.keys()].filter((registered) =>\n commandTypes.includes(registered),\n );\n\n if (alreadyRegistered.length > 0)\n throw new EmmettError(\n `Cannot register handler for commands ${alreadyRegistered.join(', ')} as they're already registered!`,\n );\n for (const commandType of commandTypes) {\n allHandlers.set(commandType, [\n commandHandler as SingleRawMessageHandlerWithoutContext<AnyMessage>,\n ]);\n }\n },\n\n subscribe<EventType extends Event>(\n eventHandler: SingleMessageHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void {\n for (const eventType of eventTypes) {\n if (!allHandlers.has(eventType)) allHandlers.set(eventType, []);\n\n allHandlers.set(eventType, [\n ...(allHandlers.get(eventType) ?? []),\n eventHandler as SingleRawMessageHandlerWithoutContext<AnyMessage>,\n ]);\n }\n },\n\n dequeue: (): ScheduledMessage[] => {\n const pending = pendingMessages;\n pendingMessages = [];\n return pending;\n },\n };\n};\n","import type { EmmettError } from '../errors';\nimport type { ProjectionDefinition } from '../projections';\nimport {\n type AnyEvent,\n type AnyMessage,\n type AnyReadEventMetadata,\n type AnyRecordedMessageMetadata,\n type BatchRecordedMessageHandlerWithContext,\n type CanHandle,\n type DefaultRecord,\n type Event,\n type GlobalPositionTypeOfRecordedMessageMetadata,\n type Message,\n type MessageHandlerResult,\n type RecordedMessage,\n type SingleMessageHandlerWithContext,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport { isBigint } from '../validation';\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type CurrentMessageProcessorPosition<CheckpointType = any> =\n | { lastCheckpoint: CheckpointType }\n | 'BEGINNING'\n | 'END';\n\nexport type GetCheckpoint<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n) => CheckpointType | null;\n\nexport const getCheckpoint = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n): CheckpointType | null => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return 'checkpoint' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.checkpoint)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.checkpoint\n : 'globalPosition' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.globalPosition)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.globalPosition\n : 'streamPosition' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.streamPosition)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.streamPosition\n : null;\n};\n\nexport const wasMessageHandled = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n checkpoint: CheckpointType | null,\n): boolean => {\n //TODO Make it smarter\n const messageCheckpoint = getCheckpoint(message);\n const checkpointBigint = checkpoint as bigint | null;\n\n return (\n messageCheckpoint !== null &&\n messageCheckpoint !== undefined &&\n checkpointBigint !== null &&\n checkpointBigint !== undefined &&\n messageCheckpoint <= checkpointBigint\n );\n};\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type MessageProcessorStartFrom<CheckpointType = any> =\n | CurrentMessageProcessorPosition<CheckpointType>\n | 'CURRENT';\n\nexport type MessageProcessorType = 'projector' | 'reactor';\nexport const MessageProcessorType = {\n PROJECTOR: 'projector' as MessageProcessorType,\n REACTOR: 'reactor' as MessageProcessorType,\n};\n\nexport type MessageProcessor<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n id: string;\n type: string;\n start: (\n options: Partial<HandlerContext>,\n ) => Promise<CurrentMessageProcessorPosition<CheckpointType> | undefined>;\n close: () => Promise<void>;\n isActive: boolean;\n handle: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n Partial<HandlerContext>\n >;\n};\n\nexport const MessageProcessor = {\n result: {\n skip: (options?: { reason?: string }): MessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): MessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type MessageProcessingScope<\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = <Result = MessageHandlerResult>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => Result | Promise<Result>;\n\nexport type Checkpointer<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n read: ReadProcessorCheckpoint<CheckpointType, HandlerContext>;\n store: StoreProcessorCheckpoint<\n MessageType,\n MessageMetadataType,\n CheckpointType,\n HandlerContext\n >;\n};\n\nexport type BaseMessageProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n type?: string;\n processorId: string;\n version?: number;\n partition?: string;\n startFrom?: MessageProcessorStartFrom<CheckpointType>;\n stopAfter?: (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n ) => boolean;\n processingScope?: MessageProcessingScope<HandlerContext>;\n checkpoints?: Checkpointer<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >;\n canHandle?: CanHandle<MessageType>;\n hooks?: {\n onStart?: OnReactorStartHook<HandlerContext>;\n onClose?: OnReactorCloseHook;\n };\n};\n\nexport type HandlerOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> =\n | {\n eachMessage: SingleRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n eachBatch?: never;\n }\n | {\n eachMessage?: never;\n eachBatch: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n };\n\nexport type OnReactorStartHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorCloseHook = () => Promise<void>;\n\nexport type ReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = BaseMessageProcessorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n> &\n HandlerOptions<MessageType, MessageMetadataType, HandlerContext>;\n\nexport type ProjectorOptions<\n EventType extends AnyEvent = AnyEvent,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = Omit<\n BaseMessageProcessorOptions<\n EventType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >,\n 'type' | 'processorId'\n> & { processorId?: string } & {\n truncateOnStart?: boolean;\n projection: ProjectionDefinition<\n EventType,\n MessageMetadataType,\n HandlerContext\n >;\n};\n\nexport const defaultProcessingMessageProcessingScope = <\n HandlerContext = never,\n Result = MessageHandlerResult,\n>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => handler(partialContext as HandlerContext);\n\nexport type ReadProcessorCheckpointResult<CheckpointType = unknown> = {\n lastCheckpoint: CheckpointType | null;\n};\n\nexport type ReadProcessorCheckpoint<\n CheckpointType = unknown,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n options: { processorId: string; partition?: string },\n context: HandlerContext,\n) => Promise<ReadProcessorCheckpointResult<CheckpointType>>;\n\nexport type StoreProcessorCheckpointResult<CheckpointType = unknown> =\n | {\n success: true;\n newCheckpoint: CheckpointType;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport type StoreProcessorCheckpoint<\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = unknown,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> =\n | ((\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: CheckpointType | null;\n partition?: string;\n },\n context: HandlerContext,\n ) => Promise<StoreProcessorCheckpointResult<CheckpointType | null>>)\n | ((\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: CheckpointType | null;\n partition?: string;\n },\n context: HandlerContext,\n ) => Promise<StoreProcessorCheckpointResult<CheckpointType>>);\n\nexport const reactor = <\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n options: ReactorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >,\n): MessageProcessor<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n> => {\n const eachMessage: SingleMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n > =\n 'eachMessage' in options && options.eachMessage\n ? options.eachMessage\n : () => Promise.resolve();\n let isActive = true;\n\n const { checkpoints, processorId, partition } = options;\n\n const processingScope =\n options.processingScope ?? defaultProcessingMessageProcessingScope;\n\n let lastCheckpoint: CheckpointType | null = null;\n\n return {\n id: options.processorId,\n type: options.type ?? MessageProcessorType.REACTOR,\n close: () =>\n options.hooks?.onClose ? options.hooks?.onClose() : Promise.resolve(),\n start: async (\n startOptions: Partial<HandlerContext>,\n ): Promise<CurrentMessageProcessorPosition<CheckpointType> | undefined> => {\n isActive = true;\n\n if (lastCheckpoint !== null)\n return {\n lastCheckpoint,\n };\n\n return await processingScope(async (context) => {\n if (options.hooks?.onStart) {\n await options.hooks?.onStart(context);\n }\n\n if (options.startFrom !== 'CURRENT' && options.startFrom)\n return options.startFrom;\n\n if (checkpoints) {\n const readResult = await checkpoints?.read(\n {\n processorId: processorId,\n partition: partition,\n },\n { ...startOptions, ...context },\n );\n lastCheckpoint = readResult.lastCheckpoint;\n }\n\n if (lastCheckpoint === null) return 'BEGINNING';\n\n return {\n lastCheckpoint,\n };\n }, startOptions);\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n messages: RecordedMessage<MessageType, MessageMetadataType>[],\n partialContext: Partial<HandlerContext>,\n ): Promise<MessageHandlerResult> => {\n if (!isActive) return Promise.resolve();\n\n return await processingScope(async (context) => {\n let result: MessageHandlerResult = undefined;\n\n for (const message of messages) {\n if (wasMessageHandled(message, lastCheckpoint)) continue;\n\n const messageProcessingResult = await eachMessage(message, context);\n\n if (checkpoints) {\n const storeCheckpointResult: StoreProcessorCheckpointResult<CheckpointType | null> =\n await checkpoints.store(\n {\n processorId: options.processorId,\n version: options.version,\n message,\n lastCheckpoint,\n partition: options.partition,\n },\n context,\n );\n\n if (storeCheckpointResult.success) {\n // TODO: Add correct handling of the storing checkpoint\n lastCheckpoint = storeCheckpointResult.newCheckpoint;\n }\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(message)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n }, partialContext);\n },\n };\n};\n\nexport const projector = <\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<EventMetaDataType>,\n>(\n options: ProjectorOptions<\n EventType,\n EventMetaDataType,\n HandlerContext,\n CheckpointType\n >,\n): MessageProcessor<\n EventType,\n EventMetaDataType,\n HandlerContext,\n CheckpointType\n> => {\n const { projection, ...rest } = options;\n\n return reactor<EventType, EventMetaDataType, HandlerContext, CheckpointType>({\n ...rest,\n type: MessageProcessorType.PROJECTOR,\n processorId: options.processorId ?? `projection:${projection.name}`,\n hooks: {\n onStart:\n (options.truncateOnStart && options.projection.truncate) ||\n options.hooks?.onStart\n ? async (context: HandlerContext) => {\n if (options.truncateOnStart && options.projection.truncate)\n await options.projection.truncate(context);\n\n if (options.hooks?.onStart) await options.hooks?.onStart(context);\n }\n : undefined,\n onClose: options.hooks?.onClose,\n },\n eachMessage: async (\n event: RecordedMessage<EventType, EventMetaDataType>,\n context: HandlerContext,\n ) => {\n if (!projection.canHandle.includes(event.type)) return;\n\n await projection.handle([event], context);\n },\n });\n};\n","import { getInMemoryDatabase, type InMemoryDatabase } from '../database';\nimport { EmmettError } from '../errors';\nimport {\n type AnyEvent,\n type AnyMessage,\n type BatchRecordedMessageHandlerWithContext,\n type MessageHandlerResult,\n type ReadEventMetadataWithGlobalPosition,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport {\n getCheckpoint,\n MessageProcessor,\n projector,\n reactor,\n type Checkpointer,\n type MessageProcessingScope,\n type ProjectorOptions,\n type ReactorOptions,\n} from './processors';\n\nexport type InMemoryProcessorHandlerContext = {\n database: InMemoryDatabase;\n};\n\nexport type InMemoryProcessor<MessageType extends AnyMessage = AnyMessage> =\n MessageProcessor<\n MessageType,\n // TODO: generalize this to support other metadata types\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n > & { database: InMemoryDatabase };\n\nexport type InMemoryProcessorEachMessageHandler<\n MessageType extends AnyMessage = AnyMessage,\n> = SingleRecordedMessageHandlerWithContext<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n>;\n\nexport type InMemoryProcessorEachBatchHandler<\n MessageType extends AnyMessage = AnyMessage,\n> = BatchRecordedMessageHandlerWithContext<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n>;\n\nexport type InMemoryProcessorConnectionOptions = {\n database?: InMemoryDatabase;\n};\n\ntype CheckpointDocument = {\n _id: string;\n lastCheckpoint: bigint | null;\n};\n\nexport type InMemoryCheckpointer<MessageType extends AnyMessage = AnyMessage> =\n Checkpointer<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n >;\n\nexport const inMemoryCheckpointer = <\n MessageType extends AnyMessage = AnyMessage,\n>(): InMemoryCheckpointer<MessageType> => {\n return {\n read: async ({ processorId }, { database }) => {\n const checkpoint = await database\n .collection<CheckpointDocument>('emt_processor_checkpoints')\n .findOne((d) => d._id === processorId);\n\n return Promise.resolve({\n lastCheckpoint: checkpoint?.lastCheckpoint ?? null,\n });\n },\n store: async (context, { database }) => {\n const { message, processorId, lastCheckpoint } = context;\n const checkpoints = database.collection<CheckpointDocument>(\n 'emt_processor_checkpoints',\n );\n\n const checkpoint = await checkpoints.findOne(\n (d) => d._id === processorId,\n );\n\n const currentPosition = checkpoint?.lastCheckpoint ?? null;\n\n const newCheckpoint: bigint | null = getCheckpoint(message);\n\n if (\n currentPosition &&\n (currentPosition === newCheckpoint ||\n currentPosition !== lastCheckpoint)\n ) {\n return {\n success: false,\n reason: currentPosition === newCheckpoint ? 'IGNORED' : 'MISMATCH',\n };\n }\n\n await checkpoints.handle(processorId, (existing) => ({\n ...(existing ?? {}),\n _id: processorId,\n lastCheckpoint: newCheckpoint,\n }));\n\n return { success: true, newCheckpoint };\n },\n };\n};\n\ntype InMemoryConnectionOptions = {\n connectionOptions?: InMemoryProcessorConnectionOptions;\n};\n\nexport type InMemoryReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n> = ReactorOptions<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n> &\n InMemoryConnectionOptions;\n\nexport type InMemoryProjectorOptions<EventType extends AnyEvent = AnyEvent> =\n ProjectorOptions<\n EventType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n > &\n InMemoryConnectionOptions;\n\nexport type InMemoryProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n> =\n | InMemoryReactorOptions<MessageType>\n | InMemoryProjectorOptions<MessageType & AnyEvent>;\n\nconst inMemoryProcessingScope = (options: {\n database: InMemoryDatabase | null;\n processorId: string;\n}): MessageProcessingScope<InMemoryProcessorHandlerContext> => {\n const processorDatabase = options.database;\n\n const processingScope: MessageProcessingScope<\n InMemoryProcessorHandlerContext\n > = <Result = MessageHandlerResult>(\n handler: (\n context: InMemoryProcessorHandlerContext,\n ) => Result | Promise<Result>,\n partialContext: Partial<InMemoryProcessorHandlerContext>,\n ) => {\n const database = processorDatabase ?? partialContext?.database;\n\n if (!database)\n throw new EmmettError(\n `InMemory processor '${options.processorId}' is missing database. Ensure that you passed it through options`,\n );\n\n return handler({ ...partialContext, database });\n };\n\n return processingScope;\n};\n\nexport const inMemoryProjector = <EventType extends AnyEvent = AnyEvent>(\n options: InMemoryProjectorOptions<EventType>,\n): InMemoryProcessor<EventType> => {\n const database = options.connectionOptions?.database ?? getInMemoryDatabase();\n\n const hooks = {\n onStart: options.hooks?.onStart,\n onClose: options.hooks?.onClose\n ? async () => {\n if (options.hooks?.onClose) await options.hooks?.onClose();\n }\n : undefined,\n };\n\n return {\n ...projector<\n EventType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n >({\n ...options,\n hooks,\n processingScope: inMemoryProcessingScope({\n database,\n processorId:\n options.processorId ?? `projection:${options.projection.name}`,\n }),\n checkpoints: inMemoryCheckpointer<EventType>(),\n }),\n database,\n };\n};\n\nexport const inMemoryReactor = <MessageType extends AnyMessage = AnyMessage>(\n options: InMemoryReactorOptions<MessageType>,\n): InMemoryProcessor<MessageType> => {\n const database = options.connectionOptions?.database ?? getInMemoryDatabase();\n\n const hooks = {\n onStart: options.hooks?.onStart,\n onClose: options.hooks?.onClose\n ? async () => {\n if (options.hooks?.onClose) await options.hooks?.onClose();\n }\n : undefined,\n };\n\n return {\n ...reactor({\n ...options,\n hooks,\n processingScope: inMemoryProcessingScope({\n database,\n processorId: options.processorId,\n }),\n checkpoints: inMemoryCheckpointer<MessageType>(),\n }),\n database,\n };\n};\n","import { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyEvent,\n AnyReadEventMetadata,\n BatchRecordedMessageHandlerWithContext,\n CanHandle,\n DefaultRecord,\n Event,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = BatchRecordedMessageHandlerWithContext<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n>;\n\nexport type TruncateProjection<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: ProjectionHandlerContext) => Promise<void>;\n\nexport interface ProjectionDefinition<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n truncate?: TruncateProjection<ProjectionHandlerContext>;\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definition: ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >,\n): ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n> => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n any,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","export const concatUint8Arrays = (chunks: Uint8Array[]): Uint8Array => {\n const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const chunk of chunks) {\n result.set(chunk, offset);\n offset += chunk.length;\n }\n return result;\n};\n","import type { ReadableStream } from 'web-streams-polyfill';\n\nexport const collect = async <T>(stream: ReadableStream<T>): Promise<T[]> => {\n const results: T[] = [];\n\n for await (const value of stream) {\n results.push(value as T);\n }\n\n return results;\n};\n","import type { Decoder } from '.';\nimport { concatUint8Arrays } from '../binaryArrays';\n\nexport class BinaryJsonDecoder<Decoded>\n implements Decoder<Uint8Array, Decoded>\n{\n private buffer: Uint8Array[] = [];\n\n addToBuffer(data: Uint8Array): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n return text.includes('\\n');\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage()) {\n return null;\n }\n\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n const delimiterIndex = text.indexOf('\\n');\n\n if (delimiterIndex === -1) {\n return null;\n }\n\n const jsonString = text.slice(0, delimiterIndex);\n const remaining = new Uint8Array(combined.buffer, delimiterIndex + 1);\n this.buffer = remaining.byteLength > 0 ? [remaining] : [];\n\n return JSON.parse(jsonString) as Decoded;\n }\n}\n","import type { Decoder } from '.';\n\nexport class StringDecoder<Decoded> implements Decoder<string, Decoded> {\n protected buffer: string[] = [];\n\n constructor(private transform: (input: string) => Decoded) {\n this.transform = transform;\n }\n\n addToBuffer(data: string): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer.some((chunk) => chunk.includes('\\n'));\n }\n\n decode(): Decoded | null {\n const completeString = this.buffer.join('');\n\n if (!this.hasCompleteMessage()) {\n if (completeString.trim().length > 0) {\n throw new Error('Unterminated string in JSON at position');\n }\n return null;\n }\n\n const delimiterIndex = completeString.indexOf('\\n');\n const message = completeString.slice(0, delimiterIndex).trim();\n this.buffer = [completeString.slice(delimiterIndex + 1)];\n\n return this.transform(message);\n }\n}\n","import { StringDecoder } from './string';\n\nexport class JsonDecoder<Decoded> extends StringDecoder<Decoded> {\n constructor() {\n super((jsonString) => JSON.parse(jsonString) as Decoded);\n }\n}\n","import type { Decoder } from '.';\n\nexport class ObjectDecoder<Decoded> implements Decoder<Decoded, Decoded> {\n private buffer: Decoded | null = null;\n\n addToBuffer(data: Decoded): void {\n this.buffer = data;\n }\n\n clearBuffer(): void {\n this.buffer = null;\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer !== null;\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage() || !this.buffer) {\n return null;\n }\n\n const data = this.buffer;\n this.clearBuffer();\n return data as Decoded;\n }\n}\n","import type { Decoder } from '.';\nimport { BinaryJsonDecoder } from './binary';\nimport { JsonDecoder } from './json';\nimport { ObjectDecoder } from './object';\n\nexport class CompositeDecoder<Decoded> implements Decoder<unknown, Decoded> {\n constructor(\n private decoders: [(data: unknown) => boolean, Decoder<unknown, Decoded>][],\n ) {}\n\n private decoderFor(data: unknown): Decoder<unknown, unknown> | null {\n const decoder = this.decoders.find((d) => d[0](data));\n\n if (!decoder) return null;\n\n return decoder[1];\n }\n\n addToBuffer(data: unknown): void {\n this.decoderFor(data)?.addToBuffer(data);\n }\n\n clearBuffer(): void {\n for (const decoder of this.decoders.map((d) => d[1])) {\n decoder.clearBuffer();\n }\n }\n\n hasCompleteMessage(): boolean {\n return this.decoders.some((d) => d[1].hasCompleteMessage());\n }\n\n decode(): Decoded | null {\n const decoder = this.decoders\n .map((d) => d[1])\n .find((d) => d.hasCompleteMessage());\n\n return decoder?.decode() ?? null;\n }\n}\n\nexport class DefaultDecoder<Decoded> extends CompositeDecoder<Decoded> {\n constructor() {\n super([\n [(data) => typeof data === 'string', new JsonDecoder<Decoded>()],\n [(data) => data instanceof Uint8Array, new BinaryJsonDecoder<Decoded>()],\n [(data) => typeof data === 'object', new ObjectDecoder<Decoded>()],\n ]);\n }\n}\n","import { ReadableStream } from 'web-streams-polyfill';\n\nexport const fromArray = <T>(chunks: T[]) =>\n new ReadableStream<T>({\n start(controller) {\n for (const chunk of chunks) controller.enqueue(chunk);\n controller.close();\n },\n });\n","import { fromArray } from './fromArray';\nexport const streamGenerators = { fromArray };\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport type { AsyncRetryOptions } from '../utils';\nimport type { Decoder } from './decoders';\nimport { DefaultDecoder } from './decoders/composite';\nimport { streamTransformations } from './transformations';\n\nconst { retry } = streamTransformations;\n\nexport const restream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n): ReadableStream<Transformed> =>\n retry(createSourceStream, handleChunk(transform, decoder), retryOptions)\n .readable;\n\nconst handleChunk =\n <Source = unknown, Transformed = Source, StreamType = Source>(\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n ) =>\n (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ): void => {\n const { done: isDone, value } = readResult;\n\n if (value) decoder.addToBuffer(value);\n\n if (!isDone && !decoder.hasCompleteMessage()) return;\n\n decodeAndTransform(decoder, transform, controller);\n };\n\nconst decodeAndTransform = <StreamType, Source, Transformed = Source>(\n decoder: Decoder<StreamType, Source>,\n transform: (input: Source) => Transformed,\n controller: TransformStreamDefaultController<Transformed>,\n) => {\n try {\n const decoded = decoder.decode();\n if (!decoded) return; // TODO: Add a proper handling of decode errors\n\n const transformed = transform(decoded);\n controller.enqueue(transformed);\n } catch (error) {\n controller.error(new Error(`Decoding error: ${error?.toString()}`));\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const filter = <Item>(filter: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n if (filter(chunk)) {\n controller.enqueue(chunk);\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const map = <From, To>(map: (item: From) => To) =>\n new TransformStream<From, To>({\n transform(chunk, controller) {\n controller.enqueue(map(chunk));\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const reduce = <I, O>(\n reducer: (accumulator: O, chunk: I) => O,\n initialValue: O,\n) => new ReduceTransformStream<I, O>(reducer, initialValue);\n\nexport class ReduceTransformStream<I, O> extends TransformStream<I, O> {\n private accumulator: O;\n private reducer: (accumulator: O, chunk: I) => O;\n\n constructor(reducer: (accumulator: O, chunk: I) => O, initialValue: O) {\n super({\n transform: (chunk) => {\n this.accumulator = this.reducer(this.accumulator, chunk);\n },\n flush: (controller) => {\n controller.enqueue(this.accumulator);\n controller.terminate();\n },\n });\n\n this.accumulator = initialValue;\n this.reducer = reducer;\n }\n}\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n TransformStream,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport { type AsyncRetryOptions, asyncRetry } from '../../utils';\n\nexport const retryStream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n): TransformStream<Source, Transformed> =>\n new TransformStream<Source, Transformed>({\n start(controller) {\n asyncRetry(\n () => onRestream(createSourceStream, handleChunk, controller),\n retryOptions,\n ).catch((error) => {\n controller.error(error);\n });\n },\n });\n\nconst onRestream = async <StreamType, Source, Transformed = Source>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n controller: TransformStreamDefaultController<Transformed>,\n): Promise<void> => {\n const sourceStream = createSourceStream();\n const reader = sourceStream.getReader();\n\n try {\n let done: boolean;\n\n do {\n const result = await reader.read();\n done = result.done;\n\n await handleChunk(result, controller);\n\n if (done) {\n controller.terminate();\n }\n } while (!done);\n } finally {\n reader.releaseLock();\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const skip = <T>(limit: number) => new SkipTransformStream<T>(limit);\n\nexport class SkipTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private skip: number;\n\n constructor(skip: number) {\n super({\n transform: (chunk, controller) => {\n this.count++;\n if (this.count > this.skip) {\n controller.enqueue(chunk);\n }\n },\n });\n\n this.skip = skip;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopAfter = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n controller.enqueue(chunk);\n\n if (stopCondition(chunk)) {\n controller.terminate();\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopOn = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n async transform(chunk, controller) {\n if (!stopCondition(chunk)) {\n controller.enqueue(chunk);\n return;\n }\n await Promise.resolve();\n controller.terminate();\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const take = <T>(limit: number) => new TakeTransformStream<T>(limit);\n\nexport class TakeTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private limit: number;\n\n constructor(limit: number) {\n super({\n transform: (chunk, controller) => {\n if (this.count < this.limit) {\n this.count++;\n controller.enqueue(chunk);\n } else {\n controller.terminate();\n }\n },\n });\n\n this.limit = limit;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const waitAtMost = <Item>(waitTimeInMs: number) =>\n new TransformStream<Item, Item>({\n start(controller) {\n const timeoutId = setTimeout(() => {\n controller.terminate();\n }, waitTimeInMs);\n\n const originalTerminate = controller.terminate.bind(controller);\n\n // Clear the timeout if the stream is terminated early\n controller.terminate = () => {\n clearTimeout(timeoutId);\n originalTerminate();\n };\n },\n transform(chunk, controller) {\n controller.enqueue(chunk);\n },\n });\n","import { filter } from './filter';\nimport { map } from './map';\nimport {\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n} from './notifyAboutNoActiveReaders';\nimport { reduce, ReduceTransformStream } from './reduce';\nimport { retryStream } from './retry';\nimport { skip, SkipTransformStream } from './skip';\nimport { stopAfter } from './stopAfter';\nimport { stopOn } from './stopOn';\nimport { take, TakeTransformStream } from './take';\nimport { waitAtMost } from './waitAtMost';\n\nexport const streamTransformations = {\n filter,\n take,\n TakeTransformStream,\n skip,\n SkipTransformStream,\n map,\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n reduce,\n ReduceTransformStream,\n retry: retryStream,\n stopAfter,\n stopOn,\n waitAtMost,\n};\n","import {\n JSONParser,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AppendToStreamOptions,\n type BeforeEventStoreCommitHandler,\n type ExpectedStreamVersion,\n type Event as Message,\n type RecordedMessage,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport {\n isSQLiteError,\n type Parameters,\n type SQLiteConnection,\n type SQLiteError,\n} from '../../connection';\nimport type {\n SQLiteEventStore,\n SQLiteReadEventMetadata,\n} from '../SQLiteEventStore';\nimport { defaultTag, messagesTable, streamsTable } from './typing';\n\nexport type AppendEventResult =\n | {\n success: true;\n nextStreamPosition: bigint;\n lastGlobalPosition: bigint;\n }\n | { success: false };\n\nexport const appendToStream = async <MessageType extends Message>(\n connection: SQLiteConnection,\n streamName: string,\n streamType: string,\n messages: MessageType[],\n options?: AppendToStreamOptions & {\n partition?: string;\n onBeforeCommit?: BeforeEventStoreCommitHandler<\n SQLiteEventStore,\n { connection: SQLiteConnection }\n >;\n },\n): Promise<AppendEventResult> => {\n if (messages.length === 0) return { success: false };\n\n const expectedStreamVersion = toExpectedVersion(\n options?.expectedStreamVersion,\n );\n\n const messagesToAppend: RecordedMessage<\n MessageType,\n SQLiteReadEventMetadata\n >[] = messages.map(\n (\n m: Message,\n i: number,\n ): RecordedMessage<MessageType, SQLiteReadEventMetadata> =>\n ({\n ...m,\n kind: m.kind ?? 'Event',\n metadata: {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(i + 1),\n ...('metadata' in m ? (m.metadata ?? {}) : {}),\n },\n }) as RecordedMessage<MessageType, SQLiteReadEventMetadata>,\n );\n\n let result: AppendEventResult;\n\n return await connection.withTransaction(async () => {\n result = await appendToStreamRaw(\n connection,\n streamName,\n streamType,\n messagesToAppend,\n {\n expectedStreamVersion,\n },\n );\n\n if (options?.onBeforeCommit)\n await options.onBeforeCommit(messagesToAppend, { connection });\n\n return result;\n });\n};\n\nconst toExpectedVersion = (\n expected: ExpectedStreamVersion | undefined,\n): bigint | null => {\n if (expected === undefined) return null;\n\n if (expected === NO_CONCURRENCY_CHECK) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_DOES_NOT_EXIST) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_EXISTS) return null;\n\n return expected as bigint;\n};\n\nconst appendToStreamRaw = async (\n connection: SQLiteConnection,\n streamId: string,\n streamType: string,\n messages: RecordedMessage[],\n options?: {\n expectedStreamVersion: bigint | null;\n partition?: string;\n },\n): Promise<AppendEventResult> => {\n let streamPosition;\n let globalPosition;\n\n try {\n let expectedStreamVersion = options?.expectedStreamVersion ?? null;\n\n if (expectedStreamVersion == null) {\n expectedStreamVersion = await getLastStreamPosition(\n connection,\n streamId,\n expectedStreamVersion,\n );\n }\n\n let position: { stream_position: string } | null;\n\n if (expectedStreamVersion === 0n) {\n position = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `INSERT INTO ${streamsTable.name}\n (stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)\n VALUES (\n ?,\n ?,\n ?,\n ?,\n '[]',\n false\n )\n RETURNING stream_position;\n `,\n [\n streamId,\n messages.length,\n options?.partition ?? streamsTable.columns.partition,\n streamType,\n ],\n );\n } else {\n position = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `UPDATE ${streamsTable.name}\n SET stream_position = stream_position + ?\n WHERE stream_id = ?\n AND partition = ?\n AND is_archived = false\n RETURNING stream_position;\n `,\n [\n messages.length,\n streamId,\n options?.partition ?? streamsTable.columns.partition,\n ],\n );\n }\n\n if (position == null) {\n throw new Error('Could not find stream position');\n }\n\n streamPosition = BigInt(position.stream_position);\n\n if (expectedStreamVersion != null) {\n const expectedStreamPositionAfterSave =\n BigInt(expectedStreamVersion) + BigInt(messages.length);\n if (streamPosition !== expectedStreamPositionAfterSave) {\n return {\n success: false,\n };\n }\n }\n\n const { sqlString, values } = buildMessageInsertQuery(\n messages,\n expectedStreamVersion,\n streamId,\n options?.partition?.toString() ?? defaultTag,\n );\n\n const returningIds = await connection.query<{\n global_position: string;\n } | null>(sqlString, values);\n\n if (\n returningIds.length === 0 ||\n !returningIds[returningIds.length - 1]?.global_position\n ) {\n throw new Error('Could not find global position');\n }\n\n globalPosition = BigInt(\n returningIds[returningIds.length - 1]!.global_position,\n );\n } catch (err: unknown) {\n if (isSQLiteError(err) && isOptimisticConcurrencyError(err)) {\n return {\n success: false,\n };\n }\n\n throw err;\n }\n\n return {\n success: true,\n nextStreamPosition: streamPosition,\n lastGlobalPosition: globalPosition,\n };\n};\n\nconst isOptimisticConcurrencyError = (error: SQLiteError): boolean => {\n return error?.errno !== undefined && error.errno === 19;\n};\n\nasync function getLastStreamPosition(\n connection: SQLiteConnection,\n streamId: string,\n expectedStreamVersion: bigint | null,\n): Promise<bigint> {\n const result = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `SELECT CAST(stream_position AS VARCHAR) AS stream_position FROM ${streamsTable.name} WHERE stream_id = ?`,\n [streamId],\n );\n\n if (result?.stream_position == null) {\n expectedStreamVersion = 0n;\n } else {\n expectedStreamVersion = BigInt(result.stream_position);\n }\n return expectedStreamVersion;\n}\n\nconst buildMessageInsertQuery = (\n messages: RecordedMessage[],\n expectedStreamVersion: bigint,\n streamId: string,\n partition: string | null | undefined,\n): {\n sqlString: string;\n values: Parameters[];\n} => {\n const query = messages.reduce(\n (\n queryBuilder: { parameterMarkers: string[]; values: Parameters[] },\n message: RecordedMessage,\n ) => {\n if (\n message.metadata?.streamPosition == null ||\n typeof message.metadata.streamPosition !== 'bigint'\n ) {\n throw new Error('Stream position is required');\n }\n\n const streamPosition =\n BigInt(message.metadata.streamPosition) + BigInt(expectedStreamVersion);\n\n queryBuilder.parameterMarkers.push(`(?,?,?,?,?,?,?,?,?,?)`);\n queryBuilder.values.push(\n streamId,\n streamPosition.toString() ?? 0,\n partition ?? defaultTag,\n message.kind === 'Event' ? 'E' : 'C',\n JSONParser.stringify(message.data),\n JSONParser.stringify(message.metadata),\n expectedStreamVersion?.toString() ?? 0,\n message.type,\n message.metadata.messageId,\n false,\n );\n\n return queryBuilder;\n },\n {\n parameterMarkers: [],\n values: [],\n },\n );\n\n const sqlString = `\n INSERT INTO ${messagesTable.name} (\n stream_id, \n stream_position, \n partition, \n message_kind,\n message_data, \n message_metadata, \n message_schema_version, \n message_type, \n message_id, \n is_archived\n ) \n VALUES ${query.parameterMarkers.join(', ')} \n RETURNING \n CAST(global_position as VARCHAR) AS global_position\n `;\n return { sqlString, values: query.values };\n};\n","export const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = 'emt:default';\n\nexport const globalNames = {\n module: `${emmettPrefix}:module:${globalTag}`,\n};\n\nconst columns = {\n partition: {\n name: 'partition',\n },\n isArchived: { name: 'is_archived' },\n};\n\nexport const streamsTable = {\n name: `${emmettPrefix}_streams`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const messagesTable = {\n name: `${emmettPrefix}_messages`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const subscriptionsTable = {\n name: `${emmettPrefix}_subscriptions`,\n};\n","import type { SQLiteConnection } from '../../connection';\nimport {\n globalTag,\n messagesTable,\n streamsTable,\n subscriptionsTable,\n} from './typing';\n\nexport const sql = (sql: string) => sql;\n\nexport const streamsTableSQL = sql(\n `CREATE TABLE IF NOT EXISTS ${streamsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL DEFAULT 0,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n stream_type TEXT NOT NULL,\n stream_metadata JSONB NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n PRIMARY KEY (stream_id, partition, is_archived),\n UNIQUE (stream_id, partition, is_archived)\n );`,\n);\n\nexport const messagesTableSQL = sql(\n `CREATE TABLE IF NOT EXISTS ${messagesTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n message_kind CHAR(1) NOT NULL DEFAULT 'E',\n message_data JSONB NOT NULL,\n message_metadata JSONB NOT NULL,\n message_schema_version TEXT NOT NULL,\n message_type TEXT NOT NULL,\n message_id TEXT NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n global_position INTEGER PRIMARY KEY,\n created DATETIME DEFAULT CURRENT_TIMESTAMP,\n UNIQUE (stream_id, stream_position, partition, is_archived)\n ); \n`,\n);\n\nexport const subscriptionsTableSQL = sql(\n `\n CREATE TABLE IF NOT EXISTS ${subscriptionsTable.name}(\n subscription_id TEXT NOT NULL,\n version INTEGER NOT NULL DEFAULT 1,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n last_processed_position BIGINT NOT NULL,\n PRIMARY KEY (subscription_id, partition, version)\n );\n`,\n);\n\nexport const schemaSQL: string[] = [\n streamsTableSQL,\n messagesTableSQL,\n subscriptionsTableSQL,\n];\n\nexport const createEventStoreSchema = async (\n db: SQLiteConnection,\n): Promise<void> => {\n for (const sql of schemaSQL) {\n await db.command(sql);\n }\n};\n","export const singleOrNull = async <T>(\n getResult: Promise<T[]>,\n): Promise<T | null> => {\n const result = await getResult;\n\n if (result.length > 1) throw new Error('Query had more than one result');\n\n return result.length > 0 ? (result[0] ?? null) : null;\n};\n\nexport const single = async <T>(getResult: Promise<T[]>): Promise<T> => {\n const result = await getResult;\n\n if (result.length === 0) throw new Error(\"Query didn't return any result\");\n\n if (result.length > 1) throw new Error('Query had more than one result');\n\n return result[0]!;\n};\n","import type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, messagesTable } from './typing';\nimport { singleOrNull } from './utils';\n\ntype ReadLastMessageGlobalPositionSqlResult = {\n global_position: string;\n};\n\nexport type ReadLastMessageGlobalPositionResult = {\n currentGlobalPosition: bigint | null;\n};\n\nexport const readLastMessageGlobalPosition = async (\n db: SQLiteConnection,\n options?: { partition?: string },\n): Promise<ReadLastMessageGlobalPositionResult> => {\n const result = await singleOrNull(\n db.query<ReadLastMessageGlobalPositionSqlResult>(\n sql(\n `SELECT global_position\n FROM ${messagesTable.name}\n WHERE partition = ? AND is_archived = FALSE\n ORDER BY global_position\n LIMIT 1`,\n ),\n [options?.partition ?? defaultTag],\n ),\n );\n\n return {\n currentGlobalPosition:\n result !== null ? BigInt(result.global_position) : null,\n };\n};\n","import {\n JSONParser,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ReadEventMetadata,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, messagesTable } from './typing';\n\ntype ReadMessagesBatchSqlResult = {\n stream_position: string;\n stream_id: string;\n message_data: string;\n message_metadata: string;\n message_schema_version: string;\n message_type: string;\n message_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport type ReadMessagesBatchOptions =\n | {\n after: bigint;\n batchSize: number;\n }\n | {\n from: bigint;\n batchSize: number;\n }\n | { to: bigint; batchSize: number }\n | { from: bigint; to: bigint };\n\nexport type ReadMessagesBatchResult<\n EventType extends Event,\n ReadEventMetadataType extends ReadEventMetadata = ReadEventMetadata,\n> = {\n currentGlobalPosition: bigint;\n messages: ReadEvent<EventType, ReadEventMetadataType>[];\n areEventsLeft: boolean;\n};\n\nexport const readMessagesBatch = async <\n MessageType extends Event,\n ReadEventMetadataType extends\n ReadEventMetadataWithGlobalPosition = ReadEventMetadataWithGlobalPosition,\n>(\n db: SQLiteConnection,\n options: ReadMessagesBatchOptions & { partition?: string },\n): Promise<ReadMessagesBatchResult<MessageType, ReadEventMetadataType>> => {\n const from =\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after + 1n\n : 0n;\n const batchSize =\n options && 'batchSize' in options\n ? options.batchSize\n : options.to - options.from;\n\n const fromCondition: string =\n from !== -0n ? `AND global_position >= ${from}` : '';\n\n const toCondition =\n 'to' in options ? `AND global_position <= ${options.to}` : '';\n\n const limitCondition =\n 'batchSize' in options ? `LIMIT ${options.batchSize}` : '';\n\n const events: ReadEvent<MessageType, ReadEventMetadataType>[] = (\n await db.query<ReadMessagesBatchSqlResult>(\n sql(\n `SELECT stream_id, stream_position, global_position, message_data, message_metadata, message_schema_version, message_type, message_id\n FROM ${messagesTable.name}\n WHERE partition = ? AND is_archived = FALSE ${fromCondition} ${toCondition}\n ORDER BY global_position\n ${limitCondition}`,\n ),\n [options?.partition ?? defaultTag],\n )\n ).map((row) => {\n const rawEvent = {\n type: row.message_type,\n data: JSONParser.parse(row.message_data),\n metadata: JSONParser.parse(row.message_metadata),\n } as unknown as MessageType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n messageId: row.message_id,\n streamName: row.stream_id,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n kind: 'Event',\n metadata: metadata as CombinedReadEventMetadata<\n MessageType,\n ReadEventMetadataType\n >,\n };\n });\n\n return events.length > 0\n ? {\n currentGlobalPosition:\n events[events.length - 1]!.metadata.globalPosition,\n messages: events,\n areEventsLeft: events.length === batchSize,\n }\n : {\n currentGlobalPosition:\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after\n : 0n,\n messages: [],\n areEventsLeft: false,\n };\n};\n","import type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, subscriptionsTable } from './typing';\nimport { singleOrNull } from './utils';\n\ntype ReadProcessorCheckpointSqlResult = {\n last_processed_position: string;\n};\n\nexport type ReadProcessorCheckpointResult = {\n lastProcessedPosition: bigint | null;\n};\n\nexport const readProcessorCheckpoint = async (\n db: SQLiteConnection,\n options: { processorId: string; partition?: string },\n): Promise<ReadProcessorCheckpointResult> => {\n const result = await singleOrNull(\n db.query<ReadProcessorCheckpointSqlResult>(\n sql(\n `SELECT last_processed_position\n FROM ${subscriptionsTable.name}\n WHERE partition = ? AND subscription_id = ?\n LIMIT 1`,\n ),\n [options?.partition ?? defaultTag, options.processorId],\n ),\n );\n\n return {\n lastProcessedPosition:\n result !== null ? BigInt(result.last_processed_position) : null,\n };\n};\n","import type {\n EmmettError,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../../connection';\nimport { readLastMessageGlobalPosition } from '../../schema/readLastMessageGlobalPosition';\nimport {\n readMessagesBatch,\n type ReadMessagesBatchOptions,\n} from '../../schema/readMessagesBatch';\n\nexport const DefaultSQLiteEventStoreProcessorBatchSize = 100;\nexport const DefaultSQLiteEventStoreProcessorPullingFrequencyInMs = 50;\n\nexport type SQLiteEventStoreMessagesBatch<EventType extends Event = Event> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type SQLiteEventStoreMessagesBatchHandlerResult = void | {\n type: 'STOP';\n reason?: string;\n error?: EmmettError;\n};\n\nexport type SQLiteEventStoreMessagesBatchHandler<\n EventType extends Event = Event,\n> = (\n messagesBatch: SQLiteEventStoreMessagesBatch<EventType>,\n) =>\n | Promise<SQLiteEventStoreMessagesBatchHandlerResult>\n | SQLiteEventStoreMessagesBatchHandlerResult;\n\nexport type SQLiteEventStoreMessageBatchPullerOptions<\n EventType extends Event = Event,\n> = {\n connection: SQLiteConnection;\n pullingFrequencyInMs: number;\n batchSize: number;\n eachBatch: SQLiteEventStoreMessagesBatchHandler<EventType>;\n};\n\nexport type SQLiteEventStoreMessageBatchPullerStartFrom =\n | { globalPosition: bigint }\n | 'BEGINNING'\n | 'END';\n\nexport type SQLiteEventStoreMessageBatchPullerStartOptions = {\n startFrom: SQLiteEventStoreMessageBatchPullerStartFrom;\n};\n\nexport type SQLiteEventStoreMessageBatchPuller = {\n isRunning: boolean;\n start(options: SQLiteEventStoreMessageBatchPullerStartOptions): Promise<void>;\n stop(): Promise<void>;\n};\n\nexport const sqliteEventStoreMessageBatchPuller = <\n EventType extends Event = Event,\n>({\n connection,\n batchSize,\n eachBatch,\n pullingFrequencyInMs,\n}: SQLiteEventStoreMessageBatchPullerOptions<EventType>): SQLiteEventStoreMessageBatchPuller => {\n let isRunning = false;\n\n let start: Promise<void>;\n\n const pullMessages = async (\n options: SQLiteEventStoreMessageBatchPullerStartOptions,\n ) => {\n const after =\n options.startFrom === 'BEGINNING'\n ? 0n\n : options.startFrom === 'END'\n ? ((await readLastMessageGlobalPosition(connection))\n .currentGlobalPosition ?? 0n)\n : options.startFrom.globalPosition;\n\n const readMessagesOptions: ReadMessagesBatchOptions = {\n after,\n batchSize,\n };\n\n let waitTime = 100;\n\n do {\n const { messages, currentGlobalPosition, areEventsLeft } =\n await readMessagesBatch<EventType>(connection, readMessagesOptions);\n\n if (messages.length > 0) {\n const result = await eachBatch({ messages });\n\n if (result && result.type === 'STOP') {\n isRunning = false;\n break;\n }\n }\n\n readMessagesOptions.after = currentGlobalPosition;\n\n await new Promise((resolve) => setTimeout(resolve, waitTime));\n\n if (!areEventsLeft) {\n waitTime = Math.min(waitTime * 2, 1000);\n } else {\n waitTime = pullingFrequencyInMs;\n }\n } while (isRunning);\n };\n\n return {\n get isRunning() {\n return isRunning;\n },\n start: (options) => {\n if (isRunning) return start;\n\n start = (async () => {\n isRunning = true;\n\n return pullMessages(options);\n })();\n\n return start;\n },\n stop: async () => {\n if (!isRunning) return;\n isRunning = false;\n await start;\n },\n };\n};\n\nexport const zipSQLiteEventStoreMessageBatchPullerStartFrom = (\n options: (SQLiteEventStoreMessageBatchPullerStartFrom | undefined)[],\n): SQLiteEventStoreMessageBatchPullerStartFrom => {\n if (\n options.length === 0 ||\n options.some((o) => o === undefined || o === 'BEGINNING')\n )\n return 'BEGINNING';\n\n if (options.every((o) => o === 'END')) return 'END';\n\n return options\n .filter((o) => o !== undefined && o !== 'BEGINNING' && o !== 'END')\n .sort((a, b) => (a > b ? 1 : -1))[0]!;\n};\n","import {\n EmmettError,\n getCheckpoint,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { sqliteConnection, type SQLiteConnection } from '../../connection';\nimport type { SQLiteProjectionDefinition } from '../projections';\nimport { readProcessorCheckpoint, storeProcessorCheckpoint } from '../schema';\nimport type { SQLiteEventStoreMessageBatchPullerStartFrom } from './messageBatchProcessing';\n\nexport type SQLiteProcessorEventsBatch<EventType extends Event = Event> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type SQLiteProcessorHandlerContext = {\n connection: SQLiteConnection;\n fileName: string;\n};\n\nexport type SQLiteProcessor<EventType extends Event = Event> = {\n id: string;\n start: (\n connection: SQLiteConnection,\n ) => Promise<SQLiteEventStoreMessageBatchPullerStartFrom | undefined>;\n isActive: boolean;\n handle: (\n messagesBatch: SQLiteProcessorEventsBatch<EventType>,\n context: { connection?: SQLiteConnection; fileName?: string },\n ) => Promise<SQLiteProcessorMessageHandlerResult>;\n};\n\nexport const SQLiteProcessor = {\n result: {\n skip: (options?: {\n reason?: string;\n }): SQLiteProcessorMessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): SQLiteProcessorMessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type SQLiteProcessorMessageHandlerResult =\n | void\n | { type: 'SKIP'; reason?: string }\n | { type: 'STOP'; reason?: string; error?: EmmettError };\n\nexport type SQLiteProcessorEachMessageHandler<EventType extends Event = Event> =\n (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n context: SQLiteProcessorHandlerContext,\n ) =>\n | Promise<SQLiteProcessorMessageHandlerResult>\n | SQLiteProcessorMessageHandlerResult;\n\nexport type SQLiteProcessorEachBatchHandler<EventType extends Event = Event> = (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[],\n context: SQLiteProcessorHandlerContext,\n) =>\n | Promise<SQLiteProcessorMessageHandlerResult>\n | SQLiteProcessorMessageHandlerResult;\n\nexport type SQLiteProcessorStartFrom =\n | SQLiteEventStoreMessageBatchPullerStartFrom\n | 'CURRENT';\n\nexport type SQLiteProcessorConnectionOptions = {\n fileName: string;\n connection?: SQLiteConnection;\n};\n\nexport type GenericSQLiteProcessorOptions<EventType extends Event = Event> = {\n processorId: string;\n version?: number;\n partition?: string;\n startFrom?: SQLiteProcessorStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n eachMessage: SQLiteProcessorEachMessageHandler<EventType>;\n connectionOptions?: SQLiteProcessorConnectionOptions;\n // TODO: Add eachBatch\n};\n\nexport type SQLiteProjectionProcessorOptions<EventType extends Event = Event> =\n {\n processorId?: string;\n version?: number;\n projection: SQLiteProjectionDefinition<EventType>;\n partition?: string;\n startFrom?: SQLiteProcessorStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n };\n\nexport type SQLiteProcessorOptions<EventType extends Event = Event> =\n | GenericSQLiteProcessorOptions<EventType>\n | SQLiteProjectionProcessorOptions<EventType>;\n\nconst genericSQLiteProcessor = <EventType extends Event = Event>(\n options: GenericSQLiteProcessorOptions<EventType>,\n): SQLiteProcessor => {\n const { eachMessage } = options;\n let isActive = true;\n //let lastProcessedPosition: number | null = null;\n\n const getDb = (context: {\n connection?: SQLiteConnection;\n fileName?: string;\n }): { connection: SQLiteConnection; fileName: string } => {\n const fileName = context.fileName ?? options.connectionOptions?.fileName;\n if (!fileName)\n throw new EmmettError(\n `SQLite processor '${options.processorId}' is missing file name. Ensure that you passed it through options`,\n );\n\n const connection =\n context.connection ??\n options.connectionOptions?.connection ??\n sqliteConnection({ fileName });\n\n return { connection, fileName };\n };\n\n return {\n id: options.processorId,\n start: async (\n connection: SQLiteConnection,\n ): Promise<SQLiteEventStoreMessageBatchPullerStartFrom | undefined> => {\n isActive = true;\n if (options.startFrom !== 'CURRENT') return options.startFrom;\n\n const { lastProcessedPosition } = await readProcessorCheckpoint(\n connection,\n {\n processorId: options.processorId,\n partition: options.partition,\n },\n );\n\n if (lastProcessedPosition === null) return 'BEGINNING';\n\n return { globalPosition: lastProcessedPosition };\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n { messages },\n context,\n ): Promise<SQLiteProcessorMessageHandlerResult> => {\n if (!isActive) return;\n\n const { connection, fileName } = getDb(context);\n\n return connection.withTransaction(async () => {\n let result: SQLiteProcessorMessageHandlerResult | undefined = undefined;\n\n let lastProcessedPosition: bigint | null = null;\n\n for (const message of messages) {\n const typedMessage = message as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >;\n\n const messageProcessingResult = await eachMessage(typedMessage, {\n connection,\n fileName,\n });\n\n const newPosition: bigint | null = getCheckpoint(typedMessage);\n\n // TODO: Add correct handling of the storing checkpoint\n await storeProcessorCheckpoint(connection, {\n processorId: options.processorId,\n version: options.version,\n lastProcessedPosition,\n newPosition,\n partition: options.partition,\n });\n\n lastProcessedPosition = typedMessage.metadata.globalPosition;\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(typedMessage)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n return result;\n });\n },\n };\n};\n\nexport const sqliteProjectionProcessor = <EventType extends Event = Event>(\n options: SQLiteProjectionProcessorOptions<EventType>,\n): SQLiteProcessor => {\n const projection = options.projection;\n\n return genericSQLiteProcessor<EventType>({\n processorId: options.processorId ?? `projection:${projection.name}`,\n eachMessage: async (event, context) => {\n if (!projection.canHandle.includes(event.type)) return;\n\n await projection.handle([event], { connection: context.connection });\n },\n ...options,\n });\n};\n\nexport const sqliteProcessor = <EventType extends Event = Event>(\n options: SQLiteProcessorOptions<EventType>,\n): SQLiteProcessor => {\n if ('projection' in options) {\n return sqliteProjectionProcessor(options);\n }\n\n return genericSQLiteProcessor(options);\n};\n","import { EmmettError, type Event } from '@event-driven-io/emmett';\nimport { sqliteConnection, type SQLiteConnection } from '../../connection';\nimport {\n DefaultSQLiteEventStoreProcessorBatchSize,\n DefaultSQLiteEventStoreProcessorPullingFrequencyInMs,\n sqliteEventStoreMessageBatchPuller,\n zipSQLiteEventStoreMessageBatchPullerStartFrom,\n type SQLiteEventStoreMessageBatchPuller,\n type SQLiteEventStoreMessagesBatchHandler,\n} from './messageBatchProcessing';\nimport {\n sqliteProcessor,\n type SQLiteProcessor,\n type SQLiteProcessorOptions,\n} from './sqliteProcessor';\n\nexport type SQLiteEventStoreConsumerConfig<\n ConsumerEventType extends Event = Event,\n> = {\n processors?: SQLiteProcessor<ConsumerEventType>[];\n pulling?: {\n batchSize?: number;\n pullingFrequencyInMs?: number;\n };\n};\nexport type SQLiteEventStoreConsumerOptions<\n ConsumerEventType extends Event = Event,\n> = SQLiteEventStoreConsumerConfig<ConsumerEventType> & {\n fileName: string;\n connection?: SQLiteConnection;\n};\n\nexport type SQLiteEventStoreConsumer<ConsumerEventType extends Event = Event> =\n Readonly<{\n isRunning: boolean;\n processors: SQLiteProcessor<ConsumerEventType>[];\n processor: <EventType extends ConsumerEventType = ConsumerEventType>(\n options: SQLiteProcessorOptions<EventType>,\n ) => SQLiteProcessor<EventType>;\n start: () => Promise<void>;\n stop: () => Promise<void>;\n close: () => Promise<void>;\n }>;\n\nexport const sqliteEventStoreConsumer = <\n ConsumerEventType extends Event = Event,\n>(\n options: SQLiteEventStoreConsumerOptions<ConsumerEventType>,\n): SQLiteEventStoreConsumer<ConsumerEventType> => {\n let isRunning = false;\n const { pulling } = options;\n const processors = options.processors ?? [];\n\n let start: Promise<void>;\n\n let currentMessagePuller: SQLiteEventStoreMessageBatchPuller | undefined;\n\n const connection =\n options.connection ?? sqliteConnection({ fileName: options.fileName });\n\n const eachBatch: SQLiteEventStoreMessagesBatchHandler<\n ConsumerEventType\n > = async (messagesBatch) => {\n const activeProcessors = processors.filter((s) => s.isActive);\n\n if (activeProcessors.length === 0)\n return {\n type: 'STOP',\n reason: 'No active processors',\n };\n\n const result = await Promise.allSettled(\n activeProcessors.map((s) => {\n // TODO: Add here filtering to only pass messages that can be handled by processor\n return s.handle(messagesBatch, {\n connection,\n fileName: options.fileName,\n });\n }),\n );\n\n return result.some(\n (r) => r.status === 'fulfilled' && r.value?.type !== 'STOP',\n )\n ? undefined\n : {\n type: 'STOP',\n };\n };\n\n const messagePooler = (currentMessagePuller =\n sqliteEventStoreMessageBatchPuller({\n connection,\n eachBatch,\n batchSize:\n pulling?.batchSize ?? DefaultSQLiteEventStoreProcessorBatchSize,\n pullingFrequencyInMs:\n pulling?.pullingFrequencyInMs ??\n DefaultSQLiteEventStoreProcessorPullingFrequencyInMs,\n }));\n\n const stop = async () => {\n if (!isRunning) return;\n isRunning = false;\n if (currentMessagePuller) {\n await currentMessagePuller.stop();\n currentMessagePuller = undefined;\n }\n await start;\n };\n\n return {\n processors,\n get isRunning() {\n return isRunning;\n },\n processor: <EventType extends ConsumerEventType = ConsumerEventType>(\n options: SQLiteProcessorOptions<EventType>,\n ): SQLiteProcessor<EventType> => {\n const processor = sqliteProcessor<EventType>(options);\n\n processors.push(processor);\n\n return processor;\n },\n start: () => {\n if (isRunning) return start;\n\n start = (async () => {\n if (processors.length === 0)\n return Promise.reject(\n new EmmettError(\n 'Cannot start consumer without at least a single processor',\n ),\n );\n\n isRunning = true;\n\n const startFrom = zipSQLiteEventStoreMessageBatchPullerStartFrom(\n await Promise.all(processors.map((o) => o.start(connection))),\n );\n\n return messagePooler.start({ startFrom });\n })();\n\n return start;\n },\n stop,\n close: async () => {\n await stop();\n\n connection.close();\n\n await new Promise((resolve) => setTimeout(resolve, 250));\n },\n };\n};\n","import {\n projection,\n type CanHandle,\n type Event,\n type ProjectionDefinition,\n type ProjectionHandler,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../connection';\nimport type { SQLiteReadEventMetadata } from '../SQLiteEventStore';\n\nexport type SQLiteProjectionHandlerContext = {\n connection: SQLiteConnection;\n};\n\nexport type SQLiteProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends SQLiteReadEventMetadata = SQLiteReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n SQLiteProjectionHandlerContext\n>;\n\nexport type SQLiteProjectionDefinition<EventType extends Event = Event> =\n ProjectionDefinition<\n EventType,\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >;\n\nexport type SQLiteProjectionHandlerOptions<EventType extends Event = Event> = {\n events: ReadEvent<EventType, SQLiteReadEventMetadata>[];\n projections: SQLiteProjectionDefinition<EventType>[];\n connection: SQLiteConnection;\n};\n\nexport const handleProjections = async <EventType extends Event = Event>(\n options: SQLiteProjectionHandlerOptions<EventType>,\n): Promise<void> => {\n const { projections: allProjections, events, connection } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n for (const projection of allProjections) {\n if (!projection.canHandle.some((type) => eventTypes.includes(type))) {\n continue;\n }\n await projection.handle(events, {\n connection,\n });\n }\n};\n\nexport const sqliteProjection = <EventType extends Event>(\n definition: SQLiteProjectionDefinition<EventType>,\n): SQLiteProjectionDefinition<EventType> =>\n projection<\n EventType,\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >(definition);\n\nexport const sqliteRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: SQLiteProjectionHandlerContext,\n ) => Promise<string[]> | string[],\n ...canHandle: CanHandle<EventType>\n): SQLiteProjectionDefinition<EventType> =>\n sqliteProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: string[] = await handle(events, context);\n\n for (const sql of sqls) await context.connection.command(sql);\n },\n });\n\nexport const sqliteRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: SQLiteProjectionHandlerContext,\n ) => Promise<string> | string,\n ...canHandle: CanHandle<EventType>\n): SQLiteProjectionDefinition<EventType> =>\n sqliteRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: string[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n","import type {\n AppendToStreamResultWithGlobalPosition,\n BeforeEventStoreCommitHandler,\n BigIntStreamPosition,\n Event,\n ProjectionRegistration,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\n\nimport {\n assertExpectedVersionMatchesCurrent,\n ExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport {\n InMemorySharedCacheSQLiteDatabase,\n InMemorySQLiteDatabase,\n sqliteConnection,\n type SQLiteConnection,\n} from '../connection';\nimport {\n sqliteEventStoreConsumer,\n type SQLiteEventStoreConsumer,\n type SQLiteEventStoreConsumerConfig,\n} from './consumers';\nimport {\n handleProjections,\n type SQLiteProjectionHandlerContext,\n} from './projections';\nimport { createEventStoreSchema } from './schema';\nimport { appendToStream } from './schema/appendToStream';\nimport { readStream } from './schema/readStream';\n\nexport type EventHandler<E extends Event = Event> = (\n eventEnvelope: ReadEvent<E>,\n) => void;\n\nexport const SQLiteEventStoreDefaultStreamVersion = 0n;\n\nexport interface SQLiteEventStore extends EventStore<SQLiteReadEventMetadata> {\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition>;\n consumer<ConsumerEventType extends Event = Event>(\n options?: SQLiteEventStoreConsumerConfig<ConsumerEventType>,\n ): SQLiteEventStoreConsumer<ConsumerEventType>;\n}\n\nexport type SQLiteReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type SQLiteReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n SQLiteReadEventMetadata\n>;\n\nexport type SQLiteEventStoreOptions = {\n fileName: // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n | InMemorySQLiteDatabase\n // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n | InMemorySharedCacheSQLiteDatabase\n | string\n | undefined;\n projections?: ProjectionRegistration<\n 'inline',\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >[];\n schema?: {\n autoMigration?: 'None' | 'CreateOrUpdate';\n };\n hooks?: {\n /**\n * This hook will be called **BEFORE** events were stored in the event store.\n * @type {BeforeEventStoreCommitHandler<SQLiteEventStore, HandlerContext>}\n */\n onBeforeCommit?: BeforeEventStoreCommitHandler<\n SQLiteEventStore,\n { connection: SQLiteConnection }\n >;\n };\n};\n\nexport const getSQLiteEventStore = (\n options: SQLiteEventStoreOptions,\n): SQLiteEventStore => {\n let schemaMigrated = false;\n let autoGenerateSchema = false;\n let database: SQLiteConnection | null;\n const fileName = options.fileName ?? InMemorySQLiteDatabase;\n\n const isInMemory: boolean =\n fileName === InMemorySQLiteDatabase ||\n fileName === InMemorySharedCacheSQLiteDatabase;\n\n const inlineProjections = (options.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n const onBeforeCommitHook = options.hooks?.onBeforeCommit;\n\n const createConnection = () => {\n if (database != null) {\n return database;\n }\n\n return sqliteConnection({\n fileName,\n });\n };\n\n const closeConnection = () => {\n if (isInMemory) {\n return;\n }\n if (database != null) {\n database.close();\n database = null;\n }\n };\n\n const withConnection = async <Result>(\n handler: (connection: SQLiteConnection) => Promise<Result>,\n ): Promise<Result> => {\n if (database == null) {\n database = createConnection();\n }\n\n try {\n await ensureSchemaExists(database);\n return await handler(database);\n } finally {\n closeConnection();\n }\n };\n\n if (options) {\n autoGenerateSchema =\n options.schema?.autoMigration === undefined ||\n options.schema?.autoMigration !== 'None';\n }\n\n const ensureSchemaExists = async (\n connection: SQLiteConnection,\n ): Promise<void> => {\n if (!autoGenerateSchema) return Promise.resolve();\n\n if (!schemaMigrated) {\n await createEventStoreSchema(connection);\n schemaMigrated = true;\n }\n\n return Promise.resolve();\n };\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const expectedStreamVersion = read?.expectedStreamVersion;\n\n let state = initialState();\n\n if (typeof streamName !== 'string') {\n throw new Error('Stream name is not string');\n }\n\n if (database == null) {\n database = createConnection();\n }\n\n const result = await withConnection((connection) =>\n readStream<EventType>(connection, streamName, options.read),\n );\n\n const currentStreamVersion = result.currentStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n SQLiteEventStoreDefaultStreamVersion,\n );\n\n for (const event of result.events) {\n if (!event) continue;\n\n state = evolve(state, event);\n }\n\n return {\n currentStreamVersion: currentStreamVersion,\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: async <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > =>\n withConnection((connection) =>\n readStream<EventType>(connection, streamName, options),\n ),\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition> => {\n if (database == null) {\n database = createConnection();\n }\n\n // TODO: This has to be smarter when we introduce urn-based resolution\n const [firstPart, ...rest] = streamName.split('-');\n\n const streamType =\n firstPart && rest.length > 0 ? firstPart : 'emt:unknown';\n\n const appendResult = await withConnection((connection) =>\n appendToStream(connection, streamName, streamType, events, {\n ...options,\n onBeforeCommit: async (messages, context) => {\n if (inlineProjections.length > 0)\n await handleProjections({\n projections: inlineProjections,\n events: messages,\n ...context,\n });\n\n if (onBeforeCommitHook) await onBeforeCommitHook(messages, context);\n },\n }),\n );\n\n if (!appendResult.success)\n throw new ExpectedVersionConflictError<bigint>(\n -1n, //TODO: Return actual version in case of error\n options?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n );\n\n return {\n nextExpectedStreamVersion: appendResult.nextStreamPosition,\n lastEventGlobalPosition: appendResult.lastGlobalPosition,\n createdNewStream:\n appendResult.nextStreamPosition >= BigInt(events.length),\n };\n },\n consumer: <ConsumerEventType extends Event = Event>(\n options?: SQLiteEventStoreConsumerConfig<ConsumerEventType>,\n ): SQLiteEventStoreConsumer<ConsumerEventType> =>\n sqliteEventStoreConsumer<ConsumerEventType>({\n ...(options ?? {}),\n fileName,\n connection: database ?? undefined,\n }),\n };\n};\n","import {\n JSONParser,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { type SQLiteConnection } from '../../connection';\nimport { SQLiteEventStoreDefaultStreamVersion } from '../SQLiteEventStore';\nimport { defaultTag, messagesTable } from './typing';\n\ntype ReadStreamSqlResult = {\n stream_position: string;\n message_data: string;\n message_metadata: string;\n message_schema_version: string;\n message_type: string;\n message_id: string;\n global_position: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n db: SQLiteConnection,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const results = await db.query<ReadStreamSqlResult>(\n `SELECT stream_id, stream_position, global_position, message_data, message_metadata, message_schema_version, message_type, message_id\n FROM ${messagesTable.name}\n WHERE stream_id = ? AND partition = ? AND is_archived = FALSE ${fromCondition} ${toCondition}\n ORDER BY stream_position ASC`,\n [streamId, options?.partition ?? defaultTag],\n );\n\n const messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n results.map((row) => {\n const rawEvent = {\n type: row.message_type,\n data: JSONParser.parse(row.message_data),\n metadata: JSONParser.parse(row.message_metadata),\n } as unknown as EventType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n messageId: row.message_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n kind: 'Event',\n metadata: metadata as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n return messages.length > 0\n ? {\n currentStreamVersion:\n messages[messages.length - 1]!.metadata.streamPosition,\n events: messages,\n streamExists: true,\n }\n : {\n currentStreamVersion: SQLiteEventStoreDefaultStreamVersion,\n events: [],\n streamExists: false,\n };\n};\n","import { isSQLiteError, type SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, subscriptionsTable } from './typing';\nimport { singleOrNull } from './utils';\n\n// for more infos see the postgresql stored procedure version\nasync function storeSubscriptionCheckpointSQLite(\n db: SQLiteConnection,\n processorId: string,\n version: number,\n position: bigint | null,\n checkPosition: bigint | null,\n partition: string,\n): Promise<0 | 1 | 2> {\n if (checkPosition !== null) {\n const updateResult = await db.command(\n sql(`\n UPDATE ${subscriptionsTable.name}\n SET last_processed_position = ?\n WHERE subscription_id = ? \n AND last_processed_position = ? \n AND partition = ?\n `),\n [position!.toString(), processorId, checkPosition.toString(), partition],\n );\n if (updateResult.changes > 0) {\n return 1;\n } else {\n const current_position = await singleOrNull(\n db.query<{ last_processed_position: bigint }>(\n sql(\n `SELECT last_processed_position FROM ${subscriptionsTable.name} \n WHERE subscription_id = ? AND partition = ?`,\n ),\n [processorId, partition],\n ),\n );\n\n if (current_position?.last_processed_position === position) {\n return 0;\n } else if (\n position !== null &&\n current_position !== null &&\n current_position?.last_processed_position > position\n ) {\n return 2;\n } else {\n return 2;\n }\n }\n } else {\n try {\n await db.command(\n sql(\n `INSERT INTO ${subscriptionsTable.name} (subscription_id, version, last_processed_position, partition) VALUES (?, ?, ?, ?)`,\n ),\n [processorId, version, position!.toString(), partition],\n );\n return 1;\n } catch (err) {\n if (!(isSQLiteError(err) && (err.errno === 19 || err.errno === 2067))) {\n throw err;\n }\n\n const current = await singleOrNull(\n db.query<{ last_processed_position: bigint }>(\n sql(\n `SELECT last_processed_position FROM ${subscriptionsTable.name} WHERE subscription_id = ? AND partition = ?`,\n ),\n [processorId, partition],\n ),\n );\n if (current?.last_processed_position === position) {\n return 0;\n } else {\n return 2;\n }\n }\n }\n}\n\nexport type StoreLastProcessedProcessorPositionResult<\n Position extends bigint | null = bigint,\n> =\n | {\n success: true;\n newPosition: Position;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport async function storeProcessorCheckpoint(\n db: SQLiteConnection,\n options: {\n processorId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedProcessorPositionResult<bigint | null>> {\n try {\n const result = await storeSubscriptionCheckpointSQLite(\n db,\n options.processorId,\n options.version ?? 1,\n options.newPosition,\n options.lastProcessedPosition,\n options.partition ?? defaultTag,\n );\n\n return result === 1\n ? { success: true, newPosition: options.newPosition }\n : { success: false, reason: result === 0 ? 'IGNORED' : 'MISMATCH' };\n } catch (error) {\n console.log(error);\n throw error;\n }\n}\n"],"mappings":";AAAA,OAAO,aAAa;AAgBb,IAAM,gBAAgB,CAAC,UAAyC;AACrE,MAAI,iBAAiB,SAAS,UAAU,OAAO;AAC7C,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAGO,IAAM,oCAAoC;AAE1C,IAAM,yBAAyB;AAO/B,IAAM,mBAAmB,CAC9B,YACqB;AACrB,QAAM,WAAW,QAAQ,YAAY;AACrC,MAAI;AAEJ,MAAI,SAAS,WAAW,OAAO,GAAG;AAChC,SAAK,IAAI,QAAQ;AAAA,MACf;AAAA,MACA,QAAQ,WAAW,QAAQ,iBAAiB,QAAQ;AAAA,IACtD;AAAA,EACF,OAAO;AACL,SAAK,IAAI,QAAQ,SAAS,QAAQ;AAAA,EACpC;AACA,KAAG,IAAI,4BAA4B;AACnC,MAAI,qBAAqB;AAEzB,SAAO;AAAA,IACL,OAAO,MAAY,GAAG,MAAM;AAAA,IAC5B,SAAS,CAACA,MAAa,WACrB,IAAI,QAA2B,CAAC,SAAS,WAAW;AAClD,SAAG;AAAA,QACDA;AAAA,QACA,UAAU,CAAC;AAAA,QACX,SAAmC,KAAmB;AACpD,cAAI,KAAK;AACP,mBAAO,GAAG;AACV;AAAA,UACF;AAEA,kBAAQ,IAAI;AAAA,QACd;AAAA,MACF;AAAA,IACF,CAAC;AAAA,IACH,OAAO,CAAIA,MAAa,WACtB,IAAI,QAAQ,CAAC,SAAS,WAAW;AAC/B,SAAG,IAAIA,MAAK,UAAU,CAAC,GAAG,CAAC,KAAmB,WAAgB;AAC5D,YAAI,KAAK;AACP,iBAAO,GAAG;AACV;AAAA,QACF;AAEA,gBAAQ,MAAM;AAAA,MAChB,CAAC;AAAA,IACH,CAAC;AAAA,IACH,aAAa,CAAIA,MAAa,WAC5B,IAAI,QAAQ,CAAC,SAAS,WAAW;AAC/B,SAAG,IAAIA,MAAK,UAAU,CAAC,GAAG,CAAC,KAAmB,WAAqB;AACjE,YAAI,KAAK;AACP,iBAAO,GAAG;AACV;AAAA,QACF;AAEA,gBAAQ,MAAM;AAAA,MAChB,CAAC;AAAA,IACH,CAAC;AAAA,IACH,iBAAiB,OAAU,OAAyB;AAClD,UAAI;AACF,YAAI,wBAAwB,GAAG;AAC7B,gBAAM,iBAAiB,EAAE;AAAA,QAC3B;AACA,cAAM,SAAS,MAAM,GAAG;AAExB,YAAI,uBAAuB,EAAG,OAAM,kBAAkB,EAAE;AACxD;AAEA,eAAO;AAAA,MACT,SAAS,KAAK;AACZ,gBAAQ,IAAI,GAAG;AAEf,YAAI,EAAE,uBAAuB,EAAG,OAAM,oBAAoB,EAAE;AAE5D,cAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;AAEA,IAAM,mBAAmB,CAAC,OACxB,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,+BAA+B,CAAC,QAAsB;AAC3D,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;AAEH,IAAM,oBAAoB,CAAC,OACzB,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,UAAU,CAAC,QAAsB;AACtC,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;AAEH,IAAM,sBAAsB,CAAC,OAC3B,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,YAAY,CAAC,QAAsB;AACxC,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;;;AE1II,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAG9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAE9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ;ACIV,IAAM,cAAN,MAAM,qBAAoB,MAAM;EAC9B;EAEP,YACE,SACA;AACA,UAAM,YACJ,WAAW,OAAO,YAAY,YAAY,eAAe,UACrD,QAAQ,YACR,SAAS,OAAO,IACd,UACA;AACR,UAAM,UACJ,WAAW,OAAO,YAAY,YAAY,aAAa,UACnD,QAAQ,UACR,SAAS,OAAO,IACd,UACA,2BAA2B,SAAS;AAE5C,UAAM,OAAO;AACb,SAAK,YAAY;AAGjB,WAAO,eAAe,MAAM,aAAY,SAAS;EACnD;EAEA,OAAc,QACZ,OACa;AACb,QAAI,iBAAiB,cAAa;AAChC,aAAO;IACT;AAEA,WAAO,IAAI,aAAY;MACrB,WACE,eAAe,SACf,MAAM,cAAc,UACpB,MAAM,cAAc,OAChB,MAAM,YACN;MACN,SAAS,MAAM,WAAW;IAC5B,CAAC;EACH;AACF;AAEO,IAAM,mBAAN,MAAM,0BAAyB,YAAY;EAChD,YACS,SACA,UACP,SACA;AACA,UAAM;MACJ,WAAW;MACX,SACE,WACA,oBAAoB,SAAS,SAAS,CAAC,2BAA2B,SAAS,SAAS,CAAC;IACzF,CAAC;AATM,SAAA,UAAA;AACA,SAAA,WAAA;AAWP,WAAO,eAAe,MAAM,kBAAiB,SAAS;EACxD;AACF;;;AWjFA,SAAS,MAAMC,aAAY;ACA3B,SAAS,MAAM,YAAY;ASA3B,OAAO,WAAW;AIAlB,SAAS,MAAMA,aAAY;AIA3B,SAAS,uBAAuB;ACAhC,SAAS,MAAMA,aAAY;ACA3B,SAAS,MAAMA,aAAY;AAC3B,SAAS,mBAAAC,wBAAuB;AeDhC,SAAS,sBAAsB;AEA/B,OAIO;ACJP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC;EAGE,mBAAAA;OAEK;ACLP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;A/CezB,IAAM,gBAAgB;AACtB,IAAM,wBACX;AACK,IAAM,uBACX;AAEK,IAAM,yBAAyB,CACpC,SACA,UACA,mBACY;AACZ,MAAI,aAAa,qBAAsB,QAAO;AAE9C,MAAI,YAAY,sBAAuB,QAAO,YAAY;AAE1D,MAAI,YAAY,cAAe,QAAO,YAAY;AAElD,SAAO,YAAY;AACrB;AAEO,IAAM,sCAAsC,CAGjD,SACA,UACA,mBACS;AACT,eAAa;AAEb,MAAI,CAAC,uBAAuB,SAAS,UAAU,cAAc;AAC3D,UAAM,IAAI,6BAA6B,SAAS,QAAQ;AAC5D;AAEO,IAAM,+BAAN,MAAM,sCAEH,iBAAiB;EACzB,YACE,SACA,UACA;AACA,UAAM,SAAS,SAAS,GAAG,UAAU,SAAS,CAAC;AAG/C,WAAO,eAAe,MAAM,8BAA6B,SAAS;EACpE;AACF;AY5DO,IAAM,aAAN,cAAyB,MAAM;EACpC,YAAY,MAAc;AACxB,UAAM,iBAAiB,IAAI,EAAE;EAC/B;AACF;AA0BO,IAAM,aAAa;EACxB,WAAW,CACT,OACA,YACG;AACH,WAAO,KAAK;MACV,SAAS,MAAM,QAAQ,IAAI,KAA6B,IAAI;;;MAG5D,CAAC,GAAG,MAAO,OAAO,MAAM,WAAW,EAAE,SAAS,IAAI;IACpD;EACF;EACA,OAAO,CACL,MACA,YACmB;AACnB,UAAM,SAAkB,KAAK,MAAM,MAAM,SAAS,OAAO;AAEzD,QAAI,SAAS,aAAa,CAAC,SAAS,UAAc,MAAM;AACtD,YAAM,IAAI,WAAW,IAAI;AAE3B,WAAO,SAAS,MACZ,QAAQ,IAAI,MAA8B,IACzC;EACP;AACF;AD5CO,IAAM,aAAa,OACxB,IACA,SACe;AACf,MAAI,SAAS,UAAa,KAAK,YAAY,EAAG,QAAO,GAAG;AAExD,SAAO;IACL,OAAO,SAAS;AACd,UAAI;AACF,cAAM,SAAS,MAAM,GAAG;AAExB,YAAI,MAAM,qBAAqB,KAAK,kBAAkB,MAAM,GAAG;AAC7D,gBAAM,IAAI;YACR,+BAA+B,WAAW,UAAU,MAAM,CAAC;UAC7D;QACF;AACA,eAAO;MACT,SAASC,QAAO;AACd,YAAI,MAAM,oBAAoB,CAAC,KAAK,iBAAiBA,MAAK,GAAG;AAC3D,eAAKA,MAAc;AACnB,iBAAO;QACT;AACA,cAAMA;MACR;IACF;IACA,QAAQ,EAAE,SAAS,EAAE;EACvB;AACF;AUnCO,IAAM,mCAAmC,CAC9C,0BAGA,UAA6D,CAAC,MAC3D,IAAI,iCAAiC,0BAA0B,OAAO;AAEpE,IAAM,mCAAN,cAAqDC,iBAG1D;EAQA,YACU,0BAGR,UAA6D,CAAC,GAC9D;AACA,UAAM;MACJ,QAAQ,CAAC,WAAW;AAClB,gBAAQ,IAAI,gCAAgC,MAAM;AAClD,aAAK,aAAa;MACpB;IACF,CAAC;AAVO,SAAA,2BAAA;AAWR,SAAK,WAAW,SAAS,YAAYC,MAAK;AAE1C,SAAK,2BAA2B;AAEhC,SAAK,cAAc,SAAS,qBAAqB,EAAE;EACrD;EAxBQ,gBAAuC;EAC/B;EACR,aAAsB;EAC9B,IAAW,uBAAuB;AAChC,WAAO,CAAC,KAAK;EACf;EAqBQ,cAAc,UAAkB;AACtC,SAAK,gBAAgB,YAAY,MAAM;AACrC,WAAK,oBAAoB;IAC3B,GAAG,QAAQ;EACb;EAEQ,eAAe;AACrB,QAAI,CAAC,KAAK,cAAe;AAEzB,kBAAc,KAAK,aAAa;AAChC,SAAK,gBAAgB;AACrB,SAAK,aAAa;AAClB,SAAK,yBAAyB,IAAI;EACpC;EAEQ,sBAAsB;AAC5B,QAAI,CAAC,KAAK,SAAS,UAAU,CAAC,KAAK,YAAY;AAC7C,WAAK,aAAa;IACpB;EACF;AACF;AK1BO,IAAM,gBAAgB,CAK3BC,aAC0B;AAE1B,SAAO,gBAAgBA,SAAQ;EAE7B,SAASA,SAAQ,SAAS,UAAU;;IAElCA,SAAQ,SAAS;MACjB,oBAAoBA,SAAQ;EAE1B,SAASA,SAAQ,SAAS,cAAc;;IAExCA,SAAQ,SAAS;MACjB,oBAAoBA,SAAQ;EAE1B,SAASA,SAAQ,SAAS,cAAc;;IAExCA,SAAQ,SAAS;MACjB;AACV;AaxDO,IAAM,SAAS,CAAOC,YAC3B,IAAIC,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,QAAID,QAAO,KAAK,GAAG;AACjB,iBAAW,QAAQ,KAAK;IAC1B;EACF;AACF,CAAC;ACPI,IAAM,MAAM,CAAWE,SAC5B,IAAID,iBAA0B;EAC5B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQC,KAAI,KAAK,CAAC;EAC/B;AACF,CAAC;ACLI,IAAM,SAAS,CACpB,SACA,iBACG,IAAI,sBAA4B,SAAS,YAAY;AAEnD,IAAM,wBAAN,cAA0CD,iBAAsB;EAC7D;EACA;EAER,YAAY,SAA0C,cAAiB;AACrE,UAAM;MACJ,WAAW,CAAC,UAAU;AACpB,aAAK,cAAc,KAAK,QAAQ,KAAK,aAAa,KAAK;MACzD;MACA,OAAO,CAAC,eAAe;AACrB,mBAAW,QAAQ,KAAK,WAAW;AACnC,mBAAW,UAAU;MACvB;IACF,CAAC;AAED,SAAK,cAAc;AACnB,SAAK,UAAU;EACjB;AACF;ACjBO,IAAM,cAAc,CAKzB,oBACAE,cAIA,eAAkC,EAAE,SAAS,MAAM,YAAY,GAAG,MAElE,IAAIF,iBAAqC;EACvC,MAAM,YAAY;AAChB;MACE,MAAM,WAAW,oBAAoBE,cAAa,UAAU;MAC5D;IACF,EAAE,MAAM,CAACC,WAAU;AACjB,iBAAW,MAAMA,MAAK;IACxB,CAAC;EACH;AACF,CAAC;AAEH,IAAM,aAAa,OACjB,oBACAD,cAIA,eACkB;AAClB,QAAM,eAAe,mBAAmB;AACxC,QAAM,SAAS,aAAa,UAAU;AAEtC,MAAI;AACF,QAAI;AAEJ,OAAG;AACD,YAAM,SAAS,MAAM,OAAO,KAAK;AACjC,aAAO,OAAO;AAEd,YAAMA,aAAY,QAAQ,UAAU;AAEpC,UAAI,MAAM;AACR,mBAAW,UAAU;MACvB;IACF,SAAS,CAAC;EACZ,UAAA;AACE,WAAO,YAAY;EACrB;AACF;ACxDO,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCF,iBAAsB;EACxD,QAAQ;EACR;EAER,YAAYI,OAAc;AACxB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,aAAK;AACL,YAAI,KAAK,QAAQ,KAAK,MAAM;AAC1B,qBAAW,QAAQ,KAAK;QAC1B;MACF;IACF,CAAC;AAED,SAAK,OAAOA;EACd;AACF;AClBO,IAAM,YAAY,CAAO,kBAC9B,IAAIJ,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;AAExB,QAAI,cAAc,KAAK,GAAG;AACxB,iBAAW,UAAU;IACvB;EACF;AACF,CAAC;ACTI,IAAM,SAAS,CAAO,kBAC3B,IAAIA,iBAA4B;EAC9B,MAAM,UAAU,OAAO,YAAY;AACjC,QAAI,CAAC,cAAc,KAAK,GAAG;AACzB,iBAAW,QAAQ,KAAK;AACxB;IACF;AACA,UAAM,QAAQ,QAAQ;AACtB,eAAW,UAAU;EACvB;AACF,CAAC;ACVI,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCA,kBAAsB;EACxD,QAAQ;EACR;EAER,YAAY,OAAe;AACzB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,YAAI,KAAK,QAAQ,KAAK,OAAO;AAC3B,eAAK;AACL,qBAAW,QAAQ,KAAK;QAC1B,OAAO;AACL,qBAAW,UAAU;QACvB;MACF;IACF,CAAC;AAED,SAAK,QAAQ;EACf;AACF;ACpBO,IAAM,aAAa,CAAO,iBAC/B,IAAIA,kBAA4B;EAC9B,MAAM,YAAY;AAChB,UAAM,YAAY,WAAW,MAAM;AACjC,iBAAW,UAAU;IACvB,GAAG,YAAY;AAEf,UAAM,oBAAoB,WAAW,UAAU,KAAK,UAAU;AAG9D,eAAW,YAAY,MAAM;AAC3B,mBAAa,SAAS;AACtB,wBAAkB;IACpB;EACF;EACA,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;EAC1B;AACF,CAAC;ACNI,IAAM,wBAAwB;EACnC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA,OAAO;EACP;EACA;EACA;AACF;AVnBA,IAAM,EAAE,OAAAK,OAAM,IAAI;;;AWClB,SAAS,MAAMC,aAAY;;;ACXpB,IAAM,eAAe;AAErB,IAAM,YAAY;AAClB,IAAM,aAAa;AAEnB,IAAM,cAAc;AAAA,EACzB,QAAQ,GAAG,YAAY,WAAW,SAAS;AAC7C;AAEA,IAAM,UAAU;AAAA,EACd,WAAW;AAAA,IACT,MAAM;AAAA,EACR;AAAA,EACA,YAAY,EAAE,MAAM,cAAc;AACpC;AAEO,IAAM,eAAe;AAAA,EAC1B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,gBAAgB;AAAA,EAC3B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,qBAAqB;AAAA,EAChC,MAAM,GAAG,YAAY;AACvB;;;ADFO,IAAM,iBAAiB,OAC5B,YACA,YACA,YACA,UACA,YAO+B;AAC/B,MAAI,SAAS,WAAW,EAAG,QAAO,EAAE,SAAS,MAAM;AAEnD,QAAM,wBAAwB;AAAA,IAC5B,SAAS;AAAA,EACX;AAEA,QAAM,mBAGA,SAAS;AAAA,IACb,CACE,GACA,OAEC;AAAA,MACC,GAAG;AAAA,MACH,MAAM,EAAE,QAAQ;AAAA,MAChB,UAAU;AAAA,QACR;AAAA,QACA,WAAWC,MAAK;AAAA,QAChB,gBAAgB,OAAO,IAAI,CAAC;AAAA,QAC5B,GAAI,cAAc,IAAK,EAAE,YAAY,CAAC,IAAK,CAAC;AAAA,MAC9C;AAAA,IACF;AAAA,EACJ;AAEA,MAAI;AAEJ,SAAO,MAAM,WAAW,gBAAgB,YAAY;AAClD,aAAS,MAAM;AAAA,MACb;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,QACE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,SAAS;AACX,YAAM,QAAQ,eAAe,kBAAkB,EAAE,WAAW,CAAC;AAE/D,WAAO;AAAA,EACT,CAAC;AACH;AAEA,IAAM,oBAAoB,CACxB,aACkB;AAClB,MAAI,aAAa,OAAW,QAAO;AAEnC,MAAI,aAAa,qBAAsB,QAAO;AAG9C,MAAI,YAAY,sBAAuB,QAAO;AAG9C,MAAI,YAAY,cAAe,QAAO;AAEtC,SAAO;AACT;AAEA,IAAM,oBAAoB,OACxB,YACA,UACA,YACA,UACA,YAI+B;AAC/B,MAAI;AACJ,MAAI;AAEJ,MAAI;AACF,QAAI,wBAAwB,SAAS,yBAAyB;AAE9D,QAAI,yBAAyB,MAAM;AACjC,8BAAwB,MAAM;AAAA,QAC5B;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAAA,IACF;AAEA,QAAI;AAEJ,QAAI,0BAA0B,IAAI;AAChC,iBAAW,MAAM,WAAW;AAAA,QAG1B,eAAe,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAYhC;AAAA,UACE;AAAA,UACA,SAAS;AAAA,UACT,SAAS,aAAa,aAAa,QAAQ;AAAA,UAC3C;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,iBAAW,MAAM,WAAW;AAAA,QAG1B,UAAU,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAO3B;AAAA,UACE,SAAS;AAAA,UACT;AAAA,UACA,SAAS,aAAa,aAAa,QAAQ;AAAA,QAC7C;AAAA,MACF;AAAA,IACF;AAEA,QAAI,YAAY,MAAM;AACpB,YAAM,IAAI,MAAM,gCAAgC;AAAA,IAClD;AAEA,qBAAiB,OAAO,SAAS,eAAe;AAEhD,QAAI,yBAAyB,MAAM;AACjC,YAAM,kCACJ,OAAO,qBAAqB,IAAI,OAAO,SAAS,MAAM;AACxD,UAAI,mBAAmB,iCAAiC;AACtD,eAAO;AAAA,UACL,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAEA,UAAM,EAAE,WAAW,OAAO,IAAI;AAAA,MAC5B;AAAA,MACA;AAAA,MACA;AAAA,MACA,SAAS,WAAW,SAAS,KAAK;AAAA,IACpC;AAEA,UAAM,eAAe,MAAM,WAAW,MAE5B,WAAW,MAAM;AAE3B,QACE,aAAa,WAAW,KACxB,CAAC,aAAa,aAAa,SAAS,CAAC,GAAG,iBACxC;AACA,YAAM,IAAI,MAAM,gCAAgC;AAAA,IAClD;AAEA,qBAAiB;AAAA,MACf,aAAa,aAAa,SAAS,CAAC,EAAG;AAAA,IACzC;AAAA,EACF,SAAS,KAAc;AACrB,QAAI,cAAc,GAAG,KAAK,6BAA6B,GAAG,GAAG;AAC3D,aAAO;AAAA,QACL,SAAS;AAAA,MACX;AAAA,IACF;AAEA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL,SAAS;AAAA,IACT,oBAAoB;AAAA,IACpB,oBAAoB;AAAA,EACtB;AACF;AAEA,IAAM,+BAA+B,CAAC,UAAgC;AACpE,SAAO,OAAO,UAAU,UAAa,MAAM,UAAU;AACvD;AAEA,eAAe,sBACb,YACA,UACA,uBACiB;AACjB,QAAM,SAAS,MAAM,WAAW;AAAA,IAG9B,mEAAmE,aAAa,IAAI;AAAA,IACpF,CAAC,QAAQ;AAAA,EACX;AAEA,MAAI,QAAQ,mBAAmB,MAAM;AACnC,4BAAwB;AAAA,EAC1B,OAAO;AACL,4BAAwB,OAAO,OAAO,eAAe;AAAA,EACvD;AACA,SAAO;AACT;AAEA,IAAM,0BAA0B,CAC9B,UACA,uBACA,UACA,cAIG;AACH,QAAM,QAAQ,SAAS;AAAA,IACrB,CACE,cACA,YACG;AACH,UACE,QAAQ,UAAU,kBAAkB,QACpC,OAAO,QAAQ,SAAS,mBAAmB,UAC3C;AACA,cAAM,IAAI,MAAM,6BAA6B;AAAA,MAC/C;AAEA,YAAM,iBACJ,OAAO,QAAQ,SAAS,cAAc,IAAI,OAAO,qBAAqB;AAExE,mBAAa,iBAAiB,KAAK,uBAAuB;AAC1D,mBAAa,OAAO;AAAA,QAClB;AAAA,QACA,eAAe,SAAS,KAAK;AAAA,QAC7B,aAAa;AAAA,QACb,QAAQ,SAAS,UAAU,MAAM;AAAA,QACjC,WAAW,UAAU,QAAQ,IAAI;AAAA,QACjC,WAAW,UAAU,QAAQ,QAAQ;AAAA,QACrC,uBAAuB,SAAS,KAAK;AAAA,QACrC,QAAQ;AAAA,QACR,QAAQ,SAAS;AAAA,QACjB;AAAA,MACF;AAEA,aAAO;AAAA,IACT;AAAA,IACA;AAAA,MACE,kBAAkB,CAAC;AAAA,MACnB,QAAQ,CAAC;AAAA,IACX;AAAA,EACF;AAEA,QAAM,YAAY;AAAA,oBACA,cAAc,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,eAYvB,MAAM,iBAAiB,KAAK,IAAI,CAAC;AAAA;AAAA;AAAA;AAI9C,SAAO,EAAE,WAAW,QAAQ,MAAM,OAAO;AAC3C;;;AErTO,IAAM,MAAM,CAACC,SAAgBA;AAE7B,IAAM,kBAAkB;AAAA,EAC7B,8BAA8B,aAAa,IAAI;AAAA;AAAA;AAAA,sEAGqB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAO/E;AAEO,IAAM,mBAAmB;AAAA,EAC9B,8BAA8B,cAAc,IAAI;AAAA;AAAA;AAAA,2EAGyB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAapF;AAEO,IAAM,wBAAwB;AAAA,EACnC;AAAA,+BAC6B,mBAAmB,IAAI;AAAA;AAAA;AAAA,iFAG2B,SAAS;AAAA;AAAA;AAAA;AAAA;AAK1F;AAEO,IAAM,YAAsB;AAAA,EACjC;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,yBAAyB,OACpC,OACkB;AAClB,aAAWA,QAAO,WAAW;AAC3B,UAAM,GAAG,QAAQA,IAAG;AAAA,EACtB;AACF;;;AClEO,IAAM,eAAe,OAC1B,cACsB;AACtB,QAAM,SAAS,MAAM;AAErB,MAAI,OAAO,SAAS,EAAG,OAAM,IAAI,MAAM,gCAAgC;AAEvE,SAAO,OAAO,SAAS,IAAK,OAAO,CAAC,KAAK,OAAQ;AACnD;;;ACKO,IAAM,gCAAgC,OAC3C,IACA,YACiD;AACjD,QAAM,SAAS,MAAM;AAAA,IACnB,GAAG;AAAA,MACD;AAAA,QACE;AAAA,gBACQ,cAAc,IAAI;AAAA;AAAA;AAAA;AAAA,MAI5B;AAAA,MACA,CAAC,SAAS,aAAa,UAAU;AAAA,IACnC;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,eAAe,IAAI;AAAA,EACvD;AACF;;;ACYO,IAAM,oBAAoB,OAK/B,IACA,YACyE;AACzE,QAAM,OACJ,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QAAQ,KAChB;AACR,QAAM,YACJ,WAAW,eAAe,UACtB,QAAQ,YACR,QAAQ,KAAK,QAAQ;AAE3B,QAAM,gBACJ,SAAS,CAAC,KAAK,0BAA0B,IAAI,KAAK;AAEpD,QAAM,cACJ,QAAQ,UAAU,0BAA0B,QAAQ,EAAE,KAAK;AAE7D,QAAM,iBACJ,eAAe,UAAU,SAAS,QAAQ,SAAS,KAAK;AAE1D,QAAM,UACJ,MAAM,GAAG;AAAA,IACP;AAAA,MACE;AAAA,kBACU,cAAc,IAAI;AAAA,yDACqB,aAAa,IAAI,WAAW;AAAA;AAAA,aAExE,cAAc;AAAA,IACrB;AAAA,IACA,CAAC,SAAS,aAAa,UAAU;AAAA,EACnC,GACA,IAAI,CAAC,QAAQ;AACb,UAAM,WAAW;AAAA,MACf,MAAM,IAAI;AAAA,MACV,MAAM,WAAW,MAAM,IAAI,YAAY;AAAA,MACvC,UAAU,WAAW,MAAM,IAAI,gBAAgB;AAAA,IACjD;AAEA,UAAM,WAAgD;AAAA,MACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,MAC1D,WAAW,IAAI;AAAA,MACf,YAAY,IAAI;AAAA,MAChB,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,IAC5C;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH,MAAM;AAAA,MACN;AAAA,IAIF;AAAA,EACF,CAAC;AAED,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,uBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC,UAAU;AAAA,IACV,eAAe,OAAO,WAAW;AAAA,EACnC,IACA;AAAA,IACE,uBACE,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QACR;AAAA,IACR,UAAU,CAAC;AAAA,IACX,eAAe;AAAA,EACjB;AACN;;;AClHO,IAAM,0BAA0B,OACrC,IACA,YAC2C;AAC3C,QAAM,SAAS,MAAM;AAAA,IACnB,GAAG;AAAA,MACD;AAAA,QACE;AAAA,kBACU,mBAAmB,IAAI;AAAA;AAAA;AAAA,MAGnC;AAAA,MACA,CAAC,SAAS,aAAa,YAAY,QAAQ,WAAW;AAAA,IACxD;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,uBAAuB,IAAI;AAAA,EAC/D;AACF;;;ACpBO,IAAM,4CAA4C;AAClD,IAAM,uDAAuD;AA4C7D,IAAM,qCAAqC,CAEhD;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF,MAAgG;AAC9F,MAAI,YAAY;AAEhB,MAAI;AAEJ,QAAM,eAAe,OACnB,YACG;AACH,UAAM,QACJ,QAAQ,cAAc,cAClB,KACA,QAAQ,cAAc,SAClB,MAAM,8BAA8B,UAAU,GAC7C,yBAAyB,KAC5B,QAAQ,UAAU;AAE1B,UAAM,sBAAgD;AAAA,MACpD;AAAA,MACA;AAAA,IACF;AAEA,QAAI,WAAW;AAEf,OAAG;AACD,YAAM,EAAE,UAAU,uBAAuB,cAAc,IACrD,MAAM,kBAA6B,YAAY,mBAAmB;AAEpE,UAAI,SAAS,SAAS,GAAG;AACvB,cAAM,SAAS,MAAM,UAAU,EAAE,SAAS,CAAC;AAE3C,YAAI,UAAU,OAAO,SAAS,QAAQ;AACpC,sBAAY;AACZ;AAAA,QACF;AAAA,MACF;AAEA,0BAAoB,QAAQ;AAE5B,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAE5D,UAAI,CAAC,eAAe;AAClB,mBAAW,KAAK,IAAI,WAAW,GAAG,GAAI;AAAA,MACxC,OAAO;AACL,mBAAW;AAAA,MACb;AAAA,IACF,SAAS;AAAA,EACX;AAEA,SAAO;AAAA,IACL,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,OAAO,CAAC,YAAY;AAClB,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,oBAAY;AAEZ,eAAO,aAAa,OAAO;AAAA,MAC7B,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA,MAAM,YAAY;AAChB,UAAI,CAAC,UAAW;AAChB,kBAAY;AACZ,YAAM;AAAA,IACR;AAAA,EACF;AACF;AAEO,IAAM,iDAAiD,CAC5D,YACgD;AAChD,MACE,QAAQ,WAAW,KACnB,QAAQ,KAAK,CAAC,MAAM,MAAM,UAAa,MAAM,WAAW;AAExD,WAAO;AAET,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,KAAK,EAAG,QAAO;AAE9C,SAAO,QACJ,OAAO,CAAC,MAAM,MAAM,UAAa,MAAM,eAAe,MAAM,KAAK,EACjE,KAAK,CAAC,GAAG,MAAO,IAAI,IAAI,IAAI,EAAG,EAAE,CAAC;AACvC;;;ACzCA,IAAM,yBAAyB,CAC7B,YACoB;AACpB,QAAM,EAAE,YAAY,IAAI;AACxB,MAAI,WAAW;AAGf,QAAM,QAAQ,CAAC,YAG2C;AACxD,UAAM,WAAW,QAAQ,YAAY,QAAQ,mBAAmB;AAChE,QAAI,CAAC;AACH,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ,WAAW;AAAA,MAC1C;AAEF,UAAM,aACJ,QAAQ,cACR,QAAQ,mBAAmB,cAC3B,iBAAiB,EAAE,SAAS,CAAC;AAE/B,WAAO,EAAE,YAAY,SAAS;AAAA,EAChC;AAEA,SAAO;AAAA,IACL,IAAI,QAAQ;AAAA,IACZ,OAAO,OACL,eACqE;AACrE,iBAAW;AACX,UAAI,QAAQ,cAAc,UAAW,QAAO,QAAQ;AAEpD,YAAM,EAAE,sBAAsB,IAAI,MAAM;AAAA,QACtC;AAAA,QACA;AAAA,UACE,aAAa,QAAQ;AAAA,UACrB,WAAW,QAAQ;AAAA,QACrB;AAAA,MACF;AAEA,UAAI,0BAA0B,KAAM,QAAO;AAE3C,aAAO,EAAE,gBAAgB,sBAAsB;AAAA,IACjD;AAAA,IACA,IAAI,WAAW;AACb,aAAO;AAAA,IACT;AAAA,IACA,QAAQ,OACN,EAAE,SAAS,GACX,YACiD;AACjD,UAAI,CAAC,SAAU;AAEf,YAAM,EAAE,YAAY,SAAS,IAAI,MAAM,OAAO;AAE9C,aAAO,WAAW,gBAAgB,YAAY;AAC5C,YAAI,SAA0D;AAE9D,YAAI,wBAAuC;AAE3C,mBAAW,WAAW,UAAU;AAC9B,gBAAM,eAAe;AAKrB,gBAAM,0BAA0B,MAAM,YAAY,cAAc;AAAA,YAC9D;AAAA,YACA;AAAA,UACF,CAAC;AAED,gBAAM,cAA6B,cAAc,YAAY;AAG7D,gBAAM,yBAAyB,YAAY;AAAA,YACzC,aAAa,QAAQ;AAAA,YACrB,SAAS,QAAQ;AAAA,YACjB;AAAA,YACA;AAAA,YACA,WAAW,QAAQ;AAAA,UACrB,CAAC;AAED,kCAAwB,aAAa,SAAS;AAE9C,cACE,2BACA,wBAAwB,SAAS,QACjC;AACA,uBAAW;AACX,qBAAS;AACT;AAAA,UACF;AAEA,cAAI,QAAQ,aAAa,QAAQ,UAAU,YAAY,GAAG;AACxD,uBAAW;AACX,qBAAS,EAAE,MAAM,QAAQ,QAAQ,yBAAyB;AAC1D;AAAA,UACF;AAEA,cACE,2BACA,wBAAwB,SAAS;AAEjC;AAAA,QACJ;AACA,eAAO;AAAA,MACT,CAAC;AAAA,IACH;AAAA,EACF;AACF;AAEO,IAAM,4BAA4B,CACvC,YACoB;AACpB,QAAMC,cAAa,QAAQ;AAE3B,SAAO,uBAAkC;AAAA,IACvC,aAAa,QAAQ,eAAe,cAAcA,YAAW,IAAI;AAAA,IACjE,aAAa,OAAO,OAAO,YAAY;AACrC,UAAI,CAACA,YAAW,UAAU,SAAS,MAAM,IAAI,EAAG;AAEhD,YAAMA,YAAW,OAAO,CAAC,KAAK,GAAG,EAAE,YAAY,QAAQ,WAAW,CAAC;AAAA,IACrE;AAAA,IACA,GAAG;AAAA,EACL,CAAC;AACH;AAEO,IAAM,kBAAkB,CAC7B,YACoB;AACpB,MAAI,gBAAgB,SAAS;AAC3B,WAAO,0BAA0B,OAAO;AAAA,EAC1C;AAEA,SAAO,uBAAuB,OAAO;AACvC;;;ACzMO,IAAM,2BAA2B,CAGtC,YACgD;AAChD,MAAI,YAAY;AAChB,QAAM,EAAE,QAAQ,IAAI;AACpB,QAAM,aAAa,QAAQ,cAAc,CAAC;AAE1C,MAAI;AAEJ,MAAI;AAEJ,QAAM,aACJ,QAAQ,cAAc,iBAAiB,EAAE,UAAU,QAAQ,SAAS,CAAC;AAEvE,QAAM,YAEF,OAAO,kBAAkB;AAC3B,UAAM,mBAAmB,WAAW,OAAO,CAAC,MAAM,EAAE,QAAQ;AAE5D,QAAI,iBAAiB,WAAW;AAC9B,aAAO;AAAA,QACL,MAAM;AAAA,QACN,QAAQ;AAAA,MACV;AAEF,UAAM,SAAS,MAAM,QAAQ;AAAA,MAC3B,iBAAiB,IAAI,CAAC,MAAM;AAE1B,eAAO,EAAE,OAAO,eAAe;AAAA,UAC7B;AAAA,UACA,UAAU,QAAQ;AAAA,QACpB,CAAC;AAAA,MACH,CAAC;AAAA,IACH;AAEA,WAAO,OAAO;AAAA,MACZ,CAAC,MAAM,EAAE,WAAW,eAAe,EAAE,OAAO,SAAS;AAAA,IACvD,IACI,SACA;AAAA,MACE,MAAM;AAAA,IACR;AAAA,EACN;AAEA,QAAM,gBAAiB,uBACrB,mCAAmC;AAAA,IACjC;AAAA,IACA;AAAA,IACA,WACE,SAAS,aAAa;AAAA,IACxB,sBACE,SAAS,wBACT;AAAA,EACJ,CAAC;AAEH,QAAM,OAAO,YAAY;AACvB,QAAI,CAAC,UAAW;AAChB,gBAAY;AACZ,QAAI,sBAAsB;AACxB,YAAM,qBAAqB,KAAK;AAChC,6BAAuB;AAAA,IACzB;AACA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL;AAAA,IACA,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,WAAW,CACTC,aAC+B;AAC/B,YAAM,YAAY,gBAA2BA,QAAO;AAEpD,iBAAW,KAAK,SAAS;AAEzB,aAAO;AAAA,IACT;AAAA,IACA,OAAO,MAAM;AACX,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,YAAI,WAAW,WAAW;AACxB,iBAAO,QAAQ;AAAA,YACb,IAAI;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAEF,oBAAY;AAEZ,cAAM,YAAY;AAAA,UAChB,MAAM,QAAQ,IAAI,WAAW,IAAI,CAAC,MAAM,EAAE,MAAM,UAAU,CAAC,CAAC;AAAA,QAC9D;AAEA,eAAO,cAAc,MAAM,EAAE,UAAU,CAAC;AAAA,MAC1C,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA;AAAA,IACA,OAAO,YAAY;AACjB,YAAM,KAAK;AAEX,iBAAW,MAAM;AAEjB,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,GAAG,CAAC;AAAA,IACzD;AAAA,EACF;AACF;;;ACvHO,IAAM,oBAAoB,OAC/B,YACkB;AAClB,QAAM,EAAE,aAAa,gBAAgB,QAAQ,WAAW,IAAI;AAE5D,QAAM,aAAa,OAAO,IAAI,CAAC,MAAM,EAAE,IAAI;AAE3C,aAAWC,eAAc,gBAAgB;AACvC,QAAI,CAACA,YAAW,UAAU,KAAK,CAAC,SAAS,WAAW,SAAS,IAAI,CAAC,GAAG;AACnE;AAAA,IACF;AACA,UAAMA,YAAW,OAAO,QAAQ;AAAA,MAC9B;AAAA,IACF,CAAC;AAAA,EACH;AACF;;;ACRO,IAAM,uCAAuC;AA+C7C,IAAM,sBAAsB,CACjC,YACqB;AACrB,MAAI,iBAAiB;AACrB,MAAI,qBAAqB;AACzB,MAAI;AACJ,QAAM,WAAW,QAAQ,YAAY;AAErC,QAAM,aACJ,aAAa,0BACb,aAAa;AAEf,QAAM,qBAAqB,QAAQ,eAAe,CAAC,GAChD,OAAO,CAAC,EAAE,KAAK,MAAM,SAAS,QAAQ,EACtC,IAAI,CAAC,EAAE,YAAAC,YAAW,MAAMA,WAAU;AAErC,QAAM,qBAAqB,QAAQ,OAAO;AAE1C,QAAM,mBAAmB,MAAM;AAC7B,QAAI,YAAY,MAAM;AACpB,aAAO;AAAA,IACT;AAEA,WAAO,iBAAiB;AAAA,MACtB;AAAA,IACF,CAAC;AAAA,EACH;AAEA,QAAM,kBAAkB,MAAM;AAC5B,QAAI,YAAY;AACd;AAAA,IACF;AACA,QAAI,YAAY,MAAM;AACpB,eAAS,MAAM;AACf,iBAAW;AAAA,IACb;AAAA,EACF;AAEA,QAAM,iBAAiB,OACrB,YACoB;AACpB,QAAI,YAAY,MAAM;AACpB,iBAAW,iBAAiB;AAAA,IAC9B;AAEA,QAAI;AACF,YAAM,mBAAmB,QAAQ;AACjC,aAAO,MAAM,QAAQ,QAAQ;AAAA,IAC/B,UAAE;AACA,sBAAgB;AAAA,IAClB;AAAA,EACF;AAEA,MAAI,SAAS;AACX,yBACE,QAAQ,QAAQ,kBAAkB,UAClC,QAAQ,QAAQ,kBAAkB;AAAA,EACtC;AAEA,QAAM,qBAAqB,OACzB,eACkB;AAClB,QAAI,CAAC,mBAAoB,QAAO,QAAQ,QAAQ;AAEhD,QAAI,CAAC,gBAAgB;AACnB,YAAM,uBAAuB,UAAU;AACvC,uBAAiB;AAAA,IACnB;AAEA,WAAO,QAAQ,QAAQ;AAAA,EACzB;AAEA,SAAO;AAAA,IACL,MAAM,gBACJ,YACAC,UAKuC;AACvC,YAAM,EAAE,QAAQ,cAAc,KAAK,IAAIA;AAEvC,YAAM,wBAAwB,MAAM;AAEpC,UAAI,QAAQ,aAAa;AAEzB,UAAI,OAAO,eAAe,UAAU;AAClC,cAAM,IAAI,MAAM,2BAA2B;AAAA,MAC7C;AAEA,UAAI,YAAY,MAAM;AACpB,mBAAW,iBAAiB;AAAA,MAC9B;AAEA,YAAM,SAAS,MAAM;AAAA,QAAe,CAAC,eACnC,WAAsB,YAAY,YAAYA,SAAQ,IAAI;AAAA,MAC5D;AAEA,YAAM,uBAAuB,OAAO;AAEpC;AAAA,QACE;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,iBAAW,SAAS,OAAO,QAAQ;AACjC,YAAI,CAAC,MAAO;AAEZ,gBAAQ,OAAO,OAAO,KAAK;AAAA,MAC7B;AAEA,aAAO;AAAA,QACL;AAAA,QACA;AAAA,QACA,cAAc,OAAO;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,YAAY,OACV,YACAA,aAIA;AAAA,MAAe,CAAC,eACd,WAAsB,YAAY,YAAYA,QAAO;AAAA,IACvD;AAAA,IAEF,gBAAgB,OACd,YACA,QACAA,aACoD;AACpD,UAAI,YAAY,MAAM;AACpB,mBAAW,iBAAiB;AAAA,MAC9B;AAGA,YAAM,CAAC,WAAW,GAAG,IAAI,IAAI,WAAW,MAAM,GAAG;AAEjD,YAAM,aACJ,aAAa,KAAK,SAAS,IAAI,YAAY;AAE7C,YAAM,eAAe,MAAM;AAAA,QAAe,CAAC,eACzC,eAAe,YAAY,YAAY,YAAY,QAAQ;AAAA,UACzD,GAAGA;AAAA,UACH,gBAAgB,OAAO,UAAU,YAAY;AAC3C,gBAAI,kBAAkB,SAAS;AAC7B,oBAAM,kBAAkB;AAAA,gBACtB,aAAa;AAAA,gBACb,QAAQ;AAAA,gBACR,GAAG;AAAA,cACL,CAAC;AAEH,gBAAI,mBAAoB,OAAM,mBAAmB,UAAU,OAAO;AAAA,UACpE;AAAA,QACF,CAAC;AAAA,MACH;AAEA,UAAI,CAAC,aAAa;AAChB,cAAM,IAAI;AAAA,UACR,CAAC;AAAA;AAAA,UACDA,UAAS,yBAAyB;AAAA,QACpC;AAEF,aAAO;AAAA,QACL,2BAA2B,aAAa;AAAA,QACxC,yBAAyB,aAAa;AAAA,QACtC,kBACE,aAAa,sBAAsB,OAAO,OAAO,MAAM;AAAA,MAC3D;AAAA,IACF;AAAA,IACA,UAAU,CACRA,aAEA,yBAA4C;AAAA,MAC1C,GAAIA,YAAW,CAAC;AAAA,MAChB;AAAA,MACA,YAAY,YAAY;AAAA,IAC1B,CAAC;AAAA,EACL;AACF;;;AC1PO,IAAM,aAAa,OACxB,IACA,UACA,YAGG;AACH,QAAM,gBACJ,WAAW,UAAU,UACjB,0BAA0B,QAAQ,IAAI,KACtC;AAEN,QAAM,KAAK;AAAA,IACT,WAAW,QAAQ,UACf,QAAQ,KACR,WAAW,cAAc,WAAW,QAAQ,WAC1C,QAAQ,OAAO,QAAQ,WACvB;AAAA,EACR;AAEA,QAAM,cAAc,CAAC,MAAM,EAAE,IAAI,0BAA0B,EAAE,KAAK;AAElE,QAAM,UAAU,MAAM,GAAG;AAAA,IACvB;AAAA,kBACc,cAAc,IAAI;AAAA,2EACuC,aAAa,IAAI,WAAW;AAAA;AAAA,IAEnG,CAAC,UAAU,SAAS,aAAa,UAAU;AAAA,EAC7C;AAEA,QAAM,WACJ,QAAQ,IAAI,CAAC,QAAQ;AACnB,UAAM,WAAW;AAAA,MACf,MAAM,IAAI;AAAA,MACV,MAAM,WAAW,MAAM,IAAI,YAAY;AAAA,MACvC,UAAU,WAAW,MAAM,IAAI,gBAAgB;AAAA,IACjD;AAEA,UAAM,WAAgD;AAAA,MACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,MAC1D,WAAW,IAAI;AAAA,MACf,YAAY;AAAA,MACZ,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,IAC5C;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH,MAAM;AAAA,MACN;AAAA,IAIF;AAAA,EACF,CAAC;AAEH,SAAO,SAAS,SAAS,IACrB;AAAA,IACE,sBACE,SAAS,SAAS,SAAS,CAAC,EAAG,SAAS;AAAA,IAC1C,QAAQ;AAAA,IACR,cAAc;AAAA,EAChB,IACA;AAAA,IACE,sBAAsB;AAAA,IACtB,QAAQ,CAAC;AAAA,IACT,cAAc;AAAA,EAChB;AACN;;;ACtFA,eAAe,kCACb,IACA,aACA,SACA,UACA,eACA,WACoB;AACpB,MAAI,kBAAkB,MAAM;AAC1B,UAAM,eAAe,MAAM,GAAG;AAAA,MAC5B,IAAI;AAAA,mBACS,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,SAKjC;AAAA,MACH,CAAC,SAAU,SAAS,GAAG,aAAa,cAAc,SAAS,GAAG,SAAS;AAAA,IACzE;AACA,QAAI,aAAa,UAAU,GAAG;AAC5B,aAAO;AAAA,IACT,OAAO;AACL,YAAM,mBAAmB,MAAM;AAAA,QAC7B,GAAG;AAAA,UACD;AAAA,YACE,uCAAuC,mBAAmB,IAAI;AAAA;AAAA,UAEhE;AAAA,UACA,CAAC,aAAa,SAAS;AAAA,QACzB;AAAA,MACF;AAEA,UAAI,kBAAkB,4BAA4B,UAAU;AAC1D,eAAO;AAAA,MACT,WACE,aAAa,QACb,qBAAqB,QACrB,kBAAkB,0BAA0B,UAC5C;AACA,eAAO;AAAA,MACT,OAAO;AACL,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF,OAAO;AACL,QAAI;AACF,YAAM,GAAG;AAAA,QACP;AAAA,UACE,eAAe,mBAAmB,IAAI;AAAA,QACxC;AAAA,QACA,CAAC,aAAa,SAAS,SAAU,SAAS,GAAG,SAAS;AAAA,MACxD;AACA,aAAO;AAAA,IACT,SAAS,KAAK;AACZ,UAAI,EAAE,cAAc,GAAG,MAAM,IAAI,UAAU,MAAM,IAAI,UAAU,QAAQ;AACrE,cAAM;AAAA,MACR;AAEA,YAAM,UAAU,MAAM;AAAA,QACpB,GAAG;AAAA,UACD;AAAA,YACE,uCAAuC,mBAAmB,IAAI;AAAA,UAChE;AAAA,UACA,CAAC,aAAa,SAAS;AAAA,QACzB;AAAA,MACF;AACA,UAAI,SAAS,4BAA4B,UAAU;AACjD,eAAO;AAAA,MACT,OAAO;AACL,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF;AACF;AAWA,eAAsB,yBACpB,IACA,SAOmE;AACnE,MAAI;AACF,UAAM,SAAS,MAAM;AAAA,MACnB;AAAA,MACA,QAAQ;AAAA,MACR,QAAQ,WAAW;AAAA,MACnB,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ,aAAa;AAAA,IACvB;AAEA,WAAO,WAAW,IACd,EAAE,SAAS,MAAM,aAAa,QAAQ,YAAY,IAClD,EAAE,SAAS,OAAO,QAAQ,WAAW,IAAI,YAAY,WAAW;AAAA,EACtE,SAAS,OAAO;AACd,YAAQ,IAAI,KAAK;AACjB,UAAM;AAAA,EACR;AACF;","names":["sql","uuid","TransformStream","error","TransformStream","uuid","message","filter","TransformStream","map","handleChunk","error","skip","retry","uuid","uuid","sql","projection","options","projection","projection","options"]}
1
+ {"version":3,"sources":["../src/connection/sqliteConnection.ts","../../emmett/src/config/plugins/index.ts","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/validation/dates.ts","../../emmett/src/typing/command.ts","../../emmett/src/typing/event.ts","../../emmett/src/typing/message.ts","../../emmett/src/typing/workflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/eventStore/afterCommit/forwardToMessageBus.ts","../../emmett/src/eventStore/events/index.ts","../../emmett/src/eventStore/eventStore.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/database/inMemoryDatabase.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/utils/iterators.ts","../../emmett/src/taskProcessing/taskProcessor.ts","../../emmett/src/utils/locking/index.ts","../../emmett/src/utils/promises.ts","../../emmett/src/utils/retry.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/database/utils.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjection.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjectionSpec.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/testing/deciderSpecification.ts","../../emmett/src/testing/wrapEventStore.ts","../../emmett/src/eventStore/subscriptions/caughtUpTransformStream.ts","../../emmett/src/eventStore/subscriptions/streamingCoordinator.ts","../../emmett/src/streaming/transformations/notifyAboutNoActiveReaders.ts","../../emmett/src/streaming/writers/writeToStream.ts","../../emmett/src/commandHandling/handleCommand.ts","../../emmett/src/commandHandling/handleCommandWithDecider.ts","../../emmett/src/messageBus/index.ts","../../emmett/src/processors/processors.ts","../../emmett/src/processors/inMemoryProcessors.ts","../../emmett/src/projections/index.ts","../../emmett/src/streaming/binaryArrays.ts","../../emmett/src/streaming/collectors/collect.ts","../../emmett/src/streaming/decoders/binary.ts","../../emmett/src/streaming/decoders/string.ts","../../emmett/src/streaming/decoders/json.ts","../../emmett/src/streaming/decoders/object.ts","../../emmett/src/streaming/decoders/composite.ts","../../emmett/src/streaming/generators/fromArray.ts","../../emmett/src/streaming/generators/index.ts","../../emmett/src/streaming/restream.ts","../../emmett/src/streaming/transformations/filter.ts","../../emmett/src/streaming/transformations/map.ts","../../emmett/src/streaming/transformations/reduce.ts","../../emmett/src/streaming/transformations/retry.ts","../../emmett/src/streaming/transformations/skip.ts","../../emmett/src/streaming/transformations/stopAfter.ts","../../emmett/src/streaming/transformations/stopOn.ts","../../emmett/src/streaming/transformations/take.ts","../../emmett/src/streaming/transformations/waitAtMost.ts","../../emmett/src/streaming/transformations/index.ts","../src/eventStore/schema/appendToStream.ts","../src/eventStore/schema/typing.ts","../src/eventStore/schema/tables.ts","../src/eventStore/schema/utils.ts","../src/eventStore/schema/readLastMessageGlobalPosition.ts","../src/eventStore/schema/readMessagesBatch.ts","../src/eventStore/schema/readProcessorCheckpoint.ts","../src/eventStore/consumers/messageBatchProcessing/index.ts","../src/eventStore/consumers/sqliteProcessor.ts","../src/eventStore/consumers/sqliteEventStoreConsumer.ts","../src/eventStore/projections/index.ts","../src/eventStore/SQLiteEventStore.ts","../src/eventStore/schema/readStream.ts","../src/eventStore/schema/storeProcessorCheckpoint.ts"],"sourcesContent":["import sqlite3 from 'sqlite3';\n\nexport type Parameters = object | string | bigint | number | boolean | null;\n\nexport type SQLiteConnection = {\n close: () => void;\n command: (sql: string, values?: Parameters[]) => Promise<sqlite3.RunResult>;\n query: <T>(sql: string, values?: Parameters[]) => Promise<T[]>;\n querySingle: <T>(sql: string, values?: Parameters[]) => Promise<T | null>;\n withTransaction: <T>(fn: () => Promise<T>) => Promise<T>;\n};\n\nexport interface SQLiteError extends Error {\n errno: number;\n}\n\nexport const isSQLiteError = (error: unknown): error is SQLiteError => {\n if (error instanceof Error && 'code' in error) {\n return true;\n }\n\n return false;\n};\n\nexport type InMemorySharedCacheSQLiteDatabase = 'file::memory:?cache=shared';\nexport const InMemorySharedCacheSQLiteDatabase = 'file::memory:?cache=shared';\nexport type InMemorySQLiteDatabase = ':memory:';\nexport const InMemorySQLiteDatabase = ':memory:';\n\ntype SQLiteConnectionOptions = {\n // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n fileName: InMemorySQLiteDatabase | string | undefined;\n};\n\nexport const sqliteConnection = (\n options: SQLiteConnectionOptions,\n): SQLiteConnection => {\n const fileName = options.fileName ?? InMemorySQLiteDatabase;\n let db: sqlite3.Database;\n\n if (fileName.startsWith('file:')) {\n db = new sqlite3.Database(\n fileName,\n sqlite3.OPEN_URI | sqlite3.OPEN_READWRITE | sqlite3.OPEN_CREATE,\n );\n } else {\n db = new sqlite3.Database(fileName);\n }\n db.run('PRAGMA journal_mode = WAL;');\n let transactionNesting = 0;\n\n return {\n close: (): void => db.close(),\n command: (sql: string, params?: Parameters[]) =>\n new Promise<sqlite3.RunResult>((resolve, reject) => {\n db.run(\n sql,\n params ?? [],\n function (this: sqlite3.RunResult, err: Error | null) {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(this);\n },\n );\n }),\n query: <T>(sql: string, params?: Parameters[]): Promise<T[]> =>\n new Promise((resolve, reject) => {\n db.all(sql, params ?? [], (err: Error | null, result: T[]) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(result);\n });\n }),\n querySingle: <T>(sql: string, params?: Parameters[]): Promise<T | null> =>\n new Promise((resolve, reject) => {\n db.get(sql, params ?? [], (err: Error | null, result: T | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve(result);\n });\n }),\n withTransaction: async <T>(fn: () => Promise<T>) => {\n try {\n if (transactionNesting++ == 0) {\n await beginTransaction(db);\n }\n const result = await fn();\n\n if (transactionNesting === 1) await commitTransaction(db);\n transactionNesting--;\n\n return result;\n } catch (err) {\n console.log(err);\n\n if (--transactionNesting === 0) await rollbackTransaction(db);\n\n throw err;\n }\n },\n };\n};\n\nconst beginTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('BEGIN IMMEDIATE TRANSACTION', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n\nconst commitTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('COMMIT', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n\nconst rollbackTransaction = (db: sqlite3.Database) =>\n new Promise<void>((resolve, reject) => {\n db.run('ROLLBACK', (err: Error | null) => {\n if (err) {\n reject(err);\n return;\n }\n\n resolve();\n });\n });\n","export type EmmettPluginConfig =\n | {\n name: string;\n register: EmmettPluginRegistration[];\n }\n | string;\n\nexport type EmmettPluginType = 'cli';\n\nexport type EmmettCliPluginRegistration = { pluginType: 'cli'; path?: string };\n\nexport type EmmettPluginRegistration = EmmettCliPluginRegistration;\n\nexport type EmmettCliCommand = {\n addCommand<CliCommand>(command: CliCommand): CliCommand;\n};\n\nexport type EmmettCliPlugin = {\n pluginType: 'cli';\n name: string;\n registerCommands: (program: EmmettCliCommand) => Promise<void> | void;\n};\n\nexport type EmmettPlugin = EmmettCliPlugin;\n\nexport const isPluginConfig = (\n plugin: Partial<EmmettPluginConfig> | string | undefined,\n): plugin is EmmettPluginConfig =>\n plugin !== undefined &&\n (typeof plugin === 'string' ||\n ('name' in plugin &&\n plugin.name !== undefined &&\n typeof plugin.name === 'string'));\n","import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const isBigint = (val: any): val is bigint =>\n typeof val === 'bigint' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : 500;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n\n public static mapFrom(\n error: Error | { message?: string; errorCode?: number },\n ): EmmettError {\n if (error instanceof EmmettError) {\n return error;\n }\n\n return new EmmettError({\n errorCode:\n 'errorCode' in error &&\n error.errorCode !== undefined &&\n error.errorCode !== null\n ? error.errorCode\n : 500,\n message: error.message ?? 'An unknown error occurred',\n });\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: 412,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\nexport class ConcurrencyInMemoryDatabaseError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 412,\n message: message ?? `Expected document state does not match current one!`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyInMemoryDatabaseError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 400,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 403,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: 404,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { ValidationError } from '../errors';\n\nexport const formatDateToUtcYYYYMMDD = (date: Date) => {\n // Use the 'en-CA' locale which formats as 'yyyy-mm-dd'\n const formatter = new Intl.DateTimeFormat('en-CA', {\n timeZone: 'UTC',\n year: 'numeric',\n month: '2-digit',\n day: '2-digit',\n });\n\n // Format the date\n return formatter.format(date);\n};\n\n// Function to validate 'yyyy-mm-dd' format\nexport const isValidYYYYMMDD = (dateString: string) => {\n const regex = /^\\d{4}-\\d{2}-\\d{2}$/;\n return regex.test(dateString);\n};\n\nexport const parseDateFromUtcYYYYMMDD = (dateString: string) => {\n const date = new Date(dateString + 'T00:00:00Z');\n\n if (!isValidYYYYMMDD(dateString)) {\n throw new ValidationError('Invalid date format, must be yyyy-mm-dd');\n }\n\n if (isNaN(date.getTime())) {\n throw new ValidationError('Invalid date format');\n }\n\n return date;\n};\n","import type { DefaultRecord } from './';\n\nexport type Command<\n CommandType extends string = string,\n CommandData extends DefaultRecord = DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: Readonly<CommandData>;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n> & { readonly kind?: 'Command' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyCommand = Command<any, any, any>;\n\nexport type CommandTypeOf<T extends Command> = T['type'];\nexport type CommandDataOf<T extends Command> = T['data'];\nexport type CommandMetaDataOf<T extends Command> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CreateCommandType<\n CommandType extends string,\n CommandData extends DefaultRecord,\n CommandMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n CommandMetaData extends undefined\n ? {\n type: CommandType;\n data: CommandData;\n metadata?: DefaultCommandMetadata | undefined;\n }\n : {\n type: CommandType;\n data: CommandData;\n metadata: CommandMetaData;\n }\n> & { readonly kind?: 'Command' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const command = <CommandType extends Command<string, any, any>>(\n ...args: CommandMetaDataOf<CommandType> extends undefined\n ? [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata?: DefaultCommandMetadata | undefined,\n ]\n : [\n type: CommandTypeOf<CommandType>,\n data: CommandDataOf<CommandType>,\n metadata: CommandMetaDataOf<CommandType>,\n ]\n): CommandType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind: 'Command' } as CommandType)\n : ({ type, data, kind: 'Command' } as CommandType);\n};\n\nexport type DefaultCommandMetadata = { now: Date };\n","import type { DefaultRecord } from './';\nimport type {\n AnyRecordedMessageMetadata,\n CombinedMessageMetadata,\n CommonRecordedMessageMetadata,\n GlobalPositionTypeOfRecordedMessageMetadata,\n RecordedMessage,\n RecordedMessageMetadata,\n RecordedMessageMetadataWithGlobalPosition,\n RecordedMessageMetadataWithoutGlobalPosition,\n StreamPositionTypeOfRecordedMessageMetadata,\n} from './message';\n\nexport type BigIntStreamPosition = bigint;\nexport type BigIntGlobalPosition = bigint;\n\nexport type Event<\n EventType extends string = string,\n EventData extends DefaultRecord = DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n> & { readonly kind?: 'Event' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyEvent = Event<any, any, any>;\n\nexport type EventTypeOf<T extends Event> = T['type'];\nexport type EventDataOf<T extends Event> = T['data'];\nexport type EventMetaDataOf<T extends Event> = T extends { metadata: infer M }\n ? M\n : undefined;\n\nexport type CreateEventType<\n EventType extends string,\n EventData extends DefaultRecord,\n EventMetaData extends DefaultRecord | undefined = undefined,\n> = Readonly<\n EventMetaData extends undefined\n ? {\n type: EventType;\n data: EventData;\n }\n : {\n type: EventType;\n data: EventData;\n metadata: EventMetaData;\n }\n> & { readonly kind?: 'Event' };\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const event = <EventType extends Event<string, any, any>>(\n ...args: EventMetaDataOf<EventType> extends undefined\n ? [type: EventTypeOf<EventType>, data: EventDataOf<EventType>]\n : [\n type: EventTypeOf<EventType>,\n data: EventDataOf<EventType>,\n metadata: EventMetaDataOf<EventType>,\n ]\n): EventType => {\n const [type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind: 'Event' } as EventType)\n : ({ type, data, kind: 'Event' } as EventType);\n};\n\nexport type CombinedReadEventMetadata<\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = CombinedMessageMetadata<EventType, EventMetaDataType>;\n\nexport type ReadEvent<\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = RecordedMessage<EventType, EventMetaDataType>;\n\nexport type AnyReadEvent<\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = ReadEvent<AnyEvent, EventMetaDataType>;\n\nexport type CommonReadEventMetadata<StreamPosition = BigIntStreamPosition> =\n CommonRecordedMessageMetadata<StreamPosition>;\n\nexport type ReadEventMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadata<GlobalPosition, StreamPosition>;\n\nexport type AnyReadEventMetadata = AnyRecordedMessageMetadata;\n\nexport type ReadEventMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = RecordedMessageMetadataWithGlobalPosition<GlobalPosition>;\n\nexport type ReadEventMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadataWithoutGlobalPosition<StreamPosition>;\n\nexport type GlobalPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n GlobalPositionTypeOfRecordedMessageMetadata<ReadEventMetadataType>;\n\nexport type StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType> =\n StreamPositionTypeOfRecordedMessageMetadata<ReadEventMetadataType>;\n","import type {\n AnyCommand,\n AnyEvent,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n Command,\n DefaultRecord,\n Event,\n} from '.';\n\nexport type Message<\n Type extends string = string,\n Data extends DefaultRecord = DefaultRecord,\n MetaData extends DefaultRecord | undefined = undefined,\n> = Command<Type, Data, MetaData> | Event<Type, Data, MetaData>;\n\nexport type AnyMessage = AnyEvent | AnyCommand;\n\nexport type MessageKindOf<T extends Message> = T['kind'];\nexport type MessageTypeOf<T extends Message> = T['type'];\nexport type MessageDataOf<T extends Message> = T['data'];\nexport type MessageMetaDataOf<T extends Message> = T extends {\n metadata: infer M;\n}\n ? M\n : undefined;\n\nexport type CanHandle<T extends Message> = MessageTypeOf<T>[];\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const message = <MessageType extends Message<string, any, any>>(\n ...args: MessageMetaDataOf<MessageType> extends undefined\n ? [\n kind: MessageKindOf<MessageType>,\n type: MessageTypeOf<MessageType>,\n data: MessageDataOf<MessageType>,\n ]\n : [\n kind: MessageKindOf<MessageType>,\n type: MessageTypeOf<MessageType>,\n data: MessageDataOf<MessageType>,\n metadata: MessageMetaDataOf<MessageType>,\n ]\n): MessageType => {\n const [kind, type, data, metadata] = args;\n\n return metadata !== undefined\n ? ({ type, data, metadata, kind } as MessageType)\n : ({ type, data, kind } as MessageType);\n};\n\nexport type CombinedMessageMetadata<\n MessageType extends Message = Message,\n MessageMetaDataType extends DefaultRecord = DefaultRecord,\n> =\n MessageMetaDataOf<MessageType> extends undefined\n ? MessageMetaDataType\n : MessageMetaDataOf<MessageType> & MessageMetaDataType;\n\nexport type CombineMetadata<\n MessageType extends Message = Message,\n MessageMetaDataType extends DefaultRecord = DefaultRecord,\n> = MessageType & {\n metadata: CombinedMessageMetadata<MessageType, MessageMetaDataType>;\n};\n\nexport type RecordedMessage<\n MessageType extends Message = Message,\n MessageMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n> = CombineMetadata<MessageType, MessageMetaDataType> & {\n kind: NonNullable<MessageKindOf<Message>>;\n};\n\nexport type CommonRecordedMessageMetadata<\n StreamPosition = BigIntStreamPosition,\n> = Readonly<{\n messageId: string;\n streamPosition: StreamPosition;\n streamName: string;\n}>;\n\nexport type WithGlobalPosition<GlobalPosition> = Readonly<{\n globalPosition: GlobalPosition;\n}>;\n\nexport type RecordedMessageMetadata<\n GlobalPosition = undefined,\n StreamPosition = BigIntStreamPosition,\n> = CommonRecordedMessageMetadata<StreamPosition> &\n // eslint-disable-next-line @typescript-eslint/no-empty-object-type\n (GlobalPosition extends undefined ? {} : WithGlobalPosition<GlobalPosition>);\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyRecordedMessageMetadata = RecordedMessageMetadata<any, any>;\n\nexport type RecordedMessageMetadataWithGlobalPosition<\n GlobalPosition = BigIntGlobalPosition,\n> = RecordedMessageMetadata<GlobalPosition>;\n\nexport type RecordedMessageMetadataWithoutGlobalPosition<\n StreamPosition = BigIntStreamPosition,\n> = RecordedMessageMetadata<undefined, StreamPosition>;\n\nexport type GlobalPositionTypeOfRecordedMessageMetadata<\n RecordedMessageMetadataType,\n> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RecordedMessageMetadataType extends RecordedMessageMetadata<infer GP, any>\n ? GP\n : never;\n\nexport type StreamPositionTypeOfRecordedMessageMetadata<\n RecordedMessageMetadataType,\n> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RecordedMessageMetadataType extends RecordedMessageMetadata<any, infer SV>\n ? SV\n : never;\n","import type { AnyCommand } from './command';\nimport type { AnyEvent } from './event';\n\n/// Inspired by https://blog.bittacklr.be/the-workflow-pattern.html\n\nexport type Workflow<\n Input extends AnyEvent | AnyCommand,\n State,\n Output extends AnyEvent | AnyCommand,\n> = {\n decide: (command: Input, state: State) => WorkflowOutput<Output>[];\n evolve: (currentState: State, event: WorkflowEvent<Output>) => State;\n initialState: () => State;\n};\n\nexport type WorkflowEvent<Output extends AnyEvent | AnyCommand> = Extract<\n Output,\n { kind?: 'Event' }\n>;\n\nexport type WorkflowCommand<Output extends AnyEvent | AnyCommand> = Extract<\n Output,\n { kind?: 'Command' }\n>;\n\nexport type WorkflowOutput<TOutput extends AnyEvent | AnyCommand> =\n | { action: 'Reply'; message: TOutput }\n | { action: 'Send'; message: WorkflowCommand<TOutput> }\n | { action: 'Publish'; message: WorkflowEvent<TOutput> }\n | {\n action: 'Schedule';\n message: TOutput;\n when: { afterInMs: number } | { at: Date };\n }\n | { action: 'Complete' }\n | { action: 'Accept' }\n | { action: 'Ignore'; reason: string }\n | { action: 'Error'; reason: string };\n\nexport const reply = <TOutput extends AnyEvent | AnyCommand>(\n message: TOutput,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Reply',\n message,\n };\n};\n\nexport const send = <TOutput extends AnyEvent | AnyCommand>(\n message: WorkflowCommand<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Send',\n message,\n };\n};\n\nexport const publish = <TOutput extends AnyEvent | AnyCommand>(\n message: WorkflowEvent<TOutput>,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Publish',\n message,\n };\n};\n\nexport const schedule = <TOutput extends AnyEvent | AnyCommand>(\n message: TOutput,\n when: { afterInMs: number } | { at: Date },\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Schedule',\n message,\n when,\n };\n};\n\nexport const complete = <\n TOutput extends AnyEvent | AnyCommand,\n>(): WorkflowOutput<TOutput> => {\n return {\n action: 'Complete',\n };\n};\n\nexport const ignore = <TOutput extends AnyEvent | AnyCommand>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Ignore',\n reason,\n };\n};\n\nexport const error = <TOutput extends AnyEvent | AnyCommand>(\n reason: string,\n): WorkflowOutput<TOutput> => {\n return {\n action: 'Error',\n reason,\n };\n};\n\nexport const accept = <\n TOutput extends AnyEvent | AnyCommand,\n>(): WorkflowOutput<TOutput> => {\n return { action: 'Accept' };\n};\n","import {\n type BatchRecordedMessageHandlerWithContext,\n type BatchRecordedMessageHandlerWithoutContext,\n type DefaultRecord,\n type Event,\n type ReadEvent,\n} from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\nexport type BeforeEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","import type { EventsPublisher } from '../../messageBus';\nimport type { DefaultRecord, Event, ReadEvent } from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\nimport type { AfterEventStoreCommitHandler } from './afterEventStoreCommitHandler';\n\nexport const forwardToMessageBus = <\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n eventPublisher: EventsPublisher,\n): AfterEventStoreCommitHandler<Store, HandlerContext> =>\n (async (\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n ): Promise<void> => {\n for (const message of messages) {\n await eventPublisher.publish(message);\n }\n }) as AfterEventStoreCommitHandler<Store, HandlerContext>;\n","import {\n event,\n type Event,\n type EventDataOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\n\nexport const GlobalStreamCaughtUpType = '__emt:GlobalStreamCaughtUp';\n\nexport type GlobalStreamCaughtUp = Event<\n '__emt:GlobalStreamCaughtUp',\n { globalPosition: bigint },\n { globalPosition: bigint }\n>;\n\nexport const isGlobalStreamCaughtUp = (\n event: Event,\n): event is GlobalStreamCaughtUp => event.type === GlobalStreamCaughtUpType;\n\nexport const caughtUpEventFrom =\n (position: bigint) =>\n (\n event: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n ): event is ReadEvent<\n GlobalStreamCaughtUp,\n ReadEventMetadataWithGlobalPosition\n > =>\n event.type === GlobalStreamCaughtUpType &&\n event.metadata?.globalPosition >= position;\n\nexport const globalStreamCaughtUp = (\n data: EventDataOf<GlobalStreamCaughtUp>,\n): GlobalStreamCaughtUp =>\n event<GlobalStreamCaughtUp>(GlobalStreamCaughtUpType, data, {\n globalPosition: data.globalPosition,\n });\n\nexport const isSubscriptionEvent = (\n event: Event,\n): event is GlobalSubscriptionEvent => isGlobalStreamCaughtUp(event);\n\nexport const isNotInternalEvent = (event: Event): boolean =>\n !isGlobalStreamCaughtUp(event);\n\nexport type GlobalSubscriptionEvent = GlobalStreamCaughtUp;\n","//import type { ReadableStream } from 'web-streams-polyfill';\nimport type {\n AnyReadEventMetadata,\n BigIntGlobalPosition,\n BigIntStreamPosition,\n CommonReadEventMetadata,\n DefaultRecord,\n Event,\n GlobalPositionTypeOfReadEventMetadata,\n ReadEvent,\n ReadEventMetadata,\n StreamPositionTypeOfReadEventMetadata,\n WithGlobalPosition,\n} from '../typing';\nimport type { AfterEventStoreCommitHandler } from './afterCommit';\n//import type { GlobalSubscriptionEvent } from './events';\nimport type { ExpectedStreamVersion } from './expectedVersion';\n\n// #region event-store\nexport interface EventStore<\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> {\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType, ReadEventMetadataType>,\n ): Promise<\n AggregateStreamResult<\n State,\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<ReadStreamResult<EventType, ReadEventMetadataType>>;\n\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >,\n ): Promise<\n AppendToStreamResult<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >\n >;\n\n // streamEvents(): ReadableStream<\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // >;\n}\n\nexport type EventStoreReadEventMetadata<Store extends EventStore> =\n Store extends EventStore<infer T>\n ? T extends CommonReadEventMetadata<infer SP>\n ? T extends WithGlobalPosition<infer GP>\n ? ReadEventMetadata<GP, SP> & T\n : ReadEventMetadata<undefined, SP> & T\n : never\n : never;\n\nexport type GlobalPositionTypeOfEventStore<Store extends EventStore> =\n GlobalPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type StreamPositionTypeOfEventStore<Store extends EventStore> =\n StreamPositionTypeOfReadEventMetadata<EventStoreReadEventMetadata<Store>>;\n\nexport type EventStoreSession<EventStoreType extends EventStore> = {\n eventStore: EventStoreType;\n close: () => Promise<void>;\n};\n\nexport interface EventStoreSessionFactory<EventStoreType extends EventStore> {\n withSession<T = unknown>(\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n ): Promise<T>;\n}\n// #endregion event-store\n\nexport const canCreateEventStoreSession = <Store extends EventStore>(\n eventStore: Store | EventStoreSessionFactory<Store>,\n): eventStore is EventStoreSessionFactory<Store> => 'withSession' in eventStore;\n\nexport const nulloSessionFactory = <EventStoreType extends EventStore>(\n eventStore: EventStoreType,\n): EventStoreSessionFactory<EventStoreType> => ({\n withSession: (callback) => {\n const nulloSession: EventStoreSession<EventStoreType> = {\n eventStore,\n close: () => Promise.resolve(),\n };\n\n return callback(nulloSession);\n },\n});\n\n////////////////////////////////////////////////////////////////////\n/// ReadStream types\n////////////////////////////////////////////////////////////////////\n\nexport type ReadStreamOptions<StreamVersion = BigIntStreamPosition> = (\n | {\n from: StreamVersion;\n }\n | { to: StreamVersion }\n | { from: StreamVersion; maxCount?: bigint }\n | {\n expectedStreamVersion: ExpectedStreamVersion<StreamVersion>;\n }\n) & {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type ReadStreamResult<\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n currentStreamVersion: StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>;\n events: ReadEvent<EventType, ReadEventMetadataType>[];\n streamExists: boolean;\n};\n\n////////////////////////////////////////////////////////////////////\n/// AggregateStream types\n////////////////////////////////////////////////////////////////////\n\ntype Evolve<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> =\n | ((currentState: State, event: EventType) => State)\n | ((\n currentState: State,\n event: ReadEvent<EventType, ReadEventMetadataType>,\n ) => State)\n | ((currentState: State, event: ReadEvent<EventType>) => State);\n\nexport type AggregateStreamOptions<\n State,\n EventType extends Event,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = {\n evolve: Evolve<State, EventType, ReadEventMetadataType>;\n initialState: () => State;\n read?: ReadStreamOptions<\n StreamPositionTypeOfReadEventMetadata<ReadEventMetadataType>\n >;\n};\n\nexport type AggregateStreamResult<\n State,\n StreamPosition = BigIntStreamPosition,\n> = {\n currentStreamVersion: StreamPosition;\n state: State;\n streamExists: boolean;\n};\n\nexport type AggregateStreamResultWithGlobalPosition<\n State,\n StreamPosition = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> =\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: true;\n lastEventGlobalPosition: GlobalPosition;\n })\n | (AggregateStreamResult<State, StreamPosition> & {\n streamExists: false;\n });\n\nexport type AggregateStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['aggregateStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// AppendToStream types\n////////////////////////////////////////////////////////////////////\n\nexport type AppendToStreamOptions<StreamVersion = BigIntStreamPosition> = {\n expectedStreamVersion?: ExpectedStreamVersion<StreamVersion>;\n};\n\nexport type AppendToStreamResult<StreamVersion = BigIntStreamPosition> = {\n nextExpectedStreamVersion: StreamVersion;\n createdNewStream: boolean;\n};\n\nexport type AppendToStreamResultWithGlobalPosition<\n StreamVersion = BigIntStreamPosition,\n GlobalPosition = BigIntGlobalPosition,\n> = AppendToStreamResult<StreamVersion> & {\n lastEventGlobalPosition: GlobalPosition;\n};\n\nexport type AppendStreamResultOfEventStore<Store extends EventStore> =\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Store['appendToStream'] extends (...args: any[]) => Promise<infer R>\n ? R\n : never;\n\n////////////////////////////////////////////////////////////////////\n/// DefaultEventStoreOptions\n////////////////////////////////////////////////////////////////////\n\nexport type DefaultEventStoreOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n /**\n * Pluggable set of hooks informing about the event store internal behaviour.\n */\n hooks?: {\n /**\n * This hook will be called **AFTER** events were stored in the event store.\n * It's designed to handle scenarios where delivery and ordering guarantees do not matter much.\n *\n * **WARNINGS:**\n *\n * 1. It will be called **EXACTLY ONCE** if append succeded.\n * 2. If the hook fails, its append **will still silently succeed**, and no error will be thrown.\n * 3. Wen process crashes after events were committed, but before the hook was called, delivery won't be retried.\n * That can lead to state inconsistencies.\n * 4. In the case of high concurrent traffic, **race conditions may cause ordering issues**.\n * For instance, where the second hook takes longer to process than the first one, ordering won't be guaranteed.\n *\n * @type {AfterEventStoreCommitHandler<Store, HandlerContext>}\n */\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n };\n};\n","import { ConcurrencyError } from '../errors';\nimport type { BigIntStreamPosition, Flavour } from '../typing';\n\nexport type ExpectedStreamVersion<VersionType = BigIntStreamPosition> =\n | ExpectedStreamVersionWithValue<VersionType>\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue<VersionType = BigIntStreamPosition> =\n Flavour<VersionType, 'StreamVersion'>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = <StreamVersion = BigIntStreamPosition>(\n current: StreamVersion | undefined,\n expected: ExpectedStreamVersion<StreamVersion>,\n defaultVersion: StreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = <\n StreamVersion = BigIntStreamPosition,\n>(\n current: StreamVersion,\n expected: ExpectedStreamVersion<StreamVersion> | undefined,\n defaultVersion: StreamVersion,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError<\n VersionType = BigIntStreamPosition,\n> extends ConcurrencyError {\n constructor(\n current: VersionType,\n expected: ExpectedStreamVersion<VersionType>,\n ) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError;\n","import { v4 as uuid } from 'uuid';\nimport {\n getInMemoryDatabase,\n type InMemoryDatabase,\n} from '../database/inMemoryDatabase';\nimport type { ProjectionRegistration } from '../projections';\nimport type {\n BigIntStreamPosition,\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport {\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultEventStoreOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { handleInMemoryProjections } from './projections/inMemory';\nimport { StreamingCoordinator } from './subscriptions';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition> & {\n database: InMemoryDatabase;\n };\n\nexport type InMemoryReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type InMemoryProjectionHandlerContext = {\n eventStore?: InMemoryEventStore;\n database?: InMemoryDatabase;\n};\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore> & {\n projections?: ProjectionRegistration<\n 'inline',\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >[];\n database?: InMemoryDatabase;\n };\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n const streamingCoordinator = StreamingCoordinator();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n // Get the database instance to be used for projections\n const database = eventStoreOptions?.database || getInMemoryDatabase();\n\n // Extract inline projections from options\n const inlineProjections = (eventStoreOptions?.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n // Create the event store object\n const eventStore: InMemoryEventStore = {\n database,\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType>(streamName, read);\n\n const events = result?.events ?? [];\n\n return {\n currentStreamVersion: BigInt(events.length),\n state: events.reduce(evolve, initialState()),\n streamExists: result.streamExists,\n };\n },\n\n readStream: <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options && 'from' in options ? options.from : 0);\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : (events?.length ?? 1),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? events\n .map(\n (e) =>\n e as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n )\n .slice(from, to)\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition: BigInt(getAllEventsCount() + index + 1),\n };\n return {\n ...event,\n kind: event.kind ?? 'Event',\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [...currentEvents, ...newEvents]);\n await streamingCoordinator.notify(newEvents);\n\n // Process projections if there are any registered\n if (inlineProjections.length > 0) {\n await handleInMemoryProjections({\n projections: inlineProjections,\n events: newEvents,\n database: eventStore.database,\n eventStore,\n });\n }\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit<InMemoryEventStore>(\n newEvents,\n eventStoreOptions?.hooks,\n );\n\n return result;\n },\n\n //streamEvents: streamingCoordinator.stream,\n };\n\n return eventStore;\n};\n","import { v7 as uuid } from 'uuid';\nimport { deepEquals } from '../utils';\nimport {\n type DatabaseHandleOptionErrors,\n type DatabaseHandleOptions,\n type DatabaseHandleResult,\n type DeleteResult,\n type Document,\n type DocumentHandler,\n type InsertOneResult,\n type OptionalUnlessRequiredIdAndVersion,\n type ReplaceOneOptions,\n type UpdateResult,\n type WithIdAndVersion,\n type WithoutId,\n} from './types';\nimport { expectedVersionValue, operationResult } from './utils';\n\nexport interface InMemoryDocumentsCollection<T extends Document> {\n handle: (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ) => Promise<DatabaseHandleResult<T>>;\n findOne: (predicate?: Predicate<T>) => Promise<T | null>;\n find: (predicate?: Predicate<T>) => Promise<T[]>;\n insertOne: (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ) => Promise<InsertOneResult>;\n deleteOne: (predicate?: Predicate<T>) => Promise<DeleteResult>;\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ) => Promise<UpdateResult>;\n}\n\nexport interface InMemoryDatabase {\n collection: <T extends Document>(\n name: string,\n ) => InMemoryDocumentsCollection<T>;\n}\n\ntype Predicate<T> = (item: T) => boolean;\ntype CollectionName = string;\n\nexport const getInMemoryDatabase = (): InMemoryDatabase => {\n const storage = new Map<CollectionName, WithIdAndVersion<Document>[]>();\n\n return {\n collection: <T extends Document, CollectionName extends string>(\n collectionName: CollectionName,\n collectionOptions: {\n errors?: DatabaseHandleOptionErrors;\n } = {},\n ): InMemoryDocumentsCollection<T> => {\n const ensureCollectionCreated = () => {\n if (!storage.has(collectionName)) storage.set(collectionName, []);\n };\n\n const errors = collectionOptions.errors;\n\n const collection = {\n collectionName,\n insertOne: async (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ): Promise<InsertOneResult> => {\n ensureCollectionCreated();\n\n const _id = (document._id as string | undefined | null) ?? uuid();\n const _version = document._version ?? 1n;\n\n const existing = await collection.findOne((c) => c._id === _id);\n\n if (existing) {\n return operationResult<InsertOneResult>(\n {\n successful: false,\n insertedId: null,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n }\n\n const documentsInCollection = storage.get(collectionName)!;\n const newDocument = { ...document, _id, _version };\n const newCollection = [...documentsInCollection, newDocument];\n storage.set(collectionName, newCollection);\n\n return operationResult<InsertOneResult>(\n {\n successful: true,\n insertedId: _id,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n },\n findOne: (predicate?: Predicate<T>): Promise<T | null> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n const firstOne = filteredDocuments?.[0] ?? null;\n\n return Promise.resolve(firstOne as T | null);\n },\n find: (predicate?: Predicate<T>): Promise<T[]> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n return Promise.resolve(filteredDocuments as T[]);\n },\n deleteOne: (predicate?: Predicate<T>): Promise<DeleteResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n if (predicate) {\n const foundIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (foundIndex === -1) {\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: false,\n matchedCount: 0,\n deletedCount: 0,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n } else {\n const newCollection = documentsInCollection.toSpliced(\n foundIndex,\n 1,\n );\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n }\n }\n\n const newCollection = documentsInCollection.slice(1);\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n },\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ): Promise<UpdateResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n const firstIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (firstIndex === undefined || firstIndex === -1) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 0,\n modifiedCount: 0,\n nextExpectedVersion: 0n,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const existing = documentsInCollection[firstIndex]!;\n\n if (\n typeof options?.expectedVersion === 'bigint' &&\n existing._version !== options.expectedVersion\n ) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 1,\n modifiedCount: 0,\n nextExpectedVersion: existing._version,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const newVersion = existing._version + 1n;\n\n const newCollection = documentsInCollection.with(firstIndex, {\n _id: existing._id,\n ...document,\n _version: newVersion,\n });\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: true,\n modifiedCount: 1,\n matchedCount: firstIndex,\n nextExpectedVersion: newVersion,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n },\n handle: async (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ): Promise<DatabaseHandleResult<T>> => {\n const { expectedVersion: version, ...operationOptions } =\n options ?? {};\n ensureCollectionCreated();\n const existing = await collection.findOne(({ _id }) => _id === id);\n\n const expectedVersion = expectedVersionValue(version);\n\n if (\n (existing == null && version === 'DOCUMENT_EXISTS') ||\n (existing == null && expectedVersion != null) ||\n (existing != null && version === 'DOCUMENT_DOES_NOT_EXIST') ||\n (existing != null &&\n expectedVersion !== null &&\n existing._version !== expectedVersion)\n ) {\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: false,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n }\n\n const result = handle(existing !== null ? { ...existing } : null);\n\n if (deepEquals(existing, result))\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n\n if (!existing && result) {\n const newDoc = { ...result, _id: id };\n const insertResult = await collection.insertOne({\n ...newDoc,\n _id: id,\n } as OptionalUnlessRequiredIdAndVersion<T>);\n return {\n ...insertResult,\n document: {\n ...newDoc,\n _version: insertResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n if (existing && !result) {\n const deleteResult = await collection.deleteOne(\n ({ _id }) => id === _id,\n );\n return { ...deleteResult, document: null };\n }\n\n if (existing && result) {\n const replaceResult = await collection.replaceOne(\n ({ _id }) => id === _id,\n result,\n {\n ...operationOptions,\n expectedVersion: expectedVersion ?? 'DOCUMENT_EXISTS',\n },\n );\n return {\n ...replaceResult,\n document: {\n ...result,\n _version: replaceResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n },\n };\n\n return collection;\n },\n };\n};\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","const isPrimitive = (value: unknown): boolean => {\n const type = typeof value;\n return (\n value === null ||\n value === undefined ||\n type === 'boolean' ||\n type === 'number' ||\n type === 'string' ||\n type === 'symbol' ||\n type === 'bigint'\n );\n};\n\nconst compareArrays = <T>(left: T[], right: T[]): boolean => {\n if (left.length !== right.length) {\n return false;\n }\n for (let i = 0; i < left.length; i++) {\n const leftHas = i in left;\n const rightHas = i in right;\n if (leftHas !== rightHas) return false;\n if (leftHas && !deepEquals(left[i], right[i])) return false;\n }\n return true;\n};\n\nconst compareDates = (left: Date, right: Date): boolean => {\n return left.getTime() === right.getTime();\n};\n\nconst compareRegExps = (left: RegExp, right: RegExp): boolean => {\n return left.toString() === right.toString();\n};\n\nconst compareErrors = (left: Error, right: Error): boolean => {\n if (left.message !== right.message || left.name !== right.name) {\n return false;\n }\n const leftKeys = Object.keys(left);\n const rightKeys = Object.keys(right);\n if (leftKeys.length !== rightKeys.length) return false;\n const rightKeySet = new Set(rightKeys);\n for (const key of leftKeys) {\n if (!rightKeySet.has(key)) return false;\n // @ts-expect-error - accessing dynamic keys\n if (!deepEquals(left[key], right[key])) return false;\n }\n return true;\n};\n\nconst compareMaps = (\n left: Map<unknown, unknown>,\n right: Map<unknown, unknown>,\n): boolean => {\n if (left.size !== right.size) return false;\n\n for (const [key, value] of left) {\n if (isPrimitive(key)) {\n if (!right.has(key) || !deepEquals(value, right.get(key))) {\n return false;\n }\n } else {\n let found = false;\n for (const [rightKey, rightValue] of right) {\n if (deepEquals(key, rightKey) && deepEquals(value, rightValue)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareSets = (left: Set<unknown>, right: Set<unknown>): boolean => {\n if (left.size !== right.size) return false;\n\n for (const leftItem of left) {\n if (isPrimitive(leftItem)) {\n if (!right.has(leftItem)) return false;\n } else {\n let found = false;\n for (const rightItem of right) {\n if (deepEquals(leftItem, rightItem)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareArrayBuffers = (\n left: ArrayBuffer,\n right: ArrayBuffer,\n): boolean => {\n if (left.byteLength !== right.byteLength) return false;\n const leftView = new Uint8Array(left);\n const rightView = new Uint8Array(right);\n for (let i = 0; i < leftView.length; i++) {\n if (leftView[i] !== rightView[i]) return false;\n }\n return true;\n};\n\nconst compareTypedArrays = (\n left: ArrayBufferView,\n right: ArrayBufferView,\n): boolean => {\n if (left.constructor !== right.constructor) return false;\n if (left.byteLength !== right.byteLength) return false;\n\n const leftArray = new Uint8Array(\n left.buffer,\n left.byteOffset,\n left.byteLength,\n );\n const rightArray = new Uint8Array(\n right.buffer,\n right.byteOffset,\n right.byteLength,\n );\n\n for (let i = 0; i < leftArray.length; i++) {\n if (leftArray[i] !== rightArray[i]) return false;\n }\n return true;\n};\n\nconst compareObjects = (\n left: Record<string, unknown>,\n right: Record<string, unknown>,\n): boolean => {\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (keys1.length !== keys2.length) {\n return false;\n }\n\n for (const key of keys1) {\n if (left[key] instanceof Function && right[key] instanceof Function) {\n continue;\n }\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nconst getType = (value: unknown): string => {\n if (value === null) return 'null';\n if (value === undefined) return 'undefined';\n\n const primitiveType = typeof value;\n if (primitiveType !== 'object') return primitiveType;\n\n if (Array.isArray(value)) return 'array';\n if (value instanceof Boolean) return 'boxed-boolean';\n if (value instanceof Number) return 'boxed-number';\n if (value instanceof String) return 'boxed-string';\n if (value instanceof Date) return 'date';\n if (value instanceof RegExp) return 'regexp';\n if (value instanceof Error) return 'error';\n if (value instanceof Map) return 'map';\n if (value instanceof Set) return 'set';\n if (value instanceof ArrayBuffer) return 'arraybuffer';\n if (value instanceof DataView) return 'dataview';\n if (value instanceof WeakMap) return 'weakmap';\n if (value instanceof WeakSet) return 'weakset';\n\n if (ArrayBuffer.isView(value)) return 'typedarray';\n\n return 'object';\n};\n\nexport const deepEquals = <T>(left: T, right: T): boolean => {\n if (left === right) return true;\n\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n const leftType = getType(left);\n const rightType = getType(right);\n\n if (leftType !== rightType) return false;\n\n switch (leftType) {\n case 'null':\n case 'undefined':\n case 'boolean':\n case 'number':\n case 'bigint':\n case 'string':\n case 'symbol':\n case 'function':\n return left === right;\n\n case 'array':\n return compareArrays(left as unknown[], right as unknown[]);\n\n case 'date':\n return compareDates(left as Date, right as Date);\n\n case 'regexp':\n return compareRegExps(left as RegExp, right as RegExp);\n\n case 'error':\n return compareErrors(left as Error, right as Error);\n\n case 'map':\n return compareMaps(\n left as Map<unknown, unknown>,\n right as Map<unknown, unknown>,\n );\n\n case 'set':\n return compareSets(left as Set<unknown>, right as Set<unknown>);\n\n case 'arraybuffer':\n return compareArrayBuffers(left as ArrayBuffer, right as ArrayBuffer);\n\n case 'dataview':\n case 'weakmap':\n case 'weakset':\n return false;\n\n case 'typedarray':\n return compareTypedArrays(\n left as ArrayBufferView,\n right as ArrayBufferView,\n );\n\n case 'boxed-boolean':\n return (left as boolean).valueOf() === (right as boolean).valueOf();\n\n case 'boxed-number':\n return (left as number).valueOf() === (right as number).valueOf();\n\n case 'boxed-string':\n return (left as string).valueOf() === (right as string).valueOf();\n\n case 'object':\n return compareObjects(\n left as Record<string, unknown>,\n right as Record<string, unknown>,\n );\n\n default:\n return false;\n }\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left !== null &&\n left !== undefined &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export const sum = (\n iterator: Iterator<number, number, number> | Iterator<number>,\n) => {\n let value,\n done: boolean | undefined,\n sum = 0;\n do {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n ({ value, done } = iterator.next());\n sum += value || 0;\n } while (!done);\n return sum;\n};\n","import { EmmettError } from '../errors';\n\nexport type TaskQueue = TaskQueueItem[];\n\nexport type TaskQueueItem = {\n task: () => Promise<void>;\n options?: EnqueueTaskOptions;\n};\n\nexport type TaskProcessorOptions = {\n maxActiveTasks: number;\n maxQueueSize: number;\n maxTaskIdleTime?: number;\n};\n\nexport type Task<T> = (context: TaskContext) => Promise<T>;\n\nexport type TaskContext = {\n ack: () => void;\n};\n\nexport type EnqueueTaskOptions = { taskGroupId?: string };\n\nexport class TaskProcessor {\n private queue: TaskQueue = [];\n private isProcessing = false;\n private activeTasks = 0;\n private activeGroups: Set<string> = new Set();\n\n constructor(private options: TaskProcessorOptions) {}\n\n enqueue<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n if (this.queue.length >= this.options.maxQueueSize) {\n return Promise.reject(\n new EmmettError(\n 'Too many pending connections. Please try again later.',\n ),\n );\n }\n\n return this.schedule(task, options);\n }\n\n waitForEndOfProcessing(): Promise<void> {\n return this.schedule(({ ack }) => Promise.resolve(ack()));\n }\n\n private schedule<T>(task: Task<T>, options?: EnqueueTaskOptions): Promise<T> {\n return promiseWithDeadline(\n (resolve, reject) => {\n const taskWithContext = () => {\n return new Promise<void>((resolveTask, failTask) => {\n const taskPromise = task({\n ack: resolveTask,\n });\n\n taskPromise.then(resolve).catch((err) => {\n // eslint-disable-next-line @typescript-eslint/prefer-promise-reject-errors\n failTask(err);\n reject(err);\n });\n });\n };\n\n this.queue.push({ task: taskWithContext, options });\n if (!this.isProcessing) {\n this.ensureProcessing();\n }\n },\n { deadline: this.options.maxTaskIdleTime },\n );\n }\n\n private ensureProcessing(): void {\n if (this.isProcessing) return;\n this.isProcessing = true;\n this.processQueue();\n }\n\n private processQueue(): void {\n try {\n while (\n this.activeTasks < this.options.maxActiveTasks &&\n this.queue.length > 0\n ) {\n const item = this.takeFirstAvailableItem();\n\n if (item === null) return;\n\n const groupId = item.options?.taskGroupId;\n\n if (groupId) {\n // Mark the group as active\n this.activeGroups.add(groupId);\n }\n\n this.activeTasks++;\n void this.executeItem(item);\n }\n } catch (error) {\n console.error(error);\n throw error;\n } finally {\n this.isProcessing = false;\n if (\n this.hasItemsToProcess() &&\n this.activeTasks < this.options.maxActiveTasks\n ) {\n this.ensureProcessing();\n }\n }\n }\n\n private async executeItem({ task, options }: TaskQueueItem): Promise<void> {\n try {\n await task();\n } finally {\n this.activeTasks--;\n\n // Mark the group as inactive after task completion\n if (options && options.taskGroupId) {\n this.activeGroups.delete(options.taskGroupId);\n }\n\n this.ensureProcessing();\n }\n }\n\n private takeFirstAvailableItem = (): TaskQueueItem | null => {\n const taskIndex = this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n );\n\n if (taskIndex === -1) {\n // All remaining tasks are blocked by active groups\n return null;\n }\n\n // Remove the task from the queue\n const [item] = this.queue.splice(taskIndex, 1);\n\n return item ?? null;\n };\n\n private hasItemsToProcess = (): boolean =>\n this.queue.findIndex(\n (item) =>\n !item.options?.taskGroupId ||\n !this.activeGroups.has(item.options.taskGroupId),\n ) !== -1;\n}\n\nconst DEFAULT_PROMISE_DEADLINE = 2147483647;\n\nconst promiseWithDeadline = <T>(\n executor: (\n resolve: (value: T | PromiseLike<T>) => void,\n reject: (reason?: unknown) => void,\n ) => void,\n options: { deadline?: number },\n) => {\n return new Promise<T>((resolve, reject) => {\n let taskStarted = false;\n\n const maxWaitingTime = options.deadline || DEFAULT_PROMISE_DEADLINE;\n\n let timeoutId: NodeJS.Timeout | null = setTimeout(() => {\n if (!taskStarted) {\n reject(\n new Error('Task was not started within the maximum waiting time'),\n );\n }\n }, maxWaitingTime);\n\n executor((value) => {\n taskStarted = true;\n if (timeoutId) {\n clearTimeout(timeoutId);\n }\n timeoutId = null;\n resolve(value);\n }, reject);\n });\n};\n","import { TaskProcessor } from '../../taskProcessing';\n\nexport type LockOptions = { lockId: number };\n\nexport type AcquireLockOptions = { lockId: string };\nexport type ReleaseLockOptions = { lockId: string };\n\nexport type Lock = {\n acquire(options: AcquireLockOptions): Promise<void>;\n tryAcquire(options: AcquireLockOptions): Promise<boolean>;\n release(options: ReleaseLockOptions): Promise<boolean>;\n withAcquire: <Result = unknown>(\n handle: () => Promise<Result>,\n options: AcquireLockOptions,\n ) => Promise<Result>;\n};\n\nexport const InProcessLock = (): Lock => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: Number.MAX_VALUE,\n maxQueueSize: Number.MAX_VALUE,\n });\n\n // Map to store ack functions of currently held locks: lockId -> ack()\n const locks = new Map<string, () => void>();\n\n return {\n async acquire({ lockId }: AcquireLockOptions): Promise<void> {\n // If the lock is already held, we just queue up another task in the same group.\n // TaskProcessor ensures tasks in the same group run one at a time.\n await new Promise<void>((resolve, reject) => {\n taskProcessor\n .enqueue(\n ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n // We do NOT call ack() here. We hold onto the lock.\n resolve();\n return Promise.resolve();\n },\n { taskGroupId: lockId },\n )\n .catch(reject);\n });\n },\n\n async tryAcquire({ lockId }: AcquireLockOptions): Promise<boolean> {\n // If lock is already held, fail immediately\n if (locks.has(lockId)) {\n return false;\n }\n\n // TODO: Check pending queue\n await this.acquire({ lockId });\n\n return true;\n },\n\n release({ lockId }: ReleaseLockOptions): Promise<boolean> {\n const ack = locks.get(lockId);\n if (ack === undefined) {\n return Promise.resolve(true);\n }\n locks.delete(lockId);\n ack();\n return Promise.resolve(true);\n },\n\n async withAcquire<Result = unknown>(\n handle: () => Promise<Result>,\n { lockId }: AcquireLockOptions,\n ): Promise<Result> {\n return taskProcessor.enqueue(\n async ({ ack }) => {\n // When this task starts, it means the previous lock (if any) was released\n // and now we have exclusive access.\n locks.set(lockId, ack);\n\n // We do NOT call ack() here. We hold onto the lock.\n try {\n return await handle();\n } finally {\n locks.delete(lockId);\n ack();\n }\n },\n { taskGroupId: lockId },\n );\n },\n };\n};\n","export const delay = (ms: number): Promise<void> => {\n return new Promise((resolve) => setTimeout(resolve, ms));\n};\n\nexport type AsyncAwaiter<T = void> = {\n wait: Promise<T>;\n resolve: (value: T | PromiseLike<T>) => void;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n reject: (reason?: any) => void;\n reset: () => void;\n};\n\n// TODO: Remove this after migrating to Node 22\nexport const asyncAwaiter = <T = void>(): AsyncAwaiter<T> => {\n const result: AsyncAwaiter<T> = {} as AsyncAwaiter<T>;\n\n (result.reset = () => {\n result.wait = new Promise<T>((res, rej) => {\n result.resolve = res;\n result.reject = rej;\n });\n })();\n\n return result;\n};\n","import retry from 'async-retry';\nimport { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\n\nexport type AsyncRetryOptions<T = unknown> = retry.Options & {\n shouldRetryResult?: (result: T) => boolean;\n shouldRetryError?: (error?: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions<T>,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n const result = await fn();\n\n if (opts?.shouldRetryResult && opts.shouldRetryResult(result)) {\n throw new EmmettError(\n `Retrying because of result: ${JSONParser.stringify(result)}`,\n );\n }\n return result;\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n return undefined as unknown as T;\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","import { ConcurrencyInMemoryDatabaseError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n DatabaseHandleOptionErrors,\n ExpectedDocumentVersion,\n ExpectedDocumentVersionGeneral,\n ExpectedDocumentVersionValue,\n OperationResult,\n} from './types';\n\nexport const isGeneralExpectedDocumentVersion = (\n version: ExpectedDocumentVersion | undefined,\n): version is ExpectedDocumentVersionGeneral => {\n return (\n version === 'DOCUMENT_DOES_NOT_EXIST' ||\n version === 'DOCUMENT_EXISTS' ||\n version === 'NO_CONCURRENCY_CHECK'\n );\n};\n\nexport const expectedVersionValue = (\n version: ExpectedDocumentVersion | undefined,\n): ExpectedDocumentVersionValue | null =>\n version === undefined || isGeneralExpectedDocumentVersion(version)\n ? null\n : version;\n\nexport const operationResult = <T extends OperationResult>(\n result: Omit<T, 'assertSuccess' | 'acknowledged' | 'assertSuccessful'>,\n options: {\n operationName: string;\n collectionName: string;\n errors?: DatabaseHandleOptionErrors;\n },\n): T => {\n const operationResult: T = {\n ...result,\n acknowledged: true,\n successful: result.successful,\n assertSuccessful: (errorMessage?: string) => {\n const { successful } = result;\n const { operationName, collectionName } = options;\n\n if (!successful)\n throw new ConcurrencyInMemoryDatabaseError(\n errorMessage ??\n `${operationName} on ${collectionName} failed. Expected document state does not match current one! Result: ${JSONParser.stringify(result)}!`,\n );\n },\n } as T;\n\n if (options.errors?.throwOnOperationFailures)\n operationResult.assertSuccessful();\n\n return operationResult;\n};\n","import type { InMemoryDatabase } from '../../../database/inMemoryDatabase';\nimport type {\n ProjectionDefinition,\n TruncateProjection,\n} from '../../../projections';\nimport type { CanHandle, Event, ReadEvent } from '../../../typing';\nimport {\n type InMemoryProjectionHandlerContext,\n type InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\nexport const DATABASE_REQUIRED_ERROR_MESSAGE =\n 'Database is required in context for InMemory projections';\n\nexport type InMemoryProjectionDefinition<EventType extends Event> =\n ProjectionDefinition<\n EventType,\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >;\n\nexport type InMemoryProjectionHandlerOptions<EventType extends Event = Event> =\n {\n projections: InMemoryProjectionDefinition<EventType>[];\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[];\n database: InMemoryDatabase;\n eventStore?: InMemoryProjectionHandlerContext['eventStore'];\n };\n\n/**\n * Handles projections for the InMemoryEventStore\n * Similar to the PostgreSQL implementation, this processes events through projections\n */\nexport const handleInMemoryProjections = async <\n EventType extends Event = Event,\n>(\n options: InMemoryProjectionHandlerOptions<EventType>,\n): Promise<void> => {\n const { projections, events, database, eventStore } = options;\n\n // Get all event types from the events batch to filter projections\n const eventTypes = events.map((e) => e.type);\n\n // Filter projections that can handle these event types\n const relevantProjections = projections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n // Process each projection\n for (const projection of relevantProjections) {\n await projection.handle(events, {\n eventStore,\n database,\n });\n }\n};\n\nexport type InMemoryWithNotNullDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = (\n document: DocumentType,\n event: ReadEvent<EventType, InMemoryReadEventMetadata>,\n) => DocumentType | null;\n\nexport type InMemoryWithNullableDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = (\n document: DocumentType | null,\n event: ReadEvent<EventType, InMemoryReadEventMetadata>,\n) => DocumentType | null;\n\nexport type InMemoryDocumentEvolve<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> =\n | InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>\n | InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n\nexport type InMemoryProjectionOptions<EventType extends Event> = {\n handle: (\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[],\n context: InMemoryProjectionHandlerContext & { database: InMemoryDatabase },\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n truncate?: TruncateProjection<\n InMemoryProjectionHandlerContext & { database: InMemoryDatabase }\n >;\n};\n\n/**\n * Creates an InMemory projection\n */\nexport const inMemoryProjection = <EventType extends Event>({\n truncate,\n handle,\n canHandle,\n}: InMemoryProjectionOptions<EventType>): InMemoryProjectionDefinition<EventType> => ({\n canHandle,\n handle: async (events, context) => {\n if (!context.database) {\n throw new Error(DATABASE_REQUIRED_ERROR_MESSAGE);\n }\n await handle(events, {\n ...context,\n database: context.database,\n });\n },\n truncate: truncate\n ? (context) => {\n if (!context.database) {\n throw new Error(DATABASE_REQUIRED_ERROR_MESSAGE);\n }\n return truncate({\n ...context,\n database: context.database,\n });\n }\n : undefined,\n});\n\n/**\n * Creates a multi-stream projection for InMemoryDatabase\n */\nexport type InMemoryMultiStreamProjectionOptions<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = {\n canHandle: CanHandle<EventType>;\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n} & (\n | {\n evolve: InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n }\n | {\n evolve: InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>;\n initialState: () => DocumentType;\n }\n);\n\n/**\n * Creates a projection that handles events across multiple streams\n */\nexport const inMemoryMultiStreamProjection = <\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n>(\n options: InMemoryMultiStreamProjectionOptions<DocumentType, EventType>,\n): InMemoryProjectionDefinition<EventType> => {\n const { collectionName, getDocumentId, canHandle } = options;\n\n return inMemoryProjection({\n handle: async (\n events: ReadEvent<EventType, InMemoryReadEventMetadata>[],\n { database },\n ) => {\n const collection = database.collection<DocumentType>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), (document) => {\n if ('initialState' in options) {\n return options.evolve(document ?? options.initialState(), event);\n } else {\n return options.evolve(document, event);\n }\n });\n }\n },\n canHandle,\n truncate: async ({\n database,\n }: InMemoryProjectionHandlerContext & { database: InMemoryDatabase }) => {\n // For InMemory database, we can't directly truncate a collection\n // So we'll delete all documents from the collection\n const collection = database.collection<DocumentType>(collectionName);\n const documents = await collection.find();\n\n for (const doc of documents) {\n if (doc && '_id' in doc) {\n const id = doc._id;\n await collection.deleteOne((d) => d._id === id);\n }\n }\n },\n });\n};\n\n/**\n * Creates a single-stream projection for InMemoryDatabase\n */\nexport type InMemorySingleStreamProjectionOptions<\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n> = {\n canHandle: CanHandle<EventType>;\n getDocumentId?: (event: ReadEvent<EventType>) => string;\n collectionName: string;\n} & (\n | {\n evolve: InMemoryWithNullableDocumentEvolve<DocumentType, EventType>;\n }\n | {\n evolve: InMemoryWithNotNullDocumentEvolve<DocumentType, EventType>;\n initialState: () => DocumentType;\n }\n);\n\n/**\n * Creates a projection that handles events from a single stream\n */\nexport const inMemorySingleStreamProjection = <\n DocumentType extends Record<string, unknown>,\n EventType extends Event,\n>(\n options: InMemorySingleStreamProjectionOptions<DocumentType, EventType>,\n): InMemoryProjectionDefinition<EventType> => {\n return inMemoryMultiStreamProjection<DocumentType, EventType>({\n ...options,\n getDocumentId:\n options.getDocumentId ?? ((event) => event.metadata.streamName),\n });\n};\n","import { v4 as uuid } from 'uuid';\nimport {\n handleInMemoryProjections,\n type InMemoryProjectionDefinition,\n} from '.';\nimport {\n getInMemoryDatabase,\n type Document,\n type InMemoryDatabase,\n} from '../../../database';\nimport { isErrorConstructor } from '../../../errors';\nimport { JSONParser } from '../../../serialization';\nimport {\n assertFails,\n AssertionError,\n assertTrue,\n type ThenThrows,\n} from '../../../testing';\nimport type { CombinedReadEventMetadata, ReadEvent } from '../../../typing';\nimport { type Event } from '../../../typing';\nimport type {\n InMemoryEventStore,\n InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\n// Define a more specific type for T that extends Document\ntype DocumentWithId = Document & { _id?: string | number };\n\nexport type InMemoryProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type InMemoryProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type InMemoryProjectionSpec<EventType extends Event> = (\n givenEvents: InMemoryProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n then: (assert: InMemoryProjectionAssert, message?: string) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type InMemoryProjectionAssert = (options: {\n database: InMemoryDatabase;\n}) => Promise<void | boolean>;\n\nexport type InMemoryProjectionSpecOptions<EventType extends Event> = {\n projection: InMemoryProjectionDefinition<EventType>;\n};\n\nexport const InMemoryProjectionSpec = {\n for: <EventType extends Event>(\n options: InMemoryProjectionSpecOptions<EventType>,\n ): InMemoryProjectionSpec<EventType> => {\n const { projection } = options;\n\n return (givenEvents: InMemoryProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, InMemoryReadEventMetadata>[] =\n [];\n\n const run = async (database: InMemoryDatabase) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: InMemoryReadEventMetadata = {\n globalPosition: ++globalPosition,\n streamPosition: globalPosition,\n streamName: event.metadata?.streamName ?? `test-${uuid()}`,\n messageId: uuid(),\n };\n\n allEvents.push({\n ...event,\n kind: 'Event',\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n InMemoryReadEventMetadata\n >,\n });\n }\n\n // Create a minimal mock EventStore implementation\n const mockEventStore = {\n database,\n aggregateStream: async () => {\n return Promise.resolve({\n state: {},\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n readStream: async () => {\n return Promise.resolve({\n events: [],\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n appendToStream: async () => {\n return Promise.resolve({\n nextExpectedStreamVersion: 0n,\n createdNewStream: false,\n });\n },\n } as InMemoryEventStore;\n\n await handleInMemoryProjections({\n events: allEvents,\n projections: [projection],\n database,\n eventStore: mockEventStore,\n });\n };\n\n return {\n then: async (\n assertFn: InMemoryProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n await run(database);\n\n const succeeded = await assertFn({ database });\n\n if (succeeded !== undefined && succeeded === false) {\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n }\n },\n thenThrows: async <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n try {\n await run(database);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n },\n};\n\n// Helper functions for creating events in stream\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n>(\n streamName: string,\n event: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>,\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends\n InMemoryReadEventMetadata = InMemoryReadEventMetadata,\n>(\n streamName: string,\n events: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[],\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\n// Assertion helpers for checking documents\nexport function documentExists<T extends DocumentWithId>(\n expected: Partial<T>,\n options: { inCollection: string; withId: string | number },\n): InMemoryProjectionAssert {\n return async ({ database }) => {\n const collection = database.collection<T>(options.inCollection);\n\n const document = await collection.findOne((doc) => {\n // Handle both string IDs and numeric IDs in a type-safe way\n const docId = '_id' in doc ? doc._id : undefined;\n return docId === options.withId;\n });\n\n if (!document) {\n assertFails(\n `Document with ID ${options.withId} does not exist in collection ${options.inCollection}`,\n );\n return Promise.resolve(false);\n }\n\n // Check that all expected properties exist with expected values\n for (const [key, value] of Object.entries(expected)) {\n const propKey = key as keyof typeof document;\n if (\n !(key in document) ||\n JSONParser.stringify(document[propKey]) !== JSONParser.stringify(value)\n ) {\n assertFails(`Property ${key} doesn't match the expected value`);\n return Promise.resolve(false);\n }\n }\n\n return Promise.resolve(true);\n };\n}\n\n// Helper for checking document contents\nexport const expectInMemoryDocuments = {\n fromCollection: <T extends DocumentWithId>(collectionName: string) => ({\n withId: (id: string | number) => ({\n toBeEqual: (expected: Partial<T>): InMemoryProjectionAssert =>\n documentExists<T>(expected, {\n inCollection: collectionName,\n withId: id,\n }),\n }),\n }),\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<void>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import { isErrorConstructor, type ErrorConstructor } from '../errors';\nimport { AssertionError, assertThatArray, assertTrue } from './assertions';\n\ntype ErrorCheck<ErrorType> = (error: ErrorType) => boolean;\n\nexport type ThenThrows<ErrorType extends Error> =\n | (() => void)\n | ((errorConstructor: ErrorConstructor<ErrorType>) => void)\n | ((errorCheck: ErrorCheck<ErrorType>) => void)\n | ((\n errorConstructor: ErrorConstructor<ErrorType>,\n errorCheck?: ErrorCheck<ErrorType>,\n ) => void);\n\nexport type DeciderSpecification<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => void;\n thenNothingHappened: () => void;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => void;\n };\n};\nexport type AsyncDeciderSpecification<Command, Event> = (\n givenEvents: Event | Event[],\n) => {\n when: (command: Command) => {\n then: (expectedEvents: Event | Event[]) => Promise<void>;\n thenNothingHappened: () => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport const DeciderSpecification = {\n for: deciderSpecificationFor,\n};\n\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Event | Event[];\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}): DeciderSpecification<Command, Event>;\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (command: Command, state: State) => Promise<Event | Event[]>;\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}): AsyncDeciderSpecification<Command, Event>;\nfunction deciderSpecificationFor<Command, Event, State>(decider: {\n decide: (\n command: Command,\n state: State,\n ) => Event | Event[] | Promise<Event | Event[]>;\n evolve: (state: State, event: Event) => State;\n initialState: () => State;\n}):\n | DeciderSpecification<Command, Event>\n | AsyncDeciderSpecification<Command, Event> {\n {\n return (givenEvents: Event | Event[]) => {\n return {\n when: (command: Command) => {\n const handle = () => {\n const existingEvents = Array.isArray(givenEvents)\n ? givenEvents\n : [givenEvents];\n\n const currentState = existingEvents.reduce<State>(\n decider.evolve,\n decider.initialState(),\n );\n\n return decider.decide(command, currentState);\n };\n\n return {\n then: (expectedEvents: Event | Event[]): void | Promise<void> => {\n const resultEvents = handle();\n\n if (resultEvents instanceof Promise) {\n return resultEvents.then((events) => {\n thenHandler(events, expectedEvents);\n });\n }\n\n thenHandler(resultEvents, expectedEvents);\n },\n thenNothingHappened: (): void | Promise<void> => {\n const resultEvents = handle();\n\n if (resultEvents instanceof Promise) {\n return resultEvents.then((events) => {\n thenNothingHappensHandler(events);\n });\n }\n\n thenNothingHappensHandler(resultEvents);\n },\n thenThrows: <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): void | Promise<void> => {\n try {\n const result = handle();\n if (result instanceof Promise) {\n return result\n .then(() => {\n throw new AssertionError(\n 'Handler did not fail as expected',\n );\n })\n .catch((error) => {\n thenThrowsErrorHandler(error, args);\n });\n }\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n thenThrowsErrorHandler(error, args);\n }\n },\n };\n },\n };\n };\n }\n}\n\nfunction thenHandler<Event>(\n events: Event | Event[],\n expectedEvents: Event | Event[],\n): void {\n const resultEventsArray = Array.isArray(events) ? events : [events];\n\n const expectedEventsArray = Array.isArray(expectedEvents)\n ? expectedEvents\n : [expectedEvents];\n\n assertThatArray(resultEventsArray).containsOnlyElementsMatching(\n expectedEventsArray,\n );\n}\n\nfunction thenNothingHappensHandler<Event>(events: Event | Event[]): void {\n const resultEventsArray = Array.isArray(events) ? events : [events];\n assertThatArray(resultEventsArray).isEmpty();\n}\n\nfunction thenThrowsErrorHandler<ErrorType extends Error>(\n error: unknown,\n args: Parameters<ThenThrows<ErrorType>>,\n): void {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n}\n","import type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n EventStore,\n EventStoreReadEventMetadata,\n ReadStreamOptions,\n ReadStreamResult,\n StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport { type Event, type EventMetaDataOf } from '../typing';\n\nexport type TestEventStream<EventType extends Event = Event> = [\n string,\n EventType[],\n];\n\nexport type EventStoreWrapper<Store extends EventStore> = Store & {\n appendedEvents: Map<string, TestEventStream>;\n setup<EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>>;\n};\n\nexport const WrapEventStore = <Store extends EventStore>(\n eventStore: Store,\n): EventStoreWrapper<Store> => {\n const appendedEvents = new Map<string, TestEventStream>();\n\n const wrapped = {\n ...eventStore,\n aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<\n AggregateStreamResult<State, StreamPositionTypeOfEventStore<Store>>\n > {\n return eventStore.aggregateStream(streamName, options);\n },\n\n async readStream<EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<\n ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >\n > {\n return (await eventStore.readStream(\n streamName,\n options,\n )) as ReadStreamResult<\n EventType,\n EventStoreReadEventMetadata<Store> & EventMetaDataOf<EventType>\n >;\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<StreamPositionTypeOfEventStore<Store>>,\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n const result = await eventStore.appendToStream(\n streamName,\n events,\n options,\n );\n\n const currentStream = appendedEvents.get(streamName) ?? [streamName, []];\n\n appendedEvents.set(streamName, [\n streamName,\n [...currentStream[1], ...events],\n ]);\n\n return result;\n },\n\n appendedEvents,\n\n setup: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n ): Promise<AppendToStreamResult<StreamPositionTypeOfEventStore<Store>>> => {\n return eventStore.appendToStream(streamName, events);\n },\n\n // streamEvents: (): ReadableStream<\n // // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n // ReadEvent<Event, ReadEventMetadataType> | GlobalSubscriptionEvent\n // > => {\n // return eventStore.streamEvents();\n // },\n };\n\n return wrapped as EventStoreWrapper<Store>;\n};\n","import { TransformStream } from 'web-streams-polyfill';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport { globalStreamCaughtUp, type GlobalSubscriptionEvent } from '../events';\n\nexport const streamTrackingGlobalPosition = (\n currentEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n) => new CaughtUpTransformStream(currentEvents);\n\nexport class CaughtUpTransformStream extends TransformStream<\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n | ReadEvent<Event, ReadEventMetadataWithGlobalPosition>\n | GlobalSubscriptionEvent\n> {\n private _currentPosition: bigint;\n private _logPosition: bigint;\n\n constructor(events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]) {\n super({\n start: (controller) => {\n let globalPosition = 0n;\n for (const event of events) {\n controller.enqueue(event);\n globalPosition = event.metadata.globalPosition;\n }\n controller.enqueue(globalStreamCaughtUp({ globalPosition }));\n },\n transform: (event, controller) => {\n this._currentPosition = event.metadata.globalPosition;\n controller.enqueue(event);\n\n if (this._currentPosition < this._logPosition) return;\n\n controller.enqueue(\n globalStreamCaughtUp({ globalPosition: this._currentPosition }),\n );\n },\n });\n\n this._currentPosition = this._logPosition =\n events.length > 0\n ? events[events.length - 1]!.metadata.globalPosition\n : 0n;\n }\n\n public set logPosition(value: bigint) {\n this._logPosition = value;\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport { notifyAboutNoActiveReadersStream } from '../../streaming/transformations/notifyAboutNoActiveReaders';\nimport { writeToStream } from '../../streaming/writers';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport {\n CaughtUpTransformStream,\n streamTrackingGlobalPosition,\n} from './caughtUpTransformStream';\n\nexport const StreamingCoordinator = () => {\n const allEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[] = [];\n const listeners = new Map<string, CaughtUpTransformStream>();\n\n return {\n notify: async (\n events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n ) => {\n if (events.length === 0) return;\n\n allEvents.push(...events);\n\n for (const listener of listeners.values()) {\n listener.logPosition =\n events[events.length - 1]!.metadata.globalPosition;\n\n await writeToStream(listener, events);\n }\n },\n\n stream: () => {\n const streamId = uuid();\n const transformStream = streamTrackingGlobalPosition(allEvents);\n\n listeners.set(streamId, transformStream);\n return transformStream.readable.pipeThrough(\n notifyAboutNoActiveReadersStream(\n (stream) => {\n if (listeners.has(stream.streamId))\n listeners.delete(stream.streamId);\n },\n { streamId },\n ),\n );\n },\n };\n};\n","import { v4 as uuid } from 'uuid';\nimport { TransformStream } from 'web-streams-polyfill';\n\nexport const notifyAboutNoActiveReadersStream = <Item>(\n onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);\n\nexport class NotifyAboutNoActiveReadersStream<Item> extends TransformStream<\n Item,\n Item\n> {\n private checkInterval: NodeJS.Timeout | null = null;\n public readonly streamId: string;\n private _isStopped: boolean = false;\n public get hasActiveSubscribers() {\n return !this._isStopped;\n }\n\n constructor(\n private onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n ) {\n super({\n cancel: (reason) => {\n console.log('Stream was canceled. Reason:', reason);\n this.stopChecking();\n },\n });\n this.streamId = options?.streamId ?? uuid();\n\n this.onNoActiveReaderCallback = onNoActiveReaderCallback;\n\n this.startChecking(options?.intervalCheckInMs ?? 20);\n }\n\n private startChecking(interval: number) {\n this.checkInterval = setInterval(() => {\n this.checkNoActiveReader();\n }, interval);\n }\n\n private stopChecking() {\n if (!this.checkInterval) return;\n\n clearInterval(this.checkInterval);\n this.checkInterval = null;\n this._isStopped = true;\n this.onNoActiveReaderCallback(this);\n }\n\n private checkNoActiveReader() {\n if (!this.readable.locked && !this._isStopped) {\n this.stopChecking();\n }\n }\n}\n","import type { ReadableStream, TransformStream } from 'web-streams-polyfill';\nimport { noMoreWritingOn } from './noMoreWritingOn';\n\nexport const writeToStream = async <In, Out = In>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<boolean> => {\n if (stream.writable.locked) return false;\n\n const writer = stream.writable.getWriter();\n await writer.ready;\n\n if (!stream.readable.locked) return false;\n\n try {\n for (const item of items) {\n await writer.write(item);\n }\n } catch (error) {\n console.log(error);\n } finally {\n await writer.close();\n }\n return true;\n};\n\nexport const writeToStreamAndStop = async <In, Out>(\n stream: TransformStream<In, Out>,\n items: In[],\n): Promise<ReadableStream<Out>> => {\n await writeToStream(stream, items);\n return await noMoreWritingOn(stream);\n};\n","import {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type StreamPositionTypeOfEventStore,\n} from '../eventStore';\nimport type { Event } from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\n\nexport const CommandHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type CommandHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromCommandHandlerRetryOptions = (\n retryOptions: CommandHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return CommandHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...CommandHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region command-handler\nexport type CommandHandlerResult<\n State,\n StreamEvent extends Event,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newState: State;\n newEvents: StreamEvent[];\n};\n\nexport type CommandHandlerOptions<State, StreamEvent extends Event> = {\n evolve: (state: State, event: StreamEvent) => State;\n initialState: () => State;\n mapToStreamId?: (id: string) => string;\n retry?: CommandHandlerRetryOptions;\n};\n\nexport type HandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] &\n (\n | {\n expectedStreamVersion?: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >;\n }\n | {\n retry?: CommandHandlerRetryOptions;\n }\n );\n\ntype CommandHandlerFunction<State, StreamEvent extends Event> = (\n state: State,\n) => StreamEvent | StreamEvent[] | Promise<StreamEvent | StreamEvent[]>;\n\nexport const CommandHandler =\n <State, StreamEvent extends Event>(\n options: CommandHandlerOptions<State, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n store: Store,\n id: string,\n handle:\n | CommandHandlerFunction<State, StreamEvent>\n | CommandHandlerFunction<State, StreamEvent>[],\n handleOptions?: HandleOptions<Store>,\n ): Promise<CommandHandlerResult<State, StreamEvent, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n CommandHandlerResult<\n State,\n StreamEvent,\n StreamPositionTypeOfEventStore<Store>\n >\n >(store, async ({ eventStore }) => {\n const { evolve, initialState } = options;\n const mapToStreamId = options.mapToStreamId ?? ((id) => id);\n\n const streamName = mapToStreamId(id);\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n State,\n StreamEvent\n >(streamName, {\n evolve,\n initialState,\n read: {\n ...(handleOptions ? handleOptions : {}),\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n currentStreamVersion,\n streamExists: _streamExists,\n ...restOfAggregationResult\n } = aggregationResult;\n\n let state = aggregationResult.state;\n\n const handlers = Array.isArray(handle) ? handle : [handle];\n let eventsToAppend: StreamEvent[] = [];\n\n // 3. Run business logic\n for (const handler of handlers) {\n const result = await handler(state);\n\n const newEvents = Array.isArray(result) ? result : [result];\n\n if (newEvents.length > 0) {\n state = newEvents.reduce(evolve, state);\n }\n\n eventsToAppend = [...eventsToAppend, ...newEvents];\n }\n\n //const newEvents = Array.isArray(result) ? result : [result];\n\n if (eventsToAppend.length === 0) {\n return {\n ...restOfAggregationResult,\n newEvents: [],\n newState: state,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n nextExpectedStreamVersion: currentStreamVersion,\n createdNewStream: false,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const expectedStreamVersion: ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n > =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? (currentStreamVersion as ExpectedStreamVersion<\n StreamPositionTypeOfEventStore<Store>\n >)\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n eventsToAppend,\n {\n ...handleOptions,\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with updated state\n return {\n ...appendResult,\n newEvents: eventsToAppend,\n newState: state,\n } as unknown as CommandHandlerResult<State, StreamEvent, Store>;\n });\n\n return result;\n },\n fromCommandHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion command-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type { EventStore } from '../eventStore';\nimport { type Command, type Event } from '../typing';\nimport type { Decider } from '../typing/decider';\nimport {\n CommandHandler,\n type CommandHandlerOptions,\n type HandleOptions,\n} from './handleCommand';\n\n// #region command-handler\n\nexport type DeciderCommandHandlerOptions<\n State,\n CommandType extends Command,\n StreamEvent extends Event,\n> = CommandHandlerOptions<State, StreamEvent> &\n Decider<State, CommandType, StreamEvent>;\n\nexport const DeciderCommandHandler =\n <State, CommandType extends Command, StreamEvent extends Event>(\n options: DeciderCommandHandlerOptions<State, CommandType, StreamEvent>,\n ) =>\n async <Store extends EventStore>(\n eventStore: Store,\n id: string,\n commands: CommandType | CommandType[],\n handleOptions?: HandleOptions<Store>,\n ) => {\n const { decide, ...rest } = options;\n\n const deciders = (Array.isArray(commands) ? commands : [commands]).map(\n (command) => (state: State) => decide(command, state),\n );\n\n return CommandHandler<State, StreamEvent>(rest)(\n eventStore,\n id,\n deciders,\n handleOptions,\n );\n };\n// #endregion command-handler\n","import { EmmettError } from '../errors';\nimport {\n type AnyCommand,\n type AnyMessage,\n type Command,\n type CommandTypeOf,\n type Event,\n type EventTypeOf,\n type Message,\n type SingleMessageHandler,\n type SingleRawMessageHandlerWithoutContext,\n} from '../typing';\n\nexport interface CommandSender {\n send<CommandType extends Command = Command>(\n command: CommandType,\n ): Promise<void>;\n}\n\nexport interface EventsPublisher {\n publish<EventType extends Event = Event>(event: EventType): Promise<void>;\n}\n\nexport type ScheduleOptions = { afterInMs: number } | { at: Date };\n\nexport interface MessageScheduler<CommandOrEvent extends Command | Event> {\n schedule<MessageType extends CommandOrEvent>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandBus extends CommandSender, MessageScheduler<Command> {}\n\nexport interface EventBus extends EventsPublisher, MessageScheduler<Event> {}\n\nexport interface MessageBus extends CommandBus, EventBus {\n schedule<MessageType extends Command | Event>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void;\n}\n\nexport interface CommandProcessor {\n handle<CommandType extends Command>(\n commandHandler: SingleMessageHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void;\n}\nexport interface EventSubscription {\n subscribe<EventType extends Event>(\n eventHandler: SingleMessageHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void;\n}\n\nexport type ScheduledMessage = {\n message: Message;\n options?: ScheduleOptions;\n};\n\nexport interface ScheduledMessageProcessor {\n dequeue(): ScheduledMessage[];\n}\n\nexport type MessageSubscription = EventSubscription | CommandProcessor;\n\nexport const getInMemoryMessageBus = (): MessageBus &\n EventSubscription &\n CommandProcessor &\n ScheduledMessageProcessor => {\n const allHandlers = new Map<\n string,\n SingleRawMessageHandlerWithoutContext<AnyMessage>[]\n >();\n let pendingMessages: ScheduledMessage[] = [];\n\n return {\n send: async <CommandType extends Command = AnyCommand>(\n command: CommandType,\n ): Promise<void> => {\n const handlers = allHandlers.get(command.type);\n\n if (handlers === undefined || handlers.length === 0)\n throw new EmmettError(\n `No handler registered for command ${command.type}!`,\n );\n\n const commandHandler = handlers[0]!;\n\n await commandHandler(command);\n },\n\n publish: async <EventType extends Event = Event>(\n event: EventType,\n ): Promise<void> => {\n const handlers = allHandlers.get(event.type) ?? [];\n\n for (const handler of handlers) {\n const eventHandler = handler;\n\n await eventHandler(event);\n }\n },\n\n schedule: <MessageType extends Message>(\n message: MessageType,\n when?: ScheduleOptions,\n ): void => {\n pendingMessages = [...pendingMessages, { message, options: when }];\n },\n\n handle: <CommandType extends Command>(\n commandHandler: SingleMessageHandler<CommandType>,\n ...commandTypes: CommandTypeOf<CommandType>[]\n ): void => {\n const alreadyRegistered = [...allHandlers.keys()].filter((registered) =>\n commandTypes.includes(registered),\n );\n\n if (alreadyRegistered.length > 0)\n throw new EmmettError(\n `Cannot register handler for commands ${alreadyRegistered.join(', ')} as they're already registered!`,\n );\n for (const commandType of commandTypes) {\n allHandlers.set(commandType, [\n commandHandler as SingleRawMessageHandlerWithoutContext<AnyMessage>,\n ]);\n }\n },\n\n subscribe<EventType extends Event>(\n eventHandler: SingleMessageHandler<EventType>,\n ...eventTypes: EventTypeOf<EventType>[]\n ): void {\n for (const eventType of eventTypes) {\n if (!allHandlers.has(eventType)) allHandlers.set(eventType, []);\n\n allHandlers.set(eventType, [\n ...(allHandlers.get(eventType) ?? []),\n eventHandler as SingleRawMessageHandlerWithoutContext<AnyMessage>,\n ]);\n }\n },\n\n dequeue: (): ScheduledMessage[] => {\n const pending = pendingMessages;\n pendingMessages = [];\n return pending;\n },\n };\n};\n","import type { EmmettError } from '../errors';\nimport type { ProjectionDefinition } from '../projections';\nimport {\n type AnyEvent,\n type AnyMessage,\n type AnyReadEventMetadata,\n type AnyRecordedMessageMetadata,\n type BatchRecordedMessageHandlerWithContext,\n type CanHandle,\n type DefaultRecord,\n type Event,\n type GlobalPositionTypeOfRecordedMessageMetadata,\n type Message,\n type MessageHandlerResult,\n type RecordedMessage,\n type SingleMessageHandlerWithContext,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport { isBigint } from '../validation';\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type CurrentMessageProcessorPosition<CheckpointType = any> =\n | { lastCheckpoint: CheckpointType }\n | 'BEGINNING'\n | 'END';\n\nexport type GetCheckpoint<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n) => CheckpointType | null;\n\nexport const getCheckpoint = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n): CheckpointType | null => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return 'checkpoint' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.checkpoint)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.checkpoint\n : 'globalPosition' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.globalPosition)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.globalPosition\n : 'streamPosition' in message.metadata &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n isBigint(message.metadata.streamPosition)\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n message.metadata.streamPosition\n : null;\n};\n\nexport const wasMessageHandled = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n checkpoint: CheckpointType | null,\n): boolean => {\n //TODO Make it smarter\n const messageCheckpoint = getCheckpoint(message);\n const checkpointBigint = checkpoint as bigint | null;\n\n return (\n messageCheckpoint !== null &&\n messageCheckpoint !== undefined &&\n checkpointBigint !== null &&\n checkpointBigint !== undefined &&\n messageCheckpoint <= checkpointBigint\n );\n};\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type MessageProcessorStartFrom<CheckpointType = any> =\n | CurrentMessageProcessorPosition<CheckpointType>\n | 'CURRENT';\n\nexport type MessageProcessorType = 'projector' | 'reactor';\nexport const MessageProcessorType = {\n PROJECTOR: 'projector' as MessageProcessorType,\n REACTOR: 'reactor' as MessageProcessorType,\n};\n\nexport type MessageProcessor<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n id: string;\n type: string;\n start: (\n options: Partial<HandlerContext>,\n ) => Promise<CurrentMessageProcessorPosition<CheckpointType> | undefined>;\n close: () => Promise<void>;\n isActive: boolean;\n handle: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n Partial<HandlerContext>\n >;\n};\n\nexport const MessageProcessor = {\n result: {\n skip: (options?: { reason?: string }): MessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): MessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type MessageProcessingScope<\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = <Result = MessageHandlerResult>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => Result | Promise<Result>;\n\nexport type Checkpointer<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n read: ReadProcessorCheckpoint<CheckpointType, HandlerContext>;\n store: StoreProcessorCheckpoint<\n MessageType,\n MessageMetadataType,\n CheckpointType,\n HandlerContext\n >;\n};\n\nexport type BaseMessageProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = {\n type?: string;\n processorId: string;\n version?: number;\n partition?: string;\n startFrom?: MessageProcessorStartFrom<CheckpointType>;\n stopAfter?: (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n ) => boolean;\n processingScope?: MessageProcessingScope<HandlerContext>;\n checkpoints?: Checkpointer<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >;\n canHandle?: CanHandle<MessageType>;\n hooks?: {\n onStart?: OnReactorStartHook<HandlerContext>;\n onClose?: OnReactorCloseHook;\n };\n};\n\nexport type HandlerOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> =\n | {\n eachMessage: SingleRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n eachBatch?: never;\n }\n | {\n eachMessage?: never;\n eachBatch: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n };\n\nexport type OnReactorStartHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorCloseHook = () => Promise<void>;\n\nexport type ReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = BaseMessageProcessorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n> &\n HandlerOptions<MessageType, MessageMetadataType, HandlerContext>;\n\nexport type ProjectorOptions<\n EventType extends AnyEvent = AnyEvent,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n> = Omit<\n BaseMessageProcessorOptions<\n EventType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >,\n 'type' | 'processorId'\n> & { processorId?: string } & {\n truncateOnStart?: boolean;\n projection: ProjectionDefinition<\n EventType,\n MessageMetadataType,\n HandlerContext\n >;\n};\n\nexport const defaultProcessingMessageProcessingScope = <\n HandlerContext = never,\n Result = MessageHandlerResult,\n>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => handler(partialContext as HandlerContext);\n\nexport type ReadProcessorCheckpointResult<CheckpointType = unknown> = {\n lastCheckpoint: CheckpointType | null;\n};\n\nexport type ReadProcessorCheckpoint<\n CheckpointType = unknown,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n options: { processorId: string; partition?: string },\n context: HandlerContext,\n) => Promise<ReadProcessorCheckpointResult<CheckpointType>>;\n\nexport type StoreProcessorCheckpointResult<CheckpointType = unknown> =\n | {\n success: true;\n newCheckpoint: CheckpointType;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport type StoreProcessorCheckpoint<\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n CheckpointType = unknown,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> =\n | ((\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: CheckpointType | null;\n partition?: string;\n },\n context: HandlerContext,\n ) => Promise<StoreProcessorCheckpointResult<CheckpointType | null>>)\n | ((\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: CheckpointType | null;\n partition?: string;\n },\n context: HandlerContext,\n ) => Promise<StoreProcessorCheckpointResult<CheckpointType>>);\n\nexport const reactor = <\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<MessageMetadataType>,\n>(\n options: ReactorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n >,\n): MessageProcessor<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n CheckpointType\n> => {\n const eachMessage: SingleMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n > =\n 'eachMessage' in options && options.eachMessage\n ? options.eachMessage\n : () => Promise.resolve();\n let isActive = true;\n\n const { checkpoints, processorId, partition } = options;\n\n const processingScope =\n options.processingScope ?? defaultProcessingMessageProcessingScope;\n\n let lastCheckpoint: CheckpointType | null = null;\n\n return {\n id: options.processorId,\n type: options.type ?? MessageProcessorType.REACTOR,\n close: () =>\n options.hooks?.onClose ? options.hooks?.onClose() : Promise.resolve(),\n start: async (\n startOptions: Partial<HandlerContext>,\n ): Promise<CurrentMessageProcessorPosition<CheckpointType> | undefined> => {\n isActive = true;\n\n if (lastCheckpoint !== null)\n return {\n lastCheckpoint,\n };\n\n return await processingScope(async (context) => {\n if (options.hooks?.onStart) {\n await options.hooks?.onStart(context);\n }\n\n if (options.startFrom !== 'CURRENT' && options.startFrom)\n return options.startFrom;\n\n if (checkpoints) {\n const readResult = await checkpoints?.read(\n {\n processorId: processorId,\n partition: partition,\n },\n { ...startOptions, ...context },\n );\n lastCheckpoint = readResult.lastCheckpoint;\n }\n\n if (lastCheckpoint === null) return 'BEGINNING';\n\n return {\n lastCheckpoint,\n };\n }, startOptions);\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n messages: RecordedMessage<MessageType, MessageMetadataType>[],\n partialContext: Partial<HandlerContext>,\n ): Promise<MessageHandlerResult> => {\n if (!isActive) return Promise.resolve();\n\n return await processingScope(async (context) => {\n let result: MessageHandlerResult = undefined;\n\n for (const message of messages) {\n if (wasMessageHandled(message, lastCheckpoint)) continue;\n\n const messageProcessingResult = await eachMessage(message, context);\n\n if (checkpoints) {\n const storeCheckpointResult: StoreProcessorCheckpointResult<CheckpointType | null> =\n await checkpoints.store(\n {\n processorId: options.processorId,\n version: options.version,\n message,\n lastCheckpoint,\n partition: options.partition,\n },\n context,\n );\n\n if (storeCheckpointResult.success) {\n // TODO: Add correct handling of the storing checkpoint\n lastCheckpoint = storeCheckpointResult.newCheckpoint;\n }\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(message)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n }, partialContext);\n },\n };\n};\n\nexport const projector = <\n EventType extends Event = Event,\n EventMetaDataType extends\n AnyRecordedMessageMetadata = AnyRecordedMessageMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n CheckpointType = GlobalPositionTypeOfRecordedMessageMetadata<EventMetaDataType>,\n>(\n options: ProjectorOptions<\n EventType,\n EventMetaDataType,\n HandlerContext,\n CheckpointType\n >,\n): MessageProcessor<\n EventType,\n EventMetaDataType,\n HandlerContext,\n CheckpointType\n> => {\n const { projection, ...rest } = options;\n\n return reactor<EventType, EventMetaDataType, HandlerContext, CheckpointType>({\n ...rest,\n type: MessageProcessorType.PROJECTOR,\n processorId: options.processorId ?? `projection:${projection.name}`,\n hooks: {\n onStart:\n (options.truncateOnStart && options.projection.truncate) ||\n options.hooks?.onStart\n ? async (context: HandlerContext) => {\n if (options.truncateOnStart && options.projection.truncate)\n await options.projection.truncate(context);\n\n if (options.hooks?.onStart) await options.hooks?.onStart(context);\n }\n : undefined,\n onClose: options.hooks?.onClose,\n },\n eachMessage: async (\n event: RecordedMessage<EventType, EventMetaDataType>,\n context: HandlerContext,\n ) => {\n if (!projection.canHandle.includes(event.type)) return;\n\n await projection.handle([event], context);\n },\n });\n};\n","import { getInMemoryDatabase, type InMemoryDatabase } from '../database';\nimport { EmmettError } from '../errors';\nimport {\n type AnyEvent,\n type AnyMessage,\n type BatchRecordedMessageHandlerWithContext,\n type MessageHandlerResult,\n type ReadEventMetadataWithGlobalPosition,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport {\n getCheckpoint,\n MessageProcessor,\n projector,\n reactor,\n type Checkpointer,\n type MessageProcessingScope,\n type ProjectorOptions,\n type ReactorOptions,\n} from './processors';\n\nexport type InMemoryProcessorHandlerContext = {\n database: InMemoryDatabase;\n};\n\nexport type InMemoryProcessor<MessageType extends AnyMessage = AnyMessage> =\n MessageProcessor<\n MessageType,\n // TODO: generalize this to support other metadata types\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n > & { database: InMemoryDatabase };\n\nexport type InMemoryProcessorEachMessageHandler<\n MessageType extends AnyMessage = AnyMessage,\n> = SingleRecordedMessageHandlerWithContext<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n>;\n\nexport type InMemoryProcessorEachBatchHandler<\n MessageType extends AnyMessage = AnyMessage,\n> = BatchRecordedMessageHandlerWithContext<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n>;\n\nexport type InMemoryProcessorConnectionOptions = {\n database?: InMemoryDatabase;\n};\n\ntype CheckpointDocument = {\n _id: string;\n lastCheckpoint: bigint | null;\n};\n\nexport type InMemoryCheckpointer<MessageType extends AnyMessage = AnyMessage> =\n Checkpointer<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n >;\n\nexport const inMemoryCheckpointer = <\n MessageType extends AnyMessage = AnyMessage,\n>(): InMemoryCheckpointer<MessageType> => {\n return {\n read: async ({ processorId }, { database }) => {\n const checkpoint = await database\n .collection<CheckpointDocument>('emt_processor_checkpoints')\n .findOne((d) => d._id === processorId);\n\n return Promise.resolve({\n lastCheckpoint: checkpoint?.lastCheckpoint ?? null,\n });\n },\n store: async (context, { database }) => {\n const { message, processorId, lastCheckpoint } = context;\n const checkpoints = database.collection<CheckpointDocument>(\n 'emt_processor_checkpoints',\n );\n\n const checkpoint = await checkpoints.findOne(\n (d) => d._id === processorId,\n );\n\n const currentPosition = checkpoint?.lastCheckpoint ?? null;\n\n const newCheckpoint: bigint | null = getCheckpoint(message);\n\n if (\n currentPosition &&\n (currentPosition === newCheckpoint ||\n currentPosition !== lastCheckpoint)\n ) {\n return {\n success: false,\n reason: currentPosition === newCheckpoint ? 'IGNORED' : 'MISMATCH',\n };\n }\n\n await checkpoints.handle(processorId, (existing) => ({\n ...(existing ?? {}),\n _id: processorId,\n lastCheckpoint: newCheckpoint,\n }));\n\n return { success: true, newCheckpoint };\n },\n };\n};\n\ntype InMemoryConnectionOptions = {\n connectionOptions?: InMemoryProcessorConnectionOptions;\n};\n\nexport type InMemoryReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n> = ReactorOptions<\n MessageType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n> &\n InMemoryConnectionOptions;\n\nexport type InMemoryProjectorOptions<EventType extends AnyEvent = AnyEvent> =\n ProjectorOptions<\n EventType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n > &\n InMemoryConnectionOptions;\n\nexport type InMemoryProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n> =\n | InMemoryReactorOptions<MessageType>\n | InMemoryProjectorOptions<MessageType & AnyEvent>;\n\nconst inMemoryProcessingScope = (options: {\n database: InMemoryDatabase | null;\n processorId: string;\n}): MessageProcessingScope<InMemoryProcessorHandlerContext> => {\n const processorDatabase = options.database;\n\n const processingScope: MessageProcessingScope<\n InMemoryProcessorHandlerContext\n > = <Result = MessageHandlerResult>(\n handler: (\n context: InMemoryProcessorHandlerContext,\n ) => Result | Promise<Result>,\n partialContext: Partial<InMemoryProcessorHandlerContext>,\n ) => {\n const database = processorDatabase ?? partialContext?.database;\n\n if (!database)\n throw new EmmettError(\n `InMemory processor '${options.processorId}' is missing database. Ensure that you passed it through options`,\n );\n\n return handler({ ...partialContext, database });\n };\n\n return processingScope;\n};\n\nexport const inMemoryProjector = <EventType extends AnyEvent = AnyEvent>(\n options: InMemoryProjectorOptions<EventType>,\n): InMemoryProcessor<EventType> => {\n const database = options.connectionOptions?.database ?? getInMemoryDatabase();\n\n const hooks = {\n onStart: options.hooks?.onStart,\n onClose: options.hooks?.onClose\n ? async () => {\n if (options.hooks?.onClose) await options.hooks?.onClose();\n }\n : undefined,\n };\n\n return {\n ...projector<\n EventType,\n ReadEventMetadataWithGlobalPosition,\n InMemoryProcessorHandlerContext\n >({\n ...options,\n hooks,\n processingScope: inMemoryProcessingScope({\n database,\n processorId:\n options.processorId ?? `projection:${options.projection.name}`,\n }),\n checkpoints: inMemoryCheckpointer<EventType>(),\n }),\n database,\n };\n};\n\nexport const inMemoryReactor = <MessageType extends AnyMessage = AnyMessage>(\n options: InMemoryReactorOptions<MessageType>,\n): InMemoryProcessor<MessageType> => {\n const database = options.connectionOptions?.database ?? getInMemoryDatabase();\n\n const hooks = {\n onStart: options.hooks?.onStart,\n onClose: options.hooks?.onClose\n ? async () => {\n if (options.hooks?.onClose) await options.hooks?.onClose();\n }\n : undefined,\n };\n\n return {\n ...reactor({\n ...options,\n hooks,\n processingScope: inMemoryProcessingScope({\n database,\n processorId: options.processorId,\n }),\n checkpoints: inMemoryCheckpointer<MessageType>(),\n }),\n database,\n };\n};\n","import { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyEvent,\n AnyReadEventMetadata,\n BatchRecordedMessageHandlerWithContext,\n CanHandle,\n DefaultRecord,\n Event,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = BatchRecordedMessageHandlerWithContext<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n>;\n\nexport type TruncateProjection<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: ProjectionHandlerContext) => Promise<void>;\n\nexport interface ProjectionDefinition<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> {\n name?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n truncate?: TruncateProjection<ProjectionHandlerContext>;\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definition: ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >,\n): ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n> => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n any,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","export const concatUint8Arrays = (chunks: Uint8Array[]): Uint8Array => {\n const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const chunk of chunks) {\n result.set(chunk, offset);\n offset += chunk.length;\n }\n return result;\n};\n","import type { ReadableStream } from 'web-streams-polyfill';\n\nexport const collect = async <T>(stream: ReadableStream<T>): Promise<T[]> => {\n const results: T[] = [];\n\n for await (const value of stream) {\n results.push(value as T);\n }\n\n return results;\n};\n","import type { Decoder } from '.';\nimport { concatUint8Arrays } from '../binaryArrays';\n\nexport class BinaryJsonDecoder<Decoded>\n implements Decoder<Uint8Array, Decoded>\n{\n private buffer: Uint8Array[] = [];\n\n addToBuffer(data: Uint8Array): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n return text.includes('\\n');\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage()) {\n return null;\n }\n\n const combined = concatUint8Arrays(this.buffer);\n const text = new TextDecoder().decode(combined);\n const delimiterIndex = text.indexOf('\\n');\n\n if (delimiterIndex === -1) {\n return null;\n }\n\n const jsonString = text.slice(0, delimiterIndex);\n const remaining = new Uint8Array(combined.buffer, delimiterIndex + 1);\n this.buffer = remaining.byteLength > 0 ? [remaining] : [];\n\n return JSON.parse(jsonString) as Decoded;\n }\n}\n","import type { Decoder } from '.';\n\nexport class StringDecoder<Decoded> implements Decoder<string, Decoded> {\n protected buffer: string[] = [];\n\n constructor(private transform: (input: string) => Decoded) {\n this.transform = transform;\n }\n\n addToBuffer(data: string): void {\n this.buffer.push(data);\n }\n\n clearBuffer(): void {\n this.buffer = [];\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer.some((chunk) => chunk.includes('\\n'));\n }\n\n decode(): Decoded | null {\n const completeString = this.buffer.join('');\n\n if (!this.hasCompleteMessage()) {\n if (completeString.trim().length > 0) {\n throw new Error('Unterminated string in JSON at position');\n }\n return null;\n }\n\n const delimiterIndex = completeString.indexOf('\\n');\n const message = completeString.slice(0, delimiterIndex).trim();\n this.buffer = [completeString.slice(delimiterIndex + 1)];\n\n return this.transform(message);\n }\n}\n","import { StringDecoder } from './string';\n\nexport class JsonDecoder<Decoded> extends StringDecoder<Decoded> {\n constructor() {\n super((jsonString) => JSON.parse(jsonString) as Decoded);\n }\n}\n","import type { Decoder } from '.';\n\nexport class ObjectDecoder<Decoded> implements Decoder<Decoded, Decoded> {\n private buffer: Decoded | null = null;\n\n addToBuffer(data: Decoded): void {\n this.buffer = data;\n }\n\n clearBuffer(): void {\n this.buffer = null;\n }\n\n hasCompleteMessage(): boolean {\n return this.buffer !== null;\n }\n\n decode(): Decoded | null {\n if (!this.hasCompleteMessage() || !this.buffer) {\n return null;\n }\n\n const data = this.buffer;\n this.clearBuffer();\n return data as Decoded;\n }\n}\n","import type { Decoder } from '.';\nimport { BinaryJsonDecoder } from './binary';\nimport { JsonDecoder } from './json';\nimport { ObjectDecoder } from './object';\n\nexport class CompositeDecoder<Decoded> implements Decoder<unknown, Decoded> {\n constructor(\n private decoders: [(data: unknown) => boolean, Decoder<unknown, Decoded>][],\n ) {}\n\n private decoderFor(data: unknown): Decoder<unknown, unknown> | null {\n const decoder = this.decoders.find((d) => d[0](data));\n\n if (!decoder) return null;\n\n return decoder[1];\n }\n\n addToBuffer(data: unknown): void {\n this.decoderFor(data)?.addToBuffer(data);\n }\n\n clearBuffer(): void {\n for (const decoder of this.decoders.map((d) => d[1])) {\n decoder.clearBuffer();\n }\n }\n\n hasCompleteMessage(): boolean {\n return this.decoders.some((d) => d[1].hasCompleteMessage());\n }\n\n decode(): Decoded | null {\n const decoder = this.decoders\n .map((d) => d[1])\n .find((d) => d.hasCompleteMessage());\n\n return decoder?.decode() ?? null;\n }\n}\n\nexport class DefaultDecoder<Decoded> extends CompositeDecoder<Decoded> {\n constructor() {\n super([\n [(data) => typeof data === 'string', new JsonDecoder<Decoded>()],\n [(data) => data instanceof Uint8Array, new BinaryJsonDecoder<Decoded>()],\n [(data) => typeof data === 'object', new ObjectDecoder<Decoded>()],\n ]);\n }\n}\n","import { ReadableStream } from 'web-streams-polyfill';\n\nexport const fromArray = <T>(chunks: T[]) =>\n new ReadableStream<T>({\n start(controller) {\n for (const chunk of chunks) controller.enqueue(chunk);\n controller.close();\n },\n });\n","import { fromArray } from './fromArray';\nexport const streamGenerators = { fromArray };\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport type { AsyncRetryOptions } from '../utils';\nimport type { Decoder } from './decoders';\nimport { DefaultDecoder } from './decoders/composite';\nimport { streamTransformations } from './transformations';\n\nconst { retry } = streamTransformations;\n\nexport const restream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n): ReadableStream<Transformed> =>\n retry(createSourceStream, handleChunk(transform, decoder), retryOptions)\n .readable;\n\nconst handleChunk =\n <Source = unknown, Transformed = Source, StreamType = Source>(\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n ) =>\n (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ): void => {\n const { done: isDone, value } = readResult;\n\n if (value) decoder.addToBuffer(value);\n\n if (!isDone && !decoder.hasCompleteMessage()) return;\n\n decodeAndTransform(decoder, transform, controller);\n };\n\nconst decodeAndTransform = <StreamType, Source, Transformed = Source>(\n decoder: Decoder<StreamType, Source>,\n transform: (input: Source) => Transformed,\n controller: TransformStreamDefaultController<Transformed>,\n) => {\n try {\n const decoded = decoder.decode();\n if (!decoded) return; // TODO: Add a proper handling of decode errors\n\n const transformed = transform(decoded);\n controller.enqueue(transformed);\n } catch (error) {\n controller.error(new Error(`Decoding error: ${error?.toString()}`));\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const filter = <Item>(filter: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n if (filter(chunk)) {\n controller.enqueue(chunk);\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const map = <From, To>(map: (item: From) => To) =>\n new TransformStream<From, To>({\n transform(chunk, controller) {\n controller.enqueue(map(chunk));\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const reduce = <I, O>(\n reducer: (accumulator: O, chunk: I) => O,\n initialValue: O,\n) => new ReduceTransformStream<I, O>(reducer, initialValue);\n\nexport class ReduceTransformStream<I, O> extends TransformStream<I, O> {\n private accumulator: O;\n private reducer: (accumulator: O, chunk: I) => O;\n\n constructor(reducer: (accumulator: O, chunk: I) => O, initialValue: O) {\n super({\n transform: (chunk) => {\n this.accumulator = this.reducer(this.accumulator, chunk);\n },\n flush: (controller) => {\n controller.enqueue(this.accumulator);\n controller.terminate();\n },\n });\n\n this.accumulator = initialValue;\n this.reducer = reducer;\n }\n}\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n TransformStream,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport { type AsyncRetryOptions, asyncRetry } from '../../utils';\n\nexport const retryStream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n): TransformStream<Source, Transformed> =>\n new TransformStream<Source, Transformed>({\n start(controller) {\n asyncRetry(\n () => onRestream(createSourceStream, handleChunk, controller),\n retryOptions,\n ).catch((error) => {\n controller.error(error);\n });\n },\n });\n\nconst onRestream = async <StreamType, Source, Transformed = Source>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n controller: TransformStreamDefaultController<Transformed>,\n): Promise<void> => {\n const sourceStream = createSourceStream();\n const reader = sourceStream.getReader();\n\n try {\n let done: boolean;\n\n do {\n const result = await reader.read();\n done = result.done;\n\n await handleChunk(result, controller);\n\n if (done) {\n controller.terminate();\n }\n } while (!done);\n } finally {\n reader.releaseLock();\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const skip = <T>(limit: number) => new SkipTransformStream<T>(limit);\n\nexport class SkipTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private skip: number;\n\n constructor(skip: number) {\n super({\n transform: (chunk, controller) => {\n this.count++;\n if (this.count > this.skip) {\n controller.enqueue(chunk);\n }\n },\n });\n\n this.skip = skip;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopAfter = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n controller.enqueue(chunk);\n\n if (stopCondition(chunk)) {\n controller.terminate();\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopOn = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n async transform(chunk, controller) {\n if (!stopCondition(chunk)) {\n controller.enqueue(chunk);\n return;\n }\n await Promise.resolve();\n controller.terminate();\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const take = <T>(limit: number) => new TakeTransformStream<T>(limit);\n\nexport class TakeTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private limit: number;\n\n constructor(limit: number) {\n super({\n transform: (chunk, controller) => {\n if (this.count < this.limit) {\n this.count++;\n controller.enqueue(chunk);\n } else {\n controller.terminate();\n }\n },\n });\n\n this.limit = limit;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const waitAtMost = <Item>(waitTimeInMs: number) =>\n new TransformStream<Item, Item>({\n start(controller) {\n const timeoutId = setTimeout(() => {\n controller.terminate();\n }, waitTimeInMs);\n\n const originalTerminate = controller.terminate.bind(controller);\n\n // Clear the timeout if the stream is terminated early\n controller.terminate = () => {\n clearTimeout(timeoutId);\n originalTerminate();\n };\n },\n transform(chunk, controller) {\n controller.enqueue(chunk);\n },\n });\n","import { filter } from './filter';\nimport { map } from './map';\nimport {\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n} from './notifyAboutNoActiveReaders';\nimport { reduce, ReduceTransformStream } from './reduce';\nimport { retryStream } from './retry';\nimport { skip, SkipTransformStream } from './skip';\nimport { stopAfter } from './stopAfter';\nimport { stopOn } from './stopOn';\nimport { take, TakeTransformStream } from './take';\nimport { waitAtMost } from './waitAtMost';\n\nexport const streamTransformations = {\n filter,\n take,\n TakeTransformStream,\n skip,\n SkipTransformStream,\n map,\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n reduce,\n ReduceTransformStream,\n retry: retryStream,\n stopAfter,\n stopOn,\n waitAtMost,\n};\n","import {\n JSONParser,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AppendToStreamOptions,\n type BeforeEventStoreCommitHandler,\n type ExpectedStreamVersion,\n type Event as Message,\n type RecordedMessage,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport {\n isSQLiteError,\n type Parameters,\n type SQLiteConnection,\n type SQLiteError,\n} from '../../connection';\nimport type {\n SQLiteEventStore,\n SQLiteReadEventMetadata,\n} from '../SQLiteEventStore';\nimport { defaultTag, messagesTable, streamsTable } from './typing';\n\nexport type AppendEventResult =\n | {\n success: true;\n nextStreamPosition: bigint;\n lastGlobalPosition: bigint;\n }\n | { success: false };\n\nexport const appendToStream = async <MessageType extends Message>(\n connection: SQLiteConnection,\n streamName: string,\n streamType: string,\n messages: MessageType[],\n options?: AppendToStreamOptions & {\n partition?: string;\n onBeforeCommit?: BeforeEventStoreCommitHandler<\n SQLiteEventStore,\n { connection: SQLiteConnection }\n >;\n },\n): Promise<AppendEventResult> => {\n if (messages.length === 0) return { success: false };\n\n const expectedStreamVersion = toExpectedVersion(\n options?.expectedStreamVersion,\n );\n\n const messagesToAppend: RecordedMessage<\n MessageType,\n SQLiteReadEventMetadata\n >[] = messages.map(\n (\n m: Message,\n i: number,\n ): RecordedMessage<MessageType, SQLiteReadEventMetadata> =>\n ({\n ...m,\n kind: m.kind ?? 'Event',\n metadata: {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(i + 1),\n ...('metadata' in m ? (m.metadata ?? {}) : {}),\n },\n }) as RecordedMessage<MessageType, SQLiteReadEventMetadata>,\n );\n\n let result: AppendEventResult;\n\n return await connection.withTransaction(async () => {\n result = await appendToStreamRaw(\n connection,\n streamName,\n streamType,\n messagesToAppend,\n {\n expectedStreamVersion,\n },\n );\n\n if (options?.onBeforeCommit)\n await options.onBeforeCommit(messagesToAppend, { connection });\n\n return result;\n });\n};\n\nconst toExpectedVersion = (\n expected: ExpectedStreamVersion | undefined,\n): bigint | null => {\n if (expected === undefined) return null;\n\n if (expected === NO_CONCURRENCY_CHECK) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_DOES_NOT_EXIST) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_EXISTS) return null;\n\n return expected as bigint;\n};\n\nconst appendToStreamRaw = async (\n connection: SQLiteConnection,\n streamId: string,\n streamType: string,\n messages: RecordedMessage[],\n options?: {\n expectedStreamVersion: bigint | null;\n partition?: string;\n },\n): Promise<AppendEventResult> => {\n let streamPosition;\n let globalPosition;\n\n try {\n let expectedStreamVersion = options?.expectedStreamVersion ?? null;\n\n if (expectedStreamVersion == null) {\n expectedStreamVersion = await getLastStreamPosition(\n connection,\n streamId,\n expectedStreamVersion,\n );\n }\n\n let position: { stream_position: string } | null;\n\n if (expectedStreamVersion === 0n) {\n position = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `INSERT INTO ${streamsTable.name}\n (stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)\n VALUES (\n ?,\n ?,\n ?,\n ?,\n '[]',\n false\n )\n RETURNING stream_position;\n `,\n [\n streamId,\n messages.length,\n options?.partition ?? streamsTable.columns.partition,\n streamType,\n ],\n );\n } else {\n position = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `UPDATE ${streamsTable.name}\n SET stream_position = stream_position + ?\n WHERE stream_id = ?\n AND partition = ?\n AND is_archived = false\n RETURNING stream_position;\n `,\n [\n messages.length,\n streamId,\n options?.partition ?? streamsTable.columns.partition,\n ],\n );\n }\n\n if (position == null) {\n throw new Error('Could not find stream position');\n }\n\n streamPosition = BigInt(position.stream_position);\n\n if (expectedStreamVersion != null) {\n const expectedStreamPositionAfterSave =\n BigInt(expectedStreamVersion) + BigInt(messages.length);\n if (streamPosition !== expectedStreamPositionAfterSave) {\n return {\n success: false,\n };\n }\n }\n\n const { sqlString, values } = buildMessageInsertQuery(\n messages,\n expectedStreamVersion,\n streamId,\n options?.partition?.toString() ?? defaultTag,\n );\n\n const returningIds = await connection.query<{\n global_position: string;\n } | null>(sqlString, values);\n\n if (\n returningIds.length === 0 ||\n !returningIds[returningIds.length - 1]?.global_position\n ) {\n throw new Error('Could not find global position');\n }\n\n globalPosition = BigInt(\n returningIds[returningIds.length - 1]!.global_position,\n );\n } catch (err: unknown) {\n if (isSQLiteError(err) && isOptimisticConcurrencyError(err)) {\n return {\n success: false,\n };\n }\n\n throw err;\n }\n\n return {\n success: true,\n nextStreamPosition: streamPosition,\n lastGlobalPosition: globalPosition,\n };\n};\n\nconst isOptimisticConcurrencyError = (error: SQLiteError): boolean => {\n return error?.errno !== undefined && error.errno === 19;\n};\n\nasync function getLastStreamPosition(\n connection: SQLiteConnection,\n streamId: string,\n expectedStreamVersion: bigint | null,\n): Promise<bigint> {\n const result = await connection.querySingle<{\n stream_position: string;\n } | null>(\n `SELECT CAST(stream_position AS VARCHAR) AS stream_position FROM ${streamsTable.name} WHERE stream_id = ?`,\n [streamId],\n );\n\n if (result?.stream_position == null) {\n expectedStreamVersion = 0n;\n } else {\n expectedStreamVersion = BigInt(result.stream_position);\n }\n return expectedStreamVersion;\n}\n\nconst buildMessageInsertQuery = (\n messages: RecordedMessage[],\n expectedStreamVersion: bigint,\n streamId: string,\n partition: string | null | undefined,\n): {\n sqlString: string;\n values: Parameters[];\n} => {\n const query = messages.reduce(\n (\n queryBuilder: { parameterMarkers: string[]; values: Parameters[] },\n message: RecordedMessage,\n ) => {\n if (\n message.metadata?.streamPosition == null ||\n typeof message.metadata.streamPosition !== 'bigint'\n ) {\n throw new Error('Stream position is required');\n }\n\n const streamPosition =\n BigInt(message.metadata.streamPosition) + BigInt(expectedStreamVersion);\n\n queryBuilder.parameterMarkers.push(`(?,?,?,?,?,?,?,?,?,?)`);\n queryBuilder.values.push(\n streamId,\n streamPosition.toString() ?? 0,\n partition ?? defaultTag,\n message.kind === 'Event' ? 'E' : 'C',\n JSONParser.stringify(message.data),\n JSONParser.stringify(message.metadata),\n expectedStreamVersion?.toString() ?? 0,\n message.type,\n message.metadata.messageId,\n false,\n );\n\n return queryBuilder;\n },\n {\n parameterMarkers: [],\n values: [],\n },\n );\n\n const sqlString = `\n INSERT INTO ${messagesTable.name} (\n stream_id, \n stream_position, \n partition, \n message_kind,\n message_data, \n message_metadata, \n message_schema_version, \n message_type, \n message_id, \n is_archived\n ) \n VALUES ${query.parameterMarkers.join(', ')} \n RETURNING \n CAST(global_position as VARCHAR) AS global_position\n `;\n return { sqlString, values: query.values };\n};\n","export const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = 'emt:default';\n\nexport const globalNames = {\n module: `${emmettPrefix}:module:${globalTag}`,\n};\n\nconst columns = {\n partition: {\n name: 'partition',\n },\n isArchived: { name: 'is_archived' },\n};\n\nexport const streamsTable = {\n name: `${emmettPrefix}_streams`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const messagesTable = {\n name: `${emmettPrefix}_messages`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const subscriptionsTable = {\n name: `${emmettPrefix}_subscriptions`,\n};\n","import type { SQLiteConnection } from '../../connection';\nimport {\n globalTag,\n messagesTable,\n streamsTable,\n subscriptionsTable,\n} from './typing';\n\nexport const sql = (sql: string) => sql;\n\nexport const streamsTableSQL = sql(\n `CREATE TABLE IF NOT EXISTS ${streamsTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL DEFAULT 0,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n stream_type TEXT NOT NULL,\n stream_metadata JSONB NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n PRIMARY KEY (stream_id, partition, is_archived),\n UNIQUE (stream_id, partition, is_archived)\n );`,\n);\n\nexport const messagesTableSQL = sql(\n `CREATE TABLE IF NOT EXISTS ${messagesTable.name}(\n stream_id TEXT NOT NULL,\n stream_position BIGINT NOT NULL,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n message_kind CHAR(1) NOT NULL DEFAULT 'E',\n message_data JSONB NOT NULL,\n message_metadata JSONB NOT NULL,\n message_schema_version TEXT NOT NULL,\n message_type TEXT NOT NULL,\n message_id TEXT NOT NULL,\n is_archived BOOLEAN NOT NULL DEFAULT FALSE,\n global_position INTEGER PRIMARY KEY,\n created DATETIME DEFAULT CURRENT_TIMESTAMP,\n UNIQUE (stream_id, stream_position, partition, is_archived)\n ); \n`,\n);\n\nexport const subscriptionsTableSQL = sql(\n `\n CREATE TABLE IF NOT EXISTS ${subscriptionsTable.name}(\n subscription_id TEXT NOT NULL,\n version INTEGER NOT NULL DEFAULT 1,\n partition TEXT NOT NULL DEFAULT '${globalTag}',\n last_processed_position BIGINT NOT NULL,\n PRIMARY KEY (subscription_id, partition, version)\n );\n`,\n);\n\nexport const schemaSQL: string[] = [\n streamsTableSQL,\n messagesTableSQL,\n subscriptionsTableSQL,\n];\n\nexport const createEventStoreSchema = async (\n db: SQLiteConnection,\n): Promise<void> => {\n for (const sql of schemaSQL) {\n await db.command(sql);\n }\n};\n","export const singleOrNull = async <T>(\n getResult: Promise<T[]>,\n): Promise<T | null> => {\n const result = await getResult;\n\n if (result.length > 1) throw new Error('Query had more than one result');\n\n return result.length > 0 ? (result[0] ?? null) : null;\n};\n\nexport const single = async <T>(getResult: Promise<T[]>): Promise<T> => {\n const result = await getResult;\n\n if (result.length === 0) throw new Error(\"Query didn't return any result\");\n\n if (result.length > 1) throw new Error('Query had more than one result');\n\n return result[0]!;\n};\n","import type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, messagesTable } from './typing';\nimport { singleOrNull } from './utils';\n\ntype ReadLastMessageGlobalPositionSqlResult = {\n global_position: string;\n};\n\nexport type ReadLastMessageGlobalPositionResult = {\n currentGlobalPosition: bigint | null;\n};\n\nexport const readLastMessageGlobalPosition = async (\n db: SQLiteConnection,\n options?: { partition?: string },\n): Promise<ReadLastMessageGlobalPositionResult> => {\n const result = await singleOrNull(\n db.query<ReadLastMessageGlobalPositionSqlResult>(\n sql(\n `SELECT global_position\n FROM ${messagesTable.name}\n WHERE partition = ? AND is_archived = FALSE\n ORDER BY global_position\n LIMIT 1`,\n ),\n [options?.partition ?? defaultTag],\n ),\n );\n\n return {\n currentGlobalPosition:\n result !== null ? BigInt(result.global_position) : null,\n };\n};\n","import {\n JSONParser,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ReadEventMetadata,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, messagesTable } from './typing';\n\ntype ReadMessagesBatchSqlResult = {\n stream_position: string;\n stream_id: string;\n message_data: string;\n message_metadata: string;\n message_schema_version: string;\n message_type: string;\n message_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport type ReadMessagesBatchOptions =\n | {\n after: bigint;\n batchSize: number;\n }\n | {\n from: bigint;\n batchSize: number;\n }\n | { to: bigint; batchSize: number }\n | { from: bigint; to: bigint };\n\nexport type ReadMessagesBatchResult<\n EventType extends Event,\n ReadEventMetadataType extends ReadEventMetadata = ReadEventMetadata,\n> = {\n currentGlobalPosition: bigint;\n messages: ReadEvent<EventType, ReadEventMetadataType>[];\n areEventsLeft: boolean;\n};\n\nexport const readMessagesBatch = async <\n MessageType extends Event,\n ReadEventMetadataType extends\n ReadEventMetadataWithGlobalPosition = ReadEventMetadataWithGlobalPosition,\n>(\n db: SQLiteConnection,\n options: ReadMessagesBatchOptions & { partition?: string },\n): Promise<ReadMessagesBatchResult<MessageType, ReadEventMetadataType>> => {\n const from =\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after + 1n\n : 0n;\n const batchSize =\n options && 'batchSize' in options\n ? options.batchSize\n : options.to - options.from;\n\n const fromCondition: string =\n from !== -0n ? `AND global_position >= ${from}` : '';\n\n const toCondition =\n 'to' in options ? `AND global_position <= ${options.to}` : '';\n\n const limitCondition =\n 'batchSize' in options ? `LIMIT ${options.batchSize}` : '';\n\n const events: ReadEvent<MessageType, ReadEventMetadataType>[] = (\n await db.query<ReadMessagesBatchSqlResult>(\n sql(\n `SELECT stream_id, stream_position, global_position, message_data, message_metadata, message_schema_version, message_type, message_id\n FROM ${messagesTable.name}\n WHERE partition = ? AND is_archived = FALSE ${fromCondition} ${toCondition}\n ORDER BY global_position\n ${limitCondition}`,\n ),\n [options?.partition ?? defaultTag],\n )\n ).map((row) => {\n const rawEvent = {\n type: row.message_type,\n data: JSONParser.parse(row.message_data),\n metadata: JSONParser.parse(row.message_metadata),\n } as unknown as MessageType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n messageId: row.message_id,\n streamName: row.stream_id,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n kind: 'Event',\n metadata: metadata as CombinedReadEventMetadata<\n MessageType,\n ReadEventMetadataType\n >,\n };\n });\n\n return events.length > 0\n ? {\n currentGlobalPosition:\n events[events.length - 1]!.metadata.globalPosition,\n messages: events,\n areEventsLeft: events.length === batchSize,\n }\n : {\n currentGlobalPosition:\n 'from' in options\n ? options.from\n : 'after' in options\n ? options.after\n : 0n,\n messages: [],\n areEventsLeft: false,\n };\n};\n","import type { SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, subscriptionsTable } from './typing';\nimport { singleOrNull } from './utils';\n\ntype ReadProcessorCheckpointSqlResult = {\n last_processed_position: string;\n};\n\nexport type ReadProcessorCheckpointResult = {\n lastProcessedPosition: bigint | null;\n};\n\nexport const readProcessorCheckpoint = async (\n db: SQLiteConnection,\n options: { processorId: string; partition?: string },\n): Promise<ReadProcessorCheckpointResult> => {\n const result = await singleOrNull(\n db.query<ReadProcessorCheckpointSqlResult>(\n sql(\n `SELECT last_processed_position\n FROM ${subscriptionsTable.name}\n WHERE partition = ? AND subscription_id = ?\n LIMIT 1`,\n ),\n [options?.partition ?? defaultTag, options.processorId],\n ),\n );\n\n return {\n lastProcessedPosition:\n result !== null ? BigInt(result.last_processed_position) : null,\n };\n};\n","import type {\n EmmettError,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../../connection';\nimport { readLastMessageGlobalPosition } from '../../schema/readLastMessageGlobalPosition';\nimport {\n readMessagesBatch,\n type ReadMessagesBatchOptions,\n} from '../../schema/readMessagesBatch';\n\nexport const DefaultSQLiteEventStoreProcessorBatchSize = 100;\nexport const DefaultSQLiteEventStoreProcessorPullingFrequencyInMs = 50;\n\nexport type SQLiteEventStoreMessagesBatch<EventType extends Event = Event> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type SQLiteEventStoreMessagesBatchHandlerResult = void | {\n type: 'STOP';\n reason?: string;\n error?: EmmettError;\n};\n\nexport type SQLiteEventStoreMessagesBatchHandler<\n EventType extends Event = Event,\n> = (\n messagesBatch: SQLiteEventStoreMessagesBatch<EventType>,\n) =>\n | Promise<SQLiteEventStoreMessagesBatchHandlerResult>\n | SQLiteEventStoreMessagesBatchHandlerResult;\n\nexport type SQLiteEventStoreMessageBatchPullerOptions<\n EventType extends Event = Event,\n> = {\n connection: SQLiteConnection;\n pullingFrequencyInMs: number;\n batchSize: number;\n eachBatch: SQLiteEventStoreMessagesBatchHandler<EventType>;\n};\n\nexport type SQLiteEventStoreMessageBatchPullerStartFrom =\n | { globalPosition: bigint }\n | 'BEGINNING'\n | 'END';\n\nexport type SQLiteEventStoreMessageBatchPullerStartOptions = {\n startFrom: SQLiteEventStoreMessageBatchPullerStartFrom;\n};\n\nexport type SQLiteEventStoreMessageBatchPuller = {\n isRunning: boolean;\n start(options: SQLiteEventStoreMessageBatchPullerStartOptions): Promise<void>;\n stop(): Promise<void>;\n};\n\nexport const sqliteEventStoreMessageBatchPuller = <\n EventType extends Event = Event,\n>({\n connection,\n batchSize,\n eachBatch,\n pullingFrequencyInMs,\n}: SQLiteEventStoreMessageBatchPullerOptions<EventType>): SQLiteEventStoreMessageBatchPuller => {\n let isRunning = false;\n\n let start: Promise<void>;\n\n const pullMessages = async (\n options: SQLiteEventStoreMessageBatchPullerStartOptions,\n ) => {\n const after =\n options.startFrom === 'BEGINNING'\n ? 0n\n : options.startFrom === 'END'\n ? ((await readLastMessageGlobalPosition(connection))\n .currentGlobalPosition ?? 0n)\n : options.startFrom.globalPosition;\n\n const readMessagesOptions: ReadMessagesBatchOptions = {\n after,\n batchSize,\n };\n\n let waitTime = 100;\n\n do {\n const { messages, currentGlobalPosition, areEventsLeft } =\n await readMessagesBatch<EventType>(connection, readMessagesOptions);\n\n if (messages.length > 0) {\n const result = await eachBatch({ messages });\n\n if (result && result.type === 'STOP') {\n isRunning = false;\n break;\n }\n }\n\n readMessagesOptions.after = currentGlobalPosition;\n\n await new Promise((resolve) => setTimeout(resolve, waitTime));\n\n if (!areEventsLeft) {\n waitTime = Math.min(waitTime * 2, 1000);\n } else {\n waitTime = pullingFrequencyInMs;\n }\n } while (isRunning);\n };\n\n return {\n get isRunning() {\n return isRunning;\n },\n start: (options) => {\n if (isRunning) return start;\n\n start = (async () => {\n isRunning = true;\n\n return pullMessages(options);\n })();\n\n return start;\n },\n stop: async () => {\n if (!isRunning) return;\n isRunning = false;\n await start;\n },\n };\n};\n\nexport const zipSQLiteEventStoreMessageBatchPullerStartFrom = (\n options: (SQLiteEventStoreMessageBatchPullerStartFrom | undefined)[],\n): SQLiteEventStoreMessageBatchPullerStartFrom => {\n if (\n options.length === 0 ||\n options.some((o) => o === undefined || o === 'BEGINNING')\n )\n return 'BEGINNING';\n\n if (options.every((o) => o === 'END')) return 'END';\n\n return options\n .filter((o) => o !== undefined && o !== 'BEGINNING' && o !== 'END')\n .sort((a, b) => (a > b ? 1 : -1))[0]!;\n};\n","import {\n EmmettError,\n getCheckpoint,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\nimport { sqliteConnection, type SQLiteConnection } from '../../connection';\nimport type { SQLiteProjectionDefinition } from '../projections';\nimport { readProcessorCheckpoint, storeProcessorCheckpoint } from '../schema';\nimport type { SQLiteEventStoreMessageBatchPullerStartFrom } from './messageBatchProcessing';\n\nexport type SQLiteProcessorEventsBatch<EventType extends Event = Event> = {\n messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[];\n};\n\nexport type SQLiteProcessorHandlerContext = {\n connection: SQLiteConnection;\n fileName: string;\n};\n\nexport type SQLiteProcessor<EventType extends Event = Event> = {\n id: string;\n start: (\n connection: SQLiteConnection,\n ) => Promise<SQLiteEventStoreMessageBatchPullerStartFrom | undefined>;\n isActive: boolean;\n handle: (\n messagesBatch: SQLiteProcessorEventsBatch<EventType>,\n context: { connection?: SQLiteConnection; fileName?: string },\n ) => Promise<SQLiteProcessorMessageHandlerResult>;\n};\n\nexport const SQLiteProcessor = {\n result: {\n skip: (options?: {\n reason?: string;\n }): SQLiteProcessorMessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): SQLiteProcessorMessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type SQLiteProcessorMessageHandlerResult =\n | void\n | { type: 'SKIP'; reason?: string }\n | { type: 'STOP'; reason?: string; error?: EmmettError };\n\nexport type SQLiteProcessorEachMessageHandler<EventType extends Event = Event> =\n (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n context: SQLiteProcessorHandlerContext,\n ) =>\n | Promise<SQLiteProcessorMessageHandlerResult>\n | SQLiteProcessorMessageHandlerResult;\n\nexport type SQLiteProcessorEachBatchHandler<EventType extends Event = Event> = (\n event: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[],\n context: SQLiteProcessorHandlerContext,\n) =>\n | Promise<SQLiteProcessorMessageHandlerResult>\n | SQLiteProcessorMessageHandlerResult;\n\nexport type SQLiteProcessorStartFrom =\n | SQLiteEventStoreMessageBatchPullerStartFrom\n | 'CURRENT';\n\nexport type SQLiteProcessorConnectionOptions = {\n fileName: string;\n connection?: SQLiteConnection;\n};\n\nexport type GenericSQLiteProcessorOptions<EventType extends Event = Event> = {\n processorId: string;\n version?: number;\n partition?: string;\n startFrom?: SQLiteProcessorStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n eachMessage: SQLiteProcessorEachMessageHandler<EventType>;\n connectionOptions?: SQLiteProcessorConnectionOptions;\n // TODO: Add eachBatch\n};\n\nexport type SQLiteProjectionProcessorOptions<EventType extends Event = Event> =\n {\n processorId?: string;\n version?: number;\n projection: SQLiteProjectionDefinition<EventType>;\n partition?: string;\n startFrom?: SQLiteProcessorStartFrom;\n stopAfter?: (\n message: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>,\n ) => boolean;\n };\n\nexport type SQLiteProcessorOptions<EventType extends Event = Event> =\n | GenericSQLiteProcessorOptions<EventType>\n | SQLiteProjectionProcessorOptions<EventType>;\n\nconst genericSQLiteProcessor = <EventType extends Event = Event>(\n options: GenericSQLiteProcessorOptions<EventType>,\n): SQLiteProcessor => {\n const { eachMessage } = options;\n let isActive = true;\n //let lastProcessedPosition: number | null = null;\n\n const getDb = (context: {\n connection?: SQLiteConnection;\n fileName?: string;\n }): { connection: SQLiteConnection; fileName: string } => {\n const fileName = context.fileName ?? options.connectionOptions?.fileName;\n if (!fileName)\n throw new EmmettError(\n `SQLite processor '${options.processorId}' is missing file name. Ensure that you passed it through options`,\n );\n\n const connection =\n context.connection ??\n options.connectionOptions?.connection ??\n sqliteConnection({ fileName });\n\n return { connection, fileName };\n };\n\n return {\n id: options.processorId,\n start: async (\n connection: SQLiteConnection,\n ): Promise<SQLiteEventStoreMessageBatchPullerStartFrom | undefined> => {\n isActive = true;\n if (options.startFrom !== 'CURRENT') return options.startFrom;\n\n const { lastProcessedPosition } = await readProcessorCheckpoint(\n connection,\n {\n processorId: options.processorId,\n partition: options.partition,\n },\n );\n\n if (lastProcessedPosition === null) return 'BEGINNING';\n\n return { globalPosition: lastProcessedPosition };\n },\n get isActive() {\n return isActive;\n },\n handle: async (\n { messages },\n context,\n ): Promise<SQLiteProcessorMessageHandlerResult> => {\n if (!isActive) return;\n\n const { connection, fileName } = getDb(context);\n\n return connection.withTransaction(async () => {\n let result: SQLiteProcessorMessageHandlerResult | undefined = undefined;\n\n let lastProcessedPosition: bigint | null = null;\n\n for (const message of messages) {\n const typedMessage = message as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >;\n\n const messageProcessingResult = await eachMessage(typedMessage, {\n connection,\n fileName,\n });\n\n const newPosition: bigint | null = getCheckpoint(typedMessage);\n\n // TODO: Add correct handling of the storing checkpoint\n await storeProcessorCheckpoint(connection, {\n processorId: options.processorId,\n version: options.version,\n lastProcessedPosition,\n newPosition,\n partition: options.partition,\n });\n\n lastProcessedPosition = typedMessage.metadata.globalPosition;\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (options.stopAfter && options.stopAfter(typedMessage)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n return result;\n });\n },\n };\n};\n\nexport const sqliteProjectionProcessor = <EventType extends Event = Event>(\n options: SQLiteProjectionProcessorOptions<EventType>,\n): SQLiteProcessor => {\n const projection = options.projection;\n\n return genericSQLiteProcessor<EventType>({\n processorId: options.processorId ?? `projection:${projection.name}`,\n eachMessage: async (event, context) => {\n if (!projection.canHandle.includes(event.type)) return;\n\n await projection.handle([event], { connection: context.connection });\n },\n ...options,\n });\n};\n\nexport const sqliteProcessor = <EventType extends Event = Event>(\n options: SQLiteProcessorOptions<EventType>,\n): SQLiteProcessor => {\n if ('projection' in options) {\n return sqliteProjectionProcessor(options);\n }\n\n return genericSQLiteProcessor(options);\n};\n","import { EmmettError, type Event } from '@event-driven-io/emmett';\nimport { sqliteConnection, type SQLiteConnection } from '../../connection';\nimport {\n DefaultSQLiteEventStoreProcessorBatchSize,\n DefaultSQLiteEventStoreProcessorPullingFrequencyInMs,\n sqliteEventStoreMessageBatchPuller,\n zipSQLiteEventStoreMessageBatchPullerStartFrom,\n type SQLiteEventStoreMessageBatchPuller,\n type SQLiteEventStoreMessagesBatchHandler,\n} from './messageBatchProcessing';\nimport {\n sqliteProcessor,\n type SQLiteProcessor,\n type SQLiteProcessorOptions,\n} from './sqliteProcessor';\n\nexport type SQLiteEventStoreConsumerConfig<\n ConsumerEventType extends Event = Event,\n> = {\n processors?: SQLiteProcessor<ConsumerEventType>[];\n pulling?: {\n batchSize?: number;\n pullingFrequencyInMs?: number;\n };\n};\nexport type SQLiteEventStoreConsumerOptions<\n ConsumerEventType extends Event = Event,\n> = SQLiteEventStoreConsumerConfig<ConsumerEventType> & {\n fileName: string;\n connection?: SQLiteConnection;\n};\n\nexport type SQLiteEventStoreConsumer<ConsumerEventType extends Event = Event> =\n Readonly<{\n isRunning: boolean;\n processors: SQLiteProcessor<ConsumerEventType>[];\n processor: <EventType extends ConsumerEventType = ConsumerEventType>(\n options: SQLiteProcessorOptions<EventType>,\n ) => SQLiteProcessor<EventType>;\n start: () => Promise<void>;\n stop: () => Promise<void>;\n close: () => Promise<void>;\n }>;\n\nexport const sqliteEventStoreConsumer = <\n ConsumerEventType extends Event = Event,\n>(\n options: SQLiteEventStoreConsumerOptions<ConsumerEventType>,\n): SQLiteEventStoreConsumer<ConsumerEventType> => {\n let isRunning = false;\n const { pulling } = options;\n const processors = options.processors ?? [];\n\n let start: Promise<void>;\n\n let currentMessagePuller: SQLiteEventStoreMessageBatchPuller | undefined;\n\n const connection =\n options.connection ?? sqliteConnection({ fileName: options.fileName });\n\n const eachBatch: SQLiteEventStoreMessagesBatchHandler<\n ConsumerEventType\n > = async (messagesBatch) => {\n const activeProcessors = processors.filter((s) => s.isActive);\n\n if (activeProcessors.length === 0)\n return {\n type: 'STOP',\n reason: 'No active processors',\n };\n\n const result = await Promise.allSettled(\n activeProcessors.map((s) => {\n // TODO: Add here filtering to only pass messages that can be handled by processor\n return s.handle(messagesBatch, {\n connection,\n fileName: options.fileName,\n });\n }),\n );\n\n return result.some(\n (r) => r.status === 'fulfilled' && r.value?.type !== 'STOP',\n )\n ? undefined\n : {\n type: 'STOP',\n };\n };\n\n const messagePooler = (currentMessagePuller =\n sqliteEventStoreMessageBatchPuller({\n connection,\n eachBatch,\n batchSize:\n pulling?.batchSize ?? DefaultSQLiteEventStoreProcessorBatchSize,\n pullingFrequencyInMs:\n pulling?.pullingFrequencyInMs ??\n DefaultSQLiteEventStoreProcessorPullingFrequencyInMs,\n }));\n\n const stop = async () => {\n if (!isRunning) return;\n isRunning = false;\n if (currentMessagePuller) {\n await currentMessagePuller.stop();\n currentMessagePuller = undefined;\n }\n await start;\n };\n\n return {\n processors,\n get isRunning() {\n return isRunning;\n },\n processor: <EventType extends ConsumerEventType = ConsumerEventType>(\n options: SQLiteProcessorOptions<EventType>,\n ): SQLiteProcessor<EventType> => {\n const processor = sqliteProcessor<EventType>(options);\n\n processors.push(processor);\n\n return processor;\n },\n start: () => {\n if (isRunning) return start;\n\n start = (async () => {\n if (processors.length === 0)\n return Promise.reject(\n new EmmettError(\n 'Cannot start consumer without at least a single processor',\n ),\n );\n\n isRunning = true;\n\n const startFrom = zipSQLiteEventStoreMessageBatchPullerStartFrom(\n await Promise.all(processors.map((o) => o.start(connection))),\n );\n\n return messagePooler.start({ startFrom });\n })();\n\n return start;\n },\n stop,\n close: async () => {\n await stop();\n\n connection.close();\n\n await new Promise((resolve) => setTimeout(resolve, 250));\n },\n };\n};\n","import {\n projection,\n type CanHandle,\n type Event,\n type ProjectionDefinition,\n type ProjectionHandler,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport type { SQLiteConnection } from '../../connection';\nimport type { SQLiteReadEventMetadata } from '../SQLiteEventStore';\n\nexport type SQLiteProjectionHandlerContext = {\n connection: SQLiteConnection;\n};\n\nexport type SQLiteProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends SQLiteReadEventMetadata = SQLiteReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n SQLiteProjectionHandlerContext\n>;\n\nexport type SQLiteProjectionDefinition<EventType extends Event = Event> =\n ProjectionDefinition<\n EventType,\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >;\n\nexport type SQLiteProjectionHandlerOptions<EventType extends Event = Event> = {\n events: ReadEvent<EventType, SQLiteReadEventMetadata>[];\n projections: SQLiteProjectionDefinition<EventType>[];\n connection: SQLiteConnection;\n};\n\nexport const handleProjections = async <EventType extends Event = Event>(\n options: SQLiteProjectionHandlerOptions<EventType>,\n): Promise<void> => {\n const { projections: allProjections, events, connection } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n for (const projection of allProjections) {\n if (!projection.canHandle.some((type) => eventTypes.includes(type))) {\n continue;\n }\n await projection.handle(events, {\n connection,\n });\n }\n};\n\nexport const sqliteProjection = <EventType extends Event>(\n definition: SQLiteProjectionDefinition<EventType>,\n): SQLiteProjectionDefinition<EventType> =>\n projection<\n EventType,\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >(definition);\n\nexport const sqliteRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: SQLiteProjectionHandlerContext,\n ) => Promise<string[]> | string[],\n ...canHandle: CanHandle<EventType>\n): SQLiteProjectionDefinition<EventType> =>\n sqliteProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: string[] = await handle(events, context);\n\n for (const sql of sqls) await context.connection.command(sql);\n },\n });\n\nexport const sqliteRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: SQLiteProjectionHandlerContext,\n ) => Promise<string> | string,\n ...canHandle: CanHandle<EventType>\n): SQLiteProjectionDefinition<EventType> =>\n sqliteRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: string[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n","import type {\n AppendToStreamResultWithGlobalPosition,\n BeforeEventStoreCommitHandler,\n BigIntStreamPosition,\n Event,\n ProjectionRegistration,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '@event-driven-io/emmett';\n\nimport {\n assertExpectedVersionMatchesCurrent,\n ExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport {\n InMemorySharedCacheSQLiteDatabase,\n InMemorySQLiteDatabase,\n sqliteConnection,\n type SQLiteConnection,\n} from '../connection';\nimport {\n sqliteEventStoreConsumer,\n type SQLiteEventStoreConsumer,\n type SQLiteEventStoreConsumerConfig,\n} from './consumers';\nimport {\n handleProjections,\n type SQLiteProjectionHandlerContext,\n} from './projections';\nimport { createEventStoreSchema } from './schema';\nimport { appendToStream } from './schema/appendToStream';\nimport { readStream } from './schema/readStream';\n\nexport type EventHandler<E extends Event = Event> = (\n eventEnvelope: ReadEvent<E>,\n) => void;\n\nexport const SQLiteEventStoreDefaultStreamVersion = 0n;\n\nexport interface SQLiteEventStore extends EventStore<SQLiteReadEventMetadata> {\n appendToStream<EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition>;\n consumer<ConsumerEventType extends Event = Event>(\n options?: SQLiteEventStoreConsumerConfig<ConsumerEventType>,\n ): SQLiteEventStoreConsumer<ConsumerEventType>;\n}\n\nexport type SQLiteReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type SQLiteReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n SQLiteReadEventMetadata\n>;\n\nexport type SQLiteEventStoreOptions = {\n fileName: // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n | InMemorySQLiteDatabase\n // eslint-disable-next-line @typescript-eslint/no-redundant-type-constituents\n | InMemorySharedCacheSQLiteDatabase\n | string\n | undefined;\n projections?: ProjectionRegistration<\n 'inline',\n SQLiteReadEventMetadata,\n SQLiteProjectionHandlerContext\n >[];\n schema?: {\n autoMigration?: 'None' | 'CreateOrUpdate';\n };\n hooks?: {\n /**\n * This hook will be called **BEFORE** events were stored in the event store.\n * @type {BeforeEventStoreCommitHandler<SQLiteEventStore, HandlerContext>}\n */\n onBeforeCommit?: BeforeEventStoreCommitHandler<\n SQLiteEventStore,\n { connection: SQLiteConnection }\n >;\n };\n};\n\nexport const getSQLiteEventStore = (\n options: SQLiteEventStoreOptions,\n): SQLiteEventStore => {\n let schemaMigrated = false;\n let autoGenerateSchema = false;\n let database: SQLiteConnection | null;\n const fileName = options.fileName ?? InMemorySQLiteDatabase;\n\n const isInMemory: boolean =\n fileName === InMemorySQLiteDatabase ||\n fileName === InMemorySharedCacheSQLiteDatabase;\n\n const inlineProjections = (options.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n const onBeforeCommitHook = options.hooks?.onBeforeCommit;\n\n const createConnection = () => {\n if (database != null) {\n return database;\n }\n\n return sqliteConnection({\n fileName,\n });\n };\n\n const closeConnection = () => {\n if (isInMemory) {\n return;\n }\n if (database != null) {\n database.close();\n database = null;\n }\n };\n\n const withConnection = async <Result>(\n handler: (connection: SQLiteConnection) => Promise<Result>,\n ): Promise<Result> => {\n if (database == null) {\n database = createConnection();\n }\n\n try {\n await ensureSchemaExists(database);\n return await handler(database);\n } finally {\n closeConnection();\n }\n };\n\n if (options) {\n autoGenerateSchema =\n options.schema?.autoMigration === undefined ||\n options.schema?.autoMigration !== 'None';\n }\n\n const ensureSchemaExists = async (\n connection: SQLiteConnection,\n ): Promise<void> => {\n if (!autoGenerateSchema) return Promise.resolve();\n\n if (!schemaMigrated) {\n await createEventStoreSchema(connection);\n schemaMigrated = true;\n }\n\n return Promise.resolve();\n };\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const expectedStreamVersion = read?.expectedStreamVersion;\n\n let state = initialState();\n\n if (typeof streamName !== 'string') {\n throw new Error('Stream name is not string');\n }\n\n if (database == null) {\n database = createConnection();\n }\n\n const result = await withConnection((connection) =>\n readStream<EventType>(connection, streamName, options.read),\n );\n\n const currentStreamVersion = result.currentStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n SQLiteEventStoreDefaultStreamVersion,\n );\n\n for (const event of result.events) {\n if (!event) continue;\n\n state = evolve(state, event);\n }\n\n return {\n currentStreamVersion: currentStreamVersion,\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: async <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions<BigIntStreamPosition>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > =>\n withConnection((connection) =>\n readStream<EventType>(connection, streamName, options),\n ),\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResultWithGlobalPosition> => {\n if (database == null) {\n database = createConnection();\n }\n\n // TODO: This has to be smarter when we introduce urn-based resolution\n const [firstPart, ...rest] = streamName.split('-');\n\n const streamType =\n firstPart && rest.length > 0 ? firstPart : 'emt:unknown';\n\n const appendResult = await withConnection((connection) =>\n appendToStream(connection, streamName, streamType, events, {\n ...options,\n onBeforeCommit: async (messages, context) => {\n if (inlineProjections.length > 0)\n await handleProjections({\n projections: inlineProjections,\n events: messages,\n ...context,\n });\n\n if (onBeforeCommitHook) await onBeforeCommitHook(messages, context);\n },\n }),\n );\n\n if (!appendResult.success)\n throw new ExpectedVersionConflictError<bigint>(\n -1n, //TODO: Return actual version in case of error\n options?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n );\n\n return {\n nextExpectedStreamVersion: appendResult.nextStreamPosition,\n lastEventGlobalPosition: appendResult.lastGlobalPosition,\n createdNewStream:\n appendResult.nextStreamPosition >= BigInt(events.length),\n };\n },\n consumer: <ConsumerEventType extends Event = Event>(\n options?: SQLiteEventStoreConsumerConfig<ConsumerEventType>,\n ): SQLiteEventStoreConsumer<ConsumerEventType> =>\n sqliteEventStoreConsumer<ConsumerEventType>({\n ...(options ?? {}),\n fileName,\n connection: database ?? undefined,\n }),\n };\n};\n","import {\n JSONParser,\n type CombinedReadEventMetadata,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { type SQLiteConnection } from '../../connection';\nimport { SQLiteEventStoreDefaultStreamVersion } from '../SQLiteEventStore';\nimport { defaultTag, messagesTable } from './typing';\n\ntype ReadStreamSqlResult = {\n stream_position: string;\n message_data: string;\n message_metadata: string;\n message_schema_version: string;\n message_type: string;\n message_id: string;\n global_position: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n db: SQLiteConnection,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const results = await db.query<ReadStreamSqlResult>(\n `SELECT stream_id, stream_position, global_position, message_data, message_metadata, message_schema_version, message_type, message_id\n FROM ${messagesTable.name}\n WHERE stream_id = ? AND partition = ? AND is_archived = FALSE ${fromCondition} ${toCondition}\n ORDER BY stream_position ASC`,\n [streamId, options?.partition ?? defaultTag],\n );\n\n const messages: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n results.map((row) => {\n const rawEvent = {\n type: row.message_type,\n data: JSONParser.parse(row.message_data),\n metadata: JSONParser.parse(row.message_metadata),\n } as unknown as EventType;\n\n const metadata: ReadEventMetadataWithGlobalPosition = {\n ...('metadata' in rawEvent ? (rawEvent.metadata ?? {}) : {}),\n messageId: row.message_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n };\n\n return {\n ...rawEvent,\n kind: 'Event',\n metadata: metadata as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n return messages.length > 0\n ? {\n currentStreamVersion:\n messages[messages.length - 1]!.metadata.streamPosition,\n events: messages,\n streamExists: true,\n }\n : {\n currentStreamVersion: SQLiteEventStoreDefaultStreamVersion,\n events: [],\n streamExists: false,\n };\n};\n","import { isSQLiteError, type SQLiteConnection } from '../../connection';\nimport { sql } from './tables';\nimport { defaultTag, subscriptionsTable } from './typing';\nimport { singleOrNull } from './utils';\n\n// for more infos see the postgresql stored procedure version\nasync function storeSubscriptionCheckpointSQLite(\n db: SQLiteConnection,\n processorId: string,\n version: number,\n position: bigint | null,\n checkPosition: bigint | null,\n partition: string,\n): Promise<0 | 1 | 2> {\n if (checkPosition !== null) {\n const updateResult = await db.command(\n sql(`\n UPDATE ${subscriptionsTable.name}\n SET last_processed_position = ?\n WHERE subscription_id = ? \n AND last_processed_position = ? \n AND partition = ?\n `),\n [position!.toString(), processorId, checkPosition.toString(), partition],\n );\n if (updateResult.changes > 0) {\n return 1;\n } else {\n const current_position = await singleOrNull(\n db.query<{ last_processed_position: bigint }>(\n sql(\n `SELECT last_processed_position FROM ${subscriptionsTable.name} \n WHERE subscription_id = ? AND partition = ?`,\n ),\n [processorId, partition],\n ),\n );\n\n if (current_position?.last_processed_position === position) {\n return 0;\n } else if (\n position !== null &&\n current_position !== null &&\n current_position?.last_processed_position > position\n ) {\n return 2;\n } else {\n return 2;\n }\n }\n } else {\n try {\n await db.command(\n sql(\n `INSERT INTO ${subscriptionsTable.name} (subscription_id, version, last_processed_position, partition) VALUES (?, ?, ?, ?)`,\n ),\n [processorId, version, position!.toString(), partition],\n );\n return 1;\n } catch (err) {\n if (!(isSQLiteError(err) && (err.errno === 19 || err.errno === 2067))) {\n throw err;\n }\n\n const current = await singleOrNull(\n db.query<{ last_processed_position: bigint }>(\n sql(\n `SELECT last_processed_position FROM ${subscriptionsTable.name} WHERE subscription_id = ? AND partition = ?`,\n ),\n [processorId, partition],\n ),\n );\n if (current?.last_processed_position === position) {\n return 0;\n } else {\n return 2;\n }\n }\n }\n}\n\nexport type StoreLastProcessedProcessorPositionResult<\n Position extends bigint | null = bigint,\n> =\n | {\n success: true;\n newPosition: Position;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' };\n\nexport async function storeProcessorCheckpoint(\n db: SQLiteConnection,\n options: {\n processorId: string;\n version: number | undefined;\n newPosition: bigint | null;\n lastProcessedPosition: bigint | null;\n partition?: string;\n },\n): Promise<StoreLastProcessedProcessorPositionResult<bigint | null>> {\n try {\n const result = await storeSubscriptionCheckpointSQLite(\n db,\n options.processorId,\n options.version ?? 1,\n options.newPosition,\n options.lastProcessedPosition,\n options.partition ?? defaultTag,\n );\n\n return result === 1\n ? { success: true, newPosition: options.newPosition }\n : { success: false, reason: result === 0 ? 'IGNORED' : 'MISMATCH' };\n } catch (error) {\n console.log(error);\n throw error;\n }\n}\n"],"mappings":";AAAA,OAAO,aAAa;AAgBb,IAAM,gBAAgB,CAAC,UAAyC;AACrE,MAAI,iBAAiB,SAAS,UAAU,OAAO;AAC7C,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAGO,IAAM,oCAAoC;AAE1C,IAAM,yBAAyB;AAO/B,IAAM,mBAAmB,CAC9B,YACqB;AACrB,QAAM,WAAW,QAAQ,YAAY;AACrC,MAAI;AAEJ,MAAI,SAAS,WAAW,OAAO,GAAG;AAChC,SAAK,IAAI,QAAQ;AAAA,MACf;AAAA,MACA,QAAQ,WAAW,QAAQ,iBAAiB,QAAQ;AAAA,IACtD;AAAA,EACF,OAAO;AACL,SAAK,IAAI,QAAQ,SAAS,QAAQ;AAAA,EACpC;AACA,KAAG,IAAI,4BAA4B;AACnC,MAAI,qBAAqB;AAEzB,SAAO;AAAA,IACL,OAAO,MAAY,GAAG,MAAM;AAAA,IAC5B,SAAS,CAACA,MAAa,WACrB,IAAI,QAA2B,CAAC,SAAS,WAAW;AAClD,SAAG;AAAA,QACDA;AAAA,QACA,UAAU,CAAC;AAAA,QACX,SAAmC,KAAmB;AACpD,cAAI,KAAK;AACP,mBAAO,GAAG;AACV;AAAA,UACF;AAEA,kBAAQ,IAAI;AAAA,QACd;AAAA,MACF;AAAA,IACF,CAAC;AAAA,IACH,OAAO,CAAIA,MAAa,WACtB,IAAI,QAAQ,CAAC,SAAS,WAAW;AAC/B,SAAG,IAAIA,MAAK,UAAU,CAAC,GAAG,CAAC,KAAmB,WAAgB;AAC5D,YAAI,KAAK;AACP,iBAAO,GAAG;AACV;AAAA,QACF;AAEA,gBAAQ,MAAM;AAAA,MAChB,CAAC;AAAA,IACH,CAAC;AAAA,IACH,aAAa,CAAIA,MAAa,WAC5B,IAAI,QAAQ,CAAC,SAAS,WAAW;AAC/B,SAAG,IAAIA,MAAK,UAAU,CAAC,GAAG,CAAC,KAAmB,WAAqB;AACjE,YAAI,KAAK;AACP,iBAAO,GAAG;AACV;AAAA,QACF;AAEA,gBAAQ,MAAM;AAAA,MAChB,CAAC;AAAA,IACH,CAAC;AAAA,IACH,iBAAiB,OAAU,OAAyB;AAClD,UAAI;AACF,YAAI,wBAAwB,GAAG;AAC7B,gBAAM,iBAAiB,EAAE;AAAA,QAC3B;AACA,cAAM,SAAS,MAAM,GAAG;AAExB,YAAI,uBAAuB,EAAG,OAAM,kBAAkB,EAAE;AACxD;AAEA,eAAO;AAAA,MACT,SAAS,KAAK;AACZ,gBAAQ,IAAI,GAAG;AAEf,YAAI,EAAE,uBAAuB,EAAG,OAAM,oBAAoB,EAAE;AAE5D,cAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;AAEA,IAAM,mBAAmB,CAAC,OACxB,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,+BAA+B,CAAC,QAAsB;AAC3D,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;AAEH,IAAM,oBAAoB,CAAC,OACzB,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,UAAU,CAAC,QAAsB;AACtC,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;AAEH,IAAM,sBAAsB,CAAC,OAC3B,IAAI,QAAc,CAAC,SAAS,WAAW;AACrC,KAAG,IAAI,YAAY,CAAC,QAAsB;AACxC,QAAI,KAAK;AACP,aAAO,GAAG;AACV;AAAA,IACF;AAEA,YAAQ;AAAA,EACV,CAAC;AACH,CAAC;;;AE1II,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAG9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ,YAAY,QAAQ;AAE9B,IAAM,WAAW,CAAC,QACvB,OAAO,QAAQ;ACIV,IAAM,cAAN,MAAM,qBAAoB,MAAM;EAC9B;EAEP,YACE,SACA;AACA,UAAM,YACJ,WAAW,OAAO,YAAY,YAAY,eAAe,UACrD,QAAQ,YACR,SAAS,OAAO,IACd,UACA;AACR,UAAM,UACJ,WAAW,OAAO,YAAY,YAAY,aAAa,UACnD,QAAQ,UACR,SAAS,OAAO,IACd,UACA,2BAA2B,SAAS;AAE5C,UAAM,OAAO;AACb,SAAK,YAAY;AAGjB,WAAO,eAAe,MAAM,aAAY,SAAS;EACnD;EAEA,OAAc,QACZ,OACa;AACb,QAAI,iBAAiB,cAAa;AAChC,aAAO;IACT;AAEA,WAAO,IAAI,aAAY;MACrB,WACE,eAAe,SACf,MAAM,cAAc,UACpB,MAAM,cAAc,OAChB,MAAM,YACN;MACN,SAAS,MAAM,WAAW;IAC5B,CAAC;EACH;AACF;AAEO,IAAM,mBAAN,MAAM,0BAAyB,YAAY;EAChD,YACS,SACA,UACP,SACA;AACA,UAAM;MACJ,WAAW;MACX,SACE,WACA,oBAAoB,SAAS,SAAS,CAAC,2BAA2B,SAAS,SAAS,CAAC;IACzF,CAAC;AATM,SAAA,UAAA;AACA,SAAA,WAAA;AAWP,WAAO,eAAe,MAAM,kBAAiB,SAAS;EACxD;AACF;;;AWjFA,SAAS,MAAMC,aAAY;ACA3B,SAAS,MAAM,YAAY;ASA3B,OAAO,WAAW;AIAlB,SAAS,MAAMA,aAAY;AIA3B,SAAS,uBAAuB;ACAhC,SAAS,MAAMA,aAAY;ACA3B,SAAS,MAAMA,aAAY;AAC3B,SAAS,mBAAAC,wBAAuB;AeDhC,SAAS,sBAAsB;AEA/B,OAIO;ACJP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC;EAGE,mBAAAA;OAEK;ACLP,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,wBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;ACAhC,SAAS,mBAAAA,yBAAuB;A/CezB,IAAM,gBAAgB;AACtB,IAAM,wBACX;AACK,IAAM,uBACX;AAEK,IAAM,yBAAyB,CACpC,SACA,UACA,mBACY;AACZ,MAAI,aAAa,qBAAsB,QAAO;AAE9C,MAAI,YAAY,sBAAuB,QAAO,YAAY;AAE1D,MAAI,YAAY,cAAe,QAAO,YAAY;AAElD,SAAO,YAAY;AACrB;AAEO,IAAM,sCAAsC,CAGjD,SACA,UACA,mBACS;AACT,eAAa;AAEb,MAAI,CAAC,uBAAuB,SAAS,UAAU,cAAc;AAC3D,UAAM,IAAI,6BAA6B,SAAS,QAAQ;AAC5D;AAEO,IAAM,+BAAN,MAAM,sCAEH,iBAAiB;EACzB,YACE,SACA,UACA;AACA,UAAM,SAAS,SAAS,GAAG,UAAU,SAAS,CAAC;AAG/C,WAAO,eAAe,MAAM,8BAA6B,SAAS;EACpE;AACF;AY5DO,IAAM,aAAN,cAAyB,MAAM;EACpC,YAAY,MAAc;AACxB,UAAM,iBAAiB,IAAI,EAAE;EAC/B;AACF;AA0BO,IAAM,aAAa;EACxB,WAAW,CACT,OACA,YACG;AACH,WAAO,KAAK;MACV,SAAS,MAAM,QAAQ,IAAI,KAA6B,IAAI;;;MAG5D,CAAC,GAAG,MAAO,OAAO,MAAM,WAAW,EAAE,SAAS,IAAI;IACpD;EACF;EACA,OAAO,CACL,MACA,YACmB;AACnB,UAAM,SAAkB,KAAK,MAAM,MAAM,SAAS,OAAO;AAEzD,QAAI,SAAS,aAAa,CAAC,SAAS,UAAc,MAAM;AACtD,YAAM,IAAI,WAAW,IAAI;AAE3B,WAAO,SAAS,MACZ,QAAQ,IAAI,MAA8B,IACzC;EACP;AACF;AD5CO,IAAM,aAAa,OACxB,IACA,SACe;AACf,MAAI,SAAS,UAAa,KAAK,YAAY,EAAG,QAAO,GAAG;AAExD,SAAO;IACL,OAAO,SAAS;AACd,UAAI;AACF,cAAM,SAAS,MAAM,GAAG;AAExB,YAAI,MAAM,qBAAqB,KAAK,kBAAkB,MAAM,GAAG;AAC7D,gBAAM,IAAI;YACR,+BAA+B,WAAW,UAAU,MAAM,CAAC;UAC7D;QACF;AACA,eAAO;MACT,SAASC,QAAO;AACd,YAAI,MAAM,oBAAoB,CAAC,KAAK,iBAAiBA,MAAK,GAAG;AAC3D,eAAKA,MAAc;AACnB,iBAAO;QACT;AACA,cAAMA;MACR;IACF;IACA,QAAQ,EAAE,SAAS,EAAE;EACvB;AACF;AUnCO,IAAM,mCAAmC,CAC9C,0BAGA,UAA6D,CAAC,MAC3D,IAAI,iCAAiC,0BAA0B,OAAO;AAEpE,IAAM,mCAAN,cAAqDC,iBAG1D;EAQA,YACU,0BAGR,UAA6D,CAAC,GAC9D;AACA,UAAM;MACJ,QAAQ,CAAC,WAAW;AAClB,gBAAQ,IAAI,gCAAgC,MAAM;AAClD,aAAK,aAAa;MACpB;IACF,CAAC;AAVO,SAAA,2BAAA;AAWR,SAAK,WAAW,SAAS,YAAYC,MAAK;AAE1C,SAAK,2BAA2B;AAEhC,SAAK,cAAc,SAAS,qBAAqB,EAAE;EACrD;EAxBQ,gBAAuC;EAC/B;EACR,aAAsB;EAC9B,IAAW,uBAAuB;AAChC,WAAO,CAAC,KAAK;EACf;EAqBQ,cAAc,UAAkB;AACtC,SAAK,gBAAgB,YAAY,MAAM;AACrC,WAAK,oBAAoB;IAC3B,GAAG,QAAQ;EACb;EAEQ,eAAe;AACrB,QAAI,CAAC,KAAK,cAAe;AAEzB,kBAAc,KAAK,aAAa;AAChC,SAAK,gBAAgB;AACrB,SAAK,aAAa;AAClB,SAAK,yBAAyB,IAAI;EACpC;EAEQ,sBAAsB;AAC5B,QAAI,CAAC,KAAK,SAAS,UAAU,CAAC,KAAK,YAAY;AAC7C,WAAK,aAAa;IACpB;EACF;AACF;AK1BO,IAAM,gBAAgB,CAK3BC,aAC0B;AAE1B,SAAO,gBAAgBA,SAAQ;EAE7B,SAASA,SAAQ,SAAS,UAAU;;IAElCA,SAAQ,SAAS;MACjB,oBAAoBA,SAAQ;EAE1B,SAASA,SAAQ,SAAS,cAAc;;IAExCA,SAAQ,SAAS;MACjB,oBAAoBA,SAAQ;EAE1B,SAASA,SAAQ,SAAS,cAAc;;IAExCA,SAAQ,SAAS;MACjB;AACV;AaxDO,IAAM,SAAS,CAAOC,YAC3B,IAAIC,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,QAAID,QAAO,KAAK,GAAG;AACjB,iBAAW,QAAQ,KAAK;IAC1B;EACF;AACF,CAAC;ACPI,IAAM,MAAM,CAAWE,SAC5B,IAAID,iBAA0B;EAC5B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQC,KAAI,KAAK,CAAC;EAC/B;AACF,CAAC;ACLI,IAAM,SAAS,CACpB,SACA,iBACG,IAAI,sBAA4B,SAAS,YAAY;AAEnD,IAAM,wBAAN,cAA0CD,iBAAsB;EAC7D;EACA;EAER,YAAY,SAA0C,cAAiB;AACrE,UAAM;MACJ,WAAW,CAAC,UAAU;AACpB,aAAK,cAAc,KAAK,QAAQ,KAAK,aAAa,KAAK;MACzD;MACA,OAAO,CAAC,eAAe;AACrB,mBAAW,QAAQ,KAAK,WAAW;AACnC,mBAAW,UAAU;MACvB;IACF,CAAC;AAED,SAAK,cAAc;AACnB,SAAK,UAAU;EACjB;AACF;ACjBO,IAAM,cAAc,CAKzB,oBACAE,cAIA,eAAkC,EAAE,SAAS,MAAM,YAAY,GAAG,MAElE,IAAIF,iBAAqC;EACvC,MAAM,YAAY;AAChB;MACE,MAAM,WAAW,oBAAoBE,cAAa,UAAU;MAC5D;IACF,EAAE,MAAM,CAACC,WAAU;AACjB,iBAAW,MAAMA,MAAK;IACxB,CAAC;EACH;AACF,CAAC;AAEH,IAAM,aAAa,OACjB,oBACAD,cAIA,eACkB;AAClB,QAAM,eAAe,mBAAmB;AACxC,QAAM,SAAS,aAAa,UAAU;AAEtC,MAAI;AACF,QAAI;AAEJ,OAAG;AACD,YAAM,SAAS,MAAM,OAAO,KAAK;AACjC,aAAO,OAAO;AAEd,YAAMA,aAAY,QAAQ,UAAU;AAEpC,UAAI,MAAM;AACR,mBAAW,UAAU;MACvB;IACF,SAAS,CAAC;EACZ,UAAA;AACE,WAAO,YAAY;EACrB;AACF;ACxDO,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCF,iBAAsB;EACxD,QAAQ;EACR;EAER,YAAYI,OAAc;AACxB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,aAAK;AACL,YAAI,KAAK,QAAQ,KAAK,MAAM;AAC1B,qBAAW,QAAQ,KAAK;QAC1B;MACF;IACF,CAAC;AAED,SAAK,OAAOA;EACd;AACF;AClBO,IAAM,YAAY,CAAO,kBAC9B,IAAIJ,iBAA4B;EAC9B,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;AAExB,QAAI,cAAc,KAAK,GAAG;AACxB,iBAAW,UAAU;IACvB;EACF;AACF,CAAC;ACTI,IAAM,SAAS,CAAO,kBAC3B,IAAIA,iBAA4B;EAC9B,MAAM,UAAU,OAAO,YAAY;AACjC,QAAI,CAAC,cAAc,KAAK,GAAG;AACzB,iBAAW,QAAQ,KAAK;AACxB;IACF;AACA,UAAM,QAAQ,QAAQ;AACtB,eAAW,UAAU;EACvB;AACF,CAAC;ACVI,IAAM,OAAO,CAAI,UAAkB,IAAI,oBAAuB,KAAK;AAEnE,IAAM,sBAAN,cAAqCA,kBAAsB;EACxD,QAAQ;EACR;EAER,YAAY,OAAe;AACzB,UAAM;MACJ,WAAW,CAAC,OAAO,eAAe;AAChC,YAAI,KAAK,QAAQ,KAAK,OAAO;AAC3B,eAAK;AACL,qBAAW,QAAQ,KAAK;QAC1B,OAAO;AACL,qBAAW,UAAU;QACvB;MACF;IACF,CAAC;AAED,SAAK,QAAQ;EACf;AACF;ACpBO,IAAM,aAAa,CAAO,iBAC/B,IAAIA,kBAA4B;EAC9B,MAAM,YAAY;AAChB,UAAM,YAAY,WAAW,MAAM;AACjC,iBAAW,UAAU;IACvB,GAAG,YAAY;AAEf,UAAM,oBAAoB,WAAW,UAAU,KAAK,UAAU;AAG9D,eAAW,YAAY,MAAM;AAC3B,mBAAa,SAAS;AACtB,wBAAkB;IACpB;EACF;EACA,UAAU,OAAO,YAAY;AAC3B,eAAW,QAAQ,KAAK;EAC1B;AACF,CAAC;ACNI,IAAM,wBAAwB;EACnC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACA,OAAO;EACP;EACA;EACA;AACF;AVnBA,IAAM,EAAE,OAAAK,OAAM,IAAI;;;AWClB,SAAS,MAAMC,aAAY;;;ACXpB,IAAM,eAAe;AAErB,IAAM,YAAY;AAClB,IAAM,aAAa;AAEnB,IAAM,cAAc;AAAA,EACzB,QAAQ,GAAG,YAAY,WAAW,SAAS;AAC7C;AAEA,IAAM,UAAU;AAAA,EACd,WAAW;AAAA,IACT,MAAM;AAAA,EACR;AAAA,EACA,YAAY,EAAE,MAAM,cAAc;AACpC;AAEO,IAAM,eAAe;AAAA,EAC1B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,gBAAgB;AAAA,EAC3B,MAAM,GAAG,YAAY;AAAA,EACrB,SAAS;AAAA,IACP,WAAW,QAAQ;AAAA,IACnB,YAAY,QAAQ;AAAA,EACtB;AACF;AAEO,IAAM,qBAAqB;AAAA,EAChC,MAAM,GAAG,YAAY;AACvB;;;ADFO,IAAM,iBAAiB,OAC5B,YACA,YACA,YACA,UACA,YAO+B;AAC/B,MAAI,SAAS,WAAW,EAAG,QAAO,EAAE,SAAS,MAAM;AAEnD,QAAM,wBAAwB;AAAA,IAC5B,SAAS;AAAA,EACX;AAEA,QAAM,mBAGA,SAAS;AAAA,IACb,CACE,GACA,OAEC;AAAA,MACC,GAAG;AAAA,MACH,MAAM,EAAE,QAAQ;AAAA,MAChB,UAAU;AAAA,QACR;AAAA,QACA,WAAWC,MAAK;AAAA,QAChB,gBAAgB,OAAO,IAAI,CAAC;AAAA,QAC5B,GAAI,cAAc,IAAK,EAAE,YAAY,CAAC,IAAK,CAAC;AAAA,MAC9C;AAAA,IACF;AAAA,EACJ;AAEA,MAAI;AAEJ,SAAO,MAAM,WAAW,gBAAgB,YAAY;AAClD,aAAS,MAAM;AAAA,MACb;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,QACE;AAAA,MACF;AAAA,IACF;AAEA,QAAI,SAAS;AACX,YAAM,QAAQ,eAAe,kBAAkB,EAAE,WAAW,CAAC;AAE/D,WAAO;AAAA,EACT,CAAC;AACH;AAEA,IAAM,oBAAoB,CACxB,aACkB;AAClB,MAAI,aAAa,OAAW,QAAO;AAEnC,MAAI,aAAa,qBAAsB,QAAO;AAG9C,MAAI,YAAY,sBAAuB,QAAO;AAG9C,MAAI,YAAY,cAAe,QAAO;AAEtC,SAAO;AACT;AAEA,IAAM,oBAAoB,OACxB,YACA,UACA,YACA,UACA,YAI+B;AAC/B,MAAI;AACJ,MAAI;AAEJ,MAAI;AACF,QAAI,wBAAwB,SAAS,yBAAyB;AAE9D,QAAI,yBAAyB,MAAM;AACjC,8BAAwB,MAAM;AAAA,QAC5B;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAAA,IACF;AAEA,QAAI;AAEJ,QAAI,0BAA0B,IAAI;AAChC,iBAAW,MAAM,WAAW;AAAA,QAG1B,eAAe,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAYhC;AAAA,UACE;AAAA,UACA,SAAS;AAAA,UACT,SAAS,aAAa,aAAa,QAAQ;AAAA,UAC3C;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,iBAAW,MAAM,WAAW;AAAA,QAG1B,UAAU,aAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAO3B;AAAA,UACE,SAAS;AAAA,UACT;AAAA,UACA,SAAS,aAAa,aAAa,QAAQ;AAAA,QAC7C;AAAA,MACF;AAAA,IACF;AAEA,QAAI,YAAY,MAAM;AACpB,YAAM,IAAI,MAAM,gCAAgC;AAAA,IAClD;AAEA,qBAAiB,OAAO,SAAS,eAAe;AAEhD,QAAI,yBAAyB,MAAM;AACjC,YAAM,kCACJ,OAAO,qBAAqB,IAAI,OAAO,SAAS,MAAM;AACxD,UAAI,mBAAmB,iCAAiC;AACtD,eAAO;AAAA,UACL,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAEA,UAAM,EAAE,WAAW,OAAO,IAAI;AAAA,MAC5B;AAAA,MACA;AAAA,MACA;AAAA,MACA,SAAS,WAAW,SAAS,KAAK;AAAA,IACpC;AAEA,UAAM,eAAe,MAAM,WAAW,MAE5B,WAAW,MAAM;AAE3B,QACE,aAAa,WAAW,KACxB,CAAC,aAAa,aAAa,SAAS,CAAC,GAAG,iBACxC;AACA,YAAM,IAAI,MAAM,gCAAgC;AAAA,IAClD;AAEA,qBAAiB;AAAA,MACf,aAAa,aAAa,SAAS,CAAC,EAAG;AAAA,IACzC;AAAA,EACF,SAAS,KAAc;AACrB,QAAI,cAAc,GAAG,KAAK,6BAA6B,GAAG,GAAG;AAC3D,aAAO;AAAA,QACL,SAAS;AAAA,MACX;AAAA,IACF;AAEA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL,SAAS;AAAA,IACT,oBAAoB;AAAA,IACpB,oBAAoB;AAAA,EACtB;AACF;AAEA,IAAM,+BAA+B,CAAC,UAAgC;AACpE,SAAO,OAAO,UAAU,UAAa,MAAM,UAAU;AACvD;AAEA,eAAe,sBACb,YACA,UACA,uBACiB;AACjB,QAAM,SAAS,MAAM,WAAW;AAAA,IAG9B,mEAAmE,aAAa,IAAI;AAAA,IACpF,CAAC,QAAQ;AAAA,EACX;AAEA,MAAI,QAAQ,mBAAmB,MAAM;AACnC,4BAAwB;AAAA,EAC1B,OAAO;AACL,4BAAwB,OAAO,OAAO,eAAe;AAAA,EACvD;AACA,SAAO;AACT;AAEA,IAAM,0BAA0B,CAC9B,UACA,uBACA,UACA,cAIG;AACH,QAAM,QAAQ,SAAS;AAAA,IACrB,CACE,cACA,YACG;AACH,UACE,QAAQ,UAAU,kBAAkB,QACpC,OAAO,QAAQ,SAAS,mBAAmB,UAC3C;AACA,cAAM,IAAI,MAAM,6BAA6B;AAAA,MAC/C;AAEA,YAAM,iBACJ,OAAO,QAAQ,SAAS,cAAc,IAAI,OAAO,qBAAqB;AAExE,mBAAa,iBAAiB,KAAK,uBAAuB;AAC1D,mBAAa,OAAO;AAAA,QAClB;AAAA,QACA,eAAe,SAAS,KAAK;AAAA,QAC7B,aAAa;AAAA,QACb,QAAQ,SAAS,UAAU,MAAM;AAAA,QACjC,WAAW,UAAU,QAAQ,IAAI;AAAA,QACjC,WAAW,UAAU,QAAQ,QAAQ;AAAA,QACrC,uBAAuB,SAAS,KAAK;AAAA,QACrC,QAAQ;AAAA,QACR,QAAQ,SAAS;AAAA,QACjB;AAAA,MACF;AAEA,aAAO;AAAA,IACT;AAAA,IACA;AAAA,MACE,kBAAkB,CAAC;AAAA,MACnB,QAAQ,CAAC;AAAA,IACX;AAAA,EACF;AAEA,QAAM,YAAY;AAAA,oBACA,cAAc,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,eAYvB,MAAM,iBAAiB,KAAK,IAAI,CAAC;AAAA;AAAA;AAAA;AAI9C,SAAO,EAAE,WAAW,QAAQ,MAAM,OAAO;AAC3C;;;AErTO,IAAM,MAAM,CAACC,SAAgBA;AAE7B,IAAM,kBAAkB;AAAA,EAC7B,8BAA8B,aAAa,IAAI;AAAA;AAAA;AAAA,sEAGqB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAO/E;AAEO,IAAM,mBAAmB;AAAA,EAC9B,8BAA8B,cAAc,IAAI;AAAA;AAAA;AAAA,2EAGyB,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAapF;AAEO,IAAM,wBAAwB;AAAA,EACnC;AAAA,+BAC6B,mBAAmB,IAAI;AAAA;AAAA;AAAA,iFAG2B,SAAS;AAAA;AAAA;AAAA;AAAA;AAK1F;AAEO,IAAM,YAAsB;AAAA,EACjC;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,yBAAyB,OACpC,OACkB;AAClB,aAAWA,QAAO,WAAW;AAC3B,UAAM,GAAG,QAAQA,IAAG;AAAA,EACtB;AACF;;;AClEO,IAAM,eAAe,OAC1B,cACsB;AACtB,QAAM,SAAS,MAAM;AAErB,MAAI,OAAO,SAAS,EAAG,OAAM,IAAI,MAAM,gCAAgC;AAEvE,SAAO,OAAO,SAAS,IAAK,OAAO,CAAC,KAAK,OAAQ;AACnD;;;ACKO,IAAM,gCAAgC,OAC3C,IACA,YACiD;AACjD,QAAM,SAAS,MAAM;AAAA,IACnB,GAAG;AAAA,MACD;AAAA,QACE;AAAA,gBACQ,cAAc,IAAI;AAAA;AAAA;AAAA;AAAA,MAI5B;AAAA,MACA,CAAC,SAAS,aAAa,UAAU;AAAA,IACnC;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,eAAe,IAAI;AAAA,EACvD;AACF;;;ACYO,IAAM,oBAAoB,OAK/B,IACA,YACyE;AACzE,QAAM,OACJ,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QAAQ,KAChB;AACR,QAAM,YACJ,WAAW,eAAe,UACtB,QAAQ,YACR,QAAQ,KAAK,QAAQ;AAE3B,QAAM,gBACJ,SAAS,CAAC,KAAK,0BAA0B,IAAI,KAAK;AAEpD,QAAM,cACJ,QAAQ,UAAU,0BAA0B,QAAQ,EAAE,KAAK;AAE7D,QAAM,iBACJ,eAAe,UAAU,SAAS,QAAQ,SAAS,KAAK;AAE1D,QAAM,UACJ,MAAM,GAAG;AAAA,IACP;AAAA,MACE;AAAA,kBACU,cAAc,IAAI;AAAA,yDACqB,aAAa,IAAI,WAAW;AAAA;AAAA,aAExE,cAAc;AAAA,IACrB;AAAA,IACA,CAAC,SAAS,aAAa,UAAU;AAAA,EACnC,GACA,IAAI,CAAC,QAAQ;AACb,UAAM,WAAW;AAAA,MACf,MAAM,IAAI;AAAA,MACV,MAAM,WAAW,MAAM,IAAI,YAAY;AAAA,MACvC,UAAU,WAAW,MAAM,IAAI,gBAAgB;AAAA,IACjD;AAEA,UAAM,WAAgD;AAAA,MACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,MAC1D,WAAW,IAAI;AAAA,MACf,YAAY,IAAI;AAAA,MAChB,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,IAC5C;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH,MAAM;AAAA,MACN;AAAA,IAIF;AAAA,EACF,CAAC;AAED,SAAO,OAAO,SAAS,IACnB;AAAA,IACE,uBACE,OAAO,OAAO,SAAS,CAAC,EAAG,SAAS;AAAA,IACtC,UAAU;AAAA,IACV,eAAe,OAAO,WAAW;AAAA,EACnC,IACA;AAAA,IACE,uBACE,UAAU,UACN,QAAQ,OACR,WAAW,UACT,QAAQ,QACR;AAAA,IACR,UAAU,CAAC;AAAA,IACX,eAAe;AAAA,EACjB;AACN;;;AClHO,IAAM,0BAA0B,OACrC,IACA,YAC2C;AAC3C,QAAM,SAAS,MAAM;AAAA,IACnB,GAAG;AAAA,MACD;AAAA,QACE;AAAA,kBACU,mBAAmB,IAAI;AAAA;AAAA;AAAA,MAGnC;AAAA,MACA,CAAC,SAAS,aAAa,YAAY,QAAQ,WAAW;AAAA,IACxD;AAAA,EACF;AAEA,SAAO;AAAA,IACL,uBACE,WAAW,OAAO,OAAO,OAAO,uBAAuB,IAAI;AAAA,EAC/D;AACF;;;ACpBO,IAAM,4CAA4C;AAClD,IAAM,uDAAuD;AA4C7D,IAAM,qCAAqC,CAEhD;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF,MAAgG;AAC9F,MAAI,YAAY;AAEhB,MAAI;AAEJ,QAAM,eAAe,OACnB,YACG;AACH,UAAM,QACJ,QAAQ,cAAc,cAClB,KACA,QAAQ,cAAc,SAClB,MAAM,8BAA8B,UAAU,GAC7C,yBAAyB,KAC5B,QAAQ,UAAU;AAE1B,UAAM,sBAAgD;AAAA,MACpD;AAAA,MACA;AAAA,IACF;AAEA,QAAI,WAAW;AAEf,OAAG;AACD,YAAM,EAAE,UAAU,uBAAuB,cAAc,IACrD,MAAM,kBAA6B,YAAY,mBAAmB;AAEpE,UAAI,SAAS,SAAS,GAAG;AACvB,cAAM,SAAS,MAAM,UAAU,EAAE,SAAS,CAAC;AAE3C,YAAI,UAAU,OAAO,SAAS,QAAQ;AACpC,sBAAY;AACZ;AAAA,QACF;AAAA,MACF;AAEA,0BAAoB,QAAQ;AAE5B,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAE5D,UAAI,CAAC,eAAe;AAClB,mBAAW,KAAK,IAAI,WAAW,GAAG,GAAI;AAAA,MACxC,OAAO;AACL,mBAAW;AAAA,MACb;AAAA,IACF,SAAS;AAAA,EACX;AAEA,SAAO;AAAA,IACL,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,OAAO,CAAC,YAAY;AAClB,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,oBAAY;AAEZ,eAAO,aAAa,OAAO;AAAA,MAC7B,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA,MAAM,YAAY;AAChB,UAAI,CAAC,UAAW;AAChB,kBAAY;AACZ,YAAM;AAAA,IACR;AAAA,EACF;AACF;AAEO,IAAM,iDAAiD,CAC5D,YACgD;AAChD,MACE,QAAQ,WAAW,KACnB,QAAQ,KAAK,CAAC,MAAM,MAAM,UAAa,MAAM,WAAW;AAExD,WAAO;AAET,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,KAAK,EAAG,QAAO;AAE9C,SAAO,QACJ,OAAO,CAAC,MAAM,MAAM,UAAa,MAAM,eAAe,MAAM,KAAK,EACjE,KAAK,CAAC,GAAG,MAAO,IAAI,IAAI,IAAI,EAAG,EAAE,CAAC;AACvC;;;ACzCA,IAAM,yBAAyB,CAC7B,YACoB;AACpB,QAAM,EAAE,YAAY,IAAI;AACxB,MAAI,WAAW;AAGf,QAAM,QAAQ,CAAC,YAG2C;AACxD,UAAM,WAAW,QAAQ,YAAY,QAAQ,mBAAmB;AAChE,QAAI,CAAC;AACH,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ,WAAW;AAAA,MAC1C;AAEF,UAAM,aACJ,QAAQ,cACR,QAAQ,mBAAmB,cAC3B,iBAAiB,EAAE,SAAS,CAAC;AAE/B,WAAO,EAAE,YAAY,SAAS;AAAA,EAChC;AAEA,SAAO;AAAA,IACL,IAAI,QAAQ;AAAA,IACZ,OAAO,OACL,eACqE;AACrE,iBAAW;AACX,UAAI,QAAQ,cAAc,UAAW,QAAO,QAAQ;AAEpD,YAAM,EAAE,sBAAsB,IAAI,MAAM;AAAA,QACtC;AAAA,QACA;AAAA,UACE,aAAa,QAAQ;AAAA,UACrB,WAAW,QAAQ;AAAA,QACrB;AAAA,MACF;AAEA,UAAI,0BAA0B,KAAM,QAAO;AAE3C,aAAO,EAAE,gBAAgB,sBAAsB;AAAA,IACjD;AAAA,IACA,IAAI,WAAW;AACb,aAAO;AAAA,IACT;AAAA,IACA,QAAQ,OACN,EAAE,SAAS,GACX,YACiD;AACjD,UAAI,CAAC,SAAU;AAEf,YAAM,EAAE,YAAY,SAAS,IAAI,MAAM,OAAO;AAE9C,aAAO,WAAW,gBAAgB,YAAY;AAC5C,YAAI,SAA0D;AAE9D,YAAI,wBAAuC;AAE3C,mBAAW,WAAW,UAAU;AAC9B,gBAAM,eAAe;AAKrB,gBAAM,0BAA0B,MAAM,YAAY,cAAc;AAAA,YAC9D;AAAA,YACA;AAAA,UACF,CAAC;AAED,gBAAM,cAA6B,cAAc,YAAY;AAG7D,gBAAM,yBAAyB,YAAY;AAAA,YACzC,aAAa,QAAQ;AAAA,YACrB,SAAS,QAAQ;AAAA,YACjB;AAAA,YACA;AAAA,YACA,WAAW,QAAQ;AAAA,UACrB,CAAC;AAED,kCAAwB,aAAa,SAAS;AAE9C,cACE,2BACA,wBAAwB,SAAS,QACjC;AACA,uBAAW;AACX,qBAAS;AACT;AAAA,UACF;AAEA,cAAI,QAAQ,aAAa,QAAQ,UAAU,YAAY,GAAG;AACxD,uBAAW;AACX,qBAAS,EAAE,MAAM,QAAQ,QAAQ,yBAAyB;AAC1D;AAAA,UACF;AAEA,cACE,2BACA,wBAAwB,SAAS;AAEjC;AAAA,QACJ;AACA,eAAO;AAAA,MACT,CAAC;AAAA,IACH;AAAA,EACF;AACF;AAEO,IAAM,4BAA4B,CACvC,YACoB;AACpB,QAAMC,cAAa,QAAQ;AAE3B,SAAO,uBAAkC;AAAA,IACvC,aAAa,QAAQ,eAAe,cAAcA,YAAW,IAAI;AAAA,IACjE,aAAa,OAAO,OAAO,YAAY;AACrC,UAAI,CAACA,YAAW,UAAU,SAAS,MAAM,IAAI,EAAG;AAEhD,YAAMA,YAAW,OAAO,CAAC,KAAK,GAAG,EAAE,YAAY,QAAQ,WAAW,CAAC;AAAA,IACrE;AAAA,IACA,GAAG;AAAA,EACL,CAAC;AACH;AAEO,IAAM,kBAAkB,CAC7B,YACoB;AACpB,MAAI,gBAAgB,SAAS;AAC3B,WAAO,0BAA0B,OAAO;AAAA,EAC1C;AAEA,SAAO,uBAAuB,OAAO;AACvC;;;ACzMO,IAAM,2BAA2B,CAGtC,YACgD;AAChD,MAAI,YAAY;AAChB,QAAM,EAAE,QAAQ,IAAI;AACpB,QAAM,aAAa,QAAQ,cAAc,CAAC;AAE1C,MAAI;AAEJ,MAAI;AAEJ,QAAM,aACJ,QAAQ,cAAc,iBAAiB,EAAE,UAAU,QAAQ,SAAS,CAAC;AAEvE,QAAM,YAEF,OAAO,kBAAkB;AAC3B,UAAM,mBAAmB,WAAW,OAAO,CAAC,MAAM,EAAE,QAAQ;AAE5D,QAAI,iBAAiB,WAAW;AAC9B,aAAO;AAAA,QACL,MAAM;AAAA,QACN,QAAQ;AAAA,MACV;AAEF,UAAM,SAAS,MAAM,QAAQ;AAAA,MAC3B,iBAAiB,IAAI,CAAC,MAAM;AAE1B,eAAO,EAAE,OAAO,eAAe;AAAA,UAC7B;AAAA,UACA,UAAU,QAAQ;AAAA,QACpB,CAAC;AAAA,MACH,CAAC;AAAA,IACH;AAEA,WAAO,OAAO;AAAA,MACZ,CAAC,MAAM,EAAE,WAAW,eAAe,EAAE,OAAO,SAAS;AAAA,IACvD,IACI,SACA;AAAA,MACE,MAAM;AAAA,IACR;AAAA,EACN;AAEA,QAAM,gBAAiB,uBACrB,mCAAmC;AAAA,IACjC;AAAA,IACA;AAAA,IACA,WACE,SAAS,aAAa;AAAA,IACxB,sBACE,SAAS,wBACT;AAAA,EACJ,CAAC;AAEH,QAAM,OAAO,YAAY;AACvB,QAAI,CAAC,UAAW;AAChB,gBAAY;AACZ,QAAI,sBAAsB;AACxB,YAAM,qBAAqB,KAAK;AAChC,6BAAuB;AAAA,IACzB;AACA,UAAM;AAAA,EACR;AAEA,SAAO;AAAA,IACL;AAAA,IACA,IAAI,YAAY;AACd,aAAO;AAAA,IACT;AAAA,IACA,WAAW,CACTC,aAC+B;AAC/B,YAAM,YAAY,gBAA2BA,QAAO;AAEpD,iBAAW,KAAK,SAAS;AAEzB,aAAO;AAAA,IACT;AAAA,IACA,OAAO,MAAM;AACX,UAAI,UAAW,QAAO;AAEtB,eAAS,YAAY;AACnB,YAAI,WAAW,WAAW;AACxB,iBAAO,QAAQ;AAAA,YACb,IAAI;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAEF,oBAAY;AAEZ,cAAM,YAAY;AAAA,UAChB,MAAM,QAAQ,IAAI,WAAW,IAAI,CAAC,MAAM,EAAE,MAAM,UAAU,CAAC,CAAC;AAAA,QAC9D;AAEA,eAAO,cAAc,MAAM,EAAE,UAAU,CAAC;AAAA,MAC1C,GAAG;AAEH,aAAO;AAAA,IACT;AAAA,IACA;AAAA,IACA,OAAO,YAAY;AACjB,YAAM,KAAK;AAEX,iBAAW,MAAM;AAEjB,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,GAAG,CAAC;AAAA,IACzD;AAAA,EACF;AACF;;;ACvHO,IAAM,oBAAoB,OAC/B,YACkB;AAClB,QAAM,EAAE,aAAa,gBAAgB,QAAQ,WAAW,IAAI;AAE5D,QAAM,aAAa,OAAO,IAAI,CAAC,MAAM,EAAE,IAAI;AAE3C,aAAWC,eAAc,gBAAgB;AACvC,QAAI,CAACA,YAAW,UAAU,KAAK,CAAC,SAAS,WAAW,SAAS,IAAI,CAAC,GAAG;AACnE;AAAA,IACF;AACA,UAAMA,YAAW,OAAO,QAAQ;AAAA,MAC9B;AAAA,IACF,CAAC;AAAA,EACH;AACF;;;ACRO,IAAM,uCAAuC;AA+C7C,IAAM,sBAAsB,CACjC,YACqB;AACrB,MAAI,iBAAiB;AACrB,MAAI,qBAAqB;AACzB,MAAI;AACJ,QAAM,WAAW,QAAQ,YAAY;AAErC,QAAM,aACJ,aAAa,0BACb,aAAa;AAEf,QAAM,qBAAqB,QAAQ,eAAe,CAAC,GAChD,OAAO,CAAC,EAAE,KAAK,MAAM,SAAS,QAAQ,EACtC,IAAI,CAAC,EAAE,YAAAC,YAAW,MAAMA,WAAU;AAErC,QAAM,qBAAqB,QAAQ,OAAO;AAE1C,QAAM,mBAAmB,MAAM;AAC7B,QAAI,YAAY,MAAM;AACpB,aAAO;AAAA,IACT;AAEA,WAAO,iBAAiB;AAAA,MACtB;AAAA,IACF,CAAC;AAAA,EACH;AAEA,QAAM,kBAAkB,MAAM;AAC5B,QAAI,YAAY;AACd;AAAA,IACF;AACA,QAAI,YAAY,MAAM;AACpB,eAAS,MAAM;AACf,iBAAW;AAAA,IACb;AAAA,EACF;AAEA,QAAM,iBAAiB,OACrB,YACoB;AACpB,QAAI,YAAY,MAAM;AACpB,iBAAW,iBAAiB;AAAA,IAC9B;AAEA,QAAI;AACF,YAAM,mBAAmB,QAAQ;AACjC,aAAO,MAAM,QAAQ,QAAQ;AAAA,IAC/B,UAAE;AACA,sBAAgB;AAAA,IAClB;AAAA,EACF;AAEA,MAAI,SAAS;AACX,yBACE,QAAQ,QAAQ,kBAAkB,UAClC,QAAQ,QAAQ,kBAAkB;AAAA,EACtC;AAEA,QAAM,qBAAqB,OACzB,eACkB;AAClB,QAAI,CAAC,mBAAoB,QAAO,QAAQ,QAAQ;AAEhD,QAAI,CAAC,gBAAgB;AACnB,YAAM,uBAAuB,UAAU;AACvC,uBAAiB;AAAA,IACnB;AAEA,WAAO,QAAQ,QAAQ;AAAA,EACzB;AAEA,SAAO;AAAA,IACL,MAAM,gBACJ,YACAC,UAKuC;AACvC,YAAM,EAAE,QAAQ,cAAc,KAAK,IAAIA;AAEvC,YAAM,wBAAwB,MAAM;AAEpC,UAAI,QAAQ,aAAa;AAEzB,UAAI,OAAO,eAAe,UAAU;AAClC,cAAM,IAAI,MAAM,2BAA2B;AAAA,MAC7C;AAEA,UAAI,YAAY,MAAM;AACpB,mBAAW,iBAAiB;AAAA,MAC9B;AAEA,YAAM,SAAS,MAAM;AAAA,QAAe,CAAC,eACnC,WAAsB,YAAY,YAAYA,SAAQ,IAAI;AAAA,MAC5D;AAEA,YAAM,uBAAuB,OAAO;AAEpC;AAAA,QACE;AAAA,QACA;AAAA,QACA;AAAA,MACF;AAEA,iBAAW,SAAS,OAAO,QAAQ;AACjC,YAAI,CAAC,MAAO;AAEZ,gBAAQ,OAAO,OAAO,KAAK;AAAA,MAC7B;AAEA,aAAO;AAAA,QACL;AAAA,QACA;AAAA,QACA,cAAc,OAAO;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,YAAY,OACV,YACAA,aAIA;AAAA,MAAe,CAAC,eACd,WAAsB,YAAY,YAAYA,QAAO;AAAA,IACvD;AAAA,IAEF,gBAAgB,OACd,YACA,QACAA,aACoD;AACpD,UAAI,YAAY,MAAM;AACpB,mBAAW,iBAAiB;AAAA,MAC9B;AAGA,YAAM,CAAC,WAAW,GAAG,IAAI,IAAI,WAAW,MAAM,GAAG;AAEjD,YAAM,aACJ,aAAa,KAAK,SAAS,IAAI,YAAY;AAE7C,YAAM,eAAe,MAAM;AAAA,QAAe,CAAC,eACzC,eAAe,YAAY,YAAY,YAAY,QAAQ;AAAA,UACzD,GAAGA;AAAA,UACH,gBAAgB,OAAO,UAAU,YAAY;AAC3C,gBAAI,kBAAkB,SAAS;AAC7B,oBAAM,kBAAkB;AAAA,gBACtB,aAAa;AAAA,gBACb,QAAQ;AAAA,gBACR,GAAG;AAAA,cACL,CAAC;AAEH,gBAAI,mBAAoB,OAAM,mBAAmB,UAAU,OAAO;AAAA,UACpE;AAAA,QACF,CAAC;AAAA,MACH;AAEA,UAAI,CAAC,aAAa;AAChB,cAAM,IAAI;AAAA,UACR,CAAC;AAAA;AAAA,UACDA,UAAS,yBAAyB;AAAA,QACpC;AAEF,aAAO;AAAA,QACL,2BAA2B,aAAa;AAAA,QACxC,yBAAyB,aAAa;AAAA,QACtC,kBACE,aAAa,sBAAsB,OAAO,OAAO,MAAM;AAAA,MAC3D;AAAA,IACF;AAAA,IACA,UAAU,CACRA,aAEA,yBAA4C;AAAA,MAC1C,GAAIA,YAAW,CAAC;AAAA,MAChB;AAAA,MACA,YAAY,YAAY;AAAA,IAC1B,CAAC;AAAA,EACL;AACF;;;AC1PO,IAAM,aAAa,OACxB,IACA,UACA,YAGG;AACH,QAAM,gBACJ,WAAW,UAAU,UACjB,0BAA0B,QAAQ,IAAI,KACtC;AAEN,QAAM,KAAK;AAAA,IACT,WAAW,QAAQ,UACf,QAAQ,KACR,WAAW,cAAc,WAAW,QAAQ,WAC1C,QAAQ,OAAO,QAAQ,WACvB;AAAA,EACR;AAEA,QAAM,cAAc,CAAC,MAAM,EAAE,IAAI,0BAA0B,EAAE,KAAK;AAElE,QAAM,UAAU,MAAM,GAAG;AAAA,IACvB;AAAA,kBACc,cAAc,IAAI;AAAA,2EACuC,aAAa,IAAI,WAAW;AAAA;AAAA,IAEnG,CAAC,UAAU,SAAS,aAAa,UAAU;AAAA,EAC7C;AAEA,QAAM,WACJ,QAAQ,IAAI,CAAC,QAAQ;AACnB,UAAM,WAAW;AAAA,MACf,MAAM,IAAI;AAAA,MACV,MAAM,WAAW,MAAM,IAAI,YAAY;AAAA,MACvC,UAAU,WAAW,MAAM,IAAI,gBAAgB;AAAA,IACjD;AAEA,UAAM,WAAgD;AAAA,MACpD,GAAI,cAAc,WAAY,SAAS,YAAY,CAAC,IAAK,CAAC;AAAA,MAC1D,WAAW,IAAI;AAAA,MACf,YAAY;AAAA,MACZ,gBAAgB,OAAO,IAAI,eAAe;AAAA,MAC1C,gBAAgB,OAAO,IAAI,eAAe;AAAA,IAC5C;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH,MAAM;AAAA,MACN;AAAA,IAIF;AAAA,EACF,CAAC;AAEH,SAAO,SAAS,SAAS,IACrB;AAAA,IACE,sBACE,SAAS,SAAS,SAAS,CAAC,EAAG,SAAS;AAAA,IAC1C,QAAQ;AAAA,IACR,cAAc;AAAA,EAChB,IACA;AAAA,IACE,sBAAsB;AAAA,IACtB,QAAQ,CAAC;AAAA,IACT,cAAc;AAAA,EAChB;AACN;;;ACtFA,eAAe,kCACb,IACA,aACA,SACA,UACA,eACA,WACoB;AACpB,MAAI,kBAAkB,MAAM;AAC1B,UAAM,eAAe,MAAM,GAAG;AAAA,MAC5B,IAAI;AAAA,mBACS,mBAAmB,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,SAKjC;AAAA,MACH,CAAC,SAAU,SAAS,GAAG,aAAa,cAAc,SAAS,GAAG,SAAS;AAAA,IACzE;AACA,QAAI,aAAa,UAAU,GAAG;AAC5B,aAAO;AAAA,IACT,OAAO;AACL,YAAM,mBAAmB,MAAM;AAAA,QAC7B,GAAG;AAAA,UACD;AAAA,YACE,uCAAuC,mBAAmB,IAAI;AAAA;AAAA,UAEhE;AAAA,UACA,CAAC,aAAa,SAAS;AAAA,QACzB;AAAA,MACF;AAEA,UAAI,kBAAkB,4BAA4B,UAAU;AAC1D,eAAO;AAAA,MACT,WACE,aAAa,QACb,qBAAqB,QACrB,kBAAkB,0BAA0B,UAC5C;AACA,eAAO;AAAA,MACT,OAAO;AACL,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF,OAAO;AACL,QAAI;AACF,YAAM,GAAG;AAAA,QACP;AAAA,UACE,eAAe,mBAAmB,IAAI;AAAA,QACxC;AAAA,QACA,CAAC,aAAa,SAAS,SAAU,SAAS,GAAG,SAAS;AAAA,MACxD;AACA,aAAO;AAAA,IACT,SAAS,KAAK;AACZ,UAAI,EAAE,cAAc,GAAG,MAAM,IAAI,UAAU,MAAM,IAAI,UAAU,QAAQ;AACrE,cAAM;AAAA,MACR;AAEA,YAAM,UAAU,MAAM;AAAA,QACpB,GAAG;AAAA,UACD;AAAA,YACE,uCAAuC,mBAAmB,IAAI;AAAA,UAChE;AAAA,UACA,CAAC,aAAa,SAAS;AAAA,QACzB;AAAA,MACF;AACA,UAAI,SAAS,4BAA4B,UAAU;AACjD,eAAO;AAAA,MACT,OAAO;AACL,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EACF;AACF;AAWA,eAAsB,yBACpB,IACA,SAOmE;AACnE,MAAI;AACF,UAAM,SAAS,MAAM;AAAA,MACnB;AAAA,MACA,QAAQ;AAAA,MACR,QAAQ,WAAW;AAAA,MACnB,QAAQ;AAAA,MACR,QAAQ;AAAA,MACR,QAAQ,aAAa;AAAA,IACvB;AAEA,WAAO,WAAW,IACd,EAAE,SAAS,MAAM,aAAa,QAAQ,YAAY,IAClD,EAAE,SAAS,OAAO,QAAQ,WAAW,IAAI,YAAY,WAAW;AAAA,EACtE,SAAS,OAAO;AACd,YAAQ,IAAI,KAAK;AACjB,UAAM;AAAA,EACR;AACF;","names":["sql","uuid","TransformStream","error","TransformStream","uuid","message","filter","TransformStream","map","handleChunk","error","skip","retry","uuid","uuid","sql","projection","options","projection","projection","options"]}