@event-driven-io/emmett-mongodb 0.21.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs ADDED
@@ -0,0 +1,416 @@
1
+ "use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; } function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; } var _class; var _class2; var _class3;// ../emmett/dist/chunk-AEEEXE2R.js
2
+ var isNumber = (val) => typeof val === "number" && val === val;
3
+ var isString = (val) => typeof val === "string";
4
+ var EmmettError = class _EmmettError extends Error {
5
+
6
+ constructor(options) {
7
+ const errorCode = options && typeof options === "object" && "errorCode" in options ? options.errorCode : isNumber(options) ? options : 500;
8
+ const message = options && typeof options === "object" && "message" in options ? options.message : isString(options) ? options : `Error with status code '${errorCode}' ocurred during Emmett processing`;
9
+ super(message);
10
+ this.errorCode = errorCode;
11
+ Object.setPrototypeOf(this, _EmmettError.prototype);
12
+ }
13
+ };
14
+ var ConcurrencyError = class _ConcurrencyError extends EmmettError {
15
+ constructor(current, expected, message) {
16
+ super({
17
+ errorCode: 412,
18
+ message: _nullishCoalesce(message, () => ( `Expected version ${expected.toString()} does not match current ${_optionalChain([current, 'optionalAccess', _ => _.toString, 'call', _2 => _2()])}`))
19
+ });
20
+ this.current = current;
21
+ this.expected = expected;
22
+ Object.setPrototypeOf(this, _ConcurrencyError.prototype);
23
+ }
24
+ };
25
+
26
+ // ../emmett/dist/index.js
27
+ var _uuid = require('uuid');
28
+ var _webstreamspolyfill = require('web-streams-polyfill');
29
+
30
+
31
+
32
+ var _asyncretry = require('async-retry'); var _asyncretry2 = _interopRequireDefault(_asyncretry);
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+ var STREAM_EXISTS = "STREAM_EXISTS";
47
+ var STREAM_DOES_NOT_EXIST = "STREAM_DOES_NOT_EXIST";
48
+ var NO_CONCURRENCY_CHECK = "NO_CONCURRENCY_CHECK";
49
+ var matchesExpectedVersion = (current, expected, defaultVersion) => {
50
+ if (expected === NO_CONCURRENCY_CHECK) return true;
51
+ if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;
52
+ if (expected == STREAM_EXISTS) return current !== defaultVersion;
53
+ return current === expected;
54
+ };
55
+ var assertExpectedVersionMatchesCurrent = (current, expected, defaultVersion) => {
56
+ expected ??= NO_CONCURRENCY_CHECK;
57
+ if (!matchesExpectedVersion(current, expected, defaultVersion))
58
+ throw new ExpectedVersionConflictError(current, expected);
59
+ };
60
+ var ExpectedVersionConflictError = class _ExpectedVersionConflictError extends ConcurrencyError {
61
+ constructor(current, expected) {
62
+ super(_optionalChain([current, 'optionalAccess', _3 => _3.toString, 'call', _4 => _4()]), _optionalChain([expected, 'optionalAccess', _5 => _5.toString, 'call', _6 => _6()]));
63
+ Object.setPrototypeOf(this, _ExpectedVersionConflictError.prototype);
64
+ }
65
+ };
66
+ var notifyAboutNoActiveReadersStream = (onNoActiveReaderCallback, options = {}) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);
67
+ var NotifyAboutNoActiveReadersStream = (_class = class extends _webstreamspolyfill.TransformStream {
68
+ constructor(onNoActiveReaderCallback, options = {}) {
69
+ super({
70
+ cancel: (reason) => {
71
+ console.log("Stream was canceled. Reason:", reason);
72
+ this.stopChecking();
73
+ }
74
+ });_class.prototype.__init.call(this);_class.prototype.__init2.call(this);;
75
+ this.onNoActiveReaderCallback = onNoActiveReaderCallback;
76
+ this.streamId = _nullishCoalesce(_optionalChain([options, 'optionalAccess', _7 => _7.streamId]), () => ( _uuid.v4.call(void 0, )));
77
+ this.onNoActiveReaderCallback = onNoActiveReaderCallback;
78
+ this.startChecking(_nullishCoalesce(_optionalChain([options, 'optionalAccess', _8 => _8.intervalCheckInMs]), () => ( 20)));
79
+ }
80
+ __init() {this.checkInterval = null}
81
+
82
+ __init2() {this._isStopped = false}
83
+ get hasActiveSubscribers() {
84
+ return !this._isStopped;
85
+ }
86
+ startChecking(interval) {
87
+ this.checkInterval = setInterval(() => {
88
+ this.checkNoActiveReader();
89
+ }, interval);
90
+ }
91
+ stopChecking() {
92
+ if (!this.checkInterval) return;
93
+ clearInterval(this.checkInterval);
94
+ this.checkInterval = null;
95
+ this._isStopped = true;
96
+ this.onNoActiveReaderCallback(this);
97
+ }
98
+ checkNoActiveReader() {
99
+ if (!this.readable.locked && !this._isStopped) {
100
+ this.stopChecking();
101
+ }
102
+ }
103
+ }, _class);
104
+ var asyncRetry = async (fn, opts) => {
105
+ if (opts === void 0 || opts.retries === 0) return fn();
106
+ return _asyncretry2.default.call(void 0,
107
+ async (bail) => {
108
+ try {
109
+ return await fn();
110
+ } catch (error2) {
111
+ if (_optionalChain([opts, 'optionalAccess', _9 => _9.shouldRetryError]) && !opts.shouldRetryError(error2)) {
112
+ bail(error2);
113
+ }
114
+ throw error2;
115
+ }
116
+ },
117
+ _nullishCoalesce(opts, () => ( { retries: 0 }))
118
+ );
119
+ };
120
+ var filter = (filter2) => new (0, _webstreamspolyfill.TransformStream)({
121
+ transform(chunk, controller) {
122
+ if (filter2(chunk)) {
123
+ controller.enqueue(chunk);
124
+ }
125
+ }
126
+ });
127
+ var map = (map2) => new (0, _webstreamspolyfill.TransformStream)({
128
+ transform(chunk, controller) {
129
+ controller.enqueue(map2(chunk));
130
+ }
131
+ });
132
+ var reduce = (reducer, initialValue) => new ReduceTransformStream(reducer, initialValue);
133
+ var ReduceTransformStream = class extends _webstreamspolyfill.TransformStream {
134
+
135
+
136
+ constructor(reducer, initialValue) {
137
+ super({
138
+ transform: (chunk) => {
139
+ this.accumulator = this.reducer(this.accumulator, chunk);
140
+ },
141
+ flush: (controller) => {
142
+ controller.enqueue(this.accumulator);
143
+ controller.terminate();
144
+ }
145
+ });
146
+ this.accumulator = initialValue;
147
+ this.reducer = reducer;
148
+ }
149
+ };
150
+ var retryStream = (createSourceStream, handleChunk2, retryOptions = { forever: true, minTimeout: 25 }) => new (0, _webstreamspolyfill.TransformStream)({
151
+ start(controller) {
152
+ asyncRetry(
153
+ () => onRestream(createSourceStream, handleChunk2, controller),
154
+ retryOptions
155
+ ).catch((error2) => {
156
+ controller.error(error2);
157
+ });
158
+ }
159
+ });
160
+ var onRestream = async (createSourceStream, handleChunk2, controller) => {
161
+ const sourceStream = createSourceStream();
162
+ const reader = sourceStream.getReader();
163
+ try {
164
+ let done;
165
+ do {
166
+ const result = await reader.read();
167
+ done = result.done;
168
+ await handleChunk2(result, controller);
169
+ if (done) {
170
+ controller.terminate();
171
+ }
172
+ } while (!done);
173
+ } finally {
174
+ reader.releaseLock();
175
+ }
176
+ };
177
+ var skip = (limit) => new SkipTransformStream(limit);
178
+ var SkipTransformStream = (_class2 = class extends _webstreamspolyfill.TransformStream {
179
+ __init3() {this.count = 0}
180
+
181
+ constructor(skip2) {
182
+ super({
183
+ transform: (chunk, controller) => {
184
+ this.count++;
185
+ if (this.count > this.skip) {
186
+ controller.enqueue(chunk);
187
+ }
188
+ }
189
+ });_class2.prototype.__init3.call(this);;
190
+ this.skip = skip2;
191
+ }
192
+ }, _class2);
193
+ var stopAfter = (stopCondition) => new (0, _webstreamspolyfill.TransformStream)({
194
+ transform(chunk, controller) {
195
+ controller.enqueue(chunk);
196
+ if (stopCondition(chunk)) {
197
+ controller.terminate();
198
+ }
199
+ }
200
+ });
201
+ var stopOn = (stopCondition) => new (0, _webstreamspolyfill.TransformStream)({
202
+ async transform(chunk, controller) {
203
+ if (!stopCondition(chunk)) {
204
+ controller.enqueue(chunk);
205
+ return;
206
+ }
207
+ await Promise.resolve();
208
+ controller.terminate();
209
+ }
210
+ });
211
+ var take = (limit) => new TakeTransformStream(limit);
212
+ var TakeTransformStream = (_class3 = class extends _webstreamspolyfill.TransformStream {
213
+ __init4() {this.count = 0}
214
+
215
+ constructor(limit) {
216
+ super({
217
+ transform: (chunk, controller) => {
218
+ if (this.count < this.limit) {
219
+ this.count++;
220
+ controller.enqueue(chunk);
221
+ } else {
222
+ controller.terminate();
223
+ }
224
+ }
225
+ });_class3.prototype.__init4.call(this);;
226
+ this.limit = limit;
227
+ }
228
+ }, _class3);
229
+ var waitAtMost = (waitTimeInMs) => new (0, _webstreamspolyfill.TransformStream)({
230
+ start(controller) {
231
+ const timeoutId = setTimeout(() => {
232
+ controller.terminate();
233
+ }, waitTimeInMs);
234
+ const originalTerminate = controller.terminate.bind(controller);
235
+ controller.terminate = () => {
236
+ clearTimeout(timeoutId);
237
+ originalTerminate();
238
+ };
239
+ },
240
+ transform(chunk, controller) {
241
+ controller.enqueue(chunk);
242
+ }
243
+ });
244
+ var streamTransformations = {
245
+ filter,
246
+ take,
247
+ TakeTransformStream,
248
+ skip,
249
+ SkipTransformStream,
250
+ map,
251
+ notifyAboutNoActiveReadersStream,
252
+ NotifyAboutNoActiveReadersStream,
253
+ reduce,
254
+ ReduceTransformStream,
255
+ retry: retryStream,
256
+ stopAfter,
257
+ stopOn,
258
+ waitAtMost
259
+ };
260
+ var { retry: retry2 } = streamTransformations;
261
+
262
+ // src/eventStore/mongoDBEventStore.ts
263
+ require('mongodb');
264
+
265
+ var MongoDBEventStoreDefaultStreamVersion = 0;
266
+ var MongoDBEventStore = class {
267
+
268
+ constructor(collection) {
269
+ this.collection = collection;
270
+ }
271
+ async readStream(streamName, options) {
272
+ const expectedStreamVersion = _optionalChain([options, 'optionalAccess', _10 => _10.expectedStreamVersion]);
273
+ const stream = await this.collection.findOne({
274
+ streamName: { $eq: streamName }
275
+ });
276
+ if (!stream) {
277
+ return {
278
+ events: [],
279
+ currentStreamVersion: MongoDBEventStoreDefaultStreamVersion,
280
+ streamExists: false
281
+ };
282
+ }
283
+ assertExpectedVersionMatchesCurrent(
284
+ stream.events.length,
285
+ expectedStreamVersion,
286
+ MongoDBEventStoreDefaultStreamVersion
287
+ );
288
+ return {
289
+ events: stream.events.slice(0, maxEventIndex(expectedStreamVersion)),
290
+ currentStreamVersion: stream.events.length,
291
+ streamExists: true
292
+ };
293
+ }
294
+ async aggregateStream(streamName, options) {
295
+ const stream = await this.readStream(streamName, _optionalChain([options, 'optionalAccess', _11 => _11.read]));
296
+ const state = stream.events.reduce(options.evolve, options.initialState());
297
+ return {
298
+ state,
299
+ currentStreamVersion: stream.currentStreamVersion,
300
+ streamExists: stream.streamExists
301
+ };
302
+ }
303
+ async appendToStream(streamName, events, options) {
304
+ let stream = await this.collection.findOne({
305
+ streamName: { $eq: streamName }
306
+ });
307
+ let currentStreamPosition = _nullishCoalesce(_optionalChain([stream, 'optionalAccess', _12 => _12.events, 'access', _13 => _13.length]), () => ( 0));
308
+ let createdNewStream = false;
309
+ if (!stream) {
310
+ const now = /* @__PURE__ */ new Date();
311
+ const result = await this.collection.insertOne({
312
+ streamName,
313
+ events: [],
314
+ createdAt: now,
315
+ updatedAt: now
316
+ });
317
+ stream = await this.collection.findOne({
318
+ _id: result.insertedId
319
+ });
320
+ createdNewStream = true;
321
+ }
322
+ const eventCreateInputs = [];
323
+ for (const event of events) {
324
+ currentStreamPosition++;
325
+ eventCreateInputs.push({
326
+ type: event.type,
327
+ data: event.data,
328
+ metadata: {
329
+ now: /* @__PURE__ */ new Date(),
330
+ eventId: _uuid.v4.call(void 0, ),
331
+ streamName,
332
+ streamPosition: BigInt(currentStreamPosition),
333
+ ..._nullishCoalesce(event.metadata, () => ( {}))
334
+ }
335
+ });
336
+ }
337
+ if (!stream) throw new Error("Failed to create stream");
338
+ assertExpectedVersionMatchesCurrent(
339
+ stream.events.length,
340
+ _optionalChain([options, 'optionalAccess', _14 => _14.expectedStreamVersion]),
341
+ MongoDBEventStoreDefaultStreamVersion
342
+ );
343
+ const updatedStream = await this.collection.findOneAndUpdate(
344
+ {
345
+ streamName: { $eq: streamName },
346
+ events: { $size: stream.events.length }
347
+ },
348
+ {
349
+ $push: { events: { $each: eventCreateInputs } },
350
+ $set: { updatedAt: /* @__PURE__ */ new Date() }
351
+ },
352
+ { returnDocument: "after" }
353
+ );
354
+ if (!updatedStream) {
355
+ const currentStream = await this.collection.findOne({
356
+ streamName: { $eq: streamName }
357
+ });
358
+ throw new ExpectedVersionConflictError(
359
+ _nullishCoalesce(_optionalChain([currentStream, 'optionalAccess', _15 => _15.events, 'access', _16 => _16.length]), () => ( -1)),
360
+ stream.events.length
361
+ );
362
+ }
363
+ const { streamType, streamId } = fromStreamName(streamName);
364
+ await executeProjections(
365
+ {
366
+ streamName,
367
+ streamType,
368
+ streamId,
369
+ streamVersion: updatedStream.events.length,
370
+ events: updatedStream.events
371
+ },
372
+ _optionalChain([options, 'optionalAccess', _17 => _17.projections])
373
+ );
374
+ return {
375
+ nextExpectedStreamVersion: updatedStream.events.length,
376
+ createdNewStream
377
+ };
378
+ }
379
+ };
380
+ var getMongoDBEventStore = (collection) => {
381
+ const eventStore = new MongoDBEventStore(collection);
382
+ return eventStore;
383
+ };
384
+ function executeProjections(params, projections) {
385
+ return Promise.all((_nullishCoalesce(projections, () => ( []))).map((project) => project(params)));
386
+ }
387
+ function maxEventIndex(expectedStreamVersion) {
388
+ if (!expectedStreamVersion) return void 0;
389
+ if (typeof expectedStreamVersion === "number") {
390
+ return expectedStreamVersion;
391
+ }
392
+ switch (expectedStreamVersion) {
393
+ case STREAM_DOES_NOT_EXIST:
394
+ return 0;
395
+ default:
396
+ return void 0;
397
+ }
398
+ }
399
+ function toStreamName(streamType, streamId) {
400
+ return `${streamType}:${streamId}`;
401
+ }
402
+ function fromStreamName(streamName) {
403
+ const parts = streamName.split(":");
404
+ return {
405
+ streamType: parts[0],
406
+ streamId: parts[1]
407
+ };
408
+ }
409
+
410
+
411
+
412
+
413
+
414
+
415
+ exports.MongoDBEventStore = MongoDBEventStore; exports.MongoDBEventStoreDefaultStreamVersion = MongoDBEventStoreDefaultStreamVersion; exports.fromStreamName = fromStreamName; exports.getMongoDBEventStore = getMongoDBEventStore; exports.toStreamName = toStreamName;
416
+ //# sourceMappingURL=index.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/eventStore/subscriptions/caughtUpTransformStream.ts","../../emmett/src/eventStore/subscriptions/streamingCoordinator.ts","../../emmett/src/streaming/transformations/notifyAboutNoActiveReaders.ts","../../emmett/src/utils/retry.ts","../../emmett/src/streaming/generators/fromArray.ts","../../emmett/src/streaming/restream.ts","../../emmett/src/streaming/transformations/filter.ts","../../emmett/src/streaming/transformations/map.ts","../../emmett/src/streaming/transformations/reduce.ts","../../emmett/src/streaming/transformations/retry.ts","../../emmett/src/streaming/transformations/skip.ts","../../emmett/src/streaming/transformations/stopAfter.ts","../../emmett/src/streaming/transformations/stopOn.ts","../../emmett/src/streaming/transformations/take.ts","../../emmett/src/streaming/transformations/waitAtMost.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/streaming/transformations/index.ts","../src/eventStore/mongoDBEventStore.ts"],"names":[],"mappings":"AAAA;ACQO,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,SAAA,GAAY,IAAA,IAAQ,GAAA;AAE9B,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,QAAA;ACQV,IAAM,YAAA,EAAN,MAAM,aAAA,QAAoB,MAAM;AFhBvC,EEiBS;AFhBT,EEkBE,WAAA,CACE,OAAA,EACA;AACA,IAAA,MAAM,UAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,YAAA,GAAe,QAAA,EACrD,OAAA,CAAQ,UAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,GAAA;AACR,IAAA,MAAM,QAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,UAAA,GAAa,QAAA,EACnD,OAAA,CAAQ,QAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,CAAA,wBAAA,EAA2B,SAAS,CAAA,kCAAA,CAAA;AAE5C,IAAA,KAAA,CAAM,OAAO,CAAA;AACb,IAAA,IAAA,CAAK,UAAA,EAAY,SAAA;AAGjB,IAAA,MAAA,CAAO,cAAA,CAAe,IAAA,EAAM,YAAA,CAAY,SAAS,CAAA;AFhCrD,EEiCE;AACF,CAAA;AAEO,IAAM,iBAAA,EAAN,MAAM,kBAAA,QAAyB,YAAY;AFjClD,EEkCE,WAAA,CACS,OAAA,EACA,QAAA,EACP,OAAA,EACA;AACA,IAAA,KAAA,CAAM;AFrCV,MEsCM,SAAA,EAAW,GAAA;AFrCjB,MEsCM,OAAA,mBACE,OAAA,UACA,CAAA,iBAAA,EAAoB,QAAA,CAAS,QAAA,CAAS,CAAC,CAAA,wBAAA,kBAA2B,OAAA,2BAAS,QAAA,mBAAS,GAAC,CAAA;AFvC7F,IAAA;AE+BW,IAAA;AACA,IAAA;AAWP,IAAA;AFvCJ,EAAA;AEyCA;AFvCA;AACA;AGzBA;ACAA;ACAA;ACAA;AACA;ACDA;ACAA;ACAA;ACAA;ACAA;ACAA;ACAA;AbsCA;AACA;AcvCA;ACAA;ACAA;ACAA;ACAA;ACiBO;AACA;AAEA;AAGA;AAOL,EAAA;AAEA,EAAA;AAEA,EAAA;AAEA,EAAA;AACF;AAEO;AAOL,EAAA;AAEA,EAAA;AACE,IAAA;AACJ;AAEO;AnBQP,EAAA;AmBDI,IAAA;AAGA,IAAA;AnBCJ,EAAA;AmBCA;Ab7DO;AAOA;ANyDP,EAAA;AMxCI,IAAA;AN0CJ,MAAA;AMxCQ,QAAA;AACA,QAAA;AN0CR,MAAA;AACA,IAAA;AMnDY,IAAA;AAWR,IAAA;AAEA,IAAA;AAEA,IAAA;ANyCJ,EAAA;AACA,iBAAA;AACA,EAAA;AACA,kBAAA;AACA,EAAA;AMhEI,IAAA;ANkEJ,EAAA;AACA,EAAA;AM5CI,IAAA;AACE,MAAA;AN8CN,IAAA;AACA,EAAA;AACA,EAAA;AM3CI,IAAA;AAEA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AN4CJ,EAAA;AACA,EAAA;AMzCI,IAAA;AACE,MAAA;AN2CN,IAAA;AACA,EAAA;AMzCA;ACpDO;AAIL,EAAA;AAEA,EAAA;AP4FF,IAAA;AO1FM,MAAA;AACE,QAAA;AP4FR,MAAA;AO1FQ,QAAA;AACE,UAAA;AP4FV,QAAA;AO1FQ,QAAA;AP4FR,MAAA;AACA,IAAA;AACA,qBAAA;AACA,EAAA;AO1FA;AGzBO;AVsHP,EAAA;AUnHM,IAAA;AACE,MAAA;AVqHR,IAAA;AACA,EAAA;AUnHE;ACPK;AX6HP,EAAA;AW1HM,IAAA;AX4HN,EAAA;AW1HE;ACLK;AAKA;AZ8HP,EAAA;AACA,EAAA;AACA,EAAA;AY3HI,IAAA;AZ6HJ,MAAA;AY3HQ,QAAA;AZ6HR,MAAA;AACA,MAAA;AY3HQ,QAAA;AACA,QAAA;AZ6HR,MAAA;AACA,IAAA;AY1HI,IAAA;AACA,IAAA;AZ4HJ,EAAA;AY1HA;ACjBO;Ab8IP,EAAA;AahIM,IAAA;AbkIN,MAAA;AACA,MAAA;AACA,IAAA;AahIQ,MAAA;AbkIR,IAAA;AACA,EAAA;AahIE;AAEF;AAQE,EAAA;AACA,EAAA;AAEA,EAAA;AACE,IAAA;AAEA,IAAA;AACE,MAAA;AACA,MAAA;AAEA,MAAA;AAEA,MAAA;AACE,QAAA;AbsHR,MAAA;AACA,IAAA;AACA,EAAA;AapHI,IAAA;AbsHJ,EAAA;AapHA;ACxDO;AAEA;Ad8KP,kBAAA;AACA,EAAA;AACA,EAAA;Ac3KI,IAAA;Ad6KJ,MAAA;Ac3KQ,QAAA;AACA,QAAA;AACE,UAAA;Ad6KV,QAAA;AACA,MAAA;AACA,IAAA;Ac1KI,IAAA;Ad4KJ,EAAA;Ac1KA;AClBO;Af+LP,EAAA;Ae5LM,IAAA;AAEA,IAAA;AACE,MAAA;Af6LR,IAAA;AACA,EAAA;Ae3LE;ACTK;AhBuMP,EAAA;AgBpMM,IAAA;AACE,MAAA;AACA,MAAA;AhBsMR,IAAA;AgBpMM,IAAA;AACA,IAAA;AhBsMN,EAAA;AgBpME;ACVK;AAEA;AjBgNP,kBAAA;AACA,EAAA;AACA,EAAA;AiB7MI,IAAA;AjB+MJ,MAAA;AiB7MQ,QAAA;AACE,UAAA;AACA,UAAA;AjB+MV,QAAA;AiB7MU,UAAA;AjB+MV,QAAA;AACA,MAAA;AACA,IAAA;AiB5MI,IAAA;AjB8MJ,EAAA;AiB5MA;ACpBO;AlBmOP,EAAA;AkBhOM,IAAA;AACE,MAAA;AlBkOR,IAAA;AkB/NM,IAAA;AAGA,IAAA;AACE,MAAA;AACA,MAAA;AlB+NR,IAAA;AACA,EAAA;AACA,EAAA;AkB7NM,IAAA;AlB+NN,EAAA;AkB7NE;AENK;ApBsOP,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AoBpOA;AXnBA;AT0PA;AACA;AqBrPA;AACA;AAEO;AAqCA;AAAsD,EAAA;AAC1C,EAAA;AAGf,IAAA;AAAkB,EAAA;AACpB,EAAA;AAME,IAAA;AAEA,IAAA;AAEE,MAAA;AAC8B,IAAA;AAGhC,IAAA;AACE,MAAA;AAAO,QAAA;AACI,QAAA;AACa,QAAA;AACR,MAAA;AAChB,IAAA;AAGF,IAAA;AAAA,MAAA;AACgB,MAAA;AACd,MAAA;AACA,IAAA;AAGF,IAAA;AAAO,MAAA;AAC8D,MAAA;AAC/B,MAAA;AACtB,IAAA;AAChB,EAAA;AACF,EAAA;AAME,IAAA;AACA,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AAC6B,MAAA;AACR,IAAA;AACvB,EAAA;AACF,EAAA;AAaE,IAAA;AAA2C,MAAA;AACX,IAAA;AAEhC,IAAA;AACA,IAAA;AAEA,IAAA;AACE,MAAA;AACA,MAAA;AAA+C,QAAA;AAC7C,QAAA;AACS,QAAA;AACE,QAAA;AACA,MAAA;AAEb,MAAA;AAAuC,QAAA;AACzB,MAAA;AAEd,MAAA;AAAmB,IAAA;AAGrB,IAAA;AACA,IAAA;AACE,MAAA;AACA,MAAA;AAAuB,QAAA;AACT,QAAA;AACA,QAAA;AACF,UAAA;AACM,UAAA;AACA,UAAA;AACd,UAAA;AAC4C,UAAA;AACrB,QAAA;AACzB,MAAA;AACD,IAAA;AAKH,IAAA;AAEA,IAAA;AAAA,MAAA;AACgB,sBAAA;AACL,MAAA;AACT,IAAA;AAMF,IAAA;AACwB,MAAA;AACpB,QAAA;AACgC,QAAA;AACQ,MAAA;AACxC,MAAA;AACA,QAAA;AACgD,QAAA;AAChB,MAAA;AAChC,MAAA;AAC0B,IAAA;AAG9B,IAAA;AACE,MAAA;AAAoD,QAAA;AACpB,MAAA;AAEhC,MAAA;AAAU,yCAAA;AACwB,QAAA;AAClB,MAAA;AAChB,IAAA;AAGF,IAAA;AAEA,IAAA;AAAM,MAAA;AACJ,QAAA;AACE,QAAA;AACA,QAAA;AACA,QAAA;AACoC,QAAA;AACd,MAAA;AACxB,sBAAA;AACS,IAAA;AAGX,IAAA;AAAO,MAAA;AAC2C,MAAA;AAChD,IAAA;AACF,EAAA;AAEJ;AAEO;AACL,EAAA;AACA,EAAA;AACF;AAEA;AAIE,EAAA;AACF;AAEA;AAGE,EAAA;AAEA,EAAA;AACE,IAAA;AAAO,EAAA;AAGT,EAAA;AAA+B,IAAA;AAE3B,MAAA;AAAO,IAAA;AAEP,MAAA;AAAO,EAAA;AAEb;AAOO;AAIL,EAAA;AACF;AAMO;AAGL,EAAA;AACA,EAAA;AAAO,IAAA;AACc,IAAA;AACF,EAAA;AAErB;ArB+IA;AACA;AACA;AACA;AACA;AACA;AACA","file":"/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","sourcesContent":[null,"import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : 500;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: 412,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 400,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: 403,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: 404,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport {\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultStreamVersionType,\n type EventStore,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { StreamingCoordinator } from './subscriptions';\n\nexport type EventHandler<E extends Event = Event> = (\n eventEnvelope: ReadEvent<E>,\n) => void;\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport const getInMemoryEventStore = (): EventStore<\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n> => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n const streamingCoordinator = StreamingCoordinator();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType>(streamName, read);\n\n const events = result?.events ?? [];\n\n return {\n currentStreamVersion: BigInt(events.length),\n state: events.reduce(evolve, initialState()),\n streamExists: result.streamExists,\n };\n },\n\n readStream: <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions,\n ): Promise<\n ReadStreamResult<\n EventType,\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n >\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options && 'from' in options ? options.from : 0);\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : (events?.length ?? 1),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? events\n .map(\n (e) =>\n e as ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n )\n .slice(from, to)\n : [];\n\n const result: ReadStreamResult<\n EventType,\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition: BigInt(getAllEventsCount() + index + 1),\n },\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [...currentEvents, ...newEvents]);\n await streamingCoordinator.notify(newEvents);\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n return result;\n },\n\n //streamEvents: streamingCoordinator.stream,\n };\n};\n","import { TransformStream } from 'web-streams-polyfill';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport { globalStreamCaughtUp, type GlobalSubscriptionEvent } from '../events';\n\nexport const streamTrackingGlobalPosition = (\n currentEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n) => new CaughtUpTransformStream(currentEvents);\n\nexport class CaughtUpTransformStream extends TransformStream<\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>,\n | ReadEvent<Event, ReadEventMetadataWithGlobalPosition>\n | GlobalSubscriptionEvent\n> {\n private _currentPosition: bigint;\n private _logPosition: bigint;\n\n constructor(events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]) {\n super({\n start: (controller) => {\n let globalPosition = 0n;\n for (const event of events) {\n controller.enqueue(event);\n globalPosition = event.metadata.globalPosition;\n }\n controller.enqueue(globalStreamCaughtUp({ globalPosition }));\n },\n transform: (event, controller) => {\n this._currentPosition = event.metadata.globalPosition;\n controller.enqueue(event);\n\n if (this._currentPosition < this._logPosition) return;\n\n controller.enqueue(\n globalStreamCaughtUp({ globalPosition: this._currentPosition }),\n );\n },\n });\n\n this._currentPosition = this._logPosition =\n events.length > 0\n ? events[events.length - 1]!.metadata.globalPosition\n : 0n;\n }\n\n public set logPosition(value: bigint) {\n this._logPosition = value;\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport { notifyAboutNoActiveReadersStream } from '../../streaming/transformations/notifyAboutNoActiveReaders';\nimport { writeToStream } from '../../streaming/writers';\nimport type {\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../../typing';\nimport {\n CaughtUpTransformStream,\n streamTrackingGlobalPosition,\n} from './caughtUpTransformStream';\n\nexport const StreamingCoordinator = () => {\n const allEvents: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[] = [];\n const listeners = new Map<string, CaughtUpTransformStream>();\n\n return {\n notify: async (\n events: ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[],\n ) => {\n if (events.length === 0) return;\n\n allEvents.push(...events);\n\n for (const listener of listeners.values()) {\n listener.logPosition =\n events[events.length - 1]!.metadata.globalPosition;\n\n await writeToStream(listener, events);\n }\n },\n\n stream: () => {\n const streamId = uuid();\n const transformStream = streamTrackingGlobalPosition(allEvents);\n\n listeners.set(streamId, transformStream);\n return transformStream.readable.pipeThrough(\n notifyAboutNoActiveReadersStream(\n (stream) => {\n if (listeners.has(stream.streamId))\n listeners.delete(stream.streamId);\n },\n { streamId },\n ),\n );\n },\n };\n};\n","import { v4 as uuid } from 'uuid';\nimport { TransformStream } from 'web-streams-polyfill';\n\nexport const notifyAboutNoActiveReadersStream = <Item>(\n onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n) => new NotifyAboutNoActiveReadersStream(onNoActiveReaderCallback, options);\n\nexport class NotifyAboutNoActiveReadersStream<Item> extends TransformStream<\n Item,\n Item\n> {\n private checkInterval: NodeJS.Timeout | null = null;\n public readonly streamId: string;\n private _isStopped: boolean = false;\n public get hasActiveSubscribers() {\n return !this._isStopped;\n }\n\n constructor(\n private onNoActiveReaderCallback: (\n stream: NotifyAboutNoActiveReadersStream<Item>,\n ) => void,\n options: { streamId?: string; intervalCheckInMs?: number } = {},\n ) {\n super({\n cancel: (reason) => {\n console.log('Stream was canceled. Reason:', reason);\n this.stopChecking();\n },\n });\n this.streamId = options?.streamId ?? uuid();\n\n this.onNoActiveReaderCallback = onNoActiveReaderCallback;\n\n this.startChecking(options?.intervalCheckInMs ?? 20);\n }\n\n private startChecking(interval: number) {\n this.checkInterval = setInterval(() => {\n this.checkNoActiveReader();\n }, interval);\n }\n\n private stopChecking() {\n if (!this.checkInterval) return;\n\n clearInterval(this.checkInterval);\n this.checkInterval = null;\n this._isStopped = true;\n this.onNoActiveReaderCallback(this);\n }\n\n private checkNoActiveReader() {\n if (!this.readable.locked && !this._isStopped) {\n this.stopChecking();\n }\n }\n}\n","import retry from 'async-retry';\n\nexport type AsyncRetryOptions = retry.Options & {\n shouldRetryError?: (error: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n return await fn();\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","import { ReadableStream } from 'web-streams-polyfill';\n\nexport const fromArray = <T>(chunks: T[]) =>\n new ReadableStream<T>({\n start(controller) {\n for (const chunk of chunks) controller.enqueue(chunk);\n controller.close();\n },\n });\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport type { AsyncRetryOptions } from '../utils';\nimport type { Decoder } from './decoders';\nimport { DefaultDecoder } from './decoders/composite';\nimport { streamTransformations } from './transformations';\n\nconst { retry } = streamTransformations;\n\nexport const restream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n): ReadableStream<Transformed> =>\n retry(createSourceStream, handleChunk(transform, decoder), retryOptions)\n .readable;\n\nconst handleChunk =\n <Source = unknown, Transformed = Source, StreamType = Source>(\n transform: (input: Source) => Transformed = (source) =>\n source as unknown as Transformed,\n decoder: Decoder<StreamType, Source> = new DefaultDecoder<Source>(),\n ) =>\n (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ): void => {\n const { done: isDone, value } = readResult;\n\n if (value) decoder.addToBuffer(value);\n\n if (!isDone && !decoder.hasCompleteMessage()) return;\n\n decodeAndTransform(decoder, transform, controller);\n };\n\nconst decodeAndTransform = <StreamType, Source, Transformed = Source>(\n decoder: Decoder<StreamType, Source>,\n transform: (input: Source) => Transformed,\n controller: TransformStreamDefaultController<Transformed>,\n) => {\n try {\n const decoded = decoder.decode();\n if (!decoded) return; // TODO: Add a proper handling of decode errors\n\n const transformed = transform(decoded);\n controller.enqueue(transformed);\n } catch (error) {\n controller.error(new Error(`Decoding error: ${error?.toString()}`));\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const filter = <Item>(filter: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n if (filter(chunk)) {\n controller.enqueue(chunk);\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const map = <From, To>(map: (item: From) => To) =>\n new TransformStream<From, To>({\n transform(chunk, controller) {\n controller.enqueue(map(chunk));\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const reduce = <I, O>(\n reducer: (accumulator: O, chunk: I) => O,\n initialValue: O,\n) => new ReduceTransformStream<I, O>(reducer, initialValue);\n\nexport class ReduceTransformStream<I, O> extends TransformStream<I, O> {\n private accumulator: O;\n private reducer: (accumulator: O, chunk: I) => O;\n\n constructor(reducer: (accumulator: O, chunk: I) => O, initialValue: O) {\n super({\n transform: (chunk) => {\n this.accumulator = this.reducer(this.accumulator, chunk);\n },\n flush: (controller) => {\n controller.enqueue(this.accumulator);\n controller.terminate();\n },\n });\n\n this.accumulator = initialValue;\n this.reducer = reducer;\n }\n}\n","import {\n type ReadableStream,\n type ReadableStreamDefaultReadResult,\n TransformStream,\n type TransformStreamDefaultController,\n} from 'web-streams-polyfill';\nimport { type AsyncRetryOptions, asyncRetry } from '../../utils';\n\nexport const retryStream = <\n Source = unknown,\n Transformed = Source,\n StreamType = Source,\n>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n retryOptions: AsyncRetryOptions = { forever: true, minTimeout: 25 },\n): TransformStream<Source, Transformed> =>\n new TransformStream<Source, Transformed>({\n start(controller) {\n asyncRetry(\n () => onRestream(createSourceStream, handleChunk, controller),\n retryOptions,\n ).catch((error) => {\n controller.error(error);\n });\n },\n });\n\nconst onRestream = async <StreamType, Source, Transformed = Source>(\n createSourceStream: () => ReadableStream<StreamType>,\n handleChunk: (\n readResult: ReadableStreamDefaultReadResult<StreamType>,\n controller: TransformStreamDefaultController<Transformed>,\n ) => Promise<void> | void,\n controller: TransformStreamDefaultController<Transformed>,\n): Promise<void> => {\n const sourceStream = createSourceStream();\n const reader = sourceStream.getReader();\n\n try {\n let done: boolean;\n\n do {\n const result = await reader.read();\n done = result.done;\n\n await handleChunk(result, controller);\n\n if (done) {\n controller.terminate();\n }\n } while (!done);\n } finally {\n reader.releaseLock();\n }\n};\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const skip = <T>(limit: number) => new SkipTransformStream<T>(limit);\n\nexport class SkipTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private skip: number;\n\n constructor(skip: number) {\n super({\n transform: (chunk, controller) => {\n this.count++;\n if (this.count > this.skip) {\n controller.enqueue(chunk);\n }\n },\n });\n\n this.skip = skip;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopAfter = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n transform(chunk, controller) {\n controller.enqueue(chunk);\n\n if (stopCondition(chunk)) {\n controller.terminate();\n }\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const stopOn = <Item>(stopCondition: (item: Item) => boolean) =>\n new TransformStream<Item, Item>({\n async transform(chunk, controller) {\n if (!stopCondition(chunk)) {\n controller.enqueue(chunk);\n return;\n }\n await Promise.resolve();\n controller.terminate();\n },\n });\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const take = <T>(limit: number) => new TakeTransformStream<T>(limit);\n\nexport class TakeTransformStream<T> extends TransformStream<T, T> {\n private count = 0;\n private limit: number;\n\n constructor(limit: number) {\n super({\n transform: (chunk, controller) => {\n if (this.count < this.limit) {\n this.count++;\n controller.enqueue(chunk);\n } else {\n controller.terminate();\n }\n },\n });\n\n this.limit = limit;\n }\n}\n","import { TransformStream } from 'web-streams-polyfill';\n\nexport const waitAtMost = <Item>(waitTimeInMs: number) =>\n new TransformStream<Item, Item>({\n start(controller) {\n const timeoutId = setTimeout(() => {\n controller.terminate();\n }, waitTimeInMs);\n\n const originalTerminate = controller.terminate.bind(controller);\n\n // Clear the timeout if the stream is terminated early\n controller.terminate = () => {\n clearTimeout(timeoutId);\n originalTerminate();\n };\n },\n transform(chunk, controller) {\n controller.enqueue(chunk);\n },\n });\n","import { ConcurrencyError } from '../errors';\nimport type { Flavour } from '../typing';\nimport type { DefaultStreamVersionType } from './eventStore';\n\nexport type ExpectedStreamVersion<VersionType = DefaultStreamVersionType> =\n | ExpectedStreamVersionWithValue<VersionType>\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue<\n VersionType = DefaultStreamVersionType,\n> = Flavour<VersionType, 'StreamVersion'>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = <\n StreamVersion = DefaultStreamVersionType,\n>(\n current: StreamVersion | undefined,\n expected: ExpectedStreamVersion<StreamVersion>,\n defaultVersion: StreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = <\n StreamVersion = DefaultStreamVersionType,\n>(\n current: StreamVersion,\n expected: ExpectedStreamVersion<StreamVersion> | undefined,\n defaultVersion: StreamVersion,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError<\n VersionType = DefaultStreamVersionType,\n> extends ConcurrencyError {\n constructor(\n current: VersionType,\n expected: ExpectedStreamVersion<VersionType>,\n ) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError;\n","import { filter } from './filter';\nimport { map } from './map';\nimport {\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n} from './notifyAboutNoActiveReaders';\nimport { reduce, ReduceTransformStream } from './reduce';\nimport { retryStream } from './retry';\nimport { skip, SkipTransformStream } from './skip';\nimport { stopAfter } from './stopAfter';\nimport { stopOn } from './stopOn';\nimport { take, TakeTransformStream } from './take';\nimport { waitAtMost } from './waitAtMost';\n\nexport const streamTransformations = {\n filter,\n take,\n TakeTransformStream,\n skip,\n SkipTransformStream,\n map,\n notifyAboutNoActiveReadersStream,\n NotifyAboutNoActiveReadersStream,\n reduce,\n ReduceTransformStream,\n retry: retryStream,\n stopAfter,\n stopOn,\n waitAtMost,\n};\n","import {\n ExpectedVersionConflictError,\n STREAM_DOES_NOT_EXIST,\n assertExpectedVersionMatchesCurrent,\n type EventStore,\n type Event,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type ReadStreamOptions,\n type ReadStreamResult,\n type ReadEvent,\n type ReadEventMetadata,\n type ExpectedStreamVersion,\n} from '@event-driven-io/emmett';\nimport { type Collection, type WithId } from 'mongodb';\nimport { v4 as uuid } from 'uuid';\n\nexport const MongoDBEventStoreDefaultStreamVersion = 0;\n\nexport type StreamType = string;\nexport type StreamName<T extends StreamType = StreamType> = `${T}:${string}`;\n\nexport type StreamNameParts<T extends StreamType = StreamType> = {\n streamType: T;\n streamId: string;\n};\n\nexport type StreamToProject<EventType extends Event> = {\n streamName: StreamName;\n streamType: StreamType;\n streamId: string;\n streamVersion: number;\n events: ReadEvent<EventType, ReadEventMetadata>[];\n};\n\nexport type Projection<EventType extends Event> = (\n stream: StreamToProject<EventType>,\n) => void | Promise<void>;\n\nexport interface EventStream<EventType extends Event = Event> {\n streamName: string;\n events: Array<ReadEvent<EventType, ReadEventMetadata>>;\n createdAt: Date;\n updatedAt: Date;\n}\nexport type EventStreamEvent<EventType extends Event = Event> =\n EventStream<EventType>['events'][number];\n\nexport interface MongoDBConnectionOptions {\n connectionString: string;\n database: string;\n collection?: string;\n}\n\nexport class MongoDBEventStore implements EventStore<number> {\n private readonly collection: Collection<EventStream>;\n\n constructor(collection: typeof this.collection) {\n this.collection = collection;\n }\n\n async readStream<EventType extends Event>(\n streamName: StreamName,\n options?: ReadStreamOptions<number>,\n ): Promise<Exclude<ReadStreamResult<EventType, number>, null>> {\n const expectedStreamVersion = options?.expectedStreamVersion;\n\n const stream = await this.collection.findOne<\n WithId<EventStream<EventType>>\n >({\n streamName: { $eq: streamName },\n });\n\n if (!stream) {\n return {\n events: [],\n currentStreamVersion: MongoDBEventStoreDefaultStreamVersion,\n streamExists: false,\n };\n }\n\n assertExpectedVersionMatchesCurrent(\n stream.events.length,\n expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n return {\n events: stream.events.slice(0, maxEventIndex(expectedStreamVersion)),\n currentStreamVersion: stream.events.length,\n streamExists: true,\n };\n }\n\n async aggregateStream<State, EventType extends Event>(\n streamName: StreamName,\n options: AggregateStreamOptions<State, EventType, number>,\n ): Promise<AggregateStreamResult<State, number>> {\n const stream = await this.readStream<EventType>(streamName, options?.read);\n const state = stream.events.reduce(options.evolve, options.initialState());\n return {\n state,\n currentStreamVersion: stream.currentStreamVersion,\n streamExists: stream.streamExists,\n };\n }\n\n async appendToStream<EventType extends Event>(\n streamName: StreamName,\n events: EventType[],\n options?: AppendToStreamOptions<number> & {\n /**\n * These will be ran after a the events have been successfully appended to\n * the stream. `appendToStream` will return after every projection is completed.\n */\n projections?: Array<Projection<EventType>>;\n },\n ): Promise<AppendToStreamResult<number>> {\n let stream = await this.collection.findOne({\n streamName: { $eq: streamName },\n });\n let currentStreamPosition = stream?.events.length ?? 0;\n let createdNewStream = false;\n\n if (!stream) {\n const now = new Date();\n const result = await this.collection.insertOne({\n streamName,\n events: [],\n createdAt: now,\n updatedAt: now,\n });\n stream = await this.collection.findOne({\n _id: result.insertedId,\n });\n createdNewStream = true;\n }\n\n const eventCreateInputs: ReadEvent[] = [];\n for (const event of events) {\n currentStreamPosition++;\n eventCreateInputs.push({\n type: event.type,\n data: event.data,\n metadata: {\n now: new Date(),\n eventId: uuid(),\n streamName,\n streamPosition: BigInt(currentStreamPosition),\n ...(event.metadata ?? {}),\n },\n });\n }\n\n // TODO: better error here, should rarely happen if ever\n // if another error was not thrown before this\n if (!stream) throw new Error('Failed to create stream');\n\n assertExpectedVersionMatchesCurrent(\n stream.events.length,\n options?.expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n // @ts-expect-error The actual `EventType` is different across each stream document,\n // but the collection was instantiated as being `EventStream<Event>`. Unlike `findOne`,\n // `findOneAndUpdate` does not allow a generic to override what the return type is.\n const updatedStream: WithId<EventStream<EventType>> | null =\n await this.collection.findOneAndUpdate(\n {\n streamName: { $eq: streamName },\n events: { $size: stream.events.length },\n },\n {\n $push: { events: { $each: eventCreateInputs } },\n $set: { updatedAt: new Date() },\n },\n { returnDocument: 'after' },\n );\n\n if (!updatedStream) {\n const currentStream = await this.collection.findOne({\n streamName: { $eq: streamName },\n });\n throw new ExpectedVersionConflictError(\n currentStream?.events.length ?? -1,\n stream.events.length,\n );\n }\n\n const { streamType, streamId } = fromStreamName(streamName);\n\n await executeProjections(\n {\n streamName,\n streamType,\n streamId,\n streamVersion: updatedStream.events.length,\n events: updatedStream.events,\n },\n options?.projections,\n );\n\n return {\n nextExpectedStreamVersion: updatedStream.events.length,\n createdNewStream,\n };\n }\n}\n\nexport const getMongoDBEventStore = (collection: Collection<EventStream>) => {\n const eventStore = new MongoDBEventStore(collection);\n return eventStore;\n};\n\nfunction executeProjections<EventType extends Event>(\n params: StreamToProject<EventType>,\n projections?: Array<Projection<EventType>>,\n) {\n return Promise.all((projections ?? []).map((project) => project(params)));\n}\n\nfunction maxEventIndex(\n expectedStreamVersion?: ExpectedStreamVersion<number>,\n): number | undefined {\n if (!expectedStreamVersion) return undefined;\n\n if (typeof expectedStreamVersion === 'number') {\n return expectedStreamVersion;\n }\n\n switch (expectedStreamVersion) {\n case STREAM_DOES_NOT_EXIST:\n return 0;\n default:\n return undefined;\n }\n}\n\n/**\n * Accepts a `streamType` (the type/category of the event stream) and an `streamId`\n * (the individual entity/object or aggregate ID) and combines them to a singular\n * `streamName` which can be used in `EventStore`.\n */\nexport function toStreamName<T extends StreamType>(\n streamType: T,\n streamId: string,\n): StreamName<T> {\n return `${streamType}:${streamId}`;\n}\n\n/**\n * Accepts a fully formatted `streamName` and returns the broken down\n * `streamType` and `streamId`.\n */\nexport function fromStreamName<T extends StreamType>(\n streamName: StreamName<T>,\n): StreamNameParts<T> {\n const parts = streamName.split(':') as [T, string];\n return {\n streamType: parts[0],\n streamId: parts[1],\n };\n}\n"]}
@@ -0,0 +1,57 @@
1
+ import { Event, ReadEvent, ReadEventMetadata, EventStore, ReadStreamOptions, ReadStreamResult, AggregateStreamOptions, AggregateStreamResult, AppendToStreamOptions, AppendToStreamResult } from '@event-driven-io/emmett';
2
+ import { Collection } from 'mongodb';
3
+
4
+ declare const MongoDBEventStoreDefaultStreamVersion = 0;
5
+ type StreamType = string;
6
+ type StreamName<T extends StreamType = StreamType> = `${T}:${string}`;
7
+ type StreamNameParts<T extends StreamType = StreamType> = {
8
+ streamType: T;
9
+ streamId: string;
10
+ };
11
+ type StreamToProject<EventType extends Event> = {
12
+ streamName: StreamName;
13
+ streamType: StreamType;
14
+ streamId: string;
15
+ streamVersion: number;
16
+ events: ReadEvent<EventType, ReadEventMetadata>[];
17
+ };
18
+ type Projection<EventType extends Event> = (stream: StreamToProject<EventType>) => void | Promise<void>;
19
+ interface EventStream<EventType extends Event = Event> {
20
+ streamName: string;
21
+ events: Array<ReadEvent<EventType, ReadEventMetadata>>;
22
+ createdAt: Date;
23
+ updatedAt: Date;
24
+ }
25
+ type EventStreamEvent<EventType extends Event = Event> = EventStream<EventType>['events'][number];
26
+ interface MongoDBConnectionOptions {
27
+ connectionString: string;
28
+ database: string;
29
+ collection?: string;
30
+ }
31
+ declare class MongoDBEventStore implements EventStore<number> {
32
+ private readonly collection;
33
+ constructor(collection: typeof this.collection);
34
+ readStream<EventType extends Event>(streamName: StreamName, options?: ReadStreamOptions<number>): Promise<Exclude<ReadStreamResult<EventType, number>, null>>;
35
+ aggregateStream<State, EventType extends Event>(streamName: StreamName, options: AggregateStreamOptions<State, EventType, number>): Promise<AggregateStreamResult<State, number>>;
36
+ appendToStream<EventType extends Event>(streamName: StreamName, events: EventType[], options?: AppendToStreamOptions<number> & {
37
+ /**
38
+ * These will be ran after a the events have been successfully appended to
39
+ * the stream. `appendToStream` will return after every projection is completed.
40
+ */
41
+ projections?: Array<Projection<EventType>>;
42
+ }): Promise<AppendToStreamResult<number>>;
43
+ }
44
+ declare const getMongoDBEventStore: (collection: Collection<EventStream>) => MongoDBEventStore;
45
+ /**
46
+ * Accepts a `streamType` (the type/category of the event stream) and an `streamId`
47
+ * (the individual entity/object or aggregate ID) and combines them to a singular
48
+ * `streamName` which can be used in `EventStore`.
49
+ */
50
+ declare function toStreamName<T extends StreamType>(streamType: T, streamId: string): StreamName<T>;
51
+ /**
52
+ * Accepts a fully formatted `streamName` and returns the broken down
53
+ * `streamType` and `streamId`.
54
+ */
55
+ declare function fromStreamName<T extends StreamType>(streamName: StreamName<T>): StreamNameParts<T>;
56
+
57
+ export { type EventStream, type EventStreamEvent, type MongoDBConnectionOptions, MongoDBEventStore, MongoDBEventStoreDefaultStreamVersion, type Projection, type StreamName, type StreamNameParts, type StreamToProject, type StreamType, fromStreamName, getMongoDBEventStore, toStreamName };
@@ -0,0 +1,57 @@
1
+ import { Event, ReadEvent, ReadEventMetadata, EventStore, ReadStreamOptions, ReadStreamResult, AggregateStreamOptions, AggregateStreamResult, AppendToStreamOptions, AppendToStreamResult } from '@event-driven-io/emmett';
2
+ import { Collection } from 'mongodb';
3
+
4
+ declare const MongoDBEventStoreDefaultStreamVersion = 0;
5
+ type StreamType = string;
6
+ type StreamName<T extends StreamType = StreamType> = `${T}:${string}`;
7
+ type StreamNameParts<T extends StreamType = StreamType> = {
8
+ streamType: T;
9
+ streamId: string;
10
+ };
11
+ type StreamToProject<EventType extends Event> = {
12
+ streamName: StreamName;
13
+ streamType: StreamType;
14
+ streamId: string;
15
+ streamVersion: number;
16
+ events: ReadEvent<EventType, ReadEventMetadata>[];
17
+ };
18
+ type Projection<EventType extends Event> = (stream: StreamToProject<EventType>) => void | Promise<void>;
19
+ interface EventStream<EventType extends Event = Event> {
20
+ streamName: string;
21
+ events: Array<ReadEvent<EventType, ReadEventMetadata>>;
22
+ createdAt: Date;
23
+ updatedAt: Date;
24
+ }
25
+ type EventStreamEvent<EventType extends Event = Event> = EventStream<EventType>['events'][number];
26
+ interface MongoDBConnectionOptions {
27
+ connectionString: string;
28
+ database: string;
29
+ collection?: string;
30
+ }
31
+ declare class MongoDBEventStore implements EventStore<number> {
32
+ private readonly collection;
33
+ constructor(collection: typeof this.collection);
34
+ readStream<EventType extends Event>(streamName: StreamName, options?: ReadStreamOptions<number>): Promise<Exclude<ReadStreamResult<EventType, number>, null>>;
35
+ aggregateStream<State, EventType extends Event>(streamName: StreamName, options: AggregateStreamOptions<State, EventType, number>): Promise<AggregateStreamResult<State, number>>;
36
+ appendToStream<EventType extends Event>(streamName: StreamName, events: EventType[], options?: AppendToStreamOptions<number> & {
37
+ /**
38
+ * These will be ran after a the events have been successfully appended to
39
+ * the stream. `appendToStream` will return after every projection is completed.
40
+ */
41
+ projections?: Array<Projection<EventType>>;
42
+ }): Promise<AppendToStreamResult<number>>;
43
+ }
44
+ declare const getMongoDBEventStore: (collection: Collection<EventStream>) => MongoDBEventStore;
45
+ /**
46
+ * Accepts a `streamType` (the type/category of the event stream) and an `streamId`
47
+ * (the individual entity/object or aggregate ID) and combines them to a singular
48
+ * `streamName` which can be used in `EventStore`.
49
+ */
50
+ declare function toStreamName<T extends StreamType>(streamType: T, streamId: string): StreamName<T>;
51
+ /**
52
+ * Accepts a fully formatted `streamName` and returns the broken down
53
+ * `streamType` and `streamId`.
54
+ */
55
+ declare function fromStreamName<T extends StreamType>(streamName: StreamName<T>): StreamNameParts<T>;
56
+
57
+ export { type EventStream, type EventStreamEvent, type MongoDBConnectionOptions, MongoDBEventStore, MongoDBEventStoreDefaultStreamVersion, type Projection, type StreamName, type StreamNameParts, type StreamToProject, type StreamType, fromStreamName, getMongoDBEventStore, toStreamName };