@event-driven-io/emmett-mongodb 0.43.0-beta.6 → 0.43.0-beta.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs.map +1 -1
- package/dist/index.js.map +1 -1
- package/package.json +10 -10
package/dist/index.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/database/inMemoryDatabase.ts","../../emmett/src/taskProcessing/executionGuards.ts","../../emmett/src/utils/retry.ts","../../emmett/src/processors/processors.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjectionSpec.ts","../../emmett/src/workflows/handleWorkflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/typing/index.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/utils/strings/hashText.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/eventStore/versioning/downcasting.ts","../../emmett/src/eventStore/versioning/upcasting.ts","../../emmett/src/projections/index.ts","../src/eventStore/mongoDBEventStore.ts","../src/eventStore/projections/mongoDBInlineProjection.ts","../src/eventStore/projections/mongoDBInlineProjectionSpec.ts","../src/eventStore/storage/mongoDBEventStoreStorage.ts"],"names":[],"mappings":"AAAA;ACQO,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,SAAA,GAAY,IAAA,IAAQ,GAAA;AAM9B,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,QAAA;ACTV,IAAM,mBAAA,EAAqB,CAEhC,MAAA,EAAA,GAC0C;AAE1C,EAAA,OACE,OAAO,OAAA,IAAW,WAAA,GAClB,MAAA,CAAO,UAAA;AFTX,EEWI,MAAA,CAAO,SAAA,CAAU,YAAA,IAAgB,MAAA;AAErC,CAAA;AAEO,IAAM,YAAA,YAAN,MAAM,aAAA,QAAoB,MAAM;AFZvC,EEaE,4BAAuB,MAAA,EAAQ;AFZjC,IEaI,eAAA,EAAiB,GAAA;AFZrB,IEaI,iBAAA,EAAmB,GAAA;AFZvB,IEaI,aAAA,EAAe,GAAA;AFZnB,IEaI,gBAAA,EAAkB,GAAA;AFZtB,IEaI,mBAAA,EAAqB;AFZzB,EEaE,EAAA;AFZF,EEcS;AFbT,EEeE,WAAA,CACE,OAAA,EACA;AACA,IAAA,MAAM,UAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,YAAA,GAAe,QAAA,EACrD,OAAA,CAAQ,UAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,YAAA,CAAY,KAAA,CAAM,mBAAA;AAC1B,IAAA,MAAM,QAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,UAAA,GAAa,QAAA,EACnD,OAAA,CAAQ,QAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,CAAA,wBAAA,EAA2B,SAAS,CAAA,kCAAA,CAAA;AAE5C,IAAA,KAAA,CAAM,OAAO,CAAA;AACb,IAAA,IAAA,CAAK,UAAA,EAAY,SAAA;AAGjB,IAAA,MAAA,CAAO,cAAA,CAAe,IAAA,EAAM,YAAA,CAAY,SAAS,CAAA;AF7BrD,EE8BE;AF7BF,EE+BE,OAAc,OAAA,CACZ,KAAA,EACa;AACb,IAAA,GAAA,CAAI,YAAA,CAAY,YAAA,CAAa,KAAK,CAAA,EAAG;AACnC,MAAA,OAAO,KAAA;AFhCb,IEiCI;AAEA,IAAA,OAAO,IAAI,YAAA,CAAY;AFjC3B,MEkCM,SAAA,EACE,YAAA,GAAe,MAAA,GACf,KAAA,CAAM,UAAA,IAAc,KAAA,EAAA,GACpB,KAAA,CAAM,UAAA,IAAc,KAAA,EAChB,KAAA,CAAM,UAAA,EACN,YAAA,CAAY,KAAA,CAAM,mBAAA;AFtC9B,MEuCM,OAAA,mBAAS,KAAA,CAAM,OAAA,UAAW;AFtChC,IEuCI,CAAC,CAAA;AFtCL,EEuCE;AFtCF,EEwCE,OAAc,YAAA,CACZ,KAAA,EACA,SAAA,EACoB;AACpB,IAAA,OACE,OAAO,MAAA,IAAU,SAAA,GACjB,MAAA,IAAU,KAAA,GACV,YAAA,GAAe,MAAA,GACf,QAAA,CAAS,KAAA,CAAM,SAAS,EAAA,GAAA,CACvB,UAAA,IAAc,KAAA,EAAA,GAAa,KAAA,CAAM,UAAA,IAAc,SAAA,CAAA;AF/CtD,EEiDE;AACF,iCAAA;AAEO,IAAM,iBAAA,EAAN,MAAM,kBAAA,QAAyB,YAAY;AFjDlD,EEkDE,WAAA,CACS,OAAA,EACA,QAAA,EACP,OAAA,EACA;AACA,IAAA,KAAA,CAAM;AFrDV,MEsDM,SAAA,EAAW,WAAA,CAAY,KAAA,CAAM,gBAAA;AFrDnC,MEsDM,OAAA,mBACE,OAAA,UACA,CAAA,iBAAA,EAAoB,QAAA,CAAS,QAAA,CAAS,CAAC,CAAA,wBAAA,kBAA2B,OAAA,6BAAS,QAAA,mBAAS,GAAC,CAAA;AFvD7F,IAAA;AE+CW,IAAA;AACA,IAAA;AAWP,IAAA;AFvDJ,EAAA;AEyDA;AFvDA;AACA;AGhDA;ACAA;ACAA;ACAA;ACAA;ACAA;ACAA;AC0DA;AAUE,EAAA;AAEA,EAAA;AACE,IAAA;AACA,IAAA;AVXJ,EAAA;AUcI,IAAA;AACA,IAAA;AVZJ,EAAA;AUcA;AC3DO;AAGA;AACA;ACNA;AACA;AAEA;AAGA;AAKL,EAAA;AAEA,EAAA;AAEA,EAAA;AAEA,EAAA;AACF;AAEO;AAKL,EAAA;AAEA,EAAA;AACE,IAAA;AACJ;AAEO;AZoCP,EAAA;AYlCI,IAAA;AAGA,IAAA;AZkCJ,EAAA;AYhCA;ACvDO;AAIL,EAAA;AACA,EAAA;AAEA,EAAA;AACF;AAEO;AAIL,EAAA;AAEA,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AACE,MAAA;AbiFN,IAAA;Aa/EI,IAAA;AbiFJ,EAAA;Aa9EE,EAAA;AAGF;AC5BO;AAOL,EAAA;AAEA,EAAA;AAGI,IAAA;AAEA,IAAA;AACA,IAAA;AdiGN,EAAA;Ac1FM,IAAA;AAEA,IAAA;Ad2FN,EAAA;ActFE,EAAA;AACE,IAAA;AAEA,IAAA;AduFJ,EAAA;AcpFE,EAAA;AACF;AC/BO;AfsHP,EAAA;AACA,EAAA;AACA,EAAA;AepHA;ACTA;AACE,EAAA;AACA,EAAA;AASF;AAEA;AACE,EAAA;AACE,IAAA;AhBuHJ,EAAA;AgBrHE,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AhBuHJ,EAAA;AgBrHE,EAAA;AACF;AAEA;AACE,EAAA;AACF;AAEA;AACE,EAAA;AACF;AAEA;AACE,EAAA;AACE,IAAA;AhBoHJ,EAAA;AgBlHE,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACE,IAAA;AAEA,IAAA;AhBmHJ,EAAA;AgBjHE,EAAA;AACF;AAEA;AAIE,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AACE,QAAA;AhB8GR,MAAA;AACA,IAAA;AgB5GM,MAAA;AACA,MAAA;AACE,QAAA;AACE,UAAA;AACA,UAAA;AhB8GV,QAAA;AACA,MAAA;AgB5GM,MAAA;AhB8GN,IAAA;AACA,EAAA;AgB5GE,EAAA;AACF;AAEA;AACE,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AhB4GN,IAAA;AgB1GM,MAAA;AACA,MAAA;AACE,QAAA;AACE,UAAA;AACA,UAAA;AhB4GV,QAAA;AACA,MAAA;AgB1GM,MAAA;AhB4GN,IAAA;AACA,EAAA;AgB1GE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACE,IAAA;AhBwGJ,EAAA;AgBtGE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AAEA,EAAA;AhBmGF,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AgBjGE,EAAA;AhBmGF,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AgBhGE,EAAA;AACE,IAAA;AhBkGJ,EAAA;AgBhGE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AAEA,EAAA;AACE,IAAA;AhB6FJ,EAAA;AgB1FE,EAAA;AACE,IAAA;AACE,MAAA;AhB4FN,IAAA;AgBzFI,IAAA;AACA,IAAA;AACE,MAAA;AhB2FN,IAAA;AACA,EAAA;AgBxFE,EAAA;AACF;AAEA;AACE,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AACF;AAEO;AACL,EAAA;AAEA,EAAA;AACE,IAAA;AhBmFJ,EAAA;AgBhFE,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AhBgFF,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AgB9EM,MAAA;AhBgFN,IAAA;AgB7EM,MAAA;AhB+EN,IAAA;AgB5EM,MAAA;AhB8EN,IAAA;AgB3EM,MAAA;AhB6EN,IAAA;AgB1EM,MAAA;AhB4EN,IAAA;AgBzEM,MAAA;AhB2EN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBxEM,MAAA;AhB0EN,IAAA;AgBvEM,MAAA;AhByEN,IAAA;AACA,IAAA;AACA,IAAA;AgBtEM,MAAA;AhBwEN,IAAA;AgBrEM,MAAA;AhBuEN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBpEM,MAAA;AhBsEN,IAAA;AgBnEM,MAAA;AhBqEN,IAAA;AgBlEM,MAAA;AhBoEN,IAAA;AgBjEM,MAAA;AhBmEN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBhEM,MAAA;AhBkEN,EAAA;AgBhEA;AAIO;AACL,EAAA;AAOF;AC/QO;AjByUP,EAAA;AiBvUI,IAAA;AjByUJ,EAAA;AiBvUA;AA0BO;AjBgTP,EAAA;AiB3SI,IAAA;AjB6SJ,sBAAA;AACA;AACA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AACA,EAAA;AiBxSI,IAAA;AAEA,IAAA;AACE,MAAA;AAEF,IAAA;AjBwSJ,EAAA;AiBpSA;ACvDA;ACIO;AnB2VP,EAAA;AmBzVI,IAAA;AnB2VJ,EAAA;AmBzVA;AAEO;AACL,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AnBwVN,IAAA;AmBtVI,IAAA;AnBwVJ,EAAA;AmBtVA;AAEO;AACL,EAAA;AACF;AAqJO;AAIL,EAAA;AACE,IAAA;AACJ;AAGO;AAIL,EAAA;AACF;AC7KO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ApBqVF,IAAA;AACA,EAAA;AoB/UE,EAAA;ApBiVF,IAAA;AACA;AACA,IAAA;AACA,IAAA;AACA,MAAA;AACA,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AoBtUA;AClDO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ArBuWF,IAAA;AACA,EAAA;AqBjWE,EAAA;ArBmWF,IAAA;AACA;AACA,IAAA;AACA,IAAA;AACA,MAAA;AACA,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AqB7VA;AAEO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ArB0UF,IAAA;AACA,EAAA;AqBxUA;ACdO;AAWL,EAAA;AAIA,EAAA;AtB4UF,IAAA;AACA,IAAA;AACA,EAAA;AsBzUE,EAAA;AACE,IAAA;AtB2UJ;AACA,MAAA;AACA,0BAAA;AACA,EAAA;AsBxUE,EAAA;AACF;AAqBO;AtBsTP,EAAA;AACA,EAAA;AsBrSE;AAEK;AtBsSP,EAAA;AACA,EAAA;AsBtRE;AAEK;AtBuRP,EAAA;AACA,EAAA;AsBrRA;AtBuRA;AACA;AuBzaA;AAAA;AACE;AASF;AvBoaA;AACA;AwBtbO;AAiDA;AAML,EAAA;AAAM,IAAA;AACJ,IAAA;AACa,IAAA;AACJ,IAAA;AACT,IAAA;AACA,IAAA;AACA,EAAA;AAGF,EAAA;AAEA,EAAA;AAAmC,IAAA;AACmB,EAAA;AAGtD,EAAA;AACE,IAAA;AAAgC,MAAA;AACW,MAAA;AACzC,MAAA;AACA,MAAA;AACS,IAAA;AACV,EAAA;AAEL;AAqDO;AAOL,EAAA;AACA,EAAA;AAEA,EAAA;AAAO,IAAA;AACC,IAAA;AACa,IAAA;AAEjB,MAAA;AAEA,MAAA;AAKA,MAAA;AACE,QAAA;AAAsB,UAAA;AACpB,UAAA;AACA,QAAA;AACF,MAAA;AAGF,MAAA;AAA2C,QAAA;AACzC,QAAA;AACM,QAAA;AACN,QAAA;AACoD,MAAA;AAGtD,MAAA;AAEM,QAAA;AACK,QAAA;AACQ,MAAA;AAEb,IAAA;AACR,EAAA;AAEJ;AxB2TA;AACA;AyB/eA;AA0DO;AAAoC,EAAA;AAIvC,IAAA;AACE,MAAA;AAEA,MAAA;AAME,QAAA;AACA,QAAA;AAAO,UAAA;AAEH,YAAA;AAEA,YAAA;AAGA,YAAA;AAAO,cAAA;AAKH,gBAAA;AAGU,kBAAA;AACgB,kBAAA;AACA,gBAAA;AAG1B,gBAAA;AAAwC,kBAAA;AACM,kBAAA;AAC5C,gBAAA;AAGF,gBAAA;AACE,kBAAA;AAEA,kBAAA;AAEA,kBAAA;AACE,oBAAA;AAAA,uCAAA;AAEI,oBAAA;AACJ,gBAAA;AAEF,kBAAA;AAAmB,gBAAA;AACrB,cAAA;AACF,cAAA;AAIE,gBAAA;AAGU,kBAAA;AACgB,kBAAA;AACA,gBAAA;AAG1B,gBAAA;AAAwC,kBAAA;AACM,kBAAA;AAC5C,gBAAA;AAGF,gBAAA;AACE,kBAAA;AACA,kBAAA;AAA2D,gBAAA;AAE3D,kBAAA;AAEA,kBAAA;AAEA,kBAAA;AACE,oBAAA;AAAA,sBAAA;AAC4B,sBAAA;AACkC,oBAAA;AAE9D,oBAAA;AAAA,kBAAA;AAGF,kBAAA;AAAA,oBAAA;AACyB,oBAAA;AACmD,kBAAA;AAG5E,kBAAA;AACE,oBAAA;AAAA,sBAAA;AAC4B,sBAAA;AACkC,oBAAA;AAC9D,kBAAA;AACF,gBAAA;AAEA,kBAAA;AAAmB,gBAAA;AACrB,cAAA;AACF,YAAA;AACF,UAAA;AACF,QAAA;AACF,MAAA;AACF,IAAA;AACF,EAAA;AAEJ;AAEO;AAMkE,EAAA;AACvE,EAAA;AAEF;AAEO;AAMkE,EAAA;AACvE,EAAA;AAEF;AAEA;AASE,EAAA;AACA,EAAA;AAAmE,IAAA;AACjE,IAAA;AACA,EAAA;AAGF,EAAA;AACF;AAEA;AAAmE,EAAA;AAMjC,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACkD,EAAA;AACnD,EAAA;AAM2B,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACoD,EAAA;AACrD,EAAA;AAM2B,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AACD,EAAA;AAMsB,IAAA;AACrB,IAAA;AACA,IAAA;AACA,IAAA;AACoC,EAAA;AACrC,EAAA;AAIsB,IAAA;AACrB,IAAA;AACA,IAAA;AACA,IAAA;AACoC,EAAA;AAE5C;AAEO;AAA8B,EAAA;AAC2B,EAAA;AAEhE;AzBsWA;AACA;A0BvkBO;AASA;AAEP;AAIE,EAAA;AAIE,IAAA;AAAO,MAAA;AAIC,MAAA;AAEiD,IAAA;AACzD,EAAA;AAMA,IAAA;AAAO,MAAA;AAC4C,MAAA;AAEM,IAAA;AACzD,EAAA;AAEA,IAAA;AACA,IAAA;AAAO,MAAA;AAEkD,MAAA;AAIzC,IAAA;AAChB,EAAA;AAEJ;AAEA;AAKE,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AACE,IAAA;AAEA,IAAA;AAEA,IAAA;AAA2B,EAAA;AAG7B,EAAA;AACF;AAEA;AAKE,EAAA;AAEA,EAAA;AAIA,EAAA;AACE,IAAA;AACA,IAAA;AAEA,IAAA;AAAkB,MAAA;AAChB,MAAA;AACA,IAAA;AACF,EAAA;AAGF,EAAA;AACF;AAEO;AAIL,EAAA;AACA,EAAA;AACA,EAAA;AAGA,EAAA;AAEA,EAAA;AAAO,IAAA;AAOH,MAAA;AAAyC,QAAA;AACvC,QAAA;AACA,MAAA;AAGF,MAAA;AAIA,MAAA;AACE,QAAA;AACA,QAAA;AAA4C,UAAA;AAC1C,UAAA;AACA,UAAA;AACA,QAAA;AACD,MAAA;AAGH,MAAA;AAAO,IAAA;AACT,EAAA;AAEJ;A1BugBA;AACA;AuBrpBO;AAqJP;AAA8E,EAAA;AAC3D,EAAA;AACA,EAAA;AACT,iBAAA;AACoB,EAAA;AACpB,EAAA;AACA,EAAA;AACD,EAAA;AAGL,IAAA;AACA,IAAA;AAIA,IAAA;AACA,IAAA;AAAwC,MAAA;AACrB,MAAA;AACiC,IAAA;AAEpD,IAAA;AAAyB,MAAA;AACvB,uBAAA;AACwB,IAAA;AAG1B,IAAA;AAAmB,MAAA;AACT,QAAA;AACyC,QAAA;AACN,QAAA;AACE,MAAA;AAC7C,IAAA;AACF,EAAA;AACF,EAAA;AAWE,IAAA;AACA,IAAA;AAEA,IAAA;AAEA,IAAA;AAAe,MAAA;AACiB,IAAA;AAGhC,IAAA;AAEA,IAAA;AACE,MAAA;AAAwC,IAAA;AAExC,MAAA;AAAqB,IAAA;AAGvB,IAAA;AACE,MAAA;AAAsC,IAAA;AAGxC,IAAA;AAGA,IAAA;AAEU,MAAA;AACK,MAAA;AACD,QAAA;AACA,QAAA;AACA,MAAA;AACZ,IAAA;AAGF,IAAA;AACE,MAAA;AAAO,QAAA;AACI,QAAA;AACa,QAAA;AACR,MAAA;AAChB,IAAA;AAGF,IAAA;AAAA,MAAA;AACkB,MAAA;AAChB,MAAA;AACA,IAAA;AAGF,IAAA;AAAe,MAAA;AACN,sBAAA;AACU,IAAA;AAGnB,IAAA;AAAO,MAAA;AACL,MAAA;AACsC,MAAA;AACxB,IAAA;AAChB,EAAA;AACF,EAAA;AAeE,IAAA;AAA0B,MAAA;AACxB,sBAAA;AACS,IAAA;AAEX,IAAA;AAEA,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AAC6B,MAAA;AACR,IAAA;AACvB,EAAA;AACF,EAAA;AAUE,IAAA;AACA,IAAA;AAEA,IAAA;AAEA,IAAA;AAAgC,MAAA;AAGI,MAAA;AAClC,QAAA;AACe,QAAA;AACD,UAAA;AACiB,UAAA;AACd,QAAA;AACf,MAAA;AACF,IAAA;AAGF,IAAA;AAGA,IAAA;AAAA,MAAA;AACE,MAAA;AACA,MAAA;AACA,IAAA;AAGF,IAAA;AAEA,IAAA;AAIE,MAAA;AAA2C,QAAA;AACzB,QAAA;AAChB,QAAA;AACkB,MAAA;AAEpB,MAAA;AAAO,QAAA;AACL,UAAA;AACc,UAAA;AACA,UAAA;AACF,YAAA;AACL,YAAA;AACiD,UAAA;AACtD,QAAA;AACF,wBAAA;AACiB,MAAA;AACnB,IAAA;AAGF,IAAA;AACA,IAAA;AAA2C,MAAA;AACI,MAAA;AACvC,QAAA;AACkB,QAAA;AACgD,MAAA;AACxE,MAAA;AACc,QAAA;AACZ,QAAA;AACqB,QAAA;AACE,QAAA;AACD,MAAA;AACxB,IAAA;AAGF,IAAA;AACE,MAAA;AAA8B,QAAA;AACQ,QAAA;AACpC,QAAA;AACQ,QAAA;AACU,QAAA;AAClB,QAAA;AACA,QAAA;AACS,MAAA;AACV,IAAA;AAGH,IAAA;AAAuC,MAAA;AACrC,QAAA;AACgC,QAAA;AACH,MAAA;AAC7B,MAAA;AACA,MAAA;AACkC,IAAA;AAGpC,IAAA;AACE,MAAA;AAAU,QAAA;AACR,yCAAA;AACkC,MAAA;AACpC,IAAA;AAGF,IAAA;AAAM,MAAA;AACJ,MAAA;AACa;AAAA;AAAA;AAAA,IAAA;AAMf,IAAA;AAAO,MAAA;AAEgD,MAAA;AAE1B,IAAA;AAC7B,EAAA;AACF,EAAA;AAGE,IAAA;AAEA,IAAA;AAEA,IAAA;AAAe,MAAA;AACiB,IAAA;AAGhC,IAAA;AAAsD,MAAA;AACvC,MAAA;AACN,IAAA;AAGT,IAAA;AAAwB,EAAA;AAC1B,kBAAA;AAKE,IAAA;AAA4C,EAAA;AAC9C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,kBAAA;AAUE,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AAAyB,EAAA;AAC3B,EAAA;AAME,IAAA;AAEA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACkB,IAAA;AAGzD,IAAA;AACE,MAAA;AAAkB,IAAA;AAGpB,IAAA;AACE,MAAA;AAAgD,IAAA;AAGlD,IAAA;AAAgC,MAAA;AAGd,MAAA;AAChB,QAAA;AACe,QAAA;AACsC,MAAA;AACrD,IAAA;AAGF,IAAA;AAAgD,EAAA;AAClD,EAAA;AAOE,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AACA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACc,IAAA;AAGrD,IAAA;AACE,MAAA;AAA6B,IAAA;AAG/B,IAAA;AACE,MAAA;AAAiD,IAAA;AAGnD,IAAA;AAAuB,MAAA;AAKL,MAAA;AAChB,QAAA;AACe,QAAA;AACsC,MAAA;AACrD,IAAA;AAGF,IAAA;AACE,MAAA;AAAoC,IAAA;AAGtC,IAAA;AACE,MAAA;AAAsC,IAAA;AAGxC,IAAA;AACE,MAAA;AAAa,QAAA;AACE,QAAA;AACb,MAAA;AAEF,MAAA;AAAuB,IAAA;AAGzB,IAAA;AAEA,IAAA;AAEgD,EAAA;AAClD,EAAA;AAME,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AACA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACc,IAAA;AAGrD,IAAA;AACE,MAAA;AAA6B,IAAA;AAG/B,IAAA;AACE,MAAA;AAAiD,IAAA;AAGnD,IAAA;AACA,IAAA;AAAO,EAAA;AACT,kBAAA;AAGE,IAAA;AACA,IAAA;AAAY,EAAA;AAEhB;AAEA;AAGE,EAAA;AAGA,EAAA;AACE,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AACyB,MAAA;AACzB,IAAA;AACF,EAAA;AAGF,EAAA;AACE,IAAA;AAAmB,MAAA;AACJ,MAAA;AACA,IAAA;AAEf,IAAA;AAAO,MAAA;AACL,MAAA;AACA,MAAA;AACyB,IAAA;AAC3B,EAAA;AAGF,EAAA;AAAO,IAAA;AACL,IAAA;AACyB,EAAA;AAE7B;AAEA;AAGE,EAAA;AAGA,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AAC0B,MAAA;AAC1B,IAAA;AACF,EAAA;AAGF,EAAA;AACE,IAAA;AAA2C,MAAA;AACD,IAAA;AAE1C,IAAA;AAAO,MAAA;AACL,MAAA;AACA,MAAA;AACyB,IAAA;AAC3B,EAAA;AAGF,EAAA;AAAO,IAAA;AACL,IAAA;AACyB,EAAA;AAE7B;AAKO;AAIL,EAAA;AACE,IAAA;AAAO,EAAA;AAGT,EAAA;AACE,IAAA;AAEE,MAAA;AAA8D,IAAA;AAEhE,IAAA;AAAO,EAAA;AAGT,EAAA;AAEE,IAAA;AACA,IAAA;AACE,MAAA;AACA,MAAA;AAAkC,IAAA;AAGpC,IAAA;AAA8D,EAAA;AAGhE,EAAA;AACF;AAEA;AAEE,EAAA;AACE,IAAA;AAAO,EAAA;AAGT,EAAA;AACF;AAWO;AAGL,EAAA;AAGA,EAAA;AACE,IAAA;AAA0D,EAAA;AAG5D,EAAA;AACF;AAOO;AAIL,EAAA;AACF;AAMO;AAGL,EAAA;AACA,EAAA;AAAO,IAAA;AACc,IAAA;AACF,EAAA;AAErB;AAMO;AAGL,EAAA;AACF;AAKO;AAGL,EAAA;AACA,EAAA;AAAO,IAAA;AACc,EAAA;AAEvB;AvBkTA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA","file":"/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","sourcesContent":[null,"import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const isBigint = (val: any): val is bigint =>\n typeof val === 'bigint' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public static readonly Codes = {\n ValidationError: 400,\n IllegalStateError: 403,\n NotFoundError: 404,\n ConcurrencyError: 412,\n InternalServerError: 500,\n };\n\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : EmmettError.Codes.InternalServerError;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n\n public static mapFrom(\n error: Error | { message?: string; errorCode?: number },\n ): EmmettError {\n if (EmmettError.isInstanceOf(error)) {\n return error;\n }\n\n return new EmmettError({\n errorCode:\n 'errorCode' in error &&\n error.errorCode !== undefined &&\n error.errorCode !== null\n ? error.errorCode\n : EmmettError.Codes.InternalServerError,\n message: error.message ?? 'An unknown error occurred',\n });\n }\n\n public static isInstanceOf<ErrorType extends EmmettError = EmmettError>(\n error: unknown,\n errorCode?: (typeof EmmettError.Codes)[keyof typeof EmmettError.Codes],\n ): error is ErrorType {\n return (\n typeof error === 'object' &&\n error !== null &&\n 'errorCode' in error &&\n isNumber(error.errorCode) &&\n (errorCode === undefined || error.errorCode === errorCode)\n );\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: EmmettError.Codes.ConcurrencyError,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\n// TODO: Make it derive from ConcurrencyError to avoid code duplication\n// Or add additional type to distinguinsh both errors\nexport class ConcurrencyInMemoryDatabaseError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.ConcurrencyError,\n message: message ?? `Expected document state does not match current one!`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyInMemoryDatabaseError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.ValidationError,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.IllegalStateError,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: EmmettError.Codes.NotFoundError,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport {\n getInMemoryDatabase,\n type InMemoryDatabase,\n} from '../database/inMemoryDatabase';\nimport { bigIntProcessorCheckpoint } from '../processors';\nimport type { ProjectionRegistration } from '../projections';\nimport type {\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n DefaultEventStoreOptions,\n EventStore,\n ReadStreamOptions,\n ReadStreamResult,\n StreamExistsResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { handleInMemoryProjections } from './projections/inMemory';\nimport { downcastRecordedMessages, upcastRecordedMessages } from './versioning';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition> & {\n database: InMemoryDatabase;\n };\n\nexport type InMemoryReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type InMemoryProjectionHandlerContext = {\n eventStore?: InMemoryEventStore;\n database?: InMemoryDatabase;\n};\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore> & {\n projections?: ProjectionRegistration<\n 'inline',\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >[];\n database?: InMemoryDatabase;\n };\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n // Get the database instance to be used for projections\n const database = eventStoreOptions?.database || getInMemoryDatabase();\n\n // Extract inline projections from options\n const inlineProjections = (eventStoreOptions?.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n // Create the event store object\n const eventStore: InMemoryEventStore = {\n database,\n async aggregateStream<\n State,\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition,\n EventPayloadType\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType, EventPayloadType>(\n streamName,\n read,\n );\n\n const events = result?.events ?? [];\n\n const state = events.reduce((s, e) => evolve(s, e), initialState());\n\n return {\n currentStreamVersion: BigInt(events.length),\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: <\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n options?: ReadStreamOptions<EventType, EventPayloadType>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options?.from ?? 0);\n const to = Number(\n options?.to ??\n (options?.maxCount\n ? (options.from ?? 0n) + options.maxCount\n : (events?.length ?? 1)),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? upcastRecordedMessages<\n EventType,\n EventPayloadType,\n ReadEventMetadataWithGlobalPosition\n >(\n events.slice(from, to) as ReadEvent<\n EventPayloadType,\n ReadEventMetadataWithGlobalPosition\n >[],\n options?.schema?.versioning,\n )\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<EventType, EventPayloadType>,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const globalPosition = BigInt(getAllEventsCount() + index + 1);\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition,\n checkpoint: bigIntProcessorCheckpoint(globalPosition),\n };\n return {\n ...event,\n kind: event.kind ?? 'Event',\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [\n ...currentEvents,\n ...downcastRecordedMessages(newEvents, options?.schema?.versioning),\n ]);\n\n // Process projections if there are any registered\n if (inlineProjections.length > 0) {\n await handleInMemoryProjections({\n projections: inlineProjections,\n events: newEvents,\n database: eventStore.database,\n eventStore,\n });\n }\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit<InMemoryEventStore>(\n newEvents,\n eventStoreOptions?.hooks,\n );\n\n return result;\n },\n\n streamExists: (streamName): Promise<StreamExistsResult> => {\n const events = streams.get(streamName);\n\n return Promise.resolve(events !== undefined && events.length > 0);\n },\n };\n\n return eventStore;\n};\n","import { v7 as uuid } from 'uuid';\nimport { deepEquals } from '../utils';\nimport type {\n DatabaseHandleOptionErrors,\n DatabaseHandleOptions,\n DatabaseHandleResult,\n DeleteResult,\n Document,\n DocumentHandler,\n InsertOneResult,\n OptionalUnlessRequiredIdAndVersion,\n ReplaceOneOptions,\n UpdateResult,\n WithIdAndVersion,\n WithoutId,\n} from './types';\nimport { expectedVersionValue, operationResult } from './utils';\n\nexport interface InMemoryDocumentsCollection<T extends Document> {\n handle: (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ) => Promise<DatabaseHandleResult<T>>;\n findOne: (predicate?: Predicate<T>) => Promise<T | null>;\n find: (predicate?: Predicate<T>) => Promise<T[]>;\n insertOne: (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ) => Promise<InsertOneResult>;\n deleteOne: (predicate?: Predicate<T>) => Promise<DeleteResult>;\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ) => Promise<UpdateResult>;\n}\n\nexport interface InMemoryDatabase {\n collection: <T extends Document>(\n name: string,\n ) => InMemoryDocumentsCollection<T>;\n}\n\ntype Predicate<T> = (item: T) => boolean;\ntype CollectionName = string;\n\nexport const getInMemoryDatabase = (): InMemoryDatabase => {\n const storage = new Map<CollectionName, WithIdAndVersion<Document>[]>();\n\n return {\n collection: <T extends Document, CollectionName extends string>(\n collectionName: CollectionName,\n collectionOptions: {\n errors?: DatabaseHandleOptionErrors;\n } = {},\n ): InMemoryDocumentsCollection<T> => {\n const ensureCollectionCreated = () => {\n if (!storage.has(collectionName)) storage.set(collectionName, []);\n };\n\n const errors = collectionOptions.errors;\n\n const collection = {\n collectionName,\n insertOne: async (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ): Promise<InsertOneResult> => {\n ensureCollectionCreated();\n\n const _id = (document._id as string | undefined | null) ?? uuid();\n const _version = document._version ?? 1n;\n\n const existing = await collection.findOne((c) => c._id === _id);\n\n if (existing) {\n return operationResult<InsertOneResult>(\n {\n successful: false,\n insertedId: null,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n }\n\n const documentsInCollection = storage.get(collectionName)!;\n const newDocument = { ...document, _id, _version };\n const newCollection = [...documentsInCollection, newDocument];\n storage.set(collectionName, newCollection);\n\n return operationResult<InsertOneResult>(\n {\n successful: true,\n insertedId: _id,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n },\n findOne: (predicate?: Predicate<T>): Promise<T | null> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n const firstOne = filteredDocuments?.[0] ?? null;\n\n return Promise.resolve(firstOne as T | null);\n },\n find: (predicate?: Predicate<T>): Promise<T[]> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n return Promise.resolve(filteredDocuments as T[]);\n },\n deleteOne: (predicate?: Predicate<T>): Promise<DeleteResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n if (predicate) {\n const foundIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (foundIndex === -1) {\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: false,\n matchedCount: 0,\n deletedCount: 0,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n } else {\n const newCollection = documentsInCollection.toSpliced(\n foundIndex,\n 1,\n );\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n }\n }\n\n const newCollection = documentsInCollection.slice(1);\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n },\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ): Promise<UpdateResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n const firstIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (firstIndex === undefined || firstIndex === -1) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 0,\n modifiedCount: 0,\n nextExpectedVersion: 0n,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const existing = documentsInCollection[firstIndex]!;\n\n if (\n typeof options?.expectedVersion === 'bigint' &&\n existing._version !== options.expectedVersion\n ) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 1,\n modifiedCount: 0,\n nextExpectedVersion: existing._version,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const newVersion = existing._version + 1n;\n\n const newCollection = documentsInCollection.with(firstIndex, {\n _id: existing._id,\n ...document,\n _version: newVersion,\n });\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: true,\n modifiedCount: 1,\n matchedCount: firstIndex,\n nextExpectedVersion: newVersion,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n },\n handle: async (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ): Promise<DatabaseHandleResult<T>> => {\n const { expectedVersion: version, ...operationOptions } =\n options ?? {};\n ensureCollectionCreated();\n const existing = await collection.findOne(({ _id }) => _id === id);\n\n const expectedVersion = expectedVersionValue(version);\n\n if (\n (existing == null && version === 'DOCUMENT_EXISTS') ||\n (existing == null && expectedVersion != null) ||\n (existing != null && version === 'DOCUMENT_DOES_NOT_EXIST') ||\n (existing != null &&\n expectedVersion !== null &&\n existing._version !== expectedVersion)\n ) {\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: false,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n }\n\n const result = handle(existing !== null ? { ...existing } : null);\n\n if (deepEquals(existing, result))\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n\n if (!existing && result) {\n const newDoc = { ...result, _id: id };\n const insertResult = await collection.insertOne({\n ...newDoc,\n _id: id,\n } as OptionalUnlessRequiredIdAndVersion<T>);\n return {\n ...insertResult,\n document: {\n ...newDoc,\n _version: insertResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n if (existing && !result) {\n const deleteResult = await collection.deleteOne(\n ({ _id }) => id === _id,\n );\n return { ...deleteResult, document: null };\n }\n\n if (existing && result) {\n const replaceResult = await collection.replaceOne(\n ({ _id }) => id === _id,\n result,\n {\n ...operationOptions,\n expectedVersion: expectedVersion ?? 'DOCUMENT_EXISTS',\n },\n );\n return {\n ...replaceResult,\n document: {\n ...result,\n _version: replaceResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n },\n };\n\n return collection;\n },\n };\n};\n","import { v7 as uuid } from 'uuid';\nimport { TaskProcessor } from './taskProcessor';\n\nexport type ExclusiveAccessGuard = {\n execute: <Result>(operation: () => Promise<Result>) => Promise<Result>;\n waitForIdle: () => Promise<void>;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardExclusiveAccess = (options?: {\n maxQueueSize?: number;\n}): ExclusiveAccessGuard => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: 1,\n maxQueueSize: options?.maxQueueSize ?? 1000,\n });\n\n return {\n execute: <Result>(operation: () => Promise<Result>): Promise<Result> =>\n taskProcessor.enqueue(async ({ ack }) => {\n try {\n return await operation();\n } finally {\n ack();\n }\n }),\n waitForIdle: () => taskProcessor.waitForEndOfProcessing(),\n stop: (options) => taskProcessor.stop(options),\n };\n};\n\nexport type BoundedAccessGuard<Resource> = {\n acquire: () => Promise<Resource>;\n release: (resource: Resource) => void;\n execute: <Result>(\n operation: (resource: Resource) => Promise<Result>,\n ) => Promise<Result>;\n waitForIdle: () => Promise<void>;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardBoundedAccess = <Resource>(\n getResource: () => Resource | Promise<Resource>,\n options: {\n maxResources: number;\n maxQueueSize?: number;\n reuseResources?: boolean;\n closeResource?: (resource: Resource) => void | Promise<void>;\n },\n): BoundedAccessGuard<Resource> => {\n let isStopped = false;\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: options.maxResources,\n maxQueueSize: options.maxQueueSize ?? 1000,\n });\n\n const resourcePool: Resource[] = [];\n const allResources = new Set<Resource>();\n const ackCallbacks = new Map<Resource, () => void>();\n\n const acquire = async (): Promise<Resource> =>\n taskProcessor.enqueue(async ({ ack }) => {\n try {\n let resource: Resource | undefined;\n\n if (options.reuseResources) {\n resource = resourcePool.pop();\n }\n\n if (!resource) {\n resource = await getResource();\n allResources.add(resource);\n }\n\n ackCallbacks.set(resource, ack);\n return resource;\n } catch (e) {\n ack();\n throw e;\n }\n });\n\n const release = (resource: Resource) => {\n const ack = ackCallbacks.get(resource);\n if (ack) {\n ackCallbacks.delete(resource);\n if (options.reuseResources) {\n resourcePool.push(resource);\n }\n ack();\n }\n };\n\n const execute = async <Result>(\n operation: (resource: Resource) => Promise<Result>,\n ): Promise<Result> => {\n const resource = await acquire();\n try {\n return await operation(resource);\n } finally {\n release(resource);\n }\n };\n\n return {\n acquire,\n release,\n execute,\n waitForIdle: () => taskProcessor.waitForEndOfProcessing(),\n stop: async (stopOptions) => {\n if (isStopped) return;\n isStopped = true;\n if (options?.closeResource) {\n const resources = [...allResources];\n allResources.clear();\n resourcePool.length = 0;\n await Promise.all(\n resources.map(\n async (resource) => await options.closeResource!(resource),\n ),\n );\n }\n\n await taskProcessor.stop(stopOptions);\n },\n };\n};\n\nexport type InitializedOnceGuard<T> = {\n ensureInitialized: () => Promise<T>;\n reset: () => void;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardInitializedOnce = <T>(\n initialize: () => Promise<T>,\n options?: {\n maxQueueSize?: number;\n maxRetries?: number;\n },\n): InitializedOnceGuard<T> => {\n let initPromise: Promise<T> | null = null;\n\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: 1,\n maxQueueSize: options?.maxQueueSize ?? 1000,\n });\n\n const ensureInitialized = async (retryCount = 0): Promise<T> => {\n if (initPromise !== null) {\n return initPromise;\n }\n\n return taskProcessor.enqueue(\n async ({ ack }) => {\n if (initPromise !== null) {\n ack();\n return initPromise;\n }\n\n try {\n const promise = initialize();\n initPromise = promise;\n const result = await promise;\n ack();\n return result;\n } catch (error) {\n initPromise = null;\n ack();\n const maxRetries = options?.maxRetries ?? 3;\n if (retryCount < maxRetries) {\n return ensureInitialized(retryCount + 1);\n }\n throw error;\n }\n },\n { taskGroupId: uuid() },\n );\n };\n\n return {\n ensureInitialized,\n reset: () => {\n initPromise = null;\n },\n stop: (options) => taskProcessor.stop(options),\n };\n};\n","import retry from 'async-retry';\nimport { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\n\nexport type AsyncRetryOptions<T = unknown> = retry.Options & {\n shouldRetryResult?: (result: T) => boolean;\n shouldRetryError?: (error?: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions<T>,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n const result = await fn();\n\n if (opts?.shouldRetryResult && opts.shouldRetryResult(result)) {\n throw new EmmettError(\n `Retrying because of result: ${JSONParser.stringify(result)}`,\n );\n }\n return result;\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n return undefined as unknown as T;\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","import { v7 as uuid } from 'uuid';\nimport type { EmmettError } from '../errors';\nimport { upcastRecordedMessage } from '../eventStore';\nimport type { ProjectionDefinition } from '../projections';\nimport {\n defaultTag,\n type AnyEvent,\n type AnyMessage,\n type AnyReadEventMetadata,\n type AnyRecordedMessageMetadata,\n type BatchRecordedMessageHandlerWithContext,\n type Brand,\n type CanHandle,\n type DefaultRecord,\n type Event,\n type Message,\n type MessageHandlerResult,\n type RecordedMessage,\n type SingleMessageHandlerWithContext,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport { bigInt } from '../utils';\nimport { onShutdown } from '../utils/shutdown';\n\nexport type CurrentMessageProcessorPosition =\n | { lastCheckpoint: ProcessorCheckpoint }\n | 'BEGINNING'\n | 'END';\n\nexport type GetCheckpoint<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n) => ProcessorCheckpoint | null;\n\nexport const getCheckpoint = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n): ProcessorCheckpoint | null => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return, @typescript-eslint/no-unsafe-member-access\n return message.metadata.checkpoint;\n};\n\nexport const wasMessageHandled = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n checkpoint: ProcessorCheckpoint | null,\n): boolean => {\n //TODO Make it smarter\n const messageCheckpoint = getCheckpoint(message);\n\n return (\n messageCheckpoint !== null &&\n messageCheckpoint !== undefined &&\n checkpoint !== null &&\n checkpoint !== undefined &&\n messageCheckpoint <= checkpoint\n );\n};\n\nexport type MessageProcessorStartFrom =\n | CurrentMessageProcessorPosition\n | 'CURRENT';\n\nexport type MessageProcessorType = 'projector' | 'reactor';\nexport const MessageProcessorType = {\n PROJECTOR: 'projector' as MessageProcessorType,\n REACTOR: 'reactor' as MessageProcessorType,\n};\n\nexport type MessageProcessor<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n id: string;\n instanceId: string;\n type: string;\n canHandle?: string[];\n init: (options: Partial<HandlerContext>) => Promise<void>;\n start: (\n options: Partial<HandlerContext>,\n ) => Promise<CurrentMessageProcessorPosition | undefined>;\n close: (closeOptions: Partial<HandlerContext>) => Promise<void>;\n isActive: boolean;\n handle: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n Partial<HandlerContext>\n >;\n};\n\nexport const MessageProcessor = {\n result: {\n skip: (options?: { reason?: string }): MessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): MessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type MessageProcessingScope<\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = <Result = MessageHandlerResult>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => Result | Promise<Result>;\n\nexport type Checkpointer<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n read: ReadProcessorCheckpoint<HandlerContext>;\n store: StoreProcessorCheckpoint<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n};\n\nexport type ProcessorHooks<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n onInit?: OnReactorInitHook<HandlerContext>;\n onStart?: OnReactorStartHook<HandlerContext>;\n onClose?: OnReactorCloseHook<HandlerContext>;\n};\n\nexport type BaseMessageProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type?: string;\n processorId: string;\n processorInstanceId?: string;\n version?: number;\n partition?: string;\n startFrom?: MessageProcessorStartFrom;\n stopAfter?: (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n ) => boolean;\n processingScope?: MessageProcessingScope<HandlerContext>;\n checkpoints?: Checkpointer<MessageType, MessageMetadataType, HandlerContext>;\n canHandle?: CanHandle<MessageType>;\n hooks?: ProcessorHooks<HandlerContext>;\n};\n\nexport type HandlerOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> =\n | {\n eachMessage: SingleRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n eachBatch?: never;\n }\n | {\n eachMessage?: never;\n eachBatch: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n };\n\nexport type OnReactorInitHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorStartHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorCloseHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type ReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n MessagePayloadType extends AnyMessage = MessageType,\n> = BaseMessageProcessorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext\n> &\n HandlerOptions<MessageType, MessageMetadataType, HandlerContext> & {\n messageOptions?: {\n schema?: {\n versioning?: { upcast?: (event: MessagePayloadType) => MessageType };\n };\n };\n };\n\nexport type ProjectorOptions<\n EventType extends AnyEvent = AnyEvent,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n> = Omit<\n BaseMessageProcessorOptions<EventType, MessageMetadataType, HandlerContext>,\n 'type' | 'processorId'\n> & { processorId?: string } & {\n truncateOnStart?: boolean;\n projection: ProjectionDefinition<\n EventType,\n MessageMetadataType,\n HandlerContext,\n EventPayloadType\n >;\n};\n\nexport const defaultProcessingMessageProcessingScope = <\n HandlerContext = never,\n Result = MessageHandlerResult,\n>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => handler(partialContext as HandlerContext);\n\nexport type ProcessorCheckpoint = Brand<string, 'ProcessorCheckpoint'>;\n\nexport const bigIntProcessorCheckpoint = (value: bigint): ProcessorCheckpoint =>\n bigInt.toNormalizedString(value) as ProcessorCheckpoint;\n\nexport const parseBigIntProcessorCheckpoint = (\n value: ProcessorCheckpoint,\n): bigint => BigInt(value);\n\nexport type ReadProcessorCheckpointResult = {\n lastCheckpoint: ProcessorCheckpoint | null;\n};\n\nexport type ReadProcessorCheckpoint<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n options: { processorId: string; partition?: string },\n context: HandlerContext,\n) => Promise<ReadProcessorCheckpointResult>;\n\nexport type StoreProcessorCheckpointResult =\n | {\n success: true;\n newCheckpoint: ProcessorCheckpoint | null;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' | 'CURRENT_AHEAD' };\n\nexport type StoreProcessorCheckpoint<\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = (\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: ProcessorCheckpoint | null;\n partition?: string;\n },\n context: HandlerContext,\n) => Promise<StoreProcessorCheckpointResult>;\n\nexport const defaultProcessorVersion = 1;\nexport const defaultProcessorPartition = defaultTag;\n\nexport const getProcessorInstanceId = (processorId: string): string =>\n `${processorId}:${uuid()}`;\n\nexport const getProjectorId = (options: { projectionName: string }): string =>\n `emt:processor:projector:${options.projectionName}`;\n\nexport const reactor = <\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n MessagePayloadType extends Message = MessageType,\n>(\n options: ReactorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n MessagePayloadType\n >,\n): MessageProcessor<MessageType, MessageMetadataType, HandlerContext> => {\n const {\n checkpoints,\n processorId,\n processorInstanceId: instanceId = getProcessorInstanceId(processorId),\n type = MessageProcessorType.REACTOR,\n version = defaultProcessorVersion,\n partition = defaultProcessorPartition,\n hooks = {},\n processingScope = defaultProcessingMessageProcessingScope,\n startFrom,\n canHandle,\n stopAfter,\n } = options;\n\n const eachMessage: SingleMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n > =\n 'eachMessage' in options && options.eachMessage\n ? options.eachMessage\n : () => Promise.resolve();\n\n let isInitiated = false;\n let isActive = false;\n\n let lastCheckpoint: ProcessorCheckpoint | null = null;\n let closeSignal: (() => void) | null = null;\n\n const init = async (initOptions: Partial<HandlerContext>): Promise<void> => {\n if (isInitiated) return;\n\n if (hooks.onInit === undefined) {\n isInitiated = true;\n return;\n }\n\n return await processingScope(async (context) => {\n await hooks.onInit!(context);\n isInitiated = true;\n }, initOptions);\n };\n\n const close = async (\n closeOptions: Partial<HandlerContext>,\n ): Promise<void> => {\n // TODO: Align when active is set to false\n // if (!isActive) return;\n\n isActive = false;\n\n if (closeSignal) {\n closeSignal();\n closeSignal = null;\n }\n\n if (hooks.onClose) {\n await processingScope(hooks.onClose, closeOptions);\n }\n };\n\n return {\n // TODO: Consider whether not make it optional or add URN prefix\n id: processorId,\n instanceId,\n type,\n canHandle,\n init,\n start: async (\n startOptions: Partial<HandlerContext>,\n ): Promise<CurrentMessageProcessorPosition | undefined> => {\n if (isActive) return;\n\n await init(startOptions);\n\n isActive = true;\n\n closeSignal = onShutdown(() => close(startOptions));\n\n if (lastCheckpoint !== null)\n return {\n lastCheckpoint,\n };\n\n return await processingScope(async (context) => {\n if (hooks.onStart) {\n await hooks.onStart(context);\n }\n\n if (startFrom && startFrom !== 'CURRENT') return startFrom;\n\n if (checkpoints) {\n const readResult = await checkpoints?.read(\n {\n processorId: processorId,\n partition,\n },\n { ...startOptions, ...context },\n );\n lastCheckpoint = readResult.lastCheckpoint;\n }\n\n if (lastCheckpoint === null) return 'BEGINNING';\n\n return {\n lastCheckpoint,\n };\n }, startOptions);\n },\n close,\n get isActive() {\n return isActive;\n },\n handle: async (\n messages: RecordedMessage<MessageType, MessageMetadataType>[],\n partialContext: Partial<HandlerContext>,\n ): Promise<MessageHandlerResult> => {\n if (!isActive) return Promise.resolve();\n\n return await processingScope(async (context) => {\n let result: MessageHandlerResult = undefined;\n\n for (const message of messages) {\n if (wasMessageHandled(message, lastCheckpoint)) continue;\n\n const upcasted = upcastRecordedMessage(\n // TODO: Make it smarter\n message as unknown as RecordedMessage<\n MessagePayloadType,\n MessageMetadataType\n >,\n options.messageOptions?.schema?.versioning,\n );\n\n if (canHandle !== undefined && !canHandle.includes(upcasted.type))\n continue;\n\n const messageProcessingResult = await eachMessage(upcasted, context);\n\n if (checkpoints) {\n const storeCheckpointResult: StoreProcessorCheckpointResult =\n await checkpoints.store(\n {\n processorId,\n version,\n message: upcasted,\n lastCheckpoint,\n partition,\n },\n context,\n );\n\n if (storeCheckpointResult.success) {\n // TODO: Add correct handling of the storing checkpoint\n lastCheckpoint = storeCheckpointResult.newCheckpoint;\n }\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (stopAfter && stopAfter(upcasted)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n }, partialContext);\n },\n };\n};\n\nexport const projector = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n>(\n options: ProjectorOptions<\n EventType,\n EventMetaDataType,\n HandlerContext,\n EventPayloadType\n >,\n): MessageProcessor<EventType, EventMetaDataType, HandlerContext> => {\n const {\n projection,\n processorId = getProjectorId({\n projectionName: projection.name ?? 'unknown',\n }),\n ...rest\n } = options;\n\n return reactor<\n EventType,\n EventMetaDataType,\n HandlerContext,\n EventPayloadType\n >({\n ...rest,\n type: MessageProcessorType.PROJECTOR,\n canHandle: projection.canHandle,\n processorId,\n messageOptions: options.projection.eventsOptions,\n hooks: {\n onInit: options.hooks?.onInit,\n onStart:\n (options.truncateOnStart && options.projection.truncate) ||\n options.hooks?.onStart\n ? async (context: HandlerContext) => {\n if (options.truncateOnStart && options.projection.truncate)\n await options.projection.truncate(context);\n\n if (options.hooks?.onStart) await options.hooks?.onStart(context);\n }\n : undefined,\n onClose: options.hooks?.onClose,\n },\n eachMessage: async (\n event: RecordedMessage<EventType, EventMetaDataType>,\n context: HandlerContext,\n ) => projection.handle([event], context),\n });\n};\n","import { v4 as uuid } from 'uuid';\nimport {\n handleInMemoryProjections,\n type InMemoryProjectionDefinition,\n} from '.';\nimport { bigIntProcessorCheckpoint } from '../../..';\nimport {\n getInMemoryDatabase,\n type Document,\n type InMemoryDatabase,\n} from '../../../database';\nimport { isErrorConstructor } from '../../../errors';\nimport { JSONParser } from '../../../serialization';\nimport {\n assertFails,\n AssertionError,\n assertTrue,\n type ThenThrows,\n} from '../../../testing';\nimport type {\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n} from '../../../typing';\nimport type {\n InMemoryEventStore,\n InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\n// Define a more specific type for T that extends Document\ntype DocumentWithId = Document & { _id?: string | number };\n\nexport type InMemoryProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type InMemoryProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type InMemoryProjectionSpec<EventType extends Event> = (\n givenEvents: InMemoryProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n then: (assert: InMemoryProjectionAssert, message?: string) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type InMemoryProjectionAssert = (options: {\n database: InMemoryDatabase;\n}) => Promise<void | boolean>;\n\nexport type InMemoryProjectionSpecOptions<EventType extends Event> = {\n projection: InMemoryProjectionDefinition<EventType>;\n};\n\nexport const InMemoryProjectionSpec = {\n for: <EventType extends Event>(\n options: InMemoryProjectionSpecOptions<EventType>,\n ): InMemoryProjectionSpec<EventType> => {\n const { projection } = options;\n\n return (givenEvents: InMemoryProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, InMemoryReadEventMetadata>[] =\n [];\n\n const run = async (database: InMemoryDatabase) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: InMemoryReadEventMetadata = {\n checkpoint: bigIntProcessorCheckpoint(++globalPosition),\n globalPosition,\n streamPosition: globalPosition,\n streamName: event.metadata?.streamName ?? `test-${uuid()}`,\n messageId: uuid(),\n };\n\n allEvents.push({\n ...event,\n kind: 'Event',\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n InMemoryReadEventMetadata\n >,\n });\n }\n\n // Create a minimal mock EventStore implementation\n const mockEventStore = {\n database,\n aggregateStream: async () => {\n return Promise.resolve({\n state: {},\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n readStream: async () => {\n return Promise.resolve({\n events: [],\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n appendToStream: async () => {\n return Promise.resolve({\n nextExpectedStreamVersion: 0n,\n createdNewStream: false,\n });\n },\n streamExists: async () => {\n return Promise.resolve(false);\n },\n } as InMemoryEventStore;\n\n await handleInMemoryProjections({\n events: allEvents,\n projections: [projection],\n database,\n eventStore: mockEventStore,\n });\n };\n\n return {\n then: async (\n assertFn: InMemoryProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n await run(database);\n\n const succeeded = await assertFn({ database });\n\n if (succeeded !== undefined && succeeded === false) {\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n }\n },\n thenThrows: async <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n try {\n await run(database);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n },\n};\n\n// Helper functions for creating events in stream\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n>(\n streamName: string,\n event: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>,\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n>(\n streamName: string,\n events: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[],\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\n// Assertion helpers for checking documents\nexport function documentExists<T extends DocumentWithId>(\n expected: Partial<T>,\n options: { inCollection: string; withId: string | number },\n): InMemoryProjectionAssert {\n return async ({ database }) => {\n const collection = database.collection<T>(options.inCollection);\n\n const document = await collection.findOne((doc) => {\n // Handle both string IDs and numeric IDs in a type-safe way\n const docId = '_id' in doc ? doc._id : undefined;\n return docId === options.withId;\n });\n\n if (!document) {\n assertFails(\n `Document with ID ${options.withId} does not exist in collection ${options.inCollection}`,\n );\n return Promise.resolve(false);\n }\n\n // Check that all expected properties exist with expected values\n for (const [key, value] of Object.entries(expected)) {\n const propKey = key as keyof typeof document;\n if (\n !(key in document) ||\n JSONParser.stringify(document[propKey]) !== JSONParser.stringify(value)\n ) {\n assertFails(`Property ${key} doesn't match the expected value`);\n return Promise.resolve(false);\n }\n }\n\n return Promise.resolve(true);\n };\n}\n\n// Helper for checking document contents\nexport const expectInMemoryDocuments = {\n fromCollection: <T extends DocumentWithId>(collectionName: string) => ({\n withId: (id: string | number) => ({\n toBeEqual: (expected: Partial<T>): InMemoryProjectionAssert =>\n documentExists<T>(expected, {\n inCollection: collectionName,\n withId: id,\n }),\n }),\n }),\n};\n","import { v7 as uuid } from 'uuid';\nimport {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type AppendToStreamOptions,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type ReadStreamOptions,\n} from '../eventStore';\nimport type {\n AnyCommand,\n AnyEvent,\n AnyReadEventMetadata,\n Event,\n RecordedMessage,\n} from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\nimport type {\n WorkflowEvent,\n WorkflowInputMessageMetadata,\n WorkflowMessageAction,\n} from './workflow';\nimport type { WorkflowOptions } from './workflowProcessor';\n\nexport const WorkflowHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type WorkflowHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromWorkflowHandlerRetryOptions = (\n retryOptions: WorkflowHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return WorkflowHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...WorkflowHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region workflow-handler\nexport type WorkflowHandlerResult<\n Output extends AnyEvent | AnyCommand,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newMessages: Output[];\n};\n\nexport type WorkflowHandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] & {\n expectedStreamVersion?: ExpectedStreamVersion;\n retry?: WorkflowHandlerRetryOptions;\n};\n\ntype WorkflowInternalState<State> = {\n userState: State;\n processedInputIds: Set<string>;\n};\n\nconst emptyHandlerResult = <\n Output extends AnyEvent | AnyCommand,\n Store extends EventStore,\n>(\n nextExpectedStreamVersion: bigint = 0n,\n): WorkflowHandlerResult<Output, Store> =>\n ({\n newMessages: [] as Output[],\n createdNewStream: false,\n nextExpectedStreamVersion,\n }) as unknown as WorkflowHandlerResult<Output, Store>;\n\nconst createInputMetadata = (\n originalMessageId: string,\n action: Extract<WorkflowMessageAction, 'InitiatedBy' | 'Received'>,\n): WorkflowInputMessageMetadata => ({\n originalMessageId,\n input: true,\n action,\n});\n\nconst tagOutputMessage = <Output extends AnyEvent | AnyCommand>(\n msg: Output,\n action: Extract<WorkflowMessageAction, 'Sent' | 'Published' | 'Scheduled'>,\n): Output => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const existingMetadata =\n 'metadata' in msg && msg.metadata ? msg.metadata : {};\n return {\n ...msg,\n metadata: {\n ...existingMetadata,\n action,\n },\n } as Output;\n};\n\nconst createWrappedInitialState = <State>(initialState: () => State) => {\n return (): WorkflowInternalState<State> => ({\n userState: initialState(),\n processedInputIds: new Set(),\n });\n};\n\nconst createWrappedEvolve = <\n Input extends AnyEvent | AnyCommand,\n Output extends AnyEvent | AnyCommand,\n State,\n>(\n evolve: (state: State, event: WorkflowEvent<Input | Output>) => State,\n workflowName: string,\n separateInputInboxFromProcessing: boolean,\n) => {\n return (\n state: WorkflowInternalState<State>,\n event: WorkflowEvent<Input | Output>,\n ): WorkflowInternalState<State> => {\n const metadata = (event as Record<string, unknown>).metadata as\n | Record<string, unknown>\n | undefined;\n\n // Track processed inputs for idempotency\n let processedInputIds = state.processedInputIds;\n if (\n metadata?.input === true &&\n typeof metadata?.originalMessageId === 'string'\n ) {\n processedInputIds = new Set(state.processedInputIds);\n processedInputIds.add(metadata.originalMessageId);\n }\n\n // In separated inbox mode, don't apply inputs to state - they're just sitting in inbox\n // Only outputs (from processing) should update state\n if (separateInputInboxFromProcessing && metadata?.input === true) {\n return {\n userState: state.userState,\n processedInputIds,\n };\n }\n\n // Strip workflow prefix from input event types\n const eventType = event.type as string;\n const eventForEvolve = eventType.startsWith(`${workflowName}:`)\n ? ({\n ...event,\n type: eventType.replace(`${workflowName}:`, ''),\n } as WorkflowEvent<Input | Output>)\n : event;\n\n return {\n userState: evolve(state.userState, eventForEvolve),\n processedInputIds,\n };\n };\n};\n\nexport const workflowStreamName = ({\n workflowName,\n workflowId,\n}: {\n workflowName: string;\n workflowId: string;\n}) => `emt:workflow:${workflowName}:${workflowId}`;\n\nexport const WorkflowHandler =\n <\n Input extends AnyEvent | AnyCommand,\n State,\n Output extends AnyEvent | AnyCommand,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n StoredMessage extends AnyEvent | AnyCommand = Output,\n >(\n options: WorkflowOptions<\n Input,\n State,\n Output,\n MessageMetadataType,\n StoredMessage\n > & {\n retry?: WorkflowHandlerRetryOptions;\n },\n ) =>\n async <Store extends EventStore>(\n store: Store,\n message: Input | RecordedMessage<Input, MessageMetadataType>,\n handleOptions?: WorkflowHandleOptions<Store>,\n ): Promise<WorkflowHandlerResult<Output, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n WorkflowHandlerResult<Output, Store>\n >(store, async ({ eventStore }) => {\n const {\n workflow: { evolve, initialState, decide, name: workflowName },\n getWorkflowId,\n } = options;\n\n const inputMessageId =\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n ('metadata' in message && message.metadata?.messageId\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n (message.metadata.messageId as string | undefined)\n : undefined) ?? uuid();\n\n const messageWithMetadata: RecordedMessage<\n Input,\n MessageMetadataType\n > = {\n ...message,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n metadata: {\n messageId: inputMessageId,\n ...(message as RecordedMessage<Input, MessageMetadataType>)\n .metadata,\n },\n } as RecordedMessage<Input, MessageMetadataType>;\n\n const workflowId = getWorkflowId(messageWithMetadata);\n\n if (!workflowId) {\n return emptyHandlerResult<Output, Store>();\n }\n\n const streamName = options.mapWorkflowId\n ? options.mapWorkflowId(workflowId)\n : workflowStreamName({ workflowName, workflowId });\n\n const messageType = messageWithMetadata.type as string;\n const hasWorkflowPrefix = messageType.startsWith(`${workflowName}:`);\n\n // Separated inbox mode: store-only path (no prefix = external input)\n if (options.separateInputInboxFromProcessing && !hasWorkflowPrefix) {\n const inputMetadata = createInputMetadata(\n inputMessageId,\n 'InitiatedBy',\n );\n\n const inputToStore = {\n type: `${workflowName}:${messageWithMetadata.type}`,\n data: messageWithMetadata.data,\n kind: messageWithMetadata.kind,\n metadata: inputMetadata,\n } as StoredMessage;\n\n const appendResult = await eventStore.appendToStream(\n streamName,\n [inputToStore] as unknown as Event[],\n {\n ...(handleOptions as AppendToStreamOptions<\n Event,\n StoredMessage & Event\n >),\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n );\n\n return {\n ...appendResult,\n newMessages: [] as Output[],\n } as unknown as WorkflowHandlerResult<Output, Store>;\n }\n\n // Wrap the evolve and initialState for idempotency tracking\n const wrappedInitialState = createWrappedInitialState(initialState);\n const wrappedEvolve = createWrappedEvolve(\n evolve,\n workflowName,\n options.separateInputInboxFromProcessing ?? false,\n ) as (\n state: WorkflowInternalState<State>,\n event: WorkflowEvent<Input | Output>,\n ) => WorkflowInternalState<State>;\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n WorkflowInternalState<State>,\n WorkflowEvent<Input | Output>,\n StoredMessage & Event\n >(streamName, {\n evolve: wrappedEvolve,\n initialState: wrappedInitialState,\n read: {\n ...(handleOptions as ReadStreamOptions<\n WorkflowEvent<Input | Output>,\n StoredMessage & Event\n >),\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const { currentStreamVersion } = aggregationResult;\n\n const { userState: state, processedInputIds } =\n aggregationResult.state;\n\n // Idempotency: skip if this input was already processed\n\n if (processedInputIds.has(inputMessageId)) {\n return emptyHandlerResult<Output, Store>(currentStreamVersion);\n }\n\n // 3. Run business logic\n // Strip workflow prefix from message type if present (for separated inbox processing)\n const messageForDecide = hasWorkflowPrefix\n ? ({\n ...messageWithMetadata,\n type: messageType.replace(`${workflowName}:`, ''),\n } as Input)\n : (messageWithMetadata as Input);\n\n const result = decide(messageForDecide, state);\n\n const inputMetadata = createInputMetadata(\n inputMessageId,\n aggregationResult.streamExists ? 'Received' : 'InitiatedBy',\n );\n\n const inputToStore = {\n type: `${workflowName}:${messageWithMetadata.type}`,\n data: messageWithMetadata.data,\n kind: messageWithMetadata.kind,\n metadata: inputMetadata,\n } as StoredMessage;\n\n const outputMessages = (\n Array.isArray(result) ? result : [result]\n ).filter((msg): msg is Output => msg !== undefined && msg !== null);\n\n const outputCommandTypes = options.outputs?.commands ?? [];\n const taggedOutputMessages = outputMessages.map((msg) => {\n const action: WorkflowMessageAction = outputCommandTypes.includes(\n msg.type as string,\n )\n ? 'Sent'\n : 'Published';\n return tagOutputMessage(msg, action);\n });\n\n const messagesToAppend =\n options.separateInputInboxFromProcessing && hasWorkflowPrefix\n ? [...taggedOutputMessages] // input already in stream\n : [inputToStore, ...taggedOutputMessages]; // normal: store input + outputs\n\n // If there are no messages to append, return early with current state\n if (messagesToAppend.length === 0) {\n return emptyHandlerResult<Output, Store>(currentStreamVersion);\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n\n const expectedStreamVersion: ExpectedStreamVersion =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? currentStreamVersion\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n // TODO: Fix this cast\n messagesToAppend as unknown as Event[],\n {\n ...(handleOptions as AppendToStreamOptions<\n Event,\n StoredMessage & Event\n >),\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with output messages only\n return {\n ...appendResult,\n newMessages: outputMessages,\n } as unknown as WorkflowHandlerResult<Output, Store>;\n });\n\n return result;\n },\n fromWorkflowHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion stream-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type {\n BatchRecordedMessageHandlerWithContext,\n BatchRecordedMessageHandlerWithoutContext,\n DefaultRecord,\n Event,\n ReadEvent,\n} from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\nexport type BeforeEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","export * from './deepReadonly';\n\nexport * from './command';\nexport * from './event';\nexport * from './message';\nexport * from './messageHandling';\n\nexport * from './decider';\n\nexport type Brand<K, T> = K & { readonly __brand: T };\nexport type Flavour<K, T> = K & { readonly __brand?: T };\n\nexport type DefaultRecord = Record<string, unknown>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyRecord = Record<string, any>;\n\nexport type NonNullable<T> = T extends null | undefined ? never : T;\n\nexport const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = `${emmettPrefix}:default`;\nexport const unknownTag = `${emmettPrefix}:unknown`;\n","import { ConcurrencyError, EmmettError } from '../errors';\nimport type { Flavour, StreamPosition } from '../typing';\n\nexport type ExpectedStreamVersion =\n | ExpectedStreamVersionWithValue\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue = Flavour<\n StreamPosition,\n 'StreamVersion'\n>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = (\n current: StreamPosition | undefined,\n expected: ExpectedStreamVersion,\n defaultVersion: StreamPosition,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = (\n current: StreamPosition,\n expected: ExpectedStreamVersion | undefined,\n defaultVersion: StreamPosition,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError extends ConcurrencyError {\n constructor(current: StreamPosition, expected: ExpectedStreamVersion) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError ||\n EmmettError.isInstanceOf<ConcurrencyError>(\n error,\n ExpectedVersionConflictError.Codes.ConcurrencyError,\n );\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","const isPrimitive = (value: unknown): boolean => {\n const type = typeof value;\n return (\n value === null ||\n value === undefined ||\n type === 'boolean' ||\n type === 'number' ||\n type === 'string' ||\n type === 'symbol' ||\n type === 'bigint'\n );\n};\n\nconst compareArrays = <T>(left: T[], right: T[]): boolean => {\n if (left.length !== right.length) {\n return false;\n }\n for (let i = 0; i < left.length; i++) {\n const leftHas = i in left;\n const rightHas = i in right;\n if (leftHas !== rightHas) return false;\n if (leftHas && !deepEquals(left[i], right[i])) return false;\n }\n return true;\n};\n\nconst compareDates = (left: Date, right: Date): boolean => {\n return left.getTime() === right.getTime();\n};\n\nconst compareRegExps = (left: RegExp, right: RegExp): boolean => {\n return left.toString() === right.toString();\n};\n\nconst compareErrors = (left: Error, right: Error): boolean => {\n if (left.message !== right.message || left.name !== right.name) {\n return false;\n }\n const leftKeys = Object.keys(left);\n const rightKeys = Object.keys(right);\n if (leftKeys.length !== rightKeys.length) return false;\n const rightKeySet = new Set(rightKeys);\n for (const key of leftKeys) {\n if (!rightKeySet.has(key)) return false;\n // @ts-expect-error - accessing dynamic keys\n if (!deepEquals(left[key], right[key])) return false;\n }\n return true;\n};\n\nconst compareMaps = (\n left: Map<unknown, unknown>,\n right: Map<unknown, unknown>,\n): boolean => {\n if (left.size !== right.size) return false;\n\n for (const [key, value] of left) {\n if (isPrimitive(key)) {\n if (!right.has(key) || !deepEquals(value, right.get(key))) {\n return false;\n }\n } else {\n let found = false;\n for (const [rightKey, rightValue] of right) {\n if (deepEquals(key, rightKey) && deepEquals(value, rightValue)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareSets = (left: Set<unknown>, right: Set<unknown>): boolean => {\n if (left.size !== right.size) return false;\n\n for (const leftItem of left) {\n if (isPrimitive(leftItem)) {\n if (!right.has(leftItem)) return false;\n } else {\n let found = false;\n for (const rightItem of right) {\n if (deepEquals(leftItem, rightItem)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareArrayBuffers = (\n left: ArrayBuffer,\n right: ArrayBuffer,\n): boolean => {\n if (left.byteLength !== right.byteLength) return false;\n const leftView = new Uint8Array(left);\n const rightView = new Uint8Array(right);\n for (let i = 0; i < leftView.length; i++) {\n if (leftView[i] !== rightView[i]) return false;\n }\n return true;\n};\n\nconst compareTypedArrays = (\n left: ArrayBufferView,\n right: ArrayBufferView,\n): boolean => {\n if (left.constructor !== right.constructor) return false;\n if (left.byteLength !== right.byteLength) return false;\n\n const leftArray = new Uint8Array(\n left.buffer,\n left.byteOffset,\n left.byteLength,\n );\n const rightArray = new Uint8Array(\n right.buffer,\n right.byteOffset,\n right.byteLength,\n );\n\n for (let i = 0; i < leftArray.length; i++) {\n if (leftArray[i] !== rightArray[i]) return false;\n }\n return true;\n};\n\nconst compareObjects = (\n left: Record<string, unknown>,\n right: Record<string, unknown>,\n): boolean => {\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (keys1.length !== keys2.length) {\n return false;\n }\n\n for (const key of keys1) {\n if (left[key] instanceof Function && right[key] instanceof Function) {\n continue;\n }\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nconst getType = (value: unknown): string => {\n if (value === null) return 'null';\n if (value === undefined) return 'undefined';\n\n const primitiveType = typeof value;\n if (primitiveType !== 'object') return primitiveType;\n\n if (Array.isArray(value)) return 'array';\n if (value instanceof Boolean) return 'boxed-boolean';\n if (value instanceof Number) return 'boxed-number';\n if (value instanceof String) return 'boxed-string';\n if (value instanceof Date) return 'date';\n if (value instanceof RegExp) return 'regexp';\n if (value instanceof Error) return 'error';\n if (value instanceof Map) return 'map';\n if (value instanceof Set) return 'set';\n if (value instanceof ArrayBuffer) return 'arraybuffer';\n if (value instanceof DataView) return 'dataview';\n if (value instanceof WeakMap) return 'weakmap';\n if (value instanceof WeakSet) return 'weakset';\n\n if (ArrayBuffer.isView(value)) return 'typedarray';\n\n return 'object';\n};\n\nexport const deepEquals = <T>(left: T, right: T): boolean => {\n if (left === right) return true;\n\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n const leftType = getType(left);\n const rightType = getType(right);\n\n if (leftType !== rightType) return false;\n\n switch (leftType) {\n case 'null':\n case 'undefined':\n case 'boolean':\n case 'number':\n case 'bigint':\n case 'string':\n case 'symbol':\n case 'function':\n return left === right;\n\n case 'array':\n return compareArrays(left as unknown[], right as unknown[]);\n\n case 'date':\n return compareDates(left as Date, right as Date);\n\n case 'regexp':\n return compareRegExps(left as RegExp, right as RegExp);\n\n case 'error':\n return compareErrors(left as Error, right as Error);\n\n case 'map':\n return compareMaps(\n left as Map<unknown, unknown>,\n right as Map<unknown, unknown>,\n );\n\n case 'set':\n return compareSets(left as Set<unknown>, right as Set<unknown>);\n\n case 'arraybuffer':\n return compareArrayBuffers(left as ArrayBuffer, right as ArrayBuffer);\n\n case 'dataview':\n case 'weakmap':\n case 'weakset':\n return false;\n\n case 'typedarray':\n return compareTypedArrays(\n left as ArrayBufferView,\n right as ArrayBufferView,\n );\n\n case 'boxed-boolean':\n return (left as boolean).valueOf() === (right as boolean).valueOf();\n\n case 'boxed-number':\n return (left as number).valueOf() === (right as number).valueOf();\n\n case 'boxed-string':\n return (left as string).valueOf() === (right as string).valueOf();\n\n case 'object':\n return compareObjects(\n left as Record<string, unknown>,\n right as Record<string, unknown>,\n );\n\n default:\n return false;\n }\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left !== null &&\n left !== undefined &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","const textEncoder = new TextEncoder();\n\nexport const hashText = async (text: string): Promise<bigint> => {\n const hashBuffer = await crypto.subtle.digest(\n 'SHA-256',\n textEncoder.encode(text),\n );\n // Create an array with a single element that is a 64-bit signed integer\n // We take the first 8 bytes (so 64 bits) of the SHA-256 hash\n const view = new BigInt64Array(hashBuffer, 0, 1);\n return view[0]!;\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (sub[ele] !== null && typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<void>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport const assertDefined = (\n value: unknown,\n message?: string | Error,\n): asserts value => {\n assertOk(value, message instanceof Error ? message.message : message);\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\n// TODO: replace with assertDefined\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import type {\n AnyMessage,\n AnyRecordedMessageMetadata,\n RecordedMessage,\n} from '../../typing';\n\nexport type MessageDowncast<\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n> =\n | ((\n message: RecordedMessage<MessageType, RecordedMessageMetadataType>,\n ) => RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>)\n | ((message: MessageType) => MessagePayloadType);\n\nexport const downcastRecordedMessage = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessage:\n | RecordedMessage<MessageType, RecordedMessageMetadataType>\n | MessageType,\n options?: {\n downcast?: MessageDowncast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessagePayloadType, RecordedMessageMetadataType> => {\n if (!options?.downcast)\n return recordedMessage as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n\n const downcasted = options.downcast(\n recordedMessage as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >,\n );\n\n return {\n ...recordedMessage,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n data: downcasted.data,\n ...('metadata' in recordedMessage || 'metadata' in downcasted\n ? {\n metadata: {\n ...('metadata' in recordedMessage\n ? (recordedMessage.metadata as object)\n : {}),\n ...('metadata' in downcasted\n ? (downcasted.metadata as object)\n : {}),\n },\n }\n : {}),\n } as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n};\n\nexport const downcastRecordedMessages = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessages:\n | RecordedMessage<MessageType, RecordedMessageMetadataType>[]\n | MessageType[],\n options?: {\n downcast?: MessageDowncast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>[] => {\n if (!options?.downcast)\n return recordedMessages as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >[];\n\n return recordedMessages.map((recordedMessage) =>\n downcastRecordedMessage(recordedMessage, options),\n );\n};\n","import type {\n AnyMessage,\n AnyRecordedMessageMetadata,\n RecordedMessage,\n} from '../../typing';\n\nexport type MessageUpcast<\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n> =\n | ((message: MessagePayloadType) => MessageType)\n | ((\n message: RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>,\n ) => RecordedMessage<MessageType, RecordedMessageMetadataType>);\n\nexport const upcastRecordedMessage = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessage:\n | RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>\n | MessagePayloadType,\n options?: {\n upcast?: MessageUpcast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessageType, RecordedMessageMetadataType> => {\n if (!options?.upcast)\n return recordedMessage as unknown as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >;\n\n const upcasted = options.upcast(\n recordedMessage as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >,\n );\n\n return {\n ...recordedMessage,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n data: upcasted.data,\n ...('metadata' in recordedMessage || 'metadata' in upcasted\n ? {\n metadata: {\n ...('metadata' in recordedMessage\n ? (recordedMessage.metadata as object)\n : {}),\n ...('metadata' in upcasted ? (upcasted.metadata as object) : {}),\n },\n }\n : {}),\n } as unknown as RecordedMessage<MessageType, RecordedMessageMetadataType>;\n};\n\nexport const upcastRecordedMessages = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessages:\n | RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>[]\n | MessagePayloadType[],\n options?: {\n upcast?: MessageUpcast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessageType, RecordedMessageMetadataType>[] => {\n if (!options?.upcast)\n return recordedMessages as unknown as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >[];\n\n return recordedMessages.map((recordedMessage) =>\n upcastRecordedMessage(recordedMessage, options),\n );\n};\n","import { EmmettError } from '../errors';\nimport type { EventStoreReadSchemaOptions } from '../eventStore';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyEvent,\n AnyReadEventMetadata,\n BatchRecordedMessageHandlerWithContext,\n CanHandle,\n DefaultRecord,\n Event,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = BatchRecordedMessageHandlerWithContext<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n>;\n\nexport type TruncateProjection<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: ProjectionHandlerContext) => Promise<void>;\n\nexport type ProjectionInitOptions<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n version: number;\n status?: 'active' | 'inactive';\n registrationType: ProjectionHandlingType;\n context: ProjectionHandlerContext;\n};\n\nexport interface ProjectionDefinition<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n> {\n name?: string;\n version?: number;\n kind?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n truncate?: TruncateProjection<ProjectionHandlerContext>;\n init?: (\n options: ProjectionInitOptions<ProjectionHandlerContext>,\n ) => void | Promise<void>;\n eventsOptions?: {\n schema?: EventStoreReadSchemaOptions<EventType, EventPayloadType>;\n };\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext,\n AnyEvent\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n>(\n definition: ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext,\n EventPayloadType\n >,\n): ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext,\n EventPayloadType\n> => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n any,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","import {\n assertExpectedVersionMatchesCurrent,\n downcastRecordedMessage,\n ExpectedVersionConflictError,\n filterProjections,\n tryPublishMessagesAfterCommit,\n upcastRecordedMessages,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type Closeable,\n type DefaultEventStoreOptions,\n type Event,\n type EventStore,\n type ProjectionRegistration,\n type ReadEvent,\n type ReadEventMetadataWithoutGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n type StreamExistsResult,\n} from '@event-driven-io/emmett';\nimport {\n MongoClient,\n type Collection,\n type Document,\n type Filter,\n type MongoClientOptions,\n type Sort,\n type UpdateFilter,\n type WithId,\n} from 'mongodb';\nimport { v4 as uuid } from 'uuid';\nimport {\n handleInlineProjections,\n MongoDBDefaultInlineProjectionName,\n type MongoDBInlineProjectionDefinition,\n type MongoDBProjectionInlineHandlerContext,\n} from './projections';\nimport {\n mongoDBEventStoreStorage,\n type MongoDBEventStoreStorage,\n type MongoDBEventStoreStorageOptions,\n} from './storage';\n\nexport const MongoDBEventStoreDefaultStreamVersion = 0n;\n\nexport type StreamType = string;\nexport type StreamName<T extends StreamType = StreamType> = `${T}:${string}`;\n\nexport type StreamNameParts<T extends StreamType = StreamType> = {\n streamType: T;\n streamId: string;\n};\n\nexport type StreamCollectionName<T extends StreamType = StreamType> =\n `emt:${T}`;\n\nexport type StreamCollectionNameParts<T extends StreamType = StreamType> = {\n streamType: T;\n};\n\nexport type MongoDBReadModelMetadata = {\n streamId: string;\n name: string;\n schemaVersion: number;\n streamPosition: bigint;\n};\n\nexport type MongoDBReadModel<Doc extends Document = Document> = Doc & {\n _metadata: MongoDBReadModelMetadata;\n};\n\nexport interface EventStream<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> {\n streamName: string;\n messages: Array<ReadEvent<EventType, EventMetaDataType>>;\n metadata: {\n streamId: string;\n streamType: StreamType;\n streamPosition: bigint;\n createdAt: Date;\n updatedAt: Date;\n };\n projections: Record<string, MongoDBReadModel>;\n}\n\nexport type MongoDBReadEventMetadata = ReadEventMetadataWithoutGlobalPosition;\n\nexport type MongoDBReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n MongoDBReadEventMetadata\n>;\n\ntype SingleProjectionQueryStreamFilter<T extends StreamType> = {\n projectionName?: string;\n} & ({ streamName: StreamName<T> } | { streamType: T; streamId?: string });\n\ntype MultiProjectionQueryStreamFilter<T extends StreamType> = {\n projectionName?: string;\n} & (\n | { streamNames: StreamName<T>[] }\n | { streamType: T; streamIds?: string[] }\n);\n\ntype MultiProjectionQueryOptions = {\n skip?: number;\n limit?: number;\n sort?: [string, 1 | -1][] | Record<string, 1 | -1>;\n};\n\n/**\n * Helpers for querying inline projections on event streams.\n */\ntype InlineProjectionQueries<T extends StreamType> = {\n /**\n * Helper for querying for a single projection. Similar to `collection.findOne`.\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamName` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n */\n findOne: <Doc extends Document>(\n streamFilter: SingleProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) => Promise<MongoDBReadModel<Doc> | null>;\n /**\n * Helper for querying for multiple projections. Similar to `collection.find`.\n *\n * ***NOTE***: If `streamFilter.streamNames` is an empty array, this function will return an empty array. If `streamFilter.streamIds` is an empty array, the `streamIds` filter will not be used.\n *\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamNames` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n * @param queryOptions - Additional query options like `skip`, `limit`, and `sort`. `sort`, similar to `projectionQuery`, will prepend each object key with the necessary projection name.\n */\n find: <Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n queryOptions?: MultiProjectionQueryOptions,\n ) => Promise<MongoDBReadModel<Doc>[]>;\n /**\n * Returns the total number of documents matching the provided filter options. Similar to `collection.countDocuments`.\n *\n * ***NOTE***: If `streamFilter.streamNames` is an empty array, this function will return `0`. If `streamFilter.streamIds` is an empty array, the `streamIds` filter will not be used.\n *\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamNames` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n */\n count: <Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) => Promise<number>;\n};\n\n/**\n * Helpers for querying projections on event streams.\n */\ntype ProjectionQueries<T extends StreamType> = {\n inline: InlineProjectionQueries<T>;\n};\n\nexport type MongoDBEventStoreClientOptions = {\n client: MongoClient;\n connectionString?: never;\n clientOptions?: never;\n};\n\nexport type MongoDBEventStoreConnectionStringOptions = {\n client?: never;\n connectionString: string;\n clientOptions?: MongoClientOptions;\n};\n\nexport type MongoDBEventStoreConnectionOptions =\n | MongoDBEventStoreClientOptions\n | MongoDBEventStoreConnectionStringOptions;\n\nexport type MongoDBEventStoreOptions = {\n projections?: ProjectionRegistration<\n 'inline',\n MongoDBReadEventMetadata,\n MongoDBProjectionInlineHandlerContext\n >[];\n storage?: MongoDBEventStoreStorageOptions;\n} & MongoDBEventStoreConnectionOptions &\n DefaultEventStoreOptions<MongoDBEventStore>;\n\nexport type MongoDBEventStore = EventStore<MongoDBReadEventMetadata> & {\n projections: ProjectionQueries<StreamType>;\n collectionFor: <EventType extends Event>(\n streamType: StreamType,\n ) => Promise<Collection<EventStream<EventType>>>;\n};\n\nclass MongoDBEventStoreImplementation implements MongoDBEventStore, Closeable {\n private readonly client: MongoClient;\n private readonly inlineProjections: MongoDBInlineProjectionDefinition[];\n private shouldManageClientLifetime: boolean;\n private isClosed: boolean = false;\n private storage: MongoDBEventStoreStorage;\n private options: MongoDBEventStoreOptions;\n public projections: ProjectionQueries<StreamType>;\n\n constructor(options: MongoDBEventStoreOptions) {\n this.options = options;\n this.client =\n 'client' in options && options.client\n ? options.client\n : new MongoClient(options.connectionString, options.clientOptions);\n this.shouldManageClientLifetime = !('client' in options);\n this.storage = mongoDBEventStoreStorage({\n storage: options.storage,\n getConnectedClient: () => this.getConnectedClient(),\n });\n this.inlineProjections = filterProjections(\n 'inline',\n options.projections ?? [],\n ) as MongoDBInlineProjectionDefinition[];\n\n this.projections = {\n inline: {\n findOne: this.findOneInlineProjection.bind(this),\n find: this.findInlineProjection.bind(this),\n count: this.countInlineProjection.bind(this),\n },\n };\n }\n\n async readStream<\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n options?: ReadStreamOptions<EventType, EventPayloadType>,\n ): Promise<\n Exclude<ReadStreamResult<EventType, MongoDBReadEventMetadata>, null>\n > {\n const { streamType } = fromStreamName(streamName);\n const expectedStreamVersion = options?.expectedStreamVersion;\n\n const collection = await this.storage.collectionFor(streamType);\n\n const filter = {\n streamName: { $eq: streamName },\n };\n\n const eventsSliceArr: number[] = [];\n\n if (options && 'from' in options) {\n eventsSliceArr.push(Number(options.from));\n } else {\n eventsSliceArr.push(0);\n }\n\n if (options && 'to' in options) {\n eventsSliceArr.push(Number(options.to));\n }\n\n const eventsSlice =\n eventsSliceArr.length > 1 ? { $slice: eventsSliceArr } : 1;\n\n const stream = await collection.findOne<\n WithId<Pick<EventStream<EventPayloadType>, 'metadata' | 'messages'>>\n >(filter, {\n useBigInt64: true,\n projection: {\n metadata: 1,\n messages: eventsSlice,\n },\n });\n\n if (!stream) {\n return {\n events: [],\n currentStreamVersion: MongoDBEventStoreDefaultStreamVersion,\n streamExists: false,\n };\n }\n\n assertExpectedVersionMatchesCurrent(\n stream.metadata.streamPosition,\n expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n const events = upcastRecordedMessages(\n stream.messages,\n options?.schema?.versioning,\n );\n\n return {\n events,\n currentStreamVersion: stream.metadata.streamPosition,\n streamExists: true,\n };\n }\n\n async aggregateStream<\n State,\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n options: AggregateStreamOptions<\n State,\n EventType,\n MongoDBReadEventMetadata,\n EventPayloadType\n >,\n ): Promise<AggregateStreamResult<State>> {\n const stream = await this.readStream<EventType, EventPayloadType>(\n streamName,\n options?.read,\n );\n const { evolve, initialState } = options;\n\n const state = stream.events.reduce(evolve, initialState());\n return {\n state,\n currentStreamVersion: stream.currentStreamVersion,\n streamExists: stream.streamExists,\n };\n }\n\n async appendToStream<\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n events: EventType[],\n options?: AppendToStreamOptions<EventType, EventPayloadType>,\n ): Promise<AppendToStreamResult> {\n const { streamId, streamType } = fromStreamName(streamName);\n const expectedStreamVersion = options?.expectedStreamVersion;\n\n const collection = await this.storage.collectionFor(streamType);\n\n const stream = await collection.findOne<\n WithId<Pick<EventStream<EventPayloadType>, 'metadata' | 'projections'>>\n >(\n { streamName: { $eq: streamName } },\n {\n useBigInt64: true,\n projection: {\n 'metadata.streamPosition': 1,\n projections: 1,\n },\n },\n );\n\n const currentStreamVersion =\n stream?.metadata.streamPosition ?? MongoDBEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n let streamOffset = currentStreamVersion;\n\n const eventsToAppend: ReadEvent<\n EventPayloadType,\n MongoDBReadEventMetadata\n >[] = events.map((event) => {\n const metadata: MongoDBReadEventMetadata = {\n messageId: uuid(),\n streamName,\n streamPosition: ++streamOffset,\n };\n return downcastRecordedMessage(\n {\n type: event.type,\n data: event.data,\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n },\n } as ReadEvent<EventType, MongoDBReadEventMetadata>,\n options?.schema?.versioning,\n );\n });\n\n const now = new Date();\n const updates: UpdateFilter<EventStream> = {\n $push: { messages: { $each: eventsToAppend } },\n $set: {\n 'metadata.updatedAt': now,\n 'metadata.streamPosition': currentStreamVersion + BigInt(events.length),\n },\n $setOnInsert: {\n streamName,\n 'metadata.streamId': streamId,\n 'metadata.streamType': streamType,\n 'metadata.createdAt': now,\n },\n };\n\n if (this.inlineProjections) {\n await handleInlineProjections({\n readModels: stream?.projections ?? {},\n streamId,\n events: eventsToAppend,\n projections: this.inlineProjections,\n collection,\n updates,\n client: {},\n });\n }\n\n const updatedStream = await collection.updateOne(\n {\n streamName: { $eq: streamName },\n 'metadata.streamPosition': currentStreamVersion,\n },\n updates,\n { useBigInt64: true, upsert: true },\n );\n\n if (!updatedStream) {\n throw new ExpectedVersionConflictError(\n currentStreamVersion,\n options?.expectedStreamVersion ?? 0n,\n );\n }\n\n await tryPublishMessagesAfterCommit<MongoDBEventStore>(\n eventsToAppend,\n this.options.hooks,\n // {\n // TODO: same context as InlineProjectionHandlerContext for mongodb?\n // },\n );\n\n return {\n nextExpectedStreamVersion:\n currentStreamVersion + BigInt(eventsToAppend.length),\n createdNewStream:\n currentStreamVersion === MongoDBEventStoreDefaultStreamVersion,\n };\n }\n\n async streamExists(streamName: StreamName): Promise<StreamExistsResult> {\n const { streamType } = fromStreamName(streamName);\n\n const collection = await this.storage.collectionFor(streamType);\n\n const filter = {\n streamName: { $eq: streamName },\n };\n\n const count = await collection.countDocuments(filter, {\n useBigInt64: true,\n limit: 1,\n });\n\n return Boolean(count > 0);\n }\n\n collectionFor = async <EventType extends Event>(\n streamType: StreamType,\n ): Promise<Collection<EventStream<EventType>>> => {\n return this.storage.collectionFor(streamType);\n };\n\n /**\n * Gracefully cleans up managed resources by the MongoDBEventStore.\n * It closes MongoDB client created for the provided connection string\n * through event store options.\n *\n * @memberof Closeable\n */\n close = (): Promise<void> => {\n if (this.isClosed) return Promise.resolve();\n\n this.isClosed = true;\n if (!this.shouldManageClientLifetime) return Promise.resolve();\n\n return this.client.close();\n };\n\n private async findOneInlineProjection<Doc extends Document>(\n streamFilter: SingleProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) {\n const { projectionName, streamName, streamType } =\n parseSingleProjectionQueryStreamFilter(streamFilter);\n const collection = await this.storage.collectionFor(streamType);\n const query = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, `projections.${projectionName}`);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $exists: true } },\n ];\n\n if (query) {\n filters.push(query);\n }\n\n if (streamName) {\n filters.push({ streamName: { $eq: streamName } });\n }\n\n const result = await collection.findOne<{\n projections: Record<typeof projectionName, MongoDBReadModel<Doc>>;\n }>(\n { $and: filters },\n {\n useBigInt64: true,\n projection: { [`projections.${projectionName}`]: 1 },\n },\n );\n\n return result?.projections?.[projectionName] ?? null;\n }\n\n private async findInlineProjection<Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n queryOptions?: MultiProjectionQueryOptions,\n ) {\n const parsedStreamFilter =\n parseMultiProjectionQueryStreamFilter(streamFilter);\n if (!parsedStreamFilter) return [];\n const { projectionName, streamNames, streamType } = parsedStreamFilter;\n\n const collection = await this.storage.collectionFor(streamType);\n const prefix = `projections.${projectionName}`;\n const projectionFilter = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, prefix);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $ne: null } },\n ];\n\n if (projectionFilter) {\n filters.push(projectionFilter);\n }\n\n if (streamNames) {\n filters.push({ streamName: { $in: streamNames } });\n }\n\n let query = collection.find<\n EventStream & {\n projections: Record<typeof projectionName, MongoDBReadModel<Doc>>;\n }\n >(\n { $and: filters },\n {\n useBigInt64: true,\n projection: { [`projections.${projectionName}`]: 1 },\n },\n );\n\n if (queryOptions?.skip) {\n query = query.skip(queryOptions.skip);\n }\n\n if (queryOptions?.limit) {\n query = query.limit(queryOptions.limit);\n }\n\n if (queryOptions?.sort) {\n const sort = prependMongoFilterWithProjectionPrefix<Sort>(\n queryOptions.sort,\n prefix,\n );\n query = query.sort(sort);\n }\n\n const streams = await query.toArray();\n\n return streams\n .map((s) => s.projections[projectionName])\n .filter((p): p is MongoDBReadModel<Doc> => !!p);\n }\n\n private async countInlineProjection<Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) {\n const parsedStreamFilter =\n parseMultiProjectionQueryStreamFilter(streamFilter);\n if (!parsedStreamFilter) return 0;\n const { projectionName, streamNames, streamType } = parsedStreamFilter;\n\n const collection = await this.storage.collectionFor(streamType);\n const prefix = `projections.${projectionName}`;\n const projectionFilter = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, prefix);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $ne: null } },\n ];\n\n if (projectionFilter) {\n filters.push(projectionFilter);\n }\n\n if (streamNames) {\n filters.push({ streamName: { $in: streamNames } });\n }\n\n const total = await collection.countDocuments({ $and: filters });\n return total;\n }\n\n private getConnectedClient = async (): Promise<MongoClient> => {\n if (!this.isClosed) await this.client.connect();\n return this.client;\n };\n}\n\nfunction parseSingleProjectionQueryStreamFilter<\n T extends StreamType = StreamType,\n>(streamFilter: SingleProjectionQueryStreamFilter<T>) {\n const projectionName =\n streamFilter.projectionName ?? MongoDBDefaultInlineProjectionName;\n\n if ('streamName' in streamFilter) {\n const { streamType } = fromStreamName(streamFilter.streamName);\n return {\n projectionName,\n streamName: streamFilter.streamName,\n streamType,\n };\n }\n\n if (streamFilter.streamId) {\n const streamName = toStreamName(\n streamFilter.streamType,\n streamFilter.streamId,\n );\n return {\n projectionName,\n streamName,\n streamType: streamFilter.streamType,\n };\n }\n\n return {\n projectionName,\n streamType: streamFilter.streamType,\n };\n}\n\nfunction parseMultiProjectionQueryStreamFilter<T extends StreamType>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n) {\n const projectionName =\n streamFilter.projectionName ?? MongoDBDefaultInlineProjectionName;\n\n if ('streamNames' in streamFilter) {\n if (streamFilter.streamNames.length == 0) return null;\n const { streamType } = fromStreamName(streamFilter.streamNames[0]!);\n return {\n projectionName,\n streamNames: streamFilter.streamNames,\n streamType,\n };\n }\n\n if (streamFilter.streamIds && streamFilter.streamIds.length > 0) {\n const streamNames = streamFilter.streamIds.map((id) =>\n toStreamName(streamFilter.streamType, id),\n );\n return {\n projectionName,\n streamNames,\n streamType: streamFilter.streamType,\n };\n }\n\n return {\n projectionName,\n streamType: streamFilter.streamType,\n };\n}\n\n/**\n * Prepends `prefix` to all object keys that don't start with a '$'\n */\nexport function prependMongoFilterWithProjectionPrefix<T, Result = T>(\n obj: T,\n prefix: string,\n): Result {\n if (typeof obj !== 'object' || obj === null || obj === undefined) {\n return obj as unknown as Result;\n }\n\n if (Array.isArray(obj)) {\n for (let i = 0; i < obj.length; i++) {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n obj[i] = prependMongoFilterWithProjectionPrefix(obj[i], prefix);\n }\n return obj as unknown as Result;\n }\n\n for (const key in obj) {\n // @ts-expect-error we're forcing `k` to be a key of `T`\n const k: keyof typeof obj = addProjectionPrefixToMongoKey(key, prefix);\n if (k !== key) {\n obj[k] = obj[key as keyof typeof obj];\n delete obj[key as keyof typeof obj];\n }\n\n obj[k] = prependMongoFilterWithProjectionPrefix(obj[k], prefix);\n }\n\n return obj as unknown as Result;\n}\n\nfunction addProjectionPrefixToMongoKey(key: string, prefix: string): string {\n // MongoDB operators\n if (key[0] === '$') {\n return key;\n }\n\n return `${prefix}${key.length > 0 ? '.' : ''}${key}`;\n}\n\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions & { client: MongoClient },\n): MongoDBEventStore;\n\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions & { connectionString: string },\n): MongoDBEventStore & Closeable;\n\n// Implementation signature covers both, using a union for `options`\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions,\n): MongoDBEventStore | Closeable {\n const impl = new MongoDBEventStoreImplementation(options);\n\n // If a client is provided externally, we don't want to allow closing it\n if ('client' in options && 'close' in impl) {\n delete (impl as Partial<MongoDBEventStoreImplementation>).close;\n }\n\n return impl;\n}\n\n/**\n * Accepts a `streamType` (the type/category of the event stream) and an `streamId`\n * (the individual entity/object or aggregate ID) and combines them to a singular\n * `streamName` which can be used in `EventStore`.\n */\nexport function toStreamName<T extends StreamType>(\n streamType: T,\n streamId: string,\n): StreamName<T> {\n return `${streamType}:${streamId}`;\n}\n\n/**\n * Accepts a fully formatted `streamName` and returns the broken down\n * `streamType` and `streamId`.\n */\nexport function fromStreamName<T extends StreamType>(\n streamName: StreamName<T>,\n): StreamNameParts<T> {\n const parts = streamName.split(':') as [T, string];\n return {\n streamType: parts[0],\n streamId: parts[1],\n };\n}\n\n/**\n * Accepts a `streamType` (the type/category of the event stream)\n * and combines them to a `collectionName` which can be used in `EventStore`.\n */\nexport function toStreamCollectionName<T extends StreamType>(\n streamType: T,\n): StreamCollectionName<T> {\n return `emt:${streamType}`;\n}\n\n/**\n * Accepts a fully formatted `streamCollectionName` and returns the parsed `streamType`.\n */\nexport function fromStreamCollectionName<T extends StreamType>(\n streamCollectionName: StreamCollectionName<T>,\n): StreamCollectionNameParts<T> {\n const parts = streamCollectionName.split(':') as [string, T];\n return {\n streamType: parts[1],\n };\n}\n","import type {\n CanHandle,\n Event,\n ProjectionDefinition,\n ProjectionHandler,\n ReadEvent,\n} from '@event-driven-io/emmett';\nimport type { Collection, Document, UpdateFilter } from 'mongodb';\nimport type {\n EventStream,\n MongoDBReadEventMetadata,\n MongoDBReadModel,\n MongoDBReadModelMetadata,\n} from '../mongoDBEventStore';\n\nexport const MongoDBDefaultInlineProjectionName = '_default';\n\nexport type MongoDBProjectionInlineHandlerContext<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n document: MongoDBReadModel | null;\n streamId: string;\n updates: UpdateFilter<EventStream<EventType, EventMetaDataType>>;\n collection: Collection<EventStream<EventType, EventMetaDataType>>;\n};\n\nexport type MongoDBInlineProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n MongoDBProjectionInlineHandlerContext\n>;\n\nexport type MongoDBInlineProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = ProjectionDefinition<\n EventType,\n EventMetaDataType,\n MongoDBProjectionInlineHandlerContext\n> & { name: string };\n\nexport type InlineProjectionHandlerOptions<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n readModels: Record<string, MongoDBReadModel>;\n events: Array<ReadEvent<EventType, EventMetaDataType>>;\n projections: MongoDBInlineProjectionDefinition<\n EventType,\n EventMetaDataType\n >[];\n streamId: string;\n collection: Collection<EventStream>;\n updates: UpdateFilter<EventStream<Event>>;\n // eslint-disable-next-line @typescript-eslint/no-empty-object-type\n client: {\n //todo: add client here\n };\n};\n\nexport const handleInlineProjections = async <\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n>(\n options: InlineProjectionHandlerOptions<EventType, EventMetaDataType>,\n): Promise<void> => {\n const {\n events,\n projections: allProjections,\n updates: update,\n streamId,\n collection,\n readModels,\n } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n for (const projection of projections) {\n await projection.handle(events, {\n document: readModels[projection.name] ?? null,\n streamId,\n collection,\n updates: update,\n });\n }\n};\n\nexport type MongoDBWithNotNullDocumentEvolve<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> =\n | ((\n document: Doc,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Doc | null)\n | ((document: Doc, event: ReadEvent<EventType>) => Promise<Doc | null>);\n\nexport type MongoDBWithNullableDocumentEvolve<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> =\n | ((\n document: Doc | null,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Doc | null)\n | ((\n document: Doc | null,\n event: ReadEvent<EventType>,\n ) => Promise<Doc | null>);\n\nexport type MongoDBInlineProjectionOptions<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n name?: string;\n schemaVersion?: number;\n canHandle: CanHandle<EventType>;\n} & (\n | {\n evolve: MongoDBWithNullableDocumentEvolve<\n Doc,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: MongoDBWithNotNullDocumentEvolve<\n Doc,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Doc;\n }\n);\n\nexport const mongoDBInlineProjection = <\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n>(\n options: MongoDBInlineProjectionOptions<Doc, EventType, EventMetaDataType>,\n): MongoDBInlineProjectionDefinition => {\n const projectionName = options.name ?? MongoDBDefaultInlineProjectionName;\n const schemaVersion = options.schemaVersion ?? 1;\n\n return {\n name: projectionName,\n canHandle: options.canHandle,\n handle: async (events, { document, updates, streamId }) => {\n if (events.length === 0) return;\n\n let state =\n 'initialState' in options\n ? (document ?? options.initialState())\n : document;\n\n for (const event of events) {\n state = await options.evolve(\n state as Doc,\n event as ReadEvent<EventType, EventMetaDataType>,\n );\n }\n\n const metadata: MongoDBReadModelMetadata = {\n streamId,\n name: projectionName,\n schemaVersion,\n streamPosition: events[events.length - 1]!.metadata.streamPosition,\n };\n\n updates.$set![`projections.${projectionName}`] =\n state !== null\n ? {\n ...state,\n _metadata: metadata,\n }\n : null;\n },\n };\n};\n","import {\n assertFails,\n AssertionError,\n assertTrue,\n deepEquals,\n isErrorConstructor,\n isSubset,\n projections,\n type Event,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { MongoClient, type Document } from 'mongodb';\nimport {\n getMongoDBEventStore,\n type MongoDBEventStore,\n type MongoDBEventStoreConnectionOptions,\n type MongoDBReadModel,\n type StreamName,\n} from '../mongoDBEventStore';\nimport {\n MongoDBDefaultInlineProjectionName,\n type MongoDBInlineProjectionDefinition,\n} from './mongoDBInlineProjection';\n\nexport type MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType extends StreamName,\n EventType extends Event,\n> = {\n streamName: StreamNameType;\n events: EventType[];\n};\n\nexport type MongoDBInlineProjectionSpec<\n StreamNameType extends StreamName,\n EventType extends Event,\n> = (\n givenStream: MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType,\n EventType\n >,\n) => {\n when: (events: EventType[]) => {\n then: (\n assert: MongoDBInlineProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type MongoDBInlineProjectionAssertOptions<\n StreamNameType extends StreamName = StreamName,\n> = {\n streamName: StreamNameType;\n eventStore: MongoDBEventStore;\n};\n\nexport type MongoDBInlineProjectionAssert<\n StreamNameType extends StreamName = StreamName,\n> = (\n options: MongoDBInlineProjectionAssertOptions<StreamNameType>,\n) => Promise<void | boolean>;\n\nexport type MongoDBInlineProjectionSpecOptions = {\n projection: MongoDBInlineProjectionDefinition;\n} & MongoDBEventStoreConnectionOptions;\n\nexport const MongoDBInlineProjectionSpec = {\n for: <StreamNameType extends StreamName, EventType extends Event>(\n options: MongoDBInlineProjectionSpecOptions,\n ): MongoDBInlineProjectionSpec<StreamNameType, EventType> => {\n {\n const { projection, ...connectionOptions } = options;\n\n return (\n givenStream: MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType,\n EventType\n >,\n ) => {\n const { streamName, events: givenEvents } = givenStream;\n return {\n when: (events: EventType[]) => {\n const allEvents = [...givenEvents, ...events];\n\n const run = (eventStore: MongoDBEventStore) =>\n eventStore.appendToStream(streamName, allEvents);\n\n return {\n then: async (\n assert: MongoDBInlineProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const client =\n 'client' in connectionOptions && connectionOptions.client\n ? connectionOptions.client\n : new MongoClient(\n connectionOptions.connectionString,\n connectionOptions.clientOptions,\n );\n\n const eventStore = getMongoDBEventStore({\n projections: projections.inline([projection]),\n client,\n });\n\n try {\n await run(eventStore);\n\n const succeeded = await assert({ eventStore, streamName });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await client.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const client =\n 'client' in connectionOptions && connectionOptions.client\n ? connectionOptions.client\n : new MongoClient(\n connectionOptions.connectionString,\n connectionOptions.clientOptions,\n );\n\n const eventStore = getMongoDBEventStore({\n projections: projections.inline([projection]),\n client,\n });\n\n try {\n await run(eventStore);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await client.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const eventInStream = <\n StreamNameType extends StreamName,\n EventType extends Event,\n>(\n streamName: StreamNameType,\n event: EventType,\n): MongoDBInlineProjectionSpecGivenEvents<StreamNameType, EventType> => ({\n streamName,\n events: [event],\n});\n\nexport const eventsInStream = <\n StreamNameType extends StreamName,\n EventType extends Event,\n>(\n streamName: StreamNameType,\n events: EventType[],\n): MongoDBInlineProjectionSpecGivenEvents<StreamNameType, EventType> => ({\n streamName,\n events,\n});\n\nconst expectReadModelToMatch = async <\n Doc extends Document = Document,\n StreamNameType extends StreamName = StreamName,\n>(\n options: MongoDBInlineProjectionAssertOptions<StreamNameType> & {\n projectionName: string;\n match: (readModel: MongoDBReadModel<Doc> | null) => boolean;\n },\n) => {\n const { streamName, projectionName, eventStore, match } = options;\n const readModel = await eventStore.projections.inline.findOne<Doc>({\n streamName,\n projectionName,\n });\n\n return match(readModel);\n};\n\nconst expectInlineReadModelWithName = (projectionName: string) => ({\n toHave:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n expected: Partial<MongoDBReadModel<Doc>> | null,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => isSubset(readModel, expected),\n }),\n toDeepEquals:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n expected: MongoDBReadModel<Doc> | null,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => deepEquals(readModel, expected),\n }),\n toMatch:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n match: (readModel: MongoDBReadModel<Doc> | null) => boolean,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match,\n }),\n notToExist:\n <\n StreamNameType extends StreamName = StreamName,\n >(): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => readModel === null,\n }),\n toExist:\n (): MongoDBInlineProjectionAssert =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => readModel !== null,\n }),\n});\n\nexport const expectInlineReadModel = {\n withName: (name: string) => expectInlineReadModelWithName(name),\n ...expectInlineReadModelWithName(MongoDBDefaultInlineProjectionName),\n};\n","import type { Event } from '@event-driven-io/emmett';\nimport type { Collection, Db, MongoClient } from 'mongodb';\nimport {\n toStreamCollectionName,\n type EventStream,\n type StreamType,\n} from '../mongoDBEventStore';\n\nexport type MongoDBEventStoreCollectionPerStreamTypeStorageOptions = {\n /**\n * The recommended setting where each stream type will be kept\n * in a separate collection type using the format: `emt_${streamType}`.\n */\n type: 'COLLECTION_PER_STREAM_TYPE';\n databaseName?: string;\n};\n\nexport type MongoDBEventStoreSingleCollectionStorageOptions = {\n /**\n * All streams will be kept withing a single MongDB collection\n * It'll either use default collection name (\"emt_streams\")\n * or provided name through 'collection' param.\n */\n type: 'SINGLE_COLLECTION';\n collectionName?: string;\n databaseName?: string;\n};\n\nexport type MongoDBEventStoreCollectionResolution = {\n databaseName?: string;\n collectionName: string;\n};\n\nexport type MongoDBEventStoreCustomStorageOptions = {\n /**\n * This is advanced option, where you specify your own collection\n * resolution function. You can do that by specifying the `collectionFor` function.\n */\n type: 'CUSTOM';\n databaseName?: string;\n collectionFor: <T extends StreamType>(\n streamType: T,\n ) => string | MongoDBEventStoreCollectionResolution;\n};\n\nexport type MongoDBEventStoreStorageOptions =\n | 'COLLECTION_PER_STREAM_TYPE'\n | 'SINGLE_COLLECTION'\n | MongoDBEventStoreSingleCollectionStorageOptions\n | MongoDBEventStoreCollectionPerStreamTypeStorageOptions\n | MongoDBEventStoreCustomStorageOptions;\n\nexport const DefaultMongoDBEventStoreStorageOptions =\n 'COLLECTION_PER_STREAM_TYPE';\n\nexport type MongoDBEventStoreStorage = {\n collectionFor: <T extends StreamType, EventType extends Event = Event>(\n streamType: T,\n ) => Promise<Collection<EventStream<EventType>>>;\n};\n\nexport const DefaultMongoDBEventStoreCollectionName = 'emt:streams';\n\nconst resolveCollectionAndDatabase = <T extends StreamType>(\n streamType: T,\n options: MongoDBEventStoreStorageOptions,\n): MongoDBEventStoreCollectionResolution => {\n if (\n options === 'SINGLE_COLLECTION' ||\n (typeof options === 'object' && options.type === 'SINGLE_COLLECTION')\n ) {\n return {\n collectionName:\n typeof options === 'object'\n ? (options.collectionName ?? DefaultMongoDBEventStoreCollectionName)\n : DefaultMongoDBEventStoreCollectionName,\n databaseName:\n typeof options === 'object' ? options.databaseName : undefined,\n };\n } else if (\n options === 'COLLECTION_PER_STREAM_TYPE' ||\n (typeof options === 'object' &&\n options.type === 'COLLECTION_PER_STREAM_TYPE')\n ) {\n return {\n collectionName: toStreamCollectionName(streamType),\n databaseName:\n typeof options === 'object' ? options.databaseName : undefined,\n };\n } else {\n const result = options.collectionFor(streamType);\n return {\n collectionName:\n typeof result === 'object' ? result.collectionName : result,\n databaseName:\n typeof result === 'object'\n ? (result.databaseName ?? options.databaseName)\n : options.databaseName,\n };\n }\n};\n\nconst getDB = async (options: {\n databaseName: string | undefined;\n dbsCache: Map<string, Db>;\n getConnectedClient: () => Promise<MongoClient>;\n}): Promise<Db> => {\n const { dbsCache, databaseName, getConnectedClient } = options;\n const safeDbName = databaseName ?? '___default';\n\n let db = dbsCache.get(safeDbName);\n\n if (!db) {\n const connectedClient = await getConnectedClient();\n\n db = connectedClient.db(databaseName);\n\n dbsCache.set(safeDbName, db);\n }\n\n return db;\n};\n\nconst collectionFor = async <EventType extends Event = Event>(options: {\n collectionName: string;\n streamCollections: Map<string, Collection<EventStream>>;\n db: Db;\n}): Promise<Collection<EventStream<EventType>>> => {\n const { collectionName, db, streamCollections } = options;\n\n let collection = streamCollections.get(collectionName) as\n | Collection<EventStream<EventType>>\n | undefined;\n\n if (!collection) {\n collection = db.collection<EventStream<EventType>>(collectionName);\n await collection.createIndex({ streamName: 1 }, { unique: true });\n\n streamCollections.set(\n collectionName,\n collection as Collection<EventStream>,\n );\n }\n\n return collection;\n};\n\nexport const mongoDBEventStoreStorage = (options: {\n storage?: MongoDBEventStoreStorageOptions | undefined;\n getConnectedClient: () => Promise<MongoClient>;\n}): MongoDBEventStoreStorage => {\n const dbsCache: Map<string, Db> = new Map();\n const streamCollections: Map<string, Collection<EventStream>> = new Map();\n const storageOptions =\n options.storage ?? DefaultMongoDBEventStoreStorageOptions;\n\n const { getConnectedClient } = options;\n\n return {\n collectionFor: async <\n T extends StreamType,\n EventType extends Event = Event,\n >(\n streamType: T,\n ): Promise<Collection<EventStream<EventType>>> => {\n const { collectionName, databaseName } = resolveCollectionAndDatabase(\n streamType,\n storageOptions,\n );\n\n let collection = streamCollections.get(collectionName) as\n | Collection<EventStream<EventType>>\n | undefined;\n\n if (!collection) {\n const db = await getDB({ databaseName, dbsCache, getConnectedClient });\n collection = await collectionFor<EventType>({\n collectionName,\n streamCollections,\n db,\n });\n }\n\n return collection;\n },\n };\n};\n"]}
|
|
1
|
+
{"version":3,"sources":["/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","../../emmett/src/validation/index.ts","../../emmett/src/errors/index.ts","../../emmett/src/eventStore/inMemoryEventStore.ts","../../emmett/src/database/inMemoryDatabase.ts","../../emmett/src/taskProcessing/executionGuards.ts","../../emmett/src/utils/retry.ts","../../emmett/src/processors/processors.ts","../../emmett/src/eventStore/projections/inMemory/inMemoryProjectionSpec.ts","../../emmett/src/workflows/handleWorkflow.ts","../../emmett/src/eventStore/afterCommit/afterEventStoreCommitHandler.ts","../../emmett/src/typing/index.ts","../../emmett/src/eventStore/expectedVersion.ts","../../emmett/src/utils/collections/duplicates.ts","../../emmett/src/utils/collections/merge.ts","../../emmett/src/utils/collections/index.ts","../../emmett/src/utils/deepEquals.ts","../../emmett/src/serialization/json/JSONParser.ts","../../emmett/src/utils/strings/hashText.ts","../../emmett/src/testing/assertions.ts","../../emmett/src/eventStore/versioning/downcasting.ts","../../emmett/src/eventStore/versioning/upcasting.ts","../../emmett/src/projections/index.ts","../src/eventStore/mongoDBEventStore.ts","../src/eventStore/projections/mongoDBInlineProjection.ts","../src/eventStore/projections/mongoDBInlineProjectionSpec.ts","../src/eventStore/storage/mongoDBEventStoreStorage.ts"],"names":[],"mappings":"AAAA;ACQO,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,SAAA,GAAY,IAAA,IAAQ,GAAA;AAM9B,IAAM,SAAA,EAAW,CAAC,GAAA,EAAA,GACvB,OAAO,IAAA,IAAQ,QAAA;ACTV,IAAM,mBAAA,EAAqB,CAEhC,MAAA,EAAA,GAC0C;AAE1C,EAAA,OACE,OAAO,OAAA,IAAW,WAAA,GAClB,MAAA,CAAO,UAAA;AFTX,EEWI,MAAA,CAAO,SAAA,CAAU,YAAA,IAAgB,MAAA;AAErC,CAAA;AAEO,IAAM,YAAA,YAAN,MAAM,aAAA,QAAoB,MAAM;AFZvC,EEaE,4BAAuB,MAAA,EAAQ;AFZjC,IEaI,eAAA,EAAiB,GAAA;AFZrB,IEaI,iBAAA,EAAmB,GAAA;AFZvB,IEaI,aAAA,EAAe,GAAA;AFZnB,IEaI,gBAAA,EAAkB,GAAA;AFZtB,IEaI,mBAAA,EAAqB;AFZzB,EEaE,EAAA;AFZF,EEcS;AFbT,EEeE,WAAA,CACE,OAAA,EACA;AACA,IAAA,MAAM,UAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,YAAA,GAAe,QAAA,EACrD,OAAA,CAAQ,UAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,YAAA,CAAY,KAAA,CAAM,mBAAA;AAC1B,IAAA,MAAM,QAAA,EACJ,QAAA,GAAW,OAAO,QAAA,IAAY,SAAA,GAAY,UAAA,GAAa,QAAA,EACnD,OAAA,CAAQ,QAAA,EACR,QAAA,CAAS,OAAO,EAAA,EACd,QAAA,EACA,CAAA,wBAAA,EAA2B,SAAS,CAAA,kCAAA,CAAA;AAE5C,IAAA,KAAA,CAAM,OAAO,CAAA;AACb,IAAA,IAAA,CAAK,UAAA,EAAY,SAAA;AAGjB,IAAA,MAAA,CAAO,cAAA,CAAe,IAAA,EAAM,YAAA,CAAY,SAAS,CAAA;AF7BrD,EE8BE;AF7BF,EE+BE,OAAc,OAAA,CACZ,KAAA,EACa;AACb,IAAA,GAAA,CAAI,YAAA,CAAY,YAAA,CAAa,KAAK,CAAA,EAAG;AACnC,MAAA,OAAO,KAAA;AFhCb,IEiCI;AAEA,IAAA,OAAO,IAAI,YAAA,CAAY;AFjC3B,MEkCM,SAAA,EACE,YAAA,GAAe,MAAA,GACf,KAAA,CAAM,UAAA,IAAc,KAAA,EAAA,GACpB,KAAA,CAAM,UAAA,IAAc,KAAA,EAChB,KAAA,CAAM,UAAA,EACN,YAAA,CAAY,KAAA,CAAM,mBAAA;AFtC9B,MEuCM,OAAA,mBAAS,KAAA,CAAM,OAAA,UAAW;AFtChC,IEuCI,CAAC,CAAA;AFtCL,EEuCE;AFtCF,EEwCE,OAAc,YAAA,CACZ,KAAA,EACA,SAAA,EACoB;AACpB,IAAA,OACE,OAAO,MAAA,IAAU,SAAA,GACjB,MAAA,IAAU,KAAA,GACV,YAAA,GAAe,MAAA,GACf,QAAA,CAAS,KAAA,CAAM,SAAS,EAAA,GAAA,CACvB,UAAA,IAAc,KAAA,EAAA,GAAa,KAAA,CAAM,UAAA,IAAc,SAAA,CAAA;AF/CtD,EEiDE;AACF,iCAAA;AAEO,IAAM,iBAAA,EAAN,MAAM,kBAAA,QAAyB,YAAY;AFjDlD,EEkDE,WAAA,CACS,OAAA,EACA,QAAA,EACP,OAAA,EACA;AACA,IAAA,KAAA,CAAM;AFrDV,MEsDM,SAAA,EAAW,WAAA,CAAY,KAAA,CAAM,gBAAA;AFrDnC,MEsDM,OAAA,mBACE,OAAA,UACA,CAAA,iBAAA,EAAoB,QAAA,CAAS,QAAA,CAAS,CAAC,CAAA,wBAAA,kBAA2B,OAAA,6BAAS,QAAA,mBAAS,GAAC,CAAA;AFvD7F,IAAA;AE+CW,IAAA;AACA,IAAA;AAWP,IAAA;AFvDJ,EAAA;AEyDA;AFvDA;AACA;AGhDA;ACAA;ACAA;ACAA;ACAA;ACAA;ACAA;AC0DA;AAUE,EAAA;AAEA,EAAA;AACE,IAAA;AACA,IAAA;AVXJ,EAAA;AUcI,IAAA;AACA,IAAA;AVZJ,EAAA;AUcA;AC3DO;AAGA;AACA;ACNA;AACA;AAEA;AAGA;AAKL,EAAA;AAEA,EAAA;AAEA,EAAA;AAEA,EAAA;AACF;AAEO;AAKL,EAAA;AAEA,EAAA;AACE,IAAA;AACJ;AAEO;AZoCP,EAAA;AYlCI,IAAA;AAGA,IAAA;AZkCJ,EAAA;AYhCA;ACvDO;AAIL,EAAA;AACA,EAAA;AAEA,EAAA;AACF;AAEO;AAIL,EAAA;AAEA,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AACE,MAAA;AbiFN,IAAA;Aa/EI,IAAA;AbiFJ,EAAA;Aa9EE,EAAA;AAGF;AC5BO;AAOL,EAAA;AAEA,EAAA;AAGI,IAAA;AAEA,IAAA;AACA,IAAA;AdiGN,EAAA;Ac1FM,IAAA;AAEA,IAAA;Ad2FN,EAAA;ActFE,EAAA;AACE,IAAA;AAEA,IAAA;AduFJ,EAAA;AcpFE,EAAA;AACF;AC/BO;AfsHP,EAAA;AACA,EAAA;AACA,EAAA;AepHA;ACTA;AACE,EAAA;AACA,EAAA;AASF;AAEA;AACE,EAAA;AACE,IAAA;AhBuHJ,EAAA;AgBrHE,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AhBuHJ,EAAA;AgBrHE,EAAA;AACF;AAEA;AACE,EAAA;AACF;AAEA;AACE,EAAA;AACF;AAEA;AACE,EAAA;AACE,IAAA;AhBoHJ,EAAA;AgBlHE,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACE,IAAA;AAEA,IAAA;AhBmHJ,EAAA;AgBjHE,EAAA;AACF;AAEA;AAIE,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AACE,QAAA;AhB8GR,MAAA;AACA,IAAA;AgB5GM,MAAA;AACA,MAAA;AACE,QAAA;AACE,UAAA;AACA,UAAA;AhB8GV,QAAA;AACA,MAAA;AgB5GM,MAAA;AhB8GN,IAAA;AACA,EAAA;AgB5GE,EAAA;AACF;AAEA;AACE,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AhB4GN,IAAA;AgB1GM,MAAA;AACA,MAAA;AACE,QAAA;AACE,UAAA;AACA,UAAA;AhB4GV,QAAA;AACA,MAAA;AgB1GM,MAAA;AhB4GN,IAAA;AACA,EAAA;AgB1GE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACE,IAAA;AhBwGJ,EAAA;AgBtGE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AAEA,EAAA;AhBmGF,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AgBjGE,EAAA;AhBmGF,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AgBhGE,EAAA;AACE,IAAA;AhBkGJ,EAAA;AgBhGE,EAAA;AACF;AAEA;AAIE,EAAA;AACA,EAAA;AAEA,EAAA;AACE,IAAA;AhB6FJ,EAAA;AgB1FE,EAAA;AACE,IAAA;AACE,MAAA;AhB4FN,IAAA;AgBzFI,IAAA;AACA,IAAA;AACE,MAAA;AhB2FN,IAAA;AACA,EAAA;AgBxFE,EAAA;AACF;AAEA;AACE,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AACF;AAEO;AACL,EAAA;AAEA,EAAA;AACE,IAAA;AhBmFJ,EAAA;AgBhFE,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AhBgFF,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AgB9EM,MAAA;AhBgFN,IAAA;AgB7EM,MAAA;AhB+EN,IAAA;AgB5EM,MAAA;AhB8EN,IAAA;AgB3EM,MAAA;AhB6EN,IAAA;AgB1EM,MAAA;AhB4EN,IAAA;AgBzEM,MAAA;AhB2EN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBxEM,MAAA;AhB0EN,IAAA;AgBvEM,MAAA;AhByEN,IAAA;AACA,IAAA;AACA,IAAA;AgBtEM,MAAA;AhBwEN,IAAA;AgBrEM,MAAA;AhBuEN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBpEM,MAAA;AhBsEN,IAAA;AgBnEM,MAAA;AhBqEN,IAAA;AgBlEM,MAAA;AhBoEN,IAAA;AgBjEM,MAAA;AhBmEN,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AgBhEM,MAAA;AhBkEN,EAAA;AgBhEA;AAIO;AACL,EAAA;AAOF;AC/QO;AjByUP,EAAA;AiBvUI,IAAA;AjByUJ,EAAA;AiBvUA;AA0BO;AjBgTP,EAAA;AiB3SI,IAAA;AjB6SJ,sBAAA;AACA;AACA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AACA,EAAA;AiBxSI,IAAA;AAEA,IAAA;AACE,MAAA;AAEF,IAAA;AjBwSJ,EAAA;AiBpSA;ACvDA;ACIO;AnB2VP,EAAA;AmBzVI,IAAA;AnB2VJ,EAAA;AmBzVA;AAEO;AACL,EAAA;AACA,EAAA;AAEA,EAAA;AACA,EAAA;AAEA,EAAA;AACE,IAAA;AACE,MAAA;AnBwVN,IAAA;AmBtVI,IAAA;AnBwVJ,EAAA;AmBtVA;AAEO;AACL,EAAA;AACF;AAqJO;AAIL,EAAA;AACE,IAAA;AACJ;AAGO;AAIL,EAAA;AACF;AC7KO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ApBqVF,IAAA;AACA,EAAA;AoB/UE,EAAA;ApBiVF,IAAA;AACA;AACA,IAAA;AACA,IAAA;AACA,MAAA;AACA,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AoBtUA;AClDO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ArBuWF,IAAA;AACA,EAAA;AqBjWE,EAAA;ArBmWF,IAAA;AACA;AACA,IAAA;AACA,IAAA;AACA,MAAA;AACA,QAAA;AACA,QAAA;AACA,MAAA;AACA,IAAA;AACA,EAAA;AqB7VA;AAEO;AAiBL,EAAA;AACE,IAAA;AAKF,EAAA;ArB0UF,IAAA;AACA,EAAA;AqBxUA;ACdO;AAWL,EAAA;AAIA,EAAA;AtB4UF,IAAA;AACA,IAAA;AACA,EAAA;AsBzUE,EAAA;AACE,IAAA;AtB2UJ;AACA,MAAA;AACA,0BAAA;AACA,EAAA;AsBxUE,EAAA;AACF;AAqBO;AtBsTP,EAAA;AACA,EAAA;AsBrSE;AAEK;AtBsSP,EAAA;AACA,EAAA;AsBtRE;AAEK;AtBuRP,EAAA;AACA,EAAA;AsBrRA;AtBuRA;AACA;AuBzaA;AAAA;AACE;AASF;AvBoaA;AACA;AwBtbO;AAiDA;AAML,EAAA;AAAM,IAAA;AACJ,IAAA;AACa,IAAA;AACJ,IAAA;AACT,IAAA;AACA,IAAA;AACA,EAAA;AAGF,EAAA;AAEA,EAAA;AAAmC,IAAA;AACmB,EAAA;AAGtD,EAAA;AACE,IAAA;AAAgC,MAAA;AACW,MAAA;AACzC,MAAA;AACA,MAAA;AACS,IAAA;AACV,EAAA;AAEL;AAqDO;AAOL,EAAA;AACA,EAAA;AAEA,EAAA;AAAO,IAAA;AACC,IAAA;AACa,IAAA;AAEjB,MAAA;AAEA,MAAA;AAKA,MAAA;AACE,QAAA;AAAsB,UAAA;AACpB,UAAA;AACA,QAAA;AACF,MAAA;AAGF,MAAA;AAA2C,QAAA;AACzC,QAAA;AACM,QAAA;AACN,QAAA;AACoD,MAAA;AAGtD,MAAA;AAEM,QAAA;AACK,QAAA;AACQ,MAAA;AAEb,IAAA;AACR,EAAA;AAEJ;AxB2TA;AACA;AyB/eA;AA0DO;AAAoC,EAAA;AAIvC,IAAA;AACE,MAAA;AAEA,MAAA;AAME,QAAA;AACA,QAAA;AAAO,UAAA;AAEH,YAAA;AAEA,YAAA;AAGA,YAAA;AAAO,cAAA;AAKH,gBAAA;AAGU,kBAAA;AACgB,kBAAA;AACA,gBAAA;AAG1B,gBAAA;AAAwC,kBAAA;AACM,kBAAA;AAC5C,gBAAA;AAGF,gBAAA;AACE,kBAAA;AAEA,kBAAA;AAEA,kBAAA;AACE,oBAAA;AAAA,uCAAA;AAEI,oBAAA;AACJ,gBAAA;AAEF,kBAAA;AAAmB,gBAAA;AACrB,cAAA;AACF,cAAA;AAIE,gBAAA;AAGU,kBAAA;AACgB,kBAAA;AACA,gBAAA;AAG1B,gBAAA;AAAwC,kBAAA;AACM,kBAAA;AAC5C,gBAAA;AAGF,gBAAA;AACE,kBAAA;AACA,kBAAA;AAA2D,gBAAA;AAE3D,kBAAA;AAEA,kBAAA;AAEA,kBAAA;AACE,oBAAA;AAAA,sBAAA;AAC4B,sBAAA;AACkC,oBAAA;AAE9D,oBAAA;AAAA,kBAAA;AAGF,kBAAA;AAAA,oBAAA;AACyB,oBAAA;AACmD,kBAAA;AAG5E,kBAAA;AACE,oBAAA;AAAA,sBAAA;AAC4B,sBAAA;AACkC,oBAAA;AAC9D,kBAAA;AACF,gBAAA;AAEA,kBAAA;AAAmB,gBAAA;AACrB,cAAA;AACF,YAAA;AACF,UAAA;AACF,QAAA;AACF,MAAA;AACF,IAAA;AACF,EAAA;AAEJ;AAEO;AAMkE,EAAA;AACvE,EAAA;AAEF;AAEO;AAMkE,EAAA;AACvE,EAAA;AAEF;AAEA;AASE,EAAA;AACA,EAAA;AAAmE,IAAA;AACjE,IAAA;AACA,EAAA;AAGF,EAAA;AACF;AAEA;AAAmE,EAAA;AAMjC,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACkD,EAAA;AACnD,EAAA;AAM2B,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACoD,EAAA;AACrD,EAAA;AAM2B,IAAA;AAC1B,IAAA;AACA,IAAA;AACA,IAAA;AACA,EAAA;AACD,EAAA;AAMsB,IAAA;AACrB,IAAA;AACA,IAAA;AACA,IAAA;AACoC,EAAA;AACrC,EAAA;AAIsB,IAAA;AACrB,IAAA;AACA,IAAA;AACA,IAAA;AACoC,EAAA;AAE5C;AAEO;AAA8B,EAAA;AAC2B,EAAA;AAEhE;AzBsWA;AACA;A0BvkBO;AASA;AAEP;AAIE,EAAA;AAIE,IAAA;AAAO,MAAA;AAIC,MAAA;AAEiD,IAAA;AACzD,EAAA;AAMA,IAAA;AAAO,MAAA;AAC4C,MAAA;AAEM,IAAA;AACzD,EAAA;AAEA,IAAA;AACA,IAAA;AAAO,MAAA;AAEkD,MAAA;AAIzC,IAAA;AAChB,EAAA;AAEJ;AAEA;AAKE,EAAA;AACA,EAAA;AAEA,EAAA;AAEA,EAAA;AACE,IAAA;AAEA,IAAA;AAEA,IAAA;AAA2B,EAAA;AAG7B,EAAA;AACF;AAEA;AAKE,EAAA;AAEA,EAAA;AAIA,EAAA;AACE,IAAA;AACA,IAAA;AAEA,IAAA;AAAkB,MAAA;AAChB,MAAA;AACA,IAAA;AACF,EAAA;AAGF,EAAA;AACF;AAEO;AAIL,EAAA;AACA,EAAA;AACA,EAAA;AAGA,EAAA;AAEA,EAAA;AAAO,IAAA;AAOH,MAAA;AAAyC,QAAA;AACvC,QAAA;AACA,MAAA;AAGF,MAAA;AAIA,MAAA;AACE,QAAA;AACA,QAAA;AAA4C,UAAA;AAC1C,UAAA;AACA,UAAA;AACA,QAAA;AACD,MAAA;AAGH,MAAA;AAAO,IAAA;AACT,EAAA;AAEJ;A1BugBA;AACA;AuBrpBO;AAqJP;AAA8E,EAAA;AAC3D,EAAA;AACA,EAAA;AACT,iBAAA;AACoB,EAAA;AACpB,EAAA;AACA,EAAA;AACD,EAAA;AAGL,IAAA;AACA,IAAA;AAIA,IAAA;AACA,IAAA;AAAwC,MAAA;AACrB,MAAA;AACiC,IAAA;AAEpD,IAAA;AAAyB,MAAA;AACvB,uBAAA;AACwB,IAAA;AAG1B,IAAA;AAAmB,MAAA;AACT,QAAA;AACyC,QAAA;AACN,QAAA;AACE,MAAA;AAC7C,IAAA;AACF,EAAA;AACF,EAAA;AAWE,IAAA;AACA,IAAA;AAEA,IAAA;AAEA,IAAA;AAAe,MAAA;AACiB,IAAA;AAGhC,IAAA;AAEA,IAAA;AACE,MAAA;AAAwC,IAAA;AAExC,MAAA;AAAqB,IAAA;AAGvB,IAAA;AACE,MAAA;AAAsC,IAAA;AAGxC,IAAA;AAGA,IAAA;AAEU,MAAA;AACK,MAAA;AACD,QAAA;AACA,QAAA;AACA,MAAA;AACZ,IAAA;AAGF,IAAA;AACE,MAAA;AAAO,QAAA;AACI,QAAA;AACa,QAAA;AACR,MAAA;AAChB,IAAA;AAGF,IAAA;AAAA,MAAA;AACkB,MAAA;AAChB,MAAA;AACA,IAAA;AAGF,IAAA;AAAe,MAAA;AACN,sBAAA;AACU,IAAA;AAGnB,IAAA;AAAO,MAAA;AACL,MAAA;AACsC,MAAA;AACxB,IAAA;AAChB,EAAA;AACF,EAAA;AAeE,IAAA;AAA0B,MAAA;AACxB,sBAAA;AACS,IAAA;AAEX,IAAA;AAEA,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AAC6B,MAAA;AACR,IAAA;AACvB,EAAA;AACF,EAAA;AAUE,IAAA;AACA,IAAA;AAEA,IAAA;AAEA,IAAA;AAAgC,MAAA;AAGI,MAAA;AAClC,QAAA;AACe,QAAA;AACD,UAAA;AACiB,UAAA;AACd,QAAA;AACf,MAAA;AACF,IAAA;AAGF,IAAA;AAGA,IAAA;AAAA,MAAA;AACE,MAAA;AACA,MAAA;AACA,IAAA;AAGF,IAAA;AAEA,IAAA;AAIE,MAAA;AAA2C,QAAA;AACzB,QAAA;AAChB,QAAA;AACkB,MAAA;AAEpB,MAAA;AAAO,QAAA;AACL,UAAA;AACc,UAAA;AACA,UAAA;AACF,YAAA;AACL,YAAA;AACiD,UAAA;AACtD,QAAA;AACF,wBAAA;AACiB,MAAA;AACnB,IAAA;AAGF,IAAA;AACA,IAAA;AAA2C,MAAA;AACI,MAAA;AACvC,QAAA;AACkB,QAAA;AACgD,MAAA;AACxE,MAAA;AACc,QAAA;AACZ,QAAA;AACqB,QAAA;AACE,QAAA;AACD,MAAA;AACxB,IAAA;AAGF,IAAA;AACE,MAAA;AAA8B,QAAA;AACQ,QAAA;AACpC,QAAA;AACQ,QAAA;AACU,QAAA;AAClB,QAAA;AACA,QAAA;AACS,MAAA;AACV,IAAA;AAGH,IAAA;AAAuC,MAAA;AACrC,QAAA;AACgC,QAAA;AACH,MAAA;AAC7B,MAAA;AACA,MAAA;AACkC,IAAA;AAGpC,IAAA;AACE,MAAA;AAAU,QAAA;AACR,yCAAA;AACkC,MAAA;AACpC,IAAA;AAGF,IAAA;AAAM,MAAA;AACJ,MAAA;AACa;AAAA;AAAA;AAAA,IAAA;AAMf,IAAA;AAAO,MAAA;AAEgD,MAAA;AAE1B,IAAA;AAC7B,EAAA;AACF,EAAA;AAGE,IAAA;AAEA,IAAA;AAEA,IAAA;AAAe,MAAA;AACiB,IAAA;AAGhC,IAAA;AAAsD,MAAA;AACvC,MAAA;AACN,IAAA;AAGT,IAAA;AAAwB,EAAA;AAC1B,kBAAA;AAKE,IAAA;AAA4C,EAAA;AAC9C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,kBAAA;AAUE,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AAAyB,EAAA;AAC3B,EAAA;AAME,IAAA;AAEA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACkB,IAAA;AAGzD,IAAA;AACE,MAAA;AAAkB,IAAA;AAGpB,IAAA;AACE,MAAA;AAAgD,IAAA;AAGlD,IAAA;AAAgC,MAAA;AAGd,MAAA;AAChB,QAAA;AACe,QAAA;AACsC,MAAA;AACrD,IAAA;AAGF,IAAA;AAAgD,EAAA;AAClD,EAAA;AAOE,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AACA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACc,IAAA;AAGrD,IAAA;AACE,MAAA;AAA6B,IAAA;AAG/B,IAAA;AACE,MAAA;AAAiD,IAAA;AAGnD,IAAA;AAAuB,MAAA;AAKL,MAAA;AAChB,QAAA;AACe,QAAA;AACsC,MAAA;AACrD,IAAA;AAGF,IAAA;AACE,MAAA;AAAoC,IAAA;AAGtC,IAAA;AACE,MAAA;AAAsC,IAAA;AAGxC,IAAA;AACE,MAAA;AAAa,QAAA;AACE,QAAA;AACb,MAAA;AAEF,MAAA;AAAuB,IAAA;AAGzB,IAAA;AAEA,IAAA;AAEgD,EAAA;AAClD,EAAA;AAME,IAAA;AAEA,IAAA;AACA,IAAA;AAEA,IAAA;AACA,IAAA;AACA,IAAA;AAKA,IAAA;AAAuC,MAAA;AACc,IAAA;AAGrD,IAAA;AACE,MAAA;AAA6B,IAAA;AAG/B,IAAA;AACE,MAAA;AAAiD,IAAA;AAGnD,IAAA;AACA,IAAA;AAAO,EAAA;AACT,kBAAA;AAGE,IAAA;AACA,IAAA;AAAY,EAAA;AAEhB;AAEA;AAGE,EAAA;AAGA,EAAA;AACE,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AACyB,MAAA;AACzB,IAAA;AACF,EAAA;AAGF,EAAA;AACE,IAAA;AAAmB,MAAA;AACJ,MAAA;AACA,IAAA;AAEf,IAAA;AAAO,MAAA;AACL,MAAA;AACA,MAAA;AACyB,IAAA;AAC3B,EAAA;AAGF,EAAA;AAAO,IAAA;AACL,IAAA;AACyB,EAAA;AAE7B;AAEA;AAGE,EAAA;AAGA,EAAA;AACE,IAAA;AACA,IAAA;AACA,IAAA;AAAO,MAAA;AACL,MAAA;AAC0B,MAAA;AAC1B,IAAA;AACF,EAAA;AAGF,EAAA;AACE,IAAA;AAA2C,MAAA;AACD,IAAA;AAE1C,IAAA;AAAO,MAAA;AACL,MAAA;AACA,MAAA;AACyB,IAAA;AAC3B,EAAA;AAGF,EAAA;AAAO,IAAA;AACL,IAAA;AACyB,EAAA;AAE7B;AAKO;AAIL,EAAA;AACE,IAAA;AAAO,EAAA;AAGT,EAAA;AACE,IAAA;AAEE,MAAA;AAA8D,IAAA;AAEhE,IAAA;AAAO,EAAA;AAGT,EAAA;AAEE,IAAA;AACA,IAAA;AACE,MAAA;AACA,MAAA;AAAkC,IAAA;AAGpC,IAAA;AAA8D,EAAA;AAGhE,EAAA;AACF;AAEA;AAEE,EAAA;AACE,IAAA;AAAO,EAAA;AAGT,EAAA;AACF;AAWO;AAGL,EAAA;AAGA,EAAA;AACE,IAAA;AAA0D,EAAA;AAG5D,EAAA;AACF;AAOO;AAIL,EAAA;AACF;AAMO;AAGL,EAAA;AACA,EAAA;AAAO,IAAA;AACc,IAAA;AACF,EAAA;AAErB;AAMO;AAGL,EAAA;AACF;AAKO;AAGL,EAAA;AACA,EAAA;AAAO,IAAA;AACc,EAAA;AAEvB;AvBkTA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA","file":"/home/runner/work/emmett/emmett/src/packages/emmett-mongodb/dist/index.cjs","sourcesContent":[null,"import { ValidationError } from '../errors';\n\nexport const enum ValidationErrors {\n NOT_A_NONEMPTY_STRING = 'NOT_A_NONEMPTY_STRING',\n NOT_A_POSITIVE_NUMBER = 'NOT_A_POSITIVE_NUMBER',\n NOT_AN_UNSIGNED_BIGINT = 'NOT_AN_UNSIGNED_BIGINT',\n}\n\nexport const isNumber = (val: unknown): val is number =>\n typeof val === 'number' && val === val;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport const isBigint = (val: any): val is bigint =>\n typeof val === 'bigint' && val === val;\n\nexport const isString = (val: unknown): val is string =>\n typeof val === 'string';\n\nexport const assertNotEmptyString = (value: unknown): string => {\n if (!isString(value) || value.length === 0) {\n throw new ValidationError(ValidationErrors.NOT_A_NONEMPTY_STRING);\n }\n return value;\n};\n\nexport const assertPositiveNumber = (value: unknown): number => {\n if (!isNumber(value) || value <= 0) {\n throw new ValidationError(ValidationErrors.NOT_A_POSITIVE_NUMBER);\n }\n return value;\n};\n\nexport const assertUnsignedBigInt = (value: string): bigint => {\n const number = BigInt(value);\n if (number < 0) {\n throw new ValidationError(ValidationErrors.NOT_AN_UNSIGNED_BIGINT);\n }\n return number;\n};\n\nexport * from './dates';\n","import { isNumber, isString } from '../validation';\n\nexport type ErrorConstructor<ErrorType extends Error> = new (\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n ...args: any[]\n) => ErrorType;\n\nexport const isErrorConstructor = <ErrorType extends Error>(\n // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\n expect: Function,\n): expect is ErrorConstructor<ErrorType> => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n return (\n typeof expect === 'function' &&\n expect.prototype &&\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n expect.prototype.constructor === expect\n );\n};\n\nexport class EmmettError extends Error {\n public static readonly Codes = {\n ValidationError: 400,\n IllegalStateError: 403,\n NotFoundError: 404,\n ConcurrencyError: 412,\n InternalServerError: 500,\n };\n\n public errorCode: number;\n\n constructor(\n options?: { errorCode: number; message?: string } | string | number,\n ) {\n const errorCode =\n options && typeof options === 'object' && 'errorCode' in options\n ? options.errorCode\n : isNumber(options)\n ? options\n : EmmettError.Codes.InternalServerError;\n const message =\n options && typeof options === 'object' && 'message' in options\n ? options.message\n : isString(options)\n ? options\n : `Error with status code '${errorCode}' ocurred during Emmett processing`;\n\n super(message);\n this.errorCode = errorCode;\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, EmmettError.prototype);\n }\n\n public static mapFrom(\n error: Error | { message?: string; errorCode?: number },\n ): EmmettError {\n if (EmmettError.isInstanceOf(error)) {\n return error;\n }\n\n return new EmmettError({\n errorCode:\n 'errorCode' in error &&\n error.errorCode !== undefined &&\n error.errorCode !== null\n ? error.errorCode\n : EmmettError.Codes.InternalServerError,\n message: error.message ?? 'An unknown error occurred',\n });\n }\n\n public static isInstanceOf<ErrorType extends EmmettError = EmmettError>(\n error: unknown,\n errorCode?: (typeof EmmettError.Codes)[keyof typeof EmmettError.Codes],\n ): error is ErrorType {\n return (\n typeof error === 'object' &&\n error !== null &&\n 'errorCode' in error &&\n isNumber(error.errorCode) &&\n (errorCode === undefined || error.errorCode === errorCode)\n );\n }\n}\n\nexport class ConcurrencyError extends EmmettError {\n constructor(\n public current: string | undefined,\n public expected: string,\n message?: string,\n ) {\n super({\n errorCode: EmmettError.Codes.ConcurrencyError,\n message:\n message ??\n `Expected version ${expected.toString()} does not match current ${current?.toString()}`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyError.prototype);\n }\n}\n\n// TODO: Make it derive from ConcurrencyError to avoid code duplication\n// Or add additional type to distinguinsh both errors\nexport class ConcurrencyInMemoryDatabaseError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.ConcurrencyError,\n message: message ?? `Expected document state does not match current one!`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ConcurrencyInMemoryDatabaseError.prototype);\n }\n}\n\nexport class ValidationError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.ValidationError,\n message: message ?? `Validation Error ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ValidationError.prototype);\n }\n}\n\nexport class IllegalStateError extends EmmettError {\n constructor(message?: string) {\n super({\n errorCode: EmmettError.Codes.IllegalStateError,\n message: message ?? `Illegal State ocurred during Emmett processing`,\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, IllegalStateError.prototype);\n }\n}\n\nexport class NotFoundError extends EmmettError {\n constructor(options?: { id: string; type: string; message?: string }) {\n super({\n errorCode: EmmettError.Codes.NotFoundError,\n message:\n options?.message ??\n (options?.id\n ? options.type\n ? `${options.type} with ${options.id} was not found during Emmett processing`\n : `State with ${options.id} was not found during Emmett processing`\n : options?.type\n ? `${options.type} was not found during Emmett processing`\n : 'State was not found during Emmett processing'),\n });\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, NotFoundError.prototype);\n }\n}\n","import { v4 as uuid } from 'uuid';\nimport {\n getInMemoryDatabase,\n type InMemoryDatabase,\n} from '../database/inMemoryDatabase';\nimport { bigIntProcessorCheckpoint } from '../processors';\nimport type { ProjectionRegistration } from '../projections';\nimport type {\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n ReadEventMetadataWithGlobalPosition,\n} from '../typing';\nimport { tryPublishMessagesAfterCommit } from './afterCommit';\nimport type {\n AggregateStreamOptions,\n AggregateStreamResult,\n AppendToStreamOptions,\n AppendToStreamResult,\n DefaultEventStoreOptions,\n EventStore,\n ReadStreamOptions,\n ReadStreamResult,\n StreamExistsResult,\n} from './eventStore';\nimport { assertExpectedVersionMatchesCurrent } from './expectedVersion';\nimport { handleInMemoryProjections } from './projections/inMemory';\nimport { downcastRecordedMessages, upcastRecordedMessages } from './versioning';\n\nexport const InMemoryEventStoreDefaultStreamVersion = 0n;\n\nexport type InMemoryEventStore =\n EventStore<ReadEventMetadataWithGlobalPosition> & {\n database: InMemoryDatabase;\n };\n\nexport type InMemoryReadEventMetadata = ReadEventMetadataWithGlobalPosition;\n\nexport type InMemoryProjectionHandlerContext = {\n eventStore?: InMemoryEventStore;\n database?: InMemoryDatabase;\n};\n\nexport type InMemoryEventStoreOptions =\n DefaultEventStoreOptions<InMemoryEventStore> & {\n projections?: ProjectionRegistration<\n 'inline',\n InMemoryReadEventMetadata,\n InMemoryProjectionHandlerContext\n >[];\n database?: InMemoryDatabase;\n };\n\nexport type InMemoryReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n>;\n\nexport const getInMemoryEventStore = (\n eventStoreOptions?: InMemoryEventStoreOptions,\n): InMemoryEventStore => {\n const streams = new Map<\n string,\n ReadEvent<Event, ReadEventMetadataWithGlobalPosition>[]\n >();\n\n const getAllEventsCount = () => {\n return Array.from<ReadEvent[]>(streams.values())\n .map((s) => s.length)\n .reduce((p, c) => p + c, 0);\n };\n\n // Get the database instance to be used for projections\n const database = eventStoreOptions?.database || getInMemoryDatabase();\n\n // Extract inline projections from options\n const inlineProjections = (eventStoreOptions?.projections ?? [])\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n // Create the event store object\n const eventStore: InMemoryEventStore = {\n database,\n async aggregateStream<\n State,\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n options: AggregateStreamOptions<\n State,\n EventType,\n ReadEventMetadataWithGlobalPosition,\n EventPayloadType\n >,\n ): Promise<AggregateStreamResult<State>> {\n const { evolve, initialState, read } = options;\n\n const result = await this.readStream<EventType, EventPayloadType>(\n streamName,\n read,\n );\n\n const events = result?.events ?? [];\n\n const state = events.reduce((s, e) => evolve(s, e), initialState());\n\n return {\n currentStreamVersion: BigInt(events.length),\n state,\n streamExists: result.streamExists,\n };\n },\n\n readStream: <\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n options?: ReadStreamOptions<EventType, EventPayloadType>,\n ): Promise<\n ReadStreamResult<EventType, ReadEventMetadataWithGlobalPosition>\n > => {\n const events = streams.get(streamName);\n const currentStreamVersion = events\n ? BigInt(events.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const from = Number(options?.from ?? 0);\n const to = Number(\n options?.to ??\n (options?.maxCount\n ? (options.from ?? 0n) + options.maxCount\n : (events?.length ?? 1)),\n );\n\n const resultEvents =\n events !== undefined && events.length > 0\n ? upcastRecordedMessages<\n EventType,\n EventPayloadType,\n ReadEventMetadataWithGlobalPosition\n >(\n events.slice(from, to) as ReadEvent<\n EventPayloadType,\n ReadEventMetadataWithGlobalPosition\n >[],\n options?.schema?.versioning,\n )\n : [];\n\n const result: ReadStreamResult<\n EventType,\n ReadEventMetadataWithGlobalPosition\n > = {\n currentStreamVersion,\n events: resultEvents,\n streamExists: events !== undefined && events.length > 0,\n };\n\n return Promise.resolve(result);\n },\n\n appendToStream: async <\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions<EventType, EventPayloadType>,\n ): Promise<AppendToStreamResult> => {\n const currentEvents = streams.get(streamName) ?? [];\n const currentStreamVersion =\n currentEvents.length > 0\n ? BigInt(currentEvents.length)\n : InMemoryEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n options?.expectedStreamVersion,\n InMemoryEventStoreDefaultStreamVersion,\n );\n\n const newEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = events.map((event, index) => {\n const globalPosition = BigInt(getAllEventsCount() + index + 1);\n const metadata: ReadEventMetadataWithGlobalPosition = {\n streamName,\n messageId: uuid(),\n streamPosition: BigInt(currentEvents.length + index + 1),\n globalPosition,\n checkpoint: bigIntProcessorCheckpoint(globalPosition),\n };\n return {\n ...event,\n kind: event.kind ?? 'Event',\n metadata: {\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n ...metadata,\n } as CombinedReadEventMetadata<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >,\n };\n });\n\n const positionOfLastEventInTheStream = BigInt(\n newEvents.slice(-1)[0]!.metadata.streamPosition,\n );\n\n streams.set(streamName, [\n ...currentEvents,\n ...downcastRecordedMessages(newEvents, options?.schema?.versioning),\n ]);\n\n // Process projections if there are any registered\n if (inlineProjections.length > 0) {\n await handleInMemoryProjections({\n projections: inlineProjections,\n events: newEvents,\n database: eventStore.database,\n eventStore,\n });\n }\n\n const result: AppendToStreamResult = {\n nextExpectedStreamVersion: positionOfLastEventInTheStream,\n createdNewStream:\n currentStreamVersion === InMemoryEventStoreDefaultStreamVersion,\n };\n\n await tryPublishMessagesAfterCommit<InMemoryEventStore>(\n newEvents,\n eventStoreOptions?.hooks,\n );\n\n return result;\n },\n\n streamExists: (streamName): Promise<StreamExistsResult> => {\n const events = streams.get(streamName);\n\n return Promise.resolve(events !== undefined && events.length > 0);\n },\n };\n\n return eventStore;\n};\n","import { v7 as uuid } from 'uuid';\nimport { deepEquals } from '../utils';\nimport type {\n DatabaseHandleOptionErrors,\n DatabaseHandleOptions,\n DatabaseHandleResult,\n DeleteResult,\n Document,\n DocumentHandler,\n InsertOneResult,\n OptionalUnlessRequiredIdAndVersion,\n ReplaceOneOptions,\n UpdateResult,\n WithIdAndVersion,\n WithoutId,\n} from './types';\nimport { expectedVersionValue, operationResult } from './utils';\n\nexport interface InMemoryDocumentsCollection<T extends Document> {\n handle: (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ) => Promise<DatabaseHandleResult<T>>;\n findOne: (predicate?: Predicate<T>) => Promise<T | null>;\n find: (predicate?: Predicate<T>) => Promise<T[]>;\n insertOne: (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ) => Promise<InsertOneResult>;\n deleteOne: (predicate?: Predicate<T>) => Promise<DeleteResult>;\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ) => Promise<UpdateResult>;\n}\n\nexport interface InMemoryDatabase {\n collection: <T extends Document>(\n name: string,\n ) => InMemoryDocumentsCollection<T>;\n}\n\ntype Predicate<T> = (item: T) => boolean;\ntype CollectionName = string;\n\nexport const getInMemoryDatabase = (): InMemoryDatabase => {\n const storage = new Map<CollectionName, WithIdAndVersion<Document>[]>();\n\n return {\n collection: <T extends Document, CollectionName extends string>(\n collectionName: CollectionName,\n collectionOptions: {\n errors?: DatabaseHandleOptionErrors;\n } = {},\n ): InMemoryDocumentsCollection<T> => {\n const ensureCollectionCreated = () => {\n if (!storage.has(collectionName)) storage.set(collectionName, []);\n };\n\n const errors = collectionOptions.errors;\n\n const collection = {\n collectionName,\n insertOne: async (\n document: OptionalUnlessRequiredIdAndVersion<T>,\n ): Promise<InsertOneResult> => {\n ensureCollectionCreated();\n\n const _id = (document._id as string | undefined | null) ?? uuid();\n const _version = document._version ?? 1n;\n\n const existing = await collection.findOne((c) => c._id === _id);\n\n if (existing) {\n return operationResult<InsertOneResult>(\n {\n successful: false,\n insertedId: null,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n }\n\n const documentsInCollection = storage.get(collectionName)!;\n const newDocument = { ...document, _id, _version };\n const newCollection = [...documentsInCollection, newDocument];\n storage.set(collectionName, newCollection);\n\n return operationResult<InsertOneResult>(\n {\n successful: true,\n insertedId: _id,\n nextExpectedVersion: _version,\n },\n { operationName: 'insertOne', collectionName, errors },\n );\n },\n findOne: (predicate?: Predicate<T>): Promise<T | null> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n const firstOne = filteredDocuments?.[0] ?? null;\n\n return Promise.resolve(firstOne as T | null);\n },\n find: (predicate?: Predicate<T>): Promise<T[]> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName);\n const filteredDocuments = predicate\n ? documentsInCollection?.filter((doc) => predicate(doc as T))\n : documentsInCollection;\n\n return Promise.resolve(filteredDocuments as T[]);\n },\n deleteOne: (predicate?: Predicate<T>): Promise<DeleteResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n if (predicate) {\n const foundIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (foundIndex === -1) {\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: false,\n matchedCount: 0,\n deletedCount: 0,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n } else {\n const newCollection = documentsInCollection.toSpliced(\n foundIndex,\n 1,\n );\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n }\n }\n\n const newCollection = documentsInCollection.slice(1);\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<DeleteResult>(\n {\n successful: true,\n matchedCount: 1,\n deletedCount: 1,\n },\n { operationName: 'deleteOne', collectionName, errors },\n ),\n );\n },\n replaceOne: (\n predicate: Predicate<T>,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ): Promise<UpdateResult> => {\n ensureCollectionCreated();\n\n const documentsInCollection = storage.get(collectionName)!;\n\n const firstIndex = documentsInCollection.findIndex((doc) =>\n predicate(doc as T),\n );\n\n if (firstIndex === undefined || firstIndex === -1) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 0,\n modifiedCount: 0,\n nextExpectedVersion: 0n,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const existing = documentsInCollection[firstIndex]!;\n\n if (\n typeof options?.expectedVersion === 'bigint' &&\n existing._version !== options.expectedVersion\n ) {\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: false,\n matchedCount: 1,\n modifiedCount: 0,\n nextExpectedVersion: existing._version,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n }\n\n const newVersion = existing._version + 1n;\n\n const newCollection = documentsInCollection.with(firstIndex, {\n _id: existing._id,\n ...document,\n _version: newVersion,\n });\n\n storage.set(collectionName, newCollection);\n\n return Promise.resolve(\n operationResult<UpdateResult>(\n {\n successful: true,\n modifiedCount: 1,\n matchedCount: firstIndex,\n nextExpectedVersion: newVersion,\n },\n { operationName: 'replaceOne', collectionName, errors },\n ),\n );\n },\n handle: async (\n id: string,\n handle: DocumentHandler<T>,\n options?: DatabaseHandleOptions,\n ): Promise<DatabaseHandleResult<T>> => {\n const { expectedVersion: version, ...operationOptions } =\n options ?? {};\n ensureCollectionCreated();\n const existing = await collection.findOne(({ _id }) => _id === id);\n\n const expectedVersion = expectedVersionValue(version);\n\n if (\n (existing == null && version === 'DOCUMENT_EXISTS') ||\n (existing == null && expectedVersion != null) ||\n (existing != null && version === 'DOCUMENT_DOES_NOT_EXIST') ||\n (existing != null &&\n expectedVersion !== null &&\n existing._version !== expectedVersion)\n ) {\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: false,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n }\n\n const result = handle(existing !== null ? { ...existing } : null);\n\n if (deepEquals(existing, result))\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n\n if (!existing && result) {\n const newDoc = { ...result, _id: id };\n const insertResult = await collection.insertOne({\n ...newDoc,\n _id: id,\n } as OptionalUnlessRequiredIdAndVersion<T>);\n return {\n ...insertResult,\n document: {\n ...newDoc,\n _version: insertResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n if (existing && !result) {\n const deleteResult = await collection.deleteOne(\n ({ _id }) => id === _id,\n );\n return { ...deleteResult, document: null };\n }\n\n if (existing && result) {\n const replaceResult = await collection.replaceOne(\n ({ _id }) => id === _id,\n result,\n {\n ...operationOptions,\n expectedVersion: expectedVersion ?? 'DOCUMENT_EXISTS',\n },\n );\n return {\n ...replaceResult,\n document: {\n ...result,\n _version: replaceResult.nextExpectedVersion,\n } as unknown as WithIdAndVersion<T>,\n };\n }\n\n return operationResult<DatabaseHandleResult<T>>(\n {\n successful: true,\n document: existing as WithIdAndVersion<T>,\n },\n { operationName: 'handle', collectionName, errors },\n );\n },\n };\n\n return collection;\n },\n };\n};\n","import { v7 as uuid } from 'uuid';\nimport { TaskProcessor } from './taskProcessor';\n\nexport type ExclusiveAccessGuard = {\n execute: <Result>(operation: () => Promise<Result>) => Promise<Result>;\n waitForIdle: () => Promise<void>;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardExclusiveAccess = (options?: {\n maxQueueSize?: number;\n}): ExclusiveAccessGuard => {\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: 1,\n maxQueueSize: options?.maxQueueSize ?? 1000,\n });\n\n return {\n execute: <Result>(operation: () => Promise<Result>): Promise<Result> =>\n taskProcessor.enqueue(async ({ ack }) => {\n try {\n return await operation();\n } finally {\n ack();\n }\n }),\n waitForIdle: () => taskProcessor.waitForEndOfProcessing(),\n stop: (options) => taskProcessor.stop(options),\n };\n};\n\nexport type BoundedAccessGuard<Resource> = {\n acquire: () => Promise<Resource>;\n release: (resource: Resource) => void;\n execute: <Result>(\n operation: (resource: Resource) => Promise<Result>,\n ) => Promise<Result>;\n waitForIdle: () => Promise<void>;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardBoundedAccess = <Resource>(\n getResource: () => Resource | Promise<Resource>,\n options: {\n maxResources: number;\n maxQueueSize?: number;\n reuseResources?: boolean;\n closeResource?: (resource: Resource) => void | Promise<void>;\n },\n): BoundedAccessGuard<Resource> => {\n let isStopped = false;\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: options.maxResources,\n maxQueueSize: options.maxQueueSize ?? 1000,\n });\n\n const resourcePool: Resource[] = [];\n const allResources = new Set<Resource>();\n const ackCallbacks = new Map<Resource, () => void>();\n\n const acquire = async (): Promise<Resource> =>\n taskProcessor.enqueue(async ({ ack }) => {\n try {\n let resource: Resource | undefined;\n\n if (options.reuseResources) {\n resource = resourcePool.pop();\n }\n\n if (!resource) {\n resource = await getResource();\n allResources.add(resource);\n }\n\n ackCallbacks.set(resource, ack);\n return resource;\n } catch (e) {\n ack();\n throw e;\n }\n });\n\n const release = (resource: Resource) => {\n const ack = ackCallbacks.get(resource);\n if (ack) {\n ackCallbacks.delete(resource);\n if (options.reuseResources) {\n resourcePool.push(resource);\n }\n ack();\n }\n };\n\n const execute = async <Result>(\n operation: (resource: Resource) => Promise<Result>,\n ): Promise<Result> => {\n const resource = await acquire();\n try {\n return await operation(resource);\n } finally {\n release(resource);\n }\n };\n\n return {\n acquire,\n release,\n execute,\n waitForIdle: () => taskProcessor.waitForEndOfProcessing(),\n stop: async (stopOptions) => {\n if (isStopped) return;\n isStopped = true;\n if (options?.closeResource) {\n const resources = [...allResources];\n allResources.clear();\n resourcePool.length = 0;\n await Promise.all(\n resources.map(\n async (resource) => await options.closeResource!(resource),\n ),\n );\n }\n\n await taskProcessor.stop(stopOptions);\n },\n };\n};\n\nexport type InitializedOnceGuard<T> = {\n ensureInitialized: () => Promise<T>;\n reset: () => void;\n stop: (options?: { force?: boolean }) => Promise<void>;\n};\n\nexport const guardInitializedOnce = <T>(\n initialize: () => Promise<T>,\n options?: {\n maxQueueSize?: number;\n maxRetries?: number;\n },\n): InitializedOnceGuard<T> => {\n let initPromise: Promise<T> | null = null;\n\n const taskProcessor = new TaskProcessor({\n maxActiveTasks: 1,\n maxQueueSize: options?.maxQueueSize ?? 1000,\n });\n\n const ensureInitialized = async (retryCount = 0): Promise<T> => {\n if (initPromise !== null) {\n return initPromise;\n }\n\n return taskProcessor.enqueue(\n async ({ ack }) => {\n if (initPromise !== null) {\n ack();\n return initPromise;\n }\n\n try {\n const promise = initialize();\n initPromise = promise;\n const result = await promise;\n ack();\n return result;\n } catch (error) {\n initPromise = null;\n ack();\n const maxRetries = options?.maxRetries ?? 3;\n if (retryCount < maxRetries) {\n return ensureInitialized(retryCount + 1);\n }\n throw error;\n }\n },\n { taskGroupId: uuid() },\n );\n };\n\n return {\n ensureInitialized,\n reset: () => {\n initPromise = null;\n },\n stop: (options) => taskProcessor.stop(options),\n };\n};\n","import retry from 'async-retry';\nimport { EmmettError } from '../errors';\nimport { JSONParser } from '../serialization';\n\nexport type AsyncRetryOptions<T = unknown> = retry.Options & {\n shouldRetryResult?: (result: T) => boolean;\n shouldRetryError?: (error?: unknown) => boolean;\n};\n\nexport const NoRetries: AsyncRetryOptions = { retries: 0 };\n\nexport const asyncRetry = async <T>(\n fn: () => Promise<T>,\n opts?: AsyncRetryOptions<T>,\n): Promise<T> => {\n if (opts === undefined || opts.retries === 0) return fn();\n\n return retry(\n async (bail) => {\n try {\n const result = await fn();\n\n if (opts?.shouldRetryResult && opts.shouldRetryResult(result)) {\n throw new EmmettError(\n `Retrying because of result: ${JSONParser.stringify(result)}`,\n );\n }\n return result;\n } catch (error) {\n if (opts?.shouldRetryError && !opts.shouldRetryError(error)) {\n bail(error as Error);\n return undefined as unknown as T;\n }\n throw error;\n }\n },\n opts ?? { retries: 0 },\n );\n};\n","import { v7 as uuid } from 'uuid';\nimport type { EmmettError } from '../errors';\nimport { upcastRecordedMessage } from '../eventStore';\nimport type { ProjectionDefinition } from '../projections';\nimport {\n defaultTag,\n type AnyEvent,\n type AnyMessage,\n type AnyReadEventMetadata,\n type AnyRecordedMessageMetadata,\n type BatchRecordedMessageHandlerWithContext,\n type Brand,\n type CanHandle,\n type DefaultRecord,\n type Event,\n type Message,\n type MessageHandlerResult,\n type RecordedMessage,\n type SingleMessageHandlerWithContext,\n type SingleRecordedMessageHandlerWithContext,\n} from '../typing';\nimport { bigInt } from '../utils';\nimport { onShutdown } from '../utils/shutdown';\n\nexport type CurrentMessageProcessorPosition =\n | { lastCheckpoint: ProcessorCheckpoint }\n | 'BEGINNING'\n | 'END';\n\nexport type GetCheckpoint<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n> = (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n) => ProcessorCheckpoint | null;\n\nexport const getCheckpoint = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n): ProcessorCheckpoint | null => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return, @typescript-eslint/no-unsafe-member-access\n return message.metadata.checkpoint;\n};\n\nexport const wasMessageHandled = <\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n>(\n message: RecordedMessage<MessageType, MessageMetadataType>,\n checkpoint: ProcessorCheckpoint | null,\n): boolean => {\n //TODO Make it smarter\n const messageCheckpoint = getCheckpoint(message);\n\n return (\n messageCheckpoint !== null &&\n messageCheckpoint !== undefined &&\n checkpoint !== null &&\n checkpoint !== undefined &&\n messageCheckpoint <= checkpoint\n );\n};\n\nexport type MessageProcessorStartFrom =\n | CurrentMessageProcessorPosition\n | 'CURRENT';\n\nexport type MessageProcessorType = 'projector' | 'reactor';\nexport const MessageProcessorType = {\n PROJECTOR: 'projector' as MessageProcessorType,\n REACTOR: 'reactor' as MessageProcessorType,\n};\n\nexport type MessageProcessor<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n id: string;\n instanceId: string;\n type: string;\n canHandle?: string[];\n init: (options: Partial<HandlerContext>) => Promise<void>;\n start: (\n options: Partial<HandlerContext>,\n ) => Promise<CurrentMessageProcessorPosition | undefined>;\n close: (closeOptions: Partial<HandlerContext>) => Promise<void>;\n isActive: boolean;\n handle: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n Partial<HandlerContext>\n >;\n};\n\nexport const MessageProcessor = {\n result: {\n skip: (options?: { reason?: string }): MessageHandlerResult => ({\n type: 'SKIP',\n ...(options ?? {}),\n }),\n stop: (options?: {\n reason?: string;\n error?: EmmettError;\n }): MessageHandlerResult => ({\n type: 'STOP',\n ...(options ?? {}),\n }),\n },\n};\n\nexport type MessageProcessingScope<\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = <Result = MessageHandlerResult>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => Result | Promise<Result>;\n\nexport type Checkpointer<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n read: ReadProcessorCheckpoint<HandlerContext>;\n store: StoreProcessorCheckpoint<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n};\n\nexport type ProcessorHooks<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n onInit?: OnReactorInitHook<HandlerContext>;\n onStart?: OnReactorStartHook<HandlerContext>;\n onClose?: OnReactorCloseHook<HandlerContext>;\n};\n\nexport type BaseMessageProcessorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type?: string;\n processorId: string;\n processorInstanceId?: string;\n version?: number;\n partition?: string;\n startFrom?: MessageProcessorStartFrom;\n stopAfter?: (\n message: RecordedMessage<MessageType, MessageMetadataType>,\n ) => boolean;\n processingScope?: MessageProcessingScope<HandlerContext>;\n checkpoints?: Checkpointer<MessageType, MessageMetadataType, HandlerContext>;\n canHandle?: CanHandle<MessageType>;\n hooks?: ProcessorHooks<HandlerContext>;\n};\n\nexport type HandlerOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n> =\n | {\n eachMessage: SingleRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n eachBatch?: never;\n }\n | {\n eachMessage?: never;\n eachBatch: BatchRecordedMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n >;\n };\n\nexport type OnReactorInitHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorStartHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type OnReactorCloseHook<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: HandlerContext) => Promise<void>;\n\nexport type ReactorOptions<\n MessageType extends AnyMessage = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n MessagePayloadType extends AnyMessage = MessageType,\n> = BaseMessageProcessorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext\n> &\n HandlerOptions<MessageType, MessageMetadataType, HandlerContext> & {\n messageOptions?: {\n schema?: {\n versioning?: { upcast?: (event: MessagePayloadType) => MessageType };\n };\n };\n };\n\nexport type ProjectorOptions<\n EventType extends AnyEvent = AnyEvent,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n> = Omit<\n BaseMessageProcessorOptions<EventType, MessageMetadataType, HandlerContext>,\n 'type' | 'processorId'\n> & { processorId?: string } & {\n truncateOnStart?: boolean;\n projection: ProjectionDefinition<\n EventType,\n MessageMetadataType,\n HandlerContext,\n EventPayloadType\n >;\n};\n\nexport const defaultProcessingMessageProcessingScope = <\n HandlerContext = never,\n Result = MessageHandlerResult,\n>(\n handler: (context: HandlerContext) => Result | Promise<Result>,\n partialContext: Partial<HandlerContext>,\n) => handler(partialContext as HandlerContext);\n\nexport type ProcessorCheckpoint = Brand<string, 'ProcessorCheckpoint'>;\n\nexport const bigIntProcessorCheckpoint = (value: bigint): ProcessorCheckpoint =>\n bigInt.toNormalizedString(value) as ProcessorCheckpoint;\n\nexport const parseBigIntProcessorCheckpoint = (\n value: ProcessorCheckpoint,\n): bigint => BigInt(value);\n\nexport type ReadProcessorCheckpointResult = {\n lastCheckpoint: ProcessorCheckpoint | null;\n};\n\nexport type ReadProcessorCheckpoint<\n HandlerContext extends DefaultRecord = DefaultRecord,\n> = (\n options: { processorId: string; partition?: string },\n context: HandlerContext,\n) => Promise<ReadProcessorCheckpointResult>;\n\nexport type StoreProcessorCheckpointResult =\n | {\n success: true;\n newCheckpoint: ProcessorCheckpoint | null;\n }\n | { success: false; reason: 'IGNORED' | 'MISMATCH' | 'CURRENT_AHEAD' };\n\nexport type StoreProcessorCheckpoint<\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = (\n options: {\n message: RecordedMessage<MessageType, MessageMetadataType>;\n processorId: string;\n version: number | undefined;\n lastCheckpoint: ProcessorCheckpoint | null;\n partition?: string;\n },\n context: HandlerContext,\n) => Promise<StoreProcessorCheckpointResult>;\n\nexport const defaultProcessorVersion = 1;\nexport const defaultProcessorPartition = defaultTag;\n\nexport const getProcessorInstanceId = (processorId: string): string =>\n `${processorId}:${uuid()}`;\n\nexport const getProjectorId = (options: { projectionName: string }): string =>\n `emt:processor:projector:${options.projectionName}`;\n\nexport const reactor = <\n MessageType extends Message = AnyMessage,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n MessagePayloadType extends Message = MessageType,\n>(\n options: ReactorOptions<\n MessageType,\n MessageMetadataType,\n HandlerContext,\n MessagePayloadType\n >,\n): MessageProcessor<MessageType, MessageMetadataType, HandlerContext> => {\n const {\n checkpoints,\n processorId,\n processorInstanceId: instanceId = getProcessorInstanceId(processorId),\n type = MessageProcessorType.REACTOR,\n version = defaultProcessorVersion,\n partition = defaultProcessorPartition,\n hooks = {},\n processingScope = defaultProcessingMessageProcessingScope,\n startFrom,\n canHandle,\n stopAfter,\n } = options;\n\n const eachMessage: SingleMessageHandlerWithContext<\n MessageType,\n MessageMetadataType,\n HandlerContext\n > =\n 'eachMessage' in options && options.eachMessage\n ? options.eachMessage\n : () => Promise.resolve();\n\n let isInitiated = false;\n let isActive = false;\n\n let lastCheckpoint: ProcessorCheckpoint | null = null;\n let closeSignal: (() => void) | null = null;\n\n const init = async (initOptions: Partial<HandlerContext>): Promise<void> => {\n if (isInitiated) return;\n\n if (hooks.onInit === undefined) {\n isInitiated = true;\n return;\n }\n\n return await processingScope(async (context) => {\n await hooks.onInit!(context);\n isInitiated = true;\n }, initOptions);\n };\n\n const close = async (\n closeOptions: Partial<HandlerContext>,\n ): Promise<void> => {\n // TODO: Align when active is set to false\n // if (!isActive) return;\n\n isActive = false;\n\n if (closeSignal) {\n closeSignal();\n closeSignal = null;\n }\n\n if (hooks.onClose) {\n await processingScope(hooks.onClose, closeOptions);\n }\n };\n\n return {\n // TODO: Consider whether not make it optional or add URN prefix\n id: processorId,\n instanceId,\n type,\n canHandle,\n init,\n start: async (\n startOptions: Partial<HandlerContext>,\n ): Promise<CurrentMessageProcessorPosition | undefined> => {\n if (isActive) return;\n\n await init(startOptions);\n\n isActive = true;\n\n closeSignal = onShutdown(() => close(startOptions));\n\n if (lastCheckpoint !== null)\n return {\n lastCheckpoint,\n };\n\n return await processingScope(async (context) => {\n if (hooks.onStart) {\n await hooks.onStart(context);\n }\n\n if (startFrom && startFrom !== 'CURRENT') return startFrom;\n\n if (checkpoints) {\n const readResult = await checkpoints?.read(\n {\n processorId: processorId,\n partition,\n },\n { ...startOptions, ...context },\n );\n lastCheckpoint = readResult.lastCheckpoint;\n }\n\n if (lastCheckpoint === null) return 'BEGINNING';\n\n return {\n lastCheckpoint,\n };\n }, startOptions);\n },\n close,\n get isActive() {\n return isActive;\n },\n handle: async (\n messages: RecordedMessage<MessageType, MessageMetadataType>[],\n partialContext: Partial<HandlerContext>,\n ): Promise<MessageHandlerResult> => {\n if (!isActive) return Promise.resolve();\n\n return await processingScope(async (context) => {\n let result: MessageHandlerResult = undefined;\n\n for (const message of messages) {\n if (wasMessageHandled(message, lastCheckpoint)) continue;\n\n const upcasted = upcastRecordedMessage(\n // TODO: Make it smarter\n message as unknown as RecordedMessage<\n MessagePayloadType,\n MessageMetadataType\n >,\n options.messageOptions?.schema?.versioning,\n );\n\n if (canHandle !== undefined && !canHandle.includes(upcasted.type))\n continue;\n\n const messageProcessingResult = await eachMessage(upcasted, context);\n\n if (checkpoints) {\n const storeCheckpointResult: StoreProcessorCheckpointResult =\n await checkpoints.store(\n {\n processorId,\n version,\n message: upcasted,\n lastCheckpoint,\n partition,\n },\n context,\n );\n\n if (storeCheckpointResult.success) {\n // TODO: Add correct handling of the storing checkpoint\n lastCheckpoint = storeCheckpointResult.newCheckpoint;\n }\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'STOP'\n ) {\n isActive = false;\n result = messageProcessingResult;\n break;\n }\n\n if (stopAfter && stopAfter(upcasted)) {\n isActive = false;\n result = { type: 'STOP', reason: 'Stop condition reached' };\n break;\n }\n\n if (\n messageProcessingResult &&\n messageProcessingResult.type === 'SKIP'\n )\n continue;\n }\n\n return result;\n }, partialContext);\n },\n };\n};\n\nexport const projector = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n HandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n>(\n options: ProjectorOptions<\n EventType,\n EventMetaDataType,\n HandlerContext,\n EventPayloadType\n >,\n): MessageProcessor<EventType, EventMetaDataType, HandlerContext> => {\n const {\n projection,\n processorId = getProjectorId({\n projectionName: projection.name ?? 'unknown',\n }),\n ...rest\n } = options;\n\n return reactor<\n EventType,\n EventMetaDataType,\n HandlerContext,\n EventPayloadType\n >({\n ...rest,\n type: MessageProcessorType.PROJECTOR,\n canHandle: projection.canHandle,\n processorId,\n messageOptions: options.projection.eventsOptions,\n hooks: {\n onInit: options.hooks?.onInit,\n onStart:\n (options.truncateOnStart && options.projection.truncate) ||\n options.hooks?.onStart\n ? async (context: HandlerContext) => {\n if (options.truncateOnStart && options.projection.truncate)\n await options.projection.truncate(context);\n\n if (options.hooks?.onStart) await options.hooks?.onStart(context);\n }\n : undefined,\n onClose: options.hooks?.onClose,\n },\n eachMessage: async (\n event: RecordedMessage<EventType, EventMetaDataType>,\n context: HandlerContext,\n ) => projection.handle([event], context),\n });\n};\n","import { v4 as uuid } from 'uuid';\nimport {\n handleInMemoryProjections,\n type InMemoryProjectionDefinition,\n} from '.';\nimport { bigIntProcessorCheckpoint } from '../../..';\nimport {\n getInMemoryDatabase,\n type Document,\n type InMemoryDatabase,\n} from '../../../database';\nimport { isErrorConstructor } from '../../../errors';\nimport { JSONParser } from '../../../serialization';\nimport {\n assertFails,\n AssertionError,\n assertTrue,\n type ThenThrows,\n} from '../../../testing';\nimport type {\n CombinedReadEventMetadata,\n Event,\n ReadEvent,\n} from '../../../typing';\nimport type {\n InMemoryEventStore,\n InMemoryReadEventMetadata,\n} from '../../inMemoryEventStore';\n\n// Define a more specific type for T that extends Document\ntype DocumentWithId = Document & { _id?: string | number };\n\nexport type InMemoryProjectionSpecEvent<\n EventType extends Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n> = EventType & {\n metadata?: Partial<EventMetaDataType>;\n};\n\nexport type InMemoryProjectionSpecWhenOptions = { numberOfTimes: number };\n\nexport type InMemoryProjectionSpec<EventType extends Event> = (\n givenEvents: InMemoryProjectionSpecEvent<EventType>[],\n) => {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n then: (assert: InMemoryProjectionAssert, message?: string) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type InMemoryProjectionAssert = (options: {\n database: InMemoryDatabase;\n}) => Promise<void | boolean>;\n\nexport type InMemoryProjectionSpecOptions<EventType extends Event> = {\n projection: InMemoryProjectionDefinition<EventType>;\n};\n\nexport const InMemoryProjectionSpec = {\n for: <EventType extends Event>(\n options: InMemoryProjectionSpecOptions<EventType>,\n ): InMemoryProjectionSpec<EventType> => {\n const { projection } = options;\n\n return (givenEvents: InMemoryProjectionSpecEvent<EventType>[]) => {\n return {\n when: (\n events: InMemoryProjectionSpecEvent<EventType>[],\n options?: InMemoryProjectionSpecWhenOptions,\n ) => {\n const allEvents: ReadEvent<EventType, InMemoryReadEventMetadata>[] =\n [];\n\n const run = async (database: InMemoryDatabase) => {\n let globalPosition = 0n;\n const numberOfTimes = options?.numberOfTimes ?? 1;\n\n for (const event of [\n ...givenEvents,\n ...Array.from({ length: numberOfTimes }).flatMap(() => events),\n ]) {\n const metadata: InMemoryReadEventMetadata = {\n checkpoint: bigIntProcessorCheckpoint(++globalPosition),\n globalPosition,\n streamPosition: globalPosition,\n streamName: event.metadata?.streamName ?? `test-${uuid()}`,\n messageId: uuid(),\n };\n\n allEvents.push({\n ...event,\n kind: 'Event',\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n } as CombinedReadEventMetadata<\n EventType,\n InMemoryReadEventMetadata\n >,\n });\n }\n\n // Create a minimal mock EventStore implementation\n const mockEventStore = {\n database,\n aggregateStream: async () => {\n return Promise.resolve({\n state: {},\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n readStream: async () => {\n return Promise.resolve({\n events: [],\n currentStreamVersion: 0n,\n streamExists: false,\n });\n },\n appendToStream: async () => {\n return Promise.resolve({\n nextExpectedStreamVersion: 0n,\n createdNewStream: false,\n });\n },\n streamExists: async () => {\n return Promise.resolve(false);\n },\n } as InMemoryEventStore;\n\n await handleInMemoryProjections({\n events: allEvents,\n projections: [projection],\n database,\n eventStore: mockEventStore,\n });\n };\n\n return {\n then: async (\n assertFn: InMemoryProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n await run(database);\n\n const succeeded = await assertFn({ database });\n\n if (succeeded !== undefined && succeeded === false) {\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n }\n },\n thenThrows: async <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const database = getInMemoryDatabase();\n try {\n await run(database);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n }\n },\n };\n },\n };\n };\n },\n};\n\n// Helper functions for creating events in stream\nexport const eventInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n>(\n streamName: string,\n event: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>,\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n } as Partial<EventMetaDataType>,\n };\n};\n\nexport const eventsInStream = <\n EventType extends Event = Event,\n EventMetaDataType extends InMemoryReadEventMetadata =\n InMemoryReadEventMetadata,\n>(\n streamName: string,\n events: InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[],\n): InMemoryProjectionSpecEvent<EventType, EventMetaDataType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\n// Assertion helpers for checking documents\nexport function documentExists<T extends DocumentWithId>(\n expected: Partial<T>,\n options: { inCollection: string; withId: string | number },\n): InMemoryProjectionAssert {\n return async ({ database }) => {\n const collection = database.collection<T>(options.inCollection);\n\n const document = await collection.findOne((doc) => {\n // Handle both string IDs and numeric IDs in a type-safe way\n const docId = '_id' in doc ? doc._id : undefined;\n return docId === options.withId;\n });\n\n if (!document) {\n assertFails(\n `Document with ID ${options.withId} does not exist in collection ${options.inCollection}`,\n );\n return Promise.resolve(false);\n }\n\n // Check that all expected properties exist with expected values\n for (const [key, value] of Object.entries(expected)) {\n const propKey = key as keyof typeof document;\n if (\n !(key in document) ||\n JSONParser.stringify(document[propKey]) !== JSONParser.stringify(value)\n ) {\n assertFails(`Property ${key} doesn't match the expected value`);\n return Promise.resolve(false);\n }\n }\n\n return Promise.resolve(true);\n };\n}\n\n// Helper for checking document contents\nexport const expectInMemoryDocuments = {\n fromCollection: <T extends DocumentWithId>(collectionName: string) => ({\n withId: (id: string | number) => ({\n toBeEqual: (expected: Partial<T>): InMemoryProjectionAssert =>\n documentExists<T>(expected, {\n inCollection: collectionName,\n withId: id,\n }),\n }),\n }),\n};\n","import { v7 as uuid } from 'uuid';\nimport {\n canCreateEventStoreSession,\n isExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n nulloSessionFactory,\n STREAM_DOES_NOT_EXIST,\n type AppendStreamResultOfEventStore,\n type AppendToStreamOptions,\n type EventStore,\n type EventStoreSession,\n type ExpectedStreamVersion,\n type ReadStreamOptions,\n} from '../eventStore';\nimport type {\n AnyCommand,\n AnyEvent,\n AnyReadEventMetadata,\n Event,\n RecordedMessage,\n} from '../typing';\nimport { asyncRetry, NoRetries, type AsyncRetryOptions } from '../utils';\nimport type {\n WorkflowEvent,\n WorkflowInputMessageMetadata,\n WorkflowMessageAction,\n} from './workflow';\nimport type { WorkflowOptions } from './workflowProcessor';\n\nexport const WorkflowHandlerStreamVersionConflictRetryOptions: AsyncRetryOptions =\n {\n retries: 3,\n minTimeout: 100,\n factor: 1.5,\n shouldRetryError: isExpectedVersionConflictError,\n };\n\nexport type WorkflowHandlerRetryOptions =\n | AsyncRetryOptions\n | { onVersionConflict: true | number | AsyncRetryOptions };\n\nconst fromWorkflowHandlerRetryOptions = (\n retryOptions: WorkflowHandlerRetryOptions | undefined,\n): AsyncRetryOptions => {\n if (retryOptions === undefined) return NoRetries;\n\n if ('onVersionConflict' in retryOptions) {\n if (typeof retryOptions.onVersionConflict === 'boolean')\n return WorkflowHandlerStreamVersionConflictRetryOptions;\n else if (typeof retryOptions.onVersionConflict === 'number')\n return {\n ...WorkflowHandlerStreamVersionConflictRetryOptions,\n retries: retryOptions.onVersionConflict,\n };\n else return retryOptions.onVersionConflict;\n }\n\n return retryOptions;\n};\n\n// #region workflow-handler\nexport type WorkflowHandlerResult<\n Output extends AnyEvent | AnyCommand,\n Store extends EventStore,\n> = AppendStreamResultOfEventStore<Store> & {\n newMessages: Output[];\n};\n\nexport type WorkflowHandleOptions<Store extends EventStore> = Parameters<\n Store['appendToStream']\n>[2] & {\n expectedStreamVersion?: ExpectedStreamVersion;\n retry?: WorkflowHandlerRetryOptions;\n};\n\ntype WorkflowInternalState<State> = {\n userState: State;\n processedInputIds: Set<string>;\n};\n\nconst emptyHandlerResult = <\n Output extends AnyEvent | AnyCommand,\n Store extends EventStore,\n>(\n nextExpectedStreamVersion: bigint = 0n,\n): WorkflowHandlerResult<Output, Store> =>\n ({\n newMessages: [] as Output[],\n createdNewStream: false,\n nextExpectedStreamVersion,\n }) as unknown as WorkflowHandlerResult<Output, Store>;\n\nconst createInputMetadata = (\n originalMessageId: string,\n action: Extract<WorkflowMessageAction, 'InitiatedBy' | 'Received'>,\n): WorkflowInputMessageMetadata => ({\n originalMessageId,\n input: true,\n action,\n});\n\nconst tagOutputMessage = <Output extends AnyEvent | AnyCommand>(\n msg: Output,\n action: Extract<WorkflowMessageAction, 'Sent' | 'Published' | 'Scheduled'>,\n): Output => {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n const existingMetadata =\n 'metadata' in msg && msg.metadata ? msg.metadata : {};\n return {\n ...msg,\n metadata: {\n ...existingMetadata,\n action,\n },\n } as Output;\n};\n\nconst createWrappedInitialState = <State>(initialState: () => State) => {\n return (): WorkflowInternalState<State> => ({\n userState: initialState(),\n processedInputIds: new Set(),\n });\n};\n\nconst createWrappedEvolve = <\n Input extends AnyEvent | AnyCommand,\n Output extends AnyEvent | AnyCommand,\n State,\n>(\n evolve: (state: State, event: WorkflowEvent<Input | Output>) => State,\n workflowName: string,\n separateInputInboxFromProcessing: boolean,\n) => {\n return (\n state: WorkflowInternalState<State>,\n event: WorkflowEvent<Input | Output>,\n ): WorkflowInternalState<State> => {\n const metadata = (event as Record<string, unknown>).metadata as\n | Record<string, unknown>\n | undefined;\n\n // Track processed inputs for idempotency\n let processedInputIds = state.processedInputIds;\n if (\n metadata?.input === true &&\n typeof metadata?.originalMessageId === 'string'\n ) {\n processedInputIds = new Set(state.processedInputIds);\n processedInputIds.add(metadata.originalMessageId);\n }\n\n // In separated inbox mode, don't apply inputs to state - they're just sitting in inbox\n // Only outputs (from processing) should update state\n if (separateInputInboxFromProcessing && metadata?.input === true) {\n return {\n userState: state.userState,\n processedInputIds,\n };\n }\n\n // Strip workflow prefix from input event types\n const eventType = event.type as string;\n const eventForEvolve = eventType.startsWith(`${workflowName}:`)\n ? ({\n ...event,\n type: eventType.replace(`${workflowName}:`, ''),\n } as WorkflowEvent<Input | Output>)\n : event;\n\n return {\n userState: evolve(state.userState, eventForEvolve),\n processedInputIds,\n };\n };\n};\n\nexport const workflowStreamName = ({\n workflowName,\n workflowId,\n}: {\n workflowName: string;\n workflowId: string;\n}) => `emt:workflow:${workflowName}:${workflowId}`;\n\nexport const WorkflowHandler =\n <\n Input extends AnyEvent | AnyCommand,\n State,\n Output extends AnyEvent | AnyCommand,\n MessageMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n StoredMessage extends AnyEvent | AnyCommand = Output,\n >(\n options: WorkflowOptions<\n Input,\n State,\n Output,\n MessageMetadataType,\n StoredMessage\n > & {\n retry?: WorkflowHandlerRetryOptions;\n },\n ) =>\n async <Store extends EventStore>(\n store: Store,\n message: Input | RecordedMessage<Input, MessageMetadataType>,\n handleOptions?: WorkflowHandleOptions<Store>,\n ): Promise<WorkflowHandlerResult<Output, Store>> =>\n asyncRetry(\n async () => {\n const result = await withSession<\n Store,\n WorkflowHandlerResult<Output, Store>\n >(store, async ({ eventStore }) => {\n const {\n workflow: { evolve, initialState, decide, name: workflowName },\n getWorkflowId,\n } = options;\n\n const inputMessageId =\n // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n ('metadata' in message && message.metadata?.messageId\n ? // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access\n (message.metadata.messageId as string | undefined)\n : undefined) ?? uuid();\n\n const messageWithMetadata: RecordedMessage<\n Input,\n MessageMetadataType\n > = {\n ...message,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n metadata: {\n messageId: inputMessageId,\n ...(message as RecordedMessage<Input, MessageMetadataType>)\n .metadata,\n },\n } as RecordedMessage<Input, MessageMetadataType>;\n\n const workflowId = getWorkflowId(messageWithMetadata);\n\n if (!workflowId) {\n return emptyHandlerResult<Output, Store>();\n }\n\n const streamName = options.mapWorkflowId\n ? options.mapWorkflowId(workflowId)\n : workflowStreamName({ workflowName, workflowId });\n\n const messageType = messageWithMetadata.type as string;\n const hasWorkflowPrefix = messageType.startsWith(`${workflowName}:`);\n\n // Separated inbox mode: store-only path (no prefix = external input)\n if (options.separateInputInboxFromProcessing && !hasWorkflowPrefix) {\n const inputMetadata = createInputMetadata(\n inputMessageId,\n 'InitiatedBy',\n );\n\n const inputToStore = {\n type: `${workflowName}:${messageWithMetadata.type}`,\n data: messageWithMetadata.data,\n kind: messageWithMetadata.kind,\n metadata: inputMetadata,\n } as StoredMessage;\n\n const appendResult = await eventStore.appendToStream(\n streamName,\n [inputToStore] as unknown as Event[],\n {\n ...(handleOptions as AppendToStreamOptions<\n Event,\n StoredMessage & Event\n >),\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n );\n\n return {\n ...appendResult,\n newMessages: [] as Output[],\n } as unknown as WorkflowHandlerResult<Output, Store>;\n }\n\n // Wrap the evolve and initialState for idempotency tracking\n const wrappedInitialState = createWrappedInitialState(initialState);\n const wrappedEvolve = createWrappedEvolve(\n evolve,\n workflowName,\n options.separateInputInboxFromProcessing ?? false,\n ) as (\n state: WorkflowInternalState<State>,\n event: WorkflowEvent<Input | Output>,\n ) => WorkflowInternalState<State>;\n\n // 1. Aggregate the stream\n const aggregationResult = await eventStore.aggregateStream<\n WorkflowInternalState<State>,\n WorkflowEvent<Input | Output>,\n StoredMessage & Event\n >(streamName, {\n evolve: wrappedEvolve,\n initialState: wrappedInitialState,\n read: {\n ...(handleOptions as ReadStreamOptions<\n WorkflowEvent<Input | Output>,\n StoredMessage & Event\n >),\n // expected stream version is passed to fail fast\n // if stream is in the wrong state\n\n expectedStreamVersion:\n handleOptions?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n },\n });\n\n // 2. Use the aggregate state\n\n const { currentStreamVersion } = aggregationResult;\n\n const { userState: state, processedInputIds } =\n aggregationResult.state;\n\n // Idempotency: skip if this input was already processed\n\n if (processedInputIds.has(inputMessageId)) {\n return emptyHandlerResult<Output, Store>(currentStreamVersion);\n }\n\n // 3. Run business logic\n // Strip workflow prefix from message type if present (for separated inbox processing)\n const messageForDecide = hasWorkflowPrefix\n ? ({\n ...messageWithMetadata,\n type: messageType.replace(`${workflowName}:`, ''),\n } as Input)\n : (messageWithMetadata as Input);\n\n const result = decide(messageForDecide, state);\n\n const inputMetadata = createInputMetadata(\n inputMessageId,\n aggregationResult.streamExists ? 'Received' : 'InitiatedBy',\n );\n\n const inputToStore = {\n type: `${workflowName}:${messageWithMetadata.type}`,\n data: messageWithMetadata.data,\n kind: messageWithMetadata.kind,\n metadata: inputMetadata,\n } as StoredMessage;\n\n const outputMessages = (\n Array.isArray(result) ? result : [result]\n ).filter((msg): msg is Output => msg !== undefined && msg !== null);\n\n const outputCommandTypes = options.outputs?.commands ?? [];\n const taggedOutputMessages = outputMessages.map((msg) => {\n const action: WorkflowMessageAction = outputCommandTypes.includes(\n msg.type as string,\n )\n ? 'Sent'\n : 'Published';\n return tagOutputMessage(msg, action);\n });\n\n const messagesToAppend =\n options.separateInputInboxFromProcessing && hasWorkflowPrefix\n ? [...taggedOutputMessages] // input already in stream\n : [inputToStore, ...taggedOutputMessages]; // normal: store input + outputs\n\n // If there are no messages to append, return early with current state\n if (messagesToAppend.length === 0) {\n return emptyHandlerResult<Output, Store>(currentStreamVersion);\n }\n\n // Either use:\n // - provided expected stream version,\n // - current stream version got from stream aggregation,\n // - or expect stream not to exists otherwise.\n\n const expectedStreamVersion: ExpectedStreamVersion =\n handleOptions?.expectedStreamVersion ??\n (aggregationResult.streamExists\n ? currentStreamVersion\n : STREAM_DOES_NOT_EXIST);\n\n // 4. Append result to the stream\n const appendResult = await eventStore.appendToStream(\n streamName,\n // TODO: Fix this cast\n messagesToAppend as unknown as Event[],\n {\n ...(handleOptions as AppendToStreamOptions<\n Event,\n StoredMessage & Event\n >),\n expectedStreamVersion,\n },\n );\n\n // 5. Return result with output messages only\n return {\n ...appendResult,\n newMessages: outputMessages,\n } as unknown as WorkflowHandlerResult<Output, Store>;\n });\n\n return result;\n },\n fromWorkflowHandlerRetryOptions(\n handleOptions && 'retry' in handleOptions\n ? handleOptions.retry\n : options.retry,\n ),\n );\n// #endregion stream-handler\n\nconst withSession = <EventStoreType extends EventStore, T = unknown>(\n eventStore: EventStoreType,\n callback: (session: EventStoreSession<EventStoreType>) => Promise<T>,\n) => {\n const sessionFactory = canCreateEventStoreSession<EventStoreType>(eventStore)\n ? eventStore\n : nulloSessionFactory<EventStoreType>(eventStore);\n\n return sessionFactory.withSession(callback);\n};\n","import type {\n BatchRecordedMessageHandlerWithContext,\n BatchRecordedMessageHandlerWithoutContext,\n DefaultRecord,\n Event,\n ReadEvent,\n} from '../../typing';\nimport type { EventStore, EventStoreReadEventMetadata } from '../eventStore';\n\nexport type AfterEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\nexport type BeforeEventStoreCommitHandler<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = HandlerContext extends undefined\n ? BatchRecordedMessageHandlerWithoutContext<\n Event,\n EventStoreReadEventMetadata<Store>\n >\n : BatchRecordedMessageHandlerWithContext<\n Event,\n EventStoreReadEventMetadata<Store>,\n NonNullable<HandlerContext>\n >;\n\ntype TryPublishMessagesAfterCommitOptions<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n> = {\n onAfterCommit?: AfterEventStoreCommitHandler<Store, HandlerContext>;\n};\n\nexport async function tryPublishMessagesAfterCommit<Store extends EventStore>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options: TryPublishMessagesAfterCommitOptions<Store, undefined> | undefined,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context: HandlerContext,\n): Promise<boolean>;\nexport async function tryPublishMessagesAfterCommit<\n Store extends EventStore,\n HandlerContext extends DefaultRecord | undefined = undefined,\n>(\n messages: ReadEvent<Event, EventStoreReadEventMetadata<Store>>[],\n options:\n | TryPublishMessagesAfterCommitOptions<Store, HandlerContext>\n | undefined,\n context?: HandlerContext,\n): Promise<boolean> {\n if (options?.onAfterCommit === undefined) return false;\n\n try {\n await options?.onAfterCommit(messages, context!);\n return true;\n } catch (error) {\n // TODO: enhance with tracing\n console.error(`Error in on after commit hook`, error);\n return false;\n }\n}\n","export * from './deepReadonly';\n\nexport * from './command';\nexport * from './event';\nexport * from './message';\nexport * from './messageHandling';\n\nexport * from './decider';\n\nexport type Brand<K, T> = K & { readonly __brand: T };\nexport type Flavour<K, T> = K & { readonly __brand?: T };\n\nexport type DefaultRecord = Record<string, unknown>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type AnyRecord = Record<string, any>;\n\nexport type NonNullable<T> = T extends null | undefined ? never : T;\n\nexport const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = `${emmettPrefix}:default`;\nexport const unknownTag = `${emmettPrefix}:unknown`;\n","import { ConcurrencyError, EmmettError } from '../errors';\nimport type { Flavour, StreamPosition } from '../typing';\n\nexport type ExpectedStreamVersion =\n | ExpectedStreamVersionWithValue\n | ExpectedStreamVersionGeneral;\n\nexport type ExpectedStreamVersionWithValue = Flavour<\n StreamPosition,\n 'StreamVersion'\n>;\n\nexport type ExpectedStreamVersionGeneral = Flavour<\n 'STREAM_EXISTS' | 'STREAM_DOES_NOT_EXIST' | 'NO_CONCURRENCY_CHECK',\n 'StreamVersion'\n>;\n\nexport const STREAM_EXISTS = 'STREAM_EXISTS' as ExpectedStreamVersionGeneral;\nexport const STREAM_DOES_NOT_EXIST =\n 'STREAM_DOES_NOT_EXIST' as ExpectedStreamVersionGeneral;\nexport const NO_CONCURRENCY_CHECK =\n 'NO_CONCURRENCY_CHECK' as ExpectedStreamVersionGeneral;\n\nexport const matchesExpectedVersion = (\n current: StreamPosition | undefined,\n expected: ExpectedStreamVersion,\n defaultVersion: StreamPosition,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === defaultVersion;\n\n if (expected == STREAM_EXISTS) return current !== defaultVersion;\n\n return current === expected;\n};\n\nexport const assertExpectedVersionMatchesCurrent = (\n current: StreamPosition,\n expected: ExpectedStreamVersion | undefined,\n defaultVersion: StreamPosition,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected, defaultVersion))\n throw new ExpectedVersionConflictError(current, expected);\n};\n\nexport class ExpectedVersionConflictError extends ConcurrencyError {\n constructor(current: StreamPosition, expected: ExpectedStreamVersion) {\n super(current?.toString(), expected?.toString());\n\n // 👇️ because we are extending a built-in class\n Object.setPrototypeOf(this, ExpectedVersionConflictError.prototype);\n }\n}\n\nexport const isExpectedVersionConflictError = (\n error: unknown,\n): error is ExpectedVersionConflictError =>\n error instanceof ExpectedVersionConflictError ||\n EmmettError.isInstanceOf<ConcurrencyError>(\n error,\n ExpectedVersionConflictError.Codes.ConcurrencyError,\n );\n","export const hasDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n) => {\n const mapped = array.map(predicate);\n const uniqueValues = new Set(mapped);\n\n return uniqueValues.size < mapped.length;\n};\n\nexport const getDuplicates = <ArrayItem, Mapped>(\n array: ArrayItem[],\n predicate: (value: ArrayItem, index: number, array: ArrayItem[]) => Mapped,\n): ArrayItem[] => {\n const map = new Map<Mapped, ArrayItem[]>();\n\n for (let i = 0; i < array.length; i++) {\n const item = array[i]!;\n const key = predicate(item, i, array);\n if (!map.has(key)) {\n map.set(key, []);\n }\n map.get(key)!.push(item);\n }\n\n return Array.from(map.values())\n .filter((group) => group.length > 1)\n .flat();\n};\n","export const merge = <T>(\n array: T[],\n item: T,\n where: (current: T) => boolean,\n onExisting: (current: T) => T,\n onNotFound: () => T | undefined = () => undefined,\n) => {\n let wasFound = false;\n\n const result = array\n // merge the existing item if matches condition\n .map((p: T) => {\n if (!where(p)) return p;\n\n wasFound = true;\n return onExisting(p);\n })\n // filter out item if undefined was returned\n // for cases of removal\n .filter((p) => p !== undefined)\n // make TypeScript happy\n .map((p) => {\n if (!p) throw Error('That should not happen');\n\n return p;\n });\n\n // if item was not found and onNotFound action is defined\n // try to generate new item\n if (!wasFound) {\n const result = onNotFound();\n\n if (result !== undefined) return [...array, item];\n }\n\n return result;\n};\n","import { getDuplicates, hasDuplicates } from './duplicates';\nimport { merge } from './merge';\n\nexport * from './merge';\n\nexport const arrayUtils = {\n merge,\n hasDuplicates,\n getDuplicates,\n};\n","const isPrimitive = (value: unknown): boolean => {\n const type = typeof value;\n return (\n value === null ||\n value === undefined ||\n type === 'boolean' ||\n type === 'number' ||\n type === 'string' ||\n type === 'symbol' ||\n type === 'bigint'\n );\n};\n\nconst compareArrays = <T>(left: T[], right: T[]): boolean => {\n if (left.length !== right.length) {\n return false;\n }\n for (let i = 0; i < left.length; i++) {\n const leftHas = i in left;\n const rightHas = i in right;\n if (leftHas !== rightHas) return false;\n if (leftHas && !deepEquals(left[i], right[i])) return false;\n }\n return true;\n};\n\nconst compareDates = (left: Date, right: Date): boolean => {\n return left.getTime() === right.getTime();\n};\n\nconst compareRegExps = (left: RegExp, right: RegExp): boolean => {\n return left.toString() === right.toString();\n};\n\nconst compareErrors = (left: Error, right: Error): boolean => {\n if (left.message !== right.message || left.name !== right.name) {\n return false;\n }\n const leftKeys = Object.keys(left);\n const rightKeys = Object.keys(right);\n if (leftKeys.length !== rightKeys.length) return false;\n const rightKeySet = new Set(rightKeys);\n for (const key of leftKeys) {\n if (!rightKeySet.has(key)) return false;\n // @ts-expect-error - accessing dynamic keys\n if (!deepEquals(left[key], right[key])) return false;\n }\n return true;\n};\n\nconst compareMaps = (\n left: Map<unknown, unknown>,\n right: Map<unknown, unknown>,\n): boolean => {\n if (left.size !== right.size) return false;\n\n for (const [key, value] of left) {\n if (isPrimitive(key)) {\n if (!right.has(key) || !deepEquals(value, right.get(key))) {\n return false;\n }\n } else {\n let found = false;\n for (const [rightKey, rightValue] of right) {\n if (deepEquals(key, rightKey) && deepEquals(value, rightValue)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareSets = (left: Set<unknown>, right: Set<unknown>): boolean => {\n if (left.size !== right.size) return false;\n\n for (const leftItem of left) {\n if (isPrimitive(leftItem)) {\n if (!right.has(leftItem)) return false;\n } else {\n let found = false;\n for (const rightItem of right) {\n if (deepEquals(leftItem, rightItem)) {\n found = true;\n break;\n }\n }\n if (!found) return false;\n }\n }\n return true;\n};\n\nconst compareArrayBuffers = (\n left: ArrayBuffer,\n right: ArrayBuffer,\n): boolean => {\n if (left.byteLength !== right.byteLength) return false;\n const leftView = new Uint8Array(left);\n const rightView = new Uint8Array(right);\n for (let i = 0; i < leftView.length; i++) {\n if (leftView[i] !== rightView[i]) return false;\n }\n return true;\n};\n\nconst compareTypedArrays = (\n left: ArrayBufferView,\n right: ArrayBufferView,\n): boolean => {\n if (left.constructor !== right.constructor) return false;\n if (left.byteLength !== right.byteLength) return false;\n\n const leftArray = new Uint8Array(\n left.buffer,\n left.byteOffset,\n left.byteLength,\n );\n const rightArray = new Uint8Array(\n right.buffer,\n right.byteOffset,\n right.byteLength,\n );\n\n for (let i = 0; i < leftArray.length; i++) {\n if (leftArray[i] !== rightArray[i]) return false;\n }\n return true;\n};\n\nconst compareObjects = (\n left: Record<string, unknown>,\n right: Record<string, unknown>,\n): boolean => {\n const keys1 = Object.keys(left);\n const keys2 = Object.keys(right);\n\n if (keys1.length !== keys2.length) {\n return false;\n }\n\n for (const key of keys1) {\n if (left[key] instanceof Function && right[key] instanceof Function) {\n continue;\n }\n\n const isEqual = deepEquals(left[key], right[key]);\n if (!isEqual) {\n return false;\n }\n }\n\n return true;\n};\n\nconst getType = (value: unknown): string => {\n if (value === null) return 'null';\n if (value === undefined) return 'undefined';\n\n const primitiveType = typeof value;\n if (primitiveType !== 'object') return primitiveType;\n\n if (Array.isArray(value)) return 'array';\n if (value instanceof Boolean) return 'boxed-boolean';\n if (value instanceof Number) return 'boxed-number';\n if (value instanceof String) return 'boxed-string';\n if (value instanceof Date) return 'date';\n if (value instanceof RegExp) return 'regexp';\n if (value instanceof Error) return 'error';\n if (value instanceof Map) return 'map';\n if (value instanceof Set) return 'set';\n if (value instanceof ArrayBuffer) return 'arraybuffer';\n if (value instanceof DataView) return 'dataview';\n if (value instanceof WeakMap) return 'weakmap';\n if (value instanceof WeakSet) return 'weakset';\n\n if (ArrayBuffer.isView(value)) return 'typedarray';\n\n return 'object';\n};\n\nexport const deepEquals = <T>(left: T, right: T): boolean => {\n if (left === right) return true;\n\n if (isEquatable(left)) {\n return left.equals(right);\n }\n\n const leftType = getType(left);\n const rightType = getType(right);\n\n if (leftType !== rightType) return false;\n\n switch (leftType) {\n case 'null':\n case 'undefined':\n case 'boolean':\n case 'number':\n case 'bigint':\n case 'string':\n case 'symbol':\n case 'function':\n return left === right;\n\n case 'array':\n return compareArrays(left as unknown[], right as unknown[]);\n\n case 'date':\n return compareDates(left as Date, right as Date);\n\n case 'regexp':\n return compareRegExps(left as RegExp, right as RegExp);\n\n case 'error':\n return compareErrors(left as Error, right as Error);\n\n case 'map':\n return compareMaps(\n left as Map<unknown, unknown>,\n right as Map<unknown, unknown>,\n );\n\n case 'set':\n return compareSets(left as Set<unknown>, right as Set<unknown>);\n\n case 'arraybuffer':\n return compareArrayBuffers(left as ArrayBuffer, right as ArrayBuffer);\n\n case 'dataview':\n case 'weakmap':\n case 'weakset':\n return false;\n\n case 'typedarray':\n return compareTypedArrays(\n left as ArrayBufferView,\n right as ArrayBufferView,\n );\n\n case 'boxed-boolean':\n return (left as boolean).valueOf() === (right as boolean).valueOf();\n\n case 'boxed-number':\n return (left as number).valueOf() === (right as number).valueOf();\n\n case 'boxed-string':\n return (left as string).valueOf() === (right as string).valueOf();\n\n case 'object':\n return compareObjects(\n left as Record<string, unknown>,\n right as Record<string, unknown>,\n );\n\n default:\n return false;\n }\n};\n\nexport type Equatable<T> = { equals: (right: T) => boolean } & T;\n\nexport const isEquatable = <T>(left: T): left is Equatable<T> => {\n return (\n left !== null &&\n left !== undefined &&\n typeof left === 'object' &&\n 'equals' in left &&\n typeof left['equals'] === 'function'\n );\n};\n","export class ParseError extends Error {\n constructor(text: string) {\n super(`Cannot parse! ${text}`);\n }\n}\n\nexport type Mapper<From, To = From> =\n | ((value: unknown) => To)\n | ((value: Partial<From>) => To)\n | ((value: From) => To)\n | ((value: Partial<To>) => To)\n | ((value: To) => To)\n | ((value: Partial<To | From>) => To)\n | ((value: To | From) => To);\n\nexport type MapperArgs<From, To = From> = Partial<From> &\n From &\n Partial<To> &\n To;\n\nexport type ParseOptions<From, To = From> = {\n reviver?: (key: string, value: unknown) => unknown;\n map?: Mapper<From, To>;\n typeCheck?: <To>(value: unknown) => value is To;\n};\n\nexport type StringifyOptions<From, To = From> = {\n map?: Mapper<From, To>;\n};\n\nexport const JSONParser = {\n stringify: <From, To = From>(\n value: From,\n options?: StringifyOptions<From, To>,\n ) => {\n return JSON.stringify(\n options?.map ? options.map(value as MapperArgs<From, To>) : value,\n //TODO: Consider adding support to DateTime and adding specific format to mark that's a bigint\n // eslint-disable-next-line @typescript-eslint/no-unsafe-return\n (_, v) => (typeof v === 'bigint' ? v.toString() : v),\n );\n },\n parse: <From, To = From>(\n text: string,\n options?: ParseOptions<From, To>,\n ): To | undefined => {\n const parsed: unknown = JSON.parse(text, options?.reviver);\n\n if (options?.typeCheck && !options?.typeCheck<To>(parsed))\n throw new ParseError(text);\n\n return options?.map\n ? options.map(parsed as MapperArgs<From, To>)\n : (parsed as To | undefined);\n },\n};\n","const textEncoder = new TextEncoder();\n\nexport const hashText = async (text: string): Promise<bigint> => {\n const hashBuffer = await crypto.subtle.digest(\n 'SHA-256',\n textEncoder.encode(text),\n );\n // Create an array with a single element that is a 64-bit signed integer\n // We take the first 8 bytes (so 64 bits) of the SHA-256 hash\n const view = new BigInt64Array(hashBuffer, 0, 1);\n return view[0]!;\n};\n","import { JSONParser } from '../serialization';\nimport type { DefaultRecord } from '../typing';\nimport { deepEquals } from '../utils';\n\nexport class AssertionError extends Error {\n constructor(message: string) {\n super(message);\n }\n}\n\nexport const isSubset = (superObj: unknown, subObj: unknown): boolean => {\n const sup = superObj as DefaultRecord;\n const sub = subObj as DefaultRecord;\n\n assertOk(sup);\n assertOk(sub);\n\n return Object.keys(sub).every((ele: string) => {\n if (sub[ele] !== null && typeof sub[ele] == 'object') {\n return isSubset(sup[ele], sub[ele]);\n }\n return sub[ele] === sup[ele];\n });\n};\n\nexport const assertFails = (message?: string) => {\n throw new AssertionError(message ?? 'That should not ever happened, right?');\n};\n\nexport const assertThrowsAsync = async <TError extends Error>(\n fun: () => Promise<unknown>,\n errorCheck?: (error: Error) => boolean,\n): Promise<TError> => {\n try {\n await fun();\n } catch (error) {\n const typedError = error as TError;\n if (typedError instanceof AssertionError || !errorCheck) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n return typedError;\n }\n\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertThrows = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError => {\n try {\n fun();\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertTrue(\n errorCheck(typedError),\n `Error doesn't match the expected condition: ${JSONParser.stringify(error)}`,\n );\n } else if (typedError instanceof AssertionError) {\n assertFalse(\n typedError instanceof AssertionError,\n \"Function didn't throw expected error\",\n );\n }\n\n return typedError;\n }\n throw new AssertionError(\"Function didn't throw expected error\");\n};\n\nexport const assertDoesNotThrow = <TError extends Error>(\n fun: () => void,\n errorCheck?: (error: Error) => boolean,\n): TError | null => {\n try {\n fun();\n return null;\n } catch (error) {\n const typedError = error as TError;\n\n if (errorCheck) {\n assertFalse(\n errorCheck(typedError),\n `Error matching the expected condition was thrown!: ${JSONParser.stringify(error)}`,\n );\n } else {\n assertFails(`Function threw an error: ${JSONParser.stringify(error)}`);\n }\n\n return typedError;\n }\n};\n\nexport const assertRejects = async <T, TError extends Error = Error>(\n promise: Promise<T>,\n errorCheck?: ((error: TError) => boolean) | TError,\n) => {\n try {\n await promise;\n throw new AssertionError(\"Function didn't throw expected error\");\n } catch (error) {\n if (!errorCheck) return;\n\n if (errorCheck instanceof Error) assertDeepEqual(error, errorCheck);\n else assertTrue(errorCheck(error as TError));\n }\n};\n\nexport const assertMatches = (\n actual: unknown,\n expected: unknown,\n message?: string,\n) => {\n if (!isSubset(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not subset of\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (!deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis not equal to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertNotDeepEqual = <T = unknown>(\n actual: T,\n expected: T,\n message?: string,\n) => {\n if (deepEquals(actual, expected))\n throw new AssertionError(\n message ??\n `subObj:\\n${JSONParser.stringify(expected)}\\nis equals to\\n${JSONParser.stringify(actual)}`,\n );\n};\n\nexport const assertThat = <T>(item: T) => {\n return {\n isEqualTo: (other: T) => assertTrue(deepEquals(item, other)),\n };\n};\n\nexport const assertDefined = (\n value: unknown,\n message?: string | Error,\n): asserts value => {\n assertOk(value, message instanceof Error ? message.message : message);\n};\n\nexport function assertFalse(\n condition: boolean,\n message?: string,\n): asserts condition is false {\n if (condition !== false)\n throw new AssertionError(message ?? `Condition is true`);\n}\n\nexport function assertTrue(\n condition: boolean,\n message?: string,\n): asserts condition is true {\n if (condition !== true)\n throw new AssertionError(message ?? `Condition is false`);\n}\n\n// TODO: replace with assertDefined\nexport function assertOk<T>(\n obj: T | null | undefined,\n message?: string,\n): asserts obj is T {\n if (!obj) throw new AssertionError(message ?? `Condition is not truthy`);\n}\n\nexport function assertEqual<T>(\n expected: T | null | undefined,\n actual: T | null | undefined,\n message?: string,\n): void {\n if (expected !== actual)\n throw new AssertionError(\n `${message ?? 'Objects are not equal'}:\\nExpected: ${JSONParser.stringify(expected)}\\nActual: ${JSONParser.stringify(actual)}`,\n );\n}\n\nexport function assertNotEqual<T>(\n obj: T | null | undefined,\n other: T | null | undefined,\n message?: string,\n): void {\n if (obj === other)\n throw new AssertionError(\n message ?? `Objects are equal: ${JSONParser.stringify(obj)}`,\n );\n}\n\nexport function assertIsNotNull<T extends object | bigint>(\n result: T | null,\n): asserts result is T {\n assertNotEqual(result, null);\n assertOk(result);\n}\n\nexport function assertIsNull<T extends object>(\n result: T | null,\n): asserts result is null {\n assertEqual(result, null);\n}\n\ntype Call = {\n arguments: unknown[];\n result: unknown;\n target: unknown;\n this: unknown;\n};\n\nexport type ArgumentMatcher = (arg: unknown) => boolean;\n\nexport const argValue =\n <T>(value: T): ArgumentMatcher =>\n (arg) =>\n deepEquals(arg, value);\n\nexport const argMatches =\n <T>(matches: (arg: T) => boolean): ArgumentMatcher =>\n (arg) =>\n matches(arg as T);\n\n// eslint-disable-next-line @typescript-eslint/no-unsafe-function-type\nexport type MockedFunction = Function & { mock?: { calls: Call[] } };\n\nexport function verifyThat(fn: MockedFunction) {\n return {\n calledTimes: (times: number) => {\n assertEqual(fn.mock?.calls?.length, times);\n },\n notCalled: () => {\n assertEqual(fn?.mock?.calls?.length, 0);\n },\n called: () => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length > 0,\n );\n },\n calledWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledOnceWith: (...args: unknown[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length === 1 &&\n fn.mock.calls.some((call) => deepEquals(call.arguments, args)),\n );\n },\n calledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertTrue(\n fn.mock?.calls.length !== undefined && fn.mock.calls.length >= 1,\n );\n assertTrue(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls.some(\n (call) =>\n call.arguments &&\n call.arguments.length >= matches.length &&\n matches.every((match, index) => match(call.arguments[index])),\n ),\n );\n },\n notCalledWithArgumentMatching: (...matches: ArgumentMatcher[]) => {\n assertFalse(\n fn.mock?.calls.length !== undefined &&\n fn.mock.calls.length >= 1 &&\n fn.mock.calls[0]!.arguments &&\n fn.mock.calls[0]!.arguments.length >= matches.length &&\n matches.every((match, index) =>\n match(fn.mock!.calls[0]!.arguments[index]),\n ),\n );\n },\n };\n}\n\nexport const assertThatArray = <T>(array: T[]) => {\n return {\n isEmpty: () =>\n assertEqual(\n array.length,\n 0,\n `Array is not empty ${JSONParser.stringify(array)}`,\n ),\n isNotEmpty: () => assertNotEqual(array.length, 0, `Array is empty`),\n hasSize: (length: number) => assertEqual(array.length, length),\n containsElements: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => deepEquals(ts, o))));\n },\n containsElementsMatching: (other: T[]) => {\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsOnlyElementsMatching: (other: T[]) => {\n assertEqual(array.length, other.length, `Arrays lengths don't match`);\n assertTrue(other.every((ts) => array.some((o) => isSubset(o, ts))));\n },\n containsExactlyInAnyOrder: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyInAnyOrderElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n assertTrue(array.every((ts) => other.some((o) => deepEquals(ts, o))));\n },\n containsExactlyElementsOf: (other: T[]) => {\n assertEqual(array.length, other.length);\n for (let i = 0; i < array.length; i++) {\n assertTrue(deepEquals(array[i], other[i]));\n }\n },\n containsExactly: (elem: T) => {\n assertEqual(array.length, 1);\n assertTrue(deepEquals(array[0], elem));\n },\n contains: (elem: T) => {\n assertTrue(array.some((a) => deepEquals(a, elem)));\n },\n containsOnlyOnceElementsOf: (other: T[]) => {\n assertTrue(\n other\n .map((o) => array.filter((a) => deepEquals(a, o)).length)\n .filter((a) => a === 1).length === other.length,\n );\n },\n containsAnyOf: (other: T[]) => {\n assertTrue(array.some((a) => other.some((o) => deepEquals(a, o))));\n },\n allMatch: (matches: (item: T) => boolean) => {\n assertTrue(array.every(matches));\n },\n anyMatches: (matches: (item: T) => boolean) => {\n assertTrue(array.some(matches));\n },\n allMatchAsync: async (\n matches: (item: T) => Promise<boolean>,\n ): Promise<void> => {\n for (const item of array) {\n assertTrue(await matches(item));\n }\n },\n };\n};\n","import type {\n AnyMessage,\n AnyRecordedMessageMetadata,\n RecordedMessage,\n} from '../../typing';\n\nexport type MessageDowncast<\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n> =\n | ((\n message: RecordedMessage<MessageType, RecordedMessageMetadataType>,\n ) => RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>)\n | ((message: MessageType) => MessagePayloadType);\n\nexport const downcastRecordedMessage = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessage:\n | RecordedMessage<MessageType, RecordedMessageMetadataType>\n | MessageType,\n options?: {\n downcast?: MessageDowncast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessagePayloadType, RecordedMessageMetadataType> => {\n if (!options?.downcast)\n return recordedMessage as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n\n const downcasted = options.downcast(\n recordedMessage as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >,\n );\n\n return {\n ...recordedMessage,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n data: downcasted.data,\n ...('metadata' in recordedMessage || 'metadata' in downcasted\n ? {\n metadata: {\n ...('metadata' in recordedMessage\n ? (recordedMessage.metadata as object)\n : {}),\n ...('metadata' in downcasted\n ? (downcasted.metadata as object)\n : {}),\n },\n }\n : {}),\n } as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n};\n\nexport const downcastRecordedMessages = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessages:\n | RecordedMessage<MessageType, RecordedMessageMetadataType>[]\n | MessageType[],\n options?: {\n downcast?: MessageDowncast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>[] => {\n if (!options?.downcast)\n return recordedMessages as unknown as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >[];\n\n return recordedMessages.map((recordedMessage) =>\n downcastRecordedMessage(recordedMessage, options),\n );\n};\n","import type {\n AnyMessage,\n AnyRecordedMessageMetadata,\n RecordedMessage,\n} from '../../typing';\n\nexport type MessageUpcast<\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n> =\n | ((message: MessagePayloadType) => MessageType)\n | ((\n message: RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>,\n ) => RecordedMessage<MessageType, RecordedMessageMetadataType>);\n\nexport const upcastRecordedMessage = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessage:\n | RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>\n | MessagePayloadType,\n options?: {\n upcast?: MessageUpcast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessageType, RecordedMessageMetadataType> => {\n if (!options?.upcast)\n return recordedMessage as unknown as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >;\n\n const upcasted = options.upcast(\n recordedMessage as RecordedMessage<\n MessagePayloadType,\n RecordedMessageMetadataType\n >,\n );\n\n return {\n ...recordedMessage,\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n data: upcasted.data,\n ...('metadata' in recordedMessage || 'metadata' in upcasted\n ? {\n metadata: {\n ...('metadata' in recordedMessage\n ? (recordedMessage.metadata as object)\n : {}),\n ...('metadata' in upcasted ? (upcasted.metadata as object) : {}),\n },\n }\n : {}),\n } as unknown as RecordedMessage<MessageType, RecordedMessageMetadataType>;\n};\n\nexport const upcastRecordedMessages = <\n MessageType extends AnyMessage,\n MessagePayloadType extends AnyMessage = MessageType,\n RecordedMessageMetadataType extends AnyRecordedMessageMetadata =\n AnyRecordedMessageMetadata,\n>(\n recordedMessages:\n | RecordedMessage<MessagePayloadType, RecordedMessageMetadataType>[]\n | MessagePayloadType[],\n options?: {\n upcast?: MessageUpcast<\n MessageType,\n MessagePayloadType,\n RecordedMessageMetadataType\n >;\n },\n): RecordedMessage<MessageType, RecordedMessageMetadataType>[] => {\n if (!options?.upcast)\n return recordedMessages as unknown as RecordedMessage<\n MessageType,\n RecordedMessageMetadataType\n >[];\n\n return recordedMessages.map((recordedMessage) =>\n upcastRecordedMessage(recordedMessage, options),\n );\n};\n","import { EmmettError } from '../errors';\nimport type { EventStoreReadSchemaOptions } from '../eventStore';\nimport { JSONParser } from '../serialization';\nimport type {\n AnyEvent,\n AnyReadEventMetadata,\n BatchRecordedMessageHandlerWithContext,\n CanHandle,\n DefaultRecord,\n Event,\n} from '../typing';\nimport { arrayUtils } from '../utils';\n\nexport type ProjectionHandlingType = 'inline' | 'async';\n\nexport type ProjectionHandler<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = BatchRecordedMessageHandlerWithContext<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n>;\n\nexport type TruncateProjection<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = (context: ProjectionHandlerContext) => Promise<void>;\n\nexport type ProjectionInitOptions<\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n version: number;\n status?: 'active' | 'inactive';\n registrationType: ProjectionHandlingType;\n context: ProjectionHandlerContext;\n};\n\nexport interface ProjectionDefinition<\n EventType extends Event = AnyEvent,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n> {\n name?: string;\n version?: number;\n kind?: string;\n canHandle: CanHandle<EventType>;\n handle: ProjectionHandler<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext\n >;\n truncate?: TruncateProjection<ProjectionHandlerContext>;\n init?: (\n options: ProjectionInitOptions<ProjectionHandlerContext>,\n ) => void | Promise<void>;\n eventsOptions?: {\n schema?: EventStoreReadSchemaOptions<EventType, EventPayloadType>;\n };\n}\n\nexport type ProjectionRegistration<\n HandlingType extends ProjectionHandlingType,\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n> = {\n type: HandlingType;\n projection: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext,\n AnyEvent\n >;\n};\n\nexport const filterProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n type: ProjectionHandlingType,\n projections: ProjectionRegistration<\n ProjectionHandlingType,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n) => {\n const inlineProjections = projections\n .filter((projection) => projection.type === type)\n .map(({ projection }) => projection);\n\n const duplicateRegistrations = arrayUtils.getDuplicates(\n inlineProjections,\n (proj) => proj.name,\n );\n\n if (duplicateRegistrations.length > 0) {\n throw new EmmettError(`You cannot register multiple projections with the same name (or without the name).\n Ensure that:\n ${JSONParser.stringify(duplicateRegistrations)}\n have different names`);\n }\n\n return inlineProjections;\n};\n\nexport const projection = <\n EventType extends Event = Event,\n EventMetaDataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n EventPayloadType extends Event = EventType,\n>(\n definition: ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext,\n EventPayloadType\n >,\n): ProjectionDefinition<\n EventType,\n EventMetaDataType,\n ProjectionHandlerContext,\n EventPayloadType\n> => definition;\n\nexport const inlineProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n any,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const asyncProjections = <\n ReadEventMetadataType extends AnyReadEventMetadata = AnyReadEventMetadata,\n ProjectionHandlerContext extends DefaultRecord = DefaultRecord,\n>(\n definitions: ProjectionDefinition<\n AnyEvent,\n ReadEventMetadataType,\n ProjectionHandlerContext\n >[],\n): ProjectionRegistration<\n 'inline',\n ReadEventMetadataType,\n ProjectionHandlerContext\n>[] =>\n definitions.map((definition) => ({\n type: 'inline',\n projection: definition,\n }));\n\nexport const projections = {\n inline: inlineProjections,\n async: asyncProjections,\n};\n","import {\n assertExpectedVersionMatchesCurrent,\n downcastRecordedMessage,\n ExpectedVersionConflictError,\n filterProjections,\n tryPublishMessagesAfterCommit,\n upcastRecordedMessages,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type Closeable,\n type DefaultEventStoreOptions,\n type Event,\n type EventStore,\n type ProjectionRegistration,\n type ReadEvent,\n type ReadEventMetadataWithoutGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n type StreamExistsResult,\n} from '@event-driven-io/emmett';\nimport {\n MongoClient,\n type Collection,\n type Document,\n type Filter,\n type MongoClientOptions,\n type Sort,\n type UpdateFilter,\n type WithId,\n} from 'mongodb';\nimport { v4 as uuid } from 'uuid';\nimport {\n handleInlineProjections,\n MongoDBDefaultInlineProjectionName,\n type MongoDBInlineProjectionDefinition,\n type MongoDBProjectionInlineHandlerContext,\n} from './projections';\nimport {\n mongoDBEventStoreStorage,\n type MongoDBEventStoreStorage,\n type MongoDBEventStoreStorageOptions,\n} from './storage';\n\nexport const MongoDBEventStoreDefaultStreamVersion = 0n;\n\nexport type StreamType = string;\nexport type StreamName<T extends StreamType = StreamType> = `${T}:${string}`;\n\nexport type StreamNameParts<T extends StreamType = StreamType> = {\n streamType: T;\n streamId: string;\n};\n\nexport type StreamCollectionName<T extends StreamType = StreamType> =\n `emt:${T}`;\n\nexport type StreamCollectionNameParts<T extends StreamType = StreamType> = {\n streamType: T;\n};\n\nexport type MongoDBReadModelMetadata = {\n streamId: string;\n name: string;\n schemaVersion: number;\n streamPosition: bigint;\n};\n\nexport type MongoDBReadModel<Doc extends Document = Document> = Doc & {\n _metadata: MongoDBReadModelMetadata;\n};\n\nexport interface EventStream<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> {\n streamName: string;\n messages: Array<ReadEvent<EventType, EventMetaDataType>>;\n metadata: {\n streamId: string;\n streamType: StreamType;\n streamPosition: bigint;\n createdAt: Date;\n updatedAt: Date;\n };\n projections: Record<string, MongoDBReadModel>;\n}\n\nexport type MongoDBReadEventMetadata = ReadEventMetadataWithoutGlobalPosition;\n\nexport type MongoDBReadEvent<EventType extends Event = Event> = ReadEvent<\n EventType,\n MongoDBReadEventMetadata\n>;\n\ntype SingleProjectionQueryStreamFilter<T extends StreamType> = {\n projectionName?: string;\n} & ({ streamName: StreamName<T> } | { streamType: T; streamId?: string });\n\ntype MultiProjectionQueryStreamFilter<T extends StreamType> = {\n projectionName?: string;\n} & (\n | { streamNames: StreamName<T>[] }\n | { streamType: T; streamIds?: string[] }\n);\n\ntype MultiProjectionQueryOptions = {\n skip?: number;\n limit?: number;\n sort?: [string, 1 | -1][] | Record<string, 1 | -1>;\n};\n\n/**\n * Helpers for querying inline projections on event streams.\n */\ntype InlineProjectionQueries<T extends StreamType> = {\n /**\n * Helper for querying for a single projection. Similar to `collection.findOne`.\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamName` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n */\n findOne: <Doc extends Document>(\n streamFilter: SingleProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) => Promise<MongoDBReadModel<Doc> | null>;\n /**\n * Helper for querying for multiple projections. Similar to `collection.find`.\n *\n * ***NOTE***: If `streamFilter.streamNames` is an empty array, this function will return an empty array. If `streamFilter.streamIds` is an empty array, the `streamIds` filter will not be used.\n *\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamNames` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n * @param queryOptions - Additional query options like `skip`, `limit`, and `sort`. `sort`, similar to `projectionQuery`, will prepend each object key with the necessary projection name.\n */\n find: <Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n queryOptions?: MultiProjectionQueryOptions,\n ) => Promise<MongoDBReadModel<Doc>[]>;\n /**\n * Returns the total number of documents matching the provided filter options. Similar to `collection.countDocuments`.\n *\n * ***NOTE***: If `streamFilter.streamNames` is an empty array, this function will return `0`. If `streamFilter.streamIds` is an empty array, the `streamIds` filter will not be used.\n *\n * @param streamFilter - A filter object for stream level fields. If `streamType` is required if `streamNames` is not provided. If `projectionName` is not provided, the default projection will be used (`MongoDBDefaultInlineProjectionName`).\n * @param projectionQuery - A MongoDB filter query based on the projection schema. Internally, this function will prepend each object key with the necessary projection name.\n */\n count: <Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) => Promise<number>;\n};\n\n/**\n * Helpers for querying projections on event streams.\n */\ntype ProjectionQueries<T extends StreamType> = {\n inline: InlineProjectionQueries<T>;\n};\n\nexport type MongoDBEventStoreClientOptions = {\n client: MongoClient;\n connectionString?: never;\n clientOptions?: never;\n};\n\nexport type MongoDBEventStoreConnectionStringOptions = {\n client?: never;\n connectionString: string;\n clientOptions?: MongoClientOptions;\n};\n\nexport type MongoDBEventStoreConnectionOptions =\n | MongoDBEventStoreClientOptions\n | MongoDBEventStoreConnectionStringOptions;\n\nexport type MongoDBEventStoreOptions = {\n projections?: ProjectionRegistration<\n 'inline',\n MongoDBReadEventMetadata,\n MongoDBProjectionInlineHandlerContext\n >[];\n storage?: MongoDBEventStoreStorageOptions;\n} & MongoDBEventStoreConnectionOptions &\n DefaultEventStoreOptions<MongoDBEventStore>;\n\nexport type MongoDBEventStore = EventStore<MongoDBReadEventMetadata> & {\n projections: ProjectionQueries<StreamType>;\n collectionFor: <EventType extends Event>(\n streamType: StreamType,\n ) => Promise<Collection<EventStream<EventType>>>;\n};\n\nclass MongoDBEventStoreImplementation implements MongoDBEventStore, Closeable {\n private readonly client: MongoClient;\n private readonly inlineProjections: MongoDBInlineProjectionDefinition[];\n private shouldManageClientLifetime: boolean;\n private isClosed: boolean = false;\n private storage: MongoDBEventStoreStorage;\n private options: MongoDBEventStoreOptions;\n public projections: ProjectionQueries<StreamType>;\n\n constructor(options: MongoDBEventStoreOptions) {\n this.options = options;\n this.client =\n 'client' in options && options.client\n ? options.client\n : new MongoClient(options.connectionString, options.clientOptions);\n this.shouldManageClientLifetime = !('client' in options);\n this.storage = mongoDBEventStoreStorage({\n storage: options.storage,\n getConnectedClient: () => this.getConnectedClient(),\n });\n this.inlineProjections = filterProjections(\n 'inline',\n options.projections ?? [],\n ) as MongoDBInlineProjectionDefinition[];\n\n this.projections = {\n inline: {\n findOne: this.findOneInlineProjection.bind(this),\n find: this.findInlineProjection.bind(this),\n count: this.countInlineProjection.bind(this),\n },\n };\n }\n\n async readStream<\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n options?: ReadStreamOptions<EventType, EventPayloadType>,\n ): Promise<\n Exclude<ReadStreamResult<EventType, MongoDBReadEventMetadata>, null>\n > {\n const { streamType } = fromStreamName(streamName);\n const expectedStreamVersion = options?.expectedStreamVersion;\n\n const collection = await this.storage.collectionFor(streamType);\n\n const filter = {\n streamName: { $eq: streamName },\n };\n\n const eventsSliceArr: number[] = [];\n\n if (options && 'from' in options) {\n eventsSliceArr.push(Number(options.from));\n } else {\n eventsSliceArr.push(0);\n }\n\n if (options && 'to' in options) {\n eventsSliceArr.push(Number(options.to));\n }\n\n const eventsSlice =\n eventsSliceArr.length > 1 ? { $slice: eventsSliceArr } : 1;\n\n const stream = await collection.findOne<\n WithId<Pick<EventStream<EventPayloadType>, 'metadata' | 'messages'>>\n >(filter, {\n useBigInt64: true,\n projection: {\n metadata: 1,\n messages: eventsSlice,\n },\n });\n\n if (!stream) {\n return {\n events: [],\n currentStreamVersion: MongoDBEventStoreDefaultStreamVersion,\n streamExists: false,\n };\n }\n\n assertExpectedVersionMatchesCurrent(\n stream.metadata.streamPosition,\n expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n const events = upcastRecordedMessages(\n stream.messages,\n options?.schema?.versioning,\n );\n\n return {\n events,\n currentStreamVersion: stream.metadata.streamPosition,\n streamExists: true,\n };\n }\n\n async aggregateStream<\n State,\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n options: AggregateStreamOptions<\n State,\n EventType,\n MongoDBReadEventMetadata,\n EventPayloadType\n >,\n ): Promise<AggregateStreamResult<State>> {\n const stream = await this.readStream<EventType, EventPayloadType>(\n streamName,\n options?.read,\n );\n const { evolve, initialState } = options;\n\n const state = stream.events.reduce(evolve, initialState());\n return {\n state,\n currentStreamVersion: stream.currentStreamVersion,\n streamExists: stream.streamExists,\n };\n }\n\n async appendToStream<\n EventType extends Event,\n EventPayloadType extends Event = EventType,\n >(\n streamName: StreamName,\n events: EventType[],\n options?: AppendToStreamOptions<EventType, EventPayloadType>,\n ): Promise<AppendToStreamResult> {\n const { streamId, streamType } = fromStreamName(streamName);\n const expectedStreamVersion = options?.expectedStreamVersion;\n\n const collection = await this.storage.collectionFor(streamType);\n\n const stream = await collection.findOne<\n WithId<Pick<EventStream<EventPayloadType>, 'metadata' | 'projections'>>\n >(\n { streamName: { $eq: streamName } },\n {\n useBigInt64: true,\n projection: {\n 'metadata.streamPosition': 1,\n projections: 1,\n },\n },\n );\n\n const currentStreamVersion =\n stream?.metadata.streamPosition ?? MongoDBEventStoreDefaultStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n MongoDBEventStoreDefaultStreamVersion,\n );\n\n let streamOffset = currentStreamVersion;\n\n const eventsToAppend: ReadEvent<\n EventPayloadType,\n MongoDBReadEventMetadata\n >[] = events.map((event) => {\n const metadata: MongoDBReadEventMetadata = {\n messageId: uuid(),\n streamName,\n streamPosition: ++streamOffset,\n };\n return downcastRecordedMessage(\n {\n type: event.type,\n data: event.data,\n metadata: {\n ...metadata,\n ...('metadata' in event ? (event.metadata ?? {}) : {}),\n },\n } as ReadEvent<EventType, MongoDBReadEventMetadata>,\n options?.schema?.versioning,\n );\n });\n\n const now = new Date();\n const updates: UpdateFilter<EventStream> = {\n $push: { messages: { $each: eventsToAppend } },\n $set: {\n 'metadata.updatedAt': now,\n 'metadata.streamPosition': currentStreamVersion + BigInt(events.length),\n },\n $setOnInsert: {\n streamName,\n 'metadata.streamId': streamId,\n 'metadata.streamType': streamType,\n 'metadata.createdAt': now,\n },\n };\n\n if (this.inlineProjections) {\n await handleInlineProjections({\n readModels: stream?.projections ?? {},\n streamId,\n events: eventsToAppend,\n projections: this.inlineProjections,\n collection,\n updates,\n client: {},\n });\n }\n\n const updatedStream = await collection.updateOne(\n {\n streamName: { $eq: streamName },\n 'metadata.streamPosition': currentStreamVersion,\n },\n updates,\n { useBigInt64: true, upsert: true },\n );\n\n if (!updatedStream) {\n throw new ExpectedVersionConflictError(\n currentStreamVersion,\n options?.expectedStreamVersion ?? 0n,\n );\n }\n\n await tryPublishMessagesAfterCommit<MongoDBEventStore>(\n eventsToAppend,\n this.options.hooks,\n // {\n // TODO: same context as InlineProjectionHandlerContext for mongodb?\n // },\n );\n\n return {\n nextExpectedStreamVersion:\n currentStreamVersion + BigInt(eventsToAppend.length),\n createdNewStream:\n currentStreamVersion === MongoDBEventStoreDefaultStreamVersion,\n };\n }\n\n async streamExists(streamName: StreamName): Promise<StreamExistsResult> {\n const { streamType } = fromStreamName(streamName);\n\n const collection = await this.storage.collectionFor(streamType);\n\n const filter = {\n streamName: { $eq: streamName },\n };\n\n const count = await collection.countDocuments(filter, {\n useBigInt64: true,\n limit: 1,\n });\n\n return Boolean(count > 0);\n }\n\n collectionFor = async <EventType extends Event>(\n streamType: StreamType,\n ): Promise<Collection<EventStream<EventType>>> => {\n return this.storage.collectionFor(streamType);\n };\n\n /**\n * Gracefully cleans up managed resources by the MongoDBEventStore.\n * It closes MongoDB client created for the provided connection string\n * through event store options.\n *\n * @memberof Closeable\n */\n close = (): Promise<void> => {\n if (this.isClosed) return Promise.resolve();\n\n this.isClosed = true;\n if (!this.shouldManageClientLifetime) return Promise.resolve();\n\n return this.client.close();\n };\n\n private async findOneInlineProjection<Doc extends Document>(\n streamFilter: SingleProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) {\n const { projectionName, streamName, streamType } =\n parseSingleProjectionQueryStreamFilter(streamFilter);\n const collection = await this.storage.collectionFor(streamType);\n const query = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, `projections.${projectionName}`);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $exists: true } },\n ];\n\n if (query) {\n filters.push(query);\n }\n\n if (streamName) {\n filters.push({ streamName: { $eq: streamName } });\n }\n\n const result = await collection.findOne<{\n projections: Record<typeof projectionName, MongoDBReadModel<Doc>>;\n }>(\n { $and: filters },\n {\n useBigInt64: true,\n projection: { [`projections.${projectionName}`]: 1 },\n },\n );\n\n return result?.projections?.[projectionName] ?? null;\n }\n\n private async findInlineProjection<Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n queryOptions?: MultiProjectionQueryOptions,\n ) {\n const parsedStreamFilter =\n parseMultiProjectionQueryStreamFilter(streamFilter);\n if (!parsedStreamFilter) return [];\n const { projectionName, streamNames, streamType } = parsedStreamFilter;\n\n const collection = await this.storage.collectionFor(streamType);\n const prefix = `projections.${projectionName}`;\n const projectionFilter = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, prefix);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $ne: null } },\n ];\n\n if (projectionFilter) {\n filters.push(projectionFilter);\n }\n\n if (streamNames) {\n filters.push({ streamName: { $in: streamNames } });\n }\n\n let query = collection.find<\n EventStream & {\n projections: Record<typeof projectionName, MongoDBReadModel<Doc>>;\n }\n >(\n { $and: filters },\n {\n useBigInt64: true,\n projection: { [`projections.${projectionName}`]: 1 },\n },\n );\n\n if (queryOptions?.skip) {\n query = query.skip(queryOptions.skip);\n }\n\n if (queryOptions?.limit) {\n query = query.limit(queryOptions.limit);\n }\n\n if (queryOptions?.sort) {\n const sort = prependMongoFilterWithProjectionPrefix<Sort>(\n queryOptions.sort,\n prefix,\n );\n query = query.sort(sort);\n }\n\n const streams = await query.toArray();\n\n return streams\n .map((s) => s.projections[projectionName])\n .filter((p): p is MongoDBReadModel<Doc> => !!p);\n }\n\n private async countInlineProjection<Doc extends Document>(\n streamFilter: MultiProjectionQueryStreamFilter<StreamType>,\n projectionQuery?: Filter<MongoDBReadModel<Doc>>,\n ) {\n const parsedStreamFilter =\n parseMultiProjectionQueryStreamFilter(streamFilter);\n if (!parsedStreamFilter) return 0;\n const { projectionName, streamNames, streamType } = parsedStreamFilter;\n\n const collection = await this.storage.collectionFor(streamType);\n const prefix = `projections.${projectionName}`;\n const projectionFilter = prependMongoFilterWithProjectionPrefix<\n Filter<MongoDBReadModel<Doc>> | undefined,\n Filter<EventStream> | undefined\n >(projectionQuery, prefix);\n\n const filters: Filter<EventStream>[] = [\n { [`projections.${projectionName}`]: { $ne: null } },\n ];\n\n if (projectionFilter) {\n filters.push(projectionFilter);\n }\n\n if (streamNames) {\n filters.push({ streamName: { $in: streamNames } });\n }\n\n const total = await collection.countDocuments({ $and: filters });\n return total;\n }\n\n private getConnectedClient = async (): Promise<MongoClient> => {\n if (!this.isClosed) await this.client.connect();\n return this.client;\n };\n}\n\nfunction parseSingleProjectionQueryStreamFilter<\n T extends StreamType = StreamType,\n>(streamFilter: SingleProjectionQueryStreamFilter<T>) {\n const projectionName =\n streamFilter.projectionName ?? MongoDBDefaultInlineProjectionName;\n\n if ('streamName' in streamFilter) {\n const { streamType } = fromStreamName(streamFilter.streamName);\n return {\n projectionName,\n streamName: streamFilter.streamName,\n streamType,\n };\n }\n\n if (streamFilter.streamId) {\n const streamName = toStreamName(\n streamFilter.streamType,\n streamFilter.streamId,\n );\n return {\n projectionName,\n streamName,\n streamType: streamFilter.streamType,\n };\n }\n\n return {\n projectionName,\n streamType: streamFilter.streamType,\n };\n}\n\nfunction parseMultiProjectionQueryStreamFilter<T extends StreamType>(\n streamFilter: MultiProjectionQueryStreamFilter<T>,\n) {\n const projectionName =\n streamFilter.projectionName ?? MongoDBDefaultInlineProjectionName;\n\n if ('streamNames' in streamFilter) {\n if (streamFilter.streamNames.length == 0) return null;\n const { streamType } = fromStreamName(streamFilter.streamNames[0]!);\n return {\n projectionName,\n streamNames: streamFilter.streamNames,\n streamType,\n };\n }\n\n if (streamFilter.streamIds && streamFilter.streamIds.length > 0) {\n const streamNames = streamFilter.streamIds.map((id) =>\n toStreamName(streamFilter.streamType, id),\n );\n return {\n projectionName,\n streamNames,\n streamType: streamFilter.streamType,\n };\n }\n\n return {\n projectionName,\n streamType: streamFilter.streamType,\n };\n}\n\n/**\n * Prepends `prefix` to all object keys that don't start with a '$'\n */\nexport function prependMongoFilterWithProjectionPrefix<T, Result = T>(\n obj: T,\n prefix: string,\n): Result {\n if (typeof obj !== 'object' || obj === null || obj === undefined) {\n return obj as unknown as Result;\n }\n\n if (Array.isArray(obj)) {\n for (let i = 0; i < obj.length; i++) {\n // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment\n obj[i] = prependMongoFilterWithProjectionPrefix(obj[i], prefix);\n }\n return obj as unknown as Result;\n }\n\n for (const key in obj) {\n // @ts-expect-error we're forcing `k` to be a key of `T`\n const k: keyof typeof obj = addProjectionPrefixToMongoKey(key, prefix);\n if (k !== key) {\n obj[k] = obj[key as keyof typeof obj];\n delete obj[key as keyof typeof obj];\n }\n\n obj[k] = prependMongoFilterWithProjectionPrefix(obj[k], prefix);\n }\n\n return obj as unknown as Result;\n}\n\nfunction addProjectionPrefixToMongoKey(key: string, prefix: string): string {\n // MongoDB operators\n if (key[0] === '$') {\n return key;\n }\n\n return `${prefix}${key.length > 0 ? '.' : ''}${key}`;\n}\n\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions & { client: MongoClient },\n): MongoDBEventStore;\n\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions & { connectionString: string },\n): MongoDBEventStore & Closeable;\n\n// Implementation signature covers both, using a union for `options`\nexport function getMongoDBEventStore(\n options: MongoDBEventStoreOptions,\n): MongoDBEventStore | Closeable {\n const impl = new MongoDBEventStoreImplementation(options);\n\n // If a client is provided externally, we don't want to allow closing it\n if ('client' in options && 'close' in impl) {\n delete (impl as Partial<MongoDBEventStoreImplementation>).close;\n }\n\n return impl;\n}\n\n/**\n * Accepts a `streamType` (the type/category of the event stream) and an `streamId`\n * (the individual entity/object or aggregate ID) and combines them to a singular\n * `streamName` which can be used in `EventStore`.\n */\nexport function toStreamName<T extends StreamType>(\n streamType: T,\n streamId: string,\n): StreamName<T> {\n return `${streamType}:${streamId}`;\n}\n\n/**\n * Accepts a fully formatted `streamName` and returns the broken down\n * `streamType` and `streamId`.\n */\nexport function fromStreamName<T extends StreamType>(\n streamName: StreamName<T>,\n): StreamNameParts<T> {\n const parts = streamName.split(':') as [T, string];\n return {\n streamType: parts[0],\n streamId: parts[1],\n };\n}\n\n/**\n * Accepts a `streamType` (the type/category of the event stream)\n * and combines them to a `collectionName` which can be used in `EventStore`.\n */\nexport function toStreamCollectionName<T extends StreamType>(\n streamType: T,\n): StreamCollectionName<T> {\n return `emt:${streamType}`;\n}\n\n/**\n * Accepts a fully formatted `streamCollectionName` and returns the parsed `streamType`.\n */\nexport function fromStreamCollectionName<T extends StreamType>(\n streamCollectionName: StreamCollectionName<T>,\n): StreamCollectionNameParts<T> {\n const parts = streamCollectionName.split(':') as [string, T];\n return {\n streamType: parts[1],\n };\n}\n","import type {\n CanHandle,\n Event,\n ProjectionDefinition,\n ProjectionHandler,\n ReadEvent,\n} from '@event-driven-io/emmett';\nimport type { Collection, Document, UpdateFilter } from 'mongodb';\nimport type {\n EventStream,\n MongoDBReadEventMetadata,\n MongoDBReadModel,\n MongoDBReadModelMetadata,\n} from '../mongoDBEventStore';\n\nexport const MongoDBDefaultInlineProjectionName = '_default';\n\nexport type MongoDBProjectionInlineHandlerContext<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n document: MongoDBReadModel | null;\n streamId: string;\n updates: UpdateFilter<EventStream<EventType, EventMetaDataType>>;\n collection: Collection<EventStream<EventType, EventMetaDataType>>;\n};\n\nexport type MongoDBInlineProjectionHandler<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = ProjectionHandler<\n EventType,\n EventMetaDataType,\n MongoDBProjectionInlineHandlerContext\n>;\n\nexport type MongoDBInlineProjectionDefinition<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = ProjectionDefinition<\n EventType,\n EventMetaDataType,\n MongoDBProjectionInlineHandlerContext\n> & { name: string };\n\nexport type InlineProjectionHandlerOptions<\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n readModels: Record<string, MongoDBReadModel>;\n events: Array<ReadEvent<EventType, EventMetaDataType>>;\n projections: MongoDBInlineProjectionDefinition<\n EventType,\n EventMetaDataType\n >[];\n streamId: string;\n collection: Collection<EventStream>;\n updates: UpdateFilter<EventStream<Event>>;\n // eslint-disable-next-line @typescript-eslint/no-empty-object-type\n client: {\n //todo: add client here\n };\n};\n\nexport const handleInlineProjections = async <\n EventType extends Event = Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n>(\n options: InlineProjectionHandlerOptions<EventType, EventMetaDataType>,\n): Promise<void> => {\n const {\n events,\n projections: allProjections,\n updates: update,\n streamId,\n collection,\n readModels,\n } = options;\n\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n for (const projection of projections) {\n await projection.handle(events, {\n document: readModels[projection.name] ?? null,\n streamId,\n collection,\n updates: update,\n });\n }\n};\n\nexport type MongoDBWithNotNullDocumentEvolve<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> =\n | ((\n document: Doc,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Doc | null)\n | ((document: Doc, event: ReadEvent<EventType>) => Promise<Doc | null>);\n\nexport type MongoDBWithNullableDocumentEvolve<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> =\n | ((\n document: Doc | null,\n event: ReadEvent<EventType, EventMetaDataType>,\n ) => Doc | null)\n | ((\n document: Doc | null,\n event: ReadEvent<EventType>,\n ) => Promise<Doc | null>);\n\nexport type MongoDBInlineProjectionOptions<\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n> = {\n name?: string;\n schemaVersion?: number;\n canHandle: CanHandle<EventType>;\n} & (\n | {\n evolve: MongoDBWithNullableDocumentEvolve<\n Doc,\n EventType,\n EventMetaDataType\n >;\n }\n | {\n evolve: MongoDBWithNotNullDocumentEvolve<\n Doc,\n EventType,\n EventMetaDataType\n >;\n initialState: () => Doc;\n }\n);\n\nexport const mongoDBInlineProjection = <\n Doc extends Document,\n EventType extends Event,\n EventMetaDataType extends MongoDBReadEventMetadata = MongoDBReadEventMetadata,\n>(\n options: MongoDBInlineProjectionOptions<Doc, EventType, EventMetaDataType>,\n): MongoDBInlineProjectionDefinition => {\n const projectionName = options.name ?? MongoDBDefaultInlineProjectionName;\n const schemaVersion = options.schemaVersion ?? 1;\n\n return {\n name: projectionName,\n canHandle: options.canHandle,\n handle: async (events, { document, updates, streamId }) => {\n if (events.length === 0) return;\n\n let state =\n 'initialState' in options\n ? (document ?? options.initialState())\n : document;\n\n for (const event of events) {\n state = await options.evolve(\n state as Doc,\n event as ReadEvent<EventType, EventMetaDataType>,\n );\n }\n\n const metadata: MongoDBReadModelMetadata = {\n streamId,\n name: projectionName,\n schemaVersion,\n streamPosition: events[events.length - 1]!.metadata.streamPosition,\n };\n\n updates.$set![`projections.${projectionName}`] =\n state !== null\n ? {\n ...state,\n _metadata: metadata,\n }\n : null;\n },\n };\n};\n","import {\n assertFails,\n AssertionError,\n assertTrue,\n deepEquals,\n isErrorConstructor,\n isSubset,\n projections,\n type Event,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { MongoClient, type Document } from 'mongodb';\nimport {\n getMongoDBEventStore,\n type MongoDBEventStore,\n type MongoDBEventStoreConnectionOptions,\n type MongoDBReadModel,\n type StreamName,\n} from '../mongoDBEventStore';\nimport {\n MongoDBDefaultInlineProjectionName,\n type MongoDBInlineProjectionDefinition,\n} from './mongoDBInlineProjection';\n\nexport type MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType extends StreamName,\n EventType extends Event,\n> = {\n streamName: StreamNameType;\n events: EventType[];\n};\n\nexport type MongoDBInlineProjectionSpec<\n StreamNameType extends StreamName,\n EventType extends Event,\n> = (\n givenStream: MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType,\n EventType\n >,\n) => {\n when: (events: EventType[]) => {\n then: (\n assert: MongoDBInlineProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type MongoDBInlineProjectionAssertOptions<\n StreamNameType extends StreamName = StreamName,\n> = {\n streamName: StreamNameType;\n eventStore: MongoDBEventStore;\n};\n\nexport type MongoDBInlineProjectionAssert<\n StreamNameType extends StreamName = StreamName,\n> = (\n options: MongoDBInlineProjectionAssertOptions<StreamNameType>,\n) => Promise<void | boolean>;\n\nexport type MongoDBInlineProjectionSpecOptions = {\n projection: MongoDBInlineProjectionDefinition;\n} & MongoDBEventStoreConnectionOptions;\n\nexport const MongoDBInlineProjectionSpec = {\n for: <StreamNameType extends StreamName, EventType extends Event>(\n options: MongoDBInlineProjectionSpecOptions,\n ): MongoDBInlineProjectionSpec<StreamNameType, EventType> => {\n {\n const { projection, ...connectionOptions } = options;\n\n return (\n givenStream: MongoDBInlineProjectionSpecGivenEvents<\n StreamNameType,\n EventType\n >,\n ) => {\n const { streamName, events: givenEvents } = givenStream;\n return {\n when: (events: EventType[]) => {\n const allEvents = [...givenEvents, ...events];\n\n const run = (eventStore: MongoDBEventStore) =>\n eventStore.appendToStream(streamName, allEvents);\n\n return {\n then: async (\n assert: MongoDBInlineProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const client =\n 'client' in connectionOptions && connectionOptions.client\n ? connectionOptions.client\n : new MongoClient(\n connectionOptions.connectionString,\n connectionOptions.clientOptions,\n );\n\n const eventStore = getMongoDBEventStore({\n projections: projections.inline([projection]),\n client,\n });\n\n try {\n await run(eventStore);\n\n const succeeded = await assert({ eventStore, streamName });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await client.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const client =\n 'client' in connectionOptions && connectionOptions.client\n ? connectionOptions.client\n : new MongoClient(\n connectionOptions.connectionString,\n connectionOptions.clientOptions,\n );\n\n const eventStore = getMongoDBEventStore({\n projections: projections.inline([projection]),\n client,\n });\n\n try {\n await run(eventStore);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await client.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const eventInStream = <\n StreamNameType extends StreamName,\n EventType extends Event,\n>(\n streamName: StreamNameType,\n event: EventType,\n): MongoDBInlineProjectionSpecGivenEvents<StreamNameType, EventType> => ({\n streamName,\n events: [event],\n});\n\nexport const eventsInStream = <\n StreamNameType extends StreamName,\n EventType extends Event,\n>(\n streamName: StreamNameType,\n events: EventType[],\n): MongoDBInlineProjectionSpecGivenEvents<StreamNameType, EventType> => ({\n streamName,\n events,\n});\n\nconst expectReadModelToMatch = async <\n Doc extends Document = Document,\n StreamNameType extends StreamName = StreamName,\n>(\n options: MongoDBInlineProjectionAssertOptions<StreamNameType> & {\n projectionName: string;\n match: (readModel: MongoDBReadModel<Doc> | null) => boolean;\n },\n) => {\n const { streamName, projectionName, eventStore, match } = options;\n const readModel = await eventStore.projections.inline.findOne<Doc>({\n streamName,\n projectionName,\n });\n\n return match(readModel);\n};\n\nconst expectInlineReadModelWithName = (projectionName: string) => ({\n toHave:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n expected: Partial<MongoDBReadModel<Doc>> | null,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => isSubset(readModel, expected),\n }),\n toDeepEquals:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n expected: MongoDBReadModel<Doc> | null,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => deepEquals(readModel, expected),\n }),\n toMatch:\n <Doc extends Document, StreamNameType extends StreamName = StreamName>(\n match: (readModel: MongoDBReadModel<Doc> | null) => boolean,\n ): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch<Doc>({\n eventStore,\n streamName,\n projectionName,\n match,\n }),\n notToExist:\n <\n StreamNameType extends StreamName = StreamName,\n >(): MongoDBInlineProjectionAssert<StreamNameType> =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => readModel === null,\n }),\n toExist:\n (): MongoDBInlineProjectionAssert =>\n ({ eventStore, streamName }) =>\n expectReadModelToMatch({\n eventStore,\n streamName,\n projectionName,\n match: (readModel) => readModel !== null,\n }),\n});\n\nexport const expectInlineReadModel = {\n withName: (name: string) => expectInlineReadModelWithName(name),\n ...expectInlineReadModelWithName(MongoDBDefaultInlineProjectionName),\n};\n","import type { Event } from '@event-driven-io/emmett';\nimport type { Collection, Db, MongoClient } from 'mongodb';\nimport {\n toStreamCollectionName,\n type EventStream,\n type StreamType,\n} from '../mongoDBEventStore';\n\nexport type MongoDBEventStoreCollectionPerStreamTypeStorageOptions = {\n /**\n * The recommended setting where each stream type will be kept\n * in a separate collection type using the format: `emt_${streamType}`.\n */\n type: 'COLLECTION_PER_STREAM_TYPE';\n databaseName?: string;\n};\n\nexport type MongoDBEventStoreSingleCollectionStorageOptions = {\n /**\n * All streams will be kept withing a single MongDB collection\n * It'll either use default collection name (\"emt_streams\")\n * or provided name through 'collection' param.\n */\n type: 'SINGLE_COLLECTION';\n collectionName?: string;\n databaseName?: string;\n};\n\nexport type MongoDBEventStoreCollectionResolution = {\n databaseName?: string;\n collectionName: string;\n};\n\nexport type MongoDBEventStoreCustomStorageOptions = {\n /**\n * This is advanced option, where you specify your own collection\n * resolution function. You can do that by specifying the `collectionFor` function.\n */\n type: 'CUSTOM';\n databaseName?: string;\n collectionFor: <T extends StreamType>(\n streamType: T,\n ) => string | MongoDBEventStoreCollectionResolution;\n};\n\nexport type MongoDBEventStoreStorageOptions =\n | 'COLLECTION_PER_STREAM_TYPE'\n | 'SINGLE_COLLECTION'\n | MongoDBEventStoreSingleCollectionStorageOptions\n | MongoDBEventStoreCollectionPerStreamTypeStorageOptions\n | MongoDBEventStoreCustomStorageOptions;\n\nexport const DefaultMongoDBEventStoreStorageOptions =\n 'COLLECTION_PER_STREAM_TYPE';\n\nexport type MongoDBEventStoreStorage = {\n collectionFor: <T extends StreamType, EventType extends Event = Event>(\n streamType: T,\n ) => Promise<Collection<EventStream<EventType>>>;\n};\n\nexport const DefaultMongoDBEventStoreCollectionName = 'emt:streams';\n\nconst resolveCollectionAndDatabase = <T extends StreamType>(\n streamType: T,\n options: MongoDBEventStoreStorageOptions,\n): MongoDBEventStoreCollectionResolution => {\n if (\n options === 'SINGLE_COLLECTION' ||\n (typeof options === 'object' && options.type === 'SINGLE_COLLECTION')\n ) {\n return {\n collectionName:\n typeof options === 'object'\n ? (options.collectionName ?? DefaultMongoDBEventStoreCollectionName)\n : DefaultMongoDBEventStoreCollectionName,\n databaseName:\n typeof options === 'object' ? options.databaseName : undefined,\n };\n } else if (\n options === 'COLLECTION_PER_STREAM_TYPE' ||\n (typeof options === 'object' &&\n options.type === 'COLLECTION_PER_STREAM_TYPE')\n ) {\n return {\n collectionName: toStreamCollectionName(streamType),\n databaseName:\n typeof options === 'object' ? options.databaseName : undefined,\n };\n } else {\n const result = options.collectionFor(streamType);\n return {\n collectionName:\n typeof result === 'object' ? result.collectionName : result,\n databaseName:\n typeof result === 'object'\n ? (result.databaseName ?? options.databaseName)\n : options.databaseName,\n };\n }\n};\n\nconst getDB = async (options: {\n databaseName: string | undefined;\n dbsCache: Map<string, Db>;\n getConnectedClient: () => Promise<MongoClient>;\n}): Promise<Db> => {\n const { dbsCache, databaseName, getConnectedClient } = options;\n const safeDbName = databaseName ?? '___default';\n\n let db = dbsCache.get(safeDbName);\n\n if (!db) {\n const connectedClient = await getConnectedClient();\n\n db = connectedClient.db(databaseName);\n\n dbsCache.set(safeDbName, db);\n }\n\n return db;\n};\n\nconst collectionFor = async <EventType extends Event = Event>(options: {\n collectionName: string;\n streamCollections: Map<string, Collection<EventStream>>;\n db: Db;\n}): Promise<Collection<EventStream<EventType>>> => {\n const { collectionName, db, streamCollections } = options;\n\n let collection = streamCollections.get(collectionName) as\n | Collection<EventStream<EventType>>\n | undefined;\n\n if (!collection) {\n collection = db.collection<EventStream<EventType>>(collectionName);\n await collection.createIndex({ streamName: 1 }, { unique: true });\n\n streamCollections.set(\n collectionName,\n collection as Collection<EventStream>,\n );\n }\n\n return collection;\n};\n\nexport const mongoDBEventStoreStorage = (options: {\n storage?: MongoDBEventStoreStorageOptions | undefined;\n getConnectedClient: () => Promise<MongoClient>;\n}): MongoDBEventStoreStorage => {\n const dbsCache: Map<string, Db> = new Map();\n const streamCollections: Map<string, Collection<EventStream>> = new Map();\n const storageOptions =\n options.storage ?? DefaultMongoDBEventStoreStorageOptions;\n\n const { getConnectedClient } = options;\n\n return {\n collectionFor: async <\n T extends StreamType,\n EventType extends Event = Event,\n >(\n streamType: T,\n ): Promise<Collection<EventStream<EventType>>> => {\n const { collectionName, databaseName } = resolveCollectionAndDatabase(\n streamType,\n storageOptions,\n );\n\n let collection = streamCollections.get(collectionName) as\n | Collection<EventStream<EventType>>\n | undefined;\n\n if (!collection) {\n const db = await getDB({ databaseName, dbsCache, getConnectedClient });\n collection = await collectionFor<EventType>({\n collectionName,\n streamCollections,\n db,\n });\n }\n\n return collection;\n },\n };\n};\n"]}
|