@event-driven-io/emmett-postgresql 0.16.0 → 0.18.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +308 -10
- package/dist/index.d.ts +308 -10
- package/dist/index.js +362 -1
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +362 -1
- package/dist/index.mjs.map +1 -1
- package/package.json +8 -8
- package/dist/chunk-2452BJTJ.js +0 -1
- package/dist/chunk-2452BJTJ.js.map +0 -1
- package/dist/chunk-45ODDXOQ.mjs +0 -1
- package/dist/chunk-45ODDXOQ.mjs.map +0 -1
- package/dist/chunk-6FBMFEVK.js +0 -7
- package/dist/chunk-6FBMFEVK.js.map +0 -1
- package/dist/chunk-ABNBEUC6.mjs +0 -2
- package/dist/chunk-ABNBEUC6.mjs.map +0 -1
- package/dist/chunk-DUXB3PBP.js +0 -94
- package/dist/chunk-DUXB3PBP.js.map +0 -1
- package/dist/chunk-EEYVFO3G.js +0 -264
- package/dist/chunk-EEYVFO3G.js.map +0 -1
- package/dist/chunk-HIX4PKIP.mjs +0 -7
- package/dist/chunk-HIX4PKIP.mjs.map +0 -1
- package/dist/chunk-HS4KMVAP.mjs +0 -2
- package/dist/chunk-HS4KMVAP.mjs.map +0 -1
- package/dist/chunk-J3JL23C4.mjs +0 -2
- package/dist/chunk-J3JL23C4.mjs.map +0 -1
- package/dist/chunk-JQ2VF3NG.js +0 -2
- package/dist/chunk-JQ2VF3NG.js.map +0 -1
- package/dist/chunk-KZTZL7GV.js +0 -2
- package/dist/chunk-KZTZL7GV.js.map +0 -1
- package/dist/chunk-NBTN5MZ6.js +0 -4
- package/dist/chunk-NBTN5MZ6.js.map +0 -1
- package/dist/chunk-PCRD6RK2.mjs +0 -4
- package/dist/chunk-PCRD6RK2.mjs.map +0 -1
- package/dist/chunk-QNOJWKPZ.mjs +0 -2
- package/dist/chunk-QNOJWKPZ.mjs.map +0 -1
- package/dist/chunk-SS2LQM3B.js +0 -2
- package/dist/chunk-SS2LQM3B.js.map +0 -1
- package/dist/chunk-ST3FNDJ5.mjs +0 -94
- package/dist/chunk-ST3FNDJ5.mjs.map +0 -1
- package/dist/chunk-UWD6GOZC.mjs +0 -264
- package/dist/chunk-UWD6GOZC.mjs.map +0 -1
- package/dist/chunk-WQQC2IS2.js +0 -2
- package/dist/chunk-WQQC2IS2.js.map +0 -1
- package/dist/eventStore/index.d.mts +0 -10
- package/dist/eventStore/index.d.ts +0 -10
- package/dist/eventStore/index.js +0 -2
- package/dist/eventStore/index.js.map +0 -1
- package/dist/eventStore/index.mjs +0 -2
- package/dist/eventStore/index.mjs.map +0 -1
- package/dist/eventStore/postgreSQLEventStore.d.mts +0 -5
- package/dist/eventStore/postgreSQLEventStore.d.ts +0 -5
- package/dist/eventStore/postgreSQLEventStore.js +0 -2
- package/dist/eventStore/postgreSQLEventStore.js.map +0 -1
- package/dist/eventStore/postgreSQLEventStore.mjs +0 -2
- package/dist/eventStore/postgreSQLEventStore.mjs.map +0 -1
- package/dist/eventStore/projections/index.d.mts +0 -5
- package/dist/eventStore/projections/index.d.ts +0 -5
- package/dist/eventStore/projections/index.js +0 -2
- package/dist/eventStore/projections/index.js.map +0 -1
- package/dist/eventStore/projections/index.mjs +0 -2
- package/dist/eventStore/projections/index.mjs.map +0 -1
- package/dist/eventStore/projections/pongo/index.d.mts +0 -5
- package/dist/eventStore/projections/pongo/index.d.ts +0 -5
- package/dist/eventStore/projections/pongo/index.js +0 -2
- package/dist/eventStore/projections/pongo/index.js.map +0 -1
- package/dist/eventStore/projections/pongo/index.mjs +0 -2
- package/dist/eventStore/projections/pongo/index.mjs.map +0 -1
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.d.mts +0 -5
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.d.ts +0 -5
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.js +0 -2
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.js.map +0 -1
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.mjs +0 -2
- package/dist/eventStore/projections/pongo/pongoProjectionSpec.mjs.map +0 -1
- package/dist/eventStore/projections/pongo/projections.d.mts +0 -5
- package/dist/eventStore/projections/pongo/projections.d.ts +0 -5
- package/dist/eventStore/projections/pongo/projections.js +0 -2
- package/dist/eventStore/projections/pongo/projections.js.map +0 -1
- package/dist/eventStore/projections/pongo/projections.mjs +0 -2
- package/dist/eventStore/projections/pongo/projections.mjs.map +0 -1
- package/dist/eventStore/projections/postgresProjectionSpec.d.mts +0 -5
- package/dist/eventStore/projections/postgresProjectionSpec.d.ts +0 -5
- package/dist/eventStore/projections/postgresProjectionSpec.js +0 -2
- package/dist/eventStore/projections/postgresProjectionSpec.js.map +0 -1
- package/dist/eventStore/projections/postgresProjectionSpec.mjs +0 -2
- package/dist/eventStore/projections/postgresProjectionSpec.mjs.map +0 -1
- package/dist/eventStore/schema/appendToStream.d.mts +0 -19
- package/dist/eventStore/schema/appendToStream.d.ts +0 -19
- package/dist/eventStore/schema/appendToStream.js +0 -2
- package/dist/eventStore/schema/appendToStream.js.map +0 -1
- package/dist/eventStore/schema/appendToStream.mjs +0 -2
- package/dist/eventStore/schema/appendToStream.mjs.map +0 -1
- package/dist/eventStore/schema/index.d.mts +0 -12
- package/dist/eventStore/schema/index.d.ts +0 -12
- package/dist/eventStore/schema/index.js +0 -2
- package/dist/eventStore/schema/index.js.map +0 -1
- package/dist/eventStore/schema/index.mjs +0 -2
- package/dist/eventStore/schema/index.mjs.map +0 -1
- package/dist/eventStore/schema/readStream.d.mts +0 -8
- package/dist/eventStore/schema/readStream.d.ts +0 -8
- package/dist/eventStore/schema/readStream.js +0 -2
- package/dist/eventStore/schema/readStream.js.map +0 -1
- package/dist/eventStore/schema/readStream.mjs +0 -2
- package/dist/eventStore/schema/readStream.mjs.map +0 -1
- package/dist/eventStore/schema/tables.d.mts +0 -15
- package/dist/eventStore/schema/tables.d.ts +0 -15
- package/dist/eventStore/schema/tables.js +0 -2
- package/dist/eventStore/schema/tables.js.map +0 -1
- package/dist/eventStore/schema/tables.mjs +0 -2
- package/dist/eventStore/schema/tables.mjs.map +0 -1
- package/dist/eventStore/schema/typing.d.mts +0 -31
- package/dist/eventStore/schema/typing.d.ts +0 -31
- package/dist/eventStore/schema/typing.js +0 -2
- package/dist/eventStore/schema/typing.js.map +0 -1
- package/dist/eventStore/schema/typing.mjs +0 -2
- package/dist/eventStore/schema/typing.mjs.map +0 -1
- package/dist/postgreSQLEventStore-loGROgA0.d.mts +0 -159
- package/dist/postgreSQLEventStore-loGROgA0.d.ts +0 -159
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/projections/pongo/pongoProjectionSpec.ts","../src/eventStore/projections/index.ts","../src/eventStore/projections/pongo/projections.ts","../src/eventStore/projections/postgresProjectionSpec.ts"],"names":["pongoClient","pongoProjection","handle","canHandle","postgreSQLProjection","events","context","connectionString","client","pongo","pongoMultiStreamProjection","collectionName","getDocumentId","evolve","collection","event","document","pongoSingleStreamProjection","dumbo","uuid","eventInStream","streamName","eventsInStream","e","newEventsInStream","PostgreSQLProjectionSpec","options","projection","dumoOptions","givenEvents","allEvents","run","pool","globalPosition","eventId","transaction","handleProjections","assert","message","succeeded","p","args","l","error","w","s","assertSQLQueryResultMatches","sql","rows","execute","result","q","expectSQL","defaultPostgreSQLProjectionOptions","allProjections","eventTypes","projections","type","definition","i","postgreSQLRawBatchSQLProjection","sqls","postgreSQLRawSQLProjection","withCollection","inDatabase","inCollection","connection","withoutId","doc","_id","without","assertDocumentsEqual","actual","expected","c","x","documentExists","assertOptions","y","documentsAreTheSame","documents","documentsMatchingHaveCount","expectedCount","documentMatchingExists","documentDoesNotExist","expectPongoDocuments","id","filter"],"mappings":"gGAAA,MAA2B,yBAO3B,OACE,eAAAA,MAKK,yBCbP,MAKO,yBCAP,OACE,eAAAA,MAGK,yBAiCA,IAAMC,EAAkB,CAA0B,CACvD,OAAAC,EACA,UAAAC,CACF,IACEC,EAAgC,CAC9B,UAAAD,EACA,OAAQ,MAAOE,EAAQC,IAAY,CACjC,GAAM,CAAE,iBAAAC,EAAkB,OAAAC,CAAO,EAAIF,EAC/BG,EAAQT,EAAYO,EAAkB,CAAE,OAAAC,CAAO,CAAC,EACtD,MAAMN,EAAOG,EAAQ,CAAE,GAAGC,EAAS,MAAAG,CAAM,CAAC,CAC5C,CACF,CAAC,EAYUC,EAA6B,CAGxC,CACA,eAAAC,EACA,cAAAC,EACA,OAAAC,EACA,UAAAV,CACF,IAIEF,EAAgB,CACd,OAAQ,MAAOI,EAAQ,CAAE,MAAAI,CAAM,IAAM,CACnC,IAAMK,EAAaL,EAAM,GAAG,EAAE,WAAqBE,CAAc,EAEjE,QAAWI,KAASV,EAClB,MAAMS,EAAW,OAAOF,EAAcG,CAAK,EAAG,MAAOC,GAC5C,MAAMH,EAAOG,EAAUD,CAAK,CACpC,CAEL,EACA,UAAAZ,CACF,CAAC,EAWUc,EAA8B,CAGzC,CACA,eAAAN,EACA,OAAAE,EACA,UAAAV,CACF,IAIEO,EAA2B,CACzB,eAAAC,EACA,cAAgBI,GAAUA,EAAM,SAAS,WACzC,OAAAF,EACA,UAAAV,CACF,CAAC,ECnHH,OACE,SAAAe,MAKK,yBAYP,OAAS,MAAMC,MAAY,OAMpB,IAAMC,EAAgB,CAC3BC,EACAN,KAEO,CACL,GAAGA,EACH,SAAU,CACR,GAAIA,EAAM,UAAY,CAAC,EACvB,WAAYA,EAAM,UAAU,YAAcM,CAC5C,CACF,GAGWC,EAAiB,CAC5BD,EACAhB,IAEOA,EAAO,IAAKkB,GAAMH,EAAcC,EAAYE,CAAC,CAAC,EAG1CC,GAAoBF,EAyBpBG,GAA2B,CACtC,IACEC,GACwC,CACxC,CACE,GAAM,CAAE,WAAAC,EAAY,GAAGC,CAAY,EAAIF,EACjC,CAAE,iBAAAnB,CAAiB,EAAIqB,EAE7B,OAAQC,IACC,CACL,KAAOxB,GAAuD,CAC5D,IAAMyB,EAGA,CAAC,EAEDC,EAAM,MAAOC,GAAgB,CACjC,IAAIC,EAAiB,GACfZ,EAAa,oBAEnB,QAAWN,IAAS,CAAC,GAAGc,EAAa,GAAGxB,CAAM,EAAG,CAC/C,IAAM6B,EAAUf,EAAK,EACrBW,EAAU,KAAK,CACb,GAAGf,EACH,SAAU,CAEN,eAAAkB,EACA,eAAgBA,EAChB,WAAAZ,EACA,QAAAa,EAEF,GAAInB,EAAM,UAAY,CAAC,CACzB,CACF,CAAC,EAEDkB,GACF,CAEA,MAAMD,EAAK,gBAAiBG,GAC1BC,EACE,CAACT,CAAU,EACXpB,EACA4B,EACAL,CACF,CACF,CACF,EAEA,MAAO,CACL,KAAM,MACJO,EACAC,IACkB,CAClB,IAAMN,EAAOd,EAAMU,CAAW,EAC9B,GAAI,CACF,MAAMG,EAAIC,CAAI,EAEd,IAAMO,EAAY,MAAMF,EAAO,CAAE,KAAAL,EAAM,iBAAAzB,CAAiB,CAAC,EAErDgC,IAAc,QAAaA,IAAc,IAC3CC,EACEF,GACE,oDACJ,CACJ,QAAE,CACA,MAAMN,EAAK,MAAM,CACnB,CACF,EACA,WAAY,SACPS,IACe,CAClB,IAAMT,EAAOd,EAAMU,CAAW,EAC9B,GAAI,CACF,YAAMG,EAAIC,CAAI,EACR,IAAIU,EAAe,kCAAkC,CAC7D,OAASC,EAAO,CACd,GAAIA,aAAiBD,EAAgB,MAAMC,EAE3C,GAAIF,EAAK,SAAW,EAAG,OAEvB,GAAI,CAACG,EAAmBH,EAAK,CAAC,CAAC,EAAG,CAChCI,EACEJ,EAAK,CAAC,EAAEE,CAAkB,EAC1B,2CAA2CA,GAAO,SAAS,CAAC,EAC9D,EACA,MACF,CAEAE,EACEF,aAAiBF,EAAK,CAAC,EACvB,yDAAyDE,GAAO,SAAS,CAAC,EAC5E,EAEIF,EAAK,CAAC,GACRI,EACEJ,EAAK,CAAC,EAAEE,CAAkB,EAC1B,2CAA2CA,GAAO,SAAS,CAAC,EAC9D,CAEJ,QAAE,CACA,MAAMX,EAAK,MAAM,CACnB,CACF,CACF,CACF,CACF,EAEJ,CACF,CACF,EAEac,EACX,CAA2BC,EAAUC,IACrC,MAAO,CAAE,KAAM,CAAE,QAAAC,CAAQ,CAAE,IAAM,CAC/B,IAAMC,EAAS,MAAMD,EAAQ,MAASF,CAAG,EAEzCI,EAAgBH,CAAI,EAAE,0BAA0B,GAAGE,EAAO,IAAI,CAChE,EAEWE,GAAY,CACvB,MAAQL,IAAc,CACpB,WAAY,CACV,YAAwCC,GACtCF,EAA4BC,EAAKC,CAAI,CACzC,CACF,EACF,EFnKO,IAAMK,GAAgE,CAC3E,YAAa,CAAC,CAChB,EAEajB,EAAoB,MAC/BkB,EACA/C,EACA4B,EACA9B,IACkB,CAClB,IAAMkD,EAAalD,EAAO,IAAKkB,GAAMA,EAAE,IAAI,EAErCiC,EAAcF,EAAe,OAAQd,GACzCA,EAAE,UAAU,KAAMiB,GAASF,EAAW,SAASE,CAAI,CAAC,CACtD,EAEMjD,EAAU,MAAM2B,EAAY,WAAW,KAAK,EAElD,QAAWR,KAAc6B,EACvB,MAAM7B,EAAW,OAAOtB,EAAQ,CAC9B,iBAAAE,EACA,OAAAC,EACA,YAAA2B,EACA,QAASA,EAAY,OACvB,CAAC,CAEL,EAEa/B,EACXsD,GAEAC,EAIED,CAAU,EAEDE,EAAkC,CAC7C1D,KAIGC,IAEHC,EAAgC,CAC9B,UAAAD,EACA,OAAQ,MAAOE,EAAQC,IAAY,CACjC,IAAMuD,EAAc,MAAM3D,EAAOG,EAAQC,CAAO,EAEhD,MAAMA,EAAQ,QAAQ,aAAauD,CAAI,CACzC,CACF,CAAC,EAEUC,GAA6B,CACxC5D,KAIGC,IAEHyD,EACE,MAAOvD,EAAQC,IAAY,CACzB,IAAMuD,EAAc,CAAC,EAErB,QAAW9C,KAASV,EAClBwD,EAAK,KAAK,MAAM3D,EAAOa,EAAOT,CAAO,CAAC,EAExC,OAAOuD,CACT,EACA,GAAG1D,CACL,EDjFF,IAAM4D,EAAiB,CACrB7D,EACAwB,IAIG,CACH,GAAM,CAAE,KAAAM,EAAM,iBAAAzB,EAAkB,WAAAyD,EAAY,aAAAC,CAAa,EAAIvC,EAE7D,OAAOM,EAAK,eAAe,MAAOkC,GAAe,CAC/C,IAAMzD,EAAQT,EAAYO,EAAkB,CAC1C,WAAA2D,CACF,CAAC,EACD,GAAI,CACF,IAAMpD,EAAaL,EAAM,GAAGuD,CAAU,EAAE,WAAWC,CAAY,EAE/D,OAAO/D,EAAOY,CAAU,CAC1B,QAAE,CACA,MAAML,EAAM,MAAM,CACpB,CACF,CAAC,CACH,EAEM0D,EACJC,GACG,CACH,GAAM,CAAE,IAAAC,EAAK,GAAGC,CAAQ,EAAIF,EAE5B,OAAOE,CACT,EAEMC,EAAuB,CAG3BC,EACAC,KAEI,QAASA,GACXC,EACED,EAAS,IACTD,EAAO,IAEP,4CAA4CC,EAAS,GAAG,aAAaD,EAAO,GAAG,EACjF,EAEKG,EAAgBR,EAAUK,CAAM,EAAGL,EAAUM,CAAQ,CAAC,GASlDG,EACX,CACE5D,EACAU,IAEDmD,GACCd,EACE,MAAOjD,GAAe,CACpB,IAAMoC,EAAS,MAAMpC,EAAW,QAC9B,WAAYY,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAoD,EAAgB5B,CAAM,EAEtBqB,EAAqBrB,EAAQlC,CAAQ,CACvC,EACA,CAAE,GAAGU,EAAS,GAAGmD,CAAc,CACjC,EAESE,EACX,CACEC,EACAtD,IAEDmD,GACCd,EACE,MAAOjD,GAAe,CACpB,IAAMoC,EAAS,MAAMpC,EAAW,KAC9B,WAAYY,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAgD,EACEM,EAAU,OACV9B,EAAO,OACP,yCACF,EAEA,QAASS,EAAI,EAAGA,EAAIqB,EAAU,OAAQrB,IACpCR,EAAgBD,CAAM,EAAE,SAAS8B,EAAUrB,CAAC,CAAE,CAElD,EACA,CAAE,GAAGjC,EAAS,GAAGmD,CAAc,CACjC,EAESI,EACX,CACEC,EACAxD,IAEDmD,GACCd,EACE,MAAOjD,GAAe,CACpB,IAAMoC,EAAS,MAAMpC,EAAW,KAC9B,WAAYY,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAgD,EACEQ,EACAhC,EAAO,OACP,yCACF,CACF,EACA,CAAE,GAAGxB,EAAS,GAAGmD,CAAc,CACjC,EAESM,EAETzD,GAEDmD,GACCd,EACE,MAAOjD,GAAe,CACpB,IAAMoC,EAAS,MAAMpC,EAAW,KAC9B,WAAYY,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAyB,EAAgBD,CAAM,EAAE,WAAW,CACrC,EACA,CAAE,GAAGxB,EAAS,GAAGmD,CAAc,CACjC,EAESO,EAET1D,GAEDmD,GACCd,EACE,MAAOjD,GAAe,CACpB,IAAMoC,EAAS,MAAMpC,EAAW,QAC9B,WAAYY,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAoD,EAAgB5B,CAAM,CACxB,EACA,CAAE,GAAGxB,EAAS,GAAGmD,CAAc,CACjC,EAESQ,GAAuB,CAClC,eAAiB1E,IACR,CACL,OAAS2E,IACA,CACL,UACEtE,GAEA4D,EAAe5D,EAAU,CACvB,OAAQsE,EACR,aAAc3E,CAChB,CAAC,EACH,QAAS,IACPwE,EAAuB,CACrB,OAAQG,EACR,aAAc3E,CAChB,CAAC,EACH,WAAY,IACVyE,EAAqB,CACnB,OAAQE,EACR,aAAc3E,CAChB,CAAC,CACL,GAEF,SACE4E,IAEO,CACL,YAAcP,GACZD,EAAyBC,EAAW,CAClC,eAAgBO,EAChB,aAAc5E,CAChB,CAAC,EACH,YAAcuE,GACZD,EAA2BC,EAAe,CACxC,eAAgBK,EAChB,aAAc5E,CAChB,CAAC,EACH,QAAS,IACPwE,EAAuB,CACrB,eAAgBI,EAChB,aAAc5E,CAChB,CAAC,EACH,WAAY,IACVyE,EAAqB,CACnB,eAAgBG,EAChB,aAAc5E,CAChB,CAAC,CACL,EAEJ,EAEJ","sourcesContent":["import { type Dumbo } from '@event-driven-io/dumbo';\nimport {\n assertDeepEqual,\n assertEqual,\n assertIsNotNull,\n assertThatArray,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoCollection,\n type PongoDocument,\n type PongoFilter,\n type WithId,\n} from '@event-driven-io/pongo';\nimport { type PostgreSQLProjectionAssert } from '..';\n\nexport type PongoAssertOptions = {\n inCollection: string;\n inDatabase?: string;\n};\n\nconst withCollection = (\n handle: (collection: PongoCollection<PongoDocument>) => Promise<void>,\n options: {\n pool: Dumbo;\n connectionString: string;\n } & PongoAssertOptions,\n) => {\n const { pool, connectionString, inDatabase, inCollection } = options;\n\n return pool.withConnection(async (connection) => {\n const pongo = pongoClient(connectionString, {\n connection,\n });\n try {\n const collection = pongo.db(inDatabase).collection(inCollection);\n\n return handle(collection);\n } finally {\n await pongo.close();\n }\n });\n};\n\nconst withoutId = <Doc extends PongoDocument | WithId<PongoDocument>>(\n doc: Doc,\n) => {\n const { _id, ...without } = doc;\n\n return without;\n};\n\nconst assertDocumentsEqual = <\n Doc extends PongoDocument | WithId<PongoDocument>,\n>(\n actual: PongoDocument,\n expected: Doc,\n) => {\n if ('_id' in expected)\n assertEqual(\n expected._id,\n actual._id,\n // eslint-disable-next-line @typescript-eslint/restrict-template-expressions\n `Document ids are not matching! Expected: ${expected._id}, actual: ${actual._id}`,\n );\n\n return assertDeepEqual(withoutId(actual), withoutId(expected));\n};\n\ntype FilterOrId<Doc extends PongoDocument | WithId<PongoDocument>> =\n | { withId: string }\n | {\n matchingFilter: PongoFilter<Doc>;\n };\n\nexport const documentExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n\n assertDocumentsEqual(result, document);\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsAreTheSame =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n documents: Doc[],\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n documents.length,\n result.length,\n 'Different Documents Count than expected',\n );\n\n for (let i = 0; i < documents.length; i++) {\n assertThatArray(result).contains(documents[i]!);\n }\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsMatchingHaveCount =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n expectedCount: number,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n expectedCount,\n result.length,\n 'Different Documents Count than expected',\n );\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentMatchingExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertThatArray(result).isNotEmpty();\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentDoesNotExist =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n },\n { ...options, ...assertOptions },\n );\n\nexport const expectPongoDocuments = {\n fromCollection: (collectionName: string) => {\n return {\n withId: (id: string) => {\n return {\n toBeEqual: <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n ) =>\n documentExists(document, {\n withId: id,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n withId: id,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n withId: id,\n inCollection: collectionName,\n }),\n };\n },\n matching: <Doc extends PongoDocument | WithId<PongoDocument>>(\n filter: PongoFilter<Doc>,\n ) => {\n return {\n toBeTheSame: (documents: Doc[]) =>\n documentsAreTheSame<Doc>(documents, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toHaveCount: (expectedCount: number) =>\n documentsMatchingHaveCount(expectedCount, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n };\n },\n };\n },\n};\n","import {\n type NodePostgresClient,\n type NodePostgresTransaction,\n type SQL,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n projection,\n type CanHandle,\n type Event,\n type ProjectionHandler,\n type ReadEvent,\n type TypedProjectionDefinition,\n} from '@event-driven-io/emmett';\nimport type { PostgresEventStoreOptions } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionHandlerContext = {\n connectionString: string;\n client: NodePostgresClient;\n execute: SQLExecutor;\n transaction: NodePostgresTransaction;\n};\n\nexport type PostgreSQLProjectionHandler<EventType extends Event = Event> =\n ProjectionHandler<EventType, PostgreSQLProjectionHandlerContext>;\n\nexport interface PostgreSQLProjectionDefinition<EventType extends Event = Event>\n extends TypedProjectionDefinition<\n EventType,\n PostgreSQLProjectionHandlerContext\n > {}\n\nexport const defaultPostgreSQLProjectionOptions: PostgresEventStoreOptions = {\n projections: [],\n};\n\nexport const handleProjections = async <EventType extends Event = Event>(\n allProjections: PostgreSQLProjectionDefinition<EventType>[],\n connectionString: string,\n transaction: NodePostgresTransaction,\n events: ReadEvent<EventType>[],\n): Promise<void> => {\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n const client = (await transaction.connection.open()) as NodePostgresClient;\n\n for (const projection of projections) {\n await projection.handle(events, {\n connectionString,\n client,\n transaction,\n execute: transaction.execute,\n });\n }\n};\n\nexport const postgreSQLProjection = <EventType extends Event>(\n definition: PostgreSQLProjectionDefinition<EventType>,\n): PostgreSQLProjectionDefinition =>\n projection<\n EventType,\n PostgreSQLProjectionHandlerContext,\n PostgreSQLProjectionDefinition<EventType>\n >(definition) as PostgreSQLProjectionDefinition;\n\nexport const postgreSQLRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL[]> | SQL[],\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: SQL[] = await handle(events, context);\n\n await context.execute.batchCommand(sqls);\n },\n });\n\nexport const postgreSQLRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL> | SQL,\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: SQL[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n\nexport * from './pongo';\nexport * from './postgresProjectionSpec';\n","import {\n type CanHandle,\n type Event,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoClient,\n type PongoDocument,\n} from '@event-driven-io/pongo';\nimport {\n postgreSQLProjection,\n type PostgreSQLProjectionDefinition,\n type PostgreSQLProjectionHandlerContext,\n} from '..';\n\nexport type PongoProjectionHandlerContext =\n PostgreSQLProjectionHandlerContext & {\n pongo: PongoClient;\n };\n\nexport type PongoDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n> =\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Document | null)\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoProjectionOptions<EventType extends Event> = {\n handle: (\n events: ReadEvent<EventType>[],\n context: PongoProjectionHandlerContext,\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoProjection = <EventType extends Event>({\n handle,\n canHandle,\n}: PongoProjectionOptions<EventType>): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const { connectionString, client } = context;\n const pongo = pongoClient(connectionString, { client });\n await handle(events, { ...context, pongo });\n },\n });\n\nexport type PongoMultiStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n> = {\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n evolve: PongoDocumentEvolve<Document, EventType>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoMultiStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n>({\n collectionName,\n getDocumentId,\n evolve,\n canHandle,\n}: PongoMultiStreamProjectionOptions<\n Document,\n EventType\n>): PostgreSQLProjectionDefinition =>\n pongoProjection({\n handle: async (events, { pongo }) => {\n const collection = pongo.db().collection<Document>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), async (document) => {\n return await evolve(document, event);\n });\n }\n },\n canHandle,\n });\n\nexport type PongoSingleStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n> = {\n collectionName: string;\n evolve: PongoDocumentEvolve<Document, EventType>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoSingleStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n>({\n collectionName,\n evolve,\n canHandle,\n}: PongoSingleStreamProjectionOptions<\n Document,\n EventType\n>): PostgreSQLProjectionDefinition =>\n pongoMultiStreamProjection({\n collectionName,\n getDocumentId: (event) => event.metadata.streamName,\n evolve,\n canHandle,\n });\n","import {\n dumbo,\n type Dumbo,\n type DumboOptions,\n type QueryResultRow,\n type SQL,\n} from '@event-driven-io/dumbo';\nimport {\n assertFails,\n AssertionError,\n assertThatArray,\n assertTrue,\n isErrorConstructor,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { handleProjections, type PostgreSQLProjectionDefinition } from '.';\n\nexport type PostgreSQLProjectionSpecEvent<EventType extends Event> =\n EventType & { metadata?: Partial<ReadEventMetadataWithGlobalPosition> };\n\nexport const eventInStream = <EventType extends Event = Event>(\n streamName: string,\n event: PostgreSQLProjectionSpecEvent<EventType>,\n): PostgreSQLProjectionSpecEvent<EventType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n },\n };\n};\n\nexport const eventsInStream = <EventType extends Event = Event>(\n streamName: string,\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n): PostgreSQLProjectionSpecEvent<EventType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\nexport type PostgreSQLProjectionSpec<EventType extends Event> = (\n givenEvents: PostgreSQLProjectionSpecEvent<EventType>[],\n) => {\n when: (events: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n then: (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type PostgreSQLProjectionAssert = (options: {\n pool: Dumbo;\n connectionString: string;\n}) => Promise<void | boolean>;\n\nexport type PostgreSQLProjectionSpecOptions = {\n projection: PostgreSQLProjectionDefinition;\n} & DumboOptions;\n\nexport const PostgreSQLProjectionSpec = {\n for: <EventType extends Event>(\n options: PostgreSQLProjectionSpecOptions,\n ): PostgreSQLProjectionSpec<EventType> => {\n {\n const { projection, ...dumoOptions } = options;\n const { connectionString } = dumoOptions;\n\n return (givenEvents: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n return {\n when: (events: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n const allEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = [];\n\n const run = async (pool: Dumbo) => {\n let globalPosition = 0n;\n const streamName = 'defaultStreamName';\n\n for (const event of [...givenEvents, ...events]) {\n const eventId = uuid();\n allEvents.push({\n ...event,\n metadata: {\n ...{\n globalPosition,\n streamPosition: globalPosition,\n streamName,\n eventId,\n },\n ...(event.metadata ?? {}),\n },\n });\n\n globalPosition++;\n }\n\n await pool.withTransaction((transaction) =>\n handleProjections(\n [projection],\n connectionString,\n transaction,\n allEvents,\n ),\n );\n };\n\n return {\n then: async (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n\n const succeeded = await assert({ pool, connectionString });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await pool.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await pool.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const assertSQLQueryResultMatches =\n <T extends QueryResultRow>(sql: SQL, rows: T[]): PostgreSQLProjectionAssert =>\n async ({ pool: { execute } }) => {\n const result = await execute.query<T>(sql);\n\n assertThatArray(rows).containsExactlyInAnyOrder(...result.rows);\n };\n\nexport const expectSQL = {\n query: (sql: SQL) => ({\n resultRows: {\n toBeTheSame: <T extends QueryResultRow>(rows: T[]) =>\n assertSQLQueryResultMatches(sql, rows),\n },\n }),\n};\n"]}
|
package/dist/chunk-KZTZL7GV.js
DELETED
|
@@ -1,2 +0,0 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }var _chunkJQ2VF3NGjs = require('./chunk-JQ2VF3NG.js');var _chunkWQQC2IS2js = require('./chunk-WQQC2IS2.js');var _chunkDUXB3PBPjs = require('./chunk-DUXB3PBP.js');var _chunkNBTN5MZ6js = require('./chunk-NBTN5MZ6.js');var _chunk6FBMFEVKjs = require('./chunk-6FBMFEVK.js');var _dumbo = require('@event-driven-io/dumbo');require('pg');var _=(t,e=_chunkJQ2VF3NGjs.p)=>{let g=_dumbo.dumbo.call(void 0, {connectionString:t,...e.connectionOptions?e.connectionOptions:{}}),d=_chunkWQQC2IS2js.b.call(void 0, g),O=e.projections.filter(({type:n})=>n==="inline").map(({projection:n})=>n);return{async aggregateStream(n,o){await d;let{evolve:i,initialState:a,read:l}=o,S=_optionalChain([l, 'optionalAccess', _2 => _2.expectedStreamVersion]),r=a(),s=await this.readStream(n,o.read);if(s===null)return null;let c=s.currentStreamVersion;b(c,S);for(let P of s.events)P&&(r=i(r,P));return{currentStreamVersion:c,state:r}},readStream:async(n,o)=>(await d,_chunkNBTN5MZ6js.a.call(void 0, g.execute,n,o)),appendToStream:async(n,o,i)=>{await d;let[a,...l]=n.split("-"),S=a&&l.length>0?a:"emt:unknown",r=await _chunkDUXB3PBPjs.b.call(void 0, g,n,S,o,{...i,preCommitHook:(s,c)=>_chunkJQ2VF3NGjs.q.call(void 0, O,t,s,c)});if(!r.success)throw new (0, _chunk6FBMFEVKjs.o)(-1n,_nullishCoalesce(_optionalChain([i, 'optionalAccess', _3 => _3.expectedStreamVersion]), () => (_chunk6FBMFEVKjs.n)));return{nextExpectedStreamVersion:r.nextStreamPosition}},close:()=>_dumbo.endPool.call(void 0, {connectionString:t})}},N=(t,e)=>e===_chunk6FBMFEVKjs.n?!0:e==_chunk6FBMFEVKjs.m?t===void 0:e==_chunk6FBMFEVKjs.l?t!==void 0:t===e,b=(t,e)=>{if(e??=_chunk6FBMFEVKjs.n,!N(t,e))throw new (0, _chunk6FBMFEVKjs.o)(t,e)};exports.a = _;
|
|
2
|
-
//# sourceMappingURL=chunk-KZTZL7GV.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/postgreSQLEventStore.ts"],"names":["dumbo","endPool","getPostgreSQLEventStore","connectionString","options","defaultPostgreSQLProjectionOptions","pool","ensureSchemaExists","createEventStoreSchema","inlineProjections","type","projection","streamName","evolve","initialState","read","expectedStreamVersion","state","result","currentStreamVersion","assertExpectedVersionMatchesCurrent","event","readStream","events","firstPart","rest","streamType","appendResult","appendToStream","client","handleProjections","t","n","matchesExpectedVersion","current","expected","a","s"],"mappings":"oOAAA,OACE,SAAAA,EACA,WAAAC,MAIK,yBAmBP,MAAe,KAmFR,IAAMC,EAA0B,CACrCC,EACAC,EAAqCC,IACd,CACvB,IAAMC,EAAON,EAAM,CACjB,iBAAAG,EACA,GAAIC,EAAQ,kBAAoBA,EAAQ,kBAAoB,CAAC,CAC/D,CAAC,EACKG,EAAqBC,EAAuBF,CAAI,EAEhDG,EAAoBL,EAAQ,YAC/B,OAAO,CAAC,CAAE,KAAAM,CAAK,IAAMA,IAAS,QAAQ,EACtC,IAAI,CAAC,CAAE,WAAAC,CAAW,IAAMA,CAAU,EAErC,MAAO,CACL,MAAM,gBACJC,EACAR,EAC8C,CAC9C,MAAMG,EACN,GAAM,CAAE,OAAAM,EAAQ,aAAAC,EAAc,KAAAC,CAAK,EAAIX,EAEjCY,EAAwBD,GAAM,sBAEhCE,EAAQH,EAAa,EAEnBI,EAAS,MAAM,KAAK,WAAsBN,EAAYR,EAAQ,IAAI,EAExE,GAAIc,IAAW,KAAM,OAAO,KAE5B,IAAMC,EAAuBD,EAAO,qBAEpCE,EACED,EACAH,CACF,EAEA,QAAWK,KAASH,EAAO,OACpBG,IAELJ,EAAQJ,EAAOI,EAAOI,CAAK,GAG7B,MAAO,CACL,qBAAsBF,EACtB,MAAAF,CACF,CACF,EAEA,WAAY,MACVL,EACAR,KAQA,MAAMG,EACCe,EAAsBhB,EAAK,QAASM,EAAYR,CAAO,GAGhE,eAAgB,MACdQ,EACAW,EACAnB,IACkC,CAClC,MAAMG,EAEN,GAAM,CAACiB,EAAW,GAAGC,CAAI,EAAIb,EAAW,MAAM,GAAG,EAE3Cc,EACJF,GAAaC,EAAK,OAAS,EAAID,EAAY,cAEvCG,EAAe,MAAMC,EACzBtB,EACAM,EACAc,EACAH,EACA,CACE,GAAGnB,EACH,cAAe,CAACyB,EAAQN,IACtBO,EACErB,EACAN,EACA0B,EACAN,CACF,CACJ,CACF,EAEA,GAAI,CAACI,EAAa,QAChB,MAAM,IAAII,EACR,CAAC,GACD3B,GAAS,uBAAyB4B,CACpC,EAEF,MAAO,CAAE,0BAA2BL,EAAa,kBAAmB,CACtE,EACA,MAAO,IAAM1B,EAAQ,CAAE,iBAAAE,CAAiB,CAAC,CAC3C,CACF,EAEM8B,EAAyB,CAC7BC,EACAC,IAEIA,IAAaH,EAA6B,GAE1CG,GAAYC,EAA8BF,IAAY,OAEtDC,GAAYE,EAAsBH,IAAY,OAE3CA,IAAYC,EAGff,EAAsC,CAC1Cc,EACAC,IACS,CAGT,GAFAA,IAAaH,EAET,CAACC,EAAuBC,EAASC,CAAQ,EAC3C,MAAM,IAAIJ,EAA6BG,EAASC,CAAQ,CAC5D","sourcesContent":["import {\n dumbo,\n endPool,\n type NodePostgresClientConnection,\n type NodePostgresConnector,\n type NodePostgresPoolClientConnection,\n} from '@event-driven-io/dumbo';\nimport {\n ExpectedVersionConflictError,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AggregateStreamOptions,\n type AggregateStreamResult,\n type AppendToStreamOptions,\n type AppendToStreamResult,\n type DefaultStreamVersionType,\n type Event,\n type EventStore,\n type ExpectedStreamVersion,\n type ProjectionRegistration,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport pg from 'pg';\nimport {\n defaultPostgreSQLProjectionOptions,\n handleProjections,\n type PostgreSQLProjectionHandlerContext,\n} from './projections';\nimport { appendToStream, createEventStoreSchema, readStream } from './schema';\n\nexport interface PostgresEventStore\n extends EventStore<\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n > {\n close(): Promise<void>;\n}\n\ntype PostgresEventStorePooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pool: pg.Pool;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: true;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n };\n\ntype PostgresEventStoreNotPooledOptions =\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n client: pg.Client;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n pooled: false;\n }\n | {\n connector?: NodePostgresConnector;\n connectionString?: string;\n database?: string;\n connection:\n | NodePostgresPoolClientConnection\n | NodePostgresClientConnection;\n pooled?: false;\n };\n\nexport type PostgresEventStoreConnectionOptions =\n | PostgresEventStorePooledOptions\n | PostgresEventStoreNotPooledOptions;\n\nexport type PostgresEventStoreOptions = {\n projections: ProjectionRegistration<\n 'inline',\n PostgreSQLProjectionHandlerContext\n >[];\n connectionOptions?: PostgresEventStoreConnectionOptions;\n};\nexport const getPostgreSQLEventStore = (\n connectionString: string,\n options: PostgresEventStoreOptions = defaultPostgreSQLProjectionOptions,\n): PostgresEventStore => {\n const pool = dumbo({\n connectionString,\n ...(options.connectionOptions ? options.connectionOptions : {}),\n });\n const ensureSchemaExists = createEventStoreSchema(pool);\n\n const inlineProjections = options.projections\n .filter(({ type }) => type === 'inline')\n .map(({ projection }) => projection);\n\n return {\n async aggregateStream<State, EventType extends Event>(\n streamName: string,\n options: AggregateStreamOptions<State, EventType>,\n ): Promise<AggregateStreamResult<State> | null> {\n await ensureSchemaExists;\n const { evolve, initialState, read } = options;\n\n const expectedStreamVersion = read?.expectedStreamVersion;\n\n let state = initialState();\n\n const result = await this.readStream<EventType>(streamName, options.read);\n\n if (result === null) return null;\n\n const currentStreamVersion = result.currentStreamVersion;\n\n assertExpectedVersionMatchesCurrent(\n currentStreamVersion,\n expectedStreamVersion,\n );\n\n for (const event of result.events) {\n if (!event) continue;\n\n state = evolve(state, event);\n }\n\n return {\n currentStreamVersion: currentStreamVersion,\n state,\n };\n },\n\n readStream: async <EventType extends Event>(\n streamName: string,\n options?: ReadStreamOptions,\n ): Promise<\n ReadStreamResult<\n EventType,\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n >\n > => {\n await ensureSchemaExists;\n return readStream<EventType>(pool.execute, streamName, options);\n },\n\n appendToStream: async <EventType extends Event>(\n streamName: string,\n events: EventType[],\n options?: AppendToStreamOptions,\n ): Promise<AppendToStreamResult> => {\n await ensureSchemaExists;\n // TODO: This has to be smarter when we introduce urn-based resolution\n const [firstPart, ...rest] = streamName.split('-');\n\n const streamType =\n firstPart && rest.length > 0 ? firstPart : 'emt:unknown';\n\n const appendResult = await appendToStream(\n pool,\n streamName,\n streamType,\n events,\n {\n ...options,\n preCommitHook: (client, events) =>\n handleProjections(\n inlineProjections,\n connectionString,\n client,\n events,\n ),\n },\n );\n\n if (!appendResult.success)\n throw new ExpectedVersionConflictError<bigint>(\n -1n, //TODO: Return actual version in case of error\n options?.expectedStreamVersion ?? NO_CONCURRENCY_CHECK,\n );\n\n return { nextExpectedStreamVersion: appendResult.nextStreamPosition };\n },\n close: () => endPool({ connectionString }),\n };\n};\n\nconst matchesExpectedVersion = (\n current: bigint | undefined,\n expected: ExpectedStreamVersion,\n): boolean => {\n if (expected === NO_CONCURRENCY_CHECK) return true;\n\n if (expected == STREAM_DOES_NOT_EXIST) return current === undefined;\n\n if (expected == STREAM_EXISTS) return current !== undefined;\n\n return current === expected;\n};\n\nconst assertExpectedVersionMatchesCurrent = (\n current: bigint | undefined,\n expected: ExpectedStreamVersion | undefined,\n): void => {\n expected ??= NO_CONCURRENCY_CHECK;\n\n if (!matchesExpectedVersion(current, expected))\n throw new ExpectedVersionConflictError(current, expected);\n};\n"]}
|
package/dist/chunk-NBTN5MZ6.js
DELETED
|
@@ -1,4 +0,0 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }var _chunkSS2LQM3Bjs = require('./chunk-SS2LQM3B.js');var _chunk6FBMFEVKjs = require('./chunk-6FBMFEVK.js');var _dumbo = require('@event-driven-io/dumbo');var u=async(m,n,t)=>{let d=t&&"from"in t?`AND stream_position >= ${t.from}`:"",i=Number(t&&"to"in t?t.to:t&&"maxCount"in t&&t.maxCount?t.from+t.maxCount:NaN),p=isNaN(i)?"":`AND stream_position <= ${i}`,a=await _dumbo.mapRows.call(void 0, m.query(_dumbo.sql.call(void 0, `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id
|
|
2
|
-
FROM ${_chunkSS2LQM3Bjs.f.name}
|
|
3
|
-
WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${d} ${p}`,n,_nullishCoalesce(_optionalChain([t, 'optionalAccess', _ => _.partition]), () => (_chunkSS2LQM3Bjs.c)))),e=>{let r=_chunk6FBMFEVKjs.j.call(void 0, e.event_type,e.event_data,e.event_metadata);return{...r,metadata:{...r.metadata,eventId:e.event_id,streamName:n,streamPosition:BigInt(e.stream_position),globalPosition:BigInt(e.global_position)}}});return a.length>0?{currentStreamVersion:a[a.length-1].metadata.streamPosition,events:a}:null};exports.a = u;
|
|
4
|
-
//# sourceMappingURL=chunk-NBTN5MZ6.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/schema/readStream.ts"],"names":["mapRows","sql","readStream","execute","streamId","options","fromCondition","to","toCondition","events","eventsTable","defaultTag","row","rawEvent","n"],"mappings":"uFAAA,OAAS,WAAAA,EAAS,OAAAC,MAA6B,yBA2BxC,IAAMC,EAAa,MACxBC,EACAC,EACAC,IAOG,CACH,IAAMC,EACJD,GAAW,SAAUA,EACjB,0BAA0BA,EAAQ,IAAI,GACtC,GAEAE,EAAK,OACTF,GAAW,OAAQA,EACfA,EAAQ,GACRA,GAAW,aAAcA,GAAWA,EAAQ,SAC1CA,EAAQ,KAAOA,EAAQ,SACvB,GACR,EAEMG,EAAe,MAAMD,CAAE,EAAqC,GAAjC,0BAA0BA,CAAE,GAEvDE,EACJ,MAAMT,EACJG,EAAQ,MACNF,EACE;AAAA,kBACQS,EAAY,IAAI;AAAA,6EAC2CJ,CAAa,IAAIE,CAAW,GAC/FJ,EACAC,GAAS,WAAaM,CACxB,CACF,EACCC,GAAQ,CACP,IAAMC,EAAWC,EACfF,EAAI,WACJA,EAAI,WACJA,EAAI,cACN,EAEA,MAAO,CACL,GAAGC,EACH,SAAU,CACR,GAAGA,EAAS,SACZ,QAASD,EAAI,SACb,WAAYR,EACZ,eAAgB,OAAOQ,EAAI,eAAe,EAC1C,eAAgB,OAAOA,EAAI,eAAe,CAC5C,CACF,CACF,CACF,EAEF,OAAOH,EAAO,OAAS,EACnB,CACE,qBACEA,EAAOA,EAAO,OAAS,CAAC,EAAG,SAAS,eACtC,OAAAA,CACF,EACA,IACN","sourcesContent":["import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n event,\n type DefaultStreamVersionType,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadStreamSqlResult<EventType extends Event> = {\n stream_position: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n execute: SQLExecutor,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<\n EventType,\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n >\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const events: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n await mapRows(\n execute.query<ReadStreamSqlResult<EventType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${fromCondition} ${toCondition}`,\n streamId,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = event<EventType>(\n row.event_type,\n row.event_data,\n row.event_metadata,\n ) as EventType;\n\n return {\n ...rawEvent,\n metadata: {\n ...rawEvent.metadata,\n eventId: row.event_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n },\n };\n },\n );\n\n return events.length > 0\n ? {\n currentStreamVersion:\n events[events.length - 1]!.metadata.streamPosition,\n events,\n }\n : null;\n};\n"]}
|
package/dist/chunk-PCRD6RK2.mjs
DELETED
|
@@ -1,4 +0,0 @@
|
|
|
1
|
-
import{c as o,f as v}from"./chunk-ABNBEUC6.mjs";import{j as s}from"./chunk-HIX4PKIP.mjs";import{mapRows as E,sql as y}from"@event-driven-io/dumbo";var u=async(m,n,t)=>{let d=t&&"from"in t?`AND stream_position >= ${t.from}`:"",i=Number(t&&"to"in t?t.to:t&&"maxCount"in t&&t.maxCount?t.from+t.maxCount:NaN),p=isNaN(i)?"":`AND stream_position <= ${i}`,a=await E(m.query(y(`SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id
|
|
2
|
-
FROM ${v.name}
|
|
3
|
-
WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${d} ${p}`,n,t?.partition??o)),e=>{let r=s(e.event_type,e.event_data,e.event_metadata);return{...r,metadata:{...r.metadata,eventId:e.event_id,streamName:n,streamPosition:BigInt(e.stream_position),globalPosition:BigInt(e.global_position)}}});return a.length>0?{currentStreamVersion:a[a.length-1].metadata.streamPosition,events:a}:null};export{u as a};
|
|
4
|
-
//# sourceMappingURL=chunk-PCRD6RK2.mjs.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/schema/readStream.ts"],"sourcesContent":["import { mapRows, sql, type SQLExecutor } from '@event-driven-io/dumbo';\nimport {\n event,\n type DefaultStreamVersionType,\n type Event,\n type EventDataOf,\n type EventMetaDataOf,\n type EventTypeOf,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ReadStreamOptions,\n type ReadStreamResult,\n} from '@event-driven-io/emmett';\nimport { defaultTag, eventsTable } from './typing';\n\ntype ReadStreamSqlResult<EventType extends Event> = {\n stream_position: string;\n event_data: EventDataOf<EventType>;\n event_metadata: EventMetaDataOf<EventType>;\n event_schema_version: string;\n event_type: EventTypeOf<EventType>;\n event_id: string;\n global_position: string;\n transaction_id: string;\n created: string;\n};\n\nexport const readStream = async <EventType extends Event>(\n execute: SQLExecutor,\n streamId: string,\n options?: ReadStreamOptions & { partition?: string },\n): Promise<\n ReadStreamResult<\n EventType,\n DefaultStreamVersionType,\n ReadEventMetadataWithGlobalPosition\n >\n> => {\n const fromCondition: string =\n options && 'from' in options\n ? `AND stream_position >= ${options.from}`\n : '';\n\n const to = Number(\n options && 'to' in options\n ? options.to\n : options && 'maxCount' in options && options.maxCount\n ? options.from + options.maxCount\n : NaN,\n );\n\n const toCondition = !isNaN(to) ? `AND stream_position <= ${to}` : '';\n\n const events: ReadEvent<EventType, ReadEventMetadataWithGlobalPosition>[] =\n await mapRows(\n execute.query<ReadStreamSqlResult<EventType>>(\n sql(\n `SELECT stream_id, stream_position, global_position, event_data, event_metadata, event_schema_version, event_type, event_id\n FROM ${eventsTable.name}\n WHERE stream_id = %L AND partition = %L AND is_archived = FALSE ${fromCondition} ${toCondition}`,\n streamId,\n options?.partition ?? defaultTag,\n ),\n ),\n (row) => {\n const rawEvent = event<EventType>(\n row.event_type,\n row.event_data,\n row.event_metadata,\n ) as EventType;\n\n return {\n ...rawEvent,\n metadata: {\n ...rawEvent.metadata,\n eventId: row.event_id,\n streamName: streamId,\n streamPosition: BigInt(row.stream_position),\n globalPosition: BigInt(row.global_position),\n },\n };\n },\n );\n\n return events.length > 0\n ? {\n currentStreamVersion:\n events[events.length - 1]!.metadata.streamPosition,\n events,\n }\n : null;\n};\n"],"mappings":"yFAAA,OAAS,WAAAA,EAAS,OAAAC,MAA6B,yBA2BxC,IAAMC,EAAa,MACxBC,EACAC,EACAC,IAOG,CACH,IAAMC,EACJD,GAAW,SAAUA,EACjB,0BAA0BA,EAAQ,IAAI,GACtC,GAEAE,EAAK,OACTF,GAAW,OAAQA,EACfA,EAAQ,GACRA,GAAW,aAAcA,GAAWA,EAAQ,SAC1CA,EAAQ,KAAOA,EAAQ,SACvB,GACR,EAEMG,EAAe,MAAMD,CAAE,EAAqC,GAAjC,0BAA0BA,CAAE,GAEvDE,EACJ,MAAMC,EACJP,EAAQ,MACNQ,EACE;AAAA,kBACQC,EAAY,IAAI;AAAA,6EAC2CN,CAAa,IAAIE,CAAW,GAC/FJ,EACAC,GAAS,WAAaQ,CACxB,CACF,EACCC,GAAQ,CACP,IAAMC,EAAWC,EACfF,EAAI,WACJA,EAAI,WACJA,EAAI,cACN,EAEA,MAAO,CACL,GAAGC,EACH,SAAU,CACR,GAAGA,EAAS,SACZ,QAASD,EAAI,SACb,WAAYV,EACZ,eAAgB,OAAOU,EAAI,eAAe,EAC1C,eAAgB,OAAOA,EAAI,eAAe,CAC5C,CACF,CACF,CACF,EAEF,OAAOL,EAAO,OAAS,EACnB,CACE,qBACEA,EAAOA,EAAO,OAAS,CAAC,EAAG,SAAS,eACtC,OAAAA,CACF,EACA,IACN","names":["mapRows","sql","readStream","execute","streamId","options","fromCondition","to","toCondition","events","mapRows","sql","eventsTable","defaultTag","row","rawEvent","n"]}
|
package/dist/chunk-QNOJWKPZ.mjs
DELETED
|
@@ -1,2 +0,0 @@
|
|
|
1
|
-
import{b as v,c as S,d as D,e as m,f as y,g as E,h as g,i as x,k as h}from"./chunk-HIX4PKIP.mjs";import"@event-driven-io/dumbo";import{pongoClient as W}from"@event-driven-io/pongo";import"@event-driven-io/dumbo";import{pongoClient as C}from"@event-driven-io/pongo";var I=({handle:t,canHandle:e})=>u({canHandle:e,handle:async(o,n)=>{let{connectionString:r,client:s}=n,p=C(r,{client:s});await t(o,{...n,pongo:p})}}),O=({collectionName:t,getDocumentId:e,evolve:o,canHandle:n})=>I({handle:async(r,{pongo:s})=>{let p=s.db().collection(t);for(let c of r)await p.handle(e(c),async i=>await o(i,c))},canHandle:n}),k=({collectionName:t,evolve:e,canHandle:o})=>O({collectionName:t,getDocumentId:n=>n.metadata.streamName,evolve:e,canHandle:o});import{dumbo as T}from"@event-driven-io/dumbo";import{v4 as H}from"uuid";var A=(t,e)=>({...e,metadata:{...e.metadata??{},streamName:e.metadata?.streamName??t}}),R=(t,e)=>e.map(o=>A(t,o)),ee=R,te={for:t=>{{let{projection:e,...o}=t,{connectionString:n}=o;return r=>({when:s=>{let p=[],c=async i=>{let P=0n,a="defaultStreamName";for(let d of[...r,...s]){let w=H();p.push({...d,metadata:{globalPosition:P,streamPosition:P,streamName:a,eventId:w,...d.metadata??{}}}),P++}await i.withTransaction(d=>j([e],n,d,p))};return{then:async(i,P)=>{let a=T(o);try{await c(a);let d=await i({pool:a,connectionString:n});d!==void 0&&d===!1&&S(P??"Projection specification didn't match the criteria")}finally{await a.close()}},thenThrows:async(...i)=>{let P=T(o);try{throw await c(P),new v("Handler did not fail as expected")}catch(a){if(a instanceof v)throw a;if(i.length===0)return;if(!x(i[0])){m(i[0](a),`Error didn't match the error condition: ${a?.toString()}`);return}m(a instanceof i[0],`Caught error is not an instance of the expected type: ${a?.toString()}`),i[1]&&m(i[1](a),`Error didn't match the error condition: ${a?.toString()}`)}finally{await P.close()}}}}})}}},b=(t,e)=>async({pool:{execute:o}})=>{let n=await o.query(t);g(e).containsExactlyInAnyOrder(...n.rows)},oe={query:t=>({resultRows:{toBeTheSame:e=>b(t,e)}})};var se={projections:[]},j=async(t,e,o,n)=>{let r=n.map(c=>c.type),s=t.filter(c=>c.canHandle.some(i=>r.includes(i))),p=await o.connection.open();for(let c of s)await c.handle(n,{connectionString:e,client:p,transaction:o,execute:o.execute})},u=t=>h(t),F=(t,...e)=>u({canHandle:e,handle:async(o,n)=>{let r=await t(o,n);await n.execute.batchCommand(r)}}),ce=(t,...e)=>F(async(o,n)=>{let r=[];for(let s of o)r.push(await t(s,n));return r},...e);var l=(t,e)=>{let{pool:o,connectionString:n,inDatabase:r,inCollection:s}=e;return o.withConnection(async p=>{let c=W(n,{connection:p});try{let i=c.db(r).collection(s);return t(i)}finally{await c.close()}})},Q=t=>{let{_id:e,...o}=t;return o},N=(t,e)=>("_id"in e&&y(e._id,t._id,`Document ids are not matching! Expected: ${e._id}, actual: ${t._id}`),D(Q(t),Q(e))),_=(t,e)=>o=>l(async n=>{let r=await n.findOne("withId"in e?{_id:e.withId}:e.matchingFilter);E(r),N(r,t)},{...e,...o}),M=(t,e)=>o=>l(async n=>{let r=await n.find("withId"in e?{_id:e.withId}:e.matchingFilter);y(t.length,r.length,"Different Documents Count than expected");for(let s=0;s<t.length;s++)g(r).contains(t[s])},{...e,...o}),q=(t,e)=>o=>l(async n=>{let r=await n.find("withId"in e?{_id:e.withId}:e.matchingFilter);y(t,r.length,"Different Documents Count than expected")},{...e,...o}),f=t=>e=>l(async o=>{let n=await o.find("withId"in t?{_id:t.withId}:t.matchingFilter);g(n).isNotEmpty()},{...t,...e}),L=t=>e=>l(async o=>{let n=await o.findOne("withId"in t?{_id:t.withId}:t.matchingFilter);E(n)},{...t,...e}),ye={fromCollection:t=>({withId:e=>({toBeEqual:o=>_(o,{withId:e,inCollection:t}),toExist:()=>f({withId:e,inCollection:t}),notToExist:()=>L({withId:e,inCollection:t})}),matching:e=>({toBeTheSame:o=>M(o,{matchingFilter:e,inCollection:t}),toHaveCount:o=>q(o,{matchingFilter:e,inCollection:t}),toExist:()=>f({matchingFilter:e,inCollection:t}),notToExist:()=>L({matchingFilter:e,inCollection:t})})})};export{_ as a,M as b,q as c,f as d,L as e,ye as f,I as g,O as h,k as i,A as j,R as k,ee as l,te as m,b as n,oe as o,se as p,j as q,u as r,F as s,ce as t};
|
|
2
|
-
//# sourceMappingURL=chunk-QNOJWKPZ.mjs.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/projections/pongo/pongoProjectionSpec.ts","../src/eventStore/projections/index.ts","../src/eventStore/projections/pongo/projections.ts","../src/eventStore/projections/postgresProjectionSpec.ts"],"sourcesContent":["import { type Dumbo } from '@event-driven-io/dumbo';\nimport {\n assertDeepEqual,\n assertEqual,\n assertIsNotNull,\n assertThatArray,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoCollection,\n type PongoDocument,\n type PongoFilter,\n type WithId,\n} from '@event-driven-io/pongo';\nimport { type PostgreSQLProjectionAssert } from '..';\n\nexport type PongoAssertOptions = {\n inCollection: string;\n inDatabase?: string;\n};\n\nconst withCollection = (\n handle: (collection: PongoCollection<PongoDocument>) => Promise<void>,\n options: {\n pool: Dumbo;\n connectionString: string;\n } & PongoAssertOptions,\n) => {\n const { pool, connectionString, inDatabase, inCollection } = options;\n\n return pool.withConnection(async (connection) => {\n const pongo = pongoClient(connectionString, {\n connection,\n });\n try {\n const collection = pongo.db(inDatabase).collection(inCollection);\n\n return handle(collection);\n } finally {\n await pongo.close();\n }\n });\n};\n\nconst withoutId = <Doc extends PongoDocument | WithId<PongoDocument>>(\n doc: Doc,\n) => {\n const { _id, ...without } = doc;\n\n return without;\n};\n\nconst assertDocumentsEqual = <\n Doc extends PongoDocument | WithId<PongoDocument>,\n>(\n actual: PongoDocument,\n expected: Doc,\n) => {\n if ('_id' in expected)\n assertEqual(\n expected._id,\n actual._id,\n // eslint-disable-next-line @typescript-eslint/restrict-template-expressions\n `Document ids are not matching! Expected: ${expected._id}, actual: ${actual._id}`,\n );\n\n return assertDeepEqual(withoutId(actual), withoutId(expected));\n};\n\ntype FilterOrId<Doc extends PongoDocument | WithId<PongoDocument>> =\n | { withId: string }\n | {\n matchingFilter: PongoFilter<Doc>;\n };\n\nexport const documentExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n\n assertDocumentsEqual(result, document);\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsAreTheSame =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n documents: Doc[],\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n documents.length,\n result.length,\n 'Different Documents Count than expected',\n );\n\n for (let i = 0; i < documents.length; i++) {\n assertThatArray(result).contains(documents[i]!);\n }\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentsMatchingHaveCount =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n expectedCount: number,\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertEqual(\n expectedCount,\n result.length,\n 'Different Documents Count than expected',\n );\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentMatchingExists =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.find(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertThatArray(result).isNotEmpty();\n },\n { ...options, ...assertOptions },\n );\n\nexport const documentDoesNotExist =\n <Doc extends PongoDocument | WithId<PongoDocument>>(\n options: PongoAssertOptions & FilterOrId<Doc>,\n ): PostgreSQLProjectionAssert =>\n (assertOptions) =>\n withCollection(\n async (collection) => {\n const result = await collection.findOne(\n 'withId' in options\n ? { _id: options.withId }\n : options.matchingFilter,\n );\n\n assertIsNotNull(result);\n },\n { ...options, ...assertOptions },\n );\n\nexport const expectPongoDocuments = {\n fromCollection: (collectionName: string) => {\n return {\n withId: (id: string) => {\n return {\n toBeEqual: <Doc extends PongoDocument | WithId<PongoDocument>>(\n document: Doc,\n ) =>\n documentExists(document, {\n withId: id,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n withId: id,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n withId: id,\n inCollection: collectionName,\n }),\n };\n },\n matching: <Doc extends PongoDocument | WithId<PongoDocument>>(\n filter: PongoFilter<Doc>,\n ) => {\n return {\n toBeTheSame: (documents: Doc[]) =>\n documentsAreTheSame<Doc>(documents, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toHaveCount: (expectedCount: number) =>\n documentsMatchingHaveCount(expectedCount, {\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n toExist: () =>\n documentMatchingExists({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n notToExist: () =>\n documentDoesNotExist({\n matchingFilter: filter,\n inCollection: collectionName,\n }),\n };\n },\n };\n },\n};\n","import {\n type NodePostgresClient,\n type NodePostgresTransaction,\n type SQL,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n projection,\n type CanHandle,\n type Event,\n type ProjectionHandler,\n type ReadEvent,\n type TypedProjectionDefinition,\n} from '@event-driven-io/emmett';\nimport type { PostgresEventStoreOptions } from '../postgreSQLEventStore';\n\nexport type PostgreSQLProjectionHandlerContext = {\n connectionString: string;\n client: NodePostgresClient;\n execute: SQLExecutor;\n transaction: NodePostgresTransaction;\n};\n\nexport type PostgreSQLProjectionHandler<EventType extends Event = Event> =\n ProjectionHandler<EventType, PostgreSQLProjectionHandlerContext>;\n\nexport interface PostgreSQLProjectionDefinition<EventType extends Event = Event>\n extends TypedProjectionDefinition<\n EventType,\n PostgreSQLProjectionHandlerContext\n > {}\n\nexport const defaultPostgreSQLProjectionOptions: PostgresEventStoreOptions = {\n projections: [],\n};\n\nexport const handleProjections = async <EventType extends Event = Event>(\n allProjections: PostgreSQLProjectionDefinition<EventType>[],\n connectionString: string,\n transaction: NodePostgresTransaction,\n events: ReadEvent<EventType>[],\n): Promise<void> => {\n const eventTypes = events.map((e) => e.type);\n\n const projections = allProjections.filter((p) =>\n p.canHandle.some((type) => eventTypes.includes(type)),\n );\n\n const client = (await transaction.connection.open()) as NodePostgresClient;\n\n for (const projection of projections) {\n await projection.handle(events, {\n connectionString,\n client,\n transaction,\n execute: transaction.execute,\n });\n }\n};\n\nexport const postgreSQLProjection = <EventType extends Event>(\n definition: PostgreSQLProjectionDefinition<EventType>,\n): PostgreSQLProjectionDefinition =>\n projection<\n EventType,\n PostgreSQLProjectionHandlerContext,\n PostgreSQLProjectionDefinition<EventType>\n >(definition) as PostgreSQLProjectionDefinition;\n\nexport const postgreSQLRawBatchSQLProjection = <EventType extends Event>(\n handle: (\n events: EventType[],\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL[]> | SQL[],\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const sqls: SQL[] = await handle(events, context);\n\n await context.execute.batchCommand(sqls);\n },\n });\n\nexport const postgreSQLRawSQLProjection = <EventType extends Event>(\n handle: (\n event: EventType,\n context: PostgreSQLProjectionHandlerContext,\n ) => Promise<SQL> | SQL,\n ...canHandle: CanHandle<EventType>\n): PostgreSQLProjectionDefinition =>\n postgreSQLRawBatchSQLProjection<EventType>(\n async (events, context) => {\n const sqls: SQL[] = [];\n\n for (const event of events) {\n sqls.push(await handle(event, context));\n }\n return sqls;\n },\n ...canHandle,\n );\n\nexport * from './pongo';\nexport * from './postgresProjectionSpec';\n","import {\n type CanHandle,\n type Event,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport {\n pongoClient,\n type PongoClient,\n type PongoDocument,\n} from '@event-driven-io/pongo';\nimport {\n postgreSQLProjection,\n type PostgreSQLProjectionDefinition,\n type PostgreSQLProjectionHandlerContext,\n} from '..';\n\nexport type PongoProjectionHandlerContext =\n PostgreSQLProjectionHandlerContext & {\n pongo: PongoClient;\n };\n\nexport type PongoDocumentEvolve<\n Document extends PongoDocument,\n EventType extends Event,\n> =\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Document | null)\n | ((\n document: Document | null,\n event: ReadEvent<EventType>,\n ) => Promise<Document | null>);\n\nexport type PongoProjectionOptions<EventType extends Event> = {\n handle: (\n events: ReadEvent<EventType>[],\n context: PongoProjectionHandlerContext,\n ) => Promise<void>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoProjection = <EventType extends Event>({\n handle,\n canHandle,\n}: PongoProjectionOptions<EventType>): PostgreSQLProjectionDefinition =>\n postgreSQLProjection<EventType>({\n canHandle,\n handle: async (events, context) => {\n const { connectionString, client } = context;\n const pongo = pongoClient(connectionString, { client });\n await handle(events, { ...context, pongo });\n },\n });\n\nexport type PongoMultiStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n> = {\n collectionName: string;\n getDocumentId: (event: ReadEvent<EventType>) => string;\n evolve: PongoDocumentEvolve<Document, EventType>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoMultiStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n>({\n collectionName,\n getDocumentId,\n evolve,\n canHandle,\n}: PongoMultiStreamProjectionOptions<\n Document,\n EventType\n>): PostgreSQLProjectionDefinition =>\n pongoProjection({\n handle: async (events, { pongo }) => {\n const collection = pongo.db().collection<Document>(collectionName);\n\n for (const event of events) {\n await collection.handle(getDocumentId(event), async (document) => {\n return await evolve(document, event);\n });\n }\n },\n canHandle,\n });\n\nexport type PongoSingleStreamProjectionOptions<\n Document extends PongoDocument,\n EventType extends Event,\n> = {\n collectionName: string;\n evolve: PongoDocumentEvolve<Document, EventType>;\n canHandle: CanHandle<EventType>;\n};\n\nexport const pongoSingleStreamProjection = <\n Document extends PongoDocument,\n EventType extends Event,\n>({\n collectionName,\n evolve,\n canHandle,\n}: PongoSingleStreamProjectionOptions<\n Document,\n EventType\n>): PostgreSQLProjectionDefinition =>\n pongoMultiStreamProjection({\n collectionName,\n getDocumentId: (event) => event.metadata.streamName,\n evolve,\n canHandle,\n });\n","import {\n dumbo,\n type Dumbo,\n type DumboOptions,\n type QueryResultRow,\n type SQL,\n} from '@event-driven-io/dumbo';\nimport {\n assertFails,\n AssertionError,\n assertThatArray,\n assertTrue,\n isErrorConstructor,\n type Event,\n type ReadEvent,\n type ReadEventMetadataWithGlobalPosition,\n type ThenThrows,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { handleProjections, type PostgreSQLProjectionDefinition } from '.';\n\nexport type PostgreSQLProjectionSpecEvent<EventType extends Event> =\n EventType & { metadata?: Partial<ReadEventMetadataWithGlobalPosition> };\n\nexport const eventInStream = <EventType extends Event = Event>(\n streamName: string,\n event: PostgreSQLProjectionSpecEvent<EventType>,\n): PostgreSQLProjectionSpecEvent<EventType> => {\n return {\n ...event,\n metadata: {\n ...(event.metadata ?? {}),\n streamName: event.metadata?.streamName ?? streamName,\n },\n };\n};\n\nexport const eventsInStream = <EventType extends Event = Event>(\n streamName: string,\n events: PostgreSQLProjectionSpecEvent<EventType>[],\n): PostgreSQLProjectionSpecEvent<EventType>[] => {\n return events.map((e) => eventInStream(streamName, e));\n};\n\nexport const newEventsInStream = eventsInStream;\n\nexport type PostgreSQLProjectionSpec<EventType extends Event> = (\n givenEvents: PostgreSQLProjectionSpecEvent<EventType>[],\n) => {\n when: (events: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n then: (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ) => Promise<void>;\n thenThrows: <ErrorType extends Error = Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ) => Promise<void>;\n };\n};\n\nexport type PostgreSQLProjectionAssert = (options: {\n pool: Dumbo;\n connectionString: string;\n}) => Promise<void | boolean>;\n\nexport type PostgreSQLProjectionSpecOptions = {\n projection: PostgreSQLProjectionDefinition;\n} & DumboOptions;\n\nexport const PostgreSQLProjectionSpec = {\n for: <EventType extends Event>(\n options: PostgreSQLProjectionSpecOptions,\n ): PostgreSQLProjectionSpec<EventType> => {\n {\n const { projection, ...dumoOptions } = options;\n const { connectionString } = dumoOptions;\n\n return (givenEvents: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n return {\n when: (events: PostgreSQLProjectionSpecEvent<EventType>[]) => {\n const allEvents: ReadEvent<\n EventType,\n ReadEventMetadataWithGlobalPosition\n >[] = [];\n\n const run = async (pool: Dumbo) => {\n let globalPosition = 0n;\n const streamName = 'defaultStreamName';\n\n for (const event of [...givenEvents, ...events]) {\n const eventId = uuid();\n allEvents.push({\n ...event,\n metadata: {\n ...{\n globalPosition,\n streamPosition: globalPosition,\n streamName,\n eventId,\n },\n ...(event.metadata ?? {}),\n },\n });\n\n globalPosition++;\n }\n\n await pool.withTransaction((transaction) =>\n handleProjections(\n [projection],\n connectionString,\n transaction,\n allEvents,\n ),\n );\n };\n\n return {\n then: async (\n assert: PostgreSQLProjectionAssert,\n message?: string,\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n\n const succeeded = await assert({ pool, connectionString });\n\n if (succeeded !== undefined && succeeded === false)\n assertFails(\n message ??\n \"Projection specification didn't match the criteria\",\n );\n } finally {\n await pool.close();\n }\n },\n thenThrows: async <ErrorType extends Error>(\n ...args: Parameters<ThenThrows<ErrorType>>\n ): Promise<void> => {\n const pool = dumbo(dumoOptions);\n try {\n await run(pool);\n throw new AssertionError('Handler did not fail as expected');\n } catch (error) {\n if (error instanceof AssertionError) throw error;\n\n if (args.length === 0) return;\n\n if (!isErrorConstructor(args[0])) {\n assertTrue(\n args[0](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n return;\n }\n\n assertTrue(\n error instanceof args[0],\n `Caught error is not an instance of the expected type: ${error?.toString()}`,\n );\n\n if (args[1]) {\n assertTrue(\n args[1](error as ErrorType),\n `Error didn't match the error condition: ${error?.toString()}`,\n );\n }\n } finally {\n await pool.close();\n }\n },\n };\n },\n };\n };\n }\n },\n};\n\nexport const assertSQLQueryResultMatches =\n <T extends QueryResultRow>(sql: SQL, rows: T[]): PostgreSQLProjectionAssert =>\n async ({ pool: { execute } }) => {\n const result = await execute.query<T>(sql);\n\n assertThatArray(rows).containsExactlyInAnyOrder(...result.rows);\n };\n\nexport const expectSQL = {\n query: (sql: SQL) => ({\n resultRows: {\n toBeTheSame: <T extends QueryResultRow>(rows: T[]) =>\n assertSQLQueryResultMatches(sql, rows),\n },\n }),\n};\n"],"mappings":"iGAAA,MAA2B,yBAO3B,OACE,eAAAA,MAKK,yBCbP,MAKO,yBCAP,OACE,eAAAC,MAGK,yBAiCA,IAAMC,EAAkB,CAA0B,CACvD,OAAAC,EACA,UAAAC,CACF,IACEC,EAAgC,CAC9B,UAAAD,EACA,OAAQ,MAAOE,EAAQC,IAAY,CACjC,GAAM,CAAE,iBAAAC,EAAkB,OAAAC,CAAO,EAAIF,EAC/BG,EAAQC,EAAYH,EAAkB,CAAE,OAAAC,CAAO,CAAC,EACtD,MAAMN,EAAOG,EAAQ,CAAE,GAAGC,EAAS,MAAAG,CAAM,CAAC,CAC5C,CACF,CAAC,EAYUE,EAA6B,CAGxC,CACA,eAAAC,EACA,cAAAC,EACA,OAAAC,EACA,UAAAX,CACF,IAIEF,EAAgB,CACd,OAAQ,MAAOI,EAAQ,CAAE,MAAAI,CAAM,IAAM,CACnC,IAAMM,EAAaN,EAAM,GAAG,EAAE,WAAqBG,CAAc,EAEjE,QAAWI,KAASX,EAClB,MAAMU,EAAW,OAAOF,EAAcG,CAAK,EAAG,MAAOC,GAC5C,MAAMH,EAAOG,EAAUD,CAAK,CACpC,CAEL,EACA,UAAAb,CACF,CAAC,EAWUe,EAA8B,CAGzC,CACA,eAAAN,EACA,OAAAE,EACA,UAAAX,CACF,IAIEQ,EAA2B,CACzB,eAAAC,EACA,cAAgBI,GAAUA,EAAM,SAAS,WACzC,OAAAF,EACA,UAAAX,CACF,CAAC,ECnHH,OACE,SAAAgB,MAKK,yBAYP,OAAS,MAAMC,MAAY,OAMpB,IAAMC,EAAgB,CAC3BC,EACAC,KAEO,CACL,GAAGA,EACH,SAAU,CACR,GAAIA,EAAM,UAAY,CAAC,EACvB,WAAYA,EAAM,UAAU,YAAcD,CAC5C,CACF,GAGWE,EAAiB,CAC5BF,EACAG,IAEOA,EAAO,IAAKC,GAAML,EAAcC,EAAYI,CAAC,CAAC,EAG1CC,GAAoBH,EAyBpBI,GAA2B,CACtC,IACEC,GACwC,CACxC,CACE,GAAM,CAAE,WAAAC,EAAY,GAAGC,CAAY,EAAIF,EACjC,CAAE,iBAAAG,CAAiB,EAAID,EAE7B,OAAQE,IACC,CACL,KAAOR,GAAuD,CAC5D,IAAMS,EAGA,CAAC,EAEDC,EAAM,MAAOC,GAAgB,CACjC,IAAIC,EAAiB,GACff,EAAa,oBAEnB,QAAWC,IAAS,CAAC,GAAGU,EAAa,GAAGR,CAAM,EAAG,CAC/C,IAAMa,EAAUC,EAAK,EACrBL,EAAU,KAAK,CACb,GAAGX,EACH,SAAU,CAEN,eAAAc,EACA,eAAgBA,EAChB,WAAAf,EACA,QAAAgB,EAEF,GAAIf,EAAM,UAAY,CAAC,CACzB,CACF,CAAC,EAEDc,GACF,CAEA,MAAMD,EAAK,gBAAiBI,GAC1BC,EACE,CAACX,CAAU,EACXE,EACAQ,EACAN,CACF,CACF,CACF,EAEA,MAAO,CACL,KAAM,MACJQ,EACAC,IACkB,CAClB,IAAMP,EAAOQ,EAAMb,CAAW,EAC9B,GAAI,CACF,MAAMI,EAAIC,CAAI,EAEd,IAAMS,EAAY,MAAMH,EAAO,CAAE,KAAAN,EAAM,iBAAAJ,CAAiB,CAAC,EAErDa,IAAc,QAAaA,IAAc,IAC3CC,EACEH,GACE,oDACJ,CACJ,QAAE,CACA,MAAMP,EAAK,MAAM,CACnB,CACF,EACA,WAAY,SACPW,IACe,CAClB,IAAMX,EAAOQ,EAAMb,CAAW,EAC9B,GAAI,CACF,YAAMI,EAAIC,CAAI,EACR,IAAIY,EAAe,kCAAkC,CAC7D,OAASC,EAAO,CACd,GAAIA,aAAiBD,EAAgB,MAAMC,EAE3C,GAAIF,EAAK,SAAW,EAAG,OAEvB,GAAI,CAACG,EAAmBH,EAAK,CAAC,CAAC,EAAG,CAChCI,EACEJ,EAAK,CAAC,EAAEE,CAAkB,EAC1B,2CAA2CA,GAAO,SAAS,CAAC,EAC9D,EACA,MACF,CAEAE,EACEF,aAAiBF,EAAK,CAAC,EACvB,yDAAyDE,GAAO,SAAS,CAAC,EAC5E,EAEIF,EAAK,CAAC,GACRI,EACEJ,EAAK,CAAC,EAAEE,CAAkB,EAC1B,2CAA2CA,GAAO,SAAS,CAAC,EAC9D,CAEJ,QAAE,CACA,MAAMb,EAAK,MAAM,CACnB,CACF,CACF,CACF,CACF,EAEJ,CACF,CACF,EAEagB,EACX,CAA2BC,EAAUC,IACrC,MAAO,CAAE,KAAM,CAAE,QAAAC,CAAQ,CAAE,IAAM,CAC/B,IAAMC,EAAS,MAAMD,EAAQ,MAASF,CAAG,EAEzCI,EAAgBH,CAAI,EAAE,0BAA0B,GAAGE,EAAO,IAAI,CAChE,EAEWE,GAAY,CACvB,MAAQL,IAAc,CACpB,WAAY,CACV,YAAwCC,GACtCF,EAA4BC,EAAKC,CAAI,CACzC,CACF,EACF,EFnKO,IAAMK,GAAgE,CAC3E,YAAa,CAAC,CAChB,EAEaC,EAAoB,MAC/BC,EACAC,EACAC,EACAC,IACkB,CAClB,IAAMC,EAAaD,EAAO,IAAKE,GAAMA,EAAE,IAAI,EAErCC,EAAcN,EAAe,OAAQO,GACzCA,EAAE,UAAU,KAAMC,GAASJ,EAAW,SAASI,CAAI,CAAC,CACtD,EAEMC,EAAU,MAAMP,EAAY,WAAW,KAAK,EAElD,QAAWQ,KAAcJ,EACvB,MAAMI,EAAW,OAAOP,EAAQ,CAC9B,iBAAAF,EACA,OAAAQ,EACA,YAAAP,EACA,QAASA,EAAY,OACvB,CAAC,CAEL,EAEaS,EACXC,GAEAC,EAIED,CAAU,EAEDE,EAAkC,CAC7CC,KAIGC,IAEHL,EAAgC,CAC9B,UAAAK,EACA,OAAQ,MAAOb,EAAQc,IAAY,CACjC,IAAMC,EAAc,MAAMH,EAAOZ,EAAQc,CAAO,EAEhD,MAAMA,EAAQ,QAAQ,aAAaC,CAAI,CACzC,CACF,CAAC,EAEUC,GAA6B,CACxCJ,KAIGC,IAEHF,EACE,MAAOX,EAAQc,IAAY,CACzB,IAAMC,EAAc,CAAC,EAErB,QAAWE,KAASjB,EAClBe,EAAK,KAAK,MAAMH,EAAOK,EAAOH,CAAO,CAAC,EAExC,OAAOC,CACT,EACA,GAAGF,CACL,EDjFF,IAAMK,EAAiB,CACrBC,EACAC,IAIG,CACH,GAAM,CAAE,KAAAC,EAAM,iBAAAC,EAAkB,WAAAC,EAAY,aAAAC,CAAa,EAAIJ,EAE7D,OAAOC,EAAK,eAAe,MAAOI,GAAe,CAC/C,IAAMC,EAAQC,EAAYL,EAAkB,CAC1C,WAAAG,CACF,CAAC,EACD,GAAI,CACF,IAAMG,EAAaF,EAAM,GAAGH,CAAU,EAAE,WAAWC,CAAY,EAE/D,OAAOL,EAAOS,CAAU,CAC1B,QAAE,CACA,MAAMF,EAAM,MAAM,CACpB,CACF,CAAC,CACH,EAEMG,EACJC,GACG,CACH,GAAM,CAAE,IAAAC,EAAK,GAAGC,CAAQ,EAAIF,EAE5B,OAAOE,CACT,EAEMC,EAAuB,CAG3BC,EACAC,KAEI,QAASA,GACXC,EACED,EAAS,IACTD,EAAO,IAEP,4CAA4CC,EAAS,GAAG,aAAaD,EAAO,GAAG,EACjF,EAEKG,EAAgBR,EAAUK,CAAM,EAAGL,EAAUM,CAAQ,CAAC,GASlDG,EACX,CACEC,EACAnB,IAEDoB,GACCtB,EACE,MAAOU,GAAe,CACpB,IAAMa,EAAS,MAAMb,EAAW,QAC9B,WAAYR,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAsB,EAAgBD,CAAM,EAEtBR,EAAqBQ,EAAQF,CAAQ,CACvC,EACA,CAAE,GAAGnB,EAAS,GAAGoB,CAAc,CACjC,EAESG,EACX,CACEC,EACAxB,IAEDoB,GACCtB,EACE,MAAOU,GAAe,CACpB,IAAMa,EAAS,MAAMb,EAAW,KAC9B,WAAYR,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAgB,EACEQ,EAAU,OACVH,EAAO,OACP,yCACF,EAEA,QAASI,EAAI,EAAGA,EAAID,EAAU,OAAQC,IACpCC,EAAgBL,CAAM,EAAE,SAASG,EAAUC,CAAC,CAAE,CAElD,EACA,CAAE,GAAGzB,EAAS,GAAGoB,CAAc,CACjC,EAESO,EACX,CACEC,EACA5B,IAEDoB,GACCtB,EACE,MAAOU,GAAe,CACpB,IAAMa,EAAS,MAAMb,EAAW,KAC9B,WAAYR,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAgB,EACEY,EACAP,EAAO,OACP,yCACF,CACF,EACA,CAAE,GAAGrB,EAAS,GAAGoB,CAAc,CACjC,EAESS,EAET7B,GAEDoB,GACCtB,EACE,MAAOU,GAAe,CACpB,IAAMa,EAAS,MAAMb,EAAW,KAC9B,WAAYR,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEA0B,EAAgBL,CAAM,EAAE,WAAW,CACrC,EACA,CAAE,GAAGrB,EAAS,GAAGoB,CAAc,CACjC,EAESU,EAET9B,GAEDoB,GACCtB,EACE,MAAOU,GAAe,CACpB,IAAMa,EAAS,MAAMb,EAAW,QAC9B,WAAYR,EACR,CAAE,IAAKA,EAAQ,MAAO,EACtBA,EAAQ,cACd,EAEAsB,EAAgBD,CAAM,CACxB,EACA,CAAE,GAAGrB,EAAS,GAAGoB,CAAc,CACjC,EAESW,GAAuB,CAClC,eAAiBC,IACR,CACL,OAASC,IACA,CACL,UACEd,GAEAD,EAAeC,EAAU,CACvB,OAAQc,EACR,aAAcD,CAChB,CAAC,EACH,QAAS,IACPH,EAAuB,CACrB,OAAQI,EACR,aAAcD,CAChB,CAAC,EACH,WAAY,IACVF,EAAqB,CACnB,OAAQG,EACR,aAAcD,CAChB,CAAC,CACL,GAEF,SACEE,IAEO,CACL,YAAcV,GACZD,EAAyBC,EAAW,CAClC,eAAgBU,EAChB,aAAcF,CAChB,CAAC,EACH,YAAcJ,GACZD,EAA2BC,EAAe,CACxC,eAAgBM,EAChB,aAAcF,CAChB,CAAC,EACH,QAAS,IACPH,EAAuB,CACrB,eAAgBK,EAChB,aAAcF,CAChB,CAAC,EACH,WAAY,IACVF,EAAqB,CACnB,eAAgBI,EAChB,aAAcF,CAChB,CAAC,CACL,EAEJ,EAEJ","names":["pongoClient","pongoClient","pongoProjection","handle","canHandle","postgreSQLProjection","events","context","connectionString","client","pongo","pongoClient","pongoMultiStreamProjection","collectionName","getDocumentId","evolve","collection","event","document","pongoSingleStreamProjection","dumbo","uuid","eventInStream","streamName","event","eventsInStream","events","e","newEventsInStream","PostgreSQLProjectionSpec","options","projection","dumoOptions","connectionString","givenEvents","allEvents","run","pool","globalPosition","eventId","uuid","transaction","handleProjections","assert","message","dumbo","succeeded","p","args","l","error","w","s","assertSQLQueryResultMatches","sql","rows","execute","result","q","expectSQL","defaultPostgreSQLProjectionOptions","handleProjections","allProjections","connectionString","transaction","events","eventTypes","e","projections","p","type","client","projection","postgreSQLProjection","definition","i","postgreSQLRawBatchSQLProjection","handle","canHandle","context","sqls","postgreSQLRawSQLProjection","event","withCollection","handle","options","pool","connectionString","inDatabase","inCollection","connection","pongo","pongoClient","collection","withoutId","doc","_id","without","assertDocumentsEqual","actual","expected","c","x","documentExists","document","assertOptions","result","y","documentsAreTheSame","documents","i","q","documentsMatchingHaveCount","expectedCount","documentMatchingExists","documentDoesNotExist","expectPongoDocuments","collectionName","id","filter"]}
|
package/dist/chunk-SS2LQM3B.js
DELETED
|
@@ -1,2 +0,0 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true});var t="emt",o= exports.b ="global",n= exports.c ="emt:default",a= exports.d ={module:`${t}:module:${o}`,tenant:`${t}:tenant:${o}`},e={partition:{name:"partition"},isArchived:{name:"is_archived"}},i= exports.e ={name:`${t}_streams`,columns:{partition:e.partition,isArchived:e.isArchived}},s= exports.f ={name:`${t}_events`,columns:{partition:e.partition,isArchived:e.isArchived}};exports.a = t; exports.b = o; exports.c = n; exports.d = a; exports.e = i; exports.f = s;
|
|
2
|
-
//# sourceMappingURL=chunk-SS2LQM3B.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/schema/typing.ts"],"names":["emmettPrefix","globalTag","defaultTag","globalNames","columns","streamsTable","eventsTable"],"mappings":"AAAO,IAAMA,EAAe,MAEfC,EAAY,SACZC,EAAa,cAEbC,EAAc,CACzB,OAAQ,GAAGH,CAAY,WAAWC,CAAS,GAC3C,OAAQ,GAAGD,CAAY,WAAWC,CAAS,EAC7C,EAEMG,EAAU,CACd,UAAW,CACT,KAAM,WACR,EACA,WAAY,CAAE,KAAM,aAAc,CACpC,EAEaC,EAAe,CAC1B,KAAM,GAAGL,CAAY,WACrB,QAAS,CACP,UAAWI,EAAQ,UACnB,WAAYA,EAAQ,UACtB,CACF,EAEaE,EAAc,CACzB,KAAM,GAAGN,CAAY,UACrB,QAAS,CACP,UAAWI,EAAQ,UACnB,WAAYA,EAAQ,UACtB,CACF","sourcesContent":["export const emmettPrefix = 'emt';\n\nexport const globalTag = 'global';\nexport const defaultTag = 'emt:default';\n\nexport const globalNames = {\n module: `${emmettPrefix}:module:${globalTag}`,\n tenant: `${emmettPrefix}:tenant:${globalTag}`,\n};\n\nconst columns = {\n partition: {\n name: 'partition',\n },\n isArchived: { name: 'is_archived' },\n};\n\nexport const streamsTable = {\n name: `${emmettPrefix}_streams`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n\nexport const eventsTable = {\n name: `${emmettPrefix}_events`,\n columns: {\n partition: columns.partition,\n isArchived: columns.isArchived,\n },\n};\n"]}
|
package/dist/chunk-ST3FNDJ5.mjs
DELETED
|
@@ -1,94 +0,0 @@
|
|
|
1
|
-
import{c as A,e as _,f as T}from"./chunk-ABNBEUC6.mjs";import{a as v,l as R,m as S,n as g}from"./chunk-HIX4PKIP.mjs";import{rawSql as N,single as b,sql as s}from"@event-driven-io/dumbo";import{v4 as I}from"uuid";var D=N(`CREATE OR REPLACE FUNCTION emt_append_event(
|
|
2
|
-
v_event_ids text[],
|
|
3
|
-
v_events_data jsonb[],
|
|
4
|
-
v_events_metadata jsonb[],
|
|
5
|
-
v_event_schema_versions text[],
|
|
6
|
-
v_event_types text[],
|
|
7
|
-
v_stream_id text,
|
|
8
|
-
v_stream_type text,
|
|
9
|
-
v_expected_stream_position bigint DEFAULT NULL,
|
|
10
|
-
v_partition text DEFAULT emt_sanitize_name('default_partition')
|
|
11
|
-
) RETURNS TABLE (
|
|
12
|
-
success boolean,
|
|
13
|
-
next_stream_position bigint,
|
|
14
|
-
last_global_position bigint,
|
|
15
|
-
transaction_id xid8
|
|
16
|
-
) LANGUAGE plpgsql
|
|
17
|
-
AS $$
|
|
18
|
-
DECLARE
|
|
19
|
-
v_next_stream_position bigint;
|
|
20
|
-
v_position bigint;
|
|
21
|
-
v_updated_rows int;
|
|
22
|
-
v_transaction_id xid8;
|
|
23
|
-
v_last_global_position bigint;
|
|
24
|
-
BEGIN
|
|
25
|
-
IF v_expected_stream_position IS NULL THEN
|
|
26
|
-
SELECT COALESCE(max(stream_position), 0) INTO v_expected_stream_position
|
|
27
|
-
FROM ${_.name}
|
|
28
|
-
WHERE stream_id = v_stream_id AND partition = v_partition;
|
|
29
|
-
END IF;
|
|
30
|
-
|
|
31
|
-
v_next_stream_position := v_expected_stream_position + array_upper(v_events_data, 1);
|
|
32
|
-
v_transaction_id := pg_current_xact_id();
|
|
33
|
-
|
|
34
|
-
WITH ev AS (
|
|
35
|
-
SELECT row_number() OVER () + v_expected_stream_position AS stream_position,
|
|
36
|
-
event_data,
|
|
37
|
-
event_metadata,
|
|
38
|
-
schema_version,
|
|
39
|
-
event_id,
|
|
40
|
-
event_type
|
|
41
|
-
FROM (
|
|
42
|
-
SELECT *
|
|
43
|
-
FROM
|
|
44
|
-
unnest(v_event_ids, v_events_data, v_events_metadata, v_event_schema_versions, v_event_types)
|
|
45
|
-
AS event(event_id, event_data, event_metadata, schema_version, event_type)
|
|
46
|
-
) AS event
|
|
47
|
-
),
|
|
48
|
-
all_events_insert AS (
|
|
49
|
-
INSERT INTO ${T.name}
|
|
50
|
-
(stream_id, stream_position, partition, event_data, event_metadata, event_schema_version, event_type, event_id, transaction_id)
|
|
51
|
-
SELECT
|
|
52
|
-
v_stream_id, ev.stream_position, v_partition, ev.event_data, ev.event_metadata, ev.schema_version, ev.event_type, ev.event_id, v_transaction_id
|
|
53
|
-
FROM ev
|
|
54
|
-
RETURNING global_position
|
|
55
|
-
)
|
|
56
|
-
SELECT
|
|
57
|
-
max(global_position) INTO v_last_global_position
|
|
58
|
-
FROM
|
|
59
|
-
all_events_insert;
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
IF v_expected_stream_position = 0 THEN
|
|
63
|
-
INSERT INTO ${_.name}
|
|
64
|
-
(stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)
|
|
65
|
-
VALUES
|
|
66
|
-
(v_stream_id, v_next_stream_position, v_partition, v_stream_type, '{}', FALSE);
|
|
67
|
-
ELSE
|
|
68
|
-
UPDATE ${_.name} as s
|
|
69
|
-
SET stream_position = v_next_stream_position
|
|
70
|
-
WHERE stream_id = v_stream_id AND stream_position = v_expected_stream_position AND partition = v_partition AND is_archived = FALSE;
|
|
71
|
-
|
|
72
|
-
get diagnostics v_updated_rows = row_count;
|
|
73
|
-
|
|
74
|
-
IF v_updated_rows = 0 THEN
|
|
75
|
-
RETURN QUERY SELECT FALSE, NULL::bigint, NULL::bigint, NULL::xid8;
|
|
76
|
-
RETURN;
|
|
77
|
-
END IF;
|
|
78
|
-
END IF;
|
|
79
|
-
|
|
80
|
-
RETURN QUERY SELECT TRUE, v_next_stream_position, v_last_global_position, v_transaction_id;
|
|
81
|
-
END;
|
|
82
|
-
$$;
|
|
83
|
-
`),h=(t,a,r,n,i)=>t.withTransaction(async e=>{let{execute:x}=e;if(n.length===0)return{success:!1,result:{success:!1}};let p;try{let o=y(i?.expectedStreamVersion),c=n.map((u,L)=>({...u,metadata:{streamName:a,eventId:I(),streamPosition:BigInt(L),...u.metadata}}));p=await C(x,a,r,c,{expectedStreamVersion:o}),i?.preCommitHook&&await i.preCommitHook(e,c)}catch(o){if(!f(o))throw o;p={success:!1,last_global_position:null,next_stream_position:null,transaction_id:null}}let{success:d,next_stream_position:m,last_global_position:l,transaction_id:E}=p;return{success:d,result:d&&m&&l&&E?{success:!0,nextStreamPosition:BigInt(m),lastGlobalPosition:BigInt(l),transactionId:E}:{success:!1}}}),y=t=>t===void 0||t===g||t==S||t==R?null:t,f=t=>t instanceof Error&&"code"in t&&t.code==="23505",C=(t,a,r,n,i)=>b(t.command(s(`SELECT * FROM emt_append_event(
|
|
84
|
-
ARRAY[%s]::text[],
|
|
85
|
-
ARRAY[%s]::jsonb[],
|
|
86
|
-
ARRAY[%s]::jsonb[],
|
|
87
|
-
ARRAY[%s]::text[],
|
|
88
|
-
ARRAY[%s]::text[],
|
|
89
|
-
%L::text,
|
|
90
|
-
%L::text,
|
|
91
|
-
%s::bigint,
|
|
92
|
-
%L::text
|
|
93
|
-
)`,n.map(e=>s("%L",e.metadata.eventId)).join(","),n.map(e=>s("%L",v.stringify(e.data))).join(","),n.map(e=>s("%L",v.stringify(e.metadata??{}))).join(","),n.map(()=>"'1'").join(","),n.map(e=>s("%L",e.type)).join(","),a,r,i?.expectedStreamVersion??"NULL",i?.partition??A)));export{D as a,h as b};
|
|
94
|
-
//# sourceMappingURL=chunk-ST3FNDJ5.mjs.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/eventStore/schema/appendToStream.ts"],"sourcesContent":["import {\n rawSql,\n single,\n sql,\n type NodePostgresPool,\n type NodePostgresTransaction,\n type SQLExecutor,\n} from '@event-driven-io/dumbo';\nimport {\n JSONParser,\n NO_CONCURRENCY_CHECK,\n STREAM_DOES_NOT_EXIST,\n STREAM_EXISTS,\n type AppendToStreamOptions,\n type Event,\n type ExpectedStreamVersion,\n type ReadEvent,\n} from '@event-driven-io/emmett';\nimport { v4 as uuid } from 'uuid';\nimport { defaultTag, eventsTable, streamsTable } from './typing';\n\nexport const appendEventsSQL = rawSql(\n `CREATE OR REPLACE FUNCTION emt_append_event(\n v_event_ids text[],\n v_events_data jsonb[],\n v_events_metadata jsonb[],\n v_event_schema_versions text[],\n v_event_types text[],\n v_stream_id text,\n v_stream_type text,\n v_expected_stream_position bigint DEFAULT NULL,\n v_partition text DEFAULT emt_sanitize_name('default_partition')\n ) RETURNS TABLE (\n success boolean,\n next_stream_position bigint,\n last_global_position bigint,\n transaction_id xid8\n ) LANGUAGE plpgsql\n AS $$\n DECLARE\n v_next_stream_position bigint;\n v_position bigint;\n v_updated_rows int;\n v_transaction_id xid8;\n v_last_global_position bigint;\n BEGIN\n IF v_expected_stream_position IS NULL THEN\n SELECT COALESCE(max(stream_position), 0) INTO v_expected_stream_position\n FROM ${streamsTable.name}\n WHERE stream_id = v_stream_id AND partition = v_partition;\n END IF;\n\n v_next_stream_position := v_expected_stream_position + array_upper(v_events_data, 1);\n v_transaction_id := pg_current_xact_id();\n\n WITH ev AS (\n SELECT row_number() OVER () + v_expected_stream_position AS stream_position, \n event_data, \n event_metadata, \n schema_version, \n event_id, \n event_type\n FROM (\n SELECT *\n FROM \n unnest(v_event_ids, v_events_data, v_events_metadata, v_event_schema_versions, v_event_types) \n AS event(event_id, event_data, event_metadata, schema_version, event_type)\n ) AS event\n ),\n all_events_insert AS (\n INSERT INTO ${eventsTable.name}\n (stream_id, stream_position, partition, event_data, event_metadata, event_schema_version, event_type, event_id, transaction_id)\n SELECT \n v_stream_id, ev.stream_position, v_partition, ev.event_data, ev.event_metadata, ev.schema_version, ev.event_type, ev.event_id, v_transaction_id\n FROM ev\n RETURNING global_position\n )\n SELECT \n max(global_position) INTO v_last_global_position \n FROM \n all_events_insert;\n\n\n IF v_expected_stream_position = 0 THEN\n INSERT INTO ${streamsTable.name}\n (stream_id, stream_position, partition, stream_type, stream_metadata, is_archived)\n VALUES\n (v_stream_id, v_next_stream_position, v_partition, v_stream_type, '{}', FALSE);\n ELSE\n UPDATE ${streamsTable.name} as s \n SET stream_position = v_next_stream_position\n WHERE stream_id = v_stream_id AND stream_position = v_expected_stream_position AND partition = v_partition AND is_archived = FALSE;\n\n get diagnostics v_updated_rows = row_count;\n\n IF v_updated_rows = 0 THEN\n RETURN QUERY SELECT FALSE, NULL::bigint, NULL::bigint, NULL::xid8;\n RETURN;\n END IF;\n END IF;\n\n RETURN QUERY SELECT TRUE, v_next_stream_position, v_last_global_position, v_transaction_id;\n END;\n $$;\n `,\n);\n\ntype AppendEventResult =\n | {\n success: true;\n nextStreamPosition: bigint;\n lastGlobalPosition: bigint;\n transactionId: string;\n }\n | { success: false };\n\nexport const appendToStream = (\n pool: NodePostgresPool,\n streamName: string,\n streamType: string,\n events: Event[],\n options?: AppendToStreamOptions & {\n partition?: string;\n preCommitHook?: (\n transaction: NodePostgresTransaction,\n events: ReadEvent[],\n ) => Promise<void>;\n },\n): Promise<AppendEventResult> =>\n pool.withTransaction<AppendEventResult>(async (transaction) => {\n const { execute } = transaction;\n\n if (events.length === 0)\n return { success: false, result: { success: false } };\n\n let appendResult: AppendEventSqlResult;\n\n try {\n const expectedStreamVersion = toExpectedVersion(\n options?.expectedStreamVersion,\n );\n\n const eventsToAppend: ReadEvent[] = events.map((e, i) => ({\n ...e,\n metadata: {\n streamName,\n eventId: uuid(),\n streamPosition: BigInt(i),\n ...e.metadata,\n },\n }));\n\n // TODO: return global positions from append raw and other generated data\n appendResult = await appendEventsRaw(\n execute,\n streamName,\n streamType,\n eventsToAppend,\n {\n expectedStreamVersion,\n },\n );\n\n if (options?.preCommitHook)\n await options.preCommitHook(transaction, eventsToAppend);\n } catch (error) {\n if (!isOptimisticConcurrencyError(error)) throw error;\n\n appendResult = {\n success: false,\n last_global_position: null,\n next_stream_position: null,\n transaction_id: null,\n };\n }\n\n const {\n success,\n next_stream_position,\n last_global_position,\n transaction_id,\n } = appendResult;\n\n return {\n success,\n result:\n success &&\n next_stream_position &&\n last_global_position &&\n transaction_id\n ? {\n success: true,\n nextStreamPosition: BigInt(next_stream_position),\n lastGlobalPosition: BigInt(last_global_position),\n transactionId: transaction_id,\n }\n : { success: false },\n };\n });\n\nconst toExpectedVersion = (\n expected: ExpectedStreamVersion | undefined,\n): bigint | null => {\n if (expected === undefined) return null;\n\n if (expected === NO_CONCURRENCY_CHECK) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_DOES_NOT_EXIST) return null;\n\n // TODO: this needs to be fixed\n if (expected == STREAM_EXISTS) return null;\n\n return expected as bigint;\n};\n\nconst isOptimisticConcurrencyError = (error: unknown): boolean =>\n error instanceof Error && 'code' in error && error.code === '23505';\n\ntype AppendEventSqlResult = {\n success: boolean;\n next_stream_position: string | null;\n last_global_position: string | null;\n transaction_id: string | null | undefined;\n};\n\nconst appendEventsRaw = (\n execute: SQLExecutor,\n streamId: string,\n streamType: string,\n events: ReadEvent[],\n options?: {\n expectedStreamVersion: bigint | null;\n partition?: string;\n },\n): Promise<AppendEventSqlResult> =>\n single(\n execute.command<AppendEventSqlResult>(\n sql(\n `SELECT * FROM emt_append_event(\n ARRAY[%s]::text[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::jsonb[],\n ARRAY[%s]::text[],\n ARRAY[%s]::text[],\n %L::text,\n %L::text,\n %s::bigint,\n %L::text\n )`,\n events.map((e) => sql('%L', e.metadata.eventId)).join(','),\n events.map((e) => sql('%L', JSONParser.stringify(e.data))).join(','),\n events\n .map((e) => sql('%L', JSONParser.stringify(e.metadata ?? {})))\n .join(','),\n events.map(() => `'1'`).join(','),\n events.map((e) => sql('%L', e.type)).join(','),\n streamId,\n streamType,\n options?.expectedStreamVersion ?? 'NULL',\n options?.partition ?? defaultTag,\n ),\n ),\n );\n"],"mappings":"qHAAA,OACE,UAAAA,EACA,UAAAC,EACA,OAAAC,MAIK,yBAWP,OAAS,MAAMC,MAAY,OAGpB,IAAMC,EAAkBC,EAC7B;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iBA0BeC,EAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAsBVC,EAAY,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,wBAchBD,EAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA,mBAKtBA,EAAa,IAAI;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAgBpC,EAWaE,EAAiB,CAC5BC,EACAC,EACAC,EACAC,EACAC,IAQAJ,EAAK,gBAAmC,MAAOK,GAAgB,CAC7D,GAAM,CAAE,QAAAC,CAAQ,EAAID,EAEpB,GAAIF,EAAO,SAAW,EACpB,MAAO,CAAE,QAAS,GAAO,OAAQ,CAAE,QAAS,EAAM,CAAE,EAEtD,IAAII,EAEJ,GAAI,CACF,IAAMC,EAAwBC,EAC5BL,GAAS,qBACX,EAEMM,EAA8BP,EAAO,IAAI,CAACQ,EAAGC,KAAO,CACxD,GAAGD,EACH,SAAU,CACR,WAAAV,EACA,QAASY,EAAK,EACd,eAAgB,OAAOD,CAAC,EACxB,GAAGD,EAAE,QACP,CACF,EAAE,EAGFJ,EAAe,MAAMO,EACnBR,EACAL,EACAC,EACAQ,EACA,CACE,sBAAAF,CACF,CACF,EAEIJ,GAAS,eACX,MAAMA,EAAQ,cAAcC,EAAaK,CAAc,CAC3D,OAASK,EAAO,CACd,GAAI,CAACC,EAA6BD,CAAK,EAAG,MAAMA,EAEhDR,EAAe,CACb,QAAS,GACT,qBAAsB,KACtB,qBAAsB,KACtB,eAAgB,IAClB,CACF,CAEA,GAAM,CACJ,QAAAU,EACA,qBAAAC,EACA,qBAAAC,EACA,eAAAC,CACF,EAAIb,EAEJ,MAAO,CACL,QAAAU,EACA,OACEA,GACAC,GACAC,GACAC,EACI,CACE,QAAS,GACT,mBAAoB,OAAOF,CAAoB,EAC/C,mBAAoB,OAAOC,CAAoB,EAC/C,cAAeC,CACjB,EACA,CAAE,QAAS,EAAM,CACzB,CACF,CAAC,EAEGX,EACJY,GAEIA,IAAa,QAEbA,IAAaC,GAGbD,GAAYE,GAGZF,GAAYG,EAAsB,KAE/BH,EAGHL,EAAgCD,GACpCA,aAAiB,OAAS,SAAUA,GAASA,EAAM,OAAS,QASxDD,EAAkB,CACtBR,EACAmB,EACAvB,EACAC,EACAC,IAKAsB,EACEpB,EAAQ,QACNqB,EACE;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,iBAWAxB,EAAO,IAAK,GAAMwB,EAAI,KAAM,EAAE,SAAS,OAAO,CAAC,EAAE,KAAK,GAAG,EACzDxB,EAAO,IAAK,GAAMwB,EAAI,KAAMC,EAAW,UAAU,EAAE,IAAI,CAAC,CAAC,EAAE,KAAK,GAAG,EACnEzB,EACG,IAAK,GAAMwB,EAAI,KAAMC,EAAW,UAAU,EAAE,UAAY,CAAC,CAAC,CAAC,CAAC,EAC5D,KAAK,GAAG,EACXzB,EAAO,IAAI,IAAM,KAAK,EAAE,KAAK,GAAG,EAChCA,EAAO,IAAK,GAAMwB,EAAI,KAAM,EAAE,IAAI,CAAC,EAAE,KAAK,GAAG,EAC7CF,EACAvB,EACAE,GAAS,uBAAyB,OAClCA,GAAS,WAAayB,CACxB,CACF,CACF","names":["rawSql","single","sql","uuid","appendEventsSQL","rawSql","streamsTable","eventsTable","appendToStream","pool","streamName","streamType","events","options","transaction","execute","appendResult","expectedStreamVersion","toExpectedVersion","eventsToAppend","e","i","uuid","appendEventsRaw","error","isOptimisticConcurrencyError","success","next_stream_position","last_global_position","transaction_id","expected","n","a","s","streamId","single","sql","p","defaultTag"]}
|
package/dist/chunk-UWD6GOZC.mjs
DELETED
|
@@ -1,264 +0,0 @@
|
|
|
1
|
-
import{b as _,c as t,e as a,f as e}from"./chunk-ABNBEUC6.mjs";import{rawSql as n}from"@event-driven-io/dumbo";var m=n(`CREATE TABLE IF NOT EXISTS ${a.name}(
|
|
2
|
-
stream_id TEXT NOT NULL,
|
|
3
|
-
stream_position BIGINT NOT NULL,
|
|
4
|
-
partition TEXT NOT NULL DEFAULT '${_}__${_}',
|
|
5
|
-
stream_type TEXT NOT NULL,
|
|
6
|
-
stream_metadata JSONB NOT NULL,
|
|
7
|
-
is_archived BOOLEAN NOT NULL DEFAULT FALSE,
|
|
8
|
-
PRIMARY KEY (stream_id, stream_position, partition, is_archived),
|
|
9
|
-
UNIQUE (stream_id, partition, is_archived)
|
|
10
|
-
) PARTITION BY LIST (partition);`),i=n(`
|
|
11
|
-
CREATE SEQUENCE IF NOT EXISTS emt_global_event_position;
|
|
12
|
-
|
|
13
|
-
CREATE TABLE IF NOT EXISTS ${e.name}(
|
|
14
|
-
stream_id TEXT NOT NULL,
|
|
15
|
-
stream_position BIGINT NOT NULL,
|
|
16
|
-
partition TEXT NOT NULL DEFAULT '${_}',
|
|
17
|
-
event_data JSONB NOT NULL,
|
|
18
|
-
event_metadata JSONB NOT NULL,
|
|
19
|
-
event_schema_version TEXT NOT NULL,
|
|
20
|
-
event_type TEXT NOT NULL,
|
|
21
|
-
event_id TEXT NOT NULL,
|
|
22
|
-
is_archived BOOLEAN NOT NULL DEFAULT FALSE,
|
|
23
|
-
global_position BIGINT DEFAULT nextval('emt_global_event_position'),
|
|
24
|
-
transaction_id XID8 NOT NULL,
|
|
25
|
-
created TIMESTAMPTZ NOT NULL DEFAULT now(),
|
|
26
|
-
PRIMARY KEY (stream_id, stream_position, partition, is_archived)
|
|
27
|
-
) PARTITION BY LIST (partition);`),I=n(`
|
|
28
|
-
CREATE TABLE IF NOT EXISTS emt_subscriptions(
|
|
29
|
-
subscription_id TEXT NOT NULL PRIMARY KEY,
|
|
30
|
-
version INT NOT NULL DEFAULT 1,
|
|
31
|
-
module TEXT NULL,
|
|
32
|
-
tenant TEXT NULL,
|
|
33
|
-
last_processed_position BIGINT NOT NULL,
|
|
34
|
-
last_processed_transaction_id BIGINT NOT NULL
|
|
35
|
-
);
|
|
36
|
-
`),o=n(`CREATE OR REPLACE FUNCTION emt_sanitize_name(input_name TEXT) RETURNS TEXT AS $$
|
|
37
|
-
BEGIN
|
|
38
|
-
RETURN REGEXP_REPLACE(input_name, '[^a-zA-Z0-9_]', '_', 'g');
|
|
39
|
-
END;
|
|
40
|
-
$$ LANGUAGE plpgsql;`),r=n(`
|
|
41
|
-
CREATE OR REPLACE FUNCTION emt_add_table_partition(tableName TEXT, partition_name TEXT) RETURNS void AS $$
|
|
42
|
-
DECLARE
|
|
43
|
-
v_main_partiton_name TEXT;
|
|
44
|
-
v_active_partiton_name TEXT;
|
|
45
|
-
v_archived_partiton_name TEXT;
|
|
46
|
-
BEGIN
|
|
47
|
-
v_main_partiton_name := emt_sanitize_name(tableName || '_' || partition_name);
|
|
48
|
-
v_active_partiton_name := emt_sanitize_name(v_main_partiton_name || '_active');
|
|
49
|
-
v_archived_partiton_name := emt_sanitize_name(v_main_partiton_name || '_archived');
|
|
50
|
-
|
|
51
|
-
-- create default events partition
|
|
52
|
-
EXECUTE format('
|
|
53
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
54
|
-
FOR VALUES IN (%L) PARTITION BY LIST (is_archived);',
|
|
55
|
-
v_main_partiton_name, tableName, partition_name
|
|
56
|
-
);
|
|
57
|
-
|
|
58
|
-
-- create default streams partition
|
|
59
|
-
EXECUTE format('
|
|
60
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
61
|
-
FOR VALUES IN (%L) PARTITION BY LIST (is_archived);',
|
|
62
|
-
v_main_partiton_name, tableName, partition_name
|
|
63
|
-
);
|
|
64
|
-
|
|
65
|
-
EXECUTE format('
|
|
66
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
67
|
-
FOR VALUES IN (FALSE);',
|
|
68
|
-
v_active_partiton_name, v_main_partiton_name
|
|
69
|
-
);
|
|
70
|
-
|
|
71
|
-
EXECUTE format('
|
|
72
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
73
|
-
FOR VALUES IN (TRUE);',
|
|
74
|
-
v_archived_partiton_name, v_main_partiton_name
|
|
75
|
-
);
|
|
76
|
-
END;
|
|
77
|
-
$$ LANGUAGE plpgsql;`),N=n(`
|
|
78
|
-
CREATE OR REPLACE FUNCTION emt_add_partition(partition_name TEXT) RETURNS void AS $$
|
|
79
|
-
BEGIN
|
|
80
|
-
PERFORM emt_add_table_partition('${e.name}', partition_name);
|
|
81
|
-
PERFORM emt_add_table_partition('${a.name}', partition_name);
|
|
82
|
-
END;
|
|
83
|
-
$$ LANGUAGE plpgsql;`),A=n(`
|
|
84
|
-
CREATE OR REPLACE FUNCTION add_module(new_module TEXT) RETURNS void AS $$
|
|
85
|
-
BEGIN
|
|
86
|
-
-- For ${e.name} table
|
|
87
|
-
EXECUTE format('
|
|
88
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
89
|
-
FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',
|
|
90
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || '${_}'), '${e.name}', new_module, '${_}'
|
|
91
|
-
);
|
|
92
|
-
|
|
93
|
-
EXECUTE format('
|
|
94
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
95
|
-
FOR VALUES IN (FALSE);',
|
|
96
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || '${_}' || '_active'), emt_sanitize_name('${e.name}_' || new_module || '__' || '${_}')
|
|
97
|
-
);
|
|
98
|
-
|
|
99
|
-
EXECUTE format('
|
|
100
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
101
|
-
FOR VALUES IN (TRUE);',
|
|
102
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || '${_}' || '_archived'), emt_sanitize_name('${e.name}_' || new_module || '__' || '${_}')
|
|
103
|
-
);
|
|
104
|
-
|
|
105
|
-
-- For ${a.name} table
|
|
106
|
-
EXECUTE format('
|
|
107
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
108
|
-
FOR VALUES IN (emt_sanitize_name(%L || ''__'' || %L)) PARTITION BY LIST (is_archived);',
|
|
109
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || '${_}'), '${a.name}', new_module, '${_}'
|
|
110
|
-
);
|
|
111
|
-
|
|
112
|
-
EXECUTE format('
|
|
113
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
114
|
-
FOR VALUES IN (FALSE);',
|
|
115
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || '${_}' || '_active'), emt_sanitize_name('${a.name}_' || new_module || '__' || '${_}')
|
|
116
|
-
);
|
|
117
|
-
|
|
118
|
-
EXECUTE format('
|
|
119
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
120
|
-
FOR VALUES IN (TRUE);',
|
|
121
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || '${_}' || '_archived'), emt_sanitize_name('${a.name}_' || new_module || '__' || '${_}')
|
|
122
|
-
);
|
|
123
|
-
END;
|
|
124
|
-
$$ LANGUAGE plpgsql;
|
|
125
|
-
`),O=n(`
|
|
126
|
-
CREATE OR REPLACE FUNCTION add_tenant(new_module TEXT, new_tenant TEXT) RETURNS void AS $$
|
|
127
|
-
BEGIN
|
|
128
|
-
-- For ${e.name} table
|
|
129
|
-
EXECUTE format('
|
|
130
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
131
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
132
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || new_tenant), '${e.name}', new_module, new_tenant
|
|
133
|
-
);
|
|
134
|
-
|
|
135
|
-
EXECUTE format('
|
|
136
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
137
|
-
FOR VALUES IN (FALSE);',
|
|
138
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${e.name}_' || new_module || '__' || new_tenant)
|
|
139
|
-
);
|
|
140
|
-
|
|
141
|
-
EXECUTE format('
|
|
142
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
143
|
-
FOR VALUES IN (TRUE);',
|
|
144
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${e.name}_' || new_module || '__' || new_tenant)
|
|
145
|
-
);
|
|
146
|
-
|
|
147
|
-
-- For ${a.name} table
|
|
148
|
-
EXECUTE format('
|
|
149
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
150
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
151
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || new_tenant), '${a.name}', new_module, new_tenant
|
|
152
|
-
);
|
|
153
|
-
|
|
154
|
-
EXECUTE format('
|
|
155
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
156
|
-
FOR VALUES IN (FALSE);',
|
|
157
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || new_tenant || '_active'), emt_sanitize_name('${a.name}_' || new_module || '__' || new_tenant)
|
|
158
|
-
);
|
|
159
|
-
|
|
160
|
-
EXECUTE format('
|
|
161
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
162
|
-
FOR VALUES IN (TRUE);',
|
|
163
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || new_tenant || '_archived'), emt_sanitize_name('${a.name}_' || new_module || '__' || new_tenant)
|
|
164
|
-
);
|
|
165
|
-
END;
|
|
166
|
-
$$ LANGUAGE plpgsql;
|
|
167
|
-
`),d=n(`
|
|
168
|
-
CREATE OR REPLACE FUNCTION add_module_for_all_tenants(new_module TEXT) RETURNS void AS $$
|
|
169
|
-
DECLARE
|
|
170
|
-
tenant_record RECORD;
|
|
171
|
-
BEGIN
|
|
172
|
-
PERFORM add_module(new_module);
|
|
173
|
-
|
|
174
|
-
FOR tenant_record IN SELECT DISTINCT tenant FROM ${e.name}
|
|
175
|
-
LOOP
|
|
176
|
-
-- For ${e.name} table
|
|
177
|
-
EXECUTE format('
|
|
178
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
179
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
180
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || tenant_record.tenant), '${e.name}', new_module, tenant_record.tenant
|
|
181
|
-
);
|
|
182
|
-
|
|
183
|
-
EXECUTE format('
|
|
184
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
185
|
-
FOR VALUES IN (FALSE);',
|
|
186
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${e.name}_' || new_module || '__' || tenant_record.tenant)
|
|
187
|
-
);
|
|
188
|
-
|
|
189
|
-
EXECUTE format('
|
|
190
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
191
|
-
FOR VALUES IN (TRUE);',
|
|
192
|
-
emt_sanitize_name('${e.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${e.name}_' || new_module || '__' || tenant_record.tenant)
|
|
193
|
-
);
|
|
194
|
-
|
|
195
|
-
-- For ${a.name} table
|
|
196
|
-
EXECUTE format('
|
|
197
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
198
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
199
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || tenant_record.tenant), '${a.name}', new_module, tenant_record.tenant
|
|
200
|
-
);
|
|
201
|
-
|
|
202
|
-
EXECUTE format('
|
|
203
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
204
|
-
FOR VALUES IN (FALSE);',
|
|
205
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || tenant_record.tenant || '_active'), emt_sanitize_name('${a.name}_' || new_module || '__' || tenant_record.tenant)
|
|
206
|
-
);
|
|
207
|
-
|
|
208
|
-
EXECUTE format('
|
|
209
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
210
|
-
FOR VALUES IN (TRUE);',
|
|
211
|
-
emt_sanitize_name('${a.name}_' || new_module || '__' || tenant_record.tenant || '_archived'), emt_sanitize_name('${a.name}_' || new_module || '__' || tenant_record.tenant)
|
|
212
|
-
);
|
|
213
|
-
END LOOP;
|
|
214
|
-
END;
|
|
215
|
-
$$ LANGUAGE plpgsql;
|
|
216
|
-
`),L=n(`
|
|
217
|
-
CREATE OR REPLACE FUNCTION add_tenant_for_all_modules(new_tenant TEXT) RETURNS void AS $$
|
|
218
|
-
DECLARE
|
|
219
|
-
module_record RECORD;
|
|
220
|
-
BEGIN
|
|
221
|
-
FOR module_record IN SELECT DISTINCT partitionname FROM pg_partman.part_config WHERE parent_table = '${e.name}'
|
|
222
|
-
LOOP
|
|
223
|
-
-- For ${e.name} table
|
|
224
|
-
EXECUTE format('
|
|
225
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
226
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
227
|
-
emt_sanitize_name('${e.name}_' || module_record.partitionname || '__' || new_tenant), '${e.name}', module_record.partitionname, new_tenant
|
|
228
|
-
);
|
|
229
|
-
|
|
230
|
-
EXECUTE format('
|
|
231
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
232
|
-
FOR VALUES IN (FALSE);',
|
|
233
|
-
emt_sanitize_name('${e.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${e.name}_' || module_record.partitionname || '__' || new_tenant)
|
|
234
|
-
);
|
|
235
|
-
|
|
236
|
-
EXECUTE format('
|
|
237
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
238
|
-
FOR VALUES IN (TRUE);',
|
|
239
|
-
emt_sanitize_name('${e.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${e.name}_' || module_record.partitionname || '__' || new_tenant)
|
|
240
|
-
);
|
|
241
|
-
|
|
242
|
-
-- For ${a.name} table
|
|
243
|
-
EXECUTE format('
|
|
244
|
-
CREATE TABLE IF NOT EXISTS %I PARTITION OF %I
|
|
245
|
-
FOR VALUES IN (emt_sanitize_name(''%s__%s'')) PARTITION BY LIST (is_archived);',
|
|
246
|
-
emt_sanitize_name('${a.name}_' || module_record.partitionname || '__' || new_tenant), '${a.name}', module_record.partitionname, new_tenant
|
|
247
|
-
);
|
|
248
|
-
|
|
249
|
-
EXECUTE format('
|
|
250
|
-
CREATE TABLE IF NOT EXISTS %I_active PARTITION OF %I
|
|
251
|
-
FOR VALUES IN (FALSE);',
|
|
252
|
-
emt_sanitize_name('${a.name}_' || module_record.partitionname || '__' || new_tenant || '_active'), emt_sanitize_name('${a.name}_' || module_record.partitionname || '__' || new_tenant)
|
|
253
|
-
);
|
|
254
|
-
|
|
255
|
-
EXECUTE format('
|
|
256
|
-
CREATE TABLE IF NOT EXISTS %I_archived PARTITION OF %I
|
|
257
|
-
FOR VALUES IN (TRUE);',
|
|
258
|
-
emt_sanitize_name('${a.name}_' || module_record.partitionname || '__' || new_tenant || '_archived'), emt_sanitize_name('${a.name}_' || module_record.partitionname || '__' || new_tenant)
|
|
259
|
-
);
|
|
260
|
-
END LOOP;
|
|
261
|
-
END;
|
|
262
|
-
$$ LANGUAGE plpgsql;
|
|
263
|
-
`),R=n(`SELECT emt_add_partition('${t}');`);export{m as a,i as b,I as c,o as d,r as e,N as f,A as g,O as h,d as i,L as j,R as k};
|
|
264
|
-
//# sourceMappingURL=chunk-UWD6GOZC.mjs.map
|