@ddd-ts/event-sourcing-firestore 0.0.0-compute-timeout-on-process.6 → 0.0.0-compute-timeout-on-process.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/_virtual/_rolldown/runtime.js +29 -0
- package/dist/firestore.event-lake.aggregate-store.js +12 -11
- package/dist/firestore.event-lake.aggregate-store.mjs +38 -0
- package/dist/firestore.event-lake.aggregate-store.mjs.map +1 -0
- package/dist/firestore.event-lake.storage-layer.d.ts +1 -1
- package/dist/firestore.event-lake.storage-layer.js +8 -7
- package/dist/firestore.event-lake.storage-layer.mjs +66 -0
- package/dist/firestore.event-lake.storage-layer.mjs.map +1 -0
- package/dist/firestore.event-lake.store.js +7 -7
- package/dist/firestore.event-lake.store.mjs +14 -0
- package/dist/firestore.event-lake.store.mjs.map +1 -0
- package/dist/firestore.event-stream.aggregate-store.js +15 -14
- package/dist/firestore.event-stream.aggregate-store.mjs +38 -0
- package/dist/firestore.event-stream.aggregate-store.mjs.map +1 -0
- package/dist/firestore.event-stream.storage-layer.d.ts +1 -1
- package/dist/firestore.event-stream.storage-layer.js +9 -8
- package/dist/firestore.event-stream.storage-layer.mjs +67 -0
- package/dist/firestore.event-stream.storage-layer.mjs.map +1 -0
- package/dist/firestore.event-stream.store.js +7 -7
- package/dist/firestore.event-stream.store.mjs +14 -0
- package/dist/firestore.event-stream.store.mjs.map +1 -0
- package/dist/firestore.projected-stream.reader.js +7 -7
- package/dist/firestore.projected-stream.reader.mjs +35 -0
- package/dist/firestore.projected-stream.reader.mjs.map +1 -0
- package/dist/firestore.projected-stream.storage-layer.js +20 -18
- package/dist/firestore.projected-stream.storage-layer.mjs +119 -0
- package/dist/firestore.projected-stream.storage-layer.mjs.map +1 -0
- package/dist/firestore.snapshotter.js +5 -5
- package/dist/firestore.snapshotter.mjs +36 -0
- package/dist/firestore.snapshotter.mjs.map +1 -0
- package/dist/index.js +29 -11
- package/dist/index.mjs +12 -0
- package/dist/projection/firestore.projector.js +48 -44
- package/dist/projection/firestore.projector.mjs +521 -0
- package/dist/projection/firestore.projector.mjs.map +1 -0
- package/package.json +16 -11
- package/dist/firestore.event-lake.aggregate-store.js.map +0 -1
- package/dist/firestore.event-lake.storage-layer.js.map +0 -1
- package/dist/firestore.event-lake.store.js.map +0 -1
- package/dist/firestore.event-stream.aggregate-store.js.map +0 -1
- package/dist/firestore.event-stream.storage-layer.js.map +0 -1
- package/dist/firestore.event-stream.store.js.map +0 -1
- package/dist/firestore.projected-stream.reader.js.map +0 -1
- package/dist/firestore.projected-stream.storage-layer.js.map +0 -1
- package/dist/firestore.snapshotter.js.map +0 -1
- package/dist/projection/firestore.projector.js.map +0 -1
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
//#region \0rolldown/runtime.js
|
|
2
|
+
var __create = Object.create;
|
|
3
|
+
var __defProp = Object.defineProperty;
|
|
4
|
+
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
5
|
+
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
6
|
+
var __getProtoOf = Object.getPrototypeOf;
|
|
7
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
8
|
+
var __copyProps = (to, from, except, desc) => {
|
|
9
|
+
if (from && typeof from === "object" || typeof from === "function") {
|
|
10
|
+
for (var keys = __getOwnPropNames(from), i = 0, n = keys.length, key; i < n; i++) {
|
|
11
|
+
key = keys[i];
|
|
12
|
+
if (!__hasOwnProp.call(to, key) && key !== except) {
|
|
13
|
+
__defProp(to, key, {
|
|
14
|
+
get: ((k) => from[k]).bind(null, key),
|
|
15
|
+
enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable
|
|
16
|
+
});
|
|
17
|
+
}
|
|
18
|
+
}
|
|
19
|
+
}
|
|
20
|
+
return to;
|
|
21
|
+
};
|
|
22
|
+
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", {
|
|
23
|
+
value: mod,
|
|
24
|
+
enumerable: true
|
|
25
|
+
}) : target, mod));
|
|
26
|
+
|
|
27
|
+
//#endregion
|
|
28
|
+
|
|
29
|
+
exports.__toESM = __toESM;
|
|
@@ -1,8 +1,9 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
const require_firestore_event_lake_storage_layer = require('./firestore.event-lake.storage-layer.js');
|
|
3
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
4
|
+
let _ddd_ts_store_firestore = require("@ddd-ts/store-firestore");
|
|
5
|
+
let firebase_admin_firestore = require("firebase-admin/firestore");
|
|
6
|
+
require("@ddd-ts/traits");
|
|
6
7
|
|
|
7
8
|
//#region src/firestore.event-lake.aggregate-store.ts
|
|
8
9
|
const MakeFirestoreEventLakeAggregateStore = (AGGREGATE) => {
|
|
@@ -13,14 +14,14 @@ const MakeFirestoreEventLakeAggregateStore = (AGGREGATE) => {
|
|
|
13
14
|
}
|
|
14
15
|
return $FirestoreEventLakeAggregateStore;
|
|
15
16
|
};
|
|
16
|
-
var FirestoreEventLakeAggregateStore = class extends FirestoreStore {
|
|
17
|
+
var FirestoreEventLakeAggregateStore = class extends _ddd_ts_store_firestore.FirestoreStore {
|
|
17
18
|
transaction;
|
|
18
19
|
lakeStore;
|
|
19
20
|
constructor(collection, serializer, eventBus, $name) {
|
|
20
21
|
super(collection, serializer, $name);
|
|
21
|
-
const storageLayer = new FirestoreEventLakeStorageLayer(collection.firestore);
|
|
22
|
-
this.transaction = new FirestoreTransactionPerformer(collection.firestore);
|
|
23
|
-
this.lakeStore = new EventLakeStore(storageLayer, serializer, eventBus);
|
|
22
|
+
const storageLayer = new require_firestore_event_lake_storage_layer.FirestoreEventLakeStorageLayer(collection.firestore);
|
|
23
|
+
this.transaction = new _ddd_ts_store_firestore.FirestoreTransactionPerformer(collection.firestore);
|
|
24
|
+
this.lakeStore = new _ddd_ts_core.EventLakeStore(storageLayer, serializer, eventBus);
|
|
24
25
|
}
|
|
25
26
|
async save(aggregate, trx) {
|
|
26
27
|
const changes = [...aggregate.changes];
|
|
@@ -34,5 +35,5 @@ var FirestoreEventLakeAggregateStore = class extends FirestoreStore {
|
|
|
34
35
|
};
|
|
35
36
|
|
|
36
37
|
//#endregion
|
|
37
|
-
|
|
38
|
-
|
|
38
|
+
exports.FirestoreEventLakeAggregateStore = FirestoreEventLakeAggregateStore;
|
|
39
|
+
exports.MakeFirestoreEventLakeAggregateStore = MakeFirestoreEventLakeAggregateStore;
|
|
@@ -0,0 +1,38 @@
|
|
|
1
|
+
import { FirestoreEventLakeStorageLayer } from "./firestore.event-lake.storage-layer.mjs";
|
|
2
|
+
import { EventLakeStore, EventOf, EventSourced, IEventBus, IEventSourced, IIdentifiable, ISerializer, Identifiable, LakeId } from "@ddd-ts/core";
|
|
3
|
+
import { FirestoreStore, FirestoreTransaction, FirestoreTransactionPerformer } from "@ddd-ts/store-firestore";
|
|
4
|
+
import { CollectionReference } from "firebase-admin/firestore";
|
|
5
|
+
import { HasTrait } from "@ddd-ts/traits";
|
|
6
|
+
|
|
7
|
+
//#region src/firestore.event-lake.aggregate-store.ts
|
|
8
|
+
const MakeFirestoreEventLakeAggregateStore = (AGGREGATE) => {
|
|
9
|
+
class $FirestoreEventLakeAggregateStore extends FirestoreEventLakeAggregateStore {
|
|
10
|
+
constructor(collection, serializer, eventBus) {
|
|
11
|
+
super(collection, serializer, eventBus, AGGREGATE.name);
|
|
12
|
+
}
|
|
13
|
+
}
|
|
14
|
+
return $FirestoreEventLakeAggregateStore;
|
|
15
|
+
};
|
|
16
|
+
var FirestoreEventLakeAggregateStore = class extends FirestoreStore {
|
|
17
|
+
transaction;
|
|
18
|
+
lakeStore;
|
|
19
|
+
constructor(collection, serializer, eventBus, $name) {
|
|
20
|
+
super(collection, serializer, $name);
|
|
21
|
+
const storageLayer = new FirestoreEventLakeStorageLayer(collection.firestore);
|
|
22
|
+
this.transaction = new FirestoreTransactionPerformer(collection.firestore);
|
|
23
|
+
this.lakeStore = new EventLakeStore(storageLayer, serializer, eventBus);
|
|
24
|
+
}
|
|
25
|
+
async save(aggregate, trx) {
|
|
26
|
+
const changes = [...aggregate.changes];
|
|
27
|
+
await this.transaction.performWith(trx, async (trx) => {
|
|
28
|
+
const lakeId = this.getLakeId(aggregate);
|
|
29
|
+
await super.save(aggregate, trx);
|
|
30
|
+
await this.lakeStore.append(lakeId, changes, trx);
|
|
31
|
+
aggregate.clearChanges();
|
|
32
|
+
});
|
|
33
|
+
}
|
|
34
|
+
};
|
|
35
|
+
|
|
36
|
+
//#endregion
|
|
37
|
+
export { FirestoreEventLakeAggregateStore, MakeFirestoreEventLakeAggregateStore };
|
|
38
|
+
//# sourceMappingURL=firestore.event-lake.aggregate-store.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-lake.aggregate-store.mjs","names":[],"sources":["../src/firestore.event-lake.aggregate-store.ts"],"sourcesContent":["import {\n ISerializer,\n IEventSourced,\n IIdentifiable,\n EventOf,\n EventLakeStore,\n LakeId,\n IChange,\n IEventBus,\n EventSourced,\n Identifiable,\n} from \"@ddd-ts/core\";\nimport {\n FirestoreStore,\n FirestoreTransaction,\n FirestoreTransactionPerformer,\n} from \"@ddd-ts/store-firestore\";\n\nimport { CollectionReference } from \"firebase-admin/firestore\";\nimport { FirestoreEventLakeStorageLayer } from \"./firestore.event-lake.storage-layer\";\nimport { HasTrait } from \"@ddd-ts/traits\";\n\nexport const MakeFirestoreEventLakeAggregateStore = <\n A extends HasTrait<typeof EventSourced> & HasTrait<typeof Identifiable>,\n>(\n AGGREGATE: A,\n) => {\n abstract class $FirestoreEventLakeAggregateStore extends FirestoreEventLakeAggregateStore<\n InstanceType<A>\n > {\n constructor(\n collection: CollectionReference,\n serializer: ISerializer<InstanceType<A>> &\n ISerializer<EventOf<InstanceType<A>>>,\n eventBus?: IEventBus,\n ) {\n super(collection, serializer, eventBus, AGGREGATE.name);\n }\n\n abstract getLakeId(instance: InstanceType<A>): LakeId;\n }\n\n return $FirestoreEventLakeAggregateStore;\n};\n\nexport abstract class FirestoreEventLakeAggregateStore<\n A extends IEventSourced & IIdentifiable,\n> extends FirestoreStore<A> {\n transaction: FirestoreTransactionPerformer;\n lakeStore: EventLakeStore<EventOf<A>>;\n constructor(\n collection: CollectionReference,\n serializer: ISerializer<EventOf<A>> & ISerializer<A>,\n eventBus?: IEventBus,\n $name?: string,\n ) {\n super(collection, serializer, $name);\n const storageLayer = new FirestoreEventLakeStorageLayer(\n collection.firestore,\n );\n this.transaction = new FirestoreTransactionPerformer(collection.firestore);\n this.lakeStore = new EventLakeStore<EventOf<A>>(\n storageLayer,\n serializer,\n eventBus,\n );\n }\n\n abstract getLakeId(instance: A): LakeId;\n\n override async save(aggregate: A, trx?: FirestoreTransaction) {\n const changes = [...aggregate.changes] as IChange<EventOf<A>>[];\n\n await this.transaction.performWith(trx, async (trx) => {\n const lakeId = this.getLakeId(aggregate);\n\n await super.save(aggregate, trx);\n await this.lakeStore.append(lakeId, changes, trx);\n aggregate.clearChanges();\n });\n }\n}\n"],"mappings":";;;;;;;AAsBA,MAAa,wCAGX,cACG;CACH,MAAe,0CAA0C,iCAEvD;EACA,YACE,YACA,YAEA,UACA;AACA,SAAM,YAAY,YAAY,UAAU,UAAU,KAAK;;;AAM3D,QAAO;;AAGT,IAAsB,mCAAtB,cAEU,eAAkB;CAC1B;CACA;CACA,YACE,YACA,YACA,UACA,OACA;AACA,QAAM,YAAY,YAAY,MAAM;EACpC,MAAM,eAAe,IAAI,+BACvB,WAAW,UACZ;AACD,OAAK,cAAc,IAAI,8BAA8B,WAAW,UAAU;AAC1E,OAAK,YAAY,IAAI,eACnB,cACA,YACA,SACD;;CAKH,MAAe,KAAK,WAAc,KAA4B;EAC5D,MAAM,UAAU,CAAC,GAAG,UAAU,QAAQ;AAEtC,QAAM,KAAK,YAAY,YAAY,KAAK,OAAO,QAAQ;GACrD,MAAM,SAAS,KAAK,UAAU,UAAU;AAExC,SAAM,MAAM,KAAK,WAAW,IAAI;AAChC,SAAM,KAAK,UAAU,OAAO,QAAQ,SAAS,IAAI;AACjD,aAAU,cAAc;IACxB"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { EventId, EventLakeStorageLayer, ISerializedChange, ISerializedFact, LakeId } from "@ddd-ts/core";
|
|
2
|
+
import { ISerializedSavedChange as ISerializedSavedChange$1 } from "@ddd-ts/core/dist/interfaces/es-event";
|
|
2
3
|
import { DefaultConverter, FirestoreTransaction } from "@ddd-ts/store-firestore";
|
|
3
4
|
import * as fb from "firebase-admin";
|
|
4
|
-
import { ISerializedSavedChange as ISerializedSavedChange$1 } from "@ddd-ts/core/dist/interfaces/es-event";
|
|
5
5
|
|
|
6
6
|
//#region src/firestore.event-lake.storage-layer.d.ts
|
|
7
7
|
declare class FirestoreEventLakeStorageLayer implements EventLakeStorageLayer {
|
|
@@ -1,11 +1,13 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
3
|
+
let _ddd_ts_store_firestore = require("@ddd-ts/store-firestore");
|
|
4
|
+
let firebase_admin = require("firebase-admin");
|
|
5
|
+
firebase_admin = require_runtime.__toESM(firebase_admin);
|
|
4
6
|
|
|
5
7
|
//#region src/firestore.event-lake.storage-layer.ts
|
|
6
|
-
const serverTimestamp =
|
|
8
|
+
const serverTimestamp = firebase_admin.firestore.FieldValue.serverTimestamp;
|
|
7
9
|
var FirestoreEventLakeStorageLayer = class {
|
|
8
|
-
constructor(firestore, converter = new DefaultConverter()) {
|
|
10
|
+
constructor(firestore, converter = new _ddd_ts_store_firestore.DefaultConverter()) {
|
|
9
11
|
this.firestore = firestore;
|
|
10
12
|
this.converter = converter;
|
|
11
13
|
}
|
|
@@ -62,5 +64,4 @@ var FirestoreEventLakeStorageLayer = class {
|
|
|
62
64
|
};
|
|
63
65
|
|
|
64
66
|
//#endregion
|
|
65
|
-
|
|
66
|
-
//# sourceMappingURL=firestore.event-lake.storage-layer.js.map
|
|
67
|
+
exports.FirestoreEventLakeStorageLayer = FirestoreEventLakeStorageLayer;
|
|
@@ -0,0 +1,66 @@
|
|
|
1
|
+
import { EventId, LakeId } from "@ddd-ts/core";
|
|
2
|
+
import { DefaultConverter, FirestoreTransaction } from "@ddd-ts/store-firestore";
|
|
3
|
+
import * as fb from "firebase-admin";
|
|
4
|
+
|
|
5
|
+
//#region src/firestore.event-lake.storage-layer.ts
|
|
6
|
+
const serverTimestamp = fb.firestore.FieldValue.serverTimestamp;
|
|
7
|
+
var FirestoreEventLakeStorageLayer = class {
|
|
8
|
+
constructor(firestore, converter = new DefaultConverter()) {
|
|
9
|
+
this.firestore = firestore;
|
|
10
|
+
this.converter = converter;
|
|
11
|
+
}
|
|
12
|
+
getCollection(lakeId) {
|
|
13
|
+
return this.firestore.collection("event-store").doc("Lakes").collection(lakeId.shardType).doc(lakeId.shardId).collection("events");
|
|
14
|
+
}
|
|
15
|
+
async append(lakeId, changes, trx) {
|
|
16
|
+
const collection = this.getCollection(lakeId);
|
|
17
|
+
const result = [];
|
|
18
|
+
for (const change of changes) {
|
|
19
|
+
const revision = trx.increment();
|
|
20
|
+
const storageChange = {
|
|
21
|
+
eventId: change.id,
|
|
22
|
+
name: change.name,
|
|
23
|
+
payload: change.payload,
|
|
24
|
+
occurredAt: serverTimestamp(),
|
|
25
|
+
version: change.version,
|
|
26
|
+
revision
|
|
27
|
+
};
|
|
28
|
+
const ref = collection.doc(change.id);
|
|
29
|
+
trx.transaction.create(ref, this.converter.toFirestore(storageChange));
|
|
30
|
+
result.push({
|
|
31
|
+
...change,
|
|
32
|
+
ref: ref.path,
|
|
33
|
+
revision,
|
|
34
|
+
occurredAt: void 0
|
|
35
|
+
});
|
|
36
|
+
}
|
|
37
|
+
return result;
|
|
38
|
+
}
|
|
39
|
+
async *read(lakeId, startAfter, endAt) {
|
|
40
|
+
const collection = this.getCollection(lakeId);
|
|
41
|
+
const [start, end] = await Promise.all([startAfter ? collection.doc(startAfter.serialize()).get() : null, endAt ? collection.doc(endAt.serialize()).get() : null]);
|
|
42
|
+
if (startAfter && !start?.exists) throw new Error(`StartAfter event not found: ${startAfter}`);
|
|
43
|
+
if (endAt && !end?.exists) throw new Error(`EndAt event not found: ${endAt}`);
|
|
44
|
+
let query = collection.orderBy("occurredAt", "asc").orderBy("revision", "asc");
|
|
45
|
+
if (start) query = query.startAfter(start);
|
|
46
|
+
if (endAt) query = query.endAt(end);
|
|
47
|
+
for await (const event of query.stream()) {
|
|
48
|
+
const e = event;
|
|
49
|
+
const data = this.converter.fromFirestore(e);
|
|
50
|
+
yield {
|
|
51
|
+
id: data.eventId,
|
|
52
|
+
ref: e.ref.path,
|
|
53
|
+
revision: data.revision,
|
|
54
|
+
name: data.name,
|
|
55
|
+
$name: data.name,
|
|
56
|
+
payload: data.payload,
|
|
57
|
+
occurredAt: data.occurredAt,
|
|
58
|
+
version: data.version ?? 1
|
|
59
|
+
};
|
|
60
|
+
}
|
|
61
|
+
}
|
|
62
|
+
};
|
|
63
|
+
|
|
64
|
+
//#endregion
|
|
65
|
+
export { FirestoreEventLakeStorageLayer };
|
|
66
|
+
//# sourceMappingURL=firestore.event-lake.storage-layer.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-lake.storage-layer.mjs","names":[],"sources":["../src/firestore.event-lake.storage-layer.ts"],"sourcesContent":["import {\n LakeId,\n EventId,\n type ISerializedChange,\n type ISerializedFact,\n type EventLakeStorageLayer,\n} from \"@ddd-ts/core\";\nimport type { ISerializedSavedChange } from \"@ddd-ts/core/dist/interfaces/es-event\";\n\nimport {\n DefaultConverter,\n FirestoreTransaction,\n} from \"@ddd-ts/store-firestore\";\nimport * as fb from \"firebase-admin\";\n\nexport const serverTimestamp = fb.firestore.FieldValue.serverTimestamp;\n\nexport class FirestoreEventLakeStorageLayer implements EventLakeStorageLayer {\n constructor(\n public readonly firestore: fb.firestore.Firestore,\n public readonly converter = new DefaultConverter(),\n ) {}\n\n getCollection(lakeId: LakeId) {\n return this.firestore\n .collection(\"event-store\")\n .doc(\"Lakes\")\n .collection(lakeId.shardType)\n .doc(lakeId.shardId)\n .collection(\"events\");\n }\n\n async append(\n lakeId: LakeId,\n changes: ISerializedChange[],\n trx: FirestoreTransaction,\n ) {\n const collection = this.getCollection(lakeId);\n\n const result: ISerializedSavedChange[] = [];\n\n for (const change of changes) {\n const revision = trx.increment();\n const storageChange = {\n eventId: change.id,\n name: change.name,\n payload: change.payload,\n occurredAt: serverTimestamp(),\n version: change.version,\n revision,\n };\n\n const ref = collection.doc(change.id);\n trx.transaction.create(ref, this.converter.toFirestore(storageChange));\n\n result.push({\n ...change,\n ref: ref.path,\n revision,\n occurredAt: undefined,\n });\n }\n\n return result;\n }\n\n async *read(\n lakeId: LakeId,\n startAfter?: EventId,\n endAt?: EventId,\n ): AsyncIterable<ISerializedFact> {\n const collection = this.getCollection(lakeId);\n\n const [start, end] = await Promise.all([\n startAfter ? collection.doc(startAfter.serialize()).get() : null,\n endAt ? collection.doc(endAt.serialize()).get() : null,\n ]);\n\n if (startAfter && !start?.exists) {\n throw new Error(`StartAfter event not found: ${startAfter}`);\n }\n\n if (endAt && !end?.exists) {\n throw new Error(`EndAt event not found: ${endAt}`);\n }\n\n let query = collection\n .orderBy(\"occurredAt\", \"asc\")\n .orderBy(\"revision\", \"asc\");\n\n if (start) {\n query = query.startAfter(start);\n }\n\n if (endAt) {\n query = query.endAt(end);\n }\n\n for await (const event of query.stream()) {\n const e = event as any as fb.firestore.QueryDocumentSnapshot<any>;\n const data = this.converter.fromFirestore(e);\n yield {\n id: data.eventId,\n ref: e.ref.path,\n revision: data.revision,\n name: data.name,\n $name: data.name,\n payload: data.payload,\n occurredAt: data.occurredAt,\n version: data.version ?? 1,\n } as ISerializedFact;\n }\n }\n}\n"],"mappings":";;;;;AAeA,MAAa,kBAAkB,GAAG,UAAU,WAAW;AAEvD,IAAa,iCAAb,MAA6E;CAC3E,YACE,AAAgB,WAChB,AAAgB,YAAY,IAAI,kBAAkB,EAClD;EAFgB;EACA;;CAGlB,cAAc,QAAgB;AAC5B,SAAO,KAAK,UACT,WAAW,cAAc,CACzB,IAAI,QAAQ,CACZ,WAAW,OAAO,UAAU,CAC5B,IAAI,OAAO,QAAQ,CACnB,WAAW,SAAS;;CAGzB,MAAM,OACJ,QACA,SACA,KACA;EACA,MAAM,aAAa,KAAK,cAAc,OAAO;EAE7C,MAAM,SAAmC,EAAE;AAE3C,OAAK,MAAM,UAAU,SAAS;GAC5B,MAAM,WAAW,IAAI,WAAW;GAChC,MAAM,gBAAgB;IACpB,SAAS,OAAO;IAChB,MAAM,OAAO;IACb,SAAS,OAAO;IAChB,YAAY,iBAAiB;IAC7B,SAAS,OAAO;IAChB;IACD;GAED,MAAM,MAAM,WAAW,IAAI,OAAO,GAAG;AACrC,OAAI,YAAY,OAAO,KAAK,KAAK,UAAU,YAAY,cAAc,CAAC;AAEtE,UAAO,KAAK;IACV,GAAG;IACH,KAAK,IAAI;IACT;IACA,YAAY;IACb,CAAC;;AAGJ,SAAO;;CAGT,OAAO,KACL,QACA,YACA,OACgC;EAChC,MAAM,aAAa,KAAK,cAAc,OAAO;EAE7C,MAAM,CAAC,OAAO,OAAO,MAAM,QAAQ,IAAI,CACrC,aAAa,WAAW,IAAI,WAAW,WAAW,CAAC,CAAC,KAAK,GAAG,MAC5D,QAAQ,WAAW,IAAI,MAAM,WAAW,CAAC,CAAC,KAAK,GAAG,KACnD,CAAC;AAEF,MAAI,cAAc,CAAC,OAAO,OACxB,OAAM,IAAI,MAAM,+BAA+B,aAAa;AAG9D,MAAI,SAAS,CAAC,KAAK,OACjB,OAAM,IAAI,MAAM,0BAA0B,QAAQ;EAGpD,IAAI,QAAQ,WACT,QAAQ,cAAc,MAAM,CAC5B,QAAQ,YAAY,MAAM;AAE7B,MAAI,MACF,SAAQ,MAAM,WAAW,MAAM;AAGjC,MAAI,MACF,SAAQ,MAAM,MAAM,IAAI;AAG1B,aAAW,MAAM,SAAS,MAAM,QAAQ,EAAE;GACxC,MAAM,IAAI;GACV,MAAM,OAAO,KAAK,UAAU,cAAc,EAAE;AAC5C,SAAM;IACJ,IAAI,KAAK;IACT,KAAK,EAAE,IAAI;IACX,UAAU,KAAK;IACf,MAAM,KAAK;IACX,OAAO,KAAK;IACZ,SAAS,KAAK;IACd,YAAY,KAAK;IACjB,SAAS,KAAK,WAAW;IAC1B"}
|
|
@@ -1,14 +1,14 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
const require_firestore_event_lake_storage_layer = require('./firestore.event-lake.storage-layer.js');
|
|
3
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
4
|
+
let firebase_admin_firestore = require("firebase-admin/firestore");
|
|
4
5
|
|
|
5
6
|
//#region src/firestore.event-lake.store.ts
|
|
6
|
-
var FirestoreEventLakeStore = class extends EventLakeStore {
|
|
7
|
+
var FirestoreEventLakeStore = class extends _ddd_ts_core.EventLakeStore {
|
|
7
8
|
constructor(firestore, serializer, eventBus) {
|
|
8
|
-
super(new FirestoreEventLakeStorageLayer(firestore), serializer, eventBus);
|
|
9
|
+
super(new require_firestore_event_lake_storage_layer.FirestoreEventLakeStorageLayer(firestore), serializer, eventBus);
|
|
9
10
|
}
|
|
10
11
|
};
|
|
11
12
|
|
|
12
13
|
//#endregion
|
|
13
|
-
|
|
14
|
-
//# sourceMappingURL=firestore.event-lake.store.js.map
|
|
14
|
+
exports.FirestoreEventLakeStore = FirestoreEventLakeStore;
|
|
@@ -0,0 +1,14 @@
|
|
|
1
|
+
import { FirestoreEventLakeStorageLayer } from "./firestore.event-lake.storage-layer.mjs";
|
|
2
|
+
import { EventLakeStore, IEsEvent, IEventBus, ISerializer } from "@ddd-ts/core";
|
|
3
|
+
import { Firestore } from "firebase-admin/firestore";
|
|
4
|
+
|
|
5
|
+
//#region src/firestore.event-lake.store.ts
|
|
6
|
+
var FirestoreEventLakeStore = class extends EventLakeStore {
|
|
7
|
+
constructor(firestore, serializer, eventBus) {
|
|
8
|
+
super(new FirestoreEventLakeStorageLayer(firestore), serializer, eventBus);
|
|
9
|
+
}
|
|
10
|
+
};
|
|
11
|
+
|
|
12
|
+
//#endregion
|
|
13
|
+
export { FirestoreEventLakeStore };
|
|
14
|
+
//# sourceMappingURL=firestore.event-lake.store.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-lake.store.mjs","names":[],"sources":["../src/firestore.event-lake.store.ts"],"sourcesContent":["import { EventLakeStore, IEsEvent, IEventBus, ISerializer } from \"@ddd-ts/core\";\nimport { FirestoreEventLakeStorageLayer } from \"./firestore.event-lake.storage-layer\";\nimport { Firestore } from \"firebase-admin/firestore\";\n\nexport class FirestoreEventLakeStore<\n Event extends IEsEvent,\n> extends EventLakeStore<Event> {\n constructor(\n firestore: Firestore,\n serializer: ISerializer<Event>,\n eventBus?: IEventBus,\n ) {\n super(new FirestoreEventLakeStorageLayer(firestore), serializer, eventBus);\n }\n}\n"],"mappings":";;;;;AAIA,IAAa,0BAAb,cAEU,eAAsB;CAC9B,YACE,WACA,YACA,UACA;AACA,QAAM,IAAI,+BAA+B,UAAU,EAAE,YAAY,SAAS"}
|
|
@@ -1,30 +1,31 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
const require_firestore_event_stream_storage_layer = require('./firestore.event-stream.storage-layer.js');
|
|
3
|
+
const require_firestore_snapshotter = require('./firestore.snapshotter.js');
|
|
4
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
5
|
+
let _ddd_ts_store_firestore = require("@ddd-ts/store-firestore");
|
|
6
|
+
let firebase_admin_firestore = require("firebase-admin/firestore");
|
|
7
|
+
require("@ddd-ts/traits");
|
|
7
8
|
|
|
8
9
|
//#region src/firestore.event-stream.aggregate-store.ts
|
|
9
10
|
const MakeFirestoreEventStreamAggregateStore = (AGGREGATE) => {
|
|
10
11
|
return class $FirestoreEventStreamAggregateStore extends FirestoreEventStreamAggregateStore {
|
|
11
12
|
constructor(firestore, serializer, eventBus) {
|
|
12
|
-
const snapshotter = new FirestoreSnapshotter(AGGREGATE.name, firestore, serializer);
|
|
13
|
+
const snapshotter = new require_firestore_snapshotter.FirestoreSnapshotter(AGGREGATE.name, firestore, serializer);
|
|
13
14
|
super(firestore, serializer, snapshotter, eventBus);
|
|
14
15
|
}
|
|
15
16
|
loadFirst(event) {
|
|
16
17
|
return AGGREGATE.loadFirst(event);
|
|
17
18
|
}
|
|
18
19
|
getStreamId(id) {
|
|
19
|
-
return StreamId.from(AGGREGATE.name, id.serialize());
|
|
20
|
+
return _ddd_ts_core.StreamId.from(AGGREGATE.name, id.serialize());
|
|
20
21
|
}
|
|
21
22
|
};
|
|
22
23
|
};
|
|
23
|
-
var FirestoreEventStreamAggregateStore = class extends EventStreamAggregateStore {
|
|
24
|
+
var FirestoreEventStreamAggregateStore = class extends _ddd_ts_core.EventStreamAggregateStore {
|
|
24
25
|
constructor(firestore, serializer, snapshotter, eventBus) {
|
|
25
|
-
const storageLayer = new FirestoreEventStreamStorageLayer(firestore);
|
|
26
|
-
const transaction = new FirestoreTransactionPerformer(firestore);
|
|
27
|
-
const streamStore = new EventStreamStore(storageLayer, serializer, eventBus);
|
|
26
|
+
const storageLayer = new require_firestore_event_stream_storage_layer.FirestoreEventStreamStorageLayer(firestore);
|
|
27
|
+
const transaction = new _ddd_ts_store_firestore.FirestoreTransactionPerformer(firestore);
|
|
28
|
+
const streamStore = new _ddd_ts_core.EventStreamStore(storageLayer, serializer, eventBus);
|
|
28
29
|
super(streamStore, transaction, snapshotter);
|
|
29
30
|
this.firestore = firestore;
|
|
30
31
|
this.serializer = serializer;
|
|
@@ -34,5 +35,5 @@ var FirestoreEventStreamAggregateStore = class extends EventStreamAggregateStore
|
|
|
34
35
|
};
|
|
35
36
|
|
|
36
37
|
//#endregion
|
|
37
|
-
|
|
38
|
-
|
|
38
|
+
exports.FirestoreEventStreamAggregateStore = FirestoreEventStreamAggregateStore;
|
|
39
|
+
exports.MakeFirestoreEventStreamAggregateStore = MakeFirestoreEventStreamAggregateStore;
|
|
@@ -0,0 +1,38 @@
|
|
|
1
|
+
import { FirestoreEventStreamStorageLayer } from "./firestore.event-stream.storage-layer.mjs";
|
|
2
|
+
import { FirestoreSnapshotter } from "./firestore.snapshotter.mjs";
|
|
3
|
+
import { EventOf, EventSourced, EventStreamAggregateStore, EventStreamStore, EventsOf, IEventSourced, IIdentifiable, ISerializer, StreamId } from "@ddd-ts/core";
|
|
4
|
+
import { FirestoreTransactionPerformer } from "@ddd-ts/store-firestore";
|
|
5
|
+
import { Firestore } from "firebase-admin/firestore";
|
|
6
|
+
import { HasTrait } from "@ddd-ts/traits";
|
|
7
|
+
|
|
8
|
+
//#region src/firestore.event-stream.aggregate-store.ts
|
|
9
|
+
const MakeFirestoreEventStreamAggregateStore = (AGGREGATE) => {
|
|
10
|
+
return class $FirestoreEventStreamAggregateStore extends FirestoreEventStreamAggregateStore {
|
|
11
|
+
constructor(firestore, serializer, eventBus) {
|
|
12
|
+
const snapshotter = new FirestoreSnapshotter(AGGREGATE.name, firestore, serializer);
|
|
13
|
+
super(firestore, serializer, snapshotter, eventBus);
|
|
14
|
+
}
|
|
15
|
+
loadFirst(event) {
|
|
16
|
+
return AGGREGATE.loadFirst(event);
|
|
17
|
+
}
|
|
18
|
+
getStreamId(id) {
|
|
19
|
+
return StreamId.from(AGGREGATE.name, id.serialize());
|
|
20
|
+
}
|
|
21
|
+
};
|
|
22
|
+
};
|
|
23
|
+
var FirestoreEventStreamAggregateStore = class extends EventStreamAggregateStore {
|
|
24
|
+
constructor(firestore, serializer, snapshotter, eventBus) {
|
|
25
|
+
const storageLayer = new FirestoreEventStreamStorageLayer(firestore);
|
|
26
|
+
const transaction = new FirestoreTransactionPerformer(firestore);
|
|
27
|
+
const streamStore = new EventStreamStore(storageLayer, serializer, eventBus);
|
|
28
|
+
super(streamStore, transaction, snapshotter);
|
|
29
|
+
this.firestore = firestore;
|
|
30
|
+
this.serializer = serializer;
|
|
31
|
+
this.snapshotter = snapshotter;
|
|
32
|
+
this.eventBus = eventBus;
|
|
33
|
+
}
|
|
34
|
+
};
|
|
35
|
+
|
|
36
|
+
//#endregion
|
|
37
|
+
export { FirestoreEventStreamAggregateStore, MakeFirestoreEventStreamAggregateStore };
|
|
38
|
+
//# sourceMappingURL=firestore.event-stream.aggregate-store.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-stream.aggregate-store.mjs","names":[],"sources":["../src/firestore.event-stream.aggregate-store.ts"],"sourcesContent":["import { HasTrait } from \"@ddd-ts/traits\";\nimport {\n StreamId,\n EventsOf,\n EventSourced,\n type Identifiable,\n type IEventBus,\n EventStreamStore,\n EventStreamAggregateStore,\n ISerializer,\n IEventSourced,\n IIdentifiable,\n EventOf,\n} from \"@ddd-ts/core\";\nimport { FirestoreTransactionPerformer } from \"@ddd-ts/store-firestore\";\n\nimport { FirestoreSnapshotter } from \"./firestore.snapshotter\";\nimport { FirestoreEventStreamStorageLayer } from \"./firestore.event-stream.storage-layer\";\nimport { Firestore } from \"firebase-admin/firestore\";\n\nexport const MakeFirestoreEventStreamAggregateStore = <\n A extends HasTrait<typeof EventSourced> & HasTrait<typeof Identifiable>,\n>(\n AGGREGATE: A,\n) => {\n return class $FirestoreEventStreamAggregateStore extends FirestoreEventStreamAggregateStore<\n InstanceType<A>\n > {\n constructor(\n firestore: Firestore,\n serializer: ISerializer<InstanceType<A>> &\n ISerializer<EventOf<InstanceType<A>>>,\n eventBus?: IEventBus,\n ) {\n const snapshotter = new FirestoreSnapshotter<InstanceType<A>>(\n AGGREGATE.name,\n firestore,\n serializer,\n );\n super(firestore, serializer, snapshotter, eventBus);\n }\n\n loadFirst(event: EventsOf<A>[number]): InstanceType<A> {\n return AGGREGATE.loadFirst(event);\n }\n\n getStreamId(id: InstanceType<A>[\"id\"]): StreamId {\n return StreamId.from(AGGREGATE.name, id.serialize());\n }\n };\n};\n\nexport abstract class FirestoreEventStreamAggregateStore<\n A extends IEventSourced & IIdentifiable,\n> extends EventStreamAggregateStore<A> {\n constructor(\n public readonly firestore: Firestore,\n public readonly serializer: ISerializer<EventOf<A>>,\n public readonly snapshotter: FirestoreSnapshotter<A>,\n public readonly eventBus?: IEventBus,\n ) {\n const storageLayer = new FirestoreEventStreamStorageLayer(firestore);\n const transaction = new FirestoreTransactionPerformer(firestore);\n const streamStore = new EventStreamStore<EventOf<A>>(\n storageLayer,\n serializer,\n eventBus,\n );\n super(streamStore, transaction, snapshotter);\n }\n}\n"],"mappings":";;;;;;;;AAoBA,MAAa,0CAGX,cACG;AACH,QAAO,MAAM,4CAA4C,mCAEvD;EACA,YACE,WACA,YAEA,UACA;GACA,MAAM,cAAc,IAAI,qBACtB,UAAU,MACV,WACA,WACD;AACD,SAAM,WAAW,YAAY,aAAa,SAAS;;EAGrD,UAAU,OAA6C;AACrD,UAAO,UAAU,UAAU,MAAM;;EAGnC,YAAY,IAAqC;AAC/C,UAAO,SAAS,KAAK,UAAU,MAAM,GAAG,WAAW,CAAC;;;;AAK1D,IAAsB,qCAAtB,cAEU,0BAA6B;CACrC,YACE,AAAgB,WAChB,AAAgB,YAChB,AAAgB,aAChB,AAAgB,UAChB;EACA,MAAM,eAAe,IAAI,iCAAiC,UAAU;EACpE,MAAM,cAAc,IAAI,8BAA8B,UAAU;EAChE,MAAM,cAAc,IAAI,iBACtB,cACA,YACA,SACD;AACD,QAAM,aAAa,aAAa,YAAY;EAZ5B;EACA;EACA;EACA"}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { EventStreamStorageLayer, ISerializedChange, ISerializedFact, StreamId } from "@ddd-ts/core";
|
|
2
|
+
import { ISerializedSavedChange as ISerializedSavedChange$1 } from "@ddd-ts/core/dist/interfaces/es-event";
|
|
2
3
|
import { DefaultConverter, FirestoreTransaction } from "@ddd-ts/store-firestore";
|
|
3
4
|
import * as fb from "firebase-admin";
|
|
4
|
-
import { ISerializedSavedChange as ISerializedSavedChange$1 } from "@ddd-ts/core/dist/interfaces/es-event";
|
|
5
5
|
|
|
6
6
|
//#region src/firestore.event-stream.storage-layer.d.ts
|
|
7
7
|
declare class FirestoreEventStreamStorageLayer implements EventStreamStorageLayer {
|
|
@@ -1,12 +1,14 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
3
|
+
let _ddd_ts_store_firestore = require("@ddd-ts/store-firestore");
|
|
4
|
+
let firebase_admin = require("firebase-admin");
|
|
5
|
+
firebase_admin = require_runtime.__toESM(firebase_admin);
|
|
6
|
+
require("@ddd-ts/core/dist/interfaces/es-event");
|
|
5
7
|
|
|
6
8
|
//#region src/firestore.event-stream.storage-layer.ts
|
|
7
|
-
const serverTimestamp =
|
|
9
|
+
const serverTimestamp = firebase_admin.firestore.FieldValue.serverTimestamp;
|
|
8
10
|
var FirestoreEventStreamStorageLayer = class {
|
|
9
|
-
constructor(firestore, converter = new DefaultConverter()) {
|
|
11
|
+
constructor(firestore, converter = new _ddd_ts_store_firestore.DefaultConverter()) {
|
|
10
12
|
this.firestore = firestore;
|
|
11
13
|
this.converter = converter;
|
|
12
14
|
}
|
|
@@ -63,5 +65,4 @@ var FirestoreEventStreamStorageLayer = class {
|
|
|
63
65
|
};
|
|
64
66
|
|
|
65
67
|
//#endregion
|
|
66
|
-
|
|
67
|
-
//# sourceMappingURL=firestore.event-stream.storage-layer.js.map
|
|
68
|
+
exports.FirestoreEventStreamStorageLayer = FirestoreEventStreamStorageLayer;
|
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
import { EventStreamStorageLayer, StreamId } from "@ddd-ts/core";
|
|
2
|
+
import { DefaultConverter, FirestoreTransaction } from "@ddd-ts/store-firestore";
|
|
3
|
+
import * as fb from "firebase-admin";
|
|
4
|
+
import { ISerializedSavedChange as ISerializedSavedChange$1 } from "@ddd-ts/core/dist/interfaces/es-event";
|
|
5
|
+
|
|
6
|
+
//#region src/firestore.event-stream.storage-layer.ts
|
|
7
|
+
const serverTimestamp = fb.firestore.FieldValue.serverTimestamp;
|
|
8
|
+
var FirestoreEventStreamStorageLayer = class {
|
|
9
|
+
constructor(firestore, converter = new DefaultConverter()) {
|
|
10
|
+
this.firestore = firestore;
|
|
11
|
+
this.converter = converter;
|
|
12
|
+
}
|
|
13
|
+
isLocalRevisionOutdatedError(error) {
|
|
14
|
+
return typeof error === "object" && error !== null && "code" in error && error.code === 6;
|
|
15
|
+
}
|
|
16
|
+
getCollection(streamId) {
|
|
17
|
+
return this.firestore.collection("event-store").doc(streamId.aggregateType).collection("streams").doc(streamId.aggregateId).collection("events");
|
|
18
|
+
}
|
|
19
|
+
async append(streamId, changes, expectedRevision, trx) {
|
|
20
|
+
const collection = this.getCollection(streamId);
|
|
21
|
+
const result = [];
|
|
22
|
+
let revision = expectedRevision + 1;
|
|
23
|
+
for (const change of changes) {
|
|
24
|
+
const storageChange = {
|
|
25
|
+
aggregateType: streamId.aggregateType,
|
|
26
|
+
eventId: change.id,
|
|
27
|
+
aggregateId: streamId.aggregateId,
|
|
28
|
+
revision,
|
|
29
|
+
name: change.name,
|
|
30
|
+
payload: change.payload,
|
|
31
|
+
occurredAt: serverTimestamp(),
|
|
32
|
+
version: change.version
|
|
33
|
+
};
|
|
34
|
+
const ref = collection.doc(`${revision}`);
|
|
35
|
+
result.push({
|
|
36
|
+
...change,
|
|
37
|
+
ref: ref.path,
|
|
38
|
+
revision,
|
|
39
|
+
occurredAt: void 0
|
|
40
|
+
});
|
|
41
|
+
trx.transaction.create(ref, this.converter.toFirestore(storageChange));
|
|
42
|
+
revision++;
|
|
43
|
+
}
|
|
44
|
+
return result;
|
|
45
|
+
}
|
|
46
|
+
async *read(streamId, startAt) {
|
|
47
|
+
const query = this.getCollection(streamId).where("revision", ">=", startAt || 0).orderBy("revision", "asc");
|
|
48
|
+
for await (const event of query.stream()) {
|
|
49
|
+
const e = event;
|
|
50
|
+
const data = this.converter.fromFirestore(e);
|
|
51
|
+
yield {
|
|
52
|
+
id: data.eventId,
|
|
53
|
+
ref: e.ref.path,
|
|
54
|
+
revision: data.revision,
|
|
55
|
+
name: data.name,
|
|
56
|
+
$name: data.name,
|
|
57
|
+
payload: data.payload,
|
|
58
|
+
occurredAt: data.occurredAt,
|
|
59
|
+
version: data.version ?? 1
|
|
60
|
+
};
|
|
61
|
+
}
|
|
62
|
+
}
|
|
63
|
+
};
|
|
64
|
+
|
|
65
|
+
//#endregion
|
|
66
|
+
export { FirestoreEventStreamStorageLayer };
|
|
67
|
+
//# sourceMappingURL=firestore.event-stream.storage-layer.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-stream.storage-layer.mjs","names":[],"sources":["../src/firestore.event-stream.storage-layer.ts"],"sourcesContent":["import {\n StreamId,\n type ISerializedChange,\n type ISerializedFact,\n EventStreamStorageLayer,\n} from \"@ddd-ts/core\";\nimport { ISerializedSavedChange } from \"@ddd-ts/core/dist/interfaces/es-event\";\n\nimport {\n DefaultConverter,\n FirestoreTransaction,\n} from \"@ddd-ts/store-firestore\";\nimport * as fb from \"firebase-admin\";\n\nexport const serverTimestamp = fb.firestore.FieldValue.serverTimestamp;\n\nexport class FirestoreEventStreamStorageLayer\n implements EventStreamStorageLayer\n{\n constructor(\n public readonly firestore: fb.firestore.Firestore,\n public readonly converter = new DefaultConverter(),\n ) {}\n\n isLocalRevisionOutdatedError(error: unknown): boolean {\n return (\n typeof error === \"object\" &&\n error !== null &&\n \"code\" in error &&\n error.code === 6\n );\n }\n\n getCollection(streamId: StreamId) {\n return this.firestore\n .collection(\"event-store\")\n .doc(streamId.aggregateType)\n .collection(\"streams\")\n .doc(streamId.aggregateId)\n .collection(\"events\");\n }\n\n async append(\n streamId: StreamId,\n changes: ISerializedChange[],\n expectedRevision: number,\n trx: FirestoreTransaction,\n ) {\n const collection = this.getCollection(streamId);\n const result: ISerializedSavedChange[] = [];\n\n let revision = expectedRevision + 1;\n for (const change of changes) {\n const storageChange = {\n aggregateType: streamId.aggregateType,\n eventId: change.id,\n aggregateId: streamId.aggregateId,\n revision: revision,\n name: change.name,\n payload: change.payload,\n occurredAt: serverTimestamp(),\n version: change.version,\n };\n\n const ref = collection.doc(`${revision}`);\n\n result.push({\n ...change,\n ref: ref.path,\n revision: revision,\n occurredAt: undefined,\n });\n\n trx.transaction.create(ref, this.converter.toFirestore(storageChange));\n revision++;\n }\n\n return result;\n }\n\n async *read(\n streamId: StreamId,\n startAt?: number,\n ): AsyncIterable<ISerializedFact> {\n const collection = this.getCollection(streamId);\n\n const query = collection\n .where(\"revision\", \">=\", startAt || 0)\n .orderBy(\"revision\", \"asc\");\n\n for await (const event of query.stream()) {\n const e = event as any as fb.firestore.QueryDocumentSnapshot<any>;\n const data = this.converter.fromFirestore(e);\n yield {\n id: data.eventId,\n ref: e.ref.path,\n revision: data.revision,\n name: data.name,\n $name: data.name,\n payload: data.payload,\n occurredAt: data.occurredAt,\n version: data.version ?? 1,\n };\n }\n }\n}\n"],"mappings":";;;;;;AAcA,MAAa,kBAAkB,GAAG,UAAU,WAAW;AAEvD,IAAa,mCAAb,MAEA;CACE,YACE,AAAgB,WAChB,AAAgB,YAAY,IAAI,kBAAkB,EAClD;EAFgB;EACA;;CAGlB,6BAA6B,OAAyB;AACpD,SACE,OAAO,UAAU,YACjB,UAAU,QACV,UAAU,SACV,MAAM,SAAS;;CAInB,cAAc,UAAoB;AAChC,SAAO,KAAK,UACT,WAAW,cAAc,CACzB,IAAI,SAAS,cAAc,CAC3B,WAAW,UAAU,CACrB,IAAI,SAAS,YAAY,CACzB,WAAW,SAAS;;CAGzB,MAAM,OACJ,UACA,SACA,kBACA,KACA;EACA,MAAM,aAAa,KAAK,cAAc,SAAS;EAC/C,MAAM,SAAmC,EAAE;EAE3C,IAAI,WAAW,mBAAmB;AAClC,OAAK,MAAM,UAAU,SAAS;GAC5B,MAAM,gBAAgB;IACpB,eAAe,SAAS;IACxB,SAAS,OAAO;IAChB,aAAa,SAAS;IACZ;IACV,MAAM,OAAO;IACb,SAAS,OAAO;IAChB,YAAY,iBAAiB;IAC7B,SAAS,OAAO;IACjB;GAED,MAAM,MAAM,WAAW,IAAI,GAAG,WAAW;AAEzC,UAAO,KAAK;IACV,GAAG;IACH,KAAK,IAAI;IACC;IACV,YAAY;IACb,CAAC;AAEF,OAAI,YAAY,OAAO,KAAK,KAAK,UAAU,YAAY,cAAc,CAAC;AACtE;;AAGF,SAAO;;CAGT,OAAO,KACL,UACA,SACgC;EAGhC,MAAM,QAFa,KAAK,cAAc,SAAS,CAG5C,MAAM,YAAY,MAAM,WAAW,EAAE,CACrC,QAAQ,YAAY,MAAM;AAE7B,aAAW,MAAM,SAAS,MAAM,QAAQ,EAAE;GACxC,MAAM,IAAI;GACV,MAAM,OAAO,KAAK,UAAU,cAAc,EAAE;AAC5C,SAAM;IACJ,IAAI,KAAK;IACT,KAAK,EAAE,IAAI;IACX,UAAU,KAAK;IACf,MAAM,KAAK;IACX,OAAO,KAAK;IACZ,SAAS,KAAK;IACd,YAAY,KAAK;IACjB,SAAS,KAAK,WAAW;IAC1B"}
|
|
@@ -1,14 +1,14 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
const require_firestore_event_stream_storage_layer = require('./firestore.event-stream.storage-layer.js');
|
|
3
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
4
|
+
let firebase_admin_firestore = require("firebase-admin/firestore");
|
|
4
5
|
|
|
5
6
|
//#region src/firestore.event-stream.store.ts
|
|
6
|
-
var FirestoreEventStreamStore = class extends EventStreamStore {
|
|
7
|
+
var FirestoreEventStreamStore = class extends _ddd_ts_core.EventStreamStore {
|
|
7
8
|
constructor(firestore, serializer, eventBus) {
|
|
8
|
-
super(new FirestoreEventStreamStorageLayer(firestore), serializer, eventBus);
|
|
9
|
+
super(new require_firestore_event_stream_storage_layer.FirestoreEventStreamStorageLayer(firestore), serializer, eventBus);
|
|
9
10
|
}
|
|
10
11
|
};
|
|
11
12
|
|
|
12
13
|
//#endregion
|
|
13
|
-
|
|
14
|
-
//# sourceMappingURL=firestore.event-stream.store.js.map
|
|
14
|
+
exports.FirestoreEventStreamStore = FirestoreEventStreamStore;
|
|
@@ -0,0 +1,14 @@
|
|
|
1
|
+
import { FirestoreEventStreamStorageLayer } from "./firestore.event-stream.storage-layer.mjs";
|
|
2
|
+
import { EventStreamStore } from "@ddd-ts/core";
|
|
3
|
+
import { Firestore } from "firebase-admin/firestore";
|
|
4
|
+
|
|
5
|
+
//#region src/firestore.event-stream.store.ts
|
|
6
|
+
var FirestoreEventStreamStore = class extends EventStreamStore {
|
|
7
|
+
constructor(firestore, serializer, eventBus) {
|
|
8
|
+
super(new FirestoreEventStreamStorageLayer(firestore), serializer, eventBus);
|
|
9
|
+
}
|
|
10
|
+
};
|
|
11
|
+
|
|
12
|
+
//#endregion
|
|
13
|
+
export { FirestoreEventStreamStore };
|
|
14
|
+
//# sourceMappingURL=firestore.event-stream.store.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.event-stream.store.mjs","names":[],"sources":["../src/firestore.event-stream.store.ts"],"sourcesContent":["import {\n EventStreamStore,\n type IEsEvent,\n type IEventBus,\n type ISerializer,\n} from \"@ddd-ts/core\";\nimport { FirestoreEventStreamStorageLayer } from \"./firestore.event-stream.storage-layer\";\nimport { Firestore } from \"firebase-admin/firestore\";\n\nexport class FirestoreEventStreamStore<\n Event extends IEsEvent,\n> extends EventStreamStore<Event> {\n constructor(\n firestore: Firestore,\n serializer: ISerializer<Event>,\n eventBus?: IEventBus,\n ) {\n super(\n new FirestoreEventStreamStorageLayer(firestore),\n serializer,\n eventBus,\n );\n }\n}\n"],"mappings":";;;;;AASA,IAAa,4BAAb,cAEU,iBAAwB;CAChC,YACE,WACA,YACA,UACA;AACA,QACE,IAAI,iCAAiC,UAAU,EAC/C,YACA,SACD"}
|
|
@@ -1,13 +1,14 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
const require_runtime = require('./_virtual/_rolldown/runtime.js');
|
|
2
|
+
const require_firestore_projected_stream_storage_layer = require('./firestore.projected-stream.storage-layer.js');
|
|
3
|
+
let _ddd_ts_core = require("@ddd-ts/core");
|
|
4
|
+
let firebase_admin_firestore = require("firebase-admin/firestore");
|
|
4
5
|
|
|
5
6
|
//#region src/firestore.projected-stream.reader.ts
|
|
6
|
-
var FirestoreProjectedStreamReader = class extends ProjectedStreamReader {
|
|
7
|
+
var FirestoreProjectedStreamReader = class extends _ddd_ts_core.ProjectedStreamReader {
|
|
7
8
|
storage;
|
|
8
9
|
serializer;
|
|
9
10
|
constructor(firestore, serializer) {
|
|
10
|
-
const storage = new FirestoreProjectedStreamStorageLayer(firestore);
|
|
11
|
+
const storage = new require_firestore_projected_stream_storage_layer.FirestoreProjectedStreamStorageLayer(firestore);
|
|
11
12
|
super(storage, serializer);
|
|
12
13
|
this.storage = storage;
|
|
13
14
|
this.serializer = serializer;
|
|
@@ -31,5 +32,4 @@ var FirestoreProjectedStreamReader = class extends ProjectedStreamReader {
|
|
|
31
32
|
};
|
|
32
33
|
|
|
33
34
|
//#endregion
|
|
34
|
-
|
|
35
|
-
//# sourceMappingURL=firestore.projected-stream.reader.js.map
|
|
35
|
+
exports.FirestoreProjectedStreamReader = FirestoreProjectedStreamReader;
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
import { FirestoreProjectedStreamStorageLayer } from "./firestore.projected-stream.storage-layer.mjs";
|
|
2
|
+
import { Cursor, IEsEvent, IFact, ISavedChange, ISerializedSavedChange, ISerializer, ProjectedStream, ProjectedStreamReader } from "@ddd-ts/core";
|
|
3
|
+
import { Firestore } from "firebase-admin/firestore";
|
|
4
|
+
|
|
5
|
+
//#region src/firestore.projected-stream.reader.ts
|
|
6
|
+
var FirestoreProjectedStreamReader = class extends ProjectedStreamReader {
|
|
7
|
+
storage;
|
|
8
|
+
serializer;
|
|
9
|
+
constructor(firestore, serializer) {
|
|
10
|
+
const storage = new FirestoreProjectedStreamStorageLayer(firestore);
|
|
11
|
+
super(storage, serializer);
|
|
12
|
+
this.storage = storage;
|
|
13
|
+
this.serializer = serializer;
|
|
14
|
+
}
|
|
15
|
+
async getCursor(savedChange) {
|
|
16
|
+
const serialized = await this.serializer.serialize(savedChange);
|
|
17
|
+
return this.storage.getCursor(serialized);
|
|
18
|
+
}
|
|
19
|
+
async get(cursor) {
|
|
20
|
+
const serialized = await this.storage.get(cursor);
|
|
21
|
+
if (!serialized) return;
|
|
22
|
+
return this.serializer.deserialize(serialized);
|
|
23
|
+
}
|
|
24
|
+
async slice(projectedStream, shard, startAfter, endAt, limit) {
|
|
25
|
+
const serialized = await this.storage.slice(projectedStream, shard, startAfter, endAt, limit);
|
|
26
|
+
return Promise.all(serialized.map((s) => this.serializer.deserialize(s)));
|
|
27
|
+
}
|
|
28
|
+
async *read(projectedStream, shard, startAfter, endAt) {
|
|
29
|
+
for await (const serialized of this.storage.read(projectedStream, shard, startAfter, endAt)) yield this.serializer.deserialize(serialized);
|
|
30
|
+
}
|
|
31
|
+
};
|
|
32
|
+
|
|
33
|
+
//#endregion
|
|
34
|
+
export { FirestoreProjectedStreamReader };
|
|
35
|
+
//# sourceMappingURL=firestore.projected-stream.reader.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"firestore.projected-stream.reader.mjs","names":[],"sources":["../src/firestore.projected-stream.reader.ts"],"sourcesContent":["import {\n Cursor,\n IEsEvent,\n IFact,\n ISavedChange,\n ISerializedSavedChange,\n ISerializer,\n ProjectedStream,\n ProjectedStreamReader,\n} from \"@ddd-ts/core\";\nimport { Firestore } from \"firebase-admin/firestore\";\nimport { FirestoreProjectedStreamStorageLayer } from \"./firestore.projected-stream.storage-layer\";\n\nexport class FirestoreProjectedStreamReader<\n Event extends IEsEvent,\n> extends ProjectedStreamReader<Event> {\n storage: FirestoreProjectedStreamStorageLayer;\n serializer: ISerializer<Event>;\n constructor(firestore: Firestore, serializer: ISerializer<Event>) {\n const storage = new FirestoreProjectedStreamStorageLayer(firestore);\n super(storage, serializer);\n this.storage = storage;\n this.serializer = serializer;\n }\n\n async getCursor(savedChange: ISavedChange<Event>) {\n const serialized = await this.serializer.serialize(savedChange);\n return this.storage.getCursor(serialized as ISerializedSavedChange);\n }\n\n async get(cursor: Cursor) {\n const serialized = await this.storage.get(cursor);\n if (!serialized) {\n return undefined;\n }\n return this.serializer.deserialize(serialized) as unknown as Promise<\n IFact<Event>\n >;\n }\n\n async slice(\n projectedStream: ProjectedStream,\n shard: string,\n startAfter?: Cursor,\n endAt?: Cursor,\n limit?: number,\n ) {\n const serialized = await this.storage.slice(\n projectedStream,\n shard,\n startAfter,\n endAt,\n limit,\n );\n return Promise.all(\n serialized.map((s) => this.serializer.deserialize(s)),\n ) as any;\n }\n\n async *read(\n projectedStream: ProjectedStream,\n shard: string,\n startAfter?: Cursor,\n endAt?: Cursor,\n ) {\n for await (const serialized of this.storage.read(\n projectedStream,\n shard,\n startAfter,\n endAt,\n )) {\n yield this.serializer.deserialize(serialized) as unknown as IFact<Event>;\n }\n }\n}\n"],"mappings":";;;;;AAaA,IAAa,iCAAb,cAEU,sBAA6B;CACrC;CACA;CACA,YAAY,WAAsB,YAAgC;EAChE,MAAM,UAAU,IAAI,qCAAqC,UAAU;AACnE,QAAM,SAAS,WAAW;AAC1B,OAAK,UAAU;AACf,OAAK,aAAa;;CAGpB,MAAM,UAAU,aAAkC;EAChD,MAAM,aAAa,MAAM,KAAK,WAAW,UAAU,YAAY;AAC/D,SAAO,KAAK,QAAQ,UAAU,WAAqC;;CAGrE,MAAM,IAAI,QAAgB;EACxB,MAAM,aAAa,MAAM,KAAK,QAAQ,IAAI,OAAO;AACjD,MAAI,CAAC,WACH;AAEF,SAAO,KAAK,WAAW,YAAY,WAAW;;CAKhD,MAAM,MACJ,iBACA,OACA,YACA,OACA,OACA;EACA,MAAM,aAAa,MAAM,KAAK,QAAQ,MACpC,iBACA,OACA,YACA,OACA,MACD;AACD,SAAO,QAAQ,IACb,WAAW,KAAK,MAAM,KAAK,WAAW,YAAY,EAAE,CAAC,CACtD;;CAGH,OAAO,KACL,iBACA,OACA,YACA,OACA;AACA,aAAW,MAAM,cAAc,KAAK,QAAQ,KAC1C,iBACA,OACA,YACA,MACD,CACC,OAAM,KAAK,WAAW,YAAY,WAAW"}
|