@event-driven-io/pongo 0.17.0-beta.33 → 0.17.0-beta.34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (46) hide show
  1. package/dist/{chunk-TTZGGAZV.js → chunk-OJVG4KXA.js} +69 -2
  2. package/dist/chunk-OJVG4KXA.js.map +1 -0
  3. package/dist/chunk-TYTEQJBC.cjs +1583 -0
  4. package/dist/chunk-TYTEQJBC.cjs.map +1 -0
  5. package/dist/{chunk-H637RRXS.js → chunk-U4NNEXNH.js} +1310 -632
  6. package/dist/chunk-U4NNEXNH.js.map +1 -0
  7. package/dist/{chunk-A4DCNQJR.cjs → chunk-UP6HTRMM.cjs} +79 -12
  8. package/dist/chunk-UP6HTRMM.cjs.map +1 -0
  9. package/dist/cli.cjs +13 -13
  10. package/dist/cli.cjs.map +1 -1
  11. package/dist/cli.js +2 -2
  12. package/dist/cli.js.map +1 -1
  13. package/dist/cloudflare.cjs +10 -10
  14. package/dist/cloudflare.d.cts +3 -2
  15. package/dist/cloudflare.d.ts +3 -2
  16. package/dist/cloudflare.js +2 -2
  17. package/dist/{index-gHRYr05w.d.ts → index-DHszkVjP.d.ts} +1 -1
  18. package/dist/{index-DxHXL62G.d.cts → index-DLDCIBgH.d.cts} +1 -1
  19. package/dist/index.cjs +26 -2
  20. package/dist/index.cjs.map +1 -1
  21. package/dist/index.d.cts +38 -16
  22. package/dist/index.d.ts +38 -16
  23. package/dist/index.js +25 -1
  24. package/dist/index.js.map +1 -1
  25. package/dist/pg.cjs +102 -16
  26. package/dist/pg.cjs.map +1 -1
  27. package/dist/pg.d.cts +2 -1
  28. package/dist/pg.d.ts +2 -1
  29. package/dist/pg.js +87 -1
  30. package/dist/pg.js.map +1 -1
  31. package/dist/{pongoCollectionSchemaComponent-B5Oatwu0.d.cts → pongoTransactionCache-BsigBOq1.d.cts} +154 -34
  32. package/dist/{pongoCollectionSchemaComponent-B5Oatwu0.d.ts → pongoTransactionCache-BsigBOq1.d.ts} +154 -34
  33. package/dist/shim.cjs +5 -5
  34. package/dist/shim.d.cts +2 -1
  35. package/dist/shim.d.ts +2 -1
  36. package/dist/shim.js +1 -1
  37. package/dist/sqlite3.cjs +10 -10
  38. package/dist/sqlite3.d.cts +3 -2
  39. package/dist/sqlite3.d.ts +3 -2
  40. package/dist/sqlite3.js +2 -2
  41. package/package.json +20 -18
  42. package/dist/chunk-A4DCNQJR.cjs.map +0 -1
  43. package/dist/chunk-BZRKCNRY.cjs +0 -905
  44. package/dist/chunk-BZRKCNRY.cjs.map +0 -1
  45. package/dist/chunk-H637RRXS.js.map +0 -1
  46. package/dist/chunk-TTZGGAZV.js.map +0 -1
package/dist/shim.d.ts CHANGED
@@ -1,7 +1,8 @@
1
1
  import { ClientSessionOptions } from 'http2';
2
2
  import { Document, Collection as Collection$1, ObjectId, ClientSession, WithSessionCallback, Db as Db$1, ReadConcern, ReadPreference, BSONSerializeOptions, WriteConcern, Hint, OptionalUnlessRequiredId, InsertOneOptions, InsertOneResult, BulkWriteOptions, InsertManyResult, AnyBulkWriteOperation, BulkWriteResult, Filter, UpdateFilter, UpdateOptions, UpdateResult, WithoutId, ReplaceOptions, DeleteOptions, DeleteResult, RenameOptions, DropCollectionOptions, WithId, FindOptions, FindCursor as FindCursor$1, OperationOptions, IndexSpecification, CreateIndexesOptions, IndexDescription, CommandOperationOptions, AbstractCursorOptions, ListIndexesCursor, IndexInformationOptions, IndexDescriptionInfo, IndexDescriptionCompact, EstimatedDocumentCountOptions, CountDocumentsOptions, EnhancedOmit, Flatten, FindOneAndDeleteOptions, ModifyResult, FindOneAndReplaceOptions, FindOneAndUpdateOptions, AggregateOptions, AggregationCursor, ChangeStreamDocument, ChangeStreamOptions, ChangeStream, UnorderedBulkOperation, OrderedBulkOperation, CountOptions, ListSearchIndexesOptions, ListSearchIndexesCursor, SearchIndexDescription } from 'mongodb';
3
- import { a as PongoDb, a9 as PongoDBCollectionOptions, H as DocumentHandler, O as HandleOptions, al as PongoHandleResult, k as AnyPongoDriver, f as PongoClientSchema, l as PongoClientOptions, e as PongoCollection } from './pongoCollectionSchemaComponent-B5Oatwu0.js';
3
+ import { a as PongoDb, ao as PongoDBCollectionOptions, T as DocumentHandler, _ as HandleOptions, aB as PongoHandleResult, o as AnyPongoDriver, j as PongoClientSchema, p as PongoClientOptions, g as PongoCollection } from './pongoTransactionCache-BsigBOq1.js';
4
4
  import '@event-driven-io/dumbo';
5
+ import 'lru-cache';
5
6
 
6
7
  declare class FindCursor<T> {
7
8
  private findDocumentsPromise;
package/dist/shim.js CHANGED
@@ -1,7 +1,7 @@
1
1
  import {
2
2
  pongoClient,
3
3
  pongoSession
4
- } from "./chunk-H637RRXS.js";
4
+ } from "./chunk-U4NNEXNH.js";
5
5
 
6
6
  // src/mongo/findCursor.ts
7
7
  var FindCursor = class {
package/dist/sqlite3.cjs CHANGED
@@ -1,14 +1,14 @@
1
1
  "use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }
2
2
 
3
3
 
4
- var _chunkA4DCNQJRcjs = require('./chunk-A4DCNQJR.cjs');
4
+ var _chunkUP6HTRMMcjs = require('./chunk-UP6HTRMM.cjs');
5
5
 
6
6
 
7
7
 
8
8
 
9
9
 
10
10
 
11
- var _chunkBZRKCNRYcjs = require('./chunk-BZRKCNRY.cjs');
11
+ var _chunkTYTEQJBCcjs = require('./chunk-TYTEQJBC.cjs');
12
12
 
13
13
  // src/storage/sqlite/sqlite3/index.ts
14
14
  var _dumbo = require('@event-driven-io/dumbo');
@@ -20,7 +20,7 @@ var sqlite3PongoDriver = {
20
20
  driverType: _sqlite3.SQLite3DriverType,
21
21
  databaseFactory: (options) => {
22
22
  const databaseName = _nullishCoalesce(options.databaseName, () => ( "db:default"));
23
- return _chunkBZRKCNRYcjs.PongoDatabase.call(void 0, {
23
+ return _chunkTYTEQJBCcjs.PongoDatabase.call(void 0, {
24
24
  ...options,
25
25
  pool: _dumbo.dumbo.call(void 0, {
26
26
  connectionString: options.connectionString,
@@ -28,27 +28,27 @@ var sqlite3PongoDriver = {
28
28
  ...options.connectionOptions,
29
29
  serialization: { serializer: options.serializer }
30
30
  }),
31
- schemaComponent: _chunkBZRKCNRYcjs.PongoDatabaseSchemaComponent.call(void 0, {
31
+ schemaComponent: _chunkTYTEQJBCcjs.PongoDatabaseSchemaComponent.call(void 0, {
32
32
  driverType: _sqlite3.SQLite3DriverType,
33
- collectionFactory: (schema) => _chunkBZRKCNRYcjs.PongoCollectionSchemaComponent.call(void 0, {
33
+ collectionFactory: (schema) => _chunkTYTEQJBCcjs.PongoCollectionSchemaComponent.call(void 0, {
34
34
  driverType: _sqlite3.SQLite3DriverType,
35
35
  definition: schema,
36
36
  migrationsOrSchemaComponents: {
37
- migrations: _chunkA4DCNQJRcjs.pongoCollectionSQLiteMigrations.call(void 0, schema.name)
37
+ migrations: _chunkUP6HTRMMcjs.pongoCollectionSQLiteMigrations.call(void 0, schema.name)
38
38
  },
39
- sqlBuilder: _chunkA4DCNQJRcjs.sqliteSQLBuilder.call(void 0,
39
+ sqlBuilder: _chunkUP6HTRMMcjs.sqliteSQLBuilder.call(void 0,
40
40
  schema.name,
41
41
  _nullishCoalesce(_optionalChain([options, 'access', _ => _.serialization, 'optionalAccess', _2 => _2.serializer]), () => ( _dumbo.JSONSerializer))
42
42
  )
43
43
  }),
44
- definition: _nullishCoalesce(_optionalChain([options, 'access', _3 => _3.schema, 'optionalAccess', _4 => _4.definition]), () => ( _chunkBZRKCNRYcjs.pongoSchema.db(databaseName, {})))
44
+ definition: _nullishCoalesce(_optionalChain([options, 'access', _3 => _3.schema, 'optionalAccess', _4 => _4.definition]), () => ( _chunkTYTEQJBCcjs.pongoSchema.db(databaseName, {})))
45
45
  }),
46
46
  databaseName
47
47
  });
48
48
  }
49
49
  };
50
50
  var useSqlite3PongoDriver = () => {
51
- _chunkBZRKCNRYcjs.pongoDriverRegistry.register(_sqlite3.SQLite3DriverType, sqlite3PongoDriver);
51
+ _chunkTYTEQJBCcjs.pongoDriverRegistry.register(_sqlite3.SQLite3DriverType, sqlite3PongoDriver);
52
52
  };
53
53
  useSqlite3PongoDriver();
54
54
 
@@ -57,5 +57,5 @@ useSqlite3PongoDriver();
57
57
 
58
58
 
59
59
 
60
- exports.pongoCollectionSQLiteMigrations = _chunkA4DCNQJRcjs.pongoCollectionSQLiteMigrations; exports.pongoDriver = sqlite3PongoDriver; exports.sqlite3Driver = sqlite3PongoDriver; exports.sqliteSQLBuilder = _chunkA4DCNQJRcjs.sqliteSQLBuilder; exports.useSqlite3PongoDriver = useSqlite3PongoDriver;
60
+ exports.pongoCollectionSQLiteMigrations = _chunkUP6HTRMMcjs.pongoCollectionSQLiteMigrations; exports.pongoDriver = sqlite3PongoDriver; exports.sqlite3Driver = sqlite3PongoDriver; exports.sqliteSQLBuilder = _chunkUP6HTRMMcjs.sqliteSQLBuilder; exports.useSqlite3PongoDriver = useSqlite3PongoDriver;
61
61
  //# sourceMappingURL=sqlite3.cjs.map
@@ -1,7 +1,8 @@
1
- export { p as pongoCollectionSQLiteMigrations, s as sqliteSQLBuilder } from './index-DxHXL62G.cjs';
1
+ export { p as pongoCollectionSQLiteMigrations, s as sqliteSQLBuilder } from './index-DLDCIBgH.cjs';
2
2
  import { SQLite3DriverType } from '@event-driven-io/dumbo/sqlite3';
3
- import { P as PongoDriver, a as PongoDb, b as PongoDriverOptions } from './pongoCollectionSchemaComponent-B5Oatwu0.cjs';
3
+ import { P as PongoDriver, a as PongoDb, b as PongoDriverOptions } from './pongoTransactionCache-BsigBOq1.cjs';
4
4
  import '@event-driven-io/dumbo';
5
+ import 'lru-cache';
5
6
 
6
7
  type SQLitePongoClientOptions = object;
7
8
  type SQLiteDatabaseDriverOptions = PongoDriverOptions<SQLitePongoClientOptions> & {
package/dist/sqlite3.d.ts CHANGED
@@ -1,7 +1,8 @@
1
- export { p as pongoCollectionSQLiteMigrations, s as sqliteSQLBuilder } from './index-gHRYr05w.js';
1
+ export { p as pongoCollectionSQLiteMigrations, s as sqliteSQLBuilder } from './index-DHszkVjP.js';
2
2
  import { SQLite3DriverType } from '@event-driven-io/dumbo/sqlite3';
3
- import { P as PongoDriver, a as PongoDb, b as PongoDriverOptions } from './pongoCollectionSchemaComponent-B5Oatwu0.js';
3
+ import { P as PongoDriver, a as PongoDb, b as PongoDriverOptions } from './pongoTransactionCache-BsigBOq1.js';
4
4
  import '@event-driven-io/dumbo';
5
+ import 'lru-cache';
5
6
 
6
7
  type SQLitePongoClientOptions = object;
7
8
  type SQLiteDatabaseDriverOptions = PongoDriverOptions<SQLitePongoClientOptions> & {
package/dist/sqlite3.js CHANGED
@@ -1,14 +1,14 @@
1
1
  import {
2
2
  pongoCollectionSQLiteMigrations,
3
3
  sqliteSQLBuilder
4
- } from "./chunk-TTZGGAZV.js";
4
+ } from "./chunk-OJVG4KXA.js";
5
5
  import {
6
6
  PongoCollectionSchemaComponent,
7
7
  PongoDatabase,
8
8
  PongoDatabaseSchemaComponent,
9
9
  pongoDriverRegistry,
10
10
  pongoSchema
11
- } from "./chunk-H637RRXS.js";
11
+ } from "./chunk-U4NNEXNH.js";
12
12
 
13
13
  // src/storage/sqlite/sqlite3/index.ts
14
14
  import { dumbo, JSONSerializer } from "@event-driven-io/dumbo";
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@event-driven-io/pongo",
3
- "version": "0.17.0-beta.33",
3
+ "version": "0.17.0-beta.34",
4
4
  "description": "Pongo - Mongo with strong consistency on top of PostgreSQL",
5
5
  "type": "module",
6
6
  "scripts": {
@@ -10,19 +10,20 @@
10
10
  "test": "run-s test:unit test:int test:e2e",
11
11
  "test:postgresql": "run-s test:unit:postgresql test:int:postgresql test:e2e:postgresql",
12
12
  "test:sqlite": "run-s test:unit:sqlite test:int:sqlite test:e2e:sqlite",
13
- "test:unit": "glob-bin -d -c \"node --import tsx --test\" **/*.unit.spec.ts",
14
- "test:unit:postgresql": "glob-bin -d -c \"node --import tsx --test\" **/postgresql/**/*unit.spec.ts",
15
- "test:unit:sqlite": "glob-bin -d -c \"node --import tsx --test\" **/sqlite/**/*unit.spec.ts",
16
- "test:int": "glob-bin -d -c \"node --import tsx --test\" **/*.int.spec.ts",
17
- "test:int:postgresql": "glob-bin -d -c \"node --import tsx --test\" **/postgresql/**/*int.spec.ts",
18
- "test:int:sqlite": "glob-bin -d -c \"node --import tsx --test\" **/sqlite/**/*int.spec.ts",
19
- "test:e2e": "glob-bin -d -c \"node --import tsx --test\" **/*.e2e.spec.ts",
20
- "test:e2e:postgresql": "glob-bin -d -c \"node --import tsx --test\" **/postgresql/**/*e2e.spec.ts",
21
- "test:e2e:sqlite": "glob-bin -d -c \"node --import tsx --test\" **/sqlite/**/*e2e.spec.ts",
22
- "test:watch": "node --import tsx --test --watch",
23
- "test:unit:watch": "glob-bin -d -c \"node --import tsx --test --watch\" **/*.unit.spec.ts",
24
- "test:int:watch": "glob-bin -d -c \"node --import tsx --test --watch\" **/*.int.spec.ts",
25
- "test:e2e:watch": "glob-bin -d -c \"node --import tsx --test --watch\" **/*.e2e.spec.ts",
13
+ "test:unit": "vitest run \".unit.spec\"",
14
+ "test:unit:postgresql": "vitest run postgresql \".unit.spec\"",
15
+ "test:unit:sqlite": "vitest run sqlite \".unit.spec\"",
16
+ "test:int": "vitest run \".int.spec\"",
17
+ "test:int:postgresql": "vitest run postgresql \".int.spec\"",
18
+ "test:int:sqlite": "vitest run sqlite \".int.spec\"",
19
+ "test:e2e": "vitest run \".e2e.spec\"",
20
+ "test:e2e:postgresql": "vitest run postgresql \".e2e.spec\"",
21
+ "test:e2e:sqlite": "vitest run sqlite \".e2e.spec\"",
22
+ "test:file": "vitest run --no-file-parallelism",
23
+ "test:watch": "vitest",
24
+ "test:unit:watch": "vitest \".unit.spec\"",
25
+ "test:int:watch": "vitest \".int.spec\"",
26
+ "test:e2e:watch": "vitest \".e2e.spec\"",
26
27
  "cli:sql:print": "tsx src/cli.ts migrate sql --collection users",
27
28
  "cli:migrate:dryRun": "tsx src/cli.ts migrate run --config src/e2e/cli-config.ts -cs postgresql://postgres:postgres@localhost:5432/postgres",
28
29
  "cli:config:print": "tsx src/cli.ts config sample --print",
@@ -131,12 +132,13 @@
131
132
  "pongo": "dist/cli.js"
132
133
  },
133
134
  "dependencies": {
134
- "@event-driven-io/dumbo": "0.13.0-beta.33",
135
+ "@event-driven-io/dumbo": "0.13.0-beta.34",
135
136
  "@types/mongodb": "^4.0.7",
136
- "uuid": "^13.0.0",
137
137
  "ansis": "^4.2.0",
138
138
  "cli-table3": "^0.6.5",
139
- "commander": "^14.0.2"
139
+ "commander": "^14.0.2",
140
+ "lru-cache": "^11.2.7",
141
+ "uuid": "^13.0.0"
140
142
  },
141
143
  "peerDependencies": {
142
144
  "@cloudflare/workers-types": "^4.20260127.0",
@@ -160,8 +162,8 @@
160
162
  },
161
163
  "devDependencies": {
162
164
  "@cloudflare/workers-types": "^4.20260127.0",
163
- "@types/pg": "^8.16.0",
164
165
  "@types/node": "^24.5.2",
166
+ "@types/pg": "^8.16.0",
165
167
  "pg": "^8.17.2",
166
168
  "sqlite3": "^5.1.7"
167
169
  }
@@ -1 +0,0 @@
1
- {"version":3,"sources":["/home/runner/work/Pongo/Pongo/src/packages/pongo/dist/chunk-A4DCNQJR.cjs","../src/storage/sqlite/core/sqlBuilder/index.ts","../src/storage/sqlite/core/sqlBuilder/filter/index.ts","../src/storage/sqlite/core/sqlBuilder/filter/queryOperators.ts","../src/storage/sqlite/core/sqlBuilder/update/index.ts"],"names":["SQL"],"mappings":"AAAA;AACE;AACA;AACA;AACA;AACA;AACF,wDAA6B;AAC7B;AACA;ACPA,+CAAyC;ADSzC;AACA;AEVA;AFYA;AACA;AGbA;AAGO,IAAM,eAAA,EAAiB,CAC5B,IAAA,EACA,QAAA,EACA,KAAA,EACA,UAAA,EAAA,GACQ;AACR,EAAA,GAAA,CAAI,KAAA,IAAS,MAAA,GAAS,KAAA,IAAS,UAAA,EAAY;AACzC,IAAA,OAAO,sBAAA,CAAuB,IAAA,EAAM,QAAA,EAAU,KAAK,CAAA;AAAA,EACrD;AAEA,EAAA,OAAA,CAAQ,QAAA,EAAU;AAAA,IAChB,KAAK,KAAA,EAAO;AACV,MAAA,MAAM,SAAA,EAAW,aAAA,CAAc,IAAI,CAAA;AAEnC,MAAA,OAAO,UAAA,CAAA;AAAA,4BAAA,EACiB,UAAA,CAAI,KAAA,CAAM,QAAQ,CAAC,CAAA,KAAA,EAAQ,KAAK,CAAA;AAAA;AAAA,2BAAA,EAEjC,UAAA,CAAI,KAAA,CAAM,QAAQ,CAAC,CAAA;AAAA;AAAA,2CAAA,EAEH,UAAA,CAAI,KAAA,CAAM,QAAQ,CAAC,CAAA;AAAA,oCAAA,EAC1B,KAAK,CAAA;AAAA;AAAA;AAAA,OAAA,CAAA;AAAA,IAIvC;AAAA,IACA,KAAK,KAAA;AAAA,IACL,KAAK,MAAA;AAAA,IACL,KAAK,KAAA;AAAA,IACL,KAAK,MAAA;AAAA,IACL,KAAK,KAAA,EAAO;AACV,MAAA,MAAM,SAAA,EAAW,aAAA,CAAc,IAAI,CAAA;AAEnC,MAAA,OAAO,UAAA,CAAA,oBAAA,EAA0B,UAAA,CAAI,KAAA,CAAM,QAAQ,CAAC,CAAA,GAAA,EAAM,UAAA,CAAI,KAAA,CAAM,6BAAA,CAAY,QAAQ,CAAC,CAAC,CAAA,CAAA,EAAI,KAAK,CAAA,CAAA;AACrG,IAAA;AACY,IAAA;AACyB,MAAA;AACpB,MAAA;AACM,MAAA;AACQ,QAAA;AAC3B,QAAA;AACF,MAAA;AAEsE,MAAA;AACxE,IAAA;AACa,IAAA;AACwB,MAAA;AACpB,MAAA;AACM,MAAA;AACQ,QAAA;AAC3B,QAAA;AACF,MAAA;AAE0E,MAAA;AAC5E,IAAA;AACmB,IAAA;AAEc,MAAA;AAC0B,QAAA;AACe,QAAA;AAEzD,MAAA;AAEoB,MAAA;AACyD,MAAA;AAC9F,IAAA;AACa,IAAA;AACwB,MAAA;AACe,MAAA;AAEe,MAAA;AACnE,IAAA;AACc,IAAA;AACuB,MAAA;AAEiD,MAAA;AACtF,IAAA;AACA,IAAA;AACqD,MAAA;AACvD,EAAA;AACF;AAMU;AACU,EAAA;AACX,IAAA;AACyC,MAAA;AACzC,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AACA,IAAA;AAC2E,MAAA;AACpE,IAAA;AACK,MAAA;AACM,MAAA;AACQ,QAAA;AAC3B,QAAA;AACF,MAAA;AACiD,MAAA;AACnD,IAAA;AACa,IAAA;AACI,MAAA;AACM,MAAA;AACQ,QAAA;AAC3B,QAAA;AACF,MAAA;AACqD,MAAA;AACvD,IAAA;AACA,IAAA;AACqD,MAAA;AACvD,EAAA;AACF;AAEgD;AAC9B,EAAA;AAClB;AHR0G;AACA;AEvG9F;AAMN;AACqB,EAAA;AAGjB,IAAA;AACN,EAAA;AACO,EAAA;AACT;AAMQ;AAC8B,EAAA;AAE3B,EAAA;AACY,IAAA;AAEf,MAAA;AACqB,QAAA;AACJ,QAAA;AACf,QAAA;AACA,QAAA;AAE4C,MAAA;AACpD,IAAA;AACO,IAAA;AACT,EAAA;AACF;AAGqE;AFwFqC;AACA;AI1ItF;AAcI;AACiB,EAAA;AACvB,IAAA;AACL,MAAA;AACuD,QAAA;AACvD,MAAA;AAC6C,QAAA;AAC7C,MAAA;AAC2C,QAAA;AAC3C,MAAA;AACwD,QAAA;AAC7D,MAAA;AACS,QAAA;AACX,IAAA;AACF,EAAA;AACAA,EAAAA;AACF;AAMgD;AAKxC;AACsB,EAAA;AAClB,EAAA;AACY,EAAA;AAC+B,IAAA;AACvD,EAAA;AACO,EAAA;AACT;AAKU;AACwC,EAAA;AAGW,IAAA;AAE3D,EAAA;AACO,EAAA;AACT;AAMU;AACyC,EAAA;AACG,IAAA;AAC0B,IAAA;AACJ,kCAAA;AACsB,oCAAA;AACtD,2BAAA;AAAA,QAAA;AAE1C,EAAA;AACO,EAAA;AACT;AJ0G0G;AACA;ACvKxGA;AAC6D,+BAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAAA;AAWY;AACP,EAAA;AACjC,IAAA;AAChC,EAAA;AACH;AAKiC;AAC6B,EAAA;AACY,EAAA;AACnD,IAAA;AACC,IAAA;AACiB,IAAA;AAE9BA,IAAAA;AACiD,4BAAA;AACf,cAAA;AAAA,oBAAA;AAE3C,EAAA;AAC4E,EAAA;AACvD,IAAA;AACP,MAAA;AAE6D,QAAA;AACvE,MAAA;AACA,MAAA;AACF,IAAA;AAEOA,IAAAA;AACwF,4BAAA;AAAA,oBAAA;AAEjG,EAAA;AAKU,EAAA;AAC6D,IAAA;AAEAA,IAAAA;AAI1B,IAAA;AAGJ,IAAA;AAEhCA,IAAAA;AACkC,aAAA;AAAA;AAEN,0BAAA;AAAA;AAAA;AAAA;AAIiB,wBAAA;AAC5B,QAAA;AAAA;AAEE,QAAA;AAAA;AAAA;AAAA;AAAA;AAAA,sBAAA;AAM5B,EAAA;AAKU,EAAA;AAC6D,IAAA;AAEAA,IAAAA;AAI1B,IAAA;AAEpCA,IAAAA;AACkC,aAAA;AAAA;AAEa,0BAAA;AAAA;AAAA;AAAA;AAIF,wBAAA;AAC5B,QAAA;AAAA;AAEE,QAAA;AAAA;AAAA;AAAA;AAAA;AAAA,sBAAA;AAM5B,EAAA;AAIU,EAAA;AAGmC,IAAA;AAGJ,IAAA;AAEhCA,IAAAA;AACkC,aAAA;AAAA;AAEN,0BAAA;AAAA;AAAA;AAGb,MAAA;AAAA,oBAAA;AAExB,EAAA;AAIU,EAAA;AAC6D,IAAA;AAEAA,IAAAA;AAI1B,IAAA;AAEpCA,IAAAA;AACuC,kBAAA;AAAA;AAEM,wBAAA;AAC5B,QAAA;AAAA;AAEE,QAAA;AAAA;AAAA;AAAA;AAAA,qBAAA;AAK5B,EAAA;AACsD,EAAA;AAGT,IAAA;AAEkC,IAAA;AAC/E,EAAA;AACmD,EAAA;AAGN,IAAA;AAEiD,IAAA;AAC9F,EAAA;AACuE,EAAA;AAG1B,IAAA;AACrB,IAAA;AAEhB,IAAA;AAC2D,MAAA;AACjE,IAAA;AAE6B,IAAA;AAET,IAAA;AACoB,MAAA;AACxC,IAAA;AAEmB,IAAA;AACqB,MAAA;AACxC,IAAA;AAEmC,IAAA;AACrC,EAAA;AAC0D,EAAA;AAGb,IAAA;AACoD,IAAA;AACjG,EAAA;AAE+E,EAAA;AAExB,EAAA;AACzD;AAK8B;ADwG4E;AACA;AACA;AACA;AACA","file":"/home/runner/work/Pongo/Pongo/src/packages/pongo/dist/chunk-A4DCNQJR.cjs","sourcesContent":[null,"import type { JSONSerializer } from '@event-driven-io/dumbo';\nimport { isSQL, SQL, sqlMigration } from '@event-driven-io/dumbo';\nimport {\n expectedVersionValue,\n type DeleteOneOptions,\n type FindOptions,\n type OptionalUnlessRequiredIdAndVersion,\n type PongoCollectionSQLBuilder,\n type PongoFilter,\n type PongoUpdate,\n type ReplaceOneOptions,\n type UpdateOneOptions,\n type WithoutId,\n} from '../../../../core';\nimport { constructFilterQuery } from './filter';\nimport { buildUpdateQuery } from './update';\n\nconst createCollection = (collectionName: string): SQL =>\n SQL`\n CREATE TABLE IF NOT EXISTS ${SQL.identifier(collectionName)} (\n _id TEXT PRIMARY KEY,\n data JSON NOT NULL,\n metadata JSON NOT NULL DEFAULT '{}',\n _version INTEGER NOT NULL DEFAULT 1,\n _partition TEXT NOT NULL DEFAULT 'png_global',\n _archived INTEGER NOT NULL DEFAULT 0,\n _created TEXT NOT NULL DEFAULT (datetime('now')),\n _updated TEXT NOT NULL DEFAULT (datetime('now'))\n )`;\n\nexport const pongoCollectionSQLiteMigrations = (collectionName: string) => [\n sqlMigration(`pongoCollection:${collectionName}:001:createtable`, [\n createCollection(collectionName),\n ]),\n];\n\nexport const sqliteSQLBuilder = (\n collectionName: string,\n serializer: JSONSerializer,\n): PongoCollectionSQLBuilder => ({\n createCollection: (): SQL => createCollection(collectionName),\n insertOne: <T>(document: OptionalUnlessRequiredIdAndVersion<T>): SQL => {\n const serialized = document;\n const id = document._id;\n const version = document._version ?? 1n;\n\n return SQL`\n INSERT OR IGNORE INTO ${SQL.identifier(collectionName)} (_id, data, _version)\n VALUES (${id}, ${serialized}, ${version})\n RETURNING _id;`;\n },\n insertMany: <T>(documents: OptionalUnlessRequiredIdAndVersion<T>[]): SQL => {\n const values = SQL.merge(\n documents.map(\n (doc) =>\n SQL`(${doc._id}, ${serializer.serialize(doc)}, ${doc._version ?? 1n})`,\n ),\n ',',\n );\n\n return SQL`\n INSERT OR IGNORE INTO ${SQL.identifier(collectionName)} (_id, data, _version) VALUES ${values}\n RETURNING _id;`;\n },\n updateOne: <T>(\n filter: PongoFilter<T> | SQL,\n update: PongoUpdate<T> | SQL,\n options?: UpdateOneOptions,\n ): SQL => {\n const expectedVersion = expectedVersionValue(options?.expectedVersion);\n const expectedVersionCheck =\n expectedVersion != null ? SQL`AND _version = ${expectedVersion}` : SQL``;\n\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n const updateQuery = isSQL(update)\n ? update\n : buildUpdateQuery(update, serializer);\n\n return SQL`\n UPDATE ${SQL.identifier(collectionName)}\n SET\n data = json_patch(${updateQuery}, json_object('_id', _id, '_version', cast(_version + 1 as TEXT))),\n _version = _version + 1,\n _updated = datetime('now')\n WHERE _id = (\n SELECT _id FROM ${SQL.identifier(collectionName)}\n ${where(filterQuery)}\n LIMIT 1\n ) ${expectedVersionCheck}\n RETURNING\n _id,\n cast(_version as TEXT) as version,\n 1 as matched,\n 1 as modified;`;\n },\n replaceOne: <T>(\n filter: PongoFilter<T> | SQL,\n document: WithoutId<T>,\n options?: ReplaceOneOptions,\n ): SQL => {\n const expectedVersion = expectedVersionValue(options?.expectedVersion);\n const expectedVersionCheck =\n expectedVersion != null ? SQL`AND _version = ${expectedVersion}` : SQL``;\n\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n\n return SQL`\n UPDATE ${SQL.identifier(collectionName)}\n SET\n data = json_patch(${serializer.serialize(document)}, json_object('_id', _id, '_version', cast(_version + 1 as TEXT))),\n _version = _version + 1,\n _updated = datetime('now')\n WHERE _id = (\n SELECT _id FROM ${SQL.identifier(collectionName)}\n ${where(filterQuery)}\n LIMIT 1\n ) ${expectedVersionCheck}\n RETURNING\n _id,\n cast(_version as TEXT) AS version,\n 1 AS matched,\n 1 AS modified;`;\n },\n updateMany: <T>(\n filter: PongoFilter<T> | SQL,\n update: PongoUpdate<T> | SQL,\n ): SQL => {\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n const updateQuery = isSQL(update)\n ? update\n : buildUpdateQuery(update, serializer);\n\n return SQL`\n UPDATE ${SQL.identifier(collectionName)}\n SET\n data = json_patch(${updateQuery}, json_object('_version', cast(_version + 1 as TEXT))),\n _version = _version + 1,\n _updated = datetime('now')\n ${where(filterQuery)}\n RETURNING _id;`;\n },\n deleteOne: <T>(\n filter: PongoFilter<T> | SQL,\n options?: DeleteOneOptions,\n ): SQL => {\n const expectedVersion = expectedVersionValue(options?.expectedVersion);\n const expectedVersionCheck =\n expectedVersion != null ? SQL`AND _version = ${expectedVersion}` : SQL``;\n\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n\n return SQL`\n DELETE FROM ${SQL.identifier(collectionName)}\n WHERE _id = (\n SELECT _id FROM ${SQL.identifier(collectionName)}\n ${where(filterQuery)}\n LIMIT 1\n ) ${expectedVersionCheck}\n RETURNING\n _id,\n 1 AS matched,\n 1 AS deleted;`;\n },\n deleteMany: <T>(filter: PongoFilter<T> | SQL): SQL => {\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n\n return SQL`DELETE FROM ${SQL.identifier(collectionName)} ${where(filterQuery)} RETURNING _id`;\n },\n findOne: <T>(filter: PongoFilter<T> | SQL): SQL => {\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n\n return SQL`SELECT data, _version FROM ${SQL.identifier(collectionName)} ${where(filterQuery)} LIMIT 1;`;\n },\n find: <T>(filter: PongoFilter<T> | SQL, options?: FindOptions): SQL => {\n const filterQuery = isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n const query: SQL[] = [];\n\n query.push(\n SQL`SELECT data, _version FROM ${SQL.identifier(collectionName)}`,\n );\n\n query.push(where(filterQuery));\n\n if (options?.limit) {\n query.push(SQL`LIMIT ${options.limit}`);\n }\n\n if (options?.skip) {\n query.push(SQL`OFFSET ${options.skip}`);\n }\n\n return SQL.merge([...query, SQL`;`]);\n },\n countDocuments: <T>(filter: PongoFilter<T> | SQL): SQL => {\n const filterQuery = SQL.check.isSQL(filter)\n ? filter\n : constructFilterQuery(filter, serializer);\n return SQL`SELECT COUNT(1) as count FROM ${SQL.identifier(collectionName)} ${where(filterQuery)};`;\n },\n rename: (newName: string): SQL =>\n SQL`ALTER TABLE ${SQL.identifier(collectionName)} RENAME TO ${SQL.identifier(newName)};`,\n drop: (targetName: string = collectionName): SQL =>\n SQL`DROP TABLE IF EXISTS ${SQL.identifier(targetName)}`,\n});\n\nconst where = (filterQuery: SQL): SQL =>\n SQL.check.isEmpty(filterQuery)\n ? SQL.EMPTY\n : SQL.merge([SQL`WHERE `, filterQuery]);\n","import type { JSONSerializer } from '@event-driven-io/dumbo';\nimport { SQL } from '@event-driven-io/dumbo';\nimport {\n hasOperators,\n objectEntries,\n QueryOperators,\n type PongoFilter,\n} from '../../../../../core';\nimport { handleOperator } from './queryOperators';\n\nexport * from './queryOperators';\n\nconst AND = 'AND';\n\nexport const constructFilterQuery = <T>(\n filter: PongoFilter<T>,\n serializer: JSONSerializer,\n): SQL =>\n SQL.merge(\n Object.entries(filter).map(([key, value]) =>\n isRecord(value)\n ? constructComplexFilterQuery(key, value, serializer)\n : handleOperator(key, '$eq', value, serializer),\n ),\n ` ${AND} `,\n );\n\nconst constructComplexFilterQuery = (\n key: string,\n value: Record<string, unknown>,\n serializer: JSONSerializer,\n): SQL => {\n const isEquality = !hasOperators(value);\n\n return SQL.merge(\n objectEntries(value).map(([nestedKey, val]) =>\n isEquality\n ? handleOperator(\n `${key}.${nestedKey}`,\n QueryOperators.$eq,\n val,\n serializer,\n )\n : handleOperator(key, nestedKey, val, serializer),\n ),\n ` ${AND} `,\n );\n};\n\nconst isRecord = (value: unknown): value is Record<string, unknown> =>\n value !== null && typeof value === 'object' && !Array.isArray(value);\n","import type { JSONSerializer } from '@event-driven-io/dumbo';\nimport { SQL } from '@event-driven-io/dumbo';\nimport { objectEntries, OperatorMap } from '../../../../../core';\n\nexport const handleOperator = (\n path: string,\n operator: string,\n value: unknown,\n serializer: JSONSerializer,\n): SQL => {\n if (path === '_id' || path === '_version') {\n return handleMetadataOperator(path, operator, value);\n }\n\n switch (operator) {\n case '$eq': {\n const jsonPath = buildJsonPath(path);\n\n return SQL`(\n json_extract(data, '${SQL.plain(jsonPath)}') = ${value}\n OR (\n json_type(data, '${SQL.plain(jsonPath)}') = 'array'\n AND EXISTS(\n SELECT 1 FROM json_each(data, '${SQL.plain(jsonPath)}')\n WHERE json_each.value = ${value}\n )\n )\n )`;\n }\n case '$gt':\n case '$gte':\n case '$lt':\n case '$lte':\n case '$ne': {\n const jsonPath = buildJsonPath(path);\n\n return SQL`json_extract(data, '${SQL.plain(jsonPath)}') ${SQL.plain(OperatorMap[operator])} ${value}`;\n }\n case '$in': {\n const jsonPath = buildJsonPath(path);\n const values = value as unknown[];\n const inClause = SQL.merge(\n values.map((v) => SQL`${v}`),\n ', ',\n );\n\n return SQL`json_extract(data, '${SQL.plain(jsonPath)}') IN (${inClause})`;\n }\n case '$nin': {\n const jsonPath = buildJsonPath(path);\n const values = value as unknown[];\n const inClause = SQL.merge(\n values.map((v) => SQL`${v}`),\n ', ',\n );\n\n return SQL`json_extract(data, '${SQL.plain(jsonPath)}') NOT IN (${inClause})`;\n }\n case '$elemMatch': {\n const subConditions = objectEntries(value as Record<string, unknown>)\n .map(([subKey, subValue]) => {\n const serializedValue = serializer.serialize(subValue);\n return `json_extract(value, '$.${subKey}') = json('${serializedValue}')`;\n })\n .join(' AND ');\n\n const jsonPath = buildJsonPath(path);\n return SQL`EXISTS(SELECT 1 FROM json_each(data, '${SQL.plain(jsonPath)}') WHERE ${SQL.plain(subConditions)})`;\n }\n case '$all': {\n const jsonPath = buildJsonPath(path);\n const serializedValue = serializer.serialize(value);\n\n return SQL`(SELECT COUNT(*) FROM json_each(json(${serializedValue})) WHERE json_each.value NOT IN (SELECT value FROM json_each(data, '${SQL.plain(jsonPath)}'))) = 0`;\n }\n case '$size': {\n const jsonPath = buildJsonPath(path);\n\n return SQL`json_array_length(json_extract(data, '${SQL.plain(jsonPath)}')) = ${value}`;\n }\n default:\n throw new Error(`Unsupported operator: ${operator}`);\n }\n};\n\nconst handleMetadataOperator = (\n fieldName: string,\n operator: string,\n value: unknown,\n): SQL => {\n switch (operator) {\n case '$eq':\n return SQL`${SQL.plain(fieldName)} = ${value}`;\n case '$gt':\n case '$gte':\n case '$lt':\n case '$lte':\n case '$ne':\n return SQL`${SQL.plain(fieldName)} ${SQL.plain(OperatorMap[operator])} ${value}`;\n case '$in': {\n const values = value as unknown[];\n const inClause = SQL.merge(\n values.map((v) => SQL`${v}`),\n ', ',\n );\n return SQL`${SQL.plain(fieldName)} IN (${inClause})`;\n }\n case '$nin': {\n const values = value as unknown[];\n const inClause = SQL.merge(\n values.map((v) => SQL`${v}`),\n ', ',\n );\n return SQL`${SQL.plain(fieldName)} NOT IN (${inClause})`;\n }\n default:\n throw new Error(`Unsupported operator: ${operator}`);\n }\n};\n\nconst buildJsonPath = (path: string): string => {\n return `$.${path}`;\n};\n","import type { JSONSerializer } from '@event-driven-io/dumbo';\nimport { SQL } from '@event-driven-io/dumbo';\nimport {\n objectEntries,\n type $inc,\n type $push,\n type $set,\n type $unset,\n type PongoUpdate,\n} from '../../../../../core';\n\nexport const buildUpdateQuery = <T>(\n update: PongoUpdate<T>,\n serializer: JSONSerializer,\n): SQL =>\n objectEntries(update).reduce(\n (currentUpdateQuery, [op, value]) => {\n switch (op) {\n case '$set':\n return buildSetQuery(value, currentUpdateQuery, serializer);\n case '$unset':\n return buildUnsetQuery(value, currentUpdateQuery);\n case '$inc':\n return buildIncQuery(value, currentUpdateQuery);\n case '$push':\n return buildPushQuery(value, currentUpdateQuery, serializer);\n default:\n return currentUpdateQuery;\n }\n },\n SQL`data`,\n );\n\nexport const buildSetQuery = <T>(\n set: $set<T>,\n currentUpdateQuery: SQL,\n serializer: JSONSerializer,\n): SQL => SQL`json_patch(${currentUpdateQuery}, ${serializer.serialize(set)})`;\n\nexport const buildUnsetQuery = <T>(\n unset: $unset<T>,\n currentUpdateQuery: SQL,\n): SQL => {\n const keys = Object.keys(unset);\n let query = currentUpdateQuery;\n for (const key of keys) {\n query = SQL`json_remove(${query}, '$.${SQL.plain(key)}')`;\n }\n return query;\n};\n\nexport const buildIncQuery = <T>(\n inc: $inc<T>,\n currentUpdateQuery: SQL,\n): SQL => {\n for (const [key, value] of Object.entries(inc)) {\n currentUpdateQuery =\n typeof value === 'bigint'\n ? SQL`json_set(${currentUpdateQuery}, '$.${SQL.plain(key)}', CAST((COALESCE(json_extract(${currentUpdateQuery}, '$.${SQL.plain(key)}'), 0) + ${value}) AS TEXT))`\n : SQL`json_set(${currentUpdateQuery}, '$.${SQL.plain(key)}', COALESCE(json_extract(${currentUpdateQuery}, '$.${SQL.plain(key)}'), 0) + ${value})`;\n }\n return currentUpdateQuery;\n};\n\nexport const buildPushQuery = <T>(\n push: $push<T>,\n currentUpdateQuery: SQL,\n serializer: JSONSerializer,\n): SQL => {\n for (const [key, value] of Object.entries(push)) {\n const serializedValue = serializer.serialize(value);\n currentUpdateQuery = SQL`json_set(${currentUpdateQuery}, '$.${SQL.plain(key)}', CASE\n WHEN json_type(json_extract(${currentUpdateQuery}, '$.${SQL.plain(key)}')) = 'array'\n THEN json_insert(json_extract(${currentUpdateQuery}, '$.${SQL.plain(key)}'), '$[#]', json(${serializedValue}))\n ELSE json_array(json(${serializedValue}))\n END)`;\n }\n return currentUpdateQuery;\n};\n"]}