@prisma-next/mongo-contract-psl 0.4.0-dev.1 → 0.4.0-dev.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,7 +3,6 @@ import { ContractConfig } from "@prisma-next/config/config-types";
3
3
  //#region src/provider.d.ts
4
4
  interface MongoContractOptions {
5
5
  readonly output?: string;
6
- readonly scalarTypeDescriptors?: ReadonlyMap<string, string>;
7
6
  }
8
7
  declare function mongoContract(schemaPath: string, options?: MongoContractOptions): ContractConfig;
9
8
  //#endregion
@@ -1 +1 @@
1
- {"version":3,"file":"provider.d.mts","names":[],"sources":["../../src/provider.ts"],"sourcesContent":[],"mappings":";;;UASiB,oBAAA;;EAAA,SAAA,qBAAoB,CAAA,EAEF,WAAA,CAAA,MAAW,EAAA,MAAA,CAAA;AAG9C;iBAAgB,aAAA,+BAA4C,uBAAuB"}
1
+ {"version":3,"file":"provider.d.mts","names":[],"sources":["../../src/provider.ts"],"sourcesContent":[],"mappings":";;;UAQiB,oBAAA;;AAAjB;AAIgB,iBAAA,aAAA,CAA4C,UAAA,EAAA,MAAuB,EAAA,OAAc,CAAd,EAAvB,oBAAqC,CAAA,EAAd,cAAc"}
@@ -1,4 +1,4 @@
1
- import { n as interpretPslDocumentToMongoContract, t as createMongoScalarTypeDescriptors } from "../scalar-type-descriptors-B0Lt0pJI.mjs";
1
+ import { t as interpretPslDocumentToMongoContract } from "../interpreter-CWOjqYYA.mjs";
2
2
  import { notOk, ok } from "@prisma-next/utils/result";
3
3
  import { parsePslDocument } from "@prisma-next/psl-parser";
4
4
  import { readFile } from "node:fs/promises";
@@ -8,7 +8,7 @@ import { resolve } from "pathe";
8
8
  //#region src/provider.ts
9
9
  function mongoContract(schemaPath, options) {
10
10
  return {
11
- source: async (_context) => {
11
+ source: async (context) => {
12
12
  const absoluteSchemaPath = resolve(schemaPath);
13
13
  let schema;
14
14
  try {
@@ -34,7 +34,8 @@ function mongoContract(schemaPath, options) {
34
34
  schema,
35
35
  sourceId: schemaPath
36
36
  }),
37
- scalarTypeDescriptors: options?.scalarTypeDescriptors ?? createMongoScalarTypeDescriptors()
37
+ scalarTypeDescriptors: context.scalarTypeDescriptors,
38
+ codecLookup: context.codecLookup
38
39
  });
39
40
  if (!interpreted.ok) return interpreted;
40
41
  return ok(interpreted.value);
@@ -1 +1 @@
1
- {"version":3,"file":"provider.mjs","names":["schema: string"],"sources":["../../src/provider.ts"],"sourcesContent":["import { readFile } from 'node:fs/promises';\nimport type { ContractConfig, ContractSourceContext } from '@prisma-next/config/config-types';\nimport { parsePslDocument } from '@prisma-next/psl-parser';\nimport { ifDefined } from '@prisma-next/utils/defined';\nimport { notOk, ok } from '@prisma-next/utils/result';\nimport { resolve } from 'pathe';\nimport { interpretPslDocumentToMongoContract } from './interpreter';\nimport { createMongoScalarTypeDescriptors } from './scalar-type-descriptors';\n\nexport interface MongoContractOptions {\n readonly output?: string;\n readonly scalarTypeDescriptors?: ReadonlyMap<string, string>;\n}\n\nexport function mongoContract(schemaPath: string, options?: MongoContractOptions): ContractConfig {\n return {\n source: async (_context: ContractSourceContext) => {\n const absoluteSchemaPath = resolve(schemaPath);\n let schema: string;\n try {\n schema = await readFile(absoluteSchemaPath, 'utf-8');\n } catch (error) {\n const message = String(error);\n return notOk({\n summary: `Failed to read Prisma schema at \"${schemaPath}\"`,\n diagnostics: [\n {\n code: 'PSL_SCHEMA_READ_FAILED',\n message,\n sourceId: schemaPath,\n },\n ],\n meta: { schemaPath, absoluteSchemaPath, cause: message },\n });\n }\n\n const document = parsePslDocument({\n schema,\n sourceId: schemaPath,\n });\n\n const interpreted = interpretPslDocumentToMongoContract({\n document,\n scalarTypeDescriptors: options?.scalarTypeDescriptors ?? createMongoScalarTypeDescriptors(),\n });\n if (!interpreted.ok) {\n return interpreted;\n }\n\n return ok(interpreted.value);\n },\n ...ifDefined('output', options?.output),\n };\n}\n"],"mappings":";;;;;;;;AAcA,SAAgB,cAAc,YAAoB,SAAgD;AAChG,QAAO;EACL,QAAQ,OAAO,aAAoC;GACjD,MAAM,qBAAqB,QAAQ,WAAW;GAC9C,IAAIA;AACJ,OAAI;AACF,aAAS,MAAM,SAAS,oBAAoB,QAAQ;YAC7C,OAAO;IACd,MAAM,UAAU,OAAO,MAAM;AAC7B,WAAO,MAAM;KACX,SAAS,oCAAoC,WAAW;KACxD,aAAa,CACX;MACE,MAAM;MACN;MACA,UAAU;MACX,CACF;KACD,MAAM;MAAE;MAAY;MAAoB,OAAO;MAAS;KACzD,CAAC;;GAQJ,MAAM,cAAc,oCAAoC;IACtD,UANe,iBAAiB;KAChC;KACA,UAAU;KACX,CAAC;IAIA,uBAAuB,SAAS,yBAAyB,kCAAkC;IAC5F,CAAC;AACF,OAAI,CAAC,YAAY,GACf,QAAO;AAGT,UAAO,GAAG,YAAY,MAAM;;EAE9B,GAAG,UAAU,UAAU,SAAS,OAAO;EACxC"}
1
+ {"version":3,"file":"provider.mjs","names":["schema: string"],"sources":["../../src/provider.ts"],"sourcesContent":["import { readFile } from 'node:fs/promises';\nimport type { ContractConfig, ContractSourceContext } from '@prisma-next/config/config-types';\nimport { parsePslDocument } from '@prisma-next/psl-parser';\nimport { ifDefined } from '@prisma-next/utils/defined';\nimport { notOk, ok } from '@prisma-next/utils/result';\nimport { resolve } from 'pathe';\nimport { interpretPslDocumentToMongoContract } from './interpreter';\n\nexport interface MongoContractOptions {\n readonly output?: string;\n}\n\nexport function mongoContract(schemaPath: string, options?: MongoContractOptions): ContractConfig {\n return {\n source: async (context: ContractSourceContext) => {\n const absoluteSchemaPath = resolve(schemaPath);\n let schema: string;\n try {\n schema = await readFile(absoluteSchemaPath, 'utf-8');\n } catch (error) {\n const message = String(error);\n return notOk({\n summary: `Failed to read Prisma schema at \"${schemaPath}\"`,\n diagnostics: [\n {\n code: 'PSL_SCHEMA_READ_FAILED',\n message,\n sourceId: schemaPath,\n },\n ],\n meta: { schemaPath, absoluteSchemaPath, cause: message },\n });\n }\n\n const document = parsePslDocument({\n schema,\n sourceId: schemaPath,\n });\n\n const interpreted = interpretPslDocumentToMongoContract({\n document,\n scalarTypeDescriptors: context.scalarTypeDescriptors,\n codecLookup: context.codecLookup,\n });\n if (!interpreted.ok) {\n return interpreted;\n }\n\n return ok(interpreted.value);\n },\n ...ifDefined('output', options?.output),\n };\n}\n"],"mappings":";;;;;;;;AAYA,SAAgB,cAAc,YAAoB,SAAgD;AAChG,QAAO;EACL,QAAQ,OAAO,YAAmC;GAChD,MAAM,qBAAqB,QAAQ,WAAW;GAC9C,IAAIA;AACJ,OAAI;AACF,aAAS,MAAM,SAAS,oBAAoB,QAAQ;YAC7C,OAAO;IACd,MAAM,UAAU,OAAO,MAAM;AAC7B,WAAO,MAAM;KACX,SAAS,oCAAoC,WAAW;KACxD,aAAa,CACX;MACE,MAAM;MACN;MACA,UAAU;MACX,CACF;KACD,MAAM;MAAE;MAAY;MAAoB,OAAO;MAAS;KACzD,CAAC;;GAQJ,MAAM,cAAc,oCAAoC;IACtD,UANe,iBAAiB;KAChC;KACA,UAAU;KACX,CAAC;IAIA,uBAAuB,QAAQ;IAC/B,aAAa,QAAQ;IACtB,CAAC;AACF,OAAI,CAAC,YAAY,GACf,QAAO;AAGT,UAAO,GAAG,YAAY,MAAM;;EAE9B,GAAG,UAAU,UAAU,SAAS,OAAO;EACxC"}
package/dist/index.d.mts CHANGED
@@ -2,16 +2,15 @@ import { Result } from "@prisma-next/utils/result";
2
2
  import { ParsePslDocumentResult } from "@prisma-next/psl-parser";
3
3
  import { ContractSourceDiagnostics } from "@prisma-next/config/config-types";
4
4
  import { Contract } from "@prisma-next/contract/types";
5
+ import { CodecLookup } from "@prisma-next/framework-components/codec";
5
6
 
6
7
  //#region src/interpreter.d.ts
7
8
  interface InterpretPslDocumentToMongoContractInput {
8
9
  readonly document: ParsePslDocumentResult;
9
10
  readonly scalarTypeDescriptors: ReadonlyMap<string, string>;
11
+ readonly codecLookup?: CodecLookup;
10
12
  }
11
13
  declare function interpretPslDocumentToMongoContract(input: InterpretPslDocumentToMongoContractInput): Result<Contract, ContractSourceDiagnostics>;
12
14
  //#endregion
13
- //#region src/scalar-type-descriptors.d.ts
14
- declare function createMongoScalarTypeDescriptors(): ReadonlyMap<string, string>;
15
- //#endregion
16
- export { type InterpretPslDocumentToMongoContractInput, createMongoScalarTypeDescriptors, interpretPslDocumentToMongoContract };
15
+ export { type InterpretPslDocumentToMongoContractInput, interpretPslDocumentToMongoContract };
17
16
  //# sourceMappingURL=index.d.mts.map
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.mts","names":[],"sources":["../src/interpreter.ts","../src/scalar-type-descriptors.ts"],"sourcesContent":[],"mappings":";;;;;;UA0BiB,wCAAA;qBACI;EADJ,SAAA,qBAAA,EAEiB,WAFuB,CAAA,MAAA,EAAA,MACpC,CAAA;AAgmBrB;AACS,iBADO,mCAAA,CACP,KAAA,EAAA,wCAAA,CAAA,EACN,MADM,CACC,QADD,EACW,yBADX,CAAA;;;iBC5nBO,gCAAA,CAAA,GAAoC"}
1
+ {"version":3,"file":"index.d.mts","names":[],"sources":["../src/interpreter.ts"],"sourcesContent":[],"mappings":";;;;;;;UA2BiB,wCAAA;qBACI;EADJ,SAAA,qBAAA,EAEiB,WAFuB,CAAA,MAAA,EAAA,MAAA,CAAA;EACpC,SAAA,WAAA,CAAA,EAEI,WAFJ;;AAEI,iBA2oBT,mCAAA,CA3oBS,KAAA,EA4oBhB,wCA5oBgB,CAAA,EA6oBtB,MA7oBsB,CA6oBf,QA7oBe,EA6oBL,yBA7oBK,CAAA"}
package/dist/index.mjs CHANGED
@@ -1,3 +1,3 @@
1
- import { n as interpretPslDocumentToMongoContract, t as createMongoScalarTypeDescriptors } from "./scalar-type-descriptors-B0Lt0pJI.mjs";
1
+ import { t as interpretPslDocumentToMongoContract } from "./interpreter-CWOjqYYA.mjs";
2
2
 
3
- export { createMongoScalarTypeDescriptors, interpretPslDocumentToMongoContract };
3
+ export { interpretPslDocumentToMongoContract };
@@ -3,16 +3,12 @@ import { notOk, ok } from "@prisma-next/utils/result";
3
3
  import { getPositionalArgument, parseQuotedStringLiteral } from "@prisma-next/psl-parser";
4
4
 
5
5
  //#region src/derive-json-schema.ts
6
- const CODEC_TO_BSON_TYPE = {
7
- "mongo/string@1": "string",
8
- "mongo/int32@1": "int",
9
- "mongo/bool@1": "bool",
10
- "mongo/date@1": "date",
11
- "mongo/objectId@1": "objectId"
12
- };
13
- function fieldToBsonSchema(field, valueObjects) {
6
+ function resolveBsonType(codecId, codecLookup) {
7
+ return (codecLookup?.get(codecId))?.targetTypes[0];
8
+ }
9
+ function fieldToBsonSchema(field, valueObjects, codecLookup) {
14
10
  if (field.type.kind === "scalar") {
15
- const bsonType = CODEC_TO_BSON_TYPE[field.type.codecId];
11
+ const bsonType = resolveBsonType(field.type.codecId, codecLookup);
16
12
  if (!bsonType) return void 0;
17
13
  if ("many" in field && field.many) return {
18
14
  bsonType: "array",
@@ -24,7 +20,7 @@ function fieldToBsonSchema(field, valueObjects) {
24
20
  if (field.type.kind === "valueObject") {
25
21
  const vo = valueObjects?.[field.type.name];
26
22
  if (!vo) return void 0;
27
- const voSchema = deriveObjectSchema(vo.fields, valueObjects);
23
+ const voSchema = deriveObjectSchema(vo.fields, valueObjects, codecLookup);
28
24
  if ("many" in field && field.many) return {
29
25
  bsonType: "array",
30
26
  items: voSchema
@@ -33,11 +29,11 @@ function fieldToBsonSchema(field, valueObjects) {
33
29
  return voSchema;
34
30
  }
35
31
  }
36
- function deriveObjectSchema(fields, valueObjects) {
32
+ function deriveObjectSchema(fields, valueObjects, codecLookup) {
37
33
  const properties = {};
38
34
  const required = [];
39
35
  for (const [fieldName, field] of Object.entries(fields)) {
40
- const schema = fieldToBsonSchema(field, valueObjects);
36
+ const schema = fieldToBsonSchema(field, valueObjects, codecLookup);
41
37
  if (schema) {
42
38
  properties[fieldName] = schema;
43
39
  if (!field.nullable) required.push(fieldName);
@@ -50,9 +46,40 @@ function deriveObjectSchema(fields, valueObjects) {
50
46
  if (required.length > 0) result["required"] = required.sort();
51
47
  return result;
52
48
  }
53
- function deriveJsonSchema(fields, valueObjects) {
49
+ function deriveJsonSchema(fields, valueObjects, codecLookup) {
50
+ return {
51
+ jsonSchema: deriveObjectSchema(fields, valueObjects, codecLookup),
52
+ validationLevel: "strict",
53
+ validationAction: "error"
54
+ };
55
+ }
56
+ function derivePolymorphicJsonSchema(baseFields, discriminatorField, variants, valueObjects, codecLookup) {
57
+ const baseSchema = deriveObjectSchema(baseFields, valueObjects, codecLookup);
58
+ const oneOf = [];
59
+ for (const variant of variants) {
60
+ const variantOnlyFields = {};
61
+ for (const [name, field] of Object.entries(variant.fields)) if (!(name in baseFields)) variantOnlyFields[name] = field;
62
+ const entry = { properties: { [discriminatorField]: { enum: [variant.discriminatorValue] } } };
63
+ const variantProperties = {};
64
+ const variantRequired = [discriminatorField];
65
+ for (const [name, field] of Object.entries(variantOnlyFields)) {
66
+ const schema = fieldToBsonSchema(field, valueObjects, codecLookup);
67
+ if (schema) {
68
+ variantProperties[name] = schema;
69
+ if (!field.nullable) variantRequired.push(name);
70
+ }
71
+ }
72
+ if (Object.keys(variantProperties).length > 0) entry["properties"] = {
73
+ ...entry["properties"],
74
+ ...variantProperties
75
+ };
76
+ entry["required"] = variantRequired.sort();
77
+ oneOf.push(entry);
78
+ }
79
+ const jsonSchema = { ...baseSchema };
80
+ if (oneOf.length > 0) jsonSchema["oneOf"] = oneOf;
54
81
  return {
55
- jsonSchema: deriveObjectSchema(fields, valueObjects),
82
+ jsonSchema,
56
83
  validationLevel: "strict",
57
84
  validationAction: "error"
58
85
  };
@@ -230,6 +257,7 @@ function resolvePolymorphism(input) {
230
257
  const { discriminatorDeclarations, baseDeclarations, modelNames, sourceId, document } = input;
231
258
  let patched = input.models;
232
259
  let roots = input.roots;
260
+ let collections = input.collections;
233
261
  const diagnostics = [];
234
262
  for (const [modelName, decl] of discriminatorDeclarations) {
235
263
  if (baseDeclarations.has(modelName)) {
@@ -243,7 +271,9 @@ function resolvePolymorphism(input) {
243
271
  }
244
272
  const model = patched[modelName];
245
273
  if (!model) continue;
246
- if (!Object.hasOwn(model.fields, decl.fieldName)) {
274
+ const pslModel = document.ast.models.find((m) => m.name === modelName);
275
+ const mappedDiscriminatorField = pslModel ? resolveFieldMappings(pslModel).pslNameToMapped.get(decl.fieldName) ?? decl.fieldName : decl.fieldName;
276
+ if (!Object.hasOwn(model.fields, mappedDiscriminatorField)) {
247
277
  diagnostics.push({
248
278
  code: "PSL_DISCRIMINATOR_FIELD_NOT_FOUND",
249
279
  message: `Discriminator field "${decl.fieldName}" is not a field on model "${modelName}"`,
@@ -270,7 +300,7 @@ function resolvePolymorphism(input) {
270
300
  ...patched,
271
301
  [modelName]: {
272
302
  ...model,
273
- discriminator: { field: decl.fieldName },
303
+ discriminator: { field: mappedDiscriminatorField },
274
304
  variants
275
305
  }
276
306
  };
@@ -323,10 +353,37 @@ function resolvePolymorphism(input) {
323
353
  [variantCollectionName]: baseDecl.baseName
324
354
  };
325
355
  else roots = Object.fromEntries(Object.entries(roots).filter(([key]) => key !== variantCollectionName));
356
+ const variantIndexes = collections[variantCollectionName]?.["indexes"] ?? [];
357
+ const baseColl = collections[baseCollection];
358
+ if (variantCollectionName !== baseCollection) {
359
+ const filtered = Object.fromEntries(Object.entries(collections).filter(([key]) => key !== variantCollectionName));
360
+ if (variantIndexes.length > 0 && baseColl) {
361
+ const baseIndexes = baseColl["indexes"] ?? [];
362
+ collections = {
363
+ ...filtered,
364
+ [baseCollection]: {
365
+ ...baseColl,
366
+ indexes: [...baseIndexes, ...variantIndexes]
367
+ }
368
+ };
369
+ } else collections = filtered;
370
+ } else if (variantIndexes.length > 0 && baseColl) {
371
+ const baseIndexes = baseColl["indexes"] ?? [];
372
+ const mergedIndexes = [...baseIndexes];
373
+ for (const idx of variantIndexes) if (!baseIndexes.some((existing) => JSON.stringify(existing) === JSON.stringify(idx))) mergedIndexes.push(idx);
374
+ if (mergedIndexes.length > baseIndexes.length) collections = {
375
+ ...collections,
376
+ [baseCollection]: {
377
+ ...baseColl,
378
+ indexes: mergedIndexes
379
+ }
380
+ };
381
+ }
326
382
  }
327
383
  return {
328
384
  models: patched,
329
385
  roots,
386
+ collections,
330
387
  diagnostics
331
388
  };
332
389
  }
@@ -603,7 +660,7 @@ function resolveNonRelationField(field, ownerName, compositeTypeNames, scalarTyp
603
660
  } : result;
604
661
  }
605
662
  function interpretPslDocumentToMongoContract(input) {
606
- const { document, scalarTypeDescriptors } = input;
663
+ const { document, scalarTypeDescriptors, codecLookup } = input;
607
664
  const sourceId = document.ast.sourceId;
608
665
  const diagnostics = [];
609
666
  const modelNames = new Set(document.ast.models.map((m) => m.name));
@@ -673,7 +730,9 @@ function interpretPslDocumentToMongoContract(input) {
673
730
  storage: { collection: collectionName }
674
731
  };
675
732
  const modelIndexes = collectIndexes(pslModel, fieldMappings, modelNames, sourceId, diagnostics);
676
- collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};
733
+ const existingColl = collections[collectionName];
734
+ if (existingColl && modelIndexes.length > 0) collections[collectionName] = { indexes: [...existingColl["indexes"] ?? [], ...modelIndexes] };
735
+ else if (!existingColl) collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};
677
736
  roots[collectionName] = pslModel.name;
678
737
  }
679
738
  const valueObjects = {};
@@ -686,10 +745,6 @@ function interpretPslDocumentToMongoContract(input) {
686
745
  }
687
746
  valueObjects[compositeType.name] = { fields };
688
747
  }
689
- for (const [, modelEntry] of Object.entries(models)) {
690
- const coll = collections[modelEntry.storage.collection];
691
- if (coll) coll["validator"] = deriveJsonSchema(modelEntry.fields, valueObjects);
692
- }
693
748
  const fkRelationsByPair = /* @__PURE__ */ new Map();
694
749
  for (const fk of allFkRelations) {
695
750
  const key = fkRelationPairKey(fk.declaringModel, fk.targetModel);
@@ -736,6 +791,7 @@ function interpretPslDocumentToMongoContract(input) {
736
791
  const polyResult = resolvePolymorphism({
737
792
  models,
738
793
  roots,
794
+ collections,
739
795
  document,
740
796
  discriminatorDeclarations,
741
797
  baseDeclarations,
@@ -746,9 +802,23 @@ function interpretPslDocumentToMongoContract(input) {
746
802
  summary: "PSL to Mongo contract interpretation failed",
747
803
  diagnostics: [...diagnostics, ...polyResult.diagnostics]
748
804
  });
805
+ const resolvedModels = polyResult.models;
806
+ const resolvedCollections = polyResult.collections;
807
+ for (const [, modelEntry] of Object.entries(resolvedModels)) {
808
+ if (modelEntry.base) continue;
809
+ const coll = resolvedCollections[modelEntry.storage.collection];
810
+ if (!coll) continue;
811
+ if (modelEntry.discriminator && modelEntry.variants) {
812
+ const variantEntries = Object.entries(modelEntry.variants).map(([variantName, { value }]) => ({
813
+ discriminatorValue: value,
814
+ fields: resolvedModels[variantName]?.fields ?? {}
815
+ }));
816
+ coll["validator"] = derivePolymorphicJsonSchema(modelEntry.fields, modelEntry.discriminator.field, variantEntries, valueObjects, codecLookup);
817
+ } else coll["validator"] = deriveJsonSchema(modelEntry.fields, valueObjects, codecLookup);
818
+ }
749
819
  const target = "mongo";
750
820
  const targetFamily = "mongo";
751
- const storageWithoutHash = { collections };
821
+ const storageWithoutHash = { collections: resolvedCollections };
752
822
  const storageHash = computeStorageHash({
753
823
  target,
754
824
  targetFamily,
@@ -777,17 +847,5 @@ function interpretPslDocumentToMongoContract(input) {
777
847
  }
778
848
 
779
849
  //#endregion
780
- //#region src/scalar-type-descriptors.ts
781
- function createMongoScalarTypeDescriptors() {
782
- return new Map([
783
- ["String", "mongo/string@1"],
784
- ["Int", "mongo/int32@1"],
785
- ["Boolean", "mongo/bool@1"],
786
- ["DateTime", "mongo/date@1"],
787
- ["ObjectId", "mongo/objectId@1"]
788
- ]);
789
- }
790
-
791
- //#endregion
792
- export { interpretPslDocumentToMongoContract as n, createMongoScalarTypeDescriptors as t };
793
- //# sourceMappingURL=scalar-type-descriptors-B0Lt0pJI.mjs.map
850
+ export { interpretPslDocumentToMongoContract as t };
851
+ //# sourceMappingURL=interpreter-CWOjqYYA.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"interpreter-CWOjqYYA.mjs","names":["properties: Record<string, unknown>","required: string[]","result: Record<string, unknown>","oneOf: Record<string, unknown>[]","variantOnlyFields: Record<string, ContractField>","entry: Record<string, unknown>","variantProperties: Record<string, unknown>","variantRequired: string[]","parts: string[]","relationName: string | undefined","fieldsArg: PslAttributeArgument | undefined","referencesArg: PslAttributeArgument | undefined","diagnostics: ContractSourceDiagnostic[]","variants: Record<string, { readonly value: string }>","collation: Record<string, unknown>","result: Record<string, 0 | 1>","indexes: MongoStorageIndex[]","defaultDirection: MongoIndexKeyDirection","weights: Record<string, number> | undefined","index: MongoStorageIndex","result: ContractField","result","models: Record<string, MongoModelEntry>","collections: Record<string, Record<string, unknown>>","roots: Record<string, string>","allFkRelations: FkRelation[]","backrelationCandidates: BackrelationCandidate[]","fields: Record<string, ContractField>","relations: Record<string, ContractReferenceRelation>","valueObjects: Record<string, ContractValueObject>","capabilities: Record<string, Record<string, boolean>>"],"sources":["../src/derive-json-schema.ts","../src/psl-helpers.ts","../src/interpreter.ts"],"sourcesContent":["import type { ContractField, ContractValueObject } from '@prisma-next/contract/types';\nimport type { CodecLookup } from '@prisma-next/framework-components/codec';\nimport type { MongoStorageValidator } from '@prisma-next/mongo-contract';\n\nfunction resolveBsonType(\n codecId: string,\n codecLookup: CodecLookup | undefined,\n): string | undefined {\n const codec = codecLookup?.get(codecId);\n return codec?.targetTypes[0];\n}\n\nfunction fieldToBsonSchema(\n field: ContractField,\n valueObjects: Record<string, ContractValueObject> | undefined,\n codecLookup: CodecLookup | undefined,\n): Record<string, unknown> | undefined {\n if (field.type.kind === 'scalar') {\n const bsonType = resolveBsonType(field.type.codecId, codecLookup);\n if (!bsonType) return undefined;\n\n if ('many' in field && field.many) {\n return { bsonType: 'array', items: { bsonType } };\n }\n\n if (field.nullable) {\n return { bsonType: ['null', bsonType] };\n }\n\n return { bsonType };\n }\n\n if (field.type.kind === 'valueObject') {\n const vo = valueObjects?.[field.type.name];\n if (!vo) return undefined;\n const voSchema = deriveObjectSchema(vo.fields, valueObjects, codecLookup);\n if ('many' in field && field.many) {\n return { bsonType: 'array', items: voSchema };\n }\n if (field.nullable) {\n return { oneOf: [{ bsonType: 'null' }, voSchema] };\n }\n return voSchema;\n }\n\n return undefined;\n}\n\nfunction deriveObjectSchema(\n fields: Record<string, ContractField>,\n valueObjects: Record<string, ContractValueObject> | undefined,\n codecLookup: CodecLookup | undefined,\n): Record<string, unknown> {\n const properties: Record<string, unknown> = {};\n const required: string[] = [];\n\n for (const [fieldName, field] of Object.entries(fields)) {\n const schema = fieldToBsonSchema(field, valueObjects, codecLookup);\n if (schema) {\n properties[fieldName] = schema;\n if (!field.nullable) {\n required.push(fieldName);\n }\n }\n }\n\n const result: Record<string, unknown> = {\n bsonType: 'object',\n properties,\n };\n if (required.length > 0) {\n result['required'] = required.sort();\n }\n return result;\n}\n\nexport function deriveJsonSchema(\n fields: Record<string, ContractField>,\n valueObjects?: Record<string, ContractValueObject>,\n codecLookup?: CodecLookup,\n): MongoStorageValidator {\n return {\n jsonSchema: deriveObjectSchema(fields, valueObjects, codecLookup),\n validationLevel: 'strict',\n validationAction: 'error',\n };\n}\n\nexport interface PolymorphicVariant {\n readonly discriminatorValue: string;\n readonly fields: Record<string, ContractField>;\n}\n\nexport function derivePolymorphicJsonSchema(\n baseFields: Record<string, ContractField>,\n discriminatorField: string,\n variants: readonly PolymorphicVariant[],\n valueObjects?: Record<string, ContractValueObject>,\n codecLookup?: CodecLookup,\n): MongoStorageValidator {\n const baseSchema = deriveObjectSchema(baseFields, valueObjects, codecLookup);\n\n const oneOf: Record<string, unknown>[] = [];\n for (const variant of variants) {\n const variantOnlyFields: Record<string, ContractField> = {};\n for (const [name, field] of Object.entries(variant.fields)) {\n if (!(name in baseFields)) {\n variantOnlyFields[name] = field;\n }\n }\n\n const entry: Record<string, unknown> = {\n properties: {\n [discriminatorField]: { enum: [variant.discriminatorValue] },\n },\n };\n\n const variantProperties: Record<string, unknown> = {};\n const variantRequired: string[] = [discriminatorField];\n for (const [name, field] of Object.entries(variantOnlyFields)) {\n const schema = fieldToBsonSchema(field, valueObjects, codecLookup);\n if (schema) {\n variantProperties[name] = schema;\n if (!field.nullable) {\n variantRequired.push(name);\n }\n }\n }\n\n if (Object.keys(variantProperties).length > 0) {\n (entry['properties'] as Record<string, unknown>) = {\n ...(entry['properties'] as Record<string, unknown>),\n ...variantProperties,\n };\n }\n entry['required'] = variantRequired.sort();\n\n oneOf.push(entry);\n }\n\n const jsonSchema = { ...baseSchema };\n if (oneOf.length > 0) {\n jsonSchema['oneOf'] = oneOf;\n }\n\n return {\n jsonSchema,\n validationLevel: 'strict',\n validationAction: 'error',\n };\n}\n","import type { PslAttribute, PslAttributeArgument } from '@prisma-next/psl-parser';\nimport { getPositionalArgument, parseQuotedStringLiteral } from '@prisma-next/psl-parser';\n\nexport { getPositionalArgument, parseQuotedStringLiteral };\n\nexport function getNamedArgument(attr: PslAttribute, name: string): string | undefined {\n const arg = attr.args.find((a) => a.kind === 'named' && a.name === name);\n return arg?.value;\n}\n\nexport function parseFieldList(value: string): readonly string[] {\n const inner = value.replace(/^\\[/, '').replace(/\\]$/, '').trim();\n if (inner.length === 0) return [];\n return splitTopLevel(inner).map((s) => s.trim());\n}\n\nexport interface ParsedIndexField {\n readonly name: string;\n readonly isWildcard: boolean;\n readonly direction?: number;\n}\n\nexport function parseIndexFieldList(value: string): readonly ParsedIndexField[] {\n const segments = parseFieldList(value);\n return segments.map(parseIndexFieldSegment);\n}\n\nfunction parseIndexFieldSegment(segment: string): ParsedIndexField {\n const wildcardMatch = segment.match(/^wildcard\\(\\s*(.*?)\\s*\\)$/);\n if (wildcardMatch) {\n const scope = wildcardMatch[1] ?? '';\n return {\n name: scope.length > 0 ? `${scope}.$**` : '$**',\n isWildcard: true,\n };\n }\n\n const modifierMatch = segment.match(/^(\\w+)\\(\\s*sort:\\s*(\\w+)\\s*\\)$/);\n if (modifierMatch) {\n const fieldName = modifierMatch[1] ?? segment;\n const sortValue = modifierMatch[2];\n return {\n name: fieldName,\n isWildcard: false,\n direction: sortValue === 'Desc' ? -1 : 1,\n };\n }\n\n return { name: segment, isWildcard: false };\n}\n\nfunction splitTopLevel(input: string): string[] {\n const parts: string[] = [];\n let depth = 0;\n let start = 0;\n for (let i = 0; i < input.length; i++) {\n const ch = input[i];\n if (ch === '(' || ch === '[' || ch === '{') depth++;\n else if (ch === ')' || ch === ']' || ch === '}') depth = Math.max(0, depth - 1);\n else if (ch === ',' && depth === 0) {\n parts.push(input.slice(start, i));\n start = i + 1;\n }\n }\n parts.push(input.slice(start));\n return parts;\n}\n\nexport function lowerFirst(value: string): string {\n if (value.length === 0) return value;\n return value[0]?.toLowerCase() + value.slice(1);\n}\n\nexport function getAttribute(\n attributes: readonly PslAttribute[],\n name: string,\n): PslAttribute | undefined {\n return attributes.find((attr) => attr.name === name);\n}\n\nexport function getMapName(attributes: readonly PslAttribute[]): string | undefined {\n const mapAttr = getAttribute(attributes, 'map');\n if (!mapAttr) return undefined;\n const arg = mapAttr.args[0];\n if (!arg) return undefined;\n return stripQuotes(arg.value);\n}\n\nexport interface ParsedRelationAttribute {\n readonly relationName?: string;\n readonly fields?: readonly string[];\n readonly references?: readonly string[];\n}\n\nexport function parseRelationAttribute(\n attributes: readonly PslAttribute[],\n): ParsedRelationAttribute | undefined {\n const relationAttr = getAttribute(attributes, 'relation');\n if (!relationAttr) return undefined;\n\n let relationName: string | undefined;\n let fieldsArg: PslAttributeArgument | undefined;\n let referencesArg: PslAttributeArgument | undefined;\n\n for (const arg of relationAttr.args) {\n if (arg.kind === 'positional') {\n relationName = stripQuotes(arg.value);\n } else if (arg.name === 'name') {\n relationName = stripQuotes(arg.value);\n } else if (arg.name === 'fields') {\n fieldsArg = arg;\n } else if (arg.name === 'references') {\n referencesArg = arg;\n }\n }\n\n const fields = fieldsArg ? parseFieldList(fieldsArg.value) : undefined;\n const references = referencesArg ? parseFieldList(referencesArg.value) : undefined;\n\n return {\n ...(relationName !== undefined ? { relationName } : {}),\n ...(fields !== undefined ? { fields } : {}),\n ...(references !== undefined ? { references } : {}),\n };\n}\n\nfunction stripQuotes(value: string): string {\n if (value.startsWith('\"') && value.endsWith('\"')) {\n return value.slice(1, -1);\n }\n return value;\n}\n","import type {\n ContractSourceDiagnostic,\n ContractSourceDiagnostics,\n} from '@prisma-next/config/config-types';\nimport { computeProfileHash, computeStorageHash } from '@prisma-next/contract/hashing';\nimport type {\n Contract,\n ContractField,\n ContractReferenceRelation,\n ContractValueObject,\n} from '@prisma-next/contract/types';\nimport type { CodecLookup } from '@prisma-next/framework-components/codec';\nimport type { MongoIndexKeyDirection, MongoStorageIndex } from '@prisma-next/mongo-contract';\nimport type { ParsePslDocumentResult, PslField, PslModel } from '@prisma-next/psl-parser';\nimport { notOk, ok, type Result } from '@prisma-next/utils/result';\nimport { deriveJsonSchema, derivePolymorphicJsonSchema } from './derive-json-schema';\nimport {\n getAttribute,\n getMapName,\n getNamedArgument,\n getPositionalArgument,\n lowerFirst,\n parseIndexFieldList,\n parseQuotedStringLiteral,\n parseRelationAttribute,\n} from './psl-helpers';\n\nexport interface InterpretPslDocumentToMongoContractInput {\n readonly document: ParsePslDocumentResult;\n readonly scalarTypeDescriptors: ReadonlyMap<string, string>;\n readonly codecLookup?: CodecLookup;\n}\n\ninterface FieldMappings {\n readonly pslNameToMapped: Map<string, string>;\n}\n\ninterface FkRelation {\n readonly declaringModel: string;\n readonly fieldName: string;\n readonly targetModel: string;\n readonly relationName?: string;\n readonly localFields: readonly string[];\n readonly targetFields: readonly string[];\n}\n\nfunction fkRelationPairKey(declaringModel: string, targetModel: string): string {\n return `${declaringModel}::${targetModel}`;\n}\n\nfunction resolveFieldMappings(model: PslModel): FieldMappings {\n const pslNameToMapped = new Map<string, string>();\n for (const field of model.fields) {\n const mapped = getMapName(field.attributes) ?? field.name;\n pslNameToMapped.set(field.name, mapped);\n }\n return { pslNameToMapped };\n}\n\nfunction resolveCollectionName(model: PslModel): string {\n return getMapName(model.attributes) ?? lowerFirst(model.name);\n}\n\ninterface MongoModelEntry {\n readonly fields: Record<string, ContractField>;\n readonly relations: Record<string, ContractReferenceRelation>;\n readonly storage: { readonly collection: string };\n readonly discriminator?: { readonly field: string };\n readonly variants?: Record<string, { readonly value: string }>;\n readonly base?: string;\n}\n\ntype DiscriminatorDeclaration = { readonly fieldName: string; readonly span: PslModel['span'] };\ntype BaseDeclaration = {\n readonly baseName: string;\n readonly value: string;\n readonly collectionName: string;\n readonly span: PslModel['span'];\n};\n\nfunction collectPolymorphismDeclarations(\n document: ParsePslDocumentResult,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): {\n discriminatorDeclarations: Map<string, DiscriminatorDeclaration>;\n baseDeclarations: Map<string, BaseDeclaration>;\n} {\n const discriminatorDeclarations = new Map<string, DiscriminatorDeclaration>();\n const baseDeclarations = new Map<string, BaseDeclaration>();\n\n for (const pslModel of document.ast.models) {\n for (const attr of pslModel.attributes) {\n if (attr.name === 'discriminator') {\n const fieldName = getPositionalArgument(attr);\n if (!fieldName) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@discriminator requires a field name argument`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const discField = pslModel.fields.find((f) => f.name === fieldName);\n if (discField && discField.typeName !== 'String') {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Discriminator field \"${fieldName}\" on model \"${pslModel.name}\" must be of type String, but is \"${discField.typeName}\"`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n discriminatorDeclarations.set(pslModel.name, { fieldName, span: attr.span });\n }\n if (attr.name === 'base') {\n const baseName = getPositionalArgument(attr, 0);\n const rawValue = getPositionalArgument(attr, 1);\n if (!baseName || !rawValue) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@base requires two arguments: base model name and discriminator value`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const value = parseQuotedStringLiteral(rawValue);\n if (value === undefined) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@base discriminator value must be a quoted string literal`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const collectionName = resolveCollectionName(pslModel);\n baseDeclarations.set(pslModel.name, { baseName, value, collectionName, span: attr.span });\n }\n }\n }\n\n return { discriminatorDeclarations, baseDeclarations };\n}\n\nfunction resolvePolymorphism(input: {\n models: Record<string, MongoModelEntry>;\n roots: Record<string, string>;\n collections: Record<string, Record<string, unknown>>;\n document: ParsePslDocumentResult;\n discriminatorDeclarations: Map<string, DiscriminatorDeclaration>;\n baseDeclarations: Map<string, BaseDeclaration>;\n modelNames: ReadonlySet<string>;\n sourceId: string;\n}): {\n models: Record<string, MongoModelEntry>;\n roots: Record<string, string>;\n collections: Record<string, Record<string, unknown>>;\n diagnostics: ContractSourceDiagnostic[];\n} {\n const { discriminatorDeclarations, baseDeclarations, modelNames, sourceId, document } = input;\n let patched = input.models;\n let roots = input.roots;\n let collections = input.collections;\n const diagnostics: ContractSourceDiagnostic[] = [];\n\n for (const [modelName, decl] of discriminatorDeclarations) {\n if (baseDeclarations.has(modelName)) {\n diagnostics.push({\n code: 'PSL_DISCRIMINATOR_AND_BASE',\n message: `Model \"${modelName}\" cannot have both @@discriminator and @@base`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n const model = patched[modelName];\n if (!model) continue;\n\n const pslModel = document.ast.models.find((m) => m.name === modelName);\n const mappedDiscriminatorField = pslModel\n ? (resolveFieldMappings(pslModel).pslNameToMapped.get(decl.fieldName) ?? decl.fieldName)\n : decl.fieldName;\n\n if (!Object.hasOwn(model.fields, mappedDiscriminatorField)) {\n diagnostics.push({\n code: 'PSL_DISCRIMINATOR_FIELD_NOT_FOUND',\n message: `Discriminator field \"${decl.fieldName}\" is not a field on model \"${modelName}\"`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n const variants: Record<string, { readonly value: string }> = {};\n for (const [variantName, baseDecl] of baseDeclarations) {\n if (baseDecl.baseName !== modelName) continue;\n variants[variantName] = { value: baseDecl.value };\n }\n\n if (Object.keys(variants).length === 0) {\n diagnostics.push({\n code: 'PSL_ORPHANED_DISCRIMINATOR',\n message: `Model \"${modelName}\" has @@discriminator but no variant models declare @@base(${modelName}, ...)`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n patched = {\n ...patched,\n [modelName]: { ...model, discriminator: { field: mappedDiscriminatorField }, variants },\n };\n }\n\n for (const [variantName, baseDecl] of baseDeclarations) {\n if (!modelNames.has(baseDecl.baseName)) {\n diagnostics.push({\n code: 'PSL_BASE_TARGET_NOT_FOUND',\n message: `Model \"${variantName}\" @@base references non-existent model \"${baseDecl.baseName}\"`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n if (!discriminatorDeclarations.has(baseDecl.baseName)) {\n diagnostics.push({\n code: 'PSL_ORPHANED_BASE',\n message: `Model \"${variantName}\" declares @@base(${baseDecl.baseName}, ...) but \"${baseDecl.baseName}\" has no @@discriminator`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n if (discriminatorDeclarations.has(variantName)) {\n continue;\n }\n\n const baseModel = patched[baseDecl.baseName];\n const variantPslModel = document.ast.models.find((m) => m.name === variantName);\n if (!variantPslModel) continue;\n const hasExplicitMap = getMapName(variantPslModel.attributes) !== undefined;\n\n if (hasExplicitMap && baseModel && baseDecl.collectionName !== baseModel.storage.collection) {\n diagnostics.push({\n code: 'PSL_MONGO_VARIANT_SEPARATE_COLLECTION',\n message: `Mongo variant \"${variantName}\" cannot use a different collection than its base \"${baseDecl.baseName}\". Mongo only supports single-collection polymorphism.`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n const baseCollection = baseModel?.storage.collection ?? baseDecl.collectionName;\n const variantModel = patched[variantName];\n if (variantModel) {\n patched = {\n ...patched,\n [variantName]: {\n ...variantModel,\n base: baseDecl.baseName,\n storage: { collection: baseCollection },\n },\n };\n }\n\n const variantCollectionName = resolveCollectionName(variantPslModel);\n if (roots[variantCollectionName] === variantName) {\n if (variantCollectionName === baseCollection && baseModel) {\n roots = { ...roots, [variantCollectionName]: baseDecl.baseName };\n } else {\n roots = Object.fromEntries(\n Object.entries(roots).filter(([key]) => key !== variantCollectionName),\n );\n }\n }\n\n const variantColl = collections[variantCollectionName];\n const variantIndexes = (variantColl?.['indexes'] ?? []) as MongoStorageIndex[];\n const baseColl = collections[baseCollection];\n\n if (variantCollectionName !== baseCollection) {\n const filtered = Object.fromEntries(\n Object.entries(collections).filter(([key]) => key !== variantCollectionName),\n );\n if (variantIndexes.length > 0 && baseColl) {\n const baseIndexes = (baseColl['indexes'] ?? []) as MongoStorageIndex[];\n collections = {\n ...filtered,\n [baseCollection]: { ...baseColl, indexes: [...baseIndexes, ...variantIndexes] },\n };\n } else {\n collections = filtered;\n }\n } else if (variantIndexes.length > 0 && baseColl) {\n const baseIndexes = (baseColl['indexes'] ?? []) as MongoStorageIndex[];\n const mergedIndexes = [...baseIndexes];\n for (const idx of variantIndexes) {\n const isDuplicate = baseIndexes.some(\n (existing) => JSON.stringify(existing) === JSON.stringify(idx),\n );\n if (!isDuplicate) {\n mergedIndexes.push(idx);\n }\n }\n if (mergedIndexes.length > baseIndexes.length) {\n collections = {\n ...collections,\n [baseCollection]: { ...baseColl, indexes: mergedIndexes },\n };\n }\n }\n }\n\n return { models: patched, roots, collections, diagnostics };\n}\n\nfunction parseIndexDirection(raw: string | undefined): MongoIndexKeyDirection {\n if (!raw) return 1;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n const num = Number(stripped);\n if (num === 1 || num === -1) return num;\n if (['text', '2dsphere', '2d', 'hashed'].includes(stripped))\n return stripped as MongoIndexKeyDirection;\n return 1;\n}\n\nfunction parseNumericArg(raw: string | undefined): number | undefined {\n if (!raw) return undefined;\n const n = Number(raw);\n return Number.isFinite(n) ? n : undefined;\n}\n\nfunction parseBooleanArg(raw: string | undefined): boolean | undefined {\n if (raw === 'true') return true;\n if (raw === 'false') return false;\n return undefined;\n}\n\nfunction parseJsonArg(raw: string | undefined): Record<string, unknown> | undefined {\n if (!raw) return undefined;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '').replace(/\\\\\"/g, '\"');\n try {\n const parsed = JSON.parse(stripped);\n if (typeof parsed === 'object' && parsed !== null && !Array.isArray(parsed)) {\n return parsed as Record<string, unknown>;\n }\n } catch {\n // not valid JSON\n }\n return undefined;\n}\n\nfunction parseCollation(\n attr: import('@prisma-next/psl-parser').PslAttribute,\n): Record<string, unknown> | null | undefined {\n const locale = stripQuotesHelper(getNamedArgument(attr, 'collationLocale'));\n if (!locale) {\n const hasAnyCollationArg =\n getNamedArgument(attr, 'collationStrength') != null ||\n getNamedArgument(attr, 'collationCaseLevel') != null ||\n getNamedArgument(attr, 'collationCaseFirst') != null ||\n getNamedArgument(attr, 'collationNumericOrdering') != null ||\n getNamedArgument(attr, 'collationAlternate') != null ||\n getNamedArgument(attr, 'collationMaxVariable') != null ||\n getNamedArgument(attr, 'collationBackwards') != null ||\n getNamedArgument(attr, 'collationNormalization') != null;\n return hasAnyCollationArg ? null : undefined;\n }\n\n const collation: Record<string, unknown> = { locale };\n const strength = parseNumericArg(getNamedArgument(attr, 'collationStrength'));\n if (strength != null) collation['strength'] = strength;\n const caseLevel = parseBooleanArg(getNamedArgument(attr, 'collationCaseLevel'));\n if (caseLevel != null) collation['caseLevel'] = caseLevel;\n const caseFirst = stripQuotesHelper(getNamedArgument(attr, 'collationCaseFirst'));\n if (caseFirst != null) collation['caseFirst'] = caseFirst;\n const numericOrdering = parseBooleanArg(getNamedArgument(attr, 'collationNumericOrdering'));\n if (numericOrdering != null) collation['numericOrdering'] = numericOrdering;\n const alternate = stripQuotesHelper(getNamedArgument(attr, 'collationAlternate'));\n if (alternate != null) collation['alternate'] = alternate;\n const maxVariable = stripQuotesHelper(getNamedArgument(attr, 'collationMaxVariable'));\n if (maxVariable != null) collation['maxVariable'] = maxVariable;\n const backwards = parseBooleanArg(getNamedArgument(attr, 'collationBackwards'));\n if (backwards != null) collation['backwards'] = backwards;\n const normalization = parseBooleanArg(getNamedArgument(attr, 'collationNormalization'));\n if (normalization != null) collation['normalization'] = normalization;\n return collation;\n}\n\nfunction stripQuotesHelper(raw: string | undefined): string | undefined {\n if (!raw) return undefined;\n return raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n}\n\nfunction parseProjectionList(\n raw: string | undefined,\n value: 0 | 1,\n): Record<string, 0 | 1> | undefined {\n if (!raw) return undefined;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n const inner = stripped.replace(/^\\[/, '').replace(/\\]$/, '').trim();\n if (inner.length === 0) return undefined;\n const fields = inner\n .split(',')\n .map((s) => s.trim())\n .filter((s) => s.length > 0);\n const result: Record<string, 0 | 1> = {};\n for (const f of fields) {\n result[f] = value;\n }\n return result;\n}\n\nfunction collectIndexes(\n pslModel: PslModel,\n fieldMappings: FieldMappings,\n modelNames: ReadonlySet<string>,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): MongoStorageIndex[] {\n const indexes: MongoStorageIndex[] = [];\n let textIndexCount = 0;\n\n for (const field of pslModel.fields) {\n if (modelNames.has(field.typeName)) continue;\n const uniqueAttr = getAttribute(field.attributes, 'unique');\n if (!uniqueAttr) continue;\n const mappedName = fieldMappings.pslNameToMapped.get(field.name) ?? field.name;\n indexes.push({\n keys: [{ field: mappedName, direction: 1 }],\n unique: true,\n });\n }\n\n for (const attr of pslModel.attributes) {\n const isIndex = attr.name === 'index';\n const isUnique = attr.name === 'unique';\n const isTextIndex = attr.name === 'textIndex';\n if (!isIndex && !isUnique && !isTextIndex) continue;\n\n const fieldsArg = getPositionalArgument(attr, 0);\n if (!fieldsArg) continue;\n const parsedFields = parseIndexFieldList(fieldsArg);\n if (parsedFields.length === 0) continue;\n\n const hasWildcard = parsedFields.some((f) => f.isWildcard);\n const wildcardCount = parsedFields.filter((f) => f.isWildcard).length;\n\n if (wildcardCount > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'An index can contain at most one wildcard() field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (isUnique && hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Unique indexes cannot use wildcard() fields',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (isTextIndex) {\n textIndexCount++;\n if (textIndexCount > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: `Only one @@textIndex is allowed per collection (model \"${pslModel.name}\")`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message:\n 'wildcard() fields cannot be combined with type: hashed/2dsphere/2d or @@textIndex',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n }\n\n const typeArg = getNamedArgument(attr, 'type');\n const defaultDirection: MongoIndexKeyDirection = isTextIndex\n ? 'text'\n : parseIndexDirection(typeArg);\n\n if (\n hasWildcard &&\n typeof defaultDirection === 'string' &&\n ['hashed', '2dsphere', '2d'].includes(defaultDirection)\n ) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: `wildcard() fields cannot be combined with type: ${defaultDirection}`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (defaultDirection === 'hashed' && parsedFields.length > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Hashed indexes must have exactly one field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const keys = parsedFields.map((pf) => {\n const mappedName = pf.isWildcard\n ? pf.name.replace(/^(.+)\\.\\$\\*\\*$/, (_, prefix: string) => {\n const mapped = fieldMappings.pslNameToMapped.get(prefix);\n return mapped ? `${mapped}.$**` : `${prefix}.$**`;\n })\n : (fieldMappings.pslNameToMapped.get(pf.name) ?? pf.name);\n const direction: MongoIndexKeyDirection =\n pf.direction != null ? (pf.direction as MongoIndexKeyDirection) : defaultDirection;\n return { field: mappedName, direction };\n });\n\n const unique = isUnique ? true : undefined;\n const sparse = isTextIndex ? undefined : parseBooleanArg(getNamedArgument(attr, 'sparse'));\n const expireAfterSeconds = isTextIndex\n ? undefined\n : parseNumericArg(getNamedArgument(attr, 'expireAfterSeconds'));\n\n if (hasWildcard && expireAfterSeconds != null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'expireAfterSeconds cannot be combined with wildcard() fields',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const partialFilterExpression = parseJsonArg(getNamedArgument(attr, 'filter'));\n\n const includeArg = getNamedArgument(attr, 'include');\n const excludeArg = getNamedArgument(attr, 'exclude');\n\n if (includeArg != null && excludeArg != null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Cannot specify both include and exclude on the same index',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if ((includeArg != null || excludeArg != null) && !hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message:\n 'include/exclude options are only valid when the index contains a wildcard() field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const wildcardProjection =\n includeArg != null\n ? parseProjectionList(includeArg, 1)\n : excludeArg != null\n ? parseProjectionList(excludeArg, 0)\n : undefined;\n\n const collation = parseCollation(attr);\n if (collation === null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'collationLocale is required when using collation options',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const rawWeights = parseJsonArg(getNamedArgument(attr, 'weights'));\n let weights: Record<string, number> | undefined;\n if (rawWeights) {\n weights = {};\n for (const [k, v] of Object.entries(rawWeights)) {\n if (typeof v === 'number') weights[k] = v;\n }\n }\n\n const rawDefaultLang = isTextIndex\n ? getNamedArgument(attr, 'language')\n : getNamedArgument(attr, 'default_language');\n const default_language = stripQuotesHelper(rawDefaultLang);\n\n const rawLangOverride = getNamedArgument(attr, 'languageOverride');\n const language_override = stripQuotesHelper(rawLangOverride);\n\n const index: MongoStorageIndex = {\n keys,\n ...(unique != null && { unique }),\n ...(sparse != null && { sparse }),\n ...(expireAfterSeconds != null && { expireAfterSeconds }),\n ...(partialFilterExpression != null && { partialFilterExpression }),\n ...(wildcardProjection != null && { wildcardProjection }),\n ...(collation != null && { collation }),\n ...(weights != null && { weights }),\n ...(default_language != null && { default_language }),\n ...(language_override != null && { language_override }),\n };\n\n indexes.push(index);\n }\n\n return indexes;\n}\n\nfunction isRelationField(field: PslField, modelNames: ReadonlySet<string>): boolean {\n return modelNames.has(field.typeName);\n}\n\nfunction resolveFieldCodecId(\n field: PslField,\n scalarTypeDescriptors: ReadonlyMap<string, string>,\n): string | undefined {\n return scalarTypeDescriptors.get(field.typeName);\n}\n\nfunction resolveNonRelationField(\n field: PslField,\n ownerName: string,\n compositeTypeNames: ReadonlySet<string>,\n scalarTypeDescriptors: ReadonlyMap<string, string>,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): ContractField | undefined {\n if (compositeTypeNames.has(field.typeName)) {\n const result: ContractField = {\n type: { kind: 'valueObject', name: field.typeName },\n nullable: field.optional,\n };\n return field.list ? { ...result, many: true } : result;\n }\n\n const codecId = resolveFieldCodecId(field, scalarTypeDescriptors);\n if (!codecId) {\n diagnostics.push({\n code: 'PSL_UNSUPPORTED_FIELD_TYPE',\n message: `Field \"${ownerName}.${field.name}\" type \"${field.typeName}\" is not supported in Mongo PSL interpreter`,\n sourceId,\n span: field.span,\n });\n return undefined;\n }\n\n const result: ContractField = {\n type: { kind: 'scalar', codecId },\n nullable: field.optional,\n };\n return field.list ? { ...result, many: true } : result;\n}\n\nexport function interpretPslDocumentToMongoContract(\n input: InterpretPslDocumentToMongoContractInput,\n): Result<Contract, ContractSourceDiagnostics> {\n const { document, scalarTypeDescriptors, codecLookup } = input;\n const sourceId = document.ast.sourceId;\n const diagnostics: ContractSourceDiagnostic[] = [];\n const modelNames = new Set(document.ast.models.map((m) => m.name));\n const compositeTypeNames = new Set(document.ast.compositeTypes.map((ct) => ct.name));\n\n const models: Record<string, MongoModelEntry> = {};\n const collections: Record<string, Record<string, unknown>> = {};\n const roots: Record<string, string> = {};\n const allFkRelations: FkRelation[] = [];\n\n interface BackrelationCandidate {\n readonly modelName: string;\n readonly fieldName: string;\n readonly targetModelName: string;\n readonly relationName?: string;\n readonly cardinality: '1:1' | '1:N';\n readonly field: PslField;\n }\n const backrelationCandidates: BackrelationCandidate[] = [];\n\n for (const pslModel of document.ast.models) {\n const collectionName = resolveCollectionName(pslModel);\n const fieldMappings = resolveFieldMappings(pslModel);\n\n const fields: Record<string, ContractField> = {};\n const relations: Record<string, ContractReferenceRelation> = {};\n\n for (const field of pslModel.fields) {\n if (isRelationField(field, modelNames)) {\n const relation = parseRelationAttribute(field.attributes);\n\n if (field.list || !(relation?.fields && relation?.references)) {\n backrelationCandidates.push({\n modelName: pslModel.name,\n fieldName: field.name,\n targetModelName: field.typeName,\n ...(relation?.relationName !== undefined\n ? { relationName: relation.relationName }\n : {}),\n cardinality: field.list ? '1:N' : '1:1',\n field,\n });\n continue;\n }\n\n if (relation?.fields && relation?.references) {\n const localMapped = relation.fields.map((f) => fieldMappings.pslNameToMapped.get(f) ?? f);\n\n const targetModel = document.ast.models.find((m) => m.name === field.typeName);\n const targetFieldMappings = targetModel ? resolveFieldMappings(targetModel) : undefined;\n const targetMapped = relation.references.map(\n (f) => targetFieldMappings?.pslNameToMapped.get(f) ?? f,\n );\n\n relations[field.name] = {\n to: field.typeName,\n cardinality: 'N:1' as const,\n on: {\n localFields: localMapped,\n targetFields: targetMapped,\n },\n };\n\n allFkRelations.push({\n declaringModel: pslModel.name,\n fieldName: field.name,\n targetModel: field.typeName,\n ...(relation.relationName !== undefined ? { relationName: relation.relationName } : {}),\n localFields: localMapped,\n targetFields: targetMapped,\n });\n }\n continue;\n }\n\n const resolved = resolveNonRelationField(\n field,\n pslModel.name,\n compositeTypeNames,\n scalarTypeDescriptors,\n sourceId,\n diagnostics,\n );\n if (!resolved) continue;\n\n const mappedName = fieldMappings.pslNameToMapped.get(field.name) ?? field.name;\n fields[mappedName] = resolved;\n }\n\n const isVariantModel = pslModel.attributes.some((attr) => attr.name === 'base');\n const hasIdField = pslModel.fields.some((f) => getAttribute(f.attributes, 'id') !== undefined);\n if (!hasIdField && !isVariantModel) {\n diagnostics.push({\n code: 'PSL_MISSING_ID_FIELD',\n message: `Model \"${pslModel.name}\" has no field with @id attribute. Every model must have exactly one @id field.`,\n sourceId,\n });\n }\n\n models[pslModel.name] = { fields, relations, storage: { collection: collectionName } };\n const modelIndexes = collectIndexes(pslModel, fieldMappings, modelNames, sourceId, diagnostics);\n const existingColl = collections[collectionName];\n if (existingColl && modelIndexes.length > 0) {\n const existingIndexes = (existingColl['indexes'] ?? []) as MongoStorageIndex[];\n collections[collectionName] = { indexes: [...existingIndexes, ...modelIndexes] };\n } else if (!existingColl) {\n collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};\n }\n roots[collectionName] = pslModel.name;\n }\n\n const valueObjects: Record<string, ContractValueObject> = {};\n for (const compositeType of document.ast.compositeTypes) {\n const fields: Record<string, ContractField> = {};\n for (const field of compositeType.fields) {\n const resolved = resolveNonRelationField(\n field,\n compositeType.name,\n compositeTypeNames,\n scalarTypeDescriptors,\n sourceId,\n diagnostics,\n );\n if (!resolved) continue;\n fields[field.name] = resolved;\n }\n valueObjects[compositeType.name] = { fields };\n }\n\n const fkRelationsByPair = new Map<string, FkRelation[]>();\n for (const fk of allFkRelations) {\n const key = fkRelationPairKey(fk.declaringModel, fk.targetModel);\n const existing = fkRelationsByPair.get(key);\n if (existing) {\n existing.push(fk);\n } else {\n fkRelationsByPair.set(key, [fk]);\n }\n }\n\n for (const candidate of backrelationCandidates) {\n const pairKey = fkRelationPairKey(candidate.targetModelName, candidate.modelName);\n const pairMatches = fkRelationsByPair.get(pairKey) ?? [];\n const matches = candidate.relationName\n ? pairMatches.filter((r) => r.relationName === candidate.relationName)\n : [...pairMatches];\n\n if (matches.length === 0) {\n diagnostics.push({\n code: 'PSL_ORPHANED_BACKRELATION',\n message: `Backrelation list field \"${candidate.modelName}.${candidate.fieldName}\" has no matching FK-side relation on model \"${candidate.targetModelName}\". Add @relation(fields: [...], references: [...]) on the FK-side relation or use an explicit join model for many-to-many.`,\n sourceId,\n span: candidate.field.span,\n });\n continue;\n }\n if (matches.length > 1) {\n diagnostics.push({\n code: 'PSL_AMBIGUOUS_BACKRELATION',\n message: `Backrelation list field \"${candidate.modelName}.${candidate.fieldName}\" matches multiple FK-side relations on model \"${candidate.targetModelName}\". Add @relation(\"...\") to both sides to disambiguate.`,\n sourceId,\n span: candidate.field.span,\n });\n continue;\n }\n\n const fk = matches[0];\n if (!fk) continue;\n const modelEntry = models[candidate.modelName];\n if (!modelEntry) continue;\n modelEntry.relations[candidate.fieldName] = {\n to: candidate.targetModelName,\n cardinality: candidate.cardinality,\n on: {\n localFields: fk.targetFields,\n targetFields: fk.localFields,\n },\n };\n }\n\n const { discriminatorDeclarations, baseDeclarations } = collectPolymorphismDeclarations(\n document,\n sourceId,\n diagnostics,\n );\n const polyResult = resolvePolymorphism({\n models,\n roots,\n collections,\n document,\n discriminatorDeclarations,\n baseDeclarations,\n modelNames,\n sourceId,\n });\n\n if (diagnostics.length > 0 || polyResult.diagnostics.length > 0) {\n return notOk({\n summary: 'PSL to Mongo contract interpretation failed',\n diagnostics: [...diagnostics, ...polyResult.diagnostics],\n });\n }\n\n const resolvedModels = polyResult.models;\n const resolvedCollections = polyResult.collections;\n\n for (const [, modelEntry] of Object.entries(resolvedModels)) {\n if (modelEntry.base) continue;\n\n const collectionName = modelEntry.storage.collection;\n const coll = resolvedCollections[collectionName];\n if (!coll) continue;\n\n if (modelEntry.discriminator && modelEntry.variants) {\n const variantEntries = Object.entries(modelEntry.variants).map(\n ([variantName, { value }]) => ({\n discriminatorValue: value,\n fields: resolvedModels[variantName]?.fields ?? {},\n }),\n );\n coll['validator'] = derivePolymorphicJsonSchema(\n modelEntry.fields,\n modelEntry.discriminator.field,\n variantEntries,\n valueObjects,\n codecLookup,\n );\n } else {\n coll['validator'] = deriveJsonSchema(modelEntry.fields, valueObjects, codecLookup);\n }\n }\n\n const target = 'mongo';\n const targetFamily = 'mongo';\n const storageWithoutHash = { collections: resolvedCollections };\n const storageHash = computeStorageHash({ target, targetFamily, storage: storageWithoutHash });\n const capabilities: Record<string, Record<string, boolean>> = {};\n\n return ok({\n targetFamily,\n target,\n roots: polyResult.roots,\n models: polyResult.models,\n ...(Object.keys(valueObjects).length > 0 ? { valueObjects } : {}),\n storage: { ...storageWithoutHash, storageHash },\n extensionPacks: {},\n capabilities,\n profileHash: computeProfileHash({ target, targetFamily, capabilities }),\n meta: {},\n });\n}\n"],"mappings":";;;;;AAIA,SAAS,gBACP,SACA,aACoB;AAEpB,SADc,aAAa,IAAI,QAAQ,GACzB,YAAY;;AAG5B,SAAS,kBACP,OACA,cACA,aACqC;AACrC,KAAI,MAAM,KAAK,SAAS,UAAU;EAChC,MAAM,WAAW,gBAAgB,MAAM,KAAK,SAAS,YAAY;AACjE,MAAI,CAAC,SAAU,QAAO;AAEtB,MAAI,UAAU,SAAS,MAAM,KAC3B,QAAO;GAAE,UAAU;GAAS,OAAO,EAAE,UAAU;GAAE;AAGnD,MAAI,MAAM,SACR,QAAO,EAAE,UAAU,CAAC,QAAQ,SAAS,EAAE;AAGzC,SAAO,EAAE,UAAU;;AAGrB,KAAI,MAAM,KAAK,SAAS,eAAe;EACrC,MAAM,KAAK,eAAe,MAAM,KAAK;AACrC,MAAI,CAAC,GAAI,QAAO;EAChB,MAAM,WAAW,mBAAmB,GAAG,QAAQ,cAAc,YAAY;AACzE,MAAI,UAAU,SAAS,MAAM,KAC3B,QAAO;GAAE,UAAU;GAAS,OAAO;GAAU;AAE/C,MAAI,MAAM,SACR,QAAO,EAAE,OAAO,CAAC,EAAE,UAAU,QAAQ,EAAE,SAAS,EAAE;AAEpD,SAAO;;;AAMX,SAAS,mBACP,QACA,cACA,aACyB;CACzB,MAAMA,aAAsC,EAAE;CAC9C,MAAMC,WAAqB,EAAE;AAE7B,MAAK,MAAM,CAAC,WAAW,UAAU,OAAO,QAAQ,OAAO,EAAE;EACvD,MAAM,SAAS,kBAAkB,OAAO,cAAc,YAAY;AAClE,MAAI,QAAQ;AACV,cAAW,aAAa;AACxB,OAAI,CAAC,MAAM,SACT,UAAS,KAAK,UAAU;;;CAK9B,MAAMC,SAAkC;EACtC,UAAU;EACV;EACD;AACD,KAAI,SAAS,SAAS,EACpB,QAAO,cAAc,SAAS,MAAM;AAEtC,QAAO;;AAGT,SAAgB,iBACd,QACA,cACA,aACuB;AACvB,QAAO;EACL,YAAY,mBAAmB,QAAQ,cAAc,YAAY;EACjE,iBAAiB;EACjB,kBAAkB;EACnB;;AAQH,SAAgB,4BACd,YACA,oBACA,UACA,cACA,aACuB;CACvB,MAAM,aAAa,mBAAmB,YAAY,cAAc,YAAY;CAE5E,MAAMC,QAAmC,EAAE;AAC3C,MAAK,MAAM,WAAW,UAAU;EAC9B,MAAMC,oBAAmD,EAAE;AAC3D,OAAK,MAAM,CAAC,MAAM,UAAU,OAAO,QAAQ,QAAQ,OAAO,CACxD,KAAI,EAAE,QAAQ,YACZ,mBAAkB,QAAQ;EAI9B,MAAMC,QAAiC,EACrC,YAAY,GACT,qBAAqB,EAAE,MAAM,CAAC,QAAQ,mBAAmB,EAAE,EAC7D,EACF;EAED,MAAMC,oBAA6C,EAAE;EACrD,MAAMC,kBAA4B,CAAC,mBAAmB;AACtD,OAAK,MAAM,CAAC,MAAM,UAAU,OAAO,QAAQ,kBAAkB,EAAE;GAC7D,MAAM,SAAS,kBAAkB,OAAO,cAAc,YAAY;AAClE,OAAI,QAAQ;AACV,sBAAkB,QAAQ;AAC1B,QAAI,CAAC,MAAM,SACT,iBAAgB,KAAK,KAAK;;;AAKhC,MAAI,OAAO,KAAK,kBAAkB,CAAC,SAAS,EAC1C,CAAC,MAAM,gBAA4C;GACjD,GAAI,MAAM;GACV,GAAG;GACJ;AAEH,QAAM,cAAc,gBAAgB,MAAM;AAE1C,QAAM,KAAK,MAAM;;CAGnB,MAAM,aAAa,EAAE,GAAG,YAAY;AACpC,KAAI,MAAM,SAAS,EACjB,YAAW,WAAW;AAGxB,QAAO;EACL;EACA,iBAAiB;EACjB,kBAAkB;EACnB;;;;;AChJH,SAAgB,iBAAiB,MAAoB,MAAkC;AAErF,QADY,KAAK,KAAK,MAAM,MAAM,EAAE,SAAS,WAAW,EAAE,SAAS,KAAK,EAC5D;;AAGd,SAAgB,eAAe,OAAkC;CAC/D,MAAM,QAAQ,MAAM,QAAQ,OAAO,GAAG,CAAC,QAAQ,OAAO,GAAG,CAAC,MAAM;AAChE,KAAI,MAAM,WAAW,EAAG,QAAO,EAAE;AACjC,QAAO,cAAc,MAAM,CAAC,KAAK,MAAM,EAAE,MAAM,CAAC;;AASlD,SAAgB,oBAAoB,OAA4C;AAE9E,QADiB,eAAe,MAAM,CACtB,IAAI,uBAAuB;;AAG7C,SAAS,uBAAuB,SAAmC;CACjE,MAAM,gBAAgB,QAAQ,MAAM,4BAA4B;AAChE,KAAI,eAAe;EACjB,MAAM,QAAQ,cAAc,MAAM;AAClC,SAAO;GACL,MAAM,MAAM,SAAS,IAAI,GAAG,MAAM,QAAQ;GAC1C,YAAY;GACb;;CAGH,MAAM,gBAAgB,QAAQ,MAAM,iCAAiC;AACrE,KAAI,cAGF,QAAO;EACL,MAHgB,cAAc,MAAM;EAIpC,YAAY;EACZ,WAJgB,cAAc,OAIL,SAAS,KAAK;EACxC;AAGH,QAAO;EAAE,MAAM;EAAS,YAAY;EAAO;;AAG7C,SAAS,cAAc,OAAyB;CAC9C,MAAMC,QAAkB,EAAE;CAC1B,IAAI,QAAQ;CACZ,IAAI,QAAQ;AACZ,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;EACrC,MAAM,KAAK,MAAM;AACjB,MAAI,OAAO,OAAO,OAAO,OAAO,OAAO,IAAK;WACnC,OAAO,OAAO,OAAO,OAAO,OAAO,IAAK,SAAQ,KAAK,IAAI,GAAG,QAAQ,EAAE;WACtE,OAAO,OAAO,UAAU,GAAG;AAClC,SAAM,KAAK,MAAM,MAAM,OAAO,EAAE,CAAC;AACjC,WAAQ,IAAI;;;AAGhB,OAAM,KAAK,MAAM,MAAM,MAAM,CAAC;AAC9B,QAAO;;AAGT,SAAgB,WAAW,OAAuB;AAChD,KAAI,MAAM,WAAW,EAAG,QAAO;AAC/B,QAAO,MAAM,IAAI,aAAa,GAAG,MAAM,MAAM,EAAE;;AAGjD,SAAgB,aACd,YACA,MAC0B;AAC1B,QAAO,WAAW,MAAM,SAAS,KAAK,SAAS,KAAK;;AAGtD,SAAgB,WAAW,YAAyD;CAClF,MAAM,UAAU,aAAa,YAAY,MAAM;AAC/C,KAAI,CAAC,QAAS,QAAO;CACrB,MAAM,MAAM,QAAQ,KAAK;AACzB,KAAI,CAAC,IAAK,QAAO;AACjB,QAAO,YAAY,IAAI,MAAM;;AAS/B,SAAgB,uBACd,YACqC;CACrC,MAAM,eAAe,aAAa,YAAY,WAAW;AACzD,KAAI,CAAC,aAAc,QAAO;CAE1B,IAAIC;CACJ,IAAIC;CACJ,IAAIC;AAEJ,MAAK,MAAM,OAAO,aAAa,KAC7B,KAAI,IAAI,SAAS,aACf,gBAAe,YAAY,IAAI,MAAM;UAC5B,IAAI,SAAS,OACtB,gBAAe,YAAY,IAAI,MAAM;UAC5B,IAAI,SAAS,SACtB,aAAY;UACH,IAAI,SAAS,aACtB,iBAAgB;CAIpB,MAAM,SAAS,YAAY,eAAe,UAAU,MAAM,GAAG;CAC7D,MAAM,aAAa,gBAAgB,eAAe,cAAc,MAAM,GAAG;AAEzE,QAAO;EACL,GAAI,iBAAiB,SAAY,EAAE,cAAc,GAAG,EAAE;EACtD,GAAI,WAAW,SAAY,EAAE,QAAQ,GAAG,EAAE;EAC1C,GAAI,eAAe,SAAY,EAAE,YAAY,GAAG,EAAE;EACnD;;AAGH,SAAS,YAAY,OAAuB;AAC1C,KAAI,MAAM,WAAW,KAAI,IAAI,MAAM,SAAS,KAAI,CAC9C,QAAO,MAAM,MAAM,GAAG,GAAG;AAE3B,QAAO;;;;;ACpFT,SAAS,kBAAkB,gBAAwB,aAA6B;AAC9E,QAAO,GAAG,eAAe,IAAI;;AAG/B,SAAS,qBAAqB,OAAgC;CAC5D,MAAM,kCAAkB,IAAI,KAAqB;AACjD,MAAK,MAAM,SAAS,MAAM,QAAQ;EAChC,MAAM,SAAS,WAAW,MAAM,WAAW,IAAI,MAAM;AACrD,kBAAgB,IAAI,MAAM,MAAM,OAAO;;AAEzC,QAAO,EAAE,iBAAiB;;AAG5B,SAAS,sBAAsB,OAAyB;AACtD,QAAO,WAAW,MAAM,WAAW,IAAI,WAAW,MAAM,KAAK;;AAoB/D,SAAS,gCACP,UACA,UACA,aAIA;CACA,MAAM,4CAA4B,IAAI,KAAuC;CAC7E,MAAM,mCAAmB,IAAI,KAA8B;AAE3D,MAAK,MAAM,YAAY,SAAS,IAAI,OAClC,MAAK,MAAM,QAAQ,SAAS,YAAY;AACtC,MAAI,KAAK,SAAS,iBAAiB;GACjC,MAAM,YAAY,sBAAsB,KAAK;AAC7C,OAAI,CAAC,WAAW;AACd,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,YAAY,SAAS,OAAO,MAAM,MAAM,EAAE,SAAS,UAAU;AACnE,OAAI,aAAa,UAAU,aAAa,UAAU;AAChD,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,wBAAwB,UAAU,cAAc,SAAS,KAAK,oCAAoC,UAAU,SAAS;KAC9H;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;AAEF,6BAA0B,IAAI,SAAS,MAAM;IAAE;IAAW,MAAM,KAAK;IAAM,CAAC;;AAE9E,MAAI,KAAK,SAAS,QAAQ;GACxB,MAAM,WAAW,sBAAsB,MAAM,EAAE;GAC/C,MAAM,WAAW,sBAAsB,MAAM,EAAE;AAC/C,OAAI,CAAC,YAAY,CAAC,UAAU;AAC1B,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,QAAQ,yBAAyB,SAAS;AAChD,OAAI,UAAU,QAAW;AACvB,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,iBAAiB,sBAAsB,SAAS;AACtD,oBAAiB,IAAI,SAAS,MAAM;IAAE;IAAU;IAAO;IAAgB,MAAM,KAAK;IAAM,CAAC;;;AAK/F,QAAO;EAAE;EAA2B;EAAkB;;AAGxD,SAAS,oBAAoB,OAc3B;CACA,MAAM,EAAE,2BAA2B,kBAAkB,YAAY,UAAU,aAAa;CACxF,IAAI,UAAU,MAAM;CACpB,IAAI,QAAQ,MAAM;CAClB,IAAI,cAAc,MAAM;CACxB,MAAMC,cAA0C,EAAE;AAElD,MAAK,MAAM,CAAC,WAAW,SAAS,2BAA2B;AACzD,MAAI,iBAAiB,IAAI,UAAU,EAAE;AACnC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,UAAU;IAC7B;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,QAAQ,QAAQ;AACtB,MAAI,CAAC,MAAO;EAEZ,MAAM,WAAW,SAAS,IAAI,OAAO,MAAM,MAAM,EAAE,SAAS,UAAU;EACtE,MAAM,2BAA2B,WAC5B,qBAAqB,SAAS,CAAC,gBAAgB,IAAI,KAAK,UAAU,IAAI,KAAK,YAC5E,KAAK;AAET,MAAI,CAAC,OAAO,OAAO,MAAM,QAAQ,yBAAyB,EAAE;AAC1D,eAAY,KAAK;IACf,MAAM;IACN,SAAS,wBAAwB,KAAK,UAAU,6BAA6B,UAAU;IACvF;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAMC,WAAuD,EAAE;AAC/D,OAAK,MAAM,CAAC,aAAa,aAAa,kBAAkB;AACtD,OAAI,SAAS,aAAa,UAAW;AACrC,YAAS,eAAe,EAAE,OAAO,SAAS,OAAO;;AAGnD,MAAI,OAAO,KAAK,SAAS,CAAC,WAAW,GAAG;AACtC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,UAAU,6DAA6D,UAAU;IACpG;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,YAAU;GACR,GAAG;IACF,YAAY;IAAE,GAAG;IAAO,eAAe,EAAE,OAAO,0BAA0B;IAAE;IAAU;GACxF;;AAGH,MAAK,MAAM,CAAC,aAAa,aAAa,kBAAkB;AACtD,MAAI,CAAC,WAAW,IAAI,SAAS,SAAS,EAAE;AACtC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,YAAY,0CAA0C,SAAS,SAAS;IAC3F;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;AAGF,MAAI,CAAC,0BAA0B,IAAI,SAAS,SAAS,EAAE;AACrD,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,YAAY,oBAAoB,SAAS,SAAS,cAAc,SAAS,SAAS;IACrG;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;AAGF,MAAI,0BAA0B,IAAI,YAAY,CAC5C;EAGF,MAAM,YAAY,QAAQ,SAAS;EACnC,MAAM,kBAAkB,SAAS,IAAI,OAAO,MAAM,MAAM,EAAE,SAAS,YAAY;AAC/E,MAAI,CAAC,gBAAiB;AAGtB,MAFuB,WAAW,gBAAgB,WAAW,KAAK,UAE5C,aAAa,SAAS,mBAAmB,UAAU,QAAQ,YAAY;AAC3F,eAAY,KAAK;IACf,MAAM;IACN,SAAS,kBAAkB,YAAY,qDAAqD,SAAS,SAAS;IAC9G;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;EAGF,MAAM,iBAAiB,WAAW,QAAQ,cAAc,SAAS;EACjE,MAAM,eAAe,QAAQ;AAC7B,MAAI,aACF,WAAU;GACR,GAAG;IACF,cAAc;IACb,GAAG;IACH,MAAM,SAAS;IACf,SAAS,EAAE,YAAY,gBAAgB;IACxC;GACF;EAGH,MAAM,wBAAwB,sBAAsB,gBAAgB;AACpE,MAAI,MAAM,2BAA2B,YACnC,KAAI,0BAA0B,kBAAkB,UAC9C,SAAQ;GAAE,GAAG;IAAQ,wBAAwB,SAAS;GAAU;MAEhE,SAAQ,OAAO,YACb,OAAO,QAAQ,MAAM,CAAC,QAAQ,CAAC,SAAS,QAAQ,sBAAsB,CACvE;EAKL,MAAM,iBADc,YAAY,yBACM,cAAc,EAAE;EACtD,MAAM,WAAW,YAAY;AAE7B,MAAI,0BAA0B,gBAAgB;GAC5C,MAAM,WAAW,OAAO,YACtB,OAAO,QAAQ,YAAY,CAAC,QAAQ,CAAC,SAAS,QAAQ,sBAAsB,CAC7E;AACD,OAAI,eAAe,SAAS,KAAK,UAAU;IACzC,MAAM,cAAe,SAAS,cAAc,EAAE;AAC9C,kBAAc;KACZ,GAAG;MACF,iBAAiB;MAAE,GAAG;MAAU,SAAS,CAAC,GAAG,aAAa,GAAG,eAAe;MAAE;KAChF;SAED,eAAc;aAEP,eAAe,SAAS,KAAK,UAAU;GAChD,MAAM,cAAe,SAAS,cAAc,EAAE;GAC9C,MAAM,gBAAgB,CAAC,GAAG,YAAY;AACtC,QAAK,MAAM,OAAO,eAIhB,KAAI,CAHgB,YAAY,MAC7B,aAAa,KAAK,UAAU,SAAS,KAAK,KAAK,UAAU,IAAI,CAC/D,CAEC,eAAc,KAAK,IAAI;AAG3B,OAAI,cAAc,SAAS,YAAY,OACrC,eAAc;IACZ,GAAG;KACF,iBAAiB;KAAE,GAAG;KAAU,SAAS;KAAe;IAC1D;;;AAKP,QAAO;EAAE,QAAQ;EAAS;EAAO;EAAa;EAAa;;AAG7D,SAAS,oBAAoB,KAAiD;AAC5E,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,WAAW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG;CAC9D,MAAM,MAAM,OAAO,SAAS;AAC5B,KAAI,QAAQ,KAAK,QAAQ,GAAI,QAAO;AACpC,KAAI;EAAC;EAAQ;EAAY;EAAM;EAAS,CAAC,SAAS,SAAS,CACzD,QAAO;AACT,QAAO;;AAGT,SAAS,gBAAgB,KAA6C;AACpE,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,IAAI,OAAO,IAAI;AACrB,QAAO,OAAO,SAAS,EAAE,GAAG,IAAI;;AAGlC,SAAS,gBAAgB,KAA8C;AACrE,KAAI,QAAQ,OAAQ,QAAO;AAC3B,KAAI,QAAQ,QAAS,QAAO;;AAI9B,SAAS,aAAa,KAA8D;AAClF,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,WAAW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG,CAAC,QAAQ,QAAQ,KAAI;AACnF,KAAI;EACF,MAAM,SAAS,KAAK,MAAM,SAAS;AACnC,MAAI,OAAO,WAAW,YAAY,WAAW,QAAQ,CAAC,MAAM,QAAQ,OAAO,CACzE,QAAO;SAEH;;AAMV,SAAS,eACP,MAC4C;CAC5C,MAAM,SAAS,kBAAkB,iBAAiB,MAAM,kBAAkB,CAAC;AAC3E,KAAI,CAAC,OAUH,QARE,iBAAiB,MAAM,oBAAoB,IAAI,QAC/C,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,2BAA2B,IAAI,QACtD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,uBAAuB,IAAI,QAClD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,yBAAyB,IAAI,OAC1B,OAAO;CAGrC,MAAMC,YAAqC,EAAE,QAAQ;CACrD,MAAM,WAAW,gBAAgB,iBAAiB,MAAM,oBAAoB,CAAC;AAC7E,KAAI,YAAY,KAAM,WAAU,cAAc;CAC9C,MAAM,YAAY,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAC/E,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,YAAY,kBAAkB,iBAAiB,MAAM,qBAAqB,CAAC;AACjF,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,kBAAkB,gBAAgB,iBAAiB,MAAM,2BAA2B,CAAC;AAC3F,KAAI,mBAAmB,KAAM,WAAU,qBAAqB;CAC5D,MAAM,YAAY,kBAAkB,iBAAiB,MAAM,qBAAqB,CAAC;AACjF,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,cAAc,kBAAkB,iBAAiB,MAAM,uBAAuB,CAAC;AACrF,KAAI,eAAe,KAAM,WAAU,iBAAiB;CACpD,MAAM,YAAY,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAC/E,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,gBAAgB,gBAAgB,iBAAiB,MAAM,yBAAyB,CAAC;AACvF,KAAI,iBAAiB,KAAM,WAAU,mBAAmB;AACxD,QAAO;;AAGT,SAAS,kBAAkB,KAA6C;AACtE,KAAI,CAAC,IAAK,QAAO;AACjB,QAAO,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG;;AAGtD,SAAS,oBACP,KACA,OACmC;AACnC,KAAI,CAAC,IAAK,QAAO;CAEjB,MAAM,QADW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG,CACvC,QAAQ,OAAO,GAAG,CAAC,QAAQ,OAAO,GAAG,CAAC,MAAM;AACnE,KAAI,MAAM,WAAW,EAAG,QAAO;CAC/B,MAAM,SAAS,MACZ,MAAM,IAAI,CACV,KAAK,MAAM,EAAE,MAAM,CAAC,CACpB,QAAQ,MAAM,EAAE,SAAS,EAAE;CAC9B,MAAMC,SAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,OACd,QAAO,KAAK;AAEd,QAAO;;AAGT,SAAS,eACP,UACA,eACA,YACA,UACA,aACqB;CACrB,MAAMC,UAA+B,EAAE;CACvC,IAAI,iBAAiB;AAErB,MAAK,MAAM,SAAS,SAAS,QAAQ;AACnC,MAAI,WAAW,IAAI,MAAM,SAAS,CAAE;AAEpC,MAAI,CADe,aAAa,MAAM,YAAY,SAAS,CAC1C;EACjB,MAAM,aAAa,cAAc,gBAAgB,IAAI,MAAM,KAAK,IAAI,MAAM;AAC1E,UAAQ,KAAK;GACX,MAAM,CAAC;IAAE,OAAO;IAAY,WAAW;IAAG,CAAC;GAC3C,QAAQ;GACT,CAAC;;AAGJ,MAAK,MAAM,QAAQ,SAAS,YAAY;EACtC,MAAM,UAAU,KAAK,SAAS;EAC9B,MAAM,WAAW,KAAK,SAAS;EAC/B,MAAM,cAAc,KAAK,SAAS;AAClC,MAAI,CAAC,WAAW,CAAC,YAAY,CAAC,YAAa;EAE3C,MAAM,YAAY,sBAAsB,MAAM,EAAE;AAChD,MAAI,CAAC,UAAW;EAChB,MAAM,eAAe,oBAAoB,UAAU;AACnD,MAAI,aAAa,WAAW,EAAG;EAE/B,MAAM,cAAc,aAAa,MAAM,MAAM,EAAE,WAAW;AAG1D,MAFsB,aAAa,QAAQ,MAAM,EAAE,WAAW,CAAC,SAE3C,GAAG;AACrB,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,YAAY,aAAa;AAC3B,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,aAAa;AACf;AACA,OAAI,iBAAiB,GAAG;AACtB,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,0DAA0D,SAAS,KAAK;KACjF;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;AAGF,OAAI,aAAa;AACf,gBAAY,KAAK;KACf,MAAM;KACN,SACE;KACF;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;;EAIJ,MAAM,UAAU,iBAAiB,MAAM,OAAO;EAC9C,MAAMC,mBAA2C,cAC7C,SACA,oBAAoB,QAAQ;AAEhC,MACE,eACA,OAAO,qBAAqB,YAC5B;GAAC;GAAU;GAAY;GAAK,CAAC,SAAS,iBAAiB,EACvD;AACA,eAAY,KAAK;IACf,MAAM;IACN,SAAS,mDAAmD;IAC5D;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,qBAAqB,YAAY,aAAa,SAAS,GAAG;AAC5D,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,OAAO,aAAa,KAAK,OAAO;AASpC,UAAO;IAAE,OARU,GAAG,aAClB,GAAG,KAAK,QAAQ,mBAAmB,GAAG,WAAmB;KACvD,MAAM,SAAS,cAAc,gBAAgB,IAAI,OAAO;AACxD,YAAO,SAAS,GAAG,OAAO,QAAQ,GAAG,OAAO;MAC5C,GACD,cAAc,gBAAgB,IAAI,GAAG,KAAK,IAAI,GAAG;IAG1B,WAD1B,GAAG,aAAa,OAAQ,GAAG,YAAuC;IAC7B;IACvC;EAEF,MAAM,SAAS,WAAW,OAAO;EACjC,MAAM,SAAS,cAAc,SAAY,gBAAgB,iBAAiB,MAAM,SAAS,CAAC;EAC1F,MAAM,qBAAqB,cACvB,SACA,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAEjE,MAAI,eAAe,sBAAsB,MAAM;AAC7C,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,0BAA0B,aAAa,iBAAiB,MAAM,SAAS,CAAC;EAE9E,MAAM,aAAa,iBAAiB,MAAM,UAAU;EACpD,MAAM,aAAa,iBAAiB,MAAM,UAAU;AAEpD,MAAI,cAAc,QAAQ,cAAc,MAAM;AAC5C,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,OAAK,cAAc,QAAQ,cAAc,SAAS,CAAC,aAAa;AAC9D,eAAY,KAAK;IACf,MAAM;IACN,SACE;IACF;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,qBACJ,cAAc,OACV,oBAAoB,YAAY,EAAE,GAClC,cAAc,OACZ,oBAAoB,YAAY,EAAE,GAClC;EAER,MAAM,YAAY,eAAe,KAAK;AACtC,MAAI,cAAc,MAAM;AACtB,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,aAAa,aAAa,iBAAiB,MAAM,UAAU,CAAC;EAClE,IAAIC;AACJ,MAAI,YAAY;AACd,aAAU,EAAE;AACZ,QAAK,MAAM,CAAC,GAAG,MAAM,OAAO,QAAQ,WAAW,CAC7C,KAAI,OAAO,MAAM,SAAU,SAAQ,KAAK;;EAO5C,MAAM,mBAAmB,kBAHF,cACnB,iBAAiB,MAAM,WAAW,GAClC,iBAAiB,MAAM,mBAAmB,CACY;EAG1D,MAAM,oBAAoB,kBADF,iBAAiB,MAAM,mBAAmB,CACN;EAE5D,MAAMC,QAA2B;GAC/B;GACA,GAAI,UAAU,QAAQ,EAAE,QAAQ;GAChC,GAAI,UAAU,QAAQ,EAAE,QAAQ;GAChC,GAAI,sBAAsB,QAAQ,EAAE,oBAAoB;GACxD,GAAI,2BAA2B,QAAQ,EAAE,yBAAyB;GAClE,GAAI,sBAAsB,QAAQ,EAAE,oBAAoB;GACxD,GAAI,aAAa,QAAQ,EAAE,WAAW;GACtC,GAAI,WAAW,QAAQ,EAAE,SAAS;GAClC,GAAI,oBAAoB,QAAQ,EAAE,kBAAkB;GACpD,GAAI,qBAAqB,QAAQ,EAAE,mBAAmB;GACvD;AAED,UAAQ,KAAK,MAAM;;AAGrB,QAAO;;AAGT,SAAS,gBAAgB,OAAiB,YAA0C;AAClF,QAAO,WAAW,IAAI,MAAM,SAAS;;AAGvC,SAAS,oBACP,OACA,uBACoB;AACpB,QAAO,sBAAsB,IAAI,MAAM,SAAS;;AAGlD,SAAS,wBACP,OACA,WACA,oBACA,uBACA,UACA,aAC2B;AAC3B,KAAI,mBAAmB,IAAI,MAAM,SAAS,EAAE;EAC1C,MAAMC,WAAwB;GAC5B,MAAM;IAAE,MAAM;IAAe,MAAM,MAAM;IAAU;GACnD,UAAU,MAAM;GACjB;AACD,SAAO,MAAM,OAAO;GAAE,GAAGC;GAAQ,MAAM;GAAM,GAAGA;;CAGlD,MAAM,UAAU,oBAAoB,OAAO,sBAAsB;AACjE,KAAI,CAAC,SAAS;AACZ,cAAY,KAAK;GACf,MAAM;GACN,SAAS,UAAU,UAAU,GAAG,MAAM,KAAK,UAAU,MAAM,SAAS;GACpE;GACA,MAAM,MAAM;GACb,CAAC;AACF;;CAGF,MAAMD,SAAwB;EAC5B,MAAM;GAAE,MAAM;GAAU;GAAS;EACjC,UAAU,MAAM;EACjB;AACD,QAAO,MAAM,OAAO;EAAE,GAAG;EAAQ,MAAM;EAAM,GAAG;;AAGlD,SAAgB,oCACd,OAC6C;CAC7C,MAAM,EAAE,UAAU,uBAAuB,gBAAgB;CACzD,MAAM,WAAW,SAAS,IAAI;CAC9B,MAAMR,cAA0C,EAAE;CAClD,MAAM,aAAa,IAAI,IAAI,SAAS,IAAI,OAAO,KAAK,MAAM,EAAE,KAAK,CAAC;CAClE,MAAM,qBAAqB,IAAI,IAAI,SAAS,IAAI,eAAe,KAAK,OAAO,GAAG,KAAK,CAAC;CAEpF,MAAMU,SAA0C,EAAE;CAClD,MAAMC,cAAuD,EAAE;CAC/D,MAAMC,QAAgC,EAAE;CACxC,MAAMC,iBAA+B,EAAE;CAUvC,MAAMC,yBAAkD,EAAE;AAE1D,MAAK,MAAM,YAAY,SAAS,IAAI,QAAQ;EAC1C,MAAM,iBAAiB,sBAAsB,SAAS;EACtD,MAAM,gBAAgB,qBAAqB,SAAS;EAEpD,MAAMC,SAAwC,EAAE;EAChD,MAAMC,YAAuD,EAAE;AAE/D,OAAK,MAAM,SAAS,SAAS,QAAQ;AACnC,OAAI,gBAAgB,OAAO,WAAW,EAAE;IACtC,MAAM,WAAW,uBAAuB,MAAM,WAAW;AAEzD,QAAI,MAAM,QAAQ,EAAE,UAAU,UAAU,UAAU,aAAa;AAC7D,4BAAuB,KAAK;MAC1B,WAAW,SAAS;MACpB,WAAW,MAAM;MACjB,iBAAiB,MAAM;MACvB,GAAI,UAAU,iBAAiB,SAC3B,EAAE,cAAc,SAAS,cAAc,GACvC,EAAE;MACN,aAAa,MAAM,OAAO,QAAQ;MAClC;MACD,CAAC;AACF;;AAGF,QAAI,UAAU,UAAU,UAAU,YAAY;KAC5C,MAAM,cAAc,SAAS,OAAO,KAAK,MAAM,cAAc,gBAAgB,IAAI,EAAE,IAAI,EAAE;KAEzF,MAAM,cAAc,SAAS,IAAI,OAAO,MAAM,MAAM,EAAE,SAAS,MAAM,SAAS;KAC9E,MAAM,sBAAsB,cAAc,qBAAqB,YAAY,GAAG;KAC9E,MAAM,eAAe,SAAS,WAAW,KACtC,MAAM,qBAAqB,gBAAgB,IAAI,EAAE,IAAI,EACvD;AAED,eAAU,MAAM,QAAQ;MACtB,IAAI,MAAM;MACV,aAAa;MACb,IAAI;OACF,aAAa;OACb,cAAc;OACf;MACF;AAED,oBAAe,KAAK;MAClB,gBAAgB,SAAS;MACzB,WAAW,MAAM;MACjB,aAAa,MAAM;MACnB,GAAI,SAAS,iBAAiB,SAAY,EAAE,cAAc,SAAS,cAAc,GAAG,EAAE;MACtF,aAAa;MACb,cAAc;MACf,CAAC;;AAEJ;;GAGF,MAAM,WAAW,wBACf,OACA,SAAS,MACT,oBACA,uBACA,UACA,YACD;AACD,OAAI,CAAC,SAAU;GAEf,MAAM,aAAa,cAAc,gBAAgB,IAAI,MAAM,KAAK,IAAI,MAAM;AAC1E,UAAO,cAAc;;EAGvB,MAAM,iBAAiB,SAAS,WAAW,MAAM,SAAS,KAAK,SAAS,OAAO;AAE/E,MAAI,CADe,SAAS,OAAO,MAAM,MAAM,aAAa,EAAE,YAAY,KAAK,KAAK,OAAU,IAC3E,CAAC,eAClB,aAAY,KAAK;GACf,MAAM;GACN,SAAS,UAAU,SAAS,KAAK;GACjC;GACD,CAAC;AAGJ,SAAO,SAAS,QAAQ;GAAE;GAAQ;GAAW,SAAS,EAAE,YAAY,gBAAgB;GAAE;EACtF,MAAM,eAAe,eAAe,UAAU,eAAe,YAAY,UAAU,YAAY;EAC/F,MAAM,eAAe,YAAY;AACjC,MAAI,gBAAgB,aAAa,SAAS,EAExC,aAAY,kBAAkB,EAAE,SAAS,CAAC,GADjB,aAAa,cAAc,EAAE,EACQ,GAAG,aAAa,EAAE;WACvE,CAAC,aACV,aAAY,kBAAkB,aAAa,SAAS,IAAI,EAAE,SAAS,cAAc,GAAG,EAAE;AAExF,QAAM,kBAAkB,SAAS;;CAGnC,MAAMC,eAAoD,EAAE;AAC5D,MAAK,MAAM,iBAAiB,SAAS,IAAI,gBAAgB;EACvD,MAAMF,SAAwC,EAAE;AAChD,OAAK,MAAM,SAAS,cAAc,QAAQ;GACxC,MAAM,WAAW,wBACf,OACA,cAAc,MACd,oBACA,uBACA,UACA,YACD;AACD,OAAI,CAAC,SAAU;AACf,UAAO,MAAM,QAAQ;;AAEvB,eAAa,cAAc,QAAQ,EAAE,QAAQ;;CAG/C,MAAM,oCAAoB,IAAI,KAA2B;AACzD,MAAK,MAAM,MAAM,gBAAgB;EAC/B,MAAM,MAAM,kBAAkB,GAAG,gBAAgB,GAAG,YAAY;EAChE,MAAM,WAAW,kBAAkB,IAAI,IAAI;AAC3C,MAAI,SACF,UAAS,KAAK,GAAG;MAEjB,mBAAkB,IAAI,KAAK,CAAC,GAAG,CAAC;;AAIpC,MAAK,MAAM,aAAa,wBAAwB;EAC9C,MAAM,UAAU,kBAAkB,UAAU,iBAAiB,UAAU,UAAU;EACjF,MAAM,cAAc,kBAAkB,IAAI,QAAQ,IAAI,EAAE;EACxD,MAAM,UAAU,UAAU,eACtB,YAAY,QAAQ,MAAM,EAAE,iBAAiB,UAAU,aAAa,GACpE,CAAC,GAAG,YAAY;AAEpB,MAAI,QAAQ,WAAW,GAAG;AACxB,eAAY,KAAK;IACf,MAAM;IACN,SAAS,4BAA4B,UAAU,UAAU,GAAG,UAAU,UAAU,+CAA+C,UAAU,gBAAgB;IACzJ;IACA,MAAM,UAAU,MAAM;IACvB,CAAC;AACF;;AAEF,MAAI,QAAQ,SAAS,GAAG;AACtB,eAAY,KAAK;IACf,MAAM;IACN,SAAS,4BAA4B,UAAU,UAAU,GAAG,UAAU,UAAU,iDAAiD,UAAU,gBAAgB;IAC3J;IACA,MAAM,UAAU,MAAM;IACvB,CAAC;AACF;;EAGF,MAAM,KAAK,QAAQ;AACnB,MAAI,CAAC,GAAI;EACT,MAAM,aAAa,OAAO,UAAU;AACpC,MAAI,CAAC,WAAY;AACjB,aAAW,UAAU,UAAU,aAAa;GAC1C,IAAI,UAAU;GACd,aAAa,UAAU;GACvB,IAAI;IACF,aAAa,GAAG;IAChB,cAAc,GAAG;IAClB;GACF;;CAGH,MAAM,EAAE,2BAA2B,qBAAqB,gCACtD,UACA,UACA,YACD;CACD,MAAM,aAAa,oBAAoB;EACrC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;EACD,CAAC;AAEF,KAAI,YAAY,SAAS,KAAK,WAAW,YAAY,SAAS,EAC5D,QAAO,MAAM;EACX,SAAS;EACT,aAAa,CAAC,GAAG,aAAa,GAAG,WAAW,YAAY;EACzD,CAAC;CAGJ,MAAM,iBAAiB,WAAW;CAClC,MAAM,sBAAsB,WAAW;AAEvC,MAAK,MAAM,GAAG,eAAe,OAAO,QAAQ,eAAe,EAAE;AAC3D,MAAI,WAAW,KAAM;EAGrB,MAAM,OAAO,oBADU,WAAW,QAAQ;AAE1C,MAAI,CAAC,KAAM;AAEX,MAAI,WAAW,iBAAiB,WAAW,UAAU;GACnD,MAAM,iBAAiB,OAAO,QAAQ,WAAW,SAAS,CAAC,KACxD,CAAC,aAAa,EAAE,cAAc;IAC7B,oBAAoB;IACpB,QAAQ,eAAe,cAAc,UAAU,EAAE;IAClD,EACF;AACD,QAAK,eAAe,4BAClB,WAAW,QACX,WAAW,cAAc,OACzB,gBACA,cACA,YACD;QAED,MAAK,eAAe,iBAAiB,WAAW,QAAQ,cAAc,YAAY;;CAItF,MAAM,SAAS;CACf,MAAM,eAAe;CACrB,MAAM,qBAAqB,EAAE,aAAa,qBAAqB;CAC/D,MAAM,cAAc,mBAAmB;EAAE;EAAQ;EAAc,SAAS;EAAoB,CAAC;CAC7F,MAAMG,eAAwD,EAAE;AAEhE,QAAO,GAAG;EACR;EACA;EACA,OAAO,WAAW;EAClB,QAAQ,WAAW;EACnB,GAAI,OAAO,KAAK,aAAa,CAAC,SAAS,IAAI,EAAE,cAAc,GAAG,EAAE;EAChE,SAAS;GAAE,GAAG;GAAoB;GAAa;EAC/C,gBAAgB,EAAE;EAClB;EACA,aAAa,mBAAmB;GAAE;GAAQ;GAAc;GAAc,CAAC;EACvE,MAAM,EAAE;EACT,CAAC"}
package/package.json CHANGED
@@ -1,16 +1,17 @@
1
1
  {
2
2
  "name": "@prisma-next/mongo-contract-psl",
3
- "version": "0.4.0-dev.1",
3
+ "version": "0.4.0-dev.3",
4
4
  "type": "module",
5
5
  "sideEffects": false,
6
6
  "description": "PSL-to-Mongo ContractIR interpreter for Prisma Next",
7
7
  "dependencies": {
8
8
  "pathe": "^2.0.3",
9
- "@prisma-next/config": "0.4.0-dev.1",
10
- "@prisma-next/contract": "0.4.0-dev.1",
11
- "@prisma-next/mongo-contract": "0.4.0-dev.1",
12
- "@prisma-next/utils": "0.4.0-dev.1",
13
- "@prisma-next/psl-parser": "0.4.0-dev.1"
9
+ "@prisma-next/config": "0.4.0-dev.3",
10
+ "@prisma-next/mongo-contract": "0.4.0-dev.3",
11
+ "@prisma-next/contract": "0.4.0-dev.3",
12
+ "@prisma-next/psl-parser": "0.4.0-dev.3",
13
+ "@prisma-next/utils": "0.4.0-dev.3",
14
+ "@prisma-next/framework-components": "0.4.0-dev.3"
14
15
  },
15
16
  "devDependencies": {
16
17
  "tsdown": "0.18.4",
@@ -1,20 +1,22 @@
1
1
  import type { ContractField, ContractValueObject } from '@prisma-next/contract/types';
2
+ import type { CodecLookup } from '@prisma-next/framework-components/codec';
2
3
  import type { MongoStorageValidator } from '@prisma-next/mongo-contract';
3
4
 
4
- const CODEC_TO_BSON_TYPE: Record<string, string> = {
5
- 'mongo/string@1': 'string',
6
- 'mongo/int32@1': 'int',
7
- 'mongo/bool@1': 'bool',
8
- 'mongo/date@1': 'date',
9
- 'mongo/objectId@1': 'objectId',
10
- };
5
+ function resolveBsonType(
6
+ codecId: string,
7
+ codecLookup: CodecLookup | undefined,
8
+ ): string | undefined {
9
+ const codec = codecLookup?.get(codecId);
10
+ return codec?.targetTypes[0];
11
+ }
11
12
 
12
13
  function fieldToBsonSchema(
13
14
  field: ContractField,
14
15
  valueObjects: Record<string, ContractValueObject> | undefined,
16
+ codecLookup: CodecLookup | undefined,
15
17
  ): Record<string, unknown> | undefined {
16
18
  if (field.type.kind === 'scalar') {
17
- const bsonType = CODEC_TO_BSON_TYPE[field.type.codecId];
19
+ const bsonType = resolveBsonType(field.type.codecId, codecLookup);
18
20
  if (!bsonType) return undefined;
19
21
 
20
22
  if ('many' in field && field.many) {
@@ -31,7 +33,7 @@ function fieldToBsonSchema(
31
33
  if (field.type.kind === 'valueObject') {
32
34
  const vo = valueObjects?.[field.type.name];
33
35
  if (!vo) return undefined;
34
- const voSchema = deriveObjectSchema(vo.fields, valueObjects);
36
+ const voSchema = deriveObjectSchema(vo.fields, valueObjects, codecLookup);
35
37
  if ('many' in field && field.many) {
36
38
  return { bsonType: 'array', items: voSchema };
37
39
  }
@@ -47,12 +49,13 @@ function fieldToBsonSchema(
47
49
  function deriveObjectSchema(
48
50
  fields: Record<string, ContractField>,
49
51
  valueObjects: Record<string, ContractValueObject> | undefined,
52
+ codecLookup: CodecLookup | undefined,
50
53
  ): Record<string, unknown> {
51
54
  const properties: Record<string, unknown> = {};
52
55
  const required: string[] = [];
53
56
 
54
57
  for (const [fieldName, field] of Object.entries(fields)) {
55
- const schema = fieldToBsonSchema(field, valueObjects);
58
+ const schema = fieldToBsonSchema(field, valueObjects, codecLookup);
56
59
  if (schema) {
57
60
  properties[fieldName] = schema;
58
61
  if (!field.nullable) {
@@ -74,9 +77,74 @@ function deriveObjectSchema(
74
77
  export function deriveJsonSchema(
75
78
  fields: Record<string, ContractField>,
76
79
  valueObjects?: Record<string, ContractValueObject>,
80
+ codecLookup?: CodecLookup,
77
81
  ): MongoStorageValidator {
78
82
  return {
79
- jsonSchema: deriveObjectSchema(fields, valueObjects),
83
+ jsonSchema: deriveObjectSchema(fields, valueObjects, codecLookup),
84
+ validationLevel: 'strict',
85
+ validationAction: 'error',
86
+ };
87
+ }
88
+
89
+ export interface PolymorphicVariant {
90
+ readonly discriminatorValue: string;
91
+ readonly fields: Record<string, ContractField>;
92
+ }
93
+
94
+ export function derivePolymorphicJsonSchema(
95
+ baseFields: Record<string, ContractField>,
96
+ discriminatorField: string,
97
+ variants: readonly PolymorphicVariant[],
98
+ valueObjects?: Record<string, ContractValueObject>,
99
+ codecLookup?: CodecLookup,
100
+ ): MongoStorageValidator {
101
+ const baseSchema = deriveObjectSchema(baseFields, valueObjects, codecLookup);
102
+
103
+ const oneOf: Record<string, unknown>[] = [];
104
+ for (const variant of variants) {
105
+ const variantOnlyFields: Record<string, ContractField> = {};
106
+ for (const [name, field] of Object.entries(variant.fields)) {
107
+ if (!(name in baseFields)) {
108
+ variantOnlyFields[name] = field;
109
+ }
110
+ }
111
+
112
+ const entry: Record<string, unknown> = {
113
+ properties: {
114
+ [discriminatorField]: { enum: [variant.discriminatorValue] },
115
+ },
116
+ };
117
+
118
+ const variantProperties: Record<string, unknown> = {};
119
+ const variantRequired: string[] = [discriminatorField];
120
+ for (const [name, field] of Object.entries(variantOnlyFields)) {
121
+ const schema = fieldToBsonSchema(field, valueObjects, codecLookup);
122
+ if (schema) {
123
+ variantProperties[name] = schema;
124
+ if (!field.nullable) {
125
+ variantRequired.push(name);
126
+ }
127
+ }
128
+ }
129
+
130
+ if (Object.keys(variantProperties).length > 0) {
131
+ (entry['properties'] as Record<string, unknown>) = {
132
+ ...(entry['properties'] as Record<string, unknown>),
133
+ ...variantProperties,
134
+ };
135
+ }
136
+ entry['required'] = variantRequired.sort();
137
+
138
+ oneOf.push(entry);
139
+ }
140
+
141
+ const jsonSchema = { ...baseSchema };
142
+ if (oneOf.length > 0) {
143
+ jsonSchema['oneOf'] = oneOf;
144
+ }
145
+
146
+ return {
147
+ jsonSchema,
80
148
  validationLevel: 'strict',
81
149
  validationAction: 'error',
82
150
  };
@@ -2,4 +2,3 @@ export {
2
2
  type InterpretPslDocumentToMongoContractInput,
3
3
  interpretPslDocumentToMongoContract,
4
4
  } from '../interpreter';
5
- export { createMongoScalarTypeDescriptors } from '../scalar-type-descriptors';
@@ -9,10 +9,11 @@ import type {
9
9
  ContractReferenceRelation,
10
10
  ContractValueObject,
11
11
  } from '@prisma-next/contract/types';
12
+ import type { CodecLookup } from '@prisma-next/framework-components/codec';
12
13
  import type { MongoIndexKeyDirection, MongoStorageIndex } from '@prisma-next/mongo-contract';
13
14
  import type { ParsePslDocumentResult, PslField, PslModel } from '@prisma-next/psl-parser';
14
15
  import { notOk, ok, type Result } from '@prisma-next/utils/result';
15
- import { deriveJsonSchema } from './derive-json-schema';
16
+ import { deriveJsonSchema, derivePolymorphicJsonSchema } from './derive-json-schema';
16
17
  import {
17
18
  getAttribute,
18
19
  getMapName,
@@ -27,6 +28,7 @@ import {
27
28
  export interface InterpretPslDocumentToMongoContractInput {
28
29
  readonly document: ParsePslDocumentResult;
29
30
  readonly scalarTypeDescriptors: ReadonlyMap<string, string>;
31
+ readonly codecLookup?: CodecLookup;
30
32
  }
31
33
 
32
34
  interface FieldMappings {
@@ -146,6 +148,7 @@ function collectPolymorphismDeclarations(
146
148
  function resolvePolymorphism(input: {
147
149
  models: Record<string, MongoModelEntry>;
148
150
  roots: Record<string, string>;
151
+ collections: Record<string, Record<string, unknown>>;
149
152
  document: ParsePslDocumentResult;
150
153
  discriminatorDeclarations: Map<string, DiscriminatorDeclaration>;
151
154
  baseDeclarations: Map<string, BaseDeclaration>;
@@ -154,11 +157,13 @@ function resolvePolymorphism(input: {
154
157
  }): {
155
158
  models: Record<string, MongoModelEntry>;
156
159
  roots: Record<string, string>;
160
+ collections: Record<string, Record<string, unknown>>;
157
161
  diagnostics: ContractSourceDiagnostic[];
158
162
  } {
159
163
  const { discriminatorDeclarations, baseDeclarations, modelNames, sourceId, document } = input;
160
164
  let patched = input.models;
161
165
  let roots = input.roots;
166
+ let collections = input.collections;
162
167
  const diagnostics: ContractSourceDiagnostic[] = [];
163
168
 
164
169
  for (const [modelName, decl] of discriminatorDeclarations) {
@@ -175,7 +180,12 @@ function resolvePolymorphism(input: {
175
180
  const model = patched[modelName];
176
181
  if (!model) continue;
177
182
 
178
- if (!Object.hasOwn(model.fields, decl.fieldName)) {
183
+ const pslModel = document.ast.models.find((m) => m.name === modelName);
184
+ const mappedDiscriminatorField = pslModel
185
+ ? (resolveFieldMappings(pslModel).pslNameToMapped.get(decl.fieldName) ?? decl.fieldName)
186
+ : decl.fieldName;
187
+
188
+ if (!Object.hasOwn(model.fields, mappedDiscriminatorField)) {
179
189
  diagnostics.push({
180
190
  code: 'PSL_DISCRIMINATOR_FIELD_NOT_FOUND',
181
191
  message: `Discriminator field "${decl.fieldName}" is not a field on model "${modelName}"`,
@@ -203,7 +213,7 @@ function resolvePolymorphism(input: {
203
213
 
204
214
  patched = {
205
215
  ...patched,
206
- [modelName]: { ...model, discriminator: { field: decl.fieldName }, variants },
216
+ [modelName]: { ...model, discriminator: { field: mappedDiscriminatorField }, variants },
207
217
  };
208
218
  }
209
219
 
@@ -270,9 +280,45 @@ function resolvePolymorphism(input: {
270
280
  );
271
281
  }
272
282
  }
283
+
284
+ const variantColl = collections[variantCollectionName];
285
+ const variantIndexes = (variantColl?.['indexes'] ?? []) as MongoStorageIndex[];
286
+ const baseColl = collections[baseCollection];
287
+
288
+ if (variantCollectionName !== baseCollection) {
289
+ const filtered = Object.fromEntries(
290
+ Object.entries(collections).filter(([key]) => key !== variantCollectionName),
291
+ );
292
+ if (variantIndexes.length > 0 && baseColl) {
293
+ const baseIndexes = (baseColl['indexes'] ?? []) as MongoStorageIndex[];
294
+ collections = {
295
+ ...filtered,
296
+ [baseCollection]: { ...baseColl, indexes: [...baseIndexes, ...variantIndexes] },
297
+ };
298
+ } else {
299
+ collections = filtered;
300
+ }
301
+ } else if (variantIndexes.length > 0 && baseColl) {
302
+ const baseIndexes = (baseColl['indexes'] ?? []) as MongoStorageIndex[];
303
+ const mergedIndexes = [...baseIndexes];
304
+ for (const idx of variantIndexes) {
305
+ const isDuplicate = baseIndexes.some(
306
+ (existing) => JSON.stringify(existing) === JSON.stringify(idx),
307
+ );
308
+ if (!isDuplicate) {
309
+ mergedIndexes.push(idx);
310
+ }
311
+ }
312
+ if (mergedIndexes.length > baseIndexes.length) {
313
+ collections = {
314
+ ...collections,
315
+ [baseCollection]: { ...baseColl, indexes: mergedIndexes },
316
+ };
317
+ }
318
+ }
273
319
  }
274
320
 
275
- return { models: patched, roots, diagnostics };
321
+ return { models: patched, roots, collections, diagnostics };
276
322
  }
277
323
 
278
324
  function parseIndexDirection(raw: string | undefined): MongoIndexKeyDirection {
@@ -636,7 +682,7 @@ function resolveNonRelationField(
636
682
  export function interpretPslDocumentToMongoContract(
637
683
  input: InterpretPslDocumentToMongoContractInput,
638
684
  ): Result<Contract, ContractSourceDiagnostics> {
639
- const { document, scalarTypeDescriptors } = input;
685
+ const { document, scalarTypeDescriptors, codecLookup } = input;
640
686
  const sourceId = document.ast.sourceId;
641
687
  const diagnostics: ContractSourceDiagnostic[] = [];
642
688
  const modelNames = new Set(document.ast.models.map((m) => m.name));
@@ -738,7 +784,13 @@ export function interpretPslDocumentToMongoContract(
738
784
 
739
785
  models[pslModel.name] = { fields, relations, storage: { collection: collectionName } };
740
786
  const modelIndexes = collectIndexes(pslModel, fieldMappings, modelNames, sourceId, diagnostics);
741
- collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};
787
+ const existingColl = collections[collectionName];
788
+ if (existingColl && modelIndexes.length > 0) {
789
+ const existingIndexes = (existingColl['indexes'] ?? []) as MongoStorageIndex[];
790
+ collections[collectionName] = { indexes: [...existingIndexes, ...modelIndexes] };
791
+ } else if (!existingColl) {
792
+ collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};
793
+ }
742
794
  roots[collectionName] = pslModel.name;
743
795
  }
744
796
 
@@ -760,15 +812,6 @@ export function interpretPslDocumentToMongoContract(
760
812
  valueObjects[compositeType.name] = { fields };
761
813
  }
762
814
 
763
- for (const [, modelEntry] of Object.entries(models)) {
764
- const collectionName = modelEntry.storage.collection;
765
- const coll = collections[collectionName];
766
- if (coll) {
767
- const validator = deriveJsonSchema(modelEntry.fields, valueObjects);
768
- coll['validator'] = validator;
769
- }
770
- }
771
-
772
815
  const fkRelationsByPair = new Map<string, FkRelation[]>();
773
816
  for (const fk of allFkRelations) {
774
817
  const key = fkRelationPairKey(fk.declaringModel, fk.targetModel);
@@ -828,6 +871,7 @@ export function interpretPslDocumentToMongoContract(
828
871
  const polyResult = resolvePolymorphism({
829
872
  models,
830
873
  roots,
874
+ collections,
831
875
  document,
832
876
  discriminatorDeclarations,
833
877
  baseDeclarations,
@@ -842,9 +886,38 @@ export function interpretPslDocumentToMongoContract(
842
886
  });
843
887
  }
844
888
 
889
+ const resolvedModels = polyResult.models;
890
+ const resolvedCollections = polyResult.collections;
891
+
892
+ for (const [, modelEntry] of Object.entries(resolvedModels)) {
893
+ if (modelEntry.base) continue;
894
+
895
+ const collectionName = modelEntry.storage.collection;
896
+ const coll = resolvedCollections[collectionName];
897
+ if (!coll) continue;
898
+
899
+ if (modelEntry.discriminator && modelEntry.variants) {
900
+ const variantEntries = Object.entries(modelEntry.variants).map(
901
+ ([variantName, { value }]) => ({
902
+ discriminatorValue: value,
903
+ fields: resolvedModels[variantName]?.fields ?? {},
904
+ }),
905
+ );
906
+ coll['validator'] = derivePolymorphicJsonSchema(
907
+ modelEntry.fields,
908
+ modelEntry.discriminator.field,
909
+ variantEntries,
910
+ valueObjects,
911
+ codecLookup,
912
+ );
913
+ } else {
914
+ coll['validator'] = deriveJsonSchema(modelEntry.fields, valueObjects, codecLookup);
915
+ }
916
+ }
917
+
845
918
  const target = 'mongo';
846
919
  const targetFamily = 'mongo';
847
- const storageWithoutHash = { collections };
920
+ const storageWithoutHash = { collections: resolvedCollections };
848
921
  const storageHash = computeStorageHash({ target, targetFamily, storage: storageWithoutHash });
849
922
  const capabilities: Record<string, Record<string, boolean>> = {};
850
923
 
package/src/provider.ts CHANGED
@@ -5,16 +5,14 @@ import { ifDefined } from '@prisma-next/utils/defined';
5
5
  import { notOk, ok } from '@prisma-next/utils/result';
6
6
  import { resolve } from 'pathe';
7
7
  import { interpretPslDocumentToMongoContract } from './interpreter';
8
- import { createMongoScalarTypeDescriptors } from './scalar-type-descriptors';
9
8
 
10
9
  export interface MongoContractOptions {
11
10
  readonly output?: string;
12
- readonly scalarTypeDescriptors?: ReadonlyMap<string, string>;
13
11
  }
14
12
 
15
13
  export function mongoContract(schemaPath: string, options?: MongoContractOptions): ContractConfig {
16
14
  return {
17
- source: async (_context: ContractSourceContext) => {
15
+ source: async (context: ContractSourceContext) => {
18
16
  const absoluteSchemaPath = resolve(schemaPath);
19
17
  let schema: string;
20
18
  try {
@@ -41,7 +39,8 @@ export function mongoContract(schemaPath: string, options?: MongoContractOptions
41
39
 
42
40
  const interpreted = interpretPslDocumentToMongoContract({
43
41
  document,
44
- scalarTypeDescriptors: options?.scalarTypeDescriptors ?? createMongoScalarTypeDescriptors(),
42
+ scalarTypeDescriptors: context.scalarTypeDescriptors,
43
+ codecLookup: context.codecLookup,
45
44
  });
46
45
  if (!interpreted.ok) {
47
46
  return interpreted;
@@ -1 +0,0 @@
1
- {"version":3,"file":"scalar-type-descriptors-B0Lt0pJI.mjs","names":["CODEC_TO_BSON_TYPE: Record<string, string>","properties: Record<string, unknown>","required: string[]","result: Record<string, unknown>","parts: string[]","relationName: string | undefined","fieldsArg: PslAttributeArgument | undefined","referencesArg: PslAttributeArgument | undefined","diagnostics: ContractSourceDiagnostic[]","variants: Record<string, { readonly value: string }>","collation: Record<string, unknown>","result: Record<string, 0 | 1>","indexes: MongoStorageIndex[]","defaultDirection: MongoIndexKeyDirection","weights: Record<string, number> | undefined","index: MongoStorageIndex","result: ContractField","result","models: Record<string, MongoModelEntry>","collections: Record<string, Record<string, unknown>>","roots: Record<string, string>","allFkRelations: FkRelation[]","backrelationCandidates: BackrelationCandidate[]","fields: Record<string, ContractField>","relations: Record<string, ContractReferenceRelation>","valueObjects: Record<string, ContractValueObject>","capabilities: Record<string, Record<string, boolean>>"],"sources":["../src/derive-json-schema.ts","../src/psl-helpers.ts","../src/interpreter.ts","../src/scalar-type-descriptors.ts"],"sourcesContent":["import type { ContractField, ContractValueObject } from '@prisma-next/contract/types';\nimport type { MongoStorageValidator } from '@prisma-next/mongo-contract';\n\nconst CODEC_TO_BSON_TYPE: Record<string, string> = {\n 'mongo/string@1': 'string',\n 'mongo/int32@1': 'int',\n 'mongo/bool@1': 'bool',\n 'mongo/date@1': 'date',\n 'mongo/objectId@1': 'objectId',\n};\n\nfunction fieldToBsonSchema(\n field: ContractField,\n valueObjects: Record<string, ContractValueObject> | undefined,\n): Record<string, unknown> | undefined {\n if (field.type.kind === 'scalar') {\n const bsonType = CODEC_TO_BSON_TYPE[field.type.codecId];\n if (!bsonType) return undefined;\n\n if ('many' in field && field.many) {\n return { bsonType: 'array', items: { bsonType } };\n }\n\n if (field.nullable) {\n return { bsonType: ['null', bsonType] };\n }\n\n return { bsonType };\n }\n\n if (field.type.kind === 'valueObject') {\n const vo = valueObjects?.[field.type.name];\n if (!vo) return undefined;\n const voSchema = deriveObjectSchema(vo.fields, valueObjects);\n if ('many' in field && field.many) {\n return { bsonType: 'array', items: voSchema };\n }\n if (field.nullable) {\n return { oneOf: [{ bsonType: 'null' }, voSchema] };\n }\n return voSchema;\n }\n\n return undefined;\n}\n\nfunction deriveObjectSchema(\n fields: Record<string, ContractField>,\n valueObjects: Record<string, ContractValueObject> | undefined,\n): Record<string, unknown> {\n const properties: Record<string, unknown> = {};\n const required: string[] = [];\n\n for (const [fieldName, field] of Object.entries(fields)) {\n const schema = fieldToBsonSchema(field, valueObjects);\n if (schema) {\n properties[fieldName] = schema;\n if (!field.nullable) {\n required.push(fieldName);\n }\n }\n }\n\n const result: Record<string, unknown> = {\n bsonType: 'object',\n properties,\n };\n if (required.length > 0) {\n result['required'] = required.sort();\n }\n return result;\n}\n\nexport function deriveJsonSchema(\n fields: Record<string, ContractField>,\n valueObjects?: Record<string, ContractValueObject>,\n): MongoStorageValidator {\n return {\n jsonSchema: deriveObjectSchema(fields, valueObjects),\n validationLevel: 'strict',\n validationAction: 'error',\n };\n}\n","import type { PslAttribute, PslAttributeArgument } from '@prisma-next/psl-parser';\nimport { getPositionalArgument, parseQuotedStringLiteral } from '@prisma-next/psl-parser';\n\nexport { getPositionalArgument, parseQuotedStringLiteral };\n\nexport function getNamedArgument(attr: PslAttribute, name: string): string | undefined {\n const arg = attr.args.find((a) => a.kind === 'named' && a.name === name);\n return arg?.value;\n}\n\nexport function parseFieldList(value: string): readonly string[] {\n const inner = value.replace(/^\\[/, '').replace(/\\]$/, '').trim();\n if (inner.length === 0) return [];\n return splitTopLevel(inner).map((s) => s.trim());\n}\n\nexport interface ParsedIndexField {\n readonly name: string;\n readonly isWildcard: boolean;\n readonly direction?: number;\n}\n\nexport function parseIndexFieldList(value: string): readonly ParsedIndexField[] {\n const segments = parseFieldList(value);\n return segments.map(parseIndexFieldSegment);\n}\n\nfunction parseIndexFieldSegment(segment: string): ParsedIndexField {\n const wildcardMatch = segment.match(/^wildcard\\(\\s*(.*?)\\s*\\)$/);\n if (wildcardMatch) {\n const scope = wildcardMatch[1] ?? '';\n return {\n name: scope.length > 0 ? `${scope}.$**` : '$**',\n isWildcard: true,\n };\n }\n\n const modifierMatch = segment.match(/^(\\w+)\\(\\s*sort:\\s*(\\w+)\\s*\\)$/);\n if (modifierMatch) {\n const fieldName = modifierMatch[1] ?? segment;\n const sortValue = modifierMatch[2];\n return {\n name: fieldName,\n isWildcard: false,\n direction: sortValue === 'Desc' ? -1 : 1,\n };\n }\n\n return { name: segment, isWildcard: false };\n}\n\nfunction splitTopLevel(input: string): string[] {\n const parts: string[] = [];\n let depth = 0;\n let start = 0;\n for (let i = 0; i < input.length; i++) {\n const ch = input[i];\n if (ch === '(' || ch === '[' || ch === '{') depth++;\n else if (ch === ')' || ch === ']' || ch === '}') depth = Math.max(0, depth - 1);\n else if (ch === ',' && depth === 0) {\n parts.push(input.slice(start, i));\n start = i + 1;\n }\n }\n parts.push(input.slice(start));\n return parts;\n}\n\nexport function lowerFirst(value: string): string {\n if (value.length === 0) return value;\n return value[0]?.toLowerCase() + value.slice(1);\n}\n\nexport function getAttribute(\n attributes: readonly PslAttribute[],\n name: string,\n): PslAttribute | undefined {\n return attributes.find((attr) => attr.name === name);\n}\n\nexport function getMapName(attributes: readonly PslAttribute[]): string | undefined {\n const mapAttr = getAttribute(attributes, 'map');\n if (!mapAttr) return undefined;\n const arg = mapAttr.args[0];\n if (!arg) return undefined;\n return stripQuotes(arg.value);\n}\n\nexport interface ParsedRelationAttribute {\n readonly relationName?: string;\n readonly fields?: readonly string[];\n readonly references?: readonly string[];\n}\n\nexport function parseRelationAttribute(\n attributes: readonly PslAttribute[],\n): ParsedRelationAttribute | undefined {\n const relationAttr = getAttribute(attributes, 'relation');\n if (!relationAttr) return undefined;\n\n let relationName: string | undefined;\n let fieldsArg: PslAttributeArgument | undefined;\n let referencesArg: PslAttributeArgument | undefined;\n\n for (const arg of relationAttr.args) {\n if (arg.kind === 'positional') {\n relationName = stripQuotes(arg.value);\n } else if (arg.name === 'name') {\n relationName = stripQuotes(arg.value);\n } else if (arg.name === 'fields') {\n fieldsArg = arg;\n } else if (arg.name === 'references') {\n referencesArg = arg;\n }\n }\n\n const fields = fieldsArg ? parseFieldList(fieldsArg.value) : undefined;\n const references = referencesArg ? parseFieldList(referencesArg.value) : undefined;\n\n return {\n ...(relationName !== undefined ? { relationName } : {}),\n ...(fields !== undefined ? { fields } : {}),\n ...(references !== undefined ? { references } : {}),\n };\n}\n\nfunction stripQuotes(value: string): string {\n if (value.startsWith('\"') && value.endsWith('\"')) {\n return value.slice(1, -1);\n }\n return value;\n}\n","import type {\n ContractSourceDiagnostic,\n ContractSourceDiagnostics,\n} from '@prisma-next/config/config-types';\nimport { computeProfileHash, computeStorageHash } from '@prisma-next/contract/hashing';\nimport type {\n Contract,\n ContractField,\n ContractReferenceRelation,\n ContractValueObject,\n} from '@prisma-next/contract/types';\nimport type { MongoIndexKeyDirection, MongoStorageIndex } from '@prisma-next/mongo-contract';\nimport type { ParsePslDocumentResult, PslField, PslModel } from '@prisma-next/psl-parser';\nimport { notOk, ok, type Result } from '@prisma-next/utils/result';\nimport { deriveJsonSchema } from './derive-json-schema';\nimport {\n getAttribute,\n getMapName,\n getNamedArgument,\n getPositionalArgument,\n lowerFirst,\n parseIndexFieldList,\n parseQuotedStringLiteral,\n parseRelationAttribute,\n} from './psl-helpers';\n\nexport interface InterpretPslDocumentToMongoContractInput {\n readonly document: ParsePslDocumentResult;\n readonly scalarTypeDescriptors: ReadonlyMap<string, string>;\n}\n\ninterface FieldMappings {\n readonly pslNameToMapped: Map<string, string>;\n}\n\ninterface FkRelation {\n readonly declaringModel: string;\n readonly fieldName: string;\n readonly targetModel: string;\n readonly relationName?: string;\n readonly localFields: readonly string[];\n readonly targetFields: readonly string[];\n}\n\nfunction fkRelationPairKey(declaringModel: string, targetModel: string): string {\n return `${declaringModel}::${targetModel}`;\n}\n\nfunction resolveFieldMappings(model: PslModel): FieldMappings {\n const pslNameToMapped = new Map<string, string>();\n for (const field of model.fields) {\n const mapped = getMapName(field.attributes) ?? field.name;\n pslNameToMapped.set(field.name, mapped);\n }\n return { pslNameToMapped };\n}\n\nfunction resolveCollectionName(model: PslModel): string {\n return getMapName(model.attributes) ?? lowerFirst(model.name);\n}\n\ninterface MongoModelEntry {\n readonly fields: Record<string, ContractField>;\n readonly relations: Record<string, ContractReferenceRelation>;\n readonly storage: { readonly collection: string };\n readonly discriminator?: { readonly field: string };\n readonly variants?: Record<string, { readonly value: string }>;\n readonly base?: string;\n}\n\ntype DiscriminatorDeclaration = { readonly fieldName: string; readonly span: PslModel['span'] };\ntype BaseDeclaration = {\n readonly baseName: string;\n readonly value: string;\n readonly collectionName: string;\n readonly span: PslModel['span'];\n};\n\nfunction collectPolymorphismDeclarations(\n document: ParsePslDocumentResult,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): {\n discriminatorDeclarations: Map<string, DiscriminatorDeclaration>;\n baseDeclarations: Map<string, BaseDeclaration>;\n} {\n const discriminatorDeclarations = new Map<string, DiscriminatorDeclaration>();\n const baseDeclarations = new Map<string, BaseDeclaration>();\n\n for (const pslModel of document.ast.models) {\n for (const attr of pslModel.attributes) {\n if (attr.name === 'discriminator') {\n const fieldName = getPositionalArgument(attr);\n if (!fieldName) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@discriminator requires a field name argument`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const discField = pslModel.fields.find((f) => f.name === fieldName);\n if (discField && discField.typeName !== 'String') {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Discriminator field \"${fieldName}\" on model \"${pslModel.name}\" must be of type String, but is \"${discField.typeName}\"`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n discriminatorDeclarations.set(pslModel.name, { fieldName, span: attr.span });\n }\n if (attr.name === 'base') {\n const baseName = getPositionalArgument(attr, 0);\n const rawValue = getPositionalArgument(attr, 1);\n if (!baseName || !rawValue) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@base requires two arguments: base model name and discriminator value`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const value = parseQuotedStringLiteral(rawValue);\n if (value === undefined) {\n diagnostics.push({\n code: 'PSL_INVALID_ATTRIBUTE_ARGUMENT',\n message: `Model \"${pslModel.name}\" @@base discriminator value must be a quoted string literal`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n const collectionName = resolveCollectionName(pslModel);\n baseDeclarations.set(pslModel.name, { baseName, value, collectionName, span: attr.span });\n }\n }\n }\n\n return { discriminatorDeclarations, baseDeclarations };\n}\n\nfunction resolvePolymorphism(input: {\n models: Record<string, MongoModelEntry>;\n roots: Record<string, string>;\n document: ParsePslDocumentResult;\n discriminatorDeclarations: Map<string, DiscriminatorDeclaration>;\n baseDeclarations: Map<string, BaseDeclaration>;\n modelNames: ReadonlySet<string>;\n sourceId: string;\n}): {\n models: Record<string, MongoModelEntry>;\n roots: Record<string, string>;\n diagnostics: ContractSourceDiagnostic[];\n} {\n const { discriminatorDeclarations, baseDeclarations, modelNames, sourceId, document } = input;\n let patched = input.models;\n let roots = input.roots;\n const diagnostics: ContractSourceDiagnostic[] = [];\n\n for (const [modelName, decl] of discriminatorDeclarations) {\n if (baseDeclarations.has(modelName)) {\n diagnostics.push({\n code: 'PSL_DISCRIMINATOR_AND_BASE',\n message: `Model \"${modelName}\" cannot have both @@discriminator and @@base`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n const model = patched[modelName];\n if (!model) continue;\n\n if (!Object.hasOwn(model.fields, decl.fieldName)) {\n diagnostics.push({\n code: 'PSL_DISCRIMINATOR_FIELD_NOT_FOUND',\n message: `Discriminator field \"${decl.fieldName}\" is not a field on model \"${modelName}\"`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n const variants: Record<string, { readonly value: string }> = {};\n for (const [variantName, baseDecl] of baseDeclarations) {\n if (baseDecl.baseName !== modelName) continue;\n variants[variantName] = { value: baseDecl.value };\n }\n\n if (Object.keys(variants).length === 0) {\n diagnostics.push({\n code: 'PSL_ORPHANED_DISCRIMINATOR',\n message: `Model \"${modelName}\" has @@discriminator but no variant models declare @@base(${modelName}, ...)`,\n sourceId,\n span: decl.span,\n });\n continue;\n }\n\n patched = {\n ...patched,\n [modelName]: { ...model, discriminator: { field: decl.fieldName }, variants },\n };\n }\n\n for (const [variantName, baseDecl] of baseDeclarations) {\n if (!modelNames.has(baseDecl.baseName)) {\n diagnostics.push({\n code: 'PSL_BASE_TARGET_NOT_FOUND',\n message: `Model \"${variantName}\" @@base references non-existent model \"${baseDecl.baseName}\"`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n if (!discriminatorDeclarations.has(baseDecl.baseName)) {\n diagnostics.push({\n code: 'PSL_ORPHANED_BASE',\n message: `Model \"${variantName}\" declares @@base(${baseDecl.baseName}, ...) but \"${baseDecl.baseName}\" has no @@discriminator`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n if (discriminatorDeclarations.has(variantName)) {\n continue;\n }\n\n const baseModel = patched[baseDecl.baseName];\n const variantPslModel = document.ast.models.find((m) => m.name === variantName);\n if (!variantPslModel) continue;\n const hasExplicitMap = getMapName(variantPslModel.attributes) !== undefined;\n\n if (hasExplicitMap && baseModel && baseDecl.collectionName !== baseModel.storage.collection) {\n diagnostics.push({\n code: 'PSL_MONGO_VARIANT_SEPARATE_COLLECTION',\n message: `Mongo variant \"${variantName}\" cannot use a different collection than its base \"${baseDecl.baseName}\". Mongo only supports single-collection polymorphism.`,\n sourceId,\n span: baseDecl.span,\n });\n continue;\n }\n\n const baseCollection = baseModel?.storage.collection ?? baseDecl.collectionName;\n const variantModel = patched[variantName];\n if (variantModel) {\n patched = {\n ...patched,\n [variantName]: {\n ...variantModel,\n base: baseDecl.baseName,\n storage: { collection: baseCollection },\n },\n };\n }\n\n const variantCollectionName = resolveCollectionName(variantPslModel);\n if (roots[variantCollectionName] === variantName) {\n if (variantCollectionName === baseCollection && baseModel) {\n roots = { ...roots, [variantCollectionName]: baseDecl.baseName };\n } else {\n roots = Object.fromEntries(\n Object.entries(roots).filter(([key]) => key !== variantCollectionName),\n );\n }\n }\n }\n\n return { models: patched, roots, diagnostics };\n}\n\nfunction parseIndexDirection(raw: string | undefined): MongoIndexKeyDirection {\n if (!raw) return 1;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n const num = Number(stripped);\n if (num === 1 || num === -1) return num;\n if (['text', '2dsphere', '2d', 'hashed'].includes(stripped))\n return stripped as MongoIndexKeyDirection;\n return 1;\n}\n\nfunction parseNumericArg(raw: string | undefined): number | undefined {\n if (!raw) return undefined;\n const n = Number(raw);\n return Number.isFinite(n) ? n : undefined;\n}\n\nfunction parseBooleanArg(raw: string | undefined): boolean | undefined {\n if (raw === 'true') return true;\n if (raw === 'false') return false;\n return undefined;\n}\n\nfunction parseJsonArg(raw: string | undefined): Record<string, unknown> | undefined {\n if (!raw) return undefined;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '').replace(/\\\\\"/g, '\"');\n try {\n const parsed = JSON.parse(stripped);\n if (typeof parsed === 'object' && parsed !== null && !Array.isArray(parsed)) {\n return parsed as Record<string, unknown>;\n }\n } catch {\n // not valid JSON\n }\n return undefined;\n}\n\nfunction parseCollation(\n attr: import('@prisma-next/psl-parser').PslAttribute,\n): Record<string, unknown> | null | undefined {\n const locale = stripQuotesHelper(getNamedArgument(attr, 'collationLocale'));\n if (!locale) {\n const hasAnyCollationArg =\n getNamedArgument(attr, 'collationStrength') != null ||\n getNamedArgument(attr, 'collationCaseLevel') != null ||\n getNamedArgument(attr, 'collationCaseFirst') != null ||\n getNamedArgument(attr, 'collationNumericOrdering') != null ||\n getNamedArgument(attr, 'collationAlternate') != null ||\n getNamedArgument(attr, 'collationMaxVariable') != null ||\n getNamedArgument(attr, 'collationBackwards') != null ||\n getNamedArgument(attr, 'collationNormalization') != null;\n return hasAnyCollationArg ? null : undefined;\n }\n\n const collation: Record<string, unknown> = { locale };\n const strength = parseNumericArg(getNamedArgument(attr, 'collationStrength'));\n if (strength != null) collation['strength'] = strength;\n const caseLevel = parseBooleanArg(getNamedArgument(attr, 'collationCaseLevel'));\n if (caseLevel != null) collation['caseLevel'] = caseLevel;\n const caseFirst = stripQuotesHelper(getNamedArgument(attr, 'collationCaseFirst'));\n if (caseFirst != null) collation['caseFirst'] = caseFirst;\n const numericOrdering = parseBooleanArg(getNamedArgument(attr, 'collationNumericOrdering'));\n if (numericOrdering != null) collation['numericOrdering'] = numericOrdering;\n const alternate = stripQuotesHelper(getNamedArgument(attr, 'collationAlternate'));\n if (alternate != null) collation['alternate'] = alternate;\n const maxVariable = stripQuotesHelper(getNamedArgument(attr, 'collationMaxVariable'));\n if (maxVariable != null) collation['maxVariable'] = maxVariable;\n const backwards = parseBooleanArg(getNamedArgument(attr, 'collationBackwards'));\n if (backwards != null) collation['backwards'] = backwards;\n const normalization = parseBooleanArg(getNamedArgument(attr, 'collationNormalization'));\n if (normalization != null) collation['normalization'] = normalization;\n return collation;\n}\n\nfunction stripQuotesHelper(raw: string | undefined): string | undefined {\n if (!raw) return undefined;\n return raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n}\n\nfunction parseProjectionList(\n raw: string | undefined,\n value: 0 | 1,\n): Record<string, 0 | 1> | undefined {\n if (!raw) return undefined;\n const stripped = raw.replace(/^[\"']/, '').replace(/[\"']$/, '');\n const inner = stripped.replace(/^\\[/, '').replace(/\\]$/, '').trim();\n if (inner.length === 0) return undefined;\n const fields = inner\n .split(',')\n .map((s) => s.trim())\n .filter((s) => s.length > 0);\n const result: Record<string, 0 | 1> = {};\n for (const f of fields) {\n result[f] = value;\n }\n return result;\n}\n\nfunction collectIndexes(\n pslModel: PslModel,\n fieldMappings: FieldMappings,\n modelNames: ReadonlySet<string>,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): MongoStorageIndex[] {\n const indexes: MongoStorageIndex[] = [];\n let textIndexCount = 0;\n\n for (const field of pslModel.fields) {\n if (modelNames.has(field.typeName)) continue;\n const uniqueAttr = getAttribute(field.attributes, 'unique');\n if (!uniqueAttr) continue;\n const mappedName = fieldMappings.pslNameToMapped.get(field.name) ?? field.name;\n indexes.push({\n keys: [{ field: mappedName, direction: 1 }],\n unique: true,\n });\n }\n\n for (const attr of pslModel.attributes) {\n const isIndex = attr.name === 'index';\n const isUnique = attr.name === 'unique';\n const isTextIndex = attr.name === 'textIndex';\n if (!isIndex && !isUnique && !isTextIndex) continue;\n\n const fieldsArg = getPositionalArgument(attr, 0);\n if (!fieldsArg) continue;\n const parsedFields = parseIndexFieldList(fieldsArg);\n if (parsedFields.length === 0) continue;\n\n const hasWildcard = parsedFields.some((f) => f.isWildcard);\n const wildcardCount = parsedFields.filter((f) => f.isWildcard).length;\n\n if (wildcardCount > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'An index can contain at most one wildcard() field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (isUnique && hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Unique indexes cannot use wildcard() fields',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (isTextIndex) {\n textIndexCount++;\n if (textIndexCount > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: `Only one @@textIndex is allowed per collection (model \"${pslModel.name}\")`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message:\n 'wildcard() fields cannot be combined with type: hashed/2dsphere/2d or @@textIndex',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n }\n\n const typeArg = getNamedArgument(attr, 'type');\n const defaultDirection: MongoIndexKeyDirection = isTextIndex\n ? 'text'\n : parseIndexDirection(typeArg);\n\n if (\n hasWildcard &&\n typeof defaultDirection === 'string' &&\n ['hashed', '2dsphere', '2d'].includes(defaultDirection)\n ) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: `wildcard() fields cannot be combined with type: ${defaultDirection}`,\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if (defaultDirection === 'hashed' && parsedFields.length > 1) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Hashed indexes must have exactly one field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const keys = parsedFields.map((pf) => {\n const mappedName = pf.isWildcard\n ? pf.name.replace(/^(.+)\\.\\$\\*\\*$/, (_, prefix: string) => {\n const mapped = fieldMappings.pslNameToMapped.get(prefix);\n return mapped ? `${mapped}.$**` : `${prefix}.$**`;\n })\n : (fieldMappings.pslNameToMapped.get(pf.name) ?? pf.name);\n const direction: MongoIndexKeyDirection =\n pf.direction != null ? (pf.direction as MongoIndexKeyDirection) : defaultDirection;\n return { field: mappedName, direction };\n });\n\n const unique = isUnique ? true : undefined;\n const sparse = isTextIndex ? undefined : parseBooleanArg(getNamedArgument(attr, 'sparse'));\n const expireAfterSeconds = isTextIndex\n ? undefined\n : parseNumericArg(getNamedArgument(attr, 'expireAfterSeconds'));\n\n if (hasWildcard && expireAfterSeconds != null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'expireAfterSeconds cannot be combined with wildcard() fields',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const partialFilterExpression = parseJsonArg(getNamedArgument(attr, 'filter'));\n\n const includeArg = getNamedArgument(attr, 'include');\n const excludeArg = getNamedArgument(attr, 'exclude');\n\n if (includeArg != null && excludeArg != null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'Cannot specify both include and exclude on the same index',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n if ((includeArg != null || excludeArg != null) && !hasWildcard) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message:\n 'include/exclude options are only valid when the index contains a wildcard() field',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const wildcardProjection =\n includeArg != null\n ? parseProjectionList(includeArg, 1)\n : excludeArg != null\n ? parseProjectionList(excludeArg, 0)\n : undefined;\n\n const collation = parseCollation(attr);\n if (collation === null) {\n diagnostics.push({\n code: 'PSL_INVALID_INDEX',\n message: 'collationLocale is required when using collation options',\n sourceId,\n span: attr.span,\n });\n continue;\n }\n\n const rawWeights = parseJsonArg(getNamedArgument(attr, 'weights'));\n let weights: Record<string, number> | undefined;\n if (rawWeights) {\n weights = {};\n for (const [k, v] of Object.entries(rawWeights)) {\n if (typeof v === 'number') weights[k] = v;\n }\n }\n\n const rawDefaultLang = isTextIndex\n ? getNamedArgument(attr, 'language')\n : getNamedArgument(attr, 'default_language');\n const default_language = stripQuotesHelper(rawDefaultLang);\n\n const rawLangOverride = getNamedArgument(attr, 'languageOverride');\n const language_override = stripQuotesHelper(rawLangOverride);\n\n const index: MongoStorageIndex = {\n keys,\n ...(unique != null && { unique }),\n ...(sparse != null && { sparse }),\n ...(expireAfterSeconds != null && { expireAfterSeconds }),\n ...(partialFilterExpression != null && { partialFilterExpression }),\n ...(wildcardProjection != null && { wildcardProjection }),\n ...(collation != null && { collation }),\n ...(weights != null && { weights }),\n ...(default_language != null && { default_language }),\n ...(language_override != null && { language_override }),\n };\n\n indexes.push(index);\n }\n\n return indexes;\n}\n\nfunction isRelationField(field: PslField, modelNames: ReadonlySet<string>): boolean {\n return modelNames.has(field.typeName);\n}\n\nfunction resolveFieldCodecId(\n field: PslField,\n scalarTypeDescriptors: ReadonlyMap<string, string>,\n): string | undefined {\n return scalarTypeDescriptors.get(field.typeName);\n}\n\nfunction resolveNonRelationField(\n field: PslField,\n ownerName: string,\n compositeTypeNames: ReadonlySet<string>,\n scalarTypeDescriptors: ReadonlyMap<string, string>,\n sourceId: string,\n diagnostics: ContractSourceDiagnostic[],\n): ContractField | undefined {\n if (compositeTypeNames.has(field.typeName)) {\n const result: ContractField = {\n type: { kind: 'valueObject', name: field.typeName },\n nullable: field.optional,\n };\n return field.list ? { ...result, many: true } : result;\n }\n\n const codecId = resolveFieldCodecId(field, scalarTypeDescriptors);\n if (!codecId) {\n diagnostics.push({\n code: 'PSL_UNSUPPORTED_FIELD_TYPE',\n message: `Field \"${ownerName}.${field.name}\" type \"${field.typeName}\" is not supported in Mongo PSL interpreter`,\n sourceId,\n span: field.span,\n });\n return undefined;\n }\n\n const result: ContractField = {\n type: { kind: 'scalar', codecId },\n nullable: field.optional,\n };\n return field.list ? { ...result, many: true } : result;\n}\n\nexport function interpretPslDocumentToMongoContract(\n input: InterpretPslDocumentToMongoContractInput,\n): Result<Contract, ContractSourceDiagnostics> {\n const { document, scalarTypeDescriptors } = input;\n const sourceId = document.ast.sourceId;\n const diagnostics: ContractSourceDiagnostic[] = [];\n const modelNames = new Set(document.ast.models.map((m) => m.name));\n const compositeTypeNames = new Set(document.ast.compositeTypes.map((ct) => ct.name));\n\n const models: Record<string, MongoModelEntry> = {};\n const collections: Record<string, Record<string, unknown>> = {};\n const roots: Record<string, string> = {};\n const allFkRelations: FkRelation[] = [];\n\n interface BackrelationCandidate {\n readonly modelName: string;\n readonly fieldName: string;\n readonly targetModelName: string;\n readonly relationName?: string;\n readonly cardinality: '1:1' | '1:N';\n readonly field: PslField;\n }\n const backrelationCandidates: BackrelationCandidate[] = [];\n\n for (const pslModel of document.ast.models) {\n const collectionName = resolveCollectionName(pslModel);\n const fieldMappings = resolveFieldMappings(pslModel);\n\n const fields: Record<string, ContractField> = {};\n const relations: Record<string, ContractReferenceRelation> = {};\n\n for (const field of pslModel.fields) {\n if (isRelationField(field, modelNames)) {\n const relation = parseRelationAttribute(field.attributes);\n\n if (field.list || !(relation?.fields && relation?.references)) {\n backrelationCandidates.push({\n modelName: pslModel.name,\n fieldName: field.name,\n targetModelName: field.typeName,\n ...(relation?.relationName !== undefined\n ? { relationName: relation.relationName }\n : {}),\n cardinality: field.list ? '1:N' : '1:1',\n field,\n });\n continue;\n }\n\n if (relation?.fields && relation?.references) {\n const localMapped = relation.fields.map((f) => fieldMappings.pslNameToMapped.get(f) ?? f);\n\n const targetModel = document.ast.models.find((m) => m.name === field.typeName);\n const targetFieldMappings = targetModel ? resolveFieldMappings(targetModel) : undefined;\n const targetMapped = relation.references.map(\n (f) => targetFieldMappings?.pslNameToMapped.get(f) ?? f,\n );\n\n relations[field.name] = {\n to: field.typeName,\n cardinality: 'N:1' as const,\n on: {\n localFields: localMapped,\n targetFields: targetMapped,\n },\n };\n\n allFkRelations.push({\n declaringModel: pslModel.name,\n fieldName: field.name,\n targetModel: field.typeName,\n ...(relation.relationName !== undefined ? { relationName: relation.relationName } : {}),\n localFields: localMapped,\n targetFields: targetMapped,\n });\n }\n continue;\n }\n\n const resolved = resolveNonRelationField(\n field,\n pslModel.name,\n compositeTypeNames,\n scalarTypeDescriptors,\n sourceId,\n diagnostics,\n );\n if (!resolved) continue;\n\n const mappedName = fieldMappings.pslNameToMapped.get(field.name) ?? field.name;\n fields[mappedName] = resolved;\n }\n\n const isVariantModel = pslModel.attributes.some((attr) => attr.name === 'base');\n const hasIdField = pslModel.fields.some((f) => getAttribute(f.attributes, 'id') !== undefined);\n if (!hasIdField && !isVariantModel) {\n diagnostics.push({\n code: 'PSL_MISSING_ID_FIELD',\n message: `Model \"${pslModel.name}\" has no field with @id attribute. Every model must have exactly one @id field.`,\n sourceId,\n });\n }\n\n models[pslModel.name] = { fields, relations, storage: { collection: collectionName } };\n const modelIndexes = collectIndexes(pslModel, fieldMappings, modelNames, sourceId, diagnostics);\n collections[collectionName] = modelIndexes.length > 0 ? { indexes: modelIndexes } : {};\n roots[collectionName] = pslModel.name;\n }\n\n const valueObjects: Record<string, ContractValueObject> = {};\n for (const compositeType of document.ast.compositeTypes) {\n const fields: Record<string, ContractField> = {};\n for (const field of compositeType.fields) {\n const resolved = resolveNonRelationField(\n field,\n compositeType.name,\n compositeTypeNames,\n scalarTypeDescriptors,\n sourceId,\n diagnostics,\n );\n if (!resolved) continue;\n fields[field.name] = resolved;\n }\n valueObjects[compositeType.name] = { fields };\n }\n\n for (const [, modelEntry] of Object.entries(models)) {\n const collectionName = modelEntry.storage.collection;\n const coll = collections[collectionName];\n if (coll) {\n const validator = deriveJsonSchema(modelEntry.fields, valueObjects);\n coll['validator'] = validator;\n }\n }\n\n const fkRelationsByPair = new Map<string, FkRelation[]>();\n for (const fk of allFkRelations) {\n const key = fkRelationPairKey(fk.declaringModel, fk.targetModel);\n const existing = fkRelationsByPair.get(key);\n if (existing) {\n existing.push(fk);\n } else {\n fkRelationsByPair.set(key, [fk]);\n }\n }\n\n for (const candidate of backrelationCandidates) {\n const pairKey = fkRelationPairKey(candidate.targetModelName, candidate.modelName);\n const pairMatches = fkRelationsByPair.get(pairKey) ?? [];\n const matches = candidate.relationName\n ? pairMatches.filter((r) => r.relationName === candidate.relationName)\n : [...pairMatches];\n\n if (matches.length === 0) {\n diagnostics.push({\n code: 'PSL_ORPHANED_BACKRELATION',\n message: `Backrelation list field \"${candidate.modelName}.${candidate.fieldName}\" has no matching FK-side relation on model \"${candidate.targetModelName}\". Add @relation(fields: [...], references: [...]) on the FK-side relation or use an explicit join model for many-to-many.`,\n sourceId,\n span: candidate.field.span,\n });\n continue;\n }\n if (matches.length > 1) {\n diagnostics.push({\n code: 'PSL_AMBIGUOUS_BACKRELATION',\n message: `Backrelation list field \"${candidate.modelName}.${candidate.fieldName}\" matches multiple FK-side relations on model \"${candidate.targetModelName}\". Add @relation(\"...\") to both sides to disambiguate.`,\n sourceId,\n span: candidate.field.span,\n });\n continue;\n }\n\n const fk = matches[0];\n if (!fk) continue;\n const modelEntry = models[candidate.modelName];\n if (!modelEntry) continue;\n modelEntry.relations[candidate.fieldName] = {\n to: candidate.targetModelName,\n cardinality: candidate.cardinality,\n on: {\n localFields: fk.targetFields,\n targetFields: fk.localFields,\n },\n };\n }\n\n const { discriminatorDeclarations, baseDeclarations } = collectPolymorphismDeclarations(\n document,\n sourceId,\n diagnostics,\n );\n const polyResult = resolvePolymorphism({\n models,\n roots,\n document,\n discriminatorDeclarations,\n baseDeclarations,\n modelNames,\n sourceId,\n });\n\n if (diagnostics.length > 0 || polyResult.diagnostics.length > 0) {\n return notOk({\n summary: 'PSL to Mongo contract interpretation failed',\n diagnostics: [...diagnostics, ...polyResult.diagnostics],\n });\n }\n\n const target = 'mongo';\n const targetFamily = 'mongo';\n const storageWithoutHash = { collections };\n const storageHash = computeStorageHash({ target, targetFamily, storage: storageWithoutHash });\n const capabilities: Record<string, Record<string, boolean>> = {};\n\n return ok({\n targetFamily,\n target,\n roots: polyResult.roots,\n models: polyResult.models,\n ...(Object.keys(valueObjects).length > 0 ? { valueObjects } : {}),\n storage: { ...storageWithoutHash, storageHash },\n extensionPacks: {},\n capabilities,\n profileHash: computeProfileHash({ target, targetFamily, capabilities }),\n meta: {},\n });\n}\n","export function createMongoScalarTypeDescriptors(): ReadonlyMap<string, string> {\n return new Map<string, string>([\n ['String', 'mongo/string@1'],\n ['Int', 'mongo/int32@1'],\n ['Boolean', 'mongo/bool@1'],\n ['DateTime', 'mongo/date@1'],\n ['ObjectId', 'mongo/objectId@1'],\n ]);\n}\n"],"mappings":";;;;;AAGA,MAAMA,qBAA6C;CACjD,kBAAkB;CAClB,iBAAiB;CACjB,gBAAgB;CAChB,gBAAgB;CAChB,oBAAoB;CACrB;AAED,SAAS,kBACP,OACA,cACqC;AACrC,KAAI,MAAM,KAAK,SAAS,UAAU;EAChC,MAAM,WAAW,mBAAmB,MAAM,KAAK;AAC/C,MAAI,CAAC,SAAU,QAAO;AAEtB,MAAI,UAAU,SAAS,MAAM,KAC3B,QAAO;GAAE,UAAU;GAAS,OAAO,EAAE,UAAU;GAAE;AAGnD,MAAI,MAAM,SACR,QAAO,EAAE,UAAU,CAAC,QAAQ,SAAS,EAAE;AAGzC,SAAO,EAAE,UAAU;;AAGrB,KAAI,MAAM,KAAK,SAAS,eAAe;EACrC,MAAM,KAAK,eAAe,MAAM,KAAK;AACrC,MAAI,CAAC,GAAI,QAAO;EAChB,MAAM,WAAW,mBAAmB,GAAG,QAAQ,aAAa;AAC5D,MAAI,UAAU,SAAS,MAAM,KAC3B,QAAO;GAAE,UAAU;GAAS,OAAO;GAAU;AAE/C,MAAI,MAAM,SACR,QAAO,EAAE,OAAO,CAAC,EAAE,UAAU,QAAQ,EAAE,SAAS,EAAE;AAEpD,SAAO;;;AAMX,SAAS,mBACP,QACA,cACyB;CACzB,MAAMC,aAAsC,EAAE;CAC9C,MAAMC,WAAqB,EAAE;AAE7B,MAAK,MAAM,CAAC,WAAW,UAAU,OAAO,QAAQ,OAAO,EAAE;EACvD,MAAM,SAAS,kBAAkB,OAAO,aAAa;AACrD,MAAI,QAAQ;AACV,cAAW,aAAa;AACxB,OAAI,CAAC,MAAM,SACT,UAAS,KAAK,UAAU;;;CAK9B,MAAMC,SAAkC;EACtC,UAAU;EACV;EACD;AACD,KAAI,SAAS,SAAS,EACpB,QAAO,cAAc,SAAS,MAAM;AAEtC,QAAO;;AAGT,SAAgB,iBACd,QACA,cACuB;AACvB,QAAO;EACL,YAAY,mBAAmB,QAAQ,aAAa;EACpD,iBAAiB;EACjB,kBAAkB;EACnB;;;;;AC5EH,SAAgB,iBAAiB,MAAoB,MAAkC;AAErF,QADY,KAAK,KAAK,MAAM,MAAM,EAAE,SAAS,WAAW,EAAE,SAAS,KAAK,EAC5D;;AAGd,SAAgB,eAAe,OAAkC;CAC/D,MAAM,QAAQ,MAAM,QAAQ,OAAO,GAAG,CAAC,QAAQ,OAAO,GAAG,CAAC,MAAM;AAChE,KAAI,MAAM,WAAW,EAAG,QAAO,EAAE;AACjC,QAAO,cAAc,MAAM,CAAC,KAAK,MAAM,EAAE,MAAM,CAAC;;AASlD,SAAgB,oBAAoB,OAA4C;AAE9E,QADiB,eAAe,MAAM,CACtB,IAAI,uBAAuB;;AAG7C,SAAS,uBAAuB,SAAmC;CACjE,MAAM,gBAAgB,QAAQ,MAAM,4BAA4B;AAChE,KAAI,eAAe;EACjB,MAAM,QAAQ,cAAc,MAAM;AAClC,SAAO;GACL,MAAM,MAAM,SAAS,IAAI,GAAG,MAAM,QAAQ;GAC1C,YAAY;GACb;;CAGH,MAAM,gBAAgB,QAAQ,MAAM,iCAAiC;AACrE,KAAI,cAGF,QAAO;EACL,MAHgB,cAAc,MAAM;EAIpC,YAAY;EACZ,WAJgB,cAAc,OAIL,SAAS,KAAK;EACxC;AAGH,QAAO;EAAE,MAAM;EAAS,YAAY;EAAO;;AAG7C,SAAS,cAAc,OAAyB;CAC9C,MAAMC,QAAkB,EAAE;CAC1B,IAAI,QAAQ;CACZ,IAAI,QAAQ;AACZ,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;EACrC,MAAM,KAAK,MAAM;AACjB,MAAI,OAAO,OAAO,OAAO,OAAO,OAAO,IAAK;WACnC,OAAO,OAAO,OAAO,OAAO,OAAO,IAAK,SAAQ,KAAK,IAAI,GAAG,QAAQ,EAAE;WACtE,OAAO,OAAO,UAAU,GAAG;AAClC,SAAM,KAAK,MAAM,MAAM,OAAO,EAAE,CAAC;AACjC,WAAQ,IAAI;;;AAGhB,OAAM,KAAK,MAAM,MAAM,MAAM,CAAC;AAC9B,QAAO;;AAGT,SAAgB,WAAW,OAAuB;AAChD,KAAI,MAAM,WAAW,EAAG,QAAO;AAC/B,QAAO,MAAM,IAAI,aAAa,GAAG,MAAM,MAAM,EAAE;;AAGjD,SAAgB,aACd,YACA,MAC0B;AAC1B,QAAO,WAAW,MAAM,SAAS,KAAK,SAAS,KAAK;;AAGtD,SAAgB,WAAW,YAAyD;CAClF,MAAM,UAAU,aAAa,YAAY,MAAM;AAC/C,KAAI,CAAC,QAAS,QAAO;CACrB,MAAM,MAAM,QAAQ,KAAK;AACzB,KAAI,CAAC,IAAK,QAAO;AACjB,QAAO,YAAY,IAAI,MAAM;;AAS/B,SAAgB,uBACd,YACqC;CACrC,MAAM,eAAe,aAAa,YAAY,WAAW;AACzD,KAAI,CAAC,aAAc,QAAO;CAE1B,IAAIC;CACJ,IAAIC;CACJ,IAAIC;AAEJ,MAAK,MAAM,OAAO,aAAa,KAC7B,KAAI,IAAI,SAAS,aACf,gBAAe,YAAY,IAAI,MAAM;UAC5B,IAAI,SAAS,OACtB,gBAAe,YAAY,IAAI,MAAM;UAC5B,IAAI,SAAS,SACtB,aAAY;UACH,IAAI,SAAS,aACtB,iBAAgB;CAIpB,MAAM,SAAS,YAAY,eAAe,UAAU,MAAM,GAAG;CAC7D,MAAM,aAAa,gBAAgB,eAAe,cAAc,MAAM,GAAG;AAEzE,QAAO;EACL,GAAI,iBAAiB,SAAY,EAAE,cAAc,GAAG,EAAE;EACtD,GAAI,WAAW,SAAY,EAAE,QAAQ,GAAG,EAAE;EAC1C,GAAI,eAAe,SAAY,EAAE,YAAY,GAAG,EAAE;EACnD;;AAGH,SAAS,YAAY,OAAuB;AAC1C,KAAI,MAAM,WAAW,KAAI,IAAI,MAAM,SAAS,KAAI,CAC9C,QAAO,MAAM,MAAM,GAAG,GAAG;AAE3B,QAAO;;;;;ACtFT,SAAS,kBAAkB,gBAAwB,aAA6B;AAC9E,QAAO,GAAG,eAAe,IAAI;;AAG/B,SAAS,qBAAqB,OAAgC;CAC5D,MAAM,kCAAkB,IAAI,KAAqB;AACjD,MAAK,MAAM,SAAS,MAAM,QAAQ;EAChC,MAAM,SAAS,WAAW,MAAM,WAAW,IAAI,MAAM;AACrD,kBAAgB,IAAI,MAAM,MAAM,OAAO;;AAEzC,QAAO,EAAE,iBAAiB;;AAG5B,SAAS,sBAAsB,OAAyB;AACtD,QAAO,WAAW,MAAM,WAAW,IAAI,WAAW,MAAM,KAAK;;AAoB/D,SAAS,gCACP,UACA,UACA,aAIA;CACA,MAAM,4CAA4B,IAAI,KAAuC;CAC7E,MAAM,mCAAmB,IAAI,KAA8B;AAE3D,MAAK,MAAM,YAAY,SAAS,IAAI,OAClC,MAAK,MAAM,QAAQ,SAAS,YAAY;AACtC,MAAI,KAAK,SAAS,iBAAiB;GACjC,MAAM,YAAY,sBAAsB,KAAK;AAC7C,OAAI,CAAC,WAAW;AACd,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,YAAY,SAAS,OAAO,MAAM,MAAM,EAAE,SAAS,UAAU;AACnE,OAAI,aAAa,UAAU,aAAa,UAAU;AAChD,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,wBAAwB,UAAU,cAAc,SAAS,KAAK,oCAAoC,UAAU,SAAS;KAC9H;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;AAEF,6BAA0B,IAAI,SAAS,MAAM;IAAE;IAAW,MAAM,KAAK;IAAM,CAAC;;AAE9E,MAAI,KAAK,SAAS,QAAQ;GACxB,MAAM,WAAW,sBAAsB,MAAM,EAAE;GAC/C,MAAM,WAAW,sBAAsB,MAAM,EAAE;AAC/C,OAAI,CAAC,YAAY,CAAC,UAAU;AAC1B,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,QAAQ,yBAAyB,SAAS;AAChD,OAAI,UAAU,QAAW;AACvB,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,UAAU,SAAS,KAAK;KACjC;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;GAEF,MAAM,iBAAiB,sBAAsB,SAAS;AACtD,oBAAiB,IAAI,SAAS,MAAM;IAAE;IAAU;IAAO;IAAgB,MAAM,KAAK;IAAM,CAAC;;;AAK/F,QAAO;EAAE;EAA2B;EAAkB;;AAGxD,SAAS,oBAAoB,OAY3B;CACA,MAAM,EAAE,2BAA2B,kBAAkB,YAAY,UAAU,aAAa;CACxF,IAAI,UAAU,MAAM;CACpB,IAAI,QAAQ,MAAM;CAClB,MAAMC,cAA0C,EAAE;AAElD,MAAK,MAAM,CAAC,WAAW,SAAS,2BAA2B;AACzD,MAAI,iBAAiB,IAAI,UAAU,EAAE;AACnC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,UAAU;IAC7B;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,QAAQ,QAAQ;AACtB,MAAI,CAAC,MAAO;AAEZ,MAAI,CAAC,OAAO,OAAO,MAAM,QAAQ,KAAK,UAAU,EAAE;AAChD,eAAY,KAAK;IACf,MAAM;IACN,SAAS,wBAAwB,KAAK,UAAU,6BAA6B,UAAU;IACvF;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAMC,WAAuD,EAAE;AAC/D,OAAK,MAAM,CAAC,aAAa,aAAa,kBAAkB;AACtD,OAAI,SAAS,aAAa,UAAW;AACrC,YAAS,eAAe,EAAE,OAAO,SAAS,OAAO;;AAGnD,MAAI,OAAO,KAAK,SAAS,CAAC,WAAW,GAAG;AACtC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,UAAU,6DAA6D,UAAU;IACpG;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,YAAU;GACR,GAAG;IACF,YAAY;IAAE,GAAG;IAAO,eAAe,EAAE,OAAO,KAAK,WAAW;IAAE;IAAU;GAC9E;;AAGH,MAAK,MAAM,CAAC,aAAa,aAAa,kBAAkB;AACtD,MAAI,CAAC,WAAW,IAAI,SAAS,SAAS,EAAE;AACtC,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,YAAY,0CAA0C,SAAS,SAAS;IAC3F;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;AAGF,MAAI,CAAC,0BAA0B,IAAI,SAAS,SAAS,EAAE;AACrD,eAAY,KAAK;IACf,MAAM;IACN,SAAS,UAAU,YAAY,oBAAoB,SAAS,SAAS,cAAc,SAAS,SAAS;IACrG;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;AAGF,MAAI,0BAA0B,IAAI,YAAY,CAC5C;EAGF,MAAM,YAAY,QAAQ,SAAS;EACnC,MAAM,kBAAkB,SAAS,IAAI,OAAO,MAAM,MAAM,EAAE,SAAS,YAAY;AAC/E,MAAI,CAAC,gBAAiB;AAGtB,MAFuB,WAAW,gBAAgB,WAAW,KAAK,UAE5C,aAAa,SAAS,mBAAmB,UAAU,QAAQ,YAAY;AAC3F,eAAY,KAAK;IACf,MAAM;IACN,SAAS,kBAAkB,YAAY,qDAAqD,SAAS,SAAS;IAC9G;IACA,MAAM,SAAS;IAChB,CAAC;AACF;;EAGF,MAAM,iBAAiB,WAAW,QAAQ,cAAc,SAAS;EACjE,MAAM,eAAe,QAAQ;AAC7B,MAAI,aACF,WAAU;GACR,GAAG;IACF,cAAc;IACb,GAAG;IACH,MAAM,SAAS;IACf,SAAS,EAAE,YAAY,gBAAgB;IACxC;GACF;EAGH,MAAM,wBAAwB,sBAAsB,gBAAgB;AACpE,MAAI,MAAM,2BAA2B,YACnC,KAAI,0BAA0B,kBAAkB,UAC9C,SAAQ;GAAE,GAAG;IAAQ,wBAAwB,SAAS;GAAU;MAEhE,SAAQ,OAAO,YACb,OAAO,QAAQ,MAAM,CAAC,QAAQ,CAAC,SAAS,QAAQ,sBAAsB,CACvE;;AAKP,QAAO;EAAE,QAAQ;EAAS;EAAO;EAAa;;AAGhD,SAAS,oBAAoB,KAAiD;AAC5E,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,WAAW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG;CAC9D,MAAM,MAAM,OAAO,SAAS;AAC5B,KAAI,QAAQ,KAAK,QAAQ,GAAI,QAAO;AACpC,KAAI;EAAC;EAAQ;EAAY;EAAM;EAAS,CAAC,SAAS,SAAS,CACzD,QAAO;AACT,QAAO;;AAGT,SAAS,gBAAgB,KAA6C;AACpE,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,IAAI,OAAO,IAAI;AACrB,QAAO,OAAO,SAAS,EAAE,GAAG,IAAI;;AAGlC,SAAS,gBAAgB,KAA8C;AACrE,KAAI,QAAQ,OAAQ,QAAO;AAC3B,KAAI,QAAQ,QAAS,QAAO;;AAI9B,SAAS,aAAa,KAA8D;AAClF,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,WAAW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG,CAAC,QAAQ,QAAQ,KAAI;AACnF,KAAI;EACF,MAAM,SAAS,KAAK,MAAM,SAAS;AACnC,MAAI,OAAO,WAAW,YAAY,WAAW,QAAQ,CAAC,MAAM,QAAQ,OAAO,CACzE,QAAO;SAEH;;AAMV,SAAS,eACP,MAC4C;CAC5C,MAAM,SAAS,kBAAkB,iBAAiB,MAAM,kBAAkB,CAAC;AAC3E,KAAI,CAAC,OAUH,QARE,iBAAiB,MAAM,oBAAoB,IAAI,QAC/C,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,2BAA2B,IAAI,QACtD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,uBAAuB,IAAI,QAClD,iBAAiB,MAAM,qBAAqB,IAAI,QAChD,iBAAiB,MAAM,yBAAyB,IAAI,OAC1B,OAAO;CAGrC,MAAMC,YAAqC,EAAE,QAAQ;CACrD,MAAM,WAAW,gBAAgB,iBAAiB,MAAM,oBAAoB,CAAC;AAC7E,KAAI,YAAY,KAAM,WAAU,cAAc;CAC9C,MAAM,YAAY,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAC/E,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,YAAY,kBAAkB,iBAAiB,MAAM,qBAAqB,CAAC;AACjF,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,kBAAkB,gBAAgB,iBAAiB,MAAM,2BAA2B,CAAC;AAC3F,KAAI,mBAAmB,KAAM,WAAU,qBAAqB;CAC5D,MAAM,YAAY,kBAAkB,iBAAiB,MAAM,qBAAqB,CAAC;AACjF,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,cAAc,kBAAkB,iBAAiB,MAAM,uBAAuB,CAAC;AACrF,KAAI,eAAe,KAAM,WAAU,iBAAiB;CACpD,MAAM,YAAY,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAC/E,KAAI,aAAa,KAAM,WAAU,eAAe;CAChD,MAAM,gBAAgB,gBAAgB,iBAAiB,MAAM,yBAAyB,CAAC;AACvF,KAAI,iBAAiB,KAAM,WAAU,mBAAmB;AACxD,QAAO;;AAGT,SAAS,kBAAkB,KAA6C;AACtE,KAAI,CAAC,IAAK,QAAO;AACjB,QAAO,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG;;AAGtD,SAAS,oBACP,KACA,OACmC;AACnC,KAAI,CAAC,IAAK,QAAO;CAEjB,MAAM,QADW,IAAI,QAAQ,SAAS,GAAG,CAAC,QAAQ,SAAS,GAAG,CACvC,QAAQ,OAAO,GAAG,CAAC,QAAQ,OAAO,GAAG,CAAC,MAAM;AACnE,KAAI,MAAM,WAAW,EAAG,QAAO;CAC/B,MAAM,SAAS,MACZ,MAAM,IAAI,CACV,KAAK,MAAM,EAAE,MAAM,CAAC,CACpB,QAAQ,MAAM,EAAE,SAAS,EAAE;CAC9B,MAAMC,SAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,OACd,QAAO,KAAK;AAEd,QAAO;;AAGT,SAAS,eACP,UACA,eACA,YACA,UACA,aACqB;CACrB,MAAMC,UAA+B,EAAE;CACvC,IAAI,iBAAiB;AAErB,MAAK,MAAM,SAAS,SAAS,QAAQ;AACnC,MAAI,WAAW,IAAI,MAAM,SAAS,CAAE;AAEpC,MAAI,CADe,aAAa,MAAM,YAAY,SAAS,CAC1C;EACjB,MAAM,aAAa,cAAc,gBAAgB,IAAI,MAAM,KAAK,IAAI,MAAM;AAC1E,UAAQ,KAAK;GACX,MAAM,CAAC;IAAE,OAAO;IAAY,WAAW;IAAG,CAAC;GAC3C,QAAQ;GACT,CAAC;;AAGJ,MAAK,MAAM,QAAQ,SAAS,YAAY;EACtC,MAAM,UAAU,KAAK,SAAS;EAC9B,MAAM,WAAW,KAAK,SAAS;EAC/B,MAAM,cAAc,KAAK,SAAS;AAClC,MAAI,CAAC,WAAW,CAAC,YAAY,CAAC,YAAa;EAE3C,MAAM,YAAY,sBAAsB,MAAM,EAAE;AAChD,MAAI,CAAC,UAAW;EAChB,MAAM,eAAe,oBAAoB,UAAU;AACnD,MAAI,aAAa,WAAW,EAAG;EAE/B,MAAM,cAAc,aAAa,MAAM,MAAM,EAAE,WAAW;AAG1D,MAFsB,aAAa,QAAQ,MAAM,EAAE,WAAW,CAAC,SAE3C,GAAG;AACrB,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,YAAY,aAAa;AAC3B,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,aAAa;AACf;AACA,OAAI,iBAAiB,GAAG;AACtB,gBAAY,KAAK;KACf,MAAM;KACN,SAAS,0DAA0D,SAAS,KAAK;KACjF;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;AAGF,OAAI,aAAa;AACf,gBAAY,KAAK;KACf,MAAM;KACN,SACE;KACF;KACA,MAAM,KAAK;KACZ,CAAC;AACF;;;EAIJ,MAAM,UAAU,iBAAiB,MAAM,OAAO;EAC9C,MAAMC,mBAA2C,cAC7C,SACA,oBAAoB,QAAQ;AAEhC,MACE,eACA,OAAO,qBAAqB,YAC5B;GAAC;GAAU;GAAY;GAAK,CAAC,SAAS,iBAAiB,EACvD;AACA,eAAY,KAAK;IACf,MAAM;IACN,SAAS,mDAAmD;IAC5D;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,MAAI,qBAAqB,YAAY,aAAa,SAAS,GAAG;AAC5D,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,OAAO,aAAa,KAAK,OAAO;AASpC,UAAO;IAAE,OARU,GAAG,aAClB,GAAG,KAAK,QAAQ,mBAAmB,GAAG,WAAmB;KACvD,MAAM,SAAS,cAAc,gBAAgB,IAAI,OAAO;AACxD,YAAO,SAAS,GAAG,OAAO,QAAQ,GAAG,OAAO;MAC5C,GACD,cAAc,gBAAgB,IAAI,GAAG,KAAK,IAAI,GAAG;IAG1B,WAD1B,GAAG,aAAa,OAAQ,GAAG,YAAuC;IAC7B;IACvC;EAEF,MAAM,SAAS,WAAW,OAAO;EACjC,MAAM,SAAS,cAAc,SAAY,gBAAgB,iBAAiB,MAAM,SAAS,CAAC;EAC1F,MAAM,qBAAqB,cACvB,SACA,gBAAgB,iBAAiB,MAAM,qBAAqB,CAAC;AAEjE,MAAI,eAAe,sBAAsB,MAAM;AAC7C,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,0BAA0B,aAAa,iBAAiB,MAAM,SAAS,CAAC;EAE9E,MAAM,aAAa,iBAAiB,MAAM,UAAU;EACpD,MAAM,aAAa,iBAAiB,MAAM,UAAU;AAEpD,MAAI,cAAc,QAAQ,cAAc,MAAM;AAC5C,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;AAGF,OAAK,cAAc,QAAQ,cAAc,SAAS,CAAC,aAAa;AAC9D,eAAY,KAAK;IACf,MAAM;IACN,SACE;IACF;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,qBACJ,cAAc,OACV,oBAAoB,YAAY,EAAE,GAClC,cAAc,OACZ,oBAAoB,YAAY,EAAE,GAClC;EAER,MAAM,YAAY,eAAe,KAAK;AACtC,MAAI,cAAc,MAAM;AACtB,eAAY,KAAK;IACf,MAAM;IACN,SAAS;IACT;IACA,MAAM,KAAK;IACZ,CAAC;AACF;;EAGF,MAAM,aAAa,aAAa,iBAAiB,MAAM,UAAU,CAAC;EAClE,IAAIC;AACJ,MAAI,YAAY;AACd,aAAU,EAAE;AACZ,QAAK,MAAM,CAAC,GAAG,MAAM,OAAO,QAAQ,WAAW,CAC7C,KAAI,OAAO,MAAM,SAAU,SAAQ,KAAK;;EAO5C,MAAM,mBAAmB,kBAHF,cACnB,iBAAiB,MAAM,WAAW,GAClC,iBAAiB,MAAM,mBAAmB,CACY;EAG1D,MAAM,oBAAoB,kBADF,iBAAiB,MAAM,mBAAmB,CACN;EAE5D,MAAMC,QAA2B;GAC/B;GACA,GAAI,UAAU,QAAQ,EAAE,QAAQ;GAChC,GAAI,UAAU,QAAQ,EAAE,QAAQ;GAChC,GAAI,sBAAsB,QAAQ,EAAE,oBAAoB;GACxD,GAAI,2BAA2B,QAAQ,EAAE,yBAAyB;GAClE,GAAI,sBAAsB,QAAQ,EAAE,oBAAoB;GACxD,GAAI,aAAa,QAAQ,EAAE,WAAW;GACtC,GAAI,WAAW,QAAQ,EAAE,SAAS;GAClC,GAAI,oBAAoB,QAAQ,EAAE,kBAAkB;GACpD,GAAI,qBAAqB,QAAQ,EAAE,mBAAmB;GACvD;AAED,UAAQ,KAAK,MAAM;;AAGrB,QAAO;;AAGT,SAAS,gBAAgB,OAAiB,YAA0C;AAClF,QAAO,WAAW,IAAI,MAAM,SAAS;;AAGvC,SAAS,oBACP,OACA,uBACoB;AACpB,QAAO,sBAAsB,IAAI,MAAM,SAAS;;AAGlD,SAAS,wBACP,OACA,WACA,oBACA,uBACA,UACA,aAC2B;AAC3B,KAAI,mBAAmB,IAAI,MAAM,SAAS,EAAE;EAC1C,MAAMC,WAAwB;GAC5B,MAAM;IAAE,MAAM;IAAe,MAAM,MAAM;IAAU;GACnD,UAAU,MAAM;GACjB;AACD,SAAO,MAAM,OAAO;GAAE,GAAGC;GAAQ,MAAM;GAAM,GAAGA;;CAGlD,MAAM,UAAU,oBAAoB,OAAO,sBAAsB;AACjE,KAAI,CAAC,SAAS;AACZ,cAAY,KAAK;GACf,MAAM;GACN,SAAS,UAAU,UAAU,GAAG,MAAM,KAAK,UAAU,MAAM,SAAS;GACpE;GACA,MAAM,MAAM;GACb,CAAC;AACF;;CAGF,MAAMD,SAAwB;EAC5B,MAAM;GAAE,MAAM;GAAU;GAAS;EACjC,UAAU,MAAM;EACjB;AACD,QAAO,MAAM,OAAO;EAAE,GAAG;EAAQ,MAAM;EAAM,GAAG;;AAGlD,SAAgB,oCACd,OAC6C;CAC7C,MAAM,EAAE,UAAU,0BAA0B;CAC5C,MAAM,WAAW,SAAS,IAAI;CAC9B,MAAMR,cAA0C,EAAE;CAClD,MAAM,aAAa,IAAI,IAAI,SAAS,IAAI,OAAO,KAAK,MAAM,EAAE,KAAK,CAAC;CAClE,MAAM,qBAAqB,IAAI,IAAI,SAAS,IAAI,eAAe,KAAK,OAAO,GAAG,KAAK,CAAC;CAEpF,MAAMU,SAA0C,EAAE;CAClD,MAAMC,cAAuD,EAAE;CAC/D,MAAMC,QAAgC,EAAE;CACxC,MAAMC,iBAA+B,EAAE;CAUvC,MAAMC,yBAAkD,EAAE;AAE1D,MAAK,MAAM,YAAY,SAAS,IAAI,QAAQ;EAC1C,MAAM,iBAAiB,sBAAsB,SAAS;EACtD,MAAM,gBAAgB,qBAAqB,SAAS;EAEpD,MAAMC,SAAwC,EAAE;EAChD,MAAMC,YAAuD,EAAE;AAE/D,OAAK,MAAM,SAAS,SAAS,QAAQ;AACnC,OAAI,gBAAgB,OAAO,WAAW,EAAE;IACtC,MAAM,WAAW,uBAAuB,MAAM,WAAW;AAEzD,QAAI,MAAM,QAAQ,EAAE,UAAU,UAAU,UAAU,aAAa;AAC7D,4BAAuB,KAAK;MAC1B,WAAW,SAAS;MACpB,WAAW,MAAM;MACjB,iBAAiB,MAAM;MACvB,GAAI,UAAU,iBAAiB,SAC3B,EAAE,cAAc,SAAS,cAAc,GACvC,EAAE;MACN,aAAa,MAAM,OAAO,QAAQ;MAClC;MACD,CAAC;AACF;;AAGF,QAAI,UAAU,UAAU,UAAU,YAAY;KAC5C,MAAM,cAAc,SAAS,OAAO,KAAK,MAAM,cAAc,gBAAgB,IAAI,EAAE,IAAI,EAAE;KAEzF,MAAM,cAAc,SAAS,IAAI,OAAO,MAAM,MAAM,EAAE,SAAS,MAAM,SAAS;KAC9E,MAAM,sBAAsB,cAAc,qBAAqB,YAAY,GAAG;KAC9E,MAAM,eAAe,SAAS,WAAW,KACtC,MAAM,qBAAqB,gBAAgB,IAAI,EAAE,IAAI,EACvD;AAED,eAAU,MAAM,QAAQ;MACtB,IAAI,MAAM;MACV,aAAa;MACb,IAAI;OACF,aAAa;OACb,cAAc;OACf;MACF;AAED,oBAAe,KAAK;MAClB,gBAAgB,SAAS;MACzB,WAAW,MAAM;MACjB,aAAa,MAAM;MACnB,GAAI,SAAS,iBAAiB,SAAY,EAAE,cAAc,SAAS,cAAc,GAAG,EAAE;MACtF,aAAa;MACb,cAAc;MACf,CAAC;;AAEJ;;GAGF,MAAM,WAAW,wBACf,OACA,SAAS,MACT,oBACA,uBACA,UACA,YACD;AACD,OAAI,CAAC,SAAU;GAEf,MAAM,aAAa,cAAc,gBAAgB,IAAI,MAAM,KAAK,IAAI,MAAM;AAC1E,UAAO,cAAc;;EAGvB,MAAM,iBAAiB,SAAS,WAAW,MAAM,SAAS,KAAK,SAAS,OAAO;AAE/E,MAAI,CADe,SAAS,OAAO,MAAM,MAAM,aAAa,EAAE,YAAY,KAAK,KAAK,OAAU,IAC3E,CAAC,eAClB,aAAY,KAAK;GACf,MAAM;GACN,SAAS,UAAU,SAAS,KAAK;GACjC;GACD,CAAC;AAGJ,SAAO,SAAS,QAAQ;GAAE;GAAQ;GAAW,SAAS,EAAE,YAAY,gBAAgB;GAAE;EACtF,MAAM,eAAe,eAAe,UAAU,eAAe,YAAY,UAAU,YAAY;AAC/F,cAAY,kBAAkB,aAAa,SAAS,IAAI,EAAE,SAAS,cAAc,GAAG,EAAE;AACtF,QAAM,kBAAkB,SAAS;;CAGnC,MAAMC,eAAoD,EAAE;AAC5D,MAAK,MAAM,iBAAiB,SAAS,IAAI,gBAAgB;EACvD,MAAMF,SAAwC,EAAE;AAChD,OAAK,MAAM,SAAS,cAAc,QAAQ;GACxC,MAAM,WAAW,wBACf,OACA,cAAc,MACd,oBACA,uBACA,UACA,YACD;AACD,OAAI,CAAC,SAAU;AACf,UAAO,MAAM,QAAQ;;AAEvB,eAAa,cAAc,QAAQ,EAAE,QAAQ;;AAG/C,MAAK,MAAM,GAAG,eAAe,OAAO,QAAQ,OAAO,EAAE;EAEnD,MAAM,OAAO,YADU,WAAW,QAAQ;AAE1C,MAAI,KAEF,MAAK,eADa,iBAAiB,WAAW,QAAQ,aAAa;;CAKvE,MAAM,oCAAoB,IAAI,KAA2B;AACzD,MAAK,MAAM,MAAM,gBAAgB;EAC/B,MAAM,MAAM,kBAAkB,GAAG,gBAAgB,GAAG,YAAY;EAChE,MAAM,WAAW,kBAAkB,IAAI,IAAI;AAC3C,MAAI,SACF,UAAS,KAAK,GAAG;MAEjB,mBAAkB,IAAI,KAAK,CAAC,GAAG,CAAC;;AAIpC,MAAK,MAAM,aAAa,wBAAwB;EAC9C,MAAM,UAAU,kBAAkB,UAAU,iBAAiB,UAAU,UAAU;EACjF,MAAM,cAAc,kBAAkB,IAAI,QAAQ,IAAI,EAAE;EACxD,MAAM,UAAU,UAAU,eACtB,YAAY,QAAQ,MAAM,EAAE,iBAAiB,UAAU,aAAa,GACpE,CAAC,GAAG,YAAY;AAEpB,MAAI,QAAQ,WAAW,GAAG;AACxB,eAAY,KAAK;IACf,MAAM;IACN,SAAS,4BAA4B,UAAU,UAAU,GAAG,UAAU,UAAU,+CAA+C,UAAU,gBAAgB;IACzJ;IACA,MAAM,UAAU,MAAM;IACvB,CAAC;AACF;;AAEF,MAAI,QAAQ,SAAS,GAAG;AACtB,eAAY,KAAK;IACf,MAAM;IACN,SAAS,4BAA4B,UAAU,UAAU,GAAG,UAAU,UAAU,iDAAiD,UAAU,gBAAgB;IAC3J;IACA,MAAM,UAAU,MAAM;IACvB,CAAC;AACF;;EAGF,MAAM,KAAK,QAAQ;AACnB,MAAI,CAAC,GAAI;EACT,MAAM,aAAa,OAAO,UAAU;AACpC,MAAI,CAAC,WAAY;AACjB,aAAW,UAAU,UAAU,aAAa;GAC1C,IAAI,UAAU;GACd,aAAa,UAAU;GACvB,IAAI;IACF,aAAa,GAAG;IAChB,cAAc,GAAG;IAClB;GACF;;CAGH,MAAM,EAAE,2BAA2B,qBAAqB,gCACtD,UACA,UACA,YACD;CACD,MAAM,aAAa,oBAAoB;EACrC;EACA;EACA;EACA;EACA;EACA;EACA;EACD,CAAC;AAEF,KAAI,YAAY,SAAS,KAAK,WAAW,YAAY,SAAS,EAC5D,QAAO,MAAM;EACX,SAAS;EACT,aAAa,CAAC,GAAG,aAAa,GAAG,WAAW,YAAY;EACzD,CAAC;CAGJ,MAAM,SAAS;CACf,MAAM,eAAe;CACrB,MAAM,qBAAqB,EAAE,aAAa;CAC1C,MAAM,cAAc,mBAAmB;EAAE;EAAQ;EAAc,SAAS;EAAoB,CAAC;CAC7F,MAAMG,eAAwD,EAAE;AAEhE,QAAO,GAAG;EACR;EACA;EACA,OAAO,WAAW;EAClB,QAAQ,WAAW;EACnB,GAAI,OAAO,KAAK,aAAa,CAAC,SAAS,IAAI,EAAE,cAAc,GAAG,EAAE;EAChE,SAAS;GAAE,GAAG;GAAoB;GAAa;EAC/C,gBAAgB,EAAE;EAClB;EACA,aAAa,mBAAmB;GAAE;GAAQ;GAAc;GAAc,CAAC;EACvE,MAAM,EAAE;EACT,CAAC;;;;;AC71BJ,SAAgB,mCAAgE;AAC9E,QAAO,IAAI,IAAoB;EAC7B,CAAC,UAAU,iBAAiB;EAC5B,CAAC,OAAO,gBAAgB;EACxB,CAAC,WAAW,eAAe;EAC3B,CAAC,YAAY,eAAe;EAC5B,CAAC,YAAY,mBAAmB;EACjC,CAAC"}
@@ -1,9 +0,0 @@
1
- export function createMongoScalarTypeDescriptors(): ReadonlyMap<string, string> {
2
- return new Map<string, string>([
3
- ['String', 'mongo/string@1'],
4
- ['Int', 'mongo/int32@1'],
5
- ['Boolean', 'mongo/bool@1'],
6
- ['DateTime', 'mongo/date@1'],
7
- ['ObjectId', 'mongo/objectId@1'],
8
- ]);
9
- }