@soda-gql/codegen 0.11.18 → 0.11.19

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -13,16 +13,6 @@ let node_crypto = require("node:crypto");
13
13
  const emitOperation = (operation, options) => {
14
14
  const lines = [];
15
15
  const schema = options.schemaDocument ? require_generator.createSchemaIndex(options.schemaDocument) : null;
16
- lines.push(`import { gql } from "${options.graphqlSystemPath}";`);
17
- if (operation.fragmentDependencies.length > 0 && options.fragmentImports) {
18
- for (const fragName of operation.fragmentDependencies) {
19
- const importPath = options.fragmentImports.get(fragName);
20
- if (importPath) {
21
- lines.push(`import { ${fragName}Fragment } from "${importPath}";`);
22
- }
23
- }
24
- }
25
- lines.push("");
26
16
  const exportName = `${operation.name}Compat`;
27
17
  const operationType = operation.kind;
28
18
  lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);
@@ -49,16 +39,6 @@ const emitFragment = (fragment, options) => {
49
39
  const lines = [];
50
40
  const schema = options.schemaDocument ? require_generator.createSchemaIndex(options.schemaDocument) : null;
51
41
  const hasVariables = fragment.variables.length > 0;
52
- lines.push(`import { gql } from "${options.graphqlSystemPath}";`);
53
- if (fragment.fragmentDependencies.length > 0 && options.fragmentImports) {
54
- for (const fragName of fragment.fragmentDependencies) {
55
- const importPath = options.fragmentImports.get(fragName);
56
- if (importPath) {
57
- lines.push(`import { ${fragName}Fragment } from "${importPath}";`);
58
- }
59
- }
60
- }
61
- lines.push("");
62
42
  const exportName = `${fragment.name}Fragment`;
63
43
  const destructure = hasVariables ? "fragment, $var" : "fragment";
64
44
  lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${destructure} }) =>`);
@@ -190,6 +170,9 @@ const emitFieldSelection = (field, indent, variableNames, schema) => {
190
170
  const selections = field.selections;
191
171
  const hasArgs = args && args.length > 0;
192
172
  const hasSelections = selections && selections.length > 0;
173
+ if (!hasArgs && !hasSelections) {
174
+ return (0, neverthrow.ok)(`${padding}${field.name}: true,`);
175
+ }
193
176
  let line = `${padding}...f.${field.name}(`;
194
177
  if (hasArgs) {
195
178
  const argsResult = emitArguments(args, variableNames);
@@ -1296,6 +1279,23 @@ const generateDefsStructure = (schemaName, categoryVars, chunkSize) => {
1296
1279
 
1297
1280
  //#endregion
1298
1281
  //#region packages/codegen/src/file.ts
1282
+ const removeDirectory = (dirPath) => {
1283
+ const targetPath = (0, node_path.resolve)(dirPath);
1284
+ try {
1285
+ (0, node_fs.rmSync)(targetPath, {
1286
+ recursive: true,
1287
+ force: true
1288
+ });
1289
+ return (0, neverthrow.ok)(undefined);
1290
+ } catch (error) {
1291
+ const message = error instanceof Error ? error.message : String(error);
1292
+ return (0, neverthrow.err)({
1293
+ code: "REMOVE_FAILED",
1294
+ message,
1295
+ outPath: targetPath
1296
+ });
1297
+ }
1298
+ };
1299
1299
  const writeModule = (outPath, contents) => {
1300
1300
  const targetPath = (0, node_path.resolve)(outPath);
1301
1301
  try {
@@ -1492,6 +1492,13 @@ export * from "./_internal";
1492
1492
  const defsPaths = [];
1493
1493
  if (categoryVars) {
1494
1494
  const outDir = (0, node_path.dirname)(outPath);
1495
+ const defsDir = (0, node_path.join)(outDir, "_defs");
1496
+ if ((0, node_fs.existsSync)(defsDir)) {
1497
+ const removeResult = removeDirectory(defsDir);
1498
+ if (removeResult.isErr()) {
1499
+ return (0, neverthrow.err)(removeResult.error);
1500
+ }
1501
+ }
1495
1502
  const combinedVars = {
1496
1503
  enums: [],
1497
1504
  inputs: [],
@@ -1 +1 @@
1
- {"version":3,"file":"index.cjs","names":["lines: string[]","createSchemaIndex","inlineFragments: ParsedInlineFragment[]","otherSelections: ParsedSelection[]","entries: string[]","argEntries: string[]","values: string[]","operations: ParsedOperation[]","fragments: ParsedFragment[]","Kind","variables: ParsedVariable[]","levels: TypeLevel[]","lists: (\"[]!\" | \"[]?\")[]","match: RegExpExecArray | null","mergedInner: \"!\" | \"?\"","mergedLists: (\"[]!\" | \"[]?\")[]","usages: VariableUsage[]","error","variables: InferredVariable[]","sorted: ParsedFragment[]","createSchemaIndex","fragments: EnrichedFragment[]","operations: EnrichedOperation[]","variables: EnrichedVariable[]","spreadVariables: InferredVariable[]","esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","generateMultiSchemaModule","defsPaths: string[]","defaultBundler"],"sources":["../src/graphql-compat/emitter.ts","../src/graphql-compat/parser.ts","../src/graphql-compat/transformer.ts","../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["/**\n * Emitter for generating TypeScript compat code from enriched operations.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport type { EnrichedFragment, EnrichedOperation } from \"./transformer\";\nimport type { GraphqlCompatError, ParsedArgument, ParsedInlineFragment, ParsedSelection, ParsedValue } from \"./types\";\n\n/**\n * Schema index for type lookups.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Options for code emission.\n */\nexport type EmitOptions = {\n /** Schema name to use in gql.schemaName() call */\n readonly schemaName: string;\n /** Import path for graphql-system module */\n readonly graphqlSystemPath: string;\n /** Map of fragment name to its import path (relative) */\n readonly fragmentImports?: ReadonlyMap<string, string>;\n /** Schema document for type lookups (required for inline fragment support) */\n readonly schemaDocument?: DocumentNode;\n};\n\n/**\n * Emit TypeScript code for an operation.\n */\nexport const emitOperation = (operation: EnrichedOperation, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Generate imports\n lines.push(`import { gql } from \"${options.graphqlSystemPath}\";`);\n\n // Add fragment imports if needed\n if (operation.fragmentDependencies.length > 0 && options.fragmentImports) {\n for (const fragName of operation.fragmentDependencies) {\n const importPath = options.fragmentImports.get(fragName);\n if (importPath) {\n lines.push(`import { ${fragName}Fragment } from \"${importPath}\";`);\n }\n }\n }\n\n lines.push(\"\");\n\n // Generate export\n const exportName = `${operation.name}Compat`;\n const operationType = operation.kind;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);\n lines.push(` ${operationType}.compat({`);\n lines.push(` name: ${JSON.stringify(operation.name)},`);\n\n // Variables\n if (operation.variables.length > 0) {\n lines.push(` variables: { ${emitVariables(operation.variables)} },`);\n }\n\n // Fields\n lines.push(` fields: ({ f, $ }) => ({`);\n const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit TypeScript code for a fragment.\n */\nexport const emitFragment = (fragment: EnrichedFragment, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n const hasVariables = fragment.variables.length > 0;\n\n // Generate imports\n lines.push(`import { gql } from \"${options.graphqlSystemPath}\";`);\n\n // Add fragment imports if needed\n if (fragment.fragmentDependencies.length > 0 && options.fragmentImports) {\n for (const fragName of fragment.fragmentDependencies) {\n const importPath = options.fragmentImports.get(fragName);\n if (importPath) {\n lines.push(`import { ${fragName}Fragment } from \"${importPath}\";`);\n }\n }\n }\n\n lines.push(\"\");\n\n // Generate export\n const exportName = `${fragment.name}Fragment`;\n\n // Include $var in destructure if fragment has variables\n const destructure = hasVariables ? \"fragment, $var\" : \"fragment\";\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${destructure} }) =>`);\n lines.push(` fragment.${fragment.onType}({`);\n\n // Variables block (if any)\n if (hasVariables) {\n lines.push(` variables: { ${emitVariables(fragment.variables)} },`);\n }\n\n // Fields - include $ in context if fragment has variables\n const fieldsContext = hasVariables ? \"{ f, $ }\" : \"{ f }\";\n lines.push(` fields: (${fieldsContext}) => ({`);\n const fieldLinesResult = emitSelections(fragment.selections, 3, fragment.variables, schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Common variable type for emission (both EnrichedVariable and InferredVariable have these fields).\n */\ntype EmittableVariable = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n};\n\n/**\n * Emit variable definitions.\n */\nconst emitVariables = (variables: readonly EmittableVariable[]): string => {\n return variables.map((v) => `...$var(${JSON.stringify(v.name)}).${v.typeName}(${JSON.stringify(v.modifier)})`).join(\", \");\n};\n\n/**\n * Emit field selections (public API).\n * Converts variable array to Set<string> and delegates to internal implementation.\n */\nconst emitSelections = (\n selections: readonly ParsedSelection[],\n indent: number,\n variables: readonly EmittableVariable[],\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const variableNames = new Set(variables.map((v) => v.name));\n return emitSelectionsInternal(selections, indent, variableNames, schema);\n};\n\n/**\n * Internal implementation for emitting field selections.\n * Takes variableNames as Set<string> for recursive calls.\n */\nconst emitSelectionsInternal = (\n selections: readonly ParsedSelection[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n\n // Separate inline fragments from other selections\n const inlineFragments: ParsedInlineFragment[] = [];\n const otherSelections: ParsedSelection[] = [];\n\n for (const sel of selections) {\n if (sel.kind === \"inlineFragment\") {\n inlineFragments.push(sel);\n } else {\n otherSelections.push(sel);\n }\n }\n\n // Emit regular selections (fields and fragment spreads)\n for (const sel of otherSelections) {\n const result = emitSingleSelection(sel, indent, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n lines.push(result.value);\n }\n\n // Emit grouped inline fragments as union selections\n if (inlineFragments.length > 0) {\n const unionResult = emitInlineFragmentsAsUnion(inlineFragments, indent, variableNames, schema);\n if (unionResult.isErr()) {\n return err(unionResult.error);\n }\n lines.push(unionResult.value);\n }\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit a single selection (field or fragment spread).\n */\nconst emitSingleSelection = (\n sel: ParsedSelection,\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n switch (sel.kind) {\n case \"field\":\n return emitFieldSelection(sel, indent, variableNames, schema);\n case \"fragmentSpread\":\n return ok(`${padding}...${sel.name}Fragment.spread(),`);\n case \"inlineFragment\":\n // This should not happen as inline fragments are handled separately\n return ok(\"\");\n }\n};\n\n/**\n * Emit inline fragments grouped as a union selection.\n * Format: { TypeA: ({ f }) => ({ ...fields }), TypeB: ({ f }) => ({ ...fields }) }\n */\nconst emitInlineFragmentsAsUnion = (\n inlineFragments: readonly ParsedInlineFragment[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Validate inline fragments have type conditions\n for (const frag of inlineFragments) {\n if (frag.onType === \"\") {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_WITHOUT_TYPE\",\n message: \"Inline fragments without type condition are not supported. Use `... on TypeName { }` syntax.\",\n });\n }\n }\n\n // Validate all inline fragments are on union types (not interfaces)\n for (const frag of inlineFragments) {\n if (schema && !schema.objects.has(frag.onType)) {\n // If it's not a known object type, it might be an interface\n // Check if any union contains this type as a member\n let isUnionMember = false;\n for (const [, unionDef] of schema.unions) {\n if (unionDef.members.has(frag.onType)) {\n isUnionMember = true;\n break;\n }\n }\n if (!isUnionMember) {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_ON_INTERFACE\",\n message: `Inline fragments on interface type \"${frag.onType}\" are not supported. Use union types instead.`,\n onType: frag.onType,\n });\n }\n }\n }\n\n // Build union member entries\n const entries: string[] = [];\n for (const frag of inlineFragments) {\n const innerPadding = \" \".repeat(indent + 1);\n const fieldsResult = emitSelectionsInternal(frag.selections, indent + 2, variableNames, schema);\n if (fieldsResult.isErr()) {\n return err(fieldsResult.error);\n }\n\n entries.push(`${innerPadding}${frag.onType}: ({ f }) => ({\n${fieldsResult.value}\n${innerPadding}}),`);\n }\n\n // Emit as spread with union callback: ...f.fieldName()({ Type: ... })\n // Note: This assumes the parent field handles the union - we emit just the union object\n return ok(`${padding}...({\n${entries.join(\"\\n\")}\n${padding}}),`);\n};\n\n/**\n * Emit a single field selection.\n */\nconst emitFieldSelection = (\n field: ParsedSelection & { kind: \"field\" },\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Extract optional fields for type narrowing\n const args = field.arguments;\n const selections = field.selections;\n const hasArgs = args && args.length > 0;\n const hasSelections = selections && selections.length > 0;\n\n let line = `${padding}...f.${field.name}(`;\n\n if (hasArgs) {\n const argsResult = emitArguments(args, variableNames);\n if (argsResult.isErr()) {\n return err(argsResult.error);\n }\n line += argsResult.value;\n }\n\n line += \")\";\n\n if (hasSelections) {\n // Check if selections contain inline fragments (union field)\n const hasInlineFragments = selections.some((s) => s.kind === \"inlineFragment\");\n\n if (hasInlineFragments) {\n // Union field: emit with union callback pattern\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += \"({\\n\";\n line += `${nestedResult.value}\\n`;\n line += `${padding}})`;\n } else {\n // Regular nested selections\n line += \"(({ f }) => ({\\n\";\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += `${nestedResult.value}\\n`;\n line += `${padding}}))`;\n }\n }\n\n line += \",\";\n\n return ok(line);\n};\n\n/**\n * Emit field arguments.\n */\nconst emitArguments = (args: readonly ParsedArgument[], variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n if (args.length === 0) {\n return ok(\"\");\n }\n\n const argEntries: string[] = [];\n for (const arg of args) {\n const result = emitValue(arg.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n argEntries.push(`${arg.name}: ${result.value}`);\n }\n return ok(`{ ${argEntries.join(\", \")} }`);\n};\n\n/**\n * Emit a value (literal or variable reference).\n */\nconst emitValue = (value: ParsedValue, variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n switch (value.kind) {\n case \"variable\":\n // Check if it's a declared variable\n if (variableNames.has(value.name)) {\n return ok(`$.${value.name}`);\n }\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `Variable \"$${value.name}\" is not declared in the operation`,\n variableName: value.name,\n });\n case \"int\":\n case \"float\":\n return ok(value.value);\n case \"string\":\n return ok(JSON.stringify(value.value));\n case \"boolean\":\n return ok(value.value ? \"true\" : \"false\");\n case \"null\":\n return ok(\"null\");\n case \"enum\":\n // Enums are emitted as string literals in soda-gql\n return ok(JSON.stringify(value.value));\n case \"list\": {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValue(v, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n case \"object\": {\n if (value.fields.length === 0) {\n return ok(\"{}\");\n }\n const entries: string[] = [];\n for (const f of value.fields) {\n const result = emitValue(f.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n return ok(`{ ${entries.join(\", \")} }`);\n }\n }\n};\n","/**\n * Parser for .graphql operation files.\n * Extracts operations and fragments from GraphQL documents.\n * @module\n */\n\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport {\n type ArgumentNode,\n type DocumentNode,\n type FieldNode,\n type FragmentDefinitionNode,\n type FragmentSpreadNode,\n type InlineFragmentNode,\n Kind,\n type OperationDefinitionNode,\n parse,\n type SelectionNode,\n type TypeNode,\n type ValueNode,\n type VariableDefinitionNode,\n} from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\n\nimport type {\n GraphqlCompatError,\n ParsedArgument,\n ParsedFieldSelection,\n ParsedFragment,\n ParsedFragmentSpread,\n ParsedInlineFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n} from \"./types\";\n\n/**\n * Parse a single .graphql file and extract operations and fragments.\n */\nexport const parseGraphqlFile = (filePath: string): Result<ParseResult, GraphqlCompatError> => {\n const resolvedPath = resolve(filePath);\n\n if (!existsSync(resolvedPath)) {\n return err({\n code: \"GRAPHQL_FILE_NOT_FOUND\",\n message: `GraphQL file not found at ${resolvedPath}`,\n filePath: resolvedPath,\n });\n }\n\n try {\n const source = readFileSync(resolvedPath, \"utf8\");\n const document = parse(source);\n return ok(extractFromDocument(document, resolvedPath));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: resolvedPath,\n });\n }\n};\n\n/**\n * Parse GraphQL source string directly.\n */\nexport const parseGraphqlSource = (source: string, sourceFile: string): Result<ParseResult, GraphqlCompatError> => {\n try {\n const document = parse(source);\n return ok(extractFromDocument(document, sourceFile));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: sourceFile,\n });\n }\n};\n\n/**\n * Extract operations and fragments from a parsed GraphQL document.\n */\nconst extractFromDocument = (document: DocumentNode, sourceFile: string): ParseResult => {\n const operations: ParsedOperation[] = [];\n const fragments: ParsedFragment[] = [];\n\n for (const definition of document.definitions) {\n if (definition.kind === Kind.OPERATION_DEFINITION) {\n const operation = extractOperation(definition, sourceFile);\n if (operation) {\n operations.push(operation);\n }\n } else if (definition.kind === Kind.FRAGMENT_DEFINITION) {\n fragments.push(extractFragment(definition, sourceFile));\n }\n }\n\n return { operations, fragments };\n};\n\n/**\n * Extract a single operation from an OperationDefinitionNode.\n */\nconst extractOperation = (node: OperationDefinitionNode, sourceFile: string): ParsedOperation | null => {\n // Anonymous operations are not supported\n if (!node.name) {\n return null;\n }\n\n const variables: ParsedVariable[] = (node.variableDefinitions ?? []).map(extractVariable);\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n kind: node.operation,\n name: node.name.value,\n variables,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a fragment from a FragmentDefinitionNode.\n */\nconst extractFragment = (node: FragmentDefinitionNode, sourceFile: string): ParsedFragment => {\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n name: node.name.value,\n onType: node.typeCondition.name.value,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a variable definition.\n */\nconst extractVariable = (node: VariableDefinitionNode): ParsedVariable => {\n const { typeName, modifier } = parseTypeNode(node.type);\n const defaultValue = node.defaultValue ? extractValue(node.defaultValue) : undefined;\n\n return {\n name: node.variable.name.value,\n typeName,\n modifier,\n // Type kind will be determined later when we have schema context\n typeKind: \"scalar\",\n defaultValue,\n };\n};\n\n/**\n * Parse a GraphQL TypeNode into type name and modifier.\n *\n * Format: inner nullability + list modifiers\n * - Inner: `!` (non-null) or `?` (nullable)\n * - List: `[]!` (non-null list) or `[]?` (nullable list)\n */\nexport const parseTypeNode = (node: TypeNode): { typeName: string; modifier: string } => {\n type TypeLevel = { kind: \"list\" | \"named\"; nonNull: boolean };\n const levels: TypeLevel[] = [];\n\n const collect = (n: TypeNode, nonNull: boolean): string => {\n if (n.kind === Kind.NON_NULL_TYPE) {\n return collect(n.type, true);\n }\n if (n.kind === Kind.LIST_TYPE) {\n levels.push({ kind: \"list\", nonNull });\n return collect(n.type, false);\n }\n levels.push({ kind: \"named\", nonNull });\n return n.name.value;\n };\n\n const typeName = collect(node, false);\n\n // Build modifier from levels (reverse order)\n let modifier = \"?\";\n for (const level of levels.slice().reverse()) {\n if (level.kind === \"named\") {\n modifier = level.nonNull ? \"!\" : \"?\";\n continue;\n }\n const listSuffix = level.nonNull ? \"[]!\" : \"[]?\";\n modifier = `${modifier}${listSuffix}`;\n }\n\n return { typeName, modifier };\n};\n\n/**\n * Extract selections from a SelectionSet.\n */\nconst extractSelections = (selections: readonly SelectionNode[]): ParsedSelection[] => {\n return selections.map(extractSelection);\n};\n\n/**\n * Extract a single selection.\n */\nconst extractSelection = (node: SelectionNode): ParsedSelection => {\n switch (node.kind) {\n case Kind.FIELD:\n return extractFieldSelection(node);\n case Kind.FRAGMENT_SPREAD:\n return extractFragmentSpread(node);\n case Kind.INLINE_FRAGMENT:\n return extractInlineFragment(node);\n }\n};\n\n/**\n * Extract a field selection.\n */\nconst extractFieldSelection = (node: FieldNode): ParsedFieldSelection => {\n const args = node.arguments?.length ? node.arguments.map(extractArgument) : undefined;\n const selections = node.selectionSet ? extractSelections(node.selectionSet.selections) : undefined;\n\n return {\n kind: \"field\",\n name: node.name.value,\n alias: node.alias?.value,\n arguments: args,\n selections,\n };\n};\n\n/**\n * Extract a fragment spread.\n */\nconst extractFragmentSpread = (node: FragmentSpreadNode): ParsedFragmentSpread => {\n return {\n kind: \"fragmentSpread\",\n name: node.name.value,\n };\n};\n\n/**\n * Extract an inline fragment.\n */\nconst extractInlineFragment = (node: InlineFragmentNode): ParsedInlineFragment => {\n return {\n kind: \"inlineFragment\",\n onType: node.typeCondition?.name.value ?? \"\",\n selections: extractSelections(node.selectionSet.selections),\n };\n};\n\n/**\n * Extract an argument.\n */\nconst extractArgument = (node: ArgumentNode): ParsedArgument => {\n return {\n name: node.name.value,\n value: extractValue(node.value),\n };\n};\n\n/**\n * Assert unreachable code path (for exhaustiveness checks).\n */\nconst assertUnreachable = (value: never): never => {\n throw new Error(`Unexpected value: ${JSON.stringify(value)}`);\n};\n\n/**\n * Extract a value (literal or variable reference).\n */\nconst extractValue = (node: ValueNode): ParsedValue => {\n switch (node.kind) {\n case Kind.VARIABLE:\n return { kind: \"variable\", name: node.name.value };\n case Kind.INT:\n return { kind: \"int\", value: node.value };\n case Kind.FLOAT:\n return { kind: \"float\", value: node.value };\n case Kind.STRING:\n return { kind: \"string\", value: node.value };\n case Kind.BOOLEAN:\n return { kind: \"boolean\", value: node.value };\n case Kind.NULL:\n return { kind: \"null\" };\n case Kind.ENUM:\n return { kind: \"enum\", value: node.value };\n case Kind.LIST:\n return { kind: \"list\", values: node.values.map(extractValue) };\n case Kind.OBJECT:\n return {\n kind: \"object\",\n fields: node.fields.map((field) => ({\n name: field.name.value,\n value: extractValue(field.value),\n })),\n };\n default:\n return assertUnreachable(node);\n }\n};\n","/**\n * Transformer for enriching parsed GraphQL operations with schema information.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport { parseTypeNode } from \"./parser\";\nimport type {\n GraphqlCompatError,\n InferredVariable,\n ParsedArgument,\n ParsedFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n} from \"./types\";\n\n/**\n * Schema index type extracted from generator.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Built-in GraphQL scalar types.\n */\nconst builtinScalarTypes = new Set([\"ID\", \"String\", \"Int\", \"Float\", \"Boolean\"]);\n\n// ============================================================================\n// Modifier Merging Utilities\n// ============================================================================\n\n/**\n * Parsed structure of a modifier for comparison and merging.\n * Example: \"![]?\" -> { inner: \"!\", lists: [\"[]?\"] }\n */\ntype ModifierStructure = {\n readonly inner: \"!\" | \"?\";\n readonly lists: readonly (\"[]!\" | \"[]?\")[];\n};\n\n/**\n * Parse a modifier string into its structural components.\n * @param modifier - Modifier string like \"!\", \"?\", \"![]!\", \"?[]?[]!\"\n * @returns Parsed structure with inner nullability and list modifiers\n */\nconst parseModifierStructure = (modifier: string): ModifierStructure => {\n // Extract inner nullability (first character)\n const inner = modifier[0] === \"!\" ? \"!\" : \"?\";\n\n // Extract list modifiers ([]! or []?)\n const lists: (\"[]!\" | \"[]?\")[] = [];\n const listPattern = /\\[\\]([!?])/g;\n let match: RegExpExecArray | null;\n while ((match = listPattern.exec(modifier)) !== null) {\n lists.push(`[]${match[1]}` as \"[]!\" | \"[]?\");\n }\n\n return { inner, lists };\n};\n\n/**\n * Rebuild modifier string from structure.\n */\nconst buildModifier = (structure: ModifierStructure): string => {\n return structure.inner + structure.lists.join(\"\");\n};\n\n/**\n * Merge two modifiers by taking the stricter constraint at each level.\n * - Non-null (!) is stricter than nullable (?)\n * - List depths must match\n *\n * @param a - First modifier\n * @param b - Second modifier\n * @returns Merged modifier or error if incompatible\n */\nexport const mergeModifiers = (a: string, b: string): { ok: true; value: string } | { ok: false; reason: string } => {\n const structA = parseModifierStructure(a);\n const structB = parseModifierStructure(b);\n\n // List depths must match\n if (structA.lists.length !== structB.lists.length) {\n return {\n ok: false,\n reason: `Incompatible list depths: \"${a}\" has ${structA.lists.length} list level(s), \"${b}\" has ${structB.lists.length}`,\n };\n }\n\n // Take stricter inner constraint (! beats ?)\n const mergedInner: \"!\" | \"?\" = structA.inner === \"!\" || structB.inner === \"!\" ? \"!\" : \"?\";\n\n // Merge each list level (! beats ?)\n const mergedLists: (\"[]!\" | \"[]?\")[] = [];\n for (let i = 0; i < structA.lists.length; i++) {\n const listA = structA.lists[i]!;\n const listB = structB.lists[i]!;\n mergedLists.push(listA === \"[]!\" || listB === \"[]!\" ? \"[]!\" : \"[]?\");\n }\n\n return { ok: true, value: buildModifier({ inner: mergedInner, lists: mergedLists }) };\n};\n\n// ============================================================================\n// Variable Collection from Selections\n// ============================================================================\n\n/**\n * Intermediate type for tracking variable usages before merging.\n */\ntype VariableUsage = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Get the expected type for a field argument from the schema.\n * Returns null if the field or argument is not found.\n */\nconst getArgumentType = (\n schema: SchemaIndex,\n parentTypeName: string,\n fieldName: string,\n argumentName: string,\n): { typeName: string; modifier: string } | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const argDef = fieldDef.arguments?.find((arg) => arg.name.value === argumentName);\n if (!argDef) return null;\n\n return parseTypeNode(argDef.type);\n};\n\n/**\n * Get the expected type for an input object field from the schema.\n */\nconst getInputFieldType = (\n schema: SchemaIndex,\n inputTypeName: string,\n fieldName: string,\n): { typeName: string; modifier: string } | null => {\n const inputRecord = schema.inputs.get(inputTypeName);\n if (!inputRecord) return null;\n\n const fieldDef = inputRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n return parseTypeNode(fieldDef.type);\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKindFromName = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) return \"scalar\";\n if (isEnumName(schema, typeName)) return \"enum\";\n if (schema.inputs.has(typeName)) return \"input\";\n return null;\n};\n\n/**\n * Extract variable usages from a parsed value, given the expected type.\n * Handles nested input objects recursively.\n */\nconst collectVariablesFromValue = (\n value: ParsedValue,\n expectedTypeName: string,\n expectedModifier: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n if (value.kind === \"variable\") {\n const typeKind = resolveTypeKindFromName(schema, expectedTypeName);\n if (!typeKind) {\n return {\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${expectedTypeName}\" for variable \"$${value.name}\"`,\n typeName: expectedTypeName,\n };\n }\n usages.push({\n name: value.name,\n typeName: expectedTypeName,\n modifier: expectedModifier,\n typeKind,\n });\n return null;\n }\n\n if (value.kind === \"object\") {\n // For object values, check each field against input type definition\n for (const field of value.fields) {\n const fieldType = getInputFieldType(schema, expectedTypeName, field.name);\n if (!fieldType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${field.name}\" on input type \"${expectedTypeName}\"`,\n typeName: expectedTypeName,\n fieldName: field.name,\n };\n }\n const error = collectVariablesFromValue(field.value, fieldType.typeName, fieldType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n }\n\n if (value.kind === \"list\") {\n // For list values, unwrap one level of list modifier\n // e.g., [ID!]! -> ID! for elements\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n const innerModifier = buildModifier({\n inner: struct.inner,\n lists: struct.lists.slice(1),\n });\n for (const item of value.values) {\n const error = collectVariablesFromValue(item, expectedTypeName, innerModifier, schema, usages);\n if (error) return error;\n }\n }\n }\n\n // Other value kinds (int, float, string, etc.) don't contain variables\n return null;\n};\n\n/**\n * Collect variable usages from field arguments.\n */\nconst collectVariablesFromArguments = (\n args: readonly ParsedArgument[],\n parentTypeName: string,\n fieldName: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n for (const arg of args) {\n const argType = getArgumentType(schema, parentTypeName, fieldName, arg.name);\n if (!argType) {\n return {\n code: \"GRAPHQL_UNKNOWN_ARGUMENT\",\n message: `Unknown argument \"${arg.name}\" on field \"${fieldName}\"`,\n fieldName,\n argumentName: arg.name,\n };\n }\n const error = collectVariablesFromValue(arg.value, argType.typeName, argType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n};\n\n/**\n * Recursively collect all variable usages from selections.\n */\nexport const collectVariableUsages = (\n selections: readonly ParsedSelection[],\n parentTypeName: string,\n schema: SchemaIndex,\n): Result<VariableUsage[], GraphqlCompatError> => {\n const usages: VariableUsage[] = [];\n\n const collect = (sels: readonly ParsedSelection[], parentType: string): GraphqlCompatError | null => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"field\": {\n // Collect from arguments\n if (sel.arguments && sel.arguments.length > 0) {\n const error = collectVariablesFromArguments(sel.arguments, parentType, sel.name, schema, usages);\n if (error) return error;\n }\n\n // Recurse into nested selections\n if (sel.selections && sel.selections.length > 0) {\n // Need to determine the field's return type for nested selections\n const fieldReturnType = getFieldReturnType(schema, parentType, sel.name);\n if (!fieldReturnType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${sel.name}\" on type \"${parentType}\"`,\n typeName: parentType,\n fieldName: sel.name,\n };\n }\n const error = collect(sel.selections, fieldReturnType);\n if (error) return error;\n }\n break;\n }\n case \"inlineFragment\": {\n // Use the inline fragment's type condition\n const error = collect(sel.selections, sel.onType);\n if (error) return error;\n break;\n }\n case \"fragmentSpread\":\n // Fragment spreads are handled separately (variables from spread fragments)\n break;\n }\n }\n return null;\n };\n\n const error = collect(selections, parentTypeName);\n if (error) return err(error);\n\n return ok(usages);\n};\n\n/**\n * Get the return type of a field (unwrapped from modifiers).\n */\nconst getFieldReturnType = (schema: SchemaIndex, parentTypeName: string, fieldName: string): string | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const { typeName } = parseTypeNode(fieldDef.type);\n return typeName;\n};\n\n/**\n * Merge multiple variable usages into a single InferredVariable.\n * Validates type compatibility and merges modifiers using stricter constraint.\n */\nexport const mergeVariableUsages = (\n variableName: string,\n usages: readonly VariableUsage[],\n): Result<InferredVariable, GraphqlCompatError> => {\n if (usages.length === 0) {\n // This shouldn't happen, but handle defensively\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `No usages found for variable \"${variableName}\"`,\n variableName,\n });\n }\n\n const first = usages[0]!;\n\n // Validate all usages have the same type name\n for (const usage of usages) {\n if (usage.typeName !== first.typeName) {\n return err({\n code: \"GRAPHQL_VARIABLE_TYPE_MISMATCH\",\n message: `Variable \"$${variableName}\" has conflicting types: \"${first.typeName}\" and \"${usage.typeName}\"`,\n variableName,\n });\n }\n }\n\n // Merge modifiers\n let mergedModifier = first.modifier;\n for (let i = 1; i < usages.length; i++) {\n const result = mergeModifiers(mergedModifier, usages[i]!.modifier);\n if (!result.ok) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" has incompatible modifiers: ${result.reason}`,\n variableName,\n });\n }\n mergedModifier = result.value;\n }\n\n return ok({\n name: variableName,\n typeName: first.typeName,\n modifier: mergedModifier,\n typeKind: first.typeKind,\n });\n};\n\n/**\n * Infer variables from collected usages.\n * Groups by variable name and merges each group.\n */\nexport const inferVariablesFromUsages = (usages: readonly VariableUsage[]): Result<InferredVariable[], GraphqlCompatError> => {\n // Group usages by variable name\n const byName = new Map<string, VariableUsage[]>();\n for (const usage of usages) {\n const existing = byName.get(usage.name);\n if (existing) {\n existing.push(usage);\n } else {\n byName.set(usage.name, [usage]);\n }\n }\n\n // Merge each group\n const variables: InferredVariable[] = [];\n for (const [name, group] of byName) {\n const result = mergeVariableUsages(name, group);\n if (result.isErr()) return err(result.error);\n variables.push(result.value);\n }\n\n // Sort by name for deterministic output\n variables.sort((a, b) => a.name.localeCompare(b.name));\n\n return ok(variables);\n};\n\n/**\n * Check if a type name is a scalar type.\n */\nconst isScalarName = (schema: SchemaIndex, name: string): boolean => builtinScalarTypes.has(name) || schema.scalars.has(name);\n\n// ============================================================================\n// Fragment Dependency Ordering\n// ============================================================================\n\n/**\n * Topologically sort fragments so dependencies come before dependents.\n * Detects circular dependencies.\n *\n * Note: Uses the existing collectFragmentDependencies function defined below.\n */\nexport const sortFragmentsByDependency = (fragments: readonly ParsedFragment[]): Result<ParsedFragment[], GraphqlCompatError> => {\n // Build dependency graph using existing function\n const graph = new Map<string, Set<string>>();\n for (const fragment of fragments) {\n const deps = collectFragmentDependenciesSet(fragment.selections);\n graph.set(fragment.name, deps);\n }\n\n const fragmentByName = new Map<string, ParsedFragment>();\n for (const f of fragments) {\n fragmentByName.set(f.name, f);\n }\n\n const sorted: ParsedFragment[] = [];\n const visited = new Set<string>();\n const visiting = new Set<string>(); // For cycle detection\n\n const visit = (name: string, path: string[]): GraphqlCompatError | null => {\n if (visited.has(name)) return null;\n\n if (visiting.has(name)) {\n // Found a cycle\n const cycleStart = path.indexOf(name);\n const cycle = path.slice(cycleStart).concat(name);\n return {\n code: \"GRAPHQL_FRAGMENT_CIRCULAR_DEPENDENCY\",\n message: `Circular dependency detected in fragments: ${cycle.join(\" -> \")}`,\n fragmentNames: cycle,\n };\n }\n\n // Fragment might not be in our list (external dependency)\n const fragment = fragmentByName.get(name);\n if (!fragment) {\n // External fragment, skip\n visited.add(name);\n return null;\n }\n\n visiting.add(name);\n const deps = graph.get(name) ?? new Set();\n\n for (const dep of deps) {\n const error = visit(dep, [...path, name]);\n if (error) return error;\n }\n\n visiting.delete(name);\n visited.add(name);\n sorted.push(fragment);\n return null;\n };\n\n for (const fragment of fragments) {\n const error = visit(fragment.name, []);\n if (error) return err(error);\n }\n\n return ok(sorted);\n};\n\n/**\n * Recursively collect fragment spread names from selections into a Set.\n * Internal helper for sortFragmentsByDependency.\n */\nconst collectFragmentDependenciesSet = (selections: readonly ParsedSelection[]): Set<string> => {\n const deps = new Set<string>();\n\n const collect = (sels: readonly ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n deps.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return deps;\n};\n\n/**\n * Check if a type name is an enum type.\n */\nconst isEnumName = (schema: SchemaIndex, name: string): boolean => schema.enums.has(name);\n\n/**\n * Enriched operation with resolved type information.\n */\nexport type EnrichedOperation = Omit<ParsedOperation, \"variables\"> & {\n readonly variables: readonly EnrichedVariable[];\n /** Fragment names used in this operation (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched fragment with resolved type information.\n */\nexport type EnrichedFragment = ParsedFragment & {\n /** Fragment names used in this fragment (for imports) */\n readonly fragmentDependencies: readonly string[];\n /** Variables inferred from field arguments in this fragment */\n readonly variables: readonly InferredVariable[];\n};\n\n/**\n * Enriched variable with resolved type kind.\n */\nexport type EnrichedVariable = Omit<ParsedVariable, \"typeKind\"> & {\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Result of transforming parsed operations.\n */\nexport type TransformResult = {\n readonly operations: readonly EnrichedOperation[];\n readonly fragments: readonly EnrichedFragment[];\n};\n\n/**\n * Options for transformation.\n */\nexport type TransformOptions = {\n /** Schema document for type resolution */\n readonly schemaDocument: DocumentNode;\n};\n\n/**\n * Transform parsed operations/fragments by enriching them with schema information.\n *\n * This resolves variable type kinds (scalar, enum, input), collects\n * fragment dependencies, and infers variables for fragments.\n */\nexport const transformParsedGraphql = (\n parsed: ParseResult,\n options: TransformOptions,\n): Result<TransformResult, GraphqlCompatError> => {\n const schema = createSchemaIndex(options.schemaDocument);\n\n // Sort fragments by dependency (dependencies first)\n const sortResult = sortFragmentsByDependency(parsed.fragments);\n if (sortResult.isErr()) {\n return err(sortResult.error);\n }\n const sortedFragments = sortResult.value;\n\n // Transform fragments in dependency order, building up resolved variables map\n const resolvedFragmentVariables = new Map<string, readonly InferredVariable[]>();\n const fragments: EnrichedFragment[] = [];\n\n for (const frag of sortedFragments) {\n const result = transformFragment(frag, schema, resolvedFragmentVariables);\n if (result.isErr()) {\n return err(result.error);\n }\n resolvedFragmentVariables.set(frag.name, result.value.variables);\n fragments.push(result.value);\n }\n\n // Transform operations\n const operations: EnrichedOperation[] = [];\n for (const op of parsed.operations) {\n const result = transformOperation(op, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n operations.push(result.value);\n }\n\n return ok({ operations, fragments });\n};\n\n/**\n * Transform a single operation.\n */\nconst transformOperation = (op: ParsedOperation, schema: SchemaIndex): Result<EnrichedOperation, GraphqlCompatError> => {\n // Resolve variable type kinds\n const variables: EnrichedVariable[] = [];\n for (const v of op.variables) {\n const typeKind = resolveTypeKind(schema, v.typeName);\n if (typeKind === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${v.typeName}\" in variable \"${v.name}\"`,\n typeName: v.typeName,\n });\n }\n variables.push({ ...v, typeKind });\n }\n\n // Collect fragment dependencies\n const fragmentDependencies = collectFragmentDependencies(op.selections);\n\n return ok({\n ...op,\n variables,\n fragmentDependencies,\n });\n};\n\n/**\n * Transform a single fragment.\n * Infers variables from field arguments and propagates variables from spread fragments.\n */\nconst transformFragment = (\n frag: ParsedFragment,\n schema: SchemaIndex,\n resolvedFragmentVariables: Map<string, readonly InferredVariable[]>,\n): Result<EnrichedFragment, GraphqlCompatError> => {\n // Collect fragment dependencies (fragments used within this fragment)\n const fragmentDependencies = collectFragmentDependencies(frag.selections);\n\n // Collect direct variable usages from this fragment's selections\n const directUsagesResult = collectVariableUsages(frag.selections, frag.onType, schema);\n if (directUsagesResult.isErr()) {\n return err(directUsagesResult.error);\n }\n const directUsages = directUsagesResult.value;\n\n // Collect variables from spread fragments\n const spreadVariables: InferredVariable[] = [];\n for (const depName of fragmentDependencies) {\n const depVariables = resolvedFragmentVariables.get(depName);\n if (depVariables) {\n spreadVariables.push(...depVariables);\n }\n // If not found, it's an external fragment - skip\n }\n\n // Combine direct usages with spread variables\n // Convert spread variables to usages for merging\n const allUsages = [\n ...directUsages,\n ...spreadVariables.map((v) => ({\n name: v.name,\n typeName: v.typeName,\n modifier: v.modifier,\n typeKind: v.typeKind,\n })),\n ];\n\n // Infer final variables\n const variablesResult = inferVariablesFromUsages(allUsages);\n if (variablesResult.isErr()) {\n return err(variablesResult.error);\n }\n\n return ok({\n ...frag,\n fragmentDependencies,\n variables: variablesResult.value,\n });\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKind = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) {\n return \"scalar\";\n }\n if (isEnumName(schema, typeName)) {\n return \"enum\";\n }\n if (schema.inputs.has(typeName)) {\n return \"input\";\n }\n return null;\n};\n\n/**\n * Collect fragment names used in selections (recursively).\n */\nconst collectFragmentDependencies = (selections: readonly import(\"./types\").ParsedSelection[]): readonly string[] => {\n const fragments = new Set<string>();\n\n const collect = (sels: readonly import(\"./types\").ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n fragments.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return [...fragments];\n};\n","import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport type { TypeFilterConfig } from \"@soda-gql/config\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Build typeFilters config for each schema\n const typeFiltersConfig = new Map<string, TypeFilterConfig>();\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.typeFilter) {\n typeFiltersConfig.set(schemaName, schemaConfig.typeFilter);\n }\n }\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n typeFilters: typeFiltersConfig.size > 0 ? typeFiltersConfig : undefined,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;;;;AAiCA,MAAa,iBAAiB,WAA8B,YAA6D;CACvH,MAAMA,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;AAGpF,OAAM,KAAK,wBAAwB,QAAQ,kBAAkB,IAAI;AAGjE,KAAI,UAAU,qBAAqB,SAAS,KAAK,QAAQ,iBAAiB;AACxE,OAAK,MAAM,YAAY,UAAU,sBAAsB;GACrD,MAAM,aAAa,QAAQ,gBAAgB,IAAI,SAAS;AACxD,OAAI,YAAY;AACd,UAAM,KAAK,YAAY,SAAS,mBAAmB,WAAW,IAAI;;;;AAKxE,OAAM,KAAK,GAAG;CAGd,MAAM,aAAa,GAAG,UAAU,KAAK;CACrC,MAAM,gBAAgB,UAAU;AAEhC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,cAAc,cAAc;AACpG,OAAM,KAAK,KAAK,cAAc,WAAW;AACzC,OAAM,KAAK,aAAa,KAAK,UAAU,UAAU,KAAK,CAAC,GAAG;AAG1D,KAAI,UAAU,UAAU,SAAS,GAAG;AAClC,QAAM,KAAK,oBAAoB,cAAc,UAAU,UAAU,CAAC,KAAK;;AAIzE,OAAM,KAAK,+BAA+B;CAC1C,MAAM,mBAAmB,eAAe,UAAU,YAAY,GAAG,UAAU,WAAW,OAAO;AAC7F,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAa,gBAAgB,UAA4B,YAA6D;CACpH,MAAMD,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CACpF,MAAM,eAAe,SAAS,UAAU,SAAS;AAGjD,OAAM,KAAK,wBAAwB,QAAQ,kBAAkB,IAAI;AAGjE,KAAI,SAAS,qBAAqB,SAAS,KAAK,QAAQ,iBAAiB;AACvE,OAAK,MAAM,YAAY,SAAS,sBAAsB;GACpD,MAAM,aAAa,QAAQ,gBAAgB,IAAI,SAAS;AACxD,OAAI,YAAY;AACd,UAAM,KAAK,YAAY,SAAS,mBAAmB,WAAW,IAAI;;;;AAKxE,OAAM,KAAK,GAAG;CAGd,MAAM,aAAa,GAAG,SAAS,KAAK;CAGpC,MAAM,cAAc,eAAe,mBAAmB;AACtD,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,YAAY,QAAQ;AAC5F,OAAM,KAAK,cAAc,SAAS,OAAO,IAAI;AAG7C,KAAI,cAAc;AAChB,QAAM,KAAK,oBAAoB,cAAc,SAAS,UAAU,CAAC,KAAK;;CAIxE,MAAM,gBAAgB,eAAe,aAAa;AAClD,OAAM,KAAK,gBAAgB,cAAc,SAAS;CAClD,MAAM,mBAAmB,eAAe,SAAS,YAAY,GAAG,SAAS,WAAW,OAAO;AAC3F,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAe7B,MAAM,iBAAiB,cAAoD;AACzE,QAAO,UAAU,KAAK,MAAM,WAAW,KAAK,UAAU,EAAE,KAAK,CAAC,IAAI,EAAE,SAAS,GAAG,KAAK,UAAU,EAAE,SAAS,CAAC,GAAG,CAAC,KAAK,KAAK;;;;;;AAO3H,MAAM,kBACJ,YACA,QACA,WACA,WACuC;CACvC,MAAM,gBAAgB,IAAI,IAAI,UAAU,KAAK,MAAM,EAAE,KAAK,CAAC;AAC3D,QAAO,uBAAuB,YAAY,QAAQ,eAAe,OAAO;;;;;;AAO1E,MAAM,0BACJ,YACA,QACA,eACA,WACuC;CACvC,MAAMD,QAAkB,EAAE;CAG1B,MAAME,kBAA0C,EAAE;CAClD,MAAMC,kBAAqC,EAAE;AAE7C,MAAK,MAAM,OAAO,YAAY;AAC5B,MAAI,IAAI,SAAS,kBAAkB;AACjC,mBAAgB,KAAK,IAAI;SACpB;AACL,mBAAgB,KAAK,IAAI;;;AAK7B,MAAK,MAAM,OAAO,iBAAiB;EACjC,MAAM,SAAS,oBAAoB,KAAK,QAAQ,eAAe,OAAO;AACtE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,QAAM,KAAK,OAAO,MAAM;;AAI1B,KAAI,gBAAgB,SAAS,GAAG;EAC9B,MAAM,cAAc,2BAA2B,iBAAiB,QAAQ,eAAe,OAAO;AAC9F,MAAI,YAAY,OAAO,EAAE;AACvB,8BAAW,YAAY,MAAM;;AAE/B,QAAM,KAAK,YAAY,MAAM;;AAG/B,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,uBACJ,KACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAEnC,SAAQ,IAAI,MAAZ;EACE,KAAK,QACH,QAAO,mBAAmB,KAAK,QAAQ,eAAe,OAAO;EAC/D,KAAK,iBACH,2BAAU,GAAG,QAAQ,KAAK,IAAI,KAAK,oBAAoB;EACzD,KAAK,iBAEH,2BAAU,GAAG;;;;;;;AAQnB,MAAM,8BACJ,iBACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAGnC,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,KAAK,WAAW,IAAI;AACtB,8BAAW;IACT,MAAM;IACN,SAAS;IACV,CAAC;;;AAKN,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,UAAU,CAAC,OAAO,QAAQ,IAAI,KAAK,OAAO,EAAE;GAG9C,IAAI,gBAAgB;AACpB,QAAK,MAAM,GAAG,aAAa,OAAO,QAAQ;AACxC,QAAI,SAAS,QAAQ,IAAI,KAAK,OAAO,EAAE;AACrC,qBAAgB;AAChB;;;AAGJ,OAAI,CAAC,eAAe;AAClB,+BAAW;KACT,MAAM;KACN,SAAS,uCAAuC,KAAK,OAAO;KAC5D,QAAQ,KAAK;KACd,CAAC;;;;CAMR,MAAMC,UAAoB,EAAE;AAC5B,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,eAAe,KAAK,OAAO,SAAS,EAAE;EAC5C,MAAM,eAAe,uBAAuB,KAAK,YAAY,SAAS,GAAG,eAAe,OAAO;AAC/F,MAAI,aAAa,OAAO,EAAE;AACxB,8BAAW,aAAa,MAAM;;AAGhC,UAAQ,KAAK,GAAG,eAAe,KAAK,OAAO;EAC7C,aAAa,MAAM;EACnB,aAAa,KAAK;;AAKlB,2BAAU,GAAG,QAAQ;EACrB,QAAQ,KAAK,KAAK,CAAC;EACnB,QAAQ,KAAK;;;;;AAMf,MAAM,sBACJ,OACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;CAGnC,MAAM,OAAO,MAAM;CACnB,MAAM,aAAa,MAAM;CACzB,MAAM,UAAU,QAAQ,KAAK,SAAS;CACtC,MAAM,gBAAgB,cAAc,WAAW,SAAS;CAExD,IAAI,OAAO,GAAG,QAAQ,OAAO,MAAM,KAAK;AAExC,KAAI,SAAS;EACX,MAAM,aAAa,cAAc,MAAM,cAAc;AACrD,MAAI,WAAW,OAAO,EAAE;AACtB,8BAAW,WAAW,MAAM;;AAE9B,UAAQ,WAAW;;AAGrB,SAAQ;AAER,KAAI,eAAe;EAEjB,MAAM,qBAAqB,WAAW,MAAM,MAAM,EAAE,SAAS,iBAAiB;AAE9E,MAAI,oBAAoB;GAEtB,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ;AACR,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;SACd;AAEL,WAAQ;GACR,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;;;AAIvB,SAAQ;AAER,2BAAU,KAAK;;;;;AAMjB,MAAM,iBAAiB,MAAiC,kBAAmE;AACzH,KAAI,KAAK,WAAW,GAAG;AACrB,4BAAU,GAAG;;CAGf,MAAMC,aAAuB,EAAE;AAC/B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,SAAS,UAAU,IAAI,OAAO,cAAc;AAClD,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,GAAG,IAAI,KAAK,IAAI,OAAO,QAAQ;;AAEjD,2BAAU,KAAK,WAAW,KAAK,KAAK,CAAC,IAAI;;;;;AAM3C,MAAM,aAAa,OAAoB,kBAAmE;AACxG,SAAQ,MAAM,MAAd;EACE,KAAK;AAEH,OAAI,cAAc,IAAI,MAAM,KAAK,EAAE;AACjC,8BAAU,KAAK,MAAM,OAAO;;AAE9B,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,MAAM,KAAK;IAClC,cAAc,MAAM;IACrB,CAAC;EACJ,KAAK;EACL,KAAK,QACH,2BAAU,MAAM,MAAM;EACxB,KAAK,SACH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,UACH,2BAAU,MAAM,QAAQ,SAAS,QAAQ;EAC3C,KAAK,OACH,2BAAU,OAAO;EACnB,KAAK,OAEH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,QAAQ;GACX,MAAMC,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,GAAG,cAAc;AAC1C,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,WAAO,KAAK,OAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;EAErC,KAAK,UAAU;AACb,OAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,8BAAU,KAAK;;GAEjB,MAAMF,UAAoB,EAAE;AAC5B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,EAAE,OAAO,cAAc;AAChD,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,YAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;AAE5C,6BAAU,KAAK,QAAQ,KAAK,KAAK,CAAC,IAAI;;;;;;;;;;;;;;;AC5X5C,MAAa,oBAAoB,aAA8D;CAC7F,MAAM,sCAAuB,SAAS;AAEtC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAW;GACT,MAAM;GACN,SAAS,6BAA6B;GACtC,UAAU;GACX,CAAC;;AAGJ,KAAI;EACF,MAAM,mCAAsB,cAAc,OAAO;EACjD,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,aAAa,CAAC;UAC/C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAa,sBAAsB,QAAgB,eAAgE;AACjH,KAAI;EACF,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,WAAW,CAAC;UAC7C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAM,uBAAuB,UAAwB,eAAoC;CACvF,MAAMG,aAAgC,EAAE;CACxC,MAAMC,YAA8B,EAAE;AAEtC,MAAK,MAAM,cAAc,SAAS,aAAa;AAC7C,MAAI,WAAW,SAASC,aAAK,sBAAsB;GACjD,MAAM,YAAY,iBAAiB,YAAY,WAAW;AAC1D,OAAI,WAAW;AACb,eAAW,KAAK,UAAU;;aAEnB,WAAW,SAASA,aAAK,qBAAqB;AACvD,aAAU,KAAK,gBAAgB,YAAY,WAAW,CAAC;;;AAI3D,QAAO;EAAE;EAAY;EAAW;;;;;AAMlC,MAAM,oBAAoB,MAA+B,eAA+C;AAEtG,KAAI,CAAC,KAAK,MAAM;AACd,SAAO;;CAGT,MAAMC,aAA+B,KAAK,uBAAuB,EAAE,EAAE,IAAI,gBAAgB;CACzF,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK;EACX,MAAM,KAAK,KAAK;EAChB;EACA;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,MAA8B,eAAuC;CAC5F,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,QAAQ,KAAK,cAAc,KAAK;EAChC;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,SAAiD;CACxE,MAAM,EAAE,UAAU,aAAa,cAAc,KAAK,KAAK;CACvD,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,aAAa,GAAG;AAE3E,QAAO;EACL,MAAM,KAAK,SAAS,KAAK;EACzB;EACA;EAEA,UAAU;EACV;EACD;;;;;;;;;AAUH,MAAa,iBAAiB,SAA2D;CAEvF,MAAMC,SAAsB,EAAE;CAE9B,MAAM,WAAW,GAAa,YAA6B;AACzD,MAAI,EAAE,SAASF,aAAK,eAAe;AACjC,UAAO,QAAQ,EAAE,MAAM,KAAK;;AAE9B,MAAI,EAAE,SAASA,aAAK,WAAW;AAC7B,UAAO,KAAK;IAAE,MAAM;IAAQ;IAAS,CAAC;AACtC,UAAO,QAAQ,EAAE,MAAM,MAAM;;AAE/B,SAAO,KAAK;GAAE,MAAM;GAAS;GAAS,CAAC;AACvC,SAAO,EAAE,KAAK;;CAGhB,MAAM,WAAW,QAAQ,MAAM,MAAM;CAGrC,IAAI,WAAW;AACf,MAAK,MAAM,SAAS,OAAO,OAAO,CAAC,SAAS,EAAE;AAC5C,MAAI,MAAM,SAAS,SAAS;AAC1B,cAAW,MAAM,UAAU,MAAM;AACjC;;EAEF,MAAM,aAAa,MAAM,UAAU,QAAQ;AAC3C,aAAW,GAAG,WAAW;;AAG3B,QAAO;EAAE;EAAU;EAAU;;;;;AAM/B,MAAM,qBAAqB,eAA4D;AACrF,QAAO,WAAW,IAAI,iBAAiB;;;;;AAMzC,MAAM,oBAAoB,SAAyC;AACjE,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,MACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;;;;;;AAOxC,MAAM,yBAAyB,SAA0C;CACvE,MAAM,OAAO,KAAK,WAAW,SAAS,KAAK,UAAU,IAAI,gBAAgB,GAAG;CAC5E,MAAM,aAAa,KAAK,eAAe,kBAAkB,KAAK,aAAa,WAAW,GAAG;AAEzF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EAChB,OAAO,KAAK,OAAO;EACnB,WAAW;EACX;EACD;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EACjB;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,QAAQ,KAAK,eAAe,KAAK,SAAS;EAC1C,YAAY,kBAAkB,KAAK,aAAa,WAAW;EAC5D;;;;;AAMH,MAAM,mBAAmB,SAAuC;AAC9D,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,OAAO,aAAa,KAAK,MAAM;EAChC;;;;;AAMH,MAAM,qBAAqB,UAAwB;AACjD,OAAM,IAAI,MAAM,qBAAqB,KAAK,UAAU,MAAM,GAAG;;;;;AAM/D,MAAM,gBAAgB,SAAiC;AACrD,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,SACR,QAAO;GAAE,MAAM;GAAY,MAAM,KAAK,KAAK;GAAO;EACpD,KAAKA,aAAK,IACR,QAAO;GAAE,MAAM;GAAO,OAAO,KAAK;GAAO;EAC3C,KAAKA,aAAK,MACR,QAAO;GAAE,MAAM;GAAS,OAAO,KAAK;GAAO;EAC7C,KAAKA,aAAK,OACR,QAAO;GAAE,MAAM;GAAU,OAAO,KAAK;GAAO;EAC9C,KAAKA,aAAK,QACR,QAAO;GAAE,MAAM;GAAW,OAAO,KAAK;GAAO;EAC/C,KAAKA,aAAK,KACR,QAAO,EAAE,MAAM,QAAQ;EACzB,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,OAAO,KAAK;GAAO;EAC5C,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,QAAQ,KAAK,OAAO,IAAI,aAAa;GAAE;EAChE,KAAKA,aAAK,OACR,QAAO;GACL,MAAM;GACN,QAAQ,KAAK,OAAO,KAAK,WAAW;IAClC,MAAM,MAAM,KAAK;IACjB,OAAO,aAAa,MAAM,MAAM;IACjC,EAAE;GACJ;EACH,QACE,QAAO,kBAAkB,KAAK;;;;;;;;;AChRpC,MAAM,qBAAqB,IAAI,IAAI;CAAC;CAAM;CAAU;CAAO;CAAS;CAAU,CAAC;;;;;;AAoB/E,MAAM,0BAA0B,aAAwC;CAEtE,MAAM,QAAQ,SAAS,OAAO,MAAM,MAAM;CAG1C,MAAMG,QAA2B,EAAE;CACnC,MAAM,cAAc;CACpB,IAAIC;AACJ,SAAQ,QAAQ,YAAY,KAAK,SAAS,MAAM,MAAM;AACpD,QAAM,KAAK,KAAK,MAAM,KAAsB;;AAG9C,QAAO;EAAE;EAAO;EAAO;;;;;AAMzB,MAAM,iBAAiB,cAAyC;AAC9D,QAAO,UAAU,QAAQ,UAAU,MAAM,KAAK,GAAG;;;;;;;;;;;AAYnD,MAAa,kBAAkB,GAAW,MAA2E;CACnH,MAAM,UAAU,uBAAuB,EAAE;CACzC,MAAM,UAAU,uBAAuB,EAAE;AAGzC,KAAI,QAAQ,MAAM,WAAW,QAAQ,MAAM,QAAQ;AACjD,SAAO;GACL,IAAI;GACJ,QAAQ,8BAA8B,EAAE,QAAQ,QAAQ,MAAM,OAAO,mBAAmB,EAAE,QAAQ,QAAQ,MAAM;GACjH;;CAIH,MAAMC,cAAyB,QAAQ,UAAU,OAAO,QAAQ,UAAU,MAAM,MAAM;CAGtF,MAAMC,cAAiC,EAAE;AACzC,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,MAAM,QAAQ,KAAK;EAC7C,MAAM,QAAQ,QAAQ,MAAM;EAC5B,MAAM,QAAQ,QAAQ,MAAM;AAC5B,cAAY,KAAK,UAAU,SAAS,UAAU,QAAQ,QAAQ,MAAM;;AAGtE,QAAO;EAAE,IAAI;EAAM,OAAO,cAAc;GAAE,OAAO;GAAa,OAAO;GAAa,CAAC;EAAE;;;;;;AAqBvF,MAAM,mBACJ,QACA,gBACA,WACA,iBACkD;CAClD,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,SAAS,SAAS,WAAW,MAAM,QAAQ,IAAI,KAAK,UAAU,aAAa;AACjF,KAAI,CAAC,OAAQ,QAAO;AAEpB,QAAO,cAAc,OAAO,KAAK;;;;;AAMnC,MAAM,qBACJ,QACA,eACA,cACkD;CAClD,MAAM,cAAc,OAAO,OAAO,IAAI,cAAc;AACpD,KAAI,CAAC,YAAa,QAAO;CAEzB,MAAM,WAAW,YAAY,OAAO,IAAI,UAAU;AAClD,KAAI,CAAC,SAAU,QAAO;AAEtB,QAAO,cAAc,SAAS,KAAK;;;;;AAMrC,MAAM,2BAA2B,QAAqB,aAAyD;AAC7G,KAAI,aAAa,QAAQ,SAAS,CAAE,QAAO;AAC3C,KAAI,WAAW,QAAQ,SAAS,CAAE,QAAO;AACzC,KAAI,OAAO,OAAO,IAAI,SAAS,CAAE,QAAO;AACxC,QAAO;;;;;;AAOT,MAAM,6BACJ,OACA,kBACA,kBACA,QACA,WAC8B;AAC9B,KAAI,MAAM,SAAS,YAAY;EAC7B,MAAM,WAAW,wBAAwB,QAAQ,iBAAiB;AAClE,MAAI,CAAC,UAAU;AACb,UAAO;IACL,MAAM;IACN,SAAS,iBAAiB,iBAAiB,mBAAmB,MAAM,KAAK;IACzE,UAAU;IACX;;AAEH,SAAO,KAAK;GACV,MAAM,MAAM;GACZ,UAAU;GACV,UAAU;GACV;GACD,CAAC;AACF,SAAO;;AAGT,KAAI,MAAM,SAAS,UAAU;AAE3B,OAAK,MAAM,SAAS,MAAM,QAAQ;GAChC,MAAM,YAAY,kBAAkB,QAAQ,kBAAkB,MAAM,KAAK;AACzE,OAAI,CAAC,WAAW;AACd,WAAO;KACL,MAAM;KACN,SAAS,kBAAkB,MAAM,KAAK,mBAAmB,iBAAiB;KAC1E,UAAU;KACV,WAAW,MAAM;KAClB;;GAEH,MAAM,QAAQ,0BAA0B,MAAM,OAAO,UAAU,UAAU,UAAU,UAAU,QAAQ,OAAO;AAC5G,OAAI,MAAO,QAAO;;AAEpB,SAAO;;AAGT,KAAI,MAAM,SAAS,QAAQ;EAGzB,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,MAAI,OAAO,MAAM,SAAS,GAAG;GAC3B,MAAM,gBAAgB,cAAc;IAClC,OAAO,OAAO;IACd,OAAO,OAAO,MAAM,MAAM,EAAE;IAC7B,CAAC;AACF,QAAK,MAAM,QAAQ,MAAM,QAAQ;IAC/B,MAAM,QAAQ,0BAA0B,MAAM,kBAAkB,eAAe,QAAQ,OAAO;AAC9F,QAAI,MAAO,QAAO;;;;AAMxB,QAAO;;;;;AAMT,MAAM,iCACJ,MACA,gBACA,WACA,QACA,WAC8B;AAC9B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,UAAU,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK;AAC5E,MAAI,CAAC,SAAS;AACZ,UAAO;IACL,MAAM;IACN,SAAS,qBAAqB,IAAI,KAAK,cAAc,UAAU;IAC/D;IACA,cAAc,IAAI;IACnB;;EAEH,MAAM,QAAQ,0BAA0B,IAAI,OAAO,QAAQ,UAAU,QAAQ,UAAU,QAAQ,OAAO;AACtG,MAAI,MAAO,QAAO;;AAEpB,QAAO;;;;;AAMT,MAAa,yBACX,YACA,gBACA,WACgD;CAChD,MAAMC,SAA0B,EAAE;CAElC,MAAM,WAAW,MAAkC,eAAkD;AACnG,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK,SAAS;AAEZ,SAAI,IAAI,aAAa,IAAI,UAAU,SAAS,GAAG;MAC7C,MAAMC,UAAQ,8BAA8B,IAAI,WAAW,YAAY,IAAI,MAAM,QAAQ,OAAO;AAChG,UAAIA,QAAO,QAAOA;;AAIpB,SAAI,IAAI,cAAc,IAAI,WAAW,SAAS,GAAG;MAE/C,MAAM,kBAAkB,mBAAmB,QAAQ,YAAY,IAAI,KAAK;AACxE,UAAI,CAAC,iBAAiB;AACpB,cAAO;QACL,MAAM;QACN,SAAS,kBAAkB,IAAI,KAAK,aAAa,WAAW;QAC5D,UAAU;QACV,WAAW,IAAI;QAChB;;MAEH,MAAMA,UAAQ,QAAQ,IAAI,YAAY,gBAAgB;AACtD,UAAIA,QAAO,QAAOA;;AAEpB;;IAEF,KAAK,kBAAkB;KAErB,MAAMA,UAAQ,QAAQ,IAAI,YAAY,IAAI,OAAO;AACjD,SAAIA,QAAO,QAAOA;AAClB;;IAEF,KAAK,iBAEH;;;AAGN,SAAO;;CAGT,MAAM,QAAQ,QAAQ,YAAY,eAAe;AACjD,KAAI,MAAO,4BAAW,MAAM;AAE5B,2BAAU,OAAO;;;;;AAMnB,MAAM,sBAAsB,QAAqB,gBAAwB,cAAqC;CAC5G,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,EAAE,aAAa,cAAc,SAAS,KAAK;AACjD,QAAO;;;;;;AAOT,MAAa,uBACX,cACA,WACiD;AACjD,KAAI,OAAO,WAAW,GAAG;AAEvB,6BAAW;GACT,MAAM;GACN,SAAS,iCAAiC,aAAa;GACvD;GACD,CAAC;;CAGJ,MAAM,QAAQ,OAAO;AAGrB,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,MAAM,aAAa,MAAM,UAAU;AACrC,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,4BAA4B,MAAM,SAAS,SAAS,MAAM,SAAS;IACvG;IACD,CAAC;;;CAKN,IAAI,iBAAiB,MAAM;AAC3B,MAAK,IAAI,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;EACtC,MAAM,SAAS,eAAe,gBAAgB,OAAO,GAAI,SAAS;AAClE,MAAI,CAAC,OAAO,IAAI;AACd,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,gCAAgC,OAAO;IAC3E;IACD,CAAC;;AAEJ,mBAAiB,OAAO;;AAG1B,2BAAU;EACR,MAAM;EACN,UAAU,MAAM;EAChB,UAAU;EACV,UAAU,MAAM;EACjB,CAAC;;;;;;AAOJ,MAAa,4BAA4B,WAAqF;CAE5H,MAAM,SAAS,IAAI,KAA8B;AACjD,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,WAAW,OAAO,IAAI,MAAM,KAAK;AACvC,MAAI,UAAU;AACZ,YAAS,KAAK,MAAM;SACf;AACL,UAAO,IAAI,MAAM,MAAM,CAAC,MAAM,CAAC;;;CAKnC,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,CAAC,MAAM,UAAU,QAAQ;EAClC,MAAM,SAAS,oBAAoB,MAAM,MAAM;AAC/C,MAAI,OAAO,OAAO,CAAE,4BAAW,OAAO,MAAM;AAC5C,YAAU,KAAK,OAAO,MAAM;;AAI9B,WAAU,MAAM,GAAG,MAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC;AAEtD,2BAAU,UAAU;;;;;AAMtB,MAAM,gBAAgB,QAAqB,SAA0B,mBAAmB,IAAI,KAAK,IAAI,OAAO,QAAQ,IAAI,KAAK;;;;;;;AAY7H,MAAa,6BAA6B,cAAuF;CAE/H,MAAM,QAAQ,IAAI,KAA0B;AAC5C,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,OAAO,+BAA+B,SAAS,WAAW;AAChE,QAAM,IAAI,SAAS,MAAM,KAAK;;CAGhC,MAAM,iBAAiB,IAAI,KAA6B;AACxD,MAAK,MAAM,KAAK,WAAW;AACzB,iBAAe,IAAI,EAAE,MAAM,EAAE;;CAG/B,MAAMC,SAA2B,EAAE;CACnC,MAAM,UAAU,IAAI,KAAa;CACjC,MAAM,WAAW,IAAI,KAAa;CAElC,MAAM,SAAS,MAAc,SAA8C;AACzE,MAAI,QAAQ,IAAI,KAAK,CAAE,QAAO;AAE9B,MAAI,SAAS,IAAI,KAAK,EAAE;GAEtB,MAAM,aAAa,KAAK,QAAQ,KAAK;GACrC,MAAM,QAAQ,KAAK,MAAM,WAAW,CAAC,OAAO,KAAK;AACjD,UAAO;IACL,MAAM;IACN,SAAS,8CAA8C,MAAM,KAAK,OAAO;IACzE,eAAe;IAChB;;EAIH,MAAM,WAAW,eAAe,IAAI,KAAK;AACzC,MAAI,CAAC,UAAU;AAEb,WAAQ,IAAI,KAAK;AACjB,UAAO;;AAGT,WAAS,IAAI,KAAK;EAClB,MAAM,OAAO,MAAM,IAAI,KAAK,IAAI,IAAI,KAAK;AAEzC,OAAK,MAAM,OAAO,MAAM;GACtB,MAAM,QAAQ,MAAM,KAAK,CAAC,GAAG,MAAM,KAAK,CAAC;AACzC,OAAI,MAAO,QAAO;;AAGpB,WAAS,OAAO,KAAK;AACrB,UAAQ,IAAI,KAAK;AACjB,SAAO,KAAK,SAAS;AACrB,SAAO;;AAGT,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,QAAQ,MAAM,SAAS,MAAM,EAAE,CAAC;AACtC,MAAI,MAAO,4BAAW,MAAM;;AAG9B,2BAAU,OAAO;;;;;;AAOnB,MAAM,kCAAkC,eAAwD;CAC9F,MAAM,OAAO,IAAI,KAAa;CAE9B,MAAM,WAAW,SAA2C;AAC1D,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,UAAK,IAAI,IAAI,KAAK;AAClB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO;;;;;AAMT,MAAM,cAAc,QAAqB,SAA0B,OAAO,MAAM,IAAI,KAAK;;;;;;;AAkDzF,MAAa,0BACX,QACA,YACgD;CAChD,MAAM,SAASC,oCAAkB,QAAQ,eAAe;CAGxD,MAAM,aAAa,0BAA0B,OAAO,UAAU;AAC9D,KAAI,WAAW,OAAO,EAAE;AACtB,6BAAW,WAAW,MAAM;;CAE9B,MAAM,kBAAkB,WAAW;CAGnC,MAAM,4BAA4B,IAAI,KAA0C;CAChF,MAAMC,YAAgC,EAAE;AAExC,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,SAAS,kBAAkB,MAAM,QAAQ,0BAA0B;AACzE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,4BAA0B,IAAI,KAAK,MAAM,OAAO,MAAM,UAAU;AAChE,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAMC,aAAkC,EAAE;AAC1C,MAAK,MAAM,MAAM,OAAO,YAAY;EAClC,MAAM,SAAS,mBAAmB,IAAI,OAAO;AAC7C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,OAAO,MAAM;;AAG/B,2BAAU;EAAE;EAAY;EAAW,CAAC;;;;;AAMtC,MAAM,sBAAsB,IAAqB,WAAuE;CAEtH,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,GAAG,WAAW;EAC5B,MAAM,WAAW,gBAAgB,QAAQ,EAAE,SAAS;AACpD,MAAI,aAAa,MAAM;AACrB,8BAAW;IACT,MAAM;IACN,SAAS,iBAAiB,EAAE,SAAS,iBAAiB,EAAE,KAAK;IAC7D,UAAU,EAAE;IACb,CAAC;;AAEJ,YAAU,KAAK;GAAE,GAAG;GAAG;GAAU,CAAC;;CAIpC,MAAM,uBAAuB,4BAA4B,GAAG,WAAW;AAEvE,2BAAU;EACR,GAAG;EACH;EACA;EACD,CAAC;;;;;;AAOJ,MAAM,qBACJ,MACA,QACA,8BACiD;CAEjD,MAAM,uBAAuB,4BAA4B,KAAK,WAAW;CAGzE,MAAM,qBAAqB,sBAAsB,KAAK,YAAY,KAAK,QAAQ,OAAO;AACtF,KAAI,mBAAmB,OAAO,EAAE;AAC9B,6BAAW,mBAAmB,MAAM;;CAEtC,MAAM,eAAe,mBAAmB;CAGxC,MAAMC,kBAAsC,EAAE;AAC9C,MAAK,MAAM,WAAW,sBAAsB;EAC1C,MAAM,eAAe,0BAA0B,IAAI,QAAQ;AAC3D,MAAI,cAAc;AAChB,mBAAgB,KAAK,GAAG,aAAa;;;CAOzC,MAAM,YAAY,CAChB,GAAG,cACH,GAAG,gBAAgB,KAAK,OAAO;EAC7B,MAAM,EAAE;EACR,UAAU,EAAE;EACZ,UAAU,EAAE;EACZ,UAAU,EAAE;EACb,EAAE,CACJ;CAGD,MAAM,kBAAkB,yBAAyB,UAAU;AAC3D,KAAI,gBAAgB,OAAO,EAAE;AAC3B,6BAAW,gBAAgB,MAAM;;AAGnC,2BAAU;EACR,GAAG;EACH;EACA,WAAW,gBAAgB;EAC5B,CAAC;;;;;AAMJ,MAAM,mBAAmB,QAAqB,aAAyD;AACrG,KAAI,aAAa,QAAQ,SAAS,EAAE;AAClC,SAAO;;AAET,KAAI,WAAW,QAAQ,SAAS,EAAE;AAChC,SAAO;;AAET,KAAI,OAAO,OAAO,IAAI,SAAS,EAAE;AAC/B,SAAO;;AAET,QAAO;;;;;AAMT,MAAM,+BAA+B,eAAgF;CACnH,MAAM,YAAY,IAAI,KAAa;CAEnC,MAAM,WAAW,SAA6D;AAC5E,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,eAAU,IAAI,IAAI,KAAK;AACvB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO,CAAC,GAAG,UAAU;;;;;ACztBvB,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,8BAAe,WAAW,EAAE;AAC1B,8BAA+B;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaC,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,mCAAoB,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,4BAAY;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,6BAAU,EAAE,SAAS,CAAC;WACf,OAAO;AACd,8BAAW;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,SAAS;AACnC,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACPN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,sCAAuB,WAAW;AAExC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAuC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,yCAA4B,cAAc,OAAO;EACvD,MAAM,8BAAiB,aAAa;AACpC,4BAAsC,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAuC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAuC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,gCAAmB,UAAU;AACnC,2BAAsC,OAAO;;AAG/C,MAAa,cAAc,yCAA8C,SAAS,CAAC,0BAAa,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC9CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,iCAAkB,SAAS;CACjC,MAAM,qCAAsB,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,mCAAoB,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,6BAAc,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,sCAAqBD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,oCAAqB,YAAY,UAAU,2BAAY,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,oCAAqB,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,iCAAkB,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,oCAAqB,aAAa,OAAO,QAAQ;AACvD,MAAI,yBAAY,WAAW,EAAE;AAC3B,8BAAW;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,qCAAsB,aAAa,OAAO,QAAQ;AACxD,OAAI,yBAAY,YAAY,EAAE;AAC5B,+BAAW;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,2BAAW,IAAI,CAAC,GAChC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,oBAAoB,IAAI,KAA+B;AAC7D,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,YAAY;AAC3B,qBAAkB,IAAI,YAAY,aAAa,WAAW;;;CAK9D,MAAM,EACJ,MAAM,cACN,aACA,iBACEC,4CAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACA,aAAa,kBAAkB,OAAO,IAAI,oBAAoB;EAC/D,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,2CAAM,8BAAuB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,yDAA2B,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,8BAAW,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,gCAAiB,QAAQ;EAK/B,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,+BAAgB,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,+BAAW,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,0DAA4B,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,6BAAW,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,8BAAc,OAAO,GACrB,8BAAc,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,6BAAW,aAAa,MAAM;;AAGhC,2BAAU;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
1
+ {"version":3,"file":"index.cjs","names":["lines: string[]","createSchemaIndex","inlineFragments: ParsedInlineFragment[]","otherSelections: ParsedSelection[]","entries: string[]","argEntries: string[]","values: string[]","operations: ParsedOperation[]","fragments: ParsedFragment[]","Kind","variables: ParsedVariable[]","levels: TypeLevel[]","lists: (\"[]!\" | \"[]?\")[]","match: RegExpExecArray | null","mergedInner: \"!\" | \"?\"","mergedLists: (\"[]!\" | \"[]?\")[]","usages: VariableUsage[]","error","variables: InferredVariable[]","sorted: ParsedFragment[]","createSchemaIndex","fragments: EnrichedFragment[]","operations: EnrichedOperation[]","variables: EnrichedVariable[]","spreadVariables: InferredVariable[]","esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","generateMultiSchemaModule","defsPaths: string[]","defaultBundler"],"sources":["../src/graphql-compat/emitter.ts","../src/graphql-compat/parser.ts","../src/graphql-compat/transformer.ts","../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["/**\n * Emitter for generating TypeScript compat code from enriched operations.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport type { EnrichedFragment, EnrichedOperation } from \"./transformer\";\nimport type { GraphqlCompatError, ParsedArgument, ParsedInlineFragment, ParsedSelection, ParsedValue } from \"./types\";\n\n/**\n * Schema index for type lookups.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Options for code emission.\n */\nexport type EmitOptions = {\n /** Schema name to use in gql.schemaName() call */\n readonly schemaName: string;\n /** Import path for graphql-system module */\n readonly graphqlSystemPath: string;\n /** Schema document for type lookups (required for inline fragment support) */\n readonly schemaDocument?: DocumentNode;\n};\n\n/**\n * Emit TypeScript code for an operation.\n */\nexport const emitOperation = (operation: EnrichedOperation, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${operation.name}Compat`;\n const operationType = operation.kind;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);\n lines.push(` ${operationType}.compat({`);\n lines.push(` name: ${JSON.stringify(operation.name)},`);\n\n // Variables\n if (operation.variables.length > 0) {\n lines.push(` variables: { ${emitVariables(operation.variables)} },`);\n }\n\n // Fields\n lines.push(` fields: ({ f, $ }) => ({`);\n const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit TypeScript code for a fragment.\n */\nexport const emitFragment = (fragment: EnrichedFragment, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n const hasVariables = fragment.variables.length > 0;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${fragment.name}Fragment`;\n\n // Include $var in destructure if fragment has variables\n const destructure = hasVariables ? \"fragment, $var\" : \"fragment\";\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${destructure} }) =>`);\n lines.push(` fragment.${fragment.onType}({`);\n\n // Variables block (if any)\n if (hasVariables) {\n lines.push(` variables: { ${emitVariables(fragment.variables)} },`);\n }\n\n // Fields - include $ in context if fragment has variables\n const fieldsContext = hasVariables ? \"{ f, $ }\" : \"{ f }\";\n lines.push(` fields: (${fieldsContext}) => ({`);\n const fieldLinesResult = emitSelections(fragment.selections, 3, fragment.variables, schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Common variable type for emission (both EnrichedVariable and InferredVariable have these fields).\n */\ntype EmittableVariable = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n};\n\n/**\n * Emit variable definitions.\n */\nconst emitVariables = (variables: readonly EmittableVariable[]): string => {\n return variables.map((v) => `...$var(${JSON.stringify(v.name)}).${v.typeName}(${JSON.stringify(v.modifier)})`).join(\", \");\n};\n\n/**\n * Emit field selections (public API).\n * Converts variable array to Set<string> and delegates to internal implementation.\n */\nconst emitSelections = (\n selections: readonly ParsedSelection[],\n indent: number,\n variables: readonly EmittableVariable[],\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const variableNames = new Set(variables.map((v) => v.name));\n return emitSelectionsInternal(selections, indent, variableNames, schema);\n};\n\n/**\n * Internal implementation for emitting field selections.\n * Takes variableNames as Set<string> for recursive calls.\n */\nconst emitSelectionsInternal = (\n selections: readonly ParsedSelection[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n\n // Separate inline fragments from other selections\n const inlineFragments: ParsedInlineFragment[] = [];\n const otherSelections: ParsedSelection[] = [];\n\n for (const sel of selections) {\n if (sel.kind === \"inlineFragment\") {\n inlineFragments.push(sel);\n } else {\n otherSelections.push(sel);\n }\n }\n\n // Emit regular selections (fields and fragment spreads)\n for (const sel of otherSelections) {\n const result = emitSingleSelection(sel, indent, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n lines.push(result.value);\n }\n\n // Emit grouped inline fragments as union selections\n if (inlineFragments.length > 0) {\n const unionResult = emitInlineFragmentsAsUnion(inlineFragments, indent, variableNames, schema);\n if (unionResult.isErr()) {\n return err(unionResult.error);\n }\n lines.push(unionResult.value);\n }\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit a single selection (field or fragment spread).\n */\nconst emitSingleSelection = (\n sel: ParsedSelection,\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n switch (sel.kind) {\n case \"field\":\n return emitFieldSelection(sel, indent, variableNames, schema);\n case \"fragmentSpread\":\n return ok(`${padding}...${sel.name}Fragment.spread(),`);\n case \"inlineFragment\":\n // This should not happen as inline fragments are handled separately\n return ok(\"\");\n }\n};\n\n/**\n * Emit inline fragments grouped as a union selection.\n * Format: { TypeA: ({ f }) => ({ ...fields }), TypeB: ({ f }) => ({ ...fields }) }\n */\nconst emitInlineFragmentsAsUnion = (\n inlineFragments: readonly ParsedInlineFragment[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Validate inline fragments have type conditions\n for (const frag of inlineFragments) {\n if (frag.onType === \"\") {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_WITHOUT_TYPE\",\n message: \"Inline fragments without type condition are not supported. Use `... on TypeName { }` syntax.\",\n });\n }\n }\n\n // Validate all inline fragments are on union types (not interfaces)\n for (const frag of inlineFragments) {\n if (schema && !schema.objects.has(frag.onType)) {\n // If it's not a known object type, it might be an interface\n // Check if any union contains this type as a member\n let isUnionMember = false;\n for (const [, unionDef] of schema.unions) {\n if (unionDef.members.has(frag.onType)) {\n isUnionMember = true;\n break;\n }\n }\n if (!isUnionMember) {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_ON_INTERFACE\",\n message: `Inline fragments on interface type \"${frag.onType}\" are not supported. Use union types instead.`,\n onType: frag.onType,\n });\n }\n }\n }\n\n // Build union member entries\n const entries: string[] = [];\n for (const frag of inlineFragments) {\n const innerPadding = \" \".repeat(indent + 1);\n const fieldsResult = emitSelectionsInternal(frag.selections, indent + 2, variableNames, schema);\n if (fieldsResult.isErr()) {\n return err(fieldsResult.error);\n }\n\n entries.push(`${innerPadding}${frag.onType}: ({ f }) => ({\n${fieldsResult.value}\n${innerPadding}}),`);\n }\n\n // Emit as spread with union callback: ...f.fieldName()({ Type: ... })\n // Note: This assumes the parent field handles the union - we emit just the union object\n return ok(`${padding}...({\n${entries.join(\"\\n\")}\n${padding}}),`);\n};\n\n/**\n * Emit a single field selection.\n */\nconst emitFieldSelection = (\n field: ParsedSelection & { kind: \"field\" },\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Extract optional fields for type narrowing\n const args = field.arguments;\n const selections = field.selections;\n const hasArgs = args && args.length > 0;\n const hasSelections = selections && selections.length > 0;\n\n // Use shorthand syntax for scalar fields (no args, no nested selections)\n if (!hasArgs && !hasSelections) {\n return ok(`${padding}${field.name}: true,`);\n }\n\n let line = `${padding}...f.${field.name}(`;\n\n if (hasArgs) {\n const argsResult = emitArguments(args, variableNames);\n if (argsResult.isErr()) {\n return err(argsResult.error);\n }\n line += argsResult.value;\n }\n\n line += \")\";\n\n if (hasSelections) {\n // Check if selections contain inline fragments (union field)\n const hasInlineFragments = selections.some((s) => s.kind === \"inlineFragment\");\n\n if (hasInlineFragments) {\n // Union field: emit with union callback pattern\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += \"({\\n\";\n line += `${nestedResult.value}\\n`;\n line += `${padding}})`;\n } else {\n // Regular nested selections\n line += \"(({ f }) => ({\\n\";\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += `${nestedResult.value}\\n`;\n line += `${padding}}))`;\n }\n }\n\n line += \",\";\n\n return ok(line);\n};\n\n/**\n * Emit field arguments.\n */\nconst emitArguments = (args: readonly ParsedArgument[], variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n if (args.length === 0) {\n return ok(\"\");\n }\n\n const argEntries: string[] = [];\n for (const arg of args) {\n const result = emitValue(arg.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n argEntries.push(`${arg.name}: ${result.value}`);\n }\n return ok(`{ ${argEntries.join(\", \")} }`);\n};\n\n/**\n * Emit a value (literal or variable reference).\n */\nconst emitValue = (value: ParsedValue, variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n switch (value.kind) {\n case \"variable\":\n // Check if it's a declared variable\n if (variableNames.has(value.name)) {\n return ok(`$.${value.name}`);\n }\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `Variable \"$${value.name}\" is not declared in the operation`,\n variableName: value.name,\n });\n case \"int\":\n case \"float\":\n return ok(value.value);\n case \"string\":\n return ok(JSON.stringify(value.value));\n case \"boolean\":\n return ok(value.value ? \"true\" : \"false\");\n case \"null\":\n return ok(\"null\");\n case \"enum\":\n // Enums are emitted as string literals in soda-gql\n return ok(JSON.stringify(value.value));\n case \"list\": {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValue(v, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n case \"object\": {\n if (value.fields.length === 0) {\n return ok(\"{}\");\n }\n const entries: string[] = [];\n for (const f of value.fields) {\n const result = emitValue(f.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n return ok(`{ ${entries.join(\", \")} }`);\n }\n }\n};\n","/**\n * Parser for .graphql operation files.\n * Extracts operations and fragments from GraphQL documents.\n * @module\n */\n\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport {\n type ArgumentNode,\n type DocumentNode,\n type FieldNode,\n type FragmentDefinitionNode,\n type FragmentSpreadNode,\n type InlineFragmentNode,\n Kind,\n type OperationDefinitionNode,\n parse,\n type SelectionNode,\n type TypeNode,\n type ValueNode,\n type VariableDefinitionNode,\n} from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\n\nimport type {\n GraphqlCompatError,\n ParsedArgument,\n ParsedFieldSelection,\n ParsedFragment,\n ParsedFragmentSpread,\n ParsedInlineFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n} from \"./types\";\n\n/**\n * Parse a single .graphql file and extract operations and fragments.\n */\nexport const parseGraphqlFile = (filePath: string): Result<ParseResult, GraphqlCompatError> => {\n const resolvedPath = resolve(filePath);\n\n if (!existsSync(resolvedPath)) {\n return err({\n code: \"GRAPHQL_FILE_NOT_FOUND\",\n message: `GraphQL file not found at ${resolvedPath}`,\n filePath: resolvedPath,\n });\n }\n\n try {\n const source = readFileSync(resolvedPath, \"utf8\");\n const document = parse(source);\n return ok(extractFromDocument(document, resolvedPath));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: resolvedPath,\n });\n }\n};\n\n/**\n * Parse GraphQL source string directly.\n */\nexport const parseGraphqlSource = (source: string, sourceFile: string): Result<ParseResult, GraphqlCompatError> => {\n try {\n const document = parse(source);\n return ok(extractFromDocument(document, sourceFile));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: sourceFile,\n });\n }\n};\n\n/**\n * Extract operations and fragments from a parsed GraphQL document.\n */\nconst extractFromDocument = (document: DocumentNode, sourceFile: string): ParseResult => {\n const operations: ParsedOperation[] = [];\n const fragments: ParsedFragment[] = [];\n\n for (const definition of document.definitions) {\n if (definition.kind === Kind.OPERATION_DEFINITION) {\n const operation = extractOperation(definition, sourceFile);\n if (operation) {\n operations.push(operation);\n }\n } else if (definition.kind === Kind.FRAGMENT_DEFINITION) {\n fragments.push(extractFragment(definition, sourceFile));\n }\n }\n\n return { operations, fragments };\n};\n\n/**\n * Extract a single operation from an OperationDefinitionNode.\n */\nconst extractOperation = (node: OperationDefinitionNode, sourceFile: string): ParsedOperation | null => {\n // Anonymous operations are not supported\n if (!node.name) {\n return null;\n }\n\n const variables: ParsedVariable[] = (node.variableDefinitions ?? []).map(extractVariable);\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n kind: node.operation,\n name: node.name.value,\n variables,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a fragment from a FragmentDefinitionNode.\n */\nconst extractFragment = (node: FragmentDefinitionNode, sourceFile: string): ParsedFragment => {\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n name: node.name.value,\n onType: node.typeCondition.name.value,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a variable definition.\n */\nconst extractVariable = (node: VariableDefinitionNode): ParsedVariable => {\n const { typeName, modifier } = parseTypeNode(node.type);\n const defaultValue = node.defaultValue ? extractValue(node.defaultValue) : undefined;\n\n return {\n name: node.variable.name.value,\n typeName,\n modifier,\n // Type kind will be determined later when we have schema context\n typeKind: \"scalar\",\n defaultValue,\n };\n};\n\n/**\n * Parse a GraphQL TypeNode into type name and modifier.\n *\n * Format: inner nullability + list modifiers\n * - Inner: `!` (non-null) or `?` (nullable)\n * - List: `[]!` (non-null list) or `[]?` (nullable list)\n */\nexport const parseTypeNode = (node: TypeNode): { typeName: string; modifier: string } => {\n type TypeLevel = { kind: \"list\" | \"named\"; nonNull: boolean };\n const levels: TypeLevel[] = [];\n\n const collect = (n: TypeNode, nonNull: boolean): string => {\n if (n.kind === Kind.NON_NULL_TYPE) {\n return collect(n.type, true);\n }\n if (n.kind === Kind.LIST_TYPE) {\n levels.push({ kind: \"list\", nonNull });\n return collect(n.type, false);\n }\n levels.push({ kind: \"named\", nonNull });\n return n.name.value;\n };\n\n const typeName = collect(node, false);\n\n // Build modifier from levels (reverse order)\n let modifier = \"?\";\n for (const level of levels.slice().reverse()) {\n if (level.kind === \"named\") {\n modifier = level.nonNull ? \"!\" : \"?\";\n continue;\n }\n const listSuffix = level.nonNull ? \"[]!\" : \"[]?\";\n modifier = `${modifier}${listSuffix}`;\n }\n\n return { typeName, modifier };\n};\n\n/**\n * Extract selections from a SelectionSet.\n */\nconst extractSelections = (selections: readonly SelectionNode[]): ParsedSelection[] => {\n return selections.map(extractSelection);\n};\n\n/**\n * Extract a single selection.\n */\nconst extractSelection = (node: SelectionNode): ParsedSelection => {\n switch (node.kind) {\n case Kind.FIELD:\n return extractFieldSelection(node);\n case Kind.FRAGMENT_SPREAD:\n return extractFragmentSpread(node);\n case Kind.INLINE_FRAGMENT:\n return extractInlineFragment(node);\n }\n};\n\n/**\n * Extract a field selection.\n */\nconst extractFieldSelection = (node: FieldNode): ParsedFieldSelection => {\n const args = node.arguments?.length ? node.arguments.map(extractArgument) : undefined;\n const selections = node.selectionSet ? extractSelections(node.selectionSet.selections) : undefined;\n\n return {\n kind: \"field\",\n name: node.name.value,\n alias: node.alias?.value,\n arguments: args,\n selections,\n };\n};\n\n/**\n * Extract a fragment spread.\n */\nconst extractFragmentSpread = (node: FragmentSpreadNode): ParsedFragmentSpread => {\n return {\n kind: \"fragmentSpread\",\n name: node.name.value,\n };\n};\n\n/**\n * Extract an inline fragment.\n */\nconst extractInlineFragment = (node: InlineFragmentNode): ParsedInlineFragment => {\n return {\n kind: \"inlineFragment\",\n onType: node.typeCondition?.name.value ?? \"\",\n selections: extractSelections(node.selectionSet.selections),\n };\n};\n\n/**\n * Extract an argument.\n */\nconst extractArgument = (node: ArgumentNode): ParsedArgument => {\n return {\n name: node.name.value,\n value: extractValue(node.value),\n };\n};\n\n/**\n * Assert unreachable code path (for exhaustiveness checks).\n */\nconst assertUnreachable = (value: never): never => {\n throw new Error(`Unexpected value: ${JSON.stringify(value)}`);\n};\n\n/**\n * Extract a value (literal or variable reference).\n */\nconst extractValue = (node: ValueNode): ParsedValue => {\n switch (node.kind) {\n case Kind.VARIABLE:\n return { kind: \"variable\", name: node.name.value };\n case Kind.INT:\n return { kind: \"int\", value: node.value };\n case Kind.FLOAT:\n return { kind: \"float\", value: node.value };\n case Kind.STRING:\n return { kind: \"string\", value: node.value };\n case Kind.BOOLEAN:\n return { kind: \"boolean\", value: node.value };\n case Kind.NULL:\n return { kind: \"null\" };\n case Kind.ENUM:\n return { kind: \"enum\", value: node.value };\n case Kind.LIST:\n return { kind: \"list\", values: node.values.map(extractValue) };\n case Kind.OBJECT:\n return {\n kind: \"object\",\n fields: node.fields.map((field) => ({\n name: field.name.value,\n value: extractValue(field.value),\n })),\n };\n default:\n return assertUnreachable(node);\n }\n};\n","/**\n * Transformer for enriching parsed GraphQL operations with schema information.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport { parseTypeNode } from \"./parser\";\nimport type {\n GraphqlCompatError,\n InferredVariable,\n ParsedArgument,\n ParsedFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n} from \"./types\";\n\n/**\n * Schema index type extracted from generator.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Built-in GraphQL scalar types.\n */\nconst builtinScalarTypes = new Set([\"ID\", \"String\", \"Int\", \"Float\", \"Boolean\"]);\n\n// ============================================================================\n// Modifier Merging Utilities\n// ============================================================================\n\n/**\n * Parsed structure of a modifier for comparison and merging.\n * Example: \"![]?\" -> { inner: \"!\", lists: [\"[]?\"] }\n */\ntype ModifierStructure = {\n readonly inner: \"!\" | \"?\";\n readonly lists: readonly (\"[]!\" | \"[]?\")[];\n};\n\n/**\n * Parse a modifier string into its structural components.\n * @param modifier - Modifier string like \"!\", \"?\", \"![]!\", \"?[]?[]!\"\n * @returns Parsed structure with inner nullability and list modifiers\n */\nconst parseModifierStructure = (modifier: string): ModifierStructure => {\n // Extract inner nullability (first character)\n const inner = modifier[0] === \"!\" ? \"!\" : \"?\";\n\n // Extract list modifiers ([]! or []?)\n const lists: (\"[]!\" | \"[]?\")[] = [];\n const listPattern = /\\[\\]([!?])/g;\n let match: RegExpExecArray | null;\n while ((match = listPattern.exec(modifier)) !== null) {\n lists.push(`[]${match[1]}` as \"[]!\" | \"[]?\");\n }\n\n return { inner, lists };\n};\n\n/**\n * Rebuild modifier string from structure.\n */\nconst buildModifier = (structure: ModifierStructure): string => {\n return structure.inner + structure.lists.join(\"\");\n};\n\n/**\n * Merge two modifiers by taking the stricter constraint at each level.\n * - Non-null (!) is stricter than nullable (?)\n * - List depths must match\n *\n * @param a - First modifier\n * @param b - Second modifier\n * @returns Merged modifier or error if incompatible\n */\nexport const mergeModifiers = (a: string, b: string): { ok: true; value: string } | { ok: false; reason: string } => {\n const structA = parseModifierStructure(a);\n const structB = parseModifierStructure(b);\n\n // List depths must match\n if (structA.lists.length !== structB.lists.length) {\n return {\n ok: false,\n reason: `Incompatible list depths: \"${a}\" has ${structA.lists.length} list level(s), \"${b}\" has ${structB.lists.length}`,\n };\n }\n\n // Take stricter inner constraint (! beats ?)\n const mergedInner: \"!\" | \"?\" = structA.inner === \"!\" || structB.inner === \"!\" ? \"!\" : \"?\";\n\n // Merge each list level (! beats ?)\n const mergedLists: (\"[]!\" | \"[]?\")[] = [];\n for (let i = 0; i < structA.lists.length; i++) {\n const listA = structA.lists[i]!;\n const listB = structB.lists[i]!;\n mergedLists.push(listA === \"[]!\" || listB === \"[]!\" ? \"[]!\" : \"[]?\");\n }\n\n return { ok: true, value: buildModifier({ inner: mergedInner, lists: mergedLists }) };\n};\n\n// ============================================================================\n// Variable Collection from Selections\n// ============================================================================\n\n/**\n * Intermediate type for tracking variable usages before merging.\n */\ntype VariableUsage = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Get the expected type for a field argument from the schema.\n * Returns null if the field or argument is not found.\n */\nconst getArgumentType = (\n schema: SchemaIndex,\n parentTypeName: string,\n fieldName: string,\n argumentName: string,\n): { typeName: string; modifier: string } | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const argDef = fieldDef.arguments?.find((arg) => arg.name.value === argumentName);\n if (!argDef) return null;\n\n return parseTypeNode(argDef.type);\n};\n\n/**\n * Get the expected type for an input object field from the schema.\n */\nconst getInputFieldType = (\n schema: SchemaIndex,\n inputTypeName: string,\n fieldName: string,\n): { typeName: string; modifier: string } | null => {\n const inputRecord = schema.inputs.get(inputTypeName);\n if (!inputRecord) return null;\n\n const fieldDef = inputRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n return parseTypeNode(fieldDef.type);\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKindFromName = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) return \"scalar\";\n if (isEnumName(schema, typeName)) return \"enum\";\n if (schema.inputs.has(typeName)) return \"input\";\n return null;\n};\n\n/**\n * Extract variable usages from a parsed value, given the expected type.\n * Handles nested input objects recursively.\n */\nconst collectVariablesFromValue = (\n value: ParsedValue,\n expectedTypeName: string,\n expectedModifier: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n if (value.kind === \"variable\") {\n const typeKind = resolveTypeKindFromName(schema, expectedTypeName);\n if (!typeKind) {\n return {\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${expectedTypeName}\" for variable \"$${value.name}\"`,\n typeName: expectedTypeName,\n };\n }\n usages.push({\n name: value.name,\n typeName: expectedTypeName,\n modifier: expectedModifier,\n typeKind,\n });\n return null;\n }\n\n if (value.kind === \"object\") {\n // For object values, check each field against input type definition\n for (const field of value.fields) {\n const fieldType = getInputFieldType(schema, expectedTypeName, field.name);\n if (!fieldType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${field.name}\" on input type \"${expectedTypeName}\"`,\n typeName: expectedTypeName,\n fieldName: field.name,\n };\n }\n const error = collectVariablesFromValue(field.value, fieldType.typeName, fieldType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n }\n\n if (value.kind === \"list\") {\n // For list values, unwrap one level of list modifier\n // e.g., [ID!]! -> ID! for elements\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n const innerModifier = buildModifier({\n inner: struct.inner,\n lists: struct.lists.slice(1),\n });\n for (const item of value.values) {\n const error = collectVariablesFromValue(item, expectedTypeName, innerModifier, schema, usages);\n if (error) return error;\n }\n }\n }\n\n // Other value kinds (int, float, string, etc.) don't contain variables\n return null;\n};\n\n/**\n * Collect variable usages from field arguments.\n */\nconst collectVariablesFromArguments = (\n args: readonly ParsedArgument[],\n parentTypeName: string,\n fieldName: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n for (const arg of args) {\n const argType = getArgumentType(schema, parentTypeName, fieldName, arg.name);\n if (!argType) {\n return {\n code: \"GRAPHQL_UNKNOWN_ARGUMENT\",\n message: `Unknown argument \"${arg.name}\" on field \"${fieldName}\"`,\n fieldName,\n argumentName: arg.name,\n };\n }\n const error = collectVariablesFromValue(arg.value, argType.typeName, argType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n};\n\n/**\n * Recursively collect all variable usages from selections.\n */\nexport const collectVariableUsages = (\n selections: readonly ParsedSelection[],\n parentTypeName: string,\n schema: SchemaIndex,\n): Result<VariableUsage[], GraphqlCompatError> => {\n const usages: VariableUsage[] = [];\n\n const collect = (sels: readonly ParsedSelection[], parentType: string): GraphqlCompatError | null => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"field\": {\n // Collect from arguments\n if (sel.arguments && sel.arguments.length > 0) {\n const error = collectVariablesFromArguments(sel.arguments, parentType, sel.name, schema, usages);\n if (error) return error;\n }\n\n // Recurse into nested selections\n if (sel.selections && sel.selections.length > 0) {\n // Need to determine the field's return type for nested selections\n const fieldReturnType = getFieldReturnType(schema, parentType, sel.name);\n if (!fieldReturnType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${sel.name}\" on type \"${parentType}\"`,\n typeName: parentType,\n fieldName: sel.name,\n };\n }\n const error = collect(sel.selections, fieldReturnType);\n if (error) return error;\n }\n break;\n }\n case \"inlineFragment\": {\n // Use the inline fragment's type condition\n const error = collect(sel.selections, sel.onType);\n if (error) return error;\n break;\n }\n case \"fragmentSpread\":\n // Fragment spreads are handled separately (variables from spread fragments)\n break;\n }\n }\n return null;\n };\n\n const error = collect(selections, parentTypeName);\n if (error) return err(error);\n\n return ok(usages);\n};\n\n/**\n * Get the return type of a field (unwrapped from modifiers).\n */\nconst getFieldReturnType = (schema: SchemaIndex, parentTypeName: string, fieldName: string): string | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const { typeName } = parseTypeNode(fieldDef.type);\n return typeName;\n};\n\n/**\n * Merge multiple variable usages into a single InferredVariable.\n * Validates type compatibility and merges modifiers using stricter constraint.\n */\nexport const mergeVariableUsages = (\n variableName: string,\n usages: readonly VariableUsage[],\n): Result<InferredVariable, GraphqlCompatError> => {\n if (usages.length === 0) {\n // This shouldn't happen, but handle defensively\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `No usages found for variable \"${variableName}\"`,\n variableName,\n });\n }\n\n const first = usages[0]!;\n\n // Validate all usages have the same type name\n for (const usage of usages) {\n if (usage.typeName !== first.typeName) {\n return err({\n code: \"GRAPHQL_VARIABLE_TYPE_MISMATCH\",\n message: `Variable \"$${variableName}\" has conflicting types: \"${first.typeName}\" and \"${usage.typeName}\"`,\n variableName,\n });\n }\n }\n\n // Merge modifiers\n let mergedModifier = first.modifier;\n for (let i = 1; i < usages.length; i++) {\n const result = mergeModifiers(mergedModifier, usages[i]!.modifier);\n if (!result.ok) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" has incompatible modifiers: ${result.reason}`,\n variableName,\n });\n }\n mergedModifier = result.value;\n }\n\n return ok({\n name: variableName,\n typeName: first.typeName,\n modifier: mergedModifier,\n typeKind: first.typeKind,\n });\n};\n\n/**\n * Infer variables from collected usages.\n * Groups by variable name and merges each group.\n */\nexport const inferVariablesFromUsages = (usages: readonly VariableUsage[]): Result<InferredVariable[], GraphqlCompatError> => {\n // Group usages by variable name\n const byName = new Map<string, VariableUsage[]>();\n for (const usage of usages) {\n const existing = byName.get(usage.name);\n if (existing) {\n existing.push(usage);\n } else {\n byName.set(usage.name, [usage]);\n }\n }\n\n // Merge each group\n const variables: InferredVariable[] = [];\n for (const [name, group] of byName) {\n const result = mergeVariableUsages(name, group);\n if (result.isErr()) return err(result.error);\n variables.push(result.value);\n }\n\n // Sort by name for deterministic output\n variables.sort((a, b) => a.name.localeCompare(b.name));\n\n return ok(variables);\n};\n\n/**\n * Check if a type name is a scalar type.\n */\nconst isScalarName = (schema: SchemaIndex, name: string): boolean => builtinScalarTypes.has(name) || schema.scalars.has(name);\n\n// ============================================================================\n// Fragment Dependency Ordering\n// ============================================================================\n\n/**\n * Topologically sort fragments so dependencies come before dependents.\n * Detects circular dependencies.\n *\n * Note: Uses the existing collectFragmentDependencies function defined below.\n */\nexport const sortFragmentsByDependency = (fragments: readonly ParsedFragment[]): Result<ParsedFragment[], GraphqlCompatError> => {\n // Build dependency graph using existing function\n const graph = new Map<string, Set<string>>();\n for (const fragment of fragments) {\n const deps = collectFragmentDependenciesSet(fragment.selections);\n graph.set(fragment.name, deps);\n }\n\n const fragmentByName = new Map<string, ParsedFragment>();\n for (const f of fragments) {\n fragmentByName.set(f.name, f);\n }\n\n const sorted: ParsedFragment[] = [];\n const visited = new Set<string>();\n const visiting = new Set<string>(); // For cycle detection\n\n const visit = (name: string, path: string[]): GraphqlCompatError | null => {\n if (visited.has(name)) return null;\n\n if (visiting.has(name)) {\n // Found a cycle\n const cycleStart = path.indexOf(name);\n const cycle = path.slice(cycleStart).concat(name);\n return {\n code: \"GRAPHQL_FRAGMENT_CIRCULAR_DEPENDENCY\",\n message: `Circular dependency detected in fragments: ${cycle.join(\" -> \")}`,\n fragmentNames: cycle,\n };\n }\n\n // Fragment might not be in our list (external dependency)\n const fragment = fragmentByName.get(name);\n if (!fragment) {\n // External fragment, skip\n visited.add(name);\n return null;\n }\n\n visiting.add(name);\n const deps = graph.get(name) ?? new Set();\n\n for (const dep of deps) {\n const error = visit(dep, [...path, name]);\n if (error) return error;\n }\n\n visiting.delete(name);\n visited.add(name);\n sorted.push(fragment);\n return null;\n };\n\n for (const fragment of fragments) {\n const error = visit(fragment.name, []);\n if (error) return err(error);\n }\n\n return ok(sorted);\n};\n\n/**\n * Recursively collect fragment spread names from selections into a Set.\n * Internal helper for sortFragmentsByDependency.\n */\nconst collectFragmentDependenciesSet = (selections: readonly ParsedSelection[]): Set<string> => {\n const deps = new Set<string>();\n\n const collect = (sels: readonly ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n deps.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return deps;\n};\n\n/**\n * Check if a type name is an enum type.\n */\nconst isEnumName = (schema: SchemaIndex, name: string): boolean => schema.enums.has(name);\n\n/**\n * Enriched operation with resolved type information.\n */\nexport type EnrichedOperation = Omit<ParsedOperation, \"variables\"> & {\n readonly variables: readonly EnrichedVariable[];\n /** Fragment names used in this operation (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched fragment with resolved type information.\n */\nexport type EnrichedFragment = ParsedFragment & {\n /** Fragment names used in this fragment (for imports) */\n readonly fragmentDependencies: readonly string[];\n /** Variables inferred from field arguments in this fragment */\n readonly variables: readonly InferredVariable[];\n};\n\n/**\n * Enriched variable with resolved type kind.\n */\nexport type EnrichedVariable = Omit<ParsedVariable, \"typeKind\"> & {\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Result of transforming parsed operations.\n */\nexport type TransformResult = {\n readonly operations: readonly EnrichedOperation[];\n readonly fragments: readonly EnrichedFragment[];\n};\n\n/**\n * Options for transformation.\n */\nexport type TransformOptions = {\n /** Schema document for type resolution */\n readonly schemaDocument: DocumentNode;\n};\n\n/**\n * Transform parsed operations/fragments by enriching them with schema information.\n *\n * This resolves variable type kinds (scalar, enum, input), collects\n * fragment dependencies, and infers variables for fragments.\n */\nexport const transformParsedGraphql = (\n parsed: ParseResult,\n options: TransformOptions,\n): Result<TransformResult, GraphqlCompatError> => {\n const schema = createSchemaIndex(options.schemaDocument);\n\n // Sort fragments by dependency (dependencies first)\n const sortResult = sortFragmentsByDependency(parsed.fragments);\n if (sortResult.isErr()) {\n return err(sortResult.error);\n }\n const sortedFragments = sortResult.value;\n\n // Transform fragments in dependency order, building up resolved variables map\n const resolvedFragmentVariables = new Map<string, readonly InferredVariable[]>();\n const fragments: EnrichedFragment[] = [];\n\n for (const frag of sortedFragments) {\n const result = transformFragment(frag, schema, resolvedFragmentVariables);\n if (result.isErr()) {\n return err(result.error);\n }\n resolvedFragmentVariables.set(frag.name, result.value.variables);\n fragments.push(result.value);\n }\n\n // Transform operations\n const operations: EnrichedOperation[] = [];\n for (const op of parsed.operations) {\n const result = transformOperation(op, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n operations.push(result.value);\n }\n\n return ok({ operations, fragments });\n};\n\n/**\n * Transform a single operation.\n */\nconst transformOperation = (op: ParsedOperation, schema: SchemaIndex): Result<EnrichedOperation, GraphqlCompatError> => {\n // Resolve variable type kinds\n const variables: EnrichedVariable[] = [];\n for (const v of op.variables) {\n const typeKind = resolveTypeKind(schema, v.typeName);\n if (typeKind === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${v.typeName}\" in variable \"${v.name}\"`,\n typeName: v.typeName,\n });\n }\n variables.push({ ...v, typeKind });\n }\n\n // Collect fragment dependencies\n const fragmentDependencies = collectFragmentDependencies(op.selections);\n\n return ok({\n ...op,\n variables,\n fragmentDependencies,\n });\n};\n\n/**\n * Transform a single fragment.\n * Infers variables from field arguments and propagates variables from spread fragments.\n */\nconst transformFragment = (\n frag: ParsedFragment,\n schema: SchemaIndex,\n resolvedFragmentVariables: Map<string, readonly InferredVariable[]>,\n): Result<EnrichedFragment, GraphqlCompatError> => {\n // Collect fragment dependencies (fragments used within this fragment)\n const fragmentDependencies = collectFragmentDependencies(frag.selections);\n\n // Collect direct variable usages from this fragment's selections\n const directUsagesResult = collectVariableUsages(frag.selections, frag.onType, schema);\n if (directUsagesResult.isErr()) {\n return err(directUsagesResult.error);\n }\n const directUsages = directUsagesResult.value;\n\n // Collect variables from spread fragments\n const spreadVariables: InferredVariable[] = [];\n for (const depName of fragmentDependencies) {\n const depVariables = resolvedFragmentVariables.get(depName);\n if (depVariables) {\n spreadVariables.push(...depVariables);\n }\n // If not found, it's an external fragment - skip\n }\n\n // Combine direct usages with spread variables\n // Convert spread variables to usages for merging\n const allUsages = [\n ...directUsages,\n ...spreadVariables.map((v) => ({\n name: v.name,\n typeName: v.typeName,\n modifier: v.modifier,\n typeKind: v.typeKind,\n })),\n ];\n\n // Infer final variables\n const variablesResult = inferVariablesFromUsages(allUsages);\n if (variablesResult.isErr()) {\n return err(variablesResult.error);\n }\n\n return ok({\n ...frag,\n fragmentDependencies,\n variables: variablesResult.value,\n });\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKind = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) {\n return \"scalar\";\n }\n if (isEnumName(schema, typeName)) {\n return \"enum\";\n }\n if (schema.inputs.has(typeName)) {\n return \"input\";\n }\n return null;\n};\n\n/**\n * Collect fragment names used in selections (recursively).\n */\nconst collectFragmentDependencies = (selections: readonly import(\"./types\").ParsedSelection[]): readonly string[] => {\n const fragments = new Set<string>();\n\n const collect = (sels: readonly import(\"./types\").ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n fragments.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return [...fragments];\n};\n","import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, rmSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const removeDirectory = (dirPath: string) => {\n const targetPath = resolve(dirPath);\n try {\n rmSync(targetPath, { recursive: true, force: true });\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"REMOVE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport type { TypeFilterConfig } from \"@soda-gql/config\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { removeDirectory, writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Build typeFilters config for each schema\n const typeFiltersConfig = new Map<string, TypeFilterConfig>();\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.typeFilter) {\n typeFiltersConfig.set(schemaName, schemaConfig.typeFilter);\n }\n }\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n typeFilters: typeFiltersConfig.size > 0 ? typeFiltersConfig : undefined,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Clean up existing _defs directory to prevent stale files\n const defsDir = join(outDir, \"_defs\");\n if (existsSync(defsDir)) {\n const removeResult = removeDirectory(defsDir);\n if (removeResult.isErr()) {\n return err(removeResult.error);\n }\n }\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;;;;AA+BA,MAAa,iBAAiB,WAA8B,YAA6D;CACvH,MAAMA,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CAKpF,MAAM,aAAa,GAAG,UAAU,KAAK;CACrC,MAAM,gBAAgB,UAAU;AAEhC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,cAAc,cAAc;AACpG,OAAM,KAAK,KAAK,cAAc,WAAW;AACzC,OAAM,KAAK,aAAa,KAAK,UAAU,UAAU,KAAK,CAAC,GAAG;AAG1D,KAAI,UAAU,UAAU,SAAS,GAAG;AAClC,QAAM,KAAK,oBAAoB,cAAc,UAAU,UAAU,CAAC,KAAK;;AAIzE,OAAM,KAAK,+BAA+B;CAC1C,MAAM,mBAAmB,eAAe,UAAU,YAAY,GAAG,UAAU,WAAW,OAAO;AAC7F,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAa,gBAAgB,UAA4B,YAA6D;CACpH,MAAMD,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CACpF,MAAM,eAAe,SAAS,UAAU,SAAS;CAKjD,MAAM,aAAa,GAAG,SAAS,KAAK;CAGpC,MAAM,cAAc,eAAe,mBAAmB;AACtD,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,YAAY,QAAQ;AAC5F,OAAM,KAAK,cAAc,SAAS,OAAO,IAAI;AAG7C,KAAI,cAAc;AAChB,QAAM,KAAK,oBAAoB,cAAc,SAAS,UAAU,CAAC,KAAK;;CAIxE,MAAM,gBAAgB,eAAe,aAAa;AAClD,OAAM,KAAK,gBAAgB,cAAc,SAAS;CAClD,MAAM,mBAAmB,eAAe,SAAS,YAAY,GAAG,SAAS,WAAW,OAAO;AAC3F,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAe7B,MAAM,iBAAiB,cAAoD;AACzE,QAAO,UAAU,KAAK,MAAM,WAAW,KAAK,UAAU,EAAE,KAAK,CAAC,IAAI,EAAE,SAAS,GAAG,KAAK,UAAU,EAAE,SAAS,CAAC,GAAG,CAAC,KAAK,KAAK;;;;;;AAO3H,MAAM,kBACJ,YACA,QACA,WACA,WACuC;CACvC,MAAM,gBAAgB,IAAI,IAAI,UAAU,KAAK,MAAM,EAAE,KAAK,CAAC;AAC3D,QAAO,uBAAuB,YAAY,QAAQ,eAAe,OAAO;;;;;;AAO1E,MAAM,0BACJ,YACA,QACA,eACA,WACuC;CACvC,MAAMD,QAAkB,EAAE;CAG1B,MAAME,kBAA0C,EAAE;CAClD,MAAMC,kBAAqC,EAAE;AAE7C,MAAK,MAAM,OAAO,YAAY;AAC5B,MAAI,IAAI,SAAS,kBAAkB;AACjC,mBAAgB,KAAK,IAAI;SACpB;AACL,mBAAgB,KAAK,IAAI;;;AAK7B,MAAK,MAAM,OAAO,iBAAiB;EACjC,MAAM,SAAS,oBAAoB,KAAK,QAAQ,eAAe,OAAO;AACtE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,QAAM,KAAK,OAAO,MAAM;;AAI1B,KAAI,gBAAgB,SAAS,GAAG;EAC9B,MAAM,cAAc,2BAA2B,iBAAiB,QAAQ,eAAe,OAAO;AAC9F,MAAI,YAAY,OAAO,EAAE;AACvB,8BAAW,YAAY,MAAM;;AAE/B,QAAM,KAAK,YAAY,MAAM;;AAG/B,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,uBACJ,KACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAEnC,SAAQ,IAAI,MAAZ;EACE,KAAK,QACH,QAAO,mBAAmB,KAAK,QAAQ,eAAe,OAAO;EAC/D,KAAK,iBACH,2BAAU,GAAG,QAAQ,KAAK,IAAI,KAAK,oBAAoB;EACzD,KAAK,iBAEH,2BAAU,GAAG;;;;;;;AAQnB,MAAM,8BACJ,iBACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAGnC,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,KAAK,WAAW,IAAI;AACtB,8BAAW;IACT,MAAM;IACN,SAAS;IACV,CAAC;;;AAKN,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,UAAU,CAAC,OAAO,QAAQ,IAAI,KAAK,OAAO,EAAE;GAG9C,IAAI,gBAAgB;AACpB,QAAK,MAAM,GAAG,aAAa,OAAO,QAAQ;AACxC,QAAI,SAAS,QAAQ,IAAI,KAAK,OAAO,EAAE;AACrC,qBAAgB;AAChB;;;AAGJ,OAAI,CAAC,eAAe;AAClB,+BAAW;KACT,MAAM;KACN,SAAS,uCAAuC,KAAK,OAAO;KAC5D,QAAQ,KAAK;KACd,CAAC;;;;CAMR,MAAMC,UAAoB,EAAE;AAC5B,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,eAAe,KAAK,OAAO,SAAS,EAAE;EAC5C,MAAM,eAAe,uBAAuB,KAAK,YAAY,SAAS,GAAG,eAAe,OAAO;AAC/F,MAAI,aAAa,OAAO,EAAE;AACxB,8BAAW,aAAa,MAAM;;AAGhC,UAAQ,KAAK,GAAG,eAAe,KAAK,OAAO;EAC7C,aAAa,MAAM;EACnB,aAAa,KAAK;;AAKlB,2BAAU,GAAG,QAAQ;EACrB,QAAQ,KAAK,KAAK,CAAC;EACnB,QAAQ,KAAK;;;;;AAMf,MAAM,sBACJ,OACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;CAGnC,MAAM,OAAO,MAAM;CACnB,MAAM,aAAa,MAAM;CACzB,MAAM,UAAU,QAAQ,KAAK,SAAS;CACtC,MAAM,gBAAgB,cAAc,WAAW,SAAS;AAGxD,KAAI,CAAC,WAAW,CAAC,eAAe;AAC9B,4BAAU,GAAG,UAAU,MAAM,KAAK,SAAS;;CAG7C,IAAI,OAAO,GAAG,QAAQ,OAAO,MAAM,KAAK;AAExC,KAAI,SAAS;EACX,MAAM,aAAa,cAAc,MAAM,cAAc;AACrD,MAAI,WAAW,OAAO,EAAE;AACtB,8BAAW,WAAW,MAAM;;AAE9B,UAAQ,WAAW;;AAGrB,SAAQ;AAER,KAAI,eAAe;EAEjB,MAAM,qBAAqB,WAAW,MAAM,MAAM,EAAE,SAAS,iBAAiB;AAE9E,MAAI,oBAAoB;GAEtB,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ;AACR,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;SACd;AAEL,WAAQ;GACR,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;;;AAIvB,SAAQ;AAER,2BAAU,KAAK;;;;;AAMjB,MAAM,iBAAiB,MAAiC,kBAAmE;AACzH,KAAI,KAAK,WAAW,GAAG;AACrB,4BAAU,GAAG;;CAGf,MAAMC,aAAuB,EAAE;AAC/B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,SAAS,UAAU,IAAI,OAAO,cAAc;AAClD,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,GAAG,IAAI,KAAK,IAAI,OAAO,QAAQ;;AAEjD,2BAAU,KAAK,WAAW,KAAK,KAAK,CAAC,IAAI;;;;;AAM3C,MAAM,aAAa,OAAoB,kBAAmE;AACxG,SAAQ,MAAM,MAAd;EACE,KAAK;AAEH,OAAI,cAAc,IAAI,MAAM,KAAK,EAAE;AACjC,8BAAU,KAAK,MAAM,OAAO;;AAE9B,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,MAAM,KAAK;IAClC,cAAc,MAAM;IACrB,CAAC;EACJ,KAAK;EACL,KAAK,QACH,2BAAU,MAAM,MAAM;EACxB,KAAK,SACH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,UACH,2BAAU,MAAM,QAAQ,SAAS,QAAQ;EAC3C,KAAK,OACH,2BAAU,OAAO;EACnB,KAAK,OAEH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,QAAQ;GACX,MAAMC,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,GAAG,cAAc;AAC1C,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,WAAO,KAAK,OAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;EAErC,KAAK,UAAU;AACb,OAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,8BAAU,KAAK;;GAEjB,MAAMF,UAAoB,EAAE;AAC5B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,EAAE,OAAO,cAAc;AAChD,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,YAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;AAE5C,6BAAU,KAAK,QAAQ,KAAK,KAAK,CAAC,IAAI;;;;;;;;;;;;;;;ACrW5C,MAAa,oBAAoB,aAA8D;CAC7F,MAAM,sCAAuB,SAAS;AAEtC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAW;GACT,MAAM;GACN,SAAS,6BAA6B;GACtC,UAAU;GACX,CAAC;;AAGJ,KAAI;EACF,MAAM,mCAAsB,cAAc,OAAO;EACjD,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,aAAa,CAAC;UAC/C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAa,sBAAsB,QAAgB,eAAgE;AACjH,KAAI;EACF,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,WAAW,CAAC;UAC7C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAM,uBAAuB,UAAwB,eAAoC;CACvF,MAAMG,aAAgC,EAAE;CACxC,MAAMC,YAA8B,EAAE;AAEtC,MAAK,MAAM,cAAc,SAAS,aAAa;AAC7C,MAAI,WAAW,SAASC,aAAK,sBAAsB;GACjD,MAAM,YAAY,iBAAiB,YAAY,WAAW;AAC1D,OAAI,WAAW;AACb,eAAW,KAAK,UAAU;;aAEnB,WAAW,SAASA,aAAK,qBAAqB;AACvD,aAAU,KAAK,gBAAgB,YAAY,WAAW,CAAC;;;AAI3D,QAAO;EAAE;EAAY;EAAW;;;;;AAMlC,MAAM,oBAAoB,MAA+B,eAA+C;AAEtG,KAAI,CAAC,KAAK,MAAM;AACd,SAAO;;CAGT,MAAMC,aAA+B,KAAK,uBAAuB,EAAE,EAAE,IAAI,gBAAgB;CACzF,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK;EACX,MAAM,KAAK,KAAK;EAChB;EACA;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,MAA8B,eAAuC;CAC5F,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,QAAQ,KAAK,cAAc,KAAK;EAChC;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,SAAiD;CACxE,MAAM,EAAE,UAAU,aAAa,cAAc,KAAK,KAAK;CACvD,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,aAAa,GAAG;AAE3E,QAAO;EACL,MAAM,KAAK,SAAS,KAAK;EACzB;EACA;EAEA,UAAU;EACV;EACD;;;;;;;;;AAUH,MAAa,iBAAiB,SAA2D;CAEvF,MAAMC,SAAsB,EAAE;CAE9B,MAAM,WAAW,GAAa,YAA6B;AACzD,MAAI,EAAE,SAASF,aAAK,eAAe;AACjC,UAAO,QAAQ,EAAE,MAAM,KAAK;;AAE9B,MAAI,EAAE,SAASA,aAAK,WAAW;AAC7B,UAAO,KAAK;IAAE,MAAM;IAAQ;IAAS,CAAC;AACtC,UAAO,QAAQ,EAAE,MAAM,MAAM;;AAE/B,SAAO,KAAK;GAAE,MAAM;GAAS;GAAS,CAAC;AACvC,SAAO,EAAE,KAAK;;CAGhB,MAAM,WAAW,QAAQ,MAAM,MAAM;CAGrC,IAAI,WAAW;AACf,MAAK,MAAM,SAAS,OAAO,OAAO,CAAC,SAAS,EAAE;AAC5C,MAAI,MAAM,SAAS,SAAS;AAC1B,cAAW,MAAM,UAAU,MAAM;AACjC;;EAEF,MAAM,aAAa,MAAM,UAAU,QAAQ;AAC3C,aAAW,GAAG,WAAW;;AAG3B,QAAO;EAAE;EAAU;EAAU;;;;;AAM/B,MAAM,qBAAqB,eAA4D;AACrF,QAAO,WAAW,IAAI,iBAAiB;;;;;AAMzC,MAAM,oBAAoB,SAAyC;AACjE,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,MACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;;;;;;AAOxC,MAAM,yBAAyB,SAA0C;CACvE,MAAM,OAAO,KAAK,WAAW,SAAS,KAAK,UAAU,IAAI,gBAAgB,GAAG;CAC5E,MAAM,aAAa,KAAK,eAAe,kBAAkB,KAAK,aAAa,WAAW,GAAG;AAEzF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EAChB,OAAO,KAAK,OAAO;EACnB,WAAW;EACX;EACD;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EACjB;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,QAAQ,KAAK,eAAe,KAAK,SAAS;EAC1C,YAAY,kBAAkB,KAAK,aAAa,WAAW;EAC5D;;;;;AAMH,MAAM,mBAAmB,SAAuC;AAC9D,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,OAAO,aAAa,KAAK,MAAM;EAChC;;;;;AAMH,MAAM,qBAAqB,UAAwB;AACjD,OAAM,IAAI,MAAM,qBAAqB,KAAK,UAAU,MAAM,GAAG;;;;;AAM/D,MAAM,gBAAgB,SAAiC;AACrD,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,SACR,QAAO;GAAE,MAAM;GAAY,MAAM,KAAK,KAAK;GAAO;EACpD,KAAKA,aAAK,IACR,QAAO;GAAE,MAAM;GAAO,OAAO,KAAK;GAAO;EAC3C,KAAKA,aAAK,MACR,QAAO;GAAE,MAAM;GAAS,OAAO,KAAK;GAAO;EAC7C,KAAKA,aAAK,OACR,QAAO;GAAE,MAAM;GAAU,OAAO,KAAK;GAAO;EAC9C,KAAKA,aAAK,QACR,QAAO;GAAE,MAAM;GAAW,OAAO,KAAK;GAAO;EAC/C,KAAKA,aAAK,KACR,QAAO,EAAE,MAAM,QAAQ;EACzB,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,OAAO,KAAK;GAAO;EAC5C,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,QAAQ,KAAK,OAAO,IAAI,aAAa;GAAE;EAChE,KAAKA,aAAK,OACR,QAAO;GACL,MAAM;GACN,QAAQ,KAAK,OAAO,KAAK,WAAW;IAClC,MAAM,MAAM,KAAK;IACjB,OAAO,aAAa,MAAM,MAAM;IACjC,EAAE;GACJ;EACH,QACE,QAAO,kBAAkB,KAAK;;;;;;;;;AChRpC,MAAM,qBAAqB,IAAI,IAAI;CAAC;CAAM;CAAU;CAAO;CAAS;CAAU,CAAC;;;;;;AAoB/E,MAAM,0BAA0B,aAAwC;CAEtE,MAAM,QAAQ,SAAS,OAAO,MAAM,MAAM;CAG1C,MAAMG,QAA2B,EAAE;CACnC,MAAM,cAAc;CACpB,IAAIC;AACJ,SAAQ,QAAQ,YAAY,KAAK,SAAS,MAAM,MAAM;AACpD,QAAM,KAAK,KAAK,MAAM,KAAsB;;AAG9C,QAAO;EAAE;EAAO;EAAO;;;;;AAMzB,MAAM,iBAAiB,cAAyC;AAC9D,QAAO,UAAU,QAAQ,UAAU,MAAM,KAAK,GAAG;;;;;;;;;;;AAYnD,MAAa,kBAAkB,GAAW,MAA2E;CACnH,MAAM,UAAU,uBAAuB,EAAE;CACzC,MAAM,UAAU,uBAAuB,EAAE;AAGzC,KAAI,QAAQ,MAAM,WAAW,QAAQ,MAAM,QAAQ;AACjD,SAAO;GACL,IAAI;GACJ,QAAQ,8BAA8B,EAAE,QAAQ,QAAQ,MAAM,OAAO,mBAAmB,EAAE,QAAQ,QAAQ,MAAM;GACjH;;CAIH,MAAMC,cAAyB,QAAQ,UAAU,OAAO,QAAQ,UAAU,MAAM,MAAM;CAGtF,MAAMC,cAAiC,EAAE;AACzC,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,MAAM,QAAQ,KAAK;EAC7C,MAAM,QAAQ,QAAQ,MAAM;EAC5B,MAAM,QAAQ,QAAQ,MAAM;AAC5B,cAAY,KAAK,UAAU,SAAS,UAAU,QAAQ,QAAQ,MAAM;;AAGtE,QAAO;EAAE,IAAI;EAAM,OAAO,cAAc;GAAE,OAAO;GAAa,OAAO;GAAa,CAAC;EAAE;;;;;;AAqBvF,MAAM,mBACJ,QACA,gBACA,WACA,iBACkD;CAClD,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,SAAS,SAAS,WAAW,MAAM,QAAQ,IAAI,KAAK,UAAU,aAAa;AACjF,KAAI,CAAC,OAAQ,QAAO;AAEpB,QAAO,cAAc,OAAO,KAAK;;;;;AAMnC,MAAM,qBACJ,QACA,eACA,cACkD;CAClD,MAAM,cAAc,OAAO,OAAO,IAAI,cAAc;AACpD,KAAI,CAAC,YAAa,QAAO;CAEzB,MAAM,WAAW,YAAY,OAAO,IAAI,UAAU;AAClD,KAAI,CAAC,SAAU,QAAO;AAEtB,QAAO,cAAc,SAAS,KAAK;;;;;AAMrC,MAAM,2BAA2B,QAAqB,aAAyD;AAC7G,KAAI,aAAa,QAAQ,SAAS,CAAE,QAAO;AAC3C,KAAI,WAAW,QAAQ,SAAS,CAAE,QAAO;AACzC,KAAI,OAAO,OAAO,IAAI,SAAS,CAAE,QAAO;AACxC,QAAO;;;;;;AAOT,MAAM,6BACJ,OACA,kBACA,kBACA,QACA,WAC8B;AAC9B,KAAI,MAAM,SAAS,YAAY;EAC7B,MAAM,WAAW,wBAAwB,QAAQ,iBAAiB;AAClE,MAAI,CAAC,UAAU;AACb,UAAO;IACL,MAAM;IACN,SAAS,iBAAiB,iBAAiB,mBAAmB,MAAM,KAAK;IACzE,UAAU;IACX;;AAEH,SAAO,KAAK;GACV,MAAM,MAAM;GACZ,UAAU;GACV,UAAU;GACV;GACD,CAAC;AACF,SAAO;;AAGT,KAAI,MAAM,SAAS,UAAU;AAE3B,OAAK,MAAM,SAAS,MAAM,QAAQ;GAChC,MAAM,YAAY,kBAAkB,QAAQ,kBAAkB,MAAM,KAAK;AACzE,OAAI,CAAC,WAAW;AACd,WAAO;KACL,MAAM;KACN,SAAS,kBAAkB,MAAM,KAAK,mBAAmB,iBAAiB;KAC1E,UAAU;KACV,WAAW,MAAM;KAClB;;GAEH,MAAM,QAAQ,0BAA0B,MAAM,OAAO,UAAU,UAAU,UAAU,UAAU,QAAQ,OAAO;AAC5G,OAAI,MAAO,QAAO;;AAEpB,SAAO;;AAGT,KAAI,MAAM,SAAS,QAAQ;EAGzB,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,MAAI,OAAO,MAAM,SAAS,GAAG;GAC3B,MAAM,gBAAgB,cAAc;IAClC,OAAO,OAAO;IACd,OAAO,OAAO,MAAM,MAAM,EAAE;IAC7B,CAAC;AACF,QAAK,MAAM,QAAQ,MAAM,QAAQ;IAC/B,MAAM,QAAQ,0BAA0B,MAAM,kBAAkB,eAAe,QAAQ,OAAO;AAC9F,QAAI,MAAO,QAAO;;;;AAMxB,QAAO;;;;;AAMT,MAAM,iCACJ,MACA,gBACA,WACA,QACA,WAC8B;AAC9B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,UAAU,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK;AAC5E,MAAI,CAAC,SAAS;AACZ,UAAO;IACL,MAAM;IACN,SAAS,qBAAqB,IAAI,KAAK,cAAc,UAAU;IAC/D;IACA,cAAc,IAAI;IACnB;;EAEH,MAAM,QAAQ,0BAA0B,IAAI,OAAO,QAAQ,UAAU,QAAQ,UAAU,QAAQ,OAAO;AACtG,MAAI,MAAO,QAAO;;AAEpB,QAAO;;;;;AAMT,MAAa,yBACX,YACA,gBACA,WACgD;CAChD,MAAMC,SAA0B,EAAE;CAElC,MAAM,WAAW,MAAkC,eAAkD;AACnG,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK,SAAS;AAEZ,SAAI,IAAI,aAAa,IAAI,UAAU,SAAS,GAAG;MAC7C,MAAMC,UAAQ,8BAA8B,IAAI,WAAW,YAAY,IAAI,MAAM,QAAQ,OAAO;AAChG,UAAIA,QAAO,QAAOA;;AAIpB,SAAI,IAAI,cAAc,IAAI,WAAW,SAAS,GAAG;MAE/C,MAAM,kBAAkB,mBAAmB,QAAQ,YAAY,IAAI,KAAK;AACxE,UAAI,CAAC,iBAAiB;AACpB,cAAO;QACL,MAAM;QACN,SAAS,kBAAkB,IAAI,KAAK,aAAa,WAAW;QAC5D,UAAU;QACV,WAAW,IAAI;QAChB;;MAEH,MAAMA,UAAQ,QAAQ,IAAI,YAAY,gBAAgB;AACtD,UAAIA,QAAO,QAAOA;;AAEpB;;IAEF,KAAK,kBAAkB;KAErB,MAAMA,UAAQ,QAAQ,IAAI,YAAY,IAAI,OAAO;AACjD,SAAIA,QAAO,QAAOA;AAClB;;IAEF,KAAK,iBAEH;;;AAGN,SAAO;;CAGT,MAAM,QAAQ,QAAQ,YAAY,eAAe;AACjD,KAAI,MAAO,4BAAW,MAAM;AAE5B,2BAAU,OAAO;;;;;AAMnB,MAAM,sBAAsB,QAAqB,gBAAwB,cAAqC;CAC5G,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,EAAE,aAAa,cAAc,SAAS,KAAK;AACjD,QAAO;;;;;;AAOT,MAAa,uBACX,cACA,WACiD;AACjD,KAAI,OAAO,WAAW,GAAG;AAEvB,6BAAW;GACT,MAAM;GACN,SAAS,iCAAiC,aAAa;GACvD;GACD,CAAC;;CAGJ,MAAM,QAAQ,OAAO;AAGrB,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,MAAM,aAAa,MAAM,UAAU;AACrC,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,4BAA4B,MAAM,SAAS,SAAS,MAAM,SAAS;IACvG;IACD,CAAC;;;CAKN,IAAI,iBAAiB,MAAM;AAC3B,MAAK,IAAI,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;EACtC,MAAM,SAAS,eAAe,gBAAgB,OAAO,GAAI,SAAS;AAClE,MAAI,CAAC,OAAO,IAAI;AACd,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,gCAAgC,OAAO;IAC3E;IACD,CAAC;;AAEJ,mBAAiB,OAAO;;AAG1B,2BAAU;EACR,MAAM;EACN,UAAU,MAAM;EAChB,UAAU;EACV,UAAU,MAAM;EACjB,CAAC;;;;;;AAOJ,MAAa,4BAA4B,WAAqF;CAE5H,MAAM,SAAS,IAAI,KAA8B;AACjD,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,WAAW,OAAO,IAAI,MAAM,KAAK;AACvC,MAAI,UAAU;AACZ,YAAS,KAAK,MAAM;SACf;AACL,UAAO,IAAI,MAAM,MAAM,CAAC,MAAM,CAAC;;;CAKnC,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,CAAC,MAAM,UAAU,QAAQ;EAClC,MAAM,SAAS,oBAAoB,MAAM,MAAM;AAC/C,MAAI,OAAO,OAAO,CAAE,4BAAW,OAAO,MAAM;AAC5C,YAAU,KAAK,OAAO,MAAM;;AAI9B,WAAU,MAAM,GAAG,MAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC;AAEtD,2BAAU,UAAU;;;;;AAMtB,MAAM,gBAAgB,QAAqB,SAA0B,mBAAmB,IAAI,KAAK,IAAI,OAAO,QAAQ,IAAI,KAAK;;;;;;;AAY7H,MAAa,6BAA6B,cAAuF;CAE/H,MAAM,QAAQ,IAAI,KAA0B;AAC5C,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,OAAO,+BAA+B,SAAS,WAAW;AAChE,QAAM,IAAI,SAAS,MAAM,KAAK;;CAGhC,MAAM,iBAAiB,IAAI,KAA6B;AACxD,MAAK,MAAM,KAAK,WAAW;AACzB,iBAAe,IAAI,EAAE,MAAM,EAAE;;CAG/B,MAAMC,SAA2B,EAAE;CACnC,MAAM,UAAU,IAAI,KAAa;CACjC,MAAM,WAAW,IAAI,KAAa;CAElC,MAAM,SAAS,MAAc,SAA8C;AACzE,MAAI,QAAQ,IAAI,KAAK,CAAE,QAAO;AAE9B,MAAI,SAAS,IAAI,KAAK,EAAE;GAEtB,MAAM,aAAa,KAAK,QAAQ,KAAK;GACrC,MAAM,QAAQ,KAAK,MAAM,WAAW,CAAC,OAAO,KAAK;AACjD,UAAO;IACL,MAAM;IACN,SAAS,8CAA8C,MAAM,KAAK,OAAO;IACzE,eAAe;IAChB;;EAIH,MAAM,WAAW,eAAe,IAAI,KAAK;AACzC,MAAI,CAAC,UAAU;AAEb,WAAQ,IAAI,KAAK;AACjB,UAAO;;AAGT,WAAS,IAAI,KAAK;EAClB,MAAM,OAAO,MAAM,IAAI,KAAK,IAAI,IAAI,KAAK;AAEzC,OAAK,MAAM,OAAO,MAAM;GACtB,MAAM,QAAQ,MAAM,KAAK,CAAC,GAAG,MAAM,KAAK,CAAC;AACzC,OAAI,MAAO,QAAO;;AAGpB,WAAS,OAAO,KAAK;AACrB,UAAQ,IAAI,KAAK;AACjB,SAAO,KAAK,SAAS;AACrB,SAAO;;AAGT,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,QAAQ,MAAM,SAAS,MAAM,EAAE,CAAC;AACtC,MAAI,MAAO,4BAAW,MAAM;;AAG9B,2BAAU,OAAO;;;;;;AAOnB,MAAM,kCAAkC,eAAwD;CAC9F,MAAM,OAAO,IAAI,KAAa;CAE9B,MAAM,WAAW,SAA2C;AAC1D,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,UAAK,IAAI,IAAI,KAAK;AAClB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO;;;;;AAMT,MAAM,cAAc,QAAqB,SAA0B,OAAO,MAAM,IAAI,KAAK;;;;;;;AAkDzF,MAAa,0BACX,QACA,YACgD;CAChD,MAAM,SAASC,oCAAkB,QAAQ,eAAe;CAGxD,MAAM,aAAa,0BAA0B,OAAO,UAAU;AAC9D,KAAI,WAAW,OAAO,EAAE;AACtB,6BAAW,WAAW,MAAM;;CAE9B,MAAM,kBAAkB,WAAW;CAGnC,MAAM,4BAA4B,IAAI,KAA0C;CAChF,MAAMC,YAAgC,EAAE;AAExC,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,SAAS,kBAAkB,MAAM,QAAQ,0BAA0B;AACzE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,4BAA0B,IAAI,KAAK,MAAM,OAAO,MAAM,UAAU;AAChE,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAMC,aAAkC,EAAE;AAC1C,MAAK,MAAM,MAAM,OAAO,YAAY;EAClC,MAAM,SAAS,mBAAmB,IAAI,OAAO;AAC7C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,OAAO,MAAM;;AAG/B,2BAAU;EAAE;EAAY;EAAW,CAAC;;;;;AAMtC,MAAM,sBAAsB,IAAqB,WAAuE;CAEtH,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,GAAG,WAAW;EAC5B,MAAM,WAAW,gBAAgB,QAAQ,EAAE,SAAS;AACpD,MAAI,aAAa,MAAM;AACrB,8BAAW;IACT,MAAM;IACN,SAAS,iBAAiB,EAAE,SAAS,iBAAiB,EAAE,KAAK;IAC7D,UAAU,EAAE;IACb,CAAC;;AAEJ,YAAU,KAAK;GAAE,GAAG;GAAG;GAAU,CAAC;;CAIpC,MAAM,uBAAuB,4BAA4B,GAAG,WAAW;AAEvE,2BAAU;EACR,GAAG;EACH;EACA;EACD,CAAC;;;;;;AAOJ,MAAM,qBACJ,MACA,QACA,8BACiD;CAEjD,MAAM,uBAAuB,4BAA4B,KAAK,WAAW;CAGzE,MAAM,qBAAqB,sBAAsB,KAAK,YAAY,KAAK,QAAQ,OAAO;AACtF,KAAI,mBAAmB,OAAO,EAAE;AAC9B,6BAAW,mBAAmB,MAAM;;CAEtC,MAAM,eAAe,mBAAmB;CAGxC,MAAMC,kBAAsC,EAAE;AAC9C,MAAK,MAAM,WAAW,sBAAsB;EAC1C,MAAM,eAAe,0BAA0B,IAAI,QAAQ;AAC3D,MAAI,cAAc;AAChB,mBAAgB,KAAK,GAAG,aAAa;;;CAOzC,MAAM,YAAY,CAChB,GAAG,cACH,GAAG,gBAAgB,KAAK,OAAO;EAC7B,MAAM,EAAE;EACR,UAAU,EAAE;EACZ,UAAU,EAAE;EACZ,UAAU,EAAE;EACb,EAAE,CACJ;CAGD,MAAM,kBAAkB,yBAAyB,UAAU;AAC3D,KAAI,gBAAgB,OAAO,EAAE;AAC3B,6BAAW,gBAAgB,MAAM;;AAGnC,2BAAU;EACR,GAAG;EACH;EACA,WAAW,gBAAgB;EAC5B,CAAC;;;;;AAMJ,MAAM,mBAAmB,QAAqB,aAAyD;AACrG,KAAI,aAAa,QAAQ,SAAS,EAAE;AAClC,SAAO;;AAET,KAAI,WAAW,QAAQ,SAAS,EAAE;AAChC,SAAO;;AAET,KAAI,OAAO,OAAO,IAAI,SAAS,EAAE;AAC/B,SAAO;;AAET,QAAO;;;;;AAMT,MAAM,+BAA+B,eAAgF;CACnH,MAAM,YAAY,IAAI,KAAa;CAEnC,MAAM,WAAW,SAA6D;AAC5E,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,eAAU,IAAI,IAAI,KAAK;AACvB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO,CAAC,GAAG,UAAU;;;;;ACztBvB,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,8BAAe,WAAW,EAAE;AAC1B,8BAA+B;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaC,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,mCAAoB,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,4BAAY;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,6BAAU,EAAE,SAAS,CAAC;WACf,OAAO;AACd,8BAAW;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,mBAAmB,YAAoB;CAClD,MAAM,oCAAqB,QAAQ;AACnC,KAAI;AACF,sBAAO,YAAY;GAAE,WAAW;GAAM,OAAO;GAAM,CAAC;AACpD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,SAAS;AACnC,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACtBN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,sCAAuB,WAAW;AAExC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAuC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,yCAA4B,cAAc,OAAO;EACvD,MAAM,8BAAiB,aAAa;AACpC,4BAAsC,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAuC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAuC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,gCAAmB,UAAU;AACnC,2BAAsC,OAAO;;AAG/C,MAAa,cAAc,yCAA8C,SAAS,CAAC,0BAAa,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC9CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,iCAAkB,SAAS;CACjC,MAAM,qCAAsB,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,mCAAoB,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,6BAAc,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,sCAAqBD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,oCAAqB,YAAY,UAAU,2BAAY,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,oCAAqB,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,iCAAkB,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,oCAAqB,aAAa,OAAO,QAAQ;AACvD,MAAI,yBAAY,WAAW,EAAE;AAC3B,8BAAW;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,qCAAsB,aAAa,OAAO,QAAQ;AACxD,OAAI,yBAAY,YAAY,EAAE;AAC5B,+BAAW;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,2BAAW,IAAI,CAAC,GAChC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,oBAAoB,IAAI,KAA+B;AAC7D,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,YAAY;AAC3B,qBAAkB,IAAI,YAAY,aAAa,WAAW;;;CAK9D,MAAM,EACJ,MAAM,cACN,aACA,iBACEC,4CAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACA,aAAa,kBAAkB,OAAO,IAAI,oBAAoB;EAC/D,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,2CAAM,8BAAuB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,yDAA2B,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,8BAAW,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,gCAAiB,QAAQ;EAG/B,MAAM,8BAAe,QAAQ,QAAQ;AACrC,8BAAe,QAAQ,EAAE;GACvB,MAAM,eAAe,gBAAgB,QAAQ;AAC7C,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;;EAOlC,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,+BAAgB,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,+BAAW,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,0DAA4B,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,6BAAW,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,8BAAc,OAAO,GACrB,8BAAc,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,6BAAW,aAAa,MAAM;;AAGhC,2BAAU;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
package/dist/index.d.cts CHANGED
@@ -491,8 +491,6 @@ type EmitOptions = {
491
491
  readonly schemaName: string;
492
492
  /** Import path for graphql-system module */
493
493
  readonly graphqlSystemPath: string;
494
- /** Map of fragment name to its import path (relative) */
495
- readonly fragmentImports?: ReadonlyMap<string, string>;
496
494
  /** Schema document for type lookups (required for inline fragment support) */
497
495
  readonly schemaDocument?: DocumentNode;
498
496
  };
@@ -581,6 +579,10 @@ type CodegenError = {
581
579
  readonly code: "INJECT_TEMPLATE_FAILED";
582
580
  readonly message: string;
583
581
  readonly outPath: string;
582
+ } | {
583
+ readonly code: "REMOVE_FAILED";
584
+ readonly message: string;
585
+ readonly outPath: string;
584
586
  };
585
587
  type CodegenSuccess = {
586
588
  readonly schemas: Record<string, {