@soda-gql/codegen 0.11.21 → 0.11.22

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -846,8 +846,16 @@ const collectFragmentDependencies = (selections) => {
846
846
  //#region packages/codegen/src/graphql-compat/emitter.ts
847
847
  /**
848
848
  * Map operation kind to root type name.
849
- */
850
- const getRootTypeName = (kind) => {
849
+ * Uses schema.operationTypes if available, falls back to standard names.
850
+ */
851
+ const getRootTypeName = (schema, kind) => {
852
+ if (schema) {
853
+ switch (kind) {
854
+ case "query": return schema.operationTypes.query ?? "Query";
855
+ case "mutation": return schema.operationTypes.mutation ?? "Mutation";
856
+ case "subscription": return schema.operationTypes.subscription ?? "Subscription";
857
+ }
858
+ }
851
859
  switch (kind) {
852
860
  case "query": return "Query";
853
861
  case "mutation": return "Mutation";
@@ -868,7 +876,7 @@ const emitOperation = (operation, options) => {
868
876
  if (operation.variables.length > 0) {
869
877
  lines.push(` variables: { ${emitVariables(operation.variables)} },`);
870
878
  }
871
- const rootTypeName = getRootTypeName(operation.kind);
879
+ const rootTypeName = getRootTypeName(schema, operation.kind);
872
880
  lines.push(` fields: ({ f, $ }) => ({`);
873
881
  const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema, rootTypeName);
874
882
  if (fieldLinesResult.isErr()) {
@@ -1 +1 @@
1
- {"version":3,"file":"index.cjs","names":["operations: ParsedOperation[]","fragments: ParsedFragment[]","Kind","variables: ParsedVariable[]","levels: TypeLevel[]","lists: (\"[]!\" | \"[]?\")[]","match: RegExpExecArray | null","mergedInner: \"!\" | \"?\"","mergedLists: (\"[]!\" | \"[]?\")[]","usages: VariableUsage[]","error","variables: InferredVariable[]","sorted: ParsedFragment[]","createSchemaIndex","fragments: EnrichedFragment[]","operations: EnrichedOperation[]","variables: EnrichedVariable[]","spreadVariables: InferredVariable[]","allUsages: VariableUsage[]","lines: string[]","createSchemaIndex","inlineFragments: ParsedInlineFragment[]","otherSelections: ParsedSelection[]","entries: string[]","values: string[]","result","argEntries: string[]","esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","generateMultiSchemaModule","defsPaths: string[]","defaultBundler"],"sources":["../src/graphql-compat/parser.ts","../src/graphql-compat/transformer.ts","../src/graphql-compat/emitter.ts","../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["/**\n * Parser for .graphql operation files.\n * Extracts operations and fragments from GraphQL documents.\n * @module\n */\n\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport {\n type ArgumentNode,\n type DocumentNode,\n type FieldNode,\n type FragmentDefinitionNode,\n type FragmentSpreadNode,\n type InlineFragmentNode,\n Kind,\n type OperationDefinitionNode,\n parse,\n type SelectionNode,\n type TypeNode,\n type ValueNode,\n type VariableDefinitionNode,\n} from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\n\nimport type {\n GraphqlCompatError,\n ParsedArgument,\n ParsedFieldSelection,\n ParsedFragment,\n ParsedFragmentSpread,\n ParsedInlineFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n TypeInfo,\n} from \"./types\";\n\n/**\n * Parse a single .graphql file and extract operations and fragments.\n */\nexport const parseGraphqlFile = (filePath: string): Result<ParseResult, GraphqlCompatError> => {\n const resolvedPath = resolve(filePath);\n\n if (!existsSync(resolvedPath)) {\n return err({\n code: \"GRAPHQL_FILE_NOT_FOUND\",\n message: `GraphQL file not found at ${resolvedPath}`,\n filePath: resolvedPath,\n });\n }\n\n try {\n const source = readFileSync(resolvedPath, \"utf8\");\n const document = parse(source);\n return ok(extractFromDocument(document, resolvedPath));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: resolvedPath,\n });\n }\n};\n\n/**\n * Parse GraphQL source string directly.\n */\nexport const parseGraphqlSource = (source: string, sourceFile: string): Result<ParseResult, GraphqlCompatError> => {\n try {\n const document = parse(source);\n return ok(extractFromDocument(document, sourceFile));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: sourceFile,\n });\n }\n};\n\n/**\n * Extract operations and fragments from a parsed GraphQL document.\n */\nconst extractFromDocument = (document: DocumentNode, sourceFile: string): ParseResult => {\n const operations: ParsedOperation[] = [];\n const fragments: ParsedFragment[] = [];\n\n for (const definition of document.definitions) {\n if (definition.kind === Kind.OPERATION_DEFINITION) {\n const operation = extractOperation(definition, sourceFile);\n if (operation) {\n operations.push(operation);\n }\n } else if (definition.kind === Kind.FRAGMENT_DEFINITION) {\n fragments.push(extractFragment(definition, sourceFile));\n }\n }\n\n return { operations, fragments };\n};\n\n/**\n * Extract a single operation from an OperationDefinitionNode.\n */\nconst extractOperation = (node: OperationDefinitionNode, sourceFile: string): ParsedOperation | null => {\n // Anonymous operations are not supported\n if (!node.name) {\n return null;\n }\n\n const variables: ParsedVariable[] = (node.variableDefinitions ?? []).map(extractVariable);\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n kind: node.operation,\n name: node.name.value,\n variables,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a fragment from a FragmentDefinitionNode.\n */\nconst extractFragment = (node: FragmentDefinitionNode, sourceFile: string): ParsedFragment => {\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n name: node.name.value,\n onType: node.typeCondition.name.value,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a variable definition.\n */\nconst extractVariable = (node: VariableDefinitionNode): ParsedVariable => {\n const { typeName, modifier } = parseTypeNode(node.type);\n const defaultValue = node.defaultValue ? extractValue(node.defaultValue) : undefined;\n\n return {\n name: node.variable.name.value,\n typeName,\n modifier,\n // Type kind will be determined later when we have schema context\n typeKind: \"scalar\",\n defaultValue,\n };\n};\n\n/**\n * Parse a GraphQL TypeNode into type name and modifier.\n *\n * Format: inner nullability + list modifiers\n * - Inner: `!` (non-null) or `?` (nullable)\n * - List: `[]!` (non-null list) or `[]?` (nullable list)\n */\nexport const parseTypeNode = (node: TypeNode): TypeInfo => {\n type TypeLevel = { kind: \"list\" | \"named\"; nonNull: boolean };\n const levels: TypeLevel[] = [];\n\n const collect = (n: TypeNode, nonNull: boolean): string => {\n if (n.kind === Kind.NON_NULL_TYPE) {\n return collect(n.type, true);\n }\n if (n.kind === Kind.LIST_TYPE) {\n levels.push({ kind: \"list\", nonNull });\n return collect(n.type, false);\n }\n levels.push({ kind: \"named\", nonNull });\n return n.name.value;\n };\n\n const typeName = collect(node, false);\n\n // Build modifier from levels (reverse order)\n let modifier = \"?\";\n for (const level of levels.slice().reverse()) {\n if (level.kind === \"named\") {\n modifier = level.nonNull ? \"!\" : \"?\";\n continue;\n }\n const listSuffix = level.nonNull ? \"[]!\" : \"[]?\";\n modifier = `${modifier}${listSuffix}`;\n }\n\n return { typeName, modifier };\n};\n\n/**\n * Extract selections from a SelectionSet.\n */\nconst extractSelections = (selections: readonly SelectionNode[]): ParsedSelection[] => {\n return selections.map(extractSelection);\n};\n\n/**\n * Extract a single selection.\n */\nconst extractSelection = (node: SelectionNode): ParsedSelection => {\n switch (node.kind) {\n case Kind.FIELD:\n return extractFieldSelection(node);\n case Kind.FRAGMENT_SPREAD:\n return extractFragmentSpread(node);\n case Kind.INLINE_FRAGMENT:\n return extractInlineFragment(node);\n }\n};\n\n/**\n * Extract a field selection.\n */\nconst extractFieldSelection = (node: FieldNode): ParsedFieldSelection => {\n const args = node.arguments?.length ? node.arguments.map(extractArgument) : undefined;\n const selections = node.selectionSet ? extractSelections(node.selectionSet.selections) : undefined;\n\n return {\n kind: \"field\",\n name: node.name.value,\n alias: node.alias?.value,\n arguments: args,\n selections,\n };\n};\n\n/**\n * Extract a fragment spread.\n */\nconst extractFragmentSpread = (node: FragmentSpreadNode): ParsedFragmentSpread => {\n return {\n kind: \"fragmentSpread\",\n name: node.name.value,\n };\n};\n\n/**\n * Extract an inline fragment.\n */\nconst extractInlineFragment = (node: InlineFragmentNode): ParsedInlineFragment => {\n return {\n kind: \"inlineFragment\",\n onType: node.typeCondition?.name.value ?? \"\",\n selections: extractSelections(node.selectionSet.selections),\n };\n};\n\n/**\n * Extract an argument.\n */\nconst extractArgument = (node: ArgumentNode): ParsedArgument => {\n return {\n name: node.name.value,\n value: extractValue(node.value),\n };\n};\n\n/**\n * Assert unreachable code path (for exhaustiveness checks).\n */\nconst assertUnreachable = (value: never): never => {\n throw new Error(`Unexpected value: ${JSON.stringify(value)}`);\n};\n\n/**\n * Extract a value (literal or variable reference).\n */\nconst extractValue = (node: ValueNode): ParsedValue => {\n switch (node.kind) {\n case Kind.VARIABLE:\n return { kind: \"variable\", name: node.name.value };\n case Kind.INT:\n return { kind: \"int\", value: node.value };\n case Kind.FLOAT:\n return { kind: \"float\", value: node.value };\n case Kind.STRING:\n return { kind: \"string\", value: node.value };\n case Kind.BOOLEAN:\n return { kind: \"boolean\", value: node.value };\n case Kind.NULL:\n return { kind: \"null\" };\n case Kind.ENUM:\n return { kind: \"enum\", value: node.value };\n case Kind.LIST:\n return { kind: \"list\", values: node.values.map(extractValue) };\n case Kind.OBJECT:\n return {\n kind: \"object\",\n fields: node.fields.map((field) => ({\n name: field.name.value,\n value: extractValue(field.value),\n })),\n };\n default:\n return assertUnreachable(node);\n }\n};\n","/**\n * Transformer for enriching parsed GraphQL operations with schema information.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport { parseTypeNode } from \"./parser\";\nimport type {\n GraphqlCompatError,\n InferredVariable,\n ParsedArgument,\n ParsedFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n TypeInfo,\n} from \"./types\";\n\n/**\n * Schema index type extracted from generator.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Built-in GraphQL scalar types.\n */\nconst builtinScalarTypes = new Set([\"ID\", \"String\", \"Int\", \"Float\", \"Boolean\"]);\n\n// ============================================================================\n// Modifier Merging Utilities\n// ============================================================================\n\n/**\n * Parsed structure of a modifier for comparison and merging.\n * Example: \"![]?\" -> { inner: \"!\", lists: [\"[]?\"] }\n */\ntype ModifierStructure = {\n readonly inner: \"!\" | \"?\";\n readonly lists: readonly (\"[]!\" | \"[]?\")[];\n};\n\n/**\n * Parse a modifier string into its structural components.\n * @param modifier - Modifier string like \"!\", \"?\", \"![]!\", \"?[]?[]!\"\n * @returns Parsed structure with inner nullability and list modifiers\n */\nconst parseModifierStructure = (modifier: string): ModifierStructure => {\n // Extract inner nullability (first character)\n const inner = modifier[0] === \"!\" ? \"!\" : \"?\";\n\n // Extract list modifiers ([]! or []?)\n const lists: (\"[]!\" | \"[]?\")[] = [];\n const listPattern = /\\[\\]([!?])/g;\n let match: RegExpExecArray | null;\n while ((match = listPattern.exec(modifier)) !== null) {\n lists.push(`[]${match[1]}` as \"[]!\" | \"[]?\");\n }\n\n return { inner, lists };\n};\n\n/**\n * Rebuild modifier string from structure.\n */\nconst buildModifier = (structure: ModifierStructure): string => {\n return structure.inner + structure.lists.join(\"\");\n};\n\n/**\n * Check if source modifier can be assigned to target modifier.\n * Implements GraphQL List Coercion: depth difference of 0 or 1 is allowed.\n *\n * Rules:\n * - A single value can be coerced into a list (depth diff = 1)\n * - At each level, non-null can be assigned to nullable (but not vice versa)\n *\n * @param source - The modifier of the value being assigned (variable's type)\n * @param target - The modifier expected by the position (field argument's type)\n * @returns true if assignment is valid\n */\nexport const isModifierAssignable = (source: string, target: string): boolean => {\n const srcStruct = parseModifierStructure(source);\n const tgtStruct = parseModifierStructure(target);\n\n const depthDiff = tgtStruct.lists.length - srcStruct.lists.length;\n\n // Depth difference must be 0 or 1 (List Coercion only wraps one level)\n if (depthDiff < 0 || depthDiff > 1) return false;\n\n // When List Coercion applies (depth diff = 1), compare inner levels only\n // The outer list level of target is satisfied by the coercion itself\n const tgtListsToCompare = depthDiff === 1 ? tgtStruct.lists.slice(1) : tgtStruct.lists;\n\n // When coercing a nullable scalar to a list, target's outer list must be nullable\n // A null scalar would produce null (not [null]), violating non-null list constraint\n // This only applies to scalars (no lists), not when coercing list to nested list\n if (depthDiff === 1 && srcStruct.lists.length === 0 && srcStruct.inner === \"?\" && tgtStruct.lists[0] === \"[]!\") {\n return false;\n }\n\n // Inner nullability: non-null can go to nullable, but not vice versa\n if (srcStruct.inner === \"?\" && tgtStruct.inner === \"!\") return false;\n\n // List level nullability: check each corresponding level\n for (let i = 0; i < srcStruct.lists.length; i++) {\n const srcList = srcStruct.lists[i]!;\n const tgtList = tgtListsToCompare[i]!;\n if (srcList === \"[]?\" && tgtList === \"[]!\") return false;\n }\n\n return true;\n};\n\n/**\n * Derive minimum modifier needed to satisfy expected modifier.\n * When List Coercion can apply, returns one level shallower.\n *\n * @param expectedModifier - The modifier expected by the field argument\n * @returns The minimum modifier the variable must have\n */\nconst deriveMinimumModifier = (expectedModifier: string): string => {\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n return buildModifier({ inner: struct.inner, lists: struct.lists.slice(1) });\n }\n return expectedModifier;\n};\n\n/**\n * Merge two modifiers by taking the stricter constraint at each level.\n * - Non-null (!) is stricter than nullable (?)\n * - List depths must match\n *\n * @param a - First modifier\n * @param b - Second modifier\n * @returns Merged modifier or error if incompatible\n */\nexport const mergeModifiers = (a: string, b: string): { ok: true; value: string } | { ok: false; reason: string } => {\n const structA = parseModifierStructure(a);\n const structB = parseModifierStructure(b);\n\n // List depths must match\n if (structA.lists.length !== structB.lists.length) {\n return {\n ok: false,\n reason: `Incompatible list depths: \"${a}\" has ${structA.lists.length} list level(s), \"${b}\" has ${structB.lists.length}`,\n };\n }\n\n // Take stricter inner constraint (! beats ?)\n const mergedInner: \"!\" | \"?\" = structA.inner === \"!\" || structB.inner === \"!\" ? \"!\" : \"?\";\n\n // Merge each list level (! beats ?)\n const mergedLists: (\"[]!\" | \"[]?\")[] = [];\n for (let i = 0; i < structA.lists.length; i++) {\n const listA = structA.lists[i]!;\n const listB = structB.lists[i]!;\n mergedLists.push(listA === \"[]!\" || listB === \"[]!\" ? \"[]!\" : \"[]?\");\n }\n\n return { ok: true, value: buildModifier({ inner: mergedInner, lists: mergedLists }) };\n};\n\n// ============================================================================\n// Variable Collection from Selections\n// ============================================================================\n\n/**\n * Intermediate type for tracking variable usages before merging.\n */\ntype VariableUsage = {\n readonly name: string;\n readonly typeName: string;\n /** The modifier expected by the field/argument position */\n readonly expectedModifier: string;\n /** The minimum modifier the variable needs (after applying List Coercion) */\n readonly minimumModifier: string;\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Get the expected type for a field argument from the schema.\n * Returns null if the field or argument is not found.\n */\nexport const getArgumentType = (\n schema: SchemaIndex,\n parentTypeName: string,\n fieldName: string,\n argumentName: string,\n): TypeInfo | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const argDef = fieldDef.arguments?.find((arg) => arg.name.value === argumentName);\n if (!argDef) return null;\n\n return parseTypeNode(argDef.type);\n};\n\n/**\n * Get the expected type for an input object field from the schema.\n */\nexport const getInputFieldType = (schema: SchemaIndex, inputTypeName: string, fieldName: string): TypeInfo | null => {\n const inputRecord = schema.inputs.get(inputTypeName);\n if (!inputRecord) return null;\n\n const fieldDef = inputRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n return parseTypeNode(fieldDef.type);\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKindFromName = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) return \"scalar\";\n if (isEnumName(schema, typeName)) return \"enum\";\n if (schema.inputs.has(typeName)) return \"input\";\n return null;\n};\n\n/**\n * Extract variable usages from a parsed value, given the expected type.\n * Handles nested input objects recursively.\n */\nconst collectVariablesFromValue = (\n value: ParsedValue,\n expectedTypeName: string,\n expectedModifier: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n if (value.kind === \"variable\") {\n const typeKind = resolveTypeKindFromName(schema, expectedTypeName);\n if (!typeKind) {\n return {\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${expectedTypeName}\" for variable \"$${value.name}\"`,\n typeName: expectedTypeName,\n };\n }\n usages.push({\n name: value.name,\n typeName: expectedTypeName,\n expectedModifier,\n minimumModifier: deriveMinimumModifier(expectedModifier),\n typeKind,\n });\n return null;\n }\n\n if (value.kind === \"object\") {\n // For object values, check each field against input type definition\n for (const field of value.fields) {\n const fieldType = getInputFieldType(schema, expectedTypeName, field.name);\n if (!fieldType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${field.name}\" on input type \"${expectedTypeName}\"`,\n typeName: expectedTypeName,\n fieldName: field.name,\n };\n }\n const error = collectVariablesFromValue(field.value, fieldType.typeName, fieldType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n }\n\n if (value.kind === \"list\") {\n // For list values, unwrap one level of list modifier\n // e.g., [ID!]! -> ID! for elements\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n const innerModifier = buildModifier({\n inner: struct.inner,\n lists: struct.lists.slice(1),\n });\n for (const item of value.values) {\n const error = collectVariablesFromValue(item, expectedTypeName, innerModifier, schema, usages);\n if (error) return error;\n }\n }\n }\n\n // Other value kinds (int, float, string, etc.) don't contain variables\n return null;\n};\n\n/**\n * Collect variable usages from field arguments.\n */\nconst collectVariablesFromArguments = (\n args: readonly ParsedArgument[],\n parentTypeName: string,\n fieldName: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n for (const arg of args) {\n const argType = getArgumentType(schema, parentTypeName, fieldName, arg.name);\n if (!argType) {\n return {\n code: \"GRAPHQL_UNKNOWN_ARGUMENT\",\n message: `Unknown argument \"${arg.name}\" on field \"${fieldName}\"`,\n fieldName,\n argumentName: arg.name,\n };\n }\n const error = collectVariablesFromValue(arg.value, argType.typeName, argType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n};\n\n/**\n * Recursively collect all variable usages from selections.\n */\nexport const collectVariableUsages = (\n selections: readonly ParsedSelection[],\n parentTypeName: string,\n schema: SchemaIndex,\n): Result<VariableUsage[], GraphqlCompatError> => {\n const usages: VariableUsage[] = [];\n\n const collect = (sels: readonly ParsedSelection[], parentType: string): GraphqlCompatError | null => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"field\": {\n // Collect from arguments\n if (sel.arguments && sel.arguments.length > 0) {\n const error = collectVariablesFromArguments(sel.arguments, parentType, sel.name, schema, usages);\n if (error) return error;\n }\n\n // Recurse into nested selections\n if (sel.selections && sel.selections.length > 0) {\n // Need to determine the field's return type for nested selections\n const fieldReturnType = getFieldReturnType(schema, parentType, sel.name);\n if (!fieldReturnType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${sel.name}\" on type \"${parentType}\"`,\n typeName: parentType,\n fieldName: sel.name,\n };\n }\n const error = collect(sel.selections, fieldReturnType);\n if (error) return error;\n }\n break;\n }\n case \"inlineFragment\": {\n // Use the inline fragment's type condition\n const error = collect(sel.selections, sel.onType);\n if (error) return error;\n break;\n }\n case \"fragmentSpread\":\n // Fragment spreads are handled separately (variables from spread fragments)\n break;\n }\n }\n return null;\n };\n\n const error = collect(selections, parentTypeName);\n if (error) return err(error);\n\n return ok(usages);\n};\n\n/**\n * Get the return type of a field (unwrapped from modifiers).\n */\nexport const getFieldReturnType = (schema: SchemaIndex, parentTypeName: string, fieldName: string): string | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const { typeName } = parseTypeNode(fieldDef.type);\n return typeName;\n};\n\n/**\n * Merge multiple variable usages into a single InferredVariable.\n * Validates type compatibility and merges modifiers using List Coercion rules.\n *\n * The algorithm:\n * 1. Validate all usages have the same type name\n * 2. Merge minimumModifiers to find the candidate (shallowest type that could work)\n * 3. Verify the candidate can satisfy ALL expected modifiers via isModifierAssignable\n */\nexport const mergeVariableUsages = (\n variableName: string,\n usages: readonly VariableUsage[],\n): Result<InferredVariable, GraphqlCompatError> => {\n if (usages.length === 0) {\n // This shouldn't happen, but handle defensively\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `No usages found for variable \"${variableName}\"`,\n variableName,\n });\n }\n\n const first = usages[0]!;\n\n // Validate all usages have the same type name\n for (const usage of usages) {\n if (usage.typeName !== first.typeName) {\n return err({\n code: \"GRAPHQL_VARIABLE_TYPE_MISMATCH\",\n message: `Variable \"$${variableName}\" has conflicting types: \"${first.typeName}\" and \"${usage.typeName}\"`,\n variableName,\n });\n }\n }\n\n // Merge minimumModifiers to find candidate\n let candidateModifier = first.minimumModifier;\n for (let i = 1; i < usages.length; i++) {\n const result = mergeModifiers(candidateModifier, usages[i]!.minimumModifier);\n if (!result.ok) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" has incompatible modifiers: ${result.reason}`,\n variableName,\n });\n }\n candidateModifier = result.value;\n }\n\n // Verify candidate satisfies all expected modifiers\n for (const usage of usages) {\n if (!isModifierAssignable(candidateModifier, usage.expectedModifier)) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" with modifier \"${candidateModifier}\" cannot satisfy expected \"${usage.expectedModifier}\"`,\n variableName,\n });\n }\n }\n\n return ok({\n name: variableName,\n typeName: first.typeName,\n modifier: candidateModifier,\n typeKind: first.typeKind,\n });\n};\n\n/**\n * Infer variables from collected usages.\n * Groups by variable name and merges each group.\n */\nexport const inferVariablesFromUsages = (usages: readonly VariableUsage[]): Result<InferredVariable[], GraphqlCompatError> => {\n // Group usages by variable name\n const byName = new Map<string, VariableUsage[]>();\n for (const usage of usages) {\n const existing = byName.get(usage.name);\n if (existing) {\n existing.push(usage);\n } else {\n byName.set(usage.name, [usage]);\n }\n }\n\n // Merge each group\n const variables: InferredVariable[] = [];\n for (const [name, group] of byName) {\n const result = mergeVariableUsages(name, group);\n if (result.isErr()) return err(result.error);\n variables.push(result.value);\n }\n\n // Sort by name for deterministic output\n variables.sort((a, b) => a.name.localeCompare(b.name));\n\n return ok(variables);\n};\n\n/**\n * Check if a type name is a scalar type.\n */\nconst isScalarName = (schema: SchemaIndex, name: string): boolean => builtinScalarTypes.has(name) || schema.scalars.has(name);\n\n// ============================================================================\n// Fragment Dependency Ordering\n// ============================================================================\n\n/**\n * Topologically sort fragments so dependencies come before dependents.\n * Detects circular dependencies.\n *\n * Note: Uses the existing collectFragmentDependencies function defined below.\n */\nexport const sortFragmentsByDependency = (fragments: readonly ParsedFragment[]): Result<ParsedFragment[], GraphqlCompatError> => {\n // Build dependency graph using existing function\n const graph = new Map<string, Set<string>>();\n for (const fragment of fragments) {\n const deps = collectFragmentDependenciesSet(fragment.selections);\n graph.set(fragment.name, deps);\n }\n\n const fragmentByName = new Map<string, ParsedFragment>();\n for (const f of fragments) {\n fragmentByName.set(f.name, f);\n }\n\n const sorted: ParsedFragment[] = [];\n const visited = new Set<string>();\n const visiting = new Set<string>(); // For cycle detection\n\n const visit = (name: string, path: string[]): GraphqlCompatError | null => {\n if (visited.has(name)) return null;\n\n if (visiting.has(name)) {\n // Found a cycle\n const cycleStart = path.indexOf(name);\n const cycle = path.slice(cycleStart).concat(name);\n return {\n code: \"GRAPHQL_FRAGMENT_CIRCULAR_DEPENDENCY\",\n message: `Circular dependency detected in fragments: ${cycle.join(\" -> \")}`,\n fragmentNames: cycle,\n };\n }\n\n // Fragment might not be in our list (external dependency)\n const fragment = fragmentByName.get(name);\n if (!fragment) {\n // External fragment, skip\n visited.add(name);\n return null;\n }\n\n visiting.add(name);\n const deps = graph.get(name) ?? new Set();\n\n for (const dep of deps) {\n const error = visit(dep, [...path, name]);\n if (error) return error;\n }\n\n visiting.delete(name);\n visited.add(name);\n sorted.push(fragment);\n return null;\n };\n\n for (const fragment of fragments) {\n const error = visit(fragment.name, []);\n if (error) return err(error);\n }\n\n return ok(sorted);\n};\n\n/**\n * Recursively collect fragment spread names from selections into a Set.\n * Internal helper for sortFragmentsByDependency.\n */\nconst collectFragmentDependenciesSet = (selections: readonly ParsedSelection[]): Set<string> => {\n const deps = new Set<string>();\n\n const collect = (sels: readonly ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n deps.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return deps;\n};\n\n/**\n * Check if a type name is an enum type.\n */\nconst isEnumName = (schema: SchemaIndex, name: string): boolean => schema.enums.has(name);\n\n/**\n * Enriched operation with resolved type information.\n */\nexport type EnrichedOperation = Omit<ParsedOperation, \"variables\"> & {\n readonly variables: readonly EnrichedVariable[];\n /** Fragment names used in this operation (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched fragment with resolved type information.\n */\nexport type EnrichedFragment = ParsedFragment & {\n /** Fragment names used in this fragment (for imports) */\n readonly fragmentDependencies: readonly string[];\n /** Variables inferred from field arguments in this fragment */\n readonly variables: readonly InferredVariable[];\n};\n\n/**\n * Enriched variable with resolved type kind.\n */\nexport type EnrichedVariable = Omit<ParsedVariable, \"typeKind\"> & {\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Result of transforming parsed operations.\n */\nexport type TransformResult = {\n readonly operations: readonly EnrichedOperation[];\n readonly fragments: readonly EnrichedFragment[];\n};\n\n/**\n * Options for transformation.\n */\nexport type TransformOptions = {\n /** Schema document for type resolution */\n readonly schemaDocument: DocumentNode;\n};\n\n/**\n * Transform parsed operations/fragments by enriching them with schema information.\n *\n * This resolves variable type kinds (scalar, enum, input), collects\n * fragment dependencies, and infers variables for fragments.\n */\nexport const transformParsedGraphql = (\n parsed: ParseResult,\n options: TransformOptions,\n): Result<TransformResult, GraphqlCompatError> => {\n const schema = createSchemaIndex(options.schemaDocument);\n\n // Sort fragments by dependency (dependencies first)\n const sortResult = sortFragmentsByDependency(parsed.fragments);\n if (sortResult.isErr()) {\n return err(sortResult.error);\n }\n const sortedFragments = sortResult.value;\n\n // Transform fragments in dependency order, building up resolved variables map\n const resolvedFragmentVariables = new Map<string, readonly InferredVariable[]>();\n const fragments: EnrichedFragment[] = [];\n\n for (const frag of sortedFragments) {\n const result = transformFragment(frag, schema, resolvedFragmentVariables);\n if (result.isErr()) {\n return err(result.error);\n }\n resolvedFragmentVariables.set(frag.name, result.value.variables);\n fragments.push(result.value);\n }\n\n // Transform operations\n const operations: EnrichedOperation[] = [];\n for (const op of parsed.operations) {\n const result = transformOperation(op, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n operations.push(result.value);\n }\n\n return ok({ operations, fragments });\n};\n\n/**\n * Transform a single operation.\n */\nconst transformOperation = (op: ParsedOperation, schema: SchemaIndex): Result<EnrichedOperation, GraphqlCompatError> => {\n // Resolve variable type kinds\n const variables: EnrichedVariable[] = [];\n for (const v of op.variables) {\n const typeKind = resolveTypeKind(schema, v.typeName);\n if (typeKind === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${v.typeName}\" in variable \"${v.name}\"`,\n typeName: v.typeName,\n });\n }\n variables.push({ ...v, typeKind });\n }\n\n // Collect fragment dependencies\n const fragmentDependencies = collectFragmentDependencies(op.selections);\n\n return ok({\n ...op,\n variables,\n fragmentDependencies,\n });\n};\n\n/**\n * Transform a single fragment.\n * Infers variables from field arguments and propagates variables from spread fragments.\n */\nconst transformFragment = (\n frag: ParsedFragment,\n schema: SchemaIndex,\n resolvedFragmentVariables: Map<string, readonly InferredVariable[]>,\n): Result<EnrichedFragment, GraphqlCompatError> => {\n // Collect fragment dependencies (fragments used within this fragment)\n const fragmentDependencies = collectFragmentDependencies(frag.selections);\n\n // Collect direct variable usages from this fragment's selections\n const directUsagesResult = collectVariableUsages(frag.selections, frag.onType, schema);\n if (directUsagesResult.isErr()) {\n return err(directUsagesResult.error);\n }\n const directUsages = directUsagesResult.value;\n\n // Collect variables from spread fragments\n const spreadVariables: InferredVariable[] = [];\n for (const depName of fragmentDependencies) {\n const depVariables = resolvedFragmentVariables.get(depName);\n if (depVariables) {\n spreadVariables.push(...depVariables);\n }\n // If not found, it's an external fragment - skip\n }\n\n // Combine direct usages with spread variables\n // Convert spread variables (InferredVariable) to usages for merging\n // For already-inferred variables, expectedModifier and minimumModifier are the same\n const allUsages: VariableUsage[] = [\n ...directUsages,\n ...spreadVariables.map((v) => ({\n name: v.name,\n typeName: v.typeName,\n expectedModifier: v.modifier,\n minimumModifier: v.modifier,\n typeKind: v.typeKind,\n })),\n ];\n\n // Infer final variables\n const variablesResult = inferVariablesFromUsages(allUsages);\n if (variablesResult.isErr()) {\n return err(variablesResult.error);\n }\n\n return ok({\n ...frag,\n fragmentDependencies,\n variables: variablesResult.value,\n });\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKind = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) {\n return \"scalar\";\n }\n if (isEnumName(schema, typeName)) {\n return \"enum\";\n }\n if (schema.inputs.has(typeName)) {\n return \"input\";\n }\n return null;\n};\n\n/**\n * Collect fragment names used in selections (recursively).\n */\nconst collectFragmentDependencies = (selections: readonly import(\"./types\").ParsedSelection[]): readonly string[] => {\n const fragments = new Set<string>();\n\n const collect = (sels: readonly import(\"./types\").ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n fragments.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return [...fragments];\n};\n","/**\n * Emitter for generating TypeScript compat code from enriched operations.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport {\n type EnrichedFragment,\n type EnrichedOperation,\n getArgumentType,\n getFieldReturnType,\n getInputFieldType,\n} from \"./transformer\";\nimport type { GraphqlCompatError, ParsedArgument, ParsedInlineFragment, ParsedSelection, ParsedValue, TypeInfo } from \"./types\";\n\n/**\n * Schema index for type lookups.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Options for code emission.\n */\nexport type EmitOptions = {\n /** Schema name to use in gql.schemaName() call */\n readonly schemaName: string;\n /** Import path for graphql-system module */\n readonly graphqlSystemPath: string;\n /** Schema document for type lookups (required for inline fragment support) */\n readonly schemaDocument?: DocumentNode;\n};\n\n/**\n * Map operation kind to root type name.\n */\nconst getRootTypeName = (kind: \"query\" | \"mutation\" | \"subscription\"): string => {\n switch (kind) {\n case \"query\":\n return \"Query\";\n case \"mutation\":\n return \"Mutation\";\n case \"subscription\":\n return \"Subscription\";\n }\n};\n\n/**\n * Emit TypeScript code for an operation.\n */\nexport const emitOperation = (operation: EnrichedOperation, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${operation.name}Compat`;\n const operationType = operation.kind;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);\n lines.push(` ${operationType}.compat({`);\n lines.push(` name: ${JSON.stringify(operation.name)},`);\n\n // Variables\n if (operation.variables.length > 0) {\n lines.push(` variables: { ${emitVariables(operation.variables)} },`);\n }\n\n // Fields - pass root type name for list coercion\n const rootTypeName = getRootTypeName(operation.kind);\n lines.push(` fields: ({ f, $ }) => ({`);\n const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema, rootTypeName);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit TypeScript code for a fragment.\n */\nexport const emitFragment = (fragment: EnrichedFragment, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n const hasVariables = fragment.variables.length > 0;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${fragment.name}Fragment`;\n\n // Include $var in destructure if fragment has variables\n const destructure = hasVariables ? \"fragment, $var\" : \"fragment\";\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${destructure} }) =>`);\n lines.push(` fragment.${fragment.onType}({`);\n\n // Variables block (if any)\n if (hasVariables) {\n lines.push(` variables: { ${emitVariables(fragment.variables)} },`);\n }\n\n // Fields - include $ in context if fragment has variables\n // Pass fragment.onType as the parent type for list coercion\n const fieldsContext = hasVariables ? \"{ f, $ }\" : \"{ f }\";\n lines.push(` fields: (${fieldsContext}) => ({`);\n const fieldLinesResult = emitSelections(fragment.selections, 3, fragment.variables, schema, fragment.onType);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Common variable type for emission (both EnrichedVariable and InferredVariable have these fields).\n */\ntype EmittableVariable = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n};\n\n/**\n * Emit variable definitions.\n */\nconst emitVariables = (variables: readonly EmittableVariable[]): string => {\n return variables.map((v) => `...$var(${JSON.stringify(v.name)}).${v.typeName}(${JSON.stringify(v.modifier)})`).join(\", \");\n};\n\n/**\n * Emit field selections (public API).\n * Converts variable array to Set<string> and delegates to internal implementation.\n */\nconst emitSelections = (\n selections: readonly ParsedSelection[],\n indent: number,\n variables: readonly EmittableVariable[],\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const variableNames = new Set(variables.map((v) => v.name));\n return emitSelectionsInternal(selections, indent, variableNames, schema, parentTypeName);\n};\n\n/**\n * Internal implementation for emitting field selections.\n * Takes variableNames as Set<string> for recursive calls.\n */\nconst emitSelectionsInternal = (\n selections: readonly ParsedSelection[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n\n // Separate inline fragments from other selections\n const inlineFragments: ParsedInlineFragment[] = [];\n const otherSelections: ParsedSelection[] = [];\n\n for (const sel of selections) {\n if (sel.kind === \"inlineFragment\") {\n inlineFragments.push(sel);\n } else {\n otherSelections.push(sel);\n }\n }\n\n // Emit regular selections (fields and fragment spreads)\n for (const sel of otherSelections) {\n const result = emitSingleSelection(sel, indent, variableNames, schema, parentTypeName);\n if (result.isErr()) {\n return err(result.error);\n }\n lines.push(result.value);\n }\n\n // Emit grouped inline fragments as union selections\n if (inlineFragments.length > 0) {\n const unionResult = emitInlineFragmentsAsUnion(inlineFragments, indent, variableNames, schema);\n if (unionResult.isErr()) {\n return err(unionResult.error);\n }\n lines.push(unionResult.value);\n }\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit a single selection (field or fragment spread).\n */\nconst emitSingleSelection = (\n sel: ParsedSelection,\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n switch (sel.kind) {\n case \"field\":\n return emitFieldSelection(sel, indent, variableNames, schema, parentTypeName);\n case \"fragmentSpread\":\n return ok(`${padding}...${sel.name}Fragment.spread(),`);\n case \"inlineFragment\":\n // This should not happen as inline fragments are handled separately\n return ok(\"\");\n }\n};\n\n/**\n * Emit inline fragments grouped as a union selection.\n * Format: { TypeA: ({ f }) => ({ ...fields }), TypeB: ({ f }) => ({ ...fields }) }\n */\nconst emitInlineFragmentsAsUnion = (\n inlineFragments: readonly ParsedInlineFragment[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Validate inline fragments have type conditions\n for (const frag of inlineFragments) {\n if (frag.onType === \"\") {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_WITHOUT_TYPE\",\n message: \"Inline fragments without type condition are not supported. Use `... on TypeName { }` syntax.\",\n });\n }\n }\n\n // Validate all inline fragments are on union types (not interfaces)\n for (const frag of inlineFragments) {\n if (schema && !schema.objects.has(frag.onType)) {\n // If it's not a known object type, it might be an interface\n // Check if any union contains this type as a member\n let isUnionMember = false;\n for (const [, unionDef] of schema.unions) {\n if (unionDef.members.has(frag.onType)) {\n isUnionMember = true;\n break;\n }\n }\n if (!isUnionMember) {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_ON_INTERFACE\",\n message: `Inline fragments on interface type \"${frag.onType}\" are not supported. Use union types instead.`,\n onType: frag.onType,\n });\n }\n }\n }\n\n // Build union member entries\n const entries: string[] = [];\n for (const frag of inlineFragments) {\n const innerPadding = \" \".repeat(indent + 1);\n // Use the inline fragment's type condition as the parent type for nested selections\n const fieldsResult = emitSelectionsInternal(frag.selections, indent + 2, variableNames, schema, frag.onType);\n if (fieldsResult.isErr()) {\n return err(fieldsResult.error);\n }\n\n entries.push(`${innerPadding}${frag.onType}: ({ f }) => ({\n${fieldsResult.value}\n${innerPadding}}),`);\n }\n\n // Emit as spread with union callback: ...f.fieldName()({ Type: ... })\n // Note: This assumes the parent field handles the union - we emit just the union object\n return ok(`${padding}...({\n${entries.join(\"\\n\")}\n${padding}}),`);\n};\n\n/**\n * Emit a single field selection.\n */\nconst emitFieldSelection = (\n field: ParsedSelection & { kind: \"field\" },\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Extract optional fields for type narrowing\n const args = field.arguments;\n const selections = field.selections;\n const hasArgs = args && args.length > 0;\n const hasSelections = selections && selections.length > 0;\n\n // Use shorthand syntax for scalar fields (no args, no nested selections)\n if (!hasArgs && !hasSelections) {\n return ok(`${padding}${field.name}: true,`);\n }\n\n let line = `${padding}...f.${field.name}(`;\n\n if (hasArgs) {\n const argsResult = emitArguments(args, variableNames, schema, parentTypeName, field.name);\n if (argsResult.isErr()) {\n return err(argsResult.error);\n }\n line += argsResult.value;\n }\n\n line += \")\";\n\n if (hasSelections) {\n // Check if selections contain inline fragments (union field)\n const hasInlineFragments = selections.some((s) => s.kind === \"inlineFragment\");\n\n // Determine nested parent type for recursive selections\n const nestedParentType =\n schema && parentTypeName ? (getFieldReturnType(schema, parentTypeName, field.name) ?? undefined) : undefined;\n\n if (hasInlineFragments) {\n // Union field: emit with union callback pattern\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema, nestedParentType);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += \"({\\n\";\n line += `${nestedResult.value}\\n`;\n line += `${padding}})`;\n } else {\n // Regular nested selections\n line += \"(({ f }) => ({\\n\";\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema, nestedParentType);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += `${nestedResult.value}\\n`;\n line += `${padding}}))`;\n }\n }\n\n line += \",\";\n\n return ok(line);\n};\n\n// ============================================================================\n// List Coercion Utilities\n// ============================================================================\n\n/**\n * Check if a modifier represents a list type (contains []).\n */\nconst isListModifier = (modifier: string): boolean => {\n return modifier.includes(\"[]\");\n};\n\n/**\n * Determine if a value needs to be wrapped in an array for list coercion.\n * Returns true if:\n * - Expected type is a list\n * - Value is NOT already a list\n * - Value is NOT a variable (runtime handles coercion)\n * - Value is NOT null\n */\nconst needsListCoercion = (value: ParsedValue, expectedModifier: string | undefined): boolean => {\n // No coercion if no expected type info\n if (!expectedModifier) return false;\n\n // No coercion if expected type is not a list\n if (!isListModifier(expectedModifier)) return false;\n\n // No coercion for variables (runtime handles this)\n if (value.kind === \"variable\") return false;\n\n // No coercion for null\n if (value.kind === \"null\") return false;\n\n // No coercion if value is already a list\n if (value.kind === \"list\") return false;\n\n return true;\n};\n\n/**\n * Extract the element type from a list type by removing the outermost list modifier.\n * For example: \"![]!\" (non-null list of non-null) → \"!\" (non-null element)\n * \"?[]![]!\" (nested lists) → \"?[]!\" (inner list type)\n * Returns null if the modifier doesn't represent a list type.\n */\nconst getListElementType = (expectedType: TypeInfo): TypeInfo | null => {\n const { modifier, typeName } = expectedType;\n // Modifier format: {inner}{list_modifiers}\n // Strip the outermost list modifier ([]! or []?)\n const listMatch = modifier.match(/^(.+?)(\\[\\][!?])$/);\n if (!listMatch || !listMatch[1]) return null;\n return { typeName, modifier: listMatch[1] };\n};\n\n/**\n * Emit a value with type context for list coercion.\n */\nconst emitValueWithType = (\n value: ParsedValue,\n expectedType: TypeInfo | null,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n // Check if list coercion is needed\n const shouldCoerce = needsListCoercion(value, expectedType?.modifier);\n\n // Handle object values with recursive type context\n if (value.kind === \"object\" && expectedType && schema) {\n return emitObjectWithType(value, expectedType.typeName, variableNames, schema, shouldCoerce);\n }\n\n // Handle list values with recursive type context for element coercion\n if (value.kind === \"list\" && expectedType && schema) {\n const elementType = getListElementType(expectedType);\n if (elementType) {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValueWithType(v, elementType, variableNames, schema);\n if (result.isErr()) return result;\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n }\n\n // Emit the value normally\n const result = emitValue(value, variableNames);\n if (result.isErr()) return result;\n\n // Wrap in array if coercion needed\n if (shouldCoerce) {\n return ok(`[${result.value}]`);\n }\n\n return result;\n};\n\n/**\n * Emit an object value with type context for recursive list coercion.\n */\nconst emitObjectWithType = (\n value: ParsedValue & { kind: \"object\" },\n inputTypeName: string,\n variableNames: Set<string>,\n schema: SchemaIndex,\n wrapInArray: boolean,\n): Result<string, GraphqlCompatError> => {\n if (value.fields.length === 0) {\n return ok(wrapInArray ? \"[{}]\" : \"{}\");\n }\n\n const entries: string[] = [];\n for (const f of value.fields) {\n // Look up field type from input object definition\n const fieldType = getInputFieldType(schema, inputTypeName, f.name);\n if (fieldType === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${f.name}\" on input type \"${inputTypeName}\"`,\n typeName: inputTypeName,\n fieldName: f.name,\n });\n }\n\n const result = emitValueWithType(f.value, fieldType, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n\n const objectStr = `{ ${entries.join(\", \")} }`;\n return ok(wrapInArray ? `[${objectStr}]` : objectStr);\n};\n\n// ============================================================================\n// Argument Emission\n// ============================================================================\n\n/**\n * Emit field arguments with type context for list coercion.\n */\nconst emitArguments = (\n args: readonly ParsedArgument[],\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n fieldName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n if (args.length === 0) {\n return ok(\"\");\n }\n\n const argEntries: string[] = [];\n for (const arg of args) {\n // Look up expected type from schema\n const expectedType =\n schema && parentTypeName && fieldName ? getArgumentType(schema, parentTypeName, fieldName, arg.name) : null;\n\n const result = emitValueWithType(arg.value, expectedType, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n argEntries.push(`${arg.name}: ${result.value}`);\n }\n return ok(`{ ${argEntries.join(\", \")} }`);\n};\n\n/**\n * Emit a value (literal or variable reference).\n */\nconst emitValue = (value: ParsedValue, variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n switch (value.kind) {\n case \"variable\":\n // Check if it's a declared variable\n if (variableNames.has(value.name)) {\n return ok(`$.${value.name}`);\n }\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `Variable \"$${value.name}\" is not declared in the operation`,\n variableName: value.name,\n });\n case \"int\":\n case \"float\":\n return ok(value.value);\n case \"string\":\n return ok(JSON.stringify(value.value));\n case \"boolean\":\n return ok(value.value ? \"true\" : \"false\");\n case \"null\":\n return ok(\"null\");\n case \"enum\":\n // Enums are emitted as string literals in soda-gql\n return ok(JSON.stringify(value.value));\n case \"list\": {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValue(v, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n case \"object\": {\n if (value.fields.length === 0) {\n return ok(\"{}\");\n }\n const entries: string[] = [];\n for (const f of value.fields) {\n const result = emitValue(f.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n return ok(`{ ${entries.join(\", \")} }`);\n }\n }\n};\n","import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, rmSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const removeDirectory = (dirPath: string) => {\n const targetPath = resolve(dirPath);\n try {\n rmSync(targetPath, { recursive: true, force: true });\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"REMOVE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport type { TypeFilterConfig } from \"@soda-gql/config\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { removeDirectory, writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Build typeFilters config for each schema\n const typeFiltersConfig = new Map<string, TypeFilterConfig>();\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.typeFilter) {\n typeFiltersConfig.set(schemaName, schemaConfig.typeFilter);\n }\n }\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n typeFilters: typeFiltersConfig.size > 0 ? typeFiltersConfig : undefined,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Clean up existing _defs directory to prevent stale files\n const defsDir = join(outDir, \"_defs\");\n if (existsSync(defsDir)) {\n const removeResult = removeDirectory(defsDir);\n if (removeResult.isErr()) {\n return err(removeResult.error);\n }\n }\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;;;;;;;;;AA2CA,MAAa,oBAAoB,aAA8D;CAC7F,MAAM,sCAAuB,SAAS;AAEtC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAW;GACT,MAAM;GACN,SAAS,6BAA6B;GACtC,UAAU;GACX,CAAC;;AAGJ,KAAI;EACF,MAAM,mCAAsB,cAAc,OAAO;EACjD,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,aAAa,CAAC;UAC/C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAa,sBAAsB,QAAgB,eAAgE;AACjH,KAAI;EACF,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,WAAW,CAAC;UAC7C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAM,uBAAuB,UAAwB,eAAoC;CACvF,MAAMA,aAAgC,EAAE;CACxC,MAAMC,YAA8B,EAAE;AAEtC,MAAK,MAAM,cAAc,SAAS,aAAa;AAC7C,MAAI,WAAW,SAASC,aAAK,sBAAsB;GACjD,MAAM,YAAY,iBAAiB,YAAY,WAAW;AAC1D,OAAI,WAAW;AACb,eAAW,KAAK,UAAU;;aAEnB,WAAW,SAASA,aAAK,qBAAqB;AACvD,aAAU,KAAK,gBAAgB,YAAY,WAAW,CAAC;;;AAI3D,QAAO;EAAE;EAAY;EAAW;;;;;AAMlC,MAAM,oBAAoB,MAA+B,eAA+C;AAEtG,KAAI,CAAC,KAAK,MAAM;AACd,SAAO;;CAGT,MAAMC,aAA+B,KAAK,uBAAuB,EAAE,EAAE,IAAI,gBAAgB;CACzF,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK;EACX,MAAM,KAAK,KAAK;EAChB;EACA;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,MAA8B,eAAuC;CAC5F,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,QAAQ,KAAK,cAAc,KAAK;EAChC;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,SAAiD;CACxE,MAAM,EAAE,UAAU,aAAa,cAAc,KAAK,KAAK;CACvD,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,aAAa,GAAG;AAE3E,QAAO;EACL,MAAM,KAAK,SAAS,KAAK;EACzB;EACA;EAEA,UAAU;EACV;EACD;;;;;;;;;AAUH,MAAa,iBAAiB,SAA6B;CAEzD,MAAMC,SAAsB,EAAE;CAE9B,MAAM,WAAW,GAAa,YAA6B;AACzD,MAAI,EAAE,SAASF,aAAK,eAAe;AACjC,UAAO,QAAQ,EAAE,MAAM,KAAK;;AAE9B,MAAI,EAAE,SAASA,aAAK,WAAW;AAC7B,UAAO,KAAK;IAAE,MAAM;IAAQ;IAAS,CAAC;AACtC,UAAO,QAAQ,EAAE,MAAM,MAAM;;AAE/B,SAAO,KAAK;GAAE,MAAM;GAAS;GAAS,CAAC;AACvC,SAAO,EAAE,KAAK;;CAGhB,MAAM,WAAW,QAAQ,MAAM,MAAM;CAGrC,IAAI,WAAW;AACf,MAAK,MAAM,SAAS,OAAO,OAAO,CAAC,SAAS,EAAE;AAC5C,MAAI,MAAM,SAAS,SAAS;AAC1B,cAAW,MAAM,UAAU,MAAM;AACjC;;EAEF,MAAM,aAAa,MAAM,UAAU,QAAQ;AAC3C,aAAW,GAAG,WAAW;;AAG3B,QAAO;EAAE;EAAU;EAAU;;;;;AAM/B,MAAM,qBAAqB,eAA4D;AACrF,QAAO,WAAW,IAAI,iBAAiB;;;;;AAMzC,MAAM,oBAAoB,SAAyC;AACjE,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,MACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;;;;;;AAOxC,MAAM,yBAAyB,SAA0C;CACvE,MAAM,OAAO,KAAK,WAAW,SAAS,KAAK,UAAU,IAAI,gBAAgB,GAAG;CAC5E,MAAM,aAAa,KAAK,eAAe,kBAAkB,KAAK,aAAa,WAAW,GAAG;AAEzF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EAChB,OAAO,KAAK,OAAO;EACnB,WAAW;EACX;EACD;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EACjB;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,QAAQ,KAAK,eAAe,KAAK,SAAS;EAC1C,YAAY,kBAAkB,KAAK,aAAa,WAAW;EAC5D;;;;;AAMH,MAAM,mBAAmB,SAAuC;AAC9D,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,OAAO,aAAa,KAAK,MAAM;EAChC;;;;;AAMH,MAAM,qBAAqB,UAAwB;AACjD,OAAM,IAAI,MAAM,qBAAqB,KAAK,UAAU,MAAM,GAAG;;;;;AAM/D,MAAM,gBAAgB,SAAiC;AACrD,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,SACR,QAAO;GAAE,MAAM;GAAY,MAAM,KAAK,KAAK;GAAO;EACpD,KAAKA,aAAK,IACR,QAAO;GAAE,MAAM;GAAO,OAAO,KAAK;GAAO;EAC3C,KAAKA,aAAK,MACR,QAAO;GAAE,MAAM;GAAS,OAAO,KAAK;GAAO;EAC7C,KAAKA,aAAK,OACR,QAAO;GAAE,MAAM;GAAU,OAAO,KAAK;GAAO;EAC9C,KAAKA,aAAK,QACR,QAAO;GAAE,MAAM;GAAW,OAAO,KAAK;GAAO;EAC/C,KAAKA,aAAK,KACR,QAAO,EAAE,MAAM,QAAQ;EACzB,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,OAAO,KAAK;GAAO;EAC5C,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,QAAQ,KAAK,OAAO,IAAI,aAAa;GAAE;EAChE,KAAKA,aAAK,OACR,QAAO;GACL,MAAM;GACN,QAAQ,KAAK,OAAO,KAAK,WAAW;IAClC,MAAM,MAAM,KAAK;IACjB,OAAO,aAAa,MAAM,MAAM;IACjC,EAAE;GACJ;EACH,QACE,QAAO,kBAAkB,KAAK;;;;;;;;;AChRpC,MAAM,qBAAqB,IAAI,IAAI;CAAC;CAAM;CAAU;CAAO;CAAS;CAAU,CAAC;;;;;;AAoB/E,MAAM,0BAA0B,aAAwC;CAEtE,MAAM,QAAQ,SAAS,OAAO,MAAM,MAAM;CAG1C,MAAMG,QAA2B,EAAE;CACnC,MAAM,cAAc;CACpB,IAAIC;AACJ,SAAQ,QAAQ,YAAY,KAAK,SAAS,MAAM,MAAM;AACpD,QAAM,KAAK,KAAK,MAAM,KAAsB;;AAG9C,QAAO;EAAE;EAAO;EAAO;;;;;AAMzB,MAAM,iBAAiB,cAAyC;AAC9D,QAAO,UAAU,QAAQ,UAAU,MAAM,KAAK,GAAG;;;;;;;;;;;;;;AAenD,MAAa,wBAAwB,QAAgB,WAA4B;CAC/E,MAAM,YAAY,uBAAuB,OAAO;CAChD,MAAM,YAAY,uBAAuB,OAAO;CAEhD,MAAM,YAAY,UAAU,MAAM,SAAS,UAAU,MAAM;AAG3D,KAAI,YAAY,KAAK,YAAY,EAAG,QAAO;CAI3C,MAAM,oBAAoB,cAAc,IAAI,UAAU,MAAM,MAAM,EAAE,GAAG,UAAU;AAKjF,KAAI,cAAc,KAAK,UAAU,MAAM,WAAW,KAAK,UAAU,UAAU,OAAO,UAAU,MAAM,OAAO,OAAO;AAC9G,SAAO;;AAIT,KAAI,UAAU,UAAU,OAAO,UAAU,UAAU,IAAK,QAAO;AAG/D,MAAK,IAAI,IAAI,GAAG,IAAI,UAAU,MAAM,QAAQ,KAAK;EAC/C,MAAM,UAAU,UAAU,MAAM;EAChC,MAAM,UAAU,kBAAkB;AAClC,MAAI,YAAY,SAAS,YAAY,MAAO,QAAO;;AAGrD,QAAO;;;;;;;;;AAUT,MAAM,yBAAyB,qBAAqC;CAClE,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,KAAI,OAAO,MAAM,SAAS,GAAG;AAC3B,SAAO,cAAc;GAAE,OAAO,OAAO;GAAO,OAAO,OAAO,MAAM,MAAM,EAAE;GAAE,CAAC;;AAE7E,QAAO;;;;;;;;;;;AAYT,MAAa,kBAAkB,GAAW,MAA2E;CACnH,MAAM,UAAU,uBAAuB,EAAE;CACzC,MAAM,UAAU,uBAAuB,EAAE;AAGzC,KAAI,QAAQ,MAAM,WAAW,QAAQ,MAAM,QAAQ;AACjD,SAAO;GACL,IAAI;GACJ,QAAQ,8BAA8B,EAAE,QAAQ,QAAQ,MAAM,OAAO,mBAAmB,EAAE,QAAQ,QAAQ,MAAM;GACjH;;CAIH,MAAMC,cAAyB,QAAQ,UAAU,OAAO,QAAQ,UAAU,MAAM,MAAM;CAGtF,MAAMC,cAAiC,EAAE;AACzC,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,MAAM,QAAQ,KAAK;EAC7C,MAAM,QAAQ,QAAQ,MAAM;EAC5B,MAAM,QAAQ,QAAQ,MAAM;AAC5B,cAAY,KAAK,UAAU,SAAS,UAAU,QAAQ,QAAQ,MAAM;;AAGtE,QAAO;EAAE,IAAI;EAAM,OAAO,cAAc;GAAE,OAAO;GAAa,OAAO;GAAa,CAAC;EAAE;;;;;;AAwBvF,MAAa,mBACX,QACA,gBACA,WACA,iBACoB;CACpB,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,SAAS,SAAS,WAAW,MAAM,QAAQ,IAAI,KAAK,UAAU,aAAa;AACjF,KAAI,CAAC,OAAQ,QAAO;AAEpB,QAAO,cAAc,OAAO,KAAK;;;;;AAMnC,MAAa,qBAAqB,QAAqB,eAAuB,cAAuC;CACnH,MAAM,cAAc,OAAO,OAAO,IAAI,cAAc;AACpD,KAAI,CAAC,YAAa,QAAO;CAEzB,MAAM,WAAW,YAAY,OAAO,IAAI,UAAU;AAClD,KAAI,CAAC,SAAU,QAAO;AAEtB,QAAO,cAAc,SAAS,KAAK;;;;;AAMrC,MAAM,2BAA2B,QAAqB,aAAyD;AAC7G,KAAI,aAAa,QAAQ,SAAS,CAAE,QAAO;AAC3C,KAAI,WAAW,QAAQ,SAAS,CAAE,QAAO;AACzC,KAAI,OAAO,OAAO,IAAI,SAAS,CAAE,QAAO;AACxC,QAAO;;;;;;AAOT,MAAM,6BACJ,OACA,kBACA,kBACA,QACA,WAC8B;AAC9B,KAAI,MAAM,SAAS,YAAY;EAC7B,MAAM,WAAW,wBAAwB,QAAQ,iBAAiB;AAClE,MAAI,CAAC,UAAU;AACb,UAAO;IACL,MAAM;IACN,SAAS,iBAAiB,iBAAiB,mBAAmB,MAAM,KAAK;IACzE,UAAU;IACX;;AAEH,SAAO,KAAK;GACV,MAAM,MAAM;GACZ,UAAU;GACV;GACA,iBAAiB,sBAAsB,iBAAiB;GACxD;GACD,CAAC;AACF,SAAO;;AAGT,KAAI,MAAM,SAAS,UAAU;AAE3B,OAAK,MAAM,SAAS,MAAM,QAAQ;GAChC,MAAM,YAAY,kBAAkB,QAAQ,kBAAkB,MAAM,KAAK;AACzE,OAAI,CAAC,WAAW;AACd,WAAO;KACL,MAAM;KACN,SAAS,kBAAkB,MAAM,KAAK,mBAAmB,iBAAiB;KAC1E,UAAU;KACV,WAAW,MAAM;KAClB;;GAEH,MAAM,QAAQ,0BAA0B,MAAM,OAAO,UAAU,UAAU,UAAU,UAAU,QAAQ,OAAO;AAC5G,OAAI,MAAO,QAAO;;AAEpB,SAAO;;AAGT,KAAI,MAAM,SAAS,QAAQ;EAGzB,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,MAAI,OAAO,MAAM,SAAS,GAAG;GAC3B,MAAM,gBAAgB,cAAc;IAClC,OAAO,OAAO;IACd,OAAO,OAAO,MAAM,MAAM,EAAE;IAC7B,CAAC;AACF,QAAK,MAAM,QAAQ,MAAM,QAAQ;IAC/B,MAAM,QAAQ,0BAA0B,MAAM,kBAAkB,eAAe,QAAQ,OAAO;AAC9F,QAAI,MAAO,QAAO;;;;AAMxB,QAAO;;;;;AAMT,MAAM,iCACJ,MACA,gBACA,WACA,QACA,WAC8B;AAC9B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,UAAU,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK;AAC5E,MAAI,CAAC,SAAS;AACZ,UAAO;IACL,MAAM;IACN,SAAS,qBAAqB,IAAI,KAAK,cAAc,UAAU;IAC/D;IACA,cAAc,IAAI;IACnB;;EAEH,MAAM,QAAQ,0BAA0B,IAAI,OAAO,QAAQ,UAAU,QAAQ,UAAU,QAAQ,OAAO;AACtG,MAAI,MAAO,QAAO;;AAEpB,QAAO;;;;;AAMT,MAAa,yBACX,YACA,gBACA,WACgD;CAChD,MAAMC,SAA0B,EAAE;CAElC,MAAM,WAAW,MAAkC,eAAkD;AACnG,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK,SAAS;AAEZ,SAAI,IAAI,aAAa,IAAI,UAAU,SAAS,GAAG;MAC7C,MAAMC,UAAQ,8BAA8B,IAAI,WAAW,YAAY,IAAI,MAAM,QAAQ,OAAO;AAChG,UAAIA,QAAO,QAAOA;;AAIpB,SAAI,IAAI,cAAc,IAAI,WAAW,SAAS,GAAG;MAE/C,MAAM,kBAAkB,mBAAmB,QAAQ,YAAY,IAAI,KAAK;AACxE,UAAI,CAAC,iBAAiB;AACpB,cAAO;QACL,MAAM;QACN,SAAS,kBAAkB,IAAI,KAAK,aAAa,WAAW;QAC5D,UAAU;QACV,WAAW,IAAI;QAChB;;MAEH,MAAMA,UAAQ,QAAQ,IAAI,YAAY,gBAAgB;AACtD,UAAIA,QAAO,QAAOA;;AAEpB;;IAEF,KAAK,kBAAkB;KAErB,MAAMA,UAAQ,QAAQ,IAAI,YAAY,IAAI,OAAO;AACjD,SAAIA,QAAO,QAAOA;AAClB;;IAEF,KAAK,iBAEH;;;AAGN,SAAO;;CAGT,MAAM,QAAQ,QAAQ,YAAY,eAAe;AACjD,KAAI,MAAO,4BAAW,MAAM;AAE5B,2BAAU,OAAO;;;;;AAMnB,MAAa,sBAAsB,QAAqB,gBAAwB,cAAqC;CACnH,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,EAAE,aAAa,cAAc,SAAS,KAAK;AACjD,QAAO;;;;;;;;;;;AAYT,MAAa,uBACX,cACA,WACiD;AACjD,KAAI,OAAO,WAAW,GAAG;AAEvB,6BAAW;GACT,MAAM;GACN,SAAS,iCAAiC,aAAa;GACvD;GACD,CAAC;;CAGJ,MAAM,QAAQ,OAAO;AAGrB,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,MAAM,aAAa,MAAM,UAAU;AACrC,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,4BAA4B,MAAM,SAAS,SAAS,MAAM,SAAS;IACvG;IACD,CAAC;;;CAKN,IAAI,oBAAoB,MAAM;AAC9B,MAAK,IAAI,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;EACtC,MAAM,SAAS,eAAe,mBAAmB,OAAO,GAAI,gBAAgB;AAC5E,MAAI,CAAC,OAAO,IAAI;AACd,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,gCAAgC,OAAO;IAC3E;IACD,CAAC;;AAEJ,sBAAoB,OAAO;;AAI7B,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,CAAC,qBAAqB,mBAAmB,MAAM,iBAAiB,EAAE;AACpE,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,mBAAmB,kBAAkB,6BAA6B,MAAM,iBAAiB;IAC7H;IACD,CAAC;;;AAIN,2BAAU;EACR,MAAM;EACN,UAAU,MAAM;EAChB,UAAU;EACV,UAAU,MAAM;EACjB,CAAC;;;;;;AAOJ,MAAa,4BAA4B,WAAqF;CAE5H,MAAM,SAAS,IAAI,KAA8B;AACjD,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,WAAW,OAAO,IAAI,MAAM,KAAK;AACvC,MAAI,UAAU;AACZ,YAAS,KAAK,MAAM;SACf;AACL,UAAO,IAAI,MAAM,MAAM,CAAC,MAAM,CAAC;;;CAKnC,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,CAAC,MAAM,UAAU,QAAQ;EAClC,MAAM,SAAS,oBAAoB,MAAM,MAAM;AAC/C,MAAI,OAAO,OAAO,CAAE,4BAAW,OAAO,MAAM;AAC5C,YAAU,KAAK,OAAO,MAAM;;AAI9B,WAAU,MAAM,GAAG,MAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC;AAEtD,2BAAU,UAAU;;;;;AAMtB,MAAM,gBAAgB,QAAqB,SAA0B,mBAAmB,IAAI,KAAK,IAAI,OAAO,QAAQ,IAAI,KAAK;;;;;;;AAY7H,MAAa,6BAA6B,cAAuF;CAE/H,MAAM,QAAQ,IAAI,KAA0B;AAC5C,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,OAAO,+BAA+B,SAAS,WAAW;AAChE,QAAM,IAAI,SAAS,MAAM,KAAK;;CAGhC,MAAM,iBAAiB,IAAI,KAA6B;AACxD,MAAK,MAAM,KAAK,WAAW;AACzB,iBAAe,IAAI,EAAE,MAAM,EAAE;;CAG/B,MAAMC,SAA2B,EAAE;CACnC,MAAM,UAAU,IAAI,KAAa;CACjC,MAAM,WAAW,IAAI,KAAa;CAElC,MAAM,SAAS,MAAc,SAA8C;AACzE,MAAI,QAAQ,IAAI,KAAK,CAAE,QAAO;AAE9B,MAAI,SAAS,IAAI,KAAK,EAAE;GAEtB,MAAM,aAAa,KAAK,QAAQ,KAAK;GACrC,MAAM,QAAQ,KAAK,MAAM,WAAW,CAAC,OAAO,KAAK;AACjD,UAAO;IACL,MAAM;IACN,SAAS,8CAA8C,MAAM,KAAK,OAAO;IACzE,eAAe;IAChB;;EAIH,MAAM,WAAW,eAAe,IAAI,KAAK;AACzC,MAAI,CAAC,UAAU;AAEb,WAAQ,IAAI,KAAK;AACjB,UAAO;;AAGT,WAAS,IAAI,KAAK;EAClB,MAAM,OAAO,MAAM,IAAI,KAAK,IAAI,IAAI,KAAK;AAEzC,OAAK,MAAM,OAAO,MAAM;GACtB,MAAM,QAAQ,MAAM,KAAK,CAAC,GAAG,MAAM,KAAK,CAAC;AACzC,OAAI,MAAO,QAAO;;AAGpB,WAAS,OAAO,KAAK;AACrB,UAAQ,IAAI,KAAK;AACjB,SAAO,KAAK,SAAS;AACrB,SAAO;;AAGT,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,QAAQ,MAAM,SAAS,MAAM,EAAE,CAAC;AACtC,MAAI,MAAO,4BAAW,MAAM;;AAG9B,2BAAU,OAAO;;;;;;AAOnB,MAAM,kCAAkC,eAAwD;CAC9F,MAAM,OAAO,IAAI,KAAa;CAE9B,MAAM,WAAW,SAA2C;AAC1D,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,UAAK,IAAI,IAAI,KAAK;AAClB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO;;;;;AAMT,MAAM,cAAc,QAAqB,SAA0B,OAAO,MAAM,IAAI,KAAK;;;;;;;AAkDzF,MAAa,0BACX,QACA,YACgD;CAChD,MAAM,SAASC,oCAAkB,QAAQ,eAAe;CAGxD,MAAM,aAAa,0BAA0B,OAAO,UAAU;AAC9D,KAAI,WAAW,OAAO,EAAE;AACtB,6BAAW,WAAW,MAAM;;CAE9B,MAAM,kBAAkB,WAAW;CAGnC,MAAM,4BAA4B,IAAI,KAA0C;CAChF,MAAMC,YAAgC,EAAE;AAExC,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,SAAS,kBAAkB,MAAM,QAAQ,0BAA0B;AACzE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,4BAA0B,IAAI,KAAK,MAAM,OAAO,MAAM,UAAU;AAChE,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAMC,aAAkC,EAAE;AAC1C,MAAK,MAAM,MAAM,OAAO,YAAY;EAClC,MAAM,SAAS,mBAAmB,IAAI,OAAO;AAC7C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,OAAO,MAAM;;AAG/B,2BAAU;EAAE;EAAY;EAAW,CAAC;;;;;AAMtC,MAAM,sBAAsB,IAAqB,WAAuE;CAEtH,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,GAAG,WAAW;EAC5B,MAAM,WAAW,gBAAgB,QAAQ,EAAE,SAAS;AACpD,MAAI,aAAa,MAAM;AACrB,8BAAW;IACT,MAAM;IACN,SAAS,iBAAiB,EAAE,SAAS,iBAAiB,EAAE,KAAK;IAC7D,UAAU,EAAE;IACb,CAAC;;AAEJ,YAAU,KAAK;GAAE,GAAG;GAAG;GAAU,CAAC;;CAIpC,MAAM,uBAAuB,4BAA4B,GAAG,WAAW;AAEvE,2BAAU;EACR,GAAG;EACH;EACA;EACD,CAAC;;;;;;AAOJ,MAAM,qBACJ,MACA,QACA,8BACiD;CAEjD,MAAM,uBAAuB,4BAA4B,KAAK,WAAW;CAGzE,MAAM,qBAAqB,sBAAsB,KAAK,YAAY,KAAK,QAAQ,OAAO;AACtF,KAAI,mBAAmB,OAAO,EAAE;AAC9B,6BAAW,mBAAmB,MAAM;;CAEtC,MAAM,eAAe,mBAAmB;CAGxC,MAAMC,kBAAsC,EAAE;AAC9C,MAAK,MAAM,WAAW,sBAAsB;EAC1C,MAAM,eAAe,0BAA0B,IAAI,QAAQ;AAC3D,MAAI,cAAc;AAChB,mBAAgB,KAAK,GAAG,aAAa;;;CAQzC,MAAMC,YAA6B,CACjC,GAAG,cACH,GAAG,gBAAgB,KAAK,OAAO;EAC7B,MAAM,EAAE;EACR,UAAU,EAAE;EACZ,kBAAkB,EAAE;EACpB,iBAAiB,EAAE;EACnB,UAAU,EAAE;EACb,EAAE,CACJ;CAGD,MAAM,kBAAkB,yBAAyB,UAAU;AAC3D,KAAI,gBAAgB,OAAO,EAAE;AAC3B,6BAAW,gBAAgB,MAAM;;AAGnC,2BAAU;EACR,GAAG;EACH;EACA,WAAW,gBAAgB;EAC5B,CAAC;;;;;AAMJ,MAAM,mBAAmB,QAAqB,aAAyD;AACrG,KAAI,aAAa,QAAQ,SAAS,EAAE;AAClC,SAAO;;AAET,KAAI,WAAW,QAAQ,SAAS,EAAE;AAChC,SAAO;;AAET,KAAI,OAAO,OAAO,IAAI,SAAS,EAAE;AAC/B,SAAO;;AAET,QAAO;;;;;AAMT,MAAM,+BAA+B,eAAgF;CACnH,MAAM,YAAY,IAAI,KAAa;CAEnC,MAAM,WAAW,SAA6D;AAC5E,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,eAAU,IAAI,IAAI,KAAK;AACvB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO,CAAC,GAAG,UAAU;;;;;;;;ACzwBvB,MAAM,mBAAmB,SAAwD;AAC/E,SAAQ,MAAR;EACE,KAAK,QACH,QAAO;EACT,KAAK,WACH,QAAO;EACT,KAAK,eACH,QAAO;;;;;;AAOb,MAAa,iBAAiB,WAA8B,YAA6D;CACvH,MAAMC,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CAKpF,MAAM,aAAa,GAAG,UAAU,KAAK;CACrC,MAAM,gBAAgB,UAAU;AAEhC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,cAAc,cAAc;AACpG,OAAM,KAAK,KAAK,cAAc,WAAW;AACzC,OAAM,KAAK,aAAa,KAAK,UAAU,UAAU,KAAK,CAAC,GAAG;AAG1D,KAAI,UAAU,UAAU,SAAS,GAAG;AAClC,QAAM,KAAK,oBAAoB,cAAc,UAAU,UAAU,CAAC,KAAK;;CAIzE,MAAM,eAAe,gBAAgB,UAAU,KAAK;AACpD,OAAM,KAAK,+BAA+B;CAC1C,MAAM,mBAAmB,eAAe,UAAU,YAAY,GAAG,UAAU,WAAW,QAAQ,aAAa;AAC3G,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAa,gBAAgB,UAA4B,YAA6D;CACpH,MAAMD,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CACpF,MAAM,eAAe,SAAS,UAAU,SAAS;CAKjD,MAAM,aAAa,GAAG,SAAS,KAAK;CAGpC,MAAM,cAAc,eAAe,mBAAmB;AACtD,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,YAAY,QAAQ;AAC5F,OAAM,KAAK,cAAc,SAAS,OAAO,IAAI;AAG7C,KAAI,cAAc;AAChB,QAAM,KAAK,oBAAoB,cAAc,SAAS,UAAU,CAAC,KAAK;;CAKxE,MAAM,gBAAgB,eAAe,aAAa;AAClD,OAAM,KAAK,gBAAgB,cAAc,SAAS;CAClD,MAAM,mBAAmB,eAAe,SAAS,YAAY,GAAG,SAAS,WAAW,QAAQ,SAAS,OAAO;AAC5G,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAe7B,MAAM,iBAAiB,cAAoD;AACzE,QAAO,UAAU,KAAK,MAAM,WAAW,KAAK,UAAU,EAAE,KAAK,CAAC,IAAI,EAAE,SAAS,GAAG,KAAK,UAAU,EAAE,SAAS,CAAC,GAAG,CAAC,KAAK,KAAK;;;;;;AAO3H,MAAM,kBACJ,YACA,QACA,WACA,QACA,mBACuC;CACvC,MAAM,gBAAgB,IAAI,IAAI,UAAU,KAAK,MAAM,EAAE,KAAK,CAAC;AAC3D,QAAO,uBAAuB,YAAY,QAAQ,eAAe,QAAQ,eAAe;;;;;;AAO1F,MAAM,0BACJ,YACA,QACA,eACA,QACA,mBACuC;CACvC,MAAMD,QAAkB,EAAE;CAG1B,MAAME,kBAA0C,EAAE;CAClD,MAAMC,kBAAqC,EAAE;AAE7C,MAAK,MAAM,OAAO,YAAY;AAC5B,MAAI,IAAI,SAAS,kBAAkB;AACjC,mBAAgB,KAAK,IAAI;SACpB;AACL,mBAAgB,KAAK,IAAI;;;AAK7B,MAAK,MAAM,OAAO,iBAAiB;EACjC,MAAM,SAAS,oBAAoB,KAAK,QAAQ,eAAe,QAAQ,eAAe;AACtF,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,QAAM,KAAK,OAAO,MAAM;;AAI1B,KAAI,gBAAgB,SAAS,GAAG;EAC9B,MAAM,cAAc,2BAA2B,iBAAiB,QAAQ,eAAe,OAAO;AAC9F,MAAI,YAAY,OAAO,EAAE;AACvB,8BAAW,YAAY,MAAM;;AAE/B,QAAM,KAAK,YAAY,MAAM;;AAG/B,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,uBACJ,KACA,QACA,eACA,QACA,mBACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAEnC,SAAQ,IAAI,MAAZ;EACE,KAAK,QACH,QAAO,mBAAmB,KAAK,QAAQ,eAAe,QAAQ,eAAe;EAC/E,KAAK,iBACH,2BAAU,GAAG,QAAQ,KAAK,IAAI,KAAK,oBAAoB;EACzD,KAAK,iBAEH,2BAAU,GAAG;;;;;;;AAQnB,MAAM,8BACJ,iBACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAGnC,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,KAAK,WAAW,IAAI;AACtB,8BAAW;IACT,MAAM;IACN,SAAS;IACV,CAAC;;;AAKN,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,UAAU,CAAC,OAAO,QAAQ,IAAI,KAAK,OAAO,EAAE;GAG9C,IAAI,gBAAgB;AACpB,QAAK,MAAM,GAAG,aAAa,OAAO,QAAQ;AACxC,QAAI,SAAS,QAAQ,IAAI,KAAK,OAAO,EAAE;AACrC,qBAAgB;AAChB;;;AAGJ,OAAI,CAAC,eAAe;AAClB,+BAAW;KACT,MAAM;KACN,SAAS,uCAAuC,KAAK,OAAO;KAC5D,QAAQ,KAAK;KACd,CAAC;;;;CAMR,MAAMC,UAAoB,EAAE;AAC5B,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,eAAe,KAAK,OAAO,SAAS,EAAE;EAE5C,MAAM,eAAe,uBAAuB,KAAK,YAAY,SAAS,GAAG,eAAe,QAAQ,KAAK,OAAO;AAC5G,MAAI,aAAa,OAAO,EAAE;AACxB,8BAAW,aAAa,MAAM;;AAGhC,UAAQ,KAAK,GAAG,eAAe,KAAK,OAAO;EAC7C,aAAa,MAAM;EACnB,aAAa,KAAK;;AAKlB,2BAAU,GAAG,QAAQ;EACrB,QAAQ,KAAK,KAAK,CAAC;EACnB,QAAQ,KAAK;;;;;AAMf,MAAM,sBACJ,OACA,QACA,eACA,QACA,mBACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;CAGnC,MAAM,OAAO,MAAM;CACnB,MAAM,aAAa,MAAM;CACzB,MAAM,UAAU,QAAQ,KAAK,SAAS;CACtC,MAAM,gBAAgB,cAAc,WAAW,SAAS;AAGxD,KAAI,CAAC,WAAW,CAAC,eAAe;AAC9B,4BAAU,GAAG,UAAU,MAAM,KAAK,SAAS;;CAG7C,IAAI,OAAO,GAAG,QAAQ,OAAO,MAAM,KAAK;AAExC,KAAI,SAAS;EACX,MAAM,aAAa,cAAc,MAAM,eAAe,QAAQ,gBAAgB,MAAM,KAAK;AACzF,MAAI,WAAW,OAAO,EAAE;AACtB,8BAAW,WAAW,MAAM;;AAE9B,UAAQ,WAAW;;AAGrB,SAAQ;AAER,KAAI,eAAe;EAEjB,MAAM,qBAAqB,WAAW,MAAM,MAAM,EAAE,SAAS,iBAAiB;EAG9E,MAAM,mBACJ,UAAU,iBAAkB,mBAAmB,QAAQ,gBAAgB,MAAM,KAAK,IAAI,YAAa;AAErG,MAAI,oBAAoB;GAEtB,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,QAAQ,iBAAiB;AAC5G,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ;AACR,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;SACd;AAEL,WAAQ;GACR,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,QAAQ,iBAAiB;AAC5G,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;;;AAIvB,SAAQ;AAER,2BAAU,KAAK;;;;;AAUjB,MAAM,kBAAkB,aAA8B;AACpD,QAAO,SAAS,SAAS,KAAK;;;;;;;;;;AAWhC,MAAM,qBAAqB,OAAoB,qBAAkD;AAE/F,KAAI,CAAC,iBAAkB,QAAO;AAG9B,KAAI,CAAC,eAAe,iBAAiB,CAAE,QAAO;AAG9C,KAAI,MAAM,SAAS,WAAY,QAAO;AAGtC,KAAI,MAAM,SAAS,OAAQ,QAAO;AAGlC,KAAI,MAAM,SAAS,OAAQ,QAAO;AAElC,QAAO;;;;;;;;AAST,MAAM,sBAAsB,iBAA4C;CACtE,MAAM,EAAE,UAAU,aAAa;CAG/B,MAAM,YAAY,SAAS,MAAM,oBAAoB;AACrD,KAAI,CAAC,aAAa,CAAC,UAAU,GAAI,QAAO;AACxC,QAAO;EAAE;EAAU,UAAU,UAAU;EAAI;;;;;AAM7C,MAAM,qBACJ,OACA,cACA,eACA,WACuC;CAEvC,MAAM,eAAe,kBAAkB,OAAO,cAAc,SAAS;AAGrE,KAAI,MAAM,SAAS,YAAY,gBAAgB,QAAQ;AACrD,SAAO,mBAAmB,OAAO,aAAa,UAAU,eAAe,QAAQ,aAAa;;AAI9F,KAAI,MAAM,SAAS,UAAU,gBAAgB,QAAQ;EACnD,MAAM,cAAc,mBAAmB,aAAa;AACpD,MAAI,aAAa;GACf,MAAMC,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAMC,WAAS,kBAAkB,GAAG,aAAa,eAAe,OAAO;AACvE,QAAIA,SAAO,OAAO,CAAE,QAAOA;AAC3B,WAAO,KAAKA,SAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;;CAKvC,MAAM,SAAS,UAAU,OAAO,cAAc;AAC9C,KAAI,OAAO,OAAO,CAAE,QAAO;AAG3B,KAAI,cAAc;AAChB,4BAAU,IAAI,OAAO,MAAM,GAAG;;AAGhC,QAAO;;;;;AAMT,MAAM,sBACJ,OACA,eACA,eACA,QACA,gBACuC;AACvC,KAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,4BAAU,cAAc,SAAS,KAAK;;CAGxC,MAAMF,UAAoB,EAAE;AAC5B,MAAK,MAAM,KAAK,MAAM,QAAQ;EAE5B,MAAM,YAAY,kBAAkB,QAAQ,eAAe,EAAE,KAAK;AAClE,MAAI,cAAc,MAAM;AACtB,8BAAW;IACT,MAAM;IACN,SAAS,kBAAkB,EAAE,KAAK,mBAAmB,cAAc;IACnE,UAAU;IACV,WAAW,EAAE;IACd,CAAC;;EAGJ,MAAM,SAAS,kBAAkB,EAAE,OAAO,WAAW,eAAe,OAAO;AAC3E,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,UAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;CAG5C,MAAM,YAAY,KAAK,QAAQ,KAAK,KAAK,CAAC;AAC1C,2BAAU,cAAc,IAAI,UAAU,KAAK,UAAU;;;;;AAUvD,MAAM,iBACJ,MACA,eACA,QACA,gBACA,cACuC;AACvC,KAAI,KAAK,WAAW,GAAG;AACrB,4BAAU,GAAG;;CAGf,MAAMG,aAAuB,EAAE;AAC/B,MAAK,MAAM,OAAO,MAAM;EAEtB,MAAM,eACJ,UAAU,kBAAkB,YAAY,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK,GAAG;EAEzG,MAAM,SAAS,kBAAkB,IAAI,OAAO,cAAc,eAAe,OAAO;AAChF,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,GAAG,IAAI,KAAK,IAAI,OAAO,QAAQ;;AAEjD,2BAAU,KAAK,WAAW,KAAK,KAAK,CAAC,IAAI;;;;;AAM3C,MAAM,aAAa,OAAoB,kBAAmE;AACxG,SAAQ,MAAM,MAAd;EACE,KAAK;AAEH,OAAI,cAAc,IAAI,MAAM,KAAK,EAAE;AACjC,8BAAU,KAAK,MAAM,OAAO;;AAE9B,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,MAAM,KAAK;IAClC,cAAc,MAAM;IACrB,CAAC;EACJ,KAAK;EACL,KAAK,QACH,2BAAU,MAAM,MAAM;EACxB,KAAK,SACH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,UACH,2BAAU,MAAM,QAAQ,SAAS,QAAQ;EAC3C,KAAK,OACH,2BAAU,OAAO;EACnB,KAAK,OAEH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,QAAQ;GACX,MAAMF,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,GAAG,cAAc;AAC1C,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,WAAO,KAAK,OAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;EAErC,KAAK,UAAU;AACb,OAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,8BAAU,KAAK;;GAEjB,MAAMD,UAAoB,EAAE;AAC5B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,EAAE,OAAO,cAAc;AAChD,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,YAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;AAE5C,6BAAU,KAAK,QAAQ,KAAK,KAAK,CAAC,IAAI;;;;;;;AC5jB5C,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,8BAAe,WAAW,EAAE;AAC1B,8BAA+B;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaI,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,mCAAoB,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,4BAAY;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,6BAAU,EAAE,SAAS,CAAC;WACf,OAAO;AACd,8BAAW;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,mBAAmB,YAAoB;CAClD,MAAM,oCAAqB,QAAQ;AACnC,KAAI;AACF,sBAAO,YAAY;GAAE,WAAW;GAAM,OAAO;GAAM,CAAC;AACpD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,SAAS;AACnC,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACtBN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,sCAAuB,WAAW;AAExC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAuC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,yCAA4B,cAAc,OAAO;EACvD,MAAM,8BAAiB,aAAa;AACpC,4BAAsC,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAuC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAuC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,gCAAmB,UAAU;AACnC,2BAAsC,OAAO;;AAG/C,MAAa,cAAc,yCAA8C,SAAS,CAAC,0BAAa,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC9CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,iCAAkB,SAAS;CACjC,MAAM,qCAAsB,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,mCAAoB,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,6BAAc,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,sCAAqBD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,oCAAqB,YAAY,UAAU,2BAAY,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,oCAAqB,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,iCAAkB,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,oCAAqB,aAAa,OAAO,QAAQ;AACvD,MAAI,yBAAY,WAAW,EAAE;AAC3B,8BAAW;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,qCAAsB,aAAa,OAAO,QAAQ;AACxD,OAAI,yBAAY,YAAY,EAAE;AAC5B,+BAAW;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,2BAAW,IAAI,CAAC,GAChC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,oBAAoB,IAAI,KAA+B;AAC7D,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,YAAY;AAC3B,qBAAkB,IAAI,YAAY,aAAa,WAAW;;;CAK9D,MAAM,EACJ,MAAM,cACN,aACA,iBACEC,4CAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACA,aAAa,kBAAkB,OAAO,IAAI,oBAAoB;EAC/D,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,2CAAM,8BAAuB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,yDAA2B,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,8BAAW,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,gCAAiB,QAAQ;EAG/B,MAAM,8BAAe,QAAQ,QAAQ;AACrC,8BAAe,QAAQ,EAAE;GACvB,MAAM,eAAe,gBAAgB,QAAQ;AAC7C,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;;EAOlC,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,+BAAgB,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,+BAAW,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,0DAA4B,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,6BAAW,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,8BAAc,OAAO,GACrB,8BAAc,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,6BAAW,aAAa,MAAM;;AAGhC,2BAAU;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
1
+ {"version":3,"file":"index.cjs","names":["operations: ParsedOperation[]","fragments: ParsedFragment[]","Kind","variables: ParsedVariable[]","levels: TypeLevel[]","lists: (\"[]!\" | \"[]?\")[]","match: RegExpExecArray | null","mergedInner: \"!\" | \"?\"","mergedLists: (\"[]!\" | \"[]?\")[]","usages: VariableUsage[]","error","variables: InferredVariable[]","sorted: ParsedFragment[]","createSchemaIndex","fragments: EnrichedFragment[]","operations: EnrichedOperation[]","variables: EnrichedVariable[]","spreadVariables: InferredVariable[]","allUsages: VariableUsage[]","lines: string[]","createSchemaIndex","inlineFragments: ParsedInlineFragment[]","otherSelections: ParsedSelection[]","entries: string[]","values: string[]","result","argEntries: string[]","esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","generateMultiSchemaModule","defsPaths: string[]","defaultBundler"],"sources":["../src/graphql-compat/parser.ts","../src/graphql-compat/transformer.ts","../src/graphql-compat/emitter.ts","../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["/**\n * Parser for .graphql operation files.\n * Extracts operations and fragments from GraphQL documents.\n * @module\n */\n\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport {\n type ArgumentNode,\n type DocumentNode,\n type FieldNode,\n type FragmentDefinitionNode,\n type FragmentSpreadNode,\n type InlineFragmentNode,\n Kind,\n type OperationDefinitionNode,\n parse,\n type SelectionNode,\n type TypeNode,\n type ValueNode,\n type VariableDefinitionNode,\n} from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\n\nimport type {\n GraphqlCompatError,\n ParsedArgument,\n ParsedFieldSelection,\n ParsedFragment,\n ParsedFragmentSpread,\n ParsedInlineFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n TypeInfo,\n} from \"./types\";\n\n/**\n * Parse a single .graphql file and extract operations and fragments.\n */\nexport const parseGraphqlFile = (filePath: string): Result<ParseResult, GraphqlCompatError> => {\n const resolvedPath = resolve(filePath);\n\n if (!existsSync(resolvedPath)) {\n return err({\n code: \"GRAPHQL_FILE_NOT_FOUND\",\n message: `GraphQL file not found at ${resolvedPath}`,\n filePath: resolvedPath,\n });\n }\n\n try {\n const source = readFileSync(resolvedPath, \"utf8\");\n const document = parse(source);\n return ok(extractFromDocument(document, resolvedPath));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: resolvedPath,\n });\n }\n};\n\n/**\n * Parse GraphQL source string directly.\n */\nexport const parseGraphqlSource = (source: string, sourceFile: string): Result<ParseResult, GraphqlCompatError> => {\n try {\n const document = parse(source);\n return ok(extractFromDocument(document, sourceFile));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: sourceFile,\n });\n }\n};\n\n/**\n * Extract operations and fragments from a parsed GraphQL document.\n */\nconst extractFromDocument = (document: DocumentNode, sourceFile: string): ParseResult => {\n const operations: ParsedOperation[] = [];\n const fragments: ParsedFragment[] = [];\n\n for (const definition of document.definitions) {\n if (definition.kind === Kind.OPERATION_DEFINITION) {\n const operation = extractOperation(definition, sourceFile);\n if (operation) {\n operations.push(operation);\n }\n } else if (definition.kind === Kind.FRAGMENT_DEFINITION) {\n fragments.push(extractFragment(definition, sourceFile));\n }\n }\n\n return { operations, fragments };\n};\n\n/**\n * Extract a single operation from an OperationDefinitionNode.\n */\nconst extractOperation = (node: OperationDefinitionNode, sourceFile: string): ParsedOperation | null => {\n // Anonymous operations are not supported\n if (!node.name) {\n return null;\n }\n\n const variables: ParsedVariable[] = (node.variableDefinitions ?? []).map(extractVariable);\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n kind: node.operation,\n name: node.name.value,\n variables,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a fragment from a FragmentDefinitionNode.\n */\nconst extractFragment = (node: FragmentDefinitionNode, sourceFile: string): ParsedFragment => {\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n name: node.name.value,\n onType: node.typeCondition.name.value,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a variable definition.\n */\nconst extractVariable = (node: VariableDefinitionNode): ParsedVariable => {\n const { typeName, modifier } = parseTypeNode(node.type);\n const defaultValue = node.defaultValue ? extractValue(node.defaultValue) : undefined;\n\n return {\n name: node.variable.name.value,\n typeName,\n modifier,\n // Type kind will be determined later when we have schema context\n typeKind: \"scalar\",\n defaultValue,\n };\n};\n\n/**\n * Parse a GraphQL TypeNode into type name and modifier.\n *\n * Format: inner nullability + list modifiers\n * - Inner: `!` (non-null) or `?` (nullable)\n * - List: `[]!` (non-null list) or `[]?` (nullable list)\n */\nexport const parseTypeNode = (node: TypeNode): TypeInfo => {\n type TypeLevel = { kind: \"list\" | \"named\"; nonNull: boolean };\n const levels: TypeLevel[] = [];\n\n const collect = (n: TypeNode, nonNull: boolean): string => {\n if (n.kind === Kind.NON_NULL_TYPE) {\n return collect(n.type, true);\n }\n if (n.kind === Kind.LIST_TYPE) {\n levels.push({ kind: \"list\", nonNull });\n return collect(n.type, false);\n }\n levels.push({ kind: \"named\", nonNull });\n return n.name.value;\n };\n\n const typeName = collect(node, false);\n\n // Build modifier from levels (reverse order)\n let modifier = \"?\";\n for (const level of levels.slice().reverse()) {\n if (level.kind === \"named\") {\n modifier = level.nonNull ? \"!\" : \"?\";\n continue;\n }\n const listSuffix = level.nonNull ? \"[]!\" : \"[]?\";\n modifier = `${modifier}${listSuffix}`;\n }\n\n return { typeName, modifier };\n};\n\n/**\n * Extract selections from a SelectionSet.\n */\nconst extractSelections = (selections: readonly SelectionNode[]): ParsedSelection[] => {\n return selections.map(extractSelection);\n};\n\n/**\n * Extract a single selection.\n */\nconst extractSelection = (node: SelectionNode): ParsedSelection => {\n switch (node.kind) {\n case Kind.FIELD:\n return extractFieldSelection(node);\n case Kind.FRAGMENT_SPREAD:\n return extractFragmentSpread(node);\n case Kind.INLINE_FRAGMENT:\n return extractInlineFragment(node);\n }\n};\n\n/**\n * Extract a field selection.\n */\nconst extractFieldSelection = (node: FieldNode): ParsedFieldSelection => {\n const args = node.arguments?.length ? node.arguments.map(extractArgument) : undefined;\n const selections = node.selectionSet ? extractSelections(node.selectionSet.selections) : undefined;\n\n return {\n kind: \"field\",\n name: node.name.value,\n alias: node.alias?.value,\n arguments: args,\n selections,\n };\n};\n\n/**\n * Extract a fragment spread.\n */\nconst extractFragmentSpread = (node: FragmentSpreadNode): ParsedFragmentSpread => {\n return {\n kind: \"fragmentSpread\",\n name: node.name.value,\n };\n};\n\n/**\n * Extract an inline fragment.\n */\nconst extractInlineFragment = (node: InlineFragmentNode): ParsedInlineFragment => {\n return {\n kind: \"inlineFragment\",\n onType: node.typeCondition?.name.value ?? \"\",\n selections: extractSelections(node.selectionSet.selections),\n };\n};\n\n/**\n * Extract an argument.\n */\nconst extractArgument = (node: ArgumentNode): ParsedArgument => {\n return {\n name: node.name.value,\n value: extractValue(node.value),\n };\n};\n\n/**\n * Assert unreachable code path (for exhaustiveness checks).\n */\nconst assertUnreachable = (value: never): never => {\n throw new Error(`Unexpected value: ${JSON.stringify(value)}`);\n};\n\n/**\n * Extract a value (literal or variable reference).\n */\nconst extractValue = (node: ValueNode): ParsedValue => {\n switch (node.kind) {\n case Kind.VARIABLE:\n return { kind: \"variable\", name: node.name.value };\n case Kind.INT:\n return { kind: \"int\", value: node.value };\n case Kind.FLOAT:\n return { kind: \"float\", value: node.value };\n case Kind.STRING:\n return { kind: \"string\", value: node.value };\n case Kind.BOOLEAN:\n return { kind: \"boolean\", value: node.value };\n case Kind.NULL:\n return { kind: \"null\" };\n case Kind.ENUM:\n return { kind: \"enum\", value: node.value };\n case Kind.LIST:\n return { kind: \"list\", values: node.values.map(extractValue) };\n case Kind.OBJECT:\n return {\n kind: \"object\",\n fields: node.fields.map((field) => ({\n name: field.name.value,\n value: extractValue(field.value),\n })),\n };\n default:\n return assertUnreachable(node);\n }\n};\n","/**\n * Transformer for enriching parsed GraphQL operations with schema information.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport { parseTypeNode } from \"./parser\";\nimport type {\n GraphqlCompatError,\n InferredVariable,\n ParsedArgument,\n ParsedFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n TypeInfo,\n} from \"./types\";\n\n/**\n * Schema index type extracted from generator.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Built-in GraphQL scalar types.\n */\nconst builtinScalarTypes = new Set([\"ID\", \"String\", \"Int\", \"Float\", \"Boolean\"]);\n\n// ============================================================================\n// Modifier Merging Utilities\n// ============================================================================\n\n/**\n * Parsed structure of a modifier for comparison and merging.\n * Example: \"![]?\" -> { inner: \"!\", lists: [\"[]?\"] }\n */\ntype ModifierStructure = {\n readonly inner: \"!\" | \"?\";\n readonly lists: readonly (\"[]!\" | \"[]?\")[];\n};\n\n/**\n * Parse a modifier string into its structural components.\n * @param modifier - Modifier string like \"!\", \"?\", \"![]!\", \"?[]?[]!\"\n * @returns Parsed structure with inner nullability and list modifiers\n */\nconst parseModifierStructure = (modifier: string): ModifierStructure => {\n // Extract inner nullability (first character)\n const inner = modifier[0] === \"!\" ? \"!\" : \"?\";\n\n // Extract list modifiers ([]! or []?)\n const lists: (\"[]!\" | \"[]?\")[] = [];\n const listPattern = /\\[\\]([!?])/g;\n let match: RegExpExecArray | null;\n while ((match = listPattern.exec(modifier)) !== null) {\n lists.push(`[]${match[1]}` as \"[]!\" | \"[]?\");\n }\n\n return { inner, lists };\n};\n\n/**\n * Rebuild modifier string from structure.\n */\nconst buildModifier = (structure: ModifierStructure): string => {\n return structure.inner + structure.lists.join(\"\");\n};\n\n/**\n * Check if source modifier can be assigned to target modifier.\n * Implements GraphQL List Coercion: depth difference of 0 or 1 is allowed.\n *\n * Rules:\n * - A single value can be coerced into a list (depth diff = 1)\n * - At each level, non-null can be assigned to nullable (but not vice versa)\n *\n * @param source - The modifier of the value being assigned (variable's type)\n * @param target - The modifier expected by the position (field argument's type)\n * @returns true if assignment is valid\n */\nexport const isModifierAssignable = (source: string, target: string): boolean => {\n const srcStruct = parseModifierStructure(source);\n const tgtStruct = parseModifierStructure(target);\n\n const depthDiff = tgtStruct.lists.length - srcStruct.lists.length;\n\n // Depth difference must be 0 or 1 (List Coercion only wraps one level)\n if (depthDiff < 0 || depthDiff > 1) return false;\n\n // When List Coercion applies (depth diff = 1), compare inner levels only\n // The outer list level of target is satisfied by the coercion itself\n const tgtListsToCompare = depthDiff === 1 ? tgtStruct.lists.slice(1) : tgtStruct.lists;\n\n // When coercing a nullable scalar to a list, target's outer list must be nullable\n // A null scalar would produce null (not [null]), violating non-null list constraint\n // This only applies to scalars (no lists), not when coercing list to nested list\n if (depthDiff === 1 && srcStruct.lists.length === 0 && srcStruct.inner === \"?\" && tgtStruct.lists[0] === \"[]!\") {\n return false;\n }\n\n // Inner nullability: non-null can go to nullable, but not vice versa\n if (srcStruct.inner === \"?\" && tgtStruct.inner === \"!\") return false;\n\n // List level nullability: check each corresponding level\n for (let i = 0; i < srcStruct.lists.length; i++) {\n const srcList = srcStruct.lists[i]!;\n const tgtList = tgtListsToCompare[i]!;\n if (srcList === \"[]?\" && tgtList === \"[]!\") return false;\n }\n\n return true;\n};\n\n/**\n * Derive minimum modifier needed to satisfy expected modifier.\n * When List Coercion can apply, returns one level shallower.\n *\n * @param expectedModifier - The modifier expected by the field argument\n * @returns The minimum modifier the variable must have\n */\nconst deriveMinimumModifier = (expectedModifier: string): string => {\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n return buildModifier({ inner: struct.inner, lists: struct.lists.slice(1) });\n }\n return expectedModifier;\n};\n\n/**\n * Merge two modifiers by taking the stricter constraint at each level.\n * - Non-null (!) is stricter than nullable (?)\n * - List depths must match\n *\n * @param a - First modifier\n * @param b - Second modifier\n * @returns Merged modifier or error if incompatible\n */\nexport const mergeModifiers = (a: string, b: string): { ok: true; value: string } | { ok: false; reason: string } => {\n const structA = parseModifierStructure(a);\n const structB = parseModifierStructure(b);\n\n // List depths must match\n if (structA.lists.length !== structB.lists.length) {\n return {\n ok: false,\n reason: `Incompatible list depths: \"${a}\" has ${structA.lists.length} list level(s), \"${b}\" has ${structB.lists.length}`,\n };\n }\n\n // Take stricter inner constraint (! beats ?)\n const mergedInner: \"!\" | \"?\" = structA.inner === \"!\" || structB.inner === \"!\" ? \"!\" : \"?\";\n\n // Merge each list level (! beats ?)\n const mergedLists: (\"[]!\" | \"[]?\")[] = [];\n for (let i = 0; i < structA.lists.length; i++) {\n const listA = structA.lists[i]!;\n const listB = structB.lists[i]!;\n mergedLists.push(listA === \"[]!\" || listB === \"[]!\" ? \"[]!\" : \"[]?\");\n }\n\n return { ok: true, value: buildModifier({ inner: mergedInner, lists: mergedLists }) };\n};\n\n// ============================================================================\n// Variable Collection from Selections\n// ============================================================================\n\n/**\n * Intermediate type for tracking variable usages before merging.\n */\ntype VariableUsage = {\n readonly name: string;\n readonly typeName: string;\n /** The modifier expected by the field/argument position */\n readonly expectedModifier: string;\n /** The minimum modifier the variable needs (after applying List Coercion) */\n readonly minimumModifier: string;\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Get the expected type for a field argument from the schema.\n * Returns null if the field or argument is not found.\n */\nexport const getArgumentType = (\n schema: SchemaIndex,\n parentTypeName: string,\n fieldName: string,\n argumentName: string,\n): TypeInfo | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const argDef = fieldDef.arguments?.find((arg) => arg.name.value === argumentName);\n if (!argDef) return null;\n\n return parseTypeNode(argDef.type);\n};\n\n/**\n * Get the expected type for an input object field from the schema.\n */\nexport const getInputFieldType = (schema: SchemaIndex, inputTypeName: string, fieldName: string): TypeInfo | null => {\n const inputRecord = schema.inputs.get(inputTypeName);\n if (!inputRecord) return null;\n\n const fieldDef = inputRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n return parseTypeNode(fieldDef.type);\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKindFromName = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) return \"scalar\";\n if (isEnumName(schema, typeName)) return \"enum\";\n if (schema.inputs.has(typeName)) return \"input\";\n return null;\n};\n\n/**\n * Extract variable usages from a parsed value, given the expected type.\n * Handles nested input objects recursively.\n */\nconst collectVariablesFromValue = (\n value: ParsedValue,\n expectedTypeName: string,\n expectedModifier: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n if (value.kind === \"variable\") {\n const typeKind = resolveTypeKindFromName(schema, expectedTypeName);\n if (!typeKind) {\n return {\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${expectedTypeName}\" for variable \"$${value.name}\"`,\n typeName: expectedTypeName,\n };\n }\n usages.push({\n name: value.name,\n typeName: expectedTypeName,\n expectedModifier,\n minimumModifier: deriveMinimumModifier(expectedModifier),\n typeKind,\n });\n return null;\n }\n\n if (value.kind === \"object\") {\n // For object values, check each field against input type definition\n for (const field of value.fields) {\n const fieldType = getInputFieldType(schema, expectedTypeName, field.name);\n if (!fieldType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${field.name}\" on input type \"${expectedTypeName}\"`,\n typeName: expectedTypeName,\n fieldName: field.name,\n };\n }\n const error = collectVariablesFromValue(field.value, fieldType.typeName, fieldType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n }\n\n if (value.kind === \"list\") {\n // For list values, unwrap one level of list modifier\n // e.g., [ID!]! -> ID! for elements\n const struct = parseModifierStructure(expectedModifier);\n if (struct.lists.length > 0) {\n const innerModifier = buildModifier({\n inner: struct.inner,\n lists: struct.lists.slice(1),\n });\n for (const item of value.values) {\n const error = collectVariablesFromValue(item, expectedTypeName, innerModifier, schema, usages);\n if (error) return error;\n }\n }\n }\n\n // Other value kinds (int, float, string, etc.) don't contain variables\n return null;\n};\n\n/**\n * Collect variable usages from field arguments.\n */\nconst collectVariablesFromArguments = (\n args: readonly ParsedArgument[],\n parentTypeName: string,\n fieldName: string,\n schema: SchemaIndex,\n usages: VariableUsage[],\n): GraphqlCompatError | null => {\n for (const arg of args) {\n const argType = getArgumentType(schema, parentTypeName, fieldName, arg.name);\n if (!argType) {\n return {\n code: \"GRAPHQL_UNKNOWN_ARGUMENT\",\n message: `Unknown argument \"${arg.name}\" on field \"${fieldName}\"`,\n fieldName,\n argumentName: arg.name,\n };\n }\n const error = collectVariablesFromValue(arg.value, argType.typeName, argType.modifier, schema, usages);\n if (error) return error;\n }\n return null;\n};\n\n/**\n * Recursively collect all variable usages from selections.\n */\nexport const collectVariableUsages = (\n selections: readonly ParsedSelection[],\n parentTypeName: string,\n schema: SchemaIndex,\n): Result<VariableUsage[], GraphqlCompatError> => {\n const usages: VariableUsage[] = [];\n\n const collect = (sels: readonly ParsedSelection[], parentType: string): GraphqlCompatError | null => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"field\": {\n // Collect from arguments\n if (sel.arguments && sel.arguments.length > 0) {\n const error = collectVariablesFromArguments(sel.arguments, parentType, sel.name, schema, usages);\n if (error) return error;\n }\n\n // Recurse into nested selections\n if (sel.selections && sel.selections.length > 0) {\n // Need to determine the field's return type for nested selections\n const fieldReturnType = getFieldReturnType(schema, parentType, sel.name);\n if (!fieldReturnType) {\n return {\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${sel.name}\" on type \"${parentType}\"`,\n typeName: parentType,\n fieldName: sel.name,\n };\n }\n const error = collect(sel.selections, fieldReturnType);\n if (error) return error;\n }\n break;\n }\n case \"inlineFragment\": {\n // Use the inline fragment's type condition\n const error = collect(sel.selections, sel.onType);\n if (error) return error;\n break;\n }\n case \"fragmentSpread\":\n // Fragment spreads are handled separately (variables from spread fragments)\n break;\n }\n }\n return null;\n };\n\n const error = collect(selections, parentTypeName);\n if (error) return err(error);\n\n return ok(usages);\n};\n\n/**\n * Get the return type of a field (unwrapped from modifiers).\n */\nexport const getFieldReturnType = (schema: SchemaIndex, parentTypeName: string, fieldName: string): string | null => {\n const objectRecord = schema.objects.get(parentTypeName);\n if (!objectRecord) return null;\n\n const fieldDef = objectRecord.fields.get(fieldName);\n if (!fieldDef) return null;\n\n const { typeName } = parseTypeNode(fieldDef.type);\n return typeName;\n};\n\n/**\n * Merge multiple variable usages into a single InferredVariable.\n * Validates type compatibility and merges modifiers using List Coercion rules.\n *\n * The algorithm:\n * 1. Validate all usages have the same type name\n * 2. Merge minimumModifiers to find the candidate (shallowest type that could work)\n * 3. Verify the candidate can satisfy ALL expected modifiers via isModifierAssignable\n */\nexport const mergeVariableUsages = (\n variableName: string,\n usages: readonly VariableUsage[],\n): Result<InferredVariable, GraphqlCompatError> => {\n if (usages.length === 0) {\n // This shouldn't happen, but handle defensively\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `No usages found for variable \"${variableName}\"`,\n variableName,\n });\n }\n\n const first = usages[0]!;\n\n // Validate all usages have the same type name\n for (const usage of usages) {\n if (usage.typeName !== first.typeName) {\n return err({\n code: \"GRAPHQL_VARIABLE_TYPE_MISMATCH\",\n message: `Variable \"$${variableName}\" has conflicting types: \"${first.typeName}\" and \"${usage.typeName}\"`,\n variableName,\n });\n }\n }\n\n // Merge minimumModifiers to find candidate\n let candidateModifier = first.minimumModifier;\n for (let i = 1; i < usages.length; i++) {\n const result = mergeModifiers(candidateModifier, usages[i]!.minimumModifier);\n if (!result.ok) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" has incompatible modifiers: ${result.reason}`,\n variableName,\n });\n }\n candidateModifier = result.value;\n }\n\n // Verify candidate satisfies all expected modifiers\n for (const usage of usages) {\n if (!isModifierAssignable(candidateModifier, usage.expectedModifier)) {\n return err({\n code: \"GRAPHQL_VARIABLE_MODIFIER_INCOMPATIBLE\",\n message: `Variable \"$${variableName}\" with modifier \"${candidateModifier}\" cannot satisfy expected \"${usage.expectedModifier}\"`,\n variableName,\n });\n }\n }\n\n return ok({\n name: variableName,\n typeName: first.typeName,\n modifier: candidateModifier,\n typeKind: first.typeKind,\n });\n};\n\n/**\n * Infer variables from collected usages.\n * Groups by variable name and merges each group.\n */\nexport const inferVariablesFromUsages = (usages: readonly VariableUsage[]): Result<InferredVariable[], GraphqlCompatError> => {\n // Group usages by variable name\n const byName = new Map<string, VariableUsage[]>();\n for (const usage of usages) {\n const existing = byName.get(usage.name);\n if (existing) {\n existing.push(usage);\n } else {\n byName.set(usage.name, [usage]);\n }\n }\n\n // Merge each group\n const variables: InferredVariable[] = [];\n for (const [name, group] of byName) {\n const result = mergeVariableUsages(name, group);\n if (result.isErr()) return err(result.error);\n variables.push(result.value);\n }\n\n // Sort by name for deterministic output\n variables.sort((a, b) => a.name.localeCompare(b.name));\n\n return ok(variables);\n};\n\n/**\n * Check if a type name is a scalar type.\n */\nconst isScalarName = (schema: SchemaIndex, name: string): boolean => builtinScalarTypes.has(name) || schema.scalars.has(name);\n\n// ============================================================================\n// Fragment Dependency Ordering\n// ============================================================================\n\n/**\n * Topologically sort fragments so dependencies come before dependents.\n * Detects circular dependencies.\n *\n * Note: Uses the existing collectFragmentDependencies function defined below.\n */\nexport const sortFragmentsByDependency = (fragments: readonly ParsedFragment[]): Result<ParsedFragment[], GraphqlCompatError> => {\n // Build dependency graph using existing function\n const graph = new Map<string, Set<string>>();\n for (const fragment of fragments) {\n const deps = collectFragmentDependenciesSet(fragment.selections);\n graph.set(fragment.name, deps);\n }\n\n const fragmentByName = new Map<string, ParsedFragment>();\n for (const f of fragments) {\n fragmentByName.set(f.name, f);\n }\n\n const sorted: ParsedFragment[] = [];\n const visited = new Set<string>();\n const visiting = new Set<string>(); // For cycle detection\n\n const visit = (name: string, path: string[]): GraphqlCompatError | null => {\n if (visited.has(name)) return null;\n\n if (visiting.has(name)) {\n // Found a cycle\n const cycleStart = path.indexOf(name);\n const cycle = path.slice(cycleStart).concat(name);\n return {\n code: \"GRAPHQL_FRAGMENT_CIRCULAR_DEPENDENCY\",\n message: `Circular dependency detected in fragments: ${cycle.join(\" -> \")}`,\n fragmentNames: cycle,\n };\n }\n\n // Fragment might not be in our list (external dependency)\n const fragment = fragmentByName.get(name);\n if (!fragment) {\n // External fragment, skip\n visited.add(name);\n return null;\n }\n\n visiting.add(name);\n const deps = graph.get(name) ?? new Set();\n\n for (const dep of deps) {\n const error = visit(dep, [...path, name]);\n if (error) return error;\n }\n\n visiting.delete(name);\n visited.add(name);\n sorted.push(fragment);\n return null;\n };\n\n for (const fragment of fragments) {\n const error = visit(fragment.name, []);\n if (error) return err(error);\n }\n\n return ok(sorted);\n};\n\n/**\n * Recursively collect fragment spread names from selections into a Set.\n * Internal helper for sortFragmentsByDependency.\n */\nconst collectFragmentDependenciesSet = (selections: readonly ParsedSelection[]): Set<string> => {\n const deps = new Set<string>();\n\n const collect = (sels: readonly ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n deps.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return deps;\n};\n\n/**\n * Check if a type name is an enum type.\n */\nconst isEnumName = (schema: SchemaIndex, name: string): boolean => schema.enums.has(name);\n\n/**\n * Enriched operation with resolved type information.\n */\nexport type EnrichedOperation = Omit<ParsedOperation, \"variables\"> & {\n readonly variables: readonly EnrichedVariable[];\n /** Fragment names used in this operation (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched fragment with resolved type information.\n */\nexport type EnrichedFragment = ParsedFragment & {\n /** Fragment names used in this fragment (for imports) */\n readonly fragmentDependencies: readonly string[];\n /** Variables inferred from field arguments in this fragment */\n readonly variables: readonly InferredVariable[];\n};\n\n/**\n * Enriched variable with resolved type kind.\n */\nexport type EnrichedVariable = Omit<ParsedVariable, \"typeKind\"> & {\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Result of transforming parsed operations.\n */\nexport type TransformResult = {\n readonly operations: readonly EnrichedOperation[];\n readonly fragments: readonly EnrichedFragment[];\n};\n\n/**\n * Options for transformation.\n */\nexport type TransformOptions = {\n /** Schema document for type resolution */\n readonly schemaDocument: DocumentNode;\n};\n\n/**\n * Transform parsed operations/fragments by enriching them with schema information.\n *\n * This resolves variable type kinds (scalar, enum, input), collects\n * fragment dependencies, and infers variables for fragments.\n */\nexport const transformParsedGraphql = (\n parsed: ParseResult,\n options: TransformOptions,\n): Result<TransformResult, GraphqlCompatError> => {\n const schema = createSchemaIndex(options.schemaDocument);\n\n // Sort fragments by dependency (dependencies first)\n const sortResult = sortFragmentsByDependency(parsed.fragments);\n if (sortResult.isErr()) {\n return err(sortResult.error);\n }\n const sortedFragments = sortResult.value;\n\n // Transform fragments in dependency order, building up resolved variables map\n const resolvedFragmentVariables = new Map<string, readonly InferredVariable[]>();\n const fragments: EnrichedFragment[] = [];\n\n for (const frag of sortedFragments) {\n const result = transformFragment(frag, schema, resolvedFragmentVariables);\n if (result.isErr()) {\n return err(result.error);\n }\n resolvedFragmentVariables.set(frag.name, result.value.variables);\n fragments.push(result.value);\n }\n\n // Transform operations\n const operations: EnrichedOperation[] = [];\n for (const op of parsed.operations) {\n const result = transformOperation(op, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n operations.push(result.value);\n }\n\n return ok({ operations, fragments });\n};\n\n/**\n * Transform a single operation.\n */\nconst transformOperation = (op: ParsedOperation, schema: SchemaIndex): Result<EnrichedOperation, GraphqlCompatError> => {\n // Resolve variable type kinds\n const variables: EnrichedVariable[] = [];\n for (const v of op.variables) {\n const typeKind = resolveTypeKind(schema, v.typeName);\n if (typeKind === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${v.typeName}\" in variable \"${v.name}\"`,\n typeName: v.typeName,\n });\n }\n variables.push({ ...v, typeKind });\n }\n\n // Collect fragment dependencies\n const fragmentDependencies = collectFragmentDependencies(op.selections);\n\n return ok({\n ...op,\n variables,\n fragmentDependencies,\n });\n};\n\n/**\n * Transform a single fragment.\n * Infers variables from field arguments and propagates variables from spread fragments.\n */\nconst transformFragment = (\n frag: ParsedFragment,\n schema: SchemaIndex,\n resolvedFragmentVariables: Map<string, readonly InferredVariable[]>,\n): Result<EnrichedFragment, GraphqlCompatError> => {\n // Collect fragment dependencies (fragments used within this fragment)\n const fragmentDependencies = collectFragmentDependencies(frag.selections);\n\n // Collect direct variable usages from this fragment's selections\n const directUsagesResult = collectVariableUsages(frag.selections, frag.onType, schema);\n if (directUsagesResult.isErr()) {\n return err(directUsagesResult.error);\n }\n const directUsages = directUsagesResult.value;\n\n // Collect variables from spread fragments\n const spreadVariables: InferredVariable[] = [];\n for (const depName of fragmentDependencies) {\n const depVariables = resolvedFragmentVariables.get(depName);\n if (depVariables) {\n spreadVariables.push(...depVariables);\n }\n // If not found, it's an external fragment - skip\n }\n\n // Combine direct usages with spread variables\n // Convert spread variables (InferredVariable) to usages for merging\n // For already-inferred variables, expectedModifier and minimumModifier are the same\n const allUsages: VariableUsage[] = [\n ...directUsages,\n ...spreadVariables.map((v) => ({\n name: v.name,\n typeName: v.typeName,\n expectedModifier: v.modifier,\n minimumModifier: v.modifier,\n typeKind: v.typeKind,\n })),\n ];\n\n // Infer final variables\n const variablesResult = inferVariablesFromUsages(allUsages);\n if (variablesResult.isErr()) {\n return err(variablesResult.error);\n }\n\n return ok({\n ...frag,\n fragmentDependencies,\n variables: variablesResult.value,\n });\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKind = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) {\n return \"scalar\";\n }\n if (isEnumName(schema, typeName)) {\n return \"enum\";\n }\n if (schema.inputs.has(typeName)) {\n return \"input\";\n }\n return null;\n};\n\n/**\n * Collect fragment names used in selections (recursively).\n */\nconst collectFragmentDependencies = (selections: readonly import(\"./types\").ParsedSelection[]): readonly string[] => {\n const fragments = new Set<string>();\n\n const collect = (sels: readonly import(\"./types\").ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n fragments.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return [...fragments];\n};\n","/**\n * Emitter for generating TypeScript compat code from enriched operations.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport {\n type EnrichedFragment,\n type EnrichedOperation,\n getArgumentType,\n getFieldReturnType,\n getInputFieldType,\n} from \"./transformer\";\nimport type { GraphqlCompatError, ParsedArgument, ParsedInlineFragment, ParsedSelection, ParsedValue, TypeInfo } from \"./types\";\n\n/**\n * Schema index for type lookups.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Options for code emission.\n */\nexport type EmitOptions = {\n /** Schema name to use in gql.schemaName() call */\n readonly schemaName: string;\n /** Import path for graphql-system module */\n readonly graphqlSystemPath: string;\n /** Schema document for type lookups (required for inline fragment support) */\n readonly schemaDocument?: DocumentNode;\n};\n\n/**\n * Map operation kind to root type name.\n * Uses schema.operationTypes if available, falls back to standard names.\n */\nconst getRootTypeName = (schema: SchemaIndex | null, kind: \"query\" | \"mutation\" | \"subscription\"): string => {\n if (schema) {\n switch (kind) {\n case \"query\":\n return schema.operationTypes.query ?? \"Query\";\n case \"mutation\":\n return schema.operationTypes.mutation ?? \"Mutation\";\n case \"subscription\":\n return schema.operationTypes.subscription ?? \"Subscription\";\n }\n }\n // Fallback when no schema is available\n switch (kind) {\n case \"query\":\n return \"Query\";\n case \"mutation\":\n return \"Mutation\";\n case \"subscription\":\n return \"Subscription\";\n }\n};\n\n/**\n * Emit TypeScript code for an operation.\n */\nexport const emitOperation = (operation: EnrichedOperation, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${operation.name}Compat`;\n const operationType = operation.kind;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);\n lines.push(` ${operationType}.compat({`);\n lines.push(` name: ${JSON.stringify(operation.name)},`);\n\n // Variables\n if (operation.variables.length > 0) {\n lines.push(` variables: { ${emitVariables(operation.variables)} },`);\n }\n\n // Fields - pass root type name for list coercion\n const rootTypeName = getRootTypeName(schema, operation.kind);\n lines.push(` fields: ({ f, $ }) => ({`);\n const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema, rootTypeName);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit TypeScript code for a fragment.\n */\nexport const emitFragment = (fragment: EnrichedFragment, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n const hasVariables = fragment.variables.length > 0;\n\n // Note: imports (gql and fragment) are handled by the caller\n\n // Generate export\n const exportName = `${fragment.name}Fragment`;\n\n // Include $var in destructure if fragment has variables\n const destructure = hasVariables ? \"fragment, $var\" : \"fragment\";\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${destructure} }) =>`);\n lines.push(` fragment.${fragment.onType}({`);\n\n // Variables block (if any)\n if (hasVariables) {\n lines.push(` variables: { ${emitVariables(fragment.variables)} },`);\n }\n\n // Fields - include $ in context if fragment has variables\n // Pass fragment.onType as the parent type for list coercion\n const fieldsContext = hasVariables ? \"{ f, $ }\" : \"{ f }\";\n lines.push(` fields: (${fieldsContext}) => ({`);\n const fieldLinesResult = emitSelections(fragment.selections, 3, fragment.variables, schema, fragment.onType);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Common variable type for emission (both EnrichedVariable and InferredVariable have these fields).\n */\ntype EmittableVariable = {\n readonly name: string;\n readonly typeName: string;\n readonly modifier: string;\n};\n\n/**\n * Emit variable definitions.\n */\nconst emitVariables = (variables: readonly EmittableVariable[]): string => {\n return variables.map((v) => `...$var(${JSON.stringify(v.name)}).${v.typeName}(${JSON.stringify(v.modifier)})`).join(\", \");\n};\n\n/**\n * Emit field selections (public API).\n * Converts variable array to Set<string> and delegates to internal implementation.\n */\nconst emitSelections = (\n selections: readonly ParsedSelection[],\n indent: number,\n variables: readonly EmittableVariable[],\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const variableNames = new Set(variables.map((v) => v.name));\n return emitSelectionsInternal(selections, indent, variableNames, schema, parentTypeName);\n};\n\n/**\n * Internal implementation for emitting field selections.\n * Takes variableNames as Set<string> for recursive calls.\n */\nconst emitSelectionsInternal = (\n selections: readonly ParsedSelection[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n\n // Separate inline fragments from other selections\n const inlineFragments: ParsedInlineFragment[] = [];\n const otherSelections: ParsedSelection[] = [];\n\n for (const sel of selections) {\n if (sel.kind === \"inlineFragment\") {\n inlineFragments.push(sel);\n } else {\n otherSelections.push(sel);\n }\n }\n\n // Emit regular selections (fields and fragment spreads)\n for (const sel of otherSelections) {\n const result = emitSingleSelection(sel, indent, variableNames, schema, parentTypeName);\n if (result.isErr()) {\n return err(result.error);\n }\n lines.push(result.value);\n }\n\n // Emit grouped inline fragments as union selections\n if (inlineFragments.length > 0) {\n const unionResult = emitInlineFragmentsAsUnion(inlineFragments, indent, variableNames, schema);\n if (unionResult.isErr()) {\n return err(unionResult.error);\n }\n lines.push(unionResult.value);\n }\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit a single selection (field or fragment spread).\n */\nconst emitSingleSelection = (\n sel: ParsedSelection,\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n switch (sel.kind) {\n case \"field\":\n return emitFieldSelection(sel, indent, variableNames, schema, parentTypeName);\n case \"fragmentSpread\":\n return ok(`${padding}...${sel.name}Fragment.spread(),`);\n case \"inlineFragment\":\n // This should not happen as inline fragments are handled separately\n return ok(\"\");\n }\n};\n\n/**\n * Emit inline fragments grouped as a union selection.\n * Format: { TypeA: ({ f }) => ({ ...fields }), TypeB: ({ f }) => ({ ...fields }) }\n */\nconst emitInlineFragmentsAsUnion = (\n inlineFragments: readonly ParsedInlineFragment[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Validate inline fragments have type conditions\n for (const frag of inlineFragments) {\n if (frag.onType === \"\") {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_WITHOUT_TYPE\",\n message: \"Inline fragments without type condition are not supported. Use `... on TypeName { }` syntax.\",\n });\n }\n }\n\n // Validate all inline fragments are on union types (not interfaces)\n for (const frag of inlineFragments) {\n if (schema && !schema.objects.has(frag.onType)) {\n // If it's not a known object type, it might be an interface\n // Check if any union contains this type as a member\n let isUnionMember = false;\n for (const [, unionDef] of schema.unions) {\n if (unionDef.members.has(frag.onType)) {\n isUnionMember = true;\n break;\n }\n }\n if (!isUnionMember) {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_ON_INTERFACE\",\n message: `Inline fragments on interface type \"${frag.onType}\" are not supported. Use union types instead.`,\n onType: frag.onType,\n });\n }\n }\n }\n\n // Build union member entries\n const entries: string[] = [];\n for (const frag of inlineFragments) {\n const innerPadding = \" \".repeat(indent + 1);\n // Use the inline fragment's type condition as the parent type for nested selections\n const fieldsResult = emitSelectionsInternal(frag.selections, indent + 2, variableNames, schema, frag.onType);\n if (fieldsResult.isErr()) {\n return err(fieldsResult.error);\n }\n\n entries.push(`${innerPadding}${frag.onType}: ({ f }) => ({\n${fieldsResult.value}\n${innerPadding}}),`);\n }\n\n // Emit as spread with union callback: ...f.fieldName()({ Type: ... })\n // Note: This assumes the parent field handles the union - we emit just the union object\n return ok(`${padding}...({\n${entries.join(\"\\n\")}\n${padding}}),`);\n};\n\n/**\n * Emit a single field selection.\n */\nconst emitFieldSelection = (\n field: ParsedSelection & { kind: \"field\" },\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Extract optional fields for type narrowing\n const args = field.arguments;\n const selections = field.selections;\n const hasArgs = args && args.length > 0;\n const hasSelections = selections && selections.length > 0;\n\n // Use shorthand syntax for scalar fields (no args, no nested selections)\n if (!hasArgs && !hasSelections) {\n return ok(`${padding}${field.name}: true,`);\n }\n\n let line = `${padding}...f.${field.name}(`;\n\n if (hasArgs) {\n const argsResult = emitArguments(args, variableNames, schema, parentTypeName, field.name);\n if (argsResult.isErr()) {\n return err(argsResult.error);\n }\n line += argsResult.value;\n }\n\n line += \")\";\n\n if (hasSelections) {\n // Check if selections contain inline fragments (union field)\n const hasInlineFragments = selections.some((s) => s.kind === \"inlineFragment\");\n\n // Determine nested parent type for recursive selections\n const nestedParentType =\n schema && parentTypeName ? (getFieldReturnType(schema, parentTypeName, field.name) ?? undefined) : undefined;\n\n if (hasInlineFragments) {\n // Union field: emit with union callback pattern\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema, nestedParentType);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += \"({\\n\";\n line += `${nestedResult.value}\\n`;\n line += `${padding}})`;\n } else {\n // Regular nested selections\n line += \"(({ f }) => ({\\n\";\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema, nestedParentType);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += `${nestedResult.value}\\n`;\n line += `${padding}}))`;\n }\n }\n\n line += \",\";\n\n return ok(line);\n};\n\n// ============================================================================\n// List Coercion Utilities\n// ============================================================================\n\n/**\n * Check if a modifier represents a list type (contains []).\n */\nconst isListModifier = (modifier: string): boolean => {\n return modifier.includes(\"[]\");\n};\n\n/**\n * Determine if a value needs to be wrapped in an array for list coercion.\n * Returns true if:\n * - Expected type is a list\n * - Value is NOT already a list\n * - Value is NOT a variable (runtime handles coercion)\n * - Value is NOT null\n */\nconst needsListCoercion = (value: ParsedValue, expectedModifier: string | undefined): boolean => {\n // No coercion if no expected type info\n if (!expectedModifier) return false;\n\n // No coercion if expected type is not a list\n if (!isListModifier(expectedModifier)) return false;\n\n // No coercion for variables (runtime handles this)\n if (value.kind === \"variable\") return false;\n\n // No coercion for null\n if (value.kind === \"null\") return false;\n\n // No coercion if value is already a list\n if (value.kind === \"list\") return false;\n\n return true;\n};\n\n/**\n * Extract the element type from a list type by removing the outermost list modifier.\n * For example: \"![]!\" (non-null list of non-null) → \"!\" (non-null element)\n * \"?[]![]!\" (nested lists) → \"?[]!\" (inner list type)\n * Returns null if the modifier doesn't represent a list type.\n */\nconst getListElementType = (expectedType: TypeInfo): TypeInfo | null => {\n const { modifier, typeName } = expectedType;\n // Modifier format: {inner}{list_modifiers}\n // Strip the outermost list modifier ([]! or []?)\n const listMatch = modifier.match(/^(.+?)(\\[\\][!?])$/);\n if (!listMatch || !listMatch[1]) return null;\n return { typeName, modifier: listMatch[1] };\n};\n\n/**\n * Emit a value with type context for list coercion.\n */\nconst emitValueWithType = (\n value: ParsedValue,\n expectedType: TypeInfo | null,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n // Check if list coercion is needed\n const shouldCoerce = needsListCoercion(value, expectedType?.modifier);\n\n // Handle object values with recursive type context\n if (value.kind === \"object\" && expectedType && schema) {\n return emitObjectWithType(value, expectedType.typeName, variableNames, schema, shouldCoerce);\n }\n\n // Handle list values with recursive type context for element coercion\n if (value.kind === \"list\" && expectedType && schema) {\n const elementType = getListElementType(expectedType);\n if (elementType) {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValueWithType(v, elementType, variableNames, schema);\n if (result.isErr()) return result;\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n }\n\n // Emit the value normally\n const result = emitValue(value, variableNames);\n if (result.isErr()) return result;\n\n // Wrap in array if coercion needed\n if (shouldCoerce) {\n return ok(`[${result.value}]`);\n }\n\n return result;\n};\n\n/**\n * Emit an object value with type context for recursive list coercion.\n */\nconst emitObjectWithType = (\n value: ParsedValue & { kind: \"object\" },\n inputTypeName: string,\n variableNames: Set<string>,\n schema: SchemaIndex,\n wrapInArray: boolean,\n): Result<string, GraphqlCompatError> => {\n if (value.fields.length === 0) {\n return ok(wrapInArray ? \"[{}]\" : \"{}\");\n }\n\n const entries: string[] = [];\n for (const f of value.fields) {\n // Look up field type from input object definition\n const fieldType = getInputFieldType(schema, inputTypeName, f.name);\n if (fieldType === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_FIELD\",\n message: `Unknown field \"${f.name}\" on input type \"${inputTypeName}\"`,\n typeName: inputTypeName,\n fieldName: f.name,\n });\n }\n\n const result = emitValueWithType(f.value, fieldType, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n\n const objectStr = `{ ${entries.join(\", \")} }`;\n return ok(wrapInArray ? `[${objectStr}]` : objectStr);\n};\n\n// ============================================================================\n// Argument Emission\n// ============================================================================\n\n/**\n * Emit field arguments with type context for list coercion.\n */\nconst emitArguments = (\n args: readonly ParsedArgument[],\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n parentTypeName: string | undefined,\n fieldName: string | undefined,\n): Result<string, GraphqlCompatError> => {\n if (args.length === 0) {\n return ok(\"\");\n }\n\n const argEntries: string[] = [];\n for (const arg of args) {\n // Look up expected type from schema\n const expectedType =\n schema && parentTypeName && fieldName ? getArgumentType(schema, parentTypeName, fieldName, arg.name) : null;\n\n const result = emitValueWithType(arg.value, expectedType, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n argEntries.push(`${arg.name}: ${result.value}`);\n }\n return ok(`{ ${argEntries.join(\", \")} }`);\n};\n\n/**\n * Emit a value (literal or variable reference).\n */\nconst emitValue = (value: ParsedValue, variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n switch (value.kind) {\n case \"variable\":\n // Check if it's a declared variable\n if (variableNames.has(value.name)) {\n return ok(`$.${value.name}`);\n }\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `Variable \"$${value.name}\" is not declared in the operation`,\n variableName: value.name,\n });\n case \"int\":\n case \"float\":\n return ok(value.value);\n case \"string\":\n return ok(JSON.stringify(value.value));\n case \"boolean\":\n return ok(value.value ? \"true\" : \"false\");\n case \"null\":\n return ok(\"null\");\n case \"enum\":\n // Enums are emitted as string literals in soda-gql\n return ok(JSON.stringify(value.value));\n case \"list\": {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValue(v, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n case \"object\": {\n if (value.fields.length === 0) {\n return ok(\"{}\");\n }\n const entries: string[] = [];\n for (const f of value.fields) {\n const result = emitValue(f.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n return ok(`{ ${entries.join(\", \")} }`);\n }\n }\n};\n","import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, rmSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const removeDirectory = (dirPath: string) => {\n const targetPath = resolve(dirPath);\n try {\n rmSync(targetPath, { recursive: true, force: true });\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"REMOVE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport type { TypeFilterConfig } from \"@soda-gql/config\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { removeDirectory, writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Build typeFilters config for each schema\n const typeFiltersConfig = new Map<string, TypeFilterConfig>();\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.typeFilter) {\n typeFiltersConfig.set(schemaName, schemaConfig.typeFilter);\n }\n }\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n typeFilters: typeFiltersConfig.size > 0 ? typeFiltersConfig : undefined,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Clean up existing _defs directory to prevent stale files\n const defsDir = join(outDir, \"_defs\");\n if (existsSync(defsDir)) {\n const removeResult = removeDirectory(defsDir);\n if (removeResult.isErr()) {\n return err(removeResult.error);\n }\n }\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;;;;;;;;;AA2CA,MAAa,oBAAoB,aAA8D;CAC7F,MAAM,sCAAuB,SAAS;AAEtC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAW;GACT,MAAM;GACN,SAAS,6BAA6B;GACtC,UAAU;GACX,CAAC;;AAGJ,KAAI;EACF,MAAM,mCAAsB,cAAc,OAAO;EACjD,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,aAAa,CAAC;UAC/C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAa,sBAAsB,QAAgB,eAAgE;AACjH,KAAI;EACF,MAAM,8BAAiB,OAAO;AAC9B,4BAAU,oBAAoB,UAAU,WAAW,CAAC;UAC7C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAW;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAM,uBAAuB,UAAwB,eAAoC;CACvF,MAAMA,aAAgC,EAAE;CACxC,MAAMC,YAA8B,EAAE;AAEtC,MAAK,MAAM,cAAc,SAAS,aAAa;AAC7C,MAAI,WAAW,SAASC,aAAK,sBAAsB;GACjD,MAAM,YAAY,iBAAiB,YAAY,WAAW;AAC1D,OAAI,WAAW;AACb,eAAW,KAAK,UAAU;;aAEnB,WAAW,SAASA,aAAK,qBAAqB;AACvD,aAAU,KAAK,gBAAgB,YAAY,WAAW,CAAC;;;AAI3D,QAAO;EAAE;EAAY;EAAW;;;;;AAMlC,MAAM,oBAAoB,MAA+B,eAA+C;AAEtG,KAAI,CAAC,KAAK,MAAM;AACd,SAAO;;CAGT,MAAMC,aAA+B,KAAK,uBAAuB,EAAE,EAAE,IAAI,gBAAgB;CACzF,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK;EACX,MAAM,KAAK,KAAK;EAChB;EACA;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,MAA8B,eAAuC;CAC5F,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,QAAQ,KAAK,cAAc,KAAK;EAChC;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,SAAiD;CACxE,MAAM,EAAE,UAAU,aAAa,cAAc,KAAK,KAAK;CACvD,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,aAAa,GAAG;AAE3E,QAAO;EACL,MAAM,KAAK,SAAS,KAAK;EACzB;EACA;EAEA,UAAU;EACV;EACD;;;;;;;;;AAUH,MAAa,iBAAiB,SAA6B;CAEzD,MAAMC,SAAsB,EAAE;CAE9B,MAAM,WAAW,GAAa,YAA6B;AACzD,MAAI,EAAE,SAASF,aAAK,eAAe;AACjC,UAAO,QAAQ,EAAE,MAAM,KAAK;;AAE9B,MAAI,EAAE,SAASA,aAAK,WAAW;AAC7B,UAAO,KAAK;IAAE,MAAM;IAAQ;IAAS,CAAC;AACtC,UAAO,QAAQ,EAAE,MAAM,MAAM;;AAE/B,SAAO,KAAK;GAAE,MAAM;GAAS;GAAS,CAAC;AACvC,SAAO,EAAE,KAAK;;CAGhB,MAAM,WAAW,QAAQ,MAAM,MAAM;CAGrC,IAAI,WAAW;AACf,MAAK,MAAM,SAAS,OAAO,OAAO,CAAC,SAAS,EAAE;AAC5C,MAAI,MAAM,SAAS,SAAS;AAC1B,cAAW,MAAM,UAAU,MAAM;AACjC;;EAEF,MAAM,aAAa,MAAM,UAAU,QAAQ;AAC3C,aAAW,GAAG,WAAW;;AAG3B,QAAO;EAAE;EAAU;EAAU;;;;;AAM/B,MAAM,qBAAqB,eAA4D;AACrF,QAAO,WAAW,IAAI,iBAAiB;;;;;AAMzC,MAAM,oBAAoB,SAAyC;AACjE,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,MACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;EACpC,KAAKA,aAAK,gBACR,QAAO,sBAAsB,KAAK;;;;;;AAOxC,MAAM,yBAAyB,SAA0C;CACvE,MAAM,OAAO,KAAK,WAAW,SAAS,KAAK,UAAU,IAAI,gBAAgB,GAAG;CAC5E,MAAM,aAAa,KAAK,eAAe,kBAAkB,KAAK,aAAa,WAAW,GAAG;AAEzF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EAChB,OAAO,KAAK,OAAO;EACnB,WAAW;EACX;EACD;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EACjB;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,QAAQ,KAAK,eAAe,KAAK,SAAS;EAC1C,YAAY,kBAAkB,KAAK,aAAa,WAAW;EAC5D;;;;;AAMH,MAAM,mBAAmB,SAAuC;AAC9D,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,OAAO,aAAa,KAAK,MAAM;EAChC;;;;;AAMH,MAAM,qBAAqB,UAAwB;AACjD,OAAM,IAAI,MAAM,qBAAqB,KAAK,UAAU,MAAM,GAAG;;;;;AAM/D,MAAM,gBAAgB,SAAiC;AACrD,SAAQ,KAAK,MAAb;EACE,KAAKA,aAAK,SACR,QAAO;GAAE,MAAM;GAAY,MAAM,KAAK,KAAK;GAAO;EACpD,KAAKA,aAAK,IACR,QAAO;GAAE,MAAM;GAAO,OAAO,KAAK;GAAO;EAC3C,KAAKA,aAAK,MACR,QAAO;GAAE,MAAM;GAAS,OAAO,KAAK;GAAO;EAC7C,KAAKA,aAAK,OACR,QAAO;GAAE,MAAM;GAAU,OAAO,KAAK;GAAO;EAC9C,KAAKA,aAAK,QACR,QAAO;GAAE,MAAM;GAAW,OAAO,KAAK;GAAO;EAC/C,KAAKA,aAAK,KACR,QAAO,EAAE,MAAM,QAAQ;EACzB,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,OAAO,KAAK;GAAO;EAC5C,KAAKA,aAAK,KACR,QAAO;GAAE,MAAM;GAAQ,QAAQ,KAAK,OAAO,IAAI,aAAa;GAAE;EAChE,KAAKA,aAAK,OACR,QAAO;GACL,MAAM;GACN,QAAQ,KAAK,OAAO,KAAK,WAAW;IAClC,MAAM,MAAM,KAAK;IACjB,OAAO,aAAa,MAAM,MAAM;IACjC,EAAE;GACJ;EACH,QACE,QAAO,kBAAkB,KAAK;;;;;;;;;AChRpC,MAAM,qBAAqB,IAAI,IAAI;CAAC;CAAM;CAAU;CAAO;CAAS;CAAU,CAAC;;;;;;AAoB/E,MAAM,0BAA0B,aAAwC;CAEtE,MAAM,QAAQ,SAAS,OAAO,MAAM,MAAM;CAG1C,MAAMG,QAA2B,EAAE;CACnC,MAAM,cAAc;CACpB,IAAIC;AACJ,SAAQ,QAAQ,YAAY,KAAK,SAAS,MAAM,MAAM;AACpD,QAAM,KAAK,KAAK,MAAM,KAAsB;;AAG9C,QAAO;EAAE;EAAO;EAAO;;;;;AAMzB,MAAM,iBAAiB,cAAyC;AAC9D,QAAO,UAAU,QAAQ,UAAU,MAAM,KAAK,GAAG;;;;;;;;;;;;;;AAenD,MAAa,wBAAwB,QAAgB,WAA4B;CAC/E,MAAM,YAAY,uBAAuB,OAAO;CAChD,MAAM,YAAY,uBAAuB,OAAO;CAEhD,MAAM,YAAY,UAAU,MAAM,SAAS,UAAU,MAAM;AAG3D,KAAI,YAAY,KAAK,YAAY,EAAG,QAAO;CAI3C,MAAM,oBAAoB,cAAc,IAAI,UAAU,MAAM,MAAM,EAAE,GAAG,UAAU;AAKjF,KAAI,cAAc,KAAK,UAAU,MAAM,WAAW,KAAK,UAAU,UAAU,OAAO,UAAU,MAAM,OAAO,OAAO;AAC9G,SAAO;;AAIT,KAAI,UAAU,UAAU,OAAO,UAAU,UAAU,IAAK,QAAO;AAG/D,MAAK,IAAI,IAAI,GAAG,IAAI,UAAU,MAAM,QAAQ,KAAK;EAC/C,MAAM,UAAU,UAAU,MAAM;EAChC,MAAM,UAAU,kBAAkB;AAClC,MAAI,YAAY,SAAS,YAAY,MAAO,QAAO;;AAGrD,QAAO;;;;;;;;;AAUT,MAAM,yBAAyB,qBAAqC;CAClE,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,KAAI,OAAO,MAAM,SAAS,GAAG;AAC3B,SAAO,cAAc;GAAE,OAAO,OAAO;GAAO,OAAO,OAAO,MAAM,MAAM,EAAE;GAAE,CAAC;;AAE7E,QAAO;;;;;;;;;;;AAYT,MAAa,kBAAkB,GAAW,MAA2E;CACnH,MAAM,UAAU,uBAAuB,EAAE;CACzC,MAAM,UAAU,uBAAuB,EAAE;AAGzC,KAAI,QAAQ,MAAM,WAAW,QAAQ,MAAM,QAAQ;AACjD,SAAO;GACL,IAAI;GACJ,QAAQ,8BAA8B,EAAE,QAAQ,QAAQ,MAAM,OAAO,mBAAmB,EAAE,QAAQ,QAAQ,MAAM;GACjH;;CAIH,MAAMC,cAAyB,QAAQ,UAAU,OAAO,QAAQ,UAAU,MAAM,MAAM;CAGtF,MAAMC,cAAiC,EAAE;AACzC,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,MAAM,QAAQ,KAAK;EAC7C,MAAM,QAAQ,QAAQ,MAAM;EAC5B,MAAM,QAAQ,QAAQ,MAAM;AAC5B,cAAY,KAAK,UAAU,SAAS,UAAU,QAAQ,QAAQ,MAAM;;AAGtE,QAAO;EAAE,IAAI;EAAM,OAAO,cAAc;GAAE,OAAO;GAAa,OAAO;GAAa,CAAC;EAAE;;;;;;AAwBvF,MAAa,mBACX,QACA,gBACA,WACA,iBACoB;CACpB,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,SAAS,SAAS,WAAW,MAAM,QAAQ,IAAI,KAAK,UAAU,aAAa;AACjF,KAAI,CAAC,OAAQ,QAAO;AAEpB,QAAO,cAAc,OAAO,KAAK;;;;;AAMnC,MAAa,qBAAqB,QAAqB,eAAuB,cAAuC;CACnH,MAAM,cAAc,OAAO,OAAO,IAAI,cAAc;AACpD,KAAI,CAAC,YAAa,QAAO;CAEzB,MAAM,WAAW,YAAY,OAAO,IAAI,UAAU;AAClD,KAAI,CAAC,SAAU,QAAO;AAEtB,QAAO,cAAc,SAAS,KAAK;;;;;AAMrC,MAAM,2BAA2B,QAAqB,aAAyD;AAC7G,KAAI,aAAa,QAAQ,SAAS,CAAE,QAAO;AAC3C,KAAI,WAAW,QAAQ,SAAS,CAAE,QAAO;AACzC,KAAI,OAAO,OAAO,IAAI,SAAS,CAAE,QAAO;AACxC,QAAO;;;;;;AAOT,MAAM,6BACJ,OACA,kBACA,kBACA,QACA,WAC8B;AAC9B,KAAI,MAAM,SAAS,YAAY;EAC7B,MAAM,WAAW,wBAAwB,QAAQ,iBAAiB;AAClE,MAAI,CAAC,UAAU;AACb,UAAO;IACL,MAAM;IACN,SAAS,iBAAiB,iBAAiB,mBAAmB,MAAM,KAAK;IACzE,UAAU;IACX;;AAEH,SAAO,KAAK;GACV,MAAM,MAAM;GACZ,UAAU;GACV;GACA,iBAAiB,sBAAsB,iBAAiB;GACxD;GACD,CAAC;AACF,SAAO;;AAGT,KAAI,MAAM,SAAS,UAAU;AAE3B,OAAK,MAAM,SAAS,MAAM,QAAQ;GAChC,MAAM,YAAY,kBAAkB,QAAQ,kBAAkB,MAAM,KAAK;AACzE,OAAI,CAAC,WAAW;AACd,WAAO;KACL,MAAM;KACN,SAAS,kBAAkB,MAAM,KAAK,mBAAmB,iBAAiB;KAC1E,UAAU;KACV,WAAW,MAAM;KAClB;;GAEH,MAAM,QAAQ,0BAA0B,MAAM,OAAO,UAAU,UAAU,UAAU,UAAU,QAAQ,OAAO;AAC5G,OAAI,MAAO,QAAO;;AAEpB,SAAO;;AAGT,KAAI,MAAM,SAAS,QAAQ;EAGzB,MAAM,SAAS,uBAAuB,iBAAiB;AACvD,MAAI,OAAO,MAAM,SAAS,GAAG;GAC3B,MAAM,gBAAgB,cAAc;IAClC,OAAO,OAAO;IACd,OAAO,OAAO,MAAM,MAAM,EAAE;IAC7B,CAAC;AACF,QAAK,MAAM,QAAQ,MAAM,QAAQ;IAC/B,MAAM,QAAQ,0BAA0B,MAAM,kBAAkB,eAAe,QAAQ,OAAO;AAC9F,QAAI,MAAO,QAAO;;;;AAMxB,QAAO;;;;;AAMT,MAAM,iCACJ,MACA,gBACA,WACA,QACA,WAC8B;AAC9B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,UAAU,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK;AAC5E,MAAI,CAAC,SAAS;AACZ,UAAO;IACL,MAAM;IACN,SAAS,qBAAqB,IAAI,KAAK,cAAc,UAAU;IAC/D;IACA,cAAc,IAAI;IACnB;;EAEH,MAAM,QAAQ,0BAA0B,IAAI,OAAO,QAAQ,UAAU,QAAQ,UAAU,QAAQ,OAAO;AACtG,MAAI,MAAO,QAAO;;AAEpB,QAAO;;;;;AAMT,MAAa,yBACX,YACA,gBACA,WACgD;CAChD,MAAMC,SAA0B,EAAE;CAElC,MAAM,WAAW,MAAkC,eAAkD;AACnG,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK,SAAS;AAEZ,SAAI,IAAI,aAAa,IAAI,UAAU,SAAS,GAAG;MAC7C,MAAMC,UAAQ,8BAA8B,IAAI,WAAW,YAAY,IAAI,MAAM,QAAQ,OAAO;AAChG,UAAIA,QAAO,QAAOA;;AAIpB,SAAI,IAAI,cAAc,IAAI,WAAW,SAAS,GAAG;MAE/C,MAAM,kBAAkB,mBAAmB,QAAQ,YAAY,IAAI,KAAK;AACxE,UAAI,CAAC,iBAAiB;AACpB,cAAO;QACL,MAAM;QACN,SAAS,kBAAkB,IAAI,KAAK,aAAa,WAAW;QAC5D,UAAU;QACV,WAAW,IAAI;QAChB;;MAEH,MAAMA,UAAQ,QAAQ,IAAI,YAAY,gBAAgB;AACtD,UAAIA,QAAO,QAAOA;;AAEpB;;IAEF,KAAK,kBAAkB;KAErB,MAAMA,UAAQ,QAAQ,IAAI,YAAY,IAAI,OAAO;AACjD,SAAIA,QAAO,QAAOA;AAClB;;IAEF,KAAK,iBAEH;;;AAGN,SAAO;;CAGT,MAAM,QAAQ,QAAQ,YAAY,eAAe;AACjD,KAAI,MAAO,4BAAW,MAAM;AAE5B,2BAAU,OAAO;;;;;AAMnB,MAAa,sBAAsB,QAAqB,gBAAwB,cAAqC;CACnH,MAAM,eAAe,OAAO,QAAQ,IAAI,eAAe;AACvD,KAAI,CAAC,aAAc,QAAO;CAE1B,MAAM,WAAW,aAAa,OAAO,IAAI,UAAU;AACnD,KAAI,CAAC,SAAU,QAAO;CAEtB,MAAM,EAAE,aAAa,cAAc,SAAS,KAAK;AACjD,QAAO;;;;;;;;;;;AAYT,MAAa,uBACX,cACA,WACiD;AACjD,KAAI,OAAO,WAAW,GAAG;AAEvB,6BAAW;GACT,MAAM;GACN,SAAS,iCAAiC,aAAa;GACvD;GACD,CAAC;;CAGJ,MAAM,QAAQ,OAAO;AAGrB,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,MAAM,aAAa,MAAM,UAAU;AACrC,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,4BAA4B,MAAM,SAAS,SAAS,MAAM,SAAS;IACvG;IACD,CAAC;;;CAKN,IAAI,oBAAoB,MAAM;AAC9B,MAAK,IAAI,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;EACtC,MAAM,SAAS,eAAe,mBAAmB,OAAO,GAAI,gBAAgB;AAC5E,MAAI,CAAC,OAAO,IAAI;AACd,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,gCAAgC,OAAO;IAC3E;IACD,CAAC;;AAEJ,sBAAoB,OAAO;;AAI7B,MAAK,MAAM,SAAS,QAAQ;AAC1B,MAAI,CAAC,qBAAqB,mBAAmB,MAAM,iBAAiB,EAAE;AACpE,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,aAAa,mBAAmB,kBAAkB,6BAA6B,MAAM,iBAAiB;IAC7H;IACD,CAAC;;;AAIN,2BAAU;EACR,MAAM;EACN,UAAU,MAAM;EAChB,UAAU;EACV,UAAU,MAAM;EACjB,CAAC;;;;;;AAOJ,MAAa,4BAA4B,WAAqF;CAE5H,MAAM,SAAS,IAAI,KAA8B;AACjD,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,WAAW,OAAO,IAAI,MAAM,KAAK;AACvC,MAAI,UAAU;AACZ,YAAS,KAAK,MAAM;SACf;AACL,UAAO,IAAI,MAAM,MAAM,CAAC,MAAM,CAAC;;;CAKnC,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,CAAC,MAAM,UAAU,QAAQ;EAClC,MAAM,SAAS,oBAAoB,MAAM,MAAM;AAC/C,MAAI,OAAO,OAAO,CAAE,4BAAW,OAAO,MAAM;AAC5C,YAAU,KAAK,OAAO,MAAM;;AAI9B,WAAU,MAAM,GAAG,MAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC;AAEtD,2BAAU,UAAU;;;;;AAMtB,MAAM,gBAAgB,QAAqB,SAA0B,mBAAmB,IAAI,KAAK,IAAI,OAAO,QAAQ,IAAI,KAAK;;;;;;;AAY7H,MAAa,6BAA6B,cAAuF;CAE/H,MAAM,QAAQ,IAAI,KAA0B;AAC5C,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,OAAO,+BAA+B,SAAS,WAAW;AAChE,QAAM,IAAI,SAAS,MAAM,KAAK;;CAGhC,MAAM,iBAAiB,IAAI,KAA6B;AACxD,MAAK,MAAM,KAAK,WAAW;AACzB,iBAAe,IAAI,EAAE,MAAM,EAAE;;CAG/B,MAAMC,SAA2B,EAAE;CACnC,MAAM,UAAU,IAAI,KAAa;CACjC,MAAM,WAAW,IAAI,KAAa;CAElC,MAAM,SAAS,MAAc,SAA8C;AACzE,MAAI,QAAQ,IAAI,KAAK,CAAE,QAAO;AAE9B,MAAI,SAAS,IAAI,KAAK,EAAE;GAEtB,MAAM,aAAa,KAAK,QAAQ,KAAK;GACrC,MAAM,QAAQ,KAAK,MAAM,WAAW,CAAC,OAAO,KAAK;AACjD,UAAO;IACL,MAAM;IACN,SAAS,8CAA8C,MAAM,KAAK,OAAO;IACzE,eAAe;IAChB;;EAIH,MAAM,WAAW,eAAe,IAAI,KAAK;AACzC,MAAI,CAAC,UAAU;AAEb,WAAQ,IAAI,KAAK;AACjB,UAAO;;AAGT,WAAS,IAAI,KAAK;EAClB,MAAM,OAAO,MAAM,IAAI,KAAK,IAAI,IAAI,KAAK;AAEzC,OAAK,MAAM,OAAO,MAAM;GACtB,MAAM,QAAQ,MAAM,KAAK,CAAC,GAAG,MAAM,KAAK,CAAC;AACzC,OAAI,MAAO,QAAO;;AAGpB,WAAS,OAAO,KAAK;AACrB,UAAQ,IAAI,KAAK;AACjB,SAAO,KAAK,SAAS;AACrB,SAAO;;AAGT,MAAK,MAAM,YAAY,WAAW;EAChC,MAAM,QAAQ,MAAM,SAAS,MAAM,EAAE,CAAC;AACtC,MAAI,MAAO,4BAAW,MAAM;;AAG9B,2BAAU,OAAO;;;;;;AAOnB,MAAM,kCAAkC,eAAwD;CAC9F,MAAM,OAAO,IAAI,KAAa;CAE9B,MAAM,WAAW,SAA2C;AAC1D,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,UAAK,IAAI,IAAI,KAAK;AAClB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO;;;;;AAMT,MAAM,cAAc,QAAqB,SAA0B,OAAO,MAAM,IAAI,KAAK;;;;;;;AAkDzF,MAAa,0BACX,QACA,YACgD;CAChD,MAAM,SAASC,oCAAkB,QAAQ,eAAe;CAGxD,MAAM,aAAa,0BAA0B,OAAO,UAAU;AAC9D,KAAI,WAAW,OAAO,EAAE;AACtB,6BAAW,WAAW,MAAM;;CAE9B,MAAM,kBAAkB,WAAW;CAGnC,MAAM,4BAA4B,IAAI,KAA0C;CAChF,MAAMC,YAAgC,EAAE;AAExC,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,SAAS,kBAAkB,MAAM,QAAQ,0BAA0B;AACzE,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,4BAA0B,IAAI,KAAK,MAAM,OAAO,MAAM,UAAU;AAChE,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAMC,aAAkC,EAAE;AAC1C,MAAK,MAAM,MAAM,OAAO,YAAY;EAClC,MAAM,SAAS,mBAAmB,IAAI,OAAO;AAC7C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,OAAO,MAAM;;AAG/B,2BAAU;EAAE;EAAY;EAAW,CAAC;;;;;AAMtC,MAAM,sBAAsB,IAAqB,WAAuE;CAEtH,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,GAAG,WAAW;EAC5B,MAAM,WAAW,gBAAgB,QAAQ,EAAE,SAAS;AACpD,MAAI,aAAa,MAAM;AACrB,8BAAW;IACT,MAAM;IACN,SAAS,iBAAiB,EAAE,SAAS,iBAAiB,EAAE,KAAK;IAC7D,UAAU,EAAE;IACb,CAAC;;AAEJ,YAAU,KAAK;GAAE,GAAG;GAAG;GAAU,CAAC;;CAIpC,MAAM,uBAAuB,4BAA4B,GAAG,WAAW;AAEvE,2BAAU;EACR,GAAG;EACH;EACA;EACD,CAAC;;;;;;AAOJ,MAAM,qBACJ,MACA,QACA,8BACiD;CAEjD,MAAM,uBAAuB,4BAA4B,KAAK,WAAW;CAGzE,MAAM,qBAAqB,sBAAsB,KAAK,YAAY,KAAK,QAAQ,OAAO;AACtF,KAAI,mBAAmB,OAAO,EAAE;AAC9B,6BAAW,mBAAmB,MAAM;;CAEtC,MAAM,eAAe,mBAAmB;CAGxC,MAAMC,kBAAsC,EAAE;AAC9C,MAAK,MAAM,WAAW,sBAAsB;EAC1C,MAAM,eAAe,0BAA0B,IAAI,QAAQ;AAC3D,MAAI,cAAc;AAChB,mBAAgB,KAAK,GAAG,aAAa;;;CAQzC,MAAMC,YAA6B,CACjC,GAAG,cACH,GAAG,gBAAgB,KAAK,OAAO;EAC7B,MAAM,EAAE;EACR,UAAU,EAAE;EACZ,kBAAkB,EAAE;EACpB,iBAAiB,EAAE;EACnB,UAAU,EAAE;EACb,EAAE,CACJ;CAGD,MAAM,kBAAkB,yBAAyB,UAAU;AAC3D,KAAI,gBAAgB,OAAO,EAAE;AAC3B,6BAAW,gBAAgB,MAAM;;AAGnC,2BAAU;EACR,GAAG;EACH;EACA,WAAW,gBAAgB;EAC5B,CAAC;;;;;AAMJ,MAAM,mBAAmB,QAAqB,aAAyD;AACrG,KAAI,aAAa,QAAQ,SAAS,EAAE;AAClC,SAAO;;AAET,KAAI,WAAW,QAAQ,SAAS,EAAE;AAChC,SAAO;;AAET,KAAI,OAAO,OAAO,IAAI,SAAS,EAAE;AAC/B,SAAO;;AAET,QAAO;;;;;AAMT,MAAM,+BAA+B,eAAgF;CACnH,MAAM,YAAY,IAAI,KAAa;CAEnC,MAAM,WAAW,SAA6D;AAC5E,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,eAAU,IAAI,IAAI,KAAK;AACvB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO,CAAC,GAAG,UAAU;;;;;;;;;ACxwBvB,MAAM,mBAAmB,QAA4B,SAAwD;AAC3G,KAAI,QAAQ;AACV,UAAQ,MAAR;GACE,KAAK,QACH,QAAO,OAAO,eAAe,SAAS;GACxC,KAAK,WACH,QAAO,OAAO,eAAe,YAAY;GAC3C,KAAK,eACH,QAAO,OAAO,eAAe,gBAAgB;;;AAInD,SAAQ,MAAR;EACE,KAAK,QACH,QAAO;EACT,KAAK,WACH,QAAO;EACT,KAAK,eACH,QAAO;;;;;;AAOb,MAAa,iBAAiB,WAA8B,YAA6D;CACvH,MAAMC,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CAKpF,MAAM,aAAa,GAAG,UAAU,KAAK;CACrC,MAAM,gBAAgB,UAAU;AAEhC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,cAAc,cAAc;AACpG,OAAM,KAAK,KAAK,cAAc,WAAW;AACzC,OAAM,KAAK,aAAa,KAAK,UAAU,UAAU,KAAK,CAAC,GAAG;AAG1D,KAAI,UAAU,UAAU,SAAS,GAAG;AAClC,QAAM,KAAK,oBAAoB,cAAc,UAAU,UAAU,CAAC,KAAK;;CAIzE,MAAM,eAAe,gBAAgB,QAAQ,UAAU,KAAK;AAC5D,OAAM,KAAK,+BAA+B;CAC1C,MAAM,mBAAmB,eAAe,UAAU,YAAY,GAAG,UAAU,WAAW,QAAQ,aAAa;AAC3G,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAa,gBAAgB,UAA4B,YAA6D;CACpH,MAAMD,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiBC,oCAAkB,QAAQ,eAAe,GAAG;CACpF,MAAM,eAAe,SAAS,UAAU,SAAS;CAKjD,MAAM,aAAa,GAAG,SAAS,KAAK;CAGpC,MAAM,cAAc,eAAe,mBAAmB;AACtD,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,YAAY,QAAQ;AAC5F,OAAM,KAAK,cAAc,SAAS,OAAO,IAAI;AAG7C,KAAI,cAAc;AAChB,QAAM,KAAK,oBAAoB,cAAc,SAAS,UAAU,CAAC,KAAK;;CAKxE,MAAM,gBAAgB,eAAe,aAAa;AAClD,OAAM,KAAK,gBAAgB,cAAc,SAAS;CAClD,MAAM,mBAAmB,eAAe,SAAS,YAAY,GAAG,SAAS,WAAW,QAAQ,SAAS,OAAO;AAC5G,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAe7B,MAAM,iBAAiB,cAAoD;AACzE,QAAO,UAAU,KAAK,MAAM,WAAW,KAAK,UAAU,EAAE,KAAK,CAAC,IAAI,EAAE,SAAS,GAAG,KAAK,UAAU,EAAE,SAAS,CAAC,GAAG,CAAC,KAAK,KAAK;;;;;;AAO3H,MAAM,kBACJ,YACA,QACA,WACA,QACA,mBACuC;CACvC,MAAM,gBAAgB,IAAI,IAAI,UAAU,KAAK,MAAM,EAAE,KAAK,CAAC;AAC3D,QAAO,uBAAuB,YAAY,QAAQ,eAAe,QAAQ,eAAe;;;;;;AAO1F,MAAM,0BACJ,YACA,QACA,eACA,QACA,mBACuC;CACvC,MAAMD,QAAkB,EAAE;CAG1B,MAAME,kBAA0C,EAAE;CAClD,MAAMC,kBAAqC,EAAE;AAE7C,MAAK,MAAM,OAAO,YAAY;AAC5B,MAAI,IAAI,SAAS,kBAAkB;AACjC,mBAAgB,KAAK,IAAI;SACpB;AACL,mBAAgB,KAAK,IAAI;;;AAK7B,MAAK,MAAM,OAAO,iBAAiB;EACjC,MAAM,SAAS,oBAAoB,KAAK,QAAQ,eAAe,QAAQ,eAAe;AACtF,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,QAAM,KAAK,OAAO,MAAM;;AAI1B,KAAI,gBAAgB,SAAS,GAAG;EAC9B,MAAM,cAAc,2BAA2B,iBAAiB,QAAQ,eAAe,OAAO;AAC9F,MAAI,YAAY,OAAO,EAAE;AACvB,8BAAW,YAAY,MAAM;;AAE/B,QAAM,KAAK,YAAY,MAAM;;AAG/B,2BAAU,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,uBACJ,KACA,QACA,eACA,QACA,mBACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAEnC,SAAQ,IAAI,MAAZ;EACE,KAAK,QACH,QAAO,mBAAmB,KAAK,QAAQ,eAAe,QAAQ,eAAe;EAC/E,KAAK,iBACH,2BAAU,GAAG,QAAQ,KAAK,IAAI,KAAK,oBAAoB;EACzD,KAAK,iBAEH,2BAAU,GAAG;;;;;;;AAQnB,MAAM,8BACJ,iBACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAGnC,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,KAAK,WAAW,IAAI;AACtB,8BAAW;IACT,MAAM;IACN,SAAS;IACV,CAAC;;;AAKN,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,UAAU,CAAC,OAAO,QAAQ,IAAI,KAAK,OAAO,EAAE;GAG9C,IAAI,gBAAgB;AACpB,QAAK,MAAM,GAAG,aAAa,OAAO,QAAQ;AACxC,QAAI,SAAS,QAAQ,IAAI,KAAK,OAAO,EAAE;AACrC,qBAAgB;AAChB;;;AAGJ,OAAI,CAAC,eAAe;AAClB,+BAAW;KACT,MAAM;KACN,SAAS,uCAAuC,KAAK,OAAO;KAC5D,QAAQ,KAAK;KACd,CAAC;;;;CAMR,MAAMC,UAAoB,EAAE;AAC5B,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,eAAe,KAAK,OAAO,SAAS,EAAE;EAE5C,MAAM,eAAe,uBAAuB,KAAK,YAAY,SAAS,GAAG,eAAe,QAAQ,KAAK,OAAO;AAC5G,MAAI,aAAa,OAAO,EAAE;AACxB,8BAAW,aAAa,MAAM;;AAGhC,UAAQ,KAAK,GAAG,eAAe,KAAK,OAAO;EAC7C,aAAa,MAAM;EACnB,aAAa,KAAK;;AAKlB,2BAAU,GAAG,QAAQ;EACrB,QAAQ,KAAK,KAAK,CAAC;EACnB,QAAQ,KAAK;;;;;AAMf,MAAM,sBACJ,OACA,QACA,eACA,QACA,mBACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;CAGnC,MAAM,OAAO,MAAM;CACnB,MAAM,aAAa,MAAM;CACzB,MAAM,UAAU,QAAQ,KAAK,SAAS;CACtC,MAAM,gBAAgB,cAAc,WAAW,SAAS;AAGxD,KAAI,CAAC,WAAW,CAAC,eAAe;AAC9B,4BAAU,GAAG,UAAU,MAAM,KAAK,SAAS;;CAG7C,IAAI,OAAO,GAAG,QAAQ,OAAO,MAAM,KAAK;AAExC,KAAI,SAAS;EACX,MAAM,aAAa,cAAc,MAAM,eAAe,QAAQ,gBAAgB,MAAM,KAAK;AACzF,MAAI,WAAW,OAAO,EAAE;AACtB,8BAAW,WAAW,MAAM;;AAE9B,UAAQ,WAAW;;AAGrB,SAAQ;AAER,KAAI,eAAe;EAEjB,MAAM,qBAAqB,WAAW,MAAM,MAAM,EAAE,SAAS,iBAAiB;EAG9E,MAAM,mBACJ,UAAU,iBAAkB,mBAAmB,QAAQ,gBAAgB,MAAM,KAAK,IAAI,YAAa;AAErG,MAAI,oBAAoB;GAEtB,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,QAAQ,iBAAiB;AAC5G,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ;AACR,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;SACd;AAEL,WAAQ;GACR,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,QAAQ,iBAAiB;AAC5G,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;AAEhC,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;;;AAIvB,SAAQ;AAER,2BAAU,KAAK;;;;;AAUjB,MAAM,kBAAkB,aAA8B;AACpD,QAAO,SAAS,SAAS,KAAK;;;;;;;;;;AAWhC,MAAM,qBAAqB,OAAoB,qBAAkD;AAE/F,KAAI,CAAC,iBAAkB,QAAO;AAG9B,KAAI,CAAC,eAAe,iBAAiB,CAAE,QAAO;AAG9C,KAAI,MAAM,SAAS,WAAY,QAAO;AAGtC,KAAI,MAAM,SAAS,OAAQ,QAAO;AAGlC,KAAI,MAAM,SAAS,OAAQ,QAAO;AAElC,QAAO;;;;;;;;AAST,MAAM,sBAAsB,iBAA4C;CACtE,MAAM,EAAE,UAAU,aAAa;CAG/B,MAAM,YAAY,SAAS,MAAM,oBAAoB;AACrD,KAAI,CAAC,aAAa,CAAC,UAAU,GAAI,QAAO;AACxC,QAAO;EAAE;EAAU,UAAU,UAAU;EAAI;;;;;AAM7C,MAAM,qBACJ,OACA,cACA,eACA,WACuC;CAEvC,MAAM,eAAe,kBAAkB,OAAO,cAAc,SAAS;AAGrE,KAAI,MAAM,SAAS,YAAY,gBAAgB,QAAQ;AACrD,SAAO,mBAAmB,OAAO,aAAa,UAAU,eAAe,QAAQ,aAAa;;AAI9F,KAAI,MAAM,SAAS,UAAU,gBAAgB,QAAQ;EACnD,MAAM,cAAc,mBAAmB,aAAa;AACpD,MAAI,aAAa;GACf,MAAMC,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAMC,WAAS,kBAAkB,GAAG,aAAa,eAAe,OAAO;AACvE,QAAIA,SAAO,OAAO,CAAE,QAAOA;AAC3B,WAAO,KAAKA,SAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;;CAKvC,MAAM,SAAS,UAAU,OAAO,cAAc;AAC9C,KAAI,OAAO,OAAO,CAAE,QAAO;AAG3B,KAAI,cAAc;AAChB,4BAAU,IAAI,OAAO,MAAM,GAAG;;AAGhC,QAAO;;;;;AAMT,MAAM,sBACJ,OACA,eACA,eACA,QACA,gBACuC;AACvC,KAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,4BAAU,cAAc,SAAS,KAAK;;CAGxC,MAAMF,UAAoB,EAAE;AAC5B,MAAK,MAAM,KAAK,MAAM,QAAQ;EAE5B,MAAM,YAAY,kBAAkB,QAAQ,eAAe,EAAE,KAAK;AAClE,MAAI,cAAc,MAAM;AACtB,8BAAW;IACT,MAAM;IACN,SAAS,kBAAkB,EAAE,KAAK,mBAAmB,cAAc;IACnE,UAAU;IACV,WAAW,EAAE;IACd,CAAC;;EAGJ,MAAM,SAAS,kBAAkB,EAAE,OAAO,WAAW,eAAe,OAAO;AAC3E,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,UAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;CAG5C,MAAM,YAAY,KAAK,QAAQ,KAAK,KAAK,CAAC;AAC1C,2BAAU,cAAc,IAAI,UAAU,KAAK,UAAU;;;;;AAUvD,MAAM,iBACJ,MACA,eACA,QACA,gBACA,cACuC;AACvC,KAAI,KAAK,WAAW,GAAG;AACrB,4BAAU,GAAG;;CAGf,MAAMG,aAAuB,EAAE;AAC/B,MAAK,MAAM,OAAO,MAAM;EAEtB,MAAM,eACJ,UAAU,kBAAkB,YAAY,gBAAgB,QAAQ,gBAAgB,WAAW,IAAI,KAAK,GAAG;EAEzG,MAAM,SAAS,kBAAkB,IAAI,OAAO,cAAc,eAAe,OAAO;AAChF,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAE1B,aAAW,KAAK,GAAG,IAAI,KAAK,IAAI,OAAO,QAAQ;;AAEjD,2BAAU,KAAK,WAAW,KAAK,KAAK,CAAC,IAAI;;;;;AAM3C,MAAM,aAAa,OAAoB,kBAAmE;AACxG,SAAQ,MAAM,MAAd;EACE,KAAK;AAEH,OAAI,cAAc,IAAI,MAAM,KAAK,EAAE;AACjC,8BAAU,KAAK,MAAM,OAAO;;AAE9B,8BAAW;IACT,MAAM;IACN,SAAS,cAAc,MAAM,KAAK;IAClC,cAAc,MAAM;IACrB,CAAC;EACJ,KAAK;EACL,KAAK,QACH,2BAAU,MAAM,MAAM;EACxB,KAAK,SACH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,UACH,2BAAU,MAAM,QAAQ,SAAS,QAAQ;EAC3C,KAAK,OACH,2BAAU,OAAO;EACnB,KAAK,OAEH,2BAAU,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,QAAQ;GACX,MAAMF,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,GAAG,cAAc;AAC1C,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,WAAO,KAAK,OAAO,MAAM;;AAE3B,6BAAU,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;EAErC,KAAK,UAAU;AACb,OAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,8BAAU,KAAK;;GAEjB,MAAMD,UAAoB,EAAE;AAC5B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,EAAE,OAAO,cAAc;AAChD,QAAI,OAAO,OAAO,EAAE;AAClB,gCAAW,OAAO,MAAM;;AAE1B,YAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;AAE5C,6BAAU,KAAK,QAAQ,KAAK,KAAK,CAAC,IAAI;;;;;;;ACxkB5C,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,8BAAe,WAAW,EAAE;AAC1B,8BAA+B;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaI,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,mCAAoB,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,4BAAY;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,6BAAU,EAAE,SAAS,CAAC;WACf,OAAO;AACd,8BAAW;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,mBAAmB,YAAoB;CAClD,MAAM,oCAAqB,QAAQ;AACnC,KAAI;AACF,sBAAO,YAAY;GAAE,WAAW;GAAM,OAAO;GAAM,CAAC;AACpD,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,oCAAqB,QAAQ;AAEnC,KAAI;AACF,gDAAkB,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,6BAAc,YAAY,SAAS;AACnC,4BAA8B,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAA+B;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACtBN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,sCAAuB,WAAW;AAExC,KAAI,yBAAY,aAAa,EAAE;AAC7B,6BAAuC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,yCAA4B,cAAc,OAAO;EACvD,MAAM,8BAAiB,aAAa;AACpC,4BAAsC,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,6BAAuC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAuC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,gCAAmB,UAAU;AACnC,2BAAsC,OAAO;;AAG/C,MAAa,cAAc,yCAA8C,SAAS,CAAC,0BAAa,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC9CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,iCAAkB,SAAS;CACjC,MAAM,qCAAsB,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,mCAAoB,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,6BAAc,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,sCAAqBD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,oCAAqB,YAAY,UAAU,2BAAY,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,oCAAqB,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,iCAAkB,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,oCAAqB,aAAa,OAAO,QAAQ;AACvD,MAAI,yBAAY,WAAW,EAAE;AAC3B,8BAAW;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,qCAAsB,aAAa,OAAO,QAAQ;AACxD,OAAI,yBAAY,YAAY,EAAE;AAC5B,+BAAW;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,2BAAW,IAAI,CAAC,GAChC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,8BAAW,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,gCAAiB,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,oBAAoB,IAAI,KAA+B;AAC7D,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,YAAY;AAC3B,qBAAkB,IAAI,YAAY,aAAa,WAAW;;;CAK9D,MAAM,EACJ,MAAM,cACN,aACA,iBACEC,4CAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACA,aAAa,kBAAkB,OAAO,IAAI,oBAAoB;EAC/D,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,2CAAM,8BAAuB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,yDAA2B,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,8BAAW,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,gCAAiB,QAAQ;EAG/B,MAAM,8BAAe,QAAQ,QAAQ;AACrC,8BAAe,QAAQ,EAAE;GACvB,MAAM,eAAe,gBAAgB,QAAQ;AAC7C,OAAI,aAAa,OAAO,EAAE;AACxB,+BAAW,aAAa,MAAM;;;EAOlC,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,+BAAgB,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,+BAAW,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,0DAA4B,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,6BAAW,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,2BAAW,UAAU,CAAC,GACnC,UAAU,QAAQ,4BAAY,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,6BAAW,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,8BAAc,OAAO,GACrB,8BAAc,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,6BAAW,aAAa,MAAM;;AAGhC,2BAAU;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
package/dist/index.d.cts CHANGED
@@ -1,5 +1,5 @@
1
1
  import { ConstDirectiveNode, DocumentNode, EnumValueDefinitionNode, FieldDefinitionNode, InputValueDefinitionNode, NamedTypeNode, TypeNode } from "graphql";
2
- import * as neverthrow0 from "neverthrow";
2
+ import * as neverthrow3 from "neverthrow";
3
3
  import { Result } from "neverthrow";
4
4
  import { TypeFilterConfig } from "@soda-gql/config";
5
5
 
@@ -640,7 +640,7 @@ type CodegenSuccess = {
640
640
  type CodegenResult = Result<CodegenSuccess, CodegenError>;
641
641
  //#endregion
642
642
  //#region packages/codegen/src/inject-template.d.ts
643
- declare const writeInjectTemplate: (outPath: string) => neverthrow0.Err<void, CodegenError> | neverthrow0.Ok<void, CodegenError>;
643
+ declare const writeInjectTemplate: (outPath: string) => neverthrow3.Err<void, CodegenError> | neverthrow3.Ok<void, CodegenError>;
644
644
  declare const getInjectTemplate: () => string;
645
645
  //#endregion
646
646
  //#region packages/codegen/src/runner.d.ts
@@ -651,12 +651,12 @@ declare const runCodegen: (options: CodegenOptions) => Promise<CodegenResult>;
651
651
  * Load a single schema file.
652
652
  * @internal Use loadSchema for public API.
653
653
  */
654
- declare const loadSingleSchema: (schemaPath: string) => neverthrow0.Err<DocumentNode, CodegenError> | neverthrow0.Ok<DocumentNode, CodegenError>;
654
+ declare const loadSingleSchema: (schemaPath: string) => neverthrow3.Err<DocumentNode, CodegenError> | neverthrow3.Ok<DocumentNode, CodegenError>;
655
655
  /**
656
656
  * Load and merge multiple schema files into a single DocumentNode.
657
657
  * Uses GraphQL's concatAST to combine definitions from all files.
658
658
  */
659
- declare const loadSchema: (schemaPaths: readonly string[]) => neverthrow0.Err<DocumentNode, CodegenError> | neverthrow0.Ok<DocumentNode, CodegenError>;
659
+ declare const loadSchema: (schemaPaths: readonly string[]) => neverthrow3.Err<DocumentNode, CodegenError> | neverthrow3.Ok<DocumentNode, CodegenError>;
660
660
  declare const hashSchema: (document: DocumentNode) => string;
661
661
  //#endregion
662
662
  export { type CodegenCliCommand, type CodegenError, type CodegenFormat, type CodegenInjectConfig, type CodegenOptions, type CodegenResult, type CodegenSchemaConfig, type CodegenSuccess, EmitOptions, EnrichedFragment, EnrichedOperation, EnrichedVariable, GeneratedFile, GraphqlCompatError, GraphqlCompatOptions, InferredVariable, ParseResult, ParsedArgument, ParsedFieldSelection, ParsedFragment, ParsedFragmentSpread, ParsedInlineFragment, ParsedObjectField, ParsedOperation, ParsedSelection, ParsedValue, ParsedVariable, TransformOptions, TransformResult, TypeInfo, collectVariableUsages, emitFragment, emitOperation, getArgumentType, getFieldReturnType, getInputFieldType, hashSchema, inferVariablesFromUsages, isModifierAssignable, loadSchema, mergeModifiers, mergeVariableUsages, parseGraphqlFile, parseGraphqlSource, parseTypeNode, runCodegen, sortFragmentsByDependency, transformParsedGraphql, writeInjectTemplate };