@soda-gql/codegen 0.11.11 → 0.11.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +672 -2
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +284 -2
- package/dist/index.d.cts.map +1 -1
- package/dist/index.d.mts +288 -6
- package/dist/index.d.mts.map +1 -1
- package/dist/index.mjs +668 -4
- package/dist/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/index.mjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.mjs","names":["esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","defsPaths: string[]","defaultBundler"],"sources":["../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;AAMA,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,aAAa,QAAQ,QAAQ;AAEnC,KAAI;AACF,MAAI,WAAW,WAAW,EAAE;AAC1B,UAAO,IAAwB;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,YAAU,QAAQ,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,gBAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,SAAO,GAAuB,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAwB;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaA,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,YAAY,QAAQ,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,SAAM,MAAM;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,UAAO,GAAG,EAAE,SAAS,CAAC;WACf,OAAO;AACd,UAAO,IAAI;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,aAAa,QAAQ,QAAQ;AAEnC,KAAI;AACF,YAAU,QAAQ,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,gBAAc,YAAY,SAAS;AACnC,SAAO,GAAuB,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAwB;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACPN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,eAAe,QAAQ,WAAW;AAExC,KAAI,CAAC,WAAW,aAAa,EAAE;AAC7B,SAAO,IAAgC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,eAAe,aAAa,cAAc,OAAO;EACvD,MAAM,WAAW,MAAM,aAAa;AACpC,SAAO,GAA+B,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAgC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAgC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,SAAS,UAAU,UAAU;AACnC,QAAO,GAA+B,OAAO;;AAG/C,MAAa,cAAc,aAAmC,WAAW,SAAS,CAAC,OAAO,MAAM,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC/CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,UAAU,QAAQ,SAAS;CACjC,MAAM,aAAa,SAAS,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,YAAY,QAAQ,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,KAAK,SAAS,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,eAAa,QAAQD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,YAAY,SAAS,YAAY,UAAU,GAAG,SAAS,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,aAAa,QAAQ,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,UAAU,QAAQ,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,aAAa,QAAQ,aAAa,OAAO,QAAQ;AACvD,MAAI,CAAC,WAAW,WAAW,EAAE;AAC3B,UAAO,IAAI;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,cAAc,QAAQ,aAAa,OAAO,QAAQ;AACxD,OAAI,CAAC,WAAW,YAAY,EAAE;AAC5B,WAAO,IAAI;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,QAAQ,GAAG,IAAI,CAAC,GAChC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,SAAS,QAAQ,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,SAAS,QAAQ,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,EACJ,MAAM,cACN,aACA,iBACE,0BAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACD,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,MAAM,OAAO,6BAAgB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,cAAc,KAAK,QAAQ,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,UAAO,IAAI,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,SAAS,QAAQ,QAAQ;EAK/B,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,WAAW,KAAK,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,WAAO,IAAI,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,eAAe,KAAK,QAAQ,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,SAAO,IAAI,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,SAAO,IAAI,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,WAAW,GAAG,OAAO,GACrB,UAAU,IAAI,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,SAAO,IAAI,aAAa,MAAM;;AAGhC,QAAO,GAAG;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
|
|
1
|
+
{"version":3,"file":"index.mjs","names":["lines: string[]","inlineFragments: ParsedInlineFragment[]","otherSelections: ParsedSelection[]","entries: string[]","argEntries: string[]","values: string[]","operations: ParsedOperation[]","fragments: ParsedFragment[]","variables: ParsedVariable[]","levels: TypeLevel[]","operations: EnrichedOperation[]","fragments: EnrichedFragment[]","variables: EnrichedVariable[]","esbuildBundler: Bundler","result: T[][]","imports: string[]","files: Array<{ relativePath: string; content: string }>","importPaths: Record<DefinitionCategory, string>","categories: DefinitionCategory[]","documents: DocumentNode[]","extensionMap: Record<string, string>","withPrefix","currentExt","schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }>","defsPaths: string[]","defaultBundler"],"sources":["../src/graphql-compat/emitter.ts","../src/graphql-compat/parser.ts","../src/graphql-compat/transformer.ts","../src/inject-template.ts","../src/bundler/esbuild.ts","../src/defs-generator.ts","../src/file.ts","../src/schema.ts","../src/runner.ts"],"sourcesContent":["/**\n * Emitter for generating TypeScript compat code from enriched operations.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport type { EnrichedFragment, EnrichedOperation, EnrichedVariable } from \"./transformer\";\nimport type { GraphqlCompatError, ParsedArgument, ParsedInlineFragment, ParsedSelection, ParsedValue } from \"./types\";\n\n/**\n * Schema index for type lookups.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Options for code emission.\n */\nexport type EmitOptions = {\n /** Schema name to use in gql.schemaName() call */\n readonly schemaName: string;\n /** Import path for graphql-system module */\n readonly graphqlSystemPath: string;\n /** Map of fragment name to its import path (relative) */\n readonly fragmentImports?: ReadonlyMap<string, string>;\n /** Schema document for type lookups (required for inline fragment support) */\n readonly schemaDocument?: DocumentNode;\n};\n\n/**\n * Emit TypeScript code for an operation.\n */\nexport const emitOperation = (operation: EnrichedOperation, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Generate imports\n lines.push(`import { gql } from \"${options.graphqlSystemPath}\";`);\n\n // Add fragment imports if needed\n if (operation.fragmentDependencies.length > 0 && options.fragmentImports) {\n for (const fragName of operation.fragmentDependencies) {\n const importPath = options.fragmentImports.get(fragName);\n if (importPath) {\n lines.push(`import { ${fragName}Fragment } from \"${importPath}\";`);\n }\n }\n }\n\n lines.push(\"\");\n\n // Generate export\n const exportName = `${operation.name}Compat`;\n const operationType = operation.kind;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ ${operationType}, $var }) =>`);\n lines.push(` ${operationType}.compat({`);\n lines.push(` name: ${JSON.stringify(operation.name)},`);\n\n // Variables\n if (operation.variables.length > 0) {\n lines.push(` variables: { ${emitVariables(operation.variables)} },`);\n }\n\n // Fields\n lines.push(` fields: ({ f, $ }) => ({`);\n const fieldLinesResult = emitSelections(operation.selections, 3, operation.variables, schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit TypeScript code for a fragment.\n */\nexport const emitFragment = (fragment: EnrichedFragment, options: EmitOptions): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n const schema = options.schemaDocument ? createSchemaIndex(options.schemaDocument) : null;\n\n // Generate imports\n lines.push(`import { gql } from \"${options.graphqlSystemPath}\";`);\n\n // Add fragment imports if needed\n if (fragment.fragmentDependencies.length > 0 && options.fragmentImports) {\n for (const fragName of fragment.fragmentDependencies) {\n const importPath = options.fragmentImports.get(fragName);\n if (importPath) {\n lines.push(`import { ${fragName}Fragment } from \"${importPath}\";`);\n }\n }\n }\n\n lines.push(\"\");\n\n // Generate export\n const exportName = `${fragment.name}Fragment`;\n\n lines.push(`export const ${exportName} = gql.${options.schemaName}(({ fragment }) =>`);\n lines.push(` fragment.${fragment.onType}({`);\n\n // Fields\n lines.push(` fields: ({ f }) => ({`);\n const fieldLinesResult = emitSelections(fragment.selections, 3, [], schema);\n if (fieldLinesResult.isErr()) {\n return err(fieldLinesResult.error);\n }\n lines.push(fieldLinesResult.value);\n lines.push(` }),`);\n\n lines.push(` }),`);\n lines.push(`);`);\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit variable definitions.\n */\nconst emitVariables = (variables: readonly EnrichedVariable[]): string => {\n return variables.map((v) => `...$var(${JSON.stringify(v.name)}).${v.typeName}(${JSON.stringify(v.modifier)})`).join(\", \");\n};\n\n/**\n * Emit field selections (public API).\n * Converts EnrichedVariable[] to Set<string> and delegates to internal implementation.\n */\nconst emitSelections = (\n selections: readonly ParsedSelection[],\n indent: number,\n variables: readonly EnrichedVariable[],\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const variableNames = new Set(variables.map((v) => v.name));\n return emitSelectionsInternal(selections, indent, variableNames, schema);\n};\n\n/**\n * Internal implementation for emitting field selections.\n * Takes variableNames as Set<string> for recursive calls.\n */\nconst emitSelectionsInternal = (\n selections: readonly ParsedSelection[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const lines: string[] = [];\n\n // Separate inline fragments from other selections\n const inlineFragments: ParsedInlineFragment[] = [];\n const otherSelections: ParsedSelection[] = [];\n\n for (const sel of selections) {\n if (sel.kind === \"inlineFragment\") {\n inlineFragments.push(sel);\n } else {\n otherSelections.push(sel);\n }\n }\n\n // Emit regular selections (fields and fragment spreads)\n for (const sel of otherSelections) {\n const result = emitSingleSelection(sel, indent, variableNames, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n lines.push(result.value);\n }\n\n // Emit grouped inline fragments as union selections\n if (inlineFragments.length > 0) {\n const unionResult = emitInlineFragmentsAsUnion(inlineFragments, indent, variableNames, schema);\n if (unionResult.isErr()) {\n return err(unionResult.error);\n }\n lines.push(unionResult.value);\n }\n\n return ok(lines.join(\"\\n\"));\n};\n\n/**\n * Emit a single selection (field or fragment spread).\n */\nconst emitSingleSelection = (\n sel: ParsedSelection,\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n switch (sel.kind) {\n case \"field\":\n return emitFieldSelection(sel, indent, variableNames, schema);\n case \"fragmentSpread\":\n return ok(`${padding}...${sel.name}Fragment.spread(),`);\n case \"inlineFragment\":\n // This should not happen as inline fragments are handled separately\n return ok(\"\");\n }\n};\n\n/**\n * Emit inline fragments grouped as a union selection.\n * Format: { TypeA: ({ f }) => ({ ...fields }), TypeB: ({ f }) => ({ ...fields }) }\n */\nconst emitInlineFragmentsAsUnion = (\n inlineFragments: readonly ParsedInlineFragment[],\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Validate inline fragments have type conditions\n for (const frag of inlineFragments) {\n if (frag.onType === \"\") {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_WITHOUT_TYPE\",\n message: \"Inline fragments without type condition are not supported. Use `... on TypeName { }` syntax.\",\n });\n }\n }\n\n // Validate all inline fragments are on union types (not interfaces)\n for (const frag of inlineFragments) {\n if (schema && !schema.objects.has(frag.onType)) {\n // If it's not a known object type, it might be an interface\n // Check if any union contains this type as a member\n let isUnionMember = false;\n for (const [, unionDef] of schema.unions) {\n if (unionDef.members.has(frag.onType)) {\n isUnionMember = true;\n break;\n }\n }\n if (!isUnionMember) {\n return err({\n code: \"GRAPHQL_INLINE_FRAGMENT_ON_INTERFACE\",\n message: `Inline fragments on interface type \"${frag.onType}\" are not supported. Use union types instead.`,\n onType: frag.onType,\n });\n }\n }\n }\n\n // Build union member entries\n const entries: string[] = [];\n for (const frag of inlineFragments) {\n const innerPadding = \" \".repeat(indent + 1);\n const fieldsResult = emitSelectionsInternal(frag.selections, indent + 2, variableNames, schema);\n if (fieldsResult.isErr()) {\n return err(fieldsResult.error);\n }\n\n entries.push(`${innerPadding}${frag.onType}: ({ f }) => ({\n${fieldsResult.value}\n${innerPadding}}),`);\n }\n\n // Emit as spread with union callback: ...f.fieldName()({ Type: ... })\n // Note: This assumes the parent field handles the union - we emit just the union object\n return ok(`${padding}...({\n${entries.join(\"\\n\")}\n${padding}}),`);\n};\n\n/**\n * Emit a single field selection.\n */\nconst emitFieldSelection = (\n field: ParsedSelection & { kind: \"field\" },\n indent: number,\n variableNames: Set<string>,\n schema: SchemaIndex | null,\n): Result<string, GraphqlCompatError> => {\n const padding = \" \".repeat(indent);\n\n // Extract optional fields for type narrowing\n const args = field.arguments;\n const selections = field.selections;\n const hasArgs = args && args.length > 0;\n const hasSelections = selections && selections.length > 0;\n\n let line = `${padding}...f.${field.name}(`;\n\n if (hasArgs) {\n const argsResult = emitArguments(args, variableNames);\n if (argsResult.isErr()) {\n return err(argsResult.error);\n }\n line += argsResult.value;\n }\n\n line += \")\";\n\n if (hasSelections) {\n // Check if selections contain inline fragments (union field)\n const hasInlineFragments = selections.some((s) => s.kind === \"inlineFragment\");\n\n if (hasInlineFragments) {\n // Union field: emit with union callback pattern\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += \"({\\n\";\n line += `${nestedResult.value}\\n`;\n line += `${padding}})`;\n } else {\n // Regular nested selections\n line += \"(({ f }) => ({\\n\";\n const nestedResult = emitSelectionsInternal(selections, indent + 1, variableNames, schema);\n if (nestedResult.isErr()) {\n return err(nestedResult.error);\n }\n line += `${nestedResult.value}\\n`;\n line += `${padding}}))`;\n }\n }\n\n line += \",\";\n\n return ok(line);\n};\n\n/**\n * Emit field arguments.\n */\nconst emitArguments = (args: readonly ParsedArgument[], variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n if (args.length === 0) {\n return ok(\"\");\n }\n\n const argEntries: string[] = [];\n for (const arg of args) {\n const result = emitValue(arg.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n argEntries.push(`${arg.name}: ${result.value}`);\n }\n return ok(`{ ${argEntries.join(\", \")} }`);\n};\n\n/**\n * Emit a value (literal or variable reference).\n */\nconst emitValue = (value: ParsedValue, variableNames: Set<string>): Result<string, GraphqlCompatError> => {\n switch (value.kind) {\n case \"variable\":\n // Check if it's a declared variable\n if (variableNames.has(value.name)) {\n return ok(`$.${value.name}`);\n }\n return err({\n code: \"GRAPHQL_UNDECLARED_VARIABLE\",\n message: `Variable \"$${value.name}\" is not declared in the operation`,\n variableName: value.name,\n });\n case \"int\":\n case \"float\":\n return ok(value.value);\n case \"string\":\n return ok(JSON.stringify(value.value));\n case \"boolean\":\n return ok(value.value ? \"true\" : \"false\");\n case \"null\":\n return ok(\"null\");\n case \"enum\":\n // Enums are emitted as string literals in soda-gql\n return ok(JSON.stringify(value.value));\n case \"list\": {\n const values: string[] = [];\n for (const v of value.values) {\n const result = emitValue(v, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n values.push(result.value);\n }\n return ok(`[${values.join(\", \")}]`);\n }\n case \"object\": {\n if (value.fields.length === 0) {\n return ok(\"{}\");\n }\n const entries: string[] = [];\n for (const f of value.fields) {\n const result = emitValue(f.value, variableNames);\n if (result.isErr()) {\n return err(result.error);\n }\n entries.push(`${f.name}: ${result.value}`);\n }\n return ok(`{ ${entries.join(\", \")} }`);\n }\n }\n};\n","/**\n * Parser for .graphql operation files.\n * Extracts operations and fragments from GraphQL documents.\n * @module\n */\n\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport {\n type ArgumentNode,\n type DocumentNode,\n type FieldNode,\n type FragmentDefinitionNode,\n type FragmentSpreadNode,\n type InlineFragmentNode,\n Kind,\n type OperationDefinitionNode,\n parse,\n type SelectionNode,\n type TypeNode,\n type ValueNode,\n type VariableDefinitionNode,\n} from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\n\nimport type {\n GraphqlCompatError,\n ParsedArgument,\n ParsedFieldSelection,\n ParsedFragment,\n ParsedFragmentSpread,\n ParsedInlineFragment,\n ParsedOperation,\n ParsedSelection,\n ParsedValue,\n ParsedVariable,\n ParseResult,\n} from \"./types\";\n\n/**\n * Parse a single .graphql file and extract operations and fragments.\n */\nexport const parseGraphqlFile = (filePath: string): Result<ParseResult, GraphqlCompatError> => {\n const resolvedPath = resolve(filePath);\n\n if (!existsSync(resolvedPath)) {\n return err({\n code: \"GRAPHQL_FILE_NOT_FOUND\",\n message: `GraphQL file not found at ${resolvedPath}`,\n filePath: resolvedPath,\n });\n }\n\n try {\n const source = readFileSync(resolvedPath, \"utf8\");\n const document = parse(source);\n return ok(extractFromDocument(document, resolvedPath));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: resolvedPath,\n });\n }\n};\n\n/**\n * Parse GraphQL source string directly.\n */\nexport const parseGraphqlSource = (source: string, sourceFile: string): Result<ParseResult, GraphqlCompatError> => {\n try {\n const document = parse(source);\n return ok(extractFromDocument(document, sourceFile));\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err({\n code: \"GRAPHQL_PARSE_ERROR\",\n message: `GraphQL parse error: ${message}`,\n filePath: sourceFile,\n });\n }\n};\n\n/**\n * Extract operations and fragments from a parsed GraphQL document.\n */\nconst extractFromDocument = (document: DocumentNode, sourceFile: string): ParseResult => {\n const operations: ParsedOperation[] = [];\n const fragments: ParsedFragment[] = [];\n\n for (const definition of document.definitions) {\n if (definition.kind === Kind.OPERATION_DEFINITION) {\n const operation = extractOperation(definition, sourceFile);\n if (operation) {\n operations.push(operation);\n }\n } else if (definition.kind === Kind.FRAGMENT_DEFINITION) {\n fragments.push(extractFragment(definition, sourceFile));\n }\n }\n\n return { operations, fragments };\n};\n\n/**\n * Extract a single operation from an OperationDefinitionNode.\n */\nconst extractOperation = (node: OperationDefinitionNode, sourceFile: string): ParsedOperation | null => {\n // Anonymous operations are not supported\n if (!node.name) {\n return null;\n }\n\n const variables: ParsedVariable[] = (node.variableDefinitions ?? []).map(extractVariable);\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n kind: node.operation,\n name: node.name.value,\n variables,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a fragment from a FragmentDefinitionNode.\n */\nconst extractFragment = (node: FragmentDefinitionNode, sourceFile: string): ParsedFragment => {\n const selections = extractSelections(node.selectionSet.selections);\n\n return {\n name: node.name.value,\n onType: node.typeCondition.name.value,\n selections,\n sourceFile,\n };\n};\n\n/**\n * Extract a variable definition.\n */\nconst extractVariable = (node: VariableDefinitionNode): ParsedVariable => {\n const { typeName, modifier } = parseTypeNode(node.type);\n const defaultValue = node.defaultValue ? extractValue(node.defaultValue) : undefined;\n\n return {\n name: node.variable.name.value,\n typeName,\n modifier,\n // Type kind will be determined later when we have schema context\n typeKind: \"scalar\",\n defaultValue,\n };\n};\n\n/**\n * Parse a GraphQL TypeNode into type name and modifier.\n *\n * Format: inner nullability + list modifiers\n * - Inner: `!` (non-null) or `?` (nullable)\n * - List: `[]!` (non-null list) or `[]?` (nullable list)\n */\nexport const parseTypeNode = (node: TypeNode): { typeName: string; modifier: string } => {\n type TypeLevel = { kind: \"list\" | \"named\"; nonNull: boolean };\n const levels: TypeLevel[] = [];\n\n const collect = (n: TypeNode, nonNull: boolean): string => {\n if (n.kind === Kind.NON_NULL_TYPE) {\n return collect(n.type, true);\n }\n if (n.kind === Kind.LIST_TYPE) {\n levels.push({ kind: \"list\", nonNull });\n return collect(n.type, false);\n }\n levels.push({ kind: \"named\", nonNull });\n return n.name.value;\n };\n\n const typeName = collect(node, false);\n\n // Build modifier from levels (reverse order)\n let modifier = \"?\";\n for (const level of levels.slice().reverse()) {\n if (level.kind === \"named\") {\n modifier = level.nonNull ? \"!\" : \"?\";\n continue;\n }\n const listSuffix = level.nonNull ? \"[]!\" : \"[]?\";\n modifier = `${modifier}${listSuffix}`;\n }\n\n return { typeName, modifier };\n};\n\n/**\n * Extract selections from a SelectionSet.\n */\nconst extractSelections = (selections: readonly SelectionNode[]): ParsedSelection[] => {\n return selections.map(extractSelection);\n};\n\n/**\n * Extract a single selection.\n */\nconst extractSelection = (node: SelectionNode): ParsedSelection => {\n switch (node.kind) {\n case Kind.FIELD:\n return extractFieldSelection(node);\n case Kind.FRAGMENT_SPREAD:\n return extractFragmentSpread(node);\n case Kind.INLINE_FRAGMENT:\n return extractInlineFragment(node);\n }\n};\n\n/**\n * Extract a field selection.\n */\nconst extractFieldSelection = (node: FieldNode): ParsedFieldSelection => {\n const args = node.arguments?.length ? node.arguments.map(extractArgument) : undefined;\n const selections = node.selectionSet ? extractSelections(node.selectionSet.selections) : undefined;\n\n return {\n kind: \"field\",\n name: node.name.value,\n alias: node.alias?.value,\n arguments: args,\n selections,\n };\n};\n\n/**\n * Extract a fragment spread.\n */\nconst extractFragmentSpread = (node: FragmentSpreadNode): ParsedFragmentSpread => {\n return {\n kind: \"fragmentSpread\",\n name: node.name.value,\n };\n};\n\n/**\n * Extract an inline fragment.\n */\nconst extractInlineFragment = (node: InlineFragmentNode): ParsedInlineFragment => {\n return {\n kind: \"inlineFragment\",\n onType: node.typeCondition?.name.value ?? \"\",\n selections: extractSelections(node.selectionSet.selections),\n };\n};\n\n/**\n * Extract an argument.\n */\nconst extractArgument = (node: ArgumentNode): ParsedArgument => {\n return {\n name: node.name.value,\n value: extractValue(node.value),\n };\n};\n\n/**\n * Assert unreachable code path (for exhaustiveness checks).\n */\nconst assertUnreachable = (value: never): never => {\n throw new Error(`Unexpected value: ${JSON.stringify(value)}`);\n};\n\n/**\n * Extract a value (literal or variable reference).\n */\nconst extractValue = (node: ValueNode): ParsedValue => {\n switch (node.kind) {\n case Kind.VARIABLE:\n return { kind: \"variable\", name: node.name.value };\n case Kind.INT:\n return { kind: \"int\", value: node.value };\n case Kind.FLOAT:\n return { kind: \"float\", value: node.value };\n case Kind.STRING:\n return { kind: \"string\", value: node.value };\n case Kind.BOOLEAN:\n return { kind: \"boolean\", value: node.value };\n case Kind.NULL:\n return { kind: \"null\" };\n case Kind.ENUM:\n return { kind: \"enum\", value: node.value };\n case Kind.LIST:\n return { kind: \"list\", values: node.values.map(extractValue) };\n case Kind.OBJECT:\n return {\n kind: \"object\",\n fields: node.fields.map((field) => ({\n name: field.name.value,\n value: extractValue(field.value),\n })),\n };\n default:\n return assertUnreachable(node);\n }\n};\n","/**\n * Transformer for enriching parsed GraphQL operations with schema information.\n * @module\n */\n\nimport type { DocumentNode } from \"graphql\";\nimport { err, ok, type Result } from \"neverthrow\";\nimport { createSchemaIndex } from \"../generator\";\nimport type { GraphqlCompatError, ParsedFragment, ParsedOperation, ParsedVariable, ParseResult } from \"./types\";\n\n/**\n * Schema index type extracted from generator.\n */\ntype SchemaIndex = ReturnType<typeof createSchemaIndex>;\n\n/**\n * Built-in GraphQL scalar types.\n */\nconst builtinScalarTypes = new Set([\"ID\", \"String\", \"Int\", \"Float\", \"Boolean\"]);\n\n/**\n * Check if a type name is a scalar type.\n */\nconst isScalarName = (schema: SchemaIndex, name: string): boolean => builtinScalarTypes.has(name) || schema.scalars.has(name);\n\n/**\n * Check if a type name is an enum type.\n */\nconst isEnumName = (schema: SchemaIndex, name: string): boolean => schema.enums.has(name);\n\n/**\n * Enriched operation with resolved type information.\n */\nexport type EnrichedOperation = Omit<ParsedOperation, \"variables\"> & {\n readonly variables: readonly EnrichedVariable[];\n /** Fragment names used in this operation (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched fragment with resolved type information.\n */\nexport type EnrichedFragment = ParsedFragment & {\n /** Fragment names used in this fragment (for imports) */\n readonly fragmentDependencies: readonly string[];\n};\n\n/**\n * Enriched variable with resolved type kind.\n */\nexport type EnrichedVariable = Omit<ParsedVariable, \"typeKind\"> & {\n readonly typeKind: \"scalar\" | \"enum\" | \"input\";\n};\n\n/**\n * Result of transforming parsed operations.\n */\nexport type TransformResult = {\n readonly operations: readonly EnrichedOperation[];\n readonly fragments: readonly EnrichedFragment[];\n};\n\n/**\n * Options for transformation.\n */\nexport type TransformOptions = {\n /** Schema document for type resolution */\n readonly schemaDocument: DocumentNode;\n};\n\n/**\n * Transform parsed operations/fragments by enriching them with schema information.\n *\n * This resolves variable type kinds (scalar, enum, input) and collects\n * fragment dependencies.\n */\nexport const transformParsedGraphql = (\n parsed: ParseResult,\n options: TransformOptions,\n): Result<TransformResult, GraphqlCompatError> => {\n const schema = createSchemaIndex(options.schemaDocument);\n\n // Transform operations\n const operations: EnrichedOperation[] = [];\n for (const op of parsed.operations) {\n const result = transformOperation(op, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n operations.push(result.value);\n }\n\n // Transform fragments\n const fragments: EnrichedFragment[] = [];\n for (const frag of parsed.fragments) {\n const result = transformFragment(frag, schema);\n if (result.isErr()) {\n return err(result.error);\n }\n fragments.push(result.value);\n }\n\n return ok({ operations, fragments });\n};\n\n/**\n * Transform a single operation.\n */\nconst transformOperation = (op: ParsedOperation, schema: SchemaIndex): Result<EnrichedOperation, GraphqlCompatError> => {\n // Resolve variable type kinds\n const variables: EnrichedVariable[] = [];\n for (const v of op.variables) {\n const typeKind = resolveTypeKind(schema, v.typeName);\n if (typeKind === null) {\n return err({\n code: \"GRAPHQL_UNKNOWN_TYPE\",\n message: `Unknown type \"${v.typeName}\" in variable \"${v.name}\"`,\n typeName: v.typeName,\n });\n }\n variables.push({ ...v, typeKind });\n }\n\n // Collect fragment dependencies\n const fragmentDependencies = collectFragmentDependencies(op.selections);\n\n return ok({\n ...op,\n variables,\n fragmentDependencies,\n });\n};\n\n/**\n * Transform a single fragment.\n */\nconst transformFragment = (frag: ParsedFragment, _schema: SchemaIndex): Result<EnrichedFragment, GraphqlCompatError> => {\n // Collect fragment dependencies (fragments used within this fragment)\n const fragmentDependencies = collectFragmentDependencies(frag.selections);\n\n return ok({\n ...frag,\n fragmentDependencies,\n });\n};\n\n/**\n * Resolve the type kind for a type name.\n */\nconst resolveTypeKind = (schema: SchemaIndex, typeName: string): \"scalar\" | \"enum\" | \"input\" | null => {\n if (isScalarName(schema, typeName)) {\n return \"scalar\";\n }\n if (isEnumName(schema, typeName)) {\n return \"enum\";\n }\n if (schema.inputs.has(typeName)) {\n return \"input\";\n }\n return null;\n};\n\n/**\n * Collect fragment names used in selections (recursively).\n */\nconst collectFragmentDependencies = (selections: readonly import(\"./types\").ParsedSelection[]): readonly string[] => {\n const fragments = new Set<string>();\n\n const collect = (sels: readonly import(\"./types\").ParsedSelection[]): void => {\n for (const sel of sels) {\n switch (sel.kind) {\n case \"fragmentSpread\":\n fragments.add(sel.name);\n break;\n case \"field\":\n if (sel.selections) {\n collect(sel.selections);\n }\n break;\n case \"inlineFragment\":\n collect(sel.selections);\n break;\n }\n }\n };\n\n collect(selections);\n return [...fragments];\n};\n","import { existsSync, mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nconst templateContents = `\\\nimport { defineScalar } from \"@soda-gql/core\";\n\nexport const scalar = {\n ...defineScalar<\"ID\", string, string>(\"ID\"),\n ...defineScalar<\"String\", string, string>(\"String\"),\n ...defineScalar<\"Int\", number, number>(\"Int\"),\n ...defineScalar<\"Float\", number, number>(\"Float\"),\n ...defineScalar<\"Boolean\", boolean, boolean>(\"Boolean\"),\n} as const;\n`;\n\nexport const writeInjectTemplate = (outPath: string) => {\n const targetPath = resolve(outPath);\n\n try {\n if (existsSync(targetPath)) {\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_EXISTS\",\n message: `Inject module already exists: ${targetPath}`,\n outPath: targetPath,\n });\n }\n\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, `${templateContents}\\n`);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"INJECT_TEMPLATE_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n\nexport const getInjectTemplate = (): string => `${templateContents}\\n`;\n","import { extname } from \"node:path\";\nimport { build } from \"esbuild\";\nimport { err, ok } from \"neverthrow\";\nimport type { Bundler } from \"./types\";\n\nexport const esbuildBundler: Bundler = {\n name: \"esbuild\",\n bundle: async ({ sourcePath, external }) => {\n try {\n const sourceExt = extname(sourcePath);\n const baseName = sourcePath.slice(0, -sourceExt.length);\n const cjsPath = `${baseName}.cjs`;\n\n await build({\n entryPoints: [sourcePath],\n outfile: cjsPath,\n format: \"cjs\",\n platform: \"node\",\n bundle: true,\n external: [...external],\n sourcemap: false,\n minify: false,\n treeShaking: false,\n });\n\n return ok({ cjsPath });\n } catch (error) {\n return err({\n code: \"EMIT_FAILED\" as const,\n message: `[esbuild] Failed to bundle: ${error instanceof Error ? error.message : String(error)}`,\n outPath: sourcePath,\n });\n }\n },\n};\n","/**\n * Definition file generator for split codegen.\n * Generates separate files for each definition category (enums, inputs, objects, unions).\n */\n\nexport type DefinitionCategory = \"enums\" | \"inputs\" | \"objects\" | \"unions\";\n\nexport type DefinitionVar = {\n readonly name: string;\n readonly code: string;\n};\n\n/**\n * Split an array into chunks of the specified size.\n */\nexport const chunkArray = <T>(array: readonly T[], size: number): T[][] => {\n if (size <= 0) {\n return [Array.from(array)];\n }\n\n const result: T[][] = [];\n for (let i = 0; i < array.length; i += size) {\n result.push(array.slice(i, i + size));\n }\n\n return result;\n};\n\n/**\n * Determine if chunking is needed based on the number of definitions.\n */\nexport const needsChunking = (vars: readonly DefinitionVar[], chunkSize: number): boolean => {\n return vars.length > chunkSize;\n};\n\ntype DefinitionFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a single definition file content.\n */\nexport const generateDefinitionFile = (options: DefinitionFileOptions): string => {\n const { category, vars, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} definitions (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} definitions\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkFileOptions = {\n readonly category: DefinitionCategory;\n readonly schemaName: string;\n readonly vars: readonly DefinitionVar[];\n readonly chunkIndex: number;\n readonly needsDefineEnum: boolean;\n};\n\n/**\n * Generate a chunk file content.\n */\nexport const generateChunkFile = (options: ChunkFileOptions): string => {\n const { category, vars, chunkIndex, needsDefineEnum } = options;\n\n if (vars.length === 0) {\n return `/**\n * ${category} chunk ${chunkIndex} (empty)\n * @generated by @soda-gql/codegen\n */\n`;\n }\n\n const imports: string[] = [];\n if (needsDefineEnum && category === \"enums\") {\n imports.push('import { defineEnum } from \"@soda-gql/core\";');\n }\n\n const importsBlock = imports.length > 0 ? `${imports.join(\"\\n\")}\\n\\n` : \"\";\n const exportStatements = vars.map((v) => `export ${v.code}`).join(\"\\n\");\n\n return `/**\n * ${category} chunk ${chunkIndex}\n * @generated by @soda-gql/codegen\n */\n${importsBlock}${exportStatements}\n`;\n};\n\ntype ChunkIndexOptions = {\n readonly category: DefinitionCategory;\n readonly chunkCount: number;\n readonly varNames: readonly string[];\n};\n\n/**\n * Generate the index file that re-exports all chunks.\n */\nexport const generateChunkIndex = (options: ChunkIndexOptions): string => {\n const { category, chunkCount } = options;\n\n const reExports = Array.from({ length: chunkCount }, (_, i) => `export * from \"./chunk-${i}\";`).join(\"\\n\");\n\n return `/**\n * ${category} index (re-exports all chunks)\n * @generated by @soda-gql/codegen\n */\n${reExports}\n`;\n};\n\nexport type ChunkedDefinitionFiles = {\n readonly indexContent: string;\n readonly chunks: ReadonlyArray<{\n readonly chunkIndex: number;\n readonly content: string;\n readonly varNames: readonly string[];\n }>;\n};\n\n/**\n * Generate chunked definition files.\n */\nexport const generateChunkedDefinitionFiles = (\n category: DefinitionCategory,\n schemaName: string,\n vars: readonly DefinitionVar[],\n chunkSize: number,\n): ChunkedDefinitionFiles => {\n const chunks = chunkArray(vars, chunkSize);\n const needsDefineEnum = category === \"enums\";\n\n const chunkContents = chunks.map((chunkVars, chunkIndex) => ({\n chunkIndex,\n content: generateChunkFile({\n category,\n schemaName,\n vars: chunkVars,\n chunkIndex,\n needsDefineEnum,\n }),\n varNames: chunkVars.map((v) => v.name),\n }));\n\n const allVarNames = vars.map((v) => v.name);\n const indexContent = generateChunkIndex({\n category,\n chunkCount: chunks.length,\n varNames: allVarNames,\n });\n\n return {\n indexContent,\n chunks: chunkContents,\n };\n};\n\ntype DefsDirectoryStructure = {\n readonly files: ReadonlyArray<{\n readonly relativePath: string;\n readonly content: string;\n }>;\n readonly importPaths: Record<DefinitionCategory, string>;\n};\n\nexport type CategoryVars = {\n readonly enums: readonly DefinitionVar[];\n readonly inputs: readonly DefinitionVar[];\n readonly objects: readonly DefinitionVar[];\n readonly unions: readonly DefinitionVar[];\n};\n\n/**\n * Generate the complete _defs directory structure.\n */\nexport const generateDefsStructure = (\n schemaName: string,\n categoryVars: CategoryVars,\n chunkSize: number,\n): DefsDirectoryStructure => {\n const files: Array<{ relativePath: string; content: string }> = [];\n const importPaths: Record<DefinitionCategory, string> = {\n enums: \"./_defs/enums\",\n inputs: \"./_defs/inputs\",\n objects: \"./_defs/objects\",\n unions: \"./_defs/unions\",\n };\n\n const categories: DefinitionCategory[] = [\"enums\", \"inputs\", \"objects\", \"unions\"];\n\n for (const category of categories) {\n const vars = categoryVars[category];\n const needsDefineEnum = category === \"enums\";\n\n if (needsChunking(vars, chunkSize)) {\n // Generate chunked files\n const chunked = generateChunkedDefinitionFiles(category, schemaName, vars, chunkSize);\n\n // Update import path to point to the directory (which has index.ts)\n importPaths[category] = `./_defs/${category}`;\n\n // Add index file\n files.push({\n relativePath: `_defs/${category}/index.ts`,\n content: chunked.indexContent,\n });\n\n // Add chunk files\n for (const chunk of chunked.chunks) {\n files.push({\n relativePath: `_defs/${category}/chunk-${chunk.chunkIndex}.ts`,\n content: chunk.content,\n });\n }\n } else {\n // Generate single file\n const content = generateDefinitionFile({\n category,\n schemaName,\n vars,\n needsDefineEnum,\n });\n\n files.push({\n relativePath: `_defs/${category}.ts`,\n content,\n });\n }\n }\n\n return { files, importPaths };\n};\n","import { mkdirSync, writeFileSync } from \"node:fs\";\nimport { dirname, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\nexport const writeModule = (outPath: string, contents: string) => {\n const targetPath = resolve(outPath);\n\n try {\n mkdirSync(dirname(targetPath), { recursive: true });\n writeFileSync(targetPath, contents);\n return ok<void, CodegenError>(undefined);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<void, CodegenError>({\n code: \"EMIT_FAILED\",\n message,\n outPath: targetPath,\n });\n }\n};\n","import { createHash } from \"node:crypto\";\nimport { existsSync, readFileSync } from \"node:fs\";\nimport { resolve } from \"node:path\";\nimport { concatAST, type DocumentNode, parse, print } from \"graphql\";\nimport { err, ok } from \"neverthrow\";\n\nimport type { CodegenError } from \"./types\";\n\n/**\n * Load a single schema file.\n * @internal Use loadSchema for public API.\n */\nexport const loadSingleSchema = (schemaPath: string) => {\n const resolvedPath = resolve(schemaPath);\n\n if (!existsSync(resolvedPath)) {\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_NOT_FOUND\",\n message: `Schema file not found at ${resolvedPath}`,\n schemaPath: resolvedPath,\n });\n }\n\n try {\n const schemaSource = readFileSync(resolvedPath, \"utf8\");\n const document = parse(schemaSource);\n return ok<DocumentNode, CodegenError>(document);\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return err<DocumentNode, CodegenError>({\n code: \"SCHEMA_INVALID\",\n message: `SchemaValidationError: ${message}`,\n schemaPath: resolvedPath,\n });\n }\n};\n\n/**\n * Load and merge multiple schema files into a single DocumentNode.\n * Uses GraphQL's concatAST to combine definitions from all files.\n */\nexport const loadSchema = (schemaPaths: readonly string[]) => {\n const documents: DocumentNode[] = [];\n\n for (const schemaPath of schemaPaths) {\n const result = loadSingleSchema(schemaPath);\n if (result.isErr()) {\n return err<DocumentNode, CodegenError>(result.error);\n }\n documents.push(result.value);\n }\n\n // Merge all documents into one\n const merged = concatAST(documents);\n return ok<DocumentNode, CodegenError>(merged);\n};\n\nexport const hashSchema = (document: DocumentNode): string => createHash(\"sha256\").update(print(document)).digest(\"hex\");\n","import { existsSync } from \"node:fs\";\nimport { basename, dirname, extname, join, relative, resolve } from \"node:path\";\nimport { err, ok } from \"neverthrow\";\nimport { defaultBundler } from \"./bundler\";\nimport { generateDefsStructure } from \"./defs-generator\";\nimport { writeModule } from \"./file\";\nimport { generateMultiSchemaModule } from \"./generator\";\nimport { hashSchema, loadSchema } from \"./schema\";\nimport type { CodegenOptions, CodegenResult, CodegenSuccess } from \"./types\";\n\nconst extensionMap: Record<string, string> = {\n \".ts\": \".js\",\n \".tsx\": \".js\",\n \".mts\": \".mjs\",\n \".cts\": \".cjs\",\n \".js\": \".js\",\n \".mjs\": \".mjs\",\n \".cjs\": \".cjs\",\n};\n\ntype ImportSpecifierOptions = {\n includeExtension?: boolean;\n};\n\nconst toImportSpecifier = (fromPath: string, targetPath: string, options?: ImportSpecifierOptions): string => {\n const fromDir = dirname(fromPath);\n const normalized = relative(fromDir, targetPath).replace(/\\\\/g, \"/\");\n const sourceExt = extname(targetPath);\n\n // When includeExtension is false (default), strip the extension entirely\n if (!options?.includeExtension) {\n if (normalized.length === 0) {\n return `./${basename(targetPath, sourceExt)}`;\n }\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n const currentExt = extname(withPrefix);\n return currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n }\n\n // When includeExtension is true, convert to runtime extension\n const runtimeExt = extensionMap[sourceExt] ?? sourceExt;\n\n if (normalized.length === 0) {\n const base = runtimeExt !== sourceExt ? basename(targetPath, sourceExt) : basename(targetPath);\n return `./${base}${runtimeExt}`;\n }\n\n const withPrefix = normalized.startsWith(\".\") ? normalized : `./${normalized}`;\n if (!runtimeExt) {\n return withPrefix;\n }\n if (withPrefix.endsWith(runtimeExt)) {\n return withPrefix;\n }\n\n const currentExt = extname(withPrefix);\n const withoutExt = currentExt ? withPrefix.slice(0, -currentExt.length) : withPrefix;\n return `${withoutExt}${runtimeExt}`;\n};\n\nexport const runCodegen = async (options: CodegenOptions): Promise<CodegenResult> => {\n const outPath = resolve(options.outPath);\n const importSpecifierOptions = { includeExtension: options.importExtension };\n\n // Validate that all schema and inject files exist\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const scalarPath = resolve(schemaConfig.inject.scalars);\n if (!existsSync(scalarPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Scalar module not found for schema '${schemaName}': ${scalarPath}`,\n injectPath: scalarPath,\n });\n }\n\n if (schemaConfig.inject.adapter) {\n const adapterPath = resolve(schemaConfig.inject.adapter);\n if (!existsSync(adapterPath)) {\n return err({\n code: \"INJECT_MODULE_NOT_FOUND\",\n message: `Adapter module not found for schema '${schemaName}': ${adapterPath}`,\n injectPath: adapterPath,\n });\n }\n }\n }\n\n // Load all schemas\n const schemas = new Map<string, import(\"graphql\").DocumentNode>();\n const schemaHashes: Record<string, { schemaHash: string; objects: number; enums: number; inputs: number; unions: number }> = {};\n\n for (const [name, schemaConfig] of Object.entries(options.schemas)) {\n const result = await loadSchema(schemaConfig.schema).match(\n (doc) => Promise.resolve(ok(doc)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (result.isErr()) {\n return err(result.error);\n }\n\n schemas.set(name, result.value);\n }\n\n // Build injection config for each schema\n const injectionConfig = new Map<\n string,\n {\n scalarImportPath: string;\n adapterImportPath?: string;\n }\n >();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n const injectConfig = schemaConfig.inject;\n\n injectionConfig.set(schemaName, {\n scalarImportPath: toImportSpecifier(outPath, resolve(injectConfig.scalars), importSpecifierOptions),\n ...(injectConfig.adapter\n ? { adapterImportPath: toImportSpecifier(outPath, resolve(injectConfig.adapter), importSpecifierOptions) }\n : {}),\n });\n }\n\n // Build defaultInputDepth and inputDepthOverrides config for each schema\n const defaultInputDepthConfig = new Map<string, number>();\n const inputDepthOverridesConfig = new Map<string, Readonly<Record<string, number>>>();\n\n for (const [schemaName, schemaConfig] of Object.entries(options.schemas)) {\n if (schemaConfig.defaultInputDepth !== undefined && schemaConfig.defaultInputDepth !== 3) {\n defaultInputDepthConfig.set(schemaName, schemaConfig.defaultInputDepth);\n }\n if (schemaConfig.inputDepthOverrides && Object.keys(schemaConfig.inputDepthOverrides).length > 0) {\n inputDepthOverridesConfig.set(schemaName, schemaConfig.inputDepthOverrides);\n }\n }\n\n // Get chunkSize config (default: 100)\n const chunkSize = options.chunkSize ?? 100;\n\n // Generate multi-schema module (this becomes _internal.ts content)\n const {\n code: internalCode,\n injectsCode,\n categoryVars,\n } = generateMultiSchemaModule(schemas, {\n injection: injectionConfig,\n defaultInputDepth: defaultInputDepthConfig.size > 0 ? defaultInputDepthConfig : undefined,\n inputDepthOverrides: inputDepthOverridesConfig.size > 0 ? inputDepthOverridesConfig : undefined,\n chunkSize,\n });\n\n // Generate index.ts wrapper (simple re-export from _internal)\n const indexCode = `/**\n * Generated by @soda-gql/codegen\n * @module\n * @generated\n */\nexport * from \"./_internal\";\n`;\n\n // Calculate individual schema stats and hashes\n for (const [name, document] of schemas.entries()) {\n const schemaIndex = (await import(\"./generator\")).createSchemaIndex(document);\n const objects = Array.from(schemaIndex.objects.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const enums = Array.from(schemaIndex.enums.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const inputs = Array.from(schemaIndex.inputs.keys()).filter((n) => !n.startsWith(\"__\")).length;\n const unions = Array.from(schemaIndex.unions.keys()).filter((n) => !n.startsWith(\"__\")).length;\n\n schemaHashes[name] = {\n schemaHash: hashSchema(document),\n objects,\n enums,\n inputs,\n unions,\n };\n }\n\n // Write _internal-injects.ts (adapter imports only, referenced by both _internal.ts and prebuilt)\n const injectsPath = join(dirname(outPath), \"_internal-injects.ts\");\n if (injectsCode) {\n const injectsWriteResult = await writeModule(injectsPath, injectsCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (injectsWriteResult.isErr()) {\n return err(injectsWriteResult.error);\n }\n }\n\n // Write _defs/ files (always enabled)\n const defsPaths: string[] = [];\n if (categoryVars) {\n const outDir = dirname(outPath);\n\n // Merge all schema categoryVars into a single combined structure\n // This ensures all definitions from all schemas go into the same defs files\n type DefinitionVar = { name: string; code: string };\n const combinedVars = {\n enums: [] as DefinitionVar[],\n inputs: [] as DefinitionVar[],\n objects: [] as DefinitionVar[],\n unions: [] as DefinitionVar[],\n };\n\n for (const vars of Object.values(categoryVars)) {\n combinedVars.enums.push(...vars.enums);\n combinedVars.inputs.push(...vars.inputs);\n combinedVars.objects.push(...vars.objects);\n combinedVars.unions.push(...vars.unions);\n }\n\n // Generate defs structure for all schemas combined\n const defsStructure = generateDefsStructure(\"combined\", combinedVars, chunkSize);\n\n for (const file of defsStructure.files) {\n const filePath = join(outDir, file.relativePath);\n\n // writeModule handles directory creation internally via mkdirSync\n const writeResult = await writeModule(filePath, file.content).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (writeResult.isErr()) {\n return err(writeResult.error);\n }\n\n defsPaths.push(filePath);\n }\n }\n\n // Write _internal.ts (implementation)\n const internalPath = join(dirname(outPath), \"_internal.ts\");\n const internalWriteResult = await writeModule(internalPath, internalCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (internalWriteResult.isErr()) {\n return err(internalWriteResult.error);\n }\n\n // Write index.ts (re-export wrapper)\n const indexWriteResult = await writeModule(outPath, indexCode).match(\n () => Promise.resolve(ok(undefined)),\n (error) => Promise.resolve(err(error)),\n );\n\n if (indexWriteResult.isErr()) {\n return err(indexWriteResult.error);\n }\n\n // Bundle the generated module\n const bundleOutcome = await defaultBundler.bundle({\n sourcePath: outPath,\n external: [\"@soda-gql/core\", \"@soda-gql/runtime\"],\n });\n const bundleResult = bundleOutcome.match(\n (result) => ok(result),\n (error) => err(error),\n );\n\n if (bundleResult.isErr()) {\n return err(bundleResult.error);\n }\n\n return ok({\n schemas: schemaHashes,\n outPath,\n internalPath,\n injectsPath,\n cjsPath: bundleResult.value.cjsPath,\n ...(defsPaths.length > 0 ? { defsPaths } : {}),\n } satisfies CodegenSuccess);\n};\n"],"mappings":";;;;;;;;;;;;AAiCA,MAAa,iBAAiB,WAA8B,YAA6D;CACvH,MAAMA,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiB,kBAAkB,QAAQ,eAAe,GAAG;AAGpF,OAAM,KAAK,wBAAwB,QAAQ,kBAAkB,IAAI;AAGjE,KAAI,UAAU,qBAAqB,SAAS,KAAK,QAAQ,iBAAiB;AACxE,OAAK,MAAM,YAAY,UAAU,sBAAsB;GACrD,MAAM,aAAa,QAAQ,gBAAgB,IAAI,SAAS;AACxD,OAAI,YAAY;AACd,UAAM,KAAK,YAAY,SAAS,mBAAmB,WAAW,IAAI;;;;AAKxE,OAAM,KAAK,GAAG;CAGd,MAAM,aAAa,GAAG,UAAU,KAAK;CACrC,MAAM,gBAAgB,UAAU;AAEhC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,MAAM,cAAc,cAAc;AACpG,OAAM,KAAK,KAAK,cAAc,WAAW;AACzC,OAAM,KAAK,aAAa,KAAK,UAAU,UAAU,KAAK,CAAC,GAAG;AAG1D,KAAI,UAAU,UAAU,SAAS,GAAG;AAClC,QAAM,KAAK,oBAAoB,cAAc,UAAU,UAAU,CAAC,KAAK;;AAIzE,OAAM,KAAK,+BAA+B;CAC1C,MAAM,mBAAmB,eAAe,UAAU,YAAY,GAAG,UAAU,WAAW,OAAO;AAC7F,KAAI,iBAAiB,OAAO,EAAE;AAC5B,SAAO,IAAI,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,QAAO,GAAG,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAa,gBAAgB,UAA4B,YAA6D;CACpH,MAAMA,QAAkB,EAAE;CAC1B,MAAM,SAAS,QAAQ,iBAAiB,kBAAkB,QAAQ,eAAe,GAAG;AAGpF,OAAM,KAAK,wBAAwB,QAAQ,kBAAkB,IAAI;AAGjE,KAAI,SAAS,qBAAqB,SAAS,KAAK,QAAQ,iBAAiB;AACvE,OAAK,MAAM,YAAY,SAAS,sBAAsB;GACpD,MAAM,aAAa,QAAQ,gBAAgB,IAAI,SAAS;AACxD,OAAI,YAAY;AACd,UAAM,KAAK,YAAY,SAAS,mBAAmB,WAAW,IAAI;;;;AAKxE,OAAM,KAAK,GAAG;CAGd,MAAM,aAAa,GAAG,SAAS,KAAK;AAEpC,OAAM,KAAK,gBAAgB,WAAW,SAAS,QAAQ,WAAW,oBAAoB;AACtF,OAAM,KAAK,cAAc,SAAS,OAAO,IAAI;AAG7C,OAAM,KAAK,4BAA4B;CACvC,MAAM,mBAAmB,eAAe,SAAS,YAAY,GAAG,EAAE,EAAE,OAAO;AAC3E,KAAI,iBAAiB,OAAO,EAAE;AAC5B,SAAO,IAAI,iBAAiB,MAAM;;AAEpC,OAAM,KAAK,iBAAiB,MAAM;AAClC,OAAM,KAAK,UAAU;AAErB,OAAM,KAAK,QAAQ;AACnB,OAAM,KAAK,KAAK;AAEhB,QAAO,GAAG,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,iBAAiB,cAAmD;AACxE,QAAO,UAAU,KAAK,MAAM,WAAW,KAAK,UAAU,EAAE,KAAK,CAAC,IAAI,EAAE,SAAS,GAAG,KAAK,UAAU,EAAE,SAAS,CAAC,GAAG,CAAC,KAAK,KAAK;;;;;;AAO3H,MAAM,kBACJ,YACA,QACA,WACA,WACuC;CACvC,MAAM,gBAAgB,IAAI,IAAI,UAAU,KAAK,MAAM,EAAE,KAAK,CAAC;AAC3D,QAAO,uBAAuB,YAAY,QAAQ,eAAe,OAAO;;;;;;AAO1E,MAAM,0BACJ,YACA,QACA,eACA,WACuC;CACvC,MAAMA,QAAkB,EAAE;CAG1B,MAAMC,kBAA0C,EAAE;CAClD,MAAMC,kBAAqC,EAAE;AAE7C,MAAK,MAAM,OAAO,YAAY;AAC5B,MAAI,IAAI,SAAS,kBAAkB;AACjC,mBAAgB,KAAK,IAAI;SACpB;AACL,mBAAgB,KAAK,IAAI;;;AAK7B,MAAK,MAAM,OAAO,iBAAiB;EACjC,MAAM,SAAS,oBAAoB,KAAK,QAAQ,eAAe,OAAO;AACtE,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAE1B,QAAM,KAAK,OAAO,MAAM;;AAI1B,KAAI,gBAAgB,SAAS,GAAG;EAC9B,MAAM,cAAc,2BAA2B,iBAAiB,QAAQ,eAAe,OAAO;AAC9F,MAAI,YAAY,OAAO,EAAE;AACvB,UAAO,IAAI,YAAY,MAAM;;AAE/B,QAAM,KAAK,YAAY,MAAM;;AAG/B,QAAO,GAAG,MAAM,KAAK,KAAK,CAAC;;;;;AAM7B,MAAM,uBACJ,KACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAEnC,SAAQ,IAAI,MAAZ;EACE,KAAK,QACH,QAAO,mBAAmB,KAAK,QAAQ,eAAe,OAAO;EAC/D,KAAK,iBACH,QAAO,GAAG,GAAG,QAAQ,KAAK,IAAI,KAAK,oBAAoB;EACzD,KAAK,iBAEH,QAAO,GAAG,GAAG;;;;;;;AAQnB,MAAM,8BACJ,iBACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;AAGnC,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,KAAK,WAAW,IAAI;AACtB,UAAO,IAAI;IACT,MAAM;IACN,SAAS;IACV,CAAC;;;AAKN,MAAK,MAAM,QAAQ,iBAAiB;AAClC,MAAI,UAAU,CAAC,OAAO,QAAQ,IAAI,KAAK,OAAO,EAAE;GAG9C,IAAI,gBAAgB;AACpB,QAAK,MAAM,GAAG,aAAa,OAAO,QAAQ;AACxC,QAAI,SAAS,QAAQ,IAAI,KAAK,OAAO,EAAE;AACrC,qBAAgB;AAChB;;;AAGJ,OAAI,CAAC,eAAe;AAClB,WAAO,IAAI;KACT,MAAM;KACN,SAAS,uCAAuC,KAAK,OAAO;KAC5D,QAAQ,KAAK;KACd,CAAC;;;;CAMR,MAAMC,UAAoB,EAAE;AAC5B,MAAK,MAAM,QAAQ,iBAAiB;EAClC,MAAM,eAAe,KAAK,OAAO,SAAS,EAAE;EAC5C,MAAM,eAAe,uBAAuB,KAAK,YAAY,SAAS,GAAG,eAAe,OAAO;AAC/F,MAAI,aAAa,OAAO,EAAE;AACxB,UAAO,IAAI,aAAa,MAAM;;AAGhC,UAAQ,KAAK,GAAG,eAAe,KAAK,OAAO;EAC7C,aAAa,MAAM;EACnB,aAAa,KAAK;;AAKlB,QAAO,GAAG,GAAG,QAAQ;EACrB,QAAQ,KAAK,KAAK,CAAC;EACnB,QAAQ,KAAK;;;;;AAMf,MAAM,sBACJ,OACA,QACA,eACA,WACuC;CACvC,MAAM,UAAU,KAAK,OAAO,OAAO;CAGnC,MAAM,OAAO,MAAM;CACnB,MAAM,aAAa,MAAM;CACzB,MAAM,UAAU,QAAQ,KAAK,SAAS;CACtC,MAAM,gBAAgB,cAAc,WAAW,SAAS;CAExD,IAAI,OAAO,GAAG,QAAQ,OAAO,MAAM,KAAK;AAExC,KAAI,SAAS;EACX,MAAM,aAAa,cAAc,MAAM,cAAc;AACrD,MAAI,WAAW,OAAO,EAAE;AACtB,UAAO,IAAI,WAAW,MAAM;;AAE9B,UAAQ,WAAW;;AAGrB,SAAQ;AAER,KAAI,eAAe;EAEjB,MAAM,qBAAqB,WAAW,MAAM,MAAM,EAAE,SAAS,iBAAiB;AAE9E,MAAI,oBAAoB;GAEtB,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,WAAO,IAAI,aAAa,MAAM;;AAEhC,WAAQ;AACR,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;SACd;AAEL,WAAQ;GACR,MAAM,eAAe,uBAAuB,YAAY,SAAS,GAAG,eAAe,OAAO;AAC1F,OAAI,aAAa,OAAO,EAAE;AACxB,WAAO,IAAI,aAAa,MAAM;;AAEhC,WAAQ,GAAG,aAAa,MAAM;AAC9B,WAAQ,GAAG,QAAQ;;;AAIvB,SAAQ;AAER,QAAO,GAAG,KAAK;;;;;AAMjB,MAAM,iBAAiB,MAAiC,kBAAmE;AACzH,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO,GAAG,GAAG;;CAGf,MAAMC,aAAuB,EAAE;AAC/B,MAAK,MAAM,OAAO,MAAM;EACtB,MAAM,SAAS,UAAU,IAAI,OAAO,cAAc;AAClD,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAE1B,aAAW,KAAK,GAAG,IAAI,KAAK,IAAI,OAAO,QAAQ;;AAEjD,QAAO,GAAG,KAAK,WAAW,KAAK,KAAK,CAAC,IAAI;;;;;AAM3C,MAAM,aAAa,OAAoB,kBAAmE;AACxG,SAAQ,MAAM,MAAd;EACE,KAAK;AAEH,OAAI,cAAc,IAAI,MAAM,KAAK,EAAE;AACjC,WAAO,GAAG,KAAK,MAAM,OAAO;;AAE9B,UAAO,IAAI;IACT,MAAM;IACN,SAAS,cAAc,MAAM,KAAK;IAClC,cAAc,MAAM;IACrB,CAAC;EACJ,KAAK;EACL,KAAK,QACH,QAAO,GAAG,MAAM,MAAM;EACxB,KAAK,SACH,QAAO,GAAG,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,UACH,QAAO,GAAG,MAAM,QAAQ,SAAS,QAAQ;EAC3C,KAAK,OACH,QAAO,GAAG,OAAO;EACnB,KAAK,OAEH,QAAO,GAAG,KAAK,UAAU,MAAM,MAAM,CAAC;EACxC,KAAK,QAAQ;GACX,MAAMC,SAAmB,EAAE;AAC3B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,GAAG,cAAc;AAC1C,QAAI,OAAO,OAAO,EAAE;AAClB,YAAO,IAAI,OAAO,MAAM;;AAE1B,WAAO,KAAK,OAAO,MAAM;;AAE3B,UAAO,GAAG,IAAI,OAAO,KAAK,KAAK,CAAC,GAAG;;EAErC,KAAK,UAAU;AACb,OAAI,MAAM,OAAO,WAAW,GAAG;AAC7B,WAAO,GAAG,KAAK;;GAEjB,MAAMF,UAAoB,EAAE;AAC5B,QAAK,MAAM,KAAK,MAAM,QAAQ;IAC5B,MAAM,SAAS,UAAU,EAAE,OAAO,cAAc;AAChD,QAAI,OAAO,OAAO,EAAE;AAClB,YAAO,IAAI,OAAO,MAAM;;AAE1B,YAAQ,KAAK,GAAG,EAAE,KAAK,IAAI,OAAO,QAAQ;;AAE5C,UAAO,GAAG,KAAK,QAAQ,KAAK,KAAK,CAAC,IAAI;;;;;;;;;;;;;;;AC1W5C,MAAa,oBAAoB,aAA8D;CAC7F,MAAM,eAAe,QAAQ,SAAS;AAEtC,KAAI,CAAC,WAAW,aAAa,EAAE;AAC7B,SAAO,IAAI;GACT,MAAM;GACN,SAAS,6BAA6B;GACtC,UAAU;GACX,CAAC;;AAGJ,KAAI;EACF,MAAM,SAAS,aAAa,cAAc,OAAO;EACjD,MAAM,WAAW,MAAM,OAAO;AAC9B,SAAO,GAAG,oBAAoB,UAAU,aAAa,CAAC;UAC/C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAI;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAa,sBAAsB,QAAgB,eAAgE;AACjH,KAAI;EACF,MAAM,WAAW,MAAM,OAAO;AAC9B,SAAO,GAAG,oBAAoB,UAAU,WAAW,CAAC;UAC7C,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAI;GACT,MAAM;GACN,SAAS,wBAAwB;GACjC,UAAU;GACX,CAAC;;;;;;AAON,MAAM,uBAAuB,UAAwB,eAAoC;CACvF,MAAMG,aAAgC,EAAE;CACxC,MAAMC,YAA8B,EAAE;AAEtC,MAAK,MAAM,cAAc,SAAS,aAAa;AAC7C,MAAI,WAAW,SAAS,KAAK,sBAAsB;GACjD,MAAM,YAAY,iBAAiB,YAAY,WAAW;AAC1D,OAAI,WAAW;AACb,eAAW,KAAK,UAAU;;aAEnB,WAAW,SAAS,KAAK,qBAAqB;AACvD,aAAU,KAAK,gBAAgB,YAAY,WAAW,CAAC;;;AAI3D,QAAO;EAAE;EAAY;EAAW;;;;;AAMlC,MAAM,oBAAoB,MAA+B,eAA+C;AAEtG,KAAI,CAAC,KAAK,MAAM;AACd,SAAO;;CAGT,MAAMC,aAA+B,KAAK,uBAAuB,EAAE,EAAE,IAAI,gBAAgB;CACzF,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK;EACX,MAAM,KAAK,KAAK;EAChB;EACA;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,MAA8B,eAAuC;CAC5F,MAAM,aAAa,kBAAkB,KAAK,aAAa,WAAW;AAElE,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,QAAQ,KAAK,cAAc,KAAK;EAChC;EACA;EACD;;;;;AAMH,MAAM,mBAAmB,SAAiD;CACxE,MAAM,EAAE,UAAU,aAAa,cAAc,KAAK,KAAK;CACvD,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,aAAa,GAAG;AAE3E,QAAO;EACL,MAAM,KAAK,SAAS,KAAK;EACzB;EACA;EAEA,UAAU;EACV;EACD;;;;;;;;;AAUH,MAAa,iBAAiB,SAA2D;CAEvF,MAAMC,SAAsB,EAAE;CAE9B,MAAM,WAAW,GAAa,YAA6B;AACzD,MAAI,EAAE,SAAS,KAAK,eAAe;AACjC,UAAO,QAAQ,EAAE,MAAM,KAAK;;AAE9B,MAAI,EAAE,SAAS,KAAK,WAAW;AAC7B,UAAO,KAAK;IAAE,MAAM;IAAQ;IAAS,CAAC;AACtC,UAAO,QAAQ,EAAE,MAAM,MAAM;;AAE/B,SAAO,KAAK;GAAE,MAAM;GAAS;GAAS,CAAC;AACvC,SAAO,EAAE,KAAK;;CAGhB,MAAM,WAAW,QAAQ,MAAM,MAAM;CAGrC,IAAI,WAAW;AACf,MAAK,MAAM,SAAS,OAAO,OAAO,CAAC,SAAS,EAAE;AAC5C,MAAI,MAAM,SAAS,SAAS;AAC1B,cAAW,MAAM,UAAU,MAAM;AACjC;;EAEF,MAAM,aAAa,MAAM,UAAU,QAAQ;AAC3C,aAAW,GAAG,WAAW;;AAG3B,QAAO;EAAE;EAAU;EAAU;;;;;AAM/B,MAAM,qBAAqB,eAA4D;AACrF,QAAO,WAAW,IAAI,iBAAiB;;;;;AAMzC,MAAM,oBAAoB,SAAyC;AACjE,SAAQ,KAAK,MAAb;EACE,KAAK,KAAK,MACR,QAAO,sBAAsB,KAAK;EACpC,KAAK,KAAK,gBACR,QAAO,sBAAsB,KAAK;EACpC,KAAK,KAAK,gBACR,QAAO,sBAAsB,KAAK;;;;;;AAOxC,MAAM,yBAAyB,SAA0C;CACvE,MAAM,OAAO,KAAK,WAAW,SAAS,KAAK,UAAU,IAAI,gBAAgB,GAAG;CAC5E,MAAM,aAAa,KAAK,eAAe,kBAAkB,KAAK,aAAa,WAAW,GAAG;AAEzF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EAChB,OAAO,KAAK,OAAO;EACnB,WAAW;EACX;EACD;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,MAAM,KAAK,KAAK;EACjB;;;;;AAMH,MAAM,yBAAyB,SAAmD;AAChF,QAAO;EACL,MAAM;EACN,QAAQ,KAAK,eAAe,KAAK,SAAS;EAC1C,YAAY,kBAAkB,KAAK,aAAa,WAAW;EAC5D;;;;;AAMH,MAAM,mBAAmB,SAAuC;AAC9D,QAAO;EACL,MAAM,KAAK,KAAK;EAChB,OAAO,aAAa,KAAK,MAAM;EAChC;;;;;AAMH,MAAM,qBAAqB,UAAwB;AACjD,OAAM,IAAI,MAAM,qBAAqB,KAAK,UAAU,MAAM,GAAG;;;;;AAM/D,MAAM,gBAAgB,SAAiC;AACrD,SAAQ,KAAK,MAAb;EACE,KAAK,KAAK,SACR,QAAO;GAAE,MAAM;GAAY,MAAM,KAAK,KAAK;GAAO;EACpD,KAAK,KAAK,IACR,QAAO;GAAE,MAAM;GAAO,OAAO,KAAK;GAAO;EAC3C,KAAK,KAAK,MACR,QAAO;GAAE,MAAM;GAAS,OAAO,KAAK;GAAO;EAC7C,KAAK,KAAK,OACR,QAAO;GAAE,MAAM;GAAU,OAAO,KAAK;GAAO;EAC9C,KAAK,KAAK,QACR,QAAO;GAAE,MAAM;GAAW,OAAO,KAAK;GAAO;EAC/C,KAAK,KAAK,KACR,QAAO,EAAE,MAAM,QAAQ;EACzB,KAAK,KAAK,KACR,QAAO;GAAE,MAAM;GAAQ,OAAO,KAAK;GAAO;EAC5C,KAAK,KAAK,KACR,QAAO;GAAE,MAAM;GAAQ,QAAQ,KAAK,OAAO,IAAI,aAAa;GAAE;EAChE,KAAK,KAAK,OACR,QAAO;GACL,MAAM;GACN,QAAQ,KAAK,OAAO,KAAK,WAAW;IAClC,MAAM,MAAM,KAAK;IACjB,OAAO,aAAa,MAAM,MAAM;IACjC,EAAE;GACJ;EACH,QACE,QAAO,kBAAkB,KAAK;;;;;;;;;AC3RpC,MAAM,qBAAqB,IAAI,IAAI;CAAC;CAAM;CAAU;CAAO;CAAS;CAAU,CAAC;;;;AAK/E,MAAM,gBAAgB,QAAqB,SAA0B,mBAAmB,IAAI,KAAK,IAAI,OAAO,QAAQ,IAAI,KAAK;;;;AAK7H,MAAM,cAAc,QAAqB,SAA0B,OAAO,MAAM,IAAI,KAAK;;;;;;;AAgDzF,MAAa,0BACX,QACA,YACgD;CAChD,MAAM,SAAS,kBAAkB,QAAQ,eAAe;CAGxD,MAAMC,aAAkC,EAAE;AAC1C,MAAK,MAAM,MAAM,OAAO,YAAY;EAClC,MAAM,SAAS,mBAAmB,IAAI,OAAO;AAC7C,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAE1B,aAAW,KAAK,OAAO,MAAM;;CAI/B,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,QAAQ,OAAO,WAAW;EACnC,MAAM,SAAS,kBAAkB,MAAM,OAAO;AAC9C,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAE1B,YAAU,KAAK,OAAO,MAAM;;AAG9B,QAAO,GAAG;EAAE;EAAY;EAAW,CAAC;;;;;AAMtC,MAAM,sBAAsB,IAAqB,WAAuE;CAEtH,MAAMC,YAAgC,EAAE;AACxC,MAAK,MAAM,KAAK,GAAG,WAAW;EAC5B,MAAM,WAAW,gBAAgB,QAAQ,EAAE,SAAS;AACpD,MAAI,aAAa,MAAM;AACrB,UAAO,IAAI;IACT,MAAM;IACN,SAAS,iBAAiB,EAAE,SAAS,iBAAiB,EAAE,KAAK;IAC7D,UAAU,EAAE;IACb,CAAC;;AAEJ,YAAU,KAAK;GAAE,GAAG;GAAG;GAAU,CAAC;;CAIpC,MAAM,uBAAuB,4BAA4B,GAAG,WAAW;AAEvE,QAAO,GAAG;EACR,GAAG;EACH;EACA;EACD,CAAC;;;;;AAMJ,MAAM,qBAAqB,MAAsB,YAAuE;CAEtH,MAAM,uBAAuB,4BAA4B,KAAK,WAAW;AAEzE,QAAO,GAAG;EACR,GAAG;EACH;EACD,CAAC;;;;;AAMJ,MAAM,mBAAmB,QAAqB,aAAyD;AACrG,KAAI,aAAa,QAAQ,SAAS,EAAE;AAClC,SAAO;;AAET,KAAI,WAAW,QAAQ,SAAS,EAAE;AAChC,SAAO;;AAET,KAAI,OAAO,OAAO,IAAI,SAAS,EAAE;AAC/B,SAAO;;AAET,QAAO;;;;;AAMT,MAAM,+BAA+B,eAAgF;CACnH,MAAM,YAAY,IAAI,KAAa;CAEnC,MAAM,WAAW,SAA6D;AAC5E,OAAK,MAAM,OAAO,MAAM;AACtB,WAAQ,IAAI,MAAZ;IACE,KAAK;AACH,eAAU,IAAI,IAAI,KAAK;AACvB;IACF,KAAK;AACH,SAAI,IAAI,YAAY;AAClB,cAAQ,IAAI,WAAW;;AAEzB;IACF,KAAK;AACH,aAAQ,IAAI,WAAW;AACvB;;;;AAKR,SAAQ,WAAW;AACnB,QAAO,CAAC,GAAG,UAAU;;;;;ACrLvB,MAAM,mBAAmB;;;;;;;;;;;AAYzB,MAAa,uBAAuB,YAAoB;CACtD,MAAM,aAAa,QAAQ,QAAQ;AAEnC,KAAI;AACF,MAAI,WAAW,WAAW,EAAE;AAC1B,UAAO,IAAwB;IAC7B,MAAM;IACN,SAAS,iCAAiC;IAC1C,SAAS;IACV,CAAC;;AAGJ,YAAU,QAAQ,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,gBAAc,YAAY,GAAG,iBAAiB,IAAI;AAClD,SAAO,GAAuB,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAwB;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;AAIN,MAAa,0BAAkC,GAAG,iBAAiB;;;;ACtCnE,MAAaC,iBAA0B;CACrC,MAAM;CACN,QAAQ,OAAO,EAAE,YAAY,eAAe;AAC1C,MAAI;GACF,MAAM,YAAY,QAAQ,WAAW;GACrC,MAAM,WAAW,WAAW,MAAM,GAAG,CAAC,UAAU,OAAO;GACvD,MAAM,UAAU,GAAG,SAAS;AAE5B,SAAM,MAAM;IACV,aAAa,CAAC,WAAW;IACzB,SAAS;IACT,QAAQ;IACR,UAAU;IACV,QAAQ;IACR,UAAU,CAAC,GAAG,SAAS;IACvB,WAAW;IACX,QAAQ;IACR,aAAa;IACd,CAAC;AAEF,UAAO,GAAG,EAAE,SAAS,CAAC;WACf,OAAO;AACd,UAAO,IAAI;IACT,MAAM;IACN,SAAS,+BAA+B,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;IAC9F,SAAS;IACV,CAAC;;;CAGP;;;;;;;ACnBD,MAAa,cAAiB,OAAqB,SAAwB;AACzE,KAAI,QAAQ,GAAG;AACb,SAAO,CAAC,MAAM,KAAK,MAAM,CAAC;;CAG5B,MAAMC,SAAgB,EAAE;AACxB,MAAK,IAAI,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,MAAM;AAC3C,SAAO,KAAK,MAAM,MAAM,GAAG,IAAI,KAAK,CAAC;;AAGvC,QAAO;;;;;AAMT,MAAa,iBAAiB,MAAgC,cAA+B;AAC3F,QAAO,KAAK,SAAS;;;;;AAavB,MAAa,0BAA0B,YAA2C;CAChF,MAAM,EAAE,UAAU,MAAM,oBAAoB;AAE5C,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS;;;;;CAMZ,MAAMC,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS;;;EAGZ,eAAe,iBAAiB;;;;;;AAelC,MAAa,qBAAqB,YAAsC;CACtE,MAAM,EAAE,UAAU,MAAM,YAAY,oBAAoB;AAExD,KAAI,KAAK,WAAW,GAAG;AACrB,SAAO;KACN,SAAS,SAAS,WAAW;;;;;CAMhC,MAAMA,UAAoB,EAAE;AAC5B,KAAI,mBAAmB,aAAa,SAAS;AAC3C,UAAQ,KAAK,iDAA+C;;CAG9D,MAAM,eAAe,QAAQ,SAAS,IAAI,GAAG,QAAQ,KAAK,KAAK,CAAC,QAAQ;CACxE,MAAM,mBAAmB,KAAK,KAAK,MAAM,UAAU,EAAE,OAAO,CAAC,KAAK,KAAK;AAEvE,QAAO;KACJ,SAAS,SAAS,WAAW;;;EAGhC,eAAe,iBAAiB;;;;;;AAalC,MAAa,sBAAsB,YAAuC;CACxE,MAAM,EAAE,UAAU,eAAe;CAEjC,MAAM,YAAY,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,GAAG,MAAM,0BAA0B,EAAE,IAAI,CAAC,KAAK,KAAK;AAE1G,QAAO;KACJ,SAAS;;;EAGZ,UAAU;;;;;;AAgBZ,MAAa,kCACX,UACA,YACA,MACA,cAC2B;CAC3B,MAAM,SAAS,WAAW,MAAM,UAAU;CAC1C,MAAM,kBAAkB,aAAa;CAErC,MAAM,gBAAgB,OAAO,KAAK,WAAW,gBAAgB;EAC3D;EACA,SAAS,kBAAkB;GACzB;GACA;GACA,MAAM;GACN;GACA;GACD,CAAC;EACF,UAAU,UAAU,KAAK,MAAM,EAAE,KAAK;EACvC,EAAE;CAEH,MAAM,cAAc,KAAK,KAAK,MAAM,EAAE,KAAK;CAC3C,MAAM,eAAe,mBAAmB;EACtC;EACA,YAAY,OAAO;EACnB,UAAU;EACX,CAAC;AAEF,QAAO;EACL;EACA,QAAQ;EACT;;;;;AAqBH,MAAa,yBACX,YACA,cACA,cAC2B;CAC3B,MAAMC,QAA0D,EAAE;CAClE,MAAMC,cAAkD;EACtD,OAAO;EACP,QAAQ;EACR,SAAS;EACT,QAAQ;EACT;CAED,MAAMC,aAAmC;EAAC;EAAS;EAAU;EAAW;EAAS;AAEjF,MAAK,MAAM,YAAY,YAAY;EACjC,MAAM,OAAO,aAAa;EAC1B,MAAM,kBAAkB,aAAa;AAErC,MAAI,cAAc,MAAM,UAAU,EAAE;GAElC,MAAM,UAAU,+BAA+B,UAAU,YAAY,MAAM,UAAU;AAGrF,eAAY,YAAY,WAAW;AAGnC,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC,SAAS,QAAQ;IAClB,CAAC;AAGF,QAAK,MAAM,SAAS,QAAQ,QAAQ;AAClC,UAAM,KAAK;KACT,cAAc,SAAS,SAAS,SAAS,MAAM,WAAW;KAC1D,SAAS,MAAM;KAChB,CAAC;;SAEC;GAEL,MAAM,UAAU,uBAAuB;IACrC;IACA;IACA;IACA;IACD,CAAC;AAEF,SAAM,KAAK;IACT,cAAc,SAAS,SAAS;IAChC;IACD,CAAC;;;AAIN,QAAO;EAAE;EAAO;EAAa;;;;;ACrP/B,MAAa,eAAe,SAAiB,aAAqB;CAChE,MAAM,aAAa,QAAQ,QAAQ;AAEnC,KAAI;AACF,YAAU,QAAQ,WAAW,EAAE,EAAE,WAAW,MAAM,CAAC;AACnD,gBAAc,YAAY,SAAS;AACnC,SAAO,GAAuB,UAAU;UACjC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAwB;GAC7B,MAAM;GACN;GACA,SAAS;GACV,CAAC;;;;;;;;;;ACPN,MAAa,oBAAoB,eAAuB;CACtD,MAAM,eAAe,QAAQ,WAAW;AAExC,KAAI,CAAC,WAAW,aAAa,EAAE;AAC7B,SAAO,IAAgC;GACrC,MAAM;GACN,SAAS,4BAA4B;GACrC,YAAY;GACb,CAAC;;AAGJ,KAAI;EACF,MAAM,eAAe,aAAa,cAAc,OAAO;EACvD,MAAM,WAAW,MAAM,aAAa;AACpC,SAAO,GAA+B,SAAS;UACxC,OAAO;EACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM;AACtE,SAAO,IAAgC;GACrC,MAAM;GACN,SAAS,0BAA0B;GACnC,YAAY;GACb,CAAC;;;;;;;AAQN,MAAa,cAAc,gBAAmC;CAC5D,MAAMC,YAA4B,EAAE;AAEpC,MAAK,MAAM,cAAc,aAAa;EACpC,MAAM,SAAS,iBAAiB,WAAW;AAC3C,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAgC,OAAO,MAAM;;AAEtD,YAAU,KAAK,OAAO,MAAM;;CAI9B,MAAM,SAAS,UAAU,UAAU;AACnC,QAAO,GAA+B,OAAO;;AAG/C,MAAa,cAAc,aAAmC,WAAW,SAAS,CAAC,OAAO,MAAM,SAAS,CAAC,CAAC,OAAO,MAAM;;;;AC/CxH,MAAMC,eAAuC;CAC3C,OAAO;CACP,QAAQ;CACR,QAAQ;CACR,QAAQ;CACR,OAAO;CACP,QAAQ;CACR,QAAQ;CACT;AAMD,MAAM,qBAAqB,UAAkB,YAAoB,YAA6C;CAC5G,MAAM,UAAU,QAAQ,SAAS;CACjC,MAAM,aAAa,SAAS,SAAS,WAAW,CAAC,QAAQ,OAAO,IAAI;CACpE,MAAM,YAAY,QAAQ,WAAW;AAGrC,KAAI,CAAC,SAAS,kBAAkB;AAC9B,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAO,KAAK,SAAS,YAAY,UAAU;;EAE7C,MAAMC,eAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;EAClE,MAAMC,eAAa,QAAQD,aAAW;AACtC,SAAOC,eAAaD,aAAW,MAAM,GAAG,CAACC,aAAW,OAAO,GAAGD;;CAIhE,MAAM,aAAa,aAAa,cAAc;AAE9C,KAAI,WAAW,WAAW,GAAG;EAC3B,MAAM,OAAO,eAAe,YAAY,SAAS,YAAY,UAAU,GAAG,SAAS,WAAW;AAC9F,SAAO,KAAK,OAAO;;CAGrB,MAAM,aAAa,WAAW,WAAW,IAAI,GAAG,aAAa,KAAK;AAClE,KAAI,CAAC,YAAY;AACf,SAAO;;AAET,KAAI,WAAW,SAAS,WAAW,EAAE;AACnC,SAAO;;CAGT,MAAM,aAAa,QAAQ,WAAW;CACtC,MAAM,aAAa,aAAa,WAAW,MAAM,GAAG,CAAC,WAAW,OAAO,GAAG;AAC1E,QAAO,GAAG,aAAa;;AAGzB,MAAa,aAAa,OAAO,YAAoD;CACnF,MAAM,UAAU,QAAQ,QAAQ,QAAQ;CACxC,MAAM,yBAAyB,EAAE,kBAAkB,QAAQ,iBAAiB;AAG5E,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,aAAa,QAAQ,aAAa,OAAO,QAAQ;AACvD,MAAI,CAAC,WAAW,WAAW,EAAE;AAC3B,UAAO,IAAI;IACT,MAAM;IACN,SAAS,uCAAuC,WAAW,KAAK;IAChE,YAAY;IACb,CAAC;;AAGJ,MAAI,aAAa,OAAO,SAAS;GAC/B,MAAM,cAAc,QAAQ,aAAa,OAAO,QAAQ;AACxD,OAAI,CAAC,WAAW,YAAY,EAAE;AAC5B,WAAO,IAAI;KACT,MAAM;KACN,SAAS,wCAAwC,WAAW,KAAK;KACjE,YAAY;KACb,CAAC;;;;CAMR,MAAM,UAAU,IAAI,KAA6C;CACjE,MAAME,eAAuH,EAAE;AAE/H,MAAK,MAAM,CAAC,MAAM,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EAClE,MAAM,SAAS,MAAM,WAAW,aAAa,OAAO,CAAC,OAClD,QAAQ,QAAQ,QAAQ,GAAG,IAAI,CAAC,GAChC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,MAAI,OAAO,OAAO,EAAE;AAClB,UAAO,IAAI,OAAO,MAAM;;AAG1B,UAAQ,IAAI,MAAM,OAAO,MAAM;;CAIjC,MAAM,kBAAkB,IAAI,KAMzB;AAEH,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;EACxE,MAAM,eAAe,aAAa;AAElC,kBAAgB,IAAI,YAAY;GAC9B,kBAAkB,kBAAkB,SAAS,QAAQ,aAAa,QAAQ,EAAE,uBAAuB;GACnG,GAAI,aAAa,UACb,EAAE,mBAAmB,kBAAkB,SAAS,QAAQ,aAAa,QAAQ,EAAE,uBAAuB,EAAE,GACxG,EAAE;GACP,CAAC;;CAIJ,MAAM,0BAA0B,IAAI,KAAqB;CACzD,MAAM,4BAA4B,IAAI,KAA+C;AAErF,MAAK,MAAM,CAAC,YAAY,iBAAiB,OAAO,QAAQ,QAAQ,QAAQ,EAAE;AACxE,MAAI,aAAa,sBAAsB,aAAa,aAAa,sBAAsB,GAAG;AACxF,2BAAwB,IAAI,YAAY,aAAa,kBAAkB;;AAEzE,MAAI,aAAa,uBAAuB,OAAO,KAAK,aAAa,oBAAoB,CAAC,SAAS,GAAG;AAChG,6BAA0B,IAAI,YAAY,aAAa,oBAAoB;;;CAK/E,MAAM,YAAY,QAAQ,aAAa;CAGvC,MAAM,EACJ,MAAM,cACN,aACA,iBACE,0BAA0B,SAAS;EACrC,WAAW;EACX,mBAAmB,wBAAwB,OAAO,IAAI,0BAA0B;EAChF,qBAAqB,0BAA0B,OAAO,IAAI,4BAA4B;EACtF;EACD,CAAC;CAGF,MAAM,YAAY;;;;;;;AASlB,MAAK,MAAM,CAAC,MAAM,aAAa,QAAQ,SAAS,EAAE;EAChD,MAAM,eAAe,MAAM,OAAO,6BAAgB,kBAAkB,SAAS;EAC7E,MAAM,UAAU,MAAM,KAAK,YAAY,QAAQ,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EAC1F,MAAM,QAAQ,MAAM,KAAK,YAAY,MAAM,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACtF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;EACxF,MAAM,SAAS,MAAM,KAAK,YAAY,OAAO,MAAM,CAAC,CAAC,QAAQ,MAAM,CAAC,EAAE,WAAW,KAAK,CAAC,CAAC;AAExF,eAAa,QAAQ;GACnB,YAAY,WAAW,SAAS;GAChC;GACA;GACA;GACA;GACD;;CAIH,MAAM,cAAc,KAAK,QAAQ,QAAQ,EAAE,uBAAuB;AAClE,KAAI,aAAa;EACf,MAAM,qBAAqB,MAAM,YAAY,aAAa,YAAY,CAAC,YAC/D,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,MAAI,mBAAmB,OAAO,EAAE;AAC9B,UAAO,IAAI,mBAAmB,MAAM;;;CAKxC,MAAMC,YAAsB,EAAE;AAC9B,KAAI,cAAc;EAChB,MAAM,SAAS,QAAQ,QAAQ;EAK/B,MAAM,eAAe;GACnB,OAAO,EAAE;GACT,QAAQ,EAAE;GACV,SAAS,EAAE;GACX,QAAQ,EAAE;GACX;AAED,OAAK,MAAM,QAAQ,OAAO,OAAO,aAAa,EAAE;AAC9C,gBAAa,MAAM,KAAK,GAAG,KAAK,MAAM;AACtC,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;AACxC,gBAAa,QAAQ,KAAK,GAAG,KAAK,QAAQ;AAC1C,gBAAa,OAAO,KAAK,GAAG,KAAK,OAAO;;EAI1C,MAAM,gBAAgB,sBAAsB,YAAY,cAAc,UAAU;AAEhF,OAAK,MAAM,QAAQ,cAAc,OAAO;GACtC,MAAM,WAAW,KAAK,QAAQ,KAAK,aAAa;GAGhD,MAAM,cAAc,MAAM,YAAY,UAAU,KAAK,QAAQ,CAAC,YACtD,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,OAAI,YAAY,OAAO,EAAE;AACvB,WAAO,IAAI,YAAY,MAAM;;AAG/B,aAAU,KAAK,SAAS;;;CAK5B,MAAM,eAAe,KAAK,QAAQ,QAAQ,EAAE,eAAe;CAC3D,MAAM,sBAAsB,MAAM,YAAY,cAAc,aAAa,CAAC,YAClE,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,KAAI,oBAAoB,OAAO,EAAE;AAC/B,SAAO,IAAI,oBAAoB,MAAM;;CAIvC,MAAM,mBAAmB,MAAM,YAAY,SAAS,UAAU,CAAC,YACvD,QAAQ,QAAQ,GAAG,UAAU,CAAC,GACnC,UAAU,QAAQ,QAAQ,IAAI,MAAM,CAAC,CACvC;AAED,KAAI,iBAAiB,OAAO,EAAE;AAC5B,SAAO,IAAI,iBAAiB,MAAM;;CAIpC,MAAM,gBAAgB,MAAMC,eAAe,OAAO;EAChD,YAAY;EACZ,UAAU,CAAC,kBAAkB,oBAAoB;EAClD,CAAC;CACF,MAAM,eAAe,cAAc,OAChC,WAAW,GAAG,OAAO,GACrB,UAAU,IAAI,MAAM,CACtB;AAED,KAAI,aAAa,OAAO,EAAE;AACxB,SAAO,IAAI,aAAa,MAAM;;AAGhC,QAAO,GAAG;EACR,SAAS;EACT;EACA;EACA;EACA,SAAS,aAAa,MAAM;EAC5B,GAAI,UAAU,SAAS,IAAI,EAAE,WAAW,GAAG,EAAE;EAC9C,CAA0B"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@soda-gql/codegen",
|
|
3
|
-
"version": "0.11.
|
|
3
|
+
"version": "0.11.13",
|
|
4
4
|
"description": "GraphQL schema code generation for soda-gql",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"private": false,
|
|
@@ -48,8 +48,8 @@
|
|
|
48
48
|
"./package.json": "./package.json"
|
|
49
49
|
},
|
|
50
50
|
"dependencies": {
|
|
51
|
-
"@soda-gql/core": "0.11.
|
|
52
|
-
"@soda-gql/runtime": "0.11.
|
|
51
|
+
"@soda-gql/core": "0.11.13",
|
|
52
|
+
"@soda-gql/runtime": "0.11.13",
|
|
53
53
|
"esbuild": "^0.24.0",
|
|
54
54
|
"graphql": "^16.8.1",
|
|
55
55
|
"neverthrow": "^8.1.1"
|