@character-foundry/character-foundry 0.4.2-dev.1765942273 → 0.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/charx.cjs +2 -2
- package/dist/charx.cjs.map +1 -1
- package/dist/charx.d.cts +9 -18
- package/dist/charx.d.ts +9 -18
- package/dist/charx.js +2 -2
- package/dist/charx.js.map +1 -1
- package/dist/exporter.cjs +2 -2
- package/dist/exporter.cjs.map +1 -1
- package/dist/exporter.d.cts +9 -18
- package/dist/exporter.d.ts +9 -18
- package/dist/exporter.js +2 -2
- package/dist/exporter.js.map +1 -1
- package/dist/federation.d.cts +9 -18
- package/dist/federation.d.ts +9 -18
- package/dist/index.cjs +2 -2
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +21 -42
- package/dist/index.d.ts +21 -42
- package/dist/index.js +2 -2
- package/dist/index.js.map +1 -1
- package/dist/loader.cjs +2 -2
- package/dist/loader.cjs.map +1 -1
- package/dist/loader.d.cts +14 -28
- package/dist/loader.d.ts +14 -28
- package/dist/loader.js +2 -2
- package/dist/loader.js.map +1 -1
- package/dist/lorebook.d.cts +17 -34
- package/dist/lorebook.d.ts +17 -34
- package/dist/normalizer.cjs +2 -2
- package/dist/normalizer.cjs.map +1 -1
- package/dist/normalizer.d.cts +30 -60
- package/dist/normalizer.d.ts +30 -60
- package/dist/normalizer.js +2 -2
- package/dist/normalizer.js.map +1 -1
- package/dist/png.cjs +2 -2
- package/dist/png.cjs.map +1 -1
- package/dist/png.d.cts +16 -32
- package/dist/png.d.ts +16 -32
- package/dist/png.js +2 -2
- package/dist/png.js.map +1 -1
- package/dist/schemas.cjs +2 -2
- package/dist/schemas.cjs.map +1 -1
- package/dist/schemas.d.cts +48 -96
- package/dist/schemas.d.ts +48 -96
- package/dist/schemas.js +2 -2
- package/dist/schemas.js.map +1 -1
- package/dist/voxta.cjs +2 -2
- package/dist/voxta.cjs.map +1 -1
- package/dist/voxta.d.cts +14 -28
- package/dist/voxta.d.ts +14 -28
- package/dist/voxta.js +2 -2
- package/dist/voxta.js.map +1 -1
- package/package.json +4 -4
package/dist/png.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/png.ts","../../png/src/parser.ts","../../png/src/builder.ts","../../png/src/crc32.ts","../../core/src/binary.ts","../../core/src/base64.ts","../../core/src/errors.ts","../../core/src/data-url.ts","../../core/src/uri.ts","../../core/src/image.ts","../../core/src/uuid.ts","../../schemas/src/common.ts","../../schemas/src/ccv2.ts","../../schemas/src/ccv3.ts","../../schemas/src/risu.ts","../../schemas/src/normalized.ts","../../schemas/src/feature-deriver.ts","../../schemas/src/detection.ts","../../schemas/src/normalizer.ts","../../schemas/src/validation.ts"],"sourcesContent":["export * from '@character-foundry/png';\n","/**\n * PNG Parser\n *\n * Extracts text chunks and character card data from PNG files.\n * Uses fflate for zTXt decompression, works in browser and Node.js.\n */\n\nimport { Inflate } from 'fflate';\nimport {\n type BinaryData,\n readUInt32BE,\n slice,\n toString,\n toLatin1,\n indexOf,\n concat,\n base64Decode,\n ParseError,\n SizeLimitError,\n} from '@character-foundry/core';\nimport {\n type CCv2Data,\n type CCv3Data,\n type Spec,\n detectSpec,\n} from '@character-foundry/schemas';\n\n/**\n * PNG signature bytes\n */\nexport const PNG_SIGNATURE = new Uint8Array([0x89, 0x50, 0x4e, 0x47, 0x0d, 0x0a, 0x1a, 0x0a]);\n\n/**\n * Maximum size for a single PNG chunk (50MB per Risu CharX spec)\n */\nexport const MAX_CHUNK_SIZE = 50 * 1024 * 1024;\n\n/**\n * Maximum inflated size for zTXt chunks (16MB - reasonable for card JSON)\n * This prevents inflation attacks where small compressed data expands to huge sizes\n */\nexport const MAX_INFLATED_SIZE = 16 * 1024 * 1024;\n\n/**\n * Synchronous inflate with size limit to prevent inflation attacks.\n * Uses streaming decompression to enforce the limit without fully decompressing.\n *\n * @param compressed - Compressed data to inflate\n * @param maxSize - Maximum allowed inflated size (defaults to MAX_INFLATED_SIZE)\n * @returns Inflated data as Uint8Array\n * @throws SizeLimitError if inflated size exceeds maxSize\n */\nfunction inflateSyncWithLimit(compressed: BinaryData, maxSize: number = MAX_INFLATED_SIZE): Uint8Array {\n const chunks: Uint8Array[] = [];\n let totalSize = 0;\n let error: Error | null = null;\n\n const inflater = new Inflate((data: Uint8Array, _final: boolean) => {\n if (error) return; // Stop processing if we've hit a limit\n\n if (data && data.length > 0) {\n totalSize += data.length;\n if (totalSize > maxSize) {\n error = new SizeLimitError(totalSize, maxSize, 'inflated zTXt chunk');\n return;\n }\n chunks.push(data);\n }\n });\n\n // Push all data at once with final=true for sync operation\n try {\n inflater.push(compressed instanceof Uint8Array ? compressed : new Uint8Array(compressed), true);\n } catch (e) {\n throw new ParseError(`Decompression failed: ${e instanceof Error ? e.message : String(e)}`, 'png');\n }\n\n if (error) {\n throw error;\n }\n\n return concat(...chunks);\n}\n\n/**\n * Text chunk keys used for character cards by various frontends\n */\nexport const TEXT_CHUNK_KEYS = [\n // v3 keys\n 'ccv3',\n 'chara_card_v3',\n // v2 keys (most common)\n 'chara',\n 'ccv2',\n 'character',\n // Alternative/legacy keys\n 'charactercard',\n 'card',\n 'CharacterCard',\n 'Chara',\n];\n\n/**\n * Parsed text chunk from PNG\n */\nexport interface TextChunk {\n keyword: string;\n text: string;\n}\n\n/**\n * Result of extracting card data from PNG\n */\nexport interface PNGExtractionResult {\n data: CCv2Data | CCv3Data;\n spec: Spec;\n extraChunks: TextChunk[];\n}\n\n/**\n * Check if data is a valid PNG (has correct signature)\n */\nexport function isPNG(data: BinaryData): boolean {\n if (data.length < 8) return false;\n for (let i = 0; i < 8; i++) {\n if (data[i] !== PNG_SIGNATURE[i]) return false;\n }\n return true;\n}\n\n/**\n * Parse all text chunks (tEXt and zTXt) from PNG buffer\n * Returns array of {keyword, text} to support multiple chunks\n */\nexport function parseTextChunks(data: BinaryData): TextChunk[] {\n const textChunks: TextChunk[] = [];\n\n // Verify PNG signature\n if (!isPNG(data)) {\n return textChunks;\n }\n\n let offset = 8; // Skip PNG signature\n\n while (offset < data.length) {\n // Read chunk length (4 bytes, big-endian)\n if (offset + 4 > data.length) break;\n const length = readUInt32BE(data, offset);\n offset += 4;\n\n // Read chunk type (4 bytes ASCII)\n if (offset + 4 > data.length) break;\n const typeBytes = slice(data, offset, offset + 4);\n const type = toLatin1(typeBytes);\n offset += 4;\n\n // Check chunk size limit before reading\n if (length > MAX_CHUNK_SIZE) {\n throw new SizeLimitError(length, MAX_CHUNK_SIZE, `PNG chunk '${type}'`);\n }\n\n // Read chunk data\n if (offset + length > data.length) break;\n const chunkData = slice(data, offset, offset + length);\n offset += length;\n\n // Skip CRC (4 bytes)\n if (offset + 4 > data.length) break;\n offset += 4;\n\n // Parse tEXt chunks\n if (type === 'tEXt') {\n const nullIndex = indexOf(chunkData, new Uint8Array([0]));\n if (nullIndex !== -1) {\n const keyword = toLatin1(slice(chunkData, 0, nullIndex));\n const text = toString(slice(chunkData, nullIndex + 1));\n textChunks.push({ keyword, text });\n }\n }\n\n // Parse zTXt chunks (compressed) with size limit protection\n if (type === 'zTXt') {\n const nullIndex = indexOf(chunkData, new Uint8Array([0]));\n if (nullIndex !== -1) {\n const keyword = toLatin1(slice(chunkData, 0, nullIndex));\n const compressionMethod = chunkData[nullIndex + 1];\n\n if (compressionMethod === 0) { // 0 = deflate/inflate\n try {\n const compressedData = slice(chunkData, nullIndex + 2);\n // Use size-limited inflate to prevent inflation attacks\n const decompressed = inflateSyncWithLimit(compressedData, MAX_INFLATED_SIZE);\n const text = toString(decompressed);\n textChunks.push({ keyword, text });\n } catch (err) {\n // Re-throw size limit errors, skip other decompression failures\n if (err instanceof SizeLimitError) {\n throw err;\n }\n // Failed to decompress zTXt chunk, skip it\n }\n }\n }\n }\n\n // Stop after IEND chunk\n if (type === 'IEND') break;\n }\n\n return textChunks;\n}\n\n/**\n * List all chunks in a PNG file\n */\nexport function listChunks(data: BinaryData): Array<{ type: string; offset: number; length: number }> {\n const chunks: Array<{ type: string; offset: number; length: number }> = [];\n\n if (!isPNG(data)) {\n return chunks;\n }\n\n let offset = 8;\n\n while (offset < data.length) {\n if (offset + 4 > data.length) break;\n const length = readUInt32BE(data, offset);\n const chunkStart = offset;\n offset += 4;\n\n if (offset + 4 > data.length) break;\n const type = toLatin1(slice(data, offset, offset + 4));\n offset += 4;\n\n chunks.push({ type, offset: chunkStart, length });\n\n // Skip data and CRC\n offset += length + 4;\n\n if (type === 'IEND') break;\n }\n\n return chunks;\n}\n\n/**\n * Try to parse JSON from chunk data (supports plain and base64)\n */\nfunction tryParseChunk(chunkData: string): unknown {\n // Try direct JSON parse first\n try {\n return JSON.parse(chunkData);\n } catch {\n // Try base64 decode then JSON parse\n try {\n const decoded = toString(base64Decode(chunkData));\n return JSON.parse(decoded);\n } catch {\n throw new ParseError('Not valid JSON or base64-encoded JSON', 'png');\n }\n }\n}\n\n/**\n * Check if parsed data has a lorebook\n */\nfunction hasLorebookInData(json: unknown): boolean {\n if (!json || typeof json !== 'object') return false;\n const obj = json as Record<string, unknown>;\n\n // Check wrapped format\n const data = obj.data as Record<string, unknown> | undefined;\n if (data?.character_book) {\n const book = data.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Extract character card JSON from PNG tEXt chunks\n * Returns card data, spec version, and any extra chunks\n */\nexport function extractFromPNG(data: BinaryData): PNGExtractionResult {\n // Validate PNG format\n if (!isPNG(data)) {\n throw new ParseError('Invalid PNG signature', 'png');\n }\n\n // Parse all text chunks\n const textChunks = parseTextChunks(data);\n\n if (textChunks.length === 0) {\n throw new ParseError('No text chunks found in PNG', 'png');\n }\n\n // Try all known keys, preferring chunks with lorebooks\n let fallbackResult: PNGExtractionResult | null = null;\n\n for (const key of TEXT_CHUNK_KEYS) {\n // Find all chunks with this keyword\n const matchingChunks = textChunks.filter(c => c.keyword === key);\n\n for (const chunk of matchingChunks) {\n try {\n const json = tryParseChunk(chunk.text);\n const spec = detectSpec(json);\n\n if (spec === 'v3' || spec === 'v2') {\n const result: PNGExtractionResult = {\n data: json as CCv2Data | CCv3Data,\n spec,\n extraChunks: textChunks.filter(c => c.keyword !== key),\n };\n\n // Prefer chunks with lorebooks\n if (hasLorebookInData(json)) {\n return result;\n }\n\n // Store as fallback if we don't find one with lorebook\n if (!fallbackResult) {\n fallbackResult = result;\n }\n }\n\n // If detectSpec failed but we have JSON that looks like a card, try to infer\n if (!spec && json && typeof json === 'object') {\n const obj = json as Record<string, unknown>;\n let inferredResult: { data: CCv2Data | CCv3Data; spec: Spec } | null = null;\n\n if (obj.spec === 'chara_card_v3' && obj.data && (obj.data as Record<string, unknown>).name) {\n inferredResult = { data: json as CCv3Data, spec: 'v3' };\n } else if (obj.spec === 'chara_card_v2' && obj.data && (obj.data as Record<string, unknown>).name) {\n inferredResult = { data: json as CCv2Data, spec: 'v2' };\n } else if (obj.name && (obj.description || obj.personality || obj.scenario)) {\n inferredResult = { data: json as CCv2Data, spec: 'v2' };\n }\n\n if (inferredResult) {\n const fullResult: PNGExtractionResult = {\n ...inferredResult,\n extraChunks: textChunks.filter(c => c.keyword !== key),\n };\n if (hasLorebookInData(json)) {\n return fullResult;\n }\n if (!fallbackResult) fallbackResult = fullResult;\n }\n }\n } catch {\n // Continue to next chunk\n }\n }\n }\n\n // If we found a valid card but no lorebook, use the fallback\n if (fallbackResult) {\n return fallbackResult;\n }\n\n throw new ParseError('No valid character card data found in PNG', 'png');\n}\n","/**\n * PNG Builder\n *\n * Embeds character card data into PNG tEXt chunks.\n * Works in browser and Node.js.\n */\n\nimport {\n type BinaryData,\n readUInt32BE,\n writeUInt32BE,\n slice,\n concat,\n fromLatin1,\n fromString,\n alloc,\n base64Encode,\n toLatin1,\n ParseError,\n} from '@character-foundry/core';\nimport type { CCv2Data, CCv3Data } from '@character-foundry/schemas';\nimport { crc32Bytes } from './crc32.js';\nimport { isPNG } from './parser.js';\n\n/**\n * Options for embedding card data into PNG\n */\nexport interface EmbedOptions {\n /**\n * Key to use for the tEXt chunk (default: 'chara')\n */\n key?: string;\n\n /**\n * Whether to base64 encode the JSON (default: true, recommended for compatibility)\n */\n base64?: boolean;\n\n /**\n * Whether to minify JSON (default: true)\n */\n minify?: boolean;\n}\n\n/**\n * Remove all tEXt and zTXt chunks from PNG buffer\n * CRITICAL: Must remove old chunks before adding new ones to prevent duplicate data\n */\nexport function removeAllTextChunks(pngBuffer: BinaryData): BinaryData {\n if (!isPNG(pngBuffer)) {\n throw new ParseError('Invalid PNG signature', 'png');\n }\n\n const chunks: BinaryData[] = [slice(pngBuffer, 0, 8)]; // Start with PNG signature\n let offset = 8;\n\n while (offset < pngBuffer.length) {\n // Read chunk length (4 bytes, big-endian)\n if (offset + 4 > pngBuffer.length) break;\n const length = readUInt32BE(pngBuffer, offset);\n const lengthBuf = slice(pngBuffer, offset, offset + 4);\n offset += 4;\n\n // Read chunk type (4 bytes ASCII)\n if (offset + 4 > pngBuffer.length) break;\n const type = toLatin1(slice(pngBuffer, offset, offset + 4));\n const typeBuf = slice(pngBuffer, offset, offset + 4);\n offset += 4;\n\n // Read chunk data + CRC\n if (offset + length + 4 > pngBuffer.length) break;\n const dataBuf = slice(pngBuffer, offset, offset + length);\n const crcBuf = slice(pngBuffer, offset + length, offset + length + 4);\n offset += length + 4;\n\n // Skip tEXt and zTXt chunks (don't add them to output)\n if (type === 'tEXt' || type === 'zTXt') {\n continue;\n }\n\n // Keep all other chunks\n chunks.push(lengthBuf, typeBuf, dataBuf, crcBuf);\n\n // Stop after IEND\n if (type === 'IEND') break;\n }\n\n return concat(...chunks);\n}\n\n/**\n * Find the offset of the IEND chunk in PNG buffer\n */\nfunction findIendOffset(pngBuffer: BinaryData): number {\n // Search backwards from the end - IEND should be near the end\n // IEND format: length(4) + \"IEND\"(4) + CRC(4) = 12 bytes\n for (let i = pngBuffer.length - 12; i >= 8; i--) {\n if (\n pngBuffer[i + 4] === 0x49 && // 'I'\n pngBuffer[i + 5] === 0x45 && // 'E'\n pngBuffer[i + 6] === 0x4e && // 'N'\n pngBuffer[i + 7] === 0x44 // 'D'\n ) {\n return i; // Start of length field\n }\n }\n return -1;\n}\n\n/**\n * Create a tEXt chunk\n */\nfunction createTextChunk(keyword: string, text: string): BinaryData {\n const keywordBuffer = fromLatin1(keyword);\n const textBuffer = fromString(text);\n\n // Chunk data: keyword + null separator + text\n const chunkData = concat(\n keywordBuffer,\n new Uint8Array([0]), // null separator\n textBuffer\n );\n\n // Chunk type\n const chunkType = fromLatin1('tEXt');\n\n // Calculate CRC32 of type + data\n const crc = crc32Bytes(concat(chunkType, chunkData));\n\n // Build length buffer\n const lengthBuffer = alloc(4);\n writeUInt32BE(lengthBuffer, chunkData.length, 0);\n\n // Assemble: length + type + data + crc\n return concat(lengthBuffer, chunkType, chunkData, crc);\n}\n\n/**\n * Inject a tEXt chunk into PNG buffer before IEND\n */\nexport function injectTextChunk(pngBuffer: BinaryData, keyword: string, text: string): BinaryData {\n const iendOffset = findIendOffset(pngBuffer);\n\n if (iendOffset === -1) {\n throw new ParseError('Invalid PNG: IEND chunk not found', 'png');\n }\n\n const textChunk = createTextChunk(keyword, text);\n\n // Insert before IEND\n const beforeIend = slice(pngBuffer, 0, iendOffset);\n const iendAndAfter = slice(pngBuffer, iendOffset);\n\n return concat(beforeIend, textChunk, iendAndAfter);\n}\n\n/**\n * Embed character card JSON into PNG tEXt chunk\n */\nexport function embedIntoPNG(\n imageBuffer: BinaryData,\n cardData: CCv2Data | CCv3Data,\n options: EmbedOptions = {}\n): BinaryData {\n const {\n key = 'chara',\n base64 = true,\n minify = true,\n } = options;\n\n // Remove all existing tEXt chunks first to prevent duplicate/stale data\n const cleanPng = removeAllTextChunks(imageBuffer);\n\n // Serialize JSON\n const json = minify\n ? JSON.stringify(cardData)\n : JSON.stringify(cardData, null, 2);\n\n // Optionally base64 encode (recommended for compatibility)\n const text = base64 ? base64Encode(fromString(json)) : json;\n\n // Inject the text chunk\n return injectTextChunk(cleanPng, key, text);\n}\n\n/**\n * Validate PNG size against limits\n */\nexport function validatePNGSize(\n buffer: BinaryData,\n limits: { max: number; warn: number }\n): { valid: boolean; warnings: string[] } {\n const sizeMB = buffer.length / (1024 * 1024);\n const warnings: string[] = [];\n\n if (sizeMB > limits.max) {\n return {\n valid: false,\n warnings: [`PNG size (${sizeMB.toFixed(2)}MB) exceeds maximum (${limits.max}MB)`],\n };\n }\n\n if (sizeMB > limits.warn) {\n warnings.push(`PNG size (${sizeMB.toFixed(2)}MB) is large (recommended: <${limits.warn}MB)`);\n }\n\n return { valid: true, warnings };\n}\n","/**\n * CRC32 Implementation for PNG Chunks\n *\n * Pure JavaScript implementation using the standard polynomial.\n */\n\nimport { type BinaryData, alloc, writeUInt32BE } from '@character-foundry/core';\n\n// Pre-computed CRC table for faster calculation\nlet CRC_TABLE: Uint32Array | null = null;\n\n/**\n * Build the CRC lookup table (lazy initialization)\n */\nfunction getCRCTable(): Uint32Array {\n if (CRC_TABLE) return CRC_TABLE;\n\n CRC_TABLE = new Uint32Array(256);\n for (let i = 0; i < 256; i++) {\n let c = i;\n for (let j = 0; j < 8; j++) {\n c = (c & 1) ? (0xEDB88320 ^ (c >>> 1)) : (c >>> 1);\n }\n CRC_TABLE[i] = c;\n }\n return CRC_TABLE;\n}\n\n/**\n * Calculate CRC32 checksum for a PNG chunk\n * @param data - Binary data to calculate CRC for\n * @returns CRC32 value as a number\n */\nexport function crc32(data: BinaryData): number {\n const table = getCRCTable();\n let crc = 0xFFFFFFFF;\n\n for (let i = 0; i < data.length; i++) {\n crc = table[(crc ^ data[i]!) & 0xFF]! ^ (crc >>> 8);\n }\n\n return (crc ^ 0xFFFFFFFF) >>> 0;\n}\n\n/**\n * Calculate CRC32 and return as a 4-byte big-endian Uint8Array\n * @param data - Binary data to calculate CRC for\n * @returns 4-byte Uint8Array containing CRC32 in big-endian format\n */\nexport function crc32Bytes(data: BinaryData): BinaryData {\n const crc = crc32(data);\n const result = alloc(4);\n writeUInt32BE(result, crc, 0);\n return result;\n}\n","/**\n * Binary Data Utilities\n *\n * Universal binary data operations using Uint8Array.\n * Works in both Node.js and browser environments.\n */\n\n/**\n * Universal binary data type (works in both environments)\n */\nexport type BinaryData = Uint8Array;\n\n/**\n * Read a 32-bit big-endian unsigned integer\n */\nexport function readUInt32BE(data: BinaryData, offset: number): number {\n return (\n (data[offset]! << 24) |\n (data[offset + 1]! << 16) |\n (data[offset + 2]! << 8) |\n data[offset + 3]!\n ) >>> 0;\n}\n\n/**\n * Write a 32-bit big-endian unsigned integer\n */\nexport function writeUInt32BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 24) & 0xff;\n data[offset + 1] = (value >>> 16) & 0xff;\n data[offset + 2] = (value >>> 8) & 0xff;\n data[offset + 3] = value & 0xff;\n}\n\n/**\n * Read a 16-bit big-endian unsigned integer\n */\nexport function readUInt16BE(data: BinaryData, offset: number): number {\n return ((data[offset]! << 8) | data[offset + 1]!) >>> 0;\n}\n\n/**\n * Write a 16-bit big-endian unsigned integer\n */\nexport function writeUInt16BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 8) & 0xff;\n data[offset + 1] = value & 0xff;\n}\n\n/**\n * Find a byte sequence in binary data\n */\nexport function indexOf(data: BinaryData, search: BinaryData, fromIndex = 0): number {\n outer: for (let i = fromIndex; i <= data.length - search.length; i++) {\n for (let j = 0; j < search.length; j++) {\n if (data[i + j] !== search[j]) continue outer;\n }\n return i;\n }\n return -1;\n}\n\n/**\n * Concatenate multiple binary arrays\n */\nexport function concat(...arrays: BinaryData[]): BinaryData {\n const totalLength = arrays.reduce((sum, arr) => sum + arr.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const arr of arrays) {\n result.set(arr, offset);\n offset += arr.length;\n }\n return result;\n}\n\n/**\n * Slice binary data (returns a view, not a copy)\n */\nexport function slice(data: BinaryData, start: number, end?: number): BinaryData {\n return data.subarray(start, end);\n}\n\n/**\n * Copy a portion of binary data (returns a new array)\n */\nexport function copy(data: BinaryData, start: number, end?: number): BinaryData {\n return data.slice(start, end);\n}\n\n/**\n * Convert string to binary (UTF-8)\n */\nexport function fromString(str: string): BinaryData {\n return new TextEncoder().encode(str);\n}\n\n/**\n * Convert binary to string (UTF-8)\n */\nexport function toString(data: BinaryData): string {\n return new TextDecoder().decode(data);\n}\n\n/**\n * Convert string to binary (Latin1 - for PNG keywords and similar)\n */\nexport function fromLatin1(str: string): BinaryData {\n const result = new Uint8Array(str.length);\n for (let i = 0; i < str.length; i++) {\n result[i] = str.charCodeAt(i) & 0xff;\n }\n return result;\n}\n\n/**\n * Convert binary to string (Latin1)\n */\nexport function toLatin1(data: BinaryData): string {\n let result = '';\n for (let i = 0; i < data.length; i++) {\n result += String.fromCharCode(data[i]!);\n }\n return result;\n}\n\n/**\n * Compare two binary arrays for equality\n */\nexport function equals(a: BinaryData, b: BinaryData): boolean {\n if (a.length !== b.length) return false;\n for (let i = 0; i < a.length; i++) {\n if (a[i] !== b[i]) return false;\n }\n return true;\n}\n\n/**\n * Create a new Uint8Array filled with zeros\n */\nexport function alloc(size: number): BinaryData {\n return new Uint8Array(size);\n}\n\n/**\n * Create a Uint8Array from an array of numbers\n */\nexport function from(data: number[] | ArrayBuffer | BinaryData): BinaryData {\n if (data instanceof Uint8Array) {\n return data;\n }\n if (data instanceof ArrayBuffer) {\n return new Uint8Array(data);\n }\n return new Uint8Array(data);\n}\n\n/**\n * Check if value is a Uint8Array\n */\nexport function isBinaryData(value: unknown): value is BinaryData {\n return value instanceof Uint8Array;\n}\n\n/**\n * Convert Node.js Buffer to Uint8Array (no-op if already Uint8Array)\n * This provides compatibility when interfacing with Node.js code\n */\nexport function toUint8Array(data: BinaryData | Buffer): BinaryData {\n if (data instanceof Uint8Array) {\n // Buffer extends Uint8Array, but we want a plain Uint8Array\n // This ensures we get a proper Uint8Array in all cases\n if (Object.getPrototypeOf(data).constructor.name === 'Buffer') {\n return new Uint8Array(data.buffer, data.byteOffset, data.byteLength);\n }\n return data;\n }\n return new Uint8Array(data);\n}\n\n/**\n * Convert binary data to hex string\n */\nexport function toHex(data: BinaryData): string {\n return Array.from(data)\n .map(b => b.toString(16).padStart(2, '0'))\n .join('');\n}\n\n/**\n * Convert hex string to binary data\n */\nexport function fromHex(hex: string): BinaryData {\n const bytes = new Uint8Array(hex.length / 2);\n for (let i = 0; i < bytes.length; i++) {\n bytes[i] = parseInt(hex.substr(i * 2, 2), 16);\n }\n return bytes;\n}\n","/**\n * Universal Base64 Encoding/Decoding\n *\n * Works in both Node.js and browser environments.\n */\n\nimport type { BinaryData } from './binary.js';\n\n/**\n * Check if we're in a Node.js environment\n */\nconst isNode = typeof process !== 'undefined' &&\n process.versions != null &&\n process.versions.node != null;\n\n/**\n * Threshold for switching to chunked encoding in browsers (1MB)\n * Below this, simple string concatenation is fast enough.\n * Above this, quadratic string growth becomes a problem.\n */\nconst LARGE_BUFFER_THRESHOLD = 1024 * 1024;\n\n/**\n * Encode binary data to base64 string\n *\n * PERFORMANCE: For large buffers (>1MB) in browsers, this automatically\n * uses the chunked implementation to avoid quadratic string concatenation.\n */\nexport function encode(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: use chunked encoding for large buffers to avoid O(n²) string growth\n if (data.length > LARGE_BUFFER_THRESHOLD) {\n return encodeChunked(data);\n }\n\n // Small buffers: simple approach is fast enough\n let binary = '';\n for (let i = 0; i < data.length; i++) {\n binary += String.fromCharCode(data[i]!);\n }\n return btoa(binary);\n}\n\n/**\n * Decode base64 string to binary data\n */\nexport function decode(base64: string): BinaryData {\n if (isNode) {\n // Node.js: use Buffer\n return new Uint8Array(Buffer.from(base64, 'base64'));\n }\n\n // Browser: use atob\n const binary = atob(base64);\n const result = new Uint8Array(binary.length);\n for (let i = 0; i < binary.length; i++) {\n result[i] = binary.charCodeAt(i);\n }\n return result;\n}\n\n/**\n * Check if a string is valid base64\n */\nexport function isBase64(str: string): boolean {\n if (str.length === 0) return false;\n // Base64 regex: only valid base64 characters, length multiple of 4 (with padding)\n const base64Regex = /^[A-Za-z0-9+/]*={0,2}$/;\n return base64Regex.test(str) && str.length % 4 === 0;\n}\n\n/**\n * Encode binary data to URL-safe base64 string\n * Replaces + with -, / with _, and removes padding\n */\nexport function encodeUrlSafe(data: BinaryData): string {\n return encode(data)\n .replace(/\\+/g, '-')\n .replace(/\\//g, '_')\n .replace(/=+$/, '');\n}\n\n/**\n * Decode URL-safe base64 string to binary data\n */\nexport function decodeUrlSafe(base64: string): BinaryData {\n // Add back padding if needed\n let padded = base64\n .replace(/-/g, '+')\n .replace(/_/g, '/');\n\n while (padded.length % 4 !== 0) {\n padded += '=';\n }\n\n return decode(padded);\n}\n\n/**\n * Chunk size for encoding large buffers (64KB)\n * Prevents stack overflow when using String.fromCharCode with spread operator\n */\nconst ENCODE_CHUNK_SIZE = 64 * 1024;\n\n/**\n * Encode binary data to base64 string with chunking for large buffers.\n * Handles buffers >10MB without stack overflow.\n *\n * @param data - Binary data to encode\n * @returns Base64 encoded string\n *\n * @example\n * ```typescript\n * const largeBuffer = new Uint8Array(20 * 1024 * 1024); // 20MB\n * const base64 = encodeChunked(largeBuffer); // No stack overflow\n * ```\n */\nexport function encodeChunked(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: process in chunks to avoid stack overflow\n const chunks: string[] = [];\n\n for (let i = 0; i < data.length; i += ENCODE_CHUNK_SIZE) {\n const chunk = data.subarray(i, Math.min(i + ENCODE_CHUNK_SIZE, data.length));\n let binary = '';\n for (let j = 0; j < chunk.length; j++) {\n binary += String.fromCharCode(chunk[j]!);\n }\n chunks.push(binary);\n }\n\n return btoa(chunks.join(''));\n}\n","/**\n * Error Classes\n *\n * Specific error types for character card operations.\n * All errors extend FoundryError for consistent handling.\n */\n\n/** Symbol to identify FoundryError instances across ESM/CJS boundaries */\nconst FOUNDRY_ERROR_MARKER = Symbol.for('@character-foundry/core:FoundryError');\n\n/**\n * Base error class for all Character Foundry errors\n */\nexport class FoundryError extends Error {\n /** @internal Marker for cross-module identification */\n readonly [FOUNDRY_ERROR_MARKER] = true;\n\n constructor(message: string, public readonly code: string) {\n super(message);\n this.name = 'FoundryError';\n // Maintains proper stack trace in V8 environments\n if (Error.captureStackTrace) {\n Error.captureStackTrace(this, this.constructor);\n }\n }\n}\n\n/**\n * Error during card parsing\n */\nexport class ParseError extends FoundryError {\n constructor(message: string, public readonly format?: string) {\n super(message, 'PARSE_ERROR');\n this.name = 'ParseError';\n }\n}\n\n/**\n * Error during card validation\n */\nexport class ValidationError extends FoundryError {\n constructor(message: string, public readonly field?: string) {\n super(message, 'VALIDATION_ERROR');\n this.name = 'ValidationError';\n }\n}\n\n/**\n * Asset not found in card or archive\n */\nexport class AssetNotFoundError extends FoundryError {\n constructor(public readonly uri: string) {\n super(`Asset not found: ${uri}`, 'ASSET_NOT_FOUND');\n this.name = 'AssetNotFoundError';\n }\n}\n\n/**\n * Format not supported for operation\n */\nexport class FormatNotSupportedError extends FoundryError {\n constructor(public readonly format: string, operation?: string) {\n const msg = operation\n ? `Format '${format}' not supported for ${operation}`\n : `Format not supported: ${format}`;\n super(msg, 'FORMAT_NOT_SUPPORTED');\n this.name = 'FormatNotSupportedError';\n }\n}\n\n/**\n * File size exceeds limits\n */\nexport class SizeLimitError extends FoundryError {\n constructor(\n public readonly actualSize: number,\n public readonly maxSize: number,\n context?: string\n ) {\n const actualMB = (actualSize / 1024 / 1024).toFixed(2);\n const maxMB = (maxSize / 1024 / 1024).toFixed(2);\n const msg = context\n ? `${context}: Size ${actualMB}MB exceeds limit ${maxMB}MB`\n : `Size ${actualMB}MB exceeds limit ${maxMB}MB`;\n super(msg, 'SIZE_LIMIT_EXCEEDED');\n this.name = 'SizeLimitError';\n }\n}\n\n/**\n * Path traversal or unsafe path detected\n */\nexport class PathTraversalError extends FoundryError {\n constructor(public readonly path: string) {\n super(`Unsafe path detected: ${path}`, 'PATH_TRAVERSAL');\n this.name = 'PathTraversalError';\n }\n}\n\n/**\n * Export operation would lose data\n */\nexport class DataLossError extends FoundryError {\n constructor(\n public readonly lostFields: string[],\n public readonly targetFormat: string\n ) {\n const fields = lostFields.slice(0, 3).join(', ');\n const more = lostFields.length > 3 ? ` and ${lostFields.length - 3} more` : '';\n super(\n `Export to ${targetFormat} would lose: ${fields}${more}`,\n 'DATA_LOSS'\n );\n this.name = 'DataLossError';\n }\n}\n\n/**\n * Check if an error is a FoundryError\n *\n * Uses Symbol.for() marker instead of instanceof to handle dual ESM/CJS package loading.\n * In dual-package environments, instanceof can fail if the error comes from a different\n * module instance (e.g., ESM vs CJS version of the same package). Symbol.for() creates\n * a global symbol shared across all module instances.\n */\nexport function isFoundryError(error: unknown): error is FoundryError {\n return (\n error instanceof Error &&\n FOUNDRY_ERROR_MARKER in error &&\n (error as Record<symbol, unknown>)[FOUNDRY_ERROR_MARKER] === true\n );\n}\n\n/**\n * Wrap unknown errors in a FoundryError\n */\nexport function wrapError(error: unknown, context?: string): FoundryError {\n if (isFoundryError(error)) {\n return error;\n }\n\n const message = error instanceof Error\n ? error.message\n : String(error);\n\n return new FoundryError(\n context ? `${context}: ${message}` : message,\n 'UNKNOWN_ERROR'\n );\n}\n","/**\n * Data URL Utilities\n *\n * Convert between Uint8Array buffers and data URLs.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n */\n\nimport type { BinaryData } from './binary.js';\nimport { encodeChunked as base64Encode, decode as base64Decode } from './base64.js';\nimport { ValidationError } from './errors.js';\n\n/**\n * Convert Uint8Array to data URL.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n *\n * @param buffer - Binary data to encode\n * @param mimeType - MIME type for the data URL (e.g., 'image/png', 'application/octet-stream')\n * @returns Data URL string\n *\n * @example\n * ```typescript\n * const png = new Uint8Array([...]);\n * const dataUrl = toDataURL(png, 'image/png');\n * // => \"data:image/png;base64,iVBORw0KGgo...\"\n * ```\n */\nexport function toDataURL(buffer: BinaryData, mimeType: string): string {\n // Use chunked encoding to handle large buffers without stack overflow\n const base64 = base64Encode(buffer);\n return `data:${mimeType};base64,${base64}`;\n}\n\n/**\n * Parse a data URL back to buffer and MIME type.\n * Validates the data URL format before parsing.\n *\n * @param dataUrl - Data URL string to parse\n * @returns Object containing the decoded buffer and MIME type\n * @throws Error if the data URL format is invalid\n *\n * @example\n * ```typescript\n * const { buffer, mimeType } = fromDataURL('data:image/png;base64,iVBORw0KGgo...');\n * // buffer: Uint8Array\n * // mimeType: 'image/png'\n * ```\n */\nexport function fromDataURL(dataUrl: string): { buffer: Uint8Array; mimeType: string } {\n // Validate data URL format\n if (!dataUrl.startsWith('data:')) {\n throw new ValidationError('Invalid data URL: must start with \"data:\"', 'dataUrl');\n }\n\n const commaIndex = dataUrl.indexOf(',');\n if (commaIndex === -1) {\n throw new ValidationError('Invalid data URL: missing comma separator', 'dataUrl');\n }\n\n const header = dataUrl.slice(5, commaIndex); // Skip 'data:'\n const data = dataUrl.slice(commaIndex + 1);\n\n // Parse header: [<mediatype>][;base64]\n let mimeType = 'text/plain';\n let isBase64 = false;\n\n const parts = header.split(';');\n for (const part of parts) {\n if (part === 'base64') {\n isBase64 = true;\n } else if (part && !part.includes('=')) {\n // MIME type (not a parameter like charset=utf-8)\n mimeType = part;\n }\n }\n\n if (!isBase64) {\n // URL-encoded text data\n throw new ValidationError('Non-base64 data URLs are not supported', 'dataUrl');\n }\n\n const buffer = base64Decode(data);\n return { buffer, mimeType };\n}\n\n/**\n * Check if a string is a valid data URL\n *\n * @param str - String to check\n * @returns true if the string is a valid data URL format\n */\nexport function isDataURL(str: string): boolean {\n if (!str.startsWith('data:')) return false;\n const commaIndex = str.indexOf(',');\n if (commaIndex === -1) return false;\n const header = str.slice(5, commaIndex);\n return header.includes('base64');\n}\n","/**\n * URI Utilities\n *\n * Handles different asset URI schemes used in character cards.\n * Supports: embeded://, embedded://, ccdefault:, https://, http://,\n * data:, file://, __asset:, asset:, chara-ext-asset_\n */\n\nexport type URIScheme =\n | 'embeded' // embeded:// (CharX standard, note intentional typo)\n | 'ccdefault' // ccdefault:\n | 'https' // https://\n | 'http' // http://\n | 'data' // data:mime;base64,...\n | 'file' // file://\n | 'internal' // Internal asset ID (UUID/string)\n | 'pngchunk' // PNG chunk reference (__asset:, asset:, chara-ext-asset_)\n | 'unknown';\n\nexport interface ParsedURI {\n scheme: URIScheme;\n originalUri: string;\n normalizedUri: string; // Normalized form of the URI\n path?: string; // For embeded://, file://\n url?: string; // For http://, https://\n data?: string; // For data: URIs\n mimeType?: string; // For data: URIs\n encoding?: string; // For data: URIs (e.g., base64)\n chunkKey?: string; // For pngchunk - the key/index to look up\n chunkCandidates?: string[]; // For pngchunk - all possible chunk keys to search\n}\n\n/**\n * Normalize a URI to its canonical form\n * Handles common typos and variant formats\n */\nexport function normalizeURI(uri: string): string {\n const trimmed = uri.trim();\n\n // Fix embedded:// -> embeded:// (common typo, CharX spec uses single 'd')\n if (trimmed.startsWith('embedded://')) {\n return 'embeded://' + trimmed.substring('embedded://'.length);\n }\n\n // Normalize PNG chunk references to pngchunk: scheme\n if (trimmed.startsWith('__asset:')) {\n const id = trimmed.substring('__asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('asset:')) {\n const id = trimmed.substring('asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_:')) {\n const id = trimmed.substring('chara-ext-asset_:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_')) {\n const id = trimmed.substring('chara-ext-asset_'.length);\n return `pngchunk:${id}`;\n }\n\n return trimmed;\n}\n\n/**\n * Parse a URI and determine its scheme and components\n */\nexport function parseURI(uri: string): ParsedURI {\n const trimmed = uri.trim();\n const normalized = normalizeURI(trimmed);\n\n // PNG chunk references (__asset:, asset:, chara-ext-asset_, pngchunk:)\n if (\n trimmed.startsWith('__asset:') ||\n trimmed.startsWith('asset:') ||\n trimmed.startsWith('chara-ext-asset_') ||\n trimmed.startsWith('pngchunk:')\n ) {\n let assetId: string;\n if (trimmed.startsWith('__asset:')) {\n assetId = trimmed.substring('__asset:'.length);\n } else if (trimmed.startsWith('asset:')) {\n assetId = trimmed.substring('asset:'.length);\n } else if (trimmed.startsWith('chara-ext-asset_:')) {\n assetId = trimmed.substring('chara-ext-asset_:'.length);\n } else if (trimmed.startsWith('pngchunk:')) {\n assetId = trimmed.substring('pngchunk:'.length);\n } else {\n assetId = trimmed.substring('chara-ext-asset_'.length);\n }\n\n // Generate all possible chunk key variations for lookup\n const candidates = [\n assetId, // \"0\" or \"filename.png\"\n trimmed, // Original URI\n `asset:${assetId}`, // \"asset:0\"\n `__asset:${assetId}`, // \"__asset:0\"\n `__asset_${assetId}`, // \"__asset_0\"\n `chara-ext-asset_${assetId}`, // \"chara-ext-asset_0\"\n `chara-ext-asset_:${assetId}`, // \"chara-ext-asset_:0\"\n `pngchunk:${assetId}`, // \"pngchunk:0\"\n ];\n\n return {\n scheme: 'pngchunk',\n originalUri: uri,\n normalizedUri: normalized,\n chunkKey: assetId,\n chunkCandidates: candidates,\n };\n }\n\n // ccdefault: - use default asset\n if (trimmed === 'ccdefault:' || trimmed.startsWith('ccdefault:')) {\n return {\n scheme: 'ccdefault',\n originalUri: uri,\n normalizedUri: normalized,\n };\n }\n\n // embeded:// or embedded:// (normalize typo)\n if (trimmed.startsWith('embeded://') || trimmed.startsWith('embedded://')) {\n const path = trimmed.startsWith('embeded://')\n ? trimmed.substring('embeded://'.length)\n : trimmed.substring('embedded://'.length);\n return {\n scheme: 'embeded',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // https://\n if (trimmed.startsWith('https://')) {\n return {\n scheme: 'https',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // http://\n if (trimmed.startsWith('http://')) {\n return {\n scheme: 'http',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // data: URIs\n if (trimmed.startsWith('data:')) {\n const parsed = parseDataURI(trimmed);\n return {\n scheme: 'data',\n originalUri: uri,\n normalizedUri: normalized,\n ...parsed,\n };\n }\n\n // file://\n if (trimmed.startsWith('file://')) {\n const path = trimmed.substring('file://'.length);\n return {\n scheme: 'file',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // Internal asset ID (alphanumeric/UUID format)\n if (/^[a-zA-Z0-9_-]+$/.test(trimmed)) {\n return {\n scheme: 'internal',\n originalUri: uri,\n normalizedUri: normalized,\n path: trimmed,\n };\n }\n\n // Unknown scheme\n return {\n scheme: 'unknown',\n originalUri: uri,\n normalizedUri: normalized,\n };\n}\n\n/**\n * Parse a data URI into its components\n * Format: data:[<mediatype>][;base64],<data>\n */\nfunction parseDataURI(uri: string): { mimeType?: string; encoding?: string; data?: string } {\n const match = uri.match(/^data:([^;,]+)?(;base64)?,(.*)$/);\n\n if (!match) {\n return {};\n }\n\n return {\n mimeType: match[1] || 'text/plain',\n encoding: match[2] ? 'base64' : undefined,\n data: match[3],\n };\n}\n\n/**\n * Check if extension is an image format\n */\nexport function isImageExt(ext: string): boolean {\n const imageExts = ['png', 'jpg', 'jpeg', 'webp', 'gif', 'avif', 'bmp', 'svg'];\n return imageExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is an audio format\n */\nexport function isAudioExt(ext: string): boolean {\n const audioExts = ['mp3', 'wav', 'ogg', 'flac', 'm4a', 'aac'];\n return audioExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is a video format\n */\nexport function isVideoExt(ext: string): boolean {\n const videoExts = ['mp4', 'webm', 'avi', 'mov', 'mkv'];\n return videoExts.includes(ext.toLowerCase());\n}\n\n/** Safe MIME types for data: URIs that can be used in href/src */\nconst SAFE_DATA_URI_MIME_TYPES = new Set([\n // Images (safe for img src)\n 'image/png',\n 'image/jpeg',\n 'image/gif',\n 'image/webp',\n 'image/avif',\n 'image/bmp',\n 'image/x-icon',\n // Audio (safe for audio src)\n 'audio/mpeg',\n 'audio/wav',\n 'audio/ogg',\n 'audio/flac',\n 'audio/mp4',\n 'audio/aac',\n // Video (safe for video src)\n 'video/mp4',\n 'video/webm',\n // Text/data (generally safe)\n 'text/plain',\n 'application/json',\n 'application/octet-stream',\n]);\n\n/** Potentially dangerous MIME types that should NOT be used in href/src */\nconst DANGEROUS_DATA_URI_MIME_TYPES = new Set([\n // Executable/script content\n 'text/html',\n 'text/javascript',\n 'application/javascript',\n 'application/x-javascript',\n 'text/css',\n 'image/svg+xml', // SVG can contain scripts\n 'application/xhtml+xml',\n 'application/xml',\n]);\n\n/**\n * Options for URI safety validation\n */\nexport interface URISafetyOptions {\n /** Allow http:// URIs (default: false) */\n allowHttp?: boolean;\n /** Allow file:// URIs (default: false) */\n allowFile?: boolean;\n /**\n * Allowed MIME types for data: URIs (default: all safe types).\n * Set to empty array to reject all data: URIs.\n * Set to undefined to use default safe list.\n */\n allowedDataMimes?: string[];\n}\n\n/**\n * Result of URI safety check with detailed information\n */\nexport interface URISafetyResult {\n /** Whether the URI is safe to use */\n safe: boolean;\n /** Reason if unsafe */\n reason?: string;\n /** Detected scheme */\n scheme: URIScheme;\n /** MIME type for data: URIs */\n mimeType?: string;\n}\n\n/**\n * Validate if a URI is safe to use (detailed version)\n *\n * @param uri - URI to validate\n * @param options - Safety options\n * @returns Detailed safety result\n */\nexport function checkURISafety(uri: string, options: URISafetyOptions = {}): URISafetyResult {\n const parsed = parseURI(uri);\n\n switch (parsed.scheme) {\n case 'embeded':\n case 'ccdefault':\n case 'internal':\n case 'https':\n case 'pngchunk':\n return { safe: true, scheme: parsed.scheme };\n\n case 'data': {\n const mimeType = parsed.mimeType || 'text/plain';\n\n // Check for explicitly dangerous MIME types\n if (DANGEROUS_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI with potentially dangerous MIME type: ${mimeType}`,\n };\n }\n\n // If custom allowed list is provided, check against it\n if (options.allowedDataMimes !== undefined) {\n if (options.allowedDataMimes.length === 0) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: 'Data URIs are not allowed',\n };\n }\n if (!options.allowedDataMimes.includes(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI MIME type not in allowed list: ${mimeType}`,\n };\n }\n }\n\n // Otherwise use default safe list\n if (!SAFE_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Unknown data URI MIME type: ${mimeType}`,\n };\n }\n\n return { safe: true, scheme: parsed.scheme, mimeType };\n }\n\n case 'http':\n if (options.allowHttp === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'HTTP URIs are not allowed' };\n\n case 'file':\n if (options.allowFile === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'File URIs are not allowed' };\n\n case 'unknown':\n default:\n return { safe: false, scheme: parsed.scheme, reason: 'Unknown URI scheme' };\n }\n}\n\n/**\n * Validate if a URI is safe to use (simple boolean version for backwards compatibility)\n *\n * @deprecated Use checkURISafety() for detailed safety information\n */\nexport function isURISafe(uri: string, options: { allowHttp?: boolean; allowFile?: boolean } = {}): boolean {\n return checkURISafety(uri, options).safe;\n}\n\n/**\n * Extract file extension from URI\n */\nexport function getExtensionFromURI(uri: string): string {\n const parsed = parseURI(uri);\n\n if (parsed.path) {\n const parts = parsed.path.split('.');\n if (parts.length > 1) {\n return parts[parts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.url) {\n const urlParts = parsed.url.split('?')[0]!.split('.');\n if (urlParts.length > 1) {\n return urlParts[urlParts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.mimeType) {\n return getExtFromMimeType(parsed.mimeType);\n }\n\n return 'unknown';\n}\n\n/**\n * Get MIME type from file extension\n */\nexport function getMimeTypeFromExt(ext: string): string {\n const extToMime: Record<string, string> = {\n // Images\n 'png': 'image/png',\n 'jpg': 'image/jpeg',\n 'jpeg': 'image/jpeg',\n 'webp': 'image/webp',\n 'gif': 'image/gif',\n 'avif': 'image/avif',\n 'svg': 'image/svg+xml',\n 'bmp': 'image/bmp',\n 'ico': 'image/x-icon',\n\n // Audio\n 'mp3': 'audio/mpeg',\n 'wav': 'audio/wav',\n 'ogg': 'audio/ogg',\n 'flac': 'audio/flac',\n 'm4a': 'audio/mp4',\n 'aac': 'audio/aac',\n\n // Video\n 'mp4': 'video/mp4',\n 'webm': 'video/webm',\n 'avi': 'video/x-msvideo',\n 'mov': 'video/quicktime',\n 'mkv': 'video/x-matroska',\n\n // Text/Data\n 'json': 'application/json',\n 'txt': 'text/plain',\n 'html': 'text/html',\n 'css': 'text/css',\n 'js': 'application/javascript',\n };\n\n return extToMime[ext.toLowerCase()] || 'application/octet-stream';\n}\n\n/**\n * Get file extension from MIME type\n */\nexport function getExtFromMimeType(mimeType: string): string {\n const mimeToExt: Record<string, string> = {\n 'image/png': 'png',\n 'image/jpeg': 'jpg',\n 'image/webp': 'webp',\n 'image/gif': 'gif',\n 'image/avif': 'avif',\n 'image/svg+xml': 'svg',\n 'image/bmp': 'bmp',\n 'image/x-icon': 'ico',\n 'audio/mpeg': 'mp3',\n 'audio/wav': 'wav',\n 'audio/ogg': 'ogg',\n 'audio/flac': 'flac',\n 'audio/mp4': 'm4a',\n 'audio/aac': 'aac',\n 'video/mp4': 'mp4',\n 'video/webm': 'webm',\n 'video/x-msvideo': 'avi',\n 'video/quicktime': 'mov',\n 'video/x-matroska': 'mkv',\n 'application/json': 'json',\n 'text/plain': 'txt',\n 'text/html': 'html',\n 'text/css': 'css',\n 'application/javascript': 'js',\n };\n\n return mimeToExt[mimeType] || 'bin';\n}\n\n/**\n * Build a data URI from binary data and MIME type\n */\nexport function buildDataURI(data: string, mimeType: string, isBase64 = true): string {\n if (isBase64) {\n return `data:${mimeType};base64,${data}`;\n }\n return `data:${mimeType},${encodeURIComponent(data)}`;\n}\n","/**\n * Image Analysis Utilities\n *\n * Detect properties of image files from binary data.\n */\n\nimport {\n type BinaryData,\n indexOf,\n fromLatin1,\n} from './binary.js';\n\n/**\n * Check if an image buffer contains animation data.\n * Supports: APNG, WebP (Animated), GIF\n */\nexport function isAnimatedImage(data: BinaryData, _mimeType?: string): boolean {\n // 1. WebP Detection\n // RIFF .... WEBP\n if (\n data.length > 12 &&\n data[0] === 0x52 && data[1] === 0x49 && data[2] === 0x46 && data[3] === 0x46 && // RIFF\n data[8] === 0x57 && data[9] === 0x45 && data[10] === 0x42 && data[11] === 0x50 // WEBP\n ) {\n // Check for VP8X chunk\n // VP8X chunk header: 'VP8X' (bytes 12-15)\n if (\n data[12] === 0x56 && data[13] === 0x50 && data[14] === 0x38 && data[15] === 0x58\n ) {\n // Flags byte is at offset 20 (16 + 4 bytes chunk size)\n // Animation bit is bit 1 (0x02)\n const flags = data[20];\n return (flags! & 0x02) !== 0;\n }\n return false;\n }\n\n // 2. PNG/APNG Detection\n // Signature: 89 50 4E 47 0D 0A 1A 0A\n if (\n data.length > 8 &&\n data[0] === 0x89 && data[1] === 0x50 && data[2] === 0x4E && data[3] === 0x47\n ) {\n // Search for 'acTL' chunk (Animation Control)\n // It must appear before IDAT.\n // Simple search: indexOf('acTL')\n // Note: theoretically 'acTL' string could appear in other data, but highly unlikely in valid PNG structure before IDAT\n // We can iterate chunks to be safe, but indexOf is faster for a quick check\n const actlSig = fromLatin1('acTL');\n const idatSig = fromLatin1('IDAT');\n \n const actlIndex = indexOf(data, actlSig);\n if (actlIndex === -1) return false;\n\n const idatIndex = indexOf(data, idatSig);\n // If acTL exists and is before the first IDAT (or IDAT not found yet), it's APNG\n return idatIndex === -1 || actlIndex < idatIndex;\n }\n\n // 3. GIF Detection\n // Signature: GIF87a or GIF89a\n if (\n data.length > 6 &&\n data[0] === 0x47 && data[1] === 0x49 && data[2] === 0x46 // GIF\n ) {\n // Check for NETSCAPE2.0 extension (looping animation)\n // This is a heuristic. Static GIFs are rare in this domain but possible.\n // Full frame counting is expensive. Presence of NETSCAPE block is a strong indicator.\n const netscape = fromLatin1('NETSCAPE2.0');\n return indexOf(data, netscape) !== -1;\n }\n\n return false;\n}\n","/**\n * UUID Generation Utilities\n *\n * Provides crypto-grade UUID v4 generation that works in Node.js,\n * browsers (secure contexts), and falls back gracefully.\n */\n\n/**\n * Format 16 random bytes as a UUID v4 string\n */\nfunction formatUUID(bytes: Uint8Array): string {\n const hex = Array.from(bytes, (b) => b.toString(16).padStart(2, '0')).join('');\n return `${hex.slice(0, 8)}-${hex.slice(8, 12)}-${hex.slice(12, 16)}-${hex.slice(16, 20)}-${hex.slice(20, 32)}`;\n}\n\n/**\n * Fallback UUID generation using Math.random()\n * Only used when crypto APIs are unavailable (rare)\n */\nfunction mathRandomUUID(): string {\n return 'xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx'.replace(/[xy]/g, (c) => {\n const r = (Math.random() * 16) | 0;\n const v = c === 'x' ? r : (r & 0x3) | 0x8;\n return v.toString(16);\n });\n}\n\n/**\n * Generate a cryptographically secure UUID v4.\n *\n * Uses crypto.randomUUID() when available (Node.js 19+, modern browsers).\n * Falls back to crypto.getRandomValues() if randomUUID is unavailable.\n * Last resort uses Math.random() (non-secure, emits warning in dev).\n *\n * @returns A valid RFC 4122 UUID v4 string\n *\n * @example\n * ```typescript\n * const id = generateUUID();\n * // => \"550e8400-e29b-41d4-a716-446655440000\"\n * ```\n */\nexport function generateUUID(): string {\n // Node.js 19+ or browser with secure context\n if (typeof crypto !== 'undefined' && typeof crypto.randomUUID === 'function') {\n return crypto.randomUUID();\n }\n\n // Fallback using crypto.getRandomValues (older Node/browsers)\n if (typeof crypto !== 'undefined' && typeof crypto.getRandomValues === 'function') {\n const bytes = new Uint8Array(16);\n crypto.getRandomValues(bytes);\n // Set version (4) and variant (RFC 4122)\n bytes[6] = (bytes[6]! & 0x0f) | 0x40; // Version 4\n bytes[8] = (bytes[8]! & 0x3f) | 0x80; // Variant 1\n return formatUUID(bytes);\n }\n\n // Last resort - non-secure fallback\n if (typeof process !== 'undefined' && process.env?.NODE_ENV === 'development') {\n console.warn('[character-foundry/core] generateUUID: Using insecure Math.random() fallback');\n }\n return mathRandomUUID();\n}\n\n/**\n * Validate if a string is a valid UUID v4\n *\n * @param uuid - String to validate\n * @returns true if valid UUID v4 format\n */\nexport function isValidUUID(uuid: string): boolean {\n return /^[0-9a-f]{8}-[0-9a-f]{4}-4[0-9a-f]{3}-[89ab][0-9a-f]{3}-[0-9a-f]{12}$/i.test(uuid);\n}\n","/**\n * Common Types\n *\n * Shared types used across all card formats.\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * ISO 8601 date string schema\n */\nexport const ISO8601Schema = z.string().datetime();\n\n/**\n * UUID string schema\n */\nexport const UUIDSchema = z.string().uuid();\n\n/**\n * Card specification version schema\n */\nexport const SpecSchema = z.enum(['v2', 'v3']);\n\n/**\n * Source format identifier schema\n */\nexport const SourceFormatSchema = z.enum([\n 'png_v2', // PNG with 'chara' chunk (v2)\n 'png_v3', // PNG with 'ccv3' chunk (v3)\n 'json_v2', // Raw JSON v2\n 'json_v3', // Raw JSON v3\n 'charx', // ZIP with card.json (v3 spec)\n 'charx_risu', // ZIP with card.json + module.risum\n 'charx_jpeg', // JPEG with appended ZIP (read-only)\n 'voxta', // VoxPkg format\n]);\n\n/**\n * Original JSON shape schema\n */\nexport const OriginalShapeSchema = z.enum(['wrapped', 'unwrapped', 'legacy']);\n\n/**\n * Asset type identifier schema\n */\nexport const AssetTypeSchema = z.enum([\n 'icon',\n 'background',\n 'emotion',\n 'user_icon',\n 'sound',\n 'video',\n 'custom',\n 'x-risu-asset',\n]);\n\n/**\n * Asset descriptor schema (v3 spec)\n */\nexport const AssetDescriptorSchema = z.object({\n type: AssetTypeSchema,\n uri: z.string(),\n name: z.string(),\n ext: z.string(),\n});\n\n/**\n * Extracted asset with binary data schema\n */\nexport const ExtractedAssetSchema = z.object({\n descriptor: AssetDescriptorSchema,\n data: z.instanceof(Uint8Array),\n mimeType: z.string(),\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * ISO 8601 date string\n */\nexport type ISO8601 = z.infer<typeof ISO8601Schema>;\n\n/**\n * UUID string\n */\nexport type UUID = z.infer<typeof UUIDSchema>;\n\n/**\n * Card specification version\n */\nexport type Spec = z.infer<typeof SpecSchema>;\n\n/**\n * Source format identifier\n */\nexport type SourceFormat = z.infer<typeof SourceFormatSchema>;\n\n/**\n * Original JSON shape\n */\nexport type OriginalShape = z.infer<typeof OriginalShapeSchema>;\n\n/**\n * Asset type identifier\n */\nexport type AssetType = z.infer<typeof AssetTypeSchema>;\n\n/**\n * Asset descriptor (v3 spec)\n */\nexport type AssetDescriptor = z.infer<typeof AssetDescriptorSchema>;\n\n/**\n * Extracted asset with binary data\n */\nexport type ExtractedAsset = z.infer<typeof ExtractedAssetSchema>;\n","/**\n * Character Card v2 Types\n *\n * Based on: https://github.com/malfoyslastname/character-card-spec-v2\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v2 cards\n */\nexport const CCv2LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n extensions: z.record(z.unknown()).optional(),\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n}).passthrough(); // Allow SillyTavern extensions like depth, probability, etc.\n\n/**\n * Character book (lorebook) schema for v2 cards\n */\nexport const CCv2CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv2LorebookEntrySchema),\n});\n\n/**\n * Character Card v2 data structure schema\n */\nexport const CCv2DataSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv2CharacterBookSchema.optional().nullable(),\n tags: z.array(z.string()).optional(),\n creator: z.string().optional(),\n character_version: z.string().optional(),\n extensions: z.record(z.unknown()).optional(),\n});\n\n/**\n * Wrapped v2 card format schema (modern tools)\n */\nexport const CCv2WrappedSchema = z.object({\n spec: z.literal('chara_card_v2'),\n spec_version: z.literal('2.0'),\n data: CCv2DataSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v2 cards\n */\nexport type CCv2LorebookEntry = z.infer<typeof CCv2LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v2 cards\n */\nexport type CCv2CharacterBook = z.infer<typeof CCv2CharacterBookSchema>;\n\n/**\n * Character Card v2 data structure\n */\nexport type CCv2Data = z.infer<typeof CCv2DataSchema>;\n\n/**\n * Wrapped v2 card format (modern tools)\n */\nexport type CCv2Wrapped = z.infer<typeof CCv2WrappedSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a wrapped v2 card\n */\nexport function isWrappedV2(data: unknown): data is CCv2Wrapped {\n return CCv2WrappedSchema.safeParse(data).success;\n}\n\n/**\n * Check if data looks like v2 card data (wrapped or unwrapped)\n */\nexport function isV2CardData(data: unknown): data is CCv2Data | CCv2Wrapped {\n return (\n CCv2WrappedSchema.safeParse(data).success ||\n CCv2DataSchema.safeParse(data).success\n );\n}\n\n/**\n * Parse and validate a wrapped v2 card\n */\nexport function parseWrappedV2(data: unknown): CCv2Wrapped {\n return CCv2WrappedSchema.parse(data);\n}\n\n/**\n * Parse and validate v2 card data\n */\nexport function parseV2Data(data: unknown): CCv2Data {\n return CCv2DataSchema.parse(data);\n}\n\n/**\n * Check if data looks like a wrapped V2 card structurally (without strict validation).\n * This is more lenient than isWrappedV2 - it just checks structure, not full schema validity.\n */\nexport function looksLikeWrappedV2(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v2' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n\n/**\n * Get v2 card data from wrapped or unwrapped format.\n *\n * Uses structural check instead of strict Zod validation to handle\n * malformed cards that have the right structure but missing/invalid fields.\n * The caller (e.g., ccv2ToCCv3) handles defaulting missing fields.\n */\nexport function getV2Data(card: CCv2Data | CCv2Wrapped): CCv2Data {\n // Use structural check - more lenient than isWrappedV2 schema validation\n if (looksLikeWrappedV2(card)) {\n return card.data as CCv2Data;\n }\n return card;\n}\n","/**\n * Character Card v3 Types\n *\n * Based on: https://github.com/kwaroran/character-card-spec-v3\n */\n\nimport { z } from 'zod';\nimport { AssetDescriptorSchema } from './common.js';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v3 cards\n */\nexport const CCv3LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific - also lenient with types since SillyTavern uses numbers for enums\n automation_id: z.string().optional(),\n role: z.union([z.enum(['system', 'user', 'assistant']), z.number().int()]).nullable().optional(),\n group: z.string().optional(),\n scan_frequency: z.number().int().nonnegative().optional(),\n probability: z.number().min(0).max(100).optional(), // Some tools use 0-100 instead of 0-1\n use_regex: z.boolean().optional(),\n depth: z.number().int().nonnegative().optional(),\n selective_logic: z.union([z.enum(['AND', 'NOT']), z.number().int()]).optional(),\n}).passthrough(); // Allow tool-specific extensions\n\n/**\n * Character book (lorebook) schema for v3 cards\n */\nexport const CCv3CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv3LorebookEntrySchema),\n});\n\n/**\n * Character Card v3 inner data structure schema.\n *\n * Note: Fields like group_only_greetings, creator, character_version, and tags\n * are technically \"required\" per V3 spec but rarely present in wild cards.\n * We use .default() to make parsing lenient while still producing valid output.\n */\nexport const CCv3DataInnerSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // \"Required\" per spec but often missing in wild - use defaults for leniency\n creator: z.string().default(''),\n character_version: z.string().default(''),\n tags: z.array(z.string()).default([]),\n group_only_greetings: z.array(z.string()).default([]),\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv3CharacterBookSchema.optional().nullable(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific\n assets: z.array(AssetDescriptorSchema).optional(),\n nickname: z.string().optional(),\n creator_notes_multilingual: z.record(z.string()).optional(),\n source: z.array(z.string()).optional(),\n creation_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n modification_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n});\n\n/**\n * Character Card v3 full structure schema\n */\nexport const CCv3DataSchema = z.object({\n spec: z.literal('chara_card_v3'),\n spec_version: z.literal('3.0'),\n data: CCv3DataInnerSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v3 cards\n */\nexport type CCv3LorebookEntry = z.infer<typeof CCv3LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v3 cards\n */\nexport type CCv3CharacterBook = z.infer<typeof CCv3CharacterBookSchema>;\n\n/**\n * Character Card v3 inner data structure\n */\nexport type CCv3DataInner = z.infer<typeof CCv3DataInnerSchema>;\n\n/**\n * Character Card v3 full structure\n */\nexport type CCv3Data = z.infer<typeof CCv3DataSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a v3 card\n */\nexport function isV3Card(data: unknown): data is CCv3Data {\n return CCv3DataSchema.safeParse(data).success;\n}\n\n/**\n * Parse and validate a v3 card\n */\nexport function parseV3Card(data: unknown): CCv3Data {\n return CCv3DataSchema.parse(data);\n}\n\n/**\n * Parse and validate v3 card inner data\n */\nexport function parseV3DataInner(data: unknown): CCv3DataInner {\n return CCv3DataInnerSchema.parse(data);\n}\n\n/**\n * Get v3 card inner data\n */\nexport function getV3Data(card: CCv3Data): CCv3DataInner {\n return card.data;\n}\n\n/**\n * Check if data looks like a V3 card structurally (without strict validation).\n * More lenient than isV3Card - just checks structure, not full schema validity.\n */\nexport function looksLikeV3Card(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v3' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n","/**\n * RisuAI Extension Types\n *\n * These extensions are preserved as opaque blobs.\n * We do NOT interpret or transform the script contents.\n */\n\n/**\n * Risu emotions mapping (v2 style)\n * Format: [name, uri][]\n */\nexport type RisuEmotions = [string, string][];\n\n/**\n * Risu additional assets (v3 style)\n * Format: [name, uri, type][]\n */\nexport type RisuAdditionalAssets = [string, string, string][];\n\n/**\n * Risu depth prompt configuration\n */\nexport interface RisuDepthPrompt {\n depth: number;\n prompt: string;\n}\n\n/**\n * Risu extensions in card.extensions.risuai\n * Preserved as opaque - we don't interpret script contents\n */\nexport interface RisuExtensions {\n // Emotion assets\n emotions?: RisuEmotions;\n additionalAssets?: RisuAdditionalAssets;\n\n // Script data - OPAQUE, do not parse\n triggerscript?: unknown;\n customScripts?: unknown;\n\n // Voice/TTS settings\n vits?: Record<string, string>;\n\n // Depth prompt\n depth_prompt?: RisuDepthPrompt;\n\n // Other Risu-specific fields\n [key: string]: unknown;\n}\n\n/**\n * CharX x_meta entry (PNG chunk metadata preservation)\n */\nexport interface CharxMetaEntry {\n type?: string; // e.g., 'WEBP', 'PNG', 'JPEG'\n [key: string]: unknown;\n}\n\n/**\n * Check if card has Risu extensions\n */\nexport function hasRisuExtensions(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n return 'risuai' in extensions || 'risu' in extensions;\n}\n\n/**\n * Check if card has Risu scripts (triggerscript or customScripts)\n */\nexport function hasRisuScripts(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n const risu = extensions.risuai as RisuExtensions | undefined;\n if (!risu) return false;\n return !!risu.triggerscript || !!risu.customScripts;\n}\n\n/**\n * Check if card has depth prompt\n * Checks both SillyTavern style (extensions.depth_prompt) and Risu style (extensions.risuai.depth_prompt)\n */\nexport function hasDepthPrompt(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n // SillyTavern top-level depth_prompt\n if ('depth_prompt' in extensions && extensions.depth_prompt) return true;\n // Risu-style depth_prompt\n const risu = extensions.risuai as RisuExtensions | undefined;\n return !!risu?.depth_prompt;\n}\n","/**\n * Normalized Card Types\n *\n * Unified view of card data regardless of source format.\n * This is a computed/virtual representation, not stored.\n */\n\nimport type { CCv3CharacterBook } from './ccv3.js';\n\n/**\n * Normalized card representation\n * Provides unified access to card data from any format\n */\nexport interface NormalizedCard {\n // Core fields (always present)\n name: string;\n description: string;\n personality: string;\n scenario: string;\n firstMes: string;\n mesExample: string;\n\n // Optional prompts\n systemPrompt?: string;\n postHistoryInstructions?: string;\n\n // Arrays\n alternateGreetings: string[];\n groupOnlyGreetings: string[];\n tags: string[];\n\n // Metadata\n creator?: string;\n creatorNotes?: string;\n characterVersion?: string;\n\n // Character book (v3 format)\n characterBook?: CCv3CharacterBook;\n\n // Extensions (preserved as-is)\n extensions: Record<string, unknown>;\n}\n\n/**\n * Create empty normalized card with defaults\n */\nexport function createEmptyNormalizedCard(): NormalizedCard {\n return {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n firstMes: '',\n mesExample: '',\n alternateGreetings: [],\n groupOnlyGreetings: [],\n tags: [],\n extensions: {},\n };\n}\n\n/**\n * Derived features extracted from card (not stored in card)\n */\nexport interface DerivedFeatures {\n // Content flags\n hasAlternateGreetings: boolean;\n alternateGreetingsCount: number;\n /** Total greetings = first_mes (1) + alternate_greetings */\n totalGreetingsCount: number;\n hasLorebook: boolean;\n lorebookEntriesCount: number;\n hasEmbeddedImages: boolean;\n embeddedImagesCount: number;\n hasGallery: boolean;\n\n // Format-specific\n hasRisuExtensions: boolean;\n hasRisuScripts: boolean;\n hasDepthPrompt: boolean;\n hasVoxtaAppearance: boolean;\n\n // Token counts (estimated)\n tokens: {\n description: number;\n personality: number;\n scenario: number;\n firstMes: number;\n mesExample: number;\n systemPrompt: number;\n total: number;\n };\n}\n\n/**\n * Create empty derived features\n */\nexport function createEmptyFeatures(): DerivedFeatures {\n return {\n hasAlternateGreetings: false,\n alternateGreetingsCount: 0,\n totalGreetingsCount: 1, // first_mes always counts as 1\n hasLorebook: false,\n lorebookEntriesCount: 0,\n hasEmbeddedImages: false,\n embeddedImagesCount: 0,\n hasGallery: false,\n hasRisuExtensions: false,\n hasRisuScripts: false,\n hasDepthPrompt: false,\n hasVoxtaAppearance: false,\n tokens: {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n },\n };\n}\n","/**\n * Feature Derivation\n *\n * Canonical feature extraction from character cards.\n * Eliminates duplicate implementations across Archive, Federation, and Architect.\n */\n\nimport type { CCv2Data } from './ccv2.js';\nimport type { CCv3DataInner } from './ccv3.js';\nimport type { DerivedFeatures } from './normalized.js';\nimport { hasRisuExtensions, hasRisuScripts, hasDepthPrompt } from './risu.js';\n\n/**\n * Derive features from a character card (V2 or V3 format).\n *\n * This is the canonical implementation - all apps should use this\n * rather than implementing their own feature detection.\n *\n * @param card - Either CCv2Data or CCv3DataInner (unwrapped)\n * @returns DerivedFeatures with all feature flags populated\n *\n * @example\n * ```typescript\n * import { deriveFeatures, parseV3Card } from '@character-foundry/schemas';\n *\n * const card = parseV3Card(data);\n * const features = deriveFeatures(card.data);\n *\n * if (features.hasLorebook) {\n * console.log(`Found ${features.lorebookEntriesCount} lorebook entries`);\n * }\n * ```\n */\nexport function deriveFeatures(card: CCv2Data | CCv3DataInner): DerivedFeatures {\n // Detect format by checking for V3-specific field\n const isV3 = 'assets' in card;\n\n // Alternate greetings\n const altGreetings = card.alternate_greetings ?? [];\n const hasAlternateGreetings = altGreetings.length > 0;\n const alternateGreetingsCount = altGreetings.length;\n // Total = first_mes (1) + alternate_greetings\n const totalGreetingsCount = 1 + alternateGreetingsCount;\n\n // Lorebook\n const characterBook = card.character_book;\n const hasLorebook = !!characterBook && characterBook.entries.length > 0;\n const lorebookEntriesCount = characterBook?.entries.length ?? 0;\n\n // Assets (V3 only) - check for visual asset types\n const assets = isV3 ? (card as CCv3DataInner).assets ?? [] : [];\n const imageAssetTypes = ['icon', 'background', 'emotion', 'custom'];\n const imageAssets = assets.filter(\n (a) =>\n imageAssetTypes.includes(a.type) ||\n ['png', 'jpg', 'jpeg', 'webp', 'gif'].includes(a.ext.toLowerCase()),\n );\n const hasGallery = imageAssets.length > 0;\n\n // Embedded images - check for data URLs in text fields\n const embeddedImageCount = countEmbeddedImages(card);\n const hasEmbeddedImages = embeddedImageCount > 0;\n\n // Extensions\n const extensions = card.extensions ?? {};\n const hasRisu = hasRisuExtensions(extensions);\n const hasScripts = hasRisuScripts(extensions);\n const hasDepth = hasDepthPrompt(extensions);\n const hasVoxta = checkVoxtaAppearance(extensions);\n\n // Token counts - initialize to zero (actual counting happens in tokenizers package)\n const tokens = {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n };\n\n return {\n hasAlternateGreetings,\n alternateGreetingsCount,\n totalGreetingsCount,\n hasLorebook,\n lorebookEntriesCount,\n hasEmbeddedImages,\n embeddedImagesCount: embeddedImageCount,\n hasGallery,\n hasRisuExtensions: hasRisu,\n hasRisuScripts: hasScripts,\n hasDepthPrompt: hasDepth,\n hasVoxtaAppearance: hasVoxta,\n tokens,\n };\n}\n\n/**\n * Count embedded images (data URLs) in card text fields.\n * Looks for base64-encoded images in description, personality, scenario, etc.\n */\nfunction countEmbeddedImages(card: CCv2Data | CCv3DataInner): number {\n const textFields = [\n card.description,\n card.personality,\n card.scenario,\n card.first_mes,\n card.mes_example,\n card.creator_notes,\n card.system_prompt,\n card.post_history_instructions,\n ...(card.alternate_greetings ?? []),\n ].filter((field): field is string => typeof field === 'string');\n\n // Add group_only_greetings if V3\n if ('group_only_greetings' in card) {\n textFields.push(...(card.group_only_greetings ?? []));\n }\n\n let count = 0;\n const dataUrlPattern = /data:image\\/[^;]+;base64,/g;\n\n for (const text of textFields) {\n const matches = text.match(dataUrlPattern);\n if (matches) {\n count += matches.length;\n }\n }\n\n return count;\n}\n\n/**\n * Check if card has Voxta appearance data.\n * Voxta stores appearance in extensions.voxta.appearance\n */\nfunction checkVoxtaAppearance(extensions: Record<string, unknown>): boolean {\n if (!extensions.voxta) return false;\n const voxta = extensions.voxta as Record<string, unknown>;\n return !!voxta.appearance;\n}\n","/**\n * Format Detection\n *\n * Detect card specification version from JSON data.\n */\n\nimport type { Spec } from './common.js';\n\n/**\n * V3-only fields that indicate a V3 card\n */\nconst V3_ONLY_FIELDS = ['group_only_greetings', 'creation_date', 'modification_date', 'assets'] as const;\n\n/**\n * Result from detailed spec detection\n */\nexport interface SpecDetectionResult {\n /** Detected spec version */\n spec: Spec | null;\n /** Confidence level of detection */\n confidence: 'high' | 'medium' | 'low';\n /** What fields/values indicated this spec */\n indicators: string[];\n /** Anomalies or inconsistencies detected */\n warnings: string[];\n}\n\n/**\n * Detect card spec version from parsed JSON\n * Returns 'v2', 'v3', or null if not recognized\n */\nexport function detectSpec(data: unknown): Spec | null {\n return detectSpecDetailed(data).spec;\n}\n\n/**\n * Detailed spec detection with confidence and reasoning.\n * Useful for debugging and logging.\n */\nexport function detectSpecDetailed(data: unknown): SpecDetectionResult {\n const result: SpecDetectionResult = {\n spec: null,\n confidence: 'low',\n indicators: [],\n warnings: [],\n };\n\n if (!data || typeof data !== 'object') {\n result.indicators.push('Input is not an object');\n return result;\n }\n\n const obj = data as Record<string, unknown>;\n const dataObj = (obj.data && typeof obj.data === 'object' ? obj.data : null) as Record<\n string,\n unknown\n > | null;\n\n // Check for explicit spec markers (HIGH confidence)\n\n // Explicit v3 spec marker\n if (obj.spec === 'chara_card_v3') {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v3\"');\n\n // Check for inconsistencies\n if (obj.spec_version && obj.spec_version !== '3.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v3 spec`);\n }\n\n return result;\n }\n\n // Explicit v2 spec marker\n if (obj.spec === 'chara_card_v2') {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v2\"');\n\n // Check for inconsistencies - V3-only fields in V2 card\n if (dataObj) {\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n result.warnings.push(`V3-only field \"${field}\" found in V2 card`);\n }\n }\n }\n\n if (obj.spec_version && obj.spec_version !== '2.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v2 spec`);\n }\n\n return result;\n }\n\n // Check spec_version field (HIGH confidence)\n if (typeof obj.spec_version === 'string') {\n if (obj.spec_version.startsWith('3')) {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"3\"`);\n return result;\n }\n if (obj.spec_version.startsWith('2')) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"2\"`);\n return result;\n }\n }\n\n if (obj.spec_version === 2.0 || obj.spec_version === 2) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version is numeric ${obj.spec_version}`);\n return result;\n }\n\n // Check for V3-only fields (MEDIUM confidence)\n if (dataObj) {\n const v3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n v3Fields.push(field);\n }\n }\n\n if (v3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields: ${v3Fields.join(', ')}`);\n return result;\n }\n }\n\n // Check root level for V3-only fields (also MEDIUM confidence)\n const rootV3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in obj) {\n rootV3Fields.push(field);\n }\n }\n if (rootV3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields at root: ${rootV3Fields.join(', ')}`);\n result.warnings.push('V3 fields found at root level instead of data object');\n return result;\n }\n\n // Wrapped format with data object (MEDIUM confidence)\n if (obj.spec && dataObj) {\n const dataName = dataObj.name;\n if (dataName && typeof dataName === 'string') {\n // Infer from spec string\n if (typeof obj.spec === 'string') {\n if (obj.spec.includes('v3') || obj.spec.includes('3')) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v3\" or \"3\"`);\n return result;\n }\n if (obj.spec.includes('v2') || obj.spec.includes('2')) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v2\" or \"2\"`);\n return result;\n }\n }\n // Default wrapped format to v3 (modern)\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push('Has wrapped format with spec and data.name');\n return result;\n }\n }\n\n // Unwrapped format - V1/V2 like structure (MEDIUM confidence)\n if (obj.name && typeof obj.name === 'string') {\n if ('description' in obj || 'personality' in obj || 'scenario' in obj) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push('Unwrapped format with name, description/personality/scenario');\n return result;\n }\n }\n\n // Check if data object has card-like structure without spec (LOW confidence)\n if (dataObj && typeof dataObj.name === 'string') {\n if ('description' in dataObj || 'personality' in dataObj) {\n result.spec = 'v2';\n result.confidence = 'low';\n result.indicators.push('Has data object with name and card fields, but no spec');\n result.warnings.push('Missing spec field');\n return result;\n }\n }\n\n result.indicators.push('No card structure detected');\n return result;\n}\n\n/**\n * Check if card has a lorebook\n */\nexport function hasLorebook(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Check wrapped format\n const wrapped = obj.data as Record<string, unknown> | undefined;\n if (wrapped?.character_book) {\n const book = wrapped.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Check if data looks like a valid card structure\n */\nexport function looksLikeCard(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Has explicit spec marker\n if (obj.spec === 'chara_card_v2' || obj.spec === 'chara_card_v3') {\n return true;\n }\n\n // Has wrapped data with name\n if (obj.data && typeof obj.data === 'object') {\n const dataObj = obj.data as Record<string, unknown>;\n if (typeof dataObj.name === 'string' && dataObj.name.length > 0) {\n return true;\n }\n }\n\n // Has unwrapped card-like structure\n if (typeof obj.name === 'string' && obj.name.length > 0) {\n if ('description' in obj || 'personality' in obj || 'first_mes' in obj) {\n return true;\n }\n }\n\n return false;\n}\n","/**\n * Card Normalizer\n *\n * Handles normalization of malformed card data from various sources.\n * Fixes common issues like wrong spec values, misplaced fields, missing required fields.\n */\n\nimport type { CCv2Data, CCv2Wrapped, CCv2CharacterBook, CCv2LorebookEntry } from './ccv2.js';\nimport type { CCv3Data, CCv3CharacterBook, CCv3LorebookEntry } from './ccv3.js';\nimport { detectSpec } from './detection.js';\n\n/**\n * Position values as numbers (non-standard) and their string equivalents\n */\nconst POSITION_MAP: Record<number, 'before_char' | 'after_char'> = {\n 0: 'before_char',\n 1: 'after_char',\n};\n\n/**\n * V3-only lorebook entry fields that should be moved to extensions for V2\n */\nconst V3_ONLY_ENTRY_FIELDS = [\n 'probability',\n 'depth',\n 'group',\n 'scan_frequency',\n 'use_regex',\n 'selective_logic',\n 'role',\n 'automation_id',\n] as const;\n\n/**\n * Required V2 card fields with their defaults\n */\nconst V2_REQUIRED_DEFAULTS: Partial<CCv2Data> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n};\n\n/**\n * Required V3 card fields with their defaults\n */\nconst V3_REQUIRED_DEFAULTS: Partial<CCv3Data['data']> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n creator: '',\n character_version: '1.0',\n tags: [],\n group_only_greetings: [],\n};\n\n/**\n * Fields that belong at root level for wrapped format\n */\nconst _ROOT_FIELDS = ['spec', 'spec_version', 'data'] as const;\n\n/**\n * Fields that belong in the data object\n */\nconst DATA_FIELDS = [\n 'name',\n 'description',\n 'personality',\n 'scenario',\n 'first_mes',\n 'mes_example',\n 'creator_notes',\n 'system_prompt',\n 'post_history_instructions',\n 'alternate_greetings',\n 'character_book',\n 'tags',\n 'creator',\n 'character_version',\n 'extensions',\n 'assets',\n 'nickname',\n 'creator_notes_multilingual',\n 'source',\n 'creation_date',\n 'modification_date',\n 'group_only_greetings',\n] as const;\n\n/**\n * Deep clone an object without mutating the original\n */\nfunction deepClone<T>(obj: T): T {\n if (obj === null || obj === undefined) {\n return obj;\n }\n if (Array.isArray(obj)) {\n return obj.map((item) => deepClone(item)) as T;\n }\n if (typeof obj === 'object') {\n const result: Record<string, unknown> = {};\n for (const [key, value] of Object.entries(obj as Record<string, unknown>)) {\n result[key] = deepClone(value);\n }\n return result as T;\n }\n return obj;\n}\n\n/**\n * Check if a timestamp is in milliseconds (13+ digits)\n */\nfunction isMilliseconds(timestamp: number): boolean {\n // Timestamps before year 2001 in seconds: < 1000000000\n // Timestamps in milliseconds are typically 13 digits: 1000000000000+\n return timestamp > 10000000000;\n}\n\n/**\n * CardNormalizer - handles normalization of malformed card data\n */\nexport const CardNormalizer = {\n /**\n * Normalize card data to valid schema format.\n *\n * Handles:\n * - Fixing spec/spec_version values\n * - Moving misplaced fields to correct locations\n * - Adding missing required fields with defaults\n * - Handling hybrid formats (fields at root AND in data object)\n *\n * @param data - Raw card data (potentially malformed)\n * @param spec - Target spec version\n * @returns Normalized card data (does not mutate input)\n */\n normalize(data: unknown, spec: 'v2' | 'v3'): CCv2Wrapped | CCv3Data {\n if (!data || typeof data !== 'object') {\n // Return minimal valid card\n if (spec === 'v3') {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: { ...V3_REQUIRED_DEFAULTS } as CCv3Data['data'],\n };\n }\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: { ...V2_REQUIRED_DEFAULTS } as CCv2Data,\n };\n }\n\n const obj = data as Record<string, unknown>;\n const result: Record<string, unknown> = {};\n\n // Build merged data object from root fields + existing data object\n const existingData = (obj.data && typeof obj.data === 'object' ? obj.data : {}) as Record<\n string,\n unknown\n >;\n const mergedData: Record<string, unknown> = {};\n\n // Copy existing data first\n for (const [key, value] of Object.entries(existingData)) {\n mergedData[key] = deepClone(value);\n }\n\n // Move any misplaced root-level data fields into data object\n // (ChubAI hybrid format fix)\n for (const field of DATA_FIELDS) {\n if (field in obj && !(field in mergedData)) {\n mergedData[field] = deepClone(obj[field]);\n }\n }\n\n // Handle character_book: null -> remove entirely\n if (mergedData.character_book === null) {\n delete mergedData.character_book;\n }\n\n // Normalize character_book if present\n if (mergedData.character_book && typeof mergedData.character_book === 'object') {\n mergedData.character_book = this.normalizeCharacterBook(\n mergedData.character_book as Record<string, unknown>,\n spec\n );\n }\n\n // Apply defaults for required fields\n const defaults = spec === 'v3' ? V3_REQUIRED_DEFAULTS : V2_REQUIRED_DEFAULTS;\n for (const [key, defaultValue] of Object.entries(defaults)) {\n if (!(key in mergedData) || mergedData[key] === undefined) {\n mergedData[key] = Array.isArray(defaultValue) ? [...defaultValue] : defaultValue;\n }\n }\n\n // Ensure arrays are actually arrays\n if (mergedData.tags && !Array.isArray(mergedData.tags)) {\n mergedData.tags = [];\n }\n if (mergedData.alternate_greetings && !Array.isArray(mergedData.alternate_greetings)) {\n mergedData.alternate_greetings = [];\n }\n if (spec === 'v3') {\n if (\n mergedData.group_only_greetings &&\n !Array.isArray(mergedData.group_only_greetings)\n ) {\n mergedData.group_only_greetings = [];\n }\n }\n\n // Build result with correct spec\n if (spec === 'v3') {\n result.spec = 'chara_card_v3';\n result.spec_version = '3.0';\n result.data = this.fixTimestampsInner(mergedData);\n } else {\n result.spec = 'chara_card_v2';\n result.spec_version = '2.0';\n result.data = mergedData;\n }\n\n return result as unknown as CCv2Wrapped | CCv3Data;\n },\n\n /**\n * Normalize a character book (lorebook).\n *\n * Handles:\n * - Ensuring required fields exist\n * - Converting numeric position values to string enums\n * - Moving V3-only fields to extensions for V2 compatibility\n *\n * @param book - Raw character book data\n * @param spec - Target spec version\n * @returns Normalized character book\n */\n normalizeCharacterBook(\n book: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2CharacterBook | CCv3CharacterBook {\n const result: Record<string, unknown> = {};\n\n // Copy book-level fields\n if (book.name !== undefined) result.name = book.name;\n if (book.description !== undefined) result.description = book.description;\n if (book.scan_depth !== undefined) result.scan_depth = book.scan_depth;\n if (book.token_budget !== undefined) result.token_budget = book.token_budget;\n if (book.recursive_scanning !== undefined)\n result.recursive_scanning = book.recursive_scanning;\n if (book.extensions !== undefined) result.extensions = deepClone(book.extensions);\n\n // Normalize entries\n const entries = Array.isArray(book.entries) ? book.entries : [];\n result.entries = entries.map((entry) =>\n this.normalizeEntry(entry as Record<string, unknown>, spec)\n );\n\n return result as unknown as CCv2CharacterBook | CCv3CharacterBook;\n },\n\n /**\n * Normalize a single lorebook entry.\n *\n * Handles:\n * - Converting numeric position to string enum\n * - Moving V3-only fields to extensions for V2\n * - Ensuring required fields exist\n *\n * @param entry - Raw entry data\n * @param spec - Target spec version\n * @returns Normalized entry\n */\n normalizeEntry(\n entry: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2LorebookEntry | CCv3LorebookEntry {\n const result: Record<string, unknown> = {};\n\n // Required fields with defaults\n result.keys = Array.isArray(entry.keys) ? [...entry.keys] : [];\n result.content = typeof entry.content === 'string' ? entry.content : '';\n result.enabled = entry.enabled !== false; // default true\n result.insertion_order =\n typeof entry.insertion_order === 'number' ? entry.insertion_order : 0;\n\n // For V2, extensions is required\n if (spec === 'v2') {\n result.extensions =\n entry.extensions && typeof entry.extensions === 'object'\n ? deepClone(entry.extensions)\n : {};\n }\n\n // Optional fields\n if (entry.case_sensitive !== undefined) result.case_sensitive = entry.case_sensitive;\n if (entry.name !== undefined) result.name = entry.name;\n if (entry.priority !== undefined) result.priority = entry.priority;\n if (entry.id !== undefined) result.id = entry.id;\n if (entry.comment !== undefined) result.comment = entry.comment;\n if (entry.selective !== undefined) result.selective = entry.selective;\n if (entry.secondary_keys !== undefined) {\n result.secondary_keys = Array.isArray(entry.secondary_keys)\n ? [...entry.secondary_keys]\n : [];\n }\n if (entry.constant !== undefined) result.constant = entry.constant;\n\n // Position: convert numeric to string enum\n if (entry.position !== undefined) {\n if (typeof entry.position === 'number') {\n result.position = POSITION_MAP[entry.position] || 'before_char';\n } else if (entry.position === 'before_char' || entry.position === 'after_char') {\n result.position = entry.position;\n }\n }\n\n // Handle V3-only fields\n if (spec === 'v3') {\n // Copy V3 fields directly\n if (entry.extensions !== undefined) result.extensions = deepClone(entry.extensions);\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n result[field] = entry[field];\n }\n }\n } else {\n // V2: Move V3-only fields to extensions\n const ext = (result.extensions || {}) as Record<string, unknown>;\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n ext[field] = entry[field];\n }\n }\n result.extensions = ext;\n }\n\n return result as unknown as CCv2LorebookEntry | CCv3LorebookEntry;\n },\n\n /**\n * Fix CharacterTavern timestamp format (milliseconds -> seconds).\n *\n * CCv3 spec requires timestamps in seconds (Unix epoch).\n * CharacterTavern exports timestamps in milliseconds.\n *\n * @param data - V3 card data\n * @returns Card data with fixed timestamps (does not mutate input)\n */\n fixTimestamps(data: CCv3Data): CCv3Data {\n const result = deepClone(data);\n result.data = this.fixTimestampsInner(\n result.data as unknown as Record<string, unknown>\n ) as unknown as CCv3Data['data'];\n return result;\n },\n\n /**\n * Internal: fix timestamps in data object\n */\n fixTimestampsInner(data: Record<string, unknown>): Record<string, unknown> {\n const result = { ...data };\n\n if (typeof result.creation_date === 'number') {\n if (isMilliseconds(result.creation_date)) {\n result.creation_date = Math.floor(result.creation_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.creation_date as number) < 0) {\n delete result.creation_date;\n }\n }\n\n if (typeof result.modification_date === 'number') {\n if (isMilliseconds(result.modification_date)) {\n result.modification_date = Math.floor(result.modification_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.modification_date as number) < 0) {\n delete result.modification_date;\n }\n }\n\n return result;\n },\n\n /**\n * Auto-detect spec and normalize.\n *\n * @param data - Raw card data\n * @returns Normalized card data, or null if not a valid card\n */\n autoNormalize(data: unknown): CCv2Wrapped | CCv3Data | null {\n const spec = detectSpec(data);\n if (!spec) return null;\n\n // V1 cards get upgraded to V2\n const targetSpec = spec === 'v3' ? 'v3' : 'v2';\n return this.normalize(data, targetSpec);\n },\n};\n\nexport type { CCv2Wrapped, CCv3Data };\n","/**\n * Validation Utilities\n *\n * Helper functions for Zod validation with Foundry error integration.\n */\n\nimport { z } from 'zod';\n\n/**\n * Convert Zod error to human-readable message\n */\nexport function zodErrorToMessage(zodError: z.ZodError, context?: string): string {\n const messages = zodError.errors.map((err) => {\n const path = err.path.length > 0 ? `${err.path.join('.')}: ` : '';\n return `${path}${err.message}`;\n });\n\n const message = messages.join('; ');\n return context ? `${context} - ${message}` : message;\n}\n\n/**\n * Get the first error field from Zod error\n */\nexport function getFirstErrorField(zodError: z.ZodError): string | undefined {\n return zodError.errors[0]?.path[0]?.toString();\n}\n\n/**\n * Safe parse with detailed error information\n */\nexport function safeParse<T>(\n schema: z.ZodSchema<T>,\n data: unknown\n): { success: true; data: T } | { success: false; error: string; field?: string } {\n const result = schema.safeParse(data);\n\n if (result.success) {\n return { success: true, data: result.data };\n }\n\n return {\n success: false,\n error: zodErrorToMessage(result.error),\n field: getFirstErrorField(result.error),\n };\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACOA,oBAAwB;;;AGQjB,SAAS,aAAa,MAAkB,QAAwB;AACrE,UACG,KAAK,MAAM,KAAM,KACjB,KAAK,SAAS,CAAC,KAAM,KACrB,KAAK,SAAS,CAAC,KAAM,IACtB,KAAK,SAAS,CAAC,OACX;AACR;AAKO,SAAS,cAAc,MAAkB,OAAe,QAAsB;AACnF,OAAK,MAAM,IAAK,UAAU,KAAM;AAChC,OAAK,SAAS,CAAC,IAAK,UAAU,KAAM;AACpC,OAAK,SAAS,CAAC,IAAK,UAAU,IAAK;AACnC,OAAK,SAAS,CAAC,IAAI,QAAQ;AAC7B;AAoBO,SAAS,QAAQ,MAAkB,QAAoB,YAAY,GAAW;AACnF,QAAO,UAAS,IAAI,WAAW,KAAK,KAAK,SAAS,OAAO,QAAQ,KAAK;AACpE,aAAS,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;AACtC,UAAI,KAAK,IAAI,CAAC,MAAM,OAAO,CAAC,EAAG,UAAS;IAC1C;AACA,WAAO;EACT;AACA,SAAO;AACT;AAKO,SAAS,UAAU,QAAkC;AAC1D,QAAM,cAAc,OAAO,OAAO,CAAC,KAAK,QAAQ,MAAM,IAAI,QAAQ,CAAC;AACnE,QAAM,SAAS,IAAI,WAAW,WAAW;AACzC,MAAI,SAAS;AACb,aAAW,OAAO,QAAQ;AACxB,WAAO,IAAI,KAAK,MAAM;AACtB,cAAU,IAAI;EAChB;AACA,SAAO;AACT;AAKO,SAAS,MAAM,MAAkB,OAAe,KAA0B;AAC/E,SAAO,KAAK,SAAS,OAAO,GAAG;AACjC;AAYO,SAAS,WAAW,KAAyB;AAClD,SAAO,IAAI,YAAY,EAAE,OAAO,GAAG;AACrC;AAKO,SAAS,SAAS,MAA0B;AACjD,SAAO,IAAI,YAAY,EAAE,OAAO,IAAI;AACtC;AAKO,SAAS,WAAW,KAAyB;AAClD,QAAM,SAAS,IAAI,WAAW,IAAI,MAAM;AACxC,WAAS,IAAI,GAAG,IAAI,IAAI,QAAQ,KAAK;AACnC,WAAO,CAAC,IAAI,IAAI,WAAW,CAAC,IAAI;EAClC;AACA,SAAO;AACT;AAKO,SAAS,SAAS,MAA0B;AACjD,MAAI,SAAS;AACb,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,cAAU,OAAO,aAAa,KAAK,CAAC,CAAE;EACxC;AACA,SAAO;AACT;AAgBO,SAAS,MAAM,MAA0B;AAC9C,SAAO,IAAI,WAAW,IAAI;AAC5B;ACnIA,IAAM,SAAS,OAAO,YAAY,eAChC,QAAQ,YAAY,QACpB,QAAQ,SAAS,QAAQ;AAO3B,IAAM,yBAAyB,OAAO;AAQ/B,SAAS,OAAO,MAA0B;AAC/C,MAAI,QAAQ;AAEV,WAAO,OAAO,KAAK,IAAI,EAAE,SAAS,QAAQ;EAC5C;AAGA,MAAI,KAAK,SAAS,wBAAwB;AACxC,WAAO,cAAc,IAAI;EAC3B;AAGA,MAAI,SAAS;AACb,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,cAAU,OAAO,aAAa,KAAK,CAAC,CAAE;EACxC;AACA,SAAO,KAAK,MAAM;AACpB;AAKO,SAAS,OAAO,QAA4B;AACjD,MAAI,QAAQ;AAEV,WAAO,IAAI,WAAW,OAAO,KAAK,QAAQ,QAAQ,CAAC;EACrD;AAGA,QAAM,SAAS,KAAK,MAAM;AAC1B,QAAM,SAAS,IAAI,WAAW,OAAO,MAAM;AAC3C,WAAS,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;AACtC,WAAO,CAAC,IAAI,OAAO,WAAW,CAAC;EACjC;AACA,SAAO;AACT;AA2CA,IAAM,oBAAoB,KAAK;AAexB,SAAS,cAAc,MAA0B;AACtD,MAAI,QAAQ;AAEV,WAAO,OAAO,KAAK,IAAI,EAAE,SAAS,QAAQ;EAC5C;AAGA,QAAM,SAAmB,CAAC;AAE1B,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK,mBAAmB;AACvD,UAAM,QAAQ,KAAK,SAAS,GAAG,KAAK,IAAI,IAAI,mBAAmB,KAAK,MAAM,CAAC;AAC3E,QAAI,SAAS;AACb,aAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;AACrC,gBAAU,OAAO,aAAa,MAAM,CAAC,CAAE;IACzC;AACA,WAAO,KAAK,MAAM;EACpB;AAEA,SAAO,KAAK,OAAO,KAAK,EAAE,CAAC;AAC7B;ACpIA,IAAM,uBAAuB,uBAAO,IAAI,sCAAsC;AAKvE,IAAM,eAAN,cAA2B,MAAM;EAItC,YAAY,SAAiC,MAAc;AACzD,UAAM,OAAO;AAD8B,SAAA,OAAA;AAE3C,SAAK,OAAO;AAEZ,QAAI,MAAM,mBAAmB;AAC3B,YAAM,kBAAkB,MAAM,KAAK,WAAW;IAChD;EACF;;EATA,CAAU,oBAAoB,IAAI;AAUpC;AAKO,IAAM,aAAN,cAAyB,aAAa;EAC3C,YAAY,SAAiC,QAAiB;AAC5D,UAAM,SAAS,aAAa;AADe,SAAA,SAAA;AAE3C,SAAK,OAAO;EACd;AACF;AAsCO,IAAM,iBAAN,cAA6B,aAAa;EAC/C,YACkB,YACA,SAChB,SACA;AACA,UAAM,YAAY,aAAa,OAAO,MAAM,QAAQ,CAAC;AACrD,UAAM,SAAS,UAAU,OAAO,MAAM,QAAQ,CAAC;AAC/C,UAAM,MAAM,UACR,GAAG,OAAO,UAAU,QAAQ,oBAAoB,KAAK,OACrD,QAAQ,QAAQ,oBAAoB,KAAK;AAC7C,UAAM,KAAK,qBAAqB;AAThB,SAAA,aAAA;AACA,SAAA,UAAA;AAShB,SAAK,OAAO;EACd;AACF;;;AKjFA,iBAAkB;ACAlB,IAAAA,cAAkB;ACAlB,IAAAA,cAAkB;AMAlB,IAAAA,cAAkB;ARSX,IAAM,gBAAgB,aAAE,OAAO,EAAE,SAAS;AAK1C,IAAM,aAAa,aAAE,OAAO,EAAE,KAAK;AAKnC,IAAM,aAAa,aAAE,KAAK,CAAC,MAAM,IAAI,CAAC;AAKtC,IAAM,qBAAqB,aAAE,KAAK;EACvC;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;AACF,CAAC;AAKM,IAAM,sBAAsB,aAAE,KAAK,CAAC,WAAW,aAAa,QAAQ,CAAC;AAKrE,IAAM,kBAAkB,aAAE,KAAK;EACpC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;AACF,CAAC;AAKM,IAAM,wBAAwB,aAAE,OAAO;EAC5C,MAAM;EACN,KAAK,aAAE,OAAO;EACd,MAAM,aAAE,OAAO;EACf,KAAK,aAAE,OAAO;AAChB,CAAC;AAKM,IAAM,uBAAuB,aAAE,OAAO;EAC3C,YAAY;EACZ,MAAM,aAAE,WAAW,UAAU;EAC7B,UAAU,aAAE,OAAO;AACrB,CAAC;AC9DM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;AAC7H,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;;EAErC,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACnC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC7B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACvC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;AAC7C,CAAC;AAKM,IAAM,oBAAoBA,YAAAA,EAAE,OAAO;EACxC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AC5DM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC3H,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,MAAMA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,UAAU,QAAQ,WAAW,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC/F,OAAOA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC3B,gBAAgBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACxD,aAAaA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,GAAG,EAAE,SAAS;;EACjD,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EAChC,OAAOA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EAC/C,iBAAiBA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,OAAO,KAAK,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS;AAChF,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AASM,IAAM,sBAAsBA,YAAAA,EAAE,OAAO;;EAE1C,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,SAASA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC9B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EACxC,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;EACpC,sBAAsBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;;EAEpD,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,QAAQA,YAAAA,EAAE,MAAM,qBAAqB,EAAE,SAAS;EAChD,UAAUA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC9B,4BAA4BA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAC1D,QAAQA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACrC,eAAeA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;EACvD,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;AAC7D,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;EACrC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AIxFD,IAAM,iBAAiB,CAAC,wBAAwB,iBAAiB,qBAAqB,QAAQ;AAoBvF,SAAS,WAAW,MAA4B;AACrD,SAAO,mBAAmB,IAAI,EAAE;AAClC;AAMO,SAAS,mBAAmB,MAAoC;AACrE,QAAM,SAA8B;IAClC,MAAM;IACN,YAAY;IACZ,YAAY,CAAC;IACb,UAAU,CAAC;EACb;AAEA,MAAI,CAAC,QAAQ,OAAO,SAAS,UAAU;AACrC,WAAO,WAAW,KAAK,wBAAwB;AAC/C,WAAO;EACT;AAEA,QAAM,MAAM;AACZ,QAAM,UAAW,IAAI,QAAQ,OAAO,IAAI,SAAS,WAAW,IAAI,OAAO;AAQvE,MAAI,IAAI,SAAS,iBAAiB;AAChC,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B;AAGtD,QAAI,IAAI,gBAAgB,IAAI,iBAAiB,OAAO;AAClD,aAAO,SAAS,KAAK,iBAAiB,IAAI,YAAY,6BAA6B;IACrF;AAEA,WAAO;EACT;AAGA,MAAI,IAAI,SAAS,iBAAiB;AAChC,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B;AAGtD,QAAI,SAAS;AACX,iBAAW,SAAS,gBAAgB;AAClC,YAAI,SAAS,SAAS;AACpB,iBAAO,SAAS,KAAK,kBAAkB,KAAK,oBAAoB;QAClE;MACF;IACF;AAEA,QAAI,IAAI,gBAAgB,IAAI,iBAAiB,OAAO;AAClD,aAAO,SAAS,KAAK,iBAAiB,IAAI,YAAY,6BAA6B;IACrF;AAEA,WAAO;EACT;AAGA,MAAI,OAAO,IAAI,iBAAiB,UAAU;AACxC,QAAI,IAAI,aAAa,WAAW,GAAG,GAAG;AACpC,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,iBAAiB,IAAI,YAAY,mBAAmB;AAC3E,aAAO;IACT;AACA,QAAI,IAAI,aAAa,WAAW,GAAG,GAAG;AACpC,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,iBAAiB,IAAI,YAAY,mBAAmB;AAC3E,aAAO;IACT;EACF;AAEA,MAAI,IAAI,iBAAiB,KAAO,IAAI,iBAAiB,GAAG;AACtD,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,2BAA2B,IAAI,YAAY,EAAE;AACpE,WAAO;EACT;AAGA,MAAI,SAAS;AACX,UAAM,WAAqB,CAAC;AAC5B,eAAW,SAAS,gBAAgB;AAClC,UAAI,SAAS,SAAS;AACpB,iBAAS,KAAK,KAAK;MACrB;IACF;AAEA,QAAI,SAAS,SAAS,GAAG;AACvB,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,uBAAuB,SAAS,KAAK,IAAI,CAAC,EAAE;AACnE,aAAO;IACT;EACF;AAGA,QAAM,eAAyB,CAAC;AAChC,aAAW,SAAS,gBAAgB;AAClC,QAAI,SAAS,KAAK;AAChB,mBAAa,KAAK,KAAK;IACzB;EACF;AACA,MAAI,aAAa,SAAS,GAAG;AAC3B,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B,aAAa,KAAK,IAAI,CAAC,EAAE;AAC/E,WAAO,SAAS,KAAK,sDAAsD;AAC3E,WAAO;EACT;AAGA,MAAI,IAAI,QAAQ,SAAS;AACvB,UAAM,WAAW,QAAQ;AACzB,QAAI,YAAY,OAAO,aAAa,UAAU;AAE5C,UAAI,OAAO,IAAI,SAAS,UAAU;AAChC,YAAI,IAAI,KAAK,SAAS,IAAI,KAAK,IAAI,KAAK,SAAS,GAAG,GAAG;AACrD,iBAAO,OAAO;AACd,iBAAO,aAAa;AACpB,iBAAO,WAAW,KAAK,eAAe,IAAI,IAAI,wBAAwB;AACtE,iBAAO;QACT;AACA,YAAI,IAAI,KAAK,SAAS,IAAI,KAAK,IAAI,KAAK,SAAS,GAAG,GAAG;AACrD,iBAAO,OAAO;AACd,iBAAO,aAAa;AACpB,iBAAO,WAAW,KAAK,eAAe,IAAI,IAAI,wBAAwB;AACtE,iBAAO;QACT;MACF;AAEA,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,4CAA4C;AACnE,aAAO;IACT;EACF;AAGA,MAAI,IAAI,QAAQ,OAAO,IAAI,SAAS,UAAU;AAC5C,QAAI,iBAAiB,OAAO,iBAAiB,OAAO,cAAc,KAAK;AACrE,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,8DAA8D;AACrF,aAAO;IACT;EACF;AAGA,MAAI,WAAW,OAAO,QAAQ,SAAS,UAAU;AAC/C,QAAI,iBAAiB,WAAW,iBAAiB,SAAS;AACxD,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,wDAAwD;AAC/E,aAAO,SAAS,KAAK,oBAAoB;AACzC,aAAO;IACT;EACF;AAEA,SAAO,WAAW,KAAK,4BAA4B;AACnD,SAAO;AACT;;;AhB3KO,IAAM,gBAAgB,IAAI,WAAW,CAAC,KAAM,IAAM,IAAM,IAAM,IAAM,IAAM,IAAM,EAAI,CAAC;AAKrF,IAAM,iBAAiB,KAAK,OAAO;AAMnC,IAAM,oBAAoB,KAAK,OAAO;AAW7C,SAAS,qBAAqB,YAAwB,UAAkB,mBAA+B;AACrG,QAAM,SAAuB,CAAC;AAC9B,MAAI,YAAY;AAChB,MAAI,QAAsB;AAE1B,QAAM,WAAW,IAAI,sBAAQ,CAAC,MAAkB,WAAoB;AAClE,QAAI,MAAO;AAEX,QAAI,QAAQ,KAAK,SAAS,GAAG;AAC3B,mBAAa,KAAK;AAClB,UAAI,YAAY,SAAS;AACvB,gBAAQ,IAAI,eAAe,WAAW,SAAS,qBAAqB;AACpE;MACF;AACA,aAAO,KAAK,IAAI;IAClB;EACF,CAAC;AAGD,MAAI;AACF,aAAS,KAAK,sBAAsB,aAAa,aAAa,IAAI,WAAW,UAAU,GAAG,IAAI;EAChG,SAAS,GAAG;AACV,UAAM,IAAI,WAAW,yBAAyB,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC,CAAC,IAAI,KAAK;EACnG;AAEA,MAAI,OAAO;AACT,UAAM;EACR;AAEA,SAAO,OAAO,GAAG,MAAM;AACzB;AAKO,IAAM,kBAAkB;;EAE7B;EACA;;EAEA;EACA;EACA;;EAEA;EACA;EACA;EACA;AACF;AAsBO,SAAS,MAAM,MAA2B;AAC/C,MAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,WAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,QAAI,KAAK,CAAC,MAAM,cAAc,CAAC,EAAG,QAAO;EAC3C;AACA,SAAO;AACT;AAMO,SAAS,gBAAgB,MAA+B;AAC7D,QAAM,aAA0B,CAAC;AAGjC,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,WAAO;EACT;AAEA,MAAI,SAAS;AAEb,SAAO,SAAS,KAAK,QAAQ;AAE3B,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,SAAS,aAAa,MAAM,MAAM;AACxC,cAAU;AAGV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,YAAY,MAAM,MAAM,QAAQ,SAAS,CAAC;AAChD,UAAM,OAAO,SAAS,SAAS;AAC/B,cAAU;AAGV,QAAI,SAAS,gBAAgB;AAC3B,YAAM,IAAI,eAAe,QAAQ,gBAAgB,cAAc,IAAI,GAAG;IACxE;AAGA,QAAI,SAAS,SAAS,KAAK,OAAQ;AACnC,UAAM,YAAY,MAAM,MAAM,QAAQ,SAAS,MAAM;AACrD,cAAU;AAGV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,cAAU;AAGV,QAAI,SAAS,QAAQ;AACnB,YAAM,YAAY,QAAQ,WAAW,IAAI,WAAW,CAAC,CAAC,CAAC,CAAC;AACxD,UAAI,cAAc,IAAI;AACpB,cAAM,UAAU,SAAS,MAAM,WAAW,GAAG,SAAS,CAAC;AACvD,cAAM,OAAO,SAAS,MAAM,WAAW,YAAY,CAAC,CAAC;AACrD,mBAAW,KAAK,EAAE,SAAS,KAAK,CAAC;MACnC;IACF;AAGA,QAAI,SAAS,QAAQ;AACnB,YAAM,YAAY,QAAQ,WAAW,IAAI,WAAW,CAAC,CAAC,CAAC,CAAC;AACxD,UAAI,cAAc,IAAI;AACpB,cAAM,UAAU,SAAS,MAAM,WAAW,GAAG,SAAS,CAAC;AACvD,cAAM,oBAAoB,UAAU,YAAY,CAAC;AAEjD,YAAI,sBAAsB,GAAG;AAC3B,cAAI;AACF,kBAAM,iBAAiB,MAAM,WAAW,YAAY,CAAC;AAErD,kBAAM,eAAe,qBAAqB,gBAAgB,iBAAiB;AAC3E,kBAAM,OAAO,SAAS,YAAY;AAClC,uBAAW,KAAK,EAAE,SAAS,KAAK,CAAC;UACnC,SAAS,KAAK;AAEZ,gBAAI,eAAe,gBAAgB;AACjC,oBAAM;YACR;UAEF;QACF;MACF;IACF;AAGA,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAO;AACT;AAKO,SAAS,WAAW,MAA2E;AACpG,QAAM,SAAkE,CAAC;AAEzE,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,WAAO;EACT;AAEA,MAAI,SAAS;AAEb,SAAO,SAAS,KAAK,QAAQ;AAC3B,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,SAAS,aAAa,MAAM,MAAM;AACxC,UAAM,aAAa;AACnB,cAAU;AAEV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,OAAO,SAAS,MAAM,MAAM,QAAQ,SAAS,CAAC,CAAC;AACrD,cAAU;AAEV,WAAO,KAAK,EAAE,MAAM,QAAQ,YAAY,OAAO,CAAC;AAGhD,cAAU,SAAS;AAEnB,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAO;AACT;AAKA,SAAS,cAAc,WAA4B;AAEjD,MAAI;AACF,WAAO,KAAK,MAAM,SAAS;EAC7B,QAAQ;AAEN,QAAI;AACF,YAAM,UAAU,SAAS,OAAa,SAAS,CAAC;AAChD,aAAO,KAAK,MAAM,OAAO;IAC3B,QAAQ;AACN,YAAM,IAAI,WAAW,yCAAyC,KAAK;IACrE;EACF;AACF;AAKA,SAAS,kBAAkB,MAAwB;AACjD,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,QAAM,MAAM;AAGZ,QAAM,OAAO,IAAI;AACjB,MAAI,MAAM,gBAAgB;AACxB,UAAM,OAAO,KAAK;AAClB,QAAI,MAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,QAAQ,SAAS,EAAG,QAAO;EACrE;AAGA,MAAI,IAAI,gBAAgB;AACtB,UAAM,OAAO,IAAI;AACjB,QAAI,MAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,QAAQ,SAAS,EAAG,QAAO;EACrE;AAEA,SAAO;AACT;AAMO,SAAS,eAAe,MAAuC;AAEpE,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,UAAM,IAAI,WAAW,yBAAyB,KAAK;EACrD;AAGA,QAAM,aAAa,gBAAgB,IAAI;AAEvC,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAM,IAAI,WAAW,+BAA+B,KAAK;EAC3D;AAGA,MAAI,iBAA6C;AAEjD,aAAW,OAAO,iBAAiB;AAEjC,UAAM,iBAAiB,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;AAE/D,eAAW,SAAS,gBAAgB;AAClC,UAAI;AACF,cAAM,OAAO,cAAc,MAAM,IAAI;AACrC,cAAM,OAAO,WAAW,IAAI;AAE5B,YAAI,SAAS,QAAQ,SAAS,MAAM;AAClC,gBAAM,SAA8B;YAClC,MAAM;YACN;YACA,aAAa,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;UACvD;AAGA,cAAI,kBAAkB,IAAI,GAAG;AAC3B,mBAAO;UACT;AAGA,cAAI,CAAC,gBAAgB;AACnB,6BAAiB;UACnB;QACF;AAGA,YAAI,CAAC,QAAQ,QAAQ,OAAO,SAAS,UAAU;AAC7C,gBAAM,MAAM;AACZ,cAAI,iBAAmE;AAEvE,cAAI,IAAI,SAAS,mBAAmB,IAAI,QAAS,IAAI,KAAiC,MAAM;AAC1F,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD,WAAW,IAAI,SAAS,mBAAmB,IAAI,QAAS,IAAI,KAAiC,MAAM;AACjG,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD,WAAW,IAAI,SAAS,IAAI,eAAe,IAAI,eAAe,IAAI,WAAW;AAC3E,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD;AAEA,cAAI,gBAAgB;AAClB,kBAAM,aAAkC;cACtC,GAAG;cACH,aAAa,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;YACvD;AACA,gBAAI,kBAAkB,IAAI,GAAG;AAC3B,qBAAO;YACT;AACA,gBAAI,CAAC,eAAgB,kBAAiB;UACxC;QACF;MACF,QAAQ;MAER;IACF;EACF;AAGA,MAAI,gBAAgB;AAClB,WAAO;EACT;AAEA,QAAM,IAAI,WAAW,6CAA6C,KAAK;AACzE;AExWA,IAAI,YAAgC;AAKpC,SAAS,cAA2B;AAClC,MAAI,UAAW,QAAO;AAEtB,cAAY,IAAI,YAAY,GAAG;AAC/B,WAAS,IAAI,GAAG,IAAI,KAAK,KAAK;AAC5B,QAAI,IAAI;AACR,aAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,UAAK,IAAI,IAAM,aAAc,MAAM,IAAO,MAAM;IAClD;AACA,cAAU,CAAC,IAAI;EACjB;AACA,SAAO;AACT;AAOO,SAAS,MAAM,MAA0B;AAC9C,QAAM,QAAQ,YAAY;AAC1B,MAAI,MAAM;AAEV,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,UAAM,OAAO,MAAM,KAAK,CAAC,KAAM,GAAI,IAAM,QAAQ;EACnD;AAEA,UAAQ,MAAM,gBAAgB;AAChC;AAOO,SAAS,WAAW,MAA8B;AACvD,QAAM,MAAM,MAAM,IAAI;AACtB,QAAM,SAAS,MAAM,CAAC;AACtB,gBAAc,QAAQ,KAAK,CAAC;AAC5B,SAAO;AACT;ADNO,SAAS,oBAAoB,WAAmC;AACrE,MAAI,CAAC,MAAM,SAAS,GAAG;AACrB,UAAM,IAAIC,WAAW,yBAAyB,KAAK;EACrD;AAEA,QAAM,SAAuB,CAACC,MAAM,WAAW,GAAG,CAAC,CAAC;AACpD,MAAI,SAAS;AAEb,SAAO,SAAS,UAAU,QAAQ;AAEhC,QAAI,SAAS,IAAI,UAAU,OAAQ;AACnC,UAAM,SAASC,aAAa,WAAW,MAAM;AAC7C,UAAM,YAAYD,MAAM,WAAW,QAAQ,SAAS,CAAC;AACrD,cAAU;AAGV,QAAI,SAAS,IAAI,UAAU,OAAQ;AACnC,UAAM,OAAOE,SAASF,MAAM,WAAW,QAAQ,SAAS,CAAC,CAAC;AAC1D,UAAM,UAAUA,MAAM,WAAW,QAAQ,SAAS,CAAC;AACnD,cAAU;AAGV,QAAI,SAAS,SAAS,IAAI,UAAU,OAAQ;AAC5C,UAAM,UAAUA,MAAM,WAAW,QAAQ,SAAS,MAAM;AACxD,UAAM,SAASA,MAAM,WAAW,SAAS,QAAQ,SAAS,SAAS,CAAC;AACpE,cAAU,SAAS;AAGnB,QAAI,SAAS,UAAU,SAAS,QAAQ;AACtC;IACF;AAGA,WAAO,KAAK,WAAW,SAAS,SAAS,MAAM;AAG/C,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAOG,OAAO,GAAG,MAAM;AACzB;AAKA,SAAS,eAAe,WAA+B;AAGrD,WAAS,IAAI,UAAU,SAAS,IAAI,KAAK,GAAG,KAAK;AAC/C,QACE,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM,IACrB;AACA,aAAO;IACT;EACF;AACA,SAAO;AACT;AAKA,SAAS,gBAAgB,SAAiB,MAA0B;AAClE,QAAM,gBAAgB,WAAW,OAAO;AACxC,QAAM,aAAa,WAAW,IAAI;AAGlC,QAAM,YAAYA;IAChB;IACA,IAAI,WAAW,CAAC,CAAC,CAAC;;IAClB;EACF;AAGA,QAAM,YAAY,WAAW,MAAM;AAGnC,QAAM,MAAM,WAAWA,OAAO,WAAW,SAAS,CAAC;AAGnD,QAAM,eAAeC,MAAM,CAAC;AAC5BC,gBAAc,cAAc,UAAU,QAAQ,CAAC;AAG/C,SAAOF,OAAO,cAAc,WAAW,WAAW,GAAG;AACvD;AAKO,SAAS,gBAAgB,WAAuB,SAAiB,MAA0B;AAChG,QAAM,aAAa,eAAe,SAAS;AAE3C,MAAI,eAAe,IAAI;AACrB,UAAM,IAAIJ,WAAW,qCAAqC,KAAK;EACjE;AAEA,QAAM,YAAY,gBAAgB,SAAS,IAAI;AAG/C,QAAM,aAAaC,MAAM,WAAW,GAAG,UAAU;AACjD,QAAM,eAAeA,MAAM,WAAW,UAAU;AAEhD,SAAOG,OAAO,YAAY,WAAW,YAAY;AACnD;AAKO,SAAS,aACd,aACA,UACA,UAAwB,CAAC,GACb;AACZ,QAAM;IACJ,MAAM;IACN,SAAS;IACT,SAAS;EACX,IAAI;AAGJ,QAAM,WAAW,oBAAoB,WAAW;AAGhD,QAAM,OAAO,SACT,KAAK,UAAU,QAAQ,IACvB,KAAK,UAAU,UAAU,MAAM,CAAC;AAGpC,QAAM,OAAO,SAAS,OAAa,WAAW,IAAI,CAAC,IAAI;AAGvD,SAAO,gBAAgB,UAAU,KAAK,IAAI;AAC5C;AAKO,SAAS,gBACd,QACA,QACwC;AACxC,QAAM,SAAS,OAAO,UAAU,OAAO;AACvC,QAAM,WAAqB,CAAC;AAE5B,MAAI,SAAS,OAAO,KAAK;AACvB,WAAO;MACL,OAAO;MACP,UAAU,CAAC,aAAa,OAAO,QAAQ,CAAC,CAAC,wBAAwB,OAAO,GAAG,KAAK;IAClF;EACF;AAEA,MAAI,SAAS,OAAO,MAAM;AACxB,aAAS,KAAK,aAAa,OAAO,QAAQ,CAAC,CAAC,+BAA+B,OAAO,IAAI,KAAK;EAC7F;AAEA,SAAO,EAAE,OAAO,MAAM,SAAS;AACjC;","names":["import_zod","z","z","ParseError","slice","readUInt32BE","toLatin1","concat","alloc","writeUInt32BE"]}
|
|
1
|
+
{"version":3,"sources":["../src/png.ts","../../png/src/parser.ts","../../png/src/builder.ts","../../png/src/crc32.ts","../../core/src/binary.ts","../../core/src/base64.ts","../../core/src/errors.ts","../../core/src/data-url.ts","../../core/src/uri.ts","../../core/src/image.ts","../../core/src/uuid.ts","../../schemas/src/common.ts","../../schemas/src/ccv2.ts","../../schemas/src/ccv3.ts","../../schemas/src/risu.ts","../../schemas/src/normalized.ts","../../schemas/src/feature-deriver.ts","../../schemas/src/detection.ts","../../schemas/src/normalizer.ts","../../schemas/src/validation.ts"],"sourcesContent":["export * from '@character-foundry/png';\n","/**\n * PNG Parser\n *\n * Extracts text chunks and character card data from PNG files.\n * Uses fflate for zTXt decompression, works in browser and Node.js.\n */\n\nimport { Inflate } from 'fflate';\nimport {\n type BinaryData,\n readUInt32BE,\n slice,\n toString,\n toLatin1,\n indexOf,\n concat,\n base64Decode,\n ParseError,\n SizeLimitError,\n} from '@character-foundry/core';\nimport {\n type CCv2Data,\n type CCv3Data,\n type Spec,\n detectSpec,\n} from '@character-foundry/schemas';\n\n/**\n * PNG signature bytes\n */\nexport const PNG_SIGNATURE = new Uint8Array([0x89, 0x50, 0x4e, 0x47, 0x0d, 0x0a, 0x1a, 0x0a]);\n\n/**\n * Maximum size for a single PNG chunk (50MB per Risu CharX spec)\n */\nexport const MAX_CHUNK_SIZE = 50 * 1024 * 1024;\n\n/**\n * Maximum inflated size for zTXt chunks (16MB - reasonable for card JSON)\n * This prevents inflation attacks where small compressed data expands to huge sizes\n */\nexport const MAX_INFLATED_SIZE = 16 * 1024 * 1024;\n\n/**\n * Synchronous inflate with size limit to prevent inflation attacks.\n * Uses streaming decompression to enforce the limit without fully decompressing.\n *\n * @param compressed - Compressed data to inflate\n * @param maxSize - Maximum allowed inflated size (defaults to MAX_INFLATED_SIZE)\n * @returns Inflated data as Uint8Array\n * @throws SizeLimitError if inflated size exceeds maxSize\n */\nfunction inflateSyncWithLimit(compressed: BinaryData, maxSize: number = MAX_INFLATED_SIZE): Uint8Array {\n const chunks: Uint8Array[] = [];\n let totalSize = 0;\n let error: Error | null = null;\n\n const inflater = new Inflate((data: Uint8Array, _final: boolean) => {\n if (error) return; // Stop processing if we've hit a limit\n\n if (data && data.length > 0) {\n totalSize += data.length;\n if (totalSize > maxSize) {\n error = new SizeLimitError(totalSize, maxSize, 'inflated zTXt chunk');\n return;\n }\n chunks.push(data);\n }\n });\n\n // Push all data at once with final=true for sync operation\n try {\n inflater.push(compressed instanceof Uint8Array ? compressed : new Uint8Array(compressed), true);\n } catch (e) {\n throw new ParseError(`Decompression failed: ${e instanceof Error ? e.message : String(e)}`, 'png');\n }\n\n if (error) {\n throw error;\n }\n\n return concat(...chunks);\n}\n\n/**\n * Text chunk keys used for character cards by various frontends\n */\nexport const TEXT_CHUNK_KEYS = [\n // v3 keys\n 'ccv3',\n 'chara_card_v3',\n // v2 keys (most common)\n 'chara',\n 'ccv2',\n 'character',\n // Alternative/legacy keys\n 'charactercard',\n 'card',\n 'CharacterCard',\n 'Chara',\n];\n\n/**\n * Parsed text chunk from PNG\n */\nexport interface TextChunk {\n keyword: string;\n text: string;\n}\n\n/**\n * Result of extracting card data from PNG\n */\nexport interface PNGExtractionResult {\n data: CCv2Data | CCv3Data;\n spec: Spec;\n extraChunks: TextChunk[];\n}\n\n/**\n * Check if data is a valid PNG (has correct signature)\n */\nexport function isPNG(data: BinaryData): boolean {\n if (data.length < 8) return false;\n for (let i = 0; i < 8; i++) {\n if (data[i] !== PNG_SIGNATURE[i]) return false;\n }\n return true;\n}\n\n/**\n * Parse all text chunks (tEXt and zTXt) from PNG buffer\n * Returns array of {keyword, text} to support multiple chunks\n */\nexport function parseTextChunks(data: BinaryData): TextChunk[] {\n const textChunks: TextChunk[] = [];\n\n // Verify PNG signature\n if (!isPNG(data)) {\n return textChunks;\n }\n\n let offset = 8; // Skip PNG signature\n\n while (offset < data.length) {\n // Read chunk length (4 bytes, big-endian)\n if (offset + 4 > data.length) break;\n const length = readUInt32BE(data, offset);\n offset += 4;\n\n // Read chunk type (4 bytes ASCII)\n if (offset + 4 > data.length) break;\n const typeBytes = slice(data, offset, offset + 4);\n const type = toLatin1(typeBytes);\n offset += 4;\n\n // Check chunk size limit before reading\n if (length > MAX_CHUNK_SIZE) {\n throw new SizeLimitError(length, MAX_CHUNK_SIZE, `PNG chunk '${type}'`);\n }\n\n // Read chunk data\n if (offset + length > data.length) break;\n const chunkData = slice(data, offset, offset + length);\n offset += length;\n\n // Skip CRC (4 bytes)\n if (offset + 4 > data.length) break;\n offset += 4;\n\n // Parse tEXt chunks\n if (type === 'tEXt') {\n const nullIndex = indexOf(chunkData, new Uint8Array([0]));\n if (nullIndex !== -1) {\n const keyword = toLatin1(slice(chunkData, 0, nullIndex));\n const text = toString(slice(chunkData, nullIndex + 1));\n textChunks.push({ keyword, text });\n }\n }\n\n // Parse zTXt chunks (compressed) with size limit protection\n if (type === 'zTXt') {\n const nullIndex = indexOf(chunkData, new Uint8Array([0]));\n if (nullIndex !== -1) {\n const keyword = toLatin1(slice(chunkData, 0, nullIndex));\n const compressionMethod = chunkData[nullIndex + 1];\n\n if (compressionMethod === 0) { // 0 = deflate/inflate\n try {\n const compressedData = slice(chunkData, nullIndex + 2);\n // Use size-limited inflate to prevent inflation attacks\n const decompressed = inflateSyncWithLimit(compressedData, MAX_INFLATED_SIZE);\n const text = toString(decompressed);\n textChunks.push({ keyword, text });\n } catch (err) {\n // Re-throw size limit errors, skip other decompression failures\n if (err instanceof SizeLimitError) {\n throw err;\n }\n // Failed to decompress zTXt chunk, skip it\n }\n }\n }\n }\n\n // Stop after IEND chunk\n if (type === 'IEND') break;\n }\n\n return textChunks;\n}\n\n/**\n * List all chunks in a PNG file\n */\nexport function listChunks(data: BinaryData): Array<{ type: string; offset: number; length: number }> {\n const chunks: Array<{ type: string; offset: number; length: number }> = [];\n\n if (!isPNG(data)) {\n return chunks;\n }\n\n let offset = 8;\n\n while (offset < data.length) {\n if (offset + 4 > data.length) break;\n const length = readUInt32BE(data, offset);\n const chunkStart = offset;\n offset += 4;\n\n if (offset + 4 > data.length) break;\n const type = toLatin1(slice(data, offset, offset + 4));\n offset += 4;\n\n chunks.push({ type, offset: chunkStart, length });\n\n // Skip data and CRC\n offset += length + 4;\n\n if (type === 'IEND') break;\n }\n\n return chunks;\n}\n\n/**\n * Try to parse JSON from chunk data (supports plain and base64)\n */\nfunction tryParseChunk(chunkData: string): unknown {\n // Try direct JSON parse first\n try {\n return JSON.parse(chunkData);\n } catch {\n // Try base64 decode then JSON parse\n try {\n const decoded = toString(base64Decode(chunkData));\n return JSON.parse(decoded);\n } catch {\n throw new ParseError('Not valid JSON or base64-encoded JSON', 'png');\n }\n }\n}\n\n/**\n * Check if parsed data has a lorebook\n */\nfunction hasLorebookInData(json: unknown): boolean {\n if (!json || typeof json !== 'object') return false;\n const obj = json as Record<string, unknown>;\n\n // Check wrapped format\n const data = obj.data as Record<string, unknown> | undefined;\n if (data?.character_book) {\n const book = data.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Extract character card JSON from PNG tEXt chunks\n * Returns card data, spec version, and any extra chunks\n */\nexport function extractFromPNG(data: BinaryData): PNGExtractionResult {\n // Validate PNG format\n if (!isPNG(data)) {\n throw new ParseError('Invalid PNG signature', 'png');\n }\n\n // Parse all text chunks\n const textChunks = parseTextChunks(data);\n\n if (textChunks.length === 0) {\n throw new ParseError('No text chunks found in PNG', 'png');\n }\n\n // Try all known keys, preferring chunks with lorebooks\n let fallbackResult: PNGExtractionResult | null = null;\n\n for (const key of TEXT_CHUNK_KEYS) {\n // Find all chunks with this keyword\n const matchingChunks = textChunks.filter(c => c.keyword === key);\n\n for (const chunk of matchingChunks) {\n try {\n const json = tryParseChunk(chunk.text);\n const spec = detectSpec(json);\n\n if (spec === 'v3' || spec === 'v2') {\n const result: PNGExtractionResult = {\n data: json as CCv2Data | CCv3Data,\n spec,\n extraChunks: textChunks.filter(c => c.keyword !== key),\n };\n\n // Prefer chunks with lorebooks\n if (hasLorebookInData(json)) {\n return result;\n }\n\n // Store as fallback if we don't find one with lorebook\n if (!fallbackResult) {\n fallbackResult = result;\n }\n }\n\n // If detectSpec failed but we have JSON that looks like a card, try to infer\n if (!spec && json && typeof json === 'object') {\n const obj = json as Record<string, unknown>;\n let inferredResult: { data: CCv2Data | CCv3Data; spec: Spec } | null = null;\n\n if (obj.spec === 'chara_card_v3' && obj.data && (obj.data as Record<string, unknown>).name) {\n inferredResult = { data: json as CCv3Data, spec: 'v3' };\n } else if (obj.spec === 'chara_card_v2' && obj.data && (obj.data as Record<string, unknown>).name) {\n inferredResult = { data: json as CCv2Data, spec: 'v2' };\n } else if (obj.name && (obj.description || obj.personality || obj.scenario)) {\n inferredResult = { data: json as CCv2Data, spec: 'v2' };\n }\n\n if (inferredResult) {\n const fullResult: PNGExtractionResult = {\n ...inferredResult,\n extraChunks: textChunks.filter(c => c.keyword !== key),\n };\n if (hasLorebookInData(json)) {\n return fullResult;\n }\n if (!fallbackResult) fallbackResult = fullResult;\n }\n }\n } catch {\n // Continue to next chunk\n }\n }\n }\n\n // If we found a valid card but no lorebook, use the fallback\n if (fallbackResult) {\n return fallbackResult;\n }\n\n throw new ParseError('No valid character card data found in PNG', 'png');\n}\n","/**\n * PNG Builder\n *\n * Embeds character card data into PNG tEXt chunks.\n * Works in browser and Node.js.\n */\n\nimport {\n type BinaryData,\n readUInt32BE,\n writeUInt32BE,\n slice,\n concat,\n fromLatin1,\n fromString,\n alloc,\n base64Encode,\n toLatin1,\n ParseError,\n} from '@character-foundry/core';\nimport type { CCv2Data, CCv3Data } from '@character-foundry/schemas';\nimport { crc32Bytes } from './crc32.js';\nimport { isPNG } from './parser.js';\n\n/**\n * Options for embedding card data into PNG\n */\nexport interface EmbedOptions {\n /**\n * Key to use for the tEXt chunk (default: 'chara')\n */\n key?: string;\n\n /**\n * Whether to base64 encode the JSON (default: true, recommended for compatibility)\n */\n base64?: boolean;\n\n /**\n * Whether to minify JSON (default: true)\n */\n minify?: boolean;\n}\n\n/**\n * Remove all tEXt and zTXt chunks from PNG buffer\n * CRITICAL: Must remove old chunks before adding new ones to prevent duplicate data\n */\nexport function removeAllTextChunks(pngBuffer: BinaryData): BinaryData {\n if (!isPNG(pngBuffer)) {\n throw new ParseError('Invalid PNG signature', 'png');\n }\n\n const chunks: BinaryData[] = [slice(pngBuffer, 0, 8)]; // Start with PNG signature\n let offset = 8;\n\n while (offset < pngBuffer.length) {\n // Read chunk length (4 bytes, big-endian)\n if (offset + 4 > pngBuffer.length) break;\n const length = readUInt32BE(pngBuffer, offset);\n const lengthBuf = slice(pngBuffer, offset, offset + 4);\n offset += 4;\n\n // Read chunk type (4 bytes ASCII)\n if (offset + 4 > pngBuffer.length) break;\n const type = toLatin1(slice(pngBuffer, offset, offset + 4));\n const typeBuf = slice(pngBuffer, offset, offset + 4);\n offset += 4;\n\n // Read chunk data + CRC\n if (offset + length + 4 > pngBuffer.length) break;\n const dataBuf = slice(pngBuffer, offset, offset + length);\n const crcBuf = slice(pngBuffer, offset + length, offset + length + 4);\n offset += length + 4;\n\n // Skip tEXt and zTXt chunks (don't add them to output)\n if (type === 'tEXt' || type === 'zTXt') {\n continue;\n }\n\n // Keep all other chunks\n chunks.push(lengthBuf, typeBuf, dataBuf, crcBuf);\n\n // Stop after IEND\n if (type === 'IEND') break;\n }\n\n return concat(...chunks);\n}\n\n/**\n * Find the offset of the IEND chunk in PNG buffer\n */\nfunction findIendOffset(pngBuffer: BinaryData): number {\n // Search backwards from the end - IEND should be near the end\n // IEND format: length(4) + \"IEND\"(4) + CRC(4) = 12 bytes\n for (let i = pngBuffer.length - 12; i >= 8; i--) {\n if (\n pngBuffer[i + 4] === 0x49 && // 'I'\n pngBuffer[i + 5] === 0x45 && // 'E'\n pngBuffer[i + 6] === 0x4e && // 'N'\n pngBuffer[i + 7] === 0x44 // 'D'\n ) {\n return i; // Start of length field\n }\n }\n return -1;\n}\n\n/**\n * Create a tEXt chunk\n */\nfunction createTextChunk(keyword: string, text: string): BinaryData {\n const keywordBuffer = fromLatin1(keyword);\n const textBuffer = fromString(text);\n\n // Chunk data: keyword + null separator + text\n const chunkData = concat(\n keywordBuffer,\n new Uint8Array([0]), // null separator\n textBuffer\n );\n\n // Chunk type\n const chunkType = fromLatin1('tEXt');\n\n // Calculate CRC32 of type + data\n const crc = crc32Bytes(concat(chunkType, chunkData));\n\n // Build length buffer\n const lengthBuffer = alloc(4);\n writeUInt32BE(lengthBuffer, chunkData.length, 0);\n\n // Assemble: length + type + data + crc\n return concat(lengthBuffer, chunkType, chunkData, crc);\n}\n\n/**\n * Inject a tEXt chunk into PNG buffer before IEND\n */\nexport function injectTextChunk(pngBuffer: BinaryData, keyword: string, text: string): BinaryData {\n const iendOffset = findIendOffset(pngBuffer);\n\n if (iendOffset === -1) {\n throw new ParseError('Invalid PNG: IEND chunk not found', 'png');\n }\n\n const textChunk = createTextChunk(keyword, text);\n\n // Insert before IEND\n const beforeIend = slice(pngBuffer, 0, iendOffset);\n const iendAndAfter = slice(pngBuffer, iendOffset);\n\n return concat(beforeIend, textChunk, iendAndAfter);\n}\n\n/**\n * Embed character card JSON into PNG tEXt chunk\n */\nexport function embedIntoPNG(\n imageBuffer: BinaryData,\n cardData: CCv2Data | CCv3Data,\n options: EmbedOptions = {}\n): BinaryData {\n const {\n key = 'chara',\n base64 = true,\n minify = true,\n } = options;\n\n // Remove all existing tEXt chunks first to prevent duplicate/stale data\n const cleanPng = removeAllTextChunks(imageBuffer);\n\n // Serialize JSON\n const json = minify\n ? JSON.stringify(cardData)\n : JSON.stringify(cardData, null, 2);\n\n // Optionally base64 encode (recommended for compatibility)\n const text = base64 ? base64Encode(fromString(json)) : json;\n\n // Inject the text chunk\n return injectTextChunk(cleanPng, key, text);\n}\n\n/**\n * Validate PNG size against limits\n */\nexport function validatePNGSize(\n buffer: BinaryData,\n limits: { max: number; warn: number }\n): { valid: boolean; warnings: string[] } {\n const sizeMB = buffer.length / (1024 * 1024);\n const warnings: string[] = [];\n\n if (sizeMB > limits.max) {\n return {\n valid: false,\n warnings: [`PNG size (${sizeMB.toFixed(2)}MB) exceeds maximum (${limits.max}MB)`],\n };\n }\n\n if (sizeMB > limits.warn) {\n warnings.push(`PNG size (${sizeMB.toFixed(2)}MB) is large (recommended: <${limits.warn}MB)`);\n }\n\n return { valid: true, warnings };\n}\n","/**\n * CRC32 Implementation for PNG Chunks\n *\n * Pure JavaScript implementation using the standard polynomial.\n */\n\nimport { type BinaryData, alloc, writeUInt32BE } from '@character-foundry/core';\n\n// Pre-computed CRC table for faster calculation\nlet CRC_TABLE: Uint32Array | null = null;\n\n/**\n * Build the CRC lookup table (lazy initialization)\n */\nfunction getCRCTable(): Uint32Array {\n if (CRC_TABLE) return CRC_TABLE;\n\n CRC_TABLE = new Uint32Array(256);\n for (let i = 0; i < 256; i++) {\n let c = i;\n for (let j = 0; j < 8; j++) {\n c = (c & 1) ? (0xEDB88320 ^ (c >>> 1)) : (c >>> 1);\n }\n CRC_TABLE[i] = c;\n }\n return CRC_TABLE;\n}\n\n/**\n * Calculate CRC32 checksum for a PNG chunk\n * @param data - Binary data to calculate CRC for\n * @returns CRC32 value as a number\n */\nexport function crc32(data: BinaryData): number {\n const table = getCRCTable();\n let crc = 0xFFFFFFFF;\n\n for (let i = 0; i < data.length; i++) {\n crc = table[(crc ^ data[i]!) & 0xFF]! ^ (crc >>> 8);\n }\n\n return (crc ^ 0xFFFFFFFF) >>> 0;\n}\n\n/**\n * Calculate CRC32 and return as a 4-byte big-endian Uint8Array\n * @param data - Binary data to calculate CRC for\n * @returns 4-byte Uint8Array containing CRC32 in big-endian format\n */\nexport function crc32Bytes(data: BinaryData): BinaryData {\n const crc = crc32(data);\n const result = alloc(4);\n writeUInt32BE(result, crc, 0);\n return result;\n}\n","/**\n * Binary Data Utilities\n *\n * Universal binary data operations using Uint8Array.\n * Works in both Node.js and browser environments.\n */\n\n/**\n * Universal binary data type (works in both environments)\n */\nexport type BinaryData = Uint8Array;\n\n/**\n * Read a 32-bit big-endian unsigned integer\n */\nexport function readUInt32BE(data: BinaryData, offset: number): number {\n return (\n (data[offset]! << 24) |\n (data[offset + 1]! << 16) |\n (data[offset + 2]! << 8) |\n data[offset + 3]!\n ) >>> 0;\n}\n\n/**\n * Write a 32-bit big-endian unsigned integer\n */\nexport function writeUInt32BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 24) & 0xff;\n data[offset + 1] = (value >>> 16) & 0xff;\n data[offset + 2] = (value >>> 8) & 0xff;\n data[offset + 3] = value & 0xff;\n}\n\n/**\n * Read a 16-bit big-endian unsigned integer\n */\nexport function readUInt16BE(data: BinaryData, offset: number): number {\n return ((data[offset]! << 8) | data[offset + 1]!) >>> 0;\n}\n\n/**\n * Write a 16-bit big-endian unsigned integer\n */\nexport function writeUInt16BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 8) & 0xff;\n data[offset + 1] = value & 0xff;\n}\n\n/**\n * Find a byte sequence in binary data\n */\nexport function indexOf(data: BinaryData, search: BinaryData, fromIndex = 0): number {\n outer: for (let i = fromIndex; i <= data.length - search.length; i++) {\n for (let j = 0; j < search.length; j++) {\n if (data[i + j] !== search[j]) continue outer;\n }\n return i;\n }\n return -1;\n}\n\n/**\n * Concatenate multiple binary arrays\n */\nexport function concat(...arrays: BinaryData[]): BinaryData {\n const totalLength = arrays.reduce((sum, arr) => sum + arr.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const arr of arrays) {\n result.set(arr, offset);\n offset += arr.length;\n }\n return result;\n}\n\n/**\n * Slice binary data (returns a view, not a copy)\n */\nexport function slice(data: BinaryData, start: number, end?: number): BinaryData {\n return data.subarray(start, end);\n}\n\n/**\n * Copy a portion of binary data (returns a new array)\n */\nexport function copy(data: BinaryData, start: number, end?: number): BinaryData {\n return data.slice(start, end);\n}\n\n/**\n * Convert string to binary (UTF-8)\n */\nexport function fromString(str: string): BinaryData {\n return new TextEncoder().encode(str);\n}\n\n/**\n * Convert binary to string (UTF-8)\n */\nexport function toString(data: BinaryData): string {\n return new TextDecoder().decode(data);\n}\n\n/**\n * Convert string to binary (Latin1 - for PNG keywords and similar)\n */\nexport function fromLatin1(str: string): BinaryData {\n const result = new Uint8Array(str.length);\n for (let i = 0; i < str.length; i++) {\n result[i] = str.charCodeAt(i) & 0xff;\n }\n return result;\n}\n\n/**\n * Convert binary to string (Latin1)\n */\nexport function toLatin1(data: BinaryData): string {\n let result = '';\n for (let i = 0; i < data.length; i++) {\n result += String.fromCharCode(data[i]!);\n }\n return result;\n}\n\n/**\n * Compare two binary arrays for equality\n */\nexport function equals(a: BinaryData, b: BinaryData): boolean {\n if (a.length !== b.length) return false;\n for (let i = 0; i < a.length; i++) {\n if (a[i] !== b[i]) return false;\n }\n return true;\n}\n\n/**\n * Create a new Uint8Array filled with zeros\n */\nexport function alloc(size: number): BinaryData {\n return new Uint8Array(size);\n}\n\n/**\n * Create a Uint8Array from an array of numbers\n */\nexport function from(data: number[] | ArrayBuffer | BinaryData): BinaryData {\n if (data instanceof Uint8Array) {\n return data;\n }\n if (data instanceof ArrayBuffer) {\n return new Uint8Array(data);\n }\n return new Uint8Array(data);\n}\n\n/**\n * Check if value is a Uint8Array\n */\nexport function isBinaryData(value: unknown): value is BinaryData {\n return value instanceof Uint8Array;\n}\n\n/**\n * Convert Node.js Buffer to Uint8Array (no-op if already Uint8Array)\n * This provides compatibility when interfacing with Node.js code\n */\nexport function toUint8Array(data: BinaryData | Buffer): BinaryData {\n if (data instanceof Uint8Array) {\n // Buffer extends Uint8Array, but we want a plain Uint8Array\n // This ensures we get a proper Uint8Array in all cases\n if (Object.getPrototypeOf(data).constructor.name === 'Buffer') {\n return new Uint8Array(data.buffer, data.byteOffset, data.byteLength);\n }\n return data;\n }\n return new Uint8Array(data);\n}\n\n/**\n * Convert binary data to hex string\n */\nexport function toHex(data: BinaryData): string {\n return Array.from(data)\n .map(b => b.toString(16).padStart(2, '0'))\n .join('');\n}\n\n/**\n * Convert hex string to binary data\n */\nexport function fromHex(hex: string): BinaryData {\n const bytes = new Uint8Array(hex.length / 2);\n for (let i = 0; i < bytes.length; i++) {\n bytes[i] = parseInt(hex.substr(i * 2, 2), 16);\n }\n return bytes;\n}\n","/**\n * Universal Base64 Encoding/Decoding\n *\n * Works in both Node.js and browser environments.\n */\n\nimport type { BinaryData } from './binary.js';\n\n/**\n * Check if we're in a Node.js environment\n */\nconst isNode = typeof process !== 'undefined' &&\n process.versions != null &&\n process.versions.node != null;\n\n/**\n * Threshold for switching to chunked encoding in browsers (1MB)\n * Below this, simple string concatenation is fast enough.\n * Above this, quadratic string growth becomes a problem.\n */\nconst LARGE_BUFFER_THRESHOLD = 1024 * 1024;\n\n/**\n * Encode binary data to base64 string\n *\n * PERFORMANCE: For large buffers (>1MB) in browsers, this automatically\n * uses the chunked implementation to avoid quadratic string concatenation.\n */\nexport function encode(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: use chunked encoding for large buffers to avoid O(n²) string growth\n if (data.length > LARGE_BUFFER_THRESHOLD) {\n return encodeChunked(data);\n }\n\n // Small buffers: simple approach is fast enough\n let binary = '';\n for (let i = 0; i < data.length; i++) {\n binary += String.fromCharCode(data[i]!);\n }\n return btoa(binary);\n}\n\n/**\n * Decode base64 string to binary data\n */\nexport function decode(base64: string): BinaryData {\n if (isNode) {\n // Node.js: use Buffer\n return new Uint8Array(Buffer.from(base64, 'base64'));\n }\n\n // Browser: use atob\n const binary = atob(base64);\n const result = new Uint8Array(binary.length);\n for (let i = 0; i < binary.length; i++) {\n result[i] = binary.charCodeAt(i);\n }\n return result;\n}\n\n/**\n * Check if a string is valid base64\n */\nexport function isBase64(str: string): boolean {\n if (str.length === 0) return false;\n // Base64 regex: only valid base64 characters, length multiple of 4 (with padding)\n const base64Regex = /^[A-Za-z0-9+/]*={0,2}$/;\n return base64Regex.test(str) && str.length % 4 === 0;\n}\n\n/**\n * Encode binary data to URL-safe base64 string\n * Replaces + with -, / with _, and removes padding\n */\nexport function encodeUrlSafe(data: BinaryData): string {\n return encode(data)\n .replace(/\\+/g, '-')\n .replace(/\\//g, '_')\n .replace(/=+$/, '');\n}\n\n/**\n * Decode URL-safe base64 string to binary data\n */\nexport function decodeUrlSafe(base64: string): BinaryData {\n // Add back padding if needed\n let padded = base64\n .replace(/-/g, '+')\n .replace(/_/g, '/');\n\n while (padded.length % 4 !== 0) {\n padded += '=';\n }\n\n return decode(padded);\n}\n\n/**\n * Chunk size for encoding large buffers (64KB)\n * Prevents stack overflow when using String.fromCharCode with spread operator\n */\nconst ENCODE_CHUNK_SIZE = 64 * 1024;\n\n/**\n * Encode binary data to base64 string with chunking for large buffers.\n * Handles buffers >10MB without stack overflow.\n *\n * @param data - Binary data to encode\n * @returns Base64 encoded string\n *\n * @example\n * ```typescript\n * const largeBuffer = new Uint8Array(20 * 1024 * 1024); // 20MB\n * const base64 = encodeChunked(largeBuffer); // No stack overflow\n * ```\n */\nexport function encodeChunked(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: process in chunks to avoid stack overflow\n const chunks: string[] = [];\n\n for (let i = 0; i < data.length; i += ENCODE_CHUNK_SIZE) {\n const chunk = data.subarray(i, Math.min(i + ENCODE_CHUNK_SIZE, data.length));\n let binary = '';\n for (let j = 0; j < chunk.length; j++) {\n binary += String.fromCharCode(chunk[j]!);\n }\n chunks.push(binary);\n }\n\n return btoa(chunks.join(''));\n}\n","/**\n * Error Classes\n *\n * Specific error types for character card operations.\n * All errors extend FoundryError for consistent handling.\n */\n\n/** Symbol to identify FoundryError instances across ESM/CJS boundaries */\nconst FOUNDRY_ERROR_MARKER = Symbol.for('@character-foundry/core:FoundryError');\n\n/**\n * Base error class for all Character Foundry errors\n */\nexport class FoundryError extends Error {\n /** @internal Marker for cross-module identification */\n readonly [FOUNDRY_ERROR_MARKER] = true;\n\n constructor(message: string, public readonly code: string) {\n super(message);\n this.name = 'FoundryError';\n // Maintains proper stack trace in V8 environments\n if (Error.captureStackTrace) {\n Error.captureStackTrace(this, this.constructor);\n }\n }\n}\n\n/**\n * Error during card parsing\n */\nexport class ParseError extends FoundryError {\n constructor(message: string, public readonly format?: string) {\n super(message, 'PARSE_ERROR');\n this.name = 'ParseError';\n }\n}\n\n/**\n * Error during card validation\n */\nexport class ValidationError extends FoundryError {\n constructor(message: string, public readonly field?: string) {\n super(message, 'VALIDATION_ERROR');\n this.name = 'ValidationError';\n }\n}\n\n/**\n * Asset not found in card or archive\n */\nexport class AssetNotFoundError extends FoundryError {\n constructor(public readonly uri: string) {\n super(`Asset not found: ${uri}`, 'ASSET_NOT_FOUND');\n this.name = 'AssetNotFoundError';\n }\n}\n\n/**\n * Format not supported for operation\n */\nexport class FormatNotSupportedError extends FoundryError {\n constructor(public readonly format: string, operation?: string) {\n const msg = operation\n ? `Format '${format}' not supported for ${operation}`\n : `Format not supported: ${format}`;\n super(msg, 'FORMAT_NOT_SUPPORTED');\n this.name = 'FormatNotSupportedError';\n }\n}\n\n/**\n * File size exceeds limits\n */\nexport class SizeLimitError extends FoundryError {\n constructor(\n public readonly actualSize: number,\n public readonly maxSize: number,\n context?: string\n ) {\n const actualMB = (actualSize / 1024 / 1024).toFixed(2);\n const maxMB = (maxSize / 1024 / 1024).toFixed(2);\n const msg = context\n ? `${context}: Size ${actualMB}MB exceeds limit ${maxMB}MB`\n : `Size ${actualMB}MB exceeds limit ${maxMB}MB`;\n super(msg, 'SIZE_LIMIT_EXCEEDED');\n this.name = 'SizeLimitError';\n }\n}\n\n/**\n * Path traversal or unsafe path detected\n */\nexport class PathTraversalError extends FoundryError {\n constructor(public readonly path: string) {\n super(`Unsafe path detected: ${path}`, 'PATH_TRAVERSAL');\n this.name = 'PathTraversalError';\n }\n}\n\n/**\n * Export operation would lose data\n */\nexport class DataLossError extends FoundryError {\n constructor(\n public readonly lostFields: string[],\n public readonly targetFormat: string\n ) {\n const fields = lostFields.slice(0, 3).join(', ');\n const more = lostFields.length > 3 ? ` and ${lostFields.length - 3} more` : '';\n super(\n `Export to ${targetFormat} would lose: ${fields}${more}`,\n 'DATA_LOSS'\n );\n this.name = 'DataLossError';\n }\n}\n\n/**\n * Check if an error is a FoundryError\n *\n * Uses Symbol.for() marker instead of instanceof to handle dual ESM/CJS package loading.\n * In dual-package environments, instanceof can fail if the error comes from a different\n * module instance (e.g., ESM vs CJS version of the same package). Symbol.for() creates\n * a global symbol shared across all module instances.\n */\nexport function isFoundryError(error: unknown): error is FoundryError {\n return (\n error instanceof Error &&\n FOUNDRY_ERROR_MARKER in error &&\n (error as Record<symbol, unknown>)[FOUNDRY_ERROR_MARKER] === true\n );\n}\n\n/**\n * Wrap unknown errors in a FoundryError\n */\nexport function wrapError(error: unknown, context?: string): FoundryError {\n if (isFoundryError(error)) {\n return error;\n }\n\n const message = error instanceof Error\n ? error.message\n : String(error);\n\n return new FoundryError(\n context ? `${context}: ${message}` : message,\n 'UNKNOWN_ERROR'\n );\n}\n","/**\n * Data URL Utilities\n *\n * Convert between Uint8Array buffers and data URLs.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n */\n\nimport type { BinaryData } from './binary.js';\nimport { encodeChunked as base64Encode, decode as base64Decode } from './base64.js';\nimport { ValidationError } from './errors.js';\n\n/**\n * Convert Uint8Array to data URL.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n *\n * @param buffer - Binary data to encode\n * @param mimeType - MIME type for the data URL (e.g., 'image/png', 'application/octet-stream')\n * @returns Data URL string\n *\n * @example\n * ```typescript\n * const png = new Uint8Array([...]);\n * const dataUrl = toDataURL(png, 'image/png');\n * // => \"data:image/png;base64,iVBORw0KGgo...\"\n * ```\n */\nexport function toDataURL(buffer: BinaryData, mimeType: string): string {\n // Use chunked encoding to handle large buffers without stack overflow\n const base64 = base64Encode(buffer);\n return `data:${mimeType};base64,${base64}`;\n}\n\n/**\n * Parse a data URL back to buffer and MIME type.\n * Validates the data URL format before parsing.\n *\n * @param dataUrl - Data URL string to parse\n * @returns Object containing the decoded buffer and MIME type\n * @throws Error if the data URL format is invalid\n *\n * @example\n * ```typescript\n * const { buffer, mimeType } = fromDataURL('data:image/png;base64,iVBORw0KGgo...');\n * // buffer: Uint8Array\n * // mimeType: 'image/png'\n * ```\n */\nexport function fromDataURL(dataUrl: string): { buffer: Uint8Array; mimeType: string } {\n // Validate data URL format\n if (!dataUrl.startsWith('data:')) {\n throw new ValidationError('Invalid data URL: must start with \"data:\"', 'dataUrl');\n }\n\n const commaIndex = dataUrl.indexOf(',');\n if (commaIndex === -1) {\n throw new ValidationError('Invalid data URL: missing comma separator', 'dataUrl');\n }\n\n const header = dataUrl.slice(5, commaIndex); // Skip 'data:'\n const data = dataUrl.slice(commaIndex + 1);\n\n // Parse header: [<mediatype>][;base64]\n let mimeType = 'text/plain';\n let isBase64 = false;\n\n const parts = header.split(';');\n for (const part of parts) {\n if (part === 'base64') {\n isBase64 = true;\n } else if (part && !part.includes('=')) {\n // MIME type (not a parameter like charset=utf-8)\n mimeType = part;\n }\n }\n\n if (!isBase64) {\n // URL-encoded text data\n throw new ValidationError('Non-base64 data URLs are not supported', 'dataUrl');\n }\n\n const buffer = base64Decode(data);\n return { buffer, mimeType };\n}\n\n/**\n * Check if a string is a valid data URL\n *\n * @param str - String to check\n * @returns true if the string is a valid data URL format\n */\nexport function isDataURL(str: string): boolean {\n if (!str.startsWith('data:')) return false;\n const commaIndex = str.indexOf(',');\n if (commaIndex === -1) return false;\n const header = str.slice(5, commaIndex);\n return header.includes('base64');\n}\n","/**\n * URI Utilities\n *\n * Handles different asset URI schemes used in character cards.\n * Supports: embeded://, embedded://, ccdefault:, https://, http://,\n * data:, file://, __asset:, asset:, chara-ext-asset_\n */\n\nexport type URIScheme =\n | 'embeded' // embeded:// (CharX standard, note intentional typo)\n | 'ccdefault' // ccdefault:\n | 'https' // https://\n | 'http' // http://\n | 'data' // data:mime;base64,...\n | 'file' // file://\n | 'internal' // Internal asset ID (UUID/string)\n | 'pngchunk' // PNG chunk reference (__asset:, asset:, chara-ext-asset_)\n | 'unknown';\n\nexport interface ParsedURI {\n scheme: URIScheme;\n originalUri: string;\n normalizedUri: string; // Normalized form of the URI\n path?: string; // For embeded://, file://\n url?: string; // For http://, https://\n data?: string; // For data: URIs\n mimeType?: string; // For data: URIs\n encoding?: string; // For data: URIs (e.g., base64)\n chunkKey?: string; // For pngchunk - the key/index to look up\n chunkCandidates?: string[]; // For pngchunk - all possible chunk keys to search\n}\n\n/**\n * Normalize a URI to its canonical form\n * Handles common typos and variant formats\n */\nexport function normalizeURI(uri: string): string {\n const trimmed = uri.trim();\n\n // Fix embedded:// -> embeded:// (common typo, CharX spec uses single 'd')\n if (trimmed.startsWith('embedded://')) {\n return 'embeded://' + trimmed.substring('embedded://'.length);\n }\n\n // Normalize PNG chunk references to pngchunk: scheme\n if (trimmed.startsWith('__asset:')) {\n const id = trimmed.substring('__asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('asset:')) {\n const id = trimmed.substring('asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_:')) {\n const id = trimmed.substring('chara-ext-asset_:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_')) {\n const id = trimmed.substring('chara-ext-asset_'.length);\n return `pngchunk:${id}`;\n }\n\n return trimmed;\n}\n\n/**\n * Parse a URI and determine its scheme and components\n */\nexport function parseURI(uri: string): ParsedURI {\n const trimmed = uri.trim();\n const normalized = normalizeURI(trimmed);\n\n // PNG chunk references (__asset:, asset:, chara-ext-asset_, pngchunk:)\n if (\n trimmed.startsWith('__asset:') ||\n trimmed.startsWith('asset:') ||\n trimmed.startsWith('chara-ext-asset_') ||\n trimmed.startsWith('pngchunk:')\n ) {\n let assetId: string;\n if (trimmed.startsWith('__asset:')) {\n assetId = trimmed.substring('__asset:'.length);\n } else if (trimmed.startsWith('asset:')) {\n assetId = trimmed.substring('asset:'.length);\n } else if (trimmed.startsWith('chara-ext-asset_:')) {\n assetId = trimmed.substring('chara-ext-asset_:'.length);\n } else if (trimmed.startsWith('pngchunk:')) {\n assetId = trimmed.substring('pngchunk:'.length);\n } else {\n assetId = trimmed.substring('chara-ext-asset_'.length);\n }\n\n // Generate all possible chunk key variations for lookup\n const candidates = [\n assetId, // \"0\" or \"filename.png\"\n trimmed, // Original URI\n `asset:${assetId}`, // \"asset:0\"\n `__asset:${assetId}`, // \"__asset:0\"\n `__asset_${assetId}`, // \"__asset_0\"\n `chara-ext-asset_${assetId}`, // \"chara-ext-asset_0\"\n `chara-ext-asset_:${assetId}`, // \"chara-ext-asset_:0\"\n `pngchunk:${assetId}`, // \"pngchunk:0\"\n ];\n\n return {\n scheme: 'pngchunk',\n originalUri: uri,\n normalizedUri: normalized,\n chunkKey: assetId,\n chunkCandidates: candidates,\n };\n }\n\n // ccdefault: - use default asset\n if (trimmed === 'ccdefault:' || trimmed.startsWith('ccdefault:')) {\n return {\n scheme: 'ccdefault',\n originalUri: uri,\n normalizedUri: normalized,\n };\n }\n\n // embeded:// or embedded:// (normalize typo)\n if (trimmed.startsWith('embeded://') || trimmed.startsWith('embedded://')) {\n const path = trimmed.startsWith('embeded://')\n ? trimmed.substring('embeded://'.length)\n : trimmed.substring('embedded://'.length);\n return {\n scheme: 'embeded',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // https://\n if (trimmed.startsWith('https://')) {\n return {\n scheme: 'https',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // http://\n if (trimmed.startsWith('http://')) {\n return {\n scheme: 'http',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // data: URIs\n if (trimmed.startsWith('data:')) {\n const parsed = parseDataURI(trimmed);\n return {\n scheme: 'data',\n originalUri: uri,\n normalizedUri: normalized,\n ...parsed,\n };\n }\n\n // file://\n if (trimmed.startsWith('file://')) {\n const path = trimmed.substring('file://'.length);\n return {\n scheme: 'file',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // Internal asset ID (alphanumeric/UUID format)\n if (/^[a-zA-Z0-9_-]+$/.test(trimmed)) {\n return {\n scheme: 'internal',\n originalUri: uri,\n normalizedUri: normalized,\n path: trimmed,\n };\n }\n\n // Unknown scheme\n return {\n scheme: 'unknown',\n originalUri: uri,\n normalizedUri: normalized,\n };\n}\n\n/**\n * Parse a data URI into its components\n * Format: data:[<mediatype>][;base64],<data>\n */\nfunction parseDataURI(uri: string): { mimeType?: string; encoding?: string; data?: string } {\n const match = uri.match(/^data:([^;,]+)?(;base64)?,(.*)$/);\n\n if (!match) {\n return {};\n }\n\n return {\n mimeType: match[1] || 'text/plain',\n encoding: match[2] ? 'base64' : undefined,\n data: match[3],\n };\n}\n\n/**\n * Check if extension is an image format\n */\nexport function isImageExt(ext: string): boolean {\n const imageExts = ['png', 'jpg', 'jpeg', 'webp', 'gif', 'avif', 'bmp', 'svg'];\n return imageExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is an audio format\n */\nexport function isAudioExt(ext: string): boolean {\n const audioExts = ['mp3', 'wav', 'ogg', 'flac', 'm4a', 'aac'];\n return audioExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is a video format\n */\nexport function isVideoExt(ext: string): boolean {\n const videoExts = ['mp4', 'webm', 'avi', 'mov', 'mkv'];\n return videoExts.includes(ext.toLowerCase());\n}\n\n/** Safe MIME types for data: URIs that can be used in href/src */\nconst SAFE_DATA_URI_MIME_TYPES = new Set([\n // Images (safe for img src)\n 'image/png',\n 'image/jpeg',\n 'image/gif',\n 'image/webp',\n 'image/avif',\n 'image/bmp',\n 'image/x-icon',\n // Audio (safe for audio src)\n 'audio/mpeg',\n 'audio/wav',\n 'audio/ogg',\n 'audio/flac',\n 'audio/mp4',\n 'audio/aac',\n // Video (safe for video src)\n 'video/mp4',\n 'video/webm',\n // Text/data (generally safe)\n 'text/plain',\n 'application/json',\n 'application/octet-stream',\n]);\n\n/** Potentially dangerous MIME types that should NOT be used in href/src */\nconst DANGEROUS_DATA_URI_MIME_TYPES = new Set([\n // Executable/script content\n 'text/html',\n 'text/javascript',\n 'application/javascript',\n 'application/x-javascript',\n 'text/css',\n 'image/svg+xml', // SVG can contain scripts\n 'application/xhtml+xml',\n 'application/xml',\n]);\n\n/**\n * Options for URI safety validation\n */\nexport interface URISafetyOptions {\n /** Allow http:// URIs (default: false) */\n allowHttp?: boolean;\n /** Allow file:// URIs (default: false) */\n allowFile?: boolean;\n /**\n * Allowed MIME types for data: URIs (default: all safe types).\n * Set to empty array to reject all data: URIs.\n * Set to undefined to use default safe list.\n */\n allowedDataMimes?: string[];\n}\n\n/**\n * Result of URI safety check with detailed information\n */\nexport interface URISafetyResult {\n /** Whether the URI is safe to use */\n safe: boolean;\n /** Reason if unsafe */\n reason?: string;\n /** Detected scheme */\n scheme: URIScheme;\n /** MIME type for data: URIs */\n mimeType?: string;\n}\n\n/**\n * Validate if a URI is safe to use (detailed version)\n *\n * @param uri - URI to validate\n * @param options - Safety options\n * @returns Detailed safety result\n */\nexport function checkURISafety(uri: string, options: URISafetyOptions = {}): URISafetyResult {\n const parsed = parseURI(uri);\n\n switch (parsed.scheme) {\n case 'embeded':\n case 'ccdefault':\n case 'internal':\n case 'https':\n case 'pngchunk':\n return { safe: true, scheme: parsed.scheme };\n\n case 'data': {\n const mimeType = parsed.mimeType || 'text/plain';\n\n // Check for explicitly dangerous MIME types\n if (DANGEROUS_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI with potentially dangerous MIME type: ${mimeType}`,\n };\n }\n\n // If custom allowed list is provided, check against it\n if (options.allowedDataMimes !== undefined) {\n if (options.allowedDataMimes.length === 0) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: 'Data URIs are not allowed',\n };\n }\n if (!options.allowedDataMimes.includes(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI MIME type not in allowed list: ${mimeType}`,\n };\n }\n }\n\n // Otherwise use default safe list\n if (!SAFE_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Unknown data URI MIME type: ${mimeType}`,\n };\n }\n\n return { safe: true, scheme: parsed.scheme, mimeType };\n }\n\n case 'http':\n if (options.allowHttp === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'HTTP URIs are not allowed' };\n\n case 'file':\n if (options.allowFile === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'File URIs are not allowed' };\n\n case 'unknown':\n default:\n return { safe: false, scheme: parsed.scheme, reason: 'Unknown URI scheme' };\n }\n}\n\n/**\n * Validate if a URI is safe to use (simple boolean version for backwards compatibility)\n *\n * @deprecated Use checkURISafety() for detailed safety information\n */\nexport function isURISafe(uri: string, options: { allowHttp?: boolean; allowFile?: boolean } = {}): boolean {\n return checkURISafety(uri, options).safe;\n}\n\n/**\n * Extract file extension from URI\n */\nexport function getExtensionFromURI(uri: string): string {\n const parsed = parseURI(uri);\n\n if (parsed.path) {\n const parts = parsed.path.split('.');\n if (parts.length > 1) {\n return parts[parts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.url) {\n const urlParts = parsed.url.split('?')[0]!.split('.');\n if (urlParts.length > 1) {\n return urlParts[urlParts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.mimeType) {\n return getExtFromMimeType(parsed.mimeType);\n }\n\n return 'unknown';\n}\n\n/**\n * Get MIME type from file extension\n */\nexport function getMimeTypeFromExt(ext: string): string {\n const extToMime: Record<string, string> = {\n // Images\n 'png': 'image/png',\n 'jpg': 'image/jpeg',\n 'jpeg': 'image/jpeg',\n 'webp': 'image/webp',\n 'gif': 'image/gif',\n 'avif': 'image/avif',\n 'svg': 'image/svg+xml',\n 'bmp': 'image/bmp',\n 'ico': 'image/x-icon',\n\n // Audio\n 'mp3': 'audio/mpeg',\n 'wav': 'audio/wav',\n 'ogg': 'audio/ogg',\n 'flac': 'audio/flac',\n 'm4a': 'audio/mp4',\n 'aac': 'audio/aac',\n\n // Video\n 'mp4': 'video/mp4',\n 'webm': 'video/webm',\n 'avi': 'video/x-msvideo',\n 'mov': 'video/quicktime',\n 'mkv': 'video/x-matroska',\n\n // Text/Data\n 'json': 'application/json',\n 'txt': 'text/plain',\n 'html': 'text/html',\n 'css': 'text/css',\n 'js': 'application/javascript',\n };\n\n return extToMime[ext.toLowerCase()] || 'application/octet-stream';\n}\n\n/**\n * Get file extension from MIME type\n */\nexport function getExtFromMimeType(mimeType: string): string {\n const mimeToExt: Record<string, string> = {\n 'image/png': 'png',\n 'image/jpeg': 'jpg',\n 'image/webp': 'webp',\n 'image/gif': 'gif',\n 'image/avif': 'avif',\n 'image/svg+xml': 'svg',\n 'image/bmp': 'bmp',\n 'image/x-icon': 'ico',\n 'audio/mpeg': 'mp3',\n 'audio/wav': 'wav',\n 'audio/ogg': 'ogg',\n 'audio/flac': 'flac',\n 'audio/mp4': 'm4a',\n 'audio/aac': 'aac',\n 'video/mp4': 'mp4',\n 'video/webm': 'webm',\n 'video/x-msvideo': 'avi',\n 'video/quicktime': 'mov',\n 'video/x-matroska': 'mkv',\n 'application/json': 'json',\n 'text/plain': 'txt',\n 'text/html': 'html',\n 'text/css': 'css',\n 'application/javascript': 'js',\n };\n\n return mimeToExt[mimeType] || 'bin';\n}\n\n/**\n * Build a data URI from binary data and MIME type\n */\nexport function buildDataURI(data: string, mimeType: string, isBase64 = true): string {\n if (isBase64) {\n return `data:${mimeType};base64,${data}`;\n }\n return `data:${mimeType},${encodeURIComponent(data)}`;\n}\n","/**\n * Image Analysis Utilities\n *\n * Detect properties of image files from binary data.\n */\n\nimport {\n type BinaryData,\n indexOf,\n fromLatin1,\n} from './binary.js';\n\n/**\n * Check if an image buffer contains animation data.\n * Supports: APNG, WebP (Animated), GIF\n */\nexport function isAnimatedImage(data: BinaryData, _mimeType?: string): boolean {\n // 1. WebP Detection\n // RIFF .... WEBP\n if (\n data.length > 12 &&\n data[0] === 0x52 && data[1] === 0x49 && data[2] === 0x46 && data[3] === 0x46 && // RIFF\n data[8] === 0x57 && data[9] === 0x45 && data[10] === 0x42 && data[11] === 0x50 // WEBP\n ) {\n // Check for VP8X chunk\n // VP8X chunk header: 'VP8X' (bytes 12-15)\n if (\n data[12] === 0x56 && data[13] === 0x50 && data[14] === 0x38 && data[15] === 0x58\n ) {\n // Flags byte is at offset 20 (16 + 4 bytes chunk size)\n // Animation bit is bit 1 (0x02)\n const flags = data[20];\n return (flags! & 0x02) !== 0;\n }\n return false;\n }\n\n // 2. PNG/APNG Detection\n // Signature: 89 50 4E 47 0D 0A 1A 0A\n if (\n data.length > 8 &&\n data[0] === 0x89 && data[1] === 0x50 && data[2] === 0x4E && data[3] === 0x47\n ) {\n // Search for 'acTL' chunk (Animation Control)\n // It must appear before IDAT.\n // Simple search: indexOf('acTL')\n // Note: theoretically 'acTL' string could appear in other data, but highly unlikely in valid PNG structure before IDAT\n // We can iterate chunks to be safe, but indexOf is faster for a quick check\n const actlSig = fromLatin1('acTL');\n const idatSig = fromLatin1('IDAT');\n \n const actlIndex = indexOf(data, actlSig);\n if (actlIndex === -1) return false;\n\n const idatIndex = indexOf(data, idatSig);\n // If acTL exists and is before the first IDAT (or IDAT not found yet), it's APNG\n return idatIndex === -1 || actlIndex < idatIndex;\n }\n\n // 3. GIF Detection\n // Signature: GIF87a or GIF89a\n if (\n data.length > 6 &&\n data[0] === 0x47 && data[1] === 0x49 && data[2] === 0x46 // GIF\n ) {\n // Check for NETSCAPE2.0 extension (looping animation)\n // This is a heuristic. Static GIFs are rare in this domain but possible.\n // Full frame counting is expensive. Presence of NETSCAPE block is a strong indicator.\n const netscape = fromLatin1('NETSCAPE2.0');\n return indexOf(data, netscape) !== -1;\n }\n\n return false;\n}\n","/**\n * UUID Generation Utilities\n *\n * Provides crypto-grade UUID v4 generation that works in Node.js,\n * browsers (secure contexts), and falls back gracefully.\n */\n\n/**\n * Format 16 random bytes as a UUID v4 string\n */\nfunction formatUUID(bytes: Uint8Array): string {\n const hex = Array.from(bytes, (b) => b.toString(16).padStart(2, '0')).join('');\n return `${hex.slice(0, 8)}-${hex.slice(8, 12)}-${hex.slice(12, 16)}-${hex.slice(16, 20)}-${hex.slice(20, 32)}`;\n}\n\n/**\n * Fallback UUID generation using Math.random()\n * Only used when crypto APIs are unavailable (rare)\n */\nfunction mathRandomUUID(): string {\n return 'xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx'.replace(/[xy]/g, (c) => {\n const r = (Math.random() * 16) | 0;\n const v = c === 'x' ? r : (r & 0x3) | 0x8;\n return v.toString(16);\n });\n}\n\n/**\n * Generate a cryptographically secure UUID v4.\n *\n * Uses crypto.randomUUID() when available (Node.js 19+, modern browsers).\n * Falls back to crypto.getRandomValues() if randomUUID is unavailable.\n * Last resort uses Math.random() (non-secure, emits warning in dev).\n *\n * @returns A valid RFC 4122 UUID v4 string\n *\n * @example\n * ```typescript\n * const id = generateUUID();\n * // => \"550e8400-e29b-41d4-a716-446655440000\"\n * ```\n */\nexport function generateUUID(): string {\n // Node.js 19+ or browser with secure context\n if (typeof crypto !== 'undefined' && typeof crypto.randomUUID === 'function') {\n return crypto.randomUUID();\n }\n\n // Fallback using crypto.getRandomValues (older Node/browsers)\n if (typeof crypto !== 'undefined' && typeof crypto.getRandomValues === 'function') {\n const bytes = new Uint8Array(16);\n crypto.getRandomValues(bytes);\n // Set version (4) and variant (RFC 4122)\n bytes[6] = (bytes[6]! & 0x0f) | 0x40; // Version 4\n bytes[8] = (bytes[8]! & 0x3f) | 0x80; // Variant 1\n return formatUUID(bytes);\n }\n\n // Last resort - non-secure fallback\n if (typeof process !== 'undefined' && process.env?.NODE_ENV === 'development') {\n console.warn('[character-foundry/core] generateUUID: Using insecure Math.random() fallback');\n }\n return mathRandomUUID();\n}\n\n/**\n * Validate if a string is a valid UUID v4\n *\n * @param uuid - String to validate\n * @returns true if valid UUID v4 format\n */\nexport function isValidUUID(uuid: string): boolean {\n return /^[0-9a-f]{8}-[0-9a-f]{4}-4[0-9a-f]{3}-[89ab][0-9a-f]{3}-[0-9a-f]{12}$/i.test(uuid);\n}\n","/**\n * Common Types\n *\n * Shared types used across all card formats.\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * ISO 8601 date string schema\n */\nexport const ISO8601Schema = z.string().datetime();\n\n/**\n * UUID string schema\n */\nexport const UUIDSchema = z.string().uuid();\n\n/**\n * Card specification version schema\n */\nexport const SpecSchema = z.enum(['v2', 'v3']);\n\n/**\n * Source format identifier schema\n */\nexport const SourceFormatSchema = z.enum([\n 'png_v2', // PNG with 'chara' chunk (v2)\n 'png_v3', // PNG with 'ccv3' chunk (v3)\n 'json_v2', // Raw JSON v2\n 'json_v3', // Raw JSON v3\n 'charx', // ZIP with card.json (v3 spec)\n 'charx_risu', // ZIP with card.json + module.risum\n 'charx_jpeg', // JPEG with appended ZIP (read-only)\n 'voxta', // VoxPkg format\n]);\n\n/**\n * Original JSON shape schema\n */\nexport const OriginalShapeSchema = z.enum(['wrapped', 'unwrapped', 'legacy']);\n\n/**\n * Asset type identifier schema\n */\nexport const AssetTypeSchema = z.enum([\n 'icon',\n 'background',\n 'emotion',\n 'user_icon',\n 'sound',\n 'video',\n 'custom',\n 'x-risu-asset',\n]);\n\n/**\n * Asset descriptor schema (v3 spec)\n */\nexport const AssetDescriptorSchema = z.object({\n type: AssetTypeSchema,\n uri: z.string(),\n name: z.string(),\n ext: z.string(),\n});\n\n/**\n * Extracted asset with binary data schema\n */\nexport const ExtractedAssetSchema = z.object({\n descriptor: AssetDescriptorSchema,\n data: z.instanceof(Uint8Array),\n mimeType: z.string(),\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * ISO 8601 date string\n */\nexport type ISO8601 = z.infer<typeof ISO8601Schema>;\n\n/**\n * UUID string\n */\nexport type UUID = z.infer<typeof UUIDSchema>;\n\n/**\n * Card specification version\n */\nexport type Spec = z.infer<typeof SpecSchema>;\n\n/**\n * Source format identifier\n */\nexport type SourceFormat = z.infer<typeof SourceFormatSchema>;\n\n/**\n * Original JSON shape\n */\nexport type OriginalShape = z.infer<typeof OriginalShapeSchema>;\n\n/**\n * Asset type identifier\n */\nexport type AssetType = z.infer<typeof AssetTypeSchema>;\n\n/**\n * Asset descriptor (v3 spec)\n */\nexport type AssetDescriptor = z.infer<typeof AssetDescriptorSchema>;\n\n/**\n * Extracted asset with binary data\n */\nexport type ExtractedAsset = z.infer<typeof ExtractedAssetSchema>;\n","/**\n * Character Card v2 Types\n *\n * Based on: https://github.com/malfoyslastname/character-card-spec-v2\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v2 cards\n */\nexport const CCv2LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n extensions: z.record(z.unknown()).optional(),\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char']), z.number().int(), z.literal('')]).nullable().optional(),\n}).passthrough(); // Allow SillyTavern extensions like depth, probability, etc.\n\n/**\n * Character book (lorebook) schema for v2 cards\n */\nexport const CCv2CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv2LorebookEntrySchema),\n});\n\n/**\n * Character Card v2 data structure schema\n */\nexport const CCv2DataSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv2CharacterBookSchema.optional().nullable(),\n tags: z.array(z.string()).optional(),\n creator: z.string().optional(),\n character_version: z.string().optional(),\n extensions: z.record(z.unknown()).optional(),\n});\n\n/**\n * Wrapped v2 card format schema (modern tools)\n */\nexport const CCv2WrappedSchema = z.object({\n spec: z.literal('chara_card_v2'),\n spec_version: z.literal('2.0'),\n data: CCv2DataSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v2 cards\n */\nexport type CCv2LorebookEntry = z.infer<typeof CCv2LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v2 cards\n */\nexport type CCv2CharacterBook = z.infer<typeof CCv2CharacterBookSchema>;\n\n/**\n * Character Card v2 data structure\n */\nexport type CCv2Data = z.infer<typeof CCv2DataSchema>;\n\n/**\n * Wrapped v2 card format (modern tools)\n */\nexport type CCv2Wrapped = z.infer<typeof CCv2WrappedSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a wrapped v2 card\n */\nexport function isWrappedV2(data: unknown): data is CCv2Wrapped {\n return CCv2WrappedSchema.safeParse(data).success;\n}\n\n/**\n * Check if data looks like v2 card data (wrapped or unwrapped)\n */\nexport function isV2CardData(data: unknown): data is CCv2Data | CCv2Wrapped {\n return (\n CCv2WrappedSchema.safeParse(data).success ||\n CCv2DataSchema.safeParse(data).success\n );\n}\n\n/**\n * Parse and validate a wrapped v2 card\n */\nexport function parseWrappedV2(data: unknown): CCv2Wrapped {\n return CCv2WrappedSchema.parse(data);\n}\n\n/**\n * Parse and validate v2 card data\n */\nexport function parseV2Data(data: unknown): CCv2Data {\n return CCv2DataSchema.parse(data);\n}\n\n/**\n * Check if data looks like a wrapped V2 card structurally (without strict validation).\n * This is more lenient than isWrappedV2 - it just checks structure, not full schema validity.\n */\nexport function looksLikeWrappedV2(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v2' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n\n/**\n * Get v2 card data from wrapped or unwrapped format.\n *\n * Uses structural check instead of strict Zod validation to handle\n * malformed cards that have the right structure but missing/invalid fields.\n * The caller (e.g., ccv2ToCCv3) handles defaulting missing fields.\n */\nexport function getV2Data(card: CCv2Data | CCv2Wrapped): CCv2Data {\n // Use structural check - more lenient than isWrappedV2 schema validation\n if (looksLikeWrappedV2(card)) {\n return card.data as CCv2Data;\n }\n return card;\n}\n","/**\n * Character Card v3 Types\n *\n * Based on: https://github.com/kwaroran/character-card-spec-v3\n */\n\nimport { z } from 'zod';\nimport { AssetDescriptorSchema } from './common.js';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v3 cards\n */\nexport const CCv3LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char']), z.number().int(), z.literal('')]).nullable().optional(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific - also lenient with types since SillyTavern uses numbers for enums\n automation_id: z.string().optional(),\n role: z.union([z.enum(['system', 'user', 'assistant']), z.number().int()]).nullable().optional(),\n group: z.string().optional(),\n scan_frequency: z.number().int().nonnegative().optional(),\n probability: z.number().min(0).max(100).optional(), // Some tools use 0-100 instead of 0-1\n use_regex: z.boolean().optional(),\n depth: z.number().int().nonnegative().optional(),\n selective_logic: z.union([z.enum(['AND', 'NOT']), z.number().int()]).optional(),\n}).passthrough(); // Allow tool-specific extensions\n\n/**\n * Character book (lorebook) schema for v3 cards\n */\nexport const CCv3CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv3LorebookEntrySchema),\n});\n\n/**\n * Character Card v3 inner data structure schema.\n *\n * Note: Fields like group_only_greetings, creator, character_version, and tags\n * are technically \"required\" per V3 spec but rarely present in wild cards.\n * We use .default() to make parsing lenient while still producing valid output.\n */\nexport const CCv3DataInnerSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // \"Required\" per spec but often missing in wild - use defaults for leniency\n creator: z.string().default(''),\n character_version: z.string().default(''),\n tags: z.array(z.string()).default([]),\n group_only_greetings: z.array(z.string()).default([]),\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv3CharacterBookSchema.optional().nullable(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific\n assets: z.array(AssetDescriptorSchema).optional(),\n nickname: z.string().optional(),\n creator_notes_multilingual: z.record(z.string()).optional(),\n source: z.array(z.string()).optional(),\n creation_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n modification_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n});\n\n/**\n * Character Card v3 full structure schema\n */\nexport const CCv3DataSchema = z.object({\n spec: z.literal('chara_card_v3'),\n spec_version: z.literal('3.0'),\n data: CCv3DataInnerSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v3 cards\n */\nexport type CCv3LorebookEntry = z.infer<typeof CCv3LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v3 cards\n */\nexport type CCv3CharacterBook = z.infer<typeof CCv3CharacterBookSchema>;\n\n/**\n * Character Card v3 inner data structure\n */\nexport type CCv3DataInner = z.infer<typeof CCv3DataInnerSchema>;\n\n/**\n * Character Card v3 full structure\n */\nexport type CCv3Data = z.infer<typeof CCv3DataSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a v3 card\n */\nexport function isV3Card(data: unknown): data is CCv3Data {\n return CCv3DataSchema.safeParse(data).success;\n}\n\n/**\n * Parse and validate a v3 card\n */\nexport function parseV3Card(data: unknown): CCv3Data {\n return CCv3DataSchema.parse(data);\n}\n\n/**\n * Parse and validate v3 card inner data\n */\nexport function parseV3DataInner(data: unknown): CCv3DataInner {\n return CCv3DataInnerSchema.parse(data);\n}\n\n/**\n * Get v3 card inner data\n */\nexport function getV3Data(card: CCv3Data): CCv3DataInner {\n return card.data;\n}\n\n/**\n * Check if data looks like a V3 card structurally (without strict validation).\n * More lenient than isV3Card - just checks structure, not full schema validity.\n */\nexport function looksLikeV3Card(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v3' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n","/**\n * RisuAI Extension Types\n *\n * These extensions are preserved as opaque blobs.\n * We do NOT interpret or transform the script contents.\n */\n\n/**\n * Risu emotions mapping (v2 style)\n * Format: [name, uri][]\n */\nexport type RisuEmotions = [string, string][];\n\n/**\n * Risu additional assets (v3 style)\n * Format: [name, uri, type][]\n */\nexport type RisuAdditionalAssets = [string, string, string][];\n\n/**\n * Risu depth prompt configuration\n */\nexport interface RisuDepthPrompt {\n depth: number;\n prompt: string;\n}\n\n/**\n * Risu extensions in card.extensions.risuai\n * Preserved as opaque - we don't interpret script contents\n */\nexport interface RisuExtensions {\n // Emotion assets\n emotions?: RisuEmotions;\n additionalAssets?: RisuAdditionalAssets;\n\n // Script data - OPAQUE, do not parse\n triggerscript?: unknown;\n customScripts?: unknown;\n\n // Voice/TTS settings\n vits?: Record<string, string>;\n\n // Depth prompt\n depth_prompt?: RisuDepthPrompt;\n\n // Other Risu-specific fields\n [key: string]: unknown;\n}\n\n/**\n * CharX x_meta entry (PNG chunk metadata preservation)\n */\nexport interface CharxMetaEntry {\n type?: string; // e.g., 'WEBP', 'PNG', 'JPEG'\n [key: string]: unknown;\n}\n\n/**\n * Check if card has Risu extensions\n */\nexport function hasRisuExtensions(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n return 'risuai' in extensions || 'risu' in extensions;\n}\n\n/**\n * Check if card has Risu scripts (triggerscript or customScripts)\n */\nexport function hasRisuScripts(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n const risu = extensions.risuai as RisuExtensions | undefined;\n if (!risu) return false;\n return !!risu.triggerscript || !!risu.customScripts;\n}\n\n/**\n * Check if card has depth prompt\n * Checks both SillyTavern style (extensions.depth_prompt) and Risu style (extensions.risuai.depth_prompt)\n */\nexport function hasDepthPrompt(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n // SillyTavern top-level depth_prompt\n if ('depth_prompt' in extensions && extensions.depth_prompt) return true;\n // Risu-style depth_prompt\n const risu = extensions.risuai as RisuExtensions | undefined;\n return !!risu?.depth_prompt;\n}\n","/**\n * Normalized Card Types\n *\n * Unified view of card data regardless of source format.\n * This is a computed/virtual representation, not stored.\n */\n\nimport type { CCv3CharacterBook } from './ccv3.js';\n\n/**\n * Normalized card representation\n * Provides unified access to card data from any format\n */\nexport interface NormalizedCard {\n // Core fields (always present)\n name: string;\n description: string;\n personality: string;\n scenario: string;\n firstMes: string;\n mesExample: string;\n\n // Optional prompts\n systemPrompt?: string;\n postHistoryInstructions?: string;\n\n // Arrays\n alternateGreetings: string[];\n groupOnlyGreetings: string[];\n tags: string[];\n\n // Metadata\n creator?: string;\n creatorNotes?: string;\n characterVersion?: string;\n\n // Character book (v3 format)\n characterBook?: CCv3CharacterBook;\n\n // Extensions (preserved as-is)\n extensions: Record<string, unknown>;\n}\n\n/**\n * Create empty normalized card with defaults\n */\nexport function createEmptyNormalizedCard(): NormalizedCard {\n return {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n firstMes: '',\n mesExample: '',\n alternateGreetings: [],\n groupOnlyGreetings: [],\n tags: [],\n extensions: {},\n };\n}\n\n/**\n * Derived features extracted from card (not stored in card)\n */\nexport interface DerivedFeatures {\n // Content flags\n hasAlternateGreetings: boolean;\n alternateGreetingsCount: number;\n /** Total greetings = first_mes (1) + alternate_greetings */\n totalGreetingsCount: number;\n hasLorebook: boolean;\n lorebookEntriesCount: number;\n hasEmbeddedImages: boolean;\n embeddedImagesCount: number;\n hasGallery: boolean;\n\n // Format-specific\n hasRisuExtensions: boolean;\n hasRisuScripts: boolean;\n hasDepthPrompt: boolean;\n hasVoxtaAppearance: boolean;\n\n // Token counts (estimated)\n tokens: {\n description: number;\n personality: number;\n scenario: number;\n firstMes: number;\n mesExample: number;\n systemPrompt: number;\n total: number;\n };\n}\n\n/**\n * Create empty derived features\n */\nexport function createEmptyFeatures(): DerivedFeatures {\n return {\n hasAlternateGreetings: false,\n alternateGreetingsCount: 0,\n totalGreetingsCount: 1, // first_mes always counts as 1\n hasLorebook: false,\n lorebookEntriesCount: 0,\n hasEmbeddedImages: false,\n embeddedImagesCount: 0,\n hasGallery: false,\n hasRisuExtensions: false,\n hasRisuScripts: false,\n hasDepthPrompt: false,\n hasVoxtaAppearance: false,\n tokens: {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n },\n };\n}\n","/**\n * Feature Derivation\n *\n * Canonical feature extraction from character cards.\n * Eliminates duplicate implementations across Archive, Federation, and Architect.\n */\n\nimport type { CCv2Data } from './ccv2.js';\nimport type { CCv3DataInner } from './ccv3.js';\nimport type { DerivedFeatures } from './normalized.js';\nimport { hasRisuExtensions, hasRisuScripts, hasDepthPrompt } from './risu.js';\n\n/**\n * Derive features from a character card (V2 or V3 format).\n *\n * This is the canonical implementation - all apps should use this\n * rather than implementing their own feature detection.\n *\n * @param card - Either CCv2Data or CCv3DataInner (unwrapped)\n * @returns DerivedFeatures with all feature flags populated\n *\n * @example\n * ```typescript\n * import { deriveFeatures, parseV3Card } from '@character-foundry/schemas';\n *\n * const card = parseV3Card(data);\n * const features = deriveFeatures(card.data);\n *\n * if (features.hasLorebook) {\n * console.log(`Found ${features.lorebookEntriesCount} lorebook entries`);\n * }\n * ```\n */\nexport function deriveFeatures(card: CCv2Data | CCv3DataInner): DerivedFeatures {\n // Detect format by checking for V3-specific field\n const isV3 = 'assets' in card;\n\n // Alternate greetings\n const altGreetings = card.alternate_greetings ?? [];\n const hasAlternateGreetings = altGreetings.length > 0;\n const alternateGreetingsCount = altGreetings.length;\n // Total = first_mes (1) + alternate_greetings\n const totalGreetingsCount = 1 + alternateGreetingsCount;\n\n // Lorebook\n const characterBook = card.character_book;\n const hasLorebook = !!characterBook && characterBook.entries.length > 0;\n const lorebookEntriesCount = characterBook?.entries.length ?? 0;\n\n // Assets (V3 only) - check for visual asset types\n const assets = isV3 ? (card as CCv3DataInner).assets ?? [] : [];\n const imageAssetTypes = ['icon', 'background', 'emotion', 'custom'];\n const imageAssets = assets.filter(\n (a) =>\n imageAssetTypes.includes(a.type) ||\n ['png', 'jpg', 'jpeg', 'webp', 'gif'].includes(a.ext.toLowerCase()),\n );\n const hasGallery = imageAssets.length > 0;\n\n // Embedded images - check for data URLs in text fields\n const embeddedImageCount = countEmbeddedImages(card);\n const hasEmbeddedImages = embeddedImageCount > 0;\n\n // Extensions\n const extensions = card.extensions ?? {};\n const hasRisu = hasRisuExtensions(extensions);\n const hasScripts = hasRisuScripts(extensions);\n const hasDepth = hasDepthPrompt(extensions);\n const hasVoxta = checkVoxtaAppearance(extensions);\n\n // Token counts - initialize to zero (actual counting happens in tokenizers package)\n const tokens = {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n };\n\n return {\n hasAlternateGreetings,\n alternateGreetingsCount,\n totalGreetingsCount,\n hasLorebook,\n lorebookEntriesCount,\n hasEmbeddedImages,\n embeddedImagesCount: embeddedImageCount,\n hasGallery,\n hasRisuExtensions: hasRisu,\n hasRisuScripts: hasScripts,\n hasDepthPrompt: hasDepth,\n hasVoxtaAppearance: hasVoxta,\n tokens,\n };\n}\n\n/**\n * Count embedded images (data URLs) in card text fields.\n * Looks for base64-encoded images in description, personality, scenario, etc.\n */\nfunction countEmbeddedImages(card: CCv2Data | CCv3DataInner): number {\n const textFields = [\n card.description,\n card.personality,\n card.scenario,\n card.first_mes,\n card.mes_example,\n card.creator_notes,\n card.system_prompt,\n card.post_history_instructions,\n ...(card.alternate_greetings ?? []),\n ].filter((field): field is string => typeof field === 'string');\n\n // Add group_only_greetings if V3\n if ('group_only_greetings' in card) {\n textFields.push(...(card.group_only_greetings ?? []));\n }\n\n let count = 0;\n const dataUrlPattern = /data:image\\/[^;]+;base64,/g;\n\n for (const text of textFields) {\n const matches = text.match(dataUrlPattern);\n if (matches) {\n count += matches.length;\n }\n }\n\n return count;\n}\n\n/**\n * Check if card has Voxta appearance data.\n * Voxta stores appearance in extensions.voxta.appearance\n */\nfunction checkVoxtaAppearance(extensions: Record<string, unknown>): boolean {\n if (!extensions.voxta) return false;\n const voxta = extensions.voxta as Record<string, unknown>;\n return !!voxta.appearance;\n}\n","/**\n * Format Detection\n *\n * Detect card specification version from JSON data.\n */\n\nimport type { Spec } from './common.js';\n\n/**\n * V3-only fields that indicate a V3 card\n */\nconst V3_ONLY_FIELDS = ['group_only_greetings', 'creation_date', 'modification_date', 'assets'] as const;\n\n/**\n * Result from detailed spec detection\n */\nexport interface SpecDetectionResult {\n /** Detected spec version */\n spec: Spec | null;\n /** Confidence level of detection */\n confidence: 'high' | 'medium' | 'low';\n /** What fields/values indicated this spec */\n indicators: string[];\n /** Anomalies or inconsistencies detected */\n warnings: string[];\n}\n\n/**\n * Detect card spec version from parsed JSON\n * Returns 'v2', 'v3', or null if not recognized\n */\nexport function detectSpec(data: unknown): Spec | null {\n return detectSpecDetailed(data).spec;\n}\n\n/**\n * Detailed spec detection with confidence and reasoning.\n * Useful for debugging and logging.\n */\nexport function detectSpecDetailed(data: unknown): SpecDetectionResult {\n const result: SpecDetectionResult = {\n spec: null,\n confidence: 'low',\n indicators: [],\n warnings: [],\n };\n\n if (!data || typeof data !== 'object') {\n result.indicators.push('Input is not an object');\n return result;\n }\n\n const obj = data as Record<string, unknown>;\n const dataObj = (obj.data && typeof obj.data === 'object' ? obj.data : null) as Record<\n string,\n unknown\n > | null;\n\n // Check for explicit spec markers (HIGH confidence)\n\n // Explicit v3 spec marker\n if (obj.spec === 'chara_card_v3') {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v3\"');\n\n // Check for inconsistencies\n if (obj.spec_version && obj.spec_version !== '3.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v3 spec`);\n }\n\n return result;\n }\n\n // Explicit v2 spec marker\n if (obj.spec === 'chara_card_v2') {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v2\"');\n\n // Check for inconsistencies - V3-only fields in V2 card\n if (dataObj) {\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n result.warnings.push(`V3-only field \"${field}\" found in V2 card`);\n }\n }\n }\n\n if (obj.spec_version && obj.spec_version !== '2.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v2 spec`);\n }\n\n return result;\n }\n\n // Check spec_version field (HIGH confidence)\n if (typeof obj.spec_version === 'string') {\n if (obj.spec_version.startsWith('3')) {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"3\"`);\n return result;\n }\n if (obj.spec_version.startsWith('2')) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"2\"`);\n return result;\n }\n }\n\n if (obj.spec_version === 2.0 || obj.spec_version === 2) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version is numeric ${obj.spec_version}`);\n return result;\n }\n\n // Check for V3-only fields (MEDIUM confidence)\n if (dataObj) {\n const v3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n v3Fields.push(field);\n }\n }\n\n if (v3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields: ${v3Fields.join(', ')}`);\n return result;\n }\n }\n\n // Check root level for V3-only fields (also MEDIUM confidence)\n const rootV3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in obj) {\n rootV3Fields.push(field);\n }\n }\n if (rootV3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields at root: ${rootV3Fields.join(', ')}`);\n result.warnings.push('V3 fields found at root level instead of data object');\n return result;\n }\n\n // Wrapped format with data object (MEDIUM confidence)\n if (obj.spec && dataObj) {\n const dataName = dataObj.name;\n if (dataName && typeof dataName === 'string') {\n // Infer from spec string\n if (typeof obj.spec === 'string') {\n if (obj.spec.includes('v3') || obj.spec.includes('3')) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v3\" or \"3\"`);\n return result;\n }\n if (obj.spec.includes('v2') || obj.spec.includes('2')) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v2\" or \"2\"`);\n return result;\n }\n }\n // Default wrapped format to v3 (modern)\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push('Has wrapped format with spec and data.name');\n return result;\n }\n }\n\n // Unwrapped format - V1/V2 like structure (MEDIUM confidence)\n if (obj.name && typeof obj.name === 'string') {\n if ('description' in obj || 'personality' in obj || 'scenario' in obj) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push('Unwrapped format with name, description/personality/scenario');\n return result;\n }\n }\n\n // Check if data object has card-like structure without spec (LOW confidence)\n if (dataObj && typeof dataObj.name === 'string') {\n if ('description' in dataObj || 'personality' in dataObj) {\n result.spec = 'v2';\n result.confidence = 'low';\n result.indicators.push('Has data object with name and card fields, but no spec');\n result.warnings.push('Missing spec field');\n return result;\n }\n }\n\n result.indicators.push('No card structure detected');\n return result;\n}\n\n/**\n * Check if card has a lorebook\n */\nexport function hasLorebook(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Check wrapped format\n const wrapped = obj.data as Record<string, unknown> | undefined;\n if (wrapped?.character_book) {\n const book = wrapped.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Check if data looks like a valid card structure\n */\nexport function looksLikeCard(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Has explicit spec marker\n if (obj.spec === 'chara_card_v2' || obj.spec === 'chara_card_v3') {\n return true;\n }\n\n // Has wrapped data with name\n if (obj.data && typeof obj.data === 'object') {\n const dataObj = obj.data as Record<string, unknown>;\n if (typeof dataObj.name === 'string' && dataObj.name.length > 0) {\n return true;\n }\n }\n\n // Has unwrapped card-like structure\n if (typeof obj.name === 'string' && obj.name.length > 0) {\n if ('description' in obj || 'personality' in obj || 'first_mes' in obj) {\n return true;\n }\n }\n\n return false;\n}\n","/**\n * Card Normalizer\n *\n * Handles normalization of malformed card data from various sources.\n * Fixes common issues like wrong spec values, misplaced fields, missing required fields.\n */\n\nimport type { CCv2Data, CCv2Wrapped, CCv2CharacterBook, CCv2LorebookEntry } from './ccv2.js';\nimport type { CCv3Data, CCv3CharacterBook, CCv3LorebookEntry } from './ccv3.js';\nimport { detectSpec } from './detection.js';\n\n/**\n * Position values as numbers (non-standard) and their string equivalents\n */\nconst POSITION_MAP: Record<number, 'before_char' | 'after_char'> = {\n 0: 'before_char',\n 1: 'after_char',\n};\n\n/**\n * V3-only lorebook entry fields that should be moved to extensions for V2\n */\nconst V3_ONLY_ENTRY_FIELDS = [\n 'probability',\n 'depth',\n 'group',\n 'scan_frequency',\n 'use_regex',\n 'selective_logic',\n 'role',\n 'automation_id',\n] as const;\n\n/**\n * Required V2 card fields with their defaults\n */\nconst V2_REQUIRED_DEFAULTS: Partial<CCv2Data> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n};\n\n/**\n * Required V3 card fields with their defaults\n */\nconst V3_REQUIRED_DEFAULTS: Partial<CCv3Data['data']> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n creator: '',\n character_version: '1.0',\n tags: [],\n group_only_greetings: [],\n};\n\n/**\n * Fields that belong at root level for wrapped format\n */\nconst _ROOT_FIELDS = ['spec', 'spec_version', 'data'] as const;\n\n/**\n * Fields that belong in the data object\n */\nconst DATA_FIELDS = [\n 'name',\n 'description',\n 'personality',\n 'scenario',\n 'first_mes',\n 'mes_example',\n 'creator_notes',\n 'system_prompt',\n 'post_history_instructions',\n 'alternate_greetings',\n 'character_book',\n 'tags',\n 'creator',\n 'character_version',\n 'extensions',\n 'assets',\n 'nickname',\n 'creator_notes_multilingual',\n 'source',\n 'creation_date',\n 'modification_date',\n 'group_only_greetings',\n] as const;\n\n/**\n * Deep clone an object without mutating the original\n */\nfunction deepClone<T>(obj: T): T {\n if (obj === null || obj === undefined) {\n return obj;\n }\n if (Array.isArray(obj)) {\n return obj.map((item) => deepClone(item)) as T;\n }\n if (typeof obj === 'object') {\n const result: Record<string, unknown> = {};\n for (const [key, value] of Object.entries(obj as Record<string, unknown>)) {\n result[key] = deepClone(value);\n }\n return result as T;\n }\n return obj;\n}\n\n/**\n * Check if a timestamp is in milliseconds (13+ digits)\n */\nfunction isMilliseconds(timestamp: number): boolean {\n // Timestamps before year 2001 in seconds: < 1000000000\n // Timestamps in milliseconds are typically 13 digits: 1000000000000+\n return timestamp > 10000000000;\n}\n\n/**\n * CardNormalizer - handles normalization of malformed card data\n */\nexport const CardNormalizer = {\n /**\n * Normalize card data to valid schema format.\n *\n * Handles:\n * - Fixing spec/spec_version values\n * - Moving misplaced fields to correct locations\n * - Adding missing required fields with defaults\n * - Handling hybrid formats (fields at root AND in data object)\n *\n * @param data - Raw card data (potentially malformed)\n * @param spec - Target spec version\n * @returns Normalized card data (does not mutate input)\n */\n normalize(data: unknown, spec: 'v2' | 'v3'): CCv2Wrapped | CCv3Data {\n if (!data || typeof data !== 'object') {\n // Return minimal valid card\n if (spec === 'v3') {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: { ...V3_REQUIRED_DEFAULTS } as CCv3Data['data'],\n };\n }\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: { ...V2_REQUIRED_DEFAULTS } as CCv2Data,\n };\n }\n\n const obj = data as Record<string, unknown>;\n const result: Record<string, unknown> = {};\n\n // Build merged data object from root fields + existing data object\n const existingData = (obj.data && typeof obj.data === 'object' ? obj.data : {}) as Record<\n string,\n unknown\n >;\n const mergedData: Record<string, unknown> = {};\n\n // Copy existing data first\n for (const [key, value] of Object.entries(existingData)) {\n mergedData[key] = deepClone(value);\n }\n\n // Move any misplaced root-level data fields into data object\n // (ChubAI hybrid format fix)\n for (const field of DATA_FIELDS) {\n if (field in obj && !(field in mergedData)) {\n mergedData[field] = deepClone(obj[field]);\n }\n }\n\n // Handle character_book: null -> remove entirely\n if (mergedData.character_book === null) {\n delete mergedData.character_book;\n }\n\n // Normalize character_book if present\n if (mergedData.character_book && typeof mergedData.character_book === 'object') {\n mergedData.character_book = this.normalizeCharacterBook(\n mergedData.character_book as Record<string, unknown>,\n spec\n );\n }\n\n // Apply defaults for required fields\n const defaults = spec === 'v3' ? V3_REQUIRED_DEFAULTS : V2_REQUIRED_DEFAULTS;\n for (const [key, defaultValue] of Object.entries(defaults)) {\n if (!(key in mergedData) || mergedData[key] === undefined) {\n mergedData[key] = Array.isArray(defaultValue) ? [...defaultValue] : defaultValue;\n }\n }\n\n // Ensure arrays are actually arrays\n if (mergedData.tags && !Array.isArray(mergedData.tags)) {\n mergedData.tags = [];\n }\n if (mergedData.alternate_greetings && !Array.isArray(mergedData.alternate_greetings)) {\n mergedData.alternate_greetings = [];\n }\n if (spec === 'v3') {\n if (\n mergedData.group_only_greetings &&\n !Array.isArray(mergedData.group_only_greetings)\n ) {\n mergedData.group_only_greetings = [];\n }\n }\n\n // Build result with correct spec\n if (spec === 'v3') {\n result.spec = 'chara_card_v3';\n result.spec_version = '3.0';\n result.data = this.fixTimestampsInner(mergedData);\n } else {\n result.spec = 'chara_card_v2';\n result.spec_version = '2.0';\n result.data = mergedData;\n }\n\n return result as unknown as CCv2Wrapped | CCv3Data;\n },\n\n /**\n * Normalize a character book (lorebook).\n *\n * Handles:\n * - Ensuring required fields exist\n * - Converting numeric position values to string enums\n * - Moving V3-only fields to extensions for V2 compatibility\n *\n * @param book - Raw character book data\n * @param spec - Target spec version\n * @returns Normalized character book\n */\n normalizeCharacterBook(\n book: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2CharacterBook | CCv3CharacterBook {\n const result: Record<string, unknown> = {};\n\n // Copy book-level fields\n if (book.name !== undefined) result.name = book.name;\n if (book.description !== undefined) result.description = book.description;\n if (book.scan_depth !== undefined) result.scan_depth = book.scan_depth;\n if (book.token_budget !== undefined) result.token_budget = book.token_budget;\n if (book.recursive_scanning !== undefined)\n result.recursive_scanning = book.recursive_scanning;\n if (book.extensions !== undefined) result.extensions = deepClone(book.extensions);\n\n // Normalize entries\n const entries = Array.isArray(book.entries) ? book.entries : [];\n result.entries = entries.map((entry) =>\n this.normalizeEntry(entry as Record<string, unknown>, spec)\n );\n\n return result as unknown as CCv2CharacterBook | CCv3CharacterBook;\n },\n\n /**\n * Normalize a single lorebook entry.\n *\n * Handles:\n * - Converting numeric position to string enum\n * - Moving V3-only fields to extensions for V2\n * - Ensuring required fields exist\n *\n * @param entry - Raw entry data\n * @param spec - Target spec version\n * @returns Normalized entry\n */\n normalizeEntry(\n entry: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2LorebookEntry | CCv3LorebookEntry {\n const result: Record<string, unknown> = {};\n\n // Required fields with defaults\n result.keys = Array.isArray(entry.keys) ? [...entry.keys] : [];\n result.content = typeof entry.content === 'string' ? entry.content : '';\n result.enabled = entry.enabled !== false; // default true\n result.insertion_order =\n typeof entry.insertion_order === 'number' ? entry.insertion_order : 0;\n\n // For V2, extensions is required\n if (spec === 'v2') {\n result.extensions =\n entry.extensions && typeof entry.extensions === 'object'\n ? deepClone(entry.extensions)\n : {};\n }\n\n // Optional fields\n if (entry.case_sensitive !== undefined) result.case_sensitive = entry.case_sensitive;\n if (entry.name !== undefined) result.name = entry.name;\n if (entry.priority !== undefined) result.priority = entry.priority;\n if (entry.id !== undefined) result.id = entry.id;\n if (entry.comment !== undefined) result.comment = entry.comment;\n if (entry.selective !== undefined) result.selective = entry.selective;\n if (entry.secondary_keys !== undefined) {\n result.secondary_keys = Array.isArray(entry.secondary_keys)\n ? [...entry.secondary_keys]\n : [];\n }\n if (entry.constant !== undefined) result.constant = entry.constant;\n\n // Position: convert numeric to string enum\n if (entry.position !== undefined) {\n if (typeof entry.position === 'number') {\n result.position = POSITION_MAP[entry.position] || 'before_char';\n } else if (entry.position === 'before_char' || entry.position === 'after_char') {\n result.position = entry.position;\n }\n }\n\n // Handle V3-only fields\n if (spec === 'v3') {\n // Copy V3 fields directly\n if (entry.extensions !== undefined) result.extensions = deepClone(entry.extensions);\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n result[field] = entry[field];\n }\n }\n } else {\n // V2: Move V3-only fields to extensions\n const ext = (result.extensions || {}) as Record<string, unknown>;\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n ext[field] = entry[field];\n }\n }\n result.extensions = ext;\n }\n\n return result as unknown as CCv2LorebookEntry | CCv3LorebookEntry;\n },\n\n /**\n * Fix CharacterTavern timestamp format (milliseconds -> seconds).\n *\n * CCv3 spec requires timestamps in seconds (Unix epoch).\n * CharacterTavern exports timestamps in milliseconds.\n *\n * @param data - V3 card data\n * @returns Card data with fixed timestamps (does not mutate input)\n */\n fixTimestamps(data: CCv3Data): CCv3Data {\n const result = deepClone(data);\n result.data = this.fixTimestampsInner(\n result.data as unknown as Record<string, unknown>\n ) as unknown as CCv3Data['data'];\n return result;\n },\n\n /**\n * Internal: fix timestamps in data object\n */\n fixTimestampsInner(data: Record<string, unknown>): Record<string, unknown> {\n const result = { ...data };\n\n if (typeof result.creation_date === 'number') {\n if (isMilliseconds(result.creation_date)) {\n result.creation_date = Math.floor(result.creation_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.creation_date as number) < 0) {\n delete result.creation_date;\n }\n }\n\n if (typeof result.modification_date === 'number') {\n if (isMilliseconds(result.modification_date)) {\n result.modification_date = Math.floor(result.modification_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.modification_date as number) < 0) {\n delete result.modification_date;\n }\n }\n\n return result;\n },\n\n /**\n * Auto-detect spec and normalize.\n *\n * @param data - Raw card data\n * @returns Normalized card data, or null if not a valid card\n */\n autoNormalize(data: unknown): CCv2Wrapped | CCv3Data | null {\n const spec = detectSpec(data);\n if (!spec) return null;\n\n // V1 cards get upgraded to V2\n const targetSpec = spec === 'v3' ? 'v3' : 'v2';\n return this.normalize(data, targetSpec);\n },\n};\n\nexport type { CCv2Wrapped, CCv3Data };\n","/**\n * Validation Utilities\n *\n * Helper functions for Zod validation with Foundry error integration.\n */\n\nimport { z } from 'zod';\n\n/**\n * Convert Zod error to human-readable message\n */\nexport function zodErrorToMessage(zodError: z.ZodError, context?: string): string {\n const messages = zodError.errors.map((err) => {\n const path = err.path.length > 0 ? `${err.path.join('.')}: ` : '';\n return `${path}${err.message}`;\n });\n\n const message = messages.join('; ');\n return context ? `${context} - ${message}` : message;\n}\n\n/**\n * Get the first error field from Zod error\n */\nexport function getFirstErrorField(zodError: z.ZodError): string | undefined {\n return zodError.errors[0]?.path[0]?.toString();\n}\n\n/**\n * Safe parse with detailed error information\n */\nexport function safeParse<T>(\n schema: z.ZodSchema<T>,\n data: unknown\n): { success: true; data: T } | { success: false; error: string; field?: string } {\n const result = schema.safeParse(data);\n\n if (result.success) {\n return { success: true, data: result.data };\n }\n\n return {\n success: false,\n error: zodErrorToMessage(result.error),\n field: getFirstErrorField(result.error),\n };\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACOA,oBAAwB;;;AGQjB,SAAS,aAAa,MAAkB,QAAwB;AACrE,UACG,KAAK,MAAM,KAAM,KACjB,KAAK,SAAS,CAAC,KAAM,KACrB,KAAK,SAAS,CAAC,KAAM,IACtB,KAAK,SAAS,CAAC,OACX;AACR;AAKO,SAAS,cAAc,MAAkB,OAAe,QAAsB;AACnF,OAAK,MAAM,IAAK,UAAU,KAAM;AAChC,OAAK,SAAS,CAAC,IAAK,UAAU,KAAM;AACpC,OAAK,SAAS,CAAC,IAAK,UAAU,IAAK;AACnC,OAAK,SAAS,CAAC,IAAI,QAAQ;AAC7B;AAoBO,SAAS,QAAQ,MAAkB,QAAoB,YAAY,GAAW;AACnF,QAAO,UAAS,IAAI,WAAW,KAAK,KAAK,SAAS,OAAO,QAAQ,KAAK;AACpE,aAAS,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;AACtC,UAAI,KAAK,IAAI,CAAC,MAAM,OAAO,CAAC,EAAG,UAAS;IAC1C;AACA,WAAO;EACT;AACA,SAAO;AACT;AAKO,SAAS,UAAU,QAAkC;AAC1D,QAAM,cAAc,OAAO,OAAO,CAAC,KAAK,QAAQ,MAAM,IAAI,QAAQ,CAAC;AACnE,QAAM,SAAS,IAAI,WAAW,WAAW;AACzC,MAAI,SAAS;AACb,aAAW,OAAO,QAAQ;AACxB,WAAO,IAAI,KAAK,MAAM;AACtB,cAAU,IAAI;EAChB;AACA,SAAO;AACT;AAKO,SAAS,MAAM,MAAkB,OAAe,KAA0B;AAC/E,SAAO,KAAK,SAAS,OAAO,GAAG;AACjC;AAYO,SAAS,WAAW,KAAyB;AAClD,SAAO,IAAI,YAAY,EAAE,OAAO,GAAG;AACrC;AAKO,SAAS,SAAS,MAA0B;AACjD,SAAO,IAAI,YAAY,EAAE,OAAO,IAAI;AACtC;AAKO,SAAS,WAAW,KAAyB;AAClD,QAAM,SAAS,IAAI,WAAW,IAAI,MAAM;AACxC,WAAS,IAAI,GAAG,IAAI,IAAI,QAAQ,KAAK;AACnC,WAAO,CAAC,IAAI,IAAI,WAAW,CAAC,IAAI;EAClC;AACA,SAAO;AACT;AAKO,SAAS,SAAS,MAA0B;AACjD,MAAI,SAAS;AACb,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,cAAU,OAAO,aAAa,KAAK,CAAC,CAAE;EACxC;AACA,SAAO;AACT;AAgBO,SAAS,MAAM,MAA0B;AAC9C,SAAO,IAAI,WAAW,IAAI;AAC5B;ACnIA,IAAM,SAAS,OAAO,YAAY,eAChC,QAAQ,YAAY,QACpB,QAAQ,SAAS,QAAQ;AAO3B,IAAM,yBAAyB,OAAO;AAQ/B,SAAS,OAAO,MAA0B;AAC/C,MAAI,QAAQ;AAEV,WAAO,OAAO,KAAK,IAAI,EAAE,SAAS,QAAQ;EAC5C;AAGA,MAAI,KAAK,SAAS,wBAAwB;AACxC,WAAO,cAAc,IAAI;EAC3B;AAGA,MAAI,SAAS;AACb,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,cAAU,OAAO,aAAa,KAAK,CAAC,CAAE;EACxC;AACA,SAAO,KAAK,MAAM;AACpB;AAKO,SAAS,OAAO,QAA4B;AACjD,MAAI,QAAQ;AAEV,WAAO,IAAI,WAAW,OAAO,KAAK,QAAQ,QAAQ,CAAC;EACrD;AAGA,QAAM,SAAS,KAAK,MAAM;AAC1B,QAAM,SAAS,IAAI,WAAW,OAAO,MAAM;AAC3C,WAAS,IAAI,GAAG,IAAI,OAAO,QAAQ,KAAK;AACtC,WAAO,CAAC,IAAI,OAAO,WAAW,CAAC;EACjC;AACA,SAAO;AACT;AA2CA,IAAM,oBAAoB,KAAK;AAexB,SAAS,cAAc,MAA0B;AACtD,MAAI,QAAQ;AAEV,WAAO,OAAO,KAAK,IAAI,EAAE,SAAS,QAAQ;EAC5C;AAGA,QAAM,SAAmB,CAAC;AAE1B,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK,mBAAmB;AACvD,UAAM,QAAQ,KAAK,SAAS,GAAG,KAAK,IAAI,IAAI,mBAAmB,KAAK,MAAM,CAAC;AAC3E,QAAI,SAAS;AACb,aAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;AACrC,gBAAU,OAAO,aAAa,MAAM,CAAC,CAAE;IACzC;AACA,WAAO,KAAK,MAAM;EACpB;AAEA,SAAO,KAAK,OAAO,KAAK,EAAE,CAAC;AAC7B;ACpIA,IAAM,uBAAuB,uBAAO,IAAI,sCAAsC;AAKvE,IAAM,eAAN,cAA2B,MAAM;EAItC,YAAY,SAAiC,MAAc;AACzD,UAAM,OAAO;AAD8B,SAAA,OAAA;AAE3C,SAAK,OAAO;AAEZ,QAAI,MAAM,mBAAmB;AAC3B,YAAM,kBAAkB,MAAM,KAAK,WAAW;IAChD;EACF;;EATA,CAAU,oBAAoB,IAAI;AAUpC;AAKO,IAAM,aAAN,cAAyB,aAAa;EAC3C,YAAY,SAAiC,QAAiB;AAC5D,UAAM,SAAS,aAAa;AADe,SAAA,SAAA;AAE3C,SAAK,OAAO;EACd;AACF;AAsCO,IAAM,iBAAN,cAA6B,aAAa;EAC/C,YACkB,YACA,SAChB,SACA;AACA,UAAM,YAAY,aAAa,OAAO,MAAM,QAAQ,CAAC;AACrD,UAAM,SAAS,UAAU,OAAO,MAAM,QAAQ,CAAC;AAC/C,UAAM,MAAM,UACR,GAAG,OAAO,UAAU,QAAQ,oBAAoB,KAAK,OACrD,QAAQ,QAAQ,oBAAoB,KAAK;AAC7C,UAAM,KAAK,qBAAqB;AAThB,SAAA,aAAA;AACA,SAAA,UAAA;AAShB,SAAK,OAAO;EACd;AACF;;;AKjFA,iBAAkB;ACAlB,IAAAA,cAAkB;ACAlB,IAAAA,cAAkB;AMAlB,IAAAA,cAAkB;ARSX,IAAM,gBAAgB,aAAE,OAAO,EAAE,SAAS;AAK1C,IAAM,aAAa,aAAE,OAAO,EAAE,KAAK;AAKnC,IAAM,aAAa,aAAE,KAAK,CAAC,MAAM,IAAI,CAAC;AAKtC,IAAM,qBAAqB,aAAE,KAAK;EACvC;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;AACF,CAAC;AAKM,IAAM,sBAAsB,aAAE,KAAK,CAAC,WAAW,aAAa,QAAQ,CAAC;AAKrE,IAAM,kBAAkB,aAAE,KAAK;EACpC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;AACF,CAAC;AAKM,IAAM,wBAAwB,aAAE,OAAO;EAC5C,MAAM;EACN,KAAK,aAAE,OAAO;EACd,MAAM,aAAE,OAAO;EACf,KAAK,aAAE,OAAO;AAChB,CAAC;AAKM,IAAM,uBAAuB,aAAE,OAAO;EAC3C,YAAY;EACZ,MAAM,aAAE,WAAW,UAAU;EAC7B,UAAU,aAAE,OAAO;AACrB,CAAC;AC9DM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,YAAY,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;AAClH,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;;EAErC,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACnC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC7B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACvC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;AAC7C,CAAC;AAKM,IAAM,oBAAoBA,YAAAA,EAAE,OAAO;EACxC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AC5DM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,YAAY,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAChH,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,MAAMA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,UAAU,QAAQ,WAAW,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC/F,OAAOA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC3B,gBAAgBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACxD,aAAaA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,GAAG,EAAE,SAAS;;EACjD,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EAChC,OAAOA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EAC/C,iBAAiBA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,OAAO,KAAK,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS;AAChF,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AASM,IAAM,sBAAsBA,YAAAA,EAAE,OAAO;;EAE1C,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,SAASA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC9B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EACxC,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;EACpC,sBAAsBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;;EAEpD,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,QAAQA,YAAAA,EAAE,MAAM,qBAAqB,EAAE,SAAS;EAChD,UAAUA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC9B,4BAA4BA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAC1D,QAAQA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACrC,eAAeA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;EACvD,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;AAC7D,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;EACrC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AIxFD,IAAM,iBAAiB,CAAC,wBAAwB,iBAAiB,qBAAqB,QAAQ;AAoBvF,SAAS,WAAW,MAA4B;AACrD,SAAO,mBAAmB,IAAI,EAAE;AAClC;AAMO,SAAS,mBAAmB,MAAoC;AACrE,QAAM,SAA8B;IAClC,MAAM;IACN,YAAY;IACZ,YAAY,CAAC;IACb,UAAU,CAAC;EACb;AAEA,MAAI,CAAC,QAAQ,OAAO,SAAS,UAAU;AACrC,WAAO,WAAW,KAAK,wBAAwB;AAC/C,WAAO;EACT;AAEA,QAAM,MAAM;AACZ,QAAM,UAAW,IAAI,QAAQ,OAAO,IAAI,SAAS,WAAW,IAAI,OAAO;AAQvE,MAAI,IAAI,SAAS,iBAAiB;AAChC,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B;AAGtD,QAAI,IAAI,gBAAgB,IAAI,iBAAiB,OAAO;AAClD,aAAO,SAAS,KAAK,iBAAiB,IAAI,YAAY,6BAA6B;IACrF;AAEA,WAAO;EACT;AAGA,MAAI,IAAI,SAAS,iBAAiB;AAChC,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B;AAGtD,QAAI,SAAS;AACX,iBAAW,SAAS,gBAAgB;AAClC,YAAI,SAAS,SAAS;AACpB,iBAAO,SAAS,KAAK,kBAAkB,KAAK,oBAAoB;QAClE;MACF;IACF;AAEA,QAAI,IAAI,gBAAgB,IAAI,iBAAiB,OAAO;AAClD,aAAO,SAAS,KAAK,iBAAiB,IAAI,YAAY,6BAA6B;IACrF;AAEA,WAAO;EACT;AAGA,MAAI,OAAO,IAAI,iBAAiB,UAAU;AACxC,QAAI,IAAI,aAAa,WAAW,GAAG,GAAG;AACpC,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,iBAAiB,IAAI,YAAY,mBAAmB;AAC3E,aAAO;IACT;AACA,QAAI,IAAI,aAAa,WAAW,GAAG,GAAG;AACpC,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,iBAAiB,IAAI,YAAY,mBAAmB;AAC3E,aAAO;IACT;EACF;AAEA,MAAI,IAAI,iBAAiB,KAAO,IAAI,iBAAiB,GAAG;AACtD,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,2BAA2B,IAAI,YAAY,EAAE;AACpE,WAAO;EACT;AAGA,MAAI,SAAS;AACX,UAAM,WAAqB,CAAC;AAC5B,eAAW,SAAS,gBAAgB;AAClC,UAAI,SAAS,SAAS;AACpB,iBAAS,KAAK,KAAK;MACrB;IACF;AAEA,QAAI,SAAS,SAAS,GAAG;AACvB,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,uBAAuB,SAAS,KAAK,IAAI,CAAC,EAAE;AACnE,aAAO;IACT;EACF;AAGA,QAAM,eAAyB,CAAC;AAChC,aAAW,SAAS,gBAAgB;AAClC,QAAI,SAAS,KAAK;AAChB,mBAAa,KAAK,KAAK;IACzB;EACF;AACA,MAAI,aAAa,SAAS,GAAG;AAC3B,WAAO,OAAO;AACd,WAAO,aAAa;AACpB,WAAO,WAAW,KAAK,+BAA+B,aAAa,KAAK,IAAI,CAAC,EAAE;AAC/E,WAAO,SAAS,KAAK,sDAAsD;AAC3E,WAAO;EACT;AAGA,MAAI,IAAI,QAAQ,SAAS;AACvB,UAAM,WAAW,QAAQ;AACzB,QAAI,YAAY,OAAO,aAAa,UAAU;AAE5C,UAAI,OAAO,IAAI,SAAS,UAAU;AAChC,YAAI,IAAI,KAAK,SAAS,IAAI,KAAK,IAAI,KAAK,SAAS,GAAG,GAAG;AACrD,iBAAO,OAAO;AACd,iBAAO,aAAa;AACpB,iBAAO,WAAW,KAAK,eAAe,IAAI,IAAI,wBAAwB;AACtE,iBAAO;QACT;AACA,YAAI,IAAI,KAAK,SAAS,IAAI,KAAK,IAAI,KAAK,SAAS,GAAG,GAAG;AACrD,iBAAO,OAAO;AACd,iBAAO,aAAa;AACpB,iBAAO,WAAW,KAAK,eAAe,IAAI,IAAI,wBAAwB;AACtE,iBAAO;QACT;MACF;AAEA,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,4CAA4C;AACnE,aAAO;IACT;EACF;AAGA,MAAI,IAAI,QAAQ,OAAO,IAAI,SAAS,UAAU;AAC5C,QAAI,iBAAiB,OAAO,iBAAiB,OAAO,cAAc,KAAK;AACrE,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,8DAA8D;AACrF,aAAO;IACT;EACF;AAGA,MAAI,WAAW,OAAO,QAAQ,SAAS,UAAU;AAC/C,QAAI,iBAAiB,WAAW,iBAAiB,SAAS;AACxD,aAAO,OAAO;AACd,aAAO,aAAa;AACpB,aAAO,WAAW,KAAK,wDAAwD;AAC/E,aAAO,SAAS,KAAK,oBAAoB;AACzC,aAAO;IACT;EACF;AAEA,SAAO,WAAW,KAAK,4BAA4B;AACnD,SAAO;AACT;;;AhB3KO,IAAM,gBAAgB,IAAI,WAAW,CAAC,KAAM,IAAM,IAAM,IAAM,IAAM,IAAM,IAAM,EAAI,CAAC;AAKrF,IAAM,iBAAiB,KAAK,OAAO;AAMnC,IAAM,oBAAoB,KAAK,OAAO;AAW7C,SAAS,qBAAqB,YAAwB,UAAkB,mBAA+B;AACrG,QAAM,SAAuB,CAAC;AAC9B,MAAI,YAAY;AAChB,MAAI,QAAsB;AAE1B,QAAM,WAAW,IAAI,sBAAQ,CAAC,MAAkB,WAAoB;AAClE,QAAI,MAAO;AAEX,QAAI,QAAQ,KAAK,SAAS,GAAG;AAC3B,mBAAa,KAAK;AAClB,UAAI,YAAY,SAAS;AACvB,gBAAQ,IAAI,eAAe,WAAW,SAAS,qBAAqB;AACpE;MACF;AACA,aAAO,KAAK,IAAI;IAClB;EACF,CAAC;AAGD,MAAI;AACF,aAAS,KAAK,sBAAsB,aAAa,aAAa,IAAI,WAAW,UAAU,GAAG,IAAI;EAChG,SAAS,GAAG;AACV,UAAM,IAAI,WAAW,yBAAyB,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC,CAAC,IAAI,KAAK;EACnG;AAEA,MAAI,OAAO;AACT,UAAM;EACR;AAEA,SAAO,OAAO,GAAG,MAAM;AACzB;AAKO,IAAM,kBAAkB;;EAE7B;EACA;;EAEA;EACA;EACA;;EAEA;EACA;EACA;EACA;AACF;AAsBO,SAAS,MAAM,MAA2B;AAC/C,MAAI,KAAK,SAAS,EAAG,QAAO;AAC5B,WAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,QAAI,KAAK,CAAC,MAAM,cAAc,CAAC,EAAG,QAAO;EAC3C;AACA,SAAO;AACT;AAMO,SAAS,gBAAgB,MAA+B;AAC7D,QAAM,aAA0B,CAAC;AAGjC,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,WAAO;EACT;AAEA,MAAI,SAAS;AAEb,SAAO,SAAS,KAAK,QAAQ;AAE3B,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,SAAS,aAAa,MAAM,MAAM;AACxC,cAAU;AAGV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,YAAY,MAAM,MAAM,QAAQ,SAAS,CAAC;AAChD,UAAM,OAAO,SAAS,SAAS;AAC/B,cAAU;AAGV,QAAI,SAAS,gBAAgB;AAC3B,YAAM,IAAI,eAAe,QAAQ,gBAAgB,cAAc,IAAI,GAAG;IACxE;AAGA,QAAI,SAAS,SAAS,KAAK,OAAQ;AACnC,UAAM,YAAY,MAAM,MAAM,QAAQ,SAAS,MAAM;AACrD,cAAU;AAGV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,cAAU;AAGV,QAAI,SAAS,QAAQ;AACnB,YAAM,YAAY,QAAQ,WAAW,IAAI,WAAW,CAAC,CAAC,CAAC,CAAC;AACxD,UAAI,cAAc,IAAI;AACpB,cAAM,UAAU,SAAS,MAAM,WAAW,GAAG,SAAS,CAAC;AACvD,cAAM,OAAO,SAAS,MAAM,WAAW,YAAY,CAAC,CAAC;AACrD,mBAAW,KAAK,EAAE,SAAS,KAAK,CAAC;MACnC;IACF;AAGA,QAAI,SAAS,QAAQ;AACnB,YAAM,YAAY,QAAQ,WAAW,IAAI,WAAW,CAAC,CAAC,CAAC,CAAC;AACxD,UAAI,cAAc,IAAI;AACpB,cAAM,UAAU,SAAS,MAAM,WAAW,GAAG,SAAS,CAAC;AACvD,cAAM,oBAAoB,UAAU,YAAY,CAAC;AAEjD,YAAI,sBAAsB,GAAG;AAC3B,cAAI;AACF,kBAAM,iBAAiB,MAAM,WAAW,YAAY,CAAC;AAErD,kBAAM,eAAe,qBAAqB,gBAAgB,iBAAiB;AAC3E,kBAAM,OAAO,SAAS,YAAY;AAClC,uBAAW,KAAK,EAAE,SAAS,KAAK,CAAC;UACnC,SAAS,KAAK;AAEZ,gBAAI,eAAe,gBAAgB;AACjC,oBAAM;YACR;UAEF;QACF;MACF;IACF;AAGA,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAO;AACT;AAKO,SAAS,WAAW,MAA2E;AACpG,QAAM,SAAkE,CAAC;AAEzE,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,WAAO;EACT;AAEA,MAAI,SAAS;AAEb,SAAO,SAAS,KAAK,QAAQ;AAC3B,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,SAAS,aAAa,MAAM,MAAM;AACxC,UAAM,aAAa;AACnB,cAAU;AAEV,QAAI,SAAS,IAAI,KAAK,OAAQ;AAC9B,UAAM,OAAO,SAAS,MAAM,MAAM,QAAQ,SAAS,CAAC,CAAC;AACrD,cAAU;AAEV,WAAO,KAAK,EAAE,MAAM,QAAQ,YAAY,OAAO,CAAC;AAGhD,cAAU,SAAS;AAEnB,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAO;AACT;AAKA,SAAS,cAAc,WAA4B;AAEjD,MAAI;AACF,WAAO,KAAK,MAAM,SAAS;EAC7B,QAAQ;AAEN,QAAI;AACF,YAAM,UAAU,SAAS,OAAa,SAAS,CAAC;AAChD,aAAO,KAAK,MAAM,OAAO;IAC3B,QAAQ;AACN,YAAM,IAAI,WAAW,yCAAyC,KAAK;IACrE;EACF;AACF;AAKA,SAAS,kBAAkB,MAAwB;AACjD,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,QAAM,MAAM;AAGZ,QAAM,OAAO,IAAI;AACjB,MAAI,MAAM,gBAAgB;AACxB,UAAM,OAAO,KAAK;AAClB,QAAI,MAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,QAAQ,SAAS,EAAG,QAAO;EACrE;AAGA,MAAI,IAAI,gBAAgB;AACtB,UAAM,OAAO,IAAI;AACjB,QAAI,MAAM,QAAQ,KAAK,OAAO,KAAK,KAAK,QAAQ,SAAS,EAAG,QAAO;EACrE;AAEA,SAAO;AACT;AAMO,SAAS,eAAe,MAAuC;AAEpE,MAAI,CAAC,MAAM,IAAI,GAAG;AAChB,UAAM,IAAI,WAAW,yBAAyB,KAAK;EACrD;AAGA,QAAM,aAAa,gBAAgB,IAAI;AAEvC,MAAI,WAAW,WAAW,GAAG;AAC3B,UAAM,IAAI,WAAW,+BAA+B,KAAK;EAC3D;AAGA,MAAI,iBAA6C;AAEjD,aAAW,OAAO,iBAAiB;AAEjC,UAAM,iBAAiB,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;AAE/D,eAAW,SAAS,gBAAgB;AAClC,UAAI;AACF,cAAM,OAAO,cAAc,MAAM,IAAI;AACrC,cAAM,OAAO,WAAW,IAAI;AAE5B,YAAI,SAAS,QAAQ,SAAS,MAAM;AAClC,gBAAM,SAA8B;YAClC,MAAM;YACN;YACA,aAAa,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;UACvD;AAGA,cAAI,kBAAkB,IAAI,GAAG;AAC3B,mBAAO;UACT;AAGA,cAAI,CAAC,gBAAgB;AACnB,6BAAiB;UACnB;QACF;AAGA,YAAI,CAAC,QAAQ,QAAQ,OAAO,SAAS,UAAU;AAC7C,gBAAM,MAAM;AACZ,cAAI,iBAAmE;AAEvE,cAAI,IAAI,SAAS,mBAAmB,IAAI,QAAS,IAAI,KAAiC,MAAM;AAC1F,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD,WAAW,IAAI,SAAS,mBAAmB,IAAI,QAAS,IAAI,KAAiC,MAAM;AACjG,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD,WAAW,IAAI,SAAS,IAAI,eAAe,IAAI,eAAe,IAAI,WAAW;AAC3E,6BAAiB,EAAE,MAAM,MAAkB,MAAM,KAAK;UACxD;AAEA,cAAI,gBAAgB;AAClB,kBAAM,aAAkC;cACtC,GAAG;cACH,aAAa,WAAW,OAAO,CAAA,MAAK,EAAE,YAAY,GAAG;YACvD;AACA,gBAAI,kBAAkB,IAAI,GAAG;AAC3B,qBAAO;YACT;AACA,gBAAI,CAAC,eAAgB,kBAAiB;UACxC;QACF;MACF,QAAQ;MAER;IACF;EACF;AAGA,MAAI,gBAAgB;AAClB,WAAO;EACT;AAEA,QAAM,IAAI,WAAW,6CAA6C,KAAK;AACzE;AExWA,IAAI,YAAgC;AAKpC,SAAS,cAA2B;AAClC,MAAI,UAAW,QAAO;AAEtB,cAAY,IAAI,YAAY,GAAG;AAC/B,WAAS,IAAI,GAAG,IAAI,KAAK,KAAK;AAC5B,QAAI,IAAI;AACR,aAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,UAAK,IAAI,IAAM,aAAc,MAAM,IAAO,MAAM;IAClD;AACA,cAAU,CAAC,IAAI;EACjB;AACA,SAAO;AACT;AAOO,SAAS,MAAM,MAA0B;AAC9C,QAAM,QAAQ,YAAY;AAC1B,MAAI,MAAM;AAEV,WAAS,IAAI,GAAG,IAAI,KAAK,QAAQ,KAAK;AACpC,UAAM,OAAO,MAAM,KAAK,CAAC,KAAM,GAAI,IAAM,QAAQ;EACnD;AAEA,UAAQ,MAAM,gBAAgB;AAChC;AAOO,SAAS,WAAW,MAA8B;AACvD,QAAM,MAAM,MAAM,IAAI;AACtB,QAAM,SAAS,MAAM,CAAC;AACtB,gBAAc,QAAQ,KAAK,CAAC;AAC5B,SAAO;AACT;ADNO,SAAS,oBAAoB,WAAmC;AACrE,MAAI,CAAC,MAAM,SAAS,GAAG;AACrB,UAAM,IAAIC,WAAW,yBAAyB,KAAK;EACrD;AAEA,QAAM,SAAuB,CAACC,MAAM,WAAW,GAAG,CAAC,CAAC;AACpD,MAAI,SAAS;AAEb,SAAO,SAAS,UAAU,QAAQ;AAEhC,QAAI,SAAS,IAAI,UAAU,OAAQ;AACnC,UAAM,SAASC,aAAa,WAAW,MAAM;AAC7C,UAAM,YAAYD,MAAM,WAAW,QAAQ,SAAS,CAAC;AACrD,cAAU;AAGV,QAAI,SAAS,IAAI,UAAU,OAAQ;AACnC,UAAM,OAAOE,SAASF,MAAM,WAAW,QAAQ,SAAS,CAAC,CAAC;AAC1D,UAAM,UAAUA,MAAM,WAAW,QAAQ,SAAS,CAAC;AACnD,cAAU;AAGV,QAAI,SAAS,SAAS,IAAI,UAAU,OAAQ;AAC5C,UAAM,UAAUA,MAAM,WAAW,QAAQ,SAAS,MAAM;AACxD,UAAM,SAASA,MAAM,WAAW,SAAS,QAAQ,SAAS,SAAS,CAAC;AACpE,cAAU,SAAS;AAGnB,QAAI,SAAS,UAAU,SAAS,QAAQ;AACtC;IACF;AAGA,WAAO,KAAK,WAAW,SAAS,SAAS,MAAM;AAG/C,QAAI,SAAS,OAAQ;EACvB;AAEA,SAAOG,OAAO,GAAG,MAAM;AACzB;AAKA,SAAS,eAAe,WAA+B;AAGrD,WAAS,IAAI,UAAU,SAAS,IAAI,KAAK,GAAG,KAAK;AAC/C,QACE,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM;IACrB,UAAU,IAAI,CAAC,MAAM,IACrB;AACA,aAAO;IACT;EACF;AACA,SAAO;AACT;AAKA,SAAS,gBAAgB,SAAiB,MAA0B;AAClE,QAAM,gBAAgB,WAAW,OAAO;AACxC,QAAM,aAAa,WAAW,IAAI;AAGlC,QAAM,YAAYA;IAChB;IACA,IAAI,WAAW,CAAC,CAAC,CAAC;;IAClB;EACF;AAGA,QAAM,YAAY,WAAW,MAAM;AAGnC,QAAM,MAAM,WAAWA,OAAO,WAAW,SAAS,CAAC;AAGnD,QAAM,eAAeC,MAAM,CAAC;AAC5BC,gBAAc,cAAc,UAAU,QAAQ,CAAC;AAG/C,SAAOF,OAAO,cAAc,WAAW,WAAW,GAAG;AACvD;AAKO,SAAS,gBAAgB,WAAuB,SAAiB,MAA0B;AAChG,QAAM,aAAa,eAAe,SAAS;AAE3C,MAAI,eAAe,IAAI;AACrB,UAAM,IAAIJ,WAAW,qCAAqC,KAAK;EACjE;AAEA,QAAM,YAAY,gBAAgB,SAAS,IAAI;AAG/C,QAAM,aAAaC,MAAM,WAAW,GAAG,UAAU;AACjD,QAAM,eAAeA,MAAM,WAAW,UAAU;AAEhD,SAAOG,OAAO,YAAY,WAAW,YAAY;AACnD;AAKO,SAAS,aACd,aACA,UACA,UAAwB,CAAC,GACb;AACZ,QAAM;IACJ,MAAM;IACN,SAAS;IACT,SAAS;EACX,IAAI;AAGJ,QAAM,WAAW,oBAAoB,WAAW;AAGhD,QAAM,OAAO,SACT,KAAK,UAAU,QAAQ,IACvB,KAAK,UAAU,UAAU,MAAM,CAAC;AAGpC,QAAM,OAAO,SAAS,OAAa,WAAW,IAAI,CAAC,IAAI;AAGvD,SAAO,gBAAgB,UAAU,KAAK,IAAI;AAC5C;AAKO,SAAS,gBACd,QACA,QACwC;AACxC,QAAM,SAAS,OAAO,UAAU,OAAO;AACvC,QAAM,WAAqB,CAAC;AAE5B,MAAI,SAAS,OAAO,KAAK;AACvB,WAAO;MACL,OAAO;MACP,UAAU,CAAC,aAAa,OAAO,QAAQ,CAAC,CAAC,wBAAwB,OAAO,GAAG,KAAK;IAClF;EACF;AAEA,MAAI,SAAS,OAAO,MAAM;AACxB,aAAS,KAAK,aAAa,OAAO,QAAQ,CAAC,CAAC,+BAA+B,OAAO,IAAI,KAAK;EAC7F;AAEA,SAAO,EAAE,OAAO,MAAM,SAAS;AACjC;","names":["import_zod","z","z","ParseError","slice","readUInt32BE","toLatin1","concat","alloc","writeUInt32BE"]}
|