@character-foundry/character-foundry 0.4.3-dev.1766103111 → 0.4.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/charx.cjs +52 -85
- package/dist/charx.cjs.map +1 -1
- package/dist/charx.d.cts +22 -22
- package/dist/charx.d.ts +22 -22
- package/dist/charx.js +52 -85
- package/dist/charx.js.map +1 -1
- package/dist/exporter.cjs +54 -104
- package/dist/exporter.cjs.map +1 -1
- package/dist/exporter.d.cts +19 -19
- package/dist/exporter.d.ts +19 -19
- package/dist/exporter.js +54 -104
- package/dist/exporter.js.map +1 -1
- package/dist/federation.cjs +36 -104
- package/dist/federation.cjs.map +1 -1
- package/dist/federation.d.cts +19 -54
- package/dist/federation.d.ts +19 -54
- package/dist/federation.js +36 -104
- package/dist/federation.js.map +1 -1
- package/dist/index.cjs +54 -104
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +29 -29
- package/dist/index.d.ts +29 -29
- package/dist/index.js +54 -104
- package/dist/index.js.map +1 -1
- package/dist/loader.cjs +31 -171
- package/dist/loader.cjs.map +1 -1
- package/dist/loader.d.cts +23 -37
- package/dist/loader.d.ts +23 -37
- package/dist/loader.js +31 -171
- package/dist/loader.js.map +1 -1
- package/dist/lorebook.d.cts +23 -23
- package/dist/lorebook.d.ts +23 -23
- package/dist/normalizer.cjs +18 -72
- package/dist/normalizer.cjs.map +1 -1
- package/dist/normalizer.d.cts +37 -37
- package/dist/normalizer.d.ts +37 -37
- package/dist/normalizer.js +18 -72
- package/dist/normalizer.js.map +1 -1
- package/dist/png.cjs +18 -72
- package/dist/png.cjs.map +1 -1
- package/dist/png.d.cts +25 -25
- package/dist/png.d.ts +25 -25
- package/dist/png.js +18 -72
- package/dist/png.js.map +1 -1
- package/dist/schemas.cjs +23 -80
- package/dist/schemas.cjs.map +1 -1
- package/dist/schemas.d.cts +67 -85
- package/dist/schemas.d.ts +67 -85
- package/dist/schemas.js +23 -80
- package/dist/schemas.js.map +1 -1
- package/dist/voxta.cjs +20 -91
- package/dist/voxta.cjs.map +1 -1
- package/dist/voxta.d.cts +23 -23
- package/dist/voxta.d.ts +23 -23
- package/dist/voxta.js +20 -91
- package/dist/voxta.js.map +1 -1
- package/package.json +5 -5
package/dist/normalizer.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/normalizer.ts","../../schemas/src/common.ts","../../schemas/src/ccv2.ts","../../schemas/src/ccv3.ts","../../schemas/src/risu.ts","../../schemas/src/normalized.ts","../../schemas/src/feature-deriver.ts","../../schemas/src/detection.ts","../../schemas/src/normalizer.ts","../../schemas/src/validation.ts","../../core/src/binary.ts","../../core/src/base64.ts","../../core/src/errors.ts","../../core/src/data-url.ts","../../core/src/uri.ts","../../core/src/image.ts","../../core/src/uuid.ts","../../normalizer/src/v2-to-v3.ts","../../normalizer/src/v3-to-v2.ts","../../normalizer/src/to-normalized.ts","../../normalizer/src/from-normalized.ts"],"sourcesContent":["export * from '@character-foundry/normalizer';\n","/**\n * Common Types\n *\n * Shared types used across all card formats.\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Preprocessing Utilities\n// ============================================================================\n\n/**\n * Preprocess timestamp values to Unix seconds.\n * Handles: ISO strings, numeric strings, milliseconds, and numbers.\n * Returns undefined for invalid/negative values (defensive).\n */\nexport function preprocessTimestamp(val: unknown): number | undefined {\n if (val === null || val === undefined) return undefined;\n\n let num: number;\n\n if (typeof val === 'number') {\n num = val;\n } else if (typeof val === 'string') {\n const trimmed = val.trim();\n if (!trimmed) return undefined;\n\n // Try parsing as number first (numeric string like \"1705314600\")\n const parsed = Number(trimmed);\n if (!isNaN(parsed)) {\n num = parsed;\n } else {\n // Try parsing as ISO date string\n const date = new Date(trimmed);\n if (isNaN(date.getTime())) return undefined;\n num = Math.floor(date.getTime() / 1000);\n }\n } else {\n return undefined;\n }\n\n // Convert milliseconds to seconds if needed (>10 billion = likely ms)\n if (num > 10_000_000_000) {\n num = Math.floor(num / 1000);\n }\n\n // Reject negative timestamps (e.g., .NET default dates)\n if (num < 0) return undefined;\n\n return num;\n}\n\n/**\n * Preprocess numeric values that may come as strings.\n * Returns undefined for invalid values.\n */\nexport function preprocessNumeric(val: unknown): number | undefined {\n if (val === null || val === undefined) return undefined;\n\n if (typeof val === 'number') {\n return isNaN(val) ? undefined : val;\n }\n\n if (typeof val === 'string') {\n const trimmed = val.trim();\n if (!trimmed) return undefined;\n const parsed = Number(trimmed);\n return isNaN(parsed) ? undefined : parsed;\n }\n\n return undefined;\n}\n\n/**\n * Known asset types for coercion\n */\nconst KNOWN_ASSET_TYPES = new Set([\n 'icon', 'background', 'emotion', 'user_icon',\n 'sound', 'video', 'custom', 'x-risu-asset',\n]);\n\n/**\n * Preprocess asset type - coerce unknown types to 'custom'.\n */\nexport function preprocessAssetType(val: unknown): string {\n if (typeof val !== 'string') return 'custom';\n return KNOWN_ASSET_TYPES.has(val) ? val : 'custom';\n}\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * ISO 8601 date string schema\n */\nexport const ISO8601Schema = z.string().datetime();\n\n/**\n * UUID string schema\n */\nexport const UUIDSchema = z.string().uuid();\n\n/**\n * Card specification version schema\n */\nexport const SpecSchema = z.enum(['v2', 'v3']);\n\n/**\n * Source format identifier schema\n */\nexport const SourceFormatSchema = z.enum([\n 'png_v2', // PNG with 'chara' chunk (v2)\n 'png_v3', // PNG with 'ccv3' chunk (v3)\n 'json_v2', // Raw JSON v2\n 'json_v3', // Raw JSON v3\n 'charx', // ZIP with card.json (v3 spec)\n 'charx_risu', // ZIP with card.json + module.risum\n 'charx_jpeg', // JPEG with appended ZIP (read-only)\n 'voxta', // VoxPkg format\n]);\n\n/**\n * Original JSON shape schema\n */\nexport const OriginalShapeSchema = z.enum(['wrapped', 'unwrapped', 'legacy']);\n\n/**\n * Asset type identifier schema.\n * Uses preprocessing to coerce unknown types to 'custom' for forward compatibility.\n */\nexport const AssetTypeSchema = z.preprocess(\n preprocessAssetType,\n z.enum([\n 'icon',\n 'background',\n 'emotion',\n 'user_icon',\n 'sound',\n 'video',\n 'custom',\n 'x-risu-asset',\n ])\n);\n\n/**\n * Asset descriptor schema (v3 spec)\n */\nexport const AssetDescriptorSchema = z.object({\n type: AssetTypeSchema,\n uri: z.string(),\n name: z.string(),\n ext: z.string(),\n});\n\n/**\n * Extracted asset with binary data schema\n */\nexport const ExtractedAssetSchema = z.object({\n descriptor: AssetDescriptorSchema,\n data: z.instanceof(Uint8Array),\n mimeType: z.string(),\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * ISO 8601 date string\n */\nexport type ISO8601 = z.infer<typeof ISO8601Schema>;\n\n/**\n * UUID string\n */\nexport type UUID = z.infer<typeof UUIDSchema>;\n\n/**\n * Card specification version\n */\nexport type Spec = z.infer<typeof SpecSchema>;\n\n/**\n * Source format identifier\n */\nexport type SourceFormat = z.infer<typeof SourceFormatSchema>;\n\n/**\n * Original JSON shape\n */\nexport type OriginalShape = z.infer<typeof OriginalShapeSchema>;\n\n/**\n * Asset type identifier\n */\nexport type AssetType = z.infer<typeof AssetTypeSchema>;\n\n/**\n * Asset descriptor (v3 spec)\n */\nexport type AssetDescriptor = z.infer<typeof AssetDescriptorSchema>;\n\n/**\n * Extracted asset with binary data\n */\nexport type ExtractedAsset = z.infer<typeof ExtractedAssetSchema>;\n","/**\n * Character Card v2 Types\n *\n * Based on: https://github.com/malfoyslastname/character-card-spec-v2\n */\n\nimport { z } from 'zod';\nimport { preprocessNumeric } from './common.js';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v2 cards\n */\nexport const CCv2LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n extensions: z.record(z.unknown()).optional(),\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n}).passthrough(); // Allow SillyTavern extensions like depth, probability, etc.\n\n/**\n * Character book (lorebook) schema for v2 cards.\n * Uses preprocessing for numeric fields that often come as strings in wild data.\n */\nexport const CCv2CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.preprocess(preprocessNumeric, z.number().int().nonnegative().optional()),\n token_budget: z.preprocess(preprocessNumeric, z.number().int().nonnegative().optional()),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv2LorebookEntrySchema),\n});\n\n/**\n * Character Card v2 data structure schema\n */\nexport const CCv2DataSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv2CharacterBookSchema.optional().nullable(),\n tags: z.array(z.string()).optional(),\n creator: z.string().optional(),\n character_version: z.string().optional(),\n extensions: z.record(z.unknown()).optional(),\n});\n\n/**\n * Wrapped v2 card format schema (modern tools)\n */\nexport const CCv2WrappedSchema = z.object({\n spec: z.literal('chara_card_v2'),\n spec_version: z.literal('2.0'),\n data: CCv2DataSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v2 cards\n */\nexport type CCv2LorebookEntry = z.infer<typeof CCv2LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v2 cards\n */\nexport type CCv2CharacterBook = z.infer<typeof CCv2CharacterBookSchema>;\n\n/**\n * Character Card v2 data structure\n */\nexport type CCv2Data = z.infer<typeof CCv2DataSchema>;\n\n/**\n * Wrapped v2 card format (modern tools)\n */\nexport type CCv2Wrapped = z.infer<typeof CCv2WrappedSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a wrapped v2 card\n */\nexport function isWrappedV2(data: unknown): data is CCv2Wrapped {\n return CCv2WrappedSchema.safeParse(data).success;\n}\n\n/**\n * Check if data looks like v2 card data (wrapped or unwrapped)\n */\nexport function isV2CardData(data: unknown): data is CCv2Data | CCv2Wrapped {\n return (\n CCv2WrappedSchema.safeParse(data).success ||\n CCv2DataSchema.safeParse(data).success\n );\n}\n\n/**\n * Parse and validate a wrapped v2 card\n */\nexport function parseWrappedV2(data: unknown): CCv2Wrapped {\n return CCv2WrappedSchema.parse(data);\n}\n\n/**\n * Parse and validate v2 card data\n */\nexport function parseV2Data(data: unknown): CCv2Data {\n return CCv2DataSchema.parse(data);\n}\n\n/**\n * Check if data looks like a wrapped V2 card structurally (without strict validation).\n * This is more lenient than isWrappedV2 - it just checks structure, not full schema validity.\n */\nexport function looksLikeWrappedV2(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v2' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n\n/**\n * Get v2 card data from wrapped or unwrapped format.\n *\n * Uses structural check instead of strict Zod validation to handle\n * malformed cards that have the right structure but missing/invalid fields.\n * The caller (e.g., ccv2ToCCv3) handles defaulting missing fields.\n */\nexport function getV2Data(card: CCv2Data | CCv2Wrapped): CCv2Data {\n // Use structural check - more lenient than isWrappedV2 schema validation\n if (looksLikeWrappedV2(card)) {\n return card.data as CCv2Data;\n }\n return card;\n}\n","/**\n * Character Card v3 Types\n *\n * Based on: https://github.com/kwaroran/character-card-spec-v3\n */\n\nimport { z } from 'zod';\nimport { AssetDescriptorSchema, preprocessTimestamp, preprocessNumeric } from './common.js';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v3 cards\n */\nexport const CCv3LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific - also lenient with types since SillyTavern uses numbers for enums\n automation_id: z.string().optional(),\n role: z.union([z.enum(['system', 'user', 'assistant']), z.number().int()]).nullable().optional(),\n group: z.string().optional(),\n scan_frequency: z.number().int().nonnegative().optional(),\n probability: z.number().min(0).max(100).optional(), // Some tools use 0-100 instead of 0-1\n use_regex: z.boolean().optional(),\n depth: z.number().int().nonnegative().optional(),\n selective_logic: z.union([z.enum(['AND', 'NOT']), z.number().int()]).optional(),\n}).passthrough(); // Allow tool-specific extensions\n\n/**\n * Character book (lorebook) schema for v3 cards.\n * Uses preprocessing for numeric fields that often come as strings in wild data.\n */\nexport const CCv3CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.preprocess(preprocessNumeric, z.number().int().nonnegative().optional()),\n token_budget: z.preprocess(preprocessNumeric, z.number().int().nonnegative().optional()),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv3LorebookEntrySchema),\n});\n\n/**\n * Character Card v3 inner data structure schema.\n *\n * Note: Fields like group_only_greetings, creator, character_version, and tags\n * are technically \"required\" per V3 spec but rarely present in wild cards.\n * We use .default() to make parsing lenient while still producing valid output.\n */\nexport const CCv3DataInnerSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // \"Required\" per spec but often missing in wild - use defaults for leniency\n creator: z.string().default(''),\n character_version: z.string().default(''),\n tags: z.array(z.string()).default([]),\n group_only_greetings: z.array(z.string()).default([]),\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv3CharacterBookSchema.optional().nullable(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific\n assets: z.array(AssetDescriptorSchema).optional(),\n nickname: z.string().optional(),\n creator_notes_multilingual: z.record(z.string()).optional(),\n source: z.array(z.string()).optional(),\n // Unix timestamps - preprocess to handle ISO strings, numeric strings, milliseconds\n creation_date: z.preprocess(preprocessTimestamp, z.number().int().nonnegative().optional()),\n modification_date: z.preprocess(preprocessTimestamp, z.number().int().nonnegative().optional()),\n});\n\n/**\n * Character Card v3 full structure schema\n */\nexport const CCv3DataSchema = z.object({\n spec: z.literal('chara_card_v3'),\n spec_version: z.literal('3.0'),\n data: CCv3DataInnerSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v3 cards\n */\nexport type CCv3LorebookEntry = z.infer<typeof CCv3LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v3 cards\n */\nexport type CCv3CharacterBook = z.infer<typeof CCv3CharacterBookSchema>;\n\n/**\n * Character Card v3 inner data structure\n */\nexport type CCv3DataInner = z.infer<typeof CCv3DataInnerSchema>;\n\n/**\n * Character Card v3 full structure\n */\nexport type CCv3Data = z.infer<typeof CCv3DataSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a v3 card\n */\nexport function isV3Card(data: unknown): data is CCv3Data {\n return CCv3DataSchema.safeParse(data).success;\n}\n\n/**\n * Parse and validate a v3 card\n */\nexport function parseV3Card(data: unknown): CCv3Data {\n return CCv3DataSchema.parse(data);\n}\n\n/**\n * Parse and validate v3 card inner data\n */\nexport function parseV3DataInner(data: unknown): CCv3DataInner {\n return CCv3DataInnerSchema.parse(data);\n}\n\n/**\n * Get v3 card inner data\n */\nexport function getV3Data(card: CCv3Data): CCv3DataInner {\n return card.data;\n}\n\n/**\n * Check if data looks like a V3 card structurally (without strict validation).\n * More lenient than isV3Card - just checks structure, not full schema validity.\n */\nexport function looksLikeV3Card(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v3' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n","/**\n * RisuAI Extension Types\n *\n * These extensions are preserved as opaque blobs.\n * We do NOT interpret or transform the script contents.\n */\n\n/**\n * Risu emotions mapping (v2 style)\n * Format: [name, uri][]\n */\nexport type RisuEmotions = [string, string][];\n\n/**\n * Risu additional assets (v3 style)\n * Format: [name, uri, type][]\n */\nexport type RisuAdditionalAssets = [string, string, string][];\n\n/**\n * Risu depth prompt configuration\n */\nexport interface RisuDepthPrompt {\n depth: number;\n prompt: string;\n}\n\n/**\n * Risu extensions in card.extensions.risuai\n * Preserved as opaque - we don't interpret script contents\n */\nexport interface RisuExtensions {\n // Emotion assets\n emotions?: RisuEmotions;\n additionalAssets?: RisuAdditionalAssets;\n\n // Script data - OPAQUE, do not parse\n triggerscript?: unknown;\n customScripts?: unknown;\n\n // Voice/TTS settings\n vits?: Record<string, string>;\n\n // Depth prompt\n depth_prompt?: RisuDepthPrompt;\n\n // Other Risu-specific fields\n [key: string]: unknown;\n}\n\n/**\n * CharX x_meta entry (PNG chunk metadata preservation)\n */\nexport interface CharxMetaEntry {\n type?: string; // e.g., 'WEBP', 'PNG', 'JPEG'\n [key: string]: unknown;\n}\n\n/**\n * Check if card has Risu extensions\n */\nexport function hasRisuExtensions(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n return 'risuai' in extensions || 'risu' in extensions;\n}\n\n/**\n * Check if card has Risu scripts (triggerscript or customScripts)\n */\nexport function hasRisuScripts(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n const risu = extensions.risuai as RisuExtensions | undefined;\n if (!risu) return false;\n return !!risu.triggerscript || !!risu.customScripts;\n}\n\n/**\n * Check if card has depth prompt\n * Checks both SillyTavern style (extensions.depth_prompt) and Risu style (extensions.risuai.depth_prompt)\n */\nexport function hasDepthPrompt(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n // SillyTavern top-level depth_prompt\n if ('depth_prompt' in extensions && extensions.depth_prompt) return true;\n // Risu-style depth_prompt\n const risu = extensions.risuai as RisuExtensions | undefined;\n return !!risu?.depth_prompt;\n}\n","/**\n * Normalized Card Types\n *\n * Unified view of card data regardless of source format.\n * This is a computed/virtual representation, not stored.\n */\n\nimport type { CCv3CharacterBook } from './ccv3.js';\n\n/**\n * Normalized card representation\n * Provides unified access to card data from any format\n */\nexport interface NormalizedCard {\n // Core fields (always present)\n name: string;\n description: string;\n personality: string;\n scenario: string;\n firstMes: string;\n mesExample: string;\n\n // Optional prompts\n systemPrompt?: string;\n postHistoryInstructions?: string;\n\n // Arrays\n alternateGreetings: string[];\n groupOnlyGreetings: string[];\n tags: string[];\n\n // Metadata\n creator?: string;\n creatorNotes?: string;\n characterVersion?: string;\n\n // Character book (v3 format)\n characterBook?: CCv3CharacterBook;\n\n // Extensions (preserved as-is)\n extensions: Record<string, unknown>;\n}\n\n/**\n * Create empty normalized card with defaults\n */\nexport function createEmptyNormalizedCard(): NormalizedCard {\n return {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n firstMes: '',\n mesExample: '',\n alternateGreetings: [],\n groupOnlyGreetings: [],\n tags: [],\n extensions: {},\n };\n}\n\n/**\n * Derived features extracted from card (not stored in card)\n */\nexport interface DerivedFeatures {\n // Content flags\n hasAlternateGreetings: boolean;\n alternateGreetingsCount: number;\n /** Total greetings = first_mes (1) + alternate_greetings */\n totalGreetingsCount: number;\n hasLorebook: boolean;\n lorebookEntriesCount: number;\n hasEmbeddedImages: boolean;\n embeddedImagesCount: number;\n hasGallery: boolean;\n\n // Format-specific\n hasRisuExtensions: boolean;\n hasRisuScripts: boolean;\n hasDepthPrompt: boolean;\n hasVoxtaAppearance: boolean;\n\n // Token counts (estimated)\n tokens: {\n description: number;\n personality: number;\n scenario: number;\n firstMes: number;\n mesExample: number;\n systemPrompt: number;\n total: number;\n };\n}\n\n/**\n * Create empty derived features\n */\nexport function createEmptyFeatures(): DerivedFeatures {\n return {\n hasAlternateGreetings: false,\n alternateGreetingsCount: 0,\n totalGreetingsCount: 1, // first_mes always counts as 1\n hasLorebook: false,\n lorebookEntriesCount: 0,\n hasEmbeddedImages: false,\n embeddedImagesCount: 0,\n hasGallery: false,\n hasRisuExtensions: false,\n hasRisuScripts: false,\n hasDepthPrompt: false,\n hasVoxtaAppearance: false,\n tokens: {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n },\n };\n}\n","/**\n * Feature Derivation\n *\n * Canonical feature extraction from character cards.\n * Eliminates duplicate implementations across Archive, Federation, and Architect.\n */\n\nimport type { CCv2Data } from './ccv2.js';\nimport type { CCv3DataInner } from './ccv3.js';\nimport type { DerivedFeatures } from './normalized.js';\nimport { hasRisuExtensions, hasRisuScripts, hasDepthPrompt } from './risu.js';\n\n/**\n * Derive features from a character card (V2 or V3 format).\n *\n * This is the canonical implementation - all apps should use this\n * rather than implementing their own feature detection.\n *\n * @param card - Either CCv2Data or CCv3DataInner (unwrapped)\n * @returns DerivedFeatures with all feature flags populated\n *\n * @example\n * ```typescript\n * import { deriveFeatures, parseV3Card } from '@character-foundry/schemas';\n *\n * const card = parseV3Card(data);\n * const features = deriveFeatures(card.data);\n *\n * if (features.hasLorebook) {\n * console.log(`Found ${features.lorebookEntriesCount} lorebook entries`);\n * }\n * ```\n */\nexport function deriveFeatures(card: CCv2Data | CCv3DataInner): DerivedFeatures {\n // Detect format by checking for V3-specific field\n const isV3 = 'assets' in card;\n\n // Alternate greetings\n const altGreetings = card.alternate_greetings ?? [];\n const hasAlternateGreetings = altGreetings.length > 0;\n const alternateGreetingsCount = altGreetings.length;\n // Total = first_mes (1) + alternate_greetings\n const totalGreetingsCount = 1 + alternateGreetingsCount;\n\n // Lorebook\n const characterBook = card.character_book;\n const hasLorebook = !!characterBook && characterBook.entries.length > 0;\n const lorebookEntriesCount = characterBook?.entries.length ?? 0;\n\n // Assets (V3 only) - check for visual asset types\n const assets = isV3 ? (card as CCv3DataInner).assets ?? [] : [];\n const imageAssetTypes = ['icon', 'background', 'emotion', 'custom'];\n const imageAssets = assets.filter(\n (a) =>\n imageAssetTypes.includes(a.type) ||\n ['png', 'jpg', 'jpeg', 'webp', 'gif'].includes(a.ext.toLowerCase()),\n );\n const hasGallery = imageAssets.length > 0;\n\n // Embedded images - check for data URLs in text fields\n const embeddedImageCount = countEmbeddedImages(card);\n const hasEmbeddedImages = embeddedImageCount > 0;\n\n // Extensions\n const extensions = card.extensions ?? {};\n const hasRisu = hasRisuExtensions(extensions);\n const hasScripts = hasRisuScripts(extensions);\n const hasDepth = hasDepthPrompt(extensions);\n const hasVoxta = checkVoxtaAppearance(extensions);\n\n // Token counts - initialize to zero (actual counting happens in tokenizers package)\n const tokens = {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n };\n\n return {\n hasAlternateGreetings,\n alternateGreetingsCount,\n totalGreetingsCount,\n hasLorebook,\n lorebookEntriesCount,\n hasEmbeddedImages,\n embeddedImagesCount: embeddedImageCount,\n hasGallery,\n hasRisuExtensions: hasRisu,\n hasRisuScripts: hasScripts,\n hasDepthPrompt: hasDepth,\n hasVoxtaAppearance: hasVoxta,\n tokens,\n };\n}\n\n/**\n * Count embedded images (data URLs) in card text fields.\n * Looks for base64-encoded images in description, personality, scenario, etc.\n */\nfunction countEmbeddedImages(card: CCv2Data | CCv3DataInner): number {\n const textFields = [\n card.description,\n card.personality,\n card.scenario,\n card.first_mes,\n card.mes_example,\n card.creator_notes,\n card.system_prompt,\n card.post_history_instructions,\n ...(card.alternate_greetings ?? []),\n ].filter((field): field is string => typeof field === 'string');\n\n // Add group_only_greetings if V3\n if ('group_only_greetings' in card) {\n textFields.push(...(card.group_only_greetings ?? []));\n }\n\n let count = 0;\n const dataUrlPattern = /data:image\\/[^;]+;base64,/g;\n\n for (const text of textFields) {\n const matches = text.match(dataUrlPattern);\n if (matches) {\n count += matches.length;\n }\n }\n\n return count;\n}\n\n/**\n * Check if card has Voxta appearance data.\n * Voxta stores appearance in extensions.voxta.appearance\n */\nfunction checkVoxtaAppearance(extensions: Record<string, unknown>): boolean {\n if (!extensions.voxta) return false;\n const voxta = extensions.voxta as Record<string, unknown>;\n return !!voxta.appearance;\n}\n","/**\n * Format Detection\n *\n * Detect card specification version from JSON data.\n */\n\nimport type { Spec } from './common.js';\n\n/**\n * V3-only fields that indicate a V3 card\n */\nconst V3_ONLY_FIELDS = ['group_only_greetings', 'creation_date', 'modification_date', 'assets'] as const;\n\n/**\n * Result from detailed spec detection\n */\nexport interface SpecDetectionResult {\n /** Detected spec version */\n spec: Spec | null;\n /** Confidence level of detection */\n confidence: 'high' | 'medium' | 'low';\n /** What fields/values indicated this spec */\n indicators: string[];\n /** Anomalies or inconsistencies detected */\n warnings: string[];\n}\n\n/**\n * Detect card spec version from parsed JSON\n * Returns 'v2', 'v3', or null if not recognized\n */\nexport function detectSpec(data: unknown): Spec | null {\n return detectSpecDetailed(data).spec;\n}\n\n/**\n * Detailed spec detection with confidence and reasoning.\n * Useful for debugging and logging.\n */\nexport function detectSpecDetailed(data: unknown): SpecDetectionResult {\n const result: SpecDetectionResult = {\n spec: null,\n confidence: 'low',\n indicators: [],\n warnings: [],\n };\n\n if (!data || typeof data !== 'object') {\n result.indicators.push('Input is not an object');\n return result;\n }\n\n const obj = data as Record<string, unknown>;\n const dataObj = (obj.data && typeof obj.data === 'object' ? obj.data : null) as Record<\n string,\n unknown\n > | null;\n\n // Check for explicit spec markers (HIGH confidence)\n\n // Explicit v3 spec marker\n if (obj.spec === 'chara_card_v3') {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v3\"');\n\n // Check for inconsistencies\n if (obj.spec_version && obj.spec_version !== '3.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v3 spec`);\n }\n\n return result;\n }\n\n // Explicit v2 spec marker\n if (obj.spec === 'chara_card_v2') {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v2\"');\n\n // Check for inconsistencies - V3-only fields in V2 card\n if (dataObj) {\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n result.warnings.push(`V3-only field \"${field}\" found in V2 card`);\n }\n }\n }\n\n if (obj.spec_version && obj.spec_version !== '2.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v2 spec`);\n }\n\n return result;\n }\n\n // Check spec_version field (HIGH confidence)\n if (typeof obj.spec_version === 'string') {\n if (obj.spec_version.startsWith('3')) {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"3\"`);\n return result;\n }\n if (obj.spec_version.startsWith('2')) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"2\"`);\n return result;\n }\n }\n\n if (obj.spec_version === 2.0 || obj.spec_version === 2) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version is numeric ${obj.spec_version}`);\n return result;\n }\n\n // Check for V3-only fields (MEDIUM confidence)\n if (dataObj) {\n const v3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n v3Fields.push(field);\n }\n }\n\n if (v3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields: ${v3Fields.join(', ')}`);\n return result;\n }\n }\n\n // Check root level for V3-only fields (also MEDIUM confidence)\n const rootV3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in obj) {\n rootV3Fields.push(field);\n }\n }\n if (rootV3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields at root: ${rootV3Fields.join(', ')}`);\n result.warnings.push('V3 fields found at root level instead of data object');\n return result;\n }\n\n // Wrapped format with data object (MEDIUM confidence)\n if (obj.spec && dataObj) {\n const dataName = dataObj.name;\n if (dataName && typeof dataName === 'string') {\n // Infer from spec string\n if (typeof obj.spec === 'string') {\n if (obj.spec.includes('v3') || obj.spec.includes('3')) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v3\" or \"3\"`);\n return result;\n }\n if (obj.spec.includes('v2') || obj.spec.includes('2')) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v2\" or \"2\"`);\n return result;\n }\n }\n // Default wrapped format to v3 (modern)\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push('Has wrapped format with spec and data.name');\n return result;\n }\n }\n\n // Unwrapped format - V1/V2 like structure (MEDIUM confidence)\n if (obj.name && typeof obj.name === 'string') {\n if ('description' in obj || 'personality' in obj || 'scenario' in obj) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push('Unwrapped format with name, description/personality/scenario');\n return result;\n }\n }\n\n // Check if data object has card-like structure without spec (LOW confidence)\n if (dataObj && typeof dataObj.name === 'string') {\n if ('description' in dataObj || 'personality' in dataObj) {\n result.spec = 'v2';\n result.confidence = 'low';\n result.indicators.push('Has data object with name and card fields, but no spec');\n result.warnings.push('Missing spec field');\n return result;\n }\n }\n\n result.indicators.push('No card structure detected');\n return result;\n}\n\n/**\n * Check if card has a lorebook\n */\nexport function hasLorebook(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Check wrapped format\n const wrapped = obj.data as Record<string, unknown> | undefined;\n if (wrapped?.character_book) {\n const book = wrapped.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Check if data looks like a valid card structure\n */\nexport function looksLikeCard(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Has explicit spec marker\n if (obj.spec === 'chara_card_v2' || obj.spec === 'chara_card_v3') {\n return true;\n }\n\n // Has wrapped data with name\n if (obj.data && typeof obj.data === 'object') {\n const dataObj = obj.data as Record<string, unknown>;\n if (typeof dataObj.name === 'string' && dataObj.name.length > 0) {\n return true;\n }\n }\n\n // Has unwrapped card-like structure\n if (typeof obj.name === 'string' && obj.name.length > 0) {\n if ('description' in obj || 'personality' in obj || 'first_mes' in obj) {\n return true;\n }\n }\n\n return false;\n}\n","/**\n * Card Normalizer\n *\n * Handles normalization of malformed card data from various sources.\n * Fixes common issues like wrong spec values, misplaced fields, missing required fields.\n */\n\nimport type { CCv2Data, CCv2Wrapped, CCv2CharacterBook, CCv2LorebookEntry } from './ccv2.js';\nimport type { CCv3Data, CCv3CharacterBook, CCv3LorebookEntry } from './ccv3.js';\nimport { detectSpec } from './detection.js';\n\n/**\n * Position values as numbers (non-standard) and their string equivalents\n */\nconst POSITION_MAP: Record<number, 'before_char' | 'after_char'> = {\n 0: 'before_char',\n 1: 'after_char',\n};\n\n/**\n * V3-only lorebook entry fields that should be moved to extensions for V2\n */\nconst V3_ONLY_ENTRY_FIELDS = [\n 'probability',\n 'depth',\n 'group',\n 'scan_frequency',\n 'use_regex',\n 'selective_logic',\n 'role',\n 'automation_id',\n] as const;\n\n/**\n * Required V2 card fields with their defaults\n */\nconst V2_REQUIRED_DEFAULTS: Partial<CCv2Data> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n};\n\n/**\n * Required V3 card fields with their defaults\n */\nconst V3_REQUIRED_DEFAULTS: Partial<CCv3Data['data']> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n creator: '',\n character_version: '1.0',\n tags: [],\n group_only_greetings: [],\n};\n\n/**\n * Fields that belong at root level for wrapped format\n */\nconst _ROOT_FIELDS = ['spec', 'spec_version', 'data'] as const;\n\n/**\n * Fields that belong in the data object\n */\nconst DATA_FIELDS = [\n 'name',\n 'description',\n 'personality',\n 'scenario',\n 'first_mes',\n 'mes_example',\n 'creator_notes',\n 'system_prompt',\n 'post_history_instructions',\n 'alternate_greetings',\n 'character_book',\n 'tags',\n 'creator',\n 'character_version',\n 'extensions',\n 'assets',\n 'nickname',\n 'creator_notes_multilingual',\n 'source',\n 'creation_date',\n 'modification_date',\n 'group_only_greetings',\n] as const;\n\n/**\n * Deep clone an object without mutating the original\n */\nfunction deepClone<T>(obj: T): T {\n if (obj === null || obj === undefined) {\n return obj;\n }\n if (Array.isArray(obj)) {\n return obj.map((item) => deepClone(item)) as T;\n }\n if (typeof obj === 'object') {\n const result: Record<string, unknown> = {};\n for (const [key, value] of Object.entries(obj as Record<string, unknown>)) {\n result[key] = deepClone(value);\n }\n return result as T;\n }\n return obj;\n}\n\n/**\n * Check if a timestamp is in milliseconds (13+ digits)\n */\nfunction isMilliseconds(timestamp: number): boolean {\n // Timestamps before year 2001 in seconds: < 1000000000\n // Timestamps in milliseconds are typically 13 digits: 1000000000000+\n return timestamp > 10000000000;\n}\n\n/**\n * CardNormalizer - handles normalization of malformed card data\n */\nexport const CardNormalizer = {\n /**\n * Normalize card data to valid schema format.\n *\n * Handles:\n * - Fixing spec/spec_version values\n * - Moving misplaced fields to correct locations\n * - Adding missing required fields with defaults\n * - Handling hybrid formats (fields at root AND in data object)\n *\n * @param data - Raw card data (potentially malformed)\n * @param spec - Target spec version\n * @returns Normalized card data (does not mutate input)\n */\n normalize(data: unknown, spec: 'v2' | 'v3'): CCv2Wrapped | CCv3Data {\n if (!data || typeof data !== 'object') {\n // Return minimal valid card\n if (spec === 'v3') {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: { ...V3_REQUIRED_DEFAULTS } as CCv3Data['data'],\n };\n }\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: { ...V2_REQUIRED_DEFAULTS } as CCv2Data,\n };\n }\n\n const obj = data as Record<string, unknown>;\n const result: Record<string, unknown> = {};\n\n // Build merged data object from root fields + existing data object\n const existingData = (obj.data && typeof obj.data === 'object' ? obj.data : {}) as Record<\n string,\n unknown\n >;\n const mergedData: Record<string, unknown> = {};\n\n // Copy existing data first\n for (const [key, value] of Object.entries(existingData)) {\n mergedData[key] = deepClone(value);\n }\n\n // Move any misplaced root-level data fields into data object\n // (ChubAI hybrid format fix)\n for (const field of DATA_FIELDS) {\n if (field in obj && !(field in mergedData)) {\n mergedData[field] = deepClone(obj[field]);\n }\n }\n\n // Handle character_book: null -> remove entirely\n if (mergedData.character_book === null) {\n delete mergedData.character_book;\n }\n\n // Normalize character_book if present\n if (mergedData.character_book && typeof mergedData.character_book === 'object') {\n mergedData.character_book = CardNormalizer.normalizeCharacterBook(\n mergedData.character_book as Record<string, unknown>,\n spec\n );\n }\n\n // Apply defaults for required fields\n const defaults = spec === 'v3' ? V3_REQUIRED_DEFAULTS : V2_REQUIRED_DEFAULTS;\n for (const [key, defaultValue] of Object.entries(defaults)) {\n if (!(key in mergedData) || mergedData[key] === undefined) {\n mergedData[key] = Array.isArray(defaultValue) ? [...defaultValue] : defaultValue;\n }\n }\n\n // Ensure arrays are actually arrays\n if (mergedData.tags && !Array.isArray(mergedData.tags)) {\n mergedData.tags = [];\n }\n if (mergedData.alternate_greetings && !Array.isArray(mergedData.alternate_greetings)) {\n mergedData.alternate_greetings = [];\n }\n if (spec === 'v3') {\n if (\n mergedData.group_only_greetings &&\n !Array.isArray(mergedData.group_only_greetings)\n ) {\n mergedData.group_only_greetings = [];\n }\n }\n\n // Build result with correct spec\n if (spec === 'v3') {\n result.spec = 'chara_card_v3';\n result.spec_version = '3.0';\n result.data = CardNormalizer.fixTimestampsInner(mergedData);\n } else {\n result.spec = 'chara_card_v2';\n result.spec_version = '2.0';\n result.data = mergedData;\n }\n\n return result as unknown as CCv2Wrapped | CCv3Data;\n },\n\n /**\n * Normalize a character book (lorebook).\n *\n * Handles:\n * - Ensuring required fields exist\n * - Converting numeric position values to string enums\n * - Moving V3-only fields to extensions for V2 compatibility\n *\n * @param book - Raw character book data\n * @param spec - Target spec version\n * @returns Normalized character book\n */\n normalizeCharacterBook(\n book: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2CharacterBook | CCv3CharacterBook {\n const result: Record<string, unknown> = {};\n\n // Copy book-level fields\n if (book.name !== undefined) result.name = book.name;\n if (book.description !== undefined) result.description = book.description;\n if (book.scan_depth !== undefined) result.scan_depth = book.scan_depth;\n if (book.token_budget !== undefined) result.token_budget = book.token_budget;\n if (book.recursive_scanning !== undefined)\n result.recursive_scanning = book.recursive_scanning;\n if (book.extensions !== undefined) result.extensions = deepClone(book.extensions);\n\n // Normalize entries\n const entries = Array.isArray(book.entries) ? book.entries : [];\n result.entries = entries.map((entry) =>\n CardNormalizer.normalizeEntry(entry as Record<string, unknown>, spec)\n );\n\n return result as unknown as CCv2CharacterBook | CCv3CharacterBook;\n },\n\n /**\n * Normalize a single lorebook entry.\n *\n * Handles:\n * - Converting numeric position to string enum\n * - Moving V3-only fields to extensions for V2\n * - Ensuring required fields exist\n *\n * @param entry - Raw entry data\n * @param spec - Target spec version\n * @returns Normalized entry\n */\n normalizeEntry(\n entry: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2LorebookEntry | CCv3LorebookEntry {\n const result: Record<string, unknown> = {};\n\n // Required fields with defaults\n result.keys = Array.isArray(entry.keys) ? [...entry.keys] : [];\n result.content = typeof entry.content === 'string' ? entry.content : '';\n result.enabled = entry.enabled !== false; // default true\n result.insertion_order =\n typeof entry.insertion_order === 'number' ? entry.insertion_order : 0;\n\n // For V2, extensions is required\n if (spec === 'v2') {\n result.extensions =\n entry.extensions && typeof entry.extensions === 'object'\n ? deepClone(entry.extensions)\n : {};\n }\n\n // Optional fields\n if (entry.case_sensitive !== undefined) result.case_sensitive = entry.case_sensitive;\n if (entry.name !== undefined) result.name = entry.name;\n if (entry.priority !== undefined) result.priority = entry.priority;\n if (entry.id !== undefined) result.id = entry.id;\n if (entry.comment !== undefined) result.comment = entry.comment;\n if (entry.selective !== undefined) result.selective = entry.selective;\n if (entry.secondary_keys !== undefined) {\n result.secondary_keys = Array.isArray(entry.secondary_keys)\n ? [...entry.secondary_keys]\n : [];\n }\n if (entry.constant !== undefined) result.constant = entry.constant;\n\n // Position: convert numeric to string enum\n if (entry.position !== undefined) {\n if (typeof entry.position === 'number') {\n result.position = POSITION_MAP[entry.position] || 'before_char';\n } else if (entry.position === 'before_char' || entry.position === 'after_char') {\n result.position = entry.position;\n }\n }\n\n // Handle V3-only fields\n if (spec === 'v3') {\n // Copy V3 fields directly\n if (entry.extensions !== undefined) result.extensions = deepClone(entry.extensions);\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n result[field] = entry[field];\n }\n }\n } else {\n // V2: Move V3-only fields to extensions\n const ext = (result.extensions || {}) as Record<string, unknown>;\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n ext[field] = entry[field];\n }\n }\n result.extensions = ext;\n }\n\n return result as unknown as CCv2LorebookEntry | CCv3LorebookEntry;\n },\n\n /**\n * Fix CharacterTavern timestamp format (milliseconds -> seconds).\n *\n * CCv3 spec requires timestamps in seconds (Unix epoch).\n * CharacterTavern exports timestamps in milliseconds.\n *\n * @param data - V3 card data\n * @returns Card data with fixed timestamps (does not mutate input)\n */\n fixTimestamps(data: CCv3Data): CCv3Data {\n const result = deepClone(data);\n result.data = CardNormalizer.fixTimestampsInner(\n result.data as unknown as Record<string, unknown>\n ) as unknown as CCv3Data['data'];\n return result;\n },\n\n /**\n * Internal: fix timestamps in data object\n */\n fixTimestampsInner(data: Record<string, unknown>): Record<string, unknown> {\n const result = { ...data };\n\n if (typeof result.creation_date === 'number') {\n if (isMilliseconds(result.creation_date)) {\n result.creation_date = Math.floor(result.creation_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.creation_date as number) < 0) {\n delete result.creation_date;\n }\n }\n\n if (typeof result.modification_date === 'number') {\n if (isMilliseconds(result.modification_date)) {\n result.modification_date = Math.floor(result.modification_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.modification_date as number) < 0) {\n delete result.modification_date;\n }\n }\n\n return result;\n },\n\n /**\n * Auto-detect spec and normalize.\n *\n * @param data - Raw card data\n * @returns Normalized card data, or null if not a valid card\n */\n autoNormalize(data: unknown): CCv2Wrapped | CCv3Data | null {\n const spec = detectSpec(data);\n if (!spec) return null;\n\n // V1 cards get upgraded to V2\n const targetSpec = spec === 'v3' ? 'v3' : 'v2';\n return CardNormalizer.normalize(data, targetSpec);\n },\n};\n\nexport type { CCv2Wrapped, CCv3Data };\n","/**\n * Validation Utilities\n *\n * Helper functions for Zod validation with Foundry error integration.\n */\n\nimport { z } from 'zod';\n\n/**\n * Convert Zod error to human-readable message\n */\nexport function zodErrorToMessage(zodError: z.ZodError, context?: string): string {\n const messages = zodError.errors.map((err) => {\n const path = err.path.length > 0 ? `${err.path.join('.')}: ` : '';\n return `${path}${err.message}`;\n });\n\n const message = messages.join('; ');\n return context ? `${context} - ${message}` : message;\n}\n\n/**\n * Get the first error field from Zod error\n */\nexport function getFirstErrorField(zodError: z.ZodError): string | undefined {\n return zodError.errors[0]?.path[0]?.toString();\n}\n\n/**\n * Safe parse with detailed error information\n */\nexport function safeParse<T>(\n schema: z.ZodSchema<T>,\n data: unknown\n): { success: true; data: T } | { success: false; error: string; field?: string } {\n const result = schema.safeParse(data);\n\n if (result.success) {\n return { success: true, data: result.data };\n }\n\n return {\n success: false,\n error: zodErrorToMessage(result.error),\n field: getFirstErrorField(result.error),\n };\n}\n","/**\n * Binary Data Utilities\n *\n * Universal binary data operations using Uint8Array.\n * Works in both Node.js and browser environments.\n */\n\n/**\n * Universal binary data type (works in both environments)\n */\nexport type BinaryData = Uint8Array;\n\n/**\n * Read a 32-bit big-endian unsigned integer\n */\nexport function readUInt32BE(data: BinaryData, offset: number): number {\n return (\n (data[offset]! << 24) |\n (data[offset + 1]! << 16) |\n (data[offset + 2]! << 8) |\n data[offset + 3]!\n ) >>> 0;\n}\n\n/**\n * Write a 32-bit big-endian unsigned integer\n */\nexport function writeUInt32BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 24) & 0xff;\n data[offset + 1] = (value >>> 16) & 0xff;\n data[offset + 2] = (value >>> 8) & 0xff;\n data[offset + 3] = value & 0xff;\n}\n\n/**\n * Read a 16-bit big-endian unsigned integer\n */\nexport function readUInt16BE(data: BinaryData, offset: number): number {\n return ((data[offset]! << 8) | data[offset + 1]!) >>> 0;\n}\n\n/**\n * Write a 16-bit big-endian unsigned integer\n */\nexport function writeUInt16BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 8) & 0xff;\n data[offset + 1] = value & 0xff;\n}\n\n/**\n * Find a byte sequence in binary data\n */\nexport function indexOf(data: BinaryData, search: BinaryData, fromIndex = 0): number {\n outer: for (let i = fromIndex; i <= data.length - search.length; i++) {\n for (let j = 0; j < search.length; j++) {\n if (data[i + j] !== search[j]) continue outer;\n }\n return i;\n }\n return -1;\n}\n\n/**\n * Concatenate multiple binary arrays\n */\nexport function concat(...arrays: BinaryData[]): BinaryData {\n const totalLength = arrays.reduce((sum, arr) => sum + arr.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const arr of arrays) {\n result.set(arr, offset);\n offset += arr.length;\n }\n return result;\n}\n\n/**\n * Slice binary data (returns a view, not a copy)\n */\nexport function slice(data: BinaryData, start: number, end?: number): BinaryData {\n return data.subarray(start, end);\n}\n\n/**\n * Copy a portion of binary data (returns a new array)\n */\nexport function copy(data: BinaryData, start: number, end?: number): BinaryData {\n return data.slice(start, end);\n}\n\n/**\n * Convert string to binary (UTF-8)\n */\nexport function fromString(str: string): BinaryData {\n return new TextEncoder().encode(str);\n}\n\n/**\n * Convert binary to string (UTF-8)\n */\nexport function toString(data: BinaryData): string {\n return new TextDecoder().decode(data);\n}\n\n/**\n * Convert string to binary (Latin1 - for PNG keywords and similar)\n */\nexport function fromLatin1(str: string): BinaryData {\n const result = new Uint8Array(str.length);\n for (let i = 0; i < str.length; i++) {\n result[i] = str.charCodeAt(i) & 0xff;\n }\n return result;\n}\n\n/**\n * Convert binary to string (Latin1)\n */\nexport function toLatin1(data: BinaryData): string {\n let result = '';\n for (let i = 0; i < data.length; i++) {\n result += String.fromCharCode(data[i]!);\n }\n return result;\n}\n\n/**\n * Compare two binary arrays for equality\n */\nexport function equals(a: BinaryData, b: BinaryData): boolean {\n if (a.length !== b.length) return false;\n for (let i = 0; i < a.length; i++) {\n if (a[i] !== b[i]) return false;\n }\n return true;\n}\n\n/**\n * Create a new Uint8Array filled with zeros\n */\nexport function alloc(size: number): BinaryData {\n return new Uint8Array(size);\n}\n\n/**\n * Create a Uint8Array from an array of numbers\n */\nexport function from(data: number[] | ArrayBuffer | BinaryData): BinaryData {\n if (data instanceof Uint8Array) {\n return data;\n }\n if (data instanceof ArrayBuffer) {\n return new Uint8Array(data);\n }\n return new Uint8Array(data);\n}\n\n/**\n * Check if value is a Uint8Array\n */\nexport function isBinaryData(value: unknown): value is BinaryData {\n return value instanceof Uint8Array;\n}\n\n/**\n * Convert Node.js Buffer to Uint8Array (no-op if already Uint8Array)\n * This provides compatibility when interfacing with Node.js code\n */\nexport function toUint8Array(data: BinaryData | Buffer): BinaryData {\n if (data instanceof Uint8Array) {\n // Buffer extends Uint8Array, but we want a plain Uint8Array\n // This ensures we get a proper Uint8Array in all cases\n if (Object.getPrototypeOf(data).constructor.name === 'Buffer') {\n return new Uint8Array(data.buffer, data.byteOffset, data.byteLength);\n }\n return data;\n }\n return new Uint8Array(data);\n}\n\n/**\n * Convert binary data to hex string\n */\nexport function toHex(data: BinaryData): string {\n return Array.from(data)\n .map(b => b.toString(16).padStart(2, '0'))\n .join('');\n}\n\n/**\n * Convert hex string to binary data\n */\nexport function fromHex(hex: string): BinaryData {\n const bytes = new Uint8Array(hex.length / 2);\n for (let i = 0; i < bytes.length; i++) {\n bytes[i] = parseInt(hex.substr(i * 2, 2), 16);\n }\n return bytes;\n}\n","/**\n * Universal Base64 Encoding/Decoding\n *\n * Works in both Node.js and browser environments.\n */\n\nimport type { BinaryData } from './binary.js';\n\n/**\n * Check if we're in a Node.js environment\n */\nconst isNode = typeof process !== 'undefined' &&\n process.versions != null &&\n process.versions.node != null;\n\n/**\n * Threshold for switching to chunked encoding in browsers (1MB)\n * Below this, simple string concatenation is fast enough.\n * Above this, quadratic string growth becomes a problem.\n */\nconst LARGE_BUFFER_THRESHOLD = 1024 * 1024;\n\n/**\n * Encode binary data to base64 string\n *\n * PERFORMANCE: For large buffers (>1MB) in browsers, this automatically\n * uses the chunked implementation to avoid quadratic string concatenation.\n */\nexport function encode(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: use chunked encoding for large buffers to avoid O(n²) string growth\n if (data.length > LARGE_BUFFER_THRESHOLD) {\n return encodeChunked(data);\n }\n\n // Small buffers: simple approach is fast enough\n let binary = '';\n for (let i = 0; i < data.length; i++) {\n binary += String.fromCharCode(data[i]!);\n }\n return btoa(binary);\n}\n\n/**\n * Decode base64 string to binary data\n */\nexport function decode(base64: string): BinaryData {\n if (isNode) {\n // Node.js: use Buffer\n return new Uint8Array(Buffer.from(base64, 'base64'));\n }\n\n // Browser: use atob\n const binary = atob(base64);\n const result = new Uint8Array(binary.length);\n for (let i = 0; i < binary.length; i++) {\n result[i] = binary.charCodeAt(i);\n }\n return result;\n}\n\n/**\n * Check if a string is valid base64\n */\nexport function isBase64(str: string): boolean {\n if (str.length === 0) return false;\n // Base64 regex: only valid base64 characters, length multiple of 4 (with padding)\n const base64Regex = /^[A-Za-z0-9+/]*={0,2}$/;\n return base64Regex.test(str) && str.length % 4 === 0;\n}\n\n/**\n * Encode binary data to URL-safe base64 string\n * Replaces + with -, / with _, and removes padding\n */\nexport function encodeUrlSafe(data: BinaryData): string {\n return encode(data)\n .replace(/\\+/g, '-')\n .replace(/\\//g, '_')\n .replace(/=+$/, '');\n}\n\n/**\n * Decode URL-safe base64 string to binary data\n */\nexport function decodeUrlSafe(base64: string): BinaryData {\n // Add back padding if needed\n let padded = base64\n .replace(/-/g, '+')\n .replace(/_/g, '/');\n\n while (padded.length % 4 !== 0) {\n padded += '=';\n }\n\n return decode(padded);\n}\n\n/**\n * Chunk size for encoding large buffers (64KB)\n * Prevents stack overflow when using String.fromCharCode with spread operator\n */\nconst ENCODE_CHUNK_SIZE = 64 * 1024;\n\n/**\n * Encode binary data to base64 string with chunking for large buffers.\n * Handles buffers >10MB without stack overflow.\n *\n * @param data - Binary data to encode\n * @returns Base64 encoded string\n *\n * @example\n * ```typescript\n * const largeBuffer = new Uint8Array(20 * 1024 * 1024); // 20MB\n * const base64 = encodeChunked(largeBuffer); // No stack overflow\n * ```\n */\nexport function encodeChunked(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: process in chunks to avoid stack overflow\n const chunks: string[] = [];\n\n for (let i = 0; i < data.length; i += ENCODE_CHUNK_SIZE) {\n const chunk = data.subarray(i, Math.min(i + ENCODE_CHUNK_SIZE, data.length));\n let binary = '';\n for (let j = 0; j < chunk.length; j++) {\n binary += String.fromCharCode(chunk[j]!);\n }\n chunks.push(binary);\n }\n\n return btoa(chunks.join(''));\n}\n","/**\n * Error Classes\n *\n * Specific error types for character card operations.\n * All errors extend FoundryError for consistent handling.\n */\n\n/** Symbol to identify FoundryError instances across ESM/CJS boundaries */\nconst FOUNDRY_ERROR_MARKER = Symbol.for('@character-foundry/core:FoundryError');\n\n/**\n * Base error class for all Character Foundry errors\n */\nexport class FoundryError extends Error {\n /** @internal Marker for cross-module identification */\n readonly [FOUNDRY_ERROR_MARKER] = true;\n\n constructor(message: string, public readonly code: string) {\n super(message);\n this.name = 'FoundryError';\n // Maintains proper stack trace in V8 environments\n if (Error.captureStackTrace) {\n Error.captureStackTrace(this, this.constructor);\n }\n }\n}\n\n/**\n * Error during card parsing\n */\nexport class ParseError extends FoundryError {\n constructor(message: string, public readonly format?: string) {\n super(message, 'PARSE_ERROR');\n this.name = 'ParseError';\n }\n}\n\n/**\n * Error during card validation\n */\nexport class ValidationError extends FoundryError {\n constructor(message: string, public readonly field?: string) {\n super(message, 'VALIDATION_ERROR');\n this.name = 'ValidationError';\n }\n}\n\n/**\n * Asset not found in card or archive\n */\nexport class AssetNotFoundError extends FoundryError {\n constructor(public readonly uri: string) {\n super(`Asset not found: ${uri}`, 'ASSET_NOT_FOUND');\n this.name = 'AssetNotFoundError';\n }\n}\n\n/**\n * Format not supported for operation\n */\nexport class FormatNotSupportedError extends FoundryError {\n constructor(public readonly format: string, operation?: string) {\n const msg = operation\n ? `Format '${format}' not supported for ${operation}`\n : `Format not supported: ${format}`;\n super(msg, 'FORMAT_NOT_SUPPORTED');\n this.name = 'FormatNotSupportedError';\n }\n}\n\n/**\n * File size exceeds limits\n */\nexport class SizeLimitError extends FoundryError {\n constructor(\n public readonly actualSize: number,\n public readonly maxSize: number,\n context?: string\n ) {\n const actualMB = (actualSize / 1024 / 1024).toFixed(2);\n const maxMB = (maxSize / 1024 / 1024).toFixed(2);\n const msg = context\n ? `${context}: Size ${actualMB}MB exceeds limit ${maxMB}MB`\n : `Size ${actualMB}MB exceeds limit ${maxMB}MB`;\n super(msg, 'SIZE_LIMIT_EXCEEDED');\n this.name = 'SizeLimitError';\n }\n}\n\n/**\n * Path traversal or unsafe path detected\n */\nexport class PathTraversalError extends FoundryError {\n constructor(public readonly path: string) {\n super(`Unsafe path detected: ${path}`, 'PATH_TRAVERSAL');\n this.name = 'PathTraversalError';\n }\n}\n\n/**\n * Export operation would lose data\n */\nexport class DataLossError extends FoundryError {\n constructor(\n public readonly lostFields: string[],\n public readonly targetFormat: string\n ) {\n const fields = lostFields.slice(0, 3).join(', ');\n const more = lostFields.length > 3 ? ` and ${lostFields.length - 3} more` : '';\n super(\n `Export to ${targetFormat} would lose: ${fields}${more}`,\n 'DATA_LOSS'\n );\n this.name = 'DataLossError';\n }\n}\n\n/**\n * Check if an error is a FoundryError\n *\n * Uses Symbol.for() marker instead of instanceof to handle dual ESM/CJS package loading.\n * In dual-package environments, instanceof can fail if the error comes from a different\n * module instance (e.g., ESM vs CJS version of the same package). Symbol.for() creates\n * a global symbol shared across all module instances.\n */\nexport function isFoundryError(error: unknown): error is FoundryError {\n return (\n error instanceof Error &&\n FOUNDRY_ERROR_MARKER in error &&\n (error as Record<symbol, unknown>)[FOUNDRY_ERROR_MARKER] === true\n );\n}\n\n/**\n * Wrap unknown errors in a FoundryError\n */\nexport function wrapError(error: unknown, context?: string): FoundryError {\n if (isFoundryError(error)) {\n return error;\n }\n\n const message = error instanceof Error\n ? error.message\n : String(error);\n\n return new FoundryError(\n context ? `${context}: ${message}` : message,\n 'UNKNOWN_ERROR'\n );\n}\n","/**\n * Data URL Utilities\n *\n * Convert between Uint8Array buffers and data URLs.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n */\n\nimport type { BinaryData } from './binary.js';\nimport { encodeChunked as base64Encode, decode as base64Decode } from './base64.js';\nimport { ValidationError } from './errors.js';\n\n/**\n * Convert Uint8Array to data URL.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n *\n * @param buffer - Binary data to encode\n * @param mimeType - MIME type for the data URL (e.g., 'image/png', 'application/octet-stream')\n * @returns Data URL string\n *\n * @example\n * ```typescript\n * const png = new Uint8Array([...]);\n * const dataUrl = toDataURL(png, 'image/png');\n * // => \"data:image/png;base64,iVBORw0KGgo...\"\n * ```\n */\nexport function toDataURL(buffer: BinaryData, mimeType: string): string {\n // Use chunked encoding to handle large buffers without stack overflow\n const base64 = base64Encode(buffer);\n return `data:${mimeType};base64,${base64}`;\n}\n\n/**\n * Parse a data URL back to buffer and MIME type.\n * Validates the data URL format before parsing.\n *\n * @param dataUrl - Data URL string to parse\n * @returns Object containing the decoded buffer and MIME type\n * @throws Error if the data URL format is invalid\n *\n * @example\n * ```typescript\n * const { buffer, mimeType } = fromDataURL('data:image/png;base64,iVBORw0KGgo...');\n * // buffer: Uint8Array\n * // mimeType: 'image/png'\n * ```\n */\nexport function fromDataURL(dataUrl: string): { buffer: Uint8Array; mimeType: string } {\n // Validate data URL format\n if (!dataUrl.startsWith('data:')) {\n throw new ValidationError('Invalid data URL: must start with \"data:\"', 'dataUrl');\n }\n\n const commaIndex = dataUrl.indexOf(',');\n if (commaIndex === -1) {\n throw new ValidationError('Invalid data URL: missing comma separator', 'dataUrl');\n }\n\n const header = dataUrl.slice(5, commaIndex); // Skip 'data:'\n const data = dataUrl.slice(commaIndex + 1);\n\n // Parse header: [<mediatype>][;base64]\n let mimeType = 'text/plain';\n let isBase64 = false;\n\n const parts = header.split(';');\n for (const part of parts) {\n if (part === 'base64') {\n isBase64 = true;\n } else if (part && !part.includes('=')) {\n // MIME type (not a parameter like charset=utf-8)\n mimeType = part;\n }\n }\n\n if (!isBase64) {\n // URL-encoded text data\n throw new ValidationError('Non-base64 data URLs are not supported', 'dataUrl');\n }\n\n const buffer = base64Decode(data);\n return { buffer, mimeType };\n}\n\n/**\n * Check if a string is a valid data URL\n *\n * @param str - String to check\n * @returns true if the string is a valid data URL format\n */\nexport function isDataURL(str: string): boolean {\n if (!str.startsWith('data:')) return false;\n const commaIndex = str.indexOf(',');\n if (commaIndex === -1) return false;\n const header = str.slice(5, commaIndex);\n return header.includes('base64');\n}\n","/**\n * URI Utilities\n *\n * Handles different asset URI schemes used in character cards.\n * Supports: embeded://, embedded://, ccdefault:, https://, http://,\n * data:, file://, __asset:, asset:, chara-ext-asset_\n */\n\nexport type URIScheme =\n | 'embeded' // embeded:// (CharX standard, note intentional typo)\n | 'ccdefault' // ccdefault:\n | 'https' // https://\n | 'http' // http://\n | 'data' // data:mime;base64,...\n | 'file' // file://\n | 'internal' // Internal asset ID (UUID/string)\n | 'pngchunk' // PNG chunk reference (__asset:, asset:, chara-ext-asset_)\n | 'unknown';\n\nexport interface ParsedURI {\n scheme: URIScheme;\n originalUri: string;\n normalizedUri: string; // Normalized form of the URI\n path?: string; // For embeded://, file://\n url?: string; // For http://, https://\n data?: string; // For data: URIs\n mimeType?: string; // For data: URIs\n encoding?: string; // For data: URIs (e.g., base64)\n chunkKey?: string; // For pngchunk - the key/index to look up\n chunkCandidates?: string[]; // For pngchunk - all possible chunk keys to search\n}\n\n/**\n * Normalize a URI to its canonical form\n * Handles common typos and variant formats\n */\nexport function normalizeURI(uri: string): string {\n const trimmed = uri.trim();\n\n // Fix embedded:// -> embeded:// (common typo, CharX spec uses single 'd')\n if (trimmed.startsWith('embedded://')) {\n return 'embeded://' + trimmed.substring('embedded://'.length);\n }\n\n // Normalize PNG chunk references to pngchunk: scheme\n if (trimmed.startsWith('__asset:')) {\n const id = trimmed.substring('__asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('asset:')) {\n const id = trimmed.substring('asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_:')) {\n const id = trimmed.substring('chara-ext-asset_:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_')) {\n const id = trimmed.substring('chara-ext-asset_'.length);\n return `pngchunk:${id}`;\n }\n\n return trimmed;\n}\n\n/**\n * Parse a URI and determine its scheme and components\n */\nexport function parseURI(uri: string): ParsedURI {\n const trimmed = uri.trim();\n const normalized = normalizeURI(trimmed);\n\n // PNG chunk references (__asset:, asset:, chara-ext-asset_, pngchunk:)\n if (\n trimmed.startsWith('__asset:') ||\n trimmed.startsWith('asset:') ||\n trimmed.startsWith('chara-ext-asset_') ||\n trimmed.startsWith('pngchunk:')\n ) {\n let assetId: string;\n if (trimmed.startsWith('__asset:')) {\n assetId = trimmed.substring('__asset:'.length);\n } else if (trimmed.startsWith('asset:')) {\n assetId = trimmed.substring('asset:'.length);\n } else if (trimmed.startsWith('chara-ext-asset_:')) {\n assetId = trimmed.substring('chara-ext-asset_:'.length);\n } else if (trimmed.startsWith('pngchunk:')) {\n assetId = trimmed.substring('pngchunk:'.length);\n } else {\n assetId = trimmed.substring('chara-ext-asset_'.length);\n }\n\n // Generate all possible chunk key variations for lookup\n const candidates = [\n assetId, // \"0\" or \"filename.png\"\n trimmed, // Original URI\n `asset:${assetId}`, // \"asset:0\"\n `__asset:${assetId}`, // \"__asset:0\"\n `__asset_${assetId}`, // \"__asset_0\"\n `chara-ext-asset_${assetId}`, // \"chara-ext-asset_0\"\n `chara-ext-asset_:${assetId}`, // \"chara-ext-asset_:0\"\n `pngchunk:${assetId}`, // \"pngchunk:0\"\n ];\n\n return {\n scheme: 'pngchunk',\n originalUri: uri,\n normalizedUri: normalized,\n chunkKey: assetId,\n chunkCandidates: candidates,\n };\n }\n\n // ccdefault: - use default asset\n if (trimmed === 'ccdefault:' || trimmed.startsWith('ccdefault:')) {\n return {\n scheme: 'ccdefault',\n originalUri: uri,\n normalizedUri: normalized,\n };\n }\n\n // embeded:// or embedded:// (normalize typo)\n if (trimmed.startsWith('embeded://') || trimmed.startsWith('embedded://')) {\n const path = trimmed.startsWith('embeded://')\n ? trimmed.substring('embeded://'.length)\n : trimmed.substring('embedded://'.length);\n return {\n scheme: 'embeded',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // https://\n if (trimmed.startsWith('https://')) {\n return {\n scheme: 'https',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // http://\n if (trimmed.startsWith('http://')) {\n return {\n scheme: 'http',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // data: URIs\n if (trimmed.startsWith('data:')) {\n const parsed = parseDataURI(trimmed);\n return {\n scheme: 'data',\n originalUri: uri,\n normalizedUri: normalized,\n ...parsed,\n };\n }\n\n // file://\n if (trimmed.startsWith('file://')) {\n const path = trimmed.substring('file://'.length);\n return {\n scheme: 'file',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // Internal asset ID (alphanumeric/UUID format)\n if (/^[a-zA-Z0-9_-]+$/.test(trimmed)) {\n return {\n scheme: 'internal',\n originalUri: uri,\n normalizedUri: normalized,\n path: trimmed,\n };\n }\n\n // Unknown scheme\n return {\n scheme: 'unknown',\n originalUri: uri,\n normalizedUri: normalized,\n };\n}\n\n/**\n * Parse a data URI into its components\n * Format: data:[<mediatype>][;base64],<data>\n */\nfunction parseDataURI(uri: string): { mimeType?: string; encoding?: string; data?: string } {\n const match = uri.match(/^data:([^;,]+)?(;base64)?,(.*)$/);\n\n if (!match) {\n return {};\n }\n\n return {\n mimeType: match[1] || 'text/plain',\n encoding: match[2] ? 'base64' : undefined,\n data: match[3],\n };\n}\n\n/**\n * Check if extension is an image format\n */\nexport function isImageExt(ext: string): boolean {\n const imageExts = ['png', 'jpg', 'jpeg', 'webp', 'gif', 'avif', 'bmp', 'svg'];\n return imageExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is an audio format\n */\nexport function isAudioExt(ext: string): boolean {\n const audioExts = ['mp3', 'wav', 'ogg', 'flac', 'm4a', 'aac'];\n return audioExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is a video format\n */\nexport function isVideoExt(ext: string): boolean {\n const videoExts = ['mp4', 'webm', 'avi', 'mov', 'mkv'];\n return videoExts.includes(ext.toLowerCase());\n}\n\n/** Safe MIME types for data: URIs that can be used in href/src */\nconst SAFE_DATA_URI_MIME_TYPES = new Set([\n // Images (safe for img src)\n 'image/png',\n 'image/jpeg',\n 'image/gif',\n 'image/webp',\n 'image/avif',\n 'image/bmp',\n 'image/x-icon',\n // Audio (safe for audio src)\n 'audio/mpeg',\n 'audio/wav',\n 'audio/ogg',\n 'audio/flac',\n 'audio/mp4',\n 'audio/aac',\n // Video (safe for video src)\n 'video/mp4',\n 'video/webm',\n // Text/data (generally safe)\n 'text/plain',\n 'application/json',\n 'application/octet-stream',\n]);\n\n/** Potentially dangerous MIME types that should NOT be used in href/src */\nconst DANGEROUS_DATA_URI_MIME_TYPES = new Set([\n // Executable/script content\n 'text/html',\n 'text/javascript',\n 'application/javascript',\n 'application/x-javascript',\n 'text/css',\n 'image/svg+xml', // SVG can contain scripts\n 'application/xhtml+xml',\n 'application/xml',\n]);\n\n/**\n * Options for URI safety validation\n */\nexport interface URISafetyOptions {\n /** Allow http:// URIs (default: false) */\n allowHttp?: boolean;\n /** Allow file:// URIs (default: false) */\n allowFile?: boolean;\n /**\n * Allowed MIME types for data: URIs (default: all safe types).\n * Set to empty array to reject all data: URIs.\n * Set to undefined to use default safe list.\n */\n allowedDataMimes?: string[];\n}\n\n/**\n * Result of URI safety check with detailed information\n */\nexport interface URISafetyResult {\n /** Whether the URI is safe to use */\n safe: boolean;\n /** Reason if unsafe */\n reason?: string;\n /** Detected scheme */\n scheme: URIScheme;\n /** MIME type for data: URIs */\n mimeType?: string;\n}\n\n/**\n * Validate if a URI is safe to use (detailed version)\n *\n * @param uri - URI to validate\n * @param options - Safety options\n * @returns Detailed safety result\n */\nexport function checkURISafety(uri: string, options: URISafetyOptions = {}): URISafetyResult {\n const parsed = parseURI(uri);\n\n switch (parsed.scheme) {\n case 'embeded':\n case 'ccdefault':\n case 'internal':\n case 'https':\n case 'pngchunk':\n return { safe: true, scheme: parsed.scheme };\n\n case 'data': {\n const mimeType = parsed.mimeType || 'text/plain';\n\n // Check for explicitly dangerous MIME types\n if (DANGEROUS_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI with potentially dangerous MIME type: ${mimeType}`,\n };\n }\n\n // If custom allowed list is provided, check against it\n if (options.allowedDataMimes !== undefined) {\n if (options.allowedDataMimes.length === 0) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: 'Data URIs are not allowed',\n };\n }\n if (!options.allowedDataMimes.includes(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI MIME type not in allowed list: ${mimeType}`,\n };\n }\n }\n\n // Otherwise use default safe list\n if (!SAFE_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Unknown data URI MIME type: ${mimeType}`,\n };\n }\n\n return { safe: true, scheme: parsed.scheme, mimeType };\n }\n\n case 'http':\n if (options.allowHttp === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'HTTP URIs are not allowed' };\n\n case 'file':\n if (options.allowFile === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'File URIs are not allowed' };\n\n case 'unknown':\n default:\n return { safe: false, scheme: parsed.scheme, reason: 'Unknown URI scheme' };\n }\n}\n\n/**\n * Validate if a URI is safe to use (simple boolean version for backwards compatibility)\n *\n * @deprecated Use checkURISafety() for detailed safety information\n */\nexport function isURISafe(uri: string, options: { allowHttp?: boolean; allowFile?: boolean } = {}): boolean {\n return checkURISafety(uri, options).safe;\n}\n\n/**\n * Extract file extension from URI\n */\nexport function getExtensionFromURI(uri: string): string {\n const parsed = parseURI(uri);\n\n if (parsed.path) {\n const parts = parsed.path.split('.');\n if (parts.length > 1) {\n return parts[parts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.url) {\n const urlParts = parsed.url.split('?')[0]!.split('.');\n if (urlParts.length > 1) {\n return urlParts[urlParts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.mimeType) {\n return getExtFromMimeType(parsed.mimeType);\n }\n\n return 'unknown';\n}\n\n/**\n * Get MIME type from file extension\n */\nexport function getMimeTypeFromExt(ext: string): string {\n const extToMime: Record<string, string> = {\n // Images\n 'png': 'image/png',\n 'jpg': 'image/jpeg',\n 'jpeg': 'image/jpeg',\n 'webp': 'image/webp',\n 'gif': 'image/gif',\n 'avif': 'image/avif',\n 'svg': 'image/svg+xml',\n 'bmp': 'image/bmp',\n 'ico': 'image/x-icon',\n\n // Audio\n 'mp3': 'audio/mpeg',\n 'wav': 'audio/wav',\n 'ogg': 'audio/ogg',\n 'flac': 'audio/flac',\n 'm4a': 'audio/mp4',\n 'aac': 'audio/aac',\n\n // Video\n 'mp4': 'video/mp4',\n 'webm': 'video/webm',\n 'avi': 'video/x-msvideo',\n 'mov': 'video/quicktime',\n 'mkv': 'video/x-matroska',\n\n // Text/Data\n 'json': 'application/json',\n 'txt': 'text/plain',\n 'html': 'text/html',\n 'css': 'text/css',\n 'js': 'application/javascript',\n };\n\n return extToMime[ext.toLowerCase()] || 'application/octet-stream';\n}\n\n/**\n * Get file extension from MIME type\n */\nexport function getExtFromMimeType(mimeType: string): string {\n const mimeToExt: Record<string, string> = {\n 'image/png': 'png',\n 'image/jpeg': 'jpg',\n 'image/webp': 'webp',\n 'image/gif': 'gif',\n 'image/avif': 'avif',\n 'image/svg+xml': 'svg',\n 'image/bmp': 'bmp',\n 'image/x-icon': 'ico',\n 'audio/mpeg': 'mp3',\n 'audio/wav': 'wav',\n 'audio/ogg': 'ogg',\n 'audio/flac': 'flac',\n 'audio/mp4': 'm4a',\n 'audio/aac': 'aac',\n 'video/mp4': 'mp4',\n 'video/webm': 'webm',\n 'video/x-msvideo': 'avi',\n 'video/quicktime': 'mov',\n 'video/x-matroska': 'mkv',\n 'application/json': 'json',\n 'text/plain': 'txt',\n 'text/html': 'html',\n 'text/css': 'css',\n 'application/javascript': 'js',\n };\n\n return mimeToExt[mimeType] || 'bin';\n}\n\n/**\n * Build a data URI from binary data and MIME type\n */\nexport function buildDataURI(data: string, mimeType: string, isBase64 = true): string {\n if (isBase64) {\n return `data:${mimeType};base64,${data}`;\n }\n return `data:${mimeType},${encodeURIComponent(data)}`;\n}\n","/**\n * Image Analysis Utilities\n *\n * Detect properties of image files from binary data.\n */\n\nimport {\n type BinaryData,\n indexOf,\n fromLatin1,\n} from './binary.js';\n\n/**\n * Check if an image buffer contains animation data.\n * Supports: APNG, WebP (Animated), GIF\n */\nexport function isAnimatedImage(data: BinaryData, _mimeType?: string): boolean {\n // 1. WebP Detection\n // RIFF .... WEBP\n if (\n data.length > 12 &&\n data[0] === 0x52 && data[1] === 0x49 && data[2] === 0x46 && data[3] === 0x46 && // RIFF\n data[8] === 0x57 && data[9] === 0x45 && data[10] === 0x42 && data[11] === 0x50 // WEBP\n ) {\n // Check for VP8X chunk\n // VP8X chunk header: 'VP8X' (bytes 12-15)\n if (\n data[12] === 0x56 && data[13] === 0x50 && data[14] === 0x38 && data[15] === 0x58\n ) {\n // Flags byte is at offset 20 (16 + 4 bytes chunk size)\n // Animation bit is bit 1 (0x02)\n const flags = data[20];\n return (flags! & 0x02) !== 0;\n }\n return false;\n }\n\n // 2. PNG/APNG Detection\n // Signature: 89 50 4E 47 0D 0A 1A 0A\n if (\n data.length > 8 &&\n data[0] === 0x89 && data[1] === 0x50 && data[2] === 0x4E && data[3] === 0x47\n ) {\n // Search for 'acTL' chunk (Animation Control)\n // It must appear before IDAT.\n // Simple search: indexOf('acTL')\n // Note: theoretically 'acTL' string could appear in other data, but highly unlikely in valid PNG structure before IDAT\n // We can iterate chunks to be safe, but indexOf is faster for a quick check\n const actlSig = fromLatin1('acTL');\n const idatSig = fromLatin1('IDAT');\n \n const actlIndex = indexOf(data, actlSig);\n if (actlIndex === -1) return false;\n\n const idatIndex = indexOf(data, idatSig);\n // If acTL exists and is before the first IDAT (or IDAT not found yet), it's APNG\n return idatIndex === -1 || actlIndex < idatIndex;\n }\n\n // 3. GIF Detection\n // Signature: GIF87a or GIF89a\n if (\n data.length > 6 &&\n data[0] === 0x47 && data[1] === 0x49 && data[2] === 0x46 // GIF\n ) {\n // Check for NETSCAPE2.0 extension (looping animation)\n // This is a heuristic. Static GIFs are rare in this domain but possible.\n // Full frame counting is expensive. Presence of NETSCAPE block is a strong indicator.\n const netscape = fromLatin1('NETSCAPE2.0');\n return indexOf(data, netscape) !== -1;\n }\n\n return false;\n}\n","/**\n * UUID Generation Utilities\n *\n * Provides crypto-grade UUID v4 generation that works in Node.js,\n * browsers (secure contexts), and falls back gracefully.\n */\n\n/**\n * Format 16 random bytes as a UUID v4 string\n */\nfunction formatUUID(bytes: Uint8Array): string {\n const hex = Array.from(bytes, (b) => b.toString(16).padStart(2, '0')).join('');\n return `${hex.slice(0, 8)}-${hex.slice(8, 12)}-${hex.slice(12, 16)}-${hex.slice(16, 20)}-${hex.slice(20, 32)}`;\n}\n\n/**\n * Fallback UUID generation using Math.random()\n * Only used when crypto APIs are unavailable (rare)\n */\nfunction mathRandomUUID(): string {\n return 'xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx'.replace(/[xy]/g, (c) => {\n const r = (Math.random() * 16) | 0;\n const v = c === 'x' ? r : (r & 0x3) | 0x8;\n return v.toString(16);\n });\n}\n\n/**\n * Generate a cryptographically secure UUID v4.\n *\n * Uses crypto.randomUUID() when available (Node.js 19+, modern browsers).\n * Falls back to crypto.getRandomValues() if randomUUID is unavailable.\n * Last resort uses Math.random() (non-secure, emits warning in dev).\n *\n * @returns A valid RFC 4122 UUID v4 string\n *\n * @example\n * ```typescript\n * const id = generateUUID();\n * // => \"550e8400-e29b-41d4-a716-446655440000\"\n * ```\n */\nexport function generateUUID(): string {\n // Node.js 19+ or browser with secure context\n if (typeof crypto !== 'undefined' && typeof crypto.randomUUID === 'function') {\n return crypto.randomUUID();\n }\n\n // Fallback using crypto.getRandomValues (older Node/browsers)\n if (typeof crypto !== 'undefined' && typeof crypto.getRandomValues === 'function') {\n const bytes = new Uint8Array(16);\n crypto.getRandomValues(bytes);\n // Set version (4) and variant (RFC 4122)\n bytes[6] = (bytes[6]! & 0x0f) | 0x40; // Version 4\n bytes[8] = (bytes[8]! & 0x3f) | 0x80; // Variant 1\n return formatUUID(bytes);\n }\n\n // Last resort - non-secure fallback\n if (typeof process !== 'undefined' && process.env?.NODE_ENV === 'development') {\n console.warn('[character-foundry/core] generateUUID: Using insecure Math.random() fallback');\n }\n return mathRandomUUID();\n}\n\n/**\n * Validate if a string is a valid UUID v4\n *\n * @param uuid - String to validate\n * @returns true if valid UUID v4 format\n */\nexport function isValidUUID(uuid: string): boolean {\n return /^[0-9a-f]{8}-[0-9a-f]{4}-4[0-9a-f]{3}-[89ab][0-9a-f]{3}-[0-9a-f]{12}$/i.test(uuid);\n}\n","/**\n * CCv2 to CCv3 Converter\n *\n * Converts CCv2 card data to CCv3 format.\n * Handles malformed cards gracefully by defaulting missing fields.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv2LorebookEntry,\n CCv3Data,\n CCv3CharacterBook,\n CCv3LorebookEntry,\n} from '@character-foundry/schemas';\nimport { getV2Data } from '@character-foundry/schemas';\n\n/**\n * Normalize position field - accepts both string literals and SillyTavern numeric values.\n * SillyTavern uses: 0 = before_char, 1 = after_char, other numbers passed through.\n */\nfunction normalizePosition(\n position: 'before_char' | 'after_char' | 'in_chat' | '' | number | null | undefined\n): 'before_char' | 'after_char' | 'in_chat' | number {\n if (position === undefined || position === null || position === '') return 'before_char';\n if (typeof position === 'string') return position;\n // SillyTavern numeric mapping - pass through as-is since schema now accepts numbers\n return position;\n}\n\n/**\n * Convert CCv2 lorebook entry to CCv3 format\n */\nfunction convertLorebookEntry(\n entry: CCv2LorebookEntry,\n index: number\n): CCv3LorebookEntry {\n return {\n keys: entry.keys || [],\n content: entry.content || '',\n extensions: entry.extensions || {},\n enabled: entry.enabled ?? true,\n insertion_order: entry.insertion_order ?? index,\n case_sensitive: entry.case_sensitive,\n name: entry.name || '',\n priority: entry.priority ?? 10,\n id: entry.id ?? index,\n comment: entry.comment || '',\n selective: entry.selective ?? false,\n secondary_keys: entry.secondary_keys || [],\n constant: entry.constant ?? false,\n position: normalizePosition(entry.position),\n };\n}\n\n/**\n * Convert CCv2 character book to CCv3 format\n */\nfunction convertCharacterBook(\n book: NonNullable<CCv2Data['character_book']>\n): CCv3CharacterBook {\n return {\n name: book.name,\n description: book.description,\n scan_depth: book.scan_depth,\n token_budget: book.token_budget,\n recursive_scanning: book.recursive_scanning,\n extensions: book.extensions,\n entries: book.entries.map((entry, i) => convertLorebookEntry(entry, i)),\n };\n}\n\n/**\n * Convert CCv2 card to CCv3 format.\n *\n * Philosophy: Be lenient on input - never completely drop data due to missing fields.\n * Defaults are applied for any missing required fields to ensure valid output.\n * Pure 1:1 field mapping - no extraction, no magic.\n *\n * @param input - CCv2 card data (wrapped or unwrapped, potentially malformed)\n * @returns CCv3 card data with all required fields populated\n */\nexport function ccv2ToCCv3(input: CCv2Data | CCv2Wrapped): CCv3Data {\n const data = getV2Data(input);\n\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: {\n // Core required fields - default to empty string if missing/undefined\n name: data.name ?? '',\n description: data.description ?? '',\n personality: data.personality ?? '',\n scenario: data.scenario ?? '',\n first_mes: data.first_mes ?? '',\n mes_example: data.mes_example ?? '',\n // Optional in V2, required in V3 - always provide defaults\n creator_notes: data.creator_notes || '',\n system_prompt: data.system_prompt || '',\n post_history_instructions: data.post_history_instructions || '',\n alternate_greetings: data.alternate_greetings || [],\n group_only_greetings: [],\n tags: data.tags || [],\n creator: data.creator || '',\n character_version: data.character_version || '',\n character_book: data.character_book\n ? convertCharacterBook(data.character_book)\n : undefined,\n extensions: data.extensions || {},\n },\n };\n}\n","/**\n * CCv3 to CCv2 Converter\n *\n * Converts CCv3 card data to CCv2 format.\n * Note: Some v3 features will be lost (group_only_greetings, etc.)\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv2CharacterBook,\n CCv2LorebookEntry,\n CCv3Data,\n CCv3LorebookEntry,\n} from '@character-foundry/schemas';\n\n/**\n * Fields lost when converting v3 to v2\n */\nexport const V3_TO_V2_LOST_FIELDS = [\n 'group_only_greetings',\n 'assets',\n];\n\n/**\n * Convert CCv3 lorebook entry to CCv2 format\n */\nfunction convertLorebookEntry(entry: CCv3LorebookEntry): CCv2LorebookEntry {\n return {\n keys: entry.keys,\n content: entry.content,\n extensions: entry.extensions || {},\n enabled: entry.enabled,\n insertion_order: entry.insertion_order,\n case_sensitive: entry.case_sensitive,\n name: entry.name,\n priority: entry.priority,\n id: entry.id,\n comment: entry.comment,\n selective: entry.selective,\n secondary_keys: entry.secondary_keys,\n constant: entry.constant,\n position: entry.position,\n };\n}\n\n/**\n * Convert CCv3 character book to CCv2 format\n */\nfunction convertCharacterBook(\n book: NonNullable<CCv3Data['data']['character_book']>\n): CCv2CharacterBook {\n return {\n name: book.name,\n description: book.description,\n scan_depth: book.scan_depth,\n token_budget: book.token_budget,\n recursive_scanning: book.recursive_scanning,\n extensions: book.extensions,\n entries: book.entries.map(convertLorebookEntry),\n };\n}\n\n/**\n * Convert CCv3 card to CCv2 unwrapped format\n *\n * @param card - CCv3 card data\n * @returns CCv2 unwrapped card data\n */\nexport function ccv3ToCCv2Data(card: CCv3Data): CCv2Data {\n const data = card.data;\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality,\n scenario: data.scenario,\n first_mes: data.first_mes,\n mes_example: data.mes_example,\n creator_notes: data.creator_notes,\n system_prompt: data.system_prompt,\n post_history_instructions: data.post_history_instructions,\n alternate_greetings: data.alternate_greetings,\n tags: data.tags,\n creator: data.creator,\n character_version: data.character_version,\n character_book: data.character_book\n ? convertCharacterBook(data.character_book)\n : undefined,\n extensions: data.extensions,\n };\n}\n\n/**\n * Convert CCv3 card to CCv2 wrapped format\n *\n * @param card - CCv3 card data\n * @returns CCv2 wrapped card data\n */\nexport function ccv3ToCCv2Wrapped(card: CCv3Data): CCv2Wrapped {\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: ccv3ToCCv2Data(card),\n };\n}\n\n/**\n * Check what would be lost converting v3 to v2\n */\nexport function checkV3ToV2Loss(card: CCv3Data): string[] {\n const lost: string[] = [];\n\n if (card.data.group_only_greetings && card.data.group_only_greetings.length > 0) {\n lost.push(`group_only_greetings (${card.data.group_only_greetings.length} entries)`);\n }\n\n if (card.data.assets && card.data.assets.length > 0) {\n lost.push(`assets (${card.data.assets.length} entries)`);\n }\n\n return lost;\n}\n","/**\n * To NormalizedCard Converter\n *\n * Converts any card format to NormalizedCard.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv3Data,\n NormalizedCard,\n} from '@character-foundry/schemas';\nimport { getV2Data, isWrappedV2, isV3Card } from '@character-foundry/schemas';\nimport { FormatNotSupportedError } from '@character-foundry/core';\n\n/**\n * Normalize CCv2 data to NormalizedCard\n */\nexport function normalizeV2(input: CCv2Data | CCv2Wrapped): NormalizedCard {\n const data = getV2Data(input);\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality ?? '', // Coerce null to empty string\n scenario: data.scenario,\n firstMes: data.first_mes,\n mesExample: data.mes_example ?? '', // Coerce null to empty string\n systemPrompt: data.system_prompt,\n postHistoryInstructions: data.post_history_instructions,\n creatorNotes: data.creator_notes,\n alternateGreetings: data.alternate_greetings || [],\n groupOnlyGreetings: [],\n tags: data.tags || [],\n creator: data.creator,\n characterVersion: data.character_version,\n characterBook: data.character_book ?? undefined,\n extensions: data.extensions || {},\n };\n}\n\n/**\n * Normalize CCv3 data to NormalizedCard\n */\nexport function normalizeV3(card: CCv3Data): NormalizedCard {\n const data = card.data;\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality ?? '', // Coerce null to empty string\n scenario: data.scenario,\n firstMes: data.first_mes,\n mesExample: data.mes_example ?? '', // Coerce null to empty string\n systemPrompt: data.system_prompt,\n postHistoryInstructions: data.post_history_instructions,\n creatorNotes: data.creator_notes,\n alternateGreetings: data.alternate_greetings || [],\n groupOnlyGreetings: data.group_only_greetings || [],\n tags: data.tags || [],\n creator: data.creator,\n characterVersion: data.character_version,\n characterBook: data.character_book ?? undefined,\n extensions: data.extensions || {},\n };\n}\n\n/**\n * Normalize any card format to NormalizedCard\n *\n * @param input - Card data in any supported format\n * @returns NormalizedCard\n */\nexport function normalize(input: CCv2Data | CCv2Wrapped | CCv3Data): NormalizedCard {\n if (isV3Card(input)) {\n return normalizeV3(input as CCv3Data);\n }\n\n if (isWrappedV2(input) || 'name' in input) {\n return normalizeV2(input as CCv2Data | CCv2Wrapped);\n }\n\n throw new FormatNotSupportedError('unknown', 'normalization');\n}\n","/**\n * From NormalizedCard Converter\n *\n * Converts NormalizedCard to other formats.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv3Data,\n NormalizedCard,\n} from '@character-foundry/schemas';\n\n/**\n * Convert NormalizedCard to CCv3\n */\nexport function denormalizeToV3(card: NormalizedCard): CCv3Data {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: {\n name: card.name,\n description: card.description,\n personality: card.personality,\n scenario: card.scenario,\n first_mes: card.firstMes,\n mes_example: card.mesExample,\n creator_notes: card.creatorNotes || '',\n system_prompt: card.systemPrompt || '',\n post_history_instructions: card.postHistoryInstructions || '',\n alternate_greetings: card.alternateGreetings,\n group_only_greetings: card.groupOnlyGreetings,\n tags: card.tags,\n creator: card.creator || '',\n character_version: card.characterVersion || '',\n character_book: card.characterBook,\n extensions: card.extensions,\n },\n };\n}\n\n/**\n * Convert character book to v2 format (ensure extensions are not undefined)\n */\nfunction convertCharacterBookToV2(book: NormalizedCard['characterBook']): CCv2Data['character_book'] {\n if (!book) return undefined;\n\n return {\n ...book,\n entries: book.entries.map((entry) => ({\n ...entry,\n extensions: entry.extensions || {},\n })),\n };\n}\n\n/**\n * Convert NormalizedCard to CCv2 unwrapped\n *\n * Note: group_only_greetings will be lost\n */\nexport function denormalizeToV2Data(card: NormalizedCard): CCv2Data {\n return {\n name: card.name,\n description: card.description,\n personality: card.personality,\n scenario: card.scenario,\n first_mes: card.firstMes,\n mes_example: card.mesExample,\n creator_notes: card.creatorNotes,\n system_prompt: card.systemPrompt,\n post_history_instructions: card.postHistoryInstructions,\n alternate_greetings: card.alternateGreetings,\n tags: card.tags,\n creator: card.creator,\n character_version: card.characterVersion,\n character_book: convertCharacterBookToV2(card.characterBook),\n extensions: card.extensions,\n };\n}\n\n/**\n * Convert NormalizedCard to CCv2 wrapped\n *\n * Note: group_only_greetings will be lost\n */\nexport function denormalizeToV2Wrapped(card: NormalizedCard): CCv2Wrapped {\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: denormalizeToV2Data(card),\n };\n}\n\n/**\n * Check what would be lost converting NormalizedCard to v2\n */\nexport function checkNormalizedToV2Loss(card: NormalizedCard): string[] {\n const lost: string[] = [];\n\n if (card.groupOnlyGreetings.length > 0) {\n lost.push(`group_only_greetings (${card.groupOnlyGreetings.length} entries)`);\n }\n\n return lost;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACMA,iBAAkB;ACAlB,IAAAA,cAAkB;ACAlB,IAAAA,cAAkB;AMAlB,IAAAA,cAAkB;ARWX,SAAS,oBAAoB,KAAkC;AACpE,MAAI,QAAQ,QAAQ,QAAQ,OAAW,QAAO;AAE9C,MAAI;AAEJ,MAAI,OAAO,QAAQ,UAAU;AAC3B,UAAM;EACR,WAAW,OAAO,QAAQ,UAAU;AAClC,UAAM,UAAU,IAAI,KAAK;AACzB,QAAI,CAAC,QAAS,QAAO;AAGrB,UAAM,SAAS,OAAO,OAAO;AAC7B,QAAI,CAAC,MAAM,MAAM,GAAG;AAClB,YAAM;IACR,OAAO;AAEL,YAAM,OAAO,IAAI,KAAK,OAAO;AAC7B,UAAI,MAAM,KAAK,QAAQ,CAAC,EAAG,QAAO;AAClC,YAAM,KAAK,MAAM,KAAK,QAAQ,IAAI,GAAI;IACxC;EACF,OAAO;AACL,WAAO;EACT;AAGA,MAAI,MAAM,MAAgB;AACxB,UAAM,KAAK,MAAM,MAAM,GAAI;EAC7B;AAGA,MAAI,MAAM,EAAG,QAAO;AAEpB,SAAO;AACT;AAMO,SAAS,kBAAkB,KAAkC;AAClE,MAAI,QAAQ,QAAQ,QAAQ,OAAW,QAAO;AAE9C,MAAI,OAAO,QAAQ,UAAU;AAC3B,WAAO,MAAM,GAAG,IAAI,SAAY;EAClC;AAEA,MAAI,OAAO,QAAQ,UAAU;AAC3B,UAAM,UAAU,IAAI,KAAK;AACzB,QAAI,CAAC,QAAS,QAAO;AACrB,UAAM,SAAS,OAAO,OAAO;AAC7B,WAAO,MAAM,MAAM,IAAI,SAAY;EACrC;AAEA,SAAO;AACT;AAKA,IAAM,oBAAoB,oBAAI,IAAI;EAChC;EAAQ;EAAc;EAAW;EACjC;EAAS;EAAS;EAAU;AAC9B,CAAC;AAKM,SAAS,oBAAoB,KAAsB;AACxD,MAAI,OAAO,QAAQ,SAAU,QAAO;AACpC,SAAO,kBAAkB,IAAI,GAAG,IAAI,MAAM;AAC5C;AASO,IAAM,gBAAgB,aAAE,OAAO,EAAE,SAAS;AAK1C,IAAM,aAAa,aAAE,OAAO,EAAE,KAAK;AAKnC,IAAM,aAAa,aAAE,KAAK,CAAC,MAAM,IAAI,CAAC;AAKtC,IAAM,qBAAqB,aAAE,KAAK;EACvC;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;AACF,CAAC;AAKM,IAAM,sBAAsB,aAAE,KAAK,CAAC,WAAW,aAAa,QAAQ,CAAC;AAMrE,IAAM,kBAAkB,aAAE;EAC/B;EACA,aAAE,KAAK;IACL;IACA;IACA;IACA;IACA;IACA;IACA;IACA;EACF,CAAC;AACH;AAKO,IAAM,wBAAwB,aAAE,OAAO;EAC5C,MAAM;EACN,KAAK,aAAE,OAAO;EACd,MAAM,aAAE,OAAO;EACf,KAAK,aAAE,OAAO;AAChB,CAAC;AAKM,IAAM,uBAAuB,aAAE,OAAO;EAC3C,YAAY;EACZ,MAAM,aAAE,WAAW,UAAU;EAC7B,UAAU,aAAE,OAAO;AACrB,CAAC;ACnJM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;AAC7H,CAAC,EAAE,YAAY;AAMR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,WAAW,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;EACrF,cAAcA,YAAAA,EAAE,WAAW,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;EACvF,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;;EAErC,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACnC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC7B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACvC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;AAC7C,CAAC;AAKM,IAAM,oBAAoBA,YAAAA,EAAE,OAAO;EACxC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AAiCM,SAAS,YAAY,MAAoC;AAC9D,SAAO,kBAAkB,UAAU,IAAI,EAAE;AAC3C;AA8BO,SAAS,mBAAmB,MAAwE;AACzG,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,QAAM,MAAM;AACZ,SACE,IAAI,SAAS,mBACb,IAAI,SAAS,QACb,OAAO,IAAI,SAAS;AAExB;AASO,SAAS,UAAU,MAAwC;AAEhE,MAAI,mBAAmB,IAAI,GAAG;AAC5B,WAAO,KAAK;EACd;AACA,SAAO;AACT;ACtJO,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC3H,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,MAAMA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,UAAU,QAAQ,WAAW,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC/F,OAAOA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC3B,gBAAgBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACxD,aAAaA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,GAAG,EAAE,SAAS;;EACjD,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EAChC,OAAOA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EAC/C,iBAAiBA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,OAAO,KAAK,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS;AAChF,CAAC,EAAE,YAAY;AAMR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,WAAW,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;EACrF,cAAcA,YAAAA,EAAE,WAAW,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;EACvF,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AASM,IAAM,sBAAsBA,YAAAA,EAAE,OAAO;;EAE1C,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,SAASA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC9B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EACxC,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;EACpC,sBAAsBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;;EAEpD,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,QAAQA,YAAAA,EAAE,MAAM,qBAAqB,EAAE,SAAS;EAChD,UAAUA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC9B,4BAA4BA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAC1D,QAAQA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EAErC,eAAeA,YAAAA,EAAE,WAAW,qBAAqBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;EAC1F,mBAAmBA,YAAAA,EAAE,WAAW,qBAAqBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS,CAAC;AAChG,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;EACrC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AAiCM,SAAS,SAAS,MAAiC;AACxD,SAAO,eAAe,UAAU,IAAI,EAAE;AACxC;;;AQ7HA,IAAM,SAAS,OAAO,YAAY,eAChC,QAAQ,YAAY,QACpB,QAAQ,SAAS,QAAQ;AAO3B,IAAM,yBAAyB,OAAO;AAsFtC,IAAM,oBAAoB,KAAK;AClG/B,IAAM,uBAAuB,uBAAO,IAAI,sCAAsC;AAKvE,IAAM,eAAN,cAA2B,MAAM;EAItC,YAAY,SAAiC,MAAc;AACzD,UAAM,OAAO;AAD8B,SAAA,OAAA;AAE3C,SAAK,OAAO;AAEZ,QAAI,MAAM,mBAAmB;AAC3B,YAAM,kBAAkB,MAAM,KAAK,WAAW;IAChD;EACF;;EATA,CAAU,oBAAoB,IAAI;AAUpC;AAmCO,IAAM,0BAAN,cAAsC,aAAa;EACxD,YAA4B,QAAgB,WAAoB;AAC9D,UAAM,MAAM,YACR,WAAW,MAAM,uBAAuB,SAAS,KACjD,yBAAyB,MAAM;AACnC,UAAM,KAAK,sBAAsB;AAJP,SAAA,SAAA;AAK1B,SAAK,OAAO;EACd;AACF;;;AK/CA,SAAS,kBACP,UACmD;AACnD,MAAI,aAAa,UAAa,aAAa,QAAQ,aAAa,GAAI,QAAO;AAC3E,MAAI,OAAO,aAAa,SAAU,QAAO;AAEzC,SAAO;AACT;AAKA,SAAS,qBACP,OACA,OACmB;AACnB,SAAO;IACL,MAAM,MAAM,QAAQ,CAAC;IACrB,SAAS,MAAM,WAAW;IAC1B,YAAY,MAAM,cAAc,CAAC;IACjC,SAAS,MAAM,WAAW;IAC1B,iBAAiB,MAAM,mBAAmB;IAC1C,gBAAgB,MAAM;IACtB,MAAM,MAAM,QAAQ;IACpB,UAAU,MAAM,YAAY;IAC5B,IAAI,MAAM,MAAM;IAChB,SAAS,MAAM,WAAW;IAC1B,WAAW,MAAM,aAAa;IAC9B,gBAAgB,MAAM,kBAAkB,CAAC;IACzC,UAAU,MAAM,YAAY;IAC5B,UAAU,kBAAkB,MAAM,QAAQ;EAC5C;AACF;AAKA,SAAS,qBACP,MACmB;AACnB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,YAAY,KAAK;IACjB,cAAc,KAAK;IACnB,oBAAoB,KAAK;IACzB,YAAY,KAAK;IACjB,SAAS,KAAK,QAAQ,IAAI,CAAC,OAAO,MAAM,qBAAqB,OAAO,CAAC,CAAC;EACxE;AACF;AAYO,SAAS,WAAW,OAAyC;AAClE,QAAM,OAAO,UAAU,KAAK;AAE5B,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM;;MAEJ,MAAM,KAAK,QAAQ;MACnB,aAAa,KAAK,eAAe;MACjC,aAAa,KAAK,eAAe;MACjC,UAAU,KAAK,YAAY;MAC3B,WAAW,KAAK,aAAa;MAC7B,aAAa,KAAK,eAAe;;MAEjC,eAAe,KAAK,iBAAiB;MACrC,eAAe,KAAK,iBAAiB;MACrC,2BAA2B,KAAK,6BAA6B;MAC7D,qBAAqB,KAAK,uBAAuB,CAAC;MAClD,sBAAsB,CAAC;MACvB,MAAM,KAAK,QAAQ,CAAC;MACpB,SAAS,KAAK,WAAW;MACzB,mBAAmB,KAAK,qBAAqB;MAC7C,gBAAgB,KAAK,iBACjB,qBAAqB,KAAK,cAAc,IACxC;MACJ,YAAY,KAAK,cAAc,CAAC;IAClC;EACF;AACF;AC5FO,IAAM,uBAAuB;EAClC;EACA;AACF;AAKA,SAASC,sBAAqB,OAA6C;AACzE,SAAO;IACL,MAAM,MAAM;IACZ,SAAS,MAAM;IACf,YAAY,MAAM,cAAc,CAAC;IACjC,SAAS,MAAM;IACf,iBAAiB,MAAM;IACvB,gBAAgB,MAAM;IACtB,MAAM,MAAM;IACZ,UAAU,MAAM;IAChB,IAAI,MAAM;IACV,SAAS,MAAM;IACf,WAAW,MAAM;IACjB,gBAAgB,MAAM;IACtB,UAAU,MAAM;IAChB,UAAU,MAAM;EAClB;AACF;AAKA,SAASC,sBACP,MACmB;AACnB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,YAAY,KAAK;IACjB,cAAc,KAAK;IACnB,oBAAoB,KAAK;IACzB,YAAY,KAAK;IACjB,SAAS,KAAK,QAAQ,IAAID,qBAAoB;EAChD;AACF;AAQO,SAAS,eAAe,MAA0B;AACvD,QAAM,OAAO,KAAK;AAElB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK;IAClB,UAAU,KAAK;IACf,WAAW,KAAK;IAChB,aAAa,KAAK;IAClB,eAAe,KAAK;IACpB,eAAe,KAAK;IACpB,2BAA2B,KAAK;IAChC,qBAAqB,KAAK;IAC1B,MAAM,KAAK;IACX,SAAS,KAAK;IACd,mBAAmB,KAAK;IACxB,gBAAgB,KAAK,iBACjBC,sBAAqB,KAAK,cAAc,IACxC;IACJ,YAAY,KAAK;EACnB;AACF;AAQO,SAAS,kBAAkB,MAA6B;AAC7D,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM,eAAe,IAAI;EAC3B;AACF;AAKO,SAAS,gBAAgB,MAA0B;AACxD,QAAM,OAAiB,CAAC;AAExB,MAAI,KAAK,KAAK,wBAAwB,KAAK,KAAK,qBAAqB,SAAS,GAAG;AAC/E,SAAK,KAAK,yBAAyB,KAAK,KAAK,qBAAqB,MAAM,WAAW;EACrF;AAEA,MAAI,KAAK,KAAK,UAAU,KAAK,KAAK,OAAO,SAAS,GAAG;AACnD,SAAK,KAAK,WAAW,KAAK,KAAK,OAAO,MAAM,WAAW;EACzD;AAEA,SAAO;AACT;ACxGO,SAAS,YAAY,OAA+C;AACzE,QAAM,OAAOC,UAAU,KAAK;AAE5B,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK,eAAe;;IACjC,UAAU,KAAK;IACf,UAAU,KAAK;IACf,YAAY,KAAK,eAAe;;IAChC,cAAc,KAAK;IACnB,yBAAyB,KAAK;IAC9B,cAAc,KAAK;IACnB,oBAAoB,KAAK,uBAAuB,CAAC;IACjD,oBAAoB,CAAC;IACrB,MAAM,KAAK,QAAQ,CAAC;IACpB,SAAS,KAAK;IACd,kBAAkB,KAAK;IACvB,eAAe,KAAK,kBAAkB;IACtC,YAAY,KAAK,cAAc,CAAC;EAClC;AACF;AAKO,SAAS,YAAY,MAAgC;AAC1D,QAAM,OAAO,KAAK;AAElB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK,eAAe;;IACjC,UAAU,KAAK;IACf,UAAU,KAAK;IACf,YAAY,KAAK,eAAe;;IAChC,cAAc,KAAK;IACnB,yBAAyB,KAAK;IAC9B,cAAc,KAAK;IACnB,oBAAoB,KAAK,uBAAuB,CAAC;IACjD,oBAAoB,KAAK,wBAAwB,CAAC;IAClD,MAAM,KAAK,QAAQ,CAAC;IACpB,SAAS,KAAK;IACd,kBAAkB,KAAK;IACvB,eAAe,KAAK,kBAAkB;IACtC,YAAY,KAAK,cAAc,CAAC;EAClC;AACF;AAQO,SAAS,UAAU,OAA0D;AAClF,MAAI,SAAS,KAAK,GAAG;AACnB,WAAO,YAAY,KAAiB;EACtC;AAEA,MAAI,YAAY,KAAK,KAAK,UAAU,OAAO;AACzC,WAAO,YAAY,KAA+B;EACpD;AAEA,QAAM,IAAI,wBAAwB,WAAW,eAAe;AAC9D;ACnEO,SAAS,gBAAgB,MAAgC;AAC9D,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM;MACJ,MAAM,KAAK;MACX,aAAa,KAAK;MAClB,aAAa,KAAK;MAClB,UAAU,KAAK;MACf,WAAW,KAAK;MAChB,aAAa,KAAK;MAClB,eAAe,KAAK,gBAAgB;MACpC,eAAe,KAAK,gBAAgB;MACpC,2BAA2B,KAAK,2BAA2B;MAC3D,qBAAqB,KAAK;MAC1B,sBAAsB,KAAK;MAC3B,MAAM,KAAK;MACX,SAAS,KAAK,WAAW;MACzB,mBAAmB,KAAK,oBAAoB;MAC5C,gBAAgB,KAAK;MACrB,YAAY,KAAK;IACnB;EACF;AACF;AAKA,SAAS,yBAAyB,MAAmE;AACnG,MAAI,CAAC,KAAM,QAAO;AAElB,SAAO;IACL,GAAG;IACH,SAAS,KAAK,QAAQ,IAAI,CAAC,WAAW;MACpC,GAAG;MACH,YAAY,MAAM,cAAc,CAAC;IACnC,EAAE;EACJ;AACF;AAOO,SAAS,oBAAoB,MAAgC;AAClE,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK;IAClB,UAAU,KAAK;IACf,WAAW,KAAK;IAChB,aAAa,KAAK;IAClB,eAAe,KAAK;IACpB,eAAe,KAAK;IACpB,2BAA2B,KAAK;IAChC,qBAAqB,KAAK;IAC1B,MAAM,KAAK;IACX,SAAS,KAAK;IACd,mBAAmB,KAAK;IACxB,gBAAgB,yBAAyB,KAAK,aAAa;IAC3D,YAAY,KAAK;EACnB;AACF;AAOO,SAAS,uBAAuB,MAAmC;AACxE,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM,oBAAoB,IAAI;EAChC;AACF;AAKO,SAAS,wBAAwB,MAAgC;AACtE,QAAM,OAAiB,CAAC;AAExB,MAAI,KAAK,mBAAmB,SAAS,GAAG;AACtC,SAAK,KAAK,yBAAyB,KAAK,mBAAmB,MAAM,WAAW;EAC9E;AAEA,SAAO;AACT;","names":["import_zod","z","z","convertLorebookEntry","convertCharacterBook","getV2Data"]}
|
|
1
|
+
{"version":3,"sources":["../src/normalizer.ts","../../schemas/src/common.ts","../../schemas/src/ccv2.ts","../../schemas/src/ccv3.ts","../../schemas/src/risu.ts","../../schemas/src/normalized.ts","../../schemas/src/feature-deriver.ts","../../schemas/src/detection.ts","../../schemas/src/normalizer.ts","../../schemas/src/validation.ts","../../core/src/binary.ts","../../core/src/base64.ts","../../core/src/errors.ts","../../core/src/data-url.ts","../../core/src/uri.ts","../../core/src/image.ts","../../core/src/uuid.ts","../../normalizer/src/v2-to-v3.ts","../../normalizer/src/v3-to-v2.ts","../../normalizer/src/to-normalized.ts","../../normalizer/src/from-normalized.ts"],"sourcesContent":["export * from '@character-foundry/normalizer';\n","/**\n * Common Types\n *\n * Shared types used across all card formats.\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * ISO 8601 date string schema\n */\nexport const ISO8601Schema = z.string().datetime();\n\n/**\n * UUID string schema\n */\nexport const UUIDSchema = z.string().uuid();\n\n/**\n * Card specification version schema\n */\nexport const SpecSchema = z.enum(['v2', 'v3']);\n\n/**\n * Source format identifier schema\n */\nexport const SourceFormatSchema = z.enum([\n 'png_v2', // PNG with 'chara' chunk (v2)\n 'png_v3', // PNG with 'ccv3' chunk (v3)\n 'json_v2', // Raw JSON v2\n 'json_v3', // Raw JSON v3\n 'charx', // ZIP with card.json (v3 spec)\n 'charx_risu', // ZIP with card.json + module.risum\n 'charx_jpeg', // JPEG with appended ZIP (read-only)\n 'voxta', // VoxPkg format\n]);\n\n/**\n * Original JSON shape schema\n */\nexport const OriginalShapeSchema = z.enum(['wrapped', 'unwrapped', 'legacy']);\n\n/**\n * Asset type identifier schema\n */\nexport const AssetTypeSchema = z.enum([\n 'icon',\n 'background',\n 'emotion',\n 'user_icon',\n 'sound',\n 'video',\n 'custom',\n 'x-risu-asset',\n]);\n\n/**\n * Asset descriptor schema (v3 spec)\n */\nexport const AssetDescriptorSchema = z.object({\n type: AssetTypeSchema,\n uri: z.string(),\n name: z.string(),\n ext: z.string(),\n});\n\n/**\n * Extracted asset with binary data schema\n */\nexport const ExtractedAssetSchema = z.object({\n descriptor: AssetDescriptorSchema,\n data: z.instanceof(Uint8Array),\n mimeType: z.string(),\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * ISO 8601 date string\n */\nexport type ISO8601 = z.infer<typeof ISO8601Schema>;\n\n/**\n * UUID string\n */\nexport type UUID = z.infer<typeof UUIDSchema>;\n\n/**\n * Card specification version\n */\nexport type Spec = z.infer<typeof SpecSchema>;\n\n/**\n * Source format identifier\n */\nexport type SourceFormat = z.infer<typeof SourceFormatSchema>;\n\n/**\n * Original JSON shape\n */\nexport type OriginalShape = z.infer<typeof OriginalShapeSchema>;\n\n/**\n * Asset type identifier\n */\nexport type AssetType = z.infer<typeof AssetTypeSchema>;\n\n/**\n * Asset descriptor (v3 spec)\n */\nexport type AssetDescriptor = z.infer<typeof AssetDescriptorSchema>;\n\n/**\n * Extracted asset with binary data\n */\nexport type ExtractedAsset = z.infer<typeof ExtractedAssetSchema>;\n","/**\n * Character Card v2 Types\n *\n * Based on: https://github.com/malfoyslastname/character-card-spec-v2\n */\n\nimport { z } from 'zod';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v2 cards\n */\nexport const CCv2LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n extensions: z.record(z.unknown()).optional(),\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n}).passthrough(); // Allow SillyTavern extensions like depth, probability, etc.\n\n/**\n * Character book (lorebook) schema for v2 cards\n */\nexport const CCv2CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv2LorebookEntrySchema),\n});\n\n/**\n * Character Card v2 data structure schema\n */\nexport const CCv2DataSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv2CharacterBookSchema.optional().nullable(),\n tags: z.array(z.string()).optional(),\n creator: z.string().optional(),\n character_version: z.string().optional(),\n extensions: z.record(z.unknown()).optional(),\n});\n\n/**\n * Wrapped v2 card format schema (modern tools)\n */\nexport const CCv2WrappedSchema = z.object({\n spec: z.literal('chara_card_v2'),\n spec_version: z.literal('2.0'),\n data: CCv2DataSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v2 cards\n */\nexport type CCv2LorebookEntry = z.infer<typeof CCv2LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v2 cards\n */\nexport type CCv2CharacterBook = z.infer<typeof CCv2CharacterBookSchema>;\n\n/**\n * Character Card v2 data structure\n */\nexport type CCv2Data = z.infer<typeof CCv2DataSchema>;\n\n/**\n * Wrapped v2 card format (modern tools)\n */\nexport type CCv2Wrapped = z.infer<typeof CCv2WrappedSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a wrapped v2 card\n */\nexport function isWrappedV2(data: unknown): data is CCv2Wrapped {\n return CCv2WrappedSchema.safeParse(data).success;\n}\n\n/**\n * Check if data looks like v2 card data (wrapped or unwrapped)\n */\nexport function isV2CardData(data: unknown): data is CCv2Data | CCv2Wrapped {\n return (\n CCv2WrappedSchema.safeParse(data).success ||\n CCv2DataSchema.safeParse(data).success\n );\n}\n\n/**\n * Parse and validate a wrapped v2 card\n */\nexport function parseWrappedV2(data: unknown): CCv2Wrapped {\n return CCv2WrappedSchema.parse(data);\n}\n\n/**\n * Parse and validate v2 card data\n */\nexport function parseV2Data(data: unknown): CCv2Data {\n return CCv2DataSchema.parse(data);\n}\n\n/**\n * Check if data looks like a wrapped V2 card structurally (without strict validation).\n * This is more lenient than isWrappedV2 - it just checks structure, not full schema validity.\n */\nexport function looksLikeWrappedV2(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v2' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n\n/**\n * Get v2 card data from wrapped or unwrapped format.\n *\n * Uses structural check instead of strict Zod validation to handle\n * malformed cards that have the right structure but missing/invalid fields.\n * The caller (e.g., ccv2ToCCv3) handles defaulting missing fields.\n */\nexport function getV2Data(card: CCv2Data | CCv2Wrapped): CCv2Data {\n // Use structural check - more lenient than isWrappedV2 schema validation\n if (looksLikeWrappedV2(card)) {\n return card.data as CCv2Data;\n }\n return card;\n}\n","/**\n * Character Card v3 Types\n *\n * Based on: https://github.com/kwaroran/character-card-spec-v3\n */\n\nimport { z } from 'zod';\nimport { AssetDescriptorSchema } from './common.js';\n\n// ============================================================================\n// Zod Schemas\n// ============================================================================\n\n/**\n * Lorebook entry schema for v3 cards\n */\nexport const CCv3LorebookEntrySchema = z.object({\n keys: z.array(z.string()).optional(), // Some tools use 'key' instead\n content: z.string(),\n enabled: z.boolean().default(true), // Default to enabled if missing\n insertion_order: z.preprocess((v) => v ?? 0, z.number().int()),\n // Optional fields - be lenient with nulls since wild data has them\n case_sensitive: z.boolean().nullable().optional(),\n name: z.string().optional(),\n priority: z.number().int().nullable().optional(),\n id: z.number().int().nullable().optional(),\n comment: z.string().nullable().optional(),\n selective: z.boolean().nullable().optional(),\n secondary_keys: z.array(z.string()).nullable().optional(),\n constant: z.boolean().nullable().optional(),\n position: z.union([z.enum(['before_char', 'after_char', 'in_chat']), z.number().int(), z.literal('')]).nullable().optional(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific - also lenient with types since SillyTavern uses numbers for enums\n automation_id: z.string().optional(),\n role: z.union([z.enum(['system', 'user', 'assistant']), z.number().int()]).nullable().optional(),\n group: z.string().optional(),\n scan_frequency: z.number().int().nonnegative().optional(),\n probability: z.number().min(0).max(100).optional(), // Some tools use 0-100 instead of 0-1\n use_regex: z.boolean().optional(),\n depth: z.number().int().nonnegative().optional(),\n selective_logic: z.union([z.enum(['AND', 'NOT']), z.number().int()]).optional(),\n}).passthrough(); // Allow tool-specific extensions\n\n/**\n * Character book (lorebook) schema for v3 cards\n */\nexport const CCv3CharacterBookSchema = z.object({\n name: z.string().optional(),\n description: z.string().optional(),\n scan_depth: z.number().int().nonnegative().optional(),\n token_budget: z.number().int().nonnegative().optional(),\n recursive_scanning: z.boolean().optional(),\n extensions: z.record(z.unknown()).optional(),\n entries: z.array(CCv3LorebookEntrySchema),\n});\n\n/**\n * Character Card v3 inner data structure schema.\n *\n * Note: Fields like group_only_greetings, creator, character_version, and tags\n * are technically \"required\" per V3 spec but rarely present in wild cards.\n * We use .default() to make parsing lenient while still producing valid output.\n */\nexport const CCv3DataInnerSchema = z.object({\n // Core fields - use .default('') to handle missing fields in malformed cards\n name: z.string().default(''),\n description: z.string().default(''),\n personality: z.string().nullable().default(''), // Can be null in wild (141 cards)\n scenario: z.string().default(''),\n first_mes: z.string().default(''),\n mes_example: z.string().nullable().default(''), // Can be null in wild (186 cards)\n // \"Required\" per spec but often missing in wild - use defaults for leniency\n creator: z.string().default(''),\n character_version: z.string().default(''),\n tags: z.array(z.string()).default([]),\n group_only_greetings: z.array(z.string()).default([]),\n // Optional fields\n creator_notes: z.string().optional(),\n system_prompt: z.string().optional(),\n post_history_instructions: z.string().optional(),\n alternate_greetings: z.array(z.string()).optional(),\n character_book: CCv3CharacterBookSchema.optional().nullable(),\n extensions: z.record(z.unknown()).optional(),\n // v3 specific\n assets: z.array(AssetDescriptorSchema).optional(),\n nickname: z.string().optional(),\n creator_notes_multilingual: z.record(z.string()).optional(),\n source: z.array(z.string()).optional(),\n creation_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n modification_date: z.number().int().nonnegative().optional(), // Unix timestamp in seconds\n});\n\n/**\n * Character Card v3 full structure schema\n */\nexport const CCv3DataSchema = z.object({\n spec: z.literal('chara_card_v3'),\n spec_version: z.literal('3.0'),\n data: CCv3DataInnerSchema,\n});\n\n// ============================================================================\n// TypeScript Types (inferred from Zod schemas)\n// ============================================================================\n\n/**\n * Lorebook entry for v3 cards\n */\nexport type CCv3LorebookEntry = z.infer<typeof CCv3LorebookEntrySchema>;\n\n/**\n * Character book (lorebook) for v3 cards\n */\nexport type CCv3CharacterBook = z.infer<typeof CCv3CharacterBookSchema>;\n\n/**\n * Character Card v3 inner data structure\n */\nexport type CCv3DataInner = z.infer<typeof CCv3DataInnerSchema>;\n\n/**\n * Character Card v3 full structure\n */\nexport type CCv3Data = z.infer<typeof CCv3DataSchema>;\n\n// ============================================================================\n// Type Guards & Parsers\n// ============================================================================\n\n/**\n * Check if data is a v3 card\n */\nexport function isV3Card(data: unknown): data is CCv3Data {\n return CCv3DataSchema.safeParse(data).success;\n}\n\n/**\n * Parse and validate a v3 card\n */\nexport function parseV3Card(data: unknown): CCv3Data {\n return CCv3DataSchema.parse(data);\n}\n\n/**\n * Parse and validate v3 card inner data\n */\nexport function parseV3DataInner(data: unknown): CCv3DataInner {\n return CCv3DataInnerSchema.parse(data);\n}\n\n/**\n * Get v3 card inner data\n */\nexport function getV3Data(card: CCv3Data): CCv3DataInner {\n return card.data;\n}\n\n/**\n * Check if data looks like a V3 card structurally (without strict validation).\n * More lenient than isV3Card - just checks structure, not full schema validity.\n */\nexport function looksLikeV3Card(data: unknown): data is { spec: string; data: Record<string, unknown> } {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n return (\n obj.spec === 'chara_card_v3' &&\n obj.data !== null &&\n typeof obj.data === 'object'\n );\n}\n","/**\n * RisuAI Extension Types\n *\n * These extensions are preserved as opaque blobs.\n * We do NOT interpret or transform the script contents.\n */\n\n/**\n * Risu emotions mapping (v2 style)\n * Format: [name, uri][]\n */\nexport type RisuEmotions = [string, string][];\n\n/**\n * Risu additional assets (v3 style)\n * Format: [name, uri, type][]\n */\nexport type RisuAdditionalAssets = [string, string, string][];\n\n/**\n * Risu depth prompt configuration\n */\nexport interface RisuDepthPrompt {\n depth: number;\n prompt: string;\n}\n\n/**\n * Risu extensions in card.extensions.risuai\n * Preserved as opaque - we don't interpret script contents\n */\nexport interface RisuExtensions {\n // Emotion assets\n emotions?: RisuEmotions;\n additionalAssets?: RisuAdditionalAssets;\n\n // Script data - OPAQUE, do not parse\n triggerscript?: unknown;\n customScripts?: unknown;\n\n // Voice/TTS settings\n vits?: Record<string, string>;\n\n // Depth prompt\n depth_prompt?: RisuDepthPrompt;\n\n // Other Risu-specific fields\n [key: string]: unknown;\n}\n\n/**\n * CharX x_meta entry (PNG chunk metadata preservation)\n */\nexport interface CharxMetaEntry {\n type?: string; // e.g., 'WEBP', 'PNG', 'JPEG'\n [key: string]: unknown;\n}\n\n/**\n * Check if card has Risu extensions\n */\nexport function hasRisuExtensions(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n return 'risuai' in extensions || 'risu' in extensions;\n}\n\n/**\n * Check if card has Risu scripts (triggerscript or customScripts)\n */\nexport function hasRisuScripts(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n const risu = extensions.risuai as RisuExtensions | undefined;\n if (!risu) return false;\n return !!risu.triggerscript || !!risu.customScripts;\n}\n\n/**\n * Check if card has depth prompt\n * Checks both SillyTavern style (extensions.depth_prompt) and Risu style (extensions.risuai.depth_prompt)\n */\nexport function hasDepthPrompt(extensions?: Record<string, unknown>): boolean {\n if (!extensions) return false;\n // SillyTavern top-level depth_prompt\n if ('depth_prompt' in extensions && extensions.depth_prompt) return true;\n // Risu-style depth_prompt\n const risu = extensions.risuai as RisuExtensions | undefined;\n return !!risu?.depth_prompt;\n}\n","/**\n * Normalized Card Types\n *\n * Unified view of card data regardless of source format.\n * This is a computed/virtual representation, not stored.\n */\n\nimport type { CCv3CharacterBook } from './ccv3.js';\n\n/**\n * Normalized card representation\n * Provides unified access to card data from any format\n */\nexport interface NormalizedCard {\n // Core fields (always present)\n name: string;\n description: string;\n personality: string;\n scenario: string;\n firstMes: string;\n mesExample: string;\n\n // Optional prompts\n systemPrompt?: string;\n postHistoryInstructions?: string;\n\n // Arrays\n alternateGreetings: string[];\n groupOnlyGreetings: string[];\n tags: string[];\n\n // Metadata\n creator?: string;\n creatorNotes?: string;\n characterVersion?: string;\n\n // Character book (v3 format)\n characterBook?: CCv3CharacterBook;\n\n // Extensions (preserved as-is)\n extensions: Record<string, unknown>;\n}\n\n/**\n * Create empty normalized card with defaults\n */\nexport function createEmptyNormalizedCard(): NormalizedCard {\n return {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n firstMes: '',\n mesExample: '',\n alternateGreetings: [],\n groupOnlyGreetings: [],\n tags: [],\n extensions: {},\n };\n}\n\n/**\n * Derived features extracted from card (not stored in card)\n */\nexport interface DerivedFeatures {\n // Content flags\n hasAlternateGreetings: boolean;\n alternateGreetingsCount: number;\n /** Total greetings = first_mes (1) + alternate_greetings */\n totalGreetingsCount: number;\n hasLorebook: boolean;\n lorebookEntriesCount: number;\n hasEmbeddedImages: boolean;\n embeddedImagesCount: number;\n hasGallery: boolean;\n\n // Format-specific\n hasRisuExtensions: boolean;\n hasRisuScripts: boolean;\n hasDepthPrompt: boolean;\n hasVoxtaAppearance: boolean;\n\n // Token counts (estimated)\n tokens: {\n description: number;\n personality: number;\n scenario: number;\n firstMes: number;\n mesExample: number;\n systemPrompt: number;\n total: number;\n };\n}\n\n/**\n * Create empty derived features\n */\nexport function createEmptyFeatures(): DerivedFeatures {\n return {\n hasAlternateGreetings: false,\n alternateGreetingsCount: 0,\n totalGreetingsCount: 1, // first_mes always counts as 1\n hasLorebook: false,\n lorebookEntriesCount: 0,\n hasEmbeddedImages: false,\n embeddedImagesCount: 0,\n hasGallery: false,\n hasRisuExtensions: false,\n hasRisuScripts: false,\n hasDepthPrompt: false,\n hasVoxtaAppearance: false,\n tokens: {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n },\n };\n}\n","/**\n * Feature Derivation\n *\n * Canonical feature extraction from character cards.\n * Eliminates duplicate implementations across Archive, Federation, and Architect.\n */\n\nimport type { CCv2Data } from './ccv2.js';\nimport type { CCv3DataInner } from './ccv3.js';\nimport type { DerivedFeatures } from './normalized.js';\nimport { hasRisuExtensions, hasRisuScripts, hasDepthPrompt } from './risu.js';\n\n/**\n * Derive features from a character card (V2 or V3 format).\n *\n * This is the canonical implementation - all apps should use this\n * rather than implementing their own feature detection.\n *\n * @param card - Either CCv2Data or CCv3DataInner (unwrapped)\n * @returns DerivedFeatures with all feature flags populated\n *\n * @example\n * ```typescript\n * import { deriveFeatures, parseV3Card } from '@character-foundry/schemas';\n *\n * const card = parseV3Card(data);\n * const features = deriveFeatures(card.data);\n *\n * if (features.hasLorebook) {\n * console.log(`Found ${features.lorebookEntriesCount} lorebook entries`);\n * }\n * ```\n */\nexport function deriveFeatures(card: CCv2Data | CCv3DataInner): DerivedFeatures {\n // Detect format by checking for V3-specific field\n const isV3 = 'assets' in card;\n\n // Alternate greetings\n const altGreetings = card.alternate_greetings ?? [];\n const hasAlternateGreetings = altGreetings.length > 0;\n const alternateGreetingsCount = altGreetings.length;\n // Total = first_mes (1) + alternate_greetings\n const totalGreetingsCount = 1 + alternateGreetingsCount;\n\n // Lorebook\n const characterBook = card.character_book;\n const hasLorebook = !!characterBook && characterBook.entries.length > 0;\n const lorebookEntriesCount = characterBook?.entries.length ?? 0;\n\n // Assets (V3 only) - check for visual asset types\n const assets = isV3 ? (card as CCv3DataInner).assets ?? [] : [];\n const imageAssetTypes = ['icon', 'background', 'emotion', 'custom'];\n const imageAssets = assets.filter(\n (a) =>\n imageAssetTypes.includes(a.type) ||\n ['png', 'jpg', 'jpeg', 'webp', 'gif'].includes(a.ext.toLowerCase()),\n );\n const hasGallery = imageAssets.length > 0;\n\n // Embedded images - check for data URLs in text fields\n const embeddedImageCount = countEmbeddedImages(card);\n const hasEmbeddedImages = embeddedImageCount > 0;\n\n // Extensions\n const extensions = card.extensions ?? {};\n const hasRisu = hasRisuExtensions(extensions);\n const hasScripts = hasRisuScripts(extensions);\n const hasDepth = hasDepthPrompt(extensions);\n const hasVoxta = checkVoxtaAppearance(extensions);\n\n // Token counts - initialize to zero (actual counting happens in tokenizers package)\n const tokens = {\n description: 0,\n personality: 0,\n scenario: 0,\n firstMes: 0,\n mesExample: 0,\n systemPrompt: 0,\n total: 0,\n };\n\n return {\n hasAlternateGreetings,\n alternateGreetingsCount,\n totalGreetingsCount,\n hasLorebook,\n lorebookEntriesCount,\n hasEmbeddedImages,\n embeddedImagesCount: embeddedImageCount,\n hasGallery,\n hasRisuExtensions: hasRisu,\n hasRisuScripts: hasScripts,\n hasDepthPrompt: hasDepth,\n hasVoxtaAppearance: hasVoxta,\n tokens,\n };\n}\n\n/**\n * Count embedded images (data URLs) in card text fields.\n * Looks for base64-encoded images in description, personality, scenario, etc.\n */\nfunction countEmbeddedImages(card: CCv2Data | CCv3DataInner): number {\n const textFields = [\n card.description,\n card.personality,\n card.scenario,\n card.first_mes,\n card.mes_example,\n card.creator_notes,\n card.system_prompt,\n card.post_history_instructions,\n ...(card.alternate_greetings ?? []),\n ].filter((field): field is string => typeof field === 'string');\n\n // Add group_only_greetings if V3\n if ('group_only_greetings' in card) {\n textFields.push(...(card.group_only_greetings ?? []));\n }\n\n let count = 0;\n const dataUrlPattern = /data:image\\/[^;]+;base64,/g;\n\n for (const text of textFields) {\n const matches = text.match(dataUrlPattern);\n if (matches) {\n count += matches.length;\n }\n }\n\n return count;\n}\n\n/**\n * Check if card has Voxta appearance data.\n * Voxta stores appearance in extensions.voxta.appearance\n */\nfunction checkVoxtaAppearance(extensions: Record<string, unknown>): boolean {\n if (!extensions.voxta) return false;\n const voxta = extensions.voxta as Record<string, unknown>;\n return !!voxta.appearance;\n}\n","/**\n * Format Detection\n *\n * Detect card specification version from JSON data.\n */\n\nimport type { Spec } from './common.js';\n\n/**\n * V3-only fields that indicate a V3 card\n */\nconst V3_ONLY_FIELDS = ['group_only_greetings', 'creation_date', 'modification_date', 'assets'] as const;\n\n/**\n * Result from detailed spec detection\n */\nexport interface SpecDetectionResult {\n /** Detected spec version */\n spec: Spec | null;\n /** Confidence level of detection */\n confidence: 'high' | 'medium' | 'low';\n /** What fields/values indicated this spec */\n indicators: string[];\n /** Anomalies or inconsistencies detected */\n warnings: string[];\n}\n\n/**\n * Detect card spec version from parsed JSON\n * Returns 'v2', 'v3', or null if not recognized\n */\nexport function detectSpec(data: unknown): Spec | null {\n return detectSpecDetailed(data).spec;\n}\n\n/**\n * Detailed spec detection with confidence and reasoning.\n * Useful for debugging and logging.\n */\nexport function detectSpecDetailed(data: unknown): SpecDetectionResult {\n const result: SpecDetectionResult = {\n spec: null,\n confidence: 'low',\n indicators: [],\n warnings: [],\n };\n\n if (!data || typeof data !== 'object') {\n result.indicators.push('Input is not an object');\n return result;\n }\n\n const obj = data as Record<string, unknown>;\n const dataObj = (obj.data && typeof obj.data === 'object' ? obj.data : null) as Record<\n string,\n unknown\n > | null;\n\n // Check for explicit spec markers (HIGH confidence)\n\n // Explicit v3 spec marker\n if (obj.spec === 'chara_card_v3') {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v3\"');\n\n // Check for inconsistencies\n if (obj.spec_version && obj.spec_version !== '3.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v3 spec`);\n }\n\n return result;\n }\n\n // Explicit v2 spec marker\n if (obj.spec === 'chara_card_v2') {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push('spec field is \"chara_card_v2\"');\n\n // Check for inconsistencies - V3-only fields in V2 card\n if (dataObj) {\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n result.warnings.push(`V3-only field \"${field}\" found in V2 card`);\n }\n }\n }\n\n if (obj.spec_version && obj.spec_version !== '2.0') {\n result.warnings.push(`spec_version \"${obj.spec_version}\" inconsistent with v2 spec`);\n }\n\n return result;\n }\n\n // Check spec_version field (HIGH confidence)\n if (typeof obj.spec_version === 'string') {\n if (obj.spec_version.startsWith('3')) {\n result.spec = 'v3';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"3\"`);\n return result;\n }\n if (obj.spec_version.startsWith('2')) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version \"${obj.spec_version}\" starts with \"2\"`);\n return result;\n }\n }\n\n if (obj.spec_version === 2.0 || obj.spec_version === 2) {\n result.spec = 'v2';\n result.confidence = 'high';\n result.indicators.push(`spec_version is numeric ${obj.spec_version}`);\n return result;\n }\n\n // Check for V3-only fields (MEDIUM confidence)\n if (dataObj) {\n const v3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in dataObj) {\n v3Fields.push(field);\n }\n }\n\n if (v3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields: ${v3Fields.join(', ')}`);\n return result;\n }\n }\n\n // Check root level for V3-only fields (also MEDIUM confidence)\n const rootV3Fields: string[] = [];\n for (const field of V3_ONLY_FIELDS) {\n if (field in obj) {\n rootV3Fields.push(field);\n }\n }\n if (rootV3Fields.length > 0) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`Has V3-only fields at root: ${rootV3Fields.join(', ')}`);\n result.warnings.push('V3 fields found at root level instead of data object');\n return result;\n }\n\n // Wrapped format with data object (MEDIUM confidence)\n if (obj.spec && dataObj) {\n const dataName = dataObj.name;\n if (dataName && typeof dataName === 'string') {\n // Infer from spec string\n if (typeof obj.spec === 'string') {\n if (obj.spec.includes('v3') || obj.spec.includes('3')) {\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v3\" or \"3\"`);\n return result;\n }\n if (obj.spec.includes('v2') || obj.spec.includes('2')) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push(`spec field \"${obj.spec}\" contains \"v2\" or \"2\"`);\n return result;\n }\n }\n // Default wrapped format to v3 (modern)\n result.spec = 'v3';\n result.confidence = 'medium';\n result.indicators.push('Has wrapped format with spec and data.name');\n return result;\n }\n }\n\n // Unwrapped format - V1/V2 like structure (MEDIUM confidence)\n if (obj.name && typeof obj.name === 'string') {\n if ('description' in obj || 'personality' in obj || 'scenario' in obj) {\n result.spec = 'v2';\n result.confidence = 'medium';\n result.indicators.push('Unwrapped format with name, description/personality/scenario');\n return result;\n }\n }\n\n // Check if data object has card-like structure without spec (LOW confidence)\n if (dataObj && typeof dataObj.name === 'string') {\n if ('description' in dataObj || 'personality' in dataObj) {\n result.spec = 'v2';\n result.confidence = 'low';\n result.indicators.push('Has data object with name and card fields, but no spec');\n result.warnings.push('Missing spec field');\n return result;\n }\n }\n\n result.indicators.push('No card structure detected');\n return result;\n}\n\n/**\n * Check if card has a lorebook\n */\nexport function hasLorebook(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Check wrapped format\n const wrapped = obj.data as Record<string, unknown> | undefined;\n if (wrapped?.character_book) {\n const book = wrapped.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n // Check unwrapped format\n if (obj.character_book) {\n const book = obj.character_book as Record<string, unknown>;\n if (Array.isArray(book.entries) && book.entries.length > 0) return true;\n }\n\n return false;\n}\n\n/**\n * Check if data looks like a valid card structure\n */\nexport function looksLikeCard(data: unknown): boolean {\n if (!data || typeof data !== 'object') return false;\n const obj = data as Record<string, unknown>;\n\n // Has explicit spec marker\n if (obj.spec === 'chara_card_v2' || obj.spec === 'chara_card_v3') {\n return true;\n }\n\n // Has wrapped data with name\n if (obj.data && typeof obj.data === 'object') {\n const dataObj = obj.data as Record<string, unknown>;\n if (typeof dataObj.name === 'string' && dataObj.name.length > 0) {\n return true;\n }\n }\n\n // Has unwrapped card-like structure\n if (typeof obj.name === 'string' && obj.name.length > 0) {\n if ('description' in obj || 'personality' in obj || 'first_mes' in obj) {\n return true;\n }\n }\n\n return false;\n}\n","/**\n * Card Normalizer\n *\n * Handles normalization of malformed card data from various sources.\n * Fixes common issues like wrong spec values, misplaced fields, missing required fields.\n */\n\nimport type { CCv2Data, CCv2Wrapped, CCv2CharacterBook, CCv2LorebookEntry } from './ccv2.js';\nimport type { CCv3Data, CCv3CharacterBook, CCv3LorebookEntry } from './ccv3.js';\nimport { detectSpec } from './detection.js';\n\n/**\n * Position values as numbers (non-standard) and their string equivalents\n */\nconst POSITION_MAP: Record<number, 'before_char' | 'after_char'> = {\n 0: 'before_char',\n 1: 'after_char',\n};\n\n/**\n * V3-only lorebook entry fields that should be moved to extensions for V2\n */\nconst V3_ONLY_ENTRY_FIELDS = [\n 'probability',\n 'depth',\n 'group',\n 'scan_frequency',\n 'use_regex',\n 'selective_logic',\n 'role',\n 'automation_id',\n] as const;\n\n/**\n * Required V2 card fields with their defaults\n */\nconst V2_REQUIRED_DEFAULTS: Partial<CCv2Data> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n};\n\n/**\n * Required V3 card fields with their defaults\n */\nconst V3_REQUIRED_DEFAULTS: Partial<CCv3Data['data']> = {\n name: '',\n description: '',\n personality: '',\n scenario: '',\n first_mes: '',\n mes_example: '',\n creator: '',\n character_version: '1.0',\n tags: [],\n group_only_greetings: [],\n};\n\n/**\n * Fields that belong at root level for wrapped format\n */\nconst _ROOT_FIELDS = ['spec', 'spec_version', 'data'] as const;\n\n/**\n * Fields that belong in the data object\n */\nconst DATA_FIELDS = [\n 'name',\n 'description',\n 'personality',\n 'scenario',\n 'first_mes',\n 'mes_example',\n 'creator_notes',\n 'system_prompt',\n 'post_history_instructions',\n 'alternate_greetings',\n 'character_book',\n 'tags',\n 'creator',\n 'character_version',\n 'extensions',\n 'assets',\n 'nickname',\n 'creator_notes_multilingual',\n 'source',\n 'creation_date',\n 'modification_date',\n 'group_only_greetings',\n] as const;\n\n/**\n * Deep clone an object without mutating the original\n */\nfunction deepClone<T>(obj: T): T {\n if (obj === null || obj === undefined) {\n return obj;\n }\n if (Array.isArray(obj)) {\n return obj.map((item) => deepClone(item)) as T;\n }\n if (typeof obj === 'object') {\n const result: Record<string, unknown> = {};\n for (const [key, value] of Object.entries(obj as Record<string, unknown>)) {\n result[key] = deepClone(value);\n }\n return result as T;\n }\n return obj;\n}\n\n/**\n * Check if a timestamp is in milliseconds (13+ digits)\n */\nfunction isMilliseconds(timestamp: number): boolean {\n // Timestamps before year 2001 in seconds: < 1000000000\n // Timestamps in milliseconds are typically 13 digits: 1000000000000+\n return timestamp > 10000000000;\n}\n\n/**\n * CardNormalizer - handles normalization of malformed card data\n */\nexport const CardNormalizer = {\n /**\n * Normalize card data to valid schema format.\n *\n * Handles:\n * - Fixing spec/spec_version values\n * - Moving misplaced fields to correct locations\n * - Adding missing required fields with defaults\n * - Handling hybrid formats (fields at root AND in data object)\n *\n * @param data - Raw card data (potentially malformed)\n * @param spec - Target spec version\n * @returns Normalized card data (does not mutate input)\n */\n normalize(data: unknown, spec: 'v2' | 'v3'): CCv2Wrapped | CCv3Data {\n if (!data || typeof data !== 'object') {\n // Return minimal valid card\n if (spec === 'v3') {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: { ...V3_REQUIRED_DEFAULTS } as CCv3Data['data'],\n };\n }\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: { ...V2_REQUIRED_DEFAULTS } as CCv2Data,\n };\n }\n\n const obj = data as Record<string, unknown>;\n const result: Record<string, unknown> = {};\n\n // Build merged data object from root fields + existing data object\n const existingData = (obj.data && typeof obj.data === 'object' ? obj.data : {}) as Record<\n string,\n unknown\n >;\n const mergedData: Record<string, unknown> = {};\n\n // Copy existing data first\n for (const [key, value] of Object.entries(existingData)) {\n mergedData[key] = deepClone(value);\n }\n\n // Move any misplaced root-level data fields into data object\n // (ChubAI hybrid format fix)\n for (const field of DATA_FIELDS) {\n if (field in obj && !(field in mergedData)) {\n mergedData[field] = deepClone(obj[field]);\n }\n }\n\n // Handle character_book: null -> remove entirely\n if (mergedData.character_book === null) {\n delete mergedData.character_book;\n }\n\n // Normalize character_book if present\n if (mergedData.character_book && typeof mergedData.character_book === 'object') {\n mergedData.character_book = this.normalizeCharacterBook(\n mergedData.character_book as Record<string, unknown>,\n spec\n );\n }\n\n // Apply defaults for required fields\n const defaults = spec === 'v3' ? V3_REQUIRED_DEFAULTS : V2_REQUIRED_DEFAULTS;\n for (const [key, defaultValue] of Object.entries(defaults)) {\n if (!(key in mergedData) || mergedData[key] === undefined) {\n mergedData[key] = Array.isArray(defaultValue) ? [...defaultValue] : defaultValue;\n }\n }\n\n // Ensure arrays are actually arrays\n if (mergedData.tags && !Array.isArray(mergedData.tags)) {\n mergedData.tags = [];\n }\n if (mergedData.alternate_greetings && !Array.isArray(mergedData.alternate_greetings)) {\n mergedData.alternate_greetings = [];\n }\n if (spec === 'v3') {\n if (\n mergedData.group_only_greetings &&\n !Array.isArray(mergedData.group_only_greetings)\n ) {\n mergedData.group_only_greetings = [];\n }\n }\n\n // Build result with correct spec\n if (spec === 'v3') {\n result.spec = 'chara_card_v3';\n result.spec_version = '3.0';\n result.data = this.fixTimestampsInner(mergedData);\n } else {\n result.spec = 'chara_card_v2';\n result.spec_version = '2.0';\n result.data = mergedData;\n }\n\n return result as unknown as CCv2Wrapped | CCv3Data;\n },\n\n /**\n * Normalize a character book (lorebook).\n *\n * Handles:\n * - Ensuring required fields exist\n * - Converting numeric position values to string enums\n * - Moving V3-only fields to extensions for V2 compatibility\n *\n * @param book - Raw character book data\n * @param spec - Target spec version\n * @returns Normalized character book\n */\n normalizeCharacterBook(\n book: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2CharacterBook | CCv3CharacterBook {\n const result: Record<string, unknown> = {};\n\n // Copy book-level fields\n if (book.name !== undefined) result.name = book.name;\n if (book.description !== undefined) result.description = book.description;\n if (book.scan_depth !== undefined) result.scan_depth = book.scan_depth;\n if (book.token_budget !== undefined) result.token_budget = book.token_budget;\n if (book.recursive_scanning !== undefined)\n result.recursive_scanning = book.recursive_scanning;\n if (book.extensions !== undefined) result.extensions = deepClone(book.extensions);\n\n // Normalize entries\n const entries = Array.isArray(book.entries) ? book.entries : [];\n result.entries = entries.map((entry) =>\n this.normalizeEntry(entry as Record<string, unknown>, spec)\n );\n\n return result as unknown as CCv2CharacterBook | CCv3CharacterBook;\n },\n\n /**\n * Normalize a single lorebook entry.\n *\n * Handles:\n * - Converting numeric position to string enum\n * - Moving V3-only fields to extensions for V2\n * - Ensuring required fields exist\n *\n * @param entry - Raw entry data\n * @param spec - Target spec version\n * @returns Normalized entry\n */\n normalizeEntry(\n entry: Record<string, unknown>,\n spec: 'v2' | 'v3'\n ): CCv2LorebookEntry | CCv3LorebookEntry {\n const result: Record<string, unknown> = {};\n\n // Required fields with defaults\n result.keys = Array.isArray(entry.keys) ? [...entry.keys] : [];\n result.content = typeof entry.content === 'string' ? entry.content : '';\n result.enabled = entry.enabled !== false; // default true\n result.insertion_order =\n typeof entry.insertion_order === 'number' ? entry.insertion_order : 0;\n\n // For V2, extensions is required\n if (spec === 'v2') {\n result.extensions =\n entry.extensions && typeof entry.extensions === 'object'\n ? deepClone(entry.extensions)\n : {};\n }\n\n // Optional fields\n if (entry.case_sensitive !== undefined) result.case_sensitive = entry.case_sensitive;\n if (entry.name !== undefined) result.name = entry.name;\n if (entry.priority !== undefined) result.priority = entry.priority;\n if (entry.id !== undefined) result.id = entry.id;\n if (entry.comment !== undefined) result.comment = entry.comment;\n if (entry.selective !== undefined) result.selective = entry.selective;\n if (entry.secondary_keys !== undefined) {\n result.secondary_keys = Array.isArray(entry.secondary_keys)\n ? [...entry.secondary_keys]\n : [];\n }\n if (entry.constant !== undefined) result.constant = entry.constant;\n\n // Position: convert numeric to string enum\n if (entry.position !== undefined) {\n if (typeof entry.position === 'number') {\n result.position = POSITION_MAP[entry.position] || 'before_char';\n } else if (entry.position === 'before_char' || entry.position === 'after_char') {\n result.position = entry.position;\n }\n }\n\n // Handle V3-only fields\n if (spec === 'v3') {\n // Copy V3 fields directly\n if (entry.extensions !== undefined) result.extensions = deepClone(entry.extensions);\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n result[field] = entry[field];\n }\n }\n } else {\n // V2: Move V3-only fields to extensions\n const ext = (result.extensions || {}) as Record<string, unknown>;\n for (const field of V3_ONLY_ENTRY_FIELDS) {\n if (entry[field] !== undefined) {\n ext[field] = entry[field];\n }\n }\n result.extensions = ext;\n }\n\n return result as unknown as CCv2LorebookEntry | CCv3LorebookEntry;\n },\n\n /**\n * Fix CharacterTavern timestamp format (milliseconds -> seconds).\n *\n * CCv3 spec requires timestamps in seconds (Unix epoch).\n * CharacterTavern exports timestamps in milliseconds.\n *\n * @param data - V3 card data\n * @returns Card data with fixed timestamps (does not mutate input)\n */\n fixTimestamps(data: CCv3Data): CCv3Data {\n const result = deepClone(data);\n result.data = this.fixTimestampsInner(\n result.data as unknown as Record<string, unknown>\n ) as unknown as CCv3Data['data'];\n return result;\n },\n\n /**\n * Internal: fix timestamps in data object\n */\n fixTimestampsInner(data: Record<string, unknown>): Record<string, unknown> {\n const result = { ...data };\n\n if (typeof result.creation_date === 'number') {\n if (isMilliseconds(result.creation_date)) {\n result.creation_date = Math.floor(result.creation_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.creation_date as number) < 0) {\n delete result.creation_date;\n }\n }\n\n if (typeof result.modification_date === 'number') {\n if (isMilliseconds(result.modification_date)) {\n result.modification_date = Math.floor(result.modification_date / 1000);\n }\n // Sanitize negative timestamps (.NET default dates like 0001-01-01)\n if ((result.modification_date as number) < 0) {\n delete result.modification_date;\n }\n }\n\n return result;\n },\n\n /**\n * Auto-detect spec and normalize.\n *\n * @param data - Raw card data\n * @returns Normalized card data, or null if not a valid card\n */\n autoNormalize(data: unknown): CCv2Wrapped | CCv3Data | null {\n const spec = detectSpec(data);\n if (!spec) return null;\n\n // V1 cards get upgraded to V2\n const targetSpec = spec === 'v3' ? 'v3' : 'v2';\n return this.normalize(data, targetSpec);\n },\n};\n\nexport type { CCv2Wrapped, CCv3Data };\n","/**\n * Validation Utilities\n *\n * Helper functions for Zod validation with Foundry error integration.\n */\n\nimport { z } from 'zod';\n\n/**\n * Convert Zod error to human-readable message\n */\nexport function zodErrorToMessage(zodError: z.ZodError, context?: string): string {\n const messages = zodError.errors.map((err) => {\n const path = err.path.length > 0 ? `${err.path.join('.')}: ` : '';\n return `${path}${err.message}`;\n });\n\n const message = messages.join('; ');\n return context ? `${context} - ${message}` : message;\n}\n\n/**\n * Get the first error field from Zod error\n */\nexport function getFirstErrorField(zodError: z.ZodError): string | undefined {\n return zodError.errors[0]?.path[0]?.toString();\n}\n\n/**\n * Safe parse with detailed error information\n */\nexport function safeParse<T>(\n schema: z.ZodSchema<T>,\n data: unknown\n): { success: true; data: T } | { success: false; error: string; field?: string } {\n const result = schema.safeParse(data);\n\n if (result.success) {\n return { success: true, data: result.data };\n }\n\n return {\n success: false,\n error: zodErrorToMessage(result.error),\n field: getFirstErrorField(result.error),\n };\n}\n","/**\n * Binary Data Utilities\n *\n * Universal binary data operations using Uint8Array.\n * Works in both Node.js and browser environments.\n */\n\n/**\n * Universal binary data type (works in both environments)\n */\nexport type BinaryData = Uint8Array;\n\n/**\n * Read a 32-bit big-endian unsigned integer\n */\nexport function readUInt32BE(data: BinaryData, offset: number): number {\n return (\n (data[offset]! << 24) |\n (data[offset + 1]! << 16) |\n (data[offset + 2]! << 8) |\n data[offset + 3]!\n ) >>> 0;\n}\n\n/**\n * Write a 32-bit big-endian unsigned integer\n */\nexport function writeUInt32BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 24) & 0xff;\n data[offset + 1] = (value >>> 16) & 0xff;\n data[offset + 2] = (value >>> 8) & 0xff;\n data[offset + 3] = value & 0xff;\n}\n\n/**\n * Read a 16-bit big-endian unsigned integer\n */\nexport function readUInt16BE(data: BinaryData, offset: number): number {\n return ((data[offset]! << 8) | data[offset + 1]!) >>> 0;\n}\n\n/**\n * Write a 16-bit big-endian unsigned integer\n */\nexport function writeUInt16BE(data: BinaryData, value: number, offset: number): void {\n data[offset] = (value >>> 8) & 0xff;\n data[offset + 1] = value & 0xff;\n}\n\n/**\n * Find a byte sequence in binary data\n */\nexport function indexOf(data: BinaryData, search: BinaryData, fromIndex = 0): number {\n outer: for (let i = fromIndex; i <= data.length - search.length; i++) {\n for (let j = 0; j < search.length; j++) {\n if (data[i + j] !== search[j]) continue outer;\n }\n return i;\n }\n return -1;\n}\n\n/**\n * Concatenate multiple binary arrays\n */\nexport function concat(...arrays: BinaryData[]): BinaryData {\n const totalLength = arrays.reduce((sum, arr) => sum + arr.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const arr of arrays) {\n result.set(arr, offset);\n offset += arr.length;\n }\n return result;\n}\n\n/**\n * Slice binary data (returns a view, not a copy)\n */\nexport function slice(data: BinaryData, start: number, end?: number): BinaryData {\n return data.subarray(start, end);\n}\n\n/**\n * Copy a portion of binary data (returns a new array)\n */\nexport function copy(data: BinaryData, start: number, end?: number): BinaryData {\n return data.slice(start, end);\n}\n\n/**\n * Convert string to binary (UTF-8)\n */\nexport function fromString(str: string): BinaryData {\n return new TextEncoder().encode(str);\n}\n\n/**\n * Convert binary to string (UTF-8)\n */\nexport function toString(data: BinaryData): string {\n return new TextDecoder().decode(data);\n}\n\n/**\n * Convert string to binary (Latin1 - for PNG keywords and similar)\n */\nexport function fromLatin1(str: string): BinaryData {\n const result = new Uint8Array(str.length);\n for (let i = 0; i < str.length; i++) {\n result[i] = str.charCodeAt(i) & 0xff;\n }\n return result;\n}\n\n/**\n * Convert binary to string (Latin1)\n */\nexport function toLatin1(data: BinaryData): string {\n let result = '';\n for (let i = 0; i < data.length; i++) {\n result += String.fromCharCode(data[i]!);\n }\n return result;\n}\n\n/**\n * Compare two binary arrays for equality\n */\nexport function equals(a: BinaryData, b: BinaryData): boolean {\n if (a.length !== b.length) return false;\n for (let i = 0; i < a.length; i++) {\n if (a[i] !== b[i]) return false;\n }\n return true;\n}\n\n/**\n * Create a new Uint8Array filled with zeros\n */\nexport function alloc(size: number): BinaryData {\n return new Uint8Array(size);\n}\n\n/**\n * Create a Uint8Array from an array of numbers\n */\nexport function from(data: number[] | ArrayBuffer | BinaryData): BinaryData {\n if (data instanceof Uint8Array) {\n return data;\n }\n if (data instanceof ArrayBuffer) {\n return new Uint8Array(data);\n }\n return new Uint8Array(data);\n}\n\n/**\n * Check if value is a Uint8Array\n */\nexport function isBinaryData(value: unknown): value is BinaryData {\n return value instanceof Uint8Array;\n}\n\n/**\n * Convert Node.js Buffer to Uint8Array (no-op if already Uint8Array)\n * This provides compatibility when interfacing with Node.js code\n */\nexport function toUint8Array(data: BinaryData | Buffer): BinaryData {\n if (data instanceof Uint8Array) {\n // Buffer extends Uint8Array, but we want a plain Uint8Array\n // This ensures we get a proper Uint8Array in all cases\n if (Object.getPrototypeOf(data).constructor.name === 'Buffer') {\n return new Uint8Array(data.buffer, data.byteOffset, data.byteLength);\n }\n return data;\n }\n return new Uint8Array(data);\n}\n\n/**\n * Convert binary data to hex string\n */\nexport function toHex(data: BinaryData): string {\n return Array.from(data)\n .map(b => b.toString(16).padStart(2, '0'))\n .join('');\n}\n\n/**\n * Convert hex string to binary data\n */\nexport function fromHex(hex: string): BinaryData {\n const bytes = new Uint8Array(hex.length / 2);\n for (let i = 0; i < bytes.length; i++) {\n bytes[i] = parseInt(hex.substr(i * 2, 2), 16);\n }\n return bytes;\n}\n","/**\n * Universal Base64 Encoding/Decoding\n *\n * Works in both Node.js and browser environments.\n */\n\nimport type { BinaryData } from './binary.js';\n\n/**\n * Check if we're in a Node.js environment\n */\nconst isNode = typeof process !== 'undefined' &&\n process.versions != null &&\n process.versions.node != null;\n\n/**\n * Threshold for switching to chunked encoding in browsers (1MB)\n * Below this, simple string concatenation is fast enough.\n * Above this, quadratic string growth becomes a problem.\n */\nconst LARGE_BUFFER_THRESHOLD = 1024 * 1024;\n\n/**\n * Encode binary data to base64 string\n *\n * PERFORMANCE: For large buffers (>1MB) in browsers, this automatically\n * uses the chunked implementation to avoid quadratic string concatenation.\n */\nexport function encode(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: use chunked encoding for large buffers to avoid O(n²) string growth\n if (data.length > LARGE_BUFFER_THRESHOLD) {\n return encodeChunked(data);\n }\n\n // Small buffers: simple approach is fast enough\n let binary = '';\n for (let i = 0; i < data.length; i++) {\n binary += String.fromCharCode(data[i]!);\n }\n return btoa(binary);\n}\n\n/**\n * Decode base64 string to binary data\n */\nexport function decode(base64: string): BinaryData {\n if (isNode) {\n // Node.js: use Buffer\n return new Uint8Array(Buffer.from(base64, 'base64'));\n }\n\n // Browser: use atob\n const binary = atob(base64);\n const result = new Uint8Array(binary.length);\n for (let i = 0; i < binary.length; i++) {\n result[i] = binary.charCodeAt(i);\n }\n return result;\n}\n\n/**\n * Check if a string is valid base64\n */\nexport function isBase64(str: string): boolean {\n if (str.length === 0) return false;\n // Base64 regex: only valid base64 characters, length multiple of 4 (with padding)\n const base64Regex = /^[A-Za-z0-9+/]*={0,2}$/;\n return base64Regex.test(str) && str.length % 4 === 0;\n}\n\n/**\n * Encode binary data to URL-safe base64 string\n * Replaces + with -, / with _, and removes padding\n */\nexport function encodeUrlSafe(data: BinaryData): string {\n return encode(data)\n .replace(/\\+/g, '-')\n .replace(/\\//g, '_')\n .replace(/=+$/, '');\n}\n\n/**\n * Decode URL-safe base64 string to binary data\n */\nexport function decodeUrlSafe(base64: string): BinaryData {\n // Add back padding if needed\n let padded = base64\n .replace(/-/g, '+')\n .replace(/_/g, '/');\n\n while (padded.length % 4 !== 0) {\n padded += '=';\n }\n\n return decode(padded);\n}\n\n/**\n * Chunk size for encoding large buffers (64KB)\n * Prevents stack overflow when using String.fromCharCode with spread operator\n */\nconst ENCODE_CHUNK_SIZE = 64 * 1024;\n\n/**\n * Encode binary data to base64 string with chunking for large buffers.\n * Handles buffers >10MB without stack overflow.\n *\n * @param data - Binary data to encode\n * @returns Base64 encoded string\n *\n * @example\n * ```typescript\n * const largeBuffer = new Uint8Array(20 * 1024 * 1024); // 20MB\n * const base64 = encodeChunked(largeBuffer); // No stack overflow\n * ```\n */\nexport function encodeChunked(data: BinaryData): string {\n if (isNode) {\n // Node.js: Buffer handles large data efficiently\n return Buffer.from(data).toString('base64');\n }\n\n // Browser: process in chunks to avoid stack overflow\n const chunks: string[] = [];\n\n for (let i = 0; i < data.length; i += ENCODE_CHUNK_SIZE) {\n const chunk = data.subarray(i, Math.min(i + ENCODE_CHUNK_SIZE, data.length));\n let binary = '';\n for (let j = 0; j < chunk.length; j++) {\n binary += String.fromCharCode(chunk[j]!);\n }\n chunks.push(binary);\n }\n\n return btoa(chunks.join(''));\n}\n","/**\n * Error Classes\n *\n * Specific error types for character card operations.\n * All errors extend FoundryError for consistent handling.\n */\n\n/** Symbol to identify FoundryError instances across ESM/CJS boundaries */\nconst FOUNDRY_ERROR_MARKER = Symbol.for('@character-foundry/core:FoundryError');\n\n/**\n * Base error class for all Character Foundry errors\n */\nexport class FoundryError extends Error {\n /** @internal Marker for cross-module identification */\n readonly [FOUNDRY_ERROR_MARKER] = true;\n\n constructor(message: string, public readonly code: string) {\n super(message);\n this.name = 'FoundryError';\n // Maintains proper stack trace in V8 environments\n if (Error.captureStackTrace) {\n Error.captureStackTrace(this, this.constructor);\n }\n }\n}\n\n/**\n * Error during card parsing\n */\nexport class ParseError extends FoundryError {\n constructor(message: string, public readonly format?: string) {\n super(message, 'PARSE_ERROR');\n this.name = 'ParseError';\n }\n}\n\n/**\n * Error during card validation\n */\nexport class ValidationError extends FoundryError {\n constructor(message: string, public readonly field?: string) {\n super(message, 'VALIDATION_ERROR');\n this.name = 'ValidationError';\n }\n}\n\n/**\n * Asset not found in card or archive\n */\nexport class AssetNotFoundError extends FoundryError {\n constructor(public readonly uri: string) {\n super(`Asset not found: ${uri}`, 'ASSET_NOT_FOUND');\n this.name = 'AssetNotFoundError';\n }\n}\n\n/**\n * Format not supported for operation\n */\nexport class FormatNotSupportedError extends FoundryError {\n constructor(public readonly format: string, operation?: string) {\n const msg = operation\n ? `Format '${format}' not supported for ${operation}`\n : `Format not supported: ${format}`;\n super(msg, 'FORMAT_NOT_SUPPORTED');\n this.name = 'FormatNotSupportedError';\n }\n}\n\n/**\n * File size exceeds limits\n */\nexport class SizeLimitError extends FoundryError {\n constructor(\n public readonly actualSize: number,\n public readonly maxSize: number,\n context?: string\n ) {\n const actualMB = (actualSize / 1024 / 1024).toFixed(2);\n const maxMB = (maxSize / 1024 / 1024).toFixed(2);\n const msg = context\n ? `${context}: Size ${actualMB}MB exceeds limit ${maxMB}MB`\n : `Size ${actualMB}MB exceeds limit ${maxMB}MB`;\n super(msg, 'SIZE_LIMIT_EXCEEDED');\n this.name = 'SizeLimitError';\n }\n}\n\n/**\n * Path traversal or unsafe path detected\n */\nexport class PathTraversalError extends FoundryError {\n constructor(public readonly path: string) {\n super(`Unsafe path detected: ${path}`, 'PATH_TRAVERSAL');\n this.name = 'PathTraversalError';\n }\n}\n\n/**\n * Export operation would lose data\n */\nexport class DataLossError extends FoundryError {\n constructor(\n public readonly lostFields: string[],\n public readonly targetFormat: string\n ) {\n const fields = lostFields.slice(0, 3).join(', ');\n const more = lostFields.length > 3 ? ` and ${lostFields.length - 3} more` : '';\n super(\n `Export to ${targetFormat} would lose: ${fields}${more}`,\n 'DATA_LOSS'\n );\n this.name = 'DataLossError';\n }\n}\n\n/**\n * Check if an error is a FoundryError\n *\n * Uses Symbol.for() marker instead of instanceof to handle dual ESM/CJS package loading.\n * In dual-package environments, instanceof can fail if the error comes from a different\n * module instance (e.g., ESM vs CJS version of the same package). Symbol.for() creates\n * a global symbol shared across all module instances.\n */\nexport function isFoundryError(error: unknown): error is FoundryError {\n return (\n error instanceof Error &&\n FOUNDRY_ERROR_MARKER in error &&\n (error as Record<symbol, unknown>)[FOUNDRY_ERROR_MARKER] === true\n );\n}\n\n/**\n * Wrap unknown errors in a FoundryError\n */\nexport function wrapError(error: unknown, context?: string): FoundryError {\n if (isFoundryError(error)) {\n return error;\n }\n\n const message = error instanceof Error\n ? error.message\n : String(error);\n\n return new FoundryError(\n context ? `${context}: ${message}` : message,\n 'UNKNOWN_ERROR'\n );\n}\n","/**\n * Data URL Utilities\n *\n * Convert between Uint8Array buffers and data URLs.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n */\n\nimport type { BinaryData } from './binary.js';\nimport { encodeChunked as base64Encode, decode as base64Decode } from './base64.js';\nimport { ValidationError } from './errors.js';\n\n/**\n * Convert Uint8Array to data URL.\n * Handles large buffers (>10MB) without stack overflow by processing in chunks.\n *\n * @param buffer - Binary data to encode\n * @param mimeType - MIME type for the data URL (e.g., 'image/png', 'application/octet-stream')\n * @returns Data URL string\n *\n * @example\n * ```typescript\n * const png = new Uint8Array([...]);\n * const dataUrl = toDataURL(png, 'image/png');\n * // => \"data:image/png;base64,iVBORw0KGgo...\"\n * ```\n */\nexport function toDataURL(buffer: BinaryData, mimeType: string): string {\n // Use chunked encoding to handle large buffers without stack overflow\n const base64 = base64Encode(buffer);\n return `data:${mimeType};base64,${base64}`;\n}\n\n/**\n * Parse a data URL back to buffer and MIME type.\n * Validates the data URL format before parsing.\n *\n * @param dataUrl - Data URL string to parse\n * @returns Object containing the decoded buffer and MIME type\n * @throws Error if the data URL format is invalid\n *\n * @example\n * ```typescript\n * const { buffer, mimeType } = fromDataURL('data:image/png;base64,iVBORw0KGgo...');\n * // buffer: Uint8Array\n * // mimeType: 'image/png'\n * ```\n */\nexport function fromDataURL(dataUrl: string): { buffer: Uint8Array; mimeType: string } {\n // Validate data URL format\n if (!dataUrl.startsWith('data:')) {\n throw new ValidationError('Invalid data URL: must start with \"data:\"', 'dataUrl');\n }\n\n const commaIndex = dataUrl.indexOf(',');\n if (commaIndex === -1) {\n throw new ValidationError('Invalid data URL: missing comma separator', 'dataUrl');\n }\n\n const header = dataUrl.slice(5, commaIndex); // Skip 'data:'\n const data = dataUrl.slice(commaIndex + 1);\n\n // Parse header: [<mediatype>][;base64]\n let mimeType = 'text/plain';\n let isBase64 = false;\n\n const parts = header.split(';');\n for (const part of parts) {\n if (part === 'base64') {\n isBase64 = true;\n } else if (part && !part.includes('=')) {\n // MIME type (not a parameter like charset=utf-8)\n mimeType = part;\n }\n }\n\n if (!isBase64) {\n // URL-encoded text data\n throw new ValidationError('Non-base64 data URLs are not supported', 'dataUrl');\n }\n\n const buffer = base64Decode(data);\n return { buffer, mimeType };\n}\n\n/**\n * Check if a string is a valid data URL\n *\n * @param str - String to check\n * @returns true if the string is a valid data URL format\n */\nexport function isDataURL(str: string): boolean {\n if (!str.startsWith('data:')) return false;\n const commaIndex = str.indexOf(',');\n if (commaIndex === -1) return false;\n const header = str.slice(5, commaIndex);\n return header.includes('base64');\n}\n","/**\n * URI Utilities\n *\n * Handles different asset URI schemes used in character cards.\n * Supports: embeded://, embedded://, ccdefault:, https://, http://,\n * data:, file://, __asset:, asset:, chara-ext-asset_\n */\n\nexport type URIScheme =\n | 'embeded' // embeded:// (CharX standard, note intentional typo)\n | 'ccdefault' // ccdefault:\n | 'https' // https://\n | 'http' // http://\n | 'data' // data:mime;base64,...\n | 'file' // file://\n | 'internal' // Internal asset ID (UUID/string)\n | 'pngchunk' // PNG chunk reference (__asset:, asset:, chara-ext-asset_)\n | 'unknown';\n\nexport interface ParsedURI {\n scheme: URIScheme;\n originalUri: string;\n normalizedUri: string; // Normalized form of the URI\n path?: string; // For embeded://, file://\n url?: string; // For http://, https://\n data?: string; // For data: URIs\n mimeType?: string; // For data: URIs\n encoding?: string; // For data: URIs (e.g., base64)\n chunkKey?: string; // For pngchunk - the key/index to look up\n chunkCandidates?: string[]; // For pngchunk - all possible chunk keys to search\n}\n\n/**\n * Normalize a URI to its canonical form\n * Handles common typos and variant formats\n */\nexport function normalizeURI(uri: string): string {\n const trimmed = uri.trim();\n\n // Fix embedded:// -> embeded:// (common typo, CharX spec uses single 'd')\n if (trimmed.startsWith('embedded://')) {\n return 'embeded://' + trimmed.substring('embedded://'.length);\n }\n\n // Normalize PNG chunk references to pngchunk: scheme\n if (trimmed.startsWith('__asset:')) {\n const id = trimmed.substring('__asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('asset:')) {\n const id = trimmed.substring('asset:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_:')) {\n const id = trimmed.substring('chara-ext-asset_:'.length);\n return `pngchunk:${id}`;\n }\n if (trimmed.startsWith('chara-ext-asset_')) {\n const id = trimmed.substring('chara-ext-asset_'.length);\n return `pngchunk:${id}`;\n }\n\n return trimmed;\n}\n\n/**\n * Parse a URI and determine its scheme and components\n */\nexport function parseURI(uri: string): ParsedURI {\n const trimmed = uri.trim();\n const normalized = normalizeURI(trimmed);\n\n // PNG chunk references (__asset:, asset:, chara-ext-asset_, pngchunk:)\n if (\n trimmed.startsWith('__asset:') ||\n trimmed.startsWith('asset:') ||\n trimmed.startsWith('chara-ext-asset_') ||\n trimmed.startsWith('pngchunk:')\n ) {\n let assetId: string;\n if (trimmed.startsWith('__asset:')) {\n assetId = trimmed.substring('__asset:'.length);\n } else if (trimmed.startsWith('asset:')) {\n assetId = trimmed.substring('asset:'.length);\n } else if (trimmed.startsWith('chara-ext-asset_:')) {\n assetId = trimmed.substring('chara-ext-asset_:'.length);\n } else if (trimmed.startsWith('pngchunk:')) {\n assetId = trimmed.substring('pngchunk:'.length);\n } else {\n assetId = trimmed.substring('chara-ext-asset_'.length);\n }\n\n // Generate all possible chunk key variations for lookup\n const candidates = [\n assetId, // \"0\" or \"filename.png\"\n trimmed, // Original URI\n `asset:${assetId}`, // \"asset:0\"\n `__asset:${assetId}`, // \"__asset:0\"\n `__asset_${assetId}`, // \"__asset_0\"\n `chara-ext-asset_${assetId}`, // \"chara-ext-asset_0\"\n `chara-ext-asset_:${assetId}`, // \"chara-ext-asset_:0\"\n `pngchunk:${assetId}`, // \"pngchunk:0\"\n ];\n\n return {\n scheme: 'pngchunk',\n originalUri: uri,\n normalizedUri: normalized,\n chunkKey: assetId,\n chunkCandidates: candidates,\n };\n }\n\n // ccdefault: - use default asset\n if (trimmed === 'ccdefault:' || trimmed.startsWith('ccdefault:')) {\n return {\n scheme: 'ccdefault',\n originalUri: uri,\n normalizedUri: normalized,\n };\n }\n\n // embeded:// or embedded:// (normalize typo)\n if (trimmed.startsWith('embeded://') || trimmed.startsWith('embedded://')) {\n const path = trimmed.startsWith('embeded://')\n ? trimmed.substring('embeded://'.length)\n : trimmed.substring('embedded://'.length);\n return {\n scheme: 'embeded',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // https://\n if (trimmed.startsWith('https://')) {\n return {\n scheme: 'https',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // http://\n if (trimmed.startsWith('http://')) {\n return {\n scheme: 'http',\n originalUri: uri,\n normalizedUri: normalized,\n url: trimmed,\n };\n }\n\n // data: URIs\n if (trimmed.startsWith('data:')) {\n const parsed = parseDataURI(trimmed);\n return {\n scheme: 'data',\n originalUri: uri,\n normalizedUri: normalized,\n ...parsed,\n };\n }\n\n // file://\n if (trimmed.startsWith('file://')) {\n const path = trimmed.substring('file://'.length);\n return {\n scheme: 'file',\n originalUri: uri,\n normalizedUri: normalized,\n path,\n };\n }\n\n // Internal asset ID (alphanumeric/UUID format)\n if (/^[a-zA-Z0-9_-]+$/.test(trimmed)) {\n return {\n scheme: 'internal',\n originalUri: uri,\n normalizedUri: normalized,\n path: trimmed,\n };\n }\n\n // Unknown scheme\n return {\n scheme: 'unknown',\n originalUri: uri,\n normalizedUri: normalized,\n };\n}\n\n/**\n * Parse a data URI into its components\n * Format: data:[<mediatype>][;base64],<data>\n */\nfunction parseDataURI(uri: string): { mimeType?: string; encoding?: string; data?: string } {\n const match = uri.match(/^data:([^;,]+)?(;base64)?,(.*)$/);\n\n if (!match) {\n return {};\n }\n\n return {\n mimeType: match[1] || 'text/plain',\n encoding: match[2] ? 'base64' : undefined,\n data: match[3],\n };\n}\n\n/**\n * Check if extension is an image format\n */\nexport function isImageExt(ext: string): boolean {\n const imageExts = ['png', 'jpg', 'jpeg', 'webp', 'gif', 'avif', 'bmp', 'svg'];\n return imageExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is an audio format\n */\nexport function isAudioExt(ext: string): boolean {\n const audioExts = ['mp3', 'wav', 'ogg', 'flac', 'm4a', 'aac'];\n return audioExts.includes(ext.toLowerCase());\n}\n\n/**\n * Check if extension is a video format\n */\nexport function isVideoExt(ext: string): boolean {\n const videoExts = ['mp4', 'webm', 'avi', 'mov', 'mkv'];\n return videoExts.includes(ext.toLowerCase());\n}\n\n/** Safe MIME types for data: URIs that can be used in href/src */\nconst SAFE_DATA_URI_MIME_TYPES = new Set([\n // Images (safe for img src)\n 'image/png',\n 'image/jpeg',\n 'image/gif',\n 'image/webp',\n 'image/avif',\n 'image/bmp',\n 'image/x-icon',\n // Audio (safe for audio src)\n 'audio/mpeg',\n 'audio/wav',\n 'audio/ogg',\n 'audio/flac',\n 'audio/mp4',\n 'audio/aac',\n // Video (safe for video src)\n 'video/mp4',\n 'video/webm',\n // Text/data (generally safe)\n 'text/plain',\n 'application/json',\n 'application/octet-stream',\n]);\n\n/** Potentially dangerous MIME types that should NOT be used in href/src */\nconst DANGEROUS_DATA_URI_MIME_TYPES = new Set([\n // Executable/script content\n 'text/html',\n 'text/javascript',\n 'application/javascript',\n 'application/x-javascript',\n 'text/css',\n 'image/svg+xml', // SVG can contain scripts\n 'application/xhtml+xml',\n 'application/xml',\n]);\n\n/**\n * Options for URI safety validation\n */\nexport interface URISafetyOptions {\n /** Allow http:// URIs (default: false) */\n allowHttp?: boolean;\n /** Allow file:// URIs (default: false) */\n allowFile?: boolean;\n /**\n * Allowed MIME types for data: URIs (default: all safe types).\n * Set to empty array to reject all data: URIs.\n * Set to undefined to use default safe list.\n */\n allowedDataMimes?: string[];\n}\n\n/**\n * Result of URI safety check with detailed information\n */\nexport interface URISafetyResult {\n /** Whether the URI is safe to use */\n safe: boolean;\n /** Reason if unsafe */\n reason?: string;\n /** Detected scheme */\n scheme: URIScheme;\n /** MIME type for data: URIs */\n mimeType?: string;\n}\n\n/**\n * Validate if a URI is safe to use (detailed version)\n *\n * @param uri - URI to validate\n * @param options - Safety options\n * @returns Detailed safety result\n */\nexport function checkURISafety(uri: string, options: URISafetyOptions = {}): URISafetyResult {\n const parsed = parseURI(uri);\n\n switch (parsed.scheme) {\n case 'embeded':\n case 'ccdefault':\n case 'internal':\n case 'https':\n case 'pngchunk':\n return { safe: true, scheme: parsed.scheme };\n\n case 'data': {\n const mimeType = parsed.mimeType || 'text/plain';\n\n // Check for explicitly dangerous MIME types\n if (DANGEROUS_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI with potentially dangerous MIME type: ${mimeType}`,\n };\n }\n\n // If custom allowed list is provided, check against it\n if (options.allowedDataMimes !== undefined) {\n if (options.allowedDataMimes.length === 0) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: 'Data URIs are not allowed',\n };\n }\n if (!options.allowedDataMimes.includes(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Data URI MIME type not in allowed list: ${mimeType}`,\n };\n }\n }\n\n // Otherwise use default safe list\n if (!SAFE_DATA_URI_MIME_TYPES.has(mimeType)) {\n return {\n safe: false,\n scheme: parsed.scheme,\n mimeType,\n reason: `Unknown data URI MIME type: ${mimeType}`,\n };\n }\n\n return { safe: true, scheme: parsed.scheme, mimeType };\n }\n\n case 'http':\n if (options.allowHttp === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'HTTP URIs are not allowed' };\n\n case 'file':\n if (options.allowFile === true) {\n return { safe: true, scheme: parsed.scheme };\n }\n return { safe: false, scheme: parsed.scheme, reason: 'File URIs are not allowed' };\n\n case 'unknown':\n default:\n return { safe: false, scheme: parsed.scheme, reason: 'Unknown URI scheme' };\n }\n}\n\n/**\n * Validate if a URI is safe to use (simple boolean version for backwards compatibility)\n *\n * @deprecated Use checkURISafety() for detailed safety information\n */\nexport function isURISafe(uri: string, options: { allowHttp?: boolean; allowFile?: boolean } = {}): boolean {\n return checkURISafety(uri, options).safe;\n}\n\n/**\n * Extract file extension from URI\n */\nexport function getExtensionFromURI(uri: string): string {\n const parsed = parseURI(uri);\n\n if (parsed.path) {\n const parts = parsed.path.split('.');\n if (parts.length > 1) {\n return parts[parts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.url) {\n const urlParts = parsed.url.split('?')[0]!.split('.');\n if (urlParts.length > 1) {\n return urlParts[urlParts.length - 1]!.toLowerCase();\n }\n }\n\n if (parsed.mimeType) {\n return getExtFromMimeType(parsed.mimeType);\n }\n\n return 'unknown';\n}\n\n/**\n * Get MIME type from file extension\n */\nexport function getMimeTypeFromExt(ext: string): string {\n const extToMime: Record<string, string> = {\n // Images\n 'png': 'image/png',\n 'jpg': 'image/jpeg',\n 'jpeg': 'image/jpeg',\n 'webp': 'image/webp',\n 'gif': 'image/gif',\n 'avif': 'image/avif',\n 'svg': 'image/svg+xml',\n 'bmp': 'image/bmp',\n 'ico': 'image/x-icon',\n\n // Audio\n 'mp3': 'audio/mpeg',\n 'wav': 'audio/wav',\n 'ogg': 'audio/ogg',\n 'flac': 'audio/flac',\n 'm4a': 'audio/mp4',\n 'aac': 'audio/aac',\n\n // Video\n 'mp4': 'video/mp4',\n 'webm': 'video/webm',\n 'avi': 'video/x-msvideo',\n 'mov': 'video/quicktime',\n 'mkv': 'video/x-matroska',\n\n // Text/Data\n 'json': 'application/json',\n 'txt': 'text/plain',\n 'html': 'text/html',\n 'css': 'text/css',\n 'js': 'application/javascript',\n };\n\n return extToMime[ext.toLowerCase()] || 'application/octet-stream';\n}\n\n/**\n * Get file extension from MIME type\n */\nexport function getExtFromMimeType(mimeType: string): string {\n const mimeToExt: Record<string, string> = {\n 'image/png': 'png',\n 'image/jpeg': 'jpg',\n 'image/webp': 'webp',\n 'image/gif': 'gif',\n 'image/avif': 'avif',\n 'image/svg+xml': 'svg',\n 'image/bmp': 'bmp',\n 'image/x-icon': 'ico',\n 'audio/mpeg': 'mp3',\n 'audio/wav': 'wav',\n 'audio/ogg': 'ogg',\n 'audio/flac': 'flac',\n 'audio/mp4': 'm4a',\n 'audio/aac': 'aac',\n 'video/mp4': 'mp4',\n 'video/webm': 'webm',\n 'video/x-msvideo': 'avi',\n 'video/quicktime': 'mov',\n 'video/x-matroska': 'mkv',\n 'application/json': 'json',\n 'text/plain': 'txt',\n 'text/html': 'html',\n 'text/css': 'css',\n 'application/javascript': 'js',\n };\n\n return mimeToExt[mimeType] || 'bin';\n}\n\n/**\n * Build a data URI from binary data and MIME type\n */\nexport function buildDataURI(data: string, mimeType: string, isBase64 = true): string {\n if (isBase64) {\n return `data:${mimeType};base64,${data}`;\n }\n return `data:${mimeType},${encodeURIComponent(data)}`;\n}\n","/**\n * Image Analysis Utilities\n *\n * Detect properties of image files from binary data.\n */\n\nimport {\n type BinaryData,\n indexOf,\n fromLatin1,\n} from './binary.js';\n\n/**\n * Check if an image buffer contains animation data.\n * Supports: APNG, WebP (Animated), GIF\n */\nexport function isAnimatedImage(data: BinaryData, _mimeType?: string): boolean {\n // 1. WebP Detection\n // RIFF .... WEBP\n if (\n data.length > 12 &&\n data[0] === 0x52 && data[1] === 0x49 && data[2] === 0x46 && data[3] === 0x46 && // RIFF\n data[8] === 0x57 && data[9] === 0x45 && data[10] === 0x42 && data[11] === 0x50 // WEBP\n ) {\n // Check for VP8X chunk\n // VP8X chunk header: 'VP8X' (bytes 12-15)\n if (\n data[12] === 0x56 && data[13] === 0x50 && data[14] === 0x38 && data[15] === 0x58\n ) {\n // Flags byte is at offset 20 (16 + 4 bytes chunk size)\n // Animation bit is bit 1 (0x02)\n const flags = data[20];\n return (flags! & 0x02) !== 0;\n }\n return false;\n }\n\n // 2. PNG/APNG Detection\n // Signature: 89 50 4E 47 0D 0A 1A 0A\n if (\n data.length > 8 &&\n data[0] === 0x89 && data[1] === 0x50 && data[2] === 0x4E && data[3] === 0x47\n ) {\n // Search for 'acTL' chunk (Animation Control)\n // It must appear before IDAT.\n // Simple search: indexOf('acTL')\n // Note: theoretically 'acTL' string could appear in other data, but highly unlikely in valid PNG structure before IDAT\n // We can iterate chunks to be safe, but indexOf is faster for a quick check\n const actlSig = fromLatin1('acTL');\n const idatSig = fromLatin1('IDAT');\n \n const actlIndex = indexOf(data, actlSig);\n if (actlIndex === -1) return false;\n\n const idatIndex = indexOf(data, idatSig);\n // If acTL exists and is before the first IDAT (or IDAT not found yet), it's APNG\n return idatIndex === -1 || actlIndex < idatIndex;\n }\n\n // 3. GIF Detection\n // Signature: GIF87a or GIF89a\n if (\n data.length > 6 &&\n data[0] === 0x47 && data[1] === 0x49 && data[2] === 0x46 // GIF\n ) {\n // Check for NETSCAPE2.0 extension (looping animation)\n // This is a heuristic. Static GIFs are rare in this domain but possible.\n // Full frame counting is expensive. Presence of NETSCAPE block is a strong indicator.\n const netscape = fromLatin1('NETSCAPE2.0');\n return indexOf(data, netscape) !== -1;\n }\n\n return false;\n}\n","/**\n * UUID Generation Utilities\n *\n * Provides crypto-grade UUID v4 generation that works in Node.js,\n * browsers (secure contexts), and falls back gracefully.\n */\n\n/**\n * Format 16 random bytes as a UUID v4 string\n */\nfunction formatUUID(bytes: Uint8Array): string {\n const hex = Array.from(bytes, (b) => b.toString(16).padStart(2, '0')).join('');\n return `${hex.slice(0, 8)}-${hex.slice(8, 12)}-${hex.slice(12, 16)}-${hex.slice(16, 20)}-${hex.slice(20, 32)}`;\n}\n\n/**\n * Fallback UUID generation using Math.random()\n * Only used when crypto APIs are unavailable (rare)\n */\nfunction mathRandomUUID(): string {\n return 'xxxxxxxx-xxxx-4xxx-yxxx-xxxxxxxxxxxx'.replace(/[xy]/g, (c) => {\n const r = (Math.random() * 16) | 0;\n const v = c === 'x' ? r : (r & 0x3) | 0x8;\n return v.toString(16);\n });\n}\n\n/**\n * Generate a cryptographically secure UUID v4.\n *\n * Uses crypto.randomUUID() when available (Node.js 19+, modern browsers).\n * Falls back to crypto.getRandomValues() if randomUUID is unavailable.\n * Last resort uses Math.random() (non-secure, emits warning in dev).\n *\n * @returns A valid RFC 4122 UUID v4 string\n *\n * @example\n * ```typescript\n * const id = generateUUID();\n * // => \"550e8400-e29b-41d4-a716-446655440000\"\n * ```\n */\nexport function generateUUID(): string {\n // Node.js 19+ or browser with secure context\n if (typeof crypto !== 'undefined' && typeof crypto.randomUUID === 'function') {\n return crypto.randomUUID();\n }\n\n // Fallback using crypto.getRandomValues (older Node/browsers)\n if (typeof crypto !== 'undefined' && typeof crypto.getRandomValues === 'function') {\n const bytes = new Uint8Array(16);\n crypto.getRandomValues(bytes);\n // Set version (4) and variant (RFC 4122)\n bytes[6] = (bytes[6]! & 0x0f) | 0x40; // Version 4\n bytes[8] = (bytes[8]! & 0x3f) | 0x80; // Variant 1\n return formatUUID(bytes);\n }\n\n // Last resort - non-secure fallback\n if (typeof process !== 'undefined' && process.env?.NODE_ENV === 'development') {\n console.warn('[character-foundry/core] generateUUID: Using insecure Math.random() fallback');\n }\n return mathRandomUUID();\n}\n\n/**\n * Validate if a string is a valid UUID v4\n *\n * @param uuid - String to validate\n * @returns true if valid UUID v4 format\n */\nexport function isValidUUID(uuid: string): boolean {\n return /^[0-9a-f]{8}-[0-9a-f]{4}-4[0-9a-f]{3}-[89ab][0-9a-f]{3}-[0-9a-f]{12}$/i.test(uuid);\n}\n","/**\n * CCv2 to CCv3 Converter\n *\n * Converts CCv2 card data to CCv3 format.\n * Handles malformed cards gracefully by defaulting missing fields.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv2LorebookEntry,\n CCv3Data,\n CCv3CharacterBook,\n CCv3LorebookEntry,\n} from '@character-foundry/schemas';\nimport { getV2Data } from '@character-foundry/schemas';\n\n/**\n * Normalize position field - accepts both string literals and SillyTavern numeric values.\n * SillyTavern uses: 0 = before_char, 1 = after_char, other numbers passed through.\n */\nfunction normalizePosition(\n position: 'before_char' | 'after_char' | 'in_chat' | '' | number | null | undefined\n): 'before_char' | 'after_char' | 'in_chat' | number {\n if (position === undefined || position === null || position === '') return 'before_char';\n if (typeof position === 'string') return position;\n // SillyTavern numeric mapping - pass through as-is since schema now accepts numbers\n return position;\n}\n\n/**\n * Convert CCv2 lorebook entry to CCv3 format\n */\nfunction convertLorebookEntry(\n entry: CCv2LorebookEntry,\n index: number\n): CCv3LorebookEntry {\n return {\n keys: entry.keys || [],\n content: entry.content || '',\n extensions: entry.extensions || {},\n enabled: entry.enabled ?? true,\n insertion_order: entry.insertion_order ?? index,\n case_sensitive: entry.case_sensitive,\n name: entry.name || '',\n priority: entry.priority ?? 10,\n id: entry.id ?? index,\n comment: entry.comment || '',\n selective: entry.selective ?? false,\n secondary_keys: entry.secondary_keys || [],\n constant: entry.constant ?? false,\n position: normalizePosition(entry.position),\n };\n}\n\n/**\n * Convert CCv2 character book to CCv3 format\n */\nfunction convertCharacterBook(\n book: NonNullable<CCv2Data['character_book']>\n): CCv3CharacterBook {\n return {\n name: book.name,\n description: book.description,\n scan_depth: book.scan_depth,\n token_budget: book.token_budget,\n recursive_scanning: book.recursive_scanning,\n extensions: book.extensions,\n entries: book.entries.map((entry, i) => convertLorebookEntry(entry, i)),\n };\n}\n\n/**\n * Convert CCv2 card to CCv3 format.\n *\n * Philosophy: Be lenient on input - never completely drop data due to missing fields.\n * Defaults are applied for any missing required fields to ensure valid output.\n * Pure 1:1 field mapping - no extraction, no magic.\n *\n * @param input - CCv2 card data (wrapped or unwrapped, potentially malformed)\n * @returns CCv3 card data with all required fields populated\n */\nexport function ccv2ToCCv3(input: CCv2Data | CCv2Wrapped): CCv3Data {\n const data = getV2Data(input);\n\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: {\n // Core required fields - default to empty string if missing/undefined\n name: data.name ?? '',\n description: data.description ?? '',\n personality: data.personality ?? '',\n scenario: data.scenario ?? '',\n first_mes: data.first_mes ?? '',\n mes_example: data.mes_example ?? '',\n // Optional in V2, required in V3 - always provide defaults\n creator_notes: data.creator_notes || '',\n system_prompt: data.system_prompt || '',\n post_history_instructions: data.post_history_instructions || '',\n alternate_greetings: data.alternate_greetings || [],\n group_only_greetings: [],\n tags: data.tags || [],\n creator: data.creator || '',\n character_version: data.character_version || '',\n character_book: data.character_book\n ? convertCharacterBook(data.character_book)\n : undefined,\n extensions: data.extensions || {},\n },\n };\n}\n","/**\n * CCv3 to CCv2 Converter\n *\n * Converts CCv3 card data to CCv2 format.\n * Note: Some v3 features will be lost (group_only_greetings, etc.)\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv2CharacterBook,\n CCv2LorebookEntry,\n CCv3Data,\n CCv3LorebookEntry,\n} from '@character-foundry/schemas';\n\n/**\n * Fields lost when converting v3 to v2\n */\nexport const V3_TO_V2_LOST_FIELDS = [\n 'group_only_greetings',\n 'assets',\n];\n\n/**\n * Convert CCv3 lorebook entry to CCv2 format\n */\nfunction convertLorebookEntry(entry: CCv3LorebookEntry): CCv2LorebookEntry {\n return {\n keys: entry.keys,\n content: entry.content,\n extensions: entry.extensions || {},\n enabled: entry.enabled,\n insertion_order: entry.insertion_order,\n case_sensitive: entry.case_sensitive,\n name: entry.name,\n priority: entry.priority,\n id: entry.id,\n comment: entry.comment,\n selective: entry.selective,\n secondary_keys: entry.secondary_keys,\n constant: entry.constant,\n position: entry.position,\n };\n}\n\n/**\n * Convert CCv3 character book to CCv2 format\n */\nfunction convertCharacterBook(\n book: NonNullable<CCv3Data['data']['character_book']>\n): CCv2CharacterBook {\n return {\n name: book.name,\n description: book.description,\n scan_depth: book.scan_depth,\n token_budget: book.token_budget,\n recursive_scanning: book.recursive_scanning,\n extensions: book.extensions,\n entries: book.entries.map(convertLorebookEntry),\n };\n}\n\n/**\n * Convert CCv3 card to CCv2 unwrapped format\n *\n * @param card - CCv3 card data\n * @returns CCv2 unwrapped card data\n */\nexport function ccv3ToCCv2Data(card: CCv3Data): CCv2Data {\n const data = card.data;\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality,\n scenario: data.scenario,\n first_mes: data.first_mes,\n mes_example: data.mes_example,\n creator_notes: data.creator_notes,\n system_prompt: data.system_prompt,\n post_history_instructions: data.post_history_instructions,\n alternate_greetings: data.alternate_greetings,\n tags: data.tags,\n creator: data.creator,\n character_version: data.character_version,\n character_book: data.character_book\n ? convertCharacterBook(data.character_book)\n : undefined,\n extensions: data.extensions,\n };\n}\n\n/**\n * Convert CCv3 card to CCv2 wrapped format\n *\n * @param card - CCv3 card data\n * @returns CCv2 wrapped card data\n */\nexport function ccv3ToCCv2Wrapped(card: CCv3Data): CCv2Wrapped {\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: ccv3ToCCv2Data(card),\n };\n}\n\n/**\n * Check what would be lost converting v3 to v2\n */\nexport function checkV3ToV2Loss(card: CCv3Data): string[] {\n const lost: string[] = [];\n\n if (card.data.group_only_greetings && card.data.group_only_greetings.length > 0) {\n lost.push(`group_only_greetings (${card.data.group_only_greetings.length} entries)`);\n }\n\n if (card.data.assets && card.data.assets.length > 0) {\n lost.push(`assets (${card.data.assets.length} entries)`);\n }\n\n return lost;\n}\n","/**\n * To NormalizedCard Converter\n *\n * Converts any card format to NormalizedCard.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv3Data,\n NormalizedCard,\n} from '@character-foundry/schemas';\nimport { getV2Data, isWrappedV2, isV3Card } from '@character-foundry/schemas';\nimport { FormatNotSupportedError } from '@character-foundry/core';\n\n/**\n * Normalize CCv2 data to NormalizedCard\n */\nexport function normalizeV2(input: CCv2Data | CCv2Wrapped): NormalizedCard {\n const data = getV2Data(input);\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality ?? '', // Coerce null to empty string\n scenario: data.scenario,\n firstMes: data.first_mes,\n mesExample: data.mes_example ?? '', // Coerce null to empty string\n systemPrompt: data.system_prompt,\n postHistoryInstructions: data.post_history_instructions,\n creatorNotes: data.creator_notes,\n alternateGreetings: data.alternate_greetings || [],\n groupOnlyGreetings: [],\n tags: data.tags || [],\n creator: data.creator,\n characterVersion: data.character_version,\n characterBook: data.character_book ?? undefined,\n extensions: data.extensions || {},\n };\n}\n\n/**\n * Normalize CCv3 data to NormalizedCard\n */\nexport function normalizeV3(card: CCv3Data): NormalizedCard {\n const data = card.data;\n\n return {\n name: data.name,\n description: data.description,\n personality: data.personality ?? '', // Coerce null to empty string\n scenario: data.scenario,\n firstMes: data.first_mes,\n mesExample: data.mes_example ?? '', // Coerce null to empty string\n systemPrompt: data.system_prompt,\n postHistoryInstructions: data.post_history_instructions,\n creatorNotes: data.creator_notes,\n alternateGreetings: data.alternate_greetings || [],\n groupOnlyGreetings: data.group_only_greetings || [],\n tags: data.tags || [],\n creator: data.creator,\n characterVersion: data.character_version,\n characterBook: data.character_book ?? undefined,\n extensions: data.extensions || {},\n };\n}\n\n/**\n * Normalize any card format to NormalizedCard\n *\n * @param input - Card data in any supported format\n * @returns NormalizedCard\n */\nexport function normalize(input: CCv2Data | CCv2Wrapped | CCv3Data): NormalizedCard {\n if (isV3Card(input)) {\n return normalizeV3(input as CCv3Data);\n }\n\n if (isWrappedV2(input) || 'name' in input) {\n return normalizeV2(input as CCv2Data | CCv2Wrapped);\n }\n\n throw new FormatNotSupportedError('unknown', 'normalization');\n}\n","/**\n * From NormalizedCard Converter\n *\n * Converts NormalizedCard to other formats.\n */\n\nimport type {\n CCv2Data,\n CCv2Wrapped,\n CCv3Data,\n NormalizedCard,\n} from '@character-foundry/schemas';\n\n/**\n * Convert NormalizedCard to CCv3\n */\nexport function denormalizeToV3(card: NormalizedCard): CCv3Data {\n return {\n spec: 'chara_card_v3',\n spec_version: '3.0',\n data: {\n name: card.name,\n description: card.description,\n personality: card.personality,\n scenario: card.scenario,\n first_mes: card.firstMes,\n mes_example: card.mesExample,\n creator_notes: card.creatorNotes || '',\n system_prompt: card.systemPrompt || '',\n post_history_instructions: card.postHistoryInstructions || '',\n alternate_greetings: card.alternateGreetings,\n group_only_greetings: card.groupOnlyGreetings,\n tags: card.tags,\n creator: card.creator || '',\n character_version: card.characterVersion || '',\n character_book: card.characterBook,\n extensions: card.extensions,\n },\n };\n}\n\n/**\n * Convert character book to v2 format (ensure extensions are not undefined)\n */\nfunction convertCharacterBookToV2(book: NormalizedCard['characterBook']): CCv2Data['character_book'] {\n if (!book) return undefined;\n\n return {\n ...book,\n entries: book.entries.map((entry) => ({\n ...entry,\n extensions: entry.extensions || {},\n })),\n };\n}\n\n/**\n * Convert NormalizedCard to CCv2 unwrapped\n *\n * Note: group_only_greetings will be lost\n */\nexport function denormalizeToV2Data(card: NormalizedCard): CCv2Data {\n return {\n name: card.name,\n description: card.description,\n personality: card.personality,\n scenario: card.scenario,\n first_mes: card.firstMes,\n mes_example: card.mesExample,\n creator_notes: card.creatorNotes,\n system_prompt: card.systemPrompt,\n post_history_instructions: card.postHistoryInstructions,\n alternate_greetings: card.alternateGreetings,\n tags: card.tags,\n creator: card.creator,\n character_version: card.characterVersion,\n character_book: convertCharacterBookToV2(card.characterBook),\n extensions: card.extensions,\n };\n}\n\n/**\n * Convert NormalizedCard to CCv2 wrapped\n *\n * Note: group_only_greetings will be lost\n */\nexport function denormalizeToV2Wrapped(card: NormalizedCard): CCv2Wrapped {\n return {\n spec: 'chara_card_v2',\n spec_version: '2.0',\n data: denormalizeToV2Data(card),\n };\n}\n\n/**\n * Check what would be lost converting NormalizedCard to v2\n */\nexport function checkNormalizedToV2Loss(card: NormalizedCard): string[] {\n const lost: string[] = [];\n\n if (card.groupOnlyGreetings.length > 0) {\n lost.push(`group_only_greetings (${card.groupOnlyGreetings.length} entries)`);\n }\n\n return lost;\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACMA,iBAAkB;ACAlB,IAAAA,cAAkB;ACAlB,IAAAA,cAAkB;AMAlB,IAAAA,cAAkB;ARSX,IAAM,gBAAgB,aAAE,OAAO,EAAE,SAAS;AAK1C,IAAM,aAAa,aAAE,OAAO,EAAE,KAAK;AAKnC,IAAM,aAAa,aAAE,KAAK,CAAC,MAAM,IAAI,CAAC;AAKtC,IAAM,qBAAqB,aAAE,KAAK;EACvC;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;EACA;;AACF,CAAC;AAKM,IAAM,sBAAsB,aAAE,KAAK,CAAC,WAAW,aAAa,QAAQ,CAAC;AAKrE,IAAM,kBAAkB,aAAE,KAAK;EACpC;EACA;EACA;EACA;EACA;EACA;EACA;EACA;AACF,CAAC;AAKM,IAAM,wBAAwB,aAAE,OAAO;EAC5C,MAAM;EACN,KAAK,aAAE,OAAO;EACd,MAAM,aAAE,OAAO;EACf,KAAK,aAAE,OAAO;AAChB,CAAC;AAKM,IAAM,uBAAuB,aAAE,OAAO;EAC3C,YAAY;EACZ,MAAM,aAAE,WAAW,UAAU;EAC7B,UAAU,aAAE,OAAO;AACrB,CAAC;AC9DM,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;AAC7H,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;;EAErC,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACnC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC7B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACvC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;AAC7C,CAAC;AAKM,IAAM,oBAAoBA,YAAAA,EAAE,OAAO;EACxC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AAiCM,SAAS,YAAY,MAAoC;AAC9D,SAAO,kBAAkB,UAAU,IAAI,EAAE;AAC3C;AA8BO,SAAS,mBAAmB,MAAwE;AACzG,MAAI,CAAC,QAAQ,OAAO,SAAS,SAAU,QAAO;AAC9C,QAAM,MAAM;AACZ,SACE,IAAI,SAAS,mBACb,IAAI,SAAS,QACb,OAAO,IAAI,SAAS;AAExB;AASO,SAAS,UAAU,MAAwC;AAEhE,MAAI,mBAAmB,IAAI,GAAG;AAC5B,WAAO,KAAK;EACd;AACA,SAAO;AACT;ACpJO,IAAM,0BAA0BC,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;;EACnC,SAASA,YAAAA,EAAE,OAAO;EAClB,SAASA,YAAAA,EAAE,QAAQ,EAAE,QAAQ,IAAI;;EACjC,iBAAiBA,YAAAA,EAAE,WAAW,CAAC,MAAM,KAAK,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC;;EAE7D,gBAAgBA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAChD,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,UAAUA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EAC/C,IAAIA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS;EACzC,SAASA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS;EACxC,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC3C,gBAAgBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS,EAAE,SAAS;EACxD,UAAUA,YAAAA,EAAE,QAAQ,EAAE,SAAS,EAAE,SAAS;EAC1C,UAAUA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,eAAe,cAAc,SAAS,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,GAAGA,YAAAA,EAAE,QAAQ,EAAE,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC3H,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,MAAMA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,UAAU,QAAQ,WAAW,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS,EAAE,SAAS;EAC/F,OAAOA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC3B,gBAAgBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACxD,aAAaA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,GAAG,EAAE,SAAS;;EACjD,WAAWA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EAChC,OAAOA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EAC/C,iBAAiBA,YAAAA,EAAE,MAAM,CAACA,YAAAA,EAAE,KAAK,CAAC,OAAO,KAAK,CAAC,GAAGA,YAAAA,EAAE,OAAO,EAAE,IAAI,CAAC,CAAC,EAAE,SAAS;AAChF,CAAC,EAAE,YAAY;AAKR,IAAM,0BAA0BA,YAAAA,EAAE,OAAO;EAC9C,MAAMA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC1B,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACjC,YAAYA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACpD,cAAcA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;EACtD,oBAAoBA,YAAAA,EAAE,QAAQ,EAAE,SAAS;EACzC,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;EAC3C,SAASA,YAAAA,EAAE,MAAM,uBAAuB;AAC1C,CAAC;AASM,IAAM,sBAAsBA,YAAAA,EAAE,OAAO;;EAE1C,MAAMA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC3B,aAAaA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAClC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;EAC7C,UAAUA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC/B,WAAWA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAChC,aAAaA,YAAAA,EAAE,OAAO,EAAE,SAAS,EAAE,QAAQ,EAAE;;;EAE7C,SAASA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EAC9B,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,QAAQ,EAAE;EACxC,MAAMA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;EACpC,sBAAsBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,QAAQ,CAAC,CAAC;;EAEpD,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,eAAeA,YAAAA,EAAE,OAAO,EAAE,SAAS;EACnC,2BAA2BA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC/C,qBAAqBA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAClD,gBAAgB,wBAAwB,SAAS,EAAE,SAAS;EAC5D,YAAYA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,QAAQ,CAAC,EAAE,SAAS;;EAE3C,QAAQA,YAAAA,EAAE,MAAM,qBAAqB,EAAE,SAAS;EAChD,UAAUA,YAAAA,EAAE,OAAO,EAAE,SAAS;EAC9B,4BAA4BA,YAAAA,EAAE,OAAOA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EAC1D,QAAQA,YAAAA,EAAE,MAAMA,YAAAA,EAAE,OAAO,CAAC,EAAE,SAAS;EACrC,eAAeA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;EACvD,mBAAmBA,YAAAA,EAAE,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,SAAS;;AAC7D,CAAC;AAKM,IAAM,iBAAiBA,YAAAA,EAAE,OAAO;EACrC,MAAMA,YAAAA,EAAE,QAAQ,eAAe;EAC/B,cAAcA,YAAAA,EAAE,QAAQ,KAAK;EAC7B,MAAM;AACR,CAAC;AAiCM,SAAS,SAAS,MAAiC;AACxD,SAAO,eAAe,UAAU,IAAI,EAAE;AACxC;;;AQ3HA,IAAM,SAAS,OAAO,YAAY,eAChC,QAAQ,YAAY,QACpB,QAAQ,SAAS,QAAQ;AAO3B,IAAM,yBAAyB,OAAO;AAsFtC,IAAM,oBAAoB,KAAK;AClG/B,IAAM,uBAAuB,uBAAO,IAAI,sCAAsC;AAKvE,IAAM,eAAN,cAA2B,MAAM;EAItC,YAAY,SAAiC,MAAc;AACzD,UAAM,OAAO;AAD8B,SAAA,OAAA;AAE3C,SAAK,OAAO;AAEZ,QAAI,MAAM,mBAAmB;AAC3B,YAAM,kBAAkB,MAAM,KAAK,WAAW;IAChD;EACF;;EATA,CAAU,oBAAoB,IAAI;AAUpC;AAmCO,IAAM,0BAAN,cAAsC,aAAa;EACxD,YAA4B,QAAgB,WAAoB;AAC9D,UAAM,MAAM,YACR,WAAW,MAAM,uBAAuB,SAAS,KACjD,yBAAyB,MAAM;AACnC,UAAM,KAAK,sBAAsB;AAJP,SAAA,SAAA;AAK1B,SAAK,OAAO;EACd;AACF;;;AK/CA,SAAS,kBACP,UACmD;AACnD,MAAI,aAAa,UAAa,aAAa,QAAQ,aAAa,GAAI,QAAO;AAC3E,MAAI,OAAO,aAAa,SAAU,QAAO;AAEzC,SAAO;AACT;AAKA,SAAS,qBACP,OACA,OACmB;AACnB,SAAO;IACL,MAAM,MAAM,QAAQ,CAAC;IACrB,SAAS,MAAM,WAAW;IAC1B,YAAY,MAAM,cAAc,CAAC;IACjC,SAAS,MAAM,WAAW;IAC1B,iBAAiB,MAAM,mBAAmB;IAC1C,gBAAgB,MAAM;IACtB,MAAM,MAAM,QAAQ;IACpB,UAAU,MAAM,YAAY;IAC5B,IAAI,MAAM,MAAM;IAChB,SAAS,MAAM,WAAW;IAC1B,WAAW,MAAM,aAAa;IAC9B,gBAAgB,MAAM,kBAAkB,CAAC;IACzC,UAAU,MAAM,YAAY;IAC5B,UAAU,kBAAkB,MAAM,QAAQ;EAC5C;AACF;AAKA,SAAS,qBACP,MACmB;AACnB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,YAAY,KAAK;IACjB,cAAc,KAAK;IACnB,oBAAoB,KAAK;IACzB,YAAY,KAAK;IACjB,SAAS,KAAK,QAAQ,IAAI,CAAC,OAAO,MAAM,qBAAqB,OAAO,CAAC,CAAC;EACxE;AACF;AAYO,SAAS,WAAW,OAAyC;AAClE,QAAM,OAAO,UAAU,KAAK;AAE5B,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM;;MAEJ,MAAM,KAAK,QAAQ;MACnB,aAAa,KAAK,eAAe;MACjC,aAAa,KAAK,eAAe;MACjC,UAAU,KAAK,YAAY;MAC3B,WAAW,KAAK,aAAa;MAC7B,aAAa,KAAK,eAAe;;MAEjC,eAAe,KAAK,iBAAiB;MACrC,eAAe,KAAK,iBAAiB;MACrC,2BAA2B,KAAK,6BAA6B;MAC7D,qBAAqB,KAAK,uBAAuB,CAAC;MAClD,sBAAsB,CAAC;MACvB,MAAM,KAAK,QAAQ,CAAC;MACpB,SAAS,KAAK,WAAW;MACzB,mBAAmB,KAAK,qBAAqB;MAC7C,gBAAgB,KAAK,iBACjB,qBAAqB,KAAK,cAAc,IACxC;MACJ,YAAY,KAAK,cAAc,CAAC;IAClC;EACF;AACF;AC5FO,IAAM,uBAAuB;EAClC;EACA;AACF;AAKA,SAASC,sBAAqB,OAA6C;AACzE,SAAO;IACL,MAAM,MAAM;IACZ,SAAS,MAAM;IACf,YAAY,MAAM,cAAc,CAAC;IACjC,SAAS,MAAM;IACf,iBAAiB,MAAM;IACvB,gBAAgB,MAAM;IACtB,MAAM,MAAM;IACZ,UAAU,MAAM;IAChB,IAAI,MAAM;IACV,SAAS,MAAM;IACf,WAAW,MAAM;IACjB,gBAAgB,MAAM;IACtB,UAAU,MAAM;IAChB,UAAU,MAAM;EAClB;AACF;AAKA,SAASC,sBACP,MACmB;AACnB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,YAAY,KAAK;IACjB,cAAc,KAAK;IACnB,oBAAoB,KAAK;IACzB,YAAY,KAAK;IACjB,SAAS,KAAK,QAAQ,IAAID,qBAAoB;EAChD;AACF;AAQO,SAAS,eAAe,MAA0B;AACvD,QAAM,OAAO,KAAK;AAElB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK;IAClB,UAAU,KAAK;IACf,WAAW,KAAK;IAChB,aAAa,KAAK;IAClB,eAAe,KAAK;IACpB,eAAe,KAAK;IACpB,2BAA2B,KAAK;IAChC,qBAAqB,KAAK;IAC1B,MAAM,KAAK;IACX,SAAS,KAAK;IACd,mBAAmB,KAAK;IACxB,gBAAgB,KAAK,iBACjBC,sBAAqB,KAAK,cAAc,IACxC;IACJ,YAAY,KAAK;EACnB;AACF;AAQO,SAAS,kBAAkB,MAA6B;AAC7D,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM,eAAe,IAAI;EAC3B;AACF;AAKO,SAAS,gBAAgB,MAA0B;AACxD,QAAM,OAAiB,CAAC;AAExB,MAAI,KAAK,KAAK,wBAAwB,KAAK,KAAK,qBAAqB,SAAS,GAAG;AAC/E,SAAK,KAAK,yBAAyB,KAAK,KAAK,qBAAqB,MAAM,WAAW;EACrF;AAEA,MAAI,KAAK,KAAK,UAAU,KAAK,KAAK,OAAO,SAAS,GAAG;AACnD,SAAK,KAAK,WAAW,KAAK,KAAK,OAAO,MAAM,WAAW;EACzD;AAEA,SAAO;AACT;ACxGO,SAAS,YAAY,OAA+C;AACzE,QAAM,OAAOC,UAAU,KAAK;AAE5B,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK,eAAe;;IACjC,UAAU,KAAK;IACf,UAAU,KAAK;IACf,YAAY,KAAK,eAAe;;IAChC,cAAc,KAAK;IACnB,yBAAyB,KAAK;IAC9B,cAAc,KAAK;IACnB,oBAAoB,KAAK,uBAAuB,CAAC;IACjD,oBAAoB,CAAC;IACrB,MAAM,KAAK,QAAQ,CAAC;IACpB,SAAS,KAAK;IACd,kBAAkB,KAAK;IACvB,eAAe,KAAK,kBAAkB;IACtC,YAAY,KAAK,cAAc,CAAC;EAClC;AACF;AAKO,SAAS,YAAY,MAAgC;AAC1D,QAAM,OAAO,KAAK;AAElB,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK,eAAe;;IACjC,UAAU,KAAK;IACf,UAAU,KAAK;IACf,YAAY,KAAK,eAAe;;IAChC,cAAc,KAAK;IACnB,yBAAyB,KAAK;IAC9B,cAAc,KAAK;IACnB,oBAAoB,KAAK,uBAAuB,CAAC;IACjD,oBAAoB,KAAK,wBAAwB,CAAC;IAClD,MAAM,KAAK,QAAQ,CAAC;IACpB,SAAS,KAAK;IACd,kBAAkB,KAAK;IACvB,eAAe,KAAK,kBAAkB;IACtC,YAAY,KAAK,cAAc,CAAC;EAClC;AACF;AAQO,SAAS,UAAU,OAA0D;AAClF,MAAI,SAAS,KAAK,GAAG;AACnB,WAAO,YAAY,KAAiB;EACtC;AAEA,MAAI,YAAY,KAAK,KAAK,UAAU,OAAO;AACzC,WAAO,YAAY,KAA+B;EACpD;AAEA,QAAM,IAAI,wBAAwB,WAAW,eAAe;AAC9D;ACnEO,SAAS,gBAAgB,MAAgC;AAC9D,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM;MACJ,MAAM,KAAK;MACX,aAAa,KAAK;MAClB,aAAa,KAAK;MAClB,UAAU,KAAK;MACf,WAAW,KAAK;MAChB,aAAa,KAAK;MAClB,eAAe,KAAK,gBAAgB;MACpC,eAAe,KAAK,gBAAgB;MACpC,2BAA2B,KAAK,2BAA2B;MAC3D,qBAAqB,KAAK;MAC1B,sBAAsB,KAAK;MAC3B,MAAM,KAAK;MACX,SAAS,KAAK,WAAW;MACzB,mBAAmB,KAAK,oBAAoB;MAC5C,gBAAgB,KAAK;MACrB,YAAY,KAAK;IACnB;EACF;AACF;AAKA,SAAS,yBAAyB,MAAmE;AACnG,MAAI,CAAC,KAAM,QAAO;AAElB,SAAO;IACL,GAAG;IACH,SAAS,KAAK,QAAQ,IAAI,CAAC,WAAW;MACpC,GAAG;MACH,YAAY,MAAM,cAAc,CAAC;IACnC,EAAE;EACJ;AACF;AAOO,SAAS,oBAAoB,MAAgC;AAClE,SAAO;IACL,MAAM,KAAK;IACX,aAAa,KAAK;IAClB,aAAa,KAAK;IAClB,UAAU,KAAK;IACf,WAAW,KAAK;IAChB,aAAa,KAAK;IAClB,eAAe,KAAK;IACpB,eAAe,KAAK;IACpB,2BAA2B,KAAK;IAChC,qBAAqB,KAAK;IAC1B,MAAM,KAAK;IACX,SAAS,KAAK;IACd,mBAAmB,KAAK;IACxB,gBAAgB,yBAAyB,KAAK,aAAa;IAC3D,YAAY,KAAK;EACnB;AACF;AAOO,SAAS,uBAAuB,MAAmC;AACxE,SAAO;IACL,MAAM;IACN,cAAc;IACd,MAAM,oBAAoB,IAAI;EAChC;AACF;AAKO,SAAS,wBAAwB,MAAgC;AACtE,QAAM,OAAiB,CAAC;AAExB,MAAI,KAAK,mBAAmB,SAAS,GAAG;AACtC,SAAK,KAAK,yBAAyB,KAAK,mBAAmB,MAAM,WAAW;EAC9E;AAEA,SAAO;AACT;","names":["import_zod","z","z","convertLorebookEntry","convertCharacterBook","getV2Data"]}
|