nextjs-studio 1.0.3 → 1.0.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/bin/nextjs-studio.js +6227 -29
- package/dist/cli/ui/standalone.tar.gz +0 -0
- package/dist/{query-builder-KXz9cPzF.d.ts → query-builder-BOu-D7a1.d.ts} +10 -0
- package/package.json +102 -101
- package/dist/bin/nextjs-studio.d.ts +0 -1
- package/dist/bin/nextjs-studio.js.map +0 -1
- package/dist/core/index.d.ts +0 -117
- package/dist/core/index.js.map +0 -1
- package/dist/core/server.d.ts +0 -135
- package/dist/core/server.js +0 -707
- package/dist/core/server.js.map +0 -1
package/dist/core/server.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/core/server.ts","../../src/shared/constants.ts","../../src/core/fs-adapter.ts","../../src/core/indexer.ts","../../src/core/parsers/parser-mdx.ts","../../src/core/frontmatter-binder.ts","../../src/core/parsers/parser-json.ts","../../src/core/schema-inferrer.ts","../../src/core/locale-parser.ts","../../src/core/content-store.ts","../../src/core/config-loader.ts","../../src/core/query-builder.ts","../../src/core/draft-filter.ts"],"sourcesContent":["/**\n * @context Core layer — server entry point at src/core/server.ts\n * @does Re-exports server-only APIs and auto-initializes the content store\n * @depends src/core/fs-adapter.ts, src/core/content-store.ts\n * @do Import from 'nextjs-studio/server' in Next.js server components\n * @dont Import from client components\n */\n\nimport path from \"node:path\";\nimport { CONTENTS_DIR } from \"../shared/constants.js\";\nimport { FsAdapter } from \"./fs-adapter.js\";\nimport { ContentIndex } from \"./indexer.js\";\nimport { hasStore, setStore } from \"./content-store.js\";\n\n// Auto-initialize the content store synchronously on import\nif (!hasStore()) {\n const dir = path.join(process.cwd(), CONTENTS_DIR);\n const index = new ContentIndex(new FsAdapter(dir));\n index.buildSync();\n setStore(index);\n}\n\nexport { FsAdapter } from \"./fs-adapter.js\";\nexport { ContentIndex } from \"./indexer.js\";\nexport { loadContent, loadContentSync } from \"./content-store.js\";\nexport { loadStudioConfig, resolveConfigPath, loadConfigFromPath } from \"./config-loader.js\";\nexport { queryCollection } from \"./query-builder.js\";\n\nexport type {\n ContentEntry,\n Collection,\n CollectionTypeMap,\n StudioConfig,\n CollectionConfig,\n QueryOptions,\n} from \"../shared/types.js\";\n\nexport type { QueryResult } from \"./query-builder.js\";\n","/**\n * @context Shared layer — constants at src/shared/constants.ts\n * @does Defines project-wide constants shared across core, CLI, and UI layers\n * @depends none\n * @do Add new shared constants here\n * @dont Import from CLI or UI; constants must be framework-agnostic\n */\n\nexport const CONTENTS_DIR = \"contents\";\nexport const CLI_PORT = 3030;\nexport const CONFIG_FILE = \"studio.config.ts\";\nexport const CONFIG_FILENAMES = [\"studio.config.ts\", \"studio.config.js\", \"studio.config.mjs\"] as const;\nexport const SUPPORTED_EXTENSIONS = [\".mdx\", \".json\"] as const;\nexport const COLLECTION_ORDER_FILE = \"collection.json\";\nexport const WATCHER_DEBOUNCE_MS = 5_000;\nexport const MEDIA_DIR = \"media\";\n\nexport const IMAGE_MIME_TYPES = [\n \"image/png\",\n \"image/jpeg\",\n \"image/gif\",\n \"image/webp\",\n \"image/svg+xml\",\n \"image/avif\",\n] as const;\n\nexport const VIDEO_MIME_TYPES = [\"video/mp4\", \"video/webm\", \"video/ogg\"] as const;\n\nexport const AUDIO_MIME_TYPES = [\n \"audio/mpeg\",\n \"audio/ogg\",\n \"audio/wav\",\n \"audio/webm\",\n \"audio/aac\",\n \"audio/flac\",\n] as const;\n\nexport const MEDIA_MIME_TYPES = [...IMAGE_MIME_TYPES, ...VIDEO_MIME_TYPES, ...AUDIO_MIME_TYPES] as const;\n\nexport const IMAGE_EXTENSIONS = [\".png\", \".jpg\", \".jpeg\", \".gif\", \".webp\", \".svg\", \".avif\"] as const;\nexport const VIDEO_EXTENSIONS = [\".mp4\", \".webm\", \".ogv\"] as const;\nexport const AUDIO_EXTENSIONS = [\".mp3\", \".ogg\", \".wav\", \".m4a\", \".aac\", \".flac\"] as const;\n","/**\n * @context Core layer — filesystem adapter at src/core/fs-adapter.ts\n * @does Implements IFsAdapter; abstracts all file read/write/list operations behind a single interface\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts\n * @do Add new I/O operations here; all file access must go through this adapter\n * @dont Import UI components, run HTTP requests, or contain business logic\n */\n\nimport fs from \"node:fs/promises\";\nimport fsSync from \"node:fs\";\nimport path from \"node:path\";\nimport type { Dirent } from \"node:fs\";\nimport type { FileInfo, DirectoryFileEntry } from \"../shared/types.js\";\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport { SUPPORTED_EXTENSIONS } from \"../shared/constants.js\";\n\nexport class FsAdapter implements IFsAdapter {\n private readonly basePath: string;\n\n constructor(basePath: string) {\n this.basePath = path.resolve(basePath);\n }\n\n private resolve(...segments: string[]): string {\n return path.resolve(this.basePath, ...segments);\n }\n\n async readFile(filePath: string): Promise<string> {\n return fs.readFile(this.resolve(filePath), \"utf-8\");\n }\n\n async writeFile(filePath: string, content: string): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, content, \"utf-8\");\n }\n\n async deleteFile(filePath: string): Promise<void> {\n await fs.unlink(this.resolve(filePath));\n }\n\n async exists(filePath: string): Promise<boolean> {\n try {\n await fs.access(this.resolve(filePath));\n return true;\n } catch {\n return false;\n }\n }\n\n async getStats(filePath: string): Promise<FileInfo> {\n const fullPath = this.resolve(filePath);\n const stats = await fs.stat(fullPath);\n return { path: filePath, size: stats.size, modifiedAt: stats.mtime };\n }\n\n async listFiles(dirPath: string, extensions?: readonly string[]): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n const filterExts = extensions ?? SUPPORTED_EXTENSIONS;\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext)))\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async listDirectories(dirPath: string): Promise<string[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isDirectory())\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n async readBuffer(filePath: string): Promise<Buffer> {\n return fs.readFile(this.resolve(filePath));\n }\n\n async writeBuffer(filePath: string, data: Buffer): Promise<void> {\n const fullPath = this.resolve(filePath);\n await fs.mkdir(path.dirname(fullPath), { recursive: true });\n await fs.writeFile(fullPath, data);\n }\n\n async listAllFiles(dirPath: string): Promise<DirectoryFileEntry[]> {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = await fs.readdir(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n const results: DirectoryFileEntry[] = [];\n for (const entry of entries) {\n if (!entry.isFile()) continue;\n const relativePath = this.join(dirPath, entry.name);\n const stats = await fs.stat(this.resolve(relativePath));\n results.push({ name: entry.name, relativePath, size: stats.size, modifiedAt: stats.mtime });\n }\n return results;\n }\n\n join(...segments: string[]): string {\n return path.join(...segments);\n }\n\n basename(filePath: string): string {\n return path.basename(filePath);\n }\n\n extname(filePath: string): string {\n return path.extname(filePath);\n }\n\n relative(from: string, to: string): string {\n return path.relative(from, to);\n }\n\n normalizeSlug(relativePath: string, ext: string): string {\n return relativePath.replace(ext, \"\").split(path.sep).join(\"/\");\n }\n\n readFileSync(filePath: string): string {\n return fsSync.readFileSync(this.resolve(filePath), \"utf-8\");\n }\n\n existsSync(filePath: string): boolean {\n return fsSync.existsSync(this.resolve(filePath));\n }\n\n listFilesSync(dirPath: string, extensions?: readonly string[]): string[] {\n const fullPath = this.resolve(dirPath);\n const filterExts = extensions ?? SUPPORTED_EXTENSIONS;\n\n let entries: Dirent[];\n try {\n entries = fsSync.readdirSync(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isFile() && filterExts.some((ext) => entry.name.endsWith(ext)))\n .map((entry) => this.join(dirPath, entry.name));\n }\n\n listDirectoriesSync(dirPath: string): string[] {\n const fullPath = this.resolve(dirPath);\n\n let entries: Dirent[];\n try {\n entries = fsSync.readdirSync(fullPath, { withFileTypes: true });\n } catch {\n return [];\n }\n\n return entries\n .filter((entry) => entry.isDirectory())\n .map((entry) => this.join(dirPath, entry.name));\n }\n}\n","/**\n * @context Core layer — content indexer at src/core/indexer.ts\n * @does Scans the contents directory, parses MDX/JSON files, and builds an in-memory index\n * @depends src/shared/types.ts, src/shared/constants.ts, src/shared/fs-adapter.interface.ts, src/core/parsers/, src/core/schema-inferrer.ts\n * @do Add new file type handling here; extend indexCollection for new collection behaviors\n * @dont Import from CLI or UI; instantiate FsAdapter; access the filesystem directly\n */\n\nimport slugify from \"@sindresorhus/slugify\";\nimport type { CollectionSchema } from \"../shared/fields.js\";\nimport type { ContentEntry, Collection, StudioConfig } from \"../shared/types.js\";\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport { COLLECTION_ORDER_FILE } from \"../shared/constants.js\";\nimport { parseMdx } from \"./parsers/parser-mdx.js\";\nimport { parseJson } from \"./parsers/parser-json.js\";\nimport { inferSchema } from \"./schema-inferrer.js\";\nimport { parseLocaleFromFilename, stripLocaleFromSlug } from \"./locale-parser.js\";\n\nexport class ContentIndex {\n private readonly entries = new Map<string, ContentEntry[]>();\n private readonly collections = new Map<string, Collection>();\n private readonly fs: IFsAdapter;\n\n constructor(fsAdapter: IFsAdapter) {\n this.fs = fsAdapter;\n }\n\n async build(config?: StudioConfig): Promise<void> {\n this.clear();\n const dirs = await this.fs.listDirectories(\".\");\n\n for (const dir of dirs) {\n const dirName = this.fs.basename(dir);\n const collectionName = slugify(dirName);\n const collectionConfig = config?.collections?.[collectionName];\n await this.indexCollection(dirName, collectionName, collectionConfig?.schema);\n }\n }\n\n buildSync(config?: StudioConfig): void {\n this.clear();\n const dirs = this.fs.listDirectoriesSync(\".\");\n\n for (const dir of dirs) {\n const dirName = this.fs.basename(dir);\n const collectionName = slugify(dirName);\n const collectionConfig = config?.collections?.[collectionName];\n this.indexCollectionSync(dirName, collectionName, collectionConfig?.schema);\n }\n }\n\n getCollection(name: string): ContentEntry[] {\n return this.entries.get(name) ?? [];\n }\n\n getCollections(): Collection[] {\n return Array.from(this.collections.values());\n }\n\n clear(): void {\n this.entries.clear();\n this.collections.clear();\n }\n\n updateEntry(collectionName: string, entry: ContentEntry): void {\n const entries = this.entries.get(collectionName) ?? [];\n const idx = entries.findIndex((e) => e.slug === entry.slug);\n if (idx >= 0) {\n entries[idx] = entry;\n } else {\n entries.push(entry);\n }\n this.entries.set(collectionName, entries);\n this.updateCollectionMeta(collectionName);\n }\n\n removeEntry(collectionName: string, slug: string): void {\n const entries = this.entries.get(collectionName);\n if (!entries) return;\n const filtered = entries.filter((e) => e.slug !== slug);\n this.entries.set(collectionName, filtered);\n this.updateCollectionMeta(collectionName);\n }\n\n private updateCollectionMeta(collectionName: string): void {\n const col = this.collections.get(collectionName);\n const entries = this.entries.get(collectionName) ?? [];\n if (col) {\n this.collections.set(collectionName, {\n ...col,\n count: entries.length,\n type: this.detectCollectionType(entries),\n });\n }\n }\n\n private async indexCollection(\n dirName: string,\n collectionName: string,\n manualSchema?: CollectionSchema,\n ): Promise<void> {\n const entries: ContentEntry[] = [];\n await this.scanDir(dirName, collectionName, dirName, entries);\n\n const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);\n const ordering = await this.readOrdering(orderPath);\n if (ordering) {\n this.applyOrdering(entries, ordering);\n }\n\n const schema = manualSchema ?? inferSchema(entries, collectionName);\n\n this.entries.set(collectionName, entries);\n this.collections.set(collectionName, {\n name: collectionName,\n type: this.detectCollectionType(entries),\n count: entries.length,\n basePath: dirName,\n schema,\n });\n }\n\n private indexCollectionSync(\n dirName: string,\n collectionName: string,\n manualSchema?: CollectionSchema,\n ): void {\n const entries: ContentEntry[] = [];\n this.scanDirSync(dirName, collectionName, dirName, entries);\n\n const orderPath = this.fs.join(dirName, COLLECTION_ORDER_FILE);\n const ordering = this.readOrderingSync(orderPath);\n if (ordering) {\n this.applyOrdering(entries, ordering);\n }\n\n const schema = manualSchema ?? inferSchema(entries, collectionName);\n\n this.entries.set(collectionName, entries);\n this.collections.set(collectionName, {\n name: collectionName,\n type: this.detectCollectionType(entries),\n count: entries.length,\n basePath: dirName,\n schema,\n });\n }\n\n private async scanDir(\n dirName: string,\n collectionName: string,\n dirPath: string,\n entries: ContentEntry[],\n ): Promise<void> {\n const subDirs = await this.fs.listDirectories(dirPath);\n for (const subDir of subDirs) {\n await this.scanDir(dirName, collectionName, subDir, entries);\n }\n\n const files = await this.fs.listFiles(dirPath);\n for (const filePath of files) {\n const fileName = this.fs.basename(filePath);\n if (fileName === COLLECTION_ORDER_FILE) continue;\n\n const ext = this.fs.extname(fileName);\n const content = await this.fs.readFile(filePath);\n const relativePath = this.fs.relative(dirName, filePath);\n const slug = this.fs\n .normalizeSlug(relativePath, ext)\n .split(\"/\")\n .map((segment) => slugify(segment))\n .join(\"/\");\n\n if (ext === \".mdx\") {\n entries.push(this.buildMdxEntry(collectionName, slug, fileName, content));\n } else if (ext === \".json\") {\n entries.push(...this.buildJsonEntries(collectionName, slug, content));\n }\n }\n }\n\n private scanDirSync(\n dirName: string,\n collectionName: string,\n dirPath: string,\n entries: ContentEntry[],\n ): void {\n const subDirs = this.fs.listDirectoriesSync(dirPath);\n for (const subDir of subDirs) {\n this.scanDirSync(dirName, collectionName, subDir, entries);\n }\n\n const files = this.fs.listFilesSync(dirPath);\n for (const filePath of files) {\n const fileName = this.fs.basename(filePath);\n if (fileName === COLLECTION_ORDER_FILE) continue;\n\n const ext = this.fs.extname(fileName);\n const content = this.fs.readFileSync(filePath);\n const relativePath = this.fs.relative(dirName, filePath);\n const slug = this.fs\n .normalizeSlug(relativePath, ext)\n .split(\"/\")\n .map((segment) => slugify(segment))\n .join(\"/\");\n\n if (ext === \".mdx\") {\n entries.push(this.buildMdxEntry(collectionName, slug, fileName, content));\n } else if (ext === \".json\") {\n entries.push(...this.buildJsonEntries(collectionName, slug, content));\n }\n }\n }\n\n private buildMdxEntry(collectionName: string, slug: string, fileName: string, content: string): ContentEntry {\n const parsed = parseMdx(content);\n const locale = parseLocaleFromFilename(fileName);\n const normalizedSlug = stripLocaleFromSlug(slug, locale);\n return {\n collection: collectionName,\n slug: normalizedSlug,\n path: `/${collectionName}/${normalizedSlug}`,\n body: parsed.body,\n data: parsed.data,\n ...(locale ? { locale } : {}),\n };\n }\n\n private buildJsonEntries(collectionName: string, slug: string, content: string): ContentEntry[] {\n const parsed = parseJson(content);\n\n if (parsed.type === \"json-array\") {\n return parsed.entries.map((data, index) => {\n const entrySlug =\n typeof data[\"slug\"] === \"string\" ? slugify(data[\"slug\"]) : `${slug}/${index}`;\n return {\n collection: collectionName,\n slug: entrySlug,\n path: `/${collectionName}/${entrySlug}`,\n data,\n };\n });\n }\n\n return [{ collection: collectionName, slug, path: `/${collectionName}/${slug}`, data: parsed.data }];\n }\n\n private async readOrdering(orderPath: string): Promise<string[] | null> {\n if (!(await this.fs.exists(orderPath))) return null;\n\n try {\n const content = await this.fs.readFile(orderPath);\n const parsed: unknown = JSON.parse(content);\n if (Array.isArray(parsed)) return parsed as string[];\n } catch (error) {\n console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);\n }\n return null;\n }\n\n private readOrderingSync(orderPath: string): string[] | null {\n if (!this.fs.existsSync(orderPath)) return null;\n\n try {\n const content = this.fs.readFileSync(orderPath);\n const parsed: unknown = JSON.parse(content);\n if (Array.isArray(parsed)) return parsed as string[];\n } catch (error) {\n console.warn(`[Nextjs Studio] Failed to parse ordering file: ${orderPath}`, error);\n }\n return null;\n }\n\n private applyOrdering(entries: ContentEntry[], ordering: string[]): void {\n const orderMap = new Map(ordering.map((slug, index) => [slug, index]));\n entries.sort((a, b) => {\n const aIndex = orderMap.get(a.slug) ?? Infinity;\n const bIndex = orderMap.get(b.slug) ?? Infinity;\n return aIndex - bIndex;\n });\n }\n\n private detectCollectionType(entries: ContentEntry[]): Collection[\"type\"] {\n if (entries.length === 0) return \"mdx\";\n const first = entries[0];\n if (first.body !== undefined) return \"mdx\";\n if (entries.length === 1 && !first.slug.includes(\"/\")) return \"json-object\";\n return \"json-array\";\n }\n}\n","/**\n * @context Core layer — MDX parser/serializer at src/core/parsers/parser-mdx.ts\n * @does Parses .mdx content into frontmatter + body, and serializes them back to MDX strings\n * @depends none (gray-matter is an external dep)\n * @do Add MDX transform steps here; both parse and serialize live here intentionally\n * @dont Access the filesystem; import from CLI or UI; handle JSON content\n */\n\nimport matter from \"gray-matter\";\nimport { bindFrontmatter } from \"../frontmatter-binder.js\";\n\nexport interface ParsedMdx {\n data: Record<string, unknown>;\n body: string;\n}\n\nexport interface ParseMdxOptions {\n /** When true, replaces {frontmatter.X} tokens in the body with actual values. */\n bindTokens?: boolean;\n}\n\n/** Convert Date objects produced by gray-matter back to ISO strings. */\nfunction normalizeDates(data: Record<string, unknown>): Record<string, unknown> {\n const result: Record<string, unknown> = {};\n for (const [key, value] of Object.entries(data)) {\n if (value instanceof Date) {\n result[key] = value.toISOString().split(\"T\")[0];\n } else if (typeof value === \"object\" && value !== null && !Array.isArray(value)) {\n result[key] = normalizeDates(value as Record<string, unknown>);\n } else {\n result[key] = value;\n }\n }\n return result;\n}\n\nexport function parseMdx(content: string, options?: ParseMdxOptions): ParsedMdx {\n const { data: rawData, content: body } = matter(content);\n const data = normalizeDates(rawData);\n const trimmed = body.trim();\n return {\n data,\n body: options?.bindTokens ? bindFrontmatter(trimmed, data) : trimmed,\n };\n}\n\nexport function serializeMdx(data: Record<string, unknown>, body: string): string {\n return matter.stringify(body, data);\n}\n","/**\n * @context Core layer — frontmatter binder at src/core/frontmatter-binder.ts\n * @does Replaces {frontmatter.X} tokens in MDX body with actual frontmatter values\n * @depends none\n * @do Add new token patterns or transformation rules here\n * @dont Import from CLI or UI; access filesystem\n */\n\nconst TOKEN_REGEX = /\\{frontmatter\\.([a-zA-Z0-9_.]+)\\}/g;\n\n/**\n * Replaces `{frontmatter.X}` tokens in the body with values from the data object.\n * Supports dot-notation for nested values (e.g. `{frontmatter.author.name}`).\n */\nexport function bindFrontmatter(body: string, data: Record<string, unknown>): string {\n return body.replace(TOKEN_REGEX, (_match, path: string) => {\n const value = resolvePath(data, path);\n if (value === undefined || value === null) return _match;\n if (typeof value === \"object\") return JSON.stringify(value);\n return String(value);\n });\n}\n\n/**\n * Extracts all frontmatter token paths from the body.\n */\nexport function extractFrontmatterTokens(body: string): string[] {\n const tokens: string[] = [];\n let match: RegExpExecArray | null;\n const regex = new RegExp(TOKEN_REGEX.source, \"g\");\n while ((match = regex.exec(body)) !== null) {\n tokens.push(match[1]!);\n }\n return tokens;\n}\n\nfunction resolvePath(obj: Record<string, unknown>, path: string): unknown {\n const keys = path.split(\".\");\n let current: unknown = obj;\n for (const key of keys) {\n if (typeof current !== \"object\" || current === null) return undefined;\n current = (current as Record<string, unknown>)[key];\n }\n return current;\n}\n","/**\n * @context Core layer — JSON parser at src/core/parsers/parser-json.ts\n * @does Parses JSON content strings into typed ParsedJson results (array or object)\n * @depends none\n * @do Extend ParsedJson variants here if new JSON structures are supported\n * @dont Access the filesystem; import from CLI or UI; contain serialization logic\n */\n\nexport interface ParsedJsonArray {\n type: \"json-array\";\n entries: Record<string, unknown>[];\n}\n\nexport interface ParsedJsonObject {\n type: \"json-object\";\n data: Record<string, unknown>;\n}\n\nexport type ParsedJson = ParsedJsonArray | ParsedJsonObject;\n\nexport function parseJson(content: string): ParsedJson {\n const parsed: unknown = JSON.parse(content);\n\n if (Array.isArray(parsed)) {\n return {\n type: \"json-array\",\n entries: parsed as Record<string, unknown>[],\n };\n }\n\n if (typeof parsed === \"object\" && parsed !== null) {\n return {\n type: \"json-object\",\n data: parsed as Record<string, unknown>,\n };\n }\n\n throw new Error(\"JSON content must be an array or object\");\n}\n","/**\n * @context Core layer — schema inferrer at src/core/schema-inferrer.ts\n * @does Infers a CollectionSchema from actual content entries when no manual schema is defined\n * @depends src/shared/types.ts, src/shared/fields.ts\n * @do Add new type detection heuristics here (e.g. color, phone)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport type { ContentEntry } from \"../shared/types.js\";\nimport type { CollectionSchema, FieldDefinition, SelectOption } from \"../shared/fields.js\";\n\n// Value detector patterns\nconst RE_ISO_DATE = /^\\d{4}-\\d{2}-\\d{2}$/;\nconst RE_ISO_DATETIME =\n /^\\d{4}-\\d{2}-\\d{2}[T ]\\d{2}:\\d{2}(:\\d{2}(\\.\\d+)?)?(Z|[+-]\\d{2}:?\\d{2})?$/;\nconst RE_EMAIL = /^[^\\s@]+@[^\\s@]+\\.[^\\s@]+$/;\nconst RE_URL = /^https?:\\/\\/.+/;\nconst LONG_TEXT_THRESHOLD = 200;\nconst RICH_TEXT_FIELD_NAMES = [\"description\", \"descriptions\", \"text\", \"content\"];\n\nfunction isISODate(value: string): boolean {\n return RE_ISO_DATE.test(value);\n}\n\nfunction isISODateTime(value: string): boolean {\n return RE_ISO_DATETIME.test(value);\n}\n\nfunction isEmail(value: string): boolean {\n return RE_EMAIL.test(value);\n}\n\nfunction isUrl(value: string): boolean {\n return RE_URL.test(value);\n}\n\nfunction inferStringField(name: string, strings: string[]): FieldDefinition {\n if (strings.every(isEmail)) return { name, type: \"email\" };\n if (strings.every(isUrl)) return { name, type: \"url\" };\n if (strings.every(isISODateTime)) return { name, type: \"date\", includeTime: true };\n if (strings.every(isISODate)) return { name, type: \"date\" };\n\n // Fields with well-known names are always treated as long/rich text\n if (RICH_TEXT_FIELD_NAMES.includes(name.toLowerCase())) return { name, type: \"long-text\" };\n\n const isLong = strings.some((s) => s.length > LONG_TEXT_THRESHOLD || s.includes(\"\\n\"));\n return { name, type: isLong ? \"long-text\" : \"text\" };\n}\n\nfunction inferArrayField(name: string, items: unknown[]): FieldDefinition {\n if (items.length === 0) return { name, type: \"array\", itemFields: [] };\n\n if (items.every((item) => typeof item === \"string\")) {\n const unique = [...new Set(items as string[])].slice(0, 50);\n const options: SelectOption[] = unique.map((v) => ({ label: v, value: v }));\n return { name, type: \"multi-select\", options };\n }\n\n if (items.every((item) => typeof item === \"object\" && item !== null && !Array.isArray(item))) {\n return { name, type: \"array\", itemFields: inferFields(items as Record<string, unknown>[]) };\n }\n\n return { name, type: \"array\", itemFields: [] };\n}\n\nfunction inferFieldDefinition(name: string, values: unknown[]): FieldDefinition {\n const present = values.filter((v) => v !== null && v !== undefined);\n\n if (present.length === 0) return { name, type: \"text\" };\n if (present.every((v) => typeof v === \"boolean\")) return { name, type: \"boolean\" };\n\n if (present.every((v) => typeof v === \"number\")) {\n const format = present.every((v) => Number.isInteger(v)) ? \"integer\" : \"decimal\";\n return { name, type: \"number\", format };\n }\n\n if (present.every((v) => typeof v === \"string\")) {\n return inferStringField(name, present as string[]);\n }\n\n if (present.every((v) => Array.isArray(v))) {\n return inferArrayField(name, (present as unknown[][]).flat());\n }\n\n if (present.every((v) => typeof v === \"object\" && v !== null && !Array.isArray(v))) {\n return { name, type: \"object\", fields: inferFields(present as Record<string, unknown>[]) };\n }\n\n return { name, type: \"text\" };\n}\n\nfunction inferFields(rows: Record<string, unknown>[]): FieldDefinition[] {\n const keySet = new Set<string>(rows.flatMap((row) => Object.keys(row)));\n return Array.from(keySet).map((key) => inferFieldDefinition(key, rows.map((row) => row[key])));\n}\n\n/**\n * Infer a `CollectionSchema` from the data of a set of content entries.\n *\n * The result is a best-effort approximation — string fields that look like\n * emails, URLs, or ISO dates get the correct semantic type. Everything else\n * falls back to `text`.\n */\nexport function inferSchema(entries: ContentEntry[], collectionName: string): CollectionSchema {\n const rows = entries.map((entry) => entry.data as Record<string, unknown>);\n return { collection: collectionName, fields: inferFields(rows) };\n}\n","/**\n * @context Core layer — locale parser at src/core/locale-parser.ts\n * @does Extracts locale codes from filenames using the convention `slug.locale.mdx`\n * @depends none\n * @do Add new locale detection strategies here\n * @dont Import from CLI or UI; access filesystem\n */\n\nconst LOCALE_REGEX = /\\.([a-z]{2}(?:-[A-Z]{2})?)\\.mdx$/;\n\n/**\n * Parses locale from a filename.\n * Supports `post.pt.mdx`, `post.en-US.mdx` patterns.\n * Returns undefined for files without a locale suffix.\n */\nexport function parseLocaleFromFilename(filename: string): string | undefined {\n const match = LOCALE_REGEX.exec(filename);\n return match?.[1];\n}\n\n/**\n * Removes the locale suffix from a slug.\n * Handles both pre-slugify (`.pt`) and post-slugify (`-pt`) formats,\n * since `@sindresorhus/slugify` converts dots to dashes.\n * `post.pt` → `post`, `post-pt` → `post`, `post` → `post`\n */\nexport function stripLocaleFromSlug(slug: string, locale?: string): string {\n if (!locale) return slug;\n const dotSuffix = `.${locale}`;\n if (slug.endsWith(dotSuffix)) return slug.slice(0, -dotSuffix.length);\n const dashSuffix = `-${locale}`;\n if (slug.endsWith(dashSuffix)) return slug.slice(0, -dashSuffix.length);\n return slug;\n}\n","/**\n * @context Core layer — content store at src/core/content-store.ts\n * @does Manages a singleton ContentIndex; exposes loadContent() and getStore() for consumers\n * @depends src/core/indexer.ts, src/shared/types.ts\n * @do Use this as the single access point for in-memory indexed content\n * @dont Import from CLI or UI; contain parsing or I/O logic; import fs-adapter at top level\n */\n\nimport type { IFsAdapter } from \"../shared/fs-adapter.interface.js\";\nimport type { StudioConfig } from \"../shared/types.js\";\nimport { ContentIndex } from \"./indexer.js\";\n\nlet store: ContentIndex | null = null;\n\nexport function getStore(): ContentIndex {\n if (!store) {\n throw new Error(\n \"Content not loaded. Auto-init requires 'nextjs-studio/server' — \" +\n \"use loadContentSync() in a server context, or queryCollection() \" +\n \"will auto-init when imported from 'nextjs-studio/server'.\",\n );\n }\n return store;\n}\n\nexport function setStore(index: ContentIndex): void {\n store = index;\n}\n\nexport function hasStore(): boolean {\n return store !== null;\n}\n\nexport async function loadContent(\n fsAdapter: IFsAdapter,\n config?: StudioConfig,\n): Promise<ContentIndex> {\n const index = new ContentIndex(fsAdapter);\n await index.build(config);\n store = index;\n return index;\n}\n\nexport function loadContentSync(\n fsAdapter: IFsAdapter,\n config?: StudioConfig,\n): ContentIndex {\n const index = new ContentIndex(fsAdapter);\n index.buildSync(config);\n store = index;\n return index;\n}\n","/**\n * @context Core layer — config loader at src/core/config-loader.ts\n * @does Resolves and loads studio.config.ts/.js from the project root using dynamic import\n * @depends src/shared/constants.ts, src/shared/types.ts\n * @do Add new config resolution strategies or validation here\n * @dont Import from CLI or UI; access content files\n */\n\nimport { existsSync } from \"node:fs\";\nimport path from \"node:path\";\nimport { pathToFileURL } from \"node:url\";\nimport type { StudioConfig } from \"../shared/types.js\";\nimport { CONFIG_FILENAMES } from \"../shared/constants.js\";\n\n/**\n * Resolves the config file path from the project root.\n * Returns undefined if no config file is found.\n */\nexport function resolveConfigPath(projectRoot: string): string | undefined {\n for (const filename of CONFIG_FILENAMES) {\n const fullPath = path.resolve(projectRoot, filename);\n if (existsSync(fullPath)) return fullPath;\n }\n return undefined;\n}\n\n/**\n * Loads the studio config from the project root.\n * Tries CONFIG_FILENAMES in order, uses dynamic import().\n * Returns empty config if no file found or loading fails.\n */\nexport async function loadStudioConfig(projectRoot: string): Promise<StudioConfig> {\n const configPath = resolveConfigPath(projectRoot);\n if (!configPath) return {};\n\n return loadConfigFromPath(configPath);\n}\n\n/**\n * Loads config from a specific file path.\n */\nexport async function loadConfigFromPath(configPath: string): Promise<StudioConfig> {\n try {\n const fileUrl = pathToFileURL(configPath).href;\n const mod = await import(/* webpackIgnore: true */ fileUrl);\n const config = mod.default ?? mod.config ?? mod;\n\n if (typeof config !== \"object\" || config === null || Array.isArray(config)) {\n return {};\n }\n\n return config as StudioConfig;\n } catch {\n return {};\n }\n}\n","/**\n * @context Core layer — query builder at src/core/query-builder.ts\n * @does Provides a fluent API to filter, sort, and paginate content entries from a collection\n * @depends src/shared/types.ts, src/core/content-store.ts\n * @do Add new query capabilities here (e.g. search, groupBy)\n * @dont Import from CLI or UI; access the filesystem; perform I/O\n */\n\nimport { filter, orderBy, get, slice } from \"lodash-es\";\nimport type { QueryOptions, EntryResult } from \"../shared/types.js\";\nimport type { CollectionTypeMap } from \"../shared/types.js\";\nimport { getStore } from \"./content-store.js\";\nimport { isDraft } from \"./draft-filter.js\";\n\n/**\n * Fluent query builder for content collections.\n * Returned by `queryCollection()` — supports both chaining and direct array usage.\n *\n * ```ts\n * Fluent chaining\n * queryCollection(\"blog\").where({ published: true }).sort(\"date\", \"desc\").limit(10).all()\n *\n * Direct array usage — all native JS array methods work\n * queryCollection(\"blog\").slice(0, 5)\n * queryCollection(\"blog\").map(post => post.title)\n * ```\n */\nexport class QueryBuilder<T = Record<string, unknown>> {\n private readonly collectionName: string;\n private options: QueryOptions = {};\n private _excludeDrafts = false;\n private _locale: string | undefined;\n\n constructor(collection: string) {\n this.collectionName = collection;\n }\n\n where(conditions: Record<string, unknown>): this {\n this.options.where = { ...this.options.where, ...conditions };\n return this;\n }\n\n sort(field: string, order: \"asc\" | \"desc\" = \"asc\"): this {\n this.options.sort = { field, order };\n return this;\n }\n\n limit(count: number): this {\n this.options.limit = count;\n return this;\n }\n\n offset(count: number): this {\n this.options.offset = count;\n return this;\n }\n\n excludeDrafts(): this {\n this._excludeDrafts = true;\n return this;\n }\n\n locale(code: string): this {\n this._locale = code;\n return this;\n }\n\n all(): T[] {\n let entries = [...getStore().getCollection(this.collectionName)];\n\n if (this._excludeDrafts) {\n entries = entries.filter((entry) => !isDraft(entry));\n }\n\n if (this._locale) {\n const loc = this._locale;\n entries = entries.filter((entry) => entry.locale === loc);\n }\n\n if (this.options.where) {\n const conditions = this.options.where;\n entries = filter(entries, (entry) =>\n Object.entries(conditions).every(([key, value]) => get(entry.data, key) === value),\n );\n }\n\n if (this.options.sort) {\n const { field, order } = this.options.sort;\n entries = orderBy(entries, [(entry) => get(entry.data, field)], [order]);\n }\n\n const start = this.options.offset ?? 0;\n const end = this.options.limit ? start + this.options.limit : undefined;\n return slice(entries, start, end).map((e) => ({\n collection: e.collection,\n slug: e.slug,\n path: e.path,\n body: e.body,\n locale: e.locale,\n ...e.data,\n }) as unknown as T);\n }\n\n first(): T | undefined {\n return this.limit(1).all()[0];\n }\n\n one(): T {\n const result = this.limit(1).all()[0];\n if (result === undefined) {\n throw new Error(`Collection \"${this.collectionName}\" is empty — expected exactly one entry.`);\n }\n return result;\n }\n\n count(): number {\n return this.all().length;\n }\n}\n\n/** Intersection type: fluent builder + full native Array<T> interface. */\nexport type QueryResult<T> = QueryBuilder<T> & T[];\n\n/**\n * Wraps a QueryBuilder in a Proxy that delegates any unknown property access\n * to the resolved array. The array result is cached and invalidated whenever\n * a fluent method (where/sort/limit/offset) is called.\n */\nconst FLUENT_METHODS = new Set([\"where\", \"sort\", \"limit\", \"offset\", \"excludeDrafts\", \"locale\"]);\n\nfunction wrapWithArrayProxy<T>(builder: QueryBuilder<T>): QueryResult<T> {\n let cache: T[] | null = null;\n\n return new Proxy(builder, {\n get(target, prop, receiver) {\n // Fluent methods: apply on target, invalidate cache, return proxy for chaining\n if (FLUENT_METHODS.has(String(prop))) {\n const method = Reflect.get(target, prop) as (...args: unknown[]) => unknown;\n return (...args: unknown[]) => {\n cache = null;\n method.apply(target, args);\n return receiver;\n };\n }\n\n // Own QueryBuilder methods (all, first, count, etc.)\n if (prop in target) {\n const value = Reflect.get(target, prop, receiver);\n return typeof value === \"function\" ? value.bind(target) : value;\n }\n\n // Delegate everything else (slice, map, filter, length, etc.) to the resolved array\n if (!cache) cache = target.all();\n const value = Reflect.get(cache, prop);\n return typeof value === \"function\" ? value.bind(cache) : value;\n },\n }) as unknown as QueryResult<T>;\n}\n\n/**\n * Entry point for querying a content collection.\n */\nexport function queryCollection<K extends keyof CollectionTypeMap>(\n name: K,\n): QueryResult<EntryResult<CollectionTypeMap[K]>>;\nexport function queryCollection(name: string): QueryResult<EntryResult>;\nexport function queryCollection(name: string): QueryResult<EntryResult> {\n return wrapWithArrayProxy(new QueryBuilder(name));\n}\n","/**\n * @context Core layer — draft filter at src/core/draft-filter.ts\n * @does Provides utilities to detect and filter draft content entries\n * @depends src/shared/types.ts\n * @do Add new draft detection heuristics here\n * @dont Import from CLI or UI; access filesystem\n */\n\nimport type { ContentEntry } from \"../shared/types.js\";\n\nexport function isDraft(entry: ContentEntry): boolean {\n return entry.data.draft === true;\n}\n\nexport function filterDrafts(entries: ContentEntry[]): ContentEntry[] {\n return entries.filter((entry) => !isDraft(entry));\n}\n"],"mappings":";AAQA,OAAOA,WAAU;;;ACAV,IAAM,eAAe;AAGrB,IAAM,mBAAmB,CAAC,oBAAoB,oBAAoB,mBAAmB;AACrF,IAAM,uBAAuB,CAAC,QAAQ,OAAO;AAC7C,IAAM,wBAAwB;AAI9B,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,aAAa,cAAc,WAAW;AAEhE,IAAM,mBAAmB;AAAA,EAC9B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF;AAEO,IAAM,mBAAmB,CAAC,GAAG,kBAAkB,GAAG,kBAAkB,GAAG,gBAAgB;;;AC7B9F,OAAO,QAAQ;AACf,OAAO,YAAY;AACnB,OAAO,UAAU;AAMV,IAAM,YAAN,MAAsC;AAAA,EAC1B;AAAA,EAEjB,YAAY,UAAkB;AAC5B,SAAK,WAAW,KAAK,QAAQ,QAAQ;AAAA,EACvC;AAAA,EAEQ,WAAW,UAA4B;AAC7C,WAAO,KAAK,QAAQ,KAAK,UAAU,GAAG,QAAQ;AAAA,EAChD;AAAA,EAEA,MAAM,SAAS,UAAmC;AAChD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,GAAG,OAAO;AAAA,EACpD;AAAA,EAEA,MAAM,UAAU,UAAkB,SAAgC;AAChE,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,SAAS,OAAO;AAAA,EAC/C;AAAA,EAEA,MAAM,WAAW,UAAiC;AAChD,UAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AAAA,EACxC;AAAA,EAEA,MAAM,OAAO,UAAoC;AAC/C,QAAI;AACF,YAAM,GAAG,OAAO,KAAK,QAAQ,QAAQ,CAAC;AACtC,aAAO;AAAA,IACT,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,SAAS,UAAqC;AAClD,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,QAAQ,MAAM,GAAG,KAAK,QAAQ;AACpC,WAAO,EAAE,MAAM,UAAU,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM;AAAA,EACrE;AAAA,EAEA,MAAM,UAAU,SAAiB,YAAmD;AAClF,UAAM,WAAW,KAAK,QAAQ,OAAO;AACrC,UAAM,aAAa,cAAc;AAEjC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,OAAO,KAAK,WAAW,KAAK,CAAC,QAAQ,MAAM,KAAK,SAAS,GAAG,CAAC,CAAC,EACtF,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,gBAAgB,SAAoC;AACxD,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,YAAY,CAAC,EACrC,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,MAAM,WAAW,UAAmC;AAClD,WAAO,GAAG,SAAS,KAAK,QAAQ,QAAQ,CAAC;AAAA,EAC3C;AAAA,EAEA,MAAM,YAAY,UAAkB,MAA6B;AAC/D,UAAM,WAAW,KAAK,QAAQ,QAAQ;AACtC,UAAM,GAAG,MAAM,KAAK,QAAQ,QAAQ,GAAG,EAAE,WAAW,KAAK,CAAC;AAC1D,UAAM,GAAG,UAAU,UAAU,IAAI;AAAA,EACnC;AAAA,EAEA,MAAM,aAAa,SAAgD;AACjE,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,MAAM,GAAG,QAAQ,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAC9D,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,UAAM,UAAgC,CAAC;AACvC,eAAW,SAAS,SAAS;AAC3B,UAAI,CAAC,MAAM,OAAO,EAAG;AACrB,YAAM,eAAe,KAAK,KAAK,SAAS,MAAM,IAAI;AAClD,YAAM,QAAQ,MAAM,GAAG,KAAK,KAAK,QAAQ,YAAY,CAAC;AACtD,cAAQ,KAAK,EAAE,MAAM,MAAM,MAAM,cAAc,MAAM,MAAM,MAAM,YAAY,MAAM,MAAM,CAAC;AAAA,IAC5F;AACA,WAAO;AAAA,EACT;AAAA,EAEA,QAAQ,UAA4B;AAClC,WAAO,KAAK,KAAK,GAAG,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,UAA0B;AACjC,WAAO,KAAK,SAAS,QAAQ;AAAA,EAC/B;AAAA,EAEA,QAAQ,UAA0B;AAChC,WAAO,KAAK,QAAQ,QAAQ;AAAA,EAC9B;AAAA,EAEA,SAAS,MAAc,IAAoB;AACzC,WAAO,KAAK,SAAS,MAAM,EAAE;AAAA,EAC/B;AAAA,EAEA,cAAc,cAAsB,KAAqB;AACvD,WAAO,aAAa,QAAQ,KAAK,EAAE,EAAE,MAAM,KAAK,GAAG,EAAE,KAAK,GAAG;AAAA,EAC/D;AAAA,EAEA,aAAa,UAA0B;AACrC,WAAO,OAAO,aAAa,KAAK,QAAQ,QAAQ,GAAG,OAAO;AAAA,EAC5D;AAAA,EAEA,WAAW,UAA2B;AACpC,WAAO,OAAO,WAAW,KAAK,QAAQ,QAAQ,CAAC;AAAA,EACjD;AAAA,EAEA,cAAc,SAAiB,YAA0C;AACvE,UAAM,WAAW,KAAK,QAAQ,OAAO;AACrC,UAAM,aAAa,cAAc;AAEjC,QAAI;AACJ,QAAI;AACF,gBAAU,OAAO,YAAY,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAChE,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,OAAO,KAAK,WAAW,KAAK,CAAC,QAAQ,MAAM,KAAK,SAAS,GAAG,CAAC,CAAC,EACtF,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AAAA,EAEA,oBAAoB,SAA2B;AAC7C,UAAM,WAAW,KAAK,QAAQ,OAAO;AAErC,QAAI;AACJ,QAAI;AACF,gBAAU,OAAO,YAAY,UAAU,EAAE,eAAe,KAAK,CAAC;AAAA,IAChE,QAAQ;AACN,aAAO,CAAC;AAAA,IACV;AAEA,WAAO,QACJ,OAAO,CAAC,UAAU,MAAM,YAAY,CAAC,EACrC,IAAI,CAAC,UAAU,KAAK,KAAK,SAAS,MAAM,IAAI,CAAC;AAAA,EAClD;AACF;;;ACvKA,OAAO,aAAa;;;ACApB,OAAO,YAAY;;;ACAnB,IAAM,cAAc;AAMb,SAAS,gBAAgB,MAAc,MAAuC;AACnF,SAAO,KAAK,QAAQ,aAAa,CAAC,QAAQC,UAAiB;AACzD,UAAM,QAAQ,YAAY,MAAMA,KAAI;AACpC,QAAI,UAAU,UAAa,UAAU,KAAM,QAAO;AAClD,QAAI,OAAO,UAAU,SAAU,QAAO,KAAK,UAAU,KAAK;AAC1D,WAAO,OAAO,KAAK;AAAA,EACrB,CAAC;AACH;AAeA,SAAS,YAAY,KAA8BC,OAAuB;AACxE,QAAM,OAAOA,MAAK,MAAM,GAAG;AAC3B,MAAI,UAAmB;AACvB,aAAW,OAAO,MAAM;AACtB,QAAI,OAAO,YAAY,YAAY,YAAY,KAAM,QAAO;AAC5D,cAAW,QAAoC,GAAG;AAAA,EACpD;AACA,SAAO;AACT;;;ADtBA,SAAS,eAAe,MAAwD;AAC9E,QAAM,SAAkC,CAAC;AACzC,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,IAAI,GAAG;AAC/C,QAAI,iBAAiB,MAAM;AACzB,aAAO,GAAG,IAAI,MAAM,YAAY,EAAE,MAAM,GAAG,EAAE,CAAC;AAAA,IAChD,WAAW,OAAO,UAAU,YAAY,UAAU,QAAQ,CAAC,MAAM,QAAQ,KAAK,GAAG;AAC/E,aAAO,GAAG,IAAI,eAAe,KAAgC;AAAA,IAC/D,OAAO;AACL,aAAO,GAAG,IAAI;AAAA,IAChB;AAAA,EACF;AACA,SAAO;AACT;AAEO,SAAS,SAAS,SAAiB,SAAsC;AAC9E,QAAM,EAAE,MAAM,SAAS,SAAS,KAAK,IAAI,OAAO,OAAO;AACvD,QAAM,OAAO,eAAe,OAAO;AACnC,QAAM,UAAU,KAAK,KAAK;AAC1B,SAAO;AAAA,IACL;AAAA,IACA,MAAM,SAAS,aAAa,gBAAgB,SAAS,IAAI,IAAI;AAAA,EAC/D;AACF;;;AExBO,SAAS,UAAU,SAA6B;AACrD,QAAM,SAAkB,KAAK,MAAM,OAAO;AAE1C,MAAI,MAAM,QAAQ,MAAM,GAAG;AACzB,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS;AAAA,IACX;AAAA,EACF;AAEA,MAAI,OAAO,WAAW,YAAY,WAAW,MAAM;AACjD,WAAO;AAAA,MACL,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,IAAI,MAAM,yCAAyC;AAC3D;;;AC1BA,IAAM,cAAc;AACpB,IAAM,kBACJ;AACF,IAAM,WAAW;AACjB,IAAM,SAAS;AACf,IAAM,sBAAsB;AAC5B,IAAM,wBAAwB,CAAC,eAAe,gBAAgB,QAAQ,SAAS;AAE/E,SAAS,UAAU,OAAwB;AACzC,SAAO,YAAY,KAAK,KAAK;AAC/B;AAEA,SAAS,cAAc,OAAwB;AAC7C,SAAO,gBAAgB,KAAK,KAAK;AACnC;AAEA,SAAS,QAAQ,OAAwB;AACvC,SAAO,SAAS,KAAK,KAAK;AAC5B;AAEA,SAAS,MAAM,OAAwB;AACrC,SAAO,OAAO,KAAK,KAAK;AAC1B;AAEA,SAAS,iBAAiB,MAAc,SAAoC;AAC1E,MAAI,QAAQ,MAAM,OAAO,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ;AACzD,MAAI,QAAQ,MAAM,KAAK,EAAG,QAAO,EAAE,MAAM,MAAM,MAAM;AACrD,MAAI,QAAQ,MAAM,aAAa,EAAG,QAAO,EAAE,MAAM,MAAM,QAAQ,aAAa,KAAK;AACjF,MAAI,QAAQ,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AAG1D,MAAI,sBAAsB,SAAS,KAAK,YAAY,CAAC,EAAG,QAAO,EAAE,MAAM,MAAM,YAAY;AAEzF,QAAM,SAAS,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,uBAAuB,EAAE,SAAS,IAAI,CAAC;AACrF,SAAO,EAAE,MAAM,MAAM,SAAS,cAAc,OAAO;AACrD;AAEA,SAAS,gBAAgB,MAAc,OAAmC;AACxE,MAAI,MAAM,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAErE,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,QAAQ,GAAG;AACnD,UAAM,SAAS,CAAC,GAAG,IAAI,IAAI,KAAiB,CAAC,EAAE,MAAM,GAAG,EAAE;AAC1D,UAAM,UAA0B,OAAO,IAAI,CAAC,OAAO,EAAE,OAAO,GAAG,OAAO,EAAE,EAAE;AAC1E,WAAO,EAAE,MAAM,MAAM,gBAAgB,QAAQ;AAAA,EAC/C;AAEA,MAAI,MAAM,MAAM,CAAC,SAAS,OAAO,SAAS,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,IAAI,CAAC,GAAG;AAC5F,WAAO,EAAE,MAAM,MAAM,SAAS,YAAY,YAAY,KAAkC,EAAE;AAAA,EAC5F;AAEA,SAAO,EAAE,MAAM,MAAM,SAAS,YAAY,CAAC,EAAE;AAC/C;AAEA,SAAS,qBAAqB,MAAc,QAAoC;AAC9E,QAAM,UAAU,OAAO,OAAO,CAAC,MAAM,MAAM,QAAQ,MAAM,MAAS;AAElE,MAAI,QAAQ,WAAW,EAAG,QAAO,EAAE,MAAM,MAAM,OAAO;AACtD,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,SAAS,EAAG,QAAO,EAAE,MAAM,MAAM,UAAU;AAEjF,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,UAAM,SAAS,QAAQ,MAAM,CAAC,MAAM,OAAO,UAAU,CAAC,CAAC,IAAI,YAAY;AACvE,WAAO,EAAE,MAAM,MAAM,UAAU,OAAO;AAAA,EACxC;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,QAAQ,GAAG;AAC/C,WAAO,iBAAiB,MAAM,OAAmB;AAAA,EACnD;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,MAAM,QAAQ,CAAC,CAAC,GAAG;AAC1C,WAAO,gBAAgB,MAAO,QAAwB,KAAK,CAAC;AAAA,EAC9D;AAEA,MAAI,QAAQ,MAAM,CAAC,MAAM,OAAO,MAAM,YAAY,MAAM,QAAQ,CAAC,MAAM,QAAQ,CAAC,CAAC,GAAG;AAClF,WAAO,EAAE,MAAM,MAAM,UAAU,QAAQ,YAAY,OAAoC,EAAE;AAAA,EAC3F;AAEA,SAAO,EAAE,MAAM,MAAM,OAAO;AAC9B;AAEA,SAAS,YAAY,MAAoD;AACvE,QAAM,SAAS,IAAI,IAAY,KAAK,QAAQ,CAAC,QAAQ,OAAO,KAAK,GAAG,CAAC,CAAC;AACtE,SAAO,MAAM,KAAK,MAAM,EAAE,IAAI,CAAC,QAAQ,qBAAqB,KAAK,KAAK,IAAI,CAAC,QAAQ,IAAI,GAAG,CAAC,CAAC,CAAC;AAC/F;AASO,SAAS,YAAY,SAAyB,gBAA0C;AAC7F,QAAM,OAAO,QAAQ,IAAI,CAAC,UAAU,MAAM,IAA+B;AACzE,SAAO,EAAE,YAAY,gBAAgB,QAAQ,YAAY,IAAI,EAAE;AACjE;;;AClGA,IAAM,eAAe;AAOd,SAAS,wBAAwB,UAAsC;AAC5E,QAAM,QAAQ,aAAa,KAAK,QAAQ;AACxC,SAAO,QAAQ,CAAC;AAClB;AAQO,SAAS,oBAAoB,MAAc,QAAyB;AACzE,MAAI,CAAC,OAAQ,QAAO;AACpB,QAAM,YAAY,IAAI,MAAM;AAC5B,MAAI,KAAK,SAAS,SAAS,EAAG,QAAO,KAAK,MAAM,GAAG,CAAC,UAAU,MAAM;AACpE,QAAM,aAAa,IAAI,MAAM;AAC7B,MAAI,KAAK,SAAS,UAAU,EAAG,QAAO,KAAK,MAAM,GAAG,CAAC,WAAW,MAAM;AACtE,SAAO;AACT;;;ALfO,IAAM,eAAN,MAAmB;AAAA,EACP,UAAU,oBAAI,IAA4B;AAAA,EAC1C,cAAc,oBAAI,IAAwB;AAAA,EAC1C;AAAA,EAEjB,YAAY,WAAuB;AACjC,SAAK,KAAK;AAAA,EACZ;AAAA,EAEA,MAAM,MAAM,QAAsC;AAChD,SAAK,MAAM;AACX,UAAM,OAAO,MAAM,KAAK,GAAG,gBAAgB,GAAG;AAE9C,eAAW,OAAO,MAAM;AACtB,YAAM,UAAU,KAAK,GAAG,SAAS,GAAG;AACpC,YAAM,iBAAiB,QAAQ,OAAO;AACtC,YAAM,mBAAmB,QAAQ,cAAc,cAAc;AAC7D,YAAM,KAAK,gBAAgB,SAAS,gBAAgB,kBAAkB,MAAM;AAAA,IAC9E;AAAA,EACF;AAAA,EAEA,UAAU,QAA6B;AACrC,SAAK,MAAM;AACX,UAAM,OAAO,KAAK,GAAG,oBAAoB,GAAG;AAE5C,eAAW,OAAO,MAAM;AACtB,YAAM,UAAU,KAAK,GAAG,SAAS,GAAG;AACpC,YAAM,iBAAiB,QAAQ,OAAO;AACtC,YAAM,mBAAmB,QAAQ,cAAc,cAAc;AAC7D,WAAK,oBAAoB,SAAS,gBAAgB,kBAAkB,MAAM;AAAA,IAC5E;AAAA,EACF;AAAA,EAEA,cAAc,MAA8B;AAC1C,WAAO,KAAK,QAAQ,IAAI,IAAI,KAAK,CAAC;AAAA,EACpC;AAAA,EAEA,iBAA+B;AAC7B,WAAO,MAAM,KAAK,KAAK,YAAY,OAAO,CAAC;AAAA,EAC7C;AAAA,EAEA,QAAc;AACZ,SAAK,QAAQ,MAAM;AACnB,SAAK,YAAY,MAAM;AAAA,EACzB;AAAA,EAEA,YAAY,gBAAwB,OAA2B;AAC7D,UAAM,UAAU,KAAK,QAAQ,IAAI,cAAc,KAAK,CAAC;AACrD,UAAM,MAAM,QAAQ,UAAU,CAAC,MAAM,EAAE,SAAS,MAAM,IAAI;AAC1D,QAAI,OAAO,GAAG;AACZ,cAAQ,GAAG,IAAI;AAAA,IACjB,OAAO;AACL,cAAQ,KAAK,KAAK;AAAA,IACpB;AACA,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,qBAAqB,cAAc;AAAA,EAC1C;AAAA,EAEA,YAAY,gBAAwB,MAAoB;AACtD,UAAM,UAAU,KAAK,QAAQ,IAAI,cAAc;AAC/C,QAAI,CAAC,QAAS;AACd,UAAM,WAAW,QAAQ,OAAO,CAAC,MAAM,EAAE,SAAS,IAAI;AACtD,SAAK,QAAQ,IAAI,gBAAgB,QAAQ;AACzC,SAAK,qBAAqB,cAAc;AAAA,EAC1C;AAAA,EAEQ,qBAAqB,gBAA8B;AACzD,UAAM,MAAM,KAAK,YAAY,IAAI,cAAc;AAC/C,UAAM,UAAU,KAAK,QAAQ,IAAI,cAAc,KAAK,CAAC;AACrD,QAAI,KAAK;AACP,WAAK,YAAY,IAAI,gBAAgB;AAAA,QACnC,GAAG;AAAA,QACH,OAAO,QAAQ;AAAA,QACf,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACzC,CAAC;AAAA,IACH;AAAA,EACF;AAAA,EAEA,MAAc,gBACZ,SACA,gBACA,cACe;AACf,UAAM,UAA0B,CAAC;AACjC,UAAM,KAAK,QAAQ,SAAS,gBAAgB,SAAS,OAAO;AAE5D,UAAM,YAAY,KAAK,GAAG,KAAK,SAAS,qBAAqB;AAC7D,UAAM,WAAW,MAAM,KAAK,aAAa,SAAS;AAClD,QAAI,UAAU;AACZ,WAAK,cAAc,SAAS,QAAQ;AAAA,IACtC;AAEA,UAAM,SAAS,gBAAgB,YAAY,SAAS,cAAc;AAElE,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,YAAY,IAAI,gBAAgB;AAAA,MACnC,MAAM;AAAA,MACN,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACvC,OAAO,QAAQ;AAAA,MACf,UAAU;AAAA,MACV;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEQ,oBACN,SACA,gBACA,cACM;AACN,UAAM,UAA0B,CAAC;AACjC,SAAK,YAAY,SAAS,gBAAgB,SAAS,OAAO;AAE1D,UAAM,YAAY,KAAK,GAAG,KAAK,SAAS,qBAAqB;AAC7D,UAAM,WAAW,KAAK,iBAAiB,SAAS;AAChD,QAAI,UAAU;AACZ,WAAK,cAAc,SAAS,QAAQ;AAAA,IACtC;AAEA,UAAM,SAAS,gBAAgB,YAAY,SAAS,cAAc;AAElE,SAAK,QAAQ,IAAI,gBAAgB,OAAO;AACxC,SAAK,YAAY,IAAI,gBAAgB;AAAA,MACnC,MAAM;AAAA,MACN,MAAM,KAAK,qBAAqB,OAAO;AAAA,MACvC,OAAO,QAAQ;AAAA,MACf,UAAU;AAAA,MACV;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAc,QACZ,SACA,gBACA,SACA,SACe;AACf,UAAM,UAAU,MAAM,KAAK,GAAG,gBAAgB,OAAO;AACrD,eAAW,UAAU,SAAS;AAC5B,YAAM,KAAK,QAAQ,SAAS,gBAAgB,QAAQ,OAAO;AAAA,IAC7D;AAEA,UAAM,QAAQ,MAAM,KAAK,GAAG,UAAU,OAAO;AAC7C,eAAW,YAAY,OAAO;AAC5B,YAAM,WAAW,KAAK,GAAG,SAAS,QAAQ;AAC1C,UAAI,aAAa,sBAAuB;AAExC,YAAM,MAAM,KAAK,GAAG,QAAQ,QAAQ;AACpC,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,QAAQ;AAC/C,YAAM,eAAe,KAAK,GAAG,SAAS,SAAS,QAAQ;AACvD,YAAM,OAAO,KAAK,GACf,cAAc,cAAc,GAAG,EAC/B,MAAM,GAAG,EACT,IAAI,CAAC,YAAY,QAAQ,OAAO,CAAC,EACjC,KAAK,GAAG;AAEX,UAAI,QAAQ,QAAQ;AAClB,gBAAQ,KAAK,KAAK,cAAc,gBAAgB,MAAM,UAAU,OAAO,CAAC;AAAA,MAC1E,WAAW,QAAQ,SAAS;AAC1B,gBAAQ,KAAK,GAAG,KAAK,iBAAiB,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACtE;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,YACN,SACA,gBACA,SACA,SACM;AACN,UAAM,UAAU,KAAK,GAAG,oBAAoB,OAAO;AACnD,eAAW,UAAU,SAAS;AAC5B,WAAK,YAAY,SAAS,gBAAgB,QAAQ,OAAO;AAAA,IAC3D;AAEA,UAAM,QAAQ,KAAK,GAAG,cAAc,OAAO;AAC3C,eAAW,YAAY,OAAO;AAC5B,YAAM,WAAW,KAAK,GAAG,SAAS,QAAQ;AAC1C,UAAI,aAAa,sBAAuB;AAExC,YAAM,MAAM,KAAK,GAAG,QAAQ,QAAQ;AACpC,YAAM,UAAU,KAAK,GAAG,aAAa,QAAQ;AAC7C,YAAM,eAAe,KAAK,GAAG,SAAS,SAAS,QAAQ;AACvD,YAAM,OAAO,KAAK,GACf,cAAc,cAAc,GAAG,EAC/B,MAAM,GAAG,EACT,IAAI,CAAC,YAAY,QAAQ,OAAO,CAAC,EACjC,KAAK,GAAG;AAEX,UAAI,QAAQ,QAAQ;AAClB,gBAAQ,KAAK,KAAK,cAAc,gBAAgB,MAAM,UAAU,OAAO,CAAC;AAAA,MAC1E,WAAW,QAAQ,SAAS;AAC1B,gBAAQ,KAAK,GAAG,KAAK,iBAAiB,gBAAgB,MAAM,OAAO,CAAC;AAAA,MACtE;AAAA,IACF;AAAA,EACF;AAAA,EAEQ,cAAc,gBAAwB,MAAc,UAAkB,SAA+B;AAC3G,UAAM,SAAS,SAAS,OAAO;AAC/B,UAAM,SAAS,wBAAwB,QAAQ;AAC/C,UAAM,iBAAiB,oBAAoB,MAAM,MAAM;AACvD,WAAO;AAAA,MACL,YAAY;AAAA,MACZ,MAAM;AAAA,MACN,MAAM,IAAI,cAAc,IAAI,cAAc;AAAA,MAC1C,MAAM,OAAO;AAAA,MACb,MAAM,OAAO;AAAA,MACb,GAAI,SAAS,EAAE,OAAO,IAAI,CAAC;AAAA,IAC7B;AAAA,EACF;AAAA,EAEQ,iBAAiB,gBAAwB,MAAc,SAAiC;AAC9F,UAAM,SAAS,UAAU,OAAO;AAEhC,QAAI,OAAO,SAAS,cAAc;AAChC,aAAO,OAAO,QAAQ,IAAI,CAAC,MAAM,UAAU;AACzC,cAAM,YACJ,OAAO,KAAK,MAAM,MAAM,WAAW,QAAQ,KAAK,MAAM,CAAC,IAAI,GAAG,IAAI,IAAI,KAAK;AAC7E,eAAO;AAAA,UACL,YAAY;AAAA,UACZ,MAAM;AAAA,UACN,MAAM,IAAI,cAAc,IAAI,SAAS;AAAA,UACrC;AAAA,QACF;AAAA,MACF,CAAC;AAAA,IACH;AAEA,WAAO,CAAC,EAAE,YAAY,gBAAgB,MAAM,MAAM,IAAI,cAAc,IAAI,IAAI,IAAI,MAAM,OAAO,KAAK,CAAC;AAAA,EACrG;AAAA,EAEA,MAAc,aAAa,WAA6C;AACtE,QAAI,CAAE,MAAM,KAAK,GAAG,OAAO,SAAS,EAAI,QAAO;AAE/C,QAAI;AACF,YAAM,UAAU,MAAM,KAAK,GAAG,SAAS,SAAS;AAChD,YAAM,SAAkB,KAAK,MAAM,OAAO;AAC1C,UAAI,MAAM,QAAQ,MAAM,EAAG,QAAO;AAAA,IACpC,SAAS,OAAO;AACd,cAAQ,KAAK,kDAAkD,SAAS,IAAI,KAAK;AAAA,IACnF;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,iBAAiB,WAAoC;AAC3D,QAAI,CAAC,KAAK,GAAG,WAAW,SAAS,EAAG,QAAO;AAE3C,QAAI;AACF,YAAM,UAAU,KAAK,GAAG,aAAa,SAAS;AAC9C,YAAM,SAAkB,KAAK,MAAM,OAAO;AAC1C,UAAI,MAAM,QAAQ,MAAM,EAAG,QAAO;AAAA,IACpC,SAAS,OAAO;AACd,cAAQ,KAAK,kDAAkD,SAAS,IAAI,KAAK;AAAA,IACnF;AACA,WAAO;AAAA,EACT;AAAA,EAEQ,cAAc,SAAyB,UAA0B;AACvE,UAAM,WAAW,IAAI,IAAI,SAAS,IAAI,CAAC,MAAM,UAAU,CAAC,MAAM,KAAK,CAAC,CAAC;AACrE,YAAQ,KAAK,CAAC,GAAG,MAAM;AACrB,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,YAAM,SAAS,SAAS,IAAI,EAAE,IAAI,KAAK;AACvC,aAAO,SAAS;AAAA,IAClB,CAAC;AAAA,EACH;AAAA,EAEQ,qBAAqB,SAA6C;AACxE,QAAI,QAAQ,WAAW,EAAG,QAAO;AACjC,UAAM,QAAQ,QAAQ,CAAC;AACvB,QAAI,MAAM,SAAS,OAAW,QAAO;AACrC,QAAI,QAAQ,WAAW,KAAK,CAAC,MAAM,KAAK,SAAS,GAAG,EAAG,QAAO;AAC9D,WAAO;AAAA,EACT;AACF;;;AMrRA,IAAI,QAA6B;AAE1B,SAAS,WAAyB;AACvC,MAAI,CAAC,OAAO;AACV,UAAM,IAAI;AAAA,MACR;AAAA,IAGF;AAAA,EACF;AACA,SAAO;AACT;AAEO,SAAS,SAAS,OAA2B;AAClD,UAAQ;AACV;AAEO,SAAS,WAAoB;AAClC,SAAO,UAAU;AACnB;AAEA,eAAsB,YACpB,WACA,QACuB;AACvB,QAAM,QAAQ,IAAI,aAAa,SAAS;AACxC,QAAM,MAAM,MAAM,MAAM;AACxB,UAAQ;AACR,SAAO;AACT;AAEO,SAAS,gBACd,WACA,QACc;AACd,QAAM,QAAQ,IAAI,aAAa,SAAS;AACxC,QAAM,UAAU,MAAM;AACtB,UAAQ;AACR,SAAO;AACT;;;AC3CA,SAAS,kBAAkB;AAC3B,OAAOC,WAAU;AACjB,SAAS,qBAAqB;AAQvB,SAAS,kBAAkB,aAAyC;AACzE,aAAW,YAAY,kBAAkB;AACvC,UAAM,WAAWC,MAAK,QAAQ,aAAa,QAAQ;AACnD,QAAI,WAAW,QAAQ,EAAG,QAAO;AAAA,EACnC;AACA,SAAO;AACT;AAOA,eAAsB,iBAAiB,aAA4C;AACjF,QAAM,aAAa,kBAAkB,WAAW;AAChD,MAAI,CAAC,WAAY,QAAO,CAAC;AAEzB,SAAO,mBAAmB,UAAU;AACtC;AAKA,eAAsB,mBAAmB,YAA2C;AAClF,MAAI;AACF,UAAM,UAAU,cAAc,UAAU,EAAE;AAC1C,UAAM,MAAM,MAAM;AAAA;AAAA,MAAiC;AAAA;AACnD,UAAM,SAAS,IAAI,WAAW,IAAI,UAAU;AAE5C,QAAI,OAAO,WAAW,YAAY,WAAW,QAAQ,MAAM,QAAQ,MAAM,GAAG;AAC1E,aAAO,CAAC;AAAA,IACV;AAEA,WAAO;AAAA,EACT,QAAQ;AACN,WAAO,CAAC;AAAA,EACV;AACF;;;AC/CA,SAAS,QAAQ,SAAS,KAAK,aAAa;;;ACErC,SAAS,QAAQ,OAA8B;AACpD,SAAO,MAAM,KAAK,UAAU;AAC9B;;;ADeO,IAAM,eAAN,MAAgD;AAAA,EACpC;AAAA,EACT,UAAwB,CAAC;AAAA,EACzB,iBAAiB;AAAA,EACjB;AAAA,EAER,YAAY,YAAoB;AAC9B,SAAK,iBAAiB;AAAA,EACxB;AAAA,EAEA,MAAM,YAA2C;AAC/C,SAAK,QAAQ,QAAQ,EAAE,GAAG,KAAK,QAAQ,OAAO,GAAG,WAAW;AAC5D,WAAO;AAAA,EACT;AAAA,EAEA,KAAK,OAAe,QAAwB,OAAa;AACvD,SAAK,QAAQ,OAAO,EAAE,OAAO,MAAM;AACnC,WAAO;AAAA,EACT;AAAA,EAEA,MAAM,OAAqB;AACzB,SAAK,QAAQ,QAAQ;AACrB,WAAO;AAAA,EACT;AAAA,EAEA,OAAO,OAAqB;AAC1B,SAAK,QAAQ,SAAS;AACtB,WAAO;AAAA,EACT;AAAA,EAEA,gBAAsB;AACpB,SAAK,iBAAiB;AACtB,WAAO;AAAA,EACT;AAAA,EAEA,OAAO,MAAoB;AACzB,SAAK,UAAU;AACf,WAAO;AAAA,EACT;AAAA,EAEA,MAAW;AACT,QAAI,UAAU,CAAC,GAAG,SAAS,EAAE,cAAc,KAAK,cAAc,CAAC;AAE/D,QAAI,KAAK,gBAAgB;AACvB,gBAAU,QAAQ,OAAO,CAAC,UAAU,CAAC,QAAQ,KAAK,CAAC;AAAA,IACrD;AAEA,QAAI,KAAK,SAAS;AAChB,YAAM,MAAM,KAAK;AACjB,gBAAU,QAAQ,OAAO,CAAC,UAAU,MAAM,WAAW,GAAG;AAAA,IAC1D;AAEA,QAAI,KAAK,QAAQ,OAAO;AACtB,YAAM,aAAa,KAAK,QAAQ;AAChC,gBAAU;AAAA,QAAO;AAAA,QAAS,CAAC,UACzB,OAAO,QAAQ,UAAU,EAAE,MAAM,CAAC,CAAC,KAAK,KAAK,MAAM,IAAI,MAAM,MAAM,GAAG,MAAM,KAAK;AAAA,MACnF;AAAA,IACF;AAEA,QAAI,KAAK,QAAQ,MAAM;AACrB,YAAM,EAAE,OAAO,MAAM,IAAI,KAAK,QAAQ;AACtC,gBAAU,QAAQ,SAAS,CAAC,CAAC,UAAU,IAAI,MAAM,MAAM,KAAK,CAAC,GAAG,CAAC,KAAK,CAAC;AAAA,IACzE;AAEA,UAAM,QAAQ,KAAK,QAAQ,UAAU;AACrC,UAAM,MAAM,KAAK,QAAQ,QAAQ,QAAQ,KAAK,QAAQ,QAAQ;AAC9D,WAAO,MAAM,SAAS,OAAO,GAAG,EAAE,IAAI,CAAC,OAAO;AAAA,MAC5C,YAAY,EAAE;AAAA,MACd,MAAM,EAAE;AAAA,MACR,MAAM,EAAE;AAAA,MACR,MAAM,EAAE;AAAA,MACR,QAAQ,EAAE;AAAA,MACV,GAAG,EAAE;AAAA,IACP,EAAkB;AAAA,EACpB;AAAA,EAEA,QAAuB;AACrB,WAAO,KAAK,MAAM,CAAC,EAAE,IAAI,EAAE,CAAC;AAAA,EAC9B;AAAA,EAEA,MAAS;AACP,UAAM,SAAS,KAAK,MAAM,CAAC,EAAE,IAAI,EAAE,CAAC;AACpC,QAAI,WAAW,QAAW;AACxB,YAAM,IAAI,MAAM,eAAe,KAAK,cAAc,+CAA0C;AAAA,IAC9F;AACA,WAAO;AAAA,EACT;AAAA,EAEA,QAAgB;AACd,WAAO,KAAK,IAAI,EAAE;AAAA,EACpB;AACF;AAUA,IAAM,iBAAiB,oBAAI,IAAI,CAAC,SAAS,QAAQ,SAAS,UAAU,iBAAiB,QAAQ,CAAC;AAE9F,SAAS,mBAAsB,SAA0C;AACvE,MAAI,QAAoB;AAExB,SAAO,IAAI,MAAM,SAAS;AAAA,IACxB,IAAI,QAAQ,MAAM,UAAU;AAE1B,UAAI,eAAe,IAAI,OAAO,IAAI,CAAC,GAAG;AACpC,cAAM,SAAS,QAAQ,IAAI,QAAQ,IAAI;AACvC,eAAO,IAAI,SAAoB;AAC7B,kBAAQ;AACR,iBAAO,MAAM,QAAQ,IAAI;AACzB,iBAAO;AAAA,QACT;AAAA,MACF;AAGA,UAAI,QAAQ,QAAQ;AAClB,cAAMC,SAAQ,QAAQ,IAAI,QAAQ,MAAM,QAAQ;AAChD,eAAO,OAAOA,WAAU,aAAaA,OAAM,KAAK,MAAM,IAAIA;AAAA,MAC5D;AAGA,UAAI,CAAC,MAAO,SAAQ,OAAO,IAAI;AAC/B,YAAM,QAAQ,QAAQ,IAAI,OAAO,IAAI;AACrC,aAAO,OAAO,UAAU,aAAa,MAAM,KAAK,KAAK,IAAI;AAAA,IAC3D;AAAA,EACF,CAAC;AACH;AASO,SAAS,gBAAgB,MAAwC;AACtE,SAAO,mBAAmB,IAAI,aAAa,IAAI,CAAC;AAClD;;;AXzJA,IAAI,CAAC,SAAS,GAAG;AACf,QAAM,MAAMC,MAAK,KAAK,QAAQ,IAAI,GAAG,YAAY;AACjD,QAAM,QAAQ,IAAI,aAAa,IAAI,UAAU,GAAG,CAAC;AACjD,QAAM,UAAU;AAChB,WAAS,KAAK;AAChB;","names":["path","path","path","path","path","value","path"]}
|