@ulrichc1/sparn 1.2.1 → 1.2.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/core/btsp-embedder.ts","../src/utils/hash.ts","../src/core/confidence-states.ts","../src/core/engram-scorer.ts","../src/utils/tokenizer.ts","../src/core/sparse-pruner.ts","../src/utils/context-parser.ts","../src/adapters/claude-code.ts","../src/adapters/generic.ts","../src/core/budget-pruner.ts","../src/core/metrics.ts","../src/core/incremental-optimizer.ts","../src/core/context-pipeline.ts","../src/core/kv-memory.ts","../src/core/sleep-compressor.ts","../src/daemon/daemon-process.ts","../src/daemon/file-tracker.ts","../src/daemon/session-watcher.ts","../src/mcp/server.ts","../src/types/config.ts","../src/utils/logger.ts"],"sourcesContent":["/**\n * BTSP Embedder - Implements behavioral timescale synaptic plasticity\n *\n * Neuroscience: One-shot learning from critical events (errors, conflicts).\n * Application: Detect high-importance patterns and mark for permanent retention.\n */\n\nimport { randomUUID } from 'node:crypto';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from '../utils/hash.js';\n\nexport interface BTSPEmbedder {\n /**\n * Detect if content contains BTSP patterns (errors, stack traces, conflicts, git diffs)\n * @param content - Content to analyze\n * @returns True if BTSP pattern detected\n */\n detectBTSP(content: string): boolean;\n\n /**\n * Create a new memory entry marked as BTSP (one-shot learned)\n * @param content - Entry content\n * @param tags - Optional tags\n * @param metadata - Optional metadata\n * @returns BTSP-marked memory entry\n */\n createBTSPEntry(\n content: string,\n tags?: string[],\n metadata?: Record<string, unknown>,\n ): MemoryEntry;\n}\n\n/**\n * Create a BTSP embedder instance\n * @returns BTSPEmbedder instance\n */\nexport function createBTSPEmbedder(): BTSPEmbedder {\n // Patterns that indicate critical events\n const BTSP_PATTERNS = [\n // Error patterns\n /\\b(error|exception|failure|fatal|critical|panic)\\b/i,\n /\\b(TypeError|ReferenceError|SyntaxError|RangeError|URIError)\\b/,\n /\\bENOENT|EACCES|ECONNREFUSED|ETIMEDOUT\\b/,\n\n // Stack trace patterns\n /^\\s+at\\s+.*\\(.*:\\d+:\\d+\\)/m, // JavaScript stack trace\n /^\\s+at\\s+.*\\.[a-zA-Z]+:\\d+/m, // Python/Ruby stack trace\n\n // Git diff new files\n /^new file mode \\d+$/m,\n /^--- \\/dev\\/null$/m,\n\n // Merge conflict markers\n /^<<<<<<< /m,\n /^=======/m,\n /^>>>>>>> /m,\n ];\n\n function detectBTSP(content: string): boolean {\n return BTSP_PATTERNS.some((pattern) => pattern.test(content));\n }\n\n function createBTSPEntry(\n content: string,\n tags: string[] = [],\n metadata: Record<string, unknown> = {},\n ): MemoryEntry {\n return {\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: Date.now(),\n score: 1.0, // Maximum initial score\n ttl: 365 * 24 * 3600, // 1 year in seconds (long retention)\n state: 'active', // Always active\n accessCount: 0,\n tags: [...tags, 'btsp'],\n metadata,\n isBTSP: true,\n };\n }\n\n return {\n detectBTSP,\n createBTSPEntry,\n };\n}\n","/**\n * Content hashing utilities.\n * Uses SHA-256 for deduplication.\n */\n\nimport { createHash } from 'node:crypto';\n\n/**\n * Generate SHA-256 hash of content for deduplication.\n *\n * @param content - Content to hash\n * @returns 64-character hex string (SHA-256)\n *\n * @example\n * ```typescript\n * const hash = hashContent('Hello world');\n * console.log(hash.length); // 64\n * ```\n */\nexport function hashContent(content: string): string {\n return createHash('sha256').update(content, 'utf8').digest('hex');\n}\n","/**\n * Confidence States - Implements multi-state synapses\n *\n * Neuroscience: Synapses exist in three states: silent, ready (potentiated), active.\n * Application: Classify memory entries by score into silent/ready/active states.\n */\n\nimport type { ConfidenceState, MemoryEntry, StateDistribution } from '../types/memory.js';\n\nexport interface ConfidenceStatesConfig {\n /** Score threshold for active state (e.g., 0.7) */\n activeThreshold: number;\n /** Score threshold for ready state (e.g., 0.3) */\n readyThreshold: number;\n}\n\nexport interface ConfidenceStates {\n /**\n * Calculate state based on entry score and BTSP flag\n * @param entry - Memory entry\n * @returns Confidence state\n */\n calculateState(entry: MemoryEntry): ConfidenceState;\n\n /**\n * Transition entry to correct state based on its score\n * @param entry - Entry to transition\n * @returns Entry with updated state\n */\n transition(entry: MemoryEntry): MemoryEntry;\n\n /**\n * Get distribution of states across all entries\n * @param entries - All memory entries\n * @returns State distribution with counts\n */\n getDistribution(entries: MemoryEntry[]): StateDistribution;\n}\n\n/**\n * Create a confidence states manager\n * @param config - States configuration\n * @returns ConfidenceStates instance\n */\nexport function createConfidenceStates(config: ConfidenceStatesConfig): ConfidenceStates {\n const { activeThreshold, readyThreshold } = config;\n\n function calculateState(entry: MemoryEntry): ConfidenceState {\n // BTSP entries are always active\n if (entry.isBTSP) {\n return 'active';\n }\n\n // State based on score thresholds\n // Active: score > 0.7\n if (entry.score > activeThreshold) {\n return 'active';\n }\n\n // Ready: 0.3 <= score <= 0.7\n if (entry.score >= readyThreshold) {\n return 'ready';\n }\n\n // Silent: score < 0.3\n return 'silent';\n }\n\n function transition(entry: MemoryEntry): MemoryEntry {\n const newState = calculateState(entry);\n\n return {\n ...entry,\n state: newState,\n };\n }\n\n function getDistribution(entries: MemoryEntry[]): StateDistribution {\n const distribution: StateDistribution = {\n silent: 0,\n ready: 0,\n active: 0,\n total: entries.length,\n };\n\n for (const entry of entries) {\n const state = calculateState(entry);\n distribution[state]++;\n }\n\n return distribution;\n }\n\n return {\n calculateState,\n transition,\n getDistribution,\n };\n}\n","/**\n * Engram Scorer - Implements engram theory (memory decay)\n *\n * Neuroscience: Memories fade over time without reinforcement.\n * Application: Apply exponential decay formula to memory scores based on age and access count.\n *\n * Formula: decay = 1 - e^(-age/TTL)\n * Score adjustment: score_new = score_old * (1 - decay) + (accessCount bonus)\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\n\nexport interface EngramScorerConfig {\n /** Default TTL in hours for new entries */\n defaultTTL: number;\n /** Decay threshold (0.0-1.0) above which entries are marked for pruning */\n decayThreshold: number;\n}\n\nexport interface EngramScorer {\n /**\n * Calculate current score for an entry based on decay and access count\n * @param entry - Memory entry to score\n * @param currentTime - Current timestamp in milliseconds (for testing)\n * @returns Updated score (0.0-1.0)\n */\n calculateScore(entry: MemoryEntry, currentTime?: number): number;\n\n /**\n * Refresh TTL to default value\n * @param entry - Entry to refresh\n * @returns Entry with refreshed TTL and timestamp\n */\n refreshTTL(entry: MemoryEntry): MemoryEntry;\n\n /**\n * Calculate decay factor (0.0-1.0) based on age and TTL\n * @param ageInSeconds - Age of entry in seconds\n * @param ttlInSeconds - TTL in seconds\n * @returns Decay factor (0.0 = fresh, 1.0 = fully decayed)\n */\n calculateDecay(ageInSeconds: number, ttlInSeconds: number): number;\n}\n\n/**\n * Create an engram scorer instance\n * @param config - Scorer configuration\n * @returns EngramScorer instance\n */\nexport function createEngramScorer(config: EngramScorerConfig): EngramScorer {\n const { defaultTTL } = config;\n\n function calculateDecay(ageInSeconds: number, ttlInSeconds: number): number {\n if (ttlInSeconds === 0) return 1.0; // Instant decay\n if (ageInSeconds <= 0) return 0.0; // Fresh entry\n\n // Exponential decay: 1 - e^(-age/TTL)\n const ratio = ageInSeconds / ttlInSeconds;\n const decay = 1 - Math.exp(-ratio);\n\n // Clamp to [0.0, 1.0]\n return Math.max(0, Math.min(1, decay));\n }\n\n function calculateScore(entry: MemoryEntry, currentTime: number = Date.now()): number {\n // Calculate age in seconds\n const ageInMilliseconds = currentTime - entry.timestamp;\n const ageInSeconds = Math.max(0, ageInMilliseconds / 1000);\n\n // Calculate decay factor\n const decay = calculateDecay(ageInSeconds, entry.ttl);\n\n // Base score reduced by decay\n let score = entry.score * (1 - decay);\n\n // Access count bonus (diminishing returns via log)\n if (entry.accessCount > 0) {\n const accessBonus = Math.log(entry.accessCount + 1) * 0.1;\n score = Math.min(1.0, score + accessBonus);\n }\n\n // BTSP entries maintain high score\n if (entry.isBTSP) {\n score = Math.max(score, 0.9);\n }\n\n return Math.max(0, Math.min(1, score));\n }\n\n function refreshTTL(entry: MemoryEntry): MemoryEntry {\n return {\n ...entry,\n ttl: defaultTTL * 3600, // Convert hours to seconds\n timestamp: Date.now(),\n };\n }\n\n return {\n calculateScore,\n refreshTTL,\n calculateDecay,\n };\n}\n","/**\n * Token estimation utilities.\n * Uses whitespace heuristic (~90% accuracy vs GPT tokenizer).\n */\n\n/**\n * Estimate token count for text using heuristic.\n *\n * Approximation: 1 token ≈ 4 chars or 0.75 words\n * Provides ~90% accuracy compared to GPT tokenizer, sufficient for optimization heuristics.\n *\n * @param text - Text to count\n * @returns Estimated token count\n *\n * @example\n * ```typescript\n * const tokens = estimateTokens('Hello world');\n * console.log(tokens); // ~2\n * ```\n */\nexport function estimateTokens(text: string): number {\n if (!text || text.length === 0) {\n return 0;\n }\n\n // Split on whitespace to get words\n const words = text.split(/\\s+/).filter((w) => w.length > 0);\n const wordCount = words.length;\n\n // Character-based estimate\n const charCount = text.length;\n const charEstimate = Math.ceil(charCount / 4);\n\n // Word-based estimate\n const wordEstimate = Math.ceil(wordCount * 0.75);\n\n // Return the maximum of both estimates (more conservative)\n return Math.max(wordEstimate, charEstimate);\n}\n","/**\n * Sparse Pruner - Implements sparse coding principle\n *\n * Neuroscience: Only 2-5% of neurons fire at any given time.\n * Application: Keep only top 5% most relevant context entries by TF-IDF score.\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\nexport interface SparsePrunerConfig {\n /** Percentage threshold for pruning (e.g., 5 = keep top 5%) */\n threshold: number;\n}\n\nexport interface SparsePruner {\n /**\n * Prune entries to keep only top N% by relevance score\n * @param entries - Memory entries to prune\n * @returns Result with kept/removed entries and token counts\n */\n prune(entries: MemoryEntry[]): PruneResult;\n\n /**\n * Calculate TF-IDF relevance score for a single entry\n * @param entry - Entry to score\n * @param allEntries - All entries for IDF calculation\n * @returns Relevance score (0.0-1.0)\n */\n scoreEntry(entry: MemoryEntry, allEntries: MemoryEntry[]): number;\n}\n\n/**\n * Create a sparse pruner instance\n * @param config - Pruner configuration\n * @returns SparsePruner instance\n */\nexport function createSparsePruner(config: SparsePrunerConfig): SparsePruner {\n const { threshold } = config;\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n function calculateTF(term: string, tokens: string[]): number {\n const count = tokens.filter((t) => t === term).length;\n // Sqrt capping to prevent common words from dominating\n return Math.sqrt(count);\n }\n\n function calculateIDF(term: string, allEntries: MemoryEntry[]): number {\n const totalDocs = allEntries.length;\n const docsWithTerm = allEntries.filter((entry) => {\n const tokens = tokenize(entry.content);\n return tokens.includes(term);\n }).length;\n\n if (docsWithTerm === 0) return 0;\n\n return Math.log(totalDocs / docsWithTerm);\n }\n\n function scoreEntry(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tokens = tokenize(entry.content);\n if (tokens.length === 0) return 0;\n\n const uniqueTerms = [...new Set(tokens)];\n let totalScore = 0;\n\n for (const term of uniqueTerms) {\n const tf = calculateTF(term, tokens);\n const idf = calculateIDF(term, allEntries);\n totalScore += tf * idf;\n }\n\n // Normalize by entry length\n return totalScore / tokens.length;\n }\n\n function prune(entries: MemoryEntry[]): PruneResult {\n if (entries.length === 0) {\n return {\n kept: [],\n removed: [],\n originalTokens: 0,\n prunedTokens: 0,\n };\n }\n\n // Calculate original token count\n const originalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Score all entries\n const scored = entries.map((entry) => ({\n entry,\n score: scoreEntry(entry, entries),\n }));\n\n // Sort by score descending\n scored.sort((a, b) => b.score - a.score);\n\n // Keep top N% (minimum 1 entry)\n const keepCount = Math.max(1, Math.ceil(entries.length * (threshold / 100)));\n const kept = scored.slice(0, keepCount).map((s) => s.entry);\n const removed = scored.slice(keepCount).map((s) => s.entry);\n\n // Calculate pruned token count\n const prunedTokens = kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n kept,\n removed,\n originalTokens,\n prunedTokens,\n };\n }\n\n return {\n prune,\n scoreEntry,\n };\n}\n","/**\n * Context Parser - Shared utilities for parsing agent contexts into memory entries\n *\n * Extracted from claude-code adapter to enable reuse across:\n * - Adapters (claude-code, generic)\n * - Real-time pipeline (streaming context)\n * - Hooks (pre-prompt, post-tool-result)\n */\n\nimport { randomUUID } from 'node:crypto';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from './hash.js';\n\n/**\n * Block type classification for Claude Code context\n */\nexport type BlockType = 'conversation' | 'tool' | 'result' | 'other';\n\n/**\n * Parse Claude Code context into memory entries\n * Handles conversation turns, tool uses, and results\n * @param context - Raw context string\n * @returns Array of memory entries\n */\nexport function parseClaudeCodeContext(context: string): MemoryEntry[] {\n const entries: MemoryEntry[] = [];\n const now = Date.now();\n\n // Split by conversation turns and tool boundaries\n const lines = context.split('\\n');\n let currentBlock: string[] = [];\n let blockType: BlockType = 'other';\n\n for (const line of lines) {\n const trimmed = line.trim();\n\n // Detect conversation turns\n if (trimmed.startsWith('User:') || trimmed.startsWith('Assistant:')) {\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'conversation';\n currentBlock.push(line);\n }\n // Detect tool calls\n else if (\n trimmed.includes('<function_calls>') ||\n trimmed.includes('<invoke>') ||\n trimmed.includes('<tool_use>')\n ) {\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'tool';\n currentBlock.push(line);\n }\n // Detect tool results\n else if (trimmed.includes('<function_results>') || trimmed.includes('</function_results>')) {\n if (currentBlock.length > 0 && blockType !== 'result') {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'result';\n currentBlock.push(line);\n }\n // Continue current block\n else if (currentBlock.length > 0) {\n currentBlock.push(line);\n }\n // Start new block if line has content\n else if (trimmed.length > 0) {\n currentBlock.push(line);\n blockType = 'other';\n }\n }\n\n // Add final block\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n }\n\n return entries.filter((e) => e.content.trim().length > 0);\n}\n\n/**\n * Create a memory entry from a content block\n * @param content - Block content\n * @param type - Block type\n * @param baseTime - Base timestamp\n * @returns Memory entry\n */\nexport function createEntry(content: string, type: BlockType, baseTime: number): MemoryEntry {\n const tags: string[] = [type];\n\n // Assign initial score based on type\n let initialScore = 0.5;\n if (type === 'conversation') initialScore = 0.8; // Prioritize conversation\n if (type === 'tool') initialScore = 0.7; // Tool calls are important\n if (type === 'result') initialScore = 0.4; // Results can be verbose\n\n return {\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: baseTime,\n score: initialScore,\n state: initialScore > 0.7 ? 'active' : initialScore > 0.3 ? 'ready' : 'silent',\n ttl: 24 * 3600, // 24 hours default\n accessCount: 0,\n tags,\n metadata: { type },\n isBTSP: false,\n };\n}\n\n/**\n * Parse generic context (fallback for non-Claude-Code agents)\n * Splits on double newlines, treats as paragraphs\n * @param context - Raw context string\n * @returns Array of memory entries\n */\nexport function parseGenericContext(context: string): MemoryEntry[] {\n const entries: MemoryEntry[] = [];\n const now = Date.now();\n\n // Split on double newlines (paragraph boundaries)\n const blocks = context.split(/\\n\\n+/);\n\n for (const block of blocks) {\n const trimmed = block.trim();\n if (trimmed.length === 0) continue;\n\n entries.push(createEntry(trimmed, 'other', now));\n }\n\n return entries;\n}\n","/**\n * Claude Code Adapter - Claude Code-specific optimization pipeline\n *\n * Optimized for Claude Code's conversation patterns, tool use, and context management.\n * Implements the same AgentAdapter interface as GenericAdapter but with Claude-specific tuning.\n */\n\nimport { createBTSPEmbedder } from '../core/btsp-embedder.js';\nimport { createConfidenceStates } from '../core/confidence-states.js';\nimport { createEngramScorer } from '../core/engram-scorer.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSparsePruner } from '../core/sparse-pruner.js';\nimport type { AgentAdapter, OptimizationResult, OptimizeOptions } from '../types/adapter.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { parseClaudeCodeContext } from '../utils/context-parser.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\n/**\n * Claude Code-specific optimization profile\n * Tuned for Claude's conversation patterns and tool use\n */\nconst CLAUDE_CODE_PROFILE = {\n // More aggressive pruning for tool results (they can be verbose)\n toolResultThreshold: 3, // Keep top 3% of tool results\n\n // Preserve conversation turns more aggressively\n conversationBoost: 1.5, // 50% boost for User/Assistant exchanges\n\n // Prioritize recent context (Claude Code sessions are typically focused)\n recentContextWindow: 10 * 60, // Last 10 minutes gets priority\n\n // BTSP patterns specific to Claude Code\n btspPatterns: [\n // Error patterns\n /\\b(error|exception|failure|fatal|critical|panic)\\b/i,\n /^\\s+at\\s+.*\\(.*:\\d+:\\d+\\)/m, // Stack traces\n /^Error:/m,\n\n // Git conflict markers\n /^<<<<<<< /m,\n /^=======/m,\n /^>>>>>>> /m,\n\n // Tool use patterns (important for context)\n /<function_calls>/,\n /<invoke>/,\n /<tool_use>/,\n\n // File operation results (often critical)\n /ENOENT|EACCES|EISDIR|EEXIST/,\n ],\n};\n\n/**\n * Create a Claude Code adapter instance\n * @param memory - KV memory store\n * @param config - Sparn configuration\n * @returns AgentAdapter instance optimized for Claude Code\n */\nexport function createClaudeCodeAdapter(memory: KVMemory, config: SparnConfig): AgentAdapter {\n // Create core modules with Claude Code-optimized settings\n const pruner = createSparsePruner({\n threshold: config.pruning.threshold,\n });\n\n const scorer = createEngramScorer(config.decay);\n const states = createConfidenceStates(config.states);\n const btsp = createBTSPEmbedder();\n\n async function optimize(\n context: string,\n options: OptimizeOptions = {},\n ): Promise<OptimizationResult> {\n const startTime = Date.now();\n\n // Parse context into entries\n // For Claude Code, we parse by conversation turns and tool uses\n const entries = parseClaudeCodeContext(context);\n\n // Apply BTSP detection with Claude Code-specific patterns\n const entriesWithBTSP = entries.map((entry) => {\n const isBTSP = CLAUDE_CODE_PROFILE.btspPatterns.some((pattern) =>\n pattern.test(entry.content),\n );\n\n if (isBTSP) {\n const btspEntry = btsp.createBTSPEntry(entry.content, [...entry.tags, 'claude-code'], {\n originalTimestamp: entry.timestamp,\n });\n // Preserve original timestamp\n return {\n ...btspEntry,\n timestamp: entry.timestamp,\n };\n }\n\n return entry;\n });\n\n // Apply conversation boost to User/Assistant exchanges\n const boostedEntries = entriesWithBTSP.map((entry) => {\n const isConversationTurn =\n entry.content.trim().startsWith('User:') || entry.content.trim().startsWith('Assistant:');\n\n if (isConversationTurn) {\n return {\n ...entry,\n score: entry.score * CLAUDE_CODE_PROFILE.conversationBoost,\n };\n }\n\n return entry;\n });\n\n // Score entries with decay\n const scoredEntries = boostedEntries.map((entry) => {\n const decayScore = scorer.calculateScore(entry);\n return {\n ...entry,\n score: decayScore,\n };\n });\n\n // Calculate states\n const entriesWithStates = scoredEntries.map((entry) => {\n const state = states.calculateState(entry);\n return {\n ...entry,\n state,\n };\n });\n\n // Prune entries (keep top N%)\n const pruneResult = pruner.prune(entriesWithStates);\n\n // Store kept entries in memory (if not dry-run)\n if (!options.dryRun) {\n for (const entry of pruneResult.kept) {\n await memory.put(entry);\n }\n\n // Record optimization stats\n await memory.recordOptimization({\n timestamp: Date.now(),\n tokens_before: pruneResult.originalTokens,\n tokens_after: pruneResult.prunedTokens,\n entries_pruned: pruneResult.removed.length,\n duration_ms: Date.now() - startTime,\n });\n }\n\n // Build optimized context from kept entries\n const optimizedContext = pruneResult.kept.map((entry) => entry.content).join('\\n');\n\n // Calculate state distribution\n const stateDistribution = states.getDistribution(pruneResult.kept);\n\n // Build result\n const result: OptimizationResult = {\n optimizedContext,\n tokensBefore: pruneResult.originalTokens,\n tokensAfter: pruneResult.prunedTokens,\n reduction:\n pruneResult.originalTokens > 0\n ? (pruneResult.originalTokens - pruneResult.prunedTokens) / pruneResult.originalTokens\n : 0,\n entriesProcessed: entries.length,\n entriesKept: pruneResult.kept.length,\n durationMs: Date.now() - startTime,\n stateDistribution,\n };\n\n // Add verbose details if requested\n if (options.verbose) {\n result.details = pruneResult.kept.map((entry) => ({\n id: entry.id,\n score: entry.score,\n state: entry.state || 'unknown',\n isBTSP: entry.tags.includes('btsp'),\n tokens: estimateTokens(entry.content),\n }));\n }\n\n return result;\n }\n\n return {\n optimize,\n };\n}\n","/**\n * Generic Adapter - Agent-agnostic optimization pipeline\n *\n * Orchestrates all 6 neuroscience modules to optimize context memory.\n */\n\nimport { randomUUID } from 'node:crypto';\nimport { createBTSPEmbedder } from '../core/btsp-embedder.js';\nimport { createConfidenceStates } from '../core/confidence-states.js';\nimport { createEngramScorer } from '../core/engram-scorer.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSparsePruner } from '../core/sparse-pruner.js';\nimport type { AgentAdapter, OptimizationResult, OptimizeOptions } from '../types/adapter.js';\nimport type { SparnConfig } from '../types/config.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from '../utils/hash.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\n/**\n * Create a generic adapter instance\n * @param memory - KV memory store\n * @param config - Sparn configuration\n * @returns AgentAdapter instance\n */\nexport function createGenericAdapter(memory: KVMemory, config: SparnConfig): AgentAdapter {\n const pruner = createSparsePruner(config.pruning);\n const scorer = createEngramScorer(config.decay);\n const states = createConfidenceStates(config.states);\n const btsp = createBTSPEmbedder();\n\n async function optimize(\n context: string,\n options: OptimizeOptions = {},\n ): Promise<OptimizationResult> {\n const startTime = Date.now();\n\n // Parse context into entries (line-based for simplicity)\n const lines = context.split('\\n').filter((line) => line.trim().length > 0);\n const entries: MemoryEntry[] = lines.map((content) => ({\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: Date.now(),\n score: btsp.detectBTSP(content) ? 1.0 : 0.5, // BTSP gets high initial score\n ttl: config.decay.defaultTTL * 3600, // Convert hours to seconds\n state: 'ready' as const,\n accessCount: 0,\n tags: [],\n metadata: {},\n isBTSP: btsp.detectBTSP(content),\n }));\n\n // Calculate original token count\n const tokensBefore = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 1: Update scores with decay\n const scoredEntries = entries.map((entry) => ({\n ...entry,\n score: scorer.calculateScore(entry),\n }));\n\n // Step 2: Transition states based on scores\n const statedEntries = scoredEntries.map((entry) => states.transition(entry));\n\n // Step 3: Apply sparse pruning\n const pruneResult = pruner.prune(statedEntries);\n\n // Step 4: Keep active and ready entries, discard silent\n const optimizedEntries = pruneResult.kept.filter(\n (e) => e.state === 'active' || e.state === 'ready',\n );\n\n // Calculate final token count\n const tokensAfter = optimizedEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Reconstruct optimized context\n const optimizedContext = optimizedEntries.map((e) => e.content).join('\\n');\n\n // Store entries in memory (if not dry run)\n if (!options.dryRun) {\n for (const entry of optimizedEntries) {\n await memory.put(entry);\n }\n\n // Record optimization statistics\n await memory.recordOptimization({\n timestamp: Date.now(),\n tokens_before: tokensBefore,\n tokens_after: tokensAfter,\n entries_pruned: entries.length - optimizedEntries.length,\n duration_ms: Date.now() - startTime,\n });\n }\n\n // Get state distribution\n const distribution = states.getDistribution(optimizedEntries);\n\n const result: OptimizationResult = {\n optimizedContext,\n tokensBefore,\n tokensAfter,\n reduction: tokensBefore > 0 ? (tokensBefore - tokensAfter) / tokensBefore : 0,\n entriesProcessed: entries.length,\n entriesKept: optimizedEntries.length,\n stateDistribution: distribution,\n durationMs: Date.now() - startTime,\n };\n\n // Add verbose details if requested\n if (options.verbose) {\n result.details = optimizedEntries.map((e) => ({\n id: e.id,\n score: e.score,\n state: e.state,\n isBTSP: e.isBTSP,\n tokens: estimateTokens(e.content),\n }));\n }\n\n return result;\n }\n\n return {\n optimize,\n };\n}\n","/**\n * Budget-Aware Pruner - Token budget optimization\n *\n * Unlike SparsePruner which keeps top N% entries, BudgetPruner fits entries\n * within a target token budget using priority scoring that combines:\n * - TF-IDF relevance\n * - Engram decay\n * - Confidence state multipliers\n * - BTSP bypass (always included)\n *\n * Target use case: Real-time optimization for Opus model (~50K token budget)\n */\n\nimport type { RealtimeConfig } from '../types/config.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { createEngramScorer } from './engram-scorer.js';\n\nexport interface BudgetPrunerConfig {\n /** Target token budget */\n tokenBudget: number;\n /** Decay configuration */\n decay: {\n defaultTTL: number;\n decayThreshold: number;\n };\n /** State multipliers */\n states: {\n activeThreshold: number;\n readyThreshold: number;\n };\n}\n\nexport interface BudgetPruner {\n /**\n * Prune entries to fit within token budget\n * @param entries - Memory entries to prune\n * @param budget - Optional override budget (uses config default if not provided)\n * @returns Result with kept/removed entries and budget utilization\n */\n pruneToFit(entries: MemoryEntry[], budget?: number): PruneResult & { budgetUtilization: number };\n\n /**\n * Calculate priority score for an entry\n * @param entry - Entry to score\n * @param allEntries - All entries for TF-IDF calculation\n * @returns Priority score (higher = more important)\n */\n priorityScore(entry: MemoryEntry, allEntries: MemoryEntry[]): number;\n}\n\n/**\n * Create a budget-aware pruner instance\n * @param config - Pruner configuration\n * @returns BudgetPruner instance\n */\nexport function createBudgetPruner(config: BudgetPrunerConfig): BudgetPruner {\n const { tokenBudget, decay } = config;\n const engramScorer = createEngramScorer(decay);\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n function calculateTF(term: string, tokens: string[]): number {\n const count = tokens.filter((t) => t === term).length;\n // Sqrt capping to prevent common words from dominating\n return Math.sqrt(count);\n }\n\n function calculateIDF(term: string, allEntries: MemoryEntry[]): number {\n const totalDocs = allEntries.length;\n const docsWithTerm = allEntries.filter((entry) => {\n const tokens = tokenize(entry.content);\n return tokens.includes(term);\n }).length;\n\n if (docsWithTerm === 0) return 0;\n\n return Math.log(totalDocs / docsWithTerm);\n }\n\n function calculateTFIDF(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tokens = tokenize(entry.content);\n if (tokens.length === 0) return 0;\n\n const uniqueTerms = [...new Set(tokens)];\n let totalScore = 0;\n\n for (const term of uniqueTerms) {\n const tf = calculateTF(term, tokens);\n const idf = calculateIDF(term, allEntries);\n totalScore += tf * idf;\n }\n\n // Normalize by entry length\n return totalScore / tokens.length;\n }\n\n function getStateMultiplier(entry: MemoryEntry): number {\n // BTSP entries get max priority (handled separately, but keep high multiplier)\n if (entry.isBTSP) return 2.0;\n\n // State-based multipliers\n switch (entry.state) {\n case 'active':\n return 2.0;\n case 'ready':\n return 1.0;\n case 'silent':\n return 0.5;\n default:\n return 1.0;\n }\n }\n\n function priorityScore(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tfidf = calculateTFIDF(entry, allEntries);\n const currentScore = engramScorer.calculateScore(entry);\n const engramDecay = 1 - currentScore; // Lower decay = higher priority\n const stateMultiplier = getStateMultiplier(entry);\n\n // Priority = TF-IDF * (1 - decay) * state_multiplier\n // This balances relevance, recency, and confidence state\n return tfidf * (1 - engramDecay) * stateMultiplier;\n }\n\n function pruneToFit(\n entries: MemoryEntry[],\n budget: number = tokenBudget,\n ): PruneResult & { budgetUtilization: number } {\n if (entries.length === 0) {\n return {\n kept: [],\n removed: [],\n originalTokens: 0,\n prunedTokens: 0,\n budgetUtilization: 0,\n };\n }\n\n // Calculate original token count\n const originalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 1: Separate BTSP entries (always included, bypass budget)\n const btspEntries = entries.filter((e) => e.isBTSP);\n const regularEntries = entries.filter((e) => !e.isBTSP);\n\n const btspTokens = btspEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 2: Score regular entries\n const scored = regularEntries.map((entry) => ({\n entry,\n score: priorityScore(entry, entries),\n tokens: estimateTokens(entry.content),\n }));\n\n // Step 3: Sort by priority score descending\n scored.sort((a, b) => b.score - a.score);\n\n // Step 4: Greedy fill until budget exceeded\n const kept: MemoryEntry[] = [...btspEntries];\n const removed: MemoryEntry[] = [];\n let currentTokens = btspTokens;\n\n for (const item of scored) {\n if (currentTokens + item.tokens <= budget) {\n kept.push(item.entry);\n currentTokens += item.tokens;\n } else {\n removed.push(item.entry);\n }\n }\n\n const budgetUtilization = budget > 0 ? currentTokens / budget : 0;\n\n return {\n kept,\n removed,\n originalTokens,\n prunedTokens: currentTokens,\n budgetUtilization,\n };\n }\n\n return {\n pruneToFit,\n priorityScore,\n };\n}\n\n/**\n * Helper to create budget pruner from RealtimeConfig\n * @param realtimeConfig - Realtime configuration\n * @param decayConfig - Decay configuration\n * @param statesConfig - States configuration\n * @returns BudgetPruner instance\n */\nexport function createBudgetPrunerFromConfig(\n realtimeConfig: RealtimeConfig,\n decayConfig: { defaultTTL: number; decayThreshold: number },\n statesConfig: { activeThreshold: number; readyThreshold: number },\n): BudgetPruner {\n return createBudgetPruner({\n tokenBudget: realtimeConfig.tokenBudget,\n decay: decayConfig,\n states: statesConfig,\n });\n}\n","/**\n * Metrics and Telemetry System\n *\n * Tracks performance metrics and optimization statistics:\n * - Optimization duration and throughput\n * - Token savings and reduction rates\n * - Memory usage and cache hit rates\n * - Daemon uptime and session counts\n */\n\nexport interface OptimizationMetric {\n timestamp: number;\n duration: number;\n tokensBefore: number;\n tokensAfter: number;\n entriesProcessed: number;\n entriesKept: number;\n cacheHitRate: number;\n memoryUsage: number;\n}\n\nexport interface DaemonMetric {\n startTime: number;\n sessionsWatched: number;\n totalOptimizations: number;\n totalTokensSaved: number;\n averageLatency: number;\n memoryUsage: number;\n}\n\nexport interface MetricsSnapshot {\n timestamp: number;\n optimization: {\n totalRuns: number;\n totalDuration: number;\n totalTokensSaved: number;\n averageReduction: number;\n p50Latency: number;\n p95Latency: number;\n p99Latency: number;\n };\n cache: {\n hitRate: number;\n totalHits: number;\n totalMisses: number;\n size: number;\n };\n daemon: {\n uptime: number;\n sessionsWatched: number;\n memoryUsage: number;\n };\n}\n\nexport interface MetricsCollector {\n /**\n * Record an optimization metric\n */\n recordOptimization(metric: OptimizationMetric): void;\n\n /**\n * Update daemon metrics\n */\n updateDaemon(metric: Partial<DaemonMetric>): void;\n\n /**\n * Get current metrics snapshot\n */\n getSnapshot(): MetricsSnapshot;\n\n /**\n * Export metrics as JSON\n */\n export(): string;\n\n /**\n * Reset all metrics\n */\n reset(): void;\n}\n\n/**\n * Create a metrics collector instance\n */\nexport function createMetricsCollector(): MetricsCollector {\n const optimizations: OptimizationMetric[] = [];\n let daemonMetrics: DaemonMetric = {\n startTime: Date.now(),\n sessionsWatched: 0,\n totalOptimizations: 0,\n totalTokensSaved: 0,\n averageLatency: 0,\n memoryUsage: 0,\n };\n\n let cacheHits = 0;\n let cacheMisses = 0;\n\n function recordOptimization(metric: OptimizationMetric): void {\n optimizations.push(metric);\n\n // Update daemon totals\n daemonMetrics.totalOptimizations++;\n daemonMetrics.totalTokensSaved += metric.tokensBefore - metric.tokensAfter;\n\n // Update cache stats\n if (metric.cacheHitRate > 0) {\n const hits = Math.round(metric.entriesProcessed * metric.cacheHitRate);\n cacheHits += hits;\n cacheMisses += metric.entriesProcessed - hits;\n }\n\n // Update average latency (moving average)\n daemonMetrics.averageLatency =\n (daemonMetrics.averageLatency * (daemonMetrics.totalOptimizations - 1) + metric.duration) /\n daemonMetrics.totalOptimizations;\n\n // Keep only last 1000 metrics in memory\n if (optimizations.length > 1000) {\n optimizations.shift();\n }\n }\n\n function updateDaemon(metric: Partial<DaemonMetric>): void {\n daemonMetrics = {\n ...daemonMetrics,\n ...metric,\n };\n }\n\n function calculatePercentile(values: number[], percentile: number): number {\n if (values.length === 0) return 0;\n\n const sorted = [...values].sort((a, b) => a - b);\n const index = Math.ceil((percentile / 100) * sorted.length) - 1;\n return sorted[index] || 0;\n }\n\n function getSnapshot(): MetricsSnapshot {\n const totalRuns = optimizations.length;\n const totalDuration = optimizations.reduce((sum, m) => sum + m.duration, 0);\n const totalTokensSaved = optimizations.reduce(\n (sum, m) => sum + (m.tokensBefore - m.tokensAfter),\n 0,\n );\n\n const totalTokensBefore = optimizations.reduce((sum, m) => sum + m.tokensBefore, 0);\n const averageReduction = totalTokensBefore > 0 ? totalTokensSaved / totalTokensBefore : 0;\n\n const durations = optimizations.map((m) => m.duration);\n\n const totalCacheQueries = cacheHits + cacheMisses;\n const hitRate = totalCacheQueries > 0 ? cacheHits / totalCacheQueries : 0;\n\n return {\n timestamp: Date.now(),\n optimization: {\n totalRuns,\n totalDuration,\n totalTokensSaved,\n averageReduction,\n p50Latency: calculatePercentile(durations, 50),\n p95Latency: calculatePercentile(durations, 95),\n p99Latency: calculatePercentile(durations, 99),\n },\n cache: {\n hitRate,\n totalHits: cacheHits,\n totalMisses: cacheMisses,\n size: optimizations.reduce((sum, m) => sum + m.entriesKept, 0),\n },\n daemon: {\n uptime: Date.now() - daemonMetrics.startTime,\n sessionsWatched: daemonMetrics.sessionsWatched,\n memoryUsage: daemonMetrics.memoryUsage,\n },\n };\n }\n\n function exportMetrics(): string {\n return JSON.stringify(getSnapshot(), null, 2);\n }\n\n function reset(): void {\n optimizations.length = 0;\n cacheHits = 0;\n cacheMisses = 0;\n daemonMetrics = {\n startTime: Date.now(),\n sessionsWatched: 0,\n totalOptimizations: 0,\n totalTokensSaved: 0,\n averageLatency: 0,\n memoryUsage: 0,\n };\n }\n\n return {\n recordOptimization,\n updateDaemon,\n getSnapshot,\n export: exportMetrics,\n reset,\n };\n}\n\n// Global metrics instance\nlet globalMetrics: MetricsCollector | null = null;\n\n/**\n * Get or create the global metrics collector\n */\nexport function getMetrics(): MetricsCollector {\n if (!globalMetrics) {\n globalMetrics = createMetricsCollector();\n }\n return globalMetrics;\n}\n","/**\n * Incremental Optimizer - Cache-based delta processing\n *\n * Optimizes performance for real-time scenarios by:\n * - Caching entry scores by content hash\n * - Only recomputing scores for new/changed entries\n * - Pre-computing and caching document frequency tables\n * - Periodically forcing full re-optimization to prevent drift\n *\n * Target: <50ms for incremental updates on 100K token contexts\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { type BudgetPrunerConfig, createBudgetPruner } from './budget-pruner.js';\nimport { getMetrics } from './metrics.js';\n\nexport interface IncrementalOptimizerConfig extends BudgetPrunerConfig {\n /** Force full re-optimization every N incremental updates */\n fullOptimizationInterval: number;\n}\n\nexport interface IncrementalOptimizerState {\n /** Entry cache keyed by content hash */\n entryCache: Map<string, { entry: MemoryEntry; score: number; timestamp: number }>;\n /** Document frequency table for IDF calculation */\n documentFrequency: Map<string, number>;\n /** Total document count for IDF */\n totalDocuments: number;\n /** Incremental update counter */\n updateCount: number;\n /** Last full optimization timestamp */\n lastFullOptimization: number;\n}\n\nexport interface IncrementalOptimizer {\n /**\n * Optimize incrementally (only process new/changed entries)\n * @param newEntries - New entries to add\n * @param budget - Optional budget override\n * @returns Prune result with budget utilization\n */\n optimizeIncremental(\n newEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number };\n\n /**\n * Optimize fully (recompute all scores)\n * @param allEntries - All entries to optimize\n * @param budget - Optional budget override\n * @returns Prune result with budget utilization\n */\n optimizeFull(\n allEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number };\n\n /**\n * Get current optimizer state (for serialization)\n * @returns Serializable state object\n */\n getState(): IncrementalOptimizerState;\n\n /**\n * Restore optimizer state (from serialization)\n * @param state - State to restore\n */\n restoreState(state: IncrementalOptimizerState): void;\n\n /**\n * Reset optimizer state (clear all caches)\n */\n reset(): void;\n\n /**\n * Get cache statistics\n * @returns Cache stats\n */\n getStats(): {\n cachedEntries: number;\n uniqueTerms: number;\n totalDocuments: number;\n updateCount: number;\n lastFullOptimization: number;\n };\n}\n\n/**\n * Create an incremental optimizer instance\n * @param config - Optimizer configuration\n * @returns IncrementalOptimizer instance\n */\nexport function createIncrementalOptimizer(\n config: IncrementalOptimizerConfig,\n): IncrementalOptimizer {\n const pruner = createBudgetPruner(config);\n const { fullOptimizationInterval } = config;\n\n // Internal state\n let state: IncrementalOptimizerState = {\n entryCache: new Map(),\n documentFrequency: new Map(),\n totalDocuments: 0,\n updateCount: 0,\n lastFullOptimization: Date.now(),\n };\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n /**\n * Update document frequency table incrementally\n */\n function updateDocumentFrequency(entries: MemoryEntry[], remove = false): void {\n for (const entry of entries) {\n const tokens = tokenize(entry.content);\n const uniqueTerms = [...new Set(tokens)];\n\n for (const term of uniqueTerms) {\n const current = state.documentFrequency.get(term) || 0;\n const updated = remove ? Math.max(0, current - 1) : current + 1;\n\n if (updated === 0) {\n state.documentFrequency.delete(term);\n } else {\n state.documentFrequency.set(term, updated);\n }\n }\n }\n\n state.totalDocuments += remove ? -entries.length : entries.length;\n state.totalDocuments = Math.max(0, state.totalDocuments);\n }\n\n /**\n * Check if entry is cached and still valid\n */\n function getCachedEntry(hash: string): MemoryEntry | null {\n const cached = state.entryCache.get(hash);\n if (!cached) return null;\n\n // Entry is valid if found in cache\n return cached.entry;\n }\n\n /**\n * Cache entry with score\n */\n function cacheEntry(entry: MemoryEntry, score: number): void {\n state.entryCache.set(entry.hash, {\n entry,\n score,\n timestamp: Date.now(),\n });\n }\n\n function optimizeIncremental(\n newEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number } {\n const startTime = Date.now();\n state.updateCount++;\n\n // Force full optimization if interval reached\n if (state.updateCount >= fullOptimizationInterval) {\n // Get all cached entries\n const allEntries = Array.from(state.entryCache.values()).map((c) => c.entry);\n return optimizeFull([...allEntries, ...newEntries], budget);\n }\n\n // Filter out already-cached entries\n const uncachedEntries: MemoryEntry[] = [];\n const cachedEntries: MemoryEntry[] = [];\n\n for (const entry of newEntries) {\n const cached = getCachedEntry(entry.hash);\n if (cached) {\n cachedEntries.push(cached);\n } else {\n uncachedEntries.push(entry);\n }\n }\n\n // Update document frequency for new entries only\n if (uncachedEntries.length > 0) {\n updateDocumentFrequency(uncachedEntries, false);\n }\n\n // Combine with cached entries for scoring context\n const allEntries = [...cachedEntries, ...uncachedEntries];\n\n // Score only uncached entries (reuse cached scores)\n for (const entry of uncachedEntries) {\n const score = pruner.priorityScore(entry, allEntries);\n cacheEntry(entry, score);\n }\n\n // Get all current entries (from cache + new)\n const currentEntries = Array.from(state.entryCache.values()).map((c) => c.entry);\n\n // Calculate tokens before\n const tokensBefore = currentEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Prune to fit budget\n const result = pruner.pruneToFit(currentEntries, budget);\n\n // Calculate tokens after\n const tokensAfter = result.kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Update cache: remove pruned entries\n for (const removed of result.removed) {\n state.entryCache.delete(removed.hash);\n }\n\n // Update document frequency to reflect removal\n if (result.removed.length > 0) {\n updateDocumentFrequency(result.removed, true);\n }\n\n // Record metrics\n const duration = Date.now() - startTime;\n const cacheHitRate = newEntries.length > 0 ? cachedEntries.length / newEntries.length : 0;\n\n getMetrics().recordOptimization({\n timestamp: Date.now(),\n duration,\n tokensBefore,\n tokensAfter,\n entriesProcessed: newEntries.length,\n entriesKept: result.kept.length,\n cacheHitRate,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n return result;\n }\n\n function optimizeFull(\n allEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number } {\n const startTime = Date.now();\n\n // Calculate tokens before\n const tokensBefore = allEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Reset state\n state.entryCache.clear();\n state.documentFrequency.clear();\n state.totalDocuments = 0;\n state.updateCount = 0;\n state.lastFullOptimization = Date.now();\n\n // Rebuild document frequency table\n updateDocumentFrequency(allEntries, false);\n\n // Score and cache all entries\n for (const entry of allEntries) {\n const score = pruner.priorityScore(entry, allEntries);\n cacheEntry(entry, score);\n }\n\n // Prune to fit budget\n const result = pruner.pruneToFit(allEntries, budget);\n\n // Calculate tokens after\n const tokensAfter = result.kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Update cache: remove pruned entries\n for (const removed of result.removed) {\n state.entryCache.delete(removed.hash);\n }\n\n // Update document frequency to reflect removal\n if (result.removed.length > 0) {\n updateDocumentFrequency(result.removed, true);\n }\n\n // Record metrics\n const duration = Date.now() - startTime;\n\n getMetrics().recordOptimization({\n timestamp: Date.now(),\n duration,\n tokensBefore,\n tokensAfter,\n entriesProcessed: allEntries.length,\n entriesKept: result.kept.length,\n cacheHitRate: 0, // Full optimization has no cache hits\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n return result;\n }\n\n function getState(): IncrementalOptimizerState {\n return {\n entryCache: new Map(state.entryCache),\n documentFrequency: new Map(state.documentFrequency),\n totalDocuments: state.totalDocuments,\n updateCount: state.updateCount,\n lastFullOptimization: state.lastFullOptimization,\n };\n }\n\n function restoreState(restoredState: IncrementalOptimizerState): void {\n state = {\n entryCache: new Map(restoredState.entryCache),\n documentFrequency: new Map(restoredState.documentFrequency),\n totalDocuments: restoredState.totalDocuments,\n updateCount: restoredState.updateCount,\n lastFullOptimization: restoredState.lastFullOptimization,\n };\n }\n\n function reset(): void {\n state = {\n entryCache: new Map(),\n documentFrequency: new Map(),\n totalDocuments: 0,\n updateCount: 0,\n lastFullOptimization: Date.now(),\n };\n }\n\n function getStats() {\n return {\n cachedEntries: state.entryCache.size,\n uniqueTerms: state.documentFrequency.size,\n totalDocuments: state.totalDocuments,\n updateCount: state.updateCount,\n lastFullOptimization: state.lastFullOptimization,\n };\n }\n\n return {\n optimizeIncremental,\n optimizeFull,\n getState,\n restoreState,\n reset,\n getStats,\n };\n}\n","/**\n * Streaming Context Pipeline - Real-time sliding window buffer\n *\n * Maintains an optimized context in real-time by:\n * - Ingesting new content as it arrives\n * - Storing entries by priority internally (for eviction decisions)\n * - Outputting in chronological order (for conversation coherence)\n * - Evicting lowest-priority entries when budget exceeded\n * - Using IncrementalOptimizer for fast delta processing\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport { parseClaudeCodeContext } from '../utils/context-parser.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport {\n createIncrementalOptimizer,\n type IncrementalOptimizerConfig,\n} from './incremental-optimizer.js';\n\nexport interface ContextPipelineConfig extends IncrementalOptimizerConfig {\n /** Sliding window size (max entries to keep) */\n windowSize: number;\n}\n\nexport interface ContextPipelineStats {\n /** Total entries ingested */\n totalIngested: number;\n /** Current entry count */\n currentEntries: number;\n /** Current token count */\n currentTokens: number;\n /** Budget utilization (0.0-1.0) */\n budgetUtilization: number;\n /** Evicted entry count */\n evictedEntries: number;\n /** Optimizer stats */\n optimizer: {\n cachedEntries: number;\n uniqueTerms: number;\n updateCount: number;\n };\n}\n\nexport interface ContextPipeline {\n /**\n * Ingest new content into the pipeline\n * @param content - Raw content string\n * @param metadata - Optional metadata to attach to entries\n * @returns Number of entries ingested\n */\n ingest(content: string, metadata?: Record<string, unknown>): number;\n\n /**\n * Get current optimized context (chronologically ordered)\n * @returns Optimized context string\n */\n getContext(): string;\n\n /**\n * Get current entries (chronologically ordered)\n * @returns Array of memory entries\n */\n getEntries(): MemoryEntry[];\n\n /**\n * Get pipeline statistics\n * @returns Pipeline stats\n */\n getStats(): ContextPipelineStats;\n\n /**\n * Clear all entries and reset state\n */\n clear(): void;\n}\n\n/**\n * Create a context pipeline instance\n * @param config - Pipeline configuration\n * @returns ContextPipeline instance\n */\nexport function createContextPipeline(config: ContextPipelineConfig): ContextPipeline {\n const optimizer = createIncrementalOptimizer(config);\n const { windowSize, tokenBudget } = config;\n\n // Internal state\n let totalIngested = 0;\n let evictedEntries = 0;\n let currentEntries: MemoryEntry[] = [];\n let budgetUtilization = 0;\n\n function ingest(content: string, metadata: Record<string, unknown> = {}): number {\n // Parse content into entries\n const newEntries = parseClaudeCodeContext(content);\n\n if (newEntries.length === 0) return 0;\n\n // Attach metadata to entries\n const entriesWithMetadata = newEntries.map((entry) => ({\n ...entry,\n metadata: { ...entry.metadata, ...metadata },\n }));\n\n // Optimize incrementally\n const result = optimizer.optimizeIncremental(entriesWithMetadata, tokenBudget);\n\n // Update statistics\n totalIngested += newEntries.length;\n evictedEntries += result.removed.length;\n currentEntries = result.kept;\n budgetUtilization = result.budgetUtilization;\n\n // Enforce window size limit (keep most recent if exceeded)\n if (currentEntries.length > windowSize) {\n // Sort by timestamp descending (newest first)\n const sorted = [...currentEntries].sort((a, b) => b.timestamp - a.timestamp);\n const toKeep = sorted.slice(0, windowSize);\n const toRemove = sorted.slice(windowSize);\n\n currentEntries = toKeep;\n evictedEntries += toRemove.length;\n }\n\n return newEntries.length;\n }\n\n function getContext(): string {\n // Sort entries chronologically (oldest first)\n const sorted = [...currentEntries].sort((a, b) => a.timestamp - b.timestamp);\n return sorted.map((e) => e.content).join('\\n\\n');\n }\n\n function getEntries(): MemoryEntry[] {\n // Return entries chronologically (oldest first)\n return [...currentEntries].sort((a, b) => a.timestamp - b.timestamp);\n }\n\n function getStats(): ContextPipelineStats {\n const optimizerStats = optimizer.getStats();\n const currentTokens = currentEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n totalIngested,\n currentEntries: currentEntries.length,\n currentTokens,\n budgetUtilization,\n evictedEntries,\n optimizer: {\n cachedEntries: optimizerStats.cachedEntries,\n uniqueTerms: optimizerStats.uniqueTerms,\n updateCount: optimizerStats.updateCount,\n },\n };\n }\n\n function clear(): void {\n totalIngested = 0;\n evictedEntries = 0;\n currentEntries = [];\n budgetUtilization = 0;\n optimizer.reset();\n }\n\n return {\n ingest,\n getContext,\n getEntries,\n getStats,\n clear,\n };\n}\n","/**\n * KV Memory Store Module\n * Implements hippocampal key-value storage with dual index/value tables.\n * Maps to: Hippocampal Key-Value — the hippocampus separates what to store from how to retrieve it.\n */\n\nimport { copyFileSync, existsSync } from 'node:fs';\nimport Database from 'better-sqlite3';\nimport type { MemoryEntry, MemoryQueryFilters } from '../types/memory.js';\n\n/**\n * Optimization statistics record.\n */\nexport interface OptimizationStats {\n id: number;\n timestamp: number;\n tokens_before: number;\n tokens_after: number;\n entries_pruned: number;\n duration_ms: number;\n}\n\n/**\n * KV Memory interface.\n */\nexport interface KVMemory {\n /** Store a memory entry */\n put(entry: MemoryEntry): Promise<void>;\n\n /** Retrieve a memory entry by ID */\n get(id: string): Promise<MemoryEntry | null>;\n\n /** Query entries by filters */\n query(filters: MemoryQueryFilters): Promise<MemoryEntry[]>;\n\n /** Delete a memory entry */\n delete(id: string): Promise<void>;\n\n /** List all entry IDs */\n list(): Promise<string[]>;\n\n /** Compact database (remove expired entries) */\n compact(): Promise<number>;\n\n /** Close database connection */\n close(): Promise<void>;\n\n /** Record optimization statistics */\n recordOptimization(stats: Omit<OptimizationStats, 'id'>): Promise<void>;\n\n /** Get all optimization statistics */\n getOptimizationStats(): Promise<OptimizationStats[]>;\n\n /** Clear all optimization statistics */\n clearOptimizationStats(): Promise<void>;\n}\n\n/**\n * Create a timestamped backup of the database\n * @param dbPath - Path to database file\n * @returns Path to backup file\n */\nfunction createBackup(dbPath: string): string {\n const timestamp = new Date().toISOString().replace(/[:.]/g, '-');\n const backupPath = `${dbPath}.backup-${timestamp}`;\n\n try {\n copyFileSync(dbPath, backupPath);\n console.log(`✓ Database backed up to: ${backupPath}`);\n return backupPath;\n } catch (error) {\n console.error(`Warning: Could not create backup: ${error}`);\n return '';\n }\n}\n\n/**\n * Create KV Memory store with SQLite backend.\n *\n * Initializes database with dual table schema:\n * - entries_index: Fast lookups (id, hash, timestamp, score, ttl, state, accessCount, isBTSP)\n * - entries_value: Content storage (id, content, tags, metadata)\n *\n * @param dbPath - Path to SQLite database file\n * @returns KVMemory instance\n */\nexport async function createKVMemory(dbPath: string): Promise<KVMemory> {\n // Detect database corruption and create backup\n let db: Database.Database;\n try {\n db = new Database(dbPath);\n\n // Quick integrity check\n const integrityCheck = db.pragma('quick_check', { simple: true });\n if (integrityCheck !== 'ok') {\n console.error('⚠ Database corruption detected!');\n\n // Create backup before attempting recovery\n if (existsSync(dbPath)) {\n const backupPath = createBackup(dbPath);\n if (backupPath) {\n console.log(`Backup created at: ${backupPath}`);\n }\n }\n\n // Try to recover\n console.log('Attempting database recovery...');\n db.close();\n db = new Database(dbPath);\n }\n } catch (error) {\n console.error('⚠ Database error detected:', error);\n\n // Create backup if database exists\n if (existsSync(dbPath)) {\n createBackup(dbPath);\n console.log('Creating new database...');\n }\n\n db = new Database(dbPath);\n }\n\n // Enable WAL mode for better concurrency\n db.pragma('journal_mode = WAL');\n\n // Create entries_index table\n db.exec(`\n CREATE TABLE IF NOT EXISTS entries_index (\n id TEXT PRIMARY KEY NOT NULL,\n hash TEXT UNIQUE NOT NULL,\n timestamp INTEGER NOT NULL,\n score REAL NOT NULL DEFAULT 0.0 CHECK(score >= 0.0 AND score <= 1.0),\n ttl INTEGER NOT NULL CHECK(ttl >= 0),\n state TEXT NOT NULL CHECK(state IN ('silent', 'ready', 'active')),\n accessCount INTEGER NOT NULL DEFAULT 0 CHECK(accessCount >= 0),\n isBTSP INTEGER NOT NULL DEFAULT 0 CHECK(isBTSP IN (0, 1)),\n created_at INTEGER NOT NULL DEFAULT (strftime('%s', 'now'))\n );\n `);\n\n // Create entries_value table\n db.exec(`\n CREATE TABLE IF NOT EXISTS entries_value (\n id TEXT PRIMARY KEY NOT NULL,\n content TEXT NOT NULL,\n tags TEXT,\n metadata TEXT,\n FOREIGN KEY (id) REFERENCES entries_index(id) ON DELETE CASCADE\n );\n `);\n\n // Create optimization_stats table\n db.exec(`\n CREATE TABLE IF NOT EXISTS optimization_stats (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp INTEGER NOT NULL DEFAULT (strftime('%s', 'now')),\n tokens_before INTEGER NOT NULL,\n tokens_after INTEGER NOT NULL,\n entries_pruned INTEGER NOT NULL,\n duration_ms INTEGER NOT NULL\n );\n `);\n\n // Create indexes\n db.exec(`\n CREATE INDEX IF NOT EXISTS idx_entries_state ON entries_index(state);\n CREATE INDEX IF NOT EXISTS idx_entries_score ON entries_index(score DESC);\n CREATE INDEX IF NOT EXISTS idx_entries_hash ON entries_index(hash);\n CREATE INDEX IF NOT EXISTS idx_entries_timestamp ON entries_index(timestamp DESC);\n CREATE INDEX IF NOT EXISTS idx_stats_timestamp ON optimization_stats(timestamp DESC);\n `);\n\n // Prepare statements for better performance\n const putIndexStmt = db.prepare(`\n INSERT OR REPLACE INTO entries_index\n (id, hash, timestamp, score, ttl, state, accessCount, isBTSP)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n const putValueStmt = db.prepare(`\n INSERT OR REPLACE INTO entries_value\n (id, content, tags, metadata)\n VALUES (?, ?, ?, ?)\n `);\n\n const getStmt = db.prepare(`\n SELECT\n i.id, i.hash, i.timestamp, i.score, i.ttl, i.state, i.accessCount, i.isBTSP,\n v.content, v.tags, v.metadata\n FROM entries_index i\n JOIN entries_value v ON i.id = v.id\n WHERE i.id = ?\n `);\n\n const deleteIndexStmt = db.prepare('DELETE FROM entries_index WHERE id = ?');\n const deleteValueStmt = db.prepare('DELETE FROM entries_value WHERE id = ?');\n\n return {\n async put(entry: MemoryEntry): Promise<void> {\n const transaction = db.transaction(() => {\n putIndexStmt.run(\n entry.id,\n entry.hash,\n entry.timestamp,\n entry.score,\n entry.ttl,\n entry.state,\n entry.accessCount,\n entry.isBTSP ? 1 : 0,\n );\n\n putValueStmt.run(\n entry.id,\n entry.content,\n JSON.stringify(entry.tags),\n JSON.stringify(entry.metadata),\n );\n });\n\n transaction();\n },\n\n async get(id: string): Promise<MemoryEntry | null> {\n const row = getStmt.get(id) as unknown;\n\n if (!row) {\n return null;\n }\n\n const r = row as {\n id: string;\n hash: string;\n timestamp: number;\n score: number;\n ttl: number;\n state: string;\n accessCount: number;\n isBTSP: number;\n content: string;\n tags: string | null;\n metadata: string | null;\n };\n\n return {\n id: r.id,\n content: r.content,\n hash: r.hash,\n timestamp: r.timestamp,\n score: r.score,\n ttl: r.ttl,\n state: r.state as 'silent' | 'ready' | 'active',\n accessCount: r.accessCount,\n tags: r.tags ? JSON.parse(r.tags) : [],\n metadata: r.metadata ? JSON.parse(r.metadata) : {},\n isBTSP: r.isBTSP === 1,\n };\n },\n\n async query(filters: MemoryQueryFilters): Promise<MemoryEntry[]> {\n let sql = `\n SELECT\n i.id, i.hash, i.timestamp, i.score, i.ttl, i.state, i.accessCount, i.isBTSP,\n v.content, v.tags, v.metadata\n FROM entries_index i\n JOIN entries_value v ON i.id = v.id\n WHERE 1=1\n `;\n\n const params: unknown[] = [];\n\n if (filters.state) {\n sql += ' AND i.state = ?';\n params.push(filters.state);\n }\n\n if (filters.minScore !== undefined) {\n sql += ' AND i.score >= ?';\n params.push(filters.minScore);\n }\n\n if (filters.maxScore !== undefined) {\n sql += ' AND i.score <= ?';\n params.push(filters.maxScore);\n }\n\n if (filters.isBTSP !== undefined) {\n sql += ' AND i.isBTSP = ?';\n params.push(filters.isBTSP ? 1 : 0);\n }\n\n sql += ' ORDER BY i.score DESC';\n\n if (filters.limit) {\n sql += ' LIMIT ?';\n params.push(filters.limit);\n }\n\n if (filters.offset) {\n sql += ' OFFSET ?';\n params.push(filters.offset);\n }\n\n const stmt = db.prepare(sql);\n const rows = stmt.all(...params) as unknown[];\n\n return rows.map((row) => {\n const r = row as {\n id: string;\n hash: string;\n timestamp: number;\n score: number;\n ttl: number;\n state: string;\n accessCount: number;\n isBTSP: number;\n content: string;\n tags: string | null;\n metadata: string | null;\n };\n\n return {\n id: r.id,\n content: r.content,\n hash: r.hash,\n timestamp: r.timestamp,\n score: r.score,\n ttl: r.ttl,\n state: r.state as 'silent' | 'ready' | 'active',\n accessCount: r.accessCount,\n tags: r.tags ? JSON.parse(r.tags) : [],\n metadata: r.metadata ? JSON.parse(r.metadata) : {},\n isBTSP: r.isBTSP === 1,\n };\n });\n },\n\n async delete(id: string): Promise<void> {\n const transaction = db.transaction(() => {\n deleteIndexStmt.run(id);\n deleteValueStmt.run(id);\n });\n\n transaction();\n },\n\n async list(): Promise<string[]> {\n const stmt = db.prepare('SELECT id FROM entries_index');\n const rows = stmt.all() as { id: string }[];\n return rows.map((r) => r.id);\n },\n\n async compact(): Promise<number> {\n const before = db.prepare('SELECT COUNT(*) as count FROM entries_index').get() as {\n count: number;\n };\n\n // Remove fully decayed entries (this will be enhanced in sleep-compressor)\n db.exec('DELETE FROM entries_index WHERE ttl <= 0');\n\n db.exec('VACUUM');\n\n const after = db.prepare('SELECT COUNT(*) as count FROM entries_index').get() as {\n count: number;\n };\n\n return before.count - after.count;\n },\n\n async close(): Promise<void> {\n db.close();\n },\n\n async recordOptimization(stats: Omit<OptimizationStats, 'id'>): Promise<void> {\n const stmt = db.prepare(`\n INSERT INTO optimization_stats (timestamp, tokens_before, tokens_after, entries_pruned, duration_ms)\n VALUES (?, ?, ?, ?, ?)\n `);\n\n stmt.run(\n stats.timestamp,\n stats.tokens_before,\n stats.tokens_after,\n stats.entries_pruned,\n stats.duration_ms,\n );\n },\n\n async getOptimizationStats(): Promise<OptimizationStats[]> {\n const stmt = db.prepare(`\n SELECT id, timestamp, tokens_before, tokens_after, entries_pruned, duration_ms\n FROM optimization_stats\n ORDER BY timestamp DESC\n `);\n\n const rows = stmt.all() as OptimizationStats[];\n return rows;\n },\n\n async clearOptimizationStats(): Promise<void> {\n db.exec('DELETE FROM optimization_stats');\n },\n };\n}\n","/**\n * Sleep Compressor - Implements sleep replay principle\n *\n * Neuroscience: During sleep, the brain consolidates memories by replaying important ones\n * and discarding irrelevant information.\n * Application: Periodic consolidation removes decayed entries and merges duplicates.\n */\n\nimport type { ConsolidateResult, DuplicateGroup } from '../types/consolidate.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { createEngramScorer } from './engram-scorer.js';\n\nexport interface SleepCompressor {\n /**\n * Consolidate entries: remove decayed, merge duplicates\n * @param entries - All memory entries\n * @returns Consolidation result\n */\n consolidate(entries: MemoryEntry[]): ConsolidateResult;\n\n /**\n * Find duplicate entries (exact hash or near-duplicate by similarity)\n * @param entries - Memory entries\n * @returns Groups of duplicates\n */\n findDuplicates(entries: MemoryEntry[]): DuplicateGroup[];\n\n /**\n * Merge duplicate entries, keeping highest score\n * @param groups - Duplicate groups\n * @returns Merged entries\n */\n mergeDuplicates(groups: DuplicateGroup[]): MemoryEntry[];\n}\n\n/**\n * Create a sleep compressor instance\n * @returns SleepCompressor instance\n */\nexport function createSleepCompressor(): SleepCompressor {\n const scorer = createEngramScorer({ defaultTTL: 24, decayThreshold: 0.95 });\n\n function consolidate(entries: MemoryEntry[]): ConsolidateResult {\n const startTime = Date.now();\n const originalCount = entries.length;\n\n // Step 1: Remove fully decayed entries (decay ≥ 0.95)\n const now = Date.now();\n const nonDecayed = entries.filter((entry) => {\n const ageInSeconds = (now - entry.timestamp) / 1000;\n const decay = scorer.calculateDecay(ageInSeconds, entry.ttl);\n return decay < 0.95; // Keep entries with decay < 0.95\n });\n\n const decayedRemoved = originalCount - nonDecayed.length;\n\n // Step 2: Find and merge duplicates\n const duplicateGroups = findDuplicates(nonDecayed);\n const merged = mergeDuplicates(duplicateGroups);\n\n // Step 3: Keep non-duplicates\n const duplicateIds = new Set(duplicateGroups.flatMap((g) => g.entries.map((e) => e.id)));\n const nonDuplicates = nonDecayed.filter((e) => !duplicateIds.has(e.id));\n\n // Combine merged duplicates with non-duplicates\n const kept = [...merged, ...nonDuplicates];\n const removed = entries.filter((e) => !kept.some((k) => k.id === e.id));\n\n const duplicatesRemoved = duplicateGroups.reduce((sum, g) => sum + (g.entries.length - 1), 0);\n\n return {\n kept,\n removed,\n entriesBefore: originalCount,\n entriesAfter: kept.length,\n decayedRemoved,\n duplicatesRemoved,\n compressionRatio: originalCount > 0 ? kept.length / originalCount : 0,\n durationMs: Date.now() - startTime,\n };\n }\n\n function findDuplicates(entries: MemoryEntry[]): DuplicateGroup[] {\n const groups: DuplicateGroup[] = [];\n const processed = new Set<string>();\n\n // Find exact hash matches\n for (let i = 0; i < entries.length; i++) {\n const entry = entries[i];\n if (!entry || processed.has(entry.id)) continue;\n\n const duplicates = entries.filter((e, idx) => idx !== i && e.hash === entry.hash);\n\n if (duplicates.length > 0) {\n const group: DuplicateGroup = {\n entries: [entry, ...duplicates],\n similarity: 1.0, // Exact match\n };\n groups.push(group);\n\n // Mark as processed\n processed.add(entry.id);\n for (const dup of duplicates) {\n processed.add(dup.id);\n }\n }\n }\n\n // Find near-duplicates (cosine similarity ≥ 0.85)\n for (let i = 0; i < entries.length; i++) {\n const entryI = entries[i];\n if (!entryI || processed.has(entryI.id)) continue;\n\n for (let j = i + 1; j < entries.length; j++) {\n const entryJ = entries[j];\n if (!entryJ || processed.has(entryJ.id)) continue;\n\n const similarity = cosineSimilarity(entryI.content, entryJ.content);\n\n if (similarity >= 0.85) {\n const group: DuplicateGroup = {\n entries: [entryI, entryJ],\n similarity,\n };\n groups.push(group);\n\n processed.add(entryI.id);\n processed.add(entryJ.id);\n break; // Move to next i\n }\n }\n }\n\n return groups;\n }\n\n function mergeDuplicates(groups: DuplicateGroup[]): MemoryEntry[] {\n const merged: MemoryEntry[] = [];\n\n for (const group of groups) {\n // Keep entry with highest score\n const sorted = [...group.entries].sort((a, b) => b.score - a.score);\n const best = sorted[0];\n if (!best) continue; // Skip empty groups\n\n // Sum access counts\n const totalAccessCount = group.entries.reduce((sum, e) => sum + e.accessCount, 0);\n\n // Merge tags\n const allTags = new Set(group.entries.flatMap((e) => e.tags));\n\n merged.push({\n ...best,\n accessCount: totalAccessCount,\n tags: Array.from(allTags),\n });\n }\n\n return merged;\n }\n\n /**\n * Calculate cosine similarity between two text strings\n * @param text1 - First text\n * @param text2 - Second text\n * @returns Similarity score (0.0-1.0)\n */\n function cosineSimilarity(text1: string, text2: string): number {\n const words1 = tokenize(text1);\n const words2 = tokenize(text2);\n\n // Build vocabulary\n const vocab = new Set([...words1, ...words2]);\n\n // Build word frequency vectors\n const vec1: Record<string, number> = {};\n const vec2: Record<string, number> = {};\n\n for (const word of vocab) {\n vec1[word] = words1.filter((w) => w === word).length;\n vec2[word] = words2.filter((w) => w === word).length;\n }\n\n // Calculate dot product and magnitudes\n let dotProduct = 0;\n let mag1 = 0;\n let mag2 = 0;\n\n for (const word of vocab) {\n const count1 = vec1[word] ?? 0;\n const count2 = vec2[word] ?? 0;\n dotProduct += count1 * count2;\n mag1 += count1 * count1;\n mag2 += count2 * count2;\n }\n\n mag1 = Math.sqrt(mag1);\n mag2 = Math.sqrt(mag2);\n\n if (mag1 === 0 || mag2 === 0) return 0;\n\n return dotProduct / (mag1 * mag2);\n }\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n return {\n consolidate,\n findDuplicates,\n mergeDuplicates,\n };\n}\n","/**\n * Daemon Process Manager - Background process lifecycle management\n *\n * Handles:\n * - Process forking and detachment\n * - PID file management\n * - Signal handling (SIGTERM, SIGINT)\n * - Daemon start/stop/status commands\n */\n\nimport { fork } from 'node:child_process';\nimport { existsSync, mkdirSync, readFileSync, unlinkSync, writeFileSync } from 'node:fs';\nimport { dirname, join } from 'node:path';\nimport { fileURLToPath } from 'node:url';\nimport { getMetrics } from '../core/metrics.js';\nimport type { SparnConfig } from '../types/config.js';\n\nexport interface DaemonCommand {\n /** Start the daemon */\n start(config: SparnConfig): Promise<DaemonStartResult>;\n\n /** Stop the daemon */\n stop(config: SparnConfig): Promise<DaemonStopResult>;\n\n /** Get daemon status */\n status(config: SparnConfig): Promise<DaemonStatusResult>;\n}\n\nexport interface DaemonStartResult {\n success: boolean;\n pid?: number;\n message: string;\n error?: string;\n}\n\nexport interface DaemonStopResult {\n success: boolean;\n message: string;\n error?: string;\n}\n\nexport interface DaemonStatusResult {\n running: boolean;\n pid?: number;\n uptime?: number;\n sessionsWatched?: number;\n tokensSaved?: number;\n message: string;\n}\n\n/**\n * Create daemon command interface\n * @returns DaemonCommand instance\n */\nexport function createDaemonCommand(): DaemonCommand {\n /**\n * Check if daemon is running\n */\n function isDaemonRunning(pidFile: string): { running: boolean; pid?: number } {\n if (!existsSync(pidFile)) {\n return { running: false };\n }\n\n try {\n const pidStr = readFileSync(pidFile, 'utf-8').trim();\n const pid = Number.parseInt(pidStr, 10);\n\n if (Number.isNaN(pid)) {\n return { running: false };\n }\n\n // Check if process exists (cross-platform)\n try {\n process.kill(pid, 0); // Signal 0 checks existence without killing\n return { running: true, pid };\n } catch {\n // Process doesn't exist, clean up stale PID file\n unlinkSync(pidFile);\n return { running: false };\n }\n } catch {\n return { running: false };\n }\n }\n\n /**\n * Write PID file\n */\n function writePidFile(pidFile: string, pid: number): void {\n // Ensure directory exists\n const dir = dirname(pidFile);\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n\n writeFileSync(pidFile, String(pid), 'utf-8');\n }\n\n /**\n * Remove PID file\n */\n function removePidFile(pidFile: string): void {\n if (existsSync(pidFile)) {\n unlinkSync(pidFile);\n }\n }\n\n async function start(config: SparnConfig): Promise<DaemonStartResult> {\n const { pidFile, logFile } = config.realtime;\n\n // Check if already running\n const status = isDaemonRunning(pidFile);\n if (status.running) {\n return {\n success: false,\n pid: status.pid,\n message: `Daemon already running (PID ${status.pid})`,\n error: 'Already running',\n };\n }\n\n try {\n // Fork child process (daemon entry point)\n const __filename = fileURLToPath(import.meta.url);\n const __dirname = dirname(__filename);\n const daemonPath = join(__dirname, 'index.js');\n\n const child = fork(daemonPath, [], {\n detached: true,\n stdio: 'ignore',\n env: {\n ...process.env,\n SPARN_CONFIG: JSON.stringify(config),\n SPARN_PID_FILE: pidFile,\n SPARN_LOG_FILE: logFile,\n },\n });\n\n // Detach from parent\n child.unref();\n\n // Write PID file\n if (child.pid) {\n writePidFile(pidFile, child.pid);\n\n return {\n success: true,\n pid: child.pid,\n message: `Daemon started (PID ${child.pid})`,\n };\n }\n\n return {\n success: false,\n message: 'Failed to start daemon (no PID)',\n error: 'No PID',\n };\n } catch (error) {\n return {\n success: false,\n message: 'Failed to start daemon',\n error: error instanceof Error ? error.message : String(error),\n };\n }\n }\n\n async function stop(config: SparnConfig): Promise<DaemonStopResult> {\n const { pidFile } = config.realtime;\n\n const status = isDaemonRunning(pidFile);\n\n if (!status.running || !status.pid) {\n return {\n success: true,\n message: 'Daemon not running',\n };\n }\n\n try {\n // Send SIGTERM\n process.kill(status.pid, 'SIGTERM');\n\n // Wait for process to exit (timeout after 5s)\n const maxWait = 5000;\n const interval = 100;\n let waited = 0;\n\n while (waited < maxWait) {\n try {\n process.kill(status.pid, 0);\n // Still running, wait\n await new Promise((resolve) => setTimeout(resolve, interval));\n waited += interval;\n } catch {\n // Process exited\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon stopped (PID ${status.pid})`,\n };\n }\n }\n\n // Timeout, force kill\n try {\n process.kill(status.pid, 'SIGKILL');\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon force killed (PID ${status.pid})`,\n };\n } catch {\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon stopped (PID ${status.pid})`,\n };\n }\n } catch (error) {\n return {\n success: false,\n message: 'Failed to stop daemon',\n error: error instanceof Error ? error.message : String(error),\n };\n }\n }\n\n async function status(config: SparnConfig): Promise<DaemonStatusResult> {\n const { pidFile } = config.realtime;\n\n const daemonStatus = isDaemonRunning(pidFile);\n\n if (!daemonStatus.running || !daemonStatus.pid) {\n return {\n running: false,\n message: 'Daemon not running',\n };\n }\n\n // Get metrics snapshot\n const metrics = getMetrics().getSnapshot();\n\n return {\n running: true,\n pid: daemonStatus.pid,\n uptime: metrics.daemon.uptime,\n sessionsWatched: metrics.daemon.sessionsWatched,\n tokensSaved: metrics.optimization.totalTokensSaved,\n message: `Daemon running (PID ${daemonStatus.pid})`,\n };\n }\n\n return {\n start,\n stop,\n status,\n };\n}\n","/**\n * File Tracker - Incremental file reading with byte position tracking\n *\n * Tracks read positions for files to enable efficient incremental reading.\n * Handles JSONL partial line buffering for incomplete writes.\n *\n * Use case: Monitor Claude Code session JSONL files and only read new lines\n * as they're appended, without re-reading the entire file.\n */\n\nimport { readFileSync, statSync } from 'node:fs';\n\nexport interface FilePosition {\n /** File path */\n path: string;\n /** Last read byte position */\n position: number;\n /** Partial line buffer (for JSONL incomplete writes) */\n partialLine: string;\n /** Last modification time */\n lastModified: number;\n /** File size at last read */\n lastSize: number;\n}\n\nexport interface FileTracker {\n /**\n * Read new content from file since last read\n * @param filePath - File to read\n * @returns New content as array of lines (empty if no new content)\n */\n readNewLines(filePath: string): string[];\n\n /**\n * Get current position for a file\n * @param filePath - File path\n * @returns File position or null if not tracked\n */\n getPosition(filePath: string): FilePosition | null;\n\n /**\n * Reset position for a file (start from beginning on next read)\n * @param filePath - File path\n */\n resetPosition(filePath: string): void;\n\n /**\n * Clear all tracked positions\n */\n clearAll(): void;\n\n /**\n * Get all tracked file paths\n * @returns Array of tracked file paths\n */\n getTrackedFiles(): string[];\n}\n\n/**\n * Create a file tracker instance\n * @returns FileTracker instance\n */\nexport function createFileTracker(): FileTracker {\n // Track positions by file path\n const positions = new Map<string, FilePosition>();\n\n function readNewLines(filePath: string): string[] {\n try {\n // Get current file stats\n const stats = statSync(filePath);\n const currentSize = stats.size;\n const currentModified = stats.mtimeMs;\n\n // Get or initialize position\n let pos = positions.get(filePath);\n\n if (!pos) {\n // First read: start from beginning\n pos = {\n path: filePath,\n position: 0,\n partialLine: '',\n lastModified: currentModified,\n lastSize: 0,\n };\n positions.set(filePath, pos);\n }\n\n // Check if file was truncated or is same size\n if (currentSize < pos.lastSize || currentSize === pos.position) {\n // File truncated or no new content\n if (currentSize < pos.lastSize) {\n // Reset position if truncated\n pos.position = 0;\n pos.partialLine = '';\n }\n return [];\n }\n\n // Read new content from last position\n const buffer = Buffer.alloc(currentSize - pos.position);\n const fd = readFileSync(filePath);\n fd.copy(buffer, 0, pos.position, currentSize);\n\n // Convert to string and combine with partial line\n const newContent = (pos.partialLine + buffer.toString('utf-8')).split('\\n');\n\n // Last element might be incomplete (no trailing newline yet)\n const partialLine = newContent.pop() || '';\n\n // Update position\n pos.position = currentSize;\n pos.partialLine = partialLine;\n pos.lastModified = currentModified;\n pos.lastSize = currentSize;\n\n // Return complete lines (filter empty)\n return newContent.filter((line) => line.trim().length > 0);\n } catch (_error) {\n // File doesn't exist or can't be read\n // Return empty array (fail silently for watcher use case)\n return [];\n }\n }\n\n function getPosition(filePath: string): FilePosition | null {\n return positions.get(filePath) || null;\n }\n\n function resetPosition(filePath: string): void {\n positions.delete(filePath);\n }\n\n function clearAll(): void {\n positions.clear();\n }\n\n function getTrackedFiles(): string[] {\n return Array.from(positions.keys());\n }\n\n return {\n readNewLines,\n getPosition,\n resetPosition,\n clearAll,\n getTrackedFiles,\n };\n}\n","/**\n * Session Watcher - Monitor Claude Code session files for changes\n *\n * Uses Node.js fs.watch to monitor ~/.claude/projects/**\\/*.jsonl files.\n * Debounces events and triggers optimization when token threshold exceeded.\n * Maintains per-session ContextPipeline instances.\n */\n\nimport { type FSWatcher, readdirSync, statSync, watch } from 'node:fs';\nimport { homedir } from 'node:os';\nimport { dirname, join } from 'node:path';\nimport { type ContextPipeline, createContextPipeline } from '../core/context-pipeline.js';\nimport { getMetrics } from '../core/metrics.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { createFileTracker } from './file-tracker.js';\n\nexport interface SessionWatcherConfig {\n /** Sparn configuration */\n config: SparnConfig;\n /** Callback when optimization triggered */\n onOptimize?: (sessionId: string, stats: SessionStats) => void;\n /** Callback on error */\n onError?: (error: Error) => void;\n}\n\nexport interface SessionStats {\n /** Session ID */\n sessionId: string;\n /** Total tokens ingested */\n totalTokens: number;\n /** Current optimized tokens */\n optimizedTokens: number;\n /** Reduction percentage */\n reduction: number;\n /** Entry count */\n entryCount: number;\n /** Budget utilization */\n budgetUtilization: number;\n}\n\nexport interface SessionWatcher {\n /**\n * Start watching Claude Code session files\n * @returns Promise that resolves when watcher is ready\n */\n start(): Promise<void>;\n\n /**\n * Stop watching and cleanup\n */\n stop(): void;\n\n /**\n * Get statistics for all sessions\n * @returns Array of session stats\n */\n getStats(): SessionStats[];\n\n /**\n * Get statistics for a specific session\n * @param sessionId - Session ID\n * @returns Session stats or null if not found\n */\n getSessionStats(sessionId: string): SessionStats | null;\n\n /**\n * Manually trigger optimization for a session\n * @param sessionId - Session ID\n */\n optimizeSession(sessionId: string): void;\n}\n\n/**\n * Create a session watcher instance\n * @param config - Watcher configuration\n * @returns SessionWatcher instance\n */\nexport function createSessionWatcher(config: SessionWatcherConfig): SessionWatcher {\n const { config: sparnConfig, onOptimize, onError } = config;\n const { realtime, decay, states } = sparnConfig;\n\n // Per-session pipelines and trackers\n const pipelines = new Map<string, ContextPipeline>();\n const fileTracker = createFileTracker();\n\n // File system watchers\n const watchers: FSWatcher[] = [];\n\n // Debounce timers per file\n const debounceTimers = new Map<string, NodeJS.Timeout>();\n\n /**\n * Get Claude Code projects directory\n */\n function getProjectsDir(): string {\n return join(homedir(), '.claude', 'projects');\n }\n\n /**\n * Extract session ID from file path\n * Example: ~/.claude/projects/my-project/abc123.jsonl -> abc123\n */\n function getSessionId(filePath: string): string {\n const filename = filePath.split(/[/\\\\]/).pop() || '';\n return filename.replace(/\\.jsonl$/, '');\n }\n\n /**\n * Get or create pipeline for session\n */\n function getPipeline(sessionId: string): ContextPipeline {\n let pipeline = pipelines.get(sessionId);\n\n if (!pipeline) {\n pipeline = createContextPipeline({\n tokenBudget: realtime.tokenBudget,\n decay,\n states,\n windowSize: realtime.windowSize,\n fullOptimizationInterval: 50, // Full re-optimization every 50 incremental updates\n });\n pipelines.set(sessionId, pipeline);\n }\n\n return pipeline;\n }\n\n /**\n * Handle file change event (debounced)\n */\n function handleFileChange(filePath: string): void {\n // Clear existing timer\n const existingTimer = debounceTimers.get(filePath);\n if (existingTimer) {\n clearTimeout(existingTimer);\n }\n\n // Set new debounced timer\n const timer = setTimeout(() => {\n try {\n // Read new lines from file\n const newLines = fileTracker.readNewLines(filePath);\n\n if (newLines.length === 0) return;\n\n // Parse JSONL content\n const content = newLines.join('\\n');\n const sessionId = getSessionId(filePath);\n const pipeline = getPipeline(sessionId);\n\n // Ingest into pipeline\n pipeline.ingest(content, { sessionId, filePath });\n\n // Check if we should trigger optimization\n const stats = pipeline.getStats();\n if (stats.currentTokens >= realtime.autoOptimizeThreshold) {\n // Update daemon metrics\n getMetrics().updateDaemon({\n sessionsWatched: pipelines.size,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n // Trigger optimization callback\n if (onOptimize) {\n const sessionStats = computeSessionStats(sessionId, pipeline);\n onOptimize(sessionId, sessionStats);\n }\n }\n } catch (error) {\n if (onError) {\n onError(error instanceof Error ? error : new Error(String(error)));\n }\n } finally {\n debounceTimers.delete(filePath);\n }\n }, realtime.debounceMs);\n\n debounceTimers.set(filePath, timer);\n }\n\n /**\n * Recursively find all JSONL files in directory\n */\n function findJsonlFiles(dir: string): string[] {\n const files: string[] = [];\n\n try {\n const entries = readdirSync(dir);\n\n for (const entry of entries) {\n const fullPath = join(dir, entry);\n const stat = statSync(fullPath);\n\n if (stat.isDirectory()) {\n // Recurse into subdirectories\n files.push(...findJsonlFiles(fullPath));\n } else if (entry.endsWith('.jsonl')) {\n // Match pattern\n const matches = realtime.watchPatterns.some((pattern) => {\n // Simple glob matching (supports **/*.jsonl)\n const regex = new RegExp(\n pattern.replace(/\\*\\*/g, '.*').replace(/\\*/g, '[^/\\\\\\\\]*').replace(/\\./g, '\\\\.'),\n );\n return regex.test(fullPath);\n });\n\n if (matches) {\n files.push(fullPath);\n }\n }\n }\n } catch (_error) {\n // Ignore errors (directory might not exist yet)\n }\n\n return files;\n }\n\n /**\n * Compute session statistics\n */\n function computeSessionStats(sessionId: string, pipeline: ContextPipeline): SessionStats {\n const stats = pipeline.getStats();\n const entries = pipeline.getEntries();\n const totalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n sessionId,\n totalTokens: stats.totalIngested,\n optimizedTokens: stats.currentTokens,\n reduction: totalTokens > 0 ? (totalTokens - stats.currentTokens) / totalTokens : 0,\n entryCount: stats.currentEntries,\n budgetUtilization: stats.budgetUtilization,\n };\n }\n\n async function start(): Promise<void> {\n const projectsDir = getProjectsDir();\n\n // Find all existing JSONL files\n const jsonlFiles = findJsonlFiles(projectsDir);\n\n // Watch each file's parent directory (fs.watch is directory-based)\n const watchedDirs = new Set<string>();\n\n for (const file of jsonlFiles) {\n const dir = dirname(file);\n\n if (!watchedDirs.has(dir)) {\n const watcher = watch(dir, { recursive: false }, (_eventType, filename) => {\n if (filename?.endsWith('.jsonl')) {\n const fullPath = join(dir, filename);\n handleFileChange(fullPath);\n }\n });\n\n watchers.push(watcher);\n watchedDirs.add(dir);\n }\n }\n\n // Also watch projects directory for new subdirectories\n const projectsWatcher = watch(projectsDir, { recursive: true }, (_eventType, filename) => {\n if (filename?.endsWith('.jsonl')) {\n const fullPath = join(projectsDir, filename);\n handleFileChange(fullPath);\n }\n });\n\n watchers.push(projectsWatcher);\n\n // Update daemon metrics\n getMetrics().updateDaemon({\n startTime: Date.now(),\n sessionsWatched: jsonlFiles.length,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n }\n\n function stop(): void {\n // Close all watchers\n for (const watcher of watchers) {\n watcher.close();\n }\n watchers.length = 0;\n\n // Clear all timers\n for (const timer of debounceTimers.values()) {\n clearTimeout(timer);\n }\n debounceTimers.clear();\n\n // Clear pipelines\n pipelines.clear();\n\n // Clear file tracker\n fileTracker.clearAll();\n }\n\n function getStats(): SessionStats[] {\n const stats: SessionStats[] = [];\n\n for (const [sessionId, pipeline] of pipelines.entries()) {\n stats.push(computeSessionStats(sessionId, pipeline));\n }\n\n return stats;\n }\n\n function getSessionStats(sessionId: string): SessionStats | null {\n const pipeline = pipelines.get(sessionId);\n if (!pipeline) return null;\n\n return computeSessionStats(sessionId, pipeline);\n }\n\n function optimizeSession(sessionId: string): void {\n const pipeline = pipelines.get(sessionId);\n if (!pipeline) return;\n\n // Get entries and force full optimization\n const entries = pipeline.getEntries();\n pipeline.clear();\n pipeline.ingest(entries.map((e) => e.content).join('\\n\\n'));\n\n // Trigger callback\n if (onOptimize) {\n const stats = computeSessionStats(sessionId, pipeline);\n onOptimize(sessionId, stats);\n }\n }\n\n return {\n start,\n stop,\n getStats,\n getSessionStats,\n optimizeSession,\n };\n}\n","/**\n * Sparn MCP Server - Model Context Protocol server implementation\n *\n * Exposes Sparn's neuroscience-inspired context optimization as MCP tools,\n * enabling integration with Claude Desktop, VS Code, and other MCP clients.\n *\n * Tools:\n * - sparn_optimize: Optimize context with configurable options\n * - sparn_stats: Get optimization statistics\n * - sparn_consolidate: Run memory consolidation (sleep replay)\n */\n\nimport { McpServer } from '@modelcontextprotocol/sdk/server/mcp.js';\nimport { z } from 'zod';\nimport { createGenericAdapter } from '../adapters/generic.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSleepCompressor } from '../core/sleep-compressor.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { DEFAULT_CONFIG } from '../types/config.js';\n\n/**\n * Options for creating the Sparn MCP server.\n */\nexport interface SparnMcpServerOptions {\n /** KV memory store instance */\n memory: KVMemory;\n /** Sparn configuration (defaults to DEFAULT_CONFIG) */\n config?: SparnConfig;\n}\n\n/**\n * Create and configure the Sparn MCP server with all tools registered.\n *\n * @param options - Server options including memory store and config\n * @returns Configured McpServer instance ready to connect to a transport\n */\nexport function createSparnMcpServer(options: SparnMcpServerOptions): McpServer {\n const { memory, config = DEFAULT_CONFIG } = options;\n\n const server = new McpServer({\n name: 'sparn',\n version: '1.1.1',\n });\n\n registerOptimizeTool(server, memory, config);\n registerStatsTool(server, memory);\n registerConsolidateTool(server, memory);\n\n return server;\n}\n\n/**\n * Register the sparn_optimize tool.\n *\n * Optimizes input context using the neuroscience-inspired pipeline:\n * BTSP detection, engram scoring, confidence states, and sparse pruning.\n */\nfunction registerOptimizeTool(server: McpServer, memory: KVMemory, config: SparnConfig): void {\n server.registerTool(\n 'sparn_optimize',\n {\n title: 'Sparn Optimize',\n description:\n 'Optimize context using neuroscience-inspired pruning. ' +\n 'Applies BTSP detection, engram scoring, confidence states, ' +\n 'and sparse pruning to reduce token usage while preserving important information.',\n inputSchema: {\n context: z.string().describe('The context text to optimize'),\n dryRun: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, do not persist changes to the memory store'),\n verbose: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, include per-entry details in the response'),\n threshold: z\n .number()\n .min(0)\n .max(100)\n .optional()\n .describe('Custom pruning threshold (1-100, overrides config)'),\n },\n },\n async ({ context, dryRun, verbose, threshold }) => {\n try {\n const effectiveConfig = threshold\n ? { ...config, pruning: { ...config.pruning, threshold } }\n : config;\n\n const adapter = createGenericAdapter(memory, effectiveConfig);\n const result = await adapter.optimize(context, {\n dryRun,\n verbose,\n threshold,\n });\n\n const response = {\n optimizedContext: result.optimizedContext,\n tokensBefore: result.tokensBefore,\n tokensAfter: result.tokensAfter,\n reduction: `${(result.reduction * 100).toFixed(1)}%`,\n entriesProcessed: result.entriesProcessed,\n entriesKept: result.entriesKept,\n durationMs: result.durationMs,\n stateDistribution: result.stateDistribution,\n ...(verbose && result.details ? { details: result.details } : {}),\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n\n/**\n * Register the sparn_stats tool.\n *\n * Returns optimization statistics from the memory store, including\n * total commands run, tokens saved, and average reduction.\n */\nfunction registerStatsTool(server: McpServer, memory: KVMemory): void {\n server.registerTool(\n 'sparn_stats',\n {\n title: 'Sparn Stats',\n description:\n 'Get optimization statistics including total commands run, ' +\n 'tokens saved, and average reduction percentage.',\n inputSchema: {\n reset: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, reset all optimization statistics'),\n },\n },\n async ({ reset }) => {\n try {\n if (reset) {\n await memory.clearOptimizationStats();\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(\n {\n message: 'Optimization statistics have been reset.',\n totalCommands: 0,\n totalTokensSaved: 0,\n averageReduction: '0.0%',\n },\n null,\n 2,\n ),\n },\n ],\n };\n }\n\n const stats = await memory.getOptimizationStats();\n const totalCommands = stats.length;\n\n const totalTokensSaved = stats.reduce(\n (sum, s) => sum + (s.tokens_before - s.tokens_after),\n 0,\n );\n\n const averageReduction =\n totalCommands > 0\n ? stats.reduce((sum, s) => {\n const reduction =\n s.tokens_before > 0 ? (s.tokens_before - s.tokens_after) / s.tokens_before : 0;\n return sum + reduction;\n }, 0) / totalCommands\n : 0;\n\n const recentOptimizations = stats.slice(0, 10).map((s) => ({\n timestamp: new Date(s.timestamp).toISOString(),\n tokensBefore: s.tokens_before,\n tokensAfter: s.tokens_after,\n entriesPruned: s.entries_pruned,\n durationMs: s.duration_ms,\n reduction: `${(\n ((s.tokens_before - s.tokens_after) / Math.max(s.tokens_before, 1)) * 100\n ).toFixed(1)}%`,\n }));\n\n const response = {\n totalCommands,\n totalTokensSaved,\n averageReduction: `${(averageReduction * 100).toFixed(1)}%`,\n recentOptimizations,\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n\n/**\n * Register the sparn_consolidate tool.\n *\n * Runs the sleep-compressor consolidation process, which removes\n * decayed entries and merges duplicates in the memory store.\n */\nfunction registerConsolidateTool(server: McpServer, memory: KVMemory): void {\n server.registerTool(\n 'sparn_consolidate',\n {\n title: 'Sparn Consolidate',\n description:\n 'Run memory consolidation (sleep replay). ' +\n 'Removes decayed entries and merges duplicates to reclaim space. ' +\n 'Inspired by the neuroscience principle of sleep-based memory consolidation.',\n },\n async () => {\n try {\n const allIds = await memory.list();\n const allEntries = await Promise.all(\n allIds.map(async (id) => {\n const entry = await memory.get(id);\n return entry;\n }),\n );\n\n const entries = allEntries.filter((e) => e !== null);\n\n const compressor = createSleepCompressor();\n const result = compressor.consolidate(entries);\n\n // Apply changes to memory store\n for (const removed of result.removed) {\n await memory.delete(removed.id);\n }\n\n for (const kept of result.kept) {\n await memory.put(kept);\n }\n\n // Run VACUUM to reclaim disk space\n await memory.compact();\n\n const response = {\n entriesBefore: result.entriesBefore,\n entriesAfter: result.entriesAfter,\n decayedRemoved: result.decayedRemoved,\n duplicatesRemoved: result.duplicatesRemoved,\n compressionRatio: `${(result.compressionRatio * 100).toFixed(1)}%`,\n durationMs: result.durationMs,\n vacuumCompleted: true,\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n","/**\n * Configuration types for Sparn behavior customization.\n */\n\n/**\n * Agent adapter type.\n */\nexport type AgentType = 'claude-code' | 'generic';\n\n/**\n * Pruning configuration.\n */\nexport interface PruningConfig {\n /** Percentage of top-scored entries to keep (1-100, default: 5) */\n threshold: number;\n\n /** Aggressiveness scale 0-100 (affects TF-IDF weighting, default: 50) */\n aggressiveness: number;\n}\n\n/**\n * Decay configuration.\n */\nexport interface DecayConfig {\n /** Default TTL in hours (default: 24) */\n defaultTTL: number;\n\n /** Decay threshold for pruning (0.0-1.0, default: 0.95) */\n decayThreshold: number;\n}\n\n/**\n * Confidence state threshold configuration.\n */\nexport interface StatesConfig {\n /** Score threshold for active state (default: 0.7) */\n activeThreshold: number;\n\n /** Score threshold for ready state (default: 0.3) */\n readyThreshold: number;\n}\n\n/**\n * UI configuration.\n */\nexport interface UIConfig {\n /** Enable colored output (default: true) */\n colors: boolean;\n\n /** Enable sound effects (default: false) */\n sounds: boolean;\n\n /** Verbose logging (default: false) */\n verbose: boolean;\n}\n\n/**\n * Real-time optimization configuration.\n */\nexport interface RealtimeConfig {\n /** Target token budget for optimized context (default: 50000) */\n tokenBudget: number;\n\n /** Token threshold that triggers auto-optimization (default: 80000) */\n autoOptimizeThreshold: number;\n\n /** File patterns to watch for changes (default: ['**\\/*.jsonl']) */\n watchPatterns: string[];\n\n /** Daemon PID file path (default: '.sparn/daemon.pid') */\n pidFile: string;\n\n /** Daemon log file path (default: '.sparn/daemon.log') */\n logFile: string;\n\n /** Debounce delay in milliseconds for file changes (default: 5000) */\n debounceMs: number;\n\n /** Enable incremental optimization (default: true) */\n incremental: boolean;\n\n /** Sliding window size for context entries (default: 500) */\n windowSize: number;\n\n /** Consolidation interval in hours, or null for disabled (default: null) */\n consolidationInterval: number | null;\n}\n\n/**\n * Complete Sparn configuration.\n */\nexport interface SparnConfig {\n pruning: PruningConfig;\n decay: DecayConfig;\n states: StatesConfig;\n agent: AgentType;\n ui: UIConfig;\n /** Auto-consolidation interval in hours, or null for manual */\n autoConsolidate: number | null;\n /** Real-time optimization settings */\n realtime: RealtimeConfig;\n}\n\n/**\n * Default configuration values.\n */\nexport const DEFAULT_CONFIG: SparnConfig = {\n pruning: {\n threshold: 5,\n aggressiveness: 50,\n },\n decay: {\n defaultTTL: 24,\n decayThreshold: 0.95,\n },\n states: {\n activeThreshold: 0.7,\n readyThreshold: 0.3,\n },\n agent: 'generic',\n ui: {\n colors: true,\n sounds: false,\n verbose: false,\n },\n autoConsolidate: null,\n realtime: {\n tokenBudget: 50000,\n autoOptimizeThreshold: 80000,\n watchPatterns: ['**/*.jsonl'],\n pidFile: '.sparn/daemon.pid',\n logFile: '.sparn/daemon.log',\n debounceMs: 5000,\n incremental: true,\n windowSize: 500,\n consolidationInterval: null,\n },\n};\n","/**\n * Logging utility.\n * Simple console wrapper with log levels.\n */\n\nexport type LogLevel = 'debug' | 'info' | 'warn' | 'error';\n\n/**\n * Logger interface.\n */\nexport interface Logger {\n debug(message: string, ...args: unknown[]): void;\n info(message: string, ...args: unknown[]): void;\n warn(message: string, ...args: unknown[]): void;\n error(message: string, ...args: unknown[]): void;\n}\n\n/**\n * Create a logger with optional verbosity control.\n *\n * @param verbose - Enable debug-level logging\n * @returns Logger instance\n */\nexport function createLogger(verbose = false): Logger {\n return {\n debug(message: string, ...args: unknown[]): void {\n if (verbose) {\n console.debug(`[DEBUG] ${message}`, ...args);\n }\n },\n info(message: string, ...args: unknown[]): void {\n console.info(`[INFO] ${message}`, ...args);\n },\n warn(message: string, ...args: unknown[]): void {\n console.warn(`[WARN] ${message}`, ...args);\n },\n error(message: string, ...args: unknown[]): void {\n console.error(`[ERROR] ${message}`, ...args);\n },\n };\n}\n"],"mappings":";AAOA,SAAS,kBAAkB;;;ACF3B,SAAS,kBAAkB;AAcpB,SAAS,YAAY,SAAyB;AACnD,SAAO,WAAW,QAAQ,EAAE,OAAO,SAAS,MAAM,EAAE,OAAO,KAAK;AAClE;;;ADgBO,SAAS,qBAAmC;AAEjD,QAAM,gBAAgB;AAAA;AAAA,IAEpB;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA;AAAA,IACA;AAAA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAEA,WAAS,WAAW,SAA0B;AAC5C,WAAO,cAAc,KAAK,CAAC,YAAY,QAAQ,KAAK,OAAO,CAAC;AAAA,EAC9D;AAEA,WAAS,gBACP,SACA,OAAiB,CAAC,GAClB,WAAoC,CAAC,GACxB;AACb,WAAO;AAAA,MACL,IAAI,WAAW;AAAA,MACf;AAAA,MACA,MAAM,YAAY,OAAO;AAAA,MACzB,WAAW,KAAK,IAAI;AAAA,MACpB,OAAO;AAAA;AAAA,MACP,KAAK,MAAM,KAAK;AAAA;AAAA,MAChB,OAAO;AAAA;AAAA,MACP,aAAa;AAAA,MACb,MAAM,CAAC,GAAG,MAAM,MAAM;AAAA,MACtB;AAAA,MACA,QAAQ;AAAA,IACV;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;;;AE3CO,SAAS,uBAAuB,QAAkD;AACvF,QAAM,EAAE,iBAAiB,eAAe,IAAI;AAE5C,WAAS,eAAe,OAAqC;AAE3D,QAAI,MAAM,QAAQ;AAChB,aAAO;AAAA,IACT;AAIA,QAAI,MAAM,QAAQ,iBAAiB;AACjC,aAAO;AAAA,IACT;AAGA,QAAI,MAAM,SAAS,gBAAgB;AACjC,aAAO;AAAA,IACT;AAGA,WAAO;AAAA,EACT;AAEA,WAAS,WAAW,OAAiC;AACnD,UAAM,WAAW,eAAe,KAAK;AAErC,WAAO;AAAA,MACL,GAAG;AAAA,MACH,OAAO;AAAA,IACT;AAAA,EACF;AAEA,WAAS,gBAAgB,SAA2C;AAClE,UAAM,eAAkC;AAAA,MACtC,QAAQ;AAAA,MACR,OAAO;AAAA,MACP,QAAQ;AAAA,MACR,OAAO,QAAQ;AAAA,IACjB;AAEA,eAAW,SAAS,SAAS;AAC3B,YAAM,QAAQ,eAAe,KAAK;AAClC,mBAAa,KAAK;AAAA,IACpB;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACjDO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,WAAW,IAAI;AAEvB,WAAS,eAAe,cAAsB,cAA8B;AAC1E,QAAI,iBAAiB,EAAG,QAAO;AAC/B,QAAI,gBAAgB,EAAG,QAAO;AAG9B,UAAM,QAAQ,eAAe;AAC7B,UAAM,QAAQ,IAAI,KAAK,IAAI,CAAC,KAAK;AAGjC,WAAO,KAAK,IAAI,GAAG,KAAK,IAAI,GAAG,KAAK,CAAC;AAAA,EACvC;AAEA,WAAS,eAAe,OAAoB,cAAsB,KAAK,IAAI,GAAW;AAEpF,UAAM,oBAAoB,cAAc,MAAM;AAC9C,UAAM,eAAe,KAAK,IAAI,GAAG,oBAAoB,GAAI;AAGzD,UAAM,QAAQ,eAAe,cAAc,MAAM,GAAG;AAGpD,QAAI,QAAQ,MAAM,SAAS,IAAI;AAG/B,QAAI,MAAM,cAAc,GAAG;AACzB,YAAM,cAAc,KAAK,IAAI,MAAM,cAAc,CAAC,IAAI;AACtD,cAAQ,KAAK,IAAI,GAAK,QAAQ,WAAW;AAAA,IAC3C;AAGA,QAAI,MAAM,QAAQ;AAChB,cAAQ,KAAK,IAAI,OAAO,GAAG;AAAA,IAC7B;AAEA,WAAO,KAAK,IAAI,GAAG,KAAK,IAAI,GAAG,KAAK,CAAC;AAAA,EACvC;AAEA,WAAS,WAAW,OAAiC;AACnD,WAAO;AAAA,MACL,GAAG;AAAA,MACH,KAAK,aAAa;AAAA;AAAA,MAClB,WAAW,KAAK,IAAI;AAAA,IACtB;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AClFO,SAAS,eAAe,MAAsB;AACnD,MAAI,CAAC,QAAQ,KAAK,WAAW,GAAG;AAC9B,WAAO;AAAA,EACT;AAGA,QAAM,QAAQ,KAAK,MAAM,KAAK,EAAE,OAAO,CAAC,MAAM,EAAE,SAAS,CAAC;AAC1D,QAAM,YAAY,MAAM;AAGxB,QAAM,YAAY,KAAK;AACvB,QAAM,eAAe,KAAK,KAAK,YAAY,CAAC;AAG5C,QAAM,eAAe,KAAK,KAAK,YAAY,IAAI;AAG/C,SAAO,KAAK,IAAI,cAAc,YAAY;AAC5C;;;ACAO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,UAAU,IAAI;AAEtB,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,WAAS,YAAY,MAAc,QAA0B;AAC3D,UAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAE/C,WAAO,KAAK,KAAK,KAAK;AAAA,EACxB;AAEA,WAAS,aAAa,MAAc,YAAmC;AACrE,UAAM,YAAY,WAAW;AAC7B,UAAM,eAAe,WAAW,OAAO,CAAC,UAAU;AAChD,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,aAAO,OAAO,SAAS,IAAI;AAAA,IAC7B,CAAC,EAAE;AAEH,QAAI,iBAAiB,EAAG,QAAO;AAE/B,WAAO,KAAK,IAAI,YAAY,YAAY;AAAA,EAC1C;AAEA,WAAS,WAAW,OAAoB,YAAmC;AACzE,UAAM,SAAS,SAAS,MAAM,OAAO;AACrC,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AACvC,QAAI,aAAa;AAEjB,eAAW,QAAQ,aAAa;AAC9B,YAAM,KAAK,YAAY,MAAM,MAAM;AACnC,YAAM,MAAM,aAAa,MAAM,UAAU;AACzC,oBAAc,KAAK;AAAA,IACrB;AAGA,WAAO,aAAa,OAAO;AAAA,EAC7B;AAEA,WAAS,MAAM,SAAqC;AAClD,QAAI,QAAQ,WAAW,GAAG;AACxB,aAAO;AAAA,QACL,MAAM,CAAC;AAAA,QACP,SAAS,CAAC;AAAA,QACV,gBAAgB;AAAA,QAChB,cAAc;AAAA,MAChB;AAAA,IACF;AAGA,UAAM,iBAAiB,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,SAAS,QAAQ,IAAI,CAAC,WAAW;AAAA,MACrC;AAAA,MACA,OAAO,WAAW,OAAO,OAAO;AAAA,IAClC,EAAE;AAGF,WAAO,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAGvC,UAAM,YAAY,KAAK,IAAI,GAAG,KAAK,KAAK,QAAQ,UAAU,YAAY,IAAI,CAAC;AAC3E,UAAM,OAAO,OAAO,MAAM,GAAG,SAAS,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAC1D,UAAM,UAAU,OAAO,MAAM,SAAS,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAG1D,UAAM,eAAe,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAE/E,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;;;ACpHA,SAAS,cAAAA,mBAAkB;AAepB,SAAS,uBAAuB,SAAgC;AACrE,QAAM,UAAyB,CAAC;AAChC,QAAM,MAAM,KAAK,IAAI;AAGrB,QAAM,QAAQ,QAAQ,MAAM,IAAI;AAChC,MAAI,eAAyB,CAAC;AAC9B,MAAI,YAAuB;AAE3B,aAAW,QAAQ,OAAO;AACxB,UAAM,UAAU,KAAK,KAAK;AAG1B,QAAI,QAAQ,WAAW,OAAO,KAAK,QAAQ,WAAW,YAAY,GAAG;AACnE,UAAI,aAAa,SAAS,GAAG;AAC3B,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAGE,QAAQ,SAAS,kBAAkB,KACnC,QAAQ,SAAS,UAAU,KAC3B,QAAQ,SAAS,YAAY,GAC7B;AACA,UAAI,aAAa,SAAS,GAAG;AAC3B,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,QAAQ,SAAS,oBAAoB,KAAK,QAAQ,SAAS,qBAAqB,GAAG;AAC1F,UAAI,aAAa,SAAS,KAAK,cAAc,UAAU;AACrD,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,aAAa,SAAS,GAAG;AAChC,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,QAAQ,SAAS,GAAG;AAC3B,mBAAa,KAAK,IAAI;AACtB,kBAAY;AAAA,IACd;AAAA,EACF;AAGA,MAAI,aAAa,SAAS,GAAG;AAC3B,YAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AAAA,EACnE;AAEA,SAAO,QAAQ,OAAO,CAAC,MAAM,EAAE,QAAQ,KAAK,EAAE,SAAS,CAAC;AAC1D;AASO,SAAS,YAAY,SAAiB,MAAiB,UAA+B;AAC3F,QAAM,OAAiB,CAAC,IAAI;AAG5B,MAAI,eAAe;AACnB,MAAI,SAAS,eAAgB,gBAAe;AAC5C,MAAI,SAAS,OAAQ,gBAAe;AACpC,MAAI,SAAS,SAAU,gBAAe;AAEtC,SAAO;AAAA,IACL,IAAIC,YAAW;AAAA,IACf;AAAA,IACA,MAAM,YAAY,OAAO;AAAA,IACzB,WAAW;AAAA,IACX,OAAO;AAAA,IACP,OAAO,eAAe,MAAM,WAAW,eAAe,MAAM,UAAU;AAAA,IACtE,KAAK,KAAK;AAAA;AAAA,IACV,aAAa;AAAA,IACb;AAAA,IACA,UAAU,EAAE,KAAK;AAAA,IACjB,QAAQ;AAAA,EACV;AACF;AAQO,SAAS,oBAAoB,SAAgC;AAClE,QAAM,UAAyB,CAAC;AAChC,QAAM,MAAM,KAAK,IAAI;AAGrB,QAAM,SAAS,QAAQ,MAAM,OAAO;AAEpC,aAAW,SAAS,QAAQ;AAC1B,UAAM,UAAU,MAAM,KAAK;AAC3B,QAAI,QAAQ,WAAW,EAAG;AAE1B,YAAQ,KAAK,YAAY,SAAS,SAAS,GAAG,CAAC;AAAA,EACjD;AAEA,SAAO;AACT;;;ACrHA,IAAM,sBAAsB;AAAA;AAAA,EAE1B,qBAAqB;AAAA;AAAA;AAAA,EAGrB,mBAAmB;AAAA;AAAA;AAAA,EAGnB,qBAAqB,KAAK;AAAA;AAAA;AAAA,EAG1B,cAAc;AAAA;AAAA,IAEZ;AAAA,IACA;AAAA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,EACF;AACF;AAQO,SAAS,wBAAwB,QAAkB,QAAmC;AAE3F,QAAM,SAAS,mBAAmB;AAAA,IAChC,WAAW,OAAO,QAAQ;AAAA,EAC5B,CAAC;AAED,QAAM,SAAS,mBAAmB,OAAO,KAAK;AAC9C,QAAM,SAAS,uBAAuB,OAAO,MAAM;AACnD,QAAM,OAAO,mBAAmB;AAEhC,iBAAe,SACb,SACA,UAA2B,CAAC,GACC;AAC7B,UAAM,YAAY,KAAK,IAAI;AAI3B,UAAM,UAAU,uBAAuB,OAAO;AAG9C,UAAM,kBAAkB,QAAQ,IAAI,CAAC,UAAU;AAC7C,YAAM,SAAS,oBAAoB,aAAa;AAAA,QAAK,CAAC,YACpD,QAAQ,KAAK,MAAM,OAAO;AAAA,MAC5B;AAEA,UAAI,QAAQ;AACV,cAAM,YAAY,KAAK,gBAAgB,MAAM,SAAS,CAAC,GAAG,MAAM,MAAM,aAAa,GAAG;AAAA,UACpF,mBAAmB,MAAM;AAAA,QAC3B,CAAC;AAED,eAAO;AAAA,UACL,GAAG;AAAA,UACH,WAAW,MAAM;AAAA,QACnB;AAAA,MACF;AAEA,aAAO;AAAA,IACT,CAAC;AAGD,UAAM,iBAAiB,gBAAgB,IAAI,CAAC,UAAU;AACpD,YAAM,qBACJ,MAAM,QAAQ,KAAK,EAAE,WAAW,OAAO,KAAK,MAAM,QAAQ,KAAK,EAAE,WAAW,YAAY;AAE1F,UAAI,oBAAoB;AACtB,eAAO;AAAA,UACL,GAAG;AAAA,UACH,OAAO,MAAM,QAAQ,oBAAoB;AAAA,QAC3C;AAAA,MACF;AAEA,aAAO;AAAA,IACT,CAAC;AAGD,UAAM,gBAAgB,eAAe,IAAI,CAAC,UAAU;AAClD,YAAM,aAAa,OAAO,eAAe,KAAK;AAC9C,aAAO;AAAA,QACL,GAAG;AAAA,QACH,OAAO;AAAA,MACT;AAAA,IACF,CAAC;AAGD,UAAM,oBAAoB,cAAc,IAAI,CAAC,UAAU;AACrD,YAAM,QAAQ,OAAO,eAAe,KAAK;AACzC,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MACF;AAAA,IACF,CAAC;AAGD,UAAM,cAAc,OAAO,MAAM,iBAAiB;AAGlD,QAAI,CAAC,QAAQ,QAAQ;AACnB,iBAAW,SAAS,YAAY,MAAM;AACpC,cAAM,OAAO,IAAI,KAAK;AAAA,MACxB;AAGA,YAAM,OAAO,mBAAmB;AAAA,QAC9B,WAAW,KAAK,IAAI;AAAA,QACpB,eAAe,YAAY;AAAA,QAC3B,cAAc,YAAY;AAAA,QAC1B,gBAAgB,YAAY,QAAQ;AAAA,QACpC,aAAa,KAAK,IAAI,IAAI;AAAA,MAC5B,CAAC;AAAA,IACH;AAGA,UAAM,mBAAmB,YAAY,KAAK,IAAI,CAAC,UAAU,MAAM,OAAO,EAAE,KAAK,IAAI;AAGjF,UAAM,oBAAoB,OAAO,gBAAgB,YAAY,IAAI;AAGjE,UAAM,SAA6B;AAAA,MACjC;AAAA,MACA,cAAc,YAAY;AAAA,MAC1B,aAAa,YAAY;AAAA,MACzB,WACE,YAAY,iBAAiB,KACxB,YAAY,iBAAiB,YAAY,gBAAgB,YAAY,iBACtE;AAAA,MACN,kBAAkB,QAAQ;AAAA,MAC1B,aAAa,YAAY,KAAK;AAAA,MAC9B,YAAY,KAAK,IAAI,IAAI;AAAA,MACzB;AAAA,IACF;AAGA,QAAI,QAAQ,SAAS;AACnB,aAAO,UAAU,YAAY,KAAK,IAAI,CAAC,WAAW;AAAA,QAChD,IAAI,MAAM;AAAA,QACV,OAAO,MAAM;AAAA,QACb,OAAO,MAAM,SAAS;AAAA,QACtB,QAAQ,MAAM,KAAK,SAAS,MAAM;AAAA,QAClC,QAAQ,eAAe,MAAM,OAAO;AAAA,MACtC,EAAE;AAAA,IACJ;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,EACF;AACF;;;ACvLA,SAAS,cAAAC,mBAAkB;AAkBpB,SAAS,qBAAqB,QAAkB,QAAmC;AACxF,QAAM,SAAS,mBAAmB,OAAO,OAAO;AAChD,QAAM,SAAS,mBAAmB,OAAO,KAAK;AAC9C,QAAM,SAAS,uBAAuB,OAAO,MAAM;AACnD,QAAM,OAAO,mBAAmB;AAEhC,iBAAe,SACb,SACA,UAA2B,CAAC,GACC;AAC7B,UAAM,YAAY,KAAK,IAAI;AAG3B,UAAM,QAAQ,QAAQ,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,EAAE,SAAS,CAAC;AACzE,UAAM,UAAyB,MAAM,IAAI,CAAC,aAAa;AAAA,MACrD,IAAIC,YAAW;AAAA,MACf;AAAA,MACA,MAAM,YAAY,OAAO;AAAA,MACzB,WAAW,KAAK,IAAI;AAAA,MACpB,OAAO,KAAK,WAAW,OAAO,IAAI,IAAM;AAAA;AAAA,MACxC,KAAK,OAAO,MAAM,aAAa;AAAA;AAAA,MAC/B,OAAO;AAAA,MACP,aAAa;AAAA,MACb,MAAM,CAAC;AAAA,MACP,UAAU,CAAC;AAAA,MACX,QAAQ,KAAK,WAAW,OAAO;AAAA,IACjC,EAAE;AAGF,UAAM,eAAe,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGlF,UAAM,gBAAgB,QAAQ,IAAI,CAAC,WAAW;AAAA,MAC5C,GAAG;AAAA,MACH,OAAO,OAAO,eAAe,KAAK;AAAA,IACpC,EAAE;AAGF,UAAM,gBAAgB,cAAc,IAAI,CAAC,UAAU,OAAO,WAAW,KAAK,CAAC;AAG3E,UAAM,cAAc,OAAO,MAAM,aAAa;AAG9C,UAAM,mBAAmB,YAAY,KAAK;AAAA,MACxC,CAAC,MAAM,EAAE,UAAU,YAAY,EAAE,UAAU;AAAA,IAC7C;AAGA,UAAM,cAAc,iBAAiB,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAG1F,UAAM,mBAAmB,iBAAiB,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,IAAI;AAGzE,QAAI,CAAC,QAAQ,QAAQ;AACnB,iBAAW,SAAS,kBAAkB;AACpC,cAAM,OAAO,IAAI,KAAK;AAAA,MACxB;AAGA,YAAM,OAAO,mBAAmB;AAAA,QAC9B,WAAW,KAAK,IAAI;AAAA,QACpB,eAAe;AAAA,QACf,cAAc;AAAA,QACd,gBAAgB,QAAQ,SAAS,iBAAiB;AAAA,QAClD,aAAa,KAAK,IAAI,IAAI;AAAA,MAC5B,CAAC;AAAA,IACH;AAGA,UAAM,eAAe,OAAO,gBAAgB,gBAAgB;AAE5D,UAAM,SAA6B;AAAA,MACjC;AAAA,MACA;AAAA,MACA;AAAA,MACA,WAAW,eAAe,KAAK,eAAe,eAAe,eAAe;AAAA,MAC5E,kBAAkB,QAAQ;AAAA,MAC1B,aAAa,iBAAiB;AAAA,MAC9B,mBAAmB;AAAA,MACnB,YAAY,KAAK,IAAI,IAAI;AAAA,IAC3B;AAGA,QAAI,QAAQ,SAAS;AACnB,aAAO,UAAU,iBAAiB,IAAI,CAAC,OAAO;AAAA,QAC5C,IAAI,EAAE;AAAA,QACN,OAAO,EAAE;AAAA,QACT,OAAO,EAAE;AAAA,QACT,QAAQ,EAAE;AAAA,QACV,QAAQ,eAAe,EAAE,OAAO;AAAA,MAClC,EAAE;AAAA,IACJ;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,EACF;AACF;;;ACpEO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,aAAa,MAAM,IAAI;AAC/B,QAAM,eAAe,mBAAmB,KAAK;AAE7C,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,WAAS,YAAY,MAAc,QAA0B;AAC3D,UAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAE/C,WAAO,KAAK,KAAK,KAAK;AAAA,EACxB;AAEA,WAAS,aAAa,MAAc,YAAmC;AACrE,UAAM,YAAY,WAAW;AAC7B,UAAM,eAAe,WAAW,OAAO,CAAC,UAAU;AAChD,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,aAAO,OAAO,SAAS,IAAI;AAAA,IAC7B,CAAC,EAAE;AAEH,QAAI,iBAAiB,EAAG,QAAO;AAE/B,WAAO,KAAK,IAAI,YAAY,YAAY;AAAA,EAC1C;AAEA,WAAS,eAAe,OAAoB,YAAmC;AAC7E,UAAM,SAAS,SAAS,MAAM,OAAO;AACrC,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AACvC,QAAI,aAAa;AAEjB,eAAW,QAAQ,aAAa;AAC9B,YAAM,KAAK,YAAY,MAAM,MAAM;AACnC,YAAM,MAAM,aAAa,MAAM,UAAU;AACzC,oBAAc,KAAK;AAAA,IACrB;AAGA,WAAO,aAAa,OAAO;AAAA,EAC7B;AAEA,WAAS,mBAAmB,OAA4B;AAEtD,QAAI,MAAM,OAAQ,QAAO;AAGzB,YAAQ,MAAM,OAAO;AAAA,MACnB,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT;AACE,eAAO;AAAA,IACX;AAAA,EACF;AAEA,WAAS,cAAc,OAAoB,YAAmC;AAC5E,UAAM,QAAQ,eAAe,OAAO,UAAU;AAC9C,UAAM,eAAe,aAAa,eAAe,KAAK;AACtD,UAAM,cAAc,IAAI;AACxB,UAAM,kBAAkB,mBAAmB,KAAK;AAIhD,WAAO,SAAS,IAAI,eAAe;AAAA,EACrC;AAEA,WAAS,WACP,SACA,SAAiB,aAC4B;AAC7C,QAAI,QAAQ,WAAW,GAAG;AACxB,aAAO;AAAA,QACL,MAAM,CAAC;AAAA,QACP,SAAS,CAAC;AAAA,QACV,gBAAgB;AAAA,QAChB,cAAc;AAAA,QACd,mBAAmB;AAAA,MACrB;AAAA,IACF;AAGA,UAAM,iBAAiB,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,cAAc,QAAQ,OAAO,CAAC,MAAM,EAAE,MAAM;AAClD,UAAM,iBAAiB,QAAQ,OAAO,CAAC,MAAM,CAAC,EAAE,MAAM;AAEtD,UAAM,aAAa,YAAY,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,SAAS,eAAe,IAAI,CAAC,WAAW;AAAA,MAC5C;AAAA,MACA,OAAO,cAAc,OAAO,OAAO;AAAA,MACnC,QAAQ,eAAe,MAAM,OAAO;AAAA,IACtC,EAAE;AAGF,WAAO,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAGvC,UAAM,OAAsB,CAAC,GAAG,WAAW;AAC3C,UAAM,UAAyB,CAAC;AAChC,QAAI,gBAAgB;AAEpB,eAAW,QAAQ,QAAQ;AACzB,UAAI,gBAAgB,KAAK,UAAU,QAAQ;AACzC,aAAK,KAAK,KAAK,KAAK;AACpB,yBAAiB,KAAK;AAAA,MACxB,OAAO;AACL,gBAAQ,KAAK,KAAK,KAAK;AAAA,MACzB;AAAA,IACF;AAEA,UAAM,oBAAoB,SAAS,IAAI,gBAAgB,SAAS;AAEhE,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA,cAAc;AAAA,MACd;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;AASO,SAAS,6BACd,gBACA,aACA,cACc;AACd,SAAO,mBAAmB;AAAA,IACxB,aAAa,eAAe;AAAA,IAC5B,OAAO;AAAA,IACP,QAAQ;AAAA,EACV,CAAC;AACH;;;AChIO,SAAS,yBAA2C;AACzD,QAAM,gBAAsC,CAAC;AAC7C,MAAI,gBAA8B;AAAA,IAChC,WAAW,KAAK,IAAI;AAAA,IACpB,iBAAiB;AAAA,IACjB,oBAAoB;AAAA,IACpB,kBAAkB;AAAA,IAClB,gBAAgB;AAAA,IAChB,aAAa;AAAA,EACf;AAEA,MAAI,YAAY;AAChB,MAAI,cAAc;AAElB,WAAS,mBAAmB,QAAkC;AAC5D,kBAAc,KAAK,MAAM;AAGzB,kBAAc;AACd,kBAAc,oBAAoB,OAAO,eAAe,OAAO;AAG/D,QAAI,OAAO,eAAe,GAAG;AAC3B,YAAM,OAAO,KAAK,MAAM,OAAO,mBAAmB,OAAO,YAAY;AACrE,mBAAa;AACb,qBAAe,OAAO,mBAAmB;AAAA,IAC3C;AAGA,kBAAc,kBACX,cAAc,kBAAkB,cAAc,qBAAqB,KAAK,OAAO,YAChF,cAAc;AAGhB,QAAI,cAAc,SAAS,KAAM;AAC/B,oBAAc,MAAM;AAAA,IACtB;AAAA,EACF;AAEA,WAAS,aAAa,QAAqC;AACzD,oBAAgB;AAAA,MACd,GAAG;AAAA,MACH,GAAG;AAAA,IACL;AAAA,EACF;AAEA,WAAS,oBAAoB,QAAkB,YAA4B;AACzE,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,SAAS,CAAC,GAAG,MAAM,EAAE,KAAK,CAAC,GAAG,MAAM,IAAI,CAAC;AAC/C,UAAM,QAAQ,KAAK,KAAM,aAAa,MAAO,OAAO,MAAM,IAAI;AAC9D,WAAO,OAAO,KAAK,KAAK;AAAA,EAC1B;AAEA,WAAS,cAA+B;AACtC,UAAM,YAAY,cAAc;AAChC,UAAM,gBAAgB,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,UAAU,CAAC;AAC1E,UAAM,mBAAmB,cAAc;AAAA,MACrC,CAAC,KAAK,MAAM,OAAO,EAAE,eAAe,EAAE;AAAA,MACtC;AAAA,IACF;AAEA,UAAM,oBAAoB,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,cAAc,CAAC;AAClF,UAAM,mBAAmB,oBAAoB,IAAI,mBAAmB,oBAAoB;AAExF,UAAM,YAAY,cAAc,IAAI,CAAC,MAAM,EAAE,QAAQ;AAErD,UAAM,oBAAoB,YAAY;AACtC,UAAM,UAAU,oBAAoB,IAAI,YAAY,oBAAoB;AAExE,WAAO;AAAA,MACL,WAAW,KAAK,IAAI;AAAA,MACpB,cAAc;AAAA,QACZ;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA,YAAY,oBAAoB,WAAW,EAAE;AAAA,QAC7C,YAAY,oBAAoB,WAAW,EAAE;AAAA,QAC7C,YAAY,oBAAoB,WAAW,EAAE;AAAA,MAC/C;AAAA,MACA,OAAO;AAAA,QACL;AAAA,QACA,WAAW;AAAA,QACX,aAAa;AAAA,QACb,MAAM,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,aAAa,CAAC;AAAA,MAC/D;AAAA,MACA,QAAQ;AAAA,QACN,QAAQ,KAAK,IAAI,IAAI,cAAc;AAAA,QACnC,iBAAiB,cAAc;AAAA,QAC/B,aAAa,cAAc;AAAA,MAC7B;AAAA,IACF;AAAA,EACF;AAEA,WAAS,gBAAwB;AAC/B,WAAO,KAAK,UAAU,YAAY,GAAG,MAAM,CAAC;AAAA,EAC9C;AAEA,WAAS,QAAc;AACrB,kBAAc,SAAS;AACvB,gBAAY;AACZ,kBAAc;AACd,oBAAgB;AAAA,MACd,WAAW,KAAK,IAAI;AAAA,MACpB,iBAAiB;AAAA,MACjB,oBAAoB;AAAA,MACpB,kBAAkB;AAAA,MAClB,gBAAgB;AAAA,MAChB,aAAa;AAAA,IACf;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ;AAAA,IACR;AAAA,EACF;AACF;AAGA,IAAI,gBAAyC;AAKtC,SAAS,aAA+B;AAC7C,MAAI,CAAC,eAAe;AAClB,oBAAgB,uBAAuB;AAAA,EACzC;AACA,SAAO;AACT;;;AC3HO,SAAS,2BACd,QACsB;AACtB,QAAM,SAAS,mBAAmB,MAAM;AACxC,QAAM,EAAE,yBAAyB,IAAI;AAGrC,MAAI,QAAmC;AAAA,IACrC,YAAY,oBAAI,IAAI;AAAA,IACpB,mBAAmB,oBAAI,IAAI;AAAA,IAC3B,gBAAgB;AAAA,IAChB,aAAa;AAAA,IACb,sBAAsB,KAAK,IAAI;AAAA,EACjC;AAEA,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAKA,WAAS,wBAAwB,SAAwB,SAAS,OAAa;AAC7E,eAAW,SAAS,SAAS;AAC3B,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,YAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AAEvC,iBAAW,QAAQ,aAAa;AAC9B,cAAM,UAAU,MAAM,kBAAkB,IAAI,IAAI,KAAK;AACrD,cAAM,UAAU,SAAS,KAAK,IAAI,GAAG,UAAU,CAAC,IAAI,UAAU;AAE9D,YAAI,YAAY,GAAG;AACjB,gBAAM,kBAAkB,OAAO,IAAI;AAAA,QACrC,OAAO;AACL,gBAAM,kBAAkB,IAAI,MAAM,OAAO;AAAA,QAC3C;AAAA,MACF;AAAA,IACF;AAEA,UAAM,kBAAkB,SAAS,CAAC,QAAQ,SAAS,QAAQ;AAC3D,UAAM,iBAAiB,KAAK,IAAI,GAAG,MAAM,cAAc;AAAA,EACzD;AAKA,WAAS,eAAe,MAAkC;AACxD,UAAM,SAAS,MAAM,WAAW,IAAI,IAAI;AACxC,QAAI,CAAC,OAAQ,QAAO;AAGpB,WAAO,OAAO;AAAA,EAChB;AAKA,WAAS,WAAW,OAAoB,OAAqB;AAC3D,UAAM,WAAW,IAAI,MAAM,MAAM;AAAA,MAC/B;AAAA,MACA;AAAA,MACA,WAAW,KAAK,IAAI;AAAA,IACtB,CAAC;AAAA,EACH;AAEA,WAAS,oBACP,YACA,QAC6C;AAC7C,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM;AAGN,QAAI,MAAM,eAAe,0BAA0B;AAEjD,YAAMC,cAAa,MAAM,KAAK,MAAM,WAAW,OAAO,CAAC,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAC3E,aAAO,aAAa,CAAC,GAAGA,aAAY,GAAG,UAAU,GAAG,MAAM;AAAA,IAC5D;AAGA,UAAM,kBAAiC,CAAC;AACxC,UAAM,gBAA+B,CAAC;AAEtC,eAAW,SAAS,YAAY;AAC9B,YAAM,SAAS,eAAe,MAAM,IAAI;AACxC,UAAI,QAAQ;AACV,sBAAc,KAAK,MAAM;AAAA,MAC3B,OAAO;AACL,wBAAgB,KAAK,KAAK;AAAA,MAC5B;AAAA,IACF;AAGA,QAAI,gBAAgB,SAAS,GAAG;AAC9B,8BAAwB,iBAAiB,KAAK;AAAA,IAChD;AAGA,UAAM,aAAa,CAAC,GAAG,eAAe,GAAG,eAAe;AAGxD,eAAW,SAAS,iBAAiB;AACnC,YAAM,QAAQ,OAAO,cAAc,OAAO,UAAU;AACpD,iBAAW,OAAO,KAAK;AAAA,IACzB;AAGA,UAAM,iBAAiB,MAAM,KAAK,MAAM,WAAW,OAAO,CAAC,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAG/E,UAAM,eAAe,eAAe,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGzF,UAAM,SAAS,OAAO,WAAW,gBAAgB,MAAM;AAGvD,UAAM,cAAc,OAAO,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,eAAW,WAAW,OAAO,SAAS;AACpC,YAAM,WAAW,OAAO,QAAQ,IAAI;AAAA,IACtC;AAGA,QAAI,OAAO,QAAQ,SAAS,GAAG;AAC7B,8BAAwB,OAAO,SAAS,IAAI;AAAA,IAC9C;AAGA,UAAM,WAAW,KAAK,IAAI,IAAI;AAC9B,UAAM,eAAe,WAAW,SAAS,IAAI,cAAc,SAAS,WAAW,SAAS;AAExF,eAAW,EAAE,mBAAmB;AAAA,MAC9B,WAAW,KAAK,IAAI;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA,kBAAkB,WAAW;AAAA,MAC7B,aAAa,OAAO,KAAK;AAAA,MACzB;AAAA,MACA,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAED,WAAO;AAAA,EACT;AAEA,WAAS,aACP,YACA,QAC6C;AAC7C,UAAM,YAAY,KAAK,IAAI;AAG3B,UAAM,eAAe,WAAW,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,UAAM,WAAW,MAAM;AACvB,UAAM,kBAAkB,MAAM;AAC9B,UAAM,iBAAiB;AACvB,UAAM,cAAc;AACpB,UAAM,uBAAuB,KAAK,IAAI;AAGtC,4BAAwB,YAAY,KAAK;AAGzC,eAAW,SAAS,YAAY;AAC9B,YAAM,QAAQ,OAAO,cAAc,OAAO,UAAU;AACpD,iBAAW,OAAO,KAAK;AAAA,IACzB;AAGA,UAAM,SAAS,OAAO,WAAW,YAAY,MAAM;AAGnD,UAAM,cAAc,OAAO,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,eAAW,WAAW,OAAO,SAAS;AACpC,YAAM,WAAW,OAAO,QAAQ,IAAI;AAAA,IACtC;AAGA,QAAI,OAAO,QAAQ,SAAS,GAAG;AAC7B,8BAAwB,OAAO,SAAS,IAAI;AAAA,IAC9C;AAGA,UAAM,WAAW,KAAK,IAAI,IAAI;AAE9B,eAAW,EAAE,mBAAmB;AAAA,MAC9B,WAAW,KAAK,IAAI;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA,kBAAkB,WAAW;AAAA,MAC7B,aAAa,OAAO,KAAK;AAAA,MACzB,cAAc;AAAA;AAAA,MACd,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAED,WAAO;AAAA,EACT;AAEA,WAAS,WAAsC;AAC7C,WAAO;AAAA,MACL,YAAY,IAAI,IAAI,MAAM,UAAU;AAAA,MACpC,mBAAmB,IAAI,IAAI,MAAM,iBAAiB;AAAA,MAClD,gBAAgB,MAAM;AAAA,MACtB,aAAa,MAAM;AAAA,MACnB,sBAAsB,MAAM;AAAA,IAC9B;AAAA,EACF;AAEA,WAAS,aAAa,eAAgD;AACpE,YAAQ;AAAA,MACN,YAAY,IAAI,IAAI,cAAc,UAAU;AAAA,MAC5C,mBAAmB,IAAI,IAAI,cAAc,iBAAiB;AAAA,MAC1D,gBAAgB,cAAc;AAAA,MAC9B,aAAa,cAAc;AAAA,MAC3B,sBAAsB,cAAc;AAAA,IACtC;AAAA,EACF;AAEA,WAAS,QAAc;AACrB,YAAQ;AAAA,MACN,YAAY,oBAAI,IAAI;AAAA,MACpB,mBAAmB,oBAAI,IAAI;AAAA,MAC3B,gBAAgB;AAAA,MAChB,aAAa;AAAA,MACb,sBAAsB,KAAK,IAAI;AAAA,IACjC;AAAA,EACF;AAEA,WAAS,WAAW;AAClB,WAAO;AAAA,MACL,eAAe,MAAM,WAAW;AAAA,MAChC,aAAa,MAAM,kBAAkB;AAAA,MACrC,gBAAgB,MAAM;AAAA,MACtB,aAAa,MAAM;AAAA,MACnB,sBAAsB,MAAM;AAAA,IAC9B;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5QO,SAAS,sBAAsB,QAAgD;AACpF,QAAM,YAAY,2BAA2B,MAAM;AACnD,QAAM,EAAE,YAAY,YAAY,IAAI;AAGpC,MAAI,gBAAgB;AACpB,MAAI,iBAAiB;AACrB,MAAI,iBAAgC,CAAC;AACrC,MAAI,oBAAoB;AAExB,WAAS,OAAO,SAAiB,WAAoC,CAAC,GAAW;AAE/E,UAAM,aAAa,uBAAuB,OAAO;AAEjD,QAAI,WAAW,WAAW,EAAG,QAAO;AAGpC,UAAM,sBAAsB,WAAW,IAAI,CAAC,WAAW;AAAA,MACrD,GAAG;AAAA,MACH,UAAU,EAAE,GAAG,MAAM,UAAU,GAAG,SAAS;AAAA,IAC7C,EAAE;AAGF,UAAM,SAAS,UAAU,oBAAoB,qBAAqB,WAAW;AAG7E,qBAAiB,WAAW;AAC5B,sBAAkB,OAAO,QAAQ;AACjC,qBAAiB,OAAO;AACxB,wBAAoB,OAAO;AAG3B,QAAI,eAAe,SAAS,YAAY;AAEtC,YAAM,SAAS,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAC3E,YAAM,SAAS,OAAO,MAAM,GAAG,UAAU;AACzC,YAAM,WAAW,OAAO,MAAM,UAAU;AAExC,uBAAiB;AACjB,wBAAkB,SAAS;AAAA,IAC7B;AAEA,WAAO,WAAW;AAAA,EACpB;AAEA,WAAS,aAAqB;AAE5B,UAAM,SAAS,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAC3E,WAAO,OAAO,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,MAAM;AAAA,EACjD;AAEA,WAAS,aAA4B;AAEnC,WAAO,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAAA,EACrE;AAEA,WAAS,WAAiC;AACxC,UAAM,iBAAiB,UAAU,SAAS;AAC1C,UAAM,gBAAgB,eAAe,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAE1F,WAAO;AAAA,MACL;AAAA,MACA,gBAAgB,eAAe;AAAA,MAC/B;AAAA,MACA;AAAA,MACA;AAAA,MACA,WAAW;AAAA,QACT,eAAe,eAAe;AAAA,QAC9B,aAAa,eAAe;AAAA,QAC5B,aAAa,eAAe;AAAA,MAC9B;AAAA,IACF;AAAA,EACF;AAEA,WAAS,QAAc;AACrB,oBAAgB;AAChB,qBAAiB;AACjB,qBAAiB,CAAC;AAClB,wBAAoB;AACpB,cAAU,MAAM;AAAA,EAClB;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACpKA,SAAS,cAAc,kBAAkB;AACzC,OAAO,cAAc;AAuDrB,SAAS,aAAa,QAAwB;AAC5C,QAAM,aAAY,oBAAI,KAAK,GAAE,YAAY,EAAE,QAAQ,SAAS,GAAG;AAC/D,QAAM,aAAa,GAAG,MAAM,WAAW,SAAS;AAEhD,MAAI;AACF,iBAAa,QAAQ,UAAU;AAC/B,YAAQ,IAAI,iCAA4B,UAAU,EAAE;AACpD,WAAO;AAAA,EACT,SAAS,OAAO;AACd,YAAQ,MAAM,qCAAqC,KAAK,EAAE;AAC1D,WAAO;AAAA,EACT;AACF;AAYA,eAAsB,eAAe,QAAmC;AAEtE,MAAI;AACJ,MAAI;AACF,SAAK,IAAI,SAAS,MAAM;AAGxB,UAAM,iBAAiB,GAAG,OAAO,eAAe,EAAE,QAAQ,KAAK,CAAC;AAChE,QAAI,mBAAmB,MAAM;AAC3B,cAAQ,MAAM,sCAAiC;AAG/C,UAAI,WAAW,MAAM,GAAG;AACtB,cAAM,aAAa,aAAa,MAAM;AACtC,YAAI,YAAY;AACd,kBAAQ,IAAI,sBAAsB,UAAU,EAAE;AAAA,QAChD;AAAA,MACF;AAGA,cAAQ,IAAI,iCAAiC;AAC7C,SAAG,MAAM;AACT,WAAK,IAAI,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,MAAM,mCAA8B,KAAK;AAGjD,QAAI,WAAW,MAAM,GAAG;AACtB,mBAAa,MAAM;AACnB,cAAQ,IAAI,0BAA0B;AAAA,IACxC;AAEA,SAAK,IAAI,SAAS,MAAM;AAAA,EAC1B;AAGA,KAAG,OAAO,oBAAoB;AAG9B,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAYP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAQP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GASP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAMP;AAGD,QAAM,eAAe,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,GAI/B;AAED,QAAM,eAAe,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,GAI/B;AAED,QAAM,UAAU,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAO1B;AAED,QAAM,kBAAkB,GAAG,QAAQ,wCAAwC;AAC3E,QAAM,kBAAkB,GAAG,QAAQ,wCAAwC;AAE3E,SAAO;AAAA,IACL,MAAM,IAAI,OAAmC;AAC3C,YAAM,cAAc,GAAG,YAAY,MAAM;AACvC,qBAAa;AAAA,UACX,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM,SAAS,IAAI;AAAA,QACrB;AAEA,qBAAa;AAAA,UACX,MAAM;AAAA,UACN,MAAM;AAAA,UACN,KAAK,UAAU,MAAM,IAAI;AAAA,UACzB,KAAK,UAAU,MAAM,QAAQ;AAAA,QAC/B;AAAA,MACF,CAAC;AAED,kBAAY;AAAA,IACd;AAAA,IAEA,MAAM,IAAI,IAAyC;AACjD,YAAM,MAAM,QAAQ,IAAI,EAAE;AAE1B,UAAI,CAAC,KAAK;AACR,eAAO;AAAA,MACT;AAEA,YAAM,IAAI;AAcV,aAAO;AAAA,QACL,IAAI,EAAE;AAAA,QACN,SAAS,EAAE;AAAA,QACX,MAAM,EAAE;AAAA,QACR,WAAW,EAAE;AAAA,QACb,OAAO,EAAE;AAAA,QACT,KAAK,EAAE;AAAA,QACP,OAAO,EAAE;AAAA,QACT,aAAa,EAAE;AAAA,QACf,MAAM,EAAE,OAAO,KAAK,MAAM,EAAE,IAAI,IAAI,CAAC;AAAA,QACrC,UAAU,EAAE,WAAW,KAAK,MAAM,EAAE,QAAQ,IAAI,CAAC;AAAA,QACjD,QAAQ,EAAE,WAAW;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,MAAM,MAAM,SAAqD;AAC/D,UAAI,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AASV,YAAM,SAAoB,CAAC;AAE3B,UAAI,QAAQ,OAAO;AACjB,eAAO;AACP,eAAO,KAAK,QAAQ,KAAK;AAAA,MAC3B;AAEA,UAAI,QAAQ,aAAa,QAAW;AAClC,eAAO;AACP,eAAO,KAAK,QAAQ,QAAQ;AAAA,MAC9B;AAEA,UAAI,QAAQ,aAAa,QAAW;AAClC,eAAO;AACP,eAAO,KAAK,QAAQ,QAAQ;AAAA,MAC9B;AAEA,UAAI,QAAQ,WAAW,QAAW;AAChC,eAAO;AACP,eAAO,KAAK,QAAQ,SAAS,IAAI,CAAC;AAAA,MACpC;AAEA,aAAO;AAEP,UAAI,QAAQ,OAAO;AACjB,eAAO;AACP,eAAO,KAAK,QAAQ,KAAK;AAAA,MAC3B;AAEA,UAAI,QAAQ,QAAQ;AAClB,eAAO;AACP,eAAO,KAAK,QAAQ,MAAM;AAAA,MAC5B;AAEA,YAAM,OAAO,GAAG,QAAQ,GAAG;AAC3B,YAAM,OAAO,KAAK,IAAI,GAAG,MAAM;AAE/B,aAAO,KAAK,IAAI,CAAC,QAAQ;AACvB,cAAM,IAAI;AAcV,eAAO;AAAA,UACL,IAAI,EAAE;AAAA,UACN,SAAS,EAAE;AAAA,UACX,MAAM,EAAE;AAAA,UACR,WAAW,EAAE;AAAA,UACb,OAAO,EAAE;AAAA,UACT,KAAK,EAAE;AAAA,UACP,OAAO,EAAE;AAAA,UACT,aAAa,EAAE;AAAA,UACf,MAAM,EAAE,OAAO,KAAK,MAAM,EAAE,IAAI,IAAI,CAAC;AAAA,UACrC,UAAU,EAAE,WAAW,KAAK,MAAM,EAAE,QAAQ,IAAI,CAAC;AAAA,UACjD,QAAQ,EAAE,WAAW;AAAA,QACvB;AAAA,MACF,CAAC;AAAA,IACH;AAAA,IAEA,MAAM,OAAO,IAA2B;AACtC,YAAM,cAAc,GAAG,YAAY,MAAM;AACvC,wBAAgB,IAAI,EAAE;AACtB,wBAAgB,IAAI,EAAE;AAAA,MACxB,CAAC;AAED,kBAAY;AAAA,IACd;AAAA,IAEA,MAAM,OAA0B;AAC9B,YAAM,OAAO,GAAG,QAAQ,8BAA8B;AACtD,YAAM,OAAO,KAAK,IAAI;AACtB,aAAO,KAAK,IAAI,CAAC,MAAM,EAAE,EAAE;AAAA,IAC7B;AAAA,IAEA,MAAM,UAA2B;AAC/B,YAAM,SAAS,GAAG,QAAQ,6CAA6C,EAAE,IAAI;AAK7E,SAAG,KAAK,0CAA0C;AAElD,SAAG,KAAK,QAAQ;AAEhB,YAAM,QAAQ,GAAG,QAAQ,6CAA6C,EAAE,IAAI;AAI5E,aAAO,OAAO,QAAQ,MAAM;AAAA,IAC9B;AAAA,IAEA,MAAM,QAAuB;AAC3B,SAAG,MAAM;AAAA,IACX;AAAA,IAEA,MAAM,mBAAmB,OAAqD;AAC5E,YAAM,OAAO,GAAG,QAAQ;AAAA;AAAA;AAAA,OAGvB;AAED,WAAK;AAAA,QACH,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,MACR;AAAA,IACF;AAAA,IAEA,MAAM,uBAAqD;AACzD,YAAM,OAAO,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAIvB;AAED,YAAM,OAAO,KAAK,IAAI;AACtB,aAAO;AAAA,IACT;AAAA,IAEA,MAAM,yBAAwC;AAC5C,SAAG,KAAK,gCAAgC;AAAA,IAC1C;AAAA,EACF;AACF;;;AC3WO,SAAS,wBAAyC;AACvD,QAAM,SAAS,mBAAmB,EAAE,YAAY,IAAI,gBAAgB,KAAK,CAAC;AAE1E,WAAS,YAAY,SAA2C;AAC9D,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM,gBAAgB,QAAQ;AAG9B,UAAM,MAAM,KAAK,IAAI;AACrB,UAAM,aAAa,QAAQ,OAAO,CAAC,UAAU;AAC3C,YAAM,gBAAgB,MAAM,MAAM,aAAa;AAC/C,YAAM,QAAQ,OAAO,eAAe,cAAc,MAAM,GAAG;AAC3D,aAAO,QAAQ;AAAA,IACjB,CAAC;AAED,UAAM,iBAAiB,gBAAgB,WAAW;AAGlD,UAAM,kBAAkB,eAAe,UAAU;AACjD,UAAM,SAAS,gBAAgB,eAAe;AAG9C,UAAM,eAAe,IAAI,IAAI,gBAAgB,QAAQ,CAAC,MAAM,EAAE,QAAQ,IAAI,CAAC,MAAM,EAAE,EAAE,CAAC,CAAC;AACvF,UAAM,gBAAgB,WAAW,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,EAAE,CAAC;AAGtE,UAAM,OAAO,CAAC,GAAG,QAAQ,GAAG,aAAa;AACzC,UAAM,UAAU,QAAQ,OAAO,CAAC,MAAM,CAAC,KAAK,KAAK,CAAC,MAAM,EAAE,OAAO,EAAE,EAAE,CAAC;AAEtE,UAAM,oBAAoB,gBAAgB,OAAO,CAAC,KAAK,MAAM,OAAO,EAAE,QAAQ,SAAS,IAAI,CAAC;AAE5F,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,eAAe;AAAA,MACf,cAAc,KAAK;AAAA,MACnB;AAAA,MACA;AAAA,MACA,kBAAkB,gBAAgB,IAAI,KAAK,SAAS,gBAAgB;AAAA,MACpE,YAAY,KAAK,IAAI,IAAI;AAAA,IAC3B;AAAA,EACF;AAEA,WAAS,eAAe,SAA0C;AAChE,UAAM,SAA2B,CAAC;AAClC,UAAM,YAAY,oBAAI,IAAY;AAGlC,aAAS,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AACvC,YAAM,QAAQ,QAAQ,CAAC;AACvB,UAAI,CAAC,SAAS,UAAU,IAAI,MAAM,EAAE,EAAG;AAEvC,YAAM,aAAa,QAAQ,OAAO,CAAC,GAAG,QAAQ,QAAQ,KAAK,EAAE,SAAS,MAAM,IAAI;AAEhF,UAAI,WAAW,SAAS,GAAG;AACzB,cAAM,QAAwB;AAAA,UAC5B,SAAS,CAAC,OAAO,GAAG,UAAU;AAAA,UAC9B,YAAY;AAAA;AAAA,QACd;AACA,eAAO,KAAK,KAAK;AAGjB,kBAAU,IAAI,MAAM,EAAE;AACtB,mBAAW,OAAO,YAAY;AAC5B,oBAAU,IAAI,IAAI,EAAE;AAAA,QACtB;AAAA,MACF;AAAA,IACF;AAGA,aAAS,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AACvC,YAAM,SAAS,QAAQ,CAAC;AACxB,UAAI,CAAC,UAAU,UAAU,IAAI,OAAO,EAAE,EAAG;AAEzC,eAAS,IAAI,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AAC3C,cAAM,SAAS,QAAQ,CAAC;AACxB,YAAI,CAAC,UAAU,UAAU,IAAI,OAAO,EAAE,EAAG;AAEzC,cAAM,aAAa,iBAAiB,OAAO,SAAS,OAAO,OAAO;AAElE,YAAI,cAAc,MAAM;AACtB,gBAAM,QAAwB;AAAA,YAC5B,SAAS,CAAC,QAAQ,MAAM;AAAA,YACxB;AAAA,UACF;AACA,iBAAO,KAAK,KAAK;AAEjB,oBAAU,IAAI,OAAO,EAAE;AACvB,oBAAU,IAAI,OAAO,EAAE;AACvB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,gBAAgB,QAAyC;AAChE,UAAM,SAAwB,CAAC;AAE/B,eAAW,SAAS,QAAQ;AAE1B,YAAM,SAAS,CAAC,GAAG,MAAM,OAAO,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAClE,YAAM,OAAO,OAAO,CAAC;AACrB,UAAI,CAAC,KAAM;AAGX,YAAM,mBAAmB,MAAM,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,aAAa,CAAC;AAGhF,YAAM,UAAU,IAAI,IAAI,MAAM,QAAQ,QAAQ,CAAC,MAAM,EAAE,IAAI,CAAC;AAE5D,aAAO,KAAK;AAAA,QACV,GAAG;AAAA,QACH,aAAa;AAAA,QACb,MAAM,MAAM,KAAK,OAAO;AAAA,MAC1B,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAQA,WAAS,iBAAiB,OAAe,OAAuB;AAC9D,UAAM,SAAS,SAAS,KAAK;AAC7B,UAAM,SAAS,SAAS,KAAK;AAG7B,UAAM,QAAQ,oBAAI,IAAI,CAAC,GAAG,QAAQ,GAAG,MAAM,CAAC;AAG5C,UAAM,OAA+B,CAAC;AACtC,UAAM,OAA+B,CAAC;AAEtC,eAAW,QAAQ,OAAO;AACxB,WAAK,IAAI,IAAI,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAC9C,WAAK,IAAI,IAAI,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAAA,IAChD;AAGA,QAAI,aAAa;AACjB,QAAI,OAAO;AACX,QAAI,OAAO;AAEX,eAAW,QAAQ,OAAO;AACxB,YAAM,SAAS,KAAK,IAAI,KAAK;AAC7B,YAAM,SAAS,KAAK,IAAI,KAAK;AAC7B,oBAAc,SAAS;AACvB,cAAQ,SAAS;AACjB,cAAQ,SAAS;AAAA,IACnB;AAEA,WAAO,KAAK,KAAK,IAAI;AACrB,WAAO,KAAK,KAAK,IAAI;AAErB,QAAI,SAAS,KAAK,SAAS,EAAG,QAAO;AAErC,WAAO,cAAc,OAAO;AAAA,EAC9B;AAEA,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC9MA,SAAS,YAAY;AACrB,SAAS,cAAAC,aAAY,WAAW,cAAc,YAAY,qBAAqB;AAC/E,SAAS,SAAS,YAAY;AAC9B,SAAS,qBAAqB;AAyCvB,SAAS,sBAAqC;AAInD,WAAS,gBAAgB,SAAqD;AAC5E,QAAI,CAACC,YAAW,OAAO,GAAG;AACxB,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAEA,QAAI;AACF,YAAM,SAAS,aAAa,SAAS,OAAO,EAAE,KAAK;AACnD,YAAM,MAAM,OAAO,SAAS,QAAQ,EAAE;AAEtC,UAAI,OAAO,MAAM,GAAG,GAAG;AACrB,eAAO,EAAE,SAAS,MAAM;AAAA,MAC1B;AAGA,UAAI;AACF,gBAAQ,KAAK,KAAK,CAAC;AACnB,eAAO,EAAE,SAAS,MAAM,IAAI;AAAA,MAC9B,QAAQ;AAEN,mBAAW,OAAO;AAClB,eAAO,EAAE,SAAS,MAAM;AAAA,MAC1B;AAAA,IACF,QAAQ;AACN,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAKA,WAAS,aAAa,SAAiB,KAAmB;AAExD,UAAM,MAAM,QAAQ,OAAO;AAC3B,QAAI,CAACA,YAAW,GAAG,GAAG;AACpB,gBAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,IACpC;AAEA,kBAAc,SAAS,OAAO,GAAG,GAAG,OAAO;AAAA,EAC7C;AAKA,WAAS,cAAc,SAAuB;AAC5C,QAAIA,YAAW,OAAO,GAAG;AACvB,iBAAW,OAAO;AAAA,IACpB;AAAA,EACF;AAEA,iBAAe,MAAM,QAAiD;AACpE,UAAM,EAAE,SAAS,QAAQ,IAAI,OAAO;AAGpC,UAAMC,UAAS,gBAAgB,OAAO;AACtC,QAAIA,QAAO,SAAS;AAClB,aAAO;AAAA,QACL,SAAS;AAAA,QACT,KAAKA,QAAO;AAAA,QACZ,SAAS,+BAA+BA,QAAO,GAAG;AAAA,QAClD,OAAO;AAAA,MACT;AAAA,IACF;AAEA,QAAI;AAEF,YAAMC,cAAa,cAAc,YAAY,GAAG;AAChD,YAAMC,aAAY,QAAQD,WAAU;AACpC,YAAM,aAAa,KAAKC,YAAW,UAAU;AAE7C,YAAM,QAAQ,KAAK,YAAY,CAAC,GAAG;AAAA,QACjC,UAAU;AAAA,QACV,OAAO;AAAA,QACP,KAAK;AAAA,UACH,GAAG,QAAQ;AAAA,UACX,cAAc,KAAK,UAAU,MAAM;AAAA,UACnC,gBAAgB;AAAA,UAChB,gBAAgB;AAAA,QAClB;AAAA,MACF,CAAC;AAGD,YAAM,MAAM;AAGZ,UAAI,MAAM,KAAK;AACb,qBAAa,SAAS,MAAM,GAAG;AAE/B,eAAO;AAAA,UACL,SAAS;AAAA,UACT,KAAK,MAAM;AAAA,UACX,SAAS,uBAAuB,MAAM,GAAG;AAAA,QAC3C;AAAA,MACF;AAEA,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO;AAAA,MACT;AAAA,IACF,SAAS,OAAO;AACd,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D;AAAA,IACF;AAAA,EACF;AAEA,iBAAe,KAAK,QAAgD;AAClE,UAAM,EAAE,QAAQ,IAAI,OAAO;AAE3B,UAAMF,UAAS,gBAAgB,OAAO;AAEtC,QAAI,CAACA,QAAO,WAAW,CAACA,QAAO,KAAK;AAClC,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,MACX;AAAA,IACF;AAEA,QAAI;AAEF,cAAQ,KAAKA,QAAO,KAAK,SAAS;AAGlC,YAAM,UAAU;AAChB,YAAM,WAAW;AACjB,UAAI,SAAS;AAEb,aAAO,SAAS,SAAS;AACvB,YAAI;AACF,kBAAQ,KAAKA,QAAO,KAAK,CAAC;AAE1B,gBAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAC5D,oBAAU;AAAA,QACZ,QAAQ;AAEN,wBAAc,OAAO;AACrB,iBAAO;AAAA,YACL,SAAS;AAAA,YACT,SAAS,uBAAuBA,QAAO,GAAG;AAAA,UAC5C;AAAA,QACF;AAAA,MACF;AAGA,UAAI;AACF,gBAAQ,KAAKA,QAAO,KAAK,SAAS;AAClC,sBAAc,OAAO;AACrB,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,4BAA4BA,QAAO,GAAG;AAAA,QACjD;AAAA,MACF,QAAQ;AACN,sBAAc,OAAO;AACrB,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,uBAAuBA,QAAO,GAAG;AAAA,QAC5C;AAAA,MACF;AAAA,IACF,SAAS,OAAO;AACd,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D;AAAA,IACF;AAAA,EACF;AAEA,iBAAe,OAAO,QAAkD;AACtE,UAAM,EAAE,QAAQ,IAAI,OAAO;AAE3B,UAAM,eAAe,gBAAgB,OAAO;AAE5C,QAAI,CAAC,aAAa,WAAW,CAAC,aAAa,KAAK;AAC9C,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,MACX;AAAA,IACF;AAGA,UAAM,UAAU,WAAW,EAAE,YAAY;AAEzC,WAAO;AAAA,MACL,SAAS;AAAA,MACT,KAAK,aAAa;AAAA,MAClB,QAAQ,QAAQ,OAAO;AAAA,MACvB,iBAAiB,QAAQ,OAAO;AAAA,MAChC,aAAa,QAAQ,aAAa;AAAA,MAClC,SAAS,uBAAuB,aAAa,GAAG;AAAA,IAClD;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACvPA,SAAS,gBAAAG,eAAc,gBAAgB;AAoDhC,SAAS,oBAAiC;AAE/C,QAAM,YAAY,oBAAI,IAA0B;AAEhD,WAAS,aAAa,UAA4B;AAChD,QAAI;AAEF,YAAM,QAAQ,SAAS,QAAQ;AAC/B,YAAM,cAAc,MAAM;AAC1B,YAAM,kBAAkB,MAAM;AAG9B,UAAI,MAAM,UAAU,IAAI,QAAQ;AAEhC,UAAI,CAAC,KAAK;AAER,cAAM;AAAA,UACJ,MAAM;AAAA,UACN,UAAU;AAAA,UACV,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,QACZ;AACA,kBAAU,IAAI,UAAU,GAAG;AAAA,MAC7B;AAGA,UAAI,cAAc,IAAI,YAAY,gBAAgB,IAAI,UAAU;AAE9D,YAAI,cAAc,IAAI,UAAU;AAE9B,cAAI,WAAW;AACf,cAAI,cAAc;AAAA,QACpB;AACA,eAAO,CAAC;AAAA,MACV;AAGA,YAAM,SAAS,OAAO,MAAM,cAAc,IAAI,QAAQ;AACtD,YAAM,KAAKA,cAAa,QAAQ;AAChC,SAAG,KAAK,QAAQ,GAAG,IAAI,UAAU,WAAW;AAG5C,YAAM,cAAc,IAAI,cAAc,OAAO,SAAS,OAAO,GAAG,MAAM,IAAI;AAG1E,YAAM,cAAc,WAAW,IAAI,KAAK;AAGxC,UAAI,WAAW;AACf,UAAI,cAAc;AAClB,UAAI,eAAe;AACnB,UAAI,WAAW;AAGf,aAAO,WAAW,OAAO,CAAC,SAAS,KAAK,KAAK,EAAE,SAAS,CAAC;AAAA,IAC3D,SAAS,QAAQ;AAGf,aAAO,CAAC;AAAA,IACV;AAAA,EACF;AAEA,WAAS,YAAY,UAAuC;AAC1D,WAAO,UAAU,IAAI,QAAQ,KAAK;AAAA,EACpC;AAEA,WAAS,cAAc,UAAwB;AAC7C,cAAU,OAAO,QAAQ;AAAA,EAC3B;AAEA,WAAS,WAAiB;AACxB,cAAU,MAAM;AAAA,EAClB;AAEA,WAAS,kBAA4B;AACnC,WAAO,MAAM,KAAK,UAAU,KAAK,CAAC;AAAA,EACpC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5IA,SAAyB,aAAa,YAAAC,WAAU,aAAa;AAC7D,SAAS,eAAe;AACxB,SAAS,WAAAC,UAAS,QAAAC,aAAY;AAoEvB,SAAS,qBAAqB,QAA8C;AACjF,QAAM,EAAE,QAAQ,aAAa,YAAY,QAAQ,IAAI;AACrD,QAAM,EAAE,UAAU,OAAO,OAAO,IAAI;AAGpC,QAAM,YAAY,oBAAI,IAA6B;AACnD,QAAM,cAAc,kBAAkB;AAGtC,QAAM,WAAwB,CAAC;AAG/B,QAAM,iBAAiB,oBAAI,IAA4B;AAKvD,WAAS,iBAAyB;AAChC,WAAOC,MAAK,QAAQ,GAAG,WAAW,UAAU;AAAA,EAC9C;AAMA,WAAS,aAAa,UAA0B;AAC9C,UAAM,WAAW,SAAS,MAAM,OAAO,EAAE,IAAI,KAAK;AAClD,WAAO,SAAS,QAAQ,YAAY,EAAE;AAAA,EACxC;AAKA,WAAS,YAAY,WAAoC;AACvD,QAAI,WAAW,UAAU,IAAI,SAAS;AAEtC,QAAI,CAAC,UAAU;AACb,iBAAW,sBAAsB;AAAA,QAC/B,aAAa,SAAS;AAAA,QACtB;AAAA,QACA;AAAA,QACA,YAAY,SAAS;AAAA,QACrB,0BAA0B;AAAA;AAAA,MAC5B,CAAC;AACD,gBAAU,IAAI,WAAW,QAAQ;AAAA,IACnC;AAEA,WAAO;AAAA,EACT;AAKA,WAAS,iBAAiB,UAAwB;AAEhD,UAAM,gBAAgB,eAAe,IAAI,QAAQ;AACjD,QAAI,eAAe;AACjB,mBAAa,aAAa;AAAA,IAC5B;AAGA,UAAM,QAAQ,WAAW,MAAM;AAC7B,UAAI;AAEF,cAAM,WAAW,YAAY,aAAa,QAAQ;AAElD,YAAI,SAAS,WAAW,EAAG;AAG3B,cAAM,UAAU,SAAS,KAAK,IAAI;AAClC,cAAM,YAAY,aAAa,QAAQ;AACvC,cAAM,WAAW,YAAY,SAAS;AAGtC,iBAAS,OAAO,SAAS,EAAE,WAAW,SAAS,CAAC;AAGhD,cAAM,QAAQ,SAAS,SAAS;AAChC,YAAI,MAAM,iBAAiB,SAAS,uBAAuB;AAEzD,qBAAW,EAAE,aAAa;AAAA,YACxB,iBAAiB,UAAU;AAAA,YAC3B,aAAa,QAAQ,YAAY,EAAE;AAAA,UACrC,CAAC;AAGD,cAAI,YAAY;AACd,kBAAM,eAAe,oBAAoB,WAAW,QAAQ;AAC5D,uBAAW,WAAW,YAAY;AAAA,UACpC;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,YAAI,SAAS;AACX,kBAAQ,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC,CAAC;AAAA,QACnE;AAAA,MACF,UAAE;AACA,uBAAe,OAAO,QAAQ;AAAA,MAChC;AAAA,IACF,GAAG,SAAS,UAAU;AAEtB,mBAAe,IAAI,UAAU,KAAK;AAAA,EACpC;AAKA,WAAS,eAAe,KAAuB;AAC7C,UAAM,QAAkB,CAAC;AAEzB,QAAI;AACF,YAAM,UAAU,YAAY,GAAG;AAE/B,iBAAW,SAAS,SAAS;AAC3B,cAAM,WAAWA,MAAK,KAAK,KAAK;AAChC,cAAM,OAAOC,UAAS,QAAQ;AAE9B,YAAI,KAAK,YAAY,GAAG;AAEtB,gBAAM,KAAK,GAAG,eAAe,QAAQ,CAAC;AAAA,QACxC,WAAW,MAAM,SAAS,QAAQ,GAAG;AAEnC,gBAAM,UAAU,SAAS,cAAc,KAAK,CAAC,YAAY;AAEvD,kBAAM,QAAQ,IAAI;AAAA,cAChB,QAAQ,QAAQ,SAAS,IAAI,EAAE,QAAQ,OAAO,WAAW,EAAE,QAAQ,OAAO,KAAK;AAAA,YACjF;AACA,mBAAO,MAAM,KAAK,QAAQ;AAAA,UAC5B,CAAC;AAED,cAAI,SAAS;AACX,kBAAM,KAAK,QAAQ;AAAA,UACrB;AAAA,QACF;AAAA,MACF;AAAA,IACF,SAAS,QAAQ;AAAA,IAEjB;AAEA,WAAO;AAAA,EACT;AAKA,WAAS,oBAAoB,WAAmB,UAAyC;AACvF,UAAM,QAAQ,SAAS,SAAS;AAChC,UAAM,UAAU,SAAS,WAAW;AACpC,UAAM,cAAc,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAEjF,WAAO;AAAA,MACL;AAAA,MACA,aAAa,MAAM;AAAA,MACnB,iBAAiB,MAAM;AAAA,MACvB,WAAW,cAAc,KAAK,cAAc,MAAM,iBAAiB,cAAc;AAAA,MACjF,YAAY,MAAM;AAAA,MAClB,mBAAmB,MAAM;AAAA,IAC3B;AAAA,EACF;AAEA,iBAAe,QAAuB;AACpC,UAAM,cAAc,eAAe;AAGnC,UAAM,aAAa,eAAe,WAAW;AAG7C,UAAM,cAAc,oBAAI,IAAY;AAEpC,eAAW,QAAQ,YAAY;AAC7B,YAAM,MAAMC,SAAQ,IAAI;AAExB,UAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAM,UAAU,MAAM,KAAK,EAAE,WAAW,MAAM,GAAG,CAAC,YAAY,aAAa;AACzE,cAAI,UAAU,SAAS,QAAQ,GAAG;AAChC,kBAAM,WAAWF,MAAK,KAAK,QAAQ;AACnC,6BAAiB,QAAQ;AAAA,UAC3B;AAAA,QACF,CAAC;AAED,iBAAS,KAAK,OAAO;AACrB,oBAAY,IAAI,GAAG;AAAA,MACrB;AAAA,IACF;AAGA,UAAM,kBAAkB,MAAM,aAAa,EAAE,WAAW,KAAK,GAAG,CAAC,YAAY,aAAa;AACxF,UAAI,UAAU,SAAS,QAAQ,GAAG;AAChC,cAAM,WAAWA,MAAK,aAAa,QAAQ;AAC3C,yBAAiB,QAAQ;AAAA,MAC3B;AAAA,IACF,CAAC;AAED,aAAS,KAAK,eAAe;AAG7B,eAAW,EAAE,aAAa;AAAA,MACxB,WAAW,KAAK,IAAI;AAAA,MACpB,iBAAiB,WAAW;AAAA,MAC5B,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAAA,EACH;AAEA,WAAS,OAAa;AAEpB,eAAW,WAAW,UAAU;AAC9B,cAAQ,MAAM;AAAA,IAChB;AACA,aAAS,SAAS;AAGlB,eAAW,SAAS,eAAe,OAAO,GAAG;AAC3C,mBAAa,KAAK;AAAA,IACpB;AACA,mBAAe,MAAM;AAGrB,cAAU,MAAM;AAGhB,gBAAY,SAAS;AAAA,EACvB;AAEA,WAAS,WAA2B;AAClC,UAAM,QAAwB,CAAC;AAE/B,eAAW,CAAC,WAAW,QAAQ,KAAK,UAAU,QAAQ,GAAG;AACvD,YAAM,KAAK,oBAAoB,WAAW,QAAQ,CAAC;AAAA,IACrD;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,gBAAgB,WAAwC;AAC/D,UAAM,WAAW,UAAU,IAAI,SAAS;AACxC,QAAI,CAAC,SAAU,QAAO;AAEtB,WAAO,oBAAoB,WAAW,QAAQ;AAAA,EAChD;AAEA,WAAS,gBAAgB,WAAyB;AAChD,UAAM,WAAW,UAAU,IAAI,SAAS;AACxC,QAAI,CAAC,SAAU;AAGf,UAAM,UAAU,SAAS,WAAW;AACpC,aAAS,MAAM;AACf,aAAS,OAAO,QAAQ,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,MAAM,CAAC;AAG1D,QAAI,YAAY;AACd,YAAM,QAAQ,oBAAoB,WAAW,QAAQ;AACrD,iBAAW,WAAW,KAAK;AAAA,IAC7B;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACxUA,SAAS,iBAAiB;AAC1B,SAAS,SAAS;;;AC6FX,IAAM,iBAA8B;AAAA,EACzC,SAAS;AAAA,IACP,WAAW;AAAA,IACX,gBAAgB;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,YAAY;AAAA,IACZ,gBAAgB;AAAA,EAClB;AAAA,EACA,QAAQ;AAAA,IACN,iBAAiB;AAAA,IACjB,gBAAgB;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,EACP,IAAI;AAAA,IACF,QAAQ;AAAA,IACR,QAAQ;AAAA,IACR,SAAS;AAAA,EACX;AAAA,EACA,iBAAiB;AAAA,EACjB,UAAU;AAAA,IACR,aAAa;AAAA,IACb,uBAAuB;AAAA,IACvB,eAAe,CAAC,YAAY;AAAA,IAC5B,SAAS;AAAA,IACT,SAAS;AAAA,IACT,YAAY;AAAA,IACZ,aAAa;AAAA,IACb,YAAY;AAAA,IACZ,uBAAuB;AAAA,EACzB;AACF;;;ADrGO,SAAS,qBAAqB,SAA2C;AAC9E,QAAM,EAAE,QAAQ,SAAS,eAAe,IAAI;AAE5C,QAAM,SAAS,IAAI,UAAU;AAAA,IAC3B,MAAM;AAAA,IACN,SAAS;AAAA,EACX,CAAC;AAED,uBAAqB,QAAQ,QAAQ,MAAM;AAC3C,oBAAkB,QAAQ,MAAM;AAChC,0BAAwB,QAAQ,MAAM;AAEtC,SAAO;AACT;AAQA,SAAS,qBAAqB,QAAmB,QAAkB,QAA2B;AAC5F,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,MAGF,aAAa;AAAA,QACX,SAAS,EAAE,OAAO,EAAE,SAAS,8BAA8B;AAAA,QAC3D,QAAQ,EACL,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,qDAAqD;AAAA,QACjE,SAAS,EACN,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,oDAAoD;AAAA,QAChE,WAAW,EACR,OAAO,EACP,IAAI,CAAC,EACL,IAAI,GAAG,EACP,SAAS,EACT,SAAS,oDAAoD;AAAA,MAClE;AAAA,IACF;AAAA,IACA,OAAO,EAAE,SAAS,QAAQ,SAAS,UAAU,MAAM;AACjD,UAAI;AACF,cAAM,kBAAkB,YACpB,EAAE,GAAG,QAAQ,SAAS,EAAE,GAAG,OAAO,SAAS,UAAU,EAAE,IACvD;AAEJ,cAAM,UAAU,qBAAqB,QAAQ,eAAe;AAC5D,cAAM,SAAS,MAAM,QAAQ,SAAS,SAAS;AAAA,UAC7C;AAAA,UACA;AAAA,UACA;AAAA,QACF,CAAC;AAED,cAAM,WAAW;AAAA,UACf,kBAAkB,OAAO;AAAA,UACzB,cAAc,OAAO;AAAA,UACrB,aAAa,OAAO;AAAA,UACpB,WAAW,IAAI,OAAO,YAAY,KAAK,QAAQ,CAAC,CAAC;AAAA,UACjD,kBAAkB,OAAO;AAAA,UACzB,aAAa,OAAO;AAAA,UACpB,YAAY,OAAO;AAAA,UACnB,mBAAmB,OAAO;AAAA,UAC1B,GAAI,WAAW,OAAO,UAAU,EAAE,SAAS,OAAO,QAAQ,IAAI,CAAC;AAAA,QACjE;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAQA,SAAS,kBAAkB,QAAmB,QAAwB;AACpE,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,MAEF,aAAa;AAAA,QACX,OAAO,EACJ,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,4CAA4C;AAAA,MAC1D;AAAA,IACF;AAAA,IACA,OAAO,EAAE,MAAM,MAAM;AACnB,UAAI;AACF,YAAI,OAAO;AACT,gBAAM,OAAO,uBAAuB;AACpC,iBAAO;AAAA,YACL,SAAS;AAAA,cACP;AAAA,gBACE,MAAM;AAAA,gBACN,MAAM,KAAK;AAAA,kBACT;AAAA,oBACE,SAAS;AAAA,oBACT,eAAe;AAAA,oBACf,kBAAkB;AAAA,oBAClB,kBAAkB;AAAA,kBACpB;AAAA,kBACA;AAAA,kBACA;AAAA,gBACF;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAEA,cAAM,QAAQ,MAAM,OAAO,qBAAqB;AAChD,cAAM,gBAAgB,MAAM;AAE5B,cAAM,mBAAmB,MAAM;AAAA,UAC7B,CAAC,KAAK,MAAM,OAAO,EAAE,gBAAgB,EAAE;AAAA,UACvC;AAAA,QACF;AAEA,cAAM,mBACJ,gBAAgB,IACZ,MAAM,OAAO,CAAC,KAAK,MAAM;AACvB,gBAAM,YACJ,EAAE,gBAAgB,KAAK,EAAE,gBAAgB,EAAE,gBAAgB,EAAE,gBAAgB;AAC/E,iBAAO,MAAM;AAAA,QACf,GAAG,CAAC,IAAI,gBACR;AAEN,cAAM,sBAAsB,MAAM,MAAM,GAAG,EAAE,EAAE,IAAI,CAAC,OAAO;AAAA,UACzD,WAAW,IAAI,KAAK,EAAE,SAAS,EAAE,YAAY;AAAA,UAC7C,cAAc,EAAE;AAAA,UAChB,aAAa,EAAE;AAAA,UACf,eAAe,EAAE;AAAA,UACjB,YAAY,EAAE;AAAA,UACd,WAAW,KACP,EAAE,gBAAgB,EAAE,gBAAgB,KAAK,IAAI,EAAE,eAAe,CAAC,IAAK,KACtE,QAAQ,CAAC,CAAC;AAAA,QACd,EAAE;AAEF,cAAM,WAAW;AAAA,UACf;AAAA,UACA;AAAA,UACA,kBAAkB,IAAI,mBAAmB,KAAK,QAAQ,CAAC,CAAC;AAAA,UACxD;AAAA,QACF;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAQA,SAAS,wBAAwB,QAAmB,QAAwB;AAC1E,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,IAGJ;AAAA,IACA,YAAY;AACV,UAAI;AACF,cAAM,SAAS,MAAM,OAAO,KAAK;AACjC,cAAM,aAAa,MAAM,QAAQ;AAAA,UAC/B,OAAO,IAAI,OAAO,OAAO;AACvB,kBAAM,QAAQ,MAAM,OAAO,IAAI,EAAE;AACjC,mBAAO;AAAA,UACT,CAAC;AAAA,QACH;AAEA,cAAM,UAAU,WAAW,OAAO,CAAC,MAAM,MAAM,IAAI;AAEnD,cAAM,aAAa,sBAAsB;AACzC,cAAM,SAAS,WAAW,YAAY,OAAO;AAG7C,mBAAW,WAAW,OAAO,SAAS;AACpC,gBAAM,OAAO,OAAO,QAAQ,EAAE;AAAA,QAChC;AAEA,mBAAW,QAAQ,OAAO,MAAM;AAC9B,gBAAM,OAAO,IAAI,IAAI;AAAA,QACvB;AAGA,cAAM,OAAO,QAAQ;AAErB,cAAM,WAAW;AAAA,UACf,eAAe,OAAO;AAAA,UACtB,cAAc,OAAO;AAAA,UACrB,gBAAgB,OAAO;AAAA,UACvB,mBAAmB,OAAO;AAAA,UAC1B,kBAAkB,IAAI,OAAO,mBAAmB,KAAK,QAAQ,CAAC,CAAC;AAAA,UAC/D,YAAY,OAAO;AAAA,UACnB,iBAAiB;AAAA,QACnB;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;;;AEnSO,SAAS,aAAa,UAAU,OAAe;AACpD,SAAO;AAAA,IACL,MAAM,YAAoB,MAAuB;AAC/C,UAAI,SAAS;AACX,gBAAQ,MAAM,WAAW,OAAO,IAAI,GAAG,IAAI;AAAA,MAC7C;AAAA,IACF;AAAA,IACA,KAAK,YAAoB,MAAuB;AAC9C,cAAQ,KAAK,UAAU,OAAO,IAAI,GAAG,IAAI;AAAA,IAC3C;AAAA,IACA,KAAK,YAAoB,MAAuB;AAC9C,cAAQ,KAAK,UAAU,OAAO,IAAI,GAAG,IAAI;AAAA,IAC3C;AAAA,IACA,MAAM,YAAoB,MAAuB;AAC/C,cAAQ,MAAM,WAAW,OAAO,IAAI,GAAG,IAAI;AAAA,IAC7C;AAAA,EACF;AACF;","names":["randomUUID","randomUUID","randomUUID","randomUUID","allEntries","existsSync","existsSync","status","__filename","__dirname","readFileSync","statSync","dirname","join","join","statSync","dirname"]}
1
+ {"version":3,"sources":["../src/core/btsp-embedder.ts","../src/utils/hash.ts","../src/core/confidence-states.ts","../src/core/engram-scorer.ts","../src/utils/tokenizer.ts","../src/core/sparse-pruner.ts","../src/utils/context-parser.ts","../src/adapters/claude-code.ts","../src/adapters/generic.ts","../src/core/budget-pruner.ts","../src/core/metrics.ts","../src/core/incremental-optimizer.ts","../src/core/context-pipeline.ts","../src/core/kv-memory.ts","../src/core/sleep-compressor.ts","../src/daemon/daemon-process.ts","../src/daemon/file-tracker.ts","../src/daemon/session-watcher.ts","../src/mcp/server.ts","../src/types/config.ts","../src/utils/logger.ts"],"sourcesContent":["/**\n * BTSP Embedder - Implements behavioral timescale synaptic plasticity\n *\n * Neuroscience: One-shot learning from critical events (errors, conflicts).\n * Application: Detect high-importance patterns and mark for permanent retention.\n */\n\nimport { randomUUID } from 'node:crypto';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from '../utils/hash.js';\n\nexport interface BTSPEmbedder {\n /**\n * Detect if content contains BTSP patterns (errors, stack traces, conflicts, git diffs)\n * @param content - Content to analyze\n * @returns True if BTSP pattern detected\n */\n detectBTSP(content: string): boolean;\n\n /**\n * Create a new memory entry marked as BTSP (one-shot learned)\n * @param content - Entry content\n * @param tags - Optional tags\n * @param metadata - Optional metadata\n * @returns BTSP-marked memory entry\n */\n createBTSPEntry(\n content: string,\n tags?: string[],\n metadata?: Record<string, unknown>,\n ): MemoryEntry;\n}\n\n/**\n * Create a BTSP embedder instance\n * @returns BTSPEmbedder instance\n */\nexport function createBTSPEmbedder(): BTSPEmbedder {\n // Patterns that indicate critical events\n const BTSP_PATTERNS = [\n // Error patterns\n /\\b(error|exception|failure|fatal|critical|panic)\\b/i,\n /\\b(TypeError|ReferenceError|SyntaxError|RangeError|URIError)\\b/,\n /\\bENOENT|EACCES|ECONNREFUSED|ETIMEDOUT\\b/,\n\n // Stack trace patterns\n /^\\s+at\\s+.*\\(.*:\\d+:\\d+\\)/m, // JavaScript stack trace\n /^\\s+at\\s+.*\\.[a-zA-Z]+:\\d+/m, // Python/Ruby stack trace\n\n // Git diff new files\n /^new file mode \\d+$/m,\n /^--- \\/dev\\/null$/m,\n\n // Merge conflict markers\n /^<<<<<<< /m,\n /^=======/m,\n /^>>>>>>> /m,\n ];\n\n function detectBTSP(content: string): boolean {\n return BTSP_PATTERNS.some((pattern) => pattern.test(content));\n }\n\n function createBTSPEntry(\n content: string,\n tags: string[] = [],\n metadata: Record<string, unknown> = {},\n ): MemoryEntry {\n return {\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: Date.now(),\n score: 1.0, // Maximum initial score\n ttl: 365 * 24 * 3600, // 1 year in seconds (long retention)\n state: 'active', // Always active\n accessCount: 0,\n tags: [...tags, 'btsp'],\n metadata,\n isBTSP: true,\n };\n }\n\n return {\n detectBTSP,\n createBTSPEntry,\n };\n}\n","/**\n * Content hashing utilities.\n * Uses SHA-256 for deduplication.\n */\n\nimport { createHash } from 'node:crypto';\n\n/**\n * Generate SHA-256 hash of content for deduplication.\n *\n * @param content - Content to hash\n * @returns 64-character hex string (SHA-256)\n *\n * @example\n * ```typescript\n * const hash = hashContent('Hello world');\n * console.log(hash.length); // 64\n * ```\n */\nexport function hashContent(content: string): string {\n return createHash('sha256').update(content, 'utf8').digest('hex');\n}\n","/**\n * Confidence States - Implements multi-state synapses\n *\n * Neuroscience: Synapses exist in three states: silent, ready (potentiated), active.\n * Application: Classify memory entries by score into silent/ready/active states.\n */\n\nimport type { ConfidenceState, MemoryEntry, StateDistribution } from '../types/memory.js';\n\nexport interface ConfidenceStatesConfig {\n /** Score threshold for active state (e.g., 0.7) */\n activeThreshold: number;\n /** Score threshold for ready state (e.g., 0.3) */\n readyThreshold: number;\n}\n\nexport interface ConfidenceStates {\n /**\n * Calculate state based on entry score and BTSP flag\n * @param entry - Memory entry\n * @returns Confidence state\n */\n calculateState(entry: MemoryEntry): ConfidenceState;\n\n /**\n * Transition entry to correct state based on its score\n * @param entry - Entry to transition\n * @returns Entry with updated state\n */\n transition(entry: MemoryEntry): MemoryEntry;\n\n /**\n * Get distribution of states across all entries\n * @param entries - All memory entries\n * @returns State distribution with counts\n */\n getDistribution(entries: MemoryEntry[]): StateDistribution;\n}\n\n/**\n * Create a confidence states manager\n * @param config - States configuration\n * @returns ConfidenceStates instance\n */\nexport function createConfidenceStates(config: ConfidenceStatesConfig): ConfidenceStates {\n const { activeThreshold, readyThreshold } = config;\n\n function calculateState(entry: MemoryEntry): ConfidenceState {\n // BTSP entries are always active\n if (entry.isBTSP) {\n return 'active';\n }\n\n // State based on score thresholds\n // Active: score > 0.7\n if (entry.score > activeThreshold) {\n return 'active';\n }\n\n // Ready: 0.3 <= score <= 0.7\n if (entry.score >= readyThreshold) {\n return 'ready';\n }\n\n // Silent: score < 0.3\n return 'silent';\n }\n\n function transition(entry: MemoryEntry): MemoryEntry {\n const newState = calculateState(entry);\n\n return {\n ...entry,\n state: newState,\n };\n }\n\n function getDistribution(entries: MemoryEntry[]): StateDistribution {\n const distribution: StateDistribution = {\n silent: 0,\n ready: 0,\n active: 0,\n total: entries.length,\n };\n\n for (const entry of entries) {\n const state = calculateState(entry);\n distribution[state]++;\n }\n\n return distribution;\n }\n\n return {\n calculateState,\n transition,\n getDistribution,\n };\n}\n","/**\n * Engram Scorer - Implements engram theory (memory decay)\n *\n * Neuroscience: Memories fade over time without reinforcement.\n * Application: Apply exponential decay formula to memory scores based on age and access count.\n *\n * Formula: decay = 1 - e^(-age/TTL)\n * Score adjustment: score_new = score_old * (1 - decay) + (accessCount bonus)\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\n\nexport interface EngramScorerConfig {\n /** Default TTL in hours for new entries */\n defaultTTL: number;\n /** Decay threshold (0.0-1.0) above which entries are marked for pruning */\n decayThreshold: number;\n}\n\nexport interface EngramScorer {\n /**\n * Calculate current score for an entry based on decay and access count\n * @param entry - Memory entry to score\n * @param currentTime - Current timestamp in milliseconds (for testing)\n * @returns Updated score (0.0-1.0)\n */\n calculateScore(entry: MemoryEntry, currentTime?: number): number;\n\n /**\n * Refresh TTL to default value\n * @param entry - Entry to refresh\n * @returns Entry with refreshed TTL and timestamp\n */\n refreshTTL(entry: MemoryEntry): MemoryEntry;\n\n /**\n * Calculate decay factor (0.0-1.0) based on age and TTL\n * @param ageInSeconds - Age of entry in seconds\n * @param ttlInSeconds - TTL in seconds\n * @returns Decay factor (0.0 = fresh, 1.0 = fully decayed)\n */\n calculateDecay(ageInSeconds: number, ttlInSeconds: number): number;\n}\n\n/**\n * Create an engram scorer instance\n * @param config - Scorer configuration\n * @returns EngramScorer instance\n */\nexport function createEngramScorer(config: EngramScorerConfig): EngramScorer {\n const { defaultTTL } = config;\n\n function calculateDecay(ageInSeconds: number, ttlInSeconds: number): number {\n if (ttlInSeconds === 0) return 1.0; // Instant decay\n if (ageInSeconds <= 0) return 0.0; // Fresh entry\n\n // Exponential decay: 1 - e^(-age/TTL)\n const ratio = ageInSeconds / ttlInSeconds;\n const decay = 1 - Math.exp(-ratio);\n\n // Clamp to [0.0, 1.0]\n return Math.max(0, Math.min(1, decay));\n }\n\n function calculateScore(entry: MemoryEntry, currentTime: number = Date.now()): number {\n // Calculate age in seconds\n const ageInMilliseconds = currentTime - entry.timestamp;\n const ageInSeconds = Math.max(0, ageInMilliseconds / 1000);\n\n // Calculate decay factor\n const decay = calculateDecay(ageInSeconds, entry.ttl);\n\n // Base score reduced by decay\n let score = entry.score * (1 - decay);\n\n // Access count bonus (diminishing returns via log)\n if (entry.accessCount > 0) {\n const accessBonus = Math.log(entry.accessCount + 1) * 0.1;\n score = Math.min(1.0, score + accessBonus);\n }\n\n // BTSP entries maintain high score\n if (entry.isBTSP) {\n score = Math.max(score, 0.9);\n }\n\n return Math.max(0, Math.min(1, score));\n }\n\n function refreshTTL(entry: MemoryEntry): MemoryEntry {\n return {\n ...entry,\n ttl: defaultTTL * 3600, // Convert hours to seconds\n timestamp: Date.now(),\n };\n }\n\n return {\n calculateScore,\n refreshTTL,\n calculateDecay,\n };\n}\n","/**\n * Token estimation utilities.\n * Uses whitespace heuristic (~90% accuracy vs GPT tokenizer).\n */\n\n/**\n * Estimate token count for text using heuristic.\n *\n * Approximation: 1 token ≈ 4 chars or 0.75 words\n * Provides ~90% accuracy compared to GPT tokenizer, sufficient for optimization heuristics.\n *\n * @param text - Text to count\n * @returns Estimated token count\n *\n * @example\n * ```typescript\n * const tokens = estimateTokens('Hello world');\n * console.log(tokens); // ~2\n * ```\n */\nexport function estimateTokens(text: string): number {\n if (!text || text.length === 0) {\n return 0;\n }\n\n // Split on whitespace to get words\n const words = text.split(/\\s+/).filter((w) => w.length > 0);\n const wordCount = words.length;\n\n // Character-based estimate\n const charCount = text.length;\n const charEstimate = Math.ceil(charCount / 4);\n\n // Word-based estimate\n const wordEstimate = Math.ceil(wordCount * 0.75);\n\n // Return the maximum of both estimates (more conservative)\n return Math.max(wordEstimate, charEstimate);\n}\n","/**\n * Sparse Pruner - Implements sparse coding principle\n *\n * Neuroscience: Only 2-5% of neurons fire at any given time.\n * Application: Keep only top 5% most relevant context entries by TF-IDF score.\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\nexport interface SparsePrunerConfig {\n /** Percentage threshold for pruning (e.g., 5 = keep top 5%) */\n threshold: number;\n}\n\nexport interface SparsePruner {\n /**\n * Prune entries to keep only top N% by relevance score\n * @param entries - Memory entries to prune\n * @returns Result with kept/removed entries and token counts\n */\n prune(entries: MemoryEntry[]): PruneResult;\n\n /**\n * Calculate TF-IDF relevance score for a single entry\n * @param entry - Entry to score\n * @param allEntries - All entries for IDF calculation\n * @returns Relevance score (0.0-1.0)\n */\n scoreEntry(entry: MemoryEntry, allEntries: MemoryEntry[]): number;\n}\n\n/**\n * Create a sparse pruner instance\n * @param config - Pruner configuration\n * @returns SparsePruner instance\n */\nexport function createSparsePruner(config: SparsePrunerConfig): SparsePruner {\n const { threshold } = config;\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n function calculateTF(term: string, tokens: string[]): number {\n const count = tokens.filter((t) => t === term).length;\n // Sqrt capping to prevent common words from dominating\n return Math.sqrt(count);\n }\n\n function calculateIDF(term: string, allEntries: MemoryEntry[]): number {\n const totalDocs = allEntries.length;\n const docsWithTerm = allEntries.filter((entry) => {\n const tokens = tokenize(entry.content);\n return tokens.includes(term);\n }).length;\n\n if (docsWithTerm === 0) return 0;\n\n return Math.log(totalDocs / docsWithTerm);\n }\n\n function scoreEntry(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tokens = tokenize(entry.content);\n if (tokens.length === 0) return 0;\n\n const uniqueTerms = [...new Set(tokens)];\n let totalScore = 0;\n\n for (const term of uniqueTerms) {\n const tf = calculateTF(term, tokens);\n const idf = calculateIDF(term, allEntries);\n totalScore += tf * idf;\n }\n\n // Normalize by entry length\n return totalScore / tokens.length;\n }\n\n function prune(entries: MemoryEntry[]): PruneResult {\n if (entries.length === 0) {\n return {\n kept: [],\n removed: [],\n originalTokens: 0,\n prunedTokens: 0,\n };\n }\n\n // Calculate original token count\n const originalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Score all entries\n const scored = entries.map((entry) => ({\n entry,\n score: scoreEntry(entry, entries),\n }));\n\n // Sort by score descending\n scored.sort((a, b) => b.score - a.score);\n\n // Keep top N% (minimum 1 entry)\n const keepCount = Math.max(1, Math.ceil(entries.length * (threshold / 100)));\n const kept = scored.slice(0, keepCount).map((s) => s.entry);\n const removed = scored.slice(keepCount).map((s) => s.entry);\n\n // Calculate pruned token count\n const prunedTokens = kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n kept,\n removed,\n originalTokens,\n prunedTokens,\n };\n }\n\n return {\n prune,\n scoreEntry,\n };\n}\n","/**\n * Context Parser - Shared utilities for parsing agent contexts into memory entries\n *\n * Extracted from claude-code adapter to enable reuse across:\n * - Adapters (claude-code, generic)\n * - Real-time pipeline (streaming context)\n * - Hooks (pre-prompt, post-tool-result)\n */\n\nimport { randomUUID } from 'node:crypto';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from './hash.js';\n\n/**\n * Block type classification for Claude Code context\n */\nexport type BlockType = 'conversation' | 'tool' | 'result' | 'other';\n\n/**\n * Parse Claude Code context into memory entries\n * Handles conversation turns, tool uses, and results\n * @param context - Raw context string\n * @returns Array of memory entries\n */\nexport function parseClaudeCodeContext(context: string): MemoryEntry[] {\n const entries: MemoryEntry[] = [];\n const now = Date.now();\n\n // Split by conversation turns and tool boundaries\n const lines = context.split('\\n');\n let currentBlock: string[] = [];\n let blockType: BlockType = 'other';\n\n for (const line of lines) {\n const trimmed = line.trim();\n\n // Detect conversation turns\n if (trimmed.startsWith('User:') || trimmed.startsWith('Assistant:')) {\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'conversation';\n currentBlock.push(line);\n }\n // Detect tool calls\n else if (\n trimmed.includes('<function_calls>') ||\n trimmed.includes('<invoke>') ||\n trimmed.includes('<tool_use>')\n ) {\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'tool';\n currentBlock.push(line);\n }\n // Detect tool results\n else if (trimmed.includes('<function_results>') || trimmed.includes('</function_results>')) {\n if (currentBlock.length > 0 && blockType !== 'result') {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n currentBlock = [];\n }\n blockType = 'result';\n currentBlock.push(line);\n }\n // Continue current block\n else if (currentBlock.length > 0) {\n currentBlock.push(line);\n }\n // Start new block if line has content\n else if (trimmed.length > 0) {\n currentBlock.push(line);\n blockType = 'other';\n }\n }\n\n // Add final block\n if (currentBlock.length > 0) {\n entries.push(createEntry(currentBlock.join('\\n'), blockType, now));\n }\n\n return entries.filter((e) => e.content.trim().length > 0);\n}\n\n/**\n * Create a memory entry from a content block\n * @param content - Block content\n * @param type - Block type\n * @param baseTime - Base timestamp\n * @returns Memory entry\n */\nexport function createEntry(content: string, type: BlockType, baseTime: number): MemoryEntry {\n const tags: string[] = [type];\n\n // Assign initial score based on type\n let initialScore = 0.5;\n if (type === 'conversation') initialScore = 0.8; // Prioritize conversation\n if (type === 'tool') initialScore = 0.7; // Tool calls are important\n if (type === 'result') initialScore = 0.4; // Results can be verbose\n\n return {\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: baseTime,\n score: initialScore,\n state: initialScore > 0.7 ? 'active' : initialScore > 0.3 ? 'ready' : 'silent',\n ttl: 24 * 3600, // 24 hours default\n accessCount: 0,\n tags,\n metadata: { type },\n isBTSP: false,\n };\n}\n\n/**\n * Parse generic context (fallback for non-Claude-Code agents)\n * Splits on double newlines, treats as paragraphs\n * @param context - Raw context string\n * @returns Array of memory entries\n */\nexport function parseGenericContext(context: string): MemoryEntry[] {\n const entries: MemoryEntry[] = [];\n const now = Date.now();\n\n // Split on double newlines (paragraph boundaries)\n const blocks = context.split(/\\n\\n+/);\n\n for (const block of blocks) {\n const trimmed = block.trim();\n if (trimmed.length === 0) continue;\n\n entries.push(createEntry(trimmed, 'other', now));\n }\n\n return entries;\n}\n","/**\n * Claude Code Adapter - Claude Code-specific optimization pipeline\n *\n * Optimized for Claude Code's conversation patterns, tool use, and context management.\n * Implements the same AgentAdapter interface as GenericAdapter but with Claude-specific tuning.\n */\n\nimport { createBTSPEmbedder } from '../core/btsp-embedder.js';\nimport { createConfidenceStates } from '../core/confidence-states.js';\nimport { createEngramScorer } from '../core/engram-scorer.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSparsePruner } from '../core/sparse-pruner.js';\nimport type { AgentAdapter, OptimizationResult, OptimizeOptions } from '../types/adapter.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { parseClaudeCodeContext } from '../utils/context-parser.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\n/**\n * Claude Code-specific optimization profile\n * Tuned for Claude's conversation patterns and tool use\n */\nconst CLAUDE_CODE_PROFILE = {\n // More aggressive pruning for tool results (they can be verbose)\n toolResultThreshold: 3, // Keep top 3% of tool results\n\n // Preserve conversation turns more aggressively\n conversationBoost: 1.5, // 50% boost for User/Assistant exchanges\n\n // Prioritize recent context (Claude Code sessions are typically focused)\n recentContextWindow: 10 * 60, // Last 10 minutes gets priority\n\n // BTSP patterns specific to Claude Code\n btspPatterns: [\n // Error patterns\n /\\b(error|exception|failure|fatal|critical|panic)\\b/i,\n /^\\s+at\\s+.*\\(.*:\\d+:\\d+\\)/m, // Stack traces\n /^Error:/m,\n\n // Git conflict markers\n /^<<<<<<< /m,\n /^=======/m,\n /^>>>>>>> /m,\n\n // Tool use patterns (important for context)\n /<function_calls>/,\n /<invoke>/,\n /<tool_use>/,\n\n // File operation results (often critical)\n /ENOENT|EACCES|EISDIR|EEXIST/,\n ],\n};\n\n/**\n * Create a Claude Code adapter instance\n * @param memory - KV memory store\n * @param config - Sparn configuration\n * @returns AgentAdapter instance optimized for Claude Code\n */\nexport function createClaudeCodeAdapter(memory: KVMemory, config: SparnConfig): AgentAdapter {\n // Create core modules with Claude Code-optimized settings\n const pruner = createSparsePruner({\n threshold: config.pruning.threshold,\n });\n\n const scorer = createEngramScorer(config.decay);\n const states = createConfidenceStates(config.states);\n const btsp = createBTSPEmbedder();\n\n async function optimize(\n context: string,\n options: OptimizeOptions = {},\n ): Promise<OptimizationResult> {\n const startTime = Date.now();\n\n // Parse context into entries\n // For Claude Code, we parse by conversation turns and tool uses\n const entries = parseClaudeCodeContext(context);\n\n // Apply BTSP detection with Claude Code-specific patterns\n const entriesWithBTSP = entries.map((entry) => {\n const isBTSP = CLAUDE_CODE_PROFILE.btspPatterns.some((pattern) =>\n pattern.test(entry.content),\n );\n\n if (isBTSP) {\n const btspEntry = btsp.createBTSPEntry(entry.content, [...entry.tags, 'claude-code'], {\n originalTimestamp: entry.timestamp,\n });\n // Preserve original timestamp\n return {\n ...btspEntry,\n timestamp: entry.timestamp,\n };\n }\n\n return entry;\n });\n\n // Apply conversation boost to User/Assistant exchanges\n const boostedEntries = entriesWithBTSP.map((entry) => {\n const isConversationTurn =\n entry.content.trim().startsWith('User:') || entry.content.trim().startsWith('Assistant:');\n\n if (isConversationTurn) {\n return {\n ...entry,\n score: entry.score * CLAUDE_CODE_PROFILE.conversationBoost,\n };\n }\n\n return entry;\n });\n\n // Score entries with decay\n const scoredEntries = boostedEntries.map((entry) => {\n const decayScore = scorer.calculateScore(entry);\n return {\n ...entry,\n score: decayScore,\n };\n });\n\n // Calculate states\n const entriesWithStates = scoredEntries.map((entry) => {\n const state = states.calculateState(entry);\n return {\n ...entry,\n state,\n };\n });\n\n // Prune entries (keep top N%)\n const pruneResult = pruner.prune(entriesWithStates);\n\n // Store kept entries in memory (if not dry-run)\n if (!options.dryRun) {\n for (const entry of pruneResult.kept) {\n await memory.put(entry);\n }\n\n // Record optimization stats\n await memory.recordOptimization({\n timestamp: Date.now(),\n tokens_before: pruneResult.originalTokens,\n tokens_after: pruneResult.prunedTokens,\n entries_pruned: pruneResult.removed.length,\n duration_ms: Date.now() - startTime,\n });\n }\n\n // Build optimized context from kept entries\n const optimizedContext = pruneResult.kept.map((entry) => entry.content).join('\\n');\n\n // Calculate state distribution\n const stateDistribution = states.getDistribution(pruneResult.kept);\n\n // Build result\n const result: OptimizationResult = {\n optimizedContext,\n tokensBefore: pruneResult.originalTokens,\n tokensAfter: pruneResult.prunedTokens,\n reduction:\n pruneResult.originalTokens > 0\n ? (pruneResult.originalTokens - pruneResult.prunedTokens) / pruneResult.originalTokens\n : 0,\n entriesProcessed: entries.length,\n entriesKept: pruneResult.kept.length,\n durationMs: Date.now() - startTime,\n stateDistribution,\n };\n\n // Add verbose details if requested\n if (options.verbose) {\n result.details = pruneResult.kept.map((entry) => ({\n id: entry.id,\n score: entry.score,\n state: entry.state || 'unknown',\n isBTSP: entry.tags.includes('btsp'),\n tokens: estimateTokens(entry.content),\n }));\n }\n\n return result;\n }\n\n return {\n optimize,\n };\n}\n","/**\n * Generic Adapter - Agent-agnostic optimization pipeline\n *\n * Orchestrates all 6 neuroscience modules to optimize context memory.\n */\n\nimport { randomUUID } from 'node:crypto';\nimport { createBTSPEmbedder } from '../core/btsp-embedder.js';\nimport { createConfidenceStates } from '../core/confidence-states.js';\nimport { createEngramScorer } from '../core/engram-scorer.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSparsePruner } from '../core/sparse-pruner.js';\nimport type { AgentAdapter, OptimizationResult, OptimizeOptions } from '../types/adapter.js';\nimport type { SparnConfig } from '../types/config.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { hashContent } from '../utils/hash.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\n\n/**\n * Create a generic adapter instance\n * @param memory - KV memory store\n * @param config - Sparn configuration\n * @returns AgentAdapter instance\n */\nexport function createGenericAdapter(memory: KVMemory, config: SparnConfig): AgentAdapter {\n const pruner = createSparsePruner(config.pruning);\n const scorer = createEngramScorer(config.decay);\n const states = createConfidenceStates(config.states);\n const btsp = createBTSPEmbedder();\n\n async function optimize(\n context: string,\n options: OptimizeOptions = {},\n ): Promise<OptimizationResult> {\n const startTime = Date.now();\n\n // Parse context into entries (line-based for simplicity)\n const lines = context.split('\\n').filter((line) => line.trim().length > 0);\n const entries: MemoryEntry[] = lines.map((content) => ({\n id: randomUUID(),\n content,\n hash: hashContent(content),\n timestamp: Date.now(),\n score: btsp.detectBTSP(content) ? 1.0 : 0.5, // BTSP gets high initial score\n ttl: config.decay.defaultTTL * 3600, // Convert hours to seconds\n state: 'ready' as const,\n accessCount: 0,\n tags: [],\n metadata: {},\n isBTSP: btsp.detectBTSP(content),\n }));\n\n // Calculate original token count\n const tokensBefore = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 1: Update scores with decay\n const scoredEntries = entries.map((entry) => ({\n ...entry,\n score: scorer.calculateScore(entry),\n }));\n\n // Step 2: Transition states based on scores\n const statedEntries = scoredEntries.map((entry) => states.transition(entry));\n\n // Step 3: Apply sparse pruning\n const pruneResult = pruner.prune(statedEntries);\n\n // Step 4: Keep active and ready entries, discard silent\n const optimizedEntries = pruneResult.kept.filter(\n (e) => e.state === 'active' || e.state === 'ready',\n );\n\n // Calculate final token count\n const tokensAfter = optimizedEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Reconstruct optimized context\n const optimizedContext = optimizedEntries.map((e) => e.content).join('\\n');\n\n // Store entries in memory (if not dry run)\n if (!options.dryRun) {\n for (const entry of optimizedEntries) {\n await memory.put(entry);\n }\n\n // Record optimization statistics\n await memory.recordOptimization({\n timestamp: Date.now(),\n tokens_before: tokensBefore,\n tokens_after: tokensAfter,\n entries_pruned: entries.length - optimizedEntries.length,\n duration_ms: Date.now() - startTime,\n });\n }\n\n // Get state distribution\n const distribution = states.getDistribution(optimizedEntries);\n\n const result: OptimizationResult = {\n optimizedContext,\n tokensBefore,\n tokensAfter,\n reduction: tokensBefore > 0 ? (tokensBefore - tokensAfter) / tokensBefore : 0,\n entriesProcessed: entries.length,\n entriesKept: optimizedEntries.length,\n stateDistribution: distribution,\n durationMs: Date.now() - startTime,\n };\n\n // Add verbose details if requested\n if (options.verbose) {\n result.details = optimizedEntries.map((e) => ({\n id: e.id,\n score: e.score,\n state: e.state,\n isBTSP: e.isBTSP,\n tokens: estimateTokens(e.content),\n }));\n }\n\n return result;\n }\n\n return {\n optimize,\n };\n}\n","/**\n * Budget-Aware Pruner - Token budget optimization\n *\n * Unlike SparsePruner which keeps top N% entries, BudgetPruner fits entries\n * within a target token budget using priority scoring that combines:\n * - TF-IDF relevance\n * - Engram decay\n * - Confidence state multipliers\n * - BTSP bypass (always included)\n *\n * Target use case: Real-time optimization for Opus model (~50K token budget)\n */\n\nimport type { RealtimeConfig } from '../types/config.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { createEngramScorer } from './engram-scorer.js';\n\nexport interface BudgetPrunerConfig {\n /** Target token budget */\n tokenBudget: number;\n /** Decay configuration */\n decay: {\n defaultTTL: number;\n decayThreshold: number;\n };\n /** State multipliers */\n states: {\n activeThreshold: number;\n readyThreshold: number;\n };\n}\n\nexport interface BudgetPruner {\n /**\n * Prune entries to fit within token budget\n * @param entries - Memory entries to prune\n * @param budget - Optional override budget (uses config default if not provided)\n * @returns Result with kept/removed entries and budget utilization\n */\n pruneToFit(entries: MemoryEntry[], budget?: number): PruneResult & { budgetUtilization: number };\n\n /**\n * Calculate priority score for an entry\n * @param entry - Entry to score\n * @param allEntries - All entries for TF-IDF calculation\n * @returns Priority score (higher = more important)\n */\n priorityScore(entry: MemoryEntry, allEntries: MemoryEntry[]): number;\n}\n\n/**\n * Create a budget-aware pruner instance\n * @param config - Pruner configuration\n * @returns BudgetPruner instance\n */\nexport function createBudgetPruner(config: BudgetPrunerConfig): BudgetPruner {\n const { tokenBudget, decay } = config;\n const engramScorer = createEngramScorer(decay);\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n function calculateTF(term: string, tokens: string[]): number {\n const count = tokens.filter((t) => t === term).length;\n // Sqrt capping to prevent common words from dominating\n return Math.sqrt(count);\n }\n\n function calculateIDF(term: string, allEntries: MemoryEntry[]): number {\n const totalDocs = allEntries.length;\n const docsWithTerm = allEntries.filter((entry) => {\n const tokens = tokenize(entry.content);\n return tokens.includes(term);\n }).length;\n\n if (docsWithTerm === 0) return 0;\n\n return Math.log(totalDocs / docsWithTerm);\n }\n\n function calculateTFIDF(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tokens = tokenize(entry.content);\n if (tokens.length === 0) return 0;\n\n const uniqueTerms = [...new Set(tokens)];\n let totalScore = 0;\n\n for (const term of uniqueTerms) {\n const tf = calculateTF(term, tokens);\n const idf = calculateIDF(term, allEntries);\n totalScore += tf * idf;\n }\n\n // Normalize by entry length\n return totalScore / tokens.length;\n }\n\n function getStateMultiplier(entry: MemoryEntry): number {\n // BTSP entries get max priority (handled separately, but keep high multiplier)\n if (entry.isBTSP) return 2.0;\n\n // State-based multipliers\n switch (entry.state) {\n case 'active':\n return 2.0;\n case 'ready':\n return 1.0;\n case 'silent':\n return 0.5;\n default:\n return 1.0;\n }\n }\n\n function priorityScore(entry: MemoryEntry, allEntries: MemoryEntry[]): number {\n const tfidf = calculateTFIDF(entry, allEntries);\n const currentScore = engramScorer.calculateScore(entry);\n const engramDecay = 1 - currentScore; // Lower decay = higher priority\n const stateMultiplier = getStateMultiplier(entry);\n\n // Priority = TF-IDF * (1 - decay) * state_multiplier\n // This balances relevance, recency, and confidence state\n return tfidf * (1 - engramDecay) * stateMultiplier;\n }\n\n function pruneToFit(\n entries: MemoryEntry[],\n budget: number = tokenBudget,\n ): PruneResult & { budgetUtilization: number } {\n if (entries.length === 0) {\n return {\n kept: [],\n removed: [],\n originalTokens: 0,\n prunedTokens: 0,\n budgetUtilization: 0,\n };\n }\n\n // Calculate original token count\n const originalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 1: Separate BTSP entries (always included, bypass budget)\n const btspEntries = entries.filter((e) => e.isBTSP);\n const regularEntries = entries.filter((e) => !e.isBTSP);\n\n const btspTokens = btspEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Step 2: Score regular entries\n const scored = regularEntries.map((entry) => ({\n entry,\n score: priorityScore(entry, entries),\n tokens: estimateTokens(entry.content),\n }));\n\n // Step 3: Sort by priority score descending\n scored.sort((a, b) => b.score - a.score);\n\n // Step 4: Greedy fill until budget exceeded\n const kept: MemoryEntry[] = [...btspEntries];\n const removed: MemoryEntry[] = [];\n let currentTokens = btspTokens;\n\n for (const item of scored) {\n if (currentTokens + item.tokens <= budget) {\n kept.push(item.entry);\n currentTokens += item.tokens;\n } else {\n removed.push(item.entry);\n }\n }\n\n const budgetUtilization = budget > 0 ? currentTokens / budget : 0;\n\n return {\n kept,\n removed,\n originalTokens,\n prunedTokens: currentTokens,\n budgetUtilization,\n };\n }\n\n return {\n pruneToFit,\n priorityScore,\n };\n}\n\n/**\n * Helper to create budget pruner from RealtimeConfig\n * @param realtimeConfig - Realtime configuration\n * @param decayConfig - Decay configuration\n * @param statesConfig - States configuration\n * @returns BudgetPruner instance\n */\nexport function createBudgetPrunerFromConfig(\n realtimeConfig: RealtimeConfig,\n decayConfig: { defaultTTL: number; decayThreshold: number },\n statesConfig: { activeThreshold: number; readyThreshold: number },\n): BudgetPruner {\n return createBudgetPruner({\n tokenBudget: realtimeConfig.tokenBudget,\n decay: decayConfig,\n states: statesConfig,\n });\n}\n","/**\n * Metrics and Telemetry System\n *\n * Tracks performance metrics and optimization statistics:\n * - Optimization duration and throughput\n * - Token savings and reduction rates\n * - Memory usage and cache hit rates\n * - Daemon uptime and session counts\n */\n\nexport interface OptimizationMetric {\n timestamp: number;\n duration: number;\n tokensBefore: number;\n tokensAfter: number;\n entriesProcessed: number;\n entriesKept: number;\n cacheHitRate: number;\n memoryUsage: number;\n}\n\nexport interface DaemonMetric {\n startTime: number;\n sessionsWatched: number;\n totalOptimizations: number;\n totalTokensSaved: number;\n averageLatency: number;\n memoryUsage: number;\n}\n\nexport interface MetricsSnapshot {\n timestamp: number;\n optimization: {\n totalRuns: number;\n totalDuration: number;\n totalTokensSaved: number;\n averageReduction: number;\n p50Latency: number;\n p95Latency: number;\n p99Latency: number;\n };\n cache: {\n hitRate: number;\n totalHits: number;\n totalMisses: number;\n size: number;\n };\n daemon: {\n uptime: number;\n sessionsWatched: number;\n memoryUsage: number;\n };\n}\n\nexport interface MetricsCollector {\n /**\n * Record an optimization metric\n */\n recordOptimization(metric: OptimizationMetric): void;\n\n /**\n * Update daemon metrics\n */\n updateDaemon(metric: Partial<DaemonMetric>): void;\n\n /**\n * Get current metrics snapshot\n */\n getSnapshot(): MetricsSnapshot;\n\n /**\n * Export metrics as JSON\n */\n export(): string;\n\n /**\n * Reset all metrics\n */\n reset(): void;\n}\n\n/**\n * Create a metrics collector instance\n */\nexport function createMetricsCollector(): MetricsCollector {\n const optimizations: OptimizationMetric[] = [];\n let daemonMetrics: DaemonMetric = {\n startTime: Date.now(),\n sessionsWatched: 0,\n totalOptimizations: 0,\n totalTokensSaved: 0,\n averageLatency: 0,\n memoryUsage: 0,\n };\n\n let cacheHits = 0;\n let cacheMisses = 0;\n\n function recordOptimization(metric: OptimizationMetric): void {\n optimizations.push(metric);\n\n // Update daemon totals\n daemonMetrics.totalOptimizations++;\n daemonMetrics.totalTokensSaved += metric.tokensBefore - metric.tokensAfter;\n\n // Update cache stats\n if (metric.cacheHitRate > 0) {\n const hits = Math.round(metric.entriesProcessed * metric.cacheHitRate);\n cacheHits += hits;\n cacheMisses += metric.entriesProcessed - hits;\n }\n\n // Update average latency (moving average)\n daemonMetrics.averageLatency =\n (daemonMetrics.averageLatency * (daemonMetrics.totalOptimizations - 1) + metric.duration) /\n daemonMetrics.totalOptimizations;\n\n // Keep only last 1000 metrics in memory\n if (optimizations.length > 1000) {\n optimizations.shift();\n }\n }\n\n function updateDaemon(metric: Partial<DaemonMetric>): void {\n daemonMetrics = {\n ...daemonMetrics,\n ...metric,\n };\n }\n\n function calculatePercentile(values: number[], percentile: number): number {\n if (values.length === 0) return 0;\n\n const sorted = [...values].sort((a, b) => a - b);\n const index = Math.ceil((percentile / 100) * sorted.length) - 1;\n return sorted[index] || 0;\n }\n\n function getSnapshot(): MetricsSnapshot {\n const totalRuns = optimizations.length;\n const totalDuration = optimizations.reduce((sum, m) => sum + m.duration, 0);\n const totalTokensSaved = optimizations.reduce(\n (sum, m) => sum + (m.tokensBefore - m.tokensAfter),\n 0,\n );\n\n const totalTokensBefore = optimizations.reduce((sum, m) => sum + m.tokensBefore, 0);\n const averageReduction = totalTokensBefore > 0 ? totalTokensSaved / totalTokensBefore : 0;\n\n const durations = optimizations.map((m) => m.duration);\n\n const totalCacheQueries = cacheHits + cacheMisses;\n const hitRate = totalCacheQueries > 0 ? cacheHits / totalCacheQueries : 0;\n\n return {\n timestamp: Date.now(),\n optimization: {\n totalRuns,\n totalDuration,\n totalTokensSaved,\n averageReduction,\n p50Latency: calculatePercentile(durations, 50),\n p95Latency: calculatePercentile(durations, 95),\n p99Latency: calculatePercentile(durations, 99),\n },\n cache: {\n hitRate,\n totalHits: cacheHits,\n totalMisses: cacheMisses,\n size: optimizations.reduce((sum, m) => sum + m.entriesKept, 0),\n },\n daemon: {\n uptime: Date.now() - daemonMetrics.startTime,\n sessionsWatched: daemonMetrics.sessionsWatched,\n memoryUsage: daemonMetrics.memoryUsage,\n },\n };\n }\n\n function exportMetrics(): string {\n return JSON.stringify(getSnapshot(), null, 2);\n }\n\n function reset(): void {\n optimizations.length = 0;\n cacheHits = 0;\n cacheMisses = 0;\n daemonMetrics = {\n startTime: Date.now(),\n sessionsWatched: 0,\n totalOptimizations: 0,\n totalTokensSaved: 0,\n averageLatency: 0,\n memoryUsage: 0,\n };\n }\n\n return {\n recordOptimization,\n updateDaemon,\n getSnapshot,\n export: exportMetrics,\n reset,\n };\n}\n\n// Global metrics instance\nlet globalMetrics: MetricsCollector | null = null;\n\n/**\n * Get or create the global metrics collector\n */\nexport function getMetrics(): MetricsCollector {\n if (!globalMetrics) {\n globalMetrics = createMetricsCollector();\n }\n return globalMetrics;\n}\n","/**\n * Incremental Optimizer - Cache-based delta processing\n *\n * Optimizes performance for real-time scenarios by:\n * - Caching entry scores by content hash\n * - Only recomputing scores for new/changed entries\n * - Pre-computing and caching document frequency tables\n * - Periodically forcing full re-optimization to prevent drift\n *\n * Target: <50ms for incremental updates on 100K token contexts\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport type { PruneResult } from '../types/pruner.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { type BudgetPrunerConfig, createBudgetPruner } from './budget-pruner.js';\nimport { getMetrics } from './metrics.js';\n\nexport interface IncrementalOptimizerConfig extends BudgetPrunerConfig {\n /** Force full re-optimization every N incremental updates */\n fullOptimizationInterval: number;\n}\n\nexport interface IncrementalOptimizerState {\n /** Entry cache keyed by content hash */\n entryCache: Map<string, { entry: MemoryEntry; score: number; timestamp: number }>;\n /** Document frequency table for IDF calculation */\n documentFrequency: Map<string, number>;\n /** Total document count for IDF */\n totalDocuments: number;\n /** Incremental update counter */\n updateCount: number;\n /** Last full optimization timestamp */\n lastFullOptimization: number;\n}\n\nexport interface IncrementalOptimizer {\n /**\n * Optimize incrementally (only process new/changed entries)\n * @param newEntries - New entries to add\n * @param budget - Optional budget override\n * @returns Prune result with budget utilization\n */\n optimizeIncremental(\n newEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number };\n\n /**\n * Optimize fully (recompute all scores)\n * @param allEntries - All entries to optimize\n * @param budget - Optional budget override\n * @returns Prune result with budget utilization\n */\n optimizeFull(\n allEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number };\n\n /**\n * Get current optimizer state (for serialization)\n * @returns Serializable state object\n */\n getState(): IncrementalOptimizerState;\n\n /**\n * Restore optimizer state (from serialization)\n * @param state - State to restore\n */\n restoreState(state: IncrementalOptimizerState): void;\n\n /**\n * Reset optimizer state (clear all caches)\n */\n reset(): void;\n\n /**\n * Get cache statistics\n * @returns Cache stats\n */\n getStats(): {\n cachedEntries: number;\n uniqueTerms: number;\n totalDocuments: number;\n updateCount: number;\n lastFullOptimization: number;\n };\n}\n\n/**\n * Create an incremental optimizer instance\n * @param config - Optimizer configuration\n * @returns IncrementalOptimizer instance\n */\nexport function createIncrementalOptimizer(\n config: IncrementalOptimizerConfig,\n): IncrementalOptimizer {\n const pruner = createBudgetPruner(config);\n const { fullOptimizationInterval } = config;\n\n // Internal state\n let state: IncrementalOptimizerState = {\n entryCache: new Map(),\n documentFrequency: new Map(),\n totalDocuments: 0,\n updateCount: 0,\n lastFullOptimization: Date.now(),\n };\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n /**\n * Update document frequency table incrementally\n */\n function updateDocumentFrequency(entries: MemoryEntry[], remove = false): void {\n for (const entry of entries) {\n const tokens = tokenize(entry.content);\n const uniqueTerms = [...new Set(tokens)];\n\n for (const term of uniqueTerms) {\n const current = state.documentFrequency.get(term) || 0;\n const updated = remove ? Math.max(0, current - 1) : current + 1;\n\n if (updated === 0) {\n state.documentFrequency.delete(term);\n } else {\n state.documentFrequency.set(term, updated);\n }\n }\n }\n\n state.totalDocuments += remove ? -entries.length : entries.length;\n state.totalDocuments = Math.max(0, state.totalDocuments);\n }\n\n /**\n * Check if entry is cached and still valid\n */\n function getCachedEntry(hash: string): MemoryEntry | null {\n const cached = state.entryCache.get(hash);\n if (!cached) return null;\n\n // Entry is valid if found in cache\n return cached.entry;\n }\n\n /**\n * Cache entry with score\n */\n function cacheEntry(entry: MemoryEntry, score: number): void {\n state.entryCache.set(entry.hash, {\n entry,\n score,\n timestamp: Date.now(),\n });\n }\n\n function optimizeIncremental(\n newEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number } {\n const startTime = Date.now();\n state.updateCount++;\n\n // Force full optimization if interval reached\n if (state.updateCount >= fullOptimizationInterval) {\n // Get all cached entries\n const allEntries = Array.from(state.entryCache.values()).map((c) => c.entry);\n return optimizeFull([...allEntries, ...newEntries], budget);\n }\n\n // Filter out already-cached entries\n const uncachedEntries: MemoryEntry[] = [];\n const cachedEntries: MemoryEntry[] = [];\n\n for (const entry of newEntries) {\n const cached = getCachedEntry(entry.hash);\n if (cached) {\n cachedEntries.push(cached);\n } else {\n uncachedEntries.push(entry);\n }\n }\n\n // Update document frequency for new entries only\n if (uncachedEntries.length > 0) {\n updateDocumentFrequency(uncachedEntries, false);\n }\n\n // Combine with cached entries for scoring context\n const allEntries = [...cachedEntries, ...uncachedEntries];\n\n // Score only uncached entries (reuse cached scores)\n for (const entry of uncachedEntries) {\n const score = pruner.priorityScore(entry, allEntries);\n cacheEntry(entry, score);\n }\n\n // Get all current entries (from cache + new)\n const currentEntries = Array.from(state.entryCache.values()).map((c) => c.entry);\n\n // Calculate tokens before\n const tokensBefore = currentEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Prune to fit budget\n const result = pruner.pruneToFit(currentEntries, budget);\n\n // Calculate tokens after\n const tokensAfter = result.kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Update cache: remove pruned entries\n for (const removed of result.removed) {\n state.entryCache.delete(removed.hash);\n }\n\n // Update document frequency to reflect removal\n if (result.removed.length > 0) {\n updateDocumentFrequency(result.removed, true);\n }\n\n // Record metrics\n const duration = Date.now() - startTime;\n const cacheHitRate = newEntries.length > 0 ? cachedEntries.length / newEntries.length : 0;\n\n getMetrics().recordOptimization({\n timestamp: Date.now(),\n duration,\n tokensBefore,\n tokensAfter,\n entriesProcessed: newEntries.length,\n entriesKept: result.kept.length,\n cacheHitRate,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n return result;\n }\n\n function optimizeFull(\n allEntries: MemoryEntry[],\n budget?: number,\n ): PruneResult & { budgetUtilization: number } {\n const startTime = Date.now();\n\n // Calculate tokens before\n const tokensBefore = allEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Reset state\n state.entryCache.clear();\n state.documentFrequency.clear();\n state.totalDocuments = 0;\n state.updateCount = 0;\n state.lastFullOptimization = Date.now();\n\n // Rebuild document frequency table\n updateDocumentFrequency(allEntries, false);\n\n // Score and cache all entries\n for (const entry of allEntries) {\n const score = pruner.priorityScore(entry, allEntries);\n cacheEntry(entry, score);\n }\n\n // Prune to fit budget\n const result = pruner.pruneToFit(allEntries, budget);\n\n // Calculate tokens after\n const tokensAfter = result.kept.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n // Update cache: remove pruned entries\n for (const removed of result.removed) {\n state.entryCache.delete(removed.hash);\n }\n\n // Update document frequency to reflect removal\n if (result.removed.length > 0) {\n updateDocumentFrequency(result.removed, true);\n }\n\n // Record metrics\n const duration = Date.now() - startTime;\n\n getMetrics().recordOptimization({\n timestamp: Date.now(),\n duration,\n tokensBefore,\n tokensAfter,\n entriesProcessed: allEntries.length,\n entriesKept: result.kept.length,\n cacheHitRate: 0, // Full optimization has no cache hits\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n return result;\n }\n\n function getState(): IncrementalOptimizerState {\n return {\n entryCache: new Map(state.entryCache),\n documentFrequency: new Map(state.documentFrequency),\n totalDocuments: state.totalDocuments,\n updateCount: state.updateCount,\n lastFullOptimization: state.lastFullOptimization,\n };\n }\n\n function restoreState(restoredState: IncrementalOptimizerState): void {\n state = {\n entryCache: new Map(restoredState.entryCache),\n documentFrequency: new Map(restoredState.documentFrequency),\n totalDocuments: restoredState.totalDocuments,\n updateCount: restoredState.updateCount,\n lastFullOptimization: restoredState.lastFullOptimization,\n };\n }\n\n function reset(): void {\n state = {\n entryCache: new Map(),\n documentFrequency: new Map(),\n totalDocuments: 0,\n updateCount: 0,\n lastFullOptimization: Date.now(),\n };\n }\n\n function getStats() {\n return {\n cachedEntries: state.entryCache.size,\n uniqueTerms: state.documentFrequency.size,\n totalDocuments: state.totalDocuments,\n updateCount: state.updateCount,\n lastFullOptimization: state.lastFullOptimization,\n };\n }\n\n return {\n optimizeIncremental,\n optimizeFull,\n getState,\n restoreState,\n reset,\n getStats,\n };\n}\n","/**\n * Streaming Context Pipeline - Real-time sliding window buffer\n *\n * Maintains an optimized context in real-time by:\n * - Ingesting new content as it arrives\n * - Storing entries by priority internally (for eviction decisions)\n * - Outputting in chronological order (for conversation coherence)\n * - Evicting lowest-priority entries when budget exceeded\n * - Using IncrementalOptimizer for fast delta processing\n */\n\nimport type { MemoryEntry } from '../types/memory.js';\nimport { parseClaudeCodeContext } from '../utils/context-parser.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport {\n createIncrementalOptimizer,\n type IncrementalOptimizerConfig,\n} from './incremental-optimizer.js';\n\nexport interface ContextPipelineConfig extends IncrementalOptimizerConfig {\n /** Sliding window size (max entries to keep) */\n windowSize: number;\n}\n\nexport interface ContextPipelineStats {\n /** Total entries ingested */\n totalIngested: number;\n /** Current entry count */\n currentEntries: number;\n /** Current token count */\n currentTokens: number;\n /** Budget utilization (0.0-1.0) */\n budgetUtilization: number;\n /** Evicted entry count */\n evictedEntries: number;\n /** Optimizer stats */\n optimizer: {\n cachedEntries: number;\n uniqueTerms: number;\n updateCount: number;\n };\n}\n\nexport interface ContextPipeline {\n /**\n * Ingest new content into the pipeline\n * @param content - Raw content string\n * @param metadata - Optional metadata to attach to entries\n * @returns Number of entries ingested\n */\n ingest(content: string, metadata?: Record<string, unknown>): number;\n\n /**\n * Get current optimized context (chronologically ordered)\n * @returns Optimized context string\n */\n getContext(): string;\n\n /**\n * Get current entries (chronologically ordered)\n * @returns Array of memory entries\n */\n getEntries(): MemoryEntry[];\n\n /**\n * Get pipeline statistics\n * @returns Pipeline stats\n */\n getStats(): ContextPipelineStats;\n\n /**\n * Clear all entries and reset state\n */\n clear(): void;\n}\n\n/**\n * Create a context pipeline instance\n * @param config - Pipeline configuration\n * @returns ContextPipeline instance\n */\nexport function createContextPipeline(config: ContextPipelineConfig): ContextPipeline {\n const optimizer = createIncrementalOptimizer(config);\n const { windowSize, tokenBudget } = config;\n\n // Internal state\n let totalIngested = 0;\n let evictedEntries = 0;\n let currentEntries: MemoryEntry[] = [];\n let budgetUtilization = 0;\n\n function ingest(content: string, metadata: Record<string, unknown> = {}): number {\n // Parse content into entries\n const newEntries = parseClaudeCodeContext(content);\n\n if (newEntries.length === 0) return 0;\n\n // Attach metadata to entries\n const entriesWithMetadata = newEntries.map((entry) => ({\n ...entry,\n metadata: { ...entry.metadata, ...metadata },\n }));\n\n // Optimize incrementally\n const result = optimizer.optimizeIncremental(entriesWithMetadata, tokenBudget);\n\n // Update statistics\n totalIngested += newEntries.length;\n evictedEntries += result.removed.length;\n currentEntries = result.kept;\n budgetUtilization = result.budgetUtilization;\n\n // Enforce window size limit (keep most recent if exceeded)\n if (currentEntries.length > windowSize) {\n // Sort by timestamp descending (newest first)\n const sorted = [...currentEntries].sort((a, b) => b.timestamp - a.timestamp);\n const toKeep = sorted.slice(0, windowSize);\n const toRemove = sorted.slice(windowSize);\n\n currentEntries = toKeep;\n evictedEntries += toRemove.length;\n }\n\n return newEntries.length;\n }\n\n function getContext(): string {\n // Sort entries chronologically (oldest first)\n const sorted = [...currentEntries].sort((a, b) => a.timestamp - b.timestamp);\n return sorted.map((e) => e.content).join('\\n\\n');\n }\n\n function getEntries(): MemoryEntry[] {\n // Return entries chronologically (oldest first)\n return [...currentEntries].sort((a, b) => a.timestamp - b.timestamp);\n }\n\n function getStats(): ContextPipelineStats {\n const optimizerStats = optimizer.getStats();\n const currentTokens = currentEntries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n totalIngested,\n currentEntries: currentEntries.length,\n currentTokens,\n budgetUtilization,\n evictedEntries,\n optimizer: {\n cachedEntries: optimizerStats.cachedEntries,\n uniqueTerms: optimizerStats.uniqueTerms,\n updateCount: optimizerStats.updateCount,\n },\n };\n }\n\n function clear(): void {\n totalIngested = 0;\n evictedEntries = 0;\n currentEntries = [];\n budgetUtilization = 0;\n optimizer.reset();\n }\n\n return {\n ingest,\n getContext,\n getEntries,\n getStats,\n clear,\n };\n}\n","/**\n * KV Memory Store Module\n * Implements hippocampal key-value storage with dual index/value tables.\n * Maps to: Hippocampal Key-Value — the hippocampus separates what to store from how to retrieve it.\n */\n\nimport { copyFileSync, existsSync } from 'node:fs';\nimport Database from 'better-sqlite3';\nimport type { MemoryEntry, MemoryQueryFilters } from '../types/memory.js';\n\n/**\n * Optimization statistics record.\n */\nexport interface OptimizationStats {\n id: number;\n timestamp: number;\n tokens_before: number;\n tokens_after: number;\n entries_pruned: number;\n duration_ms: number;\n}\n\n/**\n * KV Memory interface.\n */\nexport interface KVMemory {\n /** Store a memory entry */\n put(entry: MemoryEntry): Promise<void>;\n\n /** Retrieve a memory entry by ID */\n get(id: string): Promise<MemoryEntry | null>;\n\n /** Query entries by filters */\n query(filters: MemoryQueryFilters): Promise<MemoryEntry[]>;\n\n /** Delete a memory entry */\n delete(id: string): Promise<void>;\n\n /** List all entry IDs */\n list(): Promise<string[]>;\n\n /** Compact database (remove expired entries) */\n compact(): Promise<number>;\n\n /** Close database connection */\n close(): Promise<void>;\n\n /** Record optimization statistics */\n recordOptimization(stats: Omit<OptimizationStats, 'id'>): Promise<void>;\n\n /** Get all optimization statistics */\n getOptimizationStats(): Promise<OptimizationStats[]>;\n\n /** Clear all optimization statistics */\n clearOptimizationStats(): Promise<void>;\n}\n\n/**\n * Create a timestamped backup of the database\n * @param dbPath - Path to database file\n * @returns Path to backup file\n */\nfunction createBackup(dbPath: string): string {\n const timestamp = new Date().toISOString().replace(/[:.]/g, '-');\n const backupPath = `${dbPath}.backup-${timestamp}`;\n\n try {\n copyFileSync(dbPath, backupPath);\n console.log(`✓ Database backed up to: ${backupPath}`);\n return backupPath;\n } catch (error) {\n console.error(`Warning: Could not create backup: ${error}`);\n return '';\n }\n}\n\n/**\n * Create KV Memory store with SQLite backend.\n *\n * Initializes database with dual table schema:\n * - entries_index: Fast lookups (id, hash, timestamp, score, ttl, state, accessCount, isBTSP)\n * - entries_value: Content storage (id, content, tags, metadata)\n *\n * @param dbPath - Path to SQLite database file\n * @returns KVMemory instance\n */\nexport async function createKVMemory(dbPath: string): Promise<KVMemory> {\n // Detect database corruption and create backup\n let db: Database.Database;\n try {\n db = new Database(dbPath);\n\n // Quick integrity check\n const integrityCheck = db.pragma('quick_check', { simple: true });\n if (integrityCheck !== 'ok') {\n console.error('⚠ Database corruption detected!');\n\n // Create backup before attempting recovery\n if (existsSync(dbPath)) {\n const backupPath = createBackup(dbPath);\n if (backupPath) {\n console.log(`Backup created at: ${backupPath}`);\n }\n }\n\n // Try to recover\n console.log('Attempting database recovery...');\n db.close();\n db = new Database(dbPath);\n }\n } catch (error) {\n console.error('⚠ Database error detected:', error);\n\n // Create backup if database exists\n if (existsSync(dbPath)) {\n createBackup(dbPath);\n console.log('Creating new database...');\n }\n\n db = new Database(dbPath);\n }\n\n // Enable WAL mode for better concurrency\n db.pragma('journal_mode = WAL');\n\n // Create entries_index table\n db.exec(`\n CREATE TABLE IF NOT EXISTS entries_index (\n id TEXT PRIMARY KEY NOT NULL,\n hash TEXT UNIQUE NOT NULL,\n timestamp INTEGER NOT NULL,\n score REAL NOT NULL DEFAULT 0.0 CHECK(score >= 0.0 AND score <= 1.0),\n ttl INTEGER NOT NULL CHECK(ttl >= 0),\n state TEXT NOT NULL CHECK(state IN ('silent', 'ready', 'active')),\n accessCount INTEGER NOT NULL DEFAULT 0 CHECK(accessCount >= 0),\n isBTSP INTEGER NOT NULL DEFAULT 0 CHECK(isBTSP IN (0, 1)),\n created_at INTEGER NOT NULL DEFAULT (strftime('%s', 'now'))\n );\n `);\n\n // Create entries_value table\n db.exec(`\n CREATE TABLE IF NOT EXISTS entries_value (\n id TEXT PRIMARY KEY NOT NULL,\n content TEXT NOT NULL,\n tags TEXT,\n metadata TEXT,\n FOREIGN KEY (id) REFERENCES entries_index(id) ON DELETE CASCADE\n );\n `);\n\n // Create optimization_stats table\n db.exec(`\n CREATE TABLE IF NOT EXISTS optimization_stats (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n timestamp INTEGER NOT NULL DEFAULT (strftime('%s', 'now')),\n tokens_before INTEGER NOT NULL,\n tokens_after INTEGER NOT NULL,\n entries_pruned INTEGER NOT NULL,\n duration_ms INTEGER NOT NULL\n );\n `);\n\n // Create indexes\n db.exec(`\n CREATE INDEX IF NOT EXISTS idx_entries_state ON entries_index(state);\n CREATE INDEX IF NOT EXISTS idx_entries_score ON entries_index(score DESC);\n CREATE INDEX IF NOT EXISTS idx_entries_hash ON entries_index(hash);\n CREATE INDEX IF NOT EXISTS idx_entries_timestamp ON entries_index(timestamp DESC);\n CREATE INDEX IF NOT EXISTS idx_stats_timestamp ON optimization_stats(timestamp DESC);\n `);\n\n // Prepare statements for better performance\n const putIndexStmt = db.prepare(`\n INSERT OR REPLACE INTO entries_index\n (id, hash, timestamp, score, ttl, state, accessCount, isBTSP)\n VALUES (?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n const putValueStmt = db.prepare(`\n INSERT OR REPLACE INTO entries_value\n (id, content, tags, metadata)\n VALUES (?, ?, ?, ?)\n `);\n\n const getStmt = db.prepare(`\n SELECT\n i.id, i.hash, i.timestamp, i.score, i.ttl, i.state, i.accessCount, i.isBTSP,\n v.content, v.tags, v.metadata\n FROM entries_index i\n JOIN entries_value v ON i.id = v.id\n WHERE i.id = ?\n `);\n\n const deleteIndexStmt = db.prepare('DELETE FROM entries_index WHERE id = ?');\n const deleteValueStmt = db.prepare('DELETE FROM entries_value WHERE id = ?');\n\n return {\n async put(entry: MemoryEntry): Promise<void> {\n const transaction = db.transaction(() => {\n putIndexStmt.run(\n entry.id,\n entry.hash,\n entry.timestamp,\n entry.score,\n entry.ttl,\n entry.state,\n entry.accessCount,\n entry.isBTSP ? 1 : 0,\n );\n\n putValueStmt.run(\n entry.id,\n entry.content,\n JSON.stringify(entry.tags),\n JSON.stringify(entry.metadata),\n );\n });\n\n transaction();\n },\n\n async get(id: string): Promise<MemoryEntry | null> {\n const row = getStmt.get(id) as unknown;\n\n if (!row) {\n return null;\n }\n\n const r = row as {\n id: string;\n hash: string;\n timestamp: number;\n score: number;\n ttl: number;\n state: string;\n accessCount: number;\n isBTSP: number;\n content: string;\n tags: string | null;\n metadata: string | null;\n };\n\n return {\n id: r.id,\n content: r.content,\n hash: r.hash,\n timestamp: r.timestamp,\n score: r.score,\n ttl: r.ttl,\n state: r.state as 'silent' | 'ready' | 'active',\n accessCount: r.accessCount,\n tags: r.tags ? JSON.parse(r.tags) : [],\n metadata: r.metadata ? JSON.parse(r.metadata) : {},\n isBTSP: r.isBTSP === 1,\n };\n },\n\n async query(filters: MemoryQueryFilters): Promise<MemoryEntry[]> {\n let sql = `\n SELECT\n i.id, i.hash, i.timestamp, i.score, i.ttl, i.state, i.accessCount, i.isBTSP,\n v.content, v.tags, v.metadata\n FROM entries_index i\n JOIN entries_value v ON i.id = v.id\n WHERE 1=1\n `;\n\n const params: unknown[] = [];\n\n if (filters.state) {\n sql += ' AND i.state = ?';\n params.push(filters.state);\n }\n\n if (filters.minScore !== undefined) {\n sql += ' AND i.score >= ?';\n params.push(filters.minScore);\n }\n\n if (filters.maxScore !== undefined) {\n sql += ' AND i.score <= ?';\n params.push(filters.maxScore);\n }\n\n if (filters.isBTSP !== undefined) {\n sql += ' AND i.isBTSP = ?';\n params.push(filters.isBTSP ? 1 : 0);\n }\n\n sql += ' ORDER BY i.score DESC';\n\n if (filters.limit) {\n sql += ' LIMIT ?';\n params.push(filters.limit);\n }\n\n if (filters.offset) {\n sql += ' OFFSET ?';\n params.push(filters.offset);\n }\n\n const stmt = db.prepare(sql);\n const rows = stmt.all(...params) as unknown[];\n\n return rows.map((row) => {\n const r = row as {\n id: string;\n hash: string;\n timestamp: number;\n score: number;\n ttl: number;\n state: string;\n accessCount: number;\n isBTSP: number;\n content: string;\n tags: string | null;\n metadata: string | null;\n };\n\n return {\n id: r.id,\n content: r.content,\n hash: r.hash,\n timestamp: r.timestamp,\n score: r.score,\n ttl: r.ttl,\n state: r.state as 'silent' | 'ready' | 'active',\n accessCount: r.accessCount,\n tags: r.tags ? JSON.parse(r.tags) : [],\n metadata: r.metadata ? JSON.parse(r.metadata) : {},\n isBTSP: r.isBTSP === 1,\n };\n });\n },\n\n async delete(id: string): Promise<void> {\n const transaction = db.transaction(() => {\n deleteIndexStmt.run(id);\n deleteValueStmt.run(id);\n });\n\n transaction();\n },\n\n async list(): Promise<string[]> {\n const stmt = db.prepare('SELECT id FROM entries_index');\n const rows = stmt.all() as { id: string }[];\n return rows.map((r) => r.id);\n },\n\n async compact(): Promise<number> {\n const before = db.prepare('SELECT COUNT(*) as count FROM entries_index').get() as {\n count: number;\n };\n\n // Remove fully decayed entries (this will be enhanced in sleep-compressor)\n db.exec('DELETE FROM entries_index WHERE ttl <= 0');\n\n db.exec('VACUUM');\n\n const after = db.prepare('SELECT COUNT(*) as count FROM entries_index').get() as {\n count: number;\n };\n\n return before.count - after.count;\n },\n\n async close(): Promise<void> {\n db.close();\n },\n\n async recordOptimization(stats: Omit<OptimizationStats, 'id'>): Promise<void> {\n const stmt = db.prepare(`\n INSERT INTO optimization_stats (timestamp, tokens_before, tokens_after, entries_pruned, duration_ms)\n VALUES (?, ?, ?, ?, ?)\n `);\n\n stmt.run(\n stats.timestamp,\n stats.tokens_before,\n stats.tokens_after,\n stats.entries_pruned,\n stats.duration_ms,\n );\n },\n\n async getOptimizationStats(): Promise<OptimizationStats[]> {\n const stmt = db.prepare(`\n SELECT id, timestamp, tokens_before, tokens_after, entries_pruned, duration_ms\n FROM optimization_stats\n ORDER BY timestamp DESC\n `);\n\n const rows = stmt.all() as OptimizationStats[];\n return rows;\n },\n\n async clearOptimizationStats(): Promise<void> {\n db.exec('DELETE FROM optimization_stats');\n },\n };\n}\n","/**\n * Sleep Compressor - Implements sleep replay principle\n *\n * Neuroscience: During sleep, the brain consolidates memories by replaying important ones\n * and discarding irrelevant information.\n * Application: Periodic consolidation removes decayed entries and merges duplicates.\n */\n\nimport type { ConsolidateResult, DuplicateGroup } from '../types/consolidate.js';\nimport type { MemoryEntry } from '../types/memory.js';\nimport { createEngramScorer } from './engram-scorer.js';\n\nexport interface SleepCompressor {\n /**\n * Consolidate entries: remove decayed, merge duplicates\n * @param entries - All memory entries\n * @returns Consolidation result\n */\n consolidate(entries: MemoryEntry[]): ConsolidateResult;\n\n /**\n * Find duplicate entries (exact hash or near-duplicate by similarity)\n * @param entries - Memory entries\n * @returns Groups of duplicates\n */\n findDuplicates(entries: MemoryEntry[]): DuplicateGroup[];\n\n /**\n * Merge duplicate entries, keeping highest score\n * @param groups - Duplicate groups\n * @returns Merged entries\n */\n mergeDuplicates(groups: DuplicateGroup[]): MemoryEntry[];\n}\n\n/**\n * Create a sleep compressor instance\n * @returns SleepCompressor instance\n */\nexport function createSleepCompressor(): SleepCompressor {\n const scorer = createEngramScorer({ defaultTTL: 24, decayThreshold: 0.95 });\n\n function consolidate(entries: MemoryEntry[]): ConsolidateResult {\n const startTime = Date.now();\n const originalCount = entries.length;\n\n // Step 1: Remove fully decayed entries (decay ≥ 0.95)\n const now = Date.now();\n const nonDecayed = entries.filter((entry) => {\n const ageInSeconds = (now - entry.timestamp) / 1000;\n const decay = scorer.calculateDecay(ageInSeconds, entry.ttl);\n return decay < 0.95; // Keep entries with decay < 0.95\n });\n\n const decayedRemoved = originalCount - nonDecayed.length;\n\n // Step 2: Find and merge duplicates\n const duplicateGroups = findDuplicates(nonDecayed);\n const merged = mergeDuplicates(duplicateGroups);\n\n // Step 3: Keep non-duplicates\n const duplicateIds = new Set(duplicateGroups.flatMap((g) => g.entries.map((e) => e.id)));\n const nonDuplicates = nonDecayed.filter((e) => !duplicateIds.has(e.id));\n\n // Combine merged duplicates with non-duplicates\n const kept = [...merged, ...nonDuplicates];\n const removed = entries.filter((e) => !kept.some((k) => k.id === e.id));\n\n const duplicatesRemoved = duplicateGroups.reduce((sum, g) => sum + (g.entries.length - 1), 0);\n\n return {\n kept,\n removed,\n entriesBefore: originalCount,\n entriesAfter: kept.length,\n decayedRemoved,\n duplicatesRemoved,\n compressionRatio: originalCount > 0 ? kept.length / originalCount : 0,\n durationMs: Date.now() - startTime,\n };\n }\n\n function findDuplicates(entries: MemoryEntry[]): DuplicateGroup[] {\n const groups: DuplicateGroup[] = [];\n const processed = new Set<string>();\n\n // Find exact hash matches\n for (let i = 0; i < entries.length; i++) {\n const entry = entries[i];\n if (!entry || processed.has(entry.id)) continue;\n\n const duplicates = entries.filter((e, idx) => idx !== i && e.hash === entry.hash);\n\n if (duplicates.length > 0) {\n const group: DuplicateGroup = {\n entries: [entry, ...duplicates],\n similarity: 1.0, // Exact match\n };\n groups.push(group);\n\n // Mark as processed\n processed.add(entry.id);\n for (const dup of duplicates) {\n processed.add(dup.id);\n }\n }\n }\n\n // Find near-duplicates (cosine similarity ≥ 0.85)\n for (let i = 0; i < entries.length; i++) {\n const entryI = entries[i];\n if (!entryI || processed.has(entryI.id)) continue;\n\n for (let j = i + 1; j < entries.length; j++) {\n const entryJ = entries[j];\n if (!entryJ || processed.has(entryJ.id)) continue;\n\n const similarity = cosineSimilarity(entryI.content, entryJ.content);\n\n if (similarity >= 0.85) {\n const group: DuplicateGroup = {\n entries: [entryI, entryJ],\n similarity,\n };\n groups.push(group);\n\n processed.add(entryI.id);\n processed.add(entryJ.id);\n break; // Move to next i\n }\n }\n }\n\n return groups;\n }\n\n function mergeDuplicates(groups: DuplicateGroup[]): MemoryEntry[] {\n const merged: MemoryEntry[] = [];\n\n for (const group of groups) {\n // Keep entry with highest score\n const sorted = [...group.entries].sort((a, b) => b.score - a.score);\n const best = sorted[0];\n if (!best) continue; // Skip empty groups\n\n // Sum access counts\n const totalAccessCount = group.entries.reduce((sum, e) => sum + e.accessCount, 0);\n\n // Merge tags\n const allTags = new Set(group.entries.flatMap((e) => e.tags));\n\n merged.push({\n ...best,\n accessCount: totalAccessCount,\n tags: Array.from(allTags),\n });\n }\n\n return merged;\n }\n\n /**\n * Calculate cosine similarity between two text strings\n * @param text1 - First text\n * @param text2 - Second text\n * @returns Similarity score (0.0-1.0)\n */\n function cosineSimilarity(text1: string, text2: string): number {\n const words1 = tokenize(text1);\n const words2 = tokenize(text2);\n\n // Build vocabulary\n const vocab = new Set([...words1, ...words2]);\n\n // Build word frequency vectors\n const vec1: Record<string, number> = {};\n const vec2: Record<string, number> = {};\n\n for (const word of vocab) {\n vec1[word] = words1.filter((w) => w === word).length;\n vec2[word] = words2.filter((w) => w === word).length;\n }\n\n // Calculate dot product and magnitudes\n let dotProduct = 0;\n let mag1 = 0;\n let mag2 = 0;\n\n for (const word of vocab) {\n const count1 = vec1[word] ?? 0;\n const count2 = vec2[word] ?? 0;\n dotProduct += count1 * count2;\n mag1 += count1 * count1;\n mag2 += count2 * count2;\n }\n\n mag1 = Math.sqrt(mag1);\n mag2 = Math.sqrt(mag2);\n\n if (mag1 === 0 || mag2 === 0) return 0;\n\n return dotProduct / (mag1 * mag2);\n }\n\n function tokenize(text: string): string[] {\n return text\n .toLowerCase()\n .split(/\\s+/)\n .filter((word) => word.length > 0);\n }\n\n return {\n consolidate,\n findDuplicates,\n mergeDuplicates,\n };\n}\n","/**\n * Daemon Process Manager - Background process lifecycle management\n *\n * Handles:\n * - Process forking and detachment\n * - PID file management\n * - Signal handling (SIGTERM, SIGINT)\n * - Daemon start/stop/status commands\n */\n\nimport { fork } from 'node:child_process';\nimport { existsSync, mkdirSync, readFileSync, unlinkSync, writeFileSync } from 'node:fs';\nimport { dirname, join } from 'node:path';\nimport { fileURLToPath } from 'node:url';\nimport { getMetrics } from '../core/metrics.js';\nimport type { SparnConfig } from '../types/config.js';\n\nexport interface DaemonCommand {\n /** Start the daemon */\n start(config: SparnConfig): Promise<DaemonStartResult>;\n\n /** Stop the daemon */\n stop(config: SparnConfig): Promise<DaemonStopResult>;\n\n /** Get daemon status */\n status(config: SparnConfig): Promise<DaemonStatusResult>;\n}\n\nexport interface DaemonStartResult {\n success: boolean;\n pid?: number;\n message: string;\n error?: string;\n}\n\nexport interface DaemonStopResult {\n success: boolean;\n message: string;\n error?: string;\n}\n\nexport interface DaemonStatusResult {\n running: boolean;\n pid?: number;\n uptime?: number;\n sessionsWatched?: number;\n tokensSaved?: number;\n message: string;\n}\n\n/**\n * Create daemon command interface\n * @returns DaemonCommand instance\n */\nexport function createDaemonCommand(): DaemonCommand {\n /**\n * Check if daemon is running\n */\n function isDaemonRunning(pidFile: string): { running: boolean; pid?: number } {\n if (!existsSync(pidFile)) {\n return { running: false };\n }\n\n try {\n const pidStr = readFileSync(pidFile, 'utf-8').trim();\n const pid = Number.parseInt(pidStr, 10);\n\n if (Number.isNaN(pid)) {\n return { running: false };\n }\n\n // Check if process exists (cross-platform)\n try {\n process.kill(pid, 0); // Signal 0 checks existence without killing\n return { running: true, pid };\n } catch {\n // Process doesn't exist, clean up stale PID file\n unlinkSync(pidFile);\n return { running: false };\n }\n } catch {\n return { running: false };\n }\n }\n\n /**\n * Write PID file\n */\n function writePidFile(pidFile: string, pid: number): void {\n // Ensure directory exists\n const dir = dirname(pidFile);\n if (!existsSync(dir)) {\n mkdirSync(dir, { recursive: true });\n }\n\n writeFileSync(pidFile, String(pid), 'utf-8');\n }\n\n /**\n * Remove PID file\n */\n function removePidFile(pidFile: string): void {\n if (existsSync(pidFile)) {\n unlinkSync(pidFile);\n }\n }\n\n async function start(config: SparnConfig): Promise<DaemonStartResult> {\n const { pidFile, logFile } = config.realtime;\n\n // Check if already running\n const status = isDaemonRunning(pidFile);\n if (status.running) {\n return {\n success: false,\n pid: status.pid,\n message: `Daemon already running (PID ${status.pid})`,\n error: 'Already running',\n };\n }\n\n try {\n // Fork child process (daemon entry point)\n const __filename = fileURLToPath(import.meta.url);\n const __dirname = dirname(__filename);\n const daemonPath = join(__dirname, 'index.js');\n\n const child = fork(daemonPath, [], {\n detached: true,\n stdio: 'ignore',\n env: {\n ...process.env,\n SPARN_CONFIG: JSON.stringify(config),\n SPARN_PID_FILE: pidFile,\n SPARN_LOG_FILE: logFile,\n },\n });\n\n // Detach from parent\n child.unref();\n\n // Write PID file\n if (child.pid) {\n writePidFile(pidFile, child.pid);\n\n return {\n success: true,\n pid: child.pid,\n message: `Daemon started (PID ${child.pid})`,\n };\n }\n\n return {\n success: false,\n message: 'Failed to start daemon (no PID)',\n error: 'No PID',\n };\n } catch (error) {\n return {\n success: false,\n message: 'Failed to start daemon',\n error: error instanceof Error ? error.message : String(error),\n };\n }\n }\n\n async function stop(config: SparnConfig): Promise<DaemonStopResult> {\n const { pidFile } = config.realtime;\n\n const status = isDaemonRunning(pidFile);\n\n if (!status.running || !status.pid) {\n return {\n success: true,\n message: 'Daemon not running',\n };\n }\n\n try {\n // Send SIGTERM\n process.kill(status.pid, 'SIGTERM');\n\n // Wait for process to exit (timeout after 5s)\n const maxWait = 5000;\n const interval = 100;\n let waited = 0;\n\n while (waited < maxWait) {\n try {\n process.kill(status.pid, 0);\n // Still running, wait\n await new Promise((resolve) => setTimeout(resolve, interval));\n waited += interval;\n } catch {\n // Process exited\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon stopped (PID ${status.pid})`,\n };\n }\n }\n\n // Timeout, force kill\n try {\n process.kill(status.pid, 'SIGKILL');\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon force killed (PID ${status.pid})`,\n };\n } catch {\n removePidFile(pidFile);\n return {\n success: true,\n message: `Daemon stopped (PID ${status.pid})`,\n };\n }\n } catch (error) {\n return {\n success: false,\n message: 'Failed to stop daemon',\n error: error instanceof Error ? error.message : String(error),\n };\n }\n }\n\n async function status(config: SparnConfig): Promise<DaemonStatusResult> {\n const { pidFile } = config.realtime;\n\n const daemonStatus = isDaemonRunning(pidFile);\n\n if (!daemonStatus.running || !daemonStatus.pid) {\n return {\n running: false,\n message: 'Daemon not running',\n };\n }\n\n // Get metrics snapshot\n const metrics = getMetrics().getSnapshot();\n\n return {\n running: true,\n pid: daemonStatus.pid,\n uptime: metrics.daemon.uptime,\n sessionsWatched: metrics.daemon.sessionsWatched,\n tokensSaved: metrics.optimization.totalTokensSaved,\n message: `Daemon running (PID ${daemonStatus.pid})`,\n };\n }\n\n return {\n start,\n stop,\n status,\n };\n}\n","/**\n * File Tracker - Incremental file reading with byte position tracking\n *\n * Tracks read positions for files to enable efficient incremental reading.\n * Handles JSONL partial line buffering for incomplete writes.\n *\n * Use case: Monitor Claude Code session JSONL files and only read new lines\n * as they're appended, without re-reading the entire file.\n */\n\nimport { readFileSync, statSync } from 'node:fs';\n\nexport interface FilePosition {\n /** File path */\n path: string;\n /** Last read byte position */\n position: number;\n /** Partial line buffer (for JSONL incomplete writes) */\n partialLine: string;\n /** Last modification time */\n lastModified: number;\n /** File size at last read */\n lastSize: number;\n}\n\nexport interface FileTracker {\n /**\n * Read new content from file since last read\n * @param filePath - File to read\n * @returns New content as array of lines (empty if no new content)\n */\n readNewLines(filePath: string): string[];\n\n /**\n * Get current position for a file\n * @param filePath - File path\n * @returns File position or null if not tracked\n */\n getPosition(filePath: string): FilePosition | null;\n\n /**\n * Reset position for a file (start from beginning on next read)\n * @param filePath - File path\n */\n resetPosition(filePath: string): void;\n\n /**\n * Clear all tracked positions\n */\n clearAll(): void;\n\n /**\n * Get all tracked file paths\n * @returns Array of tracked file paths\n */\n getTrackedFiles(): string[];\n}\n\n/**\n * Create a file tracker instance\n * @returns FileTracker instance\n */\nexport function createFileTracker(): FileTracker {\n // Track positions by file path\n const positions = new Map<string, FilePosition>();\n\n function readNewLines(filePath: string): string[] {\n try {\n // Get current file stats\n const stats = statSync(filePath);\n const currentSize = stats.size;\n const currentModified = stats.mtimeMs;\n\n // Get or initialize position\n let pos = positions.get(filePath);\n\n if (!pos) {\n // First read: start from beginning\n pos = {\n path: filePath,\n position: 0,\n partialLine: '',\n lastModified: currentModified,\n lastSize: 0,\n };\n positions.set(filePath, pos);\n }\n\n // Check if file was truncated or is same size\n if (currentSize < pos.lastSize || currentSize === pos.position) {\n // File truncated or no new content\n if (currentSize < pos.lastSize) {\n // Reset position if truncated\n pos.position = 0;\n pos.partialLine = '';\n }\n return [];\n }\n\n // Read new content from last position\n const buffer = Buffer.alloc(currentSize - pos.position);\n const fd = readFileSync(filePath);\n fd.copy(buffer, 0, pos.position, currentSize);\n\n // Convert to string and combine with partial line\n const newContent = (pos.partialLine + buffer.toString('utf-8')).split('\\n');\n\n // Last element might be incomplete (no trailing newline yet)\n const partialLine = newContent.pop() || '';\n\n // Update position\n pos.position = currentSize;\n pos.partialLine = partialLine;\n pos.lastModified = currentModified;\n pos.lastSize = currentSize;\n\n // Return complete lines (filter empty)\n return newContent.filter((line) => line.trim().length > 0);\n } catch (_error) {\n // File doesn't exist or can't be read\n // Return empty array (fail silently for watcher use case)\n return [];\n }\n }\n\n function getPosition(filePath: string): FilePosition | null {\n return positions.get(filePath) || null;\n }\n\n function resetPosition(filePath: string): void {\n positions.delete(filePath);\n }\n\n function clearAll(): void {\n positions.clear();\n }\n\n function getTrackedFiles(): string[] {\n return Array.from(positions.keys());\n }\n\n return {\n readNewLines,\n getPosition,\n resetPosition,\n clearAll,\n getTrackedFiles,\n };\n}\n","/**\n * Session Watcher - Monitor Claude Code session files for changes\n *\n * Uses Node.js fs.watch to monitor ~/.claude/projects/**\\/*.jsonl files.\n * Debounces events and triggers optimization when token threshold exceeded.\n * Maintains per-session ContextPipeline instances.\n */\n\nimport { type FSWatcher, readdirSync, statSync, watch } from 'node:fs';\nimport { homedir } from 'node:os';\nimport { dirname, join } from 'node:path';\nimport { type ContextPipeline, createContextPipeline } from '../core/context-pipeline.js';\nimport { getMetrics } from '../core/metrics.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { estimateTokens } from '../utils/tokenizer.js';\nimport { createFileTracker } from './file-tracker.js';\n\nexport interface SessionWatcherConfig {\n /** Sparn configuration */\n config: SparnConfig;\n /** Callback when optimization triggered */\n onOptimize?: (sessionId: string, stats: SessionStats) => void;\n /** Callback on error */\n onError?: (error: Error) => void;\n}\n\nexport interface SessionStats {\n /** Session ID */\n sessionId: string;\n /** Total tokens ingested */\n totalTokens: number;\n /** Current optimized tokens */\n optimizedTokens: number;\n /** Reduction percentage */\n reduction: number;\n /** Entry count */\n entryCount: number;\n /** Budget utilization */\n budgetUtilization: number;\n}\n\nexport interface SessionWatcher {\n /**\n * Start watching Claude Code session files\n * @returns Promise that resolves when watcher is ready\n */\n start(): Promise<void>;\n\n /**\n * Stop watching and cleanup\n */\n stop(): void;\n\n /**\n * Get statistics for all sessions\n * @returns Array of session stats\n */\n getStats(): SessionStats[];\n\n /**\n * Get statistics for a specific session\n * @param sessionId - Session ID\n * @returns Session stats or null if not found\n */\n getSessionStats(sessionId: string): SessionStats | null;\n\n /**\n * Manually trigger optimization for a session\n * @param sessionId - Session ID\n */\n optimizeSession(sessionId: string): void;\n}\n\n/**\n * Create a session watcher instance\n * @param config - Watcher configuration\n * @returns SessionWatcher instance\n */\nexport function createSessionWatcher(config: SessionWatcherConfig): SessionWatcher {\n const { config: sparnConfig, onOptimize, onError } = config;\n const { realtime, decay, states } = sparnConfig;\n\n // Per-session pipelines and trackers\n const pipelines = new Map<string, ContextPipeline>();\n const fileTracker = createFileTracker();\n\n // File system watchers\n const watchers: FSWatcher[] = [];\n\n // Debounce timers per file\n const debounceTimers = new Map<string, NodeJS.Timeout>();\n\n /**\n * Get Claude Code projects directory\n */\n function getProjectsDir(): string {\n return join(homedir(), '.claude', 'projects');\n }\n\n /**\n * Extract session ID from file path\n * Example: ~/.claude/projects/my-project/abc123.jsonl -> abc123\n */\n function getSessionId(filePath: string): string {\n const filename = filePath.split(/[/\\\\]/).pop() || '';\n return filename.replace(/\\.jsonl$/, '');\n }\n\n /**\n * Get or create pipeline for session\n */\n function getPipeline(sessionId: string): ContextPipeline {\n let pipeline = pipelines.get(sessionId);\n\n if (!pipeline) {\n pipeline = createContextPipeline({\n tokenBudget: realtime.tokenBudget,\n decay,\n states,\n windowSize: realtime.windowSize,\n fullOptimizationInterval: 50, // Full re-optimization every 50 incremental updates\n });\n pipelines.set(sessionId, pipeline);\n }\n\n return pipeline;\n }\n\n /**\n * Handle file change event (debounced)\n */\n function handleFileChange(filePath: string): void {\n // Clear existing timer\n const existingTimer = debounceTimers.get(filePath);\n if (existingTimer) {\n clearTimeout(existingTimer);\n }\n\n // Set new debounced timer\n const timer = setTimeout(() => {\n try {\n // Read new lines from file\n const newLines = fileTracker.readNewLines(filePath);\n\n if (newLines.length === 0) return;\n\n // Parse JSONL content\n const content = newLines.join('\\n');\n const sessionId = getSessionId(filePath);\n const pipeline = getPipeline(sessionId);\n\n // Ingest into pipeline\n pipeline.ingest(content, { sessionId, filePath });\n\n // Check if we should trigger optimization\n const stats = pipeline.getStats();\n if (stats.currentTokens >= realtime.autoOptimizeThreshold) {\n // Update daemon metrics\n getMetrics().updateDaemon({\n sessionsWatched: pipelines.size,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n\n // Trigger optimization callback\n if (onOptimize) {\n const sessionStats = computeSessionStats(sessionId, pipeline);\n onOptimize(sessionId, sessionStats);\n }\n }\n } catch (error) {\n if (onError) {\n onError(error instanceof Error ? error : new Error(String(error)));\n }\n } finally {\n debounceTimers.delete(filePath);\n }\n }, realtime.debounceMs);\n\n debounceTimers.set(filePath, timer);\n }\n\n /**\n * Recursively find all JSONL files in directory\n */\n function findJsonlFiles(dir: string): string[] {\n const files: string[] = [];\n\n try {\n const entries = readdirSync(dir);\n\n for (const entry of entries) {\n const fullPath = join(dir, entry);\n const stat = statSync(fullPath);\n\n if (stat.isDirectory()) {\n // Recurse into subdirectories\n files.push(...findJsonlFiles(fullPath));\n } else if (entry.endsWith('.jsonl')) {\n // Match pattern\n const matches = realtime.watchPatterns.some((pattern) => {\n // Simple glob matching (supports **/*.jsonl)\n const regex = new RegExp(\n pattern.replace(/\\*\\*/g, '.*').replace(/\\*/g, '[^/\\\\\\\\]*').replace(/\\./g, '\\\\.'),\n );\n return regex.test(fullPath);\n });\n\n if (matches) {\n files.push(fullPath);\n }\n }\n }\n } catch (_error) {\n // Ignore errors (directory might not exist yet)\n }\n\n return files;\n }\n\n /**\n * Compute session statistics\n */\n function computeSessionStats(sessionId: string, pipeline: ContextPipeline): SessionStats {\n const stats = pipeline.getStats();\n const entries = pipeline.getEntries();\n const totalTokens = entries.reduce((sum, e) => sum + estimateTokens(e.content), 0);\n\n return {\n sessionId,\n totalTokens: stats.totalIngested,\n optimizedTokens: stats.currentTokens,\n reduction: totalTokens > 0 ? (totalTokens - stats.currentTokens) / totalTokens : 0,\n entryCount: stats.currentEntries,\n budgetUtilization: stats.budgetUtilization,\n };\n }\n\n async function start(): Promise<void> {\n const projectsDir = getProjectsDir();\n\n // Find all existing JSONL files\n const jsonlFiles = findJsonlFiles(projectsDir);\n\n // Watch each file's parent directory (fs.watch is directory-based)\n const watchedDirs = new Set<string>();\n\n for (const file of jsonlFiles) {\n const dir = dirname(file);\n\n if (!watchedDirs.has(dir)) {\n const watcher = watch(dir, { recursive: false }, (_eventType, filename) => {\n if (filename?.endsWith('.jsonl')) {\n const fullPath = join(dir, filename);\n handleFileChange(fullPath);\n }\n });\n\n watchers.push(watcher);\n watchedDirs.add(dir);\n }\n }\n\n // Also watch projects directory for new subdirectories\n const projectsWatcher = watch(projectsDir, { recursive: true }, (_eventType, filename) => {\n if (filename?.endsWith('.jsonl')) {\n const fullPath = join(projectsDir, filename);\n handleFileChange(fullPath);\n }\n });\n\n watchers.push(projectsWatcher);\n\n // Update daemon metrics\n getMetrics().updateDaemon({\n startTime: Date.now(),\n sessionsWatched: jsonlFiles.length,\n memoryUsage: process.memoryUsage().heapUsed,\n });\n }\n\n function stop(): void {\n // Close all watchers\n for (const watcher of watchers) {\n watcher.close();\n }\n watchers.length = 0;\n\n // Clear all timers\n for (const timer of debounceTimers.values()) {\n clearTimeout(timer);\n }\n debounceTimers.clear();\n\n // Clear pipelines\n pipelines.clear();\n\n // Clear file tracker\n fileTracker.clearAll();\n }\n\n function getStats(): SessionStats[] {\n const stats: SessionStats[] = [];\n\n for (const [sessionId, pipeline] of pipelines.entries()) {\n stats.push(computeSessionStats(sessionId, pipeline));\n }\n\n return stats;\n }\n\n function getSessionStats(sessionId: string): SessionStats | null {\n const pipeline = pipelines.get(sessionId);\n if (!pipeline) return null;\n\n return computeSessionStats(sessionId, pipeline);\n }\n\n function optimizeSession(sessionId: string): void {\n const pipeline = pipelines.get(sessionId);\n if (!pipeline) return;\n\n // Get entries and force full optimization\n const entries = pipeline.getEntries();\n pipeline.clear();\n pipeline.ingest(entries.map((e) => e.content).join('\\n\\n'));\n\n // Trigger callback\n if (onOptimize) {\n const stats = computeSessionStats(sessionId, pipeline);\n onOptimize(sessionId, stats);\n }\n }\n\n return {\n start,\n stop,\n getStats,\n getSessionStats,\n optimizeSession,\n };\n}\n","/**\n * Sparn MCP Server - Model Context Protocol server implementation\n *\n * Exposes Sparn's neuroscience-inspired context optimization as MCP tools,\n * enabling integration with Claude Desktop, VS Code, and other MCP clients.\n *\n * Tools:\n * - sparn_optimize: Optimize context with configurable options\n * - sparn_stats: Get optimization statistics\n * - sparn_consolidate: Run memory consolidation (sleep replay)\n */\n\nimport { McpServer } from '@modelcontextprotocol/sdk/server/mcp.js';\nimport { z } from 'zod';\nimport { createGenericAdapter } from '../adapters/generic.js';\nimport type { KVMemory } from '../core/kv-memory.js';\nimport { createSleepCompressor } from '../core/sleep-compressor.js';\nimport type { SparnConfig } from '../types/config.js';\nimport { DEFAULT_CONFIG } from '../types/config.js';\n\n/**\n * Options for creating the Sparn MCP server.\n */\nexport interface SparnMcpServerOptions {\n /** KV memory store instance */\n memory: KVMemory;\n /** Sparn configuration (defaults to DEFAULT_CONFIG) */\n config?: SparnConfig;\n}\n\n/**\n * Create and configure the Sparn MCP server with all tools registered.\n *\n * @param options - Server options including memory store and config\n * @returns Configured McpServer instance ready to connect to a transport\n */\nexport function createSparnMcpServer(options: SparnMcpServerOptions): McpServer {\n const { memory, config = DEFAULT_CONFIG } = options;\n\n const server = new McpServer({\n name: 'sparn',\n version: '1.1.1',\n });\n\n registerOptimizeTool(server, memory, config);\n registerStatsTool(server, memory);\n registerConsolidateTool(server, memory);\n\n return server;\n}\n\n/**\n * Register the sparn_optimize tool.\n *\n * Optimizes input context using the neuroscience-inspired pipeline:\n * BTSP detection, engram scoring, confidence states, and sparse pruning.\n */\nfunction registerOptimizeTool(server: McpServer, memory: KVMemory, config: SparnConfig): void {\n server.registerTool(\n 'sparn_optimize',\n {\n title: 'Sparn Optimize',\n description:\n 'Optimize context using neuroscience-inspired pruning. ' +\n 'Applies BTSP detection, engram scoring, confidence states, ' +\n 'and sparse pruning to reduce token usage while preserving important information.',\n inputSchema: {\n context: z.string().describe('The context text to optimize'),\n dryRun: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, do not persist changes to the memory store'),\n verbose: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, include per-entry details in the response'),\n threshold: z\n .number()\n .min(0)\n .max(100)\n .optional()\n .describe('Custom pruning threshold (1-100, overrides config)'),\n },\n },\n async ({ context, dryRun, verbose, threshold }) => {\n try {\n const effectiveConfig = threshold\n ? { ...config, pruning: { ...config.pruning, threshold } }\n : config;\n\n const adapter = createGenericAdapter(memory, effectiveConfig);\n const result = await adapter.optimize(context, {\n dryRun,\n verbose,\n threshold,\n });\n\n const response = {\n optimizedContext: result.optimizedContext,\n tokensBefore: result.tokensBefore,\n tokensAfter: result.tokensAfter,\n reduction: `${(result.reduction * 100).toFixed(1)}%`,\n entriesProcessed: result.entriesProcessed,\n entriesKept: result.entriesKept,\n durationMs: result.durationMs,\n stateDistribution: result.stateDistribution,\n ...(verbose && result.details ? { details: result.details } : {}),\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n\n/**\n * Register the sparn_stats tool.\n *\n * Returns optimization statistics from the memory store, including\n * total commands run, tokens saved, and average reduction.\n */\nfunction registerStatsTool(server: McpServer, memory: KVMemory): void {\n server.registerTool(\n 'sparn_stats',\n {\n title: 'Sparn Stats',\n description:\n 'Get optimization statistics including total commands run, ' +\n 'tokens saved, and average reduction percentage.',\n inputSchema: {\n reset: z\n .boolean()\n .optional()\n .default(false)\n .describe('If true, reset all optimization statistics'),\n },\n },\n async ({ reset }) => {\n try {\n if (reset) {\n await memory.clearOptimizationStats();\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(\n {\n message: 'Optimization statistics have been reset.',\n totalCommands: 0,\n totalTokensSaved: 0,\n averageReduction: '0.0%',\n },\n null,\n 2,\n ),\n },\n ],\n };\n }\n\n const stats = await memory.getOptimizationStats();\n const totalCommands = stats.length;\n\n const totalTokensSaved = stats.reduce(\n (sum, s) => sum + (s.tokens_before - s.tokens_after),\n 0,\n );\n\n const averageReduction =\n totalCommands > 0\n ? stats.reduce((sum, s) => {\n const reduction =\n s.tokens_before > 0 ? (s.tokens_before - s.tokens_after) / s.tokens_before : 0;\n return sum + reduction;\n }, 0) / totalCommands\n : 0;\n\n const recentOptimizations = stats.slice(0, 10).map((s) => ({\n timestamp: new Date(s.timestamp).toISOString(),\n tokensBefore: s.tokens_before,\n tokensAfter: s.tokens_after,\n entriesPruned: s.entries_pruned,\n durationMs: s.duration_ms,\n reduction: `${(\n ((s.tokens_before - s.tokens_after) / Math.max(s.tokens_before, 1)) * 100\n ).toFixed(1)}%`,\n }));\n\n const response = {\n totalCommands,\n totalTokensSaved,\n averageReduction: `${(averageReduction * 100).toFixed(1)}%`,\n recentOptimizations,\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n\n/**\n * Register the sparn_consolidate tool.\n *\n * Runs the sleep-compressor consolidation process, which removes\n * decayed entries and merges duplicates in the memory store.\n */\nfunction registerConsolidateTool(server: McpServer, memory: KVMemory): void {\n server.registerTool(\n 'sparn_consolidate',\n {\n title: 'Sparn Consolidate',\n description:\n 'Run memory consolidation (sleep replay). ' +\n 'Removes decayed entries and merges duplicates to reclaim space. ' +\n 'Inspired by the neuroscience principle of sleep-based memory consolidation.',\n },\n async () => {\n try {\n const allIds = await memory.list();\n const allEntries = await Promise.all(\n allIds.map(async (id) => {\n const entry = await memory.get(id);\n return entry;\n }),\n );\n\n const entries = allEntries.filter((e) => e !== null);\n\n const compressor = createSleepCompressor();\n const result = compressor.consolidate(entries);\n\n // Apply changes to memory store\n for (const removed of result.removed) {\n await memory.delete(removed.id);\n }\n\n for (const kept of result.kept) {\n await memory.put(kept);\n }\n\n // Run VACUUM to reclaim disk space\n await memory.compact();\n\n const response = {\n entriesBefore: result.entriesBefore,\n entriesAfter: result.entriesAfter,\n decayedRemoved: result.decayedRemoved,\n duplicatesRemoved: result.duplicatesRemoved,\n compressionRatio: `${(result.compressionRatio * 100).toFixed(1)}%`,\n durationMs: result.durationMs,\n vacuumCompleted: true,\n };\n\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify(response, null, 2),\n },\n ],\n };\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n return {\n content: [\n {\n type: 'text' as const,\n text: JSON.stringify({ error: message }),\n },\n ],\n isError: true,\n };\n }\n },\n );\n}\n","/**\n * Configuration types for Sparn behavior customization.\n */\n\n/**\n * Agent adapter type.\n */\nexport type AgentType = 'claude-code' | 'generic';\n\n/**\n * Pruning configuration.\n */\nexport interface PruningConfig {\n /** Percentage of top-scored entries to keep (1-100, default: 5) */\n threshold: number;\n\n /** Aggressiveness scale 0-100 (affects TF-IDF weighting, default: 50) */\n aggressiveness: number;\n}\n\n/**\n * Decay configuration.\n */\nexport interface DecayConfig {\n /** Default TTL in hours (default: 24) */\n defaultTTL: number;\n\n /** Decay threshold for pruning (0.0-1.0, default: 0.95) */\n decayThreshold: number;\n}\n\n/**\n * Confidence state threshold configuration.\n */\nexport interface StatesConfig {\n /** Score threshold for active state (default: 0.7) */\n activeThreshold: number;\n\n /** Score threshold for ready state (default: 0.3) */\n readyThreshold: number;\n}\n\n/**\n * UI configuration.\n */\nexport interface UIConfig {\n /** Enable colored output (default: true) */\n colors: boolean;\n\n /** Enable sound effects (default: false) */\n sounds: boolean;\n\n /** Verbose logging (default: false) */\n verbose: boolean;\n}\n\n/**\n * Real-time optimization configuration.\n */\nexport interface RealtimeConfig {\n /** Target token budget for optimized context (default: 50000) */\n tokenBudget: number;\n\n /** Token threshold that triggers auto-optimization (default: 60000) */\n autoOptimizeThreshold: number;\n\n /** File patterns to watch for changes (default: ['**\\/*.jsonl']) */\n watchPatterns: string[];\n\n /** Daemon PID file path (default: '.sparn/daemon.pid') */\n pidFile: string;\n\n /** Daemon log file path (default: '.sparn/daemon.log') */\n logFile: string;\n\n /** Debounce delay in milliseconds for file changes (default: 5000) */\n debounceMs: number;\n\n /** Enable incremental optimization (default: true) */\n incremental: boolean;\n\n /** Sliding window size for context entries (default: 500) */\n windowSize: number;\n\n /** Consolidation interval in hours, or null for disabled (default: null) */\n consolidationInterval: number | null;\n}\n\n/**\n * Complete Sparn configuration.\n */\nexport interface SparnConfig {\n pruning: PruningConfig;\n decay: DecayConfig;\n states: StatesConfig;\n agent: AgentType;\n ui: UIConfig;\n /** Auto-consolidation interval in hours, or null for manual */\n autoConsolidate: number | null;\n /** Real-time optimization settings */\n realtime: RealtimeConfig;\n}\n\n/**\n * Default configuration values.\n */\nexport const DEFAULT_CONFIG: SparnConfig = {\n pruning: {\n threshold: 5,\n aggressiveness: 50,\n },\n decay: {\n defaultTTL: 24,\n decayThreshold: 0.95,\n },\n states: {\n activeThreshold: 0.7,\n readyThreshold: 0.3,\n },\n agent: 'generic',\n ui: {\n colors: true,\n sounds: false,\n verbose: false,\n },\n autoConsolidate: null,\n realtime: {\n tokenBudget: 40000,\n autoOptimizeThreshold: 60000,\n watchPatterns: ['**/*.jsonl'],\n pidFile: '.sparn/daemon.pid',\n logFile: '.sparn/daemon.log',\n debounceMs: 5000,\n incremental: true,\n windowSize: 500,\n consolidationInterval: null,\n },\n};\n","/**\n * Logging utility.\n * Simple console wrapper with log levels.\n */\n\nexport type LogLevel = 'debug' | 'info' | 'warn' | 'error';\n\n/**\n * Logger interface.\n */\nexport interface Logger {\n debug(message: string, ...args: unknown[]): void;\n info(message: string, ...args: unknown[]): void;\n warn(message: string, ...args: unknown[]): void;\n error(message: string, ...args: unknown[]): void;\n}\n\n/**\n * Create a logger with optional verbosity control.\n *\n * @param verbose - Enable debug-level logging\n * @returns Logger instance\n */\nexport function createLogger(verbose = false): Logger {\n return {\n debug(message: string, ...args: unknown[]): void {\n if (verbose) {\n console.debug(`[DEBUG] ${message}`, ...args);\n }\n },\n info(message: string, ...args: unknown[]): void {\n console.info(`[INFO] ${message}`, ...args);\n },\n warn(message: string, ...args: unknown[]): void {\n console.warn(`[WARN] ${message}`, ...args);\n },\n error(message: string, ...args: unknown[]): void {\n console.error(`[ERROR] ${message}`, ...args);\n },\n };\n}\n"],"mappings":";AAOA,SAAS,kBAAkB;;;ACF3B,SAAS,kBAAkB;AAcpB,SAAS,YAAY,SAAyB;AACnD,SAAO,WAAW,QAAQ,EAAE,OAAO,SAAS,MAAM,EAAE,OAAO,KAAK;AAClE;;;ADgBO,SAAS,qBAAmC;AAEjD,QAAM,gBAAgB;AAAA;AAAA,IAEpB;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA;AAAA,IACA;AAAA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AAEA,WAAS,WAAW,SAA0B;AAC5C,WAAO,cAAc,KAAK,CAAC,YAAY,QAAQ,KAAK,OAAO,CAAC;AAAA,EAC9D;AAEA,WAAS,gBACP,SACA,OAAiB,CAAC,GAClB,WAAoC,CAAC,GACxB;AACb,WAAO;AAAA,MACL,IAAI,WAAW;AAAA,MACf;AAAA,MACA,MAAM,YAAY,OAAO;AAAA,MACzB,WAAW,KAAK,IAAI;AAAA,MACpB,OAAO;AAAA;AAAA,MACP,KAAK,MAAM,KAAK;AAAA;AAAA,MAChB,OAAO;AAAA;AAAA,MACP,aAAa;AAAA,MACb,MAAM,CAAC,GAAG,MAAM,MAAM;AAAA,MACtB;AAAA,MACA,QAAQ;AAAA,IACV;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;;;AE3CO,SAAS,uBAAuB,QAAkD;AACvF,QAAM,EAAE,iBAAiB,eAAe,IAAI;AAE5C,WAAS,eAAe,OAAqC;AAE3D,QAAI,MAAM,QAAQ;AAChB,aAAO;AAAA,IACT;AAIA,QAAI,MAAM,QAAQ,iBAAiB;AACjC,aAAO;AAAA,IACT;AAGA,QAAI,MAAM,SAAS,gBAAgB;AACjC,aAAO;AAAA,IACT;AAGA,WAAO;AAAA,EACT;AAEA,WAAS,WAAW,OAAiC;AACnD,UAAM,WAAW,eAAe,KAAK;AAErC,WAAO;AAAA,MACL,GAAG;AAAA,MACH,OAAO;AAAA,IACT;AAAA,EACF;AAEA,WAAS,gBAAgB,SAA2C;AAClE,UAAM,eAAkC;AAAA,MACtC,QAAQ;AAAA,MACR,OAAO;AAAA,MACP,QAAQ;AAAA,MACR,OAAO,QAAQ;AAAA,IACjB;AAEA,eAAW,SAAS,SAAS;AAC3B,YAAM,QAAQ,eAAe,KAAK;AAClC,mBAAa,KAAK;AAAA,IACpB;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACjDO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,WAAW,IAAI;AAEvB,WAAS,eAAe,cAAsB,cAA8B;AAC1E,QAAI,iBAAiB,EAAG,QAAO;AAC/B,QAAI,gBAAgB,EAAG,QAAO;AAG9B,UAAM,QAAQ,eAAe;AAC7B,UAAM,QAAQ,IAAI,KAAK,IAAI,CAAC,KAAK;AAGjC,WAAO,KAAK,IAAI,GAAG,KAAK,IAAI,GAAG,KAAK,CAAC;AAAA,EACvC;AAEA,WAAS,eAAe,OAAoB,cAAsB,KAAK,IAAI,GAAW;AAEpF,UAAM,oBAAoB,cAAc,MAAM;AAC9C,UAAM,eAAe,KAAK,IAAI,GAAG,oBAAoB,GAAI;AAGzD,UAAM,QAAQ,eAAe,cAAc,MAAM,GAAG;AAGpD,QAAI,QAAQ,MAAM,SAAS,IAAI;AAG/B,QAAI,MAAM,cAAc,GAAG;AACzB,YAAM,cAAc,KAAK,IAAI,MAAM,cAAc,CAAC,IAAI;AACtD,cAAQ,KAAK,IAAI,GAAK,QAAQ,WAAW;AAAA,IAC3C;AAGA,QAAI,MAAM,QAAQ;AAChB,cAAQ,KAAK,IAAI,OAAO,GAAG;AAAA,IAC7B;AAEA,WAAO,KAAK,IAAI,GAAG,KAAK,IAAI,GAAG,KAAK,CAAC;AAAA,EACvC;AAEA,WAAS,WAAW,OAAiC;AACnD,WAAO;AAAA,MACL,GAAG;AAAA,MACH,KAAK,aAAa;AAAA;AAAA,MAClB,WAAW,KAAK,IAAI;AAAA,IACtB;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AClFO,SAAS,eAAe,MAAsB;AACnD,MAAI,CAAC,QAAQ,KAAK,WAAW,GAAG;AAC9B,WAAO;AAAA,EACT;AAGA,QAAM,QAAQ,KAAK,MAAM,KAAK,EAAE,OAAO,CAAC,MAAM,EAAE,SAAS,CAAC;AAC1D,QAAM,YAAY,MAAM;AAGxB,QAAM,YAAY,KAAK;AACvB,QAAM,eAAe,KAAK,KAAK,YAAY,CAAC;AAG5C,QAAM,eAAe,KAAK,KAAK,YAAY,IAAI;AAG/C,SAAO,KAAK,IAAI,cAAc,YAAY;AAC5C;;;ACAO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,UAAU,IAAI;AAEtB,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,WAAS,YAAY,MAAc,QAA0B;AAC3D,UAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAE/C,WAAO,KAAK,KAAK,KAAK;AAAA,EACxB;AAEA,WAAS,aAAa,MAAc,YAAmC;AACrE,UAAM,YAAY,WAAW;AAC7B,UAAM,eAAe,WAAW,OAAO,CAAC,UAAU;AAChD,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,aAAO,OAAO,SAAS,IAAI;AAAA,IAC7B,CAAC,EAAE;AAEH,QAAI,iBAAiB,EAAG,QAAO;AAE/B,WAAO,KAAK,IAAI,YAAY,YAAY;AAAA,EAC1C;AAEA,WAAS,WAAW,OAAoB,YAAmC;AACzE,UAAM,SAAS,SAAS,MAAM,OAAO;AACrC,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AACvC,QAAI,aAAa;AAEjB,eAAW,QAAQ,aAAa;AAC9B,YAAM,KAAK,YAAY,MAAM,MAAM;AACnC,YAAM,MAAM,aAAa,MAAM,UAAU;AACzC,oBAAc,KAAK;AAAA,IACrB;AAGA,WAAO,aAAa,OAAO;AAAA,EAC7B;AAEA,WAAS,MAAM,SAAqC;AAClD,QAAI,QAAQ,WAAW,GAAG;AACxB,aAAO;AAAA,QACL,MAAM,CAAC;AAAA,QACP,SAAS,CAAC;AAAA,QACV,gBAAgB;AAAA,QAChB,cAAc;AAAA,MAChB;AAAA,IACF;AAGA,UAAM,iBAAiB,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,SAAS,QAAQ,IAAI,CAAC,WAAW;AAAA,MACrC;AAAA,MACA,OAAO,WAAW,OAAO,OAAO;AAAA,IAClC,EAAE;AAGF,WAAO,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAGvC,UAAM,YAAY,KAAK,IAAI,GAAG,KAAK,KAAK,QAAQ,UAAU,YAAY,IAAI,CAAC;AAC3E,UAAM,OAAO,OAAO,MAAM,GAAG,SAAS,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAC1D,UAAM,UAAU,OAAO,MAAM,SAAS,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAG1D,UAAM,eAAe,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAE/E,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;;;ACpHA,SAAS,cAAAA,mBAAkB;AAepB,SAAS,uBAAuB,SAAgC;AACrE,QAAM,UAAyB,CAAC;AAChC,QAAM,MAAM,KAAK,IAAI;AAGrB,QAAM,QAAQ,QAAQ,MAAM,IAAI;AAChC,MAAI,eAAyB,CAAC;AAC9B,MAAI,YAAuB;AAE3B,aAAW,QAAQ,OAAO;AACxB,UAAM,UAAU,KAAK,KAAK;AAG1B,QAAI,QAAQ,WAAW,OAAO,KAAK,QAAQ,WAAW,YAAY,GAAG;AACnE,UAAI,aAAa,SAAS,GAAG;AAC3B,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAGE,QAAQ,SAAS,kBAAkB,KACnC,QAAQ,SAAS,UAAU,KAC3B,QAAQ,SAAS,YAAY,GAC7B;AACA,UAAI,aAAa,SAAS,GAAG;AAC3B,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,QAAQ,SAAS,oBAAoB,KAAK,QAAQ,SAAS,qBAAqB,GAAG;AAC1F,UAAI,aAAa,SAAS,KAAK,cAAc,UAAU;AACrD,gBAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AACjE,uBAAe,CAAC;AAAA,MAClB;AACA,kBAAY;AACZ,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,aAAa,SAAS,GAAG;AAChC,mBAAa,KAAK,IAAI;AAAA,IACxB,WAES,QAAQ,SAAS,GAAG;AAC3B,mBAAa,KAAK,IAAI;AACtB,kBAAY;AAAA,IACd;AAAA,EACF;AAGA,MAAI,aAAa,SAAS,GAAG;AAC3B,YAAQ,KAAK,YAAY,aAAa,KAAK,IAAI,GAAG,WAAW,GAAG,CAAC;AAAA,EACnE;AAEA,SAAO,QAAQ,OAAO,CAAC,MAAM,EAAE,QAAQ,KAAK,EAAE,SAAS,CAAC;AAC1D;AASO,SAAS,YAAY,SAAiB,MAAiB,UAA+B;AAC3F,QAAM,OAAiB,CAAC,IAAI;AAG5B,MAAI,eAAe;AACnB,MAAI,SAAS,eAAgB,gBAAe;AAC5C,MAAI,SAAS,OAAQ,gBAAe;AACpC,MAAI,SAAS,SAAU,gBAAe;AAEtC,SAAO;AAAA,IACL,IAAIC,YAAW;AAAA,IACf;AAAA,IACA,MAAM,YAAY,OAAO;AAAA,IACzB,WAAW;AAAA,IACX,OAAO;AAAA,IACP,OAAO,eAAe,MAAM,WAAW,eAAe,MAAM,UAAU;AAAA,IACtE,KAAK,KAAK;AAAA;AAAA,IACV,aAAa;AAAA,IACb;AAAA,IACA,UAAU,EAAE,KAAK;AAAA,IACjB,QAAQ;AAAA,EACV;AACF;AAQO,SAAS,oBAAoB,SAAgC;AAClE,QAAM,UAAyB,CAAC;AAChC,QAAM,MAAM,KAAK,IAAI;AAGrB,QAAM,SAAS,QAAQ,MAAM,OAAO;AAEpC,aAAW,SAAS,QAAQ;AAC1B,UAAM,UAAU,MAAM,KAAK;AAC3B,QAAI,QAAQ,WAAW,EAAG;AAE1B,YAAQ,KAAK,YAAY,SAAS,SAAS,GAAG,CAAC;AAAA,EACjD;AAEA,SAAO;AACT;;;ACrHA,IAAM,sBAAsB;AAAA;AAAA,EAE1B,qBAAqB;AAAA;AAAA;AAAA,EAGrB,mBAAmB;AAAA;AAAA;AAAA,EAGnB,qBAAqB,KAAK;AAAA;AAAA;AAAA,EAG1B,cAAc;AAAA;AAAA,IAEZ;AAAA,IACA;AAAA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IAGA;AAAA,EACF;AACF;AAQO,SAAS,wBAAwB,QAAkB,QAAmC;AAE3F,QAAM,SAAS,mBAAmB;AAAA,IAChC,WAAW,OAAO,QAAQ;AAAA,EAC5B,CAAC;AAED,QAAM,SAAS,mBAAmB,OAAO,KAAK;AAC9C,QAAM,SAAS,uBAAuB,OAAO,MAAM;AACnD,QAAM,OAAO,mBAAmB;AAEhC,iBAAe,SACb,SACA,UAA2B,CAAC,GACC;AAC7B,UAAM,YAAY,KAAK,IAAI;AAI3B,UAAM,UAAU,uBAAuB,OAAO;AAG9C,UAAM,kBAAkB,QAAQ,IAAI,CAAC,UAAU;AAC7C,YAAM,SAAS,oBAAoB,aAAa;AAAA,QAAK,CAAC,YACpD,QAAQ,KAAK,MAAM,OAAO;AAAA,MAC5B;AAEA,UAAI,QAAQ;AACV,cAAM,YAAY,KAAK,gBAAgB,MAAM,SAAS,CAAC,GAAG,MAAM,MAAM,aAAa,GAAG;AAAA,UACpF,mBAAmB,MAAM;AAAA,QAC3B,CAAC;AAED,eAAO;AAAA,UACL,GAAG;AAAA,UACH,WAAW,MAAM;AAAA,QACnB;AAAA,MACF;AAEA,aAAO;AAAA,IACT,CAAC;AAGD,UAAM,iBAAiB,gBAAgB,IAAI,CAAC,UAAU;AACpD,YAAM,qBACJ,MAAM,QAAQ,KAAK,EAAE,WAAW,OAAO,KAAK,MAAM,QAAQ,KAAK,EAAE,WAAW,YAAY;AAE1F,UAAI,oBAAoB;AACtB,eAAO;AAAA,UACL,GAAG;AAAA,UACH,OAAO,MAAM,QAAQ,oBAAoB;AAAA,QAC3C;AAAA,MACF;AAEA,aAAO;AAAA,IACT,CAAC;AAGD,UAAM,gBAAgB,eAAe,IAAI,CAAC,UAAU;AAClD,YAAM,aAAa,OAAO,eAAe,KAAK;AAC9C,aAAO;AAAA,QACL,GAAG;AAAA,QACH,OAAO;AAAA,MACT;AAAA,IACF,CAAC;AAGD,UAAM,oBAAoB,cAAc,IAAI,CAAC,UAAU;AACrD,YAAM,QAAQ,OAAO,eAAe,KAAK;AACzC,aAAO;AAAA,QACL,GAAG;AAAA,QACH;AAAA,MACF;AAAA,IACF,CAAC;AAGD,UAAM,cAAc,OAAO,MAAM,iBAAiB;AAGlD,QAAI,CAAC,QAAQ,QAAQ;AACnB,iBAAW,SAAS,YAAY,MAAM;AACpC,cAAM,OAAO,IAAI,KAAK;AAAA,MACxB;AAGA,YAAM,OAAO,mBAAmB;AAAA,QAC9B,WAAW,KAAK,IAAI;AAAA,QACpB,eAAe,YAAY;AAAA,QAC3B,cAAc,YAAY;AAAA,QAC1B,gBAAgB,YAAY,QAAQ;AAAA,QACpC,aAAa,KAAK,IAAI,IAAI;AAAA,MAC5B,CAAC;AAAA,IACH;AAGA,UAAM,mBAAmB,YAAY,KAAK,IAAI,CAAC,UAAU,MAAM,OAAO,EAAE,KAAK,IAAI;AAGjF,UAAM,oBAAoB,OAAO,gBAAgB,YAAY,IAAI;AAGjE,UAAM,SAA6B;AAAA,MACjC;AAAA,MACA,cAAc,YAAY;AAAA,MAC1B,aAAa,YAAY;AAAA,MACzB,WACE,YAAY,iBAAiB,KACxB,YAAY,iBAAiB,YAAY,gBAAgB,YAAY,iBACtE;AAAA,MACN,kBAAkB,QAAQ;AAAA,MAC1B,aAAa,YAAY,KAAK;AAAA,MAC9B,YAAY,KAAK,IAAI,IAAI;AAAA,MACzB;AAAA,IACF;AAGA,QAAI,QAAQ,SAAS;AACnB,aAAO,UAAU,YAAY,KAAK,IAAI,CAAC,WAAW;AAAA,QAChD,IAAI,MAAM;AAAA,QACV,OAAO,MAAM;AAAA,QACb,OAAO,MAAM,SAAS;AAAA,QACtB,QAAQ,MAAM,KAAK,SAAS,MAAM;AAAA,QAClC,QAAQ,eAAe,MAAM,OAAO;AAAA,MACtC,EAAE;AAAA,IACJ;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,EACF;AACF;;;ACvLA,SAAS,cAAAC,mBAAkB;AAkBpB,SAAS,qBAAqB,QAAkB,QAAmC;AACxF,QAAM,SAAS,mBAAmB,OAAO,OAAO;AAChD,QAAM,SAAS,mBAAmB,OAAO,KAAK;AAC9C,QAAM,SAAS,uBAAuB,OAAO,MAAM;AACnD,QAAM,OAAO,mBAAmB;AAEhC,iBAAe,SACb,SACA,UAA2B,CAAC,GACC;AAC7B,UAAM,YAAY,KAAK,IAAI;AAG3B,UAAM,QAAQ,QAAQ,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,EAAE,SAAS,CAAC;AACzE,UAAM,UAAyB,MAAM,IAAI,CAAC,aAAa;AAAA,MACrD,IAAIC,YAAW;AAAA,MACf;AAAA,MACA,MAAM,YAAY,OAAO;AAAA,MACzB,WAAW,KAAK,IAAI;AAAA,MACpB,OAAO,KAAK,WAAW,OAAO,IAAI,IAAM;AAAA;AAAA,MACxC,KAAK,OAAO,MAAM,aAAa;AAAA;AAAA,MAC/B,OAAO;AAAA,MACP,aAAa;AAAA,MACb,MAAM,CAAC;AAAA,MACP,UAAU,CAAC;AAAA,MACX,QAAQ,KAAK,WAAW,OAAO;AAAA,IACjC,EAAE;AAGF,UAAM,eAAe,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGlF,UAAM,gBAAgB,QAAQ,IAAI,CAAC,WAAW;AAAA,MAC5C,GAAG;AAAA,MACH,OAAO,OAAO,eAAe,KAAK;AAAA,IACpC,EAAE;AAGF,UAAM,gBAAgB,cAAc,IAAI,CAAC,UAAU,OAAO,WAAW,KAAK,CAAC;AAG3E,UAAM,cAAc,OAAO,MAAM,aAAa;AAG9C,UAAM,mBAAmB,YAAY,KAAK;AAAA,MACxC,CAAC,MAAM,EAAE,UAAU,YAAY,EAAE,UAAU;AAAA,IAC7C;AAGA,UAAM,cAAc,iBAAiB,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAG1F,UAAM,mBAAmB,iBAAiB,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,IAAI;AAGzE,QAAI,CAAC,QAAQ,QAAQ;AACnB,iBAAW,SAAS,kBAAkB;AACpC,cAAM,OAAO,IAAI,KAAK;AAAA,MACxB;AAGA,YAAM,OAAO,mBAAmB;AAAA,QAC9B,WAAW,KAAK,IAAI;AAAA,QACpB,eAAe;AAAA,QACf,cAAc;AAAA,QACd,gBAAgB,QAAQ,SAAS,iBAAiB;AAAA,QAClD,aAAa,KAAK,IAAI,IAAI;AAAA,MAC5B,CAAC;AAAA,IACH;AAGA,UAAM,eAAe,OAAO,gBAAgB,gBAAgB;AAE5D,UAAM,SAA6B;AAAA,MACjC;AAAA,MACA;AAAA,MACA;AAAA,MACA,WAAW,eAAe,KAAK,eAAe,eAAe,eAAe;AAAA,MAC5E,kBAAkB,QAAQ;AAAA,MAC1B,aAAa,iBAAiB;AAAA,MAC9B,mBAAmB;AAAA,MACnB,YAAY,KAAK,IAAI,IAAI;AAAA,IAC3B;AAGA,QAAI,QAAQ,SAAS;AACnB,aAAO,UAAU,iBAAiB,IAAI,CAAC,OAAO;AAAA,QAC5C,IAAI,EAAE;AAAA,QACN,OAAO,EAAE;AAAA,QACT,OAAO,EAAE;AAAA,QACT,QAAQ,EAAE;AAAA,QACV,QAAQ,eAAe,EAAE,OAAO;AAAA,MAClC,EAAE;AAAA,IACJ;AAEA,WAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL;AAAA,EACF;AACF;;;ACpEO,SAAS,mBAAmB,QAA0C;AAC3E,QAAM,EAAE,aAAa,MAAM,IAAI;AAC/B,QAAM,eAAe,mBAAmB,KAAK;AAE7C,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,WAAS,YAAY,MAAc,QAA0B;AAC3D,UAAM,QAAQ,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAE/C,WAAO,KAAK,KAAK,KAAK;AAAA,EACxB;AAEA,WAAS,aAAa,MAAc,YAAmC;AACrE,UAAM,YAAY,WAAW;AAC7B,UAAM,eAAe,WAAW,OAAO,CAAC,UAAU;AAChD,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,aAAO,OAAO,SAAS,IAAI;AAAA,IAC7B,CAAC,EAAE;AAEH,QAAI,iBAAiB,EAAG,QAAO;AAE/B,WAAO,KAAK,IAAI,YAAY,YAAY;AAAA,EAC1C;AAEA,WAAS,eAAe,OAAoB,YAAmC;AAC7E,UAAM,SAAS,SAAS,MAAM,OAAO;AACrC,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AACvC,QAAI,aAAa;AAEjB,eAAW,QAAQ,aAAa;AAC9B,YAAM,KAAK,YAAY,MAAM,MAAM;AACnC,YAAM,MAAM,aAAa,MAAM,UAAU;AACzC,oBAAc,KAAK;AAAA,IACrB;AAGA,WAAO,aAAa,OAAO;AAAA,EAC7B;AAEA,WAAS,mBAAmB,OAA4B;AAEtD,QAAI,MAAM,OAAQ,QAAO;AAGzB,YAAQ,MAAM,OAAO;AAAA,MACnB,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT,KAAK;AACH,eAAO;AAAA,MACT;AACE,eAAO;AAAA,IACX;AAAA,EACF;AAEA,WAAS,cAAc,OAAoB,YAAmC;AAC5E,UAAM,QAAQ,eAAe,OAAO,UAAU;AAC9C,UAAM,eAAe,aAAa,eAAe,KAAK;AACtD,UAAM,cAAc,IAAI;AACxB,UAAM,kBAAkB,mBAAmB,KAAK;AAIhD,WAAO,SAAS,IAAI,eAAe;AAAA,EACrC;AAEA,WAAS,WACP,SACA,SAAiB,aAC4B;AAC7C,QAAI,QAAQ,WAAW,GAAG;AACxB,aAAO;AAAA,QACL,MAAM,CAAC;AAAA,QACP,SAAS,CAAC;AAAA,QACV,gBAAgB;AAAA,QAChB,cAAc;AAAA,QACd,mBAAmB;AAAA,MACrB;AAAA,IACF;AAGA,UAAM,iBAAiB,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,cAAc,QAAQ,OAAO,CAAC,MAAM,EAAE,MAAM;AAClD,UAAM,iBAAiB,QAAQ,OAAO,CAAC,MAAM,CAAC,EAAE,MAAM;AAEtD,UAAM,aAAa,YAAY,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGpF,UAAM,SAAS,eAAe,IAAI,CAAC,WAAW;AAAA,MAC5C;AAAA,MACA,OAAO,cAAc,OAAO,OAAO;AAAA,MACnC,QAAQ,eAAe,MAAM,OAAO;AAAA,IACtC,EAAE;AAGF,WAAO,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAGvC,UAAM,OAAsB,CAAC,GAAG,WAAW;AAC3C,UAAM,UAAyB,CAAC;AAChC,QAAI,gBAAgB;AAEpB,eAAW,QAAQ,QAAQ;AACzB,UAAI,gBAAgB,KAAK,UAAU,QAAQ;AACzC,aAAK,KAAK,KAAK,KAAK;AACpB,yBAAiB,KAAK;AAAA,MACxB,OAAO;AACL,gBAAQ,KAAK,KAAK,KAAK;AAAA,MACzB;AAAA,IACF;AAEA,UAAM,oBAAoB,SAAS,IAAI,gBAAgB,SAAS;AAEhE,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA,cAAc;AAAA,MACd;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,EACF;AACF;AASO,SAAS,6BACd,gBACA,aACA,cACc;AACd,SAAO,mBAAmB;AAAA,IACxB,aAAa,eAAe;AAAA,IAC5B,OAAO;AAAA,IACP,QAAQ;AAAA,EACV,CAAC;AACH;;;AChIO,SAAS,yBAA2C;AACzD,QAAM,gBAAsC,CAAC;AAC7C,MAAI,gBAA8B;AAAA,IAChC,WAAW,KAAK,IAAI;AAAA,IACpB,iBAAiB;AAAA,IACjB,oBAAoB;AAAA,IACpB,kBAAkB;AAAA,IAClB,gBAAgB;AAAA,IAChB,aAAa;AAAA,EACf;AAEA,MAAI,YAAY;AAChB,MAAI,cAAc;AAElB,WAAS,mBAAmB,QAAkC;AAC5D,kBAAc,KAAK,MAAM;AAGzB,kBAAc;AACd,kBAAc,oBAAoB,OAAO,eAAe,OAAO;AAG/D,QAAI,OAAO,eAAe,GAAG;AAC3B,YAAM,OAAO,KAAK,MAAM,OAAO,mBAAmB,OAAO,YAAY;AACrE,mBAAa;AACb,qBAAe,OAAO,mBAAmB;AAAA,IAC3C;AAGA,kBAAc,kBACX,cAAc,kBAAkB,cAAc,qBAAqB,KAAK,OAAO,YAChF,cAAc;AAGhB,QAAI,cAAc,SAAS,KAAM;AAC/B,oBAAc,MAAM;AAAA,IACtB;AAAA,EACF;AAEA,WAAS,aAAa,QAAqC;AACzD,oBAAgB;AAAA,MACd,GAAG;AAAA,MACH,GAAG;AAAA,IACL;AAAA,EACF;AAEA,WAAS,oBAAoB,QAAkB,YAA4B;AACzE,QAAI,OAAO,WAAW,EAAG,QAAO;AAEhC,UAAM,SAAS,CAAC,GAAG,MAAM,EAAE,KAAK,CAAC,GAAG,MAAM,IAAI,CAAC;AAC/C,UAAM,QAAQ,KAAK,KAAM,aAAa,MAAO,OAAO,MAAM,IAAI;AAC9D,WAAO,OAAO,KAAK,KAAK;AAAA,EAC1B;AAEA,WAAS,cAA+B;AACtC,UAAM,YAAY,cAAc;AAChC,UAAM,gBAAgB,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,UAAU,CAAC;AAC1E,UAAM,mBAAmB,cAAc;AAAA,MACrC,CAAC,KAAK,MAAM,OAAO,EAAE,eAAe,EAAE;AAAA,MACtC;AAAA,IACF;AAEA,UAAM,oBAAoB,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,cAAc,CAAC;AAClF,UAAM,mBAAmB,oBAAoB,IAAI,mBAAmB,oBAAoB;AAExF,UAAM,YAAY,cAAc,IAAI,CAAC,MAAM,EAAE,QAAQ;AAErD,UAAM,oBAAoB,YAAY;AACtC,UAAM,UAAU,oBAAoB,IAAI,YAAY,oBAAoB;AAExE,WAAO;AAAA,MACL,WAAW,KAAK,IAAI;AAAA,MACpB,cAAc;AAAA,QACZ;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA,YAAY,oBAAoB,WAAW,EAAE;AAAA,QAC7C,YAAY,oBAAoB,WAAW,EAAE;AAAA,QAC7C,YAAY,oBAAoB,WAAW,EAAE;AAAA,MAC/C;AAAA,MACA,OAAO;AAAA,QACL;AAAA,QACA,WAAW;AAAA,QACX,aAAa;AAAA,QACb,MAAM,cAAc,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,aAAa,CAAC;AAAA,MAC/D;AAAA,MACA,QAAQ;AAAA,QACN,QAAQ,KAAK,IAAI,IAAI,cAAc;AAAA,QACnC,iBAAiB,cAAc;AAAA,QAC/B,aAAa,cAAc;AAAA,MAC7B;AAAA,IACF;AAAA,EACF;AAEA,WAAS,gBAAwB;AAC/B,WAAO,KAAK,UAAU,YAAY,GAAG,MAAM,CAAC;AAAA,EAC9C;AAEA,WAAS,QAAc;AACrB,kBAAc,SAAS;AACvB,gBAAY;AACZ,kBAAc;AACd,oBAAgB;AAAA,MACd,WAAW,KAAK,IAAI;AAAA,MACpB,iBAAiB;AAAA,MACjB,oBAAoB;AAAA,MACpB,kBAAkB;AAAA,MAClB,gBAAgB;AAAA,MAChB,aAAa;AAAA,IACf;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ;AAAA,IACR;AAAA,EACF;AACF;AAGA,IAAI,gBAAyC;AAKtC,SAAS,aAA+B;AAC7C,MAAI,CAAC,eAAe;AAClB,oBAAgB,uBAAuB;AAAA,EACzC;AACA,SAAO;AACT;;;AC3HO,SAAS,2BACd,QACsB;AACtB,QAAM,SAAS,mBAAmB,MAAM;AACxC,QAAM,EAAE,yBAAyB,IAAI;AAGrC,MAAI,QAAmC;AAAA,IACrC,YAAY,oBAAI,IAAI;AAAA,IACpB,mBAAmB,oBAAI,IAAI;AAAA,IAC3B,gBAAgB;AAAA,IAChB,aAAa;AAAA,IACb,sBAAsB,KAAK,IAAI;AAAA,EACjC;AAEA,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAKA,WAAS,wBAAwB,SAAwB,SAAS,OAAa;AAC7E,eAAW,SAAS,SAAS;AAC3B,YAAM,SAAS,SAAS,MAAM,OAAO;AACrC,YAAM,cAAc,CAAC,GAAG,IAAI,IAAI,MAAM,CAAC;AAEvC,iBAAW,QAAQ,aAAa;AAC9B,cAAM,UAAU,MAAM,kBAAkB,IAAI,IAAI,KAAK;AACrD,cAAM,UAAU,SAAS,KAAK,IAAI,GAAG,UAAU,CAAC,IAAI,UAAU;AAE9D,YAAI,YAAY,GAAG;AACjB,gBAAM,kBAAkB,OAAO,IAAI;AAAA,QACrC,OAAO;AACL,gBAAM,kBAAkB,IAAI,MAAM,OAAO;AAAA,QAC3C;AAAA,MACF;AAAA,IACF;AAEA,UAAM,kBAAkB,SAAS,CAAC,QAAQ,SAAS,QAAQ;AAC3D,UAAM,iBAAiB,KAAK,IAAI,GAAG,MAAM,cAAc;AAAA,EACzD;AAKA,WAAS,eAAe,MAAkC;AACxD,UAAM,SAAS,MAAM,WAAW,IAAI,IAAI;AACxC,QAAI,CAAC,OAAQ,QAAO;AAGpB,WAAO,OAAO;AAAA,EAChB;AAKA,WAAS,WAAW,OAAoB,OAAqB;AAC3D,UAAM,WAAW,IAAI,MAAM,MAAM;AAAA,MAC/B;AAAA,MACA;AAAA,MACA,WAAW,KAAK,IAAI;AAAA,IACtB,CAAC;AAAA,EACH;AAEA,WAAS,oBACP,YACA,QAC6C;AAC7C,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM;AAGN,QAAI,MAAM,eAAe,0BAA0B;AAEjD,YAAMC,cAAa,MAAM,KAAK,MAAM,WAAW,OAAO,CAAC,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAC3E,aAAO,aAAa,CAAC,GAAGA,aAAY,GAAG,UAAU,GAAG,MAAM;AAAA,IAC5D;AAGA,UAAM,kBAAiC,CAAC;AACxC,UAAM,gBAA+B,CAAC;AAEtC,eAAW,SAAS,YAAY;AAC9B,YAAM,SAAS,eAAe,MAAM,IAAI;AACxC,UAAI,QAAQ;AACV,sBAAc,KAAK,MAAM;AAAA,MAC3B,OAAO;AACL,wBAAgB,KAAK,KAAK;AAAA,MAC5B;AAAA,IACF;AAGA,QAAI,gBAAgB,SAAS,GAAG;AAC9B,8BAAwB,iBAAiB,KAAK;AAAA,IAChD;AAGA,UAAM,aAAa,CAAC,GAAG,eAAe,GAAG,eAAe;AAGxD,eAAW,SAAS,iBAAiB;AACnC,YAAM,QAAQ,OAAO,cAAc,OAAO,UAAU;AACpD,iBAAW,OAAO,KAAK;AAAA,IACzB;AAGA,UAAM,iBAAiB,MAAM,KAAK,MAAM,WAAW,OAAO,CAAC,EAAE,IAAI,CAAC,MAAM,EAAE,KAAK;AAG/E,UAAM,eAAe,eAAe,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGzF,UAAM,SAAS,OAAO,WAAW,gBAAgB,MAAM;AAGvD,UAAM,cAAc,OAAO,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,eAAW,WAAW,OAAO,SAAS;AACpC,YAAM,WAAW,OAAO,QAAQ,IAAI;AAAA,IACtC;AAGA,QAAI,OAAO,QAAQ,SAAS,GAAG;AAC7B,8BAAwB,OAAO,SAAS,IAAI;AAAA,IAC9C;AAGA,UAAM,WAAW,KAAK,IAAI,IAAI;AAC9B,UAAM,eAAe,WAAW,SAAS,IAAI,cAAc,SAAS,WAAW,SAAS;AAExF,eAAW,EAAE,mBAAmB;AAAA,MAC9B,WAAW,KAAK,IAAI;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA,kBAAkB,WAAW;AAAA,MAC7B,aAAa,OAAO,KAAK;AAAA,MACzB;AAAA,MACA,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAED,WAAO;AAAA,EACT;AAEA,WAAS,aACP,YACA,QAC6C;AAC7C,UAAM,YAAY,KAAK,IAAI;AAG3B,UAAM,eAAe,WAAW,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,UAAM,WAAW,MAAM;AACvB,UAAM,kBAAkB,MAAM;AAC9B,UAAM,iBAAiB;AACvB,UAAM,cAAc;AACpB,UAAM,uBAAuB,KAAK,IAAI;AAGtC,4BAAwB,YAAY,KAAK;AAGzC,eAAW,SAAS,YAAY;AAC9B,YAAM,QAAQ,OAAO,cAAc,OAAO,UAAU;AACpD,iBAAW,OAAO,KAAK;AAAA,IACzB;AAGA,UAAM,SAAS,OAAO,WAAW,YAAY,MAAM;AAGnD,UAAM,cAAc,OAAO,KAAK,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAGrF,eAAW,WAAW,OAAO,SAAS;AACpC,YAAM,WAAW,OAAO,QAAQ,IAAI;AAAA,IACtC;AAGA,QAAI,OAAO,QAAQ,SAAS,GAAG;AAC7B,8BAAwB,OAAO,SAAS,IAAI;AAAA,IAC9C;AAGA,UAAM,WAAW,KAAK,IAAI,IAAI;AAE9B,eAAW,EAAE,mBAAmB;AAAA,MAC9B,WAAW,KAAK,IAAI;AAAA,MACpB;AAAA,MACA;AAAA,MACA;AAAA,MACA,kBAAkB,WAAW;AAAA,MAC7B,aAAa,OAAO,KAAK;AAAA,MACzB,cAAc;AAAA;AAAA,MACd,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAED,WAAO;AAAA,EACT;AAEA,WAAS,WAAsC;AAC7C,WAAO;AAAA,MACL,YAAY,IAAI,IAAI,MAAM,UAAU;AAAA,MACpC,mBAAmB,IAAI,IAAI,MAAM,iBAAiB;AAAA,MAClD,gBAAgB,MAAM;AAAA,MACtB,aAAa,MAAM;AAAA,MACnB,sBAAsB,MAAM;AAAA,IAC9B;AAAA,EACF;AAEA,WAAS,aAAa,eAAgD;AACpE,YAAQ;AAAA,MACN,YAAY,IAAI,IAAI,cAAc,UAAU;AAAA,MAC5C,mBAAmB,IAAI,IAAI,cAAc,iBAAiB;AAAA,MAC1D,gBAAgB,cAAc;AAAA,MAC9B,aAAa,cAAc;AAAA,MAC3B,sBAAsB,cAAc;AAAA,IACtC;AAAA,EACF;AAEA,WAAS,QAAc;AACrB,YAAQ;AAAA,MACN,YAAY,oBAAI,IAAI;AAAA,MACpB,mBAAmB,oBAAI,IAAI;AAAA,MAC3B,gBAAgB;AAAA,MAChB,aAAa;AAAA,MACb,sBAAsB,KAAK,IAAI;AAAA,IACjC;AAAA,EACF;AAEA,WAAS,WAAW;AAClB,WAAO;AAAA,MACL,eAAe,MAAM,WAAW;AAAA,MAChC,aAAa,MAAM,kBAAkB;AAAA,MACrC,gBAAgB,MAAM;AAAA,MACtB,aAAa,MAAM;AAAA,MACnB,sBAAsB,MAAM;AAAA,IAC9B;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5QO,SAAS,sBAAsB,QAAgD;AACpF,QAAM,YAAY,2BAA2B,MAAM;AACnD,QAAM,EAAE,YAAY,YAAY,IAAI;AAGpC,MAAI,gBAAgB;AACpB,MAAI,iBAAiB;AACrB,MAAI,iBAAgC,CAAC;AACrC,MAAI,oBAAoB;AAExB,WAAS,OAAO,SAAiB,WAAoC,CAAC,GAAW;AAE/E,UAAM,aAAa,uBAAuB,OAAO;AAEjD,QAAI,WAAW,WAAW,EAAG,QAAO;AAGpC,UAAM,sBAAsB,WAAW,IAAI,CAAC,WAAW;AAAA,MACrD,GAAG;AAAA,MACH,UAAU,EAAE,GAAG,MAAM,UAAU,GAAG,SAAS;AAAA,IAC7C,EAAE;AAGF,UAAM,SAAS,UAAU,oBAAoB,qBAAqB,WAAW;AAG7E,qBAAiB,WAAW;AAC5B,sBAAkB,OAAO,QAAQ;AACjC,qBAAiB,OAAO;AACxB,wBAAoB,OAAO;AAG3B,QAAI,eAAe,SAAS,YAAY;AAEtC,YAAM,SAAS,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAC3E,YAAM,SAAS,OAAO,MAAM,GAAG,UAAU;AACzC,YAAM,WAAW,OAAO,MAAM,UAAU;AAExC,uBAAiB;AACjB,wBAAkB,SAAS;AAAA,IAC7B;AAEA,WAAO,WAAW;AAAA,EACpB;AAEA,WAAS,aAAqB;AAE5B,UAAM,SAAS,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAC3E,WAAO,OAAO,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,MAAM;AAAA,EACjD;AAEA,WAAS,aAA4B;AAEnC,WAAO,CAAC,GAAG,cAAc,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,YAAY,EAAE,SAAS;AAAA,EACrE;AAEA,WAAS,WAAiC;AACxC,UAAM,iBAAiB,UAAU,SAAS;AAC1C,UAAM,gBAAgB,eAAe,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAE1F,WAAO;AAAA,MACL;AAAA,MACA,gBAAgB,eAAe;AAAA,MAC/B;AAAA,MACA;AAAA,MACA;AAAA,MACA,WAAW;AAAA,QACT,eAAe,eAAe;AAAA,QAC9B,aAAa,eAAe;AAAA,QAC5B,aAAa,eAAe;AAAA,MAC9B;AAAA,IACF;AAAA,EACF;AAEA,WAAS,QAAc;AACrB,oBAAgB;AAChB,qBAAiB;AACjB,qBAAiB,CAAC;AAClB,wBAAoB;AACpB,cAAU,MAAM;AAAA,EAClB;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACpKA,SAAS,cAAc,kBAAkB;AACzC,OAAO,cAAc;AAuDrB,SAAS,aAAa,QAAwB;AAC5C,QAAM,aAAY,oBAAI,KAAK,GAAE,YAAY,EAAE,QAAQ,SAAS,GAAG;AAC/D,QAAM,aAAa,GAAG,MAAM,WAAW,SAAS;AAEhD,MAAI;AACF,iBAAa,QAAQ,UAAU;AAC/B,YAAQ,IAAI,iCAA4B,UAAU,EAAE;AACpD,WAAO;AAAA,EACT,SAAS,OAAO;AACd,YAAQ,MAAM,qCAAqC,KAAK,EAAE;AAC1D,WAAO;AAAA,EACT;AACF;AAYA,eAAsB,eAAe,QAAmC;AAEtE,MAAI;AACJ,MAAI;AACF,SAAK,IAAI,SAAS,MAAM;AAGxB,UAAM,iBAAiB,GAAG,OAAO,eAAe,EAAE,QAAQ,KAAK,CAAC;AAChE,QAAI,mBAAmB,MAAM;AAC3B,cAAQ,MAAM,sCAAiC;AAG/C,UAAI,WAAW,MAAM,GAAG;AACtB,cAAM,aAAa,aAAa,MAAM;AACtC,YAAI,YAAY;AACd,kBAAQ,IAAI,sBAAsB,UAAU,EAAE;AAAA,QAChD;AAAA,MACF;AAGA,cAAQ,IAAI,iCAAiC;AAC7C,SAAG,MAAM;AACT,WAAK,IAAI,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,MAAM,mCAA8B,KAAK;AAGjD,QAAI,WAAW,MAAM,GAAG;AACtB,mBAAa,MAAM;AACnB,cAAQ,IAAI,0BAA0B;AAAA,IACxC;AAEA,SAAK,IAAI,SAAS,MAAM;AAAA,EAC1B;AAGA,KAAG,OAAO,oBAAoB;AAG9B,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAYP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAQP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GASP;AAGD,KAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAMP;AAGD,QAAM,eAAe,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,GAI/B;AAED,QAAM,eAAe,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,GAI/B;AAED,QAAM,UAAU,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,GAO1B;AAED,QAAM,kBAAkB,GAAG,QAAQ,wCAAwC;AAC3E,QAAM,kBAAkB,GAAG,QAAQ,wCAAwC;AAE3E,SAAO;AAAA,IACL,MAAM,IAAI,OAAmC;AAC3C,YAAM,cAAc,GAAG,YAAY,MAAM;AACvC,qBAAa;AAAA,UACX,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM;AAAA,UACN,MAAM,SAAS,IAAI;AAAA,QACrB;AAEA,qBAAa;AAAA,UACX,MAAM;AAAA,UACN,MAAM;AAAA,UACN,KAAK,UAAU,MAAM,IAAI;AAAA,UACzB,KAAK,UAAU,MAAM,QAAQ;AAAA,QAC/B;AAAA,MACF,CAAC;AAED,kBAAY;AAAA,IACd;AAAA,IAEA,MAAM,IAAI,IAAyC;AACjD,YAAM,MAAM,QAAQ,IAAI,EAAE;AAE1B,UAAI,CAAC,KAAK;AACR,eAAO;AAAA,MACT;AAEA,YAAM,IAAI;AAcV,aAAO;AAAA,QACL,IAAI,EAAE;AAAA,QACN,SAAS,EAAE;AAAA,QACX,MAAM,EAAE;AAAA,QACR,WAAW,EAAE;AAAA,QACb,OAAO,EAAE;AAAA,QACT,KAAK,EAAE;AAAA,QACP,OAAO,EAAE;AAAA,QACT,aAAa,EAAE;AAAA,QACf,MAAM,EAAE,OAAO,KAAK,MAAM,EAAE,IAAI,IAAI,CAAC;AAAA,QACrC,UAAU,EAAE,WAAW,KAAK,MAAM,EAAE,QAAQ,IAAI,CAAC;AAAA,QACjD,QAAQ,EAAE,WAAW;AAAA,MACvB;AAAA,IACF;AAAA,IAEA,MAAM,MAAM,SAAqD;AAC/D,UAAI,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AASV,YAAM,SAAoB,CAAC;AAE3B,UAAI,QAAQ,OAAO;AACjB,eAAO;AACP,eAAO,KAAK,QAAQ,KAAK;AAAA,MAC3B;AAEA,UAAI,QAAQ,aAAa,QAAW;AAClC,eAAO;AACP,eAAO,KAAK,QAAQ,QAAQ;AAAA,MAC9B;AAEA,UAAI,QAAQ,aAAa,QAAW;AAClC,eAAO;AACP,eAAO,KAAK,QAAQ,QAAQ;AAAA,MAC9B;AAEA,UAAI,QAAQ,WAAW,QAAW;AAChC,eAAO;AACP,eAAO,KAAK,QAAQ,SAAS,IAAI,CAAC;AAAA,MACpC;AAEA,aAAO;AAEP,UAAI,QAAQ,OAAO;AACjB,eAAO;AACP,eAAO,KAAK,QAAQ,KAAK;AAAA,MAC3B;AAEA,UAAI,QAAQ,QAAQ;AAClB,eAAO;AACP,eAAO,KAAK,QAAQ,MAAM;AAAA,MAC5B;AAEA,YAAM,OAAO,GAAG,QAAQ,GAAG;AAC3B,YAAM,OAAO,KAAK,IAAI,GAAG,MAAM;AAE/B,aAAO,KAAK,IAAI,CAAC,QAAQ;AACvB,cAAM,IAAI;AAcV,eAAO;AAAA,UACL,IAAI,EAAE;AAAA,UACN,SAAS,EAAE;AAAA,UACX,MAAM,EAAE;AAAA,UACR,WAAW,EAAE;AAAA,UACb,OAAO,EAAE;AAAA,UACT,KAAK,EAAE;AAAA,UACP,OAAO,EAAE;AAAA,UACT,aAAa,EAAE;AAAA,UACf,MAAM,EAAE,OAAO,KAAK,MAAM,EAAE,IAAI,IAAI,CAAC;AAAA,UACrC,UAAU,EAAE,WAAW,KAAK,MAAM,EAAE,QAAQ,IAAI,CAAC;AAAA,UACjD,QAAQ,EAAE,WAAW;AAAA,QACvB;AAAA,MACF,CAAC;AAAA,IACH;AAAA,IAEA,MAAM,OAAO,IAA2B;AACtC,YAAM,cAAc,GAAG,YAAY,MAAM;AACvC,wBAAgB,IAAI,EAAE;AACtB,wBAAgB,IAAI,EAAE;AAAA,MACxB,CAAC;AAED,kBAAY;AAAA,IACd;AAAA,IAEA,MAAM,OAA0B;AAC9B,YAAM,OAAO,GAAG,QAAQ,8BAA8B;AACtD,YAAM,OAAO,KAAK,IAAI;AACtB,aAAO,KAAK,IAAI,CAAC,MAAM,EAAE,EAAE;AAAA,IAC7B;AAAA,IAEA,MAAM,UAA2B;AAC/B,YAAM,SAAS,GAAG,QAAQ,6CAA6C,EAAE,IAAI;AAK7E,SAAG,KAAK,0CAA0C;AAElD,SAAG,KAAK,QAAQ;AAEhB,YAAM,QAAQ,GAAG,QAAQ,6CAA6C,EAAE,IAAI;AAI5E,aAAO,OAAO,QAAQ,MAAM;AAAA,IAC9B;AAAA,IAEA,MAAM,QAAuB;AAC3B,SAAG,MAAM;AAAA,IACX;AAAA,IAEA,MAAM,mBAAmB,OAAqD;AAC5E,YAAM,OAAO,GAAG,QAAQ;AAAA;AAAA;AAAA,OAGvB;AAED,WAAK;AAAA,QACH,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,QACN,MAAM;AAAA,MACR;AAAA,IACF;AAAA,IAEA,MAAM,uBAAqD;AACzD,YAAM,OAAO,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA,OAIvB;AAED,YAAM,OAAO,KAAK,IAAI;AACtB,aAAO;AAAA,IACT;AAAA,IAEA,MAAM,yBAAwC;AAC5C,SAAG,KAAK,gCAAgC;AAAA,IAC1C;AAAA,EACF;AACF;;;AC3WO,SAAS,wBAAyC;AACvD,QAAM,SAAS,mBAAmB,EAAE,YAAY,IAAI,gBAAgB,KAAK,CAAC;AAE1E,WAAS,YAAY,SAA2C;AAC9D,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM,gBAAgB,QAAQ;AAG9B,UAAM,MAAM,KAAK,IAAI;AACrB,UAAM,aAAa,QAAQ,OAAO,CAAC,UAAU;AAC3C,YAAM,gBAAgB,MAAM,MAAM,aAAa;AAC/C,YAAM,QAAQ,OAAO,eAAe,cAAc,MAAM,GAAG;AAC3D,aAAO,QAAQ;AAAA,IACjB,CAAC;AAED,UAAM,iBAAiB,gBAAgB,WAAW;AAGlD,UAAM,kBAAkB,eAAe,UAAU;AACjD,UAAM,SAAS,gBAAgB,eAAe;AAG9C,UAAM,eAAe,IAAI,IAAI,gBAAgB,QAAQ,CAAC,MAAM,EAAE,QAAQ,IAAI,CAAC,MAAM,EAAE,EAAE,CAAC,CAAC;AACvF,UAAM,gBAAgB,WAAW,OAAO,CAAC,MAAM,CAAC,aAAa,IAAI,EAAE,EAAE,CAAC;AAGtE,UAAM,OAAO,CAAC,GAAG,QAAQ,GAAG,aAAa;AACzC,UAAM,UAAU,QAAQ,OAAO,CAAC,MAAM,CAAC,KAAK,KAAK,CAAC,MAAM,EAAE,OAAO,EAAE,EAAE,CAAC;AAEtE,UAAM,oBAAoB,gBAAgB,OAAO,CAAC,KAAK,MAAM,OAAO,EAAE,QAAQ,SAAS,IAAI,CAAC;AAE5F,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,eAAe;AAAA,MACf,cAAc,KAAK;AAAA,MACnB;AAAA,MACA;AAAA,MACA,kBAAkB,gBAAgB,IAAI,KAAK,SAAS,gBAAgB;AAAA,MACpE,YAAY,KAAK,IAAI,IAAI;AAAA,IAC3B;AAAA,EACF;AAEA,WAAS,eAAe,SAA0C;AAChE,UAAM,SAA2B,CAAC;AAClC,UAAM,YAAY,oBAAI,IAAY;AAGlC,aAAS,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AACvC,YAAM,QAAQ,QAAQ,CAAC;AACvB,UAAI,CAAC,SAAS,UAAU,IAAI,MAAM,EAAE,EAAG;AAEvC,YAAM,aAAa,QAAQ,OAAO,CAAC,GAAG,QAAQ,QAAQ,KAAK,EAAE,SAAS,MAAM,IAAI;AAEhF,UAAI,WAAW,SAAS,GAAG;AACzB,cAAM,QAAwB;AAAA,UAC5B,SAAS,CAAC,OAAO,GAAG,UAAU;AAAA,UAC9B,YAAY;AAAA;AAAA,QACd;AACA,eAAO,KAAK,KAAK;AAGjB,kBAAU,IAAI,MAAM,EAAE;AACtB,mBAAW,OAAO,YAAY;AAC5B,oBAAU,IAAI,IAAI,EAAE;AAAA,QACtB;AAAA,MACF;AAAA,IACF;AAGA,aAAS,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AACvC,YAAM,SAAS,QAAQ,CAAC;AACxB,UAAI,CAAC,UAAU,UAAU,IAAI,OAAO,EAAE,EAAG;AAEzC,eAAS,IAAI,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK;AAC3C,cAAM,SAAS,QAAQ,CAAC;AACxB,YAAI,CAAC,UAAU,UAAU,IAAI,OAAO,EAAE,EAAG;AAEzC,cAAM,aAAa,iBAAiB,OAAO,SAAS,OAAO,OAAO;AAElE,YAAI,cAAc,MAAM;AACtB,gBAAM,QAAwB;AAAA,YAC5B,SAAS,CAAC,QAAQ,MAAM;AAAA,YACxB;AAAA,UACF;AACA,iBAAO,KAAK,KAAK;AAEjB,oBAAU,IAAI,OAAO,EAAE;AACvB,oBAAU,IAAI,OAAO,EAAE;AACvB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,gBAAgB,QAAyC;AAChE,UAAM,SAAwB,CAAC;AAE/B,eAAW,SAAS,QAAQ;AAE1B,YAAM,SAAS,CAAC,GAAG,MAAM,OAAO,EAAE,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;AAClE,YAAM,OAAO,OAAO,CAAC;AACrB,UAAI,CAAC,KAAM;AAGX,YAAM,mBAAmB,MAAM,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,aAAa,CAAC;AAGhF,YAAM,UAAU,IAAI,IAAI,MAAM,QAAQ,QAAQ,CAAC,MAAM,EAAE,IAAI,CAAC;AAE5D,aAAO,KAAK;AAAA,QACV,GAAG;AAAA,QACH,aAAa;AAAA,QACb,MAAM,MAAM,KAAK,OAAO;AAAA,MAC1B,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT;AAQA,WAAS,iBAAiB,OAAe,OAAuB;AAC9D,UAAM,SAAS,SAAS,KAAK;AAC7B,UAAM,SAAS,SAAS,KAAK;AAG7B,UAAM,QAAQ,oBAAI,IAAI,CAAC,GAAG,QAAQ,GAAG,MAAM,CAAC;AAG5C,UAAM,OAA+B,CAAC;AACtC,UAAM,OAA+B,CAAC;AAEtC,eAAW,QAAQ,OAAO;AACxB,WAAK,IAAI,IAAI,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAC9C,WAAK,IAAI,IAAI,OAAO,OAAO,CAAC,MAAM,MAAM,IAAI,EAAE;AAAA,IAChD;AAGA,QAAI,aAAa;AACjB,QAAI,OAAO;AACX,QAAI,OAAO;AAEX,eAAW,QAAQ,OAAO;AACxB,YAAM,SAAS,KAAK,IAAI,KAAK;AAC7B,YAAM,SAAS,KAAK,IAAI,KAAK;AAC7B,oBAAc,SAAS;AACvB,cAAQ,SAAS;AACjB,cAAQ,SAAS;AAAA,IACnB;AAEA,WAAO,KAAK,KAAK,IAAI;AACrB,WAAO,KAAK,KAAK,IAAI;AAErB,QAAI,SAAS,KAAK,SAAS,EAAG,QAAO;AAErC,WAAO,cAAc,OAAO;AAAA,EAC9B;AAEA,WAAS,SAAS,MAAwB;AACxC,WAAO,KACJ,YAAY,EACZ,MAAM,KAAK,EACX,OAAO,CAAC,SAAS,KAAK,SAAS,CAAC;AAAA,EACrC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC9MA,SAAS,YAAY;AACrB,SAAS,cAAAC,aAAY,WAAW,cAAc,YAAY,qBAAqB;AAC/E,SAAS,SAAS,YAAY;AAC9B,SAAS,qBAAqB;AAyCvB,SAAS,sBAAqC;AAInD,WAAS,gBAAgB,SAAqD;AAC5E,QAAI,CAACC,YAAW,OAAO,GAAG;AACxB,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAEA,QAAI;AACF,YAAM,SAAS,aAAa,SAAS,OAAO,EAAE,KAAK;AACnD,YAAM,MAAM,OAAO,SAAS,QAAQ,EAAE;AAEtC,UAAI,OAAO,MAAM,GAAG,GAAG;AACrB,eAAO,EAAE,SAAS,MAAM;AAAA,MAC1B;AAGA,UAAI;AACF,gBAAQ,KAAK,KAAK,CAAC;AACnB,eAAO,EAAE,SAAS,MAAM,IAAI;AAAA,MAC9B,QAAQ;AAEN,mBAAW,OAAO;AAClB,eAAO,EAAE,SAAS,MAAM;AAAA,MAC1B;AAAA,IACF,QAAQ;AACN,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,EACF;AAKA,WAAS,aAAa,SAAiB,KAAmB;AAExD,UAAM,MAAM,QAAQ,OAAO;AAC3B,QAAI,CAACA,YAAW,GAAG,GAAG;AACpB,gBAAU,KAAK,EAAE,WAAW,KAAK,CAAC;AAAA,IACpC;AAEA,kBAAc,SAAS,OAAO,GAAG,GAAG,OAAO;AAAA,EAC7C;AAKA,WAAS,cAAc,SAAuB;AAC5C,QAAIA,YAAW,OAAO,GAAG;AACvB,iBAAW,OAAO;AAAA,IACpB;AAAA,EACF;AAEA,iBAAe,MAAM,QAAiD;AACpE,UAAM,EAAE,SAAS,QAAQ,IAAI,OAAO;AAGpC,UAAMC,UAAS,gBAAgB,OAAO;AACtC,QAAIA,QAAO,SAAS;AAClB,aAAO;AAAA,QACL,SAAS;AAAA,QACT,KAAKA,QAAO;AAAA,QACZ,SAAS,+BAA+BA,QAAO,GAAG;AAAA,QAClD,OAAO;AAAA,MACT;AAAA,IACF;AAEA,QAAI;AAEF,YAAMC,cAAa,cAAc,YAAY,GAAG;AAChD,YAAMC,aAAY,QAAQD,WAAU;AACpC,YAAM,aAAa,KAAKC,YAAW,UAAU;AAE7C,YAAM,QAAQ,KAAK,YAAY,CAAC,GAAG;AAAA,QACjC,UAAU;AAAA,QACV,OAAO;AAAA,QACP,KAAK;AAAA,UACH,GAAG,QAAQ;AAAA,UACX,cAAc,KAAK,UAAU,MAAM;AAAA,UACnC,gBAAgB;AAAA,UAChB,gBAAgB;AAAA,QAClB;AAAA,MACF,CAAC;AAGD,YAAM,MAAM;AAGZ,UAAI,MAAM,KAAK;AACb,qBAAa,SAAS,MAAM,GAAG;AAE/B,eAAO;AAAA,UACL,SAAS;AAAA,UACT,KAAK,MAAM;AAAA,UACX,SAAS,uBAAuB,MAAM,GAAG;AAAA,QAC3C;AAAA,MACF;AAEA,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO;AAAA,MACT;AAAA,IACF,SAAS,OAAO;AACd,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D;AAAA,IACF;AAAA,EACF;AAEA,iBAAe,KAAK,QAAgD;AAClE,UAAM,EAAE,QAAQ,IAAI,OAAO;AAE3B,UAAMF,UAAS,gBAAgB,OAAO;AAEtC,QAAI,CAACA,QAAO,WAAW,CAACA,QAAO,KAAK;AAClC,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,MACX;AAAA,IACF;AAEA,QAAI;AAEF,cAAQ,KAAKA,QAAO,KAAK,SAAS;AAGlC,YAAM,UAAU;AAChB,YAAM,WAAW;AACjB,UAAI,SAAS;AAEb,aAAO,SAAS,SAAS;AACvB,YAAI;AACF,kBAAQ,KAAKA,QAAO,KAAK,CAAC;AAE1B,gBAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,QAAQ,CAAC;AAC5D,oBAAU;AAAA,QACZ,QAAQ;AAEN,wBAAc,OAAO;AACrB,iBAAO;AAAA,YACL,SAAS;AAAA,YACT,SAAS,uBAAuBA,QAAO,GAAG;AAAA,UAC5C;AAAA,QACF;AAAA,MACF;AAGA,UAAI;AACF,gBAAQ,KAAKA,QAAO,KAAK,SAAS;AAClC,sBAAc,OAAO;AACrB,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,4BAA4BA,QAAO,GAAG;AAAA,QACjD;AAAA,MACF,QAAQ;AACN,sBAAc,OAAO;AACrB,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,uBAAuBA,QAAO,GAAG;AAAA,QAC5C;AAAA,MACF;AAAA,IACF,SAAS,OAAO;AACd,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,QACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D;AAAA,IACF;AAAA,EACF;AAEA,iBAAe,OAAO,QAAkD;AACtE,UAAM,EAAE,QAAQ,IAAI,OAAO;AAE3B,UAAM,eAAe,gBAAgB,OAAO;AAE5C,QAAI,CAAC,aAAa,WAAW,CAAC,aAAa,KAAK;AAC9C,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS;AAAA,MACX;AAAA,IACF;AAGA,UAAM,UAAU,WAAW,EAAE,YAAY;AAEzC,WAAO;AAAA,MACL,SAAS;AAAA,MACT,KAAK,aAAa;AAAA,MAClB,QAAQ,QAAQ,OAAO;AAAA,MACvB,iBAAiB,QAAQ,OAAO;AAAA,MAChC,aAAa,QAAQ,aAAa;AAAA,MAClC,SAAS,uBAAuB,aAAa,GAAG;AAAA,IAClD;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACvPA,SAAS,gBAAAG,eAAc,gBAAgB;AAoDhC,SAAS,oBAAiC;AAE/C,QAAM,YAAY,oBAAI,IAA0B;AAEhD,WAAS,aAAa,UAA4B;AAChD,QAAI;AAEF,YAAM,QAAQ,SAAS,QAAQ;AAC/B,YAAM,cAAc,MAAM;AAC1B,YAAM,kBAAkB,MAAM;AAG9B,UAAI,MAAM,UAAU,IAAI,QAAQ;AAEhC,UAAI,CAAC,KAAK;AAER,cAAM;AAAA,UACJ,MAAM;AAAA,UACN,UAAU;AAAA,UACV,aAAa;AAAA,UACb,cAAc;AAAA,UACd,UAAU;AAAA,QACZ;AACA,kBAAU,IAAI,UAAU,GAAG;AAAA,MAC7B;AAGA,UAAI,cAAc,IAAI,YAAY,gBAAgB,IAAI,UAAU;AAE9D,YAAI,cAAc,IAAI,UAAU;AAE9B,cAAI,WAAW;AACf,cAAI,cAAc;AAAA,QACpB;AACA,eAAO,CAAC;AAAA,MACV;AAGA,YAAM,SAAS,OAAO,MAAM,cAAc,IAAI,QAAQ;AACtD,YAAM,KAAKA,cAAa,QAAQ;AAChC,SAAG,KAAK,QAAQ,GAAG,IAAI,UAAU,WAAW;AAG5C,YAAM,cAAc,IAAI,cAAc,OAAO,SAAS,OAAO,GAAG,MAAM,IAAI;AAG1E,YAAM,cAAc,WAAW,IAAI,KAAK;AAGxC,UAAI,WAAW;AACf,UAAI,cAAc;AAClB,UAAI,eAAe;AACnB,UAAI,WAAW;AAGf,aAAO,WAAW,OAAO,CAAC,SAAS,KAAK,KAAK,EAAE,SAAS,CAAC;AAAA,IAC3D,SAAS,QAAQ;AAGf,aAAO,CAAC;AAAA,IACV;AAAA,EACF;AAEA,WAAS,YAAY,UAAuC;AAC1D,WAAO,UAAU,IAAI,QAAQ,KAAK;AAAA,EACpC;AAEA,WAAS,cAAc,UAAwB;AAC7C,cAAU,OAAO,QAAQ;AAAA,EAC3B;AAEA,WAAS,WAAiB;AACxB,cAAU,MAAM;AAAA,EAClB;AAEA,WAAS,kBAA4B;AACnC,WAAO,MAAM,KAAK,UAAU,KAAK,CAAC;AAAA,EACpC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;AC5IA,SAAyB,aAAa,YAAAC,WAAU,aAAa;AAC7D,SAAS,eAAe;AACxB,SAAS,WAAAC,UAAS,QAAAC,aAAY;AAoEvB,SAAS,qBAAqB,QAA8C;AACjF,QAAM,EAAE,QAAQ,aAAa,YAAY,QAAQ,IAAI;AACrD,QAAM,EAAE,UAAU,OAAO,OAAO,IAAI;AAGpC,QAAM,YAAY,oBAAI,IAA6B;AACnD,QAAM,cAAc,kBAAkB;AAGtC,QAAM,WAAwB,CAAC;AAG/B,QAAM,iBAAiB,oBAAI,IAA4B;AAKvD,WAAS,iBAAyB;AAChC,WAAOC,MAAK,QAAQ,GAAG,WAAW,UAAU;AAAA,EAC9C;AAMA,WAAS,aAAa,UAA0B;AAC9C,UAAM,WAAW,SAAS,MAAM,OAAO,EAAE,IAAI,KAAK;AAClD,WAAO,SAAS,QAAQ,YAAY,EAAE;AAAA,EACxC;AAKA,WAAS,YAAY,WAAoC;AACvD,QAAI,WAAW,UAAU,IAAI,SAAS;AAEtC,QAAI,CAAC,UAAU;AACb,iBAAW,sBAAsB;AAAA,QAC/B,aAAa,SAAS;AAAA,QACtB;AAAA,QACA;AAAA,QACA,YAAY,SAAS;AAAA,QACrB,0BAA0B;AAAA;AAAA,MAC5B,CAAC;AACD,gBAAU,IAAI,WAAW,QAAQ;AAAA,IACnC;AAEA,WAAO;AAAA,EACT;AAKA,WAAS,iBAAiB,UAAwB;AAEhD,UAAM,gBAAgB,eAAe,IAAI,QAAQ;AACjD,QAAI,eAAe;AACjB,mBAAa,aAAa;AAAA,IAC5B;AAGA,UAAM,QAAQ,WAAW,MAAM;AAC7B,UAAI;AAEF,cAAM,WAAW,YAAY,aAAa,QAAQ;AAElD,YAAI,SAAS,WAAW,EAAG;AAG3B,cAAM,UAAU,SAAS,KAAK,IAAI;AAClC,cAAM,YAAY,aAAa,QAAQ;AACvC,cAAM,WAAW,YAAY,SAAS;AAGtC,iBAAS,OAAO,SAAS,EAAE,WAAW,SAAS,CAAC;AAGhD,cAAM,QAAQ,SAAS,SAAS;AAChC,YAAI,MAAM,iBAAiB,SAAS,uBAAuB;AAEzD,qBAAW,EAAE,aAAa;AAAA,YACxB,iBAAiB,UAAU;AAAA,YAC3B,aAAa,QAAQ,YAAY,EAAE;AAAA,UACrC,CAAC;AAGD,cAAI,YAAY;AACd,kBAAM,eAAe,oBAAoB,WAAW,QAAQ;AAC5D,uBAAW,WAAW,YAAY;AAAA,UACpC;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,YAAI,SAAS;AACX,kBAAQ,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,KAAK,CAAC,CAAC;AAAA,QACnE;AAAA,MACF,UAAE;AACA,uBAAe,OAAO,QAAQ;AAAA,MAChC;AAAA,IACF,GAAG,SAAS,UAAU;AAEtB,mBAAe,IAAI,UAAU,KAAK;AAAA,EACpC;AAKA,WAAS,eAAe,KAAuB;AAC7C,UAAM,QAAkB,CAAC;AAEzB,QAAI;AACF,YAAM,UAAU,YAAY,GAAG;AAE/B,iBAAW,SAAS,SAAS;AAC3B,cAAM,WAAWA,MAAK,KAAK,KAAK;AAChC,cAAM,OAAOC,UAAS,QAAQ;AAE9B,YAAI,KAAK,YAAY,GAAG;AAEtB,gBAAM,KAAK,GAAG,eAAe,QAAQ,CAAC;AAAA,QACxC,WAAW,MAAM,SAAS,QAAQ,GAAG;AAEnC,gBAAM,UAAU,SAAS,cAAc,KAAK,CAAC,YAAY;AAEvD,kBAAM,QAAQ,IAAI;AAAA,cAChB,QAAQ,QAAQ,SAAS,IAAI,EAAE,QAAQ,OAAO,WAAW,EAAE,QAAQ,OAAO,KAAK;AAAA,YACjF;AACA,mBAAO,MAAM,KAAK,QAAQ;AAAA,UAC5B,CAAC;AAED,cAAI,SAAS;AACX,kBAAM,KAAK,QAAQ;AAAA,UACrB;AAAA,QACF;AAAA,MACF;AAAA,IACF,SAAS,QAAQ;AAAA,IAEjB;AAEA,WAAO;AAAA,EACT;AAKA,WAAS,oBAAoB,WAAmB,UAAyC;AACvF,UAAM,QAAQ,SAAS,SAAS;AAChC,UAAM,UAAU,SAAS,WAAW;AACpC,UAAM,cAAc,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,eAAe,EAAE,OAAO,GAAG,CAAC;AAEjF,WAAO;AAAA,MACL;AAAA,MACA,aAAa,MAAM;AAAA,MACnB,iBAAiB,MAAM;AAAA,MACvB,WAAW,cAAc,KAAK,cAAc,MAAM,iBAAiB,cAAc;AAAA,MACjF,YAAY,MAAM;AAAA,MAClB,mBAAmB,MAAM;AAAA,IAC3B;AAAA,EACF;AAEA,iBAAe,QAAuB;AACpC,UAAM,cAAc,eAAe;AAGnC,UAAM,aAAa,eAAe,WAAW;AAG7C,UAAM,cAAc,oBAAI,IAAY;AAEpC,eAAW,QAAQ,YAAY;AAC7B,YAAM,MAAMC,SAAQ,IAAI;AAExB,UAAI,CAAC,YAAY,IAAI,GAAG,GAAG;AACzB,cAAM,UAAU,MAAM,KAAK,EAAE,WAAW,MAAM,GAAG,CAAC,YAAY,aAAa;AACzE,cAAI,UAAU,SAAS,QAAQ,GAAG;AAChC,kBAAM,WAAWF,MAAK,KAAK,QAAQ;AACnC,6BAAiB,QAAQ;AAAA,UAC3B;AAAA,QACF,CAAC;AAED,iBAAS,KAAK,OAAO;AACrB,oBAAY,IAAI,GAAG;AAAA,MACrB;AAAA,IACF;AAGA,UAAM,kBAAkB,MAAM,aAAa,EAAE,WAAW,KAAK,GAAG,CAAC,YAAY,aAAa;AACxF,UAAI,UAAU,SAAS,QAAQ,GAAG;AAChC,cAAM,WAAWA,MAAK,aAAa,QAAQ;AAC3C,yBAAiB,QAAQ;AAAA,MAC3B;AAAA,IACF,CAAC;AAED,aAAS,KAAK,eAAe;AAG7B,eAAW,EAAE,aAAa;AAAA,MACxB,WAAW,KAAK,IAAI;AAAA,MACpB,iBAAiB,WAAW;AAAA,MAC5B,aAAa,QAAQ,YAAY,EAAE;AAAA,IACrC,CAAC;AAAA,EACH;AAEA,WAAS,OAAa;AAEpB,eAAW,WAAW,UAAU;AAC9B,cAAQ,MAAM;AAAA,IAChB;AACA,aAAS,SAAS;AAGlB,eAAW,SAAS,eAAe,OAAO,GAAG;AAC3C,mBAAa,KAAK;AAAA,IACpB;AACA,mBAAe,MAAM;AAGrB,cAAU,MAAM;AAGhB,gBAAY,SAAS;AAAA,EACvB;AAEA,WAAS,WAA2B;AAClC,UAAM,QAAwB,CAAC;AAE/B,eAAW,CAAC,WAAW,QAAQ,KAAK,UAAU,QAAQ,GAAG;AACvD,YAAM,KAAK,oBAAoB,WAAW,QAAQ,CAAC;AAAA,IACrD;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,gBAAgB,WAAwC;AAC/D,UAAM,WAAW,UAAU,IAAI,SAAS;AACxC,QAAI,CAAC,SAAU,QAAO;AAEtB,WAAO,oBAAoB,WAAW,QAAQ;AAAA,EAChD;AAEA,WAAS,gBAAgB,WAAyB;AAChD,UAAM,WAAW,UAAU,IAAI,SAAS;AACxC,QAAI,CAAC,SAAU;AAGf,UAAM,UAAU,SAAS,WAAW;AACpC,aAAS,MAAM;AACf,aAAS,OAAO,QAAQ,IAAI,CAAC,MAAM,EAAE,OAAO,EAAE,KAAK,MAAM,CAAC;AAG1D,QAAI,YAAY;AACd,YAAM,QAAQ,oBAAoB,WAAW,QAAQ;AACrD,iBAAW,WAAW,KAAK;AAAA,IAC7B;AAAA,EACF;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;;;ACxUA,SAAS,iBAAiB;AAC1B,SAAS,SAAS;;;AC6FX,IAAM,iBAA8B;AAAA,EACzC,SAAS;AAAA,IACP,WAAW;AAAA,IACX,gBAAgB;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,YAAY;AAAA,IACZ,gBAAgB;AAAA,EAClB;AAAA,EACA,QAAQ;AAAA,IACN,iBAAiB;AAAA,IACjB,gBAAgB;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,EACP,IAAI;AAAA,IACF,QAAQ;AAAA,IACR,QAAQ;AAAA,IACR,SAAS;AAAA,EACX;AAAA,EACA,iBAAiB;AAAA,EACjB,UAAU;AAAA,IACR,aAAa;AAAA,IACb,uBAAuB;AAAA,IACvB,eAAe,CAAC,YAAY;AAAA,IAC5B,SAAS;AAAA,IACT,SAAS;AAAA,IACT,YAAY;AAAA,IACZ,aAAa;AAAA,IACb,YAAY;AAAA,IACZ,uBAAuB;AAAA,EACzB;AACF;;;ADrGO,SAAS,qBAAqB,SAA2C;AAC9E,QAAM,EAAE,QAAQ,SAAS,eAAe,IAAI;AAE5C,QAAM,SAAS,IAAI,UAAU;AAAA,IAC3B,MAAM;AAAA,IACN,SAAS;AAAA,EACX,CAAC;AAED,uBAAqB,QAAQ,QAAQ,MAAM;AAC3C,oBAAkB,QAAQ,MAAM;AAChC,0BAAwB,QAAQ,MAAM;AAEtC,SAAO;AACT;AAQA,SAAS,qBAAqB,QAAmB,QAAkB,QAA2B;AAC5F,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,MAGF,aAAa;AAAA,QACX,SAAS,EAAE,OAAO,EAAE,SAAS,8BAA8B;AAAA,QAC3D,QAAQ,EACL,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,qDAAqD;AAAA,QACjE,SAAS,EACN,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,oDAAoD;AAAA,QAChE,WAAW,EACR,OAAO,EACP,IAAI,CAAC,EACL,IAAI,GAAG,EACP,SAAS,EACT,SAAS,oDAAoD;AAAA,MAClE;AAAA,IACF;AAAA,IACA,OAAO,EAAE,SAAS,QAAQ,SAAS,UAAU,MAAM;AACjD,UAAI;AACF,cAAM,kBAAkB,YACpB,EAAE,GAAG,QAAQ,SAAS,EAAE,GAAG,OAAO,SAAS,UAAU,EAAE,IACvD;AAEJ,cAAM,UAAU,qBAAqB,QAAQ,eAAe;AAC5D,cAAM,SAAS,MAAM,QAAQ,SAAS,SAAS;AAAA,UAC7C;AAAA,UACA;AAAA,UACA;AAAA,QACF,CAAC;AAED,cAAM,WAAW;AAAA,UACf,kBAAkB,OAAO;AAAA,UACzB,cAAc,OAAO;AAAA,UACrB,aAAa,OAAO;AAAA,UACpB,WAAW,IAAI,OAAO,YAAY,KAAK,QAAQ,CAAC,CAAC;AAAA,UACjD,kBAAkB,OAAO;AAAA,UACzB,aAAa,OAAO;AAAA,UACpB,YAAY,OAAO;AAAA,UACnB,mBAAmB,OAAO;AAAA,UAC1B,GAAI,WAAW,OAAO,UAAU,EAAE,SAAS,OAAO,QAAQ,IAAI,CAAC;AAAA,QACjE;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAQA,SAAS,kBAAkB,QAAmB,QAAwB;AACpE,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,MAEF,aAAa;AAAA,QACX,OAAO,EACJ,QAAQ,EACR,SAAS,EACT,QAAQ,KAAK,EACb,SAAS,4CAA4C;AAAA,MAC1D;AAAA,IACF;AAAA,IACA,OAAO,EAAE,MAAM,MAAM;AACnB,UAAI;AACF,YAAI,OAAO;AACT,gBAAM,OAAO,uBAAuB;AACpC,iBAAO;AAAA,YACL,SAAS;AAAA,cACP;AAAA,gBACE,MAAM;AAAA,gBACN,MAAM,KAAK;AAAA,kBACT;AAAA,oBACE,SAAS;AAAA,oBACT,eAAe;AAAA,oBACf,kBAAkB;AAAA,oBAClB,kBAAkB;AAAA,kBACpB;AAAA,kBACA;AAAA,kBACA;AAAA,gBACF;AAAA,cACF;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAEA,cAAM,QAAQ,MAAM,OAAO,qBAAqB;AAChD,cAAM,gBAAgB,MAAM;AAE5B,cAAM,mBAAmB,MAAM;AAAA,UAC7B,CAAC,KAAK,MAAM,OAAO,EAAE,gBAAgB,EAAE;AAAA,UACvC;AAAA,QACF;AAEA,cAAM,mBACJ,gBAAgB,IACZ,MAAM,OAAO,CAAC,KAAK,MAAM;AACvB,gBAAM,YACJ,EAAE,gBAAgB,KAAK,EAAE,gBAAgB,EAAE,gBAAgB,EAAE,gBAAgB;AAC/E,iBAAO,MAAM;AAAA,QACf,GAAG,CAAC,IAAI,gBACR;AAEN,cAAM,sBAAsB,MAAM,MAAM,GAAG,EAAE,EAAE,IAAI,CAAC,OAAO;AAAA,UACzD,WAAW,IAAI,KAAK,EAAE,SAAS,EAAE,YAAY;AAAA,UAC7C,cAAc,EAAE;AAAA,UAChB,aAAa,EAAE;AAAA,UACf,eAAe,EAAE;AAAA,UACjB,YAAY,EAAE;AAAA,UACd,WAAW,KACP,EAAE,gBAAgB,EAAE,gBAAgB,KAAK,IAAI,EAAE,eAAe,CAAC,IAAK,KACtE,QAAQ,CAAC,CAAC;AAAA,QACd,EAAE;AAEF,cAAM,WAAW;AAAA,UACf;AAAA,UACA;AAAA,UACA,kBAAkB,IAAI,mBAAmB,KAAK,QAAQ,CAAC,CAAC;AAAA,UACxD;AAAA,QACF;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAQA,SAAS,wBAAwB,QAAmB,QAAwB;AAC1E,SAAO;AAAA,IACL;AAAA,IACA;AAAA,MACE,OAAO;AAAA,MACP,aACE;AAAA,IAGJ;AAAA,IACA,YAAY;AACV,UAAI;AACF,cAAM,SAAS,MAAM,OAAO,KAAK;AACjC,cAAM,aAAa,MAAM,QAAQ;AAAA,UAC/B,OAAO,IAAI,OAAO,OAAO;AACvB,kBAAM,QAAQ,MAAM,OAAO,IAAI,EAAE;AACjC,mBAAO;AAAA,UACT,CAAC;AAAA,QACH;AAEA,cAAM,UAAU,WAAW,OAAO,CAAC,MAAM,MAAM,IAAI;AAEnD,cAAM,aAAa,sBAAsB;AACzC,cAAM,SAAS,WAAW,YAAY,OAAO;AAG7C,mBAAW,WAAW,OAAO,SAAS;AACpC,gBAAM,OAAO,OAAO,QAAQ,EAAE;AAAA,QAChC;AAEA,mBAAW,QAAQ,OAAO,MAAM;AAC9B,gBAAM,OAAO,IAAI,IAAI;AAAA,QACvB;AAGA,cAAM,OAAO,QAAQ;AAErB,cAAM,WAAW;AAAA,UACf,eAAe,OAAO;AAAA,UACtB,cAAc,OAAO;AAAA,UACrB,gBAAgB,OAAO;AAAA,UACvB,mBAAmB,OAAO;AAAA,UAC1B,kBAAkB,IAAI,OAAO,mBAAmB,KAAK,QAAQ,CAAC,CAAC;AAAA,UAC/D,YAAY,OAAO;AAAA,UACnB,iBAAiB;AAAA,QACnB;AAEA,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,UAAU,MAAM,CAAC;AAAA,YACxC;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAAS,OAAO;AACd,cAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACrE,eAAO;AAAA,UACL,SAAS;AAAA,YACP;AAAA,cACE,MAAM;AAAA,cACN,MAAM,KAAK,UAAU,EAAE,OAAO,QAAQ,CAAC;AAAA,YACzC;AAAA,UACF;AAAA,UACA,SAAS;AAAA,QACX;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;;;AEnSO,SAAS,aAAa,UAAU,OAAe;AACpD,SAAO;AAAA,IACL,MAAM,YAAoB,MAAuB;AAC/C,UAAI,SAAS;AACX,gBAAQ,MAAM,WAAW,OAAO,IAAI,GAAG,IAAI;AAAA,MAC7C;AAAA,IACF;AAAA,IACA,KAAK,YAAoB,MAAuB;AAC9C,cAAQ,KAAK,UAAU,OAAO,IAAI,GAAG,IAAI;AAAA,IAC3C;AAAA,IACA,KAAK,YAAoB,MAAuB;AAC9C,cAAQ,KAAK,UAAU,OAAO,IAAI,GAAG,IAAI;AAAA,IAC3C;AAAA,IACA,MAAM,YAAoB,MAAuB;AAC/C,cAAQ,MAAM,WAAW,OAAO,IAAI,GAAG,IAAI;AAAA,IAC7C;AAAA,EACF;AACF;","names":["randomUUID","randomUUID","randomUUID","randomUUID","allEntries","existsSync","existsSync","status","__filename","__dirname","readFileSync","statSync","dirname","join","join","statSync","dirname"]}