@stackmemoryai/stackmemory 0.3.14 → 0.3.15
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cli/commands/storage-tier.js +179 -0
- package/dist/cli/commands/storage-tier.js.map +7 -0
- package/dist/cli/index.js +1 -1
- package/dist/cli/index.js.map +1 -1
- package/dist/core/storage/two-tier-storage.js +762 -0
- package/dist/core/storage/two-tier-storage.js.map +7 -0
- package/package.json +1 -1
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
{
|
|
2
|
+
"version": 3,
|
|
3
|
+
"sources": ["../../../src/core/storage/two-tier-storage.ts"],
|
|
4
|
+
"sourcesContent": ["/**\n * Two-Tier Storage System for StackMemory\n * Implements STA-414: Local and Remote Storage with Migration\n *\n * Local Tiers:\n * - Young: < 1 day (complete retention in RAM)\n * - Mature: 1-7 days (selective retention with LZ4)\n * - Old: 7-30 days (critical only with ZSTD)\n *\n * Remote Tier:\n * - Infinite retention with TimeSeries DB + S3\n * - Monthly partitioning\n * - Cost-aware migration\n */\n\nimport Database from 'better-sqlite3';\nimport {\n S3Client,\n PutObjectCommand,\n GetObjectCommand,\n} from '@aws-sdk/client-s3';\nimport type { RedisClientType } from 'redis';\nimport { createClient as createRedisClient } from 'redis';\nimport { Pool } from 'pg';\nimport * as zlib from 'zlib';\nimport { promisify } from 'util';\nimport { v4 as uuidv4 } from 'uuid';\nimport { Logger } from '../monitoring/logger.js';\nimport type { Frame, Event, Anchor } from '../context/frame-manager.js';\n\n// LZ4 would be installed separately: npm install lz4\n// For now we'll use a placeholder\nconst lz4 = {\n encode: (data: Buffer) => data, // Placeholder\n decode: (data: Buffer) => data, // Placeholder\n};\n\nconst gzip = promisify(zlib.gzip);\nconst gunzip = promisify(zlib.gunzip);\n\nexport enum StorageTier {\n YOUNG = 'young', // < 1 day, in RAM\n MATURE = 'mature', // 1-7 days, LZ4\n OLD = 'old', // 7-30 days, ZSTD\n REMOTE = 'remote', // > 30 days, S3\n}\n\nexport interface TierConfig {\n name: StorageTier;\n maxAgeHours: number;\n compressionType: 'none' | 'lz4' | 'zstd' | 'gzip';\n retentionPolicy: 'complete' | 'selective' | 'critical';\n maxSizeMB: number;\n}\n\nexport interface MigrationTrigger {\n type: 'age' | 'size' | 'importance';\n threshold: number;\n action: 'migrate' | 'compress' | 'delete';\n}\n\nexport interface OfflineQueueItem {\n id: string;\n data: any;\n priority: 'high' | 'normal';\n timestamp: number;\n}\n\nexport interface TwoTierConfig {\n local: {\n dbPath: string;\n maxSizeGB: number;\n tiers: TierConfig[];\n };\n remote: {\n redis?: {\n url: string;\n ttlSeconds: number;\n };\n timeseries?: {\n connectionString: string;\n };\n s3: {\n bucket: string;\n region: string;\n accessKeyId?: string;\n secretAccessKey?: string;\n };\n };\n migration: {\n triggers: MigrationTrigger[];\n batchSize: number;\n intervalMs: number;\n offlineQueuePath?: string;\n };\n}\n\nexport interface StorageStats {\n localUsageMB: number;\n remoteUsageMB: number;\n tierDistribution: Record<StorageTier, number>;\n compressionRatio: number;\n migrationsPending: number;\n lastMigration: Date | null;\n}\n\nexport class TwoTierStorageSystem {\n private db: Database.Database;\n private redisClient?: RedisClientType;\n private timeseriesPool?: Pool;\n private s3Client: S3Client;\n private logger: Logger;\n private config: TwoTierConfig;\n private migrationTimer?: NodeJS.Timeout;\n private offlineQueue: OfflineQueueItem[] = [];\n private stats: StorageStats;\n\n constructor(config: TwoTierConfig) {\n this.config = config;\n this.logger = new Logger('TwoTierStorage');\n\n // Initialize local SQLite\n this.db = new Database(config.local.dbPath);\n this.initializeLocalStorage();\n\n // Initialize S3 client\n this.s3Client = new S3Client({\n region: config.remote.s3.region,\n credentials:\n config.remote.s3.accessKeyId && config.remote.s3.secretAccessKey\n ? {\n accessKeyId: config.remote.s3.accessKeyId,\n secretAccessKey: config.remote.s3.secretAccessKey,\n }\n : undefined,\n });\n\n // Initialize stats\n this.stats = {\n localUsageMB: 0,\n remoteUsageMB: 0,\n tierDistribution: {\n [StorageTier.YOUNG]: 0,\n [StorageTier.MATURE]: 0,\n [StorageTier.OLD]: 0,\n [StorageTier.REMOTE]: 0,\n },\n compressionRatio: 1.0,\n migrationsPending: 0,\n lastMigration: null,\n };\n }\n\n private initializeLocalStorage(): void {\n // Enable WAL mode for better concurrency\n this.db.pragma('journal_mode = WAL');\n this.db.pragma('synchronous = NORMAL');\n\n // Create storage tables with tier information\n this.db.exec(`\n CREATE TABLE IF NOT EXISTS storage_items (\n id TEXT PRIMARY KEY,\n frame_id TEXT NOT NULL,\n tier TEXT NOT NULL,\n data BLOB NOT NULL,\n metadata TEXT,\n size_bytes INTEGER,\n importance_score REAL DEFAULT 0.5,\n access_count INTEGER DEFAULT 0,\n last_accessed DATETIME DEFAULT CURRENT_TIMESTAMP,\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP,\n compressed BOOLEAN DEFAULT FALSE,\n compression_type TEXT\n );\n \n CREATE INDEX IF NOT EXISTS idx_tier_age ON storage_items (tier, created_at);\n CREATE INDEX IF NOT EXISTS idx_frame ON storage_items (frame_id);\n CREATE INDEX IF NOT EXISTS idx_importance ON storage_items (importance_score DESC);\n \n CREATE TABLE IF NOT EXISTS migration_queue (\n id TEXT PRIMARY KEY,\n item_id TEXT NOT NULL,\n source_tier TEXT NOT NULL,\n target_tier TEXT NOT NULL,\n priority INTEGER DEFAULT 5,\n attempts INTEGER DEFAULT 0,\n status TEXT DEFAULT 'pending',\n created_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n \n CREATE INDEX IF NOT EXISTS idx_status_priority ON migration_queue (status, priority DESC);\n \n CREATE TABLE IF NOT EXISTS storage_metrics (\n id INTEGER PRIMARY KEY AUTOINCREMENT,\n tier TEXT NOT NULL,\n item_count INTEGER,\n total_size_mb REAL,\n avg_compression_ratio REAL,\n measured_at DATETIME DEFAULT CURRENT_TIMESTAMP\n );\n `);\n }\n\n async initialize(): Promise<void> {\n try {\n // Initialize Redis for hot cache\n if (this.config.remote.redis?.url) {\n this.redisClient = createRedisClient({\n url: this.config.remote.redis.url,\n });\n await this.redisClient.connect();\n this.logger.info('Redis connected for hot cache');\n }\n\n // Initialize TimeSeries DB\n if (this.config.remote.timeseries?.connectionString) {\n this.timeseriesPool = new Pool({\n connectionString: this.config.remote.timeseries.connectionString,\n max: 5,\n });\n await this.initializeTimeseriesSchema();\n this.logger.info('TimeSeries DB connected');\n }\n\n // Start migration worker\n this.startMigrationWorker();\n\n // Load offline queue if exists\n await this.loadOfflineQueue();\n\n // Calculate initial stats\n await this.updateStats();\n\n this.logger.info('Two-tier storage system initialized');\n } catch (error) {\n this.logger.error('Failed to initialize storage', { error });\n throw error;\n }\n }\n\n private async initializeTimeseriesSchema(): Promise<void> {\n if (!this.timeseriesPool) return;\n\n const client = await this.timeseriesPool.connect();\n try {\n await client.query(`\n CREATE TABLE IF NOT EXISTS frame_timeseries (\n time TIMESTAMPTZ NOT NULL,\n frame_id TEXT NOT NULL,\n project_id TEXT NOT NULL,\n data JSONB,\n metrics JSONB,\n importance_score REAL\n );\n \n SELECT create_hypertable('frame_timeseries', 'time', \n chunk_time_interval => INTERVAL '1 month',\n if_not_exists => TRUE\n );\n \n CREATE INDEX IF NOT EXISTS idx_frame_time \n ON frame_timeseries (frame_id, time DESC);\n `);\n } finally {\n client.release();\n }\n }\n\n /**\n * Store a frame with automatic tier selection\n */\n async storeFrame(\n frame: Frame,\n events: Event[],\n anchors: Anchor[]\n ): Promise<string> {\n const storageId = uuidv4();\n const data = { frame, events, anchors };\n const tier = this.selectTier(frame);\n\n // Calculate importance score\n const importanceScore = this.calculateImportance(frame, events, anchors);\n\n // Compress based on tier\n const compressed = await this.compressData(data, tier);\n\n // Store locally first\n const stmt = this.db.prepare(`\n INSERT INTO storage_items (\n id, frame_id, tier, data, metadata, size_bytes, \n importance_score, compressed, compression_type\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)\n `);\n\n stmt.run(\n storageId,\n frame.frame_id,\n tier,\n compressed.data,\n JSON.stringify({\n originalSize: compressed.originalSize,\n compressedSize: compressed.compressedSize,\n }),\n compressed.compressedSize,\n importanceScore,\n compressed.compressed ? 1 : 0,\n compressed.compressionType\n );\n\n // Also store in Redis if young and available\n if (tier === StorageTier.YOUNG && this.redisClient) {\n await this.redisClient.setex(\n `frame:${frame.frame_id}`,\n this.config.remote.redis?.ttlSeconds || 300,\n JSON.stringify(data)\n );\n }\n\n // Queue for remote upload if important\n if (importanceScore > 0.7) {\n await this.queueRemoteUpload(storageId, data, 'high');\n }\n\n return storageId;\n }\n\n /**\n * Retrieve a frame from any tier\n */\n async retrieveFrame(frameId: string): Promise<any> {\n // Check Redis first\n if (this.redisClient) {\n try {\n const cached = await this.redisClient.get(`frame:${frameId}`);\n if (cached) {\n this.updateAccessCount(frameId);\n return JSON.parse(cached);\n }\n } catch (error) {\n this.logger.warn('Redis retrieval failed', { frameId, error });\n }\n }\n\n // Check local storage\n const local = this.db\n .prepare(\n `\n SELECT data, compressed, compression_type \n FROM storage_items \n WHERE frame_id = ?\n ORDER BY created_at DESC\n LIMIT 1\n `\n )\n .get(frameId);\n\n if (local) {\n this.updateAccessCount(frameId);\n const data = local.compressed\n ? await this.decompressData(local.data, local.compression_type)\n : JSON.parse(local.data);\n\n // Promote to Redis if frequently accessed\n if (this.redisClient) {\n await this.redisClient.setex(\n `frame:${frameId}`,\n 300, // 5 minute cache\n JSON.stringify(data)\n );\n }\n\n return data;\n }\n\n // Check remote storage\n return this.retrieveFromRemote(frameId);\n }\n\n /**\n * Select appropriate tier based on frame age and characteristics\n */\n private selectTier(frame: Frame): StorageTier {\n if (!frame.created_at || isNaN(frame.created_at)) {\n this.logger.warn('Invalid frame timestamp, defaulting to YOUNG tier');\n return StorageTier.YOUNG;\n }\n\n const ageHours = (Date.now() - frame.created_at) / (1000 * 60 * 60);\n\n if (ageHours < 24) return StorageTier.YOUNG;\n if (ageHours < 168) return StorageTier.MATURE; // 7 days\n if (ageHours < 720) return StorageTier.OLD; // 30 days\n return StorageTier.REMOTE;\n }\n\n /**\n * Calculate importance score for migration decisions\n */\n private calculateImportance(\n frame: Frame,\n events: Event[],\n anchors: Anchor[]\n ): number {\n let score = 0.5; // Base score\n\n // Increase for frames with decisions\n const decisions = anchors.filter((a) => a.type === 'DECISION');\n score += decisions.length * 0.1;\n\n // Increase for frames with many events\n score += Math.min(events.length * 0.01, 0.2);\n\n // Increase for recent frames\n const ageHours = (Date.now() - frame.created_at) / (1000 * 60 * 60);\n if (ageHours < 24) score += 0.2;\n else if (ageHours < 168) score += 0.1;\n\n // Increase for frames with errors\n const errors = events.filter((e) => e.event_type === 'error');\n if (errors.length > 0) score += 0.2;\n\n return Math.min(score, 1.0);\n }\n\n /**\n * Compress data based on tier configuration\n */\n private async compressData(data: any, tier: StorageTier): Promise<any> {\n const json = JSON.stringify(data);\n const originalSize = Buffer.byteLength(json);\n\n const tierConfig = this.config.local.tiers.find((t) => t.name === tier);\n if (!tierConfig || tierConfig.compressionType === 'none') {\n return {\n data: Buffer.from(json),\n originalSize,\n compressedSize: originalSize,\n compressed: false,\n compressionType: 'none',\n };\n }\n\n let compressed: Buffer;\n switch (tierConfig.compressionType) {\n case 'lz4':\n compressed = lz4.encode(Buffer.from(json));\n break;\n case 'zstd':\n // For ZSTD, we'll use gzip as a placeholder (install zstd-codec for real impl)\n compressed = await gzip(json, { level: 9 });\n break;\n case 'gzip':\n compressed = await gzip(json);\n break;\n default:\n compressed = Buffer.from(json);\n }\n\n return {\n data: compressed,\n originalSize,\n compressedSize: compressed.length,\n compressed: true,\n compressionType: tierConfig.compressionType,\n };\n }\n\n /**\n * Decompress data\n */\n private async decompressData(\n data: Buffer,\n compressionType: string\n ): Promise<any> {\n let decompressed: Buffer;\n\n switch (compressionType) {\n case 'lz4':\n decompressed = lz4.decode(data);\n break;\n case 'zstd':\n case 'gzip':\n decompressed = await gunzip(data);\n break;\n default:\n decompressed = data;\n }\n\n return JSON.parse(decompressed.toString());\n }\n\n /**\n * Start background migration worker\n */\n private startMigrationWorker(): void {\n this.migrationTimer = setInterval(async () => {\n try {\n await this.processMigrations();\n await this.checkMigrationTriggers();\n await this.processOfflineQueue();\n } catch (error) {\n this.logger.error('Migration worker error', { error });\n }\n }, this.config.migration.intervalMs);\n\n this.logger.info('Migration worker started');\n }\n\n /**\n * Process pending migrations\n */\n private async processMigrations(): Promise<void> {\n const pending = this.db\n .prepare(\n `\n SELECT * FROM migration_queue \n WHERE status = 'pending' \n ORDER BY priority DESC, created_at ASC \n LIMIT ?\n `\n )\n .all(this.config.migration.batchSize);\n\n for (const migration of pending) {\n try {\n await this.executeMigration(migration);\n\n // Mark as completed\n this.db\n .prepare(\n `\n UPDATE migration_queue \n SET status = 'completed' \n WHERE id = ?\n `\n )\n .run(migration.id);\n } catch (error) {\n this.logger.error('Migration failed', { migration, error });\n\n // Update attempts\n this.db\n .prepare(\n `\n UPDATE migration_queue \n SET attempts = attempts + 1,\n status = CASE WHEN attempts >= 3 THEN 'failed' ELSE 'pending' END\n WHERE id = ?\n `\n )\n .run(migration.id);\n }\n }\n\n this.stats.lastMigration = new Date();\n }\n\n /**\n * Check and trigger migrations based on rules\n */\n private async checkMigrationTriggers(): Promise<void> {\n // Check age-based triggers\n for (const trigger of this.config.migration.triggers) {\n if (trigger.type === 'age') {\n const items = this.db\n .prepare(\n `\n SELECT id, frame_id, tier \n FROM storage_items \n WHERE julianday('now') - julianday(created_at) > ?\n AND tier != ?\n LIMIT 100\n `\n )\n .all(trigger.threshold / 24, StorageTier.REMOTE);\n\n for (const item of items) {\n this.queueMigration(item.id, item.tier, StorageTier.REMOTE);\n }\n }\n\n // Check size-based triggers\n if (trigger.type === 'size') {\n const stats = this.db\n .prepare(\n `\n SELECT SUM(size_bytes) as total_size \n FROM storage_items \n WHERE tier IN ('young', 'mature')\n `\n )\n .get();\n\n if (stats.total_size > trigger.threshold * 1024 * 1024) {\n // Migrate oldest items\n const items = this.db\n .prepare(\n `\n SELECT id, tier FROM storage_items \n WHERE tier IN ('young', 'mature')\n ORDER BY created_at ASC \n LIMIT 50\n `\n )\n .all();\n\n for (const item of items) {\n const targetTier =\n item.tier === StorageTier.YOUNG\n ? StorageTier.MATURE\n : StorageTier.OLD;\n this.queueMigration(item.id, item.tier, targetTier);\n }\n }\n }\n }\n }\n\n /**\n * Queue a migration\n */\n private queueMigration(\n itemId: string,\n sourceTier: string,\n targetTier: string,\n priority: number = 5\n ): void {\n const id = uuidv4();\n this.db\n .prepare(\n `\n INSERT INTO migration_queue (id, item_id, source_tier, target_tier, priority)\n VALUES (?, ?, ?, ?, ?)\n `\n )\n .run(id, itemId, sourceTier, targetTier, priority);\n\n this.stats.migrationsPending++;\n }\n\n /**\n * Execute a single migration\n */\n private async executeMigration(migration: any): Promise<void> {\n // Get item data\n const item = this.db\n .prepare(\n `\n SELECT * FROM storage_items WHERE id = ?\n `\n )\n .get(migration.item_id);\n\n if (!item) {\n throw new Error(`Item not found: ${migration.item_id}`);\n }\n\n // Decompress if needed\n const data = item.compressed\n ? await this.decompressData(item.data, item.compression_type)\n : JSON.parse(item.data);\n\n // Upload to remote\n if (migration.target_tier === StorageTier.REMOTE) {\n await this.uploadToS3(item.frame_id, data);\n\n // Delete from local after successful upload\n this.db\n .prepare(\n `\n DELETE FROM storage_items WHERE id = ?\n `\n )\n .run(migration.item_id);\n } else {\n // Re-compress for new tier\n const compressed = await this.compressData(data, migration.target_tier);\n\n // Update local storage\n this.db\n .prepare(\n `\n UPDATE storage_items \n SET tier = ?, data = ?, size_bytes = ?, \n compressed = ?, compression_type = ?\n WHERE id = ?\n `\n )\n .run(\n migration.target_tier,\n compressed.data,\n compressed.compressedSize,\n compressed.compressed ? 1 : 0,\n compressed.compressionType,\n migration.item_id\n );\n }\n\n this.logger.info('Migration completed', {\n itemId: migration.item_id,\n from: migration.source_tier,\n to: migration.target_tier,\n });\n }\n\n /**\n * Upload data to S3\n */\n private async uploadToS3(frameId: string, data: any): Promise<void> {\n const date = new Date();\n const partition = `${date.getFullYear()}/${String(date.getMonth() + 1).padStart(2, '0')}`;\n const key = `frames/${partition}/${frameId}.json.gz`;\n\n const compressed = await gzip(JSON.stringify(data));\n\n const command = new PutObjectCommand({\n Bucket: this.config.remote.s3.bucket,\n Key: key,\n Body: compressed,\n ContentType: 'application/json',\n ContentEncoding: 'gzip',\n Metadata: {\n frameId,\n uploadedAt: date.toISOString(),\n },\n });\n\n await this.s3Client.send(command);\n }\n\n /**\n * Retrieve from remote storage\n */\n private async retrieveFromRemote(frameId: string): Promise<any> {\n // Try TimeSeries DB first\n if (this.timeseriesPool) {\n const client = await this.timeseriesPool.connect();\n try {\n const result = await client.query(\n `\n SELECT data FROM frame_timeseries \n WHERE frame_id = $1 \n ORDER BY time DESC \n LIMIT 1\n `,\n [frameId]\n );\n\n if (result.rows.length > 0) {\n return result.rows[0].data;\n }\n } finally {\n client.release();\n }\n }\n\n // Try S3\n // Would need to search multiple partitions - simplified for now\n const date = new Date();\n for (let i = 0; i < 3; i++) {\n const checkDate = new Date(date);\n checkDate.setMonth(checkDate.getMonth() - i);\n const partition = `${checkDate.getFullYear()}/${String(checkDate.getMonth() + 1).padStart(2, '0')}`;\n const key = `frames/${partition}/${frameId}.json.gz`;\n\n try {\n const command = new GetObjectCommand({\n Bucket: this.config.remote.s3.bucket,\n Key: key,\n });\n\n const response = await this.s3Client.send(command);\n if (!response.Body) continue;\n const body = await response.Body.transformToByteArray();\n const decompressed = await gunzip(Buffer.from(body));\n return JSON.parse(decompressed.toString());\n } catch {\n // Continue searching - frame not in this partition\n }\n }\n\n return null;\n }\n\n /**\n * Queue for offline upload\n */\n private async queueRemoteUpload(\n id: string,\n data: any,\n priority: 'high' | 'normal'\n ): Promise<void> {\n this.offlineQueue.push({ \n id, \n data, \n priority, \n timestamp: Date.now() \n });\n\n // Persist queue if configured\n if (this.config.migration.offlineQueuePath) {\n await this.saveOfflineQueue();\n }\n }\n\n /**\n * Process offline upload queue\n */\n private async processOfflineQueue(): Promise<void> {\n if (this.offlineQueue.length === 0) return;\n\n // Check if online\n const isOnline = await this.checkConnectivity();\n if (!isOnline) return;\n\n // Process high priority first\n this.offlineQueue.sort((a, b) => {\n if (a.priority !== b.priority) {\n return a.priority === 'high' ? -1 : 1;\n }\n return a.timestamp - b.timestamp;\n });\n\n const batch = this.offlineQueue.splice(0, 10);\n for (const item of batch) {\n try {\n await this.uploadToS3(item.id, item.data);\n } catch {\n // Re-queue on failure\n this.offlineQueue.push(item);\n }\n }\n\n await this.saveOfflineQueue();\n }\n\n /**\n * Check connectivity to remote services\n */\n private async checkConnectivity(): Promise<boolean> {\n try {\n // Simple S3 head bucket check\n const response = await fetch('https://s3.amazonaws.com');\n return response.ok;\n } catch {\n return false;\n }\n }\n\n /**\n * Update access count for cache promotion\n */\n private updateAccessCount(frameId: string): void {\n this.db\n .prepare(\n `\n UPDATE storage_items \n SET access_count = access_count + 1,\n last_accessed = CURRENT_TIMESTAMP\n WHERE frame_id = ?\n `\n )\n .run(frameId);\n }\n\n /**\n * Save offline queue to disk\n */\n private async saveOfflineQueue(): Promise<void> {\n if (!this.config.migration.offlineQueuePath) return;\n\n const fs = await import('fs/promises');\n await fs.writeFile(\n this.config.migration.offlineQueuePath,\n JSON.stringify(this.offlineQueue),\n 'utf-8'\n );\n }\n\n /**\n * Load offline queue from disk\n */\n private async loadOfflineQueue(): Promise<void> {\n if (!this.config.migration.offlineQueuePath) return;\n\n const fs = await import('fs/promises');\n try {\n const data = await fs.readFile(\n this.config.migration.offlineQueuePath,\n 'utf-8'\n );\n this.offlineQueue = JSON.parse(data);\n this.logger.info(\n `Loaded ${this.offlineQueue.length} items from offline queue`\n );\n } catch {\n // Queue file doesn't exist yet - this is normal on first run\n }\n }\n\n /**\n * Update storage statistics\n */\n private async updateStats(): Promise<void> {\n // Calculate local usage\n const localStats = this.db\n .prepare(\n `\n SELECT \n tier,\n COUNT(*) as count,\n SUM(size_bytes) / 1048576.0 as size_mb\n FROM storage_items\n GROUP BY tier\n `\n )\n .all();\n\n this.stats.localUsageMB = 0;\n for (const stat of localStats) {\n this.stats.tierDistribution[stat.tier as StorageTier] = stat.count;\n this.stats.localUsageMB += stat.size_mb;\n }\n\n // Calculate compression ratio\n const compressionStats = this.db\n .prepare(\n `\n SELECT \n AVG(CAST(json_extract(metadata, '$.originalSize') AS REAL) / \n CAST(json_extract(metadata, '$.compressedSize') AS REAL)) as ratio\n FROM storage_items\n WHERE compressed = 1\n `\n )\n .get();\n\n this.stats.compressionRatio = compressionStats?.ratio || 1.0;\n\n // Count pending migrations\n const pending = this.db\n .prepare(\n `\n SELECT COUNT(*) as count \n FROM migration_queue \n WHERE status = 'pending'\n `\n )\n .get();\n\n this.stats.migrationsPending = pending.count;\n\n // Save metrics\n this.db\n .prepare(\n `\n INSERT INTO storage_metrics (tier, item_count, total_size_mb, avg_compression_ratio)\n VALUES ('all', ?, ?, ?)\n `\n )\n .run(\n Object.values(this.stats.tierDistribution).reduce((a, b) => a + b, 0),\n this.stats.localUsageMB,\n this.stats.compressionRatio\n );\n }\n\n /**\n * Get current storage statistics\n */\n async getStats(): Promise<StorageStats> {\n await this.updateStats();\n return { ...this.stats };\n }\n\n /**\n * Cleanup and shutdown\n */\n async shutdown(): Promise<void> {\n if (this.migrationTimer) {\n clearInterval(this.migrationTimer);\n }\n\n await this.saveOfflineQueue();\n\n if (this.redisClient) {\n await this.redisClient.quit();\n }\n\n if (this.timeseriesPool) {\n await this.timeseriesPool.end();\n }\n\n this.db.close();\n\n this.logger.info('Two-tier storage system shut down');\n }\n}\n\n// Export default configuration\nexport const defaultTwoTierConfig: TwoTierConfig = {\n local: {\n dbPath: '~/.stackmemory/two-tier.db',\n maxSizeGB: 2,\n tiers: [\n {\n name: StorageTier.YOUNG,\n maxAgeHours: 24,\n compressionType: 'none',\n retentionPolicy: 'complete',\n maxSizeMB: 500,\n },\n {\n name: StorageTier.MATURE,\n maxAgeHours: 168,\n compressionType: 'lz4',\n retentionPolicy: 'selective',\n maxSizeMB: 1000,\n },\n {\n name: StorageTier.OLD,\n maxAgeHours: 720,\n compressionType: 'zstd',\n retentionPolicy: 'critical',\n maxSizeMB: 500,\n },\n ],\n },\n remote: {\n s3: {\n bucket: process.env.S3_BUCKET || 'stackmemory-storage',\n region: process.env.AWS_REGION || 'us-east-1',\n },\n },\n migration: {\n triggers: [\n { type: 'age', threshold: 720, action: 'migrate' }, // 30 days\n { type: 'size', threshold: 1500, action: 'migrate' }, // 1.5GB\n { type: 'importance', threshold: 0.3, action: 'delete' }, // Low importance\n ],\n batchSize: 50,\n intervalMs: 60000, // 1 minute\n offlineQueuePath: '~/.stackmemory/offline-queue.json',\n },\n};\n"],
|
|
5
|
+
"mappings": "AAeA,OAAO,cAAc;AACrB;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,OACK;AAEP,SAAS,gBAAgB,yBAAyB;AAClD,SAAS,YAAY;AACrB,YAAY,UAAU;AACtB,SAAS,iBAAiB;AAC1B,SAAS,MAAM,cAAc;AAC7B,SAAS,cAAc;AAKvB,MAAM,MAAM;AAAA,EACV,QAAQ,CAAC,SAAiB;AAAA;AAAA,EAC1B,QAAQ,CAAC,SAAiB;AAAA;AAC5B;AAEA,MAAM,OAAO,UAAU,KAAK,IAAI;AAChC,MAAM,SAAS,UAAU,KAAK,MAAM;AAE7B,IAAK,cAAL,kBAAKA,iBAAL;AACL,EAAAA,aAAA,WAAQ;AACR,EAAAA,aAAA,YAAS;AACT,EAAAA,aAAA,SAAM;AACN,EAAAA,aAAA,YAAS;AAJC,SAAAA;AAAA,GAAA;AAkEL,MAAM,qBAAqB;AAAA,EACxB;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA,eAAmC,CAAC;AAAA,EACpC;AAAA,EAER,YAAY,QAAuB;AACjC,SAAK,SAAS;AACd,SAAK,SAAS,IAAI,OAAO,gBAAgB;AAGzC,SAAK,KAAK,IAAI,SAAS,OAAO,MAAM,MAAM;AAC1C,SAAK,uBAAuB;AAG5B,SAAK,WAAW,IAAI,SAAS;AAAA,MAC3B,QAAQ,OAAO,OAAO,GAAG;AAAA,MACzB,aACE,OAAO,OAAO,GAAG,eAAe,OAAO,OAAO,GAAG,kBAC7C;AAAA,QACE,aAAa,OAAO,OAAO,GAAG;AAAA,QAC9B,iBAAiB,OAAO,OAAO,GAAG;AAAA,MACpC,IACA;AAAA,IACR,CAAC;AAGD,SAAK,QAAQ;AAAA,MACX,cAAc;AAAA,MACd,eAAe;AAAA,MACf,kBAAkB;AAAA,QAChB,CAAC,mBAAiB,GAAG;AAAA,QACrB,CAAC,qBAAkB,GAAG;AAAA,QACtB,CAAC,eAAe,GAAG;AAAA,QACnB,CAAC,qBAAkB,GAAG;AAAA,MACxB;AAAA,MACA,kBAAkB;AAAA,MAClB,mBAAmB;AAAA,MACnB,eAAe;AAAA,IACjB;AAAA,EACF;AAAA,EAEQ,yBAA+B;AAErC,SAAK,GAAG,OAAO,oBAAoB;AACnC,SAAK,GAAG,OAAO,sBAAsB;AAGrC,SAAK,GAAG,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,KAyCZ;AAAA,EACH;AAAA,EAEA,MAAM,aAA4B;AAChC,QAAI;AAEF,UAAI,KAAK,OAAO,OAAO,OAAO,KAAK;AACjC,aAAK,cAAc,kBAAkB;AAAA,UACnC,KAAK,KAAK,OAAO,OAAO,MAAM;AAAA,QAChC,CAAC;AACD,cAAM,KAAK,YAAY,QAAQ;AAC/B,aAAK,OAAO,KAAK,+BAA+B;AAAA,MAClD;AAGA,UAAI,KAAK,OAAO,OAAO,YAAY,kBAAkB;AACnD,aAAK,iBAAiB,IAAI,KAAK;AAAA,UAC7B,kBAAkB,KAAK,OAAO,OAAO,WAAW;AAAA,UAChD,KAAK;AAAA,QACP,CAAC;AACD,cAAM,KAAK,2BAA2B;AACtC,aAAK,OAAO,KAAK,yBAAyB;AAAA,MAC5C;AAGA,WAAK,qBAAqB;AAG1B,YAAM,KAAK,iBAAiB;AAG5B,YAAM,KAAK,YAAY;AAEvB,WAAK,OAAO,KAAK,qCAAqC;AAAA,IACxD,SAAS,OAAO;AACd,WAAK,OAAO,MAAM,gCAAgC,EAAE,MAAM,CAAC;AAC3D,YAAM;AAAA,IACR;AAAA,EACF;AAAA,EAEA,MAAc,6BAA4C;AACxD,QAAI,CAAC,KAAK,eAAgB;AAE1B,UAAM,SAAS,MAAM,KAAK,eAAe,QAAQ;AACjD,QAAI;AACF,YAAM,OAAO,MAAM;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,OAiBlB;AAAA,IACH,UAAE;AACA,aAAO,QAAQ;AAAA,IACjB;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,WACJ,OACA,QACA,SACiB;AACjB,UAAM,YAAY,OAAO;AACzB,UAAM,OAAO,EAAE,OAAO,QAAQ,QAAQ;AACtC,UAAM,OAAO,KAAK,WAAW,KAAK;AAGlC,UAAM,kBAAkB,KAAK,oBAAoB,OAAO,QAAQ,OAAO;AAGvE,UAAM,aAAa,MAAM,KAAK,aAAa,MAAM,IAAI;AAGrD,UAAM,OAAO,KAAK,GAAG,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,KAK5B;AAED,SAAK;AAAA,MACH;AAAA,MACA,MAAM;AAAA,MACN;AAAA,MACA,WAAW;AAAA,MACX,KAAK,UAAU;AAAA,QACb,cAAc,WAAW;AAAA,QACzB,gBAAgB,WAAW;AAAA,MAC7B,CAAC;AAAA,MACD,WAAW;AAAA,MACX;AAAA,MACA,WAAW,aAAa,IAAI;AAAA,MAC5B,WAAW;AAAA,IACb;AAGA,QAAI,SAAS,uBAAqB,KAAK,aAAa;AAClD,YAAM,KAAK,YAAY;AAAA,QACrB,SAAS,MAAM,QAAQ;AAAA,QACvB,KAAK,OAAO,OAAO,OAAO,cAAc;AAAA,QACxC,KAAK,UAAU,IAAI;AAAA,MACrB;AAAA,IACF;AAGA,QAAI,kBAAkB,KAAK;AACzB,YAAM,KAAK,kBAAkB,WAAW,MAAM,MAAM;AAAA,IACtD;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,cAAc,SAA+B;AAEjD,QAAI,KAAK,aAAa;AACpB,UAAI;AACF,cAAM,SAAS,MAAM,KAAK,YAAY,IAAI,SAAS,OAAO,EAAE;AAC5D,YAAI,QAAQ;AACV,eAAK,kBAAkB,OAAO;AAC9B,iBAAO,KAAK,MAAM,MAAM;AAAA,QAC1B;AAAA,MACF,SAAS,OAAO;AACd,aAAK,OAAO,KAAK,0BAA0B,EAAE,SAAS,MAAM,CAAC;AAAA,MAC/D;AAAA,IACF;AAGA,UAAM,QAAQ,KAAK,GAChB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAOF,EACC,IAAI,OAAO;AAEd,QAAI,OAAO;AACT,WAAK,kBAAkB,OAAO;AAC9B,YAAM,OAAO,MAAM,aACf,MAAM,KAAK,eAAe,MAAM,MAAM,MAAM,gBAAgB,IAC5D,KAAK,MAAM,MAAM,IAAI;AAGzB,UAAI,KAAK,aAAa;AACpB,cAAM,KAAK,YAAY;AAAA,UACrB,SAAS,OAAO;AAAA,UAChB;AAAA;AAAA,UACA,KAAK,UAAU,IAAI;AAAA,QACrB;AAAA,MACF;AAEA,aAAO;AAAA,IACT;AAGA,WAAO,KAAK,mBAAmB,OAAO;AAAA,EACxC;AAAA;AAAA;AAAA;AAAA,EAKQ,WAAW,OAA2B;AAC5C,QAAI,CAAC,MAAM,cAAc,MAAM,MAAM,UAAU,GAAG;AAChD,WAAK,OAAO,KAAK,mDAAmD;AACpE,aAAO;AAAA,IACT;AAEA,UAAM,YAAY,KAAK,IAAI,IAAI,MAAM,eAAe,MAAO,KAAK;AAEhE,QAAI,WAAW,GAAI,QAAO;AAC1B,QAAI,WAAW,IAAK,QAAO;AAC3B,QAAI,WAAW,IAAK,QAAO;AAC3B,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKQ,oBACN,OACA,QACA,SACQ;AACR,QAAI,QAAQ;AAGZ,UAAM,YAAY,QAAQ,OAAO,CAAC,MAAM,EAAE,SAAS,UAAU;AAC7D,aAAS,UAAU,SAAS;AAG5B,aAAS,KAAK,IAAI,OAAO,SAAS,MAAM,GAAG;AAG3C,UAAM,YAAY,KAAK,IAAI,IAAI,MAAM,eAAe,MAAO,KAAK;AAChE,QAAI,WAAW,GAAI,UAAS;AAAA,aACnB,WAAW,IAAK,UAAS;AAGlC,UAAM,SAAS,OAAO,OAAO,CAAC,MAAM,EAAE,eAAe,OAAO;AAC5D,QAAI,OAAO,SAAS,EAAG,UAAS;AAEhC,WAAO,KAAK,IAAI,OAAO,CAAG;AAAA,EAC5B;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,aAAa,MAAW,MAAiC;AACrE,UAAM,OAAO,KAAK,UAAU,IAAI;AAChC,UAAM,eAAe,OAAO,WAAW,IAAI;AAE3C,UAAM,aAAa,KAAK,OAAO,MAAM,MAAM,KAAK,CAAC,MAAM,EAAE,SAAS,IAAI;AACtE,QAAI,CAAC,cAAc,WAAW,oBAAoB,QAAQ;AACxD,aAAO;AAAA,QACL,MAAM,OAAO,KAAK,IAAI;AAAA,QACtB;AAAA,QACA,gBAAgB;AAAA,QAChB,YAAY;AAAA,QACZ,iBAAiB;AAAA,MACnB;AAAA,IACF;AAEA,QAAI;AACJ,YAAQ,WAAW,iBAAiB;AAAA,MAClC,KAAK;AACH,qBAAa,IAAI,OAAO,OAAO,KAAK,IAAI,CAAC;AACzC;AAAA,MACF,KAAK;AAEH,qBAAa,MAAM,KAAK,MAAM,EAAE,OAAO,EAAE,CAAC;AAC1C;AAAA,MACF,KAAK;AACH,qBAAa,MAAM,KAAK,IAAI;AAC5B;AAAA,MACF;AACE,qBAAa,OAAO,KAAK,IAAI;AAAA,IACjC;AAEA,WAAO;AAAA,MACL,MAAM;AAAA,MACN;AAAA,MACA,gBAAgB,WAAW;AAAA,MAC3B,YAAY;AAAA,MACZ,iBAAiB,WAAW;AAAA,IAC9B;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,eACZ,MACA,iBACc;AACd,QAAI;AAEJ,YAAQ,iBAAiB;AAAA,MACvB,KAAK;AACH,uBAAe,IAAI,OAAO,IAAI;AAC9B;AAAA,MACF,KAAK;AAAA,MACL,KAAK;AACH,uBAAe,MAAM,OAAO,IAAI;AAChC;AAAA,MACF;AACE,uBAAe;AAAA,IACnB;AAEA,WAAO,KAAK,MAAM,aAAa,SAAS,CAAC;AAAA,EAC3C;AAAA;AAAA;AAAA;AAAA,EAKQ,uBAA6B;AACnC,SAAK,iBAAiB,YAAY,YAAY;AAC5C,UAAI;AACF,cAAM,KAAK,kBAAkB;AAC7B,cAAM,KAAK,uBAAuB;AAClC,cAAM,KAAK,oBAAoB;AAAA,MACjC,SAAS,OAAO;AACd,aAAK,OAAO,MAAM,0BAA0B,EAAE,MAAM,CAAC;AAAA,MACvD;AAAA,IACF,GAAG,KAAK,OAAO,UAAU,UAAU;AAEnC,SAAK,OAAO,KAAK,0BAA0B;AAAA,EAC7C;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,oBAAmC;AAC/C,UAAM,UAAU,KAAK,GAClB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAMF,EACC,IAAI,KAAK,OAAO,UAAU,SAAS;AAEtC,eAAW,aAAa,SAAS;AAC/B,UAAI;AACF,cAAM,KAAK,iBAAiB,SAAS;AAGrC,aAAK,GACF;AAAA,UACC;AAAA;AAAA;AAAA;AAAA;AAAA,QAKF,EACC,IAAI,UAAU,EAAE;AAAA,MACrB,SAAS,OAAO;AACd,aAAK,OAAO,MAAM,oBAAoB,EAAE,WAAW,MAAM,CAAC;AAG1D,aAAK,GACF;AAAA,UACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAMF,EACC,IAAI,UAAU,EAAE;AAAA,MACrB;AAAA,IACF;AAEA,SAAK,MAAM,gBAAgB,oBAAI,KAAK;AAAA,EACtC;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,yBAAwC;AAEpD,eAAW,WAAW,KAAK,OAAO,UAAU,UAAU;AACpD,UAAI,QAAQ,SAAS,OAAO;AAC1B,cAAM,QAAQ,KAAK,GAChB;AAAA,UACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,QAOF,EACC,IAAI,QAAQ,YAAY,IAAI,qBAAkB;AAEjD,mBAAW,QAAQ,OAAO;AACxB,eAAK,eAAe,KAAK,IAAI,KAAK,MAAM,qBAAkB;AAAA,QAC5D;AAAA,MACF;AAGA,UAAI,QAAQ,SAAS,QAAQ;AAC3B,cAAM,QAAQ,KAAK,GAChB;AAAA,UACC;AAAA;AAAA;AAAA;AAAA;AAAA,QAKF,EACC,IAAI;AAEP,YAAI,MAAM,aAAa,QAAQ,YAAY,OAAO,MAAM;AAEtD,gBAAM,QAAQ,KAAK,GAChB;AAAA,YACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,UAMF,EACC,IAAI;AAEP,qBAAW,QAAQ,OAAO;AACxB,kBAAM,aACJ,KAAK,SAAS,sBACV,wBACA;AACN,iBAAK,eAAe,KAAK,IAAI,KAAK,MAAM,UAAU;AAAA,UACpD;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,eACN,QACA,YACA,YACA,WAAmB,GACb;AACN,UAAM,KAAK,OAAO;AAClB,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC,IAAI,IAAI,QAAQ,YAAY,YAAY,QAAQ;AAEnD,SAAK,MAAM;AAAA,EACb;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,iBAAiB,WAA+B;AAE5D,UAAM,OAAO,KAAK,GACf;AAAA,MACC;AAAA;AAAA;AAAA,IAGF,EACC,IAAI,UAAU,OAAO;AAExB,QAAI,CAAC,MAAM;AACT,YAAM,IAAI,MAAM,mBAAmB,UAAU,OAAO,EAAE;AAAA,IACxD;AAGA,UAAM,OAAO,KAAK,aACd,MAAM,KAAK,eAAe,KAAK,MAAM,KAAK,gBAAgB,IAC1D,KAAK,MAAM,KAAK,IAAI;AAGxB,QAAI,UAAU,gBAAgB,uBAAoB;AAChD,YAAM,KAAK,WAAW,KAAK,UAAU,IAAI;AAGzC,WAAK,GACF;AAAA,QACC;AAAA;AAAA;AAAA,MAGF,EACC,IAAI,UAAU,OAAO;AAAA,IAC1B,OAAO;AAEL,YAAM,aAAa,MAAM,KAAK,aAAa,MAAM,UAAU,WAAW;AAGtE,WAAK,GACF;AAAA,QACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,MAMF,EACC;AAAA,QACC,UAAU;AAAA,QACV,WAAW;AAAA,QACX,WAAW;AAAA,QACX,WAAW,aAAa,IAAI;AAAA,QAC5B,WAAW;AAAA,QACX,UAAU;AAAA,MACZ;AAAA,IACJ;AAEA,SAAK,OAAO,KAAK,uBAAuB;AAAA,MACtC,QAAQ,UAAU;AAAA,MAClB,MAAM,UAAU;AAAA,MAChB,IAAI,UAAU;AAAA,IAChB,CAAC;AAAA,EACH;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,WAAW,SAAiB,MAA0B;AAClE,UAAM,OAAO,oBAAI,KAAK;AACtB,UAAM,YAAY,GAAG,KAAK,YAAY,CAAC,IAAI,OAAO,KAAK,SAAS,IAAI,CAAC,EAAE,SAAS,GAAG,GAAG,CAAC;AACvF,UAAM,MAAM,UAAU,SAAS,IAAI,OAAO;AAE1C,UAAM,aAAa,MAAM,KAAK,KAAK,UAAU,IAAI,CAAC;AAElD,UAAM,UAAU,IAAI,iBAAiB;AAAA,MACnC,QAAQ,KAAK,OAAO,OAAO,GAAG;AAAA,MAC9B,KAAK;AAAA,MACL,MAAM;AAAA,MACN,aAAa;AAAA,MACb,iBAAiB;AAAA,MACjB,UAAU;AAAA,QACR;AAAA,QACA,YAAY,KAAK,YAAY;AAAA,MAC/B;AAAA,IACF,CAAC;AAED,UAAM,KAAK,SAAS,KAAK,OAAO;AAAA,EAClC;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,mBAAmB,SAA+B;AAE9D,QAAI,KAAK,gBAAgB;AACvB,YAAM,SAAS,MAAM,KAAK,eAAe,QAAQ;AACjD,UAAI;AACF,cAAM,SAAS,MAAM,OAAO;AAAA,UAC1B;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,UAMA,CAAC,OAAO;AAAA,QACV;AAEA,YAAI,OAAO,KAAK,SAAS,GAAG;AAC1B,iBAAO,OAAO,KAAK,CAAC,EAAE;AAAA,QACxB;AAAA,MACF,UAAE;AACA,eAAO,QAAQ;AAAA,MACjB;AAAA,IACF;AAIA,UAAM,OAAO,oBAAI,KAAK;AACtB,aAAS,IAAI,GAAG,IAAI,GAAG,KAAK;AAC1B,YAAM,YAAY,IAAI,KAAK,IAAI;AAC/B,gBAAU,SAAS,UAAU,SAAS,IAAI,CAAC;AAC3C,YAAM,YAAY,GAAG,UAAU,YAAY,CAAC,IAAI,OAAO,UAAU,SAAS,IAAI,CAAC,EAAE,SAAS,GAAG,GAAG,CAAC;AACjG,YAAM,MAAM,UAAU,SAAS,IAAI,OAAO;AAE1C,UAAI;AACF,cAAM,UAAU,IAAI,iBAAiB;AAAA,UACnC,QAAQ,KAAK,OAAO,OAAO,GAAG;AAAA,UAC9B,KAAK;AAAA,QACP,CAAC;AAED,cAAM,WAAW,MAAM,KAAK,SAAS,KAAK,OAAO;AACjD,YAAI,CAAC,SAAS,KAAM;AACpB,cAAM,OAAO,MAAM,SAAS,KAAK,qBAAqB;AACtD,cAAM,eAAe,MAAM,OAAO,OAAO,KAAK,IAAI,CAAC;AACnD,eAAO,KAAK,MAAM,aAAa,SAAS,CAAC;AAAA,MAC3C,QAAQ;AAAA,MAER;AAAA,IACF;AAEA,WAAO;AAAA,EACT;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,kBACZ,IACA,MACA,UACe;AACf,SAAK,aAAa,KAAK;AAAA,MACrB;AAAA,MACA;AAAA,MACA;AAAA,MACA,WAAW,KAAK,IAAI;AAAA,IACtB,CAAC;AAGD,QAAI,KAAK,OAAO,UAAU,kBAAkB;AAC1C,YAAM,KAAK,iBAAiB;AAAA,IAC9B;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,sBAAqC;AACjD,QAAI,KAAK,aAAa,WAAW,EAAG;AAGpC,UAAM,WAAW,MAAM,KAAK,kBAAkB;AAC9C,QAAI,CAAC,SAAU;AAGf,SAAK,aAAa,KAAK,CAAC,GAAG,MAAM;AAC/B,UAAI,EAAE,aAAa,EAAE,UAAU;AAC7B,eAAO,EAAE,aAAa,SAAS,KAAK;AAAA,MACtC;AACA,aAAO,EAAE,YAAY,EAAE;AAAA,IACzB,CAAC;AAED,UAAM,QAAQ,KAAK,aAAa,OAAO,GAAG,EAAE;AAC5C,eAAW,QAAQ,OAAO;AACxB,UAAI;AACF,cAAM,KAAK,WAAW,KAAK,IAAI,KAAK,IAAI;AAAA,MAC1C,QAAQ;AAEN,aAAK,aAAa,KAAK,IAAI;AAAA,MAC7B;AAAA,IACF;AAEA,UAAM,KAAK,iBAAiB;AAAA,EAC9B;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,oBAAsC;AAClD,QAAI;AAEF,YAAM,WAAW,MAAM,MAAM,0BAA0B;AACvD,aAAO,SAAS;AAAA,IAClB,QAAQ;AACN,aAAO;AAAA,IACT;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKQ,kBAAkB,SAAuB;AAC/C,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAMF,EACC,IAAI,OAAO;AAAA,EAChB;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,mBAAkC;AAC9C,QAAI,CAAC,KAAK,OAAO,UAAU,iBAAkB;AAE7C,UAAM,KAAK,MAAM,OAAO,aAAa;AACrC,UAAM,GAAG;AAAA,MACP,KAAK,OAAO,UAAU;AAAA,MACtB,KAAK,UAAU,KAAK,YAAY;AAAA,MAChC;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,mBAAkC;AAC9C,QAAI,CAAC,KAAK,OAAO,UAAU,iBAAkB;AAE7C,UAAM,KAAK,MAAM,OAAO,aAAa;AACrC,QAAI;AACF,YAAM,OAAO,MAAM,GAAG;AAAA,QACpB,KAAK,OAAO,UAAU;AAAA,QACtB;AAAA,MACF;AACA,WAAK,eAAe,KAAK,MAAM,IAAI;AACnC,WAAK,OAAO;AAAA,QACV,UAAU,KAAK,aAAa,MAAM;AAAA,MACpC;AAAA,IACF,QAAQ;AAAA,IAER;AAAA,EACF;AAAA;AAAA;AAAA;AAAA,EAKA,MAAc,cAA6B;AAEzC,UAAM,aAAa,KAAK,GACrB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAQF,EACC,IAAI;AAEP,SAAK,MAAM,eAAe;AAC1B,eAAW,QAAQ,YAAY;AAC7B,WAAK,MAAM,iBAAiB,KAAK,IAAmB,IAAI,KAAK;AAC7D,WAAK,MAAM,gBAAgB,KAAK;AAAA,IAClC;AAGA,UAAM,mBAAmB,KAAK,GAC3B;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,IAOF,EACC,IAAI;AAEP,SAAK,MAAM,mBAAmB,kBAAkB,SAAS;AAGzD,UAAM,UAAU,KAAK,GAClB;AAAA,MACC;AAAA;AAAA;AAAA;AAAA;AAAA,IAKF,EACC,IAAI;AAEP,SAAK,MAAM,oBAAoB,QAAQ;AAGvC,SAAK,GACF;AAAA,MACC;AAAA;AAAA;AAAA;AAAA,IAIF,EACC;AAAA,MACC,OAAO,OAAO,KAAK,MAAM,gBAAgB,EAAE,OAAO,CAAC,GAAG,MAAM,IAAI,GAAG,CAAC;AAAA,MACpE,KAAK,MAAM;AAAA,MACX,KAAK,MAAM;AAAA,IACb;AAAA,EACJ;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,WAAkC;AACtC,UAAM,KAAK,YAAY;AACvB,WAAO,EAAE,GAAG,KAAK,MAAM;AAAA,EACzB;AAAA;AAAA;AAAA;AAAA,EAKA,MAAM,WAA0B;AAC9B,QAAI,KAAK,gBAAgB;AACvB,oBAAc,KAAK,cAAc;AAAA,IACnC;AAEA,UAAM,KAAK,iBAAiB;AAE5B,QAAI,KAAK,aAAa;AACpB,YAAM,KAAK,YAAY,KAAK;AAAA,IAC9B;AAEA,QAAI,KAAK,gBAAgB;AACvB,YAAM,KAAK,eAAe,IAAI;AAAA,IAChC;AAEA,SAAK,GAAG,MAAM;AAEd,SAAK,OAAO,KAAK,mCAAmC;AAAA,EACtD;AACF;AAGO,MAAM,uBAAsC;AAAA,EACjD,OAAO;AAAA,IACL,QAAQ;AAAA,IACR,WAAW;AAAA,IACX,OAAO;AAAA,MACL;AAAA,QACE,MAAM;AAAA,QACN,aAAa;AAAA,QACb,iBAAiB;AAAA,QACjB,iBAAiB;AAAA,QACjB,WAAW;AAAA,MACb;AAAA,MACA;AAAA,QACE,MAAM;AAAA,QACN,aAAa;AAAA,QACb,iBAAiB;AAAA,QACjB,iBAAiB;AAAA,QACjB,WAAW;AAAA,MACb;AAAA,MACA;AAAA,QACE,MAAM;AAAA,QACN,aAAa;AAAA,QACb,iBAAiB;AAAA,QACjB,iBAAiB;AAAA,QACjB,WAAW;AAAA,MACb;AAAA,IACF;AAAA,EACF;AAAA,EACA,QAAQ;AAAA,IACN,IAAI;AAAA,MACF,QAAQ,QAAQ,IAAI,aAAa;AAAA,MACjC,QAAQ,QAAQ,IAAI,cAAc;AAAA,IACpC;AAAA,EACF;AAAA,EACA,WAAW;AAAA,IACT,UAAU;AAAA,MACR,EAAE,MAAM,OAAO,WAAW,KAAK,QAAQ,UAAU;AAAA;AAAA,MACjD,EAAE,MAAM,QAAQ,WAAW,MAAM,QAAQ,UAAU;AAAA;AAAA,MACnD,EAAE,MAAM,cAAc,WAAW,KAAK,QAAQ,SAAS;AAAA;AAAA,IACzD;AAAA,IACA,WAAW;AAAA,IACX,YAAY;AAAA;AAAA,IACZ,kBAAkB;AAAA,EACpB;AACF;",
|
|
6
|
+
"names": ["StorageTier"]
|
|
7
|
+
}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@stackmemoryai/stackmemory",
|
|
3
|
-
"version": "0.3.
|
|
3
|
+
"version": "0.3.15",
|
|
4
4
|
"description": "Lossless memory runtime for AI coding tools - organizes context as a call stack instead of linear chat logs, with team collaboration and infinite retention",
|
|
5
5
|
"engines": {
|
|
6
6
|
"node": ">=20.0.0",
|