@usewhisper/mcp-server 0.1.0 → 0.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (70) hide show
  1. package/README.md +26 -24
  2. package/dist/autosubscribe-6EDKPBE2.js +4068 -0
  3. package/dist/autosubscribe-GHO6YR5A.js +4068 -0
  4. package/dist/autosubscribe-ISDETQIB.js +436 -0
  5. package/dist/autosubscribe-ISDETQIB.js.map +1 -0
  6. package/dist/chunk-3WGYBAYR.js +8387 -0
  7. package/dist/chunk-52VJYCZ7.js +455 -0
  8. package/dist/chunk-5KBZQHDL.js +189 -0
  9. package/dist/chunk-5KIJNY6Z.js +370 -0
  10. package/dist/chunk-7SN3CKDK.js +1076 -0
  11. package/dist/chunk-B3VWOHUA.js +271 -0
  12. package/dist/chunk-C57DHKTL.js +459 -0
  13. package/dist/chunk-EI5CE3EY.js +616 -0
  14. package/dist/chunk-FTWUJBAH.js +387 -0
  15. package/dist/chunk-FTWUJBAH.js.map +1 -0
  16. package/dist/chunk-H3HSKH2P.js +4841 -0
  17. package/dist/chunk-JO3ORBZD.js +616 -0
  18. package/dist/chunk-L6DXSM2U.js +457 -0
  19. package/dist/chunk-L6DXSM2U.js.map +1 -0
  20. package/dist/chunk-LMEYV4JD.js +368 -0
  21. package/dist/chunk-MEFLJ4PV.js +8385 -0
  22. package/dist/chunk-OBLI4FE4.js +276 -0
  23. package/dist/chunk-OBLI4FE4.js.map +1 -0
  24. package/dist/chunk-PPGYJJED.js +271 -0
  25. package/dist/chunk-QGM4M3NI.js +37 -0
  26. package/dist/chunk-T7KMSTWP.js +399 -0
  27. package/dist/chunk-TWEIYHI6.js +399 -0
  28. package/dist/chunk-UYWE7HSU.js +369 -0
  29. package/dist/chunk-UYWE7HSU.js.map +1 -0
  30. package/dist/chunk-X2DL2GWT.js +33 -0
  31. package/dist/chunk-X2DL2GWT.js.map +1 -0
  32. package/dist/chunk-X7HNNNJJ.js +1079 -0
  33. package/dist/consolidation-2GCKI4RE.js +220 -0
  34. package/dist/consolidation-4JOPW6BG.js +220 -0
  35. package/dist/consolidation-FOVQTWNQ.js +222 -0
  36. package/dist/consolidation-IFQ52E44.js +210 -0
  37. package/dist/consolidation-IFQ52E44.js.map +1 -0
  38. package/dist/context-sharing-4ITCNKG4.js +307 -0
  39. package/dist/context-sharing-6CCFIAKL.js +276 -0
  40. package/dist/context-sharing-6CCFIAKL.js.map +1 -0
  41. package/dist/context-sharing-GYKLXHZA.js +307 -0
  42. package/dist/context-sharing-PH64JTXS.js +308 -0
  43. package/dist/context-sharing-Y6LTZZOF.js +307 -0
  44. package/dist/cost-optimization-6OIKRSBV.js +196 -0
  45. package/dist/cost-optimization-6OIKRSBV.js.map +1 -0
  46. package/dist/cost-optimization-7DVSTL6R.js +307 -0
  47. package/dist/cost-optimization-BH5NAX33.js +287 -0
  48. package/dist/cost-optimization-BH5NAX33.js.map +1 -0
  49. package/dist/cost-optimization-F3L5BS5F.js +303 -0
  50. package/dist/ingest-2LPTWUUM.js +16 -0
  51. package/dist/ingest-7T5FAZNC.js +15 -0
  52. package/dist/ingest-EBNIE7XB.js +15 -0
  53. package/dist/ingest-FSHT5BCS.js +15 -0
  54. package/dist/ingest-QE2BTV72.js +15 -0
  55. package/dist/ingest-QE2BTV72.js.map +1 -0
  56. package/dist/oracle-3RLQF3DP.js +259 -0
  57. package/dist/oracle-FKRTQUUG.js +282 -0
  58. package/dist/oracle-J47QCSEW.js +263 -0
  59. package/dist/oracle-MDP5MZRC.js +257 -0
  60. package/dist/oracle-MDP5MZRC.js.map +1 -0
  61. package/dist/search-BLVHWLWC.js +14 -0
  62. package/dist/search-CZ5NYL5B.js +13 -0
  63. package/dist/search-CZ5NYL5B.js.map +1 -0
  64. package/dist/search-EG6TYWWW.js +13 -0
  65. package/dist/search-I22QQA7T.js +13 -0
  66. package/dist/search-T7H5G6DW.js +13 -0
  67. package/dist/server.d.ts +2 -0
  68. package/dist/server.js +914 -1503
  69. package/dist/server.js.map +1 -1
  70. package/package.json +6 -7
@@ -0,0 +1,287 @@
1
+ // src/engine/cost-optimization.ts
2
+ import Anthropic from "@anthropic-ai/sdk";
3
+ var MODELS = {
4
+ haiku: {
5
+ model: "claude-haiku-4.5",
6
+ maxTokens: 4096,
7
+ temperature: 0,
8
+ costPerMillion: 0.25
9
+ // $0.25 per million input tokens
10
+ },
11
+ sonnet: {
12
+ model: "claude-sonnet-4.5",
13
+ maxTokens: 8192,
14
+ temperature: 0,
15
+ costPerMillion: 3
16
+ // $3.00 per million input tokens
17
+ },
18
+ opus: {
19
+ model: "claude-opus-4.5",
20
+ maxTokens: 16384,
21
+ temperature: 0,
22
+ costPerMillion: 15
23
+ // $15.00 per million input tokens
24
+ }
25
+ };
26
+ var TASK_MODEL_MAP = {
27
+ temporal_parsing: "haiku",
28
+ // Fast, simple parsing
29
+ simple_classification: "haiku",
30
+ // Fast classification
31
+ memory_extraction: "sonnet",
32
+ // Needs accuracy for disambiguation
33
+ relation_detection: "sonnet",
34
+ // Needs reasoning
35
+ consolidation: "sonnet",
36
+ // Needs to merge intelligently
37
+ summarization: "haiku",
38
+ // Fast summarization
39
+ complex_reasoning: "opus"
40
+ // Deep reasoning tasks
41
+ };
42
+ function getOptimalModel(taskType, options = {}) {
43
+ if (options.forceModel) {
44
+ return MODELS[options.forceModel];
45
+ }
46
+ let tier = TASK_MODEL_MAP[taskType];
47
+ if (options.minQuality && tier === "haiku") {
48
+ tier = "sonnet";
49
+ }
50
+ return MODELS[tier];
51
+ }
52
+ function estimateCost(params) {
53
+ const modelConfig = getOptimalModel(params.taskType, { forceModel: params.model });
54
+ const inputCost = params.inputTokens / 1e6 * modelConfig.costPerMillion;
55
+ const outputCostPerMillion = modelConfig.costPerMillion * 5;
56
+ const outputCost = params.outputTokens / 1e6 * outputCostPerMillion;
57
+ return {
58
+ model: modelConfig.model,
59
+ inputCost,
60
+ outputCost,
61
+ totalCost: inputCost + outputCost
62
+ };
63
+ }
64
+ async function smartLLMCall(params) {
65
+ const { taskType, prompt, systemPrompt, maxTokens, temperature, forceModel } = params;
66
+ const modelConfig = getOptimalModel(taskType, { forceModel });
67
+ const anthropic = new Anthropic({
68
+ apiKey: process.env.ANTHROPIC_API_KEY || ""
69
+ });
70
+ const messages = [{ role: "user", content: prompt }];
71
+ if (systemPrompt) {
72
+ }
73
+ const response = await anthropic.messages.create({
74
+ model: modelConfig.model,
75
+ max_tokens: maxTokens || modelConfig.maxTokens,
76
+ temperature: temperature !== void 0 ? temperature : modelConfig.temperature,
77
+ messages
78
+ });
79
+ const textContent = response.content.find((c) => c.type === "text");
80
+ const responseText = textContent && textContent.type === "text" ? textContent.text : "";
81
+ const tokensUsed = {
82
+ input: response.usage.input_tokens,
83
+ output: response.usage.output_tokens
84
+ };
85
+ const cost = estimateCost({
86
+ taskType,
87
+ inputTokens: tokensUsed.input,
88
+ outputTokens: tokensUsed.output,
89
+ model: forceModel
90
+ });
91
+ return {
92
+ response: responseText,
93
+ model: modelConfig.model,
94
+ tokensUsed,
95
+ cost: cost.totalCost
96
+ };
97
+ }
98
+ async function batchOptimize(params) {
99
+ const { items, processFn, batchSize = 10, delayMs = 100 } = params;
100
+ const results = [];
101
+ for (let i = 0; i < items.length; i += batchSize) {
102
+ const batch = items.slice(i, i + batchSize);
103
+ const batchResults = await Promise.all(batch.map(processFn));
104
+ results.push(...batchResults);
105
+ if (i + batchSize < items.length) {
106
+ await new Promise((resolve) => setTimeout(resolve, delayMs));
107
+ }
108
+ }
109
+ return results;
110
+ }
111
+ var costRecords = [];
112
+ function trackCost(record) {
113
+ costRecords.push({
114
+ ...record,
115
+ timestamp: /* @__PURE__ */ new Date()
116
+ });
117
+ }
118
+ function getCostSummary(params) {
119
+ const { since, groupBy = "taskType" } = params;
120
+ const filtered = since ? costRecords.filter((r) => r.timestamp >= since) : costRecords;
121
+ const grouped = {};
122
+ for (const record of filtered) {
123
+ const key = record[groupBy];
124
+ if (!grouped[key]) {
125
+ grouped[key] = [];
126
+ }
127
+ grouped[key].push(record);
128
+ }
129
+ const summary = {};
130
+ for (const [key, records] of Object.entries(grouped)) {
131
+ const totalCost = records.reduce((sum, r) => sum + r.cost, 0);
132
+ summary[key] = {
133
+ calls: records.length,
134
+ totalCost,
135
+ avgCost: totalCost / records.length
136
+ };
137
+ }
138
+ return summary;
139
+ }
140
+ function calculateSavings(params) {
141
+ const { since } = params;
142
+ const filtered = since ? costRecords.filter((r) => r.timestamp >= since) : costRecords;
143
+ const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);
144
+ const opusCost = filtered.reduce((sum, r) => {
145
+ const cost = estimateCost({
146
+ taskType: r.taskType,
147
+ inputTokens: r.inputTokens,
148
+ outputTokens: r.outputTokens,
149
+ model: "opus"
150
+ });
151
+ return sum + cost.totalCost;
152
+ }, 0);
153
+ const savings = opusCost - actualCost;
154
+ const savingsPercent = opusCost > 0 ? savings / opusCost * 100 : 0;
155
+ return {
156
+ actualCost,
157
+ opusCost,
158
+ savings,
159
+ savingsPercent
160
+ };
161
+ }
162
+ function recommendModelUpgrades(params) {
163
+ const { errorRates, threshold = 0.05 } = params;
164
+ const recommendations = [];
165
+ for (const [taskType, errorRate] of Object.entries(errorRates)) {
166
+ if (errorRate > threshold) {
167
+ const currentModel = TASK_MODEL_MAP[taskType];
168
+ let recommendedModel;
169
+ if (currentModel === "haiku") {
170
+ recommendedModel = "sonnet";
171
+ } else if (currentModel === "sonnet") {
172
+ recommendedModel = "opus";
173
+ } else {
174
+ continue;
175
+ }
176
+ recommendations.push({
177
+ taskType,
178
+ currentModel,
179
+ recommendedModel
180
+ });
181
+ }
182
+ }
183
+ return recommendations;
184
+ }
185
+ async function getCostBreakdown(params) {
186
+ const { groupBy, startDate, endDate } = params;
187
+ let filtered = [...costRecords];
188
+ if (startDate) {
189
+ filtered = filtered.filter((r) => r.timestamp >= startDate);
190
+ }
191
+ if (endDate) {
192
+ filtered = filtered.filter((r) => r.timestamp <= endDate);
193
+ }
194
+ const groups = {};
195
+ for (const record of filtered) {
196
+ let key;
197
+ switch (groupBy) {
198
+ case "model":
199
+ key = record.model;
200
+ break;
201
+ case "task":
202
+ key = record.taskType;
203
+ break;
204
+ case "day":
205
+ key = record.timestamp.toISOString().split("T")[0];
206
+ break;
207
+ case "hour":
208
+ key = record.timestamp.toISOString().slice(0, 13) + ":00";
209
+ break;
210
+ default:
211
+ key = record.taskType;
212
+ }
213
+ if (!groups[key]) {
214
+ groups[key] = { cost: 0, requests: 0 };
215
+ }
216
+ groups[key].cost += record.cost;
217
+ groups[key].requests += 1;
218
+ }
219
+ const totalCost = filtered.reduce((sum, r) => sum + r.cost, 0);
220
+ const totalRequests = filtered.length;
221
+ return { groups, totalCost, totalRequests };
222
+ }
223
+ async function getSavingsReport(params) {
224
+ const { startDate, endDate } = params;
225
+ let filtered = [...costRecords];
226
+ if (startDate) {
227
+ filtered = filtered.filter((r) => r.timestamp >= startDate);
228
+ }
229
+ if (endDate) {
230
+ filtered = filtered.filter((r) => r.timestamp <= endDate);
231
+ }
232
+ const period = {
233
+ start: filtered.length > 0 ? filtered[0].timestamp : /* @__PURE__ */ new Date(),
234
+ end: filtered.length > 0 ? filtered[filtered.length - 1].timestamp : /* @__PURE__ */ new Date()
235
+ };
236
+ const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);
237
+ let opusOnlyCost = 0;
238
+ const requests = { total: filtered.length, haiku: 0, sonnet: 0, opus: 0 };
239
+ for (const record of filtered) {
240
+ opusOnlyCost += estimateCost({
241
+ taskType: record.taskType,
242
+ inputTokens: record.inputTokens,
243
+ outputTokens: record.outputTokens,
244
+ model: "opus"
245
+ }).totalCost;
246
+ if (record.model.includes("haiku")) {
247
+ requests.haiku++;
248
+ } else if (record.model.includes("sonnet")) {
249
+ requests.sonnet++;
250
+ } else if (record.model.includes("opus")) {
251
+ requests.opus++;
252
+ }
253
+ }
254
+ const savings = opusOnlyCost - actualCost;
255
+ const savingsPercentage = opusOnlyCost > 0 ? savings / opusOnlyCost * 100 : 0;
256
+ let recommendation = "";
257
+ if (savingsPercentage > 50) {
258
+ recommendation = "Excellent! Your model selection is highly optimized.";
259
+ } else if (savingsPercentage > 30) {
260
+ recommendation = "Good savings. Consider using Haiku for simpler tasks.";
261
+ } else {
262
+ recommendation = "Consider reviewing task complexity to better match models.";
263
+ }
264
+ return {
265
+ period,
266
+ actualCost,
267
+ opusOnlyCost,
268
+ savings,
269
+ savingsPercentage,
270
+ requests,
271
+ recommendation
272
+ };
273
+ }
274
+ export {
275
+ MODELS,
276
+ batchOptimize,
277
+ calculateSavings,
278
+ estimateCost,
279
+ getCostBreakdown,
280
+ getCostSummary,
281
+ getOptimalModel,
282
+ getSavingsReport,
283
+ recommendModelUpgrades,
284
+ smartLLMCall,
285
+ trackCost
286
+ };
287
+ //# sourceMappingURL=cost-optimization-BH5NAX33.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../src/engine/cost-optimization.ts"],"sourcesContent":["/**\n * Cost Optimization - Smart model selection based on task complexity\n * Reduces API costs by 60-80% without sacrificing quality\n */\n\nimport Anthropic from \"@anthropic-ai/sdk\";\n\nexport type ModelTier = \"haiku\" | \"sonnet\" | \"opus\";\nexport type TaskType =\n | \"temporal_parsing\"\n | \"memory_extraction\"\n | \"relation_detection\"\n | \"complex_reasoning\"\n | \"consolidation\"\n | \"simple_classification\"\n | \"summarization\";\n\nexport interface ModelConfig {\n model: string;\n maxTokens: number;\n temperature: number;\n costPerMillion: number; // Input tokens\n}\n\n/**\n * Model configurations\n */\nexport const MODELS: Record<ModelTier, ModelConfig> = {\n haiku: {\n model: \"claude-haiku-4.5\",\n maxTokens: 4096,\n temperature: 0.0,\n costPerMillion: 0.25, // $0.25 per million input tokens\n },\n sonnet: {\n model: \"claude-sonnet-4.5\",\n maxTokens: 8192,\n temperature: 0.0,\n costPerMillion: 3.0, // $3.00 per million input tokens\n },\n opus: {\n model: \"claude-opus-4.5\",\n maxTokens: 16384,\n temperature: 0.0,\n costPerMillion: 15.0, // $15.00 per million input tokens\n },\n};\n\n/**\n * Task complexity → Model tier mapping\n */\nconst TASK_MODEL_MAP: Record<TaskType, ModelTier> = {\n temporal_parsing: \"haiku\", // Fast, simple parsing\n simple_classification: \"haiku\", // Fast classification\n memory_extraction: \"sonnet\", // Needs accuracy for disambiguation\n relation_detection: \"sonnet\", // Needs reasoning\n consolidation: \"sonnet\", // Needs to merge intelligently\n summarization: \"haiku\", // Fast summarization\n complex_reasoning: \"opus\", // Deep reasoning tasks\n};\n\n/**\n * Get optimal model for task\n */\nexport function getOptimalModel(\n taskType: TaskType,\n options: {\n forceModel?: ModelTier;\n minQuality?: boolean; // Force higher quality\n } = {}\n): ModelConfig {\n if (options.forceModel) {\n return MODELS[options.forceModel];\n }\n\n let tier = TASK_MODEL_MAP[taskType];\n\n // Upgrade if min quality requested\n if (options.minQuality && tier === \"haiku\") {\n tier = \"sonnet\";\n }\n\n return MODELS[tier];\n}\n\n/**\n * Estimate cost for a task\n */\nexport function estimateCost(params: {\n taskType: TaskType;\n inputTokens: number;\n outputTokens: number;\n model?: ModelTier;\n}): {\n model: string;\n inputCost: number;\n outputCost: number;\n totalCost: number;\n} {\n const modelConfig = getOptimalModel(params.taskType, { forceModel: params.model });\n\n const inputCost = (params.inputTokens / 1_000_000) * modelConfig.costPerMillion;\n\n // Output tokens are 5x more expensive\n const outputCostPerMillion = modelConfig.costPerMillion * 5;\n const outputCost = (params.outputTokens / 1_000_000) * outputCostPerMillion;\n\n return {\n model: modelConfig.model,\n inputCost,\n outputCost,\n totalCost: inputCost + outputCost,\n };\n}\n\n/**\n * Smart LLM call with automatic model selection\n */\nexport async function smartLLMCall(params: {\n taskType: TaskType;\n prompt: string;\n systemPrompt?: string;\n maxTokens?: number;\n temperature?: number;\n forceModel?: ModelTier;\n}): Promise<{\n response: string;\n model: string;\n tokensUsed: {\n input: number;\n output: number;\n };\n cost: number;\n}> {\n const { taskType, prompt, systemPrompt, maxTokens, temperature, forceModel } = params;\n\n const modelConfig = getOptimalModel(taskType, { forceModel });\n\n const anthropic = new Anthropic({\n apiKey: process.env.ANTHROPIC_API_KEY || \"\",\n });\n\n const messages: any[] = [{ role: \"user\", content: prompt }];\n\n if (systemPrompt) {\n // Anthropic uses system parameter, not message\n // This is simplified - use actual SDK approach\n }\n\n const response = await anthropic.messages.create({\n model: modelConfig.model,\n max_tokens: maxTokens || modelConfig.maxTokens,\n temperature: temperature !== undefined ? temperature : modelConfig.temperature,\n messages,\n });\n\n const textContent = response.content.find((c) => c.type === \"text\");\n const responseText = textContent && textContent.type === \"text\" ? textContent.text : \"\";\n\n const tokensUsed = {\n input: response.usage.input_tokens,\n output: response.usage.output_tokens,\n };\n\n const cost = estimateCost({\n taskType,\n inputTokens: tokensUsed.input,\n outputTokens: tokensUsed.output,\n model: forceModel,\n });\n\n return {\n response: responseText,\n model: modelConfig.model,\n tokensUsed,\n cost: cost.totalCost,\n };\n}\n\n/**\n * Batch optimization - group similar tasks\n */\nexport async function batchOptimize<T>(params: {\n items: T[];\n processFn: (item: T) => Promise<any>;\n batchSize?: number;\n delayMs?: number;\n}): Promise<any[]> {\n const { items, processFn, batchSize = 10, delayMs = 100 } = params;\n\n const results: any[] = [];\n\n for (let i = 0; i < items.length; i += batchSize) {\n const batch = items.slice(i, i + batchSize);\n\n const batchResults = await Promise.all(batch.map(processFn));\n\n results.push(...batchResults);\n\n // Small delay between batches\n if (i + batchSize < items.length) {\n await new Promise((resolve) => setTimeout(resolve, delayMs));\n }\n }\n\n return results;\n}\n\n/**\n * Cost tracking\n */\ninterface CostRecord {\n taskType: TaskType;\n model: string;\n inputTokens: number;\n outputTokens: number;\n cost: number;\n timestamp: Date;\n}\n\nconst costRecords: CostRecord[] = [];\n\nexport function trackCost(record: Omit<CostRecord, \"timestamp\">): void {\n costRecords.push({\n ...record,\n timestamp: new Date(),\n });\n}\n\nexport function getCostSummary(params: {\n since?: Date;\n groupBy?: \"taskType\" | \"model\";\n}): Record<string, { calls: number; totalCost: number; avgCost: number }> {\n const { since, groupBy = \"taskType\" } = params;\n\n const filtered = since\n ? costRecords.filter((r) => r.timestamp >= since)\n : costRecords;\n\n const grouped: Record<string, CostRecord[]> = {};\n\n for (const record of filtered) {\n const key = record[groupBy];\n if (!grouped[key]) {\n grouped[key] = [];\n }\n grouped[key].push(record);\n }\n\n const summary: Record<string, { calls: number; totalCost: number; avgCost: number }> = {};\n\n for (const [key, records] of Object.entries(grouped)) {\n const totalCost = records.reduce((sum, r) => sum + r.cost, 0);\n summary[key] = {\n calls: records.length,\n totalCost,\n avgCost: totalCost / records.length,\n };\n }\n\n return summary;\n}\n\n/**\n * Savings calculator - compare with always using Opus\n */\nexport function calculateSavings(params: {\n since?: Date;\n}): {\n actualCost: number;\n opusCost: number;\n savings: number;\n savingsPercent: number;\n} {\n const { since } = params;\n\n const filtered = since\n ? costRecords.filter((r) => r.timestamp >= since)\n : costRecords;\n\n const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);\n\n // Calculate what it would cost with Opus for everything\n const opusCost = filtered.reduce((sum, r) => {\n const cost = estimateCost({\n taskType: r.taskType,\n inputTokens: r.inputTokens,\n outputTokens: r.outputTokens,\n model: \"opus\",\n });\n return sum + cost.totalCost;\n }, 0);\n\n const savings = opusCost - actualCost;\n const savingsPercent = opusCost > 0 ? (savings / opusCost) * 100 : 0;\n\n return {\n actualCost,\n opusCost,\n savings,\n savingsPercent,\n };\n}\n\n/**\n * Recommend model upgrades based on error rates\n */\nexport function recommendModelUpgrades(params: {\n errorRates: Record<TaskType, number>; // Task → error rate\n threshold?: number;\n}): Array<{ taskType: TaskType; currentModel: ModelTier; recommendedModel: ModelTier }> {\n const { errorRates, threshold = 0.05 } = params;\n\n const recommendations: Array<{\n taskType: TaskType;\n currentModel: ModelTier;\n recommendedModel: ModelTier;\n }> = [];\n\n for (const [taskType, errorRate] of Object.entries(errorRates)) {\n if (errorRate > threshold) {\n const currentModel = TASK_MODEL_MAP[taskType as TaskType];\n\n let recommendedModel: ModelTier;\n if (currentModel === \"haiku\") {\n recommendedModel = \"sonnet\";\n } else if (currentModel === \"sonnet\") {\n recommendedModel = \"opus\";\n } else {\n continue; // Already using Opus\n }\n\n recommendations.push({\n taskType: taskType as TaskType,\n currentModel,\n recommendedModel,\n });\n }\n }\n\n return recommendations;\n}\n\n/**\n * Get cost breakdown by different dimensions\n */\nexport async function getCostBreakdown(params: {\n orgId: string;\n projectId?: string;\n groupBy: \"model\" | \"task\" | \"day\" | \"hour\";\n startDate?: Date;\n endDate?: Date;\n}): Promise<{\n groups: Record<string, { cost: number; requests: number }>;\n totalCost: number;\n totalRequests: number;\n}> {\n const { groupBy, startDate, endDate } = params;\n\n let filtered = [...costRecords];\n\n if (startDate) {\n filtered = filtered.filter((r) => r.timestamp >= startDate);\n }\n if (endDate) {\n filtered = filtered.filter((r) => r.timestamp <= endDate);\n }\n\n const groups: Record<string, { cost: number; requests: number }> = {};\n\n for (const record of filtered) {\n let key: string;\n switch (groupBy) {\n case \"model\":\n key = record.model;\n break;\n case \"task\":\n key = record.taskType;\n break;\n case \"day\":\n key = record.timestamp.toISOString().split(\"T\")[0];\n break;\n case \"hour\":\n key = record.timestamp.toISOString().slice(0, 13) + \":00\";\n break;\n default:\n key = record.taskType;\n }\n\n if (!groups[key]) {\n groups[key] = { cost: 0, requests: 0 };\n }\n groups[key].cost += record.cost;\n groups[key].requests += 1;\n }\n\n const totalCost = filtered.reduce((sum, r) => sum + r.cost, 0);\n const totalRequests = filtered.length;\n\n return { groups, totalCost, totalRequests };\n}\n\n/**\n * Get savings report\n */\nexport async function getSavingsReport(params: {\n orgId?: string;\n projectId?: string;\n startDate?: Date;\n endDate?: Date;\n}): Promise<{\n period: { start: Date; end: Date };\n actualCost: number;\n opusOnlyCost: number;\n savings: number;\n savingsPercentage: number;\n requests: { total: number; haiku: number; sonnet: number; opus: number };\n recommendation: string;\n}> {\n const { startDate, endDate } = params;\n\n let filtered = [...costRecords];\n\n if (startDate) {\n filtered = filtered.filter((r) => r.timestamp >= startDate);\n }\n if (endDate) {\n filtered = filtered.filter((r) => r.timestamp <= endDate);\n }\n\n const period = {\n start: filtered.length > 0 ? filtered[0].timestamp : new Date(),\n end: filtered.length > 0 ? filtered[filtered.length - 1].timestamp : new Date(),\n };\n\n const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);\n\n let opusOnlyCost = 0;\n const requests = { total: filtered.length, haiku: 0, sonnet: 0, opus: 0 };\n\n for (const record of filtered) {\n opusOnlyCost += estimateCost({\n taskType: record.taskType,\n inputTokens: record.inputTokens,\n outputTokens: record.outputTokens,\n model: \"opus\",\n }).totalCost;\n\n if (record.model.includes(\"haiku\")) {\n requests.haiku++;\n } else if (record.model.includes(\"sonnet\")) {\n requests.sonnet++;\n } else if (record.model.includes(\"opus\")) {\n requests.opus++;\n }\n }\n\n const savings = opusOnlyCost - actualCost;\n const savingsPercentage = opusOnlyCost > 0 ? (savings / opusOnlyCost) * 100 : 0;\n\n let recommendation = \"\";\n if (savingsPercentage > 50) {\n recommendation = \"Excellent! Your model selection is highly optimized.\";\n } else if (savingsPercentage > 30) {\n recommendation = \"Good savings. Consider using Haiku for simpler tasks.\";\n } else {\n recommendation = \"Consider reviewing task complexity to better match models.\";\n }\n\n return {\n period,\n actualCost,\n opusOnlyCost,\n savings,\n savingsPercentage,\n requests,\n recommendation,\n };\n}\n"],"mappings":";AAKA,OAAO,eAAe;AAsBf,IAAM,SAAyC;AAAA,EACpD,OAAO;AAAA,IACL,OAAO;AAAA,IACP,WAAW;AAAA,IACX,aAAa;AAAA,IACb,gBAAgB;AAAA;AAAA,EAClB;AAAA,EACA,QAAQ;AAAA,IACN,OAAO;AAAA,IACP,WAAW;AAAA,IACX,aAAa;AAAA,IACb,gBAAgB;AAAA;AAAA,EAClB;AAAA,EACA,MAAM;AAAA,IACJ,OAAO;AAAA,IACP,WAAW;AAAA,IACX,aAAa;AAAA,IACb,gBAAgB;AAAA;AAAA,EAClB;AACF;AAKA,IAAM,iBAA8C;AAAA,EAClD,kBAAkB;AAAA;AAAA,EAClB,uBAAuB;AAAA;AAAA,EACvB,mBAAmB;AAAA;AAAA,EACnB,oBAAoB;AAAA;AAAA,EACpB,eAAe;AAAA;AAAA,EACf,eAAe;AAAA;AAAA,EACf,mBAAmB;AAAA;AACrB;AAKO,SAAS,gBACd,UACA,UAGI,CAAC,GACQ;AACb,MAAI,QAAQ,YAAY;AACtB,WAAO,OAAO,QAAQ,UAAU;AAAA,EAClC;AAEA,MAAI,OAAO,eAAe,QAAQ;AAGlC,MAAI,QAAQ,cAAc,SAAS,SAAS;AAC1C,WAAO;AAAA,EACT;AAEA,SAAO,OAAO,IAAI;AACpB;AAKO,SAAS,aAAa,QAU3B;AACA,QAAM,cAAc,gBAAgB,OAAO,UAAU,EAAE,YAAY,OAAO,MAAM,CAAC;AAEjF,QAAM,YAAa,OAAO,cAAc,MAAa,YAAY;AAGjE,QAAM,uBAAuB,YAAY,iBAAiB;AAC1D,QAAM,aAAc,OAAO,eAAe,MAAa;AAEvD,SAAO;AAAA,IACL,OAAO,YAAY;AAAA,IACnB;AAAA,IACA;AAAA,IACA,WAAW,YAAY;AAAA,EACzB;AACF;AAKA,eAAsB,aAAa,QAehC;AACD,QAAM,EAAE,UAAU,QAAQ,cAAc,WAAW,aAAa,WAAW,IAAI;AAE/E,QAAM,cAAc,gBAAgB,UAAU,EAAE,WAAW,CAAC;AAE5D,QAAM,YAAY,IAAI,UAAU;AAAA,IAC9B,QAAQ,QAAQ,IAAI,qBAAqB;AAAA,EAC3C,CAAC;AAED,QAAM,WAAkB,CAAC,EAAE,MAAM,QAAQ,SAAS,OAAO,CAAC;AAE1D,MAAI,cAAc;AAAA,EAGlB;AAEA,QAAM,WAAW,MAAM,UAAU,SAAS,OAAO;AAAA,IAC/C,OAAO,YAAY;AAAA,IACnB,YAAY,aAAa,YAAY;AAAA,IACrC,aAAa,gBAAgB,SAAY,cAAc,YAAY;AAAA,IACnE;AAAA,EACF,CAAC;AAED,QAAM,cAAc,SAAS,QAAQ,KAAK,CAAC,MAAM,EAAE,SAAS,MAAM;AAClE,QAAM,eAAe,eAAe,YAAY,SAAS,SAAS,YAAY,OAAO;AAErF,QAAM,aAAa;AAAA,IACjB,OAAO,SAAS,MAAM;AAAA,IACtB,QAAQ,SAAS,MAAM;AAAA,EACzB;AAEA,QAAM,OAAO,aAAa;AAAA,IACxB;AAAA,IACA,aAAa,WAAW;AAAA,IACxB,cAAc,WAAW;AAAA,IACzB,OAAO;AAAA,EACT,CAAC;AAED,SAAO;AAAA,IACL,UAAU;AAAA,IACV,OAAO,YAAY;AAAA,IACnB;AAAA,IACA,MAAM,KAAK;AAAA,EACb;AACF;AAKA,eAAsB,cAAiB,QAKpB;AACjB,QAAM,EAAE,OAAO,WAAW,YAAY,IAAI,UAAU,IAAI,IAAI;AAE5D,QAAM,UAAiB,CAAC;AAExB,WAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK,WAAW;AAChD,UAAM,QAAQ,MAAM,MAAM,GAAG,IAAI,SAAS;AAE1C,UAAM,eAAe,MAAM,QAAQ,IAAI,MAAM,IAAI,SAAS,CAAC;AAE3D,YAAQ,KAAK,GAAG,YAAY;AAG5B,QAAI,IAAI,YAAY,MAAM,QAAQ;AAChC,YAAM,IAAI,QAAQ,CAAC,YAAY,WAAW,SAAS,OAAO,CAAC;AAAA,IAC7D;AAAA,EACF;AAEA,SAAO;AACT;AAcA,IAAM,cAA4B,CAAC;AAE5B,SAAS,UAAU,QAA6C;AACrE,cAAY,KAAK;AAAA,IACf,GAAG;AAAA,IACH,WAAW,oBAAI,KAAK;AAAA,EACtB,CAAC;AACH;AAEO,SAAS,eAAe,QAG2C;AACxE,QAAM,EAAE,OAAO,UAAU,WAAW,IAAI;AAExC,QAAM,WAAW,QACb,YAAY,OAAO,CAAC,MAAM,EAAE,aAAa,KAAK,IAC9C;AAEJ,QAAM,UAAwC,CAAC;AAE/C,aAAW,UAAU,UAAU;AAC7B,UAAM,MAAM,OAAO,OAAO;AAC1B,QAAI,CAAC,QAAQ,GAAG,GAAG;AACjB,cAAQ,GAAG,IAAI,CAAC;AAAA,IAClB;AACA,YAAQ,GAAG,EAAE,KAAK,MAAM;AAAA,EAC1B;AAEA,QAAM,UAAiF,CAAC;AAExF,aAAW,CAAC,KAAK,OAAO,KAAK,OAAO,QAAQ,OAAO,GAAG;AACpD,UAAM,YAAY,QAAQ,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,MAAM,CAAC;AAC5D,YAAQ,GAAG,IAAI;AAAA,MACb,OAAO,QAAQ;AAAA,MACf;AAAA,MACA,SAAS,YAAY,QAAQ;AAAA,IAC/B;AAAA,EACF;AAEA,SAAO;AACT;AAKO,SAAS,iBAAiB,QAO/B;AACA,QAAM,EAAE,MAAM,IAAI;AAElB,QAAM,WAAW,QACb,YAAY,OAAO,CAAC,MAAM,EAAE,aAAa,KAAK,IAC9C;AAEJ,QAAM,aAAa,SAAS,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,MAAM,CAAC;AAG9D,QAAM,WAAW,SAAS,OAAO,CAAC,KAAK,MAAM;AAC3C,UAAM,OAAO,aAAa;AAAA,MACxB,UAAU,EAAE;AAAA,MACZ,aAAa,EAAE;AAAA,MACf,cAAc,EAAE;AAAA,MAChB,OAAO;AAAA,IACT,CAAC;AACD,WAAO,MAAM,KAAK;AAAA,EACpB,GAAG,CAAC;AAEJ,QAAM,UAAU,WAAW;AAC3B,QAAM,iBAAiB,WAAW,IAAK,UAAU,WAAY,MAAM;AAEnE,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;AAKO,SAAS,uBAAuB,QAGiD;AACtF,QAAM,EAAE,YAAY,YAAY,KAAK,IAAI;AAEzC,QAAM,kBAID,CAAC;AAEN,aAAW,CAAC,UAAU,SAAS,KAAK,OAAO,QAAQ,UAAU,GAAG;AAC9D,QAAI,YAAY,WAAW;AACzB,YAAM,eAAe,eAAe,QAAoB;AAExD,UAAI;AACJ,UAAI,iBAAiB,SAAS;AAC5B,2BAAmB;AAAA,MACrB,WAAW,iBAAiB,UAAU;AACpC,2BAAmB;AAAA,MACrB,OAAO;AACL;AAAA,MACF;AAEA,sBAAgB,KAAK;AAAA,QACnB;AAAA,QACA;AAAA,QACA;AAAA,MACF,CAAC;AAAA,IACH;AAAA,EACF;AAEA,SAAO;AACT;AAKA,eAAsB,iBAAiB,QAUpC;AACD,QAAM,EAAE,SAAS,WAAW,QAAQ,IAAI;AAExC,MAAI,WAAW,CAAC,GAAG,WAAW;AAE9B,MAAI,WAAW;AACb,eAAW,SAAS,OAAO,CAAC,MAAM,EAAE,aAAa,SAAS;AAAA,EAC5D;AACA,MAAI,SAAS;AACX,eAAW,SAAS,OAAO,CAAC,MAAM,EAAE,aAAa,OAAO;AAAA,EAC1D;AAEA,QAAM,SAA6D,CAAC;AAEpE,aAAW,UAAU,UAAU;AAC7B,QAAI;AACJ,YAAQ,SAAS;AAAA,MACf,KAAK;AACH,cAAM,OAAO;AACb;AAAA,MACF,KAAK;AACH,cAAM,OAAO;AACb;AAAA,MACF,KAAK;AACH,cAAM,OAAO,UAAU,YAAY,EAAE,MAAM,GAAG,EAAE,CAAC;AACjD;AAAA,MACF,KAAK;AACH,cAAM,OAAO,UAAU,YAAY,EAAE,MAAM,GAAG,EAAE,IAAI;AACpD;AAAA,MACF;AACE,cAAM,OAAO;AAAA,IACjB;AAEA,QAAI,CAAC,OAAO,GAAG,GAAG;AAChB,aAAO,GAAG,IAAI,EAAE,MAAM,GAAG,UAAU,EAAE;AAAA,IACvC;AACA,WAAO,GAAG,EAAE,QAAQ,OAAO;AAC3B,WAAO,GAAG,EAAE,YAAY;AAAA,EAC1B;AAEA,QAAM,YAAY,SAAS,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,MAAM,CAAC;AAC7D,QAAM,gBAAgB,SAAS;AAE/B,SAAO,EAAE,QAAQ,WAAW,cAAc;AAC5C;AAKA,eAAsB,iBAAiB,QAapC;AACD,QAAM,EAAE,WAAW,QAAQ,IAAI;AAE/B,MAAI,WAAW,CAAC,GAAG,WAAW;AAE9B,MAAI,WAAW;AACb,eAAW,SAAS,OAAO,CAAC,MAAM,EAAE,aAAa,SAAS;AAAA,EAC5D;AACA,MAAI,SAAS;AACX,eAAW,SAAS,OAAO,CAAC,MAAM,EAAE,aAAa,OAAO;AAAA,EAC1D;AAEA,QAAM,SAAS;AAAA,IACb,OAAO,SAAS,SAAS,IAAI,SAAS,CAAC,EAAE,YAAY,oBAAI,KAAK;AAAA,IAC9D,KAAK,SAAS,SAAS,IAAI,SAAS,SAAS,SAAS,CAAC,EAAE,YAAY,oBAAI,KAAK;AAAA,EAChF;AAEA,QAAM,aAAa,SAAS,OAAO,CAAC,KAAK,MAAM,MAAM,EAAE,MAAM,CAAC;AAE9D,MAAI,eAAe;AACnB,QAAM,WAAW,EAAE,OAAO,SAAS,QAAQ,OAAO,GAAG,QAAQ,GAAG,MAAM,EAAE;AAExE,aAAW,UAAU,UAAU;AAC7B,oBAAgB,aAAa;AAAA,MAC3B,UAAU,OAAO;AAAA,MACjB,aAAa,OAAO;AAAA,MACpB,cAAc,OAAO;AAAA,MACrB,OAAO;AAAA,IACT,CAAC,EAAE;AAEH,QAAI,OAAO,MAAM,SAAS,OAAO,GAAG;AAClC,eAAS;AAAA,IACX,WAAW,OAAO,MAAM,SAAS,QAAQ,GAAG;AAC1C,eAAS;AAAA,IACX,WAAW,OAAO,MAAM,SAAS,MAAM,GAAG;AACxC,eAAS;AAAA,IACX;AAAA,EACF;AAEA,QAAM,UAAU,eAAe;AAC/B,QAAM,oBAAoB,eAAe,IAAK,UAAU,eAAgB,MAAM;AAE9E,MAAI,iBAAiB;AACrB,MAAI,oBAAoB,IAAI;AAC1B,qBAAiB;AAAA,EACnB,WAAW,oBAAoB,IAAI;AACjC,qBAAiB;AAAA,EACnB,OAAO;AACL,qBAAiB;AAAA,EACnB;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;","names":[]}
@@ -0,0 +1,303 @@
1
+ import {
2
+ Anthropic
3
+ } from "./chunk-H3HSKH2P.js";
4
+ import "./chunk-QGM4M3NI.js";
5
+
6
+ // ../src/engine/cost-optimization.ts
7
+ var MODELS = {
8
+ haiku: {
9
+ model: "claude-haiku-4.5",
10
+ maxTokens: 4096,
11
+ temperature: 0,
12
+ costPerMillion: 0.25
13
+ // $0.25 per million input tokens
14
+ },
15
+ sonnet: {
16
+ model: "claude-sonnet-4.5",
17
+ maxTokens: 8192,
18
+ temperature: 0,
19
+ costPerMillion: 3
20
+ // $3.00 per million input tokens
21
+ },
22
+ opus: {
23
+ model: "claude-opus-4.5",
24
+ maxTokens: 16384,
25
+ temperature: 0,
26
+ costPerMillion: 15
27
+ // $15.00 per million input tokens
28
+ }
29
+ };
30
+ var TASK_MODEL_MAP = {
31
+ temporal_parsing: "haiku",
32
+ // Fast, simple parsing
33
+ simple_classification: "haiku",
34
+ // Fast classification
35
+ memory_extraction: "sonnet",
36
+ // Needs accuracy for disambiguation
37
+ relation_detection: "sonnet",
38
+ // Needs reasoning
39
+ consolidation: "sonnet",
40
+ // Needs to merge intelligently
41
+ summarization: "haiku",
42
+ // Fast summarization
43
+ complex_reasoning: "opus"
44
+ // Deep reasoning tasks
45
+ };
46
+ function getOptimalModel(taskType, options = {}) {
47
+ if (options.forceModel) {
48
+ return MODELS[options.forceModel];
49
+ }
50
+ let tier = TASK_MODEL_MAP[taskType];
51
+ if (options.minQuality && tier === "haiku") {
52
+ tier = "sonnet";
53
+ }
54
+ return MODELS[tier];
55
+ }
56
+ function estimateCost(params) {
57
+ const modelConfig = getOptimalModel(params.taskType, { forceModel: params.model });
58
+ const inputCost = params.inputTokens / 1e6 * modelConfig.costPerMillion;
59
+ const outputCostPerMillion = modelConfig.costPerMillion * 5;
60
+ const outputCost = params.outputTokens / 1e6 * outputCostPerMillion;
61
+ return {
62
+ model: modelConfig.model,
63
+ inputCost,
64
+ outputCost,
65
+ totalCost: inputCost + outputCost
66
+ };
67
+ }
68
+ async function smartLLMCall(params) {
69
+ const { taskType, prompt, systemPrompt, maxTokens, temperature, forceModel } = params;
70
+ const modelConfig = getOptimalModel(taskType, { forceModel });
71
+ const anthropic = new Anthropic({
72
+ apiKey: process.env.ANTHROPIC_API_KEY || ""
73
+ });
74
+ const messages = [{ role: "user", content: prompt }];
75
+ if (systemPrompt) {
76
+ }
77
+ const response = await anthropic.messages.create({
78
+ model: modelConfig.model,
79
+ max_tokens: maxTokens || modelConfig.maxTokens,
80
+ temperature: temperature !== void 0 ? temperature : modelConfig.temperature,
81
+ messages
82
+ });
83
+ const textContent = response.content.find((c) => c.type === "text");
84
+ const responseText = textContent && textContent.type === "text" ? textContent.text : "";
85
+ const tokensUsed = {
86
+ input: response.usage.input_tokens,
87
+ output: response.usage.output_tokens
88
+ };
89
+ const cost = estimateCost({
90
+ taskType,
91
+ inputTokens: tokensUsed.input,
92
+ outputTokens: tokensUsed.output,
93
+ model: forceModel
94
+ });
95
+ return {
96
+ response: responseText,
97
+ model: modelConfig.model,
98
+ tokensUsed,
99
+ cost: cost.totalCost
100
+ };
101
+ }
102
+ async function batchOptimize(params) {
103
+ const { items, processFn, batchSize = 10, delayMs = 100 } = params;
104
+ const results = [];
105
+ for (let i = 0; i < items.length; i += batchSize) {
106
+ const batch = items.slice(i, i + batchSize);
107
+ const batchResults = await Promise.all(batch.map(processFn));
108
+ results.push(...batchResults);
109
+ if (i + batchSize < items.length) {
110
+ await new Promise((resolve) => setTimeout(resolve, delayMs));
111
+ }
112
+ }
113
+ return results;
114
+ }
115
+ var costRecords = [];
116
+ function trackCost(record) {
117
+ costRecords.push({
118
+ ...record,
119
+ timestamp: /* @__PURE__ */ new Date()
120
+ });
121
+ }
122
+ async function getCostSummary(params) {
123
+ const { startDate, endDate } = params;
124
+ let filtered = [...costRecords];
125
+ if (startDate) {
126
+ filtered = filtered.filter((r) => r.timestamp >= startDate);
127
+ }
128
+ if (endDate) {
129
+ filtered = filtered.filter((r) => r.timestamp <= endDate);
130
+ }
131
+ const period = {
132
+ start: filtered.length > 0 ? filtered[0].timestamp : /* @__PURE__ */ new Date(),
133
+ end: filtered.length > 0 ? filtered[filtered.length - 1].timestamp : /* @__PURE__ */ new Date()
134
+ };
135
+ const totalCost = filtered.reduce((sum, r) => sum + r.cost, 0);
136
+ const totalRequests = filtered.length;
137
+ const costByModel = {};
138
+ const costByTask = {};
139
+ for (const record of filtered) {
140
+ costByModel[record.model] = (costByModel[record.model] || 0) + record.cost;
141
+ costByTask[record.taskType] = (costByTask[record.taskType] || 0) + record.cost;
142
+ }
143
+ const avgCostPerRequest = totalRequests > 0 ? totalCost / totalRequests : 0;
144
+ const daysDiff = period.end.getTime() - period.start.getTime();
145
+ const days = daysDiff > 0 ? daysDiff / (1e3 * 60 * 60 * 24) : 1;
146
+ const estimatedMonthlyCost = totalCost / days * 30;
147
+ return {
148
+ period,
149
+ totalCost,
150
+ totalRequests,
151
+ costByModel,
152
+ costByTask,
153
+ avgCostPerRequest,
154
+ estimatedMonthlyCost
155
+ };
156
+ }
157
+ function calculateSavings(params) {
158
+ const { since } = params;
159
+ const filtered = since ? costRecords.filter((r) => r.timestamp >= since) : costRecords;
160
+ const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);
161
+ const opusCost = filtered.reduce((sum, r) => {
162
+ const cost = estimateCost({
163
+ taskType: r.taskType,
164
+ inputTokens: r.inputTokens,
165
+ outputTokens: r.outputTokens,
166
+ model: "opus"
167
+ });
168
+ return sum + cost.totalCost;
169
+ }, 0);
170
+ const savings = opusCost - actualCost;
171
+ const savingsPercent = opusCost > 0 ? savings / opusCost * 100 : 0;
172
+ return {
173
+ actualCost,
174
+ opusCost,
175
+ savings,
176
+ savingsPercent
177
+ };
178
+ }
179
+ function recommendModelUpgrades(params) {
180
+ const { errorRates, threshold = 0.05 } = params;
181
+ const recommendations = [];
182
+ for (const [taskType, errorRate] of Object.entries(errorRates)) {
183
+ if (errorRate > threshold) {
184
+ const currentModel = TASK_MODEL_MAP[taskType];
185
+ let recommendedModel;
186
+ if (currentModel === "haiku") {
187
+ recommendedModel = "sonnet";
188
+ } else if (currentModel === "sonnet") {
189
+ recommendedModel = "opus";
190
+ } else {
191
+ continue;
192
+ }
193
+ recommendations.push({
194
+ taskType,
195
+ currentModel,
196
+ recommendedModel
197
+ });
198
+ }
199
+ }
200
+ return recommendations;
201
+ }
202
+ async function getCostBreakdown(params) {
203
+ const { groupBy, startDate, endDate } = params;
204
+ let filtered = [...costRecords];
205
+ if (startDate) {
206
+ filtered = filtered.filter((r) => r.timestamp >= startDate);
207
+ }
208
+ if (endDate) {
209
+ filtered = filtered.filter((r) => r.timestamp <= endDate);
210
+ }
211
+ const groups = {};
212
+ for (const record of filtered) {
213
+ let key;
214
+ switch (groupBy) {
215
+ case "model":
216
+ key = record.model;
217
+ break;
218
+ case "task":
219
+ key = record.taskType;
220
+ break;
221
+ case "day":
222
+ key = record.timestamp.toISOString().split("T")[0];
223
+ break;
224
+ case "hour":
225
+ key = record.timestamp.toISOString().slice(0, 13) + ":00";
226
+ break;
227
+ default:
228
+ key = record.taskType;
229
+ }
230
+ if (!groups[key]) {
231
+ groups[key] = { cost: 0, requests: 0 };
232
+ }
233
+ groups[key].cost += record.cost;
234
+ groups[key].requests += 1;
235
+ }
236
+ const totalCost = filtered.reduce((sum, r) => sum + r.cost, 0);
237
+ const totalRequests = filtered.length;
238
+ return { groups, totalCost, totalRequests };
239
+ }
240
+ async function getSavingsReport(params) {
241
+ const { startDate, endDate } = params;
242
+ let filtered = [...costRecords];
243
+ if (startDate) {
244
+ filtered = filtered.filter((r) => r.timestamp >= startDate);
245
+ }
246
+ if (endDate) {
247
+ filtered = filtered.filter((r) => r.timestamp <= endDate);
248
+ }
249
+ const period = {
250
+ start: filtered.length > 0 ? filtered[0].timestamp : /* @__PURE__ */ new Date(),
251
+ end: filtered.length > 0 ? filtered[filtered.length - 1].timestamp : /* @__PURE__ */ new Date()
252
+ };
253
+ const actualCost = filtered.reduce((sum, r) => sum + r.cost, 0);
254
+ let opusOnlyCost = 0;
255
+ const requests = { total: filtered.length, haiku: 0, sonnet: 0, opus: 0 };
256
+ for (const record of filtered) {
257
+ opusOnlyCost += estimateCost({
258
+ taskType: record.taskType,
259
+ inputTokens: record.inputTokens,
260
+ outputTokens: record.outputTokens,
261
+ model: "opus"
262
+ }).totalCost;
263
+ if (record.model.includes("haiku")) {
264
+ requests.haiku++;
265
+ } else if (record.model.includes("sonnet")) {
266
+ requests.sonnet++;
267
+ } else if (record.model.includes("opus")) {
268
+ requests.opus++;
269
+ }
270
+ }
271
+ const savings = opusOnlyCost - actualCost;
272
+ const savingsPercentage = opusOnlyCost > 0 ? savings / opusOnlyCost * 100 : 0;
273
+ let recommendation = "";
274
+ if (savingsPercentage > 50) {
275
+ recommendation = "Excellent! Your model selection is highly optimized.";
276
+ } else if (savingsPercentage > 30) {
277
+ recommendation = "Good savings. Consider using Haiku for simpler tasks.";
278
+ } else {
279
+ recommendation = "Consider reviewing task complexity to better match models.";
280
+ }
281
+ return {
282
+ period,
283
+ actualCost,
284
+ opusOnlyCost,
285
+ savings,
286
+ savingsPercentage,
287
+ requests,
288
+ recommendation
289
+ };
290
+ }
291
+ export {
292
+ MODELS,
293
+ batchOptimize,
294
+ calculateSavings,
295
+ estimateCost,
296
+ getCostBreakdown,
297
+ getCostSummary,
298
+ getOptimalModel,
299
+ getSavingsReport,
300
+ recommendModelUpgrades,
301
+ smartLLMCall,
302
+ trackCost
303
+ };
@@ -0,0 +1,16 @@
1
+ import {
2
+ ingestChunk,
3
+ ingestChunksBatch,
4
+ ingestSession,
5
+ updateMemory
6
+ } from "./chunk-C57DHKTL.js";
7
+ import "./chunk-5KIJNY6Z.js";
8
+ import "./chunk-3WGYBAYR.js";
9
+ import "./chunk-H3HSKH2P.js";
10
+ import "./chunk-QGM4M3NI.js";
11
+ export {
12
+ ingestChunk,
13
+ ingestChunksBatch,
14
+ ingestSession,
15
+ updateMemory
16
+ };
@@ -0,0 +1,15 @@
1
+ import {
2
+ ingestChunk,
3
+ ingestChunksBatch,
4
+ ingestSession,
5
+ updateMemory
6
+ } from "./chunk-EI5CE3EY.js";
7
+ import "./chunk-5KBZQHDL.js";
8
+ import "./chunk-3WGYBAYR.js";
9
+ import "./chunk-QGM4M3NI.js";
10
+ export {
11
+ ingestChunk,
12
+ ingestChunksBatch,
13
+ ingestSession,
14
+ updateMemory
15
+ };
@@ -0,0 +1,15 @@
1
+ import {
2
+ ingestChunk,
3
+ ingestChunksBatch,
4
+ ingestSession,
5
+ updateMemory
6
+ } from "./chunk-52VJYCZ7.js";
7
+ import "./chunk-LMEYV4JD.js";
8
+ import "./chunk-3WGYBAYR.js";
9
+ import "./chunk-QGM4M3NI.js";
10
+ export {
11
+ ingestChunk,
12
+ ingestChunksBatch,
13
+ ingestSession,
14
+ updateMemory
15
+ };
@@ -0,0 +1,15 @@
1
+ import {
2
+ ingestChunk,
3
+ ingestChunksBatch,
4
+ ingestSession,
5
+ updateMemory
6
+ } from "./chunk-JO3ORBZD.js";
7
+ import "./chunk-5KBZQHDL.js";
8
+ import "./chunk-MEFLJ4PV.js";
9
+ import "./chunk-QGM4M3NI.js";
10
+ export {
11
+ ingestChunk,
12
+ ingestChunksBatch,
13
+ ingestSession,
14
+ updateMemory
15
+ };