empathy-framework 5.0.3__py3-none-any.whl → 5.1.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (57) hide show
  1. {empathy_framework-5.0.3.dist-info → empathy_framework-5.1.0.dist-info}/METADATA +259 -142
  2. {empathy_framework-5.0.3.dist-info → empathy_framework-5.1.0.dist-info}/RECORD +56 -26
  3. empathy_framework-5.1.0.dist-info/licenses/LICENSE +201 -0
  4. empathy_framework-5.1.0.dist-info/licenses/LICENSE_CHANGE_ANNOUNCEMENT.md +101 -0
  5. empathy_os/__init__.py +1 -1
  6. empathy_os/cli/commands/batch.py +5 -5
  7. empathy_os/cli/commands/routing.py +1 -1
  8. empathy_os/cli/commands/workflow.py +2 -1
  9. empathy_os/cli/parsers/cache 2.py +65 -0
  10. empathy_os/cli_minimal.py +3 -3
  11. empathy_os/cli_router 2.py +416 -0
  12. empathy_os/dashboard/__init__.py +1 -2
  13. empathy_os/dashboard/app 2.py +512 -0
  14. empathy_os/dashboard/app.py +1 -1
  15. empathy_os/dashboard/simple_server 2.py +403 -0
  16. empathy_os/dashboard/standalone_server 2.py +536 -0
  17. empathy_os/memory/types 2.py +441 -0
  18. empathy_os/models/__init__.py +19 -0
  19. empathy_os/models/adaptive_routing 2.py +437 -0
  20. empathy_os/models/auth_cli.py +444 -0
  21. empathy_os/models/auth_strategy.py +450 -0
  22. empathy_os/project_index/scanner_parallel 2.py +291 -0
  23. empathy_os/telemetry/agent_coordination 2.py +478 -0
  24. empathy_os/telemetry/agent_coordination.py +3 -3
  25. empathy_os/telemetry/agent_tracking 2.py +350 -0
  26. empathy_os/telemetry/agent_tracking.py +1 -2
  27. empathy_os/telemetry/approval_gates 2.py +563 -0
  28. empathy_os/telemetry/event_streaming 2.py +405 -0
  29. empathy_os/telemetry/event_streaming.py +3 -3
  30. empathy_os/telemetry/feedback_loop 2.py +557 -0
  31. empathy_os/telemetry/feedback_loop.py +1 -1
  32. empathy_os/vscode_bridge 2.py +173 -0
  33. empathy_os/workflows/__init__.py +8 -0
  34. empathy_os/workflows/autonomous_test_gen.py +569 -0
  35. empathy_os/workflows/bug_predict.py +45 -0
  36. empathy_os/workflows/code_review.py +92 -22
  37. empathy_os/workflows/document_gen.py +594 -62
  38. empathy_os/workflows/llm_base.py +363 -0
  39. empathy_os/workflows/perf_audit.py +69 -0
  40. empathy_os/workflows/progressive/README 2.md +454 -0
  41. empathy_os/workflows/progressive/__init__ 2.py +92 -0
  42. empathy_os/workflows/progressive/cli 2.py +242 -0
  43. empathy_os/workflows/progressive/core 2.py +488 -0
  44. empathy_os/workflows/progressive/orchestrator 2.py +701 -0
  45. empathy_os/workflows/progressive/reports 2.py +528 -0
  46. empathy_os/workflows/progressive/telemetry 2.py +280 -0
  47. empathy_os/workflows/progressive/test_gen 2.py +514 -0
  48. empathy_os/workflows/progressive/workflow 2.py +628 -0
  49. empathy_os/workflows/release_prep.py +54 -0
  50. empathy_os/workflows/security_audit.py +154 -79
  51. empathy_os/workflows/test_gen.py +60 -0
  52. empathy_os/workflows/test_gen_behavioral.py +477 -0
  53. empathy_os/workflows/test_gen_parallel.py +341 -0
  54. empathy_framework-5.0.3.dist-info/licenses/LICENSE +0 -139
  55. {empathy_framework-5.0.3.dist-info → empathy_framework-5.1.0.dist-info}/WHEEL +0 -0
  56. {empathy_framework-5.0.3.dist-info → empathy_framework-5.1.0.dist-info}/entry_points.txt +0 -0
  57. {empathy_framework-5.0.3.dist-info → empathy_framework-5.1.0.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,628 @@
1
+ """Base class for progressive workflows with tier escalation.
2
+
3
+ This module provides the foundation for workflows that support progressive
4
+ tier escalation, handling retry logic, escalation decisions, cost management,
5
+ and approval prompts.
6
+ """
7
+
8
+ import logging
9
+ from datetime import datetime
10
+ from typing import Any
11
+
12
+ from empathy_os.workflows.progressive.core import (
13
+ EscalationConfig,
14
+ FailureAnalysis,
15
+ ProgressiveWorkflowResult,
16
+ Tier,
17
+ TierResult,
18
+ )
19
+ from empathy_os.workflows.progressive.orchestrator import MetaOrchestrator
20
+ from empathy_os.workflows.progressive.telemetry import ProgressiveTelemetry
21
+
22
+ logger = logging.getLogger(__name__)
23
+
24
+
25
+ class BudgetExceededError(Exception):
26
+ """Raised when execution cost exceeds configured budget."""
27
+ pass
28
+
29
+
30
+ class UserCancelledError(Exception):
31
+ """Raised when user cancels execution during approval prompt."""
32
+ pass
33
+
34
+
35
+ class ProgressiveWorkflow:
36
+ """Base class for workflows with progressive tier escalation.
37
+
38
+ Implements the core progressive escalation logic:
39
+ 1. Start with cheap tier
40
+ 2. Analyze results with multi-signal failure detection
41
+ 3. Escalate to capable tier if needed
42
+ 4. Use LLM-guided retries with stagnation detection
43
+ 5. Escalate to premium tier if capable tier stagnates
44
+ 6. Request human review if premium tier fails
45
+
46
+ Subclasses should implement:
47
+ - _execute_tier_impl(): Tier-specific execution logic
48
+ - _analyze_item(): Item-specific quality analysis
49
+
50
+ Example:
51
+ class MyProgressiveWorkflow(ProgressiveWorkflow):
52
+ def _execute_tier_impl(self, tier, items, context):
53
+ # Generate items using appropriate model
54
+ return generated_items
55
+
56
+ def _analyze_item(self, item):
57
+ # Analyze item quality
58
+ return FailureAnalysis(...)
59
+
60
+ Attributes:
61
+ config: Escalation configuration
62
+ tier_results: List of tier execution results
63
+ meta_orchestrator: Meta-agent for orchestration decisions
64
+ """
65
+
66
+ def __init__(self, config: EscalationConfig | None = None, user_id: str | None = None):
67
+ """Initialize progressive workflow.
68
+
69
+ Args:
70
+ config: Escalation configuration (uses defaults if None)
71
+ user_id: Optional user identifier for telemetry (will be hashed)
72
+ """
73
+ self.config = config or EscalationConfig()
74
+ self.tier_results: list[TierResult] = []
75
+ self.meta_orchestrator = MetaOrchestrator()
76
+ self.user_id = user_id
77
+ self.telemetry: ProgressiveTelemetry | None = None # Initialized per workflow
78
+
79
+ def execute(self, **kwargs) -> ProgressiveWorkflowResult:
80
+ """Execute workflow with progressive tier escalation.
81
+
82
+ This is the main entry point. Subclasses typically override this
83
+ to provide workflow-specific logic, then call _execute_progressive()
84
+ to handle the escalation.
85
+
86
+ Args:
87
+ **kwargs: Workflow-specific parameters
88
+
89
+ Returns:
90
+ Complete workflow results with progression history
91
+
92
+ Raises:
93
+ BudgetExceededError: If cost exceeds budget
94
+ UserCancelledError: If user declines approval
95
+ """
96
+ raise NotImplementedError("Subclasses must implement execute()")
97
+
98
+ def _execute_progressive(
99
+ self,
100
+ items: list[Any],
101
+ workflow_name: str,
102
+ **kwargs
103
+ ) -> ProgressiveWorkflowResult:
104
+ """Execute items with progressive tier escalation.
105
+
106
+ Core progressive escalation loop:
107
+ 1. Execute at current tier
108
+ 2. Analyze results
109
+ 3. Separate successful and failed items
110
+ 4. Decide: escalate, retry, or complete
111
+ 5. Repeat with failed items at next tier
112
+
113
+ Args:
114
+ items: Items to process (functions, files, etc.)
115
+ workflow_name: Name of workflow for reporting
116
+ **kwargs: Additional parameters passed to tier execution
117
+
118
+ Returns:
119
+ Complete workflow results
120
+ """
121
+ # Initialize telemetry for this workflow
122
+ self.telemetry = ProgressiveTelemetry(workflow_name, self.user_id)
123
+
124
+ if not self.config.enabled:
125
+ # Progressive escalation disabled, use default tier
126
+ logger.info("Progressive escalation disabled, using default tier")
127
+ return self._execute_single_tier(items, workflow_name, **kwargs)
128
+
129
+ # Estimate cost and request approval
130
+ estimated_cost = self._estimate_total_cost(len(items))
131
+ if not self._request_approval(
132
+ f"Execute {workflow_name} on {len(items)} items",
133
+ estimated_cost
134
+ ):
135
+ raise UserCancelledError("User declined to proceed")
136
+
137
+ # Start with cheapest tier
138
+ current_tier = self.config.tiers[0]
139
+ remaining_items = items
140
+ context: dict[str, Any] | None = None
141
+
142
+ while remaining_items and current_tier:
143
+ logger.info(
144
+ f"Executing {len(remaining_items)} items at {current_tier.value} tier"
145
+ )
146
+
147
+ # Execute at current tier
148
+ tier_result = self._execute_tier(
149
+ current_tier,
150
+ remaining_items,
151
+ context,
152
+ **kwargs
153
+ )
154
+
155
+ self.tier_results.append(tier_result)
156
+
157
+ # Track tier execution in telemetry
158
+ if self.telemetry:
159
+ self.telemetry.track_tier_execution(
160
+ tier_result=tier_result,
161
+ attempt=tier_result.attempt,
162
+ escalated=False, # Will update if escalation happens
163
+ )
164
+
165
+ # Check budget
166
+ self._check_budget()
167
+
168
+ # Separate successful and failed items
169
+ successful = [
170
+ item for item in tier_result.generated_items
171
+ if item.get("quality_score", 0) >= 80
172
+ ]
173
+ failed = [
174
+ item for item in tier_result.generated_items
175
+ if item.get("quality_score", 0) < 80
176
+ ]
177
+
178
+ logger.info(
179
+ f"{current_tier.value} tier: {len(successful)}/{len(tier_result.generated_items)} "
180
+ f"successful (CQS={tier_result.quality_score:.1f})"
181
+ )
182
+
183
+ # Update remaining items (partial escalation)
184
+ remaining_items = failed
185
+
186
+ # Decide: retry, escalate, or complete
187
+ if not remaining_items:
188
+ # All items successful
189
+ break
190
+
191
+ should_escalate, reason = self._should_escalate(
192
+ current_tier,
193
+ tier_result,
194
+ attempt=tier_result.attempt
195
+ )
196
+
197
+ if should_escalate:
198
+ # Escalate to next tier
199
+ next_tier = self._get_next_tier(current_tier)
200
+
201
+ if next_tier is None:
202
+ # No higher tier available
203
+ logger.warning(
204
+ f"Cannot escalate beyond {current_tier.value} tier, "
205
+ f"{len(remaining_items)} items incomplete"
206
+ )
207
+ tier_result.escalated = True
208
+ tier_result.escalation_reason = "No higher tier available"
209
+ break
210
+
211
+ logger.info(
212
+ f"Escalating {len(remaining_items)} items from "
213
+ f"{current_tier.value} to {next_tier.value}: {reason}"
214
+ )
215
+
216
+ # Track escalation in telemetry
217
+ if self.telemetry:
218
+ current_cost = sum(r.cost for r in self.tier_results)
219
+ self.telemetry.track_escalation(
220
+ from_tier=current_tier,
221
+ to_tier=next_tier,
222
+ reason=reason,
223
+ item_count=len(remaining_items),
224
+ current_cost=current_cost,
225
+ )
226
+
227
+ # Build context for next tier
228
+ context = {
229
+ "previous_tier": current_tier,
230
+ "previous_cqs": tier_result.quality_score,
231
+ "failures": failed,
232
+ "examples": tier_result.generated_items[-3:], # Last 3 attempts
233
+ "reason": reason
234
+ }
235
+
236
+ # Request approval for escalation
237
+ escalation_cost = self._estimate_tier_cost(next_tier, len(remaining_items))
238
+ if not self._request_escalation_approval(
239
+ current_tier,
240
+ next_tier,
241
+ len(remaining_items),
242
+ escalation_cost
243
+ ):
244
+ logger.info("User declined escalation, stopping")
245
+ break
246
+
247
+ tier_result.escalated = True
248
+ tier_result.escalation_reason = reason
249
+ current_tier = next_tier
250
+
251
+ else:
252
+ # No escalation needed (retry at same tier or success)
253
+ break
254
+
255
+ # Compile final result
256
+ task_id = f"{workflow_name}-{datetime.now().strftime('%Y%m%d-%H%M%S')}"
257
+
258
+ # Should always have at least one tier result
259
+ assert self.tier_results, "No tier results generated"
260
+
261
+ result = ProgressiveWorkflowResult(
262
+ workflow_name=workflow_name,
263
+ task_id=task_id,
264
+ tier_results=self.tier_results,
265
+ final_result=self.tier_results[-1],
266
+ total_cost=sum(r.cost for r in self.tier_results),
267
+ total_duration=sum(r.duration for r in self.tier_results),
268
+ success=len(remaining_items) == 0
269
+ )
270
+
271
+ # Track workflow completion in telemetry
272
+ if self.telemetry:
273
+ self.telemetry.track_workflow_completion(result)
274
+
275
+ return result
276
+
277
+ def _execute_single_tier(
278
+ self,
279
+ items: list[Any],
280
+ workflow_name: str,
281
+ **kwargs
282
+ ) -> ProgressiveWorkflowResult:
283
+ """Execute without progressive escalation (single tier).
284
+
285
+ Used when progressive escalation is disabled.
286
+
287
+ Args:
288
+ items: Items to process
289
+ workflow_name: Workflow name
290
+ **kwargs: Additional parameters
291
+
292
+ Returns:
293
+ Workflow results with single tier
294
+ """
295
+ # Use middle tier (capable) as default
296
+ default_tier = Tier.CAPABLE
297
+
298
+ tier_result = self._execute_tier(default_tier, items, None, **kwargs)
299
+ self.tier_results.append(tier_result)
300
+
301
+ task_id = f"{workflow_name}-{datetime.now().strftime('%Y%m%d-%H%M%S')}"
302
+
303
+ return ProgressiveWorkflowResult(
304
+ workflow_name=workflow_name,
305
+ task_id=task_id,
306
+ tier_results=[tier_result],
307
+ final_result=tier_result,
308
+ total_cost=tier_result.cost,
309
+ total_duration=tier_result.duration,
310
+ success=tier_result.quality_score >= 80
311
+ )
312
+
313
+ def _execute_tier(
314
+ self,
315
+ tier: Tier,
316
+ items: list[Any],
317
+ context: dict[str, Any] | None,
318
+ **kwargs
319
+ ) -> TierResult:
320
+ """Execute items at a specific tier.
321
+
322
+ Wrapper that handles timing, cost tracking, and error handling.
323
+ Delegates actual execution to _execute_tier_impl().
324
+
325
+ Args:
326
+ tier: Which tier to execute at
327
+ items: Items to process
328
+ context: Context from previous tier (if escalating)
329
+ **kwargs: Additional parameters
330
+
331
+ Returns:
332
+ Tier execution result
333
+ """
334
+ start_time = datetime.now()
335
+
336
+ try:
337
+ # Let subclass handle actual execution
338
+ generated_items = self._execute_tier_impl(tier, items, context, **kwargs)
339
+
340
+ # Analyze overall quality
341
+ failure_analysis = self._analyze_tier_result(generated_items)
342
+
343
+ # Calculate cost (placeholder - will be implemented)
344
+ cost = self._calculate_tier_cost(tier, len(items))
345
+
346
+ duration = (datetime.now() - start_time).total_seconds()
347
+
348
+ return TierResult(
349
+ tier=tier,
350
+ model=self._get_model_for_tier(tier),
351
+ attempt=1, # Simplified for now
352
+ timestamp=start_time,
353
+ generated_items=generated_items,
354
+ failure_analysis=failure_analysis,
355
+ cost=cost,
356
+ duration=duration
357
+ )
358
+
359
+ except Exception as e:
360
+ logger.exception(f"Error executing tier {tier.value}: {e}")
361
+ # Return failed result
362
+ duration = (datetime.now() - start_time).total_seconds()
363
+ return TierResult(
364
+ tier=tier,
365
+ model=self._get_model_for_tier(tier),
366
+ attempt=1,
367
+ timestamp=start_time,
368
+ generated_items=[],
369
+ failure_analysis=FailureAnalysis(),
370
+ cost=0.0,
371
+ duration=duration,
372
+ escalated=True,
373
+ escalation_reason=f"Execution error: {str(e)}"
374
+ )
375
+
376
+ def _execute_tier_impl(
377
+ self,
378
+ tier: Tier,
379
+ items: list[Any],
380
+ context: dict[str, Any] | None,
381
+ **kwargs
382
+ ) -> list[dict[str, Any]]:
383
+ """Execute items at specific tier (to be implemented by subclasses).
384
+
385
+ Args:
386
+ tier: Which tier to execute at
387
+ items: Items to process
388
+ context: Context from previous tier
389
+ **kwargs: Additional parameters
390
+
391
+ Returns:
392
+ List of generated items with quality scores
393
+ """
394
+ raise NotImplementedError("Subclasses must implement _execute_tier_impl()")
395
+
396
+ def _analyze_tier_result(self, generated_items: list[dict[str, Any]]) -> FailureAnalysis:
397
+ """Analyze overall quality of tier execution.
398
+
399
+ Args:
400
+ generated_items: Items generated at this tier
401
+
402
+ Returns:
403
+ Aggregated failure analysis
404
+ """
405
+ if not generated_items:
406
+ return FailureAnalysis()
407
+
408
+ # Aggregate metrics across all items
409
+ total_items = len(generated_items)
410
+ passed = sum(1 for item in generated_items if item.get("passed", False))
411
+ syntax_errors = sum(len(item.get("syntax_errors", [])) for item in generated_items)
412
+
413
+ avg_coverage = sum(item.get("coverage", 0) for item in generated_items) / total_items
414
+ avg_assertions = sum(item.get("assertions", 0) for item in generated_items) / total_items
415
+ avg_confidence = sum(item.get("confidence", 0) for item in generated_items) / total_items
416
+
417
+ return FailureAnalysis(
418
+ syntax_errors=[SyntaxError(f"Syntax error {i}") for i in range(min(syntax_errors, 10))],
419
+ test_pass_rate=passed / total_items if total_items > 0 else 0.0,
420
+ coverage_percent=avg_coverage,
421
+ assertion_depth=avg_assertions,
422
+ confidence_score=avg_confidence
423
+ )
424
+
425
+ def _should_escalate(
426
+ self,
427
+ tier: Tier,
428
+ result: TierResult,
429
+ attempt: int
430
+ ) -> tuple[bool, str]:
431
+ """Determine if escalation is needed.
432
+
433
+ Uses meta-orchestrator to make intelligent escalation decisions
434
+ based on tier, quality score, and attempt number.
435
+
436
+ Args:
437
+ tier: Current tier
438
+ result: Tier execution result
439
+ attempt: Attempt number at this tier
440
+
441
+ Returns:
442
+ Tuple of (should_escalate, reason)
443
+ """
444
+ return self.meta_orchestrator.should_escalate(
445
+ tier,
446
+ result,
447
+ attempt,
448
+ self.config
449
+ )
450
+
451
+ def _get_next_tier(self, current_tier: Tier) -> Tier | None:
452
+ """Get the next tier in the progression.
453
+
454
+ Args:
455
+ current_tier: Current tier
456
+
457
+ Returns:
458
+ Next tier, or None if at highest tier
459
+ """
460
+ try:
461
+ current_index = self.config.tiers.index(current_tier)
462
+ if current_index < len(self.config.tiers) - 1:
463
+ return self.config.tiers[current_index + 1]
464
+ except ValueError:
465
+ pass
466
+
467
+ return None
468
+
469
+ def _estimate_total_cost(self, item_count: int) -> float:
470
+ """Estimate total cost with probabilistic escalation.
471
+
472
+ Args:
473
+ item_count: Number of items to process
474
+
475
+ Returns:
476
+ Estimated total cost in USD
477
+ """
478
+ # Base cost: all items at cheap tier
479
+ cheap_cost = self._estimate_tier_cost(Tier.CHEAP, item_count)
480
+
481
+ # Estimated escalation (30% to capable, 10% to premium)
482
+ capable_cost = self._estimate_tier_cost(Tier.CAPABLE, int(item_count * 0.3))
483
+ premium_cost = self._estimate_tier_cost(Tier.PREMIUM, int(item_count * 0.1))
484
+
485
+ return cheap_cost + capable_cost + premium_cost
486
+
487
+ def _estimate_tier_cost(self, tier: Tier, item_count: int) -> float:
488
+ """Estimate cost for specific tier.
489
+
490
+ Args:
491
+ tier: Which tier
492
+ item_count: Number of items
493
+
494
+ Returns:
495
+ Estimated cost in USD
496
+ """
497
+ # Cost per item (approximate, based on typical token usage)
498
+ COST_PER_ITEM = {
499
+ Tier.CHEAP: 0.003, # ~$0.003 per item (gpt-4o-mini)
500
+ Tier.CAPABLE: 0.015, # ~$0.015 per item (claude-3-5-sonnet)
501
+ Tier.PREMIUM: 0.05 # ~$0.05 per item (claude-opus-4)
502
+ }
503
+
504
+ return COST_PER_ITEM[tier] * item_count
505
+
506
+ def _calculate_tier_cost(self, tier: Tier, item_count: int) -> float:
507
+ """Calculate actual cost for tier execution.
508
+
509
+ TODO: Implement based on actual token usage.
510
+
511
+ Args:
512
+ tier: Which tier
513
+ item_count: Number of items processed
514
+
515
+ Returns:
516
+ Actual cost in USD
517
+ """
518
+ # For now, use estimate
519
+ return self._estimate_tier_cost(tier, item_count)
520
+
521
+ def _request_approval(self, message: str, estimated_cost: float) -> bool:
522
+ """Request user approval for execution.
523
+
524
+ Args:
525
+ message: Description of what will be executed
526
+ estimated_cost: Estimated cost in USD
527
+
528
+ Returns:
529
+ True if approved, False if declined
530
+ """
531
+ # Check auto-approve threshold
532
+ if self.config.auto_approve_under and estimated_cost <= self.config.auto_approve_under:
533
+ logger.info(f"Auto-approved: ${estimated_cost:.2f} <= ${self.config.auto_approve_under:.2f}")
534
+ return True
535
+
536
+ # Check if under default threshold ($1.00)
537
+ threshold = 1.00
538
+ if estimated_cost <= threshold:
539
+ return True
540
+
541
+ # Prompt user
542
+ print("\n⚠️ Cost Estimate:")
543
+ print(f" {message}")
544
+ print(f" Estimated total: ${estimated_cost:.2f}")
545
+ print(f" (Exceeds threshold of ${threshold:.2f})")
546
+ print()
547
+
548
+ response = input("Proceed? [y/N]: ").strip().lower()
549
+ return response == 'y'
550
+
551
+ def _request_escalation_approval(
552
+ self,
553
+ from_tier: Tier,
554
+ to_tier: Tier,
555
+ item_count: int,
556
+ additional_cost: float
557
+ ) -> bool:
558
+ """Request approval for tier escalation.
559
+
560
+ Args:
561
+ from_tier: Current tier
562
+ to_tier: Target tier
563
+ item_count: Number of items to escalate
564
+ additional_cost: Additional cost for escalation
565
+
566
+ Returns:
567
+ True if approved, False if declined
568
+ """
569
+ # Check auto-approve
570
+ total_cost = sum(r.cost for r in self.tier_results) + additional_cost
571
+ if self.config.auto_approve_under and total_cost <= self.config.auto_approve_under:
572
+ logger.info(f"Auto-approved escalation: total ${total_cost:.2f}")
573
+ return True
574
+
575
+ # Prompt user
576
+ print("\n⚠️ Escalation needed:")
577
+ print(f" {item_count} items from {from_tier.value} → {to_tier.value}")
578
+ print(f" Additional cost: ~${additional_cost:.2f}")
579
+ print(f" Total so far: ${sum(r.cost for r in self.tier_results):.2f}")
580
+ print()
581
+
582
+ response = input("Proceed? [Y/n]: ").strip().lower()
583
+ return response != 'n'
584
+
585
+ def _check_budget(self) -> None:
586
+ """Check if budget has been exceeded.
587
+
588
+ Raises:
589
+ BudgetExceededError: If abort_on_budget_exceeded is True
590
+ """
591
+ current_cost = sum(r.cost for r in self.tier_results)
592
+
593
+ if current_cost > self.config.max_cost:
594
+ # Track budget exceeded event
595
+ if self.telemetry:
596
+ action = "abort" if self.config.abort_on_budget_exceeded else "warn"
597
+ self.telemetry.track_budget_exceeded(
598
+ current_cost=current_cost,
599
+ max_budget=self.config.max_cost,
600
+ action=action,
601
+ )
602
+
603
+ if self.config.abort_on_budget_exceeded:
604
+ raise BudgetExceededError(
605
+ f"Cost ${current_cost:.2f} exceeds budget ${self.config.max_cost:.2f}"
606
+ )
607
+ elif self.config.warn_on_budget_exceeded:
608
+ logger.warning(
609
+ f"Cost ${current_cost:.2f} exceeds budget ${self.config.max_cost:.2f}"
610
+ )
611
+
612
+ def _get_model_for_tier(self, tier: Tier) -> str:
613
+ """Get model name for specific tier.
614
+
615
+ Args:
616
+ tier: Which tier
617
+
618
+ Returns:
619
+ Model name (e.g., "gpt-4o-mini")
620
+ """
621
+ # TODO: Make this configurable
622
+ MODEL_MAP = {
623
+ Tier.CHEAP: "gpt-4o-mini",
624
+ Tier.CAPABLE: "claude-3-5-sonnet",
625
+ Tier.PREMIUM: "claude-opus-4"
626
+ }
627
+
628
+ return MODEL_MAP.get(tier, "claude-3-5-sonnet")