crackerjack 0.38.0__py3-none-any.whl → 0.38.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of crackerjack might be problematic. Click here for more details.
- crackerjack/__main__.py +79 -2
- crackerjack/cli/options.py +2 -2
- crackerjack/core/workflow_orchestrator.py +66 -5
- crackerjack/managers/publish_manager.py +3 -3
- crackerjack/managers/test_manager.py +118 -9
- crackerjack/mcp/file_monitor.py +12 -2
- crackerjack/mcp/rate_limiter.py +9 -5
- crackerjack/mcp/service_watchdog.py +7 -4
- crackerjack/mcp/tools/intelligence_tools.py +1 -1
- crackerjack/mcp/tools/proactive_tools.py +1 -1
- crackerjack/orchestration/execution_strategies.py +3 -3
- crackerjack/plugins/hooks.py +1 -0
- crackerjack/plugins/loader.py +1 -1
- crackerjack/services/dependency_analyzer.py +4 -4
- crackerjack/services/error_pattern_analyzer.py +4 -4
- crackerjack/services/parallel_executor.py +2 -2
- crackerjack/services/performance_benchmarks.py +3 -1
- crackerjack/services/performance_cache.py +4 -4
- crackerjack/services/quality_intelligence.py +1 -1
- {crackerjack-0.38.0.dist-info → crackerjack-0.38.2.dist-info}/METADATA +3 -2
- {crackerjack-0.38.0.dist-info → crackerjack-0.38.2.dist-info}/RECORD +24 -24
- {crackerjack-0.38.0.dist-info → crackerjack-0.38.2.dist-info}/WHEEL +0 -0
- {crackerjack-0.38.0.dist-info → crackerjack-0.38.2.dist-info}/entry_points.txt +0 -0
- {crackerjack-0.38.0.dist-info → crackerjack-0.38.2.dist-info}/licenses/LICENSE +0 -0
crackerjack/__main__.py
CHANGED
|
@@ -11,6 +11,7 @@ from crackerjack.services.git import GitService
|
|
|
11
11
|
|
|
12
12
|
from .cli import (
|
|
13
13
|
CLI_OPTIONS,
|
|
14
|
+
BumpOption,
|
|
14
15
|
create_options,
|
|
15
16
|
handle_interactive_mode,
|
|
16
17
|
handle_standard_mode,
|
|
@@ -1219,9 +1220,9 @@ def main(
|
|
|
1219
1220
|
update_precommit: bool = CLI_OPTIONS["update_precommit"],
|
|
1220
1221
|
verbose: bool = CLI_OPTIONS["verbose"],
|
|
1221
1222
|
debug: bool = CLI_OPTIONS["debug"],
|
|
1222
|
-
publish:
|
|
1223
|
+
publish: BumpOption | None = CLI_OPTIONS["publish"],
|
|
1223
1224
|
all: str | None = CLI_OPTIONS["all"],
|
|
1224
|
-
bump:
|
|
1225
|
+
bump: BumpOption | None = CLI_OPTIONS["bump"],
|
|
1225
1226
|
strip_code: bool = CLI_OPTIONS["strip_code"],
|
|
1226
1227
|
run_tests: bool = CLI_OPTIONS["run_tests"],
|
|
1227
1228
|
benchmark: bool = CLI_OPTIONS["benchmark"],
|
|
@@ -1466,6 +1467,10 @@ def _process_all_commands(local_vars: t.Any, console: t.Any, options: t.Any) ->
|
|
|
1466
1467
|
):
|
|
1467
1468
|
return False
|
|
1468
1469
|
|
|
1470
|
+
# Handle coverage status command
|
|
1471
|
+
if not _handle_coverage_status(local_vars["coverage_status"], console, options):
|
|
1472
|
+
return False
|
|
1473
|
+
|
|
1469
1474
|
# Handle documentation and analysis commands
|
|
1470
1475
|
return _handle_analysis_commands(local_vars, console, options)
|
|
1471
1476
|
|
|
@@ -1538,6 +1543,78 @@ def _handle_specialized_analytics(local_vars: t.Any, console: t.Any) -> bool:
|
|
|
1538
1543
|
return _handle_enterprise_features(local_vars, console)
|
|
1539
1544
|
|
|
1540
1545
|
|
|
1546
|
+
def _handle_coverage_status(
|
|
1547
|
+
coverage_status: bool, console: t.Any, options: t.Any
|
|
1548
|
+
) -> bool:
|
|
1549
|
+
"""Handle coverage status display command."""
|
|
1550
|
+
if not coverage_status:
|
|
1551
|
+
return True
|
|
1552
|
+
|
|
1553
|
+
try:
|
|
1554
|
+
from pathlib import Path
|
|
1555
|
+
|
|
1556
|
+
from crackerjack.managers.test_manager import TestManager
|
|
1557
|
+
|
|
1558
|
+
# Use current working directory as package path
|
|
1559
|
+
pkg_path = Path.cwd()
|
|
1560
|
+
|
|
1561
|
+
# Create test manager directly
|
|
1562
|
+
test_manager = TestManager(console, pkg_path)
|
|
1563
|
+
|
|
1564
|
+
console.print("[cyan]📊[/cyan] Coverage Status Report")
|
|
1565
|
+
console.print("=" * 50)
|
|
1566
|
+
|
|
1567
|
+
# Get coverage information
|
|
1568
|
+
coverage_info = test_manager.get_coverage()
|
|
1569
|
+
coverage_percent = coverage_info.get("coverage_percent", 0.0)
|
|
1570
|
+
coverage_source = coverage_info.get("source", "unknown")
|
|
1571
|
+
|
|
1572
|
+
if coverage_percent > 0:
|
|
1573
|
+
console.print(
|
|
1574
|
+
f"[green]Current Coverage:[/green] {coverage_percent:.2f}% (from {coverage_source})"
|
|
1575
|
+
)
|
|
1576
|
+
else:
|
|
1577
|
+
console.print(
|
|
1578
|
+
"[yellow]Current Coverage:[/yellow] No coverage data available"
|
|
1579
|
+
)
|
|
1580
|
+
|
|
1581
|
+
# Show status message if available
|
|
1582
|
+
status_message = coverage_info.get("message")
|
|
1583
|
+
if status_message:
|
|
1584
|
+
console.print(f"[dim]{status_message}[/dim]")
|
|
1585
|
+
|
|
1586
|
+
# Try to get more detailed coverage report
|
|
1587
|
+
coverage_report = test_manager.get_coverage_report()
|
|
1588
|
+
if coverage_report:
|
|
1589
|
+
console.print(f"[cyan]Details:[/cyan] {coverage_report}")
|
|
1590
|
+
|
|
1591
|
+
# Show coverage ratchet status if available
|
|
1592
|
+
try:
|
|
1593
|
+
ratchet_status = test_manager.get_coverage_ratchet_status()
|
|
1594
|
+
if ratchet_status:
|
|
1595
|
+
next_milestone = ratchet_status.get("next_milestone")
|
|
1596
|
+
if next_milestone:
|
|
1597
|
+
console.print(f"[cyan]Next Milestone:[/cyan] {next_milestone:.0f}%")
|
|
1598
|
+
|
|
1599
|
+
milestones = ratchet_status.get("milestones_achieved", [])
|
|
1600
|
+
if milestones:
|
|
1601
|
+
console.print(
|
|
1602
|
+
f"[green]Milestones Achieved:[/green] {len(milestones)}"
|
|
1603
|
+
)
|
|
1604
|
+
except Exception:
|
|
1605
|
+
pass # Ignore ratchet status errors
|
|
1606
|
+
|
|
1607
|
+
console.print()
|
|
1608
|
+
return False # Exit after showing status
|
|
1609
|
+
|
|
1610
|
+
except Exception as e:
|
|
1611
|
+
console.print(f"[red]❌[/red] Failed to get coverage status: {e}")
|
|
1612
|
+
import traceback
|
|
1613
|
+
|
|
1614
|
+
console.print(f"[dim]{traceback.format_exc()}[/dim]")
|
|
1615
|
+
return False
|
|
1616
|
+
|
|
1617
|
+
|
|
1541
1618
|
def _handle_enterprise_features(local_vars: t.Any, console: t.Any) -> bool:
|
|
1542
1619
|
"""Handle enterprise features."""
|
|
1543
1620
|
# Handle enterprise optimizer
|
crackerjack/cli/options.py
CHANGED
|
@@ -934,8 +934,8 @@ def create_options(
|
|
|
934
934
|
update_precommit: bool,
|
|
935
935
|
verbose: bool,
|
|
936
936
|
debug: bool,
|
|
937
|
-
publish:
|
|
938
|
-
bump:
|
|
937
|
+
publish: BumpOption | None,
|
|
938
|
+
bump: BumpOption | None,
|
|
939
939
|
benchmark: bool,
|
|
940
940
|
test_workers: int,
|
|
941
941
|
test_timeout: int,
|
|
@@ -66,6 +66,7 @@ class WorkflowPipeline:
|
|
|
66
66
|
self._cache = get_performance_cache()
|
|
67
67
|
|
|
68
68
|
# Initialize quality intelligence for advanced decision making
|
|
69
|
+
self._quality_intelligence: QualityIntelligenceService | None
|
|
69
70
|
try:
|
|
70
71
|
quality_baseline = EnhancedQualityBaselineService()
|
|
71
72
|
self._quality_intelligence = QualityIntelligenceService(quality_baseline)
|
|
@@ -74,6 +75,7 @@ class WorkflowPipeline:
|
|
|
74
75
|
self._quality_intelligence = None
|
|
75
76
|
|
|
76
77
|
# Initialize performance benchmarking for workflow analysis
|
|
78
|
+
self._performance_benchmarks: PerformanceBenchmarkService | None
|
|
77
79
|
try:
|
|
78
80
|
self._performance_benchmarks = PerformanceBenchmarkService(
|
|
79
81
|
console, pkg_path
|
|
@@ -241,7 +243,7 @@ class WorkflowPipeline:
|
|
|
241
243
|
self.logger.warning(f"Failed to auto-start Zuban LSP server: {e}")
|
|
242
244
|
|
|
243
245
|
def _get_zuban_lsp_config(
|
|
244
|
-
self, options: OptionsProtocol, config:
|
|
246
|
+
self, options: OptionsProtocol, config: t.Any
|
|
245
247
|
) -> tuple[int, str]:
|
|
246
248
|
"""Get Zuban LSP configuration values."""
|
|
247
249
|
if config:
|
|
@@ -384,7 +386,7 @@ class WorkflowPipeline:
|
|
|
384
386
|
|
|
385
387
|
def _gather_performance_metrics(
|
|
386
388
|
self, workflow_id: str, duration: float, success: bool
|
|
387
|
-
) -> dict:
|
|
389
|
+
) -> dict[str, t.Any]:
|
|
388
390
|
"""Gather performance metrics from workflow execution."""
|
|
389
391
|
return {
|
|
390
392
|
"workflow_id": workflow_id,
|
|
@@ -622,7 +624,7 @@ class WorkflowPipeline:
|
|
|
622
624
|
comprehensive_passed,
|
|
623
625
|
) = await self._run_main_quality_phases_async(options, workflow_id)
|
|
624
626
|
|
|
625
|
-
return await self.
|
|
627
|
+
return await self._handle_ai_workflow_completion(
|
|
626
628
|
options, iteration, testing_passed, comprehensive_passed, workflow_id
|
|
627
629
|
)
|
|
628
630
|
|
|
@@ -650,7 +652,7 @@ class WorkflowPipeline:
|
|
|
650
652
|
self._mark_code_cleaning_complete()
|
|
651
653
|
return True
|
|
652
654
|
|
|
653
|
-
async def
|
|
655
|
+
async def _handle_ai_workflow_completion(
|
|
654
656
|
self,
|
|
655
657
|
options: OptionsProtocol,
|
|
656
658
|
iteration: int,
|
|
@@ -1871,7 +1873,66 @@ class WorkflowPipeline:
|
|
|
1871
1873
|
) -> bool:
|
|
1872
1874
|
with phase_monitor(workflow_id, "testing") as monitor:
|
|
1873
1875
|
monitor.record_sequential_op()
|
|
1874
|
-
|
|
1876
|
+
test_result = self._run_testing_phase(options)
|
|
1877
|
+
|
|
1878
|
+
# Execute coverage improvement if boost_coverage is enabled and tests passed
|
|
1879
|
+
if test_result and getattr(options, "boost_coverage", False):
|
|
1880
|
+
await self._execute_coverage_improvement(options)
|
|
1881
|
+
|
|
1882
|
+
return test_result
|
|
1883
|
+
|
|
1884
|
+
async def _execute_coverage_improvement(self, options: OptionsProtocol) -> None:
|
|
1885
|
+
"""Execute coverage improvement when boost_coverage is enabled."""
|
|
1886
|
+
try:
|
|
1887
|
+
from crackerjack.orchestration.coverage_improvement import (
|
|
1888
|
+
create_coverage_improvement_orchestrator,
|
|
1889
|
+
)
|
|
1890
|
+
|
|
1891
|
+
coverage_orchestrator = await create_coverage_improvement_orchestrator(
|
|
1892
|
+
self.pkg_path, console=self.console
|
|
1893
|
+
)
|
|
1894
|
+
|
|
1895
|
+
should_improve = await coverage_orchestrator.should_improve_coverage()
|
|
1896
|
+
if not should_improve:
|
|
1897
|
+
self.console.print(
|
|
1898
|
+
"[dim]📈 Coverage at 100% - no improvement needed[/dim]"
|
|
1899
|
+
)
|
|
1900
|
+
return
|
|
1901
|
+
|
|
1902
|
+
# Create agent context for coverage improvement
|
|
1903
|
+
from crackerjack.agents.base import AgentContext
|
|
1904
|
+
from crackerjack.services.filesystem_enhanced import FileSystemService
|
|
1905
|
+
|
|
1906
|
+
filesystem_service = FileSystemService(self.pkg_path)
|
|
1907
|
+
agent_context = AgentContext(
|
|
1908
|
+
pkg_path=self.pkg_path,
|
|
1909
|
+
filesystem=filesystem_service,
|
|
1910
|
+
console=self.console,
|
|
1911
|
+
)
|
|
1912
|
+
|
|
1913
|
+
result = await coverage_orchestrator.execute_coverage_improvement(
|
|
1914
|
+
agent_context
|
|
1915
|
+
)
|
|
1916
|
+
|
|
1917
|
+
if result["status"] == "completed":
|
|
1918
|
+
self.console.print(
|
|
1919
|
+
f"[green]📈[/green] Coverage improvement: {len(result.get('fixes_applied', []))} "
|
|
1920
|
+
f"tests created in {len(result.get('files_modified', []))} files"
|
|
1921
|
+
)
|
|
1922
|
+
elif result["status"] == "skipped":
|
|
1923
|
+
self.console.print(
|
|
1924
|
+
f"[dim]📈 Coverage improvement skipped: {result.get('reason', 'Unknown')}[/dim]"
|
|
1925
|
+
)
|
|
1926
|
+
else:
|
|
1927
|
+
self.console.print(
|
|
1928
|
+
"[yellow]⚠️[/yellow] Coverage improvement completed with issues"
|
|
1929
|
+
)
|
|
1930
|
+
|
|
1931
|
+
except Exception as e:
|
|
1932
|
+
self.console.print(
|
|
1933
|
+
f"[yellow]⚠️[/yellow] Coverage improvement failed: {str(e)}"
|
|
1934
|
+
)
|
|
1935
|
+
self.logger.warning(f"Coverage improvement error: {e}")
|
|
1875
1936
|
|
|
1876
1937
|
async def _execute_standard_hooks_workflow_monitored(
|
|
1877
1938
|
self, options: OptionsProtocol, workflow_id: str
|
|
@@ -156,7 +156,7 @@ class PublishManagerImpl:
|
|
|
156
156
|
self.console.print(f"[red]❌[/ red] Version bump failed: {e}")
|
|
157
157
|
raise
|
|
158
158
|
|
|
159
|
-
def _prompt_for_version_type(self, recommendation=None) -> str:
|
|
159
|
+
def _prompt_for_version_type(self, recommendation: t.Any = None) -> str:
|
|
160
160
|
try:
|
|
161
161
|
from rich.prompt import Prompt
|
|
162
162
|
|
|
@@ -178,7 +178,7 @@ class PublishManagerImpl:
|
|
|
178
178
|
)
|
|
179
179
|
return "patch"
|
|
180
180
|
|
|
181
|
-
def _get_version_recommendation(self):
|
|
181
|
+
def _get_version_recommendation(self) -> t.Any:
|
|
182
182
|
"""Get AI-powered version bump recommendation based on git history."""
|
|
183
183
|
try:
|
|
184
184
|
import asyncio
|
|
@@ -217,7 +217,7 @@ class PublishManagerImpl:
|
|
|
217
217
|
self.console.print(f"[yellow]⚠️[/yellow] Version analysis failed: {e}")
|
|
218
218
|
return None
|
|
219
219
|
|
|
220
|
-
def _display_version_analysis(self, recommendation):
|
|
220
|
+
def _display_version_analysis(self, recommendation: t.Any) -> None:
|
|
221
221
|
"""Display version analysis in a compact format."""
|
|
222
222
|
if not recommendation:
|
|
223
223
|
return
|
|
@@ -132,21 +132,72 @@ class TestManager:
|
|
|
132
132
|
try:
|
|
133
133
|
status = self.coverage_ratchet.get_status_report()
|
|
134
134
|
|
|
135
|
-
if
|
|
135
|
+
# Check if we have actual coverage data from coverage.json even if ratchet is not initialized
|
|
136
|
+
coverage_json_path = self.pkg_path / "coverage.json"
|
|
137
|
+
direct_coverage = None
|
|
138
|
+
|
|
139
|
+
if coverage_json_path.exists():
|
|
140
|
+
try:
|
|
141
|
+
import json
|
|
142
|
+
|
|
143
|
+
with coverage_json_path.open() as f:
|
|
144
|
+
data = json.load(f)
|
|
145
|
+
# Check for totals field first (newer format)
|
|
146
|
+
direct_coverage = data.get("totals", {}).get("percent_covered")
|
|
147
|
+
|
|
148
|
+
# If no totals, calculate from files data (standard pytest-cov format)
|
|
149
|
+
if direct_coverage is None and "files" in data:
|
|
150
|
+
total_statements = 0
|
|
151
|
+
total_covered = 0
|
|
152
|
+
|
|
153
|
+
for file_data in data["files"].values():
|
|
154
|
+
summary = file_data.get("summary", {})
|
|
155
|
+
statements = summary.get("num_statements", 0)
|
|
156
|
+
covered = summary.get("covered_lines", 0)
|
|
157
|
+
total_statements += statements
|
|
158
|
+
total_covered += covered
|
|
159
|
+
|
|
160
|
+
if total_statements > 0:
|
|
161
|
+
direct_coverage = (
|
|
162
|
+
total_covered / total_statements
|
|
163
|
+
) * 100
|
|
164
|
+
except (json.JSONDecodeError, KeyError):
|
|
165
|
+
pass # Fall back to ratchet data
|
|
166
|
+
|
|
167
|
+
# If ratchet is not initialized but we have direct coverage data, use it
|
|
168
|
+
if (
|
|
169
|
+
not status or status.get("status") == "not_initialized"
|
|
170
|
+
) and direct_coverage is not None:
|
|
171
|
+
return {
|
|
172
|
+
"status": "coverage_available",
|
|
173
|
+
"coverage_percent": direct_coverage,
|
|
174
|
+
"message": "Coverage data available from coverage.json",
|
|
175
|
+
"source": "coverage.json",
|
|
176
|
+
}
|
|
177
|
+
|
|
178
|
+
# If ratchet is not initialized and no direct coverage, return not initialized
|
|
179
|
+
if not status or status.get("status") == "not_initialized":
|
|
136
180
|
return {
|
|
137
181
|
"status": "not_initialized",
|
|
138
182
|
"coverage_percent": 0.0,
|
|
139
183
|
"message": "Coverage ratchet not initialized",
|
|
140
184
|
}
|
|
141
185
|
|
|
186
|
+
# Use ratchet data, but prefer direct coverage if available and different
|
|
187
|
+
ratchet_coverage = status.get("current_coverage", 0.0)
|
|
188
|
+
final_coverage = (
|
|
189
|
+
direct_coverage if direct_coverage is not None else ratchet_coverage
|
|
190
|
+
)
|
|
191
|
+
|
|
142
192
|
return {
|
|
143
193
|
"status": "active",
|
|
144
|
-
"coverage_percent":
|
|
194
|
+
"coverage_percent": final_coverage,
|
|
145
195
|
"target_coverage": status.get("target_coverage", 100.0),
|
|
146
196
|
"next_milestone": status.get("next_milestone"),
|
|
147
197
|
"progress_percent": status.get("progress_percent", 0.0),
|
|
148
198
|
"last_updated": status.get("last_updated"),
|
|
149
199
|
"milestones_achieved": status.get("milestones_achieved", []),
|
|
200
|
+
"source": "coverage.json" if direct_coverage is not None else "ratchet",
|
|
150
201
|
}
|
|
151
202
|
except Exception as e:
|
|
152
203
|
return {
|
|
@@ -235,25 +286,83 @@ class TestManager:
|
|
|
235
286
|
current_coverage = None
|
|
236
287
|
coverage_json_path = self.pkg_path / "coverage.json"
|
|
237
288
|
|
|
289
|
+
# Primary: Try to extract from coverage.json
|
|
238
290
|
if coverage_json_path.exists():
|
|
239
|
-
|
|
240
|
-
|
|
241
|
-
|
|
291
|
+
try:
|
|
292
|
+
with coverage_json_path.open() as f:
|
|
293
|
+
data = json.load(f)
|
|
294
|
+
# Check for totals field first (newer format)
|
|
295
|
+
current_coverage = data.get("totals", {}).get("percent_covered")
|
|
296
|
+
|
|
297
|
+
# If no totals, calculate from files data (standard pytest-cov format)
|
|
298
|
+
if current_coverage is None and "files" in data:
|
|
299
|
+
total_statements = 0
|
|
300
|
+
total_covered = 0
|
|
301
|
+
|
|
302
|
+
for file_data in data["files"].values():
|
|
303
|
+
summary = file_data.get("summary", {})
|
|
304
|
+
statements = summary.get("num_statements", 0)
|
|
305
|
+
covered = summary.get("covered_lines", 0)
|
|
306
|
+
total_statements += statements
|
|
307
|
+
total_covered += covered
|
|
308
|
+
|
|
309
|
+
if total_statements > 0:
|
|
310
|
+
current_coverage = (
|
|
311
|
+
total_covered / total_statements
|
|
312
|
+
) * 100
|
|
313
|
+
|
|
314
|
+
if current_coverage is not None:
|
|
315
|
+
self.console.print(
|
|
316
|
+
f"[dim]📊 Coverage extracted from coverage.json: {current_coverage:.2f}%[/dim]"
|
|
317
|
+
)
|
|
318
|
+
except (json.JSONDecodeError, KeyError) as e:
|
|
319
|
+
self.console.print(
|
|
320
|
+
f"[yellow]⚠️[/yellow] Failed to parse coverage.json: {e}"
|
|
321
|
+
)
|
|
242
322
|
|
|
243
|
-
#
|
|
323
|
+
# Secondary: Try ratchet result if coverage.json failed
|
|
244
324
|
if current_coverage is None:
|
|
245
325
|
current_coverage = ratchet_result.get("current_coverage")
|
|
326
|
+
if current_coverage is not None:
|
|
327
|
+
self.console.print(
|
|
328
|
+
f"[dim]📊 Coverage from ratchet result: {current_coverage:.2f}%[/dim]"
|
|
329
|
+
)
|
|
246
330
|
|
|
247
|
-
#
|
|
331
|
+
# Tertiary: Try coverage service, but only accept non-zero values
|
|
248
332
|
if current_coverage is None:
|
|
249
333
|
coverage_info = self.get_coverage()
|
|
250
|
-
|
|
334
|
+
fallback_coverage = coverage_info.get("coverage_percent")
|
|
335
|
+
# Only use fallback if it's meaningful (>0) or if no coverage.json exists
|
|
336
|
+
if fallback_coverage and (
|
|
337
|
+
fallback_coverage > 0 or not coverage_json_path.exists()
|
|
338
|
+
):
|
|
339
|
+
current_coverage = fallback_coverage
|
|
340
|
+
self.console.print(
|
|
341
|
+
f"[dim]📊 Coverage from service fallback: {current_coverage:.2f}%[/dim]"
|
|
342
|
+
)
|
|
343
|
+
else:
|
|
344
|
+
self.console.print(
|
|
345
|
+
"[yellow]⚠️[/yellow] Skipping 0.0% fallback when coverage.json exists"
|
|
346
|
+
)
|
|
251
347
|
|
|
252
|
-
if
|
|
348
|
+
# Only update badge if we have valid coverage data
|
|
349
|
+
if current_coverage is not None and current_coverage >= 0:
|
|
253
350
|
if self._coverage_badge_service.should_update_badge(current_coverage):
|
|
254
351
|
self._coverage_badge_service.update_readme_coverage_badge(
|
|
255
352
|
current_coverage
|
|
256
353
|
)
|
|
354
|
+
self.console.print(
|
|
355
|
+
f"[green]✅[/green] Badge updated to {current_coverage:.2f}%"
|
|
356
|
+
)
|
|
357
|
+
else:
|
|
358
|
+
self.console.print(
|
|
359
|
+
f"[dim]📊 Badge unchanged (current: {current_coverage:.2f}%)[/dim]"
|
|
360
|
+
)
|
|
361
|
+
else:
|
|
362
|
+
self.console.print(
|
|
363
|
+
"[yellow]⚠️[/yellow] No valid coverage data found for badge update"
|
|
364
|
+
)
|
|
365
|
+
|
|
257
366
|
except Exception as e:
|
|
258
367
|
# Don't fail the test process if badge update fails
|
|
259
368
|
self.console.print(f"[yellow]⚠️[/yellow] Badge update failed: {e}")
|
crackerjack/mcp/file_monitor.py
CHANGED
|
@@ -5,10 +5,20 @@ import typing as t
|
|
|
5
5
|
from collections.abc import Callable
|
|
6
6
|
from pathlib import Path
|
|
7
7
|
|
|
8
|
+
# Type aliases for watchdog types
|
|
9
|
+
FileSystemEvent: t.Any
|
|
10
|
+
FileSystemEventHandler: t.Any
|
|
11
|
+
Observer: t.Any
|
|
12
|
+
WATCHDOG_AVAILABLE: bool
|
|
13
|
+
|
|
8
14
|
try:
|
|
9
|
-
from watchdog.events import FileSystemEvent
|
|
10
|
-
from watchdog.
|
|
15
|
+
from watchdog.events import FileSystemEvent as WatchdogFileSystemEvent
|
|
16
|
+
from watchdog.events import FileSystemEventHandler as WatchdogFileSystemEventHandler
|
|
17
|
+
from watchdog.observers import Observer as WatchdogObserver
|
|
11
18
|
|
|
19
|
+
FileSystemEvent = WatchdogFileSystemEvent
|
|
20
|
+
FileSystemEventHandler = WatchdogFileSystemEventHandler
|
|
21
|
+
Observer = WatchdogObserver
|
|
12
22
|
WATCHDOG_AVAILABLE = True
|
|
13
23
|
except ImportError:
|
|
14
24
|
# Type stubs for when watchdog is not available
|
crackerjack/mcp/rate_limiter.py
CHANGED
|
@@ -35,14 +35,18 @@ class RateLimiter:
|
|
|
35
35
|
self.requests_per_hour = requests_per_hour
|
|
36
36
|
|
|
37
37
|
self.minute_windows: dict[str, deque[float]] = defaultdict(
|
|
38
|
-
lambda: deque(maxlen=requests_per_minute), # type: ignore[misc]
|
|
38
|
+
lambda: deque[float](maxlen=requests_per_minute), # type: ignore[arg-type,misc]
|
|
39
39
|
)
|
|
40
40
|
self.hour_windows: dict[str, deque[float]] = defaultdict(
|
|
41
|
-
lambda: deque(maxlen=requests_per_hour), # type: ignore[misc]
|
|
41
|
+
lambda: deque[float](maxlen=requests_per_hour), # type: ignore[arg-type,misc]
|
|
42
42
|
)
|
|
43
43
|
|
|
44
|
-
self.global_minute_window: deque[float] = deque(
|
|
45
|
-
|
|
44
|
+
self.global_minute_window: deque[float] = deque[float](
|
|
45
|
+
maxlen=requests_per_minute * 10
|
|
46
|
+
)
|
|
47
|
+
self.global_hour_window: deque[float] = deque[float](
|
|
48
|
+
maxlen=requests_per_hour * 10
|
|
49
|
+
)
|
|
46
50
|
|
|
47
51
|
self._lock = asyncio.Lock()
|
|
48
52
|
|
|
@@ -126,7 +130,7 @@ class RateLimiter:
|
|
|
126
130
|
self._remove_expired_entries(self.global_minute_window, minute_cutoff)
|
|
127
131
|
self._remove_expired_entries(self.global_hour_window, hour_cutoff)
|
|
128
132
|
|
|
129
|
-
def _remove_expired_entries(self, window: deque, cutoff: float) -> None:
|
|
133
|
+
def _remove_expired_entries(self, window: deque[float], cutoff: float) -> None:
|
|
130
134
|
while window and window[0] < cutoff:
|
|
131
135
|
window.popleft()
|
|
132
136
|
|
|
@@ -3,6 +3,7 @@ import socket
|
|
|
3
3
|
import subprocess
|
|
4
4
|
import sys
|
|
5
5
|
import time
|
|
6
|
+
from contextlib import suppress
|
|
6
7
|
from typing import Any
|
|
7
8
|
|
|
8
9
|
import aiohttp
|
|
@@ -34,7 +35,7 @@ class ServiceConfig:
|
|
|
34
35
|
self.max_restarts = max_restarts
|
|
35
36
|
self.restart_window = restart_window
|
|
36
37
|
|
|
37
|
-
self.process: subprocess.Popen[
|
|
38
|
+
self.process: subprocess.Popen[str] | None = None
|
|
38
39
|
self.restart_count = 0
|
|
39
40
|
self.restart_timestamps: list[float] = []
|
|
40
41
|
self.last_health_check = 0.0
|
|
@@ -125,7 +126,7 @@ class ServiceWatchdog:
|
|
|
125
126
|
return False
|
|
126
127
|
|
|
127
128
|
async def _launch_service_process(self, service: ServiceConfig) -> bool:
|
|
128
|
-
service.process = subprocess.Popen(
|
|
129
|
+
service.process = subprocess.Popen[str](
|
|
129
130
|
service.command,
|
|
130
131
|
stdout=subprocess.PIPE,
|
|
131
132
|
stderr=subprocess.PIPE,
|
|
@@ -137,6 +138,8 @@ class ServiceWatchdog:
|
|
|
137
138
|
return await self._check_process_startup_success(service)
|
|
138
139
|
|
|
139
140
|
async def _check_process_startup_success(self, service: ServiceConfig) -> bool:
|
|
141
|
+
if service.process is None:
|
|
142
|
+
return False
|
|
140
143
|
exit_code = service.process.poll()
|
|
141
144
|
if exit_code is not None:
|
|
142
145
|
return await self._handle_process_died(service, exit_code)
|
|
@@ -147,6 +150,8 @@ class ServiceWatchdog:
|
|
|
147
150
|
service: ServiceConfig,
|
|
148
151
|
exit_code: int,
|
|
149
152
|
) -> bool:
|
|
153
|
+
if service.process is None:
|
|
154
|
+
return False
|
|
150
155
|
stdout, stderr = service.process.communicate()
|
|
151
156
|
error_msg = f"Process died (exit: {exit_code})"
|
|
152
157
|
if stderr and stderr.strip():
|
|
@@ -441,8 +446,6 @@ class ServiceWatchdog:
|
|
|
441
446
|
message: str,
|
|
442
447
|
) -> None:
|
|
443
448
|
if self.event_queue:
|
|
444
|
-
from contextlib import suppress
|
|
445
|
-
|
|
446
449
|
with suppress(Exception):
|
|
447
450
|
event = {
|
|
448
451
|
"type": event_type,
|
|
@@ -119,7 +119,7 @@ def _create_validation_results(file_path: str) -> dict[str, t.Any]:
|
|
|
119
119
|
|
|
120
120
|
|
|
121
121
|
def _create_pattern_suggestions(problem_context: str) -> dict[str, t.Any]:
|
|
122
|
-
pattern_suggestions = {
|
|
122
|
+
pattern_suggestions: dict[str, t.Any] = {
|
|
123
123
|
"context": problem_context,
|
|
124
124
|
"recommended_patterns": [],
|
|
125
125
|
"implementation_guidance": [],
|
|
@@ -262,10 +262,10 @@ class OrchestrationPlanner:
|
|
|
262
262
|
hook_plans=hook_plans,
|
|
263
263
|
test_plan=test_plan,
|
|
264
264
|
ai_plan=ai_plan,
|
|
265
|
-
estimated_total_duration=
|
|
266
|
-
int(plan["estimated_duration"]) for plan in hook_plans
|
|
265
|
+
estimated_total_duration=float(
|
|
266
|
+
sum(int(float(str(plan["estimated_duration"]))) for plan in hook_plans)
|
|
267
267
|
)
|
|
268
|
-
+ test_plan["estimated_duration"],
|
|
268
|
+
+ float(test_plan["estimated_duration"]),
|
|
269
269
|
)
|
|
270
270
|
|
|
271
271
|
def _estimate_strategy_duration(self, strategy: HookStrategy) -> float:
|
crackerjack/plugins/hooks.py
CHANGED
crackerjack/plugins/loader.py
CHANGED
|
@@ -67,10 +67,10 @@ class DependencyEdge:
|
|
|
67
67
|
class DependencyGraph:
|
|
68
68
|
"""Complete dependency graph data structure."""
|
|
69
69
|
|
|
70
|
-
nodes: dict[str, DependencyNode] = field(default_factory=dict
|
|
70
|
+
nodes: dict[str, DependencyNode] = field(default_factory=dict)
|
|
71
71
|
edges: list[DependencyEdge] = field(default_factory=list)
|
|
72
|
-
clusters: dict[str, list[str]] = field(default_factory=dict
|
|
73
|
-
metrics: dict[str, t.Any] = field(default_factory=dict
|
|
72
|
+
clusters: dict[str, list[str]] = field(default_factory=dict)
|
|
73
|
+
metrics: dict[str, t.Any] = field(default_factory=dict)
|
|
74
74
|
generated_at: datetime = field(default_factory=datetime.now)
|
|
75
75
|
|
|
76
76
|
def to_dict(self) -> dict[str, t.Any]:
|
|
@@ -193,7 +193,7 @@ class DependencyAnalyzer:
|
|
|
193
193
|
edges = self.dependency_graph.edges
|
|
194
194
|
|
|
195
195
|
# Basic metrics
|
|
196
|
-
metrics = {
|
|
196
|
+
metrics: dict[str, t.Any] = {
|
|
197
197
|
"total_nodes": len(nodes),
|
|
198
198
|
"total_edges": len(edges),
|
|
199
199
|
"total_clusters": len(self.dependency_graph.clusters),
|
|
@@ -157,7 +157,7 @@ class ErrorPatternAnalyzer:
|
|
|
157
157
|
error_type = pattern.error_type
|
|
158
158
|
count = pattern.count
|
|
159
159
|
|
|
160
|
-
file_error_counts[file_path][error_type] += count
|
|
160
|
+
file_error_counts[file_path][error_type] += float(count)
|
|
161
161
|
max_value = max(max_value, file_error_counts[file_path][error_type])
|
|
162
162
|
|
|
163
163
|
# Create heat map cells
|
|
@@ -175,7 +175,7 @@ class ErrorPatternAnalyzer:
|
|
|
175
175
|
for error_type in error_types:
|
|
176
176
|
count = file_error_counts[file_path].get(error_type, 0)
|
|
177
177
|
if count > 0:
|
|
178
|
-
intensity: float = count / max_value if max_value > 0 else 0
|
|
178
|
+
intensity: float = count / max_value if max_value > 0 else 0.0
|
|
179
179
|
severity = self._get_severity_for_type(error_type)
|
|
180
180
|
|
|
181
181
|
cells.append(
|
|
@@ -313,7 +313,7 @@ class ErrorPatternAnalyzer:
|
|
|
313
313
|
self, time_label: str, error_type: str, count: float, max_value: float
|
|
314
314
|
) -> HeatMapCell:
|
|
315
315
|
"""Create a single temporal heatmap cell."""
|
|
316
|
-
intensity: float = count / max_value if max_value > 0 else 0
|
|
316
|
+
intensity: float = count / max_value if max_value > 0 else 0.0
|
|
317
317
|
severity = self._get_severity_for_type(error_type)
|
|
318
318
|
|
|
319
319
|
return HeatMapCell(
|
|
@@ -405,7 +405,7 @@ class ErrorPatternAnalyzer:
|
|
|
405
405
|
self, function_id: str, error_type: str, count: float, max_value: float
|
|
406
406
|
) -> HeatMapCell:
|
|
407
407
|
"""Create a single function heatmap cell."""
|
|
408
|
-
intensity: float = count / max_value if max_value > 0 else 0
|
|
408
|
+
intensity: float = count / max_value if max_value > 0 else 0.0
|
|
409
409
|
severity = self._get_severity_for_type(error_type)
|
|
410
410
|
|
|
411
411
|
return HeatMapCell(
|
|
@@ -239,7 +239,7 @@ class ParallelHookExecutor:
|
|
|
239
239
|
f"Hook {hooks[i].name} failed with exception: {result}"
|
|
240
240
|
)
|
|
241
241
|
else:
|
|
242
|
-
processed_results.append(result)
|
|
242
|
+
processed_results.append(t.cast(ExecutionResult, result))
|
|
243
243
|
|
|
244
244
|
successful = sum(1 for r in processed_results if r.success)
|
|
245
245
|
self._logger.info(
|
|
@@ -307,7 +307,7 @@ class AsyncCommandExecutor:
|
|
|
307
307
|
)
|
|
308
308
|
processed_results.append(error_result)
|
|
309
309
|
else:
|
|
310
|
-
processed_results.append(result)
|
|
310
|
+
processed_results.append(t.cast(ExecutionResult, result))
|
|
311
311
|
|
|
312
312
|
successful = sum(1 for r in processed_results if r.success)
|
|
313
313
|
self._logger.info(
|
|
@@ -7,6 +7,8 @@ from dataclasses import dataclass, field
|
|
|
7
7
|
from datetime import datetime
|
|
8
8
|
from pathlib import Path
|
|
9
9
|
|
|
10
|
+
from rich.console import Console
|
|
11
|
+
|
|
10
12
|
from crackerjack.services.logging import get_logger
|
|
11
13
|
from crackerjack.services.memory_optimizer import get_memory_optimizer
|
|
12
14
|
from crackerjack.services.performance_cache import get_performance_cache
|
|
@@ -301,7 +303,7 @@ class PerformanceBenchmarker:
|
|
|
301
303
|
class PerformanceBenchmarkService:
|
|
302
304
|
"""Service wrapper for performance benchmarking in workflow orchestration."""
|
|
303
305
|
|
|
304
|
-
def __init__(self, console, pkg_path):
|
|
306
|
+
def __init__(self, console: Console, pkg_path: Path) -> None:
|
|
305
307
|
self._console = console
|
|
306
308
|
self._pkg_path = pkg_path
|
|
307
309
|
self._benchmarker = PerformanceBenchmarker()
|
|
@@ -268,7 +268,7 @@ class GitOperationCache:
|
|
|
268
268
|
) -> None:
|
|
269
269
|
key = self._make_repo_key(repo_path, "branch_info")
|
|
270
270
|
invalidation_keys = {f"git_repo: {repo_path}"}
|
|
271
|
-
self.cache.set
|
|
271
|
+
self.cache.set(key, branch_info, ttl_seconds, invalidation_keys)
|
|
272
272
|
|
|
273
273
|
def get_file_status(self, repo_path: Path) -> t.Any:
|
|
274
274
|
key = self._make_repo_key(repo_path, "file_status")
|
|
@@ -282,7 +282,7 @@ class GitOperationCache:
|
|
|
282
282
|
) -> None:
|
|
283
283
|
key = self._make_repo_key(repo_path, "file_status")
|
|
284
284
|
invalidation_keys = {f"git_repo: {repo_path}", "git_files"}
|
|
285
|
-
self.cache.set
|
|
285
|
+
self.cache.set(key, file_status, ttl_seconds, invalidation_keys)
|
|
286
286
|
|
|
287
287
|
def invalidate_repo(self, repo_path: Path) -> None:
|
|
288
288
|
self.cache.invalidate(f"git_repo: {repo_path}")
|
|
@@ -312,7 +312,7 @@ class FileSystemCache:
|
|
|
312
312
|
) -> None:
|
|
313
313
|
key = self._make_file_key(file_path, "stats")
|
|
314
314
|
invalidation_keys = {f"file: {file_path}"}
|
|
315
|
-
self.cache.set
|
|
315
|
+
self.cache.set(key, stats, ttl_seconds, invalidation_keys)
|
|
316
316
|
|
|
317
317
|
def invalidate_file(self, file_path: Path) -> None:
|
|
318
318
|
self.cache.invalidate(f"file: {file_path}")
|
|
@@ -352,7 +352,7 @@ class CommandResultCache:
|
|
|
352
352
|
if cwd:
|
|
353
353
|
invalidation_keys.add(f"cwd: {cwd}")
|
|
354
354
|
|
|
355
|
-
self.cache.set
|
|
355
|
+
self.cache.set(key, result, ttl_seconds, invalidation_keys)
|
|
356
356
|
|
|
357
357
|
def invalidate_commands(self) -> None:
|
|
358
358
|
self.cache.invalidate("commands")
|
|
@@ -264,7 +264,7 @@ class QualityIntelligenceService:
|
|
|
264
264
|
metric_name: str,
|
|
265
265
|
values: list[float],
|
|
266
266
|
timestamps: list[t.Any],
|
|
267
|
-
stats_data: dict[str,
|
|
267
|
+
stats_data: dict[str, t.Any],
|
|
268
268
|
) -> list[QualityAnomaly]:
|
|
269
269
|
"""Identify outlier anomalies based on z-scores."""
|
|
270
270
|
anomalies = []
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: crackerjack
|
|
3
|
-
Version: 0.38.
|
|
3
|
+
Version: 0.38.2
|
|
4
4
|
Summary: Crackerjack Python project management tool
|
|
5
5
|
Project-URL: documentation, https://github.com/lesleslie/crackerjack
|
|
6
6
|
Project-URL: homepage, https://github.com/lesleslie/crackerjack
|
|
@@ -67,13 +67,14 @@ Description-Content-Type: text/markdown
|
|
|
67
67
|
|
|
68
68
|
# Crackerjack: Advanced AI-Driven Python Development Platform
|
|
69
69
|
|
|
70
|
+
[](https://github.com/lesleslie/crackerjack)
|
|
70
71
|
[](https://www.python.org/downloads/)
|
|
71
72
|
[](https://pytest.org)
|
|
72
73
|
[](https://github.com/astral-sh/ruff)
|
|
73
74
|
[](https://github.com/astral-sh/uv)
|
|
74
75
|
[](https://github.com/pre-commit/pre-commit)
|
|
75
76
|
[](https://opensource.org/licenses/BSD-3-Clause)
|
|
76
|
-

|
|
77
78
|
|
|
78
79
|
## 🎯 Purpose
|
|
79
80
|
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
crackerjack/__init__.py,sha256=DajG9zHB8qBdgdiKMumrrssUbKeMXmtIQ3oOaSTb46Y,1426
|
|
2
|
-
crackerjack/__main__.py,sha256=
|
|
2
|
+
crackerjack/__main__.py,sha256=cAM0ezqddeO5oC3QMr8zpUor-RN0lIG-Zja_WeatDiY,55176
|
|
3
3
|
crackerjack/api.py,sha256=PyCRaZHvKWdu62_2O4t_HcEfKNBdqyrfPdonS_PNn4c,21495
|
|
4
4
|
crackerjack/code_cleaner.py,sha256=M1zVaq31uW0nOkPneKR8kfR3892gyyVx0VhFgRaxsj4,44338
|
|
5
5
|
crackerjack/dynamic_config.py,sha256=4c8Fts9vyH8Tdon_47OFVT1iTBINSzSgB0WoeSvpzII,22418
|
|
@@ -34,7 +34,7 @@ crackerjack/cli/cache_handlers_enhanced.py,sha256=6X5rYSo1l-rj9eb7eB8mpA-6BlUagy
|
|
|
34
34
|
crackerjack/cli/facade.py,sha256=e4_oB04awqEijI3yqiYAZGc6x09uMBa4ih0SsXpgMuY,3751
|
|
35
35
|
crackerjack/cli/handlers.py,sha256=mYhwMLUKid6mQLff0ScpcnhP0yUS9IzOIMdM7VLkUCc,17178
|
|
36
36
|
crackerjack/cli/interactive.py,sha256=E7WgzgNRlgOuKLbi5Io3RQ3BBXS8sIDiUQcJh6b2x9I,17583
|
|
37
|
-
crackerjack/cli/options.py,sha256=
|
|
37
|
+
crackerjack/cli/options.py,sha256=U31B1UqxtHe4bzW_IEOw0QPCjdj8MbxUvVlL-hCRt9g,36560
|
|
38
38
|
crackerjack/cli/utils.py,sha256=XC7dT8GNidhORjUe2p2hQOpZgCi2KvVCNu6g3azzgqY,584
|
|
39
39
|
crackerjack/config/__init__.py,sha256=b0481N2f_JvGufMPcbo5IXu2VjYd111r1BHw0oD3x7o,330
|
|
40
40
|
crackerjack/config/global_lock_config.py,sha256=PyonoA2__HKEEvn6SD-gEYh9iD7xIAdts2C3lNoTWhw,1996
|
|
@@ -54,7 +54,7 @@ crackerjack/core/service_watchdog.py,sha256=Ttj1imOxvUea4Tkf5JO1e2dQtGIK7D-bX1xO
|
|
|
54
54
|
crackerjack/core/session_coordinator.py,sha256=TgoGE9DfXe2x-OkH93Ld9dX9ROjx2_mZFkGXen-z5YI,15680
|
|
55
55
|
crackerjack/core/timeout_manager.py,sha256=_sbEsfYDwWx7y0Pn89QCoAZ5DpWIbCdtR9qkG_Kqj5E,15013
|
|
56
56
|
crackerjack/core/websocket_lifecycle.py,sha256=74kn6ugu6FLlDQhCNSPgqguCFwRoT1WFOvtl8G2OyFc,12860
|
|
57
|
-
crackerjack/core/workflow_orchestrator.py,sha256=
|
|
57
|
+
crackerjack/core/workflow_orchestrator.py,sha256=eyK_mtfwIDHxkcHSM7wc_Q21Xtb1wQ71aagwKeIyWEs,78517
|
|
58
58
|
crackerjack/docs/INDEX.md,sha256=a6CGFEeL5DX_FRft_JFWd0nOxoBmCSSp-QHIC3B7ato,342
|
|
59
59
|
crackerjack/docs/generated/api/API_REFERENCE.md,sha256=mWoqImZA7AhDvRqqF1MhUo70g_pnZr3NoBeZQRotqN8,155816
|
|
60
60
|
crackerjack/docs/generated/api/CLI_REFERENCE.md,sha256=ikuG0hO5EjIiQlJtAUnvEuAhXDa-JHPULPXNNmUwvk4,2805
|
|
@@ -84,10 +84,10 @@ crackerjack/intelligence/integration.py,sha256=vVaC2Fp5RbbABpaohCePzGw1XANuRztGl
|
|
|
84
84
|
crackerjack/managers/__init__.py,sha256=PFWccXx4hDQA76T02idAViOLVD-aPeVpgjdfSkh_Dmk,298
|
|
85
85
|
crackerjack/managers/async_hook_manager.py,sha256=c0HFR98sFwfk0uZ3NmAe_6OVZpBrq9I570V8A2DoIxw,5129
|
|
86
86
|
crackerjack/managers/hook_manager.py,sha256=_FT0ngwPwujqg0KZGpLz-pP07mwDmptJ5pVkiy5yS8k,7820
|
|
87
|
-
crackerjack/managers/publish_manager.py,sha256=
|
|
87
|
+
crackerjack/managers/publish_manager.py,sha256=E0jqHuscfn89pI2Ely0R6xyi3EGTvYFQeFTWyRmdjBM,22067
|
|
88
88
|
crackerjack/managers/test_command_builder.py,sha256=1TlPzddNcDDxRORH6UvAudcbRc6hKwFyknSEVLkiWAo,3459
|
|
89
89
|
crackerjack/managers/test_executor.py,sha256=2837Ti4OaNsmLxnmELjbQ18hmfL0-Z2EW-W2UeFSDcE,13871
|
|
90
|
-
crackerjack/managers/test_manager.py,sha256=
|
|
90
|
+
crackerjack/managers/test_manager.py,sha256=ClApfL9e3K1qe3vVcVbC8Y84CMJO55mwPegXiaTw9g8,18638
|
|
91
91
|
crackerjack/managers/test_manager_backup.py,sha256=CR8D7WZ68ZrTADFqYJtVDUWnlznJXJNriPIdsp6ZB1E,37932
|
|
92
92
|
crackerjack/managers/test_progress.py,sha256=B1013ygUk2nAo37whDXNA7n-FYdsEO4qj17fuDm_fdg,3058
|
|
93
93
|
crackerjack/mcp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -96,13 +96,13 @@ crackerjack/mcp/client_runner.py,sha256=Mrcsw7oTk1JgIno7OuNlVVM28UtmGjQYItpToTVO
|
|
|
96
96
|
crackerjack/mcp/context.py,sha256=T1rhScFxbqi0Kaf6o8NusVUIwc8o7YGxcsLdmlWVsPI,28350
|
|
97
97
|
crackerjack/mcp/dashboard.py,sha256=eApQekZe1DfKBrlDW5yobyqaN_gdh17ZQlqGmut5TCs,19602
|
|
98
98
|
crackerjack/mcp/enhanced_progress_monitor.py,sha256=xhACaEjreDw1cLObJmhr5ttowl5mRqeohmMzQfRQj3w,16448
|
|
99
|
-
crackerjack/mcp/file_monitor.py,sha256=
|
|
99
|
+
crackerjack/mcp/file_monitor.py,sha256=PVwbWtk7_cUyc9QLUHgKWtoHs88qvO5Fu-8S0BdcbxA,13203
|
|
100
100
|
crackerjack/mcp/progress_components.py,sha256=URtnrNv1QQzUKY3msO_cBrYHjPlzVDqFbZa_trfZHXY,20535
|
|
101
101
|
crackerjack/mcp/progress_monitor.py,sha256=3TxQ_Og1md8zuNrH9cltOa8RQf_5lHNiVh9DiBkvbBM,37546
|
|
102
|
-
crackerjack/mcp/rate_limiter.py,sha256=
|
|
102
|
+
crackerjack/mcp/rate_limiter.py,sha256=LDefC3CLawIhpyFfFXPmhR-8yZCAkQYbV2sll9JrnD8,12406
|
|
103
103
|
crackerjack/mcp/server.py,sha256=-AdHHoFNAtoPM1A3_M2giZbIbggUhhvv16uIXk2xqNo,403
|
|
104
104
|
crackerjack/mcp/server_core.py,sha256=o93_j8nmGRIl-FL-Y5cKCDDJh3J7vTi1SmxIOdoqj0w,11301
|
|
105
|
-
crackerjack/mcp/service_watchdog.py,sha256=
|
|
105
|
+
crackerjack/mcp/service_watchdog.py,sha256=B4hTrdgRZTf9wDHQWplyKio6AxePvpR8gmid511cBAY,16938
|
|
106
106
|
crackerjack/mcp/state.py,sha256=pNclwUxGTH5YiRMq8hB15B9_rpX9aJ-IVnhMw7q9dtI,14427
|
|
107
107
|
crackerjack/mcp/task_manager.py,sha256=hxuZGbRHwEBO-W3UTbg7Lx7gMhTcPRaAiQ8I8VIFs6g,8713
|
|
108
108
|
crackerjack/mcp/websocket_server.py,sha256=Dsw_orJOSeO7t7zRiVErh8GLp5jAaGab6shOeQTynbw,203
|
|
@@ -112,9 +112,9 @@ crackerjack/mcp/tools/error_analyzer.py,sha256=8ap3_TIqEfrN0n49w_wJFRE3zu7eHB9E8
|
|
|
112
112
|
crackerjack/mcp/tools/execution_tools.py,sha256=iMtcTm45v04wNgfiIdzOjMXyB-lZ9x-yUtN9SuLGxNI,10410
|
|
113
113
|
crackerjack/mcp/tools/execution_tools_backup.py,sha256=sWOtid00SkuonFLT2XCtljOGZiibbaIXzEdLWsPrXyI,33351
|
|
114
114
|
crackerjack/mcp/tools/intelligence_tool_registry.py,sha256=eAOBzW3EZTKSz_MeYVzgjD6a1xVZFSXUdgOsgcQLSBk,1262
|
|
115
|
-
crackerjack/mcp/tools/intelligence_tools.py,sha256=
|
|
115
|
+
crackerjack/mcp/tools/intelligence_tools.py,sha256=HRjX6i3wKKniUbffYZFjRP2PdxsTmDwi4D0SAlsyQDs,9124
|
|
116
116
|
crackerjack/mcp/tools/monitoring_tools.py,sha256=5RYbytBmLjJdz1sE33xYchdrO9LgzGwSvsqRsZIV7oA,21804
|
|
117
|
-
crackerjack/mcp/tools/proactive_tools.py,sha256=
|
|
117
|
+
crackerjack/mcp/tools/proactive_tools.py,sha256=EKTCNS8285p6f3p3bd3AdEGN6OWMvAWn8ArTWt_pgSc,12972
|
|
118
118
|
crackerjack/mcp/tools/progress_tools.py,sha256=S4IkENegGaH06EYNotJ3m2Zuky2OMk-iSd3rLs9T9vs,6930
|
|
119
119
|
crackerjack/mcp/tools/utility_tools.py,sha256=zt4tES5uhyvZfMdgtStlqItWCBA4MWqGCypQ5u_HPds,10367
|
|
120
120
|
crackerjack/mcp/tools/workflow_executor.py,sha256=IvHPyvak1UP1MsWgrYScoFvIe52gL7cdPlec0mAHkFc,19628
|
|
@@ -141,12 +141,12 @@ crackerjack/monitoring/websocket_server.py,sha256=tmCYh5Qp58iicS_txj5OTdmt3I85M5
|
|
|
141
141
|
crackerjack/orchestration/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
142
142
|
crackerjack/orchestration/advanced_orchestrator.py,sha256=RFvyJuUrL2-Vrii4XPX1NNtF-tcBWbbnyokqSe6sxmQ,34000
|
|
143
143
|
crackerjack/orchestration/coverage_improvement.py,sha256=1G88wEvCCL3Gxa1O6jg2U_F3suwJqpu-KKbNJf1e7TI,6745
|
|
144
|
-
crackerjack/orchestration/execution_strategies.py,sha256=
|
|
144
|
+
crackerjack/orchestration/execution_strategies.py,sha256=M3hplgaCULMwE9TiqP9rWvivnrqdSR6DJTzY_rvueww,12482
|
|
145
145
|
crackerjack/orchestration/test_progress_streamer.py,sha256=Yu6uHuhoCvX6SZP0QNG3Yt8Q4s2tufEHr40o16QU98c,22541
|
|
146
146
|
crackerjack/plugins/__init__.py,sha256=B7hy9b9amJVbYLHgIz8kgTI29j-vYxsUY_sZ5ISbXU0,386
|
|
147
147
|
crackerjack/plugins/base.py,sha256=VFk-xNsgjSlmzJ_iPQALhkr7cguiOtEd3XSR9CcCPkc,5732
|
|
148
|
-
crackerjack/plugins/hooks.py,sha256=
|
|
149
|
-
crackerjack/plugins/loader.py,sha256=
|
|
148
|
+
crackerjack/plugins/hooks.py,sha256=Much8c_KNbvYMqcBSke90pRWuga5BZvefmGY8GK3j44,7629
|
|
149
|
+
crackerjack/plugins/loader.py,sha256=uj6Qm0mOnjnRcALwi9mrqwSyw0Zv2mkdWdHD_5GJMdw,10826
|
|
150
150
|
crackerjack/plugins/managers.py,sha256=3kQlxjvcHyHDgZIdr-JZBO1kqz2asqA4kf2XVAA1K6A,8824
|
|
151
151
|
crackerjack/security/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
152
152
|
crackerjack/security/audit.py,sha256=NZJRREvjeKJMhusNZkL8GnBf4XiQFnTwKGcu6NqmvgI,6728
|
|
@@ -165,13 +165,13 @@ crackerjack/services/contextual_ai_assistant.py,sha256=6Pnb2r824c4JYkP5mtCH8sJ2O
|
|
|
165
165
|
crackerjack/services/coverage_badge_service.py,sha256=gzC3LEyVLt4rpf378APsFVRGgNwkLc50w9S_9xEPPoM,6645
|
|
166
166
|
crackerjack/services/coverage_ratchet.py,sha256=eKxmFyg-7Rnctnk--6P-yNNOFhoKwzTxd4iCJ52dtZE,13439
|
|
167
167
|
crackerjack/services/debug.py,sha256=kTy2vy6JjjNReNn4z7xmhXDVIj8xImaE67kYi-Pfb3o,23227
|
|
168
|
-
crackerjack/services/dependency_analyzer.py,sha256=
|
|
168
|
+
crackerjack/services/dependency_analyzer.py,sha256=28u5zfV_2foVKmjfVbr9ATbxIPq7CPeyOiZJJNw9_ZU,16304
|
|
169
169
|
crackerjack/services/dependency_monitor.py,sha256=x_PpRUt-LbzHlTyoSt2kiXTQn2fqAiqouiwFSAU9SGg,20560
|
|
170
170
|
crackerjack/services/documentation_generator.py,sha256=xDJ2dv3Q_HTGCPJADKPVtYzcIpSOhuH5qUxPU2ttICE,16482
|
|
171
171
|
crackerjack/services/documentation_service.py,sha256=V2fY01u_tP_ep87KMGxIocgP6hfpwbf-YyVqssLQSns,24792
|
|
172
172
|
crackerjack/services/enhanced_filesystem.py,sha256=IT-obfj2U2Sfy0iJOq0ZAiczEN6LdHf6VAhnugQe1mQ,15487
|
|
173
173
|
crackerjack/services/enterprise_optimizer.py,sha256=q6srIGxe18N9zH-MNhZ9R34sWnr_-ybGqOWnKyI17Zk,33798
|
|
174
|
-
crackerjack/services/error_pattern_analyzer.py,sha256=
|
|
174
|
+
crackerjack/services/error_pattern_analyzer.py,sha256=YKFQi_nXx5dq-Wvi6AngKEaA-Bv5ggTQ2B2ML0AtF0c,24415
|
|
175
175
|
crackerjack/services/file_hasher.py,sha256=eReytwwK-_-B8JBnpwytDC52cKKgg4qpaxaZKcQjD-0,5211
|
|
176
176
|
crackerjack/services/filesystem.py,sha256=nmL3mYqylS_BSQpwFbC7EMHoA44K5qUxa9CPg1QFZvc,17480
|
|
177
177
|
crackerjack/services/git.py,sha256=g0D9K7PFGWiv_-CbPXEhGtxJVJccEsshYNEmGstjXss,12716
|
|
@@ -185,16 +185,16 @@ crackerjack/services/logging.py,sha256=aHHrK7Hnke-pqulyJHIoqfymHVyPc5Fxp0WQX_cvb
|
|
|
185
185
|
crackerjack/services/lsp_client.py,sha256=6XPnNbKwDWaU5DEJWBCU7w1gii4glkTYgURulzYtWlY,22299
|
|
186
186
|
crackerjack/services/memory_optimizer.py,sha256=GBRkzWNR1O7lkPV1EmGKMqtaJbRWOKavYYxsdTmKxSg,13328
|
|
187
187
|
crackerjack/services/metrics.py,sha256=0_tiaK2-lBY0Y5MrS3Rrzyxft_JziGEEPgUXW486rX0,23122
|
|
188
|
-
crackerjack/services/parallel_executor.py,sha256=
|
|
188
|
+
crackerjack/services/parallel_executor.py,sha256=cizgEnEeORp0SfBlkfhlLJyWMb1I87KUgR8ulB_5d90,13654
|
|
189
189
|
crackerjack/services/pattern_cache.py,sha256=fdD-HPoImUnd07rlGuDfyFWG86teWpNVr2LacJkCZ5Q,11097
|
|
190
190
|
crackerjack/services/pattern_detector.py,sha256=soB9zXsjsNFo5pzMpFhFehNOemPTCQh-ARdKlHi46U0,18032
|
|
191
|
-
crackerjack/services/performance_benchmarks.py,sha256=
|
|
192
|
-
crackerjack/services/performance_cache.py,sha256=
|
|
191
|
+
crackerjack/services/performance_benchmarks.py,sha256=ssgPyDSl_l_VT-OYGkm55R8GmoQprW0tWCOE7DQQr_I,11383
|
|
192
|
+
crackerjack/services/performance_cache.py,sha256=Q7ziM-6kbbF3O6J89svVdm8fH-36Q3qQFo0jgR7BKhQ,11830
|
|
193
193
|
crackerjack/services/performance_monitor.py,sha256=cDu6FoiUX2VpslA3vG4JL6YQyIk1JzDcD4g5zR8NdjI,20214
|
|
194
194
|
crackerjack/services/predictive_analytics.py,sha256=OcZYP0p2_zZuyWqbqzLHDoHTxVfrPm3VtiY-Yq9y7xc,17843
|
|
195
195
|
crackerjack/services/quality_baseline.py,sha256=ImC1XTFTfrwFIDDDX1F0V7sGquAXI21oVxgjOSYIYFQ,8229
|
|
196
196
|
crackerjack/services/quality_baseline_enhanced.py,sha256=mBxzXD4lkegPc_q_KNJyorBk0cezud40lOWYDBWeNGU,24064
|
|
197
|
-
crackerjack/services/quality_intelligence.py,sha256=
|
|
197
|
+
crackerjack/services/quality_intelligence.py,sha256=xXbfLODo4Ymgv0UsXivGJSoRjhC91gnNR9KcjWYOl8c,28393
|
|
198
198
|
crackerjack/services/regex_patterns.py,sha256=iis9gSzXZtnX14lODGfSUsf7bcCRTw7rdSgog5DKFNk,117341
|
|
199
199
|
crackerjack/services/regex_utils.py,sha256=e7AD59_L-T5-oOxzqsGgrLd94uxRE9aKnwasZkohwI8,14966
|
|
200
200
|
crackerjack/services/secure_path_utils.py,sha256=aHsLwxDch42DidPYtTL_ko40g2rhbXDLnRhcx2LlGk8,16688
|
|
@@ -224,8 +224,8 @@ crackerjack/tools/validate_input_validator_patterns.py,sha256=NN7smYlXWrHLQXTb-8
|
|
|
224
224
|
crackerjack/tools/validate_regex_patterns.py,sha256=J7GG9EP1fASpRIsG8qRPeiCSkdCwmk0sdo29GgoJ6w8,5863
|
|
225
225
|
crackerjack/ui/__init__.py,sha256=eMb1OeTU-dSLICAACn0YdYB4Amdr8wHckjKfn0wOIZE,37
|
|
226
226
|
crackerjack/ui/server_panels.py,sha256=F5IH6SNN06BaZQMsFx_D-OA286aojmaFPJ5kvvSRv_c,4232
|
|
227
|
-
crackerjack-0.38.
|
|
228
|
-
crackerjack-0.38.
|
|
229
|
-
crackerjack-0.38.
|
|
230
|
-
crackerjack-0.38.
|
|
231
|
-
crackerjack-0.38.
|
|
227
|
+
crackerjack-0.38.2.dist-info/METADATA,sha256=jK0SSBuvUD_i8vsriaVZ0mA9YCnU_--kqIRUInhQArI,38082
|
|
228
|
+
crackerjack-0.38.2.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
|
229
|
+
crackerjack-0.38.2.dist-info/entry_points.txt,sha256=AJKNft0WXm9xoGUJ3Trl-iXHOWxRAYbagQiza3AILr4,57
|
|
230
|
+
crackerjack-0.38.2.dist-info/licenses/LICENSE,sha256=fDt371P6_6sCu7RyqiZH_AhT1LdN3sN1zjBtqEhDYCk,1531
|
|
231
|
+
crackerjack-0.38.2.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|