@accelerationguy/accel 1.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CLAUDE.md +19 -0
- package/LICENSE +33 -0
- package/README.md +275 -0
- package/bin/install.js +661 -0
- package/docs/getting-started.md +164 -0
- package/docs/module-guide.md +139 -0
- package/modules/drive/LICENSE +21 -0
- package/modules/drive/PAUL-VS-GSD.md +171 -0
- package/modules/drive/README.md +555 -0
- package/modules/drive/assets/terminal.svg +67 -0
- package/modules/drive/bin/install.js +210 -0
- package/modules/drive/integration.js +76 -0
- package/modules/drive/package.json +38 -0
- package/modules/drive/src/commands/add-phase.md +36 -0
- package/modules/drive/src/commands/apply.md +83 -0
- package/modules/drive/src/commands/assumptions.md +37 -0
- package/modules/drive/src/commands/audit.md +57 -0
- package/modules/drive/src/commands/complete-milestone.md +36 -0
- package/modules/drive/src/commands/config.md +175 -0
- package/modules/drive/src/commands/consider-issues.md +41 -0
- package/modules/drive/src/commands/discover.md +48 -0
- package/modules/drive/src/commands/discuss-milestone.md +33 -0
- package/modules/drive/src/commands/discuss.md +34 -0
- package/modules/drive/src/commands/flows.md +73 -0
- package/modules/drive/src/commands/handoff.md +201 -0
- package/modules/drive/src/commands/help.md +525 -0
- package/modules/drive/src/commands/init.md +54 -0
- package/modules/drive/src/commands/map-codebase.md +34 -0
- package/modules/drive/src/commands/milestone.md +34 -0
- package/modules/drive/src/commands/pause.md +44 -0
- package/modules/drive/src/commands/plan-fix.md +216 -0
- package/modules/drive/src/commands/plan.md +36 -0
- package/modules/drive/src/commands/progress.md +138 -0
- package/modules/drive/src/commands/register.md +29 -0
- package/modules/drive/src/commands/remove-phase.md +37 -0
- package/modules/drive/src/commands/research-phase.md +209 -0
- package/modules/drive/src/commands/research.md +47 -0
- package/modules/drive/src/commands/resume.md +49 -0
- package/modules/drive/src/commands/status.md +78 -0
- package/modules/drive/src/commands/unify.md +87 -0
- package/modules/drive/src/commands/verify.md +60 -0
- package/modules/drive/src/references/checkpoints.md +234 -0
- package/modules/drive/src/references/context-management.md +219 -0
- package/modules/drive/src/references/git-strategy.md +206 -0
- package/modules/drive/src/references/loop-phases.md +254 -0
- package/modules/drive/src/references/plan-format.md +263 -0
- package/modules/drive/src/references/quality-principles.md +152 -0
- package/modules/drive/src/references/research-quality-control.md +247 -0
- package/modules/drive/src/references/sonarqube-integration.md +244 -0
- package/modules/drive/src/references/specialized-workflow-integration.md +186 -0
- package/modules/drive/src/references/subagent-criteria.md +179 -0
- package/modules/drive/src/references/tdd.md +219 -0
- package/modules/drive/src/references/work-units.md +161 -0
- package/modules/drive/src/rules/commands.md +108 -0
- package/modules/drive/src/rules/references.md +107 -0
- package/modules/drive/src/rules/style.md +123 -0
- package/modules/drive/src/rules/templates.md +51 -0
- package/modules/drive/src/rules/workflows.md +133 -0
- package/modules/drive/src/templates/CONTEXT.md +88 -0
- package/modules/drive/src/templates/DEBUG.md +164 -0
- package/modules/drive/src/templates/DISCOVERY.md +148 -0
- package/modules/drive/src/templates/HANDOFF.md +77 -0
- package/modules/drive/src/templates/ISSUES.md +93 -0
- package/modules/drive/src/templates/MILESTONES.md +167 -0
- package/modules/drive/src/templates/PLAN.md +328 -0
- package/modules/drive/src/templates/PROJECT.md +219 -0
- package/modules/drive/src/templates/RESEARCH.md +130 -0
- package/modules/drive/src/templates/ROADMAP.md +328 -0
- package/modules/drive/src/templates/SPECIAL-FLOWS.md +70 -0
- package/modules/drive/src/templates/STATE.md +210 -0
- package/modules/drive/src/templates/SUMMARY.md +221 -0
- package/modules/drive/src/templates/UAT-ISSUES.md +139 -0
- package/modules/drive/src/templates/codebase/architecture.md +259 -0
- package/modules/drive/src/templates/codebase/concerns.md +329 -0
- package/modules/drive/src/templates/codebase/conventions.md +311 -0
- package/modules/drive/src/templates/codebase/integrations.md +284 -0
- package/modules/drive/src/templates/codebase/stack.md +190 -0
- package/modules/drive/src/templates/codebase/structure.md +287 -0
- package/modules/drive/src/templates/codebase/testing.md +484 -0
- package/modules/drive/src/templates/config.md +181 -0
- package/modules/drive/src/templates/milestone-archive.md +236 -0
- package/modules/drive/src/templates/milestone-context.md +190 -0
- package/modules/drive/src/templates/paul-json.md +147 -0
- package/modules/drive/src/vector-config/PAUL +26 -0
- package/modules/drive/src/vector-config/PAUL.manifest +11 -0
- package/modules/drive/src/workflows/apply-phase.md +393 -0
- package/modules/drive/src/workflows/audit-plan.md +344 -0
- package/modules/drive/src/workflows/complete-milestone.md +479 -0
- package/modules/drive/src/workflows/configure-special-flows.md +283 -0
- package/modules/drive/src/workflows/consider-issues.md +172 -0
- package/modules/drive/src/workflows/create-milestone.md +268 -0
- package/modules/drive/src/workflows/debug.md +292 -0
- package/modules/drive/src/workflows/discovery.md +187 -0
- package/modules/drive/src/workflows/discuss-milestone.md +245 -0
- package/modules/drive/src/workflows/discuss-phase.md +231 -0
- package/modules/drive/src/workflows/init-project.md +698 -0
- package/modules/drive/src/workflows/map-codebase.md +459 -0
- package/modules/drive/src/workflows/pause-work.md +259 -0
- package/modules/drive/src/workflows/phase-assumptions.md +181 -0
- package/modules/drive/src/workflows/plan-phase.md +385 -0
- package/modules/drive/src/workflows/quality-gate.md +263 -0
- package/modules/drive/src/workflows/register-manifest.md +107 -0
- package/modules/drive/src/workflows/research.md +241 -0
- package/modules/drive/src/workflows/resume-project.md +200 -0
- package/modules/drive/src/workflows/roadmap-management.md +334 -0
- package/modules/drive/src/workflows/transition-phase.md +368 -0
- package/modules/drive/src/workflows/unify-phase.md +290 -0
- package/modules/drive/src/workflows/verify-work.md +241 -0
- package/modules/forge/README.md +281 -0
- package/modules/forge/bin/install.js +200 -0
- package/modules/forge/package.json +32 -0
- package/modules/forge/skillsmith/rules/checklists-rules.md +42 -0
- package/modules/forge/skillsmith/rules/context-rules.md +43 -0
- package/modules/forge/skillsmith/rules/entry-point-rules.md +44 -0
- package/modules/forge/skillsmith/rules/frameworks-rules.md +43 -0
- package/modules/forge/skillsmith/rules/tasks-rules.md +52 -0
- package/modules/forge/skillsmith/rules/templates-rules.md +43 -0
- package/modules/forge/skillsmith/skillsmith.md +82 -0
- package/modules/forge/skillsmith/tasks/audit.md +277 -0
- package/modules/forge/skillsmith/tasks/discover.md +145 -0
- package/modules/forge/skillsmith/tasks/distill.md +276 -0
- package/modules/forge/skillsmith/tasks/scaffold.md +349 -0
- package/modules/forge/specs/checklists.md +193 -0
- package/modules/forge/specs/context.md +223 -0
- package/modules/forge/specs/entry-point.md +320 -0
- package/modules/forge/specs/frameworks.md +228 -0
- package/modules/forge/specs/rules.md +245 -0
- package/modules/forge/specs/tasks.md +344 -0
- package/modules/forge/specs/templates.md +335 -0
- package/modules/forge/terminal.svg +70 -0
- package/modules/ignition/README.md +245 -0
- package/modules/ignition/bin/install.js +184 -0
- package/modules/ignition/checklists/planning-quality.md +55 -0
- package/modules/ignition/data/application/config.md +21 -0
- package/modules/ignition/data/application/guide.md +51 -0
- package/modules/ignition/data/application/skill-loadout.md +11 -0
- package/modules/ignition/data/campaign/config.md +18 -0
- package/modules/ignition/data/campaign/guide.md +36 -0
- package/modules/ignition/data/campaign/skill-loadout.md +10 -0
- package/modules/ignition/data/client/config.md +18 -0
- package/modules/ignition/data/client/guide.md +36 -0
- package/modules/ignition/data/client/skill-loadout.md +11 -0
- package/modules/ignition/data/utility/config.md +18 -0
- package/modules/ignition/data/utility/guide.md +31 -0
- package/modules/ignition/data/utility/skill-loadout.md +8 -0
- package/modules/ignition/data/workflow/config.md +19 -0
- package/modules/ignition/data/workflow/guide.md +41 -0
- package/modules/ignition/data/workflow/skill-loadout.md +10 -0
- package/modules/ignition/integration.js +54 -0
- package/modules/ignition/package.json +35 -0
- package/modules/ignition/seed.md +81 -0
- package/modules/ignition/tasks/add-type.md +164 -0
- package/modules/ignition/tasks/graduate.md +182 -0
- package/modules/ignition/tasks/ideate.md +221 -0
- package/modules/ignition/tasks/launch.md +137 -0
- package/modules/ignition/tasks/status.md +71 -0
- package/modules/ignition/templates/planning-application.md +193 -0
- package/modules/ignition/templates/planning-campaign.md +138 -0
- package/modules/ignition/templates/planning-client.md +149 -0
- package/modules/ignition/templates/planning-utility.md +112 -0
- package/modules/ignition/templates/planning-workflow.md +125 -0
- package/modules/ignition/terminal.svg +74 -0
- package/modules/mission-control/CONTEXT-CONTINUITY-SPEC.md +293 -0
- package/modules/mission-control/CONTEXT-ENGINEERING-GUIDE.md +282 -0
- package/modules/mission-control/README.md +91 -0
- package/modules/mission-control/assets/terminal.svg +80 -0
- package/modules/mission-control/examples/entities.example.json +133 -0
- package/modules/mission-control/examples/projects.example.json +318 -0
- package/modules/mission-control/examples/state.example.json +183 -0
- package/modules/mission-control/examples/vector.example.json +245 -0
- package/modules/mission-control/mission-control/checklists/install-verification.md +46 -0
- package/modules/mission-control/mission-control/frameworks/framework-registry.md +83 -0
- package/modules/mission-control/mission-control/mission-control.md +83 -0
- package/modules/mission-control/mission-control/tasks/insights.md +73 -0
- package/modules/mission-control/mission-control/tasks/install.md +194 -0
- package/modules/mission-control/mission-control/tasks/status.md +125 -0
- package/modules/mission-control/schemas/entities.schema.json +89 -0
- package/modules/mission-control/schemas/projects.schema.json +221 -0
- package/modules/mission-control/schemas/state.schema.json +108 -0
- package/modules/mission-control/schemas/vector.schema.json +200 -0
- package/modules/momentum/README.md +678 -0
- package/modules/momentum/bin/install.js +563 -0
- package/modules/momentum/integration.js +131 -0
- package/modules/momentum/package.json +42 -0
- package/modules/momentum/schemas/entities.schema.json +89 -0
- package/modules/momentum/schemas/projects.schema.json +221 -0
- package/modules/momentum/schemas/state.schema.json +108 -0
- package/modules/momentum/src/commands/audit-claude-md.md +31 -0
- package/modules/momentum/src/commands/audit.md +33 -0
- package/modules/momentum/src/commands/groom.md +35 -0
- package/modules/momentum/src/commands/history.md +27 -0
- package/modules/momentum/src/commands/pulse.md +33 -0
- package/modules/momentum/src/commands/scaffold.md +33 -0
- package/modules/momentum/src/commands/status.md +28 -0
- package/modules/momentum/src/commands/surface-convert.md +35 -0
- package/modules/momentum/src/commands/surface-create.md +34 -0
- package/modules/momentum/src/commands/surface-list.md +27 -0
- package/modules/momentum/src/commands/vector-hygiene.md +33 -0
- package/modules/momentum/src/framework/context/momentum-principles.md +71 -0
- package/modules/momentum/src/framework/frameworks/audit-strategies.md +53 -0
- package/modules/momentum/src/framework/frameworks/satellite-registration.md +44 -0
- package/modules/momentum/src/framework/tasks/audit-claude-md.md +68 -0
- package/modules/momentum/src/framework/tasks/audit.md +64 -0
- package/modules/momentum/src/framework/tasks/groom.md +164 -0
- package/modules/momentum/src/framework/tasks/history.md +34 -0
- package/modules/momentum/src/framework/tasks/pulse.md +83 -0
- package/modules/momentum/src/framework/tasks/scaffold.md +202 -0
- package/modules/momentum/src/framework/tasks/status.md +35 -0
- package/modules/momentum/src/framework/tasks/surface-convert.md +143 -0
- package/modules/momentum/src/framework/tasks/surface-create.md +184 -0
- package/modules/momentum/src/framework/tasks/surface-list.md +42 -0
- package/modules/momentum/src/framework/tasks/vector-hygiene.md +160 -0
- package/modules/momentum/src/framework/templates/workspace-json.md +96 -0
- package/modules/momentum/src/hooks/_template.py +129 -0
- package/modules/momentum/src/hooks/active-hook.py +178 -0
- package/modules/momentum/src/hooks/backlog-hook.py +115 -0
- package/modules/momentum/src/hooks/mission-control-insights.py +169 -0
- package/modules/momentum/src/hooks/momentum-pulse-check.py +351 -0
- package/modules/momentum/src/hooks/operator.py +53 -0
- package/modules/momentum/src/hooks/psmm-injector.py +67 -0
- package/modules/momentum/src/hooks/satellite-detection.py +248 -0
- package/modules/momentum/src/packages/momentum-mcp/index.js +119 -0
- package/modules/momentum/src/packages/momentum-mcp/package.json +10 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/entities.js +226 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/operator.js +106 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/projects.js +322 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/psmm.js +206 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/state.js +199 -0
- package/modules/momentum/src/packages/momentum-mcp/tools/surfaces.js +404 -0
- package/modules/momentum/src/skill/momentum.md +111 -0
- package/modules/momentum/src/tasks/groom.md +164 -0
- package/modules/momentum/src/templates/operator.json +66 -0
- package/modules/momentum/src/templates/workspace.json +111 -0
- package/modules/momentum/terminal.svg +77 -0
- package/modules/radar/README.md +1552 -0
- package/modules/radar/commands/audit.md +233 -0
- package/modules/radar/commands/guardrails.md +194 -0
- package/modules/radar/commands/init.md +207 -0
- package/modules/radar/commands/playbook.md +176 -0
- package/modules/radar/commands/remediate.md +156 -0
- package/modules/radar/commands/report.md +172 -0
- package/modules/radar/commands/resume.md +176 -0
- package/modules/radar/commands/status.md +148 -0
- package/modules/radar/commands/transform.md +205 -0
- package/modules/radar/commands/validate.md +177 -0
- package/modules/radar/docs/ARCHITECTURE.md +336 -0
- package/modules/radar/docs/GETTING-STARTED.md +287 -0
- package/modules/radar/docs/standards/agents.md +197 -0
- package/modules/radar/docs/standards/commands.md +250 -0
- package/modules/radar/docs/standards/domains.md +191 -0
- package/modules/radar/docs/standards/personas.md +211 -0
- package/modules/radar/docs/standards/rules.md +218 -0
- package/modules/radar/docs/standards/runtime.md +445 -0
- package/modules/radar/docs/standards/schemas.md +269 -0
- package/modules/radar/docs/standards/tools.md +273 -0
- package/modules/radar/docs/standards/workflows.md +254 -0
- package/modules/radar/docs/terminal.svg +72 -0
- package/modules/radar/docs/validation/convention-compliance-report.md +183 -0
- package/modules/radar/docs/validation/cross-reference-report.md +195 -0
- package/modules/radar/docs/validation/validation-summary.md +118 -0
- package/modules/radar/docs/validation/version-manifest.yaml +363 -0
- package/modules/radar/install.sh +711 -0
- package/modules/radar/integration.js +53 -0
- package/modules/radar/src/core/agents/architect.md +25 -0
- package/modules/radar/src/core/agents/compliance-officer.md +25 -0
- package/modules/radar/src/core/agents/data-engineer.md +25 -0
- package/modules/radar/src/core/agents/devils-advocate.md +22 -0
- package/modules/radar/src/core/agents/performance-engineer.md +25 -0
- package/modules/radar/src/core/agents/principal-engineer.md +23 -0
- package/modules/radar/src/core/agents/reality-gap-analyst.md +22 -0
- package/modules/radar/src/core/agents/security-engineer.md +25 -0
- package/modules/radar/src/core/agents/senior-app-engineer.md +25 -0
- package/modules/radar/src/core/agents/sre.md +25 -0
- package/modules/radar/src/core/agents/staff-engineer.md +23 -0
- package/modules/radar/src/core/agents/test-engineer.md +25 -0
- package/modules/radar/src/core/personas/architect.md +111 -0
- package/modules/radar/src/core/personas/compliance-officer.md +104 -0
- package/modules/radar/src/core/personas/data-engineer.md +113 -0
- package/modules/radar/src/core/personas/devils-advocate.md +105 -0
- package/modules/radar/src/core/personas/performance-engineer.md +119 -0
- package/modules/radar/src/core/personas/principal-engineer.md +119 -0
- package/modules/radar/src/core/personas/reality-gap-analyst.md +111 -0
- package/modules/radar/src/core/personas/security-engineer.md +108 -0
- package/modules/radar/src/core/personas/senior-app-engineer.md +111 -0
- package/modules/radar/src/core/personas/sre.md +117 -0
- package/modules/radar/src/core/personas/staff-engineer.md +109 -0
- package/modules/radar/src/core/personas/test-engineer.md +109 -0
- package/modules/radar/src/core/workflows/disagreement-resolution.md +183 -0
- package/modules/radar/src/core/workflows/phase-0-context.md +148 -0
- package/modules/radar/src/core/workflows/phase-1-reconnaissance.md +169 -0
- package/modules/radar/src/core/workflows/phase-2-domain-audits.md +190 -0
- package/modules/radar/src/core/workflows/phase-3-cross-domain.md +177 -0
- package/modules/radar/src/core/workflows/phase-4-adversarial-review.md +165 -0
- package/modules/radar/src/core/workflows/phase-5-report.md +189 -0
- package/modules/radar/src/core/workflows/phase-checkpoint.md +222 -0
- package/modules/radar/src/core/workflows/session-handoff.md +152 -0
- package/modules/radar/src/domains/00-context.md +201 -0
- package/modules/radar/src/domains/01-architecture.md +248 -0
- package/modules/radar/src/domains/02-data.md +224 -0
- package/modules/radar/src/domains/03-correctness.md +230 -0
- package/modules/radar/src/domains/04-security.md +274 -0
- package/modules/radar/src/domains/05-compliance.md +228 -0
- package/modules/radar/src/domains/06-testing.md +228 -0
- package/modules/radar/src/domains/07-reliability.md +246 -0
- package/modules/radar/src/domains/08-performance.md +247 -0
- package/modules/radar/src/domains/09-maintainability.md +271 -0
- package/modules/radar/src/domains/10-operability.md +250 -0
- package/modules/radar/src/domains/11-change-risk.md +246 -0
- package/modules/radar/src/domains/12-team-risk.md +221 -0
- package/modules/radar/src/domains/13-risk-synthesis.md +202 -0
- package/modules/radar/src/rules/agent-boundaries.md +78 -0
- package/modules/radar/src/rules/disagreement-protocol.md +76 -0
- package/modules/radar/src/rules/epistemic-hygiene.md +78 -0
- package/modules/radar/src/schemas/confidence.md +185 -0
- package/modules/radar/src/schemas/disagreement.md +238 -0
- package/modules/radar/src/schemas/finding.md +287 -0
- package/modules/radar/src/schemas/report-section.md +150 -0
- package/modules/radar/src/schemas/signal.md +108 -0
- package/modules/radar/src/tools/checkov.md +463 -0
- package/modules/radar/src/tools/git-history.md +581 -0
- package/modules/radar/src/tools/gitleaks.md +447 -0
- package/modules/radar/src/tools/grype.md +611 -0
- package/modules/radar/src/tools/semgrep.md +378 -0
- package/modules/radar/src/tools/sonarqube.md +550 -0
- package/modules/radar/src/tools/syft.md +539 -0
- package/modules/radar/src/tools/trivy.md +439 -0
- package/modules/radar/src/transform/agents/change-risk-modeler.md +24 -0
- package/modules/radar/src/transform/agents/execution-validator.md +24 -0
- package/modules/radar/src/transform/agents/guardrail-generator.md +24 -0
- package/modules/radar/src/transform/agents/pedagogy-agent.md +24 -0
- package/modules/radar/src/transform/agents/remediation-architect.md +24 -0
- package/modules/radar/src/transform/personas/change-risk-modeler.md +95 -0
- package/modules/radar/src/transform/personas/execution-validator.md +95 -0
- package/modules/radar/src/transform/personas/guardrail-generator.md +103 -0
- package/modules/radar/src/transform/personas/pedagogy-agent.md +105 -0
- package/modules/radar/src/transform/personas/remediation-architect.md +95 -0
- package/modules/radar/src/transform/rules/change-risk-rules.md +87 -0
- package/modules/radar/src/transform/rules/safety-governance.md +87 -0
- package/modules/radar/src/transform/schemas/change-risk.md +139 -0
- package/modules/radar/src/transform/schemas/intervention-level.md +207 -0
- package/modules/radar/src/transform/schemas/playbook.md +205 -0
- package/modules/radar/src/transform/schemas/verification-plan.md +134 -0
- package/modules/radar/src/transform/workflows/phase-6-remediation.md +148 -0
- package/modules/radar/src/transform/workflows/phase-7-risk-validation.md +161 -0
- package/modules/radar/src/transform/workflows/phase-8-execution-planning.md +159 -0
- package/modules/radar/src/transform/workflows/transform-safety.md +158 -0
- package/modules/vector/.vector-template/sessions/.gitkeep +0 -0
- package/modules/vector/.vector-template/vector.json +72 -0
- package/modules/vector/AUDIT-CLAUDEMD.md +154 -0
- package/modules/vector/INSTALL.md +185 -0
- package/modules/vector/LICENSE +21 -0
- package/modules/vector/README.md +409 -0
- package/modules/vector/VECTOR-BLOCK.md +57 -0
- package/modules/vector/assets/terminal.svg +68 -0
- package/modules/vector/bin/install.js +455 -0
- package/modules/vector/bin/migrate-v1-to-v2.sh +492 -0
- package/modules/vector/commands/help.md +46 -0
- package/modules/vector/hooks/vector-hook.py +775 -0
- package/modules/vector/mcp/index.js +118 -0
- package/modules/vector/mcp/package.json +10 -0
- package/modules/vector/mcp/tools/decisions.js +269 -0
- package/modules/vector/mcp/tools/domains.js +361 -0
- package/modules/vector/mcp/tools/staging.js +252 -0
- package/modules/vector/mcp/tools/vector-json.js +647 -0
- package/modules/vector/package.json +38 -0
- package/modules/vector/schemas/vector.schema.json +237 -0
- package/package.json +39 -0
- package/shared/branding/branding.js +70 -0
- package/shared/config/defaults.json +59 -0
- package/shared/events/README.md +175 -0
- package/shared/events/event-bus.js +134 -0
- package/shared/events/event_bus.py +255 -0
- package/shared/events/integrations.js +161 -0
- package/shared/events/schemas/audit-complete.schema.json +21 -0
- package/shared/events/schemas/phase-progress.schema.json +23 -0
- package/shared/events/schemas/plan-created.schema.json +21 -0
|
@@ -0,0 +1,1552 @@
|
|
|
1
|
+
# Radar
|
|
2
|
+
|
|
3
|
+
**Multi-Agent Codebase Audit System — Accelerate Module**
|
|
4
|
+
|
|
5
|
+
<p align="center">
|
|
6
|
+
<img src="docs/terminal.svg" alt="Radar terminal demo" width="700"/>
|
|
7
|
+
</p>
|
|
8
|
+
|
|
9
|
+
A multi-session, multi-agent codebase audit system built on [Claude Code](https://claude.ai/code). Radar deploys 12 senior engineering personas across 14 audit domains to produce epistemically rigorous, adversarially reviewed findings on any codebase.
|
|
10
|
+
|
|
11
|
+
---
|
|
12
|
+
|
|
13
|
+
## Quick Install
|
|
14
|
+
|
|
15
|
+
```bash
|
|
16
|
+
curl -sSL https://raw.githubusercontent.com/accelerationguy/accelerate/main/modules/radar/install.sh | bash
|
|
17
|
+
```
|
|
18
|
+
|
|
19
|
+
Then run `/radar:audit` in Claude Code.
|
|
20
|
+
|
|
21
|
+
> **Requires:** [Claude Code](https://claude.ai/code) installed (`~/.claude/` directory must exist)
|
|
22
|
+
|
|
23
|
+
---
|
|
24
|
+
|
|
25
|
+
## Contents
|
|
26
|
+
|
|
27
|
+
- [Identity](#identity)
|
|
28
|
+
- [What Radar Is](#what-radar-is)
|
|
29
|
+
- [Core Philosophy](#core-philosophy)
|
|
30
|
+
- [The 14 Audit Domains](#the-14-audit-domains)
|
|
31
|
+
- [The Agent Team](#the-agent-team)
|
|
32
|
+
- [Execution Phases](#execution-phases)
|
|
33
|
+
- [The Three Output Layers](#the-three-output-layers)
|
|
34
|
+
- [Radar Transform](#radar-transform)
|
|
35
|
+
- [Intervention Levels](#intervention-levels)
|
|
36
|
+
- [The Transformation Model](#the-transformation-model)
|
|
37
|
+
- [Transform Agent Team](#transform-agent-team)
|
|
38
|
+
- [Transform Execution Phases](#transform-execution-phases)
|
|
39
|
+
- [Commands](#commands)
|
|
40
|
+
- [Change Risk in Remediation](#change-risk-in-remediation)
|
|
41
|
+
- [Safety & Liability Framework](#safety--liability-framework)
|
|
42
|
+
- [Drive Integration](#drive-integration)
|
|
43
|
+
- [Pattern Corpus & Feedback Loop](#pattern-corpus--feedback-loop)
|
|
44
|
+
- [The Formal Epistemic Schema](#the-formal-epistemic-schema)
|
|
45
|
+
- [Disagreement Resolution System](#disagreement-resolution-system)
|
|
46
|
+
- [Disagreement Visualization Model](#disagreement-visualization-model)
|
|
47
|
+
- [Reality Gap Framework](#reality-gap-framework)
|
|
48
|
+
- [Language-Specific Failure Models](#language-specific-failure-models)
|
|
49
|
+
- [Tooling Stack](#tooling-stack)
|
|
50
|
+
- [Output Format](#output-format)
|
|
51
|
+
- [Signal Categories](#signal-categories)
|
|
52
|
+
- [Installation & Runtime](#v02--installation--runtime)
|
|
53
|
+
- [Ecosystem](#ecosystem)
|
|
54
|
+
|
|
55
|
+
---
|
|
56
|
+
|
|
57
|
+
## Ecosystem
|
|
58
|
+
|
|
59
|
+
Radar is part of a broader Claude Code extension ecosystem:
|
|
60
|
+
|
|
61
|
+
| System | What It Does | Link |
|
|
62
|
+
|--------|-------------|------|
|
|
63
|
+
| **Radar** | Multi-agent codebase auditing — diagnosis + controlled evolution | You are here |
|
|
64
|
+
| **Momentum** | Builder's Automated State Engine — workspace lifecycle, health tracking, drift prevention | [GitHub](https://github.com/accelerationguy/accelerate) |
|
|
65
|
+
| **Vector** | Context Augmentation & Reinforcement Layer — dynamic rules loaded JIT by intent | [GitHub](https://github.com/accelerationguy/accelerate) |
|
|
66
|
+
| **Drive** | Project orchestration — Plan, Apply, Unify Loop | [GitHub](https://github.com/accelerationguy/accelerate) |
|
|
67
|
+
| **CC Strategic AI** | Skool community — courses, community, live support | [Skool](https://accelerationguy.com) |
|
|
68
|
+
|
|
69
|
+
---
|
|
70
|
+
|
|
71
|
+
## Identity
|
|
72
|
+
|
|
73
|
+
**Radar** — the Accelerate module for comprehensive codebase auditing. Radar scans across 14 domains using 12 specialized agent personas — detecting risks, vulnerabilities, and architectural issues that no single reviewer could catch.
|
|
74
|
+
|
|
75
|
+
Radar is the shield for codebases.
|
|
76
|
+
|
|
77
|
+
The name reflects what the system does — it sweeps the full surface area of a codebase, detecting signals across every domain:
|
|
78
|
+
|
|
79
|
+
- **Multi-agent** — 12 specialized personas, each an expert in a specific engineering domain
|
|
80
|
+
- **Epistemic** — Built on a formal schema for how knowledge is structured, challenged, and trusted under uncertainty
|
|
81
|
+
- **Governance** — Compliance, oversight, and the Principal Engineer's role as epistemic governor of the entire audit
|
|
82
|
+
- **Intelligence** — AI-powered domain expert agents producing findings no single human could match in breadth
|
|
83
|
+
- **Systematic** — Not a tool, not a script. A coordinated system of systems with defined phases, roles, and accountability
|
|
84
|
+
|
|
85
|
+
---
|
|
86
|
+
|
|
87
|
+
## What Radar Is
|
|
88
|
+
|
|
89
|
+
Radar is a **multi-session, multi-agent codebase audit system** built on Claude Code. It deploys a team of senior engineering personas — each an expert in a specific domain — to conduct a comprehensive analysis of any application codebase.
|
|
90
|
+
|
|
91
|
+
It is not a linter. It is not a static analyzer. It is not a report generator.
|
|
92
|
+
|
|
93
|
+
It is **an AI Principal Engineer** — a machine that performs disciplined doubt.
|
|
94
|
+
|
|
95
|
+
Radar answers five fundamental questions that senior engineers ask when they walk into an unknown codebase:
|
|
96
|
+
|
|
97
|
+
1. Can this system be **trusted**?
|
|
98
|
+
2. Can it survive **change**?
|
|
99
|
+
3. Can it **scale**?
|
|
100
|
+
4. Can it be **operated safely**?
|
|
101
|
+
5. Can new engineers **understand** it?
|
|
102
|
+
|
|
103
|
+
Every audit produces findings across all domains, uses structured epistemic reasoning, cross-validates through adversarial review, and synthesizes into actionable, severity-ranked reports.
|
|
104
|
+
|
|
105
|
+
---
|
|
106
|
+
|
|
107
|
+
## Core Philosophy
|
|
108
|
+
|
|
109
|
+
> Senior engineers don't just find bugs — they find **future failures**.
|
|
110
|
+
|
|
111
|
+
Three principles distinguish Radar from conventional code analysis:
|
|
112
|
+
|
|
113
|
+
1. **Disciplined Doubt Over Coherent Confidence** — Most AI systems optimize for helpfulness and clean narratives. Radar optimizes for correctness under uncertainty, asymmetric risk detection, and institutional memory of failure patterns.
|
|
114
|
+
|
|
115
|
+
2. **The Principal Builds the Story. The Devil's Advocate Breaks It. Truth Lives in the Tension.** — No finding survives without challenge. No conclusion is trusted without adversarial review. Disagreement is signal, not noise.
|
|
116
|
+
|
|
117
|
+
3. **Evidence > Assumptions > Code > Documentation** — The epistemic schema enforces a strict separation between observations, interpretations, and judgments. No shortcuts. No opinion soup.
|
|
118
|
+
|
|
119
|
+
---
|
|
120
|
+
|
|
121
|
+
## The 14 Audit Domains
|
|
122
|
+
|
|
123
|
+
Radar audits across 14 domains (0-13). Every audit produces findings in each domain — even if the finding is "no major issues found." Nothing is optional.
|
|
124
|
+
|
|
125
|
+
### Domain 0 — Context & Intent
|
|
126
|
+
|
|
127
|
+
**Must come first. Without this, every other audit is partially blind.**
|
|
128
|
+
|
|
129
|
+
| Aspect | Details |
|
|
130
|
+
|---|---|
|
|
131
|
+
| Owner | Principal Engineer |
|
|
132
|
+
| Purpose | Establish what the system does, who uses it, what constraints exist |
|
|
133
|
+
|
|
134
|
+
What to audit:
|
|
135
|
+
- What problem is this system solving?
|
|
136
|
+
- Who are the users?
|
|
137
|
+
- What are success/failure criteria?
|
|
138
|
+
- What data does it handle?
|
|
139
|
+
- What are its failure modes?
|
|
140
|
+
- What is the business criticality?
|
|
141
|
+
- What constraints exist (regulatory, cost, latency, team size)?
|
|
142
|
+
- What is explicitly OUT of scope?
|
|
143
|
+
|
|
144
|
+
### Domain 1 — Architecture & System Design
|
|
145
|
+
|
|
146
|
+
| Aspect | Details |
|
|
147
|
+
|---|---|
|
|
148
|
+
| Owner | Architect Agent |
|
|
149
|
+
| Why it matters | Architecture determines how easy bugs are to introduce, how expensive changes are, and whether scale is possible without rewrites |
|
|
150
|
+
|
|
151
|
+
What to audit:
|
|
152
|
+
- System boundaries and module responsibilities
|
|
153
|
+
- Dependency direction and layering consistency
|
|
154
|
+
- Coupling vs cohesion
|
|
155
|
+
- Domain modeling quality
|
|
156
|
+
- Data flow clarity
|
|
157
|
+
- Presence of clear architectural pattern (hexagonal, layered, microservices, etc.)
|
|
158
|
+
- Boundary enforcement vs violation
|
|
159
|
+
- "God modules" or "utility dumping grounds"
|
|
160
|
+
- Domain logic mixed with infrastructure
|
|
161
|
+
|
|
162
|
+
Key questions:
|
|
163
|
+
- Is there a clear architectural pattern?
|
|
164
|
+
- Are boundaries enforced or violated?
|
|
165
|
+
- Are "god modules" present?
|
|
166
|
+
- Is domain logic mixed with infrastructure?
|
|
167
|
+
|
|
168
|
+
### Domain 2 — Data & State Integrity
|
|
169
|
+
|
|
170
|
+
| Aspect | Details |
|
|
171
|
+
|---|---|
|
|
172
|
+
| Owner | Data Engineer Agent |
|
|
173
|
+
| Why it matters | Most catastrophic failures are data bugs, not code bugs — corrupt state, irreversible migrations, silent data loss, inconsistent derived data |
|
|
174
|
+
|
|
175
|
+
What to audit:
|
|
176
|
+
- Data models and schemas
|
|
177
|
+
- Schema evolution and migrations
|
|
178
|
+
- Backward compatibility
|
|
179
|
+
- Referential integrity
|
|
180
|
+
- Eventual consistency guarantees
|
|
181
|
+
- State transitions and invariants
|
|
182
|
+
- Data loss risks
|
|
183
|
+
|
|
184
|
+
### Domain 3 — Correctness & Logic
|
|
185
|
+
|
|
186
|
+
| Aspect | Details |
|
|
187
|
+
|---|---|
|
|
188
|
+
| Owner | Senior Application Engineer Agent |
|
|
189
|
+
| Why it matters | Most production incidents are boring logic bugs, not exotic failures |
|
|
190
|
+
|
|
191
|
+
What to audit:
|
|
192
|
+
- Logic errors and edge case handling
|
|
193
|
+
- Error propagation (swallowed, ignored, or properly handled)
|
|
194
|
+
- Concurrency correctness
|
|
195
|
+
- Input validation
|
|
196
|
+
- Data consistency
|
|
197
|
+
- Assumption documentation (documented or implicit)
|
|
198
|
+
- Invariant enforcement
|
|
199
|
+
- Retry safety (idempotency)
|
|
200
|
+
|
|
201
|
+
### Domain 4 — Security
|
|
202
|
+
|
|
203
|
+
| Aspect | Details |
|
|
204
|
+
|---|---|
|
|
205
|
+
| Owner | Security Engineer Agent |
|
|
206
|
+
| Why it matters | Security failures are existential risks |
|
|
207
|
+
|
|
208
|
+
What to audit:
|
|
209
|
+
- Authentication and authorization (AuthN/AuthZ)
|
|
210
|
+
- Secrets handling
|
|
211
|
+
- Input sanitization
|
|
212
|
+
- Injection risks (SQL, XSS, command, etc.)
|
|
213
|
+
- Dependency vulnerabilities
|
|
214
|
+
- Cryptography misuse
|
|
215
|
+
- Supply chain risk
|
|
216
|
+
- Trust boundaries
|
|
217
|
+
- Least-privilege enforcement
|
|
218
|
+
- Sensitive data in logs
|
|
219
|
+
|
|
220
|
+
### Domain 5 — Compliance Privacy & Governance
|
|
221
|
+
|
|
222
|
+
| Aspect | Details |
|
|
223
|
+
|---|---|
|
|
224
|
+
| Owner | Compliance Officer Agent |
|
|
225
|
+
| Why it matters | Compliance failures lead to fines, lawsuits, and shutdowns |
|
|
226
|
+
|
|
227
|
+
What to audit:
|
|
228
|
+
- PII handling and data classification
|
|
229
|
+
- Data retention policies
|
|
230
|
+
- Encryption at rest and in transit
|
|
231
|
+
- Audit logging
|
|
232
|
+
- Consent tracking
|
|
233
|
+
- Where personal data is stored
|
|
234
|
+
- How personal data is deleted
|
|
235
|
+
- Whether access is auditable
|
|
236
|
+
|
|
237
|
+
### Domain 6 — Testing Strategy & Verification
|
|
238
|
+
|
|
239
|
+
| Aspect | Details |
|
|
240
|
+
|---|---|
|
|
241
|
+
| Owner | Test Engineer Agent |
|
|
242
|
+
| Why it matters | Senior engineers don't ask "Do you have tests?" — they ask "Would these tests catch the most expensive failures?" |
|
|
243
|
+
|
|
244
|
+
What to audit:
|
|
245
|
+
- Test pyramid shape (unit vs integration vs e2e balance)
|
|
246
|
+
- Determinism vs flakiness
|
|
247
|
+
- What ISN'T tested (gaps)
|
|
248
|
+
- Mutation resistance
|
|
249
|
+
- Contract testing presence
|
|
250
|
+
- Tests as documentation
|
|
251
|
+
- Failure coverage (do tests cover failure paths, not just happy paths?)
|
|
252
|
+
|
|
253
|
+
### Domain 7 — Reliability & Resilience
|
|
254
|
+
|
|
255
|
+
| Aspect | Details |
|
|
256
|
+
|---|---|
|
|
257
|
+
| Owner | SRE Agent |
|
|
258
|
+
| Why it matters | Production systems fail constantly — good ones degrade gracefully |
|
|
259
|
+
|
|
260
|
+
What to audit:
|
|
261
|
+
- Failure handling patterns
|
|
262
|
+
- Retry strategies (bounded? with backoff?)
|
|
263
|
+
- Timeouts
|
|
264
|
+
- Circuit breakers
|
|
265
|
+
- Startup/shutdown safety
|
|
266
|
+
- What happens when dependencies fail
|
|
267
|
+
- Whether failures are noisy or silent
|
|
268
|
+
- State recoverability
|
|
269
|
+
|
|
270
|
+
### Domain 8 — Scalability & Performance
|
|
271
|
+
|
|
272
|
+
| Aspect | Details |
|
|
273
|
+
|---|---|
|
|
274
|
+
| Owner | Performance Engineer Agent |
|
|
275
|
+
| Why it matters | Scaling failures are often design bugs, not hardware limits |
|
|
276
|
+
|
|
277
|
+
What to audit:
|
|
278
|
+
- Algorithmic complexity
|
|
279
|
+
- N+1 queries
|
|
280
|
+
- Caching strategy
|
|
281
|
+
- Resource usage patterns
|
|
282
|
+
- Async vs blocking behavior
|
|
283
|
+
- What grows with user count
|
|
284
|
+
- What grows with data size
|
|
285
|
+
- Where bottlenecks exist
|
|
286
|
+
- Whether backpressure is implemented
|
|
287
|
+
|
|
288
|
+
### Domain 9 — Maintainability & Code Health
|
|
289
|
+
|
|
290
|
+
| Aspect | Details |
|
|
291
|
+
|---|---|
|
|
292
|
+
| Owner | Senior Application Engineer Agent |
|
|
293
|
+
| Why it matters | Maintenance cost dominates total lifecycle cost |
|
|
294
|
+
|
|
295
|
+
What to audit:
|
|
296
|
+
- Code smells and duplication
|
|
297
|
+
- Naming clarity
|
|
298
|
+
- Documentation accuracy
|
|
299
|
+
- Whether intent is obvious
|
|
300
|
+
- Whether tests are meaningful or superficial
|
|
301
|
+
- Tech debt interest accruing
|
|
302
|
+
|
|
303
|
+
### Domain 10 — Operability & Developer Experience
|
|
304
|
+
|
|
305
|
+
| Aspect | Details |
|
|
306
|
+
|---|---|
|
|
307
|
+
| Owner | SRE Agent |
|
|
308
|
+
| Why it matters | Many "great codebases" fail in production because ops was ignored |
|
|
309
|
+
|
|
310
|
+
What to audit:
|
|
311
|
+
- CI/CD pipelines
|
|
312
|
+
- Rollback safety
|
|
313
|
+
- Feature flags
|
|
314
|
+
- Observability (logging, metrics, tracing)
|
|
315
|
+
- Ownership clarity
|
|
316
|
+
- Debuggability
|
|
317
|
+
- Local dev friction
|
|
318
|
+
- Can this be safely deployed on Friday?
|
|
319
|
+
- Can incidents be diagnosed quickly?
|
|
320
|
+
- Who owns what?
|
|
321
|
+
|
|
322
|
+
### Domain 11 — Change Risk & Evolvability
|
|
323
|
+
|
|
324
|
+
| Aspect | Details |
|
|
325
|
+
|---|---|
|
|
326
|
+
| Owner | Staff Engineer Agent |
|
|
327
|
+
| Why it matters | "How dangerous is it to touch this code?" predicts velocity decay, team burnout, and rewrite pressure |
|
|
328
|
+
|
|
329
|
+
What to audit:
|
|
330
|
+
- Change amplification (how many files does one change touch?)
|
|
331
|
+
- Refactor safety
|
|
332
|
+
- Blast radius analysis
|
|
333
|
+
- Modularity health
|
|
334
|
+
- How risky future changes are
|
|
335
|
+
|
|
336
|
+
### Domain 12 — Team Ownership & Knowledge Risk
|
|
337
|
+
|
|
338
|
+
| Aspect | Details |
|
|
339
|
+
|---|---|
|
|
340
|
+
| Owner | Staff Engineer Agent |
|
|
341
|
+
| Why it matters | Systems fail socially before they fail technically |
|
|
342
|
+
|
|
343
|
+
What to audit:
|
|
344
|
+
- Code authorship concentration
|
|
345
|
+
- Bus factor per module
|
|
346
|
+
- Tribal knowledge hotspots
|
|
347
|
+
- Documentation debt
|
|
348
|
+
- Review culture artifacts
|
|
349
|
+
- Knowledge silos
|
|
350
|
+
|
|
351
|
+
### Domain 13 — Risk Synthesis & Forecasting
|
|
352
|
+
|
|
353
|
+
| Aspect | Details |
|
|
354
|
+
|---|---|
|
|
355
|
+
| Owner | Principal Engineer Agent |
|
|
356
|
+
| Why it matters | Senior engineers think in predictions: "This will break in 3 months," "This will fail at 10x traffic," "This is safe unless compliance changes" |
|
|
357
|
+
|
|
358
|
+
What to synthesize:
|
|
359
|
+
- Likelihood x impact for all findings
|
|
360
|
+
- Time-to-failure predictions
|
|
361
|
+
- "What breaks first" analysis
|
|
362
|
+
- Risk acceptance vs remediation recommendations
|
|
363
|
+
- Cross-domain emergent risks
|
|
364
|
+
|
|
365
|
+
---
|
|
366
|
+
|
|
367
|
+
## The Agent Team
|
|
368
|
+
|
|
369
|
+
Radar deploys a minimal-complete set of agent personas. Each exists because removing it would leave a blind spot. No overlap, no bloat.
|
|
370
|
+
|
|
371
|
+
### 1. Principal Engineer
|
|
372
|
+
|
|
373
|
+
**Role:** Epistemic governor of the entire audit
|
|
374
|
+
|
|
375
|
+
The Principal Engineer is NOT the best coder, architect, or the most knowledgeable in every domain. They are the person accountable for the **correctness of collective reasoning**.
|
|
376
|
+
|
|
377
|
+
**Core mental models:**
|
|
378
|
+
- Thinks in systems of systems — "What behavior emerges from these interactions?"
|
|
379
|
+
- Separates facts, interpretations, and judgments (enforces the epistemic schema)
|
|
380
|
+
- Actively manages uncertainty — budgets it rather than eliminating it
|
|
381
|
+
- Thinks in time horizons (immediate, near-term, long-term, hypothetical)
|
|
382
|
+
- Optimizes organizational attention, not code
|
|
383
|
+
|
|
384
|
+
**Responsibilities:**
|
|
385
|
+
1. Define audit scope and non-goals (Phase 0)
|
|
386
|
+
2. Calibrate severity scales
|
|
387
|
+
3. Resolve cross-domain conflicts
|
|
388
|
+
4. Synthesize narrative
|
|
389
|
+
5. Forecast future failure
|
|
390
|
+
6. Translate findings for multiple audiences
|
|
391
|
+
7. Own the final call
|
|
392
|
+
8. Explicitly respond to every Devil's Advocate critique
|
|
393
|
+
|
|
394
|
+
**Must never:**
|
|
395
|
+
- Introduce new raw findings late
|
|
396
|
+
- Re-run tools
|
|
397
|
+
- Argue minutiae
|
|
398
|
+
|
|
399
|
+
They reason, arbitrate, and narrate.
|
|
400
|
+
|
|
401
|
+
**Active in:** Phase 0 (Context), Phase 5 (Synthesis)
|
|
402
|
+
|
|
403
|
+
### 2. Architect
|
|
404
|
+
|
|
405
|
+
**Domains:** 1 (Architecture & System Design)
|
|
406
|
+
|
|
407
|
+
Evaluates structural patterns, boundaries, dependency direction, coupling, cohesion, and whether the architecture can support the system's actual requirements.
|
|
408
|
+
|
|
409
|
+
### 3. Data Engineer
|
|
410
|
+
|
|
411
|
+
**Domains:** 2 (Data & State Integrity)
|
|
412
|
+
|
|
413
|
+
Evaluates data models, schema evolution, migrations, referential integrity, consistency guarantees, and state transition safety.
|
|
414
|
+
|
|
415
|
+
### 4. Security Engineer
|
|
416
|
+
|
|
417
|
+
**Domains:** 4 (Security)
|
|
418
|
+
|
|
419
|
+
Evaluates AuthN/AuthZ, secrets handling, injection risks, supply chain, cryptography, trust boundaries, and attacker models.
|
|
420
|
+
|
|
421
|
+
### 5. Compliance Officer
|
|
422
|
+
|
|
423
|
+
**Domains:** 5 (Compliance Privacy & Governance)
|
|
424
|
+
|
|
425
|
+
Evaluates PII handling, data retention, encryption, audit logging, consent tracking, and regulatory exposure.
|
|
426
|
+
|
|
427
|
+
### 6. Senior Application Engineer
|
|
428
|
+
|
|
429
|
+
**Domains:** 3 (Correctness & Logic), 9 (Maintainability & Code Health)
|
|
430
|
+
|
|
431
|
+
Evaluates logic correctness, error handling, edge cases, idempotency, code smells, naming, duplication, and intent clarity.
|
|
432
|
+
|
|
433
|
+
### 7. SRE (Site Reliability Engineer)
|
|
434
|
+
|
|
435
|
+
**Domains:** 7 (Reliability & Resilience), 10 (Operability & DevEx)
|
|
436
|
+
|
|
437
|
+
Evaluates failure handling, retries, timeouts, circuit breakers, CI/CD, rollback safety, observability, and operational readiness.
|
|
438
|
+
|
|
439
|
+
### 8. Performance Engineer
|
|
440
|
+
|
|
441
|
+
**Domains:** 8 (Scalability & Performance)
|
|
442
|
+
|
|
443
|
+
Evaluates algorithmic complexity, N+1 queries, caching, resource usage, async behavior, bottlenecks, and backpressure.
|
|
444
|
+
|
|
445
|
+
### 9. Test Engineer
|
|
446
|
+
|
|
447
|
+
**Domains:** 6 (Testing Strategy & Verification)
|
|
448
|
+
|
|
449
|
+
Evaluates test pyramid, determinism, mutation resistance, contract tests, failure coverage, and test-as-documentation quality.
|
|
450
|
+
|
|
451
|
+
### 10. Staff Engineer
|
|
452
|
+
|
|
453
|
+
**Domains:** 11 (Change Risk & Evolvability), 12 (Team Ownership & Knowledge Risk)
|
|
454
|
+
|
|
455
|
+
Evaluates change amplification, refactor safety, blast radius, bus factor, knowledge silos, and documentation debt. This is a synthesis-heavy role that draws on git history and social signals.
|
|
456
|
+
|
|
457
|
+
### 11. Reality Gap Analyst
|
|
458
|
+
|
|
459
|
+
**Purpose:** Detect divergence between "code as written" and "system as run"
|
|
460
|
+
|
|
461
|
+
This agent often finds: "The audit is technically correct but operationally wrong." Audits config files, environment-specific behavior, feature flags, deployment manifests, runtime overrides, and kill switches.
|
|
462
|
+
|
|
463
|
+
See [Reality Gap Framework](#reality-gap-framework) for full details.
|
|
464
|
+
|
|
465
|
+
### 12. Devil's Advocate Reviewer
|
|
466
|
+
|
|
467
|
+
**Purpose:** Hunt collective blind spots
|
|
468
|
+
|
|
469
|
+
The Devil's Advocate is NOT a contrarian, NOT "the negative one," NOT "the security pessimist." They are the agent that hunts collective blind spots.
|
|
470
|
+
|
|
471
|
+
**Why this role exists:**
|
|
472
|
+
Every audit naturally develops consensus gravity, optimism bias, tool bias ("the scanner didn't find anything"), and authority bias. The Devil's Advocate exists to break coherence.
|
|
473
|
+
|
|
474
|
+
**Core mental models:**
|
|
475
|
+
- Assume the model is wrong: "If this report is wrong, *how* would it be wrong?"
|
|
476
|
+
- Attack confidence, not just conclusions — target high-confidence claims, clean narratives, and areas with little disagreement
|
|
477
|
+
- Seek asymmetric failure: "What failure would be disproportionately damaging relative to how little we're talking about it?"
|
|
478
|
+
- Use inversion relentlessly: "Under what conditions does this become unsafe?"
|
|
479
|
+
- Do NOT propose solutions — solutions dilute the critique
|
|
480
|
+
|
|
481
|
+
**Outputs:**
|
|
482
|
+
1. Most confident claim I distrust
|
|
483
|
+
2. Least discussed but highest-impact risk
|
|
484
|
+
3. Assumptions that must hold for conclusions to be true
|
|
485
|
+
4. Evidence that was overweighted
|
|
486
|
+
5. Evidence that was ignored or unavailable
|
|
487
|
+
6. Alternate narrative that fits the data
|
|
488
|
+
|
|
489
|
+
**Critical rule:** If the Devil's Advocate panel is empty, the system is broken.
|
|
490
|
+
|
|
491
|
+
---
|
|
492
|
+
|
|
493
|
+
## Execution Phases
|
|
494
|
+
|
|
495
|
+
Radar Core executes in six diagnostic phases (0-5). Order matters. Phases 6-8 (the Transform pipeline) extend diagnosis into remediation — see [Transform Execution Phases](#transform-execution-phases).
|
|
496
|
+
|
|
497
|
+
### Phase 0 — Context & Threat Modeling
|
|
498
|
+
|
|
499
|
+
**Agent:** Principal Engineer
|
|
500
|
+
|
|
501
|
+
Establish intent, constraints, risk profile, and non-goals. Without this phase, all audits are shallow.
|
|
502
|
+
|
|
503
|
+
Inputs: Repository, documentation, README, deployment configs, any available business context.
|
|
504
|
+
|
|
505
|
+
Outputs: Audit scope document, threat model, risk profile, explicit non-goals.
|
|
506
|
+
|
|
507
|
+
### Phase 1 — Automated Signal Gathering
|
|
508
|
+
|
|
509
|
+
**Agents:** Tool runners (non-reasoning heavy)
|
|
510
|
+
|
|
511
|
+
Run automated tools. Gather signals across six orthogonal dimensions. No opinions yet. Just evidence.
|
|
512
|
+
|
|
513
|
+
**Signal dimensions:**
|
|
514
|
+
1. **Structure** — SonarQube, Semgrep, dependency graphs
|
|
515
|
+
2. **Behavior** — Profilers, async analysis
|
|
516
|
+
3. **History** — Git churn, file age vs modification frequency, author concentration
|
|
517
|
+
4. **Dependencies** — Trivy, Syft, Grype, OpenSSF Scorecard
|
|
518
|
+
5. **Policy posture** — Checkov, Gitleaks, Semgrep
|
|
519
|
+
6. **Runtime contracts** — OpenAPI/gRPC schema validation, backward compatibility
|
|
520
|
+
|
|
521
|
+
Outputs: Normalized signal data tagged with severity, confidence, blast radius, and domain relevance.
|
|
522
|
+
|
|
523
|
+
### Phase 2 — Deep Domain Audits
|
|
524
|
+
|
|
525
|
+
**Agents:** Architect, Data Engineer, Security Engineer, Compliance Officer, Senior Application Engineer, SRE, Performance Engineer, Test Engineer
|
|
526
|
+
|
|
527
|
+
Each agent audits ONLY their assigned domains. Each receives the same Phase 1 evidence. Each produces independent findings using the formal epistemic schema.
|
|
528
|
+
|
|
529
|
+
Sessions can run in parallel. Each produces a structured findings file.
|
|
530
|
+
|
|
531
|
+
### Phase 3 — Change Risk, Team Risk & Reality Gap
|
|
532
|
+
|
|
533
|
+
**Agents:** Staff Engineer, Reality Gap Analyst
|
|
534
|
+
|
|
535
|
+
Synthesis-heavy roles that draw on Phase 2 findings + git history + configuration analysis.
|
|
536
|
+
|
|
537
|
+
The Staff Engineer evaluates change risk and ownership risk.
|
|
538
|
+
The Reality Gap Analyst checks for divergence between code-as-written and system-as-run.
|
|
539
|
+
|
|
540
|
+
### Phase 4 — Adversarial Review ("Fresh Eyes")
|
|
541
|
+
|
|
542
|
+
**Agent:** Devil's Advocate Reviewer
|
|
543
|
+
|
|
544
|
+
Goal: Invalidate conclusions, not agree. A fresh session whose only job is to challenge assumptions, attack confidence, and surface what was missed.
|
|
545
|
+
|
|
546
|
+
The Devil's Advocate reads ALL domain findings and produces their critique.
|
|
547
|
+
|
|
548
|
+
### Phase 5 — Synthesis & Report Generation
|
|
549
|
+
|
|
550
|
+
**Agent:** Principal Engineer
|
|
551
|
+
|
|
552
|
+
The Principal reads all domain findings + Devil's Advocate critique. For every disagreement, they must explicitly respond. Silence is not allowed.
|
|
553
|
+
|
|
554
|
+
Produces the final Radar report:
|
|
555
|
+
1. Executive Risk Summary
|
|
556
|
+
2. Architecture Narrative
|
|
557
|
+
3. Findings by Domain (severity-ranked)
|
|
558
|
+
4. Cross-Validation Notes (disagreements and resolutions)
|
|
559
|
+
5. Remediation Roadmap
|
|
560
|
+
6. Long-Term Structural Risks
|
|
561
|
+
7. "What Would Break First at 10x Scale"
|
|
562
|
+
|
|
563
|
+
---
|
|
564
|
+
|
|
565
|
+
## The Three Output Layers
|
|
566
|
+
|
|
567
|
+
Radar produces three distinct output layers. Each has a different purpose, different mutability, and different consumer.
|
|
568
|
+
|
|
569
|
+
**Layer A — Diagnostic Artifact (Truth Layer)**
|
|
570
|
+
|
|
571
|
+
The audit itself. Immutable, reproducible, epistemically versioned. Phases 0-5 produce Layer A. It is the forensic record of what was found, how it was found, and how confident the system is.
|
|
572
|
+
|
|
573
|
+
Layer A is never mutated. Once a finding is produced, it exists permanently in the audit record. Subsequent analysis may reinterpret it, but the original observation stands.
|
|
574
|
+
|
|
575
|
+
**Layer B — Remediation Knowledge (Instruction Layer)**
|
|
576
|
+
|
|
577
|
+
Derived from Layer A. Framework-specific. Pattern-based. Educational. This is where playbooks live.
|
|
578
|
+
|
|
579
|
+
Layer B answers: "Given what was found, how should it be fixed?" But not generically — parametrically. Every remediation is expressed at four layers of specificity (see [The Transformation Model](#the-transformation-model)) and carries both human-readable markdown and machine-consumable structured data.
|
|
580
|
+
|
|
581
|
+
Layer B is derived. It cannot exist without Layer A. It cannot contradict Layer A. If a finding changes, Layer B artifacts that reference it must be regenerated.
|
|
582
|
+
|
|
583
|
+
**Layer C — Change Orchestration (Execution Layer)**
|
|
584
|
+
|
|
585
|
+
Dependency-aware execution plans. Change graphs. Risk scoring. Verification gates. This is operational sequencing, not documentation.
|
|
586
|
+
|
|
587
|
+
Layer C answers: "In what order should fixes be applied, with what safety checks, and what happens if something goes wrong?"
|
|
588
|
+
|
|
589
|
+
Layer C is the Drive integration point. Radar Transform produces Drive-ready project artifacts — phased remediation plans with dependency ordering, risk-scored task definitions, and verification gates. Radar does not execute changes. Drive does.
|
|
590
|
+
|
|
591
|
+
**The Pipeline:**
|
|
592
|
+
|
|
593
|
+
```
|
|
594
|
+
Layer A (Diagnosis) → Layer B (Knowledge) → Layer C (Orchestration)
|
|
595
|
+
Immutable truth Derived instruction Operational planning
|
|
596
|
+
Phases 0-5 Phases 6-7 Phase 8
|
|
597
|
+
```
|
|
598
|
+
|
|
599
|
+
---
|
|
600
|
+
|
|
601
|
+
## Radar Transform
|
|
602
|
+
|
|
603
|
+
**The Controlled Evolution Engine**
|
|
604
|
+
|
|
605
|
+
Radar Core is a diagnosis system. It finds problems with epistemic rigor and forensic-grade traceability. But diagnosis alone is incomplete. A report that says "you have 47 findings" without actionable, risk-scored, dependency-ordered remediation is a paper tiger.
|
|
606
|
+
|
|
607
|
+
Radar Transform is the second-order reasoning system that converts diagnostic findings into AI-consumable transformation artifacts — playbooks, remediation plans, guardrails, and Drive-integrated execution plans.
|
|
608
|
+
|
|
609
|
+
**Core Principle:** *Diagnosis is decentralized. Intervention is centralized.*
|
|
610
|
+
|
|
611
|
+
In the diagnostic pipeline, 12 agents work independently across 14 domains. Each agent audits its own territory. Independence prevents groupthink.
|
|
612
|
+
|
|
613
|
+
In the Transform pipeline, 5 agents coordinate intervention. Remediation cannot be done in isolation — a fix to the authentication system affects security, architecture, testing, and deployment. Transform agents see the full picture and orchestrate change holistically.
|
|
614
|
+
|
|
615
|
+
**What Transform Is Not:**
|
|
616
|
+
|
|
617
|
+
- It is not bolt-on post-processing. It is a second-order reasoning system with its own agents, workflows, and risk modeling.
|
|
618
|
+
- It is not auto-remediation. Transform proposes changes; humans (via Drive) execute them.
|
|
619
|
+
- It is not generic advice. Transform produces parametric, framework-specific, project-contextualized remediation.
|
|
620
|
+
|
|
621
|
+
**What Transform Consumes:**
|
|
622
|
+
|
|
623
|
+
All Layer A outputs — findings, domain knowledge, confidence scores, disagreements, cross-validation notes. Transform agents have full visibility into the diagnostic record.
|
|
624
|
+
|
|
625
|
+
**What Transform Produces:**
|
|
626
|
+
|
|
627
|
+
- Layer B: Remediation knowledge (playbooks, patterns, guardrails, educational context)
|
|
628
|
+
- Layer C: Change orchestration (dependency graphs, risk scores, verification plans, Drive projects)
|
|
629
|
+
|
|
630
|
+
---
|
|
631
|
+
|
|
632
|
+
## Intervention Levels
|
|
633
|
+
|
|
634
|
+
Every Transform output is explicitly classified by intervention level. This is not optional. Intervention levels gate what the system is allowed to produce and how much confidence is required.
|
|
635
|
+
|
|
636
|
+
| Level | Name | Definition | Confidence Required |
|
|
637
|
+
|-------|------|------------|---------------------|
|
|
638
|
+
| 1 | **Suggesting** | "Consider this pattern." Informational. No action implied. | Any |
|
|
639
|
+
| 2 | **Planning** | "Here's how to fix this." Structured plan. Human decides whether to act. | Medium+ |
|
|
640
|
+
| 3 | **Authorizing** | "This change is recommended with confidence X." Risk-scored, gate-checked. | High |
|
|
641
|
+
| 4 | **Executing** | "Apply this change." Only via Drive with verification gates. Never auto-applied by Radar. | High + Low change risk |
|
|
642
|
+
|
|
643
|
+
**Safety Principle:** *Default to the lowest intervention level that serves the user. Escalation requires evidence.*
|
|
644
|
+
|
|
645
|
+
A finding with medium confidence gets Suggesting or Planning. Never Authorizing. A change with high blast radius stays at Planning even with high confidence. Intervention levels are a ratchet — evidence raises them, uncertainty lowers them.
|
|
646
|
+
|
|
647
|
+
**Why This Matters:**
|
|
648
|
+
|
|
649
|
+
Without explicit intervention levels, every AI code tool produces the same thing: confident-sounding suggestions with no accountability framework. Radar Transform distinguishes between "you might want to look at this" and "we are confident this change should be made, here is the risk assessment, and here is how to verify it worked."
|
|
650
|
+
|
|
651
|
+
---
|
|
652
|
+
|
|
653
|
+
## The Transformation Model
|
|
654
|
+
|
|
655
|
+
Remediation knowledge is parametric, not generic. Generic advice ("use parameterized queries") is useless to a developer staring at a specific codebase. The 4-layer transformation model ensures every piece of remediation is grounded at every level of specificity.
|
|
656
|
+
|
|
657
|
+
### Layer 1 — Abstract Pattern
|
|
658
|
+
|
|
659
|
+
The universal principle. Language-agnostic. Framework-agnostic.
|
|
660
|
+
|
|
661
|
+
*Example:* "Unbounded retries are dangerous. They amplify load during partial outages and can cause cascading failures."
|
|
662
|
+
|
|
663
|
+
### Layer 2 — Framework Mapping
|
|
664
|
+
|
|
665
|
+
The principle applied to a specific framework.
|
|
666
|
+
|
|
667
|
+
*Example:* "In Laravel, use `retryUntil()` with a deadline. In Express, use the `p-retry` library with exponential backoff. In Spring Boot, use `@Retryable` with `maxAttempts` and `backoff`."
|
|
668
|
+
|
|
669
|
+
### Layer 3 — Language Mapping
|
|
670
|
+
|
|
671
|
+
The implementation pattern in the project's language.
|
|
672
|
+
|
|
673
|
+
*Example:* "PHP retry with exponential backoff: `$delay = min($baseDelay * (2 ** $attempt), $maxDelay)`"
|
|
674
|
+
|
|
675
|
+
### Layer 4 — Project Context
|
|
676
|
+
|
|
677
|
+
The specific files, functions, and call sites in this codebase.
|
|
678
|
+
|
|
679
|
+
*Example:* "Files `app/Services/PaymentGateway.php` (line 47), `app/Jobs/SyncInventory.php` (line 112), and `app/Http/Controllers/WebhookController.php` (line 89) — these 3 call sites use unbounded `retry()` loops. Each needs exponential backoff with a circuit breaker."
|
|
680
|
+
|
|
681
|
+
### Why Generic Remediation Is Fluff
|
|
682
|
+
|
|
683
|
+
A playbook that only contains Layer 1 is an essay. A playbook that contains all four layers is a work order. The difference between "you should fix your retries" and "here are the 3 files, here is the pattern for your framework, and here is how to verify it works" is the difference between a blog post and engineering guidance.
|
|
684
|
+
|
|
685
|
+
---
|
|
686
|
+
|
|
687
|
+
## Transform Agent Team
|
|
688
|
+
|
|
689
|
+
Transform deploys 5 specialized agents. Each exists because removing it would leave a gap in the remediation pipeline.
|
|
690
|
+
|
|
691
|
+
**Design Principle:** *Remediation must be centralized. Domain agents diagnose independently. Transform agents coordinate intervention.*
|
|
692
|
+
|
|
693
|
+
### Remediation Architect
|
|
694
|
+
|
|
695
|
+
**Role:** Translates diagnosis into structured change plans.
|
|
696
|
+
|
|
697
|
+
**Consumes:** All findings + domain knowledge + framework context.
|
|
698
|
+
**Produces:** Remediation playbooks at all 4 transformation layers. Change dependency graphs. Sequenced fix orders.
|
|
699
|
+
|
|
700
|
+
**Why they exist:** Someone has to synthesize 47 findings into "fix these 5 things in this order and the other 42 findings improve as side effects." The Remediation Architect identifies root causes, groups related findings, and sequences changes by dependency and impact.
|
|
701
|
+
|
|
702
|
+
### Change Risk Modeler
|
|
703
|
+
|
|
704
|
+
**Role:** Scores blast radius, coupling, regression probability, and architectural tension for every proposed change.
|
|
705
|
+
|
|
706
|
+
**Consumes:** Change plan + codebase structure + git history + test coverage.
|
|
707
|
+
**Produces:** Per-change risk scores. Risk-adjusted priority ordering. Change risk assessment reports.
|
|
708
|
+
|
|
709
|
+
**Why they exist:** Every fix introduces new risk. A change that fixes a security vulnerability but breaks 14 integration tests is not a net improvement. The Change Risk Modeler ensures that remediation doesn't create more problems than it solves.
|
|
710
|
+
|
|
711
|
+
### Pedagogy Agent
|
|
712
|
+
|
|
713
|
+
**Role:** Explains fixes for AI-assisted developers.
|
|
714
|
+
|
|
715
|
+
**Consumes:** Remediation plan + framework context + project patterns.
|
|
716
|
+
**Produces:** Educational context at all 4 transformation layers. Before/after examples. "Why this matters" explanations. Best-practice rationale.
|
|
717
|
+
|
|
718
|
+
**Why they exist:** The fastest-growing segment of developers uses AI assistants for code generation. They can implement a fix but may not understand *why*. Without pedagogical context, fixes get applied without understanding, and the same patterns recur. The Pedagogy Agent ensures that every fix teaches something.
|
|
719
|
+
|
|
720
|
+
### Guardrail Generator
|
|
721
|
+
|
|
722
|
+
**Role:** Writes project rules for future AI usage.
|
|
723
|
+
|
|
724
|
+
**Consumes:** Patterns + findings + project conventions.
|
|
725
|
+
**Produces:** `.claude/CLAUDE.md` rules. `.cursorrules` files. Linter configurations. Pre-commit hooks. Custom Semgrep rules.
|
|
726
|
+
|
|
727
|
+
**Why they exist:** The highest-leverage output of an audit isn't a report — it's a set of rules that prevent the same problems from recurring. The Guardrail Generator translates audit findings into machine-enforceable constraints.
|
|
728
|
+
|
|
729
|
+
### Execution Validator
|
|
730
|
+
|
|
731
|
+
**Role:** Defines verification plans — how to prove fixes work.
|
|
732
|
+
|
|
733
|
+
**Consumes:** Change plan + test infrastructure + deployment configuration.
|
|
734
|
+
**Produces:** Per-change verification steps. Expected outcomes. Rollback criteria. Test specifications.
|
|
735
|
+
|
|
736
|
+
**Why they exist:** A fix without a verification plan is faith-based engineering. The Execution Validator ensures that every change can be proven correct before it's considered complete.
|
|
737
|
+
|
|
738
|
+
---
|
|
739
|
+
|
|
740
|
+
## Transform Execution Phases
|
|
741
|
+
|
|
742
|
+
Radar Transform executes in three phases (6-8), extending the Core diagnostic pipeline (0-5).
|
|
743
|
+
|
|
744
|
+
### Phase 6 — Remediation Synthesis
|
|
745
|
+
|
|
746
|
+
**Agents:** Remediation Architect, Pedagogy Agent
|
|
747
|
+
|
|
748
|
+
**Input:** Complete Layer A diagnostic record (all findings, domain knowledge, disagreement resolutions, confidence scores).
|
|
749
|
+
|
|
750
|
+
**Process:**
|
|
751
|
+
1. Remediation Architect groups findings by root cause and dependency
|
|
752
|
+
2. Remediation Architect produces playbooks at all 4 transformation layers
|
|
753
|
+
3. Pedagogy Agent enriches playbooks with educational context, before/after examples, and best-practice rationale
|
|
754
|
+
4. Each playbook is classified by intervention level
|
|
755
|
+
|
|
756
|
+
**Output:** Layer B remediation playbooks (human-readable markdown + machine-consumable YAML). Pattern library updates.
|
|
757
|
+
|
|
758
|
+
### Phase 7 — Change Risk Validation
|
|
759
|
+
|
|
760
|
+
**Agents:** Change Risk Modeler, Guardrail Generator
|
|
761
|
+
|
|
762
|
+
**Input:** Phase 6 playbooks + codebase structure + git history + test coverage.
|
|
763
|
+
|
|
764
|
+
**Process:**
|
|
765
|
+
1. Change Risk Modeler scores every proposed change across 4 dimensions (blast radius, coupling risk, regression probability, architectural tension)
|
|
766
|
+
2. Changes exceeding risk thresholds are flagged for downgrade (Authorizing → Planning) or rejection
|
|
767
|
+
3. Guardrail Generator produces project rules from pattern analysis
|
|
768
|
+
4. Final risk-adjusted priority ordering is established
|
|
769
|
+
|
|
770
|
+
**Output:** Risk-scored change plan. Generated guardrail files. Risk assessment report.
|
|
771
|
+
|
|
772
|
+
### Phase 8 — Execution Planning (Drive Handoff)
|
|
773
|
+
|
|
774
|
+
**Agents:** Execution Validator
|
|
775
|
+
|
|
776
|
+
**Input:** Risk-scored change plan from Phase 7 + test infrastructure + deployment configuration.
|
|
777
|
+
|
|
778
|
+
**Process:**
|
|
779
|
+
1. Execution Validator defines verification steps for every proposed change
|
|
780
|
+
2. System generates Drive-compatible project artifacts:
|
|
781
|
+
- `PROJECT.md` — Project definition with audit reference
|
|
782
|
+
- `ROADMAP.md` — Phased remediation plan with risk ordering
|
|
783
|
+
- Phased plans with dependency sequencing, verification gates, and rollback criteria
|
|
784
|
+
3. Risk scores and intervention levels are embedded in Drive task definitions
|
|
785
|
+
|
|
786
|
+
**Output:** Layer C execution artifacts. Complete Drive project ready for user's AI assistant to execute.
|
|
787
|
+
|
|
788
|
+
**Critical:** Radar Transform does NOT execute changes. Phase 8 produces a plan. The user's AI assistant, operating through Drive, executes the plan with human oversight at every gate.
|
|
789
|
+
|
|
790
|
+
---
|
|
791
|
+
|
|
792
|
+
## Commands
|
|
793
|
+
|
|
794
|
+
Radar is invoked through slash commands — guided wizard experiences that delegate to workflows, present options, and manage the full audit-to-remediation pipeline.
|
|
795
|
+
|
|
796
|
+
### Core Commands (Diagnostic)
|
|
797
|
+
|
|
798
|
+
| Command | Purpose |
|
|
799
|
+
|---------|---------|
|
|
800
|
+
| `/radar:audit` | Initiate a full diagnostic audit — guided wizard that configures scope, runs tools, orchestrates agents, and produces findings |
|
|
801
|
+
| `/radar:resume` | Resume an interrupted audit from the last completed phase |
|
|
802
|
+
| `/radar:status` | Show current audit position, phase progress, and next action |
|
|
803
|
+
| `/radar:report` | Generate or regenerate the final diagnostic report from completed findings |
|
|
804
|
+
|
|
805
|
+
### Transform Commands (Evolution)
|
|
806
|
+
|
|
807
|
+
| Command | Purpose |
|
|
808
|
+
|---------|---------|
|
|
809
|
+
| `/radar:remediate` | Generate remediation knowledge (Layer B) from diagnostic findings |
|
|
810
|
+
| `/radar:transform` | Generate execution plans (Layer C) from remediation knowledge |
|
|
811
|
+
| `/radar:playbook` | View or regenerate remediation playbooks for specific findings |
|
|
812
|
+
| `/radar:guardrails` | Generate project rules (`.claude/CLAUDE.md`, linter configs) from audit findings |
|
|
813
|
+
|
|
814
|
+
All commands use a wizard UX pattern: numbered options, cancel/back at every decision point, and clear confirmation before executing phases that consume significant resources.
|
|
815
|
+
|
|
816
|
+
---
|
|
817
|
+
|
|
818
|
+
## Change Risk in Remediation
|
|
819
|
+
|
|
820
|
+
Remediation introduces new risk. This is the fundamental tension of automated code evolution: the system that finds problems must not create worse ones.
|
|
821
|
+
|
|
822
|
+
> Automated technical debt migration is worse than manual debt — it happens faster and with less understanding.
|
|
823
|
+
|
|
824
|
+
Every proposed change is scored across four dimensions:
|
|
825
|
+
|
|
826
|
+
### Blast Radius
|
|
827
|
+
|
|
828
|
+
How much breaks if the fix is wrong.
|
|
829
|
+
|
|
830
|
+
A change to a utility function called from 200 locations has extreme blast radius. A change to a leaf function called from one test has minimal blast radius. Blast radius is not severity — a low-severity change can have massive blast radius.
|
|
831
|
+
|
|
832
|
+
### Coupling Risk
|
|
833
|
+
|
|
834
|
+
Does the fix create new dependencies.
|
|
835
|
+
|
|
836
|
+
Moving from inline SQL to an ORM introduces a dependency on the ORM. Extracting a function into a shared utility creates coupling between previously independent modules. Coupling risk asks: "Does this fix make the system harder to change in the future?"
|
|
837
|
+
|
|
838
|
+
### Regression Probability
|
|
839
|
+
|
|
840
|
+
Does the fix break existing behavior.
|
|
841
|
+
|
|
842
|
+
Measured by: test coverage of affected code paths, complexity of the change, number of implicit contracts that might be violated. A well-tested function with a simple change has low regression probability. An untested function with a complex refactor has high regression probability.
|
|
843
|
+
|
|
844
|
+
### Architectural Tension
|
|
845
|
+
|
|
846
|
+
Does the fix fight the existing design.
|
|
847
|
+
|
|
848
|
+
Introducing a message queue into a synchronous request/response system creates architectural tension. The fix may be correct in isolation but inappropriate for the system as currently designed. Architectural tension asks: "Does this fix require changing the system's fundamental assumptions?"
|
|
849
|
+
|
|
850
|
+
---
|
|
851
|
+
|
|
852
|
+
## Safety & Liability Framework
|
|
853
|
+
|
|
854
|
+
Moving from Advisor to Architectural Actor requires a formal safety framework. Radar Core is an advisor — it reports findings. Radar Transform is an architectural actor — it proposes specific changes to codebases. The liability profile is fundamentally different.
|
|
855
|
+
|
|
856
|
+
### Conservative Bias
|
|
857
|
+
|
|
858
|
+
**Default to the lowest intervention level that serves the user.**
|
|
859
|
+
|
|
860
|
+
When uncertain, suggest. Don't plan. When the plan is uncertain, don't authorize. When authorization is uncertain, don't execute. The cost of under-intervening (user applies a fix manually) is low. The cost of over-intervening (system proposes a change that breaks production) is catastrophic.
|
|
861
|
+
|
|
862
|
+
### Confidence Thresholds
|
|
863
|
+
|
|
864
|
+
**Do not generate remediation if finding confidence is below threshold.**
|
|
865
|
+
|
|
866
|
+
| Intervention Level | Minimum Finding Confidence | Minimum Evidence Sources |
|
|
867
|
+
|--------------------|----------------------------|--------------------------|
|
|
868
|
+
| Suggesting | Low | 1 |
|
|
869
|
+
| Planning | Medium | 2 |
|
|
870
|
+
| Authorizing | High | 3+ |
|
|
871
|
+
| Executing (via Drive) | High | 3+ with cross-validation |
|
|
872
|
+
|
|
873
|
+
A finding with low confidence and a single evidence source produces, at most, a suggestion. Not a plan. Not a change. A suggestion.
|
|
874
|
+
|
|
875
|
+
### Unsafe Context Flagging
|
|
876
|
+
|
|
877
|
+
**Flag when change risk exceeds acceptable bounds.**
|
|
878
|
+
|
|
879
|
+
If any change risk dimension (blast radius, coupling, regression probability, architectural tension) exceeds the "high" threshold, the system must:
|
|
880
|
+
|
|
881
|
+
1. Flag the change as unsafe
|
|
882
|
+
2. Downgrade intervention level to Suggesting (regardless of confidence)
|
|
883
|
+
3. Explain why the change is risky
|
|
884
|
+
4. Recommend human architectural review before proceeding
|
|
885
|
+
|
|
886
|
+
### No Auto-Execution
|
|
887
|
+
|
|
888
|
+
**Radar Transform NEVER applies changes. Ever.**
|
|
889
|
+
|
|
890
|
+
Transform produces plans. Drive executes plans. The user approves every execution step. There is no bypass, no override, no "trusted mode" that allows Radar to modify a codebase directly. This is a hard architectural boundary, not a configuration option.
|
|
891
|
+
|
|
892
|
+
### When the System Must Refuse
|
|
893
|
+
|
|
894
|
+
Radar Transform refuses to generate remediation when:
|
|
895
|
+
|
|
896
|
+
1. Finding confidence is insufficient for the requested intervention level
|
|
897
|
+
2. Change risk exceeds acceptable bounds and no downgrade is possible
|
|
898
|
+
3. The codebase lacks sufficient test coverage to verify proposed changes
|
|
899
|
+
4. Multiple high-severity disagreements remain unresolved for the affected findings
|
|
900
|
+
5. The remediation would require changes to systems outside the audit scope
|
|
901
|
+
|
|
902
|
+
Refusal is a feature. A system that always produces output, regardless of certainty, is not safe — it is reckless.
|
|
903
|
+
|
|
904
|
+
---
|
|
905
|
+
|
|
906
|
+
## Drive Integration
|
|
907
|
+
|
|
908
|
+
Layer C generates Drive-compatible project artifacts. This is how Radar hands off to the user's AI assistant for execution.
|
|
909
|
+
|
|
910
|
+
### What Gets Generated
|
|
911
|
+
|
|
912
|
+
A complete Drive project, ready to execute:
|
|
913
|
+
|
|
914
|
+
| Artifact | Contents |
|
|
915
|
+
|----------|----------|
|
|
916
|
+
| `PROJECT.md` | Project definition referencing the Radar audit, codebase target, and remediation scope |
|
|
917
|
+
| `ROADMAP.md` | Phased remediation plan with dependency ordering and verification gates |
|
|
918
|
+
| Phase plans | Per-phase PLAN.md files with tasks, acceptance criteria, and risk metadata |
|
|
919
|
+
|
|
920
|
+
### Phased Remediation
|
|
921
|
+
|
|
922
|
+
Changes are sequenced by dependency, not severity. A critical finding that depends on a medium finding being fixed first cannot be prioritized first. Radar Transform produces dependency-aware phases:
|
|
923
|
+
|
|
924
|
+
1. **Foundation changes** — Shared utilities, configurations, and infrastructure that other fixes depend on
|
|
925
|
+
2. **High-impact, low-risk changes** — Quick wins that reduce the finding count and validate the remediation pipeline
|
|
926
|
+
3. **High-impact, high-risk changes** — Major structural changes with verification gates at every step
|
|
927
|
+
4. **Cleanup and hardening** — Guardrail installation, documentation updates, and monitoring configuration
|
|
928
|
+
|
|
929
|
+
### Verification Gates
|
|
930
|
+
|
|
931
|
+
Every Drive phase includes verification criteria:
|
|
932
|
+
|
|
933
|
+
- **Pre-change verification:** Confirm the codebase is in the expected state before applying changes
|
|
934
|
+
- **Post-change verification:** Run tests, check behavior, validate that the fix actually fixes the finding
|
|
935
|
+
- **Regression verification:** Confirm that unrelated functionality still works
|
|
936
|
+
- **Rollback criteria:** Define when and how to undo a change that fails verification
|
|
937
|
+
|
|
938
|
+
### Risk Metadata in Tasks
|
|
939
|
+
|
|
940
|
+
Every Drive task carries Radar risk metadata:
|
|
941
|
+
|
|
942
|
+
- Intervention level (suggesting/planning/authorizing/executing)
|
|
943
|
+
- Change risk scores (blast radius, coupling, regression, architectural tension)
|
|
944
|
+
- Finding confidence and evidence sources
|
|
945
|
+
- Verification plan reference
|
|
946
|
+
|
|
947
|
+
The user's AI assistant (operating through Drive) can use this metadata to calibrate its own behavior — being more careful with high-risk tasks and more autonomous with low-risk ones.
|
|
948
|
+
|
|
949
|
+
### The Handoff Principle
|
|
950
|
+
|
|
951
|
+
*Radar proposes. Drive disposes.*
|
|
952
|
+
|
|
953
|
+
Radar Transform produces the most informed, risk-scored, dependency-ordered remediation plan it can. Then it stops. The user's AI assistant, with human oversight through Drive's checkpoint system, executes. The separation is absolute. Radar never crosses into execution territory.
|
|
954
|
+
|
|
955
|
+
---
|
|
956
|
+
|
|
957
|
+
## Pattern Corpus & Feedback Loop
|
|
958
|
+
|
|
959
|
+
The long-term value of Radar is not any single audit. It is the accumulation of verified architectural knowledge over time.
|
|
960
|
+
|
|
961
|
+
### The Feedback Loop
|
|
962
|
+
|
|
963
|
+
```
|
|
964
|
+
Anti-Pattern (found) → Correct Pattern (prescribed) → Verified Improvement (confirmed)
|
|
965
|
+
```
|
|
966
|
+
|
|
967
|
+
Every time a Radar audit finds a problem and Transform produces a remediation that is successfully applied and verified, the system accumulates:
|
|
968
|
+
|
|
969
|
+
1. **A confirmed anti-pattern** — with real-world evidence, not textbook examples
|
|
970
|
+
2. **A proven remediation** — with framework-specific implementation, not generic advice
|
|
971
|
+
3. **A verification methodology** — how to prove the fix works, not just that it compiles
|
|
972
|
+
|
|
973
|
+
### The Pattern Corpus
|
|
974
|
+
|
|
975
|
+
Over time, this accumulates into a proprietary failure-pattern corpus:
|
|
976
|
+
|
|
977
|
+
- **Cross-project patterns** — "This authentication anti-pattern appears in 40% of Laravel codebases we've audited"
|
|
978
|
+
- **Framework-specific remediations** — "This is how to fix unbounded retries in Express.js, validated across 12 projects"
|
|
979
|
+
- **Risk calibration data** — "Changes to authentication middleware have a 23% regression rate without dedicated test coverage"
|
|
980
|
+
- **Verified migrations** — "Moving from inline SQL to Eloquent ORM: these are the 7 things that break"
|
|
981
|
+
|
|
982
|
+
### Why This Matters
|
|
983
|
+
|
|
984
|
+
Most AI systems analyze code in isolation. Every audit starts from zero. Radar rehabilitates — and each rehabilitation makes the next one better.
|
|
985
|
+
|
|
986
|
+
The pattern corpus is the moat. The difference between a generic "you have SQL injection" and "we've fixed this exact pattern 47 times, here is the framework-specific playbook with a 94% first-attempt success rate" is the difference between a tool and an institution.
|
|
987
|
+
|
|
988
|
+
---
|
|
989
|
+
|
|
990
|
+
## The Formal Epistemic Schema
|
|
991
|
+
|
|
992
|
+
This is the intellectual core of Radar — the formal spine that prevents the system from becoming a pile of clever prompts.
|
|
993
|
+
|
|
994
|
+
### Core Principle
|
|
995
|
+
|
|
996
|
+
**All findings must be decomposed into epistemic layers.** No agent is allowed to output a conclusion without explicitly passing through these layers.
|
|
997
|
+
|
|
998
|
+
### The 7-Layer Epistemic Stack
|
|
999
|
+
|
|
1000
|
+
Every finding is a structured object with seven layers:
|
|
1001
|
+
|
|
1002
|
+
#### Layer 1 — Observation (Raw Signal)
|
|
1003
|
+
|
|
1004
|
+
What exists independently of interpretation.
|
|
1005
|
+
|
|
1006
|
+
- "Function `retryRequest()` retries on HTTP 500"
|
|
1007
|
+
- "Config flag `ENABLE_LEGACY_FLOW=true` in prod"
|
|
1008
|
+
- "Table `users` lacks a unique constraint on email"
|
|
1009
|
+
|
|
1010
|
+
**Rules:** No adjectives. No risk language. Tool outputs live here.
|
|
1011
|
+
|
|
1012
|
+
#### Layer 2 — Evidence Source
|
|
1013
|
+
|
|
1014
|
+
Why we believe the observation is real.
|
|
1015
|
+
|
|
1016
|
+
Fields:
|
|
1017
|
+
- Source type (static analysis, config file, runtime metric, log, commit history)
|
|
1018
|
+
- Tool or artifact name
|
|
1019
|
+
- Location (file, line, environment)
|
|
1020
|
+
- Freshness (static / historical / live)
|
|
1021
|
+
|
|
1022
|
+
**Purpose:** Prevents tool bias, hallucinated certainty, and overweighting single sources.
|
|
1023
|
+
|
|
1024
|
+
#### Layer 3 — Interpretation (Mechanism)
|
|
1025
|
+
|
|
1026
|
+
What this observation means in context.
|
|
1027
|
+
|
|
1028
|
+
- "Unbounded retries can amplify load during partial outages"
|
|
1029
|
+
- "Legacy flow bypasses new validation logic"
|
|
1030
|
+
- "Duplicate emails can be created under race conditions"
|
|
1031
|
+
|
|
1032
|
+
**Rules:** Must explain causal mechanism. No value judgment yet. Multiple interpretations allowed.
|
|
1033
|
+
|
|
1034
|
+
#### Layer 4 — Assumptions
|
|
1035
|
+
|
|
1036
|
+
What must be true for the interpretation to hold.
|
|
1037
|
+
|
|
1038
|
+
- "Service receives concurrent requests"
|
|
1039
|
+
- "Flag is enabled in all regions"
|
|
1040
|
+
- "Email uniqueness is required by business logic"
|
|
1041
|
+
|
|
1042
|
+
**This is where the Devil's Advocate attacks.**
|
|
1043
|
+
|
|
1044
|
+
#### Layer 5 — Risk Statement
|
|
1045
|
+
|
|
1046
|
+
What could go wrong if the interpretation is correct.
|
|
1047
|
+
|
|
1048
|
+
Format: *If [interpretation], then [failure mode], impacting [asset]*
|
|
1049
|
+
|
|
1050
|
+
Example: "Retry storms could overwhelm downstream services, causing cascading failures"
|
|
1051
|
+
|
|
1052
|
+
#### Layer 6 — Impact & Likelihood
|
|
1053
|
+
|
|
1054
|
+
Severity modeling, not vibes.
|
|
1055
|
+
|
|
1056
|
+
Fields:
|
|
1057
|
+
- **Impact domain:** security, data integrity, availability, compliance, velocity
|
|
1058
|
+
- **Impact magnitude:** low, moderate, high, critical, existential
|
|
1059
|
+
- **Likelihood:** rare, unlikely, possible, likely, frequent
|
|
1060
|
+
- **Time horizon:** immediate, near-term, long-term, hypothetical
|
|
1061
|
+
- **Blast radius:** localized, service-level, systemic, org-wide/legal/existential
|
|
1062
|
+
|
|
1063
|
+
#### Layer 7 — Judgment (Decision-Oriented)
|
|
1064
|
+
|
|
1065
|
+
What should be done about it.
|
|
1066
|
+
|
|
1067
|
+
Options: Must fix | Should fix | Accept risk | Monitor | Out of scope
|
|
1068
|
+
|
|
1069
|
+
**Rules:**
|
|
1070
|
+
- Judgment is explicitly separated from facts
|
|
1071
|
+
- Principal Engineer owns this layer
|
|
1072
|
+
- Devil's Advocate may challenge but not decide
|
|
1073
|
+
|
|
1074
|
+
### Confidence Modeling
|
|
1075
|
+
|
|
1076
|
+
Each finding carries a **confidence vector**, not a scalar.
|
|
1077
|
+
|
|
1078
|
+
**Confidence dimensions:**
|
|
1079
|
+
- Evidence diversity (1 tool vs many)
|
|
1080
|
+
- Signal freshness (static vs runtime)
|
|
1081
|
+
- Assumption fragility
|
|
1082
|
+
- Historical precedent (known failure pattern?)
|
|
1083
|
+
|
|
1084
|
+
This enables statements like: "High-impact, low-confidence risk — validate before remediation." That's senior-level nuance.
|
|
1085
|
+
|
|
1086
|
+
### Epistemic Hygiene Rules (System-Wide Invariants)
|
|
1087
|
+
|
|
1088
|
+
These are non-negotiable:
|
|
1089
|
+
|
|
1090
|
+
1. No risk statements without observations
|
|
1091
|
+
2. No judgments without risk modeling
|
|
1092
|
+
3. No confidence without evidence
|
|
1093
|
+
4. No synthesis without acknowledging uncertainty
|
|
1094
|
+
5. No "clean narrative" without Devil's Advocate response
|
|
1095
|
+
|
|
1096
|
+
---
|
|
1097
|
+
|
|
1098
|
+
## Disagreement Resolution System
|
|
1099
|
+
|
|
1100
|
+
### Why Disagreements Happen
|
|
1101
|
+
|
|
1102
|
+
Usually because of:
|
|
1103
|
+
- Different threat models
|
|
1104
|
+
- Different time horizons
|
|
1105
|
+
- Different failure memories
|
|
1106
|
+
- Different tolerance for risk
|
|
1107
|
+
|
|
1108
|
+
Radar surfaces these differences. It does not hide them.
|
|
1109
|
+
|
|
1110
|
+
### Resolution Mental Models
|
|
1111
|
+
|
|
1112
|
+
Senior engineers do not vote. They reason under uncertainty using five canonical models:
|
|
1113
|
+
|
|
1114
|
+
**Model 1 — Evidence Dominance**
|
|
1115
|
+
"Which claim is better supported by independent signals?"
|
|
1116
|
+
Weight: number of tools, signal diversity, historical precedent.
|
|
1117
|
+
|
|
1118
|
+
**Model 2 — Risk Asymmetry**
|
|
1119
|
+
"If we're wrong, who pays and how badly?"
|
|
1120
|
+
Security and data risks often override performance disagreements.
|
|
1121
|
+
|
|
1122
|
+
**Model 3 — Reversibility**
|
|
1123
|
+
"How hard is it to undo this decision?"
|
|
1124
|
+
Irreversible decisions get stricter scrutiny.
|
|
1125
|
+
|
|
1126
|
+
**Model 4 — Time-to-Failure**
|
|
1127
|
+
"Which concern manifests first?"
|
|
1128
|
+
Near-term risks outrank theoretical long-term ones.
|
|
1129
|
+
|
|
1130
|
+
**Model 5 — Blast Radius**
|
|
1131
|
+
"How much breaks if this is wrong?"
|
|
1132
|
+
Localized risk < systemic risk.
|
|
1133
|
+
|
|
1134
|
+
### Disagreement as First-Class Object
|
|
1135
|
+
|
|
1136
|
+
Every disagreement is a structured record:
|
|
1137
|
+
|
|
1138
|
+
```
|
|
1139
|
+
Disagreement {
|
|
1140
|
+
id
|
|
1141
|
+
finding_id
|
|
1142
|
+
epistemic_layer_disputed // interpretation, assumptions, impact, likelihood, judgment
|
|
1143
|
+
agents_involved
|
|
1144
|
+
positions[] // one per agent, each with claim + evidence + assumptions + confidence
|
|
1145
|
+
root_cause // from closed set (see below)
|
|
1146
|
+
resolution_model_applied
|
|
1147
|
+
principal_response // REQUIRED - silence is not allowed
|
|
1148
|
+
status // open, mitigated, accepted_risk, deferred, out_of_scope
|
|
1149
|
+
}
|
|
1150
|
+
```
|
|
1151
|
+
|
|
1152
|
+
**Root cause taxonomy (closed set):**
|
|
1153
|
+
- Threat model mismatch
|
|
1154
|
+
- Time horizon mismatch
|
|
1155
|
+
- Evidence availability mismatch
|
|
1156
|
+
- Risk tolerance mismatch
|
|
1157
|
+
- Domain boundary mismatch
|
|
1158
|
+
- Optimism vs pessimism bias
|
|
1159
|
+
- Tool trust bias
|
|
1160
|
+
|
|
1161
|
+
### Resolution Protocol
|
|
1162
|
+
|
|
1163
|
+
The Principal must explicitly respond to every Devil's Advocate critique and every unresolved disagreement:
|
|
1164
|
+
|
|
1165
|
+
1. Acknowledge it explicitly
|
|
1166
|
+
2. Choose a resolution model (evidence dominance, risk asymmetry, reversibility, time-to-failure, accept risk)
|
|
1167
|
+
3. Record rationale
|
|
1168
|
+
4. Assign follow-up if needed
|
|
1169
|
+
|
|
1170
|
+
**Status states:** Open | Mitigated by evidence | Accepted risk | Deferred pending validation | Out of scope
|
|
1171
|
+
|
|
1172
|
+
No silent disappearance allowed.
|
|
1173
|
+
|
|
1174
|
+
### Critical Anti-Patterns
|
|
1175
|
+
|
|
1176
|
+
1. Auto-resolving disagreements
|
|
1177
|
+
2. Averaging opinions
|
|
1178
|
+
3. Forcing consensus language
|
|
1179
|
+
4. Hiding disagreement in footnotes
|
|
1180
|
+
5. Treating Devil's Advocate as optional
|
|
1181
|
+
|
|
1182
|
+
These destroy trust.
|
|
1183
|
+
|
|
1184
|
+
---
|
|
1185
|
+
|
|
1186
|
+
## Disagreement Visualization Model
|
|
1187
|
+
|
|
1188
|
+
### Philosophy
|
|
1189
|
+
|
|
1190
|
+
Agreement is cheap. Disagreement is where risk hides.
|
|
1191
|
+
|
|
1192
|
+
The goal is not convergence. The goal is **epistemic transparency** — showing leadership where understanding is weakest relative to risk.
|
|
1193
|
+
|
|
1194
|
+
### Five Visualization Axes
|
|
1195
|
+
|
|
1196
|
+
**Axis 1 — Severity vs Disagreement Intensity**
|
|
1197
|
+
|
|
1198
|
+
| | Low Disagreement | High Disagreement |
|
|
1199
|
+
|---|---|---|
|
|
1200
|
+
| **High Severity** | ACT | CRITICAL ATTENTION |
|
|
1201
|
+
| **Low Severity** | Ignore | Investigate lightly |
|
|
1202
|
+
|
|
1203
|
+
**Axis 2 — Confidence Asymmetry**
|
|
1204
|
+
Shows when Agent A has high confidence and Agent B has low confidence, revealing overconfidence risk and evidence imbalance.
|
|
1205
|
+
|
|
1206
|
+
**Axis 3 — Evidence Diversity**
|
|
1207
|
+
How many independent evidence sources support each side. Senior engineers often weight one runtime log + historical incident over three static tool outputs.
|
|
1208
|
+
|
|
1209
|
+
**Axis 4 — Time Horizon**
|
|
1210
|
+
Disagreements often aren't about *if*, but *when*. Visualizes: immediate, near-term, long-term, hypothetical.
|
|
1211
|
+
|
|
1212
|
+
**Axis 5 — Blast Radius**
|
|
1213
|
+
Localized, service-level, systemic, org-wide/legal/existential. This axis often breaks ties.
|
|
1214
|
+
|
|
1215
|
+
### Canonical Views
|
|
1216
|
+
|
|
1217
|
+
**1. Disagreement Heatmap (Executive View)**
|
|
1218
|
+
Rows = findings, color intensity = severity x disagreement x confidence gap. Tells leadership: "Where are we least sure about the most important things?"
|
|
1219
|
+
|
|
1220
|
+
**2. Epistemic Stack Diff (Per Finding)**
|
|
1221
|
+
Shows which layers are agreed vs disputed:
|
|
1222
|
+
```
|
|
1223
|
+
Observation [agreed]
|
|
1224
|
+
Interpretation [disputed]
|
|
1225
|
+
Assumptions [disputed]
|
|
1226
|
+
Impact [disputed]
|
|
1227
|
+
Judgment [deferred]
|
|
1228
|
+
```
|
|
1229
|
+
|
|
1230
|
+
**3. Agent Position Overlay**
|
|
1231
|
+
For high-risk findings, plots each agent's position across impact, likelihood, and time horizon. Clusters vs outliers — outliers matter.
|
|
1232
|
+
|
|
1233
|
+
**4. Devil's Advocate Focus Panel**
|
|
1234
|
+
Dedicated view: findings where Devil's Advocate dissents, confidence vs evidence delta, Principal's response. If this panel is empty, the system is broken.
|
|
1235
|
+
|
|
1236
|
+
**5. Assumption Fragility Graph**
|
|
1237
|
+
Shows which assumptions multiple conclusions rely on and which are weakest/unverified. Identifies single-point epistemic failures.
|
|
1238
|
+
|
|
1239
|
+
---
|
|
1240
|
+
|
|
1241
|
+
## Reality Gap Framework
|
|
1242
|
+
|
|
1243
|
+
### Definition
|
|
1244
|
+
|
|
1245
|
+
**Reality Gap = Difference between system behavior as inferred from code and behavior as it actually executes in production.**
|
|
1246
|
+
|
|
1247
|
+
Most incidents live here. Most audits miss it entirely.
|
|
1248
|
+
|
|
1249
|
+
### The Four Divergence Vectors
|
|
1250
|
+
|
|
1251
|
+
1. **Configuration** — what's configured differently than the code assumes
|
|
1252
|
+
2. **Environment** — what's different between dev/staging/prod
|
|
1253
|
+
3. **Runtime Control Planes** — what external systems alter behavior
|
|
1254
|
+
4. **Human Intervention** — what manual processes bypass safeguards
|
|
1255
|
+
|
|
1256
|
+
### Reality Gap Domains
|
|
1257
|
+
|
|
1258
|
+
**RG-1: Configuration Drift**
|
|
1259
|
+
- Environment variables, YAML/JSON/HCL configs
|
|
1260
|
+
- Default vs overridden values
|
|
1261
|
+
- Secrets managers, region-specific configs
|
|
1262
|
+
- Failure patterns: safe defaults overridden unsafely, flags enabled in prod only, test env != prod env
|
|
1263
|
+
|
|
1264
|
+
**RG-2: Feature Flags & Kill Switches**
|
|
1265
|
+
- Flag inventory, ownership, lifetimes
|
|
1266
|
+
- Conditional code paths
|
|
1267
|
+
- Failure patterns: permanent "temporary" flags, untested flag combinations, flag-dependent logic bypassing invariants
|
|
1268
|
+
- Senior question: "What code runs *only* when things go wrong?"
|
|
1269
|
+
|
|
1270
|
+
**RG-3: Deployment & Infrastructure Overlay**
|
|
1271
|
+
- Kubernetes manifests, Terraform/CloudFormation
|
|
1272
|
+
- Sidecars, proxies, service meshes, init containers, CronJobs
|
|
1273
|
+
- Failure patterns: resource limits different from assumptions, hidden retries in proxies, timeouts enforced outside app code
|
|
1274
|
+
|
|
1275
|
+
**RG-4: Runtime Behavior vs Static Intent**
|
|
1276
|
+
- Logs vs code paths, metrics vs expectations
|
|
1277
|
+
- Observability coverage gaps, disabled instrumentation
|
|
1278
|
+
- Failure patterns: dead code paths that are actually live, code never exercised in tests but hot in prod, silent failure paths
|
|
1279
|
+
|
|
1280
|
+
**RG-5: Operational Overrides & Human Actions**
|
|
1281
|
+
- Hotfix mechanisms, manual scripts, admin endpoints
|
|
1282
|
+
- One-off migrations, emergency patches
|
|
1283
|
+
- Failure patterns: undocumented operational workflows, scripts with production authority, manual fixes that bypass safeguards
|
|
1284
|
+
- This is where tribal knowledge hides
|
|
1285
|
+
|
|
1286
|
+
### Reality Gap Agent Outputs
|
|
1287
|
+
|
|
1288
|
+
1. Assumed vs Actual Behavior Table
|
|
1289
|
+
2. Code Paths Active Only in Production
|
|
1290
|
+
3. Flags & Configs That Change Control Flow
|
|
1291
|
+
4. Invisible Dependencies (proxies, retries, meshes)
|
|
1292
|
+
5. Highest-Risk Mismatches
|
|
1293
|
+
|
|
1294
|
+
Reality Gap findings are encoded using the same 7-layer epistemic schema. Uniform representation enables powerful synthesis.
|
|
1295
|
+
|
|
1296
|
+
---
|
|
1297
|
+
|
|
1298
|
+
## Language-Specific Failure Models
|
|
1299
|
+
|
|
1300
|
+
Each language/runtime has unique ways to fail that generic analysis will never catch. Radar detects the project's language(s) and applies ecosystem-specific failure pattern catalogs.
|
|
1301
|
+
|
|
1302
|
+
**Core principle:** Every runtime lies to you in a different way.
|
|
1303
|
+
|
|
1304
|
+
### JVM (Java / Kotlin / Scala)
|
|
1305
|
+
|
|
1306
|
+
Hidden failures: GC pressure from object churn, thread pool exhaustion, blocking calls inside async/reactive flows, memory leaks via static references, classloader leaks, poor equals/hashCode implementations, overuse of synchronized vs fine-grained locks.
|
|
1307
|
+
|
|
1308
|
+
Audit questions: Are allocations proportional to request volume? Are thread pools bounded and observable? Are blocking I/O calls hidden in async paths? Are caches unbounded?
|
|
1309
|
+
|
|
1310
|
+
### Python
|
|
1311
|
+
|
|
1312
|
+
Hidden failures: GIL-induced throughput collapse, async code that isn't actually async, silent exception swallowing, mutable default arguments, heavy reliance on global state, memory leaks via reference cycles, CPU-bound work in request threads.
|
|
1313
|
+
|
|
1314
|
+
Audit questions: Is concurrency real or illusionary? Are asyncio boundaries respected? Is CPU work isolated? Are retries idempotent?
|
|
1315
|
+
|
|
1316
|
+
### JavaScript / TypeScript (Node, Browser)
|
|
1317
|
+
|
|
1318
|
+
Hidden failures: Event loop blocking, unhandled promise rejections, inconsistent async error handling, memory leaks via closures, excessive JSON serialization, dependency bloat (supply chain risk), TypeScript "any" erosion.
|
|
1319
|
+
|
|
1320
|
+
Audit questions: Can one slow request stall all others? Are async errors centrally handled? Is type safety enforced or aspirational? Are libraries pinned and audited?
|
|
1321
|
+
|
|
1322
|
+
### Go
|
|
1323
|
+
|
|
1324
|
+
Hidden failures: Goroutine leaks, context cancellation ignored, channel deadlocks, unbounded fan-out, hidden blocking syscalls, overuse of global state.
|
|
1325
|
+
|
|
1326
|
+
Audit questions: Are all goroutines bounded? Is context propagated everywhere? Are channels closed correctly? Is backpressure implemented?
|
|
1327
|
+
|
|
1328
|
+
### Rust
|
|
1329
|
+
|
|
1330
|
+
Hidden failures: Unsafe blocks without justification, overly complex lifetimes (maintainability risk), panic paths in prod, blocking calls in async runtimes, premature optimization.
|
|
1331
|
+
|
|
1332
|
+
Audit questions: Why is unsafe needed here? Are panics recoverable? Is async runtime respected? Is complexity justified?
|
|
1333
|
+
|
|
1334
|
+
### Databases (Cross-Cutting)
|
|
1335
|
+
|
|
1336
|
+
Hidden failures: N+1 queries, lock contention, missing indexes, overloaded migrations, weak isolation assumptions.
|
|
1337
|
+
|
|
1338
|
+
### Distributed Systems (Cross-Language)
|
|
1339
|
+
|
|
1340
|
+
Hidden failures: Clock skew, partial failures, retry storms, cascading timeouts, eventual consistency violations.
|
|
1341
|
+
|
|
1342
|
+
---
|
|
1343
|
+
|
|
1344
|
+
## Tooling Stack
|
|
1345
|
+
|
|
1346
|
+
### Must-Have (Core)
|
|
1347
|
+
|
|
1348
|
+
| Tool | What It Does | Domains Served | Cost |
|
|
1349
|
+
|---|---|---|---|
|
|
1350
|
+
| **SonarQube** | Code smells, bugs, maintainability, duplication, complexity analysis | 1, 3, 6, 9 | Free (Community Edition) |
|
|
1351
|
+
| **Semgrep** | Security-focused SAST — XSS, SQL injection, IDOR, hardcoded secrets, business logic vulnerabilities. 20,000+ rules, 30+ languages, 10-second median scan | 1, 3, 4, 5, 6, 9 | Free (OSS) / Paid (Pro) |
|
|
1352
|
+
| **Trivy** | All-in-one security scanner — OS packages, app dependencies, IaC files, license compliance | 4, 5 | Free |
|
|
1353
|
+
| **Gitleaks** | Secrets detection — scans git history for API keys, passwords, tokens. 160+ secret patterns | 4, 5 | Free |
|
|
1354
|
+
|
|
1355
|
+
### High Value
|
|
1356
|
+
|
|
1357
|
+
| Tool | What It Does | Domains Served | Cost |
|
|
1358
|
+
|---|---|---|---|
|
|
1359
|
+
| **Checkov** | IaC security scanner — Terraform, CloudFormation, K8s, Helm, Dockerfiles. 3,000+ policies covering CIS benchmarks | 4, 5 | Free |
|
|
1360
|
+
| **Syft** | SBOM generation — complete package inventory across all ecosystems (containers, filesystems, archives) | 4, 5 | Free |
|
|
1361
|
+
| **Grype** | Vulnerability scanning — matches SBOM inventory against CVE databases. Paired with Syft for full supply chain analysis | 4, 5 | Free |
|
|
1362
|
+
| **Git History Miner** | Git log mining — file churn rates, author concentration, change coupling, file age vs modification frequency | 11, 12 | Free (built-in) |
|
|
1363
|
+
|
|
1364
|
+
### Optional (Paid Enhancement)
|
|
1365
|
+
|
|
1366
|
+
| Tool | What It Does | Domains Served | Cost |
|
|
1367
|
+
|---|---|---|---|
|
|
1368
|
+
| **CodeScene** | Hotspot analysis, code churn + complexity correlation, author concentration, change coupling, knowledge distribution, CodeHealth score (25+ factors) | 1, 9, 11, 12 | ~EUR 18/mo/author |
|
|
1369
|
+
|
|
1370
|
+
### Useful
|
|
1371
|
+
|
|
1372
|
+
| Tool | What It Does | Domains Served | Cost |
|
|
1373
|
+
|---|---|---|---|
|
|
1374
|
+
| **OpenSSF Scorecard** | Scores open source projects 0-10 on security heuristics (branch protection, dependency pinning, CI tests, vulnerability disclosure). Scans top 1M projects weekly | 4, 12 | Free |
|
|
1375
|
+
| **depcruise / Madge** | JavaScript/TypeScript dependency graph visualization and validation | 1 | Free |
|
|
1376
|
+
| **Language linters** | ESLint (JS/TS), Pylint (Python), RuboCop (Ruby), Clippy (Rust), staticcheck (Go), SpotBugs (JVM) | 3, 9 | Free |
|
|
1377
|
+
|
|
1378
|
+
### Deferred (Tier 3)
|
|
1379
|
+
|
|
1380
|
+
| Tool | What It Does | Notes |
|
|
1381
|
+
|---|---|---|
|
|
1382
|
+
| **Structure101 / Lattix** | Architecture visualization and dependency management | Enterprise-licensed, expensive. Claude can do much of this analysis directly for codebases under 100k LOC |
|
|
1383
|
+
| **CodeClimate** | Automated code review for maintainability | Overlaps significantly with SonarQube |
|
|
1384
|
+
| **Snyk** | Full platform (SAST, SCA, container, IaC) | Enterprise SaaS. Trivy covers most use cases for free |
|
|
1385
|
+
|
|
1386
|
+
### Custom (Future Runtime)
|
|
1387
|
+
|
|
1388
|
+
**Signal Normalization Layer**
|
|
1389
|
+
Tools speak different languages. The signal schema (already specified) converts all findings into: severity, confidence, blast radius, domain relevance.
|
|
1390
|
+
|
|
1391
|
+
**Cross-Signal Correlation Engine**
|
|
1392
|
+
Where the system becomes exceptional:
|
|
1393
|
+
- High churn + low tests = change risk
|
|
1394
|
+
- Async code + blocking calls = latent perf bug
|
|
1395
|
+
- PII fields + logs = compliance risk
|
|
1396
|
+
|
|
1397
|
+
---
|
|
1398
|
+
|
|
1399
|
+
## Output Format
|
|
1400
|
+
|
|
1401
|
+
Radar output spans three layers (see [The Three Output Layers](#the-three-output-layers)). The Core diagnostic pipeline (Phases 0-5) produces Layer A. The Transform pipeline (Phases 6-8) produces Layers B and C.
|
|
1402
|
+
|
|
1403
|
+
### Layer A — Diagnostic Report
|
|
1404
|
+
|
|
1405
|
+
Every Radar Core run produces:
|
|
1406
|
+
|
|
1407
|
+
1. **Executive Risk Summary** — One-page overview for leadership
|
|
1408
|
+
2. **Architecture Narrative** — How the system is built and why
|
|
1409
|
+
3. **Findings by Domain (Severity-Ranked)** — All 14 domains, each with epistemic-schema-structured findings
|
|
1410
|
+
4. **Cross-Validation Notes** — All disagreements, their root causes, and the Principal's resolutions
|
|
1411
|
+
5. **Remediation Roadmap** — Prioritized action plan
|
|
1412
|
+
6. **Long-Term Structural Risks** — What degrades over time
|
|
1413
|
+
7. **"What Would Break First at 10x Scale"** — Predictive failure analysis
|
|
1414
|
+
|
|
1415
|
+
### Layer B — Remediation Knowledge (Transform)
|
|
1416
|
+
|
|
1417
|
+
When the Transform pipeline runs, it additionally produces:
|
|
1418
|
+
|
|
1419
|
+
- **Remediation Playbooks** — Per-finding playbooks at all 4 transformation layers (abstract → framework → language → project)
|
|
1420
|
+
- **Best-Practice Patterns** — Correct patterns paired with detected anti-patterns
|
|
1421
|
+
- **Educational Context** — Before/after examples and rationale for AI-assisted developers
|
|
1422
|
+
- **Generated Guardrails** — Project rules (`.claude/CLAUDE.md`, `.cursorrules`, linter configs)
|
|
1423
|
+
|
|
1424
|
+
### Layer C — Change Orchestration (Transform)
|
|
1425
|
+
|
|
1426
|
+
The execution planning phase produces:
|
|
1427
|
+
|
|
1428
|
+
- **Change Dependency Graph** — Dependency-ordered sequence of all proposed changes
|
|
1429
|
+
- **Risk Assessment Report** — Per-change scores across blast radius, coupling, regression, and architectural tension
|
|
1430
|
+
- **Verification Plan** — Per-change verification steps with expected outcomes and rollback criteria
|
|
1431
|
+
- **Drive Project** — Complete, ready-to-execute Drive project with phased plans, risk metadata, and verification gates
|
|
1432
|
+
|
|
1433
|
+
---
|
|
1434
|
+
|
|
1435
|
+
## Signal Categories
|
|
1436
|
+
|
|
1437
|
+
Phase 1 gathers evidence across six orthogonal dimensions:
|
|
1438
|
+
|
|
1439
|
+
1. **Structure** — Static code analysis, architecture patterns, dependency graphs
|
|
1440
|
+
2. **Behavior** — Performance profiles, async analysis, flamegraphs
|
|
1441
|
+
3. **History** — Git churn, file age, author concentration, bug density over time
|
|
1442
|
+
4. **Dependencies** — Vulnerability scans, SBOM, supply chain health scores
|
|
1443
|
+
5. **Policy Posture** — IaC compliance, secrets exposure, governance policies
|
|
1444
|
+
6. **Runtime Contracts** — OpenAPI/gRPC schemas, backward compatibility checks
|
|
1445
|
+
|
|
1446
|
+
---
|
|
1447
|
+
|
|
1448
|
+
## v0.1 Status
|
|
1449
|
+
|
|
1450
|
+
Radar v0.1.0 is a **validated specification set** — 90 markdown files across 8 component types that define the complete multi-agent audit system.
|
|
1451
|
+
|
|
1452
|
+
| Component Type | Count | Location |
|
|
1453
|
+
|----------------|-------|----------|
|
|
1454
|
+
| Domains | 14 | `src/domains/` |
|
|
1455
|
+
| Schemas | 9 (5 shared + 4 transform) | `src/schemas/` + `src/transform/schemas/` |
|
|
1456
|
+
| Rules | 5 (3 shared + 2 transform) | `src/rules/` + `src/transform/rules/` |
|
|
1457
|
+
| Tools | 8 | `src/tools/` |
|
|
1458
|
+
| Personas | 17 (12 core + 5 transform) | `src/core/personas/` + `src/transform/personas/` |
|
|
1459
|
+
| Agents | 17 (12 core + 5 transform) | `src/core/agents/` + `src/transform/agents/` |
|
|
1460
|
+
| Workflows | 12 (8 core + 4 transform) | `src/core/workflows/` + `src/transform/workflows/` |
|
|
1461
|
+
| Commands | 8 (4 core + 4 transform) | `commands/` |
|
|
1462
|
+
| **Total** | **90 files** | **~15,000 lines** |
|
|
1463
|
+
|
|
1464
|
+
### What v0.1 Is
|
|
1465
|
+
|
|
1466
|
+
The specifications are the blueprints. They define every agent's identity, every domain's knowledge, every schema's contract, every tool's integration, every workflow's orchestration, and every command's UX. All cross-references have been validated (310 references, 0 broken). All files conform to their component type conventions. A version-lock manifest (SHA-256 content hashes) provides traceability for reproducible audit compositions.
|
|
1467
|
+
|
|
1468
|
+
### What v0.1 Is Not
|
|
1469
|
+
|
|
1470
|
+
v0.1 does not include runtime execution — actually orchestrating Claude Code sessions to run audits on real codebases. That requires a session orchestration layer, tool execution runtime, artifact persistence layer, and report generation engine. The specifications tell the runtime what to build. The runtime is the next milestone.
|
|
1471
|
+
|
|
1472
|
+
---
|
|
1473
|
+
|
|
1474
|
+
## v0.2 — Installation & Runtime
|
|
1475
|
+
|
|
1476
|
+
v0.2 makes Radar installable and runnable as a Claude Code extension. The framework installs to `~/.claude/radar/` and commands install to `~/.claude/commands/radar/`, mirroring the Drive framework installation pattern.
|
|
1477
|
+
|
|
1478
|
+
### Directory Structure
|
|
1479
|
+
|
|
1480
|
+
```
|
|
1481
|
+
~/.claude/radar/ # Framework files (installed)
|
|
1482
|
+
├── core/
|
|
1483
|
+
│ ├── agents/ # 12 Core agent manifests
|
|
1484
|
+
│ ├── personas/ # 12 Core persona specs
|
|
1485
|
+
│ └── workflows/ # 8 Core orchestration workflows
|
|
1486
|
+
├── transform/
|
|
1487
|
+
│ ├── agents/ # 5 Transform agent manifests
|
|
1488
|
+
│ ├── personas/ # 5 Transform persona specs
|
|
1489
|
+
│ ├── workflows/ # 4 Transform workflows
|
|
1490
|
+
│ ├── schemas/ # 4 Transform schemas
|
|
1491
|
+
│ └── rules/ # 2 Transform rules
|
|
1492
|
+
├── domains/ # 14 audit domain knowledge modules
|
|
1493
|
+
├── schemas/ # 5 shared schemas
|
|
1494
|
+
├── rules/ # 3 shared rules
|
|
1495
|
+
└── tools/ # 8 tool adapter specifications
|
|
1496
|
+
|
|
1497
|
+
~/.claude/commands/radar/ # Slash commands (installed)
|
|
1498
|
+
├── audit.md # /radar:audit
|
|
1499
|
+
├── resume.md # /radar:resume
|
|
1500
|
+
├── status.md # /radar:status
|
|
1501
|
+
├── report.md # /radar:report
|
|
1502
|
+
├── transform.md # /radar:transform
|
|
1503
|
+
├── remediate.md # /radar:remediate
|
|
1504
|
+
├── playbook.md # /radar:playbook
|
|
1505
|
+
└── guardrails.md # /radar:guardrails
|
|
1506
|
+
```
|
|
1507
|
+
|
|
1508
|
+
### Installation
|
|
1509
|
+
|
|
1510
|
+
**Prerequisites:** [Claude Code](https://claude.ai/code) installed (`~/.claude/` directory must exist)
|
|
1511
|
+
|
|
1512
|
+
**One-command install (recommended):**
|
|
1513
|
+
|
|
1514
|
+
```bash
|
|
1515
|
+
curl -sSL https://raw.githubusercontent.com/accelerationguy/accelerate/main/modules/radar/install.sh | bash
|
|
1516
|
+
```
|
|
1517
|
+
|
|
1518
|
+
**Or clone and install locally:**
|
|
1519
|
+
|
|
1520
|
+
```bash
|
|
1521
|
+
git clone https://github.com/accelerationguy/accelerate.git
|
|
1522
|
+
cd radar
|
|
1523
|
+
bash install.sh
|
|
1524
|
+
```
|
|
1525
|
+
|
|
1526
|
+
The installer:
|
|
1527
|
+
1. Copies 82 framework files to `~/.claude/radar/`
|
|
1528
|
+
2. Installs 8 slash commands to `~/.claude/commands/radar/`
|
|
1529
|
+
3. Walks you through installing each OSS analysis tool (Y/N per tool)
|
|
1530
|
+
4. Verifies each installed tool works
|
|
1531
|
+
5. Auto-skips tools already installed on re-run
|
|
1532
|
+
|
|
1533
|
+
**Tool installation methods:**
|
|
1534
|
+
- **Python tools** (Semgrep, Checkov): Installed via venv to `~/.local/share/radar/venvs/` — no sudo, PEP 668 compatible
|
|
1535
|
+
- **Binary tools** (Trivy, Gitleaks, Syft, Grype): Downloaded via curl to `~/.local/bin/`
|
|
1536
|
+
- **SonarQube**: Guided setup — Docker (local server) or SonarQube Cloud, with smart detection of existing Docker installations
|
|
1537
|
+
- **Git history**: Always available (uses built-in git commands)
|
|
1538
|
+
|
|
1539
|
+
After installation, run `/radar:audit` in Claude Code to start your first audit.
|
|
1540
|
+
|
|
1541
|
+
For a complete walkthrough, see [Getting Started](docs/GETTING-STARTED.md).
|
|
1542
|
+
|
|
1543
|
+
### v0.2 Phases
|
|
1544
|
+
|
|
1545
|
+
| Phase | Name | Status |
|
|
1546
|
+
|-------|------|--------|
|
|
1547
|
+
| 9 | Command Conversion | Complete — commands relocated, `@` references rewritten, `allowed-tools` added |
|
|
1548
|
+
| 10 | Install System | Complete — interactive installer with dual-mode (local + curl\|bash), venv-based Python tools, smart SonarQube detection |
|
|
1549
|
+
| 11 | Project Init & Validation | Complete — `/radar:init` + `/radar:validate` commands, `.radar/` project setup |
|
|
1550
|
+
| 12 | Multi-Session UX | Complete — phase checkpoints (continue/pause/abort), session tracking, estimated remaining work |
|
|
1551
|
+
| 13 | Getting Started | Complete — Getting Started guide, command reference |
|
|
1552
|
+
|