agentboot 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (78) hide show
  1. package/.github/ISSUE_TEMPLATE/persona-request.md +62 -0
  2. package/.github/ISSUE_TEMPLATE/quality-feedback.md +67 -0
  3. package/.github/workflows/cla.yml +25 -0
  4. package/.github/workflows/validate.yml +49 -0
  5. package/.idea/agentboot.iml +9 -0
  6. package/.idea/misc.xml +6 -0
  7. package/.idea/modules.xml +8 -0
  8. package/.idea/vcs.xml +6 -0
  9. package/CLA.md +98 -0
  10. package/CLAUDE.md +230 -0
  11. package/CONTRIBUTING.md +168 -0
  12. package/LICENSE +191 -0
  13. package/NOTICE +4 -0
  14. package/PERSONAS.md +156 -0
  15. package/README.md +172 -0
  16. package/agentboot.config.json +207 -0
  17. package/bin/agentboot.js +17 -0
  18. package/core/gotchas/README.md +35 -0
  19. package/core/instructions/baseline.instructions.md +133 -0
  20. package/core/instructions/security.instructions.md +186 -0
  21. package/core/personas/code-reviewer/SKILL.md +175 -0
  22. package/core/personas/code-reviewer/persona.config.json +11 -0
  23. package/core/personas/security-reviewer/SKILL.md +233 -0
  24. package/core/personas/security-reviewer/persona.config.json +11 -0
  25. package/core/personas/test-data-expert/SKILL.md +234 -0
  26. package/core/personas/test-data-expert/persona.config.json +10 -0
  27. package/core/personas/test-generator/SKILL.md +262 -0
  28. package/core/personas/test-generator/persona.config.json +10 -0
  29. package/core/traits/audit-trail.md +182 -0
  30. package/core/traits/confidence-signaling.md +172 -0
  31. package/core/traits/critical-thinking.md +129 -0
  32. package/core/traits/schema-awareness.md +132 -0
  33. package/core/traits/source-citation.md +174 -0
  34. package/core/traits/structured-output.md +199 -0
  35. package/docs/ci-cd-automation.md +548 -0
  36. package/docs/claude-code-reference/README.md +21 -0
  37. package/docs/claude-code-reference/agentboot-coverage.md +484 -0
  38. package/docs/claude-code-reference/feature-inventory.md +906 -0
  39. package/docs/cli-commands-audit.md +112 -0
  40. package/docs/cli-design.md +924 -0
  41. package/docs/concepts.md +1117 -0
  42. package/docs/config-schema-audit.md +121 -0
  43. package/docs/configuration.md +645 -0
  44. package/docs/delivery-methods.md +758 -0
  45. package/docs/developer-onboarding.md +342 -0
  46. package/docs/extending.md +448 -0
  47. package/docs/getting-started.md +298 -0
  48. package/docs/knowledge-layer.md +464 -0
  49. package/docs/marketplace.md +822 -0
  50. package/docs/org-connection.md +570 -0
  51. package/docs/plans/architecture.md +2429 -0
  52. package/docs/plans/design.md +2018 -0
  53. package/docs/plans/prd.md +1862 -0
  54. package/docs/plans/stack-rank.md +261 -0
  55. package/docs/plans/technical-spec.md +2755 -0
  56. package/docs/privacy-and-safety.md +807 -0
  57. package/docs/prompt-optimization.md +1071 -0
  58. package/docs/test-plan.md +972 -0
  59. package/docs/third-party-ecosystem.md +496 -0
  60. package/domains/compliance-template/README.md +173 -0
  61. package/domains/compliance-template/traits/compliance-aware.md +228 -0
  62. package/examples/enterprise/agentboot.config.json +184 -0
  63. package/examples/minimal/agentboot.config.json +46 -0
  64. package/package.json +63 -0
  65. package/repos.json +1 -0
  66. package/scripts/cli.ts +1069 -0
  67. package/scripts/compile.ts +1000 -0
  68. package/scripts/dev-sync.ts +149 -0
  69. package/scripts/lib/config.ts +137 -0
  70. package/scripts/lib/frontmatter.ts +61 -0
  71. package/scripts/sync.ts +687 -0
  72. package/scripts/validate.ts +421 -0
  73. package/tests/REGRESSION-PLAN.md +705 -0
  74. package/tests/TEST-PLAN.md +111 -0
  75. package/tests/cli.test.ts +705 -0
  76. package/tests/pipeline.test.ts +608 -0
  77. package/tests/validate.test.ts +278 -0
  78. package/tsconfig.json +62 -0
@@ -0,0 +1,1862 @@
1
+ # AgentBoot — Product Requirements Document
2
+
3
+ **Version:** 1.0.0-draft
4
+ **Date:** 2026-03-19
5
+ **Author:** AgentBoot Team
6
+ **License:** Apache 2.0
7
+ **Status:** Draft — awaiting review
8
+
9
+ ---
10
+
11
+ ## Table of Contents
12
+
13
+ 1. [Executive Summary](#1-executive-summary)
14
+ 2. [Problem Statement](#2-problem-statement)
15
+ 3. [Target Users](#3-target-users)
16
+ 4. [Product Vision](#4-product-vision)
17
+ 5. [Design Principles](#5-design-principles)
18
+ 6. [Feature Requirements](#6-feature-requirements)
19
+ 7. [Non-Goals](#7-non-goals)
20
+ 8. [Success Metrics](#8-success-metrics)
21
+ 9. [Open Questions](#9-open-questions)
22
+ 10. [Glossary](#10-glossary)
23
+
24
+ ---
25
+
26
+ ## 1. Executive Summary
27
+
28
+ ### What AgentBoot Is
29
+
30
+ AgentBoot is a build tool and governance framework for AI agent behavior in software
31
+ engineering organizations. It provides convention-over-configuration defaults for
32
+ composing, distributing, and managing AI personas across development teams — so that
33
+ every engineer in an organization gets consistent, high-quality AI agent behavior
34
+ from day one without building it themselves.
35
+
36
+ ### The Problem
37
+
38
+ Organizations adopting agentic development tools (Claude Code, GitHub Copilot, Cursor,
39
+ Gemini CLI) face a fragmentation crisis. Every team writes its own CLAUDE.md, its own
40
+ rules, its own agents — independently, inconsistently, and without governance. There is
41
+ no standard for distributing agent behavior across repos, no mechanism for enforcing
42
+ compliance guardrails, no way to measure whether AI personas are delivering value, and
43
+ no path for sharing improvements across teams. The result is duplicated effort, drifting
44
+ quality, compliance gaps, and developer frustration.
45
+
46
+ ### The Solution
47
+
48
+ AgentBoot applies the same discipline to AI agent behavior that Spring Boot applied to
49
+ Java application configuration: sensible defaults, composable building blocks, a scope
50
+ hierarchy that mirrors organizational structure, and a build-then-distribute pipeline
51
+ that propagates governance automatically. Persona definitions are treated as code —
52
+ version-controlled, reviewed in pull requests, tested, linted, and distributed through
53
+ a hub-and-spoke model. The framework is agent-agnostic in content (using open standards
54
+ like agentskills.io) while being Claude Code-first in delivery (leveraging the richest
55
+ extensibility surface available).
56
+
57
+ ### Target User
58
+
59
+ AgentBoot serves six distinct user segments across the organization — from individual
60
+ developers who consume personas, to platform/DevEx teams who build and distribute them,
61
+ to IT/Security teams who enforce compliance, to engineering leaders who measure ROI.
62
+ Each segment interacts with AgentBoot through a different surface (plugin, CLI, managed
63
+ settings, dashboard) tailored to their role.
64
+
65
+ ### Origin
66
+
67
+ AgentBoot started with personal and family projects — building AI personas for small
68
+ codebases where the author needed structured review, test generation, and domain
69
+ expertise. The patterns grew organically as the number of projects increased, then
70
+ were adapted for workplace use when it became clear that the same governance gaps
71
+ existed at organizational scale. What began as a personal toolkit evolved into a
72
+ framework when the realization hit: every team adopting agentic development was going
73
+ to build this anyway, and most would build it poorly.
74
+
75
+ ---
76
+
77
+ ## 2. Problem Statement
78
+
79
+ ### 2.1 The Fragmentation Problem
80
+
81
+ Every team that adopts AI coding tools ends up doing the same thing: writing a
82
+ CLAUDE.md, defining some rules, maybe adding a few custom agents. Most of these are
83
+ mediocre. New teams do not know where to start. Teams in the same organization drift
84
+ apart. Nobody shares anything.
85
+
86
+ Meanwhile, the teams that invest in building it well — composable traits, structured
87
+ reviewers, compliance guardrails, a distribution pipeline — cannot share their work
88
+ because it is buried in proprietary context.
89
+
90
+ ### 2.2 Specific Pain Points
91
+
92
+ **Scattered, inconsistent agent configuration.** In a typical organization with 20+
93
+ repositories, each repo has its own CLAUDE.md written independently. An audit of such
94
+ organizations typically reveals that 67% or more of repos have some version of the same
95
+ instructions (e.g., "use TypeScript strict mode"), but expressed in 4+ different
96
+ wordings, maintained independently, and never synchronized. When best practice evolves,
97
+ there is no mechanism to propagate the improvement.
98
+
99
+ **No governance structure.** Agent behavior is not reviewed, versioned, or audited. A
100
+ developer can change a CLAUDE.md with no pull request, no review, and no record of what
101
+ changed or why. When an agent produces a bad result, there is no diff to examine. When a
102
+ compliance rule needs to be added, there is no guarantee it reaches every repo.
103
+
104
+ **No distribution mechanism.** Organizations lack a way to push agent behavior updates
105
+ to all their repos simultaneously. Each repo is an island. A platform team that writes
106
+ a better code reviewer has no standard path to deploy it across 50 repos without
107
+ manually touching each one.
108
+
109
+ **No compliance enforcement.** Regulated industries (healthcare, finance, government)
110
+ need non-negotiable guardrails — PHI scanning, credential blocking, audit logging — that
111
+ cannot be disabled by any developer. Current agent platforms offer this capability
112
+ (managed settings, hooks), but there is no framework for generating, testing, and
113
+ deploying these guardrails systematically.
114
+
115
+ **No metrics or measurement.** Organizations cannot answer basic questions: Which
116
+ personas are used most? What do they cost? What is the false positive rate? Are they
117
+ improving over time? Without measurement, prompt optimization is guesswork, budget
118
+ justification is impossible, and the investment case for agentic development is based
119
+ on anecdote rather than evidence.
120
+
121
+ **Developer resistance.** Skeptics and hold-outs will not adopt AI tools voluntarily.
122
+ Without a mechanism for zero-friction deployment (managed settings, repo-embedded
123
+ configuration) and a privacy model that builds trust (no prompt surveillance), a
124
+ significant fraction of the engineering organization will never engage.
125
+
126
+ **No sharing or reuse.** Knowledge that one team gains — a PostgreSQL gotcha, a
127
+ compliance trait, a well-tuned reviewer — stays siloed. There is no marketplace, no
128
+ contribution model, and no standard format that enables sharing across teams or across
129
+ organizations.
130
+
131
+ **Context bloat.** CLAUDE.md files grow unbounded because no one prunes them. An audit
132
+ of a typical mid-size organization reveals average CLAUDE.md files of 230+ lines, with
133
+ the longest exceeding 800 lines. Most of this content is duplicated across repos, vague
134
+ ("follow best practices"), or stale ("TODO: update this section").
135
+
136
+ **Vendor lock-in anxiety.** Teams hesitate to invest deeply in agent configuration for
137
+ one platform because they worry about being locked in. If the organization switches
138
+ from Claude Code to Copilot (or vice versa), will they have to rebuild everything?
139
+
140
+ **No onboarding path.** New developers join the organization and face a wall of AI
141
+ tooling they have never used. They type vague prompts, get vague results, conclude
142
+ "AI does not work for me," and abandon the tools. There is no guided path from
143
+ first interaction to productive use.
144
+
145
+ ### 2.3 Why These Problems Persist
146
+
147
+ These problems persist because they are coordination problems, not technology problems.
148
+ Any single team can solve them for themselves — but the solution does not scale without
149
+ a framework. Writing a good CLAUDE.md is easy. Writing a governance system that keeps
150
+ 50 repos consistent, enforces compliance, measures effectiveness, and onboards new
151
+ developers is hard. That is what AgentBoot provides.
152
+
153
+ ---
154
+
155
+ ## 3. Target Users
156
+
157
+ AgentBoot serves six distinct user segments. Each has different needs, different pain
158
+ points, and a different definition of success.
159
+
160
+ ### 3.1 Developer
161
+
162
+ **Profile:** Individual contributor who writes code daily. May or may not already use
163
+ AI coding tools. Ranges from junior to staff level.
164
+
165
+ **Needs:**
166
+ - AI personas that work immediately when they clone a repo — zero setup
167
+ - Clear invocation patterns (`/review-code`, `/gen-tests`) that give structured results
168
+ - Privacy: confidence that their conversations, questions, and false starts are private
169
+ - The ability to ask follow-up questions about persona findings
170
+ - Control over their experience: disable tips, choose models, set preferences
171
+
172
+ **Pain points:**
173
+ - Does not know where to start with AI tools
174
+ - Gets vague results from vague prompts but does not know how to improve
175
+ - Fears being judged for questions asked to AI ("what is a foreign key?")
176
+ - Persona output is confusing — unclear severity levels, no actionable suggestions
177
+ - Multiple tools (Claude Code, Copilot, Cursor) with different configuration mechanisms
178
+
179
+ **What success looks like:**
180
+ - Clones a repo, types `/review-code`, gets a structured review with actionable findings
181
+ - Trusts that their AI conversations are private (the "IDE vs PR" mental model)
182
+ - Gradually discovers more personas and becomes more effective over weeks
183
+ - Never runs `agentboot` directly — the framework is invisible to them
184
+
185
+ ### 3.2 Platform / DevEx Team
186
+
187
+ **Profile:** Engineers responsible for developer tooling, productivity, and
188
+ infrastructure. Typically 1-5 people in a mid-size org. May be a dedicated DevEx team
189
+ or a platform engineering team with tooling responsibilities.
190
+
191
+ **Needs:**
192
+ - A central repository for managing all AI persona definitions
193
+ - A build pipeline that compiles, validates, and distributes personas
194
+ - Multi-platform support (some teams use Claude Code, others use Copilot)
195
+ - Scaffolding tools for creating new personas, traits, domains, and gotchas
196
+ - Discovery tools for finding and consolidating existing agentic content across repos
197
+ - Testing tools that validate persona behavior before deployment
198
+ - A marketplace for sharing and discovering community content
199
+
200
+ **Pain points:**
201
+ - Every team has written their own CLAUDE.md independently — massive duplication
202
+ - No way to push an update to 50 repos without touching each one
203
+ - No standard format for defining agent behavior
204
+ - Cannot test whether a persona change improves or regresses behavior
205
+ - Teams resist centralization because they fear losing control of their configuration
206
+
207
+ **What success looks like:**
208
+ - Single `agentboot build && agentboot sync` deploys updates to every registered repo
209
+ - New personas and traits go through PR review with automated validation
210
+ - `agentboot discover` consolidates scattered agentic content into a governed hub
211
+ - Teams can extend (but not break) the org-wide configuration through the scope hierarchy
212
+ - Monthly testing cost under $200 for the entire persona suite
213
+
214
+ ### 3.3 Engineering Leader
215
+
216
+ **Profile:** VP of Engineering, CTO, Director of Engineering. Responsible for
217
+ engineering investment decisions, team productivity, and technical strategy.
218
+
219
+ **Needs:**
220
+ - Evidence that AI tool investment is delivering ROI
221
+ - Adoption metrics: how many developers are using the tools, how often
222
+ - Cost metrics: total spend, spend per team, spend per persona
223
+ - Quality metrics: are bug escape rates improving? Is review turnaround faster?
224
+ - Confidence that governance scales as the organization grows
225
+
226
+ **Pain points:**
227
+ - Cannot justify AI tool budgets without data
228
+ - Does not know whether teams are actually using the tools or ignoring them
229
+ - Worries about compliance exposure in regulated domains
230
+ - Has seen AI initiatives fail because of poor adoption and lack of governance
231
+
232
+ **What success looks like:**
233
+ - Monthly dashboard showing adoption rates, cost breakdown, and ROI indicators
234
+ - Evidence of improved PR review turnaround, test coverage, and bug escape rates
235
+ - Confidence that compliance guardrails are active on every repo
236
+ - Can demonstrate value to the board with concrete metrics, not anecdotes
237
+
238
+ ### 3.4 IT / Security
239
+
240
+ **Profile:** IT operations, security engineering, or compliance team. Responsible for
241
+ device management, security policy, and regulatory compliance.
242
+
243
+ **Needs:**
244
+ - Non-overridable guardrails that no developer can disable (HARD guardrails)
245
+ - Deployment via MDM (Jamf, Intune, JumpCloud) or server-managed settings
246
+ - Audit trail for compliance reporting
247
+ - PHI/PII scanning hooks that run on every interaction
248
+ - Credential blocking and secret detection
249
+ - Zero-touch deployment: developers should not have to opt in to compliance
250
+
251
+ **Pain points:**
252
+ - No way to enforce AI compliance across the organization
253
+ - Developer-configured AI tools are a compliance blind spot
254
+ - Cannot guarantee that PHI scanning runs on every Claude Code session
255
+ - No audit evidence of AI governance for regulatory reporting
256
+
257
+ **What success looks like:**
258
+ - Managed settings deployed via MDM — compliance hooks active on every machine
259
+ - HARD guardrails that cannot be overridden at any scope level
260
+ - Audit trail of persona invocations (without individual prompt content)
261
+ - Clean compliance audits with evidence of AI governance
262
+ - Escalation mechanism for genuinely harmful content (flags, not transcripts)
263
+
264
+ ### 3.5 Skeptic / Hold-out
265
+
266
+ **Profile:** Experienced engineer who is resistant to AI tools. May have philosophical
267
+ objections, may have had bad experiences, or may simply prefer their existing workflow.
268
+ This segment exists in every organization and will never be zero.
269
+
270
+ **Needs:**
271
+ - Proof of value without disrupting their workflow
272
+ - Opt-in experience with no forced adoption
273
+ - Honest assessment of AI limitations, not hype
274
+ - Assurance that AI usage is not monitored or judged
275
+
276
+ **Pain points:**
277
+ - Feels pressured to use AI tools they do not trust
278
+ - Fears that AI metrics will be used in performance reviews
279
+ - Finds AI output noisy, inaccurate, or unhelpful
280
+ - Does not want to learn yet another tool
281
+
282
+ **What success looks like:**
283
+ - Compliance guardrails activate automatically without their action (managed settings)
284
+ - Persona-generated PR reviews appear in their workflow without them invoking anything
285
+ - Over time, sees the value in structured AI review and begins invoking personas manually
286
+ - Is never shamed, ranked, or penalized for low AI usage
287
+ - Can cleanly remove AgentBoot with one command if they decide it is not for them
288
+
289
+ ### 3.6 Non-Engineer
290
+
291
+ **Profile:** Product manager, designer, compliance officer, marketing professional.
292
+ Does not use the terminal. Needs AI assistance for structured tasks (document review,
293
+ compliance checking, content analysis).
294
+
295
+ **Needs:**
296
+ - GUI-first experience with structured forms, not slash commands
297
+ - Role-specific personas (compliance review, document analysis, content checking)
298
+ - Same privacy protections as engineers
299
+
300
+ **Pain points:**
301
+ - Terminal-based tools are inaccessible
302
+ - Does not understand "install a plugin" or "run a command"
303
+ - AI tools feel designed exclusively for engineers
304
+
305
+ **What success looks like:**
306
+ - Opens a desktop application (Cowork), sees role-specific tools in a sidebar
307
+ - Fills out a form, gets structured output — never touches a terminal
308
+ - Same personas, same quality, same governance as the engineering tools
309
+
310
+ ---
311
+
312
+ ## 4. Product Vision
313
+
314
+ ### 4.1 Positioning
315
+
316
+ AgentBoot is the Spring Boot of AI agent governance: an opinionated, drop-in foundation
317
+ that gives engineering organizations consistent AI agent behavior from day one without
318
+ building it themselves.
319
+
320
+ The analogy is precise:
321
+
322
+ | Spring Boot | AgentBoot |
323
+ |---|---|
324
+ | Convention over configuration for Java apps | Convention over configuration for AI agent behavior |
325
+ | Sensible defaults, override what is different | Sensible persona defaults, extend for your domain |
326
+ | Starter dependencies compose features | Traits compose behavioral building blocks |
327
+ | Auto-configuration detects and configures | `agentboot setup` detects role, tools, org and configures |
328
+ | Spring Boot Actuator for observability | Structured telemetry and `/insights` for measurement |
329
+ | Spring Initializr for scaffolding | `agentboot setup` and `agentboot add` for scaffolding |
330
+
331
+ ### 4.2 The Build Tool Mental Model
332
+
333
+ AgentBoot is a build tool, not a runtime framework. It produces artifacts that are
334
+ consumed by AI agent platforms. The pipeline is: validate, compile, distribute.
335
+
336
+ ```
337
+ Source files Build step Distributed artifacts
338
+ (traits, personas, → agentboot build → (.claude/, copilot-
339
+ instructions, gotchas) instructions.md, plugins)
340
+ ```
341
+
342
+ This distinction matters because:
343
+ - AgentBoot has zero runtime footprint — it produces files and exits
344
+ - The distributed artifacts work without AgentBoot installed
345
+ - Any platform that reads Markdown, YAML, or JSON can consume the output
346
+ - The build step is where governance happens (validation, composition, linting)
347
+ - Teams can inspect, understand, and modify the distributed output
348
+
349
+ ### 4.3 Where AgentBoot Is Going
350
+
351
+ **Phase 1 (Foundation):** Build system, CLI, core personas, hub-and-spoke distribution,
352
+ scope hierarchy. The minimum viable governance system.
353
+
354
+ **Phase 2 (Native Distribution):** Claude Code plugin packaging, private marketplace,
355
+ managed settings generation. Native delivery through the platform's own mechanisms.
356
+
357
+ **Phase 3 (Cross-Platform and Enterprise):** MCP server for multi-agent organizations,
358
+ cross-platform output generation (Copilot, Cursor, Gemini), structured knowledge store.
359
+
360
+ **Phase 4 (Ecosystem):** Public marketplace, community contributions, domain layers for
361
+ regulated industries, SuperClaude partnership for trait format standardization.
362
+
363
+ ### 4.4 The Long-Term Vision
364
+
365
+ AgentBoot becomes the place where AI agent governance content lives — the "npm of AI
366
+ personas." Not because of marketing, but because that is where the traits, personas,
367
+ gotchas, and domain layers are. Organizations that need governance adopt AgentBoot
368
+ because it is the standard. Contributors share their knowledge because it reaches the
369
+ most people. The ecosystem grows because useful content attracts users who become
370
+ contributors who add more useful content.
371
+
372
+ ---
373
+
374
+ ## 5. Design Principles
375
+
376
+ Ordered by priority. These principles inform every design decision and resolve
377
+ ambiguity when requirements conflict.
378
+
379
+ ### 5.1 Composition Over Inheritance (Priority: Highest)
380
+
381
+ Personas compose traits. They do not inherit from each other. If two personas share
382
+ behavior, that behavior belongs in a trait that both compose.
383
+
384
+ This is the foundational design constraint, established as Design Principle #1 from
385
+ the earliest AgentBoot designs. Object-oriented inheritance applied to personas
386
+ ("security-reviewer extends code-reviewer extends base-reviewer") creates fragile
387
+ chains where changes to a parent have unpredictable effects on children. Composition
388
+ is flat, predictable, and inspectable.
389
+
390
+ **Implication:** There is no persona hierarchy. There is no `extends` keyword. Traits
391
+ are the unit of reuse. Personas are the unit of deployment.
392
+
393
+ ### 5.2 Scoped Conventions (Priority: Highest)
394
+
395
+ AgentBoot models organizations as a four-level hierarchy: Org, Group, Team, Repo.
396
+ More specific scopes layer on top of general ones. For optional behaviors, team
397
+ overrides group overrides core. For mandatory behaviors, inheritance is top-down
398
+ (org wins).
399
+
400
+ This hierarchy mirrors how real organizations structure responsibility and governance.
401
+ It ensures that governance propagates downward automatically (a new team gets all
402
+ org-level and group-level configuration without manual setup) while preserving team
403
+ autonomy on things that are genuinely team-specific.
404
+
405
+ **Implication:** The scope hierarchy is not optional. Every piece of configuration has
406
+ a scope. Conflict resolution is deterministic: mandatory rules cascade down, optional
407
+ rules override up.
408
+
409
+ ### 5.3 Prompts as Code (Priority: High)
410
+
411
+ AI agent behavior is infrastructure: defined in files, stored in version control,
412
+ reviewed in pull requests, with a complete audit history. This is the same shift that
413
+ happened with Infrastructure as Code (Terraform, Pulumi) and Configuration as Code
414
+ (Kubernetes manifests, GitHub Actions workflows).
415
+
416
+ Every change to an agent's behavior is a pull request with a description and a review.
417
+ Traits and personas have version numbers. The sync pipeline propagates updates
418
+ automatically after the PR merges.
419
+
420
+ **Implication:** All persona definitions live in a Git repository. All changes go
421
+ through PR review. The build system enforces validation rules that block invalid
422
+ configurations from merging.
423
+
424
+ ### 5.4 Privacy First (Priority: High)
425
+
426
+ AgentBoot will never collect, transmit, or surface raw developer prompts to
427
+ organizational dashboards, managers, or analytics pipelines. This is a design
428
+ invariant, not a configuration option.
429
+
430
+ The organization gets aggregate patterns, not transcripts. High rephrase rates are
431
+ framed as persona quality problems, not developer intelligence problems. The
432
+ `rawPrompts` configuration section contains three `false` fields that cannot be set
433
+ to `true`. They exist to make the design intent explicit.
434
+
435
+ AgentBoot is honest about the boundaries: this commitment covers what AgentBoot does.
436
+ It does not and cannot override what the API provider or the organization does
437
+ independently through other channels (Compliance API, network monitoring, data exports).
438
+
439
+ **Implication:** No telemetry field contains prompt text. No dashboard shows individual
440
+ developer activity. No feature ranks, shames, gamifies, or surveys developer AI usage.
441
+ The escalation exception (genuinely harmful content) flags the category, never the
442
+ transcript.
443
+
444
+ ### 5.5 Agent-Agnostic Content, Claude Code-First Delivery (Priority: High)
445
+
446
+ Content (traits, personas, gotchas, instructions) is written in agent-agnostic formats
447
+ (Markdown, agentskills.io SKILL.md). Delivery (plugins, managed settings, hooks,
448
+ native agent frontmatter) leverages Claude Code's extensibility surface to the fullest.
449
+
450
+ If a feature "just does not work without Claude Code" — that is acceptable. Non-CC
451
+ platforms get the content (SKILL.md, instruction files, MCP servers); CC platforms get
452
+ the full governance surface (hooks, managed settings, subagent isolation, agent memory).
453
+ AgentBoot documents this gap honestly rather than pretending all agents are equal.
454
+
455
+ **Implication:** Marketplace content is stored in agent-agnostic format. The build
456
+ system produces platform-specific output. CC-specific enhancements go in metadata
457
+ (persona.config.json), not in the content itself.
458
+
459
+ ### 5.6 Easy Exit Builds Trust for Easy Entry (Priority: High)
460
+
461
+ `agentboot uninstall` is a first-class command, not an afterthought. It tracks what it
462
+ manages (via a manifest), archives what it replaces, and restores what was there before.
463
+ An organization that knows they can cleanly remove the tool in one command is more
464
+ willing to try it.
465
+
466
+ No vendor lock-in, no orphaned files, no scavenger hunt. The uninstall command removes
467
+ exactly the files AgentBoot manages and preserves everything the team authored locally.
468
+
469
+ **Implication:** Every synced file is tracked in `.agentboot-manifest.json`. Original
470
+ files are archived during setup. The uninstall path is tested in integration tests.
471
+
472
+ ### 5.7 Convention Over Configuration (Priority: Medium-High)
473
+
474
+ AgentBoot ships with sensible defaults for everything. Organizations configure what is
475
+ different about their organization, not everything from scratch. A single config file
476
+ (`agentboot.config.json`) and a `build && sync` command are sufficient to get started.
477
+
478
+ **Implication:** Core personas and traits ship ready to deploy. The setup wizard
479
+ determines the right configuration based on role detection, not manual specification.
480
+ The "happy path" requires minimal configuration.
481
+
482
+ ### 5.8 Build-Time Composition, Not Runtime Resolution (Priority: Medium)
483
+
484
+ Traits are composed into personas at build time. The compiled output is complete and
485
+ standalone. Agents receive a single file (or a set of files with @imports for CC-native
486
+ output) with all trait content already resolved. No runtime resolution, no dynamic
487
+ includes on platforms that do not support them.
488
+
489
+ The exception is Claude Code-native output, where `@import` in CLAUDE.md composes
490
+ trait files at load time — but these files are pre-generated by the build system,
491
+ not resolved dynamically.
492
+
493
+ **Implication:** The cross-platform SKILL.md output is fully self-contained. The
494
+ CC-native output uses @imports but all imported files are generated and present.
495
+ No runtime API calls to resolve trait definitions.
496
+
497
+ ### 5.9 Honest Limitations (Priority: Medium)
498
+
499
+ AgentBoot documents what works, what does not, and what the actual limitations are on
500
+ each platform. Compliance hooks exist on Claude Code but not on Copilot or Cursor.
501
+ Managed settings are CC-only. Subagent isolation requires `context: fork`, which is
502
+ CC-only. These gaps are documented per platform rather than hidden.
503
+
504
+ **Implication:** Per-platform feature matrices are part of the documentation.
505
+ Non-CC delivery is labeled "best-effort." No feature is marketed as cross-platform
506
+ unless it actually works cross-platform.
507
+
508
+ ### 5.10 Hub-and-Spoke Distribution (Priority: Medium)
509
+
510
+ One central repository (the hub) contains the source of truth for all persona
511
+ definitions. Target repositories (spokes) receive compiled artifacts via the sync
512
+ pipeline. Spokes are passive — they receive governance, they do not produce it.
513
+ Teams customize through the hub's scope hierarchy, not by committing persona files
514
+ directly to their repos.
515
+
516
+ **Implication:** The sync pipeline is one-way (hub to spoke). Local modifications
517
+ to synced files are detected and warned about. The hub repository is the authoritative
518
+ source for governance decisions.
519
+
520
+ ---
521
+
522
+ ## 6. Feature Requirements
523
+
524
+ ### 6.1 Build System
525
+
526
+ #### 6.1.1 Validation (`agentboot validate`)
527
+
528
+ Pre-build validation that catches configuration errors before compilation. Runs five
529
+ checks: persona existence (every persona referenced in config has a directory),
530
+ trait reference resolution (every trait referenced by a persona exists), SKILL.md
531
+ frontmatter validity (required fields present, values in allowed ranges), PERSONAS.md
532
+ synchronization (generated registry matches actual persona definitions), and secret
533
+ scanning (no credentials, API keys, or internal URLs in persona content).
534
+
535
+ **Who it serves:** Platform teams and CI pipelines. Validation runs automatically as
536
+ the first stage of the build pipeline and as a standalone CI gate on every PR to the
537
+ personas repository.
538
+
539
+ **Why it matters:** Without validation, invalid configurations silently produce broken
540
+ output. A missing trait reference could result in an empty persona. A credential
541
+ accidentally included in a trait file could be synced to every repo in the
542
+ organization. Validation catches these errors before they propagate.
543
+
544
+ #### 6.1.2 Compilation (`agentboot build`)
545
+
546
+ The core build step. Loads `agentboot.config.json`, resolves trait references from
547
+ each persona's `persona.config.json`, and produces two distinct compilation targets:
548
+ cross-platform output and Claude Code-native output.
549
+
550
+ **Cross-platform output** produces standalone SKILL.md files (agentskills.io format
551
+ with traits inlined), copilot-instructions.md, and generic CLAUDE.md. This output
552
+ works on any platform that reads Markdown instruction files.
553
+
554
+ **Claude Code-native output** produces the full `.claude/` directory structure:
555
+ agents with rich frontmatter (model, permissionMode, maxTurns, disallowedTools,
556
+ mcpServers, hooks, memory, isolation), skills with `context: fork`, rules with
557
+ `paths:` frontmatter, separate trait files for `@import` composition, a CLAUDE.md
558
+ using `@imports`, settings.json with hook entries, and `.mcp.json` with server
559
+ configurations.
560
+
561
+ **Who it serves:** Platform teams running the build pipeline. CI systems that produce
562
+ artifacts on merge.
563
+
564
+ **Why it matters:** The two-target compilation is what makes AgentBoot both
565
+ cross-platform and CC-optimized. Organizations choose which format to deploy per repo.
566
+ Claude Code users get the full governance surface (hooks, managed settings, isolation).
567
+ Non-CC users get the content (personas, instructions, rules) without the enforcement
568
+ mechanisms.
569
+
570
+ #### 6.1.3 Sync (`agentboot sync`)
571
+
572
+ Distributes compiled output to target repositories. Reads `repos.json`, which lists
573
+ each target repo with its platform and team assignment. Merges scopes according to the
574
+ four-level hierarchy (org, group, team, repo), with team-level configuration layering
575
+ on top of group and org configuration. Writes the appropriate output format based on
576
+ each repo's declared platform.
577
+
578
+ Sync supports multiple modes: local filesystem writes, GitHub API (creates PRs via
579
+ the API), and GitLab API (creates merge requests). A dry-run mode shows what would
580
+ change without writing anything. Each sync writes an `.agentboot-manifest.json`
581
+ tracking every managed file and its content hash, enabling clean uninstall and
582
+ drift detection.
583
+
584
+ **Who it serves:** Platform teams and CI pipelines. In steady state, sync runs
585
+ automatically on merge to the personas repo's main branch.
586
+
587
+ **Why it matters:** Sync is the mechanism that turns the hub-and-spoke model from
588
+ theory into practice. Without it, every repo update is manual. With it, a single
589
+ PR to the personas repo propagates governance to every registered repo
590
+ automatically — with human review at the PR merge point as the governance checkpoint.
591
+
592
+ #### 6.1.4 Scope Merging
593
+
594
+ The scope merge algorithm implements the four-level hierarchy. For each target repo,
595
+ the sync step composes: org-level configuration (always active), group-level
596
+ configuration (based on the repo's group assignment), team-level configuration (based
597
+ on the repo's team assignment), and repo-level overrides.
598
+
599
+ Optional behaviors follow "most specific wins" — a team can override a group default.
600
+ Mandatory behaviors follow "most general wins" — an org can mark traits or personas
601
+ as required, and no lower scope can disable them. A team-level configuration that
602
+ attempts to disable a mandatory org-level guardrail causes a build failure.
603
+
604
+ **Who it serves:** Organizations with multiple groups and teams that need both
605
+ centralized governance and team-level customization.
606
+
607
+ **Why it matters:** Without scope merging, organizations face a false choice between
608
+ centralized control (which frustrates teams) and team autonomy (which breaks
609
+ governance). The scope hierarchy provides both: mandatory rules propagate down,
610
+ optional customization layers up.
611
+
612
+ ### 6.2 CLI
613
+
614
+ #### 6.2.1 Interactive Setup Wizard (`agentboot setup`)
615
+
616
+ The primary entry point for all users. An interactive wizard that asks a series of
617
+ questions to determine the user's role (developer, platform team, IT/security,
618
+ exploring), their AI tools (Claude Code, Copilot, Cursor, mixed), their organization
619
+ context (existing AgentBoot setup, no setup, solo), team size, compliance requirements,
620
+ and device management infrastructure.
621
+
622
+ Based on the answers, the wizard executes the appropriate setup: Quick Start (solo
623
+ developer — deploys files directly to the current repo), Standard Setup (platform
624
+ team — scaffolds the org personas repo with config, personas, and traits), or
625
+ Enterprise Setup (IT — generates managed settings for MDM deployment plus marketplace
626
+ template).
627
+
628
+ The wizard auto-detects as much as possible before asking: git remote for org name,
629
+ `.claude/` in the current repo for existing setup, managed settings on disk for
630
+ existing governance, installed Claude Code version, and platform tools in PATH.
631
+
632
+ **Who it serves:** Every first-time user, regardless of role.
633
+
634
+ **Why it matters:** The setup wizard is the difference between "read the docs for
635
+ 30 minutes then configure" and "answer 3-7 questions and you are running." Progressive
636
+ disclosure ensures that solo developers are not overwhelmed with enterprise options,
637
+ while platform teams get the full scaffolding they need.
638
+
639
+ #### 6.2.2 Discovery (`agentboot discover`)
640
+
641
+ Scans repositories, machines, and configurations for existing AI agent content —
642
+ CLAUDE.md files, Copilot instructions, Cursor rules, MCP servers, hooks, skills,
643
+ agents, and anything else that represents agentic maturity the organization already
644
+ has. Supports scanning a single repo, a directory of repos, all repos in a GitHub
645
+ organization (via API), local machine configuration, and managed settings paths.
646
+
647
+ Discovery produces five actions: a detailed inventory report (Markdown), a classify-
648
+ and-ingest pipeline (decompose existing content into traits, personas, gotchas, and
649
+ instructions), an overlap analysis (find near-duplicate content across repos), a
650
+ migration plan (what becomes what, with estimated reduction), and a config export
651
+ (generate `agentboot.config.json` from the discovered repo structure).
652
+
653
+ Discovery and migration are non-destructive. They never modify, move, or delete
654
+ existing files. They create new files in the AgentBoot personas repo. The originals
655
+ stay untouched. The organization reviews, tests, and deploys at their pace via PR.
656
+
657
+ **Who it serves:** Platform teams evaluating AgentBoot for an organization that already
658
+ has scattered agentic content.
659
+
660
+ **Why it matters:** Most organizations adopting AgentBoot are not starting from zero.
661
+ They have 20+ repos with independent CLAUDE.md files, custom agents, and instruction
662
+ files. Without discovery, consolidation is a manual audit. With it, AgentBoot finds
663
+ everything, identifies duplication, and produces a migration plan that typically
664
+ achieves 80-85% line reduction through deduplication and centralization.
665
+
666
+ #### 6.2.3 Build and Distribution Commands
667
+
668
+ `agentboot build` compiles personas with options for specific output formats
669
+ (claude-code, copilot, cross-platform, plugin, all) and a validate-only dry run.
670
+ `agentboot sync` distributes to repos with support for local filesystem, GitHub API,
671
+ and GitLab API modes, plus single-repo targeting and dry-run. `agentboot export`
672
+ generates distributable artifacts (plugin directory, marketplace scaffold, managed
673
+ settings for MDM, reusable GitHub Actions workflow, MCP server package).
674
+ `agentboot publish` pushes the plugin to a configured marketplace with version bumping.
675
+
676
+ **Who it serves:** Platform teams managing the build and distribution pipeline.
677
+
678
+ **Why it matters:** These commands form the core build pipeline. They are designed to
679
+ work in both interactive (developer workstation) and non-interactive (CI/CD) modes,
680
+ with every command supporting `--non-interactive` flags.
681
+
682
+ #### 6.2.4 Scaffolding (`agentboot add`)
683
+
684
+ Scaffolds new components: personas, traits, domains, gotchas, hooks, and repo
685
+ registrations. Each scaffold creates the appropriate directory structure with template
686
+ files that follow the format standards.
687
+
688
+ The `add prompt` subcommand ingests raw prompt text from multiple sources (inline text,
689
+ file, clipboard, URL, batch decomposition of an existing CLAUDE.md). It classifies the
690
+ content (trait, gotcha, persona rule, or always-on instruction), formats it to the
691
+ appropriate standard, and saves it to the correct location. A dry-run mode previews
692
+ without writing.
693
+
694
+ **Who it serves:** Platform teams creating new content and anyone contributing to the
695
+ personas repository.
696
+
697
+ **Why it matters:** Scaffolding reduces the barrier to adding new governance content.
698
+ The `add prompt` command is particularly valuable during migration: it takes the
699
+ scattered, unstructured content found by `discover` and transforms it into properly
700
+ formatted AgentBoot content.
701
+
702
+ #### 6.2.5 Diagnostics (`agentboot doctor`, `agentboot status`)
703
+
704
+ `agentboot doctor` checks environment health: Claude Code version, Node.js version,
705
+ git version, configuration validity, persona completeness, sync status, plugin state,
706
+ and managed settings status. Every problem includes a fix command.
707
+
708
+ `agentboot doctor --diagnose` runs a layered isolation test to pinpoint whether a
709
+ problem is in AgentBoot core, org config, org traits, org personas, or org extensions.
710
+ Each layer adds one piece of the stack; the layer where it breaks is the layer that
711
+ has the bug.
712
+
713
+ `agentboot status` provides a deployment dashboard: enabled personas and traits, repo
714
+ registration status with platform and sync version, plugin state, and managed settings
715
+ status.
716
+
717
+ **Who it serves:** Anyone troubleshooting issues. Platform teams monitoring deployment.
718
+
719
+ **Why it matters:** "Is this my bug or AgentBoot's bug?" is a question every
720
+ organization will ask. The diagnostic tools answer it systematically rather than
721
+ requiring manual investigation.
722
+
723
+ #### 6.2.6 Linting (`agentboot lint`)
724
+
725
+ Static analysis of persona definitions. Checks for vague language ("be thorough",
726
+ "try to"), token budget violations, credentials in prompt text, conflicting
727
+ instructions across traits, unused traits, and security issues. Configurable severity
728
+ thresholds (error, warning, info).
729
+
730
+ Token budget analysis calculates per-persona context cost and flags personas that
731
+ exceed configurable limits. This prevents context bloat — the gradual growth of prompt
732
+ content that wastes tokens without adding value.
733
+
734
+ **Who it serves:** Platform teams authoring personas, and CI pipelines gating merges.
735
+
736
+ **Why it matters:** Prompt quality is the product. A governance framework that produces
737
+ vague, bloated, or conflicting prompts is actively wasting money and developer trust.
738
+ Linting enforces a minimum quality bar automatically.
739
+
740
+ #### 6.2.7 Testing (`agentboot test`)
741
+
742
+ Multi-type test runner supporting deterministic tests (free, every commit),
743
+ behavioral tests (LLM invocation, ~$0.50/test), snapshot regression tests (compare
744
+ output across versions), and LLM-as-judge evaluation (Opus evaluates quality on
745
+ multiple dimensions). Supports cost caps, per-persona targeting, CI mode with exit
746
+ codes and JSON output, and flake tolerance (2-of-3 pass).
747
+
748
+ Full details in section 6.7 (Testing).
749
+
750
+ **Who it serves:** Platform teams validating persona behavior before deployment.
751
+
752
+ **Why it matters:** AI persona output is non-deterministic. Traditional testing
753
+ approaches (exact output matching) do not work. The multi-layer test pyramid provides
754
+ confidence at each level: schema tests are free and fast, behavioral tests verify
755
+ that personas catch known bugs, snapshot tests detect regressions, and LLM-as-judge
756
+ evaluates qualitative dimensions that pattern matching cannot assess.
757
+
758
+ #### 6.2.8 Additional Commands
759
+
760
+ `agentboot connect` provides developer self-service for connecting to an organization's
761
+ marketplace (CLI equivalent of `/agentboot:connect`). `agentboot upgrade` pulls the
762
+ latest AgentBoot core into an organization's personas repo. `agentboot validate` is the
763
+ CI-friendly validation-only command with strict mode. `agentboot search` queries the
764
+ marketplace for traits, gotchas, personas, and domains. `agentboot metrics` reads
765
+ telemetry and reports per-persona, per-team, and per-period statistics.
766
+ `agentboot cost-estimate` projects per-persona monthly cost across the organization.
767
+ `agentboot review` generates guided human review sessions with curated samples.
768
+ `agentboot issue` streamlines bug reporting with auto-attached diagnostics and
769
+ privacy-safe redaction. `agentboot uninstall` provides clean removal with manifest
770
+ tracking, original file restoration, and mixed content handling.
771
+
772
+ **Who it serves:** Various segments depending on the command.
773
+
774
+ **Why it matters:** Each command addresses a specific workflow gap. Together, they form
775
+ a complete CLI that covers the full lifecycle from setup through daily use to removal.
776
+
777
+ ### 6.3 Persona System
778
+
779
+ #### 6.3.1 Traits
780
+
781
+ Reusable behavioral building blocks. Each trait captures a single aspect of how an
782
+ agent should think or communicate — a cognitive stance, an output discipline, or an
783
+ epistemic commitment. Traits support weight configurations (HIGH / MEDIUM / LOW,
784
+ mapping to a 0.0-1.0 numeric scale) and are composed at build time.
785
+
786
+ Core traits include: critical-thinking (skepticism calibration), structured-output
787
+ (consistent output formatting), source-citation (evidence requirements), confidence-
788
+ signaling (uncertainty communication), audit-trail (compliance logging), and
789
+ schema-awareness (data structure understanding). Additional traits include
790
+ creative-suggestion (proactive improvement ideas) and cost-awareness (resource
791
+ efficiency).
792
+
793
+ The trait weight system is a calibration system, not a priority system. At HIGH, the
794
+ threshold for surfacing a concern is very low. At LOW, the threshold is high. The
795
+ weight does not override the severity floor — at any weight, CRITICAL findings always
796
+ surface. The numeric scale (0.0-1.0) provides finer calibration for advanced use cases
797
+ while the named weights (HIGH/MEDIUM/LOW) serve as the simplified interface.
798
+
799
+ **Who it serves:** Platform teams authoring personas, and the marketplace community
800
+ sharing behavioral building blocks.
801
+
802
+ **Why it matters:** Before traits, every persona independently specified its approach
803
+ to things like skepticism, output structure, and evidence requirements. In practice,
804
+ this meant the same concepts were expressed slightly differently in every persona.
805
+ Traits solve this: write the behavior once, compose it everywhere, improve it in one
806
+ place.
807
+
808
+ #### 6.3.2 Personas
809
+
810
+ Complete, deployable agents composed from traits plus a specialized system prompt that
811
+ defines identity, operating context, and mandate. Use the agentskills.io SKILL.md
812
+ format (Markdown with YAML frontmatter). The frontmatter specifies ID, version, traits,
813
+ scope, and output format. The prose is the system prompt.
814
+
815
+ Core V1 personas: code-reviewer (code review against team standards), security-
816
+ reviewer (vulnerability analysis, OWASP, auth, data handling), test-generator
817
+ (unit and integration test generation from function signatures), and test-data-expert
818
+ (realistic synthetic test data generation).
819
+
820
+ Each persona has a `persona.config.json` that specifies build metadata: which traits
821
+ to compose, at what weight, the target model, permission mode, tool restrictions,
822
+ MCP servers, hooks, and autonomy level.
823
+
824
+ **Who it serves:** All developers (as consumers) and platform teams (as authors).
825
+
826
+ **Why it matters:** Personas are the user-facing product. Everything else in AgentBoot
827
+ — traits, the build system, the scope hierarchy, the distribution pipeline — exists to
828
+ make personas work well. A persona that catches SQL injection, explains the risk,
829
+ suggests the fix, and cites the relevant standard is the experience that makes
830
+ developers trust and adopt the system.
831
+
832
+ #### 6.3.3 Gotchas Rules
833
+
834
+ Path-scoped instructions that encode hard-won operational knowledge. Each gotchas file
835
+ is a Markdown file with `paths:` frontmatter that limits when the rule activates. When
836
+ a developer is working on files that match the glob pattern, the gotchas content is
837
+ automatically included in the agent's context. When working on unrelated files, the
838
+ gotchas are invisible — zero context cost.
839
+
840
+ Gotchas capture the kind of information that lives in one engineer's head until they
841
+ leave: "PostgreSQL partitions do NOT inherit `relrowsecurity`," "UUID primary keys
842
+ cause exponential INSERT slowdown," "Lambda cold starts are 10x worse with VPC
843
+ networking." Technology-specific, not org-specific — making them highly shareable.
844
+
845
+ **Who it serves:** All developers (as consumers), platform teams and individual
846
+ contributors (as authors), and the marketplace community (as shareable knowledge).
847
+
848
+ **Why it matters:** Gotchas rules capture knowledge at the exact moment it is needed.
849
+ A developer writing a database migration sees the PostgreSQL gotchas. A developer
850
+ writing a Lambda handler sees the serverless gotchas. No one has to remember to consult
851
+ a wiki page or ask the right person.
852
+
853
+ #### 6.3.4 Always-On Instructions
854
+
855
+ Universal guardrails distributed to every repo regardless of persona configuration.
856
+ These are the org-wide rules that apply to every AI interaction: security baselines,
857
+ output format standards, coding conventions. They load at session start and remain
858
+ active throughout.
859
+
860
+ **Who it serves:** The entire organization. IT/Security teams who need guarantees
861
+ that certain rules are always active.
862
+
863
+ **Why it matters:** Some rules should never be opt-in. A security baseline, a coding
864
+ standard, a compliance requirement — these must apply everywhere. Always-on
865
+ instructions are the mechanism for universal governance.
866
+
867
+ #### 6.3.5 Domain Layers
868
+
869
+ Complete packages of traits, personas, gotchas, and instructions for a specific
870
+ compliance regime or technology stack. A healthcare domain layer includes PHI-awareness
871
+ traits, HIPAA reviewer personas, patient data gotchas, and compliance-specific
872
+ always-on instructions. A fintech domain layer includes PCI-DSS compliance traits,
873
+ financial data reviewers, and transaction security gotchas.
874
+
875
+ Domain layers are the highest-effort contribution to the marketplace but also the
876
+ highest value. They package everything an organization needs to govern AI behavior
877
+ in a regulated domain.
878
+
879
+ **Who it serves:** Organizations in regulated industries (healthcare, finance,
880
+ government) that need domain-specific compliance guardrails.
881
+
882
+ **Why it matters:** Regulated industries cannot adopt AI tools without compliance
883
+ guardrails. Building these guardrails from scratch is expensive and error-prone.
884
+ Domain layers provide a tested, community-maintained foundation that organizations
885
+ can adopt and extend.
886
+
887
+ #### 6.3.6 Per-Persona Extensions
888
+
889
+ Teams can extend base personas without forking them. An extension file adds
890
+ domain-specific rules to a base persona. The base definition stays unmodified and
891
+ receives upstream improvements automatically. This prevents the anti-pattern of
892
+ copying a persona, modifying it, and maintaining a fork that diverges over time.
893
+
894
+ **Who it serves:** Teams that need team-specific customization of org-wide personas.
895
+
896
+ **Why it matters:** Without extensions, teams face a false choice: use the org persona
897
+ as-is (too generic) or fork it (diverges and rots). Extensions provide the middle path:
898
+ customize behavior while staying on the upstream upgrade path.
899
+
900
+ ### 6.4 Delivery and Distribution
901
+
902
+ #### 6.4.1 Claude Code Plugin
903
+
904
+ The primary delivery method for Claude Code users. AgentBoot output is packaged as a
905
+ CC plugin — a directory containing agents, skills, hooks, rules, MCP configuration,
906
+ and settings — distributable through a private marketplace.
907
+
908
+ The organization creates a private marketplace (a Git repository with a
909
+ marketplace.json catalog). Engineers install with a single command:
910
+ `/plugin marketplace add org/personas` and `/plugin install org-plugin`. IT can
911
+ force-enable the plugin via managed settings, requiring zero developer action.
912
+
913
+ Plugin packaging is produced by `agentboot export --format plugin`. The plugin name
914
+ becomes the namespace prefix for all skills (e.g., `/acme:review-code`).
915
+
916
+ **Who it serves:** All Claude Code users in the organization.
917
+
918
+ **Why it matters:** The plugin is the native CC distribution mechanism. It supports
919
+ namespace isolation (preventing conflicts), version pinning, force-enable via managed
920
+ settings, and live reload without restart. It is the path of least resistance for CC
921
+ organizations.
922
+
923
+ #### 6.4.2 Marketplace
924
+
925
+ Three-layer marketplace ecosystem: Core (maintained by AgentBoot, Apache 2.0),
926
+ Verified (community-contributed, reviewed by maintainers, meets quality standards),
927
+ and Community (unreviewed, valid format only, buyer-beware).
928
+
929
+ Full details in section 6.9 (Marketplace).
930
+
931
+ **Who it serves:** All users who need pre-built governance content.
932
+
933
+ **Why it matters:** The marketplace is the mechanism that turns AgentBoot from a
934
+ framework into an ecosystem. A framework solves the governance problem for one
935
+ organization. An ecosystem solves it for the industry.
936
+
937
+ #### 6.4.3 Managed Settings
938
+
939
+ AgentBoot generates managed settings artifacts for deployment via MDM (Jamf, Intune,
940
+ JumpCloud, Kandji) or Anthropic's server-managed settings. Managed settings are
941
+ deployed to OS-level paths and cannot be overridden by any user or project
942
+ configuration.
943
+
944
+ Managed settings carry HARD guardrails only: compliance hooks, credential blocking,
945
+ audit logging, and forced plugin installation. They are not used for persona delivery
946
+ (that is what the plugin and sync are for). This separation keeps compliance and
947
+ convenience in distinct channels.
948
+
949
+ `agentboot export --format managed` produces the artifacts. IT deploys them through
950
+ their existing MDM infrastructure.
951
+
952
+ **Who it serves:** IT/Security teams responsible for compliance enforcement.
953
+
954
+ **Why it matters:** HARD guardrails must be non-overridable. A PHI scanning hook in a
955
+ healthcare organization cannot be disabled by a developer who finds it inconvenient.
956
+ Managed settings are the only mechanism that provides OS-level enforcement.
957
+
958
+ #### 6.4.4 MCP Server
959
+
960
+ AgentBoot exposed as an MCP (Model Context Protocol) server that any MCP-compatible
961
+ agent can consume. Provides persona invocation, trait lookup, governance status, and
962
+ knowledge base access as MCP tools and resources.
963
+
964
+ The MCP server is the cross-platform bridge. It works in Claude Code, Copilot, Cursor,
965
+ Gemini CLI, and any other MCP client. Organizations with mixed toolchains use the MCP
966
+ server to ensure everyone gets governed personas regardless of which tool they use.
967
+
968
+ **Who it serves:** Multi-agent organizations with engineers using different AI tools.
969
+
970
+ **Why it matters:** Without MCP, each platform needs its own delivery mechanism (plugin
971
+ for CC, instruction files for Copilot, rules for Cursor). MCP provides a single
972
+ channel that serves all platforms identically with the same persona definitions and
973
+ invocation interface.
974
+
975
+ #### 6.4.5 Direct Sync
976
+
977
+ The fallback/bootstrap delivery method. AgentBoot's sync script writes compiled files
978
+ directly to target repos' `.claude/`, `.github/`, and `.cursor/` directories. When a
979
+ developer clones the repo, the personas are already there with zero setup.
980
+
981
+ This is the simplest mental model (files in a directory), works offline, and requires
982
+ no plugin system dependency. It scales well to dozens of repos but creates sync PR
983
+ noise at hundreds.
984
+
985
+ **Who it serves:** All developers in repos registered for sync, and organizations that
986
+ cannot use the plugin system.
987
+
988
+ **Why it matters:** The "repo already has it" model provides the lowest-friction
989
+ developer experience. The developer never installs anything — governance is in the
990
+ repo they are already working in.
991
+
992
+ #### 6.4.6 Non-Claude Code Delivery
993
+
994
+ For Copilot: generates `copilot-instructions.md`, path-scoped `.instructions.md`
995
+ files, `.prompt.md` slash command files, and agentskills.io SKILL.md files. For
996
+ Cursor: generates `.cursor/rules/` files, `.cursorrules`, and SKILL.md files. For
997
+ Gemini CLI: generates `GEMINI.md` and SKILL.md files.
998
+
999
+ The `repos.json` platform field determines which format each repo receives. The MCP
1000
+ server is the equalizer for multi-platform organizations.
1001
+
1002
+ **Who it serves:** Organizations using tools other than Claude Code.
1003
+
1004
+ **Why it matters:** Honest cross-platform support is a competitive differentiator.
1005
+ AgentBoot does not pretend all platforms are equal — it documents the governance gaps
1006
+ per platform — but it does provide the best possible experience on each one.
1007
+
1008
+ ### 6.5 Privacy and Telemetry
1009
+
1010
+ #### 6.5.1 Three-Tier Privacy Model
1011
+
1012
+ **Tier 1 (Private):** Raw prompts, conversation history, session transcripts, files
1013
+ read during exploration. These stay on the developer's machine. AgentBoot does not
1014
+ read, transmit, or reference them. Retention is session duration or developer's
1015
+ choice.
1016
+
1017
+ **Tier 2 (Privileged):** Aggregated patterns extracted by LLM analysis via the
1018
+ `/insights` skill. The analysis uses the same Claude API trust boundary the developer
1019
+ already uses for every prompt. The developer sees insights first and approves what gets
1020
+ shared. Patterns, not transcripts, are the output.
1021
+
1022
+ **Tier 3 (Organizational):** Persona output metrics — invocation counts, cost,
1023
+ findings severity distribution, model usage, and duration. Anonymized by default (no
1024
+ developer ID). Team-level attribution via the scope field enables budget tracking
1025
+ without identifying individuals.
1026
+
1027
+ **Who it serves:** All users. Developers get privacy. The organization gets metrics.
1028
+
1029
+ **Why it matters:** The privacy model is a product differentiator. In a market where
1030
+ enterprises are deploying AI monitoring tools, AgentBoot takes the opposite stance:
1031
+ it helps organizations improve their AI governance without being the tool that surveys
1032
+ their developers. The best prompt optimization system is one that developers feed
1033
+ willingly because they trust it.
1034
+
1035
+ #### 6.5.2 `/insights` Skill
1036
+
1037
+ A skill that sends session transcripts to the Claude API (Haiku for speed/cost) for
1038
+ pattern extraction. Presents personal prompt insights to the developer: session counts,
1039
+ most-used personas, common topics, rephrase rates, cost, and improvement suggestions.
1040
+ The developer optionally approves sharing anonymized patterns to the org aggregate.
1041
+
1042
+ The analysis prompt is explicitly designed to extract patterns, not judge. It frames
1043
+ everything as persona improvement opportunities, not developer deficiencies. It does
1044
+ not quote developer prompts, judge question quality, identify knowledge deficiencies,
1045
+ or produce output that could embarrass the developer if shared.
1046
+
1047
+ **Who it serves:** Developers (personal analytics) and the organization (aggregate
1048
+ patterns for persona optimization).
1049
+
1050
+ **Why it matters:** Without insights, prompt optimization is guesswork. With insights,
1051
+ the organization knows "auth patterns are asked about 89 times across 23 developers"
1052
+ and can act on it (create an auth-patterns skill). The mechanism respects privacy while
1053
+ enabling organizational learning.
1054
+
1055
+ #### 6.5.3 Org Dashboard
1056
+
1057
+ Aggregate metrics visible to engineering leaders and platform teams: active seats,
1058
+ persona invocations, cost per team, model mix, persona effectiveness (rephrase rates,
1059
+ false positive rates), common knowledge gaps, and cost efficiency analysis.
1060
+
1061
+ The dashboard shows investment metrics (cost, adoption, ROI) and outcome metrics (PR
1062
+ quality, bug rates, coverage). It never shows process metrics (what developers typed,
1063
+ how many times they rephrased, what they asked about).
1064
+
1065
+ Optional per-developer usage tracking with three formats: `false` (no developer
1066
+ identity, the default), `hashed` (consistent anonymous ID), and `email` (full
1067
+ attribution). Full attribution requires clear communication to the team and is
1068
+ intended for cost allocation and license optimization, not surveillance.
1069
+
1070
+ **Who it serves:** Engineering leaders, platform teams, and org owners.
1071
+
1072
+ **Why it matters:** Leaders need evidence that the AI investment is delivering value.
1073
+ The dashboard provides it without violating developer privacy.
1074
+
1075
+ #### 6.5.4 Escalation Exception
1076
+
1077
+ One exception to prompt privacy: genuinely harmful content (attempted exfiltration,
1078
+ guardrail circumvention, harassment, malware generation). The system flags the
1079
+ category locally first (the developer sees the flag), then reports the flag (category
1080
+ and date, not the transcript) to a designated compliance contact. The compliance team
1081
+ can request the transcript through a formal process (like a legal hold), not through
1082
+ the analytics pipeline.
1083
+
1084
+ Implemented via a `UserPromptSubmit` hook using Haiku for fast evaluation. The prompt
1085
+ is explicitly scoped to harmful categories only — not quality, intelligence, or
1086
+ competence.
1087
+
1088
+ **Who it serves:** IT/Security and compliance teams.
1089
+
1090
+ **Why it matters:** Complete privacy absolutism is irresponsible in an enterprise
1091
+ context. The escalation exception provides a narrow, transparent mechanism for
1092
+ genuinely harmful content while maintaining the broader privacy commitment.
1093
+
1094
+ #### 6.5.5 Telemetry Configuration
1095
+
1096
+ Privacy configuration in `agentboot.config.json` includes telemetry settings (enabled,
1097
+ includeDevId, devIdFormat, includeCost, includeScope, destination), insights settings
1098
+ (enabled, autoShareAnonymized, escalation categories and contact), and rawPrompts
1099
+ settings (collect: false, transmit: false, surfaceToOrg: false — fields that cannot be
1100
+ set to true, existing to make design intent explicit).
1101
+
1102
+ **Who it serves:** Platform teams configuring privacy settings for their organization.
1103
+
1104
+ **Why it matters:** Privacy is not a binary. Different organizations have different
1105
+ requirements. The configuration provides a spectrum from fully anonymous to fully
1106
+ attributed, with clear documentation of the implications of each setting.
1107
+
1108
+ ### 6.6 Knowledge Layer
1109
+
1110
+ #### 6.6.1 Stage 1: Flat Files (Default)
1111
+
1112
+ The current and default knowledge system. Markdown files loaded into context:
1113
+ always-on instructions at session start, path-scoped rules when matching files are
1114
+ read, trait content composed at build time, and skills loaded on invocation.
1115
+
1116
+ Works for 5-50 knowledge items. Zero infrastructure. Version-controlled in git.
1117
+ Human-readable and editable. Most organizations stay at this stage permanently.
1118
+
1119
+ **Who it serves:** All organizations.
1120
+
1121
+ **Why it matters:** Flat files are the right default. They require no infrastructure,
1122
+ no databases, no embedding APIs. They work immediately and scale to the needs of most
1123
+ teams.
1124
+
1125
+ #### 6.6.2 Stage 2: Structured Knowledge Store
1126
+
1127
+ A queryable layer on top of flat files. Files gain optional frontmatter for structured
1128
+ fields (type, technology, tags, severity, domain, learned_from). The build system
1129
+ generates a SQLite index from this frontmatter. An MCP server exposes tag/category
1130
+ queries: search by technology and severity, get by ID, find related items, and list
1131
+ by type and tag.
1132
+
1133
+ Handles 50-500 knowledge items. Zero new infrastructure (SQLite is a single file
1134
+ shipped with the MCP server). The flat files remain the source of truth — the index
1135
+ is derived, not authoritative.
1136
+
1137
+ **Who it serves:** Mature organizations with extensive accumulated knowledge (gotchas,
1138
+ ADRs, incident learnings, standards, patterns).
1139
+
1140
+ **Why it matters:** When an organization has 200 gotchas and 50 ADRs, loading
1141
+ everything into context is wasteful. Structured queries allow personas to retrieve
1142
+ only the relevant knowledge for the current task.
1143
+
1144
+ #### 6.6.3 Stage 3: Vector / RAG
1145
+
1146
+ Semantic retrieval via embeddings. Instead of querying by tags and categories, the
1147
+ persona describes what it is looking at and the knowledge base returns the most
1148
+ semantically relevant items. Built on sqlite-vss (vector search extension for SQLite),
1149
+ keeping zero new infrastructure.
1150
+
1151
+ The key use case is context-aware review: "this code is doing token refresh validation"
1152
+ retrieves the incident report about token expiry race conditions, even though the code
1153
+ does not mention "race condition." The connection is semantic, not keyword-based.
1154
+
1155
+ Handles 500+ knowledge items. Requires embedding API calls (incremental, only changed
1156
+ items). Cost is minimal (~$0.20 for 1,000 items).
1157
+
1158
+ **Who it serves:** Compliance-heavy industries where accumulated knowledge (incidents,
1159
+ ADRs, patterns) is as valuable as the code itself.
1160
+
1161
+ **Why it matters:** The progression from rules to knowledge to organizational memory
1162
+ is the long-term value of the knowledge layer. A security reviewer that cites last
1163
+ year's incident report is qualitatively different from one that only checks rules.
1164
+
1165
+ #### 6.6.4 Stable MCP Interface
1166
+
1167
+ The MCP interface stays identical across all three stages. Personas do not know or care
1168
+ whether the backing store is flat files, SQLite, or pgvector. They call the same MCP
1169
+ tools. An organization can upgrade from Stage 2 to Stage 3 without rewriting any
1170
+ persona definitions.
1171
+
1172
+ **Who it serves:** Platform teams managing knowledge infrastructure.
1173
+
1174
+ **Why it matters:** The abstraction boundary between personas and the knowledge store
1175
+ means organizations can evolve their knowledge infrastructure without touching their
1176
+ persona definitions. This is why MCP-first matters — the interface is the contract.
1177
+
1178
+ ### 6.7 Testing
1179
+
1180
+ #### 6.7.1 Unit and Schema Tests (Layer 1)
1181
+
1182
+ Config validation against JSON schema, frontmatter parsing for all persona SKILL.md
1183
+ files, trait composition (injection markers, weight resolution, missing references,
1184
+ circular dependencies), lint rules (vague language, token budget, credentials,
1185
+ conflicts), and sync logic (scope merging, output format selection, manifest generation,
1186
+ PERSONAS.md registry).
1187
+
1188
+ Free. Runs on every commit. Must pass to merge. Uses vitest.
1189
+
1190
+ **Who it serves:** AgentBoot core maintainers and platform teams with custom content.
1191
+
1192
+ **Why it matters:** The foundation of the test pyramid. Catches configuration errors,
1193
+ format violations, and composition bugs before they reach compilation.
1194
+
1195
+ #### 6.7.2 Integration Tests (Layer 2)
1196
+
1197
+ Full build pipeline (validate, compile, sync produces expected output), CC-native
1198
+ output validation (agent frontmatter, @imports, paths: frontmatter, settings.json
1199
+ hooks, .mcp.json), cross-platform output validation (standalone SKILL.md, copilot-
1200
+ instructions.md), plugin export validation (structure, claude plugin validate),
1201
+ discover and ingest (finds files, identifies duplicates, generates migration plan,
1202
+ non-destructive guarantee), and uninstall (removes managed files only, preserves
1203
+ local files, handles mixed content, restores archive).
1204
+
1205
+ Free. Runs on every commit and PR. Uses vitest with temp directories and test repo
1206
+ fixtures.
1207
+
1208
+ **Who it serves:** AgentBoot core maintainers.
1209
+
1210
+ **Why it matters:** Integration tests verify that the full pipeline works end-to-end.
1211
+ A unit test might pass on trait composition, but the integration test catches when the
1212
+ compiled output misses a hook entry in settings.json.
1213
+
1214
+ #### 6.7.3 Behavioral Tests (Layer 3)
1215
+
1216
+ LLM invocation with known inputs, asserting on output patterns. Test cases defined in
1217
+ YAML format specifying persona, model (Haiku for cost), setup files, prompts, and
1218
+ assertions (findings_min/max, severity_includes/excludes, text_matches/excludes,
1219
+ confidence_min, output_contains, output_structure, json_schema, token_max,
1220
+ duration_max_ms).
1221
+
1222
+ Uses `claude -p` with structured JSON output, bypass permissions, and no session
1223
+ persistence. Flake tolerance (2-of-3 runs) handles non-determinism. Cost: ~$5/PR
1224
+ for 4 personas.
1225
+
1226
+ **Who it serves:** Platform teams validating persona behavior before deployment.
1227
+
1228
+ **Why it matters:** The core novel challenge in testing AI personas. Traditional
1229
+ tests cannot verify that a persona catches a SQL injection vulnerability. Behavioral
1230
+ tests use the cheapest model (Haiku) to verify that the prompt structure elicits the
1231
+ right behavior — if it works on Haiku, it will work better on Sonnet/Opus.
1232
+
1233
+ #### 6.7.4 Snapshot / Regression Tests (Layer 4)
1234
+
1235
+ Compare persona output across versions to detect regressions. Snapshots store
1236
+ structured summaries (finding counts, patterns, tokens, duration), not full prose.
1237
+ Run after persona prompt changes, trait updates, model changes, and weekly for provider
1238
+ drift detection.
1239
+
1240
+ Flags differences for human review: MATCH (identical), CHANGED (new findings — is the
1241
+ change correct?), and REGRESSION (previously caught findings now missed — investigate).
1242
+
1243
+ Cost: ~$5 per persona change.
1244
+
1245
+ **Who it serves:** Platform teams managing persona quality over time.
1246
+
1247
+ **Why it matters:** Without regression testing, a persona improvement might silently
1248
+ break an existing capability. Snapshot tests make these regressions visible before
1249
+ deployment.
1250
+
1251
+ #### 6.7.5 LLM-as-Judge (Layer 5)
1252
+
1253
+ A separate LLM call (Opus as judge) evaluates persona output on qualitative dimensions:
1254
+ completeness, accuracy, specificity, prioritization, and tone. Test cases include
1255
+ ground truth (known issues in the code) so the judge can evaluate whether the persona
1256
+ found what it should have found.
1257
+
1258
+ Run on major persona rewrites, new personas, model migrations, and quarterly quality
1259
+ audits. Cost: ~$20 per run.
1260
+
1261
+ **Who it serves:** Platform teams making major persona changes.
1262
+
1263
+ **Why it matters:** Pattern matching cannot evaluate whether a review is "thorough" or
1264
+ whether the tone is "professional." LLM-as-judge applies consistent qualitative
1265
+ evaluation at scale, filling the gap between structural assertions and human review.
1266
+
1267
+ #### 6.7.6 Human Review (Layer 6)
1268
+
1269
+ Guided human review sessions via `agentboot review`. Presents randomly sampled persona
1270
+ outputs from the last 7 days with structured questions (accuracy, severity, additions,
1271
+ removals). Takes 10-15 minutes per persona.
1272
+
1273
+ Triggers: new persona ships (3-5 real PRs), major trait change (before/after), quarterly
1274
+ audit (random sample of 20 outputs), quality escalation (specific disputed finding).
1275
+
1276
+ **Who it serves:** Platform teams and domain experts.
1277
+
1278
+ **Why it matters:** Humans make judgment calls that no automated test can make. The
1279
+ review tool makes this sustainable (structured, guided, 1 hour/month for 4 personas)
1280
+ rather than ad hoc.
1281
+
1282
+ #### 6.7.7 Mutation Testing
1283
+
1284
+ Deliberately introduces known bugs into persona prompts and verifies that tests catch
1285
+ the regressions. Validates that the test suite actually detects the problems it claims
1286
+ to detect.
1287
+
1288
+ **Who it serves:** Platform teams validating test suite quality.
1289
+
1290
+ **Why it matters:** "Who tests the tests?" Mutation testing answers this question by
1291
+ proving that behavioral tests catch regressions when they should.
1292
+
1293
+ ### 6.8 Developer Onboarding
1294
+
1295
+ #### 6.8.1 First-Session Welcome
1296
+
1297
+ A brief orientation (~80 tokens) generated as part of the CLAUDE.md that AgentBoot
1298
+ syncs. Lists available personas with invocation commands, provides 3-4 tips for
1299
+ effective use, mentions the privacy commitment, and directs new users to `/learn`.
1300
+ Loads once per session and fades into background context after a few sessions.
1301
+
1302
+ **Who it serves:** Developers opening Claude Code in an AgentBoot-governed repo for
1303
+ the first time.
1304
+
1305
+ **Why it matters:** The first interaction determines whether a developer engages or
1306
+ abandons. A wall of text loses them. A brief, actionable orientation converts them.
1307
+
1308
+ #### 6.8.2 `/learn` Skill
1309
+
1310
+ A context-aware training assistant, not a static tutorial. Provides a topic browser
1311
+ (Getting Started, Going Deeper, Quick Reference), answers specific questions
1312
+ ("how do I review one file?"), and delivers contextual explanations (severity levels,
1313
+ persona concepts, prompting tips).
1314
+
1315
+ The skill uses `disable-model-invocation: true` for static content and normal model
1316
+ invocation for free-form questions.
1317
+
1318
+ **Who it serves:** Developers at any skill level who want to understand the persona
1319
+ system better.
1320
+
1321
+ **Why it matters:** The `/learn` skill provides help at the moment of need. Instead of
1322
+ sending developers to documentation, it answers their question in context.
1323
+
1324
+ #### 6.8.3 Curated External Resources
1325
+
1326
+ AgentBoot does not build training content. It curates links to the best existing
1327
+ resources organized by skill level (Beginner, Intermediate, Advanced, Cost Management).
1328
+ Sources include Anthropic documentation, community guides, and AgentBoot's own
1329
+ reference documentation.
1330
+
1331
+ The resource list lives in a skill (updatable without rebuilding) and is checked
1332
+ periodically for link freshness.
1333
+
1334
+ **Who it serves:** Developers who want to go deeper than `/learn` covers.
1335
+
1336
+ **Why it matters:** AgentBoot should not duplicate existing high-quality training
1337
+ content. Curation is the right approach — point people to the best resources rather
1338
+ than building inferior alternatives.
1339
+
1340
+ #### 6.8.4 Onboarding Checklist
1341
+
1342
+ A generated checklist (`agentboot onboarding-checklist`) built from the organization's
1343
+ actual AgentBoot configuration. Includes: Claude Code installation verification,
1344
+ plugin installation check, first code review, first test generation, persona
1345
+ exploration, personal preference setup, and first `/insights` run.
1346
+
1347
+ Exportable as Markdown or email format for sharing with new team members.
1348
+
1349
+ **Who it serves:** Platform teams onboarding new developers.
1350
+
1351
+ **Why it matters:** A generic onboarding guide is less effective than one built from
1352
+ the organization's actual configuration. The generated checklist uses real persona
1353
+ names, real skill invocations, and real marketplace information.
1354
+
1355
+ #### 6.8.5 Contextual Tips
1356
+
1357
+ Optional tips that appear in persona output when patterns suggest the developer is new:
1358
+ first invocation of a persona, vague prompts, repeated rephrasing. Rate-limited (max
1359
+ 1 per session), disable-able by the developer, and generated by the persona itself
1360
+ (part of the persona prompt, not a separate system).
1361
+
1362
+ **Who it serves:** New developers learning to use personas effectively.
1363
+
1364
+ **Why it matters:** Tips at the moment of need are more effective than documentation.
1365
+ A developer who types "review this" and gets a tip about being more specific learns
1366
+ faster than one who reads a guide about prompting.
1367
+
1368
+ #### 6.8.6 Org-Authored Tips
1369
+
1370
+ The organization's platform team can add custom onboarding content in an `onboarding/`
1371
+ directory in the personas repo: welcome messages, tips specific to the org's stack and
1372
+ conventions, and links to internal resources. This content is synced alongside persona
1373
+ files.
1374
+
1375
+ **Who it serves:** Organizations with institutional knowledge to transfer.
1376
+
1377
+ **Why it matters:** Org-authored tips capture what a senior engineer would tell a new
1378
+ hire sitting next to them: "the security reviewer is strict about SQL parameterization
1379
+ because we had a production incident." This knowledge transfer is version-controlled,
1380
+ available 24/7, and scales beyond one-on-one conversations.
1381
+
1382
+ ### 6.9 Marketplace
1383
+
1384
+ #### 6.9.1 Core Layer (Layer 1)
1385
+
1386
+ Traits and personas maintained by the AgentBoot project. The reference implementations.
1387
+ Apache 2.0 licensed. Tested in CI on every commit. Included when you run
1388
+ `agentboot setup`.
1389
+
1390
+ Contents: core traits (critical-thinking, structured-output, source-citation,
1391
+ confidence-signaling, audit-trail, schema-awareness), core personas (code-reviewer,
1392
+ security-reviewer, test-generator, test-data-expert), and baseline instructions.
1393
+
1394
+ **Who it serves:** Every AgentBoot user.
1395
+
1396
+ **Why it matters:** The core layer is the foundation. It provides immediate value
1397
+ without any community contribution and establishes the quality bar for the ecosystem.
1398
+
1399
+ #### 6.9.2 Verified Layer (Layer 2)
1400
+
1401
+ Community-contributed content reviewed by AgentBoot maintainers and meeting quality
1402
+ standards. Listed in the official marketplace.
1403
+
1404
+ Quality requirements: reviewed by at least one maintainer, follows format standards,
1405
+ has behavioral tests (at least deterministic), documentation (README, use case,
1406
+ configuration), Apache 2.0 or MIT license, no org-specific content, and attribution
1407
+ to contributor.
1408
+
1409
+ Review process: contributor opens PR, automated checks run (lint, test, format, license
1410
+ scan), maintainer reviews for quality, generalizability, and overlap, accepted content
1411
+ is merged with attribution.
1412
+
1413
+ **Who it serves:** Organizations looking for pre-built governance content for common
1414
+ use cases.
1415
+
1416
+ **Why it matters:** The Verified layer is where the marketplace becomes valuable beyond
1417
+ core. A healthcare organization can install a PHI-awareness trait that has been reviewed,
1418
+ tested, and used by other healthcare organizations — rather than building it from scratch.
1419
+
1420
+ #### 6.9.3 Community Layer (Layer 3)
1421
+
1422
+ Anything published to a public marketplace. Not reviewed by AgentBoot maintainers.
1423
+ Only requirement: valid frontmatter / `agentboot.domain.json` and declared license.
1424
+
1425
+ **Who it serves:** Anyone looking for specialized or experimental governance content.
1426
+
1427
+ **Why it matters:** Low barrier to entry enables experimentation and niche use cases.
1428
+ Community content may be excellent or terrible — the explicit labeling ensures users
1429
+ understand the trust level.
1430
+
1431
+ #### 6.9.4 Contribution Model
1432
+
1433
+ Contributions flow through `agentboot publish` (one-command sharing that strips
1434
+ org-specific content, validates format, generates README, and opens a PR). Attribution
1435
+ is permanent and visible in content frontmatter, contributor profiles, and build output.
1436
+
1437
+ Motivations served: professional reputation (usage stats, linkable profiles), reciprocity
1438
+ (attribution on everything installed), content improvement (feedback from production
1439
+ usage), and org visibility (contributing organizations listed publicly).
1440
+
1441
+ AgentBoot does not gamify contributions — no leaderboards, badges, streak counters,
1442
+ or points. These attract gaming behavior, not quality.
1443
+
1444
+ **Who it serves:** Anyone with governance knowledge to share.
1445
+
1446
+ **Why it matters:** The biggest barrier to contribution is friction, not motivation.
1447
+ A developer who has a great PostgreSQL gotchas file will share it if it takes 30
1448
+ seconds. They will not if it takes 30 minutes. `agentboot publish` makes sharing
1449
+ as close to zero-friction as possible.
1450
+
1451
+ #### 6.9.5 Plugin Packaging
1452
+
1453
+ Each logical marketplace grouping (core, healthcare domain, fintech domain,
1454
+ infrastructure gotchas) becomes a CC plugin installable independently. Cross-listing
1455
+ with complementary projects (such as SuperClaude) enables ecosystem convergence without
1456
+ bundling dependencies.
1457
+
1458
+ **Who it serves:** Claude Code users who want to pick and choose governance content.
1459
+
1460
+ **Why it matters:** Plugin packaging turns marketplace content into one-command
1461
+ installations. A developer can install `ab-healthcare` and immediately have PHI-aware
1462
+ review personas.
1463
+
1464
+ #### 6.9.6 Trust and Quality
1465
+
1466
+ Quality mechanisms: version pinning (orgs pin to specific versions of marketplace
1467
+ content), license scanning (incompatible licenses block the build), trait isolation
1468
+ (a bad community trait cannot break a core persona), trust signals (badges, download
1469
+ count, update freshness, test coverage, compatible versions).
1470
+
1471
+ **Who it serves:** Organizations evaluating marketplace content for production use.
1472
+
1473
+ **Why it matters:** Trust is the currency of a marketplace. Without quality signals,
1474
+ organizations will not use community content in production. The three-layer structure
1475
+ (Core, Verified, Community) with explicit trust levels addresses this.
1476
+
1477
+ ---
1478
+
1479
+ ## 7. Non-Goals
1480
+
1481
+ AgentBoot explicitly does not do the following. These boundaries are intentional
1482
+ design decisions, not missing features.
1483
+
1484
+ ### 7.1 AgentBoot Is Not a Learning Management System
1485
+
1486
+ AgentBoot provides contextual assists and curated resource links, not courses, modules,
1487
+ progress tracking, or certificates. It does not track who completed what training. It
1488
+ does not gate persona access behind training completion. The `/learn` skill is a
1489
+ helpful assistant, not a course catalog.
1490
+
1491
+ The onboarding problem is real, but the solution is lightweight help at the moment of
1492
+ need, not a formal training program. Organizations that need formal AI training should
1493
+ use dedicated LMS platforms.
1494
+
1495
+ ### 7.2 AgentBoot Is Not a Surveillance Tool
1496
+
1497
+ AgentBoot does not collect, transmit, or surface raw developer prompts. It does not
1498
+ rank developers by AI skill or prompt quality. It does not gamify usage with
1499
+ leaderboards, badges, or streaks. It does not correlate AI usage with performance
1500
+ reviews. It does not make AI usage mandatory.
1501
+
1502
+ The privacy model is a design invariant, not a configuration option. An organization
1503
+ that wants conversation monitoring has that capability through their API provider's
1504
+ Compliance API and enterprise data governance tools. AgentBoot is not that channel.
1505
+
1506
+ ### 7.3 AgentBoot Does Not Compete with Content Libraries
1507
+
1508
+ AgentBoot is the governance and distribution layer, not a content library competing
1509
+ with projects like SuperClaude, ArcKit, spec-kit, or Trail of Bits configurations.
1510
+ These projects produce content (traits, agents, skills). AgentBoot distributes,
1511
+ governs, and orchestrates content.
1512
+
1513
+ The recommended relationship is marketplace curation (point to upstream, cross-list)
1514
+ rather than bundling. AgentBoot works without any third-party content; third-party
1515
+ content works without AgentBoot. They are complementary, not competitive.
1516
+
1517
+ ### 7.4 AgentBoot Is Not a Runtime Framework
1518
+
1519
+ AgentBoot is a build tool. It produces artifacts and exits. It has zero runtime
1520
+ footprint. The distributed output works without AgentBoot installed. There is no
1521
+ "AgentBoot runtime" that must be running for personas to function.
1522
+
1523
+ The exception is the MCP server, which is a separate, optional package
1524
+ (`@agentboot/mcp-server`) that runs as a process. Even this is opt-in and not required
1525
+ for core functionality.
1526
+
1527
+ ### 7.5 AgentBoot Does Not Promise Universal Platform Parity
1528
+
1529
+ AgentBoot is honest about platform capabilities. Claude Code gets the full governance
1530
+ surface (hooks, managed settings, subagent isolation, agent memory, plugin system).
1531
+ Non-CC platforms get the content (personas, instructions, rules) but not the
1532
+ enforcement mechanisms.
1533
+
1534
+ AgentBoot documents these gaps per platform rather than pretending all agents are equal
1535
+ or promising features it cannot deliver. The compliance story for non-CC platforms is:
1536
+ instruction-based refusal plus CI-based review plus organizational policy.
1537
+
1538
+ ### 7.6 AgentBoot Does Not Manage AI Models
1539
+
1540
+ AgentBoot specifies which model a persona should use (in persona.config.json and agent
1541
+ frontmatter), but it does not manage model access, API keys, billing, or rate limiting.
1542
+ Those are the responsibility of the AI platform (Anthropic Console, GitHub Settings,
1543
+ Cursor settings).
1544
+
1545
+ ### 7.7 AgentBoot Does Not Replace Code Review
1546
+
1547
+ AI code review is a complement to human code review, not a replacement. AgentBoot
1548
+ personas produce findings that humans evaluate and act on. Even at the highest autonomy
1549
+ level (Phase 3: Autonomous), the output is reviewed post-hoc by humans.
1550
+
1551
+ The test plan explicitly states: "Agents test agents, but humans always decide.
1552
+ Automation removes burden, not judgment."
1553
+
1554
+ ### 7.8 AgentBoot Does Not Do Dynamic Agent Orchestration
1555
+
1556
+ AgentBoot does not provide a runtime agent orchestration system with message passing,
1557
+ state machines, or complex multi-agent workflows. The `/review` meta-skill is a
1558
+ lightweight routing layer that reads a configuration file and dispatches to the
1559
+ appropriate persona. The persona arbitrator (V2+) resolves conflicts between
1560
+ concurrent reviewers. Neither constitutes a full orchestration framework.
1561
+
1562
+ ---
1563
+
1564
+ ## 8. Success Metrics
1565
+
1566
+ ### 8.1 Adoption Metrics
1567
+
1568
+ | Metric | Measurement | Target |
1569
+ |---|---|---|
1570
+ | Organizations using AgentBoot | GitHub template uses, marketplace installs | Growth rate |
1571
+ | Active developers per org | Persona invocation telemetry (anonymous) | 70%+ of licensed seats |
1572
+ | Persona invocations per day (org-wide) | Telemetry aggregate | Trending upward |
1573
+ | Time from setup to first persona invocation | Wizard completion to first `/review-code` | Under 10 minutes |
1574
+ | Repos governed per org | `agentboot status` repo count | Growth rate |
1575
+ | Skeptic conversion rate | Developers who start at zero usage and reach 1+ sessions/week | 30%+ within 90 days |
1576
+
1577
+ ### 8.2 Contribution Metrics
1578
+
1579
+ | Metric | Measurement | Target |
1580
+ |---|---|---|
1581
+ | Marketplace items (Verified) | Marketplace repo count | Growth rate |
1582
+ | Community contributors | Unique PR authors to marketplace | Growth rate |
1583
+ | Contribution-to-install ratio | Items contributed vs. items installed | Healthy ratio (not vanity) |
1584
+ | Domain layers available | Complete domain packages in marketplace | 3+ by V2 |
1585
+ | Trait reuse rate | How many personas compose each trait (average) | 3+ personas per core trait |
1586
+
1587
+ ### 8.3 Quality Metrics
1588
+
1589
+ | Metric | Measurement | Target |
1590
+ |---|---|---|
1591
+ | Persona false positive rate | Rephrase rates, finding dismissal rates | Under 20% |
1592
+ | Bug escape rate delta | Production bugs that a persona should have caught (before/after) | Measurable reduction |
1593
+ | Test coverage delta | Coverage change in repos with test-generator persona | Measurable increase |
1594
+ | PR review turnaround | Time from PR open to first AI review | Under 5 minutes |
1595
+ | Behavioral test pass rate | CI test results | 95%+ |
1596
+ | Lint pass rate on first commit | How often persona changes pass lint without fixes | 80%+ |
1597
+
1598
+ ### 8.4 Cost Metrics
1599
+
1600
+ | Metric | Measurement | Target |
1601
+ |---|---|---|
1602
+ | Avg cost per persona invocation | Telemetry (tokens * model rate) | Trending downward |
1603
+ | Model mix efficiency | % of invocations on Haiku/Sonnet vs. Opus | 80%+ on Sonnet or cheaper |
1604
+ | Automated test cost per month | CI billing | Under $200 for 4 personas |
1605
+ | Token budget compliance | % of personas within token budget | 95%+ |
1606
+ | Cost per finding | Total cost / total findings | Trending downward |
1607
+
1608
+ ### 8.5 Governance Metrics
1609
+
1610
+ | Metric | Measurement | Target |
1611
+ |---|---|---|
1612
+ | Sync coverage | % of repos receiving governance updates | 100% of registered repos |
1613
+ | Managed settings deployment | % of developer machines with HARD guardrails | 100% of managed machines |
1614
+ | Compliance hook activation | % of sessions where compliance hooks are active | 100% on managed machines |
1615
+ | ADR exception count | Approved exceptions via ADR governance | Low and stable |
1616
+ | Escalation flag rate | Harmful content flags per period | Low (zero is ideal) |
1617
+
1618
+ ### 8.6 Developer Experience Metrics
1619
+
1620
+ | Metric | Measurement | Target |
1621
+ |---|---|---|
1622
+ | Setup completion rate | % of users who complete `agentboot setup` | 90%+ |
1623
+ | Uninstall rate | % of orgs that uninstall within 90 days | Under 10% |
1624
+ | `/learn` invocations | How often developers seek help | Trending downward (learning) |
1625
+ | Tip dismissal rate | How often contextual tips are disabled | Under 30% |
1626
+ | NPS (Net Promoter Score) | Developer survey | Positive |
1627
+
1628
+ ---
1629
+
1630
+ ## 9. Open Questions
1631
+
1632
+ Open questions discovered during PRD authoring have been resolved or deferred.
1633
+ See the internal tracking document for remaining items.
1634
+
1635
+ ---
1636
+
1637
+ ## 10. Glossary
1638
+
1639
+ ### Agent
1640
+
1641
+ A Claude Code concept. An agent is a custom AI assistant defined in
1642
+ `.claude/agents/{name}/CLAUDE.md` with rich frontmatter (model, permissionMode,
1643
+ maxTurns, disallowedTools, mcpServers, hooks, memory, isolation). Agents are distinct
1644
+ from the base Claude Code assistant. In AgentBoot, personas are compiled into agents
1645
+ for CC-native output.
1646
+
1647
+ ### ADR (Architecture Decision Record)
1648
+
1649
+ A formal exception governance mechanism. When a developer intentionally deviates from
1650
+ a persona's recommendation, an ADR documents the rationale, gets reviewer approval,
1651
+ and becomes a permanent record. The persona learns to accept the deviation for that
1652
+ specific case. ADRs handle permanent, approved deviations (as opposed to temporary
1653
+ elevation for debugging).
1654
+
1655
+ ### Agent-Agnostic
1656
+
1657
+ Content that works across multiple AI agent platforms without modification. Traits
1658
+ (Markdown), personas (SKILL.md via agentskills.io), and gotchas (Markdown with glob
1659
+ patterns) are agent-agnostic. Hooks, managed settings, and agent frontmatter are
1660
+ agent-specific (Claude Code only).
1661
+
1662
+ ### agentskills.io
1663
+
1664
+ An open standard for AI agent skill definitions. Uses SKILL.md format (Markdown with
1665
+ YAML frontmatter). Supported by 26+ agent platforms. AgentBoot uses agentskills.io as
1666
+ the cross-platform persona format.
1667
+
1668
+ ### Always-On Instructions
1669
+
1670
+ Universal guardrails distributed to every repo regardless of persona configuration.
1671
+ These load at session start and remain active throughout. Used for org-wide rules that
1672
+ apply to every AI interaction (security baselines, coding standards, compliance
1673
+ requirements).
1674
+
1675
+ ### Autonomy Progression
1676
+
1677
+ A three-phase model for persona independence. Phase 1 (Advisory): persona produces
1678
+ findings, human decides. Phase 2 (Auto-approve): persona applies low-risk fixes
1679
+ automatically, high-risk findings require human review. Phase 3 (Autonomous): persona
1680
+ operates independently, human reviews post-hoc. Promotion between phases is a
1681
+ governance decision requiring evidence (false-positive rates, test coverage, team
1682
+ approval).
1683
+
1684
+ ### Build Pipeline
1685
+
1686
+ The three-stage process: validate (pre-build checks), compile (resolve traits, produce
1687
+ output), sync (distribute to repos). The pipeline is: validate, compile, sync.
1688
+
1689
+ ### CC-First Delivery
1690
+
1691
+ The principle that Claude Code is the primary delivery target. Content is agent-agnostic
1692
+ (portable Markdown). Delivery leverages CC's full feature surface (plugins, hooks,
1693
+ managed settings, MCP, agents, skills, rules). Non-CC platforms get the content but
1694
+ not the enforcement mechanisms.
1695
+
1696
+ ### Cowork
1697
+
1698
+ Anthropic's desktop application for non-technical users. Cowork plugins are the same
1699
+ format as Claude Code plugins but appear in a GUI with form-based input rather than
1700
+ slash commands. AgentBoot personas packaged as plugins work in both CC and Cowork.
1701
+
1702
+ ### Compilation Target
1703
+
1704
+ One of two output formats produced by `agentboot build`. The cross-platform target
1705
+ produces standalone SKILL.md files with traits inlined. The CC-native target produces
1706
+ the full `.claude/` directory structure with @imports, rich frontmatter, hooks, and
1707
+ MCP configuration.
1708
+
1709
+ ### Convention Over Configuration
1710
+
1711
+ The principle that AgentBoot ships with sensible defaults for everything. Organizations
1712
+ configure what is different about their situation, not everything from scratch. Borrowed
1713
+ from the Spring Boot design philosophy.
1714
+
1715
+ ### Domain Layer
1716
+
1717
+ A complete package of traits, personas, gotchas, and instructions for a specific
1718
+ compliance regime or technology stack. Examples: healthcare-compliance (PHI, HIPAA,
1719
+ FHIR), fintech-compliance (PCI-DSS, SOX), govtech-fedramp. Domain layers are the
1720
+ highest-value, highest-effort marketplace contribution.
1721
+
1722
+ ### Gotcha (Gotchas Rule)
1723
+
1724
+ A path-scoped instruction that encodes hard-won operational knowledge. Activated when
1725
+ a developer works on files matching the glob pattern. Invisible (zero context cost)
1726
+ when working on unrelated files. Technology-specific, not org-specific — making them
1727
+ highly shareable. Example: "PostgreSQL partitions do NOT inherit relrowsecurity."
1728
+
1729
+ ### HARD Guardrail
1730
+
1731
+ A non-overridable compliance rule deployed via MDM (managed settings) or marked
1732
+ `required: true` in the org config. Cannot be elevated, overridden, or disabled at any
1733
+ scope level. A team-level config that attempts to disable a HARD guardrail causes a
1734
+ build failure. Used for rules where violation is a compliance incident (PHI scanning,
1735
+ credential blocking, audit logging).
1736
+
1737
+ ### Hub-and-Spoke Distribution
1738
+
1739
+ The distribution model. One central repository (the hub, the personas repo) contains
1740
+ the source of truth. Target repositories (spokes, the actual codebases) receive
1741
+ compiled artifacts via the sync pipeline. One-way flow: hub publishes, spokes receive.
1742
+ Spokes do not produce governance.
1743
+
1744
+ ### Managed Settings
1745
+
1746
+ Claude Code configuration files deployed to OS-level paths
1747
+ (`/Library/Application Support/ClaudeCode/` on macOS, `/etc/claude-code/` on Linux)
1748
+ via MDM. Cannot be overridden by any user or project setting. Used for HARD guardrails
1749
+ and forced plugin installation.
1750
+
1751
+ ### Marketplace
1752
+
1753
+ The three-layer ecosystem for sharing governance content. Core (maintained by AgentBoot),
1754
+ Verified (community-contributed, reviewed), Community (unreviewed, buyer-beware).
1755
+ Physically, a marketplace is a Git repository with a marketplace.json catalog.
1756
+
1757
+ ### MCP (Model Context Protocol)
1758
+
1759
+ A protocol for AI agents to interact with external tools and data sources. MCP servers
1760
+ expose tools and resources that agents can consume. Supported by Claude Code, Copilot,
1761
+ Cursor, and Gemini CLI. AgentBoot uses MCP for cross-platform persona serving and
1762
+ knowledge base access.
1763
+
1764
+ ### Persona
1765
+
1766
+ A complete, deployable agent. A composition of traits plus a specialized system prompt
1767
+ that defines the agent's identity, operating context, and mandate. Personas compose
1768
+ traits; they do not inherit from each other. In AgentBoot, personas are defined as
1769
+ SKILL.md files and compiled into platform-specific output.
1770
+
1771
+ ### persona.config.json
1772
+
1773
+ Build metadata for a persona. Specifies which traits to compose, at what weight, the
1774
+ target model, permission mode, tool restrictions, MCP servers, hooks, and autonomy
1775
+ level. Read by the compile step to generate output.
1776
+
1777
+ ### Persona Arbitrator
1778
+
1779
+ A dedicated persona (V2+) that resolves conflicts when multiple reviewer personas
1780
+ produce contradictory findings on the same code. Only invoked when the `/review`
1781
+ meta-skill detects conflicting findings. Not invoked on every review.
1782
+
1783
+ ### Plugin
1784
+
1785
+ A Claude Code distribution unit that bundles agents, skills, hooks, rules, MCP
1786
+ configuration, and settings into a single installable package. Distributed via
1787
+ marketplaces (public or private). The primary delivery method for CC users.
1788
+
1789
+ ### Prompts as Code
1790
+
1791
+ The principle that AI agent behavior is treated as infrastructure: defined in files,
1792
+ stored in version control, reviewed in pull requests, tested, linted, and measured.
1793
+ Analogous to Infrastructure as Code (Terraform) and Configuration as Code (Kubernetes).
1794
+
1795
+ ### Scope Hierarchy
1796
+
1797
+ The four-level organizational model: Org, Group, Team, Repo. More specific scopes
1798
+ layer on top of general ones. Optional behaviors follow "most specific wins." Mandatory
1799
+ behaviors follow "most general wins." Determines how configuration is merged during
1800
+ sync.
1801
+
1802
+ ### Self-Improvement Reflections
1803
+
1804
+ An optional mechanism where personas write brief reflections after completing their
1805
+ task. Reflections accumulate into a dataset revealing patterns. Three phases:
1806
+ Phase A (humans edit personas from observation), Phase B (reflections + review skill),
1807
+ Phase C (automated accuracy tracking).
1808
+
1809
+ ### SKILL.md
1810
+
1811
+ The agentskills.io format for persona definitions. Markdown with YAML frontmatter
1812
+ (name, description, traits, scope, output format) followed by the system prompt in
1813
+ prose. The cross-platform persona format.
1814
+
1815
+ ### SME Discoverability Fragment
1816
+
1817
+ A lightweight always-on CLAUDE.md section (~100 tokens) auto-generated by the build
1818
+ system. Lists all available personas and how to invoke them. Makes personas
1819
+ discoverable without consulting external documentation.
1820
+
1821
+ ### SOFT Guardrail
1822
+
1823
+ An important default that can be temporarily elevated. Deployed via the shared repo
1824
+ (not MDM). Elevation mechanism: developer invokes `/elevate` with a reason, receives
1825
+ time-bounded bypass (default 30 minutes), audit log entry created, guardrail
1826
+ automatically re-engages on TTL expiry.
1827
+
1828
+ ### Structured Telemetry
1829
+
1830
+ Persona invocation metrics emitted as structured JSON (GELF/NDJSON format). Fields
1831
+ include persona_id, model, scope, input/output tokens, cost, findings, duration, and
1832
+ timestamp. No developer ID by default. No prompt text. Human-queryable with `jq`.
1833
+
1834
+ ### Team Champion
1835
+
1836
+ A designated engineer on each team (typically tech lead or senior IC) who runs sync,
1837
+ reviews sync PRs, files quality feedback, onboards teammates, and proposes governance
1838
+ improvements. A rotating responsibility taking minutes per week in steady state.
1839
+
1840
+ ### Trait
1841
+
1842
+ A reusable behavioral building block for an AI persona. Captures a single aspect of
1843
+ how an agent should think or communicate. Supports weight configurations
1844
+ (HIGH/MEDIUM/LOW or 0.0-1.0). Composed at build time. The DRY principle applied to
1845
+ AI behavior. Write once, compose everywhere, improve in one place.
1846
+
1847
+ ### Trait Weight
1848
+
1849
+ A calibration system for traits that support variable intensity. Named weights
1850
+ (HIGH/MEDIUM/LOW) map to numeric values (0.7/0.5/0.3). Additional values: OFF (0.0)
1851
+ and MAX (1.0). The weight adjusts the threshold for action, not the type of action.
1852
+ At any weight, CRITICAL findings always surface.
1853
+
1854
+ ### Two-Channel MDM Distribution
1855
+
1856
+ Enterprise distribution model separating non-negotiable enforcement (Channel 1: MDM-
1857
+ deployed managed settings for HARD guardrails) from team-customizable configuration
1858
+ (Channel 2: Git-based hub-and-spoke for SOFT guardrails, traits, personas, skills).
1859
+
1860
+ ---
1861
+
1862
+ *End of Product Requirements Document.*