agentboot 0.1.0 → 0.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +8 -7
- package/agentboot.config.json +4 -1
- package/package.json +2 -2
- package/scripts/cli.ts +42 -14
- package/scripts/compile.ts +30 -7
- package/scripts/dev-sync.ts +1 -1
- package/scripts/lib/config.ts +17 -1
- package/scripts/validate.ts +12 -7
- package/.github/ISSUE_TEMPLATE/persona-request.md +0 -62
- package/.github/ISSUE_TEMPLATE/quality-feedback.md +0 -67
- package/.github/workflows/cla.yml +0 -25
- package/.github/workflows/validate.yml +0 -49
- package/.idea/agentboot.iml +0 -9
- package/.idea/misc.xml +0 -6
- package/.idea/modules.xml +0 -8
- package/.idea/vcs.xml +0 -6
- package/CLAUDE.md +0 -230
- package/CONTRIBUTING.md +0 -168
- package/PERSONAS.md +0 -156
- package/core/instructions/baseline.instructions.md +0 -133
- package/core/instructions/security.instructions.md +0 -186
- package/core/personas/code-reviewer/SKILL.md +0 -175
- package/core/personas/security-reviewer/SKILL.md +0 -233
- package/core/personas/test-data-expert/SKILL.md +0 -234
- package/core/personas/test-generator/SKILL.md +0 -262
- package/core/traits/audit-trail.md +0 -182
- package/core/traits/confidence-signaling.md +0 -172
- package/core/traits/critical-thinking.md +0 -129
- package/core/traits/schema-awareness.md +0 -132
- package/core/traits/source-citation.md +0 -174
- package/core/traits/structured-output.md +0 -199
- package/docs/ci-cd-automation.md +0 -548
- package/docs/claude-code-reference/README.md +0 -21
- package/docs/claude-code-reference/agentboot-coverage.md +0 -484
- package/docs/claude-code-reference/feature-inventory.md +0 -906
- package/docs/cli-commands-audit.md +0 -112
- package/docs/cli-design.md +0 -924
- package/docs/concepts.md +0 -1117
- package/docs/config-schema-audit.md +0 -121
- package/docs/configuration.md +0 -645
- package/docs/delivery-methods.md +0 -758
- package/docs/developer-onboarding.md +0 -342
- package/docs/extending.md +0 -448
- package/docs/getting-started.md +0 -298
- package/docs/knowledge-layer.md +0 -464
- package/docs/marketplace.md +0 -822
- package/docs/org-connection.md +0 -570
- package/docs/plans/architecture.md +0 -2429
- package/docs/plans/design.md +0 -2018
- package/docs/plans/prd.md +0 -1862
- package/docs/plans/stack-rank.md +0 -261
- package/docs/plans/technical-spec.md +0 -2755
- package/docs/privacy-and-safety.md +0 -807
- package/docs/prompt-optimization.md +0 -1071
- package/docs/test-plan.md +0 -972
- package/docs/third-party-ecosystem.md +0 -496
- package/domains/compliance-template/README.md +0 -173
- package/domains/compliance-template/traits/compliance-aware.md +0 -228
- package/examples/enterprise/agentboot.config.json +0 -184
- package/examples/minimal/agentboot.config.json +0 -46
- package/tests/REGRESSION-PLAN.md +0 -705
- package/tests/TEST-PLAN.md +0 -111
- package/tests/cli.test.ts +0 -705
- package/tests/pipeline.test.ts +0 -608
- package/tests/validate.test.ts +0 -278
- package/tsconfig.json +0 -62
package/docs/plans/prd.md
DELETED
|
@@ -1,1862 +0,0 @@
|
|
|
1
|
-
# AgentBoot — Product Requirements Document
|
|
2
|
-
|
|
3
|
-
**Version:** 1.0.0-draft
|
|
4
|
-
**Date:** 2026-03-19
|
|
5
|
-
**Author:** AgentBoot Team
|
|
6
|
-
**License:** Apache 2.0
|
|
7
|
-
**Status:** Draft — awaiting review
|
|
8
|
-
|
|
9
|
-
---
|
|
10
|
-
|
|
11
|
-
## Table of Contents
|
|
12
|
-
|
|
13
|
-
1. [Executive Summary](#1-executive-summary)
|
|
14
|
-
2. [Problem Statement](#2-problem-statement)
|
|
15
|
-
3. [Target Users](#3-target-users)
|
|
16
|
-
4. [Product Vision](#4-product-vision)
|
|
17
|
-
5. [Design Principles](#5-design-principles)
|
|
18
|
-
6. [Feature Requirements](#6-feature-requirements)
|
|
19
|
-
7. [Non-Goals](#7-non-goals)
|
|
20
|
-
8. [Success Metrics](#8-success-metrics)
|
|
21
|
-
9. [Open Questions](#9-open-questions)
|
|
22
|
-
10. [Glossary](#10-glossary)
|
|
23
|
-
|
|
24
|
-
---
|
|
25
|
-
|
|
26
|
-
## 1. Executive Summary
|
|
27
|
-
|
|
28
|
-
### What AgentBoot Is
|
|
29
|
-
|
|
30
|
-
AgentBoot is a build tool and governance framework for AI agent behavior in software
|
|
31
|
-
engineering organizations. It provides convention-over-configuration defaults for
|
|
32
|
-
composing, distributing, and managing AI personas across development teams — so that
|
|
33
|
-
every engineer in an organization gets consistent, high-quality AI agent behavior
|
|
34
|
-
from day one without building it themselves.
|
|
35
|
-
|
|
36
|
-
### The Problem
|
|
37
|
-
|
|
38
|
-
Organizations adopting agentic development tools (Claude Code, GitHub Copilot, Cursor,
|
|
39
|
-
Gemini CLI) face a fragmentation crisis. Every team writes its own CLAUDE.md, its own
|
|
40
|
-
rules, its own agents — independently, inconsistently, and without governance. There is
|
|
41
|
-
no standard for distributing agent behavior across repos, no mechanism for enforcing
|
|
42
|
-
compliance guardrails, no way to measure whether AI personas are delivering value, and
|
|
43
|
-
no path for sharing improvements across teams. The result is duplicated effort, drifting
|
|
44
|
-
quality, compliance gaps, and developer frustration.
|
|
45
|
-
|
|
46
|
-
### The Solution
|
|
47
|
-
|
|
48
|
-
AgentBoot applies the same discipline to AI agent behavior that Spring Boot applied to
|
|
49
|
-
Java application configuration: sensible defaults, composable building blocks, a scope
|
|
50
|
-
hierarchy that mirrors organizational structure, and a build-then-distribute pipeline
|
|
51
|
-
that propagates governance automatically. Persona definitions are treated as code —
|
|
52
|
-
version-controlled, reviewed in pull requests, tested, linted, and distributed through
|
|
53
|
-
a hub-and-spoke model. The framework is agent-agnostic in content (using open standards
|
|
54
|
-
like agentskills.io) while being Claude Code-first in delivery (leveraging the richest
|
|
55
|
-
extensibility surface available).
|
|
56
|
-
|
|
57
|
-
### Target User
|
|
58
|
-
|
|
59
|
-
AgentBoot serves six distinct user segments across the organization — from individual
|
|
60
|
-
developers who consume personas, to platform/DevEx teams who build and distribute them,
|
|
61
|
-
to IT/Security teams who enforce compliance, to engineering leaders who measure ROI.
|
|
62
|
-
Each segment interacts with AgentBoot through a different surface (plugin, CLI, managed
|
|
63
|
-
settings, dashboard) tailored to their role.
|
|
64
|
-
|
|
65
|
-
### Origin
|
|
66
|
-
|
|
67
|
-
AgentBoot started with personal and family projects — building AI personas for small
|
|
68
|
-
codebases where the author needed structured review, test generation, and domain
|
|
69
|
-
expertise. The patterns grew organically as the number of projects increased, then
|
|
70
|
-
were adapted for workplace use when it became clear that the same governance gaps
|
|
71
|
-
existed at organizational scale. What began as a personal toolkit evolved into a
|
|
72
|
-
framework when the realization hit: every team adopting agentic development was going
|
|
73
|
-
to build this anyway, and most would build it poorly.
|
|
74
|
-
|
|
75
|
-
---
|
|
76
|
-
|
|
77
|
-
## 2. Problem Statement
|
|
78
|
-
|
|
79
|
-
### 2.1 The Fragmentation Problem
|
|
80
|
-
|
|
81
|
-
Every team that adopts AI coding tools ends up doing the same thing: writing a
|
|
82
|
-
CLAUDE.md, defining some rules, maybe adding a few custom agents. Most of these are
|
|
83
|
-
mediocre. New teams do not know where to start. Teams in the same organization drift
|
|
84
|
-
apart. Nobody shares anything.
|
|
85
|
-
|
|
86
|
-
Meanwhile, the teams that invest in building it well — composable traits, structured
|
|
87
|
-
reviewers, compliance guardrails, a distribution pipeline — cannot share their work
|
|
88
|
-
because it is buried in proprietary context.
|
|
89
|
-
|
|
90
|
-
### 2.2 Specific Pain Points
|
|
91
|
-
|
|
92
|
-
**Scattered, inconsistent agent configuration.** In a typical organization with 20+
|
|
93
|
-
repositories, each repo has its own CLAUDE.md written independently. An audit of such
|
|
94
|
-
organizations typically reveals that 67% or more of repos have some version of the same
|
|
95
|
-
instructions (e.g., "use TypeScript strict mode"), but expressed in 4+ different
|
|
96
|
-
wordings, maintained independently, and never synchronized. When best practice evolves,
|
|
97
|
-
there is no mechanism to propagate the improvement.
|
|
98
|
-
|
|
99
|
-
**No governance structure.** Agent behavior is not reviewed, versioned, or audited. A
|
|
100
|
-
developer can change a CLAUDE.md with no pull request, no review, and no record of what
|
|
101
|
-
changed or why. When an agent produces a bad result, there is no diff to examine. When a
|
|
102
|
-
compliance rule needs to be added, there is no guarantee it reaches every repo.
|
|
103
|
-
|
|
104
|
-
**No distribution mechanism.** Organizations lack a way to push agent behavior updates
|
|
105
|
-
to all their repos simultaneously. Each repo is an island. A platform team that writes
|
|
106
|
-
a better code reviewer has no standard path to deploy it across 50 repos without
|
|
107
|
-
manually touching each one.
|
|
108
|
-
|
|
109
|
-
**No compliance enforcement.** Regulated industries (healthcare, finance, government)
|
|
110
|
-
need non-negotiable guardrails — PHI scanning, credential blocking, audit logging — that
|
|
111
|
-
cannot be disabled by any developer. Current agent platforms offer this capability
|
|
112
|
-
(managed settings, hooks), but there is no framework for generating, testing, and
|
|
113
|
-
deploying these guardrails systematically.
|
|
114
|
-
|
|
115
|
-
**No metrics or measurement.** Organizations cannot answer basic questions: Which
|
|
116
|
-
personas are used most? What do they cost? What is the false positive rate? Are they
|
|
117
|
-
improving over time? Without measurement, prompt optimization is guesswork, budget
|
|
118
|
-
justification is impossible, and the investment case for agentic development is based
|
|
119
|
-
on anecdote rather than evidence.
|
|
120
|
-
|
|
121
|
-
**Developer resistance.** Skeptics and hold-outs will not adopt AI tools voluntarily.
|
|
122
|
-
Without a mechanism for zero-friction deployment (managed settings, repo-embedded
|
|
123
|
-
configuration) and a privacy model that builds trust (no prompt surveillance), a
|
|
124
|
-
significant fraction of the engineering organization will never engage.
|
|
125
|
-
|
|
126
|
-
**No sharing or reuse.** Knowledge that one team gains — a PostgreSQL gotcha, a
|
|
127
|
-
compliance trait, a well-tuned reviewer — stays siloed. There is no marketplace, no
|
|
128
|
-
contribution model, and no standard format that enables sharing across teams or across
|
|
129
|
-
organizations.
|
|
130
|
-
|
|
131
|
-
**Context bloat.** CLAUDE.md files grow unbounded because no one prunes them. An audit
|
|
132
|
-
of a typical mid-size organization reveals average CLAUDE.md files of 230+ lines, with
|
|
133
|
-
the longest exceeding 800 lines. Most of this content is duplicated across repos, vague
|
|
134
|
-
("follow best practices"), or stale ("TODO: update this section").
|
|
135
|
-
|
|
136
|
-
**Vendor lock-in anxiety.** Teams hesitate to invest deeply in agent configuration for
|
|
137
|
-
one platform because they worry about being locked in. If the organization switches
|
|
138
|
-
from Claude Code to Copilot (or vice versa), will they have to rebuild everything?
|
|
139
|
-
|
|
140
|
-
**No onboarding path.** New developers join the organization and face a wall of AI
|
|
141
|
-
tooling they have never used. They type vague prompts, get vague results, conclude
|
|
142
|
-
"AI does not work for me," and abandon the tools. There is no guided path from
|
|
143
|
-
first interaction to productive use.
|
|
144
|
-
|
|
145
|
-
### 2.3 Why These Problems Persist
|
|
146
|
-
|
|
147
|
-
These problems persist because they are coordination problems, not technology problems.
|
|
148
|
-
Any single team can solve them for themselves — but the solution does not scale without
|
|
149
|
-
a framework. Writing a good CLAUDE.md is easy. Writing a governance system that keeps
|
|
150
|
-
50 repos consistent, enforces compliance, measures effectiveness, and onboards new
|
|
151
|
-
developers is hard. That is what AgentBoot provides.
|
|
152
|
-
|
|
153
|
-
---
|
|
154
|
-
|
|
155
|
-
## 3. Target Users
|
|
156
|
-
|
|
157
|
-
AgentBoot serves six distinct user segments. Each has different needs, different pain
|
|
158
|
-
points, and a different definition of success.
|
|
159
|
-
|
|
160
|
-
### 3.1 Developer
|
|
161
|
-
|
|
162
|
-
**Profile:** Individual contributor who writes code daily. May or may not already use
|
|
163
|
-
AI coding tools. Ranges from junior to staff level.
|
|
164
|
-
|
|
165
|
-
**Needs:**
|
|
166
|
-
- AI personas that work immediately when they clone a repo — zero setup
|
|
167
|
-
- Clear invocation patterns (`/review-code`, `/gen-tests`) that give structured results
|
|
168
|
-
- Privacy: confidence that their conversations, questions, and false starts are private
|
|
169
|
-
- The ability to ask follow-up questions about persona findings
|
|
170
|
-
- Control over their experience: disable tips, choose models, set preferences
|
|
171
|
-
|
|
172
|
-
**Pain points:**
|
|
173
|
-
- Does not know where to start with AI tools
|
|
174
|
-
- Gets vague results from vague prompts but does not know how to improve
|
|
175
|
-
- Fears being judged for questions asked to AI ("what is a foreign key?")
|
|
176
|
-
- Persona output is confusing — unclear severity levels, no actionable suggestions
|
|
177
|
-
- Multiple tools (Claude Code, Copilot, Cursor) with different configuration mechanisms
|
|
178
|
-
|
|
179
|
-
**What success looks like:**
|
|
180
|
-
- Clones a repo, types `/review-code`, gets a structured review with actionable findings
|
|
181
|
-
- Trusts that their AI conversations are private (the "IDE vs PR" mental model)
|
|
182
|
-
- Gradually discovers more personas and becomes more effective over weeks
|
|
183
|
-
- Never runs `agentboot` directly — the framework is invisible to them
|
|
184
|
-
|
|
185
|
-
### 3.2 Platform / DevEx Team
|
|
186
|
-
|
|
187
|
-
**Profile:** Engineers responsible for developer tooling, productivity, and
|
|
188
|
-
infrastructure. Typically 1-5 people in a mid-size org. May be a dedicated DevEx team
|
|
189
|
-
or a platform engineering team with tooling responsibilities.
|
|
190
|
-
|
|
191
|
-
**Needs:**
|
|
192
|
-
- A central repository for managing all AI persona definitions
|
|
193
|
-
- A build pipeline that compiles, validates, and distributes personas
|
|
194
|
-
- Multi-platform support (some teams use Claude Code, others use Copilot)
|
|
195
|
-
- Scaffolding tools for creating new personas, traits, domains, and gotchas
|
|
196
|
-
- Discovery tools for finding and consolidating existing agentic content across repos
|
|
197
|
-
- Testing tools that validate persona behavior before deployment
|
|
198
|
-
- A marketplace for sharing and discovering community content
|
|
199
|
-
|
|
200
|
-
**Pain points:**
|
|
201
|
-
- Every team has written their own CLAUDE.md independently — massive duplication
|
|
202
|
-
- No way to push an update to 50 repos without touching each one
|
|
203
|
-
- No standard format for defining agent behavior
|
|
204
|
-
- Cannot test whether a persona change improves or regresses behavior
|
|
205
|
-
- Teams resist centralization because they fear losing control of their configuration
|
|
206
|
-
|
|
207
|
-
**What success looks like:**
|
|
208
|
-
- Single `agentboot build && agentboot sync` deploys updates to every registered repo
|
|
209
|
-
- New personas and traits go through PR review with automated validation
|
|
210
|
-
- `agentboot discover` consolidates scattered agentic content into a governed hub
|
|
211
|
-
- Teams can extend (but not break) the org-wide configuration through the scope hierarchy
|
|
212
|
-
- Monthly testing cost under $200 for the entire persona suite
|
|
213
|
-
|
|
214
|
-
### 3.3 Engineering Leader
|
|
215
|
-
|
|
216
|
-
**Profile:** VP of Engineering, CTO, Director of Engineering. Responsible for
|
|
217
|
-
engineering investment decisions, team productivity, and technical strategy.
|
|
218
|
-
|
|
219
|
-
**Needs:**
|
|
220
|
-
- Evidence that AI tool investment is delivering ROI
|
|
221
|
-
- Adoption metrics: how many developers are using the tools, how often
|
|
222
|
-
- Cost metrics: total spend, spend per team, spend per persona
|
|
223
|
-
- Quality metrics: are bug escape rates improving? Is review turnaround faster?
|
|
224
|
-
- Confidence that governance scales as the organization grows
|
|
225
|
-
|
|
226
|
-
**Pain points:**
|
|
227
|
-
- Cannot justify AI tool budgets without data
|
|
228
|
-
- Does not know whether teams are actually using the tools or ignoring them
|
|
229
|
-
- Worries about compliance exposure in regulated domains
|
|
230
|
-
- Has seen AI initiatives fail because of poor adoption and lack of governance
|
|
231
|
-
|
|
232
|
-
**What success looks like:**
|
|
233
|
-
- Monthly dashboard showing adoption rates, cost breakdown, and ROI indicators
|
|
234
|
-
- Evidence of improved PR review turnaround, test coverage, and bug escape rates
|
|
235
|
-
- Confidence that compliance guardrails are active on every repo
|
|
236
|
-
- Can demonstrate value to the board with concrete metrics, not anecdotes
|
|
237
|
-
|
|
238
|
-
### 3.4 IT / Security
|
|
239
|
-
|
|
240
|
-
**Profile:** IT operations, security engineering, or compliance team. Responsible for
|
|
241
|
-
device management, security policy, and regulatory compliance.
|
|
242
|
-
|
|
243
|
-
**Needs:**
|
|
244
|
-
- Non-overridable guardrails that no developer can disable (HARD guardrails)
|
|
245
|
-
- Deployment via MDM (Jamf, Intune, JumpCloud) or server-managed settings
|
|
246
|
-
- Audit trail for compliance reporting
|
|
247
|
-
- PHI/PII scanning hooks that run on every interaction
|
|
248
|
-
- Credential blocking and secret detection
|
|
249
|
-
- Zero-touch deployment: developers should not have to opt in to compliance
|
|
250
|
-
|
|
251
|
-
**Pain points:**
|
|
252
|
-
- No way to enforce AI compliance across the organization
|
|
253
|
-
- Developer-configured AI tools are a compliance blind spot
|
|
254
|
-
- Cannot guarantee that PHI scanning runs on every Claude Code session
|
|
255
|
-
- No audit evidence of AI governance for regulatory reporting
|
|
256
|
-
|
|
257
|
-
**What success looks like:**
|
|
258
|
-
- Managed settings deployed via MDM — compliance hooks active on every machine
|
|
259
|
-
- HARD guardrails that cannot be overridden at any scope level
|
|
260
|
-
- Audit trail of persona invocations (without individual prompt content)
|
|
261
|
-
- Clean compliance audits with evidence of AI governance
|
|
262
|
-
- Escalation mechanism for genuinely harmful content (flags, not transcripts)
|
|
263
|
-
|
|
264
|
-
### 3.5 Skeptic / Hold-out
|
|
265
|
-
|
|
266
|
-
**Profile:** Experienced engineer who is resistant to AI tools. May have philosophical
|
|
267
|
-
objections, may have had bad experiences, or may simply prefer their existing workflow.
|
|
268
|
-
This segment exists in every organization and will never be zero.
|
|
269
|
-
|
|
270
|
-
**Needs:**
|
|
271
|
-
- Proof of value without disrupting their workflow
|
|
272
|
-
- Opt-in experience with no forced adoption
|
|
273
|
-
- Honest assessment of AI limitations, not hype
|
|
274
|
-
- Assurance that AI usage is not monitored or judged
|
|
275
|
-
|
|
276
|
-
**Pain points:**
|
|
277
|
-
- Feels pressured to use AI tools they do not trust
|
|
278
|
-
- Fears that AI metrics will be used in performance reviews
|
|
279
|
-
- Finds AI output noisy, inaccurate, or unhelpful
|
|
280
|
-
- Does not want to learn yet another tool
|
|
281
|
-
|
|
282
|
-
**What success looks like:**
|
|
283
|
-
- Compliance guardrails activate automatically without their action (managed settings)
|
|
284
|
-
- Persona-generated PR reviews appear in their workflow without them invoking anything
|
|
285
|
-
- Over time, sees the value in structured AI review and begins invoking personas manually
|
|
286
|
-
- Is never shamed, ranked, or penalized for low AI usage
|
|
287
|
-
- Can cleanly remove AgentBoot with one command if they decide it is not for them
|
|
288
|
-
|
|
289
|
-
### 3.6 Non-Engineer
|
|
290
|
-
|
|
291
|
-
**Profile:** Product manager, designer, compliance officer, marketing professional.
|
|
292
|
-
Does not use the terminal. Needs AI assistance for structured tasks (document review,
|
|
293
|
-
compliance checking, content analysis).
|
|
294
|
-
|
|
295
|
-
**Needs:**
|
|
296
|
-
- GUI-first experience with structured forms, not slash commands
|
|
297
|
-
- Role-specific personas (compliance review, document analysis, content checking)
|
|
298
|
-
- Same privacy protections as engineers
|
|
299
|
-
|
|
300
|
-
**Pain points:**
|
|
301
|
-
- Terminal-based tools are inaccessible
|
|
302
|
-
- Does not understand "install a plugin" or "run a command"
|
|
303
|
-
- AI tools feel designed exclusively for engineers
|
|
304
|
-
|
|
305
|
-
**What success looks like:**
|
|
306
|
-
- Opens a desktop application (Cowork), sees role-specific tools in a sidebar
|
|
307
|
-
- Fills out a form, gets structured output — never touches a terminal
|
|
308
|
-
- Same personas, same quality, same governance as the engineering tools
|
|
309
|
-
|
|
310
|
-
---
|
|
311
|
-
|
|
312
|
-
## 4. Product Vision
|
|
313
|
-
|
|
314
|
-
### 4.1 Positioning
|
|
315
|
-
|
|
316
|
-
AgentBoot is the Spring Boot of AI agent governance: an opinionated, drop-in foundation
|
|
317
|
-
that gives engineering organizations consistent AI agent behavior from day one without
|
|
318
|
-
building it themselves.
|
|
319
|
-
|
|
320
|
-
The analogy is precise:
|
|
321
|
-
|
|
322
|
-
| Spring Boot | AgentBoot |
|
|
323
|
-
|---|---|
|
|
324
|
-
| Convention over configuration for Java apps | Convention over configuration for AI agent behavior |
|
|
325
|
-
| Sensible defaults, override what is different | Sensible persona defaults, extend for your domain |
|
|
326
|
-
| Starter dependencies compose features | Traits compose behavioral building blocks |
|
|
327
|
-
| Auto-configuration detects and configures | `agentboot setup` detects role, tools, org and configures |
|
|
328
|
-
| Spring Boot Actuator for observability | Structured telemetry and `/insights` for measurement |
|
|
329
|
-
| Spring Initializr for scaffolding | `agentboot setup` and `agentboot add` for scaffolding |
|
|
330
|
-
|
|
331
|
-
### 4.2 The Build Tool Mental Model
|
|
332
|
-
|
|
333
|
-
AgentBoot is a build tool, not a runtime framework. It produces artifacts that are
|
|
334
|
-
consumed by AI agent platforms. The pipeline is: validate, compile, distribute.
|
|
335
|
-
|
|
336
|
-
```
|
|
337
|
-
Source files Build step Distributed artifacts
|
|
338
|
-
(traits, personas, → agentboot build → (.claude/, copilot-
|
|
339
|
-
instructions, gotchas) instructions.md, plugins)
|
|
340
|
-
```
|
|
341
|
-
|
|
342
|
-
This distinction matters because:
|
|
343
|
-
- AgentBoot has zero runtime footprint — it produces files and exits
|
|
344
|
-
- The distributed artifacts work without AgentBoot installed
|
|
345
|
-
- Any platform that reads Markdown, YAML, or JSON can consume the output
|
|
346
|
-
- The build step is where governance happens (validation, composition, linting)
|
|
347
|
-
- Teams can inspect, understand, and modify the distributed output
|
|
348
|
-
|
|
349
|
-
### 4.3 Where AgentBoot Is Going
|
|
350
|
-
|
|
351
|
-
**Phase 1 (Foundation):** Build system, CLI, core personas, hub-and-spoke distribution,
|
|
352
|
-
scope hierarchy. The minimum viable governance system.
|
|
353
|
-
|
|
354
|
-
**Phase 2 (Native Distribution):** Claude Code plugin packaging, private marketplace,
|
|
355
|
-
managed settings generation. Native delivery through the platform's own mechanisms.
|
|
356
|
-
|
|
357
|
-
**Phase 3 (Cross-Platform and Enterprise):** MCP server for multi-agent organizations,
|
|
358
|
-
cross-platform output generation (Copilot, Cursor, Gemini), structured knowledge store.
|
|
359
|
-
|
|
360
|
-
**Phase 4 (Ecosystem):** Public marketplace, community contributions, domain layers for
|
|
361
|
-
regulated industries, SuperClaude partnership for trait format standardization.
|
|
362
|
-
|
|
363
|
-
### 4.4 The Long-Term Vision
|
|
364
|
-
|
|
365
|
-
AgentBoot becomes the place where AI agent governance content lives — the "npm of AI
|
|
366
|
-
personas." Not because of marketing, but because that is where the traits, personas,
|
|
367
|
-
gotchas, and domain layers are. Organizations that need governance adopt AgentBoot
|
|
368
|
-
because it is the standard. Contributors share their knowledge because it reaches the
|
|
369
|
-
most people. The ecosystem grows because useful content attracts users who become
|
|
370
|
-
contributors who add more useful content.
|
|
371
|
-
|
|
372
|
-
---
|
|
373
|
-
|
|
374
|
-
## 5. Design Principles
|
|
375
|
-
|
|
376
|
-
Ordered by priority. These principles inform every design decision and resolve
|
|
377
|
-
ambiguity when requirements conflict.
|
|
378
|
-
|
|
379
|
-
### 5.1 Composition Over Inheritance (Priority: Highest)
|
|
380
|
-
|
|
381
|
-
Personas compose traits. They do not inherit from each other. If two personas share
|
|
382
|
-
behavior, that behavior belongs in a trait that both compose.
|
|
383
|
-
|
|
384
|
-
This is the foundational design constraint, established as Design Principle #1 from
|
|
385
|
-
the earliest AgentBoot designs. Object-oriented inheritance applied to personas
|
|
386
|
-
("security-reviewer extends code-reviewer extends base-reviewer") creates fragile
|
|
387
|
-
chains where changes to a parent have unpredictable effects on children. Composition
|
|
388
|
-
is flat, predictable, and inspectable.
|
|
389
|
-
|
|
390
|
-
**Implication:** There is no persona hierarchy. There is no `extends` keyword. Traits
|
|
391
|
-
are the unit of reuse. Personas are the unit of deployment.
|
|
392
|
-
|
|
393
|
-
### 5.2 Scoped Conventions (Priority: Highest)
|
|
394
|
-
|
|
395
|
-
AgentBoot models organizations as a four-level hierarchy: Org, Group, Team, Repo.
|
|
396
|
-
More specific scopes layer on top of general ones. For optional behaviors, team
|
|
397
|
-
overrides group overrides core. For mandatory behaviors, inheritance is top-down
|
|
398
|
-
(org wins).
|
|
399
|
-
|
|
400
|
-
This hierarchy mirrors how real organizations structure responsibility and governance.
|
|
401
|
-
It ensures that governance propagates downward automatically (a new team gets all
|
|
402
|
-
org-level and group-level configuration without manual setup) while preserving team
|
|
403
|
-
autonomy on things that are genuinely team-specific.
|
|
404
|
-
|
|
405
|
-
**Implication:** The scope hierarchy is not optional. Every piece of configuration has
|
|
406
|
-
a scope. Conflict resolution is deterministic: mandatory rules cascade down, optional
|
|
407
|
-
rules override up.
|
|
408
|
-
|
|
409
|
-
### 5.3 Prompts as Code (Priority: High)
|
|
410
|
-
|
|
411
|
-
AI agent behavior is infrastructure: defined in files, stored in version control,
|
|
412
|
-
reviewed in pull requests, with a complete audit history. This is the same shift that
|
|
413
|
-
happened with Infrastructure as Code (Terraform, Pulumi) and Configuration as Code
|
|
414
|
-
(Kubernetes manifests, GitHub Actions workflows).
|
|
415
|
-
|
|
416
|
-
Every change to an agent's behavior is a pull request with a description and a review.
|
|
417
|
-
Traits and personas have version numbers. The sync pipeline propagates updates
|
|
418
|
-
automatically after the PR merges.
|
|
419
|
-
|
|
420
|
-
**Implication:** All persona definitions live in a Git repository. All changes go
|
|
421
|
-
through PR review. The build system enforces validation rules that block invalid
|
|
422
|
-
configurations from merging.
|
|
423
|
-
|
|
424
|
-
### 5.4 Privacy First (Priority: High)
|
|
425
|
-
|
|
426
|
-
AgentBoot will never collect, transmit, or surface raw developer prompts to
|
|
427
|
-
organizational dashboards, managers, or analytics pipelines. This is a design
|
|
428
|
-
invariant, not a configuration option.
|
|
429
|
-
|
|
430
|
-
The organization gets aggregate patterns, not transcripts. High rephrase rates are
|
|
431
|
-
framed as persona quality problems, not developer intelligence problems. The
|
|
432
|
-
`rawPrompts` configuration section contains three `false` fields that cannot be set
|
|
433
|
-
to `true`. They exist to make the design intent explicit.
|
|
434
|
-
|
|
435
|
-
AgentBoot is honest about the boundaries: this commitment covers what AgentBoot does.
|
|
436
|
-
It does not and cannot override what the API provider or the organization does
|
|
437
|
-
independently through other channels (Compliance API, network monitoring, data exports).
|
|
438
|
-
|
|
439
|
-
**Implication:** No telemetry field contains prompt text. No dashboard shows individual
|
|
440
|
-
developer activity. No feature ranks, shames, gamifies, or surveys developer AI usage.
|
|
441
|
-
The escalation exception (genuinely harmful content) flags the category, never the
|
|
442
|
-
transcript.
|
|
443
|
-
|
|
444
|
-
### 5.5 Agent-Agnostic Content, Claude Code-First Delivery (Priority: High)
|
|
445
|
-
|
|
446
|
-
Content (traits, personas, gotchas, instructions) is written in agent-agnostic formats
|
|
447
|
-
(Markdown, agentskills.io SKILL.md). Delivery (plugins, managed settings, hooks,
|
|
448
|
-
native agent frontmatter) leverages Claude Code's extensibility surface to the fullest.
|
|
449
|
-
|
|
450
|
-
If a feature "just does not work without Claude Code" — that is acceptable. Non-CC
|
|
451
|
-
platforms get the content (SKILL.md, instruction files, MCP servers); CC platforms get
|
|
452
|
-
the full governance surface (hooks, managed settings, subagent isolation, agent memory).
|
|
453
|
-
AgentBoot documents this gap honestly rather than pretending all agents are equal.
|
|
454
|
-
|
|
455
|
-
**Implication:** Marketplace content is stored in agent-agnostic format. The build
|
|
456
|
-
system produces platform-specific output. CC-specific enhancements go in metadata
|
|
457
|
-
(persona.config.json), not in the content itself.
|
|
458
|
-
|
|
459
|
-
### 5.6 Easy Exit Builds Trust for Easy Entry (Priority: High)
|
|
460
|
-
|
|
461
|
-
`agentboot uninstall` is a first-class command, not an afterthought. It tracks what it
|
|
462
|
-
manages (via a manifest), archives what it replaces, and restores what was there before.
|
|
463
|
-
An organization that knows they can cleanly remove the tool in one command is more
|
|
464
|
-
willing to try it.
|
|
465
|
-
|
|
466
|
-
No vendor lock-in, no orphaned files, no scavenger hunt. The uninstall command removes
|
|
467
|
-
exactly the files AgentBoot manages and preserves everything the team authored locally.
|
|
468
|
-
|
|
469
|
-
**Implication:** Every synced file is tracked in `.agentboot-manifest.json`. Original
|
|
470
|
-
files are archived during setup. The uninstall path is tested in integration tests.
|
|
471
|
-
|
|
472
|
-
### 5.7 Convention Over Configuration (Priority: Medium-High)
|
|
473
|
-
|
|
474
|
-
AgentBoot ships with sensible defaults for everything. Organizations configure what is
|
|
475
|
-
different about their organization, not everything from scratch. A single config file
|
|
476
|
-
(`agentboot.config.json`) and a `build && sync` command are sufficient to get started.
|
|
477
|
-
|
|
478
|
-
**Implication:** Core personas and traits ship ready to deploy. The setup wizard
|
|
479
|
-
determines the right configuration based on role detection, not manual specification.
|
|
480
|
-
The "happy path" requires minimal configuration.
|
|
481
|
-
|
|
482
|
-
### 5.8 Build-Time Composition, Not Runtime Resolution (Priority: Medium)
|
|
483
|
-
|
|
484
|
-
Traits are composed into personas at build time. The compiled output is complete and
|
|
485
|
-
standalone. Agents receive a single file (or a set of files with @imports for CC-native
|
|
486
|
-
output) with all trait content already resolved. No runtime resolution, no dynamic
|
|
487
|
-
includes on platforms that do not support them.
|
|
488
|
-
|
|
489
|
-
The exception is Claude Code-native output, where `@import` in CLAUDE.md composes
|
|
490
|
-
trait files at load time — but these files are pre-generated by the build system,
|
|
491
|
-
not resolved dynamically.
|
|
492
|
-
|
|
493
|
-
**Implication:** The cross-platform SKILL.md output is fully self-contained. The
|
|
494
|
-
CC-native output uses @imports but all imported files are generated and present.
|
|
495
|
-
No runtime API calls to resolve trait definitions.
|
|
496
|
-
|
|
497
|
-
### 5.9 Honest Limitations (Priority: Medium)
|
|
498
|
-
|
|
499
|
-
AgentBoot documents what works, what does not, and what the actual limitations are on
|
|
500
|
-
each platform. Compliance hooks exist on Claude Code but not on Copilot or Cursor.
|
|
501
|
-
Managed settings are CC-only. Subagent isolation requires `context: fork`, which is
|
|
502
|
-
CC-only. These gaps are documented per platform rather than hidden.
|
|
503
|
-
|
|
504
|
-
**Implication:** Per-platform feature matrices are part of the documentation.
|
|
505
|
-
Non-CC delivery is labeled "best-effort." No feature is marketed as cross-platform
|
|
506
|
-
unless it actually works cross-platform.
|
|
507
|
-
|
|
508
|
-
### 5.10 Hub-and-Spoke Distribution (Priority: Medium)
|
|
509
|
-
|
|
510
|
-
One central repository (the hub) contains the source of truth for all persona
|
|
511
|
-
definitions. Target repositories (spokes) receive compiled artifacts via the sync
|
|
512
|
-
pipeline. Spokes are passive — they receive governance, they do not produce it.
|
|
513
|
-
Teams customize through the hub's scope hierarchy, not by committing persona files
|
|
514
|
-
directly to their repos.
|
|
515
|
-
|
|
516
|
-
**Implication:** The sync pipeline is one-way (hub to spoke). Local modifications
|
|
517
|
-
to synced files are detected and warned about. The hub repository is the authoritative
|
|
518
|
-
source for governance decisions.
|
|
519
|
-
|
|
520
|
-
---
|
|
521
|
-
|
|
522
|
-
## 6. Feature Requirements
|
|
523
|
-
|
|
524
|
-
### 6.1 Build System
|
|
525
|
-
|
|
526
|
-
#### 6.1.1 Validation (`agentboot validate`)
|
|
527
|
-
|
|
528
|
-
Pre-build validation that catches configuration errors before compilation. Runs five
|
|
529
|
-
checks: persona existence (every persona referenced in config has a directory),
|
|
530
|
-
trait reference resolution (every trait referenced by a persona exists), SKILL.md
|
|
531
|
-
frontmatter validity (required fields present, values in allowed ranges), PERSONAS.md
|
|
532
|
-
synchronization (generated registry matches actual persona definitions), and secret
|
|
533
|
-
scanning (no credentials, API keys, or internal URLs in persona content).
|
|
534
|
-
|
|
535
|
-
**Who it serves:** Platform teams and CI pipelines. Validation runs automatically as
|
|
536
|
-
the first stage of the build pipeline and as a standalone CI gate on every PR to the
|
|
537
|
-
personas repository.
|
|
538
|
-
|
|
539
|
-
**Why it matters:** Without validation, invalid configurations silently produce broken
|
|
540
|
-
output. A missing trait reference could result in an empty persona. A credential
|
|
541
|
-
accidentally included in a trait file could be synced to every repo in the
|
|
542
|
-
organization. Validation catches these errors before they propagate.
|
|
543
|
-
|
|
544
|
-
#### 6.1.2 Compilation (`agentboot build`)
|
|
545
|
-
|
|
546
|
-
The core build step. Loads `agentboot.config.json`, resolves trait references from
|
|
547
|
-
each persona's `persona.config.json`, and produces two distinct compilation targets:
|
|
548
|
-
cross-platform output and Claude Code-native output.
|
|
549
|
-
|
|
550
|
-
**Cross-platform output** produces standalone SKILL.md files (agentskills.io format
|
|
551
|
-
with traits inlined), copilot-instructions.md, and generic CLAUDE.md. This output
|
|
552
|
-
works on any platform that reads Markdown instruction files.
|
|
553
|
-
|
|
554
|
-
**Claude Code-native output** produces the full `.claude/` directory structure:
|
|
555
|
-
agents with rich frontmatter (model, permissionMode, maxTurns, disallowedTools,
|
|
556
|
-
mcpServers, hooks, memory, isolation), skills with `context: fork`, rules with
|
|
557
|
-
`paths:` frontmatter, separate trait files for `@import` composition, a CLAUDE.md
|
|
558
|
-
using `@imports`, settings.json with hook entries, and `.mcp.json` with server
|
|
559
|
-
configurations.
|
|
560
|
-
|
|
561
|
-
**Who it serves:** Platform teams running the build pipeline. CI systems that produce
|
|
562
|
-
artifacts on merge.
|
|
563
|
-
|
|
564
|
-
**Why it matters:** The two-target compilation is what makes AgentBoot both
|
|
565
|
-
cross-platform and CC-optimized. Organizations choose which format to deploy per repo.
|
|
566
|
-
Claude Code users get the full governance surface (hooks, managed settings, isolation).
|
|
567
|
-
Non-CC users get the content (personas, instructions, rules) without the enforcement
|
|
568
|
-
mechanisms.
|
|
569
|
-
|
|
570
|
-
#### 6.1.3 Sync (`agentboot sync`)
|
|
571
|
-
|
|
572
|
-
Distributes compiled output to target repositories. Reads `repos.json`, which lists
|
|
573
|
-
each target repo with its platform and team assignment. Merges scopes according to the
|
|
574
|
-
four-level hierarchy (org, group, team, repo), with team-level configuration layering
|
|
575
|
-
on top of group and org configuration. Writes the appropriate output format based on
|
|
576
|
-
each repo's declared platform.
|
|
577
|
-
|
|
578
|
-
Sync supports multiple modes: local filesystem writes, GitHub API (creates PRs via
|
|
579
|
-
the API), and GitLab API (creates merge requests). A dry-run mode shows what would
|
|
580
|
-
change without writing anything. Each sync writes an `.agentboot-manifest.json`
|
|
581
|
-
tracking every managed file and its content hash, enabling clean uninstall and
|
|
582
|
-
drift detection.
|
|
583
|
-
|
|
584
|
-
**Who it serves:** Platform teams and CI pipelines. In steady state, sync runs
|
|
585
|
-
automatically on merge to the personas repo's main branch.
|
|
586
|
-
|
|
587
|
-
**Why it matters:** Sync is the mechanism that turns the hub-and-spoke model from
|
|
588
|
-
theory into practice. Without it, every repo update is manual. With it, a single
|
|
589
|
-
PR to the personas repo propagates governance to every registered repo
|
|
590
|
-
automatically — with human review at the PR merge point as the governance checkpoint.
|
|
591
|
-
|
|
592
|
-
#### 6.1.4 Scope Merging
|
|
593
|
-
|
|
594
|
-
The scope merge algorithm implements the four-level hierarchy. For each target repo,
|
|
595
|
-
the sync step composes: org-level configuration (always active), group-level
|
|
596
|
-
configuration (based on the repo's group assignment), team-level configuration (based
|
|
597
|
-
on the repo's team assignment), and repo-level overrides.
|
|
598
|
-
|
|
599
|
-
Optional behaviors follow "most specific wins" — a team can override a group default.
|
|
600
|
-
Mandatory behaviors follow "most general wins" — an org can mark traits or personas
|
|
601
|
-
as required, and no lower scope can disable them. A team-level configuration that
|
|
602
|
-
attempts to disable a mandatory org-level guardrail causes a build failure.
|
|
603
|
-
|
|
604
|
-
**Who it serves:** Organizations with multiple groups and teams that need both
|
|
605
|
-
centralized governance and team-level customization.
|
|
606
|
-
|
|
607
|
-
**Why it matters:** Without scope merging, organizations face a false choice between
|
|
608
|
-
centralized control (which frustrates teams) and team autonomy (which breaks
|
|
609
|
-
governance). The scope hierarchy provides both: mandatory rules propagate down,
|
|
610
|
-
optional customization layers up.
|
|
611
|
-
|
|
612
|
-
### 6.2 CLI
|
|
613
|
-
|
|
614
|
-
#### 6.2.1 Interactive Setup Wizard (`agentboot setup`)
|
|
615
|
-
|
|
616
|
-
The primary entry point for all users. An interactive wizard that asks a series of
|
|
617
|
-
questions to determine the user's role (developer, platform team, IT/security,
|
|
618
|
-
exploring), their AI tools (Claude Code, Copilot, Cursor, mixed), their organization
|
|
619
|
-
context (existing AgentBoot setup, no setup, solo), team size, compliance requirements,
|
|
620
|
-
and device management infrastructure.
|
|
621
|
-
|
|
622
|
-
Based on the answers, the wizard executes the appropriate setup: Quick Start (solo
|
|
623
|
-
developer — deploys files directly to the current repo), Standard Setup (platform
|
|
624
|
-
team — scaffolds the org personas repo with config, personas, and traits), or
|
|
625
|
-
Enterprise Setup (IT — generates managed settings for MDM deployment plus marketplace
|
|
626
|
-
template).
|
|
627
|
-
|
|
628
|
-
The wizard auto-detects as much as possible before asking: git remote for org name,
|
|
629
|
-
`.claude/` in the current repo for existing setup, managed settings on disk for
|
|
630
|
-
existing governance, installed Claude Code version, and platform tools in PATH.
|
|
631
|
-
|
|
632
|
-
**Who it serves:** Every first-time user, regardless of role.
|
|
633
|
-
|
|
634
|
-
**Why it matters:** The setup wizard is the difference between "read the docs for
|
|
635
|
-
30 minutes then configure" and "answer 3-7 questions and you are running." Progressive
|
|
636
|
-
disclosure ensures that solo developers are not overwhelmed with enterprise options,
|
|
637
|
-
while platform teams get the full scaffolding they need.
|
|
638
|
-
|
|
639
|
-
#### 6.2.2 Discovery (`agentboot discover`)
|
|
640
|
-
|
|
641
|
-
Scans repositories, machines, and configurations for existing AI agent content —
|
|
642
|
-
CLAUDE.md files, Copilot instructions, Cursor rules, MCP servers, hooks, skills,
|
|
643
|
-
agents, and anything else that represents agentic maturity the organization already
|
|
644
|
-
has. Supports scanning a single repo, a directory of repos, all repos in a GitHub
|
|
645
|
-
organization (via API), local machine configuration, and managed settings paths.
|
|
646
|
-
|
|
647
|
-
Discovery produces five actions: a detailed inventory report (Markdown), a classify-
|
|
648
|
-
and-ingest pipeline (decompose existing content into traits, personas, gotchas, and
|
|
649
|
-
instructions), an overlap analysis (find near-duplicate content across repos), a
|
|
650
|
-
migration plan (what becomes what, with estimated reduction), and a config export
|
|
651
|
-
(generate `agentboot.config.json` from the discovered repo structure).
|
|
652
|
-
|
|
653
|
-
Discovery and migration are non-destructive. They never modify, move, or delete
|
|
654
|
-
existing files. They create new files in the AgentBoot personas repo. The originals
|
|
655
|
-
stay untouched. The organization reviews, tests, and deploys at their pace via PR.
|
|
656
|
-
|
|
657
|
-
**Who it serves:** Platform teams evaluating AgentBoot for an organization that already
|
|
658
|
-
has scattered agentic content.
|
|
659
|
-
|
|
660
|
-
**Why it matters:** Most organizations adopting AgentBoot are not starting from zero.
|
|
661
|
-
They have 20+ repos with independent CLAUDE.md files, custom agents, and instruction
|
|
662
|
-
files. Without discovery, consolidation is a manual audit. With it, AgentBoot finds
|
|
663
|
-
everything, identifies duplication, and produces a migration plan that typically
|
|
664
|
-
achieves 80-85% line reduction through deduplication and centralization.
|
|
665
|
-
|
|
666
|
-
#### 6.2.3 Build and Distribution Commands
|
|
667
|
-
|
|
668
|
-
`agentboot build` compiles personas with options for specific output formats
|
|
669
|
-
(claude-code, copilot, cross-platform, plugin, all) and a validate-only dry run.
|
|
670
|
-
`agentboot sync` distributes to repos with support for local filesystem, GitHub API,
|
|
671
|
-
and GitLab API modes, plus single-repo targeting and dry-run. `agentboot export`
|
|
672
|
-
generates distributable artifacts (plugin directory, marketplace scaffold, managed
|
|
673
|
-
settings for MDM, reusable GitHub Actions workflow, MCP server package).
|
|
674
|
-
`agentboot publish` pushes the plugin to a configured marketplace with version bumping.
|
|
675
|
-
|
|
676
|
-
**Who it serves:** Platform teams managing the build and distribution pipeline.
|
|
677
|
-
|
|
678
|
-
**Why it matters:** These commands form the core build pipeline. They are designed to
|
|
679
|
-
work in both interactive (developer workstation) and non-interactive (CI/CD) modes,
|
|
680
|
-
with every command supporting `--non-interactive` flags.
|
|
681
|
-
|
|
682
|
-
#### 6.2.4 Scaffolding (`agentboot add`)
|
|
683
|
-
|
|
684
|
-
Scaffolds new components: personas, traits, domains, gotchas, hooks, and repo
|
|
685
|
-
registrations. Each scaffold creates the appropriate directory structure with template
|
|
686
|
-
files that follow the format standards.
|
|
687
|
-
|
|
688
|
-
The `add prompt` subcommand ingests raw prompt text from multiple sources (inline text,
|
|
689
|
-
file, clipboard, URL, batch decomposition of an existing CLAUDE.md). It classifies the
|
|
690
|
-
content (trait, gotcha, persona rule, or always-on instruction), formats it to the
|
|
691
|
-
appropriate standard, and saves it to the correct location. A dry-run mode previews
|
|
692
|
-
without writing.
|
|
693
|
-
|
|
694
|
-
**Who it serves:** Platform teams creating new content and anyone contributing to the
|
|
695
|
-
personas repository.
|
|
696
|
-
|
|
697
|
-
**Why it matters:** Scaffolding reduces the barrier to adding new governance content.
|
|
698
|
-
The `add prompt` command is particularly valuable during migration: it takes the
|
|
699
|
-
scattered, unstructured content found by `discover` and transforms it into properly
|
|
700
|
-
formatted AgentBoot content.
|
|
701
|
-
|
|
702
|
-
#### 6.2.5 Diagnostics (`agentboot doctor`, `agentboot status`)
|
|
703
|
-
|
|
704
|
-
`agentboot doctor` checks environment health: Claude Code version, Node.js version,
|
|
705
|
-
git version, configuration validity, persona completeness, sync status, plugin state,
|
|
706
|
-
and managed settings status. Every problem includes a fix command.
|
|
707
|
-
|
|
708
|
-
`agentboot doctor --diagnose` runs a layered isolation test to pinpoint whether a
|
|
709
|
-
problem is in AgentBoot core, org config, org traits, org personas, or org extensions.
|
|
710
|
-
Each layer adds one piece of the stack; the layer where it breaks is the layer that
|
|
711
|
-
has the bug.
|
|
712
|
-
|
|
713
|
-
`agentboot status` provides a deployment dashboard: enabled personas and traits, repo
|
|
714
|
-
registration status with platform and sync version, plugin state, and managed settings
|
|
715
|
-
status.
|
|
716
|
-
|
|
717
|
-
**Who it serves:** Anyone troubleshooting issues. Platform teams monitoring deployment.
|
|
718
|
-
|
|
719
|
-
**Why it matters:** "Is this my bug or AgentBoot's bug?" is a question every
|
|
720
|
-
organization will ask. The diagnostic tools answer it systematically rather than
|
|
721
|
-
requiring manual investigation.
|
|
722
|
-
|
|
723
|
-
#### 6.2.6 Linting (`agentboot lint`)
|
|
724
|
-
|
|
725
|
-
Static analysis of persona definitions. Checks for vague language ("be thorough",
|
|
726
|
-
"try to"), token budget violations, credentials in prompt text, conflicting
|
|
727
|
-
instructions across traits, unused traits, and security issues. Configurable severity
|
|
728
|
-
thresholds (error, warning, info).
|
|
729
|
-
|
|
730
|
-
Token budget analysis calculates per-persona context cost and flags personas that
|
|
731
|
-
exceed configurable limits. This prevents context bloat — the gradual growth of prompt
|
|
732
|
-
content that wastes tokens without adding value.
|
|
733
|
-
|
|
734
|
-
**Who it serves:** Platform teams authoring personas, and CI pipelines gating merges.
|
|
735
|
-
|
|
736
|
-
**Why it matters:** Prompt quality is the product. A governance framework that produces
|
|
737
|
-
vague, bloated, or conflicting prompts is actively wasting money and developer trust.
|
|
738
|
-
Linting enforces a minimum quality bar automatically.
|
|
739
|
-
|
|
740
|
-
#### 6.2.7 Testing (`agentboot test`)
|
|
741
|
-
|
|
742
|
-
Multi-type test runner supporting deterministic tests (free, every commit),
|
|
743
|
-
behavioral tests (LLM invocation, ~$0.50/test), snapshot regression tests (compare
|
|
744
|
-
output across versions), and LLM-as-judge evaluation (Opus evaluates quality on
|
|
745
|
-
multiple dimensions). Supports cost caps, per-persona targeting, CI mode with exit
|
|
746
|
-
codes and JSON output, and flake tolerance (2-of-3 pass).
|
|
747
|
-
|
|
748
|
-
Full details in section 6.7 (Testing).
|
|
749
|
-
|
|
750
|
-
**Who it serves:** Platform teams validating persona behavior before deployment.
|
|
751
|
-
|
|
752
|
-
**Why it matters:** AI persona output is non-deterministic. Traditional testing
|
|
753
|
-
approaches (exact output matching) do not work. The multi-layer test pyramid provides
|
|
754
|
-
confidence at each level: schema tests are free and fast, behavioral tests verify
|
|
755
|
-
that personas catch known bugs, snapshot tests detect regressions, and LLM-as-judge
|
|
756
|
-
evaluates qualitative dimensions that pattern matching cannot assess.
|
|
757
|
-
|
|
758
|
-
#### 6.2.8 Additional Commands
|
|
759
|
-
|
|
760
|
-
`agentboot connect` provides developer self-service for connecting to an organization's
|
|
761
|
-
marketplace (CLI equivalent of `/agentboot:connect`). `agentboot upgrade` pulls the
|
|
762
|
-
latest AgentBoot core into an organization's personas repo. `agentboot validate` is the
|
|
763
|
-
CI-friendly validation-only command with strict mode. `agentboot search` queries the
|
|
764
|
-
marketplace for traits, gotchas, personas, and domains. `agentboot metrics` reads
|
|
765
|
-
telemetry and reports per-persona, per-team, and per-period statistics.
|
|
766
|
-
`agentboot cost-estimate` projects per-persona monthly cost across the organization.
|
|
767
|
-
`agentboot review` generates guided human review sessions with curated samples.
|
|
768
|
-
`agentboot issue` streamlines bug reporting with auto-attached diagnostics and
|
|
769
|
-
privacy-safe redaction. `agentboot uninstall` provides clean removal with manifest
|
|
770
|
-
tracking, original file restoration, and mixed content handling.
|
|
771
|
-
|
|
772
|
-
**Who it serves:** Various segments depending on the command.
|
|
773
|
-
|
|
774
|
-
**Why it matters:** Each command addresses a specific workflow gap. Together, they form
|
|
775
|
-
a complete CLI that covers the full lifecycle from setup through daily use to removal.
|
|
776
|
-
|
|
777
|
-
### 6.3 Persona System
|
|
778
|
-
|
|
779
|
-
#### 6.3.1 Traits
|
|
780
|
-
|
|
781
|
-
Reusable behavioral building blocks. Each trait captures a single aspect of how an
|
|
782
|
-
agent should think or communicate — a cognitive stance, an output discipline, or an
|
|
783
|
-
epistemic commitment. Traits support weight configurations (HIGH / MEDIUM / LOW,
|
|
784
|
-
mapping to a 0.0-1.0 numeric scale) and are composed at build time.
|
|
785
|
-
|
|
786
|
-
Core traits include: critical-thinking (skepticism calibration), structured-output
|
|
787
|
-
(consistent output formatting), source-citation (evidence requirements), confidence-
|
|
788
|
-
signaling (uncertainty communication), audit-trail (compliance logging), and
|
|
789
|
-
schema-awareness (data structure understanding). Additional traits include
|
|
790
|
-
creative-suggestion (proactive improvement ideas) and cost-awareness (resource
|
|
791
|
-
efficiency).
|
|
792
|
-
|
|
793
|
-
The trait weight system is a calibration system, not a priority system. At HIGH, the
|
|
794
|
-
threshold for surfacing a concern is very low. At LOW, the threshold is high. The
|
|
795
|
-
weight does not override the severity floor — at any weight, CRITICAL findings always
|
|
796
|
-
surface. The numeric scale (0.0-1.0) provides finer calibration for advanced use cases
|
|
797
|
-
while the named weights (HIGH/MEDIUM/LOW) serve as the simplified interface.
|
|
798
|
-
|
|
799
|
-
**Who it serves:** Platform teams authoring personas, and the marketplace community
|
|
800
|
-
sharing behavioral building blocks.
|
|
801
|
-
|
|
802
|
-
**Why it matters:** Before traits, every persona independently specified its approach
|
|
803
|
-
to things like skepticism, output structure, and evidence requirements. In practice,
|
|
804
|
-
this meant the same concepts were expressed slightly differently in every persona.
|
|
805
|
-
Traits solve this: write the behavior once, compose it everywhere, improve it in one
|
|
806
|
-
place.
|
|
807
|
-
|
|
808
|
-
#### 6.3.2 Personas
|
|
809
|
-
|
|
810
|
-
Complete, deployable agents composed from traits plus a specialized system prompt that
|
|
811
|
-
defines identity, operating context, and mandate. Use the agentskills.io SKILL.md
|
|
812
|
-
format (Markdown with YAML frontmatter). The frontmatter specifies ID, version, traits,
|
|
813
|
-
scope, and output format. The prose is the system prompt.
|
|
814
|
-
|
|
815
|
-
Core V1 personas: code-reviewer (code review against team standards), security-
|
|
816
|
-
reviewer (vulnerability analysis, OWASP, auth, data handling), test-generator
|
|
817
|
-
(unit and integration test generation from function signatures), and test-data-expert
|
|
818
|
-
(realistic synthetic test data generation).
|
|
819
|
-
|
|
820
|
-
Each persona has a `persona.config.json` that specifies build metadata: which traits
|
|
821
|
-
to compose, at what weight, the target model, permission mode, tool restrictions,
|
|
822
|
-
MCP servers, hooks, and autonomy level.
|
|
823
|
-
|
|
824
|
-
**Who it serves:** All developers (as consumers) and platform teams (as authors).
|
|
825
|
-
|
|
826
|
-
**Why it matters:** Personas are the user-facing product. Everything else in AgentBoot
|
|
827
|
-
— traits, the build system, the scope hierarchy, the distribution pipeline — exists to
|
|
828
|
-
make personas work well. A persona that catches SQL injection, explains the risk,
|
|
829
|
-
suggests the fix, and cites the relevant standard is the experience that makes
|
|
830
|
-
developers trust and adopt the system.
|
|
831
|
-
|
|
832
|
-
#### 6.3.3 Gotchas Rules
|
|
833
|
-
|
|
834
|
-
Path-scoped instructions that encode hard-won operational knowledge. Each gotchas file
|
|
835
|
-
is a Markdown file with `paths:` frontmatter that limits when the rule activates. When
|
|
836
|
-
a developer is working on files that match the glob pattern, the gotchas content is
|
|
837
|
-
automatically included in the agent's context. When working on unrelated files, the
|
|
838
|
-
gotchas are invisible — zero context cost.
|
|
839
|
-
|
|
840
|
-
Gotchas capture the kind of information that lives in one engineer's head until they
|
|
841
|
-
leave: "PostgreSQL partitions do NOT inherit `relrowsecurity`," "UUID primary keys
|
|
842
|
-
cause exponential INSERT slowdown," "Lambda cold starts are 10x worse with VPC
|
|
843
|
-
networking." Technology-specific, not org-specific — making them highly shareable.
|
|
844
|
-
|
|
845
|
-
**Who it serves:** All developers (as consumers), platform teams and individual
|
|
846
|
-
contributors (as authors), and the marketplace community (as shareable knowledge).
|
|
847
|
-
|
|
848
|
-
**Why it matters:** Gotchas rules capture knowledge at the exact moment it is needed.
|
|
849
|
-
A developer writing a database migration sees the PostgreSQL gotchas. A developer
|
|
850
|
-
writing a Lambda handler sees the serverless gotchas. No one has to remember to consult
|
|
851
|
-
a wiki page or ask the right person.
|
|
852
|
-
|
|
853
|
-
#### 6.3.4 Always-On Instructions
|
|
854
|
-
|
|
855
|
-
Universal guardrails distributed to every repo regardless of persona configuration.
|
|
856
|
-
These are the org-wide rules that apply to every AI interaction: security baselines,
|
|
857
|
-
output format standards, coding conventions. They load at session start and remain
|
|
858
|
-
active throughout.
|
|
859
|
-
|
|
860
|
-
**Who it serves:** The entire organization. IT/Security teams who need guarantees
|
|
861
|
-
that certain rules are always active.
|
|
862
|
-
|
|
863
|
-
**Why it matters:** Some rules should never be opt-in. A security baseline, a coding
|
|
864
|
-
standard, a compliance requirement — these must apply everywhere. Always-on
|
|
865
|
-
instructions are the mechanism for universal governance.
|
|
866
|
-
|
|
867
|
-
#### 6.3.5 Domain Layers
|
|
868
|
-
|
|
869
|
-
Complete packages of traits, personas, gotchas, and instructions for a specific
|
|
870
|
-
compliance regime or technology stack. A healthcare domain layer includes PHI-awareness
|
|
871
|
-
traits, HIPAA reviewer personas, patient data gotchas, and compliance-specific
|
|
872
|
-
always-on instructions. A fintech domain layer includes PCI-DSS compliance traits,
|
|
873
|
-
financial data reviewers, and transaction security gotchas.
|
|
874
|
-
|
|
875
|
-
Domain layers are the highest-effort contribution to the marketplace but also the
|
|
876
|
-
highest value. They package everything an organization needs to govern AI behavior
|
|
877
|
-
in a regulated domain.
|
|
878
|
-
|
|
879
|
-
**Who it serves:** Organizations in regulated industries (healthcare, finance,
|
|
880
|
-
government) that need domain-specific compliance guardrails.
|
|
881
|
-
|
|
882
|
-
**Why it matters:** Regulated industries cannot adopt AI tools without compliance
|
|
883
|
-
guardrails. Building these guardrails from scratch is expensive and error-prone.
|
|
884
|
-
Domain layers provide a tested, community-maintained foundation that organizations
|
|
885
|
-
can adopt and extend.
|
|
886
|
-
|
|
887
|
-
#### 6.3.6 Per-Persona Extensions
|
|
888
|
-
|
|
889
|
-
Teams can extend base personas without forking them. An extension file adds
|
|
890
|
-
domain-specific rules to a base persona. The base definition stays unmodified and
|
|
891
|
-
receives upstream improvements automatically. This prevents the anti-pattern of
|
|
892
|
-
copying a persona, modifying it, and maintaining a fork that diverges over time.
|
|
893
|
-
|
|
894
|
-
**Who it serves:** Teams that need team-specific customization of org-wide personas.
|
|
895
|
-
|
|
896
|
-
**Why it matters:** Without extensions, teams face a false choice: use the org persona
|
|
897
|
-
as-is (too generic) or fork it (diverges and rots). Extensions provide the middle path:
|
|
898
|
-
customize behavior while staying on the upstream upgrade path.
|
|
899
|
-
|
|
900
|
-
### 6.4 Delivery and Distribution
|
|
901
|
-
|
|
902
|
-
#### 6.4.1 Claude Code Plugin
|
|
903
|
-
|
|
904
|
-
The primary delivery method for Claude Code users. AgentBoot output is packaged as a
|
|
905
|
-
CC plugin — a directory containing agents, skills, hooks, rules, MCP configuration,
|
|
906
|
-
and settings — distributable through a private marketplace.
|
|
907
|
-
|
|
908
|
-
The organization creates a private marketplace (a Git repository with a
|
|
909
|
-
marketplace.json catalog). Engineers install with a single command:
|
|
910
|
-
`/plugin marketplace add org/personas` and `/plugin install org-plugin`. IT can
|
|
911
|
-
force-enable the plugin via managed settings, requiring zero developer action.
|
|
912
|
-
|
|
913
|
-
Plugin packaging is produced by `agentboot export --format plugin`. The plugin name
|
|
914
|
-
becomes the namespace prefix for all skills (e.g., `/acme:review-code`).
|
|
915
|
-
|
|
916
|
-
**Who it serves:** All Claude Code users in the organization.
|
|
917
|
-
|
|
918
|
-
**Why it matters:** The plugin is the native CC distribution mechanism. It supports
|
|
919
|
-
namespace isolation (preventing conflicts), version pinning, force-enable via managed
|
|
920
|
-
settings, and live reload without restart. It is the path of least resistance for CC
|
|
921
|
-
organizations.
|
|
922
|
-
|
|
923
|
-
#### 6.4.2 Marketplace
|
|
924
|
-
|
|
925
|
-
Three-layer marketplace ecosystem: Core (maintained by AgentBoot, Apache 2.0),
|
|
926
|
-
Verified (community-contributed, reviewed by maintainers, meets quality standards),
|
|
927
|
-
and Community (unreviewed, valid format only, buyer-beware).
|
|
928
|
-
|
|
929
|
-
Full details in section 6.9 (Marketplace).
|
|
930
|
-
|
|
931
|
-
**Who it serves:** All users who need pre-built governance content.
|
|
932
|
-
|
|
933
|
-
**Why it matters:** The marketplace is the mechanism that turns AgentBoot from a
|
|
934
|
-
framework into an ecosystem. A framework solves the governance problem for one
|
|
935
|
-
organization. An ecosystem solves it for the industry.
|
|
936
|
-
|
|
937
|
-
#### 6.4.3 Managed Settings
|
|
938
|
-
|
|
939
|
-
AgentBoot generates managed settings artifacts for deployment via MDM (Jamf, Intune,
|
|
940
|
-
JumpCloud, Kandji) or Anthropic's server-managed settings. Managed settings are
|
|
941
|
-
deployed to OS-level paths and cannot be overridden by any user or project
|
|
942
|
-
configuration.
|
|
943
|
-
|
|
944
|
-
Managed settings carry HARD guardrails only: compliance hooks, credential blocking,
|
|
945
|
-
audit logging, and forced plugin installation. They are not used for persona delivery
|
|
946
|
-
(that is what the plugin and sync are for). This separation keeps compliance and
|
|
947
|
-
convenience in distinct channels.
|
|
948
|
-
|
|
949
|
-
`agentboot export --format managed` produces the artifacts. IT deploys them through
|
|
950
|
-
their existing MDM infrastructure.
|
|
951
|
-
|
|
952
|
-
**Who it serves:** IT/Security teams responsible for compliance enforcement.
|
|
953
|
-
|
|
954
|
-
**Why it matters:** HARD guardrails must be non-overridable. A PHI scanning hook in a
|
|
955
|
-
healthcare organization cannot be disabled by a developer who finds it inconvenient.
|
|
956
|
-
Managed settings are the only mechanism that provides OS-level enforcement.
|
|
957
|
-
|
|
958
|
-
#### 6.4.4 MCP Server
|
|
959
|
-
|
|
960
|
-
AgentBoot exposed as an MCP (Model Context Protocol) server that any MCP-compatible
|
|
961
|
-
agent can consume. Provides persona invocation, trait lookup, governance status, and
|
|
962
|
-
knowledge base access as MCP tools and resources.
|
|
963
|
-
|
|
964
|
-
The MCP server is the cross-platform bridge. It works in Claude Code, Copilot, Cursor,
|
|
965
|
-
Gemini CLI, and any other MCP client. Organizations with mixed toolchains use the MCP
|
|
966
|
-
server to ensure everyone gets governed personas regardless of which tool they use.
|
|
967
|
-
|
|
968
|
-
**Who it serves:** Multi-agent organizations with engineers using different AI tools.
|
|
969
|
-
|
|
970
|
-
**Why it matters:** Without MCP, each platform needs its own delivery mechanism (plugin
|
|
971
|
-
for CC, instruction files for Copilot, rules for Cursor). MCP provides a single
|
|
972
|
-
channel that serves all platforms identically with the same persona definitions and
|
|
973
|
-
invocation interface.
|
|
974
|
-
|
|
975
|
-
#### 6.4.5 Direct Sync
|
|
976
|
-
|
|
977
|
-
The fallback/bootstrap delivery method. AgentBoot's sync script writes compiled files
|
|
978
|
-
directly to target repos' `.claude/`, `.github/`, and `.cursor/` directories. When a
|
|
979
|
-
developer clones the repo, the personas are already there with zero setup.
|
|
980
|
-
|
|
981
|
-
This is the simplest mental model (files in a directory), works offline, and requires
|
|
982
|
-
no plugin system dependency. It scales well to dozens of repos but creates sync PR
|
|
983
|
-
noise at hundreds.
|
|
984
|
-
|
|
985
|
-
**Who it serves:** All developers in repos registered for sync, and organizations that
|
|
986
|
-
cannot use the plugin system.
|
|
987
|
-
|
|
988
|
-
**Why it matters:** The "repo already has it" model provides the lowest-friction
|
|
989
|
-
developer experience. The developer never installs anything — governance is in the
|
|
990
|
-
repo they are already working in.
|
|
991
|
-
|
|
992
|
-
#### 6.4.6 Non-Claude Code Delivery
|
|
993
|
-
|
|
994
|
-
For Copilot: generates `copilot-instructions.md`, path-scoped `.instructions.md`
|
|
995
|
-
files, `.prompt.md` slash command files, and agentskills.io SKILL.md files. For
|
|
996
|
-
Cursor: generates `.cursor/rules/` files, `.cursorrules`, and SKILL.md files. For
|
|
997
|
-
Gemini CLI: generates `GEMINI.md` and SKILL.md files.
|
|
998
|
-
|
|
999
|
-
The `repos.json` platform field determines which format each repo receives. The MCP
|
|
1000
|
-
server is the equalizer for multi-platform organizations.
|
|
1001
|
-
|
|
1002
|
-
**Who it serves:** Organizations using tools other than Claude Code.
|
|
1003
|
-
|
|
1004
|
-
**Why it matters:** Honest cross-platform support is a competitive differentiator.
|
|
1005
|
-
AgentBoot does not pretend all platforms are equal — it documents the governance gaps
|
|
1006
|
-
per platform — but it does provide the best possible experience on each one.
|
|
1007
|
-
|
|
1008
|
-
### 6.5 Privacy and Telemetry
|
|
1009
|
-
|
|
1010
|
-
#### 6.5.1 Three-Tier Privacy Model
|
|
1011
|
-
|
|
1012
|
-
**Tier 1 (Private):** Raw prompts, conversation history, session transcripts, files
|
|
1013
|
-
read during exploration. These stay on the developer's machine. AgentBoot does not
|
|
1014
|
-
read, transmit, or reference them. Retention is session duration or developer's
|
|
1015
|
-
choice.
|
|
1016
|
-
|
|
1017
|
-
**Tier 2 (Privileged):** Aggregated patterns extracted by LLM analysis via the
|
|
1018
|
-
`/insights` skill. The analysis uses the same Claude API trust boundary the developer
|
|
1019
|
-
already uses for every prompt. The developer sees insights first and approves what gets
|
|
1020
|
-
shared. Patterns, not transcripts, are the output.
|
|
1021
|
-
|
|
1022
|
-
**Tier 3 (Organizational):** Persona output metrics — invocation counts, cost,
|
|
1023
|
-
findings severity distribution, model usage, and duration. Anonymized by default (no
|
|
1024
|
-
developer ID). Team-level attribution via the scope field enables budget tracking
|
|
1025
|
-
without identifying individuals.
|
|
1026
|
-
|
|
1027
|
-
**Who it serves:** All users. Developers get privacy. The organization gets metrics.
|
|
1028
|
-
|
|
1029
|
-
**Why it matters:** The privacy model is a product differentiator. In a market where
|
|
1030
|
-
enterprises are deploying AI monitoring tools, AgentBoot takes the opposite stance:
|
|
1031
|
-
it helps organizations improve their AI governance without being the tool that surveys
|
|
1032
|
-
their developers. The best prompt optimization system is one that developers feed
|
|
1033
|
-
willingly because they trust it.
|
|
1034
|
-
|
|
1035
|
-
#### 6.5.2 `/insights` Skill
|
|
1036
|
-
|
|
1037
|
-
A skill that sends session transcripts to the Claude API (Haiku for speed/cost) for
|
|
1038
|
-
pattern extraction. Presents personal prompt insights to the developer: session counts,
|
|
1039
|
-
most-used personas, common topics, rephrase rates, cost, and improvement suggestions.
|
|
1040
|
-
The developer optionally approves sharing anonymized patterns to the org aggregate.
|
|
1041
|
-
|
|
1042
|
-
The analysis prompt is explicitly designed to extract patterns, not judge. It frames
|
|
1043
|
-
everything as persona improvement opportunities, not developer deficiencies. It does
|
|
1044
|
-
not quote developer prompts, judge question quality, identify knowledge deficiencies,
|
|
1045
|
-
or produce output that could embarrass the developer if shared.
|
|
1046
|
-
|
|
1047
|
-
**Who it serves:** Developers (personal analytics) and the organization (aggregate
|
|
1048
|
-
patterns for persona optimization).
|
|
1049
|
-
|
|
1050
|
-
**Why it matters:** Without insights, prompt optimization is guesswork. With insights,
|
|
1051
|
-
the organization knows "auth patterns are asked about 89 times across 23 developers"
|
|
1052
|
-
and can act on it (create an auth-patterns skill). The mechanism respects privacy while
|
|
1053
|
-
enabling organizational learning.
|
|
1054
|
-
|
|
1055
|
-
#### 6.5.3 Org Dashboard
|
|
1056
|
-
|
|
1057
|
-
Aggregate metrics visible to engineering leaders and platform teams: active seats,
|
|
1058
|
-
persona invocations, cost per team, model mix, persona effectiveness (rephrase rates,
|
|
1059
|
-
false positive rates), common knowledge gaps, and cost efficiency analysis.
|
|
1060
|
-
|
|
1061
|
-
The dashboard shows investment metrics (cost, adoption, ROI) and outcome metrics (PR
|
|
1062
|
-
quality, bug rates, coverage). It never shows process metrics (what developers typed,
|
|
1063
|
-
how many times they rephrased, what they asked about).
|
|
1064
|
-
|
|
1065
|
-
Optional per-developer usage tracking with three formats: `false` (no developer
|
|
1066
|
-
identity, the default), `hashed` (consistent anonymous ID), and `email` (full
|
|
1067
|
-
attribution). Full attribution requires clear communication to the team and is
|
|
1068
|
-
intended for cost allocation and license optimization, not surveillance.
|
|
1069
|
-
|
|
1070
|
-
**Who it serves:** Engineering leaders, platform teams, and org owners.
|
|
1071
|
-
|
|
1072
|
-
**Why it matters:** Leaders need evidence that the AI investment is delivering value.
|
|
1073
|
-
The dashboard provides it without violating developer privacy.
|
|
1074
|
-
|
|
1075
|
-
#### 6.5.4 Escalation Exception
|
|
1076
|
-
|
|
1077
|
-
One exception to prompt privacy: genuinely harmful content (attempted exfiltration,
|
|
1078
|
-
guardrail circumvention, harassment, malware generation). The system flags the
|
|
1079
|
-
category locally first (the developer sees the flag), then reports the flag (category
|
|
1080
|
-
and date, not the transcript) to a designated compliance contact. The compliance team
|
|
1081
|
-
can request the transcript through a formal process (like a legal hold), not through
|
|
1082
|
-
the analytics pipeline.
|
|
1083
|
-
|
|
1084
|
-
Implemented via a `UserPromptSubmit` hook using Haiku for fast evaluation. The prompt
|
|
1085
|
-
is explicitly scoped to harmful categories only — not quality, intelligence, or
|
|
1086
|
-
competence.
|
|
1087
|
-
|
|
1088
|
-
**Who it serves:** IT/Security and compliance teams.
|
|
1089
|
-
|
|
1090
|
-
**Why it matters:** Complete privacy absolutism is irresponsible in an enterprise
|
|
1091
|
-
context. The escalation exception provides a narrow, transparent mechanism for
|
|
1092
|
-
genuinely harmful content while maintaining the broader privacy commitment.
|
|
1093
|
-
|
|
1094
|
-
#### 6.5.5 Telemetry Configuration
|
|
1095
|
-
|
|
1096
|
-
Privacy configuration in `agentboot.config.json` includes telemetry settings (enabled,
|
|
1097
|
-
includeDevId, devIdFormat, includeCost, includeScope, destination), insights settings
|
|
1098
|
-
(enabled, autoShareAnonymized, escalation categories and contact), and rawPrompts
|
|
1099
|
-
settings (collect: false, transmit: false, surfaceToOrg: false — fields that cannot be
|
|
1100
|
-
set to true, existing to make design intent explicit).
|
|
1101
|
-
|
|
1102
|
-
**Who it serves:** Platform teams configuring privacy settings for their organization.
|
|
1103
|
-
|
|
1104
|
-
**Why it matters:** Privacy is not a binary. Different organizations have different
|
|
1105
|
-
requirements. The configuration provides a spectrum from fully anonymous to fully
|
|
1106
|
-
attributed, with clear documentation of the implications of each setting.
|
|
1107
|
-
|
|
1108
|
-
### 6.6 Knowledge Layer
|
|
1109
|
-
|
|
1110
|
-
#### 6.6.1 Stage 1: Flat Files (Default)
|
|
1111
|
-
|
|
1112
|
-
The current and default knowledge system. Markdown files loaded into context:
|
|
1113
|
-
always-on instructions at session start, path-scoped rules when matching files are
|
|
1114
|
-
read, trait content composed at build time, and skills loaded on invocation.
|
|
1115
|
-
|
|
1116
|
-
Works for 5-50 knowledge items. Zero infrastructure. Version-controlled in git.
|
|
1117
|
-
Human-readable and editable. Most organizations stay at this stage permanently.
|
|
1118
|
-
|
|
1119
|
-
**Who it serves:** All organizations.
|
|
1120
|
-
|
|
1121
|
-
**Why it matters:** Flat files are the right default. They require no infrastructure,
|
|
1122
|
-
no databases, no embedding APIs. They work immediately and scale to the needs of most
|
|
1123
|
-
teams.
|
|
1124
|
-
|
|
1125
|
-
#### 6.6.2 Stage 2: Structured Knowledge Store
|
|
1126
|
-
|
|
1127
|
-
A queryable layer on top of flat files. Files gain optional frontmatter for structured
|
|
1128
|
-
fields (type, technology, tags, severity, domain, learned_from). The build system
|
|
1129
|
-
generates a SQLite index from this frontmatter. An MCP server exposes tag/category
|
|
1130
|
-
queries: search by technology and severity, get by ID, find related items, and list
|
|
1131
|
-
by type and tag.
|
|
1132
|
-
|
|
1133
|
-
Handles 50-500 knowledge items. Zero new infrastructure (SQLite is a single file
|
|
1134
|
-
shipped with the MCP server). The flat files remain the source of truth — the index
|
|
1135
|
-
is derived, not authoritative.
|
|
1136
|
-
|
|
1137
|
-
**Who it serves:** Mature organizations with extensive accumulated knowledge (gotchas,
|
|
1138
|
-
ADRs, incident learnings, standards, patterns).
|
|
1139
|
-
|
|
1140
|
-
**Why it matters:** When an organization has 200 gotchas and 50 ADRs, loading
|
|
1141
|
-
everything into context is wasteful. Structured queries allow personas to retrieve
|
|
1142
|
-
only the relevant knowledge for the current task.
|
|
1143
|
-
|
|
1144
|
-
#### 6.6.3 Stage 3: Vector / RAG
|
|
1145
|
-
|
|
1146
|
-
Semantic retrieval via embeddings. Instead of querying by tags and categories, the
|
|
1147
|
-
persona describes what it is looking at and the knowledge base returns the most
|
|
1148
|
-
semantically relevant items. Built on sqlite-vss (vector search extension for SQLite),
|
|
1149
|
-
keeping zero new infrastructure.
|
|
1150
|
-
|
|
1151
|
-
The key use case is context-aware review: "this code is doing token refresh validation"
|
|
1152
|
-
retrieves the incident report about token expiry race conditions, even though the code
|
|
1153
|
-
does not mention "race condition." The connection is semantic, not keyword-based.
|
|
1154
|
-
|
|
1155
|
-
Handles 500+ knowledge items. Requires embedding API calls (incremental, only changed
|
|
1156
|
-
items). Cost is minimal (~$0.20 for 1,000 items).
|
|
1157
|
-
|
|
1158
|
-
**Who it serves:** Compliance-heavy industries where accumulated knowledge (incidents,
|
|
1159
|
-
ADRs, patterns) is as valuable as the code itself.
|
|
1160
|
-
|
|
1161
|
-
**Why it matters:** The progression from rules to knowledge to organizational memory
|
|
1162
|
-
is the long-term value of the knowledge layer. A security reviewer that cites last
|
|
1163
|
-
year's incident report is qualitatively different from one that only checks rules.
|
|
1164
|
-
|
|
1165
|
-
#### 6.6.4 Stable MCP Interface
|
|
1166
|
-
|
|
1167
|
-
The MCP interface stays identical across all three stages. Personas do not know or care
|
|
1168
|
-
whether the backing store is flat files, SQLite, or pgvector. They call the same MCP
|
|
1169
|
-
tools. An organization can upgrade from Stage 2 to Stage 3 without rewriting any
|
|
1170
|
-
persona definitions.
|
|
1171
|
-
|
|
1172
|
-
**Who it serves:** Platform teams managing knowledge infrastructure.
|
|
1173
|
-
|
|
1174
|
-
**Why it matters:** The abstraction boundary between personas and the knowledge store
|
|
1175
|
-
means organizations can evolve their knowledge infrastructure without touching their
|
|
1176
|
-
persona definitions. This is why MCP-first matters — the interface is the contract.
|
|
1177
|
-
|
|
1178
|
-
### 6.7 Testing
|
|
1179
|
-
|
|
1180
|
-
#### 6.7.1 Unit and Schema Tests (Layer 1)
|
|
1181
|
-
|
|
1182
|
-
Config validation against JSON schema, frontmatter parsing for all persona SKILL.md
|
|
1183
|
-
files, trait composition (injection markers, weight resolution, missing references,
|
|
1184
|
-
circular dependencies), lint rules (vague language, token budget, credentials,
|
|
1185
|
-
conflicts), and sync logic (scope merging, output format selection, manifest generation,
|
|
1186
|
-
PERSONAS.md registry).
|
|
1187
|
-
|
|
1188
|
-
Free. Runs on every commit. Must pass to merge. Uses vitest.
|
|
1189
|
-
|
|
1190
|
-
**Who it serves:** AgentBoot core maintainers and platform teams with custom content.
|
|
1191
|
-
|
|
1192
|
-
**Why it matters:** The foundation of the test pyramid. Catches configuration errors,
|
|
1193
|
-
format violations, and composition bugs before they reach compilation.
|
|
1194
|
-
|
|
1195
|
-
#### 6.7.2 Integration Tests (Layer 2)
|
|
1196
|
-
|
|
1197
|
-
Full build pipeline (validate, compile, sync produces expected output), CC-native
|
|
1198
|
-
output validation (agent frontmatter, @imports, paths: frontmatter, settings.json
|
|
1199
|
-
hooks, .mcp.json), cross-platform output validation (standalone SKILL.md, copilot-
|
|
1200
|
-
instructions.md), plugin export validation (structure, claude plugin validate),
|
|
1201
|
-
discover and ingest (finds files, identifies duplicates, generates migration plan,
|
|
1202
|
-
non-destructive guarantee), and uninstall (removes managed files only, preserves
|
|
1203
|
-
local files, handles mixed content, restores archive).
|
|
1204
|
-
|
|
1205
|
-
Free. Runs on every commit and PR. Uses vitest with temp directories and test repo
|
|
1206
|
-
fixtures.
|
|
1207
|
-
|
|
1208
|
-
**Who it serves:** AgentBoot core maintainers.
|
|
1209
|
-
|
|
1210
|
-
**Why it matters:** Integration tests verify that the full pipeline works end-to-end.
|
|
1211
|
-
A unit test might pass on trait composition, but the integration test catches when the
|
|
1212
|
-
compiled output misses a hook entry in settings.json.
|
|
1213
|
-
|
|
1214
|
-
#### 6.7.3 Behavioral Tests (Layer 3)
|
|
1215
|
-
|
|
1216
|
-
LLM invocation with known inputs, asserting on output patterns. Test cases defined in
|
|
1217
|
-
YAML format specifying persona, model (Haiku for cost), setup files, prompts, and
|
|
1218
|
-
assertions (findings_min/max, severity_includes/excludes, text_matches/excludes,
|
|
1219
|
-
confidence_min, output_contains, output_structure, json_schema, token_max,
|
|
1220
|
-
duration_max_ms).
|
|
1221
|
-
|
|
1222
|
-
Uses `claude -p` with structured JSON output, bypass permissions, and no session
|
|
1223
|
-
persistence. Flake tolerance (2-of-3 runs) handles non-determinism. Cost: ~$5/PR
|
|
1224
|
-
for 4 personas.
|
|
1225
|
-
|
|
1226
|
-
**Who it serves:** Platform teams validating persona behavior before deployment.
|
|
1227
|
-
|
|
1228
|
-
**Why it matters:** The core novel challenge in testing AI personas. Traditional
|
|
1229
|
-
tests cannot verify that a persona catches a SQL injection vulnerability. Behavioral
|
|
1230
|
-
tests use the cheapest model (Haiku) to verify that the prompt structure elicits the
|
|
1231
|
-
right behavior — if it works on Haiku, it will work better on Sonnet/Opus.
|
|
1232
|
-
|
|
1233
|
-
#### 6.7.4 Snapshot / Regression Tests (Layer 4)
|
|
1234
|
-
|
|
1235
|
-
Compare persona output across versions to detect regressions. Snapshots store
|
|
1236
|
-
structured summaries (finding counts, patterns, tokens, duration), not full prose.
|
|
1237
|
-
Run after persona prompt changes, trait updates, model changes, and weekly for provider
|
|
1238
|
-
drift detection.
|
|
1239
|
-
|
|
1240
|
-
Flags differences for human review: MATCH (identical), CHANGED (new findings — is the
|
|
1241
|
-
change correct?), and REGRESSION (previously caught findings now missed — investigate).
|
|
1242
|
-
|
|
1243
|
-
Cost: ~$5 per persona change.
|
|
1244
|
-
|
|
1245
|
-
**Who it serves:** Platform teams managing persona quality over time.
|
|
1246
|
-
|
|
1247
|
-
**Why it matters:** Without regression testing, a persona improvement might silently
|
|
1248
|
-
break an existing capability. Snapshot tests make these regressions visible before
|
|
1249
|
-
deployment.
|
|
1250
|
-
|
|
1251
|
-
#### 6.7.5 LLM-as-Judge (Layer 5)
|
|
1252
|
-
|
|
1253
|
-
A separate LLM call (Opus as judge) evaluates persona output on qualitative dimensions:
|
|
1254
|
-
completeness, accuracy, specificity, prioritization, and tone. Test cases include
|
|
1255
|
-
ground truth (known issues in the code) so the judge can evaluate whether the persona
|
|
1256
|
-
found what it should have found.
|
|
1257
|
-
|
|
1258
|
-
Run on major persona rewrites, new personas, model migrations, and quarterly quality
|
|
1259
|
-
audits. Cost: ~$20 per run.
|
|
1260
|
-
|
|
1261
|
-
**Who it serves:** Platform teams making major persona changes.
|
|
1262
|
-
|
|
1263
|
-
**Why it matters:** Pattern matching cannot evaluate whether a review is "thorough" or
|
|
1264
|
-
whether the tone is "professional." LLM-as-judge applies consistent qualitative
|
|
1265
|
-
evaluation at scale, filling the gap between structural assertions and human review.
|
|
1266
|
-
|
|
1267
|
-
#### 6.7.6 Human Review (Layer 6)
|
|
1268
|
-
|
|
1269
|
-
Guided human review sessions via `agentboot review`. Presents randomly sampled persona
|
|
1270
|
-
outputs from the last 7 days with structured questions (accuracy, severity, additions,
|
|
1271
|
-
removals). Takes 10-15 minutes per persona.
|
|
1272
|
-
|
|
1273
|
-
Triggers: new persona ships (3-5 real PRs), major trait change (before/after), quarterly
|
|
1274
|
-
audit (random sample of 20 outputs), quality escalation (specific disputed finding).
|
|
1275
|
-
|
|
1276
|
-
**Who it serves:** Platform teams and domain experts.
|
|
1277
|
-
|
|
1278
|
-
**Why it matters:** Humans make judgment calls that no automated test can make. The
|
|
1279
|
-
review tool makes this sustainable (structured, guided, 1 hour/month for 4 personas)
|
|
1280
|
-
rather than ad hoc.
|
|
1281
|
-
|
|
1282
|
-
#### 6.7.7 Mutation Testing
|
|
1283
|
-
|
|
1284
|
-
Deliberately introduces known bugs into persona prompts and verifies that tests catch
|
|
1285
|
-
the regressions. Validates that the test suite actually detects the problems it claims
|
|
1286
|
-
to detect.
|
|
1287
|
-
|
|
1288
|
-
**Who it serves:** Platform teams validating test suite quality.
|
|
1289
|
-
|
|
1290
|
-
**Why it matters:** "Who tests the tests?" Mutation testing answers this question by
|
|
1291
|
-
proving that behavioral tests catch regressions when they should.
|
|
1292
|
-
|
|
1293
|
-
### 6.8 Developer Onboarding
|
|
1294
|
-
|
|
1295
|
-
#### 6.8.1 First-Session Welcome
|
|
1296
|
-
|
|
1297
|
-
A brief orientation (~80 tokens) generated as part of the CLAUDE.md that AgentBoot
|
|
1298
|
-
syncs. Lists available personas with invocation commands, provides 3-4 tips for
|
|
1299
|
-
effective use, mentions the privacy commitment, and directs new users to `/learn`.
|
|
1300
|
-
Loads once per session and fades into background context after a few sessions.
|
|
1301
|
-
|
|
1302
|
-
**Who it serves:** Developers opening Claude Code in an AgentBoot-governed repo for
|
|
1303
|
-
the first time.
|
|
1304
|
-
|
|
1305
|
-
**Why it matters:** The first interaction determines whether a developer engages or
|
|
1306
|
-
abandons. A wall of text loses them. A brief, actionable orientation converts them.
|
|
1307
|
-
|
|
1308
|
-
#### 6.8.2 `/learn` Skill
|
|
1309
|
-
|
|
1310
|
-
A context-aware training assistant, not a static tutorial. Provides a topic browser
|
|
1311
|
-
(Getting Started, Going Deeper, Quick Reference), answers specific questions
|
|
1312
|
-
("how do I review one file?"), and delivers contextual explanations (severity levels,
|
|
1313
|
-
persona concepts, prompting tips).
|
|
1314
|
-
|
|
1315
|
-
The skill uses `disable-model-invocation: true` for static content and normal model
|
|
1316
|
-
invocation for free-form questions.
|
|
1317
|
-
|
|
1318
|
-
**Who it serves:** Developers at any skill level who want to understand the persona
|
|
1319
|
-
system better.
|
|
1320
|
-
|
|
1321
|
-
**Why it matters:** The `/learn` skill provides help at the moment of need. Instead of
|
|
1322
|
-
sending developers to documentation, it answers their question in context.
|
|
1323
|
-
|
|
1324
|
-
#### 6.8.3 Curated External Resources
|
|
1325
|
-
|
|
1326
|
-
AgentBoot does not build training content. It curates links to the best existing
|
|
1327
|
-
resources organized by skill level (Beginner, Intermediate, Advanced, Cost Management).
|
|
1328
|
-
Sources include Anthropic documentation, community guides, and AgentBoot's own
|
|
1329
|
-
reference documentation.
|
|
1330
|
-
|
|
1331
|
-
The resource list lives in a skill (updatable without rebuilding) and is checked
|
|
1332
|
-
periodically for link freshness.
|
|
1333
|
-
|
|
1334
|
-
**Who it serves:** Developers who want to go deeper than `/learn` covers.
|
|
1335
|
-
|
|
1336
|
-
**Why it matters:** AgentBoot should not duplicate existing high-quality training
|
|
1337
|
-
content. Curation is the right approach — point people to the best resources rather
|
|
1338
|
-
than building inferior alternatives.
|
|
1339
|
-
|
|
1340
|
-
#### 6.8.4 Onboarding Checklist
|
|
1341
|
-
|
|
1342
|
-
A generated checklist (`agentboot onboarding-checklist`) built from the organization's
|
|
1343
|
-
actual AgentBoot configuration. Includes: Claude Code installation verification,
|
|
1344
|
-
plugin installation check, first code review, first test generation, persona
|
|
1345
|
-
exploration, personal preference setup, and first `/insights` run.
|
|
1346
|
-
|
|
1347
|
-
Exportable as Markdown or email format for sharing with new team members.
|
|
1348
|
-
|
|
1349
|
-
**Who it serves:** Platform teams onboarding new developers.
|
|
1350
|
-
|
|
1351
|
-
**Why it matters:** A generic onboarding guide is less effective than one built from
|
|
1352
|
-
the organization's actual configuration. The generated checklist uses real persona
|
|
1353
|
-
names, real skill invocations, and real marketplace information.
|
|
1354
|
-
|
|
1355
|
-
#### 6.8.5 Contextual Tips
|
|
1356
|
-
|
|
1357
|
-
Optional tips that appear in persona output when patterns suggest the developer is new:
|
|
1358
|
-
first invocation of a persona, vague prompts, repeated rephrasing. Rate-limited (max
|
|
1359
|
-
1 per session), disable-able by the developer, and generated by the persona itself
|
|
1360
|
-
(part of the persona prompt, not a separate system).
|
|
1361
|
-
|
|
1362
|
-
**Who it serves:** New developers learning to use personas effectively.
|
|
1363
|
-
|
|
1364
|
-
**Why it matters:** Tips at the moment of need are more effective than documentation.
|
|
1365
|
-
A developer who types "review this" and gets a tip about being more specific learns
|
|
1366
|
-
faster than one who reads a guide about prompting.
|
|
1367
|
-
|
|
1368
|
-
#### 6.8.6 Org-Authored Tips
|
|
1369
|
-
|
|
1370
|
-
The organization's platform team can add custom onboarding content in an `onboarding/`
|
|
1371
|
-
directory in the personas repo: welcome messages, tips specific to the org's stack and
|
|
1372
|
-
conventions, and links to internal resources. This content is synced alongside persona
|
|
1373
|
-
files.
|
|
1374
|
-
|
|
1375
|
-
**Who it serves:** Organizations with institutional knowledge to transfer.
|
|
1376
|
-
|
|
1377
|
-
**Why it matters:** Org-authored tips capture what a senior engineer would tell a new
|
|
1378
|
-
hire sitting next to them: "the security reviewer is strict about SQL parameterization
|
|
1379
|
-
because we had a production incident." This knowledge transfer is version-controlled,
|
|
1380
|
-
available 24/7, and scales beyond one-on-one conversations.
|
|
1381
|
-
|
|
1382
|
-
### 6.9 Marketplace
|
|
1383
|
-
|
|
1384
|
-
#### 6.9.1 Core Layer (Layer 1)
|
|
1385
|
-
|
|
1386
|
-
Traits and personas maintained by the AgentBoot project. The reference implementations.
|
|
1387
|
-
Apache 2.0 licensed. Tested in CI on every commit. Included when you run
|
|
1388
|
-
`agentboot setup`.
|
|
1389
|
-
|
|
1390
|
-
Contents: core traits (critical-thinking, structured-output, source-citation,
|
|
1391
|
-
confidence-signaling, audit-trail, schema-awareness), core personas (code-reviewer,
|
|
1392
|
-
security-reviewer, test-generator, test-data-expert), and baseline instructions.
|
|
1393
|
-
|
|
1394
|
-
**Who it serves:** Every AgentBoot user.
|
|
1395
|
-
|
|
1396
|
-
**Why it matters:** The core layer is the foundation. It provides immediate value
|
|
1397
|
-
without any community contribution and establishes the quality bar for the ecosystem.
|
|
1398
|
-
|
|
1399
|
-
#### 6.9.2 Verified Layer (Layer 2)
|
|
1400
|
-
|
|
1401
|
-
Community-contributed content reviewed by AgentBoot maintainers and meeting quality
|
|
1402
|
-
standards. Listed in the official marketplace.
|
|
1403
|
-
|
|
1404
|
-
Quality requirements: reviewed by at least one maintainer, follows format standards,
|
|
1405
|
-
has behavioral tests (at least deterministic), documentation (README, use case,
|
|
1406
|
-
configuration), Apache 2.0 or MIT license, no org-specific content, and attribution
|
|
1407
|
-
to contributor.
|
|
1408
|
-
|
|
1409
|
-
Review process: contributor opens PR, automated checks run (lint, test, format, license
|
|
1410
|
-
scan), maintainer reviews for quality, generalizability, and overlap, accepted content
|
|
1411
|
-
is merged with attribution.
|
|
1412
|
-
|
|
1413
|
-
**Who it serves:** Organizations looking for pre-built governance content for common
|
|
1414
|
-
use cases.
|
|
1415
|
-
|
|
1416
|
-
**Why it matters:** The Verified layer is where the marketplace becomes valuable beyond
|
|
1417
|
-
core. A healthcare organization can install a PHI-awareness trait that has been reviewed,
|
|
1418
|
-
tested, and used by other healthcare organizations — rather than building it from scratch.
|
|
1419
|
-
|
|
1420
|
-
#### 6.9.3 Community Layer (Layer 3)
|
|
1421
|
-
|
|
1422
|
-
Anything published to a public marketplace. Not reviewed by AgentBoot maintainers.
|
|
1423
|
-
Only requirement: valid frontmatter / `agentboot.domain.json` and declared license.
|
|
1424
|
-
|
|
1425
|
-
**Who it serves:** Anyone looking for specialized or experimental governance content.
|
|
1426
|
-
|
|
1427
|
-
**Why it matters:** Low barrier to entry enables experimentation and niche use cases.
|
|
1428
|
-
Community content may be excellent or terrible — the explicit labeling ensures users
|
|
1429
|
-
understand the trust level.
|
|
1430
|
-
|
|
1431
|
-
#### 6.9.4 Contribution Model
|
|
1432
|
-
|
|
1433
|
-
Contributions flow through `agentboot publish` (one-command sharing that strips
|
|
1434
|
-
org-specific content, validates format, generates README, and opens a PR). Attribution
|
|
1435
|
-
is permanent and visible in content frontmatter, contributor profiles, and build output.
|
|
1436
|
-
|
|
1437
|
-
Motivations served: professional reputation (usage stats, linkable profiles), reciprocity
|
|
1438
|
-
(attribution on everything installed), content improvement (feedback from production
|
|
1439
|
-
usage), and org visibility (contributing organizations listed publicly).
|
|
1440
|
-
|
|
1441
|
-
AgentBoot does not gamify contributions — no leaderboards, badges, streak counters,
|
|
1442
|
-
or points. These attract gaming behavior, not quality.
|
|
1443
|
-
|
|
1444
|
-
**Who it serves:** Anyone with governance knowledge to share.
|
|
1445
|
-
|
|
1446
|
-
**Why it matters:** The biggest barrier to contribution is friction, not motivation.
|
|
1447
|
-
A developer who has a great PostgreSQL gotchas file will share it if it takes 30
|
|
1448
|
-
seconds. They will not if it takes 30 minutes. `agentboot publish` makes sharing
|
|
1449
|
-
as close to zero-friction as possible.
|
|
1450
|
-
|
|
1451
|
-
#### 6.9.5 Plugin Packaging
|
|
1452
|
-
|
|
1453
|
-
Each logical marketplace grouping (core, healthcare domain, fintech domain,
|
|
1454
|
-
infrastructure gotchas) becomes a CC plugin installable independently. Cross-listing
|
|
1455
|
-
with complementary projects (such as SuperClaude) enables ecosystem convergence without
|
|
1456
|
-
bundling dependencies.
|
|
1457
|
-
|
|
1458
|
-
**Who it serves:** Claude Code users who want to pick and choose governance content.
|
|
1459
|
-
|
|
1460
|
-
**Why it matters:** Plugin packaging turns marketplace content into one-command
|
|
1461
|
-
installations. A developer can install `ab-healthcare` and immediately have PHI-aware
|
|
1462
|
-
review personas.
|
|
1463
|
-
|
|
1464
|
-
#### 6.9.6 Trust and Quality
|
|
1465
|
-
|
|
1466
|
-
Quality mechanisms: version pinning (orgs pin to specific versions of marketplace
|
|
1467
|
-
content), license scanning (incompatible licenses block the build), trait isolation
|
|
1468
|
-
(a bad community trait cannot break a core persona), trust signals (badges, download
|
|
1469
|
-
count, update freshness, test coverage, compatible versions).
|
|
1470
|
-
|
|
1471
|
-
**Who it serves:** Organizations evaluating marketplace content for production use.
|
|
1472
|
-
|
|
1473
|
-
**Why it matters:** Trust is the currency of a marketplace. Without quality signals,
|
|
1474
|
-
organizations will not use community content in production. The three-layer structure
|
|
1475
|
-
(Core, Verified, Community) with explicit trust levels addresses this.
|
|
1476
|
-
|
|
1477
|
-
---
|
|
1478
|
-
|
|
1479
|
-
## 7. Non-Goals
|
|
1480
|
-
|
|
1481
|
-
AgentBoot explicitly does not do the following. These boundaries are intentional
|
|
1482
|
-
design decisions, not missing features.
|
|
1483
|
-
|
|
1484
|
-
### 7.1 AgentBoot Is Not a Learning Management System
|
|
1485
|
-
|
|
1486
|
-
AgentBoot provides contextual assists and curated resource links, not courses, modules,
|
|
1487
|
-
progress tracking, or certificates. It does not track who completed what training. It
|
|
1488
|
-
does not gate persona access behind training completion. The `/learn` skill is a
|
|
1489
|
-
helpful assistant, not a course catalog.
|
|
1490
|
-
|
|
1491
|
-
The onboarding problem is real, but the solution is lightweight help at the moment of
|
|
1492
|
-
need, not a formal training program. Organizations that need formal AI training should
|
|
1493
|
-
use dedicated LMS platforms.
|
|
1494
|
-
|
|
1495
|
-
### 7.2 AgentBoot Is Not a Surveillance Tool
|
|
1496
|
-
|
|
1497
|
-
AgentBoot does not collect, transmit, or surface raw developer prompts. It does not
|
|
1498
|
-
rank developers by AI skill or prompt quality. It does not gamify usage with
|
|
1499
|
-
leaderboards, badges, or streaks. It does not correlate AI usage with performance
|
|
1500
|
-
reviews. It does not make AI usage mandatory.
|
|
1501
|
-
|
|
1502
|
-
The privacy model is a design invariant, not a configuration option. An organization
|
|
1503
|
-
that wants conversation monitoring has that capability through their API provider's
|
|
1504
|
-
Compliance API and enterprise data governance tools. AgentBoot is not that channel.
|
|
1505
|
-
|
|
1506
|
-
### 7.3 AgentBoot Does Not Compete with Content Libraries
|
|
1507
|
-
|
|
1508
|
-
AgentBoot is the governance and distribution layer, not a content library competing
|
|
1509
|
-
with projects like SuperClaude, ArcKit, spec-kit, or Trail of Bits configurations.
|
|
1510
|
-
These projects produce content (traits, agents, skills). AgentBoot distributes,
|
|
1511
|
-
governs, and orchestrates content.
|
|
1512
|
-
|
|
1513
|
-
The recommended relationship is marketplace curation (point to upstream, cross-list)
|
|
1514
|
-
rather than bundling. AgentBoot works without any third-party content; third-party
|
|
1515
|
-
content works without AgentBoot. They are complementary, not competitive.
|
|
1516
|
-
|
|
1517
|
-
### 7.4 AgentBoot Is Not a Runtime Framework
|
|
1518
|
-
|
|
1519
|
-
AgentBoot is a build tool. It produces artifacts and exits. It has zero runtime
|
|
1520
|
-
footprint. The distributed output works without AgentBoot installed. There is no
|
|
1521
|
-
"AgentBoot runtime" that must be running for personas to function.
|
|
1522
|
-
|
|
1523
|
-
The exception is the MCP server, which is a separate, optional package
|
|
1524
|
-
(`@agentboot/mcp-server`) that runs as a process. Even this is opt-in and not required
|
|
1525
|
-
for core functionality.
|
|
1526
|
-
|
|
1527
|
-
### 7.5 AgentBoot Does Not Promise Universal Platform Parity
|
|
1528
|
-
|
|
1529
|
-
AgentBoot is honest about platform capabilities. Claude Code gets the full governance
|
|
1530
|
-
surface (hooks, managed settings, subagent isolation, agent memory, plugin system).
|
|
1531
|
-
Non-CC platforms get the content (personas, instructions, rules) but not the
|
|
1532
|
-
enforcement mechanisms.
|
|
1533
|
-
|
|
1534
|
-
AgentBoot documents these gaps per platform rather than pretending all agents are equal
|
|
1535
|
-
or promising features it cannot deliver. The compliance story for non-CC platforms is:
|
|
1536
|
-
instruction-based refusal plus CI-based review plus organizational policy.
|
|
1537
|
-
|
|
1538
|
-
### 7.6 AgentBoot Does Not Manage AI Models
|
|
1539
|
-
|
|
1540
|
-
AgentBoot specifies which model a persona should use (in persona.config.json and agent
|
|
1541
|
-
frontmatter), but it does not manage model access, API keys, billing, or rate limiting.
|
|
1542
|
-
Those are the responsibility of the AI platform (Anthropic Console, GitHub Settings,
|
|
1543
|
-
Cursor settings).
|
|
1544
|
-
|
|
1545
|
-
### 7.7 AgentBoot Does Not Replace Code Review
|
|
1546
|
-
|
|
1547
|
-
AI code review is a complement to human code review, not a replacement. AgentBoot
|
|
1548
|
-
personas produce findings that humans evaluate and act on. Even at the highest autonomy
|
|
1549
|
-
level (Phase 3: Autonomous), the output is reviewed post-hoc by humans.
|
|
1550
|
-
|
|
1551
|
-
The test plan explicitly states: "Agents test agents, but humans always decide.
|
|
1552
|
-
Automation removes burden, not judgment."
|
|
1553
|
-
|
|
1554
|
-
### 7.8 AgentBoot Does Not Do Dynamic Agent Orchestration
|
|
1555
|
-
|
|
1556
|
-
AgentBoot does not provide a runtime agent orchestration system with message passing,
|
|
1557
|
-
state machines, or complex multi-agent workflows. The `/review` meta-skill is a
|
|
1558
|
-
lightweight routing layer that reads a configuration file and dispatches to the
|
|
1559
|
-
appropriate persona. The persona arbitrator (V2+) resolves conflicts between
|
|
1560
|
-
concurrent reviewers. Neither constitutes a full orchestration framework.
|
|
1561
|
-
|
|
1562
|
-
---
|
|
1563
|
-
|
|
1564
|
-
## 8. Success Metrics
|
|
1565
|
-
|
|
1566
|
-
### 8.1 Adoption Metrics
|
|
1567
|
-
|
|
1568
|
-
| Metric | Measurement | Target |
|
|
1569
|
-
|---|---|---|
|
|
1570
|
-
| Organizations using AgentBoot | GitHub template uses, marketplace installs | Growth rate |
|
|
1571
|
-
| Active developers per org | Persona invocation telemetry (anonymous) | 70%+ of licensed seats |
|
|
1572
|
-
| Persona invocations per day (org-wide) | Telemetry aggregate | Trending upward |
|
|
1573
|
-
| Time from setup to first persona invocation | Wizard completion to first `/review-code` | Under 10 minutes |
|
|
1574
|
-
| Repos governed per org | `agentboot status` repo count | Growth rate |
|
|
1575
|
-
| Skeptic conversion rate | Developers who start at zero usage and reach 1+ sessions/week | 30%+ within 90 days |
|
|
1576
|
-
|
|
1577
|
-
### 8.2 Contribution Metrics
|
|
1578
|
-
|
|
1579
|
-
| Metric | Measurement | Target |
|
|
1580
|
-
|---|---|---|
|
|
1581
|
-
| Marketplace items (Verified) | Marketplace repo count | Growth rate |
|
|
1582
|
-
| Community contributors | Unique PR authors to marketplace | Growth rate |
|
|
1583
|
-
| Contribution-to-install ratio | Items contributed vs. items installed | Healthy ratio (not vanity) |
|
|
1584
|
-
| Domain layers available | Complete domain packages in marketplace | 3+ by V2 |
|
|
1585
|
-
| Trait reuse rate | How many personas compose each trait (average) | 3+ personas per core trait |
|
|
1586
|
-
|
|
1587
|
-
### 8.3 Quality Metrics
|
|
1588
|
-
|
|
1589
|
-
| Metric | Measurement | Target |
|
|
1590
|
-
|---|---|---|
|
|
1591
|
-
| Persona false positive rate | Rephrase rates, finding dismissal rates | Under 20% |
|
|
1592
|
-
| Bug escape rate delta | Production bugs that a persona should have caught (before/after) | Measurable reduction |
|
|
1593
|
-
| Test coverage delta | Coverage change in repos with test-generator persona | Measurable increase |
|
|
1594
|
-
| PR review turnaround | Time from PR open to first AI review | Under 5 minutes |
|
|
1595
|
-
| Behavioral test pass rate | CI test results | 95%+ |
|
|
1596
|
-
| Lint pass rate on first commit | How often persona changes pass lint without fixes | 80%+ |
|
|
1597
|
-
|
|
1598
|
-
### 8.4 Cost Metrics
|
|
1599
|
-
|
|
1600
|
-
| Metric | Measurement | Target |
|
|
1601
|
-
|---|---|---|
|
|
1602
|
-
| Avg cost per persona invocation | Telemetry (tokens * model rate) | Trending downward |
|
|
1603
|
-
| Model mix efficiency | % of invocations on Haiku/Sonnet vs. Opus | 80%+ on Sonnet or cheaper |
|
|
1604
|
-
| Automated test cost per month | CI billing | Under $200 for 4 personas |
|
|
1605
|
-
| Token budget compliance | % of personas within token budget | 95%+ |
|
|
1606
|
-
| Cost per finding | Total cost / total findings | Trending downward |
|
|
1607
|
-
|
|
1608
|
-
### 8.5 Governance Metrics
|
|
1609
|
-
|
|
1610
|
-
| Metric | Measurement | Target |
|
|
1611
|
-
|---|---|---|
|
|
1612
|
-
| Sync coverage | % of repos receiving governance updates | 100% of registered repos |
|
|
1613
|
-
| Managed settings deployment | % of developer machines with HARD guardrails | 100% of managed machines |
|
|
1614
|
-
| Compliance hook activation | % of sessions where compliance hooks are active | 100% on managed machines |
|
|
1615
|
-
| ADR exception count | Approved exceptions via ADR governance | Low and stable |
|
|
1616
|
-
| Escalation flag rate | Harmful content flags per period | Low (zero is ideal) |
|
|
1617
|
-
|
|
1618
|
-
### 8.6 Developer Experience Metrics
|
|
1619
|
-
|
|
1620
|
-
| Metric | Measurement | Target |
|
|
1621
|
-
|---|---|---|
|
|
1622
|
-
| Setup completion rate | % of users who complete `agentboot setup` | 90%+ |
|
|
1623
|
-
| Uninstall rate | % of orgs that uninstall within 90 days | Under 10% |
|
|
1624
|
-
| `/learn` invocations | How often developers seek help | Trending downward (learning) |
|
|
1625
|
-
| Tip dismissal rate | How often contextual tips are disabled | Under 30% |
|
|
1626
|
-
| NPS (Net Promoter Score) | Developer survey | Positive |
|
|
1627
|
-
|
|
1628
|
-
---
|
|
1629
|
-
|
|
1630
|
-
## 9. Open Questions
|
|
1631
|
-
|
|
1632
|
-
Open questions discovered during PRD authoring have been resolved or deferred.
|
|
1633
|
-
See the internal tracking document for remaining items.
|
|
1634
|
-
|
|
1635
|
-
---
|
|
1636
|
-
|
|
1637
|
-
## 10. Glossary
|
|
1638
|
-
|
|
1639
|
-
### Agent
|
|
1640
|
-
|
|
1641
|
-
A Claude Code concept. An agent is a custom AI assistant defined in
|
|
1642
|
-
`.claude/agents/{name}/CLAUDE.md` with rich frontmatter (model, permissionMode,
|
|
1643
|
-
maxTurns, disallowedTools, mcpServers, hooks, memory, isolation). Agents are distinct
|
|
1644
|
-
from the base Claude Code assistant. In AgentBoot, personas are compiled into agents
|
|
1645
|
-
for CC-native output.
|
|
1646
|
-
|
|
1647
|
-
### ADR (Architecture Decision Record)
|
|
1648
|
-
|
|
1649
|
-
A formal exception governance mechanism. When a developer intentionally deviates from
|
|
1650
|
-
a persona's recommendation, an ADR documents the rationale, gets reviewer approval,
|
|
1651
|
-
and becomes a permanent record. The persona learns to accept the deviation for that
|
|
1652
|
-
specific case. ADRs handle permanent, approved deviations (as opposed to temporary
|
|
1653
|
-
elevation for debugging).
|
|
1654
|
-
|
|
1655
|
-
### Agent-Agnostic
|
|
1656
|
-
|
|
1657
|
-
Content that works across multiple AI agent platforms without modification. Traits
|
|
1658
|
-
(Markdown), personas (SKILL.md via agentskills.io), and gotchas (Markdown with glob
|
|
1659
|
-
patterns) are agent-agnostic. Hooks, managed settings, and agent frontmatter are
|
|
1660
|
-
agent-specific (Claude Code only).
|
|
1661
|
-
|
|
1662
|
-
### agentskills.io
|
|
1663
|
-
|
|
1664
|
-
An open standard for AI agent skill definitions. Uses SKILL.md format (Markdown with
|
|
1665
|
-
YAML frontmatter). Supported by 26+ agent platforms. AgentBoot uses agentskills.io as
|
|
1666
|
-
the cross-platform persona format.
|
|
1667
|
-
|
|
1668
|
-
### Always-On Instructions
|
|
1669
|
-
|
|
1670
|
-
Universal guardrails distributed to every repo regardless of persona configuration.
|
|
1671
|
-
These load at session start and remain active throughout. Used for org-wide rules that
|
|
1672
|
-
apply to every AI interaction (security baselines, coding standards, compliance
|
|
1673
|
-
requirements).
|
|
1674
|
-
|
|
1675
|
-
### Autonomy Progression
|
|
1676
|
-
|
|
1677
|
-
A three-phase model for persona independence. Phase 1 (Advisory): persona produces
|
|
1678
|
-
findings, human decides. Phase 2 (Auto-approve): persona applies low-risk fixes
|
|
1679
|
-
automatically, high-risk findings require human review. Phase 3 (Autonomous): persona
|
|
1680
|
-
operates independently, human reviews post-hoc. Promotion between phases is a
|
|
1681
|
-
governance decision requiring evidence (false-positive rates, test coverage, team
|
|
1682
|
-
approval).
|
|
1683
|
-
|
|
1684
|
-
### Build Pipeline
|
|
1685
|
-
|
|
1686
|
-
The three-stage process: validate (pre-build checks), compile (resolve traits, produce
|
|
1687
|
-
output), sync (distribute to repos). The pipeline is: validate, compile, sync.
|
|
1688
|
-
|
|
1689
|
-
### CC-First Delivery
|
|
1690
|
-
|
|
1691
|
-
The principle that Claude Code is the primary delivery target. Content is agent-agnostic
|
|
1692
|
-
(portable Markdown). Delivery leverages CC's full feature surface (plugins, hooks,
|
|
1693
|
-
managed settings, MCP, agents, skills, rules). Non-CC platforms get the content but
|
|
1694
|
-
not the enforcement mechanisms.
|
|
1695
|
-
|
|
1696
|
-
### Cowork
|
|
1697
|
-
|
|
1698
|
-
Anthropic's desktop application for non-technical users. Cowork plugins are the same
|
|
1699
|
-
format as Claude Code plugins but appear in a GUI with form-based input rather than
|
|
1700
|
-
slash commands. AgentBoot personas packaged as plugins work in both CC and Cowork.
|
|
1701
|
-
|
|
1702
|
-
### Compilation Target
|
|
1703
|
-
|
|
1704
|
-
One of two output formats produced by `agentboot build`. The cross-platform target
|
|
1705
|
-
produces standalone SKILL.md files with traits inlined. The CC-native target produces
|
|
1706
|
-
the full `.claude/` directory structure with @imports, rich frontmatter, hooks, and
|
|
1707
|
-
MCP configuration.
|
|
1708
|
-
|
|
1709
|
-
### Convention Over Configuration
|
|
1710
|
-
|
|
1711
|
-
The principle that AgentBoot ships with sensible defaults for everything. Organizations
|
|
1712
|
-
configure what is different about their situation, not everything from scratch. Borrowed
|
|
1713
|
-
from the Spring Boot design philosophy.
|
|
1714
|
-
|
|
1715
|
-
### Domain Layer
|
|
1716
|
-
|
|
1717
|
-
A complete package of traits, personas, gotchas, and instructions for a specific
|
|
1718
|
-
compliance regime or technology stack. Examples: healthcare-compliance (PHI, HIPAA,
|
|
1719
|
-
FHIR), fintech-compliance (PCI-DSS, SOX), govtech-fedramp. Domain layers are the
|
|
1720
|
-
highest-value, highest-effort marketplace contribution.
|
|
1721
|
-
|
|
1722
|
-
### Gotcha (Gotchas Rule)
|
|
1723
|
-
|
|
1724
|
-
A path-scoped instruction that encodes hard-won operational knowledge. Activated when
|
|
1725
|
-
a developer works on files matching the glob pattern. Invisible (zero context cost)
|
|
1726
|
-
when working on unrelated files. Technology-specific, not org-specific — making them
|
|
1727
|
-
highly shareable. Example: "PostgreSQL partitions do NOT inherit relrowsecurity."
|
|
1728
|
-
|
|
1729
|
-
### HARD Guardrail
|
|
1730
|
-
|
|
1731
|
-
A non-overridable compliance rule deployed via MDM (managed settings) or marked
|
|
1732
|
-
`required: true` in the org config. Cannot be elevated, overridden, or disabled at any
|
|
1733
|
-
scope level. A team-level config that attempts to disable a HARD guardrail causes a
|
|
1734
|
-
build failure. Used for rules where violation is a compliance incident (PHI scanning,
|
|
1735
|
-
credential blocking, audit logging).
|
|
1736
|
-
|
|
1737
|
-
### Hub-and-Spoke Distribution
|
|
1738
|
-
|
|
1739
|
-
The distribution model. One central repository (the hub, the personas repo) contains
|
|
1740
|
-
the source of truth. Target repositories (spokes, the actual codebases) receive
|
|
1741
|
-
compiled artifacts via the sync pipeline. One-way flow: hub publishes, spokes receive.
|
|
1742
|
-
Spokes do not produce governance.
|
|
1743
|
-
|
|
1744
|
-
### Managed Settings
|
|
1745
|
-
|
|
1746
|
-
Claude Code configuration files deployed to OS-level paths
|
|
1747
|
-
(`/Library/Application Support/ClaudeCode/` on macOS, `/etc/claude-code/` on Linux)
|
|
1748
|
-
via MDM. Cannot be overridden by any user or project setting. Used for HARD guardrails
|
|
1749
|
-
and forced plugin installation.
|
|
1750
|
-
|
|
1751
|
-
### Marketplace
|
|
1752
|
-
|
|
1753
|
-
The three-layer ecosystem for sharing governance content. Core (maintained by AgentBoot),
|
|
1754
|
-
Verified (community-contributed, reviewed), Community (unreviewed, buyer-beware).
|
|
1755
|
-
Physically, a marketplace is a Git repository with a marketplace.json catalog.
|
|
1756
|
-
|
|
1757
|
-
### MCP (Model Context Protocol)
|
|
1758
|
-
|
|
1759
|
-
A protocol for AI agents to interact with external tools and data sources. MCP servers
|
|
1760
|
-
expose tools and resources that agents can consume. Supported by Claude Code, Copilot,
|
|
1761
|
-
Cursor, and Gemini CLI. AgentBoot uses MCP for cross-platform persona serving and
|
|
1762
|
-
knowledge base access.
|
|
1763
|
-
|
|
1764
|
-
### Persona
|
|
1765
|
-
|
|
1766
|
-
A complete, deployable agent. A composition of traits plus a specialized system prompt
|
|
1767
|
-
that defines the agent's identity, operating context, and mandate. Personas compose
|
|
1768
|
-
traits; they do not inherit from each other. In AgentBoot, personas are defined as
|
|
1769
|
-
SKILL.md files and compiled into platform-specific output.
|
|
1770
|
-
|
|
1771
|
-
### persona.config.json
|
|
1772
|
-
|
|
1773
|
-
Build metadata for a persona. Specifies which traits to compose, at what weight, the
|
|
1774
|
-
target model, permission mode, tool restrictions, MCP servers, hooks, and autonomy
|
|
1775
|
-
level. Read by the compile step to generate output.
|
|
1776
|
-
|
|
1777
|
-
### Persona Arbitrator
|
|
1778
|
-
|
|
1779
|
-
A dedicated persona (V2+) that resolves conflicts when multiple reviewer personas
|
|
1780
|
-
produce contradictory findings on the same code. Only invoked when the `/review`
|
|
1781
|
-
meta-skill detects conflicting findings. Not invoked on every review.
|
|
1782
|
-
|
|
1783
|
-
### Plugin
|
|
1784
|
-
|
|
1785
|
-
A Claude Code distribution unit that bundles agents, skills, hooks, rules, MCP
|
|
1786
|
-
configuration, and settings into a single installable package. Distributed via
|
|
1787
|
-
marketplaces (public or private). The primary delivery method for CC users.
|
|
1788
|
-
|
|
1789
|
-
### Prompts as Code
|
|
1790
|
-
|
|
1791
|
-
The principle that AI agent behavior is treated as infrastructure: defined in files,
|
|
1792
|
-
stored in version control, reviewed in pull requests, tested, linted, and measured.
|
|
1793
|
-
Analogous to Infrastructure as Code (Terraform) and Configuration as Code (Kubernetes).
|
|
1794
|
-
|
|
1795
|
-
### Scope Hierarchy
|
|
1796
|
-
|
|
1797
|
-
The four-level organizational model: Org, Group, Team, Repo. More specific scopes
|
|
1798
|
-
layer on top of general ones. Optional behaviors follow "most specific wins." Mandatory
|
|
1799
|
-
behaviors follow "most general wins." Determines how configuration is merged during
|
|
1800
|
-
sync.
|
|
1801
|
-
|
|
1802
|
-
### Self-Improvement Reflections
|
|
1803
|
-
|
|
1804
|
-
An optional mechanism where personas write brief reflections after completing their
|
|
1805
|
-
task. Reflections accumulate into a dataset revealing patterns. Three phases:
|
|
1806
|
-
Phase A (humans edit personas from observation), Phase B (reflections + review skill),
|
|
1807
|
-
Phase C (automated accuracy tracking).
|
|
1808
|
-
|
|
1809
|
-
### SKILL.md
|
|
1810
|
-
|
|
1811
|
-
The agentskills.io format for persona definitions. Markdown with YAML frontmatter
|
|
1812
|
-
(name, description, traits, scope, output format) followed by the system prompt in
|
|
1813
|
-
prose. The cross-platform persona format.
|
|
1814
|
-
|
|
1815
|
-
### SME Discoverability Fragment
|
|
1816
|
-
|
|
1817
|
-
A lightweight always-on CLAUDE.md section (~100 tokens) auto-generated by the build
|
|
1818
|
-
system. Lists all available personas and how to invoke them. Makes personas
|
|
1819
|
-
discoverable without consulting external documentation.
|
|
1820
|
-
|
|
1821
|
-
### SOFT Guardrail
|
|
1822
|
-
|
|
1823
|
-
An important default that can be temporarily elevated. Deployed via the shared repo
|
|
1824
|
-
(not MDM). Elevation mechanism: developer invokes `/elevate` with a reason, receives
|
|
1825
|
-
time-bounded bypass (default 30 minutes), audit log entry created, guardrail
|
|
1826
|
-
automatically re-engages on TTL expiry.
|
|
1827
|
-
|
|
1828
|
-
### Structured Telemetry
|
|
1829
|
-
|
|
1830
|
-
Persona invocation metrics emitted as structured JSON (GELF/NDJSON format). Fields
|
|
1831
|
-
include persona_id, model, scope, input/output tokens, cost, findings, duration, and
|
|
1832
|
-
timestamp. No developer ID by default. No prompt text. Human-queryable with `jq`.
|
|
1833
|
-
|
|
1834
|
-
### Team Champion
|
|
1835
|
-
|
|
1836
|
-
A designated engineer on each team (typically tech lead or senior IC) who runs sync,
|
|
1837
|
-
reviews sync PRs, files quality feedback, onboards teammates, and proposes governance
|
|
1838
|
-
improvements. A rotating responsibility taking minutes per week in steady state.
|
|
1839
|
-
|
|
1840
|
-
### Trait
|
|
1841
|
-
|
|
1842
|
-
A reusable behavioral building block for an AI persona. Captures a single aspect of
|
|
1843
|
-
how an agent should think or communicate. Supports weight configurations
|
|
1844
|
-
(HIGH/MEDIUM/LOW or 0.0-1.0). Composed at build time. The DRY principle applied to
|
|
1845
|
-
AI behavior. Write once, compose everywhere, improve in one place.
|
|
1846
|
-
|
|
1847
|
-
### Trait Weight
|
|
1848
|
-
|
|
1849
|
-
A calibration system for traits that support variable intensity. Named weights
|
|
1850
|
-
(HIGH/MEDIUM/LOW) map to numeric values (0.7/0.5/0.3). Additional values: OFF (0.0)
|
|
1851
|
-
and MAX (1.0). The weight adjusts the threshold for action, not the type of action.
|
|
1852
|
-
At any weight, CRITICAL findings always surface.
|
|
1853
|
-
|
|
1854
|
-
### Two-Channel MDM Distribution
|
|
1855
|
-
|
|
1856
|
-
Enterprise distribution model separating non-negotiable enforcement (Channel 1: MDM-
|
|
1857
|
-
deployed managed settings for HARD guardrails) from team-customizable configuration
|
|
1858
|
-
(Channel 2: Git-based hub-and-spoke for SOFT guardrails, traits, personas, skills).
|
|
1859
|
-
|
|
1860
|
-
---
|
|
1861
|
-
|
|
1862
|
-
*End of Product Requirements Document.*
|