@fro.bot/systematic 2.3.3 → 2.4.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (71) hide show
  1. package/README.md +12 -13
  2. package/agents/design/design-implementation-reviewer.md +2 -19
  3. package/agents/design/design-iterator.md +2 -31
  4. package/agents/design/figma-design-sync.md +2 -22
  5. package/agents/docs/ankane-readme-writer.md +2 -19
  6. package/agents/document-review/adversarial-document-reviewer.md +3 -2
  7. package/agents/document-review/coherence-reviewer.md +5 -7
  8. package/agents/document-review/design-lens-reviewer.md +3 -4
  9. package/agents/document-review/feasibility-reviewer.md +3 -4
  10. package/agents/document-review/product-lens-reviewer.md +25 -6
  11. package/agents/document-review/scope-guardian-reviewer.md +3 -4
  12. package/agents/document-review/security-lens-reviewer.md +3 -4
  13. package/agents/research/best-practices-researcher.md +4 -21
  14. package/agents/research/framework-docs-researcher.md +2 -19
  15. package/agents/research/git-history-analyzer.md +2 -19
  16. package/agents/research/issue-intelligence-analyst.md +2 -24
  17. package/agents/research/learnings-researcher.md +7 -28
  18. package/agents/research/repo-research-analyst.md +3 -32
  19. package/agents/research/slack-researcher.md +128 -0
  20. package/agents/review/agent-native-reviewer.md +109 -195
  21. package/agents/review/architecture-strategist.md +3 -19
  22. package/agents/review/cli-agent-readiness-reviewer.md +1 -27
  23. package/agents/review/code-simplicity-reviewer.md +5 -19
  24. package/agents/review/data-integrity-guardian.md +3 -19
  25. package/agents/review/data-migration-expert.md +3 -19
  26. package/agents/review/deployment-verification-agent.md +3 -19
  27. package/agents/review/pattern-recognition-specialist.md +4 -20
  28. package/agents/review/performance-oracle.md +3 -31
  29. package/agents/review/project-standards-reviewer.md +5 -5
  30. package/agents/review/schema-drift-detector.md +3 -19
  31. package/agents/review/security-sentinel.md +3 -25
  32. package/agents/review/testing-reviewer.md +3 -3
  33. package/agents/workflow/pr-comment-resolver.md +54 -22
  34. package/agents/workflow/spec-flow-analyzer.md +2 -25
  35. package/package.json +1 -1
  36. package/skills/agent-native-architecture/SKILL.md +28 -27
  37. package/skills/agent-native-architecture/references/agent-execution-patterns.md +3 -3
  38. package/skills/agent-native-architecture/references/agent-native-testing.md +1 -1
  39. package/skills/agent-native-architecture/references/mobile-patterns.md +1 -1
  40. package/skills/andrew-kane-gem-writer/SKILL.md +5 -5
  41. package/skills/ce-brainstorm/SKILL.md +43 -181
  42. package/skills/ce-compound/SKILL.md +143 -89
  43. package/skills/ce-compound-refresh/SKILL.md +48 -5
  44. package/skills/ce-ideate/SKILL.md +27 -242
  45. package/skills/ce-plan/SKILL.md +165 -81
  46. package/skills/ce-review/SKILL.md +348 -125
  47. package/skills/ce-review/references/findings-schema.json +5 -0
  48. package/skills/ce-review/references/persona-catalog.md +2 -2
  49. package/skills/ce-review/references/resolve-base.sh +5 -2
  50. package/skills/ce-review/references/subagent-template.md +25 -3
  51. package/skills/ce-work/SKILL.md +95 -242
  52. package/skills/ce-work-beta/SKILL.md +154 -301
  53. package/skills/dhh-rails-style/SKILL.md +13 -12
  54. package/skills/document-review/SKILL.md +56 -109
  55. package/skills/document-review/references/findings-schema.json +0 -23
  56. package/skills/document-review/references/subagent-template.md +13 -18
  57. package/skills/dspy-ruby/SKILL.md +8 -8
  58. package/skills/every-style-editor/SKILL.md +3 -2
  59. package/skills/frontend-design/SKILL.md +2 -3
  60. package/skills/git-commit/SKILL.md +1 -1
  61. package/skills/git-commit-push-pr/SKILL.md +81 -265
  62. package/skills/git-worktree/SKILL.md +20 -21
  63. package/skills/lfg/SKILL.md +10 -17
  64. package/skills/onboarding/SKILL.md +2 -2
  65. package/skills/onboarding/scripts/inventory.mjs +31 -7
  66. package/skills/proof/SKILL.md +134 -28
  67. package/skills/resolve-pr-feedback/SKILL.md +7 -2
  68. package/skills/setup/SKILL.md +1 -1
  69. package/skills/test-browser/SKILL.md +10 -11
  70. package/skills/test-xcode/SKILL.md +6 -3
  71. package/dist/lib/manifest.d.ts +0 -39
@@ -1,17 +1,41 @@
1
1
  ---
2
2
  name: document-review
3
3
  description: Review requirements or plan documents using parallel persona agents that surface role-specific issues. Use when a requirements document or plan document exists and the user wants to improve it.
4
+ argument-hint: "[mode:headless] [path/to/document.md]"
4
5
  ---
5
6
 
6
7
  # Document Review
7
8
 
8
9
  Review requirements or plan documents through multi-persona analysis. Dispatches specialized reviewer agents in parallel, auto-fixes quality issues, and presents strategic questions for user decision.
9
10
 
11
+ ## Phase 0: Detect Mode
12
+
13
+ Check the skill arguments for `mode:headless`. Arguments may contain a document path, `mode:headless`, or both. Tokens starting with `mode:` are flags, not file paths -- strip them from the arguments and use the remaining token (if any) as the document path for Phase 1.
14
+
15
+ If `mode:headless` is present, set **headless mode** for the rest of the workflow.
16
+
17
+ **Headless mode** changes the interaction model, not the classification boundaries. Document-review still applies the same judgment about what has one clear correct fix vs. what needs user judgment. The only difference is how non-auto findings are delivered:
18
+ - `auto` fixes are applied silently (same as interactive)
19
+ - `present` findings are returned as structured text for the caller to handle -- no question prompts, no interactive approval
20
+ - Phase 5 returns immediately with "Review complete" (no refine/complete question)
21
+
22
+ The caller receives findings with their original classifications intact and decides what to do with them.
23
+
24
+ Callers invoke headless mode by including `mode:headless` in the skill arguments, e.g.:
25
+ ```
26
+ Skill("systematic:document-review", "mode:headless docs/plans/my-plan.md")
27
+ ```
28
+
29
+
30
+ If `mode:headless` is not present, the skill runs in its default interactive mode with no behavior change.
31
+
10
32
  ## Phase 1: Get and Analyze Document
11
33
 
12
34
  **If a document path is provided:** Read it, then proceed.
13
35
 
14
- **If no document is specified:** Ask which document to review, or find the most recent in `docs/brainstorms/` or `docs/plans/` using a file-search/glob tool (e.g., Glob in OpenCode).
36
+ **If no document is specified (interactive mode):** Ask which document to review, or find the most recent in `docs/brainstorms/` or `docs/plans/` using a file-search/glob tool (e.g., Glob in OpenCode).
37
+
38
+ **If no document is specified (headless mode):** Output "Review failed: headless mode requires a document path. Re-invoke with: Skill(\"systematic:document-review\", \"mode:headless <path>\")" without dispatching agents.
15
39
 
16
40
  ### Classify Document Type
17
41
 
@@ -23,11 +47,19 @@ After reading, classify the document:
23
47
 
24
48
  Analyze the document content to determine which conditional personas to activate. Check for these signals:
25
49
 
26
- **product-lens** -- activate when the document contains:
27
- - User-facing features, user stories, or customer-focused language
28
- - Market claims, competitive positioning, or business justification
29
- - Scope decisions, prioritization language, or priority tiers with feature assignments
30
- - Requirements with user/customer/business outcome focus
50
+ **product-lens** -- activate when the document makes challengeable claims about what to build and why, or when the proposed work carries strategic weight beyond the immediate problem. The system's users may be end users, developers, operators, maintainers, or any other audience -- the criteria are domain-agnostic. Check for either leg:
51
+
52
+ *Leg 1 — Premise claims:* The document stakes a position on what to build or why that a knowledgeable stakeholder could reasonably challenge -- not merely describing a task or restating known requirements:
53
+ - Problem framing where the stated need is non-obvious or debatable, not self-evident from existing context
54
+ - Solution selection where alternatives plausibly exist (implicit or explicit)
55
+ - Prioritization decisions that explicitly rank what gets built vs deferred
56
+ - Goal statements that predict specific user outcomes, not just restate constraints or describe deliverables
57
+
58
+ *Leg 2 — Strategic weight:* The proposed work could affect system trajectory, user perception, or competitive positioning, even if the premise is sound:
59
+ - Changes that shape how the system is perceived or what it becomes known for
60
+ - Complexity or simplicity bets that affect adoption, onboarding, or cognitive load
61
+ - Work that opens or closes future directions (path dependencies, architectural commitments)
62
+ - Opportunity cost implications -- building this means not building something else
31
63
 
32
64
  **design-lens** -- activate when the document contains:
33
65
  - UI/UX references, frontend components, or visual design language
@@ -48,6 +80,12 @@ Analyze the document content to determine which conditional personas to activate
48
80
  - Scope boundary language that seems misaligned with stated goals
49
81
  - Goals that don't clearly connect to requirements
50
82
 
83
+ **adversarial** -- activate when the document contains:
84
+ - More than 5 distinct requirements or implementation units
85
+ - Explicit architectural or scope decisions with stated rationale
86
+ - High-stakes domains (auth, payments, data migrations, external integrations)
87
+ - Proposals of new abstractions, frameworks, or significant architectural patterns
88
+
51
89
  ## Phase 2: Announce and Dispatch Personas
52
90
 
53
91
  ### Announce the Review Team
@@ -73,15 +111,16 @@ Add activated conditional personas:
73
111
  - `systematic:document-review:design-lens-reviewer`
74
112
  - `systematic:document-review:security-lens-reviewer`
75
113
  - `systematic:document-review:scope-guardian-reviewer`
114
+ - `systematic:document-review:adversarial-document-reviewer`
76
115
 
77
116
  ### Dispatch
78
117
 
79
- Dispatch all agents in **parallel** using the platform's task/agent tool (e.g., Agent tool in OpenCode, spawn in Codex). Each agent receives the prompt built from the [subagent template](./references/subagent-template.md) with these variables filled:
118
+ Dispatch all agents in **parallel** using the platform's task/agent tool (e.g., Agent tool in OpenCode, spawn in Codex). Omit the `mode` parameter so the user's configured permission settings apply. Each agent receives the prompt built from the subagent template included below with these variables filled:
80
119
 
81
120
  | Variable | Value |
82
121
  |----------|-------|
83
122
  | `{persona_file}` | Full content of the agent's markdown file |
84
- | `{schema}` | Content of [findings-schema.json](./references/findings-schema.json) |
123
+ | `{schema}` | Content of the findings schema included below |
85
124
  | `{document_type}` | "requirements" or "plan" from Phase 1 classification |
86
125
  | `{document_path}` | Path to the document |
87
126
  | `{document_content}` | Full text of the document |
@@ -90,112 +129,20 @@ Pass each agent the **full document** -- do not split into sections.
90
129
 
91
130
  **Error handling:** If an agent fails or times out, proceed with findings from agents that completed. Note the failed agent in the Coverage section. Do not block the entire review on a single agent failure.
92
131
 
93
- **Dispatch limit:** Even at maximum (6 agents), use parallel dispatch. These are document reviewers with bounded scope reading a single document -- parallel is safe and fast.
94
-
95
- ## Phase 3: Synthesize Findings
96
-
97
- Process findings from all agents through this pipeline. **Order matters** -- each step depends on the previous.
98
-
99
- ### 3.1 Validate
100
-
101
- Check each agent's returned JSON against [findings-schema.json](./references/findings-schema.json):
102
- - Drop findings missing any required field defined in the schema
103
- - Drop findings with invalid enum values
104
- - Note the agent name for any malformed output in the Coverage section
105
-
106
- ### 3.2 Confidence Gate
107
-
108
- Suppress findings below 0.50 confidence. Store them as residual concerns for potential promotion in step 3.4.
109
-
110
- ### 3.3 Deduplicate
111
-
112
- Fingerprint each finding using `normalize(section) + normalize(title)`. Normalization: lowercase, strip punctuation, collapse whitespace.
113
-
114
- When fingerprints match across personas:
115
- - If the findings recommend **opposing actions** (e.g., one says cut, the other says keep), do not merge -- preserve both for contradiction resolution in 3.5
116
- - Otherwise merge: keep the highest severity, keep the highest confidence, union all evidence arrays, note all agreeing reviewers (e.g., "coherence, feasibility")
132
+ **Dispatch limit:** Even at maximum (7 agents), use parallel dispatch. These are document reviewers with bounded scope reading a single document -- parallel is safe and fast.
117
133
 
118
- ### 3.4 Promote Residual Concerns
134
+ ## Phases 3-5: Synthesis, Presentation, and Next Action
119
135
 
120
- Scan the residual concerns (findings suppressed in 3.2) for:
121
- - **Cross-persona corroboration**: A residual concern from Persona A overlaps with an above-threshold finding from Persona B. Promote at P2 with confidence 0.55-0.65.
122
- - **Concrete blocking risks**: A residual concern describes a specific, concrete risk that would block implementation. Promote at P2 with confidence 0.55.
136
+ After all dispatched agents return, read `references/synthesis-and-presentation.md` for the synthesis pipeline (validate, gate, dedup, promote, resolve contradictions, route by autofix class), auto-fix application, finding presentation, and next-action menu. Do not load this file before agent dispatch completes.
123
137
 
124
- ### 3.5 Resolve Contradictions
125
-
126
- When personas disagree on the same section:
127
- - Create a **combined finding** presenting both perspectives
128
- - Set `autofix_class: present`
129
- - Frame as a tradeoff, not a verdict
130
-
131
- Specific conflict patterns:
132
- - Coherence says "keep for consistency" + scope-guardian says "cut for simplicity" -> combined finding, let user decide
133
- - Feasibility says "this is impossible" + product-lens says "this is essential" -> P1 finding framed as a tradeoff
134
- - Multiple personas flag the same issue -> merge into single finding, note consensus, increase confidence
135
-
136
- ### 3.6 Route by Autofix Class
137
-
138
- | Autofix Class | Route |
139
- |---------------|-------|
140
- | `auto` | Apply automatically -- local deterministic fix (terminology, formatting, cross-references) |
141
- | `present` | Present to user for judgment |
142
-
143
- Demote any `auto` finding that lacks a `suggested_fix` to `present` -- the orchestrator cannot apply a fix without concrete replacement text.
144
-
145
- ### 3.7 Sort
146
-
147
- Sort findings for presentation: P0 -> P1 -> P2 -> P3, then by confidence (descending), then by document order (section position).
148
-
149
- ## Phase 4: Apply and Present
150
-
151
- ### Apply Auto-fixes
152
-
153
- Apply all `auto` findings to the document in a **single pass**:
154
- - Edit the document inline using the platform's edit tool
155
- - Track what was changed for the "Auto-fixes Applied" section
156
- - Do not ask for approval -- these are unambiguously correct (terminology fixes, formatting, cross-references)
157
-
158
- ### Present Remaining Findings
159
-
160
- Present all other findings to the user using the format from [review-output-template.md](./references/review-output-template.md):
161
- - Group by severity (P0 -> P3)
162
- - Include the Coverage table showing which personas ran
163
- - Show auto-fixes that were applied
164
- - Include residual concerns and deferred questions if any
165
-
166
- Brief summary at the top: "Applied N auto-fixes. M findings to consider (X at P0/P1)."
167
-
168
- ### Protected Artifacts
169
-
170
- During synthesis, discard any finding that recommends deleting or removing files in:
171
- - `docs/brainstorms/`
172
- - `docs/plans/`
173
- - `docs/solutions/`
174
-
175
- These are pipeline artifacts and must not be flagged for removal.
176
-
177
- ## Phase 5: Next Action
178
-
179
- Use the platform's blocking question tool when available (question in OpenCode, request_user_input in Codex, ask_user in Gemini). Otherwise present numbered options and wait for the user's reply.
180
-
181
- Offer:
182
-
183
- 1. **Refine again** -- another review pass
184
- 2. **Review complete** -- document is ready
185
-
186
- After 2 refinement passes, recommend completion -- diminishing returns are likely. But if the user wants to continue, allow it.
187
-
188
- Return "Review complete" as the terminal signal for callers.
138
+ ---
189
139
 
190
- ## What NOT to Do
140
+ ## Included References
191
141
 
192
- - Do not rewrite the entire document
193
- - Do not add new sections or requirements the user didn't discuss
194
- - Do not over-engineer or add complexity
195
- - Do not create separate review files or add metadata sections
196
- - Do not modify any of the 4 caller skills (ce-brainstorm, ce-plan, ce-plan-beta, deepen-plan-beta)
142
+ ### Subagent Template
197
143
 
198
- ## Iteration Guidance
144
+ @./references/subagent-template.md
199
145
 
200
- On subsequent passes, re-dispatch personas and re-synthesize. The auto-fix mechanism and confidence gating prevent the same findings from recurring once fixed. If findings are repetitive across passes, recommend completion.
146
+ ### Findings Schema
201
147
 
148
+ @./references/findings-schema.json
@@ -82,28 +82,5 @@
82
82
  "description": "Questions that should be resolved in a later workflow stage (planning, implementation)",
83
83
  "items": { "type": "string" }
84
84
  }
85
- },
86
-
87
- "_meta": {
88
- "confidence_thresholds": {
89
- "suppress": "Below 0.50 -- do not report. Finding is speculative noise.",
90
- "flag": "0.50-0.69 -- include only when the persona's calibration says the issue is actionable at that confidence.",
91
- "report": "0.70+ -- report with full confidence."
92
- },
93
- "severity_definitions": {
94
- "P0": "Contradictions or gaps that would cause building the wrong thing. Must fix before proceeding.",
95
- "P1": "Significant gap likely hit during planning or implementation. Should fix.",
96
- "P2": "Moderate issue with meaningful downside. Fix if straightforward.",
97
- "P3": "Minor improvement. User's discretion."
98
- },
99
- "autofix_classes": {
100
- "_principle": "Autofix class is independent of severity. A P1 finding can be auto if the fix is obvious. The test: is there one clear correct fix, or does resolving this require judgment?",
101
- "auto": "One clear correct fix -- applied silently. Includes both internal reconciliation (summary/detail mismatches, wrong counts, stale cross-references, terminology drift) and additions mechanically implied by other content (missing steps, unstated thresholds, completeness gaps where the correct content is obvious). Must include suggested_fix.",
102
- "present": "Requires individual user judgment -- strategic questions, design tradeoffs, or findings where reasonable people could disagree on the right action."
103
- },
104
- "finding_types": {
105
- "error": "Something the document says that is wrong -- contradictions, incorrect statements, design tensions, incoherent tradeoffs. These are mistakes in what exists.",
106
- "omission": "Something the document forgot to say -- missing mechanical steps, absent list entries, undefined thresholds, forgotten cross-references. These are gaps in completeness."
107
- }
108
85
  }
109
86
  }
@@ -19,20 +19,25 @@ Return ONLY valid JSON matching the findings schema below. No prose, no markdown
19
19
  {schema}
20
20
 
21
21
  Rules:
22
+ - You are a leaf reviewer inside an already-running systematic review workflow. Do not invoke systematic skills or agents unless this template explicitly instructs you to. Perform your analysis directly and return findings in the required output format only.
22
23
  - Suppress any finding below your stated confidence floor (see your Confidence calibration section).
23
24
  - Every finding MUST include at least one evidence item -- a direct quote from the document.
24
25
  - You are operationally read-only. Analyze the document and produce findings. Do not edit the document, create files, or make changes. You may use non-mutating tools (file reads, glob, grep, git log) to gather context about the codebase when evaluating feasibility or existing patterns.
25
26
  - Set `finding_type` for every finding:
26
27
  - `error`: Something the document says that is wrong -- contradictions, incorrect statements, design tensions, incoherent tradeoffs.
27
28
  - `omission`: Something the document forgot to say -- missing mechanical steps, absent list entries, undefined thresholds, forgotten cross-references.
28
- - Set `autofix_class` based on whether there is one clear correct fix, not on severity. A P1 finding can be `auto` if the fix is obvious:
29
- - `auto`: One clear correct fix. Applied silently without asking. The test: is there only one reasonable way to resolve this? If yes, it is auto. Two categories:
30
- - Internal reconciliation: one part of the document is authoritative over another -- reconcile toward the authority. Examples: summary/detail mismatches, wrong counts, missing list entries derivable from elsewhere, stale cross-references, terminology drift, prose/diagram contradictions where prose is authoritative.
31
- - Implied additions: the correct content is mechanically obvious from the document's own context. Examples: adding a missing implementation step implied by other content, defining a threshold implied but never stated, completeness gaps where what to add is clear.
32
- Always include `suggested_fix` for auto findings.
33
- NOT auto (the gap is clear but more than one reasonable fix exists): choosing an implementation approach when the document states a need without constraining how (e.g., "support offline mode" could mean service workers, local-first database, or queue-and-sync -- there is no single obvious answer), changing scope or priority where the author may have weighed tradeoffs the reviewer can't see (e.g., promoting a P2 to P1, or cutting a feature the document intentionally keeps at a lower tier).
34
- - `present`: Requires judgment -- strategic questions, tradeoffs, design tensions where reasonable people could disagree, findings where the right action is unclear.
35
- - `suggested_fix` is required for `auto` findings. For `present` findings, `suggested_fix` is optional -- include it only when the fix is obvious, and frame as a question when the right action is unclear.
29
+ - Set `autofix_class` based on whether there is one clear correct fix, not on severity or importance:
30
+ - `auto`: One clear correct fix, applied silently. This includes trivial fixes AND substantive ones:
31
+ - Internal reconciliation -- one document part authoritative over another (summary/detail mismatches, wrong counts, stale cross-references, terminology drift)
32
+ - Implied additions -- correct content mechanically obvious from the document (missing steps, unstated thresholds, completeness gaps)
33
+ - Codebase-pattern-resolved -- an established codebase pattern resolves ambiguity (cite the specific file/function in `why_it_matters`)
34
+ - Incorrect behavior -- the document describes behavior that is factually wrong, and the correct behavior is obvious from context or the codebase
35
+ - Missing standard security measures -- HTTPS enforcement, checksum verification, input sanitization, private IP rejection, or other controls with known implementations where omission is clearly a bug
36
+ - Incomplete technical descriptions -- the accurate/complete version is directly derivable from the codebase
37
+ - Missing requirements that follow mechanically from the document's own explicit, concrete decisions (not high-level goals -- a goal can be satisfied by multiple valid requirements)
38
+ The test is not "is this fix important?" but "is there more than one reasonable way to fix this?" If a competent implementer would arrive at the same fix independently, it is auto -- even if the fix is substantive. Always include `suggested_fix`. NOT auto if more than one reasonable fix exists or if scope/priority judgment is involved.
39
+ - `present`: Requires user judgment -- genuinely multiple valid approaches where the right choice depends on priorities, tradeoffs, or context the reviewer does not have. Examples: architectural choices with real tradeoffs, scope decisions, feature prioritization, UX design choices.
40
+ - `suggested_fix` is required for `auto` findings. For `present` findings, include only when the fix is obvious.
36
41
  - If you find no issues, return an empty findings array. Still populate residual_risks and deferred_questions if applicable.
37
42
  - Use your suppress conditions. Do not flag issues that belong to other personas.
38
43
  </output-contract>
@@ -45,13 +50,3 @@ Document content:
45
50
  {document_content}
46
51
  </review-context>
47
52
  ```
48
-
49
- ## Variable Reference
50
-
51
- | Variable | Source | Description |
52
- |----------|--------|-------------|
53
- | `{persona_file}` | Agent markdown file content | The full persona definition (identity, analysis protocol, calibration, suppress conditions) |
54
- | `{schema}` | `references/findings-schema.json` content | The JSON schema reviewers must conform to |
55
- | `{document_type}` | Orchestrator classification | Either "requirements" or "plan" |
56
- | `{document_path}` | Skill input | Path to the document being reviewed |
57
- | `{document_content}` | File read | The full document text |
@@ -647,14 +647,14 @@ end
647
647
 
648
648
  ## Resources
649
649
 
650
- - [core-concepts.md](./references/core-concepts.md) — Signatures, modules, predictors, type system deep-dive
651
- - [toolsets.md](./references/toolsets.md) — Tools::Base, Tools::Toolset DSL, type safety, testing
652
- - [providers.md](./references/providers.md) — Provider adapters, RubyLLM, fiber-local LM context, compatibility matrix
653
- - [optimization.md](./references/optimization.md) — MIPROv2, GEPA, evaluation framework, storage system
654
- - [observability.md](./references/observability.md) — Event system, dspy-o11y gems, Langfuse, score reporting
655
- - [signature-template.rb](./assets/signature-template.rb) — Signature scaffold with T::Enum, Date/Time, defaults, union types
656
- - [module-template.rb](./assets/module-template.rb) — Module scaffold with .call(), lifecycle callbacks, fiber-local LM
657
- - [config-template.rb](./assets/config-template.rb) — Rails initializer with RubyLLM, observability, feature flags
650
+ - `references/core-concepts.md` — Signatures, modules, predictors, type system deep-dive
651
+ - `references/toolsets.md` — Tools::Base, Tools::Toolset DSL, type safety, testing
652
+ - `references/providers.md` — Provider adapters, RubyLLM, fiber-local LM context, compatibility matrix
653
+ - `references/optimization.md` — MIPROv2, GEPA, evaluation framework, storage system
654
+ - `references/observability.md` — Event system, dspy-o11y gems, Langfuse, score reporting
655
+ - `assets/signature-template.rb` — Signature scaffold with T::Enum, Date/Time, defaults, union types
656
+ - `assets/module-template.rb` — Module scaffold with .call(), lifecycle callbacks, fiber-local LM
657
+ - `assets/config-template.rb` — Rails initializer with RubyLLM, observability, feature flags
658
658
 
659
659
  ## Key URLs
660
660
 
@@ -44,7 +44,7 @@ Review each paragraph systematically, checking for:
44
44
  - Word choice and usage (overused words, passive voice)
45
45
  - Adherence to Every style guide rules
46
46
 
47
- Reference the complete [EVERY_WRITE_STYLE.md](./references/EVERY_WRITE_STYLE.md) for specific rules when in doubt.
47
+ Reference the complete style guide at `references/EVERY_WRITE_STYLE.md` for specific rules when in doubt.
48
48
 
49
49
  ### Step 3: Mechanical Review
50
50
 
@@ -99,7 +99,7 @@ FINAL RECOMMENDATIONS
99
99
 
100
100
  ## Style Guide Reference
101
101
 
102
- The complete Every style guide is included in [EVERY_WRITE_STYLE.md](./references/EVERY_WRITE_STYLE.md). Key areas to focus on:
102
+ The complete Every style guide is at `references/EVERY_WRITE_STYLE.md`. Key areas to focus on:
103
103
 
104
104
  - **Quick Rules**: Title case for headlines, sentence case elsewhere
105
105
  - **Tone**: Active voice, avoid overused words (actually, very, just), be specific
@@ -132,3 +132,4 @@ Based on Every's style guide, pay special attention to:
132
132
  - Word usage (fewer vs. less, they vs. them)
133
133
  - Company references (singular "it", teams as plural "they")
134
134
  - Job title capitalization
135
+
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: frontend-design
3
- description: Build web interfaces with genuine design quality, not AI slop. Use for any frontend work - landing pages, web apps, dashboards, admin panels, components, interactive experiences. Activates for both greenfield builds and modifications to existing applications. Detects existing design systems and respects them. Covers composition, typography, color, motion, and copy. Verifies results via screenshots before declaring done.
3
+ description: 'Build web interfaces with genuine design quality, not AI slop. Use for any frontend work - landing pages, web apps, dashboards, admin panels, components, interactive experiences. Activates for both greenfield builds and modifications to existing applications. Detects existing design systems and respects them. Covers composition, typography, color, motion, and copy. Verifies results via screenshots before declaring done.'
4
4
  ---
5
5
 
6
6
  # Frontend Design
@@ -230,7 +230,7 @@ Use the first available option:
230
230
 
231
231
  1. **Existing project browser tooling** -- if Playwright, Puppeteer, Cypress, or similar is already in the project's dependencies, use it. Do not introduce new dependencies just for verification.
232
232
  2. **Browser MCP tools** -- if browser automation tools (e.g., claude-in-chrome) are available in the agent's environment, use them.
233
- 3. **agent-browser CLI** -- if nothing else is available, this is the default. Load the `agent-browser` skill for installation and usage instructions.
233
+ 3. **agent-browser CLI** -- if nothing else is available and `agent-browser` is installed, use it. If not installed, inform the user: "`agent-browser` is not installed. Run `/ce-setup` to install required dependencies." Then skip to the next option.
234
234
  4. **Mental review** -- if no browser access is possible (headless CI, no permissions to install), apply the litmus checks as a self-review and note that visual verification was skipped.
235
235
 
236
236
  ### What to Assess
@@ -256,4 +256,3 @@ For greenfield work, commit to a bold aesthetic direction. Consider the tone: br
256
256
  Ask: what makes this unforgettable? What is the one thing someone will remember?
257
257
 
258
258
  Match implementation complexity to the aesthetic vision. Maximalist designs need elaborate code with extensive animations and effects. Minimalist designs need restraint, precision, and careful attention to spacing, typography, and subtle details. Elegance comes from executing the vision well, not from intensity.
259
-
@@ -65,7 +65,7 @@ If the current branch from the context above is empty, the repository is in deta
65
65
 
66
66
  Follow this priority order:
67
67
 
68
- 1. **Repo conventions already in context** -- If project instructions (AGENTS.md, AGENTS.md, or similar) are already loaded and specify commit message conventions, follow those. Do not re-read these files; they are loaded at session start.
68
+ 1. **Repo conventions already in context** -- If project instructions (AGENTS.md, or similar) are already loaded and specify commit message conventions, follow those. Do not re-read these files; they are loaded at session start.
69
69
  2. **Recent commit history** -- If no explicit convention is documented, examine the 10 most recent commits from Step 1. If a clear pattern emerges (e.g., conventional commits, ticket prefixes, emoji prefixes), match that pattern.
70
70
  3. **Default: conventional commits** -- If neither source provides a pattern, use conventional commit format: `type(scope): description` where type is one of `feat`, `fix`, `docs`, `refactor`, `test`, `chore`, `perf`, `ci`, `style`, `build`.
71
71