@fredcallagan/arn-spark 5.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude-plugin/plugin.json +9 -0
- package/.opencode/plugins/arn-spark.js +272 -0
- package/package.json +17 -0
- package/plugins/arn-spark/.claude-plugin/plugin.json +9 -0
- package/plugins/arn-spark/LICENSE +21 -0
- package/plugins/arn-spark/README.md +25 -0
- package/plugins/arn-spark/agents/arn-spark-brand-strategist.md +299 -0
- package/plugins/arn-spark/agents/arn-spark-dev-env-builder.md +228 -0
- package/plugins/arn-spark/agents/arn-spark-doctor.md +92 -0
- package/plugins/arn-spark/agents/arn-spark-forensic-investigator.md +181 -0
- package/plugins/arn-spark/agents/arn-spark-market-researcher.md +232 -0
- package/plugins/arn-spark/agents/arn-spark-marketing-pm.md +225 -0
- package/plugins/arn-spark/agents/arn-spark-persona-architect.md +259 -0
- package/plugins/arn-spark/agents/arn-spark-persona-impersonator.md +183 -0
- package/plugins/arn-spark/agents/arn-spark-product-strategist.md +191 -0
- package/plugins/arn-spark/agents/arn-spark-prototype-builder.md +497 -0
- package/plugins/arn-spark/agents/arn-spark-scaffolder.md +228 -0
- package/plugins/arn-spark/agents/arn-spark-spike-runner.md +209 -0
- package/plugins/arn-spark/agents/arn-spark-style-capture.md +196 -0
- package/plugins/arn-spark/agents/arn-spark-tech-evaluator.md +229 -0
- package/plugins/arn-spark/agents/arn-spark-ui-interactor.md +235 -0
- package/plugins/arn-spark/agents/arn-spark-use-case-writer.md +280 -0
- package/plugins/arn-spark/agents/arn-spark-ux-judge.md +215 -0
- package/plugins/arn-spark/agents/arn-spark-ux-specialist.md +200 -0
- package/plugins/arn-spark/agents/arn-spark-visual-sketcher.md +285 -0
- package/plugins/arn-spark/agents/arn-spark-visual-test-engineer.md +224 -0
- package/plugins/arn-spark/references/copilot-tools.md +62 -0
- package/plugins/arn-spark/skills/arn-brainstorming/SKILL.md +520 -0
- package/plugins/arn-spark/skills/arn-brainstorming/references/add-feature-flow.md +155 -0
- package/plugins/arn-spark/skills/arn-spark-arch-vision/SKILL.md +226 -0
- package/plugins/arn-spark/skills/arn-spark-arch-vision/references/architecture-vision-template.md +153 -0
- package/plugins/arn-spark/skills/arn-spark-arch-vision/references/technology-evaluation-guide.md +86 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype/SKILL.md +471 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype/references/clickable-prototype-criteria.md +65 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype/references/journey-template.md +62 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype/references/review-report-template.md +75 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype/references/showcase-capture-guide.md +213 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype-teams/SKILL.md +642 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype-teams/references/debate-protocol.md +242 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype-teams/references/debate-review-report-template.md +161 -0
- package/plugins/arn-spark/skills/arn-spark-clickable-prototype-teams/references/expert-interaction-review-template.md +152 -0
- package/plugins/arn-spark/skills/arn-spark-concept-review/SKILL.md +350 -0
- package/plugins/arn-spark/skills/arn-spark-concept-review/references/conflict-resolution-protocol.md +145 -0
- package/plugins/arn-spark/skills/arn-spark-concept-review/references/review-report-template.md +185 -0
- package/plugins/arn-spark/skills/arn-spark-dev-setup/SKILL.md +366 -0
- package/plugins/arn-spark/skills/arn-spark-dev-setup/references/dev-setup-checklist.md +84 -0
- package/plugins/arn-spark/skills/arn-spark-dev-setup/references/dev-setup-template.md +205 -0
- package/plugins/arn-spark/skills/arn-spark-discover/SKILL.md +303 -0
- package/plugins/arn-spark/skills/arn-spark-discover/references/competitive-landscape-template.md +87 -0
- package/plugins/arn-spark/skills/arn-spark-discover/references/discovery-questions.md +120 -0
- package/plugins/arn-spark/skills/arn-spark-discover/references/persona-profile-template.md +97 -0
- package/plugins/arn-spark/skills/arn-spark-discover/references/product-concept-template.md +253 -0
- package/plugins/arn-spark/skills/arn-spark-ensure-config/SKILL.md +23 -0
- package/plugins/arn-spark/skills/arn-spark-ensure-config/references/ensure-config.md +388 -0
- package/plugins/arn-spark/skills/arn-spark-ensure-config/references/step-0-fast-path.md +25 -0
- package/plugins/arn-spark/skills/arn-spark-ensure-config/scripts/cache-check.sh +127 -0
- package/plugins/arn-spark/skills/arn-spark-feature-extract/SKILL.md +483 -0
- package/plugins/arn-spark/skills/arn-spark-feature-extract/references/feature-backlog-template.md +176 -0
- package/plugins/arn-spark/skills/arn-spark-feature-extract/references/feature-entry-template.md +209 -0
- package/plugins/arn-spark/skills/arn-spark-help/SKILL.md +149 -0
- package/plugins/arn-spark/skills/arn-spark-help/references/pipeline-map.md +211 -0
- package/plugins/arn-spark/skills/arn-spark-init/SKILL.md +312 -0
- package/plugins/arn-spark/skills/arn-spark-init/references/agent-models-presets/all-opus.md +23 -0
- package/plugins/arn-spark/skills/arn-spark-init/references/agent-models-presets/balanced.md +23 -0
- package/plugins/arn-spark/skills/arn-spark-init/references/bkt-setup.md +55 -0
- package/plugins/arn-spark/skills/arn-spark-init/references/jira-mcp-setup.md +61 -0
- package/plugins/arn-spark/skills/arn-spark-init/references/platform-labels.md +97 -0
- package/plugins/arn-spark/skills/arn-spark-naming/SKILL.md +275 -0
- package/plugins/arn-spark/skills/arn-spark-naming/references/creative-brief-template.md +146 -0
- package/plugins/arn-spark/skills/arn-spark-naming/references/naming-methodology.md +237 -0
- package/plugins/arn-spark/skills/arn-spark-naming/references/naming-report-template.md +122 -0
- package/plugins/arn-spark/skills/arn-spark-naming/references/trademark-databases.md +88 -0
- package/plugins/arn-spark/skills/arn-spark-naming/references/whois-server-map.md +164 -0
- package/plugins/arn-spark/skills/arn-spark-naming/scripts/whois-check.js +502 -0
- package/plugins/arn-spark/skills/arn-spark-naming/scripts/whois-check.py +533 -0
- package/plugins/arn-spark/skills/arn-spark-prototype-lock/SKILL.md +260 -0
- package/plugins/arn-spark/skills/arn-spark-prototype-lock/references/lock-report-template.md +68 -0
- package/plugins/arn-spark/skills/arn-spark-prototype-lock/references/pretooluse-hook-template.json +35 -0
- package/plugins/arn-spark/skills/arn-spark-prototype-lock/references/prototype-guardrail-rules.md +38 -0
- package/plugins/arn-spark/skills/arn-spark-report/SKILL.md +144 -0
- package/plugins/arn-spark/skills/arn-spark-report/references/issue-template.md +81 -0
- package/plugins/arn-spark/skills/arn-spark-report/references/spark-knowledge-base.md +293 -0
- package/plugins/arn-spark/skills/arn-spark-scaffold/SKILL.md +239 -0
- package/plugins/arn-spark/skills/arn-spark-scaffold/references/scaffold-checklist.md +79 -0
- package/plugins/arn-spark/skills/arn-spark-scaffold/references/scaffold-summary-template.md +74 -0
- package/plugins/arn-spark/skills/arn-spark-spike/SKILL.md +209 -0
- package/plugins/arn-spark/skills/arn-spark-spike/references/spike-report-template.md +123 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype/SKILL.md +362 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype/references/review-report-template.md +65 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype/references/showcase-capture-guide.md +153 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype/references/static-prototype-criteria.md +54 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype-teams/SKILL.md +518 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype-teams/references/debate-protocol.md +230 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype-teams/references/debate-review-report-template.md +148 -0
- package/plugins/arn-spark/skills/arn-spark-static-prototype-teams/references/expert-visual-review-template.md +130 -0
- package/plugins/arn-spark/skills/arn-spark-stress-competitive/SKILL.md +166 -0
- package/plugins/arn-spark/skills/arn-spark-stress-competitive/references/competitive-report-template.md +139 -0
- package/plugins/arn-spark/skills/arn-spark-stress-competitive/references/gap-analysis-framework.md +111 -0
- package/plugins/arn-spark/skills/arn-spark-stress-interview/SKILL.md +257 -0
- package/plugins/arn-spark/skills/arn-spark-stress-interview/references/interview-protocol.md +140 -0
- package/plugins/arn-spark/skills/arn-spark-stress-interview/references/interview-report-template.md +165 -0
- package/plugins/arn-spark/skills/arn-spark-stress-interview/references/persona-casting-spec.md +138 -0
- package/plugins/arn-spark/skills/arn-spark-stress-premortem/SKILL.md +181 -0
- package/plugins/arn-spark/skills/arn-spark-stress-premortem/references/premortem-protocol.md +112 -0
- package/plugins/arn-spark/skills/arn-spark-stress-premortem/references/premortem-report-template.md +158 -0
- package/plugins/arn-spark/skills/arn-spark-stress-prfaq/SKILL.md +206 -0
- package/plugins/arn-spark/skills/arn-spark-stress-prfaq/references/prfaq-report-template.md +139 -0
- package/plugins/arn-spark/skills/arn-spark-stress-prfaq/references/prfaq-workflow.md +118 -0
- package/plugins/arn-spark/skills/arn-spark-style-explore/SKILL.md +281 -0
- package/plugins/arn-spark/skills/arn-spark-style-explore/references/style-brief-template.md +198 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases/SKILL.md +359 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases/references/expert-review-template.md +94 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases/references/review-protocol.md +150 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases/references/use-case-index-template.md +108 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases/references/use-case-template.md +125 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases-teams/SKILL.md +306 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases-teams/references/debate-protocol.md +272 -0
- package/plugins/arn-spark/skills/arn-spark-use-cases-teams/references/review-report-template.md +112 -0
- package/plugins/arn-spark/skills/arn-spark-visual-readiness/SKILL.md +293 -0
- package/plugins/arn-spark/skills/arn-spark-visual-readiness/references/readiness-checklist.md +196 -0
- package/plugins/arn-spark/skills/arn-spark-visual-sketch/SKILL.md +376 -0
- package/plugins/arn-spark/skills/arn-spark-visual-sketch/references/aesthetic-philosophy.md +210 -0
- package/plugins/arn-spark/skills/arn-spark-visual-sketch/references/sketch-gallery-guide.md +282 -0
- package/plugins/arn-spark/skills/arn-spark-visual-sketch/references/visual-direction-template.md +174 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/SKILL.md +447 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/references/baseline-capture-script-template.js +89 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/references/journey-schema.md +375 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/references/spike-checklist.md +122 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/references/strategy-layers-guide.md +132 -0
- package/plugins/arn-spark/skills/arn-spark-visual-strategy/references/visual-strategy-template.md +141 -0
package/plugins/arn-spark/skills/arn-spark-clickable-prototype-teams/references/debate-protocol.md
ADDED
|
@@ -0,0 +1,242 @@
|
|
|
1
|
+
# Interaction Review Debate Protocol
|
|
2
|
+
|
|
3
|
+
This document defines the structured debate process for team-based interaction review in `arn-spark-clickable-prototype-teams`. Two expert reviewers -- a product strategist and a UX specialist -- independently score interaction criteria against journey screenshots and the interaction report, then cross-review each other's scores and findings to surface insights, resolve disagreements, and produce richer feedback than mechanical lower-of-two scoring.
|
|
4
|
+
|
|
5
|
+
The skill acts as the **facilitator**: it orchestrates the debate phases, passes file paths between agents, synthesizes the debate report, detects divergence, manages resolution, and presents results to the user. The facilitator does not participate in the scoring itself.
|
|
6
|
+
|
|
7
|
+
## Team Roles
|
|
8
|
+
|
|
9
|
+
| Role | Agent | Perspective |
|
|
10
|
+
|------|-------|-------------|
|
|
11
|
+
| Builder | `arn-spark-prototype-builder` | Creates prototype screens (not part of debate) |
|
|
12
|
+
| Interactor | `arn-spark-ui-interactor` | Tests journeys via Playwright (not part of debate) |
|
|
13
|
+
| Interaction Strategist | `arn-spark-product-strategist` | Navigation patterns, screen reachability, flow coherence, product goal coverage, journey completability from a product perspective |
|
|
14
|
+
| Interaction Flow Reviewer | `arn-spark-ux-specialist` | Interaction clarity, state changes visibility, transition smoothness, form element functionality, error state handling, journey experience |
|
|
15
|
+
| Facilitator | The skill itself | Orchestrates debate, synthesizes report, manages divergence |
|
|
16
|
+
| Judge | `arn-spark-ux-judge` | Independent interactive verdict (not part of debate) |
|
|
17
|
+
|
|
18
|
+
## Debate Modes
|
|
19
|
+
|
|
20
|
+
### Divergence Mode (Default)
|
|
21
|
+
|
|
22
|
+
Cross-review (Phase 2) triggers only when any criterion score differs by >= 2 points between experts. When experts mostly agree (all scores within 1 point), Phase 2 is skipped and combined scores use the lower of each pair (identical to base skill behavior). This mode saves tokens on cycles where the experts align.
|
|
23
|
+
|
|
24
|
+
### Standard Mode
|
|
25
|
+
|
|
26
|
+
Full cross-review every cycle, regardless of score agreement. Produces richer debate findings but costs more tokens. Use when the project has complex interaction patterns where expert dialogue adds value even on criteria they numerically agree about.
|
|
27
|
+
|
|
28
|
+
## Execution Modes
|
|
29
|
+
|
|
30
|
+
**Important:** All execution modes use the same file-based review output. Each expert writes its review to a file -- this works identically in Agent Teams mode and sequential mode. The execution mode selection is based ONLY on whether Agent Teams is supported. File-based output does NOT affect mode selection and does NOT favor sequential over Agent Teams. When Agent Teams is supported, always use Agent Teams mode -- it is faster because experts run in parallel.
|
|
31
|
+
|
|
32
|
+
### Agent Teams Mode (Preferred)
|
|
33
|
+
|
|
34
|
+
**When:** Agent Teams is supported by your platform.
|
|
35
|
+
|
|
36
|
+
Both experts are spawned as teammates. Phase 1 runs in parallel -- each expert writes to its own file simultaneously with no contention. Phase 2 uses Teams communication to coordinate cross-review -- each expert reads the other's completed file and writes its cross-review to a separate file.
|
|
37
|
+
|
|
38
|
+
### Sequential Mode (Fallback)
|
|
39
|
+
|
|
40
|
+
**When:** Agent Teams is NOT enabled.
|
|
41
|
+
|
|
42
|
+
The skill simulates the debate through sequential expert invocations, manually passing file paths between agents so each can read the other's review. Produces the same logical result as Agent Teams mode but with serialized invocations.
|
|
43
|
+
|
|
44
|
+
### Single-Reviewer Mode
|
|
45
|
+
|
|
46
|
+
**When:** `arn-spark-ux-specialist` is unavailable.
|
|
47
|
+
|
|
48
|
+
No debate occurs. The product strategist reviews independently. Strategist scores become the combined scores directly. The debate report notes "Single-Reviewer Mode" throughout. The skill suggests using `/arn-spark-clickable-prototype` instead, which handles single-reviewer identically.
|
|
49
|
+
|
|
50
|
+
## Debate Phases
|
|
51
|
+
|
|
52
|
+
### Phase 1: Independent Scoring
|
|
53
|
+
|
|
54
|
+
Both experts independently score ALL criteria against the journey screenshots and interaction report. Neither sees the other's scores during this phase.
|
|
55
|
+
|
|
56
|
+
**Product strategist focus areas (interaction perspective):**
|
|
57
|
+
- Can every screen be reached from the hub navigation?
|
|
58
|
+
- Does the navigation flow match the product concept's intended user journeys?
|
|
59
|
+
- Can each defined user journey be completed from start to finish?
|
|
60
|
+
- Do navigation elements consistently indicate the current location?
|
|
61
|
+
- Are all screens present that the product concept requires?
|
|
62
|
+
- Is the screen organization (functional area grouping) logical for the target user?
|
|
63
|
+
- Do error states provide helpful guidance aligned with the product's tone?
|
|
64
|
+
- If visual grounding assets exist:
|
|
65
|
+
- References: Do screen layouts and flow feel align with the reference direction?
|
|
66
|
+
- Designs: Do screen layouts match design mockups in structure and component placement?
|
|
67
|
+
- Brand: Do brand elements appear correctly across all screens?
|
|
68
|
+
|
|
69
|
+
**UX specialist focus areas (interaction perspective):**
|
|
70
|
+
- Do all clickable elements respond with visual feedback?
|
|
71
|
+
- Do form elements function correctly (inputs accept text, dropdowns open, checkboxes toggle)?
|
|
72
|
+
- Are interactive state changes (selected, active, expanded, collapsed) visually clear?
|
|
73
|
+
- Are page transitions smooth and not jarring?
|
|
74
|
+
- At each step of a journey, is the next action obvious to the user?
|
|
75
|
+
- Can users navigate back without relying on browser controls?
|
|
76
|
+
- Is there a dead end anywhere (a screen with no path forward or back)?
|
|
77
|
+
- Are component interactions consistent across screens (same component, same behavior)?
|
|
78
|
+
- Does responsive behavior work correctly (if applicable)?
|
|
79
|
+
- Are there JavaScript errors or broken assets during normal interaction?
|
|
80
|
+
|
|
81
|
+
**Phase 1 file output:** Each expert writes their review to a file using the expert interaction review template (`${CLAUDE_PLUGIN_ROOT}/skills/arn-spark-clickable-prototype-teams/references/expert-interaction-review-template.md`). The skill tells each agent the exact file path to write to:
|
|
82
|
+
- Product strategist -> `prototypes/clickable/reviews/round-N-strategist-review.md`
|
|
83
|
+
- UX specialist -> `prototypes/clickable/reviews/round-N-ux-review.md`
|
|
84
|
+
|
|
85
|
+
The agent returns a brief summary in conversation -- the full review is in the file. Downstream steps read from the file, not from conversation context.
|
|
86
|
+
|
|
87
|
+
### Divergence Check (Performed by Skill)
|
|
88
|
+
|
|
89
|
+
After Phase 1, the skill reads both review files and extracts per-criterion scores from the "Per-Criterion Scores" table.
|
|
90
|
+
|
|
91
|
+
**In divergence mode:** Calculate `|strategist_score - ux_score|` for each criterion.
|
|
92
|
+
- If max divergence < 2: Skip Phase 2. Combined score per criterion = `min(strategist, ux)`. Present to user: "Experts scored within 1 point on all criteria. No divergence detected -- skipping cross-review."
|
|
93
|
+
- If any divergence >= 2: Proceed to Phase 2. Present to user: "Divergence detected on [N] criteria (difference >= 2 points): [list criteria names and score pairs]. Triggering cross-review."
|
|
94
|
+
|
|
95
|
+
**In standard mode:** Always proceed to Phase 2 regardless of score differences.
|
|
96
|
+
|
|
97
|
+
### Phase 2: Cross-Review
|
|
98
|
+
|
|
99
|
+
Each expert reads the other's Phase 1 file and responds per-criterion.
|
|
100
|
+
|
|
101
|
+
**Instructions for each expert during cross-review:**
|
|
102
|
+
|
|
103
|
+
For each criterion:
|
|
104
|
+
- **Agree** -- the other expert's score is valid. Optionally adjust own score (up or down) with reasoning.
|
|
105
|
+
- **Disagree** -- the other expert's score is incorrect. Maintain own score with counter-evidence (reference specific journey steps, screenshots, or interaction behaviors).
|
|
106
|
+
- **New concern prompted** -- the other expert's review reveals something not previously noticed. Add observation.
|
|
107
|
+
|
|
108
|
+
For each journey:
|
|
109
|
+
- Note whether you agree with the other expert's journey assessment (Complete/Partial/Failed).
|
|
110
|
+
- If you disagree on a journey outcome, provide specific step-level evidence.
|
|
111
|
+
|
|
112
|
+
Focus on DIVERGENT criteria (score difference >= 2). For criteria with small differences (<= 1 point), a brief acknowledgment suffices.
|
|
113
|
+
|
|
114
|
+
**Phase 2 file output:** Each expert writes their cross-review to a separate file:
|
|
115
|
+
- Product strategist -> `prototypes/clickable/reviews/round-N-strategist-cross-review.md`
|
|
116
|
+
- UX specialist -> `prototypes/clickable/reviews/round-N-ux-cross-review.md`
|
|
117
|
+
|
|
118
|
+
In sequential mode (where the UX specialist writes Phase 1 + Phase 2 combined), the combined output goes to `round-N-ux-review.md` (a single file with both sections).
|
|
119
|
+
|
|
120
|
+
### Phase 3: Synthesis (Performed by Skill)
|
|
121
|
+
|
|
122
|
+
The skill reads all review files written by the experts -- never from conversation context. The files to read are:
|
|
123
|
+
- `prototypes/clickable/reviews/round-N-strategist-review.md` (Phase 1)
|
|
124
|
+
- `prototypes/clickable/reviews/round-N-ux-review.md` (Phase 1, or Phase 1 + Phase 2 combined in sequential mode)
|
|
125
|
+
- `prototypes/clickable/reviews/round-N-strategist-cross-review.md` (Phase 2, if written separately)
|
|
126
|
+
- `prototypes/clickable/reviews/round-N-ux-cross-review.md` (Phase 2, Agent Teams mode only)
|
|
127
|
+
|
|
128
|
+
For each criterion, categorize:
|
|
129
|
+
|
|
130
|
+
**Consensus:** Both experts scored the same, or one adjusted their score in cross-review to match the other. Combined score = the agreed score.
|
|
131
|
+
|
|
132
|
+
**Additions:** One expert scored lower with specific feedback, the other did not dispute the lower score in cross-review (neither agreed nor disagreed). Combined score = the lower score.
|
|
133
|
+
|
|
134
|
+
**Disagreements:** Both experts maintained different scores after cross-review -- one raised a concern and the other explicitly disagreed. These require user resolution in Phase 4.
|
|
135
|
+
|
|
136
|
+
**No-debate:** Criteria where Phase 2 was skipped (divergence mode, no divergence detected). Combined score = `min(strategist, ux)`.
|
|
137
|
+
|
|
138
|
+
Also synthesize journey assessments: if experts disagree on whether a journey completed, note the disagreement in the debate report and include it in the resolution step if the journey outcome affects a criterion score.
|
|
139
|
+
|
|
140
|
+
Write the debate review report using the debate review report template. Save to `prototypes/clickable/reviews/round-N-cycle-M-debate-report.md`.
|
|
141
|
+
|
|
142
|
+
### Phase 4: Resolution (Conditional)
|
|
143
|
+
|
|
144
|
+
**Trigger:** One or more unresolved disagreements exist after Phase 3.
|
|
145
|
+
|
|
146
|
+
For each unresolved disagreement, present to the user:
|
|
147
|
+
|
|
148
|
+
"Expert disagreement on **[Criterion Name]** (criterion #[N]):
|
|
149
|
+
- **Product Strategist:** Score [X] -- [reasoning and evidence with journey reference]
|
|
150
|
+
- **UX Specialist:** Score [Y] -- [reasoning and evidence with journey reference]
|
|
151
|
+
- **Trade-off:** [what each score optimizes for]
|
|
152
|
+
|
|
153
|
+
What score should this criterion receive?"
|
|
154
|
+
|
|
155
|
+
Record user decisions. Update the debate report with resolutions. The resolved score becomes the final combined score for that criterion.
|
|
156
|
+
|
|
157
|
+
## Sequential Mode Invocation Detail
|
|
158
|
+
|
|
159
|
+
When Agent Teams is not enabled, the skill simulates the debate with 3 sequential invocations per round:
|
|
160
|
+
|
|
161
|
+
**Invocation 1 -- Product Strategist Phase 1:**
|
|
162
|
+
|
|
163
|
+
Invoke the `arn-spark-product-strategist` agent via the Task tool, passing the model from `.arness/agent-models/spark.md` as the `model` parameter (see `plugins/arn-spark/skills/arn-spark-ensure-config/references/ensure-config.md` "Dispatch convention" for fallback). Context:
|
|
164
|
+
- Journey screenshots from interaction testing step
|
|
165
|
+
- Interaction report from `arn-spark-ui-interactor`
|
|
166
|
+
- All criteria with descriptions, scoring scale, and threshold
|
|
167
|
+
- Style brief, product concept, visual grounding assets (with category context)
|
|
168
|
+
- Expert interaction review template path: `${CLAUDE_PLUGIN_ROOT}/skills/arn-spark-clickable-prototype-teams/references/expert-interaction-review-template.md`
|
|
169
|
+
- File path to write to: `prototypes/clickable/reviews/round-N-strategist-review.md`
|
|
170
|
+
- Instruction: "Score every criterion independently against the journey screenshots and interaction report. Assess every journey for completability. Write your complete review to the specified file path using the expert interaction review template. Return a brief summary in conversation."
|
|
171
|
+
|
|
172
|
+
**Invocation 2 -- UX Specialist Phase 1 + Phase 2 Combined:**
|
|
173
|
+
|
|
174
|
+
Invoke the `arn-spark-ux-specialist` agent via the Task tool, passing the model from `.arness/agent-models/spark.md` as the `model` parameter (see `plugins/arn-spark/skills/arn-spark-ensure-config/references/ensure-config.md` "Dispatch convention" for fallback). Context:
|
|
175
|
+
- Same inputs as Invocation 1
|
|
176
|
+
- The strategist's file path to read: `prototypes/clickable/reviews/round-N-strategist-review.md`
|
|
177
|
+
- File path to write to: `prototypes/clickable/reviews/round-N-ux-review.md`
|
|
178
|
+
- Instruction: "First, score every criterion independently from the UX perspective and assess every journey, using the Phase 1 format. Then, read the strategist's review at the specified file path and respond to each criterion score and journey assessment using the Phase 2 cross-review format: agree (optionally adjust your score), disagree (with counter-evidence), or note new concerns. Write your complete review (Phase 1 + Phase 2 combined) to the specified file path. Return a brief summary in conversation."
|
|
179
|
+
|
|
180
|
+
**Invocation 3 -- Product Strategist Phase 2:**
|
|
181
|
+
|
|
182
|
+
Invoke the `arn-spark-product-strategist` agent via the Task tool, passing the model from `.arness/agent-models/spark.md` as the `model` parameter (see `plugins/arn-spark/skills/arn-spark-ensure-config/references/ensure-config.md` "Dispatch convention" for fallback). Context:
|
|
183
|
+
- The UX specialist's file path to read: `prototypes/clickable/reviews/round-N-ux-review.md`
|
|
184
|
+
- Expert interaction review template path
|
|
185
|
+
- File path to write to: `prototypes/clickable/reviews/round-N-strategist-cross-review.md`
|
|
186
|
+
- Instruction: "Read the UX specialist's review at the specified file path. The UX specialist has scored the criteria, assessed journeys, and also responded to your review. Respond to their scores, journey assessments, and cross-review using the Phase 2 format. Write your cross-review to the specified file path. Return a brief summary in conversation."
|
|
187
|
+
|
|
188
|
+
The skill synthesizes by reading all three review files (not from conversation context).
|
|
189
|
+
|
|
190
|
+
**Note on sequential asymmetry:** In sequential mode, the UX specialist sees the strategist's review before writing their own Phase 1 review. Instruct the UX specialist to "score independently first" to minimize anchoring bias. The synthesis step normalizes the output regardless of invocation order.
|
|
191
|
+
|
|
192
|
+
## Agent Teams Mode Invocation Detail
|
|
193
|
+
|
|
194
|
+
**Phase 1:** Spawn both experts simultaneously as teammates. Each receives:
|
|
195
|
+
- Journey screenshots, interaction report, criteria, scoring parameters, style brief, product concept, visual grounding assets
|
|
196
|
+
- Expert interaction review template path
|
|
197
|
+
- File path to write to: strategist -> `round-N-strategist-review.md`, UX specialist -> `round-N-ux-review.md`
|
|
198
|
+
- Instruction: "Score every criterion independently. Assess every journey. Write the complete review to the specified file path using the expert interaction review template. Do not communicate with other teammates during this phase."
|
|
199
|
+
|
|
200
|
+
Both produce Phase 1 reviews independently and write them to their respective files.
|
|
201
|
+
|
|
202
|
+
**Runtime verification:** After Phase 1, the skill checks that BOTH review files exist and contain per-criterion scores. If one file is missing (Agent Teams silently failed to spawn one expert), invoke the missing expert sequentially and note the issue in the debate report: "Agent Teams Phase 1 partial failure: [agent] did not produce its review file. Invoked sequentially as fallback."
|
|
203
|
+
|
|
204
|
+
**Phase 2:** Share file paths through Teams communication:
|
|
205
|
+
- Tell the UX specialist to read the strategist's file: `round-N-strategist-review.md`
|
|
206
|
+
- Tell the strategist to read the UX specialist's file: `round-N-ux-review.md`
|
|
207
|
+
- Each reads the other's file and writes their cross-review to a separate file:
|
|
208
|
+
- Product strategist -> `round-N-strategist-cross-review.md`
|
|
209
|
+
- UX specialist -> `round-N-ux-cross-review.md`
|
|
210
|
+
- Each responds using the Phase 2 cross-review format from the expert interaction review template
|
|
211
|
+
|
|
212
|
+
The skill synthesizes by reading all four review files (not from conversation context).
|
|
213
|
+
|
|
214
|
+
## Invocation Counts per Cycle
|
|
215
|
+
|
|
216
|
+
| Execution Mode | Debate Mode | Divergence Found | Expert Invocations | Notes |
|
|
217
|
+
|---------------|-------------|------------------|-------------------|-------|
|
|
218
|
+
| Agent Teams | Standard | N/A (always) | 4 (2 P1 + 2 P2) | Parallel within phases |
|
|
219
|
+
| Agent Teams | Divergence | Yes | 4 (2 P1 + 2 P2) | Same as standard |
|
|
220
|
+
| Agent Teams | Divergence | No | 2 (P1 only) | Phase 2 skipped |
|
|
221
|
+
| Sequential | Standard | N/A (always) | 3 (strat P1, UX P1+P2, strat P2) | |
|
|
222
|
+
| Sequential | Divergence | Yes | 3 | |
|
|
223
|
+
| Sequential | Divergence | No | 2 (strat P1, UX P1) | Phase 2 skipped |
|
|
224
|
+
| Single-reviewer | Any | N/A | 1 | No debate |
|
|
225
|
+
|
|
226
|
+
For max_cycles=3 with Agent Teams + standard mode: up to 12 expert invocations + 3 build cycles + 3 interaction testing cycles + judge + showcase.
|
|
227
|
+
|
|
228
|
+
## Skill's Facilitation Responsibilities
|
|
229
|
+
|
|
230
|
+
The `arn-spark-clickable-prototype-teams` skill (not the agents) is responsible for:
|
|
231
|
+
|
|
232
|
+
1. **Agent Teams verification:** Check env var in Step 1 AND verify both experts write files after Phase 1. If one expert's file is missing, invoke sequentially as fallback and log the issue.
|
|
233
|
+
2. **Mode detection:** Record execution mode (agent_teams / sequential) and debate mode (divergence / standard) from Step 1 and Step 3
|
|
234
|
+
3. **Directory setup:** Create `prototypes/clickable/reviews/` if it does not exist before invoking experts
|
|
235
|
+
4. **File path assignment:** Tell each expert agent the exact file path to write to and the expert interaction review template path
|
|
236
|
+
5. **Phase orchestration:** Run Phase 1, perform divergence check, conditionally run Phase 2, telling each expert to read the other's review file (by file path, not by passing content through conversation)
|
|
237
|
+
6. **Synthesis:** Read all expert review files and categorize per criterion into consensus, additions, disagreements. Also compare per-journey assessments. Never rely on the expert's conversation summary -- always read the file.
|
|
238
|
+
7. **Score computation:** For each criterion, compute the final combined score based on debate outcome (consensus: agreed score, additions: lower score, disagreements: user-resolved score, no-debate: min of two)
|
|
239
|
+
8. **Conflict detection:** Identify disagreements (both criterion-level and journey-level) and present to user for resolution with both positions and evidence
|
|
240
|
+
9. **Report writing:** Produce the debate review report per template and save to file
|
|
241
|
+
10. **Budget management:** Never exceed the user's configured max_cycles
|
|
242
|
+
11. **User communication:** Present divergence status, debate summaries, journey agreement status, and resolution requests clearly between phases
|
|
@@ -0,0 +1,161 @@
|
|
|
1
|
+
# Debate Review Report Template -- Clickable Prototype Teams
|
|
2
|
+
|
|
3
|
+
Use this template for each cycle's debate review report produced by the `arn-spark-clickable-prototype-teams` skill. The skill populates this template after synthesizing the expert debate outputs (Phase 1 scores + Phase 2 cross-review responses) into categorized findings with final combined scores.
|
|
4
|
+
|
|
5
|
+
## Instructions for the Skill
|
|
6
|
+
|
|
7
|
+
When populating this template:
|
|
8
|
+
|
|
9
|
+
- Every section MUST appear, even if empty (write "None" for empty sections)
|
|
10
|
+
- **Consensus:** both experts agree on the score, or one adjusted to match in cross-review
|
|
11
|
+
- **Additions:** one expert scored lower with feedback, the other did not dispute
|
|
12
|
+
- **Disagreements:** experts explicitly disagreed in cross-review and disagreement persists
|
|
13
|
+
- **No-debate:** Phase 2 was skipped (divergence mode, no divergence detected) -- combined = min(strategist, ux)
|
|
14
|
+
- In single-reviewer mode (no UX specialist): all scores are from the strategist alone. Omit Disagreements and note "Single-Reviewer Mode" throughout.
|
|
15
|
+
- Save each report to `prototypes/clickable/reviews/round-N-cycle-M-debate-report.md`
|
|
16
|
+
- Also copy to `prototypes/clickable/v[M]/review-report.md` for version-local access
|
|
17
|
+
|
|
18
|
+
---
|
|
19
|
+
|
|
20
|
+
## Template
|
|
21
|
+
|
|
22
|
+
```markdown
|
|
23
|
+
# Debate Review Report: Cycle [M], Round [N]
|
|
24
|
+
|
|
25
|
+
## Debate Participants
|
|
26
|
+
|
|
27
|
+
| Role | Agent | Status |
|
|
28
|
+
|------|-------|--------|
|
|
29
|
+
| Interaction Strategist | arn-spark-product-strategist | Participated |
|
|
30
|
+
| Interaction Flow Reviewer | arn-spark-ux-specialist | Participated / Unavailable |
|
|
31
|
+
|
|
32
|
+
## Configuration
|
|
33
|
+
|
|
34
|
+
- **Debate mode:** [Divergence / Standard]
|
|
35
|
+
- **Execution mode:** [Agent Teams / Sequential / Single-Reviewer]
|
|
36
|
+
- **Phase 2 triggered:** [Yes -- [N] criteria diverged by >= 2 / No -- all within 1 point / Yes -- standard mode (always)]
|
|
37
|
+
- **Divergent criteria:** [list names, or "None"]
|
|
38
|
+
- **Scoring scale:** [1-N]
|
|
39
|
+
- **Minimum threshold:** [T]
|
|
40
|
+
|
|
41
|
+
## Criterion Scores
|
|
42
|
+
|
|
43
|
+
| # | Criterion | Strategist | UX Specialist | Combined | Status | Category |
|
|
44
|
+
|---|-----------|-----------|---------------|----------|--------|----------|
|
|
45
|
+
| 1 | [name] | [score] | [score] | [combined] | PASS/FAIL | [Consensus/Addition/Disagreement/No-debate] |
|
|
46
|
+
| 2 | [name] | [score] | [score] | [combined] | PASS/FAIL | [category] |
|
|
47
|
+
| ... | ... | ... | ... | ... | ... | ... |
|
|
48
|
+
|
|
49
|
+
## Visual Grounding Comparison
|
|
50
|
+
|
|
51
|
+
**Assets provided to reviewers:**
|
|
52
|
+
| Category | Count | Source |
|
|
53
|
+
|----------|-------|--------|
|
|
54
|
+
| References | [N] | [URL captures, user screenshots] |
|
|
55
|
+
| Designs | [N] | [Figma exports, Canva exports, manual mockups] |
|
|
56
|
+
| Brand | [N] | [logos, guidelines] |
|
|
57
|
+
|
|
58
|
+
**Comparison notes:**
|
|
59
|
+
- **Reference alignment:** [How well screen layouts and flow feel match the inspirational direction]
|
|
60
|
+
- **Design fidelity:** [How closely screen layouts match the design mockups -- only if designs exist]
|
|
61
|
+
- **Brand compliance:** [Whether brand elements appear correctly across screens -- only if brand assets exist]
|
|
62
|
+
|
|
63
|
+
[If no visual grounding assets: "No visual grounding assets provided. Review based on style brief text only."]
|
|
64
|
+
|
|
65
|
+
## Journey Results Summary
|
|
66
|
+
|
|
67
|
+
| # | Journey | Strategist Assessment | UX Assessment | Agreed | Issues |
|
|
68
|
+
|---|---------|----------------------|---------------|--------|--------|
|
|
69
|
+
| 1 | [name] | [Complete/Partial/Failed] | [Complete/Partial/Failed] | [Yes/No] | [brief summary] |
|
|
70
|
+
| 2 | [name] | [Complete/Partial/Failed] | [Complete/Partial/Failed] | [Yes/No] | [brief summary] |
|
|
71
|
+
| ... | ... | ... | ... | ... | ... |
|
|
72
|
+
|
|
73
|
+
[If single-reviewer: only one assessment column, "Agreed" column reads "N/A"]
|
|
74
|
+
|
|
75
|
+
## Debate Findings
|
|
76
|
+
|
|
77
|
+
### Consensus Criteria
|
|
78
|
+
|
|
79
|
+
[Criteria where both experts agreed or one adjusted to match]
|
|
80
|
+
|
|
81
|
+
**Criterion [N]: [Name]** -- Combined [X]/[scale]
|
|
82
|
+
- **Strategist:** [brief reasoning with journey/screenshot reference]
|
|
83
|
+
- **UX Specialist:** [brief reasoning with journey/screenshot reference]
|
|
84
|
+
- **Outcome:** Both agree. [Any shared feedback for builder.]
|
|
85
|
+
|
|
86
|
+
[Repeat for each consensus criterion, or "None"]
|
|
87
|
+
|
|
88
|
+
### Addition Criteria
|
|
89
|
+
|
|
90
|
+
[Criteria where one expert raised feedback the other did not dispute]
|
|
91
|
+
|
|
92
|
+
**Criterion [N]: [Name]** -- Combined [X]/[scale] (raised by [Strategist / UX Specialist])
|
|
93
|
+
- **Lower scorer:** [agent] scored [X] -- [reasoning with journey evidence]
|
|
94
|
+
- **Higher scorer:** [agent] scored [Y] -- did not dispute
|
|
95
|
+
- **Outcome:** Lower score used. Builder feedback: [specific suggestion with journey step reference]
|
|
96
|
+
|
|
97
|
+
[Repeat for each addition criterion, or "None"]
|
|
98
|
+
|
|
99
|
+
### Disagreement Criteria
|
|
100
|
+
|
|
101
|
+
[Criteria where experts explicitly disagreed after cross-review]
|
|
102
|
+
|
|
103
|
+
**Criterion [N]: [Name]**
|
|
104
|
+
- **Strategist:** Score [X] -- [position + reasoning + journey/screenshot evidence]
|
|
105
|
+
- **UX Specialist:** Score [Y] -- [position + reasoning + journey/screenshot evidence]
|
|
106
|
+
- **Trade-off:** [what each score optimizes for]
|
|
107
|
+
- **Resolution:** [User decided: score [Z] because [reasoning] / Pending user input]
|
|
108
|
+
|
|
109
|
+
[Repeat for each disagreement criterion, or "None"]
|
|
110
|
+
|
|
111
|
+
### No-Debate Criteria
|
|
112
|
+
|
|
113
|
+
[Criteria where Phase 2 was skipped -- divergence mode only]
|
|
114
|
+
|
|
115
|
+
[If Phase 2 was skipped:] All criteria scored within 1 point. Combined = min(strategist, ux). No cross-review was performed.
|
|
116
|
+
|
|
117
|
+
[If Phase 2 ran:] N/A -- all criteria were included in the debate.
|
|
118
|
+
|
|
119
|
+
## Failing Criteria
|
|
120
|
+
|
|
121
|
+
### [Criterion Name] -- Combined [X]/[scale]
|
|
122
|
+
- **Strategist feedback:** [specific observation and suggestion with journey/screen reference]
|
|
123
|
+
- **UX specialist feedback:** [specific observation and suggestion with journey/screen reference]
|
|
124
|
+
- **Journey evidence:** [which journey step exposed the issue, with screenshot reference]
|
|
125
|
+
- **Debate insight:** [anything surfaced during cross-review that adds context beyond individual feedback]
|
|
126
|
+
- **Priority:** [Critical / Important]
|
|
127
|
+
|
|
128
|
+
[Repeat for each failing criterion]
|
|
129
|
+
|
|
130
|
+
## Passing Criteria Highlights
|
|
131
|
+
[Brief notes on particularly strong aspects]
|
|
132
|
+
|
|
133
|
+
## Summary
|
|
134
|
+
|
|
135
|
+
- **Passing:** [N] of [M] criteria meet threshold
|
|
136
|
+
- **Failing:** [N] criteria below threshold
|
|
137
|
+
- **Journeys:** [X] of [Y] completed successfully (agreed by both experts)
|
|
138
|
+
- **Phase 2 triggered:** [Yes / No]
|
|
139
|
+
- **Consensus criteria:** [N]
|
|
140
|
+
- **Addition criteria:** [N]
|
|
141
|
+
- **Disagreement criteria:** [N] ([N] resolved by user)
|
|
142
|
+
- **No-debate criteria:** [N]
|
|
143
|
+
- **Verdict:** PROCEED TO NEXT CYCLE / ALL CRITERIA PASS -- PROCEED TO JUDGE
|
|
144
|
+
|
|
145
|
+
## Recommended Focus for Next Cycle
|
|
146
|
+
|
|
147
|
+
[If failing: ordered list of what to fix, most critical first, incorporating debate insights and journey evidence. Each item includes the debate context so the builder understands WHY, not just what to fix.]
|
|
148
|
+
|
|
149
|
+
1. **[Criterion Name]:** [specific fix] -- Journey [N], Step [M]: [what went wrong]. [debate context: both experts agreed / strategist flagged X while UX specialist noted Y / user resolved in favor of Z]
|
|
150
|
+
2. ...
|
|
151
|
+
```
|
|
152
|
+
|
|
153
|
+
## Usage Notes
|
|
154
|
+
|
|
155
|
+
- The skill writes this report to `prototypes/clickable/reviews/round-N-cycle-M-debate-report.md` after each debate cycle
|
|
156
|
+
- Also copy to `prototypes/clickable/v[M]/review-report.md` for version-local access
|
|
157
|
+
- The "Recommended Focus for Next Cycle" section is the primary output fed to the builder for the next cycle -- it must be actionable and specific, enriched by debate context and journey evidence
|
|
158
|
+
- In single-reviewer mode, the Category column is always "Single-reviewer", the Disagreements section reads "N/A -- single-reviewer mode", and the UX Specialist column shows "N/A"
|
|
159
|
+
- When Phase 2 is skipped in divergence mode, all criteria are categorized as "No-debate" and the Debate Findings section reflects this
|
|
160
|
+
- The Journey Results Summary table captures expert agreement on journey outcomes -- if experts disagree on whether a journey completed, this is noted and may trigger additional debate or user resolution
|
|
161
|
+
- When writing the final report (`prototypes/clickable/final-report.md`), aggregate all per-cycle debate reports with a summary of the debate arc: how scores evolved, what diverged, what converged, what the user decided
|
|
@@ -0,0 +1,152 @@
|
|
|
1
|
+
# Expert Interaction Review Template
|
|
2
|
+
|
|
3
|
+
This template defines the file format that expert agents (`arn-spark-product-strategist`, `arn-spark-ux-specialist`) use when writing their interaction review reports to disk during `arn-spark-clickable-prototype-teams` debate cycles. Writing reviews to files ensures they survive context compression and provides a full audit trail.
|
|
4
|
+
|
|
5
|
+
## File Naming Convention
|
|
6
|
+
|
|
7
|
+
All review files go in `prototypes/clickable/reviews/`. Create the directory if it does not exist.
|
|
8
|
+
|
|
9
|
+
```
|
|
10
|
+
prototypes/clickable/reviews/
|
|
11
|
+
├── round-N-strategist-review.md ← Product strategist Phase 1
|
|
12
|
+
├── round-N-ux-review.md ← UX specialist Phase 1 (+ Phase 2 in sequential mode)
|
|
13
|
+
├── round-N-strategist-cross-review.md ← Product strategist Phase 2 response
|
|
14
|
+
├── round-N-ux-cross-review.md ← UX specialist Phase 2 response (Agent Teams only)
|
|
15
|
+
└── round-N-cycle-M-debate-report.md ← Synthesized debate report (written by skill)
|
|
16
|
+
```
|
|
17
|
+
|
|
18
|
+
Where `N` is the overall round number and `M` is the cycle number within the current validation run.
|
|
19
|
+
|
|
20
|
+
---
|
|
21
|
+
|
|
22
|
+
## Phase 1 Template (Independent Scoring)
|
|
23
|
+
|
|
24
|
+
When an expert writes their Phase 1 review, the file must follow this structure:
|
|
25
|
+
|
|
26
|
+
```markdown
|
|
27
|
+
# Interaction Review: Round [N] -- [Product Strategist / UX Specialist]
|
|
28
|
+
|
|
29
|
+
**Agent:** [arn-spark-product-strategist / arn-spark-ux-specialist]
|
|
30
|
+
**Phase:** Phase 1: Independent Scoring
|
|
31
|
+
**Execution mode:** [Agent Teams / Sequential / Single-Reviewer]
|
|
32
|
+
**Criteria scored:** [count]
|
|
33
|
+
**Journey screenshots reviewed:** [count]
|
|
34
|
+
**Journeys assessed:** [count]
|
|
35
|
+
**Version:** v[X]
|
|
36
|
+
|
|
37
|
+
---
|
|
38
|
+
|
|
39
|
+
## Per-Criterion Scores
|
|
40
|
+
|
|
41
|
+
| # | Criterion | Score | Evidence |
|
|
42
|
+
|---|-----------|-------|----------|
|
|
43
|
+
| 1 | [name] | [X]/[scale] | [1-2 sentence observation grounded in specific journey screenshot evidence] |
|
|
44
|
+
| 2 | [name] | [X]/[scale] | [evidence] |
|
|
45
|
+
| ... | ... | ... | ... |
|
|
46
|
+
|
|
47
|
+
## Per-Journey Assessment
|
|
48
|
+
|
|
49
|
+
| # | Journey | Steps | Completed | Issues | Key Screenshots |
|
|
50
|
+
|---|---------|-------|-----------|--------|----------------|
|
|
51
|
+
| 1 | [name] | [total] | [count] | [brief summary or "None"] | [screenshot filenames] |
|
|
52
|
+
| 2 | [name] | [total] | [count] | [brief summary or "None"] | [screenshot filenames] |
|
|
53
|
+
| ... | ... | ... | ... | ... | ... |
|
|
54
|
+
|
|
55
|
+
## Failing Criteria Detail
|
|
56
|
+
|
|
57
|
+
### [Criterion Name] -- [X]/[scale]
|
|
58
|
+
|
|
59
|
+
- **Observation:** [What specifically is wrong -- reference journey step, screenshot, screen, or interaction]
|
|
60
|
+
- **Expected:** [What the criteria description requires]
|
|
61
|
+
- **Journey evidence:** [Which journey step exposed the issue, with screenshot reference]
|
|
62
|
+
- **Suggestion:** [Specific actionable improvement for the builder]
|
|
63
|
+
- **Priority:** [Critical / Important]
|
|
64
|
+
|
|
65
|
+
[Repeat for each criterion below threshold]
|
|
66
|
+
|
|
67
|
+
## Passing Criteria Highlights
|
|
68
|
+
|
|
69
|
+
[Brief notes on particularly strong aspects, optional]
|
|
70
|
+
|
|
71
|
+
## Cross-Cutting Observations
|
|
72
|
+
|
|
73
|
+
- **Navigation coherence:** [Overall observation about screen reachability and navigation flow]
|
|
74
|
+
- **Interaction quality:** [Overall observation about interactive element responsiveness and feedback]
|
|
75
|
+
- **Journey experience:** [Overall observation about journey flow, clarity, and completability]
|
|
76
|
+
- **Visual consistency:** [Overall observation about style consistency across screens]
|
|
77
|
+
- **Missing elements:** [Anything expected but not present in the prototype]
|
|
78
|
+
```
|
|
79
|
+
|
|
80
|
+
---
|
|
81
|
+
|
|
82
|
+
## Phase 2 Template (Cross-Review)
|
|
83
|
+
|
|
84
|
+
When an expert writes their Phase 2 cross-review, the file must follow this structure. In sequential mode where the UX specialist writes Phase 1 + Phase 2 combined, append this section after the Phase 1 content in the same file.
|
|
85
|
+
|
|
86
|
+
```markdown
|
|
87
|
+
## Cross-Review Response
|
|
88
|
+
|
|
89
|
+
### Response to [Product Strategist / UX Specialist]'s Scores
|
|
90
|
+
|
|
91
|
+
| # | Criterion | Their Score | My Score | Response | Adjusted Score |
|
|
92
|
+
|---|-----------|-------------|----------|----------|----------------|
|
|
93
|
+
| 1 | [name] | [X] | [Y] | [Agree/Disagree/New concern] | [new score or unchanged] |
|
|
94
|
+
| 2 | [name] | [X] | [Y] | [Agree/Disagree/New concern] | [new score or unchanged] |
|
|
95
|
+
| ... | ... | ... | ... | ... | ... |
|
|
96
|
+
|
|
97
|
+
### Journey Assessment Comparison
|
|
98
|
+
|
|
99
|
+
| # | Journey | Their Assessment | My Assessment | Response |
|
|
100
|
+
|---|---------|-----------------|---------------|----------|
|
|
101
|
+
| 1 | [name] | [Completed/Partial/Failed] | [Completed/Partial/Failed] | [Agree/Disagree with reason] |
|
|
102
|
+
| ... | ... | ... | ... | ... |
|
|
103
|
+
|
|
104
|
+
### Detailed Responses (for divergent criteria)
|
|
105
|
+
|
|
106
|
+
**Criterion [N]: [Name]**
|
|
107
|
+
- **Their score:** [X] -- "[their evidence summary]"
|
|
108
|
+
- **My score:** [Y]
|
|
109
|
+
- **Response:** [Agree / Disagree / New concern]
|
|
110
|
+
- **Reasoning:** [Specific counter-evidence or supporting evidence, referencing journey steps and screenshots]
|
|
111
|
+
- **Adjusted score:** [new score, or same if maintaining position]
|
|
112
|
+
|
|
113
|
+
[Repeat for each criterion with score difference >= 2, or all criteria in standard mode]
|
|
114
|
+
|
|
115
|
+
### New Concerns Prompted by Other Expert's Review
|
|
116
|
+
|
|
117
|
+
- [Description of something their review revealed that was not noticed in Phase 1, with journey/screenshot references]
|
|
118
|
+
```
|
|
119
|
+
|
|
120
|
+
---
|
|
121
|
+
|
|
122
|
+
## Instructions for Expert Agents
|
|
123
|
+
|
|
124
|
+
When instructed to write an interaction review:
|
|
125
|
+
|
|
126
|
+
1. Read all journey screenshots provided (visually, via multimodal capabilities)
|
|
127
|
+
2. Read the interaction report from `arn-spark-ui-interactor` for journey completion data
|
|
128
|
+
3. Read the criteria list and scoring scale
|
|
129
|
+
4. Read the style brief and product concept for context
|
|
130
|
+
5. Read visual grounding assets if provided (with their category context: references=inspirational direction, designs=specification targets, brand=constraints)
|
|
131
|
+
6. Score EVERY criterion -- do not skip or combine criteria
|
|
132
|
+
7. Assess EVERY journey -- note completion status, issues, and key screenshots
|
|
133
|
+
8. For each score, provide specific evidence grounded in what you observe in the journey screenshots and interaction report
|
|
134
|
+
9. Read this template to understand the expected file format
|
|
135
|
+
10. Write your review to the exact file path specified by the caller
|
|
136
|
+
11. Return a brief summary in conversation (criteria scored, count below threshold, journeys assessed, top concerns) -- the full detail is in the file
|
|
137
|
+
|
|
138
|
+
The file contains the COMPLETE review with all scores, journey assessments, and evidence. The conversation summary is just an acknowledgment -- downstream steps read from the file, not from conversation context.
|
|
139
|
+
|
|
140
|
+
## Instructions for the Skill (Facilitator)
|
|
141
|
+
|
|
142
|
+
When orchestrating expert interaction reviews:
|
|
143
|
+
|
|
144
|
+
1. Create the `prototypes/clickable/reviews/` directory if it does not exist before invoking any expert
|
|
145
|
+
2. Tell each expert agent the exact file path to write to AND the path to this template
|
|
146
|
+
3. Tell each expert the criteria list, scoring scale, threshold, and all reference documents (style brief, product concept, visual grounding assets with categories)
|
|
147
|
+
4. Provide all journey screenshots from the interaction testing step AND the interaction report
|
|
148
|
+
5. When invoking for cross-review (Phase 2), tell the expert to READ the other expert's file by providing the file path -- do not pass the file content through conversation
|
|
149
|
+
6. After each expert completes, read the review file (not the conversation summary) to extract scores for divergence calculation and synthesis
|
|
150
|
+
7. Extract per-criterion scores from the "Per-Criterion Scores" table in each review file
|
|
151
|
+
8. Extract per-journey assessments from the "Per-Journey Assessment" table to compare journey-level agreement
|
|
152
|
+
9. When synthesizing the debate report, read ALL review files from the current round -- never rely on conversation context
|