@tgoodington/intuition 11.3.0 → 11.4.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@tgoodington/intuition",
|
|
3
|
-
"version": "11.
|
|
3
|
+
"version": "11.4.0",
|
|
4
4
|
"description": "Domain-adaptive workflow system for Claude Code. Includes the Enuncia pipeline (discovery, compose, design, execute, verify) and the classic pipeline (prompt, outline, assemble, detail, build, test, implement).",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"claude-code",
|
|
@@ -1,9 +1,9 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: intuition-enuncia-verify
|
|
3
|
-
description: Integration and verification for code projects.
|
|
3
|
+
description: Integration and verification for code projects. Walks the user through every manual step until the app is online, then systematically tests every interaction surface from a UX perspective. Not satisfied until the user can access the landing page AND every button, link, and flow works as expected.
|
|
4
4
|
model: opus
|
|
5
|
-
tools: Read, Write, Edit, Glob, Grep, Task, AskUserQuestion, Bash, mcp__ide__getDiagnostics
|
|
6
|
-
allowed-tools: Read, Write, Edit, Glob, Grep, Task, Bash, mcp__ide__getDiagnostics
|
|
5
|
+
tools: Read, Write, Edit, Glob, Grep, Task, AskUserQuestion, Bash, Agent, WebFetch, mcp__ide__getDiagnostics
|
|
6
|
+
allowed-tools: Read, Write, Edit, Glob, Grep, Task, Bash, Agent, WebFetch, mcp__ide__getDiagnostics
|
|
7
7
|
---
|
|
8
8
|
|
|
9
9
|
# Verify Protocol
|
|
@@ -14,21 +14,26 @@ Deliver something to the user through an experience that places them as creative
|
|
|
14
14
|
|
|
15
15
|
## SKILL GOAL
|
|
16
16
|
|
|
17
|
-
|
|
17
|
+
Two jobs, done relentlessly:
|
|
18
18
|
|
|
19
|
-
|
|
19
|
+
1. **Get it online.** Wire the code in, figure out every prerequisite, walk the user through every manual step, and do not stop until the app is live and the user can access the landing page in their browser (or equivalent entry point). No "it compiles" — it must be RUNNING and REACHABLE.
|
|
20
|
+
|
|
21
|
+
2. **Prove every interaction works.** Systematically navigate the live application as a real user would. Click every button. Follow every link. Submit every form. Walk every flow. Verify from a UX perspective — not just "does the endpoint return 200" but "does the user see what they should see and can they do what they should be able to do." Not satisfied until every implemented interaction surface works as expected.
|
|
22
|
+
|
|
23
|
+
No mocks. No synthetic verification. The real system, used the way a real user uses it.
|
|
20
24
|
|
|
21
25
|
## CRITICAL RULES
|
|
22
26
|
|
|
23
27
|
1. You MUST read `.project-memory-state.json` and resolve context_path before anything else.
|
|
24
28
|
2. You MUST read `{context_path}/discovery_brief.md`, `{context_path}/tasks.json`, `{context_path}/build_output.json`, and `docs/project_notes/project_map.md`.
|
|
25
29
|
3. You MUST integrate before anything else. Code that isn't wired in can't run.
|
|
26
|
-
4. You MUST NOT
|
|
27
|
-
5. You MUST NOT
|
|
28
|
-
6. You MUST NOT
|
|
29
|
-
7. You MUST
|
|
30
|
-
8. You MUST
|
|
31
|
-
9. You MUST
|
|
30
|
+
4. You MUST NOT begin UX validation until the app is online and the user confirms they can access it.
|
|
31
|
+
5. You MUST NOT consider Phase 1 complete until the landing page (or primary entry point) is reachable and the user confirms it.
|
|
32
|
+
6. You MUST NOT consider Phase 2 complete until every implemented interaction surface has been tested from a UX perspective.
|
|
33
|
+
7. You MUST NOT fix failures that violate user decisions from the specs. Escalate immediately.
|
|
34
|
+
8. You MUST delegate integration tasks and code fixes to subagents. Do not write code yourself.
|
|
35
|
+
9. You MUST verify against the discovery brief after UX validation — does the system deliver the North Star?
|
|
36
|
+
10. You MUST update `docs/project_notes/project_map.md` if integration reveals new information.
|
|
32
37
|
|
|
33
38
|
## CONTEXT PATH RESOLUTION
|
|
34
39
|
|
|
@@ -45,24 +50,27 @@ No mocks. No "verified against synthetic data." Either it works or it doesn't.
|
|
|
45
50
|
## PROTOCOL
|
|
46
51
|
|
|
47
52
|
```
|
|
48
|
-
Phase 1: Get it
|
|
53
|
+
Phase 1: Get it online
|
|
49
54
|
Step 1: Read context
|
|
50
55
|
Step 2: Integration — wire everything together
|
|
51
56
|
Step 3: Toolchain — compile, type-check, lint
|
|
52
|
-
Step 4:
|
|
53
|
-
Step 5: Assisted setup —
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
Step 7:
|
|
58
|
-
Step 8:
|
|
59
|
-
Step 9:
|
|
60
|
-
Step 10:
|
|
57
|
+
Step 4: Prerequisites — what does the system need to actually start?
|
|
58
|
+
Step 5: Assisted setup — work through every manual step with the user
|
|
59
|
+
Step 6: Go live — start the app and verify it's reachable
|
|
60
|
+
|
|
61
|
+
Phase 2: UX validation
|
|
62
|
+
Step 7: Build the interaction map
|
|
63
|
+
Step 8: Systematic walkthrough — test every interaction surface
|
|
64
|
+
Step 9: Fix cycle
|
|
65
|
+
Step 10: Final verification against discovery brief
|
|
66
|
+
Step 11: Exit
|
|
61
67
|
```
|
|
62
68
|
|
|
63
69
|
---
|
|
64
70
|
|
|
65
|
-
## PHASE 1: GET IT
|
|
71
|
+
## PHASE 1: GET IT ONLINE
|
|
72
|
+
|
|
73
|
+
The only acceptable outcome of Phase 1 is: the app is running and the user can access the landing page (or primary entry point) in their browser or client.
|
|
66
74
|
|
|
67
75
|
### STEP 1: READ CONTEXT
|
|
68
76
|
|
|
@@ -74,7 +82,7 @@ Read these files:
|
|
|
74
82
|
|
|
75
83
|
From build_output.json, extract: all files created and modified, task statuses, any escalated issues or deviations.
|
|
76
84
|
|
|
77
|
-
From tasks.json, extract: experience slices (these become the basis for
|
|
85
|
+
From tasks.json, extract: experience slices (these become the basis for the interaction map in Phase 2).
|
|
78
86
|
|
|
79
87
|
#### Gate Check
|
|
80
88
|
|
|
@@ -129,9 +137,9 @@ Also run `mcp__ide__getDiagnostics` to catch IDE-visible issues.
|
|
|
129
137
|
|
|
130
138
|
If any step fails, classify and fix before proceeding.
|
|
131
139
|
|
|
132
|
-
### STEP 4:
|
|
140
|
+
### STEP 4: PREREQUISITES
|
|
133
141
|
|
|
134
|
-
|
|
142
|
+
Figure out everything the system needs to actually start and run — not just compile.
|
|
135
143
|
|
|
136
144
|
#### 4a. Research Prerequisites
|
|
137
145
|
|
|
@@ -153,41 +161,19 @@ For each dependency, report: what it is, where in the code it's referenced, whet
|
|
|
153
161
|
|
|
154
162
|
From the researcher's findings plus context from the discovery brief (which describes the deployment environment), build a concrete readiness checklist. Group items by category.
|
|
155
163
|
|
|
156
|
-
Format:
|
|
157
|
-
|
|
158
|
-
```
|
|
159
|
-
## Readiness Checklist
|
|
160
|
-
|
|
161
|
-
To get this system running, here's what needs to be set up:
|
|
162
|
-
|
|
163
|
-
### [Category: e.g., Database]
|
|
164
|
-
- [ ] [Specific action — e.g., "Create PostgreSQL database 'staff_coverage'"]
|
|
165
|
-
- [ ] [Next action — e.g., "Run migrations: alembic upgrade head"]
|
|
166
|
-
|
|
167
|
-
### [Category: e.g., External Services]
|
|
168
|
-
- [ ] [Specific action]
|
|
169
|
-
- I can help with: [what you can assist with — e.g., "generating the config file, writing the migration"]
|
|
170
|
-
- You'll need to: [what requires human action — e.g., "create the Azure AD app registration, grant admin consent"]
|
|
171
|
-
|
|
172
|
-
### [Category: e.g., Environment]
|
|
173
|
-
- [ ] [Specific action]
|
|
174
|
-
|
|
175
|
-
...
|
|
176
|
-
```
|
|
177
|
-
|
|
178
164
|
For each item, be specific about:
|
|
179
165
|
- **What** needs to happen (exact commands, exact config values where known)
|
|
180
166
|
- **Where** it's referenced in the code (so the user can verify)
|
|
181
|
-
- **What you can
|
|
167
|
+
- **What you can do** vs. **what requires their action** (admin portals, credentials, infrastructure access)
|
|
182
168
|
|
|
183
169
|
#### 4c. Present to User
|
|
184
170
|
|
|
185
171
|
Present the readiness checklist via AskUserQuestion:
|
|
186
172
|
|
|
187
173
|
```
|
|
188
|
-
Question: "[The readiness checklist
|
|
174
|
+
Question: "[The readiness checklist]
|
|
189
175
|
|
|
190
|
-
Let's work through these. Which would you like to tackle first, or is anything already set up?"
|
|
176
|
+
Let's work through these one at a time. Which would you like to tackle first, or is anything already set up?"
|
|
191
177
|
Header: "Getting It Running"
|
|
192
178
|
```
|
|
193
179
|
|
|
@@ -195,173 +181,214 @@ Header: "Getting It Running"
|
|
|
195
181
|
|
|
196
182
|
Work through the checklist with the user interactively. For each item:
|
|
197
183
|
|
|
198
|
-
- If you can do it (write config files, run migrations, generate boilerplate):
|
|
199
|
-
- If it requires their action (portal configuration, credential creation, infrastructure provisioning): give
|
|
200
|
-
- If it requires both
|
|
184
|
+
- **If you can do it** (write config files, run migrations, generate boilerplate, set up .env): do it and confirm.
|
|
185
|
+
- **If it requires their action** (portal configuration, credential creation, infrastructure provisioning): give exact step-by-step instructions and wait for confirmation.
|
|
186
|
+
- **If it requires both**: do your part, then tell them exactly what's left.
|
|
201
187
|
|
|
202
|
-
After each item is addressed, try to
|
|
188
|
+
After each item is addressed, try to verify it works:
|
|
203
189
|
- After database setup: try connecting and running a basic query
|
|
204
190
|
- After API credentials: try a test request to the service
|
|
205
191
|
- After environment config: try importing/starting the app
|
|
206
192
|
|
|
207
|
-
When something fails, diagnose and help fix it before moving on.
|
|
193
|
+
When something fails, diagnose and help fix it before moving on. Do NOT skip items and hope they work later.
|
|
194
|
+
|
|
195
|
+
### STEP 6: GO LIVE
|
|
196
|
+
|
|
197
|
+
This is the moment of truth. Start the application and verify it's actually reachable.
|
|
198
|
+
|
|
199
|
+
#### 6a. Start the Application
|
|
208
200
|
|
|
209
|
-
|
|
201
|
+
Run the start/dev command for the application. Monitor the output for errors.
|
|
210
202
|
|
|
211
|
-
|
|
203
|
+
If the app fails to start:
|
|
204
|
+
1. Read the error output carefully
|
|
205
|
+
2. Diagnose the root cause
|
|
206
|
+
3. Fix it (or help the user fix it if it requires their action)
|
|
207
|
+
4. Try again
|
|
208
|
+
5. Repeat until the app starts successfully
|
|
209
|
+
|
|
210
|
+
#### 6b. Verify Reachability
|
|
211
|
+
|
|
212
|
+
Once the app appears to be running:
|
|
213
|
+
|
|
214
|
+
1. **Hit the landing page** — use WebFetch or curl to request the primary URL (e.g., `http://localhost:3000`). Verify you get a real response, not an error page.
|
|
215
|
+
2. **Check for common startup issues** — port conflicts, missing environment variables that only matter at request time, lazy initialization failures.
|
|
216
|
+
3. **Ask the user to confirm** — present via AskUserQuestion:
|
|
212
217
|
|
|
213
218
|
```
|
|
214
|
-
Question: "
|
|
215
|
-
|
|
219
|
+
Question: "The app is running. Can you access it at [URL]? Can you see the landing page?
|
|
220
|
+
|
|
221
|
+
If something looks wrong, describe what you see and I'll help fix it."
|
|
222
|
+
Header: "Is It Online?"
|
|
216
223
|
Options:
|
|
217
|
-
- "
|
|
218
|
-
- "
|
|
224
|
+
- "Yes — I can see the landing page"
|
|
225
|
+
- "It loads but something is wrong"
|
|
226
|
+
- "I can't access it"
|
|
219
227
|
```
|
|
220
228
|
|
|
221
|
-
|
|
229
|
+
#### 6c. Resolve Until Online
|
|
230
|
+
|
|
231
|
+
If the user reports issues, work through them. Common problems:
|
|
232
|
+
- CORS issues (browser can reach it but API calls fail)
|
|
233
|
+
- Missing static assets (page loads but looks broken)
|
|
234
|
+
- Authentication redirects blocking access
|
|
235
|
+
- Database connection failures on first real request
|
|
236
|
+
- Missing seed data causing empty/error states
|
|
237
|
+
|
|
238
|
+
Do NOT proceed to Phase 2 until the user confirms they can access the landing page and it looks right. This is a hard gate. If it takes 10 rounds of fixing, so be it.
|
|
222
239
|
|
|
223
240
|
---
|
|
224
241
|
|
|
225
|
-
## PHASE 2:
|
|
242
|
+
## PHASE 2: UX VALIDATION
|
|
226
243
|
|
|
227
|
-
|
|
244
|
+
The app is online. Now systematically verify that every implemented interaction works from a real user's perspective.
|
|
228
245
|
|
|
229
|
-
|
|
246
|
+
This is NOT writing automated test files. This is YOU walking through the application as a user would — fetching pages, analyzing what's rendered, verifying links go where they should, checking that actions produce the expected results.
|
|
230
247
|
|
|
231
|
-
|
|
248
|
+
### STEP 7: BUILD THE INTERACTION MAP
|
|
232
249
|
|
|
233
|
-
|
|
234
|
-
- **Main entry points**: Do the primary routes/endpoints/commands return non-error responses?
|
|
235
|
-
- **Core dependencies**: Does the app actually talk to its database, APIs, etc.? (Verify with a request that exercises a real dependency path)
|
|
236
|
-
- **Happy path**: One simple request through the main flow — does it complete end-to-end?
|
|
250
|
+
Before testing, build a complete map of every interaction surface that was implemented.
|
|
237
251
|
|
|
238
|
-
####
|
|
252
|
+
#### 7a. Inventory from Specs
|
|
239
253
|
|
|
240
|
-
|
|
254
|
+
Read the experience slices from `tasks.json` and the discovery brief. For each slice, extract:
|
|
255
|
+
- **Pages/routes** the user visits
|
|
256
|
+
- **Actions** the user takes (buttons clicked, forms submitted, links followed)
|
|
257
|
+
- **Expected outcomes** (what the user should see after each action)
|
|
241
258
|
|
|
242
|
-
|
|
243
|
-
You are writing smoke tests against a LIVE, RUNNING system. The app is already up — you are testing it from the outside.
|
|
259
|
+
#### 7b. Inventory from Code
|
|
244
260
|
|
|
245
|
-
|
|
246
|
-
Test conventions: [naming, directory from existing tests]
|
|
247
|
-
App URL / entry point: [how to reach the running system]
|
|
261
|
+
Spawn an `intuition-researcher` agent:
|
|
248
262
|
|
|
249
|
-
|
|
250
|
-
-
|
|
251
|
-
-
|
|
252
|
-
-
|
|
253
|
-
-
|
|
263
|
+
"Analyze the codebase to build a complete interaction map. Find:
|
|
264
|
+
- Every route/page/screen defined in the app
|
|
265
|
+
- Every navigation link (where it appears, where it points)
|
|
266
|
+
- Every button and what it triggers
|
|
267
|
+
- Every form and what it submits to
|
|
268
|
+
- Every interactive element (dropdowns, modals, toggles, tabs, etc.)
|
|
269
|
+
- Every API endpoint that backs a UI interaction
|
|
270
|
+
|
|
271
|
+
Report as a structured list: [page/route] → [interaction element] → [expected behavior]"
|
|
272
|
+
|
|
273
|
+
#### 7c. Merge and Present
|
|
274
|
+
|
|
275
|
+
Merge the spec-based inventory with the code-based inventory into a single interaction map. Present to the user via AskUserQuestion:
|
|
254
276
|
|
|
255
|
-
Rules:
|
|
256
|
-
- The system is ALREADY RUNNING. Tests make real requests to it.
|
|
257
|
-
- NO mocks. NO in-memory databases. NO test servers. You hit the live app.
|
|
258
|
-
- If a test needs data to exist, create it through the app's own API first (setup), then clean it up after (teardown).
|
|
259
|
-
- Each test should take < 10 seconds.
|
|
260
|
-
- If a test fails, it means the live system is broken — not that a mock is misconfigured.
|
|
261
277
|
```
|
|
278
|
+
Question: "Here's every interaction surface I'll be testing:
|
|
262
279
|
|
|
263
|
-
|
|
280
|
+
[The interaction map — organized by page/route, listing every link, button, form, and interactive element]
|
|
264
281
|
|
|
265
|
-
|
|
282
|
+
Anything I should add or skip?"
|
|
283
|
+
Header: "Interaction Map"
|
|
284
|
+
```
|
|
266
285
|
|
|
267
|
-
|
|
286
|
+
### STEP 8: SYSTEMATIC WALKTHROUGH
|
|
268
287
|
|
|
269
|
-
|
|
288
|
+
Work through the interaction map methodically. For each page/route:
|
|
270
289
|
|
|
271
|
-
|
|
290
|
+
#### 8a. Load the Page
|
|
272
291
|
|
|
273
|
-
|
|
274
|
-
- **
|
|
275
|
-
- **
|
|
292
|
+
Use WebFetch to load the page. Analyze what comes back:
|
|
293
|
+
- **Does the page render?** (non-error HTTP status, meaningful HTML content)
|
|
294
|
+
- **Are key elements present?** (navigation, expected headings, expected content sections)
|
|
295
|
+
- **Are there broken references?** (missing images, broken CSS/JS links, 404 resources)
|
|
276
296
|
|
|
277
|
-
####
|
|
297
|
+
#### 8b. Test Every Link
|
|
278
298
|
|
|
279
|
-
|
|
299
|
+
For every navigation link on the page:
|
|
300
|
+
- Follow it (WebFetch the target URL)
|
|
301
|
+
- Verify it resolves to the correct destination (not a 404, not a wrong page)
|
|
302
|
+
- Verify the destination page renders correctly
|
|
280
303
|
|
|
281
|
-
|
|
282
|
-
|
|
283
|
-
|
|
284
|
-
|
|
285
|
-
|
|
286
|
-
|
|
287
|
-
|
|
288
|
-
|
|
289
|
-
|
|
290
|
-
|
|
291
|
-
|
|
292
|
-
|
|
293
|
-
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
## Rules
|
|
298
|
-
- The system is ALREADY RUNNING. Tests make real requests to it.
|
|
299
|
-
- NO mocks of any kind. The app, database, and services are all live.
|
|
300
|
-
- Test the journey from the stakeholder's perspective using real entry points (HTTP routes, CLI commands, public APIs).
|
|
301
|
-
- If a test needs data, create it through the app's API first (setup), clean up after (teardown).
|
|
302
|
-
- Assert against acceptance criteria from the spec, not implementation details.
|
|
303
|
-
- Each test should tell a story: "the admin does X, the system does Y, the result is Z"
|
|
304
|
-
- If a slice requires UI interaction you can't automate, test the API layer that backs it.
|
|
305
|
-
- Do NOT read source code to determine expected behavior — the spec defines what should happen.
|
|
306
|
-
|
|
307
|
-
## Spec Sources (read these for expected behavior)
|
|
308
|
-
- Discovery brief: {context_path}/discovery_brief.md
|
|
309
|
-
- Tasks: {context_path}/tasks.json
|
|
310
|
-
```
|
|
304
|
+
#### 8c. Test Every Button and Action
|
|
305
|
+
|
|
306
|
+
For every button and interactive element:
|
|
307
|
+
- Determine what it does (from the code analysis in Step 7)
|
|
308
|
+
- If it triggers an API call: make that API call with appropriate test data and verify the response
|
|
309
|
+
- If it submits a form: submit the form with valid test data and verify the result
|
|
310
|
+
- If it toggles UI state: verify the underlying mechanism works (e.g., the API endpoint that backs a toggle)
|
|
311
|
+
|
|
312
|
+
#### 8d. Test Every Form
|
|
313
|
+
|
|
314
|
+
For every form on the page:
|
|
315
|
+
- **Valid submission**: Submit with valid data. Verify success response, data persistence, and any expected side effects (emails, state changes, redirects).
|
|
316
|
+
- **Required fields**: Verify that submitting with missing required fields produces appropriate validation feedback.
|
|
317
|
+
- **Edge cases**: Test with boundary values if the spec defines constraints.
|
|
318
|
+
|
|
319
|
+
#### 8e. Test User Flows End-to-End
|
|
311
320
|
|
|
312
|
-
|
|
321
|
+
For each experience slice, walk through the complete user journey:
|
|
322
|
+
1. Start where the user starts
|
|
323
|
+
2. Navigate as the user would (following links, not jumping directly to URLs)
|
|
324
|
+
3. Perform each action in the flow
|
|
325
|
+
4. Verify each intermediate state
|
|
326
|
+
5. Confirm the final outcome matches the acceptance criteria
|
|
313
327
|
|
|
314
|
-
|
|
328
|
+
#### 8f. Report Progress
|
|
315
329
|
|
|
316
|
-
|
|
330
|
+
After completing each page/route, briefly report status: what passed, what failed, what needs attention. Group issues for the fix cycle rather than interrupting the walkthrough for each problem (unless something is blocking further testing).
|
|
331
|
+
|
|
332
|
+
### STEP 9: FIX CYCLE
|
|
333
|
+
|
|
334
|
+
After the walkthrough, address every issue found.
|
|
335
|
+
|
|
336
|
+
#### Issue Classification
|
|
317
337
|
|
|
318
338
|
| Classification | Action |
|
|
319
339
|
|---|---|
|
|
320
|
-
| **
|
|
321
|
-
| **
|
|
322
|
-
| **
|
|
323
|
-
| **
|
|
340
|
+
| **Broken link** (404, wrong destination) | Fix via `intuition-code-writer` |
|
|
341
|
+
| **Non-functional button** (click does nothing, wrong API call) | Fix via `intuition-code-writer` |
|
|
342
|
+
| **Form submission failure** (validation error on valid data, wrong endpoint, missing handler) | Fix via `intuition-code-writer` |
|
|
343
|
+
| **Missing page/route** (implemented in code but not accessible) | Fix via `intuition-code-writer` — likely a routing issue |
|
|
344
|
+
| **Missing content** (page loads but expected elements are absent) | Fix via `intuition-code-writer` |
|
|
345
|
+
| **Broken user flow** (individual steps work but the end-to-end journey breaks) | Diagnose where the flow breaks, fix the connection point |
|
|
346
|
+
| **Visual/layout issue** (content renders but is clearly broken — overlapping elements, invisible text, unusable layout) | Fix via `intuition-code-writer` |
|
|
347
|
+
| **Data issue** (correct behavior but empty/wrong data shown) | Check seeds, migrations, API responses — fix the data pipeline |
|
|
324
348
|
| **Environment/config issue** (service not reachable, credentials wrong) | Help user diagnose and fix |
|
|
325
|
-
| **Spec violation** (
|
|
326
|
-
| **Test regression** (existing test broke) | Diagnose: is the test outdated or the new code wrong? Escalate if ambiguous |
|
|
349
|
+
| **Spec violation** (interaction works but does the wrong thing per spec) | Escalate: "Spec says X, but the app does Y" |
|
|
327
350
|
| **Violates user decision** | STOP — escalate immediately |
|
|
328
351
|
|
|
329
352
|
#### Fix Process
|
|
330
353
|
|
|
331
|
-
1.
|
|
332
|
-
|
|
333
|
-
|
|
334
|
-
|
|
335
|
-
|
|
336
|
-
|
|
354
|
+
1. Present ALL found issues to the user, grouped by severity:
|
|
355
|
+
- **Blocking**: User flows that don't work at all
|
|
356
|
+
- **Broken**: Individual interactions that fail
|
|
357
|
+
- **Degraded**: Things that work but poorly (wrong content, bad layout, missing feedback)
|
|
358
|
+
2. Fix blocking issues first, then broken, then degraded
|
|
359
|
+
3. For each fix: delegate to `intuition-code-writer`, then re-test the specific interaction on the live system
|
|
360
|
+
4. Max 3 fix attempts per issue — then escalate to user
|
|
361
|
+
5. After all fixes: **re-run the full walkthrough** on affected pages to verify fixes didn't break other interactions
|
|
362
|
+
6. Repeat until clean or all remaining issues are escalated
|
|
337
363
|
|
|
338
|
-
### STEP
|
|
364
|
+
### STEP 10: FINAL VERIFICATION
|
|
339
365
|
|
|
340
|
-
After
|
|
366
|
+
After the walkthrough is clean (all interactions work):
|
|
341
367
|
|
|
342
|
-
**North Star check**: Walk through the brief's North Star statement. For each stakeholder:
|
|
368
|
+
**North Star check**: Walk through the discovery brief's North Star statement. For each stakeholder:
|
|
343
369
|
- Can they do what the brief says they should be able to do — on the live system?
|
|
344
370
|
- Does the system honor the constraints?
|
|
345
371
|
- Would this satisfy the North Star as written?
|
|
346
372
|
|
|
347
|
-
If something drifts, flag it: "
|
|
373
|
+
If something drifts, flag it: "All interactions work, but [specific concern about North Star alignment]."
|
|
348
374
|
|
|
349
375
|
**Update `docs/project_notes/project_map.md`** if integration or testing revealed anything new.
|
|
350
376
|
|
|
351
|
-
### STEP
|
|
377
|
+
### STEP 11: EXIT
|
|
352
378
|
|
|
353
|
-
**Update state.** Read `.project-memory-state.json`. Target active context. Set: `status` → `"complete"`, `workflow.verify.completed` → `true`, `workflow.verify.completed_at` → current ISO timestamp. Set on root: `last_handoff` → current ISO timestamp, `last_handoff_transition` → `"verify_to_complete"`.
|
|
379
|
+
**Update state.** Read `.project-memory-state.json`. Target active context. Set: `status` → `"complete"`, `workflow.verify.completed` → `true`, `workflow.verify.completed_at` → current ISO timestamp. Set on root: `last_handoff` → current ISO timestamp, `last_handoff_transition` → `"verify_to_complete"`. Write back.
|
|
354
380
|
|
|
355
381
|
**Present results** via AskUserQuestion:
|
|
356
382
|
|
|
357
383
|
```
|
|
358
|
-
Question: "Verification complete — tested against the live system.
|
|
359
|
-
|
|
360
|
-
**
|
|
361
|
-
**
|
|
362
|
-
**
|
|
363
|
-
**
|
|
364
|
-
**
|
|
384
|
+
Question: "Verification complete — every interaction tested against the live system.
|
|
385
|
+
|
|
386
|
+
**Online**: [URL — confirmed accessible]
|
|
387
|
+
**Pages tested**: [N pages/routes]
|
|
388
|
+
**Links verified**: [N links — N working, N fixed, N escalated]
|
|
389
|
+
**Buttons/actions verified**: [N — N working, N fixed, N escalated]
|
|
390
|
+
**Forms verified**: [N — N working, N fixed, N escalated]
|
|
391
|
+
**User flows verified**: [N experience slices — N working, N fixed, N escalated]
|
|
365
392
|
**North Star alignment**: [met / concerns]
|
|
366
393
|
|
|
367
394
|
[If escalated issues exist, list them]
|
|
@@ -375,7 +402,7 @@ Options:
|
|
|
375
402
|
- "Done — no commit"
|
|
376
403
|
```
|
|
377
404
|
|
|
378
|
-
If committing: stage files from build output + integration changes +
|
|
405
|
+
If committing: stage files from build output + integration changes + fixes, commit with descriptive message, optionally push.
|
|
379
406
|
|
|
380
407
|
**Route.** "Workflow complete. Run `/clear` then `/intuition-enuncia-start` to see project status."
|
|
381
408
|
|
|
@@ -388,14 +415,14 @@ When verifying on a branch:
|
|
|
388
415
|
|
|
389
416
|
## RESUME LOGIC
|
|
390
417
|
|
|
391
|
-
1. If Phase 1 completed (
|
|
392
|
-
2. If
|
|
418
|
+
1. If Phase 1 completed (app confirmed online) but UX walkthrough hasn't started: skip to Step 7.
|
|
419
|
+
2. If interaction map exists but walkthrough incomplete: "Found interaction map from a previous session. Resuming walkthrough."
|
|
393
420
|
3. Otherwise fresh start from Step 1.
|
|
394
421
|
|
|
395
422
|
## VOICE
|
|
396
423
|
|
|
397
|
-
- **
|
|
398
|
-
- **
|
|
399
|
-
- **
|
|
400
|
-
- **
|
|
424
|
+
- **Relentless** — not satisfied until the app is online AND every interaction works
|
|
425
|
+
- **User-perspective** — think like the person clicking, not the person who wrote the code
|
|
426
|
+
- **Evidence-driven** — "I clicked X, expected Y, got Z" for every issue
|
|
427
|
+
- **Pragmatic** — fix what's broken, escalate what's beyond scope, report clearly
|
|
401
428
|
- **Brief-anchored** — the discovery foundation is the ultimate measure of success
|