get-shit-done-cc 1.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (49) hide show
  1. package/LICENSE +21 -0
  2. package/README.md +192 -0
  3. package/bin/install.js +53 -0
  4. package/commands/gsd/add-phase.md +201 -0
  5. package/commands/gsd/complete-milestone.md +105 -0
  6. package/commands/gsd/discuss-milestone.md +45 -0
  7. package/commands/gsd/discuss-phase.md +47 -0
  8. package/commands/gsd/execute-plan.md +108 -0
  9. package/commands/gsd/help.md +252 -0
  10. package/commands/gsd/insert-phase.md +218 -0
  11. package/commands/gsd/list-phase-assumptions.md +49 -0
  12. package/commands/gsd/new-milestone.md +58 -0
  13. package/commands/gsd/new-project.md +177 -0
  14. package/commands/gsd/pause-work.md +123 -0
  15. package/commands/gsd/plan-phase.md +60 -0
  16. package/commands/gsd/progress.md +182 -0
  17. package/commands/gsd/resume-work.md +50 -0
  18. package/get-shit-done/references/checkpoints.md +594 -0
  19. package/get-shit-done/references/cli-automation.md +527 -0
  20. package/get-shit-done/references/git-integration.md +126 -0
  21. package/get-shit-done/references/plan-format.md +397 -0
  22. package/get-shit-done/references/principles.md +97 -0
  23. package/get-shit-done/references/questioning.md +138 -0
  24. package/get-shit-done/references/research-pitfalls.md +215 -0
  25. package/get-shit-done/references/scope-estimation.md +451 -0
  26. package/get-shit-done/templates/config.json +17 -0
  27. package/get-shit-done/templates/context.md +385 -0
  28. package/get-shit-done/templates/continue-here.md +78 -0
  29. package/get-shit-done/templates/issues.md +32 -0
  30. package/get-shit-done/templates/milestone-archive.md +123 -0
  31. package/get-shit-done/templates/milestone.md +115 -0
  32. package/get-shit-done/templates/phase-prompt.md +290 -0
  33. package/get-shit-done/templates/project.md +207 -0
  34. package/get-shit-done/templates/research-prompt.md +133 -0
  35. package/get-shit-done/templates/roadmap.md +196 -0
  36. package/get-shit-done/templates/state.md +226 -0
  37. package/get-shit-done/templates/summary.md +200 -0
  38. package/get-shit-done/workflows/complete-milestone.md +490 -0
  39. package/get-shit-done/workflows/create-milestone.md +379 -0
  40. package/get-shit-done/workflows/create-roadmap.md +443 -0
  41. package/get-shit-done/workflows/discuss-milestone.md +144 -0
  42. package/get-shit-done/workflows/discuss-phase.md +254 -0
  43. package/get-shit-done/workflows/execute-phase.md +1261 -0
  44. package/get-shit-done/workflows/list-phase-assumptions.md +178 -0
  45. package/get-shit-done/workflows/plan-phase.md +783 -0
  46. package/get-shit-done/workflows/research-phase.md +293 -0
  47. package/get-shit-done/workflows/resume-project.md +248 -0
  48. package/get-shit-done/workflows/transition.md +488 -0
  49. package/package.json +30 -0
@@ -0,0 +1,215 @@
1
+ <research_pitfalls>
2
+
3
+ <purpose>
4
+ This document catalogs research mistakes discovered in production use, providing specific patterns to avoid and verification strategies to prevent recurrence.
5
+ </purpose>
6
+
7
+ <known_pitfalls>
8
+
9
+ <pitfall_config_scope>
10
+ **What**: Assuming global configuration means no project-scoping exists
11
+ **Example**: Concluding "MCP servers are configured GLOBALLY only" while missing project-scoped `.mcp.json`
12
+ **Why it happens**: Not explicitly checking all known configuration patterns
13
+ **Prevention**:
14
+ ```xml
15
+ <verification_checklist>
16
+ **CRITICAL**: Verify ALL configuration scopes:
17
+ □ User/global scope - System-wide configuration
18
+ □ Project scope - Project-level configuration files
19
+ □ Local scope - Project-specific user overrides
20
+ □ Workspace scope - IDE/tool workspace settings
21
+ □ Environment scope - Environment variables
22
+ </verification_checklist>
23
+ ```
24
+ </pitfall_config_scope>
25
+
26
+ <pitfall_search_vagueness>
27
+ **What**: Asking researchers to "search for documentation" without specifying where
28
+ **Example**: "Research MCP documentation" → finds outdated community blog instead of official docs
29
+ **Why it happens**: Vague research instructions don't specify exact sources
30
+ **Prevention**:
31
+ ```xml
32
+ <sources>
33
+ Official sources (use WebFetch):
34
+ - https://exact-url-to-official-docs
35
+ - https://exact-url-to-api-reference
36
+
37
+ Search queries (use WebSearch):
38
+ - "specific search query {current_year}"
39
+ - "another specific query {current_year}"
40
+ </sources>
41
+ ```
42
+ </pitfall_search_vagueness>
43
+
44
+ <pitfall_deprecated_features>
45
+ **What**: Finding archived/old documentation and concluding feature doesn't exist
46
+ **Example**: Finding 2022 docs saying "feature not supported" when current version added it
47
+ **Why it happens**: Not checking multiple sources or recent updates
48
+ **Prevention**:
49
+ ```xml
50
+ <verification_checklist>
51
+ □ Check current official documentation
52
+ □ Review changelog/release notes for recent updates
53
+ □ Verify version numbers and publication dates
54
+ □ Cross-reference multiple authoritative sources
55
+ </verification_checklist>
56
+ ```
57
+ </pitfall_deprecated_features>
58
+
59
+ <pitfall_tool_variations>
60
+ **What**: Conflating capabilities across different tools/environments
61
+ **Example**: "Claude Desktop supports X" ≠ "Claude Code supports X"
62
+ **Why it happens**: Not explicitly checking each environment separately
63
+ **Prevention**:
64
+ ```xml
65
+ <verification_checklist>
66
+ □ Claude Desktop capabilities
67
+ □ Claude Code capabilities
68
+ □ VS Code extension capabilities
69
+ □ API/SDK capabilities
70
+ Document which environment supports which features
71
+ </verification_checklist>
72
+ ```
73
+ </pitfall_tool_variations>
74
+
75
+ <pitfall_negative_claims>
76
+ **What**: Making definitive "X is not possible" statements without official source verification
77
+ **Example**: "Folder-scoped MCP configuration is not supported" (missing `.mcp.json`)
78
+ **Why it happens**: Drawing conclusions from absence of evidence rather than evidence of absence
79
+ **Prevention**:
80
+ ```xml
81
+ <critical_claims_audit>
82
+ For any "X is not possible" or "Y is the only way" statement:
83
+ - [ ] Is this verified by official documentation stating it explicitly?
84
+ - [ ] Have I checked for recent updates that might change this?
85
+ - [ ] Have I verified all possible approaches/mechanisms?
86
+ - [ ] Am I confusing "I didn't find it" with "it doesn't exist"?
87
+ </critical_claims_audit>
88
+ ```
89
+ </pitfall_negative_claims>
90
+
91
+ <pitfall_missing_enumeration>
92
+ **What**: Investigating open-ended scope without enumerating known possibilities first
93
+ **Example**: "Research configuration options" instead of listing specific options to verify
94
+ **Why it happens**: Not creating explicit checklist of items to investigate
95
+ **Prevention**:
96
+ ```xml
97
+ <verification_checklist>
98
+ Enumerate ALL known options FIRST:
99
+ □ Option 1: [specific item]
100
+ □ Option 2: [specific item]
101
+ □ Option 3: [specific item]
102
+ □ Check for additional unlisted options
103
+
104
+ For each option above, document:
105
+ - Existence (confirmed/not found/unclear)
106
+ - Official source URL
107
+ - Current status (active/deprecated/beta)
108
+ </verification_checklist>
109
+ ```
110
+ </pitfall_missing_enumeration>
111
+
112
+ <pitfall_single_source>
113
+ **What**: Relying on a single source for critical claims
114
+ **Example**: Using only Stack Overflow answer from 2021 for current best practices
115
+ **Why it happens**: Not cross-referencing multiple authoritative sources
116
+ **Prevention**:
117
+ ```xml
118
+ <source_verification>
119
+ For critical claims, require multiple sources:
120
+ - [ ] Official documentation (primary)
121
+ - [ ] Release notes/changelog (for currency)
122
+ - [ ] Additional authoritative source (for verification)
123
+ - [ ] Contradiction check (ensure sources agree)
124
+ </source_verification>
125
+ ```
126
+ </pitfall_single_source>
127
+
128
+ <pitfall_assumed_completeness>
129
+ **What**: Assuming search results are complete and authoritative
130
+ **Example**: First Google result is outdated but assumed current
131
+ **Why it happens**: Not verifying publication dates and source authority
132
+ **Prevention**:
133
+ ```xml
134
+ <source_verification>
135
+ For each source consulted:
136
+ - [ ] Publication/update date verified (prefer recent/current)
137
+ - [ ] Source authority confirmed (official docs, not blogs)
138
+ - [ ] Version relevance checked (matches current version)
139
+ - [ ] Multiple search queries tried (not just one)
140
+ </source_verification>
141
+ ```
142
+ </pitfall_assumed_completeness>
143
+ </known_pitfalls>
144
+
145
+ <red_flags>
146
+
147
+ <red_flag_zero_not_found>
148
+ **Warning**: Every investigation succeeds perfectly
149
+ **Problem**: Real research encounters dead ends, ambiguity, and unknowns
150
+ **Action**: Expect honest reporting of limitations, contradictions, and gaps
151
+ </red_flag_zero_not_found>
152
+
153
+ <red_flag_no_confidence>
154
+ **Warning**: All findings presented as equally certain
155
+ **Problem**: Can't distinguish verified facts from educated guesses
156
+ **Action**: Require confidence levels (High/Medium/Low) for key findings
157
+ </red_flag_no_confidence>
158
+
159
+ <red_flag_missing_urls>
160
+ **Warning**: "According to documentation..." without specific URL
161
+ **Problem**: Can't verify claims or check for updates
162
+ **Action**: Require actual URLs for all official documentation claims
163
+ </red_flag_missing_urls>
164
+
165
+ <red_flag_no_evidence>
166
+ **Warning**: "X cannot do Y" or "Z is the only way" without citation
167
+ **Problem**: Strong claims require strong evidence
168
+ **Action**: Flag for verification against official sources
169
+ </red_flag_no_evidence>
170
+
171
+ <red_flag_incomplete_enum>
172
+ **Warning**: Verification checklist lists 4 items, output covers 2
173
+ **Problem**: Systematic gaps in coverage
174
+ **Action**: Ensure all enumerated items addressed or marked "not found"
175
+ </red_flag_incomplete_enum>
176
+ </red_flags>
177
+
178
+ <continuous_improvement>
179
+
180
+ When research gaps occur:
181
+
182
+ 1. **Document the gap**
183
+ - What was missed or incorrect?
184
+ - What was the actual correct information?
185
+ - What was the impact?
186
+
187
+ 2. **Root cause analysis**
188
+ - Why wasn't it caught?
189
+ - Which verification step would have prevented it?
190
+ - What pattern does this reveal?
191
+
192
+ 3. **Update this document**
193
+ - Add new pitfall entry
194
+ - Update relevant checklists
195
+ - Share lesson learned
196
+ </continuous_improvement>
197
+
198
+ <quick_reference>
199
+
200
+ Before submitting research, verify:
201
+
202
+ - [ ] All enumerated items investigated (not just some)
203
+ - [ ] Negative claims verified with official docs
204
+ - [ ] Multiple sources cross-referenced for critical claims
205
+ - [ ] URLs provided for all official documentation
206
+ - [ ] Publication dates checked (prefer recent/current)
207
+ - [ ] Tool/environment-specific variations documented
208
+ - [ ] Confidence levels assigned honestly
209
+ - [ ] Assumptions distinguished from verified facts
210
+ - [ ] "What might I have missed?" review completed
211
+
212
+ **Living Document**: Update after each significant research gap
213
+ **Lessons From**: MCP configuration research gap (missed `.mcp.json`)
214
+ </quick_reference>
215
+ </research_pitfalls>
@@ -0,0 +1,451 @@
1
+ <scope_estimation>
2
+ Plans must maintain consistent quality from first task to last. This requires understanding the **quality degradation curve** and splitting aggressively to stay in the peak quality zone.
3
+
4
+ <quality_degradation_curve>
5
+
6
+ **Critical insight:** Claude doesn't degrade at arbitrary percentages - it degrades when it *perceives* context pressure and enters "completion mode."
7
+
8
+ ```
9
+ Context Usage │ Quality Level │ Claude's Mental State
10
+ ─────────────────────────────────────────────────────────
11
+ 0-30% │ ████████ PEAK │ "I can be thorough and comprehensive"
12
+ │ │ No anxiety, full detail, best work
13
+
14
+ 30-50% │ ██████ GOOD │ "Still have room, maintaining quality"
15
+ │ │ Engaged, confident, solid work
16
+
17
+ 50-70% │ ███ DEGRADING │ "Getting tight, need to be efficient"
18
+ │ │ Efficiency mode, compression begins
19
+
20
+ 70%+ │ █ POOR │ "Running out, must finish quickly"
21
+ │ │ Self-lobotomization, rushed, minimal
22
+ ```
23
+
24
+ **The 40-50% inflection point:**
25
+
26
+ This is where quality breaks. Claude sees context mounting and thinks "I'd better conserve now or I won't finish." Result: The classic mid-execution statement "I'll complete the remaining tasks more concisely" = quality crash.
27
+
28
+ **The fundamental rule:** Stop BEFORE quality degrades, not at context limit.
29
+ </quality_degradation_curve>
30
+
31
+ <context_target>
32
+
33
+ **Plans should complete within ~50% of context usage.**
34
+
35
+ Why 50% not 80%?
36
+ - Huge safety buffer
37
+ - No context anxiety possible
38
+ - Quality maintained from start to finish
39
+ - Room for unexpected complexity
40
+ - Space for iteration and fixes
41
+
42
+ **If you target 80%, you're planning for failure.** By the time you hit 80%, you've already spent 40% in degradation mode.
43
+ </context_target>
44
+
45
+ <task_rule>
46
+
47
+ **Each plan should contain 2-3 tasks maximum. Context usage matters more than task count.**
48
+
49
+ **The real measure: Stay under 50% context usage.**
50
+
51
+ Task count is a proxy for context. Adjust based on task complexity:
52
+
53
+ **Simple tasks (CRUD, config, basic features):**
54
+ - 3 tasks is fine
55
+ - Each burns ~10-15% context
56
+ - Total: ~30-45% → Safe
57
+
58
+ **Complex tasks (auth, payments, architecture, integrations):**
59
+ - Stick to 2 tasks
60
+ - Each burns ~20-30% context
61
+ - Total: ~40-50% → At limit
62
+
63
+ **Very complex tasks (migrations, major refactors, novel patterns):**
64
+ - Consider 1-2 tasks only
65
+ - Each can burn 30-40% context
66
+ - Splitting to 1 task/plan is valid for high complexity
67
+
68
+ **Context estimation by task type:**
69
+
70
+ **Task 1 (0-15% context for simple, 0-30% for complex):**
71
+ - Fresh context
72
+ - Peak quality
73
+ - Comprehensive implementation
74
+ - Full testing
75
+
76
+ **Task 2 (15-35% context for simple, 30-50% for complex):**
77
+ - Still good quality
78
+ - Context pressure manageable
79
+ - Natural stopping point for complex work
80
+
81
+ **Task 3 (35-50% context for simple only):**
82
+ - Only include for simple tasks
83
+ - Skip for complex work
84
+ - Better to split complex work at 2 tasks
85
+
86
+ **Task 4+ (50%+ context):**
87
+ - NEVER do this
88
+ - Quality guaranteed to degrade
89
+ - Should have split earlier
90
+
91
+ **When in doubt: Default to 2 tasks.** Better to have an extra plan than degraded quality.
92
+
93
+ **The principle:** Each plan completes within 50% context. Task count is flexible based on complexity.
94
+ </task_rule>
95
+
96
+ <split_signals>
97
+
98
+ <always_split>
99
+
100
+ **1. More than 3 tasks**
101
+ - Even if tasks seem small
102
+ - Each additional task increases degradation risk
103
+ - Split into logical groups of 2-3
104
+
105
+ **2. Multiple subsystems**
106
+ ```
107
+ ❌ Bad (1 plan):
108
+ - Database schema (3 files)
109
+ - API routes (5 files)
110
+ - UI components (8 files)
111
+ Total: 16 files, 1 plan → guaranteed degradation
112
+
113
+ ✅ Good (3 plans):
114
+ - 01-01-PLAN.md: Database schema (3 files, 2 tasks)
115
+ - 01-02-PLAN.md: API routes (5 files, 3 tasks)
116
+ - 01-03-PLAN.md: UI components (8 files, 3 tasks)
117
+ Total: 16 files, 3 plans → consistent quality
118
+ ```
119
+
120
+ **3. Any task with >5 file modifications**
121
+ - Large tasks burn context fast
122
+ - Split by file groups or logical units
123
+ - Better: 3 plans of 2 files each vs 1 plan of 6 files
124
+
125
+ **4. Checkpoint + implementation work**
126
+ - Checkpoints require user interaction (context preserved)
127
+ - Implementation after checkpoint should be separate plan
128
+
129
+ ✅ Good split:
130
+ - 02-01-PLAN.md: Setup (checkpoint: decision on auth provider)
131
+ - 02-02-PLAN.md: Implement chosen auth solution
132
+
133
+ **5. Research + implementation**
134
+ - Research produces FINDINGS.md (separate plan)
135
+ - Implementation consumes FINDINGS.md (separate plan)
136
+ - Clear boundary, clean handoff
137
+ </always_split>
138
+
139
+ <consider_splitting>
140
+
141
+ **1. Estimated >5 files modified total**
142
+ - Context from reading existing code
143
+ - Context from diffs
144
+ - Context from responses
145
+ - Adds up faster than expected
146
+
147
+ **2. Complex domains (auth, payments, data modeling)**
148
+ - These require careful thinking
149
+ - Burns more context per task than simple CRUD
150
+ - Split more aggressively
151
+
152
+ **3. Any uncertainty about approach**
153
+ - "Figure out X" phase separate from "implement X" phase
154
+ - Don't mix exploration and implementation
155
+
156
+ **4. Natural semantic boundaries**
157
+ - Setup → Core → Features
158
+ - Backend → Frontend
159
+ - Configuration → Implementation → Testing
160
+ </consider_splitting>
161
+ </split_signals>
162
+
163
+ <splitting_strategies>
164
+
165
+ <by_subsystem>
166
+
167
+ **Phase:** "Authentication System"
168
+
169
+ **Split:**
170
+ ```
171
+ - 03-01-PLAN.md: Database models (User, Session tables + relations)
172
+ - 03-02-PLAN.md: Auth API (register, login, logout endpoints)
173
+ - 03-03-PLAN.md: Protected routes (middleware, JWT validation)
174
+ - 03-04-PLAN.md: UI components (login form, registration form)
175
+ ```
176
+
177
+ Each plan: 2-3 tasks, single subsystem, clean commits.
178
+ </by_subsystem>
179
+
180
+ <by_dependency>
181
+
182
+ **Phase:** "Payment Integration"
183
+
184
+ **Split:**
185
+ ```
186
+ - 04-01-PLAN.md: Stripe setup (webhook endpoints via API, env vars, test mode)
187
+ - 04-02-PLAN.md: Subscription logic (plans, checkout, customer portal)
188
+ - 04-03-PLAN.md: Frontend integration (pricing page, payment flow)
189
+ ```
190
+
191
+ Later plans depend on earlier completion. Sequential execution, fresh context each time.
192
+ </by_dependency>
193
+
194
+ <by_complexity>
195
+
196
+ **Phase:** "Dashboard Buildout"
197
+
198
+ **Split:**
199
+ ```
200
+ - 05-01-PLAN.md: Layout shell (simple: sidebar, header, routing)
201
+ - 05-02-PLAN.md: Data fetching (moderate: TanStack Query setup, API integration)
202
+ - 05-03-PLAN.md: Data visualization (complex: charts, tables, real-time updates)
203
+ ```
204
+
205
+ Complex work gets its own plan with full context budget.
206
+ </by_complexity>
207
+
208
+ <by_verification_points>
209
+
210
+ **Phase:** "Deployment Pipeline"
211
+
212
+ **Split:**
213
+ ```
214
+ - 06-01-PLAN.md: Vercel setup (deploy via CLI, configure domains)
215
+ → Ends with checkpoint:human-verify "check xyz.vercel.app loads"
216
+
217
+ - 06-02-PLAN.md: Environment config (secrets via CLI, env vars)
218
+ → Autonomous (no checkpoints) → subagent execution
219
+
220
+ - 06-03-PLAN.md: CI/CD (GitHub Actions, preview deploys)
221
+ → Ends with checkpoint:human-verify "check PR preview works"
222
+ ```
223
+
224
+ Verification checkpoints create natural boundaries. Autonomous plans between checkpoints execute via subagent with fresh context.
225
+ </by_verification_points>
226
+ </splitting_strategies>
227
+
228
+ <autonomous_vs_interactive>
229
+
230
+ **Critical optimization:** Plans without checkpoints don't need main context.
231
+
232
+ <autonomous_plans>
233
+ - Contains only `type="auto"` tasks
234
+ - No user interaction needed
235
+ - **Execute via subagent with fresh 200k context**
236
+ - Impossible to degrade (always starts at 0%)
237
+ - Creates SUMMARY, commits, reports back
238
+ - Can run in parallel (multiple subagents)
239
+ </autonomous_plans>
240
+
241
+ <interactive_plans>
242
+ - Contains `checkpoint:human-verify` or `checkpoint:decision` tasks
243
+ - Requires user interaction
244
+ - Must execute in main context
245
+ - Still target 50% context (2-3 tasks)
246
+
247
+ **Planning guidance:** If splitting a phase, try to:
248
+ - Group autonomous work together (→ subagent)
249
+ - Separate interactive work (→ main context)
250
+ - Maximize autonomous plans (more fresh contexts)
251
+
252
+ Example:
253
+ ```
254
+ Phase: Feature X
255
+ - 07-01-PLAN.md: Backend (autonomous) → subagent
256
+ - 07-02-PLAN.md: Frontend (autonomous) → subagent
257
+ - 07-03-PLAN.md: Integration test (has checkpoint:human-verify) → main context
258
+ ```
259
+
260
+ Two fresh contexts, one interactive verification. Perfect.
261
+ </interactive_plans>
262
+ </autonomous_vs_interactive>
263
+
264
+ <anti_patterns>
265
+
266
+ <antipattern_comprehensive>
267
+
268
+ ```
269
+ Plan: "Complete Authentication System"
270
+ Tasks:
271
+ 1. Database models
272
+ 2. Migration files
273
+ 3. Auth API endpoints
274
+ 4. JWT utilities
275
+ 5. Protected route middleware
276
+ 6. Password hashing
277
+ 7. Login form component
278
+ 8. Registration form component
279
+
280
+ Result: 8 tasks, 80%+ context, degradation at task 4-5
281
+ ```
282
+
283
+ **Why this fails:**
284
+ - Task 1-3: Good quality
285
+ - Task 4-5: "I'll do these concisely" = degradation begins
286
+ - Task 6-8: Rushed, minimal, poor quality
287
+ </antipattern_comprehensive>
288
+
289
+ <pattern_atomic>
290
+
291
+ ```
292
+ Split into 4 plans:
293
+
294
+ Plan 1: "Auth Database Models" (2 tasks)
295
+ - Database schema (User, Session)
296
+ - Migration files
297
+
298
+ Plan 2: "Auth API Core" (3 tasks)
299
+ - Register endpoint
300
+ - Login endpoint
301
+ - JWT utilities
302
+
303
+ Plan 3: "Auth API Protection" (2 tasks)
304
+ - Protected route middleware
305
+ - Logout endpoint
306
+
307
+ Plan 4: "Auth UI Components" (2 tasks)
308
+ - Login form
309
+ - Registration form
310
+ ```
311
+
312
+ **Why this succeeds:**
313
+ - Each plan: 2-3 tasks, 30-40% context
314
+ - All tasks: Peak quality throughout
315
+ - Git history: 4 focused commits
316
+ - Easy to verify each piece
317
+ - Rollback is surgical
318
+ </pattern_atomic>
319
+
320
+ <antipattern_efficiency_trap>
321
+
322
+ ```
323
+ Thinking: "These tasks are small, let's do 6 to be efficient"
324
+
325
+ Result: Task 1-2 are good, task 3-4 begin degrading, task 5-6 are rushed
326
+ ```
327
+
328
+ **Why this fails:** You're optimizing for fewer plans, not quality. The "efficiency" is false - poor quality requires more rework.
329
+ </antipattern_efficiency_trap>
330
+
331
+ <pattern_quality_first>
332
+
333
+ ```
334
+ Thinking: "These tasks are small, but let's do 2-3 to guarantee quality"
335
+
336
+ Result: All tasks peak quality, clean commits, no rework needed
337
+ ```
338
+
339
+ **Why this succeeds:** You optimize for quality, which is true efficiency. No rework = faster overall.
340
+ </pattern_quality_first>
341
+ </anti_patterns>
342
+
343
+ <estimating_context>
344
+
345
+ **Rough heuristics for plan size:**
346
+
347
+ <file_counts>
348
+ - 0-3 files modified: Small task (~10-15% context)
349
+ - 4-6 files modified: Medium task (~20-30% context)
350
+ - 7+ files modified: Large task (~40%+ context) - split this
351
+ </file_counts>
352
+
353
+ <complexity>
354
+ - Simple CRUD: ~15% per task
355
+ - Business logic: ~25% per task
356
+ - Complex algorithms: ~40% per task
357
+ - Domain modeling: ~35% per task
358
+ </complexity>
359
+
360
+ <two_task_plan>
361
+ - 2 simple tasks: ~30% total ✅ Plenty of room
362
+ - 2 medium tasks: ~50% total ✅ At target
363
+ - 2 complex tasks: ~80% total ❌ Too tight, split
364
+ </two_task_plan>
365
+
366
+ <three_task_plan>
367
+ - 3 simple tasks: ~45% total ✅ Good
368
+ - 3 medium tasks: ~75% total ⚠️ Pushing it
369
+ - 3 complex tasks: 120% total ❌ Impossible, split
370
+
371
+ **Conservative principle:** When in doubt, split. Better to have an extra plan than degraded quality.
372
+ </three_task_plan>
373
+ </estimating_context>
374
+
375
+ <atomic_commits>
376
+
377
+ **What we're optimizing for:** Beautiful git history where each commit is:
378
+ - Focused (2-3 related changes)
379
+ - Complete (fully implemented, tested)
380
+ - Documented (clear commit message)
381
+ - Reviewable (small enough to understand)
382
+ - Revertable (surgical rollback possible)
383
+
384
+ **Bad git history (large plans):**
385
+ ```
386
+ feat(auth): Complete authentication system
387
+ - Added 16 files
388
+ - Modified 8 files
389
+ - 1200 lines changed
390
+ - Contains: models, API, UI, middleware, utilities
391
+ ```
392
+
393
+ Impossible to review, hard to understand, can't revert without losing everything.
394
+
395
+ **Good git history (atomic plans):**
396
+ ```
397
+ feat(auth-01): Add User and Session database models
398
+ - Added schema files
399
+ - Added migration
400
+ - 45 lines changed
401
+
402
+ feat(auth-02): Implement register and login API endpoints
403
+ - Added /api/auth/register
404
+ - Added /api/auth/login
405
+ - Added JWT utilities
406
+ - 120 lines changed
407
+
408
+ feat(auth-03): Add protected route middleware
409
+ - Added middleware/auth.ts
410
+ - Added tests
411
+ - 60 lines changed
412
+
413
+ feat(auth-04): Build login and registration forms
414
+ - Added LoginForm component
415
+ - Added RegisterForm component
416
+ - 90 lines changed
417
+ ```
418
+
419
+ Each commit tells a story. Each is reviewable. Each is revertable. This is craftsmanship.
420
+ </atomic_commits>
421
+
422
+ <quality_assurance>
423
+
424
+ **The guarantee:** When you follow the 2-3 task rule with 50% context target:
425
+
426
+ 1. **Consistency:** First task has same quality as last task
427
+ 2. **Thoroughness:** No "I'll complete X concisely" degradation
428
+ 3. **Documentation:** Full context budget for comments/tests
429
+ 4. **Error handling:** Space for proper validation and edge cases
430
+ 5. **Testing:** Room for comprehensive test coverage
431
+
432
+ **The cost:** More plans to manage.
433
+
434
+ **The benefit:** Consistent excellence. No rework. Clean history. Maintainable code.
435
+
436
+ **The trade-off is worth it.**
437
+ </quality_assurance>
438
+
439
+ <summary>
440
+
441
+ **2-3 tasks, 50% context target:**
442
+ - All tasks: Peak quality
443
+ - Git: Atomic, surgical commits
444
+ - Quality: Consistent excellence
445
+ - Autonomous plans: Subagent execution (fresh context)
446
+
447
+ **The principle:** Aggressive atomicity. More plans, smaller scope, consistent quality.
448
+
449
+ **The rule:** If in doubt, split. Quality over consolidation. Always.
450
+ </summary>
451
+ </scope_estimation>
@@ -0,0 +1,17 @@
1
+ {
2
+ "mode": "interactive",
3
+ "gates": {
4
+ "confirm_project": true,
5
+ "confirm_phases": true,
6
+ "confirm_roadmap": true,
7
+ "confirm_breakdown": true,
8
+ "confirm_plan": true,
9
+ "execute_next_plan": true,
10
+ "issues_review": true,
11
+ "confirm_transition": true
12
+ },
13
+ "safety": {
14
+ "always_confirm_destructive": true,
15
+ "always_confirm_external_services": true
16
+ }
17
+ }