autodoc-agent-kit 1.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (187) hide show
  1. package/README.md +362 -0
  2. package/package.json +49 -0
  3. package/src/core/module.yaml +5 -0
  4. package/src/modules/design/module.yaml +9 -0
  5. package/src/modules/design/skills/brand-guidelines/LICENSE.txt +202 -0
  6. package/src/modules/design/skills/brand-guidelines/SKILL.md +73 -0
  7. package/src/modules/design/skills/frontend-design/LICENSE.txt +177 -0
  8. package/src/modules/design/skills/frontend-design/SKILL.md +42 -0
  9. package/src/modules/design/skills/web-artifacts-builder/SKILL.md +229 -0
  10. package/src/modules/devops/module.yaml +10 -0
  11. package/src/modules/devops/skills/devops-helper/SKILL.md +60 -0
  12. package/src/modules/devops/skills/k8s-helm/SKILL.md +360 -0
  13. package/src/modules/devops/skills/monitoring-observability/SKILL.md +240 -0
  14. package/src/modules/devops/skills/security-auditor/SKILL.md +105 -0
  15. package/src/modules/engineering/module.yaml +22 -0
  16. package/src/modules/engineering/skills/ai-sdk/SKILL.md +314 -0
  17. package/src/modules/engineering/skills/api-designer/SKILL.md +77 -0
  18. package/src/modules/engineering/skills/code-reviewer/SKILL.md +71 -0
  19. package/src/modules/engineering/skills/db-architect/SKILL.md +50 -0
  20. package/src/modules/engineering/skills/debugger/SKILL.md +59 -0
  21. package/src/modules/engineering/skills/docs-generator/SKILL.md +51 -0
  22. package/src/modules/engineering/skills/git-workflow/SKILL.md +258 -0
  23. package/src/modules/engineering/skills/mcp-builder/LICENSE.txt +202 -0
  24. package/src/modules/engineering/skills/mcp-builder/SKILL.md +236 -0
  25. package/src/modules/engineering/skills/mcp-builder/reference/evaluation.md +602 -0
  26. package/src/modules/engineering/skills/mcp-builder/reference/mcp_best_practices.md +249 -0
  27. package/src/modules/engineering/skills/mcp-builder/reference/node_mcp_server.md +970 -0
  28. package/src/modules/engineering/skills/mcp-builder/reference/python_mcp_server.md +719 -0
  29. package/src/modules/engineering/skills/mcp-builder/scripts/connections.py +151 -0
  30. package/src/modules/engineering/skills/mcp-builder/scripts/evaluation.py +373 -0
  31. package/src/modules/engineering/skills/mcp-builder/scripts/example_evaluation.xml +22 -0
  32. package/src/modules/engineering/skills/mcp-builder/scripts/requirements.txt +2 -0
  33. package/src/modules/engineering/skills/nextjs-15/SKILL.md +312 -0
  34. package/src/modules/engineering/skills/perf-optimizer/SKILL.md +60 -0
  35. package/src/modules/engineering/skills/react-19/SKILL.md +257 -0
  36. package/src/modules/engineering/skills/refactorer/SKILL.md +60 -0
  37. package/src/modules/engineering/skills/skill-authoring-workflow/SKILL.md +183 -0
  38. package/src/modules/engineering/skills/skill-creator/LICENSE.txt +202 -0
  39. package/src/modules/engineering/skills/skill-creator/SKILL.md +356 -0
  40. package/src/modules/engineering/skills/skill-creator/references/output-patterns.md +82 -0
  41. package/src/modules/engineering/skills/skill-creator/references/workflows.md +28 -0
  42. package/src/modules/engineering/skills/skill-creator/scripts/__pycache__/quick_validate.cpython-313.pyc +0 -0
  43. package/src/modules/engineering/skills/skill-creator/scripts/init_skill.py +303 -0
  44. package/src/modules/engineering/skills/skill-creator/scripts/package_skill.py +110 -0
  45. package/src/modules/engineering/skills/skill-creator/scripts/quick_validate.py +95 -0
  46. package/src/modules/engineering/skills/typescript/SKILL.md +231 -0
  47. package/src/modules/engineering/skills/zod-4/SKILL.md +223 -0
  48. package/src/modules/product/module.yaml +51 -0
  49. package/src/modules/product/skills/acquisition-channel-advisor/SKILL.md +643 -0
  50. package/src/modules/product/skills/acquisition-channel-advisor/examples/conversation-flow.md +531 -0
  51. package/src/modules/product/skills/ai-shaped-readiness-advisor/SKILL.md +923 -0
  52. package/src/modules/product/skills/altitude-horizon-framework/SKILL.md +250 -0
  53. package/src/modules/product/skills/altitude-horizon-framework/examples/sample.md +85 -0
  54. package/src/modules/product/skills/business-health-diagnostic/SKILL.md +783 -0
  55. package/src/modules/product/skills/company-research/SKILL.md +385 -0
  56. package/src/modules/product/skills/company-research/examples/sample.md +164 -0
  57. package/src/modules/product/skills/company-research/template.md +60 -0
  58. package/src/modules/product/skills/context-engineering-advisor/SKILL.md +763 -0
  59. package/src/modules/product/skills/customer-journey-map/SKILL.md +346 -0
  60. package/src/modules/product/skills/customer-journey-map/examples/meta-product-manager-skills.md +40 -0
  61. package/src/modules/product/skills/customer-journey-map/examples/sample.md +33 -0
  62. package/src/modules/product/skills/customer-journey-map/template.md +28 -0
  63. package/src/modules/product/skills/customer-journey-mapping-workshop/SKILL.md +523 -0
  64. package/src/modules/product/skills/director-readiness-advisor/SKILL.md +351 -0
  65. package/src/modules/product/skills/director-readiness-advisor/examples/conversation-flow.md +96 -0
  66. package/src/modules/product/skills/discovery-interview-prep/SKILL.md +410 -0
  67. package/src/modules/product/skills/discovery-process/SKILL.md +504 -0
  68. package/src/modules/product/skills/discovery-process/examples/sample.md +60 -0
  69. package/src/modules/product/skills/discovery-process/template.md +39 -0
  70. package/src/modules/product/skills/eol-message/SKILL.md +348 -0
  71. package/src/modules/product/skills/eol-message/examples/sample.md +87 -0
  72. package/src/modules/product/skills/eol-message/template.md +74 -0
  73. package/src/modules/product/skills/epic-breakdown-advisor/SKILL.md +665 -0
  74. package/src/modules/product/skills/epic-hypothesis/SKILL.md +277 -0
  75. package/src/modules/product/skills/epic-hypothesis/examples/sample.md +104 -0
  76. package/src/modules/product/skills/epic-hypothesis/template.md +30 -0
  77. package/src/modules/product/skills/executive-onboarding-playbook/SKILL.md +280 -0
  78. package/src/modules/product/skills/executive-onboarding-playbook/examples/sample.md +116 -0
  79. package/src/modules/product/skills/feature-investment-advisor/SKILL.md +639 -0
  80. package/src/modules/product/skills/feature-investment-advisor/examples/conversation-flow.md +538 -0
  81. package/src/modules/product/skills/finance-based-pricing-advisor/SKILL.md +763 -0
  82. package/src/modules/product/skills/finance-metrics-quickref/SKILL.md +309 -0
  83. package/src/modules/product/skills/jobs-to-be-done/SKILL.md +370 -0
  84. package/src/modules/product/skills/jobs-to-be-done/examples/sample.md +80 -0
  85. package/src/modules/product/skills/jobs-to-be-done/template.md +65 -0
  86. package/src/modules/product/skills/lean-ux-canvas/SKILL.md +561 -0
  87. package/src/modules/product/skills/lean-ux-canvas/examples/sample.md +88 -0
  88. package/src/modules/product/skills/lean-ux-canvas/template.md +32 -0
  89. package/src/modules/product/skills/opportunity-solution-tree/SKILL.md +420 -0
  90. package/src/modules/product/skills/opportunity-solution-tree/examples/sample.md +104 -0
  91. package/src/modules/product/skills/opportunity-solution-tree/template.md +33 -0
  92. package/src/modules/product/skills/pestel-analysis/SKILL.md +376 -0
  93. package/src/modules/product/skills/pestel-analysis/examples/sample.md +143 -0
  94. package/src/modules/product/skills/pestel-analysis/template.md +53 -0
  95. package/src/modules/product/skills/pol-probe/SKILL.md +217 -0
  96. package/src/modules/product/skills/pol-probe/examples/sample.md +136 -0
  97. package/src/modules/product/skills/pol-probe/template.md +59 -0
  98. package/src/modules/product/skills/pol-probe-advisor/SKILL.md +492 -0
  99. package/src/modules/product/skills/positioning-statement/SKILL.md +230 -0
  100. package/src/modules/product/skills/positioning-statement/examples/sample.md +51 -0
  101. package/src/modules/product/skills/positioning-statement/template.md +25 -0
  102. package/src/modules/product/skills/positioning-workshop/SKILL.md +424 -0
  103. package/src/modules/product/skills/prd-development/SKILL.md +655 -0
  104. package/src/modules/product/skills/prd-development/examples/sample.md +43 -0
  105. package/src/modules/product/skills/prd-development/template.md +55 -0
  106. package/src/modules/product/skills/press-release/SKILL.md +269 -0
  107. package/src/modules/product/skills/press-release/examples/sample.md +73 -0
  108. package/src/modules/product/skills/press-release/template.md +39 -0
  109. package/src/modules/product/skills/prioritization-advisor/SKILL.md +448 -0
  110. package/src/modules/product/skills/problem-framing-canvas/SKILL.md +466 -0
  111. package/src/modules/product/skills/problem-framing-canvas/examples/sample.md +58 -0
  112. package/src/modules/product/skills/problem-framing-canvas/template.md +22 -0
  113. package/src/modules/product/skills/problem-statement/SKILL.md +246 -0
  114. package/src/modules/product/skills/problem-statement/examples/sample.md +82 -0
  115. package/src/modules/product/skills/problem-statement/template.md +37 -0
  116. package/src/modules/product/skills/product-strategy-session/SKILL.md +426 -0
  117. package/src/modules/product/skills/product-strategy-session/examples/sample.md +67 -0
  118. package/src/modules/product/skills/product-strategy-session/template.md +38 -0
  119. package/src/modules/product/skills/proto-persona/SKILL.md +326 -0
  120. package/src/modules/product/skills/proto-persona/examples/sample.md +97 -0
  121. package/src/modules/product/skills/proto-persona/template.md +45 -0
  122. package/src/modules/product/skills/recommendation-canvas/SKILL.md +375 -0
  123. package/src/modules/product/skills/recommendation-canvas/examples/sample.md +94 -0
  124. package/src/modules/product/skills/recommendation-canvas/template.md +86 -0
  125. package/src/modules/product/skills/roadmap-planning/SKILL.md +505 -0
  126. package/src/modules/product/skills/roadmap-planning/examples/sample.md +62 -0
  127. package/src/modules/product/skills/roadmap-planning/template.md +30 -0
  128. package/src/modules/product/skills/saas-economics-efficiency-metrics/SKILL.md +694 -0
  129. package/src/modules/product/skills/saas-economics-efficiency-metrics/examples/cash-trap.md +365 -0
  130. package/src/modules/product/skills/saas-economics-efficiency-metrics/examples/healthy-unit-economics.md +279 -0
  131. package/src/modules/product/skills/saas-economics-efficiency-metrics/template.md +263 -0
  132. package/src/modules/product/skills/saas-revenue-growth-metrics/SKILL.md +630 -0
  133. package/src/modules/product/skills/saas-revenue-growth-metrics/examples/healthy-saas.md +131 -0
  134. package/src/modules/product/skills/saas-revenue-growth-metrics/examples/warning-signs.md +229 -0
  135. package/src/modules/product/skills/saas-revenue-growth-metrics/template.md +192 -0
  136. package/src/modules/product/skills/storyboard/SKILL.md +252 -0
  137. package/src/modules/product/skills/storyboard/examples/sample.md +71 -0
  138. package/src/modules/product/skills/storyboard/template.md +41 -0
  139. package/src/modules/product/skills/tam-sam-som-calculator/SKILL.md +392 -0
  140. package/src/modules/product/skills/tam-sam-som-calculator/examples/sample.md +142 -0
  141. package/src/modules/product/skills/tam-sam-som-calculator/scripts/market-sizing.py +95 -0
  142. package/src/modules/product/skills/tam-sam-som-calculator/template.md +35 -0
  143. package/src/modules/product/skills/user-story/SKILL.md +272 -0
  144. package/src/modules/product/skills/user-story/examples/sample.md +110 -0
  145. package/src/modules/product/skills/user-story/scripts/user-story-template.py +65 -0
  146. package/src/modules/product/skills/user-story/template.md +32 -0
  147. package/src/modules/product/skills/user-story-mapping/SKILL.md +285 -0
  148. package/src/modules/product/skills/user-story-mapping/examples/sample.md +77 -0
  149. package/src/modules/product/skills/user-story-mapping/template.md +41 -0
  150. package/src/modules/product/skills/user-story-mapping-workshop/SKILL.md +477 -0
  151. package/src/modules/product/skills/user-story-mapping-workshop/template.md +28 -0
  152. package/src/modules/product/skills/user-story-splitting/SKILL.md +303 -0
  153. package/src/modules/product/skills/user-story-splitting/examples/sample.md +147 -0
  154. package/src/modules/product/skills/user-story-splitting/template.md +37 -0
  155. package/src/modules/product/skills/vp-cpo-readiness-advisor/SKILL.md +409 -0
  156. package/src/modules/product/skills/vp-cpo-readiness-advisor/examples/conversation-flow.md +95 -0
  157. package/src/modules/product/skills/workshop-facilitation/SKILL.md +87 -0
  158. package/src/modules/productivity/module.yaml +9 -0
  159. package/src/modules/productivity/skills/doc-coauthoring/SKILL.md +375 -0
  160. package/src/modules/productivity/skills/internal-comms/LICENSE.txt +202 -0
  161. package/src/modules/productivity/skills/internal-comms/SKILL.md +32 -0
  162. package/src/modules/productivity/skills/internal-comms/examples/3p-updates.md +47 -0
  163. package/src/modules/productivity/skills/internal-comms/examples/company-newsletter.md +65 -0
  164. package/src/modules/productivity/skills/internal-comms/examples/faq-answers.md +30 -0
  165. package/src/modules/productivity/skills/internal-comms/examples/general-comms.md +16 -0
  166. package/src/modules/productivity/skills/technical-writing/SKILL.md +266 -0
  167. package/src/modules/qa/module.yaml +9 -0
  168. package/src/modules/qa/skills/test-strategy/SKILL.md +263 -0
  169. package/src/modules/qa/skills/test-writer/SKILL.md +57 -0
  170. package/src/modules/qa/skills/webapp-testing/LICENSE.txt +202 -0
  171. package/src/modules/qa/skills/webapp-testing/SKILL.md +96 -0
  172. package/src/modules/qa/skills/webapp-testing/examples/console_logging.py +35 -0
  173. package/src/modules/qa/skills/webapp-testing/examples/element_discovery.py +40 -0
  174. package/src/modules/qa/skills/webapp-testing/examples/static_html_automation.py +33 -0
  175. package/src/modules/qa/skills/webapp-testing/scripts/with_server.py +106 -0
  176. package/tools/autodoc-npx-wrapper.js +34 -0
  177. package/tools/cli/autodoc-cli.js +55 -0
  178. package/tools/cli/commands/install.js +36 -0
  179. package/tools/cli/commands/status.js +35 -0
  180. package/tools/cli/commands/uninstall.js +60 -0
  181. package/tools/cli/installers/lib/core/installer.js +164 -0
  182. package/tools/cli/installers/lib/core/manifest.js +49 -0
  183. package/tools/cli/installers/lib/ide/manager.js +112 -0
  184. package/tools/cli/installers/lib/ide/platform-codes.yaml +207 -0
  185. package/tools/cli/installers/lib/modules/manager.js +59 -0
  186. package/tools/cli/lib/ui.js +199 -0
  187. package/tools/cli/lib/welcome.js +82 -0
@@ -0,0 +1,763 @@
1
+ ---
2
+ name: context-engineering-advisor
3
+ description: Diagnose context stuffing vs. context engineering. Use when an AI workflow feels bloated, brittle, or hard to steer reliably.
4
+ intent: >-
5
+ Guide product managers through diagnosing whether they're doing **context stuffing** (jamming volume without intent) or **context engineering** (shaping structure for attention). Use this to identify context boundaries, fix "Context Hoarding Disorder," and implement tactical practices like bounded domains, episodic retrieval, and the Research→Plan→Reset→Implement cycle.
6
+ type: interactive
7
+ theme: ai-agents
8
+ best_for:
9
+ - "Diagnosing context stuffing vs. context engineering in your AI workflows"
10
+ - "Building better memory and retrieval architecture for AI agents"
11
+ - "Improving AI output quality through structured context design"
12
+ scenarios:
13
+ - "My AI outputs are mediocre even though I'm giving it lots of information — diagnose what's wrong"
14
+ - "I want to architect context properly for a multi-step AI workflow in my product team"
15
+ estimated_time: "15-20 min"
16
+ ---
17
+
18
+ ## Purpose
19
+
20
+ Guide product managers through diagnosing whether they're doing **context stuffing** (jamming volume without intent) or **context engineering** (shaping structure for attention). Use this to identify context boundaries, fix "Context Hoarding Disorder," and implement tactical practices like bounded domains, episodic retrieval, and the Research→Plan→Reset→Implement cycle.
21
+
22
+ **Key Distinction:** Context stuffing assumes volume = quality ("paste the entire PRD"). Context engineering treats AI attention as a scarce resource and allocates it deliberately.
23
+
24
+ This is not about prompt writing—it's about **designing the information architecture** that grounds AI in reality without overwhelming it with noise.
25
+
26
+ ## Key Concepts
27
+
28
+ ### The Paradigm Shift: Parametric → Contextual Intelligence
29
+
30
+ **The Fundamental Problem:**
31
+ - LLMs have **parametric knowledge** (encoded during training) = static, outdated, non-attributable
32
+ - When asked about proprietary data, real-time info, or user preferences → forced to hallucinate or admit ignorance
33
+ - **Context engineering** bridges the gap between static training and dynamic reality
34
+
35
+ **PM's Role Shift:** From feature builder → **architect of informational ecosystems** that ground AI in reality
36
+
37
+ ---
38
+
39
+ ### Context Stuffing vs. Context Engineering
40
+
41
+ | Dimension | Context Stuffing | Context Engineering |
42
+ |-----------|------------------|---------------------|
43
+ | **Mindset** | Volume = quality | Structure = quality |
44
+ | **Approach** | "Add everything just in case" | "What decision am I making?" |
45
+ | **Persistence** | Persist all context | Retrieve with intent |
46
+ | **Agent Chains** | Share everything between agents | Bounded context per agent |
47
+ | **Failure Response** | Retry until it works | Fix the structure |
48
+ | **Economic Model** | Context as storage | Context as attention (scarce resource) |
49
+
50
+ **Critical Metaphor:** Context stuffing is like bringing your entire file cabinet to a meeting. Context engineering is bringing only the 3 documents relevant to today's decision.
51
+
52
+ ---
53
+
54
+ ### The Anti-Pattern: Context Stuffing
55
+
56
+ **Five Markers of Context Stuffing:**
57
+ 1. **Reflexively expanding context windows** — "Just add more tokens!"
58
+ 2. **Persisting everything "just in case"** — No clear retention criteria
59
+ 3. **Chaining agents without boundaries** — Agent A passes everything to Agent B to Agent C
60
+ 4. **Adding evaluations to mask inconsistency** — "We'll just retry until it's right"
61
+ 5. **Normalized retries** — "It works if you run it 3 times" becomes acceptable
62
+
63
+ **Why It Fails:**
64
+ - **Reasoning Noise:** Thousands of irrelevant files compete for attention, degrading multi-hop logic
65
+ - **Context Rot:** Dead ends, past errors, irrelevant data accumulate → goal drift
66
+ - **Lost in the Middle:** Models prioritize beginning (primacy) and end (recency), ignore middle
67
+ - **Economic Waste:** Every query becomes expensive without accuracy gains
68
+ - **Quantitative Degradation:** Accuracy drops below 20% when context exceeds ~32k tokens
69
+
70
+ **The Hidden Costs:**
71
+ - Escalating token consumption
72
+ - Diluted attention across irrelevant material
73
+ - Reduced output confidence
74
+ - Cascading retries that waste time and money
75
+
76
+ ---
77
+
78
+ ### Real Context Engineering: Core Principles
79
+
80
+ **Five Foundational Principles:**
81
+ 1. **Context without shape becomes noise**
82
+ 2. **Structure > Volume**
83
+ 3. **Retrieve with intent, not completeness**
84
+ 4. **Small working contexts** (like short-term memory)
85
+ 5. **Context Compaction:** Maximize density of relevant information per token
86
+
87
+ **Quantitative Framework:**
88
+ ```
89
+ Efficiency = (Accuracy × Coherence) / (Tokens × Latency)
90
+ ```
91
+
92
+ **Key Finding:** Using RAG with 25% of available tokens preserves 95% accuracy while significantly reducing latency and cost.
93
+
94
+ ---
95
+
96
+ ### The 5 Diagnostic Questions (Detect Context Hoarding Disorder)
97
+
98
+ Ask these to identify context stuffing:
99
+
100
+ 1. **What specific decision does this support?** — If you can't answer, you don't need it
101
+ 2. **Can retrieval replace persistence?** — Just-in-time beats always-available
102
+ 3. **Who owns the context boundary?** — If no one, it'll grow forever
103
+ 4. **What fails if we exclude this?** — If nothing breaks, delete it
104
+ 5. **Are we fixing structure or avoiding it?** — Stuffing context often masks bad information architecture
105
+
106
+ ---
107
+
108
+ ### Memory Architecture: Two-Layer System
109
+
110
+ **Short-Term (Conversational) Memory:**
111
+ - Immediate interaction history for follow-up questions
112
+ - Challenge: Space management → older parts summarized or truncated
113
+ - Lifespan: Single session
114
+
115
+ **Long-Term (Persistent) Memory:**
116
+ - User preferences, key facts across sessions → deep personalization
117
+ - Implemented via vector database (semantic retrieval)
118
+ - Two types:
119
+ - **Declarative Memory:** Facts ("I'm vegan")
120
+ - **Procedural Memory:** Behavioral patterns ("I debug by checking logs first")
121
+ - Lifespan: Persistent across sessions
122
+
123
+ **LLM-Powered ETL:** Models generate their own memories by identifying signals, consolidating with existing data, updating database automatically.
124
+
125
+ ---
126
+
127
+ ### The Research → Plan → Reset → Implement Cycle
128
+
129
+ **The Context Rot Solution:**
130
+
131
+ 1. **Research:** Agent gathers data → large, chaotic context window (noise + dead ends)
132
+ 2. **Plan:** Agent synthesizes into high-density SPEC.md or PLAN.md (Source of Truth)
133
+ 3. **Reset:** **Clear entire context window** (prevents context rot)
134
+ 4. **Implement:** Fresh session using **only** the high-density plan as context
135
+
136
+ **Why This Works:** Context rot is eliminated; agent starts clean with compressed, high-signal context.
137
+
138
+ ---
139
+
140
+ ### Anti-Patterns (What This Is NOT)
141
+
142
+ - **Not about choosing AI tools** — Claude vs. ChatGPT doesn't matter; architecture matters
143
+ - **Not about writing better prompts** — This is systems design, not copywriting
144
+ - **Not about adding more tokens** — "Infinite context" narratives are marketing, not engineering reality
145
+ - **Not about replacing human judgment** — Context engineering amplifies judgment, doesn't eliminate it
146
+
147
+ ---
148
+
149
+ ### When to Use This Skill
150
+
151
+ ✅ **Use this when:**
152
+ - You're pasting entire PRDs/codebases into AI and getting vague responses
153
+ - AI outputs are inconsistent ("works sometimes, not others")
154
+ - You're burning tokens without seeing accuracy improvements
155
+ - You suspect you're "context stuffing" but don't know how to fix it
156
+ - You need to design context architecture for an AI product feature
157
+
158
+ ❌ **Don't use this when:**
159
+ - You're just getting started with AI (start with basic prompts first)
160
+ - You're looking for tool recommendations (this is about architecture, not tooling)
161
+ - Your AI usage is working well (if it ain't broke, don't fix it)
162
+
163
+ ---
164
+
165
+ ### Facilitation Source of Truth
166
+
167
+ Use [`workshop-facilitation`](../workshop-facilitation/SKILL.md) as the default interaction protocol for this skill.
168
+
169
+ It defines:
170
+ - session heads-up + entry mode (Guided, Context dump, Best guess)
171
+ - one-question turns with plain-language prompts
172
+ - progress labels (for example, Context Qx/8 and Scoring Qx/5)
173
+ - interruption handling and pause/resume behavior
174
+ - numbered recommendations at decision points
175
+ - quick-select numbered response options for regular questions (include `Other (specify)` when useful)
176
+
177
+ This file defines the domain-specific assessment content. If there is a conflict, follow this file's domain logic.
178
+
179
+ ## Application
180
+
181
+ This interactive skill uses **adaptive questioning** to diagnose context stuffing, identify boundaries, and provide tactical implementation guidance.
182
+
183
+ ---
184
+
185
+ ### Step 0: Gather Context
186
+
187
+ **Agent asks:**
188
+
189
+ Before we diagnose your context practices, let's gather information:
190
+
191
+ **Current AI Usage:**
192
+ - What AI tools/systems do you use? (ChatGPT, Claude, custom agents, etc.)
193
+ - What PM tasks do you use AI for? (PRD writing, user research synthesis, discovery, etc.)
194
+ - How do you provide context? (paste docs, reference files, use projects/memory)
195
+
196
+ **Symptoms:**
197
+ - Are AI outputs inconsistent? (works sometimes, not others)
198
+ - Are you retrying prompts multiple times to get good results?
199
+ - Are responses vague or hedged despite providing "all the context"?
200
+ - Are token costs escalating without accuracy improvements?
201
+
202
+ **System Architecture (if applicable):**
203
+ - Do you have custom AI agents or workflows?
204
+ - How is context shared between agents?
205
+ - Do you use RAG, vector databases, or memory systems?
206
+
207
+ **You can describe briefly or paste examples.**
208
+
209
+ ---
210
+
211
+ ### Step 1: Diagnose Context Stuffing Symptoms
212
+
213
+ **Agent asks:**
214
+
215
+ Let's assess whether you're experiencing **context stuffing**. Which of these symptoms do you recognize?
216
+
217
+ **Select all that apply:**
218
+
219
+ 1. **"I paste entire documents into AI"** — Full PRDs, complete user interview transcripts, entire codebases
220
+ 2. **"AI gives vague, hedged responses despite having 'all the context'"** — Responses like "it depends," "consider these options," non-committal
221
+ 3. **"I have to retry prompts 3+ times to get usable output"** — Inconsistency is normalized
222
+ 4. **"Token costs are escalating but accuracy isn't improving"** — Spending more, getting same or worse results
223
+ 5. **"I keep adding more context hoping it'll help"** — Reflexive expansion without strategy
224
+ 6. **"My agents pass everything to each other"** — Agent A → Agent B → Agent C with full context chain
225
+ 7. **"I don't have clear criteria for what to include/exclude"** — No context boundary definitions
226
+ 8. **"None of these—my AI usage is working well"** — Skip to advanced optimization
227
+
228
+ **User response:** [Select symptoms]
229
+
230
+ **Agent analyzes:**
231
+
232
+ Based on your selections:
233
+ - **0-1 symptoms:** Healthy context practices; proceed to optimization
234
+ - **2-3 symptoms:** Early context stuffing; address before it scales
235
+ - **4+ symptoms:** Active Context Hoarding Disorder; immediate intervention needed
236
+
237
+ **Agent proceeds to diagnostic questions.**
238
+
239
+ ---
240
+
241
+ ### Step 2: Diagnostic Question 1 — What Specific Decision Does This Support?
242
+
243
+ **Agent asks:**
244
+
245
+ Let's start with the most critical question: **For each piece of context you're including, what specific decision does it support?**
246
+
247
+ **Example Context Analysis:**
248
+
249
+ Imagine you're asking AI to "help with discovery planning." You're providing:
250
+ - Entire PRD (20 pages)
251
+ - 50 user interview transcripts (full)
252
+ - Competitive analysis doc (15 pages)
253
+ - Team meeting notes from last 3 months
254
+
255
+ **Question:** What decision are you making right now?
256
+
257
+ **Offer 3 scenarios:**
258
+
259
+ 1. **"I'm deciding which user segment to interview first"**
260
+ - **Context needed:** User segments from PRD (2 paragraphs), prior interview themes (1 page synthesis), not full transcripts
261
+ - **Context NOT needed:** Meeting notes, full competitive analysis, full PRD
262
+
263
+ 2. **"I'm deciding which discovery questions to ask in interviews"**
264
+ - **Context needed:** Research objectives (from PRD), past interview insights (synthesis), Jobs-to-be-Done framework
265
+ - **Context NOT needed:** Full competitive analysis, full meeting notes
266
+
267
+ 3. **"I'm not sure what decision I'm making—I just want AI to 'understand my product'"**
268
+ - **Problem:** No specific decision = context stuffing trap
269
+ - **Fix:** Define the decision first, then select context
270
+
271
+ **Agent recommends:**
272
+
273
+ **Best Practice:** Before adding context, complete this sentence:
274
+ > "I need this context because I'm deciding [specific decision], and without [specific information], I can't make that decision."
275
+
276
+ If you can't complete that sentence, you don't need the context.
277
+
278
+ **User response:** [Describe their decision + context]
279
+
280
+ **Agent validates:** Does the context directly support the stated decision? If not, recommend trimming.
281
+
282
+ ---
283
+
284
+ ### Step 3: Diagnostic Question 2 — Can Retrieval Replace Persistence?
285
+
286
+ **Agent asks:**
287
+
288
+ **Second question: Is this information you always need, or something you can retrieve just-in-time?**
289
+
290
+ **The Distinction:**
291
+
292
+ **Always-Needed (Persist):**
293
+ - Core product constraints (technical, regulatory, strategic)
294
+ - User preferences that apply to every interaction
295
+ - Critical definitions (operational glossary)
296
+ - Non-negotiable rules
297
+
298
+ **Episodic (Retrieve on-demand):**
299
+ - Project-specific details (this epic, this sprint)
300
+ - Historical data (past PRDs, old interview transcripts)
301
+ - Contextual facts (competitive analysis, market research)
302
+ - Temporary decisions
303
+
304
+ **Key Insight:** Just-in-time retrieval beats always-available. Don't persist what you can retrieve.
305
+
306
+ **Offer 3 options:**
307
+
308
+ 1. **"Most of my context is always-needed (core constraints, user prefs)"**
309
+ - **Assessment:** Good instinct; verify with Question 4 (what fails if excluded?)
310
+ - **Recommendation:** Build constraints registry and operational glossary (persist these)
311
+
312
+ 2. **"Most of my context is episodic (project details, historical data)"**
313
+ - **Assessment:** Perfect candidate for RAG or retrieval
314
+ - **Recommendation:** Implement semantic search; retrieve only relevant chunks for each query
315
+
316
+ 3. **"I'm not sure which is which—I persist everything to be safe"**
317
+ - **Assessment:** Classic Context Hoarding Disorder symptom
318
+ - **Fix:** Apply Question 4 test to each piece of context
319
+
320
+ **Agent recommends:**
321
+
322
+ **Rule of Thumb:**
323
+ - **Persist:** Information referenced in 80%+ of interactions
324
+ - **Retrieve:** Information referenced in <20% of interactions
325
+ - **Gray zone (20-80%):** Depends on retrieval latency vs. context window cost
326
+
327
+ **User response:** [Categorize their context]
328
+
329
+ **Agent provides:** Specific recommendations on what to persist vs. retrieve.
330
+
331
+ ---
332
+
333
+ ### Step 4: Diagnostic Question 3 — Who Owns the Context Boundary?
334
+
335
+ **Agent asks:**
336
+
337
+ **Third question: Who is responsible for defining what belongs in vs. out of your AI's context?**
338
+
339
+ **The Ownership Problem:**
340
+
341
+ If **no one** owns the context boundary, it will grow indefinitely. Every PM will add "just one more thing," and six months later, you're stuffing 100k tokens per query.
342
+
343
+ **Offer 3 options:**
344
+
345
+ 1. **"I own the boundary (solo PM or small team)"**
346
+ - **Assessment:** Good—you can make fast decisions
347
+ - **Recommendation:** Document your boundary criteria (use Questions 1-5 as framework)
348
+
349
+ 2. **"My team shares ownership (collaborative boundary definition)"**
350
+ - **Assessment:** Can work if formalized
351
+ - **Recommendation:** Create a "Context Manifest" doc: what's always included, what's retrieved, what's excluded (and why)
352
+
353
+ 3. **"No one owns it—it's ad-hoc / implicit"**
354
+ - **Assessment:** Critical risk; boundary will expand uncontrollably
355
+ - **Fix:** Assign explicit ownership; schedule quarterly context audits
356
+
357
+ **Agent recommends:**
358
+
359
+ **Best Practice: Create a Context Manifest**
360
+
361
+ ```markdown
362
+ # Context Manifest: [Product/Feature Name]
363
+
364
+ ## Always Persisted (Core Context)
365
+ - Product constraints (technical, regulatory)
366
+ - User preferences (role, permissions, preferences)
367
+ - Operational glossary (20 key terms)
368
+
369
+ ## Retrieved On-Demand (Episodic Context)
370
+ - Historical PRDs (retrieve via semantic search)
371
+ - User interview transcripts (retrieve relevant quotes)
372
+ - Competitive analysis (retrieve when explicitly needed)
373
+
374
+ ## Excluded (Out of Scope)
375
+ - Meeting notes older than 30 days (no longer relevant)
376
+ - Full codebase (use code search instead)
377
+ - Marketing materials (not decision-relevant)
378
+
379
+ ## Boundary Owner: [Name]
380
+ ## Last Reviewed: [Date]
381
+ ## Next Review: [Date + 90 days]
382
+ ```
383
+
384
+ **User response:** [Describe current ownership model]
385
+
386
+ **Agent provides:** Recommendation on formalizing ownership + template for Context Manifest.
387
+
388
+ ---
389
+
390
+ ### Step 5: Diagnostic Question 4 — What Fails if We Exclude This?
391
+
392
+ **Agent asks:**
393
+
394
+ **Fourth question: For each piece of context, what specific failure mode occurs if you exclude it?**
395
+
396
+ This is the **falsification test**. If you can't identify a concrete failure, you don't need the context.
397
+
398
+ **Offer 3 scenarios:**
399
+
400
+ 1. **"If I exclude product constraints, AI will recommend infeasible solutions"**
401
+ - **Failure Mode:** Clear and concrete
402
+ - **Assessment:** Valid reason to persist constraints
403
+
404
+ 2. **"If I exclude historical PRDs, AI won't understand our product evolution"**
405
+ - **Failure Mode:** Vague and hypothetical
406
+ - **Assessment:** Historical context rarely needed for current decisions
407
+ - **Fix:** Retrieve PRDs only when explicitly referencing past decisions
408
+
409
+ 3. **"If I exclude this, I'm not sure anything would break—I just include it to be thorough"**
410
+ - **Failure Mode:** None identified
411
+ - **Assessment:** Context stuffing; delete immediately
412
+
413
+ **Agent recommends:**
414
+
415
+ **The Falsification Protocol:**
416
+
417
+ For each context element, complete this statement:
418
+ > "If I exclude [context element], then [specific failure] will occur in [specific scenario]."
419
+
420
+ **Examples:**
421
+ - ✅ Good: "If I exclude GDPR constraints, AI will recommend features that violate EU privacy law."
422
+ - ❌ Bad: "If I exclude this PRD, AI might not fully understand the product." (Vague)
423
+
424
+ **User response:** [Apply falsification test to their context]
425
+
426
+ **Agent provides:** List of context elements to delete (no concrete failure identified).
427
+
428
+ ---
429
+
430
+ ### Step 6: Diagnostic Question 5 — Are We Fixing Structure or Avoiding It?
431
+
432
+ **Agent asks:**
433
+
434
+ **Fifth question: Is adding more context solving a problem, or masking a deeper structural issue?**
435
+
436
+ **The Root Cause Question:**
437
+
438
+ Context stuffing often hides bad information architecture. Instead of fixing messy, ambiguous documents, teams add more documents hoping AI will "figure it out."
439
+
440
+ **Offer 3 options:**
441
+
442
+ 1. **"I'm adding context because our docs are poorly structured/ambiguous"**
443
+ - **Assessment:** You're masking a structural problem
444
+ - **Fix:** Clean up the docs first (remove ambiguity, add constraints, define terms)
445
+ - **Example:** Instead of pasting 5 conflicting PRDs, reconcile them into 1 Source of Truth
446
+
447
+ 2. **"I'm adding context because we don't have a shared operational glossary"**
448
+ - **Assessment:** You're compensating for missing foundations
449
+ - **Fix:** Build the glossary (20-30 key terms); AI can reference it reliably
450
+ - **Example:** Define "active user," "churn," "engagement" unambiguously
451
+
452
+ 3. **"I'm adding context because our constraints aren't documented"**
453
+ - **Assessment:** You're avoiding constraint engineering
454
+ - **Fix:** Create constraints registry (technical, regulatory, strategic)
455
+ - **Example:** Document "We won't build mobile apps" vs. explaining it in every prompt
456
+
457
+ **Agent recommends:**
458
+
459
+ **The Structural Health Test:**
460
+
461
+ If you're adding context to compensate for:
462
+ - **Ambiguous documentation** → Fix the docs, don't add more
463
+ - **Undefined terms** → Build operational glossary
464
+ - **Undocumented constraints** → Create constraints registry
465
+ - **Conflicting information** → Reconcile into Source of Truth
466
+
467
+ **User response:** [Identify structural issues]
468
+
469
+ **Agent provides:** Prioritized list of structural fixes before adding more context.
470
+
471
+ ---
472
+
473
+ ### Step 7: Define Memory Architecture
474
+
475
+ **Agent asks:**
476
+
477
+ Based on your context needs, let's design a **two-layer memory architecture**.
478
+
479
+ **Your Context Profile (from previous steps):**
480
+ - Always-needed context: [Summary from Q2]
481
+ - Episodic context: [Summary from Q2]
482
+ - Boundary owner: [From Q3]
483
+ - Validated essentials: [From Q4]
484
+ - Structural fixes needed: [From Q5]
485
+
486
+ **Recommended Architecture:**
487
+
488
+ **Short-Term (Conversational) Memory:**
489
+ - **What it stores:** Immediate interaction history for follow-up questions
490
+ - **Lifespan:** Single session
491
+ - **Management:** Summarize or truncate older parts to avoid crowding
492
+ - **Your specific needs:** [Agent customizes based on user's workflow]
493
+
494
+ **Long-Term (Persistent) Memory:**
495
+ - **What it stores:** User preferences, core constraints, operational glossary
496
+ - **Lifespan:** Persistent across sessions
497
+ - **Implementation:** Vector database (semantic retrieval)
498
+ - **Two types:**
499
+ - **Declarative Memory:** Facts (e.g., "We follow HIPAA regulations")
500
+ - **Procedural Memory:** Behavioral patterns (e.g., "Always validate feasibility before usability")
501
+ - **Your specific needs:** [Agent customizes]
502
+
503
+ **Retrieval Strategy (Episodic Context):**
504
+ - **What it retrieves:** Historical PRDs, user interviews, competitive analysis
505
+ - **Method:** Semantic search triggered by query intent
506
+ - **Optimization:** Contextual Retrieval (Anthropic) — prepend explanatory context to each chunk before embedding
507
+ - **Your specific needs:** [Agent customizes]
508
+
509
+ **Agent offers:**
510
+
511
+ Would you like me to:
512
+ 1. **Generate a Context Architecture Blueprint** for your specific use case?
513
+ 2. **Provide implementation guidance** (tools, techniques, best practices)?
514
+ 3. **Design a retrieval strategy** for your episodic context?
515
+
516
+ **User response:** [Selection]
517
+
518
+ ---
519
+
520
+ ### Step 8: Implement Research → Plan → Reset → Implement Cycle
521
+
522
+ **Agent asks:**
523
+
524
+ Now let's implement the **Research → Plan → Reset → Implement** cycle to prevent context rot.
525
+
526
+ **The Problem:** As agents research, context windows grow chaotic—filled with dead ends, errors, and noise. This dilutes attention and causes goal drift.
527
+
528
+ **The Solution:** Compress research into a high-density plan, then **clear the context window** before implementing.
529
+
530
+ **The Four-Phase Cycle:**
531
+
532
+ **Phase 1: Research (Chaotic Context Allowed)**
533
+ - Agent gathers data from multiple sources
534
+ - Context window grows large and messy (this is expected)
535
+ - Dead ends, failed hypotheses, and noise accumulate
536
+ - **Goal:** Comprehensive information gathering
537
+
538
+ **Phase 2: Plan (Synthesis)**
539
+ - Agent synthesizes research into a high-density SPEC.md or PLAN.md
540
+ - This becomes the **Source of Truth** for implementation
541
+ - **Key elements:**
542
+ - Decision made
543
+ - Evidence supporting decision
544
+ - Constraints applied
545
+ - Next steps (sequenced)
546
+ - **Format:** Structured, concise, unambiguous
547
+
548
+ **Phase 3: Reset (Clear Context Window)**
549
+ - **Critical step:** Clear the entire context window
550
+ - Delete all research artifacts, dead ends, errors
551
+ - This prevents context rot from poisoning implementation
552
+
553
+ **Phase 4: Implement (Fresh Session with Plan Only)**
554
+ - Start a new session with **only the high-density plan** as context
555
+ - Agent has clean, focused attention on execution
556
+ - No noise from research phase
557
+
558
+ **Agent offers 3 options:**
559
+
560
+ 1. **"I want a template for the PLAN.md format"**
561
+ - Agent provides structured template for high-density plans
562
+
563
+ 2. **"I want to see an example of this cycle in action"**
564
+ - Agent walks through concrete PM use case (e.g., discovery planning)
565
+
566
+ 3. **"I'm ready to implement this in my workflow"**
567
+ - Agent provides step-by-step implementation guide
568
+
569
+ **User response:** [Selection]
570
+
571
+ **Agent provides:** Tailored guidance based on selection.
572
+
573
+ ---
574
+
575
+ ### Step 9: Action Plan & Next Steps
576
+
577
+ **Agent synthesizes:**
578
+
579
+ Based on your context engineering assessment, here's your action plan:
580
+
581
+ **Immediate Fixes (This Week):**
582
+ 1. [Delete context with no falsifiable failure mode from Q4]
583
+ 2. [Apply Research→Plan→Reset→Implement to your next AI task]
584
+ 3. [Document context boundary in Context Manifest]
585
+
586
+ **Foundation Building (Next 2 Weeks):**
587
+ 1. [Build constraints registry with 20+ entries]
588
+ 2. [Create operational glossary with 20-30 key terms]
589
+ 3. [Implement two-layer memory architecture]
590
+
591
+ **Long-Term Optimization (Next Month):**
592
+ 1. [Set up semantic retrieval for episodic context]
593
+ 2. [Assign context boundary owner + quarterly audit schedule]
594
+ 3. [Implement Contextual Retrieval (Anthropic) for RAG]
595
+
596
+ **Success Metrics:**
597
+ - Token usage down 50%+ (less context stuffing)
598
+ - Output consistency up (less retry/regeneration)
599
+ - Response quality up (sharper, less hedged answers)
600
+ - Context window stable (no unbounded growth)
601
+
602
+ **Agent offers:**
603
+
604
+ Would you like me to:
605
+ 1. **Generate specific implementation docs** (Context Manifest, PLAN.md template, etc.)?
606
+ 2. **Provide advanced techniques** (Contextual Retrieval, LLM-powered ETL)?
607
+ 3. **Review your current context setup** (provide feedback on specific prompts/workflows)?
608
+
609
+ ---
610
+
611
+ ## Examples
612
+
613
+ ### Example 1: Solo PM Context Stuffing → Engineering
614
+
615
+ **Context:**
616
+ - Solo PM at early-stage startup
617
+ - Using Claude Projects for PRD writing
618
+ - Pasting entire PRDs (20 pages) + all user interviews (50 transcripts) every time
619
+ - Getting vague, inconsistent responses
620
+
621
+ **Assessment:**
622
+ - Symptoms: Hedged responses, normalized retries (4+ symptoms)
623
+ - Q1 (Decision): "I just want AI to understand my product" (no specific decision)
624
+ - Q2 (Persist/Retrieve): Persisting everything (no retrieval strategy)
625
+ - Q3 (Ownership): No formal owner (solo PM, ad-hoc)
626
+ - Q4 (Failure): Can't identify concrete failures for most context
627
+ - Q5 (Structure): Avoiding constraint documentation
628
+
629
+ **Diagnosis:** Active Context Hoarding Disorder
630
+
631
+ **Intervention:**
632
+ 1. **Immediate:** Delete all context that fails Q4 test → keeps 20% of original
633
+ 2. **Week 1:** Build constraints registry (10 technical constraints, 5 strategic)
634
+ 3. **Week 2:** Create operational glossary (25 terms)
635
+ 4. **Week 3:** Implement Research→Plan→Reset→Implement for next PRD
636
+
637
+ **Outcome:** Token usage down 70%, output quality up significantly, responses crisp and actionable.
638
+
639
+ ---
640
+
641
+ ### Example 2: Growth-Stage Team with Agent Chains
642
+
643
+ **Context:**
644
+ - Product team with 5 PMs
645
+ - Custom AI agents for discovery synthesis
646
+ - Agent A (research) → Agent B (synthesis) → Agent C (recommendations)
647
+ - Each agent passes full context to next → context window explodes to 100k+ tokens
648
+
649
+ **Assessment:**
650
+ - Symptoms: Escalating token costs, inconsistent outputs (3 symptoms)
651
+ - Q1 (Decision): Each agent has clear decision, but passes unnecessary context
652
+ - Q2 (Persist/Retrieve): Mixing persistent and episodic without strategy
653
+ - Q3 (Ownership): No explicit owner; each PM adds context
654
+ - Q4 (Failure): Agents pass "just in case" context with no falsifiable failure
655
+ - Q5 (Structure): Missing Context Manifest
656
+
657
+ **Diagnosis:** Agent orchestration without boundaries
658
+
659
+ **Intervention:**
660
+ 1. **Immediate:** Define bounded context per agent (Agent A outputs only 2-page synthesis to Agent B, not full research)
661
+ 2. **Week 1:** Assign context boundary owner (Lead PM)
662
+ 3. **Week 2:** Create Context Manifest (what persists, what's retrieved, what's excluded)
663
+ 4. **Week 3:** Implement Research→Plan→Reset→Implement between Agent B and Agent C
664
+
665
+ **Outcome:** Token usage down 60%, agent chain reliability up, costs reduced by 50%.
666
+
667
+ ---
668
+
669
+ ### Example 3: Enterprise with RAG but No Context Engineering
670
+
671
+ **Context:**
672
+ - Large enterprise with vector database RAG system
673
+ - "Stuff the whole knowledge base" approach (10,000+ documents)
674
+ - Retrieval returns 50+ chunks per query → floods context window
675
+ - Accuracy declining as knowledge base grows
676
+
677
+ **Assessment:**
678
+ - Symptoms: Vague responses despite "complete knowledge," normalized retries (2 symptoms)
679
+ - Q1 (Decision): Decisions clear, but retrieval has no intent (returns everything)
680
+ - Q2 (Persist/Retrieve): Good instinct to retrieve, but no filtering
681
+ - Q3 (Ownership): Engineering owns RAG, Product doesn't own context boundaries
682
+ - Q4 (Failure): Can't identify why 50 chunks needed vs. 5
683
+ - Q5 (Structure): Knowledge base has no structure (flat documents, no metadata)
684
+
685
+ **Diagnosis:** Retrieval without intent (RAG as context stuffing)
686
+
687
+ **Intervention:**
688
+ 1. **Immediate:** Limit retrieval to top 5 chunks per query (down from 50)
689
+ 2. **Week 1:** Implement Contextual Retrieval (Anthropic) — prepend explanatory context to each chunk during indexing
690
+ 3. **Week 2:** Add metadata to documents (category, recency, authority)
691
+ 4. **Week 3:** Product team defines retrieval intent per query type (discovery = customer insights, feasibility = technical constraints)
692
+
693
+ **Outcome:** Accuracy up 35% (from Anthropic benchmark), latency down 60%, token usage down 80%.
694
+
695
+ ---
696
+
697
+ ## Common Pitfalls
698
+
699
+ ### 1. **"Infinite Context" Marketing vs. Engineering Reality**
700
+ **Failure Mode:** Believing "1 million token context windows" means you should use all of them.
701
+
702
+ **Consequence:** Reasoning Noise degrades performance; accuracy drops below 20% past ~32k tokens.
703
+
704
+ **Fix:** Context windows are not free. Treat tokens as scarce; optimize for density, not volume.
705
+
706
+ ---
707
+
708
+ ### 2. **Retrying Instead of Restructuring**
709
+ **Failure Mode:** "It works if I run it 3 times" → normalizing retries instead of fixing structure.
710
+
711
+ **Consequence:** Wastes time and money; masks deeper context rot issues.
712
+
713
+ **Fix:** If retries are common, your context structure is broken. Apply Q5 (fix structure, don't add volume).
714
+
715
+ ---
716
+
717
+ ### 3. **No Context Boundary Owner**
718
+ **Failure Mode:** Ad-hoc, implicit context decisions → unbounded growth.
719
+
720
+ **Consequence:** Six months later, every query stuffs 100k tokens per interaction.
721
+
722
+ **Fix:** Assign explicit ownership; create Context Manifest; schedule quarterly audits.
723
+
724
+ ---
725
+
726
+ ### 4. **Mixing Always-Needed with Episodic**
727
+ **Failure Mode:** Persisting historical data that should be retrieved on-demand.
728
+
729
+ **Consequence:** Context window crowded with irrelevant information; attention diluted.
730
+
731
+ **Fix:** Apply Q2 test: persist only what's needed in 80%+ of interactions; retrieve the rest.
732
+
733
+ ---
734
+
735
+ ### 5. **Skipping the Reset Phase**
736
+ **Failure Mode:** Never clearing context window during Research→Plan→Implement cycle.
737
+
738
+ **Consequence:** Context rot accumulates; goal drift; dead ends poison implementation.
739
+
740
+ **Fix:** Mandatory Reset phase after Plan; start implementation with only high-density plan as context.
741
+
742
+ ---
743
+
744
+ ## References
745
+
746
+ ### Related Skills
747
+ - **[ai-shaped-readiness-advisor](../ai-shaped-readiness-advisor/SKILL.md)** (Interactive) — Context Design is Competency #1 of AI-shaped work
748
+ - **[problem-statement](../problem-statement/SKILL.md)** (Component) — Evidence-based framing requires context engineering
749
+ - **[epic-hypothesis](../epic-hypothesis/SKILL.md)** (Component) — Testable hypotheses depend on clear constraints (part of context)
750
+ - **[pol-probe-advisor](../pol-probe-advisor/SKILL.md)** (Interactive) — Validation experiments benefit from context engineering (define what AI needs to know)
751
+
752
+ ### External Frameworks
753
+ - **Dean Peters** — [*Context Stuffing Is Not Context Engineering*](https://deanpeters.substack.com/p/context-stuffing-is-not-context-engineering) (Dean Peters' Substack, 2026)
754
+ - **Teresa Torres** — *Continuous Discovery Habits* (Context Engineering as one of 5 new AI PM disciplines)
755
+ - **Marty Cagan** — *Empowered* (Feasibility risk in AI era includes understanding "physics of AI")
756
+ - **Anthropic** — [Contextual Retrieval whitepaper](https://www.anthropic.com/news/contextual-retrieval) (35% failure rate reduction)
757
+ - **Google** — Context engineering whitepaper on LLM-powered memory systems
758
+
759
+ ### Technical References
760
+ - **RAG (Retrieval-Augmented Generation)** — Standard technique for episodic context retrieval
761
+ - **Vector Databases** — Semantic search for long-term memory (Pinecone, Weaviate, Chroma)
762
+ - **Contextual Retrieval (Anthropic)** — Prepend explanatory context to chunks before embedding
763
+ - **LLM-as-Judge** — Automated evaluation of context quality