startup-ideation-kit 1.0.0 → 2.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (42) hide show
  1. package/README.md +46 -34
  2. package/bin/cli.js +7 -1
  3. package/package.json +7 -3
  4. package/skills/sk-competitors/SKILL.md +284 -0
  5. package/skills/sk-competitors/references/honesty-protocol.md +72 -0
  6. package/skills/sk-competitors/references/research-principles.md +54 -0
  7. package/skills/sk-competitors/references/research-scaling.md +106 -0
  8. package/skills/sk-competitors/references/research-synthesis.md +237 -0
  9. package/skills/sk-competitors/references/research-wave-1-profiles-pricing.md +186 -0
  10. package/skills/sk-competitors/references/research-wave-2-sentiment-mining.md +189 -0
  11. package/skills/sk-competitors/references/research-wave-3-gtm-signals.md +192 -0
  12. package/skills/sk-competitors/references/verification-agent.md +126 -0
  13. package/skills/sk-export/SKILL.md +36 -12
  14. package/skills/sk-leads/SKILL.md +9 -8
  15. package/skills/sk-money/SKILL.md +7 -6
  16. package/skills/sk-niche/SKILL.md +3 -3
  17. package/skills/sk-offer/SKILL.md +15 -6
  18. package/skills/sk-pitch/SKILL.md +461 -0
  19. package/skills/sk-pitch/references/honesty-protocol.md +62 -0
  20. package/skills/sk-pitch/references/pitch-frameworks.md +261 -0
  21. package/skills/sk-pitch/references/research-principles.md +64 -0
  22. package/skills/sk-pitch/references/research-scaling.md +96 -0
  23. package/skills/sk-pitch/references/research-synthesis.md +423 -0
  24. package/skills/sk-pitch/references/research-wave-1-audience-narrative.md +164 -0
  25. package/skills/sk-pitch/references/research-wave-2-competitive-framing.md +159 -0
  26. package/skills/sk-pitch/references/verification-agent.md +129 -0
  27. package/skills/sk-positioning/SKILL.md +318 -0
  28. package/skills/sk-positioning/references/frameworks.md +132 -0
  29. package/skills/sk-positioning/references/honesty-protocol.md +72 -0
  30. package/skills/sk-positioning/references/research-principles.md +64 -0
  31. package/skills/sk-positioning/references/research-scaling.md +96 -0
  32. package/skills/sk-positioning/references/research-synthesis.md +419 -0
  33. package/skills/sk-positioning/references/research-wave-1-alternatives.md +236 -0
  34. package/skills/sk-positioning/references/research-wave-2-market-frame.md +208 -0
  35. package/skills/sk-positioning/references/verification-agent.md +128 -0
  36. package/skills/sk-skills/SKILL.md +9 -8
  37. package/skills/sk-validate/SKILL.md +8 -6
  38. package/skills/startupkit/SKILL.md +39 -17
  39. package/templates/competitors-template.md +43 -0
  40. package/templates/pitch-template.md +48 -0
  41. package/templates/positioning-template.md +51 -0
  42. package/templates/session-template.md +26 -7
@@ -0,0 +1,106 @@
1
+ # Research Scaling Protocol
2
+
3
+ Dynamic scaling adjusts research depth based on market complexity and user preference. Evaluated after intake, before research begins.
4
+
5
+ ## Complexity Score
6
+
7
+ Assess three factors from the intake data:
8
+
9
+ | Factor | Low (1) | Medium (2) | High (3) |
10
+ |--------|---------|------------|----------|
11
+ | **Market breadth** | Ultra-niche, few players, well-defined segment | Defined market, moderate competition | Broad market, many segments, diverse players |
12
+ | **Known competitors** | 0-2 identified | 3-5 identified | 6+ identified |
13
+ | **Geographic scope** | Single country | Regional (e.g., Europe, North America) | Global or multi-region |
14
+
15
+ **Complexity score** = sum of the three factors (range: 3-9)
16
+
17
+ ## Research Depth Tiers
18
+
19
+ | Tier | Score Range | Manual Trigger | Description |
20
+ |------|-----------|----------------|-------------|
21
+ | **Light** | 3-4 | User says "light", "quick", or "fast research" | Quick scan, fewer agents, 2-3 search rounds |
22
+ | **Standard** | 5-7 | Default (no override needed) | Current behavior, balanced depth |
23
+ | **Deep** | 8-9 | User says "deep", "thorough", or "deep research" | More agents, 5-6 search rounds, extra coverage |
24
+
25
+ **Manual override always wins.** If the user requests "light" on a score-9 market, use Light. If they request "deep" on a score-3 market, use Deep.
26
+
27
+ ## User Communication
28
+
29
+ After calculating the score, show this to the user:
30
+
31
+ ```
32
+ ## Research Depth
33
+
34
+ Based on your intake, I've assessed the research complexity:
35
+
36
+ | Factor | Assessment | Score |
37
+ |------------------|---------------------|-------|
38
+ | Market breadth | {description} | {1-3} |
39
+ | Known competitors| {N} identified | {1-3} |
40
+ | Geographic scope | {description} | {1-3} |
41
+
42
+ **Complexity score: {X}/9 — recommended depth: {Light/Standard/Deep}**
43
+
44
+ You can override this. Here's what each depth means:
45
+
46
+ | Depth | Agents | Searches per agent | Best for |
47
+ |--------------|--------|--------------------|-----------------------------------------------|
48
+ | **Light** | {N} | 2-3 rounds | Quick scan, niche markets, time-sensitive decisions |
49
+ | **Standard** | {N} | 3-4 rounds | Most cases, balanced depth vs. speed |
50
+ | **Deep** | {N} | 5-6 rounds | Crowded markets, high-stakes decisions, thorough due diligence |
51
+
52
+ → Type **light**, **deep**, or **ok** to accept the recommendation.
53
+ ```
54
+
55
+ The agent counts shown should reflect the actual numbers for this skill (see Wave Configuration below).
56
+
57
+ ## Wave Configuration: startup-competitors
58
+
59
+ ### Light (3-4 score or user override)
60
+
61
+ **Wave 1: Competitor Profiles + Pricing** (1 agent)
62
+ - A1: Competitor Profiles & Pricing (merge A1+A2 into one agent, cover profiles and pricing together)
63
+
64
+ **Wave 2: Customer Sentiment** (1 agent)
65
+ - B1: Review & Community Mining (merge B1+B2 into one agent, cover reviews and forums together)
66
+
67
+ **Wave 3: GTM & Strategic Signals** (1 agent)
68
+ - C1: GTM & Growth Signals (merge C1+C2 into one agent, cover GTM and strategic signals together)
69
+
70
+ **Total: 3 agents** (vs. 6 Standard), 2-3 search rounds per agent
71
+
72
+ ### Standard (5-7 score, default)
73
+
74
+ No changes to current wave structure:
75
+ - Wave 1: 2 agents (A1, A2)
76
+ - Wave 2: 2 agents (B1, B2)
77
+ - Wave 3: 2 agents (C1, C2)
78
+
79
+ **Total: 6 agents**, 3-4 search rounds per agent
80
+
81
+ ### Deep (8-9 score or user override)
82
+
83
+ **Wave 1: Competitor Profiles + Pricing** (3 agents)
84
+ - A1: Competitor Deep-Dives (unchanged)
85
+ - A2: Pricing Intelligence (unchanged)
86
+ - A3: Adjacent Competitor Profiles (NEW: profile 3-5 adjacent/emerging competitors not in the direct set, including recent launches and stealth startups)
87
+
88
+ **Wave 2: Customer Sentiment** (3 agents)
89
+ - B1: Review Mining (unchanged)
90
+ - B2: Forum & Community Mining (unchanged)
91
+ - B3: Social Media Sentiment (NEW: mine Twitter/X, LinkedIn, and YouTube for competitor mentions, sentiment patterns, and influencer opinions)
92
+
93
+ **Wave 3: GTM & Strategic Signals** (3 agents)
94
+ - C1: Go-to-Market Analysis (unchanged)
95
+ - C2: Strategic & Growth Signals (unchanged)
96
+ - C3: Tech Stack & Product Analysis (NEW: analyze competitors' technology choices, API ecosystems, integration depth, and technical moats)
97
+
98
+ **Total: 9 agents**, 5-6 search rounds per agent
99
+
100
+ ## PROGRESS.md
101
+
102
+ Record the selected tier in PROGRESS.md:
103
+
104
+ ```markdown
105
+ - **Research Depth:** {Light/Standard/Deep} (score: {X}/9, {override: user request / auto})
106
+ ```
@@ -0,0 +1,237 @@
1
+ # Synthesis & Battle Cards
2
+
3
+ After ALL waves complete (6 agents), synthesize the raw findings into polished deliverables. This step creates the real analytical value — it connects dots across data sources to surface opportunities that individual research pieces can't reveal on their own.
4
+
5
+ ## Synthesis Protocol
6
+
7
+ ### Before Writing
8
+
9
+ 1. Read ALL raw files in `{project-name}/raw/` before writing anything
10
+ 2. Look for patterns across sources — what themes repeat?
11
+ 3. Identify contradictions between sources and explain which you trust more
12
+ 4. Connect the dots: pricing gaps + customer complaints + hiring signals = opportunities
13
+
14
+ ### Cross-Wave Connections to Look For
15
+
16
+ These are the high-value insights that come from combining data:
17
+
18
+ - **Complaint + Pricing = Opportunity:** Customers complain about a feature AND the competitor charges a premium for it → undercut with better value
19
+ - **Hiring + Product Direction = Threat:** Competitor hiring AI engineers + recent AI mentions in changelog → they're coming for that space
20
+ - **Churn Signal + Switching Cost = Wedge:** People want to leave but data portability is hard → build easy migration as a differentiator
21
+ - **Content Gap + Search Volume = Quick Win:** Nobody ranks for a high-volume term → own it early
22
+ - **Review Pattern + Missing Feature = MVP Feature:** Multiple competitors lack something customers want → build it first
23
+ - **Funding + Team Size = Reality Check:** Well-funded competitor with 100+ engineers → don't compete on features, compete on focus
24
+
25
+ ### Confidence Rating
26
+
27
+ Rate every major claim:
28
+ - **High:** Multiple Tier 1/2 sources agree, recent data
29
+ - **Medium:** Some evidence but gaps, or sources partially disagree
30
+ - **Low:** Limited data, mostly inferred, or data older than 12 months
31
+
32
+ ---
33
+
34
+ ## Output File: competitors-report.md
35
+
36
+ Structure:
37
+
38
+ ```markdown
39
+ # Competitive Intelligence Report: {market/product}
40
+ *Skill: startup-competitors | Generated: {date}*
41
+
42
+ ## Executive Summary
43
+ {5 sentences max: market concentration, key finding, biggest opportunity, biggest risk, overall assessment}
44
+
45
+ ## Market Concentration
46
+ - **Structure:** fragmented / consolidating / dominated
47
+ - **Number of active players:** {count}
48
+ - **Funding concentration:** {is money flowing in or drying up?}
49
+ - **Entry barriers:** low / medium / high — {why}
50
+
51
+ ## Key Players at a Glance
52
+ | Competitor | Stage | Funding | Strength | Weakness | Threat |
53
+ |-----------|-------|---------|----------|----------|--------|
54
+ | ... | ... | ... | ... | ... | H/M/L |
55
+
56
+ ## Adjacent Solutions & Substitutes
57
+ {Broader platforms, manual alternatives, and tools from neighboring categories that compete for the same budget or job. Include: what job they solve, why buyers consider them, and how they compare to direct competitors.}
58
+
59
+ ## Strategic Opportunities
60
+ For each opportunity:
61
+ ### Opportunity: {name}
62
+ - **What:** {description}
63
+ - **Evidence:** {data points from research}
64
+ - **Confidence:** High / Medium / Low
65
+ - **How to exploit:** {specific recommendation}
66
+
67
+ ## Strategic Risks
68
+ For each risk:
69
+ ### Risk: {name}
70
+ - **What:** {description}
71
+ - **Evidence:** {data points}
72
+ - **Severity:** High / Medium / Low
73
+ - **Mitigation:** {how to protect against it}
74
+
75
+ ## Competitive Moat Assessment
76
+ Evaluate the market on 5 moat dimensions:
77
+ | Moat Type | Present in Market? | Who Has It | Strength |
78
+ |----------|-------------------|-----------|----------|
79
+ | Network effects | yes / no | {who} | strong / weak |
80
+ | Switching costs | yes / no | {who} | strong / weak |
81
+ | Data moat | yes / no | {who} | strong / weak |
82
+ | Brand/trust | yes / no | {who} | strong / weak |
83
+ | Economies of scale | yes / no | {who} | strong / weak |
84
+
85
+ {Paragraph explaining what this means for a new entrant}
86
+
87
+ ## Data Gaps & Research Limitations
88
+ Aggregate all data gaps from raw research files into a single section. For each gap:
89
+ - What data is missing
90
+ - Why it matters for decision-making
91
+ - How to fill it (specific actions the founder can take)
92
+
93
+ This section is mandatory — every competitive analysis has blind spots. Being explicit about them builds trust and prevents false confidence.
94
+
95
+ ## Red Flags
96
+ - {flag 1 — things that should worry the founder}
97
+ - {flag 2}
98
+
99
+ ## Yellow Flags
100
+ - {flag 1 — things to watch}
101
+ - {flag 2}
102
+ ```
103
+
104
+ ---
105
+
106
+ ## Output File: competitive-matrix.md
107
+
108
+ ```markdown
109
+ # Competitive Feature Matrix: {market}
110
+
111
+ ## Feature Comparison
112
+ | Feature | {Your Product} | {Comp 1} | {Comp 2} | {Comp 3} | {Comp 4} |
113
+ |---------|---------------|----------|----------|----------|----------|
114
+ | {feature 1} | {rating} | {rating} | ... | ... | ... |
115
+ | {feature 2} | ... | ... | ... | ... | ... |
116
+
117
+ Rating scale: Strong / Adequate / Weak / Missing / Unknown
118
+
119
+ ## Gap Analysis
120
+ Features where no competitor excels (all Weak or Missing):
121
+ - {gap 1} — {opportunity implication}
122
+ - {gap 2} — {opportunity implication}
123
+
124
+ ## Differentiation Opportunities
125
+ Based on the matrix, the clearest paths to differentiation:
126
+ 1. {opportunity — what to build and why}
127
+ 2. {opportunity}
128
+ 3. {opportunity}
129
+ ```
130
+
131
+ ---
132
+
133
+ ## Output File: pricing-landscape.md
134
+
135
+ ```markdown
136
+ # Pricing Landscape: {market}
137
+
138
+ ## Market Pricing Overview
139
+ - **Dominant value metric:** {what most charge for}
140
+ - **Price range:** {lowest — highest for comparable tiers}
141
+ - **Median price point:** {for standard tier}
142
+ - **Free tier prevalence:** {X of Y competitors offer free}
143
+
144
+ ## Tier-by-Tier Comparison
145
+ | | {Comp 1} | {Comp 2} | {Comp 3} | {Comp 4} |
146
+ |---|----------|----------|----------|----------|
147
+ | Free tier | {what's included} | ... | ... | ... |
148
+ | Entry tier | ${price} — {limits} | ... | ... | ... |
149
+ | Mid tier | ${price} — {limits} | ... | ... | ... |
150
+ | Top tier | ${price} — {limits} | ... | ... | ... |
151
+ | Enterprise | {custom?} | ... | ... | ... |
152
+
153
+ ## Value Metric Analysis
154
+ | Competitor | Value Metric | Why It Works/Doesn't | Impact on Scaling |
155
+ |-----------|-------------|---------------------|-------------------|
156
+ | ... | per seat | {analysis} | {how costs grow} |
157
+
158
+ ## Pricing Psychology in Use
159
+ | Tactic | Used By | How |
160
+ |--------|---------|-----|
161
+ | Anchoring | {who} | {details} |
162
+ | Decoy tier | {who} | {details} |
163
+ | Charm pricing | {who} | {details} |
164
+ | Annual lock-in | {who} | {discount %} |
165
+
166
+ ## Switching Cost Matrix
167
+ | Competitor | Technical Cost | Contractual Cost | Emotional Cost | Overall |
168
+ |-----------|---------------|-----------------|----------------|---------|
169
+ | ... | H/M/L | H/M/L | H/M/L | H/M/L |
170
+
171
+ ## Pricing Whitespace
172
+ {Where there's room to position on price — underserved segments, untried models, price points nobody occupies}
173
+
174
+ ## Recommendations
175
+ - **If competing on price:** {strategy}
176
+ - **If competing on value:** {strategy}
177
+ - **If competing on model:** {alternative pricing approach that no competitor uses}
178
+ ```
179
+
180
+ ---
181
+
182
+ ## Post-Synthesis Verification
183
+
184
+ After writing all deliverables and battle cards, run the Verification Agent protocol. See `references/verification-agent.md` for the full process. The verification step checks all deliverables for unlabeled claims, internal contradictions, confidence rating consistency, and startup-competitors-specific coherence (battle card vs. report consistency, matrix vs. profiles alignment, pricing landscape vs. profiles consistency, cross-deliverable opportunity/risk traceability).
185
+
186
+ ---
187
+
188
+ ## Output File: battle-cards/{competitor-name}.md
189
+
190
+ One battle card per competitor. Keep each to ONE page — these are reference tools for quick use, not deep research docs.
191
+
192
+ ```markdown
193
+ # Battle Card: {Competitor Name}
194
+ *Last updated: {date}*
195
+
196
+ ## At a Glance
197
+ - **What they do:** {one sentence}
198
+ - **Founded:** {year} | **Funding:** {total} | **Team:** ~{size}
199
+ - **Price:** {range} | **Model:** {value metric}
200
+ - **Best for:** {their ideal customer}
201
+
202
+ ## Their Strengths (be honest)
203
+ - {strength 1 — with evidence}
204
+ - {strength 2}
205
+ - {strength 3}
206
+
207
+ ## Their Weaknesses (your openings)
208
+ - {weakness 1 — with evidence from reviews/forums}
209
+ - {weakness 2}
210
+ - {weakness 3}
211
+
212
+ ## How to Win Against Them
213
+ Specific talking points when a prospect is evaluating both:
214
+ - **When they say "{objection},"** respond: {counter with evidence}
215
+ - **When they say "{objection},"** respond: {counter}
216
+ - **Lead with:** {your strongest differentiator vs. this specific competitor}
217
+
218
+ ## When They Win Over You
219
+ Be honest about when the competitor is the better choice:
220
+ - {scenario 1 — e.g., "Enterprise teams needing SSO and audit logs"}
221
+ - {scenario 2}
222
+
223
+ ## Their Customers' Top Complaint
224
+ "{verbatim quote from review}" — {source}
225
+ This matters because: {strategic implication}
226
+
227
+ ## Key Vulnerability
228
+ {The single biggest weakness you can exploit — with evidence}
229
+
230
+ ## Churn Signals
231
+ Why their customers leave:
232
+ - {reason 1} — frequency: common / occasional
233
+ - {reason 2}
234
+
235
+ ## Watch For
236
+ {What this competitor is likely to do next based on strategic signals}
237
+ ```
@@ -0,0 +1,186 @@
1
+ # Wave 1: Competitor Profiles + Pricing Intelligence
2
+
3
+ Read `research-principles.md` first.
4
+
5
+ ---
6
+
7
+ ## Agent A1: Competitor Deep-Dives
8
+
9
+ ```
10
+ Research task: Deep analysis of direct competitors for {product description}
11
+ Context: {product summary from intake}
12
+ Known competitors: {list from intake, if any}
13
+
14
+ RESEARCH PROTOCOL — identify and profile 5-8 direct competitors:
15
+
16
+ ROUND 1 — Identify competitors (4-5 searches):
17
+ - "{problem} software/app/tool {current year}"
18
+ - "best {product category} tools {current year}"
19
+ - "{known competitor 1} vs alternatives"
20
+ - "G2 {product category} grid"
21
+ - "{product category} Product Hunt"
22
+ - "top {product category} startups"
23
+ - "{problem} solutions" OR "how do {customer type} currently handle {problem}"
24
+ (this catches adjacent solutions — inventory tools, platforms with overlapping features, manual/offline alternatives that compete for the same budget)
25
+
26
+ ROUND 2 — Deep-dive each competitor (2-3 searches per competitor):
27
+ - Visit their website: capture positioning, features, messaging, social proof
28
+ - "{competitor name} review G2 Capterra"
29
+ - "{competitor name} crunchbase funding"
30
+ - "{competitor name} linkedin employees" (team size signals)
31
+ - "{competitor name} changelog" or "{competitor name} updates {current year}"
32
+
33
+ ROUND 3 — Competitive dynamics (2-3 searches):
34
+ - "{product category} market share"
35
+ - "{competitor 1} vs {competitor 2}" comparison articles
36
+ - "{product category} landscape {current year}"
37
+
38
+ For EACH competitor, build a complete profile:
39
+
40
+ ## {Competitor Name}
41
+ - **Website:** {url}
42
+ - **Founded:** {year}
43
+ - **Headquarters:** {location}
44
+ - **Team size:** {estimate from LinkedIn/Crunchbase}
45
+ - **Funding:** {total raised, last round, lead investors}
46
+ - **Stage:** bootstrapped / seed / Series A / Series B+ / public
47
+ - **Estimated revenue:** {if available, or proxy estimate}
48
+
49
+ ### Product
50
+ - **Tagline:** {their actual tagline/positioning statement}
51
+ - **Core offering:** {what they sell in one sentence}
52
+ - **Key features:** {top 5-8 features}
53
+ - **Tech stack signals:** {any public info}
54
+ - **Integrations:** {key integrations}
55
+ - **Platform:** {web / mobile / desktop / API}
56
+
57
+ ### Market Position
58
+ - **Target customer:** {who they serve — be specific}
59
+ - **Positioning:** {how they describe themselves}
60
+ - **Key differentiator:** {what they claim makes them unique}
61
+ - **Social proof:** {notable customers, case studies, logos}
62
+
63
+ ### Traction Signals
64
+ - **G2/Capterra:** {review count and average rating}
65
+ - **Product Hunt:** {launch date, upvotes}
66
+ - **Social media:** {follower counts, engagement level}
67
+ - **Job postings:** {number and type}
68
+ - **Web traffic signals:** {if available from Similarweb/press mentions}
69
+ - **Notable customers:** {logos or case studies}
70
+
71
+ ### Strengths
72
+ - {strength 1 — based on evidence, not speculation}
73
+ - {strength 2}
74
+ - {strength 3}
75
+
76
+ ### Weaknesses
77
+ - {weakness 1 — based on reviews, gaps, complaints}
78
+ - {weakness 2}
79
+ - {weakness 3}
80
+
81
+ ### Threat Level: Low / Medium / High
82
+ - {why — with evidence}
83
+
84
+ ---
85
+
86
+ After all profiles:
87
+
88
+ ## Landscape Summary
89
+ - **Total competitors identified:** {number profiled + number found but not profiled}
90
+ - **Market concentration:** fragmented / consolidating / dominated by 1-2 players
91
+ - **Average funding level:** {across profiled competitors}
92
+ - **Common positioning themes:** {what most competitors emphasize}
93
+ - **Gaps in the market:** {what no competitor does well}
94
+
95
+ ## Adjacent Solutions
96
+ Products that aren't direct competitors but compete for the same budget or solve an overlapping problem. These matter because customers often choose "good enough" adjacent tools over a dedicated solution.
97
+ - {adjacent solution 1} — what it does, how it overlaps, why someone might pick it instead
98
+ - {adjacent solution 2}
99
+ Include: broader platforms with partial feature overlap, manual/offline alternatives, tools from adjacent categories that could expand into this space.
100
+
101
+ ## Data Gaps
102
+ - [What you couldn't find and why it matters]
103
+
104
+ Save to: {project-name}/raw/competitor-profiles.md
105
+ ```
106
+
107
+ ---
108
+
109
+ ## Agent A2: Pricing Intelligence
110
+
111
+ ```
112
+ Research task: Pricing reverse-engineering for competitors in {product category}
113
+ Context: {product summary from intake}
114
+ Competitors to analyze: {list from A1 if available, otherwise discover during research}
115
+
116
+ RESEARCH PROTOCOL:
117
+
118
+ ROUND 1 — Capture pricing pages (1 search per competitor):
119
+ - Visit each competitor's pricing page directly
120
+ - Screenshot or capture: tiers, prices, feature lists, CTAs
121
+ - Note: annual vs monthly pricing, currency, any free tier
122
+
123
+ ROUND 2 — Deep pricing analysis (2-3 searches):
124
+ - "{competitor name} pricing" (for third-party breakdowns)
125
+ - "{competitor name} pricing changes" (for pricing history)
126
+ - "{product category} pricing comparison {current year}"
127
+ - "{competitor name} enterprise pricing" (often hidden)
128
+
129
+ ROUND 3 — Value metric analysis (1-2 searches):
130
+ - "{product category} pricing model" (per-seat vs usage vs flat)
131
+ - "how much does {competitor name} cost" (real user discussions)
132
+
133
+ For EACH competitor, analyze:
134
+
135
+ ## {Competitor Name} — Pricing Breakdown
136
+
137
+ ### Pricing Model
138
+ - **Value metric:** {what they charge for — per seat / per usage / flat / hybrid}
139
+ - **Why this metric:** {how it aligns with value delivered}
140
+ - **How it scales:** {does price grow linearly with usage? Are there volume discounts?}
141
+
142
+ ### Tier Structure
143
+ | | {Tier 1} | {Tier 2} | {Tier 3} | {Enterprise} |
144
+ |---|----------|----------|----------|-------------|
145
+ | Price (monthly) | | | | |
146
+ | Price (annual) | | | | |
147
+ | Annual discount | | | | |
148
+ | {Key feature 1} | | | | |
149
+ | {Key feature 2} | | | | |
150
+ | {Key feature 3} | | | | |
151
+ | {Key limit 1} | | | | |
152
+ | Target persona | | | | |
153
+
154
+ ### Pricing Psychology
155
+ - **Anchoring:** {do they use a high-price tier to make mid-tier attractive?}
156
+ - **Decoy effect:** {is there a tier designed to push people to a specific plan?}
157
+ - **Charm pricing:** {$49 vs $50? $99 vs $100?}
158
+ - **Social proof on pricing:** {which tier is "most popular"?}
159
+ - **Free tier strategy:** {what's free and what's gated?}
160
+ - **Annual lock-in:** {discount size, refund policy}
161
+
162
+ ### Switching Costs
163
+ - **Technical:** {data export? API migration? Integration rewiring?}
164
+ - **Contractual:** {annual contracts? Cancellation penalties?}
165
+ - **Emotional:** {brand loyalty? Learning curve for alternatives?}
166
+ - **Data portability:** {can you export your data easily?}
167
+
168
+ ---
169
+
170
+ After all competitors:
171
+
172
+ ## Pricing Landscape Summary
173
+ - **Dominant value metric:** {what most charge for}
174
+ - **Price range:** {lowest to highest for comparable tiers}
175
+ - **Median price point:** {for the most common tier}
176
+ - **Free tier prevalence:** {how many offer free plans}
177
+ - **Annual discount range:** {typical discounts}
178
+ - **Pricing whitespace:** {where there's room to position — underserved price points or models}
179
+ - **Switching cost patterns:** {are switching costs high or low in this market?}
180
+
181
+ ## Data Gaps
182
+ - [Competitors with hidden/custom pricing]
183
+ - [Enterprise pricing that couldn't be found]
184
+
185
+ Save to: {project-name}/raw/pricing-intelligence.md
186
+ ```