@sylphx/flow 2.10.0 → 2.12.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (27) hide show
  1. package/CHANGELOG.md +12 -0
  2. package/assets/slash-commands/review-account-security.md +19 -29
  3. package/assets/slash-commands/review-admin.md +21 -32
  4. package/assets/slash-commands/review-auth.md +19 -25
  5. package/assets/slash-commands/review-billing.md +21 -25
  6. package/assets/slash-commands/review-code-quality.md +29 -49
  7. package/assets/slash-commands/review-data-architecture.md +26 -18
  8. package/assets/slash-commands/review-database.md +22 -21
  9. package/assets/slash-commands/review-delivery.md +25 -50
  10. package/assets/slash-commands/review-discovery.md +17 -28
  11. package/assets/slash-commands/review-growth.md +18 -40
  12. package/assets/slash-commands/review-i18n.md +18 -27
  13. package/assets/slash-commands/review-ledger.md +23 -20
  14. package/assets/slash-commands/review-observability.md +27 -41
  15. package/assets/slash-commands/review-operability.md +20 -32
  16. package/assets/slash-commands/review-performance.md +19 -34
  17. package/assets/slash-commands/review-pricing.md +19 -27
  18. package/assets/slash-commands/review-privacy.md +23 -28
  19. package/assets/slash-commands/review-pwa.md +22 -33
  20. package/assets/slash-commands/review-referral.md +27 -40
  21. package/assets/slash-commands/review-security.md +25 -32
  22. package/assets/slash-commands/review-seo.md +26 -46
  23. package/assets/slash-commands/review-storage.md +21 -21
  24. package/assets/slash-commands/review-support.md +27 -41
  25. package/assets/slash-commands/review-trust-safety.md +42 -0
  26. package/assets/slash-commands/review-uiux.md +25 -42
  27. package/package.json +1 -1
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-discovery
3
- description: Review discovery - competitive research, features, pricing opportunities
3
+ description: Review discovery - competitive research, opportunities, market positioning
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -8,42 +8,31 @@ agent: coder
8
8
 
9
9
  ## Mandate
10
10
 
11
- * Perform a **deep, thorough review** to discover opportunities in this codebase.
11
+ * Perform a **deep, thorough review** to discover opportunities for this product.
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **This review IS exploration** — think broadly and creatively about what could be.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Review Requirements: Explore Beyond the Spec
19
+ * Research across the product's full stack as needed
19
20
 
20
- * **Feature design review**: define success criteria, journeys, state model, auth/entitlements, instrumentation; propose competitiveness improvements within constraints.
21
- * **Pricing/monetization review**: packaging/entitlements, lifecycle semantics, legal/tax/invoice implications; propose conversion/churn improvements within constraints.
22
- * **Competitive research**: features, extensibility, guidance patterns, pricing/packaging norms; convert insights into testable acceptance criteria.
21
+ ## Non-Negotiables
23
22
 
24
- ### Discovery Areas
23
+ * None — this is pure exploration
25
24
 
26
- * What features are competitors offering that we lack?
27
- * What pricing models are common in the market?
28
- * What UX patterns are users expecting?
29
- * What integrations would add value?
30
- * What automation opportunities exist?
31
- * What self-service capabilities are missing?
25
+ ## Context
32
26
 
33
- ### Analysis Framework
27
+ Discovery is about finding what's missing, what's possible, and what would make the product significantly more competitive. It's not about fixing bugs — it's about identifying opportunities that don't yet exist.
34
28
 
35
- * Market positioning
36
- * Feature gap analysis
37
- * Pricing benchmarking
38
- * User journey optimization
39
- * Technical debt opportunities
40
- * Scalability considerations
29
+ Look at competitors, market trends, user expectations, and technological possibilities. What would make users choose this product over alternatives? What capabilities would unlock new business models?
41
30
 
42
- ## Verification Checklist
31
+ ## Driving Questions
43
32
 
44
- - [ ] Competitor analysis completed
45
- - [ ] Feature gaps identified
46
- - [ ] Pricing reviewed
47
- - [ ] UX patterns researched
48
- - [ ] Integration opportunities listed
49
- - [ ] Recommendations prioritized
33
+ * What are competitors doing that we're not?
34
+ * What do users expect based on industry standards that we lack?
35
+ * What integrations would add significant value?
36
+ * What pricing models are common in the market and how do we compare?
37
+ * What technical capabilities could enable new business models?
38
+ * What would make this product a category leader rather than a follower?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-growth
3
- description: Review growth - onboarding, viral mechanics, retention
3
+ description: Review growth - activation, retention, virality
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,51 +12,29 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify growth opportunities that don't yet exist.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Growth System (Onboarding, Share/Viral, Retention)
19
+ * **Analytics**: PostHog
20
+ * **Framework**: Next.js
19
21
 
20
- * The review must produce a coherent, measurable growth system for activation, sharing/virality, and retention, aligned with compliance and anti-abuse constraints.
22
+ ## Non-Negotiables
21
23
 
22
- ### Onboarding
24
+ * Sharing/virality mechanics must be consent-aware
25
+ * Growth instrumentation must not violate privacy constraints
23
26
 
24
- * Onboarding must be:
25
- * Outcome-oriented
26
- * Localized
27
- * Accessible
28
- * Instrumented
27
+ ## Context
29
28
 
30
- ### Sharing/Virality
29
+ Growth isn't about tricks — it's about removing friction from value delivery. Users who quickly experience value stay; users who don't, leave.
31
30
 
32
- * Sharing/virality must be:
33
- * Consent-aware
34
- * Abuse-resistant
35
- * Measurable end-to-end
31
+ The review should consider: what's preventing users from reaching their "aha moment" faster? What would make them want to share? What brings them back? These aren't features to add — they're fundamental product questions.
36
32
 
37
- ### Retention
33
+ ## Driving Questions
38
34
 
39
- * Retention must be:
40
- * Intentionally engineered
41
- * Monitored
42
- * Protected against regressions
43
-
44
- ### Growth Best Practices
45
-
46
- * Clear value proposition on landing
47
- * Frictionless signup
48
- * Time-to-value optimization
49
- * Activation milestones defined
50
- * Re-engagement campaigns
51
- * Churn prediction/prevention
52
- * NPS/satisfaction tracking
53
-
54
- ## Verification Checklist
55
-
56
- - [ ] Onboarding flow exists
57
- - [ ] Onboarding instrumented
58
- - [ ] Sharing mechanisms work
59
- - [ ] Sharing is consent-aware
60
- - [ ] Retention metrics tracked
61
- - [ ] Re-engagement exists
62
- - [ ] Growth metrics dashboarded
35
+ * Where do users drop off before experiencing value?
36
+ * What would cut time-to-value in half?
37
+ * Why would a user tell someone else about this product?
38
+ * What brings users back after their first session?
39
+ * What signals predict churn before it happens?
40
+ * What would a 10x better onboarding look like?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-i18n
3
- description: Review i18n - locales, routing, canonicalization, UGC
3
+ description: Review i18n - localization, routing, translation quality
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,39 +12,30 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify what would make the product feel native to each locale.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Supported Locales
19
+ * **i18n**: next-intl
20
+ * **Framework**: Next.js
19
21
 
20
- `en`, `zh-Hans`, `zh-Hant`, `es`, `ja`, `ko`, `de`, `fr`, `pt-BR`, `it`, `nl`, `pl`, `tr`, `id`, `th`, `vi`
22
+ ## Non-Negotiables
21
23
 
22
- ### URL Strategy: Prefix Except Default
23
-
24
- * English is default and non-prefixed.
25
- * `/en/*` must not exist; permanently redirect to non-prefixed equivalent.
26
- * All non-default locales are `/<locale>/...`.
27
-
28
- ### Globalization Rules
29
-
30
- * Intl formatting for dates, numbers, currency
31
- * Explicit fallback rules
24
+ * `/en/*` must not exist (permanently redirect to non-prefixed)
32
25
  * Missing translation keys must fail build
33
26
  * No hardcoded user-facing strings outside localization
34
27
 
35
- ### UGC Canonicalization
28
+ ## Context
29
+
30
+ Internationalization isn't just translation — it's making the product feel native to each market. Bad i18n is obvious to users and signals that they're second-class citizens. Good i18n is invisible.
36
31
 
37
- * Separate UI language from content language.
38
- * Exactly one canonical URL per UGC resource determined by content language.
39
- * No indexable locale-prefixed duplicates unless primary content is truly localized; otherwise redirect to canonical.
40
- * Canonical/hreflang/sitemap must reflect only true localized variants.
32
+ Consider: dates, numbers, currency, pluralization, text direction, cultural norms. Does the product feel like it was built for each locale, or does it feel like a translation of an English product?
41
33
 
42
- ## Verification Checklist
34
+ ## Driving Questions
43
35
 
44
- - [ ] All locales supported
45
- - [ ] `/en/*` returns 301 redirect
46
- - [ ] Missing keys fail build
47
- - [ ] No hardcoded strings
48
- - [ ] Intl formatting used
49
- - [ ] UGC has single canonical URL
50
- - [ ] hreflang tags correct
36
+ * What would make the product feel native to a non-English user?
37
+ * Where do translations feel awkward or machine-generated?
38
+ * What cultural assumptions are baked into the UX that don't translate?
39
+ * How painful is the translation workflow for adding new strings?
40
+ * What locales are we missing that represent real market opportunity?
41
+ * Where do we fall back to English in ways users would notice?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-ledger
3
- description: Review ledger - financial-grade balance system, immutable ledger
3
+ description: Review ledger - balance systems, financial integrity, reconciliation
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,29 +12,32 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify financial integrity risks before they become real problems.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Financial-Grade Balance System (Only if "balance/credits/wallet" exists)
19
+ * **Payments**: Stripe
20
+ * **Database**: Neon (Postgres)
21
+ * **ORM**: Drizzle
19
22
 
20
- * Any balance concept must be implemented as an **immutable ledger** (append-only source of truth), not a mutable balance field.
21
- * Deterministic precision (no floats), idempotent posting, concurrency safety, transactional integrity, and auditability are required.
22
- * Monetary flows must be currency-based and reconcilable with Stripe; credits (if used) must be governed as non-cash entitlements.
23
+ ## Non-Negotiables
23
24
 
24
- ### Ledger Requirements
25
+ * Balances must be immutable ledger (append-only), not mutable fields
26
+ * No floating-point for money (use deterministic precision)
27
+ * All financial mutations must be idempotent
28
+ * Monetary flows must be reconcilable with Stripe
25
29
 
26
- * Append-only transaction log
27
- * Double-entry bookkeeping where applicable
28
- * Idempotent transaction posting (idempotency keys)
29
- * Concurrency-safe balance calculations
30
- * Audit trail for all mutations
31
- * Reconciliation with external systems (Stripe)
30
+ ## Context
32
31
 
33
- ## Verification Checklist
32
+ Financial systems are unforgiving. A bug that creates or destroys money — even briefly — is a serious incident. Users trust us with their money; that trust is easily lost and hard to regain.
34
33
 
35
- - [ ] Immutable ledger pattern used (not mutable balance)
36
- - [ ] No floating point for money
37
- - [ ] Idempotent posting implemented
38
- - [ ] Concurrency safety verified
39
- - [ ] Full audit trail exists
40
- - [ ] Reconciliation with Stripe possible
34
+ If balance/credits/wallet exists, it must be bulletproof. If it doesn't exist yet, consider whether the current design would support adding it correctly. Retrofitting financial integrity is painful.
35
+
36
+ ## Driving Questions
37
+
38
+ * Does a balance/credits system exist, and is it implemented correctly?
39
+ * Where could money be created or destroyed by a bug?
40
+ * What happens during concurrent transactions?
41
+ * How would we detect if balances drifted from reality?
42
+ * Can we prove every balance by replaying the ledger?
43
+ * What financial edge cases (refunds, disputes, chargebacks) aren't handled?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-observability
3
- description: Review observability - logs, Sentry, correlation IDs, alerting
3
+ description: Review observability - logging, tracing, alerting, debugging
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,44 +12,30 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify the production issues we can't debug today.
15
16
 
16
- ## Review Scope
17
-
18
- ### Observability and Alerting (Mandatory)
19
-
20
- * Structured logs and correlation IDs must exist end-to-end (request/job/webhook) with consistent traceability
21
- * Define critical-path SLO/SLI posture
22
- * Define actionable alerts for:
23
- * Webhook failures
24
- * Ledger/entitlement drift
25
- * Authentication attacks
26
- * Abuse spikes
27
- * Drift detection
28
-
29
- ### Logging Best Practices
30
-
31
- * Structured JSON logs
32
- * Correlation ID in all logs
33
- * Request ID propagation
34
- * User context (anonymized where needed)
35
- * Error stack traces
36
- * Performance timing
37
- * No PII in logs
38
-
39
- ### Monitoring
40
-
41
- * Sentry for error tracking
42
- * PostHog for product analytics
43
- * Server metrics (latency, throughput, errors)
44
- * Database query performance
45
- * External service health
46
-
47
- ## Verification Checklist
48
-
49
- - [ ] Structured logging implemented
50
- - [ ] Correlation IDs propagate
51
- - [ ] Sentry configured
52
- - [ ] SLO/SLI defined
53
- - [ ] Alerts for critical failures
54
- - [ ] No PII in logs
55
- - [ ] Dashboards for key metrics
17
+ ## Tech Stack
18
+
19
+ * **Error Tracking**: Sentry
20
+ * **Analytics**: PostHog
21
+ * **Platform**: Vercel
22
+
23
+ ## Non-Negotiables
24
+
25
+ * Correlation IDs must exist end-to-end (request → job → webhook)
26
+ * Alerts must exist for critical failures (webhook failures, auth attacks, drift)
27
+
28
+ ## Context
29
+
30
+ Observability is about answering questions when things go wrong. It's 3am, something is broken, users are complaining — can you figure out what happened? How fast?
31
+
32
+ Good observability makes debugging easy. Bad observability means you're guessing, adding log lines, redeploying, and hoping. Consider: what questions would you need to answer during an incident, and can you answer them today?
33
+
34
+ ## Driving Questions
35
+
36
+ * If something breaks in production right now, how would we find out?
37
+ * What blind spots exist where errors go unnoticed?
38
+ * How long would it take to trace a user's request through the entire system?
39
+ * What alerts exist, and do they fire for the right things?
40
+ * Where do we have noise that's training people to ignore alerts?
41
+ * What production issue in the last month was hard to debug, and why?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-operability
3
- description: Review operability - async workflows, DLQ, retries, drift detection
3
+ description: Review operability - workflows, retries, DLQ, incident response
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,43 +12,31 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify what will break at 3am and how we'd fix it.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Async/Workflows Governance (Hard Requirement)
19
+ * **Workflows**: Upstash Workflows + QStash
20
+ * **Cache**: Upstash Redis
21
+ * **Platform**: Vercel
19
22
 
20
- * Define idempotency and deduplication posture
21
- * Define controlled retries/backoff
22
- * **Dead-letter handling must exist and be observable and operable**
23
- * **Safe replay must be supported**
24
- * Side-effects (email/billing/ledger/entitlements) must be governed such that they are either proven effectively-once or safely re-entrant without duplicated economic/security consequences
23
+ ## Non-Negotiables
25
24
 
26
- ### Drift Detection (Hard Requirement)
25
+ * Dead-letter handling must exist and be operable (visible, replayable)
26
+ * Side-effects (email, billing, ledger) must be idempotent or safely re-entrant
27
+ * Drift alerts must have remediation playbooks
27
28
 
28
- * Drift alerts must have a defined remediation playbook (automated fix or operator workflow)
29
- * Each remediation must be auditable and support post-incident traceability
29
+ ## Context
30
30
 
31
- ### Release Safety
31
+ Operability is about running the system in production — not just building it. Systems fail. Jobs get stuck. State drifts. The question is: when something goes wrong, can an operator fix it without deploying code?
32
32
 
33
- * Define safe rollout posture with backward compatibility
34
- * Rollback expectations for billing/ledger/auth changes
33
+ Consider the operator experience during an incident. What tools do they have? What runbooks exist? Can they safely retry failed jobs? Can they detect and fix drift?
35
34
 
36
- ### Operability Best Practices
35
+ ## Driving Questions
37
36
 
38
- * Health check endpoints
39
- * Graceful shutdown
40
- * Circuit breakers for external services
41
- * Timeout configuration
42
- * Retry with exponential backoff
43
- * Bulk operation controls
44
- * Maintenance mode capability
45
-
46
- ## Verification Checklist
47
-
48
- - [ ] Async jobs idempotent
49
- - [ ] DLQ exists and observable
50
- - [ ] Safe replay supported
51
- - [ ] Drift detection implemented
52
- - [ ] Remediation playbooks exist
53
- - [ ] Rollback strategy defined
54
- - [ ] Health checks present
37
+ * What happens when a job fails permanently?
38
+ * How would an operator know something is stuck?
39
+ * Can failed workflows be safely replayed without duplicating side-effects?
40
+ * What drift can occur between systems, and how would we detect it?
41
+ * What's the rollback plan if a deploy breaks something critical?
42
+ * What runbooks exist, and what runbooks should exist but don't?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-performance
3
- description: Review performance - budgets, Core Web Vitals, caching
3
+ description: Review performance - speed, Core Web Vitals, bottlenecks
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,45 +12,30 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify what's making the product feel slow.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Performance Requirements
19
+ * **Framework**: Next.js (SSR/ISR/Static)
20
+ * **Platform**: Vercel
21
+ * **Tooling**: Bun
19
22
 
20
- * Performance must be **measurable and regression-resistant**:
21
- * Define and enforce performance budgets for key journeys
22
- * Define caching boundaries and correctness requirements across SSR/ISR/static and service worker behavior
23
- * Monitor Core Web Vitals and server latency
24
- * Alert on regressions
23
+ ## Non-Negotiables
25
24
 
26
- ### Performance Budget Verification
25
+ * Core Web Vitals must meet thresholds (LCP < 2.5s, CLS < 0.1, INP < 200ms)
26
+ * Performance regressions must be detectable
27
27
 
28
- * **Performance budget verification** for key journeys (including Core Web Vitals-related thresholds) with release-blocking regression detection
28
+ ## Context
29
29
 
30
- ### Core Web Vitals
30
+ Performance is a feature. Slow products feel broken, even when they're correct. Users don't read loading spinners — they leave. Every 100ms of latency costs engagement.
31
31
 
32
- * LCP (Largest Contentful Paint) < 2.5s
33
- * FID (First Input Delay) < 100ms
34
- * CLS (Cumulative Layout Shift) < 0.1
35
- * INP (Interaction to Next Paint) < 200ms
32
+ Don't just measure understand. Where does time go? What's blocking the critical path? What would make the product feel instant? Sometimes small architectural changes have bigger impact than optimization.
36
33
 
37
- ### Performance Best Practices
34
+ ## Driving Questions
38
35
 
39
- * Image optimization (WebP, lazy loading)
40
- * Code splitting
41
- * Tree shaking
42
- * Bundle size monitoring
43
- * Font optimization
44
- * Critical CSS
45
- * Preloading key resources
46
- * Server response time < 200ms
47
-
48
- ## Verification Checklist
49
-
50
- - [ ] Performance budgets defined
51
- - [ ] Core Web Vitals within targets
52
- - [ ] Regression detection active
53
- - [ ] Caching boundaries defined
54
- - [ ] Images optimized
55
- - [ ] Bundle size acceptable
56
- - [ ] No render-blocking resources
36
+ * What makes the product feel slow to users?
37
+ * Where are the biggest bottlenecks in the critical user journeys?
38
+ * What's in the critical rendering path that shouldn't be?
39
+ * How large is the JavaScript bundle, and what's bloating it?
40
+ * What database queries are slow, and why?
41
+ * If we could make one thing 10x faster, what would have the most impact?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-pricing
3
- description: Review pricing - pricing governance, grandfathering, migrations
3
+ description: Review pricing - strategy, packaging, monetization
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -8,41 +8,33 @@ agent: coder
8
8
 
9
9
  ## Mandate
10
10
 
11
- * Perform a **deep, thorough review** of pricing governance in this codebase.
11
+ * Perform a **deep, thorough review** of pricing in this codebase.
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify monetization opportunities and pricing friction.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Stripe Pricing Governance (Stripe-first, not Dashboard-first)
19
+ * **Payments**: Stripe
19
20
 
20
- * Stripe is the system-of-record for products, prices, subscriptions, invoices, and disputes; internal systems must not contradict Stripe truth.
21
- * Pricing changes must be performed by creating new Stripe Prices and updating the "active sellable price" policy; historical prices must remain immutable for existing subscriptions unless an approved migration is executed.
22
- * Default pricing change policy is **grandfathering**: existing subscribers keep their current price; new customers use the currently active sellable price.
21
+ ## Non-Negotiables
23
22
 
24
- ### Pricing Admin Requirements
23
+ * Stripe is system-of-record; internal systems must not contradict Stripe truth
24
+ * Pricing changes must create new Stripe Prices (historical prices immutable)
25
+ * Non-admin Stripe Dashboard changes must be detectable (drift)
25
26
 
26
- * An operational-grade Pricing Admin must exist to manage:
27
- * Creation of new Stripe Prices
28
- * Activation/deactivation of sellable prices
29
- * Controlled bulk subscription migrations (optional)
30
- * All actions must be governed by RBAC, step-up controls, and audit logs.
31
- * Stripe Dashboard is treated as monitoring/emergency access; non-admin Stripe changes must be detectable (drift), alertable, and remediable.
27
+ ## Context
32
28
 
33
- ### Pricing Strategy
29
+ Pricing is strategy, not just configuration. The right pricing captures value, reduces friction, and aligns incentives. The wrong pricing leaves money on the table or drives users away.
34
30
 
35
- * Clear tier differentiation
36
- * Feature gating by plan
37
- * Upgrade/downgrade paths defined
38
- * Proration handling
39
- * Trial-to-paid conversion flow
31
+ Consider the entire monetization journey: how users discover value, how they decide to pay, how they upgrade/downgrade. Where is there friction? Where are we undercharging? Where are we losing conversions?
40
32
 
41
- ## Verification Checklist
33
+ ## Driving Questions
42
34
 
43
- - [ ] Stripe is system-of-record
44
- - [ ] New prices don't mutate existing
45
- - [ ] Grandfathering policy implemented
46
- - [ ] Pricing admin exists with RBAC
47
- - [ ] Stripe Dashboard drift detectable
48
- - [ ] Upgrade/downgrade paths work
35
+ * How does our pricing compare to competitors?
36
+ * Where do users abandon the upgrade flow?
37
+ * What would make upgrading feel like an obvious decision?
38
+ * How do we communicate value at each pricing tier?
39
+ * What pricing experiments would teach us the most?
40
+ * If we could change one thing about pricing, what would have the biggest impact?
@@ -1,6 +1,6 @@
1
1
  ---
2
2
  name: review-privacy
3
- description: Review privacy - consent, PII handling, data lifecycle, GDPR
3
+ description: Review privacy - consent, PII, data lifecycle, compliance
4
4
  agent: coder
5
5
  ---
6
6
 
@@ -12,39 +12,34 @@ agent: coder
12
12
  * **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
13
13
  * Deliverables must be stated as **findings, gaps, and actionable recommendations**.
14
14
  * **Single-pass delivery**: no deferrals; deliver a complete assessment.
15
+ * **Explore beyond the spec**: identify compliance gaps and privacy improvements.
15
16
 
16
- ## Review Scope
17
+ ## Tech Stack
17
18
 
18
- ### Consent Governance (Release-Blocking)
19
+ * **Analytics**: PostHog
20
+ * **Email**: Resend
21
+ * **Tag Management**: GTM (marketing only)
22
+ * **Observability**: Sentry
19
23
 
20
- * Analytics (PostHog) and marketing/newsletter communications (Resend) must be governed by consent and user preferences.
21
- * Marketing tags (including GTM and Google Ads) must not load or fire without the appropriate consent.
22
- * Without consent, tracking and marketing sends must not occur, except for strictly necessary service communications.
23
- * Event schemas and attributes must follow data minimization, with explicit PII classification and handling rules.
24
+ ## Non-Negotiables
24
25
 
25
- ### PII and Sensitive Data Controls (Hard Requirement)
26
+ * Analytics and marketing must not fire before user consent
27
+ * PII must not leak into logs, Sentry, PostHog, or third-party services
28
+ * Account deletion must propagate to all third-party processors
29
+ * Marketing tags (GTM, Google Ads) must not load without consent
30
+ * Conversion tracking must be server-truth aligned, idempotent, and deduplicated
26
31
 
27
- * PII rules apply to logs, Sentry, PostHog, support tooling, email systems, and marketing tags/conversion payloads.
28
- * A consistent scrubbing/redaction standard must exist, and must be covered by automated tests to prevent leakage to third parties.
32
+ ## Context
29
33
 
30
- ### Data Lifecycle
34
+ Privacy isn't just compliance — it's trust. Users share data expecting it to be handled responsibly. Every log line, every analytics event, every third-party integration is a potential privacy leak.
31
35
 
32
- * Define deletion/deactivation semantics
33
- * Deletion propagation to third parties
34
- * Export where applicable
35
- * DR/backup posture aligned with retention
36
- * **Define data classification, retention periods, deletion propagation to third-party processors, and explicit exceptions** (legal/tax/anti-fraud)
37
- * Ensure external disclosures match behavior
36
+ The review should verify that actual behavior matches stated policy. If the privacy policy says "we don't track without consent," does the code actually enforce that? Mismatches are not just bugs — they're trust violations.
38
37
 
39
- ### Behavioral Consistency
38
+ ## Driving Questions
40
39
 
41
- * **Behavioral consistency is required**: policy and disclosures must match actual behavior across UI, data handling, logging/observability, analytics, support operations, and marketing tags; mismatches are release-blocking.
42
-
43
- ## Verification Checklist
44
-
45
- - [ ] Consent gates analytics/marketing
46
- - [ ] No tracking without consent
47
- - [ ] PII scrubbed from logs/Sentry/PostHog
48
- - [ ] Data deletion flow works
49
- - [ ] Deletion propagates to third parties
50
- - [ ] Privacy policy matches actual behavior
40
+ * Does the consent implementation actually block tracking, or just record preference?
41
+ * Where does PII leak that we haven't noticed?
42
+ * If a user requests data deletion, what actually gets deleted vs. retained?
43
+ * Does the privacy policy accurately reflect what the code actually does?
44
+ * How would we handle a GDPR data subject access request today?
45
+ * What data are we collecting that we don't actually need?