@sylphx/flow 2.10.0 → 2.12.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/assets/slash-commands/review-account-security.md +19 -29
- package/assets/slash-commands/review-admin.md +21 -32
- package/assets/slash-commands/review-auth.md +19 -25
- package/assets/slash-commands/review-billing.md +21 -25
- package/assets/slash-commands/review-code-quality.md +29 -49
- package/assets/slash-commands/review-data-architecture.md +26 -18
- package/assets/slash-commands/review-database.md +22 -21
- package/assets/slash-commands/review-delivery.md +25 -50
- package/assets/slash-commands/review-discovery.md +17 -28
- package/assets/slash-commands/review-growth.md +18 -40
- package/assets/slash-commands/review-i18n.md +18 -27
- package/assets/slash-commands/review-ledger.md +23 -20
- package/assets/slash-commands/review-observability.md +27 -41
- package/assets/slash-commands/review-operability.md +20 -32
- package/assets/slash-commands/review-performance.md +19 -34
- package/assets/slash-commands/review-pricing.md +19 -27
- package/assets/slash-commands/review-privacy.md +23 -28
- package/assets/slash-commands/review-pwa.md +22 -33
- package/assets/slash-commands/review-referral.md +27 -40
- package/assets/slash-commands/review-security.md +25 -32
- package/assets/slash-commands/review-seo.md +26 -46
- package/assets/slash-commands/review-storage.md +21 -21
- package/assets/slash-commands/review-support.md +27 -41
- package/assets/slash-commands/review-trust-safety.md +42 -0
- package/assets/slash-commands/review-uiux.md +25 -42
- package/package.json +1 -1
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-discovery
|
|
3
|
-
description: Review discovery - competitive research,
|
|
3
|
+
description: Review discovery - competitive research, opportunities, market positioning
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -8,42 +8,31 @@ agent: coder
|
|
|
8
8
|
|
|
9
9
|
## Mandate
|
|
10
10
|
|
|
11
|
-
* Perform a **deep, thorough review** to discover opportunities
|
|
11
|
+
* Perform a **deep, thorough review** to discover opportunities for this product.
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **This review IS exploration** — think broadly and creatively about what could be.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* Research across the product's full stack as needed
|
|
19
20
|
|
|
20
|
-
|
|
21
|
-
* **Pricing/monetization review**: packaging/entitlements, lifecycle semantics, legal/tax/invoice implications; propose conversion/churn improvements within constraints.
|
|
22
|
-
* **Competitive research**: features, extensibility, guidance patterns, pricing/packaging norms; convert insights into testable acceptance criteria.
|
|
21
|
+
## Non-Negotiables
|
|
23
22
|
|
|
24
|
-
|
|
23
|
+
* None — this is pure exploration
|
|
25
24
|
|
|
26
|
-
|
|
27
|
-
* What pricing models are common in the market?
|
|
28
|
-
* What UX patterns are users expecting?
|
|
29
|
-
* What integrations would add value?
|
|
30
|
-
* What automation opportunities exist?
|
|
31
|
-
* What self-service capabilities are missing?
|
|
25
|
+
## Context
|
|
32
26
|
|
|
33
|
-
|
|
27
|
+
Discovery is about finding what's missing, what's possible, and what would make the product significantly more competitive. It's not about fixing bugs — it's about identifying opportunities that don't yet exist.
|
|
34
28
|
|
|
35
|
-
|
|
36
|
-
* Feature gap analysis
|
|
37
|
-
* Pricing benchmarking
|
|
38
|
-
* User journey optimization
|
|
39
|
-
* Technical debt opportunities
|
|
40
|
-
* Scalability considerations
|
|
29
|
+
Look at competitors, market trends, user expectations, and technological possibilities. What would make users choose this product over alternatives? What capabilities would unlock new business models?
|
|
41
30
|
|
|
42
|
-
##
|
|
31
|
+
## Driving Questions
|
|
43
32
|
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
33
|
+
* What are competitors doing that we're not?
|
|
34
|
+
* What do users expect based on industry standards that we lack?
|
|
35
|
+
* What integrations would add significant value?
|
|
36
|
+
* What pricing models are common in the market and how do we compare?
|
|
37
|
+
* What technical capabilities could enable new business models?
|
|
38
|
+
* What would make this product a category leader rather than a follower?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-growth
|
|
3
|
-
description: Review growth -
|
|
3
|
+
description: Review growth - activation, retention, virality
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,51 +12,29 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify growth opportunities that don't yet exist.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Analytics**: PostHog
|
|
20
|
+
* **Framework**: Next.js
|
|
19
21
|
|
|
20
|
-
|
|
22
|
+
## Non-Negotiables
|
|
21
23
|
|
|
22
|
-
|
|
24
|
+
* Sharing/virality mechanics must be consent-aware
|
|
25
|
+
* Growth instrumentation must not violate privacy constraints
|
|
23
26
|
|
|
24
|
-
|
|
25
|
-
* Outcome-oriented
|
|
26
|
-
* Localized
|
|
27
|
-
* Accessible
|
|
28
|
-
* Instrumented
|
|
27
|
+
## Context
|
|
29
28
|
|
|
30
|
-
|
|
29
|
+
Growth isn't about tricks — it's about removing friction from value delivery. Users who quickly experience value stay; users who don't, leave.
|
|
31
30
|
|
|
32
|
-
|
|
33
|
-
* Consent-aware
|
|
34
|
-
* Abuse-resistant
|
|
35
|
-
* Measurable end-to-end
|
|
31
|
+
The review should consider: what's preventing users from reaching their "aha moment" faster? What would make them want to share? What brings them back? These aren't features to add — they're fundamental product questions.
|
|
36
32
|
|
|
37
|
-
|
|
33
|
+
## Driving Questions
|
|
38
34
|
|
|
39
|
-
*
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
* Clear value proposition on landing
|
|
47
|
-
* Frictionless signup
|
|
48
|
-
* Time-to-value optimization
|
|
49
|
-
* Activation milestones defined
|
|
50
|
-
* Re-engagement campaigns
|
|
51
|
-
* Churn prediction/prevention
|
|
52
|
-
* NPS/satisfaction tracking
|
|
53
|
-
|
|
54
|
-
## Verification Checklist
|
|
55
|
-
|
|
56
|
-
- [ ] Onboarding flow exists
|
|
57
|
-
- [ ] Onboarding instrumented
|
|
58
|
-
- [ ] Sharing mechanisms work
|
|
59
|
-
- [ ] Sharing is consent-aware
|
|
60
|
-
- [ ] Retention metrics tracked
|
|
61
|
-
- [ ] Re-engagement exists
|
|
62
|
-
- [ ] Growth metrics dashboarded
|
|
35
|
+
* Where do users drop off before experiencing value?
|
|
36
|
+
* What would cut time-to-value in half?
|
|
37
|
+
* Why would a user tell someone else about this product?
|
|
38
|
+
* What brings users back after their first session?
|
|
39
|
+
* What signals predict churn before it happens?
|
|
40
|
+
* What would a 10x better onboarding look like?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-i18n
|
|
3
|
-
description: Review i18n -
|
|
3
|
+
description: Review i18n - localization, routing, translation quality
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,39 +12,30 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify what would make the product feel native to each locale.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **i18n**: next-intl
|
|
20
|
+
* **Framework**: Next.js
|
|
19
21
|
|
|
20
|
-
|
|
22
|
+
## Non-Negotiables
|
|
21
23
|
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
* English is default and non-prefixed.
|
|
25
|
-
* `/en/*` must not exist; permanently redirect to non-prefixed equivalent.
|
|
26
|
-
* All non-default locales are `/<locale>/...`.
|
|
27
|
-
|
|
28
|
-
### Globalization Rules
|
|
29
|
-
|
|
30
|
-
* Intl formatting for dates, numbers, currency
|
|
31
|
-
* Explicit fallback rules
|
|
24
|
+
* `/en/*` must not exist (permanently redirect to non-prefixed)
|
|
32
25
|
* Missing translation keys must fail build
|
|
33
26
|
* No hardcoded user-facing strings outside localization
|
|
34
27
|
|
|
35
|
-
|
|
28
|
+
## Context
|
|
29
|
+
|
|
30
|
+
Internationalization isn't just translation — it's making the product feel native to each market. Bad i18n is obvious to users and signals that they're second-class citizens. Good i18n is invisible.
|
|
36
31
|
|
|
37
|
-
|
|
38
|
-
* Exactly one canonical URL per UGC resource determined by content language.
|
|
39
|
-
* No indexable locale-prefixed duplicates unless primary content is truly localized; otherwise redirect to canonical.
|
|
40
|
-
* Canonical/hreflang/sitemap must reflect only true localized variants.
|
|
32
|
+
Consider: dates, numbers, currency, pluralization, text direction, cultural norms. Does the product feel like it was built for each locale, or does it feel like a translation of an English product?
|
|
41
33
|
|
|
42
|
-
##
|
|
34
|
+
## Driving Questions
|
|
43
35
|
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
- [ ] hreflang tags correct
|
|
36
|
+
* What would make the product feel native to a non-English user?
|
|
37
|
+
* Where do translations feel awkward or machine-generated?
|
|
38
|
+
* What cultural assumptions are baked into the UX that don't translate?
|
|
39
|
+
* How painful is the translation workflow for adding new strings?
|
|
40
|
+
* What locales are we missing that represent real market opportunity?
|
|
41
|
+
* Where do we fall back to English in ways users would notice?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-ledger
|
|
3
|
-
description: Review ledger -
|
|
3
|
+
description: Review ledger - balance systems, financial integrity, reconciliation
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,29 +12,32 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify financial integrity risks before they become real problems.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Payments**: Stripe
|
|
20
|
+
* **Database**: Neon (Postgres)
|
|
21
|
+
* **ORM**: Drizzle
|
|
19
22
|
|
|
20
|
-
|
|
21
|
-
* Deterministic precision (no floats), idempotent posting, concurrency safety, transactional integrity, and auditability are required.
|
|
22
|
-
* Monetary flows must be currency-based and reconcilable with Stripe; credits (if used) must be governed as non-cash entitlements.
|
|
23
|
+
## Non-Negotiables
|
|
23
24
|
|
|
24
|
-
|
|
25
|
+
* Balances must be immutable ledger (append-only), not mutable fields
|
|
26
|
+
* No floating-point for money (use deterministic precision)
|
|
27
|
+
* All financial mutations must be idempotent
|
|
28
|
+
* Monetary flows must be reconcilable with Stripe
|
|
25
29
|
|
|
26
|
-
|
|
27
|
-
* Double-entry bookkeeping where applicable
|
|
28
|
-
* Idempotent transaction posting (idempotency keys)
|
|
29
|
-
* Concurrency-safe balance calculations
|
|
30
|
-
* Audit trail for all mutations
|
|
31
|
-
* Reconciliation with external systems (Stripe)
|
|
30
|
+
## Context
|
|
32
31
|
|
|
33
|
-
|
|
32
|
+
Financial systems are unforgiving. A bug that creates or destroys money — even briefly — is a serious incident. Users trust us with their money; that trust is easily lost and hard to regain.
|
|
34
33
|
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
34
|
+
If balance/credits/wallet exists, it must be bulletproof. If it doesn't exist yet, consider whether the current design would support adding it correctly. Retrofitting financial integrity is painful.
|
|
35
|
+
|
|
36
|
+
## Driving Questions
|
|
37
|
+
|
|
38
|
+
* Does a balance/credits system exist, and is it implemented correctly?
|
|
39
|
+
* Where could money be created or destroyed by a bug?
|
|
40
|
+
* What happens during concurrent transactions?
|
|
41
|
+
* How would we detect if balances drifted from reality?
|
|
42
|
+
* Can we prove every balance by replaying the ledger?
|
|
43
|
+
* What financial edge cases (refunds, disputes, chargebacks) aren't handled?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-observability
|
|
3
|
-
description: Review observability -
|
|
3
|
+
description: Review observability - logging, tracing, alerting, debugging
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,44 +12,30 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify the production issues we can't debug today.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
*
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
*
|
|
36
|
-
*
|
|
37
|
-
*
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
* Sentry for error tracking
|
|
42
|
-
* PostHog for product analytics
|
|
43
|
-
* Server metrics (latency, throughput, errors)
|
|
44
|
-
* Database query performance
|
|
45
|
-
* External service health
|
|
46
|
-
|
|
47
|
-
## Verification Checklist
|
|
48
|
-
|
|
49
|
-
- [ ] Structured logging implemented
|
|
50
|
-
- [ ] Correlation IDs propagate
|
|
51
|
-
- [ ] Sentry configured
|
|
52
|
-
- [ ] SLO/SLI defined
|
|
53
|
-
- [ ] Alerts for critical failures
|
|
54
|
-
- [ ] No PII in logs
|
|
55
|
-
- [ ] Dashboards for key metrics
|
|
17
|
+
## Tech Stack
|
|
18
|
+
|
|
19
|
+
* **Error Tracking**: Sentry
|
|
20
|
+
* **Analytics**: PostHog
|
|
21
|
+
* **Platform**: Vercel
|
|
22
|
+
|
|
23
|
+
## Non-Negotiables
|
|
24
|
+
|
|
25
|
+
* Correlation IDs must exist end-to-end (request → job → webhook)
|
|
26
|
+
* Alerts must exist for critical failures (webhook failures, auth attacks, drift)
|
|
27
|
+
|
|
28
|
+
## Context
|
|
29
|
+
|
|
30
|
+
Observability is about answering questions when things go wrong. It's 3am, something is broken, users are complaining — can you figure out what happened? How fast?
|
|
31
|
+
|
|
32
|
+
Good observability makes debugging easy. Bad observability means you're guessing, adding log lines, redeploying, and hoping. Consider: what questions would you need to answer during an incident, and can you answer them today?
|
|
33
|
+
|
|
34
|
+
## Driving Questions
|
|
35
|
+
|
|
36
|
+
* If something breaks in production right now, how would we find out?
|
|
37
|
+
* What blind spots exist where errors go unnoticed?
|
|
38
|
+
* How long would it take to trace a user's request through the entire system?
|
|
39
|
+
* What alerts exist, and do they fire for the right things?
|
|
40
|
+
* Where do we have noise that's training people to ignore alerts?
|
|
41
|
+
* What production issue in the last month was hard to debug, and why?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-operability
|
|
3
|
-
description: Review operability -
|
|
3
|
+
description: Review operability - workflows, retries, DLQ, incident response
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,43 +12,31 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify what will break at 3am and how we'd fix it.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Workflows**: Upstash Workflows + QStash
|
|
20
|
+
* **Cache**: Upstash Redis
|
|
21
|
+
* **Platform**: Vercel
|
|
19
22
|
|
|
20
|
-
|
|
21
|
-
* Define controlled retries/backoff
|
|
22
|
-
* **Dead-letter handling must exist and be observable and operable**
|
|
23
|
-
* **Safe replay must be supported**
|
|
24
|
-
* Side-effects (email/billing/ledger/entitlements) must be governed such that they are either proven effectively-once or safely re-entrant without duplicated economic/security consequences
|
|
23
|
+
## Non-Negotiables
|
|
25
24
|
|
|
26
|
-
|
|
25
|
+
* Dead-letter handling must exist and be operable (visible, replayable)
|
|
26
|
+
* Side-effects (email, billing, ledger) must be idempotent or safely re-entrant
|
|
27
|
+
* Drift alerts must have remediation playbooks
|
|
27
28
|
|
|
28
|
-
|
|
29
|
-
* Each remediation must be auditable and support post-incident traceability
|
|
29
|
+
## Context
|
|
30
30
|
|
|
31
|
-
|
|
31
|
+
Operability is about running the system in production — not just building it. Systems fail. Jobs get stuck. State drifts. The question is: when something goes wrong, can an operator fix it without deploying code?
|
|
32
32
|
|
|
33
|
-
|
|
34
|
-
* Rollback expectations for billing/ledger/auth changes
|
|
33
|
+
Consider the operator experience during an incident. What tools do they have? What runbooks exist? Can they safely retry failed jobs? Can they detect and fix drift?
|
|
35
34
|
|
|
36
|
-
|
|
35
|
+
## Driving Questions
|
|
37
36
|
|
|
38
|
-
*
|
|
39
|
-
*
|
|
40
|
-
*
|
|
41
|
-
*
|
|
42
|
-
*
|
|
43
|
-
*
|
|
44
|
-
* Maintenance mode capability
|
|
45
|
-
|
|
46
|
-
## Verification Checklist
|
|
47
|
-
|
|
48
|
-
- [ ] Async jobs idempotent
|
|
49
|
-
- [ ] DLQ exists and observable
|
|
50
|
-
- [ ] Safe replay supported
|
|
51
|
-
- [ ] Drift detection implemented
|
|
52
|
-
- [ ] Remediation playbooks exist
|
|
53
|
-
- [ ] Rollback strategy defined
|
|
54
|
-
- [ ] Health checks present
|
|
37
|
+
* What happens when a job fails permanently?
|
|
38
|
+
* How would an operator know something is stuck?
|
|
39
|
+
* Can failed workflows be safely replayed without duplicating side-effects?
|
|
40
|
+
* What drift can occur between systems, and how would we detect it?
|
|
41
|
+
* What's the rollback plan if a deploy breaks something critical?
|
|
42
|
+
* What runbooks exist, and what runbooks should exist but don't?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-performance
|
|
3
|
-
description: Review performance -
|
|
3
|
+
description: Review performance - speed, Core Web Vitals, bottlenecks
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,45 +12,30 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify what's making the product feel slow.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Framework**: Next.js (SSR/ISR/Static)
|
|
20
|
+
* **Platform**: Vercel
|
|
21
|
+
* **Tooling**: Bun
|
|
19
22
|
|
|
20
|
-
|
|
21
|
-
* Define and enforce performance budgets for key journeys
|
|
22
|
-
* Define caching boundaries and correctness requirements across SSR/ISR/static and service worker behavior
|
|
23
|
-
* Monitor Core Web Vitals and server latency
|
|
24
|
-
* Alert on regressions
|
|
23
|
+
## Non-Negotiables
|
|
25
24
|
|
|
26
|
-
|
|
25
|
+
* Core Web Vitals must meet thresholds (LCP < 2.5s, CLS < 0.1, INP < 200ms)
|
|
26
|
+
* Performance regressions must be detectable
|
|
27
27
|
|
|
28
|
-
|
|
28
|
+
## Context
|
|
29
29
|
|
|
30
|
-
|
|
30
|
+
Performance is a feature. Slow products feel broken, even when they're correct. Users don't read loading spinners — they leave. Every 100ms of latency costs engagement.
|
|
31
31
|
|
|
32
|
-
|
|
33
|
-
* FID (First Input Delay) < 100ms
|
|
34
|
-
* CLS (Cumulative Layout Shift) < 0.1
|
|
35
|
-
* INP (Interaction to Next Paint) < 200ms
|
|
32
|
+
Don't just measure — understand. Where does time go? What's blocking the critical path? What would make the product feel instant? Sometimes small architectural changes have bigger impact than optimization.
|
|
36
33
|
|
|
37
|
-
|
|
34
|
+
## Driving Questions
|
|
38
35
|
|
|
39
|
-
*
|
|
40
|
-
*
|
|
41
|
-
*
|
|
42
|
-
*
|
|
43
|
-
*
|
|
44
|
-
*
|
|
45
|
-
* Preloading key resources
|
|
46
|
-
* Server response time < 200ms
|
|
47
|
-
|
|
48
|
-
## Verification Checklist
|
|
49
|
-
|
|
50
|
-
- [ ] Performance budgets defined
|
|
51
|
-
- [ ] Core Web Vitals within targets
|
|
52
|
-
- [ ] Regression detection active
|
|
53
|
-
- [ ] Caching boundaries defined
|
|
54
|
-
- [ ] Images optimized
|
|
55
|
-
- [ ] Bundle size acceptable
|
|
56
|
-
- [ ] No render-blocking resources
|
|
36
|
+
* What makes the product feel slow to users?
|
|
37
|
+
* Where are the biggest bottlenecks in the critical user journeys?
|
|
38
|
+
* What's in the critical rendering path that shouldn't be?
|
|
39
|
+
* How large is the JavaScript bundle, and what's bloating it?
|
|
40
|
+
* What database queries are slow, and why?
|
|
41
|
+
* If we could make one thing 10x faster, what would have the most impact?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-pricing
|
|
3
|
-
description: Review pricing -
|
|
3
|
+
description: Review pricing - strategy, packaging, monetization
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -8,41 +8,33 @@ agent: coder
|
|
|
8
8
|
|
|
9
9
|
## Mandate
|
|
10
10
|
|
|
11
|
-
* Perform a **deep, thorough review** of pricing
|
|
11
|
+
* Perform a **deep, thorough review** of pricing in this codebase.
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify monetization opportunities and pricing friction.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Payments**: Stripe
|
|
19
20
|
|
|
20
|
-
|
|
21
|
-
* Pricing changes must be performed by creating new Stripe Prices and updating the "active sellable price" policy; historical prices must remain immutable for existing subscriptions unless an approved migration is executed.
|
|
22
|
-
* Default pricing change policy is **grandfathering**: existing subscribers keep their current price; new customers use the currently active sellable price.
|
|
21
|
+
## Non-Negotiables
|
|
23
22
|
|
|
24
|
-
|
|
23
|
+
* Stripe is system-of-record; internal systems must not contradict Stripe truth
|
|
24
|
+
* Pricing changes must create new Stripe Prices (historical prices immutable)
|
|
25
|
+
* Non-admin Stripe Dashboard changes must be detectable (drift)
|
|
25
26
|
|
|
26
|
-
|
|
27
|
-
* Creation of new Stripe Prices
|
|
28
|
-
* Activation/deactivation of sellable prices
|
|
29
|
-
* Controlled bulk subscription migrations (optional)
|
|
30
|
-
* All actions must be governed by RBAC, step-up controls, and audit logs.
|
|
31
|
-
* Stripe Dashboard is treated as monitoring/emergency access; non-admin Stripe changes must be detectable (drift), alertable, and remediable.
|
|
27
|
+
## Context
|
|
32
28
|
|
|
33
|
-
|
|
29
|
+
Pricing is strategy, not just configuration. The right pricing captures value, reduces friction, and aligns incentives. The wrong pricing leaves money on the table or drives users away.
|
|
34
30
|
|
|
35
|
-
|
|
36
|
-
* Feature gating by plan
|
|
37
|
-
* Upgrade/downgrade paths defined
|
|
38
|
-
* Proration handling
|
|
39
|
-
* Trial-to-paid conversion flow
|
|
31
|
+
Consider the entire monetization journey: how users discover value, how they decide to pay, how they upgrade/downgrade. Where is there friction? Where are we undercharging? Where are we losing conversions?
|
|
40
32
|
|
|
41
|
-
##
|
|
33
|
+
## Driving Questions
|
|
42
34
|
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
35
|
+
* How does our pricing compare to competitors?
|
|
36
|
+
* Where do users abandon the upgrade flow?
|
|
37
|
+
* What would make upgrading feel like an obvious decision?
|
|
38
|
+
* How do we communicate value at each pricing tier?
|
|
39
|
+
* What pricing experiments would teach us the most?
|
|
40
|
+
* If we could change one thing about pricing, what would have the biggest impact?
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: review-privacy
|
|
3
|
-
description: Review privacy - consent, PII
|
|
3
|
+
description: Review privacy - consent, PII, data lifecycle, compliance
|
|
4
4
|
agent: coder
|
|
5
5
|
---
|
|
6
6
|
|
|
@@ -12,39 +12,34 @@ agent: coder
|
|
|
12
12
|
* **Delegate to multiple workers** to research different aspects in parallel; you act as the **final gate** to synthesize and verify quality.
|
|
13
13
|
* Deliverables must be stated as **findings, gaps, and actionable recommendations**.
|
|
14
14
|
* **Single-pass delivery**: no deferrals; deliver a complete assessment.
|
|
15
|
+
* **Explore beyond the spec**: identify compliance gaps and privacy improvements.
|
|
15
16
|
|
|
16
|
-
##
|
|
17
|
+
## Tech Stack
|
|
17
18
|
|
|
18
|
-
|
|
19
|
+
* **Analytics**: PostHog
|
|
20
|
+
* **Email**: Resend
|
|
21
|
+
* **Tag Management**: GTM (marketing only)
|
|
22
|
+
* **Observability**: Sentry
|
|
19
23
|
|
|
20
|
-
|
|
21
|
-
* Marketing tags (including GTM and Google Ads) must not load or fire without the appropriate consent.
|
|
22
|
-
* Without consent, tracking and marketing sends must not occur, except for strictly necessary service communications.
|
|
23
|
-
* Event schemas and attributes must follow data minimization, with explicit PII classification and handling rules.
|
|
24
|
+
## Non-Negotiables
|
|
24
25
|
|
|
25
|
-
|
|
26
|
+
* Analytics and marketing must not fire before user consent
|
|
27
|
+
* PII must not leak into logs, Sentry, PostHog, or third-party services
|
|
28
|
+
* Account deletion must propagate to all third-party processors
|
|
29
|
+
* Marketing tags (GTM, Google Ads) must not load without consent
|
|
30
|
+
* Conversion tracking must be server-truth aligned, idempotent, and deduplicated
|
|
26
31
|
|
|
27
|
-
|
|
28
|
-
* A consistent scrubbing/redaction standard must exist, and must be covered by automated tests to prevent leakage to third parties.
|
|
32
|
+
## Context
|
|
29
33
|
|
|
30
|
-
|
|
34
|
+
Privacy isn't just compliance — it's trust. Users share data expecting it to be handled responsibly. Every log line, every analytics event, every third-party integration is a potential privacy leak.
|
|
31
35
|
|
|
32
|
-
|
|
33
|
-
* Deletion propagation to third parties
|
|
34
|
-
* Export where applicable
|
|
35
|
-
* DR/backup posture aligned with retention
|
|
36
|
-
* **Define data classification, retention periods, deletion propagation to third-party processors, and explicit exceptions** (legal/tax/anti-fraud)
|
|
37
|
-
* Ensure external disclosures match behavior
|
|
36
|
+
The review should verify that actual behavior matches stated policy. If the privacy policy says "we don't track without consent," does the code actually enforce that? Mismatches are not just bugs — they're trust violations.
|
|
38
37
|
|
|
39
|
-
|
|
38
|
+
## Driving Questions
|
|
40
39
|
|
|
41
|
-
*
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
- [ ] PII scrubbed from logs/Sentry/PostHog
|
|
48
|
-
- [ ] Data deletion flow works
|
|
49
|
-
- [ ] Deletion propagates to third parties
|
|
50
|
-
- [ ] Privacy policy matches actual behavior
|
|
40
|
+
* Does the consent implementation actually block tracking, or just record preference?
|
|
41
|
+
* Where does PII leak that we haven't noticed?
|
|
42
|
+
* If a user requests data deletion, what actually gets deleted vs. retained?
|
|
43
|
+
* Does the privacy policy accurately reflect what the code actually does?
|
|
44
|
+
* How would we handle a GDPR data subject access request today?
|
|
45
|
+
* What data are we collecting that we don't actually need?
|