@productbrain/mcp 0.0.1-beta.51 → 0.0.1-beta.52

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2949,7 +2949,7 @@ var IMPLEMENTATION_REVIEW_WORKFLOW = {
2949
2949
  - \`verify\` \u2014 glossary code mappings, cross-references vs codebase
2950
2950
  - \`review-against-rules\` (domain) \u2014 business rule compliance
2951
2951
  - \`context action=gather task="implementation review for [BET-ID]"\` \u2014 related knowledge
2952
- - Context7 \`query-docs\` \u2014 Vitest, Convex, Node.js best practices (resolve-library-id first)
2952
+ - Context7 \`query-docs\` \u2014 relevant testing and framework best practices (resolve-library-id first)
2953
2953
  - \`session-wrapup\` \u2014 before close
2954
2954
 
2955
2955
  ## Sub-Agent Classification (for test failures)
@@ -2990,7 +2990,7 @@ When reviewing test results: (a) test staleness \u2014 code changed, test not up
2990
2990
  label: "Code & Test Honesty",
2991
2991
  type: "open",
2992
2992
  instruction: "Spawn sub-agents to review implementation and tests. Did we meet high standards? Are tests validating real behavior or did we edit them just to pass?",
2993
- facilitatorGuidance: "Spawn 1\u20132 sub-agents in parallel: (1) explore agent \u2014 code review, architecture boundaries, type-safety. (2) generalPurpose agent \u2014 test review: are tests asserting real behavior or were they edited to pass? Use Context7 (query-docs) for Vitest/Convex best practices if needed. Classify any test failures: staleness, regression, or flaky. Synthesize: implementation grade, test honesty verdict, refactoring suggestions.",
2993
+ facilitatorGuidance: "Spawn 1\u20132 sub-agents in parallel: (1) explore agent \u2014 code review, architecture boundaries, type-safety. (2) generalPurpose agent \u2014 test review: are tests asserting real behavior or were they edited to pass? Use Context7 (query-docs) for relevant testing and framework best practices if needed. Classify any test failures: staleness, regression, or flaky. Synthesize: implementation grade, test honesty verdict, refactoring suggestions.",
2994
2994
  outputSchema: {
2995
2995
  field: "codeAndTests",
2996
2996
  description: "Implementation grade, test honesty, refactor suggestions",
@@ -3045,7 +3045,7 @@ When reviewing test results: (a) test staleness \u2014 code changed, test not up
3045
3045
  kbOutputCollection: "insights",
3046
3046
  kbOutputTemplate: {
3047
3047
  nameTemplate: "Implementation Review: {scope} \u2014 {date}",
3048
- descriptionField: "rationale"
3048
+ descriptionField: "description"
3049
3049
  },
3050
3050
  errorRecovery: `If anything goes wrong during implementation review:
3051
3051
 
@@ -3088,11 +3088,7 @@ var PROCESS_CHANGE_WORKFLOW = {
3088
3088
  - **Conversational, not ceremonial.** No round headers needed unless the user wants structure.
3089
3089
  - **One question at a time.** Ask what you need, capture, move on.
3090
3090
  - **If the user gives rich context upfront**, skip straight to capture and linking. Don't ask questions they already answered.
3091
- - **Show what you captured.** Always confirm what went to the Chain before closing.
3092
-
3093
- ## Error Recovery
3094
-
3095
- If any tool fails: explain briefly, capture in conversation, offer to retry later. The workflow must never block the user's real work.`,
3091
+ - **Show what you captured.** Always confirm what went to the Chain before closing.`,
3096
3092
  rounds: [
3097
3093
  {
3098
3094
  id: "orient",
@@ -11049,4 +11045,4 @@ export {
11049
11045
  SERVER_VERSION,
11050
11046
  createProductBrainServer
11051
11047
  };
11052
- //# sourceMappingURL=chunk-LHDBWH6M.js.map
11048
+ //# sourceMappingURL=chunk-SCSEEERT.js.map