opencode-multiagent 0.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (106) hide show
  1. package/LICENSE +21 -0
  2. package/README.md +209 -0
  3. package/agents/advisor.md +57 -0
  4. package/agents/auditor.md +45 -0
  5. package/agents/critic.md +127 -0
  6. package/agents/deep-worker.md +65 -0
  7. package/agents/devil.md +36 -0
  8. package/agents/executor.md +141 -0
  9. package/agents/heavy-worker.md +68 -0
  10. package/agents/lead.md +155 -0
  11. package/agents/librarian.md +62 -0
  12. package/agents/planner.md +121 -0
  13. package/agents/qa.md +50 -0
  14. package/agents/quick.md +65 -0
  15. package/agents/reviewer.md +55 -0
  16. package/agents/scout.md +58 -0
  17. package/agents/scribe.md +78 -0
  18. package/agents/strategist.md +63 -0
  19. package/agents/ui-heavy-worker.md +62 -0
  20. package/agents/ui-worker.md +69 -0
  21. package/agents/validator.md +47 -0
  22. package/agents/worker.md +68 -0
  23. package/commands/execute.md +14 -0
  24. package/commands/init-deep.md +18 -0
  25. package/commands/init.md +18 -0
  26. package/commands/inspect.md +13 -0
  27. package/commands/plan.md +15 -0
  28. package/commands/quality.md +14 -0
  29. package/commands/review.md +14 -0
  30. package/commands/status.md +15 -0
  31. package/defaults/agent-settings.json +102 -0
  32. package/defaults/agent-settings.schema.json +25 -0
  33. package/defaults/flags.json +35 -0
  34. package/defaults/flags.schema.json +119 -0
  35. package/defaults/mcp-defaults.json +47 -0
  36. package/defaults/mcp-defaults.schema.json +38 -0
  37. package/defaults/profiles.json +53 -0
  38. package/defaults/profiles.schema.json +60 -0
  39. package/defaults/team-profiles.json +83 -0
  40. package/examples/opencode.json +4 -0
  41. package/examples/opencode.with-overrides.json +23 -0
  42. package/package.json +62 -0
  43. package/skills/advanced-evaluation/SKILL.md +454 -0
  44. package/skills/advanced-evaluation/manifest.json +20 -0
  45. package/skills/cek-context-engineering/SKILL.md +1261 -0
  46. package/skills/cek-context-engineering/manifest.json +17 -0
  47. package/skills/cek-prompt-engineering/SKILL.md +559 -0
  48. package/skills/cek-prompt-engineering/manifest.json +17 -0
  49. package/skills/cek-test-prompt/SKILL.md +714 -0
  50. package/skills/cek-test-prompt/manifest.json +17 -0
  51. package/skills/cek-thought-based-reasoning/SKILL.md +658 -0
  52. package/skills/cek-thought-based-reasoning/manifest.json +17 -0
  53. package/skills/context-degradation/SKILL.md +231 -0
  54. package/skills/context-degradation/manifest.json +17 -0
  55. package/skills/debate/SKILL.md +316 -0
  56. package/skills/debate/manifest.json +19 -0
  57. package/skills/design-first/SKILL.md +5 -0
  58. package/skills/design-first/manifest.json +20 -0
  59. package/skills/dispatching-parallel-agents/SKILL.md +180 -0
  60. package/skills/dispatching-parallel-agents/manifest.json +18 -0
  61. package/skills/drift-analysis/SKILL.md +324 -0
  62. package/skills/drift-analysis/manifest.json +19 -0
  63. package/skills/evaluation/SKILL.md +5 -0
  64. package/skills/evaluation/manifest.json +19 -0
  65. package/skills/executing-plans/SKILL.md +70 -0
  66. package/skills/executing-plans/manifest.json +17 -0
  67. package/skills/handoff-protocols/SKILL.md +5 -0
  68. package/skills/handoff-protocols/manifest.json +19 -0
  69. package/skills/parallel-investigation/SKILL.md +206 -0
  70. package/skills/parallel-investigation/manifest.json +18 -0
  71. package/skills/reflexion-critique/SKILL.md +477 -0
  72. package/skills/reflexion-critique/manifest.json +17 -0
  73. package/skills/reflexion-reflect/SKILL.md +650 -0
  74. package/skills/reflexion-reflect/manifest.json +17 -0
  75. package/skills/root-cause-analysis/SKILL.md +5 -0
  76. package/skills/root-cause-analysis/manifest.json +20 -0
  77. package/skills/sadd-judge-with-debate/SKILL.md +426 -0
  78. package/skills/sadd-judge-with-debate/manifest.json +17 -0
  79. package/skills/structured-code-review/SKILL.md +5 -0
  80. package/skills/structured-code-review/manifest.json +18 -0
  81. package/skills/task-decomposition/SKILL.md +5 -0
  82. package/skills/task-decomposition/manifest.json +20 -0
  83. package/skills/verification-before-completion/SKILL.md +5 -0
  84. package/skills/verification-before-completion/manifest.json +22 -0
  85. package/skills/verification-gates/SKILL.md +281 -0
  86. package/skills/verification-gates/manifest.json +19 -0
  87. package/src/control-plane.ts +21 -0
  88. package/src/index.ts +8 -0
  89. package/src/opencode-multiagent/compiler.ts +168 -0
  90. package/src/opencode-multiagent/constants.ts +178 -0
  91. package/src/opencode-multiagent/file-lock.ts +90 -0
  92. package/src/opencode-multiagent/hooks.ts +599 -0
  93. package/src/opencode-multiagent/log.ts +12 -0
  94. package/src/opencode-multiagent/mailbox.ts +287 -0
  95. package/src/opencode-multiagent/markdown.ts +99 -0
  96. package/src/opencode-multiagent/mcp.ts +35 -0
  97. package/src/opencode-multiagent/policy.ts +67 -0
  98. package/src/opencode-multiagent/quality.ts +140 -0
  99. package/src/opencode-multiagent/runtime.ts +55 -0
  100. package/src/opencode-multiagent/skills.ts +144 -0
  101. package/src/opencode-multiagent/supervision.ts +156 -0
  102. package/src/opencode-multiagent/task-manager.ts +148 -0
  103. package/src/opencode-multiagent/team-manager.ts +219 -0
  104. package/src/opencode-multiagent/team-tools.ts +359 -0
  105. package/src/opencode-multiagent/telemetry.ts +124 -0
  106. package/src/opencode-multiagent/utils.ts +54 -0
@@ -0,0 +1,426 @@
1
+ ---
2
+ name: sadd-judge-with-debate
3
+ description: Evaluate solutions through multi-round debate between independent judges until consensus
4
+ argument-hint: Solution path(s) and evaluation criteria
5
+ ---
6
+
7
+ # judge-with-debate
8
+
9
+ <task>
10
+ Evaluate solutions through multi-agent debate where independent judges analyze, challenge each other's assessments, and iteratively refine their evaluations until reaching consensus or maximum rounds.
11
+ </task>
12
+
13
+ <context>
14
+ This command implements the Multi-Agent Debate pattern for high-quality evaluation where multiple perspectives and rigorous argumentation improve assessment accuracy. Unlike single-pass evaluation, debate forces judges to defend their positions with evidence and consider counter-arguments.
15
+ </context>
16
+
17
+ ## Pattern: Debate-Based Evaluation
18
+
19
+ This command implements iterative multi-judge debate:
20
+
21
+ ```
22
+ Phase 0: Setup
23
+ mkdir -p .specs/reports
24
+
25
+ Phase 1: Independent Analysis
26
+ ┌─ Judge 1 → {name}.1.md ─┐
27
+ Solution ┼─ Judge 2 → {name}.2.md ─┼─┐
28
+ └─ Judge 3 → {name}.3.md ─┘ │
29
+
30
+ Phase 2: Debate Round (iterative) │
31
+ Each judge reads others' reports │
32
+ ↓ │
33
+ Argue + Defend + Challenge │
34
+ ↓ │
35
+ Revise if convinced ─────────────┤
36
+ ↓ │
37
+ Check consensus │
38
+ ├─ Yes → Final Report │
39
+ └─ No → Next Round ─────────┘
40
+ ```
41
+
42
+ ## Process
43
+
44
+ ### Setup: Create Reports Directory
45
+
46
+ Before starting evaluation, ensure the reports directory exists:
47
+
48
+ ```bash
49
+ mkdir -p .specs/reports
50
+ ```
51
+
52
+ **Report naming convention:** `.specs/reports/{solution-name}-{YYYY-MM-DD}.[1|2|3].md`
53
+
54
+ Where:
55
+ - `{solution-name}` - Derived from solution filename (e.g., `users-api` from `src/api/users.ts`)
56
+ - `{YYYY-MM-DD}` - Current date
57
+ - `[1|2|3]` - Judge number
58
+
59
+ ### Phase 1: Independent Analysis
60
+
61
+ Launch **3 independent judge agents in parallel** (recommended: Opus for rigor):
62
+
63
+ 1. Each judge receives:
64
+ - Path to solution(s) being evaluated
65
+ - Evaluation criteria with weights
66
+ - Clear rubric for scoring
67
+ 2. Each produces **independent assessment** saved to `.specs/reports/{solution-name}-{date}.[1|2|3].md`
68
+ 3. Reports must include:
69
+ - Per-criterion scores with evidence
70
+ - Specific quotes/examples supporting ratings
71
+ - Overall weighted score
72
+ - Key strengths and weaknesses
73
+
74
+ **Key principle:** Independence in initial analysis prevents groupthink.
75
+
76
+ **Prompt template for initial judges:**
77
+
78
+ ```markdown
79
+ You are Judge {N} evaluating a solution independently.
80
+
81
+ <solution_path>
82
+ {path to solution file(s)}
83
+ </solution_path>
84
+
85
+ <task_description>
86
+ {what the solution was supposed to accomplish}
87
+ </task_description>
88
+
89
+ <evaluation_criteria>
90
+ {criteria with descriptions and weights}
91
+ </evaluation_criteria>
92
+
93
+ <output_file>
94
+ .specs/reports/{solution-name}-{date}.{N}.md
95
+ </output_file>
96
+
97
+ Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology and execute using following criteria.
98
+
99
+ Instructions:
100
+ 1. Read the solution thoroughly
101
+ 2. For each criterion:
102
+ - Find specific evidence (quote exact text)
103
+ - Score on the defined scale
104
+ - Justify with concrete examples
105
+ 3. Calculate weighted overall score
106
+ 4. Write comprehensive report to {output_file}
107
+ 5. Generate verification 5 questions about your evaluation.
108
+ 6. Answer verification questions:
109
+ - Re-examine solutions for each question
110
+ - Find counter-evidence if it exists
111
+ - Check for systematic bias (length, confidence, etc.)
112
+ 7. Revise your report file and update it accordingly.
113
+
114
+ Add to report begining `Done by Judge {N}`
115
+ ```
116
+
117
+ ### Phase 2: Debate Rounds (Iterative)
118
+
119
+ For each debate round (max 3 rounds):
120
+
121
+ Launch **3 debate agents in parallel**:
122
+
123
+ 1. Each judge agent receives:
124
+ - Path to their own previous report (`.specs/reports/{solution-name}-{date}.[1|2|3].md`)
125
+ - Paths to other judges' reports (`.specs/reports/{solution-name}-{date}.[1|2|3].md`)
126
+ - The original solution
127
+ 2. Each judge:
128
+ - Identifies disagreements with other judges (>1 point score gap on any criterion)
129
+ - Defends their own ratings with evidence
130
+ - Challenges other judges' ratings they disagree with
131
+ - Considers counter-arguments
132
+ - Revises their assessment if convinced
133
+ 3. Updates their report file with new section: `## Debate Round {R}`
134
+ 4. After they reply, if they reached agreement move to Phase 3: Consensus Report
135
+
136
+ **Key principle:** Judges communicate only through filesystem - orchestrator doesn't mediate and don't read reports files itself, it can overflow your context.
137
+
138
+ **Prompt template for debate judges:**
139
+
140
+ ```markdown
141
+ You are Judge {N} in debate round {R}.
142
+
143
+ <your_previous_report>
144
+ {path to .specs/reports/{solution-name}-{date}.{N}.md}
145
+ </your_previous_report>
146
+
147
+ <other_judges_reports>
148
+ Judge 1: .specs/reports/{solution-name}-{date}.1.md
149
+ ...
150
+ </other_judges_reports>
151
+
152
+ <task_description>
153
+ {what the solution was supposed to accomplish}
154
+ </task_description>
155
+
156
+ <solution_path>
157
+ {path to solution}
158
+ </solution_path>
159
+
160
+ <output_file>
161
+ .specs/reports/{solution-name}-{date}.{N}.md (append to existing file)
162
+ </output_file>
163
+
164
+ Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology principles.
165
+
166
+ Instructions:
167
+ 1. Read your previous assessment from {your_previous_report}
168
+ 2. Read all other judges' reports
169
+ 3. Identify disagreements (where your scores differ by >1 point)
170
+ 4. For each major disagreement:
171
+ - State the disagreement clearly
172
+ - Defend your position with evidence
173
+ - Challenge the other judge's position with counter-evidence
174
+ - Consider whether their evidence changes your view
175
+ 5. Update your report file by APPENDING:
176
+ 6. Reply whether you are reached agreement, and with which judge. Include revisited scores and criteria scores.
177
+
178
+ ---
179
+
180
+ ## Debate Round {R}
181
+
182
+ ### Disagreements Identified
183
+
184
+ **Disagreement with Judge {X} on Criterion "{Name}"**
185
+ - My score: {my_score}/5
186
+ - Their score: {their_score}/5
187
+ - My defense: [quote evidence supporting my score]
188
+ - My challenge: [what did they miss or misinterpret?]
189
+
190
+ [Repeat for each disagreement]
191
+
192
+ ### Revised Assessment
193
+
194
+ After considering other judges' arguments:
195
+ - **Criterion "{Name}"**: [Maintained {X}/5 | Revised from {X} to {Y}/5]
196
+ - Reason for change: [what convinced me] OR
197
+ - Reason maintained: [why I stand by original score]
198
+
199
+ [Repeat for changed/maintained scores]
200
+
201
+ **New Weighted Score**: {updated_total}/5.0
202
+
203
+ ## Evidences
204
+ [specific quotes]
205
+
206
+ ---
207
+
208
+ CRITICAL:
209
+ - Only revise if you find their evidence compelling
210
+ - Defend your original scores if you still believe them
211
+ - Quote specific evidence from the solution
212
+ ```
213
+
214
+ ### Consensus Check
215
+
216
+ After each debate round, check for consensus:
217
+
218
+ **Consensus achieved if:**
219
+ - All judges' overall scores within 0.5 points of each other
220
+ - No criterion has >1 point disagreement across any two judges
221
+ - All judges explicitly state they accept the consensus
222
+
223
+ **If no consensus after 3 rounds:**
224
+ - Report persistent disagreements
225
+ - Provide all judge reports for human review
226
+ - Flag that automated evaluation couldn't reach consensus
227
+
228
+ **Orchestration Instructions:**
229
+
230
+ **Step 1: Run Independent Analysis (Round 1)**
231
+
232
+ 1. Launch 3 judge agents in parallel (Judge 1, 2, 3)
233
+ 2. Each writes their independent assessment to `.specs/reports/{solution-name}-{date}.[1|2|3].md`
234
+ 3. Wait for all 3 agents to complete
235
+
236
+ **Step 2: Check for Consensus**
237
+
238
+ Let's work through this systematically to ensure accurate consensus detection.
239
+
240
+ Read all three reports and extract:
241
+ - Each judge's overall weighted score
242
+ - Each judge's score for every criterion
243
+
244
+ Check consensus step by step:
245
+ 1. First, extract all overall scores from each report and list them explicitly
246
+ 2. Calculate the difference between the highest and lowest overall scores
247
+ - If difference ≤ 0.5 points → overall consensus achieved
248
+ - If difference > 0.5 points → no consensus yet
249
+ 3. Next, for each criterion, list all three judges' scores side by side
250
+ 4. For each criterion, calculate the difference between highest and lowest scores
251
+ - If any criterion has difference > 1.0 point → no consensus on that criterion
252
+ 5. Finally, verify consensus is achieved only if BOTH conditions are met:
253
+ - Overall scores within 0.5 points
254
+ - All criterion scores within 1.0 point
255
+
256
+ **Step 3: Decision Point**
257
+
258
+ - **If consensus achieved**: Go to Step 5 (Generate Consensus Report)
259
+ - **If no consensus AND round < 3**: Go to Step 4 (Run Debate Round)
260
+ - **If no consensus AND round = 3**: Go to Step 6 (Report No Consensus)
261
+
262
+ **Step 4: Run Debate Round**
263
+
264
+ 1. Increment round counter (round = round + 1)
265
+ 2. Launch 3 judge agents in parallel
266
+ 3. Each agent reads:
267
+ - Their own previous report from filesystem
268
+ - Other judges' reports from filesystem
269
+ - Original solution
270
+ 4. Each agent appends "Debate Round {R}" section to their own report file
271
+ 5. Wait for all 3 agents to complete
272
+ 6. Go back to Step 2 (Check for Consensus)
273
+
274
+ **Step 5: Reply with Report**
275
+
276
+ Let's synthesize the evaluation results step by step.
277
+
278
+ 1. Read all final reports carefully
279
+ 2. Before generating the report, analyze the following:
280
+ - What is the consensus status (achieved or not)?
281
+ - What were the key points of agreement across all judges?
282
+ - What were the main areas of disagreement, if any?
283
+ - How did the debate rounds change the evaluations?
284
+ 3. Reply to user with a report that contains:
285
+ - If there is consensus:
286
+ - Consensus scores (average of all judges)
287
+ - Consensus strengths/weaknesses
288
+ - Number of rounds to reach consensus
289
+ - Final recommendation with clear justification
290
+ - If there is no consensus:
291
+ - All judges' final scores showing disagreements
292
+ - Specific criteria where consensus wasn't reached
293
+ - Analysis of why consensus couldn't be reached
294
+ - Flag for human review
295
+ 4. Command complete
296
+
297
+ ### Phase 3: Consensus Report
298
+
299
+ If consensus achieved, synthesize the final report by working through each section methodically:
300
+
301
+ ```markdown
302
+ # Consensus Evaluation Report
303
+
304
+ Let's compile the final consensus by analyzing each component systematically.
305
+
306
+ ## Consensus Scores
307
+
308
+ First, let's consolidate all judges' final scores:
309
+
310
+ | Criterion | Judge 1 | Judge 2 | Judge 3 | Final |
311
+ |-----------|---------|---------|---------|-------|
312
+ | {Name} | {X}/5 | {X}/5 | {X}/5 | {X}/5 |
313
+ ...
314
+
315
+ **Consensus Overall Score**: {avg}/5.0
316
+
317
+ ## Consensus Strengths
318
+ [Review each judge's identified strengths and extract the common themes that all judges agreed upon]
319
+
320
+ ## Consensus Weaknesses
321
+ [Review each judge's identified weaknesses and extract the common themes that all judges agreed upon]
322
+
323
+ ## Debate Summary
324
+ Let's trace how consensus was reached:
325
+ - Rounds to consensus: {N}
326
+ - Initial disagreements: {list with specific criteria and score gaps}
327
+ - How resolved: {for each disagreement, explain what evidence or argument led to resolution}
328
+
329
+ ## Final Recommendation
330
+ Based on the consensus scores and the key strengths/weaknesses identified:
331
+ {Pass/Fail/Needs Revision with clear justification tied to the evidence}
332
+ ```
333
+
334
+ <output>
335
+ The command produces:
336
+
337
+ 1. **Reports directory**: `.specs/reports/` (created if not exists)
338
+ 2. **Initial reports**: `.specs/reports/{solution-name}-{date}.1.md`, `.specs/reports/{solution-name}-{date}.2.md`, `.specs/reports/{solution-name}-{date}.3.md`
339
+ 3. **Debate updates**: Appended sections in each report file per round
340
+ 4. **Final synthesis**: Replied to user (consensus or disagreement summary)
341
+ </output>
342
+
343
+ ## Best Practices
344
+
345
+ ### Evaluation Criteria
346
+
347
+ Choose 3-5 weighted criteria relevant to the solution type:
348
+
349
+ **Code evaluation:**
350
+ - Correctness (30%) - Does it work? Handles edge cases?
351
+ - Design Quality (25%) - Clean architecture? Maintainable?
352
+ - Efficiency (20%) - Performance considerations?
353
+ - Code Quality (15%) - Readable? Well-documented?
354
+ - Testing (10%) - Test coverage? Test quality?
355
+
356
+ **Design/Architecture evaluation:**
357
+ - Completeness (30%) - All requirements addressed?
358
+ - Feasibility (25%) - Can it actually be built?
359
+ - Scalability (20%) - Handles growth?
360
+ - Simplicity (15%) - Appropriately simple?
361
+ - Documentation (10%) - Clear and comprehensive?
362
+
363
+ **Documentation evaluation:**
364
+ - Accuracy (35%) - Technically correct?
365
+ - Completeness (30%) - Covers all necessary topics?
366
+ - Clarity (20%) - Easy to understand?
367
+ - Usability (15%) - Helpful examples? Good structure?
368
+
369
+ ### Common Pitfalls
370
+
371
+ ❌ **Judges create new reports instead of appending** - Loses debate history
372
+ ❌ **Orchestrator passes reports between judges** - Violates filesystem communication principle
373
+ ❌ **Weak initial assessments** - Garbage in, garbage out
374
+ ❌ **Too many debate rounds** - Diminishing returns after 3 rounds
375
+ ❌ **Sycophancy in debate** - Judges agree too easily without real evidence
376
+
377
+ ✅ **Judges append to their own report file**
378
+ ✅ **Judges read other reports from filesystem directly**
379
+ ✅ **Strong evidence-based initial assessments**
380
+ ✅ **Maximum 3 debate rounds**
381
+ ✅ **Require evidence for changing positions**
382
+
383
+ ## Example Usage
384
+
385
+ ### Evaluating an API Implementation
386
+
387
+ ```bash
388
+ /judge-with-debate \
389
+ --solution "src/api/users.ts" \
390
+ --task "Implement REST API for user management" \
391
+ --criteria "correctness:30,design:25,security:20,performance:15,docs:10"
392
+ ```
393
+
394
+ **Round 1 outputs** (assuming date 2025-01-15):
395
+ - `.specs/reports/users-api-2025-01-15.1.md` - Judge 1 scores correctness 4/5, security 3/5
396
+ - `.specs/reports/users-api-2025-01-15.2.md` - Judge 2 scores correctness 4/5, security 5/5
397
+ - `.specs/reports/users-api-2025-01-15.3.md` - Judge 3 scores correctness 5/5, security 4/5
398
+
399
+ **Disagreement detected:** Security scores range from 3-5
400
+
401
+ **Round 2 debate:**
402
+ - Judge 1 defends 3/5: "Missing rate limiting, input validation incomplete"
403
+ - Judge 2 challenges: "Rate limiting exists in middleware (line 45)"
404
+ - Judge 1 revises to 4/5: "Missed middleware, but input validation still weak"
405
+ - Judge 3 defends 4/5: "Input validation adequate for requirements"
406
+
407
+ **Round 2 outputs:**
408
+ - All judges now 4-5/5 on security (within 1 point)
409
+ - Disagreement on input validation remains
410
+
411
+ **Round 3 debate:**
412
+ - Judges examine specific validation code
413
+ - Judge 2 revises to 4/5: "Upon re-examination, email validation regex is weak"
414
+ - Consensus: Security = 4/5
415
+
416
+ **Final consensus:**
417
+ ```
418
+ Correctness: 4.3/5
419
+ Design: 4.5/5
420
+ Security: 4.0/5 (3 rounds to consensus)
421
+ Performance: 4.7/5
422
+ Documentation: 4.0/5
423
+
424
+ Overall: 4.3/5 - PASS
425
+ ```
426
+
@@ -0,0 +1,17 @@
1
+ {
2
+ "name": "sadd-judge-with-debate",
3
+ "version": "1.0.0",
4
+ "description": "Judge-with-debate workflow for converging on a robust decision",
5
+ "triggers": [
6
+ "judge with debate",
7
+ "debate judges",
8
+ "multi round judging",
9
+ "consensus judge",
10
+ "llm judge"
11
+ ],
12
+ "applicable_agents": [
13
+ "critic"
14
+ ],
15
+ "max_context_tokens": 2200,
16
+ "entry_file": "SKILL.md"
17
+ }
@@ -0,0 +1,5 @@
1
+ # Structured Code Review
2
+
3
+ - Check correctness first, then regressions, then maintainability.
4
+ - Look for missing tests, edge cases, and unsafe assumptions.
5
+ - Return findings by severity with concise evidence.
@@ -0,0 +1,18 @@
1
+ {
2
+ "name": "structured-code-review",
3
+ "version": "1.0.0",
4
+ "description": "Structured checklist for local code review passes",
5
+ "triggers": [
6
+ "review",
7
+ "code review",
8
+ "check code",
9
+ "examine"
10
+ ],
11
+ "applicable_agents": [
12
+ "reviewer",
13
+ "worker",
14
+ "heavy-worker"
15
+ ],
16
+ "max_context_tokens": 2000,
17
+ "entry_file": "SKILL.md"
18
+ }
@@ -0,0 +1,5 @@
1
+ # Task Decomposition
2
+
3
+ - Turn the request into bounded slices with clear outputs.
4
+ - Order steps by dependency and verification needs.
5
+ - Call out blockers, ownership, and completion checks.
@@ -0,0 +1,20 @@
1
+ {
2
+ "name": "task-decomposition",
3
+ "version": "1.0.0",
4
+ "description": "Break larger requests into clear executable subtasks",
5
+ "triggers": [
6
+ "decompose",
7
+ "break down",
8
+ "split",
9
+ "tasks",
10
+ "planning",
11
+ "subtasks"
12
+ ],
13
+ "applicable_agents": [
14
+ "planner",
15
+ "executor",
16
+ "worker"
17
+ ],
18
+ "max_context_tokens": 1500,
19
+ "entry_file": "SKILL.md"
20
+ }
@@ -0,0 +1,5 @@
1
+ # Verification Before Completion
2
+
3
+ - Run the smallest command that proves the claim.
4
+ - Read the fresh output before saying work is done.
5
+ - Report failures exactly and rerun after changes.
@@ -0,0 +1,22 @@
1
+ {
2
+ "name": "verification-before-completion",
3
+ "version": "1.0.0",
4
+ "description": "Verification checklist before claiming work complete",
5
+ "triggers": [
6
+ "verify",
7
+ "complete",
8
+ "done",
9
+ "finish",
10
+ "check",
11
+ "validation"
12
+ ],
13
+ "applicable_agents": [
14
+ "worker",
15
+ "heavy-worker",
16
+ "deep-worker",
17
+ "ui-worker",
18
+ "ui-heavy-worker"
19
+ ],
20
+ "max_context_tokens": 2000,
21
+ "entry_file": "SKILL.md"
22
+ }