opencode-swarm-plugin 0.43.0 → 0.44.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (208) hide show
  1. package/bin/cass.characterization.test.ts +422 -0
  2. package/bin/swarm.serve.test.ts +6 -4
  3. package/bin/swarm.test.ts +68 -0
  4. package/bin/swarm.ts +81 -8
  5. package/dist/compaction-prompt-scoring.js +139 -0
  6. package/dist/contributor-tools.d.ts +42 -0
  7. package/dist/contributor-tools.d.ts.map +1 -0
  8. package/dist/eval-capture.js +12811 -0
  9. package/dist/hive.d.ts.map +1 -1
  10. package/dist/index.d.ts +12 -0
  11. package/dist/index.d.ts.map +1 -1
  12. package/dist/index.js +7728 -62590
  13. package/dist/plugin.js +23833 -78695
  14. package/dist/sessions/agent-discovery.d.ts +59 -0
  15. package/dist/sessions/agent-discovery.d.ts.map +1 -0
  16. package/dist/sessions/index.d.ts +10 -0
  17. package/dist/sessions/index.d.ts.map +1 -0
  18. package/dist/swarm-orchestrate.d.ts.map +1 -1
  19. package/dist/swarm-prompts.d.ts.map +1 -1
  20. package/dist/swarm-review.d.ts.map +1 -1
  21. package/package.json +17 -5
  22. package/.changeset/swarm-insights-data-layer.md +0 -63
  23. package/.hive/analysis/eval-failure-analysis-2025-12-25.md +0 -331
  24. package/.hive/analysis/session-data-quality-audit.md +0 -320
  25. package/.hive/eval-results.json +0 -483
  26. package/.hive/issues.jsonl +0 -138
  27. package/.hive/memories.jsonl +0 -729
  28. package/.opencode/eval-history.jsonl +0 -327
  29. package/.turbo/turbo-build.log +0 -9
  30. package/CHANGELOG.md +0 -2255
  31. package/SCORER-ANALYSIS.md +0 -598
  32. package/docs/analysis/subagent-coordination-patterns.md +0 -902
  33. package/docs/analysis-socratic-planner-pattern.md +0 -504
  34. package/docs/planning/ADR-001-monorepo-structure.md +0 -171
  35. package/docs/planning/ADR-002-package-extraction.md +0 -393
  36. package/docs/planning/ADR-003-performance-improvements.md +0 -451
  37. package/docs/planning/ADR-004-message-queue-features.md +0 -187
  38. package/docs/planning/ADR-005-devtools-observability.md +0 -202
  39. package/docs/planning/ADR-007-swarm-enhancements-worktree-review.md +0 -168
  40. package/docs/planning/ADR-008-worker-handoff-protocol.md +0 -293
  41. package/docs/planning/ADR-009-oh-my-opencode-patterns.md +0 -353
  42. package/docs/planning/ROADMAP.md +0 -368
  43. package/docs/semantic-memory-cli-syntax.md +0 -123
  44. package/docs/swarm-mail-architecture.md +0 -1147
  45. package/docs/testing/context-recovery-test.md +0 -470
  46. package/evals/ARCHITECTURE.md +0 -1189
  47. package/evals/README.md +0 -768
  48. package/evals/compaction-prompt.eval.ts +0 -149
  49. package/evals/compaction-resumption.eval.ts +0 -289
  50. package/evals/coordinator-behavior.eval.ts +0 -307
  51. package/evals/coordinator-session.eval.ts +0 -154
  52. package/evals/evalite.config.ts.bak +0 -15
  53. package/evals/example.eval.ts +0 -31
  54. package/evals/fixtures/compaction-cases.ts +0 -350
  55. package/evals/fixtures/compaction-prompt-cases.ts +0 -311
  56. package/evals/fixtures/coordinator-sessions.ts +0 -328
  57. package/evals/fixtures/decomposition-cases.ts +0 -105
  58. package/evals/lib/compaction-loader.test.ts +0 -248
  59. package/evals/lib/compaction-loader.ts +0 -320
  60. package/evals/lib/data-loader.evalite-test.ts +0 -289
  61. package/evals/lib/data-loader.test.ts +0 -345
  62. package/evals/lib/data-loader.ts +0 -281
  63. package/evals/lib/llm.ts +0 -115
  64. package/evals/scorers/compaction-prompt-scorers.ts +0 -145
  65. package/evals/scorers/compaction-scorers.ts +0 -305
  66. package/evals/scorers/coordinator-discipline.evalite-test.ts +0 -539
  67. package/evals/scorers/coordinator-discipline.ts +0 -325
  68. package/evals/scorers/index.test.ts +0 -146
  69. package/evals/scorers/index.ts +0 -328
  70. package/evals/scorers/outcome-scorers.evalite-test.ts +0 -27
  71. package/evals/scorers/outcome-scorers.ts +0 -349
  72. package/evals/swarm-decomposition.eval.ts +0 -121
  73. package/examples/commands/swarm.md +0 -745
  74. package/examples/plugin-wrapper-template.ts +0 -2426
  75. package/examples/skills/hive-workflow/SKILL.md +0 -212
  76. package/examples/skills/skill-creator/SKILL.md +0 -223
  77. package/examples/skills/swarm-coordination/SKILL.md +0 -292
  78. package/global-skills/cli-builder/SKILL.md +0 -344
  79. package/global-skills/cli-builder/references/advanced-patterns.md +0 -244
  80. package/global-skills/learning-systems/SKILL.md +0 -644
  81. package/global-skills/skill-creator/LICENSE.txt +0 -202
  82. package/global-skills/skill-creator/SKILL.md +0 -352
  83. package/global-skills/skill-creator/references/output-patterns.md +0 -82
  84. package/global-skills/skill-creator/references/workflows.md +0 -28
  85. package/global-skills/swarm-coordination/SKILL.md +0 -995
  86. package/global-skills/swarm-coordination/references/coordinator-patterns.md +0 -235
  87. package/global-skills/swarm-coordination/references/strategies.md +0 -138
  88. package/global-skills/system-design/SKILL.md +0 -213
  89. package/global-skills/testing-patterns/SKILL.md +0 -430
  90. package/global-skills/testing-patterns/references/dependency-breaking-catalog.md +0 -586
  91. package/opencode-swarm-plugin-0.30.7.tgz +0 -0
  92. package/opencode-swarm-plugin-0.31.0.tgz +0 -0
  93. package/scripts/cleanup-test-memories.ts +0 -346
  94. package/scripts/init-skill.ts +0 -222
  95. package/scripts/migrate-unknown-sessions.ts +0 -349
  96. package/scripts/validate-skill.ts +0 -204
  97. package/src/agent-mail.ts +0 -1724
  98. package/src/anti-patterns.test.ts +0 -1167
  99. package/src/anti-patterns.ts +0 -448
  100. package/src/compaction-capture.integration.test.ts +0 -257
  101. package/src/compaction-hook.test.ts +0 -838
  102. package/src/compaction-hook.ts +0 -1204
  103. package/src/compaction-observability.integration.test.ts +0 -139
  104. package/src/compaction-observability.test.ts +0 -187
  105. package/src/compaction-observability.ts +0 -324
  106. package/src/compaction-prompt-scorers.test.ts +0 -475
  107. package/src/compaction-prompt-scoring.ts +0 -300
  108. package/src/dashboard.test.ts +0 -611
  109. package/src/dashboard.ts +0 -462
  110. package/src/error-enrichment.test.ts +0 -403
  111. package/src/error-enrichment.ts +0 -219
  112. package/src/eval-capture.test.ts +0 -1015
  113. package/src/eval-capture.ts +0 -929
  114. package/src/eval-gates.test.ts +0 -306
  115. package/src/eval-gates.ts +0 -218
  116. package/src/eval-history.test.ts +0 -508
  117. package/src/eval-history.ts +0 -214
  118. package/src/eval-learning.test.ts +0 -378
  119. package/src/eval-learning.ts +0 -360
  120. package/src/eval-runner.test.ts +0 -223
  121. package/src/eval-runner.ts +0 -402
  122. package/src/export-tools.test.ts +0 -476
  123. package/src/export-tools.ts +0 -257
  124. package/src/hive.integration.test.ts +0 -2241
  125. package/src/hive.ts +0 -1628
  126. package/src/index.ts +0 -935
  127. package/src/learning.integration.test.ts +0 -1815
  128. package/src/learning.ts +0 -1079
  129. package/src/logger.test.ts +0 -189
  130. package/src/logger.ts +0 -135
  131. package/src/mandate-promotion.test.ts +0 -473
  132. package/src/mandate-promotion.ts +0 -239
  133. package/src/mandate-storage.integration.test.ts +0 -601
  134. package/src/mandate-storage.test.ts +0 -578
  135. package/src/mandate-storage.ts +0 -794
  136. package/src/mandates.ts +0 -540
  137. package/src/memory-tools.test.ts +0 -195
  138. package/src/memory-tools.ts +0 -344
  139. package/src/memory.integration.test.ts +0 -334
  140. package/src/memory.test.ts +0 -158
  141. package/src/memory.ts +0 -527
  142. package/src/model-selection.test.ts +0 -188
  143. package/src/model-selection.ts +0 -68
  144. package/src/observability-tools.test.ts +0 -359
  145. package/src/observability-tools.ts +0 -871
  146. package/src/output-guardrails.test.ts +0 -438
  147. package/src/output-guardrails.ts +0 -381
  148. package/src/pattern-maturity.test.ts +0 -1160
  149. package/src/pattern-maturity.ts +0 -525
  150. package/src/planning-guardrails.test.ts +0 -491
  151. package/src/planning-guardrails.ts +0 -438
  152. package/src/plugin.ts +0 -23
  153. package/src/post-compaction-tracker.test.ts +0 -251
  154. package/src/post-compaction-tracker.ts +0 -237
  155. package/src/query-tools.test.ts +0 -636
  156. package/src/query-tools.ts +0 -324
  157. package/src/rate-limiter.integration.test.ts +0 -466
  158. package/src/rate-limiter.ts +0 -774
  159. package/src/replay-tools.test.ts +0 -496
  160. package/src/replay-tools.ts +0 -240
  161. package/src/repo-crawl.integration.test.ts +0 -441
  162. package/src/repo-crawl.ts +0 -610
  163. package/src/schemas/cell-events.test.ts +0 -347
  164. package/src/schemas/cell-events.ts +0 -807
  165. package/src/schemas/cell.ts +0 -257
  166. package/src/schemas/evaluation.ts +0 -166
  167. package/src/schemas/index.test.ts +0 -199
  168. package/src/schemas/index.ts +0 -286
  169. package/src/schemas/mandate.ts +0 -232
  170. package/src/schemas/swarm-context.ts +0 -115
  171. package/src/schemas/task.ts +0 -161
  172. package/src/schemas/worker-handoff.test.ts +0 -302
  173. package/src/schemas/worker-handoff.ts +0 -131
  174. package/src/skills.integration.test.ts +0 -1192
  175. package/src/skills.test.ts +0 -643
  176. package/src/skills.ts +0 -1549
  177. package/src/storage.integration.test.ts +0 -341
  178. package/src/storage.ts +0 -884
  179. package/src/structured.integration.test.ts +0 -817
  180. package/src/structured.test.ts +0 -1046
  181. package/src/structured.ts +0 -762
  182. package/src/swarm-decompose.test.ts +0 -188
  183. package/src/swarm-decompose.ts +0 -1302
  184. package/src/swarm-deferred.integration.test.ts +0 -157
  185. package/src/swarm-deferred.test.ts +0 -38
  186. package/src/swarm-insights.test.ts +0 -214
  187. package/src/swarm-insights.ts +0 -459
  188. package/src/swarm-mail.integration.test.ts +0 -970
  189. package/src/swarm-mail.ts +0 -739
  190. package/src/swarm-orchestrate.integration.test.ts +0 -282
  191. package/src/swarm-orchestrate.test.ts +0 -548
  192. package/src/swarm-orchestrate.ts +0 -3084
  193. package/src/swarm-prompts.test.ts +0 -1270
  194. package/src/swarm-prompts.ts +0 -2077
  195. package/src/swarm-research.integration.test.ts +0 -701
  196. package/src/swarm-research.test.ts +0 -698
  197. package/src/swarm-research.ts +0 -472
  198. package/src/swarm-review.integration.test.ts +0 -285
  199. package/src/swarm-review.test.ts +0 -879
  200. package/src/swarm-review.ts +0 -709
  201. package/src/swarm-strategies.ts +0 -407
  202. package/src/swarm-worktree.test.ts +0 -501
  203. package/src/swarm-worktree.ts +0 -575
  204. package/src/swarm.integration.test.ts +0 -2377
  205. package/src/swarm.ts +0 -38
  206. package/src/tool-adapter.integration.test.ts +0 -1221
  207. package/src/tool-availability.ts +0 -461
  208. package/tsconfig.json +0 -28
@@ -1,320 +0,0 @@
1
- /**
2
- * Compaction Data Loader
3
- *
4
- * Loads COMPACTION events from session JSONL files for use in evals.
5
- *
6
- * Features:
7
- * - Lazy loading with early termination for large datasets
8
- * - Filtering by compaction_type, sessionIds, and limit
9
- * - Graceful error handling (skips invalid lines)
10
- * - Type-safe with Zod validation
11
- *
12
- * @module compaction-loader
13
- */
14
- import * as fs from "node:fs";
15
- import { createInterface } from "node:readline";
16
- import * as path from "node:path";
17
- import type { CoordinatorEvent } from "../../src/eval-capture.js";
18
- import { CoordinatorEventSchema } from "../../src/eval-capture.js";
19
-
20
- /**
21
- * Compaction event - subset of CoordinatorEvent with event_type === "COMPACTION"
22
- */
23
- export type CompactionEvent = Extract<
24
- CoordinatorEvent,
25
- { event_type: "COMPACTION" }
26
- >;
27
-
28
- /**
29
- * Compaction session - session with only COMPACTION events
30
- */
31
- export interface CompactionSession {
32
- session_id: string;
33
- epic_id: string;
34
- start_time: string;
35
- end_time: string;
36
- events: CompactionEvent[];
37
- }
38
-
39
- /**
40
- * Load options
41
- */
42
- export interface LoadOptions {
43
- /** Filter by compaction_type */
44
- compaction_type?:
45
- | "detection_complete"
46
- | "prompt_generated"
47
- | "context_injected"
48
- | "resumption_started"
49
- | "tool_call_tracked";
50
- /** Filter by session IDs */
51
- sessionIds?: string[];
52
- /** Limit number of results */
53
- limit?: number;
54
- }
55
-
56
- /**
57
- * Load COMPACTION events from session JSONL files
58
- *
59
- * Reads all .jsonl files in the session directory, parses events,
60
- * and returns only COMPACTION events matching the filters.
61
- *
62
- * @param sessionDir - Path to session directory (default: ~/.config/swarm-tools/sessions)
63
- * @param options - Filter options
64
- * @returns Array of compaction events
65
- *
66
- * @example
67
- * // Load all COMPACTION events
68
- * const events = await loadCompactionEvents("/path/to/sessions");
69
- *
70
- * @example
71
- * // Load only detection_complete events
72
- * const events = await loadCompactionEvents("/path/to/sessions", {
73
- * compaction_type: "detection_complete",
74
- * });
75
- *
76
- * @example
77
- * // Load events from specific sessions
78
- * const events = await loadCompactionEvents("/path/to/sessions", {
79
- * sessionIds: ["session-1", "session-2"],
80
- * limit: 10,
81
- * });
82
- */
83
- export async function loadCompactionEvents(
84
- sessionDir: string,
85
- options?: LoadOptions,
86
- ): Promise<CompactionEvent[]> {
87
- const { compaction_type, sessionIds, limit } = options ?? {};
88
-
89
- // Check if directory exists
90
- if (!fs.existsSync(sessionDir)) {
91
- return [];
92
- }
93
-
94
- let files: string[];
95
- try {
96
- // Read all .jsonl files
97
- files = fs.readdirSync(sessionDir).filter((f) => f.endsWith(".jsonl"));
98
- } catch (error) {
99
- // Directory exists but can't be read - log and return empty
100
- console.warn(`Failed to read session directory ${sessionDir}:`, error);
101
- return [];
102
- }
103
-
104
- // Filter by sessionIds if provided
105
- const targetFiles = sessionIds
106
- ? files.filter((f) => sessionIds.includes(f.replace(".jsonl", "")))
107
- : files;
108
-
109
- const events: CompactionEvent[] = [];
110
-
111
- for (const file of targetFiles) {
112
- const filePath = path.join(sessionDir, file);
113
-
114
- try {
115
- // Stream large files line-by-line to avoid loading entire file into memory
116
- const shouldStream = limit && limit < 100; // For small limits, streaming is overkill
117
-
118
- if (shouldStream) {
119
- // Use streaming for better memory efficiency
120
- const found = await loadFromFileStream(filePath, {
121
- compaction_type,
122
- remainingLimit: limit - events.length,
123
- });
124
- events.push(...found);
125
- } else {
126
- // For small files or no limit, read entire file (faster)
127
- const content = fs.readFileSync(filePath, "utf-8");
128
- const lines = content.trim().split("\n").filter(Boolean);
129
-
130
- for (const line of lines) {
131
- const event = parseLine(line);
132
- if (event && event.event_type === "COMPACTION") {
133
- // Filter by compaction_type if provided
134
- if (!compaction_type || event.compaction_type === compaction_type) {
135
- events.push(event);
136
-
137
- // Apply limit early to avoid processing unnecessary files
138
- if (limit && events.length >= limit) {
139
- return events.slice(0, limit);
140
- }
141
- }
142
- }
143
- }
144
- }
145
-
146
- // Early termination if limit reached
147
- if (limit && events.length >= limit) {
148
- return events.slice(0, limit);
149
- }
150
- } catch (error) {
151
- // Log file read errors but continue processing other files
152
- console.warn(`Failed to read session file ${filePath}:`, error);
153
- }
154
- }
155
-
156
- return limit ? events.slice(0, limit) : events;
157
- }
158
-
159
- /**
160
- * Parse a JSONL line into a CoordinatorEvent
161
- *
162
- * @param line - JSONL line to parse
163
- * @returns Parsed and validated event, or null if invalid
164
- */
165
- function parseLine(line: string): CoordinatorEvent | null {
166
- try {
167
- const parsed = JSON.parse(line);
168
- return CoordinatorEventSchema.parse(parsed);
169
- } catch {
170
- // Invalid JSON or failed validation - skip silently
171
- return null;
172
- }
173
- }
174
-
175
- /**
176
- * Load COMPACTION events from a file using streaming (for large files)
177
- *
178
- * @param filePath - Path to session JSONL file
179
- * @param options - Filter options
180
- * @returns Array of matching compaction events
181
- */
182
- async function loadFromFileStream(
183
- filePath: string,
184
- options: {
185
- compaction_type?: LoadOptions["compaction_type"];
186
- remainingLimit?: number;
187
- },
188
- ): Promise<CompactionEvent[]> {
189
- const { compaction_type, remainingLimit } = options;
190
- const events: CompactionEvent[] = [];
191
-
192
- const fileStream = fs.createReadStream(filePath, { encoding: "utf-8" });
193
- const rl = createInterface({ input: fileStream, crlfDelay: Number.POSITIVE_INFINITY });
194
-
195
- for await (const line of rl) {
196
- const event = parseLine(line);
197
- if (event && event.event_type === "COMPACTION") {
198
- if (!compaction_type || event.compaction_type === compaction_type) {
199
- events.push(event);
200
-
201
- // Early termination for streaming
202
- if (remainingLimit && events.length >= remainingLimit) {
203
- rl.close();
204
- fileStream.close();
205
- break;
206
- }
207
- }
208
- }
209
- }
210
-
211
- return events;
212
- }
213
-
214
- /**
215
- * Load COMPACTION sessions grouped by session_id
216
- *
217
- * Groups COMPACTION events by session_id and returns session metadata.
218
- *
219
- * @param sessionDir - Path to session directory
220
- * @param options - Filter options
221
- * @returns Array of compaction sessions
222
- *
223
- * @example
224
- * // Load all sessions with COMPACTION events
225
- * const sessions = await loadCompactionSessions("/path/to/sessions");
226
- *
227
- * @example
228
- * // Load sessions with specific compaction_type
229
- * const sessions = await loadCompactionSessions("/path/to/sessions", {
230
- * compaction_type: "prompt_generated",
231
- * });
232
- */
233
- export async function loadCompactionSessions(
234
- sessionDir: string,
235
- options?: LoadOptions,
236
- ): Promise<CompactionSession[]> {
237
- const events = await loadCompactionEvents(sessionDir, options);
238
-
239
- if (events.length === 0) {
240
- return [];
241
- }
242
-
243
- // Group events by session_id
244
- const sessionMap = new Map<string, CompactionEvent[]>();
245
-
246
- for (const event of events) {
247
- const existing = sessionMap.get(event.session_id);
248
- if (existing) {
249
- existing.push(event);
250
- } else {
251
- sessionMap.set(event.session_id, [event]);
252
- }
253
- }
254
-
255
- // Build sessions with metadata
256
- const sessions: CompactionSession[] = [];
257
-
258
- for (const [sessionId, sessionEvents] of sessionMap.entries()) {
259
- if (sessionEvents.length === 0) {
260
- continue;
261
- }
262
-
263
- // Get epic_id from first event
264
- const epicId = sessionEvents[0].epic_id;
265
-
266
- // Get timestamps
267
- const timestamps = sessionEvents.map((e) => new Date(e.timestamp).getTime());
268
- const startTime = new Date(Math.min(...timestamps)).toISOString();
269
- const endTime = new Date(Math.max(...timestamps)).toISOString();
270
-
271
- sessions.push({
272
- session_id: sessionId,
273
- epic_id: epicId,
274
- start_time: startTime,
275
- end_time: endTime,
276
- events: sessionEvents,
277
- });
278
- }
279
-
280
- // Apply limit
281
- return options?.limit ? sessions.slice(0, options.limit) : sessions;
282
- }
283
-
284
- /**
285
- * Load COMPACTION events from default session directory
286
- *
287
- * Convenience wrapper that uses the default ~/.config/swarm-tools/sessions directory.
288
- *
289
- * @param options - Filter options
290
- * @returns Array of compaction events
291
- *
292
- * @example
293
- * // Load recent compaction events
294
- * const events = await loadDefaultCompactionEvents({ limit: 10 });
295
- */
296
- export async function loadDefaultCompactionEvents(
297
- options?: LoadOptions,
298
- ): Promise<CompactionEvent[]> {
299
- const { getSessionDir } = await import("../../src/eval-capture.js");
300
- return loadCompactionEvents(getSessionDir(), options);
301
- }
302
-
303
- /**
304
- * Load COMPACTION sessions from default session directory
305
- *
306
- * Convenience wrapper that uses the default ~/.config/swarm-tools/sessions directory.
307
- *
308
- * @param options - Filter options
309
- * @returns Array of compaction sessions
310
- *
311
- * @example
312
- * // Load all compaction sessions
313
- * const sessions = await loadDefaultCompactionSessions();
314
- */
315
- export async function loadDefaultCompactionSessions(
316
- options?: LoadOptions,
317
- ): Promise<CompactionSession[]> {
318
- const { getSessionDir } = await import("../../src/eval-capture.js");
319
- return loadCompactionSessions(getSessionDir(), options);
320
- }
@@ -1,289 +0,0 @@
1
- /**
2
- * Data Loader Tests
3
- *
4
- * Tests the PGlite-backed eval data loader functions.
5
- * Uses a real in-memory PGlite database for accurate testing.
6
- */
7
- import { describe, it, expect, beforeAll, afterAll } from "bun:test";
8
- import {
9
- loadEvalCases,
10
- hasRealEvalData,
11
- getEvalDataSummary,
12
- } from "./data-loader.js";
13
- import {
14
- appendEvent,
15
- getDatabase,
16
- closeDatabase,
17
- type DecompositionGeneratedEvent,
18
- type SubtaskOutcomeEvent,
19
- } from "swarm-mail";
20
- import * as fs from "node:fs";
21
- import * as path from "node:path";
22
- import * as os from "node:os";
23
-
24
- const TEST_PROJECT_KEY = "test-project-eval-loader";
25
-
26
- // Create a unique temp directory for this test run
27
- let testDir: string;
28
-
29
- describe("Data Loader", () => {
30
- beforeAll(async () => {
31
- // Create temp directory for test database
32
- testDir = fs.mkdtempSync(path.join(os.tmpdir(), "eval-loader-test-"));
33
-
34
- // Initialize database by getting it (lazy init)
35
- await getDatabase(testDir);
36
- });
37
-
38
- afterAll(async () => {
39
- await closeDatabase(testDir);
40
- // Clean up temp directory
41
- fs.rmSync(testDir, { recursive: true, force: true });
42
- });
43
-
44
- describe("loadEvalCases", () => {
45
- it("transforms eval records to EvalCase format", async () => {
46
- // Insert a decomposition event
47
- const decompositionEvent: DecompositionGeneratedEvent = {
48
- type: "decomposition_generated",
49
- timestamp: Date.now(),
50
- project_key: TEST_PROJECT_KEY,
51
- epic_id: "epic-load-1",
52
- task: "Add authentication",
53
- context: "Next.js app",
54
- strategy: "feature-based",
55
- epic_title: "Auth Epic",
56
- subtasks: [
57
- { title: "OAuth setup", files: ["src/auth/oauth.ts"], priority: 1 },
58
- {
59
- title: "Session management",
60
- files: ["src/auth/session.ts"],
61
- priority: 2,
62
- },
63
- ],
64
- };
65
- await appendEvent(decompositionEvent, testDir);
66
-
67
- // Insert outcome events for both subtasks
68
- const outcome1: SubtaskOutcomeEvent = {
69
- type: "subtask_outcome",
70
- timestamp: Date.now(),
71
- project_key: TEST_PROJECT_KEY,
72
- epic_id: "epic-load-1",
73
- bead_id: "epic-load-1.1",
74
- planned_files: ["src/auth/oauth.ts"],
75
- actual_files: ["src/auth/oauth.ts"],
76
- duration_ms: 5000,
77
- error_count: 0,
78
- retry_count: 0,
79
- success: true,
80
- };
81
- await appendEvent(outcome1, testDir);
82
-
83
- const outcome2: SubtaskOutcomeEvent = {
84
- type: "subtask_outcome",
85
- timestamp: Date.now(),
86
- project_key: TEST_PROJECT_KEY,
87
- epic_id: "epic-load-1",
88
- bead_id: "epic-load-1.2",
89
- planned_files: ["src/auth/session.ts"],
90
- actual_files: ["src/auth/session.ts"],
91
- duration_ms: 3000,
92
- error_count: 0,
93
- retry_count: 0,
94
- success: true,
95
- };
96
- await appendEvent(outcome2, testDir);
97
-
98
- const cases = await loadEvalCases(TEST_PROJECT_KEY, {
99
- projectPath: testDir,
100
- });
101
-
102
- expect(cases.length).toBeGreaterThanOrEqual(1);
103
- const authCase = cases.find((c) => c.input.task === "Add authentication");
104
- expect(authCase).toBeDefined();
105
- expect(authCase!.input.context).toBe("Next.js app");
106
- expect(authCase!.expected.minSubtasks).toBe(2);
107
- expect(authCase!.expected.maxSubtasks).toBe(2);
108
- expect(authCase!.expected.requiredFiles).toContain("src/auth/oauth.ts");
109
- expect(authCase!.expected.requiredFiles).toContain("src/auth/session.ts");
110
- expect(authCase!.actual).toBeDefined();
111
- });
112
-
113
- it("filters by success when successOnly is true", async () => {
114
- // Insert a successful decomposition
115
- const successEvent: DecompositionGeneratedEvent = {
116
- type: "decomposition_generated",
117
- timestamp: Date.now(),
118
- project_key: TEST_PROJECT_KEY,
119
- epic_id: "epic-success-filter",
120
- task: "Success task for filter",
121
- strategy: "feature-based",
122
- epic_title: "Success Epic",
123
- subtasks: [{ title: "Sub", files: ["src/success.ts"], priority: 1 }],
124
- };
125
- await appendEvent(successEvent, testDir);
126
-
127
- // Mark it successful
128
- const successOutcome: SubtaskOutcomeEvent = {
129
- type: "subtask_outcome",
130
- timestamp: Date.now(),
131
- project_key: TEST_PROJECT_KEY,
132
- epic_id: "epic-success-filter",
133
- bead_id: "epic-success-filter.1",
134
- planned_files: ["src/success.ts"],
135
- actual_files: ["src/success.ts"],
136
- duration_ms: 1000,
137
- error_count: 0,
138
- retry_count: 0,
139
- success: true,
140
- };
141
- await appendEvent(successOutcome, testDir);
142
-
143
- // Insert a failed decomposition
144
- const failEvent: DecompositionGeneratedEvent = {
145
- type: "decomposition_generated",
146
- timestamp: Date.now(),
147
- project_key: TEST_PROJECT_KEY,
148
- epic_id: "epic-fail-filter",
149
- task: "Failed task for filter",
150
- strategy: "feature-based",
151
- epic_title: "Failed Epic",
152
- subtasks: [{ title: "Sub", files: ["src/fail.ts"], priority: 1 }],
153
- };
154
- await appendEvent(failEvent, testDir);
155
-
156
- // Mark it failed
157
- const failOutcome: SubtaskOutcomeEvent = {
158
- type: "subtask_outcome",
159
- timestamp: Date.now(),
160
- project_key: TEST_PROJECT_KEY,
161
- epic_id: "epic-fail-filter",
162
- bead_id: "epic-fail-filter.1",
163
- planned_files: ["src/fail.ts"],
164
- actual_files: [],
165
- duration_ms: 500,
166
- error_count: 3,
167
- retry_count: 2,
168
- success: false,
169
- };
170
- await appendEvent(failOutcome, testDir);
171
-
172
- const successCases = await loadEvalCases(TEST_PROJECT_KEY, {
173
- successOnly: true,
174
- projectPath: testDir,
175
- });
176
-
177
- // Should only include successful cases
178
- const failedCase = successCases.find(
179
- (c) => c.input.task === "Failed task for filter",
180
- );
181
- expect(failedCase).toBeUndefined();
182
- });
183
-
184
- it("passes strategy filter to getEvalRecords", async () => {
185
- // Insert file-based decomposition
186
- const fileBasedEvent: DecompositionGeneratedEvent = {
187
- type: "decomposition_generated",
188
- timestamp: Date.now(),
189
- project_key: TEST_PROJECT_KEY,
190
- epic_id: "epic-file-based",
191
- task: "File-based task",
192
- strategy: "file-based",
193
- epic_title: "File Epic",
194
- subtasks: [{ title: "Sub", files: ["src/file.ts"], priority: 1 }],
195
- };
196
- await appendEvent(fileBasedEvent, testDir);
197
-
198
- const fileBasedCases = await loadEvalCases(TEST_PROJECT_KEY, {
199
- strategy: "file-based",
200
- projectPath: testDir,
201
- });
202
-
203
- // All returned cases should be file-based
204
- for (const c of fileBasedCases) {
205
- expect(c.actual?.strategy).toBe("file-based");
206
- }
207
- });
208
-
209
- it("passes limit to getEvalRecords", async () => {
210
- const cases = await loadEvalCases(TEST_PROJECT_KEY, {
211
- limit: 2,
212
- projectPath: testDir,
213
- });
214
-
215
- expect(cases.length).toBeLessThanOrEqual(2);
216
- });
217
-
218
- it("handles records with no context", async () => {
219
- const noContextEvent: DecompositionGeneratedEvent = {
220
- type: "decomposition_generated",
221
- timestamp: Date.now(),
222
- project_key: TEST_PROJECT_KEY,
223
- epic_id: "epic-no-context",
224
- task: "Task without context",
225
- // context is undefined
226
- strategy: "feature-based",
227
- epic_title: "No Context Epic",
228
- subtasks: [{ title: "Sub", files: [], priority: 1 }],
229
- };
230
- await appendEvent(noContextEvent, testDir);
231
-
232
- const cases = await loadEvalCases(TEST_PROJECT_KEY, {
233
- projectPath: testDir,
234
- });
235
- const noContextCase = cases.find(
236
- (c) => c.input.task === "Task without context",
237
- );
238
-
239
- expect(noContextCase).toBeDefined();
240
- expect(noContextCase!.input.context).toBeUndefined();
241
- });
242
- });
243
-
244
- describe("hasRealEvalData", () => {
245
- it("returns true when enough records exist", async () => {
246
- // We've inserted several records above, should have enough
247
- const hasData = await hasRealEvalData(TEST_PROJECT_KEY, 1, testDir);
248
- expect(hasData).toBe(true);
249
- });
250
-
251
- it("returns false when not enough records exist", async () => {
252
- // Use a project key with no data
253
- const hasData = await hasRealEvalData("nonexistent-project", 5, testDir);
254
- expect(hasData).toBe(false);
255
- });
256
-
257
- it("uses custom minRecords threshold", async () => {
258
- // Should have at least 1 record
259
- const hasData = await hasRealEvalData(TEST_PROJECT_KEY, 1, testDir);
260
- expect(hasData).toBe(true);
261
-
262
- // Should not have 1000 records
263
- const hasLotsOfData = await hasRealEvalData(
264
- TEST_PROJECT_KEY,
265
- 1000,
266
- testDir,
267
- );
268
- expect(hasLotsOfData).toBe(false);
269
- });
270
- });
271
-
272
- describe("getEvalDataSummary", () => {
273
- it("returns formatted summary with hasEnoughData flag", async () => {
274
- const summary = await getEvalDataSummary(TEST_PROJECT_KEY, testDir);
275
-
276
- expect(summary.totalRecords).toBeGreaterThanOrEqual(1);
277
- expect(typeof summary.successRate).toBe("number");
278
- expect(typeof summary.byStrategy).toBe("object");
279
- expect(typeof summary.hasEnoughData).toBe("boolean");
280
- });
281
-
282
- it("sets hasEnoughData based on record count", async () => {
283
- // Empty project should not have enough data
284
- const emptySummary = await getEvalDataSummary("empty-project", testDir);
285
- expect(emptySummary.hasEnoughData).toBe(false);
286
- expect(emptySummary.totalRecords).toBe(0);
287
- });
288
- });
289
- });