opencode-swarm-plugin 0.43.0 → 0.44.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/cass.characterization.test.ts +422 -0
- package/bin/swarm.serve.test.ts +6 -4
- package/bin/swarm.test.ts +68 -0
- package/bin/swarm.ts +81 -8
- package/dist/compaction-prompt-scoring.js +139 -0
- package/dist/contributor-tools.d.ts +42 -0
- package/dist/contributor-tools.d.ts.map +1 -0
- package/dist/eval-capture.js +12811 -0
- package/dist/hive.d.ts.map +1 -1
- package/dist/index.d.ts +12 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +7728 -62590
- package/dist/plugin.js +23833 -78695
- package/dist/sessions/agent-discovery.d.ts +59 -0
- package/dist/sessions/agent-discovery.d.ts.map +1 -0
- package/dist/sessions/index.d.ts +10 -0
- package/dist/sessions/index.d.ts.map +1 -0
- package/dist/swarm-orchestrate.d.ts.map +1 -1
- package/dist/swarm-prompts.d.ts.map +1 -1
- package/dist/swarm-review.d.ts.map +1 -1
- package/package.json +17 -5
- package/.changeset/swarm-insights-data-layer.md +0 -63
- package/.hive/analysis/eval-failure-analysis-2025-12-25.md +0 -331
- package/.hive/analysis/session-data-quality-audit.md +0 -320
- package/.hive/eval-results.json +0 -483
- package/.hive/issues.jsonl +0 -138
- package/.hive/memories.jsonl +0 -729
- package/.opencode/eval-history.jsonl +0 -327
- package/.turbo/turbo-build.log +0 -9
- package/CHANGELOG.md +0 -2255
- package/SCORER-ANALYSIS.md +0 -598
- package/docs/analysis/subagent-coordination-patterns.md +0 -902
- package/docs/analysis-socratic-planner-pattern.md +0 -504
- package/docs/planning/ADR-001-monorepo-structure.md +0 -171
- package/docs/planning/ADR-002-package-extraction.md +0 -393
- package/docs/planning/ADR-003-performance-improvements.md +0 -451
- package/docs/planning/ADR-004-message-queue-features.md +0 -187
- package/docs/planning/ADR-005-devtools-observability.md +0 -202
- package/docs/planning/ADR-007-swarm-enhancements-worktree-review.md +0 -168
- package/docs/planning/ADR-008-worker-handoff-protocol.md +0 -293
- package/docs/planning/ADR-009-oh-my-opencode-patterns.md +0 -353
- package/docs/planning/ROADMAP.md +0 -368
- package/docs/semantic-memory-cli-syntax.md +0 -123
- package/docs/swarm-mail-architecture.md +0 -1147
- package/docs/testing/context-recovery-test.md +0 -470
- package/evals/ARCHITECTURE.md +0 -1189
- package/evals/README.md +0 -768
- package/evals/compaction-prompt.eval.ts +0 -149
- package/evals/compaction-resumption.eval.ts +0 -289
- package/evals/coordinator-behavior.eval.ts +0 -307
- package/evals/coordinator-session.eval.ts +0 -154
- package/evals/evalite.config.ts.bak +0 -15
- package/evals/example.eval.ts +0 -31
- package/evals/fixtures/compaction-cases.ts +0 -350
- package/evals/fixtures/compaction-prompt-cases.ts +0 -311
- package/evals/fixtures/coordinator-sessions.ts +0 -328
- package/evals/fixtures/decomposition-cases.ts +0 -105
- package/evals/lib/compaction-loader.test.ts +0 -248
- package/evals/lib/compaction-loader.ts +0 -320
- package/evals/lib/data-loader.evalite-test.ts +0 -289
- package/evals/lib/data-loader.test.ts +0 -345
- package/evals/lib/data-loader.ts +0 -281
- package/evals/lib/llm.ts +0 -115
- package/evals/scorers/compaction-prompt-scorers.ts +0 -145
- package/evals/scorers/compaction-scorers.ts +0 -305
- package/evals/scorers/coordinator-discipline.evalite-test.ts +0 -539
- package/evals/scorers/coordinator-discipline.ts +0 -325
- package/evals/scorers/index.test.ts +0 -146
- package/evals/scorers/index.ts +0 -328
- package/evals/scorers/outcome-scorers.evalite-test.ts +0 -27
- package/evals/scorers/outcome-scorers.ts +0 -349
- package/evals/swarm-decomposition.eval.ts +0 -121
- package/examples/commands/swarm.md +0 -745
- package/examples/plugin-wrapper-template.ts +0 -2426
- package/examples/skills/hive-workflow/SKILL.md +0 -212
- package/examples/skills/skill-creator/SKILL.md +0 -223
- package/examples/skills/swarm-coordination/SKILL.md +0 -292
- package/global-skills/cli-builder/SKILL.md +0 -344
- package/global-skills/cli-builder/references/advanced-patterns.md +0 -244
- package/global-skills/learning-systems/SKILL.md +0 -644
- package/global-skills/skill-creator/LICENSE.txt +0 -202
- package/global-skills/skill-creator/SKILL.md +0 -352
- package/global-skills/skill-creator/references/output-patterns.md +0 -82
- package/global-skills/skill-creator/references/workflows.md +0 -28
- package/global-skills/swarm-coordination/SKILL.md +0 -995
- package/global-skills/swarm-coordination/references/coordinator-patterns.md +0 -235
- package/global-skills/swarm-coordination/references/strategies.md +0 -138
- package/global-skills/system-design/SKILL.md +0 -213
- package/global-skills/testing-patterns/SKILL.md +0 -430
- package/global-skills/testing-patterns/references/dependency-breaking-catalog.md +0 -586
- package/opencode-swarm-plugin-0.30.7.tgz +0 -0
- package/opencode-swarm-plugin-0.31.0.tgz +0 -0
- package/scripts/cleanup-test-memories.ts +0 -346
- package/scripts/init-skill.ts +0 -222
- package/scripts/migrate-unknown-sessions.ts +0 -349
- package/scripts/validate-skill.ts +0 -204
- package/src/agent-mail.ts +0 -1724
- package/src/anti-patterns.test.ts +0 -1167
- package/src/anti-patterns.ts +0 -448
- package/src/compaction-capture.integration.test.ts +0 -257
- package/src/compaction-hook.test.ts +0 -838
- package/src/compaction-hook.ts +0 -1204
- package/src/compaction-observability.integration.test.ts +0 -139
- package/src/compaction-observability.test.ts +0 -187
- package/src/compaction-observability.ts +0 -324
- package/src/compaction-prompt-scorers.test.ts +0 -475
- package/src/compaction-prompt-scoring.ts +0 -300
- package/src/dashboard.test.ts +0 -611
- package/src/dashboard.ts +0 -462
- package/src/error-enrichment.test.ts +0 -403
- package/src/error-enrichment.ts +0 -219
- package/src/eval-capture.test.ts +0 -1015
- package/src/eval-capture.ts +0 -929
- package/src/eval-gates.test.ts +0 -306
- package/src/eval-gates.ts +0 -218
- package/src/eval-history.test.ts +0 -508
- package/src/eval-history.ts +0 -214
- package/src/eval-learning.test.ts +0 -378
- package/src/eval-learning.ts +0 -360
- package/src/eval-runner.test.ts +0 -223
- package/src/eval-runner.ts +0 -402
- package/src/export-tools.test.ts +0 -476
- package/src/export-tools.ts +0 -257
- package/src/hive.integration.test.ts +0 -2241
- package/src/hive.ts +0 -1628
- package/src/index.ts +0 -935
- package/src/learning.integration.test.ts +0 -1815
- package/src/learning.ts +0 -1079
- package/src/logger.test.ts +0 -189
- package/src/logger.ts +0 -135
- package/src/mandate-promotion.test.ts +0 -473
- package/src/mandate-promotion.ts +0 -239
- package/src/mandate-storage.integration.test.ts +0 -601
- package/src/mandate-storage.test.ts +0 -578
- package/src/mandate-storage.ts +0 -794
- package/src/mandates.ts +0 -540
- package/src/memory-tools.test.ts +0 -195
- package/src/memory-tools.ts +0 -344
- package/src/memory.integration.test.ts +0 -334
- package/src/memory.test.ts +0 -158
- package/src/memory.ts +0 -527
- package/src/model-selection.test.ts +0 -188
- package/src/model-selection.ts +0 -68
- package/src/observability-tools.test.ts +0 -359
- package/src/observability-tools.ts +0 -871
- package/src/output-guardrails.test.ts +0 -438
- package/src/output-guardrails.ts +0 -381
- package/src/pattern-maturity.test.ts +0 -1160
- package/src/pattern-maturity.ts +0 -525
- package/src/planning-guardrails.test.ts +0 -491
- package/src/planning-guardrails.ts +0 -438
- package/src/plugin.ts +0 -23
- package/src/post-compaction-tracker.test.ts +0 -251
- package/src/post-compaction-tracker.ts +0 -237
- package/src/query-tools.test.ts +0 -636
- package/src/query-tools.ts +0 -324
- package/src/rate-limiter.integration.test.ts +0 -466
- package/src/rate-limiter.ts +0 -774
- package/src/replay-tools.test.ts +0 -496
- package/src/replay-tools.ts +0 -240
- package/src/repo-crawl.integration.test.ts +0 -441
- package/src/repo-crawl.ts +0 -610
- package/src/schemas/cell-events.test.ts +0 -347
- package/src/schemas/cell-events.ts +0 -807
- package/src/schemas/cell.ts +0 -257
- package/src/schemas/evaluation.ts +0 -166
- package/src/schemas/index.test.ts +0 -199
- package/src/schemas/index.ts +0 -286
- package/src/schemas/mandate.ts +0 -232
- package/src/schemas/swarm-context.ts +0 -115
- package/src/schemas/task.ts +0 -161
- package/src/schemas/worker-handoff.test.ts +0 -302
- package/src/schemas/worker-handoff.ts +0 -131
- package/src/skills.integration.test.ts +0 -1192
- package/src/skills.test.ts +0 -643
- package/src/skills.ts +0 -1549
- package/src/storage.integration.test.ts +0 -341
- package/src/storage.ts +0 -884
- package/src/structured.integration.test.ts +0 -817
- package/src/structured.test.ts +0 -1046
- package/src/structured.ts +0 -762
- package/src/swarm-decompose.test.ts +0 -188
- package/src/swarm-decompose.ts +0 -1302
- package/src/swarm-deferred.integration.test.ts +0 -157
- package/src/swarm-deferred.test.ts +0 -38
- package/src/swarm-insights.test.ts +0 -214
- package/src/swarm-insights.ts +0 -459
- package/src/swarm-mail.integration.test.ts +0 -970
- package/src/swarm-mail.ts +0 -739
- package/src/swarm-orchestrate.integration.test.ts +0 -282
- package/src/swarm-orchestrate.test.ts +0 -548
- package/src/swarm-orchestrate.ts +0 -3084
- package/src/swarm-prompts.test.ts +0 -1270
- package/src/swarm-prompts.ts +0 -2077
- package/src/swarm-research.integration.test.ts +0 -701
- package/src/swarm-research.test.ts +0 -698
- package/src/swarm-research.ts +0 -472
- package/src/swarm-review.integration.test.ts +0 -285
- package/src/swarm-review.test.ts +0 -879
- package/src/swarm-review.ts +0 -709
- package/src/swarm-strategies.ts +0 -407
- package/src/swarm-worktree.test.ts +0 -501
- package/src/swarm-worktree.ts +0 -575
- package/src/swarm.integration.test.ts +0 -2377
- package/src/swarm.ts +0 -38
- package/src/tool-adapter.integration.test.ts +0 -1221
- package/src/tool-availability.ts +0 -461
- package/tsconfig.json +0 -28
|
@@ -1,320 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Compaction Data Loader
|
|
3
|
-
*
|
|
4
|
-
* Loads COMPACTION events from session JSONL files for use in evals.
|
|
5
|
-
*
|
|
6
|
-
* Features:
|
|
7
|
-
* - Lazy loading with early termination for large datasets
|
|
8
|
-
* - Filtering by compaction_type, sessionIds, and limit
|
|
9
|
-
* - Graceful error handling (skips invalid lines)
|
|
10
|
-
* - Type-safe with Zod validation
|
|
11
|
-
*
|
|
12
|
-
* @module compaction-loader
|
|
13
|
-
*/
|
|
14
|
-
import * as fs from "node:fs";
|
|
15
|
-
import { createInterface } from "node:readline";
|
|
16
|
-
import * as path from "node:path";
|
|
17
|
-
import type { CoordinatorEvent } from "../../src/eval-capture.js";
|
|
18
|
-
import { CoordinatorEventSchema } from "../../src/eval-capture.js";
|
|
19
|
-
|
|
20
|
-
/**
|
|
21
|
-
* Compaction event - subset of CoordinatorEvent with event_type === "COMPACTION"
|
|
22
|
-
*/
|
|
23
|
-
export type CompactionEvent = Extract<
|
|
24
|
-
CoordinatorEvent,
|
|
25
|
-
{ event_type: "COMPACTION" }
|
|
26
|
-
>;
|
|
27
|
-
|
|
28
|
-
/**
|
|
29
|
-
* Compaction session - session with only COMPACTION events
|
|
30
|
-
*/
|
|
31
|
-
export interface CompactionSession {
|
|
32
|
-
session_id: string;
|
|
33
|
-
epic_id: string;
|
|
34
|
-
start_time: string;
|
|
35
|
-
end_time: string;
|
|
36
|
-
events: CompactionEvent[];
|
|
37
|
-
}
|
|
38
|
-
|
|
39
|
-
/**
|
|
40
|
-
* Load options
|
|
41
|
-
*/
|
|
42
|
-
export interface LoadOptions {
|
|
43
|
-
/** Filter by compaction_type */
|
|
44
|
-
compaction_type?:
|
|
45
|
-
| "detection_complete"
|
|
46
|
-
| "prompt_generated"
|
|
47
|
-
| "context_injected"
|
|
48
|
-
| "resumption_started"
|
|
49
|
-
| "tool_call_tracked";
|
|
50
|
-
/** Filter by session IDs */
|
|
51
|
-
sessionIds?: string[];
|
|
52
|
-
/** Limit number of results */
|
|
53
|
-
limit?: number;
|
|
54
|
-
}
|
|
55
|
-
|
|
56
|
-
/**
|
|
57
|
-
* Load COMPACTION events from session JSONL files
|
|
58
|
-
*
|
|
59
|
-
* Reads all .jsonl files in the session directory, parses events,
|
|
60
|
-
* and returns only COMPACTION events matching the filters.
|
|
61
|
-
*
|
|
62
|
-
* @param sessionDir - Path to session directory (default: ~/.config/swarm-tools/sessions)
|
|
63
|
-
* @param options - Filter options
|
|
64
|
-
* @returns Array of compaction events
|
|
65
|
-
*
|
|
66
|
-
* @example
|
|
67
|
-
* // Load all COMPACTION events
|
|
68
|
-
* const events = await loadCompactionEvents("/path/to/sessions");
|
|
69
|
-
*
|
|
70
|
-
* @example
|
|
71
|
-
* // Load only detection_complete events
|
|
72
|
-
* const events = await loadCompactionEvents("/path/to/sessions", {
|
|
73
|
-
* compaction_type: "detection_complete",
|
|
74
|
-
* });
|
|
75
|
-
*
|
|
76
|
-
* @example
|
|
77
|
-
* // Load events from specific sessions
|
|
78
|
-
* const events = await loadCompactionEvents("/path/to/sessions", {
|
|
79
|
-
* sessionIds: ["session-1", "session-2"],
|
|
80
|
-
* limit: 10,
|
|
81
|
-
* });
|
|
82
|
-
*/
|
|
83
|
-
export async function loadCompactionEvents(
|
|
84
|
-
sessionDir: string,
|
|
85
|
-
options?: LoadOptions,
|
|
86
|
-
): Promise<CompactionEvent[]> {
|
|
87
|
-
const { compaction_type, sessionIds, limit } = options ?? {};
|
|
88
|
-
|
|
89
|
-
// Check if directory exists
|
|
90
|
-
if (!fs.existsSync(sessionDir)) {
|
|
91
|
-
return [];
|
|
92
|
-
}
|
|
93
|
-
|
|
94
|
-
let files: string[];
|
|
95
|
-
try {
|
|
96
|
-
// Read all .jsonl files
|
|
97
|
-
files = fs.readdirSync(sessionDir).filter((f) => f.endsWith(".jsonl"));
|
|
98
|
-
} catch (error) {
|
|
99
|
-
// Directory exists but can't be read - log and return empty
|
|
100
|
-
console.warn(`Failed to read session directory ${sessionDir}:`, error);
|
|
101
|
-
return [];
|
|
102
|
-
}
|
|
103
|
-
|
|
104
|
-
// Filter by sessionIds if provided
|
|
105
|
-
const targetFiles = sessionIds
|
|
106
|
-
? files.filter((f) => sessionIds.includes(f.replace(".jsonl", "")))
|
|
107
|
-
: files;
|
|
108
|
-
|
|
109
|
-
const events: CompactionEvent[] = [];
|
|
110
|
-
|
|
111
|
-
for (const file of targetFiles) {
|
|
112
|
-
const filePath = path.join(sessionDir, file);
|
|
113
|
-
|
|
114
|
-
try {
|
|
115
|
-
// Stream large files line-by-line to avoid loading entire file into memory
|
|
116
|
-
const shouldStream = limit && limit < 100; // For small limits, streaming is overkill
|
|
117
|
-
|
|
118
|
-
if (shouldStream) {
|
|
119
|
-
// Use streaming for better memory efficiency
|
|
120
|
-
const found = await loadFromFileStream(filePath, {
|
|
121
|
-
compaction_type,
|
|
122
|
-
remainingLimit: limit - events.length,
|
|
123
|
-
});
|
|
124
|
-
events.push(...found);
|
|
125
|
-
} else {
|
|
126
|
-
// For small files or no limit, read entire file (faster)
|
|
127
|
-
const content = fs.readFileSync(filePath, "utf-8");
|
|
128
|
-
const lines = content.trim().split("\n").filter(Boolean);
|
|
129
|
-
|
|
130
|
-
for (const line of lines) {
|
|
131
|
-
const event = parseLine(line);
|
|
132
|
-
if (event && event.event_type === "COMPACTION") {
|
|
133
|
-
// Filter by compaction_type if provided
|
|
134
|
-
if (!compaction_type || event.compaction_type === compaction_type) {
|
|
135
|
-
events.push(event);
|
|
136
|
-
|
|
137
|
-
// Apply limit early to avoid processing unnecessary files
|
|
138
|
-
if (limit && events.length >= limit) {
|
|
139
|
-
return events.slice(0, limit);
|
|
140
|
-
}
|
|
141
|
-
}
|
|
142
|
-
}
|
|
143
|
-
}
|
|
144
|
-
}
|
|
145
|
-
|
|
146
|
-
// Early termination if limit reached
|
|
147
|
-
if (limit && events.length >= limit) {
|
|
148
|
-
return events.slice(0, limit);
|
|
149
|
-
}
|
|
150
|
-
} catch (error) {
|
|
151
|
-
// Log file read errors but continue processing other files
|
|
152
|
-
console.warn(`Failed to read session file ${filePath}:`, error);
|
|
153
|
-
}
|
|
154
|
-
}
|
|
155
|
-
|
|
156
|
-
return limit ? events.slice(0, limit) : events;
|
|
157
|
-
}
|
|
158
|
-
|
|
159
|
-
/**
|
|
160
|
-
* Parse a JSONL line into a CoordinatorEvent
|
|
161
|
-
*
|
|
162
|
-
* @param line - JSONL line to parse
|
|
163
|
-
* @returns Parsed and validated event, or null if invalid
|
|
164
|
-
*/
|
|
165
|
-
function parseLine(line: string): CoordinatorEvent | null {
|
|
166
|
-
try {
|
|
167
|
-
const parsed = JSON.parse(line);
|
|
168
|
-
return CoordinatorEventSchema.parse(parsed);
|
|
169
|
-
} catch {
|
|
170
|
-
// Invalid JSON or failed validation - skip silently
|
|
171
|
-
return null;
|
|
172
|
-
}
|
|
173
|
-
}
|
|
174
|
-
|
|
175
|
-
/**
|
|
176
|
-
* Load COMPACTION events from a file using streaming (for large files)
|
|
177
|
-
*
|
|
178
|
-
* @param filePath - Path to session JSONL file
|
|
179
|
-
* @param options - Filter options
|
|
180
|
-
* @returns Array of matching compaction events
|
|
181
|
-
*/
|
|
182
|
-
async function loadFromFileStream(
|
|
183
|
-
filePath: string,
|
|
184
|
-
options: {
|
|
185
|
-
compaction_type?: LoadOptions["compaction_type"];
|
|
186
|
-
remainingLimit?: number;
|
|
187
|
-
},
|
|
188
|
-
): Promise<CompactionEvent[]> {
|
|
189
|
-
const { compaction_type, remainingLimit } = options;
|
|
190
|
-
const events: CompactionEvent[] = [];
|
|
191
|
-
|
|
192
|
-
const fileStream = fs.createReadStream(filePath, { encoding: "utf-8" });
|
|
193
|
-
const rl = createInterface({ input: fileStream, crlfDelay: Number.POSITIVE_INFINITY });
|
|
194
|
-
|
|
195
|
-
for await (const line of rl) {
|
|
196
|
-
const event = parseLine(line);
|
|
197
|
-
if (event && event.event_type === "COMPACTION") {
|
|
198
|
-
if (!compaction_type || event.compaction_type === compaction_type) {
|
|
199
|
-
events.push(event);
|
|
200
|
-
|
|
201
|
-
// Early termination for streaming
|
|
202
|
-
if (remainingLimit && events.length >= remainingLimit) {
|
|
203
|
-
rl.close();
|
|
204
|
-
fileStream.close();
|
|
205
|
-
break;
|
|
206
|
-
}
|
|
207
|
-
}
|
|
208
|
-
}
|
|
209
|
-
}
|
|
210
|
-
|
|
211
|
-
return events;
|
|
212
|
-
}
|
|
213
|
-
|
|
214
|
-
/**
|
|
215
|
-
* Load COMPACTION sessions grouped by session_id
|
|
216
|
-
*
|
|
217
|
-
* Groups COMPACTION events by session_id and returns session metadata.
|
|
218
|
-
*
|
|
219
|
-
* @param sessionDir - Path to session directory
|
|
220
|
-
* @param options - Filter options
|
|
221
|
-
* @returns Array of compaction sessions
|
|
222
|
-
*
|
|
223
|
-
* @example
|
|
224
|
-
* // Load all sessions with COMPACTION events
|
|
225
|
-
* const sessions = await loadCompactionSessions("/path/to/sessions");
|
|
226
|
-
*
|
|
227
|
-
* @example
|
|
228
|
-
* // Load sessions with specific compaction_type
|
|
229
|
-
* const sessions = await loadCompactionSessions("/path/to/sessions", {
|
|
230
|
-
* compaction_type: "prompt_generated",
|
|
231
|
-
* });
|
|
232
|
-
*/
|
|
233
|
-
export async function loadCompactionSessions(
|
|
234
|
-
sessionDir: string,
|
|
235
|
-
options?: LoadOptions,
|
|
236
|
-
): Promise<CompactionSession[]> {
|
|
237
|
-
const events = await loadCompactionEvents(sessionDir, options);
|
|
238
|
-
|
|
239
|
-
if (events.length === 0) {
|
|
240
|
-
return [];
|
|
241
|
-
}
|
|
242
|
-
|
|
243
|
-
// Group events by session_id
|
|
244
|
-
const sessionMap = new Map<string, CompactionEvent[]>();
|
|
245
|
-
|
|
246
|
-
for (const event of events) {
|
|
247
|
-
const existing = sessionMap.get(event.session_id);
|
|
248
|
-
if (existing) {
|
|
249
|
-
existing.push(event);
|
|
250
|
-
} else {
|
|
251
|
-
sessionMap.set(event.session_id, [event]);
|
|
252
|
-
}
|
|
253
|
-
}
|
|
254
|
-
|
|
255
|
-
// Build sessions with metadata
|
|
256
|
-
const sessions: CompactionSession[] = [];
|
|
257
|
-
|
|
258
|
-
for (const [sessionId, sessionEvents] of sessionMap.entries()) {
|
|
259
|
-
if (sessionEvents.length === 0) {
|
|
260
|
-
continue;
|
|
261
|
-
}
|
|
262
|
-
|
|
263
|
-
// Get epic_id from first event
|
|
264
|
-
const epicId = sessionEvents[0].epic_id;
|
|
265
|
-
|
|
266
|
-
// Get timestamps
|
|
267
|
-
const timestamps = sessionEvents.map((e) => new Date(e.timestamp).getTime());
|
|
268
|
-
const startTime = new Date(Math.min(...timestamps)).toISOString();
|
|
269
|
-
const endTime = new Date(Math.max(...timestamps)).toISOString();
|
|
270
|
-
|
|
271
|
-
sessions.push({
|
|
272
|
-
session_id: sessionId,
|
|
273
|
-
epic_id: epicId,
|
|
274
|
-
start_time: startTime,
|
|
275
|
-
end_time: endTime,
|
|
276
|
-
events: sessionEvents,
|
|
277
|
-
});
|
|
278
|
-
}
|
|
279
|
-
|
|
280
|
-
// Apply limit
|
|
281
|
-
return options?.limit ? sessions.slice(0, options.limit) : sessions;
|
|
282
|
-
}
|
|
283
|
-
|
|
284
|
-
/**
|
|
285
|
-
* Load COMPACTION events from default session directory
|
|
286
|
-
*
|
|
287
|
-
* Convenience wrapper that uses the default ~/.config/swarm-tools/sessions directory.
|
|
288
|
-
*
|
|
289
|
-
* @param options - Filter options
|
|
290
|
-
* @returns Array of compaction events
|
|
291
|
-
*
|
|
292
|
-
* @example
|
|
293
|
-
* // Load recent compaction events
|
|
294
|
-
* const events = await loadDefaultCompactionEvents({ limit: 10 });
|
|
295
|
-
*/
|
|
296
|
-
export async function loadDefaultCompactionEvents(
|
|
297
|
-
options?: LoadOptions,
|
|
298
|
-
): Promise<CompactionEvent[]> {
|
|
299
|
-
const { getSessionDir } = await import("../../src/eval-capture.js");
|
|
300
|
-
return loadCompactionEvents(getSessionDir(), options);
|
|
301
|
-
}
|
|
302
|
-
|
|
303
|
-
/**
|
|
304
|
-
* Load COMPACTION sessions from default session directory
|
|
305
|
-
*
|
|
306
|
-
* Convenience wrapper that uses the default ~/.config/swarm-tools/sessions directory.
|
|
307
|
-
*
|
|
308
|
-
* @param options - Filter options
|
|
309
|
-
* @returns Array of compaction sessions
|
|
310
|
-
*
|
|
311
|
-
* @example
|
|
312
|
-
* // Load all compaction sessions
|
|
313
|
-
* const sessions = await loadDefaultCompactionSessions();
|
|
314
|
-
*/
|
|
315
|
-
export async function loadDefaultCompactionSessions(
|
|
316
|
-
options?: LoadOptions,
|
|
317
|
-
): Promise<CompactionSession[]> {
|
|
318
|
-
const { getSessionDir } = await import("../../src/eval-capture.js");
|
|
319
|
-
return loadCompactionSessions(getSessionDir(), options);
|
|
320
|
-
}
|
|
@@ -1,289 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Data Loader Tests
|
|
3
|
-
*
|
|
4
|
-
* Tests the PGlite-backed eval data loader functions.
|
|
5
|
-
* Uses a real in-memory PGlite database for accurate testing.
|
|
6
|
-
*/
|
|
7
|
-
import { describe, it, expect, beforeAll, afterAll } from "bun:test";
|
|
8
|
-
import {
|
|
9
|
-
loadEvalCases,
|
|
10
|
-
hasRealEvalData,
|
|
11
|
-
getEvalDataSummary,
|
|
12
|
-
} from "./data-loader.js";
|
|
13
|
-
import {
|
|
14
|
-
appendEvent,
|
|
15
|
-
getDatabase,
|
|
16
|
-
closeDatabase,
|
|
17
|
-
type DecompositionGeneratedEvent,
|
|
18
|
-
type SubtaskOutcomeEvent,
|
|
19
|
-
} from "swarm-mail";
|
|
20
|
-
import * as fs from "node:fs";
|
|
21
|
-
import * as path from "node:path";
|
|
22
|
-
import * as os from "node:os";
|
|
23
|
-
|
|
24
|
-
const TEST_PROJECT_KEY = "test-project-eval-loader";
|
|
25
|
-
|
|
26
|
-
// Create a unique temp directory for this test run
|
|
27
|
-
let testDir: string;
|
|
28
|
-
|
|
29
|
-
describe("Data Loader", () => {
|
|
30
|
-
beforeAll(async () => {
|
|
31
|
-
// Create temp directory for test database
|
|
32
|
-
testDir = fs.mkdtempSync(path.join(os.tmpdir(), "eval-loader-test-"));
|
|
33
|
-
|
|
34
|
-
// Initialize database by getting it (lazy init)
|
|
35
|
-
await getDatabase(testDir);
|
|
36
|
-
});
|
|
37
|
-
|
|
38
|
-
afterAll(async () => {
|
|
39
|
-
await closeDatabase(testDir);
|
|
40
|
-
// Clean up temp directory
|
|
41
|
-
fs.rmSync(testDir, { recursive: true, force: true });
|
|
42
|
-
});
|
|
43
|
-
|
|
44
|
-
describe("loadEvalCases", () => {
|
|
45
|
-
it("transforms eval records to EvalCase format", async () => {
|
|
46
|
-
// Insert a decomposition event
|
|
47
|
-
const decompositionEvent: DecompositionGeneratedEvent = {
|
|
48
|
-
type: "decomposition_generated",
|
|
49
|
-
timestamp: Date.now(),
|
|
50
|
-
project_key: TEST_PROJECT_KEY,
|
|
51
|
-
epic_id: "epic-load-1",
|
|
52
|
-
task: "Add authentication",
|
|
53
|
-
context: "Next.js app",
|
|
54
|
-
strategy: "feature-based",
|
|
55
|
-
epic_title: "Auth Epic",
|
|
56
|
-
subtasks: [
|
|
57
|
-
{ title: "OAuth setup", files: ["src/auth/oauth.ts"], priority: 1 },
|
|
58
|
-
{
|
|
59
|
-
title: "Session management",
|
|
60
|
-
files: ["src/auth/session.ts"],
|
|
61
|
-
priority: 2,
|
|
62
|
-
},
|
|
63
|
-
],
|
|
64
|
-
};
|
|
65
|
-
await appendEvent(decompositionEvent, testDir);
|
|
66
|
-
|
|
67
|
-
// Insert outcome events for both subtasks
|
|
68
|
-
const outcome1: SubtaskOutcomeEvent = {
|
|
69
|
-
type: "subtask_outcome",
|
|
70
|
-
timestamp: Date.now(),
|
|
71
|
-
project_key: TEST_PROJECT_KEY,
|
|
72
|
-
epic_id: "epic-load-1",
|
|
73
|
-
bead_id: "epic-load-1.1",
|
|
74
|
-
planned_files: ["src/auth/oauth.ts"],
|
|
75
|
-
actual_files: ["src/auth/oauth.ts"],
|
|
76
|
-
duration_ms: 5000,
|
|
77
|
-
error_count: 0,
|
|
78
|
-
retry_count: 0,
|
|
79
|
-
success: true,
|
|
80
|
-
};
|
|
81
|
-
await appendEvent(outcome1, testDir);
|
|
82
|
-
|
|
83
|
-
const outcome2: SubtaskOutcomeEvent = {
|
|
84
|
-
type: "subtask_outcome",
|
|
85
|
-
timestamp: Date.now(),
|
|
86
|
-
project_key: TEST_PROJECT_KEY,
|
|
87
|
-
epic_id: "epic-load-1",
|
|
88
|
-
bead_id: "epic-load-1.2",
|
|
89
|
-
planned_files: ["src/auth/session.ts"],
|
|
90
|
-
actual_files: ["src/auth/session.ts"],
|
|
91
|
-
duration_ms: 3000,
|
|
92
|
-
error_count: 0,
|
|
93
|
-
retry_count: 0,
|
|
94
|
-
success: true,
|
|
95
|
-
};
|
|
96
|
-
await appendEvent(outcome2, testDir);
|
|
97
|
-
|
|
98
|
-
const cases = await loadEvalCases(TEST_PROJECT_KEY, {
|
|
99
|
-
projectPath: testDir,
|
|
100
|
-
});
|
|
101
|
-
|
|
102
|
-
expect(cases.length).toBeGreaterThanOrEqual(1);
|
|
103
|
-
const authCase = cases.find((c) => c.input.task === "Add authentication");
|
|
104
|
-
expect(authCase).toBeDefined();
|
|
105
|
-
expect(authCase!.input.context).toBe("Next.js app");
|
|
106
|
-
expect(authCase!.expected.minSubtasks).toBe(2);
|
|
107
|
-
expect(authCase!.expected.maxSubtasks).toBe(2);
|
|
108
|
-
expect(authCase!.expected.requiredFiles).toContain("src/auth/oauth.ts");
|
|
109
|
-
expect(authCase!.expected.requiredFiles).toContain("src/auth/session.ts");
|
|
110
|
-
expect(authCase!.actual).toBeDefined();
|
|
111
|
-
});
|
|
112
|
-
|
|
113
|
-
it("filters by success when successOnly is true", async () => {
|
|
114
|
-
// Insert a successful decomposition
|
|
115
|
-
const successEvent: DecompositionGeneratedEvent = {
|
|
116
|
-
type: "decomposition_generated",
|
|
117
|
-
timestamp: Date.now(),
|
|
118
|
-
project_key: TEST_PROJECT_KEY,
|
|
119
|
-
epic_id: "epic-success-filter",
|
|
120
|
-
task: "Success task for filter",
|
|
121
|
-
strategy: "feature-based",
|
|
122
|
-
epic_title: "Success Epic",
|
|
123
|
-
subtasks: [{ title: "Sub", files: ["src/success.ts"], priority: 1 }],
|
|
124
|
-
};
|
|
125
|
-
await appendEvent(successEvent, testDir);
|
|
126
|
-
|
|
127
|
-
// Mark it successful
|
|
128
|
-
const successOutcome: SubtaskOutcomeEvent = {
|
|
129
|
-
type: "subtask_outcome",
|
|
130
|
-
timestamp: Date.now(),
|
|
131
|
-
project_key: TEST_PROJECT_KEY,
|
|
132
|
-
epic_id: "epic-success-filter",
|
|
133
|
-
bead_id: "epic-success-filter.1",
|
|
134
|
-
planned_files: ["src/success.ts"],
|
|
135
|
-
actual_files: ["src/success.ts"],
|
|
136
|
-
duration_ms: 1000,
|
|
137
|
-
error_count: 0,
|
|
138
|
-
retry_count: 0,
|
|
139
|
-
success: true,
|
|
140
|
-
};
|
|
141
|
-
await appendEvent(successOutcome, testDir);
|
|
142
|
-
|
|
143
|
-
// Insert a failed decomposition
|
|
144
|
-
const failEvent: DecompositionGeneratedEvent = {
|
|
145
|
-
type: "decomposition_generated",
|
|
146
|
-
timestamp: Date.now(),
|
|
147
|
-
project_key: TEST_PROJECT_KEY,
|
|
148
|
-
epic_id: "epic-fail-filter",
|
|
149
|
-
task: "Failed task for filter",
|
|
150
|
-
strategy: "feature-based",
|
|
151
|
-
epic_title: "Failed Epic",
|
|
152
|
-
subtasks: [{ title: "Sub", files: ["src/fail.ts"], priority: 1 }],
|
|
153
|
-
};
|
|
154
|
-
await appendEvent(failEvent, testDir);
|
|
155
|
-
|
|
156
|
-
// Mark it failed
|
|
157
|
-
const failOutcome: SubtaskOutcomeEvent = {
|
|
158
|
-
type: "subtask_outcome",
|
|
159
|
-
timestamp: Date.now(),
|
|
160
|
-
project_key: TEST_PROJECT_KEY,
|
|
161
|
-
epic_id: "epic-fail-filter",
|
|
162
|
-
bead_id: "epic-fail-filter.1",
|
|
163
|
-
planned_files: ["src/fail.ts"],
|
|
164
|
-
actual_files: [],
|
|
165
|
-
duration_ms: 500,
|
|
166
|
-
error_count: 3,
|
|
167
|
-
retry_count: 2,
|
|
168
|
-
success: false,
|
|
169
|
-
};
|
|
170
|
-
await appendEvent(failOutcome, testDir);
|
|
171
|
-
|
|
172
|
-
const successCases = await loadEvalCases(TEST_PROJECT_KEY, {
|
|
173
|
-
successOnly: true,
|
|
174
|
-
projectPath: testDir,
|
|
175
|
-
});
|
|
176
|
-
|
|
177
|
-
// Should only include successful cases
|
|
178
|
-
const failedCase = successCases.find(
|
|
179
|
-
(c) => c.input.task === "Failed task for filter",
|
|
180
|
-
);
|
|
181
|
-
expect(failedCase).toBeUndefined();
|
|
182
|
-
});
|
|
183
|
-
|
|
184
|
-
it("passes strategy filter to getEvalRecords", async () => {
|
|
185
|
-
// Insert file-based decomposition
|
|
186
|
-
const fileBasedEvent: DecompositionGeneratedEvent = {
|
|
187
|
-
type: "decomposition_generated",
|
|
188
|
-
timestamp: Date.now(),
|
|
189
|
-
project_key: TEST_PROJECT_KEY,
|
|
190
|
-
epic_id: "epic-file-based",
|
|
191
|
-
task: "File-based task",
|
|
192
|
-
strategy: "file-based",
|
|
193
|
-
epic_title: "File Epic",
|
|
194
|
-
subtasks: [{ title: "Sub", files: ["src/file.ts"], priority: 1 }],
|
|
195
|
-
};
|
|
196
|
-
await appendEvent(fileBasedEvent, testDir);
|
|
197
|
-
|
|
198
|
-
const fileBasedCases = await loadEvalCases(TEST_PROJECT_KEY, {
|
|
199
|
-
strategy: "file-based",
|
|
200
|
-
projectPath: testDir,
|
|
201
|
-
});
|
|
202
|
-
|
|
203
|
-
// All returned cases should be file-based
|
|
204
|
-
for (const c of fileBasedCases) {
|
|
205
|
-
expect(c.actual?.strategy).toBe("file-based");
|
|
206
|
-
}
|
|
207
|
-
});
|
|
208
|
-
|
|
209
|
-
it("passes limit to getEvalRecords", async () => {
|
|
210
|
-
const cases = await loadEvalCases(TEST_PROJECT_KEY, {
|
|
211
|
-
limit: 2,
|
|
212
|
-
projectPath: testDir,
|
|
213
|
-
});
|
|
214
|
-
|
|
215
|
-
expect(cases.length).toBeLessThanOrEqual(2);
|
|
216
|
-
});
|
|
217
|
-
|
|
218
|
-
it("handles records with no context", async () => {
|
|
219
|
-
const noContextEvent: DecompositionGeneratedEvent = {
|
|
220
|
-
type: "decomposition_generated",
|
|
221
|
-
timestamp: Date.now(),
|
|
222
|
-
project_key: TEST_PROJECT_KEY,
|
|
223
|
-
epic_id: "epic-no-context",
|
|
224
|
-
task: "Task without context",
|
|
225
|
-
// context is undefined
|
|
226
|
-
strategy: "feature-based",
|
|
227
|
-
epic_title: "No Context Epic",
|
|
228
|
-
subtasks: [{ title: "Sub", files: [], priority: 1 }],
|
|
229
|
-
};
|
|
230
|
-
await appendEvent(noContextEvent, testDir);
|
|
231
|
-
|
|
232
|
-
const cases = await loadEvalCases(TEST_PROJECT_KEY, {
|
|
233
|
-
projectPath: testDir,
|
|
234
|
-
});
|
|
235
|
-
const noContextCase = cases.find(
|
|
236
|
-
(c) => c.input.task === "Task without context",
|
|
237
|
-
);
|
|
238
|
-
|
|
239
|
-
expect(noContextCase).toBeDefined();
|
|
240
|
-
expect(noContextCase!.input.context).toBeUndefined();
|
|
241
|
-
});
|
|
242
|
-
});
|
|
243
|
-
|
|
244
|
-
describe("hasRealEvalData", () => {
|
|
245
|
-
it("returns true when enough records exist", async () => {
|
|
246
|
-
// We've inserted several records above, should have enough
|
|
247
|
-
const hasData = await hasRealEvalData(TEST_PROJECT_KEY, 1, testDir);
|
|
248
|
-
expect(hasData).toBe(true);
|
|
249
|
-
});
|
|
250
|
-
|
|
251
|
-
it("returns false when not enough records exist", async () => {
|
|
252
|
-
// Use a project key with no data
|
|
253
|
-
const hasData = await hasRealEvalData("nonexistent-project", 5, testDir);
|
|
254
|
-
expect(hasData).toBe(false);
|
|
255
|
-
});
|
|
256
|
-
|
|
257
|
-
it("uses custom minRecords threshold", async () => {
|
|
258
|
-
// Should have at least 1 record
|
|
259
|
-
const hasData = await hasRealEvalData(TEST_PROJECT_KEY, 1, testDir);
|
|
260
|
-
expect(hasData).toBe(true);
|
|
261
|
-
|
|
262
|
-
// Should not have 1000 records
|
|
263
|
-
const hasLotsOfData = await hasRealEvalData(
|
|
264
|
-
TEST_PROJECT_KEY,
|
|
265
|
-
1000,
|
|
266
|
-
testDir,
|
|
267
|
-
);
|
|
268
|
-
expect(hasLotsOfData).toBe(false);
|
|
269
|
-
});
|
|
270
|
-
});
|
|
271
|
-
|
|
272
|
-
describe("getEvalDataSummary", () => {
|
|
273
|
-
it("returns formatted summary with hasEnoughData flag", async () => {
|
|
274
|
-
const summary = await getEvalDataSummary(TEST_PROJECT_KEY, testDir);
|
|
275
|
-
|
|
276
|
-
expect(summary.totalRecords).toBeGreaterThanOrEqual(1);
|
|
277
|
-
expect(typeof summary.successRate).toBe("number");
|
|
278
|
-
expect(typeof summary.byStrategy).toBe("object");
|
|
279
|
-
expect(typeof summary.hasEnoughData).toBe("boolean");
|
|
280
|
-
});
|
|
281
|
-
|
|
282
|
-
it("sets hasEnoughData based on record count", async () => {
|
|
283
|
-
// Empty project should not have enough data
|
|
284
|
-
const emptySummary = await getEvalDataSummary("empty-project", testDir);
|
|
285
|
-
expect(emptySummary.hasEnoughData).toBe(false);
|
|
286
|
-
expect(emptySummary.totalRecords).toBe(0);
|
|
287
|
-
});
|
|
288
|
-
});
|
|
289
|
-
});
|