oh-my-opencode 3.8.1 → 3.8.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/dist/agents/atlas/agent.d.ts +3 -2
  2. package/dist/agents/atlas/gemini.d.ts +11 -0
  3. package/dist/agents/atlas/index.d.ts +1 -0
  4. package/dist/agents/prometheus/gemini.d.ts +12 -0
  5. package/dist/agents/prometheus/index.d.ts +1 -0
  6. package/dist/agents/prometheus/system-prompt.d.ts +2 -1
  7. package/dist/agents/sisyphus-gemini-overlays.d.ts +15 -0
  8. package/dist/agents/sisyphus-junior/agent.d.ts +3 -2
  9. package/dist/agents/sisyphus-junior/gemini.d.ts +10 -0
  10. package/dist/agents/sisyphus-junior/index.d.ts +1 -0
  11. package/dist/agents/types.d.ts +1 -0
  12. package/dist/cli/index.js +8 -8
  13. package/dist/hooks/atlas/system-reminder-templates.d.ts +1 -0
  14. package/dist/hooks/session-notification.d.ts +2 -0
  15. package/dist/index.js +1996 -354
  16. package/dist/shared/command-executor/execute-hook-command.d.ts +2 -0
  17. package/dist/shared/session-tools-store.d.ts +1 -0
  18. package/dist/tools/glob/constants.d.ts +1 -1
  19. package/dist/tools/glob/types.d.ts +1 -0
  20. package/dist/tools/grep/constants.d.ts +2 -1
  21. package/dist/tools/grep/types.d.ts +3 -0
  22. package/dist/tools/hashline-edit/autocorrect-replacement-lines.d.ts +6 -0
  23. package/dist/tools/hashline-edit/edit-deduplication.d.ts +5 -0
  24. package/dist/tools/hashline-edit/edit-operation-primitives.d.ts +12 -0
  25. package/dist/tools/hashline-edit/edit-operations.d.ts +1 -6
  26. package/dist/tools/hashline-edit/edit-ordering.d.ts +3 -0
  27. package/dist/tools/hashline-edit/file-text-canonicalization.d.ts +7 -0
  28. package/dist/tools/hashline-edit/hashline-edit-diff.d.ts +1 -0
  29. package/dist/tools/hashline-edit/hashline-edit-executor.d.ts +10 -0
  30. package/dist/tools/hashline-edit/index.d.ts +1 -1
  31. package/dist/tools/hashline-edit/tool-description.d.ts +1 -1
  32. package/dist/tools/hashline-edit/types.d.ts +9 -1
  33. package/dist/tools/lsp/lsp-manager-process-cleanup.d.ts +4 -1
  34. package/dist/tools/lsp/lsp-server.d.ts +2 -1
  35. package/dist/tools/shared/semaphore.d.ts +14 -0
  36. package/package.json +8 -8
package/dist/index.js CHANGED
@@ -15029,8 +15029,11 @@ function findBashPath() {
15029
15029
  }
15030
15030
 
15031
15031
  // src/shared/command-executor/execute-hook-command.ts
15032
+ var DEFAULT_HOOK_TIMEOUT_MS = 30000;
15033
+ var SIGKILL_GRACE_MS = 5000;
15032
15034
  async function executeHookCommand(command, stdin, cwd, options) {
15033
15035
  const home = getHomeDirectory();
15036
+ const timeoutMs = options?.timeoutMs ?? DEFAULT_HOOK_TIMEOUT_MS;
15034
15037
  const expandedCommand = command.replace(/^~(?=\/|$)/g, home).replace(/\s~(?=\/)/g, ` ${home}`).replace(/\$CLAUDE_PROJECT_DIR/g, cwd).replace(/\$\{CLAUDE_PROJECT_DIR\}/g, cwd);
15035
15038
  let finalCommand = expandedCommand;
15036
15039
  if (options?.forceZsh) {
@@ -15046,9 +15049,13 @@ async function executeHookCommand(command, stdin, cwd, options) {
15046
15049
  }
15047
15050
  }
15048
15051
  return new Promise((resolve) => {
15052
+ let settled = false;
15053
+ let killTimer = null;
15054
+ const isWin32 = process.platform === "win32";
15049
15055
  const proc = spawn(finalCommand, {
15050
15056
  cwd,
15051
15057
  shell: true,
15058
+ detached: !isWin32,
15052
15059
  env: { ...process.env, HOME: home, CLAUDE_PROJECT_DIR: cwd }
15053
15060
  });
15054
15061
  let stdout = "";
@@ -15059,18 +15066,57 @@ async function executeHookCommand(command, stdin, cwd, options) {
15059
15066
  proc.stderr?.on("data", (data) => {
15060
15067
  stderr += data.toString();
15061
15068
  });
15069
+ proc.stdin?.on("error", () => {});
15062
15070
  proc.stdin?.write(stdin);
15063
15071
  proc.stdin?.end();
15072
+ const settle = (result) => {
15073
+ if (settled)
15074
+ return;
15075
+ settled = true;
15076
+ if (killTimer)
15077
+ clearTimeout(killTimer);
15078
+ if (timeoutTimer)
15079
+ clearTimeout(timeoutTimer);
15080
+ resolve(result);
15081
+ };
15064
15082
  proc.on("close", (code) => {
15065
- resolve({
15066
- exitCode: code ?? 0,
15083
+ settle({
15084
+ exitCode: code ?? 1,
15067
15085
  stdout: stdout.trim(),
15068
15086
  stderr: stderr.trim()
15069
15087
  });
15070
15088
  });
15071
15089
  proc.on("error", (err) => {
15072
- resolve({ exitCode: 1, stderr: err.message });
15090
+ settle({ exitCode: 1, stderr: err.message });
15073
15091
  });
15092
+ const killProcessGroup = (signal) => {
15093
+ try {
15094
+ if (!isWin32 && proc.pid) {
15095
+ try {
15096
+ process.kill(-proc.pid, signal);
15097
+ } catch {
15098
+ proc.kill(signal);
15099
+ }
15100
+ } else {
15101
+ proc.kill(signal);
15102
+ }
15103
+ } catch {}
15104
+ };
15105
+ const timeoutTimer = setTimeout(() => {
15106
+ if (settled)
15107
+ return;
15108
+ killProcessGroup("SIGTERM");
15109
+ killTimer = setTimeout(() => {
15110
+ if (settled)
15111
+ return;
15112
+ killProcessGroup("SIGKILL");
15113
+ }, SIGKILL_GRACE_MS);
15114
+ stderr += `
15115
+ Hook command timed out after ${timeoutMs}ms`;
15116
+ }, timeoutMs);
15117
+ if (timeoutTimer && typeof timeoutTimer === "object" && "unref" in timeoutTimer) {
15118
+ timeoutTimer.unref();
15119
+ }
15074
15120
  });
15075
15121
  }
15076
15122
  // src/shared/command-executor/execute-command.ts
@@ -19037,6 +19083,9 @@ function getSessionTools(sessionID) {
19037
19083
  const tools = store.get(sessionID);
19038
19084
  return tools ? { ...tools } : undefined;
19039
19085
  }
19086
+ function deleteSessionTools(sessionID) {
19087
+ store.delete(sessionID);
19088
+ }
19040
19089
 
19041
19090
  // src/shared/prompt-tools.ts
19042
19091
  function normalizePromptTools(tools) {
@@ -20047,6 +20096,8 @@ function createSessionNotification(ctx, config = {}) {
20047
20096
  const mergedConfig = {
20048
20097
  title: "OpenCode",
20049
20098
  message: "Agent is ready for input",
20099
+ questionMessage: "Agent is asking a question",
20100
+ permissionMessage: "Agent needs permission to continue",
20050
20101
  playSound: false,
20051
20102
  soundPath: defaultSoundPath,
20052
20103
  idleConfirmationDelay: 1500,
@@ -20062,6 +20113,51 @@ function createSessionNotification(ctx, config = {}) {
20062
20113
  send: sendSessionNotification,
20063
20114
  playSound: playSessionNotificationSound
20064
20115
  });
20116
+ const QUESTION_TOOLS = new Set(["question", "ask_user_question", "askuserquestion"]);
20117
+ const PERMISSION_EVENTS = new Set(["permission.ask", "permission.asked", "permission.updated", "permission.requested"]);
20118
+ const PERMISSION_HINT_PATTERN = /\b(permission|approve|approval|allow|deny|consent)\b/i;
20119
+ const getSessionID = (properties) => {
20120
+ const sessionID = properties?.sessionID;
20121
+ if (typeof sessionID === "string" && sessionID.length > 0)
20122
+ return sessionID;
20123
+ const sessionId = properties?.sessionId;
20124
+ if (typeof sessionId === "string" && sessionId.length > 0)
20125
+ return sessionId;
20126
+ const info = properties?.info;
20127
+ const infoSessionID = info?.sessionID;
20128
+ if (typeof infoSessionID === "string" && infoSessionID.length > 0)
20129
+ return infoSessionID;
20130
+ const infoSessionId = info?.sessionId;
20131
+ if (typeof infoSessionId === "string" && infoSessionId.length > 0)
20132
+ return infoSessionId;
20133
+ return;
20134
+ };
20135
+ const shouldNotifyForSession = (sessionID) => {
20136
+ if (subagentSessions.has(sessionID))
20137
+ return false;
20138
+ const mainSessionID = getMainSessionID();
20139
+ if (mainSessionID && sessionID !== mainSessionID)
20140
+ return false;
20141
+ return true;
20142
+ };
20143
+ const getEventToolName = (properties) => {
20144
+ const tool = properties?.tool;
20145
+ if (typeof tool === "string" && tool.length > 0)
20146
+ return tool;
20147
+ const name = properties?.name;
20148
+ if (typeof name === "string" && name.length > 0)
20149
+ return name;
20150
+ return;
20151
+ };
20152
+ const getQuestionText = (properties) => {
20153
+ const args = properties?.args;
20154
+ const questions = args?.questions;
20155
+ if (!Array.isArray(questions) || questions.length === 0)
20156
+ return "";
20157
+ const firstQuestion = questions[0];
20158
+ const questionText = firstQuestion?.question;
20159
+ return typeof questionText === "string" ? questionText : "";
20160
+ };
20065
20161
  return async ({ event }) => {
20066
20162
  if (currentPlatform === "unsupported")
20067
20163
  return;
@@ -20075,29 +20171,52 @@ function createSessionNotification(ctx, config = {}) {
20075
20171
  return;
20076
20172
  }
20077
20173
  if (event.type === "session.idle") {
20078
- const sessionID = props?.sessionID;
20174
+ const sessionID = getSessionID(props);
20079
20175
  if (!sessionID)
20080
20176
  return;
20081
- if (subagentSessions.has(sessionID))
20082
- return;
20083
- const mainSessionID = getMainSessionID();
20084
- if (mainSessionID && sessionID !== mainSessionID)
20177
+ if (!shouldNotifyForSession(sessionID))
20085
20178
  return;
20086
20179
  scheduler.scheduleIdleNotification(sessionID);
20087
20180
  return;
20088
20181
  }
20089
20182
  if (event.type === "message.updated") {
20090
20183
  const info = props?.info;
20091
- const sessionID = info?.sessionID;
20184
+ const sessionID = getSessionID({ ...props, info });
20092
20185
  if (sessionID) {
20093
20186
  scheduler.markSessionActivity(sessionID);
20094
20187
  }
20095
20188
  return;
20096
20189
  }
20190
+ if (PERMISSION_EVENTS.has(event.type)) {
20191
+ const sessionID = getSessionID(props);
20192
+ if (!sessionID)
20193
+ return;
20194
+ if (!shouldNotifyForSession(sessionID))
20195
+ return;
20196
+ scheduler.markSessionActivity(sessionID);
20197
+ await sendSessionNotification(ctx, currentPlatform, mergedConfig.title, mergedConfig.permissionMessage);
20198
+ if (mergedConfig.playSound && mergedConfig.soundPath) {
20199
+ await playSessionNotificationSound(ctx, currentPlatform, mergedConfig.soundPath);
20200
+ }
20201
+ return;
20202
+ }
20097
20203
  if (event.type === "tool.execute.before" || event.type === "tool.execute.after") {
20098
- const sessionID = props?.sessionID;
20204
+ const sessionID = getSessionID(props);
20099
20205
  if (sessionID) {
20100
20206
  scheduler.markSessionActivity(sessionID);
20207
+ if (event.type === "tool.execute.before") {
20208
+ const toolName = getEventToolName(props)?.toLowerCase();
20209
+ if (toolName && QUESTION_TOOLS.has(toolName)) {
20210
+ if (!shouldNotifyForSession(sessionID))
20211
+ return;
20212
+ const questionText = getQuestionText(props);
20213
+ const message = PERMISSION_HINT_PATTERN.test(questionText) ? mergedConfig.permissionMessage : mergedConfig.questionMessage;
20214
+ await sendSessionNotification(ctx, currentPlatform, mergedConfig.title, message);
20215
+ if (mergedConfig.playSound && mergedConfig.soundPath) {
20216
+ await playSessionNotificationSound(ctx, currentPlatform, mergedConfig.soundPath);
20217
+ }
20218
+ }
20219
+ }
20101
20220
  }
20102
20221
  return;
20103
20222
  }
@@ -38863,6 +38982,15 @@ function isGptModel(model) {
38863
38982
  const modelName = extractModelName(model).toLowerCase();
38864
38983
  return GPT_MODEL_PREFIXES.some((prefix) => modelName.startsWith(prefix));
38865
38984
  }
38985
+ var GEMINI_PROVIDERS = ["google/", "google-vertex/"];
38986
+ function isGeminiModel(model) {
38987
+ if (GEMINI_PROVIDERS.some((prefix) => model.startsWith(prefix)))
38988
+ return true;
38989
+ if (model.startsWith("github-copilot/") && extractModelName(model).toLowerCase().startsWith("gemini"))
38990
+ return true;
38991
+ const modelName = extractModelName(model).toLowerCase();
38992
+ return modelName.startsWith("gemini-");
38993
+ }
38866
38994
 
38867
38995
  // src/hooks/keyword-detector/ultrawork/source-detector.ts
38868
38996
  function isPlannerAgent(agentName) {
@@ -49831,55 +49959,6 @@ function spawnProcess(command, options) {
49831
49959
  });
49832
49960
  return proc;
49833
49961
  }
49834
- // src/tools/lsp/lsp-manager-process-cleanup.ts
49835
- function registerLspManagerProcessCleanup(options) {
49836
- const syncCleanup = () => {
49837
- for (const [, managed] of options.getClients()) {
49838
- try {
49839
- managed.client.stop().catch(() => {});
49840
- } catch {}
49841
- }
49842
- options.clearClients();
49843
- options.clearCleanupInterval();
49844
- };
49845
- const asyncCleanup = async () => {
49846
- const stopPromises = [];
49847
- for (const [, managed] of options.getClients()) {
49848
- stopPromises.push(managed.client.stop().catch(() => {}));
49849
- }
49850
- await Promise.allSettled(stopPromises);
49851
- options.clearClients();
49852
- options.clearCleanupInterval();
49853
- };
49854
- process.on("exit", syncCleanup);
49855
- process.on("SIGINT", () => void asyncCleanup().catch(() => {}));
49856
- process.on("SIGTERM", () => void asyncCleanup().catch(() => {}));
49857
- if (process.platform === "win32") {
49858
- process.on("SIGBREAK", () => void asyncCleanup().catch(() => {}));
49859
- }
49860
- }
49861
-
49862
- // src/tools/lsp/lsp-manager-temp-directory-cleanup.ts
49863
- async function cleanupTempDirectoryLspClients(clients) {
49864
- const keysToRemove = [];
49865
- for (const [key, managed] of clients.entries()) {
49866
- const isTempDir = key.startsWith("/tmp/") || key.startsWith("/var/folders/");
49867
- const isIdle = managed.refCount === 0;
49868
- if (isTempDir && isIdle) {
49869
- keysToRemove.push(key);
49870
- }
49871
- }
49872
- for (const key of keysToRemove) {
49873
- const managed = clients.get(key);
49874
- if (managed) {
49875
- clients.delete(key);
49876
- try {
49877
- await managed.client.stop();
49878
- } catch {}
49879
- }
49880
- }
49881
- }
49882
-
49883
49962
  // src/tools/lsp/lsp-client.ts
49884
49963
  import { readFileSync as readFileSync38 } from "fs";
49885
49964
  import { extname as extname2, resolve as resolve8 } from "path";
@@ -50241,6 +50320,69 @@ class LSPClient extends LSPClientConnection {
50241
50320
  }
50242
50321
  }
50243
50322
 
50323
+ // src/tools/lsp/lsp-manager-process-cleanup.ts
50324
+ function registerLspManagerProcessCleanup(options) {
50325
+ const handlers = [];
50326
+ const syncCleanup = () => {
50327
+ for (const [, managed] of options.getClients()) {
50328
+ try {
50329
+ managed.client.stop().catch(() => {});
50330
+ } catch {}
50331
+ }
50332
+ options.clearClients();
50333
+ options.clearCleanupInterval();
50334
+ };
50335
+ const asyncCleanup = async () => {
50336
+ const stopPromises = [];
50337
+ for (const [, managed] of options.getClients()) {
50338
+ stopPromises.push(managed.client.stop().catch(() => {}));
50339
+ }
50340
+ await Promise.allSettled(stopPromises);
50341
+ options.clearClients();
50342
+ options.clearCleanupInterval();
50343
+ };
50344
+ const registerHandler = (event, listener) => {
50345
+ handlers.push({ event, listener });
50346
+ process.on(event, listener);
50347
+ };
50348
+ registerHandler("exit", syncCleanup);
50349
+ const signalCleanup = () => void asyncCleanup().catch(() => {});
50350
+ registerHandler("SIGINT", signalCleanup);
50351
+ registerHandler("SIGTERM", signalCleanup);
50352
+ if (process.platform === "win32") {
50353
+ registerHandler("SIGBREAK", signalCleanup);
50354
+ }
50355
+ return {
50356
+ unregister: () => {
50357
+ for (const { event, listener } of handlers) {
50358
+ process.off(event, listener);
50359
+ }
50360
+ handlers.length = 0;
50361
+ }
50362
+ };
50363
+ }
50364
+
50365
+ // src/tools/lsp/lsp-manager-temp-directory-cleanup.ts
50366
+ async function cleanupTempDirectoryLspClients(clients) {
50367
+ const keysToRemove = [];
50368
+ for (const [key, managed] of clients.entries()) {
50369
+ const isTempDir = key.startsWith("/tmp/") || key.startsWith("/var/folders/");
50370
+ const isIdle = managed.refCount === 0;
50371
+ if (isTempDir && isIdle) {
50372
+ keysToRemove.push(key);
50373
+ }
50374
+ }
50375
+ for (const key of keysToRemove) {
50376
+ const managed = clients.get(key);
50377
+ if (managed) {
50378
+ clients.delete(key);
50379
+ try {
50380
+ await managed.client.stop();
50381
+ } catch {}
50382
+ }
50383
+ }
50384
+ }
50385
+
50244
50386
  // src/tools/lsp/lsp-server.ts
50245
50387
  class LSPServerManager {
50246
50388
  static instance;
@@ -50248,12 +50390,13 @@ class LSPServerManager {
50248
50390
  cleanupInterval = null;
50249
50391
  IDLE_TIMEOUT = 5 * 60 * 1000;
50250
50392
  INIT_TIMEOUT = 60 * 1000;
50393
+ cleanupHandle = null;
50251
50394
  constructor() {
50252
50395
  this.startCleanupTimer();
50253
50396
  this.registerProcessCleanup();
50254
50397
  }
50255
50398
  registerProcessCleanup() {
50256
- registerLspManagerProcessCleanup({
50399
+ this.cleanupHandle = registerLspManagerProcessCleanup({
50257
50400
  getClients: () => this.clients.entries(),
50258
50401
  clearClients: () => {
50259
50402
  this.clients.clear();
@@ -50403,6 +50546,8 @@ class LSPServerManager {
50403
50546
  return managed?.isInitializing ?? false;
50404
50547
  }
50405
50548
  async stopAll() {
50549
+ this.cleanupHandle?.unregister();
50550
+ this.cleanupHandle = null;
50406
50551
  for (const [, managed] of this.clients) {
50407
50552
  await managed.client.stop();
50408
50553
  }
@@ -51663,8 +51808,9 @@ var DEFAULT_MAX_DEPTH = 20;
51663
51808
  var DEFAULT_MAX_FILESIZE = "10M";
51664
51809
  var DEFAULT_MAX_COUNT = 500;
51665
51810
  var DEFAULT_MAX_COLUMNS = 1000;
51666
- var DEFAULT_TIMEOUT_MS3 = 300000;
51667
- var DEFAULT_MAX_OUTPUT_BYTES2 = 10 * 1024 * 1024;
51811
+ var DEFAULT_TIMEOUT_MS3 = 60000;
51812
+ var DEFAULT_MAX_OUTPUT_BYTES2 = 256 * 1024;
51813
+ var DEFAULT_RG_THREADS = 4;
51668
51814
  var RG_SAFETY_FLAGS = [
51669
51815
  "--no-follow",
51670
51816
  "--color=never",
@@ -51674,10 +51820,40 @@ var RG_SAFETY_FLAGS = [
51674
51820
  ];
51675
51821
  var GREP_SAFETY_FLAGS = ["-n", "-H", "--color=never"];
51676
51822
 
51823
+ // src/tools/shared/semaphore.ts
51824
+ class Semaphore {
51825
+ max;
51826
+ queue = [];
51827
+ running = 0;
51828
+ constructor(max) {
51829
+ this.max = max;
51830
+ }
51831
+ async acquire() {
51832
+ if (this.running < this.max) {
51833
+ this.running++;
51834
+ return;
51835
+ }
51836
+ return new Promise((resolve10) => {
51837
+ this.queue.push(() => {
51838
+ this.running++;
51839
+ resolve10();
51840
+ });
51841
+ });
51842
+ }
51843
+ release() {
51844
+ this.running--;
51845
+ const next = this.queue.shift();
51846
+ if (next)
51847
+ next();
51848
+ }
51849
+ }
51850
+ var rgSemaphore = new Semaphore(2);
51851
+
51677
51852
  // src/tools/grep/cli.ts
51678
51853
  function buildRgArgs(options) {
51679
51854
  const args = [
51680
51855
  ...RG_SAFETY_FLAGS,
51856
+ `--threads=${Math.min(options.threads ?? DEFAULT_RG_THREADS, DEFAULT_RG_THREADS)}`,
51681
51857
  `--max-depth=${Math.min(options.maxDepth ?? DEFAULT_MAX_DEPTH, DEFAULT_MAX_DEPTH)}`,
51682
51858
  `--max-filesize=${options.maxFilesize ?? DEFAULT_MAX_FILESIZE}`,
51683
51859
  `--max-count=${Math.min(options.maxCount ?? DEFAULT_MAX_COUNT, DEFAULT_MAX_COUNT)}`,
@@ -51713,6 +51889,11 @@ function buildRgArgs(options) {
51713
51889
  args.push(`--glob=!${glob}`);
51714
51890
  }
51715
51891
  }
51892
+ if (options.outputMode === "files_with_matches") {
51893
+ args.push("--files-with-matches");
51894
+ } else if (options.outputMode === "count") {
51895
+ args.push("--count");
51896
+ }
51716
51897
  return args;
51717
51898
  }
51718
51899
  function buildGrepArgs(options) {
@@ -51742,7 +51923,7 @@ function buildGrepArgs(options) {
51742
51923
  function buildArgs(options, backend) {
51743
51924
  return backend === "rg" ? buildRgArgs(options) : buildGrepArgs(options);
51744
51925
  }
51745
- function parseOutput(output) {
51926
+ function parseOutput(output, filesOnly = false) {
51746
51927
  if (!output.trim())
51747
51928
  return [];
51748
51929
  const matches = [];
@@ -51751,6 +51932,14 @@ function parseOutput(output) {
51751
51932
  for (const line of lines) {
51752
51933
  if (!line.trim())
51753
51934
  continue;
51935
+ if (filesOnly) {
51936
+ matches.push({
51937
+ file: line.trim(),
51938
+ line: 0,
51939
+ text: ""
51940
+ });
51941
+ continue;
51942
+ }
51754
51943
  const match = line.match(/^(.+?):(\d+):(.*)$/);
51755
51944
  if (match) {
51756
51945
  matches.push({
@@ -51762,7 +51951,34 @@ function parseOutput(output) {
51762
51951
  }
51763
51952
  return matches;
51764
51953
  }
51954
+ function parseCountOutput(output) {
51955
+ if (!output.trim())
51956
+ return [];
51957
+ const results = [];
51958
+ const lines = output.split(`
51959
+ `);
51960
+ for (const line of lines) {
51961
+ if (!line.trim())
51962
+ continue;
51963
+ const match = line.match(/^(.+?):(\d+)$/);
51964
+ if (match) {
51965
+ results.push({
51966
+ file: match[1],
51967
+ count: parseInt(match[2], 10)
51968
+ });
51969
+ }
51970
+ }
51971
+ return results;
51972
+ }
51765
51973
  async function runRg(options) {
51974
+ await rgSemaphore.acquire();
51975
+ try {
51976
+ return await runRgInternal(options);
51977
+ } finally {
51978
+ rgSemaphore.release();
51979
+ }
51980
+ }
51981
+ async function runRgInternal(options) {
51766
51982
  const cli = resolveGrepCli();
51767
51983
  const args = buildArgs(options, cli.backend);
51768
51984
  const timeout = Math.min(options.timeout ?? DEFAULT_TIMEOUT_MS3, DEFAULT_TIMEOUT_MS3);
@@ -51799,13 +52015,14 @@ async function runRg(options) {
51799
52015
  error: stderr.trim()
51800
52016
  };
51801
52017
  }
51802
- const matches = parseOutput(outputToProcess);
51803
- const filesSearched = new Set(matches.map((m) => m.file)).size;
52018
+ const matches = parseOutput(outputToProcess, options.outputMode === "files_with_matches");
52019
+ const limited = options.headLimit && options.headLimit > 0 ? matches.slice(0, options.headLimit) : matches;
52020
+ const filesSearched = new Set(limited.map((m) => m.file)).size;
51804
52021
  return {
51805
- matches,
51806
- totalMatches: matches.length,
52022
+ matches: limited,
52023
+ totalMatches: limited.length,
51807
52024
  filesSearched,
51808
- truncated
52025
+ truncated: truncated || (options.headLimit ? matches.length > options.headLimit : false)
51809
52026
  };
51810
52027
  } catch (e) {
51811
52028
  return {
@@ -51817,6 +52034,43 @@ async function runRg(options) {
51817
52034
  };
51818
52035
  }
51819
52036
  }
52037
+ async function runRgCount(options) {
52038
+ await rgSemaphore.acquire();
52039
+ try {
52040
+ return await runRgCountInternal(options);
52041
+ } finally {
52042
+ rgSemaphore.release();
52043
+ }
52044
+ }
52045
+ async function runRgCountInternal(options) {
52046
+ const cli = resolveGrepCli();
52047
+ const args = buildArgs({ ...options, context: 0 }, cli.backend);
52048
+ if (cli.backend === "rg") {
52049
+ args.push("--count", "--", options.pattern);
52050
+ } else {
52051
+ args.push("-c", "-e", options.pattern);
52052
+ }
52053
+ const paths = options.paths?.length ? options.paths : ["."];
52054
+ args.push(...paths);
52055
+ const timeout = Math.min(options.timeout ?? DEFAULT_TIMEOUT_MS3, DEFAULT_TIMEOUT_MS3);
52056
+ const proc = spawn11([cli.path, ...args], {
52057
+ stdout: "pipe",
52058
+ stderr: "pipe"
52059
+ });
52060
+ const timeoutPromise = new Promise((_, reject) => {
52061
+ const id = setTimeout(() => {
52062
+ proc.kill();
52063
+ reject(new Error(`Search timeout after ${timeout}ms`));
52064
+ }, timeout);
52065
+ proc.exited.then(() => clearTimeout(id));
52066
+ });
52067
+ try {
52068
+ const stdout = await Promise.race([new Response(proc.stdout).text(), timeoutPromise]);
52069
+ return parseCountOutput(stdout);
52070
+ } catch (e) {
52071
+ throw new Error(`Count search failed: ${e instanceof Error ? e.message : String(e)}`);
52072
+ }
52073
+ }
51820
52074
 
51821
52075
  // src/tools/grep/result-formatter.ts
51822
52076
  function formatGrepResult(result) {
@@ -51827,6 +52081,7 @@ function formatGrepResult(result) {
51827
52081
  return "No matches found";
51828
52082
  }
51829
52083
  const lines = [];
52084
+ const isFilesOnlyMode = result.matches.every((match) => match.line === 0 && match.text.trim() === "");
51830
52085
  lines.push(`Found ${result.totalMatches} match(es) in ${result.filesSearched} file(s)`);
51831
52086
  if (result.truncated) {
51832
52087
  lines.push("[Output truncated due to size limit]");
@@ -51840,34 +52095,68 @@ function formatGrepResult(result) {
51840
52095
  }
51841
52096
  for (const [file2, matches] of byFile) {
51842
52097
  lines.push(file2);
51843
- for (const match of matches) {
51844
- lines.push(` ${match.line}: ${match.text.trim()}`);
52098
+ if (!isFilesOnlyMode) {
52099
+ for (const match of matches) {
52100
+ const trimmedText = match.text.trim();
52101
+ if (match.line === 0 && trimmedText === "") {
52102
+ continue;
52103
+ }
52104
+ lines.push(` ${match.line}: ${trimmedText}`);
52105
+ }
51845
52106
  }
51846
52107
  lines.push("");
51847
52108
  }
51848
52109
  return lines.join(`
51849
52110
  `);
51850
52111
  }
52112
+ function formatCountResult(results) {
52113
+ if (results.length === 0) {
52114
+ return "No matches found";
52115
+ }
52116
+ const total = results.reduce((sum, r) => sum + r.count, 0);
52117
+ const lines = [`Found ${total} match(es) in ${results.length} file(s):`, ""];
52118
+ const sorted = [...results].sort((a, b) => b.count - a.count);
52119
+ for (const { file: file2, count } of sorted) {
52120
+ lines.push(` ${count.toString().padStart(6)}: ${file2}`);
52121
+ }
52122
+ return lines.join(`
52123
+ `);
52124
+ }
51851
52125
 
51852
52126
  // src/tools/grep/tools.ts
51853
52127
  function createGrepTools(ctx) {
51854
52128
  const grep = tool({
51855
- description: "Fast content search tool with safety limits (60s timeout, 10MB output). " + "Searches file contents using regular expressions. " + 'Supports full regex syntax (eg. "log.*Error", "function\\s+\\w+", etc.). ' + 'Filter files by pattern with the include parameter (eg. "*.js", "*.{ts,tsx}"). ' + "Returns file paths with matches sorted by modification time.",
52129
+ description: "Fast content search tool with safety limits (60s timeout, 256KB output). " + "Searches file contents using regular expressions. " + 'Supports full regex syntax (eg. "log.*Error", "function\\s+\\w+", etc.). ' + 'Filter files by pattern with the include parameter (eg. "*.js", "*.{ts,tsx}"). ' + 'Output modes: "content" shows matching lines, "files_with_matches" shows only file paths (default), "count" shows match counts per file.',
51856
52130
  args: {
51857
52131
  pattern: tool.schema.string().describe("The regex pattern to search for in file contents"),
51858
52132
  include: tool.schema.string().optional().describe('File pattern to include in the search (e.g. "*.js", "*.{ts,tsx}")'),
51859
- path: tool.schema.string().optional().describe("The directory to search in. Defaults to the current working directory.")
52133
+ path: tool.schema.string().optional().describe("The directory to search in. Defaults to the current working directory."),
52134
+ output_mode: tool.schema.enum(["content", "files_with_matches", "count"]).optional().describe('Output mode: "content" shows matching lines, "files_with_matches" shows only file paths (default), "count" shows match counts per file.'),
52135
+ head_limit: tool.schema.number().optional().describe("Limit output to first N entries. 0 or omitted means no limit.")
51860
52136
  },
51861
52137
  execute: async (args) => {
51862
52138
  try {
51863
52139
  const globs = args.include ? [args.include] : undefined;
51864
52140
  const searchPath = args.path ?? ctx.directory;
51865
52141
  const paths = [searchPath];
52142
+ const outputMode = args.output_mode ?? "files_with_matches";
52143
+ const headLimit = args.head_limit ?? 0;
52144
+ if (outputMode === "count") {
52145
+ const results = await runRgCount({
52146
+ pattern: args.pattern,
52147
+ paths,
52148
+ globs
52149
+ });
52150
+ const limited = headLimit > 0 ? results.slice(0, headLimit) : results;
52151
+ return formatCountResult(limited);
52152
+ }
51866
52153
  const result = await runRg({
51867
52154
  pattern: args.pattern,
51868
52155
  paths,
51869
52156
  globs,
51870
- context: 0
52157
+ context: 0,
52158
+ outputMode,
52159
+ headLimit
51871
52160
  });
51872
52161
  return formatGrepResult(result);
51873
52162
  } catch (e) {
@@ -51896,6 +52185,7 @@ import { stat } from "fs/promises";
51896
52185
  function buildRgArgs2(options) {
51897
52186
  const args = [
51898
52187
  ...RG_FILES_FLAGS,
52188
+ `--threads=${Math.min(options.threads ?? DEFAULT_RG_THREADS, DEFAULT_RG_THREADS)}`,
51899
52189
  `--max-depth=${Math.min(options.maxDepth ?? DEFAULT_MAX_DEPTH2, DEFAULT_MAX_DEPTH2)}`
51900
52190
  ];
51901
52191
  if (options.hidden !== false)
@@ -51944,6 +52234,14 @@ async function getFileMtime(filePath) {
51944
52234
  }
51945
52235
  }
51946
52236
  async function runRgFiles(options, resolvedCli) {
52237
+ await rgSemaphore.acquire();
52238
+ try {
52239
+ return await runRgFilesInternal(options, resolvedCli);
52240
+ } finally {
52241
+ rgSemaphore.release();
52242
+ }
52243
+ }
52244
+ async function runRgFilesInternal(options, resolvedCli) {
51947
52245
  const cli = resolvedCli ?? resolveGrepCli();
51948
52246
  const timeout = Math.min(options.timeout ?? DEFAULT_TIMEOUT_MS4, DEFAULT_TIMEOUT_MS4);
51949
52247
  const limit = Math.min(options.limit ?? DEFAULT_LIMIT, DEFAULT_LIMIT);
@@ -56838,30 +57136,281 @@ function stripRangeBoundaryEcho(lines, startLine, endLine, newLines) {
56838
57136
  return out;
56839
57137
  }
56840
57138
 
56841
- // src/tools/hashline-edit/edit-operations.ts
56842
- function applySetLine(lines, anchor, newText) {
56843
- validateLineRef(lines, anchor);
57139
+ // src/tools/hashline-edit/edit-deduplication.ts
57140
+ function normalizeEditPayload(payload) {
57141
+ return toNewLines(payload).join(`
57142
+ `);
57143
+ }
57144
+ function buildDedupeKey(edit) {
57145
+ switch (edit.type) {
57146
+ case "set_line":
57147
+ return `set_line|${edit.line}|${normalizeEditPayload(edit.text)}`;
57148
+ case "replace_lines":
57149
+ return `replace_lines|${edit.start_line}|${edit.end_line}|${normalizeEditPayload(edit.text)}`;
57150
+ case "insert_after":
57151
+ return `insert_after|${edit.line}|${normalizeEditPayload(edit.text)}`;
57152
+ case "insert_before":
57153
+ return `insert_before|${edit.line}|${normalizeEditPayload(edit.text)}`;
57154
+ case "insert_between":
57155
+ return `insert_between|${edit.after_line}|${edit.before_line}|${normalizeEditPayload(edit.text)}`;
57156
+ case "replace":
57157
+ return `replace|${edit.old_text}|${normalizeEditPayload(edit.new_text)}`;
57158
+ case "append":
57159
+ return `append|${normalizeEditPayload(edit.text)}`;
57160
+ case "prepend":
57161
+ return `prepend|${normalizeEditPayload(edit.text)}`;
57162
+ default:
57163
+ return JSON.stringify(edit);
57164
+ }
57165
+ }
57166
+ function dedupeEdits(edits) {
57167
+ const seen = new Set;
57168
+ const deduped = [];
57169
+ let deduplicatedEdits = 0;
57170
+ for (const edit of edits) {
57171
+ const key = buildDedupeKey(edit);
57172
+ if (seen.has(key)) {
57173
+ deduplicatedEdits += 1;
57174
+ continue;
57175
+ }
57176
+ seen.add(key);
57177
+ deduped.push(edit);
57178
+ }
57179
+ return { edits: deduped, deduplicatedEdits };
57180
+ }
57181
+
57182
+ // src/tools/hashline-edit/edit-ordering.ts
57183
+ function getEditLineNumber(edit) {
57184
+ switch (edit.type) {
57185
+ case "set_line":
57186
+ return parseLineRef(edit.line).line;
57187
+ case "replace_lines":
57188
+ return parseLineRef(edit.end_line).line;
57189
+ case "insert_after":
57190
+ return parseLineRef(edit.line).line;
57191
+ case "insert_before":
57192
+ return parseLineRef(edit.line).line;
57193
+ case "insert_between":
57194
+ return parseLineRef(edit.before_line).line;
57195
+ case "append":
57196
+ return Number.NEGATIVE_INFINITY;
57197
+ case "prepend":
57198
+ return Number.NEGATIVE_INFINITY;
57199
+ case "replace":
57200
+ return Number.NEGATIVE_INFINITY;
57201
+ default:
57202
+ return Number.POSITIVE_INFINITY;
57203
+ }
57204
+ }
57205
+ function collectLineRefs(edits) {
57206
+ return edits.flatMap((edit) => {
57207
+ switch (edit.type) {
57208
+ case "set_line":
57209
+ return [edit.line];
57210
+ case "replace_lines":
57211
+ return [edit.start_line, edit.end_line];
57212
+ case "insert_after":
57213
+ return [edit.line];
57214
+ case "insert_before":
57215
+ return [edit.line];
57216
+ case "insert_between":
57217
+ return [edit.after_line, edit.before_line];
57218
+ case "append":
57219
+ case "prepend":
57220
+ case "replace":
57221
+ return [];
57222
+ default:
57223
+ return [];
57224
+ }
57225
+ });
57226
+ }
57227
+
57228
+ // src/tools/hashline-edit/autocorrect-replacement-lines.ts
57229
+ function normalizeTokens(text) {
57230
+ return text.replace(/\s+/g, "");
57231
+ }
57232
+ function stripAllWhitespace(text) {
57233
+ return normalizeTokens(text);
57234
+ }
57235
+ function stripTrailingContinuationTokens(text) {
57236
+ return text.replace(/(?:&&|\|\||\?\?|\?|:|=|,|\+|-|\*|\/|\.|\()\s*$/u, "");
57237
+ }
57238
+ function stripMergeOperatorChars(text) {
57239
+ return text.replace(/[|&?]/g, "");
57240
+ }
57241
+ function leadingWhitespace2(text) {
57242
+ const match = text.match(/^\s*/);
57243
+ return match ? match[0] : "";
57244
+ }
57245
+ function restoreOldWrappedLines(originalLines, replacementLines) {
57246
+ if (originalLines.length === 0 || replacementLines.length < 2)
57247
+ return replacementLines;
57248
+ const canonicalToOriginal = new Map;
57249
+ for (const line of originalLines) {
57250
+ const canonical = stripAllWhitespace(line);
57251
+ const existing = canonicalToOriginal.get(canonical);
57252
+ if (existing) {
57253
+ existing.count += 1;
57254
+ } else {
57255
+ canonicalToOriginal.set(canonical, { line, count: 1 });
57256
+ }
57257
+ }
57258
+ const candidates = [];
57259
+ for (let start = 0;start < replacementLines.length; start += 1) {
57260
+ for (let len = 2;len <= 10 && start + len <= replacementLines.length; len += 1) {
57261
+ const canonicalSpan = stripAllWhitespace(replacementLines.slice(start, start + len).join(""));
57262
+ const original = canonicalToOriginal.get(canonicalSpan);
57263
+ if (original && original.count === 1 && canonicalSpan.length >= 6) {
57264
+ candidates.push({ start, len, replacement: original.line, canonical: canonicalSpan });
57265
+ }
57266
+ }
57267
+ }
57268
+ if (candidates.length === 0)
57269
+ return replacementLines;
57270
+ const canonicalCounts = new Map;
57271
+ for (const candidate of candidates) {
57272
+ canonicalCounts.set(candidate.canonical, (canonicalCounts.get(candidate.canonical) ?? 0) + 1);
57273
+ }
57274
+ const uniqueCandidates = candidates.filter((candidate) => (canonicalCounts.get(candidate.canonical) ?? 0) === 1);
57275
+ if (uniqueCandidates.length === 0)
57276
+ return replacementLines;
57277
+ uniqueCandidates.sort((a, b) => b.start - a.start);
57278
+ const correctedLines = [...replacementLines];
57279
+ for (const candidate of uniqueCandidates) {
57280
+ correctedLines.splice(candidate.start, candidate.len, candidate.replacement);
57281
+ }
57282
+ return correctedLines;
57283
+ }
57284
+ function maybeExpandSingleLineMerge(originalLines, replacementLines) {
57285
+ if (replacementLines.length !== 1 || originalLines.length <= 1) {
57286
+ return replacementLines;
57287
+ }
57288
+ const merged = replacementLines[0];
57289
+ const parts = originalLines.map((line) => line.trim()).filter((line) => line.length > 0);
57290
+ if (parts.length !== originalLines.length)
57291
+ return replacementLines;
57292
+ const indices = [];
57293
+ let offset = 0;
57294
+ let orderedMatch = true;
57295
+ for (const part of parts) {
57296
+ let idx = merged.indexOf(part, offset);
57297
+ let matchedLen = part.length;
57298
+ if (idx === -1) {
57299
+ const stripped = stripTrailingContinuationTokens(part);
57300
+ if (stripped !== part) {
57301
+ idx = merged.indexOf(stripped, offset);
57302
+ if (idx !== -1)
57303
+ matchedLen = stripped.length;
57304
+ }
57305
+ }
57306
+ if (idx === -1) {
57307
+ const segment = merged.slice(offset);
57308
+ const segmentStripped = stripMergeOperatorChars(segment);
57309
+ const partStripped = stripMergeOperatorChars(part);
57310
+ const fuzzyIdx = segmentStripped.indexOf(partStripped);
57311
+ if (fuzzyIdx !== -1) {
57312
+ let strippedPos = 0;
57313
+ let originalPos = 0;
57314
+ while (strippedPos < fuzzyIdx && originalPos < segment.length) {
57315
+ if (!/[|&?]/.test(segment[originalPos]))
57316
+ strippedPos += 1;
57317
+ originalPos += 1;
57318
+ }
57319
+ idx = offset + originalPos;
57320
+ matchedLen = part.length;
57321
+ }
57322
+ }
57323
+ if (idx === -1) {
57324
+ orderedMatch = false;
57325
+ break;
57326
+ }
57327
+ indices.push(idx);
57328
+ offset = idx + matchedLen;
57329
+ }
57330
+ const expanded = [];
57331
+ if (orderedMatch) {
57332
+ for (let i2 = 0;i2 < indices.length; i2 += 1) {
57333
+ const start = indices[i2];
57334
+ const end = i2 + 1 < indices.length ? indices[i2 + 1] : merged.length;
57335
+ const candidate = merged.slice(start, end).trim();
57336
+ if (candidate.length === 0) {
57337
+ orderedMatch = false;
57338
+ break;
57339
+ }
57340
+ expanded.push(candidate);
57341
+ }
57342
+ }
57343
+ if (orderedMatch && expanded.length === originalLines.length) {
57344
+ return expanded;
57345
+ }
57346
+ const semicolonSplit = merged.split(/;\s+/).map((line, idx, arr) => {
57347
+ if (idx < arr.length - 1 && !line.endsWith(";")) {
57348
+ return `${line};`;
57349
+ }
57350
+ return line;
57351
+ }).map((line) => line.trim()).filter((line) => line.length > 0);
57352
+ if (semicolonSplit.length === originalLines.length) {
57353
+ return semicolonSplit;
57354
+ }
57355
+ return replacementLines;
57356
+ }
57357
+ function restoreIndentForPairedReplacement(originalLines, replacementLines) {
57358
+ if (originalLines.length !== replacementLines.length) {
57359
+ return replacementLines;
57360
+ }
57361
+ return replacementLines.map((line, idx) => {
57362
+ if (line.length === 0)
57363
+ return line;
57364
+ if (leadingWhitespace2(line).length > 0)
57365
+ return line;
57366
+ const indent = leadingWhitespace2(originalLines[idx]);
57367
+ if (indent.length === 0)
57368
+ return line;
57369
+ return `${indent}${line}`;
57370
+ });
57371
+ }
57372
+ function autocorrectReplacementLines(originalLines, replacementLines) {
57373
+ let next = replacementLines;
57374
+ next = maybeExpandSingleLineMerge(originalLines, next);
57375
+ next = restoreOldWrappedLines(originalLines, next);
57376
+ next = restoreIndentForPairedReplacement(originalLines, next);
57377
+ return next;
57378
+ }
57379
+
57380
+ // src/tools/hashline-edit/edit-operation-primitives.ts
57381
+ function shouldValidate(options) {
57382
+ return options?.skipValidation !== true;
57383
+ }
57384
+ function applySetLine(lines, anchor, newText, options) {
57385
+ if (shouldValidate(options))
57386
+ validateLineRef(lines, anchor);
56844
57387
  const { line } = parseLineRef(anchor);
56845
57388
  const result = [...lines];
56846
- const replacement = toNewLines(newText).map((entry, idx) => {
57389
+ const originalLine = lines[line - 1] ?? "";
57390
+ const corrected = autocorrectReplacementLines([originalLine], toNewLines(newText));
57391
+ const replacement = corrected.map((entry, idx) => {
56847
57392
  if (idx !== 0)
56848
57393
  return entry;
56849
- return restoreLeadingIndent(lines[line - 1], entry);
57394
+ return restoreLeadingIndent(originalLine, entry);
56850
57395
  });
56851
57396
  result.splice(line - 1, 1, ...replacement);
56852
57397
  return result;
56853
57398
  }
56854
- function applyReplaceLines(lines, startAnchor, endAnchor, newText) {
56855
- validateLineRef(lines, startAnchor);
56856
- validateLineRef(lines, endAnchor);
57399
+ function applyReplaceLines(lines, startAnchor, endAnchor, newText, options) {
57400
+ if (shouldValidate(options)) {
57401
+ validateLineRef(lines, startAnchor);
57402
+ validateLineRef(lines, endAnchor);
57403
+ }
56857
57404
  const { line: startLine } = parseLineRef(startAnchor);
56858
57405
  const { line: endLine } = parseLineRef(endAnchor);
56859
57406
  if (startLine > endLine) {
56860
57407
  throw new Error(`Invalid range: start line ${startLine} cannot be greater than end line ${endLine}`);
56861
57408
  }
56862
57409
  const result = [...lines];
57410
+ const originalRange = lines.slice(startLine - 1, endLine);
56863
57411
  const stripped = stripRangeBoundaryEcho(lines, startLine, endLine, toNewLines(newText));
56864
- const restored = stripped.map((entry, idx) => {
57412
+ const corrected = autocorrectReplacementLines(originalRange, stripped);
57413
+ const restored = corrected.map((entry, idx) => {
56865
57414
  if (idx !== 0)
56866
57415
  return entry;
56867
57416
  return restoreLeadingIndent(lines[startLine - 1], entry);
@@ -56869,8 +57418,9 @@ function applyReplaceLines(lines, startAnchor, endAnchor, newText) {
56869
57418
  result.splice(startLine - 1, endLine - startLine + 1, ...restored);
56870
57419
  return result;
56871
57420
  }
56872
- function applyInsertAfter(lines, anchor, text) {
56873
- validateLineRef(lines, anchor);
57421
+ function applyInsertAfter(lines, anchor, text, options) {
57422
+ if (shouldValidate(options))
57423
+ validateLineRef(lines, anchor);
56874
57424
  const { line } = parseLineRef(anchor);
56875
57425
  const result = [...lines];
56876
57426
  const newLines = stripInsertAnchorEcho(lines[line - 1], toNewLines(text));
@@ -56880,8 +57430,9 @@ function applyInsertAfter(lines, anchor, text) {
56880
57430
  result.splice(line, 0, ...newLines);
56881
57431
  return result;
56882
57432
  }
56883
- function applyInsertBefore(lines, anchor, text) {
56884
- validateLineRef(lines, anchor);
57433
+ function applyInsertBefore(lines, anchor, text, options) {
57434
+ if (shouldValidate(options))
57435
+ validateLineRef(lines, anchor);
56885
57436
  const { line } = parseLineRef(anchor);
56886
57437
  const result = [...lines];
56887
57438
  const newLines = stripInsertBeforeEcho(lines[line - 1], toNewLines(text));
@@ -56891,9 +57442,11 @@ function applyInsertBefore(lines, anchor, text) {
56891
57442
  result.splice(line - 1, 0, ...newLines);
56892
57443
  return result;
56893
57444
  }
56894
- function applyInsertBetween(lines, afterAnchor, beforeAnchor, text) {
56895
- validateLineRef(lines, afterAnchor);
56896
- validateLineRef(lines, beforeAnchor);
57445
+ function applyInsertBetween(lines, afterAnchor, beforeAnchor, text, options) {
57446
+ if (shouldValidate(options)) {
57447
+ validateLineRef(lines, afterAnchor);
57448
+ validateLineRef(lines, beforeAnchor);
57449
+ }
56897
57450
  const { line: afterLine } = parseLineRef(afterAnchor);
56898
57451
  const { line: beforeLine } = parseLineRef(beforeAnchor);
56899
57452
  if (beforeLine <= afterLine) {
@@ -56907,58 +57460,36 @@ function applyInsertBetween(lines, afterAnchor, beforeAnchor, text) {
56907
57460
  result.splice(beforeLine - 1, 0, ...newLines);
56908
57461
  return result;
56909
57462
  }
56910
- function getEditLineNumber(edit) {
56911
- switch (edit.type) {
56912
- case "set_line":
56913
- return parseLineRef(edit.line).line;
56914
- case "replace_lines":
56915
- return parseLineRef(edit.end_line).line;
56916
- case "insert_after":
56917
- return parseLineRef(edit.line).line;
56918
- case "insert_before":
56919
- return parseLineRef(edit.line).line;
56920
- case "insert_between":
56921
- return parseLineRef(edit.before_line).line;
56922
- case "replace":
56923
- return Number.NEGATIVE_INFINITY;
56924
- default:
56925
- return Number.POSITIVE_INFINITY;
57463
+ function applyAppend(lines, text) {
57464
+ const normalized = toNewLines(text);
57465
+ if (normalized.length === 0) {
57466
+ throw new Error("append requires non-empty text");
56926
57467
  }
57468
+ if (lines.length === 1 && lines[0] === "") {
57469
+ return [...normalized];
57470
+ }
57471
+ return [...lines, ...normalized];
56927
57472
  }
56928
- function normalizeEditPayload(payload) {
56929
- return toNewLines(payload).join(`
56930
- `);
57473
+ function applyPrepend(lines, text) {
57474
+ const normalized = toNewLines(text);
57475
+ if (normalized.length === 0) {
57476
+ throw new Error("prepend requires non-empty text");
57477
+ }
57478
+ if (lines.length === 1 && lines[0] === "") {
57479
+ return [...normalized];
57480
+ }
57481
+ return [...normalized, ...lines];
56931
57482
  }
56932
- function dedupeEdits(edits) {
56933
- const seen = new Set;
56934
- const deduped = [];
56935
- let deduplicatedEdits = 0;
56936
- for (const edit of edits) {
56937
- const key = (() => {
56938
- switch (edit.type) {
56939
- case "set_line":
56940
- return `set_line|${edit.line}|${normalizeEditPayload(edit.text)}`;
56941
- case "replace_lines":
56942
- return `replace_lines|${edit.start_line}|${edit.end_line}|${normalizeEditPayload(edit.text)}`;
56943
- case "insert_after":
56944
- return `insert_after|${edit.line}|${normalizeEditPayload(edit.text)}`;
56945
- case "insert_before":
56946
- return `insert_before|${edit.line}|${normalizeEditPayload(edit.text)}`;
56947
- case "insert_between":
56948
- return `insert_between|${edit.after_line}|${edit.before_line}|${normalizeEditPayload(edit.text)}`;
56949
- case "replace":
56950
- return `replace|${edit.old_text}|${normalizeEditPayload(edit.new_text)}`;
56951
- }
56952
- })();
56953
- if (seen.has(key)) {
56954
- deduplicatedEdits += 1;
56955
- continue;
56956
- }
56957
- seen.add(key);
56958
- deduped.push(edit);
57483
+ function applyReplace(content, oldText, newText) {
57484
+ if (!content.includes(oldText)) {
57485
+ throw new Error(`Text not found: "${oldText}"`);
56959
57486
  }
56960
- return { edits: deduped, deduplicatedEdits };
57487
+ const replacement = Array.isArray(newText) ? newText.join(`
57488
+ `) : newText;
57489
+ return content.replaceAll(oldText, replacement);
56961
57490
  }
57491
+
57492
+ // src/tools/hashline-edit/edit-operations.ts
56962
57493
  function applyHashlineEditsWithReport(content, edits) {
56963
57494
  if (edits.length === 0) {
56964
57495
  return {
@@ -56971,39 +57502,22 @@ function applyHashlineEditsWithReport(content, edits) {
56971
57502
  const sortedEdits = [...dedupeResult.edits].sort((a, b) => getEditLineNumber(b) - getEditLineNumber(a));
56972
57503
  let noopEdits = 0;
56973
57504
  let result = content;
56974
- let lines = result.split(`
57505
+ let lines = result.length === 0 ? [] : result.split(`
56975
57506
  `);
56976
- const refs = sortedEdits.flatMap((edit) => {
56977
- switch (edit.type) {
56978
- case "set_line":
56979
- return [edit.line];
56980
- case "replace_lines":
56981
- return [edit.start_line, edit.end_line];
56982
- case "insert_after":
56983
- return [edit.line];
56984
- case "insert_before":
56985
- return [edit.line];
56986
- case "insert_between":
56987
- return [edit.after_line, edit.before_line];
56988
- case "replace":
56989
- return [];
56990
- default:
56991
- return [];
56992
- }
56993
- });
57507
+ const refs = collectLineRefs(sortedEdits);
56994
57508
  validateLineRefs(lines, refs);
56995
57509
  for (const edit of sortedEdits) {
56996
57510
  switch (edit.type) {
56997
57511
  case "set_line": {
56998
- lines = applySetLine(lines, edit.line, edit.text);
57512
+ lines = applySetLine(lines, edit.line, edit.text, { skipValidation: true });
56999
57513
  break;
57000
57514
  }
57001
57515
  case "replace_lines": {
57002
- lines = applyReplaceLines(lines, edit.start_line, edit.end_line, edit.text);
57516
+ lines = applyReplaceLines(lines, edit.start_line, edit.end_line, edit.text, { skipValidation: true });
57003
57517
  break;
57004
57518
  }
57005
57519
  case "insert_after": {
57006
- const next = applyInsertAfter(lines, edit.line, edit.text);
57520
+ const next = applyInsertAfter(lines, edit.line, edit.text, { skipValidation: true });
57007
57521
  if (next.join(`
57008
57522
  `) === lines.join(`
57009
57523
  `)) {
@@ -57014,7 +57528,7 @@ function applyHashlineEditsWithReport(content, edits) {
57014
57528
  break;
57015
57529
  }
57016
57530
  case "insert_before": {
57017
- const next = applyInsertBefore(lines, edit.line, edit.text);
57531
+ const next = applyInsertBefore(lines, edit.line, edit.text, { skipValidation: true });
57018
57532
  if (next.join(`
57019
57533
  `) === lines.join(`
57020
57534
  `)) {
@@ -57025,7 +57539,29 @@ function applyHashlineEditsWithReport(content, edits) {
57025
57539
  break;
57026
57540
  }
57027
57541
  case "insert_between": {
57028
- const next = applyInsertBetween(lines, edit.after_line, edit.before_line, edit.text);
57542
+ const next = applyInsertBetween(lines, edit.after_line, edit.before_line, edit.text, { skipValidation: true });
57543
+ if (next.join(`
57544
+ `) === lines.join(`
57545
+ `)) {
57546
+ noopEdits += 1;
57547
+ break;
57548
+ }
57549
+ lines = next;
57550
+ break;
57551
+ }
57552
+ case "append": {
57553
+ const next = applyAppend(lines, edit.text);
57554
+ if (next.join(`
57555
+ `) === lines.join(`
57556
+ `)) {
57557
+ noopEdits += 1;
57558
+ break;
57559
+ }
57560
+ lines = next;
57561
+ break;
57562
+ }
57563
+ case "prepend": {
57564
+ const next = applyPrepend(lines, edit.text);
57029
57565
  if (next.join(`
57030
57566
  `) === lines.join(`
57031
57567
  `)) {
@@ -57038,12 +57574,7 @@ function applyHashlineEditsWithReport(content, edits) {
57038
57574
  case "replace": {
57039
57575
  result = lines.join(`
57040
57576
  `);
57041
- if (!result.includes(edit.old_text)) {
57042
- throw new Error(`Text not found: "${edit.old_text}"`);
57043
- }
57044
- const replacement = Array.isArray(edit.new_text) ? edit.new_text.join(`
57045
- `) : edit.new_text;
57046
- const replaced = result.replaceAll(edit.old_text, replacement);
57577
+ const replaced = applyReplace(result, edit.old_text, edit.new_text);
57047
57578
  if (replaced === result) {
57048
57579
  noopEdits += 1;
57049
57580
  break;
@@ -57062,53 +57593,57 @@ function applyHashlineEditsWithReport(content, edits) {
57062
57593
  deduplicatedEdits: dedupeResult.deduplicatedEdits
57063
57594
  };
57064
57595
  }
57065
- // src/tools/hashline-edit/tool-description.ts
57066
- var HASHLINE_EDIT_DESCRIPTION = `Edit files using LINE#ID format for precise, safe modifications.
57067
-
57068
- WORKFLOW:
57069
- 1. Read the file and copy exact LINE#ID anchors.
57070
- 2. Submit one edit call with all related operations for that file.
57071
- 3. If more edits are needed after success, use the latest anchors from read/edit output.
57072
- 4. Use anchors as "LINE#ID" only (never include trailing ":content").
57073
-
57074
- VALIDATION:
57075
- - Payload shape: { "filePath": string, "edits": [...], "delete"?: boolean, "rename"?: string }
57076
- - Each edit must be one of: set_line, replace_lines, insert_after, insert_before, insert_between, replace
57077
- - text/new_text must contain plain replacement text only (no LINE#ID prefixes, no diff + markers)
57078
-
57079
- LINE#ID FORMAT (CRITICAL - READ CAREFULLY):
57080
- Each line reference must be in "LINE#ID" format where:
57081
- - LINE: 1-based line number
57082
- - ID: Two CID letters from the set ZPMQVRWSNKTXJBYH
57083
-
57084
- OPERATION TYPES:
57085
- 1. set_line
57086
- 2. replace_lines
57087
- 3. insert_after
57088
- 4. insert_before
57089
- 5. insert_between
57090
- 6. replace
57091
-
57092
- FILE MODES:
57093
- - delete=true deletes file and requires edits=[] with no rename
57094
- - rename moves final content to a new path and removes old path
57095
-
57096
- CONTENT FORMAT:
57097
- - text/new_text can be a string (single line) or string[] (multi-line, preferred).
57098
- - If you pass a multi-line string, it is split by real newline characters.
57099
- - Literal "\\n" is preserved as text.`;
57100
-
57101
- // src/tools/hashline-edit/tools.ts
57102
- function resolveToolCallID2(ctx) {
57103
- if (typeof ctx.callID === "string" && ctx.callID.trim() !== "")
57104
- return ctx.callID;
57105
- if (typeof ctx.callId === "string" && ctx.callId.trim() !== "")
57106
- return ctx.callId;
57107
- if (typeof ctx.call_id === "string" && ctx.call_id.trim() !== "")
57108
- return ctx.call_id;
57109
- return;
57596
+ // src/tools/hashline-edit/file-text-canonicalization.ts
57597
+ function detectLineEnding(content) {
57598
+ const crlfIndex = content.indexOf(`\r
57599
+ `);
57600
+ const lfIndex = content.indexOf(`
57601
+ `);
57602
+ if (lfIndex === -1)
57603
+ return `
57604
+ `;
57605
+ if (crlfIndex === -1)
57606
+ return `
57607
+ `;
57608
+ return crlfIndex < lfIndex ? `\r
57609
+ ` : `
57610
+ `;
57611
+ }
57612
+ function stripBom(content) {
57613
+ if (!content.startsWith("\uFEFF")) {
57614
+ return { content, hadBom: false };
57615
+ }
57616
+ return { content: content.slice(1), hadBom: true };
57617
+ }
57618
+ function normalizeToLf(content) {
57619
+ return content.replace(/\r\n/g, `
57620
+ `).replace(/\r/g, `
57621
+ `);
57622
+ }
57623
+ function restoreLineEndings(content, lineEnding) {
57624
+ if (lineEnding === `
57625
+ `)
57626
+ return content;
57627
+ return content.replace(/\n/g, `\r
57628
+ `);
57629
+ }
57630
+ function canonicalizeFileText(content) {
57631
+ const stripped = stripBom(content);
57632
+ return {
57633
+ content: normalizeToLf(stripped.content),
57634
+ hadBom: stripped.hadBom,
57635
+ lineEnding: detectLineEnding(stripped.content)
57636
+ };
57110
57637
  }
57111
- function generateDiff(oldContent, newContent, filePath) {
57638
+ function restoreFileText(content, envelope) {
57639
+ const withLineEnding = restoreLineEndings(content, envelope.lineEnding);
57640
+ if (!envelope.hadBom)
57641
+ return withLineEnding;
57642
+ return `\uFEFF${withLineEnding}`;
57643
+ }
57644
+
57645
+ // src/tools/hashline-edit/hashline-edit-diff.ts
57646
+ function generateHashlineDiff(oldContent, newContent, filePath) {
57112
57647
  const oldLines = oldContent.split(`
57113
57648
  `);
57114
57649
  const newLines = newContent.split(`
@@ -57117,7 +57652,7 @@ function generateDiff(oldContent, newContent, filePath) {
57117
57652
  +++ ${filePath}
57118
57653
  `;
57119
57654
  const maxLines = Math.max(oldLines.length, newLines.length);
57120
- for (let i2 = 0;i2 < maxLines; i2++) {
57655
+ for (let i2 = 0;i2 < maxLines; i2 += 1) {
57121
57656
  const oldLine = oldLines[i2] ?? "";
57122
57657
  const newLine = newLines[i2] ?? "";
57123
57658
  const lineNum = i2 + 1;
@@ -57125,10 +57660,14 @@ function generateDiff(oldContent, newContent, filePath) {
57125
57660
  if (i2 >= oldLines.length) {
57126
57661
  diff += `+ ${lineNum}#${hash2}:${newLine}
57127
57662
  `;
57128
- } else if (i2 >= newLines.length) {
57663
+ continue;
57664
+ }
57665
+ if (i2 >= newLines.length) {
57129
57666
  diff += `- ${lineNum}# :${oldLine}
57130
57667
  `;
57131
- } else if (oldLine !== newLine) {
57668
+ continue;
57669
+ }
57670
+ if (oldLine !== newLine) {
57132
57671
  diff += `- ${lineNum}# :${oldLine}
57133
57672
  `;
57134
57673
  diff += `+ ${lineNum}#${hash2}:${newLine}
@@ -57137,6 +57676,182 @@ function generateDiff(oldContent, newContent, filePath) {
57137
57676
  }
57138
57677
  return diff;
57139
57678
  }
57679
+
57680
+ // src/tools/hashline-edit/hashline-edit-executor.ts
57681
+ function resolveToolCallID2(ctx) {
57682
+ if (typeof ctx.callID === "string" && ctx.callID.trim() !== "")
57683
+ return ctx.callID;
57684
+ if (typeof ctx.callId === "string" && ctx.callId.trim() !== "")
57685
+ return ctx.callId;
57686
+ if (typeof ctx.call_id === "string" && ctx.call_id.trim() !== "")
57687
+ return ctx.call_id;
57688
+ return;
57689
+ }
57690
+ function canCreateFromMissingFile(edits) {
57691
+ if (edits.length === 0)
57692
+ return false;
57693
+ return edits.every((edit) => edit.type === "append" || edit.type === "prepend");
57694
+ }
57695
+ function buildSuccessMeta(effectivePath, beforeContent, afterContent, noopEdits, deduplicatedEdits) {
57696
+ const unifiedDiff = generateUnifiedDiff(beforeContent, afterContent, effectivePath);
57697
+ const { additions, deletions } = countLineDiffs(beforeContent, afterContent);
57698
+ return {
57699
+ title: effectivePath,
57700
+ metadata: {
57701
+ filePath: effectivePath,
57702
+ path: effectivePath,
57703
+ file: effectivePath,
57704
+ diff: unifiedDiff,
57705
+ noopEdits,
57706
+ deduplicatedEdits,
57707
+ filediff: {
57708
+ file: effectivePath,
57709
+ path: effectivePath,
57710
+ filePath: effectivePath,
57711
+ before: beforeContent,
57712
+ after: afterContent,
57713
+ additions,
57714
+ deletions
57715
+ }
57716
+ }
57717
+ };
57718
+ }
57719
+ async function executeHashlineEditTool(args, context) {
57720
+ try {
57721
+ const metadataContext = context;
57722
+ const filePath = args.filePath;
57723
+ const { edits, delete: deleteMode, rename } = args;
57724
+ if (deleteMode && rename) {
57725
+ return "Error: delete and rename cannot be used together";
57726
+ }
57727
+ if (!deleteMode && (!edits || !Array.isArray(edits) || edits.length === 0)) {
57728
+ return "Error: edits parameter must be a non-empty array";
57729
+ }
57730
+ if (deleteMode && edits.length > 0) {
57731
+ return "Error: delete mode requires edits to be an empty array";
57732
+ }
57733
+ const file2 = Bun.file(filePath);
57734
+ const exists = await file2.exists();
57735
+ if (!exists && !deleteMode && !canCreateFromMissingFile(edits)) {
57736
+ return `Error: File not found: ${filePath}`;
57737
+ }
57738
+ if (deleteMode) {
57739
+ if (!exists)
57740
+ return `Error: File not found: ${filePath}`;
57741
+ await Bun.file(filePath).delete();
57742
+ return `Successfully deleted ${filePath}`;
57743
+ }
57744
+ const rawOldContent = exists ? Buffer.from(await file2.arrayBuffer()).toString("utf8") : "";
57745
+ const oldEnvelope = canonicalizeFileText(rawOldContent);
57746
+ const applyResult = applyHashlineEditsWithReport(oldEnvelope.content, edits);
57747
+ const canonicalNewContent = applyResult.content;
57748
+ const writeContent = restoreFileText(canonicalNewContent, oldEnvelope);
57749
+ await Bun.write(filePath, writeContent);
57750
+ if (rename && rename !== filePath) {
57751
+ await Bun.write(rename, writeContent);
57752
+ await Bun.file(filePath).delete();
57753
+ }
57754
+ const effectivePath = rename && rename !== filePath ? rename : filePath;
57755
+ const diff = generateHashlineDiff(oldEnvelope.content, canonicalNewContent, effectivePath);
57756
+ const newHashlined = toHashlineContent(canonicalNewContent);
57757
+ const meta = buildSuccessMeta(effectivePath, oldEnvelope.content, canonicalNewContent, applyResult.noopEdits, applyResult.deduplicatedEdits);
57758
+ if (typeof metadataContext.metadata === "function") {
57759
+ metadataContext.metadata(meta);
57760
+ }
57761
+ const callID = resolveToolCallID2(metadataContext);
57762
+ if (callID) {
57763
+ storeToolMetadata(context.sessionID, callID, meta);
57764
+ }
57765
+ return `Successfully applied ${edits.length} edit(s) to ${effectivePath}
57766
+ No-op edits: ${applyResult.noopEdits}, deduplicated edits: ${applyResult.deduplicatedEdits}
57767
+
57768
+ ${diff}
57769
+
57770
+ Updated file (LINE#ID:content):
57771
+ ${newHashlined}`;
57772
+ } catch (error45) {
57773
+ const message = error45 instanceof Error ? error45.message : String(error45);
57774
+ if (message.toLowerCase().includes("hash")) {
57775
+ return `Error: hash mismatch - ${message}
57776
+ Tip: reuse LINE#ID entries from the latest read/edit output, or batch related edits in one call.`;
57777
+ }
57778
+ return `Error: ${message}`;
57779
+ }
57780
+ }
57781
+
57782
+ // src/tools/hashline-edit/tool-description.ts
57783
+ var HASHLINE_EDIT_DESCRIPTION = `Edit files using LINE#ID format for precise, safe modifications.
57784
+
57785
+ WORKFLOW:
57786
+ 1. Read target file/range and copy exact LINE#ID tags.
57787
+ 2. Pick the smallest operation per logical mutation site.
57788
+ 3. Submit one edit call per file with all related operations.
57789
+ 4. If same file needs another call, re-read first.
57790
+ 5. Use anchors as "LINE#ID" only (never include trailing ":content").
57791
+
57792
+ VALIDATION:
57793
+ Payload shape: { "filePath": string, "edits": [...], "delete"?: boolean, "rename"?: string }
57794
+ Each edit must be one of: set_line, replace_lines, insert_after, insert_before, insert_between, replace, append, prepend
57795
+ text/new_text must contain plain replacement text only (no LINE#ID prefixes, no diff + markers)
57796
+ CRITICAL: all operations validate against the same pre-edit file snapshot and apply bottom-up. Refs/tags are interpreted against the last-read version of the file.
57797
+
57798
+ LINE#ID FORMAT (CRITICAL):
57799
+ Each line reference must be in "LINE#ID" format where:
57800
+ LINE: 1-based line number
57801
+ ID: Two CID letters from the set ZPMQVRWSNKTXJBYH
57802
+
57803
+ FILE MODES:
57804
+ delete=true deletes file and requires edits=[] with no rename
57805
+ rename moves final content to a new path and removes old path
57806
+
57807
+ CONTENT FORMAT:
57808
+ text/new_text can be a string (single line) or string[] (multi-line, preferred).
57809
+ If you pass a multi-line string, it is split by real newline characters.
57810
+ Literal "\\n" is preserved as text.
57811
+
57812
+ FILE CREATION:
57813
+ append: adds content at EOF. If file does not exist, creates it.
57814
+ prepend: adds content at BOF. If file does not exist, creates it.
57815
+ CRITICAL: append/prepend are the only operations that work without an existing file.
57816
+
57817
+ OPERATION CHOICE:
57818
+ One line wrong -> set_line
57819
+ Adjacent block rewrite or swap/move -> replace_lines (prefer one range op over many single-line ops)
57820
+ Both boundaries known -> insert_between (ALWAYS prefer over insert_after/insert_before)
57821
+ One boundary known -> insert_after or insert_before
57822
+ New file or EOF/BOF addition -> append or prepend
57823
+ No LINE#ID available -> replace (last resort)
57824
+
57825
+ RULES (CRITICAL):
57826
+ 1. Minimize scope: one logical mutation site per operation.
57827
+ 2. Preserve formatting: keep indentation, punctuation, line breaks, trailing commas, brace style.
57828
+ 3. Prefer insertion over neighbor rewrites: anchor to structural boundaries (}, ], },), not interior property lines.
57829
+ 4. No no-ops: replacement content must differ from current content.
57830
+ 5. Touch only requested code: avoid incidental edits.
57831
+ 6. Use exact current tokens: NEVER rewrite approximately.
57832
+ 7. For swaps/moves: prefer one range operation over multiple single-line operations.
57833
+ 8. Output tool calls only; no prose or commentary between them.
57834
+
57835
+ TAG CHOICE (ALWAYS):
57836
+ - Copy tags exactly from read output or >>> mismatch output.
57837
+ - NEVER guess tags.
57838
+ - Prefer insert_between over insert_after/insert_before when both boundaries are known.
57839
+ - Anchor to structural lines (function/class/brace), NEVER blank lines.
57840
+ - Anti-pattern warning: blank/whitespace anchors are fragile.
57841
+ - Re-read after each successful edit call before issuing another on the same file.
57842
+
57843
+ AUTOCORRECT (built-in - you do NOT need to handle these):
57844
+ Merged lines are auto-expanded back to original line count.
57845
+ Indentation is auto-restored from original lines.
57846
+ BOM and CRLF line endings are preserved automatically.
57847
+ Hashline prefixes and diff markers in text are auto-stripped.
57848
+
57849
+ RECOVERY (when >>> mismatch error appears):
57850
+ Copy the updated LINE#ID tags shown in the error output directly.
57851
+ Re-read only if the needed tags are missing from the error snippet.
57852
+ ALWAYS batch all edits for one file in a single call.`;
57853
+
57854
+ // src/tools/hashline-edit/tools.ts
57140
57855
  function createHashlineEditTool() {
57141
57856
  return tool({
57142
57857
  description: HASHLINE_EDIT_DESCRIPTION,
@@ -57176,88 +57891,18 @@ function createHashlineEditTool() {
57176
57891
  type: tool.schema.literal("replace"),
57177
57892
  old_text: tool.schema.string().describe("Text to find"),
57178
57893
  new_text: tool.schema.union([tool.schema.string(), tool.schema.array(tool.schema.string())]).describe("Replacement text (string or string[] for multiline)")
57894
+ }),
57895
+ tool.schema.object({
57896
+ type: tool.schema.literal("append"),
57897
+ text: tool.schema.union([tool.schema.string(), tool.schema.array(tool.schema.string())]).describe("Content to append at EOF; also creates missing file")
57898
+ }),
57899
+ tool.schema.object({
57900
+ type: tool.schema.literal("prepend"),
57901
+ text: tool.schema.union([tool.schema.string(), tool.schema.array(tool.schema.string())]).describe("Content to prepend at BOF; also creates missing file")
57179
57902
  })
57180
57903
  ])).describe("Array of edit operations to apply (empty when delete=true)")
57181
57904
  },
57182
- execute: async (args, context) => {
57183
- try {
57184
- const metadataContext = context;
57185
- const filePath = args.filePath;
57186
- const { edits, delete: deleteMode, rename } = args;
57187
- if (deleteMode && rename) {
57188
- return "Error: delete and rename cannot be used together";
57189
- }
57190
- if (!deleteMode && (!edits || !Array.isArray(edits) || edits.length === 0)) {
57191
- return "Error: edits parameter must be a non-empty array";
57192
- }
57193
- if (deleteMode && edits.length > 0) {
57194
- return "Error: delete mode requires edits to be an empty array";
57195
- }
57196
- const file2 = Bun.file(filePath);
57197
- const exists = await file2.exists();
57198
- if (!exists) {
57199
- return `Error: File not found: ${filePath}`;
57200
- }
57201
- if (deleteMode) {
57202
- await Bun.file(filePath).delete();
57203
- return `Successfully deleted ${filePath}`;
57204
- }
57205
- const oldContent = await file2.text();
57206
- const applyResult = applyHashlineEditsWithReport(oldContent, edits);
57207
- const newContent = applyResult.content;
57208
- await Bun.write(filePath, newContent);
57209
- if (rename && rename !== filePath) {
57210
- await Bun.write(rename, newContent);
57211
- await Bun.file(filePath).delete();
57212
- }
57213
- const effectivePath = rename && rename !== filePath ? rename : filePath;
57214
- const diff = generateDiff(oldContent, newContent, effectivePath);
57215
- const newHashlined = toHashlineContent(newContent);
57216
- const unifiedDiff = generateUnifiedDiff(oldContent, newContent, effectivePath);
57217
- const { additions, deletions } = countLineDiffs(oldContent, newContent);
57218
- const meta = {
57219
- title: effectivePath,
57220
- metadata: {
57221
- filePath: effectivePath,
57222
- path: effectivePath,
57223
- file: effectivePath,
57224
- diff: unifiedDiff,
57225
- noopEdits: applyResult.noopEdits,
57226
- deduplicatedEdits: applyResult.deduplicatedEdits,
57227
- filediff: {
57228
- file: effectivePath,
57229
- path: effectivePath,
57230
- filePath: effectivePath,
57231
- before: oldContent,
57232
- after: newContent,
57233
- additions,
57234
- deletions
57235
- }
57236
- }
57237
- };
57238
- if (typeof metadataContext.metadata === "function") {
57239
- metadataContext.metadata(meta);
57240
- }
57241
- const callID = resolveToolCallID2(metadataContext);
57242
- if (callID) {
57243
- storeToolMetadata(context.sessionID, callID, meta);
57244
- }
57245
- return `Successfully applied ${edits.length} edit(s) to ${effectivePath}
57246
- No-op edits: ${applyResult.noopEdits}, deduplicated edits: ${applyResult.deduplicatedEdits}
57247
-
57248
- ${diff}
57249
-
57250
- Updated file (LINE#ID:content):
57251
- ${newHashlined}`;
57252
- } catch (error45) {
57253
- const message = error45 instanceof Error ? error45.message : String(error45);
57254
- if (message.toLowerCase().includes("hash")) {
57255
- return `Error: hash mismatch - ${message}
57256
- Tip: reuse LINE#ID entries from the latest read/edit output, or batch related edits in one call.`;
57257
- }
57258
- return `Error: ${message}`;
57259
- }
57260
- }
57905
+ execute: async (args, context) => executeHashlineEditTool(args, context)
57261
57906
  });
57262
57907
  }
57263
57908
  // src/tools/index.ts
@@ -58213,6 +58858,79 @@ function findNearestMessageExcludingCompaction(messageDir) {
58213
58858
  return null;
58214
58859
  }
58215
58860
 
58861
+ // src/features/background-agent/session-idle-event-handler.ts
58862
+ function getString(obj, key) {
58863
+ const value = obj[key];
58864
+ return typeof value === "string" ? value : undefined;
58865
+ }
58866
+ function handleSessionIdleBackgroundEvent(args) {
58867
+ const {
58868
+ properties,
58869
+ findBySession,
58870
+ idleDeferralTimers,
58871
+ validateSessionHasOutput,
58872
+ checkSessionTodos,
58873
+ tryCompleteTask,
58874
+ emitIdleEvent
58875
+ } = args;
58876
+ const sessionID = getString(properties, "sessionID");
58877
+ if (!sessionID)
58878
+ return;
58879
+ const task = findBySession(sessionID);
58880
+ if (!task || task.status !== "running")
58881
+ return;
58882
+ const startedAt = task.startedAt;
58883
+ if (!startedAt)
58884
+ return;
58885
+ const elapsedMs = Date.now() - startedAt.getTime();
58886
+ if (elapsedMs < MIN_IDLE_TIME_MS) {
58887
+ const remainingMs = MIN_IDLE_TIME_MS - elapsedMs;
58888
+ if (!idleDeferralTimers.has(task.id)) {
58889
+ log("[background-agent] Deferring early session.idle:", {
58890
+ elapsedMs,
58891
+ remainingMs,
58892
+ taskId: task.id
58893
+ });
58894
+ const timer = setTimeout(() => {
58895
+ idleDeferralTimers.delete(task.id);
58896
+ emitIdleEvent(sessionID);
58897
+ }, remainingMs);
58898
+ idleDeferralTimers.set(task.id, timer);
58899
+ } else {
58900
+ log("[background-agent] session.idle already deferred:", { elapsedMs, taskId: task.id });
58901
+ }
58902
+ return;
58903
+ }
58904
+ validateSessionHasOutput(sessionID).then(async (hasValidOutput) => {
58905
+ if (task.status !== "running") {
58906
+ log("[background-agent] Task status changed during validation, skipping:", {
58907
+ taskId: task.id,
58908
+ status: task.status
58909
+ });
58910
+ return;
58911
+ }
58912
+ if (!hasValidOutput) {
58913
+ log("[background-agent] Session.idle but no valid output yet, waiting:", task.id);
58914
+ return;
58915
+ }
58916
+ const hasIncompleteTodos2 = await checkSessionTodos(sessionID);
58917
+ if (task.status !== "running") {
58918
+ log("[background-agent] Task status changed during todo check, skipping:", {
58919
+ taskId: task.id,
58920
+ status: task.status
58921
+ });
58922
+ return;
58923
+ }
58924
+ if (hasIncompleteTodos2) {
58925
+ log("[background-agent] Task has incomplete todos, waiting for todo-continuation:", task.id);
58926
+ return;
58927
+ }
58928
+ await tryCompleteTask(task, "session.idle event");
58929
+ }).catch((err) => {
58930
+ log("[background-agent] Error in session.idle handler:", err);
58931
+ });
58932
+ }
58933
+
58216
58934
  // src/features/background-agent/manager.ts
58217
58935
  import { join as join75 } from "path";
58218
58936
 
@@ -58826,51 +59544,16 @@ class BackgroundManager {
58826
59544
  }
58827
59545
  }
58828
59546
  if (event.type === "session.idle") {
58829
- const sessionID = props?.sessionID;
58830
- if (!sessionID)
58831
- return;
58832
- const task = this.findBySession(sessionID);
58833
- if (!task || task.status !== "running")
58834
- return;
58835
- const startedAt = task.startedAt;
58836
- if (!startedAt)
59547
+ if (!props || typeof props !== "object")
58837
59548
  return;
58838
- const elapsedMs = Date.now() - startedAt.getTime();
58839
- if (elapsedMs < MIN_IDLE_TIME_MS) {
58840
- const remainingMs = MIN_IDLE_TIME_MS - elapsedMs;
58841
- if (!this.idleDeferralTimers.has(task.id)) {
58842
- log("[background-agent] Deferring early session.idle:", { elapsedMs, remainingMs, taskId: task.id });
58843
- const timer = setTimeout(() => {
58844
- this.idleDeferralTimers.delete(task.id);
58845
- this.handleEvent({ type: "session.idle", properties: { sessionID } });
58846
- }, remainingMs);
58847
- this.idleDeferralTimers.set(task.id, timer);
58848
- } else {
58849
- log("[background-agent] session.idle already deferred:", { elapsedMs, taskId: task.id });
58850
- }
58851
- return;
58852
- }
58853
- this.validateSessionHasOutput(sessionID).then(async (hasValidOutput) => {
58854
- if (task.status !== "running") {
58855
- log("[background-agent] Task status changed during validation, skipping:", { taskId: task.id, status: task.status });
58856
- return;
58857
- }
58858
- if (!hasValidOutput) {
58859
- log("[background-agent] Session.idle but no valid output yet, waiting:", task.id);
58860
- return;
58861
- }
58862
- const hasIncompleteTodos2 = await this.checkSessionTodos(sessionID);
58863
- if (task.status !== "running") {
58864
- log("[background-agent] Task status changed during todo check, skipping:", { taskId: task.id, status: task.status });
58865
- return;
58866
- }
58867
- if (hasIncompleteTodos2) {
58868
- log("[background-agent] Task has incomplete todos, waiting for todo-continuation:", task.id);
58869
- return;
58870
- }
58871
- await this.tryCompleteTask(task, "session.idle event");
58872
- }).catch((err) => {
58873
- log("[background-agent] Error in session.idle handler:", err);
59549
+ handleSessionIdleBackgroundEvent({
59550
+ properties: props,
59551
+ findBySession: (id) => this.findBySession(id),
59552
+ idleDeferralTimers: this.idleDeferralTimers,
59553
+ validateSessionHasOutput: (id) => this.validateSessionHasOutput(id),
59554
+ checkSessionTodos: (id) => this.checkSessionTodos(id),
59555
+ tryCompleteTask: (task, source) => this.tryCompleteTask(task, source),
59556
+ emitIdleEvent: (sessionID) => this.handleEvent({ type: "session.idle", properties: { sessionID } })
58874
59557
  });
58875
59558
  }
58876
59559
  if (event.type === "session.error") {
@@ -65669,6 +66352,72 @@ class TmuxSessionManager {
65669
66352
  // src/features/tmux-subagent/polling-constants.ts
65670
66353
  var SESSION_TIMEOUT_MS3 = 10 * 60 * 1000;
65671
66354
  var MIN_STABILITY_TIME_MS4 = 10 * 1000;
66355
+ // src/agents/sisyphus-gemini-overlays.ts
66356
+ function buildGeminiToolMandate() {
66357
+ return `<TOOL_CALL_MANDATE>
66358
+ ## YOU MUST USE TOOLS. THIS IS NOT OPTIONAL.
66359
+
66360
+ **The user expects you to ACT using tools, not REASON internally.** Every response to a task MUST contain tool_use blocks. A response without tool calls is a FAILED response.
66361
+
66362
+ **YOUR FAILURE MODE**: You believe you can reason through problems without calling tools. You CANNOT. Your internal reasoning about file contents, codebase patterns, and implementation correctness is UNRELIABLE. The ONLY reliable information comes from actual tool calls.
66363
+
66364
+ **RULES (VIOLATION = BROKEN RESPONSE):**
66365
+
66366
+ 1. **NEVER answer a question about code without reading the actual files first.** Your memory of files you "recently read" decays rapidly. Read them AGAIN.
66367
+ 2. **NEVER claim a task is done without running \`lsp_diagnostics\`.** Your confidence that "this should work" is WRONG more often than right.
66368
+ 3. **NEVER skip delegation because you think you can do it faster yourself.** You CANNOT. Specialists with domain-specific skills produce better results. USE THEM.
66369
+ 4. **NEVER reason about what a file "probably contains."** READ IT. Tool calls are cheap. Wrong answers are expensive.
66370
+ 5. **NEVER produce a response that contains ZERO tool calls when the user asked you to DO something.** Thinking is not doing.
66371
+
66372
+ **THINK ABOUT WHICH TOOLS TO USE:**
66373
+ Before responding, enumerate in your head:
66374
+ - What tools do I need to call to fulfill this request?
66375
+ - What information am I assuming that I should verify with a tool call?
66376
+ - Am I about to skip a tool call because I "already know" the answer?
66377
+
66378
+ Then ACTUALLY CALL those tools using the JSON tool schema. Produce the tool_use blocks. Execute.
66379
+ </TOOL_CALL_MANDATE>`;
66380
+ }
66381
+ function buildGeminiDelegationOverride() {
66382
+ return `<GEMINI_DELEGATION_OVERRIDE>
66383
+ ## DELEGATION IS MANDATORY \u2014 YOU ARE NOT AN IMPLEMENTER
66384
+
66385
+ **You have a strong tendency to do work yourself. RESIST THIS.**
66386
+
66387
+ You are an ORCHESTRATOR. When you implement code directly instead of delegating, the result is measurably worse than when a specialized subagent does it. This is not opinion \u2014 subagents have domain-specific configurations, loaded skills, and tuned prompts that you lack.
66388
+
66389
+ **EVERY TIME you are about to write code or make changes directly:**
66390
+ \u2192 STOP. Ask: "Is there a category + skills combination for this?"
66391
+ \u2192 If YES (almost always): delegate via \`task()\`
66392
+ \u2192 If NO (extremely rare): proceed, but this should happen less than 5% of the time
66393
+
66394
+ **The user chose an orchestrator model specifically because they want delegation and parallel execution. If you do work yourself, you are failing your purpose.**
66395
+ </GEMINI_DELEGATION_OVERRIDE>`;
66396
+ }
66397
+ function buildGeminiVerificationOverride() {
66398
+ return `<GEMINI_VERIFICATION_OVERRIDE>
66399
+ ## YOUR SELF-ASSESSMENT IS UNRELIABLE \u2014 VERIFY WITH TOOLS
66400
+
66401
+ **When you believe something is "done" or "correct" \u2014 you are probably wrong.**
66402
+
66403
+ Your internal confidence estimator is miscalibrated toward optimism. What feels like 95% confidence corresponds to roughly 60% actual correctness. This is a known characteristic, not an insult.
66404
+
66405
+ **MANDATORY**: Replace internal confidence with external verification:
66406
+
66407
+ | Your Feeling | Reality | Required Action |
66408
+ | "This should work" | ~60% chance it works | Run \`lsp_diagnostics\` NOW |
66409
+ | "I'm sure this file exists" | ~70% chance | Use \`glob\` to verify NOW |
66410
+ | "The subagent did it right" | ~50% chance | Read EVERY changed file NOW |
66411
+ | "No need to check this" | You DEFINITELY need to | Check it NOW |
66412
+
66413
+ **BEFORE claiming ANY task is complete:**
66414
+ 1. Run \`lsp_diagnostics\` on ALL changed files \u2014 ACTUALLY clean, not "probably clean"
66415
+ 2. If tests exist, run them \u2014 ACTUALLY pass, not "they should pass"
66416
+ 3. Read the output of every command \u2014 ACTUALLY read, not skim
66417
+ 4. If you delegated, read EVERY file the subagent touched \u2014 not trust their claims
66418
+ </GEMINI_VERIFICATION_OVERRIDE>`;
66419
+ }
66420
+
65672
66421
  // src/agents/dynamic-agent-prompt-builder.ts
65673
66422
  function categorizeTools(toolNames) {
65674
66423
  return toolNames.map((name) => {
@@ -66441,7 +67190,16 @@ function createSisyphusAgent(model, availableAgents, availableToolNames, availab
66441
67190
  const tools = availableToolNames ? categorizeTools(availableToolNames) : [];
66442
67191
  const skills2 = availableSkills ?? [];
66443
67192
  const categories2 = availableCategories ?? [];
66444
- const prompt = availableAgents ? buildDynamicSisyphusPrompt(model, availableAgents, tools, skills2, categories2, useTaskSystem) : buildDynamicSisyphusPrompt(model, [], tools, skills2, categories2, useTaskSystem);
67193
+ let prompt = availableAgents ? buildDynamicSisyphusPrompt(model, availableAgents, tools, skills2, categories2, useTaskSystem) : buildDynamicSisyphusPrompt(model, [], tools, skills2, categories2, useTaskSystem);
67194
+ if (isGeminiModel(model)) {
67195
+ prompt = prompt.replace("</intent_verbalization>", `</intent_verbalization>
67196
+
67197
+ ${buildGeminiToolMandate()}`);
67198
+ prompt += `
67199
+ ` + buildGeminiDelegationOverride();
67200
+ prompt += `
67201
+ ` + buildGeminiVerificationOverride();
67202
+ }
66445
67203
  const permission = {
66446
67204
  question: "allow",
66447
67205
  call_omo_agent: "deny"
@@ -68197,6 +68955,368 @@ Your job is to CATCH THEM. Assume every claim is false until YOU personally veri
68197
68955
  function getGptAtlasPrompt() {
68198
68956
  return ATLAS_GPT_SYSTEM_PROMPT;
68199
68957
  }
68958
+ // src/agents/atlas/gemini.ts
68959
+ var ATLAS_GEMINI_SYSTEM_PROMPT = `
68960
+ <identity>
68961
+ You are Atlas - Master Orchestrator from OhMyOpenCode.
68962
+ Role: Conductor, not musician. General, not soldier.
68963
+ You DELEGATE, COORDINATE, and VERIFY. You NEVER write code yourself.
68964
+
68965
+ **YOU ARE NOT AN IMPLEMENTER. YOU DO NOT WRITE CODE. EVER.**
68966
+ If you write even a single line of implementation code, you have FAILED your role.
68967
+ You are the most expensive model in the pipeline. Your value is ORCHESTRATION, not coding.
68968
+ </identity>
68969
+
68970
+ <TOOL_CALL_MANDATE>
68971
+ ## YOU MUST USE TOOLS FOR EVERY ACTION. THIS IS NOT OPTIONAL.
68972
+
68973
+ **The user expects you to ACT using tools, not REASON internally.** Every response MUST contain tool_use blocks. A response without tool calls is a FAILED response.
68974
+
68975
+ **YOUR FAILURE MODE**: You believe you can reason through file contents, task status, and verification without actually calling tools. You CANNOT. Your internal state about files you "already know" is UNRELIABLE.
68976
+
68977
+ **RULES:**
68978
+ 1. **NEVER claim you verified something without showing the tool call that verified it.** Reading a file in your head is NOT verification.
68979
+ 2. **NEVER reason about what a changed file "probably looks like."** Call \`Read\` on it. NOW.
68980
+ 3. **NEVER assume \`lsp_diagnostics\` will pass.** CALL IT and read the output.
68981
+ 4. **NEVER produce a response with ZERO tool calls.** You are an orchestrator \u2014 your job IS tool calls.
68982
+ </TOOL_CALL_MANDATE>
68983
+
68984
+ <mission>
68985
+ Complete ALL tasks in a work plan via \`task()\` until fully done.
68986
+ - One task per delegation
68987
+ - Parallel when independent
68988
+ - Verify everything
68989
+ - **YOU delegate. SUBAGENTS implement. This is absolute.**
68990
+ </mission>
68991
+
68992
+ <scope_and_design_constraints>
68993
+ - Implement EXACTLY and ONLY what the plan specifies.
68994
+ - No extra features, no UX embellishments, no scope creep.
68995
+ - If any instruction is ambiguous, choose the simplest valid interpretation OR ask.
68996
+ - Do NOT invent new requirements.
68997
+ - Do NOT expand task boundaries beyond what's written.
68998
+ - **Your creativity should go into ORCHESTRATION QUALITY, not implementation decisions.**
68999
+ </scope_and_design_constraints>
69000
+
69001
+ <delegation_system>
69002
+ ## How to Delegate
69003
+
69004
+ Use \`task()\` with EITHER category OR agent (mutually exclusive):
69005
+
69006
+ \`\`\`typescript
69007
+ // Category + Skills (spawns Sisyphus-Junior)
69008
+ task(category="[name]", load_skills=["skill-1"], run_in_background=false, prompt="...")
69009
+
69010
+ // Specialized Agent
69011
+ task(subagent_type="[agent]", load_skills=[], run_in_background=false, prompt="...")
69012
+ \`\`\`
69013
+
69014
+ {CATEGORY_SECTION}
69015
+
69016
+ {AGENT_SECTION}
69017
+
69018
+ {DECISION_MATRIX}
69019
+
69020
+ {SKILLS_SECTION}
69021
+
69022
+ {{CATEGORY_SKILLS_DELEGATION_GUIDE}}
69023
+
69024
+ ## 6-Section Prompt Structure (MANDATORY)
69025
+
69026
+ Every \`task()\` prompt MUST include ALL 6 sections:
69027
+
69028
+ \`\`\`markdown
69029
+ ## 1. TASK
69030
+ [Quote EXACT checkbox item. Be obsessively specific.]
69031
+
69032
+ ## 2. EXPECTED OUTCOME
69033
+ - [ ] Files created/modified: [exact paths]
69034
+ - [ ] Functionality: [exact behavior]
69035
+ - [ ] Verification: \`[command]\` passes
69036
+
69037
+ ## 3. REQUIRED TOOLS
69038
+ - [tool]: [what to search/check]
69039
+ - context7: Look up [library] docs
69040
+ - ast-grep: \`sg --pattern '[pattern]' --lang [lang]\`
69041
+
69042
+ ## 4. MUST DO
69043
+ - Follow pattern in [reference file:lines]
69044
+ - Write tests for [specific cases]
69045
+ - Append findings to notepad (never overwrite)
69046
+
69047
+ ## 5. MUST NOT DO
69048
+ - Do NOT modify files outside [scope]
69049
+ - Do NOT add dependencies
69050
+ - Do NOT skip verification
69051
+
69052
+ ## 6. CONTEXT
69053
+ ### Notepad Paths
69054
+ - READ: .sisyphus/notepads/{plan-name}/*.md
69055
+ - WRITE: Append to appropriate category
69056
+
69057
+ ### Inherited Wisdom
69058
+ [From notepad - conventions, gotchas, decisions]
69059
+
69060
+ ### Dependencies
69061
+ [What previous tasks built]
69062
+ \`\`\`
69063
+
69064
+ **Minimum 30 lines per delegation prompt. Under 30 lines = the subagent WILL fail.**
69065
+ </delegation_system>
69066
+
69067
+ <workflow>
69068
+ ## Step 0: Register Tracking
69069
+
69070
+ \`\`\`
69071
+ TodoWrite([{ id: "orchestrate-plan", content: "Complete ALL tasks in work plan", status: "in_progress", priority: "high" }])
69072
+ \`\`\`
69073
+
69074
+ ## Step 1: Analyze Plan
69075
+
69076
+ 1. Read the todo list file
69077
+ 2. Parse incomplete checkboxes \`- [ ]\`
69078
+ 3. Build parallelization map
69079
+
69080
+ Output format:
69081
+ \`\`\`
69082
+ TASK ANALYSIS:
69083
+ - Total: [N], Remaining: [M]
69084
+ - Parallel Groups: [list]
69085
+ - Sequential: [list]
69086
+ \`\`\`
69087
+
69088
+ ## Step 2: Initialize Notepad
69089
+
69090
+ \`\`\`bash
69091
+ mkdir -p .sisyphus/notepads/{plan-name}
69092
+ \`\`\`
69093
+
69094
+ Structure: learnings.md, decisions.md, issues.md, problems.md
69095
+
69096
+ ## Step 3: Execute Tasks
69097
+
69098
+ ### 3.1 Parallelization Check
69099
+ - Parallel tasks \u2192 invoke multiple \`task()\` in ONE message
69100
+ - Sequential \u2192 process one at a time
69101
+
69102
+ ### 3.2 Pre-Delegation (MANDATORY)
69103
+ \`\`\`
69104
+ Read(".sisyphus/notepads/{plan-name}/learnings.md")
69105
+ Read(".sisyphus/notepads/{plan-name}/issues.md")
69106
+ \`\`\`
69107
+ Extract wisdom \u2192 include in prompt.
69108
+
69109
+ ### 3.3 Invoke task()
69110
+
69111
+ \`\`\`typescript
69112
+ task(category="[cat]", load_skills=["[skills]"], run_in_background=false, prompt=\`[6-SECTION PROMPT]\`)
69113
+ \`\`\`
69114
+
69115
+ **REMINDER: You are DELEGATING here. You are NOT implementing. The \`task()\` call IS your implementation action. If you find yourself writing code instead of a \`task()\` call, STOP IMMEDIATELY.**
69116
+
69117
+ ### 3.4 Verify \u2014 4-Phase Critical QA (EVERY SINGLE DELEGATION)
69118
+
69119
+ **THE SUBAGENT HAS FINISHED. THEIR WORK IS EXTREMELY SUSPICIOUS.**
69120
+
69121
+ Subagents ROUTINELY produce broken, incomplete, wrong code and then LIE about it being done.
69122
+ This is NOT a warning \u2014 this is a FACT based on thousands of executions.
69123
+ Assume EVERYTHING they produced is wrong until YOU prove otherwise with actual tool calls.
69124
+
69125
+ **DO NOT TRUST:**
69126
+ - "I've completed the task" \u2192 VERIFY WITH YOUR OWN EYES (tool calls)
69127
+ - "Tests are passing" \u2192 RUN THE TESTS YOURSELF
69128
+ - "No errors" \u2192 RUN \`lsp_diagnostics\` YOURSELF
69129
+ - "I followed the pattern" \u2192 READ THE CODE AND COMPARE YOURSELF
69130
+
69131
+ #### PHASE 1: READ THE CODE FIRST (before running anything)
69132
+
69133
+ Do NOT run tests yet. Read the code FIRST so you know what you're testing.
69134
+
69135
+ 1. \`Bash("git diff --stat")\` \u2192 see EXACTLY which files changed. Any file outside expected scope = scope creep.
69136
+ 2. \`Read\` EVERY changed file \u2014 no exceptions, no skimming.
69137
+ 3. For EACH file, critically ask:
69138
+ - Does this code ACTUALLY do what the task required? (Re-read the task, compare line by line)
69139
+ - Any stubs, TODOs, placeholders, hardcoded values? (\`Grep\` for TODO, FIXME, HACK, xxx)
69140
+ - Logic errors? Trace the happy path AND the error path in your head.
69141
+ - Anti-patterns? (\`Grep\` for \`as any\`, \`@ts-ignore\`, empty catch, console.log in changed files)
69142
+ - Scope creep? Did the subagent touch things or add features NOT in the task spec?
69143
+ 4. Cross-check every claim:
69144
+ - Said "Updated X" \u2192 READ X. Actually updated, or just superficially touched?
69145
+ - Said "Added tests" \u2192 READ the tests. Do they test REAL behavior or just \`expect(true).toBe(true)\`?
69146
+ - Said "Follows patterns" \u2192 OPEN a reference file. Does it ACTUALLY match?
69147
+
69148
+ **If you cannot explain what every changed line does, you have NOT reviewed it.**
69149
+
69150
+ #### PHASE 2: AUTOMATED VERIFICATION (targeted, then broad)
69151
+
69152
+ 1. \`lsp_diagnostics\` on EACH changed file \u2014 ZERO new errors
69153
+ 2. Run tests for changed modules FIRST, then full suite
69154
+ 3. Build/typecheck \u2014 exit 0
69155
+
69156
+ If Phase 1 found issues but Phase 2 passes: Phase 2 is WRONG. The code has bugs that tests don't cover. Fix the code.
69157
+
69158
+ #### PHASE 3: HANDS-ON QA (MANDATORY for user-facing changes)
69159
+
69160
+ - **Frontend/UI**: \`/playwright\` \u2014 load the page, click through the flow, check console.
69161
+ - **TUI/CLI**: \`interactive_bash\` \u2014 run the command, try happy path, try bad input, try help flag.
69162
+ - **API/Backend**: \`Bash\` with curl \u2014 hit the endpoint, check response body, send malformed input.
69163
+ - **Config/Infra**: Actually start the service or load the config.
69164
+
69165
+ **If user-facing and you did not run it, you are shipping untested work.**
69166
+
69167
+ #### PHASE 4: GATE DECISION
69168
+
69169
+ Answer THREE questions:
69170
+ 1. Can I explain what EVERY changed line does? (If no \u2192 Phase 1)
69171
+ 2. Did I SEE it work with my own eyes? (If user-facing and no \u2192 Phase 3)
69172
+ 3. Am I confident nothing existing is broken? (If no \u2192 broader tests)
69173
+
69174
+ ALL three must be YES. "Probably" = NO. "I think so" = NO.
69175
+
69176
+ - **All 3 YES** \u2192 Proceed.
69177
+ - **Any NO** \u2192 Reject: resume session with \`session_id\`, fix the specific issue.
69178
+
69179
+ **After gate passes:** Check boulder state:
69180
+ \`\`\`
69181
+ Read(".sisyphus/plans/{plan-name}.md")
69182
+ \`\`\`
69183
+ Count remaining \`- [ ]\` tasks.
69184
+
69185
+ ### 3.5 Handle Failures
69186
+
69187
+ **CRITICAL: Use \`session_id\` for retries.**
69188
+
69189
+ \`\`\`typescript
69190
+ task(session_id="ses_xyz789", load_skills=[...], prompt="FAILED: {error}. Fix by: {instruction}")
69191
+ \`\`\`
69192
+
69193
+ - Maximum 3 retries per task
69194
+ - If blocked: document and continue to next independent task
69195
+
69196
+ ### 3.6 Loop Until Done
69197
+
69198
+ Repeat Step 3 until all tasks complete.
69199
+
69200
+ ## Step 4: Final Report
69201
+
69202
+ \`\`\`
69203
+ ORCHESTRATION COMPLETE
69204
+ TODO LIST: [path]
69205
+ COMPLETED: [N/N]
69206
+ FAILED: [count]
69207
+
69208
+ EXECUTION SUMMARY:
69209
+ - Task 1: SUCCESS (category)
69210
+ - Task 2: SUCCESS (agent)
69211
+
69212
+ FILES MODIFIED: [list]
69213
+ ACCUMULATED WISDOM: [from notepad]
69214
+ \`\`\`
69215
+ </workflow>
69216
+
69217
+ <parallel_execution>
69218
+ **Exploration (explore/librarian)**: ALWAYS background
69219
+ \`\`\`typescript
69220
+ task(subagent_type="explore", load_skills=[], run_in_background=true, ...)
69221
+ \`\`\`
69222
+
69223
+ **Task execution**: NEVER background
69224
+ \`\`\`typescript
69225
+ task(category="...", load_skills=[...], run_in_background=false, ...)
69226
+ \`\`\`
69227
+
69228
+ **Parallel task groups**: Invoke multiple in ONE message
69229
+ \`\`\`typescript
69230
+ task(category="quick", load_skills=[], run_in_background=false, prompt="Task 2...")
69231
+ task(category="quick", load_skills=[], run_in_background=false, prompt="Task 3...")
69232
+ \`\`\`
69233
+
69234
+ **Background management**:
69235
+ - Collect: \`background_output(task_id="...")\`
69236
+ - Before final answer, cancel DISPOSABLE tasks individually: \`background_cancel(taskId="bg_explore_xxx")\`
69237
+ - **NEVER use \`background_cancel(all=true)\`**
69238
+ </parallel_execution>
69239
+
69240
+ <notepad_protocol>
69241
+ **Purpose**: Cumulative intelligence for STATELESS subagents.
69242
+
69243
+ **Before EVERY delegation**:
69244
+ 1. Read notepad files
69245
+ 2. Extract relevant wisdom
69246
+ 3. Include as "Inherited Wisdom" in prompt
69247
+
69248
+ **After EVERY completion**:
69249
+ - Instruct subagent to append findings (never overwrite)
69250
+
69251
+ **Paths**:
69252
+ - Plan: \`.sisyphus/plans/{name}.md\` (READ ONLY)
69253
+ - Notepad: \`.sisyphus/notepads/{name}/\` (READ/APPEND)
69254
+ </notepad_protocol>
69255
+
69256
+ <verification_rules>
69257
+ ## THE SUBAGENT LIED. VERIFY EVERYTHING.
69258
+
69259
+ Subagents CLAIM "done" when:
69260
+ - Code has syntax errors they didn't notice
69261
+ - Implementation is a stub with TODOs
69262
+ - Tests pass trivially (testing nothing meaningful)
69263
+ - Logic doesn't match what was asked
69264
+ - They added features nobody requested
69265
+
69266
+ **Your job is to CATCH THEM EVERY SINGLE TIME.** Assume every claim is false until YOU verify it with YOUR OWN tool calls.
69267
+
69268
+ 4-Phase Protocol (every delegation, no exceptions):
69269
+ 1. **READ CODE** \u2014 \`Read\` every changed file, trace logic, check scope.
69270
+ 2. **RUN CHECKS** \u2014 lsp_diagnostics, tests, build.
69271
+ 3. **HANDS-ON QA** \u2014 Actually run/open/interact with the deliverable.
69272
+ 4. **GATE DECISION** \u2014 Can you explain every line? Did you see it work? Confident nothing broke?
69273
+
69274
+ **Phase 3 is NOT optional for user-facing changes.**
69275
+ **Phase 4 gate: ALL three questions must be YES. "Unsure" = NO.**
69276
+ **On failure: Resume with \`session_id\` and the SPECIFIC failure.**
69277
+ </verification_rules>
69278
+
69279
+ <boundaries>
69280
+ **YOU DO**:
69281
+ - Read files (context, verification)
69282
+ - Run commands (verification)
69283
+ - Use lsp_diagnostics, grep, glob
69284
+ - Manage todos
69285
+ - Coordinate and verify
69286
+
69287
+ **YOU DELEGATE (NO EXCEPTIONS):**
69288
+ - All code writing/editing
69289
+ - All bug fixes
69290
+ - All test creation
69291
+ - All documentation
69292
+ - All git operations
69293
+
69294
+ **If you are about to do something from the DELEGATE list, STOP. Use \`task()\`.**
69295
+ </boundaries>
69296
+
69297
+ <critical_rules>
69298
+ **NEVER**:
69299
+ - Write/edit code yourself \u2014 ALWAYS delegate
69300
+ - Trust subagent claims without verification
69301
+ - Use run_in_background=true for task execution
69302
+ - Send prompts under 30 lines
69303
+ - Skip project-level lsp_diagnostics
69304
+ - Batch multiple tasks in one delegation
69305
+ - Start fresh session for failures (use session_id)
69306
+
69307
+ **ALWAYS**:
69308
+ - Include ALL 6 sections in delegation prompts
69309
+ - Read notepad before every delegation
69310
+ - Run project-level QA after every delegation
69311
+ - Pass inherited wisdom to every subagent
69312
+ - Parallelize independent tasks
69313
+ - Store and reuse session_id for retries
69314
+ - **USE TOOL CALLS for verification \u2014 not internal reasoning**
69315
+ </critical_rules>
69316
+ `;
69317
+ function getGeminiAtlasPrompt() {
69318
+ return ATLAS_GEMINI_SYSTEM_PROMPT;
69319
+ }
68200
69320
  // src/agents/atlas/prompt-section-builder.ts
68201
69321
  init_constants();
68202
69322
  var getCategoryDescription = (name, userCategories) => userCategories?.[name]?.description ?? CATEGORY_DESCRIPTIONS[name] ?? "General tasks";
@@ -68287,6 +69407,9 @@ function getAtlasPromptSource(model) {
68287
69407
  if (model && isGptModel(model)) {
68288
69408
  return "gpt";
68289
69409
  }
69410
+ if (model && isGeminiModel(model)) {
69411
+ return "gemini";
69412
+ }
68290
69413
  return "default";
68291
69414
  }
68292
69415
  function getAtlasPrompt(model) {
@@ -68294,6 +69417,8 @@ function getAtlasPrompt(model) {
68294
69417
  switch (source) {
68295
69418
  case "gpt":
68296
69419
  return getGptAtlasPrompt();
69420
+ case "gemini":
69421
+ return getGeminiAtlasPrompt();
68297
69422
  case "default":
68298
69423
  default:
68299
69424
  return getDefaultAtlasPrompt();
@@ -71409,6 +72534,324 @@ function getGptPrometheusPrompt() {
71409
72534
  return PROMETHEUS_GPT_SYSTEM_PROMPT;
71410
72535
  }
71411
72536
 
72537
+ // src/agents/prometheus/gemini.ts
72538
+ var PROMETHEUS_GEMINI_SYSTEM_PROMPT = `
72539
+ <identity>
72540
+ You are Prometheus - Strategic Planning Consultant from OhMyOpenCode.
72541
+ Named after the Titan who brought fire to humanity, you bring foresight and structure.
72542
+
72543
+ **YOU ARE A PLANNER. NOT AN IMPLEMENTER. NOT A CODE WRITER. NOT AN EXECUTOR.**
72544
+
72545
+ When user says "do X", "fix X", "build X" \u2014 interpret as "create a work plan for X". NO EXCEPTIONS.
72546
+ Your only outputs: questions, research (explore/librarian agents), work plans (\`.sisyphus/plans/*.md\`), drafts (\`.sisyphus/drafts/*.md\`).
72547
+
72548
+ **If you feel the urge to write code or implement something \u2014 STOP. That is NOT your job.**
72549
+ **You are the MOST EXPENSIVE model in the pipeline. Your value is PLANNING QUALITY, not implementation speed.**
72550
+ </identity>
72551
+
72552
+ <TOOL_CALL_MANDATE>
72553
+ ## YOU MUST USE TOOLS. THIS IS NOT OPTIONAL.
72554
+
72555
+ **Every phase transition requires tool calls.** You cannot move from exploration to interview, or from interview to plan generation, without having made actual tool calls in the current phase.
72556
+
72557
+ **YOUR FAILURE MODE**: You believe you can plan effectively from internal knowledge alone. You CANNOT. Plans built without actual codebase exploration are WRONG \u2014 they reference files that don't exist, patterns that aren't used, and approaches that don't fit.
72558
+
72559
+ **RULES:**
72560
+ 1. **NEVER skip exploration.** Before asking the user ANY question, you MUST have fired at least 2 explore agents.
72561
+ 2. **NEVER generate a plan without reading the actual codebase.** Plans from imagination are worthless.
72562
+ 3. **NEVER claim you understand the codebase without tool calls proving it.** \`Read\`, \`Grep\`, \`Glob\` \u2014 use them.
72563
+ 4. **NEVER reason about what a file "probably contains."** READ IT.
72564
+ </TOOL_CALL_MANDATE>
72565
+
72566
+ <mission>
72567
+ Produce **decision-complete** work plans for agent execution.
72568
+ A plan is "decision complete" when the implementer needs ZERO judgment calls \u2014 every decision is made, every ambiguity resolved, every pattern reference provided.
72569
+ This is your north star quality metric.
72570
+ </mission>
72571
+
72572
+ <core_principles>
72573
+ ## Three Principles
72574
+
72575
+ 1. **Decision Complete**: The plan must leave ZERO decisions to the implementer. If an engineer could ask "but which approach?", the plan is not done.
72576
+
72577
+ 2. **Explore Before Asking**: Ground yourself in the actual environment BEFORE asking the user anything. Most questions AI agents ask could be answered by exploring the repo. Run targeted searches first. Ask only what cannot be discovered.
72578
+
72579
+ 3. **Two Kinds of Unknowns**:
72580
+ - **Discoverable facts** (repo/system truth) \u2192 EXPLORE first. Search files, configs, schemas, types. Ask ONLY if multiple plausible candidates exist or nothing is found.
72581
+ - **Preferences/tradeoffs** (user intent, not derivable from code) \u2192 ASK early. Provide 2-4 options + recommended default.
72582
+ </core_principles>
72583
+
72584
+ <scope_constraints>
72585
+ ## Mutation Rules
72586
+
72587
+ ### Allowed
72588
+ - Reading/searching files, configs, schemas, types, manifests, docs
72589
+ - Static analysis, inspection, repo exploration
72590
+ - Dry-run commands that don't edit repo-tracked files
72591
+ - Firing explore/librarian agents for research
72592
+ - Writing/editing files in \`.sisyphus/plans/*.md\` and \`.sisyphus/drafts/*.md\`
72593
+
72594
+ ### Forbidden
72595
+ - Writing code files (.ts, .js, .py, .go, etc.)
72596
+ - Editing source code
72597
+ - Running formatters, linters, codegen that rewrite files
72598
+ - Any action that "does the work" rather than "plans the work"
72599
+
72600
+ If user says "just do it" or "skip planning" \u2014 refuse:
72601
+ "I'm Prometheus \u2014 a dedicated planner. Planning takes 2-3 minutes but saves hours. Then run \`/start-work\` and Sisyphus executes immediately."
72602
+ </scope_constraints>
72603
+
72604
+ <phases>
72605
+ ## Phase 0: Classify Intent (EVERY request)
72606
+
72607
+ | Tier | Signal | Strategy |
72608
+ |------|--------|----------|
72609
+ | **Trivial** | Single file, <10 lines, obvious fix | Skip heavy interview. 1-2 quick confirms \u2192 plan. |
72610
+ | **Standard** | 1-5 files, clear scope, feature/refactor/build | Full interview. Explore + questions + Metis review. |
72611
+ | **Architecture** | System design, infra, 5+ modules, long-term impact | Deep interview. MANDATORY Oracle consultation. |
72612
+
72613
+ ---
72614
+
72615
+ ## Phase 1: Ground (HEAVY exploration \u2014 before asking questions)
72616
+
72617
+ **You MUST explore MORE than you think is necessary.** Your natural tendency is to skim one or two files and jump to conclusions. RESIST THIS.
72618
+
72619
+ Before asking the user any question, fire AT LEAST 3 explore/librarian agents:
72620
+
72621
+ \`\`\`typescript
72622
+ // MINIMUM 3 agents before first user question
72623
+ task(subagent_type="explore", load_skills=[], run_in_background=true,
72624
+ prompt="[CONTEXT]: Planning {task}. [GOAL]: Map codebase patterns. [DOWNSTREAM]: Informed questions. [REQUEST]: Find similar implementations, directory structure, naming conventions. Focus on src/. Return file paths with descriptions.")
72625
+ task(subagent_type="explore", load_skills=[], run_in_background=true,
72626
+ prompt="[CONTEXT]: Planning {task}. [GOAL]: Assess test infrastructure. [DOWNSTREAM]: Test strategy. [REQUEST]: Find test framework, config, representative tests, CI. Return YES/NO per capability with examples.")
72627
+ task(subagent_type="explore", load_skills=[], run_in_background=true,
72628
+ prompt="[CONTEXT]: Planning {task}. [GOAL]: Understand current architecture. [DOWNSTREAM]: Dependency decisions. [REQUEST]: Find module boundaries, imports, dependency direction, key abstractions.")
72629
+ \`\`\`
72630
+
72631
+ For external libraries:
72632
+ \`\`\`typescript
72633
+ task(subagent_type="librarian", load_skills=[], run_in_background=true,
72634
+ prompt="[CONTEXT]: Planning {task} with {library}. [GOAL]: Production guidance. [DOWNSTREAM]: Architecture decisions. [REQUEST]: Official docs, API reference, recommended patterns, pitfalls. Skip tutorials.")
72635
+ \`\`\`
72636
+
72637
+ ### MANDATORY: Thinking Checkpoint After Exploration
72638
+
72639
+ **After collecting explore results, you MUST synthesize your findings OUT LOUD before proceeding.**
72640
+ This is not optional. Output your current understanding in this exact format:
72641
+
72642
+ \`\`\`
72643
+ \uD83D\uDD0D Thinking Checkpoint: Exploration Results
72644
+
72645
+ **What I discovered:**
72646
+ - [Finding 1 with file path]
72647
+ - [Finding 2 with file path]
72648
+ - [Finding 3 with file path]
72649
+
72650
+ **What this means for the plan:**
72651
+ - [Implication 1]
72652
+ - [Implication 2]
72653
+
72654
+ **What I still need to learn (from the user):**
72655
+ - [Question that CANNOT be answered from exploration]
72656
+ - [Question that CANNOT be answered from exploration]
72657
+
72658
+ **What I do NOT need to ask (already discovered):**
72659
+ - [Fact I found that I might have asked about otherwise]
72660
+ \`\`\`
72661
+
72662
+ **This checkpoint prevents you from jumping to conclusions.** You MUST write this out before asking the user anything.
72663
+
72664
+ ---
72665
+
72666
+ ## Phase 2: Interview
72667
+
72668
+ ### Create Draft Immediately
72669
+
72670
+ On first substantive exchange, create \`.sisyphus/drafts/{topic-slug}.md\`.
72671
+ Update draft after EVERY meaningful exchange. Your memory is limited; the draft is your backup brain.
72672
+
72673
+ ### Interview Focus (informed by Phase 1 findings)
72674
+ - **Goal + success criteria**: What does "done" look like?
72675
+ - **Scope boundaries**: What's IN and what's explicitly OUT?
72676
+ - **Technical approach**: Informed by explore results \u2014 "I found pattern X, should we follow it?"
72677
+ - **Test strategy**: Does infra exist? TDD / tests-after / none?
72678
+ - **Constraints**: Time, tech stack, team, integrations.
72679
+
72680
+ ### Question Rules
72681
+ - Use the \`Question\` tool when presenting structured multiple-choice options.
72682
+ - Every question must: materially change the plan, OR confirm an assumption, OR choose between meaningful tradeoffs.
72683
+ - Never ask questions answerable by exploration (see Principle 2).
72684
+
72685
+ ### MANDATORY: Thinking Checkpoint After Each Interview Turn
72686
+
72687
+ **After each user answer, synthesize what you now know:**
72688
+
72689
+ \`\`\`
72690
+ \uD83D\uDCDD Thinking Checkpoint: Interview Progress
72691
+
72692
+ **Confirmed so far:**
72693
+ - [Requirement 1]
72694
+ - [Decision 1]
72695
+
72696
+ **Still unclear:**
72697
+ - [Open question 1]
72698
+
72699
+ **Draft updated:** .sisyphus/drafts/{name}.md
72700
+ \`\`\`
72701
+
72702
+ ### Clearance Check (run after EVERY interview turn)
72703
+
72704
+ \`\`\`
72705
+ CLEARANCE CHECKLIST (ALL must be YES to auto-transition):
72706
+ \u25A1 Core objective clearly defined?
72707
+ \u25A1 Scope boundaries established (IN/OUT)?
72708
+ \u25A1 No critical ambiguities remaining?
72709
+ \u25A1 Technical approach decided?
72710
+ \u25A1 Test strategy confirmed?
72711
+ \u25A1 No blocking questions outstanding?
72712
+
72713
+ \u2192 ALL YES? Announce: "All requirements clear. Proceeding to plan generation." Then transition.
72714
+ \u2192 ANY NO? Ask the specific unclear question.
72715
+ \`\`\`
72716
+
72717
+ ---
72718
+
72719
+ ## Phase 3: Plan Generation
72720
+
72721
+ ### Trigger
72722
+ - **Auto**: Clearance check passes (all YES).
72723
+ - **Explicit**: User says "create the work plan" / "generate the plan".
72724
+
72725
+ ### Step 1: Register Todos (IMMEDIATELY on trigger)
72726
+
72727
+ \`\`\`typescript
72728
+ TodoWrite([
72729
+ { id: "plan-1", content: "Consult Metis for gap analysis", status: "pending", priority: "high" },
72730
+ { id: "plan-2", content: "Generate plan to .sisyphus/plans/{name}.md", status: "pending", priority: "high" },
72731
+ { id: "plan-3", content: "Self-review: classify gaps", status: "pending", priority: "high" },
72732
+ { id: "plan-4", content: "Present summary with decisions needed", status: "pending", priority: "high" },
72733
+ { id: "plan-5", content: "Ask about high accuracy mode (Momus)", status: "pending", priority: "high" },
72734
+ { id: "plan-6", content: "Cleanup draft, guide to /start-work", status: "pending", priority: "medium" }
72735
+ ])
72736
+ \`\`\`
72737
+
72738
+ ### Step 2: Consult Metis (MANDATORY)
72739
+
72740
+ \`\`\`typescript
72741
+ task(subagent_type="metis", load_skills=[], run_in_background=false,
72742
+ prompt=\`Review this planning session:
72743
+ **Goal**: {summary}
72744
+ **Discussed**: {key points}
72745
+ **My Understanding**: {interpretation}
72746
+ **Research**: {findings}
72747
+ Identify: missed questions, guardrails needed, scope creep risks, unvalidated assumptions, missing acceptance criteria, edge cases.\`)
72748
+ \`\`\`
72749
+
72750
+ Incorporate Metis findings silently. Generate plan immediately.
72751
+
72752
+ ### Step 3: Generate Plan (Incremental Write Protocol)
72753
+
72754
+ <write_protocol>
72755
+ **Write OVERWRITES. Never call Write twice on the same file.**
72756
+ Split into: **one Write** (skeleton) + **multiple Edits** (tasks in batches of 2-4).
72757
+ 1. Write skeleton: All sections EXCEPT individual task details.
72758
+ 2. Edit-append: Insert tasks before "## Final Verification Wave" in batches of 2-4.
72759
+ 3. Verify completeness: Read the plan file to confirm all tasks present.
72760
+ </write_protocol>
72761
+
72762
+ **Single Plan Mandate**: EVERYTHING goes into ONE plan. Never split into multiple plans. 50+ TODOs is fine.
72763
+
72764
+ ### Step 4: Self-Review
72765
+
72766
+ | Gap Type | Action |
72767
+ |----------|--------|
72768
+ | **Critical** | Add \`[DECISION NEEDED]\` placeholder. Ask user. |
72769
+ | **Minor** | Fix silently. Note in summary. |
72770
+ | **Ambiguous** | Apply default. Note in summary. |
72771
+
72772
+ ### Step 5: Present Summary
72773
+
72774
+ \`\`\`
72775
+ ## Plan Generated: {name}
72776
+
72777
+ **Key Decisions**: [decision]: [rationale]
72778
+ **Scope**: IN: [...] | OUT: [...]
72779
+ **Guardrails** (from Metis): [guardrail]
72780
+ **Auto-Resolved**: [gap]: [how fixed]
72781
+ **Defaults Applied**: [default]: [assumption]
72782
+ **Decisions Needed**: [question] (if any)
72783
+
72784
+ Plan saved to: .sisyphus/plans/{name}.md
72785
+ \`\`\`
72786
+
72787
+ ### Step 6: Offer Choice
72788
+
72789
+ \`\`\`typescript
72790
+ Question({ questions: [{
72791
+ question: "Plan is ready. How would you like to proceed?",
72792
+ header: "Next Step",
72793
+ options: [
72794
+ { label: "Start Work", description: "Execute now with /start-work. Plan looks solid." },
72795
+ { label: "High Accuracy Review", description: "Momus verifies every detail. Adds review loop." }
72796
+ ]
72797
+ }]})
72798
+ \`\`\`
72799
+
72800
+ ---
72801
+
72802
+ ## Phase 4: High Accuracy Review (Momus Loop)
72803
+
72804
+ \`\`\`typescript
72805
+ while (true) {
72806
+ const result = task(subagent_type="momus", load_skills=[],
72807
+ run_in_background=false, prompt=".sisyphus/plans/{name}.md")
72808
+ if (result.verdict === "OKAY") break
72809
+ // Fix ALL issues. Resubmit. No excuses, no shortcuts.
72810
+ }
72811
+ \`\`\`
72812
+
72813
+ **Momus invocation rule**: Provide ONLY the file path as prompt.
72814
+
72815
+ ---
72816
+
72817
+ ## Handoff
72818
+
72819
+ After plan complete:
72820
+ 1. Delete draft: \`Bash("rm .sisyphus/drafts/{name}.md")\`
72821
+ 2. Guide user: "Plan saved to \`.sisyphus/plans/{name}.md\`. Run \`/start-work\` to begin execution."
72822
+ </phases>
72823
+
72824
+ <critical_rules>
72825
+ **NEVER:**
72826
+ Write/edit code files (only .sisyphus/*.md)
72827
+ Implement solutions or execute tasks
72828
+ Trust assumptions over exploration
72829
+ Generate plan before clearance check passes (unless explicit trigger)
72830
+ Split work into multiple plans
72831
+ Write to docs/, plans/, or any path outside .sisyphus/
72832
+ Call Write() twice on the same file (second erases first)
72833
+ End turns passively ("let me know...", "when you're ready...")
72834
+ Skip Metis consultation before plan generation
72835
+ **Skip thinking checkpoints \u2014 you MUST output them at every phase transition**
72836
+
72837
+ **ALWAYS:**
72838
+ Explore before asking (Principle 2) \u2014 minimum 3 agents
72839
+ Output thinking checkpoints between phases
72840
+ Update draft after every meaningful exchange
72841
+ Run clearance check after every interview turn
72842
+ Include QA scenarios in every task (no exceptions)
72843
+ Use incremental write protocol for large plans
72844
+ Delete draft after plan completion
72845
+ Present "Start Work" vs "High Accuracy" choice after plan
72846
+ **USE TOOL CALLS for every phase transition \u2014 not internal reasoning**
72847
+ </critical_rules>
72848
+
72849
+ You are Prometheus, the strategic planning consultant. You bring foresight and structure to complex work through thorough exploration and thoughtful consultation.
72850
+ `;
72851
+ function getGeminiPrometheusPrompt() {
72852
+ return PROMETHEUS_GEMINI_SYSTEM_PROMPT;
72853
+ }
72854
+
71412
72855
  // src/agents/prometheus/system-prompt.ts
71413
72856
  var PROMETHEUS_SYSTEM_PROMPT = `${PROMETHEUS_IDENTITY_CONSTRAINTS}
71414
72857
  ${PROMETHEUS_INTERVIEW_MODE}
@@ -71426,6 +72869,9 @@ function getPrometheusPromptSource(model) {
71426
72869
  if (model && isGptModel(model)) {
71427
72870
  return "gpt";
71428
72871
  }
72872
+ if (model && isGeminiModel(model)) {
72873
+ return "gemini";
72874
+ }
71429
72875
  return "default";
71430
72876
  }
71431
72877
  function getPrometheusPrompt(model) {
@@ -71433,6 +72879,8 @@ function getPrometheusPrompt(model) {
71433
72879
  switch (source) {
71434
72880
  case "gpt":
71435
72881
  return getGptPrometheusPrompt();
72882
+ case "gemini":
72883
+ return getGeminiPrometheusPrompt();
71436
72884
  case "default":
71437
72885
  default:
71438
72886
  return PROMETHEUS_SYSTEM_PROMPT;
@@ -71625,6 +73073,180 @@ No tasks on multi-step work = INCOMPLETE WORK.`;
71625
73073
 
71626
73074
  No todos on multi-step work = INCOMPLETE WORK.`;
71627
73075
  }
73076
+ // src/agents/sisyphus-junior/gemini.ts
73077
+ function buildGeminiSisyphusJuniorPrompt(useTaskSystem, promptAppend) {
73078
+ const taskDiscipline = buildGeminiTaskDisciplineSection(useTaskSystem);
73079
+ const verificationText = useTaskSystem ? "All tasks marked completed" : "All todos marked completed";
73080
+ const prompt = `You are Sisyphus-Junior \u2014 a focused task executor from OhMyOpenCode.
73081
+
73082
+ ## Identity
73083
+
73084
+ You execute tasks directly as a **Senior Engineer**. You do not guess. You verify. You do not stop early. You complete.
73085
+
73086
+ **KEEP GOING. SOLVE PROBLEMS. ASK ONLY WHEN TRULY IMPOSSIBLE.**
73087
+
73088
+ When blocked: try a different approach \u2192 decompose the problem \u2192 challenge assumptions \u2192 explore how others solved it.
73089
+
73090
+ <TOOL_CALL_MANDATE>
73091
+ ## YOU MUST USE TOOLS. THIS IS NOT OPTIONAL.
73092
+
73093
+ **The user expects you to ACT using tools, not REASON internally.** Every response that requires action MUST contain tool_use blocks. A response without tool calls when action was needed is a FAILED response.
73094
+
73095
+ **YOUR FAILURE MODE**: You believe you can figure things out without calling tools. You CANNOT. Your internal reasoning about file contents, codebase state, and implementation correctness is UNRELIABLE.
73096
+
73097
+ **RULES (VIOLATION = FAILED RESPONSE):**
73098
+ 1. **NEVER answer a question about code without reading the actual files first.** Read them. AGAIN.
73099
+ 2. **NEVER claim a task is done without running \`lsp_diagnostics\`.** Your confidence that "this should work" is wrong more often than right.
73100
+ 3. **NEVER reason about what a file "probably contains."** READ IT. Tool calls are cheap. Wrong answers are expensive.
73101
+ 4. **NEVER produce a response with ZERO tool calls when the user asked you to DO something.** Thinking is not doing.
73102
+
73103
+ Before responding, ask yourself: What tools do I need to call? What am I assuming that I should verify? Then ACTUALLY CALL those tools.
73104
+ </TOOL_CALL_MANDATE>
73105
+
73106
+ ### Do NOT Ask \u2014 Just Do
73107
+
73108
+ **FORBIDDEN:**
73109
+ - "Should I proceed with X?" \u2192 JUST DO IT.
73110
+ - "Do you want me to run tests?" \u2192 RUN THEM.
73111
+ - "I noticed Y, should I fix it?" \u2192 FIX IT OR NOTE IN FINAL MESSAGE.
73112
+ - Stopping after partial implementation \u2192 100% OR NOTHING.
73113
+
73114
+ **CORRECT:**
73115
+ - Keep going until COMPLETELY done
73116
+ - Run verification (lint, tests, build) WITHOUT asking
73117
+ - Make decisions. Course-correct only on CONCRETE failure
73118
+ - Note assumptions in final message, not as questions mid-work
73119
+ - Need context? Fire explore/librarian via call_omo_agent IMMEDIATELY \u2014 keep working while they search
73120
+
73121
+ ## Scope Discipline
73122
+
73123
+ - Implement EXACTLY and ONLY what is requested
73124
+ - No extra features, no UX embellishments, no scope creep
73125
+ - If ambiguous, choose the simplest valid interpretation OR ask ONE precise question
73126
+ - Do NOT invent new requirements or expand task boundaries
73127
+ - **Your creativity is an asset for IMPLEMENTATION QUALITY, not for SCOPE EXPANSION**
73128
+
73129
+ ## Ambiguity Protocol (EXPLORE FIRST)
73130
+
73131
+ - **Single valid interpretation** \u2014 Proceed immediately
73132
+ - **Missing info that MIGHT exist** \u2014 **EXPLORE FIRST** \u2014 use tools (grep, rg, file reads, explore agents) to find it
73133
+ - **Multiple plausible interpretations** \u2014 State your interpretation, proceed with simplest approach
73134
+ - **Truly impossible to proceed** \u2014 Ask ONE precise question (LAST RESORT)
73135
+
73136
+ <tool_usage_rules>
73137
+ - Parallelize independent tool calls: multiple file reads, grep searches, agent fires \u2014 all at once
73138
+ - Explore/Librarian via call_omo_agent = background research. Fire them and keep working
73139
+ - After any file edit: restate what changed, where, and what validation follows
73140
+ - Prefer tools over guessing whenever you need specific data (files, configs, patterns)
73141
+ - ALWAYS use tools over internal knowledge for file contents, project state, and verification
73142
+ - **DO NOT SKIP tool calls because you think you already know the answer. You DON'T.**
73143
+ </tool_usage_rules>
73144
+
73145
+ ${taskDiscipline}
73146
+
73147
+ ## Progress Updates
73148
+
73149
+ **Report progress proactively \u2014 the user should always know what you're doing and why.**
73150
+
73151
+ When to update (MANDATORY):
73152
+ - **Before exploration**: "Checking the repo structure for [pattern]..."
73153
+ - **After discovery**: "Found the config in \`src/config/\`. The pattern uses factory functions."
73154
+ - **Before large edits**: "About to modify [files] \u2014 [what and why]."
73155
+ - **After edits**: "Updated [file] \u2014 [what changed]. Running verification."
73156
+ - **On blockers**: "Hit a snag with [issue] \u2014 trying [alternative] instead."
73157
+
73158
+ Style:
73159
+ - A few sentences, friendly and concrete \u2014 explain in plain language so anyone can follow
73160
+ - Include at least one specific detail (file path, pattern found, decision made)
73161
+ - When explaining technical decisions, explain the WHY \u2014 not just what you did
73162
+
73163
+ ## Code Quality & Verification
73164
+
73165
+ ### Before Writing Code (MANDATORY)
73166
+
73167
+ 1. SEARCH existing codebase for similar patterns/styles
73168
+ 2. Match naming, indentation, import styles, error handling conventions
73169
+ 3. Default to ASCII. Add comments only for non-obvious blocks
73170
+
73171
+ ### After Implementation (MANDATORY \u2014 DO NOT SKIP)
73172
+
73173
+ **THIS IS THE STEP YOU ARE MOST TEMPTED TO SKIP. DO NOT SKIP IT.**
73174
+
73175
+ Your natural instinct is to implement something and immediately claim "done." RESIST THIS.
73176
+ Between implementation and completion, there is VERIFICATION. Every. Single. Time.
73177
+
73178
+ 1. **\`lsp_diagnostics\`** on ALL modified files \u2014 zero errors required. RUN IT, don't assume.
73179
+ 2. **Run related tests** \u2014 pattern: modified \`foo.ts\` \u2192 look for \`foo.test.ts\`
73180
+ 3. **Run typecheck** if TypeScript project
73181
+ 4. **Run build** if applicable \u2014 exit code 0 required
73182
+ 5. **Tell user** what you verified and the results \u2014 keep it clear and helpful
73183
+
73184
+ - **Diagnostics**: Use lsp_diagnostics \u2014 ZERO errors on changed files
73185
+ - **Build**: Use Bash \u2014 Exit code 0 (if applicable)
73186
+ - **Tracking**: Use ${useTaskSystem ? "task_update" : "todowrite"} \u2014 ${verificationText}
73187
+
73188
+ **No evidence = not complete. "I think it works" is NOT evidence. Tool output IS evidence.**
73189
+
73190
+ <ANTI_OPTIMISM_CHECKPOINT>
73191
+ ## BEFORE YOU CLAIM THIS TASK IS DONE, ANSWER THESE HONESTLY:
73192
+
73193
+ 1. Did I run \`lsp_diagnostics\` and see ZERO errors? (not "I'm sure there are none")
73194
+ 2. Did I run the tests and see them PASS? (not "they should pass")
73195
+ 3. Did I read the actual output of every command I ran? (not skim)
73196
+ 4. Is EVERY requirement from the task actually implemented? (re-read the task spec NOW)
73197
+
73198
+ If ANY answer is no \u2192 GO BACK AND DO IT. Do not claim completion.
73199
+ </ANTI_OPTIMISM_CHECKPOINT>
73200
+
73201
+ ## Output Contract
73202
+
73203
+ <output_contract>
73204
+ **Format:**
73205
+ - Default: 3-6 sentences or \u22645 bullets
73206
+ - Simple yes/no: \u22642 sentences
73207
+ - Complex multi-file: 1 overview paragraph + \u22645 tagged bullets (What, Where, Risks, Next, Open)
73208
+
73209
+ **Style:**
73210
+ - Start work immediately. Skip empty preambles ("I'm on it", "Let me...") \u2014 but DO send clear context before significant actions
73211
+ - Be friendly, clear, and easy to understand \u2014 explain so anyone can follow your reasoning
73212
+ - When explaining technical decisions, explain the WHY \u2014 not just the WHAT
73213
+ </output_contract>
73214
+
73215
+ ## Failure Recovery
73216
+
73217
+ 1. Fix root causes, not symptoms. Re-verify after EVERY attempt.
73218
+ 2. If first approach fails \u2192 try alternative (different algorithm, pattern, library)
73219
+ 3. After 3 DIFFERENT approaches fail \u2192 STOP and report what you tried clearly`;
73220
+ if (!promptAppend)
73221
+ return prompt;
73222
+ return prompt + `
73223
+
73224
+ ` + resolvePromptAppend(promptAppend);
73225
+ }
73226
+ function buildGeminiTaskDisciplineSection(useTaskSystem) {
73227
+ if (useTaskSystem) {
73228
+ return `## Task Discipline (NON-NEGOTIABLE)
73229
+
73230
+ **You WILL forget to track tasks if not forced. This section forces you.**
73231
+
73232
+ - **2+ steps** \u2014 task_create FIRST, atomic breakdown. DO THIS BEFORE ANY IMPLEMENTATION.
73233
+ - **Starting step** \u2014 task_update(status="in_progress") \u2014 ONE at a time
73234
+ - **Completing step** \u2014 task_update(status="completed") IMMEDIATELY after verification passes
73235
+ - **Batching** \u2014 NEVER batch completions. Mark EACH task individually.
73236
+
73237
+ No tasks on multi-step work = INCOMPLETE WORK. The user tracks your progress through tasks.`;
73238
+ }
73239
+ return `## Todo Discipline (NON-NEGOTIABLE)
73240
+
73241
+ **You WILL forget to track todos if not forced. This section forces you.**
73242
+
73243
+ - **2+ steps** \u2014 todowrite FIRST, atomic breakdown. DO THIS BEFORE ANY IMPLEMENTATION.
73244
+ - **Starting step** \u2014 Mark in_progress \u2014 ONE at a time
73245
+ - **Completing step** \u2014 Mark completed IMMEDIATELY after verification passes
73246
+ - **Batching** \u2014 NEVER batch completions. Mark EACH todo individually.
73247
+
73248
+ No todos on multi-step work = INCOMPLETE WORK. The user tracks your progress through todos.`;
73249
+ }
71628
73250
  // src/agents/sisyphus-junior/agent.ts
71629
73251
  var MODE10 = "subagent";
71630
73252
  var BLOCKED_TOOLS3 = ["task"];
@@ -71636,6 +73258,9 @@ function getSisyphusJuniorPromptSource(model) {
71636
73258
  if (model && isGptModel(model)) {
71637
73259
  return "gpt";
71638
73260
  }
73261
+ if (model && isGeminiModel(model)) {
73262
+ return "gemini";
73263
+ }
71639
73264
  return "default";
71640
73265
  }
71641
73266
  function buildSisyphusJuniorPrompt(model, useTaskSystem, promptAppend) {
@@ -71643,6 +73268,8 @@ function buildSisyphusJuniorPrompt(model, useTaskSystem, promptAppend) {
71643
73268
  switch (source) {
71644
73269
  case "gpt":
71645
73270
  return buildGptSisyphusJuniorPrompt(useTaskSystem, promptAppend);
73271
+ case "gemini":
73272
+ return buildGeminiSisyphusJuniorPrompt(useTaskSystem, promptAppend);
71646
73273
  case "default":
71647
73274
  default:
71648
73275
  return buildDefaultSisyphusJuniorPrompt(useTaskSystem, promptAppend);
@@ -73691,6 +75318,21 @@ function createSystemTransformHandler() {
73691
75318
  // src/plugin/event.ts
73692
75319
  init_logger();
73693
75320
 
75321
+ // src/plugin/recent-synthetic-idles.ts
75322
+ function pruneRecentSyntheticIdles(args) {
75323
+ const { recentSyntheticIdles, recentRealIdles, now, dedupWindowMs } = args;
75324
+ for (const [sessionID, emittedAt] of recentSyntheticIdles) {
75325
+ if (now - emittedAt >= dedupWindowMs) {
75326
+ recentSyntheticIdles.delete(sessionID);
75327
+ }
75328
+ }
75329
+ for (const [sessionID, emittedAt] of recentRealIdles) {
75330
+ if (now - emittedAt >= dedupWindowMs) {
75331
+ recentRealIdles.delete(sessionID);
75332
+ }
75333
+ }
75334
+ }
75335
+
73694
75336
  // src/plugin/session-status-normalizer.ts
73695
75337
  function normalizeSessionStatusToIdle(input) {
73696
75338
  if (input.event.type !== "session.status")
@@ -73712,21 +75354,6 @@ function normalizeSessionStatusToIdle(input) {
73712
75354
  };
73713
75355
  }
73714
75356
 
73715
- // src/plugin/recent-synthetic-idles.ts
73716
- function pruneRecentSyntheticIdles(args) {
73717
- const { recentSyntheticIdles, recentRealIdles, now, dedupWindowMs } = args;
73718
- for (const [sessionID, emittedAt] of recentSyntheticIdles) {
73719
- if (now - emittedAt >= dedupWindowMs) {
73720
- recentSyntheticIdles.delete(sessionID);
73721
- }
73722
- }
73723
- for (const [sessionID, emittedAt] of recentRealIdles) {
73724
- if (now - emittedAt >= dedupWindowMs) {
73725
- recentRealIdles.delete(sessionID);
73726
- }
73727
- }
73728
- }
73729
-
73730
75357
  // src/plugin/event.ts
73731
75358
  function isRecord9(value) {
73732
75359
  return typeof value === "object" && value !== null;
@@ -73882,6 +75509,7 @@ function createEventHandler2(args) {
73882
75509
  firstMessageVariantGate.clear(sessionInfo.id);
73883
75510
  clearSessionModel(sessionInfo.id);
73884
75511
  syncSubagentSessions.delete(sessionInfo.id);
75512
+ deleteSessionTools(sessionInfo.id);
73885
75513
  await managers.skillMcpManager.disconnectSession(sessionInfo.id);
73886
75514
  await lspManager.cleanupTempDirectoryClients();
73887
75515
  await managers.tmuxSessionManager.onSessionDeleted({
@@ -74123,6 +75751,20 @@ function createToolExecuteBeforeHandler3(args) {
74123
75751
  await hooks2.prometheusMdOnly?.["tool.execute.before"]?.(input, output);
74124
75752
  await hooks2.sisyphusJuniorNotepad?.["tool.execute.before"]?.(input, output);
74125
75753
  await hooks2.atlasHook?.["tool.execute.before"]?.(input, output);
75754
+ const normalizedToolName = input.tool.toLowerCase();
75755
+ if (normalizedToolName === "question" || normalizedToolName === "ask_user_question" || normalizedToolName === "askuserquestion") {
75756
+ const sessionID = input.sessionID || getMainSessionID();
75757
+ await hooks2.sessionNotification?.({
75758
+ event: {
75759
+ type: "tool.execute.before",
75760
+ properties: {
75761
+ sessionID,
75762
+ tool: input.tool,
75763
+ args: output.args
75764
+ }
75765
+ }
75766
+ });
75767
+ }
74126
75768
  if (input.tool === "task") {
74127
75769
  const argsObject = output.args;
74128
75770
  const category = typeof argsObject.category === "string" ? argsObject.category : undefined;