oh-my-opencode-gpt-slim 0.1.3 → 0.1.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli/index.js CHANGED
@@ -2145,7 +2145,7 @@ var package_default;
2145
2145
  var init_package = __esm(() => {
2146
2146
  package_default = {
2147
2147
  name: "oh-my-opencode-gpt-slim",
2148
- version: "0.1.3",
2148
+ version: "0.1.4",
2149
2149
  description: "GPT-optimized lean fork of oh-my-openagent \u2014 33 hooks removed, 5 tools removed, Sisyphus prompt rewritten based on OpenAI Codex prompt.md",
2150
2150
  main: "dist/index.js",
2151
2151
  types: "dist/index.d.ts",
@@ -2223,15 +2223,15 @@ var init_package = __esm(() => {
2223
2223
  typescript: "^5.7.3"
2224
2224
  },
2225
2225
  optionalDependencies: {
2226
- "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.3",
2227
- "oh-my-opencode-gpt-slim-darwin-x64": "0.1.3",
2228
- "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.3",
2229
- "oh-my-opencode-gpt-slim-linux-arm64": "0.1.3",
2230
- "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.3",
2231
- "oh-my-opencode-gpt-slim-linux-x64": "0.1.3",
2232
- "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.3",
2233
- "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.3",
2234
- "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.3"
2226
+ "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.4",
2227
+ "oh-my-opencode-gpt-slim-darwin-x64": "0.1.4",
2228
+ "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.4",
2229
+ "oh-my-opencode-gpt-slim-linux-arm64": "0.1.4",
2230
+ "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.4",
2231
+ "oh-my-opencode-gpt-slim-linux-x64": "0.1.4",
2232
+ "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.4",
2233
+ "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.4",
2234
+ "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.4"
2235
2235
  },
2236
2236
  overrides: {
2237
2237
  "@opencode-ai/sdk": "^1.2.17"
@@ -9686,6 +9686,7 @@ function createEventState() {
9686
9686
  currentAgent: null,
9687
9687
  currentModel: null,
9688
9688
  currentVariant: null,
9689
+ currentReasoningEffort: null,
9689
9690
  currentMessageRole: null,
9690
9691
  agentColorsByName: {},
9691
9692
  partTypesById: {},
@@ -9968,13 +9969,13 @@ var displayChars = {
9968
9969
 
9969
9970
  // src/cli/run/output-renderer.ts
9970
9971
  var import_picocolors6 = __toESM(require_picocolors(), 1);
9971
- function renderAgentHeader(agent, model, variant, agentColorsByName) {
9972
+ function renderAgentHeader(agent, model, variant, reasoningEffort, agentColorsByName) {
9972
9973
  if (!agent && !model)
9973
9974
  return;
9974
9975
  const agentLabel = agent ? import_picocolors6.default.bold(colorizeWithProfileColor(agent, agentColorsByName[agent])) : "";
9975
9976
  const modelBase = model ?? "";
9976
- const variantSuffix = variant ? ` (${variant})` : "";
9977
- const modelLabel = model ? import_picocolors6.default.dim(`${modelBase}${variantSuffix}`) : "";
9977
+ const modelDetails = formatModelDetails(variant, reasoningEffort);
9978
+ const modelLabel = model ? import_picocolors6.default.dim(`${modelBase}${modelDetails}`) : "";
9978
9979
  process.stdout.write(`
9979
9980
  `);
9980
9981
  if (modelLabel) {
@@ -9988,6 +9989,10 @@ function renderAgentHeader(agent, model, variant, agentColorsByName) {
9988
9989
  process.stdout.write(`
9989
9990
  `);
9990
9991
  }
9992
+ function formatModelDetails(variant, reasoningEffort) {
9993
+ const details = [variant, reasoningEffort ? `effort: ${reasoningEffort}` : null].filter(Boolean);
9994
+ return details.length > 0 ? ` (${details.join(", ")})` : "";
9995
+ }
9991
9996
  function openThinkBlock() {
9992
9997
  process.stdout.write(`
9993
9998
  ${import_picocolors6.default.dim("\u2503 Thinking:")} `);
@@ -10066,9 +10071,9 @@ function renderCompletionMetaLine(state, messageID) {
10066
10071
  const elapsedSec = startedAt ? ((Date.now() - startedAt) / 1000).toFixed(1) : "0.0";
10067
10072
  const agent = state.currentAgent ?? "assistant";
10068
10073
  const model = state.currentModel ?? "unknown-model";
10069
- const variant = state.currentVariant ? ` (${state.currentVariant})` : "";
10074
+ const modelDetails = formatModelDetails(state.currentVariant, state.currentReasoningEffort);
10070
10075
  process.stdout.write(import_picocolors7.default.dim(`
10071
- ${displayChars.treeEnd} ${agent} \xB7 ${model}${variant} \xB7 ${elapsedSec}s
10076
+ ${displayChars.treeEnd} ${agent} \xB7 ${model}${modelDetails} \xB7 ${elapsedSec}s
10072
10077
  `));
10073
10078
  state.completionMetaPrintedByMessageId[messageID] = true;
10074
10079
  }
@@ -10254,11 +10259,13 @@ function handleMessageUpdated(ctx, payload, state) {
10254
10259
  const agent = props?.info?.agent ?? null;
10255
10260
  const model = props?.info?.modelID ?? null;
10256
10261
  const variant = props?.info?.variant ?? null;
10257
- if (agent !== state.currentAgent || model !== state.currentModel || variant !== state.currentVariant) {
10262
+ const reasoningEffort = props?.info?.reasoningEffort ?? null;
10263
+ if (agent !== state.currentAgent || model !== state.currentModel || variant !== state.currentVariant || reasoningEffort !== state.currentReasoningEffort) {
10258
10264
  state.currentAgent = agent;
10259
10265
  state.currentModel = model;
10260
10266
  state.currentVariant = variant;
10261
- renderAgentHeader(agent, model, variant, state.agentColorsByName);
10267
+ state.currentReasoningEffort = reasoningEffort;
10268
+ renderAgentHeader(agent, model, variant, reasoningEffort, state.agentColorsByName);
10262
10269
  }
10263
10270
  }
10264
10271
  function handleToolExecute(ctx, payload, state) {
@@ -17,6 +17,8 @@ export interface EventState {
17
17
  currentModel: string | null;
18
18
  /** Current model variant from the latest assistant message */
19
19
  currentVariant: string | null;
20
+ /** Current model reasoning effort from the latest assistant message */
21
+ currentReasoningEffort: string | null;
20
22
  /** Current message role (user/assistant) — used to filter user messages from display */
21
23
  currentMessageRole: string | null;
22
24
  /** Agent profile colors keyed by display name */
@@ -1,4 +1,5 @@
1
- export declare function renderAgentHeader(agent: string | null, model: string | null, variant: string | null, agentColorsByName: Record<string, string>): void;
1
+ export declare function renderAgentHeader(agent: string | null, model: string | null, variant: string | null, reasoningEffort: string | null, agentColorsByName: Record<string, string>): void;
2
+ export declare function formatModelDetails(variant: string | null, reasoningEffort: string | null): string;
2
3
  export declare function openThinkBlock(): void;
3
4
  export declare function closeThinkBlock(): void;
4
5
  export declare function writePaddedText(text: string, atLineStart: boolean): {
@@ -68,6 +68,7 @@ export interface MessageUpdatedProps {
68
68
  providerID?: string;
69
69
  agent?: string;
70
70
  variant?: string;
71
+ reasoningEffort?: string;
71
72
  };
72
73
  }
73
74
  export interface MessagePartUpdatedProps {
package/dist/index.js CHANGED
@@ -12308,7 +12308,7 @@ function initConfigContext(binary, version) {
12308
12308
  // package.json
12309
12309
  var package_default = {
12310
12310
  name: "oh-my-opencode-gpt-slim",
12311
- version: "0.1.3",
12311
+ version: "0.1.4",
12312
12312
  description: "GPT-optimized lean fork of oh-my-openagent \u2014 33 hooks removed, 5 tools removed, Sisyphus prompt rewritten based on OpenAI Codex prompt.md",
12313
12313
  main: "dist/index.js",
12314
12314
  types: "dist/index.d.ts",
@@ -12386,15 +12386,15 @@ var package_default = {
12386
12386
  typescript: "^5.7.3"
12387
12387
  },
12388
12388
  optionalDependencies: {
12389
- "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.3",
12390
- "oh-my-opencode-gpt-slim-darwin-x64": "0.1.3",
12391
- "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.3",
12392
- "oh-my-opencode-gpt-slim-linux-arm64": "0.1.3",
12393
- "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.3",
12394
- "oh-my-opencode-gpt-slim-linux-x64": "0.1.3",
12395
- "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.3",
12396
- "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.3",
12397
- "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.3"
12389
+ "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.4",
12390
+ "oh-my-opencode-gpt-slim-darwin-x64": "0.1.4",
12391
+ "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.4",
12392
+ "oh-my-opencode-gpt-slim-linux-arm64": "0.1.4",
12393
+ "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.4",
12394
+ "oh-my-opencode-gpt-slim-linux-x64": "0.1.4",
12395
+ "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.4",
12396
+ "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.4",
12397
+ "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.4"
12398
12398
  },
12399
12399
  overrides: {
12400
12400
  "@opencode-ai/sdk": "^1.2.17"
@@ -17144,6 +17144,32 @@ var CATEGORY_MODEL_REQUIREMENTS = {
17144
17144
  ]
17145
17145
  }
17146
17146
  };
17147
+
17148
+ // src/shared/agent-variant.ts
17149
+ function resolveAgentReasoningEffort(config, agentName) {
17150
+ if (!agentName) {
17151
+ return;
17152
+ }
17153
+ const agentOverrides = config.agents;
17154
+ const agentOverride = agentOverrides ? agentOverrides[agentName] ?? Object.entries(agentOverrides).find(([key]) => key.toLowerCase() === agentName.toLowerCase())?.[1] : undefined;
17155
+ if (!agentOverride) {
17156
+ return;
17157
+ }
17158
+ if (agentOverride.reasoningEffort) {
17159
+ return agentOverride.reasoningEffort;
17160
+ }
17161
+ const categoryName = agentOverride.category;
17162
+ if (!categoryName) {
17163
+ return;
17164
+ }
17165
+ return config.categories?.[categoryName]?.reasoningEffort;
17166
+ }
17167
+ function applyAgentReasoningEffort(config, agentName, message) {
17168
+ const reasoningEffort = resolveAgentReasoningEffort(config, agentName);
17169
+ if (reasoningEffort !== undefined && message.reasoningEffort === undefined) {
17170
+ message.reasoningEffort = reasoningEffort;
17171
+ }
17172
+ }
17147
17173
  // src/shared/session-cursor.ts
17148
17174
  var sessionCursors = new Map;
17149
17175
  function buildMessageKey(message, index) {
@@ -79103,6 +79129,7 @@ function createChatMessageHandler2(args) {
79103
79129
  if (firstMessageVariantGate.shouldOverride(input.sessionID)) {
79104
79130
  firstMessageVariantGate.markApplied(input.sessionID);
79105
79131
  }
79132
+ applyAgentReasoningEffort(pluginConfig, input.agent, output.message);
79106
79133
  if (!isRuntimeFallbackEnabled) {
79107
79134
  await hooks2.modelFallback?.["chat.message"]?.(input, output);
79108
79135
  }
@@ -1,5 +1,6 @@
1
1
  import type { OhMyOpenCodeConfig } from "../config";
2
2
  export declare function resolveAgentVariant(config: OhMyOpenCodeConfig, agentName?: string): string | undefined;
3
+ export declare function resolveAgentReasoningEffort(config: OhMyOpenCodeConfig, agentName?: string): string | undefined;
3
4
  export declare function resolveVariantForModel(config: OhMyOpenCodeConfig, agentName: string, currentModel: {
4
5
  providerID: string;
5
6
  modelID: string;
@@ -7,3 +8,6 @@ export declare function resolveVariantForModel(config: OhMyOpenCodeConfig, agent
7
8
  export declare function applyAgentVariant(config: OhMyOpenCodeConfig, agentName: string | undefined, message: {
8
9
  variant?: string;
9
10
  }): void;
11
+ export declare function applyAgentReasoningEffort(config: OhMyOpenCodeConfig, agentName: string | undefined, message: {
12
+ reasoningEffort?: string;
13
+ }): void;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "oh-my-opencode-gpt-slim",
3
- "version": "0.1.3",
3
+ "version": "0.1.4",
4
4
  "description": "GPT-optimized lean fork of oh-my-openagent — 33 hooks removed, 5 tools removed, Sisyphus prompt rewritten based on OpenAI Codex prompt.md",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
@@ -78,15 +78,15 @@
78
78
  "typescript": "^5.7.3"
79
79
  },
80
80
  "optionalDependencies": {
81
- "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.3",
82
- "oh-my-opencode-gpt-slim-darwin-x64": "0.1.3",
83
- "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.3",
84
- "oh-my-opencode-gpt-slim-linux-arm64": "0.1.3",
85
- "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.3",
86
- "oh-my-opencode-gpt-slim-linux-x64": "0.1.3",
87
- "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.3",
88
- "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.3",
89
- "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.3"
81
+ "oh-my-opencode-gpt-slim-darwin-arm64": "0.1.4",
82
+ "oh-my-opencode-gpt-slim-darwin-x64": "0.1.4",
83
+ "oh-my-opencode-gpt-slim-darwin-x64-baseline": "0.1.4",
84
+ "oh-my-opencode-gpt-slim-linux-arm64": "0.1.4",
85
+ "oh-my-opencode-gpt-slim-linux-arm64-musl": "0.1.4",
86
+ "oh-my-opencode-gpt-slim-linux-x64": "0.1.4",
87
+ "oh-my-opencode-gpt-slim-linux-x64-baseline": "0.1.4",
88
+ "oh-my-opencode-gpt-slim-linux-x64-musl": "0.1.4",
89
+ "oh-my-opencode-gpt-slim-linux-x64-musl-baseline": "0.1.4"
90
90
  },
91
91
  "overrides": {
92
92
  "@opencode-ai/sdk": "^1.2.17"