replicas-engine 0.1.120 → 0.1.121

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/src/index.js CHANGED
@@ -1151,7 +1151,7 @@ function parseReplicasConfigString(content, filename) {
1151
1151
  }
1152
1152
 
1153
1153
  // ../shared/src/engine/environment.ts
1154
- var DAYTONA_SNAPSHOT_ID = "23-04-2026-islington-v1";
1154
+ var DAYTONA_SNAPSHOT_ID = "23-04-2026-islington-v2";
1155
1155
 
1156
1156
  // ../shared/src/engine/types.ts
1157
1157
  var DEFAULT_CHAT_TITLES = {
@@ -2793,7 +2793,7 @@ import { existsSync as existsSync6 } from "fs";
2793
2793
  import { join as join11 } from "path";
2794
2794
  import { homedir as homedir8 } from "os";
2795
2795
  import { parse as parseToml, stringify as stringifyToml } from "smol-toml";
2796
- var DEFAULT_MODEL = "gpt-5.4";
2796
+ var DEFAULT_MODEL = "gpt-5.5";
2797
2797
  var CODEX_CONFIG_PATH = join11(homedir8(), ".codex", "config.toml");
2798
2798
  function isLinearThoughtEvent2(event) {
2799
2799
  return event.content.type === "thought";
@@ -3235,7 +3235,7 @@ You will also receive the chatId so you can send follow-up messages or clean up
3235
3235
  {
3236
3236
  provider: providerEnum.describe(providerDesc),
3237
3237
  prompt: z.string().describe("The full prompt/instructions for the subagent. Be detailed - it has no context from your conversation."),
3238
- model: z.string().optional().describe(codexAvailable ? "Model override. Claude: opus, sonnet, haiku. Codex: gpt-5.4, gpt-5.3-codex, etc." : "Model override. Claude: opus, sonnet, haiku."),
3238
+ model: z.string().optional().describe(codexAvailable ? "Model override. Claude: opus, sonnet, haiku. Codex: gpt-5.5, gpt-5.4, gpt-5.3-codex, etc." : "Model override. Claude: opus, sonnet, haiku."),
3239
3239
  thinking_level: z.enum(["low", "medium", "high", "max"]).optional().describe(
3240
3240
  "Controls how much thinking/reasoning the subagent applies. low = light thinking, medium = moderate, high = deep reasoning, max = maximum effort. Defaults: Claude = high, Codex = medium."
3241
3241
  ),
@@ -3472,7 +3472,7 @@ function getUsingToolsSection() {
3472
3472
  function getDelegationSection(codexAvailable) {
3473
3473
  const providerList = codexAvailable ? "claude, codex, or relay" : "claude or relay";
3474
3474
  const spawnDesc = `Create a new subagent with a specific provider (${providerList}), send it a prompt, and wait for its response. Returns the chatId and the agent's final response. You can set a custom timeout via the timeout_minutes parameter (default: 10 minutes).`;
3475
- const agentSelectionLines = codexAvailable ? `Use provider 'codex' for heavy code writing, implementation, and large refactors. Suggested models: gpt-5.4 (default), gpt-5.3-codex.
3475
+ const agentSelectionLines = codexAvailable ? `Use provider 'codex' for heavy code writing, implementation, and large refactors. Suggested models: gpt-5.5 (default), gpt-5.4, gpt-5.3-codex.
3476
3476
 
3477
3477
  Use provider 'claude' for codebase exploration, code review, planning, complex debugging, and tasks requiring nuanced architectural understanding. Suggested models: opus (default), sonnet (faster).` : `Use provider 'claude' for all tasks including code writing, codebase exploration, code review, planning, complex debugging, and tasks requiring nuanced architectural understanding. Suggested models: opus (default), sonnet (faster).`;
3478
3478
  return `# Delegation
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "replicas-engine",
3
- "version": "0.1.120",
3
+ "version": "0.1.121",
4
4
  "description": "Lightweight API server for Replicas workspaces",
5
5
  "type": "module",
6
6
  "main": "dist/src/index.js",
@@ -29,7 +29,7 @@
29
29
  "dependencies": {
30
30
  "@anthropic-ai/claude-agent-sdk": "0.2.112",
31
31
  "@hono/node-server": "^1.19.5",
32
- "@openai/codex-sdk": "^0.111.0",
32
+ "@openai/codex-sdk": "^0.124.0",
33
33
  "hono": "^4.10.3",
34
34
  "smol-toml": "^1.6.0",
35
35
  "yaml": "^2.8.2",