@roo-code/types 1.108.0 → 1.109.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -202,6 +202,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
202
202
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
203
203
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
204
204
  RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
205
+ RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
205
206
  RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
206
207
  RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
207
208
  RooCodeEventName2["ModeChanged"] = "modeChanged";
@@ -261,6 +262,7 @@ var rooCodeEventsSchema = z3.object({
261
262
  ["taskModeSwitched" /* TaskModeSwitched */]: z3.tuple([z3.string(), z3.string()]),
262
263
  ["taskAskResponded" /* TaskAskResponded */]: z3.tuple([z3.string()]),
263
264
  ["taskUserMessage" /* TaskUserMessage */]: z3.tuple([z3.string()]),
265
+ ["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: z3.tuple([z3.string(), z3.array(queuedMessageSchema)]),
264
266
  ["taskToolFailed" /* TaskToolFailed */]: z3.tuple([z3.string(), toolNamesSchema, z3.string()]),
265
267
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z3.tuple([z3.string(), tokenUsageSchema, toolUsageSchema]),
266
268
  ["modeChanged" /* ModeChanged */]: z3.tuple([z3.string()]),
@@ -366,6 +368,11 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
366
368
  payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
367
369
  taskId: z3.number().optional()
368
370
  }),
371
+ z3.object({
372
+ eventName: z3.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
373
+ payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
374
+ taskId: z3.number().optional()
375
+ }),
369
376
  // Task Analytics
370
377
  z3.object({
371
378
  eventName: z3.literal("taskToolFailed" /* TaskToolFailed */),
@@ -619,6 +626,38 @@ var anthropicModels = {
619
626
  }
620
627
  ]
621
628
  },
629
+ "claude-opus-4-6": {
630
+ maxTokens: 128e3,
631
+ // Overridden to 8k if `enableReasoningEffort` is false.
632
+ contextWindow: 2e5,
633
+ // Default 200K, extendable to 1M with beta flag
634
+ supportsImages: true,
635
+ supportsPromptCache: true,
636
+ inputPrice: 5,
637
+ // $5 per million input tokens (≤200K context)
638
+ outputPrice: 25,
639
+ // $25 per million output tokens (≤200K context)
640
+ cacheWritesPrice: 6.25,
641
+ // $6.25 per million tokens
642
+ cacheReadsPrice: 0.5,
643
+ // $0.50 per million tokens
644
+ supportsReasoningBudget: true,
645
+ // Tiered pricing for extended context (requires beta flag)
646
+ tiers: [
647
+ {
648
+ contextWindow: 1e6,
649
+ // 1M tokens with beta flag
650
+ inputPrice: 10,
651
+ // $10 per million input tokens (>200K context)
652
+ outputPrice: 37.5,
653
+ // $37.50 per million output tokens (>200K context)
654
+ cacheWritesPrice: 12.5,
655
+ // $12.50 per million tokens (>200K context)
656
+ cacheReadsPrice: 1
657
+ // $1.00 per million tokens (>200K context)
658
+ }
659
+ ]
660
+ },
622
661
  "claude-opus-4-5-20251101": {
623
662
  maxTokens: 32e3,
624
663
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1003,6 +1042,40 @@ var bedrockModels = {
1003
1042
  maxCachePoints: 4,
1004
1043
  cachableFields: ["system", "messages", "tools"]
1005
1044
  },
1045
+ "anthropic.claude-opus-4-6-v1": {
1046
+ maxTokens: 8192,
1047
+ contextWindow: 2e5,
1048
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
1049
+ supportsImages: true,
1050
+ supportsPromptCache: true,
1051
+ supportsReasoningBudget: true,
1052
+ inputPrice: 5,
1053
+ // $5 per million input tokens (≤200K context)
1054
+ outputPrice: 25,
1055
+ // $25 per million output tokens (≤200K context)
1056
+ cacheWritesPrice: 6.25,
1057
+ // $6.25 per million tokens
1058
+ cacheReadsPrice: 0.5,
1059
+ // $0.50 per million tokens
1060
+ minTokensPerCachePoint: 1024,
1061
+ maxCachePoints: 4,
1062
+ cachableFields: ["system", "messages", "tools"],
1063
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
1064
+ tiers: [
1065
+ {
1066
+ contextWindow: 1e6,
1067
+ // 1M tokens with beta flag
1068
+ inputPrice: 10,
1069
+ // $10 per million input tokens (>200K context)
1070
+ outputPrice: 37.5,
1071
+ // $37.50 per million output tokens (>200K context)
1072
+ cacheWritesPrice: 12.5,
1073
+ // $12.50 per million tokens (>200K context)
1074
+ cacheReadsPrice: 1
1075
+ // $1.00 per million tokens (>200K context)
1076
+ }
1077
+ ]
1078
+ },
1006
1079
  "anthropic.claude-opus-4-5-20251101-v1:0": {
1007
1080
  maxTokens: 8192,
1008
1081
  contextWindow: 2e5,
@@ -1349,13 +1422,15 @@ var BEDROCK_REGIONS = [
1349
1422
  ].sort((a, b) => a.value.localeCompare(b.value));
1350
1423
  var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1351
1424
  "anthropic.claude-sonnet-4-20250514-v1:0",
1352
- "anthropic.claude-sonnet-4-5-20250929-v1:0"
1425
+ "anthropic.claude-sonnet-4-5-20250929-v1:0",
1426
+ "anthropic.claude-opus-4-6-v1"
1353
1427
  ];
1354
1428
  var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1355
1429
  "anthropic.claude-sonnet-4-20250514-v1:0",
1356
1430
  "anthropic.claude-sonnet-4-5-20250929-v1:0",
1357
1431
  "anthropic.claude-haiku-4-5-20251001-v1:0",
1358
- "anthropic.claude-opus-4-5-20251101-v1:0"
1432
+ "anthropic.claude-opus-4-5-20251101-v1:0",
1433
+ "anthropic.claude-opus-4-6-v1"
1359
1434
  ];
1360
1435
  var BEDROCK_SERVICE_TIER_MODEL_IDS = [
1361
1436
  // Amazon Nova models
@@ -1984,6 +2059,16 @@ var fireworksModels = {
1984
2059
  cacheReadsPrice: 0.15,
1985
2060
  description: "The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems."
1986
2061
  },
2062
+ "accounts/fireworks/models/kimi-k2p5": {
2063
+ maxTokens: 16384,
2064
+ contextWindow: 262144,
2065
+ supportsImages: true,
2066
+ supportsPromptCache: true,
2067
+ inputPrice: 0.6,
2068
+ outputPrice: 3,
2069
+ cacheReadsPrice: 0.1,
2070
+ description: "Kimi K2.5 is Moonshot AI's flagship agentic model and a new SOTA open model. It unifies vision and text, thinking and non-thinking modes, and single-agent and multi-agent execution into one model. Fireworks enables users to control the reasoning behavior and inspect its reasoning history for greater transparency."
2071
+ },
1987
2072
  "accounts/fireworks/models/minimax-m2": {
1988
2073
  maxTokens: 4096,
1989
2074
  contextWindow: 204800,
@@ -3175,7 +3260,7 @@ var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3175
3260
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3176
3261
 
3177
3262
  // src/providers/openai-codex.ts
3178
- var openAiCodexDefaultModelId = "gpt-5.2-codex";
3263
+ var openAiCodexDefaultModelId = "gpt-5.3-codex";
3179
3264
  var openAiCodexModels = {
3180
3265
  "gpt-5.1-codex-max": {
3181
3266
  maxTokens: 128e3,
@@ -3207,6 +3292,20 @@ var openAiCodexModels = {
3207
3292
  supportsTemperature: false,
3208
3293
  description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3209
3294
  },
3295
+ "gpt-5.3-codex": {
3296
+ maxTokens: 128e3,
3297
+ contextWindow: 4e5,
3298
+ includedTools: ["apply_patch"],
3299
+ excludedTools: ["apply_diff", "write_to_file"],
3300
+ supportsImages: true,
3301
+ supportsPromptCache: true,
3302
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3303
+ reasoningEffort: "medium",
3304
+ inputPrice: 0,
3305
+ outputPrice: 0,
3306
+ supportsTemperature: false,
3307
+ description: "GPT-5.3 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3308
+ },
3210
3309
  "gpt-5.2-codex": {
3211
3310
  maxTokens: 128e3,
3212
3311
  contextWindow: 4e5,
@@ -3349,8 +3448,9 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3349
3448
  "anthropic/claude-sonnet-4.5",
3350
3449
  "anthropic/claude-opus-4",
3351
3450
  "anthropic/claude-opus-4.1",
3352
- "anthropic/claude-haiku-4.5",
3353
3451
  "anthropic/claude-opus-4.5",
3452
+ "anthropic/claude-opus-4.6",
3453
+ "anthropic/claude-haiku-4.5",
3354
3454
  "google/gemini-2.5-flash-preview",
3355
3455
  "google/gemini-2.5-flash-preview:thinking",
3356
3456
  "google/gemini-2.5-flash-preview-05-20",
@@ -3370,9 +3470,10 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3370
3470
  "anthropic/claude-3.7-sonnet:beta",
3371
3471
  "anthropic/claude-opus-4",
3372
3472
  "anthropic/claude-opus-4.1",
3473
+ "anthropic/claude-opus-4.5",
3474
+ "anthropic/claude-opus-4.6",
3373
3475
  "anthropic/claude-sonnet-4",
3374
3476
  "anthropic/claude-sonnet-4.5",
3375
- "anthropic/claude-opus-4.5",
3376
3477
  "anthropic/claude-haiku-4.5",
3377
3478
  "google/gemini-2.5-pro-preview",
3378
3479
  "google/gemini-2.5-pro",
@@ -3830,6 +3931,37 @@ var vertexModels = {
3830
3931
  cacheReadsPrice: 0.1,
3831
3932
  supportsReasoningBudget: true
3832
3933
  },
3934
+ "claude-opus-4-6": {
3935
+ maxTokens: 8192,
3936
+ contextWindow: 2e5,
3937
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
3938
+ supportsImages: true,
3939
+ supportsPromptCache: true,
3940
+ inputPrice: 5,
3941
+ // $5 per million input tokens (≤200K context)
3942
+ outputPrice: 25,
3943
+ // $25 per million output tokens (≤200K context)
3944
+ cacheWritesPrice: 6.25,
3945
+ // $6.25 per million tokens
3946
+ cacheReadsPrice: 0.5,
3947
+ // $0.50 per million tokens
3948
+ supportsReasoningBudget: true,
3949
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
3950
+ tiers: [
3951
+ {
3952
+ contextWindow: 1e6,
3953
+ // 1M tokens with beta flag
3954
+ inputPrice: 10,
3955
+ // $10 per million input tokens (>200K context)
3956
+ outputPrice: 37.5,
3957
+ // $37.50 per million output tokens (>200K context)
3958
+ cacheWritesPrice: 12.5,
3959
+ // $12.50 per million tokens (>200K context)
3960
+ cacheReadsPrice: 1
3961
+ // $1.00 per million tokens (>200K context)
3962
+ }
3963
+ ]
3964
+ },
3833
3965
  "claude-opus-4-5@20251101": {
3834
3966
  maxTokens: 8192,
3835
3967
  contextWindow: 2e5,
@@ -4019,7 +4151,11 @@ var vertexModels = {
4019
4151
  description: "Kimi K2 Thinking Model with 256K context window."
4020
4152
  }
4021
4153
  };
4022
- var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
4154
+ var VERTEX_1M_CONTEXT_MODEL_IDS = [
4155
+ "claude-sonnet-4@20250514",
4156
+ "claude-sonnet-4-5@20250929",
4157
+ "claude-opus-4-6"
4158
+ ];
4023
4159
  var VERTEX_REGIONS = [
4024
4160
  { value: "global", label: "global" },
4025
4161
  { value: "us-central1", label: "us-central1" },
@@ -4267,6 +4403,8 @@ var xaiModels = {
4267
4403
  cacheWritesPrice: 0.05,
4268
4404
  cacheReadsPrice: 0.05,
4269
4405
  description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning",
4406
+ supportsReasoningEffort: ["low", "high"],
4407
+ reasoningEffort: "low",
4270
4408
  includedTools: ["search_replace"],
4271
4409
  excludedTools: ["apply_diff"]
4272
4410
  },
@@ -4293,6 +4431,8 @@ var xaiModels = {
4293
4431
  cacheWritesPrice: 0.05,
4294
4432
  cacheReadsPrice: 0.05,
4295
4433
  description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning",
4434
+ supportsReasoningEffort: ["low", "high"],
4435
+ reasoningEffort: "low",
4296
4436
  includedTools: ["search_replace"],
4297
4437
  excludedTools: ["apply_diff"]
4298
4438
  },
@@ -4362,6 +4502,8 @@ var VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
4362
4502
  "anthropic/claude-3.7-sonnet",
4363
4503
  "anthropic/claude-opus-4",
4364
4504
  "anthropic/claude-opus-4.1",
4505
+ "anthropic/claude-opus-4.5",
4506
+ "anthropic/claude-opus-4.6",
4365
4507
  "anthropic/claude-sonnet-4",
4366
4508
  "openai/gpt-4.1",
4367
4509
  "openai/gpt-4.1-mini",
@@ -4399,6 +4541,8 @@ var VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS = /* @__PURE__ */ new Set([
4399
4541
  "anthropic/claude-3.7-sonnet",
4400
4542
  "anthropic/claude-opus-4",
4401
4543
  "anthropic/claude-opus-4.1",
4544
+ "anthropic/claude-opus-4.5",
4545
+ "anthropic/claude-opus-4.6",
4402
4546
  "anthropic/claude-sonnet-4",
4403
4547
  "google/gemini-1.5-flash",
4404
4548
  "google/gemini-1.5-pro",
@@ -6742,12 +6886,10 @@ var taskCommandSchema = z19.discriminatedUnion("commandName", [
6742
6886
  })
6743
6887
  }),
6744
6888
  z19.object({
6745
- commandName: z19.literal("CancelTask" /* CancelTask */),
6746
- data: z19.string()
6889
+ commandName: z19.literal("CancelTask" /* CancelTask */)
6747
6890
  }),
6748
6891
  z19.object({
6749
- commandName: z19.literal("CloseTask" /* CloseTask */),
6750
- data: z19.string()
6892
+ commandName: z19.literal("CloseTask" /* CloseTask */)
6751
6893
  }),
6752
6894
  z19.object({
6753
6895
  commandName: z19.literal("ResumeTask" /* ResumeTask */),