@roo-code/types 1.107.0 → 1.109.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -626,6 +626,38 @@ var anthropicModels = {
626
626
  }
627
627
  ]
628
628
  },
629
+ "claude-opus-4-6": {
630
+ maxTokens: 128e3,
631
+ // Overridden to 8k if `enableReasoningEffort` is false.
632
+ contextWindow: 2e5,
633
+ // Default 200K, extendable to 1M with beta flag
634
+ supportsImages: true,
635
+ supportsPromptCache: true,
636
+ inputPrice: 5,
637
+ // $5 per million input tokens (≤200K context)
638
+ outputPrice: 25,
639
+ // $25 per million output tokens (≤200K context)
640
+ cacheWritesPrice: 6.25,
641
+ // $6.25 per million tokens
642
+ cacheReadsPrice: 0.5,
643
+ // $0.50 per million tokens
644
+ supportsReasoningBudget: true,
645
+ // Tiered pricing for extended context (requires beta flag)
646
+ tiers: [
647
+ {
648
+ contextWindow: 1e6,
649
+ // 1M tokens with beta flag
650
+ inputPrice: 10,
651
+ // $10 per million input tokens (>200K context)
652
+ outputPrice: 37.5,
653
+ // $37.50 per million output tokens (>200K context)
654
+ cacheWritesPrice: 12.5,
655
+ // $12.50 per million tokens (>200K context)
656
+ cacheReadsPrice: 1
657
+ // $1.00 per million tokens (>200K context)
658
+ }
659
+ ]
660
+ },
629
661
  "claude-opus-4-5-20251101": {
630
662
  maxTokens: 32e3,
631
663
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1010,6 +1042,40 @@ var bedrockModels = {
1010
1042
  maxCachePoints: 4,
1011
1043
  cachableFields: ["system", "messages", "tools"]
1012
1044
  },
1045
+ "anthropic.claude-opus-4-6-v1": {
1046
+ maxTokens: 8192,
1047
+ contextWindow: 2e5,
1048
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
1049
+ supportsImages: true,
1050
+ supportsPromptCache: true,
1051
+ supportsReasoningBudget: true,
1052
+ inputPrice: 5,
1053
+ // $5 per million input tokens (≤200K context)
1054
+ outputPrice: 25,
1055
+ // $25 per million output tokens (≤200K context)
1056
+ cacheWritesPrice: 6.25,
1057
+ // $6.25 per million tokens
1058
+ cacheReadsPrice: 0.5,
1059
+ // $0.50 per million tokens
1060
+ minTokensPerCachePoint: 1024,
1061
+ maxCachePoints: 4,
1062
+ cachableFields: ["system", "messages", "tools"],
1063
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
1064
+ tiers: [
1065
+ {
1066
+ contextWindow: 1e6,
1067
+ // 1M tokens with beta flag
1068
+ inputPrice: 10,
1069
+ // $10 per million input tokens (>200K context)
1070
+ outputPrice: 37.5,
1071
+ // $37.50 per million output tokens (>200K context)
1072
+ cacheWritesPrice: 12.5,
1073
+ // $12.50 per million tokens (>200K context)
1074
+ cacheReadsPrice: 1
1075
+ // $1.00 per million tokens (>200K context)
1076
+ }
1077
+ ]
1078
+ },
1013
1079
  "anthropic.claude-opus-4-5-20251101-v1:0": {
1014
1080
  maxTokens: 8192,
1015
1081
  contextWindow: 2e5,
@@ -1356,13 +1422,15 @@ var BEDROCK_REGIONS = [
1356
1422
  ].sort((a, b) => a.value.localeCompare(b.value));
1357
1423
  var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1358
1424
  "anthropic.claude-sonnet-4-20250514-v1:0",
1359
- "anthropic.claude-sonnet-4-5-20250929-v1:0"
1425
+ "anthropic.claude-sonnet-4-5-20250929-v1:0",
1426
+ "anthropic.claude-opus-4-6-v1"
1360
1427
  ];
1361
1428
  var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1362
1429
  "anthropic.claude-sonnet-4-20250514-v1:0",
1363
1430
  "anthropic.claude-sonnet-4-5-20250929-v1:0",
1364
1431
  "anthropic.claude-haiku-4-5-20251001-v1:0",
1365
- "anthropic.claude-opus-4-5-20251101-v1:0"
1432
+ "anthropic.claude-opus-4-5-20251101-v1:0",
1433
+ "anthropic.claude-opus-4-6-v1"
1366
1434
  ];
1367
1435
  var BEDROCK_SERVICE_TIER_MODEL_IDS = [
1368
1436
  // Amazon Nova models
@@ -1991,6 +2059,16 @@ var fireworksModels = {
1991
2059
  cacheReadsPrice: 0.15,
1992
2060
  description: "The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems."
1993
2061
  },
2062
+ "accounts/fireworks/models/kimi-k2p5": {
2063
+ maxTokens: 16384,
2064
+ contextWindow: 262144,
2065
+ supportsImages: true,
2066
+ supportsPromptCache: true,
2067
+ inputPrice: 0.6,
2068
+ outputPrice: 3,
2069
+ cacheReadsPrice: 0.1,
2070
+ description: "Kimi K2.5 is Moonshot AI's flagship agentic model and a new SOTA open model. It unifies vision and text, thinking and non-thinking modes, and single-agent and multi-agent execution into one model. Fireworks enables users to control the reasoning behavior and inspect its reasoning history for greater transparency."
2071
+ },
1994
2072
  "accounts/fireworks/models/minimax-m2": {
1995
2073
  maxTokens: 4096,
1996
2074
  contextWindow: 204800,
@@ -3182,7 +3260,7 @@ var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3182
3260
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3183
3261
 
3184
3262
  // src/providers/openai-codex.ts
3185
- var openAiCodexDefaultModelId = "gpt-5.2-codex";
3263
+ var openAiCodexDefaultModelId = "gpt-5.3-codex";
3186
3264
  var openAiCodexModels = {
3187
3265
  "gpt-5.1-codex-max": {
3188
3266
  maxTokens: 128e3,
@@ -3214,6 +3292,20 @@ var openAiCodexModels = {
3214
3292
  supportsTemperature: false,
3215
3293
  description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3216
3294
  },
3295
+ "gpt-5.3-codex": {
3296
+ maxTokens: 128e3,
3297
+ contextWindow: 4e5,
3298
+ includedTools: ["apply_patch"],
3299
+ excludedTools: ["apply_diff", "write_to_file"],
3300
+ supportsImages: true,
3301
+ supportsPromptCache: true,
3302
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3303
+ reasoningEffort: "medium",
3304
+ inputPrice: 0,
3305
+ outputPrice: 0,
3306
+ supportsTemperature: false,
3307
+ description: "GPT-5.3 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3308
+ },
3217
3309
  "gpt-5.2-codex": {
3218
3310
  maxTokens: 128e3,
3219
3311
  contextWindow: 4e5,
@@ -3356,8 +3448,9 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3356
3448
  "anthropic/claude-sonnet-4.5",
3357
3449
  "anthropic/claude-opus-4",
3358
3450
  "anthropic/claude-opus-4.1",
3359
- "anthropic/claude-haiku-4.5",
3360
3451
  "anthropic/claude-opus-4.5",
3452
+ "anthropic/claude-opus-4.6",
3453
+ "anthropic/claude-haiku-4.5",
3361
3454
  "google/gemini-2.5-flash-preview",
3362
3455
  "google/gemini-2.5-flash-preview:thinking",
3363
3456
  "google/gemini-2.5-flash-preview-05-20",
@@ -3377,9 +3470,10 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3377
3470
  "anthropic/claude-3.7-sonnet:beta",
3378
3471
  "anthropic/claude-opus-4",
3379
3472
  "anthropic/claude-opus-4.1",
3473
+ "anthropic/claude-opus-4.5",
3474
+ "anthropic/claude-opus-4.6",
3380
3475
  "anthropic/claude-sonnet-4",
3381
3476
  "anthropic/claude-sonnet-4.5",
3382
- "anthropic/claude-opus-4.5",
3383
3477
  "anthropic/claude-haiku-4.5",
3384
3478
  "google/gemini-2.5-pro-preview",
3385
3479
  "google/gemini-2.5-pro",
@@ -3837,6 +3931,37 @@ var vertexModels = {
3837
3931
  cacheReadsPrice: 0.1,
3838
3932
  supportsReasoningBudget: true
3839
3933
  },
3934
+ "claude-opus-4-6": {
3935
+ maxTokens: 8192,
3936
+ contextWindow: 2e5,
3937
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
3938
+ supportsImages: true,
3939
+ supportsPromptCache: true,
3940
+ inputPrice: 5,
3941
+ // $5 per million input tokens (≤200K context)
3942
+ outputPrice: 25,
3943
+ // $25 per million output tokens (≤200K context)
3944
+ cacheWritesPrice: 6.25,
3945
+ // $6.25 per million tokens
3946
+ cacheReadsPrice: 0.5,
3947
+ // $0.50 per million tokens
3948
+ supportsReasoningBudget: true,
3949
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
3950
+ tiers: [
3951
+ {
3952
+ contextWindow: 1e6,
3953
+ // 1M tokens with beta flag
3954
+ inputPrice: 10,
3955
+ // $10 per million input tokens (>200K context)
3956
+ outputPrice: 37.5,
3957
+ // $37.50 per million output tokens (>200K context)
3958
+ cacheWritesPrice: 12.5,
3959
+ // $12.50 per million tokens (>200K context)
3960
+ cacheReadsPrice: 1
3961
+ // $1.00 per million tokens (>200K context)
3962
+ }
3963
+ ]
3964
+ },
3840
3965
  "claude-opus-4-5@20251101": {
3841
3966
  maxTokens: 8192,
3842
3967
  contextWindow: 2e5,
@@ -4026,7 +4151,11 @@ var vertexModels = {
4026
4151
  description: "Kimi K2 Thinking Model with 256K context window."
4027
4152
  }
4028
4153
  };
4029
- var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
4154
+ var VERTEX_1M_CONTEXT_MODEL_IDS = [
4155
+ "claude-sonnet-4@20250514",
4156
+ "claude-sonnet-4-5@20250929",
4157
+ "claude-opus-4-6"
4158
+ ];
4030
4159
  var VERTEX_REGIONS = [
4031
4160
  { value: "global", label: "global" },
4032
4161
  { value: "us-central1", label: "us-central1" },
@@ -4274,6 +4403,8 @@ var xaiModels = {
4274
4403
  cacheWritesPrice: 0.05,
4275
4404
  cacheReadsPrice: 0.05,
4276
4405
  description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning",
4406
+ supportsReasoningEffort: ["low", "high"],
4407
+ reasoningEffort: "low",
4277
4408
  includedTools: ["search_replace"],
4278
4409
  excludedTools: ["apply_diff"]
4279
4410
  },
@@ -4300,6 +4431,8 @@ var xaiModels = {
4300
4431
  cacheWritesPrice: 0.05,
4301
4432
  cacheReadsPrice: 0.05,
4302
4433
  description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning",
4434
+ supportsReasoningEffort: ["low", "high"],
4435
+ reasoningEffort: "low",
4303
4436
  includedTools: ["search_replace"],
4304
4437
  excludedTools: ["apply_diff"]
4305
4438
  },
@@ -4369,6 +4502,8 @@ var VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
4369
4502
  "anthropic/claude-3.7-sonnet",
4370
4503
  "anthropic/claude-opus-4",
4371
4504
  "anthropic/claude-opus-4.1",
4505
+ "anthropic/claude-opus-4.5",
4506
+ "anthropic/claude-opus-4.6",
4372
4507
  "anthropic/claude-sonnet-4",
4373
4508
  "openai/gpt-4.1",
4374
4509
  "openai/gpt-4.1-mini",
@@ -4406,6 +4541,8 @@ var VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS = /* @__PURE__ */ new Set([
4406
4541
  "anthropic/claude-3.7-sonnet",
4407
4542
  "anthropic/claude-opus-4",
4408
4543
  "anthropic/claude-opus-4.1",
4544
+ "anthropic/claude-opus-4.5",
4545
+ "anthropic/claude-opus-4.6",
4409
4546
  "anthropic/claude-sonnet-4",
4410
4547
  "google/gemini-1.5-flash",
4411
4548
  "google/gemini-1.5-pro",