@roo-code/types 1.94.0 → 1.96.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -289,7 +289,6 @@ __export(index_exports, {
289
289
  serviceTiers: () => serviceTiers,
290
290
  shareResponseSchema: () => shareResponseSchema,
291
291
  shouldReportApiErrorToTelemetry: () => shouldReportApiErrorToTelemetry,
292
- shouldUseSingleFileRead: () => shouldUseSingleFileRead,
293
292
  staticAppPropertiesSchema: () => staticAppPropertiesSchema,
294
293
  suggestionItemSchema: () => suggestionItemSchema,
295
294
  taskBridgeCommandSchema: () => taskBridgeCommandSchema,
@@ -498,7 +497,8 @@ var toolNames = [
498
497
  "codebase_search",
499
498
  "update_todo_list",
500
499
  "run_slash_command",
501
- "generate_image"
500
+ "generate_image",
501
+ "custom_tool"
502
502
  ];
503
503
  var toolNamesSchema = import_zod2.z.enum(toolNames);
504
504
  var toolUsageSchema = import_zod2.z.record(
@@ -1160,6 +1160,7 @@ var basetenModels = {
1160
1160
  contextWindow: 163840,
1161
1161
  supportsImages: false,
1162
1162
  supportsPromptCache: false,
1163
+ supportsNativeTools: true,
1163
1164
  inputPrice: 2.55,
1164
1165
  outputPrice: 5.95,
1165
1166
  cacheWritesPrice: 0,
@@ -1171,6 +1172,7 @@ var basetenModels = {
1171
1172
  contextWindow: 163840,
1172
1173
  supportsImages: false,
1173
1174
  supportsPromptCache: false,
1175
+ supportsNativeTools: true,
1174
1176
  inputPrice: 2.55,
1175
1177
  outputPrice: 5.95,
1176
1178
  cacheWritesPrice: 0,
@@ -1182,6 +1184,7 @@ var basetenModels = {
1182
1184
  contextWindow: 163840,
1183
1185
  supportsImages: false,
1184
1186
  supportsPromptCache: false,
1187
+ supportsNativeTools: true,
1185
1188
  inputPrice: 0.77,
1186
1189
  outputPrice: 0.77,
1187
1190
  cacheWritesPrice: 0,
@@ -1193,6 +1196,7 @@ var basetenModels = {
1193
1196
  contextWindow: 163840,
1194
1197
  supportsImages: false,
1195
1198
  supportsPromptCache: false,
1199
+ supportsNativeTools: true,
1196
1200
  inputPrice: 0.5,
1197
1201
  outputPrice: 1.5,
1198
1202
  cacheWritesPrice: 0,
@@ -1211,11 +1215,24 @@ var basetenModels = {
1211
1215
  cacheReadsPrice: 0,
1212
1216
  description: "DeepSeek's hybrid reasoning model with efficient long context scaling with GPT-5 level performance"
1213
1217
  },
1218
+ "openai/gpt-oss-120b": {
1219
+ maxTokens: 16384,
1220
+ contextWindow: 128072,
1221
+ supportsImages: false,
1222
+ supportsPromptCache: false,
1223
+ supportsNativeTools: true,
1224
+ inputPrice: 0.1,
1225
+ outputPrice: 0.5,
1226
+ cacheWritesPrice: 0,
1227
+ cacheReadsPrice: 0,
1228
+ description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
1229
+ },
1214
1230
  "Qwen/Qwen3-235B-A22B-Instruct-2507": {
1215
1231
  maxTokens: 16384,
1216
1232
  contextWindow: 262144,
1217
1233
  supportsImages: false,
1218
1234
  supportsPromptCache: false,
1235
+ supportsNativeTools: true,
1219
1236
  inputPrice: 0.22,
1220
1237
  outputPrice: 0.8,
1221
1238
  cacheWritesPrice: 0,
@@ -1227,24 +1244,13 @@ var basetenModels = {
1227
1244
  contextWindow: 262144,
1228
1245
  supportsImages: false,
1229
1246
  supportsPromptCache: false,
1247
+ supportsNativeTools: true,
1230
1248
  inputPrice: 0.38,
1231
1249
  outputPrice: 1.53,
1232
1250
  cacheWritesPrice: 0,
1233
1251
  cacheReadsPrice: 0,
1234
1252
  description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
1235
1253
  },
1236
- "openai/gpt-oss-120b": {
1237
- maxTokens: 16384,
1238
- contextWindow: 128072,
1239
- supportsImages: false,
1240
- supportsPromptCache: false,
1241
- supportsNativeTools: true,
1242
- inputPrice: 0.1,
1243
- outputPrice: 0.5,
1244
- cacheWritesPrice: 0,
1245
- cacheReadsPrice: 0,
1246
- description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
1247
- },
1248
1254
  "moonshotai/Kimi-K2-Instruct-0905": {
1249
1255
  maxTokens: 16384,
1250
1256
  contextWindow: 262e3,
@@ -1716,22 +1722,6 @@ var bedrockModels = {
1716
1722
  outputPrice: 0.6,
1717
1723
  description: "Amazon Titan Text Express"
1718
1724
  },
1719
- "amazon.titan-text-embeddings-v1:0": {
1720
- maxTokens: 8192,
1721
- contextWindow: 8e3,
1722
- supportsImages: false,
1723
- supportsPromptCache: false,
1724
- inputPrice: 0.1,
1725
- description: "Amazon Titan Text Embeddings"
1726
- },
1727
- "amazon.titan-text-embeddings-v2:0": {
1728
- maxTokens: 8192,
1729
- contextWindow: 8e3,
1730
- supportsImages: false,
1731
- supportsPromptCache: false,
1732
- inputPrice: 0.02,
1733
- description: "Amazon Titan Text Embeddings V2"
1734
- },
1735
1725
  "moonshot.kimi-k2-thinking": {
1736
1726
  maxTokens: 32e3,
1737
1727
  contextWindow: 262144,
@@ -1865,7 +1855,7 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1865
1855
  var cerebrasDefaultModelId = "gpt-oss-120b";
1866
1856
  var cerebrasModels = {
1867
1857
  "zai-glm-4.6": {
1868
- maxTokens: 8192,
1858
+ maxTokens: 16384,
1869
1859
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1870
1860
  contextWindow: 131072,
1871
1861
  supportsImages: false,
@@ -1877,7 +1867,7 @@ var cerebrasModels = {
1877
1867
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1878
1868
  },
1879
1869
  "qwen-3-235b-a22b-instruct-2507": {
1880
- maxTokens: 8192,
1870
+ maxTokens: 16384,
1881
1871
  // Conservative default to avoid premature rate limiting
1882
1872
  contextWindow: 64e3,
1883
1873
  supportsImages: false,
@@ -1889,7 +1879,7 @@ var cerebrasModels = {
1889
1879
  description: "Intelligent model with ~1400 tokens/s"
1890
1880
  },
1891
1881
  "llama-3.3-70b": {
1892
- maxTokens: 8192,
1882
+ maxTokens: 16384,
1893
1883
  // Conservative default to avoid premature rate limiting
1894
1884
  contextWindow: 64e3,
1895
1885
  supportsImages: false,
@@ -1901,7 +1891,7 @@ var cerebrasModels = {
1901
1891
  description: "Powerful model with ~2600 tokens/s"
1902
1892
  },
1903
1893
  "qwen-3-32b": {
1904
- maxTokens: 8192,
1894
+ maxTokens: 16384,
1905
1895
  // Conservative default to avoid premature rate limiting
1906
1896
  contextWindow: 64e3,
1907
1897
  supportsImages: false,
@@ -1913,7 +1903,7 @@ var cerebrasModels = {
1913
1903
  description: "SOTA coding performance with ~2500 tokens/s"
1914
1904
  },
1915
1905
  "gpt-oss-120b": {
1916
- maxTokens: 8192,
1906
+ maxTokens: 16384,
1917
1907
  // Conservative default to avoid premature rate limiting
1918
1908
  contextWindow: 64e3,
1919
1909
  supportsImages: false,
@@ -2551,6 +2541,7 @@ var featherlessModels = {
2551
2541
  contextWindow: 32678,
2552
2542
  supportsImages: false,
2553
2543
  supportsPromptCache: false,
2544
+ supportsNativeTools: true,
2554
2545
  inputPrice: 0,
2555
2546
  outputPrice: 0,
2556
2547
  description: "DeepSeek V3 0324 model."
@@ -2560,6 +2551,7 @@ var featherlessModels = {
2560
2551
  contextWindow: 32678,
2561
2552
  supportsImages: false,
2562
2553
  supportsPromptCache: false,
2554
+ supportsNativeTools: true,
2563
2555
  inputPrice: 0,
2564
2556
  outputPrice: 0,
2565
2557
  description: "DeepSeek R1 0528 model."
@@ -2579,6 +2571,7 @@ var featherlessModels = {
2579
2571
  contextWindow: 32678,
2580
2572
  supportsImages: false,
2581
2573
  supportsPromptCache: false,
2574
+ supportsNativeTools: true,
2582
2575
  inputPrice: 0,
2583
2576
  outputPrice: 0,
2584
2577
  description: "GPT-OSS 120B model."
@@ -2594,7 +2587,7 @@ var featherlessModels = {
2594
2587
  description: "Qwen3 Coder 480B A35B Instruct model."
2595
2588
  }
2596
2589
  };
2597
- var featherlessDefaultModelId = "deepseek-ai/DeepSeek-R1-0528";
2590
+ var featherlessDefaultModelId = "moonshotai/Kimi-K2-Instruct";
2598
2591
 
2599
2592
  // src/providers/fireworks.ts
2600
2593
  var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct-0905";
@@ -3039,33 +3032,6 @@ var groqModels = {
3039
3032
  outputPrice: 0.34,
3040
3033
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
3041
3034
  },
3042
- "meta-llama/llama-4-maverick-17b-128e-instruct": {
3043
- maxTokens: 8192,
3044
- contextWindow: 131072,
3045
- supportsImages: false,
3046
- supportsPromptCache: false,
3047
- inputPrice: 0.2,
3048
- outputPrice: 0.6,
3049
- description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
3050
- },
3051
- "mistral-saba-24b": {
3052
- maxTokens: 8192,
3053
- contextWindow: 32768,
3054
- supportsImages: false,
3055
- supportsPromptCache: false,
3056
- inputPrice: 0.79,
3057
- outputPrice: 0.79,
3058
- description: "Mistral Saba 24B model, 32K context."
3059
- },
3060
- "qwen-qwq-32b": {
3061
- maxTokens: 8192,
3062
- contextWindow: 131072,
3063
- supportsImages: false,
3064
- supportsPromptCache: false,
3065
- inputPrice: 0.29,
3066
- outputPrice: 0.39,
3067
- description: "Alibaba Qwen QwQ 32B model, 128K context."
3068
- },
3069
3035
  "qwen/qwen3-32b": {
3070
3036
  maxTokens: 8192,
3071
3037
  contextWindow: 131072,
@@ -3077,26 +3043,6 @@ var groqModels = {
3077
3043
  outputPrice: 0.59,
3078
3044
  description: "Alibaba Qwen 3 32B model, 128K context."
3079
3045
  },
3080
- "deepseek-r1-distill-llama-70b": {
3081
- maxTokens: 8192,
3082
- contextWindow: 131072,
3083
- supportsImages: false,
3084
- supportsPromptCache: false,
3085
- inputPrice: 0.75,
3086
- outputPrice: 0.99,
3087
- description: "DeepSeek R1 Distill Llama 70B model, 128K context."
3088
- },
3089
- "moonshotai/kimi-k2-instruct": {
3090
- maxTokens: 16384,
3091
- contextWindow: 131072,
3092
- supportsImages: false,
3093
- supportsPromptCache: true,
3094
- inputPrice: 1,
3095
- outputPrice: 3,
3096
- cacheReadsPrice: 0.5,
3097
- // 50% discount for cached input tokens
3098
- description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
3099
- },
3100
3046
  "moonshotai/kimi-k2-instruct-0905": {
3101
3047
  maxTokens: 16384,
3102
3048
  contextWindow: 262144,
@@ -4171,15 +4117,6 @@ var sambaNovaModels = {
4171
4117
  outputPrice: 4.5,
4172
4118
  description: "DeepSeek V3.1 model with 32K context window."
4173
4119
  },
4174
- "DeepSeek-R1-Distill-Llama-70B": {
4175
- maxTokens: 8192,
4176
- contextWindow: 131072,
4177
- supportsImages: false,
4178
- supportsPromptCache: false,
4179
- inputPrice: 0.7,
4180
- outputPrice: 1.4,
4181
- description: "DeepSeek R1 distilled Llama 70B model with 128K context window."
4182
- },
4183
4120
  "Llama-4-Maverick-17B-128E-Instruct": {
4184
4121
  maxTokens: 8192,
4185
4122
  contextWindow: 131072,
@@ -4191,15 +4128,6 @@ var sambaNovaModels = {
4191
4128
  outputPrice: 1.8,
4192
4129
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
4193
4130
  },
4194
- "Llama-3.3-Swallow-70B-Instruct-v0.4": {
4195
- maxTokens: 8192,
4196
- contextWindow: 16384,
4197
- supportsImages: false,
4198
- supportsPromptCache: false,
4199
- inputPrice: 0.6,
4200
- outputPrice: 1.2,
4201
- description: "Tokyotech Llama 3.3 Swallow 70B Instruct v0.4 model with 16K context window."
4202
- },
4203
4131
  "Qwen3-32B": {
4204
4132
  maxTokens: 8192,
4205
4133
  contextWindow: 8192,
@@ -4732,6 +4660,7 @@ var vertexModels = {
4732
4660
  contextWindow: 131072,
4733
4661
  supportsImages: false,
4734
4662
  supportsPromptCache: false,
4663
+ supportsNativeTools: true,
4735
4664
  inputPrice: 0.35,
4736
4665
  outputPrice: 1.15,
4737
4666
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4741,6 +4670,7 @@ var vertexModels = {
4741
4670
  contextWindow: 163840,
4742
4671
  supportsImages: false,
4743
4672
  supportsPromptCache: false,
4673
+ supportsNativeTools: true,
4744
4674
  inputPrice: 1.35,
4745
4675
  outputPrice: 5.4,
4746
4676
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4750,6 +4680,7 @@ var vertexModels = {
4750
4680
  contextWindow: 163840,
4751
4681
  supportsImages: false,
4752
4682
  supportsPromptCache: false,
4683
+ supportsNativeTools: true,
4753
4684
  inputPrice: 0.6,
4754
4685
  outputPrice: 1.7,
4755
4686
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4759,6 +4690,7 @@ var vertexModels = {
4759
4690
  contextWindow: 131072,
4760
4691
  supportsImages: false,
4761
4692
  supportsPromptCache: false,
4693
+ supportsNativeTools: true,
4762
4694
  inputPrice: 0.15,
4763
4695
  outputPrice: 0.6,
4764
4696
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4768,6 +4700,7 @@ var vertexModels = {
4768
4700
  contextWindow: 131072,
4769
4701
  supportsImages: false,
4770
4702
  supportsPromptCache: false,
4703
+ supportsNativeTools: true,
4771
4704
  inputPrice: 0.075,
4772
4705
  outputPrice: 0.3,
4773
4706
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4777,6 +4710,7 @@ var vertexModels = {
4777
4710
  contextWindow: 262144,
4778
4711
  supportsImages: false,
4779
4712
  supportsPromptCache: false,
4713
+ supportsNativeTools: true,
4780
4714
  inputPrice: 1,
4781
4715
  outputPrice: 4,
4782
4716
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4786,6 +4720,7 @@ var vertexModels = {
4786
4720
  contextWindow: 262144,
4787
4721
  supportsImages: false,
4788
4722
  supportsPromptCache: false,
4723
+ supportsNativeTools: true,
4789
4724
  inputPrice: 0.25,
4790
4725
  outputPrice: 1,
4791
4726
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
@@ -5226,6 +5161,7 @@ var vercelAiGatewayDefaultModelInfo = {
5226
5161
  contextWindow: 2e5,
5227
5162
  supportsImages: true,
5228
5163
  supportsPromptCache: true,
5164
+ supportsNativeTools: true,
5229
5165
  inputPrice: 3,
5230
5166
  outputPrice: 15,
5231
5167
  cacheWritesPrice: 3.75,
@@ -5238,7 +5174,7 @@ var VERCEL_AI_GATEWAY_DEFAULT_TEMPERATURE = 0.7;
5238
5174
  var internationalZAiDefaultModelId = "glm-4.6";
5239
5175
  var internationalZAiModels = {
5240
5176
  "glm-4.5": {
5241
- maxTokens: 98304,
5177
+ maxTokens: 16384,
5242
5178
  contextWindow: 131072,
5243
5179
  supportsImages: false,
5244
5180
  supportsPromptCache: true,
@@ -5251,7 +5187,7 @@ var internationalZAiModels = {
5251
5187
  description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
5252
5188
  },
5253
5189
  "glm-4.5-air": {
5254
- maxTokens: 98304,
5190
+ maxTokens: 16384,
5255
5191
  contextWindow: 131072,
5256
5192
  supportsImages: false,
5257
5193
  supportsPromptCache: true,
@@ -5264,7 +5200,7 @@ var internationalZAiModels = {
5264
5200
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
5265
5201
  },
5266
5202
  "glm-4.5-x": {
5267
- maxTokens: 98304,
5203
+ maxTokens: 16384,
5268
5204
  contextWindow: 131072,
5269
5205
  supportsImages: false,
5270
5206
  supportsPromptCache: true,
@@ -5277,7 +5213,7 @@ var internationalZAiModels = {
5277
5213
  description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
5278
5214
  },
5279
5215
  "glm-4.5-airx": {
5280
- maxTokens: 98304,
5216
+ maxTokens: 16384,
5281
5217
  contextWindow: 131072,
5282
5218
  supportsImages: false,
5283
5219
  supportsPromptCache: true,
@@ -5290,7 +5226,7 @@ var internationalZAiModels = {
5290
5226
  description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
5291
5227
  },
5292
5228
  "glm-4.5-flash": {
5293
- maxTokens: 98304,
5229
+ maxTokens: 16384,
5294
5230
  contextWindow: 131072,
5295
5231
  supportsImages: false,
5296
5232
  supportsPromptCache: true,
@@ -5316,7 +5252,7 @@ var internationalZAiModels = {
5316
5252
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
5317
5253
  },
5318
5254
  "glm-4.6": {
5319
- maxTokens: 98304,
5255
+ maxTokens: 16384,
5320
5256
  contextWindow: 2e5,
5321
5257
  supportsImages: false,
5322
5258
  supportsPromptCache: true,
@@ -5328,8 +5264,24 @@ var internationalZAiModels = {
5328
5264
  cacheReadsPrice: 0.11,
5329
5265
  description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
5330
5266
  },
5267
+ "glm-4.7": {
5268
+ maxTokens: 16384,
5269
+ contextWindow: 2e5,
5270
+ supportsImages: false,
5271
+ supportsPromptCache: true,
5272
+ supportsNativeTools: true,
5273
+ defaultToolProtocol: "native",
5274
+ supportsReasoningEffort: ["disable", "medium"],
5275
+ reasoningEffort: "medium",
5276
+ preserveReasoning: true,
5277
+ inputPrice: 0.6,
5278
+ outputPrice: 2.2,
5279
+ cacheWritesPrice: 0,
5280
+ cacheReadsPrice: 0.11,
5281
+ description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
5282
+ },
5331
5283
  "glm-4-32b-0414-128k": {
5332
- maxTokens: 98304,
5284
+ maxTokens: 16384,
5333
5285
  contextWindow: 131072,
5334
5286
  supportsImages: false,
5335
5287
  supportsPromptCache: false,
@@ -5345,7 +5297,7 @@ var internationalZAiModels = {
5345
5297
  var mainlandZAiDefaultModelId = "glm-4.6";
5346
5298
  var mainlandZAiModels = {
5347
5299
  "glm-4.5": {
5348
- maxTokens: 98304,
5300
+ maxTokens: 16384,
5349
5301
  contextWindow: 131072,
5350
5302
  supportsImages: false,
5351
5303
  supportsPromptCache: true,
@@ -5358,7 +5310,7 @@ var mainlandZAiModels = {
5358
5310
  description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
5359
5311
  },
5360
5312
  "glm-4.5-air": {
5361
- maxTokens: 98304,
5313
+ maxTokens: 16384,
5362
5314
  contextWindow: 131072,
5363
5315
  supportsImages: false,
5364
5316
  supportsPromptCache: true,
@@ -5371,7 +5323,7 @@ var mainlandZAiModels = {
5371
5323
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
5372
5324
  },
5373
5325
  "glm-4.5-x": {
5374
- maxTokens: 98304,
5326
+ maxTokens: 16384,
5375
5327
  contextWindow: 131072,
5376
5328
  supportsImages: false,
5377
5329
  supportsPromptCache: true,
@@ -5384,7 +5336,7 @@ var mainlandZAiModels = {
5384
5336
  description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
5385
5337
  },
5386
5338
  "glm-4.5-airx": {
5387
- maxTokens: 98304,
5339
+ maxTokens: 16384,
5388
5340
  contextWindow: 131072,
5389
5341
  supportsImages: false,
5390
5342
  supportsPromptCache: true,
@@ -5397,7 +5349,7 @@ var mainlandZAiModels = {
5397
5349
  description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
5398
5350
  },
5399
5351
  "glm-4.5-flash": {
5400
- maxTokens: 98304,
5352
+ maxTokens: 16384,
5401
5353
  contextWindow: 131072,
5402
5354
  supportsImages: false,
5403
5355
  supportsPromptCache: true,
@@ -5423,7 +5375,7 @@ var mainlandZAiModels = {
5423
5375
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
5424
5376
  },
5425
5377
  "glm-4.6": {
5426
- maxTokens: 98304,
5378
+ maxTokens: 16384,
5427
5379
  contextWindow: 204800,
5428
5380
  supportsImages: false,
5429
5381
  supportsPromptCache: true,
@@ -5434,6 +5386,22 @@ var mainlandZAiModels = {
5434
5386
  cacheWritesPrice: 0,
5435
5387
  cacheReadsPrice: 0.057,
5436
5388
  description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
5389
+ },
5390
+ "glm-4.7": {
5391
+ maxTokens: 16384,
5392
+ contextWindow: 204800,
5393
+ supportsImages: false,
5394
+ supportsPromptCache: true,
5395
+ supportsNativeTools: true,
5396
+ defaultToolProtocol: "native",
5397
+ supportsReasoningEffort: ["disable", "medium"],
5398
+ reasoningEffort: "medium",
5399
+ preserveReasoning: true,
5400
+ inputPrice: 0.29,
5401
+ outputPrice: 1.14,
5402
+ cacheWritesPrice: 0,
5403
+ cacheReadsPrice: 0.057,
5404
+ description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
5437
5405
  }
5438
5406
  };
5439
5407
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5483,6 +5451,8 @@ var minimaxModels = {
5483
5451
  supportsPromptCache: true,
5484
5452
  supportsNativeTools: true,
5485
5453
  defaultToolProtocol: "native",
5454
+ includedTools: ["search_and_replace"],
5455
+ excludedTools: ["apply_diff"],
5486
5456
  preserveReasoning: true,
5487
5457
  inputPrice: 0.3,
5488
5458
  outputPrice: 1.2,
@@ -5497,12 +5467,30 @@ var minimaxModels = {
5497
5467
  supportsPromptCache: true,
5498
5468
  supportsNativeTools: true,
5499
5469
  defaultToolProtocol: "native",
5470
+ includedTools: ["search_and_replace"],
5471
+ excludedTools: ["apply_diff"],
5500
5472
  preserveReasoning: true,
5501
5473
  inputPrice: 0.3,
5502
5474
  outputPrice: 1.2,
5503
5475
  cacheWritesPrice: 0.375,
5504
5476
  cacheReadsPrice: 0.03,
5505
5477
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
5478
+ },
5479
+ "MiniMax-M2.1": {
5480
+ maxTokens: 16384,
5481
+ contextWindow: 192e3,
5482
+ supportsImages: false,
5483
+ supportsPromptCache: true,
5484
+ supportsNativeTools: true,
5485
+ defaultToolProtocol: "native",
5486
+ includedTools: ["search_and_replace"],
5487
+ excludedTools: ["apply_diff"],
5488
+ preserveReasoning: true,
5489
+ inputPrice: 0.3,
5490
+ outputPrice: 1.2,
5491
+ cacheWritesPrice: 0.375,
5492
+ cacheReadsPrice: 0.03,
5493
+ description: "MiniMax M2.1 builds on M2 with improved overall performance for agentic coding tasks and significantly faster response times."
5506
5494
  }
5507
5495
  };
5508
5496
  var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
@@ -5684,8 +5672,7 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
5684
5672
  openRouterApiKey: import_zod8.z.string().optional(),
5685
5673
  openRouterModelId: import_zod8.z.string().optional(),
5686
5674
  openRouterBaseUrl: import_zod8.z.string().optional(),
5687
- openRouterSpecificProvider: import_zod8.z.string().optional(),
5688
- openRouterUseMiddleOutTransform: import_zod8.z.boolean().optional()
5675
+ openRouterSpecificProvider: import_zod8.z.string().optional()
5689
5676
  });
5690
5677
  var bedrockSchema = apiModelIdProviderModelSchema.extend({
5691
5678
  awsAccessKey: import_zod8.z.string().optional(),
@@ -6165,7 +6152,8 @@ var experimentIds = [
6165
6152
  "preventFocusDisruption",
6166
6153
  "imageGeneration",
6167
6154
  "runSlashCommand",
6168
- "multipleNativeToolCalls"
6155
+ "multipleNativeToolCalls",
6156
+ "customTools"
6169
6157
  ];
6170
6158
  var experimentIdsSchema = import_zod10.z.enum(experimentIds);
6171
6159
  var experimentsSchema = import_zod10.z.object({
@@ -6174,7 +6162,8 @@ var experimentsSchema = import_zod10.z.object({
6174
6162
  preventFocusDisruption: import_zod10.z.boolean().optional(),
6175
6163
  imageGeneration: import_zod10.z.boolean().optional(),
6176
6164
  runSlashCommand: import_zod10.z.boolean().optional(),
6177
- multipleNativeToolCalls: import_zod10.z.boolean().optional()
6165
+ multipleNativeToolCalls: import_zod10.z.boolean().optional(),
6166
+ customTools: import_zod10.z.boolean().optional()
6178
6167
  });
6179
6168
 
6180
6169
  // src/telemetry.ts
@@ -6814,7 +6803,6 @@ var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].fil
6814
6803
  var isGlobalStateKey = (key) => GLOBAL_STATE_KEYS.includes(key);
6815
6804
  var EVALS_SETTINGS = {
6816
6805
  apiProvider: "openrouter",
6817
- openRouterUseMiddleOutTransform: false,
6818
6806
  lastShownAnnouncementId: "jul-09-2025-3-23-0",
6819
6807
  pinnedApiConfigs: {},
6820
6808
  autoApprovalEnabled: true,
@@ -7493,11 +7481,6 @@ var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
7493
7481
  })
7494
7482
  ]);
7495
7483
 
7496
- // src/single-file-read-models.ts
7497
- function shouldUseSingleFileRead(modelId) {
7498
- return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
7499
- }
7500
-
7501
7484
  // src/todo.ts
7502
7485
  var import_zod20 = require("zod");
7503
7486
  var todoStatusSchema = import_zod20.z.enum(["pending", "in_progress", "completed"]);
@@ -7806,7 +7789,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7806
7789
  serviceTiers,
7807
7790
  shareResponseSchema,
7808
7791
  shouldReportApiErrorToTelemetry,
7809
- shouldUseSingleFileRead,
7810
7792
  staticAppPropertiesSchema,
7811
7793
  suggestionItemSchema,
7812
7794
  taskBridgeCommandSchema,