@roo-code/types 1.94.0 → 1.96.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -164,7 +164,8 @@ var toolNames = [
164
164
  "codebase_search",
165
165
  "update_todo_list",
166
166
  "run_slash_command",
167
- "generate_image"
167
+ "generate_image",
168
+ "custom_tool"
168
169
  ];
169
170
  var toolNamesSchema = z2.enum(toolNames);
170
171
  var toolUsageSchema = z2.record(
@@ -826,6 +827,7 @@ var basetenModels = {
826
827
  contextWindow: 163840,
827
828
  supportsImages: false,
828
829
  supportsPromptCache: false,
830
+ supportsNativeTools: true,
829
831
  inputPrice: 2.55,
830
832
  outputPrice: 5.95,
831
833
  cacheWritesPrice: 0,
@@ -837,6 +839,7 @@ var basetenModels = {
837
839
  contextWindow: 163840,
838
840
  supportsImages: false,
839
841
  supportsPromptCache: false,
842
+ supportsNativeTools: true,
840
843
  inputPrice: 2.55,
841
844
  outputPrice: 5.95,
842
845
  cacheWritesPrice: 0,
@@ -848,6 +851,7 @@ var basetenModels = {
848
851
  contextWindow: 163840,
849
852
  supportsImages: false,
850
853
  supportsPromptCache: false,
854
+ supportsNativeTools: true,
851
855
  inputPrice: 0.77,
852
856
  outputPrice: 0.77,
853
857
  cacheWritesPrice: 0,
@@ -859,6 +863,7 @@ var basetenModels = {
859
863
  contextWindow: 163840,
860
864
  supportsImages: false,
861
865
  supportsPromptCache: false,
866
+ supportsNativeTools: true,
862
867
  inputPrice: 0.5,
863
868
  outputPrice: 1.5,
864
869
  cacheWritesPrice: 0,
@@ -877,11 +882,24 @@ var basetenModels = {
877
882
  cacheReadsPrice: 0,
878
883
  description: "DeepSeek's hybrid reasoning model with efficient long context scaling with GPT-5 level performance"
879
884
  },
885
+ "openai/gpt-oss-120b": {
886
+ maxTokens: 16384,
887
+ contextWindow: 128072,
888
+ supportsImages: false,
889
+ supportsPromptCache: false,
890
+ supportsNativeTools: true,
891
+ inputPrice: 0.1,
892
+ outputPrice: 0.5,
893
+ cacheWritesPrice: 0,
894
+ cacheReadsPrice: 0,
895
+ description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
896
+ },
880
897
  "Qwen/Qwen3-235B-A22B-Instruct-2507": {
881
898
  maxTokens: 16384,
882
899
  contextWindow: 262144,
883
900
  supportsImages: false,
884
901
  supportsPromptCache: false,
902
+ supportsNativeTools: true,
885
903
  inputPrice: 0.22,
886
904
  outputPrice: 0.8,
887
905
  cacheWritesPrice: 0,
@@ -893,24 +911,13 @@ var basetenModels = {
893
911
  contextWindow: 262144,
894
912
  supportsImages: false,
895
913
  supportsPromptCache: false,
914
+ supportsNativeTools: true,
896
915
  inputPrice: 0.38,
897
916
  outputPrice: 1.53,
898
917
  cacheWritesPrice: 0,
899
918
  cacheReadsPrice: 0,
900
919
  description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
901
920
  },
902
- "openai/gpt-oss-120b": {
903
- maxTokens: 16384,
904
- contextWindow: 128072,
905
- supportsImages: false,
906
- supportsPromptCache: false,
907
- supportsNativeTools: true,
908
- inputPrice: 0.1,
909
- outputPrice: 0.5,
910
- cacheWritesPrice: 0,
911
- cacheReadsPrice: 0,
912
- description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
913
- },
914
921
  "moonshotai/Kimi-K2-Instruct-0905": {
915
922
  maxTokens: 16384,
916
923
  contextWindow: 262e3,
@@ -1382,22 +1389,6 @@ var bedrockModels = {
1382
1389
  outputPrice: 0.6,
1383
1390
  description: "Amazon Titan Text Express"
1384
1391
  },
1385
- "amazon.titan-text-embeddings-v1:0": {
1386
- maxTokens: 8192,
1387
- contextWindow: 8e3,
1388
- supportsImages: false,
1389
- supportsPromptCache: false,
1390
- inputPrice: 0.1,
1391
- description: "Amazon Titan Text Embeddings"
1392
- },
1393
- "amazon.titan-text-embeddings-v2:0": {
1394
- maxTokens: 8192,
1395
- contextWindow: 8e3,
1396
- supportsImages: false,
1397
- supportsPromptCache: false,
1398
- inputPrice: 0.02,
1399
- description: "Amazon Titan Text Embeddings V2"
1400
- },
1401
1392
  "moonshot.kimi-k2-thinking": {
1402
1393
  maxTokens: 32e3,
1403
1394
  contextWindow: 262144,
@@ -1531,7 +1522,7 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1531
1522
  var cerebrasDefaultModelId = "gpt-oss-120b";
1532
1523
  var cerebrasModels = {
1533
1524
  "zai-glm-4.6": {
1534
- maxTokens: 8192,
1525
+ maxTokens: 16384,
1535
1526
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1536
1527
  contextWindow: 131072,
1537
1528
  supportsImages: false,
@@ -1543,7 +1534,7 @@ var cerebrasModels = {
1543
1534
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1544
1535
  },
1545
1536
  "qwen-3-235b-a22b-instruct-2507": {
1546
- maxTokens: 8192,
1537
+ maxTokens: 16384,
1547
1538
  // Conservative default to avoid premature rate limiting
1548
1539
  contextWindow: 64e3,
1549
1540
  supportsImages: false,
@@ -1555,7 +1546,7 @@ var cerebrasModels = {
1555
1546
  description: "Intelligent model with ~1400 tokens/s"
1556
1547
  },
1557
1548
  "llama-3.3-70b": {
1558
- maxTokens: 8192,
1549
+ maxTokens: 16384,
1559
1550
  // Conservative default to avoid premature rate limiting
1560
1551
  contextWindow: 64e3,
1561
1552
  supportsImages: false,
@@ -1567,7 +1558,7 @@ var cerebrasModels = {
1567
1558
  description: "Powerful model with ~2600 tokens/s"
1568
1559
  },
1569
1560
  "qwen-3-32b": {
1570
- maxTokens: 8192,
1561
+ maxTokens: 16384,
1571
1562
  // Conservative default to avoid premature rate limiting
1572
1563
  contextWindow: 64e3,
1573
1564
  supportsImages: false,
@@ -1579,7 +1570,7 @@ var cerebrasModels = {
1579
1570
  description: "SOTA coding performance with ~2500 tokens/s"
1580
1571
  },
1581
1572
  "gpt-oss-120b": {
1582
- maxTokens: 8192,
1573
+ maxTokens: 16384,
1583
1574
  // Conservative default to avoid premature rate limiting
1584
1575
  contextWindow: 64e3,
1585
1576
  supportsImages: false,
@@ -2217,6 +2208,7 @@ var featherlessModels = {
2217
2208
  contextWindow: 32678,
2218
2209
  supportsImages: false,
2219
2210
  supportsPromptCache: false,
2211
+ supportsNativeTools: true,
2220
2212
  inputPrice: 0,
2221
2213
  outputPrice: 0,
2222
2214
  description: "DeepSeek V3 0324 model."
@@ -2226,6 +2218,7 @@ var featherlessModels = {
2226
2218
  contextWindow: 32678,
2227
2219
  supportsImages: false,
2228
2220
  supportsPromptCache: false,
2221
+ supportsNativeTools: true,
2229
2222
  inputPrice: 0,
2230
2223
  outputPrice: 0,
2231
2224
  description: "DeepSeek R1 0528 model."
@@ -2245,6 +2238,7 @@ var featherlessModels = {
2245
2238
  contextWindow: 32678,
2246
2239
  supportsImages: false,
2247
2240
  supportsPromptCache: false,
2241
+ supportsNativeTools: true,
2248
2242
  inputPrice: 0,
2249
2243
  outputPrice: 0,
2250
2244
  description: "GPT-OSS 120B model."
@@ -2260,7 +2254,7 @@ var featherlessModels = {
2260
2254
  description: "Qwen3 Coder 480B A35B Instruct model."
2261
2255
  }
2262
2256
  };
2263
- var featherlessDefaultModelId = "deepseek-ai/DeepSeek-R1-0528";
2257
+ var featherlessDefaultModelId = "moonshotai/Kimi-K2-Instruct";
2264
2258
 
2265
2259
  // src/providers/fireworks.ts
2266
2260
  var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct-0905";
@@ -2705,33 +2699,6 @@ var groqModels = {
2705
2699
  outputPrice: 0.34,
2706
2700
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
2707
2701
  },
2708
- "meta-llama/llama-4-maverick-17b-128e-instruct": {
2709
- maxTokens: 8192,
2710
- contextWindow: 131072,
2711
- supportsImages: false,
2712
- supportsPromptCache: false,
2713
- inputPrice: 0.2,
2714
- outputPrice: 0.6,
2715
- description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
2716
- },
2717
- "mistral-saba-24b": {
2718
- maxTokens: 8192,
2719
- contextWindow: 32768,
2720
- supportsImages: false,
2721
- supportsPromptCache: false,
2722
- inputPrice: 0.79,
2723
- outputPrice: 0.79,
2724
- description: "Mistral Saba 24B model, 32K context."
2725
- },
2726
- "qwen-qwq-32b": {
2727
- maxTokens: 8192,
2728
- contextWindow: 131072,
2729
- supportsImages: false,
2730
- supportsPromptCache: false,
2731
- inputPrice: 0.29,
2732
- outputPrice: 0.39,
2733
- description: "Alibaba Qwen QwQ 32B model, 128K context."
2734
- },
2735
2702
  "qwen/qwen3-32b": {
2736
2703
  maxTokens: 8192,
2737
2704
  contextWindow: 131072,
@@ -2743,26 +2710,6 @@ var groqModels = {
2743
2710
  outputPrice: 0.59,
2744
2711
  description: "Alibaba Qwen 3 32B model, 128K context."
2745
2712
  },
2746
- "deepseek-r1-distill-llama-70b": {
2747
- maxTokens: 8192,
2748
- contextWindow: 131072,
2749
- supportsImages: false,
2750
- supportsPromptCache: false,
2751
- inputPrice: 0.75,
2752
- outputPrice: 0.99,
2753
- description: "DeepSeek R1 Distill Llama 70B model, 128K context."
2754
- },
2755
- "moonshotai/kimi-k2-instruct": {
2756
- maxTokens: 16384,
2757
- contextWindow: 131072,
2758
- supportsImages: false,
2759
- supportsPromptCache: true,
2760
- inputPrice: 1,
2761
- outputPrice: 3,
2762
- cacheReadsPrice: 0.5,
2763
- // 50% discount for cached input tokens
2764
- description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
2765
- },
2766
2713
  "moonshotai/kimi-k2-instruct-0905": {
2767
2714
  maxTokens: 16384,
2768
2715
  contextWindow: 262144,
@@ -3837,15 +3784,6 @@ var sambaNovaModels = {
3837
3784
  outputPrice: 4.5,
3838
3785
  description: "DeepSeek V3.1 model with 32K context window."
3839
3786
  },
3840
- "DeepSeek-R1-Distill-Llama-70B": {
3841
- maxTokens: 8192,
3842
- contextWindow: 131072,
3843
- supportsImages: false,
3844
- supportsPromptCache: false,
3845
- inputPrice: 0.7,
3846
- outputPrice: 1.4,
3847
- description: "DeepSeek R1 distilled Llama 70B model with 128K context window."
3848
- },
3849
3787
  "Llama-4-Maverick-17B-128E-Instruct": {
3850
3788
  maxTokens: 8192,
3851
3789
  contextWindow: 131072,
@@ -3857,15 +3795,6 @@ var sambaNovaModels = {
3857
3795
  outputPrice: 1.8,
3858
3796
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
3859
3797
  },
3860
- "Llama-3.3-Swallow-70B-Instruct-v0.4": {
3861
- maxTokens: 8192,
3862
- contextWindow: 16384,
3863
- supportsImages: false,
3864
- supportsPromptCache: false,
3865
- inputPrice: 0.6,
3866
- outputPrice: 1.2,
3867
- description: "Tokyotech Llama 3.3 Swallow 70B Instruct v0.4 model with 16K context window."
3868
- },
3869
3798
  "Qwen3-32B": {
3870
3799
  maxTokens: 8192,
3871
3800
  contextWindow: 8192,
@@ -4398,6 +4327,7 @@ var vertexModels = {
4398
4327
  contextWindow: 131072,
4399
4328
  supportsImages: false,
4400
4329
  supportsPromptCache: false,
4330
+ supportsNativeTools: true,
4401
4331
  inputPrice: 0.35,
4402
4332
  outputPrice: 1.15,
4403
4333
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4407,6 +4337,7 @@ var vertexModels = {
4407
4337
  contextWindow: 163840,
4408
4338
  supportsImages: false,
4409
4339
  supportsPromptCache: false,
4340
+ supportsNativeTools: true,
4410
4341
  inputPrice: 1.35,
4411
4342
  outputPrice: 5.4,
4412
4343
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4416,6 +4347,7 @@ var vertexModels = {
4416
4347
  contextWindow: 163840,
4417
4348
  supportsImages: false,
4418
4349
  supportsPromptCache: false,
4350
+ supportsNativeTools: true,
4419
4351
  inputPrice: 0.6,
4420
4352
  outputPrice: 1.7,
4421
4353
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4425,6 +4357,7 @@ var vertexModels = {
4425
4357
  contextWindow: 131072,
4426
4358
  supportsImages: false,
4427
4359
  supportsPromptCache: false,
4360
+ supportsNativeTools: true,
4428
4361
  inputPrice: 0.15,
4429
4362
  outputPrice: 0.6,
4430
4363
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4434,6 +4367,7 @@ var vertexModels = {
4434
4367
  contextWindow: 131072,
4435
4368
  supportsImages: false,
4436
4369
  supportsPromptCache: false,
4370
+ supportsNativeTools: true,
4437
4371
  inputPrice: 0.075,
4438
4372
  outputPrice: 0.3,
4439
4373
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4443,6 +4377,7 @@ var vertexModels = {
4443
4377
  contextWindow: 262144,
4444
4378
  supportsImages: false,
4445
4379
  supportsPromptCache: false,
4380
+ supportsNativeTools: true,
4446
4381
  inputPrice: 1,
4447
4382
  outputPrice: 4,
4448
4383
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4452,6 +4387,7 @@ var vertexModels = {
4452
4387
  contextWindow: 262144,
4453
4388
  supportsImages: false,
4454
4389
  supportsPromptCache: false,
4390
+ supportsNativeTools: true,
4455
4391
  inputPrice: 0.25,
4456
4392
  outputPrice: 1,
4457
4393
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
@@ -4892,6 +4828,7 @@ var vercelAiGatewayDefaultModelInfo = {
4892
4828
  contextWindow: 2e5,
4893
4829
  supportsImages: true,
4894
4830
  supportsPromptCache: true,
4831
+ supportsNativeTools: true,
4895
4832
  inputPrice: 3,
4896
4833
  outputPrice: 15,
4897
4834
  cacheWritesPrice: 3.75,
@@ -4904,7 +4841,7 @@ var VERCEL_AI_GATEWAY_DEFAULT_TEMPERATURE = 0.7;
4904
4841
  var internationalZAiDefaultModelId = "glm-4.6";
4905
4842
  var internationalZAiModels = {
4906
4843
  "glm-4.5": {
4907
- maxTokens: 98304,
4844
+ maxTokens: 16384,
4908
4845
  contextWindow: 131072,
4909
4846
  supportsImages: false,
4910
4847
  supportsPromptCache: true,
@@ -4917,7 +4854,7 @@ var internationalZAiModels = {
4917
4854
  description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
4918
4855
  },
4919
4856
  "glm-4.5-air": {
4920
- maxTokens: 98304,
4857
+ maxTokens: 16384,
4921
4858
  contextWindow: 131072,
4922
4859
  supportsImages: false,
4923
4860
  supportsPromptCache: true,
@@ -4930,7 +4867,7 @@ var internationalZAiModels = {
4930
4867
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
4931
4868
  },
4932
4869
  "glm-4.5-x": {
4933
- maxTokens: 98304,
4870
+ maxTokens: 16384,
4934
4871
  contextWindow: 131072,
4935
4872
  supportsImages: false,
4936
4873
  supportsPromptCache: true,
@@ -4943,7 +4880,7 @@ var internationalZAiModels = {
4943
4880
  description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
4944
4881
  },
4945
4882
  "glm-4.5-airx": {
4946
- maxTokens: 98304,
4883
+ maxTokens: 16384,
4947
4884
  contextWindow: 131072,
4948
4885
  supportsImages: false,
4949
4886
  supportsPromptCache: true,
@@ -4956,7 +4893,7 @@ var internationalZAiModels = {
4956
4893
  description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
4957
4894
  },
4958
4895
  "glm-4.5-flash": {
4959
- maxTokens: 98304,
4896
+ maxTokens: 16384,
4960
4897
  contextWindow: 131072,
4961
4898
  supportsImages: false,
4962
4899
  supportsPromptCache: true,
@@ -4982,7 +4919,7 @@ var internationalZAiModels = {
4982
4919
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
4983
4920
  },
4984
4921
  "glm-4.6": {
4985
- maxTokens: 98304,
4922
+ maxTokens: 16384,
4986
4923
  contextWindow: 2e5,
4987
4924
  supportsImages: false,
4988
4925
  supportsPromptCache: true,
@@ -4994,8 +4931,24 @@ var internationalZAiModels = {
4994
4931
  cacheReadsPrice: 0.11,
4995
4932
  description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
4996
4933
  },
4934
+ "glm-4.7": {
4935
+ maxTokens: 16384,
4936
+ contextWindow: 2e5,
4937
+ supportsImages: false,
4938
+ supportsPromptCache: true,
4939
+ supportsNativeTools: true,
4940
+ defaultToolProtocol: "native",
4941
+ supportsReasoningEffort: ["disable", "medium"],
4942
+ reasoningEffort: "medium",
4943
+ preserveReasoning: true,
4944
+ inputPrice: 0.6,
4945
+ outputPrice: 2.2,
4946
+ cacheWritesPrice: 0,
4947
+ cacheReadsPrice: 0.11,
4948
+ description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4949
+ },
4997
4950
  "glm-4-32b-0414-128k": {
4998
- maxTokens: 98304,
4951
+ maxTokens: 16384,
4999
4952
  contextWindow: 131072,
5000
4953
  supportsImages: false,
5001
4954
  supportsPromptCache: false,
@@ -5011,7 +4964,7 @@ var internationalZAiModels = {
5011
4964
  var mainlandZAiDefaultModelId = "glm-4.6";
5012
4965
  var mainlandZAiModels = {
5013
4966
  "glm-4.5": {
5014
- maxTokens: 98304,
4967
+ maxTokens: 16384,
5015
4968
  contextWindow: 131072,
5016
4969
  supportsImages: false,
5017
4970
  supportsPromptCache: true,
@@ -5024,7 +4977,7 @@ var mainlandZAiModels = {
5024
4977
  description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
5025
4978
  },
5026
4979
  "glm-4.5-air": {
5027
- maxTokens: 98304,
4980
+ maxTokens: 16384,
5028
4981
  contextWindow: 131072,
5029
4982
  supportsImages: false,
5030
4983
  supportsPromptCache: true,
@@ -5037,7 +4990,7 @@ var mainlandZAiModels = {
5037
4990
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
5038
4991
  },
5039
4992
  "glm-4.5-x": {
5040
- maxTokens: 98304,
4993
+ maxTokens: 16384,
5041
4994
  contextWindow: 131072,
5042
4995
  supportsImages: false,
5043
4996
  supportsPromptCache: true,
@@ -5050,7 +5003,7 @@ var mainlandZAiModels = {
5050
5003
  description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
5051
5004
  },
5052
5005
  "glm-4.5-airx": {
5053
- maxTokens: 98304,
5006
+ maxTokens: 16384,
5054
5007
  contextWindow: 131072,
5055
5008
  supportsImages: false,
5056
5009
  supportsPromptCache: true,
@@ -5063,7 +5016,7 @@ var mainlandZAiModels = {
5063
5016
  description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
5064
5017
  },
5065
5018
  "glm-4.5-flash": {
5066
- maxTokens: 98304,
5019
+ maxTokens: 16384,
5067
5020
  contextWindow: 131072,
5068
5021
  supportsImages: false,
5069
5022
  supportsPromptCache: true,
@@ -5089,7 +5042,7 @@ var mainlandZAiModels = {
5089
5042
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
5090
5043
  },
5091
5044
  "glm-4.6": {
5092
- maxTokens: 98304,
5045
+ maxTokens: 16384,
5093
5046
  contextWindow: 204800,
5094
5047
  supportsImages: false,
5095
5048
  supportsPromptCache: true,
@@ -5100,6 +5053,22 @@ var mainlandZAiModels = {
5100
5053
  cacheWritesPrice: 0,
5101
5054
  cacheReadsPrice: 0.057,
5102
5055
  description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
5056
+ },
5057
+ "glm-4.7": {
5058
+ maxTokens: 16384,
5059
+ contextWindow: 204800,
5060
+ supportsImages: false,
5061
+ supportsPromptCache: true,
5062
+ supportsNativeTools: true,
5063
+ defaultToolProtocol: "native",
5064
+ supportsReasoningEffort: ["disable", "medium"],
5065
+ reasoningEffort: "medium",
5066
+ preserveReasoning: true,
5067
+ inputPrice: 0.29,
5068
+ outputPrice: 1.14,
5069
+ cacheWritesPrice: 0,
5070
+ cacheReadsPrice: 0.057,
5071
+ description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
5103
5072
  }
5104
5073
  };
5105
5074
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5149,6 +5118,8 @@ var minimaxModels = {
5149
5118
  supportsPromptCache: true,
5150
5119
  supportsNativeTools: true,
5151
5120
  defaultToolProtocol: "native",
5121
+ includedTools: ["search_and_replace"],
5122
+ excludedTools: ["apply_diff"],
5152
5123
  preserveReasoning: true,
5153
5124
  inputPrice: 0.3,
5154
5125
  outputPrice: 1.2,
@@ -5163,12 +5134,30 @@ var minimaxModels = {
5163
5134
  supportsPromptCache: true,
5164
5135
  supportsNativeTools: true,
5165
5136
  defaultToolProtocol: "native",
5137
+ includedTools: ["search_and_replace"],
5138
+ excludedTools: ["apply_diff"],
5166
5139
  preserveReasoning: true,
5167
5140
  inputPrice: 0.3,
5168
5141
  outputPrice: 1.2,
5169
5142
  cacheWritesPrice: 0.375,
5170
5143
  cacheReadsPrice: 0.03,
5171
5144
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
5145
+ },
5146
+ "MiniMax-M2.1": {
5147
+ maxTokens: 16384,
5148
+ contextWindow: 192e3,
5149
+ supportsImages: false,
5150
+ supportsPromptCache: true,
5151
+ supportsNativeTools: true,
5152
+ defaultToolProtocol: "native",
5153
+ includedTools: ["search_and_replace"],
5154
+ excludedTools: ["apply_diff"],
5155
+ preserveReasoning: true,
5156
+ inputPrice: 0.3,
5157
+ outputPrice: 1.2,
5158
+ cacheWritesPrice: 0.375,
5159
+ cacheReadsPrice: 0.03,
5160
+ description: "MiniMax M2.1 builds on M2 with improved overall performance for agentic coding tasks and significantly faster response times."
5172
5161
  }
5173
5162
  };
5174
5163
  var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
@@ -5350,8 +5339,7 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
5350
5339
  openRouterApiKey: z8.string().optional(),
5351
5340
  openRouterModelId: z8.string().optional(),
5352
5341
  openRouterBaseUrl: z8.string().optional(),
5353
- openRouterSpecificProvider: z8.string().optional(),
5354
- openRouterUseMiddleOutTransform: z8.boolean().optional()
5342
+ openRouterSpecificProvider: z8.string().optional()
5355
5343
  });
5356
5344
  var bedrockSchema = apiModelIdProviderModelSchema.extend({
5357
5345
  awsAccessKey: z8.string().optional(),
@@ -5831,7 +5819,8 @@ var experimentIds = [
5831
5819
  "preventFocusDisruption",
5832
5820
  "imageGeneration",
5833
5821
  "runSlashCommand",
5834
- "multipleNativeToolCalls"
5822
+ "multipleNativeToolCalls",
5823
+ "customTools"
5835
5824
  ];
5836
5825
  var experimentIdsSchema = z10.enum(experimentIds);
5837
5826
  var experimentsSchema = z10.object({
@@ -5840,7 +5829,8 @@ var experimentsSchema = z10.object({
5840
5829
  preventFocusDisruption: z10.boolean().optional(),
5841
5830
  imageGeneration: z10.boolean().optional(),
5842
5831
  runSlashCommand: z10.boolean().optional(),
5843
- multipleNativeToolCalls: z10.boolean().optional()
5832
+ multipleNativeToolCalls: z10.boolean().optional(),
5833
+ customTools: z10.boolean().optional()
5844
5834
  });
5845
5835
 
5846
5836
  // src/telemetry.ts
@@ -6480,7 +6470,6 @@ var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].fil
6480
6470
  var isGlobalStateKey = (key) => GLOBAL_STATE_KEYS.includes(key);
6481
6471
  var EVALS_SETTINGS = {
6482
6472
  apiProvider: "openrouter",
6483
- openRouterUseMiddleOutTransform: false,
6484
6473
  lastShownAnnouncementId: "jul-09-2025-3-23-0",
6485
6474
  pinnedApiConfigs: {},
6486
6475
  autoApprovalEnabled: true,
@@ -7159,11 +7148,6 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7159
7148
  })
7160
7149
  ]);
7161
7150
 
7162
- // src/single-file-read-models.ts
7163
- function shouldUseSingleFileRead(modelId) {
7164
- return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
7165
- }
7166
-
7167
7151
  // src/todo.ts
7168
7152
  import { z as z21 } from "zod";
7169
7153
  var todoStatusSchema = z21.enum(["pending", "in_progress", "completed"]);
@@ -7471,7 +7455,6 @@ export {
7471
7455
  serviceTiers,
7472
7456
  shareResponseSchema,
7473
7457
  shouldReportApiErrorToTelemetry,
7474
- shouldUseSingleFileRead,
7475
7458
  staticAppPropertiesSchema,
7476
7459
  suggestionItemSchema,
7477
7460
  taskBridgeCommandSchema,