@roo-code/types 1.94.0 → 1.96.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +107 -125
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +175 -201
- package/dist/index.d.ts +175 -201
- package/dist/index.js +107 -124
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.js
CHANGED
|
@@ -164,7 +164,8 @@ var toolNames = [
|
|
|
164
164
|
"codebase_search",
|
|
165
165
|
"update_todo_list",
|
|
166
166
|
"run_slash_command",
|
|
167
|
-
"generate_image"
|
|
167
|
+
"generate_image",
|
|
168
|
+
"custom_tool"
|
|
168
169
|
];
|
|
169
170
|
var toolNamesSchema = z2.enum(toolNames);
|
|
170
171
|
var toolUsageSchema = z2.record(
|
|
@@ -826,6 +827,7 @@ var basetenModels = {
|
|
|
826
827
|
contextWindow: 163840,
|
|
827
828
|
supportsImages: false,
|
|
828
829
|
supportsPromptCache: false,
|
|
830
|
+
supportsNativeTools: true,
|
|
829
831
|
inputPrice: 2.55,
|
|
830
832
|
outputPrice: 5.95,
|
|
831
833
|
cacheWritesPrice: 0,
|
|
@@ -837,6 +839,7 @@ var basetenModels = {
|
|
|
837
839
|
contextWindow: 163840,
|
|
838
840
|
supportsImages: false,
|
|
839
841
|
supportsPromptCache: false,
|
|
842
|
+
supportsNativeTools: true,
|
|
840
843
|
inputPrice: 2.55,
|
|
841
844
|
outputPrice: 5.95,
|
|
842
845
|
cacheWritesPrice: 0,
|
|
@@ -848,6 +851,7 @@ var basetenModels = {
|
|
|
848
851
|
contextWindow: 163840,
|
|
849
852
|
supportsImages: false,
|
|
850
853
|
supportsPromptCache: false,
|
|
854
|
+
supportsNativeTools: true,
|
|
851
855
|
inputPrice: 0.77,
|
|
852
856
|
outputPrice: 0.77,
|
|
853
857
|
cacheWritesPrice: 0,
|
|
@@ -859,6 +863,7 @@ var basetenModels = {
|
|
|
859
863
|
contextWindow: 163840,
|
|
860
864
|
supportsImages: false,
|
|
861
865
|
supportsPromptCache: false,
|
|
866
|
+
supportsNativeTools: true,
|
|
862
867
|
inputPrice: 0.5,
|
|
863
868
|
outputPrice: 1.5,
|
|
864
869
|
cacheWritesPrice: 0,
|
|
@@ -877,11 +882,24 @@ var basetenModels = {
|
|
|
877
882
|
cacheReadsPrice: 0,
|
|
878
883
|
description: "DeepSeek's hybrid reasoning model with efficient long context scaling with GPT-5 level performance"
|
|
879
884
|
},
|
|
885
|
+
"openai/gpt-oss-120b": {
|
|
886
|
+
maxTokens: 16384,
|
|
887
|
+
contextWindow: 128072,
|
|
888
|
+
supportsImages: false,
|
|
889
|
+
supportsPromptCache: false,
|
|
890
|
+
supportsNativeTools: true,
|
|
891
|
+
inputPrice: 0.1,
|
|
892
|
+
outputPrice: 0.5,
|
|
893
|
+
cacheWritesPrice: 0,
|
|
894
|
+
cacheReadsPrice: 0,
|
|
895
|
+
description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
|
|
896
|
+
},
|
|
880
897
|
"Qwen/Qwen3-235B-A22B-Instruct-2507": {
|
|
881
898
|
maxTokens: 16384,
|
|
882
899
|
contextWindow: 262144,
|
|
883
900
|
supportsImages: false,
|
|
884
901
|
supportsPromptCache: false,
|
|
902
|
+
supportsNativeTools: true,
|
|
885
903
|
inputPrice: 0.22,
|
|
886
904
|
outputPrice: 0.8,
|
|
887
905
|
cacheWritesPrice: 0,
|
|
@@ -893,24 +911,13 @@ var basetenModels = {
|
|
|
893
911
|
contextWindow: 262144,
|
|
894
912
|
supportsImages: false,
|
|
895
913
|
supportsPromptCache: false,
|
|
914
|
+
supportsNativeTools: true,
|
|
896
915
|
inputPrice: 0.38,
|
|
897
916
|
outputPrice: 1.53,
|
|
898
917
|
cacheWritesPrice: 0,
|
|
899
918
|
cacheReadsPrice: 0,
|
|
900
919
|
description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
|
|
901
920
|
},
|
|
902
|
-
"openai/gpt-oss-120b": {
|
|
903
|
-
maxTokens: 16384,
|
|
904
|
-
contextWindow: 128072,
|
|
905
|
-
supportsImages: false,
|
|
906
|
-
supportsPromptCache: false,
|
|
907
|
-
supportsNativeTools: true,
|
|
908
|
-
inputPrice: 0.1,
|
|
909
|
-
outputPrice: 0.5,
|
|
910
|
-
cacheWritesPrice: 0,
|
|
911
|
-
cacheReadsPrice: 0,
|
|
912
|
-
description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
|
|
913
|
-
},
|
|
914
921
|
"moonshotai/Kimi-K2-Instruct-0905": {
|
|
915
922
|
maxTokens: 16384,
|
|
916
923
|
contextWindow: 262e3,
|
|
@@ -1382,22 +1389,6 @@ var bedrockModels = {
|
|
|
1382
1389
|
outputPrice: 0.6,
|
|
1383
1390
|
description: "Amazon Titan Text Express"
|
|
1384
1391
|
},
|
|
1385
|
-
"amazon.titan-text-embeddings-v1:0": {
|
|
1386
|
-
maxTokens: 8192,
|
|
1387
|
-
contextWindow: 8e3,
|
|
1388
|
-
supportsImages: false,
|
|
1389
|
-
supportsPromptCache: false,
|
|
1390
|
-
inputPrice: 0.1,
|
|
1391
|
-
description: "Amazon Titan Text Embeddings"
|
|
1392
|
-
},
|
|
1393
|
-
"amazon.titan-text-embeddings-v2:0": {
|
|
1394
|
-
maxTokens: 8192,
|
|
1395
|
-
contextWindow: 8e3,
|
|
1396
|
-
supportsImages: false,
|
|
1397
|
-
supportsPromptCache: false,
|
|
1398
|
-
inputPrice: 0.02,
|
|
1399
|
-
description: "Amazon Titan Text Embeddings V2"
|
|
1400
|
-
},
|
|
1401
1392
|
"moonshot.kimi-k2-thinking": {
|
|
1402
1393
|
maxTokens: 32e3,
|
|
1403
1394
|
contextWindow: 262144,
|
|
@@ -1531,7 +1522,7 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1531
1522
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1532
1523
|
var cerebrasModels = {
|
|
1533
1524
|
"zai-glm-4.6": {
|
|
1534
|
-
maxTokens:
|
|
1525
|
+
maxTokens: 16384,
|
|
1535
1526
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1536
1527
|
contextWindow: 131072,
|
|
1537
1528
|
supportsImages: false,
|
|
@@ -1543,7 +1534,7 @@ var cerebrasModels = {
|
|
|
1543
1534
|
description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
|
|
1544
1535
|
},
|
|
1545
1536
|
"qwen-3-235b-a22b-instruct-2507": {
|
|
1546
|
-
maxTokens:
|
|
1537
|
+
maxTokens: 16384,
|
|
1547
1538
|
// Conservative default to avoid premature rate limiting
|
|
1548
1539
|
contextWindow: 64e3,
|
|
1549
1540
|
supportsImages: false,
|
|
@@ -1555,7 +1546,7 @@ var cerebrasModels = {
|
|
|
1555
1546
|
description: "Intelligent model with ~1400 tokens/s"
|
|
1556
1547
|
},
|
|
1557
1548
|
"llama-3.3-70b": {
|
|
1558
|
-
maxTokens:
|
|
1549
|
+
maxTokens: 16384,
|
|
1559
1550
|
// Conservative default to avoid premature rate limiting
|
|
1560
1551
|
contextWindow: 64e3,
|
|
1561
1552
|
supportsImages: false,
|
|
@@ -1567,7 +1558,7 @@ var cerebrasModels = {
|
|
|
1567
1558
|
description: "Powerful model with ~2600 tokens/s"
|
|
1568
1559
|
},
|
|
1569
1560
|
"qwen-3-32b": {
|
|
1570
|
-
maxTokens:
|
|
1561
|
+
maxTokens: 16384,
|
|
1571
1562
|
// Conservative default to avoid premature rate limiting
|
|
1572
1563
|
contextWindow: 64e3,
|
|
1573
1564
|
supportsImages: false,
|
|
@@ -1579,7 +1570,7 @@ var cerebrasModels = {
|
|
|
1579
1570
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
1580
1571
|
},
|
|
1581
1572
|
"gpt-oss-120b": {
|
|
1582
|
-
maxTokens:
|
|
1573
|
+
maxTokens: 16384,
|
|
1583
1574
|
// Conservative default to avoid premature rate limiting
|
|
1584
1575
|
contextWindow: 64e3,
|
|
1585
1576
|
supportsImages: false,
|
|
@@ -2217,6 +2208,7 @@ var featherlessModels = {
|
|
|
2217
2208
|
contextWindow: 32678,
|
|
2218
2209
|
supportsImages: false,
|
|
2219
2210
|
supportsPromptCache: false,
|
|
2211
|
+
supportsNativeTools: true,
|
|
2220
2212
|
inputPrice: 0,
|
|
2221
2213
|
outputPrice: 0,
|
|
2222
2214
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2226,6 +2218,7 @@ var featherlessModels = {
|
|
|
2226
2218
|
contextWindow: 32678,
|
|
2227
2219
|
supportsImages: false,
|
|
2228
2220
|
supportsPromptCache: false,
|
|
2221
|
+
supportsNativeTools: true,
|
|
2229
2222
|
inputPrice: 0,
|
|
2230
2223
|
outputPrice: 0,
|
|
2231
2224
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2245,6 +2238,7 @@ var featherlessModels = {
|
|
|
2245
2238
|
contextWindow: 32678,
|
|
2246
2239
|
supportsImages: false,
|
|
2247
2240
|
supportsPromptCache: false,
|
|
2241
|
+
supportsNativeTools: true,
|
|
2248
2242
|
inputPrice: 0,
|
|
2249
2243
|
outputPrice: 0,
|
|
2250
2244
|
description: "GPT-OSS 120B model."
|
|
@@ -2260,7 +2254,7 @@ var featherlessModels = {
|
|
|
2260
2254
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
2261
2255
|
}
|
|
2262
2256
|
};
|
|
2263
|
-
var featherlessDefaultModelId = "
|
|
2257
|
+
var featherlessDefaultModelId = "moonshotai/Kimi-K2-Instruct";
|
|
2264
2258
|
|
|
2265
2259
|
// src/providers/fireworks.ts
|
|
2266
2260
|
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct-0905";
|
|
@@ -2705,33 +2699,6 @@ var groqModels = {
|
|
|
2705
2699
|
outputPrice: 0.34,
|
|
2706
2700
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
2707
2701
|
},
|
|
2708
|
-
"meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
2709
|
-
maxTokens: 8192,
|
|
2710
|
-
contextWindow: 131072,
|
|
2711
|
-
supportsImages: false,
|
|
2712
|
-
supportsPromptCache: false,
|
|
2713
|
-
inputPrice: 0.2,
|
|
2714
|
-
outputPrice: 0.6,
|
|
2715
|
-
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
2716
|
-
},
|
|
2717
|
-
"mistral-saba-24b": {
|
|
2718
|
-
maxTokens: 8192,
|
|
2719
|
-
contextWindow: 32768,
|
|
2720
|
-
supportsImages: false,
|
|
2721
|
-
supportsPromptCache: false,
|
|
2722
|
-
inputPrice: 0.79,
|
|
2723
|
-
outputPrice: 0.79,
|
|
2724
|
-
description: "Mistral Saba 24B model, 32K context."
|
|
2725
|
-
},
|
|
2726
|
-
"qwen-qwq-32b": {
|
|
2727
|
-
maxTokens: 8192,
|
|
2728
|
-
contextWindow: 131072,
|
|
2729
|
-
supportsImages: false,
|
|
2730
|
-
supportsPromptCache: false,
|
|
2731
|
-
inputPrice: 0.29,
|
|
2732
|
-
outputPrice: 0.39,
|
|
2733
|
-
description: "Alibaba Qwen QwQ 32B model, 128K context."
|
|
2734
|
-
},
|
|
2735
2702
|
"qwen/qwen3-32b": {
|
|
2736
2703
|
maxTokens: 8192,
|
|
2737
2704
|
contextWindow: 131072,
|
|
@@ -2743,26 +2710,6 @@ var groqModels = {
|
|
|
2743
2710
|
outputPrice: 0.59,
|
|
2744
2711
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
2745
2712
|
},
|
|
2746
|
-
"deepseek-r1-distill-llama-70b": {
|
|
2747
|
-
maxTokens: 8192,
|
|
2748
|
-
contextWindow: 131072,
|
|
2749
|
-
supportsImages: false,
|
|
2750
|
-
supportsPromptCache: false,
|
|
2751
|
-
inputPrice: 0.75,
|
|
2752
|
-
outputPrice: 0.99,
|
|
2753
|
-
description: "DeepSeek R1 Distill Llama 70B model, 128K context."
|
|
2754
|
-
},
|
|
2755
|
-
"moonshotai/kimi-k2-instruct": {
|
|
2756
|
-
maxTokens: 16384,
|
|
2757
|
-
contextWindow: 131072,
|
|
2758
|
-
supportsImages: false,
|
|
2759
|
-
supportsPromptCache: true,
|
|
2760
|
-
inputPrice: 1,
|
|
2761
|
-
outputPrice: 3,
|
|
2762
|
-
cacheReadsPrice: 0.5,
|
|
2763
|
-
// 50% discount for cached input tokens
|
|
2764
|
-
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
2765
|
-
},
|
|
2766
2713
|
"moonshotai/kimi-k2-instruct-0905": {
|
|
2767
2714
|
maxTokens: 16384,
|
|
2768
2715
|
contextWindow: 262144,
|
|
@@ -3837,15 +3784,6 @@ var sambaNovaModels = {
|
|
|
3837
3784
|
outputPrice: 4.5,
|
|
3838
3785
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
3839
3786
|
},
|
|
3840
|
-
"DeepSeek-R1-Distill-Llama-70B": {
|
|
3841
|
-
maxTokens: 8192,
|
|
3842
|
-
contextWindow: 131072,
|
|
3843
|
-
supportsImages: false,
|
|
3844
|
-
supportsPromptCache: false,
|
|
3845
|
-
inputPrice: 0.7,
|
|
3846
|
-
outputPrice: 1.4,
|
|
3847
|
-
description: "DeepSeek R1 distilled Llama 70B model with 128K context window."
|
|
3848
|
-
},
|
|
3849
3787
|
"Llama-4-Maverick-17B-128E-Instruct": {
|
|
3850
3788
|
maxTokens: 8192,
|
|
3851
3789
|
contextWindow: 131072,
|
|
@@ -3857,15 +3795,6 @@ var sambaNovaModels = {
|
|
|
3857
3795
|
outputPrice: 1.8,
|
|
3858
3796
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
3859
3797
|
},
|
|
3860
|
-
"Llama-3.3-Swallow-70B-Instruct-v0.4": {
|
|
3861
|
-
maxTokens: 8192,
|
|
3862
|
-
contextWindow: 16384,
|
|
3863
|
-
supportsImages: false,
|
|
3864
|
-
supportsPromptCache: false,
|
|
3865
|
-
inputPrice: 0.6,
|
|
3866
|
-
outputPrice: 1.2,
|
|
3867
|
-
description: "Tokyotech Llama 3.3 Swallow 70B Instruct v0.4 model with 16K context window."
|
|
3868
|
-
},
|
|
3869
3798
|
"Qwen3-32B": {
|
|
3870
3799
|
maxTokens: 8192,
|
|
3871
3800
|
contextWindow: 8192,
|
|
@@ -4398,6 +4327,7 @@ var vertexModels = {
|
|
|
4398
4327
|
contextWindow: 131072,
|
|
4399
4328
|
supportsImages: false,
|
|
4400
4329
|
supportsPromptCache: false,
|
|
4330
|
+
supportsNativeTools: true,
|
|
4401
4331
|
inputPrice: 0.35,
|
|
4402
4332
|
outputPrice: 1.15,
|
|
4403
4333
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4407,6 +4337,7 @@ var vertexModels = {
|
|
|
4407
4337
|
contextWindow: 163840,
|
|
4408
4338
|
supportsImages: false,
|
|
4409
4339
|
supportsPromptCache: false,
|
|
4340
|
+
supportsNativeTools: true,
|
|
4410
4341
|
inputPrice: 1.35,
|
|
4411
4342
|
outputPrice: 5.4,
|
|
4412
4343
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4416,6 +4347,7 @@ var vertexModels = {
|
|
|
4416
4347
|
contextWindow: 163840,
|
|
4417
4348
|
supportsImages: false,
|
|
4418
4349
|
supportsPromptCache: false,
|
|
4350
|
+
supportsNativeTools: true,
|
|
4419
4351
|
inputPrice: 0.6,
|
|
4420
4352
|
outputPrice: 1.7,
|
|
4421
4353
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4425,6 +4357,7 @@ var vertexModels = {
|
|
|
4425
4357
|
contextWindow: 131072,
|
|
4426
4358
|
supportsImages: false,
|
|
4427
4359
|
supportsPromptCache: false,
|
|
4360
|
+
supportsNativeTools: true,
|
|
4428
4361
|
inputPrice: 0.15,
|
|
4429
4362
|
outputPrice: 0.6,
|
|
4430
4363
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4434,6 +4367,7 @@ var vertexModels = {
|
|
|
4434
4367
|
contextWindow: 131072,
|
|
4435
4368
|
supportsImages: false,
|
|
4436
4369
|
supportsPromptCache: false,
|
|
4370
|
+
supportsNativeTools: true,
|
|
4437
4371
|
inputPrice: 0.075,
|
|
4438
4372
|
outputPrice: 0.3,
|
|
4439
4373
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4443,6 +4377,7 @@ var vertexModels = {
|
|
|
4443
4377
|
contextWindow: 262144,
|
|
4444
4378
|
supportsImages: false,
|
|
4445
4379
|
supportsPromptCache: false,
|
|
4380
|
+
supportsNativeTools: true,
|
|
4446
4381
|
inputPrice: 1,
|
|
4447
4382
|
outputPrice: 4,
|
|
4448
4383
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4452,6 +4387,7 @@ var vertexModels = {
|
|
|
4452
4387
|
contextWindow: 262144,
|
|
4453
4388
|
supportsImages: false,
|
|
4454
4389
|
supportsPromptCache: false,
|
|
4390
|
+
supportsNativeTools: true,
|
|
4455
4391
|
inputPrice: 0.25,
|
|
4456
4392
|
outputPrice: 1,
|
|
4457
4393
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
@@ -4892,6 +4828,7 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
4892
4828
|
contextWindow: 2e5,
|
|
4893
4829
|
supportsImages: true,
|
|
4894
4830
|
supportsPromptCache: true,
|
|
4831
|
+
supportsNativeTools: true,
|
|
4895
4832
|
inputPrice: 3,
|
|
4896
4833
|
outputPrice: 15,
|
|
4897
4834
|
cacheWritesPrice: 3.75,
|
|
@@ -4904,7 +4841,7 @@ var VERCEL_AI_GATEWAY_DEFAULT_TEMPERATURE = 0.7;
|
|
|
4904
4841
|
var internationalZAiDefaultModelId = "glm-4.6";
|
|
4905
4842
|
var internationalZAiModels = {
|
|
4906
4843
|
"glm-4.5": {
|
|
4907
|
-
maxTokens:
|
|
4844
|
+
maxTokens: 16384,
|
|
4908
4845
|
contextWindow: 131072,
|
|
4909
4846
|
supportsImages: false,
|
|
4910
4847
|
supportsPromptCache: true,
|
|
@@ -4917,7 +4854,7 @@ var internationalZAiModels = {
|
|
|
4917
4854
|
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
4918
4855
|
},
|
|
4919
4856
|
"glm-4.5-air": {
|
|
4920
|
-
maxTokens:
|
|
4857
|
+
maxTokens: 16384,
|
|
4921
4858
|
contextWindow: 131072,
|
|
4922
4859
|
supportsImages: false,
|
|
4923
4860
|
supportsPromptCache: true,
|
|
@@ -4930,7 +4867,7 @@ var internationalZAiModels = {
|
|
|
4930
4867
|
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
4931
4868
|
},
|
|
4932
4869
|
"glm-4.5-x": {
|
|
4933
|
-
maxTokens:
|
|
4870
|
+
maxTokens: 16384,
|
|
4934
4871
|
contextWindow: 131072,
|
|
4935
4872
|
supportsImages: false,
|
|
4936
4873
|
supportsPromptCache: true,
|
|
@@ -4943,7 +4880,7 @@ var internationalZAiModels = {
|
|
|
4943
4880
|
description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
|
|
4944
4881
|
},
|
|
4945
4882
|
"glm-4.5-airx": {
|
|
4946
|
-
maxTokens:
|
|
4883
|
+
maxTokens: 16384,
|
|
4947
4884
|
contextWindow: 131072,
|
|
4948
4885
|
supportsImages: false,
|
|
4949
4886
|
supportsPromptCache: true,
|
|
@@ -4956,7 +4893,7 @@ var internationalZAiModels = {
|
|
|
4956
4893
|
description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
|
|
4957
4894
|
},
|
|
4958
4895
|
"glm-4.5-flash": {
|
|
4959
|
-
maxTokens:
|
|
4896
|
+
maxTokens: 16384,
|
|
4960
4897
|
contextWindow: 131072,
|
|
4961
4898
|
supportsImages: false,
|
|
4962
4899
|
supportsPromptCache: true,
|
|
@@ -4982,7 +4919,7 @@ var internationalZAiModels = {
|
|
|
4982
4919
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
4983
4920
|
},
|
|
4984
4921
|
"glm-4.6": {
|
|
4985
|
-
maxTokens:
|
|
4922
|
+
maxTokens: 16384,
|
|
4986
4923
|
contextWindow: 2e5,
|
|
4987
4924
|
supportsImages: false,
|
|
4988
4925
|
supportsPromptCache: true,
|
|
@@ -4994,8 +4931,24 @@ var internationalZAiModels = {
|
|
|
4994
4931
|
cacheReadsPrice: 0.11,
|
|
4995
4932
|
description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
|
|
4996
4933
|
},
|
|
4934
|
+
"glm-4.7": {
|
|
4935
|
+
maxTokens: 16384,
|
|
4936
|
+
contextWindow: 2e5,
|
|
4937
|
+
supportsImages: false,
|
|
4938
|
+
supportsPromptCache: true,
|
|
4939
|
+
supportsNativeTools: true,
|
|
4940
|
+
defaultToolProtocol: "native",
|
|
4941
|
+
supportsReasoningEffort: ["disable", "medium"],
|
|
4942
|
+
reasoningEffort: "medium",
|
|
4943
|
+
preserveReasoning: true,
|
|
4944
|
+
inputPrice: 0.6,
|
|
4945
|
+
outputPrice: 2.2,
|
|
4946
|
+
cacheWritesPrice: 0,
|
|
4947
|
+
cacheReadsPrice: 0.11,
|
|
4948
|
+
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
4949
|
+
},
|
|
4997
4950
|
"glm-4-32b-0414-128k": {
|
|
4998
|
-
maxTokens:
|
|
4951
|
+
maxTokens: 16384,
|
|
4999
4952
|
contextWindow: 131072,
|
|
5000
4953
|
supportsImages: false,
|
|
5001
4954
|
supportsPromptCache: false,
|
|
@@ -5011,7 +4964,7 @@ var internationalZAiModels = {
|
|
|
5011
4964
|
var mainlandZAiDefaultModelId = "glm-4.6";
|
|
5012
4965
|
var mainlandZAiModels = {
|
|
5013
4966
|
"glm-4.5": {
|
|
5014
|
-
maxTokens:
|
|
4967
|
+
maxTokens: 16384,
|
|
5015
4968
|
contextWindow: 131072,
|
|
5016
4969
|
supportsImages: false,
|
|
5017
4970
|
supportsPromptCache: true,
|
|
@@ -5024,7 +4977,7 @@ var mainlandZAiModels = {
|
|
|
5024
4977
|
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
5025
4978
|
},
|
|
5026
4979
|
"glm-4.5-air": {
|
|
5027
|
-
maxTokens:
|
|
4980
|
+
maxTokens: 16384,
|
|
5028
4981
|
contextWindow: 131072,
|
|
5029
4982
|
supportsImages: false,
|
|
5030
4983
|
supportsPromptCache: true,
|
|
@@ -5037,7 +4990,7 @@ var mainlandZAiModels = {
|
|
|
5037
4990
|
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
5038
4991
|
},
|
|
5039
4992
|
"glm-4.5-x": {
|
|
5040
|
-
maxTokens:
|
|
4993
|
+
maxTokens: 16384,
|
|
5041
4994
|
contextWindow: 131072,
|
|
5042
4995
|
supportsImages: false,
|
|
5043
4996
|
supportsPromptCache: true,
|
|
@@ -5050,7 +5003,7 @@ var mainlandZAiModels = {
|
|
|
5050
5003
|
description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
|
|
5051
5004
|
},
|
|
5052
5005
|
"glm-4.5-airx": {
|
|
5053
|
-
maxTokens:
|
|
5006
|
+
maxTokens: 16384,
|
|
5054
5007
|
contextWindow: 131072,
|
|
5055
5008
|
supportsImages: false,
|
|
5056
5009
|
supportsPromptCache: true,
|
|
@@ -5063,7 +5016,7 @@ var mainlandZAiModels = {
|
|
|
5063
5016
|
description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
|
|
5064
5017
|
},
|
|
5065
5018
|
"glm-4.5-flash": {
|
|
5066
|
-
maxTokens:
|
|
5019
|
+
maxTokens: 16384,
|
|
5067
5020
|
contextWindow: 131072,
|
|
5068
5021
|
supportsImages: false,
|
|
5069
5022
|
supportsPromptCache: true,
|
|
@@ -5089,7 +5042,7 @@ var mainlandZAiModels = {
|
|
|
5089
5042
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
5090
5043
|
},
|
|
5091
5044
|
"glm-4.6": {
|
|
5092
|
-
maxTokens:
|
|
5045
|
+
maxTokens: 16384,
|
|
5093
5046
|
contextWindow: 204800,
|
|
5094
5047
|
supportsImages: false,
|
|
5095
5048
|
supportsPromptCache: true,
|
|
@@ -5100,6 +5053,22 @@ var mainlandZAiModels = {
|
|
|
5100
5053
|
cacheWritesPrice: 0,
|
|
5101
5054
|
cacheReadsPrice: 0.057,
|
|
5102
5055
|
description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
|
|
5056
|
+
},
|
|
5057
|
+
"glm-4.7": {
|
|
5058
|
+
maxTokens: 16384,
|
|
5059
|
+
contextWindow: 204800,
|
|
5060
|
+
supportsImages: false,
|
|
5061
|
+
supportsPromptCache: true,
|
|
5062
|
+
supportsNativeTools: true,
|
|
5063
|
+
defaultToolProtocol: "native",
|
|
5064
|
+
supportsReasoningEffort: ["disable", "medium"],
|
|
5065
|
+
reasoningEffort: "medium",
|
|
5066
|
+
preserveReasoning: true,
|
|
5067
|
+
inputPrice: 0.29,
|
|
5068
|
+
outputPrice: 1.14,
|
|
5069
|
+
cacheWritesPrice: 0,
|
|
5070
|
+
cacheReadsPrice: 0.057,
|
|
5071
|
+
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
5103
5072
|
}
|
|
5104
5073
|
};
|
|
5105
5074
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5149,6 +5118,8 @@ var minimaxModels = {
|
|
|
5149
5118
|
supportsPromptCache: true,
|
|
5150
5119
|
supportsNativeTools: true,
|
|
5151
5120
|
defaultToolProtocol: "native",
|
|
5121
|
+
includedTools: ["search_and_replace"],
|
|
5122
|
+
excludedTools: ["apply_diff"],
|
|
5152
5123
|
preserveReasoning: true,
|
|
5153
5124
|
inputPrice: 0.3,
|
|
5154
5125
|
outputPrice: 1.2,
|
|
@@ -5163,12 +5134,30 @@ var minimaxModels = {
|
|
|
5163
5134
|
supportsPromptCache: true,
|
|
5164
5135
|
supportsNativeTools: true,
|
|
5165
5136
|
defaultToolProtocol: "native",
|
|
5137
|
+
includedTools: ["search_and_replace"],
|
|
5138
|
+
excludedTools: ["apply_diff"],
|
|
5166
5139
|
preserveReasoning: true,
|
|
5167
5140
|
inputPrice: 0.3,
|
|
5168
5141
|
outputPrice: 1.2,
|
|
5169
5142
|
cacheWritesPrice: 0.375,
|
|
5170
5143
|
cacheReadsPrice: 0.03,
|
|
5171
5144
|
description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
|
|
5145
|
+
},
|
|
5146
|
+
"MiniMax-M2.1": {
|
|
5147
|
+
maxTokens: 16384,
|
|
5148
|
+
contextWindow: 192e3,
|
|
5149
|
+
supportsImages: false,
|
|
5150
|
+
supportsPromptCache: true,
|
|
5151
|
+
supportsNativeTools: true,
|
|
5152
|
+
defaultToolProtocol: "native",
|
|
5153
|
+
includedTools: ["search_and_replace"],
|
|
5154
|
+
excludedTools: ["apply_diff"],
|
|
5155
|
+
preserveReasoning: true,
|
|
5156
|
+
inputPrice: 0.3,
|
|
5157
|
+
outputPrice: 1.2,
|
|
5158
|
+
cacheWritesPrice: 0.375,
|
|
5159
|
+
cacheReadsPrice: 0.03,
|
|
5160
|
+
description: "MiniMax M2.1 builds on M2 with improved overall performance for agentic coding tasks and significantly faster response times."
|
|
5172
5161
|
}
|
|
5173
5162
|
};
|
|
5174
5163
|
var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
|
|
@@ -5350,8 +5339,7 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
|
5350
5339
|
openRouterApiKey: z8.string().optional(),
|
|
5351
5340
|
openRouterModelId: z8.string().optional(),
|
|
5352
5341
|
openRouterBaseUrl: z8.string().optional(),
|
|
5353
|
-
openRouterSpecificProvider: z8.string().optional()
|
|
5354
|
-
openRouterUseMiddleOutTransform: z8.boolean().optional()
|
|
5342
|
+
openRouterSpecificProvider: z8.string().optional()
|
|
5355
5343
|
});
|
|
5356
5344
|
var bedrockSchema = apiModelIdProviderModelSchema.extend({
|
|
5357
5345
|
awsAccessKey: z8.string().optional(),
|
|
@@ -5831,7 +5819,8 @@ var experimentIds = [
|
|
|
5831
5819
|
"preventFocusDisruption",
|
|
5832
5820
|
"imageGeneration",
|
|
5833
5821
|
"runSlashCommand",
|
|
5834
|
-
"multipleNativeToolCalls"
|
|
5822
|
+
"multipleNativeToolCalls",
|
|
5823
|
+
"customTools"
|
|
5835
5824
|
];
|
|
5836
5825
|
var experimentIdsSchema = z10.enum(experimentIds);
|
|
5837
5826
|
var experimentsSchema = z10.object({
|
|
@@ -5840,7 +5829,8 @@ var experimentsSchema = z10.object({
|
|
|
5840
5829
|
preventFocusDisruption: z10.boolean().optional(),
|
|
5841
5830
|
imageGeneration: z10.boolean().optional(),
|
|
5842
5831
|
runSlashCommand: z10.boolean().optional(),
|
|
5843
|
-
multipleNativeToolCalls: z10.boolean().optional()
|
|
5832
|
+
multipleNativeToolCalls: z10.boolean().optional(),
|
|
5833
|
+
customTools: z10.boolean().optional()
|
|
5844
5834
|
});
|
|
5845
5835
|
|
|
5846
5836
|
// src/telemetry.ts
|
|
@@ -6480,7 +6470,6 @@ var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].fil
|
|
|
6480
6470
|
var isGlobalStateKey = (key) => GLOBAL_STATE_KEYS.includes(key);
|
|
6481
6471
|
var EVALS_SETTINGS = {
|
|
6482
6472
|
apiProvider: "openrouter",
|
|
6483
|
-
openRouterUseMiddleOutTransform: false,
|
|
6484
6473
|
lastShownAnnouncementId: "jul-09-2025-3-23-0",
|
|
6485
6474
|
pinnedApiConfigs: {},
|
|
6486
6475
|
autoApprovalEnabled: true,
|
|
@@ -7159,11 +7148,6 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
|
|
|
7159
7148
|
})
|
|
7160
7149
|
]);
|
|
7161
7150
|
|
|
7162
|
-
// src/single-file-read-models.ts
|
|
7163
|
-
function shouldUseSingleFileRead(modelId) {
|
|
7164
|
-
return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
|
|
7165
|
-
}
|
|
7166
|
-
|
|
7167
7151
|
// src/todo.ts
|
|
7168
7152
|
import { z as z21 } from "zod";
|
|
7169
7153
|
var todoStatusSchema = z21.enum(["pending", "in_progress", "completed"]);
|
|
@@ -7471,7 +7455,6 @@ export {
|
|
|
7471
7455
|
serviceTiers,
|
|
7472
7456
|
shareResponseSchema,
|
|
7473
7457
|
shouldReportApiErrorToTelemetry,
|
|
7474
|
-
shouldUseSingleFileRead,
|
|
7475
7458
|
staticAppPropertiesSchema,
|
|
7476
7459
|
suggestionItemSchema,
|
|
7477
7460
|
taskBridgeCommandSchema,
|