@roo-code/types 1.94.0 → 1.96.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +107 -125
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +175 -201
- package/dist/index.d.ts +175 -201
- package/dist/index.js +107 -124
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.cjs
CHANGED
|
@@ -289,7 +289,6 @@ __export(index_exports, {
|
|
|
289
289
|
serviceTiers: () => serviceTiers,
|
|
290
290
|
shareResponseSchema: () => shareResponseSchema,
|
|
291
291
|
shouldReportApiErrorToTelemetry: () => shouldReportApiErrorToTelemetry,
|
|
292
|
-
shouldUseSingleFileRead: () => shouldUseSingleFileRead,
|
|
293
292
|
staticAppPropertiesSchema: () => staticAppPropertiesSchema,
|
|
294
293
|
suggestionItemSchema: () => suggestionItemSchema,
|
|
295
294
|
taskBridgeCommandSchema: () => taskBridgeCommandSchema,
|
|
@@ -498,7 +497,8 @@ var toolNames = [
|
|
|
498
497
|
"codebase_search",
|
|
499
498
|
"update_todo_list",
|
|
500
499
|
"run_slash_command",
|
|
501
|
-
"generate_image"
|
|
500
|
+
"generate_image",
|
|
501
|
+
"custom_tool"
|
|
502
502
|
];
|
|
503
503
|
var toolNamesSchema = import_zod2.z.enum(toolNames);
|
|
504
504
|
var toolUsageSchema = import_zod2.z.record(
|
|
@@ -1160,6 +1160,7 @@ var basetenModels = {
|
|
|
1160
1160
|
contextWindow: 163840,
|
|
1161
1161
|
supportsImages: false,
|
|
1162
1162
|
supportsPromptCache: false,
|
|
1163
|
+
supportsNativeTools: true,
|
|
1163
1164
|
inputPrice: 2.55,
|
|
1164
1165
|
outputPrice: 5.95,
|
|
1165
1166
|
cacheWritesPrice: 0,
|
|
@@ -1171,6 +1172,7 @@ var basetenModels = {
|
|
|
1171
1172
|
contextWindow: 163840,
|
|
1172
1173
|
supportsImages: false,
|
|
1173
1174
|
supportsPromptCache: false,
|
|
1175
|
+
supportsNativeTools: true,
|
|
1174
1176
|
inputPrice: 2.55,
|
|
1175
1177
|
outputPrice: 5.95,
|
|
1176
1178
|
cacheWritesPrice: 0,
|
|
@@ -1182,6 +1184,7 @@ var basetenModels = {
|
|
|
1182
1184
|
contextWindow: 163840,
|
|
1183
1185
|
supportsImages: false,
|
|
1184
1186
|
supportsPromptCache: false,
|
|
1187
|
+
supportsNativeTools: true,
|
|
1185
1188
|
inputPrice: 0.77,
|
|
1186
1189
|
outputPrice: 0.77,
|
|
1187
1190
|
cacheWritesPrice: 0,
|
|
@@ -1193,6 +1196,7 @@ var basetenModels = {
|
|
|
1193
1196
|
contextWindow: 163840,
|
|
1194
1197
|
supportsImages: false,
|
|
1195
1198
|
supportsPromptCache: false,
|
|
1199
|
+
supportsNativeTools: true,
|
|
1196
1200
|
inputPrice: 0.5,
|
|
1197
1201
|
outputPrice: 1.5,
|
|
1198
1202
|
cacheWritesPrice: 0,
|
|
@@ -1211,11 +1215,24 @@ var basetenModels = {
|
|
|
1211
1215
|
cacheReadsPrice: 0,
|
|
1212
1216
|
description: "DeepSeek's hybrid reasoning model with efficient long context scaling with GPT-5 level performance"
|
|
1213
1217
|
},
|
|
1218
|
+
"openai/gpt-oss-120b": {
|
|
1219
|
+
maxTokens: 16384,
|
|
1220
|
+
contextWindow: 128072,
|
|
1221
|
+
supportsImages: false,
|
|
1222
|
+
supportsPromptCache: false,
|
|
1223
|
+
supportsNativeTools: true,
|
|
1224
|
+
inputPrice: 0.1,
|
|
1225
|
+
outputPrice: 0.5,
|
|
1226
|
+
cacheWritesPrice: 0,
|
|
1227
|
+
cacheReadsPrice: 0,
|
|
1228
|
+
description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
|
|
1229
|
+
},
|
|
1214
1230
|
"Qwen/Qwen3-235B-A22B-Instruct-2507": {
|
|
1215
1231
|
maxTokens: 16384,
|
|
1216
1232
|
contextWindow: 262144,
|
|
1217
1233
|
supportsImages: false,
|
|
1218
1234
|
supportsPromptCache: false,
|
|
1235
|
+
supportsNativeTools: true,
|
|
1219
1236
|
inputPrice: 0.22,
|
|
1220
1237
|
outputPrice: 0.8,
|
|
1221
1238
|
cacheWritesPrice: 0,
|
|
@@ -1227,24 +1244,13 @@ var basetenModels = {
|
|
|
1227
1244
|
contextWindow: 262144,
|
|
1228
1245
|
supportsImages: false,
|
|
1229
1246
|
supportsPromptCache: false,
|
|
1247
|
+
supportsNativeTools: true,
|
|
1230
1248
|
inputPrice: 0.38,
|
|
1231
1249
|
outputPrice: 1.53,
|
|
1232
1250
|
cacheWritesPrice: 0,
|
|
1233
1251
|
cacheReadsPrice: 0,
|
|
1234
1252
|
description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
|
|
1235
1253
|
},
|
|
1236
|
-
"openai/gpt-oss-120b": {
|
|
1237
|
-
maxTokens: 16384,
|
|
1238
|
-
contextWindow: 128072,
|
|
1239
|
-
supportsImages: false,
|
|
1240
|
-
supportsPromptCache: false,
|
|
1241
|
-
supportsNativeTools: true,
|
|
1242
|
-
inputPrice: 0.1,
|
|
1243
|
-
outputPrice: 0.5,
|
|
1244
|
-
cacheWritesPrice: 0,
|
|
1245
|
-
cacheReadsPrice: 0,
|
|
1246
|
-
description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
|
|
1247
|
-
},
|
|
1248
1254
|
"moonshotai/Kimi-K2-Instruct-0905": {
|
|
1249
1255
|
maxTokens: 16384,
|
|
1250
1256
|
contextWindow: 262e3,
|
|
@@ -1716,22 +1722,6 @@ var bedrockModels = {
|
|
|
1716
1722
|
outputPrice: 0.6,
|
|
1717
1723
|
description: "Amazon Titan Text Express"
|
|
1718
1724
|
},
|
|
1719
|
-
"amazon.titan-text-embeddings-v1:0": {
|
|
1720
|
-
maxTokens: 8192,
|
|
1721
|
-
contextWindow: 8e3,
|
|
1722
|
-
supportsImages: false,
|
|
1723
|
-
supportsPromptCache: false,
|
|
1724
|
-
inputPrice: 0.1,
|
|
1725
|
-
description: "Amazon Titan Text Embeddings"
|
|
1726
|
-
},
|
|
1727
|
-
"amazon.titan-text-embeddings-v2:0": {
|
|
1728
|
-
maxTokens: 8192,
|
|
1729
|
-
contextWindow: 8e3,
|
|
1730
|
-
supportsImages: false,
|
|
1731
|
-
supportsPromptCache: false,
|
|
1732
|
-
inputPrice: 0.02,
|
|
1733
|
-
description: "Amazon Titan Text Embeddings V2"
|
|
1734
|
-
},
|
|
1735
1725
|
"moonshot.kimi-k2-thinking": {
|
|
1736
1726
|
maxTokens: 32e3,
|
|
1737
1727
|
contextWindow: 262144,
|
|
@@ -1865,7 +1855,7 @@ var BEDROCK_SERVICE_TIER_PRICING = {
|
|
|
1865
1855
|
var cerebrasDefaultModelId = "gpt-oss-120b";
|
|
1866
1856
|
var cerebrasModels = {
|
|
1867
1857
|
"zai-glm-4.6": {
|
|
1868
|
-
maxTokens:
|
|
1858
|
+
maxTokens: 16384,
|
|
1869
1859
|
// Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
|
|
1870
1860
|
contextWindow: 131072,
|
|
1871
1861
|
supportsImages: false,
|
|
@@ -1877,7 +1867,7 @@ var cerebrasModels = {
|
|
|
1877
1867
|
description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
|
|
1878
1868
|
},
|
|
1879
1869
|
"qwen-3-235b-a22b-instruct-2507": {
|
|
1880
|
-
maxTokens:
|
|
1870
|
+
maxTokens: 16384,
|
|
1881
1871
|
// Conservative default to avoid premature rate limiting
|
|
1882
1872
|
contextWindow: 64e3,
|
|
1883
1873
|
supportsImages: false,
|
|
@@ -1889,7 +1879,7 @@ var cerebrasModels = {
|
|
|
1889
1879
|
description: "Intelligent model with ~1400 tokens/s"
|
|
1890
1880
|
},
|
|
1891
1881
|
"llama-3.3-70b": {
|
|
1892
|
-
maxTokens:
|
|
1882
|
+
maxTokens: 16384,
|
|
1893
1883
|
// Conservative default to avoid premature rate limiting
|
|
1894
1884
|
contextWindow: 64e3,
|
|
1895
1885
|
supportsImages: false,
|
|
@@ -1901,7 +1891,7 @@ var cerebrasModels = {
|
|
|
1901
1891
|
description: "Powerful model with ~2600 tokens/s"
|
|
1902
1892
|
},
|
|
1903
1893
|
"qwen-3-32b": {
|
|
1904
|
-
maxTokens:
|
|
1894
|
+
maxTokens: 16384,
|
|
1905
1895
|
// Conservative default to avoid premature rate limiting
|
|
1906
1896
|
contextWindow: 64e3,
|
|
1907
1897
|
supportsImages: false,
|
|
@@ -1913,7 +1903,7 @@ var cerebrasModels = {
|
|
|
1913
1903
|
description: "SOTA coding performance with ~2500 tokens/s"
|
|
1914
1904
|
},
|
|
1915
1905
|
"gpt-oss-120b": {
|
|
1916
|
-
maxTokens:
|
|
1906
|
+
maxTokens: 16384,
|
|
1917
1907
|
// Conservative default to avoid premature rate limiting
|
|
1918
1908
|
contextWindow: 64e3,
|
|
1919
1909
|
supportsImages: false,
|
|
@@ -2551,6 +2541,7 @@ var featherlessModels = {
|
|
|
2551
2541
|
contextWindow: 32678,
|
|
2552
2542
|
supportsImages: false,
|
|
2553
2543
|
supportsPromptCache: false,
|
|
2544
|
+
supportsNativeTools: true,
|
|
2554
2545
|
inputPrice: 0,
|
|
2555
2546
|
outputPrice: 0,
|
|
2556
2547
|
description: "DeepSeek V3 0324 model."
|
|
@@ -2560,6 +2551,7 @@ var featherlessModels = {
|
|
|
2560
2551
|
contextWindow: 32678,
|
|
2561
2552
|
supportsImages: false,
|
|
2562
2553
|
supportsPromptCache: false,
|
|
2554
|
+
supportsNativeTools: true,
|
|
2563
2555
|
inputPrice: 0,
|
|
2564
2556
|
outputPrice: 0,
|
|
2565
2557
|
description: "DeepSeek R1 0528 model."
|
|
@@ -2579,6 +2571,7 @@ var featherlessModels = {
|
|
|
2579
2571
|
contextWindow: 32678,
|
|
2580
2572
|
supportsImages: false,
|
|
2581
2573
|
supportsPromptCache: false,
|
|
2574
|
+
supportsNativeTools: true,
|
|
2582
2575
|
inputPrice: 0,
|
|
2583
2576
|
outputPrice: 0,
|
|
2584
2577
|
description: "GPT-OSS 120B model."
|
|
@@ -2594,7 +2587,7 @@ var featherlessModels = {
|
|
|
2594
2587
|
description: "Qwen3 Coder 480B A35B Instruct model."
|
|
2595
2588
|
}
|
|
2596
2589
|
};
|
|
2597
|
-
var featherlessDefaultModelId = "
|
|
2590
|
+
var featherlessDefaultModelId = "moonshotai/Kimi-K2-Instruct";
|
|
2598
2591
|
|
|
2599
2592
|
// src/providers/fireworks.ts
|
|
2600
2593
|
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct-0905";
|
|
@@ -3039,33 +3032,6 @@ var groqModels = {
|
|
|
3039
3032
|
outputPrice: 0.34,
|
|
3040
3033
|
description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
|
|
3041
3034
|
},
|
|
3042
|
-
"meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
3043
|
-
maxTokens: 8192,
|
|
3044
|
-
contextWindow: 131072,
|
|
3045
|
-
supportsImages: false,
|
|
3046
|
-
supportsPromptCache: false,
|
|
3047
|
-
inputPrice: 0.2,
|
|
3048
|
-
outputPrice: 0.6,
|
|
3049
|
-
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
3050
|
-
},
|
|
3051
|
-
"mistral-saba-24b": {
|
|
3052
|
-
maxTokens: 8192,
|
|
3053
|
-
contextWindow: 32768,
|
|
3054
|
-
supportsImages: false,
|
|
3055
|
-
supportsPromptCache: false,
|
|
3056
|
-
inputPrice: 0.79,
|
|
3057
|
-
outputPrice: 0.79,
|
|
3058
|
-
description: "Mistral Saba 24B model, 32K context."
|
|
3059
|
-
},
|
|
3060
|
-
"qwen-qwq-32b": {
|
|
3061
|
-
maxTokens: 8192,
|
|
3062
|
-
contextWindow: 131072,
|
|
3063
|
-
supportsImages: false,
|
|
3064
|
-
supportsPromptCache: false,
|
|
3065
|
-
inputPrice: 0.29,
|
|
3066
|
-
outputPrice: 0.39,
|
|
3067
|
-
description: "Alibaba Qwen QwQ 32B model, 128K context."
|
|
3068
|
-
},
|
|
3069
3035
|
"qwen/qwen3-32b": {
|
|
3070
3036
|
maxTokens: 8192,
|
|
3071
3037
|
contextWindow: 131072,
|
|
@@ -3077,26 +3043,6 @@ var groqModels = {
|
|
|
3077
3043
|
outputPrice: 0.59,
|
|
3078
3044
|
description: "Alibaba Qwen 3 32B model, 128K context."
|
|
3079
3045
|
},
|
|
3080
|
-
"deepseek-r1-distill-llama-70b": {
|
|
3081
|
-
maxTokens: 8192,
|
|
3082
|
-
contextWindow: 131072,
|
|
3083
|
-
supportsImages: false,
|
|
3084
|
-
supportsPromptCache: false,
|
|
3085
|
-
inputPrice: 0.75,
|
|
3086
|
-
outputPrice: 0.99,
|
|
3087
|
-
description: "DeepSeek R1 Distill Llama 70B model, 128K context."
|
|
3088
|
-
},
|
|
3089
|
-
"moonshotai/kimi-k2-instruct": {
|
|
3090
|
-
maxTokens: 16384,
|
|
3091
|
-
contextWindow: 131072,
|
|
3092
|
-
supportsImages: false,
|
|
3093
|
-
supportsPromptCache: true,
|
|
3094
|
-
inputPrice: 1,
|
|
3095
|
-
outputPrice: 3,
|
|
3096
|
-
cacheReadsPrice: 0.5,
|
|
3097
|
-
// 50% discount for cached input tokens
|
|
3098
|
-
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
3099
|
-
},
|
|
3100
3046
|
"moonshotai/kimi-k2-instruct-0905": {
|
|
3101
3047
|
maxTokens: 16384,
|
|
3102
3048
|
contextWindow: 262144,
|
|
@@ -4171,15 +4117,6 @@ var sambaNovaModels = {
|
|
|
4171
4117
|
outputPrice: 4.5,
|
|
4172
4118
|
description: "DeepSeek V3.1 model with 32K context window."
|
|
4173
4119
|
},
|
|
4174
|
-
"DeepSeek-R1-Distill-Llama-70B": {
|
|
4175
|
-
maxTokens: 8192,
|
|
4176
|
-
contextWindow: 131072,
|
|
4177
|
-
supportsImages: false,
|
|
4178
|
-
supportsPromptCache: false,
|
|
4179
|
-
inputPrice: 0.7,
|
|
4180
|
-
outputPrice: 1.4,
|
|
4181
|
-
description: "DeepSeek R1 distilled Llama 70B model with 128K context window."
|
|
4182
|
-
},
|
|
4183
4120
|
"Llama-4-Maverick-17B-128E-Instruct": {
|
|
4184
4121
|
maxTokens: 8192,
|
|
4185
4122
|
contextWindow: 131072,
|
|
@@ -4191,15 +4128,6 @@ var sambaNovaModels = {
|
|
|
4191
4128
|
outputPrice: 1.8,
|
|
4192
4129
|
description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
|
|
4193
4130
|
},
|
|
4194
|
-
"Llama-3.3-Swallow-70B-Instruct-v0.4": {
|
|
4195
|
-
maxTokens: 8192,
|
|
4196
|
-
contextWindow: 16384,
|
|
4197
|
-
supportsImages: false,
|
|
4198
|
-
supportsPromptCache: false,
|
|
4199
|
-
inputPrice: 0.6,
|
|
4200
|
-
outputPrice: 1.2,
|
|
4201
|
-
description: "Tokyotech Llama 3.3 Swallow 70B Instruct v0.4 model with 16K context window."
|
|
4202
|
-
},
|
|
4203
4131
|
"Qwen3-32B": {
|
|
4204
4132
|
maxTokens: 8192,
|
|
4205
4133
|
contextWindow: 8192,
|
|
@@ -4732,6 +4660,7 @@ var vertexModels = {
|
|
|
4732
4660
|
contextWindow: 131072,
|
|
4733
4661
|
supportsImages: false,
|
|
4734
4662
|
supportsPromptCache: false,
|
|
4663
|
+
supportsNativeTools: true,
|
|
4735
4664
|
inputPrice: 0.35,
|
|
4736
4665
|
outputPrice: 1.15,
|
|
4737
4666
|
description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
|
|
@@ -4741,6 +4670,7 @@ var vertexModels = {
|
|
|
4741
4670
|
contextWindow: 163840,
|
|
4742
4671
|
supportsImages: false,
|
|
4743
4672
|
supportsPromptCache: false,
|
|
4673
|
+
supportsNativeTools: true,
|
|
4744
4674
|
inputPrice: 1.35,
|
|
4745
4675
|
outputPrice: 5.4,
|
|
4746
4676
|
description: "DeepSeek R1 (0528). Available in us-central1"
|
|
@@ -4750,6 +4680,7 @@ var vertexModels = {
|
|
|
4750
4680
|
contextWindow: 163840,
|
|
4751
4681
|
supportsImages: false,
|
|
4752
4682
|
supportsPromptCache: false,
|
|
4683
|
+
supportsNativeTools: true,
|
|
4753
4684
|
inputPrice: 0.6,
|
|
4754
4685
|
outputPrice: 1.7,
|
|
4755
4686
|
description: "DeepSeek V3.1. Available in us-west2"
|
|
@@ -4759,6 +4690,7 @@ var vertexModels = {
|
|
|
4759
4690
|
contextWindow: 131072,
|
|
4760
4691
|
supportsImages: false,
|
|
4761
4692
|
supportsPromptCache: false,
|
|
4693
|
+
supportsNativeTools: true,
|
|
4762
4694
|
inputPrice: 0.15,
|
|
4763
4695
|
outputPrice: 0.6,
|
|
4764
4696
|
description: "OpenAI gpt-oss 120B. Available in us-central1"
|
|
@@ -4768,6 +4700,7 @@ var vertexModels = {
|
|
|
4768
4700
|
contextWindow: 131072,
|
|
4769
4701
|
supportsImages: false,
|
|
4770
4702
|
supportsPromptCache: false,
|
|
4703
|
+
supportsNativeTools: true,
|
|
4771
4704
|
inputPrice: 0.075,
|
|
4772
4705
|
outputPrice: 0.3,
|
|
4773
4706
|
description: "OpenAI gpt-oss 20B. Available in us-central1"
|
|
@@ -4777,6 +4710,7 @@ var vertexModels = {
|
|
|
4777
4710
|
contextWindow: 262144,
|
|
4778
4711
|
supportsImages: false,
|
|
4779
4712
|
supportsPromptCache: false,
|
|
4713
|
+
supportsNativeTools: true,
|
|
4780
4714
|
inputPrice: 1,
|
|
4781
4715
|
outputPrice: 4,
|
|
4782
4716
|
description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
|
|
@@ -4786,6 +4720,7 @@ var vertexModels = {
|
|
|
4786
4720
|
contextWindow: 262144,
|
|
4787
4721
|
supportsImages: false,
|
|
4788
4722
|
supportsPromptCache: false,
|
|
4723
|
+
supportsNativeTools: true,
|
|
4789
4724
|
inputPrice: 0.25,
|
|
4790
4725
|
outputPrice: 1,
|
|
4791
4726
|
description: "Qwen3 235B A22B Instruct. Available in us-south1"
|
|
@@ -5226,6 +5161,7 @@ var vercelAiGatewayDefaultModelInfo = {
|
|
|
5226
5161
|
contextWindow: 2e5,
|
|
5227
5162
|
supportsImages: true,
|
|
5228
5163
|
supportsPromptCache: true,
|
|
5164
|
+
supportsNativeTools: true,
|
|
5229
5165
|
inputPrice: 3,
|
|
5230
5166
|
outputPrice: 15,
|
|
5231
5167
|
cacheWritesPrice: 3.75,
|
|
@@ -5238,7 +5174,7 @@ var VERCEL_AI_GATEWAY_DEFAULT_TEMPERATURE = 0.7;
|
|
|
5238
5174
|
var internationalZAiDefaultModelId = "glm-4.6";
|
|
5239
5175
|
var internationalZAiModels = {
|
|
5240
5176
|
"glm-4.5": {
|
|
5241
|
-
maxTokens:
|
|
5177
|
+
maxTokens: 16384,
|
|
5242
5178
|
contextWindow: 131072,
|
|
5243
5179
|
supportsImages: false,
|
|
5244
5180
|
supportsPromptCache: true,
|
|
@@ -5251,7 +5187,7 @@ var internationalZAiModels = {
|
|
|
5251
5187
|
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
5252
5188
|
},
|
|
5253
5189
|
"glm-4.5-air": {
|
|
5254
|
-
maxTokens:
|
|
5190
|
+
maxTokens: 16384,
|
|
5255
5191
|
contextWindow: 131072,
|
|
5256
5192
|
supportsImages: false,
|
|
5257
5193
|
supportsPromptCache: true,
|
|
@@ -5264,7 +5200,7 @@ var internationalZAiModels = {
|
|
|
5264
5200
|
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
5265
5201
|
},
|
|
5266
5202
|
"glm-4.5-x": {
|
|
5267
|
-
maxTokens:
|
|
5203
|
+
maxTokens: 16384,
|
|
5268
5204
|
contextWindow: 131072,
|
|
5269
5205
|
supportsImages: false,
|
|
5270
5206
|
supportsPromptCache: true,
|
|
@@ -5277,7 +5213,7 @@ var internationalZAiModels = {
|
|
|
5277
5213
|
description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
|
|
5278
5214
|
},
|
|
5279
5215
|
"glm-4.5-airx": {
|
|
5280
|
-
maxTokens:
|
|
5216
|
+
maxTokens: 16384,
|
|
5281
5217
|
contextWindow: 131072,
|
|
5282
5218
|
supportsImages: false,
|
|
5283
5219
|
supportsPromptCache: true,
|
|
@@ -5290,7 +5226,7 @@ var internationalZAiModels = {
|
|
|
5290
5226
|
description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
|
|
5291
5227
|
},
|
|
5292
5228
|
"glm-4.5-flash": {
|
|
5293
|
-
maxTokens:
|
|
5229
|
+
maxTokens: 16384,
|
|
5294
5230
|
contextWindow: 131072,
|
|
5295
5231
|
supportsImages: false,
|
|
5296
5232
|
supportsPromptCache: true,
|
|
@@ -5316,7 +5252,7 @@ var internationalZAiModels = {
|
|
|
5316
5252
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
5317
5253
|
},
|
|
5318
5254
|
"glm-4.6": {
|
|
5319
|
-
maxTokens:
|
|
5255
|
+
maxTokens: 16384,
|
|
5320
5256
|
contextWindow: 2e5,
|
|
5321
5257
|
supportsImages: false,
|
|
5322
5258
|
supportsPromptCache: true,
|
|
@@ -5328,8 +5264,24 @@ var internationalZAiModels = {
|
|
|
5328
5264
|
cacheReadsPrice: 0.11,
|
|
5329
5265
|
description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
|
|
5330
5266
|
},
|
|
5267
|
+
"glm-4.7": {
|
|
5268
|
+
maxTokens: 16384,
|
|
5269
|
+
contextWindow: 2e5,
|
|
5270
|
+
supportsImages: false,
|
|
5271
|
+
supportsPromptCache: true,
|
|
5272
|
+
supportsNativeTools: true,
|
|
5273
|
+
defaultToolProtocol: "native",
|
|
5274
|
+
supportsReasoningEffort: ["disable", "medium"],
|
|
5275
|
+
reasoningEffort: "medium",
|
|
5276
|
+
preserveReasoning: true,
|
|
5277
|
+
inputPrice: 0.6,
|
|
5278
|
+
outputPrice: 2.2,
|
|
5279
|
+
cacheWritesPrice: 0,
|
|
5280
|
+
cacheReadsPrice: 0.11,
|
|
5281
|
+
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
5282
|
+
},
|
|
5331
5283
|
"glm-4-32b-0414-128k": {
|
|
5332
|
-
maxTokens:
|
|
5284
|
+
maxTokens: 16384,
|
|
5333
5285
|
contextWindow: 131072,
|
|
5334
5286
|
supportsImages: false,
|
|
5335
5287
|
supportsPromptCache: false,
|
|
@@ -5345,7 +5297,7 @@ var internationalZAiModels = {
|
|
|
5345
5297
|
var mainlandZAiDefaultModelId = "glm-4.6";
|
|
5346
5298
|
var mainlandZAiModels = {
|
|
5347
5299
|
"glm-4.5": {
|
|
5348
|
-
maxTokens:
|
|
5300
|
+
maxTokens: 16384,
|
|
5349
5301
|
contextWindow: 131072,
|
|
5350
5302
|
supportsImages: false,
|
|
5351
5303
|
supportsPromptCache: true,
|
|
@@ -5358,7 +5310,7 @@ var mainlandZAiModels = {
|
|
|
5358
5310
|
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
5359
5311
|
},
|
|
5360
5312
|
"glm-4.5-air": {
|
|
5361
|
-
maxTokens:
|
|
5313
|
+
maxTokens: 16384,
|
|
5362
5314
|
contextWindow: 131072,
|
|
5363
5315
|
supportsImages: false,
|
|
5364
5316
|
supportsPromptCache: true,
|
|
@@ -5371,7 +5323,7 @@ var mainlandZAiModels = {
|
|
|
5371
5323
|
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
5372
5324
|
},
|
|
5373
5325
|
"glm-4.5-x": {
|
|
5374
|
-
maxTokens:
|
|
5326
|
+
maxTokens: 16384,
|
|
5375
5327
|
contextWindow: 131072,
|
|
5376
5328
|
supportsImages: false,
|
|
5377
5329
|
supportsPromptCache: true,
|
|
@@ -5384,7 +5336,7 @@ var mainlandZAiModels = {
|
|
|
5384
5336
|
description: "GLM-4.5-X is a high-performance variant optimized for strong reasoning with ultra-fast responses."
|
|
5385
5337
|
},
|
|
5386
5338
|
"glm-4.5-airx": {
|
|
5387
|
-
maxTokens:
|
|
5339
|
+
maxTokens: 16384,
|
|
5388
5340
|
contextWindow: 131072,
|
|
5389
5341
|
supportsImages: false,
|
|
5390
5342
|
supportsPromptCache: true,
|
|
@@ -5397,7 +5349,7 @@ var mainlandZAiModels = {
|
|
|
5397
5349
|
description: "GLM-4.5-AirX is a lightweight, ultra-fast variant delivering strong performance with lower cost."
|
|
5398
5350
|
},
|
|
5399
5351
|
"glm-4.5-flash": {
|
|
5400
|
-
maxTokens:
|
|
5352
|
+
maxTokens: 16384,
|
|
5401
5353
|
contextWindow: 131072,
|
|
5402
5354
|
supportsImages: false,
|
|
5403
5355
|
supportsPromptCache: true,
|
|
@@ -5423,7 +5375,7 @@ var mainlandZAiModels = {
|
|
|
5423
5375
|
description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
|
|
5424
5376
|
},
|
|
5425
5377
|
"glm-4.6": {
|
|
5426
|
-
maxTokens:
|
|
5378
|
+
maxTokens: 16384,
|
|
5427
5379
|
contextWindow: 204800,
|
|
5428
5380
|
supportsImages: false,
|
|
5429
5381
|
supportsPromptCache: true,
|
|
@@ -5434,6 +5386,22 @@ var mainlandZAiModels = {
|
|
|
5434
5386
|
cacheWritesPrice: 0,
|
|
5435
5387
|
cacheReadsPrice: 0.057,
|
|
5436
5388
|
description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
|
|
5389
|
+
},
|
|
5390
|
+
"glm-4.7": {
|
|
5391
|
+
maxTokens: 16384,
|
|
5392
|
+
contextWindow: 204800,
|
|
5393
|
+
supportsImages: false,
|
|
5394
|
+
supportsPromptCache: true,
|
|
5395
|
+
supportsNativeTools: true,
|
|
5396
|
+
defaultToolProtocol: "native",
|
|
5397
|
+
supportsReasoningEffort: ["disable", "medium"],
|
|
5398
|
+
reasoningEffort: "medium",
|
|
5399
|
+
preserveReasoning: true,
|
|
5400
|
+
inputPrice: 0.29,
|
|
5401
|
+
outputPrice: 1.14,
|
|
5402
|
+
cacheWritesPrice: 0,
|
|
5403
|
+
cacheReadsPrice: 0.057,
|
|
5404
|
+
description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
|
|
5437
5405
|
}
|
|
5438
5406
|
};
|
|
5439
5407
|
var ZAI_DEFAULT_TEMPERATURE = 0.6;
|
|
@@ -5483,6 +5451,8 @@ var minimaxModels = {
|
|
|
5483
5451
|
supportsPromptCache: true,
|
|
5484
5452
|
supportsNativeTools: true,
|
|
5485
5453
|
defaultToolProtocol: "native",
|
|
5454
|
+
includedTools: ["search_and_replace"],
|
|
5455
|
+
excludedTools: ["apply_diff"],
|
|
5486
5456
|
preserveReasoning: true,
|
|
5487
5457
|
inputPrice: 0.3,
|
|
5488
5458
|
outputPrice: 1.2,
|
|
@@ -5497,12 +5467,30 @@ var minimaxModels = {
|
|
|
5497
5467
|
supportsPromptCache: true,
|
|
5498
5468
|
supportsNativeTools: true,
|
|
5499
5469
|
defaultToolProtocol: "native",
|
|
5470
|
+
includedTools: ["search_and_replace"],
|
|
5471
|
+
excludedTools: ["apply_diff"],
|
|
5500
5472
|
preserveReasoning: true,
|
|
5501
5473
|
inputPrice: 0.3,
|
|
5502
5474
|
outputPrice: 1.2,
|
|
5503
5475
|
cacheWritesPrice: 0.375,
|
|
5504
5476
|
cacheReadsPrice: 0.03,
|
|
5505
5477
|
description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
|
|
5478
|
+
},
|
|
5479
|
+
"MiniMax-M2.1": {
|
|
5480
|
+
maxTokens: 16384,
|
|
5481
|
+
contextWindow: 192e3,
|
|
5482
|
+
supportsImages: false,
|
|
5483
|
+
supportsPromptCache: true,
|
|
5484
|
+
supportsNativeTools: true,
|
|
5485
|
+
defaultToolProtocol: "native",
|
|
5486
|
+
includedTools: ["search_and_replace"],
|
|
5487
|
+
excludedTools: ["apply_diff"],
|
|
5488
|
+
preserveReasoning: true,
|
|
5489
|
+
inputPrice: 0.3,
|
|
5490
|
+
outputPrice: 1.2,
|
|
5491
|
+
cacheWritesPrice: 0.375,
|
|
5492
|
+
cacheReadsPrice: 0.03,
|
|
5493
|
+
description: "MiniMax M2.1 builds on M2 with improved overall performance for agentic coding tasks and significantly faster response times."
|
|
5506
5494
|
}
|
|
5507
5495
|
};
|
|
5508
5496
|
var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
|
|
@@ -5684,8 +5672,7 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
|
|
|
5684
5672
|
openRouterApiKey: import_zod8.z.string().optional(),
|
|
5685
5673
|
openRouterModelId: import_zod8.z.string().optional(),
|
|
5686
5674
|
openRouterBaseUrl: import_zod8.z.string().optional(),
|
|
5687
|
-
openRouterSpecificProvider: import_zod8.z.string().optional()
|
|
5688
|
-
openRouterUseMiddleOutTransform: import_zod8.z.boolean().optional()
|
|
5675
|
+
openRouterSpecificProvider: import_zod8.z.string().optional()
|
|
5689
5676
|
});
|
|
5690
5677
|
var bedrockSchema = apiModelIdProviderModelSchema.extend({
|
|
5691
5678
|
awsAccessKey: import_zod8.z.string().optional(),
|
|
@@ -6165,7 +6152,8 @@ var experimentIds = [
|
|
|
6165
6152
|
"preventFocusDisruption",
|
|
6166
6153
|
"imageGeneration",
|
|
6167
6154
|
"runSlashCommand",
|
|
6168
|
-
"multipleNativeToolCalls"
|
|
6155
|
+
"multipleNativeToolCalls",
|
|
6156
|
+
"customTools"
|
|
6169
6157
|
];
|
|
6170
6158
|
var experimentIdsSchema = import_zod10.z.enum(experimentIds);
|
|
6171
6159
|
var experimentsSchema = import_zod10.z.object({
|
|
@@ -6174,7 +6162,8 @@ var experimentsSchema = import_zod10.z.object({
|
|
|
6174
6162
|
preventFocusDisruption: import_zod10.z.boolean().optional(),
|
|
6175
6163
|
imageGeneration: import_zod10.z.boolean().optional(),
|
|
6176
6164
|
runSlashCommand: import_zod10.z.boolean().optional(),
|
|
6177
|
-
multipleNativeToolCalls: import_zod10.z.boolean().optional()
|
|
6165
|
+
multipleNativeToolCalls: import_zod10.z.boolean().optional(),
|
|
6166
|
+
customTools: import_zod10.z.boolean().optional()
|
|
6178
6167
|
});
|
|
6179
6168
|
|
|
6180
6169
|
// src/telemetry.ts
|
|
@@ -6814,7 +6803,6 @@ var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].fil
|
|
|
6814
6803
|
var isGlobalStateKey = (key) => GLOBAL_STATE_KEYS.includes(key);
|
|
6815
6804
|
var EVALS_SETTINGS = {
|
|
6816
6805
|
apiProvider: "openrouter",
|
|
6817
|
-
openRouterUseMiddleOutTransform: false,
|
|
6818
6806
|
lastShownAnnouncementId: "jul-09-2025-3-23-0",
|
|
6819
6807
|
pinnedApiConfigs: {},
|
|
6820
6808
|
autoApprovalEnabled: true,
|
|
@@ -7493,11 +7481,6 @@ var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
|
|
|
7493
7481
|
})
|
|
7494
7482
|
]);
|
|
7495
7483
|
|
|
7496
|
-
// src/single-file-read-models.ts
|
|
7497
|
-
function shouldUseSingleFileRead(modelId) {
|
|
7498
|
-
return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
|
|
7499
|
-
}
|
|
7500
|
-
|
|
7501
7484
|
// src/todo.ts
|
|
7502
7485
|
var import_zod20 = require("zod");
|
|
7503
7486
|
var todoStatusSchema = import_zod20.z.enum(["pending", "in_progress", "completed"]);
|
|
@@ -7806,7 +7789,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
|
|
|
7806
7789
|
serviceTiers,
|
|
7807
7790
|
shareResponseSchema,
|
|
7808
7791
|
shouldReportApiErrorToTelemetry,
|
|
7809
|
-
shouldUseSingleFileRead,
|
|
7810
7792
|
staticAppPropertiesSchema,
|
|
7811
7793
|
suggestionItemSchema,
|
|
7812
7794
|
taskBridgeCommandSchema,
|