@roo-code/types 1.92.0 → 1.94.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -32,12 +32,12 @@ __export(index_exports, {
32
32
  BEDROCK_REGIONS: () => BEDROCK_REGIONS,
33
33
  BEDROCK_SERVICE_TIER_MODEL_IDS: () => BEDROCK_SERVICE_TIER_MODEL_IDS,
34
34
  BEDROCK_SERVICE_TIER_PRICING: () => BEDROCK_SERVICE_TIER_PRICING,
35
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS: () => CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
36
35
  CODEBASE_INDEX_DEFAULTS: () => CODEBASE_INDEX_DEFAULTS,
37
36
  CONSENT_COOKIE_NAME: () => CONSENT_COOKIE_NAME,
38
37
  CONTEXT_MANAGEMENT_EVENTS: () => CONTEXT_MANAGEMENT_EVENTS,
39
38
  COOKIE_CONSENT_EVENTS: () => COOKIE_CONSENT_EVENTS,
40
39
  ConnectionState: () => ConnectionState,
40
+ ConsecutiveMistakeError: () => ConsecutiveMistakeError,
41
41
  DEEP_SEEK_DEFAULT_TEMPERATURE: () => DEEP_SEEK_DEFAULT_TEMPERATURE,
42
42
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS: () => DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
43
43
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT: () => DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
@@ -78,6 +78,7 @@ __export(index_exports, {
78
78
  MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
79
79
  MODELS_BY_PROVIDER: () => MODELS_BY_PROVIDER,
80
80
  MOONSHOT_DEFAULT_TEMPERATURE: () => MOONSHOT_DEFAULT_TEMPERATURE,
81
+ NATIVE_TOOL_DEFAULTS: () => NATIVE_TOOL_DEFAULTS,
81
82
  OPENAI_AZURE_AI_INFERENCE_PATH: () => OPENAI_AZURE_AI_INFERENCE_PATH,
82
83
  OPENAI_NATIVE_DEFAULT_TEMPERATURE: () => OPENAI_NATIVE_DEFAULT_TEMPERATURE,
83
84
  OPENROUTER_DEFAULT_PROVIDER_NAME: () => OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -103,6 +104,7 @@ __export(index_exports, {
103
104
  VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS: () => VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS,
104
105
  VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS: () => VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS,
105
106
  VERCEL_AI_GATEWAY_VISION_ONLY_MODELS: () => VERCEL_AI_GATEWAY_VISION_ONLY_MODELS,
107
+ VERTEX_1M_CONTEXT_MODEL_IDS: () => VERTEX_1M_CONTEXT_MODEL_IDS,
106
108
  VERTEX_REGIONS: () => VERTEX_REGIONS,
107
109
  ZAI_DEFAULT_TEMPERATURE: () => ZAI_DEFAULT_TEMPERATURE,
108
110
  ackSchema: () => ackSchema,
@@ -122,6 +124,7 @@ __export(index_exports, {
122
124
  chutesModels: () => chutesModels,
123
125
  claudeCodeDefaultModelId: () => claudeCodeDefaultModelId,
124
126
  claudeCodeModels: () => claudeCodeModels,
127
+ claudeCodeReasoningConfig: () => claudeCodeReasoningConfig,
125
128
  clineAskSchema: () => clineAskSchema,
126
129
  clineAsks: () => clineAsks,
127
130
  clineMessageSchema: () => clineMessageSchema,
@@ -136,7 +139,6 @@ __export(index_exports, {
136
139
  commandIds: () => commandIds,
137
140
  contextCondenseSchema: () => contextCondenseSchema,
138
141
  contextTruncationSchema: () => contextTruncationSchema,
139
- convertModelNameForVertex: () => convertModelNameForVertex,
140
142
  customModePromptsSchema: () => customModePromptsSchema,
141
143
  customModesSettingsSchema: () => customModesSettingsSchema,
142
144
  customProviders: () => customProviders,
@@ -145,6 +147,7 @@ __export(index_exports, {
145
147
  deepInfraDefaultModelInfo: () => deepInfraDefaultModelInfo,
146
148
  deepSeekDefaultModelId: () => deepSeekDefaultModelId,
147
149
  deepSeekModels: () => deepSeekModels,
150
+ defineCustomTool: () => defineCustomTool,
148
151
  discriminatedProviderSettingsWithIdSchema: () => discriminatedProviderSettingsWithIdSchema,
149
152
  doubaoDefaultModelId: () => doubaoDefaultModelId,
150
153
  doubaoDefaultModelInfo: () => doubaoDefaultModelInfo,
@@ -158,6 +161,8 @@ __export(index_exports, {
158
161
  extensionBridgeEventSchema: () => extensionBridgeEventSchema,
159
162
  extensionInstanceSchema: () => extensionInstanceSchema,
160
163
  extractApiProviderErrorProperties: () => extractApiProviderErrorProperties,
164
+ extractConsecutiveMistakeErrorProperties: () => extractConsecutiveMistakeErrorProperties,
165
+ extractMessageFromJsonPayload: () => extractMessageFromJsonPayload,
161
166
  fauxProviders: () => fauxProviders,
162
167
  featherlessDefaultModelId: () => featherlessDefaultModelId,
163
168
  featherlessModels: () => featherlessModels,
@@ -167,7 +172,6 @@ __export(index_exports, {
167
172
  geminiDefaultModelId: () => geminiDefaultModelId,
168
173
  geminiModels: () => geminiModels,
169
174
  getApiProtocol: () => getApiProtocol,
170
- getClaudeCodeModelId: () => getClaudeCodeModelId,
171
175
  getEffectiveProtocol: () => getEffectiveProtocol,
172
176
  getErrorMessage: () => getErrorMessage,
173
177
  getErrorStatusCode: () => getErrorStatusCode,
@@ -192,6 +196,7 @@ __export(index_exports, {
192
196
  ioIntelligenceModels: () => ioIntelligenceModels,
193
197
  ipcMessageSchema: () => ipcMessageSchema,
194
198
  isApiProviderError: () => isApiProviderError,
199
+ isConsecutiveMistakeError: () => isConsecutiveMistakeError,
195
200
  isContextManagementEvent: () => isContextManagementEvent,
196
201
  isCustomProvider: () => isCustomProvider,
197
202
  isDynamicProvider: () => isDynamicProvider,
@@ -239,6 +244,7 @@ __export(index_exports, {
239
244
  moonshotDefaultModelId: () => moonshotDefaultModelId,
240
245
  moonshotModels: () => moonshotModels,
241
246
  nonBlockingAsks: () => nonBlockingAsks,
247
+ normalizeClaudeCodeModelId: () => normalizeClaudeCodeModelId,
242
248
  ollamaDefaultModelId: () => ollamaDefaultModelId,
243
249
  ollamaDefaultModelInfo: () => ollamaDefaultModelInfo,
244
250
  openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
@@ -251,6 +257,7 @@ __export(index_exports, {
251
257
  organizationDefaultSettingsSchema: () => organizationDefaultSettingsSchema,
252
258
  organizationFeaturesSchema: () => organizationFeaturesSchema,
253
259
  organizationSettingsSchema: () => organizationSettingsSchema,
260
+ parametersSchema: () => import_v4.z,
254
261
  promptComponentSchema: () => promptComponentSchema,
255
262
  providerNames: () => providerNames,
256
263
  providerNamesSchema: () => providerNamesSchema,
@@ -476,6 +483,7 @@ var toolNames = [
476
483
  "apply_diff",
477
484
  "search_and_replace",
478
485
  "search_replace",
486
+ "edit_file",
479
487
  "apply_patch",
480
488
  "search_files",
481
489
  "list_files",
@@ -504,6 +512,10 @@ var TOOL_PROTOCOL = {
504
512
  XML: "xml",
505
513
  NATIVE: "native"
506
514
  };
515
+ var NATIVE_TOOL_DEFAULTS = {
516
+ supportsNativeTools: true,
517
+ defaultToolProtocol: TOOL_PROTOCOL.NATIVE
518
+ };
507
519
  function isNativeProtocol(protocol) {
508
520
  return protocol === TOOL_PROTOCOL.NATIVE;
509
521
  }
@@ -898,6 +910,7 @@ var anthropicModels = {
898
910
  supportsImages: true,
899
911
  supportsPromptCache: true,
900
912
  supportsNativeTools: true,
913
+ defaultToolProtocol: "native",
901
914
  inputPrice: 3,
902
915
  // $3 per million input tokens (≤200K context)
903
916
  outputPrice: 15,
@@ -931,6 +944,7 @@ var anthropicModels = {
931
944
  supportsImages: true,
932
945
  supportsPromptCache: true,
933
946
  supportsNativeTools: true,
947
+ defaultToolProtocol: "native",
934
948
  inputPrice: 3,
935
949
  // $3 per million input tokens (≤200K context)
936
950
  outputPrice: 15,
@@ -963,6 +977,7 @@ var anthropicModels = {
963
977
  supportsImages: true,
964
978
  supportsPromptCache: true,
965
979
  supportsNativeTools: true,
980
+ defaultToolProtocol: "native",
966
981
  inputPrice: 5,
967
982
  // $5 per million input tokens
968
983
  outputPrice: 25,
@@ -980,6 +995,7 @@ var anthropicModels = {
980
995
  supportsImages: true,
981
996
  supportsPromptCache: true,
982
997
  supportsNativeTools: true,
998
+ defaultToolProtocol: "native",
983
999
  inputPrice: 15,
984
1000
  // $15 per million input tokens
985
1001
  outputPrice: 75,
@@ -997,6 +1013,7 @@ var anthropicModels = {
997
1013
  supportsImages: true,
998
1014
  supportsPromptCache: true,
999
1015
  supportsNativeTools: true,
1016
+ defaultToolProtocol: "native",
1000
1017
  inputPrice: 15,
1001
1018
  // $15 per million input tokens
1002
1019
  outputPrice: 75,
@@ -1014,6 +1031,7 @@ var anthropicModels = {
1014
1031
  supportsImages: true,
1015
1032
  supportsPromptCache: true,
1016
1033
  supportsNativeTools: true,
1034
+ defaultToolProtocol: "native",
1017
1035
  inputPrice: 3,
1018
1036
  // $3 per million input tokens
1019
1037
  outputPrice: 15,
@@ -1032,6 +1050,7 @@ var anthropicModels = {
1032
1050
  supportsImages: true,
1033
1051
  supportsPromptCache: true,
1034
1052
  supportsNativeTools: true,
1053
+ defaultToolProtocol: "native",
1035
1054
  inputPrice: 3,
1036
1055
  // $3 per million input tokens
1037
1056
  outputPrice: 15,
@@ -1047,6 +1066,7 @@ var anthropicModels = {
1047
1066
  supportsImages: true,
1048
1067
  supportsPromptCache: true,
1049
1068
  supportsNativeTools: true,
1069
+ defaultToolProtocol: "native",
1050
1070
  inputPrice: 3,
1051
1071
  // $3 per million input tokens
1052
1072
  outputPrice: 15,
@@ -1062,6 +1082,7 @@ var anthropicModels = {
1062
1082
  supportsImages: false,
1063
1083
  supportsPromptCache: true,
1064
1084
  supportsNativeTools: true,
1085
+ defaultToolProtocol: "native",
1065
1086
  inputPrice: 1,
1066
1087
  outputPrice: 5,
1067
1088
  cacheWritesPrice: 1.25,
@@ -1073,6 +1094,7 @@ var anthropicModels = {
1073
1094
  supportsImages: true,
1074
1095
  supportsPromptCache: true,
1075
1096
  supportsNativeTools: true,
1097
+ defaultToolProtocol: "native",
1076
1098
  inputPrice: 15,
1077
1099
  outputPrice: 75,
1078
1100
  cacheWritesPrice: 18.75,
@@ -1084,6 +1106,7 @@ var anthropicModels = {
1084
1106
  supportsImages: true,
1085
1107
  supportsPromptCache: true,
1086
1108
  supportsNativeTools: true,
1109
+ defaultToolProtocol: "native",
1087
1110
  inputPrice: 0.25,
1088
1111
  outputPrice: 1.25,
1089
1112
  cacheWritesPrice: 0.3,
@@ -1095,6 +1118,7 @@ var anthropicModels = {
1095
1118
  supportsImages: true,
1096
1119
  supportsPromptCache: true,
1097
1120
  supportsNativeTools: true,
1121
+ defaultToolProtocol: "native",
1098
1122
  inputPrice: 1,
1099
1123
  outputPrice: 5,
1100
1124
  cacheWritesPrice: 1.25,
@@ -1247,6 +1271,7 @@ var bedrockModels = {
1247
1271
  supportsPromptCache: true,
1248
1272
  supportsReasoningBudget: true,
1249
1273
  supportsNativeTools: true,
1274
+ defaultToolProtocol: "native",
1250
1275
  inputPrice: 3,
1251
1276
  outputPrice: 15,
1252
1277
  cacheWritesPrice: 3.75,
@@ -1340,6 +1365,7 @@ var bedrockModels = {
1340
1365
  supportsPromptCache: true,
1341
1366
  supportsReasoningBudget: true,
1342
1367
  supportsNativeTools: true,
1368
+ defaultToolProtocol: "native",
1343
1369
  inputPrice: 3,
1344
1370
  outputPrice: 15,
1345
1371
  cacheWritesPrice: 3.75,
@@ -1355,6 +1381,7 @@ var bedrockModels = {
1355
1381
  supportsPromptCache: true,
1356
1382
  supportsReasoningBudget: true,
1357
1383
  supportsNativeTools: true,
1384
+ defaultToolProtocol: "native",
1358
1385
  inputPrice: 15,
1359
1386
  outputPrice: 75,
1360
1387
  cacheWritesPrice: 18.75,
@@ -1370,6 +1397,7 @@ var bedrockModels = {
1370
1397
  supportsPromptCache: true,
1371
1398
  supportsReasoningBudget: true,
1372
1399
  supportsNativeTools: true,
1400
+ defaultToolProtocol: "native",
1373
1401
  inputPrice: 5,
1374
1402
  outputPrice: 25,
1375
1403
  cacheWritesPrice: 6.25,
@@ -1385,6 +1413,7 @@ var bedrockModels = {
1385
1413
  supportsPromptCache: true,
1386
1414
  supportsReasoningBudget: true,
1387
1415
  supportsNativeTools: true,
1416
+ defaultToolProtocol: "native",
1388
1417
  inputPrice: 15,
1389
1418
  outputPrice: 75,
1390
1419
  cacheWritesPrice: 18.75,
@@ -1400,6 +1429,7 @@ var bedrockModels = {
1400
1429
  supportsPromptCache: true,
1401
1430
  supportsReasoningBudget: true,
1402
1431
  supportsNativeTools: true,
1432
+ defaultToolProtocol: "native",
1403
1433
  inputPrice: 3,
1404
1434
  outputPrice: 15,
1405
1435
  cacheWritesPrice: 3.75,
@@ -1414,6 +1444,7 @@ var bedrockModels = {
1414
1444
  supportsImages: true,
1415
1445
  supportsPromptCache: true,
1416
1446
  supportsNativeTools: true,
1447
+ defaultToolProtocol: "native",
1417
1448
  inputPrice: 3,
1418
1449
  outputPrice: 15,
1419
1450
  cacheWritesPrice: 3.75,
@@ -1428,6 +1459,7 @@ var bedrockModels = {
1428
1459
  supportsImages: false,
1429
1460
  supportsPromptCache: true,
1430
1461
  supportsNativeTools: true,
1462
+ defaultToolProtocol: "native",
1431
1463
  inputPrice: 0.8,
1432
1464
  outputPrice: 4,
1433
1465
  cacheWritesPrice: 1,
@@ -1443,6 +1475,7 @@ var bedrockModels = {
1443
1475
  supportsPromptCache: true,
1444
1476
  supportsReasoningBudget: true,
1445
1477
  supportsNativeTools: true,
1478
+ defaultToolProtocol: "native",
1446
1479
  inputPrice: 1,
1447
1480
  outputPrice: 5,
1448
1481
  cacheWritesPrice: 1.25,
@@ -1459,6 +1492,7 @@ var bedrockModels = {
1459
1492
  supportsImages: true,
1460
1493
  supportsPromptCache: false,
1461
1494
  supportsNativeTools: true,
1495
+ defaultToolProtocol: "native",
1462
1496
  inputPrice: 3,
1463
1497
  outputPrice: 15
1464
1498
  },
@@ -1468,6 +1502,7 @@ var bedrockModels = {
1468
1502
  supportsImages: true,
1469
1503
  supportsPromptCache: false,
1470
1504
  supportsNativeTools: true,
1505
+ defaultToolProtocol: "native",
1471
1506
  inputPrice: 15,
1472
1507
  outputPrice: 75
1473
1508
  },
@@ -1477,6 +1512,7 @@ var bedrockModels = {
1477
1512
  supportsImages: true,
1478
1513
  supportsPromptCache: false,
1479
1514
  supportsNativeTools: true,
1515
+ defaultToolProtocol: "native",
1480
1516
  inputPrice: 3,
1481
1517
  outputPrice: 15
1482
1518
  },
@@ -1486,6 +1522,7 @@ var bedrockModels = {
1486
1522
  supportsImages: true,
1487
1523
  supportsPromptCache: false,
1488
1524
  supportsNativeTools: true,
1525
+ defaultToolProtocol: "native",
1489
1526
  inputPrice: 0.25,
1490
1527
  outputPrice: 1.25
1491
1528
  },
@@ -1495,6 +1532,7 @@ var bedrockModels = {
1495
1532
  supportsImages: false,
1496
1533
  supportsPromptCache: false,
1497
1534
  supportsNativeTools: true,
1535
+ defaultToolProtocol: "native",
1498
1536
  inputPrice: 8,
1499
1537
  outputPrice: 24,
1500
1538
  description: "Claude 2.1"
@@ -1505,6 +1543,7 @@ var bedrockModels = {
1505
1543
  supportsImages: false,
1506
1544
  supportsPromptCache: false,
1507
1545
  supportsNativeTools: true,
1546
+ defaultToolProtocol: "native",
1508
1547
  inputPrice: 8,
1509
1548
  outputPrice: 24,
1510
1549
  description: "Claude 2.0"
@@ -1515,6 +1554,7 @@ var bedrockModels = {
1515
1554
  supportsImages: false,
1516
1555
  supportsPromptCache: false,
1517
1556
  supportsNativeTools: true,
1557
+ defaultToolProtocol: "native",
1518
1558
  inputPrice: 0.8,
1519
1559
  outputPrice: 2.4,
1520
1560
  description: "Claude Instant"
@@ -1831,6 +1871,7 @@ var cerebrasModels = {
1831
1871
  supportsImages: false,
1832
1872
  supportsPromptCache: false,
1833
1873
  supportsNativeTools: true,
1874
+ defaultToolProtocol: "native",
1834
1875
  inputPrice: 0,
1835
1876
  outputPrice: 0,
1836
1877
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
@@ -1842,6 +1883,7 @@ var cerebrasModels = {
1842
1883
  supportsImages: false,
1843
1884
  supportsPromptCache: false,
1844
1885
  supportsNativeTools: true,
1886
+ defaultToolProtocol: "native",
1845
1887
  inputPrice: 0,
1846
1888
  outputPrice: 0,
1847
1889
  description: "Intelligent model with ~1400 tokens/s"
@@ -1853,6 +1895,7 @@ var cerebrasModels = {
1853
1895
  supportsImages: false,
1854
1896
  supportsPromptCache: false,
1855
1897
  supportsNativeTools: true,
1898
+ defaultToolProtocol: "native",
1856
1899
  inputPrice: 0,
1857
1900
  outputPrice: 0,
1858
1901
  description: "Powerful model with ~2600 tokens/s"
@@ -1864,6 +1907,7 @@ var cerebrasModels = {
1864
1907
  supportsImages: false,
1865
1908
  supportsPromptCache: false,
1866
1909
  supportsNativeTools: true,
1910
+ defaultToolProtocol: "native",
1867
1911
  inputPrice: 0,
1868
1912
  outputPrice: 0,
1869
1913
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1875,6 +1919,7 @@ var cerebrasModels = {
1875
1919
  supportsImages: false,
1876
1920
  supportsPromptCache: false,
1877
1921
  supportsNativeTools: true,
1922
+ defaultToolProtocol: "native",
1878
1923
  inputPrice: 0,
1879
1924
  outputPrice: 0,
1880
1925
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1889,6 +1934,8 @@ var chutesModels = {
1889
1934
  contextWindow: 163840,
1890
1935
  supportsImages: false,
1891
1936
  supportsPromptCache: false,
1937
+ supportsNativeTools: true,
1938
+ defaultToolProtocol: "native",
1892
1939
  inputPrice: 0,
1893
1940
  outputPrice: 0,
1894
1941
  description: "DeepSeek R1 0528 model."
@@ -1898,6 +1945,8 @@ var chutesModels = {
1898
1945
  contextWindow: 163840,
1899
1946
  supportsImages: false,
1900
1947
  supportsPromptCache: false,
1948
+ supportsNativeTools: true,
1949
+ defaultToolProtocol: "native",
1901
1950
  inputPrice: 0,
1902
1951
  outputPrice: 0,
1903
1952
  description: "DeepSeek R1 model."
@@ -1907,6 +1956,8 @@ var chutesModels = {
1907
1956
  contextWindow: 163840,
1908
1957
  supportsImages: false,
1909
1958
  supportsPromptCache: false,
1959
+ supportsNativeTools: true,
1960
+ defaultToolProtocol: "native",
1910
1961
  inputPrice: 0,
1911
1962
  outputPrice: 0,
1912
1963
  description: "DeepSeek V3 model."
@@ -1916,6 +1967,8 @@ var chutesModels = {
1916
1967
  contextWindow: 163840,
1917
1968
  supportsImages: false,
1918
1969
  supportsPromptCache: false,
1970
+ supportsNativeTools: true,
1971
+ defaultToolProtocol: "native",
1919
1972
  inputPrice: 0,
1920
1973
  outputPrice: 0,
1921
1974
  description: "DeepSeek V3.1 model."
@@ -1925,6 +1978,8 @@ var chutesModels = {
1925
1978
  contextWindow: 163840,
1926
1979
  supportsImages: false,
1927
1980
  supportsPromptCache: false,
1981
+ supportsNativeTools: true,
1982
+ defaultToolProtocol: "native",
1928
1983
  inputPrice: 0.23,
1929
1984
  outputPrice: 0.9,
1930
1985
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1934,6 +1989,8 @@ var chutesModels = {
1934
1989
  contextWindow: 163840,
1935
1990
  supportsImages: false,
1936
1991
  supportsPromptCache: false,
1992
+ supportsNativeTools: true,
1993
+ defaultToolProtocol: "native",
1937
1994
  inputPrice: 1,
1938
1995
  outputPrice: 3,
1939
1996
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1943,6 +2000,8 @@ var chutesModels = {
1943
2000
  contextWindow: 163840,
1944
2001
  supportsImages: false,
1945
2002
  supportsPromptCache: false,
2003
+ supportsNativeTools: true,
2004
+ defaultToolProtocol: "native",
1946
2005
  inputPrice: 0.25,
1947
2006
  outputPrice: 0.35,
1948
2007
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1954,6 +2013,8 @@ var chutesModels = {
1954
2013
  // From Groq
1955
2014
  supportsImages: false,
1956
2015
  supportsPromptCache: false,
2016
+ supportsNativeTools: true,
2017
+ defaultToolProtocol: "native",
1957
2018
  inputPrice: 0,
1958
2019
  outputPrice: 0,
1959
2020
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1963,6 +2024,8 @@ var chutesModels = {
1963
2024
  contextWindow: 512e3,
1964
2025
  supportsImages: false,
1965
2026
  supportsPromptCache: false,
2027
+ supportsNativeTools: true,
2028
+ defaultToolProtocol: "native",
1966
2029
  inputPrice: 0,
1967
2030
  outputPrice: 0,
1968
2031
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1972,6 +2035,8 @@ var chutesModels = {
1972
2035
  contextWindow: 128e3,
1973
2036
  supportsImages: false,
1974
2037
  supportsPromptCache: false,
2038
+ supportsNativeTools: true,
2039
+ defaultToolProtocol: "native",
1975
2040
  inputPrice: 0,
1976
2041
  outputPrice: 0,
1977
2042
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1981,6 +2046,8 @@ var chutesModels = {
1981
2046
  contextWindow: 131072,
1982
2047
  supportsImages: false,
1983
2048
  supportsPromptCache: false,
2049
+ supportsNativeTools: true,
2050
+ defaultToolProtocol: "native",
1984
2051
  inputPrice: 0,
1985
2052
  outputPrice: 0,
1986
2053
  description: "Unsloth Gemma 3 12B IT model."
@@ -1990,6 +2057,8 @@ var chutesModels = {
1990
2057
  contextWindow: 131072,
1991
2058
  supportsImages: false,
1992
2059
  supportsPromptCache: false,
2060
+ supportsNativeTools: true,
2061
+ defaultToolProtocol: "native",
1993
2062
  inputPrice: 0,
1994
2063
  outputPrice: 0,
1995
2064
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1999,6 +2068,8 @@ var chutesModels = {
1999
2068
  contextWindow: 131072,
2000
2069
  supportsImages: false,
2001
2070
  supportsPromptCache: false,
2071
+ supportsNativeTools: true,
2072
+ defaultToolProtocol: "native",
2002
2073
  inputPrice: 0,
2003
2074
  outputPrice: 0,
2004
2075
  description: "Unsloth Gemma 3 4B IT model."
@@ -2008,6 +2079,8 @@ var chutesModels = {
2008
2079
  contextWindow: 131072,
2009
2080
  supportsImages: false,
2010
2081
  supportsPromptCache: false,
2082
+ supportsNativeTools: true,
2083
+ defaultToolProtocol: "native",
2011
2084
  inputPrice: 0,
2012
2085
  outputPrice: 0,
2013
2086
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -2017,6 +2090,8 @@ var chutesModels = {
2017
2090
  contextWindow: 131072,
2018
2091
  supportsImages: false,
2019
2092
  supportsPromptCache: false,
2093
+ supportsNativeTools: true,
2094
+ defaultToolProtocol: "native",
2020
2095
  inputPrice: 0,
2021
2096
  outputPrice: 0,
2022
2097
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -2026,6 +2101,8 @@ var chutesModels = {
2026
2101
  contextWindow: 256e3,
2027
2102
  supportsImages: false,
2028
2103
  supportsPromptCache: false,
2104
+ supportsNativeTools: true,
2105
+ defaultToolProtocol: "native",
2029
2106
  inputPrice: 0,
2030
2107
  outputPrice: 0,
2031
2108
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -2035,6 +2112,8 @@ var chutesModels = {
2035
2112
  contextWindow: 163840,
2036
2113
  supportsImages: false,
2037
2114
  supportsPromptCache: false,
2115
+ supportsNativeTools: true,
2116
+ defaultToolProtocol: "native",
2038
2117
  inputPrice: 0,
2039
2118
  outputPrice: 0,
2040
2119
  description: "DeepSeek V3 Base model."
@@ -2044,6 +2123,8 @@ var chutesModels = {
2044
2123
  contextWindow: 163840,
2045
2124
  supportsImages: false,
2046
2125
  supportsPromptCache: false,
2126
+ supportsNativeTools: true,
2127
+ defaultToolProtocol: "native",
2047
2128
  inputPrice: 0,
2048
2129
  outputPrice: 0,
2049
2130
  description: "DeepSeek R1 Zero model."
@@ -2053,6 +2134,8 @@ var chutesModels = {
2053
2134
  contextWindow: 163840,
2054
2135
  supportsImages: false,
2055
2136
  supportsPromptCache: false,
2137
+ supportsNativeTools: true,
2138
+ defaultToolProtocol: "native",
2056
2139
  inputPrice: 0,
2057
2140
  outputPrice: 0,
2058
2141
  description: "DeepSeek V3 (0324) model."
@@ -2062,6 +2145,8 @@ var chutesModels = {
2062
2145
  contextWindow: 262144,
2063
2146
  supportsImages: false,
2064
2147
  supportsPromptCache: false,
2148
+ supportsNativeTools: true,
2149
+ defaultToolProtocol: "native",
2065
2150
  inputPrice: 0,
2066
2151
  outputPrice: 0,
2067
2152
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -2071,6 +2156,8 @@ var chutesModels = {
2071
2156
  contextWindow: 40960,
2072
2157
  supportsImages: false,
2073
2158
  supportsPromptCache: false,
2159
+ supportsNativeTools: true,
2160
+ defaultToolProtocol: "native",
2074
2161
  inputPrice: 0,
2075
2162
  outputPrice: 0,
2076
2163
  description: "Qwen3 235B A22B model."
@@ -2080,6 +2167,8 @@ var chutesModels = {
2080
2167
  contextWindow: 40960,
2081
2168
  supportsImages: false,
2082
2169
  supportsPromptCache: false,
2170
+ supportsNativeTools: true,
2171
+ defaultToolProtocol: "native",
2083
2172
  inputPrice: 0,
2084
2173
  outputPrice: 0,
2085
2174
  description: "Qwen3 32B model."
@@ -2089,6 +2178,8 @@ var chutesModels = {
2089
2178
  contextWindow: 40960,
2090
2179
  supportsImages: false,
2091
2180
  supportsPromptCache: false,
2181
+ supportsNativeTools: true,
2182
+ defaultToolProtocol: "native",
2092
2183
  inputPrice: 0,
2093
2184
  outputPrice: 0,
2094
2185
  description: "Qwen3 30B A3B model."
@@ -2098,6 +2189,8 @@ var chutesModels = {
2098
2189
  contextWindow: 40960,
2099
2190
  supportsImages: false,
2100
2191
  supportsPromptCache: false,
2192
+ supportsNativeTools: true,
2193
+ defaultToolProtocol: "native",
2101
2194
  inputPrice: 0,
2102
2195
  outputPrice: 0,
2103
2196
  description: "Qwen3 14B model."
@@ -2107,6 +2200,8 @@ var chutesModels = {
2107
2200
  contextWindow: 40960,
2108
2201
  supportsImages: false,
2109
2202
  supportsPromptCache: false,
2203
+ supportsNativeTools: true,
2204
+ defaultToolProtocol: "native",
2110
2205
  inputPrice: 0,
2111
2206
  outputPrice: 0,
2112
2207
  description: "Qwen3 8B model."
@@ -2116,6 +2211,8 @@ var chutesModels = {
2116
2211
  contextWindow: 163840,
2117
2212
  supportsImages: false,
2118
2213
  supportsPromptCache: false,
2214
+ supportsNativeTools: true,
2215
+ defaultToolProtocol: "native",
2119
2216
  inputPrice: 0,
2120
2217
  outputPrice: 0,
2121
2218
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -2125,6 +2222,8 @@ var chutesModels = {
2125
2222
  contextWindow: 163840,
2126
2223
  supportsImages: false,
2127
2224
  supportsPromptCache: false,
2225
+ supportsNativeTools: true,
2226
+ defaultToolProtocol: "native",
2128
2227
  inputPrice: 0,
2129
2228
  outputPrice: 0,
2130
2229
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -2134,6 +2233,8 @@ var chutesModels = {
2134
2233
  contextWindow: 151329,
2135
2234
  supportsImages: false,
2136
2235
  supportsPromptCache: false,
2236
+ supportsNativeTools: true,
2237
+ defaultToolProtocol: "native",
2137
2238
  inputPrice: 0,
2138
2239
  outputPrice: 0,
2139
2240
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -2143,6 +2244,8 @@ var chutesModels = {
2143
2244
  contextWindow: 131072,
2144
2245
  supportsImages: false,
2145
2246
  supportsPromptCache: false,
2247
+ supportsNativeTools: true,
2248
+ defaultToolProtocol: "native",
2146
2249
  inputPrice: 0,
2147
2250
  outputPrice: 0,
2148
2251
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -2152,6 +2255,8 @@ var chutesModels = {
2152
2255
  contextWindow: 131072,
2153
2256
  supportsImages: false,
2154
2257
  supportsPromptCache: false,
2258
+ supportsNativeTools: true,
2259
+ defaultToolProtocol: "native",
2155
2260
  inputPrice: 1,
2156
2261
  outputPrice: 3,
2157
2262
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -2161,6 +2266,8 @@ var chutesModels = {
2161
2266
  contextWindow: 202752,
2162
2267
  supportsImages: false,
2163
2268
  supportsPromptCache: false,
2269
+ supportsNativeTools: true,
2270
+ defaultToolProtocol: "native",
2164
2271
  inputPrice: 0,
2165
2272
  outputPrice: 0,
2166
2273
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -2171,6 +2278,8 @@ var chutesModels = {
2171
2278
  contextWindow: 202752,
2172
2279
  supportsImages: false,
2173
2280
  supportsPromptCache: false,
2281
+ supportsNativeTools: true,
2282
+ defaultToolProtocol: "native",
2174
2283
  inputPrice: 1.15,
2175
2284
  outputPrice: 3.25,
2176
2285
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -2180,6 +2289,8 @@ var chutesModels = {
2180
2289
  contextWindow: 128e3,
2181
2290
  supportsImages: false,
2182
2291
  supportsPromptCache: false,
2292
+ supportsNativeTools: true,
2293
+ defaultToolProtocol: "native",
2183
2294
  inputPrice: 0,
2184
2295
  outputPrice: 0,
2185
2296
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -2189,6 +2300,8 @@ var chutesModels = {
2189
2300
  contextWindow: 262144,
2190
2301
  supportsImages: false,
2191
2302
  supportsPromptCache: false,
2303
+ supportsNativeTools: true,
2304
+ defaultToolProtocol: "native",
2192
2305
  inputPrice: 0,
2193
2306
  outputPrice: 0,
2194
2307
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -2198,6 +2311,8 @@ var chutesModels = {
2198
2311
  contextWindow: 75e3,
2199
2312
  supportsImages: false,
2200
2313
  supportsPromptCache: false,
2314
+ supportsNativeTools: true,
2315
+ defaultToolProtocol: "native",
2201
2316
  inputPrice: 0.1481,
2202
2317
  outputPrice: 0.5926,
2203
2318
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -2207,6 +2322,8 @@ var chutesModels = {
2207
2322
  contextWindow: 262144,
2208
2323
  supportsImages: false,
2209
2324
  supportsPromptCache: false,
2325
+ supportsNativeTools: true,
2326
+ defaultToolProtocol: "native",
2210
2327
  inputPrice: 0.1999,
2211
2328
  outputPrice: 0.8001,
2212
2329
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -2216,6 +2333,8 @@ var chutesModels = {
2216
2333
  contextWindow: 262144,
2217
2334
  supportsImages: false,
2218
2335
  supportsPromptCache: false,
2336
+ supportsNativeTools: true,
2337
+ defaultToolProtocol: "native",
2219
2338
  inputPrice: 0.077968332,
2220
2339
  outputPrice: 0.31202496,
2221
2340
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -2225,6 +2344,8 @@ var chutesModels = {
2225
2344
  contextWindow: 131072,
2226
2345
  supportsImages: false,
2227
2346
  supportsPromptCache: false,
2347
+ supportsNativeTools: true,
2348
+ defaultToolProtocol: "native",
2228
2349
  inputPrice: 0,
2229
2350
  outputPrice: 0,
2230
2351
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -2234,6 +2355,8 @@ var chutesModels = {
2234
2355
  contextWindow: 131072,
2235
2356
  supportsImages: false,
2236
2357
  supportsPromptCache: false,
2358
+ supportsNativeTools: true,
2359
+ defaultToolProtocol: "native",
2237
2360
  inputPrice: 0,
2238
2361
  outputPrice: 0,
2239
2362
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2243,6 +2366,8 @@ var chutesModels = {
2243
2366
  contextWindow: 262144,
2244
2367
  supportsImages: true,
2245
2368
  supportsPromptCache: false,
2369
+ supportsNativeTools: true,
2370
+ defaultToolProtocol: "native",
2246
2371
  inputPrice: 0.16,
2247
2372
  outputPrice: 0.65,
2248
2373
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2251,138 +2376,70 @@ var chutesModels = {
2251
2376
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2252
2377
 
2253
2378
  // src/providers/claude-code.ts
2254
- var VERTEX_DATE_PATTERN = /-(\d{8})$/;
2255
- function convertModelNameForVertex(modelName) {
2256
- return modelName.replace(VERTEX_DATE_PATTERN, "@$1");
2257
- }
2258
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2259
- var CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS = 16e3;
2260
- function getClaudeCodeModelId(baseModelId, useVertex = false) {
2261
- return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
2262
- }
2379
+ var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2263
2380
  var claudeCodeModels = {
2264
- "claude-sonnet-4-5": {
2265
- ...anthropicModels["claude-sonnet-4-5"],
2266
- supportsImages: false,
2267
- supportsPromptCache: true,
2268
- // Claude Code does report cache tokens
2269
- supportsReasoningEffort: false,
2270
- supportsReasoningBudget: false,
2271
- requiredReasoningBudget: false,
2272
- // Claude Code manages its own tools and temperature via the CLI
2273
- supportsNativeTools: false,
2274
- supportsTemperature: false
2275
- },
2276
- "claude-sonnet-4-5-20250929[1m]": {
2277
- ...anthropicModels["claude-sonnet-4-5"],
2278
- contextWindow: 1e6,
2279
- // 1M token context window (requires [1m] suffix)
2280
- supportsImages: false,
2281
- supportsPromptCache: true,
2282
- // Claude Code does report cache tokens
2283
- supportsReasoningEffort: false,
2284
- supportsReasoningBudget: false,
2285
- requiredReasoningBudget: false,
2286
- // Claude Code manages its own tools and temperature via the CLI
2287
- supportsNativeTools: false,
2288
- supportsTemperature: false
2289
- },
2290
- "claude-sonnet-4-20250514": {
2291
- ...anthropicModels["claude-sonnet-4-20250514"],
2292
- supportsImages: false,
2293
- supportsPromptCache: true,
2294
- // Claude Code does report cache tokens
2295
- supportsReasoningEffort: false,
2296
- supportsReasoningBudget: false,
2297
- requiredReasoningBudget: false,
2298
- // Claude Code manages its own tools and temperature via the CLI
2299
- supportsNativeTools: false,
2300
- supportsTemperature: false
2301
- },
2302
- "claude-opus-4-5-20251101": {
2303
- ...anthropicModels["claude-opus-4-5-20251101"],
2304
- supportsImages: false,
2305
- supportsPromptCache: true,
2306
- // Claude Code does report cache tokens
2307
- supportsReasoningEffort: false,
2308
- supportsReasoningBudget: false,
2309
- requiredReasoningBudget: false,
2310
- // Claude Code manages its own tools and temperature via the CLI
2311
- supportsNativeTools: false,
2312
- supportsTemperature: false
2313
- },
2314
- "claude-opus-4-1-20250805": {
2315
- ...anthropicModels["claude-opus-4-1-20250805"],
2316
- supportsImages: false,
2317
- supportsPromptCache: true,
2318
- // Claude Code does report cache tokens
2319
- supportsReasoningEffort: false,
2320
- supportsReasoningBudget: false,
2321
- requiredReasoningBudget: false,
2322
- // Claude Code manages its own tools and temperature via the CLI
2323
- supportsNativeTools: false,
2324
- supportsTemperature: false
2325
- },
2326
- "claude-opus-4-20250514": {
2327
- ...anthropicModels["claude-opus-4-20250514"],
2328
- supportsImages: false,
2329
- supportsPromptCache: true,
2330
- // Claude Code does report cache tokens
2331
- supportsReasoningEffort: false,
2332
- supportsReasoningBudget: false,
2333
- requiredReasoningBudget: false,
2334
- // Claude Code manages its own tools and temperature via the CLI
2335
- supportsNativeTools: false,
2336
- supportsTemperature: false
2337
- },
2338
- "claude-3-7-sonnet-20250219": {
2339
- ...anthropicModels["claude-3-7-sonnet-20250219"],
2340
- supportsImages: false,
2341
- supportsPromptCache: true,
2342
- // Claude Code does report cache tokens
2343
- supportsReasoningEffort: false,
2344
- supportsReasoningBudget: false,
2345
- requiredReasoningBudget: false,
2346
- // Claude Code manages its own tools and temperature via the CLI
2347
- supportsNativeTools: false,
2348
- supportsTemperature: false
2349
- },
2350
- "claude-3-5-sonnet-20241022": {
2351
- ...anthropicModels["claude-3-5-sonnet-20241022"],
2352
- supportsImages: false,
2381
+ "claude-haiku-4-5": {
2382
+ maxTokens: 32768,
2383
+ contextWindow: 2e5,
2384
+ supportsImages: true,
2353
2385
  supportsPromptCache: true,
2354
- // Claude Code does report cache tokens
2355
- supportsReasoningEffort: false,
2356
- supportsReasoningBudget: false,
2357
- requiredReasoningBudget: false,
2358
- // Claude Code manages its own tools and temperature via the CLI
2359
- supportsNativeTools: false,
2360
- supportsTemperature: false
2386
+ supportsNativeTools: true,
2387
+ defaultToolProtocol: "native",
2388
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2389
+ reasoningEffort: "medium",
2390
+ description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2361
2391
  },
2362
- "claude-3-5-haiku-20241022": {
2363
- ...anthropicModels["claude-3-5-haiku-20241022"],
2364
- supportsImages: false,
2392
+ "claude-sonnet-4-5": {
2393
+ maxTokens: 32768,
2394
+ contextWindow: 2e5,
2395
+ supportsImages: true,
2365
2396
  supportsPromptCache: true,
2366
- // Claude Code does report cache tokens
2367
- supportsReasoningEffort: false,
2368
- supportsReasoningBudget: false,
2369
- requiredReasoningBudget: false,
2370
- // Claude Code manages its own tools and temperature via the CLI
2371
- supportsNativeTools: false,
2372
- supportsTemperature: false
2397
+ supportsNativeTools: true,
2398
+ defaultToolProtocol: "native",
2399
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2400
+ reasoningEffort: "medium",
2401
+ description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2373
2402
  },
2374
- "claude-haiku-4-5-20251001": {
2375
- ...anthropicModels["claude-haiku-4-5-20251001"],
2376
- supportsImages: false,
2403
+ "claude-opus-4-5": {
2404
+ maxTokens: 32768,
2405
+ contextWindow: 2e5,
2406
+ supportsImages: true,
2377
2407
  supportsPromptCache: true,
2378
- // Claude Code does report cache tokens
2379
- supportsReasoningEffort: false,
2380
- supportsReasoningBudget: false,
2381
- requiredReasoningBudget: false,
2382
- // Claude Code manages its own tools and temperature via the CLI
2383
- supportsNativeTools: false,
2384
- supportsTemperature: false
2408
+ supportsNativeTools: true,
2409
+ defaultToolProtocol: "native",
2410
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2411
+ reasoningEffort: "medium",
2412
+ description: "Claude Opus 4.5 - Most capable with thinking"
2413
+ }
2414
+ };
2415
+ var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2416
+ var MODEL_FAMILY_PATTERNS = [
2417
+ // Opus models (any version) → claude-opus-4-5
2418
+ { pattern: /opus/i, target: "claude-opus-4-5" },
2419
+ // Haiku models (any version) → claude-haiku-4-5
2420
+ { pattern: /haiku/i, target: "claude-haiku-4-5" },
2421
+ // Sonnet models (any version) → claude-sonnet-4-5
2422
+ { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2423
+ ];
2424
+ function normalizeClaudeCodeModelId(modelId) {
2425
+ if (Object.hasOwn(claudeCodeModels, modelId)) {
2426
+ return modelId;
2385
2427
  }
2428
+ const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2429
+ if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2430
+ return withoutDate;
2431
+ }
2432
+ for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2433
+ if (pattern.test(modelId)) {
2434
+ return target;
2435
+ }
2436
+ }
2437
+ return claudeCodeDefaultModelId;
2438
+ }
2439
+ var claudeCodeReasoningConfig = {
2440
+ low: { budgetTokens: 16e3 },
2441
+ medium: { budgetTokens: 32e3 },
2442
+ high: { budgetTokens: 64e3 }
2386
2443
  };
2387
2444
 
2388
2445
  // src/providers/deepseek.ts
@@ -2414,6 +2471,7 @@ var deepSeekModels = {
2414
2471
  supportsPromptCache: true,
2415
2472
  supportsNativeTools: true,
2416
2473
  defaultToolProtocol: "native",
2474
+ preserveReasoning: true,
2417
2475
  inputPrice: 0.28,
2418
2476
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2419
2477
  outputPrice: 0.42,
@@ -2425,7 +2483,7 @@ var deepSeekModels = {
2425
2483
  description: `DeepSeek-V3.2 (Thinking Mode) achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 8K output tokens. Supports JSON output, tool calls, and chat prefix completion (beta).`
2426
2484
  }
2427
2485
  };
2428
- var DEEP_SEEK_DEFAULT_TEMPERATURE = 0;
2486
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.3;
2429
2487
 
2430
2488
  // src/providers/doubao.ts
2431
2489
  var doubaoDefaultModelId = "doubao-seed-1-6-250615";
@@ -2436,6 +2494,7 @@ var doubaoModels = {
2436
2494
  supportsImages: true,
2437
2495
  supportsPromptCache: true,
2438
2496
  supportsNativeTools: true,
2497
+ defaultToolProtocol: "native",
2439
2498
  inputPrice: 1e-4,
2440
2499
  // $0.0001 per million tokens (cache miss)
2441
2500
  outputPrice: 4e-4,
@@ -2452,6 +2511,7 @@ var doubaoModels = {
2452
2511
  supportsImages: true,
2453
2512
  supportsPromptCache: true,
2454
2513
  supportsNativeTools: true,
2514
+ defaultToolProtocol: "native",
2455
2515
  inputPrice: 2e-4,
2456
2516
  // $0.0002 per million tokens
2457
2517
  outputPrice: 8e-4,
@@ -2468,6 +2528,7 @@ var doubaoModels = {
2468
2528
  supportsImages: true,
2469
2529
  supportsPromptCache: true,
2470
2530
  supportsNativeTools: true,
2531
+ defaultToolProtocol: "native",
2471
2532
  inputPrice: 15e-5,
2472
2533
  // $0.00015 per million tokens
2473
2534
  outputPrice: 6e-4,
@@ -2544,6 +2605,7 @@ var fireworksModels = {
2544
2605
  supportsImages: false,
2545
2606
  supportsPromptCache: true,
2546
2607
  supportsNativeTools: true,
2608
+ defaultToolProtocol: "native",
2547
2609
  inputPrice: 0.6,
2548
2610
  outputPrice: 2.5,
2549
2611
  cacheReadsPrice: 0.15,
@@ -2555,6 +2617,7 @@ var fireworksModels = {
2555
2617
  supportsImages: false,
2556
2618
  supportsPromptCache: false,
2557
2619
  supportsNativeTools: true,
2620
+ defaultToolProtocol: "native",
2558
2621
  inputPrice: 0.6,
2559
2622
  outputPrice: 2.5,
2560
2623
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2565,6 +2628,7 @@ var fireworksModels = {
2565
2628
  supportsImages: false,
2566
2629
  supportsPromptCache: false,
2567
2630
  supportsNativeTools: true,
2631
+ defaultToolProtocol: "native",
2568
2632
  inputPrice: 0.3,
2569
2633
  outputPrice: 1.2,
2570
2634
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2575,6 +2639,7 @@ var fireworksModels = {
2575
2639
  supportsImages: false,
2576
2640
  supportsPromptCache: false,
2577
2641
  supportsNativeTools: true,
2642
+ defaultToolProtocol: "native",
2578
2643
  inputPrice: 0.22,
2579
2644
  outputPrice: 0.88,
2580
2645
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2585,6 +2650,7 @@ var fireworksModels = {
2585
2650
  supportsImages: false,
2586
2651
  supportsPromptCache: false,
2587
2652
  supportsNativeTools: true,
2653
+ defaultToolProtocol: "native",
2588
2654
  inputPrice: 0.45,
2589
2655
  outputPrice: 1.8,
2590
2656
  description: "Qwen3's most agentic code model to date."
@@ -2595,6 +2661,7 @@ var fireworksModels = {
2595
2661
  supportsImages: false,
2596
2662
  supportsPromptCache: false,
2597
2663
  supportsNativeTools: true,
2664
+ defaultToolProtocol: "native",
2598
2665
  inputPrice: 3,
2599
2666
  outputPrice: 8,
2600
2667
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2605,6 +2672,7 @@ var fireworksModels = {
2605
2672
  supportsImages: false,
2606
2673
  supportsPromptCache: false,
2607
2674
  supportsNativeTools: true,
2675
+ defaultToolProtocol: "native",
2608
2676
  inputPrice: 0.9,
2609
2677
  outputPrice: 0.9,
2610
2678
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2615,6 +2683,7 @@ var fireworksModels = {
2615
2683
  supportsImages: false,
2616
2684
  supportsPromptCache: false,
2617
2685
  supportsNativeTools: true,
2686
+ defaultToolProtocol: "native",
2618
2687
  inputPrice: 0.56,
2619
2688
  outputPrice: 1.68,
2620
2689
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2625,6 +2694,7 @@ var fireworksModels = {
2625
2694
  supportsImages: false,
2626
2695
  supportsPromptCache: false,
2627
2696
  supportsNativeTools: true,
2697
+ defaultToolProtocol: "native",
2628
2698
  inputPrice: 0.55,
2629
2699
  outputPrice: 2.19,
2630
2700
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2635,6 +2705,7 @@ var fireworksModels = {
2635
2705
  supportsImages: false,
2636
2706
  supportsPromptCache: false,
2637
2707
  supportsNativeTools: true,
2708
+ defaultToolProtocol: "native",
2638
2709
  inputPrice: 0.55,
2639
2710
  outputPrice: 2.19,
2640
2711
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2645,6 +2716,7 @@ var fireworksModels = {
2645
2716
  supportsImages: false,
2646
2717
  supportsPromptCache: false,
2647
2718
  supportsNativeTools: true,
2719
+ defaultToolProtocol: "native",
2648
2720
  inputPrice: 0.55,
2649
2721
  outputPrice: 2.19,
2650
2722
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2655,6 +2727,7 @@ var fireworksModels = {
2655
2727
  supportsImages: false,
2656
2728
  supportsPromptCache: false,
2657
2729
  supportsNativeTools: true,
2730
+ defaultToolProtocol: "native",
2658
2731
  inputPrice: 0.07,
2659
2732
  outputPrice: 0.3,
2660
2733
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2665,6 +2738,7 @@ var fireworksModels = {
2665
2738
  supportsImages: false,
2666
2739
  supportsPromptCache: false,
2667
2740
  supportsNativeTools: true,
2741
+ defaultToolProtocol: "native",
2668
2742
  inputPrice: 0.15,
2669
2743
  outputPrice: 0.6,
2670
2744
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2672,16 +2746,19 @@ var fireworksModels = {
2672
2746
  };
2673
2747
 
2674
2748
  // src/providers/gemini.ts
2675
- var geminiDefaultModelId = "gemini-2.5-pro";
2749
+ var geminiDefaultModelId = "gemini-3-pro-preview";
2676
2750
  var geminiModels = {
2677
2751
  "gemini-3-pro-preview": {
2678
2752
  maxTokens: 65536,
2679
2753
  contextWindow: 1048576,
2680
2754
  supportsImages: true,
2681
2755
  supportsNativeTools: true,
2756
+ defaultToolProtocol: "native",
2682
2757
  supportsPromptCache: true,
2683
2758
  supportsReasoningEffort: ["low", "high"],
2684
2759
  reasoningEffort: "low",
2760
+ includedTools: ["write_file", "edit_file"],
2761
+ excludedTools: ["apply_diff"],
2685
2762
  supportsTemperature: true,
2686
2763
  defaultTemperature: 1,
2687
2764
  inputPrice: 4,
@@ -2699,13 +2776,34 @@ var geminiModels = {
2699
2776
  }
2700
2777
  ]
2701
2778
  },
2779
+ "gemini-3-flash-preview": {
2780
+ maxTokens: 65536,
2781
+ contextWindow: 1048576,
2782
+ supportsImages: true,
2783
+ supportsNativeTools: true,
2784
+ defaultToolProtocol: "native",
2785
+ supportsPromptCache: true,
2786
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2787
+ reasoningEffort: "medium",
2788
+ includedTools: ["write_file", "edit_file"],
2789
+ excludedTools: ["apply_diff"],
2790
+ supportsTemperature: true,
2791
+ defaultTemperature: 1,
2792
+ inputPrice: 0.3,
2793
+ outputPrice: 2.5,
2794
+ cacheReadsPrice: 0.075,
2795
+ cacheWritesPrice: 1
2796
+ },
2702
2797
  // 2.5 Pro models
2703
2798
  "gemini-2.5-pro": {
2704
2799
  maxTokens: 64e3,
2705
2800
  contextWindow: 1048576,
2706
2801
  supportsImages: true,
2707
2802
  supportsNativeTools: true,
2803
+ defaultToolProtocol: "native",
2708
2804
  supportsPromptCache: true,
2805
+ includedTools: ["write_file", "edit_file"],
2806
+ excludedTools: ["apply_diff"],
2709
2807
  inputPrice: 2.5,
2710
2808
  // This is the pricing for prompts above 200k tokens.
2711
2809
  outputPrice: 15,
@@ -2734,7 +2832,10 @@ var geminiModels = {
2734
2832
  contextWindow: 1048576,
2735
2833
  supportsImages: true,
2736
2834
  supportsNativeTools: true,
2835
+ defaultToolProtocol: "native",
2737
2836
  supportsPromptCache: true,
2837
+ includedTools: ["write_file", "edit_file"],
2838
+ excludedTools: ["apply_diff"],
2738
2839
  inputPrice: 2.5,
2739
2840
  // This is the pricing for prompts above 200k tokens.
2740
2841
  outputPrice: 15,
@@ -2762,7 +2863,10 @@ var geminiModels = {
2762
2863
  contextWindow: 1048576,
2763
2864
  supportsImages: true,
2764
2865
  supportsNativeTools: true,
2866
+ defaultToolProtocol: "native",
2765
2867
  supportsPromptCache: true,
2868
+ includedTools: ["write_file", "edit_file"],
2869
+ excludedTools: ["apply_diff"],
2766
2870
  inputPrice: 2.5,
2767
2871
  // This is the pricing for prompts above 200k tokens.
2768
2872
  outputPrice: 15,
@@ -2788,7 +2892,10 @@ var geminiModels = {
2788
2892
  contextWindow: 1048576,
2789
2893
  supportsImages: true,
2790
2894
  supportsNativeTools: true,
2895
+ defaultToolProtocol: "native",
2791
2896
  supportsPromptCache: true,
2897
+ includedTools: ["write_file", "edit_file"],
2898
+ excludedTools: ["apply_diff"],
2792
2899
  inputPrice: 2.5,
2793
2900
  // This is the pricing for prompts above 200k tokens.
2794
2901
  outputPrice: 15,
@@ -2817,7 +2924,10 @@ var geminiModels = {
2817
2924
  contextWindow: 1048576,
2818
2925
  supportsImages: true,
2819
2926
  supportsNativeTools: true,
2927
+ defaultToolProtocol: "native",
2820
2928
  supportsPromptCache: true,
2929
+ includedTools: ["write_file", "edit_file"],
2930
+ excludedTools: ["apply_diff"],
2821
2931
  inputPrice: 0.3,
2822
2932
  outputPrice: 2.5,
2823
2933
  cacheReadsPrice: 0.075,
@@ -2830,7 +2940,10 @@ var geminiModels = {
2830
2940
  contextWindow: 1048576,
2831
2941
  supportsImages: true,
2832
2942
  supportsNativeTools: true,
2943
+ defaultToolProtocol: "native",
2833
2944
  supportsPromptCache: true,
2945
+ includedTools: ["write_file", "edit_file"],
2946
+ excludedTools: ["apply_diff"],
2834
2947
  inputPrice: 0.3,
2835
2948
  outputPrice: 2.5,
2836
2949
  cacheReadsPrice: 0.075,
@@ -2843,7 +2956,10 @@ var geminiModels = {
2843
2956
  contextWindow: 1048576,
2844
2957
  supportsImages: true,
2845
2958
  supportsNativeTools: true,
2959
+ defaultToolProtocol: "native",
2846
2960
  supportsPromptCache: true,
2961
+ includedTools: ["write_file", "edit_file"],
2962
+ excludedTools: ["apply_diff"],
2847
2963
  inputPrice: 0.3,
2848
2964
  outputPrice: 2.5,
2849
2965
  cacheReadsPrice: 0.075,
@@ -2857,7 +2973,10 @@ var geminiModels = {
2857
2973
  contextWindow: 1048576,
2858
2974
  supportsImages: true,
2859
2975
  supportsNativeTools: true,
2976
+ defaultToolProtocol: "native",
2860
2977
  supportsPromptCache: true,
2978
+ includedTools: ["write_file", "edit_file"],
2979
+ excludedTools: ["apply_diff"],
2861
2980
  inputPrice: 0.1,
2862
2981
  outputPrice: 0.4,
2863
2982
  cacheReadsPrice: 0.025,
@@ -2870,7 +2989,10 @@ var geminiModels = {
2870
2989
  contextWindow: 1048576,
2871
2990
  supportsImages: true,
2872
2991
  supportsNativeTools: true,
2992
+ defaultToolProtocol: "native",
2873
2993
  supportsPromptCache: true,
2994
+ includedTools: ["write_file", "edit_file"],
2995
+ excludedTools: ["apply_diff"],
2874
2996
  inputPrice: 0.1,
2875
2997
  outputPrice: 0.4,
2876
2998
  cacheReadsPrice: 0.025,
@@ -2890,6 +3012,7 @@ var groqModels = {
2890
3012
  supportsImages: false,
2891
3013
  supportsPromptCache: false,
2892
3014
  supportsNativeTools: true,
3015
+ defaultToolProtocol: "native",
2893
3016
  inputPrice: 0.05,
2894
3017
  outputPrice: 0.08,
2895
3018
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2900,6 +3023,7 @@ var groqModels = {
2900
3023
  supportsImages: false,
2901
3024
  supportsPromptCache: false,
2902
3025
  supportsNativeTools: true,
3026
+ defaultToolProtocol: "native",
2903
3027
  inputPrice: 0.59,
2904
3028
  outputPrice: 0.79,
2905
3029
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2910,6 +3034,7 @@ var groqModels = {
2910
3034
  supportsImages: false,
2911
3035
  supportsPromptCache: false,
2912
3036
  supportsNativeTools: true,
3037
+ defaultToolProtocol: "native",
2913
3038
  inputPrice: 0.11,
2914
3039
  outputPrice: 0.34,
2915
3040
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2947,6 +3072,7 @@ var groqModels = {
2947
3072
  supportsImages: false,
2948
3073
  supportsPromptCache: false,
2949
3074
  supportsNativeTools: true,
3075
+ defaultToolProtocol: "native",
2950
3076
  inputPrice: 0.29,
2951
3077
  outputPrice: 0.59,
2952
3078
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2977,6 +3103,7 @@ var groqModels = {
2977
3103
  supportsImages: false,
2978
3104
  supportsPromptCache: true,
2979
3105
  supportsNativeTools: true,
3106
+ defaultToolProtocol: "native",
2980
3107
  inputPrice: 0.6,
2981
3108
  outputPrice: 2.5,
2982
3109
  cacheReadsPrice: 0.15,
@@ -2988,6 +3115,7 @@ var groqModels = {
2988
3115
  supportsImages: false,
2989
3116
  supportsPromptCache: false,
2990
3117
  supportsNativeTools: true,
3118
+ defaultToolProtocol: "native",
2991
3119
  inputPrice: 0.15,
2992
3120
  outputPrice: 0.75,
2993
3121
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2998,6 +3126,7 @@ var groqModels = {
2998
3126
  supportsImages: false,
2999
3127
  supportsPromptCache: false,
3000
3128
  supportsNativeTools: true,
3129
+ defaultToolProtocol: "native",
3001
3130
  inputPrice: 0.1,
3002
3131
  outputPrice: 0.5,
3003
3132
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -3061,6 +3190,7 @@ var litellmDefaultModelInfo = {
3061
3190
  supportsImages: true,
3062
3191
  supportsPromptCache: true,
3063
3192
  supportsNativeTools: true,
3193
+ defaultToolProtocol: "native",
3064
3194
  inputPrice: 3,
3065
3195
  outputPrice: 15,
3066
3196
  cacheWritesPrice: 3.75,
@@ -3075,6 +3205,8 @@ var lMStudioDefaultModelInfo = {
3075
3205
  contextWindow: 2e5,
3076
3206
  supportsImages: true,
3077
3207
  supportsPromptCache: true,
3208
+ supportsNativeTools: true,
3209
+ defaultToolProtocol: "native",
3078
3210
  inputPrice: 0,
3079
3211
  outputPrice: 0,
3080
3212
  cacheWritesPrice: 0,
@@ -3091,6 +3223,7 @@ var mistralModels = {
3091
3223
  supportsImages: true,
3092
3224
  supportsPromptCache: false,
3093
3225
  supportsNativeTools: true,
3226
+ defaultToolProtocol: "native",
3094
3227
  inputPrice: 2,
3095
3228
  outputPrice: 5
3096
3229
  },
@@ -3100,6 +3233,7 @@ var mistralModels = {
3100
3233
  supportsImages: true,
3101
3234
  supportsPromptCache: false,
3102
3235
  supportsNativeTools: true,
3236
+ defaultToolProtocol: "native",
3103
3237
  inputPrice: 0.4,
3104
3238
  outputPrice: 2
3105
3239
  },
@@ -3109,6 +3243,7 @@ var mistralModels = {
3109
3243
  supportsImages: true,
3110
3244
  supportsPromptCache: false,
3111
3245
  supportsNativeTools: true,
3246
+ defaultToolProtocol: "native",
3112
3247
  inputPrice: 0.4,
3113
3248
  outputPrice: 2
3114
3249
  },
@@ -3118,6 +3253,7 @@ var mistralModels = {
3118
3253
  supportsImages: false,
3119
3254
  supportsPromptCache: false,
3120
3255
  supportsNativeTools: true,
3256
+ defaultToolProtocol: "native",
3121
3257
  inputPrice: 0.3,
3122
3258
  outputPrice: 0.9
3123
3259
  },
@@ -3127,6 +3263,7 @@ var mistralModels = {
3127
3263
  supportsImages: false,
3128
3264
  supportsPromptCache: false,
3129
3265
  supportsNativeTools: true,
3266
+ defaultToolProtocol: "native",
3130
3267
  inputPrice: 2,
3131
3268
  outputPrice: 6
3132
3269
  },
@@ -3136,6 +3273,7 @@ var mistralModels = {
3136
3273
  supportsImages: false,
3137
3274
  supportsPromptCache: false,
3138
3275
  supportsNativeTools: true,
3276
+ defaultToolProtocol: "native",
3139
3277
  inputPrice: 0.1,
3140
3278
  outputPrice: 0.1
3141
3279
  },
@@ -3145,6 +3283,7 @@ var mistralModels = {
3145
3283
  supportsImages: false,
3146
3284
  supportsPromptCache: false,
3147
3285
  supportsNativeTools: true,
3286
+ defaultToolProtocol: "native",
3148
3287
  inputPrice: 0.04,
3149
3288
  outputPrice: 0.04
3150
3289
  },
@@ -3154,6 +3293,7 @@ var mistralModels = {
3154
3293
  supportsImages: false,
3155
3294
  supportsPromptCache: false,
3156
3295
  supportsNativeTools: true,
3296
+ defaultToolProtocol: "native",
3157
3297
  inputPrice: 0.2,
3158
3298
  outputPrice: 0.6
3159
3299
  },
@@ -3163,6 +3303,7 @@ var mistralModels = {
3163
3303
  supportsImages: true,
3164
3304
  supportsPromptCache: false,
3165
3305
  supportsNativeTools: true,
3306
+ defaultToolProtocol: "native",
3166
3307
  inputPrice: 2,
3167
3308
  outputPrice: 6
3168
3309
  }
@@ -3178,6 +3319,7 @@ var moonshotModels = {
3178
3319
  supportsImages: false,
3179
3320
  supportsPromptCache: true,
3180
3321
  supportsNativeTools: true,
3322
+ defaultToolProtocol: "native",
3181
3323
  inputPrice: 0.6,
3182
3324
  // $0.60 per million tokens (cache miss)
3183
3325
  outputPrice: 2.5,
@@ -3194,6 +3336,7 @@ var moonshotModels = {
3194
3336
  supportsImages: false,
3195
3337
  supportsPromptCache: true,
3196
3338
  supportsNativeTools: true,
3339
+ defaultToolProtocol: "native",
3197
3340
  inputPrice: 0.6,
3198
3341
  outputPrice: 2.5,
3199
3342
  cacheReadsPrice: 0.15,
@@ -3205,6 +3348,7 @@ var moonshotModels = {
3205
3348
  supportsImages: false,
3206
3349
  supportsPromptCache: true,
3207
3350
  supportsNativeTools: true,
3351
+ defaultToolProtocol: "native",
3208
3352
  inputPrice: 2.4,
3209
3353
  // $2.40 per million tokens (cache miss)
3210
3354
  outputPrice: 10,
@@ -3224,6 +3368,7 @@ var moonshotModels = {
3224
3368
  // Text-only (no image/vision support)
3225
3369
  supportsPromptCache: true,
3226
3370
  supportsNativeTools: true,
3371
+ defaultToolProtocol: "native",
3227
3372
  inputPrice: 0.6,
3228
3373
  // $0.60 per million tokens (cache miss)
3229
3374
  outputPrice: 2.5,
@@ -3263,6 +3408,7 @@ var openAiNativeModels = {
3263
3408
  maxTokens: 128e3,
3264
3409
  contextWindow: 4e5,
3265
3410
  supportsNativeTools: true,
3411
+ defaultToolProtocol: "native",
3266
3412
  includedTools: ["apply_patch"],
3267
3413
  excludedTools: ["apply_diff", "write_to_file"],
3268
3414
  supportsImages: true,
@@ -3281,6 +3427,7 @@ var openAiNativeModels = {
3281
3427
  maxTokens: 128e3,
3282
3428
  contextWindow: 4e5,
3283
3429
  supportsNativeTools: true,
3430
+ defaultToolProtocol: "native",
3284
3431
  includedTools: ["apply_patch"],
3285
3432
  excludedTools: ["apply_diff", "write_to_file"],
3286
3433
  supportsImages: true,
@@ -3303,6 +3450,7 @@ var openAiNativeModels = {
3303
3450
  maxTokens: 16384,
3304
3451
  contextWindow: 128e3,
3305
3452
  supportsNativeTools: true,
3453
+ defaultToolProtocol: "native",
3306
3454
  includedTools: ["apply_patch"],
3307
3455
  excludedTools: ["apply_diff", "write_to_file"],
3308
3456
  supportsImages: true,
@@ -3316,6 +3464,7 @@ var openAiNativeModels = {
3316
3464
  maxTokens: 128e3,
3317
3465
  contextWindow: 4e5,
3318
3466
  supportsNativeTools: true,
3467
+ defaultToolProtocol: "native",
3319
3468
  includedTools: ["apply_patch"],
3320
3469
  excludedTools: ["apply_diff", "write_to_file"],
3321
3470
  supportsImages: true,
@@ -3338,6 +3487,7 @@ var openAiNativeModels = {
3338
3487
  maxTokens: 128e3,
3339
3488
  contextWindow: 4e5,
3340
3489
  supportsNativeTools: true,
3490
+ defaultToolProtocol: "native",
3341
3491
  includedTools: ["apply_patch"],
3342
3492
  excludedTools: ["apply_diff", "write_to_file"],
3343
3493
  supportsImages: true,
@@ -3356,6 +3506,7 @@ var openAiNativeModels = {
3356
3506
  maxTokens: 128e3,
3357
3507
  contextWindow: 4e5,
3358
3508
  supportsNativeTools: true,
3509
+ defaultToolProtocol: "native",
3359
3510
  includedTools: ["apply_patch"],
3360
3511
  excludedTools: ["apply_diff", "write_to_file"],
3361
3512
  supportsImages: true,
@@ -3373,6 +3524,7 @@ var openAiNativeModels = {
3373
3524
  maxTokens: 128e3,
3374
3525
  contextWindow: 4e5,
3375
3526
  supportsNativeTools: true,
3527
+ defaultToolProtocol: "native",
3376
3528
  includedTools: ["apply_patch"],
3377
3529
  excludedTools: ["apply_diff", "write_to_file"],
3378
3530
  supportsImages: true,
@@ -3394,6 +3546,7 @@ var openAiNativeModels = {
3394
3546
  maxTokens: 128e3,
3395
3547
  contextWindow: 4e5,
3396
3548
  supportsNativeTools: true,
3549
+ defaultToolProtocol: "native",
3397
3550
  includedTools: ["apply_patch"],
3398
3551
  excludedTools: ["apply_diff", "write_to_file"],
3399
3552
  supportsImages: true,
@@ -3415,6 +3568,7 @@ var openAiNativeModels = {
3415
3568
  maxTokens: 128e3,
3416
3569
  contextWindow: 4e5,
3417
3570
  supportsNativeTools: true,
3571
+ defaultToolProtocol: "native",
3418
3572
  includedTools: ["apply_patch"],
3419
3573
  excludedTools: ["apply_diff", "write_to_file"],
3420
3574
  supportsImages: true,
@@ -3432,6 +3586,7 @@ var openAiNativeModels = {
3432
3586
  maxTokens: 128e3,
3433
3587
  contextWindow: 4e5,
3434
3588
  supportsNativeTools: true,
3589
+ defaultToolProtocol: "native",
3435
3590
  includedTools: ["apply_patch"],
3436
3591
  excludedTools: ["apply_diff", "write_to_file"],
3437
3592
  supportsImages: true,
@@ -3450,6 +3605,7 @@ var openAiNativeModels = {
3450
3605
  maxTokens: 128e3,
3451
3606
  contextWindow: 4e5,
3452
3607
  supportsNativeTools: true,
3608
+ defaultToolProtocol: "native",
3453
3609
  includedTools: ["apply_patch"],
3454
3610
  excludedTools: ["apply_diff", "write_to_file"],
3455
3611
  supportsImages: true,
@@ -3463,6 +3619,7 @@ var openAiNativeModels = {
3463
3619
  maxTokens: 32768,
3464
3620
  contextWindow: 1047576,
3465
3621
  supportsNativeTools: true,
3622
+ defaultToolProtocol: "native",
3466
3623
  includedTools: ["apply_patch"],
3467
3624
  excludedTools: ["apply_diff", "write_to_file"],
3468
3625
  supportsImages: true,
@@ -3479,6 +3636,7 @@ var openAiNativeModels = {
3479
3636
  maxTokens: 32768,
3480
3637
  contextWindow: 1047576,
3481
3638
  supportsNativeTools: true,
3639
+ defaultToolProtocol: "native",
3482
3640
  includedTools: ["apply_patch"],
3483
3641
  excludedTools: ["apply_diff", "write_to_file"],
3484
3642
  supportsImages: true,
@@ -3495,6 +3653,7 @@ var openAiNativeModels = {
3495
3653
  maxTokens: 32768,
3496
3654
  contextWindow: 1047576,
3497
3655
  supportsNativeTools: true,
3656
+ defaultToolProtocol: "native",
3498
3657
  includedTools: ["apply_patch"],
3499
3658
  excludedTools: ["apply_diff", "write_to_file"],
3500
3659
  supportsImages: true,
@@ -3511,6 +3670,7 @@ var openAiNativeModels = {
3511
3670
  maxTokens: 1e5,
3512
3671
  contextWindow: 2e5,
3513
3672
  supportsNativeTools: true,
3673
+ defaultToolProtocol: "native",
3514
3674
  supportsImages: true,
3515
3675
  supportsPromptCache: true,
3516
3676
  inputPrice: 2,
@@ -3528,6 +3688,7 @@ var openAiNativeModels = {
3528
3688
  maxTokens: 1e5,
3529
3689
  contextWindow: 2e5,
3530
3690
  supportsNativeTools: true,
3691
+ defaultToolProtocol: "native",
3531
3692
  supportsImages: true,
3532
3693
  supportsPromptCache: true,
3533
3694
  inputPrice: 2,
@@ -3540,6 +3701,7 @@ var openAiNativeModels = {
3540
3701
  maxTokens: 1e5,
3541
3702
  contextWindow: 2e5,
3542
3703
  supportsNativeTools: true,
3704
+ defaultToolProtocol: "native",
3543
3705
  supportsImages: true,
3544
3706
  supportsPromptCache: true,
3545
3707
  inputPrice: 2,
@@ -3552,6 +3714,7 @@ var openAiNativeModels = {
3552
3714
  maxTokens: 1e5,
3553
3715
  contextWindow: 2e5,
3554
3716
  supportsNativeTools: true,
3717
+ defaultToolProtocol: "native",
3555
3718
  supportsImages: true,
3556
3719
  supportsPromptCache: true,
3557
3720
  inputPrice: 1.1,
@@ -3569,6 +3732,7 @@ var openAiNativeModels = {
3569
3732
  maxTokens: 1e5,
3570
3733
  contextWindow: 2e5,
3571
3734
  supportsNativeTools: true,
3735
+ defaultToolProtocol: "native",
3572
3736
  supportsImages: true,
3573
3737
  supportsPromptCache: true,
3574
3738
  inputPrice: 1.1,
@@ -3581,6 +3745,7 @@ var openAiNativeModels = {
3581
3745
  maxTokens: 1e5,
3582
3746
  contextWindow: 2e5,
3583
3747
  supportsNativeTools: true,
3748
+ defaultToolProtocol: "native",
3584
3749
  supportsImages: true,
3585
3750
  supportsPromptCache: true,
3586
3751
  inputPrice: 1.1,
@@ -3593,6 +3758,7 @@ var openAiNativeModels = {
3593
3758
  maxTokens: 1e5,
3594
3759
  contextWindow: 2e5,
3595
3760
  supportsNativeTools: true,
3761
+ defaultToolProtocol: "native",
3596
3762
  supportsImages: false,
3597
3763
  supportsPromptCache: true,
3598
3764
  inputPrice: 1.1,
@@ -3606,6 +3772,7 @@ var openAiNativeModels = {
3606
3772
  maxTokens: 1e5,
3607
3773
  contextWindow: 2e5,
3608
3774
  supportsNativeTools: true,
3775
+ defaultToolProtocol: "native",
3609
3776
  supportsImages: false,
3610
3777
  supportsPromptCache: true,
3611
3778
  inputPrice: 1.1,
@@ -3618,6 +3785,7 @@ var openAiNativeModels = {
3618
3785
  maxTokens: 1e5,
3619
3786
  contextWindow: 2e5,
3620
3787
  supportsNativeTools: true,
3788
+ defaultToolProtocol: "native",
3621
3789
  supportsImages: false,
3622
3790
  supportsPromptCache: true,
3623
3791
  inputPrice: 1.1,
@@ -3630,6 +3798,7 @@ var openAiNativeModels = {
3630
3798
  maxTokens: 1e5,
3631
3799
  contextWindow: 2e5,
3632
3800
  supportsNativeTools: true,
3801
+ defaultToolProtocol: "native",
3633
3802
  supportsImages: true,
3634
3803
  supportsPromptCache: true,
3635
3804
  inputPrice: 15,
@@ -3641,6 +3810,7 @@ var openAiNativeModels = {
3641
3810
  maxTokens: 32768,
3642
3811
  contextWindow: 128e3,
3643
3812
  supportsNativeTools: true,
3813
+ defaultToolProtocol: "native",
3644
3814
  supportsImages: true,
3645
3815
  supportsPromptCache: true,
3646
3816
  inputPrice: 15,
@@ -3652,6 +3822,7 @@ var openAiNativeModels = {
3652
3822
  maxTokens: 65536,
3653
3823
  contextWindow: 128e3,
3654
3824
  supportsNativeTools: true,
3825
+ defaultToolProtocol: "native",
3655
3826
  supportsImages: true,
3656
3827
  supportsPromptCache: true,
3657
3828
  inputPrice: 1.1,
@@ -3663,6 +3834,7 @@ var openAiNativeModels = {
3663
3834
  maxTokens: 16384,
3664
3835
  contextWindow: 128e3,
3665
3836
  supportsNativeTools: true,
3837
+ defaultToolProtocol: "native",
3666
3838
  supportsImages: true,
3667
3839
  supportsPromptCache: true,
3668
3840
  inputPrice: 2.5,
@@ -3677,6 +3849,7 @@ var openAiNativeModels = {
3677
3849
  maxTokens: 16384,
3678
3850
  contextWindow: 128e3,
3679
3851
  supportsNativeTools: true,
3852
+ defaultToolProtocol: "native",
3680
3853
  supportsImages: true,
3681
3854
  supportsPromptCache: true,
3682
3855
  inputPrice: 0.15,
@@ -3691,6 +3864,7 @@ var openAiNativeModels = {
3691
3864
  maxTokens: 16384,
3692
3865
  contextWindow: 2e5,
3693
3866
  supportsNativeTools: true,
3867
+ defaultToolProtocol: "native",
3694
3868
  supportsImages: false,
3695
3869
  supportsPromptCache: false,
3696
3870
  inputPrice: 1.5,
@@ -3704,6 +3878,7 @@ var openAiNativeModels = {
3704
3878
  maxTokens: 128e3,
3705
3879
  contextWindow: 4e5,
3706
3880
  supportsNativeTools: true,
3881
+ defaultToolProtocol: "native",
3707
3882
  includedTools: ["apply_patch"],
3708
3883
  excludedTools: ["apply_diff", "write_to_file"],
3709
3884
  supportsImages: true,
@@ -3725,6 +3900,7 @@ var openAiNativeModels = {
3725
3900
  maxTokens: 128e3,
3726
3901
  contextWindow: 4e5,
3727
3902
  supportsNativeTools: true,
3903
+ defaultToolProtocol: "native",
3728
3904
  includedTools: ["apply_patch"],
3729
3905
  excludedTools: ["apply_diff", "write_to_file"],
3730
3906
  supportsImages: true,
@@ -3746,6 +3922,7 @@ var openAiNativeModels = {
3746
3922
  maxTokens: 128e3,
3747
3923
  contextWindow: 4e5,
3748
3924
  supportsNativeTools: true,
3925
+ defaultToolProtocol: "native",
3749
3926
  includedTools: ["apply_patch"],
3750
3927
  excludedTools: ["apply_diff", "write_to_file"],
3751
3928
  supportsImages: true,
@@ -3768,7 +3945,8 @@ var openAiModelInfoSaneDefaults = {
3768
3945
  supportsPromptCache: false,
3769
3946
  inputPrice: 0,
3770
3947
  outputPrice: 0,
3771
- supportsNativeTools: true
3948
+ supportsNativeTools: true,
3949
+ defaultToolProtocol: "native"
3772
3950
  };
3773
3951
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3774
3952
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3855,6 +4033,8 @@ var qwenCodeModels = {
3855
4033
  contextWindow: 1e6,
3856
4034
  supportsImages: false,
3857
4035
  supportsPromptCache: false,
4036
+ supportsNativeTools: true,
4037
+ defaultToolProtocol: "native",
3858
4038
  inputPrice: 0,
3859
4039
  outputPrice: 0,
3860
4040
  cacheWritesPrice: 0,
@@ -3866,6 +4046,8 @@ var qwenCodeModels = {
3866
4046
  contextWindow: 1e6,
3867
4047
  supportsImages: false,
3868
4048
  supportsPromptCache: false,
4049
+ supportsNativeTools: true,
4050
+ defaultToolProtocol: "native",
3869
4051
  inputPrice: 0,
3870
4052
  outputPrice: 0,
3871
4053
  cacheWritesPrice: 0,
@@ -3881,6 +4063,8 @@ var requestyDefaultModelInfo = {
3881
4063
  contextWindow: 2e5,
3882
4064
  supportsImages: true,
3883
4065
  supportsPromptCache: true,
4066
+ supportsNativeTools: true,
4067
+ defaultToolProtocol: "native",
3884
4068
  inputPrice: 3,
3885
4069
  outputPrice: 15,
3886
4070
  cacheWritesPrice: 3.75,
@@ -3937,6 +4121,7 @@ var sambaNovaModels = {
3937
4121
  supportsImages: false,
3938
4122
  supportsPromptCache: false,
3939
4123
  supportsNativeTools: true,
4124
+ defaultToolProtocol: "native",
3940
4125
  inputPrice: 0.1,
3941
4126
  outputPrice: 0.2,
3942
4127
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3947,6 +4132,7 @@ var sambaNovaModels = {
3947
4132
  supportsImages: false,
3948
4133
  supportsPromptCache: false,
3949
4134
  supportsNativeTools: true,
4135
+ defaultToolProtocol: "native",
3950
4136
  inputPrice: 0.6,
3951
4137
  outputPrice: 1.2,
3952
4138
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3958,6 +4144,7 @@ var sambaNovaModels = {
3958
4144
  supportsPromptCache: false,
3959
4145
  supportsReasoningBudget: true,
3960
4146
  supportsNativeTools: true,
4147
+ defaultToolProtocol: "native",
3961
4148
  inputPrice: 5,
3962
4149
  outputPrice: 7,
3963
4150
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3968,6 +4155,7 @@ var sambaNovaModels = {
3968
4155
  supportsImages: false,
3969
4156
  supportsPromptCache: false,
3970
4157
  supportsNativeTools: true,
4158
+ defaultToolProtocol: "native",
3971
4159
  inputPrice: 3,
3972
4160
  outputPrice: 4.5,
3973
4161
  description: "DeepSeek V3 model with 32K context window."
@@ -3978,6 +4166,7 @@ var sambaNovaModels = {
3978
4166
  supportsImages: false,
3979
4167
  supportsPromptCache: false,
3980
4168
  supportsNativeTools: true,
4169
+ defaultToolProtocol: "native",
3981
4170
  inputPrice: 3,
3982
4171
  outputPrice: 4.5,
3983
4172
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3997,6 +4186,7 @@ var sambaNovaModels = {
3997
4186
  supportsImages: true,
3998
4187
  supportsPromptCache: false,
3999
4188
  supportsNativeTools: true,
4189
+ defaultToolProtocol: "native",
4000
4190
  inputPrice: 0.63,
4001
4191
  outputPrice: 1.8,
4002
4192
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -4016,6 +4206,7 @@ var sambaNovaModels = {
4016
4206
  supportsImages: false,
4017
4207
  supportsPromptCache: false,
4018
4208
  supportsNativeTools: true,
4209
+ defaultToolProtocol: "native",
4019
4210
  inputPrice: 0.4,
4020
4211
  outputPrice: 0.8,
4021
4212
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -4026,6 +4217,7 @@ var sambaNovaModels = {
4026
4217
  supportsImages: false,
4027
4218
  supportsPromptCache: false,
4028
4219
  supportsNativeTools: true,
4220
+ defaultToolProtocol: "native",
4029
4221
  inputPrice: 0.22,
4030
4222
  outputPrice: 0.59,
4031
4223
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -4054,9 +4246,12 @@ var vertexModels = {
4054
4246
  contextWindow: 1048576,
4055
4247
  supportsImages: true,
4056
4248
  supportsNativeTools: true,
4249
+ defaultToolProtocol: "native",
4057
4250
  supportsPromptCache: true,
4058
4251
  supportsReasoningEffort: ["low", "high"],
4059
4252
  reasoningEffort: "low",
4253
+ includedTools: ["write_file", "edit_file"],
4254
+ excludedTools: ["apply_diff"],
4060
4255
  supportsTemperature: true,
4061
4256
  defaultTemperature: 1,
4062
4257
  inputPrice: 4,
@@ -4074,12 +4269,33 @@ var vertexModels = {
4074
4269
  }
4075
4270
  ]
4076
4271
  },
4272
+ "gemini-3-flash-preview": {
4273
+ maxTokens: 65536,
4274
+ contextWindow: 1048576,
4275
+ supportsImages: true,
4276
+ supportsNativeTools: true,
4277
+ defaultToolProtocol: "native",
4278
+ supportsPromptCache: true,
4279
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
4280
+ reasoningEffort: "medium",
4281
+ includedTools: ["write_file", "edit_file"],
4282
+ excludedTools: ["apply_diff"],
4283
+ supportsTemperature: true,
4284
+ defaultTemperature: 1,
4285
+ inputPrice: 0.3,
4286
+ outputPrice: 2.5,
4287
+ cacheReadsPrice: 0.075,
4288
+ cacheWritesPrice: 1
4289
+ },
4077
4290
  "gemini-2.5-flash-preview-05-20:thinking": {
4078
4291
  maxTokens: 65535,
4079
4292
  contextWindow: 1048576,
4080
4293
  supportsImages: true,
4081
4294
  supportsNativeTools: true,
4295
+ defaultToolProtocol: "native",
4082
4296
  supportsPromptCache: true,
4297
+ includedTools: ["write_file", "edit_file"],
4298
+ excludedTools: ["apply_diff"],
4083
4299
  inputPrice: 0.15,
4084
4300
  outputPrice: 3.5,
4085
4301
  maxThinkingTokens: 24576,
@@ -4091,7 +4307,10 @@ var vertexModels = {
4091
4307
  contextWindow: 1048576,
4092
4308
  supportsImages: true,
4093
4309
  supportsNativeTools: true,
4310
+ defaultToolProtocol: "native",
4094
4311
  supportsPromptCache: true,
4312
+ includedTools: ["write_file", "edit_file"],
4313
+ excludedTools: ["apply_diff"],
4095
4314
  inputPrice: 0.15,
4096
4315
  outputPrice: 0.6
4097
4316
  },
@@ -4100,7 +4319,10 @@ var vertexModels = {
4100
4319
  contextWindow: 1048576,
4101
4320
  supportsImages: true,
4102
4321
  supportsNativeTools: true,
4322
+ defaultToolProtocol: "native",
4103
4323
  supportsPromptCache: true,
4324
+ includedTools: ["write_file", "edit_file"],
4325
+ excludedTools: ["apply_diff"],
4104
4326
  inputPrice: 0.3,
4105
4327
  outputPrice: 2.5,
4106
4328
  cacheReadsPrice: 0.075,
@@ -4113,7 +4335,10 @@ var vertexModels = {
4113
4335
  contextWindow: 1048576,
4114
4336
  supportsImages: true,
4115
4337
  supportsNativeTools: true,
4338
+ defaultToolProtocol: "native",
4116
4339
  supportsPromptCache: false,
4340
+ includedTools: ["write_file", "edit_file"],
4341
+ excludedTools: ["apply_diff"],
4117
4342
  inputPrice: 0.15,
4118
4343
  outputPrice: 3.5,
4119
4344
  maxThinkingTokens: 24576,
@@ -4125,7 +4350,10 @@ var vertexModels = {
4125
4350
  contextWindow: 1048576,
4126
4351
  supportsImages: true,
4127
4352
  supportsNativeTools: true,
4353
+ defaultToolProtocol: "native",
4128
4354
  supportsPromptCache: false,
4355
+ includedTools: ["write_file", "edit_file"],
4356
+ excludedTools: ["apply_diff"],
4129
4357
  inputPrice: 0.15,
4130
4358
  outputPrice: 0.6
4131
4359
  },
@@ -4134,7 +4362,10 @@ var vertexModels = {
4134
4362
  contextWindow: 1048576,
4135
4363
  supportsImages: true,
4136
4364
  supportsNativeTools: true,
4365
+ defaultToolProtocol: "native",
4137
4366
  supportsPromptCache: true,
4367
+ includedTools: ["write_file", "edit_file"],
4368
+ excludedTools: ["apply_diff"],
4138
4369
  inputPrice: 2.5,
4139
4370
  outputPrice: 15
4140
4371
  },
@@ -4143,7 +4374,10 @@ var vertexModels = {
4143
4374
  contextWindow: 1048576,
4144
4375
  supportsImages: true,
4145
4376
  supportsNativeTools: true,
4377
+ defaultToolProtocol: "native",
4146
4378
  supportsPromptCache: true,
4379
+ includedTools: ["write_file", "edit_file"],
4380
+ excludedTools: ["apply_diff"],
4147
4381
  inputPrice: 2.5,
4148
4382
  outputPrice: 15
4149
4383
  },
@@ -4152,7 +4386,10 @@ var vertexModels = {
4152
4386
  contextWindow: 1048576,
4153
4387
  supportsImages: true,
4154
4388
  supportsNativeTools: true,
4389
+ defaultToolProtocol: "native",
4155
4390
  supportsPromptCache: true,
4391
+ includedTools: ["write_file", "edit_file"],
4392
+ excludedTools: ["apply_diff"],
4156
4393
  inputPrice: 2.5,
4157
4394
  outputPrice: 15,
4158
4395
  maxThinkingTokens: 32768,
@@ -4163,7 +4400,10 @@ var vertexModels = {
4163
4400
  contextWindow: 1048576,
4164
4401
  supportsImages: true,
4165
4402
  supportsNativeTools: true,
4403
+ defaultToolProtocol: "native",
4166
4404
  supportsPromptCache: true,
4405
+ includedTools: ["write_file", "edit_file"],
4406
+ excludedTools: ["apply_diff"],
4167
4407
  inputPrice: 2.5,
4168
4408
  outputPrice: 15,
4169
4409
  maxThinkingTokens: 32768,
@@ -4189,7 +4429,10 @@ var vertexModels = {
4189
4429
  contextWindow: 1048576,
4190
4430
  supportsImages: true,
4191
4431
  supportsNativeTools: true,
4432
+ defaultToolProtocol: "native",
4192
4433
  supportsPromptCache: false,
4434
+ includedTools: ["write_file", "edit_file"],
4435
+ excludedTools: ["apply_diff"],
4193
4436
  inputPrice: 0,
4194
4437
  outputPrice: 0
4195
4438
  },
@@ -4198,7 +4441,10 @@ var vertexModels = {
4198
4441
  contextWindow: 2097152,
4199
4442
  supportsImages: true,
4200
4443
  supportsNativeTools: true,
4444
+ defaultToolProtocol: "native",
4201
4445
  supportsPromptCache: false,
4446
+ includedTools: ["write_file", "edit_file"],
4447
+ excludedTools: ["apply_diff"],
4202
4448
  inputPrice: 0,
4203
4449
  outputPrice: 0
4204
4450
  },
@@ -4207,7 +4453,10 @@ var vertexModels = {
4207
4453
  contextWindow: 1048576,
4208
4454
  supportsImages: true,
4209
4455
  supportsNativeTools: true,
4456
+ defaultToolProtocol: "native",
4210
4457
  supportsPromptCache: true,
4458
+ includedTools: ["write_file", "edit_file"],
4459
+ excludedTools: ["apply_diff"],
4211
4460
  inputPrice: 0.15,
4212
4461
  outputPrice: 0.6
4213
4462
  },
@@ -4216,7 +4465,10 @@ var vertexModels = {
4216
4465
  contextWindow: 1048576,
4217
4466
  supportsImages: true,
4218
4467
  supportsNativeTools: true,
4468
+ defaultToolProtocol: "native",
4219
4469
  supportsPromptCache: false,
4470
+ includedTools: ["write_file", "edit_file"],
4471
+ excludedTools: ["apply_diff"],
4220
4472
  inputPrice: 0.075,
4221
4473
  outputPrice: 0.3
4222
4474
  },
@@ -4225,7 +4477,10 @@ var vertexModels = {
4225
4477
  contextWindow: 32768,
4226
4478
  supportsImages: true,
4227
4479
  supportsNativeTools: true,
4480
+ defaultToolProtocol: "native",
4228
4481
  supportsPromptCache: false,
4482
+ includedTools: ["write_file", "edit_file"],
4483
+ excludedTools: ["apply_diff"],
4229
4484
  inputPrice: 0,
4230
4485
  outputPrice: 0
4231
4486
  },
@@ -4234,7 +4489,10 @@ var vertexModels = {
4234
4489
  contextWindow: 1048576,
4235
4490
  supportsImages: true,
4236
4491
  supportsNativeTools: true,
4492
+ defaultToolProtocol: "native",
4237
4493
  supportsPromptCache: true,
4494
+ includedTools: ["write_file", "edit_file"],
4495
+ excludedTools: ["apply_diff"],
4238
4496
  inputPrice: 0.075,
4239
4497
  outputPrice: 0.3
4240
4498
  },
@@ -4243,37 +4501,86 @@ var vertexModels = {
4243
4501
  contextWindow: 2097152,
4244
4502
  supportsImages: true,
4245
4503
  supportsNativeTools: true,
4504
+ defaultToolProtocol: "native",
4246
4505
  supportsPromptCache: false,
4506
+ includedTools: ["write_file", "edit_file"],
4507
+ excludedTools: ["apply_diff"],
4247
4508
  inputPrice: 1.25,
4248
4509
  outputPrice: 5
4249
4510
  },
4250
4511
  "claude-sonnet-4@20250514": {
4251
4512
  maxTokens: 8192,
4252
4513
  contextWindow: 2e5,
4514
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4253
4515
  supportsImages: true,
4254
4516
  supportsPromptCache: true,
4517
+ supportsNativeTools: true,
4518
+ defaultToolProtocol: "native",
4255
4519
  inputPrice: 3,
4520
+ // $3 per million input tokens (≤200K context)
4256
4521
  outputPrice: 15,
4522
+ // $15 per million output tokens (≤200K context)
4257
4523
  cacheWritesPrice: 3.75,
4524
+ // $3.75 per million tokens
4258
4525
  cacheReadsPrice: 0.3,
4259
- supportsReasoningBudget: true
4526
+ // $0.30 per million tokens
4527
+ supportsReasoningBudget: true,
4528
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
4529
+ tiers: [
4530
+ {
4531
+ contextWindow: 1e6,
4532
+ // 1M tokens with beta flag
4533
+ inputPrice: 6,
4534
+ // $6 per million input tokens (>200K context)
4535
+ outputPrice: 22.5,
4536
+ // $22.50 per million output tokens (>200K context)
4537
+ cacheWritesPrice: 7.5,
4538
+ // $7.50 per million tokens (>200K context)
4539
+ cacheReadsPrice: 0.6
4540
+ // $0.60 per million tokens (>200K context)
4541
+ }
4542
+ ]
4260
4543
  },
4261
4544
  "claude-sonnet-4-5@20250929": {
4262
4545
  maxTokens: 8192,
4263
4546
  contextWindow: 2e5,
4547
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4264
4548
  supportsImages: true,
4265
4549
  supportsPromptCache: true,
4550
+ supportsNativeTools: true,
4551
+ defaultToolProtocol: "native",
4266
4552
  inputPrice: 3,
4553
+ // $3 per million input tokens (≤200K context)
4267
4554
  outputPrice: 15,
4555
+ // $15 per million output tokens (≤200K context)
4268
4556
  cacheWritesPrice: 3.75,
4557
+ // $3.75 per million tokens
4269
4558
  cacheReadsPrice: 0.3,
4270
- supportsReasoningBudget: true
4559
+ // $0.30 per million tokens
4560
+ supportsReasoningBudget: true,
4561
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
4562
+ tiers: [
4563
+ {
4564
+ contextWindow: 1e6,
4565
+ // 1M tokens with beta flag
4566
+ inputPrice: 6,
4567
+ // $6 per million input tokens (>200K context)
4568
+ outputPrice: 22.5,
4569
+ // $22.50 per million output tokens (>200K context)
4570
+ cacheWritesPrice: 7.5,
4571
+ // $7.50 per million tokens (>200K context)
4572
+ cacheReadsPrice: 0.6
4573
+ // $0.60 per million tokens (>200K context)
4574
+ }
4575
+ ]
4271
4576
  },
4272
4577
  "claude-haiku-4-5@20251001": {
4273
4578
  maxTokens: 8192,
4274
4579
  contextWindow: 2e5,
4275
4580
  supportsImages: true,
4276
4581
  supportsPromptCache: true,
4582
+ supportsNativeTools: true,
4583
+ defaultToolProtocol: "native",
4277
4584
  inputPrice: 1,
4278
4585
  outputPrice: 5,
4279
4586
  cacheWritesPrice: 1.25,
@@ -4285,6 +4592,8 @@ var vertexModels = {
4285
4592
  contextWindow: 2e5,
4286
4593
  supportsImages: true,
4287
4594
  supportsPromptCache: true,
4595
+ supportsNativeTools: true,
4596
+ defaultToolProtocol: "native",
4288
4597
  inputPrice: 5,
4289
4598
  outputPrice: 25,
4290
4599
  cacheWritesPrice: 6.25,
@@ -4296,6 +4605,8 @@ var vertexModels = {
4296
4605
  contextWindow: 2e5,
4297
4606
  supportsImages: true,
4298
4607
  supportsPromptCache: true,
4608
+ supportsNativeTools: true,
4609
+ defaultToolProtocol: "native",
4299
4610
  inputPrice: 15,
4300
4611
  outputPrice: 75,
4301
4612
  cacheWritesPrice: 18.75,
@@ -4307,6 +4618,8 @@ var vertexModels = {
4307
4618
  contextWindow: 2e5,
4308
4619
  supportsImages: true,
4309
4620
  supportsPromptCache: true,
4621
+ supportsNativeTools: true,
4622
+ defaultToolProtocol: "native",
4310
4623
  inputPrice: 15,
4311
4624
  outputPrice: 75,
4312
4625
  cacheWritesPrice: 18.75,
@@ -4317,6 +4630,8 @@ var vertexModels = {
4317
4630
  contextWindow: 2e5,
4318
4631
  supportsImages: true,
4319
4632
  supportsPromptCache: true,
4633
+ supportsNativeTools: true,
4634
+ defaultToolProtocol: "native",
4320
4635
  inputPrice: 3,
4321
4636
  outputPrice: 15,
4322
4637
  cacheWritesPrice: 3.75,
@@ -4329,6 +4644,8 @@ var vertexModels = {
4329
4644
  contextWindow: 2e5,
4330
4645
  supportsImages: true,
4331
4646
  supportsPromptCache: true,
4647
+ supportsNativeTools: true,
4648
+ defaultToolProtocol: "native",
4332
4649
  inputPrice: 3,
4333
4650
  outputPrice: 15,
4334
4651
  cacheWritesPrice: 3.75,
@@ -4339,6 +4656,8 @@ var vertexModels = {
4339
4656
  contextWindow: 2e5,
4340
4657
  supportsImages: true,
4341
4658
  supportsPromptCache: true,
4659
+ supportsNativeTools: true,
4660
+ defaultToolProtocol: "native",
4342
4661
  inputPrice: 3,
4343
4662
  outputPrice: 15,
4344
4663
  cacheWritesPrice: 3.75,
@@ -4349,6 +4668,8 @@ var vertexModels = {
4349
4668
  contextWindow: 2e5,
4350
4669
  supportsImages: true,
4351
4670
  supportsPromptCache: true,
4671
+ supportsNativeTools: true,
4672
+ defaultToolProtocol: "native",
4352
4673
  inputPrice: 3,
4353
4674
  outputPrice: 15,
4354
4675
  cacheWritesPrice: 3.75,
@@ -4359,6 +4680,8 @@ var vertexModels = {
4359
4680
  contextWindow: 2e5,
4360
4681
  supportsImages: false,
4361
4682
  supportsPromptCache: true,
4683
+ supportsNativeTools: true,
4684
+ defaultToolProtocol: "native",
4362
4685
  inputPrice: 1,
4363
4686
  outputPrice: 5,
4364
4687
  cacheWritesPrice: 1.25,
@@ -4369,6 +4692,8 @@ var vertexModels = {
4369
4692
  contextWindow: 2e5,
4370
4693
  supportsImages: true,
4371
4694
  supportsPromptCache: true,
4695
+ supportsNativeTools: true,
4696
+ defaultToolProtocol: "native",
4372
4697
  inputPrice: 15,
4373
4698
  outputPrice: 75,
4374
4699
  cacheWritesPrice: 18.75,
@@ -4379,6 +4704,8 @@ var vertexModels = {
4379
4704
  contextWindow: 2e5,
4380
4705
  supportsImages: true,
4381
4706
  supportsPromptCache: true,
4707
+ supportsNativeTools: true,
4708
+ defaultToolProtocol: "native",
4382
4709
  inputPrice: 0.25,
4383
4710
  outputPrice: 1.25,
4384
4711
  cacheWritesPrice: 0.3,
@@ -4389,7 +4716,10 @@ var vertexModels = {
4389
4716
  contextWindow: 1048576,
4390
4717
  supportsImages: true,
4391
4718
  supportsNativeTools: true,
4719
+ defaultToolProtocol: "native",
4392
4720
  supportsPromptCache: true,
4721
+ includedTools: ["write_file", "edit_file"],
4722
+ excludedTools: ["apply_diff"],
4393
4723
  inputPrice: 0.1,
4394
4724
  outputPrice: 0.4,
4395
4725
  cacheReadsPrice: 0.025,
@@ -4461,6 +4791,7 @@ var vertexModels = {
4461
4791
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4462
4792
  }
4463
4793
  };
4794
+ var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
4464
4795
  var VERTEX_REGIONS = [
4465
4796
  { value: "global", label: "global" },
4466
4797
  { value: "us-central1", label: "us-central1" },
@@ -4691,6 +5022,7 @@ var xaiModels = {
4691
5022
  supportsImages: true,
4692
5023
  supportsPromptCache: true,
4693
5024
  supportsNativeTools: true,
5025
+ defaultToolProtocol: "native",
4694
5026
  inputPrice: 0.2,
4695
5027
  outputPrice: 1.5,
4696
5028
  cacheWritesPrice: 0.02,
@@ -4705,6 +5037,7 @@ var xaiModels = {
4705
5037
  supportsImages: true,
4706
5038
  supportsPromptCache: true,
4707
5039
  supportsNativeTools: true,
5040
+ defaultToolProtocol: "native",
4708
5041
  inputPrice: 0.2,
4709
5042
  outputPrice: 0.5,
4710
5043
  cacheWritesPrice: 0.05,
@@ -4719,6 +5052,7 @@ var xaiModels = {
4719
5052
  supportsImages: true,
4720
5053
  supportsPromptCache: true,
4721
5054
  supportsNativeTools: true,
5055
+ defaultToolProtocol: "native",
4722
5056
  inputPrice: 0.2,
4723
5057
  outputPrice: 0.5,
4724
5058
  cacheWritesPrice: 0.05,
@@ -4733,6 +5067,7 @@ var xaiModels = {
4733
5067
  supportsImages: true,
4734
5068
  supportsPromptCache: true,
4735
5069
  supportsNativeTools: true,
5070
+ defaultToolProtocol: "native",
4736
5071
  inputPrice: 0.2,
4737
5072
  outputPrice: 0.5,
4738
5073
  cacheWritesPrice: 0.05,
@@ -4747,6 +5082,7 @@ var xaiModels = {
4747
5082
  supportsImages: true,
4748
5083
  supportsPromptCache: true,
4749
5084
  supportsNativeTools: true,
5085
+ defaultToolProtocol: "native",
4750
5086
  inputPrice: 0.2,
4751
5087
  outputPrice: 0.5,
4752
5088
  cacheWritesPrice: 0.05,
@@ -4761,6 +5097,7 @@ var xaiModels = {
4761
5097
  supportsImages: true,
4762
5098
  supportsPromptCache: true,
4763
5099
  supportsNativeTools: true,
5100
+ defaultToolProtocol: "native",
4764
5101
  inputPrice: 3,
4765
5102
  outputPrice: 15,
4766
5103
  cacheWritesPrice: 0.75,
@@ -4775,6 +5112,7 @@ var xaiModels = {
4775
5112
  supportsImages: true,
4776
5113
  supportsPromptCache: true,
4777
5114
  supportsNativeTools: true,
5115
+ defaultToolProtocol: "native",
4778
5116
  inputPrice: 0.3,
4779
5117
  outputPrice: 0.5,
4780
5118
  cacheWritesPrice: 0.07,
@@ -4791,6 +5129,7 @@ var xaiModels = {
4791
5129
  supportsImages: true,
4792
5130
  supportsPromptCache: true,
4793
5131
  supportsNativeTools: true,
5132
+ defaultToolProtocol: "native",
4794
5133
  inputPrice: 3,
4795
5134
  outputPrice: 15,
4796
5135
  cacheWritesPrice: 0.75,
@@ -4904,6 +5243,7 @@ var internationalZAiModels = {
4904
5243
  supportsImages: false,
4905
5244
  supportsPromptCache: true,
4906
5245
  supportsNativeTools: true,
5246
+ defaultToolProtocol: "native",
4907
5247
  inputPrice: 0.6,
4908
5248
  outputPrice: 2.2,
4909
5249
  cacheWritesPrice: 0,
@@ -4916,6 +5256,7 @@ var internationalZAiModels = {
4916
5256
  supportsImages: false,
4917
5257
  supportsPromptCache: true,
4918
5258
  supportsNativeTools: true,
5259
+ defaultToolProtocol: "native",
4919
5260
  inputPrice: 0.2,
4920
5261
  outputPrice: 1.1,
4921
5262
  cacheWritesPrice: 0,
@@ -4928,6 +5269,7 @@ var internationalZAiModels = {
4928
5269
  supportsImages: false,
4929
5270
  supportsPromptCache: true,
4930
5271
  supportsNativeTools: true,
5272
+ defaultToolProtocol: "native",
4931
5273
  inputPrice: 2.2,
4932
5274
  outputPrice: 8.9,
4933
5275
  cacheWritesPrice: 0,
@@ -4940,6 +5282,7 @@ var internationalZAiModels = {
4940
5282
  supportsImages: false,
4941
5283
  supportsPromptCache: true,
4942
5284
  supportsNativeTools: true,
5285
+ defaultToolProtocol: "native",
4943
5286
  inputPrice: 1.1,
4944
5287
  outputPrice: 4.5,
4945
5288
  cacheWritesPrice: 0,
@@ -4952,6 +5295,7 @@ var internationalZAiModels = {
4952
5295
  supportsImages: false,
4953
5296
  supportsPromptCache: true,
4954
5297
  supportsNativeTools: true,
5298
+ defaultToolProtocol: "native",
4955
5299
  inputPrice: 0,
4956
5300
  outputPrice: 0,
4957
5301
  cacheWritesPrice: 0,
@@ -4964,6 +5308,7 @@ var internationalZAiModels = {
4964
5308
  supportsImages: true,
4965
5309
  supportsPromptCache: true,
4966
5310
  supportsNativeTools: true,
5311
+ defaultToolProtocol: "native",
4967
5312
  inputPrice: 0.6,
4968
5313
  outputPrice: 1.8,
4969
5314
  cacheWritesPrice: 0,
@@ -4976,6 +5321,7 @@ var internationalZAiModels = {
4976
5321
  supportsImages: false,
4977
5322
  supportsPromptCache: true,
4978
5323
  supportsNativeTools: true,
5324
+ defaultToolProtocol: "native",
4979
5325
  inputPrice: 0.6,
4980
5326
  outputPrice: 2.2,
4981
5327
  cacheWritesPrice: 0,
@@ -4988,6 +5334,7 @@ var internationalZAiModels = {
4988
5334
  supportsImages: false,
4989
5335
  supportsPromptCache: false,
4990
5336
  supportsNativeTools: true,
5337
+ defaultToolProtocol: "native",
4991
5338
  inputPrice: 0.1,
4992
5339
  outputPrice: 0.1,
4993
5340
  cacheWritesPrice: 0,
@@ -5003,6 +5350,7 @@ var mainlandZAiModels = {
5003
5350
  supportsImages: false,
5004
5351
  supportsPromptCache: true,
5005
5352
  supportsNativeTools: true,
5353
+ defaultToolProtocol: "native",
5006
5354
  inputPrice: 0.29,
5007
5355
  outputPrice: 1.14,
5008
5356
  cacheWritesPrice: 0,
@@ -5015,6 +5363,7 @@ var mainlandZAiModels = {
5015
5363
  supportsImages: false,
5016
5364
  supportsPromptCache: true,
5017
5365
  supportsNativeTools: true,
5366
+ defaultToolProtocol: "native",
5018
5367
  inputPrice: 0.1,
5019
5368
  outputPrice: 0.6,
5020
5369
  cacheWritesPrice: 0,
@@ -5027,6 +5376,7 @@ var mainlandZAiModels = {
5027
5376
  supportsImages: false,
5028
5377
  supportsPromptCache: true,
5029
5378
  supportsNativeTools: true,
5379
+ defaultToolProtocol: "native",
5030
5380
  inputPrice: 0.29,
5031
5381
  outputPrice: 1.14,
5032
5382
  cacheWritesPrice: 0,
@@ -5039,6 +5389,7 @@ var mainlandZAiModels = {
5039
5389
  supportsImages: false,
5040
5390
  supportsPromptCache: true,
5041
5391
  supportsNativeTools: true,
5392
+ defaultToolProtocol: "native",
5042
5393
  inputPrice: 0.1,
5043
5394
  outputPrice: 0.6,
5044
5395
  cacheWritesPrice: 0,
@@ -5051,6 +5402,7 @@ var mainlandZAiModels = {
5051
5402
  supportsImages: false,
5052
5403
  supportsPromptCache: true,
5053
5404
  supportsNativeTools: true,
5405
+ defaultToolProtocol: "native",
5054
5406
  inputPrice: 0,
5055
5407
  outputPrice: 0,
5056
5408
  cacheWritesPrice: 0,
@@ -5063,6 +5415,7 @@ var mainlandZAiModels = {
5063
5415
  supportsImages: true,
5064
5416
  supportsPromptCache: true,
5065
5417
  supportsNativeTools: true,
5418
+ defaultToolProtocol: "native",
5066
5419
  inputPrice: 0.29,
5067
5420
  outputPrice: 0.93,
5068
5421
  cacheWritesPrice: 0,
@@ -5075,6 +5428,7 @@ var mainlandZAiModels = {
5075
5428
  supportsImages: false,
5076
5429
  supportsPromptCache: true,
5077
5430
  supportsNativeTools: true,
5431
+ defaultToolProtocol: "native",
5078
5432
  inputPrice: 0.29,
5079
5433
  outputPrice: 1.14,
5080
5434
  cacheWritesPrice: 0,
@@ -5128,6 +5482,7 @@ var minimaxModels = {
5128
5482
  supportsImages: false,
5129
5483
  supportsPromptCache: true,
5130
5484
  supportsNativeTools: true,
5485
+ defaultToolProtocol: "native",
5131
5486
  preserveReasoning: true,
5132
5487
  inputPrice: 0.3,
5133
5488
  outputPrice: 1.2,
@@ -5141,6 +5496,7 @@ var minimaxModels = {
5141
5496
  supportsImages: false,
5142
5497
  supportsPromptCache: true,
5143
5498
  supportsNativeTools: true,
5499
+ defaultToolProtocol: "native",
5144
5500
  preserveReasoning: true,
5145
5501
  inputPrice: 0.3,
5146
5502
  outputPrice: 1.2,
@@ -5323,10 +5679,7 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5323
5679
  anthropicBeta1MContext: import_zod8.z.boolean().optional()
5324
5680
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5325
5681
  });
5326
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
5327
- claudeCodePath: import_zod8.z.string().optional(),
5328
- claudeCodeMaxOutputTokens: import_zod8.z.number().int().min(1).max(2e5).optional()
5329
- });
5682
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5330
5683
  var openRouterSchema = baseProviderSettingsSchema.extend({
5331
5684
  openRouterApiKey: import_zod8.z.string().optional(),
5332
5685
  openRouterModelId: import_zod8.z.string().optional(),
@@ -5362,7 +5715,9 @@ var vertexSchema = apiModelIdProviderModelSchema.extend({
5362
5715
  vertexProjectId: import_zod8.z.string().optional(),
5363
5716
  vertexRegion: import_zod8.z.string().optional(),
5364
5717
  enableUrlContext: import_zod8.z.boolean().optional(),
5365
- enableGrounding: import_zod8.z.boolean().optional()
5718
+ enableGrounding: import_zod8.z.boolean().optional(),
5719
+ vertex1MContext: import_zod8.z.boolean().optional()
5720
+ // Enable 'context-1m-2025-08-07' beta for 1M context window.
5366
5721
  });
5367
5722
  var openAiSchema = baseProviderSettingsSchema.extend({
5368
5723
  openAiBaseUrl: import_zod8.z.string().optional(),
@@ -5779,6 +6134,16 @@ var historyItemSchema = import_zod9.z.object({
5779
6134
  size: import_zod9.z.number().optional(),
5780
6135
  workspace: import_zod9.z.string().optional(),
5781
6136
  mode: import_zod9.z.string().optional(),
6137
+ /**
6138
+ * The tool protocol used by this task. Once a task uses tools with a specific
6139
+ * protocol (XML or Native), it is permanently locked to that protocol.
6140
+ *
6141
+ * - "xml": Tool calls are parsed from XML text (no tool IDs)
6142
+ * - "native": Tool calls come as tool_call chunks with IDs
6143
+ *
6144
+ * This ensures task resumption works correctly even when NTC settings change.
6145
+ */
6146
+ toolProtocol: import_zod9.z.enum(["xml", "native"]).optional(),
5782
6147
  status: import_zod9.z.enum(["active", "completed", "delegated"]).optional(),
5783
6148
  delegatedToId: import_zod9.z.string().optional(),
5784
6149
  // Last child this parent delegated to
@@ -6001,11 +6366,44 @@ function getErrorStatusCode(error) {
6001
6366
  }
6002
6367
  return void 0;
6003
6368
  }
6369
+ function extractMessageFromJsonPayload(message) {
6370
+ const jsonStartIndex = message.indexOf("{");
6371
+ if (jsonStartIndex === -1) {
6372
+ return void 0;
6373
+ }
6374
+ const potentialJson = message.slice(jsonStartIndex);
6375
+ try {
6376
+ const parsed = JSON.parse(potentialJson);
6377
+ if (parsed?.error?.message && typeof parsed.error.message === "string") {
6378
+ return parsed.error.message;
6379
+ }
6380
+ if (parsed?.message && typeof parsed.message === "string") {
6381
+ return parsed.message;
6382
+ }
6383
+ } catch {
6384
+ }
6385
+ return void 0;
6386
+ }
6004
6387
  function getErrorMessage(error) {
6388
+ let message;
6005
6389
  if (isOpenAISdkError(error)) {
6006
- return error.error?.metadata?.raw || error.error?.message || error.message;
6390
+ message = error.error?.metadata?.raw || error.error?.message || error.message;
6391
+ } else if (error instanceof Error) {
6392
+ message = error.message;
6393
+ } else if (typeof error === "object" && error !== null && "message" in error) {
6394
+ const msgValue = error.message;
6395
+ if (typeof msgValue === "string") {
6396
+ message = msgValue;
6397
+ }
6007
6398
  }
6008
- return void 0;
6399
+ if (!message) {
6400
+ return void 0;
6401
+ }
6402
+ const extractedMessage = extractMessageFromJsonPayload(message);
6403
+ if (extractedMessage) {
6404
+ return extractedMessage;
6405
+ }
6406
+ return message;
6009
6407
  }
6010
6408
  function shouldReportApiErrorToTelemetry(errorCode, errorMessage) {
6011
6409
  if (errorCode !== void 0 && EXPECTED_API_ERROR_CODES.has(errorCode)) {
@@ -6041,6 +6439,31 @@ function extractApiProviderErrorProperties(error) {
6041
6439
  ...error.errorCode !== void 0 && { errorCode: error.errorCode }
6042
6440
  };
6043
6441
  }
6442
+ var ConsecutiveMistakeError = class extends Error {
6443
+ constructor(message, taskId, consecutiveMistakeCount, consecutiveMistakeLimit, reason = "unknown", provider, modelId) {
6444
+ super(message);
6445
+ this.taskId = taskId;
6446
+ this.consecutiveMistakeCount = consecutiveMistakeCount;
6447
+ this.consecutiveMistakeLimit = consecutiveMistakeLimit;
6448
+ this.reason = reason;
6449
+ this.provider = provider;
6450
+ this.modelId = modelId;
6451
+ this.name = "ConsecutiveMistakeError";
6452
+ }
6453
+ };
6454
+ function isConsecutiveMistakeError(error) {
6455
+ return error instanceof Error && error.name === "ConsecutiveMistakeError" && "taskId" in error && "consecutiveMistakeCount" in error && "consecutiveMistakeLimit" in error;
6456
+ }
6457
+ function extractConsecutiveMistakeErrorProperties(error) {
6458
+ return {
6459
+ taskId: error.taskId,
6460
+ consecutiveMistakeCount: error.consecutiveMistakeCount,
6461
+ consecutiveMistakeLimit: error.consecutiveMistakeLimit,
6462
+ reason: error.reason,
6463
+ ...error.provider !== void 0 && { provider: error.provider },
6464
+ ...error.modelId !== void 0 && { modelId: error.modelId }
6465
+ };
6466
+ }
6044
6467
 
6045
6468
  // src/mode.ts
6046
6469
  var import_zod12 = require("zod");
@@ -6537,7 +6960,7 @@ var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6537
6960
  terminalShellIntegrationTimeout: import_zod16.z.number().int().nonnegative().optional()
6538
6961
  })
6539
6962
  );
6540
- var workspaceTaskVisibilitySchema = import_zod16.z.enum(["all", "list-only", "full-lockdown"]);
6963
+ var workspaceTaskVisibilitySchema = import_zod16.z.enum(["all", "list-only", "admins-and-creator", "creator", "full-lockdown"]);
6541
6964
  var organizationCloudSettingsSchema = import_zod16.z.object({
6542
6965
  recordTaskMessages: import_zod16.z.boolean().optional(),
6543
6966
  enableTaskSharing: import_zod16.z.boolean().optional(),
@@ -6925,6 +7348,12 @@ var COOKIE_CONSENT_EVENTS = {
6925
7348
  CHANGED: "cookieConsentChanged"
6926
7349
  };
6927
7350
 
7351
+ // src/custom-tool.ts
7352
+ var import_v4 = require("zod/v4");
7353
+ function defineCustomTool(definition) {
7354
+ return definition;
7355
+ }
7356
+
6928
7357
  // src/followup.ts
6929
7358
  var import_zod17 = require("zod");
6930
7359
  var suggestionItemSchema = import_zod17.z.object({
@@ -7120,12 +7549,12 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7120
7549
  BEDROCK_REGIONS,
7121
7550
  BEDROCK_SERVICE_TIER_MODEL_IDS,
7122
7551
  BEDROCK_SERVICE_TIER_PRICING,
7123
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
7124
7552
  CODEBASE_INDEX_DEFAULTS,
7125
7553
  CONSENT_COOKIE_NAME,
7126
7554
  CONTEXT_MANAGEMENT_EVENTS,
7127
7555
  COOKIE_CONSENT_EVENTS,
7128
7556
  ConnectionState,
7557
+ ConsecutiveMistakeError,
7129
7558
  DEEP_SEEK_DEFAULT_TEMPERATURE,
7130
7559
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
7131
7560
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
@@ -7166,6 +7595,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7166
7595
  MISTRAL_DEFAULT_TEMPERATURE,
7167
7596
  MODELS_BY_PROVIDER,
7168
7597
  MOONSHOT_DEFAULT_TEMPERATURE,
7598
+ NATIVE_TOOL_DEFAULTS,
7169
7599
  OPENAI_AZURE_AI_INFERENCE_PATH,
7170
7600
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
7171
7601
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -7191,6 +7621,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7191
7621
  VERCEL_AI_GATEWAY_PROMPT_CACHING_MODELS,
7192
7622
  VERCEL_AI_GATEWAY_VISION_AND_TOOLS_MODELS,
7193
7623
  VERCEL_AI_GATEWAY_VISION_ONLY_MODELS,
7624
+ VERTEX_1M_CONTEXT_MODEL_IDS,
7194
7625
  VERTEX_REGIONS,
7195
7626
  ZAI_DEFAULT_TEMPERATURE,
7196
7627
  ackSchema,
@@ -7210,6 +7641,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7210
7641
  chutesModels,
7211
7642
  claudeCodeDefaultModelId,
7212
7643
  claudeCodeModels,
7644
+ claudeCodeReasoningConfig,
7213
7645
  clineAskSchema,
7214
7646
  clineAsks,
7215
7647
  clineMessageSchema,
@@ -7224,7 +7656,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7224
7656
  commandIds,
7225
7657
  contextCondenseSchema,
7226
7658
  contextTruncationSchema,
7227
- convertModelNameForVertex,
7228
7659
  customModePromptsSchema,
7229
7660
  customModesSettingsSchema,
7230
7661
  customProviders,
@@ -7233,6 +7664,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7233
7664
  deepInfraDefaultModelInfo,
7234
7665
  deepSeekDefaultModelId,
7235
7666
  deepSeekModels,
7667
+ defineCustomTool,
7236
7668
  discriminatedProviderSettingsWithIdSchema,
7237
7669
  doubaoDefaultModelId,
7238
7670
  doubaoDefaultModelInfo,
@@ -7246,6 +7678,8 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7246
7678
  extensionBridgeEventSchema,
7247
7679
  extensionInstanceSchema,
7248
7680
  extractApiProviderErrorProperties,
7681
+ extractConsecutiveMistakeErrorProperties,
7682
+ extractMessageFromJsonPayload,
7249
7683
  fauxProviders,
7250
7684
  featherlessDefaultModelId,
7251
7685
  featherlessModels,
@@ -7255,7 +7689,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7255
7689
  geminiDefaultModelId,
7256
7690
  geminiModels,
7257
7691
  getApiProtocol,
7258
- getClaudeCodeModelId,
7259
7692
  getEffectiveProtocol,
7260
7693
  getErrorMessage,
7261
7694
  getErrorStatusCode,
@@ -7280,6 +7713,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7280
7713
  ioIntelligenceModels,
7281
7714
  ipcMessageSchema,
7282
7715
  isApiProviderError,
7716
+ isConsecutiveMistakeError,
7283
7717
  isContextManagementEvent,
7284
7718
  isCustomProvider,
7285
7719
  isDynamicProvider,
@@ -7327,6 +7761,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7327
7761
  moonshotDefaultModelId,
7328
7762
  moonshotModels,
7329
7763
  nonBlockingAsks,
7764
+ normalizeClaudeCodeModelId,
7330
7765
  ollamaDefaultModelId,
7331
7766
  ollamaDefaultModelInfo,
7332
7767
  openAiModelInfoSaneDefaults,
@@ -7339,6 +7774,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7339
7774
  organizationDefaultSettingsSchema,
7340
7775
  organizationFeaturesSchema,
7341
7776
  organizationSettingsSchema,
7777
+ parametersSchema,
7342
7778
  promptComponentSchema,
7343
7779
  providerNames,
7344
7780
  providerNamesSchema,