@roo-code/types 1.92.0 → 1.93.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -32,7 +32,6 @@ __export(index_exports, {
32
32
  BEDROCK_REGIONS: () => BEDROCK_REGIONS,
33
33
  BEDROCK_SERVICE_TIER_MODEL_IDS: () => BEDROCK_SERVICE_TIER_MODEL_IDS,
34
34
  BEDROCK_SERVICE_TIER_PRICING: () => BEDROCK_SERVICE_TIER_PRICING,
35
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS: () => CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
36
35
  CODEBASE_INDEX_DEFAULTS: () => CODEBASE_INDEX_DEFAULTS,
37
36
  CONSENT_COOKIE_NAME: () => CONSENT_COOKIE_NAME,
38
37
  CONTEXT_MANAGEMENT_EVENTS: () => CONTEXT_MANAGEMENT_EVENTS,
@@ -122,6 +121,7 @@ __export(index_exports, {
122
121
  chutesModels: () => chutesModels,
123
122
  claudeCodeDefaultModelId: () => claudeCodeDefaultModelId,
124
123
  claudeCodeModels: () => claudeCodeModels,
124
+ claudeCodeReasoningConfig: () => claudeCodeReasoningConfig,
125
125
  clineAskSchema: () => clineAskSchema,
126
126
  clineAsks: () => clineAsks,
127
127
  clineMessageSchema: () => clineMessageSchema,
@@ -136,7 +136,6 @@ __export(index_exports, {
136
136
  commandIds: () => commandIds,
137
137
  contextCondenseSchema: () => contextCondenseSchema,
138
138
  contextTruncationSchema: () => contextTruncationSchema,
139
- convertModelNameForVertex: () => convertModelNameForVertex,
140
139
  customModePromptsSchema: () => customModePromptsSchema,
141
140
  customModesSettingsSchema: () => customModesSettingsSchema,
142
141
  customProviders: () => customProviders,
@@ -158,6 +157,7 @@ __export(index_exports, {
158
157
  extensionBridgeEventSchema: () => extensionBridgeEventSchema,
159
158
  extensionInstanceSchema: () => extensionInstanceSchema,
160
159
  extractApiProviderErrorProperties: () => extractApiProviderErrorProperties,
160
+ extractMessageFromJsonPayload: () => extractMessageFromJsonPayload,
161
161
  fauxProviders: () => fauxProviders,
162
162
  featherlessDefaultModelId: () => featherlessDefaultModelId,
163
163
  featherlessModels: () => featherlessModels,
@@ -167,7 +167,6 @@ __export(index_exports, {
167
167
  geminiDefaultModelId: () => geminiDefaultModelId,
168
168
  geminiModels: () => geminiModels,
169
169
  getApiProtocol: () => getApiProtocol,
170
- getClaudeCodeModelId: () => getClaudeCodeModelId,
171
170
  getEffectiveProtocol: () => getEffectiveProtocol,
172
171
  getErrorMessage: () => getErrorMessage,
173
172
  getErrorStatusCode: () => getErrorStatusCode,
@@ -239,6 +238,7 @@ __export(index_exports, {
239
238
  moonshotDefaultModelId: () => moonshotDefaultModelId,
240
239
  moonshotModels: () => moonshotModels,
241
240
  nonBlockingAsks: () => nonBlockingAsks,
241
+ normalizeClaudeCodeModelId: () => normalizeClaudeCodeModelId,
242
242
  ollamaDefaultModelId: () => ollamaDefaultModelId,
243
243
  ollamaDefaultModelInfo: () => ollamaDefaultModelInfo,
244
244
  openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
@@ -898,6 +898,7 @@ var anthropicModels = {
898
898
  supportsImages: true,
899
899
  supportsPromptCache: true,
900
900
  supportsNativeTools: true,
901
+ defaultToolProtocol: "native",
901
902
  inputPrice: 3,
902
903
  // $3 per million input tokens (≤200K context)
903
904
  outputPrice: 15,
@@ -931,6 +932,7 @@ var anthropicModels = {
931
932
  supportsImages: true,
932
933
  supportsPromptCache: true,
933
934
  supportsNativeTools: true,
935
+ defaultToolProtocol: "native",
934
936
  inputPrice: 3,
935
937
  // $3 per million input tokens (≤200K context)
936
938
  outputPrice: 15,
@@ -963,6 +965,7 @@ var anthropicModels = {
963
965
  supportsImages: true,
964
966
  supportsPromptCache: true,
965
967
  supportsNativeTools: true,
968
+ defaultToolProtocol: "native",
966
969
  inputPrice: 5,
967
970
  // $5 per million input tokens
968
971
  outputPrice: 25,
@@ -980,6 +983,7 @@ var anthropicModels = {
980
983
  supportsImages: true,
981
984
  supportsPromptCache: true,
982
985
  supportsNativeTools: true,
986
+ defaultToolProtocol: "native",
983
987
  inputPrice: 15,
984
988
  // $15 per million input tokens
985
989
  outputPrice: 75,
@@ -997,6 +1001,7 @@ var anthropicModels = {
997
1001
  supportsImages: true,
998
1002
  supportsPromptCache: true,
999
1003
  supportsNativeTools: true,
1004
+ defaultToolProtocol: "native",
1000
1005
  inputPrice: 15,
1001
1006
  // $15 per million input tokens
1002
1007
  outputPrice: 75,
@@ -1014,6 +1019,7 @@ var anthropicModels = {
1014
1019
  supportsImages: true,
1015
1020
  supportsPromptCache: true,
1016
1021
  supportsNativeTools: true,
1022
+ defaultToolProtocol: "native",
1017
1023
  inputPrice: 3,
1018
1024
  // $3 per million input tokens
1019
1025
  outputPrice: 15,
@@ -1032,6 +1038,7 @@ var anthropicModels = {
1032
1038
  supportsImages: true,
1033
1039
  supportsPromptCache: true,
1034
1040
  supportsNativeTools: true,
1041
+ defaultToolProtocol: "native",
1035
1042
  inputPrice: 3,
1036
1043
  // $3 per million input tokens
1037
1044
  outputPrice: 15,
@@ -1047,6 +1054,7 @@ var anthropicModels = {
1047
1054
  supportsImages: true,
1048
1055
  supportsPromptCache: true,
1049
1056
  supportsNativeTools: true,
1057
+ defaultToolProtocol: "native",
1050
1058
  inputPrice: 3,
1051
1059
  // $3 per million input tokens
1052
1060
  outputPrice: 15,
@@ -1062,6 +1070,7 @@ var anthropicModels = {
1062
1070
  supportsImages: false,
1063
1071
  supportsPromptCache: true,
1064
1072
  supportsNativeTools: true,
1073
+ defaultToolProtocol: "native",
1065
1074
  inputPrice: 1,
1066
1075
  outputPrice: 5,
1067
1076
  cacheWritesPrice: 1.25,
@@ -1073,6 +1082,7 @@ var anthropicModels = {
1073
1082
  supportsImages: true,
1074
1083
  supportsPromptCache: true,
1075
1084
  supportsNativeTools: true,
1085
+ defaultToolProtocol: "native",
1076
1086
  inputPrice: 15,
1077
1087
  outputPrice: 75,
1078
1088
  cacheWritesPrice: 18.75,
@@ -1084,6 +1094,7 @@ var anthropicModels = {
1084
1094
  supportsImages: true,
1085
1095
  supportsPromptCache: true,
1086
1096
  supportsNativeTools: true,
1097
+ defaultToolProtocol: "native",
1087
1098
  inputPrice: 0.25,
1088
1099
  outputPrice: 1.25,
1089
1100
  cacheWritesPrice: 0.3,
@@ -1095,6 +1106,7 @@ var anthropicModels = {
1095
1106
  supportsImages: true,
1096
1107
  supportsPromptCache: true,
1097
1108
  supportsNativeTools: true,
1109
+ defaultToolProtocol: "native",
1098
1110
  inputPrice: 1,
1099
1111
  outputPrice: 5,
1100
1112
  cacheWritesPrice: 1.25,
@@ -1247,6 +1259,7 @@ var bedrockModels = {
1247
1259
  supportsPromptCache: true,
1248
1260
  supportsReasoningBudget: true,
1249
1261
  supportsNativeTools: true,
1262
+ defaultToolProtocol: "native",
1250
1263
  inputPrice: 3,
1251
1264
  outputPrice: 15,
1252
1265
  cacheWritesPrice: 3.75,
@@ -1340,6 +1353,7 @@ var bedrockModels = {
1340
1353
  supportsPromptCache: true,
1341
1354
  supportsReasoningBudget: true,
1342
1355
  supportsNativeTools: true,
1356
+ defaultToolProtocol: "native",
1343
1357
  inputPrice: 3,
1344
1358
  outputPrice: 15,
1345
1359
  cacheWritesPrice: 3.75,
@@ -1355,6 +1369,7 @@ var bedrockModels = {
1355
1369
  supportsPromptCache: true,
1356
1370
  supportsReasoningBudget: true,
1357
1371
  supportsNativeTools: true,
1372
+ defaultToolProtocol: "native",
1358
1373
  inputPrice: 15,
1359
1374
  outputPrice: 75,
1360
1375
  cacheWritesPrice: 18.75,
@@ -1370,6 +1385,7 @@ var bedrockModels = {
1370
1385
  supportsPromptCache: true,
1371
1386
  supportsReasoningBudget: true,
1372
1387
  supportsNativeTools: true,
1388
+ defaultToolProtocol: "native",
1373
1389
  inputPrice: 5,
1374
1390
  outputPrice: 25,
1375
1391
  cacheWritesPrice: 6.25,
@@ -1385,6 +1401,7 @@ var bedrockModels = {
1385
1401
  supportsPromptCache: true,
1386
1402
  supportsReasoningBudget: true,
1387
1403
  supportsNativeTools: true,
1404
+ defaultToolProtocol: "native",
1388
1405
  inputPrice: 15,
1389
1406
  outputPrice: 75,
1390
1407
  cacheWritesPrice: 18.75,
@@ -1400,6 +1417,7 @@ var bedrockModels = {
1400
1417
  supportsPromptCache: true,
1401
1418
  supportsReasoningBudget: true,
1402
1419
  supportsNativeTools: true,
1420
+ defaultToolProtocol: "native",
1403
1421
  inputPrice: 3,
1404
1422
  outputPrice: 15,
1405
1423
  cacheWritesPrice: 3.75,
@@ -1414,6 +1432,7 @@ var bedrockModels = {
1414
1432
  supportsImages: true,
1415
1433
  supportsPromptCache: true,
1416
1434
  supportsNativeTools: true,
1435
+ defaultToolProtocol: "native",
1417
1436
  inputPrice: 3,
1418
1437
  outputPrice: 15,
1419
1438
  cacheWritesPrice: 3.75,
@@ -1428,6 +1447,7 @@ var bedrockModels = {
1428
1447
  supportsImages: false,
1429
1448
  supportsPromptCache: true,
1430
1449
  supportsNativeTools: true,
1450
+ defaultToolProtocol: "native",
1431
1451
  inputPrice: 0.8,
1432
1452
  outputPrice: 4,
1433
1453
  cacheWritesPrice: 1,
@@ -1443,6 +1463,7 @@ var bedrockModels = {
1443
1463
  supportsPromptCache: true,
1444
1464
  supportsReasoningBudget: true,
1445
1465
  supportsNativeTools: true,
1466
+ defaultToolProtocol: "native",
1446
1467
  inputPrice: 1,
1447
1468
  outputPrice: 5,
1448
1469
  cacheWritesPrice: 1.25,
@@ -1459,6 +1480,7 @@ var bedrockModels = {
1459
1480
  supportsImages: true,
1460
1481
  supportsPromptCache: false,
1461
1482
  supportsNativeTools: true,
1483
+ defaultToolProtocol: "native",
1462
1484
  inputPrice: 3,
1463
1485
  outputPrice: 15
1464
1486
  },
@@ -1468,6 +1490,7 @@ var bedrockModels = {
1468
1490
  supportsImages: true,
1469
1491
  supportsPromptCache: false,
1470
1492
  supportsNativeTools: true,
1493
+ defaultToolProtocol: "native",
1471
1494
  inputPrice: 15,
1472
1495
  outputPrice: 75
1473
1496
  },
@@ -1477,6 +1500,7 @@ var bedrockModels = {
1477
1500
  supportsImages: true,
1478
1501
  supportsPromptCache: false,
1479
1502
  supportsNativeTools: true,
1503
+ defaultToolProtocol: "native",
1480
1504
  inputPrice: 3,
1481
1505
  outputPrice: 15
1482
1506
  },
@@ -1486,6 +1510,7 @@ var bedrockModels = {
1486
1510
  supportsImages: true,
1487
1511
  supportsPromptCache: false,
1488
1512
  supportsNativeTools: true,
1513
+ defaultToolProtocol: "native",
1489
1514
  inputPrice: 0.25,
1490
1515
  outputPrice: 1.25
1491
1516
  },
@@ -1495,6 +1520,7 @@ var bedrockModels = {
1495
1520
  supportsImages: false,
1496
1521
  supportsPromptCache: false,
1497
1522
  supportsNativeTools: true,
1523
+ defaultToolProtocol: "native",
1498
1524
  inputPrice: 8,
1499
1525
  outputPrice: 24,
1500
1526
  description: "Claude 2.1"
@@ -1505,6 +1531,7 @@ var bedrockModels = {
1505
1531
  supportsImages: false,
1506
1532
  supportsPromptCache: false,
1507
1533
  supportsNativeTools: true,
1534
+ defaultToolProtocol: "native",
1508
1535
  inputPrice: 8,
1509
1536
  outputPrice: 24,
1510
1537
  description: "Claude 2.0"
@@ -1515,6 +1542,7 @@ var bedrockModels = {
1515
1542
  supportsImages: false,
1516
1543
  supportsPromptCache: false,
1517
1544
  supportsNativeTools: true,
1545
+ defaultToolProtocol: "native",
1518
1546
  inputPrice: 0.8,
1519
1547
  outputPrice: 2.4,
1520
1548
  description: "Claude Instant"
@@ -1831,6 +1859,7 @@ var cerebrasModels = {
1831
1859
  supportsImages: false,
1832
1860
  supportsPromptCache: false,
1833
1861
  supportsNativeTools: true,
1862
+ defaultToolProtocol: "native",
1834
1863
  inputPrice: 0,
1835
1864
  outputPrice: 0,
1836
1865
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
@@ -1842,6 +1871,7 @@ var cerebrasModels = {
1842
1871
  supportsImages: false,
1843
1872
  supportsPromptCache: false,
1844
1873
  supportsNativeTools: true,
1874
+ defaultToolProtocol: "native",
1845
1875
  inputPrice: 0,
1846
1876
  outputPrice: 0,
1847
1877
  description: "Intelligent model with ~1400 tokens/s"
@@ -1853,6 +1883,7 @@ var cerebrasModels = {
1853
1883
  supportsImages: false,
1854
1884
  supportsPromptCache: false,
1855
1885
  supportsNativeTools: true,
1886
+ defaultToolProtocol: "native",
1856
1887
  inputPrice: 0,
1857
1888
  outputPrice: 0,
1858
1889
  description: "Powerful model with ~2600 tokens/s"
@@ -1864,6 +1895,7 @@ var cerebrasModels = {
1864
1895
  supportsImages: false,
1865
1896
  supportsPromptCache: false,
1866
1897
  supportsNativeTools: true,
1898
+ defaultToolProtocol: "native",
1867
1899
  inputPrice: 0,
1868
1900
  outputPrice: 0,
1869
1901
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1875,6 +1907,7 @@ var cerebrasModels = {
1875
1907
  supportsImages: false,
1876
1908
  supportsPromptCache: false,
1877
1909
  supportsNativeTools: true,
1910
+ defaultToolProtocol: "native",
1878
1911
  inputPrice: 0,
1879
1912
  outputPrice: 0,
1880
1913
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1889,6 +1922,8 @@ var chutesModels = {
1889
1922
  contextWindow: 163840,
1890
1923
  supportsImages: false,
1891
1924
  supportsPromptCache: false,
1925
+ supportsNativeTools: true,
1926
+ defaultToolProtocol: "native",
1892
1927
  inputPrice: 0,
1893
1928
  outputPrice: 0,
1894
1929
  description: "DeepSeek R1 0528 model."
@@ -1898,6 +1933,8 @@ var chutesModels = {
1898
1933
  contextWindow: 163840,
1899
1934
  supportsImages: false,
1900
1935
  supportsPromptCache: false,
1936
+ supportsNativeTools: true,
1937
+ defaultToolProtocol: "native",
1901
1938
  inputPrice: 0,
1902
1939
  outputPrice: 0,
1903
1940
  description: "DeepSeek R1 model."
@@ -1907,6 +1944,8 @@ var chutesModels = {
1907
1944
  contextWindow: 163840,
1908
1945
  supportsImages: false,
1909
1946
  supportsPromptCache: false,
1947
+ supportsNativeTools: true,
1948
+ defaultToolProtocol: "native",
1910
1949
  inputPrice: 0,
1911
1950
  outputPrice: 0,
1912
1951
  description: "DeepSeek V3 model."
@@ -1916,6 +1955,8 @@ var chutesModels = {
1916
1955
  contextWindow: 163840,
1917
1956
  supportsImages: false,
1918
1957
  supportsPromptCache: false,
1958
+ supportsNativeTools: true,
1959
+ defaultToolProtocol: "native",
1919
1960
  inputPrice: 0,
1920
1961
  outputPrice: 0,
1921
1962
  description: "DeepSeek V3.1 model."
@@ -1925,6 +1966,8 @@ var chutesModels = {
1925
1966
  contextWindow: 163840,
1926
1967
  supportsImages: false,
1927
1968
  supportsPromptCache: false,
1969
+ supportsNativeTools: true,
1970
+ defaultToolProtocol: "native",
1928
1971
  inputPrice: 0.23,
1929
1972
  outputPrice: 0.9,
1930
1973
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1934,6 +1977,8 @@ var chutesModels = {
1934
1977
  contextWindow: 163840,
1935
1978
  supportsImages: false,
1936
1979
  supportsPromptCache: false,
1980
+ supportsNativeTools: true,
1981
+ defaultToolProtocol: "native",
1937
1982
  inputPrice: 1,
1938
1983
  outputPrice: 3,
1939
1984
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1943,6 +1988,8 @@ var chutesModels = {
1943
1988
  contextWindow: 163840,
1944
1989
  supportsImages: false,
1945
1990
  supportsPromptCache: false,
1991
+ supportsNativeTools: true,
1992
+ defaultToolProtocol: "native",
1946
1993
  inputPrice: 0.25,
1947
1994
  outputPrice: 0.35,
1948
1995
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1954,6 +2001,8 @@ var chutesModels = {
1954
2001
  // From Groq
1955
2002
  supportsImages: false,
1956
2003
  supportsPromptCache: false,
2004
+ supportsNativeTools: true,
2005
+ defaultToolProtocol: "native",
1957
2006
  inputPrice: 0,
1958
2007
  outputPrice: 0,
1959
2008
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1963,6 +2012,8 @@ var chutesModels = {
1963
2012
  contextWindow: 512e3,
1964
2013
  supportsImages: false,
1965
2014
  supportsPromptCache: false,
2015
+ supportsNativeTools: true,
2016
+ defaultToolProtocol: "native",
1966
2017
  inputPrice: 0,
1967
2018
  outputPrice: 0,
1968
2019
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1972,6 +2023,8 @@ var chutesModels = {
1972
2023
  contextWindow: 128e3,
1973
2024
  supportsImages: false,
1974
2025
  supportsPromptCache: false,
2026
+ supportsNativeTools: true,
2027
+ defaultToolProtocol: "native",
1975
2028
  inputPrice: 0,
1976
2029
  outputPrice: 0,
1977
2030
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1981,6 +2034,8 @@ var chutesModels = {
1981
2034
  contextWindow: 131072,
1982
2035
  supportsImages: false,
1983
2036
  supportsPromptCache: false,
2037
+ supportsNativeTools: true,
2038
+ defaultToolProtocol: "native",
1984
2039
  inputPrice: 0,
1985
2040
  outputPrice: 0,
1986
2041
  description: "Unsloth Gemma 3 12B IT model."
@@ -1990,6 +2045,8 @@ var chutesModels = {
1990
2045
  contextWindow: 131072,
1991
2046
  supportsImages: false,
1992
2047
  supportsPromptCache: false,
2048
+ supportsNativeTools: true,
2049
+ defaultToolProtocol: "native",
1993
2050
  inputPrice: 0,
1994
2051
  outputPrice: 0,
1995
2052
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1999,6 +2056,8 @@ var chutesModels = {
1999
2056
  contextWindow: 131072,
2000
2057
  supportsImages: false,
2001
2058
  supportsPromptCache: false,
2059
+ supportsNativeTools: true,
2060
+ defaultToolProtocol: "native",
2002
2061
  inputPrice: 0,
2003
2062
  outputPrice: 0,
2004
2063
  description: "Unsloth Gemma 3 4B IT model."
@@ -2008,6 +2067,8 @@ var chutesModels = {
2008
2067
  contextWindow: 131072,
2009
2068
  supportsImages: false,
2010
2069
  supportsPromptCache: false,
2070
+ supportsNativeTools: true,
2071
+ defaultToolProtocol: "native",
2011
2072
  inputPrice: 0,
2012
2073
  outputPrice: 0,
2013
2074
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -2017,6 +2078,8 @@ var chutesModels = {
2017
2078
  contextWindow: 131072,
2018
2079
  supportsImages: false,
2019
2080
  supportsPromptCache: false,
2081
+ supportsNativeTools: true,
2082
+ defaultToolProtocol: "native",
2020
2083
  inputPrice: 0,
2021
2084
  outputPrice: 0,
2022
2085
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -2026,6 +2089,8 @@ var chutesModels = {
2026
2089
  contextWindow: 256e3,
2027
2090
  supportsImages: false,
2028
2091
  supportsPromptCache: false,
2092
+ supportsNativeTools: true,
2093
+ defaultToolProtocol: "native",
2029
2094
  inputPrice: 0,
2030
2095
  outputPrice: 0,
2031
2096
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -2035,6 +2100,8 @@ var chutesModels = {
2035
2100
  contextWindow: 163840,
2036
2101
  supportsImages: false,
2037
2102
  supportsPromptCache: false,
2103
+ supportsNativeTools: true,
2104
+ defaultToolProtocol: "native",
2038
2105
  inputPrice: 0,
2039
2106
  outputPrice: 0,
2040
2107
  description: "DeepSeek V3 Base model."
@@ -2044,6 +2111,8 @@ var chutesModels = {
2044
2111
  contextWindow: 163840,
2045
2112
  supportsImages: false,
2046
2113
  supportsPromptCache: false,
2114
+ supportsNativeTools: true,
2115
+ defaultToolProtocol: "native",
2047
2116
  inputPrice: 0,
2048
2117
  outputPrice: 0,
2049
2118
  description: "DeepSeek R1 Zero model."
@@ -2053,6 +2122,8 @@ var chutesModels = {
2053
2122
  contextWindow: 163840,
2054
2123
  supportsImages: false,
2055
2124
  supportsPromptCache: false,
2125
+ supportsNativeTools: true,
2126
+ defaultToolProtocol: "native",
2056
2127
  inputPrice: 0,
2057
2128
  outputPrice: 0,
2058
2129
  description: "DeepSeek V3 (0324) model."
@@ -2062,6 +2133,8 @@ var chutesModels = {
2062
2133
  contextWindow: 262144,
2063
2134
  supportsImages: false,
2064
2135
  supportsPromptCache: false,
2136
+ supportsNativeTools: true,
2137
+ defaultToolProtocol: "native",
2065
2138
  inputPrice: 0,
2066
2139
  outputPrice: 0,
2067
2140
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -2071,6 +2144,8 @@ var chutesModels = {
2071
2144
  contextWindow: 40960,
2072
2145
  supportsImages: false,
2073
2146
  supportsPromptCache: false,
2147
+ supportsNativeTools: true,
2148
+ defaultToolProtocol: "native",
2074
2149
  inputPrice: 0,
2075
2150
  outputPrice: 0,
2076
2151
  description: "Qwen3 235B A22B model."
@@ -2080,6 +2155,8 @@ var chutesModels = {
2080
2155
  contextWindow: 40960,
2081
2156
  supportsImages: false,
2082
2157
  supportsPromptCache: false,
2158
+ supportsNativeTools: true,
2159
+ defaultToolProtocol: "native",
2083
2160
  inputPrice: 0,
2084
2161
  outputPrice: 0,
2085
2162
  description: "Qwen3 32B model."
@@ -2089,6 +2166,8 @@ var chutesModels = {
2089
2166
  contextWindow: 40960,
2090
2167
  supportsImages: false,
2091
2168
  supportsPromptCache: false,
2169
+ supportsNativeTools: true,
2170
+ defaultToolProtocol: "native",
2092
2171
  inputPrice: 0,
2093
2172
  outputPrice: 0,
2094
2173
  description: "Qwen3 30B A3B model."
@@ -2098,6 +2177,8 @@ var chutesModels = {
2098
2177
  contextWindow: 40960,
2099
2178
  supportsImages: false,
2100
2179
  supportsPromptCache: false,
2180
+ supportsNativeTools: true,
2181
+ defaultToolProtocol: "native",
2101
2182
  inputPrice: 0,
2102
2183
  outputPrice: 0,
2103
2184
  description: "Qwen3 14B model."
@@ -2107,6 +2188,8 @@ var chutesModels = {
2107
2188
  contextWindow: 40960,
2108
2189
  supportsImages: false,
2109
2190
  supportsPromptCache: false,
2191
+ supportsNativeTools: true,
2192
+ defaultToolProtocol: "native",
2110
2193
  inputPrice: 0,
2111
2194
  outputPrice: 0,
2112
2195
  description: "Qwen3 8B model."
@@ -2116,6 +2199,8 @@ var chutesModels = {
2116
2199
  contextWindow: 163840,
2117
2200
  supportsImages: false,
2118
2201
  supportsPromptCache: false,
2202
+ supportsNativeTools: true,
2203
+ defaultToolProtocol: "native",
2119
2204
  inputPrice: 0,
2120
2205
  outputPrice: 0,
2121
2206
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -2125,6 +2210,8 @@ var chutesModels = {
2125
2210
  contextWindow: 163840,
2126
2211
  supportsImages: false,
2127
2212
  supportsPromptCache: false,
2213
+ supportsNativeTools: true,
2214
+ defaultToolProtocol: "native",
2128
2215
  inputPrice: 0,
2129
2216
  outputPrice: 0,
2130
2217
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -2134,6 +2221,8 @@ var chutesModels = {
2134
2221
  contextWindow: 151329,
2135
2222
  supportsImages: false,
2136
2223
  supportsPromptCache: false,
2224
+ supportsNativeTools: true,
2225
+ defaultToolProtocol: "native",
2137
2226
  inputPrice: 0,
2138
2227
  outputPrice: 0,
2139
2228
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -2143,6 +2232,8 @@ var chutesModels = {
2143
2232
  contextWindow: 131072,
2144
2233
  supportsImages: false,
2145
2234
  supportsPromptCache: false,
2235
+ supportsNativeTools: true,
2236
+ defaultToolProtocol: "native",
2146
2237
  inputPrice: 0,
2147
2238
  outputPrice: 0,
2148
2239
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -2152,6 +2243,8 @@ var chutesModels = {
2152
2243
  contextWindow: 131072,
2153
2244
  supportsImages: false,
2154
2245
  supportsPromptCache: false,
2246
+ supportsNativeTools: true,
2247
+ defaultToolProtocol: "native",
2155
2248
  inputPrice: 1,
2156
2249
  outputPrice: 3,
2157
2250
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -2161,6 +2254,8 @@ var chutesModels = {
2161
2254
  contextWindow: 202752,
2162
2255
  supportsImages: false,
2163
2256
  supportsPromptCache: false,
2257
+ supportsNativeTools: true,
2258
+ defaultToolProtocol: "native",
2164
2259
  inputPrice: 0,
2165
2260
  outputPrice: 0,
2166
2261
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -2171,6 +2266,8 @@ var chutesModels = {
2171
2266
  contextWindow: 202752,
2172
2267
  supportsImages: false,
2173
2268
  supportsPromptCache: false,
2269
+ supportsNativeTools: true,
2270
+ defaultToolProtocol: "native",
2174
2271
  inputPrice: 1.15,
2175
2272
  outputPrice: 3.25,
2176
2273
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -2180,6 +2277,8 @@ var chutesModels = {
2180
2277
  contextWindow: 128e3,
2181
2278
  supportsImages: false,
2182
2279
  supportsPromptCache: false,
2280
+ supportsNativeTools: true,
2281
+ defaultToolProtocol: "native",
2183
2282
  inputPrice: 0,
2184
2283
  outputPrice: 0,
2185
2284
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -2189,6 +2288,8 @@ var chutesModels = {
2189
2288
  contextWindow: 262144,
2190
2289
  supportsImages: false,
2191
2290
  supportsPromptCache: false,
2291
+ supportsNativeTools: true,
2292
+ defaultToolProtocol: "native",
2192
2293
  inputPrice: 0,
2193
2294
  outputPrice: 0,
2194
2295
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -2198,6 +2299,8 @@ var chutesModels = {
2198
2299
  contextWindow: 75e3,
2199
2300
  supportsImages: false,
2200
2301
  supportsPromptCache: false,
2302
+ supportsNativeTools: true,
2303
+ defaultToolProtocol: "native",
2201
2304
  inputPrice: 0.1481,
2202
2305
  outputPrice: 0.5926,
2203
2306
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -2207,6 +2310,8 @@ var chutesModels = {
2207
2310
  contextWindow: 262144,
2208
2311
  supportsImages: false,
2209
2312
  supportsPromptCache: false,
2313
+ supportsNativeTools: true,
2314
+ defaultToolProtocol: "native",
2210
2315
  inputPrice: 0.1999,
2211
2316
  outputPrice: 0.8001,
2212
2317
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -2216,6 +2321,8 @@ var chutesModels = {
2216
2321
  contextWindow: 262144,
2217
2322
  supportsImages: false,
2218
2323
  supportsPromptCache: false,
2324
+ supportsNativeTools: true,
2325
+ defaultToolProtocol: "native",
2219
2326
  inputPrice: 0.077968332,
2220
2327
  outputPrice: 0.31202496,
2221
2328
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -2225,6 +2332,8 @@ var chutesModels = {
2225
2332
  contextWindow: 131072,
2226
2333
  supportsImages: false,
2227
2334
  supportsPromptCache: false,
2335
+ supportsNativeTools: true,
2336
+ defaultToolProtocol: "native",
2228
2337
  inputPrice: 0,
2229
2338
  outputPrice: 0,
2230
2339
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -2234,6 +2343,8 @@ var chutesModels = {
2234
2343
  contextWindow: 131072,
2235
2344
  supportsImages: false,
2236
2345
  supportsPromptCache: false,
2346
+ supportsNativeTools: true,
2347
+ defaultToolProtocol: "native",
2237
2348
  inputPrice: 0,
2238
2349
  outputPrice: 0,
2239
2350
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2243,6 +2354,8 @@ var chutesModels = {
2243
2354
  contextWindow: 262144,
2244
2355
  supportsImages: true,
2245
2356
  supportsPromptCache: false,
2357
+ supportsNativeTools: true,
2358
+ defaultToolProtocol: "native",
2246
2359
  inputPrice: 0.16,
2247
2360
  outputPrice: 0.65,
2248
2361
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2251,139 +2364,71 @@ var chutesModels = {
2251
2364
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2252
2365
 
2253
2366
  // src/providers/claude-code.ts
2254
- var VERTEX_DATE_PATTERN = /-(\d{8})$/;
2255
- function convertModelNameForVertex(modelName) {
2256
- return modelName.replace(VERTEX_DATE_PATTERN, "@$1");
2257
- }
2258
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2259
- var CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS = 16e3;
2260
- function getClaudeCodeModelId(baseModelId, useVertex = false) {
2261
- return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
2262
- }
2367
+ var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2263
2368
  var claudeCodeModels = {
2264
- "claude-sonnet-4-5": {
2265
- ...anthropicModels["claude-sonnet-4-5"],
2266
- supportsImages: false,
2267
- supportsPromptCache: true,
2268
- // Claude Code does report cache tokens
2269
- supportsReasoningEffort: false,
2270
- supportsReasoningBudget: false,
2271
- requiredReasoningBudget: false,
2272
- // Claude Code manages its own tools and temperature via the CLI
2273
- supportsNativeTools: false,
2274
- supportsTemperature: false
2275
- },
2276
- "claude-sonnet-4-5-20250929[1m]": {
2277
- ...anthropicModels["claude-sonnet-4-5"],
2278
- contextWindow: 1e6,
2279
- // 1M token context window (requires [1m] suffix)
2280
- supportsImages: false,
2281
- supportsPromptCache: true,
2282
- // Claude Code does report cache tokens
2283
- supportsReasoningEffort: false,
2284
- supportsReasoningBudget: false,
2285
- requiredReasoningBudget: false,
2286
- // Claude Code manages its own tools and temperature via the CLI
2287
- supportsNativeTools: false,
2288
- supportsTemperature: false
2289
- },
2290
- "claude-sonnet-4-20250514": {
2291
- ...anthropicModels["claude-sonnet-4-20250514"],
2292
- supportsImages: false,
2293
- supportsPromptCache: true,
2294
- // Claude Code does report cache tokens
2295
- supportsReasoningEffort: false,
2296
- supportsReasoningBudget: false,
2297
- requiredReasoningBudget: false,
2298
- // Claude Code manages its own tools and temperature via the CLI
2299
- supportsNativeTools: false,
2300
- supportsTemperature: false
2301
- },
2302
- "claude-opus-4-5-20251101": {
2303
- ...anthropicModels["claude-opus-4-5-20251101"],
2304
- supportsImages: false,
2305
- supportsPromptCache: true,
2306
- // Claude Code does report cache tokens
2307
- supportsReasoningEffort: false,
2308
- supportsReasoningBudget: false,
2309
- requiredReasoningBudget: false,
2310
- // Claude Code manages its own tools and temperature via the CLI
2311
- supportsNativeTools: false,
2312
- supportsTemperature: false
2313
- },
2314
- "claude-opus-4-1-20250805": {
2315
- ...anthropicModels["claude-opus-4-1-20250805"],
2316
- supportsImages: false,
2317
- supportsPromptCache: true,
2318
- // Claude Code does report cache tokens
2319
- supportsReasoningEffort: false,
2320
- supportsReasoningBudget: false,
2321
- requiredReasoningBudget: false,
2322
- // Claude Code manages its own tools and temperature via the CLI
2323
- supportsNativeTools: false,
2324
- supportsTemperature: false
2325
- },
2326
- "claude-opus-4-20250514": {
2327
- ...anthropicModels["claude-opus-4-20250514"],
2328
- supportsImages: false,
2329
- supportsPromptCache: true,
2330
- // Claude Code does report cache tokens
2331
- supportsReasoningEffort: false,
2332
- supportsReasoningBudget: false,
2333
- requiredReasoningBudget: false,
2334
- // Claude Code manages its own tools and temperature via the CLI
2335
- supportsNativeTools: false,
2336
- supportsTemperature: false
2337
- },
2338
- "claude-3-7-sonnet-20250219": {
2339
- ...anthropicModels["claude-3-7-sonnet-20250219"],
2340
- supportsImages: false,
2341
- supportsPromptCache: true,
2342
- // Claude Code does report cache tokens
2343
- supportsReasoningEffort: false,
2344
- supportsReasoningBudget: false,
2345
- requiredReasoningBudget: false,
2346
- // Claude Code manages its own tools and temperature via the CLI
2347
- supportsNativeTools: false,
2348
- supportsTemperature: false
2349
- },
2350
- "claude-3-5-sonnet-20241022": {
2351
- ...anthropicModels["claude-3-5-sonnet-20241022"],
2352
- supportsImages: false,
2369
+ "claude-haiku-4-5": {
2370
+ maxTokens: 32768,
2371
+ contextWindow: 2e5,
2372
+ supportsImages: true,
2353
2373
  supportsPromptCache: true,
2354
- // Claude Code does report cache tokens
2355
- supportsReasoningEffort: false,
2356
- supportsReasoningBudget: false,
2357
- requiredReasoningBudget: false,
2358
- // Claude Code manages its own tools and temperature via the CLI
2359
- supportsNativeTools: false,
2360
- supportsTemperature: false
2374
+ supportsNativeTools: true,
2375
+ defaultToolProtocol: "native",
2376
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2377
+ reasoningEffort: "medium",
2378
+ description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2361
2379
  },
2362
- "claude-3-5-haiku-20241022": {
2363
- ...anthropicModels["claude-3-5-haiku-20241022"],
2364
- supportsImages: false,
2380
+ "claude-sonnet-4-5": {
2381
+ maxTokens: 32768,
2382
+ contextWindow: 2e5,
2383
+ supportsImages: true,
2365
2384
  supportsPromptCache: true,
2366
- // Claude Code does report cache tokens
2367
- supportsReasoningEffort: false,
2368
- supportsReasoningBudget: false,
2369
- requiredReasoningBudget: false,
2370
- // Claude Code manages its own tools and temperature via the CLI
2371
- supportsNativeTools: false,
2372
- supportsTemperature: false
2385
+ supportsNativeTools: true,
2386
+ defaultToolProtocol: "native",
2387
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2388
+ reasoningEffort: "medium",
2389
+ description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2373
2390
  },
2374
- "claude-haiku-4-5-20251001": {
2375
- ...anthropicModels["claude-haiku-4-5-20251001"],
2376
- supportsImages: false,
2391
+ "claude-opus-4-5": {
2392
+ maxTokens: 32768,
2393
+ contextWindow: 2e5,
2394
+ supportsImages: true,
2377
2395
  supportsPromptCache: true,
2378
- // Claude Code does report cache tokens
2379
- supportsReasoningEffort: false,
2380
- supportsReasoningBudget: false,
2381
- requiredReasoningBudget: false,
2382
- // Claude Code manages its own tools and temperature via the CLI
2383
- supportsNativeTools: false,
2384
- supportsTemperature: false
2396
+ supportsNativeTools: true,
2397
+ defaultToolProtocol: "native",
2398
+ supportsReasoningEffort: ["disable", "low", "medium", "high"],
2399
+ reasoningEffort: "medium",
2400
+ description: "Claude Opus 4.5 - Most capable with thinking"
2385
2401
  }
2386
2402
  };
2403
+ var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2404
+ var MODEL_FAMILY_PATTERNS = [
2405
+ // Opus models (any version) → claude-opus-4-5
2406
+ { pattern: /opus/i, target: "claude-opus-4-5" },
2407
+ // Haiku models (any version) → claude-haiku-4-5
2408
+ { pattern: /haiku/i, target: "claude-haiku-4-5" },
2409
+ // Sonnet models (any version) → claude-sonnet-4-5
2410
+ { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2411
+ ];
2412
+ function normalizeClaudeCodeModelId(modelId) {
2413
+ if (Object.hasOwn(claudeCodeModels, modelId)) {
2414
+ return modelId;
2415
+ }
2416
+ const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2417
+ if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2418
+ return withoutDate;
2419
+ }
2420
+ for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2421
+ if (pattern.test(modelId)) {
2422
+ return target;
2423
+ }
2424
+ }
2425
+ return claudeCodeDefaultModelId;
2426
+ }
2427
+ var claudeCodeReasoningConfig = {
2428
+ low: { budgetTokens: 16e3 },
2429
+ medium: { budgetTokens: 32e3 },
2430
+ high: { budgetTokens: 64e3 }
2431
+ };
2387
2432
 
2388
2433
  // src/providers/deepseek.ts
2389
2434
  var deepSeekDefaultModelId = "deepseek-chat";
@@ -2414,6 +2459,7 @@ var deepSeekModels = {
2414
2459
  supportsPromptCache: true,
2415
2460
  supportsNativeTools: true,
2416
2461
  defaultToolProtocol: "native",
2462
+ preserveReasoning: true,
2417
2463
  inputPrice: 0.28,
2418
2464
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2419
2465
  outputPrice: 0.42,
@@ -2425,7 +2471,7 @@ var deepSeekModels = {
2425
2471
  description: `DeepSeek-V3.2 (Thinking Mode) achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 8K output tokens. Supports JSON output, tool calls, and chat prefix completion (beta).`
2426
2472
  }
2427
2473
  };
2428
- var DEEP_SEEK_DEFAULT_TEMPERATURE = 0;
2474
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.3;
2429
2475
 
2430
2476
  // src/providers/doubao.ts
2431
2477
  var doubaoDefaultModelId = "doubao-seed-1-6-250615";
@@ -2436,6 +2482,7 @@ var doubaoModels = {
2436
2482
  supportsImages: true,
2437
2483
  supportsPromptCache: true,
2438
2484
  supportsNativeTools: true,
2485
+ defaultToolProtocol: "native",
2439
2486
  inputPrice: 1e-4,
2440
2487
  // $0.0001 per million tokens (cache miss)
2441
2488
  outputPrice: 4e-4,
@@ -2452,6 +2499,7 @@ var doubaoModels = {
2452
2499
  supportsImages: true,
2453
2500
  supportsPromptCache: true,
2454
2501
  supportsNativeTools: true,
2502
+ defaultToolProtocol: "native",
2455
2503
  inputPrice: 2e-4,
2456
2504
  // $0.0002 per million tokens
2457
2505
  outputPrice: 8e-4,
@@ -2468,6 +2516,7 @@ var doubaoModels = {
2468
2516
  supportsImages: true,
2469
2517
  supportsPromptCache: true,
2470
2518
  supportsNativeTools: true,
2519
+ defaultToolProtocol: "native",
2471
2520
  inputPrice: 15e-5,
2472
2521
  // $0.00015 per million tokens
2473
2522
  outputPrice: 6e-4,
@@ -2544,6 +2593,7 @@ var fireworksModels = {
2544
2593
  supportsImages: false,
2545
2594
  supportsPromptCache: true,
2546
2595
  supportsNativeTools: true,
2596
+ defaultToolProtocol: "native",
2547
2597
  inputPrice: 0.6,
2548
2598
  outputPrice: 2.5,
2549
2599
  cacheReadsPrice: 0.15,
@@ -2555,6 +2605,7 @@ var fireworksModels = {
2555
2605
  supportsImages: false,
2556
2606
  supportsPromptCache: false,
2557
2607
  supportsNativeTools: true,
2608
+ defaultToolProtocol: "native",
2558
2609
  inputPrice: 0.6,
2559
2610
  outputPrice: 2.5,
2560
2611
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2565,6 +2616,7 @@ var fireworksModels = {
2565
2616
  supportsImages: false,
2566
2617
  supportsPromptCache: false,
2567
2618
  supportsNativeTools: true,
2619
+ defaultToolProtocol: "native",
2568
2620
  inputPrice: 0.3,
2569
2621
  outputPrice: 1.2,
2570
2622
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2575,6 +2627,7 @@ var fireworksModels = {
2575
2627
  supportsImages: false,
2576
2628
  supportsPromptCache: false,
2577
2629
  supportsNativeTools: true,
2630
+ defaultToolProtocol: "native",
2578
2631
  inputPrice: 0.22,
2579
2632
  outputPrice: 0.88,
2580
2633
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2585,6 +2638,7 @@ var fireworksModels = {
2585
2638
  supportsImages: false,
2586
2639
  supportsPromptCache: false,
2587
2640
  supportsNativeTools: true,
2641
+ defaultToolProtocol: "native",
2588
2642
  inputPrice: 0.45,
2589
2643
  outputPrice: 1.8,
2590
2644
  description: "Qwen3's most agentic code model to date."
@@ -2595,6 +2649,7 @@ var fireworksModels = {
2595
2649
  supportsImages: false,
2596
2650
  supportsPromptCache: false,
2597
2651
  supportsNativeTools: true,
2652
+ defaultToolProtocol: "native",
2598
2653
  inputPrice: 3,
2599
2654
  outputPrice: 8,
2600
2655
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2605,6 +2660,7 @@ var fireworksModels = {
2605
2660
  supportsImages: false,
2606
2661
  supportsPromptCache: false,
2607
2662
  supportsNativeTools: true,
2663
+ defaultToolProtocol: "native",
2608
2664
  inputPrice: 0.9,
2609
2665
  outputPrice: 0.9,
2610
2666
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2615,6 +2671,7 @@ var fireworksModels = {
2615
2671
  supportsImages: false,
2616
2672
  supportsPromptCache: false,
2617
2673
  supportsNativeTools: true,
2674
+ defaultToolProtocol: "native",
2618
2675
  inputPrice: 0.56,
2619
2676
  outputPrice: 1.68,
2620
2677
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2625,6 +2682,7 @@ var fireworksModels = {
2625
2682
  supportsImages: false,
2626
2683
  supportsPromptCache: false,
2627
2684
  supportsNativeTools: true,
2685
+ defaultToolProtocol: "native",
2628
2686
  inputPrice: 0.55,
2629
2687
  outputPrice: 2.19,
2630
2688
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2635,6 +2693,7 @@ var fireworksModels = {
2635
2693
  supportsImages: false,
2636
2694
  supportsPromptCache: false,
2637
2695
  supportsNativeTools: true,
2696
+ defaultToolProtocol: "native",
2638
2697
  inputPrice: 0.55,
2639
2698
  outputPrice: 2.19,
2640
2699
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2645,6 +2704,7 @@ var fireworksModels = {
2645
2704
  supportsImages: false,
2646
2705
  supportsPromptCache: false,
2647
2706
  supportsNativeTools: true,
2707
+ defaultToolProtocol: "native",
2648
2708
  inputPrice: 0.55,
2649
2709
  outputPrice: 2.19,
2650
2710
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2655,6 +2715,7 @@ var fireworksModels = {
2655
2715
  supportsImages: false,
2656
2716
  supportsPromptCache: false,
2657
2717
  supportsNativeTools: true,
2718
+ defaultToolProtocol: "native",
2658
2719
  inputPrice: 0.07,
2659
2720
  outputPrice: 0.3,
2660
2721
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2665,6 +2726,7 @@ var fireworksModels = {
2665
2726
  supportsImages: false,
2666
2727
  supportsPromptCache: false,
2667
2728
  supportsNativeTools: true,
2729
+ defaultToolProtocol: "native",
2668
2730
  inputPrice: 0.15,
2669
2731
  outputPrice: 0.6,
2670
2732
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2672,16 +2734,18 @@ var fireworksModels = {
2672
2734
  };
2673
2735
 
2674
2736
  // src/providers/gemini.ts
2675
- var geminiDefaultModelId = "gemini-2.5-pro";
2737
+ var geminiDefaultModelId = "gemini-3-pro-preview";
2676
2738
  var geminiModels = {
2677
2739
  "gemini-3-pro-preview": {
2678
2740
  maxTokens: 65536,
2679
2741
  contextWindow: 1048576,
2680
2742
  supportsImages: true,
2681
2743
  supportsNativeTools: true,
2744
+ defaultToolProtocol: "native",
2682
2745
  supportsPromptCache: true,
2683
2746
  supportsReasoningEffort: ["low", "high"],
2684
2747
  reasoningEffort: "low",
2748
+ includedTools: ["write_file", "edit_file"],
2685
2749
  supportsTemperature: true,
2686
2750
  defaultTemperature: 1,
2687
2751
  inputPrice: 4,
@@ -2699,13 +2763,32 @@ var geminiModels = {
2699
2763
  }
2700
2764
  ]
2701
2765
  },
2766
+ "gemini-3-flash-preview": {
2767
+ maxTokens: 65536,
2768
+ contextWindow: 1048576,
2769
+ supportsImages: true,
2770
+ supportsNativeTools: true,
2771
+ defaultToolProtocol: "native",
2772
+ supportsPromptCache: true,
2773
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2774
+ reasoningEffort: "medium",
2775
+ includedTools: ["write_file", "edit_file"],
2776
+ supportsTemperature: true,
2777
+ defaultTemperature: 1,
2778
+ inputPrice: 0.3,
2779
+ outputPrice: 2.5,
2780
+ cacheReadsPrice: 0.075,
2781
+ cacheWritesPrice: 1
2782
+ },
2702
2783
  // 2.5 Pro models
2703
2784
  "gemini-2.5-pro": {
2704
2785
  maxTokens: 64e3,
2705
2786
  contextWindow: 1048576,
2706
2787
  supportsImages: true,
2707
2788
  supportsNativeTools: true,
2789
+ defaultToolProtocol: "native",
2708
2790
  supportsPromptCache: true,
2791
+ includedTools: ["write_file", "edit_file"],
2709
2792
  inputPrice: 2.5,
2710
2793
  // This is the pricing for prompts above 200k tokens.
2711
2794
  outputPrice: 15,
@@ -2734,7 +2817,9 @@ var geminiModels = {
2734
2817
  contextWindow: 1048576,
2735
2818
  supportsImages: true,
2736
2819
  supportsNativeTools: true,
2820
+ defaultToolProtocol: "native",
2737
2821
  supportsPromptCache: true,
2822
+ includedTools: ["write_file", "edit_file"],
2738
2823
  inputPrice: 2.5,
2739
2824
  // This is the pricing for prompts above 200k tokens.
2740
2825
  outputPrice: 15,
@@ -2762,7 +2847,9 @@ var geminiModels = {
2762
2847
  contextWindow: 1048576,
2763
2848
  supportsImages: true,
2764
2849
  supportsNativeTools: true,
2850
+ defaultToolProtocol: "native",
2765
2851
  supportsPromptCache: true,
2852
+ includedTools: ["write_file", "edit_file"],
2766
2853
  inputPrice: 2.5,
2767
2854
  // This is the pricing for prompts above 200k tokens.
2768
2855
  outputPrice: 15,
@@ -2788,7 +2875,9 @@ var geminiModels = {
2788
2875
  contextWindow: 1048576,
2789
2876
  supportsImages: true,
2790
2877
  supportsNativeTools: true,
2878
+ defaultToolProtocol: "native",
2791
2879
  supportsPromptCache: true,
2880
+ includedTools: ["write_file", "edit_file"],
2792
2881
  inputPrice: 2.5,
2793
2882
  // This is the pricing for prompts above 200k tokens.
2794
2883
  outputPrice: 15,
@@ -2817,7 +2906,9 @@ var geminiModels = {
2817
2906
  contextWindow: 1048576,
2818
2907
  supportsImages: true,
2819
2908
  supportsNativeTools: true,
2909
+ defaultToolProtocol: "native",
2820
2910
  supportsPromptCache: true,
2911
+ includedTools: ["write_file", "edit_file"],
2821
2912
  inputPrice: 0.3,
2822
2913
  outputPrice: 2.5,
2823
2914
  cacheReadsPrice: 0.075,
@@ -2830,7 +2921,9 @@ var geminiModels = {
2830
2921
  contextWindow: 1048576,
2831
2922
  supportsImages: true,
2832
2923
  supportsNativeTools: true,
2924
+ defaultToolProtocol: "native",
2833
2925
  supportsPromptCache: true,
2926
+ includedTools: ["write_file", "edit_file"],
2834
2927
  inputPrice: 0.3,
2835
2928
  outputPrice: 2.5,
2836
2929
  cacheReadsPrice: 0.075,
@@ -2843,7 +2936,9 @@ var geminiModels = {
2843
2936
  contextWindow: 1048576,
2844
2937
  supportsImages: true,
2845
2938
  supportsNativeTools: true,
2939
+ defaultToolProtocol: "native",
2846
2940
  supportsPromptCache: true,
2941
+ includedTools: ["write_file", "edit_file"],
2847
2942
  inputPrice: 0.3,
2848
2943
  outputPrice: 2.5,
2849
2944
  cacheReadsPrice: 0.075,
@@ -2857,7 +2952,9 @@ var geminiModels = {
2857
2952
  contextWindow: 1048576,
2858
2953
  supportsImages: true,
2859
2954
  supportsNativeTools: true,
2955
+ defaultToolProtocol: "native",
2860
2956
  supportsPromptCache: true,
2957
+ includedTools: ["write_file", "edit_file"],
2861
2958
  inputPrice: 0.1,
2862
2959
  outputPrice: 0.4,
2863
2960
  cacheReadsPrice: 0.025,
@@ -2870,7 +2967,9 @@ var geminiModels = {
2870
2967
  contextWindow: 1048576,
2871
2968
  supportsImages: true,
2872
2969
  supportsNativeTools: true,
2970
+ defaultToolProtocol: "native",
2873
2971
  supportsPromptCache: true,
2972
+ includedTools: ["write_file", "edit_file"],
2874
2973
  inputPrice: 0.1,
2875
2974
  outputPrice: 0.4,
2876
2975
  cacheReadsPrice: 0.025,
@@ -2890,6 +2989,7 @@ var groqModels = {
2890
2989
  supportsImages: false,
2891
2990
  supportsPromptCache: false,
2892
2991
  supportsNativeTools: true,
2992
+ defaultToolProtocol: "native",
2893
2993
  inputPrice: 0.05,
2894
2994
  outputPrice: 0.08,
2895
2995
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2900,6 +3000,7 @@ var groqModels = {
2900
3000
  supportsImages: false,
2901
3001
  supportsPromptCache: false,
2902
3002
  supportsNativeTools: true,
3003
+ defaultToolProtocol: "native",
2903
3004
  inputPrice: 0.59,
2904
3005
  outputPrice: 0.79,
2905
3006
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2910,6 +3011,7 @@ var groqModels = {
2910
3011
  supportsImages: false,
2911
3012
  supportsPromptCache: false,
2912
3013
  supportsNativeTools: true,
3014
+ defaultToolProtocol: "native",
2913
3015
  inputPrice: 0.11,
2914
3016
  outputPrice: 0.34,
2915
3017
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2947,6 +3049,7 @@ var groqModels = {
2947
3049
  supportsImages: false,
2948
3050
  supportsPromptCache: false,
2949
3051
  supportsNativeTools: true,
3052
+ defaultToolProtocol: "native",
2950
3053
  inputPrice: 0.29,
2951
3054
  outputPrice: 0.59,
2952
3055
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2977,6 +3080,7 @@ var groqModels = {
2977
3080
  supportsImages: false,
2978
3081
  supportsPromptCache: true,
2979
3082
  supportsNativeTools: true,
3083
+ defaultToolProtocol: "native",
2980
3084
  inputPrice: 0.6,
2981
3085
  outputPrice: 2.5,
2982
3086
  cacheReadsPrice: 0.15,
@@ -2988,6 +3092,7 @@ var groqModels = {
2988
3092
  supportsImages: false,
2989
3093
  supportsPromptCache: false,
2990
3094
  supportsNativeTools: true,
3095
+ defaultToolProtocol: "native",
2991
3096
  inputPrice: 0.15,
2992
3097
  outputPrice: 0.75,
2993
3098
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2998,6 +3103,7 @@ var groqModels = {
2998
3103
  supportsImages: false,
2999
3104
  supportsPromptCache: false,
3000
3105
  supportsNativeTools: true,
3106
+ defaultToolProtocol: "native",
3001
3107
  inputPrice: 0.1,
3002
3108
  outputPrice: 0.5,
3003
3109
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -3061,6 +3167,7 @@ var litellmDefaultModelInfo = {
3061
3167
  supportsImages: true,
3062
3168
  supportsPromptCache: true,
3063
3169
  supportsNativeTools: true,
3170
+ defaultToolProtocol: "native",
3064
3171
  inputPrice: 3,
3065
3172
  outputPrice: 15,
3066
3173
  cacheWritesPrice: 3.75,
@@ -3091,6 +3198,7 @@ var mistralModels = {
3091
3198
  supportsImages: true,
3092
3199
  supportsPromptCache: false,
3093
3200
  supportsNativeTools: true,
3201
+ defaultToolProtocol: "native",
3094
3202
  inputPrice: 2,
3095
3203
  outputPrice: 5
3096
3204
  },
@@ -3100,6 +3208,7 @@ var mistralModels = {
3100
3208
  supportsImages: true,
3101
3209
  supportsPromptCache: false,
3102
3210
  supportsNativeTools: true,
3211
+ defaultToolProtocol: "native",
3103
3212
  inputPrice: 0.4,
3104
3213
  outputPrice: 2
3105
3214
  },
@@ -3109,6 +3218,7 @@ var mistralModels = {
3109
3218
  supportsImages: true,
3110
3219
  supportsPromptCache: false,
3111
3220
  supportsNativeTools: true,
3221
+ defaultToolProtocol: "native",
3112
3222
  inputPrice: 0.4,
3113
3223
  outputPrice: 2
3114
3224
  },
@@ -3118,6 +3228,7 @@ var mistralModels = {
3118
3228
  supportsImages: false,
3119
3229
  supportsPromptCache: false,
3120
3230
  supportsNativeTools: true,
3231
+ defaultToolProtocol: "native",
3121
3232
  inputPrice: 0.3,
3122
3233
  outputPrice: 0.9
3123
3234
  },
@@ -3127,6 +3238,7 @@ var mistralModels = {
3127
3238
  supportsImages: false,
3128
3239
  supportsPromptCache: false,
3129
3240
  supportsNativeTools: true,
3241
+ defaultToolProtocol: "native",
3130
3242
  inputPrice: 2,
3131
3243
  outputPrice: 6
3132
3244
  },
@@ -3136,6 +3248,7 @@ var mistralModels = {
3136
3248
  supportsImages: false,
3137
3249
  supportsPromptCache: false,
3138
3250
  supportsNativeTools: true,
3251
+ defaultToolProtocol: "native",
3139
3252
  inputPrice: 0.1,
3140
3253
  outputPrice: 0.1
3141
3254
  },
@@ -3145,6 +3258,7 @@ var mistralModels = {
3145
3258
  supportsImages: false,
3146
3259
  supportsPromptCache: false,
3147
3260
  supportsNativeTools: true,
3261
+ defaultToolProtocol: "native",
3148
3262
  inputPrice: 0.04,
3149
3263
  outputPrice: 0.04
3150
3264
  },
@@ -3154,6 +3268,7 @@ var mistralModels = {
3154
3268
  supportsImages: false,
3155
3269
  supportsPromptCache: false,
3156
3270
  supportsNativeTools: true,
3271
+ defaultToolProtocol: "native",
3157
3272
  inputPrice: 0.2,
3158
3273
  outputPrice: 0.6
3159
3274
  },
@@ -3163,6 +3278,7 @@ var mistralModels = {
3163
3278
  supportsImages: true,
3164
3279
  supportsPromptCache: false,
3165
3280
  supportsNativeTools: true,
3281
+ defaultToolProtocol: "native",
3166
3282
  inputPrice: 2,
3167
3283
  outputPrice: 6
3168
3284
  }
@@ -3178,6 +3294,7 @@ var moonshotModels = {
3178
3294
  supportsImages: false,
3179
3295
  supportsPromptCache: true,
3180
3296
  supportsNativeTools: true,
3297
+ defaultToolProtocol: "native",
3181
3298
  inputPrice: 0.6,
3182
3299
  // $0.60 per million tokens (cache miss)
3183
3300
  outputPrice: 2.5,
@@ -3194,6 +3311,7 @@ var moonshotModels = {
3194
3311
  supportsImages: false,
3195
3312
  supportsPromptCache: true,
3196
3313
  supportsNativeTools: true,
3314
+ defaultToolProtocol: "native",
3197
3315
  inputPrice: 0.6,
3198
3316
  outputPrice: 2.5,
3199
3317
  cacheReadsPrice: 0.15,
@@ -3205,6 +3323,7 @@ var moonshotModels = {
3205
3323
  supportsImages: false,
3206
3324
  supportsPromptCache: true,
3207
3325
  supportsNativeTools: true,
3326
+ defaultToolProtocol: "native",
3208
3327
  inputPrice: 2.4,
3209
3328
  // $2.40 per million tokens (cache miss)
3210
3329
  outputPrice: 10,
@@ -3224,6 +3343,7 @@ var moonshotModels = {
3224
3343
  // Text-only (no image/vision support)
3225
3344
  supportsPromptCache: true,
3226
3345
  supportsNativeTools: true,
3346
+ defaultToolProtocol: "native",
3227
3347
  inputPrice: 0.6,
3228
3348
  // $0.60 per million tokens (cache miss)
3229
3349
  outputPrice: 2.5,
@@ -3263,6 +3383,7 @@ var openAiNativeModels = {
3263
3383
  maxTokens: 128e3,
3264
3384
  contextWindow: 4e5,
3265
3385
  supportsNativeTools: true,
3386
+ defaultToolProtocol: "native",
3266
3387
  includedTools: ["apply_patch"],
3267
3388
  excludedTools: ["apply_diff", "write_to_file"],
3268
3389
  supportsImages: true,
@@ -3281,6 +3402,7 @@ var openAiNativeModels = {
3281
3402
  maxTokens: 128e3,
3282
3403
  contextWindow: 4e5,
3283
3404
  supportsNativeTools: true,
3405
+ defaultToolProtocol: "native",
3284
3406
  includedTools: ["apply_patch"],
3285
3407
  excludedTools: ["apply_diff", "write_to_file"],
3286
3408
  supportsImages: true,
@@ -3303,6 +3425,7 @@ var openAiNativeModels = {
3303
3425
  maxTokens: 16384,
3304
3426
  contextWindow: 128e3,
3305
3427
  supportsNativeTools: true,
3428
+ defaultToolProtocol: "native",
3306
3429
  includedTools: ["apply_patch"],
3307
3430
  excludedTools: ["apply_diff", "write_to_file"],
3308
3431
  supportsImages: true,
@@ -3316,6 +3439,7 @@ var openAiNativeModels = {
3316
3439
  maxTokens: 128e3,
3317
3440
  contextWindow: 4e5,
3318
3441
  supportsNativeTools: true,
3442
+ defaultToolProtocol: "native",
3319
3443
  includedTools: ["apply_patch"],
3320
3444
  excludedTools: ["apply_diff", "write_to_file"],
3321
3445
  supportsImages: true,
@@ -3338,6 +3462,7 @@ var openAiNativeModels = {
3338
3462
  maxTokens: 128e3,
3339
3463
  contextWindow: 4e5,
3340
3464
  supportsNativeTools: true,
3465
+ defaultToolProtocol: "native",
3341
3466
  includedTools: ["apply_patch"],
3342
3467
  excludedTools: ["apply_diff", "write_to_file"],
3343
3468
  supportsImages: true,
@@ -3356,6 +3481,7 @@ var openAiNativeModels = {
3356
3481
  maxTokens: 128e3,
3357
3482
  contextWindow: 4e5,
3358
3483
  supportsNativeTools: true,
3484
+ defaultToolProtocol: "native",
3359
3485
  includedTools: ["apply_patch"],
3360
3486
  excludedTools: ["apply_diff", "write_to_file"],
3361
3487
  supportsImages: true,
@@ -3373,6 +3499,7 @@ var openAiNativeModels = {
3373
3499
  maxTokens: 128e3,
3374
3500
  contextWindow: 4e5,
3375
3501
  supportsNativeTools: true,
3502
+ defaultToolProtocol: "native",
3376
3503
  includedTools: ["apply_patch"],
3377
3504
  excludedTools: ["apply_diff", "write_to_file"],
3378
3505
  supportsImages: true,
@@ -3394,6 +3521,7 @@ var openAiNativeModels = {
3394
3521
  maxTokens: 128e3,
3395
3522
  contextWindow: 4e5,
3396
3523
  supportsNativeTools: true,
3524
+ defaultToolProtocol: "native",
3397
3525
  includedTools: ["apply_patch"],
3398
3526
  excludedTools: ["apply_diff", "write_to_file"],
3399
3527
  supportsImages: true,
@@ -3415,6 +3543,7 @@ var openAiNativeModels = {
3415
3543
  maxTokens: 128e3,
3416
3544
  contextWindow: 4e5,
3417
3545
  supportsNativeTools: true,
3546
+ defaultToolProtocol: "native",
3418
3547
  includedTools: ["apply_patch"],
3419
3548
  excludedTools: ["apply_diff", "write_to_file"],
3420
3549
  supportsImages: true,
@@ -3432,6 +3561,7 @@ var openAiNativeModels = {
3432
3561
  maxTokens: 128e3,
3433
3562
  contextWindow: 4e5,
3434
3563
  supportsNativeTools: true,
3564
+ defaultToolProtocol: "native",
3435
3565
  includedTools: ["apply_patch"],
3436
3566
  excludedTools: ["apply_diff", "write_to_file"],
3437
3567
  supportsImages: true,
@@ -3450,6 +3580,7 @@ var openAiNativeModels = {
3450
3580
  maxTokens: 128e3,
3451
3581
  contextWindow: 4e5,
3452
3582
  supportsNativeTools: true,
3583
+ defaultToolProtocol: "native",
3453
3584
  includedTools: ["apply_patch"],
3454
3585
  excludedTools: ["apply_diff", "write_to_file"],
3455
3586
  supportsImages: true,
@@ -3463,6 +3594,7 @@ var openAiNativeModels = {
3463
3594
  maxTokens: 32768,
3464
3595
  contextWindow: 1047576,
3465
3596
  supportsNativeTools: true,
3597
+ defaultToolProtocol: "native",
3466
3598
  includedTools: ["apply_patch"],
3467
3599
  excludedTools: ["apply_diff", "write_to_file"],
3468
3600
  supportsImages: true,
@@ -3479,6 +3611,7 @@ var openAiNativeModels = {
3479
3611
  maxTokens: 32768,
3480
3612
  contextWindow: 1047576,
3481
3613
  supportsNativeTools: true,
3614
+ defaultToolProtocol: "native",
3482
3615
  includedTools: ["apply_patch"],
3483
3616
  excludedTools: ["apply_diff", "write_to_file"],
3484
3617
  supportsImages: true,
@@ -3495,6 +3628,7 @@ var openAiNativeModels = {
3495
3628
  maxTokens: 32768,
3496
3629
  contextWindow: 1047576,
3497
3630
  supportsNativeTools: true,
3631
+ defaultToolProtocol: "native",
3498
3632
  includedTools: ["apply_patch"],
3499
3633
  excludedTools: ["apply_diff", "write_to_file"],
3500
3634
  supportsImages: true,
@@ -3511,6 +3645,7 @@ var openAiNativeModels = {
3511
3645
  maxTokens: 1e5,
3512
3646
  contextWindow: 2e5,
3513
3647
  supportsNativeTools: true,
3648
+ defaultToolProtocol: "native",
3514
3649
  supportsImages: true,
3515
3650
  supportsPromptCache: true,
3516
3651
  inputPrice: 2,
@@ -3528,6 +3663,7 @@ var openAiNativeModels = {
3528
3663
  maxTokens: 1e5,
3529
3664
  contextWindow: 2e5,
3530
3665
  supportsNativeTools: true,
3666
+ defaultToolProtocol: "native",
3531
3667
  supportsImages: true,
3532
3668
  supportsPromptCache: true,
3533
3669
  inputPrice: 2,
@@ -3540,6 +3676,7 @@ var openAiNativeModels = {
3540
3676
  maxTokens: 1e5,
3541
3677
  contextWindow: 2e5,
3542
3678
  supportsNativeTools: true,
3679
+ defaultToolProtocol: "native",
3543
3680
  supportsImages: true,
3544
3681
  supportsPromptCache: true,
3545
3682
  inputPrice: 2,
@@ -3552,6 +3689,7 @@ var openAiNativeModels = {
3552
3689
  maxTokens: 1e5,
3553
3690
  contextWindow: 2e5,
3554
3691
  supportsNativeTools: true,
3692
+ defaultToolProtocol: "native",
3555
3693
  supportsImages: true,
3556
3694
  supportsPromptCache: true,
3557
3695
  inputPrice: 1.1,
@@ -3569,6 +3707,7 @@ var openAiNativeModels = {
3569
3707
  maxTokens: 1e5,
3570
3708
  contextWindow: 2e5,
3571
3709
  supportsNativeTools: true,
3710
+ defaultToolProtocol: "native",
3572
3711
  supportsImages: true,
3573
3712
  supportsPromptCache: true,
3574
3713
  inputPrice: 1.1,
@@ -3581,6 +3720,7 @@ var openAiNativeModels = {
3581
3720
  maxTokens: 1e5,
3582
3721
  contextWindow: 2e5,
3583
3722
  supportsNativeTools: true,
3723
+ defaultToolProtocol: "native",
3584
3724
  supportsImages: true,
3585
3725
  supportsPromptCache: true,
3586
3726
  inputPrice: 1.1,
@@ -3593,6 +3733,7 @@ var openAiNativeModels = {
3593
3733
  maxTokens: 1e5,
3594
3734
  contextWindow: 2e5,
3595
3735
  supportsNativeTools: true,
3736
+ defaultToolProtocol: "native",
3596
3737
  supportsImages: false,
3597
3738
  supportsPromptCache: true,
3598
3739
  inputPrice: 1.1,
@@ -3606,6 +3747,7 @@ var openAiNativeModels = {
3606
3747
  maxTokens: 1e5,
3607
3748
  contextWindow: 2e5,
3608
3749
  supportsNativeTools: true,
3750
+ defaultToolProtocol: "native",
3609
3751
  supportsImages: false,
3610
3752
  supportsPromptCache: true,
3611
3753
  inputPrice: 1.1,
@@ -3618,6 +3760,7 @@ var openAiNativeModels = {
3618
3760
  maxTokens: 1e5,
3619
3761
  contextWindow: 2e5,
3620
3762
  supportsNativeTools: true,
3763
+ defaultToolProtocol: "native",
3621
3764
  supportsImages: false,
3622
3765
  supportsPromptCache: true,
3623
3766
  inputPrice: 1.1,
@@ -3630,6 +3773,7 @@ var openAiNativeModels = {
3630
3773
  maxTokens: 1e5,
3631
3774
  contextWindow: 2e5,
3632
3775
  supportsNativeTools: true,
3776
+ defaultToolProtocol: "native",
3633
3777
  supportsImages: true,
3634
3778
  supportsPromptCache: true,
3635
3779
  inputPrice: 15,
@@ -3641,6 +3785,7 @@ var openAiNativeModels = {
3641
3785
  maxTokens: 32768,
3642
3786
  contextWindow: 128e3,
3643
3787
  supportsNativeTools: true,
3788
+ defaultToolProtocol: "native",
3644
3789
  supportsImages: true,
3645
3790
  supportsPromptCache: true,
3646
3791
  inputPrice: 15,
@@ -3652,6 +3797,7 @@ var openAiNativeModels = {
3652
3797
  maxTokens: 65536,
3653
3798
  contextWindow: 128e3,
3654
3799
  supportsNativeTools: true,
3800
+ defaultToolProtocol: "native",
3655
3801
  supportsImages: true,
3656
3802
  supportsPromptCache: true,
3657
3803
  inputPrice: 1.1,
@@ -3663,6 +3809,7 @@ var openAiNativeModels = {
3663
3809
  maxTokens: 16384,
3664
3810
  contextWindow: 128e3,
3665
3811
  supportsNativeTools: true,
3812
+ defaultToolProtocol: "native",
3666
3813
  supportsImages: true,
3667
3814
  supportsPromptCache: true,
3668
3815
  inputPrice: 2.5,
@@ -3677,6 +3824,7 @@ var openAiNativeModels = {
3677
3824
  maxTokens: 16384,
3678
3825
  contextWindow: 128e3,
3679
3826
  supportsNativeTools: true,
3827
+ defaultToolProtocol: "native",
3680
3828
  supportsImages: true,
3681
3829
  supportsPromptCache: true,
3682
3830
  inputPrice: 0.15,
@@ -3691,6 +3839,7 @@ var openAiNativeModels = {
3691
3839
  maxTokens: 16384,
3692
3840
  contextWindow: 2e5,
3693
3841
  supportsNativeTools: true,
3842
+ defaultToolProtocol: "native",
3694
3843
  supportsImages: false,
3695
3844
  supportsPromptCache: false,
3696
3845
  inputPrice: 1.5,
@@ -3704,6 +3853,7 @@ var openAiNativeModels = {
3704
3853
  maxTokens: 128e3,
3705
3854
  contextWindow: 4e5,
3706
3855
  supportsNativeTools: true,
3856
+ defaultToolProtocol: "native",
3707
3857
  includedTools: ["apply_patch"],
3708
3858
  excludedTools: ["apply_diff", "write_to_file"],
3709
3859
  supportsImages: true,
@@ -3725,6 +3875,7 @@ var openAiNativeModels = {
3725
3875
  maxTokens: 128e3,
3726
3876
  contextWindow: 4e5,
3727
3877
  supportsNativeTools: true,
3878
+ defaultToolProtocol: "native",
3728
3879
  includedTools: ["apply_patch"],
3729
3880
  excludedTools: ["apply_diff", "write_to_file"],
3730
3881
  supportsImages: true,
@@ -3746,6 +3897,7 @@ var openAiNativeModels = {
3746
3897
  maxTokens: 128e3,
3747
3898
  contextWindow: 4e5,
3748
3899
  supportsNativeTools: true,
3900
+ defaultToolProtocol: "native",
3749
3901
  includedTools: ["apply_patch"],
3750
3902
  excludedTools: ["apply_diff", "write_to_file"],
3751
3903
  supportsImages: true,
@@ -3768,7 +3920,8 @@ var openAiModelInfoSaneDefaults = {
3768
3920
  supportsPromptCache: false,
3769
3921
  inputPrice: 0,
3770
3922
  outputPrice: 0,
3771
- supportsNativeTools: true
3923
+ supportsNativeTools: true,
3924
+ defaultToolProtocol: "native"
3772
3925
  };
3773
3926
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3774
3927
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3881,6 +4034,8 @@ var requestyDefaultModelInfo = {
3881
4034
  contextWindow: 2e5,
3882
4035
  supportsImages: true,
3883
4036
  supportsPromptCache: true,
4037
+ supportsNativeTools: true,
4038
+ defaultToolProtocol: "native",
3884
4039
  inputPrice: 3,
3885
4040
  outputPrice: 15,
3886
4041
  cacheWritesPrice: 3.75,
@@ -3937,6 +4092,7 @@ var sambaNovaModels = {
3937
4092
  supportsImages: false,
3938
4093
  supportsPromptCache: false,
3939
4094
  supportsNativeTools: true,
4095
+ defaultToolProtocol: "native",
3940
4096
  inputPrice: 0.1,
3941
4097
  outputPrice: 0.2,
3942
4098
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3947,6 +4103,7 @@ var sambaNovaModels = {
3947
4103
  supportsImages: false,
3948
4104
  supportsPromptCache: false,
3949
4105
  supportsNativeTools: true,
4106
+ defaultToolProtocol: "native",
3950
4107
  inputPrice: 0.6,
3951
4108
  outputPrice: 1.2,
3952
4109
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3958,6 +4115,7 @@ var sambaNovaModels = {
3958
4115
  supportsPromptCache: false,
3959
4116
  supportsReasoningBudget: true,
3960
4117
  supportsNativeTools: true,
4118
+ defaultToolProtocol: "native",
3961
4119
  inputPrice: 5,
3962
4120
  outputPrice: 7,
3963
4121
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3968,6 +4126,7 @@ var sambaNovaModels = {
3968
4126
  supportsImages: false,
3969
4127
  supportsPromptCache: false,
3970
4128
  supportsNativeTools: true,
4129
+ defaultToolProtocol: "native",
3971
4130
  inputPrice: 3,
3972
4131
  outputPrice: 4.5,
3973
4132
  description: "DeepSeek V3 model with 32K context window."
@@ -3978,6 +4137,7 @@ var sambaNovaModels = {
3978
4137
  supportsImages: false,
3979
4138
  supportsPromptCache: false,
3980
4139
  supportsNativeTools: true,
4140
+ defaultToolProtocol: "native",
3981
4141
  inputPrice: 3,
3982
4142
  outputPrice: 4.5,
3983
4143
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3997,6 +4157,7 @@ var sambaNovaModels = {
3997
4157
  supportsImages: true,
3998
4158
  supportsPromptCache: false,
3999
4159
  supportsNativeTools: true,
4160
+ defaultToolProtocol: "native",
4000
4161
  inputPrice: 0.63,
4001
4162
  outputPrice: 1.8,
4002
4163
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -4016,6 +4177,7 @@ var sambaNovaModels = {
4016
4177
  supportsImages: false,
4017
4178
  supportsPromptCache: false,
4018
4179
  supportsNativeTools: true,
4180
+ defaultToolProtocol: "native",
4019
4181
  inputPrice: 0.4,
4020
4182
  outputPrice: 0.8,
4021
4183
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -4026,6 +4188,7 @@ var sambaNovaModels = {
4026
4188
  supportsImages: false,
4027
4189
  supportsPromptCache: false,
4028
4190
  supportsNativeTools: true,
4191
+ defaultToolProtocol: "native",
4029
4192
  inputPrice: 0.22,
4030
4193
  outputPrice: 0.59,
4031
4194
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -4054,9 +4217,11 @@ var vertexModels = {
4054
4217
  contextWindow: 1048576,
4055
4218
  supportsImages: true,
4056
4219
  supportsNativeTools: true,
4220
+ defaultToolProtocol: "native",
4057
4221
  supportsPromptCache: true,
4058
4222
  supportsReasoningEffort: ["low", "high"],
4059
4223
  reasoningEffort: "low",
4224
+ includedTools: ["write_file", "edit_file"],
4060
4225
  supportsTemperature: true,
4061
4226
  defaultTemperature: 1,
4062
4227
  inputPrice: 4,
@@ -4074,12 +4239,31 @@ var vertexModels = {
4074
4239
  }
4075
4240
  ]
4076
4241
  },
4242
+ "gemini-3-flash-preview": {
4243
+ maxTokens: 65536,
4244
+ contextWindow: 1048576,
4245
+ supportsImages: true,
4246
+ supportsNativeTools: true,
4247
+ defaultToolProtocol: "native",
4248
+ supportsPromptCache: true,
4249
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
4250
+ reasoningEffort: "medium",
4251
+ includedTools: ["write_file", "edit_file"],
4252
+ supportsTemperature: true,
4253
+ defaultTemperature: 1,
4254
+ inputPrice: 0.3,
4255
+ outputPrice: 2.5,
4256
+ cacheReadsPrice: 0.075,
4257
+ cacheWritesPrice: 1
4258
+ },
4077
4259
  "gemini-2.5-flash-preview-05-20:thinking": {
4078
4260
  maxTokens: 65535,
4079
4261
  contextWindow: 1048576,
4080
4262
  supportsImages: true,
4081
4263
  supportsNativeTools: true,
4264
+ defaultToolProtocol: "native",
4082
4265
  supportsPromptCache: true,
4266
+ includedTools: ["write_file", "edit_file"],
4083
4267
  inputPrice: 0.15,
4084
4268
  outputPrice: 3.5,
4085
4269
  maxThinkingTokens: 24576,
@@ -4091,7 +4275,9 @@ var vertexModels = {
4091
4275
  contextWindow: 1048576,
4092
4276
  supportsImages: true,
4093
4277
  supportsNativeTools: true,
4278
+ defaultToolProtocol: "native",
4094
4279
  supportsPromptCache: true,
4280
+ includedTools: ["write_file", "edit_file"],
4095
4281
  inputPrice: 0.15,
4096
4282
  outputPrice: 0.6
4097
4283
  },
@@ -4100,7 +4286,9 @@ var vertexModels = {
4100
4286
  contextWindow: 1048576,
4101
4287
  supportsImages: true,
4102
4288
  supportsNativeTools: true,
4289
+ defaultToolProtocol: "native",
4103
4290
  supportsPromptCache: true,
4291
+ includedTools: ["write_file", "edit_file"],
4104
4292
  inputPrice: 0.3,
4105
4293
  outputPrice: 2.5,
4106
4294
  cacheReadsPrice: 0.075,
@@ -4113,7 +4301,9 @@ var vertexModels = {
4113
4301
  contextWindow: 1048576,
4114
4302
  supportsImages: true,
4115
4303
  supportsNativeTools: true,
4304
+ defaultToolProtocol: "native",
4116
4305
  supportsPromptCache: false,
4306
+ includedTools: ["write_file", "edit_file"],
4117
4307
  inputPrice: 0.15,
4118
4308
  outputPrice: 3.5,
4119
4309
  maxThinkingTokens: 24576,
@@ -4125,7 +4315,9 @@ var vertexModels = {
4125
4315
  contextWindow: 1048576,
4126
4316
  supportsImages: true,
4127
4317
  supportsNativeTools: true,
4318
+ defaultToolProtocol: "native",
4128
4319
  supportsPromptCache: false,
4320
+ includedTools: ["write_file", "edit_file"],
4129
4321
  inputPrice: 0.15,
4130
4322
  outputPrice: 0.6
4131
4323
  },
@@ -4134,7 +4326,9 @@ var vertexModels = {
4134
4326
  contextWindow: 1048576,
4135
4327
  supportsImages: true,
4136
4328
  supportsNativeTools: true,
4329
+ defaultToolProtocol: "native",
4137
4330
  supportsPromptCache: true,
4331
+ includedTools: ["write_file", "edit_file"],
4138
4332
  inputPrice: 2.5,
4139
4333
  outputPrice: 15
4140
4334
  },
@@ -4143,7 +4337,9 @@ var vertexModels = {
4143
4337
  contextWindow: 1048576,
4144
4338
  supportsImages: true,
4145
4339
  supportsNativeTools: true,
4340
+ defaultToolProtocol: "native",
4146
4341
  supportsPromptCache: true,
4342
+ includedTools: ["write_file", "edit_file"],
4147
4343
  inputPrice: 2.5,
4148
4344
  outputPrice: 15
4149
4345
  },
@@ -4152,7 +4348,9 @@ var vertexModels = {
4152
4348
  contextWindow: 1048576,
4153
4349
  supportsImages: true,
4154
4350
  supportsNativeTools: true,
4351
+ defaultToolProtocol: "native",
4155
4352
  supportsPromptCache: true,
4353
+ includedTools: ["write_file", "edit_file"],
4156
4354
  inputPrice: 2.5,
4157
4355
  outputPrice: 15,
4158
4356
  maxThinkingTokens: 32768,
@@ -4163,7 +4361,9 @@ var vertexModels = {
4163
4361
  contextWindow: 1048576,
4164
4362
  supportsImages: true,
4165
4363
  supportsNativeTools: true,
4364
+ defaultToolProtocol: "native",
4166
4365
  supportsPromptCache: true,
4366
+ includedTools: ["write_file", "edit_file"],
4167
4367
  inputPrice: 2.5,
4168
4368
  outputPrice: 15,
4169
4369
  maxThinkingTokens: 32768,
@@ -4189,7 +4389,9 @@ var vertexModels = {
4189
4389
  contextWindow: 1048576,
4190
4390
  supportsImages: true,
4191
4391
  supportsNativeTools: true,
4392
+ defaultToolProtocol: "native",
4192
4393
  supportsPromptCache: false,
4394
+ includedTools: ["write_file", "edit_file"],
4193
4395
  inputPrice: 0,
4194
4396
  outputPrice: 0
4195
4397
  },
@@ -4198,7 +4400,9 @@ var vertexModels = {
4198
4400
  contextWindow: 2097152,
4199
4401
  supportsImages: true,
4200
4402
  supportsNativeTools: true,
4403
+ defaultToolProtocol: "native",
4201
4404
  supportsPromptCache: false,
4405
+ includedTools: ["write_file", "edit_file"],
4202
4406
  inputPrice: 0,
4203
4407
  outputPrice: 0
4204
4408
  },
@@ -4207,7 +4411,9 @@ var vertexModels = {
4207
4411
  contextWindow: 1048576,
4208
4412
  supportsImages: true,
4209
4413
  supportsNativeTools: true,
4414
+ defaultToolProtocol: "native",
4210
4415
  supportsPromptCache: true,
4416
+ includedTools: ["write_file", "edit_file"],
4211
4417
  inputPrice: 0.15,
4212
4418
  outputPrice: 0.6
4213
4419
  },
@@ -4216,7 +4422,9 @@ var vertexModels = {
4216
4422
  contextWindow: 1048576,
4217
4423
  supportsImages: true,
4218
4424
  supportsNativeTools: true,
4425
+ defaultToolProtocol: "native",
4219
4426
  supportsPromptCache: false,
4427
+ includedTools: ["write_file", "edit_file"],
4220
4428
  inputPrice: 0.075,
4221
4429
  outputPrice: 0.3
4222
4430
  },
@@ -4225,7 +4433,9 @@ var vertexModels = {
4225
4433
  contextWindow: 32768,
4226
4434
  supportsImages: true,
4227
4435
  supportsNativeTools: true,
4436
+ defaultToolProtocol: "native",
4228
4437
  supportsPromptCache: false,
4438
+ includedTools: ["write_file", "edit_file"],
4229
4439
  inputPrice: 0,
4230
4440
  outputPrice: 0
4231
4441
  },
@@ -4234,7 +4444,9 @@ var vertexModels = {
4234
4444
  contextWindow: 1048576,
4235
4445
  supportsImages: true,
4236
4446
  supportsNativeTools: true,
4447
+ defaultToolProtocol: "native",
4237
4448
  supportsPromptCache: true,
4449
+ includedTools: ["write_file", "edit_file"],
4238
4450
  inputPrice: 0.075,
4239
4451
  outputPrice: 0.3
4240
4452
  },
@@ -4243,7 +4455,9 @@ var vertexModels = {
4243
4455
  contextWindow: 2097152,
4244
4456
  supportsImages: true,
4245
4457
  supportsNativeTools: true,
4458
+ defaultToolProtocol: "native",
4246
4459
  supportsPromptCache: false,
4460
+ includedTools: ["write_file", "edit_file"],
4247
4461
  inputPrice: 1.25,
4248
4462
  outputPrice: 5
4249
4463
  },
@@ -4389,7 +4603,9 @@ var vertexModels = {
4389
4603
  contextWindow: 1048576,
4390
4604
  supportsImages: true,
4391
4605
  supportsNativeTools: true,
4606
+ defaultToolProtocol: "native",
4392
4607
  supportsPromptCache: true,
4608
+ includedTools: ["write_file", "edit_file"],
4393
4609
  inputPrice: 0.1,
4394
4610
  outputPrice: 0.4,
4395
4611
  cacheReadsPrice: 0.025,
@@ -4691,6 +4907,7 @@ var xaiModels = {
4691
4907
  supportsImages: true,
4692
4908
  supportsPromptCache: true,
4693
4909
  supportsNativeTools: true,
4910
+ defaultToolProtocol: "native",
4694
4911
  inputPrice: 0.2,
4695
4912
  outputPrice: 1.5,
4696
4913
  cacheWritesPrice: 0.02,
@@ -4705,6 +4922,7 @@ var xaiModels = {
4705
4922
  supportsImages: true,
4706
4923
  supportsPromptCache: true,
4707
4924
  supportsNativeTools: true,
4925
+ defaultToolProtocol: "native",
4708
4926
  inputPrice: 0.2,
4709
4927
  outputPrice: 0.5,
4710
4928
  cacheWritesPrice: 0.05,
@@ -4719,6 +4937,7 @@ var xaiModels = {
4719
4937
  supportsImages: true,
4720
4938
  supportsPromptCache: true,
4721
4939
  supportsNativeTools: true,
4940
+ defaultToolProtocol: "native",
4722
4941
  inputPrice: 0.2,
4723
4942
  outputPrice: 0.5,
4724
4943
  cacheWritesPrice: 0.05,
@@ -4733,6 +4952,7 @@ var xaiModels = {
4733
4952
  supportsImages: true,
4734
4953
  supportsPromptCache: true,
4735
4954
  supportsNativeTools: true,
4955
+ defaultToolProtocol: "native",
4736
4956
  inputPrice: 0.2,
4737
4957
  outputPrice: 0.5,
4738
4958
  cacheWritesPrice: 0.05,
@@ -4747,6 +4967,7 @@ var xaiModels = {
4747
4967
  supportsImages: true,
4748
4968
  supportsPromptCache: true,
4749
4969
  supportsNativeTools: true,
4970
+ defaultToolProtocol: "native",
4750
4971
  inputPrice: 0.2,
4751
4972
  outputPrice: 0.5,
4752
4973
  cacheWritesPrice: 0.05,
@@ -4761,6 +4982,7 @@ var xaiModels = {
4761
4982
  supportsImages: true,
4762
4983
  supportsPromptCache: true,
4763
4984
  supportsNativeTools: true,
4985
+ defaultToolProtocol: "native",
4764
4986
  inputPrice: 3,
4765
4987
  outputPrice: 15,
4766
4988
  cacheWritesPrice: 0.75,
@@ -4775,6 +4997,7 @@ var xaiModels = {
4775
4997
  supportsImages: true,
4776
4998
  supportsPromptCache: true,
4777
4999
  supportsNativeTools: true,
5000
+ defaultToolProtocol: "native",
4778
5001
  inputPrice: 0.3,
4779
5002
  outputPrice: 0.5,
4780
5003
  cacheWritesPrice: 0.07,
@@ -4791,6 +5014,7 @@ var xaiModels = {
4791
5014
  supportsImages: true,
4792
5015
  supportsPromptCache: true,
4793
5016
  supportsNativeTools: true,
5017
+ defaultToolProtocol: "native",
4794
5018
  inputPrice: 3,
4795
5019
  outputPrice: 15,
4796
5020
  cacheWritesPrice: 0.75,
@@ -4904,6 +5128,7 @@ var internationalZAiModels = {
4904
5128
  supportsImages: false,
4905
5129
  supportsPromptCache: true,
4906
5130
  supportsNativeTools: true,
5131
+ defaultToolProtocol: "native",
4907
5132
  inputPrice: 0.6,
4908
5133
  outputPrice: 2.2,
4909
5134
  cacheWritesPrice: 0,
@@ -4916,6 +5141,7 @@ var internationalZAiModels = {
4916
5141
  supportsImages: false,
4917
5142
  supportsPromptCache: true,
4918
5143
  supportsNativeTools: true,
5144
+ defaultToolProtocol: "native",
4919
5145
  inputPrice: 0.2,
4920
5146
  outputPrice: 1.1,
4921
5147
  cacheWritesPrice: 0,
@@ -4928,6 +5154,7 @@ var internationalZAiModels = {
4928
5154
  supportsImages: false,
4929
5155
  supportsPromptCache: true,
4930
5156
  supportsNativeTools: true,
5157
+ defaultToolProtocol: "native",
4931
5158
  inputPrice: 2.2,
4932
5159
  outputPrice: 8.9,
4933
5160
  cacheWritesPrice: 0,
@@ -4940,6 +5167,7 @@ var internationalZAiModels = {
4940
5167
  supportsImages: false,
4941
5168
  supportsPromptCache: true,
4942
5169
  supportsNativeTools: true,
5170
+ defaultToolProtocol: "native",
4943
5171
  inputPrice: 1.1,
4944
5172
  outputPrice: 4.5,
4945
5173
  cacheWritesPrice: 0,
@@ -4952,6 +5180,7 @@ var internationalZAiModels = {
4952
5180
  supportsImages: false,
4953
5181
  supportsPromptCache: true,
4954
5182
  supportsNativeTools: true,
5183
+ defaultToolProtocol: "native",
4955
5184
  inputPrice: 0,
4956
5185
  outputPrice: 0,
4957
5186
  cacheWritesPrice: 0,
@@ -4964,6 +5193,7 @@ var internationalZAiModels = {
4964
5193
  supportsImages: true,
4965
5194
  supportsPromptCache: true,
4966
5195
  supportsNativeTools: true,
5196
+ defaultToolProtocol: "native",
4967
5197
  inputPrice: 0.6,
4968
5198
  outputPrice: 1.8,
4969
5199
  cacheWritesPrice: 0,
@@ -4976,6 +5206,7 @@ var internationalZAiModels = {
4976
5206
  supportsImages: false,
4977
5207
  supportsPromptCache: true,
4978
5208
  supportsNativeTools: true,
5209
+ defaultToolProtocol: "native",
4979
5210
  inputPrice: 0.6,
4980
5211
  outputPrice: 2.2,
4981
5212
  cacheWritesPrice: 0,
@@ -4988,6 +5219,7 @@ var internationalZAiModels = {
4988
5219
  supportsImages: false,
4989
5220
  supportsPromptCache: false,
4990
5221
  supportsNativeTools: true,
5222
+ defaultToolProtocol: "native",
4991
5223
  inputPrice: 0.1,
4992
5224
  outputPrice: 0.1,
4993
5225
  cacheWritesPrice: 0,
@@ -5003,6 +5235,7 @@ var mainlandZAiModels = {
5003
5235
  supportsImages: false,
5004
5236
  supportsPromptCache: true,
5005
5237
  supportsNativeTools: true,
5238
+ defaultToolProtocol: "native",
5006
5239
  inputPrice: 0.29,
5007
5240
  outputPrice: 1.14,
5008
5241
  cacheWritesPrice: 0,
@@ -5015,6 +5248,7 @@ var mainlandZAiModels = {
5015
5248
  supportsImages: false,
5016
5249
  supportsPromptCache: true,
5017
5250
  supportsNativeTools: true,
5251
+ defaultToolProtocol: "native",
5018
5252
  inputPrice: 0.1,
5019
5253
  outputPrice: 0.6,
5020
5254
  cacheWritesPrice: 0,
@@ -5027,6 +5261,7 @@ var mainlandZAiModels = {
5027
5261
  supportsImages: false,
5028
5262
  supportsPromptCache: true,
5029
5263
  supportsNativeTools: true,
5264
+ defaultToolProtocol: "native",
5030
5265
  inputPrice: 0.29,
5031
5266
  outputPrice: 1.14,
5032
5267
  cacheWritesPrice: 0,
@@ -5039,6 +5274,7 @@ var mainlandZAiModels = {
5039
5274
  supportsImages: false,
5040
5275
  supportsPromptCache: true,
5041
5276
  supportsNativeTools: true,
5277
+ defaultToolProtocol: "native",
5042
5278
  inputPrice: 0.1,
5043
5279
  outputPrice: 0.6,
5044
5280
  cacheWritesPrice: 0,
@@ -5051,6 +5287,7 @@ var mainlandZAiModels = {
5051
5287
  supportsImages: false,
5052
5288
  supportsPromptCache: true,
5053
5289
  supportsNativeTools: true,
5290
+ defaultToolProtocol: "native",
5054
5291
  inputPrice: 0,
5055
5292
  outputPrice: 0,
5056
5293
  cacheWritesPrice: 0,
@@ -5063,6 +5300,7 @@ var mainlandZAiModels = {
5063
5300
  supportsImages: true,
5064
5301
  supportsPromptCache: true,
5065
5302
  supportsNativeTools: true,
5303
+ defaultToolProtocol: "native",
5066
5304
  inputPrice: 0.29,
5067
5305
  outputPrice: 0.93,
5068
5306
  cacheWritesPrice: 0,
@@ -5075,6 +5313,7 @@ var mainlandZAiModels = {
5075
5313
  supportsImages: false,
5076
5314
  supportsPromptCache: true,
5077
5315
  supportsNativeTools: true,
5316
+ defaultToolProtocol: "native",
5078
5317
  inputPrice: 0.29,
5079
5318
  outputPrice: 1.14,
5080
5319
  cacheWritesPrice: 0,
@@ -5128,6 +5367,7 @@ var minimaxModels = {
5128
5367
  supportsImages: false,
5129
5368
  supportsPromptCache: true,
5130
5369
  supportsNativeTools: true,
5370
+ defaultToolProtocol: "native",
5131
5371
  preserveReasoning: true,
5132
5372
  inputPrice: 0.3,
5133
5373
  outputPrice: 1.2,
@@ -5141,6 +5381,7 @@ var minimaxModels = {
5141
5381
  supportsImages: false,
5142
5382
  supportsPromptCache: true,
5143
5383
  supportsNativeTools: true,
5384
+ defaultToolProtocol: "native",
5144
5385
  preserveReasoning: true,
5145
5386
  inputPrice: 0.3,
5146
5387
  outputPrice: 1.2,
@@ -5323,10 +5564,7 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5323
5564
  anthropicBeta1MContext: import_zod8.z.boolean().optional()
5324
5565
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5325
5566
  });
5326
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
5327
- claudeCodePath: import_zod8.z.string().optional(),
5328
- claudeCodeMaxOutputTokens: import_zod8.z.number().int().min(1).max(2e5).optional()
5329
- });
5567
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5330
5568
  var openRouterSchema = baseProviderSettingsSchema.extend({
5331
5569
  openRouterApiKey: import_zod8.z.string().optional(),
5332
5570
  openRouterModelId: import_zod8.z.string().optional(),
@@ -6001,11 +6239,44 @@ function getErrorStatusCode(error) {
6001
6239
  }
6002
6240
  return void 0;
6003
6241
  }
6242
+ function extractMessageFromJsonPayload(message) {
6243
+ const jsonStartIndex = message.indexOf("{");
6244
+ if (jsonStartIndex === -1) {
6245
+ return void 0;
6246
+ }
6247
+ const potentialJson = message.slice(jsonStartIndex);
6248
+ try {
6249
+ const parsed = JSON.parse(potentialJson);
6250
+ if (parsed?.error?.message && typeof parsed.error.message === "string") {
6251
+ return parsed.error.message;
6252
+ }
6253
+ if (parsed?.message && typeof parsed.message === "string") {
6254
+ return parsed.message;
6255
+ }
6256
+ } catch {
6257
+ }
6258
+ return void 0;
6259
+ }
6004
6260
  function getErrorMessage(error) {
6261
+ let message;
6005
6262
  if (isOpenAISdkError(error)) {
6006
- return error.error?.metadata?.raw || error.error?.message || error.message;
6263
+ message = error.error?.metadata?.raw || error.error?.message || error.message;
6264
+ } else if (error instanceof Error) {
6265
+ message = error.message;
6266
+ } else if (typeof error === "object" && error !== null && "message" in error) {
6267
+ const msgValue = error.message;
6268
+ if (typeof msgValue === "string") {
6269
+ message = msgValue;
6270
+ }
6007
6271
  }
6008
- return void 0;
6272
+ if (!message) {
6273
+ return void 0;
6274
+ }
6275
+ const extractedMessage = extractMessageFromJsonPayload(message);
6276
+ if (extractedMessage) {
6277
+ return extractedMessage;
6278
+ }
6279
+ return message;
6009
6280
  }
6010
6281
  function shouldReportApiErrorToTelemetry(errorCode, errorMessage) {
6011
6282
  if (errorCode !== void 0 && EXPECTED_API_ERROR_CODES.has(errorCode)) {
@@ -6537,7 +6808,7 @@ var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6537
6808
  terminalShellIntegrationTimeout: import_zod16.z.number().int().nonnegative().optional()
6538
6809
  })
6539
6810
  );
6540
- var workspaceTaskVisibilitySchema = import_zod16.z.enum(["all", "list-only", "full-lockdown"]);
6811
+ var workspaceTaskVisibilitySchema = import_zod16.z.enum(["all", "list-only", "admins-and-creator", "creator", "full-lockdown"]);
6541
6812
  var organizationCloudSettingsSchema = import_zod16.z.object({
6542
6813
  recordTaskMessages: import_zod16.z.boolean().optional(),
6543
6814
  enableTaskSharing: import_zod16.z.boolean().optional(),
@@ -7120,7 +7391,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7120
7391
  BEDROCK_REGIONS,
7121
7392
  BEDROCK_SERVICE_TIER_MODEL_IDS,
7122
7393
  BEDROCK_SERVICE_TIER_PRICING,
7123
- CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
7124
7394
  CODEBASE_INDEX_DEFAULTS,
7125
7395
  CONSENT_COOKIE_NAME,
7126
7396
  CONTEXT_MANAGEMENT_EVENTS,
@@ -7210,6 +7480,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7210
7480
  chutesModels,
7211
7481
  claudeCodeDefaultModelId,
7212
7482
  claudeCodeModels,
7483
+ claudeCodeReasoningConfig,
7213
7484
  clineAskSchema,
7214
7485
  clineAsks,
7215
7486
  clineMessageSchema,
@@ -7224,7 +7495,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7224
7495
  commandIds,
7225
7496
  contextCondenseSchema,
7226
7497
  contextTruncationSchema,
7227
- convertModelNameForVertex,
7228
7498
  customModePromptsSchema,
7229
7499
  customModesSettingsSchema,
7230
7500
  customProviders,
@@ -7246,6 +7516,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7246
7516
  extensionBridgeEventSchema,
7247
7517
  extensionInstanceSchema,
7248
7518
  extractApiProviderErrorProperties,
7519
+ extractMessageFromJsonPayload,
7249
7520
  fauxProviders,
7250
7521
  featherlessDefaultModelId,
7251
7522
  featherlessModels,
@@ -7255,7 +7526,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7255
7526
  geminiDefaultModelId,
7256
7527
  geminiModels,
7257
7528
  getApiProtocol,
7258
- getClaudeCodeModelId,
7259
7529
  getEffectiveProtocol,
7260
7530
  getErrorMessage,
7261
7531
  getErrorStatusCode,
@@ -7327,6 +7597,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7327
7597
  moonshotDefaultModelId,
7328
7598
  moonshotModels,
7329
7599
  nonBlockingAsks,
7600
+ normalizeClaudeCodeModelId,
7330
7601
  ollamaDefaultModelId,
7331
7602
  ollamaDefaultModelInfo,
7332
7603
  openAiModelInfoSaneDefaults,